From 3e3b676356b5ab65ab0e37f97930b01d6471540e Mon Sep 17 00:00:00 2001 From: Andrew Kelley Date: Sat, 2 Oct 2021 10:45:56 -0700 Subject: [PATCH] Merge branch 'llvm13' (cherry picked from commit dde0adcb363f3a3f306c0fc9eaec511cc3b74965) Signed-off-by: Jan200101 --- build.zig | 39 + ci/drone/drone.yml | 10 +- cmake/Findclang.cmake | 40 +- cmake/Findlld.cmake | 26 +- cmake/Findllvm.cmake | 85 +- lib/include/__clang_cuda_device_functions.h | 276 +- lib/include/__clang_cuda_math.h | 4 +- lib/include/__clang_cuda_runtime_wrapper.h | 7 +- lib/include/__clang_hip_cmath.h | 366 +- lib/include/__clang_hip_libdevice_declares.h | 36 +- lib/include/__clang_hip_math.h | 50 +- lib/include/__clang_hip_runtime_wrapper.h | 87 +- lib/include/altivec.h | 1588 +- lib/include/amxintrin.h | 253 +- lib/include/arm_acle.h | 55 + lib/include/arm_neon.h | 1242 +- lib/include/arm_sve.h | 17687 +- lib/include/avx512fintrin.h | 214 +- lib/include/builtins.h | 16 + lib/include/cuda_wrappers/complex | 8 + lib/include/hexagon_circ_brev_intrinsics.h | 298 + lib/include/hexagon_protos.h | 8450 + lib/include/hexagon_types.h | 2653 + lib/include/hvx_hexagon_protos.h | 4392 + lib/include/immintrin.h | 10 +- lib/include/intrin.h | 52 +- lib/include/keylockerintrin.h | 30 + lib/include/opencl-c-base.h | 147 +- lib/include/opencl-c.h | 860 +- .../__clang_openmp_device_functions.h | 72 +- lib/include/openmp_wrappers/cmath | 54 + lib/include/openmp_wrappers/complex | 6 - lib/include/openmp_wrappers/complex.h | 1 - lib/include/openmp_wrappers/math.h | 10 + lib/include/openmp_wrappers/new | 26 +- lib/include/ppc_wrappers/xmmintrin.h | 9 +- lib/include/riscv_vector.h | 123865 +++++++++++++++ lib/include/uintrintrin.h | 7 + lib/include/vaesintrin.h | 41 +- lib/include/vecintrin.h | 283 +- lib/include/vpclmulqdqintrin.h | 2 + lib/include/wasm_simd128.h | 1045 +- .../include/aarch64-linux-gnu/bits/floatn.h | 97 + .../include/aarch64-linux-musl/bits/posix.h | 2 + .../aarch64_be-linux-gnu/bits/floatn.h | 97 + .../include/arm-linux-gnueabi/bits/floatn.h | 52 - .../include/arm-linux-gnueabihf/bits/floatn.h | 52 - lib/libc/include/arm-linux-musl/bits/posix.h | 2 - .../include/armeb-linux-gnueabi/bits/floatn.h | 52 - .../armeb-linux-gnueabihf/bits/floatn.h | 52 - .../include/csky-linux-gnueabi/bits/floatn.h | 52 - .../csky-linux-gnueabi/bits/struct_rwlock.h | 61 - .../csky-linux-gnueabi/bits/wordsize.h | 21 - .../csky-linux-gnueabihf/bits/floatn.h | 52 - .../csky-linux-gnueabihf/bits/struct_rwlock.h | 61 - .../csky-linux-gnueabihf/bits/wordsize.h | 21 - lib/libc/include/generic-glibc/bits/floatn.h | 71 +- .../generic-glibc/bits/struct_rwlock.h | 44 +- .../include/generic-glibc/bits/wordsize.h | 20 +- lib/libc/include/generic-musl/bits/posix.h | 4 +- lib/libc/include/i386-linux-musl/bits/posix.h | 2 - lib/libc/include/m68k-linux-gnu/bits/a.out.h | 3 + .../include/m68k-linux-gnu/bits/endianness.h | 11 + lib/libc/include/m68k-linux-gnu/bits/fcntl.h | 54 + lib/libc/include/m68k-linux-gnu/bits/fenv.h | 131 + .../m68k-linux-gnu/bits/flt-eval-method.h | 25 + .../bits/fp-logb.h} | 13 +- .../include/m68k-linux-gnu/bits/iscanonical.h | 54 + lib/libc/include/m68k-linux-gnu/bits/link.h | 57 + .../bits/long-double.h} | 12 +- lib/libc/include/m68k-linux-gnu/bits/poll.h | 49 + .../bits/procfs-id.h} | 14 +- lib/libc/include/m68k-linux-gnu/bits/procfs.h | 34 + .../m68k-linux-gnu/bits/pthreadtypes-arch.h | 37 + .../include/m68k-linux-gnu/bits/semaphore.h | 35 + lib/libc/include/m68k-linux-gnu/bits/setjmp.h | 46 + .../include/m68k-linux-gnu/bits/sockaddr.h | 42 + .../include/m68k-linux-gnu/bits/struct_stat.h | 127 + lib/libc/include/m68k-linux-gnu/fpu_control.h | 118 + .../include/m68k-linux-gnu/gnu/lib-names.h | 31 + lib/libc/include/m68k-linux-gnu/gnu/stubs.h | 16 + lib/libc/include/m68k-linux-gnu/sys/reg.h | 101 + .../include/m68k-linux-gnu/sys/ucontext.h | 130 + lib/libc/include/m68k-linux-gnu/sys/user.h | 65 + .../include/m68k-linux-musl/bits/alltypes.h | 424 + .../bits/fcntl.h | 8 +- lib/libc/include/m68k-linux-musl/bits/fenv.h | 29 + lib/libc/include/m68k-linux-musl/bits/float.h | 39 + .../include/m68k-linux-musl/bits/ipcstat.h | 1 + lib/libc/include/m68k-linux-musl/bits/msg.h | 18 + .../include/m68k-linux-musl/bits/ptrace.h | 2 + lib/libc/include/m68k-linux-musl/bits/reg.h | 45 + lib/libc/include/m68k-linux-musl/bits/sem.h | 13 + .../include/m68k-linux-musl/bits/setjmp.h | 1 + lib/libc/include/m68k-linux-musl/bits/shm.h | 31 + .../include/m68k-linux-musl/bits/signal.h | 140 + lib/libc/include/m68k-linux-musl/bits/stat.h | 25 + .../include/m68k-linux-musl/bits/stdint.h | 20 + .../include/m68k-linux-musl/bits/syscall.h | 824 + lib/libc/include/m68k-linux-musl/bits/user.h | 38 + .../bits/struct_rwlock.h | 44 +- .../bits/wordsize.h | 20 +- lib/libc/include/mips-linux-musl/bits/posix.h | 2 - .../bits/struct_rwlock.h | 44 +- .../mips64-linux-gnuabi64/bits/wordsize.h | 31 + .../bits/struct_rwlock.h | 44 +- .../mips64-linux-gnuabin32/bits/wordsize.h | 31 + .../bits/struct_rwlock.h | 44 +- .../mips64el-linux-gnuabi64/bits/wordsize.h | 31 + .../bits/struct_rwlock.h | 71 + .../mips64el-linux-gnuabin32/bits/wordsize.h | 31 + .../mipsel-linux-gnu/bits/struct_rwlock.h | 71 + .../include/mipsel-linux-gnu/bits/wordsize.h | 31 + .../include/powerpc-linux-musl/bits/posix.h | 2 - .../include/powerpc64-linux-musl/bits/posix.h | 2 + .../include/riscv64-linux-gnu/bits/floatn.h | 97 + .../include/riscv64-linux-musl/bits/fcntl.h | 38 - .../include/s390x-linux-gnu/bits/floatn.h | 97 + .../include/s390x-linux-musl/bits/posix.h | 2 + .../include/sparc-linux-gnu/bits/floatn.h | 97 + .../include/sparcv9-linux-gnu/bits/floatn.h | 97 + .../include/x86_64-linux-musl/bits/fcntl.h | 40 - .../include/x86_64-linux-musl/bits/posix.h | 2 + .../include/__algorithm/adjacent_find.h | 51 + lib/libcxx/include/__algorithm/all_of.h | 37 + lib/libcxx/include/__algorithm/any_of.h | 37 + .../include/__algorithm/binary_search.h | 61 + lib/libcxx/include/__algorithm/clamp.h | 52 + lib/libcxx/include/__algorithm/comp.h | 97 + .../include/__algorithm/comp_ref_type.h | 87 + lib/libcxx/include/__algorithm/copy.h | 82 + .../include/__algorithm/copy_backward.h | 84 + lib/libcxx/include/__algorithm/copy_if.h | 48 + lib/libcxx/include/__algorithm/copy_n.h | 72 + lib/libcxx/include/__algorithm/count.h | 40 + lib/libcxx/include/__algorithm/count_if.h | 40 + lib/libcxx/include/__algorithm/equal.h | 90 + lib/libcxx/include/__algorithm/equal_range.h | 87 + lib/libcxx/include/__algorithm/fill.h | 55 + lib/libcxx/include/__algorithm/fill_n.h | 47 + lib/libcxx/include/__algorithm/find.h | 37 + lib/libcxx/include/__algorithm/find_end.h | 154 + .../include/__algorithm/find_first_of.h | 57 + lib/libcxx/include/__algorithm/find_if.h | 37 + lib/libcxx/include/__algorithm/find_if_not.h | 37 + lib/libcxx/include/__algorithm/for_each.h | 37 + lib/libcxx/include/__algorithm/for_each_n.h | 47 + lib/libcxx/include/__algorithm/generate.h | 36 + lib/libcxx/include/__algorithm/generate_n.h | 40 + .../include/__algorithm/half_positive.h | 54 + lib/libcxx/include/__algorithm/includes.h | 67 + .../include/__algorithm/inplace_merge.h | 231 + lib/libcxx/include/__algorithm/is_heap.h | 48 + .../include/__algorithm/is_heap_until.h | 65 + .../include/__algorithm/is_partitioned.h | 43 + .../include/__algorithm/is_permutation.h | 168 + lib/libcxx/include/__algorithm/is_sorted.h | 48 + .../include/__algorithm/is_sorted_until.h | 55 + lib/libcxx/include/__algorithm/iter_swap.h | 37 + .../__algorithm/lexicographical_compare.h | 68 + lib/libcxx/include/__algorithm/lower_bound.h | 72 + lib/libcxx/include/__algorithm/make_heap.h | 64 + lib/libcxx/include/__algorithm/max.h | 70 + lib/libcxx/include/__algorithm/max_element.h | 58 + lib/libcxx/include/__algorithm/merge.h | 76 + lib/libcxx/include/__algorithm/min.h | 70 + lib/libcxx/include/__algorithm/min_element.h | 57 + lib/libcxx/include/__algorithm/minmax.h | 101 + .../include/__algorithm/minmax_element.h | 90 + lib/libcxx/include/__algorithm/mismatch.h | 72 + lib/libcxx/include/__algorithm/move.h | 83 + .../include/__algorithm/move_backward.h | 84 + .../include/__algorithm/next_permutation.h | 77 + lib/libcxx/include/__algorithm/none_of.h | 37 + lib/libcxx/include/__algorithm/nth_element.h | 244 + lib/libcxx/include/__algorithm/partial_sort.h | 71 + .../include/__algorithm/partial_sort_copy.h | 77 + lib/libcxx/include/__algorithm/partition.h | 88 + .../include/__algorithm/partition_copy.h | 52 + .../include/__algorithm/partition_point.h | 51 + lib/libcxx/include/__algorithm/pop_heap.h | 62 + .../include/__algorithm/prev_permutation.h | 77 + lib/libcxx/include/__algorithm/push_heap.h | 75 + lib/libcxx/include/__algorithm/remove.h | 50 + lib/libcxx/include/__algorithm/remove_copy.h | 43 + .../include/__algorithm/remove_copy_if.h | 43 + lib/libcxx/include/__algorithm/remove_if.h | 51 + lib/libcxx/include/__algorithm/replace.h | 37 + lib/libcxx/include/__algorithm/replace_copy.h | 41 + .../include/__algorithm/replace_copy_if.h | 41 + lib/libcxx/include/__algorithm/replace_if.h | 37 + lib/libcxx/include/__algorithm/reverse.h | 61 + lib/libcxx/include/__algorithm/reverse_copy.h | 37 + lib/libcxx/include/__algorithm/rotate.h | 205 + lib/libcxx/include/__algorithm/rotate_copy.h | 38 + lib/libcxx/include/__algorithm/sample.h | 101 + lib/libcxx/include/__algorithm/search.h | 131 + lib/libcxx/include/__algorithm/search_n.h | 116 + .../include/__algorithm/set_difference.h | 77 + .../include/__algorithm/set_intersection.h | 74 + .../__algorithm/set_symmetric_difference.h | 82 + lib/libcxx/include/__algorithm/set_union.h | 77 + lib/libcxx/include/__algorithm/shift_left.h | 61 + lib/libcxx/include/__algorithm/shift_right.h | 106 + lib/libcxx/include/__algorithm/shuffle.h | 127 + lib/libcxx/include/__algorithm/sift_down.h | 84 + lib/libcxx/include/__algorithm/sort.h | 530 + lib/libcxx/include/__algorithm/sort_heap.h | 58 + .../include/__algorithm/stable_partition.h | 305 + lib/libcxx/include/__algorithm/stable_sort.h | 235 + lib/libcxx/include/__algorithm/swap_ranges.h | 37 + lib/libcxx/include/__algorithm/transform.h | 48 + lib/libcxx/include/__algorithm/unique.h | 63 + lib/libcxx/include/__algorithm/unique_copy.h | 114 + lib/libcxx/include/__algorithm/unwrap_iter.h | 87 + lib/libcxx/include/__algorithm/upper_bound.h | 72 + lib/libcxx/include/__availability | 68 +- lib/libcxx/include/__bit_reference | 36 +- lib/libcxx/include/__bits | 3 +- lib/libcxx/include/__bsd_locale_fallbacks.h | 4 +- lib/libcxx/include/__config | 276 +- lib/libcxx/include/__config_site.in | 40 - lib/libcxx/include/__debug | 2 +- lib/libcxx/include/__errc | 2 +- lib/libcxx/include/__format/format_error.h | 56 + .../include/__format/format_parse_context.h | 113 + lib/libcxx/include/__function_like.h | 56 + .../include/__functional/binary_function.h | 31 + .../include/__functional/binary_negate.h | 50 + lib/libcxx/include/__functional/bind.h | 386 + lib/libcxx/include/__functional/bind_front.h | 52 + lib/libcxx/include/__functional/binder1st.h | 54 + lib/libcxx/include/__functional/binder2nd.h | 54 + .../include/__functional/default_searcher.h | 56 + .../function.h} | 1276 +- lib/libcxx/include/__functional/hash.h | 873 + lib/libcxx/include/__functional/identity.h | 37 + lib/libcxx/include/__functional/invoke.h | 100 + .../include/__functional/is_transparent.h | 36 + lib/libcxx/include/__functional/mem_fn.h | 161 + lib/libcxx/include/__functional/mem_fun_ref.h | 173 + lib/libcxx/include/__functional/not_fn.h | 47 + lib/libcxx/include/__functional/operations.h | 729 + .../include/__functional/perfect_forward.h | 88 + .../__functional/pointer_to_binary_function.h | 46 + .../__functional/pointer_to_unary_function.h | 46 + .../include/__functional/ranges_operations.h | 97 + .../include/__functional/reference_wrapper.h | 223 + .../include/__functional/unary_function.h | 34 + .../include/__functional/unary_negate.h | 47 + lib/libcxx/include/__functional/unwrap_ref.h | 62 + .../include/__functional/weak_result_type.h | 481 + lib/libcxx/include/__functional_base | 645 +- lib/libcxx/include/__functional_base_03 | 223 - lib/libcxx/include/__hash_table | 66 +- lib/libcxx/include/__iterator/access.h | 134 + lib/libcxx/include/__iterator/advance.h | 200 + .../include/__iterator/back_insert_iterator.h | 75 + .../include/__iterator/common_iterator.h | 301 + lib/libcxx/include/__iterator/concepts.h | 272 + .../include/__iterator/counted_iterator.h | 306 + lib/libcxx/include/__iterator/data.h | 56 + .../include/__iterator/default_sentinel.h | 35 + lib/libcxx/include/__iterator/distance.h | 56 + lib/libcxx/include/__iterator/empty.h | 49 + .../include/__iterator/erase_if_container.h | 45 + .../__iterator/front_insert_iterator.h | 75 + .../include/__iterator/incrementable_traits.h | 77 + .../include/__iterator/insert_iterator.h | 77 + .../include/__iterator/istream_iterator.h | 103 + .../include/__iterator/istreambuf_iterator.h | 110 + lib/libcxx/include/__iterator/iter_move.h | 91 + lib/libcxx/include/__iterator/iter_swap.h | 107 + lib/libcxx/include/__iterator/iterator.h | 40 + .../include/__iterator/iterator_traits.h | 500 + lib/libcxx/include/__iterator/move_iterator.h | 189 + lib/libcxx/include/__iterator/next.h | 87 + .../include/__iterator/ostream_iterator.h | 75 + .../include/__iterator/ostreambuf_iterator.h | 81 + lib/libcxx/include/__iterator/prev.h | 79 + lib/libcxx/include/__iterator/projected.h | 45 + .../include/__iterator/readable_traits.h | 91 + .../include/__iterator/reverse_access.h | 109 + .../include/__iterator/reverse_iterator.h | 239 + lib/libcxx/include/__iterator/size.h | 58 + lib/libcxx/include/__iterator/wrap_iter.h | 300 + lib/libcxx/include/__libcpp_version | 2 +- lib/libcxx/include/__locale | 26 +- .../include/__memory/{base.h => addressof.h} | 37 +- .../{utilities.h => allocation_guard.h} | 7 +- lib/libcxx/include/__memory/allocator.h | 254 + lib/libcxx/include/__memory/allocator_arg_t.h | 78 + .../include/__memory/allocator_traits.h | 12 +- lib/libcxx/include/__memory/auto_ptr.h | 86 + lib/libcxx/include/__memory/compressed_pair.h | 201 + lib/libcxx/include/__memory/construct_at.h | 59 + lib/libcxx/include/__memory/pointer_safety.h | 57 + lib/libcxx/include/__memory/pointer_traits.h | 51 +- .../include/__memory/raw_storage_iterator.h | 73 + lib/libcxx/include/__memory/shared_ptr.h | 1879 + .../include/__memory/temporary_buffer.h | 89 + .../__memory/uninitialized_algorithms.h | 261 + lib/libcxx/include/__memory/unique_ptr.h | 773 + lib/libcxx/include/__memory/uses_allocator.h | 60 + lib/libcxx/include/__mutex_base | 8 +- lib/libcxx/include/__node_handle | 3 +- lib/libcxx/include/__nullptr | 4 +- .../__random/uniform_int_distribution.h | 316 + lib/libcxx/include/__ranges/access.h | 222 + lib/libcxx/include/__ranges/all.h | 86 + lib/libcxx/include/__ranges/common_view.h | 113 + lib/libcxx/include/__ranges/concepts.h | 138 + lib/libcxx/include/__ranges/copyable_box.h | 175 + lib/libcxx/include/__ranges/dangling.h | 47 + lib/libcxx/include/__ranges/data.h | 86 + lib/libcxx/include/__ranges/drop_view.h | 131 + lib/libcxx/include/__ranges/empty.h | 86 + lib/libcxx/include/__ranges/empty_view.h | 46 + .../include/__ranges/enable_borrowed_range.h | 46 + lib/libcxx/include/__ranges/enable_view.h | 42 + .../include/__ranges/non_propagating_cache.h | 99 + lib/libcxx/include/__ranges/ref_view.h | 87 + lib/libcxx/include/__ranges/size.h | 132 + lib/libcxx/include/__ranges/subrange.h | 267 + lib/libcxx/include/__ranges/transform_view.h | 408 + lib/libcxx/include/__ranges/view_interface.h | 198 + lib/libcxx/include/__split_buffer | 13 +- lib/libcxx/include/__std_stream | 6 +- lib/libcxx/include/__string | 72 +- lib/libcxx/include/__support/ibm/gettod_zos.h | 53 + .../include/__support/ibm/locale_mgmt_zos.h | 53 + lib/libcxx/include/__support/ibm/nanosleep.h | 60 +- lib/libcxx/include/__support/ibm/xlocale.h | 76 +- .../include/__support/openbsd/xlocale.h | 6 +- .../__support/win32/limits_msvc_win32.h | 2 +- .../include/__support/win32/locale_win32.h | 23 +- lib/libcxx/include/__threading_support | 23 +- lib/libcxx/include/__tree | 12 +- lib/libcxx/include/__tuple | 4 +- lib/libcxx/include/__utility/__decay_copy.h | 39 + lib/libcxx/include/__utility/as_const.h | 38 + lib/libcxx/include/__utility/cmp.h | 107 + lib/libcxx/include/__utility/declval.h | 39 + lib/libcxx/include/__utility/exchange.h | 40 + lib/libcxx/include/__utility/forward.h | 42 + lib/libcxx/include/__utility/in_place.h | 63 + .../include/__utility/integer_sequence.h | 83 + lib/libcxx/include/__utility/move.h | 52 + lib/libcxx/include/__utility/pair.h | 585 + .../include/__utility/piecewise_construct.h | 34 + lib/libcxx/include/__utility/rel_ops.h | 67 + lib/libcxx/include/__utility/swap.h | 55 + lib/libcxx/include/__utility/to_underlying.h | 45 + lib/libcxx/include/__variant/monostate.h | 65 + lib/libcxx/include/algorithm | 5328 +- lib/libcxx/include/any | 7 +- lib/libcxx/include/array | 15 +- lib/libcxx/include/atomic | 77 +- lib/libcxx/include/barrier | 9 +- lib/libcxx/include/bit | 82 +- lib/libcxx/include/bitset | 24 +- lib/libcxx/include/ccomplex | 2 +- lib/libcxx/include/cctype | 30 +- lib/libcxx/include/cerrno | 2 +- lib/libcxx/include/cfenv | 30 +- lib/libcxx/include/cfloat | 2 +- lib/libcxx/include/charconv | 137 +- lib/libcxx/include/chrono | 33 +- lib/libcxx/include/cinttypes | 16 +- lib/libcxx/include/ciso646 | 2 +- lib/libcxx/include/climits | 2 +- lib/libcxx/include/clocale | 8 +- lib/libcxx/include/cmath | 428 +- lib/libcxx/include/codecvt | 2 +- lib/libcxx/include/compare | 570 +- lib/libcxx/include/complex | 6 +- lib/libcxx/include/complex.h | 4 +- lib/libcxx/include/concepts | 304 +- lib/libcxx/include/condition_variable | 2 +- lib/libcxx/include/csetjmp | 6 +- lib/libcxx/include/csignal | 8 +- lib/libcxx/include/cstdarg | 4 +- lib/libcxx/include/cstdbool | 2 +- lib/libcxx/include/cstddef | 14 +- lib/libcxx/include/cstdint | 72 +- lib/libcxx/include/cstdio | 102 +- lib/libcxx/include/cstdlib | 88 +- lib/libcxx/include/cstring | 48 +- lib/libcxx/include/ctgmath | 2 +- lib/libcxx/include/ctime | 34 +- lib/libcxx/include/ctype.h | 2 +- lib/libcxx/include/cwchar | 130 +- lib/libcxx/include/cwctype | 46 +- lib/libcxx/include/deque | 50 +- lib/libcxx/include/errno.h | 6 +- lib/libcxx/include/exception | 18 +- lib/libcxx/include/experimental/__config | 4 - lib/libcxx/include/experimental/__memory | 27 +- lib/libcxx/include/experimental/functional | 45 +- lib/libcxx/include/experimental/iterator | 3 + .../include/experimental/propagate_const | 2 +- lib/libcxx/include/experimental/simd | 4 +- lib/libcxx/include/experimental/type_traits | 2 +- lib/libcxx/include/ext/__hash | 3 +- lib/libcxx/include/ext/hash_map | 6 +- lib/libcxx/include/ext/hash_set | 4 +- lib/libcxx/include/filesystem | 142 +- lib/libcxx/include/float.h | 2 +- lib/libcxx/include/format | 86 + lib/libcxx/include/forward_list | 61 +- lib/libcxx/include/fstream | 51 +- lib/libcxx/include/functional | 2786 +- lib/libcxx/include/future | 55 +- lib/libcxx/include/initializer_list | 4 +- lib/libcxx/include/inttypes.h | 2 +- lib/libcxx/include/iomanip | 20 +- lib/libcxx/include/ios | 26 +- lib/libcxx/include/iosfwd | 15 +- lib/libcxx/include/iostream | 6 +- lib/libcxx/include/istream | 97 +- lib/libcxx/include/iterator | 1910 +- lib/libcxx/include/latch | 2 +- lib/libcxx/include/limits | 2 +- lib/libcxx/include/limits.h | 2 +- lib/libcxx/include/list | 92 +- lib/libcxx/include/locale | 79 +- lib/libcxx/include/locale.h | 2 +- lib/libcxx/include/map | 86 +- lib/libcxx/include/math.h | 320 +- lib/libcxx/include/memory | 3475 +- lib/libcxx/include/module.modulemap | 293 +- lib/libcxx/include/mutex | 17 +- lib/libcxx/include/new | 6 +- lib/libcxx/include/numbers | 37 +- lib/libcxx/include/numeric | 11 +- lib/libcxx/include/optional | 193 +- lib/libcxx/include/ostream | 121 +- lib/libcxx/include/queue | 124 +- lib/libcxx/include/random | 102 +- lib/libcxx/include/ranges | 209 + lib/libcxx/include/ratio | 12 +- lib/libcxx/include/regex | 36 +- lib/libcxx/include/scoped_allocator | 7 +- lib/libcxx/include/semaphore | 6 +- lib/libcxx/include/set | 98 +- lib/libcxx/include/setjmp.h | 2 +- lib/libcxx/include/shared_mutex | 8 +- lib/libcxx/include/span | 33 +- lib/libcxx/include/sstream | 50 +- lib/libcxx/include/stack | 36 +- lib/libcxx/include/stdbool.h | 2 +- lib/libcxx/include/stddef.h | 2 +- lib/libcxx/include/stdexcept | 4 +- lib/libcxx/include/stdint.h | 2 +- lib/libcxx/include/stdio.h | 2 +- lib/libcxx/include/stdlib.h | 16 +- lib/libcxx/include/streambuf | 8 +- lib/libcxx/include/string | 374 +- lib/libcxx/include/string.h | 2 +- lib/libcxx/include/string_view | 86 +- lib/libcxx/include/strstream | 14 +- lib/libcxx/include/system_error | 15 +- lib/libcxx/include/tgmath.h | 4 +- lib/libcxx/include/thread | 31 +- lib/libcxx/include/tuple | 1081 +- lib/libcxx/include/type_traits | 548 +- lib/libcxx/include/typeindex | 6 +- lib/libcxx/include/typeinfo | 54 +- lib/libcxx/include/unordered_map | 51 +- lib/libcxx/include/unordered_set | 45 +- lib/libcxx/include/utility | 1469 +- lib/libcxx/include/valarray | 231 +- lib/libcxx/include/variant | 140 +- lib/libcxx/include/vector | 124 +- lib/libcxx/include/version | 58 +- lib/libcxx/include/wchar.h | 4 +- lib/libcxx/include/wctype.h | 14 +- lib/libcxx/src/any.cpp | 6 +- lib/libcxx/src/charconv.cpp | 4 +- lib/libcxx/src/chrono.cpp | 31 +- lib/libcxx/src/condition_variable.cpp | 8 +- lib/libcxx/src/debug.cpp | 2 +- .../src/experimental/memory_resource.cpp | 21 +- .../src/filesystem/directory_iterator.cpp | 10 +- lib/libcxx/src/filesystem/filesystem_common.h | 201 +- lib/libcxx/src/filesystem/operations.cpp | 377 +- lib/libcxx/src/filesystem/posix_compat.h | 521 + lib/libcxx/src/format.cpp | 19 + lib/libcxx/src/functional.cpp | 4 +- lib/libcxx/src/future.cpp | 10 +- lib/libcxx/src/include/config_elast.h | 4 + lib/libcxx/src/include/refstring.h | 8 +- .../include/sso_allocator.h} | 10 +- lib/libcxx/src/ios.cpp | 14 +- lib/libcxx/src/locale.cpp | 167 +- lib/libcxx/src/memory.cpp | 33 +- lib/libcxx/src/mutex.cpp | 24 +- lib/libcxx/src/mutex_destructor.cpp | 2 +- lib/libcxx/src/new.cpp | 48 +- lib/libcxx/src/optional.cpp | 8 +- lib/libcxx/src/random.cpp | 2 +- lib/libcxx/src/string.cpp | 8 +- lib/libcxx/src/support/ibm/xlocale_zos.cpp | 137 + .../support/runtime/exception_fallback.ipp | 50 +- .../src/support/runtime/exception_glibcxx.ipp | 8 +- .../support/runtime/exception_libcxxabi.ipp | 4 +- .../support/runtime/exception_libcxxrt.ipp | 4 +- .../src/support/runtime/exception_msvc.ipp | 50 +- .../runtime/exception_pointer_cxxabi.ipp | 12 +- .../runtime/exception_pointer_glibcxx.ipp | 14 +- .../runtime/exception_pointer_msvc.ipp | 24 +- .../exception_pointer_unimplemented.ipp | 12 +- .../support/runtime/new_handler_fallback.ipp | 4 +- .../src/support/runtime/stdexcept_default.ipp | 30 +- lib/libcxx/src/support/win32/support.cpp | 5 +- lib/libcxx/src/support/win32/thread_win32.cpp | 44 +- lib/libcxx/src/system_error.cpp | 34 +- lib/libcxx/src/thread.cpp | 4 +- lib/libcxx/src/typeinfo.cpp | 6 +- lib/libcxxabi/src/cxa_default_handlers.cpp | 4 +- lib/libcxxabi/src/cxa_exception.cpp | 4 +- lib/libcxxabi/src/cxa_exception.h | 2 +- lib/libcxxabi/src/cxa_handlers.cpp | 16 +- lib/libcxxabi/src/cxa_handlers.h | 4 +- lib/libcxxabi/src/cxa_personality.cpp | 95 +- lib/libcxxabi/src/demangle/ItaniumDemangle.h | 26 +- lib/libcxxabi/src/demangle/StringView.h | 13 +- lib/libcxxabi/src/private_typeinfo.cpp | 4 +- lib/libcxxabi/src/private_typeinfo.h | 2 +- lib/libcxxabi/src/stdlib_exception.cpp | 20 +- lib/libcxxabi/src/stdlib_new_delete.cpp | 52 +- lib/libcxxabi/src/stdlib_stdexcept.cpp | 26 +- lib/libcxxabi/src/stdlib_typeinfo.cpp | 12 +- lib/libunwind/include/__libunwind_config.h | 17 +- lib/libunwind/include/libunwind.h | 47 +- lib/libunwind/src/DwarfInstructions.hpp | 17 +- lib/libunwind/src/Registers.hpp | 123 +- lib/libunwind/src/Unwind-EHABI.cpp | 2 + lib/libunwind/src/UnwindCursor.hpp | 16 +- lib/libunwind/src/UnwindRegistersRestore.S | 384 +- lib/libunwind/src/UnwindRegistersSave.S | 420 +- lib/libunwind/src/assembly.h | 58 +- lib/libunwind/src/config.h | 11 +- lib/libunwind/src/libunwind.cpp | 15 +- lib/std/bit_set.zig | 4 + lib/std/fmt.zig | 4 - lib/std/hash/auto_hash.zig | 3 - lib/std/target.zig | 8 + lib/std/target/aarch64.zig | 126 +- lib/std/target/amdgpu.zig | 263 +- lib/std/target/arc.zig | 39 + lib/std/target/arm.zig | 20 +- lib/std/target/hexagon.zig | 35 + lib/std/target/nvptx.zig | 26 + lib/std/target/powerpc.zig | 35 +- lib/std/target/riscv.zig | 2 +- lib/std/target/systemz.zig | 76 + lib/std/target/wasm.zig | 8 - lib/std/target/x86.zig | 167 +- lib/tsan/interception/interception.h | 35 +- lib/tsan/interception/interception_linux.cpp | 10 +- lib/tsan/interception/interception_linux.h | 10 +- lib/tsan/interception/interception_win.cpp | 6 +- .../sanitizer_common/sanitizer_addrhashmap.h | 106 +- .../sanitizer_common/sanitizer_allocator.cpp | 25 +- .../sanitizer_common/sanitizer_allocator.h | 6 +- .../sanitizer_allocator_checks.h | 10 +- .../sanitizer_allocator_combined.h | 8 +- .../sanitizer_allocator_local_cache.h | 19 +- .../sanitizer_allocator_primary32.h | 8 +- .../sanitizer_allocator_primary64.h | 256 +- .../sanitizer_allocator_report.cpp | 8 + .../sanitizer_allocator_report.h | 1 + .../sanitizer_allocator_secondary.h | 16 +- .../sanitizer_allocator_size_class_map.h | 2 +- lib/tsan/sanitizer_common/sanitizer_atomic.h | 4 +- .../sanitizer_common/sanitizer_atomic_clang.h | 14 +- .../sanitizer_atomic_clang_mips.h | 18 +- .../sanitizer_atomic_clang_other.h | 24 +- .../sanitizer_atomic_clang_x86.h | 6 +- .../sanitizer_common/sanitizer_atomic_msvc.h | 36 +- .../sanitizer_common/sanitizer_common.cpp | 25 +- lib/tsan/sanitizer_common/sanitizer_common.h | 192 +- .../sanitizer_common_interceptors.inc | 312 +- .../sanitizer_common_interceptors_format.inc | 6 + .../sanitizer_common_interceptors_ioctl.inc | 15 +- .../sanitizer_common_interface.inc | 1 + .../sanitizer_common_libcdep.cpp | 60 +- .../sanitizer_common_nolibc.cpp | 5 +- .../sanitizer_common_syscalls.inc | 14 +- .../sanitizer_deadlock_detector1.cpp | 4 +- .../sanitizer_deadlock_detector2.cpp | 34 +- .../sanitizer_deadlock_detector_interface.h | 6 + lib/tsan/sanitizer_common/sanitizer_errno.h | 3 +- .../sanitizer_common/sanitizer_errno_codes.h | 1 + lib/tsan/sanitizer_common/sanitizer_file.cpp | 40 +- lib/tsan/sanitizer_common/sanitizer_file.h | 1 + .../sanitizer_common/sanitizer_flag_parser.h | 2 +- lib/tsan/sanitizer_common/sanitizer_flags.cpp | 16 +- lib/tsan/sanitizer_common/sanitizer_flags.h | 4 + lib/tsan/sanitizer_common/sanitizer_flags.inc | 21 +- .../sanitizer_common/sanitizer_fuchsia.cpp | 122 +- lib/tsan/sanitizer_common/sanitizer_fuchsia.h | 2 + .../sanitizer_common/sanitizer_getauxval.h | 5 +- .../sanitizer_interface_internal.h | 4 + .../sanitizer_internal_defs.h | 24 +- lib/tsan/sanitizer_common/sanitizer_libc.h | 3 +- .../sanitizer_common/sanitizer_libignore.cpp | 4 +- lib/tsan/sanitizer_common/sanitizer_linux.cpp | 291 +- lib/tsan/sanitizer_common/sanitizer_linux.h | 14 +- .../sanitizer_linux_libcdep.cpp | 612 +- .../sanitizer_local_address_space_view.h | 2 +- lib/tsan/sanitizer_common/sanitizer_mac.cpp | 320 +- lib/tsan/sanitizer_common/sanitizer_mac.h | 18 +- .../sanitizer_common/sanitizer_malloc_mac.inc | 6 +- lib/tsan/sanitizer_common/sanitizer_mutex.cpp | 225 + lib/tsan/sanitizer_common/sanitizer_mutex.h | 368 +- .../sanitizer_common/sanitizer_netbsd.cpp | 22 +- .../sanitizer_common/sanitizer_openbsd.cpp | 115 - .../sanitizer_common/sanitizer_platform.h | 68 +- .../sanitizer_platform_interceptors.h | 374 +- .../sanitizer_platform_limits_freebsd.cpp | 5 +- .../sanitizer_platform_limits_netbsd.cpp | 155 +- .../sanitizer_platform_limits_netbsd.h | 6 +- .../sanitizer_platform_limits_openbsd.cpp | 279 - .../sanitizer_platform_limits_openbsd.h | 382 - .../sanitizer_platform_limits_posix.cpp | 119 +- .../sanitizer_platform_limits_posix.h | 28 +- .../sanitizer_platform_limits_solaris.cpp | 3 +- lib/tsan/sanitizer_common/sanitizer_posix.cpp | 16 +- lib/tsan/sanitizer_common/sanitizer_posix.h | 6 +- .../sanitizer_posix_libcdep.cpp | 40 +- .../sanitizer_common/sanitizer_printf.cpp | 84 +- .../sanitizer_common/sanitizer_procmaps.h | 2 +- .../sanitizer_procmaps_bsd.cpp | 31 +- .../sanitizer_procmaps_common.cpp | 4 +- .../sanitizer_procmaps_mac.cpp | 4 +- .../sanitizer_procmaps_solaris.cpp | 7 +- lib/tsan/sanitizer_common/sanitizer_ptrauth.h | 20 + .../sanitizer_common/sanitizer_quarantine.h | 3 +- lib/tsan/sanitizer_common/sanitizer_rtems.cpp | 283 - lib/tsan/sanitizer_common/sanitizer_rtems.h | 20 - .../sanitizer_signal_interceptors.inc | 5 +- .../sanitizer_common/sanitizer_solaris.cpp | 32 +- .../sanitizer_common/sanitizer_stackdepot.cpp | 9 +- .../sanitizer_common/sanitizer_stackdepot.h | 1 + .../sanitizer_stackdepotbase.h | 20 +- .../sanitizer_common/sanitizer_stacktrace.cpp | 40 +- .../sanitizer_common/sanitizer_stacktrace.h | 59 +- .../sanitizer_stacktrace_libcdep.cpp | 174 +- .../sanitizer_stacktrace_printer.cpp | 92 +- .../sanitizer_stacktrace_printer.h | 4 +- .../sanitizer_common/sanitizer_stoptheworld.h | 13 +- .../sanitizer_stoptheworld_fuchsia.cpp | 3 +- .../sanitizer_stoptheworld_linux_libcdep.cpp | 106 +- .../sanitizer_stoptheworld_mac.cpp | 20 +- .../sanitizer_stoptheworld_netbsd_libcdep.cpp | 16 +- .../sanitizer_suppressions.cpp | 4 +- .../sanitizer_symbolizer_internal.h | 7 +- .../sanitizer_symbolizer_libbacktrace.h | 2 +- .../sanitizer_symbolizer_libcdep.cpp | 11 +- .../sanitizer_symbolizer_mac.cpp | 21 +- .../sanitizer_symbolizer_mac.h | 4 +- .../sanitizer_symbolizer_markup.cpp | 24 +- .../sanitizer_symbolizer_posix_libcdep.cpp | 19 +- .../sanitizer_symbolizer_report.cpp | 35 +- .../sanitizer_symbolizer_rtems.h | 40 - .../sanitizer_symbolizer_win.cpp | 20 +- .../sanitizer_syscall_generic.inc | 2 +- .../sanitizer_syscalls_netbsd.inc | 119 +- .../sanitizer_termination.cpp | 33 +- .../sanitizer_thread_registry.cpp | 41 +- .../sanitizer_thread_registry.h | 30 +- .../sanitizer_thread_safety.h | 42 + .../sanitizer_tls_get_addr.cpp | 86 +- .../sanitizer_common/sanitizer_tls_get_addr.h | 21 +- .../sanitizer_common/sanitizer_unwind_win.cpp | 11 + lib/tsan/sanitizer_common/sanitizer_win.cpp | 130 +- lib/tsan/tsan_clock.cpp | 66 +- lib/tsan/tsan_clock.h | 16 +- lib/tsan/tsan_defs.h | 30 +- lib/tsan/tsan_dense_alloc.h | 33 +- lib/tsan/tsan_external.cpp | 11 +- lib/tsan/tsan_flags.cpp | 2 +- lib/tsan/tsan_interceptors.h | 18 +- lib/tsan/tsan_interceptors_mac.cpp | 10 +- lib/tsan/tsan_interceptors_mach_vm.cpp | 19 +- lib/tsan/tsan_interceptors_posix.cpp | 238 +- lib/tsan/tsan_interface.cpp | 13 +- lib/tsan/tsan_interface.h | 12 +- lib/tsan/tsan_interface_ann.cpp | 9 +- lib/tsan/tsan_interface_atomic.cpp | 74 +- lib/tsan/tsan_interface_inl.h | 23 +- lib/tsan/tsan_interface_java.cpp | 1 - lib/tsan/tsan_mman.cpp | 7 +- lib/tsan/tsan_mutex.cpp | 289 - lib/tsan/tsan_mutex.h | 90 - lib/tsan/tsan_platform.h | 199 +- lib/tsan/tsan_platform_linux.cpp | 46 +- lib/tsan/tsan_platform_mac.cpp | 9 +- lib/tsan/tsan_platform_posix.cpp | 69 +- lib/tsan/tsan_report.cpp | 20 +- lib/tsan/tsan_rtl.cpp | 214 +- lib/tsan/tsan_rtl.h | 47 +- lib/tsan/tsan_rtl_mutex.cpp | 75 +- lib/tsan/tsan_rtl_report.cpp | 64 +- lib/tsan/tsan_rtl_thread.cpp | 25 +- lib/tsan/tsan_stack_trace.cpp | 8 +- lib/tsan/tsan_stat.cpp | 186 - lib/tsan/tsan_stat.h | 191 - lib/tsan/tsan_sync.cpp | 15 +- lib/tsan/tsan_sync.h | 9 +- lib/tsan/tsan_trace.h | 5 +- lib/tsan/tsan_update_shadow_word_inl.h | 12 +- lib/tsan/ubsan/ubsan_flags.h | 2 - lib/tsan/ubsan/ubsan_platform.h | 8 +- src/Compilation.zig | 4 +- src/clang.zig | 79 +- src/codegen/llvm.zig | 226 + src/codegen/llvm/bindings.zig | 1 + src/config.zig.in | 4 + src/libcxx.zig | 26 +- src/libtsan.zig | 10 +- src/stage1/codegen.cpp | 39 +- src/stage1/target.cpp | 8 +- src/target.zig | 160 + src/zig_clang.cpp | 602 +- src/zig_clang.h | 79 +- src/zig_clang_cc1_main.cpp | 15 +- src/zig_clang_cc1as_main.cpp | 61 +- src/zig_clang_driver.cpp | 106 +- src/zig_llvm-ar.cpp | 275 +- src/zig_llvm.cpp | 77 +- src/zig_llvm.h | 43 +- test/behavior/byteswap.zig | 3 - test/behavior/shuffle.zig | 3 +- test/standalone/install_raw_hex/build.zig | 0 tools/process_headers.zig | 10 + tools/update_cpu_features.zig | 2 - 739 files changed, 200334 insertions(+), 33553 deletions(-) create mode 100644 lib/include/builtins.h create mode 100644 lib/include/hexagon_circ_brev_intrinsics.h create mode 100644 lib/include/hexagon_protos.h create mode 100644 lib/include/hexagon_types.h create mode 100644 lib/include/hvx_hexagon_protos.h create mode 100644 lib/include/riscv_vector.h create mode 100644 lib/libc/include/aarch64-linux-gnu/bits/floatn.h create mode 100644 lib/libc/include/aarch64-linux-musl/bits/posix.h create mode 100644 lib/libc/include/aarch64_be-linux-gnu/bits/floatn.h delete mode 100644 lib/libc/include/arm-linux-gnueabi/bits/floatn.h delete mode 100644 lib/libc/include/arm-linux-gnueabihf/bits/floatn.h delete mode 100644 lib/libc/include/arm-linux-musl/bits/posix.h delete mode 100644 lib/libc/include/armeb-linux-gnueabi/bits/floatn.h delete mode 100644 lib/libc/include/armeb-linux-gnueabihf/bits/floatn.h delete mode 100644 lib/libc/include/csky-linux-gnueabi/bits/floatn.h delete mode 100644 lib/libc/include/csky-linux-gnueabi/bits/struct_rwlock.h delete mode 100644 lib/libc/include/csky-linux-gnueabi/bits/wordsize.h delete mode 100644 lib/libc/include/csky-linux-gnueabihf/bits/floatn.h delete mode 100644 lib/libc/include/csky-linux-gnueabihf/bits/struct_rwlock.h delete mode 100644 lib/libc/include/csky-linux-gnueabihf/bits/wordsize.h delete mode 100644 lib/libc/include/i386-linux-musl/bits/posix.h create mode 100644 lib/libc/include/m68k-linux-gnu/bits/a.out.h create mode 100644 lib/libc/include/m68k-linux-gnu/bits/endianness.h create mode 100644 lib/libc/include/m68k-linux-gnu/bits/fcntl.h create mode 100644 lib/libc/include/m68k-linux-gnu/bits/fenv.h create mode 100644 lib/libc/include/m68k-linux-gnu/bits/flt-eval-method.h rename lib/libc/include/{armeb-linux-gnueabi/bits/wordsize.h => m68k-linux-gnu/bits/fp-logb.h} (72%) create mode 100644 lib/libc/include/m68k-linux-gnu/bits/iscanonical.h create mode 100644 lib/libc/include/m68k-linux-gnu/bits/link.h rename lib/libc/include/{arm-linux-gnueabihf/bits/wordsize.h => m68k-linux-gnu/bits/long-double.h} (68%) create mode 100644 lib/libc/include/m68k-linux-gnu/bits/poll.h rename lib/libc/include/{armeb-linux-gnueabihf/bits/wordsize.h => m68k-linux-gnu/bits/procfs-id.h} (69%) create mode 100644 lib/libc/include/m68k-linux-gnu/bits/procfs.h create mode 100644 lib/libc/include/m68k-linux-gnu/bits/pthreadtypes-arch.h create mode 100644 lib/libc/include/m68k-linux-gnu/bits/semaphore.h create mode 100644 lib/libc/include/m68k-linux-gnu/bits/setjmp.h create mode 100644 lib/libc/include/m68k-linux-gnu/bits/sockaddr.h create mode 100644 lib/libc/include/m68k-linux-gnu/bits/struct_stat.h create mode 100644 lib/libc/include/m68k-linux-gnu/fpu_control.h create mode 100644 lib/libc/include/m68k-linux-gnu/gnu/lib-names.h create mode 100644 lib/libc/include/m68k-linux-gnu/gnu/stubs.h create mode 100644 lib/libc/include/m68k-linux-gnu/sys/reg.h create mode 100644 lib/libc/include/m68k-linux-gnu/sys/ucontext.h create mode 100644 lib/libc/include/m68k-linux-gnu/sys/user.h create mode 100644 lib/libc/include/m68k-linux-musl/bits/alltypes.h rename lib/libc/include/{i386-linux-musl => m68k-linux-musl}/bits/fcntl.h (83%) create mode 100644 lib/libc/include/m68k-linux-musl/bits/fenv.h create mode 100644 lib/libc/include/m68k-linux-musl/bits/float.h create mode 100644 lib/libc/include/m68k-linux-musl/bits/ipcstat.h create mode 100644 lib/libc/include/m68k-linux-musl/bits/msg.h create mode 100644 lib/libc/include/m68k-linux-musl/bits/ptrace.h create mode 100644 lib/libc/include/m68k-linux-musl/bits/reg.h create mode 100644 lib/libc/include/m68k-linux-musl/bits/sem.h create mode 100644 lib/libc/include/m68k-linux-musl/bits/setjmp.h create mode 100644 lib/libc/include/m68k-linux-musl/bits/shm.h create mode 100644 lib/libc/include/m68k-linux-musl/bits/signal.h create mode 100644 lib/libc/include/m68k-linux-musl/bits/stat.h create mode 100644 lib/libc/include/m68k-linux-musl/bits/stdint.h create mode 100644 lib/libc/include/m68k-linux-musl/bits/syscall.h create mode 100644 lib/libc/include/m68k-linux-musl/bits/user.h rename lib/libc/include/{armeb-linux-gnueabi => mips-linux-gnu}/bits/struct_rwlock.h (64%) rename lib/libc/include/{arm-linux-gnueabi => mips-linux-gnu}/bits/wordsize.h (67%) delete mode 100644 lib/libc/include/mips-linux-musl/bits/posix.h rename lib/libc/include/{arm-linux-gnueabi => mips64-linux-gnuabi64}/bits/struct_rwlock.h (64%) create mode 100644 lib/libc/include/mips64-linux-gnuabi64/bits/wordsize.h rename lib/libc/include/{arm-linux-gnueabihf => mips64-linux-gnuabin32}/bits/struct_rwlock.h (64%) create mode 100644 lib/libc/include/mips64-linux-gnuabin32/bits/wordsize.h rename lib/libc/include/{armeb-linux-gnueabihf => mips64el-linux-gnuabi64}/bits/struct_rwlock.h (64%) create mode 100644 lib/libc/include/mips64el-linux-gnuabi64/bits/wordsize.h create mode 100644 lib/libc/include/mips64el-linux-gnuabin32/bits/struct_rwlock.h create mode 100644 lib/libc/include/mips64el-linux-gnuabin32/bits/wordsize.h create mode 100644 lib/libc/include/mipsel-linux-gnu/bits/struct_rwlock.h create mode 100644 lib/libc/include/mipsel-linux-gnu/bits/wordsize.h delete mode 100644 lib/libc/include/powerpc-linux-musl/bits/posix.h create mode 100644 lib/libc/include/powerpc64-linux-musl/bits/posix.h create mode 100644 lib/libc/include/riscv64-linux-gnu/bits/floatn.h delete mode 100644 lib/libc/include/riscv64-linux-musl/bits/fcntl.h create mode 100644 lib/libc/include/s390x-linux-gnu/bits/floatn.h create mode 100644 lib/libc/include/s390x-linux-musl/bits/posix.h create mode 100644 lib/libc/include/sparc-linux-gnu/bits/floatn.h create mode 100644 lib/libc/include/sparcv9-linux-gnu/bits/floatn.h delete mode 100644 lib/libc/include/x86_64-linux-musl/bits/fcntl.h create mode 100644 lib/libc/include/x86_64-linux-musl/bits/posix.h create mode 100644 lib/libcxx/include/__algorithm/adjacent_find.h create mode 100644 lib/libcxx/include/__algorithm/all_of.h create mode 100644 lib/libcxx/include/__algorithm/any_of.h create mode 100644 lib/libcxx/include/__algorithm/binary_search.h create mode 100644 lib/libcxx/include/__algorithm/clamp.h create mode 100644 lib/libcxx/include/__algorithm/comp.h create mode 100644 lib/libcxx/include/__algorithm/comp_ref_type.h create mode 100644 lib/libcxx/include/__algorithm/copy.h create mode 100644 lib/libcxx/include/__algorithm/copy_backward.h create mode 100644 lib/libcxx/include/__algorithm/copy_if.h create mode 100644 lib/libcxx/include/__algorithm/copy_n.h create mode 100644 lib/libcxx/include/__algorithm/count.h create mode 100644 lib/libcxx/include/__algorithm/count_if.h create mode 100644 lib/libcxx/include/__algorithm/equal.h create mode 100644 lib/libcxx/include/__algorithm/equal_range.h create mode 100644 lib/libcxx/include/__algorithm/fill.h create mode 100644 lib/libcxx/include/__algorithm/fill_n.h create mode 100644 lib/libcxx/include/__algorithm/find.h create mode 100644 lib/libcxx/include/__algorithm/find_end.h create mode 100644 lib/libcxx/include/__algorithm/find_first_of.h create mode 100644 lib/libcxx/include/__algorithm/find_if.h create mode 100644 lib/libcxx/include/__algorithm/find_if_not.h create mode 100644 lib/libcxx/include/__algorithm/for_each.h create mode 100644 lib/libcxx/include/__algorithm/for_each_n.h create mode 100644 lib/libcxx/include/__algorithm/generate.h create mode 100644 lib/libcxx/include/__algorithm/generate_n.h create mode 100644 lib/libcxx/include/__algorithm/half_positive.h create mode 100644 lib/libcxx/include/__algorithm/includes.h create mode 100644 lib/libcxx/include/__algorithm/inplace_merge.h create mode 100644 lib/libcxx/include/__algorithm/is_heap.h create mode 100644 lib/libcxx/include/__algorithm/is_heap_until.h create mode 100644 lib/libcxx/include/__algorithm/is_partitioned.h create mode 100644 lib/libcxx/include/__algorithm/is_permutation.h create mode 100644 lib/libcxx/include/__algorithm/is_sorted.h create mode 100644 lib/libcxx/include/__algorithm/is_sorted_until.h create mode 100644 lib/libcxx/include/__algorithm/iter_swap.h create mode 100644 lib/libcxx/include/__algorithm/lexicographical_compare.h create mode 100644 lib/libcxx/include/__algorithm/lower_bound.h create mode 100644 lib/libcxx/include/__algorithm/make_heap.h create mode 100644 lib/libcxx/include/__algorithm/max.h create mode 100644 lib/libcxx/include/__algorithm/max_element.h create mode 100644 lib/libcxx/include/__algorithm/merge.h create mode 100644 lib/libcxx/include/__algorithm/min.h create mode 100644 lib/libcxx/include/__algorithm/min_element.h create mode 100644 lib/libcxx/include/__algorithm/minmax.h create mode 100644 lib/libcxx/include/__algorithm/minmax_element.h create mode 100644 lib/libcxx/include/__algorithm/mismatch.h create mode 100644 lib/libcxx/include/__algorithm/move.h create mode 100644 lib/libcxx/include/__algorithm/move_backward.h create mode 100644 lib/libcxx/include/__algorithm/next_permutation.h create mode 100644 lib/libcxx/include/__algorithm/none_of.h create mode 100644 lib/libcxx/include/__algorithm/nth_element.h create mode 100644 lib/libcxx/include/__algorithm/partial_sort.h create mode 100644 lib/libcxx/include/__algorithm/partial_sort_copy.h create mode 100644 lib/libcxx/include/__algorithm/partition.h create mode 100644 lib/libcxx/include/__algorithm/partition_copy.h create mode 100644 lib/libcxx/include/__algorithm/partition_point.h create mode 100644 lib/libcxx/include/__algorithm/pop_heap.h create mode 100644 lib/libcxx/include/__algorithm/prev_permutation.h create mode 100644 lib/libcxx/include/__algorithm/push_heap.h create mode 100644 lib/libcxx/include/__algorithm/remove.h create mode 100644 lib/libcxx/include/__algorithm/remove_copy.h create mode 100644 lib/libcxx/include/__algorithm/remove_copy_if.h create mode 100644 lib/libcxx/include/__algorithm/remove_if.h create mode 100644 lib/libcxx/include/__algorithm/replace.h create mode 100644 lib/libcxx/include/__algorithm/replace_copy.h create mode 100644 lib/libcxx/include/__algorithm/replace_copy_if.h create mode 100644 lib/libcxx/include/__algorithm/replace_if.h create mode 100644 lib/libcxx/include/__algorithm/reverse.h create mode 100644 lib/libcxx/include/__algorithm/reverse_copy.h create mode 100644 lib/libcxx/include/__algorithm/rotate.h create mode 100644 lib/libcxx/include/__algorithm/rotate_copy.h create mode 100644 lib/libcxx/include/__algorithm/sample.h create mode 100644 lib/libcxx/include/__algorithm/search.h create mode 100644 lib/libcxx/include/__algorithm/search_n.h create mode 100644 lib/libcxx/include/__algorithm/set_difference.h create mode 100644 lib/libcxx/include/__algorithm/set_intersection.h create mode 100644 lib/libcxx/include/__algorithm/set_symmetric_difference.h create mode 100644 lib/libcxx/include/__algorithm/set_union.h create mode 100644 lib/libcxx/include/__algorithm/shift_left.h create mode 100644 lib/libcxx/include/__algorithm/shift_right.h create mode 100644 lib/libcxx/include/__algorithm/shuffle.h create mode 100644 lib/libcxx/include/__algorithm/sift_down.h create mode 100644 lib/libcxx/include/__algorithm/sort.h create mode 100644 lib/libcxx/include/__algorithm/sort_heap.h create mode 100644 lib/libcxx/include/__algorithm/stable_partition.h create mode 100644 lib/libcxx/include/__algorithm/stable_sort.h create mode 100644 lib/libcxx/include/__algorithm/swap_ranges.h create mode 100644 lib/libcxx/include/__algorithm/transform.h create mode 100644 lib/libcxx/include/__algorithm/unique.h create mode 100644 lib/libcxx/include/__algorithm/unique_copy.h create mode 100644 lib/libcxx/include/__algorithm/unwrap_iter.h create mode 100644 lib/libcxx/include/__algorithm/upper_bound.h delete mode 100644 lib/libcxx/include/__config_site.in create mode 100644 lib/libcxx/include/__format/format_error.h create mode 100644 lib/libcxx/include/__format/format_parse_context.h create mode 100644 lib/libcxx/include/__function_like.h create mode 100644 lib/libcxx/include/__functional/binary_function.h create mode 100644 lib/libcxx/include/__functional/binary_negate.h create mode 100644 lib/libcxx/include/__functional/bind.h create mode 100644 lib/libcxx/include/__functional/bind_front.h create mode 100644 lib/libcxx/include/__functional/binder1st.h create mode 100644 lib/libcxx/include/__functional/binder2nd.h create mode 100644 lib/libcxx/include/__functional/default_searcher.h rename lib/libcxx/include/{__functional_03 => __functional/function.h} (52%) create mode 100644 lib/libcxx/include/__functional/hash.h create mode 100644 lib/libcxx/include/__functional/identity.h create mode 100644 lib/libcxx/include/__functional/invoke.h create mode 100644 lib/libcxx/include/__functional/is_transparent.h create mode 100644 lib/libcxx/include/__functional/mem_fn.h create mode 100644 lib/libcxx/include/__functional/mem_fun_ref.h create mode 100644 lib/libcxx/include/__functional/not_fn.h create mode 100644 lib/libcxx/include/__functional/operations.h create mode 100644 lib/libcxx/include/__functional/perfect_forward.h create mode 100644 lib/libcxx/include/__functional/pointer_to_binary_function.h create mode 100644 lib/libcxx/include/__functional/pointer_to_unary_function.h create mode 100644 lib/libcxx/include/__functional/ranges_operations.h create mode 100644 lib/libcxx/include/__functional/reference_wrapper.h create mode 100644 lib/libcxx/include/__functional/unary_function.h create mode 100644 lib/libcxx/include/__functional/unary_negate.h create mode 100644 lib/libcxx/include/__functional/unwrap_ref.h create mode 100644 lib/libcxx/include/__functional/weak_result_type.h delete mode 100644 lib/libcxx/include/__functional_base_03 create mode 100644 lib/libcxx/include/__iterator/access.h create mode 100644 lib/libcxx/include/__iterator/advance.h create mode 100644 lib/libcxx/include/__iterator/back_insert_iterator.h create mode 100644 lib/libcxx/include/__iterator/common_iterator.h create mode 100644 lib/libcxx/include/__iterator/concepts.h create mode 100644 lib/libcxx/include/__iterator/counted_iterator.h create mode 100644 lib/libcxx/include/__iterator/data.h create mode 100644 lib/libcxx/include/__iterator/default_sentinel.h create mode 100644 lib/libcxx/include/__iterator/distance.h create mode 100644 lib/libcxx/include/__iterator/empty.h create mode 100644 lib/libcxx/include/__iterator/erase_if_container.h create mode 100644 lib/libcxx/include/__iterator/front_insert_iterator.h create mode 100644 lib/libcxx/include/__iterator/incrementable_traits.h create mode 100644 lib/libcxx/include/__iterator/insert_iterator.h create mode 100644 lib/libcxx/include/__iterator/istream_iterator.h create mode 100644 lib/libcxx/include/__iterator/istreambuf_iterator.h create mode 100644 lib/libcxx/include/__iterator/iter_move.h create mode 100644 lib/libcxx/include/__iterator/iter_swap.h create mode 100644 lib/libcxx/include/__iterator/iterator.h create mode 100644 lib/libcxx/include/__iterator/iterator_traits.h create mode 100644 lib/libcxx/include/__iterator/move_iterator.h create mode 100644 lib/libcxx/include/__iterator/next.h create mode 100644 lib/libcxx/include/__iterator/ostream_iterator.h create mode 100644 lib/libcxx/include/__iterator/ostreambuf_iterator.h create mode 100644 lib/libcxx/include/__iterator/prev.h create mode 100644 lib/libcxx/include/__iterator/projected.h create mode 100644 lib/libcxx/include/__iterator/readable_traits.h create mode 100644 lib/libcxx/include/__iterator/reverse_access.h create mode 100644 lib/libcxx/include/__iterator/reverse_iterator.h create mode 100644 lib/libcxx/include/__iterator/size.h create mode 100644 lib/libcxx/include/__iterator/wrap_iter.h rename lib/libcxx/include/__memory/{base.h => addressof.h} (72%) rename lib/libcxx/include/__memory/{utilities.h => allocation_guard.h} (94%) create mode 100644 lib/libcxx/include/__memory/allocator.h create mode 100644 lib/libcxx/include/__memory/allocator_arg_t.h create mode 100644 lib/libcxx/include/__memory/auto_ptr.h create mode 100644 lib/libcxx/include/__memory/compressed_pair.h create mode 100644 lib/libcxx/include/__memory/construct_at.h create mode 100644 lib/libcxx/include/__memory/pointer_safety.h create mode 100644 lib/libcxx/include/__memory/raw_storage_iterator.h create mode 100644 lib/libcxx/include/__memory/shared_ptr.h create mode 100644 lib/libcxx/include/__memory/temporary_buffer.h create mode 100644 lib/libcxx/include/__memory/uninitialized_algorithms.h create mode 100644 lib/libcxx/include/__memory/unique_ptr.h create mode 100644 lib/libcxx/include/__memory/uses_allocator.h create mode 100644 lib/libcxx/include/__random/uniform_int_distribution.h create mode 100644 lib/libcxx/include/__ranges/access.h create mode 100644 lib/libcxx/include/__ranges/all.h create mode 100644 lib/libcxx/include/__ranges/common_view.h create mode 100644 lib/libcxx/include/__ranges/concepts.h create mode 100644 lib/libcxx/include/__ranges/copyable_box.h create mode 100644 lib/libcxx/include/__ranges/dangling.h create mode 100644 lib/libcxx/include/__ranges/data.h create mode 100644 lib/libcxx/include/__ranges/drop_view.h create mode 100644 lib/libcxx/include/__ranges/empty.h create mode 100644 lib/libcxx/include/__ranges/empty_view.h create mode 100644 lib/libcxx/include/__ranges/enable_borrowed_range.h create mode 100644 lib/libcxx/include/__ranges/enable_view.h create mode 100644 lib/libcxx/include/__ranges/non_propagating_cache.h create mode 100644 lib/libcxx/include/__ranges/ref_view.h create mode 100644 lib/libcxx/include/__ranges/size.h create mode 100644 lib/libcxx/include/__ranges/subrange.h create mode 100644 lib/libcxx/include/__ranges/transform_view.h create mode 100644 lib/libcxx/include/__ranges/view_interface.h create mode 100644 lib/libcxx/include/__support/ibm/gettod_zos.h create mode 100644 lib/libcxx/include/__support/ibm/locale_mgmt_zos.h create mode 100644 lib/libcxx/include/__utility/__decay_copy.h create mode 100644 lib/libcxx/include/__utility/as_const.h create mode 100644 lib/libcxx/include/__utility/cmp.h create mode 100644 lib/libcxx/include/__utility/declval.h create mode 100644 lib/libcxx/include/__utility/exchange.h create mode 100644 lib/libcxx/include/__utility/forward.h create mode 100644 lib/libcxx/include/__utility/in_place.h create mode 100644 lib/libcxx/include/__utility/integer_sequence.h create mode 100644 lib/libcxx/include/__utility/move.h create mode 100644 lib/libcxx/include/__utility/pair.h create mode 100644 lib/libcxx/include/__utility/piecewise_construct.h create mode 100644 lib/libcxx/include/__utility/rel_ops.h create mode 100644 lib/libcxx/include/__utility/swap.h create mode 100644 lib/libcxx/include/__utility/to_underlying.h create mode 100644 lib/libcxx/include/__variant/monostate.h create mode 100644 lib/libcxx/include/format create mode 100644 lib/libcxx/include/ranges create mode 100644 lib/libcxx/src/filesystem/posix_compat.h create mode 100644 lib/libcxx/src/format.cpp rename lib/libcxx/{include/__sso_allocator => src/include/sso_allocator.h} (89%) create mode 100644 lib/libcxx/src/support/ibm/xlocale_zos.cpp create mode 100644 lib/std/target/arc.zig create mode 100644 lib/tsan/sanitizer_common/sanitizer_mutex.cpp delete mode 100644 lib/tsan/sanitizer_common/sanitizer_rtems.cpp delete mode 100644 lib/tsan/sanitizer_common/sanitizer_rtems.h delete mode 100644 lib/tsan/sanitizer_common/sanitizer_symbolizer_rtems.h create mode 100644 lib/tsan/sanitizer_common/sanitizer_thread_safety.h delete mode 100644 lib/tsan/tsan_mutex.cpp delete mode 100644 lib/tsan/tsan_mutex.h delete mode 100644 lib/tsan/tsan_stat.cpp delete mode 100644 lib/tsan/tsan_stat.h create mode 100644 test/standalone/install_raw_hex/build.zig diff --git a/build.zig b/build.zig index 14a46b3f5..387723358 100644 --- a/build.zig +++ b/build.zig @@ -61,6 +61,27 @@ pub fn build(b: *Builder) !void { const omit_stage2 = b.option(bool, "omit-stage2", "Do not include stage2 behind a feature flag inside stage1") orelse false; const static_llvm = b.option(bool, "static-llvm", "Disable integration with system-installed LLVM, Clang, LLD, and libc++") orelse false; const enable_llvm = b.option(bool, "enable-llvm", "Build self-hosted compiler with LLVM backend enabled") orelse (is_stage1 or static_llvm); + const llvm_has_m68k = b.option( + bool, + "llvm-has-m68k", + "Whether LLVM has the experimental target m68k enabled", + ) orelse false; + const llvm_has_csky = b.option( + bool, + "llvm-has-csky", + "Whether LLVM has the experimental target csky enabled", + ) orelse false; + const llvm_has_ve = b.option( + bool, + "llvm-has-ve", + "Whether LLVM has the experimental target ve enabled", + ) orelse false; + const llvm_has_arc = b.option( + bool, + "llvm-has-arc", + "Whether LLVM has the experimental target arc enabled", + ) orelse false; + const enable_macos_sdk = b.option(bool, "enable-macos-sdk", "Run tests requiring presence of macOS SDK and frameworks") orelse false; const config_h_path_option = b.option([]const u8, "config_h", "Path to the generated config.h"); if (!skip_install_lib_files) { @@ -104,6 +125,20 @@ pub fn build(b: *Builder) !void { toolchain_step.dependOn(&exe.step); b.default_step.dependOn(&exe.step); + if (target.isWindows() and target.getAbi() == .gnu) { + // LTO is currently broken on mingw, this can be removed when it's fixed. + exe.want_lto = false; + test_stage2.want_lto = false; + } + + exe.addBuildOption(u32, "mem_leak_frames", mem_leak_frames); + exe.addBuildOption(bool, "skip_non_native", skip_non_native); + exe.addBuildOption(bool, "have_llvm", enable_llvm); + exe.addBuildOption(bool, "llvm_has_m68k", llvm_has_m68k); + exe.addBuildOption(bool, "llvm_has_csky", llvm_has_csky); + exe.addBuildOption(bool, "llvm_has_ve", llvm_has_ve); + exe.addBuildOption(bool, "llvm_has_arc", llvm_has_arc); + exe.addBuildOption(u32, "mem_leak_frames", mem_leak_frames); exe.addBuildOption(bool, "skip_non_native", skip_non_native); exe.addBuildOption(bool, "have_llvm", enable_llvm); @@ -237,6 +272,10 @@ pub fn build(b: *Builder) !void { test_stage2.addBuildOption(bool, "is_stage1", is_stage1); test_stage2.addBuildOption(bool, "omit_stage2", omit_stage2); test_stage2.addBuildOption(bool, "have_llvm", enable_llvm); + test_stage2.addBuildOption(bool, "llvm_has_m68k", llvm_has_m68k); + test_stage2.addBuildOption(bool, "llvm_has_csky", llvm_has_csky); + test_stage2.addBuildOption(bool, "llvm_has_ve", llvm_has_ve); + test_stage2.addBuildOption(bool, "llvm_has_arc", llvm_has_arc); test_stage2.addBuildOption(bool, "enable_qemu", is_qemu_enabled); test_stage2.addBuildOption(bool, "enable_wine", is_wine_enabled); test_stage2.addBuildOption(bool, "enable_wasmtime", is_wasmtime_enabled); diff --git a/ci/drone/drone.yml b/ci/drone/drone.yml index bfdc5a377..42be01120 100644 --- a/ci/drone/drone.yml +++ b/ci/drone/drone.yml @@ -7,28 +7,28 @@ platform: steps: - name: build - image: ziglang/static-base:llvm12-aarch64-5 + image: ziglang/static-base:llvm13-aarch64-1 commands: - ./ci/drone/linux_script_build - name: test-1 depends_on: - build - image: ziglang/static-base:llvm12-aarch64-5 + image: ziglang/static-base:llvm13-aarch64-1 commands: - ./ci/drone/linux_script_test 1 - name: test-2 depends_on: - build - image: ziglang/static-base:llvm12-aarch64-5 + image: ziglang/static-base:llvm13-aarch64-1 commands: - ./ci/drone/linux_script_test 2 - name: test-3 depends_on: - build - image: ziglang/static-base:llvm12-aarch64-5 + image: ziglang/static-base:llvm13-aarch64-1 commands: - ./ci/drone/linux_script_test 3 @@ -38,7 +38,7 @@ steps: - test-1 - test-2 - test-3 - image: ziglang/static-base:llvm12-aarch64-5 + image: ziglang/static-base:llvm13-aarch64-1 environment: SRHT_OAUTH_TOKEN: from_secret: SRHT_OAUTH_TOKEN diff --git a/cmake/Findclang.cmake b/cmake/Findclang.cmake index 3ce52df89..10a0205be 100644 --- a/cmake/Findclang.cmake +++ b/cmake/Findclang.cmake @@ -9,27 +9,31 @@ find_path(CLANG_INCLUDE_DIRS NAMES clang/Frontend/ASTUnit.h PATHS - /usr/lib/llvm/12/include - /usr/lib/llvm-12/include - /usr/lib/llvm-12.0/include - /usr/local/llvm120/include - /usr/local/llvm12/include + /usr/lib/llvm/13/include + /usr/lib/llvm-13/include + /usr/lib/llvm-13.0/include + /usr/local/llvm130/include + /usr/local/llvm13/include + /usr/local/opt/llvm@13/include + /opt/homebrew/opt/llvm@13/include /mingw64/include ) if(ZIG_PREFER_CLANG_CPP_DYLIB) find_library(CLANG_LIBRARIES NAMES - clang-cpp-12.0 - clang-cpp120 + clang-cpp-13.0 + clang-cpp130 clang-cpp PATHS ${CLANG_LIBDIRS} - /usr/lib/llvm/12/lib - /usr/lib/llvm/12/lib64 - /usr/lib/llvm-12/lib - /usr/local/llvm120/lib - /usr/local/llvm12/lib + /usr/lib/llvm/13/lib + /usr/lib/llvm/13/lib64 + /usr/lib/llvm-13/lib + /usr/local/llvm130/lib + /usr/local/llvm13/lib + /usr/local/opt/llvm@13/lib + /opt/homebrew/opt/llvm@13/lib ) endif() @@ -39,11 +43,13 @@ if(NOT CLANG_LIBRARIES) find_library(CLANG_${_prettylibname_}_LIB NAMES ${_libname_} PATHS ${CLANG_LIBDIRS} - /usr/lib/llvm/12/lib - /usr/lib/llvm-12/lib - /usr/lib/llvm-12.0/lib - /usr/local/llvm120/lib - /usr/local/llvm12/lib + /usr/lib/llvm/13/lib + /usr/lib/llvm-13/lib + /usr/lib/llvm-13.0/lib + /usr/local/llvm130/lib + /usr/local/llvm13/lib + /usr/local/opt/llvm@13/lib + /opt/homebrew/opt/llvm@13/lib /mingw64/lib /c/msys64/mingw64/lib c:\\msys64\\mingw64\\lib diff --git a/cmake/Findlld.cmake b/cmake/Findlld.cmake index 72724ecd1..1a0de6595 100644 --- a/cmake/Findlld.cmake +++ b/cmake/Findlld.cmake @@ -8,16 +8,20 @@ find_path(LLD_INCLUDE_DIRS NAMES lld/Common/Driver.h PATHS - /usr/lib/llvm-12/include - /usr/local/llvm120/include - /usr/local/llvm12/include + /usr/lib/llvm-13/include + /usr/local/llvm130/include + /usr/local/llvm13/include + /usr/local/opt/llvm@13/include + /opt/homebrew/opt/llvm@13/include /mingw64/include) -find_library(LLD_LIBRARY NAMES lld-12.0 lld120 lld +find_library(LLD_LIBRARY NAMES lld-13.0 lld130 lld PATHS - /usr/lib/llvm-12/lib - /usr/local/llvm120/lib - /usr/local/llvm12/lib + /usr/lib/llvm-13/lib + /usr/local/llvm130/lib + /usr/local/llvm13/lib + /usr/local/opt/llvm@13/lib + /opt/homebrew/opt/llvm@13/lib ) if(EXISTS ${LLD_LIBRARY}) set(LLD_LIBRARIES ${LLD_LIBRARY}) @@ -27,9 +31,11 @@ else() find_library(LLD_${_prettylibname_}_LIB NAMES ${_libname_} PATHS ${LLD_LIBDIRS} - /usr/lib/llvm-12/lib - /usr/local/llvm120/lib - /usr/local/llvm12/lib + /usr/lib/llvm-13/lib + /usr/local/llvm130/lib + /usr/local/llvm13/lib + /usr/local/opt/llvm@13/lib + /opt/homebrew/opt/llvm@13/lib /mingw64/lib /c/msys64/mingw64/lib c:/msys64/mingw64/lib) diff --git a/cmake/Findllvm.cmake b/cmake/Findllvm.cmake index bfde645ca..47947b4a8 100644 --- a/cmake/Findllvm.cmake +++ b/cmake/Findllvm.cmake @@ -9,37 +9,41 @@ find_path(LLVM_INCLUDE_DIRS NAMES llvm/IR/IRBuilder.h PATHS - /usr/lib/llvm/12/include - /usr/lib/llvm-12/include - /usr/lib/llvm-12.0/include - /usr/local/llvm12/include - /usr/local/llvm120/include + /usr/lib/llvm/13/include + /usr/lib/llvm-13/include + /usr/lib/llvm-13.0/include + /usr/local/llvm13/include + /usr/local/llvm130/include + /usr/local/opt/llvm@13/include + /opt/homebrew/opt/llvm@13/include /mingw64/include ) if(ZIG_PREFER_CLANG_CPP_DYLIB) find_library(LLVM_LIBRARIES NAMES - LLVM-12.0 - LLVM-12 - LLVM-120 + LLVM-13.0 + LLVM-13 + LLVM-130 LLVM PATHS ${LLVM_LIBDIRS} - /usr/lib/llvm/12/lib - /usr/lib/llvm/12/lib64 - /usr/lib/llvm-12/lib - /usr/local/llvm12/lib - /usr/local/llvm120/lib + /usr/lib/llvm/13/lib + /usr/lib/llvm/13/lib64 + /usr/lib/llvm-13/lib + /usr/local/llvm13/lib + /usr/local/llvm130/lib + /usr/local/opt/llvm@13/lib + /opt/homebrew/opt/llvm@13/lib ) find_program(LLVM_CONFIG_EXE - NAMES llvm-config-12 llvm-config-12.0 llvm-config120 llvm-config12 llvm-config + NAMES llvm-config-13 llvm-config-13.0 llvm-config130 llvm-config13 llvm-config PATHS "/mingw64/bin" "/c/msys64/mingw64/bin" "c:/msys64/mingw64/bin" - "C:/Libraries/llvm-12.0.0/bin") + "C:/Libraries/llvm-13.0.0/bin") if ("${LLVM_CONFIG_EXE}" STREQUAL "LLVM_CONFIG_EXE-NOTFOUND") message(FATAL_ERROR "unable to find llvm-config") @@ -54,23 +58,23 @@ if(ZIG_PREFER_CLANG_CPP_DYLIB) OUTPUT_VARIABLE LLVM_CONFIG_VERSION OUTPUT_STRIP_TRAILING_WHITESPACE) - if("${LLVM_CONFIG_VERSION}" VERSION_LESS 12) - message(FATAL_ERROR "expected LLVM 12.x but found ${LLVM_CONFIG_VERSION} using ${LLVM_CONFIG_EXE}") + if("${LLVM_CONFIG_VERSION}" VERSION_LESS 13) + message(FATAL_ERROR "expected LLVM 13.x but found ${LLVM_CONFIG_VERSION} using ${LLVM_CONFIG_EXE}") endif() - if("${LLVM_CONFIG_VERSION}" VERSION_EQUAL 13) - message(FATAL_ERROR "expected LLVM 12.x but found ${LLVM_CONFIG_VERSION} using ${LLVM_CONFIG_EXE}") + if("${LLVM_CONFIG_VERSION}" VERSION_EQUAL 14) + message(FATAL_ERROR "expected LLVM 13.x but found ${LLVM_CONFIG_VERSION} using ${LLVM_CONFIG_EXE}") endif() - if("${LLVM_CONFIG_VERSION}" VERSION_GREATER 13) - message(FATAL_ERROR "expected LLVM 12.x but found ${LLVM_CONFIG_VERSION} using ${LLVM_CONFIG_EXE}") + if("${LLVM_CONFIG_VERSION}" VERSION_GREATER 14) + message(FATAL_ERROR "expected LLVM 13.x but found ${LLVM_CONFIG_VERSION} using ${LLVM_CONFIG_EXE}") endif() elseif(ZIG_USE_LLVM_CONFIG) find_program(LLVM_CONFIG_EXE - NAMES llvm-config-12 llvm-config-12.0 llvm-config120 llvm-config12 llvm-config + NAMES llvm-config-13 llvm-config-13.0 llvm-config130 llvm-config13 llvm-config PATHS "/mingw64/bin" "/c/msys64/mingw64/bin" "c:/msys64/mingw64/bin" - "C:/Libraries/llvm-12.0.0/bin") + "C:/Libraries/llvm-13.0.0/bin") if ("${LLVM_CONFIG_EXE}" STREQUAL "LLVM_CONFIG_EXE-NOTFOUND") message(FATAL_ERROR "unable to find llvm-config") @@ -85,14 +89,14 @@ elseif(ZIG_USE_LLVM_CONFIG) OUTPUT_VARIABLE LLVM_CONFIG_VERSION OUTPUT_STRIP_TRAILING_WHITESPACE) - if("${LLVM_CONFIG_VERSION}" VERSION_LESS 12) - message(FATAL_ERROR "expected LLVM 12.x but found ${LLVM_CONFIG_VERSION} using ${LLVM_CONFIG_EXE}") + if("${LLVM_CONFIG_VERSION}" VERSION_LESS 13) + message(FATAL_ERROR "expected LLVM 13.x but found ${LLVM_CONFIG_VERSION} using ${LLVM_CONFIG_EXE}") endif() - if("${LLVM_CONFIG_VERSION}" VERSION_EQUAL 13) - message(FATAL_ERROR "expected LLVM 12.x but found ${LLVM_CONFIG_VERSION} using ${LLVM_CONFIG_EXE}") + if("${LLVM_CONFIG_VERSION}" VERSION_EQUAL 14) + message(FATAL_ERROR "expected LLVM 13.x but found ${LLVM_CONFIG_VERSION} using ${LLVM_CONFIG_EXE}") endif() - if("${LLVM_CONFIG_VERSION}" VERSION_GREATER 13) - message(FATAL_ERROR "expected LLVM 12.x but found ${LLVM_CONFIG_VERSION} using ${LLVM_CONFIG_EXE}") + if("${LLVM_CONFIG_VERSION}" VERSION_GREATER 14) + message(FATAL_ERROR "expected LLVM 13.x but found ${LLVM_CONFIG_VERSION} using ${LLVM_CONFIG_EXE}") endif() execute_process( @@ -166,7 +170,7 @@ elseif(ZIG_USE_LLVM_CONFIG) set(LLVM_LIBRARIES ${LLVM_LIBRARIES} ${LLVM_SYSTEM_LIBS}) if(NOT LLVM_LIBRARIES) - find_library(LLVM_LIBRARIES NAMES LLVM LLVM-12 LLVM-12.0) + find_library(LLVM_LIBRARIES NAMES LLVM LLVM-13 LLVM-13.0) endif() link_directories("${CMAKE_PREFIX_PATH}/lib") @@ -180,11 +184,13 @@ else() find_library(LLVM_${_prettylibname_}_LIB NAMES ${_libname_} PATHS ${LLVM_LIBDIRS} - /usr/lib/llvm/12/lib - /usr/lib/llvm-12/lib - /usr/lib/llvm-12.0/lib - /usr/local/llvm120/lib - /usr/local/llvm12/lib + /usr/lib/llvm/13/lib + /usr/lib/llvm-13/lib + /usr/lib/llvm-13.0/lib + /usr/local/llvm130/lib + /usr/local/llvm13/lib + /usr/local/opt/llvm@13/lib + /opt/homebrew/opt/llvm@13/lib /mingw64/lib /c/msys64/mingw64/lib c:\\msys64\\mingw64\\lib) @@ -213,6 +219,7 @@ else() FIND_AND_ADD_LLVM_LIB(LLVMWebAssemblyAsmParser) FIND_AND_ADD_LLVM_LIB(LLVMWebAssemblyCodeGen) FIND_AND_ADD_LLVM_LIB(LLVMWebAssemblyDesc) + FIND_AND_ADD_LLVM_LIB(LLVMWebAssemblyUtils) FIND_AND_ADD_LLVM_LIB(LLVMWebAssemblyInfo) FIND_AND_ADD_LLVM_LIB(LLVMSystemZDisassembler) FIND_AND_ADD_LLVM_LIB(LLVMSystemZAsmParser) @@ -288,11 +295,12 @@ else() FIND_AND_ADD_LLVM_LIB(LLVMOrcJIT) FIND_AND_ADD_LLVM_LIB(LLVMMCJIT) FIND_AND_ADD_LLVM_LIB(LLVMJITLink) - FIND_AND_ADD_LLVM_LIB(LLVMOrcTargetProcess) - FIND_AND_ADD_LLVM_LIB(LLVMOrcShared) FIND_AND_ADD_LLVM_LIB(LLVMInterpreter) FIND_AND_ADD_LLVM_LIB(LLVMExecutionEngine) FIND_AND_ADD_LLVM_LIB(LLVMRuntimeDyld) + FIND_AND_ADD_LLVM_LIB(LLVMOrcTargetProcess) + FIND_AND_ADD_LLVM_LIB(LLVMOrcShared) + FIND_AND_ADD_LLVM_LIB(LLVMDWP) FIND_AND_ADD_LLVM_LIB(LLVMSymbolize) FIND_AND_ADD_LLVM_LIB(LLVMDebugInfoPDB) FIND_AND_ADD_LLVM_LIB(LLVMDebugInfoGSYM) @@ -305,7 +313,6 @@ else() FIND_AND_ADD_LLVM_LIB(LLVMCFGuard) FIND_AND_ADD_LLVM_LIB(LLVMCoroutines) FIND_AND_ADD_LLVM_LIB(LLVMObjCARCOpts) - FIND_AND_ADD_LLVM_LIB(LLVMHelloNew) FIND_AND_ADD_LLVM_LIB(LLVMipo) FIND_AND_ADD_LLVM_LIB(LLVMVectorize) FIND_AND_ADD_LLVM_LIB(LLVMLinker) @@ -317,6 +324,7 @@ else() FIND_AND_ADD_LLVM_LIB(LLVMGlobalISel) FIND_AND_ADD_LLVM_LIB(LLVMMIRParser) FIND_AND_ADD_LLVM_LIB(LLVMAsmPrinter) + FIND_AND_ADD_LLVM_LIB(LLVMDebugInfoMSF) FIND_AND_ADD_LLVM_LIB(LLVMDebugInfoDWARF) FIND_AND_ADD_LLVM_LIB(LLVMSelectionDAG) FIND_AND_ADD_LLVM_LIB(LLVMCodeGen) @@ -338,7 +346,6 @@ else() FIND_AND_ADD_LLVM_LIB(LLVMMCParser) FIND_AND_ADD_LLVM_LIB(LLVMMC) FIND_AND_ADD_LLVM_LIB(LLVMDebugInfoCodeView) - FIND_AND_ADD_LLVM_LIB(LLVMDebugInfoMSF) FIND_AND_ADD_LLVM_LIB(LLVMBitReader) FIND_AND_ADD_LLVM_LIB(LLVMCore) FIND_AND_ADD_LLVM_LIB(LLVMRemarks) diff --git a/lib/include/__clang_cuda_device_functions.h b/lib/include/__clang_cuda_device_functions.h index f801e5426..cc4e1a4dd 100644 --- a/lib/include/__clang_cuda_device_functions.h +++ b/lib/include/__clang_cuda_device_functions.h @@ -34,10 +34,12 @@ __DEVICE__ unsigned long long __brevll(unsigned long long __a) { return __nv_brevll(__a); } #if defined(__cplusplus) -__DEVICE__ void __brkpt() { asm volatile("brkpt;"); } +__DEVICE__ void __brkpt() { __asm__ __volatile__("brkpt;"); } __DEVICE__ void __brkpt(int __a) { __brkpt(); } #else -__DEVICE__ void __attribute__((overloadable)) __brkpt(void) { asm volatile("brkpt;"); } +__DEVICE__ void __attribute__((overloadable)) __brkpt(void) { + __asm__ __volatile__("brkpt;"); +} __DEVICE__ void __attribute__((overloadable)) __brkpt(int __a) { __brkpt(); } #endif __DEVICE__ unsigned int __byte_perm(unsigned int __a, unsigned int __b, @@ -507,7 +509,7 @@ __DEVICE__ float __powf(float __a, float __b) { } // Parameter must have a known integer value. -#define __prof_trigger(__a) asm __volatile__("pmevent \t%0;" ::"i"(__a)) +#define __prof_trigger(__a) __asm__ __volatile__("pmevent \t%0;" ::"i"(__a)) __DEVICE__ int __rhadd(int __a, int __b) { return __nv_rhadd(__a, __b); } __DEVICE__ unsigned int __sad(int __a, int __b, unsigned int __c) { return __nv_sad(__a, __b, __c); @@ -526,7 +528,7 @@ __DEVICE__ float __tanf(float __a) { return __nv_fast_tanf(__a); } __DEVICE__ void __threadfence(void) { __nvvm_membar_gl(); } __DEVICE__ void __threadfence_block(void) { __nvvm_membar_cta(); }; __DEVICE__ void __threadfence_system(void) { __nvvm_membar_sys(); }; -__DEVICE__ void __trap(void) { asm volatile("trap;"); } +__DEVICE__ void __trap(void) { __asm__ __volatile__("trap;"); } __DEVICE__ unsigned int __uAtomicAdd(unsigned int *__p, unsigned int __v) { return __nvvm_atom_add_gen_i((int *)__p, __v); } @@ -1051,122 +1053,136 @@ __DEVICE__ unsigned int __bool2mask(unsigned int __a, int shift) { } __DEVICE__ unsigned int __vabs2(unsigned int __a) { unsigned int r; - asm("vabsdiff2.s32.s32.s32 %0,%1,%2,%3;" - : "=r"(r) - : "r"(__a), "r"(0), "r"(0)); + __asm__("vabsdiff2.s32.s32.s32 %0,%1,%2,%3;" + : "=r"(r) + : "r"(__a), "r"(0), "r"(0)); return r; } __DEVICE__ unsigned int __vabs4(unsigned int __a) { unsigned int r; - asm("vabsdiff4.s32.s32.s32 %0,%1,%2,%3;" - : "=r"(r) - : "r"(__a), "r"(0), "r"(0)); + __asm__("vabsdiff4.s32.s32.s32 %0,%1,%2,%3;" + : "=r"(r) + : "r"(__a), "r"(0), "r"(0)); return r; } __DEVICE__ unsigned int __vabsdiffs2(unsigned int __a, unsigned int __b) { unsigned int r; - asm("vabsdiff2.s32.s32.s32 %0,%1,%2,%3;" - : "=r"(r) - : "r"(__a), "r"(__b), "r"(0)); + __asm__("vabsdiff2.s32.s32.s32 %0,%1,%2,%3;" + : "=r"(r) + : "r"(__a), "r"(__b), "r"(0)); return r; } __DEVICE__ unsigned int __vabsdiffs4(unsigned int __a, unsigned int __b) { unsigned int r; - asm("vabsdiff4.s32.s32.s32 %0,%1,%2,%3;" - : "=r"(r) - : "r"(__a), "r"(__b), "r"(0)); + __asm__("vabsdiff4.s32.s32.s32 %0,%1,%2,%3;" + : "=r"(r) + : "r"(__a), "r"(__b), "r"(0)); return r; } __DEVICE__ unsigned int __vabsdiffu2(unsigned int __a, unsigned int __b) { unsigned int r; - asm("vabsdiff2.u32.u32.u32 %0,%1,%2,%3;" - : "=r"(r) - : "r"(__a), "r"(__b), "r"(0)); + __asm__("vabsdiff2.u32.u32.u32 %0,%1,%2,%3;" + : "=r"(r) + : "r"(__a), "r"(__b), "r"(0)); return r; } __DEVICE__ unsigned int __vabsdiffu4(unsigned int __a, unsigned int __b) { unsigned int r; - asm("vabsdiff4.u32.u32.u32 %0,%1,%2,%3;" - : "=r"(r) - : "r"(__a), "r"(__b), "r"(0)); + __asm__("vabsdiff4.u32.u32.u32 %0,%1,%2,%3;" + : "=r"(r) + : "r"(__a), "r"(__b), "r"(0)); return r; } __DEVICE__ unsigned int __vabsss2(unsigned int __a) { unsigned int r; - asm("vabsdiff2.s32.s32.s32.sat %0,%1,%2,%3;" - : "=r"(r) - : "r"(__a), "r"(0), "r"(0)); + __asm__("vabsdiff2.s32.s32.s32.sat %0,%1,%2,%3;" + : "=r"(r) + : "r"(__a), "r"(0), "r"(0)); return r; } __DEVICE__ unsigned int __vabsss4(unsigned int __a) { unsigned int r; - asm("vabsdiff4.s32.s32.s32.sat %0,%1,%2,%3;" - : "=r"(r) - : "r"(__a), "r"(0), "r"(0)); + __asm__("vabsdiff4.s32.s32.s32.sat %0,%1,%2,%3;" + : "=r"(r) + : "r"(__a), "r"(0), "r"(0)); return r; } __DEVICE__ unsigned int __vadd2(unsigned int __a, unsigned int __b) { unsigned int r; - asm("vadd2.u32.u32.u32 %0,%1,%2,%3;" : "=r"(r) : "r"(__a), "r"(__b), "r"(0)); + __asm__("vadd2.u32.u32.u32 %0,%1,%2,%3;" + : "=r"(r) + : "r"(__a), "r"(__b), "r"(0)); return r; } __DEVICE__ unsigned int __vadd4(unsigned int __a, unsigned int __b) { unsigned int r; - asm("vadd4.u32.u32.u32 %0,%1,%2,%3;" : "=r"(r) : "r"(__a), "r"(__b), "r"(0)); + __asm__("vadd4.u32.u32.u32 %0,%1,%2,%3;" + : "=r"(r) + : "r"(__a), "r"(__b), "r"(0)); return r; } __DEVICE__ unsigned int __vaddss2(unsigned int __a, unsigned int __b) { unsigned int r; - asm("vadd2.s32.s32.s32.sat %0,%1,%2,%3;" - : "=r"(r) - : "r"(__a), "r"(__b), "r"(0)); + __asm__("vadd2.s32.s32.s32.sat %0,%1,%2,%3;" + : "=r"(r) + : "r"(__a), "r"(__b), "r"(0)); return r; } __DEVICE__ unsigned int __vaddss4(unsigned int __a, unsigned int __b) { unsigned int r; - asm("vadd4.s32.s32.s32.sat %0,%1,%2,%3;" - : "=r"(r) - : "r"(__a), "r"(__b), "r"(0)); + __asm__("vadd4.s32.s32.s32.sat %0,%1,%2,%3;" + : "=r"(r) + : "r"(__a), "r"(__b), "r"(0)); return r; } __DEVICE__ unsigned int __vaddus2(unsigned int __a, unsigned int __b) { unsigned int r; - asm("vadd2.u32.u32.u32.sat %0,%1,%2,%3;" - : "=r"(r) - : "r"(__a), "r"(__b), "r"(0)); + __asm__("vadd2.u32.u32.u32.sat %0,%1,%2,%3;" + : "=r"(r) + : "r"(__a), "r"(__b), "r"(0)); return r; } __DEVICE__ unsigned int __vaddus4(unsigned int __a, unsigned int __b) { unsigned int r; - asm("vadd4.u32.u32.u32.sat %0,%1,%2,%3;" - : "=r"(r) - : "r"(__a), "r"(__b), "r"(0)); + __asm__("vadd4.u32.u32.u32.sat %0,%1,%2,%3;" + : "=r"(r) + : "r"(__a), "r"(__b), "r"(0)); return r; } __DEVICE__ unsigned int __vavgs2(unsigned int __a, unsigned int __b) { unsigned int r; - asm("vavrg2.s32.s32.s32 %0,%1,%2,%3;" : "=r"(r) : "r"(__a), "r"(__b), "r"(0)); + __asm__("vavrg2.s32.s32.s32 %0,%1,%2,%3;" + : "=r"(r) + : "r"(__a), "r"(__b), "r"(0)); return r; } __DEVICE__ unsigned int __vavgs4(unsigned int __a, unsigned int __b) { unsigned int r; - asm("vavrg4.s32.s32.s32 %0,%1,%2,%3;" : "=r"(r) : "r"(__a), "r"(__b), "r"(0)); + __asm__("vavrg4.s32.s32.s32 %0,%1,%2,%3;" + : "=r"(r) + : "r"(__a), "r"(__b), "r"(0)); return r; } __DEVICE__ unsigned int __vavgu2(unsigned int __a, unsigned int __b) { unsigned int r; - asm("vavrg2.u32.u32.u32 %0,%1,%2,%3;" : "=r"(r) : "r"(__a), "r"(__b), "r"(0)); + __asm__("vavrg2.u32.u32.u32 %0,%1,%2,%3;" + : "=r"(r) + : "r"(__a), "r"(__b), "r"(0)); return r; } __DEVICE__ unsigned int __vavgu4(unsigned int __a, unsigned int __b) { unsigned int r; - asm("vavrg4.u32.u32.u32 %0,%1,%2,%3;" : "=r"(r) : "r"(__a), "r"(__b), "r"(0)); + __asm__("vavrg4.u32.u32.u32 %0,%1,%2,%3;" + : "=r"(r) + : "r"(__a), "r"(__b), "r"(0)); return r; } __DEVICE__ unsigned int __vseteq2(unsigned int __a, unsigned int __b) { unsigned int r; - asm("vset2.u32.u32.eq %0,%1,%2,%3;" : "=r"(r) : "r"(__a), "r"(__b), "r"(0)); + __asm__("vset2.u32.u32.eq %0,%1,%2,%3;" + : "=r"(r) + : "r"(__a), "r"(__b), "r"(0)); return r; } __DEVICE__ unsigned int __vcmpeq2(unsigned int __a, unsigned int __b) { @@ -1174,7 +1190,9 @@ __DEVICE__ unsigned int __vcmpeq2(unsigned int __a, unsigned int __b) { } __DEVICE__ unsigned int __vseteq4(unsigned int __a, unsigned int __b) { unsigned int r; - asm("vset4.u32.u32.eq %0,%1,%2,%3;" : "=r"(r) : "r"(__a), "r"(__b), "r"(0)); + __asm__("vset4.u32.u32.eq %0,%1,%2,%3;" + : "=r"(r) + : "r"(__a), "r"(__b), "r"(0)); return r; } __DEVICE__ unsigned int __vcmpeq4(unsigned int __a, unsigned int __b) { @@ -1182,7 +1200,9 @@ __DEVICE__ unsigned int __vcmpeq4(unsigned int __a, unsigned int __b) { } __DEVICE__ unsigned int __vsetges2(unsigned int __a, unsigned int __b) { unsigned int r; - asm("vset2.s32.s32.ge %0,%1,%2,%3;" : "=r"(r) : "r"(__a), "r"(__b), "r"(0)); + __asm__("vset2.s32.s32.ge %0,%1,%2,%3;" + : "=r"(r) + : "r"(__a), "r"(__b), "r"(0)); return r; } __DEVICE__ unsigned int __vcmpges2(unsigned int __a, unsigned int __b) { @@ -1190,7 +1210,9 @@ __DEVICE__ unsigned int __vcmpges2(unsigned int __a, unsigned int __b) { } __DEVICE__ unsigned int __vsetges4(unsigned int __a, unsigned int __b) { unsigned int r; - asm("vset4.s32.s32.ge %0,%1,%2,%3;" : "=r"(r) : "r"(__a), "r"(__b), "r"(0)); + __asm__("vset4.s32.s32.ge %0,%1,%2,%3;" + : "=r"(r) + : "r"(__a), "r"(__b), "r"(0)); return r; } __DEVICE__ unsigned int __vcmpges4(unsigned int __a, unsigned int __b) { @@ -1198,7 +1220,9 @@ __DEVICE__ unsigned int __vcmpges4(unsigned int __a, unsigned int __b) { } __DEVICE__ unsigned int __vsetgeu2(unsigned int __a, unsigned int __b) { unsigned int r; - asm("vset2.u32.u32.ge %0,%1,%2,%3;" : "=r"(r) : "r"(__a), "r"(__b), "r"(0)); + __asm__("vset2.u32.u32.ge %0,%1,%2,%3;" + : "=r"(r) + : "r"(__a), "r"(__b), "r"(0)); return r; } __DEVICE__ unsigned int __vcmpgeu2(unsigned int __a, unsigned int __b) { @@ -1206,7 +1230,9 @@ __DEVICE__ unsigned int __vcmpgeu2(unsigned int __a, unsigned int __b) { } __DEVICE__ unsigned int __vsetgeu4(unsigned int __a, unsigned int __b) { unsigned int r; - asm("vset4.u32.u32.ge %0,%1,%2,%3;" : "=r"(r) : "r"(__a), "r"(__b), "r"(0)); + __asm__("vset4.u32.u32.ge %0,%1,%2,%3;" + : "=r"(r) + : "r"(__a), "r"(__b), "r"(0)); return r; } __DEVICE__ unsigned int __vcmpgeu4(unsigned int __a, unsigned int __b) { @@ -1214,7 +1240,9 @@ __DEVICE__ unsigned int __vcmpgeu4(unsigned int __a, unsigned int __b) { } __DEVICE__ unsigned int __vsetgts2(unsigned int __a, unsigned int __b) { unsigned int r; - asm("vset2.s32.s32.gt %0,%1,%2,%3;" : "=r"(r) : "r"(__a), "r"(__b), "r"(0)); + __asm__("vset2.s32.s32.gt %0,%1,%2,%3;" + : "=r"(r) + : "r"(__a), "r"(__b), "r"(0)); return r; } __DEVICE__ unsigned int __vcmpgts2(unsigned int __a, unsigned int __b) { @@ -1222,7 +1250,9 @@ __DEVICE__ unsigned int __vcmpgts2(unsigned int __a, unsigned int __b) { } __DEVICE__ unsigned int __vsetgts4(unsigned int __a, unsigned int __b) { unsigned int r; - asm("vset4.s32.s32.gt %0,%1,%2,%3;" : "=r"(r) : "r"(__a), "r"(__b), "r"(0)); + __asm__("vset4.s32.s32.gt %0,%1,%2,%3;" + : "=r"(r) + : "r"(__a), "r"(__b), "r"(0)); return r; } __DEVICE__ unsigned int __vcmpgts4(unsigned int __a, unsigned int __b) { @@ -1230,7 +1260,9 @@ __DEVICE__ unsigned int __vcmpgts4(unsigned int __a, unsigned int __b) { } __DEVICE__ unsigned int __vsetgtu2(unsigned int __a, unsigned int __b) { unsigned int r; - asm("vset2.u32.u32.gt %0,%1,%2,%3;" : "=r"(r) : "r"(__a), "r"(__b), "r"(0)); + __asm__("vset2.u32.u32.gt %0,%1,%2,%3;" + : "=r"(r) + : "r"(__a), "r"(__b), "r"(0)); return r; } __DEVICE__ unsigned int __vcmpgtu2(unsigned int __a, unsigned int __b) { @@ -1238,7 +1270,9 @@ __DEVICE__ unsigned int __vcmpgtu2(unsigned int __a, unsigned int __b) { } __DEVICE__ unsigned int __vsetgtu4(unsigned int __a, unsigned int __b) { unsigned int r; - asm("vset4.u32.u32.gt %0,%1,%2,%3;" : "=r"(r) : "r"(__a), "r"(__b), "r"(0)); + __asm__("vset4.u32.u32.gt %0,%1,%2,%3;" + : "=r"(r) + : "r"(__a), "r"(__b), "r"(0)); return r; } __DEVICE__ unsigned int __vcmpgtu4(unsigned int __a, unsigned int __b) { @@ -1246,7 +1280,9 @@ __DEVICE__ unsigned int __vcmpgtu4(unsigned int __a, unsigned int __b) { } __DEVICE__ unsigned int __vsetles2(unsigned int __a, unsigned int __b) { unsigned int r; - asm("vset2.s32.s32.le %0,%1,%2,%3;" : "=r"(r) : "r"(__a), "r"(__b), "r"(0)); + __asm__("vset2.s32.s32.le %0,%1,%2,%3;" + : "=r"(r) + : "r"(__a), "r"(__b), "r"(0)); return r; } __DEVICE__ unsigned int __vcmples2(unsigned int __a, unsigned int __b) { @@ -1254,7 +1290,9 @@ __DEVICE__ unsigned int __vcmples2(unsigned int __a, unsigned int __b) { } __DEVICE__ unsigned int __vsetles4(unsigned int __a, unsigned int __b) { unsigned int r; - asm("vset4.s32.s32.le %0,%1,%2,%3;" : "=r"(r) : "r"(__a), "r"(__b), "r"(0)); + __asm__("vset4.s32.s32.le %0,%1,%2,%3;" + : "=r"(r) + : "r"(__a), "r"(__b), "r"(0)); return r; } __DEVICE__ unsigned int __vcmples4(unsigned int __a, unsigned int __b) { @@ -1262,7 +1300,9 @@ __DEVICE__ unsigned int __vcmples4(unsigned int __a, unsigned int __b) { } __DEVICE__ unsigned int __vsetleu2(unsigned int __a, unsigned int __b) { unsigned int r; - asm("vset2.u32.u32.le %0,%1,%2,%3;" : "=r"(r) : "r"(__a), "r"(__b), "r"(0)); + __asm__("vset2.u32.u32.le %0,%1,%2,%3;" + : "=r"(r) + : "r"(__a), "r"(__b), "r"(0)); return r; } __DEVICE__ unsigned int __vcmpleu2(unsigned int __a, unsigned int __b) { @@ -1270,7 +1310,9 @@ __DEVICE__ unsigned int __vcmpleu2(unsigned int __a, unsigned int __b) { } __DEVICE__ unsigned int __vsetleu4(unsigned int __a, unsigned int __b) { unsigned int r; - asm("vset4.u32.u32.le %0,%1,%2,%3;" : "=r"(r) : "r"(__a), "r"(__b), "r"(0)); + __asm__("vset4.u32.u32.le %0,%1,%2,%3;" + : "=r"(r) + : "r"(__a), "r"(__b), "r"(0)); return r; } __DEVICE__ unsigned int __vcmpleu4(unsigned int __a, unsigned int __b) { @@ -1278,7 +1320,9 @@ __DEVICE__ unsigned int __vcmpleu4(unsigned int __a, unsigned int __b) { } __DEVICE__ unsigned int __vsetlts2(unsigned int __a, unsigned int __b) { unsigned int r; - asm("vset2.s32.s32.lt %0,%1,%2,%3;" : "=r"(r) : "r"(__a), "r"(__b), "r"(0)); + __asm__("vset2.s32.s32.lt %0,%1,%2,%3;" + : "=r"(r) + : "r"(__a), "r"(__b), "r"(0)); return r; } __DEVICE__ unsigned int __vcmplts2(unsigned int __a, unsigned int __b) { @@ -1286,7 +1330,9 @@ __DEVICE__ unsigned int __vcmplts2(unsigned int __a, unsigned int __b) { } __DEVICE__ unsigned int __vsetlts4(unsigned int __a, unsigned int __b) { unsigned int r; - asm("vset4.s32.s32.lt %0,%1,%2,%3;" : "=r"(r) : "r"(__a), "r"(__b), "r"(0)); + __asm__("vset4.s32.s32.lt %0,%1,%2,%3;" + : "=r"(r) + : "r"(__a), "r"(__b), "r"(0)); return r; } __DEVICE__ unsigned int __vcmplts4(unsigned int __a, unsigned int __b) { @@ -1294,7 +1340,9 @@ __DEVICE__ unsigned int __vcmplts4(unsigned int __a, unsigned int __b) { } __DEVICE__ unsigned int __vsetltu2(unsigned int __a, unsigned int __b) { unsigned int r; - asm("vset2.u32.u32.lt %0,%1,%2,%3;" : "=r"(r) : "r"(__a), "r"(__b), "r"(0)); + __asm__("vset2.u32.u32.lt %0,%1,%2,%3;" + : "=r"(r) + : "r"(__a), "r"(__b), "r"(0)); return r; } __DEVICE__ unsigned int __vcmpltu2(unsigned int __a, unsigned int __b) { @@ -1302,7 +1350,9 @@ __DEVICE__ unsigned int __vcmpltu2(unsigned int __a, unsigned int __b) { } __DEVICE__ unsigned int __vsetltu4(unsigned int __a, unsigned int __b) { unsigned int r; - asm("vset4.u32.u32.lt %0,%1,%2,%3;" : "=r"(r) : "r"(__a), "r"(__b), "r"(0)); + __asm__("vset4.u32.u32.lt %0,%1,%2,%3;" + : "=r"(r) + : "r"(__a), "r"(__b), "r"(0)); return r; } __DEVICE__ unsigned int __vcmpltu4(unsigned int __a, unsigned int __b) { @@ -1310,7 +1360,9 @@ __DEVICE__ unsigned int __vcmpltu4(unsigned int __a, unsigned int __b) { } __DEVICE__ unsigned int __vsetne2(unsigned int __a, unsigned int __b) { unsigned int r; - asm("vset2.u32.u32.ne %0,%1,%2,%3;" : "=r"(r) : "r"(__a), "r"(__b), "r"(0)); + __asm__("vset2.u32.u32.ne %0,%1,%2,%3;" + : "=r"(r) + : "r"(__a), "r"(__b), "r"(0)); return r; } __DEVICE__ unsigned int __vcmpne2(unsigned int __a, unsigned int __b) { @@ -1318,7 +1370,9 @@ __DEVICE__ unsigned int __vcmpne2(unsigned int __a, unsigned int __b) { } __DEVICE__ unsigned int __vsetne4(unsigned int __a, unsigned int __b) { unsigned int r; - asm("vset4.u32.u32.ne %0,%1,%2,%3;" : "=r"(r) : "r"(__a), "r"(__b), "r"(0)); + __asm__("vset4.u32.u32.ne %0,%1,%2,%3;" + : "=r"(r) + : "r"(__a), "r"(__b), "r"(0)); return r; } __DEVICE__ unsigned int __vcmpne4(unsigned int __a, unsigned int __b) { @@ -1345,94 +1399,112 @@ __DEVICE__ unsigned int __vmaxs2(unsigned int __a, unsigned int __b) { unsigned mask = __vcmpgts2(__a, __b); r = (__a & mask) | (__b & ~mask); } else { - asm("vmax2.s32.s32.s32 %0,%1,%2,%3;" - : "=r"(r) - : "r"(__a), "r"(__b), "r"(0)); + __asm__("vmax2.s32.s32.s32 %0,%1,%2,%3;" + : "=r"(r) + : "r"(__a), "r"(__b), "r"(0)); } return r; } __DEVICE__ unsigned int __vmaxs4(unsigned int __a, unsigned int __b) { unsigned int r; - asm("vmax4.s32.s32.s32 %0,%1,%2,%3;" : "=r"(r) : "r"(__a), "r"(__b), "r"(0)); + __asm__("vmax4.s32.s32.s32 %0,%1,%2,%3;" + : "=r"(r) + : "r"(__a), "r"(__b), "r"(0)); return r; } __DEVICE__ unsigned int __vmaxu2(unsigned int __a, unsigned int __b) { unsigned int r; - asm("vmax2.u32.u32.u32 %0,%1,%2,%3;" : "=r"(r) : "r"(__a), "r"(__b), "r"(0)); + __asm__("vmax2.u32.u32.u32 %0,%1,%2,%3;" + : "=r"(r) + : "r"(__a), "r"(__b), "r"(0)); return r; } __DEVICE__ unsigned int __vmaxu4(unsigned int __a, unsigned int __b) { unsigned int r; - asm("vmax4.u32.u32.u32 %0,%1,%2,%3;" : "=r"(r) : "r"(__a), "r"(__b), "r"(0)); + __asm__("vmax4.u32.u32.u32 %0,%1,%2,%3;" + : "=r"(r) + : "r"(__a), "r"(__b), "r"(0)); return r; } __DEVICE__ unsigned int __vmins2(unsigned int __a, unsigned int __b) { unsigned int r; - asm("vmin2.s32.s32.s32 %0,%1,%2,%3;" : "=r"(r) : "r"(__a), "r"(__b), "r"(0)); + __asm__("vmin2.s32.s32.s32 %0,%1,%2,%3;" + : "=r"(r) + : "r"(__a), "r"(__b), "r"(0)); return r; } __DEVICE__ unsigned int __vmins4(unsigned int __a, unsigned int __b) { unsigned int r; - asm("vmin4.s32.s32.s32 %0,%1,%2,%3;" : "=r"(r) : "r"(__a), "r"(__b), "r"(0)); + __asm__("vmin4.s32.s32.s32 %0,%1,%2,%3;" + : "=r"(r) + : "r"(__a), "r"(__b), "r"(0)); return r; } __DEVICE__ unsigned int __vminu2(unsigned int __a, unsigned int __b) { unsigned int r; - asm("vmin2.u32.u32.u32 %0,%1,%2,%3;" : "=r"(r) : "r"(__a), "r"(__b), "r"(0)); + __asm__("vmin2.u32.u32.u32 %0,%1,%2,%3;" + : "=r"(r) + : "r"(__a), "r"(__b), "r"(0)); return r; } __DEVICE__ unsigned int __vminu4(unsigned int __a, unsigned int __b) { unsigned int r; - asm("vmin4.u32.u32.u32 %0,%1,%2,%3;" : "=r"(r) : "r"(__a), "r"(__b), "r"(0)); + __asm__("vmin4.u32.u32.u32 %0,%1,%2,%3;" + : "=r"(r) + : "r"(__a), "r"(__b), "r"(0)); return r; } __DEVICE__ unsigned int __vsads2(unsigned int __a, unsigned int __b) { unsigned int r; - asm("vabsdiff2.s32.s32.s32.add %0,%1,%2,%3;" - : "=r"(r) - : "r"(__a), "r"(__b), "r"(0)); + __asm__("vabsdiff2.s32.s32.s32.add %0,%1,%2,%3;" + : "=r"(r) + : "r"(__a), "r"(__b), "r"(0)); return r; } __DEVICE__ unsigned int __vsads4(unsigned int __a, unsigned int __b) { unsigned int r; - asm("vabsdiff4.s32.s32.s32.add %0,%1,%2,%3;" - : "=r"(r) - : "r"(__a), "r"(__b), "r"(0)); + __asm__("vabsdiff4.s32.s32.s32.add %0,%1,%2,%3;" + : "=r"(r) + : "r"(__a), "r"(__b), "r"(0)); return r; } __DEVICE__ unsigned int __vsadu2(unsigned int __a, unsigned int __b) { unsigned int r; - asm("vabsdiff2.u32.u32.u32.add %0,%1,%2,%3;" - : "=r"(r) - : "r"(__a), "r"(__b), "r"(0)); + __asm__("vabsdiff2.u32.u32.u32.add %0,%1,%2,%3;" + : "=r"(r) + : "r"(__a), "r"(__b), "r"(0)); return r; } __DEVICE__ unsigned int __vsadu4(unsigned int __a, unsigned int __b) { unsigned int r; - asm("vabsdiff4.u32.u32.u32.add %0,%1,%2,%3;" - : "=r"(r) - : "r"(__a), "r"(__b), "r"(0)); + __asm__("vabsdiff4.u32.u32.u32.add %0,%1,%2,%3;" + : "=r"(r) + : "r"(__a), "r"(__b), "r"(0)); return r; } __DEVICE__ unsigned int __vsub2(unsigned int __a, unsigned int __b) { unsigned int r; - asm("vsub2.u32.u32.u32 %0,%1,%2,%3;" : "=r"(r) : "r"(__a), "r"(__b), "r"(0)); + __asm__("vsub2.u32.u32.u32 %0,%1,%2,%3;" + : "=r"(r) + : "r"(__a), "r"(__b), "r"(0)); return r; } __DEVICE__ unsigned int __vneg2(unsigned int __a) { return __vsub2(0, __a); } __DEVICE__ unsigned int __vsub4(unsigned int __a, unsigned int __b) { unsigned int r; - asm("vsub4.u32.u32.u32 %0,%1,%2,%3;" : "=r"(r) : "r"(__a), "r"(__b), "r"(0)); + __asm__("vsub4.u32.u32.u32 %0,%1,%2,%3;" + : "=r"(r) + : "r"(__a), "r"(__b), "r"(0)); return r; } __DEVICE__ unsigned int __vneg4(unsigned int __a) { return __vsub4(0, __a); } __DEVICE__ unsigned int __vsubss2(unsigned int __a, unsigned int __b) { unsigned int r; - asm("vsub2.s32.s32.s32.sat %0,%1,%2,%3;" - : "=r"(r) - : "r"(__a), "r"(__b), "r"(0)); + __asm__("vsub2.s32.s32.s32.sat %0,%1,%2,%3;" + : "=r"(r) + : "r"(__a), "r"(__b), "r"(0)); return r; } __DEVICE__ unsigned int __vnegss2(unsigned int __a) { @@ -1440,9 +1512,9 @@ __DEVICE__ unsigned int __vnegss2(unsigned int __a) { } __DEVICE__ unsigned int __vsubss4(unsigned int __a, unsigned int __b) { unsigned int r; - asm("vsub4.s32.s32.s32.sat %0,%1,%2,%3;" - : "=r"(r) - : "r"(__a), "r"(__b), "r"(0)); + __asm__("vsub4.s32.s32.s32.sat %0,%1,%2,%3;" + : "=r"(r) + : "r"(__a), "r"(__b), "r"(0)); return r; } __DEVICE__ unsigned int __vnegss4(unsigned int __a) { @@ -1450,16 +1522,16 @@ __DEVICE__ unsigned int __vnegss4(unsigned int __a) { } __DEVICE__ unsigned int __vsubus2(unsigned int __a, unsigned int __b) { unsigned int r; - asm("vsub2.u32.u32.u32.sat %0,%1,%2,%3;" - : "=r"(r) - : "r"(__a), "r"(__b), "r"(0)); + __asm__("vsub2.u32.u32.u32.sat %0,%1,%2,%3;" + : "=r"(r) + : "r"(__a), "r"(__b), "r"(0)); return r; } __DEVICE__ unsigned int __vsubus4(unsigned int __a, unsigned int __b) { unsigned int r; - asm("vsub4.u32.u32.u32.sat %0,%1,%2,%3;" - : "=r"(r) - : "r"(__a), "r"(__b), "r"(0)); + __asm__("vsub4.u32.u32.u32.sat %0,%1,%2,%3;" + : "=r"(r) + : "r"(__a), "r"(__b), "r"(0)); return r; } #endif // CUDA_VERSION >= 9020 diff --git a/lib/include/__clang_cuda_math.h b/lib/include/__clang_cuda_math.h index acb26ad34..538556f39 100644 --- a/lib/include/__clang_cuda_math.h +++ b/lib/include/__clang_cuda_math.h @@ -166,6 +166,8 @@ __DEVICE__ long long llrint(double __a) { return __nv_llrint(__a); } __DEVICE__ long long llrintf(float __a) { return __nv_llrintf(__a); } __DEVICE__ long long llround(double __a) { return __nv_llround(__a); } __DEVICE__ long long llroundf(float __a) { return __nv_llroundf(__a); } +__DEVICE__ double round(double __a) { return __nv_round(__a); } +__DEVICE__ float roundf(float __a) { return __nv_roundf(__a); } __DEVICE__ double log(double __a) { return __nv_log(__a); } __DEVICE__ double log10(double __a) { return __nv_log10(__a); } __DEVICE__ float log10f(float __a) { return __nv_log10f(__a); } @@ -270,8 +272,6 @@ __DEVICE__ float rnorm4df(float __a, float __b, float __c, float __d) { __DEVICE__ float rnormf(int __dim, const float *__t) { return __nv_rnormf(__dim, __t); } -__DEVICE__ double round(double __a) { return __nv_round(__a); } -__DEVICE__ float roundf(float __a) { return __nv_roundf(__a); } __DEVICE__ double rsqrt(double __a) { return __nv_rsqrt(__a); } __DEVICE__ float rsqrtf(float __a) { return __nv_rsqrtf(__a); } __DEVICE__ double scalbn(double __a, int __b) { return __nv_scalbn(__a, __b); } diff --git a/lib/include/__clang_cuda_runtime_wrapper.h b/lib/include/__clang_cuda_runtime_wrapper.h index f88c39a9b..f401964bd 100644 --- a/lib/include/__clang_cuda_runtime_wrapper.h +++ b/lib/include/__clang_cuda_runtime_wrapper.h @@ -349,9 +349,14 @@ extern "C" { __device__ int vprintf(const char *, const char *); __device__ void free(void *) __attribute((nothrow)); __device__ void *malloc(size_t) __attribute((nothrow)) __attribute__((malloc)); + +// __assertfail() used to have a `noreturn` attribute. Unfortunately that +// contributed to triggering the longstanding bug in ptxas when assert was used +// in sufficiently convoluted code. See +// https://bugs.llvm.org/show_bug.cgi?id=27738 for the details. __device__ void __assertfail(const char *__message, const char *__file, unsigned __line, const char *__function, - size_t __charSize) __attribute__((noreturn)); + size_t __charSize); // In order for standard assert() macro on linux to work we need to // provide device-side __assert_fail() diff --git a/lib/include/__clang_hip_cmath.h b/lib/include/__clang_hip_cmath.h index cd22a2df9..d488db0a9 100644 --- a/lib/include/__clang_hip_cmath.h +++ b/lib/include/__clang_hip_cmath.h @@ -10,10 +10,11 @@ #ifndef __CLANG_HIP_CMATH_H__ #define __CLANG_HIP_CMATH_H__ -#if !defined(__HIP__) +#if !defined(__HIP__) && !defined(__OPENMP_AMDGCN__) #error "This file is for HIP and OpenMP AMDGCN device compilation only." #endif +#if !defined(__HIPCC_RTC__) #if defined(__cplusplus) #include #include @@ -21,102 +22,162 @@ #endif #include #include +#endif // !defined(__HIPCC_RTC__) #pragma push_macro("__DEVICE__") +#pragma push_macro("__CONSTEXPR__") +#ifdef __OPENMP_AMDGCN__ +#define __DEVICE__ static __attribute__((always_inline, nothrow)) +#define __CONSTEXPR__ constexpr +#else #define __DEVICE__ static __device__ inline __attribute__((always_inline)) +#define __CONSTEXPR__ +#endif // __OPENMP_AMDGCN__ // Start with functions that cannot be defined by DEF macros below. #if defined(__cplusplus) -__DEVICE__ double abs(double __x) { return ::fabs(__x); } -__DEVICE__ float abs(float __x) { return ::fabsf(__x); } -__DEVICE__ long long abs(long long __n) { return ::llabs(__n); } -__DEVICE__ long abs(long __n) { return ::labs(__n); } -__DEVICE__ float fma(float __x, float __y, float __z) { +#if defined __OPENMP_AMDGCN__ +__DEVICE__ __CONSTEXPR__ float fabs(float __x) { return ::fabsf(__x); } +__DEVICE__ __CONSTEXPR__ float sin(float __x) { return ::sinf(__x); } +__DEVICE__ __CONSTEXPR__ float cos(float __x) { return ::cosf(__x); } +#endif +__DEVICE__ __CONSTEXPR__ double abs(double __x) { return ::fabs(__x); } +__DEVICE__ __CONSTEXPR__ float abs(float __x) { return ::fabsf(__x); } +__DEVICE__ __CONSTEXPR__ long long abs(long long __n) { return ::llabs(__n); } +__DEVICE__ __CONSTEXPR__ long abs(long __n) { return ::labs(__n); } +__DEVICE__ __CONSTEXPR__ float fma(float __x, float __y, float __z) { return ::fmaf(__x, __y, __z); } -__DEVICE__ int fpclassify(float __x) { +#if !defined(__HIPCC_RTC__) +// The value returned by fpclassify is platform dependent, therefore it is not +// supported by hipRTC. +__DEVICE__ __CONSTEXPR__ int fpclassify(float __x) { return __builtin_fpclassify(FP_NAN, FP_INFINITE, FP_NORMAL, FP_SUBNORMAL, FP_ZERO, __x); } -__DEVICE__ int fpclassify(double __x) { +__DEVICE__ __CONSTEXPR__ int fpclassify(double __x) { return __builtin_fpclassify(FP_NAN, FP_INFINITE, FP_NORMAL, FP_SUBNORMAL, FP_ZERO, __x); } -__DEVICE__ float frexp(float __arg, int *__exp) { +#endif // !defined(__HIPCC_RTC__) + +__DEVICE__ __CONSTEXPR__ float frexp(float __arg, int *__exp) { return ::frexpf(__arg, __exp); } -__DEVICE__ bool isfinite(float __x) { return ::__finitef(__x); } -__DEVICE__ bool isfinite(double __x) { return ::__finite(__x); } -__DEVICE__ bool isgreater(float __x, float __y) { + +#if defined(__OPENMP_AMDGCN__) +// For OpenMP we work around some old system headers that have non-conforming +// `isinf(float)` and `isnan(float)` implementations that return an `int`. We do +// this by providing two versions of these functions, differing only in the +// return type. To avoid conflicting definitions we disable implicit base +// function generation. That means we will end up with two specializations, one +// per type, but only one has a base function defined by the system header. +#pragma omp begin declare variant match( \ + implementation = {extension(disable_implicit_base)}) + +// FIXME: We lack an extension to customize the mangling of the variants, e.g., +// add a suffix. This means we would clash with the names of the variants +// (note that we do not create implicit base functions here). To avoid +// this clash we add a new trait to some of them that is always true +// (this is LLVM after all ;)). It will only influence the mangled name +// of the variants inside the inner region and avoid the clash. +#pragma omp begin declare variant match(implementation = {vendor(llvm)}) + +__DEVICE__ __CONSTEXPR__ int isinf(float __x) { return ::__isinff(__x); } +__DEVICE__ __CONSTEXPR__ int isinf(double __x) { return ::__isinf(__x); } +__DEVICE__ __CONSTEXPR__ int isfinite(float __x) { return ::__finitef(__x); } +__DEVICE__ __CONSTEXPR__ int isfinite(double __x) { return ::__finite(__x); } +__DEVICE__ __CONSTEXPR__ int isnan(float __x) { return ::__isnanf(__x); } +__DEVICE__ __CONSTEXPR__ int isnan(double __x) { return ::__isnan(__x); } + +#pragma omp end declare variant +#endif // defined(__OPENMP_AMDGCN__) + +__DEVICE__ __CONSTEXPR__ bool isinf(float __x) { return ::__isinff(__x); } +__DEVICE__ __CONSTEXPR__ bool isinf(double __x) { return ::__isinf(__x); } +__DEVICE__ __CONSTEXPR__ bool isfinite(float __x) { return ::__finitef(__x); } +__DEVICE__ __CONSTEXPR__ bool isfinite(double __x) { return ::__finite(__x); } +__DEVICE__ __CONSTEXPR__ bool isnan(float __x) { return ::__isnanf(__x); } +__DEVICE__ __CONSTEXPR__ bool isnan(double __x) { return ::__isnan(__x); } + +#if defined(__OPENMP_AMDGCN__) +#pragma omp end declare variant +#endif // defined(__OPENMP_AMDGCN__) + +__DEVICE__ __CONSTEXPR__ bool isgreater(float __x, float __y) { return __builtin_isgreater(__x, __y); } -__DEVICE__ bool isgreater(double __x, double __y) { +__DEVICE__ __CONSTEXPR__ bool isgreater(double __x, double __y) { return __builtin_isgreater(__x, __y); } -__DEVICE__ bool isgreaterequal(float __x, float __y) { +__DEVICE__ __CONSTEXPR__ bool isgreaterequal(float __x, float __y) { return __builtin_isgreaterequal(__x, __y); } -__DEVICE__ bool isgreaterequal(double __x, double __y) { +__DEVICE__ __CONSTEXPR__ bool isgreaterequal(double __x, double __y) { return __builtin_isgreaterequal(__x, __y); } -__DEVICE__ bool isinf(float __x) { return ::__isinff(__x); } -__DEVICE__ bool isinf(double __x) { return ::__isinf(__x); } -__DEVICE__ bool isless(float __x, float __y) { +__DEVICE__ __CONSTEXPR__ bool isless(float __x, float __y) { return __builtin_isless(__x, __y); } -__DEVICE__ bool isless(double __x, double __y) { +__DEVICE__ __CONSTEXPR__ bool isless(double __x, double __y) { return __builtin_isless(__x, __y); } -__DEVICE__ bool islessequal(float __x, float __y) { +__DEVICE__ __CONSTEXPR__ bool islessequal(float __x, float __y) { return __builtin_islessequal(__x, __y); } -__DEVICE__ bool islessequal(double __x, double __y) { +__DEVICE__ __CONSTEXPR__ bool islessequal(double __x, double __y) { return __builtin_islessequal(__x, __y); } -__DEVICE__ bool islessgreater(float __x, float __y) { +__DEVICE__ __CONSTEXPR__ bool islessgreater(float __x, float __y) { return __builtin_islessgreater(__x, __y); } -__DEVICE__ bool islessgreater(double __x, double __y) { +__DEVICE__ __CONSTEXPR__ bool islessgreater(double __x, double __y) { return __builtin_islessgreater(__x, __y); } -__DEVICE__ bool isnan(float __x) { return ::__isnanf(__x); } -__DEVICE__ bool isnan(double __x) { return ::__isnan(__x); } -__DEVICE__ bool isnormal(float __x) { return __builtin_isnormal(__x); } -__DEVICE__ bool isnormal(double __x) { return __builtin_isnormal(__x); } -__DEVICE__ bool isunordered(float __x, float __y) { +__DEVICE__ __CONSTEXPR__ bool isnormal(float __x) { + return __builtin_isnormal(__x); +} +__DEVICE__ __CONSTEXPR__ bool isnormal(double __x) { + return __builtin_isnormal(__x); +} +__DEVICE__ __CONSTEXPR__ bool isunordered(float __x, float __y) { return __builtin_isunordered(__x, __y); } -__DEVICE__ bool isunordered(double __x, double __y) { +__DEVICE__ __CONSTEXPR__ bool isunordered(double __x, double __y) { return __builtin_isunordered(__x, __y); } -__DEVICE__ float modf(float __x, float *__iptr) { return ::modff(__x, __iptr); } -__DEVICE__ float pow(float __base, int __iexp) { +__DEVICE__ __CONSTEXPR__ float modf(float __x, float *__iptr) { + return ::modff(__x, __iptr); +} +__DEVICE__ __CONSTEXPR__ float pow(float __base, int __iexp) { return ::powif(__base, __iexp); } -__DEVICE__ double pow(double __base, int __iexp) { +__DEVICE__ __CONSTEXPR__ double pow(double __base, int __iexp) { return ::powi(__base, __iexp); } -__DEVICE__ float remquo(float __x, float __y, int *__quo) { +__DEVICE__ __CONSTEXPR__ float remquo(float __x, float __y, int *__quo) { return ::remquof(__x, __y, __quo); } -__DEVICE__ float scalbln(float __x, long int __n) { +__DEVICE__ __CONSTEXPR__ float scalbln(float __x, long int __n) { return ::scalblnf(__x, __n); } -__DEVICE__ bool signbit(float __x) { return ::__signbitf(__x); } -__DEVICE__ bool signbit(double __x) { return ::__signbit(__x); } +__DEVICE__ __CONSTEXPR__ bool signbit(float __x) { return ::__signbitf(__x); } +__DEVICE__ __CONSTEXPR__ bool signbit(double __x) { return ::__signbit(__x); } // Notably missing above is nexttoward. We omit it because // ocml doesn't provide an implementation, and we don't want to be in the // business of implementing tricky libm functions in this header. // Other functions. -__DEVICE__ _Float16 fma(_Float16 __x, _Float16 __y, _Float16 __z) { +__DEVICE__ __CONSTEXPR__ _Float16 fma(_Float16 __x, _Float16 __y, + _Float16 __z) { return __ocml_fma_f16(__x, __y, __z); } -__DEVICE__ _Float16 pow(_Float16 __base, int __iexp) { +__DEVICE__ __CONSTEXPR__ _Float16 pow(_Float16 __base, int __iexp) { return __ocml_pown_f16(__base, __iexp); } +#ifndef __OPENMP_AMDGCN__ // BEGIN DEF_FUN and HIP_OVERLOAD // BEGIN DEF_FUN @@ -127,18 +188,19 @@ __DEVICE__ _Float16 pow(_Float16 __base, int __iexp) { // Define cmath functions with float argument and returns __retty. #define __DEF_FUN1(__retty, __func) \ - __DEVICE__ \ - __retty __func(float __x) { return __func##f(__x); } + __DEVICE__ __CONSTEXPR__ __retty __func(float __x) { return __func##f(__x); } // Define cmath functions with two float arguments and returns __retty. #define __DEF_FUN2(__retty, __func) \ - __DEVICE__ \ - __retty __func(float __x, float __y) { return __func##f(__x, __y); } + __DEVICE__ __CONSTEXPR__ __retty __func(float __x, float __y) { \ + return __func##f(__x, __y); \ + } // Define cmath functions with a float and an int argument and returns __retty. #define __DEF_FUN2_FI(__retty, __func) \ - __DEVICE__ \ - __retty __func(float __x, int __y) { return __func##f(__x, __y); } + __DEVICE__ __CONSTEXPR__ __retty __func(float __x, int __y) { \ + return __func##f(__x, __y); \ + } __DEF_FUN1(float, acos) __DEF_FUN1(float, acosh) @@ -207,11 +269,117 @@ template struct __hip_enable_if {}; template struct __hip_enable_if { typedef __T type; }; +namespace __hip { +template struct is_integral { + enum { value = 0 }; +}; +template <> struct is_integral { + enum { value = 1 }; +}; +template <> struct is_integral { + enum { value = 1 }; +}; +template <> struct is_integral { + enum { value = 1 }; +}; +template <> struct is_integral { + enum { value = 1 }; +}; +template <> struct is_integral { + enum { value = 1 }; +}; +template <> struct is_integral { + enum { value = 1 }; +}; +template <> struct is_integral { + enum { value = 1 }; +}; +template <> struct is_integral { + enum { value = 1 }; +}; +template <> struct is_integral { + enum { value = 1 }; +}; +template <> struct is_integral { + enum { value = 1 }; +}; +template <> struct is_integral { + enum { value = 1 }; +}; +template <> struct is_integral { + enum { value = 1 }; +}; +template <> struct is_integral { + enum { value = 1 }; +}; + +// ToDo: specializes is_arithmetic<_Float16> +template struct is_arithmetic { + enum { value = 0 }; +}; +template <> struct is_arithmetic { + enum { value = 1 }; +}; +template <> struct is_arithmetic { + enum { value = 1 }; +}; +template <> struct is_arithmetic { + enum { value = 1 }; +}; +template <> struct is_arithmetic { + enum { value = 1 }; +}; +template <> struct is_arithmetic { + enum { value = 1 }; +}; +template <> struct is_arithmetic { + enum { value = 1 }; +}; +template <> struct is_arithmetic { + enum { value = 1 }; +}; +template <> struct is_arithmetic { + enum { value = 1 }; +}; +template <> struct is_arithmetic { + enum { value = 1 }; +}; +template <> struct is_arithmetic { + enum { value = 1 }; +}; +template <> struct is_arithmetic { + enum { value = 1 }; +}; +template <> struct is_arithmetic { + enum { value = 1 }; +}; +template <> struct is_arithmetic { + enum { value = 1 }; +}; +template <> struct is_arithmetic { + enum { value = 1 }; +}; +template <> struct is_arithmetic { + enum { value = 1 }; +}; + +struct true_type { + static const __constant__ bool value = true; +}; +struct false_type { + static const __constant__ bool value = false; +}; + +template struct is_same : public false_type {}; +template struct is_same<__T, __T> : public true_type {}; + +template struct add_rvalue_reference { typedef __T &&type; }; + +template typename add_rvalue_reference<__T>::type declval(); + // decltype is only available in C++11 and above. #if __cplusplus >= 201103L // __hip_promote -namespace __hip { - template struct __numeric_type { static void __test(...); static _Float16 __test(_Float16); @@ -227,8 +395,8 @@ template struct __numeric_type { // No support for long double, use double instead. static double __test(long double); - typedef decltype(__test(std::declval<_Tp>())) type; - static const bool value = !std::is_same::value; + typedef decltype(__test(declval<_Tp>())) type; + static const bool value = !is_same::value; }; template <> struct __numeric_type { static const bool value = true; }; @@ -271,18 +439,17 @@ public: template class __promote : public __promote_imp<_A1, _A2, _A3> {}; - -} // namespace __hip #endif //__cplusplus >= 201103L +} // namespace __hip // __HIP_OVERLOAD1 is used to resolve function calls with integer argument to // avoid compilation error due to ambibuity. e.g. floor(5) is resolved with // floor(double). #define __HIP_OVERLOAD1(__retty, __fn) \ template \ - __DEVICE__ typename __hip_enable_if::is_integer, \ - __retty>::type \ - __fn(__T __x) { \ + __DEVICE__ __CONSTEXPR__ \ + typename __hip_enable_if<__hip::is_integral<__T>::value, __retty>::type \ + __fn(__T __x) { \ return ::__fn((double)__x); \ } @@ -292,9 +459,8 @@ class __promote : public __promote_imp<_A1, _A2, _A3> {}; #if __cplusplus >= 201103L #define __HIP_OVERLOAD2(__retty, __fn) \ template \ - __DEVICE__ typename __hip_enable_if< \ - std::numeric_limits<__T1>::is_specialized && \ - std::numeric_limits<__T2>::is_specialized, \ + __DEVICE__ __CONSTEXPR__ typename __hip_enable_if< \ + __hip::is_arithmetic<__T1>::value && __hip::is_arithmetic<__T2>::value, \ typename __hip::__promote<__T1, __T2>::type>::type \ __fn(__T1 __x, __T2 __y) { \ typedef typename __hip::__promote<__T1, __T2>::type __result_type; \ @@ -303,16 +469,15 @@ class __promote : public __promote_imp<_A1, _A2, _A3> {}; #else #define __HIP_OVERLOAD2(__retty, __fn) \ template \ - __DEVICE__ \ - typename __hip_enable_if::is_specialized && \ - std::numeric_limits<__T2>::is_specialized, \ + __DEVICE__ __CONSTEXPR__ \ + typename __hip_enable_if<__hip::is_arithmetic<__T1>::value && \ + __hip::is_arithmetic<__T2>::value, \ __retty>::type \ __fn(__T1 __x, __T2 __y) { \ return __fn((double)__x, (double)__y); \ } #endif -__HIP_OVERLOAD1(double, abs) __HIP_OVERLOAD1(double, acos) __HIP_OVERLOAD1(double, acosh) __HIP_OVERLOAD1(double, asin) @@ -336,7 +501,9 @@ __HIP_OVERLOAD1(double, floor) __HIP_OVERLOAD2(double, fmax) __HIP_OVERLOAD2(double, fmin) __HIP_OVERLOAD2(double, fmod) +#if !defined(__HIPCC_RTC__) __HIP_OVERLOAD1(int, fpclassify) +#endif // !defined(__HIPCC_RTC__) __HIP_OVERLOAD2(double, hypot) __HIP_OVERLOAD1(int, ilogb) __HIP_OVERLOAD1(bool, isfinite) @@ -381,10 +548,9 @@ __HIP_OVERLOAD2(double, min) // Additional Overloads that don't quite match HIP_OVERLOAD. #if __cplusplus >= 201103L template -__DEVICE__ typename __hip_enable_if< - std::numeric_limits<__T1>::is_specialized && - std::numeric_limits<__T2>::is_specialized && - std::numeric_limits<__T3>::is_specialized, +__DEVICE__ __CONSTEXPR__ typename __hip_enable_if< + __hip::is_arithmetic<__T1>::value && __hip::is_arithmetic<__T2>::value && + __hip::is_arithmetic<__T3>::value, typename __hip::__promote<__T1, __T2, __T3>::type>::type fma(__T1 __x, __T2 __y, __T3 __z) { typedef typename __hip::__promote<__T1, __T2, __T3>::type __result_type; @@ -392,10 +558,10 @@ fma(__T1 __x, __T2 __y, __T3 __z) { } #else template -__DEVICE__ - typename __hip_enable_if::is_specialized && - std::numeric_limits<__T2>::is_specialized && - std::numeric_limits<__T3>::is_specialized, +__DEVICE__ __CONSTEXPR__ + typename __hip_enable_if<__hip::is_arithmetic<__T1>::value && + __hip::is_arithmetic<__T2>::value && + __hip::is_arithmetic<__T3>::value, double>::type fma(__T1 __x, __T2 __y, __T3 __z) { return ::fma((double)__x, (double)__y, (double)__z); @@ -403,31 +569,31 @@ __DEVICE__ #endif template -__DEVICE__ - typename __hip_enable_if::is_integer, double>::type +__DEVICE__ __CONSTEXPR__ + typename __hip_enable_if<__hip::is_integral<__T>::value, double>::type frexp(__T __x, int *__exp) { return ::frexp((double)__x, __exp); } template -__DEVICE__ - typename __hip_enable_if::is_integer, double>::type +__DEVICE__ __CONSTEXPR__ + typename __hip_enable_if<__hip::is_integral<__T>::value, double>::type ldexp(__T __x, int __exp) { return ::ldexp((double)__x, __exp); } template -__DEVICE__ - typename __hip_enable_if::is_integer, double>::type +__DEVICE__ __CONSTEXPR__ + typename __hip_enable_if<__hip::is_integral<__T>::value, double>::type modf(__T __x, double *__exp) { return ::modf((double)__x, __exp); } #if __cplusplus >= 201103L template -__DEVICE__ - typename __hip_enable_if::is_specialized && - std::numeric_limits<__T2>::is_specialized, +__DEVICE__ __CONSTEXPR__ + typename __hip_enable_if<__hip::is_arithmetic<__T1>::value && + __hip::is_arithmetic<__T2>::value, typename __hip::__promote<__T1, __T2>::type>::type remquo(__T1 __x, __T2 __y, int *__quo) { typedef typename __hip::__promote<__T1, __T2>::type __result_type; @@ -435,9 +601,9 @@ __DEVICE__ } #else template -__DEVICE__ - typename __hip_enable_if::is_specialized && - std::numeric_limits<__T2>::is_specialized, +__DEVICE__ __CONSTEXPR__ + typename __hip_enable_if<__hip::is_arithmetic<__T1>::value && + __hip::is_arithmetic<__T2>::value, double>::type remquo(__T1 __x, __T2 __y, int *__quo) { return ::remquo((double)__x, (double)__y, __quo); @@ -445,15 +611,15 @@ __DEVICE__ #endif template -__DEVICE__ - typename __hip_enable_if::is_integer, double>::type +__DEVICE__ __CONSTEXPR__ + typename __hip_enable_if<__hip::is_integral<__T>::value, double>::type scalbln(__T __x, long int __exp) { return ::scalbln((double)__x, __exp); } template -__DEVICE__ - typename __hip_enable_if::is_integer, double>::type +__DEVICE__ __CONSTEXPR__ + typename __hip_enable_if<__hip::is_integral<__T>::value, double>::type scalbn(__T __x, int __exp) { return ::scalbn((double)__x, __exp); } @@ -465,17 +631,20 @@ __DEVICE__ // END DEF_FUN and HIP_OVERLOAD +#endif // ifndef __OPENMP_AMDGCN__ #endif // defined(__cplusplus) +#ifndef __OPENMP_AMDGCN__ // Define these overloads inside the namespace our standard library uses. +#if !defined(__HIPCC_RTC__) #ifdef _LIBCPP_BEGIN_NAMESPACE_STD _LIBCPP_BEGIN_NAMESPACE_STD #else namespace std { #ifdef _GLIBCXX_BEGIN_NAMESPACE_VERSION _GLIBCXX_BEGIN_NAMESPACE_VERSION -#endif -#endif +#endif // _GLIBCXX_BEGIN_NAMESPACE_VERSION +#endif // _LIBCPP_BEGIN_NAMESPACE_STD // Pull the new overloads we defined above into namespace std. // using ::abs; - This may be considered for C++. @@ -620,11 +789,13 @@ _LIBCPP_END_NAMESPACE_STD #else #ifdef _GLIBCXX_BEGIN_NAMESPACE_VERSION _GLIBCXX_END_NAMESPACE_VERSION -#endif +#endif // _GLIBCXX_BEGIN_NAMESPACE_VERSION } // namespace std -#endif +#endif // _LIBCPP_END_NAMESPACE_STD +#endif // !defined(__HIPCC_RTC__) // Define device-side math functions from on MSVC. +#if !defined(__HIPCC_RTC__) #if defined(_MSC_VER) // Before VS2019, `` is also included in `` and other headers. @@ -636,29 +807,36 @@ _GLIBCXX_END_NAMESPACE_VERSION #if defined(__cplusplus) extern "C" { #endif // defined(__cplusplus) -__DEVICE__ __attribute__((overloadable)) double _Cosh(double x, double y) { +__DEVICE__ __CONSTEXPR__ __attribute__((overloadable)) double _Cosh(double x, + double y) { return cosh(x) * y; } -__DEVICE__ __attribute__((overloadable)) float _FCosh(float x, float y) { +__DEVICE__ __CONSTEXPR__ __attribute__((overloadable)) float _FCosh(float x, + float y) { return coshf(x) * y; } -__DEVICE__ __attribute__((overloadable)) short _Dtest(double *p) { +__DEVICE__ __CONSTEXPR__ __attribute__((overloadable)) short _Dtest(double *p) { return fpclassify(*p); } -__DEVICE__ __attribute__((overloadable)) short _FDtest(float *p) { +__DEVICE__ __CONSTEXPR__ __attribute__((overloadable)) short _FDtest(float *p) { return fpclassify(*p); } -__DEVICE__ __attribute__((overloadable)) double _Sinh(double x, double y) { +__DEVICE__ __CONSTEXPR__ __attribute__((overloadable)) double _Sinh(double x, + double y) { return sinh(x) * y; } -__DEVICE__ __attribute__((overloadable)) float _FSinh(float x, float y) { +__DEVICE__ __CONSTEXPR__ __attribute__((overloadable)) float _FSinh(float x, + float y) { return sinhf(x) * y; } #if defined(__cplusplus) } #endif // defined(__cplusplus) #endif // defined(_MSC_VER) +#endif // !defined(__HIPCC_RTC__) +#endif // ifndef __OPENMP_AMDGCN__ #pragma pop_macro("__DEVICE__") +#pragma pop_macro("__CONSTEXPR__") #endif // __CLANG_HIP_CMATH_H__ diff --git a/lib/include/__clang_hip_libdevice_declares.h b/lib/include/__clang_hip_libdevice_declares.h index ac98907ad..8be848ba2 100644 --- a/lib/include/__clang_hip_libdevice_declares.h +++ b/lib/include/__clang_hip_libdevice_declares.h @@ -138,14 +138,22 @@ __device__ __attribute__((const)) float __ocml_fma_rtn_f32(float, float, float); __device__ __attribute__((const)) float __ocml_fma_rtp_f32(float, float, float); __device__ __attribute__((const)) float __ocml_fma_rtz_f32(float, float, float); -__device__ __attribute__((const)) float -__llvm_amdgcn_cos_f32(float) __asm("llvm.amdgcn.cos.f32"); -__device__ __attribute__((const)) float -__llvm_amdgcn_rcp_f32(float) __asm("llvm.amdgcn.rcp.f32"); -__device__ __attribute__((const)) float -__llvm_amdgcn_rsq_f32(float) __asm("llvm.amdgcn.rsq.f32"); -__device__ __attribute__((const)) float -__llvm_amdgcn_sin_f32(float) __asm("llvm.amdgcn.sin.f32"); +__device__ inline __attribute__((const)) float +__llvm_amdgcn_cos_f32(float __x) { + return __builtin_amdgcn_cosf(__x); +} +__device__ inline __attribute__((const)) float +__llvm_amdgcn_rcp_f32(float __x) { + return __builtin_amdgcn_rcpf(__x); +} +__device__ inline __attribute__((const)) float +__llvm_amdgcn_rsq_f32(float __x) { + return __builtin_amdgcn_rsqf(__x); +} +__device__ inline __attribute__((const)) float +__llvm_amdgcn_sin_f32(float __x) { + return __builtin_amdgcn_sinf(__x); +} // END INTRINSICS // END FLOAT @@ -269,10 +277,14 @@ __device__ __attribute__((const)) double __ocml_fma_rtp_f64(double, double, __device__ __attribute__((const)) double __ocml_fma_rtz_f64(double, double, double); -__device__ __attribute__((const)) double -__llvm_amdgcn_rcp_f64(double) __asm("llvm.amdgcn.rcp.f64"); -__device__ __attribute__((const)) double -__llvm_amdgcn_rsq_f64(double) __asm("llvm.amdgcn.rsq.f64"); +__device__ inline __attribute__((const)) double +__llvm_amdgcn_rcp_f64(double __x) { + return __builtin_amdgcn_rcp(__x); +} +__device__ inline __attribute__((const)) double +__llvm_amdgcn_rsq_f64(double __x) { + return __builtin_amdgcn_rsq(__x); +} __device__ __attribute__((const)) _Float16 __ocml_ceil_f16(_Float16); __device__ _Float16 __ocml_cos_f16(_Float16); diff --git a/lib/include/__clang_hip_math.h b/lib/include/__clang_hip_math.h index 14d91c66b..ef7e087b8 100644 --- a/lib/include/__clang_hip_math.h +++ b/lib/include/__clang_hip_math.h @@ -9,26 +9,40 @@ #ifndef __CLANG_HIP_MATH_H__ #define __CLANG_HIP_MATH_H__ -#if !defined(__HIP__) +#if !defined(__HIP__) && !defined(__OPENMP_AMDGCN__) #error "This file is for HIP and OpenMP AMDGCN device compilation only." #endif +#if !defined(__HIPCC_RTC__) #if defined(__cplusplus) #include #endif #include #include +#ifdef __OPENMP_AMDGCN__ +#include +#endif +#endif // !defined(__HIPCC_RTC__) #pragma push_macro("__DEVICE__") + +#ifdef __OPENMP_AMDGCN__ +#define __DEVICE__ static inline __attribute__((always_inline, nothrow)) +#else #define __DEVICE__ static __device__ inline __attribute__((always_inline)) +#endif // A few functions return bool type starting only in C++11. #pragma push_macro("__RETURN_TYPE") +#ifdef __OPENMP_AMDGCN__ +#define __RETURN_TYPE int +#else #if defined(__cplusplus) #define __RETURN_TYPE bool #else #define __RETURN_TYPE int #endif +#endif // __OPENMP_AMDGCN__ #if defined (__cplusplus) && __cplusplus < 201103L // emulate static_assert on type sizes @@ -36,7 +50,7 @@ template struct __compare_result{}; template<> struct __compare_result { - static const bool valid; + static const __device__ bool valid; }; __DEVICE__ @@ -247,6 +261,9 @@ float fmodf(float __x, float __y) { return __ocml_fmod_f32(__x, __y); } __DEVICE__ float frexpf(float __x, int *__nptr) { int __tmp; +#ifdef __OPENMP_AMDGCN__ +#pragma omp allocate(__tmp) allocator(omp_thread_mem_alloc) +#endif float __r = __ocml_frexp_f32(__x, (__attribute__((address_space(5))) int *)&__tmp); *__nptr = __tmp; @@ -332,6 +349,9 @@ long int lroundf(float __x) { return __ocml_round_f32(__x); } __DEVICE__ float modff(float __x, float *__iptr) { float __tmp; +#ifdef __OPENMP_AMDGCN__ +#pragma omp allocate(__tmp) allocator(omp_thread_mem_alloc) +#endif float __r = __ocml_modf_f32(__x, (__attribute__((address_space(5))) float *)&__tmp); *__iptr = __tmp; @@ -412,6 +432,9 @@ float remainderf(float __x, float __y) { __DEVICE__ float remquof(float __x, float __y, int *__quo) { int __tmp; +#ifdef __OPENMP_AMDGCN__ +#pragma omp allocate(__tmp) allocator(omp_thread_mem_alloc) +#endif float __r = __ocml_remquo_f32( __x, __y, (__attribute__((address_space(5))) int *)&__tmp); *__quo = __tmp; @@ -468,6 +491,9 @@ __RETURN_TYPE __signbitf(float __x) { return __ocml_signbit_f32(__x); } __DEVICE__ void sincosf(float __x, float *__sinptr, float *__cosptr) { float __tmp; +#ifdef __OPENMP_AMDGCN__ +#pragma omp allocate(__tmp) allocator(omp_thread_mem_alloc) +#endif *__sinptr = __ocml_sincos_f32(__x, (__attribute__((address_space(5))) float *)&__tmp); *__cosptr = __tmp; @@ -476,6 +502,9 @@ void sincosf(float __x, float *__sinptr, float *__cosptr) { __DEVICE__ void sincospif(float __x, float *__sinptr, float *__cosptr) { float __tmp; +#ifdef __OPENMP_AMDGCN__ +#pragma omp allocate(__tmp) allocator(omp_thread_mem_alloc) +#endif *__sinptr = __ocml_sincospi_f32( __x, (__attribute__((address_space(5))) float *)&__tmp); *__cosptr = __tmp; @@ -788,6 +817,9 @@ double fmod(double __x, double __y) { return __ocml_fmod_f64(__x, __y); } __DEVICE__ double frexp(double __x, int *__nptr) { int __tmp; +#ifdef __OPENMP_AMDGCN__ +#pragma omp allocate(__tmp) allocator(omp_thread_mem_alloc) +#endif double __r = __ocml_frexp_f64(__x, (__attribute__((address_space(5))) int *)&__tmp); *__nptr = __tmp; @@ -872,6 +904,9 @@ long int lround(double __x) { return __ocml_round_f64(__x); } __DEVICE__ double modf(double __x, double *__iptr) { double __tmp; +#ifdef __OPENMP_AMDGCN__ +#pragma omp allocate(__tmp) allocator(omp_thread_mem_alloc) +#endif double __r = __ocml_modf_f64(__x, (__attribute__((address_space(5))) double *)&__tmp); *__iptr = __tmp; @@ -960,6 +995,9 @@ double remainder(double __x, double __y) { __DEVICE__ double remquo(double __x, double __y, int *__quo) { int __tmp; +#ifdef __OPENMP_AMDGCN__ +#pragma omp allocate(__tmp) allocator(omp_thread_mem_alloc) +#endif double __r = __ocml_remquo_f64( __x, __y, (__attribute__((address_space(5))) int *)&__tmp); *__quo = __tmp; @@ -1018,6 +1056,9 @@ double sin(double __x) { return __ocml_sin_f64(__x); } __DEVICE__ void sincos(double __x, double *__sinptr, double *__cosptr) { double __tmp; +#ifdef __OPENMP_AMDGCN__ +#pragma omp allocate(__tmp) allocator(omp_thread_mem_alloc) +#endif *__sinptr = __ocml_sincos_f64( __x, (__attribute__((address_space(5))) double *)&__tmp); *__cosptr = __tmp; @@ -1026,6 +1067,9 @@ void sincos(double __x, double *__sinptr, double *__cosptr) { __DEVICE__ void sincospi(double __x, double *__sinptr, double *__cosptr) { double __tmp; +#ifdef __OPENMP_AMDGCN__ +#pragma omp allocate(__tmp) allocator(omp_thread_mem_alloc) +#endif *__sinptr = __ocml_sincospi_f64( __x, (__attribute__((address_space(5))) double *)&__tmp); *__cosptr = __tmp; @@ -1260,6 +1304,7 @@ float min(float __x, float __y) { return fminf(__x, __y); } __DEVICE__ double min(double __x, double __y) { return fmin(__x, __y); } +#if !defined(__HIPCC_RTC__) && !defined(__OPENMP_AMDGCN__) __host__ inline static int min(int __arg1, int __arg2) { return std::min(__arg1, __arg2); } @@ -1267,6 +1312,7 @@ __host__ inline static int min(int __arg1, int __arg2) { __host__ inline static int max(int __arg1, int __arg2) { return std::max(__arg1, __arg2); } +#endif // !defined(__HIPCC_RTC__) && !defined(__OPENMP_AMDGCN__) #endif #pragma pop_macro("__DEVICE__") diff --git a/lib/include/__clang_hip_runtime_wrapper.h b/lib/include/__clang_hip_runtime_wrapper.h index 81a16a265..73021d256 100644 --- a/lib/include/__clang_hip_runtime_wrapper.h +++ b/lib/include/__clang_hip_runtime_wrapper.h @@ -18,52 +18,107 @@ #if __HIP__ -#include -#include -#include - #define __host__ __attribute__((host)) #define __device__ __attribute__((device)) #define __global__ __attribute__((global)) #define __shared__ __attribute__((shared)) #define __constant__ __attribute__((constant)) +#define __managed__ __attribute__((managed)) #if !defined(__cplusplus) || __cplusplus < 201103L #define nullptr NULL; #endif +#ifdef __cplusplus +extern "C" { + __attribute__((__visibility__("default"))) + __attribute__((weak)) + __attribute__((noreturn)) + __device__ void __cxa_pure_virtual(void) { + __builtin_trap(); + } + __attribute__((__visibility__("default"))) + __attribute__((weak)) + __attribute__((noreturn)) + __device__ void __cxa_deleted_virtual(void) { + __builtin_trap(); + } +} +#endif //__cplusplus + +#if !defined(__HIPCC_RTC__) +#include +#include +#include +#else +typedef __SIZE_TYPE__ size_t; +// Define macros which are needed to declare HIP device API's without standard +// C/C++ headers. This is for readability so that these API's can be written +// the same way as non-hipRTC use case. These macros need to be popped so that +// they do not pollute users' name space. +#pragma push_macro("NULL") +#pragma push_macro("uint32_t") +#pragma push_macro("uint64_t") +#pragma push_macro("CHAR_BIT") +#pragma push_macro("INT_MAX") +#define NULL (void *)0 +#define uint32_t __UINT32_TYPE__ +#define uint64_t __UINT64_TYPE__ +#define CHAR_BIT __CHAR_BIT__ +#define INT_MAX __INTMAX_MAX__ +#endif // __HIPCC_RTC__ + +typedef __SIZE_TYPE__ __hip_size_t; + +#ifdef __cplusplus +extern "C" { +#endif //__cplusplus + #if __HIP_ENABLE_DEVICE_MALLOC__ -extern "C" __device__ void *__hip_malloc(size_t __size); -extern "C" __device__ void *__hip_free(void *__ptr); -static inline __device__ void *malloc(size_t __size) { +__device__ void *__hip_malloc(__hip_size_t __size); +__device__ void *__hip_free(void *__ptr); +__attribute__((weak)) inline __device__ void *malloc(__hip_size_t __size) { return __hip_malloc(__size); } -static inline __device__ void *free(void *__ptr) { return __hip_free(__ptr); } +__attribute__((weak)) inline __device__ void *free(void *__ptr) { + return __hip_free(__ptr); +} #else -static inline __device__ void *malloc(size_t __size) { +__attribute__((weak)) inline __device__ void *malloc(__hip_size_t __size) { __builtin_trap(); - return nullptr; + return (void *)0; } -static inline __device__ void *free(void *__ptr) { +__attribute__((weak)) inline __device__ void *free(void *__ptr) { __builtin_trap(); - return nullptr; + return (void *)0; } #endif +#ifdef __cplusplus +} // extern "C" +#endif //__cplusplus + #include <__clang_hip_libdevice_declares.h> #include <__clang_hip_math.h> -#if !_OPENMP || __HIP_ENABLE_CUDA_WRAPPER_FOR_OPENMP__ +#if defined(__HIPCC_RTC__) +#include <__clang_hip_cmath.h> +#else #include <__clang_cuda_math_forward_declares.h> #include <__clang_hip_cmath.h> #include <__clang_cuda_complex_builtins.h> - #include #include #include -#endif // !_OPENMP || __HIP_ENABLE_CUDA_WRAPPER_FOR_OPENMP__ +#endif // __HIPCC_RTC__ #define __CLANG_HIP_RUNTIME_WRAPPER_INCLUDED__ 1 - +#if defined(__HIPCC_RTC__) +#pragma pop_macro("NULL") +#pragma pop_macro("uint32_t") +#pragma pop_macro("uint64_t") +#pragma pop_macro("CHAR_BIT") +#pragma pop_macro("INT_MAX") +#endif // __HIPCC_RTC__ #endif // __HIP__ #endif // __CLANG_HIP_RUNTIME_WRAPPER_H__ diff --git a/lib/include/altivec.h b/lib/include/altivec.h index 4d50d47d5..0dd8c8593 100644 --- a/lib/include/altivec.h +++ b/lib/include/altivec.h @@ -41,9 +41,7 @@ #define __ATTRS_o_ai __attribute__((__overloadable__, __always_inline__)) -#ifdef __POWER9_VECTOR__ #include -#endif static __inline__ vector signed char __ATTRS_o_ai vec_perm( vector signed char __a, vector signed char __b, vector unsigned char __c); @@ -126,7 +124,7 @@ vec_abs(vector signed int __a) { return __builtin_altivec_vmaxsw(__a, -__a); } -#if defined(__POWER8_VECTOR__) && defined(__powerpc64__) +#ifdef __POWER8_VECTOR__ static __inline__ vector signed long long __ATTRS_o_ai vec_abs(vector signed long long __a) { return __builtin_altivec_vmaxsd(__a, -__a); @@ -284,7 +282,7 @@ vec_add(vector unsigned int __a, vector bool int __b) { return __a + (vector unsigned int)__b; } -#if defined(__POWER8_VECTOR__) && defined(__powerpc64__) +#ifdef __POWER8_VECTOR__ static __inline__ vector signed long long __ATTRS_o_ai vec_add(vector signed long long __a, vector signed long long __b) { return __a + __b; @@ -295,6 +293,7 @@ vec_add(vector unsigned long long __a, vector unsigned long long __b) { return __a + __b; } +#ifdef __SIZEOF_INT128__ static __inline__ vector signed __int128 __ATTRS_o_ai vec_add(vector signed __int128 __a, vector signed __int128 __b) { return __a + __b; @@ -304,7 +303,37 @@ static __inline__ vector unsigned __int128 __ATTRS_o_ai vec_add(vector unsigned __int128 __a, vector unsigned __int128 __b) { return __a + __b; } -#endif // defined(__POWER8_VECTOR__) && defined(__powerpc64__) +#endif + +static __inline__ vector unsigned char __attribute__((__always_inline__)) +vec_add_u128(vector unsigned char __a, vector unsigned char __b) { + return __builtin_altivec_vadduqm(__a, __b); +} +#elif defined(__VSX__) +static __inline__ vector signed long long __ATTRS_o_ai +vec_add(vector signed long long __a, vector signed long long __b) { +#ifdef __LITTLE_ENDIAN__ + // Little endian systems on CPU's prior to Power8 don't really exist + // so scalarizing is fine. + return __a + __b; +#else + vector unsigned int __res = + (vector unsigned int)__a + (vector unsigned int)__b; + vector unsigned int __carry = __builtin_altivec_vaddcuw( + (vector unsigned int)__a, (vector unsigned int)__b); + __carry = __builtin_shufflevector((vector unsigned char)__carry, + (vector unsigned char)__carry, 0, 0, 0, 7, + 0, 0, 0, 0, 0, 0, 0, 15, 0, 0, 0, 0); + return (vector signed long long)(__res + __carry); +#endif +} + +static __inline__ vector unsigned long long __ATTRS_o_ai +vec_add(vector unsigned long long __a, vector unsigned long long __b) { + return (vector unsigned long long)vec_add((vector signed long long)__a, + (vector signed long long)__b); +} +#endif // __POWER8_VECTOR__ static __inline__ vector float __ATTRS_o_ai vec_add(vector float __a, vector float __b) { @@ -320,7 +349,8 @@ static __inline__ vector double __ATTRS_o_ai vec_add(vector double __a, /* vec_adde */ -#if defined(__POWER8_VECTOR__) && defined(__powerpc64__) +#ifdef __POWER8_VECTOR__ +#ifdef __SIZEOF_INT128__ static __inline__ vector signed __int128 __ATTRS_o_ai vec_adde(vector signed __int128 __a, vector signed __int128 __b, vector signed __int128 __c) { @@ -334,6 +364,13 @@ vec_adde(vector unsigned __int128 __a, vector unsigned __int128 __b, } #endif +static __inline__ vector unsigned char __attribute__((__always_inline__)) +vec_adde_u128(vector unsigned char __a, vector unsigned char __b, + vector unsigned char __c) { + return (vector unsigned char)__builtin_altivec_vaddeuqm(__a, __b, __c); +} +#endif + static __inline__ vector signed int __ATTRS_o_ai vec_adde(vector signed int __a, vector signed int __b, vector signed int __c) { @@ -352,7 +389,8 @@ vec_adde(vector unsigned int __a, vector unsigned int __b, /* vec_addec */ -#if defined(__POWER8_VECTOR__) && defined(__powerpc64__) +#ifdef __POWER8_VECTOR__ +#ifdef __SIZEOF_INT128__ static __inline__ vector signed __int128 __ATTRS_o_ai vec_addec(vector signed __int128 __a, vector signed __int128 __b, vector signed __int128 __c) { @@ -364,7 +402,15 @@ vec_addec(vector unsigned __int128 __a, vector unsigned __int128 __b, vector unsigned __int128 __c) { return __builtin_altivec_vaddecuq(__a, __b, __c); } +#endif + +static __inline__ vector unsigned char __attribute__((__always_inline__)) +vec_addec_u128(vector unsigned char __a, vector unsigned char __b, + vector unsigned char __c) { + return (vector unsigned char)__builtin_altivec_vaddecuq(__a, __b, __c); +} +#ifdef __powerpc64__ static __inline__ vector signed int __ATTRS_o_ai vec_addec(vector signed int __a, vector signed int __b, vector signed int __c) { @@ -407,8 +453,8 @@ vec_addec(vector unsigned int __a, vector unsigned int __b, vector unsigned int ret = { __result[0], __result[1], __result[2], __result[3] }; return ret; } - -#endif +#endif // __powerpc64__ +#endif // __POWER8_VECTOR__ /* vec_vaddubm */ @@ -534,7 +580,8 @@ vec_addc(vector unsigned int __a, vector unsigned int __b) { return __builtin_altivec_vaddcuw(__a, __b); } -#if defined(__POWER8_VECTOR__) && defined(__powerpc64__) +#ifdef __POWER8_VECTOR__ +#ifdef __SIZEOF_INT128__ static __inline__ vector signed __int128 __ATTRS_o_ai vec_addc(vector signed __int128 __a, vector signed __int128 __b) { return (vector signed __int128)__builtin_altivec_vaddcuq( @@ -545,6 +592,12 @@ static __inline__ vector unsigned __int128 __ATTRS_o_ai vec_addc(vector unsigned __int128 __a, vector unsigned __int128 __b) { return __builtin_altivec_vaddcuq(__a, __b); } +#endif + +static __inline__ vector unsigned char __attribute__((__always_inline__)) +vec_addc_u128(vector unsigned char __a, vector unsigned char __b) { + return (vector unsigned char)__builtin_altivec_vaddcuq(__a, __b); +} #endif // defined(__POWER8_VECTOR__) && defined(__powerpc64__) /* vec_vaddcuw */ @@ -748,7 +801,8 @@ vec_vadduws(vector unsigned int __a, vector bool int __b) { return __builtin_altivec_vadduws(__a, (vector unsigned int)__b); } -#if defined(__POWER8_VECTOR__) && defined(__powerpc64__) +#if defined(__POWER8_VECTOR__) && defined(__powerpc64__) && \ + defined(__SIZEOF_INT128__) /* vec_vadduqm */ static __inline__ vector signed __int128 __ATTRS_o_ai @@ -1598,6 +1652,17 @@ static __inline__ vector double __ATTRS_o_ai vec_ceil(vector double __a) { } #endif +/* vec_roundp */ +static __inline__ vector float __ATTRS_o_ai vec_roundp(vector float __a) { + return vec_ceil(__a); +} + +#ifdef __VSX__ +static __inline__ vector double __ATTRS_o_ai vec_roundp(vector double __a) { + return vec_ceil(__a); +} +#endif + /* vec_vrfip */ static __inline__ vector float __attribute__((__always_inline__)) @@ -1690,7 +1755,31 @@ vec_cmpeq(vector bool long long __a, vector bool long long __b) { return (vector bool long long)__builtin_altivec_vcmpequd( (vector long long)__a, (vector long long)__b); } +#elif defined(__VSX__) +static __inline__ vector bool long long __ATTRS_o_ai +vec_cmpeq(vector signed long long __a, vector signed long long __b) { + vector bool int __wordcmp = + vec_cmpeq((vector signed int)__a, (vector signed int)__b); +#ifdef __LITTLE_ENDIAN__ + __wordcmp &= __builtin_shufflevector(__wordcmp, __wordcmp, 3, 0, 1, 2); + return (vector bool long long)__builtin_shufflevector(__wordcmp, __wordcmp, 1, + 1, 3, 3); +#else + __wordcmp &= __builtin_shufflevector(__wordcmp, __wordcmp, 1, 2, 3, 0); + return (vector bool long long)__builtin_shufflevector(__wordcmp, __wordcmp, 0, + 0, 2, 2); +#endif +} + +static __inline__ vector bool long long __ATTRS_o_ai +vec_cmpeq(vector unsigned long long __a, vector unsigned long long __b) { + return vec_cmpeq((vector signed long long)__a, (vector signed long long)__b); +} +static __inline__ vector bool long long __ATTRS_o_ai +vec_cmpeq(vector bool long long __a, vector bool long long __b) { + return vec_cmpeq((vector signed long long)__a, (vector signed long long)__b); +} #endif static __inline__ vector bool int __ATTRS_o_ai vec_cmpeq(vector float __a, @@ -1709,7 +1798,7 @@ vec_cmpeq(vector double __a, vector double __b) { } #endif -#ifdef __POWER10_VECTOR__ +#if defined(__POWER10_VECTOR__) && defined(__SIZEOF_INT128__) static __inline__ vector bool __int128 __ATTRS_o_ai vec_cmpeq(vector signed __int128 __a, vector signed __int128 __b) { return (vector bool __int128)__builtin_altivec_vcmpequq( @@ -1786,7 +1875,7 @@ vec_cmpne(vector float __a, vector float __b) { (vector int)__b); } -#ifdef __POWER10_VECTOR__ +#if defined(__POWER10_VECTOR__) && defined(__SIZEOF_INT128__) static __inline__ vector bool __int128 __ATTRS_o_ai vec_cmpne(vector unsigned __int128 __a, vector unsigned __int128 __b) { return (vector bool __int128) ~(__builtin_altivec_vcmpequq( @@ -1884,6 +1973,7 @@ vec_parity_lsbb(vector signed int __a) { return __builtin_altivec_vprtybw(__a); } +#ifdef __SIZEOF_INT128__ static __inline__ vector unsigned __int128 __ATTRS_o_ai vec_parity_lsbb(vector unsigned __int128 __a) { return __builtin_altivec_vprtybq(__a); @@ -1893,6 +1983,7 @@ static __inline__ vector unsigned __int128 __ATTRS_o_ai vec_parity_lsbb(vector signed __int128 __a) { return __builtin_altivec_vprtybq(__a); } +#endif static __inline__ vector unsigned long long __ATTRS_o_ai vec_parity_lsbb(vector unsigned long long __a) { @@ -1976,6 +2067,24 @@ vec_cmpne(vector unsigned long long __a, vector unsigned long long __b) { return (vector bool long long) ~(__builtin_altivec_vcmpequd((vector long long)__a, (vector long long)__b)); } +#elif defined(__VSX__) +static __inline__ vector bool long long __ATTRS_o_ai +vec_cmpne(vector bool long long __a, vector bool long long __b) { + return (vector bool long long)~( + vec_cmpeq((vector signed long long)__a, (vector signed long long)__b)); +} + +static __inline__ vector bool long long __ATTRS_o_ai +vec_cmpne(vector signed long long __a, vector signed long long __b) { + return (vector bool long long)~( + vec_cmpeq((vector signed long long)__a, (vector signed long long)__b)); +} + +static __inline__ vector bool long long __ATTRS_o_ai +vec_cmpne(vector unsigned long long __a, vector unsigned long long __b) { + return (vector bool long long)~( + vec_cmpeq((vector signed long long)__a, (vector signed long long)__b)); +} #endif #ifdef __VSX__ @@ -2028,6 +2137,46 @@ static __inline__ vector bool long long __ATTRS_o_ai vec_cmpgt(vector unsigned long long __a, vector unsigned long long __b) { return (vector bool long long)__builtin_altivec_vcmpgtud(__a, __b); } +#elif defined(__VSX__) +static __inline__ vector bool long long __ATTRS_o_ai +vec_cmpgt(vector signed long long __a, vector signed long long __b) { + vector signed int __sgtw = (vector signed int)vec_cmpgt( + (vector signed int)__a, (vector signed int)__b); + vector unsigned int __ugtw = (vector unsigned int)vec_cmpgt( + (vector unsigned int)__a, (vector unsigned int)__b); + vector unsigned int __eqw = (vector unsigned int)vec_cmpeq( + (vector signed int)__a, (vector signed int)__b); +#ifdef __LITTLE_ENDIAN__ + __ugtw = __builtin_shufflevector(__ugtw, __ugtw, 3, 0, 1, 2) & __eqw; + __sgtw |= (vector signed int)__ugtw; + return (vector bool long long)__builtin_shufflevector(__sgtw, __sgtw, 1, 1, 3, + 3); +#else + __ugtw = __builtin_shufflevector(__ugtw, __ugtw, 1, 2, 3, 0) & __eqw; + __sgtw |= (vector signed int)__ugtw; + return (vector bool long long)__builtin_shufflevector(__sgtw, __sgtw, 0, 0, 2, + 2); +#endif +} + +static __inline__ vector bool long long __ATTRS_o_ai +vec_cmpgt(vector unsigned long long __a, vector unsigned long long __b) { + vector unsigned int __ugtw = (vector unsigned int)vec_cmpgt( + (vector unsigned int)__a, (vector unsigned int)__b); + vector unsigned int __eqw = (vector unsigned int)vec_cmpeq( + (vector signed int)__a, (vector signed int)__b); +#ifdef __LITTLE_ENDIAN__ + __eqw = __builtin_shufflevector(__ugtw, __ugtw, 3, 0, 1, 2) & __eqw; + __ugtw |= __eqw; + return (vector bool long long)__builtin_shufflevector(__ugtw, __ugtw, 1, 1, 3, + 3); +#else + __eqw = __builtin_shufflevector(__ugtw, __ugtw, 1, 2, 3, 0) & __eqw; + __ugtw |= __eqw; + return (vector bool long long)__builtin_shufflevector(__ugtw, __ugtw, 0, 0, 2, + 2); +#endif +} #endif static __inline__ vector bool int __ATTRS_o_ai vec_cmpgt(vector float __a, @@ -2046,7 +2195,7 @@ vec_cmpgt(vector double __a, vector double __b) { } #endif -#ifdef __POWER10_VECTOR__ +#if defined(__POWER10_VECTOR__) && defined(__SIZEOF_INT128__) static __inline__ vector bool __int128 __ATTRS_o_ai vec_cmpgt(vector signed __int128 __a, vector signed __int128 __b) { return (vector bool __int128)__builtin_altivec_vcmpgtsq( @@ -2106,9 +2255,7 @@ static __inline__ vector bool long long __ATTRS_o_ai vec_cmpge(vector double __a, vector double __b) { return (vector bool long long)__builtin_vsx_xvcmpgedp(__a, __b); } -#endif -#ifdef __POWER8_VECTOR__ static __inline__ vector bool long long __ATTRS_o_ai vec_cmpge(vector signed long long __a, vector signed long long __b) { return ~(vec_cmpgt(__b, __a)); @@ -2120,7 +2267,7 @@ vec_cmpge(vector unsigned long long __a, vector unsigned long long __b) { } #endif -#ifdef __POWER10_VECTOR__ +#if defined(__POWER10_VECTOR__) && defined(__SIZEOF_INT128__) static __inline__ vector bool __int128 __ATTRS_o_ai vec_cmpge(vector signed __int128 __a, vector signed __int128 __b) { return ~(vec_cmpgt(__b, __a)); @@ -2230,9 +2377,7 @@ static __inline__ vector bool long long __ATTRS_o_ai vec_cmple(vector double __a, vector double __b) { return vec_cmpge(__b, __a); } -#endif -#ifdef __POWER8_VECTOR__ static __inline__ vector bool long long __ATTRS_o_ai vec_cmple(vector signed long long __a, vector signed long long __b) { return vec_cmpge(__b, __a); @@ -2244,7 +2389,7 @@ vec_cmple(vector unsigned long long __a, vector unsigned long long __b) { } #endif -#ifdef __POWER10_VECTOR__ +#if defined(__POWER10_VECTOR__) && defined(__SIZEOF_INT128__) static __inline__ vector bool __int128 __ATTRS_o_ai vec_cmple(vector signed __int128 __a, vector signed __int128 __b) { return vec_cmpge(__b, __a); @@ -2300,7 +2445,7 @@ vec_cmplt(vector double __a, vector double __b) { } #endif -#ifdef __POWER10_VECTOR__ +#if defined(__POWER10_VECTOR__) && defined(__SIZEOF_INT128__) static __inline__ vector bool __int128 __ATTRS_o_ai vec_cmplt(vector signed __int128 __a, vector signed __int128 __b) { return vec_cmpgt(__b, __a); @@ -2312,7 +2457,7 @@ vec_cmplt(vector unsigned __int128 __a, vector unsigned __int128 __b) { } #endif -#ifdef __POWER8_VECTOR__ +#ifdef __VSX__ static __inline__ vector bool long long __ATTRS_o_ai vec_cmplt(vector signed long long __a, vector signed long long __b) { return vec_cmpgt(__b, __a); @@ -2322,7 +2467,9 @@ static __inline__ vector bool long long __ATTRS_o_ai vec_cmplt(vector unsigned long long __a, vector unsigned long long __b) { return vec_cmpgt(__b, __a); } +#endif +#ifdef __POWER8_VECTOR__ /* vec_popcnt */ static __inline__ vector signed char __ATTRS_o_ai @@ -2358,6 +2505,7 @@ vec_popcnt(vector unsigned long long __a) { return __builtin_altivec_vpopcntd(__a); } +#define vec_vclz vec_cntlz /* vec_cntlz */ static __inline__ vector signed char __ATTRS_o_ai @@ -2870,6 +3018,7 @@ static __inline__ vector float __ATTRS_o_ai vec_xl_len(const float *__a, size_t return (vector float)__builtin_vsx_lxvl(__a, (__b << 56)); } +#ifdef __SIZEOF_INT128__ static __inline__ vector signed __int128 __ATTRS_o_ai vec_xl_len(const signed __int128 *__a, size_t __b) { return (vector signed __int128)__builtin_vsx_lxvl(__a, (__b << 56)); @@ -2879,6 +3028,7 @@ static __inline__ vector unsigned __int128 __ATTRS_o_ai vec_xl_len(const unsigned __int128 *__a, size_t __b) { return (vector unsigned __int128)__builtin_vsx_lxvl(__a, (__b << 56)); } +#endif static __inline__ vector signed long long __ATTRS_o_ai vec_xl_len(const signed long long *__a, size_t __b) { @@ -2946,6 +3096,7 @@ static __inline__ void __ATTRS_o_ai vec_xst_len(vector float __a, float *__b, return __builtin_vsx_stxvl((vector int)__a, __b, (__c << 56)); } +#ifdef __SIZEOF_INT128__ static __inline__ void __ATTRS_o_ai vec_xst_len(vector signed __int128 __a, signed __int128 *__b, size_t __c) { @@ -2957,6 +3108,7 @@ static __inline__ void __ATTRS_o_ai vec_xst_len(vector unsigned __int128 __a, size_t __c) { return __builtin_vsx_stxvl((vector int)__a, __b, (__c << 56)); } +#endif static __inline__ void __ATTRS_o_ai vec_xst_len(vector signed long long __a, signed long long *__b, @@ -2991,6 +3143,15 @@ static __inline__ void __ATTRS_o_ai vec_xst_len_r(vector unsigned char __a, #endif #endif +#if defined(__POWER9_VECTOR__) && defined(__powerpc64__) +#define __vec_ldrmb(PTR, CNT) vec_xl_len_r((const unsigned char *)(PTR), (CNT)) +#define __vec_strmb(PTR, CNT, VAL) \ + vec_xst_len_r((VAL), (unsigned char *)(PTR), (CNT)) +#else +#define __vec_ldrmb __builtin_vsx_ldrmb +#define __vec_strmb __builtin_vsx_strmb +#endif + /* vec_cpsgn */ #ifdef __VSX__ @@ -3008,6 +3169,23 @@ static __inline__ vector double __ATTRS_o_ai vec_cpsgn(vector double __a, /* vec_ctf */ #ifdef __VSX__ +// There are some functions that have different signatures with the XL compiler +// from those in Clang/GCC and documented in the PVIPR. This macro ensures that +// the XL-compatible signatures are used for those functions. +#ifdef __XL_COMPAT_ALTIVEC__ +#define vec_ctf(__a, __b) \ + _Generic((__a), vector int \ + : (vector float)__builtin_altivec_vcfsx((vector int)(__a), (__b)), \ + vector unsigned int \ + : (vector float)__builtin_altivec_vcfux((vector unsigned int)(__a), \ + (__b)), \ + vector unsigned long long \ + : (__builtin_vsx_xvcvuxdsp((vector unsigned long long)(__a)) * \ + (vector float)(vector unsigned)((0x7f - (__b)) << 23)), \ + vector signed long long \ + : (__builtin_vsx_xvcvsxdsp((vector signed long long)(__a)) * \ + (vector float)(vector unsigned)((0x7f - (__b)) << 23))) +#else // __XL_COMPAT_ALTIVEC__ #define vec_ctf(__a, __b) \ _Generic((__a), vector int \ : (vector float)__builtin_altivec_vcfsx((vector int)(__a), (__b)), \ @@ -3024,6 +3202,7 @@ static __inline__ vector double __ATTRS_o_ai vec_cpsgn(vector double __a, vector double) * \ (vector double)(vector unsigned long long)((0x3ffULL - (__b)) \ << 52))) +#endif // __XL_COMPAT_ALTIVEC__ #else #define vec_ctf(__a, __b) \ _Generic((__a), vector int \ @@ -3033,10 +3212,32 @@ static __inline__ vector double __ATTRS_o_ai vec_cpsgn(vector double __a, (__b))) #endif +/* vec_ctd */ +#ifdef __VSX__ +#define vec_ctd(__a, __b) \ + _Generic((__a), vector signed int \ + : (vec_doublee((vector signed int)(__a)) * \ + (vector double)(vector unsigned long long)((0x3ffULL - (__b)) \ + << 52)), \ + vector unsigned int \ + : (vec_doublee((vector unsigned int)(__a)) * \ + (vector double)(vector unsigned long long)((0x3ffULL - (__b)) \ + << 52)), \ + vector unsigned long long \ + : (__builtin_convertvector((vector unsigned long long)(__a), \ + vector double) * \ + (vector double)(vector unsigned long long)((0x3ffULL - (__b)) \ + << 52)), \ + vector signed long long \ + : (__builtin_convertvector((vector signed long long)(__a), \ + vector double) * \ + (vector double)(vector unsigned long long)((0x3ffULL - (__b)) \ + << 52))) +#endif // __VSX__ + /* vec_vcfsx */ #define vec_vcfux __builtin_altivec_vcfux - /* vec_vcfux */ #define vec_vcfsx(__a, __b) __builtin_altivec_vcfsx((vector int)(__a), (__b)) @@ -3044,6 +3245,19 @@ static __inline__ vector double __ATTRS_o_ai vec_cpsgn(vector double __a, /* vec_cts */ #ifdef __VSX__ +#ifdef __XL_COMPAT_ALTIVEC__ +#define vec_cts(__a, __b) \ + _Generic((__a), vector float \ + : __builtin_altivec_vctsxs((vector float)(__a), (__b)), \ + vector double \ + : __extension__({ \ + vector double __ret = \ + (vector double)(__a) * \ + (vector double)(vector unsigned long long)((0x3ffULL + (__b)) \ + << 52); \ + __builtin_vsx_xvcvdpsxws(__ret); \ + })) +#else // __XL_COMPAT_ALTIVEC__ #define vec_cts(__a, __b) \ _Generic((__a), vector float \ : __builtin_altivec_vctsxs((vector float)(__a), (__b)), \ @@ -3055,6 +3269,7 @@ static __inline__ vector double __ATTRS_o_ai vec_cpsgn(vector double __a, << 52); \ __builtin_convertvector(__ret, vector signed long long); \ })) +#endif // __XL_COMPAT_ALTIVEC__ #else #define vec_cts __builtin_altivec_vctsxs #endif @@ -3066,6 +3281,19 @@ static __inline__ vector double __ATTRS_o_ai vec_cpsgn(vector double __a, /* vec_ctu */ #ifdef __VSX__ +#ifdef __XL_COMPAT_ALTIVEC__ +#define vec_ctu(__a, __b) \ + _Generic((__a), vector float \ + : __builtin_altivec_vctuxs((vector float)(__a), (__b)), \ + vector double \ + : __extension__({ \ + vector double __ret = \ + (vector double)(__a) * \ + (vector double)(vector unsigned long long)((0x3ffULL + __b) \ + << 52); \ + __builtin_vsx_xvcvdpuxws(__ret); \ + })) +#else // __XL_COMPAT_ALTIVEC__ #define vec_ctu(__a, __b) \ _Generic((__a), vector float \ : __builtin_altivec_vctuxs((vector float)(__a), (__b)), \ @@ -3077,10 +3305,95 @@ static __inline__ vector double __ATTRS_o_ai vec_cpsgn(vector double __a, << 52); \ __builtin_convertvector(__ret, vector unsigned long long); \ })) +#endif // __XL_COMPAT_ALTIVEC__ #else #define vec_ctu __builtin_altivec_vctuxs #endif +#ifdef __LITTLE_ENDIAN__ +/* vec_ctsl */ + +#ifdef __VSX__ +#define vec_ctsl(__a, __b) \ + _Generic((__a), vector float \ + : __extension__({ \ + vector float __ret = \ + (vector float)(__a) * \ + (vector float)(vector unsigned)((0x7f + (__b)) << 23); \ + __builtin_vsx_xvcvspsxds( \ + __builtin_vsx_xxsldwi(__ret, __ret, 1)); \ + }), \ + vector double \ + : __extension__({ \ + vector double __ret = \ + (vector double)(__a) * \ + (vector double)(vector unsigned long long)((0x3ffULL + __b) \ + << 52); \ + __builtin_convertvector(__ret, vector signed long long); \ + })) + +/* vec_ctul */ + +#define vec_ctul(__a, __b) \ + _Generic((__a), vector float \ + : __extension__({ \ + vector float __ret = \ + (vector float)(__a) * \ + (vector float)(vector unsigned)((0x7f + (__b)) << 23); \ + __builtin_vsx_xvcvspuxds( \ + __builtin_vsx_xxsldwi(__ret, __ret, 1)); \ + }), \ + vector double \ + : __extension__({ \ + vector double __ret = \ + (vector double)(__a) * \ + (vector double)(vector unsigned long long)((0x3ffULL + __b) \ + << 52); \ + __builtin_convertvector(__ret, vector unsigned long long); \ + })) +#endif +#else // __LITTLE_ENDIAN__ +/* vec_ctsl */ + +#ifdef __VSX__ +#define vec_ctsl(__a, __b) \ + _Generic((__a), vector float \ + : __extension__({ \ + vector float __ret = \ + (vector float)(__a) * \ + (vector float)(vector unsigned)((0x7f + (__b)) << 23); \ + __builtin_vsx_xvcvspsxds(__ret); \ + }), \ + vector double \ + : __extension__({ \ + vector double __ret = \ + (vector double)(__a) * \ + (vector double)(vector unsigned long long)((0x3ffULL + __b) \ + << 52); \ + __builtin_convertvector(__ret, vector signed long long); \ + })) + +/* vec_ctul */ + +#define vec_ctul(__a, __b) \ + _Generic((__a), vector float \ + : __extension__({ \ + vector float __ret = \ + (vector float)(__a) * \ + (vector float)(vector unsigned)((0x7f + (__b)) << 23); \ + __builtin_vsx_xvcvspuxds(__ret); \ + }), \ + vector double \ + : __extension__({ \ + vector double __ret = \ + (vector double)(__a) * \ + (vector double)(vector unsigned long long)((0x3ffULL + __b) \ + << 52); \ + __builtin_convertvector(__ret, vector unsigned long long); \ + })) +#endif +#endif // __LITTLE_ENDIAN__ + /* vec_vctuxs */ #define vec_vctuxs __builtin_altivec_vctuxs @@ -3114,7 +3427,7 @@ vec_signextll(vector signed int __a) { } #endif -#ifdef __POWER10_VECTOR__ +#if defined(__POWER10_VECTOR__) && defined(__SIZEOF_INT128__) static __inline__ vector signed __int128 __ATTRS_o_ai vec_signextq(vector signed long long __a) { return __builtin_altivec_vextsd2q(__a); @@ -3399,6 +3712,15 @@ vec_doubleo(vector float __a) { return __builtin_vsx_xvcvspdp(vec_sld(__a, __a, 4)); #endif } + +/* vec_cvf */ +static __inline__ vector double __ATTRS_o_ai vec_cvf(vector float __a) { + return vec_doublee(__a); +} + +static __inline__ vector float __ATTRS_o_ai vec_cvf(vector double __a) { + return vec_floate(__a); +} #endif /* vec_div */ @@ -3481,6 +3803,7 @@ vec_dive(vector unsigned long long __a, vector unsigned long long __b) { return __builtin_altivec_vdiveud(__a, __b); } +#ifdef __SIZEOF_INT128__ static __inline__ vector unsigned __int128 __ATTRS_o_ai vec_dive(vector unsigned __int128 __a, vector unsigned __int128 __b) { return __builtin_altivec_vdiveuq(__a, __b); @@ -3491,8 +3814,9 @@ vec_dive(vector signed __int128 __a, vector signed __int128 __b) { return __builtin_altivec_vdivesq(__a, __b); } #endif +#endif -#ifdef __POWER10_VECTOR__ +#if defined(__POWER10_VECTOR__) && defined(__SIZEOF_INT128__) static __inline__ vector unsigned __int128 __ATTRS_o_ai vec_div(vector unsigned __int128 __a, vector unsigned __int128 __b) { return __a / __b; @@ -3661,6 +3985,17 @@ static __inline__ vector double __ATTRS_o_ai vec_floor(vector double __a) { } #endif +/* vec_roundm */ +static __inline__ vector float __ATTRS_o_ai vec_roundm(vector float __a) { + return vec_floor(__a); +} + +#ifdef __VSX__ +static __inline__ vector double __ATTRS_o_ai vec_roundm(vector double __a) { + return vec_floor(__a); +} +#endif + /* vec_vrfim */ static __inline__ vector float __attribute__((__always_inline__)) @@ -3671,251 +4006,251 @@ vec_vrfim(vector float __a) { /* vec_ld */ static __inline__ vector signed char __ATTRS_o_ai -vec_ld(int __a, const vector signed char *__b) { +vec_ld(long __a, const vector signed char *__b) { return (vector signed char)__builtin_altivec_lvx(__a, __b); } static __inline__ vector signed char __ATTRS_o_ai -vec_ld(int __a, const signed char *__b) { +vec_ld(long __a, const signed char *__b) { return (vector signed char)__builtin_altivec_lvx(__a, __b); } static __inline__ vector unsigned char __ATTRS_o_ai -vec_ld(int __a, const vector unsigned char *__b) { +vec_ld(long __a, const vector unsigned char *__b) { return (vector unsigned char)__builtin_altivec_lvx(__a, __b); } static __inline__ vector unsigned char __ATTRS_o_ai -vec_ld(int __a, const unsigned char *__b) { +vec_ld(long __a, const unsigned char *__b) { return (vector unsigned char)__builtin_altivec_lvx(__a, __b); } static __inline__ vector bool char __ATTRS_o_ai -vec_ld(int __a, const vector bool char *__b) { +vec_ld(long __a, const vector bool char *__b) { return (vector bool char)__builtin_altivec_lvx(__a, __b); } -static __inline__ vector short __ATTRS_o_ai vec_ld(int __a, +static __inline__ vector short __ATTRS_o_ai vec_ld(long __a, const vector short *__b) { return (vector short)__builtin_altivec_lvx(__a, __b); } -static __inline__ vector short __ATTRS_o_ai vec_ld(int __a, const short *__b) { +static __inline__ vector short __ATTRS_o_ai vec_ld(long __a, const short *__b) { return (vector short)__builtin_altivec_lvx(__a, __b); } static __inline__ vector unsigned short __ATTRS_o_ai -vec_ld(int __a, const vector unsigned short *__b) { +vec_ld(long __a, const vector unsigned short *__b) { return (vector unsigned short)__builtin_altivec_lvx(__a, __b); } static __inline__ vector unsigned short __ATTRS_o_ai -vec_ld(int __a, const unsigned short *__b) { +vec_ld(long __a, const unsigned short *__b) { return (vector unsigned short)__builtin_altivec_lvx(__a, __b); } static __inline__ vector bool short __ATTRS_o_ai -vec_ld(int __a, const vector bool short *__b) { +vec_ld(long __a, const vector bool short *__b) { return (vector bool short)__builtin_altivec_lvx(__a, __b); } -static __inline__ vector pixel __ATTRS_o_ai vec_ld(int __a, +static __inline__ vector pixel __ATTRS_o_ai vec_ld(long __a, const vector pixel *__b) { return (vector pixel)__builtin_altivec_lvx(__a, __b); } -static __inline__ vector int __ATTRS_o_ai vec_ld(int __a, +static __inline__ vector int __ATTRS_o_ai vec_ld(long __a, const vector int *__b) { return (vector int)__builtin_altivec_lvx(__a, __b); } -static __inline__ vector int __ATTRS_o_ai vec_ld(int __a, const int *__b) { +static __inline__ vector int __ATTRS_o_ai vec_ld(long __a, const int *__b) { return (vector int)__builtin_altivec_lvx(__a, __b); } static __inline__ vector unsigned int __ATTRS_o_ai -vec_ld(int __a, const vector unsigned int *__b) { +vec_ld(long __a, const vector unsigned int *__b) { return (vector unsigned int)__builtin_altivec_lvx(__a, __b); } static __inline__ vector unsigned int __ATTRS_o_ai -vec_ld(int __a, const unsigned int *__b) { +vec_ld(long __a, const unsigned int *__b) { return (vector unsigned int)__builtin_altivec_lvx(__a, __b); } static __inline__ vector bool int __ATTRS_o_ai -vec_ld(int __a, const vector bool int *__b) { +vec_ld(long __a, const vector bool int *__b) { return (vector bool int)__builtin_altivec_lvx(__a, __b); } -static __inline__ vector float __ATTRS_o_ai vec_ld(int __a, +static __inline__ vector float __ATTRS_o_ai vec_ld(long __a, const vector float *__b) { return (vector float)__builtin_altivec_lvx(__a, __b); } -static __inline__ vector float __ATTRS_o_ai vec_ld(int __a, const float *__b) { +static __inline__ vector float __ATTRS_o_ai vec_ld(long __a, const float *__b) { return (vector float)__builtin_altivec_lvx(__a, __b); } /* vec_lvx */ static __inline__ vector signed char __ATTRS_o_ai -vec_lvx(int __a, const vector signed char *__b) { +vec_lvx(long __a, const vector signed char *__b) { return (vector signed char)__builtin_altivec_lvx(__a, __b); } static __inline__ vector signed char __ATTRS_o_ai -vec_lvx(int __a, const signed char *__b) { +vec_lvx(long __a, const signed char *__b) { return (vector signed char)__builtin_altivec_lvx(__a, __b); } static __inline__ vector unsigned char __ATTRS_o_ai -vec_lvx(int __a, const vector unsigned char *__b) { +vec_lvx(long __a, const vector unsigned char *__b) { return (vector unsigned char)__builtin_altivec_lvx(__a, __b); } static __inline__ vector unsigned char __ATTRS_o_ai -vec_lvx(int __a, const unsigned char *__b) { +vec_lvx(long __a, const unsigned char *__b) { return (vector unsigned char)__builtin_altivec_lvx(__a, __b); } static __inline__ vector bool char __ATTRS_o_ai -vec_lvx(int __a, const vector bool char *__b) { +vec_lvx(long __a, const vector bool char *__b) { return (vector bool char)__builtin_altivec_lvx(__a, __b); } -static __inline__ vector short __ATTRS_o_ai vec_lvx(int __a, +static __inline__ vector short __ATTRS_o_ai vec_lvx(long __a, const vector short *__b) { return (vector short)__builtin_altivec_lvx(__a, __b); } -static __inline__ vector short __ATTRS_o_ai vec_lvx(int __a, const short *__b) { +static __inline__ vector short __ATTRS_o_ai vec_lvx(long __a, const short *__b) { return (vector short)__builtin_altivec_lvx(__a, __b); } static __inline__ vector unsigned short __ATTRS_o_ai -vec_lvx(int __a, const vector unsigned short *__b) { +vec_lvx(long __a, const vector unsigned short *__b) { return (vector unsigned short)__builtin_altivec_lvx(__a, __b); } static __inline__ vector unsigned short __ATTRS_o_ai -vec_lvx(int __a, const unsigned short *__b) { +vec_lvx(long __a, const unsigned short *__b) { return (vector unsigned short)__builtin_altivec_lvx(__a, __b); } static __inline__ vector bool short __ATTRS_o_ai -vec_lvx(int __a, const vector bool short *__b) { +vec_lvx(long __a, const vector bool short *__b) { return (vector bool short)__builtin_altivec_lvx(__a, __b); } -static __inline__ vector pixel __ATTRS_o_ai vec_lvx(int __a, +static __inline__ vector pixel __ATTRS_o_ai vec_lvx(long __a, const vector pixel *__b) { return (vector pixel)__builtin_altivec_lvx(__a, __b); } -static __inline__ vector int __ATTRS_o_ai vec_lvx(int __a, +static __inline__ vector int __ATTRS_o_ai vec_lvx(long __a, const vector int *__b) { return (vector int)__builtin_altivec_lvx(__a, __b); } -static __inline__ vector int __ATTRS_o_ai vec_lvx(int __a, const int *__b) { +static __inline__ vector int __ATTRS_o_ai vec_lvx(long __a, const int *__b) { return (vector int)__builtin_altivec_lvx(__a, __b); } static __inline__ vector unsigned int __ATTRS_o_ai -vec_lvx(int __a, const vector unsigned int *__b) { +vec_lvx(long __a, const vector unsigned int *__b) { return (vector unsigned int)__builtin_altivec_lvx(__a, __b); } static __inline__ vector unsigned int __ATTRS_o_ai -vec_lvx(int __a, const unsigned int *__b) { +vec_lvx(long __a, const unsigned int *__b) { return (vector unsigned int)__builtin_altivec_lvx(__a, __b); } static __inline__ vector bool int __ATTRS_o_ai -vec_lvx(int __a, const vector bool int *__b) { +vec_lvx(long __a, const vector bool int *__b) { return (vector bool int)__builtin_altivec_lvx(__a, __b); } -static __inline__ vector float __ATTRS_o_ai vec_lvx(int __a, +static __inline__ vector float __ATTRS_o_ai vec_lvx(long __a, const vector float *__b) { return (vector float)__builtin_altivec_lvx(__a, __b); } -static __inline__ vector float __ATTRS_o_ai vec_lvx(int __a, const float *__b) { +static __inline__ vector float __ATTRS_o_ai vec_lvx(long __a, const float *__b) { return (vector float)__builtin_altivec_lvx(__a, __b); } /* vec_lde */ static __inline__ vector signed char __ATTRS_o_ai -vec_lde(int __a, const signed char *__b) { +vec_lde(long __a, const signed char *__b) { return (vector signed char)__builtin_altivec_lvebx(__a, __b); } static __inline__ vector unsigned char __ATTRS_o_ai -vec_lde(int __a, const unsigned char *__b) { +vec_lde(long __a, const unsigned char *__b) { return (vector unsigned char)__builtin_altivec_lvebx(__a, __b); } -static __inline__ vector short __ATTRS_o_ai vec_lde(int __a, const short *__b) { +static __inline__ vector short __ATTRS_o_ai vec_lde(long __a, const short *__b) { return (vector short)__builtin_altivec_lvehx(__a, __b); } static __inline__ vector unsigned short __ATTRS_o_ai -vec_lde(int __a, const unsigned short *__b) { +vec_lde(long __a, const unsigned short *__b) { return (vector unsigned short)__builtin_altivec_lvehx(__a, __b); } -static __inline__ vector int __ATTRS_o_ai vec_lde(int __a, const int *__b) { +static __inline__ vector int __ATTRS_o_ai vec_lde(long __a, const int *__b) { return (vector int)__builtin_altivec_lvewx(__a, __b); } static __inline__ vector unsigned int __ATTRS_o_ai -vec_lde(int __a, const unsigned int *__b) { +vec_lde(long __a, const unsigned int *__b) { return (vector unsigned int)__builtin_altivec_lvewx(__a, __b); } -static __inline__ vector float __ATTRS_o_ai vec_lde(int __a, const float *__b) { +static __inline__ vector float __ATTRS_o_ai vec_lde(long __a, const float *__b) { return (vector float)__builtin_altivec_lvewx(__a, __b); } /* vec_lvebx */ static __inline__ vector signed char __ATTRS_o_ai -vec_lvebx(int __a, const signed char *__b) { +vec_lvebx(long __a, const signed char *__b) { return (vector signed char)__builtin_altivec_lvebx(__a, __b); } static __inline__ vector unsigned char __ATTRS_o_ai -vec_lvebx(int __a, const unsigned char *__b) { +vec_lvebx(long __a, const unsigned char *__b) { return (vector unsigned char)__builtin_altivec_lvebx(__a, __b); } /* vec_lvehx */ -static __inline__ vector short __ATTRS_o_ai vec_lvehx(int __a, +static __inline__ vector short __ATTRS_o_ai vec_lvehx(long __a, const short *__b) { return (vector short)__builtin_altivec_lvehx(__a, __b); } static __inline__ vector unsigned short __ATTRS_o_ai -vec_lvehx(int __a, const unsigned short *__b) { +vec_lvehx(long __a, const unsigned short *__b) { return (vector unsigned short)__builtin_altivec_lvehx(__a, __b); } /* vec_lvewx */ -static __inline__ vector int __ATTRS_o_ai vec_lvewx(int __a, const int *__b) { +static __inline__ vector int __ATTRS_o_ai vec_lvewx(long __a, const int *__b) { return (vector int)__builtin_altivec_lvewx(__a, __b); } static __inline__ vector unsigned int __ATTRS_o_ai -vec_lvewx(int __a, const unsigned int *__b) { +vec_lvewx(long __a, const unsigned int *__b) { return (vector unsigned int)__builtin_altivec_lvewx(__a, __b); } -static __inline__ vector float __ATTRS_o_ai vec_lvewx(int __a, +static __inline__ vector float __ATTRS_o_ai vec_lvewx(long __a, const float *__b) { return (vector float)__builtin_altivec_lvewx(__a, __b); } @@ -3923,179 +4258,179 @@ static __inline__ vector float __ATTRS_o_ai vec_lvewx(int __a, /* vec_ldl */ static __inline__ vector signed char __ATTRS_o_ai -vec_ldl(int __a, const vector signed char *__b) { +vec_ldl(long __a, const vector signed char *__b) { return (vector signed char)__builtin_altivec_lvxl(__a, __b); } static __inline__ vector signed char __ATTRS_o_ai -vec_ldl(int __a, const signed char *__b) { +vec_ldl(long __a, const signed char *__b) { return (vector signed char)__builtin_altivec_lvxl(__a, __b); } static __inline__ vector unsigned char __ATTRS_o_ai -vec_ldl(int __a, const vector unsigned char *__b) { +vec_ldl(long __a, const vector unsigned char *__b) { return (vector unsigned char)__builtin_altivec_lvxl(__a, __b); } static __inline__ vector unsigned char __ATTRS_o_ai -vec_ldl(int __a, const unsigned char *__b) { +vec_ldl(long __a, const unsigned char *__b) { return (vector unsigned char)__builtin_altivec_lvxl(__a, __b); } static __inline__ vector bool char __ATTRS_o_ai -vec_ldl(int __a, const vector bool char *__b) { +vec_ldl(long __a, const vector bool char *__b) { return (vector bool char)__builtin_altivec_lvxl(__a, __b); } -static __inline__ vector short __ATTRS_o_ai vec_ldl(int __a, +static __inline__ vector short __ATTRS_o_ai vec_ldl(long __a, const vector short *__b) { return (vector short)__builtin_altivec_lvxl(__a, __b); } -static __inline__ vector short __ATTRS_o_ai vec_ldl(int __a, const short *__b) { +static __inline__ vector short __ATTRS_o_ai vec_ldl(long __a, const short *__b) { return (vector short)__builtin_altivec_lvxl(__a, __b); } static __inline__ vector unsigned short __ATTRS_o_ai -vec_ldl(int __a, const vector unsigned short *__b) { +vec_ldl(long __a, const vector unsigned short *__b) { return (vector unsigned short)__builtin_altivec_lvxl(__a, __b); } static __inline__ vector unsigned short __ATTRS_o_ai -vec_ldl(int __a, const unsigned short *__b) { +vec_ldl(long __a, const unsigned short *__b) { return (vector unsigned short)__builtin_altivec_lvxl(__a, __b); } static __inline__ vector bool short __ATTRS_o_ai -vec_ldl(int __a, const vector bool short *__b) { +vec_ldl(long __a, const vector bool short *__b) { return (vector bool short)__builtin_altivec_lvxl(__a, __b); } -static __inline__ vector pixel __ATTRS_o_ai vec_ldl(int __a, +static __inline__ vector pixel __ATTRS_o_ai vec_ldl(long __a, const vector pixel *__b) { return (vector pixel short)__builtin_altivec_lvxl(__a, __b); } -static __inline__ vector int __ATTRS_o_ai vec_ldl(int __a, +static __inline__ vector int __ATTRS_o_ai vec_ldl(long __a, const vector int *__b) { return (vector int)__builtin_altivec_lvxl(__a, __b); } -static __inline__ vector int __ATTRS_o_ai vec_ldl(int __a, const int *__b) { +static __inline__ vector int __ATTRS_o_ai vec_ldl(long __a, const int *__b) { return (vector int)__builtin_altivec_lvxl(__a, __b); } static __inline__ vector unsigned int __ATTRS_o_ai -vec_ldl(int __a, const vector unsigned int *__b) { +vec_ldl(long __a, const vector unsigned int *__b) { return (vector unsigned int)__builtin_altivec_lvxl(__a, __b); } static __inline__ vector unsigned int __ATTRS_o_ai -vec_ldl(int __a, const unsigned int *__b) { +vec_ldl(long __a, const unsigned int *__b) { return (vector unsigned int)__builtin_altivec_lvxl(__a, __b); } static __inline__ vector bool int __ATTRS_o_ai -vec_ldl(int __a, const vector bool int *__b) { +vec_ldl(long __a, const vector bool int *__b) { return (vector bool int)__builtin_altivec_lvxl(__a, __b); } -static __inline__ vector float __ATTRS_o_ai vec_ldl(int __a, +static __inline__ vector float __ATTRS_o_ai vec_ldl(long __a, const vector float *__b) { return (vector float)__builtin_altivec_lvxl(__a, __b); } -static __inline__ vector float __ATTRS_o_ai vec_ldl(int __a, const float *__b) { +static __inline__ vector float __ATTRS_o_ai vec_ldl(long __a, const float *__b) { return (vector float)__builtin_altivec_lvxl(__a, __b); } /* vec_lvxl */ static __inline__ vector signed char __ATTRS_o_ai -vec_lvxl(int __a, const vector signed char *__b) { +vec_lvxl(long __a, const vector signed char *__b) { return (vector signed char)__builtin_altivec_lvxl(__a, __b); } static __inline__ vector signed char __ATTRS_o_ai -vec_lvxl(int __a, const signed char *__b) { +vec_lvxl(long __a, const signed char *__b) { return (vector signed char)__builtin_altivec_lvxl(__a, __b); } static __inline__ vector unsigned char __ATTRS_o_ai -vec_lvxl(int __a, const vector unsigned char *__b) { +vec_lvxl(long __a, const vector unsigned char *__b) { return (vector unsigned char)__builtin_altivec_lvxl(__a, __b); } static __inline__ vector unsigned char __ATTRS_o_ai -vec_lvxl(int __a, const unsigned char *__b) { +vec_lvxl(long __a, const unsigned char *__b) { return (vector unsigned char)__builtin_altivec_lvxl(__a, __b); } static __inline__ vector bool char __ATTRS_o_ai -vec_lvxl(int __a, const vector bool char *__b) { +vec_lvxl(long __a, const vector bool char *__b) { return (vector bool char)__builtin_altivec_lvxl(__a, __b); } -static __inline__ vector short __ATTRS_o_ai vec_lvxl(int __a, +static __inline__ vector short __ATTRS_o_ai vec_lvxl(long __a, const vector short *__b) { return (vector short)__builtin_altivec_lvxl(__a, __b); } -static __inline__ vector short __ATTRS_o_ai vec_lvxl(int __a, +static __inline__ vector short __ATTRS_o_ai vec_lvxl(long __a, const short *__b) { return (vector short)__builtin_altivec_lvxl(__a, __b); } static __inline__ vector unsigned short __ATTRS_o_ai -vec_lvxl(int __a, const vector unsigned short *__b) { +vec_lvxl(long __a, const vector unsigned short *__b) { return (vector unsigned short)__builtin_altivec_lvxl(__a, __b); } static __inline__ vector unsigned short __ATTRS_o_ai -vec_lvxl(int __a, const unsigned short *__b) { +vec_lvxl(long __a, const unsigned short *__b) { return (vector unsigned short)__builtin_altivec_lvxl(__a, __b); } static __inline__ vector bool short __ATTRS_o_ai -vec_lvxl(int __a, const vector bool short *__b) { +vec_lvxl(long __a, const vector bool short *__b) { return (vector bool short)__builtin_altivec_lvxl(__a, __b); } -static __inline__ vector pixel __ATTRS_o_ai vec_lvxl(int __a, +static __inline__ vector pixel __ATTRS_o_ai vec_lvxl(long __a, const vector pixel *__b) { return (vector pixel)__builtin_altivec_lvxl(__a, __b); } -static __inline__ vector int __ATTRS_o_ai vec_lvxl(int __a, +static __inline__ vector int __ATTRS_o_ai vec_lvxl(long __a, const vector int *__b) { return (vector int)__builtin_altivec_lvxl(__a, __b); } -static __inline__ vector int __ATTRS_o_ai vec_lvxl(int __a, const int *__b) { +static __inline__ vector int __ATTRS_o_ai vec_lvxl(long __a, const int *__b) { return (vector int)__builtin_altivec_lvxl(__a, __b); } static __inline__ vector unsigned int __ATTRS_o_ai -vec_lvxl(int __a, const vector unsigned int *__b) { +vec_lvxl(long __a, const vector unsigned int *__b) { return (vector unsigned int)__builtin_altivec_lvxl(__a, __b); } static __inline__ vector unsigned int __ATTRS_o_ai -vec_lvxl(int __a, const unsigned int *__b) { +vec_lvxl(long __a, const unsigned int *__b) { return (vector unsigned int)__builtin_altivec_lvxl(__a, __b); } static __inline__ vector bool int __ATTRS_o_ai -vec_lvxl(int __a, const vector bool int *__b) { +vec_lvxl(long __a, const vector bool int *__b) { return (vector bool int)__builtin_altivec_lvxl(__a, __b); } -static __inline__ vector float __ATTRS_o_ai vec_lvxl(int __a, +static __inline__ vector float __ATTRS_o_ai vec_lvxl(long __a, const vector float *__b) { return (vector float)__builtin_altivec_lvxl(__a, __b); } -static __inline__ vector float __ATTRS_o_ai vec_lvxl(int __a, +static __inline__ vector float __ATTRS_o_ai vec_lvxl(long __a, const float *__b) { return (vector float)__builtin_altivec_lvxl(__a, __b); } @@ -5695,7 +6030,7 @@ vec_msum(vector unsigned short __a, vector unsigned short __b, /* vec_msumc */ -#ifdef __POWER10_VECTOR__ +#if defined(__POWER10_VECTOR__) && defined(__SIZEOF_INT128__) static __inline__ vector unsigned __int128 __ATTRS_o_ai vec_msumc(vector unsigned long long __a, vector unsigned long long __b, vector unsigned __int128 __c) { @@ -5929,7 +6264,7 @@ vec_mule(vector unsigned int __a, vector unsigned int __b) { } #endif -#ifdef __POWER10_VECTOR__ +#if defined(__POWER10_VECTOR__) && defined(__SIZEOF_INT128__) static __inline__ vector signed __int128 __ATTRS_o_ai vec_mule(vector signed long long __a, vector signed long long __b) { #ifdef __LITTLE_ENDIAN__ @@ -6075,7 +6410,7 @@ vec_mulo(vector unsigned int __a, vector unsigned int __b) { } #endif -#ifdef __POWER10_VECTOR__ +#if defined(__POWER10_VECTOR__) && defined(__SIZEOF_INT128__) static __inline__ vector signed __int128 __ATTRS_o_ai vec_mulo(vector signed long long __a, vector signed long long __b) { #ifdef __LITTLE_ENDIAN__ @@ -7927,7 +8262,7 @@ vec_rl(vector unsigned long long __a, vector unsigned long long __b) { } #endif -#ifdef __POWER10_VECTOR__ +#if defined(__POWER10_VECTOR__) && defined(__SIZEOF_INT128__) static __inline__ vector signed __int128 __ATTRS_o_ai vec_rl(vector signed __int128 __a, vector unsigned __int128 __b) { return (__b << __a)|(__b >> ((__CHAR_BIT__ * sizeof(vector signed __int128)) - __a)); @@ -7954,7 +8289,7 @@ vec_rlmi(vector unsigned long long __a, vector unsigned long long __b, } #endif -#ifdef __POWER10_VECTOR__ +#if defined(__POWER10_VECTOR__) && defined(__SIZEOF_INT128__) static __inline__ vector unsigned __int128 __ATTRS_o_ai vec_rlmi(vector unsigned __int128 __a, vector unsigned __int128 __b, vector unsigned __int128 __c) { @@ -7985,7 +8320,7 @@ vec_rlnm(vector unsigned long long __a, vector unsigned long long __b, } #endif -#ifdef __POWER10_VECTOR__ +#if defined(__POWER10_VECTOR__) && defined(__SIZEOF_INT128__) static __inline__ vector unsigned __int128 __ATTRS_o_ai vec_rlnm(vector unsigned __int128 __a, vector unsigned __int128 __b, vector unsigned __int128 __c) { @@ -8060,11 +8395,7 @@ vec_vrlw(vector unsigned int __a, vector unsigned int __b) { /* vec_round */ static __inline__ vector float __ATTRS_o_ai vec_round(vector float __a) { -#ifdef __VSX__ - return __builtin_vsx_xvrspi(__a); -#else return __builtin_altivec_vrfin(__a); -#endif } #ifdef __VSX__ @@ -8082,6 +8413,16 @@ static __inline__ vector double __ATTRS_o_ai vec_rint(vector double __a) { return __builtin_vsx_xvrdpic(__a); } +/* vec_roundc */ + +static __inline__ vector float __ATTRS_o_ai vec_roundc(vector float __a) { + return __builtin_vsx_xvrspic(__a); +} + +static __inline__ vector double __ATTRS_o_ai vec_roundc(vector double __a) { + return __builtin_vsx_xvrdpic(__a); +} + /* vec_nearbyint */ static __inline__ vector float __ATTRS_o_ai vec_nearbyint(vector float __a) { @@ -8128,6 +8469,16 @@ static __inline__ vector double __ATTRS_o_ai vec_rsqrte(vector double __a) { } #endif +static vector float __ATTRS_o_ai vec_rsqrt(vector float __a) { + return __builtin_ppc_rsqrtf(__a); +} + +#ifdef __VSX__ +static vector double __ATTRS_o_ai vec_rsqrt(vector double __a) { + return __builtin_ppc_rsqrtd(__a); +} +#endif + /* vec_vrsqrtefp */ static __inline__ __vector float __attribute__((__always_inline__)) @@ -8484,6 +8835,52 @@ vec_sl(vector unsigned long long __a, vector unsigned long long __b) { __CHAR_BIT__)); } +static __inline__ vector long long __ATTRS_o_ai +vec_sl(vector long long __a, vector unsigned long long __b) { + return (vector long long)vec_sl((vector unsigned long long)__a, __b); +} +#else +static __inline__ vector unsigned char __ATTRS_o_ai +vec_vspltb(vector unsigned char __a, unsigned char __b); +static __inline__ vector unsigned long long __ATTRS_o_ai +vec_sl(vector unsigned long long __a, vector unsigned long long __b) { + __b %= (vector unsigned long long)(sizeof(unsigned long long) * __CHAR_BIT__); + + // Big endian element one (the right doubleword) can be left shifted as-is. + // The other element needs to be swapped into the right doubleword and + // shifted. Then the right doublewords of the two result vectors are merged. + vector signed long long __rightelt = + (vector signed long long)__builtin_altivec_vslo((vector signed int)__a, + (vector signed int)__b); +#ifdef __LITTLE_ENDIAN__ + __rightelt = (vector signed long long)__builtin_altivec_vsl( + (vector signed int)__rightelt, + (vector signed int)vec_vspltb((vector unsigned char)__b, 0)); +#else + __rightelt = (vector signed long long)__builtin_altivec_vsl( + (vector signed int)__rightelt, + (vector signed int)vec_vspltb((vector unsigned char)__b, 15)); +#endif + __a = __builtin_shufflevector(__a, __a, 1, 0); + __b = __builtin_shufflevector(__b, __b, 1, 0); + vector signed long long __leftelt = + (vector signed long long)__builtin_altivec_vslo((vector signed int)__a, + (vector signed int)__b); +#ifdef __LITTLE_ENDIAN__ + __leftelt = (vector signed long long)__builtin_altivec_vsl( + (vector signed int)__leftelt, + (vector signed int)vec_vspltb((vector unsigned char)__b, 0)); + return (vector unsigned long long)__builtin_shufflevector(__rightelt, + __leftelt, 0, 2); +#else + __leftelt = (vector signed long long)__builtin_altivec_vsl( + (vector signed int)__leftelt, + (vector signed int)vec_vspltb((vector unsigned char)__b, 15)); + return (vector unsigned long long)__builtin_shufflevector(__leftelt, + __rightelt, 1, 3); +#endif +} + static __inline__ vector long long __ATTRS_o_ai vec_sl(vector long long __a, vector unsigned long long __b) { return (vector long long)vec_sl((vector unsigned long long)__a, __b); @@ -8850,8 +9247,13 @@ static __inline__ vector unsigned int __ATTRS_o_ai vec_sldw( return vec_sld(__a, __b, ((__c << 2) & 0x0F)); } -#ifdef __VSX__ -static __inline__ vector signed long long __ATTRS_o_ai +static __inline__ vector float __ATTRS_o_ai vec_sldw( + vector float __a, vector float __b, unsigned const int __c) { + return vec_sld(__a, __b, ((__c << 2) & 0x0F)); +} + +#ifdef __VSX__ +static __inline__ vector signed long long __ATTRS_o_ai vec_sldw(vector signed long long __a, vector signed long long __b, unsigned const int __c) { return vec_sld(__a, __b, ((__c << 2) & 0x0F)); @@ -8862,6 +9264,11 @@ vec_sldw(vector unsigned long long __a, vector unsigned long long __b, unsigned const int __c) { return vec_sld(__a, __b, ((__c << 2) & 0x0F)); } + +static __inline__ vector double __ATTRS_o_ai vec_sldw( + vector double __a, vector double __b, unsigned const int __c) { + return vec_sld(__a, __b, ((__c << 2) & 0x0F)); +} #endif #ifdef __POWER9_VECTOR__ @@ -9939,6 +10346,50 @@ vec_sr(vector unsigned long long __a, vector unsigned long long __b) { __CHAR_BIT__)); } +static __inline__ vector long long __ATTRS_o_ai +vec_sr(vector long long __a, vector unsigned long long __b) { + return (vector long long)vec_sr((vector unsigned long long)__a, __b); +} +#else +static __inline__ vector unsigned long long __ATTRS_o_ai +vec_sr(vector unsigned long long __a, vector unsigned long long __b) { + __b %= (vector unsigned long long)(sizeof(unsigned long long) * __CHAR_BIT__); + + // Big endian element zero (the left doubleword) can be right shifted as-is. + // However the shift amount must be in the right doubleword. + // The other element needs to be swapped into the left doubleword and + // shifted. Then the left doublewords of the two result vectors are merged. + vector unsigned long long __swapshift = + __builtin_shufflevector(__b, __b, 1, 0); + vector unsigned long long __leftelt = + (vector unsigned long long)__builtin_altivec_vsro( + (vector signed int)__a, (vector signed int)__swapshift); +#ifdef __LITTLE_ENDIAN__ + __leftelt = (vector unsigned long long)__builtin_altivec_vsr( + (vector signed int)__leftelt, + (vector signed int)vec_vspltb((vector unsigned char)__swapshift, 0)); +#else + __leftelt = (vector unsigned long long)__builtin_altivec_vsr( + (vector signed int)__leftelt, + (vector signed int)vec_vspltb((vector unsigned char)__swapshift, 15)); +#endif + __a = __builtin_shufflevector(__a, __a, 1, 0); + vector unsigned long long __rightelt = + (vector unsigned long long)__builtin_altivec_vsro((vector signed int)__a, + (vector signed int)__b); +#ifdef __LITTLE_ENDIAN__ + __rightelt = (vector unsigned long long)__builtin_altivec_vsr( + (vector signed int)__rightelt, + (vector signed int)vec_vspltb((vector unsigned char)__b, 0)); + return __builtin_shufflevector(__rightelt, __leftelt, 1, 3); +#else + __rightelt = (vector unsigned long long)__builtin_altivec_vsr( + (vector signed int)__rightelt, + (vector signed int)vec_vspltb((vector unsigned char)__b, 15)); + return __builtin_shufflevector(__leftelt, __rightelt, 0, 2); +#endif +} + static __inline__ vector long long __ATTRS_o_ai vec_sr(vector long long __a, vector unsigned long long __b) { return (vector long long)vec_sr((vector unsigned long long)__a, __b); @@ -10029,6 +10480,18 @@ static __inline__ vector unsigned long long __ATTRS_o_ai vec_sra(vector unsigned long long __a, vector unsigned long long __b) { return (vector unsigned long long)((vector signed long long)__a >> __b); } +#else +static __inline__ vector signed long long __ATTRS_o_ai +vec_sra(vector signed long long __a, vector unsigned long long __b) { + __b %= (vector unsigned long long)(sizeof(unsigned long long) * __CHAR_BIT__); + return __a >> __b; +} + +static __inline__ vector unsigned long long __ATTRS_o_ai +vec_sra(vector unsigned long long __a, vector unsigned long long __b) { + __b %= (vector unsigned long long)(sizeof(unsigned long long) * __CHAR_BIT__); + return (vector unsigned long long)((vector signed long long)__a >> __b); +} #endif /* vec_vsrab */ @@ -10635,420 +11098,420 @@ static __inline__ vector float __ATTRS_o_ai vec_vsro(vector float __a, /* vec_st */ -static __inline__ void __ATTRS_o_ai vec_st(vector signed char __a, int __b, +static __inline__ void __ATTRS_o_ai vec_st(vector signed char __a, long __b, vector signed char *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_st(vector signed char __a, int __b, +static __inline__ void __ATTRS_o_ai vec_st(vector signed char __a, long __b, signed char *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_st(vector unsigned char __a, int __b, +static __inline__ void __ATTRS_o_ai vec_st(vector unsigned char __a, long __b, vector unsigned char *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_st(vector unsigned char __a, int __b, +static __inline__ void __ATTRS_o_ai vec_st(vector unsigned char __a, long __b, unsigned char *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_st(vector bool char __a, int __b, +static __inline__ void __ATTRS_o_ai vec_st(vector bool char __a, long __b, signed char *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_st(vector bool char __a, int __b, +static __inline__ void __ATTRS_o_ai vec_st(vector bool char __a, long __b, unsigned char *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_st(vector bool char __a, int __b, +static __inline__ void __ATTRS_o_ai vec_st(vector bool char __a, long __b, vector bool char *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_st(vector short __a, int __b, +static __inline__ void __ATTRS_o_ai vec_st(vector short __a, long __b, vector short *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_st(vector short __a, int __b, +static __inline__ void __ATTRS_o_ai vec_st(vector short __a, long __b, short *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_st(vector unsigned short __a, int __b, +static __inline__ void __ATTRS_o_ai vec_st(vector unsigned short __a, long __b, vector unsigned short *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_st(vector unsigned short __a, int __b, +static __inline__ void __ATTRS_o_ai vec_st(vector unsigned short __a, long __b, unsigned short *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_st(vector bool short __a, int __b, +static __inline__ void __ATTRS_o_ai vec_st(vector bool short __a, long __b, short *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_st(vector bool short __a, int __b, +static __inline__ void __ATTRS_o_ai vec_st(vector bool short __a, long __b, unsigned short *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_st(vector bool short __a, int __b, +static __inline__ void __ATTRS_o_ai vec_st(vector bool short __a, long __b, vector bool short *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_st(vector pixel __a, int __b, +static __inline__ void __ATTRS_o_ai vec_st(vector pixel __a, long __b, short *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_st(vector pixel __a, int __b, +static __inline__ void __ATTRS_o_ai vec_st(vector pixel __a, long __b, unsigned short *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_st(vector pixel __a, int __b, +static __inline__ void __ATTRS_o_ai vec_st(vector pixel __a, long __b, vector pixel *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_st(vector int __a, int __b, +static __inline__ void __ATTRS_o_ai vec_st(vector int __a, long __b, vector int *__c) { __builtin_altivec_stvx(__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_st(vector int __a, int __b, int *__c) { +static __inline__ void __ATTRS_o_ai vec_st(vector int __a, long __b, int *__c) { __builtin_altivec_stvx(__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_st(vector unsigned int __a, int __b, +static __inline__ void __ATTRS_o_ai vec_st(vector unsigned int __a, long __b, vector unsigned int *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_st(vector unsigned int __a, int __b, +static __inline__ void __ATTRS_o_ai vec_st(vector unsigned int __a, long __b, unsigned int *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_st(vector bool int __a, int __b, +static __inline__ void __ATTRS_o_ai vec_st(vector bool int __a, long __b, int *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_st(vector bool int __a, int __b, +static __inline__ void __ATTRS_o_ai vec_st(vector bool int __a, long __b, unsigned int *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_st(vector bool int __a, int __b, +static __inline__ void __ATTRS_o_ai vec_st(vector bool int __a, long __b, vector bool int *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_st(vector float __a, int __b, +static __inline__ void __ATTRS_o_ai vec_st(vector float __a, long __b, vector float *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_st(vector float __a, int __b, +static __inline__ void __ATTRS_o_ai vec_st(vector float __a, long __b, float *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } /* vec_stvx */ -static __inline__ void __ATTRS_o_ai vec_stvx(vector signed char __a, int __b, +static __inline__ void __ATTRS_o_ai vec_stvx(vector signed char __a, long __b, vector signed char *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_stvx(vector signed char __a, int __b, +static __inline__ void __ATTRS_o_ai vec_stvx(vector signed char __a, long __b, signed char *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_stvx(vector unsigned char __a, int __b, +static __inline__ void __ATTRS_o_ai vec_stvx(vector unsigned char __a, long __b, vector unsigned char *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_stvx(vector unsigned char __a, int __b, +static __inline__ void __ATTRS_o_ai vec_stvx(vector unsigned char __a, long __b, unsigned char *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_stvx(vector bool char __a, int __b, +static __inline__ void __ATTRS_o_ai vec_stvx(vector bool char __a, long __b, signed char *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_stvx(vector bool char __a, int __b, +static __inline__ void __ATTRS_o_ai vec_stvx(vector bool char __a, long __b, unsigned char *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_stvx(vector bool char __a, int __b, +static __inline__ void __ATTRS_o_ai vec_stvx(vector bool char __a, long __b, vector bool char *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_stvx(vector short __a, int __b, +static __inline__ void __ATTRS_o_ai vec_stvx(vector short __a, long __b, vector short *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_stvx(vector short __a, int __b, +static __inline__ void __ATTRS_o_ai vec_stvx(vector short __a, long __b, short *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_stvx(vector unsigned short __a, int __b, +static __inline__ void __ATTRS_o_ai vec_stvx(vector unsigned short __a, long __b, vector unsigned short *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_stvx(vector unsigned short __a, int __b, +static __inline__ void __ATTRS_o_ai vec_stvx(vector unsigned short __a, long __b, unsigned short *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_stvx(vector bool short __a, int __b, +static __inline__ void __ATTRS_o_ai vec_stvx(vector bool short __a, long __b, short *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_stvx(vector bool short __a, int __b, +static __inline__ void __ATTRS_o_ai vec_stvx(vector bool short __a, long __b, unsigned short *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_stvx(vector bool short __a, int __b, +static __inline__ void __ATTRS_o_ai vec_stvx(vector bool short __a, long __b, vector bool short *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_stvx(vector pixel __a, int __b, +static __inline__ void __ATTRS_o_ai vec_stvx(vector pixel __a, long __b, short *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_stvx(vector pixel __a, int __b, +static __inline__ void __ATTRS_o_ai vec_stvx(vector pixel __a, long __b, unsigned short *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_stvx(vector pixel __a, int __b, +static __inline__ void __ATTRS_o_ai vec_stvx(vector pixel __a, long __b, vector pixel *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_stvx(vector int __a, int __b, +static __inline__ void __ATTRS_o_ai vec_stvx(vector int __a, long __b, vector int *__c) { __builtin_altivec_stvx(__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_stvx(vector int __a, int __b, +static __inline__ void __ATTRS_o_ai vec_stvx(vector int __a, long __b, int *__c) { __builtin_altivec_stvx(__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_stvx(vector unsigned int __a, int __b, +static __inline__ void __ATTRS_o_ai vec_stvx(vector unsigned int __a, long __b, vector unsigned int *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_stvx(vector unsigned int __a, int __b, +static __inline__ void __ATTRS_o_ai vec_stvx(vector unsigned int __a, long __b, unsigned int *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_stvx(vector bool int __a, int __b, +static __inline__ void __ATTRS_o_ai vec_stvx(vector bool int __a, long __b, int *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_stvx(vector bool int __a, int __b, +static __inline__ void __ATTRS_o_ai vec_stvx(vector bool int __a, long __b, unsigned int *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_stvx(vector bool int __a, int __b, +static __inline__ void __ATTRS_o_ai vec_stvx(vector bool int __a, long __b, vector bool int *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_stvx(vector float __a, int __b, +static __inline__ void __ATTRS_o_ai vec_stvx(vector float __a, long __b, vector float *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_stvx(vector float __a, int __b, +static __inline__ void __ATTRS_o_ai vec_stvx(vector float __a, long __b, float *__c) { __builtin_altivec_stvx((vector int)__a, __b, __c); } /* vec_ste */ -static __inline__ void __ATTRS_o_ai vec_ste(vector signed char __a, int __b, +static __inline__ void __ATTRS_o_ai vec_ste(vector signed char __a, long __b, signed char *__c) { __builtin_altivec_stvebx((vector char)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_ste(vector unsigned char __a, int __b, +static __inline__ void __ATTRS_o_ai vec_ste(vector unsigned char __a, long __b, unsigned char *__c) { __builtin_altivec_stvebx((vector char)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_ste(vector bool char __a, int __b, +static __inline__ void __ATTRS_o_ai vec_ste(vector bool char __a, long __b, signed char *__c) { __builtin_altivec_stvebx((vector char)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_ste(vector bool char __a, int __b, +static __inline__ void __ATTRS_o_ai vec_ste(vector bool char __a, long __b, unsigned char *__c) { __builtin_altivec_stvebx((vector char)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_ste(vector short __a, int __b, +static __inline__ void __ATTRS_o_ai vec_ste(vector short __a, long __b, short *__c) { __builtin_altivec_stvehx(__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_ste(vector unsigned short __a, int __b, +static __inline__ void __ATTRS_o_ai vec_ste(vector unsigned short __a, long __b, unsigned short *__c) { __builtin_altivec_stvehx((vector short)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_ste(vector bool short __a, int __b, +static __inline__ void __ATTRS_o_ai vec_ste(vector bool short __a, long __b, short *__c) { __builtin_altivec_stvehx((vector short)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_ste(vector bool short __a, int __b, +static __inline__ void __ATTRS_o_ai vec_ste(vector bool short __a, long __b, unsigned short *__c) { __builtin_altivec_stvehx((vector short)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_ste(vector pixel __a, int __b, +static __inline__ void __ATTRS_o_ai vec_ste(vector pixel __a, long __b, short *__c) { __builtin_altivec_stvehx((vector short)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_ste(vector pixel __a, int __b, +static __inline__ void __ATTRS_o_ai vec_ste(vector pixel __a, long __b, unsigned short *__c) { __builtin_altivec_stvehx((vector short)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_ste(vector int __a, int __b, int *__c) { +static __inline__ void __ATTRS_o_ai vec_ste(vector int __a, long __b, int *__c) { __builtin_altivec_stvewx(__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_ste(vector unsigned int __a, int __b, +static __inline__ void __ATTRS_o_ai vec_ste(vector unsigned int __a, long __b, unsigned int *__c) { __builtin_altivec_stvewx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_ste(vector bool int __a, int __b, +static __inline__ void __ATTRS_o_ai vec_ste(vector bool int __a, long __b, int *__c) { __builtin_altivec_stvewx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_ste(vector bool int __a, int __b, +static __inline__ void __ATTRS_o_ai vec_ste(vector bool int __a, long __b, unsigned int *__c) { __builtin_altivec_stvewx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_ste(vector float __a, int __b, +static __inline__ void __ATTRS_o_ai vec_ste(vector float __a, long __b, float *__c) { __builtin_altivec_stvewx((vector int)__a, __b, __c); } /* vec_stvebx */ -static __inline__ void __ATTRS_o_ai vec_stvebx(vector signed char __a, int __b, +static __inline__ void __ATTRS_o_ai vec_stvebx(vector signed char __a, long __b, signed char *__c) { __builtin_altivec_stvebx((vector char)__a, __b, __c); } static __inline__ void __ATTRS_o_ai vec_stvebx(vector unsigned char __a, - int __b, unsigned char *__c) { + long __b, unsigned char *__c) { __builtin_altivec_stvebx((vector char)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_stvebx(vector bool char __a, int __b, +static __inline__ void __ATTRS_o_ai vec_stvebx(vector bool char __a, long __b, signed char *__c) { __builtin_altivec_stvebx((vector char)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_stvebx(vector bool char __a, int __b, +static __inline__ void __ATTRS_o_ai vec_stvebx(vector bool char __a, long __b, unsigned char *__c) { __builtin_altivec_stvebx((vector char)__a, __b, __c); } /* vec_stvehx */ -static __inline__ void __ATTRS_o_ai vec_stvehx(vector short __a, int __b, +static __inline__ void __ATTRS_o_ai vec_stvehx(vector short __a, long __b, short *__c) { __builtin_altivec_stvehx(__a, __b, __c); } static __inline__ void __ATTRS_o_ai vec_stvehx(vector unsigned short __a, - int __b, unsigned short *__c) { + long __b, unsigned short *__c) { __builtin_altivec_stvehx((vector short)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_stvehx(vector bool short __a, int __b, +static __inline__ void __ATTRS_o_ai vec_stvehx(vector bool short __a, long __b, short *__c) { __builtin_altivec_stvehx((vector short)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_stvehx(vector bool short __a, int __b, +static __inline__ void __ATTRS_o_ai vec_stvehx(vector bool short __a, long __b, unsigned short *__c) { __builtin_altivec_stvehx((vector short)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_stvehx(vector pixel __a, int __b, +static __inline__ void __ATTRS_o_ai vec_stvehx(vector pixel __a, long __b, short *__c) { __builtin_altivec_stvehx((vector short)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_stvehx(vector pixel __a, int __b, +static __inline__ void __ATTRS_o_ai vec_stvehx(vector pixel __a, long __b, unsigned short *__c) { __builtin_altivec_stvehx((vector short)__a, __b, __c); } /* vec_stvewx */ -static __inline__ void __ATTRS_o_ai vec_stvewx(vector int __a, int __b, +static __inline__ void __ATTRS_o_ai vec_stvewx(vector int __a, long __b, int *__c) { __builtin_altivec_stvewx(__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_stvewx(vector unsigned int __a, int __b, +static __inline__ void __ATTRS_o_ai vec_stvewx(vector unsigned int __a, long __b, unsigned int *__c) { __builtin_altivec_stvewx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_stvewx(vector bool int __a, int __b, +static __inline__ void __ATTRS_o_ai vec_stvewx(vector bool int __a, long __b, int *__c) { __builtin_altivec_stvewx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_stvewx(vector bool int __a, int __b, +static __inline__ void __ATTRS_o_ai vec_stvewx(vector bool int __a, long __b, unsigned int *__c) { __builtin_altivec_stvewx((vector int)__a, __b, __c); } -static __inline__ void __ATTRS_o_ai vec_stvewx(vector float __a, int __b, +static __inline__ void __ATTRS_o_ai vec_stvewx(vector float __a, long __b, float *__c) { __builtin_altivec_stvewx((vector int)__a, __b, __c); } @@ -11409,7 +11872,8 @@ vec_sub(vector unsigned int __a, vector bool int __b) { return __a - (vector unsigned int)__b; } -#if defined(__POWER8_VECTOR__) && defined(__powerpc64__) +#if defined(__POWER8_VECTOR__) && defined(__powerpc64__) && \ + defined(__SIZEOF_INT128__) static __inline__ vector signed __int128 __ATTRS_o_ai vec_sub(vector signed __int128 __a, vector signed __int128 __b) { return __a - __b; @@ -11419,7 +11883,8 @@ static __inline__ vector unsigned __int128 __ATTRS_o_ai vec_sub(vector unsigned __int128 __a, vector unsigned __int128 __b) { return __a - __b; } -#endif // defined(__POWER8_VECTOR__) && defined(__powerpc64__) +#endif // defined(__POWER8_VECTOR__) && defined(__powerpc64__) && + // defined(__SIZEOF_INT128__) #ifdef __VSX__ static __inline__ vector signed long long __ATTRS_o_ai @@ -11567,7 +12032,8 @@ vec_subc(vector unsigned int __a, vector unsigned int __b) { return __builtin_altivec_vsubcuw(__a, __b); } -#if defined(__POWER8_VECTOR__) && defined(__powerpc64__) +#ifdef __POWER8_VECTOR__ +#ifdef __SIZEOF_INT128__ static __inline__ vector unsigned __int128 __ATTRS_o_ai vec_subc(vector unsigned __int128 __a, vector unsigned __int128 __b) { return __builtin_altivec_vsubcuq(__a, __b); @@ -11577,7 +12043,13 @@ static __inline__ vector signed __int128 __ATTRS_o_ai vec_subc(vector signed __int128 __a, vector signed __int128 __b) { return __builtin_altivec_vsubcuq(__a, __b); } -#endif // defined(__POWER8_VECTOR__) && defined(__powerpc64__) +#endif + +static __inline__ vector unsigned char __attribute__((__always_inline__)) +vec_subc_u128(vector unsigned char __a, vector unsigned char __b) { + return (vector unsigned char)__builtin_altivec_vsubcuq(__a, __b); +} +#endif // __POWER8_VECTOR__ /* vec_vsubcuw */ @@ -11780,9 +12252,10 @@ vec_vsubuws(vector unsigned int __a, vector bool int __b) { return __builtin_altivec_vsubuws(__a, (vector unsigned int)__b); } -#if defined(__POWER8_VECTOR__) && defined(__powerpc64__) +#ifdef __POWER8_VECTOR__ /* vec_vsubuqm */ +#ifdef __SIZEOF_INT128__ static __inline__ vector signed __int128 __ATTRS_o_ai vec_vsubuqm(vector signed __int128 __a, vector signed __int128 __b) { return __a - __b; @@ -11792,10 +12265,16 @@ static __inline__ vector unsigned __int128 __ATTRS_o_ai vec_vsubuqm(vector unsigned __int128 __a, vector unsigned __int128 __b) { return __a - __b; } +#endif -/* vec_vsubeuqm */ +static __inline__ vector unsigned char __attribute__((__always_inline__)) +vec_sub_u128(vector unsigned char __a, vector unsigned char __b) { + return __builtin_altivec_vsubuqm(__a, __b); +} +/* vec_vsubeuqm */ +#ifdef __SIZEOF_INT128__ static __inline__ vector signed __int128 __ATTRS_o_ai vec_vsubeuqm(vector signed __int128 __a, vector signed __int128 __b, vector signed __int128 __c) { @@ -11819,9 +12298,17 @@ vec_sube(vector unsigned __int128 __a, vector unsigned __int128 __b, vector unsigned __int128 __c) { return __builtin_altivec_vsubeuqm(__a, __b, __c); } +#endif + +static __inline__ vector unsigned char __attribute__((__always_inline__)) +vec_sube_u128(vector unsigned char __a, vector unsigned char __b, + vector unsigned char __c) { + return (vector unsigned char)__builtin_altivec_vsubeuqm(__a, __b, __c); +} /* vec_vsubcuq */ +#ifdef __SIZEOF_INT128__ static __inline__ vector signed __int128 __ATTRS_o_ai vec_vsubcuq(vector signed __int128 __a, vector signed __int128 __b) { return __builtin_altivec_vsubcuq(__a, __b); @@ -11845,7 +12332,9 @@ vec_vsubecuq(vector unsigned __int128 __a, vector unsigned __int128 __b, vector unsigned __int128 __c) { return __builtin_altivec_vsubecuq(__a, __b, __c); } +#endif +#ifdef __powerpc64__ static __inline__ vector signed int __ATTRS_o_ai vec_subec(vector signed int __a, vector signed int __b, vector signed int __c) { @@ -11857,7 +12346,9 @@ vec_subec(vector unsigned int __a, vector unsigned int __b, vector unsigned int __c) { return vec_addec(__a, ~__b, __c); } +#endif +#ifdef __SIZEOF_INT128__ static __inline__ vector signed __int128 __ATTRS_o_ai vec_subec(vector signed __int128 __a, vector signed __int128 __b, vector signed __int128 __c) { @@ -11869,7 +12360,14 @@ vec_subec(vector unsigned __int128 __a, vector unsigned __int128 __b, vector unsigned __int128 __c) { return __builtin_altivec_vsubecuq(__a, __b, __c); } -#endif // defined(__POWER8_VECTOR__) && defined(__powerpc64__) +#endif + +static __inline__ vector unsigned char __attribute__((__always_inline__)) +vec_subec_u128(vector unsigned char __a, vector unsigned char __b, + vector unsigned char __c) { + return (vector unsigned char)__builtin_altivec_vsubecuq(__a, __b, __c); +} +#endif // __POWER8_VECTOR__ static __inline__ vector signed int __ATTRS_o_ai vec_sube(vector signed int __a, vector signed int __b, @@ -12012,6 +12510,17 @@ static __inline__ vector double __ATTRS_o_ai vec_trunc(vector double __a) { } #endif +/* vec_roundz */ +static __inline__ vector float __ATTRS_o_ai vec_roundz(vector float __a) { + return vec_trunc(__a); +} + +#ifdef __VSX__ +static __inline__ vector double __ATTRS_o_ai vec_roundz(vector double __a) { + return vec_trunc(__a); +} +#endif + /* vec_vrfiz */ static __inline__ vector float __attribute__((__always_inline__)) @@ -12023,6 +12532,13 @@ vec_vrfiz(vector float __a) { /* The vector unpack instructions all have a big-endian bias, so for little endian we must reverse the meanings of "high" and "low." */ +#ifdef __LITTLE_ENDIAN__ +#define vec_vupkhpx(__a) __builtin_altivec_vupklpx((vector short)(__a)) +#define vec_vupklpx(__a) __builtin_altivec_vupkhpx((vector short)(__a)) +#else +#define vec_vupkhpx(__a) __builtin_altivec_vupkhpx((vector short)(__a)) +#define vec_vupklpx(__a) __builtin_altivec_vupklpx((vector short)(__a)) +#endif static __inline__ vector short __ATTRS_o_ai vec_unpackh(vector signed char __a) { @@ -12558,6 +13074,16 @@ static __inline__ void __ATTRS_o_ai vec_vsx_st(vector unsigned char __a, #ifdef __VSX__ #define vec_xxpermdi __builtin_vsx_xxpermdi #define vec_xxsldwi __builtin_vsx_xxsldwi +#define vec_permi(__a, __b, __c) \ + _Generic((__a), vector signed long long \ + : __builtin_shufflevector((__a), (__b), (((__c) >> 1) & 0x1), \ + (((__c)&0x1) + 2)), \ + vector unsigned long long \ + : __builtin_shufflevector((__a), (__b), (((__c) >> 1) & 0x1), \ + (((__c)&0x1) + 2)), \ + vector double \ + : __builtin_shufflevector((__a), (__b), (((__c) >> 1) & 0x1), \ + (((__c)&0x1) + 2))) #endif /* vec_xor */ @@ -12915,73 +13441,75 @@ vec_vxor(vector bool long long __a, vector bool long long __b) { /* vec_extract */ static __inline__ signed char __ATTRS_o_ai vec_extract(vector signed char __a, - int __b) { - return __a[__b]; + unsigned int __b) { + return __a[__b & 0xf]; } static __inline__ unsigned char __ATTRS_o_ai -vec_extract(vector unsigned char __a, int __b) { - return __a[__b]; +vec_extract(vector unsigned char __a, unsigned int __b) { + return __a[__b & 0xf]; } static __inline__ unsigned char __ATTRS_o_ai vec_extract(vector bool char __a, - int __b) { - return __a[__b]; + unsigned int __b) { + return __a[__b & 0xf]; } static __inline__ signed short __ATTRS_o_ai vec_extract(vector signed short __a, - int __b) { - return __a[__b]; + unsigned int __b) { + return __a[__b & 0x7]; } static __inline__ unsigned short __ATTRS_o_ai -vec_extract(vector unsigned short __a, int __b) { - return __a[__b]; +vec_extract(vector unsigned short __a, unsigned int __b) { + return __a[__b & 0x7]; } static __inline__ unsigned short __ATTRS_o_ai vec_extract(vector bool short __a, - int __b) { - return __a[__b]; + unsigned int __b) { + return __a[__b & 0x7]; } static __inline__ signed int __ATTRS_o_ai vec_extract(vector signed int __a, - int __b) { - return __a[__b]; + unsigned int __b) { + return __a[__b & 0x3]; } static __inline__ unsigned int __ATTRS_o_ai vec_extract(vector unsigned int __a, - int __b) { - return __a[__b]; + unsigned int __b) { + return __a[__b & 0x3]; } static __inline__ unsigned int __ATTRS_o_ai vec_extract(vector bool int __a, - int __b) { - return __a[__b]; + unsigned int __b) { + return __a[__b & 0x3]; } #ifdef __VSX__ static __inline__ signed long long __ATTRS_o_ai -vec_extract(vector signed long long __a, int __b) { - return __a[__b]; +vec_extract(vector signed long long __a, unsigned int __b) { + return __a[__b & 0x1]; } static __inline__ unsigned long long __ATTRS_o_ai -vec_extract(vector unsigned long long __a, int __b) { - return __a[__b]; +vec_extract(vector unsigned long long __a, unsigned int __b) { + return __a[__b & 0x1]; } static __inline__ unsigned long long __ATTRS_o_ai -vec_extract(vector bool long long __a, int __b) { - return __a[__b]; +vec_extract(vector bool long long __a, unsigned int __b) { + return __a[__b & 0x1]; } -static __inline__ double __ATTRS_o_ai vec_extract(vector double __a, int __b) { - return __a[__b]; +static __inline__ double __ATTRS_o_ai vec_extract(vector double __a, + unsigned int __b) { + return __a[__b & 0x1]; } #endif -static __inline__ float __ATTRS_o_ai vec_extract(vector float __a, int __b) { - return __a[__b]; +static __inline__ float __ATTRS_o_ai vec_extract(vector float __a, + unsigned int __b) { + return __a[__b & 0x3]; } #ifdef __POWER9_VECTOR__ @@ -14022,49 +14550,71 @@ static __inline__ void __ATTRS_o_ai vec_stvrxl(vector float __a, int __b, static __inline__ vector signed char __ATTRS_o_ai vec_promote(signed char __a, int __b) { vector signed char __res = (vector signed char)(0); - __res[__b] = __a; + __res[__b & 0x7] = __a; return __res; } static __inline__ vector unsigned char __ATTRS_o_ai vec_promote(unsigned char __a, int __b) { vector unsigned char __res = (vector unsigned char)(0); - __res[__b] = __a; + __res[__b & 0x7] = __a; return __res; } static __inline__ vector short __ATTRS_o_ai vec_promote(short __a, int __b) { vector short __res = (vector short)(0); - __res[__b] = __a; + __res[__b & 0x7] = __a; return __res; } static __inline__ vector unsigned short __ATTRS_o_ai vec_promote(unsigned short __a, int __b) { vector unsigned short __res = (vector unsigned short)(0); - __res[__b] = __a; + __res[__b & 0x7] = __a; return __res; } static __inline__ vector int __ATTRS_o_ai vec_promote(int __a, int __b) { vector int __res = (vector int)(0); - __res[__b] = __a; + __res[__b & 0x3] = __a; return __res; } static __inline__ vector unsigned int __ATTRS_o_ai vec_promote(unsigned int __a, int __b) { vector unsigned int __res = (vector unsigned int)(0); - __res[__b] = __a; + __res[__b & 0x3] = __a; return __res; } static __inline__ vector float __ATTRS_o_ai vec_promote(float __a, int __b) { vector float __res = (vector float)(0); - __res[__b] = __a; + __res[__b & 0x3] = __a; + return __res; +} + +#ifdef __VSX__ +static __inline__ vector double __ATTRS_o_ai vec_promote(double __a, int __b) { + vector double __res = (vector double)(0); + __res[__b & 0x1] = __a; + return __res; +} + +static __inline__ vector signed long long __ATTRS_o_ai +vec_promote(signed long long __a, int __b) { + vector signed long long __res = (vector signed long long)(0); + __res[__b & 0x1] = __a; return __res; } +static __inline__ vector unsigned long long __ATTRS_o_ai +vec_promote(unsigned long long __a, int __b) { + vector unsigned long long __res = (vector unsigned long long)(0); + __res[__b & 0x1] = __a; + return __res; +} +#endif + /* vec_splats */ static __inline__ vector signed char __ATTRS_o_ai vec_splats(signed char __a) { @@ -14105,7 +14655,8 @@ vec_splats(unsigned long long __a) { return (vector unsigned long long)(__a); } -#if defined(__POWER8_VECTOR__) && defined(__powerpc64__) +#if defined(__POWER8_VECTOR__) && defined(__powerpc64__) && \ + defined(__SIZEOF_INT128__) static __inline__ vector signed __int128 __ATTRS_o_ai vec_splats(signed __int128 __a) { return (vector signed __int128)(__a); @@ -14258,7 +14809,7 @@ static __inline__ int __ATTRS_o_ai vec_all_eq(vector bool int __a, (vector int)__b); } -#ifdef __POWER8_VECTOR__ +#ifdef __VSX__ static __inline__ int __ATTRS_o_ai vec_all_eq(vector signed long long __a, vector signed long long __b) { return __builtin_altivec_vcmpequd_p(__CR6_LT, __a, __b); @@ -14316,7 +14867,7 @@ static __inline__ int __ATTRS_o_ai vec_all_eq(vector double __a, } #endif -#ifdef __POWER10_VECTOR__ +#if defined(__POWER10_VECTOR__) && defined(__SIZEOF_INT128__) static __inline__ int __ATTRS_o_ai vec_all_eq(vector signed __int128 __a, vector signed __int128 __b) { return __builtin_altivec_vcmpequq_p(__CR6_LT, __a, __b); @@ -14352,8 +14903,7 @@ static __inline__ int __ATTRS_o_ai vec_all_ge(vector unsigned char __a, static __inline__ int __ATTRS_o_ai vec_all_ge(vector bool char __a, vector signed char __b) { - return __builtin_altivec_vcmpgtub_p(__CR6_EQ, (vector unsigned char)__b, - (vector unsigned char)__a); + return __builtin_altivec_vcmpgtsb_p(__CR6_EQ, __b, (vector signed char)__a); } static __inline__ int __ATTRS_o_ai vec_all_ge(vector bool char __a, @@ -14390,8 +14940,7 @@ static __inline__ int __ATTRS_o_ai vec_all_ge(vector unsigned short __a, static __inline__ int __ATTRS_o_ai vec_all_ge(vector bool short __a, vector short __b) { - return __builtin_altivec_vcmpgtuh_p(__CR6_EQ, (vector unsigned short)__b, - (vector unsigned short)__a); + return __builtin_altivec_vcmpgtsh_p(__CR6_EQ, __b, (vector signed short)__a); } static __inline__ int __ATTRS_o_ai vec_all_ge(vector bool short __a, @@ -14427,8 +14976,7 @@ static __inline__ int __ATTRS_o_ai vec_all_ge(vector unsigned int __a, static __inline__ int __ATTRS_o_ai vec_all_ge(vector bool int __a, vector int __b) { - return __builtin_altivec_vcmpgtuw_p(__CR6_EQ, (vector unsigned int)__b, - (vector unsigned int)__a); + return __builtin_altivec_vcmpgtsw_p(__CR6_EQ, __b, (vector signed int)__a); } static __inline__ int __ATTRS_o_ai vec_all_ge(vector bool int __a, @@ -14442,7 +14990,7 @@ static __inline__ int __ATTRS_o_ai vec_all_ge(vector bool int __a, (vector unsigned int)__a); } -#ifdef __POWER8_VECTOR__ +#ifdef __VSX__ static __inline__ int __ATTRS_o_ai vec_all_ge(vector signed long long __a, vector signed long long __b) { return __builtin_altivec_vcmpgtsd_p(__CR6_EQ, __b, __a); @@ -14466,8 +15014,8 @@ static __inline__ int __ATTRS_o_ai vec_all_ge(vector unsigned long long __a, static __inline__ int __ATTRS_o_ai vec_all_ge(vector bool long long __a, vector signed long long __b) { - return __builtin_altivec_vcmpgtud_p(__CR6_EQ, (vector unsigned long long)__b, - (vector unsigned long long)__a); + return __builtin_altivec_vcmpgtsd_p(__CR6_EQ, __b, + (vector signed long long)__a); } static __inline__ int __ATTRS_o_ai vec_all_ge(vector bool long long __a, @@ -14499,7 +15047,7 @@ static __inline__ int __ATTRS_o_ai vec_all_ge(vector double __a, } #endif -#ifdef __POWER10_VECTOR__ +#if defined(__POWER10_VECTOR__) && defined(__SIZEOF_INT128__) static __inline__ int __ATTRS_o_ai vec_all_ge(vector signed __int128 __a, vector signed __int128 __b) { return __builtin_altivec_vcmpgtsq_p(__CR6_EQ, __b, __a); @@ -14535,8 +15083,7 @@ static __inline__ int __ATTRS_o_ai vec_all_gt(vector unsigned char __a, static __inline__ int __ATTRS_o_ai vec_all_gt(vector bool char __a, vector signed char __b) { - return __builtin_altivec_vcmpgtub_p(__CR6_LT, (vector unsigned char)__a, - (vector unsigned char)__b); + return __builtin_altivec_vcmpgtsb_p(__CR6_LT, (vector signed char)__a, __b); } static __inline__ int __ATTRS_o_ai vec_all_gt(vector bool char __a, @@ -14573,8 +15120,7 @@ static __inline__ int __ATTRS_o_ai vec_all_gt(vector unsigned short __a, static __inline__ int __ATTRS_o_ai vec_all_gt(vector bool short __a, vector short __b) { - return __builtin_altivec_vcmpgtuh_p(__CR6_LT, (vector unsigned short)__a, - (vector unsigned short)__b); + return __builtin_altivec_vcmpgtsh_p(__CR6_LT, (vector signed short)__a, __b); } static __inline__ int __ATTRS_o_ai vec_all_gt(vector bool short __a, @@ -14610,8 +15156,7 @@ static __inline__ int __ATTRS_o_ai vec_all_gt(vector unsigned int __a, static __inline__ int __ATTRS_o_ai vec_all_gt(vector bool int __a, vector int __b) { - return __builtin_altivec_vcmpgtuw_p(__CR6_LT, (vector unsigned int)__a, - (vector unsigned int)__b); + return __builtin_altivec_vcmpgtsw_p(__CR6_LT, (vector signed int)__a, __b); } static __inline__ int __ATTRS_o_ai vec_all_gt(vector bool int __a, @@ -14625,7 +15170,7 @@ static __inline__ int __ATTRS_o_ai vec_all_gt(vector bool int __a, (vector unsigned int)__b); } -#ifdef __POWER8_VECTOR__ +#ifdef __VSX__ static __inline__ int __ATTRS_o_ai vec_all_gt(vector signed long long __a, vector signed long long __b) { return __builtin_altivec_vcmpgtsd_p(__CR6_LT, __a, __b); @@ -14649,8 +15194,8 @@ static __inline__ int __ATTRS_o_ai vec_all_gt(vector unsigned long long __a, static __inline__ int __ATTRS_o_ai vec_all_gt(vector bool long long __a, vector signed long long __b) { - return __builtin_altivec_vcmpgtud_p(__CR6_LT, (vector unsigned long long)__a, - (vector unsigned long long)__b); + return __builtin_altivec_vcmpgtsd_p(__CR6_LT, (vector signed long long)__a, + __b); } static __inline__ int __ATTRS_o_ai vec_all_gt(vector bool long long __a, @@ -14682,7 +15227,7 @@ static __inline__ int __ATTRS_o_ai vec_all_gt(vector double __a, } #endif -#ifdef __POWER10_VECTOR__ +#if defined(__POWER10_VECTOR__) && defined(__SIZEOF_INT128__) static __inline__ int __ATTRS_o_ai vec_all_gt(vector signed __int128 __a, vector signed __int128 __b) { return __builtin_altivec_vcmpgtsq_p(__CR6_LT, __a, __b); @@ -14725,8 +15270,7 @@ static __inline__ int __ATTRS_o_ai vec_all_le(vector unsigned char __a, static __inline__ int __ATTRS_o_ai vec_all_le(vector bool char __a, vector signed char __b) { - return __builtin_altivec_vcmpgtub_p(__CR6_EQ, (vector unsigned char)__a, - (vector unsigned char)__b); + return __builtin_altivec_vcmpgtsb_p(__CR6_EQ, (vector signed char)__a, __b); } static __inline__ int __ATTRS_o_ai vec_all_le(vector bool char __a, @@ -14763,8 +15307,7 @@ static __inline__ int __ATTRS_o_ai vec_all_le(vector unsigned short __a, static __inline__ int __ATTRS_o_ai vec_all_le(vector bool short __a, vector short __b) { - return __builtin_altivec_vcmpgtuh_p(__CR6_EQ, (vector unsigned short)__a, - (vector unsigned short)__b); + return __builtin_altivec_vcmpgtsh_p(__CR6_EQ, (vector signed short)__a, __b); } static __inline__ int __ATTRS_o_ai vec_all_le(vector bool short __a, @@ -14800,8 +15343,7 @@ static __inline__ int __ATTRS_o_ai vec_all_le(vector unsigned int __a, static __inline__ int __ATTRS_o_ai vec_all_le(vector bool int __a, vector int __b) { - return __builtin_altivec_vcmpgtuw_p(__CR6_EQ, (vector unsigned int)__a, - (vector unsigned int)__b); + return __builtin_altivec_vcmpgtsw_p(__CR6_EQ, (vector signed int)__a, __b); } static __inline__ int __ATTRS_o_ai vec_all_le(vector bool int __a, @@ -14815,7 +15357,7 @@ static __inline__ int __ATTRS_o_ai vec_all_le(vector bool int __a, (vector unsigned int)__b); } -#ifdef __POWER8_VECTOR__ +#ifdef __VSX__ static __inline__ int __ATTRS_o_ai vec_all_le(vector signed long long __a, vector signed long long __b) { return __builtin_altivec_vcmpgtsd_p(__CR6_EQ, __a, __b); @@ -14840,8 +15382,8 @@ static __inline__ int __ATTRS_o_ai vec_all_le(vector unsigned long long __a, static __inline__ int __ATTRS_o_ai vec_all_le(vector bool long long __a, vector signed long long __b) { - return __builtin_altivec_vcmpgtud_p(__CR6_EQ, (vector unsigned long long)__a, - (vector unsigned long long)__b); + return __builtin_altivec_vcmpgtsd_p(__CR6_EQ, (vector signed long long)__a, + __b); } static __inline__ int __ATTRS_o_ai vec_all_le(vector bool long long __a, @@ -14873,7 +15415,7 @@ static __inline__ int __ATTRS_o_ai vec_all_le(vector double __a, } #endif -#ifdef __POWER10_VECTOR__ +#if defined(__POWER10_VECTOR__) && defined(__SIZEOF_INT128__) static __inline__ int __ATTRS_o_ai vec_all_le(vector signed __int128 __a, vector signed __int128 __b) { return __builtin_altivec_vcmpgtsq_p(__CR6_EQ, __a, __b); @@ -14909,8 +15451,7 @@ static __inline__ int __ATTRS_o_ai vec_all_lt(vector unsigned char __a, static __inline__ int __ATTRS_o_ai vec_all_lt(vector bool char __a, vector signed char __b) { - return __builtin_altivec_vcmpgtub_p(__CR6_LT, (vector unsigned char)__b, - (vector unsigned char)__a); + return __builtin_altivec_vcmpgtsb_p(__CR6_LT, __b, (vector signed char)__a); } static __inline__ int __ATTRS_o_ai vec_all_lt(vector bool char __a, @@ -14947,8 +15488,7 @@ static __inline__ int __ATTRS_o_ai vec_all_lt(vector unsigned short __a, static __inline__ int __ATTRS_o_ai vec_all_lt(vector bool short __a, vector short __b) { - return __builtin_altivec_vcmpgtuh_p(__CR6_LT, (vector unsigned short)__b, - (vector unsigned short)__a); + return __builtin_altivec_vcmpgtsh_p(__CR6_LT, __b, (vector signed short)__a); } static __inline__ int __ATTRS_o_ai vec_all_lt(vector bool short __a, @@ -14984,8 +15524,7 @@ static __inline__ int __ATTRS_o_ai vec_all_lt(vector unsigned int __a, static __inline__ int __ATTRS_o_ai vec_all_lt(vector bool int __a, vector int __b) { - return __builtin_altivec_vcmpgtuw_p(__CR6_LT, (vector unsigned int)__b, - (vector unsigned int)__a); + return __builtin_altivec_vcmpgtsw_p(__CR6_LT, __b, (vector signed int)__a); } static __inline__ int __ATTRS_o_ai vec_all_lt(vector bool int __a, @@ -14999,7 +15538,7 @@ static __inline__ int __ATTRS_o_ai vec_all_lt(vector bool int __a, (vector unsigned int)__a); } -#ifdef __POWER8_VECTOR__ +#ifdef __VSX__ static __inline__ int __ATTRS_o_ai vec_all_lt(vector signed long long __a, vector signed long long __b) { return __builtin_altivec_vcmpgtsd_p(__CR6_LT, __b, __a); @@ -15024,8 +15563,8 @@ static __inline__ int __ATTRS_o_ai vec_all_lt(vector unsigned long long __a, static __inline__ int __ATTRS_o_ai vec_all_lt(vector bool long long __a, vector signed long long __b) { - return __builtin_altivec_vcmpgtud_p(__CR6_LT, (vector unsigned long long)__b, - (vector unsigned long long)__a); + return __builtin_altivec_vcmpgtsd_p(__CR6_LT, __b, + (vector signed long long)__a); } static __inline__ int __ATTRS_o_ai vec_all_lt(vector bool long long __a, @@ -15057,7 +15596,7 @@ static __inline__ int __ATTRS_o_ai vec_all_lt(vector double __a, } #endif -#ifdef __POWER10_VECTOR__ +#if defined(__POWER10_VECTOR__) && defined(__SIZEOF_INT128__) static __inline__ int __ATTRS_o_ai vec_all_lt(vector signed __int128 __a, vector signed __int128 __b) { return __builtin_altivec_vcmpgtsq_p(__CR6_LT, __b, __a); @@ -15214,7 +15753,7 @@ static __inline__ int __ATTRS_o_ai vec_all_ne(vector bool int __a, (vector int)__b); } -#ifdef __POWER8_VECTOR__ +#ifdef __VSX__ static __inline__ int __ATTRS_o_ai vec_all_ne(vector signed long long __a, vector signed long long __b) { return __builtin_altivec_vcmpequd_p(__CR6_EQ, __a, __b); @@ -15273,7 +15812,7 @@ static __inline__ int __ATTRS_o_ai vec_all_ne(vector double __a, } #endif -#ifdef __POWER10_VECTOR__ +#if defined(__POWER10_VECTOR__) && defined(__SIZEOF_INT128__) static __inline__ int __ATTRS_o_ai vec_all_ne(vector signed __int128 __a, vector signed __int128 __b) { return __builtin_altivec_vcmpequq_p(__CR6_EQ, __a, __b); @@ -15323,25 +15862,57 @@ static __inline__ int __ATTRS_o_ai vec_all_ngt(vector double __a, /* vec_all_nle */ -static __inline__ int __attribute__((__always_inline__)) +static __inline__ int __ATTRS_o_ai vec_all_nle(vector float __a, vector float __b) { +#ifdef __VSX__ + return __builtin_vsx_xvcmpgesp_p(__CR6_EQ, __b, __a); +#else return __builtin_altivec_vcmpgefp_p(__CR6_EQ, __b, __a); +#endif +} + +#ifdef __VSX__ +static __inline__ int __ATTRS_o_ai vec_all_nle(vector double __a, + vector double __b) { + return __builtin_vsx_xvcmpgedp_p(__CR6_EQ, __b, __a); } +#endif /* vec_all_nlt */ -static __inline__ int __attribute__((__always_inline__)) +static __inline__ int __ATTRS_o_ai vec_all_nlt(vector float __a, vector float __b) { +#ifdef __VSX__ + return __builtin_vsx_xvcmpgtsp_p(__CR6_EQ, __b, __a); +#else return __builtin_altivec_vcmpgtfp_p(__CR6_EQ, __b, __a); +#endif } +#ifdef __VSX__ +static __inline__ int __ATTRS_o_ai vec_all_nlt(vector double __a, + vector double __b) { + return __builtin_vsx_xvcmpgtdp_p(__CR6_EQ, __b, __a); +} +#endif + /* vec_all_numeric */ -static __inline__ int __attribute__((__always_inline__)) +static __inline__ int __ATTRS_o_ai vec_all_numeric(vector float __a) { +#ifdef __VSX__ + return __builtin_vsx_xvcmpeqsp_p(__CR6_LT, __a, __a); +#else return __builtin_altivec_vcmpeqfp_p(__CR6_LT, __a, __a); +#endif } +#ifdef __VSX__ +static __inline__ int __ATTRS_o_ai vec_all_numeric(vector double __a) { + return __builtin_vsx_xvcmpeqdp_p(__CR6_LT, __a, __a); +} +#endif + /* vec_any_eq */ static __inline__ int __ATTRS_o_ai vec_any_eq(vector signed char __a, @@ -15471,7 +16042,7 @@ static __inline__ int __ATTRS_o_ai vec_any_eq(vector bool int __a, (vector int)__b); } -#ifdef __POWER8_VECTOR__ +#ifdef __VSX__ static __inline__ int __ATTRS_o_ai vec_any_eq(vector signed long long __a, vector signed long long __b) { return __builtin_altivec_vcmpequd_p(__CR6_EQ_REV, __a, __b); @@ -15530,7 +16101,7 @@ static __inline__ int __ATTRS_o_ai vec_any_eq(vector double __a, } #endif -#ifdef __POWER10_VECTOR__ +#if defined(__POWER10_VECTOR__) && defined(__SIZEOF_INT128__) static __inline__ int __ATTRS_o_ai vec_any_eq(vector signed __int128 __a, vector signed __int128 __b) { return __builtin_altivec_vcmpequq_p(__CR6_EQ_REV, __a, __b); @@ -15568,8 +16139,8 @@ static __inline__ int __ATTRS_o_ai vec_any_ge(vector unsigned char __a, static __inline__ int __ATTRS_o_ai vec_any_ge(vector bool char __a, vector signed char __b) { - return __builtin_altivec_vcmpgtub_p(__CR6_LT_REV, (vector unsigned char)__b, - (vector unsigned char)__a); + return __builtin_altivec_vcmpgtsb_p(__CR6_LT_REV, __b, + (vector signed char)__a); } static __inline__ int __ATTRS_o_ai vec_any_ge(vector bool char __a, @@ -15607,8 +16178,8 @@ static __inline__ int __ATTRS_o_ai vec_any_ge(vector unsigned short __a, static __inline__ int __ATTRS_o_ai vec_any_ge(vector bool short __a, vector short __b) { - return __builtin_altivec_vcmpgtuh_p(__CR6_LT_REV, (vector unsigned short)__b, - (vector unsigned short)__a); + return __builtin_altivec_vcmpgtsh_p(__CR6_LT_REV, __b, + (vector signed short)__a); } static __inline__ int __ATTRS_o_ai vec_any_ge(vector bool short __a, @@ -15645,8 +16216,8 @@ static __inline__ int __ATTRS_o_ai vec_any_ge(vector unsigned int __a, static __inline__ int __ATTRS_o_ai vec_any_ge(vector bool int __a, vector int __b) { - return __builtin_altivec_vcmpgtuw_p(__CR6_LT_REV, (vector unsigned int)__b, - (vector unsigned int)__a); + return __builtin_altivec_vcmpgtsw_p(__CR6_LT_REV, __b, + (vector signed int)__a); } static __inline__ int __ATTRS_o_ai vec_any_ge(vector bool int __a, @@ -15661,7 +16232,7 @@ static __inline__ int __ATTRS_o_ai vec_any_ge(vector bool int __a, (vector unsigned int)__a); } -#ifdef __POWER8_VECTOR__ +#ifdef __VSX__ static __inline__ int __ATTRS_o_ai vec_any_ge(vector signed long long __a, vector signed long long __b) { return __builtin_altivec_vcmpgtsd_p(__CR6_LT_REV, __b, __a); @@ -15686,9 +16257,8 @@ static __inline__ int __ATTRS_o_ai vec_any_ge(vector unsigned long long __a, static __inline__ int __ATTRS_o_ai vec_any_ge(vector bool long long __a, vector signed long long __b) { - return __builtin_altivec_vcmpgtud_p(__CR6_LT_REV, - (vector unsigned long long)__b, - (vector unsigned long long)__a); + return __builtin_altivec_vcmpgtsd_p(__CR6_LT_REV, __b, + (vector signed long long)__a); } static __inline__ int __ATTRS_o_ai vec_any_ge(vector bool long long __a, @@ -15721,7 +16291,7 @@ static __inline__ int __ATTRS_o_ai vec_any_ge(vector double __a, } #endif -#ifdef __POWER10_VECTOR__ +#if defined(__POWER10_VECTOR__) && defined(__SIZEOF_INT128__) static __inline__ int __ATTRS_o_ai vec_any_ge(vector signed __int128 __a, vector signed __int128 __b) { return __builtin_altivec_vcmpgtsq_p(__CR6_LT_REV, __b, __a); @@ -15759,8 +16329,8 @@ static __inline__ int __ATTRS_o_ai vec_any_gt(vector unsigned char __a, static __inline__ int __ATTRS_o_ai vec_any_gt(vector bool char __a, vector signed char __b) { - return __builtin_altivec_vcmpgtub_p(__CR6_EQ_REV, (vector unsigned char)__a, - (vector unsigned char)__b); + return __builtin_altivec_vcmpgtsb_p(__CR6_EQ_REV, (vector signed char)__a, + __b); } static __inline__ int __ATTRS_o_ai vec_any_gt(vector bool char __a, @@ -15798,8 +16368,8 @@ static __inline__ int __ATTRS_o_ai vec_any_gt(vector unsigned short __a, static __inline__ int __ATTRS_o_ai vec_any_gt(vector bool short __a, vector short __b) { - return __builtin_altivec_vcmpgtuh_p(__CR6_EQ_REV, (vector unsigned short)__a, - (vector unsigned short)__b); + return __builtin_altivec_vcmpgtsh_p(__CR6_EQ_REV, (vector signed short)__a, + __b); } static __inline__ int __ATTRS_o_ai vec_any_gt(vector bool short __a, @@ -15836,8 +16406,8 @@ static __inline__ int __ATTRS_o_ai vec_any_gt(vector unsigned int __a, static __inline__ int __ATTRS_o_ai vec_any_gt(vector bool int __a, vector int __b) { - return __builtin_altivec_vcmpgtuw_p(__CR6_EQ_REV, (vector unsigned int)__a, - (vector unsigned int)__b); + return __builtin_altivec_vcmpgtsw_p(__CR6_EQ_REV, (vector signed int)__a, + __b); } static __inline__ int __ATTRS_o_ai vec_any_gt(vector bool int __a, @@ -15852,7 +16422,7 @@ static __inline__ int __ATTRS_o_ai vec_any_gt(vector bool int __a, (vector unsigned int)__b); } -#ifdef __POWER8_VECTOR__ +#ifdef __VSX__ static __inline__ int __ATTRS_o_ai vec_any_gt(vector signed long long __a, vector signed long long __b) { return __builtin_altivec_vcmpgtsd_p(__CR6_EQ_REV, __a, __b); @@ -15877,9 +16447,8 @@ static __inline__ int __ATTRS_o_ai vec_any_gt(vector unsigned long long __a, static __inline__ int __ATTRS_o_ai vec_any_gt(vector bool long long __a, vector signed long long __b) { - return __builtin_altivec_vcmpgtud_p(__CR6_EQ_REV, - (vector unsigned long long)__a, - (vector unsigned long long)__b); + return __builtin_altivec_vcmpgtsd_p(__CR6_EQ_REV, + (vector signed long long)__a, __b); } static __inline__ int __ATTRS_o_ai vec_any_gt(vector bool long long __a, @@ -15912,7 +16481,7 @@ static __inline__ int __ATTRS_o_ai vec_any_gt(vector double __a, } #endif -#ifdef __POWER10_VECTOR__ +#if defined(__POWER10_VECTOR__) && defined(__SIZEOF_INT128__) static __inline__ int __ATTRS_o_ai vec_any_gt(vector signed __int128 __a, vector signed __int128 __b) { return __builtin_altivec_vcmpgtsq_p(__CR6_EQ_REV, __a, __b); @@ -15950,8 +16519,8 @@ static __inline__ int __ATTRS_o_ai vec_any_le(vector unsigned char __a, static __inline__ int __ATTRS_o_ai vec_any_le(vector bool char __a, vector signed char __b) { - return __builtin_altivec_vcmpgtub_p(__CR6_LT_REV, (vector unsigned char)__a, - (vector unsigned char)__b); + return __builtin_altivec_vcmpgtsb_p(__CR6_LT_REV, (vector signed char)__a, + __b); } static __inline__ int __ATTRS_o_ai vec_any_le(vector bool char __a, @@ -15989,8 +16558,8 @@ static __inline__ int __ATTRS_o_ai vec_any_le(vector unsigned short __a, static __inline__ int __ATTRS_o_ai vec_any_le(vector bool short __a, vector short __b) { - return __builtin_altivec_vcmpgtuh_p(__CR6_LT_REV, (vector unsigned short)__a, - (vector unsigned short)__b); + return __builtin_altivec_vcmpgtsh_p(__CR6_LT_REV, (vector signed short)__a, + __b); } static __inline__ int __ATTRS_o_ai vec_any_le(vector bool short __a, @@ -16027,8 +16596,8 @@ static __inline__ int __ATTRS_o_ai vec_any_le(vector unsigned int __a, static __inline__ int __ATTRS_o_ai vec_any_le(vector bool int __a, vector int __b) { - return __builtin_altivec_vcmpgtuw_p(__CR6_LT_REV, (vector unsigned int)__a, - (vector unsigned int)__b); + return __builtin_altivec_vcmpgtsw_p(__CR6_LT_REV, (vector signed int)__a, + __b); } static __inline__ int __ATTRS_o_ai vec_any_le(vector bool int __a, @@ -16043,7 +16612,7 @@ static __inline__ int __ATTRS_o_ai vec_any_le(vector bool int __a, (vector unsigned int)__b); } -#ifdef __POWER8_VECTOR__ +#ifdef __VSX__ static __inline__ int __ATTRS_o_ai vec_any_le(vector signed long long __a, vector signed long long __b) { return __builtin_altivec_vcmpgtsd_p(__CR6_LT_REV, __a, __b); @@ -16068,9 +16637,8 @@ static __inline__ int __ATTRS_o_ai vec_any_le(vector unsigned long long __a, static __inline__ int __ATTRS_o_ai vec_any_le(vector bool long long __a, vector signed long long __b) { - return __builtin_altivec_vcmpgtud_p(__CR6_LT_REV, - (vector unsigned long long)__a, - (vector unsigned long long)__b); + return __builtin_altivec_vcmpgtsd_p(__CR6_LT_REV, + (vector signed long long)__a, __b); } static __inline__ int __ATTRS_o_ai vec_any_le(vector bool long long __a, @@ -16103,7 +16671,7 @@ static __inline__ int __ATTRS_o_ai vec_any_le(vector double __a, } #endif -#ifdef __POWER10_VECTOR__ +#if defined(__POWER10_VECTOR__) && defined(__SIZEOF_INT128__) static __inline__ int __ATTRS_o_ai vec_any_le(vector signed __int128 __a, vector signed __int128 __b) { return __builtin_altivec_vcmpgtsq_p(__CR6_LT_REV, __a, __b); @@ -16141,8 +16709,8 @@ static __inline__ int __ATTRS_o_ai vec_any_lt(vector unsigned char __a, static __inline__ int __ATTRS_o_ai vec_any_lt(vector bool char __a, vector signed char __b) { - return __builtin_altivec_vcmpgtub_p(__CR6_EQ_REV, (vector unsigned char)__b, - (vector unsigned char)__a); + return __builtin_altivec_vcmpgtsb_p(__CR6_EQ_REV, __b, + (vector signed char)__a); } static __inline__ int __ATTRS_o_ai vec_any_lt(vector bool char __a, @@ -16180,8 +16748,8 @@ static __inline__ int __ATTRS_o_ai vec_any_lt(vector unsigned short __a, static __inline__ int __ATTRS_o_ai vec_any_lt(vector bool short __a, vector short __b) { - return __builtin_altivec_vcmpgtuh_p(__CR6_EQ_REV, (vector unsigned short)__b, - (vector unsigned short)__a); + return __builtin_altivec_vcmpgtsh_p(__CR6_EQ_REV, __b, + (vector signed short)__a); } static __inline__ int __ATTRS_o_ai vec_any_lt(vector bool short __a, @@ -16218,8 +16786,8 @@ static __inline__ int __ATTRS_o_ai vec_any_lt(vector unsigned int __a, static __inline__ int __ATTRS_o_ai vec_any_lt(vector bool int __a, vector int __b) { - return __builtin_altivec_vcmpgtuw_p(__CR6_EQ_REV, (vector unsigned int)__b, - (vector unsigned int)__a); + return __builtin_altivec_vcmpgtsw_p(__CR6_EQ_REV, __b, + (vector signed int)__a); } static __inline__ int __ATTRS_o_ai vec_any_lt(vector bool int __a, @@ -16234,7 +16802,7 @@ static __inline__ int __ATTRS_o_ai vec_any_lt(vector bool int __a, (vector unsigned int)__a); } -#ifdef __POWER8_VECTOR__ +#ifdef __VSX__ static __inline__ int __ATTRS_o_ai vec_any_lt(vector signed long long __a, vector signed long long __b) { return __builtin_altivec_vcmpgtsd_p(__CR6_EQ_REV, __b, __a); @@ -16259,9 +16827,8 @@ static __inline__ int __ATTRS_o_ai vec_any_lt(vector unsigned long long __a, static __inline__ int __ATTRS_o_ai vec_any_lt(vector bool long long __a, vector signed long long __b) { - return __builtin_altivec_vcmpgtud_p(__CR6_EQ_REV, - (vector unsigned long long)__b, - (vector unsigned long long)__a); + return __builtin_altivec_vcmpgtsd_p(__CR6_EQ_REV, __b, + (vector signed long long)__a); } static __inline__ int __ATTRS_o_ai vec_any_lt(vector bool long long __a, @@ -16294,7 +16861,7 @@ static __inline__ int __ATTRS_o_ai vec_any_lt(vector double __a, } #endif -#ifdef __POWER10_VECTOR__ +#if defined(__POWER10_VECTOR__) && defined(__SIZEOF_INT128__) static __inline__ int __ATTRS_o_ai vec_any_lt(vector signed __int128 __a, vector signed __int128 __b) { return __builtin_altivec_vcmpgtsq_p(__CR6_EQ_REV, __b, __a); @@ -16308,10 +16875,18 @@ static __inline__ int __ATTRS_o_ai vec_any_lt(vector unsigned __int128 __a, /* vec_any_nan */ -static __inline__ int __attribute__((__always_inline__)) -vec_any_nan(vector float __a) { +static __inline__ int __ATTRS_o_ai vec_any_nan(vector float __a) { +#ifdef __VSX__ + return __builtin_vsx_xvcmpeqsp_p(__CR6_LT_REV, __a, __a); +#else return __builtin_altivec_vcmpeqfp_p(__CR6_LT_REV, __a, __a); +#endif } +#ifdef __VSX__ +static __inline__ int __ATTRS_o_ai vec_any_nan(vector double __a) { + return __builtin_vsx_xvcmpeqdp_p(__CR6_LT_REV, __a, __a); +} +#endif /* vec_any_ne */ @@ -16442,7 +17017,7 @@ static __inline__ int __ATTRS_o_ai vec_any_ne(vector bool int __a, (vector int)__b); } -#ifdef __POWER8_VECTOR__ +#ifdef __VSX__ static __inline__ int __ATTRS_o_ai vec_any_ne(vector signed long long __a, vector signed long long __b) { return __builtin_altivec_vcmpequd_p(__CR6_LT_REV, __a, __b); @@ -16501,7 +17076,7 @@ static __inline__ int __ATTRS_o_ai vec_any_ne(vector double __a, } #endif -#ifdef __POWER10_VECTOR__ +#if defined(__POWER10_VECTOR__) && defined(__SIZEOF_INT128__) static __inline__ int __ATTRS_o_ai vec_any_ne(vector signed __int128 __a, vector signed __int128 __b) { return __builtin_altivec_vcmpequq_p(__CR6_LT_REV, __a, __b); @@ -16515,38 +17090,91 @@ static __inline__ int __ATTRS_o_ai vec_any_ne(vector unsigned __int128 __a, /* vec_any_nge */ -static __inline__ int __attribute__((__always_inline__)) -vec_any_nge(vector float __a, vector float __b) { +static __inline__ int __ATTRS_o_ai vec_any_nge(vector float __a, + vector float __b) { +#ifdef __VSX__ + return __builtin_vsx_xvcmpgesp_p(__CR6_LT_REV, __a, __b); +#else return __builtin_altivec_vcmpgefp_p(__CR6_LT_REV, __a, __b); +#endif } +#ifdef __VSX__ +static __inline__ int __ATTRS_o_ai vec_any_nge(vector double __a, + vector double __b) { + return __builtin_vsx_xvcmpgedp_p(__CR6_LT_REV, __a, __b); +} +#endif + /* vec_any_ngt */ -static __inline__ int __attribute__((__always_inline__)) -vec_any_ngt(vector float __a, vector float __b) { +static __inline__ int __ATTRS_o_ai vec_any_ngt(vector float __a, + vector float __b) { +#ifdef __VSX__ + return __builtin_vsx_xvcmpgtsp_p(__CR6_LT_REV, __a, __b); +#else return __builtin_altivec_vcmpgtfp_p(__CR6_LT_REV, __a, __b); +#endif } +#ifdef __VSX__ +static __inline__ int __ATTRS_o_ai vec_any_ngt(vector double __a, + vector double __b) { + return __builtin_vsx_xvcmpgtdp_p(__CR6_LT_REV, __a, __b); +} +#endif + /* vec_any_nle */ -static __inline__ int __attribute__((__always_inline__)) -vec_any_nle(vector float __a, vector float __b) { +static __inline__ int __ATTRS_o_ai vec_any_nle(vector float __a, + vector float __b) { +#ifdef __VSX__ + return __builtin_vsx_xvcmpgesp_p(__CR6_LT_REV, __b, __a); +#else return __builtin_altivec_vcmpgefp_p(__CR6_LT_REV, __b, __a); +#endif } +#ifdef __VSX__ +static __inline__ int __ATTRS_o_ai vec_any_nle(vector double __a, + vector double __b) { + return __builtin_vsx_xvcmpgedp_p(__CR6_LT_REV, __b, __a); +} +#endif + /* vec_any_nlt */ -static __inline__ int __attribute__((__always_inline__)) -vec_any_nlt(vector float __a, vector float __b) { +static __inline__ int __ATTRS_o_ai vec_any_nlt(vector float __a, + vector float __b) { +#ifdef __VSX__ + return __builtin_vsx_xvcmpgtsp_p(__CR6_LT_REV, __b, __a); +#else return __builtin_altivec_vcmpgtfp_p(__CR6_LT_REV, __b, __a); +#endif } +#ifdef __VSX__ +static __inline__ int __ATTRS_o_ai vec_any_nlt(vector double __a, + vector double __b) { + return __builtin_vsx_xvcmpgtdp_p(__CR6_LT_REV, __b, __a); +} +#endif + /* vec_any_numeric */ -static __inline__ int __attribute__((__always_inline__)) -vec_any_numeric(vector float __a) { +static __inline__ int __ATTRS_o_ai vec_any_numeric(vector float __a) { +#ifdef __VSX__ + return __builtin_vsx_xvcmpeqsp_p(__CR6_EQ_REV, __a, __a); +#else return __builtin_altivec_vcmpeqfp_p(__CR6_EQ_REV, __a, __a); +#endif +} + +#ifdef __VSX__ +static __inline__ int __ATTRS_o_ai vec_any_numeric(vector double __a) { + return __builtin_vsx_xvcmpeqdp_p(__CR6_EQ_REV, __a, __a); } +#endif /* vec_any_out */ @@ -16696,6 +17324,16 @@ vec_vgbbd(vector unsigned char __a) { return __builtin_altivec_vgbbd(__a); } +static __inline__ vector signed long long __ATTRS_o_ai +vec_gbb(vector signed long long __a) { + return __builtin_altivec_vgbbd((vector unsigned char)__a); +} + +static __inline__ vector unsigned long long __ATTRS_o_ai +vec_gbb(vector unsigned long long __a) { + return __builtin_altivec_vgbbd((vector unsigned char)__a); +} + static __inline__ vector long long __ATTRS_o_ai vec_vbpermq(vector signed char __a, vector signed char __b) { return __builtin_altivec_vbpermq((vector unsigned char)__a, @@ -16707,7 +17345,7 @@ vec_vbpermq(vector unsigned char __a, vector unsigned char __b) { return __builtin_altivec_vbpermq(__a, __b); } -#ifdef __powerpc64__ +#if defined(__powerpc64__) && defined(__SIZEOF_INT128__) static __inline__ vector unsigned long long __attribute__((__always_inline__)) vec_bperm(vector unsigned __int128 __a, vector unsigned char __b) { return __builtin_altivec_vbpermq((vector unsigned char)__a, @@ -16882,7 +17520,8 @@ vec_revb(vector double __a) { } #endif /* End __VSX__ */ -#if defined(__POWER8_VECTOR__) && defined(__powerpc64__) +#if defined(__POWER8_VECTOR__) && defined(__powerpc64__) && \ + defined(__SIZEOF_INT128__) static __inline__ vector signed __int128 __ATTRS_o_ai vec_revb(vector signed __int128 __a) { vector unsigned char __indices = @@ -16904,6 +17543,8 @@ vec_revb(vector unsigned __int128 __a) { /* vec_xl */ +#define vec_xld2 vec_xl +#define vec_xlw4 vec_xl typedef vector signed char unaligned_vec_schar __attribute__((aligned(1))); typedef vector unsigned char unaligned_vec_uchar __attribute__((aligned(1))); typedef vector signed short unaligned_vec_sshort __attribute__((aligned(1))); @@ -16912,41 +17553,41 @@ typedef vector signed int unaligned_vec_sint __attribute__((aligned(1))); typedef vector unsigned int unaligned_vec_uint __attribute__((aligned(1))); typedef vector float unaligned_vec_float __attribute__((aligned(1))); -static inline __ATTRS_o_ai vector signed char vec_xl(signed long long __offset, +static inline __ATTRS_o_ai vector signed char vec_xl(ptrdiff_t __offset, const signed char *__ptr) { return *(unaligned_vec_schar *)(__ptr + __offset); } static inline __ATTRS_o_ai vector unsigned char -vec_xl(signed long long __offset, const unsigned char *__ptr) { +vec_xl(ptrdiff_t __offset, const unsigned char *__ptr) { return *(unaligned_vec_uchar*)(__ptr + __offset); } -static inline __ATTRS_o_ai vector signed short vec_xl(signed long long __offset, - const signed short *__ptr) { +static inline __ATTRS_o_ai vector signed short +vec_xl(ptrdiff_t __offset, const signed short *__ptr) { signed char *__addr = (signed char *)__ptr + __offset; return *(unaligned_vec_sshort *)__addr; } static inline __ATTRS_o_ai vector unsigned short -vec_xl(signed long long __offset, const unsigned short *__ptr) { +vec_xl(ptrdiff_t __offset, const unsigned short *__ptr) { signed char *__addr = (signed char *)__ptr + __offset; return *(unaligned_vec_ushort *)__addr; } -static inline __ATTRS_o_ai vector signed int vec_xl(signed long long __offset, +static inline __ATTRS_o_ai vector signed int vec_xl(ptrdiff_t __offset, const signed int *__ptr) { signed char *__addr = (signed char *)__ptr + __offset; return *(unaligned_vec_sint *)__addr; } -static inline __ATTRS_o_ai vector unsigned int vec_xl(signed long long __offset, - const unsigned int *__ptr) { +static inline __ATTRS_o_ai vector unsigned int +vec_xl(ptrdiff_t __offset, const unsigned int *__ptr) { signed char *__addr = (signed char *)__ptr + __offset; return *(unaligned_vec_uint *)__addr; } -static inline __ATTRS_o_ai vector float vec_xl(signed long long __offset, +static inline __ATTRS_o_ai vector float vec_xl(ptrdiff_t __offset, const float *__ptr) { signed char *__addr = (signed char *)__ptr + __offset; return *(unaligned_vec_float *)__addr; @@ -16958,36 +17599,37 @@ typedef vector unsigned long long unaligned_vec_ull __attribute__((aligned(1))); typedef vector double unaligned_vec_double __attribute__((aligned(1))); static inline __ATTRS_o_ai vector signed long long -vec_xl(signed long long __offset, const signed long long *__ptr) { +vec_xl(ptrdiff_t __offset, const signed long long *__ptr) { signed char *__addr = (signed char *)__ptr + __offset; return *(unaligned_vec_sll *)__addr; } static inline __ATTRS_o_ai vector unsigned long long -vec_xl(signed long long __offset, const unsigned long long *__ptr) { +vec_xl(ptrdiff_t __offset, const unsigned long long *__ptr) { signed char *__addr = (signed char *)__ptr + __offset; return *(unaligned_vec_ull *)__addr; } -static inline __ATTRS_o_ai vector double vec_xl(signed long long __offset, +static inline __ATTRS_o_ai vector double vec_xl(ptrdiff_t __offset, const double *__ptr) { signed char *__addr = (signed char *)__ptr + __offset; return *(unaligned_vec_double *)__addr; } #endif -#if defined(__POWER8_VECTOR__) && defined(__powerpc64__) +#if defined(__POWER8_VECTOR__) && defined(__powerpc64__) && \ + defined(__SIZEOF_INT128__) typedef vector signed __int128 unaligned_vec_si128 __attribute__((aligned(1))); typedef vector unsigned __int128 unaligned_vec_ui128 __attribute__((aligned(1))); static inline __ATTRS_o_ai vector signed __int128 -vec_xl(signed long long __offset, const signed __int128 *__ptr) { +vec_xl(ptrdiff_t __offset, const signed __int128 *__ptr) { signed char *__addr = (signed char *)__ptr + __offset; return *(unaligned_vec_si128 *)__addr; } static inline __ATTRS_o_ai vector unsigned __int128 -vec_xl(signed long long __offset, const unsigned __int128 *__ptr) { +vec_xl(ptrdiff_t __offset, const unsigned __int128 *__ptr) { signed char *__addr = (signed char *)__ptr + __offset; return *(unaligned_vec_ui128 *)__addr; } @@ -16997,27 +17639,27 @@ vec_xl(signed long long __offset, const unsigned __int128 *__ptr) { #ifdef __LITTLE_ENDIAN__ static __inline__ vector signed char __ATTRS_o_ai -vec_xl_be(signed long long __offset, const signed char *__ptr) { +vec_xl_be(ptrdiff_t __offset, const signed char *__ptr) { vector signed char __vec = (vector signed char)__builtin_vsx_lxvd2x_be(__offset, __ptr); return __builtin_shufflevector(__vec, __vec, 7, 6, 5, 4, 3, 2, 1, 0, 15, 14, 13, 12, 11, 10, 9, 8); } static __inline__ vector unsigned char __ATTRS_o_ai -vec_xl_be(signed long long __offset, const unsigned char *__ptr) { +vec_xl_be(ptrdiff_t __offset, const unsigned char *__ptr) { vector unsigned char __vec = (vector unsigned char)__builtin_vsx_lxvd2x_be(__offset, __ptr); return __builtin_shufflevector(__vec, __vec, 7, 6, 5, 4, 3, 2, 1, 0, 15, 14, 13, 12, 11, 10, 9, 8); } -static __inline__ vector signed short __ATTRS_o_ai -vec_xl_be(signed long long __offset, const signed short *__ptr) { +static __inline__ vector signed short __ATTRS_o_ai +vec_xl_be(ptrdiff_t __offset, const signed short *__ptr) { vector signed short __vec = (vector signed short)__builtin_vsx_lxvd2x_be(__offset, __ptr); return __builtin_shufflevector(__vec, __vec, 3, 2, 1, 0, 7, 6, 5, 4); } static __inline__ vector unsigned short __ATTRS_o_ai -vec_xl_be(signed long long __offset, const unsigned short *__ptr) { +vec_xl_be(ptrdiff_t __offset, const unsigned short *__ptr) { vector unsigned short __vec = (vector unsigned short)__builtin_vsx_lxvd2x_be(__offset, __ptr); return __builtin_shufflevector(__vec, __vec, 3, 2, 1, 0, 7, 6, 5, 4); } @@ -17054,7 +17696,8 @@ vec_xl_be(signed long long __offset, const double *__ptr) { } #endif -#if defined(__POWER8_VECTOR__) && defined(__powerpc64__) +#if defined(__POWER8_VECTOR__) && defined(__powerpc64__) && \ + defined(__SIZEOF_INT128__) static __inline__ vector signed __int128 __ATTRS_o_ai vec_xl_be(signed long long __offset, const signed __int128 *__ptr) { return vec_xl(__offset, __ptr); @@ -17069,98 +17712,153 @@ vec_xl_be(signed long long __offset, const unsigned __int128 *__ptr) { #define vec_xl_be vec_xl #endif -#if defined(__POWER10_VECTOR__) && defined(__VSX__) +#if defined(__POWER10_VECTOR__) && defined(__VSX__) && \ + defined(__SIZEOF_INT128__) /* vect_xl_sext */ static __inline__ vector unsigned __int128 __ATTRS_o_ai -vec_xl_sext(signed long long __offset, const signed char *__pointer) { +vec_xl_sext(ptrdiff_t __offset, const signed char *__pointer) { return (vector unsigned __int128)*(__pointer + __offset); } static __inline__ vector unsigned __int128 __ATTRS_o_ai -vec_xl_sext(signed long long __offset, const signed short *__pointer) { +vec_xl_sext(ptrdiff_t __offset, const signed short *__pointer) { return (vector unsigned __int128)*(__pointer + __offset); } static __inline__ vector unsigned __int128 __ATTRS_o_ai -vec_xl_sext(signed long long __offset, const signed int *__pointer) { +vec_xl_sext(ptrdiff_t __offset, const signed int *__pointer) { return (vector unsigned __int128)*(__pointer + __offset); } static __inline__ vector unsigned __int128 __ATTRS_o_ai -vec_xl_sext(signed long long __offset, const signed long long *__pointer) { +vec_xl_sext(ptrdiff_t __offset, const signed long long *__pointer) { return (vector unsigned __int128)*(__pointer + __offset); } /* vec_xl_zext */ static __inline__ vector unsigned __int128 __ATTRS_o_ai -vec_xl_zext(signed long long __offset, const unsigned char *__pointer) { +vec_xl_zext(ptrdiff_t __offset, const unsigned char *__pointer) { return (vector unsigned __int128)*(__pointer + __offset); } static __inline__ vector unsigned __int128 __ATTRS_o_ai -vec_xl_zext(signed long long __offset, const unsigned short *__pointer) { +vec_xl_zext(ptrdiff_t __offset, const unsigned short *__pointer) { return (vector unsigned __int128)*(__pointer + __offset); } static __inline__ vector unsigned __int128 __ATTRS_o_ai -vec_xl_zext(signed long long __offset, const unsigned int *__pointer) { +vec_xl_zext(ptrdiff_t __offset, const unsigned int *__pointer) { return (vector unsigned __int128)*(__pointer + __offset); } static __inline__ vector unsigned __int128 __ATTRS_o_ai -vec_xl_zext(signed long long __offset, const unsigned long long *__pointer) { +vec_xl_zext(ptrdiff_t __offset, const unsigned long long *__pointer) { return (vector unsigned __int128)*(__pointer + __offset); } #endif +/* vec_xlds */ +#ifdef __VSX__ +static __inline__ vector signed long long __ATTRS_o_ai +vec_xlds(ptrdiff_t __offset, const signed long long *__ptr) { + signed long long *__addr = (signed long long*)((signed char *)__ptr + __offset); + return (vector signed long long) *__addr; +} + +static __inline__ vector unsigned long long __ATTRS_o_ai +vec_xlds(ptrdiff_t __offset, const unsigned long long *__ptr) { + unsigned long long *__addr = (unsigned long long *)((signed char *)__ptr + __offset); + return (unaligned_vec_ull) *__addr; +} + +static __inline__ vector double __ATTRS_o_ai vec_xlds(ptrdiff_t __offset, + const double *__ptr) { + double *__addr = (double*)((signed char *)__ptr + __offset); + return (unaligned_vec_double) *__addr; +} + +/* vec_load_splats */ +static __inline__ vector signed int __ATTRS_o_ai +vec_load_splats(signed long long __offset, const signed int *__ptr) { + signed int *__addr = (signed int*)((signed char *)__ptr + __offset); + return (vector signed int)*__addr; +} + +static __inline__ vector signed int __ATTRS_o_ai +vec_load_splats(unsigned long long __offset, const signed int *__ptr) { + signed int *__addr = (signed int*)((signed char *)__ptr + __offset); + return (vector signed int)*__addr; +} + +static __inline__ vector unsigned int __ATTRS_o_ai +vec_load_splats(signed long long __offset, const unsigned int *__ptr) { + unsigned int *__addr = (unsigned int*)((signed char *)__ptr + __offset); + return (vector unsigned int)*__addr; +} + +static __inline__ vector unsigned int __ATTRS_o_ai +vec_load_splats(unsigned long long __offset, const unsigned int *__ptr) { + unsigned int *__addr = (unsigned int*)((signed char *)__ptr + __offset); + return (vector unsigned int)*__addr; +} + +static __inline__ vector float __ATTRS_o_ai +vec_load_splats(signed long long __offset, const float *__ptr) { + float *__addr = (float*)((signed char *)__ptr + __offset); + return (vector float)*__addr; +} + +static __inline__ vector float __ATTRS_o_ai +vec_load_splats(unsigned long long __offset, const float *__ptr) { + float *__addr = (float*)((signed char *)__ptr + __offset); + return (vector float)*__addr; +} +#endif + /* vec_xst */ -static inline __ATTRS_o_ai void vec_xst(vector signed char __vec, - signed long long __offset, - signed char *__ptr) { +#define vec_xstd2 vec_xst +#define vec_xstw4 vec_xst +static inline __ATTRS_o_ai void +vec_xst(vector signed char __vec, ptrdiff_t __offset, signed char *__ptr) { *(unaligned_vec_schar *)(__ptr + __offset) = __vec; } -static inline __ATTRS_o_ai void vec_xst(vector unsigned char __vec, - signed long long __offset, - unsigned char *__ptr) { +static inline __ATTRS_o_ai void +vec_xst(vector unsigned char __vec, ptrdiff_t __offset, unsigned char *__ptr) { *(unaligned_vec_uchar *)(__ptr + __offset) = __vec; } -static inline __ATTRS_o_ai void vec_xst(vector signed short __vec, - signed long long __offset, - signed short *__ptr) { +static inline __ATTRS_o_ai void +vec_xst(vector signed short __vec, ptrdiff_t __offset, signed short *__ptr) { signed char *__addr = (signed char *)__ptr + __offset; *(unaligned_vec_sshort *)__addr = __vec; } static inline __ATTRS_o_ai void vec_xst(vector unsigned short __vec, - signed long long __offset, + ptrdiff_t __offset, unsigned short *__ptr) { signed char *__addr = (signed char *)__ptr + __offset; *(unaligned_vec_ushort *)__addr = __vec; } static inline __ATTRS_o_ai void vec_xst(vector signed int __vec, - signed long long __offset, - signed int *__ptr) { + ptrdiff_t __offset, signed int *__ptr) { signed char *__addr = (signed char *)__ptr + __offset; *(unaligned_vec_sint *)__addr = __vec; } -static inline __ATTRS_o_ai void vec_xst(vector unsigned int __vec, - signed long long __offset, - unsigned int *__ptr) { +static inline __ATTRS_o_ai void +vec_xst(vector unsigned int __vec, ptrdiff_t __offset, unsigned int *__ptr) { signed char *__addr = (signed char *)__ptr + __offset; *(unaligned_vec_uint *)__addr = __vec; } -static inline __ATTRS_o_ai void vec_xst(vector float __vec, - signed long long __offset, +static inline __ATTRS_o_ai void vec_xst(vector float __vec, ptrdiff_t __offset, float *__ptr) { signed char *__addr = (signed char *)__ptr + __offset; *(unaligned_vec_float *)__addr = __vec; @@ -17168,37 +17866,37 @@ static inline __ATTRS_o_ai void vec_xst(vector float __vec, #ifdef __VSX__ static inline __ATTRS_o_ai void vec_xst(vector signed long long __vec, - signed long long __offset, + ptrdiff_t __offset, signed long long *__ptr) { signed char *__addr = (signed char *)__ptr + __offset; *(unaligned_vec_sll *)__addr = __vec; } static inline __ATTRS_o_ai void vec_xst(vector unsigned long long __vec, - signed long long __offset, + ptrdiff_t __offset, unsigned long long *__ptr) { signed char *__addr = (signed char *)__ptr + __offset; *(unaligned_vec_ull *)__addr = __vec; } -static inline __ATTRS_o_ai void vec_xst(vector double __vec, - signed long long __offset, +static inline __ATTRS_o_ai void vec_xst(vector double __vec, ptrdiff_t __offset, double *__ptr) { signed char *__addr = (signed char *)__ptr + __offset; *(unaligned_vec_double *)__addr = __vec; } #endif -#if defined(__POWER8_VECTOR__) && defined(__powerpc64__) +#if defined(__POWER8_VECTOR__) && defined(__powerpc64__) && \ + defined(__SIZEOF_INT128__) static inline __ATTRS_o_ai void vec_xst(vector signed __int128 __vec, - signed long long __offset, + ptrdiff_t __offset, signed __int128 *__ptr) { signed char *__addr = (signed char *)__ptr + __offset; *(unaligned_vec_si128 *)__addr = __vec; } static inline __ATTRS_o_ai void vec_xst(vector unsigned __int128 __vec, - signed long long __offset, + ptrdiff_t __offset, unsigned __int128 *__ptr) { signed char *__addr = (signed char *)__ptr + __offset; *(unaligned_vec_ui128 *)__addr = __vec; @@ -17207,51 +17905,52 @@ static inline __ATTRS_o_ai void vec_xst(vector unsigned __int128 __vec, /* vec_xst_trunc */ -#if defined(__POWER10_VECTOR__) && defined(__VSX__) +#if defined(__POWER10_VECTOR__) && defined(__VSX__) && \ + defined(__SIZEOF_INT128__) static inline __ATTRS_o_ai void vec_xst_trunc(vector signed __int128 __vec, - signed long long __offset, + ptrdiff_t __offset, signed char *__ptr) { *(__ptr + __offset) = (signed char)__vec[0]; } static inline __ATTRS_o_ai void vec_xst_trunc(vector unsigned __int128 __vec, - signed long long __offset, + ptrdiff_t __offset, unsigned char *__ptr) { *(__ptr + __offset) = (unsigned char)__vec[0]; } static inline __ATTRS_o_ai void vec_xst_trunc(vector signed __int128 __vec, - signed long long __offset, + ptrdiff_t __offset, signed short *__ptr) { *(__ptr + __offset) = (signed short)__vec[0]; } static inline __ATTRS_o_ai void vec_xst_trunc(vector unsigned __int128 __vec, - signed long long __offset, + ptrdiff_t __offset, unsigned short *__ptr) { *(__ptr + __offset) = (unsigned short)__vec[0]; } static inline __ATTRS_o_ai void vec_xst_trunc(vector signed __int128 __vec, - signed long long __offset, + ptrdiff_t __offset, signed int *__ptr) { *(__ptr + __offset) = (signed int)__vec[0]; } static inline __ATTRS_o_ai void vec_xst_trunc(vector unsigned __int128 __vec, - signed long long __offset, + ptrdiff_t __offset, unsigned int *__ptr) { *(__ptr + __offset) = (unsigned int)__vec[0]; } static inline __ATTRS_o_ai void vec_xst_trunc(vector signed __int128 __vec, - signed long long __offset, + ptrdiff_t __offset, signed long long *__ptr) { *(__ptr + __offset) = (signed long long)__vec[0]; } static inline __ATTRS_o_ai void vec_xst_trunc(vector unsigned __int128 __vec, - signed long long __offset, + ptrdiff_t __offset, unsigned long long *__ptr) { *(__ptr + __offset) = (unsigned long long)__vec[0]; } @@ -17336,7 +18035,8 @@ static __inline__ void __ATTRS_o_ai vec_xst_be(vector double __vec, } #endif -#if defined(__POWER8_VECTOR__) && defined(__powerpc64__) +#if defined(__POWER8_VECTOR__) && defined(__powerpc64__) && \ + defined(__SIZEOF_INT128__) static __inline__ void __ATTRS_o_ai vec_xst_be(vector signed __int128 __vec, signed long long __offset, signed __int128 *__ptr) { @@ -17375,7 +18075,7 @@ static vector double __ATTRS_o_ai vec_neg(vector double __a) { #endif -#if defined(__POWER8_VECTOR__) && defined(__powerpc64__) +#ifdef __VSX__ static vector long long __ATTRS_o_ai vec_neg(vector long long __a) { return -__a; } @@ -17404,7 +18104,7 @@ static vector double __ATTRS_o_ai vec_nabs(vector double __a) { #endif -#if defined(__POWER8_VECTOR__) && defined(__powerpc64__) +#ifdef __POWER8_VECTOR__ static vector long long __ATTRS_o_ai vec_nabs(vector long long __a) { return __builtin_altivec_vminsd(__a, -__a); } @@ -17422,6 +18122,18 @@ static vector signed char __ATTRS_o_ai vec_nabs(vector signed char __a) { return __builtin_altivec_vminsb(__a, -__a); } +static vector float __ATTRS_o_ai vec_recipdiv(vector float __a, + vector float __b) { + return __builtin_ppc_recipdivf(__a, __b); +} + +#ifdef __VSX__ +static vector double __ATTRS_o_ai vec_recipdiv(vector double __a, + vector double __b) { + return __builtin_ppc_recipdivd(__a, __b); +} +#endif + #ifdef __POWER10_VECTOR__ /* vec_extractm */ @@ -17446,10 +18158,12 @@ vec_extractm(vector unsigned long long __a) { return __builtin_altivec_vextractdm(__a); } +#ifdef __SIZEOF_INT128__ static __inline__ unsigned int __ATTRS_o_ai vec_extractm(vector unsigned __int128 __a) { return __builtin_altivec_vextractqm(__a); } +#endif /* vec_expandm */ @@ -17473,10 +18187,12 @@ vec_expandm(vector unsigned long long __a) { return __builtin_altivec_vexpanddm(__a); } +#ifdef __SIZEOF_INT128__ static __inline__ vector unsigned __int128 __ATTRS_o_ai vec_expandm(vector unsigned __int128 __a) { return __builtin_altivec_vexpandqm(__a); } +#endif /* vec_cntm */ @@ -17512,10 +18228,12 @@ vec_gendm(unsigned long long __bm) { return __builtin_altivec_mtvsrdm(__bm); } +#ifdef __SIZEOF_INT128__ static __inline__ vector unsigned __int128 __ATTRS_o_ai vec_genqm(unsigned long long __bm) { return __builtin_altivec_mtvsrqm(__bm); } +#endif /* vec_pdep */ @@ -17544,6 +18262,7 @@ vec_cfuge(vector unsigned long long __a, vector unsigned long long __b) { /* vec_ternarylogic */ #ifdef __VSX__ +#ifdef __SIZEOF_INT128__ #define vec_ternarylogic(__a, __b, __c, __imm) \ _Generic((__a), vector unsigned char \ : __builtin_vsx_xxeval((vector unsigned long long)(__a), \ @@ -17565,6 +18284,25 @@ vec_cfuge(vector unsigned long long __a, vector unsigned long long __b) { : __builtin_vsx_xxeval((vector unsigned long long)(__a), \ (vector unsigned long long)(__b), \ (vector unsigned long long)(__c), (__imm))) +#else +#define vec_ternarylogic(__a, __b, __c, __imm) \ + _Generic((__a), vector unsigned char \ + : __builtin_vsx_xxeval((vector unsigned long long)(__a), \ + (vector unsigned long long)(__b), \ + (vector unsigned long long)(__c), (__imm)), \ + vector unsigned short \ + : __builtin_vsx_xxeval((vector unsigned long long)(__a), \ + (vector unsigned long long)(__b), \ + (vector unsigned long long)(__c), (__imm)), \ + vector unsigned int \ + : __builtin_vsx_xxeval((vector unsigned long long)(__a), \ + (vector unsigned long long)(__b), \ + (vector unsigned long long)(__c), (__imm)), \ + vector unsigned long long \ + : __builtin_vsx_xxeval((vector unsigned long long)(__a), \ + (vector unsigned long long)(__b), \ + (vector unsigned long long)(__c), (__imm))) +#endif /* __SIZEOF_INT128__ */ #endif /* __VSX__ */ /* vec_genpcvm */ @@ -17657,6 +18395,7 @@ vec_mod(vector unsigned long long __a, vector unsigned long long __b) { return __a % __b; } +#ifdef __SIZEOF_INT128__ static __inline__ vector signed __int128 __ATTRS_o_ai vec_mod(vector signed __int128 __a, vector signed __int128 __b) { return __a % __b; @@ -17666,6 +18405,7 @@ static __inline__ vector unsigned __int128 __ATTRS_o_ai vec_mod(vector unsigned __int128 __a, vector unsigned __int128 __b) { return __a % __b; } +#endif /* vec_sldbi */ @@ -18188,6 +18928,7 @@ static __inline__ int __ATTRS_o_ai vec_strir_p(vector signed short __a) { /* vs[l | r | ra] */ +#ifdef __SIZEOF_INT128__ static __inline__ vector unsigned __int128 __ATTRS_o_ai vec_sl(vector unsigned __int128 __a, vector unsigned __int128 __b) { return __a << (__b % (vector unsigned __int128)(sizeof(unsigned __int128) * @@ -18232,6 +18973,7 @@ vec_sra(vector signed __int128 __a, vector unsigned __int128 __b) { __CHAR_BIT__)); } +#endif /* __SIZEOF_INT128__ */ #endif /* __POWER10_VECTOR__ */ #undef __ATTRS_o_ai diff --git a/lib/include/amxintrin.h b/lib/include/amxintrin.h index 823c7ca1f..ec601a58e 100644 --- a/lib/include/amxintrin.h +++ b/lib/include/amxintrin.h @@ -15,8 +15,13 @@ #define __AMXINTRIN_H #ifdef __x86_64__ +/* Define the default attributes for the functions in this file. */ #define __DEFAULT_FN_ATTRS_TILE \ __attribute__((__always_inline__, __nodebug__, __target__("amx-tile"))) +#define __DEFAULT_FN_ATTRS_INT8 \ + __attribute__((__always_inline__, __nodebug__, __target__("amx-int8"))) +#define __DEFAULT_FN_ATTRS_BF16 \ + __attribute__((__always_inline__, __nodebug__, __target__("amx-bf16"))) /// Load tile configuration from a 64-byte memory location specified by /// "mem_addr". The tile configuration includes the tile type palette, the @@ -25,7 +30,7 @@ /// config and the tile data, and the tiles are zeroed. Any invalid /// configurations will result in #GP fault. /// -/// \headerfile +/// \headerfile /// /// This intrinsic corresponds to the LDTILECFG instruction. /// @@ -41,7 +46,7 @@ _tile_loadconfig(const void *__config) { /// palette, the number of bytes per row, and the number of rows. If tiles /// are not configured, all zeroes will be stored to memory. /// -/// \headerfile +/// \headerfile /// /// This intrinsic corresponds to the STTILECFG instruction. /// @@ -55,7 +60,7 @@ _tile_storeconfig(void *__config) { /// Release the tile configuration to return to the init state, which /// releases all storage it currently holds. /// -/// \headerfile +/// \headerfile /// /// This intrinsic corresponds to the TILERELEASE instruction. static __inline__ void __DEFAULT_FN_ATTRS_TILE _tile_release(void) { @@ -66,7 +71,7 @@ static __inline__ void __DEFAULT_FN_ATTRS_TILE _tile_release(void) { /// destination tile "dst" using the tile configuration previously configured /// via "_tile_loadconfig". /// -/// \headerfile +/// \headerfile /// /// This intrinsic corresponds to the TILELOADD instruction. /// @@ -86,7 +91,7 @@ static __inline__ void __DEFAULT_FN_ATTRS_TILE _tile_release(void) { /// that the data will likely not be reused in the near future and the data /// caching can be optimized accordingly. /// -/// \headerfile +/// \headerfile /// /// This intrinsic corresponds to the TILELOADDT1 instruction. /// @@ -104,7 +109,7 @@ static __inline__ void __DEFAULT_FN_ATTRS_TILE _tile_release(void) { /// "stride" using the tile configuration previously configured via /// "_tile_loadconfig". /// -/// \headerfile +/// \headerfile /// /// This intrinsic corresponds to the TILESTORED instruction. /// @@ -119,7 +124,7 @@ static __inline__ void __DEFAULT_FN_ATTRS_TILE _tile_release(void) { /// Zero the tile specified by "tdest". /// -/// \headerfile +/// \headerfile /// /// This intrinsic corresponds to the TILEZERO instruction. /// @@ -133,7 +138,7 @@ static __inline__ void __DEFAULT_FN_ATTRS_TILE _tile_release(void) { /// results. Sum these 4 results with the corresponding 32-bit integer in "dst", /// and store the 32-bit result back to tile "dst". /// -/// \headerfile +/// \headerfile /// /// This intrinsic corresponds to the TDPBSSD instruction. /// @@ -152,7 +157,7 @@ static __inline__ void __DEFAULT_FN_ATTRS_TILE _tile_release(void) { /// 32-bit results. Sum these 4 results with the corresponding 32-bit integer /// in "dst", and store the 32-bit result back to tile "dst". /// -/// \headerfile +/// \headerfile /// /// This intrinsic corresponds to the TDPBSUD instruction. /// @@ -171,7 +176,7 @@ static __inline__ void __DEFAULT_FN_ATTRS_TILE _tile_release(void) { /// results. Sum these 4 results with the corresponding 32-bit integer in "dst", /// and store the 32-bit result back to tile "dst". /// -/// \headerfile +/// \headerfile /// /// This intrinsic corresponds to the TDPBUSD instruction. /// @@ -190,7 +195,7 @@ static __inline__ void __DEFAULT_FN_ATTRS_TILE _tile_release(void) { /// 32-bit results. Sum these 4 results with the corresponding 32-bit integer in /// "dst", and store the 32-bit result back to tile "dst". /// -/// \headerfile +/// \headerfile /// /// This intrinsic corresponds to the TDPBUUD instruction. /// @@ -208,7 +213,7 @@ static __inline__ void __DEFAULT_FN_ATTRS_TILE _tile_release(void) { /// elements with elements in "dst", and store the 32-bit result back to tile /// "dst". /// -/// \headerfile +/// \headerfile /// /// This intrinsic corresponds to the TDPBF16PS instruction. /// @@ -221,10 +226,12 @@ static __inline__ void __DEFAULT_FN_ATTRS_TILE _tile_release(void) { #define _tile_dpbf16ps(dst, src0, src1) \ __builtin_ia32_tdpbf16ps((dst), (src0), (src1)) -#define __DEFAULT_FN_ATTRS_INT8 \ - __attribute__((__always_inline__, __nodebug__, __target__("amx-int8"))) - +/// AMX tile register size can be configured, the maximum size is 16x64=1024 +/// bytes. Since there is no 2D type in llvm IR, we use vector type to +/// represent 2D tile and the fixed size is maximum amx tile register size. typedef int _tile1024i __attribute__((__vector_size__(1024), __aligned__(64))); + +/// This is internal intrinsic. C/C++ user should avoid calling it directly. static __inline__ _tile1024i __DEFAULT_FN_ATTRS_INT8 _tile_loadd_internal(unsigned short m, unsigned short n, const void *base, __SIZE_TYPE__ stride) { @@ -232,12 +239,43 @@ _tile_loadd_internal(unsigned short m, unsigned short n, const void *base, (__SIZE_TYPE__)(stride)); } +/// This is internal intrinsic. C/C++ user should avoid calling it directly. +static __inline__ _tile1024i __DEFAULT_FN_ATTRS_INT8 +_tile_loaddt1_internal(unsigned short m, unsigned short n, const void *base, + __SIZE_TYPE__ stride) { + return __builtin_ia32_tileloaddt164_internal(m, n, base, + (__SIZE_TYPE__)(stride)); +} + +/// This is internal intrinsic. C/C++ user should avoid calling it directly. static __inline__ _tile1024i __DEFAULT_FN_ATTRS_INT8 _tile_dpbssd_internal(unsigned short m, unsigned short n, unsigned short k, _tile1024i dst, _tile1024i src1, _tile1024i src2) { return __builtin_ia32_tdpbssd_internal(m, n, k, dst, src1, src2); } +/// This is internal intrinsic. C/C++ user should avoid calling it directly. +static __inline__ _tile1024i __DEFAULT_FN_ATTRS_INT8 +_tile_dpbsud_internal(unsigned short m, unsigned short n, unsigned short k, + _tile1024i dst, _tile1024i src1, _tile1024i src2) { + return __builtin_ia32_tdpbsud_internal(m, n, k, dst, src1, src2); +} + +/// This is internal intrinsic. C/C++ user should avoid calling it directly. +static __inline__ _tile1024i __DEFAULT_FN_ATTRS_INT8 +_tile_dpbusd_internal(unsigned short m, unsigned short n, unsigned short k, + _tile1024i dst, _tile1024i src1, _tile1024i src2) { + return __builtin_ia32_tdpbusd_internal(m, n, k, dst, src1, src2); +} + +/// This is internal intrinsic. C/C++ user should avoid calling it directly. +static __inline__ _tile1024i __DEFAULT_FN_ATTRS_INT8 +_tile_dpbuud_internal(unsigned short m, unsigned short n, unsigned short k, + _tile1024i dst, _tile1024i src1, _tile1024i src2) { + return __builtin_ia32_tdpbuud_internal(m, n, k, dst, src1, src2); +} + +/// This is internal intrinsic. C/C++ user should avoid calling it directly. static __inline__ void __DEFAULT_FN_ATTRS_INT8 _tile_stored_internal(unsigned short m, unsigned short n, void *base, __SIZE_TYPE__ stride, _tile1024i tile) { @@ -245,34 +283,211 @@ _tile_stored_internal(unsigned short m, unsigned short n, void *base, (__SIZE_TYPE__)(stride), tile); } +/// This is internal intrinsic. C/C++ user should avoid calling it directly. +static __inline__ _tile1024i __DEFAULT_FN_ATTRS_BF16 +_tile_dpbf16ps_internal(unsigned short m, unsigned short n, unsigned short k, + _tile1024i dst, _tile1024i src1, _tile1024i src2) { + return __builtin_ia32_tdpbf16ps_internal(m, n, k, dst, src1, src2); +} + +/// This struct pack the shape and tile data together for user. We suggest +/// initializing the struct as early as possible, because compiler depends +/// on the shape information to do configure. The constant value is preferred +/// for optimization by compiler. typedef struct __tile1024i_str { const unsigned short row; const unsigned short col; _tile1024i tile; } __tile1024i; +/// Load tile rows from memory specifieid by "base" address and "stride" into +/// destination tile "dst". +/// +/// \headerfile +/// +/// This intrinsic corresponds to the TILELOADD instruction. +/// +/// \param dst +/// A destination tile. Max size is 1024 Bytes. +/// \param base +/// A pointer to base address. +/// \param stride +/// The stride between the rows' data to be loaded in memory. __DEFAULT_FN_ATTRS_TILE static void __tile_loadd(__tile1024i *dst, const void *base, __SIZE_TYPE__ stride) { dst->tile = _tile_loadd_internal(dst->row, dst->col, base, stride); } +/// Load tile rows from memory specifieid by "base" address and "stride" into +/// destination tile "dst". This intrinsic provides a hint to the implementation +/// that the data will likely not be reused in the near future and the data +/// caching can be optimized accordingly. +/// +/// \headerfile +/// +/// This intrinsic corresponds to the TILELOADDT1 instruction. +/// +/// \param dst +/// A destination tile. Max size is 1024 Bytes. +/// \param base +/// A pointer to base address. +/// \param stride +/// The stride between the rows' data to be loaded in memory. +__DEFAULT_FN_ATTRS_TILE +static void __tile_stream_loadd(__tile1024i *dst, const void *base, + __SIZE_TYPE__ stride) { + dst->tile = _tile_loaddt1_internal(dst->row, dst->col, base, stride); +} + +/// Compute dot-product of bytes in tiles with a source/destination accumulator. +/// Multiply groups of 4 adjacent pairs of signed 8-bit integers in src0 with +/// corresponding signed 8-bit integers in src1, producing 4 intermediate 32-bit +/// results. Sum these 4 results with the corresponding 32-bit integer in "dst", +/// and store the 32-bit result back to tile "dst". +/// +/// \headerfile +/// +/// This intrinsic corresponds to the TDPBSSD instruction. +/// +/// \param dst +/// The destination tile. Max size is 1024 Bytes. +/// \param src0 +/// The 1st source tile. Max size is 1024 Bytes. +/// \param src1 +/// The 2nd source tile. Max size is 1024 Bytes. __DEFAULT_FN_ATTRS_INT8 -static void __tile_dpbssd(__tile1024i *dst, __tile1024i src1, - __tile1024i src2) { - dst->tile = _tile_dpbssd_internal(src1.row, src2.col, src1.col, dst->tile, - src1.tile, src2.tile); +static void __tile_dpbssd(__tile1024i *dst, __tile1024i src0, + __tile1024i src1) { + dst->tile = _tile_dpbssd_internal(src0.row, src1.col, src0.col, dst->tile, + src0.tile, src1.tile); } +/// Compute dot-product of bytes in tiles with a source/destination accumulator. +/// Multiply groups of 4 adjacent pairs of signed 8-bit integers in src0 with +/// corresponding unsigned 8-bit integers in src1, producing 4 intermediate +/// 32-bit results. Sum these 4 results with the corresponding 32-bit integer +/// in "dst", and store the 32-bit result back to tile "dst". +/// +/// \headerfile +/// +/// This intrinsic corresponds to the TDPBSUD instruction. +/// +/// \param dst +/// The destination tile. Max size is 1024 Bytes. +/// \param src0 +/// The 1st source tile. Max size is 1024 Bytes. +/// \param src1 +/// The 2nd source tile. Max size is 1024 Bytes. +__DEFAULT_FN_ATTRS_INT8 +static void __tile_dpbsud(__tile1024i *dst, __tile1024i src0, + __tile1024i src1) { + dst->tile = _tile_dpbsud_internal(src0.row, src1.col, src0.col, dst->tile, + src0.tile, src1.tile); +} + +/// Compute dot-product of bytes in tiles with a source/destination accumulator. +/// Multiply groups of 4 adjacent pairs of unsigned 8-bit integers in src0 with +/// corresponding signed 8-bit integers in src1, producing 4 intermediate 32-bit +/// results. Sum these 4 results with the corresponding 32-bit integer in "dst", +/// and store the 32-bit result back to tile "dst". +/// +/// \headerfile +/// +/// This intrinsic corresponds to the TDPBUSD instruction. +/// +/// \param dst +/// The destination tile. Max size is 1024 Bytes. +/// \param src0 +/// The 1st source tile. Max size is 1024 Bytes. +/// \param src1 +/// The 2nd source tile. Max size is 1024 Bytes. +__DEFAULT_FN_ATTRS_INT8 +static void __tile_dpbusd(__tile1024i *dst, __tile1024i src0, + __tile1024i src1) { + dst->tile = _tile_dpbusd_internal(src0.row, src1.col, src0.col, dst->tile, + src0.tile, src1.tile); +} + +/// Compute dot-product of bytes in tiles with a source/destination accumulator. +/// Multiply groups of 4 adjacent pairs of unsigned 8-bit integers in src0 with +/// corresponding unsigned 8-bit integers in src1, producing 4 intermediate +/// 32-bit results. Sum these 4 results with the corresponding 32-bit integer in +/// "dst", and store the 32-bit result back to tile "dst". +/// +/// \headerfile +/// +/// This intrinsic corresponds to the TDPBUUD instruction. +/// +/// \param dst +/// The destination tile. Max size is 1024 Bytes. +/// \param src0 +/// The 1st source tile. Max size is 1024 Bytes. +/// \param src1 +/// The 2nd source tile. Max size is 1024 Bytes. +__DEFAULT_FN_ATTRS_INT8 +static void __tile_dpbuud(__tile1024i *dst, __tile1024i src0, + __tile1024i src1) { + dst->tile = _tile_dpbuud_internal(src0.row, src1.col, src0.col, dst->tile, + src0.tile, src1.tile); +} + +/// Store the tile specified by "src" to memory specifieid by "base" address and +/// "stride". +/// +/// \headerfile +/// +/// This intrinsic corresponds to the TILESTORED instruction. +/// +/// \param dst +/// A destination tile. Max size is 1024 Bytes. +/// \param base +/// A pointer to base address. +/// \param stride +/// The stride between the rows' data to be stored in memory. __DEFAULT_FN_ATTRS_TILE static void __tile_stored(void *base, __SIZE_TYPE__ stride, __tile1024i src) { _tile_stored_internal(src.row, src.col, base, stride, src.tile); } +/// Zero the tile specified by "dst". +/// +/// \headerfile +/// +/// This intrinsic corresponds to the TILEZERO instruction. +/// +/// \param dst +/// The destination tile to be zero. Max size is 1024 Bytes. __DEFAULT_FN_ATTRS_TILE static void __tile_zero(__tile1024i *dst) { dst->tile = __builtin_ia32_tilezero_internal(dst->row, dst->col); } +/// Compute dot-product of BF16 (16-bit) floating-point pairs in tiles src0 and +/// src1, accumulating the intermediate single-precision (32-bit) floating-point +/// elements with elements in "dst", and store the 32-bit result back to tile +/// "dst". +/// +/// \headerfile +/// +/// This intrinsic corresponds to the TDPBF16PS instruction. +/// +/// \param dst +/// The destination tile. Max size is 1024 Bytes. +/// \param src0 +/// The 1st source tile. Max size is 1024 Bytes. +/// \param src1 +/// The 2nd source tile. Max size is 1024 Bytes. +__DEFAULT_FN_ATTRS_BF16 +static void __tile_dpbf16ps(__tile1024i *dst, __tile1024i src0, + __tile1024i src1) { + dst->tile = _tile_dpbf16ps_internal(src0.row, src1.col, src0.col, dst->tile, + src0.tile, src1.tile); +} + +#undef __DEFAULT_FN_ATTRS_TILE +#undef __DEFAULT_FN_ATTRS_INT8 +#undef __DEFAULT_FN_ATTRS_BF16 + #endif /* __x86_64__ */ #endif /* __AMXINTRIN_H */ diff --git a/lib/include/arm_acle.h b/lib/include/arm_acle.h index c156d89c1..45fac248d 100644 --- a/lib/include/arm_acle.h +++ b/lib/include/arm_acle.h @@ -639,6 +639,49 @@ __jcvt(double __a) { } #endif +/* Armv8.5-A FP rounding intrinsics */ +#if __ARM_64BIT_STATE && defined(__ARM_FEATURE_FRINT) +static __inline__ float __attribute__((__always_inline__, __nodebug__)) +__frint32zf(float __a) { + return __builtin_arm_frint32zf(__a); +} + +static __inline__ double __attribute__((__always_inline__, __nodebug__)) +__frint32z(double __a) { + return __builtin_arm_frint32z(__a); +} + +static __inline__ float __attribute__((__always_inline__, __nodebug__)) +__frint64zf(float __a) { + return __builtin_arm_frint64zf(__a); +} + +static __inline__ double __attribute__((__always_inline__, __nodebug__)) +__frint64z(double __a) { + return __builtin_arm_frint64z(__a); +} + +static __inline__ float __attribute__((__always_inline__, __nodebug__)) +__frint32xf(float __a) { + return __builtin_arm_frint32xf(__a); +} + +static __inline__ double __attribute__((__always_inline__, __nodebug__)) +__frint32x(double __a) { + return __builtin_arm_frint32x(__a); +} + +static __inline__ float __attribute__((__always_inline__, __nodebug__)) +__frint64xf(float __a) { + return __builtin_arm_frint64xf(__a); +} + +static __inline__ double __attribute__((__always_inline__, __nodebug__)) +__frint64x(double __a) { + return __builtin_arm_frint64x(__a); +} +#endif + /* Armv8.7-A load/store 64-byte intrinsics */ #if __ARM_64BIT_STATE && defined(__ARM_FEATURE_LS64) typedef struct { @@ -709,6 +752,18 @@ __arm_st64bv0(void *__addr, data512_t __value) { #endif /* __ARM_FEATURE_TME */ +/* Armv8.5-A Random number generation intrinsics */ +#if __ARM_64BIT_STATE && defined(__ARM_FEATURE_RNG) +static __inline__ int __attribute__((__always_inline__, __nodebug__)) +__rndr(uint64_t *__p) { + return __builtin_arm_rndr(__p); +} +static __inline__ int __attribute__((__always_inline__, __nodebug__)) +__rndrrs(uint64_t *__p) { + return __builtin_arm_rndrrs(__p); +} +#endif + #if defined(__cplusplus) } #endif diff --git a/lib/include/arm_neon.h b/lib/include/arm_neon.h index dbb65b8a4..2448870ce 100644 --- a/lib/include/arm_neon.h +++ b/lib/include/arm_neon.h @@ -2432,6 +2432,96 @@ __ai int16x4_t vadd_s16(int16x4_t __p0, int16x4_t __p1) { } #endif +#ifdef __LITTLE_ENDIAN__ +__ai poly8x8_t vadd_p8(poly8x8_t __p0, poly8x8_t __p1) { + poly8x8_t __ret; + __ret = (poly8x8_t) __builtin_neon_vadd_v((int8x8_t)__p0, (int8x8_t)__p1, 4); + return __ret; +} +#else +__ai poly8x8_t vadd_p8(poly8x8_t __p0, poly8x8_t __p1) { + poly8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0); + poly8x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0); + poly8x8_t __ret; + __ret = (poly8x8_t) __builtin_neon_vadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 4); + __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); + return __ret; +} +#endif + +__ai poly64x1_t vadd_p64(poly64x1_t __p0, poly64x1_t __p1) { + poly64x1_t __ret; + __ret = (poly64x1_t) __builtin_neon_vadd_v((int8x8_t)__p0, (int8x8_t)__p1, 6); + return __ret; +} +#ifdef __LITTLE_ENDIAN__ +__ai poly16x4_t vadd_p16(poly16x4_t __p0, poly16x4_t __p1) { + poly16x4_t __ret; + __ret = (poly16x4_t) __builtin_neon_vadd_v((int8x8_t)__p0, (int8x8_t)__p1, 5); + return __ret; +} +#else +__ai poly16x4_t vadd_p16(poly16x4_t __p0, poly16x4_t __p1) { + poly16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0); + poly16x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0); + poly16x4_t __ret; + __ret = (poly16x4_t) __builtin_neon_vadd_v((int8x8_t)__rev0, (int8x8_t)__rev1, 5); + __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); + return __ret; +} +#endif + +#ifdef __LITTLE_ENDIAN__ +__ai poly8x16_t vaddq_p8(poly8x16_t __p0, poly8x16_t __p1) { + poly8x16_t __ret; + __ret = (poly8x16_t) __builtin_neon_vaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 36); + return __ret; +} +#else +__ai poly8x16_t vaddq_p8(poly8x16_t __p0, poly8x16_t __p1) { + poly8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); + poly8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); + poly8x16_t __ret; + __ret = (poly8x16_t) __builtin_neon_vaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 36); + __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); + return __ret; +} +#endif + +#ifdef __LITTLE_ENDIAN__ +__ai poly64x2_t vaddq_p64(poly64x2_t __p0, poly64x2_t __p1) { + poly64x2_t __ret; + __ret = (poly64x2_t) __builtin_neon_vaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 38); + return __ret; +} +#else +__ai poly64x2_t vaddq_p64(poly64x2_t __p0, poly64x2_t __p1) { + poly64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0); + poly64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0); + poly64x2_t __ret; + __ret = (poly64x2_t) __builtin_neon_vaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 38); + __ret = __builtin_shufflevector(__ret, __ret, 1, 0); + return __ret; +} +#endif + +#ifdef __LITTLE_ENDIAN__ +__ai poly16x8_t vaddq_p16(poly16x8_t __p0, poly16x8_t __p1) { + poly16x8_t __ret; + __ret = (poly16x8_t) __builtin_neon_vaddq_v((int8x16_t)__p0, (int8x16_t)__p1, 37); + return __ret; +} +#else +__ai poly16x8_t vaddq_p16(poly16x8_t __p0, poly16x8_t __p1) { + poly16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0); + poly16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0); + poly16x8_t __ret; + __ret = (poly16x8_t) __builtin_neon_vaddq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 37); + __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); + return __ret; +} +#endif + #ifdef __LITTLE_ENDIAN__ __ai uint16x4_t vaddhn_u32(uint32x4_t __p0, uint32x4_t __p1) { uint16x4_t __ret; @@ -4552,6 +4642,54 @@ __ai uint16x4_t vcle_s16(int16x4_t __p0, int16x4_t __p1) { } #endif +#ifdef __LITTLE_ENDIAN__ +__ai int8x16_t vclsq_u8(uint8x16_t __p0) { + int8x16_t __ret; + __ret = (int8x16_t) __builtin_neon_vclsq_v((int8x16_t)__p0, 32); + return __ret; +} +#else +__ai int8x16_t vclsq_u8(uint8x16_t __p0) { + uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); + int8x16_t __ret; + __ret = (int8x16_t) __builtin_neon_vclsq_v((int8x16_t)__rev0, 32); + __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); + return __ret; +} +#endif + +#ifdef __LITTLE_ENDIAN__ +__ai int32x4_t vclsq_u32(uint32x4_t __p0) { + int32x4_t __ret; + __ret = (int32x4_t) __builtin_neon_vclsq_v((int8x16_t)__p0, 34); + return __ret; +} +#else +__ai int32x4_t vclsq_u32(uint32x4_t __p0) { + uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0); + int32x4_t __ret; + __ret = (int32x4_t) __builtin_neon_vclsq_v((int8x16_t)__rev0, 34); + __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); + return __ret; +} +#endif + +#ifdef __LITTLE_ENDIAN__ +__ai int16x8_t vclsq_u16(uint16x8_t __p0) { + int16x8_t __ret; + __ret = (int16x8_t) __builtin_neon_vclsq_v((int8x16_t)__p0, 33); + return __ret; +} +#else +__ai int16x8_t vclsq_u16(uint16x8_t __p0) { + uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0); + int16x8_t __ret; + __ret = (int16x8_t) __builtin_neon_vclsq_v((int8x16_t)__rev0, 33); + __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); + return __ret; +} +#endif + #ifdef __LITTLE_ENDIAN__ __ai int8x16_t vclsq_s8(int8x16_t __p0) { int8x16_t __ret; @@ -4600,6 +4738,54 @@ __ai int16x8_t vclsq_s16(int16x8_t __p0) { } #endif +#ifdef __LITTLE_ENDIAN__ +__ai int8x8_t vcls_u8(uint8x8_t __p0) { + int8x8_t __ret; + __ret = (int8x8_t) __builtin_neon_vcls_v((int8x8_t)__p0, 0); + return __ret; +} +#else +__ai int8x8_t vcls_u8(uint8x8_t __p0) { + uint8x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0); + int8x8_t __ret; + __ret = (int8x8_t) __builtin_neon_vcls_v((int8x8_t)__rev0, 0); + __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); + return __ret; +} +#endif + +#ifdef __LITTLE_ENDIAN__ +__ai int32x2_t vcls_u32(uint32x2_t __p0) { + int32x2_t __ret; + __ret = (int32x2_t) __builtin_neon_vcls_v((int8x8_t)__p0, 2); + return __ret; +} +#else +__ai int32x2_t vcls_u32(uint32x2_t __p0) { + uint32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0); + int32x2_t __ret; + __ret = (int32x2_t) __builtin_neon_vcls_v((int8x8_t)__rev0, 2); + __ret = __builtin_shufflevector(__ret, __ret, 1, 0); + return __ret; +} +#endif + +#ifdef __LITTLE_ENDIAN__ +__ai int16x4_t vcls_u16(uint16x4_t __p0) { + int16x4_t __ret; + __ret = (int16x4_t) __builtin_neon_vcls_v((int8x8_t)__p0, 1); + return __ret; +} +#else +__ai int16x4_t vcls_u16(uint16x4_t __p0) { + uint16x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0); + int16x4_t __ret; + __ret = (int16x4_t) __builtin_neon_vcls_v((int8x8_t)__rev0, 1); + __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); + return __ret; +} +#endif + #ifdef __LITTLE_ENDIAN__ __ai int8x8_t vcls_s8(int8x8_t __p0) { int8x8_t __ret; @@ -34889,7 +35075,7 @@ __ai uint32x2_t vcvtp_u32_f32(float32x2_t __p0) { #endif #endif -#if __ARM_ARCH >= 8 && defined(__ARM_FEATURE_CRYPTO) +#if __ARM_ARCH >= 8 && defined(__ARM_FEATURE_AES) #ifdef __LITTLE_ENDIAN__ __ai uint8x16_t vaesdq_u8(uint8x16_t __p0, uint8x16_t __p1) { uint8x16_t __ret; @@ -34956,168 +35142,6 @@ __ai uint8x16_t vaesmcq_u8(uint8x16_t __p0) { } #endif -#ifdef __LITTLE_ENDIAN__ -__ai uint32x4_t vsha1cq_u32(uint32x4_t __p0, uint32_t __p1, uint32x4_t __p2) { - uint32x4_t __ret; - __ret = (uint32x4_t) __builtin_neon_vsha1cq_u32(__p0, __p1, __p2); - return __ret; -} -#else -__ai uint32x4_t vsha1cq_u32(uint32x4_t __p0, uint32_t __p1, uint32x4_t __p2) { - uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0); - uint32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0); - uint32x4_t __ret; - __ret = (uint32x4_t) __builtin_neon_vsha1cq_u32(__rev0, __p1, __rev2); - __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); - return __ret; -} -#endif - -__ai uint32_t vsha1h_u32(uint32_t __p0) { - uint32_t __ret; - __ret = (uint32_t) __builtin_neon_vsha1h_u32(__p0); - return __ret; -} -#ifdef __LITTLE_ENDIAN__ -__ai uint32x4_t vsha1mq_u32(uint32x4_t __p0, uint32_t __p1, uint32x4_t __p2) { - uint32x4_t __ret; - __ret = (uint32x4_t) __builtin_neon_vsha1mq_u32(__p0, __p1, __p2); - return __ret; -} -#else -__ai uint32x4_t vsha1mq_u32(uint32x4_t __p0, uint32_t __p1, uint32x4_t __p2) { - uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0); - uint32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0); - uint32x4_t __ret; - __ret = (uint32x4_t) __builtin_neon_vsha1mq_u32(__rev0, __p1, __rev2); - __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); - return __ret; -} -#endif - -#ifdef __LITTLE_ENDIAN__ -__ai uint32x4_t vsha1pq_u32(uint32x4_t __p0, uint32_t __p1, uint32x4_t __p2) { - uint32x4_t __ret; - __ret = (uint32x4_t) __builtin_neon_vsha1pq_u32(__p0, __p1, __p2); - return __ret; -} -#else -__ai uint32x4_t vsha1pq_u32(uint32x4_t __p0, uint32_t __p1, uint32x4_t __p2) { - uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0); - uint32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0); - uint32x4_t __ret; - __ret = (uint32x4_t) __builtin_neon_vsha1pq_u32(__rev0, __p1, __rev2); - __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); - return __ret; -} -#endif - -#ifdef __LITTLE_ENDIAN__ -__ai uint32x4_t vsha1su0q_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) { - uint32x4_t __ret; - __ret = (uint32x4_t) __builtin_neon_vsha1su0q_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 50); - return __ret; -} -#else -__ai uint32x4_t vsha1su0q_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) { - uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0); - uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0); - uint32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0); - uint32x4_t __ret; - __ret = (uint32x4_t) __builtin_neon_vsha1su0q_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 50); - __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); - return __ret; -} -#endif - -#ifdef __LITTLE_ENDIAN__ -__ai uint32x4_t vsha1su1q_u32(uint32x4_t __p0, uint32x4_t __p1) { - uint32x4_t __ret; - __ret = (uint32x4_t) __builtin_neon_vsha1su1q_v((int8x16_t)__p0, (int8x16_t)__p1, 50); - return __ret; -} -#else -__ai uint32x4_t vsha1su1q_u32(uint32x4_t __p0, uint32x4_t __p1) { - uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0); - uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0); - uint32x4_t __ret; - __ret = (uint32x4_t) __builtin_neon_vsha1su1q_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50); - __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); - return __ret; -} -#endif - -#ifdef __LITTLE_ENDIAN__ -__ai uint32x4_t vsha256hq_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) { - uint32x4_t __ret; - __ret = (uint32x4_t) __builtin_neon_vsha256hq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 50); - return __ret; -} -#else -__ai uint32x4_t vsha256hq_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) { - uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0); - uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0); - uint32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0); - uint32x4_t __ret; - __ret = (uint32x4_t) __builtin_neon_vsha256hq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 50); - __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); - return __ret; -} -#endif - -#ifdef __LITTLE_ENDIAN__ -__ai uint32x4_t vsha256h2q_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) { - uint32x4_t __ret; - __ret = (uint32x4_t) __builtin_neon_vsha256h2q_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 50); - return __ret; -} -#else -__ai uint32x4_t vsha256h2q_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) { - uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0); - uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0); - uint32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0); - uint32x4_t __ret; - __ret = (uint32x4_t) __builtin_neon_vsha256h2q_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 50); - __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); - return __ret; -} -#endif - -#ifdef __LITTLE_ENDIAN__ -__ai uint32x4_t vsha256su0q_u32(uint32x4_t __p0, uint32x4_t __p1) { - uint32x4_t __ret; - __ret = (uint32x4_t) __builtin_neon_vsha256su0q_v((int8x16_t)__p0, (int8x16_t)__p1, 50); - return __ret; -} -#else -__ai uint32x4_t vsha256su0q_u32(uint32x4_t __p0, uint32x4_t __p1) { - uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0); - uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0); - uint32x4_t __ret; - __ret = (uint32x4_t) __builtin_neon_vsha256su0q_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50); - __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); - return __ret; -} -#endif - -#ifdef __LITTLE_ENDIAN__ -__ai uint32x4_t vsha256su1q_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) { - uint32x4_t __ret; - __ret = (uint32x4_t) __builtin_neon_vsha256su1q_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 50); - return __ret; -} -#else -__ai uint32x4_t vsha256su1q_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) { - uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0); - uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0); - uint32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0); - uint32x4_t __ret; - __ret = (uint32x4_t) __builtin_neon_vsha256su1q_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 50); - __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); - return __ret; -} -#endif - #endif #if __ARM_ARCH >= 8 && defined(__ARM_FEATURE_DIRECTED_ROUNDING) #ifdef __LITTLE_ENDIAN__ @@ -35683,6 +35707,759 @@ __ai float16x4_t vminnm_f16(float16x4_t __p0, float16x4_t __p1) { } #endif +#endif +#if __ARM_ARCH >= 8 && defined(__ARM_FEATURE_SHA2) +#ifdef __LITTLE_ENDIAN__ +__ai uint32x4_t vsha1cq_u32(uint32x4_t __p0, uint32_t __p1, uint32x4_t __p2) { + uint32x4_t __ret; + __ret = (uint32x4_t) __builtin_neon_vsha1cq_u32(__p0, __p1, __p2); + return __ret; +} +#else +__ai uint32x4_t vsha1cq_u32(uint32x4_t __p0, uint32_t __p1, uint32x4_t __p2) { + uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0); + uint32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0); + uint32x4_t __ret; + __ret = (uint32x4_t) __builtin_neon_vsha1cq_u32(__rev0, __p1, __rev2); + __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); + return __ret; +} +#endif + +__ai uint32_t vsha1h_u32(uint32_t __p0) { + uint32_t __ret; + __ret = (uint32_t) __builtin_neon_vsha1h_u32(__p0); + return __ret; +} +#ifdef __LITTLE_ENDIAN__ +__ai uint32x4_t vsha1mq_u32(uint32x4_t __p0, uint32_t __p1, uint32x4_t __p2) { + uint32x4_t __ret; + __ret = (uint32x4_t) __builtin_neon_vsha1mq_u32(__p0, __p1, __p2); + return __ret; +} +#else +__ai uint32x4_t vsha1mq_u32(uint32x4_t __p0, uint32_t __p1, uint32x4_t __p2) { + uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0); + uint32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0); + uint32x4_t __ret; + __ret = (uint32x4_t) __builtin_neon_vsha1mq_u32(__rev0, __p1, __rev2); + __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); + return __ret; +} +#endif + +#ifdef __LITTLE_ENDIAN__ +__ai uint32x4_t vsha1pq_u32(uint32x4_t __p0, uint32_t __p1, uint32x4_t __p2) { + uint32x4_t __ret; + __ret = (uint32x4_t) __builtin_neon_vsha1pq_u32(__p0, __p1, __p2); + return __ret; +} +#else +__ai uint32x4_t vsha1pq_u32(uint32x4_t __p0, uint32_t __p1, uint32x4_t __p2) { + uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0); + uint32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0); + uint32x4_t __ret; + __ret = (uint32x4_t) __builtin_neon_vsha1pq_u32(__rev0, __p1, __rev2); + __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); + return __ret; +} +#endif + +#ifdef __LITTLE_ENDIAN__ +__ai uint32x4_t vsha1su0q_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) { + uint32x4_t __ret; + __ret = (uint32x4_t) __builtin_neon_vsha1su0q_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 50); + return __ret; +} +#else +__ai uint32x4_t vsha1su0q_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) { + uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0); + uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0); + uint32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0); + uint32x4_t __ret; + __ret = (uint32x4_t) __builtin_neon_vsha1su0q_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 50); + __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); + return __ret; +} +#endif + +#ifdef __LITTLE_ENDIAN__ +__ai uint32x4_t vsha1su1q_u32(uint32x4_t __p0, uint32x4_t __p1) { + uint32x4_t __ret; + __ret = (uint32x4_t) __builtin_neon_vsha1su1q_v((int8x16_t)__p0, (int8x16_t)__p1, 50); + return __ret; +} +#else +__ai uint32x4_t vsha1su1q_u32(uint32x4_t __p0, uint32x4_t __p1) { + uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0); + uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0); + uint32x4_t __ret; + __ret = (uint32x4_t) __builtin_neon_vsha1su1q_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50); + __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); + return __ret; +} +#endif + +#ifdef __LITTLE_ENDIAN__ +__ai uint32x4_t vsha256hq_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) { + uint32x4_t __ret; + __ret = (uint32x4_t) __builtin_neon_vsha256hq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 50); + return __ret; +} +#else +__ai uint32x4_t vsha256hq_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) { + uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0); + uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0); + uint32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0); + uint32x4_t __ret; + __ret = (uint32x4_t) __builtin_neon_vsha256hq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 50); + __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); + return __ret; +} +#endif + +#ifdef __LITTLE_ENDIAN__ +__ai uint32x4_t vsha256h2q_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) { + uint32x4_t __ret; + __ret = (uint32x4_t) __builtin_neon_vsha256h2q_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 50); + return __ret; +} +#else +__ai uint32x4_t vsha256h2q_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) { + uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0); + uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0); + uint32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0); + uint32x4_t __ret; + __ret = (uint32x4_t) __builtin_neon_vsha256h2q_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 50); + __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); + return __ret; +} +#endif + +#ifdef __LITTLE_ENDIAN__ +__ai uint32x4_t vsha256su0q_u32(uint32x4_t __p0, uint32x4_t __p1) { + uint32x4_t __ret; + __ret = (uint32x4_t) __builtin_neon_vsha256su0q_v((int8x16_t)__p0, (int8x16_t)__p1, 50); + return __ret; +} +#else +__ai uint32x4_t vsha256su0q_u32(uint32x4_t __p0, uint32x4_t __p1) { + uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0); + uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0); + uint32x4_t __ret; + __ret = (uint32x4_t) __builtin_neon_vsha256su0q_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50); + __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); + return __ret; +} +#endif + +#ifdef __LITTLE_ENDIAN__ +__ai uint32x4_t vsha256su1q_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) { + uint32x4_t __ret; + __ret = (uint32x4_t) __builtin_neon_vsha256su1q_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 50); + return __ret; +} +#else +__ai uint32x4_t vsha256su1q_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) { + uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0); + uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0); + uint32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0); + uint32x4_t __ret; + __ret = (uint32x4_t) __builtin_neon_vsha256su1q_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 50); + __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); + return __ret; +} +#endif + +#endif +#if __ARM_ARCH >= 8 && defined(__ARM_FEATURE_SHA3) && defined(__aarch64__) +#ifdef __LITTLE_ENDIAN__ +__ai uint8x16_t vbcaxq_u8(uint8x16_t __p0, uint8x16_t __p1, uint8x16_t __p2) { + uint8x16_t __ret; + __ret = (uint8x16_t) __builtin_neon_vbcaxq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 48); + return __ret; +} +#else +__ai uint8x16_t vbcaxq_u8(uint8x16_t __p0, uint8x16_t __p1, uint8x16_t __p2) { + uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); + uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); + uint8x16_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); + uint8x16_t __ret; + __ret = (uint8x16_t) __builtin_neon_vbcaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 48); + __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); + return __ret; +} +#endif + +#ifdef __LITTLE_ENDIAN__ +__ai uint32x4_t vbcaxq_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) { + uint32x4_t __ret; + __ret = (uint32x4_t) __builtin_neon_vbcaxq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 50); + return __ret; +} +#else +__ai uint32x4_t vbcaxq_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) { + uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0); + uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0); + uint32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0); + uint32x4_t __ret; + __ret = (uint32x4_t) __builtin_neon_vbcaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 50); + __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); + return __ret; +} +#endif + +#ifdef __LITTLE_ENDIAN__ +__ai uint64x2_t vbcaxq_u64(uint64x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) { + uint64x2_t __ret; + __ret = (uint64x2_t) __builtin_neon_vbcaxq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 51); + return __ret; +} +#else +__ai uint64x2_t vbcaxq_u64(uint64x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) { + uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0); + uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0); + uint64x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0); + uint64x2_t __ret; + __ret = (uint64x2_t) __builtin_neon_vbcaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 51); + __ret = __builtin_shufflevector(__ret, __ret, 1, 0); + return __ret; +} +#endif + +#ifdef __LITTLE_ENDIAN__ +__ai uint16x8_t vbcaxq_u16(uint16x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) { + uint16x8_t __ret; + __ret = (uint16x8_t) __builtin_neon_vbcaxq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 49); + return __ret; +} +#else +__ai uint16x8_t vbcaxq_u16(uint16x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) { + uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0); + uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0); + uint16x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0); + uint16x8_t __ret; + __ret = (uint16x8_t) __builtin_neon_vbcaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 49); + __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); + return __ret; +} +#endif + +#ifdef __LITTLE_ENDIAN__ +__ai int8x16_t vbcaxq_s8(int8x16_t __p0, int8x16_t __p1, int8x16_t __p2) { + int8x16_t __ret; + __ret = (int8x16_t) __builtin_neon_vbcaxq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 32); + return __ret; +} +#else +__ai int8x16_t vbcaxq_s8(int8x16_t __p0, int8x16_t __p1, int8x16_t __p2) { + int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); + int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); + int8x16_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); + int8x16_t __ret; + __ret = (int8x16_t) __builtin_neon_vbcaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 32); + __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); + return __ret; +} +#endif + +#ifdef __LITTLE_ENDIAN__ +__ai int32x4_t vbcaxq_s32(int32x4_t __p0, int32x4_t __p1, int32x4_t __p2) { + int32x4_t __ret; + __ret = (int32x4_t) __builtin_neon_vbcaxq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 34); + return __ret; +} +#else +__ai int32x4_t vbcaxq_s32(int32x4_t __p0, int32x4_t __p1, int32x4_t __p2) { + int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0); + int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0); + int32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0); + int32x4_t __ret; + __ret = (int32x4_t) __builtin_neon_vbcaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 34); + __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); + return __ret; +} +#endif + +#ifdef __LITTLE_ENDIAN__ +__ai int64x2_t vbcaxq_s64(int64x2_t __p0, int64x2_t __p1, int64x2_t __p2) { + int64x2_t __ret; + __ret = (int64x2_t) __builtin_neon_vbcaxq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 35); + return __ret; +} +#else +__ai int64x2_t vbcaxq_s64(int64x2_t __p0, int64x2_t __p1, int64x2_t __p2) { + int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0); + int64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0); + int64x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0); + int64x2_t __ret; + __ret = (int64x2_t) __builtin_neon_vbcaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 35); + __ret = __builtin_shufflevector(__ret, __ret, 1, 0); + return __ret; +} +#endif + +#ifdef __LITTLE_ENDIAN__ +__ai int16x8_t vbcaxq_s16(int16x8_t __p0, int16x8_t __p1, int16x8_t __p2) { + int16x8_t __ret; + __ret = (int16x8_t) __builtin_neon_vbcaxq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 33); + return __ret; +} +#else +__ai int16x8_t vbcaxq_s16(int16x8_t __p0, int16x8_t __p1, int16x8_t __p2) { + int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0); + int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0); + int16x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0); + int16x8_t __ret; + __ret = (int16x8_t) __builtin_neon_vbcaxq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 33); + __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); + return __ret; +} +#endif + +#ifdef __LITTLE_ENDIAN__ +__ai uint8x16_t veor3q_u8(uint8x16_t __p0, uint8x16_t __p1, uint8x16_t __p2) { + uint8x16_t __ret; + __ret = (uint8x16_t) __builtin_neon_veor3q_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 48); + return __ret; +} +#else +__ai uint8x16_t veor3q_u8(uint8x16_t __p0, uint8x16_t __p1, uint8x16_t __p2) { + uint8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); + uint8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); + uint8x16_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); + uint8x16_t __ret; + __ret = (uint8x16_t) __builtin_neon_veor3q_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 48); + __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); + return __ret; +} +#endif + +#ifdef __LITTLE_ENDIAN__ +__ai uint32x4_t veor3q_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) { + uint32x4_t __ret; + __ret = (uint32x4_t) __builtin_neon_veor3q_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 50); + return __ret; +} +#else +__ai uint32x4_t veor3q_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) { + uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0); + uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0); + uint32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0); + uint32x4_t __ret; + __ret = (uint32x4_t) __builtin_neon_veor3q_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 50); + __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); + return __ret; +} +#endif + +#ifdef __LITTLE_ENDIAN__ +__ai uint64x2_t veor3q_u64(uint64x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) { + uint64x2_t __ret; + __ret = (uint64x2_t) __builtin_neon_veor3q_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 51); + return __ret; +} +#else +__ai uint64x2_t veor3q_u64(uint64x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) { + uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0); + uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0); + uint64x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0); + uint64x2_t __ret; + __ret = (uint64x2_t) __builtin_neon_veor3q_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 51); + __ret = __builtin_shufflevector(__ret, __ret, 1, 0); + return __ret; +} +#endif + +#ifdef __LITTLE_ENDIAN__ +__ai uint16x8_t veor3q_u16(uint16x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) { + uint16x8_t __ret; + __ret = (uint16x8_t) __builtin_neon_veor3q_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 49); + return __ret; +} +#else +__ai uint16x8_t veor3q_u16(uint16x8_t __p0, uint16x8_t __p1, uint16x8_t __p2) { + uint16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0); + uint16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0); + uint16x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0); + uint16x8_t __ret; + __ret = (uint16x8_t) __builtin_neon_veor3q_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 49); + __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); + return __ret; +} +#endif + +#ifdef __LITTLE_ENDIAN__ +__ai int8x16_t veor3q_s8(int8x16_t __p0, int8x16_t __p1, int8x16_t __p2) { + int8x16_t __ret; + __ret = (int8x16_t) __builtin_neon_veor3q_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 32); + return __ret; +} +#else +__ai int8x16_t veor3q_s8(int8x16_t __p0, int8x16_t __p1, int8x16_t __p2) { + int8x16_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); + int8x16_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); + int8x16_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); + int8x16_t __ret; + __ret = (int8x16_t) __builtin_neon_veor3q_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 32); + __ret = __builtin_shufflevector(__ret, __ret, 15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0); + return __ret; +} +#endif + +#ifdef __LITTLE_ENDIAN__ +__ai int32x4_t veor3q_s32(int32x4_t __p0, int32x4_t __p1, int32x4_t __p2) { + int32x4_t __ret; + __ret = (int32x4_t) __builtin_neon_veor3q_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 34); + return __ret; +} +#else +__ai int32x4_t veor3q_s32(int32x4_t __p0, int32x4_t __p1, int32x4_t __p2) { + int32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0); + int32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0); + int32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0); + int32x4_t __ret; + __ret = (int32x4_t) __builtin_neon_veor3q_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 34); + __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); + return __ret; +} +#endif + +#ifdef __LITTLE_ENDIAN__ +__ai int64x2_t veor3q_s64(int64x2_t __p0, int64x2_t __p1, int64x2_t __p2) { + int64x2_t __ret; + __ret = (int64x2_t) __builtin_neon_veor3q_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 35); + return __ret; +} +#else +__ai int64x2_t veor3q_s64(int64x2_t __p0, int64x2_t __p1, int64x2_t __p2) { + int64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0); + int64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0); + int64x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0); + int64x2_t __ret; + __ret = (int64x2_t) __builtin_neon_veor3q_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 35); + __ret = __builtin_shufflevector(__ret, __ret, 1, 0); + return __ret; +} +#endif + +#ifdef __LITTLE_ENDIAN__ +__ai int16x8_t veor3q_s16(int16x8_t __p0, int16x8_t __p1, int16x8_t __p2) { + int16x8_t __ret; + __ret = (int16x8_t) __builtin_neon_veor3q_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 33); + return __ret; +} +#else +__ai int16x8_t veor3q_s16(int16x8_t __p0, int16x8_t __p1, int16x8_t __p2) { + int16x8_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 7, 6, 5, 4, 3, 2, 1, 0); + int16x8_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 7, 6, 5, 4, 3, 2, 1, 0); + int16x8_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 7, 6, 5, 4, 3, 2, 1, 0); + int16x8_t __ret; + __ret = (int16x8_t) __builtin_neon_veor3q_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 33); + __ret = __builtin_shufflevector(__ret, __ret, 7, 6, 5, 4, 3, 2, 1, 0); + return __ret; +} +#endif + +#ifdef __LITTLE_ENDIAN__ +__ai uint64x2_t vrax1q_u64(uint64x2_t __p0, uint64x2_t __p1) { + uint64x2_t __ret; + __ret = (uint64x2_t) __builtin_neon_vrax1q_v((int8x16_t)__p0, (int8x16_t)__p1, 51); + return __ret; +} +#else +__ai uint64x2_t vrax1q_u64(uint64x2_t __p0, uint64x2_t __p1) { + uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0); + uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0); + uint64x2_t __ret; + __ret = (uint64x2_t) __builtin_neon_vrax1q_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51); + __ret = __builtin_shufflevector(__ret, __ret, 1, 0); + return __ret; +} +#endif + +#ifdef __LITTLE_ENDIAN__ +#define vxarq_u64(__p0, __p1, __p2) __extension__ ({ \ + uint64x2_t __s0 = __p0; \ + uint64x2_t __s1 = __p1; \ + uint64x2_t __ret; \ + __ret = (uint64x2_t) __builtin_neon_vxarq_v((int8x16_t)__s0, (int8x16_t)__s1, __p2, 51); \ + __ret; \ +}) +#else +#define vxarq_u64(__p0, __p1, __p2) __extension__ ({ \ + uint64x2_t __s0 = __p0; \ + uint64x2_t __s1 = __p1; \ + uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 1, 0); \ + uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 1, 0); \ + uint64x2_t __ret; \ + __ret = (uint64x2_t) __builtin_neon_vxarq_v((int8x16_t)__rev0, (int8x16_t)__rev1, __p2, 51); \ + __ret = __builtin_shufflevector(__ret, __ret, 1, 0); \ + __ret; \ +}) +#endif + +#endif +#if __ARM_ARCH >= 8 && defined(__ARM_FEATURE_SHA512) && defined(__aarch64__) +#ifdef __LITTLE_ENDIAN__ +__ai uint64x2_t vsha512hq_u64(uint64x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) { + uint64x2_t __ret; + __ret = (uint64x2_t) __builtin_neon_vsha512hq_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 51); + return __ret; +} +#else +__ai uint64x2_t vsha512hq_u64(uint64x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) { + uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0); + uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0); + uint64x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0); + uint64x2_t __ret; + __ret = (uint64x2_t) __builtin_neon_vsha512hq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 51); + __ret = __builtin_shufflevector(__ret, __ret, 1, 0); + return __ret; +} +#endif + +#ifdef __LITTLE_ENDIAN__ +__ai uint64x2_t vsha512h2q_u64(uint64x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) { + uint64x2_t __ret; + __ret = (uint64x2_t) __builtin_neon_vsha512h2q_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 51); + return __ret; +} +#else +__ai uint64x2_t vsha512h2q_u64(uint64x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) { + uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0); + uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0); + uint64x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0); + uint64x2_t __ret; + __ret = (uint64x2_t) __builtin_neon_vsha512h2q_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 51); + __ret = __builtin_shufflevector(__ret, __ret, 1, 0); + return __ret; +} +#endif + +#ifdef __LITTLE_ENDIAN__ +__ai uint64x2_t vsha512su0q_u64(uint64x2_t __p0, uint64x2_t __p1) { + uint64x2_t __ret; + __ret = (uint64x2_t) __builtin_neon_vsha512su0q_v((int8x16_t)__p0, (int8x16_t)__p1, 51); + return __ret; +} +#else +__ai uint64x2_t vsha512su0q_u64(uint64x2_t __p0, uint64x2_t __p1) { + uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0); + uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0); + uint64x2_t __ret; + __ret = (uint64x2_t) __builtin_neon_vsha512su0q_v((int8x16_t)__rev0, (int8x16_t)__rev1, 51); + __ret = __builtin_shufflevector(__ret, __ret, 1, 0); + return __ret; +} +#endif + +#ifdef __LITTLE_ENDIAN__ +__ai uint64x2_t vsha512su1q_u64(uint64x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) { + uint64x2_t __ret; + __ret = (uint64x2_t) __builtin_neon_vsha512su1q_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 51); + return __ret; +} +#else +__ai uint64x2_t vsha512su1q_u64(uint64x2_t __p0, uint64x2_t __p1, uint64x2_t __p2) { + uint64x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0); + uint64x2_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 1, 0); + uint64x2_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 1, 0); + uint64x2_t __ret; + __ret = (uint64x2_t) __builtin_neon_vsha512su1q_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 51); + __ret = __builtin_shufflevector(__ret, __ret, 1, 0); + return __ret; +} +#endif + +#endif +#if __ARM_ARCH >= 8 && defined(__ARM_FEATURE_SM3) && defined(__aarch64__) +#ifdef __LITTLE_ENDIAN__ +__ai uint32x4_t vsm3partw1q_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) { + uint32x4_t __ret; + __ret = (uint32x4_t) __builtin_neon_vsm3partw1q_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 50); + return __ret; +} +#else +__ai uint32x4_t vsm3partw1q_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) { + uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0); + uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0); + uint32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0); + uint32x4_t __ret; + __ret = (uint32x4_t) __builtin_neon_vsm3partw1q_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 50); + __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); + return __ret; +} +#endif + +#ifdef __LITTLE_ENDIAN__ +__ai uint32x4_t vsm3partw2q_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) { + uint32x4_t __ret; + __ret = (uint32x4_t) __builtin_neon_vsm3partw2q_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 50); + return __ret; +} +#else +__ai uint32x4_t vsm3partw2q_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) { + uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0); + uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0); + uint32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0); + uint32x4_t __ret; + __ret = (uint32x4_t) __builtin_neon_vsm3partw2q_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 50); + __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); + return __ret; +} +#endif + +#ifdef __LITTLE_ENDIAN__ +__ai uint32x4_t vsm3ss1q_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) { + uint32x4_t __ret; + __ret = (uint32x4_t) __builtin_neon_vsm3ss1q_v((int8x16_t)__p0, (int8x16_t)__p1, (int8x16_t)__p2, 50); + return __ret; +} +#else +__ai uint32x4_t vsm3ss1q_u32(uint32x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) { + uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0); + uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0); + uint32x4_t __rev2; __rev2 = __builtin_shufflevector(__p2, __p2, 3, 2, 1, 0); + uint32x4_t __ret; + __ret = (uint32x4_t) __builtin_neon_vsm3ss1q_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, 50); + __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); + return __ret; +} +#endif + +#ifdef __LITTLE_ENDIAN__ +#define vsm3tt1aq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \ + uint32x4_t __s0 = __p0; \ + uint32x4_t __s1 = __p1; \ + uint32x4_t __s2 = __p2; \ + uint32x4_t __ret; \ + __ret = (uint32x4_t) __builtin_neon_vsm3tt1aq_v((int8x16_t)__s0, (int8x16_t)__s1, (int8x16_t)__s2, __p3, 50); \ + __ret; \ +}) +#else +#define vsm3tt1aq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \ + uint32x4_t __s0 = __p0; \ + uint32x4_t __s1 = __p1; \ + uint32x4_t __s2 = __p2; \ + uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \ + uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \ + uint32x4_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \ + uint32x4_t __ret; \ + __ret = (uint32x4_t) __builtin_neon_vsm3tt1aq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, __p3, 50); \ + __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \ + __ret; \ +}) +#endif + +#ifdef __LITTLE_ENDIAN__ +#define vsm3tt1bq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \ + uint32x4_t __s0 = __p0; \ + uint32x4_t __s1 = __p1; \ + uint32x4_t __s2 = __p2; \ + uint32x4_t __ret; \ + __ret = (uint32x4_t) __builtin_neon_vsm3tt1bq_v((int8x16_t)__s0, (int8x16_t)__s1, (int8x16_t)__s2, __p3, 50); \ + __ret; \ +}) +#else +#define vsm3tt1bq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \ + uint32x4_t __s0 = __p0; \ + uint32x4_t __s1 = __p1; \ + uint32x4_t __s2 = __p2; \ + uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \ + uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \ + uint32x4_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \ + uint32x4_t __ret; \ + __ret = (uint32x4_t) __builtin_neon_vsm3tt1bq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, __p3, 50); \ + __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \ + __ret; \ +}) +#endif + +#ifdef __LITTLE_ENDIAN__ +#define vsm3tt2aq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \ + uint32x4_t __s0 = __p0; \ + uint32x4_t __s1 = __p1; \ + uint32x4_t __s2 = __p2; \ + uint32x4_t __ret; \ + __ret = (uint32x4_t) __builtin_neon_vsm3tt2aq_v((int8x16_t)__s0, (int8x16_t)__s1, (int8x16_t)__s2, __p3, 50); \ + __ret; \ +}) +#else +#define vsm3tt2aq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \ + uint32x4_t __s0 = __p0; \ + uint32x4_t __s1 = __p1; \ + uint32x4_t __s2 = __p2; \ + uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \ + uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \ + uint32x4_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \ + uint32x4_t __ret; \ + __ret = (uint32x4_t) __builtin_neon_vsm3tt2aq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, __p3, 50); \ + __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \ + __ret; \ +}) +#endif + +#ifdef __LITTLE_ENDIAN__ +#define vsm3tt2bq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \ + uint32x4_t __s0 = __p0; \ + uint32x4_t __s1 = __p1; \ + uint32x4_t __s2 = __p2; \ + uint32x4_t __ret; \ + __ret = (uint32x4_t) __builtin_neon_vsm3tt2bq_v((int8x16_t)__s0, (int8x16_t)__s1, (int8x16_t)__s2, __p3, 50); \ + __ret; \ +}) +#else +#define vsm3tt2bq_u32(__p0, __p1, __p2, __p3) __extension__ ({ \ + uint32x4_t __s0 = __p0; \ + uint32x4_t __s1 = __p1; \ + uint32x4_t __s2 = __p2; \ + uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__s0, __s0, 3, 2, 1, 0); \ + uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__s1, __s1, 3, 2, 1, 0); \ + uint32x4_t __rev2; __rev2 = __builtin_shufflevector(__s2, __s2, 3, 2, 1, 0); \ + uint32x4_t __ret; \ + __ret = (uint32x4_t) __builtin_neon_vsm3tt2bq_v((int8x16_t)__rev0, (int8x16_t)__rev1, (int8x16_t)__rev2, __p3, 50); \ + __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); \ + __ret; \ +}) +#endif + +#endif +#if __ARM_ARCH >= 8 && defined(__ARM_FEATURE_SM4) && defined(__aarch64__) +#ifdef __LITTLE_ENDIAN__ +__ai uint32x4_t vsm4eq_u32(uint32x4_t __p0, uint32x4_t __p1) { + uint32x4_t __ret; + __ret = (uint32x4_t) __builtin_neon_vsm4eq_v((int8x16_t)__p0, (int8x16_t)__p1, 50); + return __ret; +} +#else +__ai uint32x4_t vsm4eq_u32(uint32x4_t __p0, uint32x4_t __p1) { + uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0); + uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0); + uint32x4_t __ret; + __ret = (uint32x4_t) __builtin_neon_vsm4eq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50); + __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); + return __ret; +} +#endif + +#ifdef __LITTLE_ENDIAN__ +__ai uint32x4_t vsm4ekeyq_u32(uint32x4_t __p0, uint32x4_t __p1) { + uint32x4_t __ret; + __ret = (uint32x4_t) __builtin_neon_vsm4ekeyq_v((int8x16_t)__p0, (int8x16_t)__p1, 50); + return __ret; +} +#else +__ai uint32x4_t vsm4ekeyq_u32(uint32x4_t __p0, uint32x4_t __p1) { + uint32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0); + uint32x4_t __rev1; __rev1 = __builtin_shufflevector(__p1, __p1, 3, 2, 1, 0); + uint32x4_t __ret; + __ret = (uint32x4_t) __builtin_neon_vsm4ekeyq_v((int8x16_t)__rev0, (int8x16_t)__rev1, 50); + __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); + return __ret; +} +#endif + #endif #if __ARM_ARCH >= 8 && defined(__aarch64__) #ifdef __LITTLE_ENDIAN__ @@ -37962,6 +38739,136 @@ __ai float64x1_t vrndx_f64(float64x1_t __p0) { __ret = (float64x1_t) __builtin_neon_vrndx_v((int8x8_t)__p0, 10); return __ret; } +#endif +#if __ARM_ARCH >= 8 && defined(__aarch64__) && defined(__ARM_FEATURE_FRINT) +#ifdef __LITTLE_ENDIAN__ +__ai float32x4_t vrnd32xq_f32(float32x4_t __p0) { + float32x4_t __ret; + __ret = (float32x4_t) __builtin_neon_vrnd32xq_v((int8x16_t)__p0, 41); + return __ret; +} +#else +__ai float32x4_t vrnd32xq_f32(float32x4_t __p0) { + float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0); + float32x4_t __ret; + __ret = (float32x4_t) __builtin_neon_vrnd32xq_v((int8x16_t)__rev0, 41); + __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); + return __ret; +} +#endif + +#ifdef __LITTLE_ENDIAN__ +__ai float32x2_t vrnd32x_f32(float32x2_t __p0) { + float32x2_t __ret; + __ret = (float32x2_t) __builtin_neon_vrnd32x_v((int8x8_t)__p0, 9); + return __ret; +} +#else +__ai float32x2_t vrnd32x_f32(float32x2_t __p0) { + float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0); + float32x2_t __ret; + __ret = (float32x2_t) __builtin_neon_vrnd32x_v((int8x8_t)__rev0, 9); + __ret = __builtin_shufflevector(__ret, __ret, 1, 0); + return __ret; +} +#endif + +#ifdef __LITTLE_ENDIAN__ +__ai float32x4_t vrnd32zq_f32(float32x4_t __p0) { + float32x4_t __ret; + __ret = (float32x4_t) __builtin_neon_vrnd32zq_v((int8x16_t)__p0, 41); + return __ret; +} +#else +__ai float32x4_t vrnd32zq_f32(float32x4_t __p0) { + float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0); + float32x4_t __ret; + __ret = (float32x4_t) __builtin_neon_vrnd32zq_v((int8x16_t)__rev0, 41); + __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); + return __ret; +} +#endif + +#ifdef __LITTLE_ENDIAN__ +__ai float32x2_t vrnd32z_f32(float32x2_t __p0) { + float32x2_t __ret; + __ret = (float32x2_t) __builtin_neon_vrnd32z_v((int8x8_t)__p0, 9); + return __ret; +} +#else +__ai float32x2_t vrnd32z_f32(float32x2_t __p0) { + float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0); + float32x2_t __ret; + __ret = (float32x2_t) __builtin_neon_vrnd32z_v((int8x8_t)__rev0, 9); + __ret = __builtin_shufflevector(__ret, __ret, 1, 0); + return __ret; +} +#endif + +#ifdef __LITTLE_ENDIAN__ +__ai float32x4_t vrnd64xq_f32(float32x4_t __p0) { + float32x4_t __ret; + __ret = (float32x4_t) __builtin_neon_vrnd64xq_v((int8x16_t)__p0, 41); + return __ret; +} +#else +__ai float32x4_t vrnd64xq_f32(float32x4_t __p0) { + float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0); + float32x4_t __ret; + __ret = (float32x4_t) __builtin_neon_vrnd64xq_v((int8x16_t)__rev0, 41); + __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); + return __ret; +} +#endif + +#ifdef __LITTLE_ENDIAN__ +__ai float32x2_t vrnd64x_f32(float32x2_t __p0) { + float32x2_t __ret; + __ret = (float32x2_t) __builtin_neon_vrnd64x_v((int8x8_t)__p0, 9); + return __ret; +} +#else +__ai float32x2_t vrnd64x_f32(float32x2_t __p0) { + float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0); + float32x2_t __ret; + __ret = (float32x2_t) __builtin_neon_vrnd64x_v((int8x8_t)__rev0, 9); + __ret = __builtin_shufflevector(__ret, __ret, 1, 0); + return __ret; +} +#endif + +#ifdef __LITTLE_ENDIAN__ +__ai float32x4_t vrnd64zq_f32(float32x4_t __p0) { + float32x4_t __ret; + __ret = (float32x4_t) __builtin_neon_vrnd64zq_v((int8x16_t)__p0, 41); + return __ret; +} +#else +__ai float32x4_t vrnd64zq_f32(float32x4_t __p0) { + float32x4_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 3, 2, 1, 0); + float32x4_t __ret; + __ret = (float32x4_t) __builtin_neon_vrnd64zq_v((int8x16_t)__rev0, 41); + __ret = __builtin_shufflevector(__ret, __ret, 3, 2, 1, 0); + return __ret; +} +#endif + +#ifdef __LITTLE_ENDIAN__ +__ai float32x2_t vrnd64z_f32(float32x2_t __p0) { + float32x2_t __ret; + __ret = (float32x2_t) __builtin_neon_vrnd64z_v((int8x8_t)__p0, 9); + return __ret; +} +#else +__ai float32x2_t vrnd64z_f32(float32x2_t __p0) { + float32x2_t __rev0; __rev0 = __builtin_shufflevector(__p0, __p0, 1, 0); + float32x2_t __ret; + __ret = (float32x2_t) __builtin_neon_vrnd64z_v((int8x8_t)__rev0, 9); + __ret = __builtin_shufflevector(__ret, __ret, 1, 0); + return __ret; +} +#endif + #endif #if __ARM_ARCH >= 8 && defined(__aarch64__) && defined(__ARM_FEATURE_NUMERIC_MAXMIN) #ifdef __LITTLE_ENDIAN__ @@ -42710,7 +43617,7 @@ __ai float32x2_t vfms_f32(float32x2_t __p0, float32x2_t __p1, float32x2_t __p2) #endif #endif -#if defined(__ARM_FEATURE_FP16FML) && defined(__aarch64__) +#if defined(__ARM_FEATURE_FP16_FML) && defined(__aarch64__) #ifdef __LITTLE_ENDIAN__ __ai float32x4_t vfmlalq_high_f16(float32x4_t __p0, float16x8_t __p1, float16x8_t __p2) { float32x4_t __ret; @@ -46786,6 +47693,11 @@ __ai int64_t vaddd_s64(int64_t __p0, int64_t __p1) { __ret = (int64_t) __builtin_neon_vaddd_s64(__p0, __p1); return __ret; } +__ai poly128_t vaddq_p128(poly128_t __p0, poly128_t __p1) { + poly128_t __ret; + __ret = (poly128_t) __builtin_neon_vaddq_p128(__p0, __p1); + return __ret; +} #ifdef __LITTLE_ENDIAN__ __ai uint16x8_t vaddhn_high_u32(uint16x4_t __p0, uint32x4_t __p1, uint32x4_t __p2) { uint16x8_t __ret; @@ -67227,7 +68139,7 @@ __ai float32x4_t vcvtq_low_f32_bf16(bfloat16x8_t __p0) { #endif #endif -#if defined(__ARM_FEATURE_FP16FML) && defined(__aarch64__) +#if defined(__ARM_FEATURE_FP16_FML) && defined(__aarch64__) #ifdef __LITTLE_ENDIAN__ #define vfmlalq_lane_high_f16(__p0_869, __p1_869, __p2_869, __p3_869) __extension__ ({ \ float32x4_t __s0_869 = __p0_869; \ diff --git a/lib/include/arm_sve.h b/lib/include/arm_sve.h index 8a03f9da5..909634a56 100644 --- a/lib/include/arm_sve.h +++ b/lib/include/arm_sve.h @@ -132,7 +132,9 @@ enum svprfop }; /* Function attributes */ -#define __aio static inline __attribute__((__always_inline__, __nodebug__, __overloadable__)) +#define __ai static __inline__ __attribute__((__always_inline__, __nodebug__)) + +#define __aio static __inline__ __attribute__((__always_inline__, __nodebug__, __overloadable__)) #define svreinterpret_s8_s8(...) __builtin_sve_reinterpret_s8_s8(__VA_ARGS__) #define svreinterpret_s8_s16(...) __builtin_sve_reinterpret_s8_s16(__VA_ARGS__) @@ -946,3825 +948,7644 @@ __aio svfloat64_t svreinterpret_f64(svfloat64_t op) { return __builtin_sve_reinterpret_f64_f64(op); } -#define svabd_n_f64_m(...) __builtin_sve_svabd_n_f64_m(__VA_ARGS__) -#define svabd_n_f32_m(...) __builtin_sve_svabd_n_f32_m(__VA_ARGS__) -#define svabd_n_f16_m(...) __builtin_sve_svabd_n_f16_m(__VA_ARGS__) -#define svabd_n_f64_x(...) __builtin_sve_svabd_n_f64_x(__VA_ARGS__) -#define svabd_n_f32_x(...) __builtin_sve_svabd_n_f32_x(__VA_ARGS__) -#define svabd_n_f16_x(...) __builtin_sve_svabd_n_f16_x(__VA_ARGS__) -#define svabd_n_f64_z(...) __builtin_sve_svabd_n_f64_z(__VA_ARGS__) -#define svabd_n_f32_z(...) __builtin_sve_svabd_n_f32_z(__VA_ARGS__) -#define svabd_n_f16_z(...) __builtin_sve_svabd_n_f16_z(__VA_ARGS__) -#define svabd_n_s8_m(...) __builtin_sve_svabd_n_s8_m(__VA_ARGS__) -#define svabd_n_s32_m(...) __builtin_sve_svabd_n_s32_m(__VA_ARGS__) -#define svabd_n_s64_m(...) __builtin_sve_svabd_n_s64_m(__VA_ARGS__) -#define svabd_n_s16_m(...) __builtin_sve_svabd_n_s16_m(__VA_ARGS__) -#define svabd_n_s8_x(...) __builtin_sve_svabd_n_s8_x(__VA_ARGS__) -#define svabd_n_s32_x(...) __builtin_sve_svabd_n_s32_x(__VA_ARGS__) -#define svabd_n_s64_x(...) __builtin_sve_svabd_n_s64_x(__VA_ARGS__) -#define svabd_n_s16_x(...) __builtin_sve_svabd_n_s16_x(__VA_ARGS__) -#define svabd_n_s8_z(...) __builtin_sve_svabd_n_s8_z(__VA_ARGS__) -#define svabd_n_s32_z(...) __builtin_sve_svabd_n_s32_z(__VA_ARGS__) -#define svabd_n_s64_z(...) __builtin_sve_svabd_n_s64_z(__VA_ARGS__) -#define svabd_n_s16_z(...) __builtin_sve_svabd_n_s16_z(__VA_ARGS__) -#define svabd_n_u8_m(...) __builtin_sve_svabd_n_u8_m(__VA_ARGS__) -#define svabd_n_u32_m(...) __builtin_sve_svabd_n_u32_m(__VA_ARGS__) -#define svabd_n_u64_m(...) __builtin_sve_svabd_n_u64_m(__VA_ARGS__) -#define svabd_n_u16_m(...) __builtin_sve_svabd_n_u16_m(__VA_ARGS__) -#define svabd_n_u8_x(...) __builtin_sve_svabd_n_u8_x(__VA_ARGS__) -#define svabd_n_u32_x(...) __builtin_sve_svabd_n_u32_x(__VA_ARGS__) -#define svabd_n_u64_x(...) __builtin_sve_svabd_n_u64_x(__VA_ARGS__) -#define svabd_n_u16_x(...) __builtin_sve_svabd_n_u16_x(__VA_ARGS__) -#define svabd_n_u8_z(...) __builtin_sve_svabd_n_u8_z(__VA_ARGS__) -#define svabd_n_u32_z(...) __builtin_sve_svabd_n_u32_z(__VA_ARGS__) -#define svabd_n_u64_z(...) __builtin_sve_svabd_n_u64_z(__VA_ARGS__) -#define svabd_n_u16_z(...) __builtin_sve_svabd_n_u16_z(__VA_ARGS__) -#define svabd_f64_m(...) __builtin_sve_svabd_f64_m(__VA_ARGS__) -#define svabd_f32_m(...) __builtin_sve_svabd_f32_m(__VA_ARGS__) -#define svabd_f16_m(...) __builtin_sve_svabd_f16_m(__VA_ARGS__) -#define svabd_f64_x(...) __builtin_sve_svabd_f64_x(__VA_ARGS__) -#define svabd_f32_x(...) __builtin_sve_svabd_f32_x(__VA_ARGS__) -#define svabd_f16_x(...) __builtin_sve_svabd_f16_x(__VA_ARGS__) -#define svabd_f64_z(...) __builtin_sve_svabd_f64_z(__VA_ARGS__) -#define svabd_f32_z(...) __builtin_sve_svabd_f32_z(__VA_ARGS__) -#define svabd_f16_z(...) __builtin_sve_svabd_f16_z(__VA_ARGS__) -#define svabd_s8_m(...) __builtin_sve_svabd_s8_m(__VA_ARGS__) -#define svabd_s32_m(...) __builtin_sve_svabd_s32_m(__VA_ARGS__) -#define svabd_s64_m(...) __builtin_sve_svabd_s64_m(__VA_ARGS__) -#define svabd_s16_m(...) __builtin_sve_svabd_s16_m(__VA_ARGS__) -#define svabd_s8_x(...) __builtin_sve_svabd_s8_x(__VA_ARGS__) -#define svabd_s32_x(...) __builtin_sve_svabd_s32_x(__VA_ARGS__) -#define svabd_s64_x(...) __builtin_sve_svabd_s64_x(__VA_ARGS__) -#define svabd_s16_x(...) __builtin_sve_svabd_s16_x(__VA_ARGS__) -#define svabd_s8_z(...) __builtin_sve_svabd_s8_z(__VA_ARGS__) -#define svabd_s32_z(...) __builtin_sve_svabd_s32_z(__VA_ARGS__) -#define svabd_s64_z(...) __builtin_sve_svabd_s64_z(__VA_ARGS__) -#define svabd_s16_z(...) __builtin_sve_svabd_s16_z(__VA_ARGS__) -#define svabd_u8_m(...) __builtin_sve_svabd_u8_m(__VA_ARGS__) -#define svabd_u32_m(...) __builtin_sve_svabd_u32_m(__VA_ARGS__) -#define svabd_u64_m(...) __builtin_sve_svabd_u64_m(__VA_ARGS__) -#define svabd_u16_m(...) __builtin_sve_svabd_u16_m(__VA_ARGS__) -#define svabd_u8_x(...) __builtin_sve_svabd_u8_x(__VA_ARGS__) -#define svabd_u32_x(...) __builtin_sve_svabd_u32_x(__VA_ARGS__) -#define svabd_u64_x(...) __builtin_sve_svabd_u64_x(__VA_ARGS__) -#define svabd_u16_x(...) __builtin_sve_svabd_u16_x(__VA_ARGS__) -#define svabd_u8_z(...) __builtin_sve_svabd_u8_z(__VA_ARGS__) -#define svabd_u32_z(...) __builtin_sve_svabd_u32_z(__VA_ARGS__) -#define svabd_u64_z(...) __builtin_sve_svabd_u64_z(__VA_ARGS__) -#define svabd_u16_z(...) __builtin_sve_svabd_u16_z(__VA_ARGS__) -#define svabs_f64_m(...) __builtin_sve_svabs_f64_m(__VA_ARGS__) -#define svabs_f32_m(...) __builtin_sve_svabs_f32_m(__VA_ARGS__) -#define svabs_f16_m(...) __builtin_sve_svabs_f16_m(__VA_ARGS__) -#define svabs_f64_x(...) __builtin_sve_svabs_f64_x(__VA_ARGS__) -#define svabs_f32_x(...) __builtin_sve_svabs_f32_x(__VA_ARGS__) -#define svabs_f16_x(...) __builtin_sve_svabs_f16_x(__VA_ARGS__) -#define svabs_f64_z(...) __builtin_sve_svabs_f64_z(__VA_ARGS__) -#define svabs_f32_z(...) __builtin_sve_svabs_f32_z(__VA_ARGS__) -#define svabs_f16_z(...) __builtin_sve_svabs_f16_z(__VA_ARGS__) -#define svabs_s8_m(...) __builtin_sve_svabs_s8_m(__VA_ARGS__) -#define svabs_s32_m(...) __builtin_sve_svabs_s32_m(__VA_ARGS__) -#define svabs_s64_m(...) __builtin_sve_svabs_s64_m(__VA_ARGS__) -#define svabs_s16_m(...) __builtin_sve_svabs_s16_m(__VA_ARGS__) -#define svabs_s8_x(...) __builtin_sve_svabs_s8_x(__VA_ARGS__) -#define svabs_s32_x(...) __builtin_sve_svabs_s32_x(__VA_ARGS__) -#define svabs_s64_x(...) __builtin_sve_svabs_s64_x(__VA_ARGS__) -#define svabs_s16_x(...) __builtin_sve_svabs_s16_x(__VA_ARGS__) -#define svabs_s8_z(...) __builtin_sve_svabs_s8_z(__VA_ARGS__) -#define svabs_s32_z(...) __builtin_sve_svabs_s32_z(__VA_ARGS__) -#define svabs_s64_z(...) __builtin_sve_svabs_s64_z(__VA_ARGS__) -#define svabs_s16_z(...) __builtin_sve_svabs_s16_z(__VA_ARGS__) -#define svacge_n_f64(...) __builtin_sve_svacge_n_f64(__VA_ARGS__) -#define svacge_n_f32(...) __builtin_sve_svacge_n_f32(__VA_ARGS__) -#define svacge_n_f16(...) __builtin_sve_svacge_n_f16(__VA_ARGS__) -#define svacge_f64(...) __builtin_sve_svacge_f64(__VA_ARGS__) -#define svacge_f32(...) __builtin_sve_svacge_f32(__VA_ARGS__) -#define svacge_f16(...) __builtin_sve_svacge_f16(__VA_ARGS__) -#define svacgt_n_f64(...) __builtin_sve_svacgt_n_f64(__VA_ARGS__) -#define svacgt_n_f32(...) __builtin_sve_svacgt_n_f32(__VA_ARGS__) -#define svacgt_n_f16(...) __builtin_sve_svacgt_n_f16(__VA_ARGS__) -#define svacgt_f64(...) __builtin_sve_svacgt_f64(__VA_ARGS__) -#define svacgt_f32(...) __builtin_sve_svacgt_f32(__VA_ARGS__) -#define svacgt_f16(...) __builtin_sve_svacgt_f16(__VA_ARGS__) -#define svacle_n_f64(...) __builtin_sve_svacle_n_f64(__VA_ARGS__) -#define svacle_n_f32(...) __builtin_sve_svacle_n_f32(__VA_ARGS__) -#define svacle_n_f16(...) __builtin_sve_svacle_n_f16(__VA_ARGS__) -#define svacle_f64(...) __builtin_sve_svacle_f64(__VA_ARGS__) -#define svacle_f32(...) __builtin_sve_svacle_f32(__VA_ARGS__) -#define svacle_f16(...) __builtin_sve_svacle_f16(__VA_ARGS__) -#define svaclt_n_f64(...) __builtin_sve_svaclt_n_f64(__VA_ARGS__) -#define svaclt_n_f32(...) __builtin_sve_svaclt_n_f32(__VA_ARGS__) -#define svaclt_n_f16(...) __builtin_sve_svaclt_n_f16(__VA_ARGS__) -#define svaclt_f64(...) __builtin_sve_svaclt_f64(__VA_ARGS__) -#define svaclt_f32(...) __builtin_sve_svaclt_f32(__VA_ARGS__) -#define svaclt_f16(...) __builtin_sve_svaclt_f16(__VA_ARGS__) -#define svadd_n_f64_m(...) __builtin_sve_svadd_n_f64_m(__VA_ARGS__) -#define svadd_n_f32_m(...) __builtin_sve_svadd_n_f32_m(__VA_ARGS__) -#define svadd_n_f16_m(...) __builtin_sve_svadd_n_f16_m(__VA_ARGS__) -#define svadd_n_f64_x(...) __builtin_sve_svadd_n_f64_x(__VA_ARGS__) -#define svadd_n_f32_x(...) __builtin_sve_svadd_n_f32_x(__VA_ARGS__) -#define svadd_n_f16_x(...) __builtin_sve_svadd_n_f16_x(__VA_ARGS__) -#define svadd_n_f64_z(...) __builtin_sve_svadd_n_f64_z(__VA_ARGS__) -#define svadd_n_f32_z(...) __builtin_sve_svadd_n_f32_z(__VA_ARGS__) -#define svadd_n_f16_z(...) __builtin_sve_svadd_n_f16_z(__VA_ARGS__) -#define svadd_n_u8_m(...) __builtin_sve_svadd_n_u8_m(__VA_ARGS__) -#define svadd_n_u32_m(...) __builtin_sve_svadd_n_u32_m(__VA_ARGS__) -#define svadd_n_u64_m(...) __builtin_sve_svadd_n_u64_m(__VA_ARGS__) -#define svadd_n_u16_m(...) __builtin_sve_svadd_n_u16_m(__VA_ARGS__) -#define svadd_n_s8_m(...) __builtin_sve_svadd_n_s8_m(__VA_ARGS__) -#define svadd_n_s32_m(...) __builtin_sve_svadd_n_s32_m(__VA_ARGS__) -#define svadd_n_s64_m(...) __builtin_sve_svadd_n_s64_m(__VA_ARGS__) -#define svadd_n_s16_m(...) __builtin_sve_svadd_n_s16_m(__VA_ARGS__) -#define svadd_n_u8_x(...) __builtin_sve_svadd_n_u8_x(__VA_ARGS__) -#define svadd_n_u32_x(...) __builtin_sve_svadd_n_u32_x(__VA_ARGS__) -#define svadd_n_u64_x(...) __builtin_sve_svadd_n_u64_x(__VA_ARGS__) -#define svadd_n_u16_x(...) __builtin_sve_svadd_n_u16_x(__VA_ARGS__) -#define svadd_n_s8_x(...) __builtin_sve_svadd_n_s8_x(__VA_ARGS__) -#define svadd_n_s32_x(...) __builtin_sve_svadd_n_s32_x(__VA_ARGS__) -#define svadd_n_s64_x(...) __builtin_sve_svadd_n_s64_x(__VA_ARGS__) -#define svadd_n_s16_x(...) __builtin_sve_svadd_n_s16_x(__VA_ARGS__) -#define svadd_n_u8_z(...) __builtin_sve_svadd_n_u8_z(__VA_ARGS__) -#define svadd_n_u32_z(...) __builtin_sve_svadd_n_u32_z(__VA_ARGS__) -#define svadd_n_u64_z(...) __builtin_sve_svadd_n_u64_z(__VA_ARGS__) -#define svadd_n_u16_z(...) __builtin_sve_svadd_n_u16_z(__VA_ARGS__) -#define svadd_n_s8_z(...) __builtin_sve_svadd_n_s8_z(__VA_ARGS__) -#define svadd_n_s32_z(...) __builtin_sve_svadd_n_s32_z(__VA_ARGS__) -#define svadd_n_s64_z(...) __builtin_sve_svadd_n_s64_z(__VA_ARGS__) -#define svadd_n_s16_z(...) __builtin_sve_svadd_n_s16_z(__VA_ARGS__) -#define svadd_f64_m(...) __builtin_sve_svadd_f64_m(__VA_ARGS__) -#define svadd_f32_m(...) __builtin_sve_svadd_f32_m(__VA_ARGS__) -#define svadd_f16_m(...) __builtin_sve_svadd_f16_m(__VA_ARGS__) -#define svadd_f64_x(...) __builtin_sve_svadd_f64_x(__VA_ARGS__) -#define svadd_f32_x(...) __builtin_sve_svadd_f32_x(__VA_ARGS__) -#define svadd_f16_x(...) __builtin_sve_svadd_f16_x(__VA_ARGS__) -#define svadd_f64_z(...) __builtin_sve_svadd_f64_z(__VA_ARGS__) -#define svadd_f32_z(...) __builtin_sve_svadd_f32_z(__VA_ARGS__) -#define svadd_f16_z(...) __builtin_sve_svadd_f16_z(__VA_ARGS__) -#define svadd_u8_m(...) __builtin_sve_svadd_u8_m(__VA_ARGS__) -#define svadd_u32_m(...) __builtin_sve_svadd_u32_m(__VA_ARGS__) -#define svadd_u64_m(...) __builtin_sve_svadd_u64_m(__VA_ARGS__) -#define svadd_u16_m(...) __builtin_sve_svadd_u16_m(__VA_ARGS__) -#define svadd_s8_m(...) __builtin_sve_svadd_s8_m(__VA_ARGS__) -#define svadd_s32_m(...) __builtin_sve_svadd_s32_m(__VA_ARGS__) -#define svadd_s64_m(...) __builtin_sve_svadd_s64_m(__VA_ARGS__) -#define svadd_s16_m(...) __builtin_sve_svadd_s16_m(__VA_ARGS__) -#define svadd_u8_x(...) __builtin_sve_svadd_u8_x(__VA_ARGS__) -#define svadd_u32_x(...) __builtin_sve_svadd_u32_x(__VA_ARGS__) -#define svadd_u64_x(...) __builtin_sve_svadd_u64_x(__VA_ARGS__) -#define svadd_u16_x(...) __builtin_sve_svadd_u16_x(__VA_ARGS__) -#define svadd_s8_x(...) __builtin_sve_svadd_s8_x(__VA_ARGS__) -#define svadd_s32_x(...) __builtin_sve_svadd_s32_x(__VA_ARGS__) -#define svadd_s64_x(...) __builtin_sve_svadd_s64_x(__VA_ARGS__) -#define svadd_s16_x(...) __builtin_sve_svadd_s16_x(__VA_ARGS__) -#define svadd_u8_z(...) __builtin_sve_svadd_u8_z(__VA_ARGS__) -#define svadd_u32_z(...) __builtin_sve_svadd_u32_z(__VA_ARGS__) -#define svadd_u64_z(...) __builtin_sve_svadd_u64_z(__VA_ARGS__) -#define svadd_u16_z(...) __builtin_sve_svadd_u16_z(__VA_ARGS__) -#define svadd_s8_z(...) __builtin_sve_svadd_s8_z(__VA_ARGS__) -#define svadd_s32_z(...) __builtin_sve_svadd_s32_z(__VA_ARGS__) -#define svadd_s64_z(...) __builtin_sve_svadd_s64_z(__VA_ARGS__) -#define svadd_s16_z(...) __builtin_sve_svadd_s16_z(__VA_ARGS__) -#define svadda_f64(...) __builtin_sve_svadda_f64(__VA_ARGS__) -#define svadda_f32(...) __builtin_sve_svadda_f32(__VA_ARGS__) -#define svadda_f16(...) __builtin_sve_svadda_f16(__VA_ARGS__) -#define svaddv_s8(...) __builtin_sve_svaddv_s8(__VA_ARGS__) -#define svaddv_s32(...) __builtin_sve_svaddv_s32(__VA_ARGS__) -#define svaddv_s64(...) __builtin_sve_svaddv_s64(__VA_ARGS__) -#define svaddv_s16(...) __builtin_sve_svaddv_s16(__VA_ARGS__) -#define svaddv_u8(...) __builtin_sve_svaddv_u8(__VA_ARGS__) -#define svaddv_u32(...) __builtin_sve_svaddv_u32(__VA_ARGS__) -#define svaddv_u64(...) __builtin_sve_svaddv_u64(__VA_ARGS__) -#define svaddv_u16(...) __builtin_sve_svaddv_u16(__VA_ARGS__) -#define svaddv_f64(...) __builtin_sve_svaddv_f64(__VA_ARGS__) -#define svaddv_f32(...) __builtin_sve_svaddv_f32(__VA_ARGS__) -#define svaddv_f16(...) __builtin_sve_svaddv_f16(__VA_ARGS__) -#define svadrb_u32base_u32offset(...) __builtin_sve_svadrb_u32base_u32offset(__VA_ARGS__) -#define svadrb_u64base_u64offset(...) __builtin_sve_svadrb_u64base_u64offset(__VA_ARGS__) -#define svadrb_u32base_s32offset(...) __builtin_sve_svadrb_u32base_s32offset(__VA_ARGS__) -#define svadrb_u64base_s64offset(...) __builtin_sve_svadrb_u64base_s64offset(__VA_ARGS__) -#define svadrd_u32base_u32index(...) __builtin_sve_svadrd_u32base_u32index(__VA_ARGS__) -#define svadrd_u64base_u64index(...) __builtin_sve_svadrd_u64base_u64index(__VA_ARGS__) -#define svadrd_u32base_s32index(...) __builtin_sve_svadrd_u32base_s32index(__VA_ARGS__) -#define svadrd_u64base_s64index(...) __builtin_sve_svadrd_u64base_s64index(__VA_ARGS__) -#define svadrh_u32base_u32index(...) __builtin_sve_svadrh_u32base_u32index(__VA_ARGS__) -#define svadrh_u64base_u64index(...) __builtin_sve_svadrh_u64base_u64index(__VA_ARGS__) -#define svadrh_u32base_s32index(...) __builtin_sve_svadrh_u32base_s32index(__VA_ARGS__) -#define svadrh_u64base_s64index(...) __builtin_sve_svadrh_u64base_s64index(__VA_ARGS__) -#define svadrw_u32base_u32index(...) __builtin_sve_svadrw_u32base_u32index(__VA_ARGS__) -#define svadrw_u64base_u64index(...) __builtin_sve_svadrw_u64base_u64index(__VA_ARGS__) -#define svadrw_u32base_s32index(...) __builtin_sve_svadrw_u32base_s32index(__VA_ARGS__) -#define svadrw_u64base_s64index(...) __builtin_sve_svadrw_u64base_s64index(__VA_ARGS__) -#define svand_b_z(...) __builtin_sve_svand_b_z(__VA_ARGS__) -#define svand_n_u8_m(...) __builtin_sve_svand_n_u8_m(__VA_ARGS__) -#define svand_n_u32_m(...) __builtin_sve_svand_n_u32_m(__VA_ARGS__) -#define svand_n_u64_m(...) __builtin_sve_svand_n_u64_m(__VA_ARGS__) -#define svand_n_u16_m(...) __builtin_sve_svand_n_u16_m(__VA_ARGS__) -#define svand_n_s8_m(...) __builtin_sve_svand_n_s8_m(__VA_ARGS__) -#define svand_n_s32_m(...) __builtin_sve_svand_n_s32_m(__VA_ARGS__) -#define svand_n_s64_m(...) __builtin_sve_svand_n_s64_m(__VA_ARGS__) -#define svand_n_s16_m(...) __builtin_sve_svand_n_s16_m(__VA_ARGS__) -#define svand_n_u8_x(...) __builtin_sve_svand_n_u8_x(__VA_ARGS__) -#define svand_n_u32_x(...) __builtin_sve_svand_n_u32_x(__VA_ARGS__) -#define svand_n_u64_x(...) __builtin_sve_svand_n_u64_x(__VA_ARGS__) -#define svand_n_u16_x(...) __builtin_sve_svand_n_u16_x(__VA_ARGS__) -#define svand_n_s8_x(...) __builtin_sve_svand_n_s8_x(__VA_ARGS__) -#define svand_n_s32_x(...) __builtin_sve_svand_n_s32_x(__VA_ARGS__) -#define svand_n_s64_x(...) __builtin_sve_svand_n_s64_x(__VA_ARGS__) -#define svand_n_s16_x(...) __builtin_sve_svand_n_s16_x(__VA_ARGS__) -#define svand_n_u8_z(...) __builtin_sve_svand_n_u8_z(__VA_ARGS__) -#define svand_n_u32_z(...) __builtin_sve_svand_n_u32_z(__VA_ARGS__) -#define svand_n_u64_z(...) __builtin_sve_svand_n_u64_z(__VA_ARGS__) -#define svand_n_u16_z(...) __builtin_sve_svand_n_u16_z(__VA_ARGS__) -#define svand_n_s8_z(...) __builtin_sve_svand_n_s8_z(__VA_ARGS__) -#define svand_n_s32_z(...) __builtin_sve_svand_n_s32_z(__VA_ARGS__) -#define svand_n_s64_z(...) __builtin_sve_svand_n_s64_z(__VA_ARGS__) -#define svand_n_s16_z(...) __builtin_sve_svand_n_s16_z(__VA_ARGS__) -#define svand_u8_m(...) __builtin_sve_svand_u8_m(__VA_ARGS__) -#define svand_u32_m(...) __builtin_sve_svand_u32_m(__VA_ARGS__) -#define svand_u64_m(...) __builtin_sve_svand_u64_m(__VA_ARGS__) -#define svand_u16_m(...) __builtin_sve_svand_u16_m(__VA_ARGS__) -#define svand_s8_m(...) __builtin_sve_svand_s8_m(__VA_ARGS__) -#define svand_s32_m(...) __builtin_sve_svand_s32_m(__VA_ARGS__) -#define svand_s64_m(...) __builtin_sve_svand_s64_m(__VA_ARGS__) -#define svand_s16_m(...) __builtin_sve_svand_s16_m(__VA_ARGS__) -#define svand_u8_x(...) __builtin_sve_svand_u8_x(__VA_ARGS__) -#define svand_u32_x(...) __builtin_sve_svand_u32_x(__VA_ARGS__) -#define svand_u64_x(...) __builtin_sve_svand_u64_x(__VA_ARGS__) -#define svand_u16_x(...) __builtin_sve_svand_u16_x(__VA_ARGS__) -#define svand_s8_x(...) __builtin_sve_svand_s8_x(__VA_ARGS__) -#define svand_s32_x(...) __builtin_sve_svand_s32_x(__VA_ARGS__) -#define svand_s64_x(...) __builtin_sve_svand_s64_x(__VA_ARGS__) -#define svand_s16_x(...) __builtin_sve_svand_s16_x(__VA_ARGS__) -#define svand_u8_z(...) __builtin_sve_svand_u8_z(__VA_ARGS__) -#define svand_u32_z(...) __builtin_sve_svand_u32_z(__VA_ARGS__) -#define svand_u64_z(...) __builtin_sve_svand_u64_z(__VA_ARGS__) -#define svand_u16_z(...) __builtin_sve_svand_u16_z(__VA_ARGS__) -#define svand_s8_z(...) __builtin_sve_svand_s8_z(__VA_ARGS__) -#define svand_s32_z(...) __builtin_sve_svand_s32_z(__VA_ARGS__) -#define svand_s64_z(...) __builtin_sve_svand_s64_z(__VA_ARGS__) -#define svand_s16_z(...) __builtin_sve_svand_s16_z(__VA_ARGS__) -#define svandv_u8(...) __builtin_sve_svandv_u8(__VA_ARGS__) -#define svandv_u32(...) __builtin_sve_svandv_u32(__VA_ARGS__) -#define svandv_u64(...) __builtin_sve_svandv_u64(__VA_ARGS__) -#define svandv_u16(...) __builtin_sve_svandv_u16(__VA_ARGS__) -#define svandv_s8(...) __builtin_sve_svandv_s8(__VA_ARGS__) -#define svandv_s32(...) __builtin_sve_svandv_s32(__VA_ARGS__) -#define svandv_s64(...) __builtin_sve_svandv_s64(__VA_ARGS__) -#define svandv_s16(...) __builtin_sve_svandv_s16(__VA_ARGS__) -#define svasr_n_s8_m(...) __builtin_sve_svasr_n_s8_m(__VA_ARGS__) -#define svasr_n_s32_m(...) __builtin_sve_svasr_n_s32_m(__VA_ARGS__) -#define svasr_n_s64_m(...) __builtin_sve_svasr_n_s64_m(__VA_ARGS__) -#define svasr_n_s16_m(...) __builtin_sve_svasr_n_s16_m(__VA_ARGS__) -#define svasr_n_s8_x(...) __builtin_sve_svasr_n_s8_x(__VA_ARGS__) -#define svasr_n_s32_x(...) __builtin_sve_svasr_n_s32_x(__VA_ARGS__) -#define svasr_n_s64_x(...) __builtin_sve_svasr_n_s64_x(__VA_ARGS__) -#define svasr_n_s16_x(...) __builtin_sve_svasr_n_s16_x(__VA_ARGS__) -#define svasr_n_s8_z(...) __builtin_sve_svasr_n_s8_z(__VA_ARGS__) -#define svasr_n_s32_z(...) __builtin_sve_svasr_n_s32_z(__VA_ARGS__) -#define svasr_n_s64_z(...) __builtin_sve_svasr_n_s64_z(__VA_ARGS__) -#define svasr_n_s16_z(...) __builtin_sve_svasr_n_s16_z(__VA_ARGS__) -#define svasr_s8_m(...) __builtin_sve_svasr_s8_m(__VA_ARGS__) -#define svasr_s32_m(...) __builtin_sve_svasr_s32_m(__VA_ARGS__) -#define svasr_s64_m(...) __builtin_sve_svasr_s64_m(__VA_ARGS__) -#define svasr_s16_m(...) __builtin_sve_svasr_s16_m(__VA_ARGS__) -#define svasr_s8_x(...) __builtin_sve_svasr_s8_x(__VA_ARGS__) -#define svasr_s32_x(...) __builtin_sve_svasr_s32_x(__VA_ARGS__) -#define svasr_s64_x(...) __builtin_sve_svasr_s64_x(__VA_ARGS__) -#define svasr_s16_x(...) __builtin_sve_svasr_s16_x(__VA_ARGS__) -#define svasr_s8_z(...) __builtin_sve_svasr_s8_z(__VA_ARGS__) -#define svasr_s32_z(...) __builtin_sve_svasr_s32_z(__VA_ARGS__) -#define svasr_s64_z(...) __builtin_sve_svasr_s64_z(__VA_ARGS__) -#define svasr_s16_z(...) __builtin_sve_svasr_s16_z(__VA_ARGS__) -#define svasr_wide_n_s8_m(...) __builtin_sve_svasr_wide_n_s8_m(__VA_ARGS__) -#define svasr_wide_n_s32_m(...) __builtin_sve_svasr_wide_n_s32_m(__VA_ARGS__) -#define svasr_wide_n_s16_m(...) __builtin_sve_svasr_wide_n_s16_m(__VA_ARGS__) -#define svasr_wide_n_s8_x(...) __builtin_sve_svasr_wide_n_s8_x(__VA_ARGS__) -#define svasr_wide_n_s32_x(...) __builtin_sve_svasr_wide_n_s32_x(__VA_ARGS__) -#define svasr_wide_n_s16_x(...) __builtin_sve_svasr_wide_n_s16_x(__VA_ARGS__) -#define svasr_wide_n_s8_z(...) __builtin_sve_svasr_wide_n_s8_z(__VA_ARGS__) -#define svasr_wide_n_s32_z(...) __builtin_sve_svasr_wide_n_s32_z(__VA_ARGS__) -#define svasr_wide_n_s16_z(...) __builtin_sve_svasr_wide_n_s16_z(__VA_ARGS__) -#define svasr_wide_s8_m(...) __builtin_sve_svasr_wide_s8_m(__VA_ARGS__) -#define svasr_wide_s32_m(...) __builtin_sve_svasr_wide_s32_m(__VA_ARGS__) -#define svasr_wide_s16_m(...) __builtin_sve_svasr_wide_s16_m(__VA_ARGS__) -#define svasr_wide_s8_x(...) __builtin_sve_svasr_wide_s8_x(__VA_ARGS__) -#define svasr_wide_s32_x(...) __builtin_sve_svasr_wide_s32_x(__VA_ARGS__) -#define svasr_wide_s16_x(...) __builtin_sve_svasr_wide_s16_x(__VA_ARGS__) -#define svasr_wide_s8_z(...) __builtin_sve_svasr_wide_s8_z(__VA_ARGS__) -#define svasr_wide_s32_z(...) __builtin_sve_svasr_wide_s32_z(__VA_ARGS__) -#define svasr_wide_s16_z(...) __builtin_sve_svasr_wide_s16_z(__VA_ARGS__) -#define svasrd_n_s8_m(...) __builtin_sve_svasrd_n_s8_m(__VA_ARGS__) -#define svasrd_n_s32_m(...) __builtin_sve_svasrd_n_s32_m(__VA_ARGS__) -#define svasrd_n_s64_m(...) __builtin_sve_svasrd_n_s64_m(__VA_ARGS__) -#define svasrd_n_s16_m(...) __builtin_sve_svasrd_n_s16_m(__VA_ARGS__) -#define svasrd_n_s8_x(...) __builtin_sve_svasrd_n_s8_x(__VA_ARGS__) -#define svasrd_n_s32_x(...) __builtin_sve_svasrd_n_s32_x(__VA_ARGS__) -#define svasrd_n_s64_x(...) __builtin_sve_svasrd_n_s64_x(__VA_ARGS__) -#define svasrd_n_s16_x(...) __builtin_sve_svasrd_n_s16_x(__VA_ARGS__) -#define svasrd_n_s8_z(...) __builtin_sve_svasrd_n_s8_z(__VA_ARGS__) -#define svasrd_n_s32_z(...) __builtin_sve_svasrd_n_s32_z(__VA_ARGS__) -#define svasrd_n_s64_z(...) __builtin_sve_svasrd_n_s64_z(__VA_ARGS__) -#define svasrd_n_s16_z(...) __builtin_sve_svasrd_n_s16_z(__VA_ARGS__) -#define svbic_b_z(...) __builtin_sve_svbic_b_z(__VA_ARGS__) -#define svbic_n_u8_m(...) __builtin_sve_svbic_n_u8_m(__VA_ARGS__) -#define svbic_n_u32_m(...) __builtin_sve_svbic_n_u32_m(__VA_ARGS__) -#define svbic_n_u64_m(...) __builtin_sve_svbic_n_u64_m(__VA_ARGS__) -#define svbic_n_u16_m(...) __builtin_sve_svbic_n_u16_m(__VA_ARGS__) -#define svbic_n_s8_m(...) __builtin_sve_svbic_n_s8_m(__VA_ARGS__) -#define svbic_n_s32_m(...) __builtin_sve_svbic_n_s32_m(__VA_ARGS__) -#define svbic_n_s64_m(...) __builtin_sve_svbic_n_s64_m(__VA_ARGS__) -#define svbic_n_s16_m(...) __builtin_sve_svbic_n_s16_m(__VA_ARGS__) -#define svbic_n_u8_x(...) __builtin_sve_svbic_n_u8_x(__VA_ARGS__) -#define svbic_n_u32_x(...) __builtin_sve_svbic_n_u32_x(__VA_ARGS__) -#define svbic_n_u64_x(...) __builtin_sve_svbic_n_u64_x(__VA_ARGS__) -#define svbic_n_u16_x(...) __builtin_sve_svbic_n_u16_x(__VA_ARGS__) -#define svbic_n_s8_x(...) __builtin_sve_svbic_n_s8_x(__VA_ARGS__) -#define svbic_n_s32_x(...) __builtin_sve_svbic_n_s32_x(__VA_ARGS__) -#define svbic_n_s64_x(...) __builtin_sve_svbic_n_s64_x(__VA_ARGS__) -#define svbic_n_s16_x(...) __builtin_sve_svbic_n_s16_x(__VA_ARGS__) -#define svbic_n_u8_z(...) __builtin_sve_svbic_n_u8_z(__VA_ARGS__) -#define svbic_n_u32_z(...) __builtin_sve_svbic_n_u32_z(__VA_ARGS__) -#define svbic_n_u64_z(...) __builtin_sve_svbic_n_u64_z(__VA_ARGS__) -#define svbic_n_u16_z(...) __builtin_sve_svbic_n_u16_z(__VA_ARGS__) -#define svbic_n_s8_z(...) __builtin_sve_svbic_n_s8_z(__VA_ARGS__) -#define svbic_n_s32_z(...) __builtin_sve_svbic_n_s32_z(__VA_ARGS__) -#define svbic_n_s64_z(...) __builtin_sve_svbic_n_s64_z(__VA_ARGS__) -#define svbic_n_s16_z(...) __builtin_sve_svbic_n_s16_z(__VA_ARGS__) -#define svbic_u8_m(...) __builtin_sve_svbic_u8_m(__VA_ARGS__) -#define svbic_u32_m(...) __builtin_sve_svbic_u32_m(__VA_ARGS__) -#define svbic_u64_m(...) __builtin_sve_svbic_u64_m(__VA_ARGS__) -#define svbic_u16_m(...) __builtin_sve_svbic_u16_m(__VA_ARGS__) -#define svbic_s8_m(...) __builtin_sve_svbic_s8_m(__VA_ARGS__) -#define svbic_s32_m(...) __builtin_sve_svbic_s32_m(__VA_ARGS__) -#define svbic_s64_m(...) __builtin_sve_svbic_s64_m(__VA_ARGS__) -#define svbic_s16_m(...) __builtin_sve_svbic_s16_m(__VA_ARGS__) -#define svbic_u8_x(...) __builtin_sve_svbic_u8_x(__VA_ARGS__) -#define svbic_u32_x(...) __builtin_sve_svbic_u32_x(__VA_ARGS__) -#define svbic_u64_x(...) __builtin_sve_svbic_u64_x(__VA_ARGS__) -#define svbic_u16_x(...) __builtin_sve_svbic_u16_x(__VA_ARGS__) -#define svbic_s8_x(...) __builtin_sve_svbic_s8_x(__VA_ARGS__) -#define svbic_s32_x(...) __builtin_sve_svbic_s32_x(__VA_ARGS__) -#define svbic_s64_x(...) __builtin_sve_svbic_s64_x(__VA_ARGS__) -#define svbic_s16_x(...) __builtin_sve_svbic_s16_x(__VA_ARGS__) -#define svbic_u8_z(...) __builtin_sve_svbic_u8_z(__VA_ARGS__) -#define svbic_u32_z(...) __builtin_sve_svbic_u32_z(__VA_ARGS__) -#define svbic_u64_z(...) __builtin_sve_svbic_u64_z(__VA_ARGS__) -#define svbic_u16_z(...) __builtin_sve_svbic_u16_z(__VA_ARGS__) -#define svbic_s8_z(...) __builtin_sve_svbic_s8_z(__VA_ARGS__) -#define svbic_s32_z(...) __builtin_sve_svbic_s32_z(__VA_ARGS__) -#define svbic_s64_z(...) __builtin_sve_svbic_s64_z(__VA_ARGS__) -#define svbic_s16_z(...) __builtin_sve_svbic_s16_z(__VA_ARGS__) -#define svbrka_b_m(...) __builtin_sve_svbrka_b_m(__VA_ARGS__) -#define svbrka_b_z(...) __builtin_sve_svbrka_b_z(__VA_ARGS__) -#define svbrkb_b_m(...) __builtin_sve_svbrkb_b_m(__VA_ARGS__) -#define svbrkb_b_z(...) __builtin_sve_svbrkb_b_z(__VA_ARGS__) -#define svbrkn_b_z(...) __builtin_sve_svbrkn_b_z(__VA_ARGS__) -#define svbrkpa_b_z(...) __builtin_sve_svbrkpa_b_z(__VA_ARGS__) -#define svbrkpb_b_z(...) __builtin_sve_svbrkpb_b_z(__VA_ARGS__) -#define svcadd_f64_m(...) __builtin_sve_svcadd_f64_m(__VA_ARGS__) -#define svcadd_f32_m(...) __builtin_sve_svcadd_f32_m(__VA_ARGS__) -#define svcadd_f16_m(...) __builtin_sve_svcadd_f16_m(__VA_ARGS__) -#define svcadd_f64_x(...) __builtin_sve_svcadd_f64_x(__VA_ARGS__) -#define svcadd_f32_x(...) __builtin_sve_svcadd_f32_x(__VA_ARGS__) -#define svcadd_f16_x(...) __builtin_sve_svcadd_f16_x(__VA_ARGS__) -#define svcadd_f64_z(...) __builtin_sve_svcadd_f64_z(__VA_ARGS__) -#define svcadd_f32_z(...) __builtin_sve_svcadd_f32_z(__VA_ARGS__) -#define svcadd_f16_z(...) __builtin_sve_svcadd_f16_z(__VA_ARGS__) -#define svclasta_n_u8(...) __builtin_sve_svclasta_n_u8(__VA_ARGS__) -#define svclasta_n_u32(...) __builtin_sve_svclasta_n_u32(__VA_ARGS__) -#define svclasta_n_u64(...) __builtin_sve_svclasta_n_u64(__VA_ARGS__) -#define svclasta_n_u16(...) __builtin_sve_svclasta_n_u16(__VA_ARGS__) -#define svclasta_n_s8(...) __builtin_sve_svclasta_n_s8(__VA_ARGS__) -#define svclasta_n_f64(...) __builtin_sve_svclasta_n_f64(__VA_ARGS__) -#define svclasta_n_f32(...) __builtin_sve_svclasta_n_f32(__VA_ARGS__) -#define svclasta_n_f16(...) __builtin_sve_svclasta_n_f16(__VA_ARGS__) -#define svclasta_n_s32(...) __builtin_sve_svclasta_n_s32(__VA_ARGS__) -#define svclasta_n_s64(...) __builtin_sve_svclasta_n_s64(__VA_ARGS__) -#define svclasta_n_s16(...) __builtin_sve_svclasta_n_s16(__VA_ARGS__) -#define svclasta_u8(...) __builtin_sve_svclasta_u8(__VA_ARGS__) -#define svclasta_u32(...) __builtin_sve_svclasta_u32(__VA_ARGS__) -#define svclasta_u64(...) __builtin_sve_svclasta_u64(__VA_ARGS__) -#define svclasta_u16(...) __builtin_sve_svclasta_u16(__VA_ARGS__) -#define svclasta_s8(...) __builtin_sve_svclasta_s8(__VA_ARGS__) -#define svclasta_f64(...) __builtin_sve_svclasta_f64(__VA_ARGS__) -#define svclasta_f32(...) __builtin_sve_svclasta_f32(__VA_ARGS__) -#define svclasta_f16(...) __builtin_sve_svclasta_f16(__VA_ARGS__) -#define svclasta_s32(...) __builtin_sve_svclasta_s32(__VA_ARGS__) -#define svclasta_s64(...) __builtin_sve_svclasta_s64(__VA_ARGS__) -#define svclasta_s16(...) __builtin_sve_svclasta_s16(__VA_ARGS__) -#define svclastb_n_u8(...) __builtin_sve_svclastb_n_u8(__VA_ARGS__) -#define svclastb_n_u32(...) __builtin_sve_svclastb_n_u32(__VA_ARGS__) -#define svclastb_n_u64(...) __builtin_sve_svclastb_n_u64(__VA_ARGS__) -#define svclastb_n_u16(...) __builtin_sve_svclastb_n_u16(__VA_ARGS__) -#define svclastb_n_s8(...) __builtin_sve_svclastb_n_s8(__VA_ARGS__) -#define svclastb_n_f64(...) __builtin_sve_svclastb_n_f64(__VA_ARGS__) -#define svclastb_n_f32(...) __builtin_sve_svclastb_n_f32(__VA_ARGS__) -#define svclastb_n_f16(...) __builtin_sve_svclastb_n_f16(__VA_ARGS__) -#define svclastb_n_s32(...) __builtin_sve_svclastb_n_s32(__VA_ARGS__) -#define svclastb_n_s64(...) __builtin_sve_svclastb_n_s64(__VA_ARGS__) -#define svclastb_n_s16(...) __builtin_sve_svclastb_n_s16(__VA_ARGS__) -#define svclastb_u8(...) __builtin_sve_svclastb_u8(__VA_ARGS__) -#define svclastb_u32(...) __builtin_sve_svclastb_u32(__VA_ARGS__) -#define svclastb_u64(...) __builtin_sve_svclastb_u64(__VA_ARGS__) -#define svclastb_u16(...) __builtin_sve_svclastb_u16(__VA_ARGS__) -#define svclastb_s8(...) __builtin_sve_svclastb_s8(__VA_ARGS__) -#define svclastb_f64(...) __builtin_sve_svclastb_f64(__VA_ARGS__) -#define svclastb_f32(...) __builtin_sve_svclastb_f32(__VA_ARGS__) -#define svclastb_f16(...) __builtin_sve_svclastb_f16(__VA_ARGS__) -#define svclastb_s32(...) __builtin_sve_svclastb_s32(__VA_ARGS__) -#define svclastb_s64(...) __builtin_sve_svclastb_s64(__VA_ARGS__) -#define svclastb_s16(...) __builtin_sve_svclastb_s16(__VA_ARGS__) -#define svcls_s8_m(...) __builtin_sve_svcls_s8_m(__VA_ARGS__) -#define svcls_s32_m(...) __builtin_sve_svcls_s32_m(__VA_ARGS__) -#define svcls_s64_m(...) __builtin_sve_svcls_s64_m(__VA_ARGS__) -#define svcls_s16_m(...) __builtin_sve_svcls_s16_m(__VA_ARGS__) -#define svcls_s8_x(...) __builtin_sve_svcls_s8_x(__VA_ARGS__) -#define svcls_s32_x(...) __builtin_sve_svcls_s32_x(__VA_ARGS__) -#define svcls_s64_x(...) __builtin_sve_svcls_s64_x(__VA_ARGS__) -#define svcls_s16_x(...) __builtin_sve_svcls_s16_x(__VA_ARGS__) -#define svcls_s8_z(...) __builtin_sve_svcls_s8_z(__VA_ARGS__) -#define svcls_s32_z(...) __builtin_sve_svcls_s32_z(__VA_ARGS__) -#define svcls_s64_z(...) __builtin_sve_svcls_s64_z(__VA_ARGS__) -#define svcls_s16_z(...) __builtin_sve_svcls_s16_z(__VA_ARGS__) -#define svclz_u8_m(...) __builtin_sve_svclz_u8_m(__VA_ARGS__) -#define svclz_u32_m(...) __builtin_sve_svclz_u32_m(__VA_ARGS__) -#define svclz_u64_m(...) __builtin_sve_svclz_u64_m(__VA_ARGS__) -#define svclz_u16_m(...) __builtin_sve_svclz_u16_m(__VA_ARGS__) -#define svclz_s8_m(...) __builtin_sve_svclz_s8_m(__VA_ARGS__) -#define svclz_s32_m(...) __builtin_sve_svclz_s32_m(__VA_ARGS__) -#define svclz_s64_m(...) __builtin_sve_svclz_s64_m(__VA_ARGS__) -#define svclz_s16_m(...) __builtin_sve_svclz_s16_m(__VA_ARGS__) -#define svclz_u8_x(...) __builtin_sve_svclz_u8_x(__VA_ARGS__) -#define svclz_u32_x(...) __builtin_sve_svclz_u32_x(__VA_ARGS__) -#define svclz_u64_x(...) __builtin_sve_svclz_u64_x(__VA_ARGS__) -#define svclz_u16_x(...) __builtin_sve_svclz_u16_x(__VA_ARGS__) -#define svclz_s8_x(...) __builtin_sve_svclz_s8_x(__VA_ARGS__) -#define svclz_s32_x(...) __builtin_sve_svclz_s32_x(__VA_ARGS__) -#define svclz_s64_x(...) __builtin_sve_svclz_s64_x(__VA_ARGS__) -#define svclz_s16_x(...) __builtin_sve_svclz_s16_x(__VA_ARGS__) -#define svclz_u8_z(...) __builtin_sve_svclz_u8_z(__VA_ARGS__) -#define svclz_u32_z(...) __builtin_sve_svclz_u32_z(__VA_ARGS__) -#define svclz_u64_z(...) __builtin_sve_svclz_u64_z(__VA_ARGS__) -#define svclz_u16_z(...) __builtin_sve_svclz_u16_z(__VA_ARGS__) -#define svclz_s8_z(...) __builtin_sve_svclz_s8_z(__VA_ARGS__) -#define svclz_s32_z(...) __builtin_sve_svclz_s32_z(__VA_ARGS__) -#define svclz_s64_z(...) __builtin_sve_svclz_s64_z(__VA_ARGS__) -#define svclz_s16_z(...) __builtin_sve_svclz_s16_z(__VA_ARGS__) -#define svcmla_f64_m(...) __builtin_sve_svcmla_f64_m(__VA_ARGS__) -#define svcmla_f32_m(...) __builtin_sve_svcmla_f32_m(__VA_ARGS__) -#define svcmla_f16_m(...) __builtin_sve_svcmla_f16_m(__VA_ARGS__) -#define svcmla_f64_x(...) __builtin_sve_svcmla_f64_x(__VA_ARGS__) -#define svcmla_f32_x(...) __builtin_sve_svcmla_f32_x(__VA_ARGS__) -#define svcmla_f16_x(...) __builtin_sve_svcmla_f16_x(__VA_ARGS__) -#define svcmla_f64_z(...) __builtin_sve_svcmla_f64_z(__VA_ARGS__) -#define svcmla_f32_z(...) __builtin_sve_svcmla_f32_z(__VA_ARGS__) -#define svcmla_f16_z(...) __builtin_sve_svcmla_f16_z(__VA_ARGS__) -#define svcmla_lane_f32(...) __builtin_sve_svcmla_lane_f32(__VA_ARGS__) -#define svcmla_lane_f16(...) __builtin_sve_svcmla_lane_f16(__VA_ARGS__) -#define svcmpeq_n_f64(...) __builtin_sve_svcmpeq_n_f64(__VA_ARGS__) -#define svcmpeq_n_f32(...) __builtin_sve_svcmpeq_n_f32(__VA_ARGS__) -#define svcmpeq_n_f16(...) __builtin_sve_svcmpeq_n_f16(__VA_ARGS__) -#define svcmpeq_n_u8(...) __builtin_sve_svcmpeq_n_u8(__VA_ARGS__) -#define svcmpeq_n_u32(...) __builtin_sve_svcmpeq_n_u32(__VA_ARGS__) -#define svcmpeq_n_u64(...) __builtin_sve_svcmpeq_n_u64(__VA_ARGS__) -#define svcmpeq_n_u16(...) __builtin_sve_svcmpeq_n_u16(__VA_ARGS__) -#define svcmpeq_n_s8(...) __builtin_sve_svcmpeq_n_s8(__VA_ARGS__) -#define svcmpeq_n_s32(...) __builtin_sve_svcmpeq_n_s32(__VA_ARGS__) -#define svcmpeq_n_s64(...) __builtin_sve_svcmpeq_n_s64(__VA_ARGS__) -#define svcmpeq_n_s16(...) __builtin_sve_svcmpeq_n_s16(__VA_ARGS__) -#define svcmpeq_u8(...) __builtin_sve_svcmpeq_u8(__VA_ARGS__) -#define svcmpeq_u32(...) __builtin_sve_svcmpeq_u32(__VA_ARGS__) -#define svcmpeq_u64(...) __builtin_sve_svcmpeq_u64(__VA_ARGS__) -#define svcmpeq_u16(...) __builtin_sve_svcmpeq_u16(__VA_ARGS__) -#define svcmpeq_s8(...) __builtin_sve_svcmpeq_s8(__VA_ARGS__) -#define svcmpeq_s32(...) __builtin_sve_svcmpeq_s32(__VA_ARGS__) -#define svcmpeq_s64(...) __builtin_sve_svcmpeq_s64(__VA_ARGS__) -#define svcmpeq_s16(...) __builtin_sve_svcmpeq_s16(__VA_ARGS__) -#define svcmpeq_f64(...) __builtin_sve_svcmpeq_f64(__VA_ARGS__) -#define svcmpeq_f32(...) __builtin_sve_svcmpeq_f32(__VA_ARGS__) -#define svcmpeq_f16(...) __builtin_sve_svcmpeq_f16(__VA_ARGS__) -#define svcmpeq_wide_n_s8(...) __builtin_sve_svcmpeq_wide_n_s8(__VA_ARGS__) -#define svcmpeq_wide_n_s32(...) __builtin_sve_svcmpeq_wide_n_s32(__VA_ARGS__) -#define svcmpeq_wide_n_s16(...) __builtin_sve_svcmpeq_wide_n_s16(__VA_ARGS__) -#define svcmpeq_wide_s8(...) __builtin_sve_svcmpeq_wide_s8(__VA_ARGS__) -#define svcmpeq_wide_s32(...) __builtin_sve_svcmpeq_wide_s32(__VA_ARGS__) -#define svcmpeq_wide_s16(...) __builtin_sve_svcmpeq_wide_s16(__VA_ARGS__) -#define svcmpge_n_f64(...) __builtin_sve_svcmpge_n_f64(__VA_ARGS__) -#define svcmpge_n_f32(...) __builtin_sve_svcmpge_n_f32(__VA_ARGS__) -#define svcmpge_n_f16(...) __builtin_sve_svcmpge_n_f16(__VA_ARGS__) -#define svcmpge_n_s8(...) __builtin_sve_svcmpge_n_s8(__VA_ARGS__) -#define svcmpge_n_s32(...) __builtin_sve_svcmpge_n_s32(__VA_ARGS__) -#define svcmpge_n_s64(...) __builtin_sve_svcmpge_n_s64(__VA_ARGS__) -#define svcmpge_n_s16(...) __builtin_sve_svcmpge_n_s16(__VA_ARGS__) -#define svcmpge_n_u8(...) __builtin_sve_svcmpge_n_u8(__VA_ARGS__) -#define svcmpge_n_u32(...) __builtin_sve_svcmpge_n_u32(__VA_ARGS__) -#define svcmpge_n_u64(...) __builtin_sve_svcmpge_n_u64(__VA_ARGS__) -#define svcmpge_n_u16(...) __builtin_sve_svcmpge_n_u16(__VA_ARGS__) -#define svcmpge_s8(...) __builtin_sve_svcmpge_s8(__VA_ARGS__) -#define svcmpge_s32(...) __builtin_sve_svcmpge_s32(__VA_ARGS__) -#define svcmpge_s64(...) __builtin_sve_svcmpge_s64(__VA_ARGS__) -#define svcmpge_s16(...) __builtin_sve_svcmpge_s16(__VA_ARGS__) -#define svcmpge_f64(...) __builtin_sve_svcmpge_f64(__VA_ARGS__) -#define svcmpge_f32(...) __builtin_sve_svcmpge_f32(__VA_ARGS__) -#define svcmpge_f16(...) __builtin_sve_svcmpge_f16(__VA_ARGS__) -#define svcmpge_u8(...) __builtin_sve_svcmpge_u8(__VA_ARGS__) -#define svcmpge_u32(...) __builtin_sve_svcmpge_u32(__VA_ARGS__) -#define svcmpge_u64(...) __builtin_sve_svcmpge_u64(__VA_ARGS__) -#define svcmpge_u16(...) __builtin_sve_svcmpge_u16(__VA_ARGS__) -#define svcmpge_wide_n_s8(...) __builtin_sve_svcmpge_wide_n_s8(__VA_ARGS__) -#define svcmpge_wide_n_s32(...) __builtin_sve_svcmpge_wide_n_s32(__VA_ARGS__) -#define svcmpge_wide_n_s16(...) __builtin_sve_svcmpge_wide_n_s16(__VA_ARGS__) -#define svcmpge_wide_n_u8(...) __builtin_sve_svcmpge_wide_n_u8(__VA_ARGS__) -#define svcmpge_wide_n_u32(...) __builtin_sve_svcmpge_wide_n_u32(__VA_ARGS__) -#define svcmpge_wide_n_u16(...) __builtin_sve_svcmpge_wide_n_u16(__VA_ARGS__) -#define svcmpge_wide_s8(...) __builtin_sve_svcmpge_wide_s8(__VA_ARGS__) -#define svcmpge_wide_s32(...) __builtin_sve_svcmpge_wide_s32(__VA_ARGS__) -#define svcmpge_wide_s16(...) __builtin_sve_svcmpge_wide_s16(__VA_ARGS__) -#define svcmpge_wide_u8(...) __builtin_sve_svcmpge_wide_u8(__VA_ARGS__) -#define svcmpge_wide_u32(...) __builtin_sve_svcmpge_wide_u32(__VA_ARGS__) -#define svcmpge_wide_u16(...) __builtin_sve_svcmpge_wide_u16(__VA_ARGS__) -#define svcmpgt_n_f64(...) __builtin_sve_svcmpgt_n_f64(__VA_ARGS__) -#define svcmpgt_n_f32(...) __builtin_sve_svcmpgt_n_f32(__VA_ARGS__) -#define svcmpgt_n_f16(...) __builtin_sve_svcmpgt_n_f16(__VA_ARGS__) -#define svcmpgt_n_s8(...) __builtin_sve_svcmpgt_n_s8(__VA_ARGS__) -#define svcmpgt_n_s32(...) __builtin_sve_svcmpgt_n_s32(__VA_ARGS__) -#define svcmpgt_n_s64(...) __builtin_sve_svcmpgt_n_s64(__VA_ARGS__) -#define svcmpgt_n_s16(...) __builtin_sve_svcmpgt_n_s16(__VA_ARGS__) -#define svcmpgt_n_u8(...) __builtin_sve_svcmpgt_n_u8(__VA_ARGS__) -#define svcmpgt_n_u32(...) __builtin_sve_svcmpgt_n_u32(__VA_ARGS__) -#define svcmpgt_n_u64(...) __builtin_sve_svcmpgt_n_u64(__VA_ARGS__) -#define svcmpgt_n_u16(...) __builtin_sve_svcmpgt_n_u16(__VA_ARGS__) -#define svcmpgt_s8(...) __builtin_sve_svcmpgt_s8(__VA_ARGS__) -#define svcmpgt_s32(...) __builtin_sve_svcmpgt_s32(__VA_ARGS__) -#define svcmpgt_s64(...) __builtin_sve_svcmpgt_s64(__VA_ARGS__) -#define svcmpgt_s16(...) __builtin_sve_svcmpgt_s16(__VA_ARGS__) -#define svcmpgt_f64(...) __builtin_sve_svcmpgt_f64(__VA_ARGS__) -#define svcmpgt_f32(...) __builtin_sve_svcmpgt_f32(__VA_ARGS__) -#define svcmpgt_f16(...) __builtin_sve_svcmpgt_f16(__VA_ARGS__) -#define svcmpgt_u8(...) __builtin_sve_svcmpgt_u8(__VA_ARGS__) -#define svcmpgt_u32(...) __builtin_sve_svcmpgt_u32(__VA_ARGS__) -#define svcmpgt_u64(...) __builtin_sve_svcmpgt_u64(__VA_ARGS__) -#define svcmpgt_u16(...) __builtin_sve_svcmpgt_u16(__VA_ARGS__) -#define svcmpgt_wide_n_s8(...) __builtin_sve_svcmpgt_wide_n_s8(__VA_ARGS__) -#define svcmpgt_wide_n_s32(...) __builtin_sve_svcmpgt_wide_n_s32(__VA_ARGS__) -#define svcmpgt_wide_n_s16(...) __builtin_sve_svcmpgt_wide_n_s16(__VA_ARGS__) -#define svcmpgt_wide_n_u8(...) __builtin_sve_svcmpgt_wide_n_u8(__VA_ARGS__) -#define svcmpgt_wide_n_u32(...) __builtin_sve_svcmpgt_wide_n_u32(__VA_ARGS__) -#define svcmpgt_wide_n_u16(...) __builtin_sve_svcmpgt_wide_n_u16(__VA_ARGS__) -#define svcmpgt_wide_s8(...) __builtin_sve_svcmpgt_wide_s8(__VA_ARGS__) -#define svcmpgt_wide_s32(...) __builtin_sve_svcmpgt_wide_s32(__VA_ARGS__) -#define svcmpgt_wide_s16(...) __builtin_sve_svcmpgt_wide_s16(__VA_ARGS__) -#define svcmpgt_wide_u8(...) __builtin_sve_svcmpgt_wide_u8(__VA_ARGS__) -#define svcmpgt_wide_u32(...) __builtin_sve_svcmpgt_wide_u32(__VA_ARGS__) -#define svcmpgt_wide_u16(...) __builtin_sve_svcmpgt_wide_u16(__VA_ARGS__) -#define svcmple_n_f64(...) __builtin_sve_svcmple_n_f64(__VA_ARGS__) -#define svcmple_n_f32(...) __builtin_sve_svcmple_n_f32(__VA_ARGS__) -#define svcmple_n_f16(...) __builtin_sve_svcmple_n_f16(__VA_ARGS__) -#define svcmple_n_s8(...) __builtin_sve_svcmple_n_s8(__VA_ARGS__) -#define svcmple_n_s32(...) __builtin_sve_svcmple_n_s32(__VA_ARGS__) -#define svcmple_n_s64(...) __builtin_sve_svcmple_n_s64(__VA_ARGS__) -#define svcmple_n_s16(...) __builtin_sve_svcmple_n_s16(__VA_ARGS__) -#define svcmple_n_u8(...) __builtin_sve_svcmple_n_u8(__VA_ARGS__) -#define svcmple_n_u32(...) __builtin_sve_svcmple_n_u32(__VA_ARGS__) -#define svcmple_n_u64(...) __builtin_sve_svcmple_n_u64(__VA_ARGS__) -#define svcmple_n_u16(...) __builtin_sve_svcmple_n_u16(__VA_ARGS__) -#define svcmple_s8(...) __builtin_sve_svcmple_s8(__VA_ARGS__) -#define svcmple_s32(...) __builtin_sve_svcmple_s32(__VA_ARGS__) -#define svcmple_s64(...) __builtin_sve_svcmple_s64(__VA_ARGS__) -#define svcmple_s16(...) __builtin_sve_svcmple_s16(__VA_ARGS__) -#define svcmple_f64(...) __builtin_sve_svcmple_f64(__VA_ARGS__) -#define svcmple_f32(...) __builtin_sve_svcmple_f32(__VA_ARGS__) -#define svcmple_f16(...) __builtin_sve_svcmple_f16(__VA_ARGS__) -#define svcmple_u8(...) __builtin_sve_svcmple_u8(__VA_ARGS__) -#define svcmple_u32(...) __builtin_sve_svcmple_u32(__VA_ARGS__) -#define svcmple_u64(...) __builtin_sve_svcmple_u64(__VA_ARGS__) -#define svcmple_u16(...) __builtin_sve_svcmple_u16(__VA_ARGS__) -#define svcmple_wide_n_s8(...) __builtin_sve_svcmple_wide_n_s8(__VA_ARGS__) -#define svcmple_wide_n_s32(...) __builtin_sve_svcmple_wide_n_s32(__VA_ARGS__) -#define svcmple_wide_n_s16(...) __builtin_sve_svcmple_wide_n_s16(__VA_ARGS__) -#define svcmple_wide_n_u8(...) __builtin_sve_svcmple_wide_n_u8(__VA_ARGS__) -#define svcmple_wide_n_u32(...) __builtin_sve_svcmple_wide_n_u32(__VA_ARGS__) -#define svcmple_wide_n_u16(...) __builtin_sve_svcmple_wide_n_u16(__VA_ARGS__) -#define svcmple_wide_s8(...) __builtin_sve_svcmple_wide_s8(__VA_ARGS__) -#define svcmple_wide_s32(...) __builtin_sve_svcmple_wide_s32(__VA_ARGS__) -#define svcmple_wide_s16(...) __builtin_sve_svcmple_wide_s16(__VA_ARGS__) -#define svcmple_wide_u8(...) __builtin_sve_svcmple_wide_u8(__VA_ARGS__) -#define svcmple_wide_u32(...) __builtin_sve_svcmple_wide_u32(__VA_ARGS__) -#define svcmple_wide_u16(...) __builtin_sve_svcmple_wide_u16(__VA_ARGS__) -#define svcmplt_n_u8(...) __builtin_sve_svcmplt_n_u8(__VA_ARGS__) -#define svcmplt_n_u32(...) __builtin_sve_svcmplt_n_u32(__VA_ARGS__) -#define svcmplt_n_u64(...) __builtin_sve_svcmplt_n_u64(__VA_ARGS__) -#define svcmplt_n_u16(...) __builtin_sve_svcmplt_n_u16(__VA_ARGS__) -#define svcmplt_n_f64(...) __builtin_sve_svcmplt_n_f64(__VA_ARGS__) -#define svcmplt_n_f32(...) __builtin_sve_svcmplt_n_f32(__VA_ARGS__) -#define svcmplt_n_f16(...) __builtin_sve_svcmplt_n_f16(__VA_ARGS__) -#define svcmplt_n_s8(...) __builtin_sve_svcmplt_n_s8(__VA_ARGS__) -#define svcmplt_n_s32(...) __builtin_sve_svcmplt_n_s32(__VA_ARGS__) -#define svcmplt_n_s64(...) __builtin_sve_svcmplt_n_s64(__VA_ARGS__) -#define svcmplt_n_s16(...) __builtin_sve_svcmplt_n_s16(__VA_ARGS__) -#define svcmplt_u8(...) __builtin_sve_svcmplt_u8(__VA_ARGS__) -#define svcmplt_u32(...) __builtin_sve_svcmplt_u32(__VA_ARGS__) -#define svcmplt_u64(...) __builtin_sve_svcmplt_u64(__VA_ARGS__) -#define svcmplt_u16(...) __builtin_sve_svcmplt_u16(__VA_ARGS__) -#define svcmplt_s8(...) __builtin_sve_svcmplt_s8(__VA_ARGS__) -#define svcmplt_s32(...) __builtin_sve_svcmplt_s32(__VA_ARGS__) -#define svcmplt_s64(...) __builtin_sve_svcmplt_s64(__VA_ARGS__) -#define svcmplt_s16(...) __builtin_sve_svcmplt_s16(__VA_ARGS__) -#define svcmplt_f64(...) __builtin_sve_svcmplt_f64(__VA_ARGS__) -#define svcmplt_f32(...) __builtin_sve_svcmplt_f32(__VA_ARGS__) -#define svcmplt_f16(...) __builtin_sve_svcmplt_f16(__VA_ARGS__) -#define svcmplt_wide_n_u8(...) __builtin_sve_svcmplt_wide_n_u8(__VA_ARGS__) -#define svcmplt_wide_n_u32(...) __builtin_sve_svcmplt_wide_n_u32(__VA_ARGS__) -#define svcmplt_wide_n_u16(...) __builtin_sve_svcmplt_wide_n_u16(__VA_ARGS__) -#define svcmplt_wide_n_s8(...) __builtin_sve_svcmplt_wide_n_s8(__VA_ARGS__) -#define svcmplt_wide_n_s32(...) __builtin_sve_svcmplt_wide_n_s32(__VA_ARGS__) -#define svcmplt_wide_n_s16(...) __builtin_sve_svcmplt_wide_n_s16(__VA_ARGS__) -#define svcmplt_wide_u8(...) __builtin_sve_svcmplt_wide_u8(__VA_ARGS__) -#define svcmplt_wide_u32(...) __builtin_sve_svcmplt_wide_u32(__VA_ARGS__) -#define svcmplt_wide_u16(...) __builtin_sve_svcmplt_wide_u16(__VA_ARGS__) -#define svcmplt_wide_s8(...) __builtin_sve_svcmplt_wide_s8(__VA_ARGS__) -#define svcmplt_wide_s32(...) __builtin_sve_svcmplt_wide_s32(__VA_ARGS__) -#define svcmplt_wide_s16(...) __builtin_sve_svcmplt_wide_s16(__VA_ARGS__) -#define svcmpne_n_f64(...) __builtin_sve_svcmpne_n_f64(__VA_ARGS__) -#define svcmpne_n_f32(...) __builtin_sve_svcmpne_n_f32(__VA_ARGS__) -#define svcmpne_n_f16(...) __builtin_sve_svcmpne_n_f16(__VA_ARGS__) -#define svcmpne_n_u8(...) __builtin_sve_svcmpne_n_u8(__VA_ARGS__) -#define svcmpne_n_u32(...) __builtin_sve_svcmpne_n_u32(__VA_ARGS__) -#define svcmpne_n_u64(...) __builtin_sve_svcmpne_n_u64(__VA_ARGS__) -#define svcmpne_n_u16(...) __builtin_sve_svcmpne_n_u16(__VA_ARGS__) -#define svcmpne_n_s8(...) __builtin_sve_svcmpne_n_s8(__VA_ARGS__) -#define svcmpne_n_s32(...) __builtin_sve_svcmpne_n_s32(__VA_ARGS__) -#define svcmpne_n_s64(...) __builtin_sve_svcmpne_n_s64(__VA_ARGS__) -#define svcmpne_n_s16(...) __builtin_sve_svcmpne_n_s16(__VA_ARGS__) -#define svcmpne_u8(...) __builtin_sve_svcmpne_u8(__VA_ARGS__) -#define svcmpne_u32(...) __builtin_sve_svcmpne_u32(__VA_ARGS__) -#define svcmpne_u64(...) __builtin_sve_svcmpne_u64(__VA_ARGS__) -#define svcmpne_u16(...) __builtin_sve_svcmpne_u16(__VA_ARGS__) -#define svcmpne_s8(...) __builtin_sve_svcmpne_s8(__VA_ARGS__) -#define svcmpne_s32(...) __builtin_sve_svcmpne_s32(__VA_ARGS__) -#define svcmpne_s64(...) __builtin_sve_svcmpne_s64(__VA_ARGS__) -#define svcmpne_s16(...) __builtin_sve_svcmpne_s16(__VA_ARGS__) -#define svcmpne_f64(...) __builtin_sve_svcmpne_f64(__VA_ARGS__) -#define svcmpne_f32(...) __builtin_sve_svcmpne_f32(__VA_ARGS__) -#define svcmpne_f16(...) __builtin_sve_svcmpne_f16(__VA_ARGS__) -#define svcmpne_wide_n_s8(...) __builtin_sve_svcmpne_wide_n_s8(__VA_ARGS__) -#define svcmpne_wide_n_s32(...) __builtin_sve_svcmpne_wide_n_s32(__VA_ARGS__) -#define svcmpne_wide_n_s16(...) __builtin_sve_svcmpne_wide_n_s16(__VA_ARGS__) -#define svcmpne_wide_s8(...) __builtin_sve_svcmpne_wide_s8(__VA_ARGS__) -#define svcmpne_wide_s32(...) __builtin_sve_svcmpne_wide_s32(__VA_ARGS__) -#define svcmpne_wide_s16(...) __builtin_sve_svcmpne_wide_s16(__VA_ARGS__) -#define svcmpuo_n_f64(...) __builtin_sve_svcmpuo_n_f64(__VA_ARGS__) -#define svcmpuo_n_f32(...) __builtin_sve_svcmpuo_n_f32(__VA_ARGS__) -#define svcmpuo_n_f16(...) __builtin_sve_svcmpuo_n_f16(__VA_ARGS__) -#define svcmpuo_f64(...) __builtin_sve_svcmpuo_f64(__VA_ARGS__) -#define svcmpuo_f32(...) __builtin_sve_svcmpuo_f32(__VA_ARGS__) -#define svcmpuo_f16(...) __builtin_sve_svcmpuo_f16(__VA_ARGS__) -#define svcnot_u8_m(...) __builtin_sve_svcnot_u8_m(__VA_ARGS__) -#define svcnot_u32_m(...) __builtin_sve_svcnot_u32_m(__VA_ARGS__) -#define svcnot_u64_m(...) __builtin_sve_svcnot_u64_m(__VA_ARGS__) -#define svcnot_u16_m(...) __builtin_sve_svcnot_u16_m(__VA_ARGS__) -#define svcnot_s8_m(...) __builtin_sve_svcnot_s8_m(__VA_ARGS__) -#define svcnot_s32_m(...) __builtin_sve_svcnot_s32_m(__VA_ARGS__) -#define svcnot_s64_m(...) __builtin_sve_svcnot_s64_m(__VA_ARGS__) -#define svcnot_s16_m(...) __builtin_sve_svcnot_s16_m(__VA_ARGS__) -#define svcnot_u8_x(...) __builtin_sve_svcnot_u8_x(__VA_ARGS__) -#define svcnot_u32_x(...) __builtin_sve_svcnot_u32_x(__VA_ARGS__) -#define svcnot_u64_x(...) __builtin_sve_svcnot_u64_x(__VA_ARGS__) -#define svcnot_u16_x(...) __builtin_sve_svcnot_u16_x(__VA_ARGS__) -#define svcnot_s8_x(...) __builtin_sve_svcnot_s8_x(__VA_ARGS__) -#define svcnot_s32_x(...) __builtin_sve_svcnot_s32_x(__VA_ARGS__) -#define svcnot_s64_x(...) __builtin_sve_svcnot_s64_x(__VA_ARGS__) -#define svcnot_s16_x(...) __builtin_sve_svcnot_s16_x(__VA_ARGS__) -#define svcnot_u8_z(...) __builtin_sve_svcnot_u8_z(__VA_ARGS__) -#define svcnot_u32_z(...) __builtin_sve_svcnot_u32_z(__VA_ARGS__) -#define svcnot_u64_z(...) __builtin_sve_svcnot_u64_z(__VA_ARGS__) -#define svcnot_u16_z(...) __builtin_sve_svcnot_u16_z(__VA_ARGS__) -#define svcnot_s8_z(...) __builtin_sve_svcnot_s8_z(__VA_ARGS__) -#define svcnot_s32_z(...) __builtin_sve_svcnot_s32_z(__VA_ARGS__) -#define svcnot_s64_z(...) __builtin_sve_svcnot_s64_z(__VA_ARGS__) -#define svcnot_s16_z(...) __builtin_sve_svcnot_s16_z(__VA_ARGS__) -#define svcnt_u8_m(...) __builtin_sve_svcnt_u8_m(__VA_ARGS__) -#define svcnt_u32_m(...) __builtin_sve_svcnt_u32_m(__VA_ARGS__) -#define svcnt_u64_m(...) __builtin_sve_svcnt_u64_m(__VA_ARGS__) -#define svcnt_u16_m(...) __builtin_sve_svcnt_u16_m(__VA_ARGS__) -#define svcnt_s8_m(...) __builtin_sve_svcnt_s8_m(__VA_ARGS__) -#define svcnt_f64_m(...) __builtin_sve_svcnt_f64_m(__VA_ARGS__) -#define svcnt_f32_m(...) __builtin_sve_svcnt_f32_m(__VA_ARGS__) -#define svcnt_f16_m(...) __builtin_sve_svcnt_f16_m(__VA_ARGS__) -#define svcnt_s32_m(...) __builtin_sve_svcnt_s32_m(__VA_ARGS__) -#define svcnt_s64_m(...) __builtin_sve_svcnt_s64_m(__VA_ARGS__) -#define svcnt_s16_m(...) __builtin_sve_svcnt_s16_m(__VA_ARGS__) -#define svcnt_u8_x(...) __builtin_sve_svcnt_u8_x(__VA_ARGS__) -#define svcnt_u32_x(...) __builtin_sve_svcnt_u32_x(__VA_ARGS__) -#define svcnt_u64_x(...) __builtin_sve_svcnt_u64_x(__VA_ARGS__) -#define svcnt_u16_x(...) __builtin_sve_svcnt_u16_x(__VA_ARGS__) -#define svcnt_s8_x(...) __builtin_sve_svcnt_s8_x(__VA_ARGS__) -#define svcnt_f64_x(...) __builtin_sve_svcnt_f64_x(__VA_ARGS__) -#define svcnt_f32_x(...) __builtin_sve_svcnt_f32_x(__VA_ARGS__) -#define svcnt_f16_x(...) __builtin_sve_svcnt_f16_x(__VA_ARGS__) -#define svcnt_s32_x(...) __builtin_sve_svcnt_s32_x(__VA_ARGS__) -#define svcnt_s64_x(...) __builtin_sve_svcnt_s64_x(__VA_ARGS__) -#define svcnt_s16_x(...) __builtin_sve_svcnt_s16_x(__VA_ARGS__) -#define svcnt_u8_z(...) __builtin_sve_svcnt_u8_z(__VA_ARGS__) -#define svcnt_u32_z(...) __builtin_sve_svcnt_u32_z(__VA_ARGS__) -#define svcnt_u64_z(...) __builtin_sve_svcnt_u64_z(__VA_ARGS__) -#define svcnt_u16_z(...) __builtin_sve_svcnt_u16_z(__VA_ARGS__) -#define svcnt_s8_z(...) __builtin_sve_svcnt_s8_z(__VA_ARGS__) -#define svcnt_f64_z(...) __builtin_sve_svcnt_f64_z(__VA_ARGS__) -#define svcnt_f32_z(...) __builtin_sve_svcnt_f32_z(__VA_ARGS__) -#define svcnt_f16_z(...) __builtin_sve_svcnt_f16_z(__VA_ARGS__) -#define svcnt_s32_z(...) __builtin_sve_svcnt_s32_z(__VA_ARGS__) -#define svcnt_s64_z(...) __builtin_sve_svcnt_s64_z(__VA_ARGS__) -#define svcnt_s16_z(...) __builtin_sve_svcnt_s16_z(__VA_ARGS__) -#define svcntb(...) __builtin_sve_svcntb(__VA_ARGS__) -#define svcntb_pat(...) __builtin_sve_svcntb_pat(__VA_ARGS__) -#define svcntd(...) __builtin_sve_svcntd(__VA_ARGS__) -#define svcntd_pat(...) __builtin_sve_svcntd_pat(__VA_ARGS__) -#define svcnth(...) __builtin_sve_svcnth(__VA_ARGS__) -#define svcnth_pat(...) __builtin_sve_svcnth_pat(__VA_ARGS__) -#define svcntp_b8(...) __builtin_sve_svcntp_b8(__VA_ARGS__) -#define svcntp_b32(...) __builtin_sve_svcntp_b32(__VA_ARGS__) -#define svcntp_b64(...) __builtin_sve_svcntp_b64(__VA_ARGS__) -#define svcntp_b16(...) __builtin_sve_svcntp_b16(__VA_ARGS__) -#define svcntw(...) __builtin_sve_svcntw(__VA_ARGS__) -#define svcntw_pat(...) __builtin_sve_svcntw_pat(__VA_ARGS__) -#define svcompact_u32(...) __builtin_sve_svcompact_u32(__VA_ARGS__) -#define svcompact_u64(...) __builtin_sve_svcompact_u64(__VA_ARGS__) -#define svcompact_f64(...) __builtin_sve_svcompact_f64(__VA_ARGS__) -#define svcompact_f32(...) __builtin_sve_svcompact_f32(__VA_ARGS__) -#define svcompact_s32(...) __builtin_sve_svcompact_s32(__VA_ARGS__) -#define svcompact_s64(...) __builtin_sve_svcompact_s64(__VA_ARGS__) -#define svcreate2_u8(...) __builtin_sve_svcreate2_u8(__VA_ARGS__) -#define svcreate2_u32(...) __builtin_sve_svcreate2_u32(__VA_ARGS__) -#define svcreate2_u64(...) __builtin_sve_svcreate2_u64(__VA_ARGS__) -#define svcreate2_u16(...) __builtin_sve_svcreate2_u16(__VA_ARGS__) -#define svcreate2_s8(...) __builtin_sve_svcreate2_s8(__VA_ARGS__) -#define svcreate2_f64(...) __builtin_sve_svcreate2_f64(__VA_ARGS__) -#define svcreate2_f32(...) __builtin_sve_svcreate2_f32(__VA_ARGS__) -#define svcreate2_f16(...) __builtin_sve_svcreate2_f16(__VA_ARGS__) -#define svcreate2_s32(...) __builtin_sve_svcreate2_s32(__VA_ARGS__) -#define svcreate2_s64(...) __builtin_sve_svcreate2_s64(__VA_ARGS__) -#define svcreate2_s16(...) __builtin_sve_svcreate2_s16(__VA_ARGS__) -#define svcreate3_u8(...) __builtin_sve_svcreate3_u8(__VA_ARGS__) -#define svcreate3_u32(...) __builtin_sve_svcreate3_u32(__VA_ARGS__) -#define svcreate3_u64(...) __builtin_sve_svcreate3_u64(__VA_ARGS__) -#define svcreate3_u16(...) __builtin_sve_svcreate3_u16(__VA_ARGS__) -#define svcreate3_s8(...) __builtin_sve_svcreate3_s8(__VA_ARGS__) -#define svcreate3_f64(...) __builtin_sve_svcreate3_f64(__VA_ARGS__) -#define svcreate3_f32(...) __builtin_sve_svcreate3_f32(__VA_ARGS__) -#define svcreate3_f16(...) __builtin_sve_svcreate3_f16(__VA_ARGS__) -#define svcreate3_s32(...) __builtin_sve_svcreate3_s32(__VA_ARGS__) -#define svcreate3_s64(...) __builtin_sve_svcreate3_s64(__VA_ARGS__) -#define svcreate3_s16(...) __builtin_sve_svcreate3_s16(__VA_ARGS__) -#define svcreate4_u8(...) __builtin_sve_svcreate4_u8(__VA_ARGS__) -#define svcreate4_u32(...) __builtin_sve_svcreate4_u32(__VA_ARGS__) -#define svcreate4_u64(...) __builtin_sve_svcreate4_u64(__VA_ARGS__) -#define svcreate4_u16(...) __builtin_sve_svcreate4_u16(__VA_ARGS__) -#define svcreate4_s8(...) __builtin_sve_svcreate4_s8(__VA_ARGS__) -#define svcreate4_f64(...) __builtin_sve_svcreate4_f64(__VA_ARGS__) -#define svcreate4_f32(...) __builtin_sve_svcreate4_f32(__VA_ARGS__) -#define svcreate4_f16(...) __builtin_sve_svcreate4_f16(__VA_ARGS__) -#define svcreate4_s32(...) __builtin_sve_svcreate4_s32(__VA_ARGS__) -#define svcreate4_s64(...) __builtin_sve_svcreate4_s64(__VA_ARGS__) -#define svcreate4_s16(...) __builtin_sve_svcreate4_s16(__VA_ARGS__) -#define svcvt_f16_f32_m(...) __builtin_sve_svcvt_f16_f32_m(__VA_ARGS__) -#define svcvt_f16_f32_x(...) __builtin_sve_svcvt_f16_f32_x(__VA_ARGS__) -#define svcvt_f16_f32_z(...) __builtin_sve_svcvt_f16_f32_z(__VA_ARGS__) -#define svcvt_f16_f64_m(...) __builtin_sve_svcvt_f16_f64_m(__VA_ARGS__) -#define svcvt_f16_f64_x(...) __builtin_sve_svcvt_f16_f64_x(__VA_ARGS__) -#define svcvt_f16_f64_z(...) __builtin_sve_svcvt_f16_f64_z(__VA_ARGS__) -#define svcvt_f16_s16_m(...) __builtin_sve_svcvt_f16_s16_m(__VA_ARGS__) -#define svcvt_f16_s16_x(...) __builtin_sve_svcvt_f16_s16_x(__VA_ARGS__) -#define svcvt_f16_s16_z(...) __builtin_sve_svcvt_f16_s16_z(__VA_ARGS__) -#define svcvt_f16_s32_m(...) __builtin_sve_svcvt_f16_s32_m(__VA_ARGS__) -#define svcvt_f16_s32_x(...) __builtin_sve_svcvt_f16_s32_x(__VA_ARGS__) -#define svcvt_f16_s32_z(...) __builtin_sve_svcvt_f16_s32_z(__VA_ARGS__) -#define svcvt_f16_s64_m(...) __builtin_sve_svcvt_f16_s64_m(__VA_ARGS__) -#define svcvt_f16_s64_x(...) __builtin_sve_svcvt_f16_s64_x(__VA_ARGS__) -#define svcvt_f16_s64_z(...) __builtin_sve_svcvt_f16_s64_z(__VA_ARGS__) -#define svcvt_f16_u16_m(...) __builtin_sve_svcvt_f16_u16_m(__VA_ARGS__) -#define svcvt_f16_u16_x(...) __builtin_sve_svcvt_f16_u16_x(__VA_ARGS__) -#define svcvt_f16_u16_z(...) __builtin_sve_svcvt_f16_u16_z(__VA_ARGS__) -#define svcvt_f16_u32_m(...) __builtin_sve_svcvt_f16_u32_m(__VA_ARGS__) -#define svcvt_f16_u32_x(...) __builtin_sve_svcvt_f16_u32_x(__VA_ARGS__) -#define svcvt_f16_u32_z(...) __builtin_sve_svcvt_f16_u32_z(__VA_ARGS__) -#define svcvt_f16_u64_m(...) __builtin_sve_svcvt_f16_u64_m(__VA_ARGS__) -#define svcvt_f16_u64_x(...) __builtin_sve_svcvt_f16_u64_x(__VA_ARGS__) -#define svcvt_f16_u64_z(...) __builtin_sve_svcvt_f16_u64_z(__VA_ARGS__) -#define svcvt_f32_f16_m(...) __builtin_sve_svcvt_f32_f16_m(__VA_ARGS__) -#define svcvt_f32_f16_x(...) __builtin_sve_svcvt_f32_f16_x(__VA_ARGS__) -#define svcvt_f32_f16_z(...) __builtin_sve_svcvt_f32_f16_z(__VA_ARGS__) -#define svcvt_f32_f64_m(...) __builtin_sve_svcvt_f32_f64_m(__VA_ARGS__) -#define svcvt_f32_f64_x(...) __builtin_sve_svcvt_f32_f64_x(__VA_ARGS__) -#define svcvt_f32_f64_z(...) __builtin_sve_svcvt_f32_f64_z(__VA_ARGS__) -#define svcvt_f32_s32_m(...) __builtin_sve_svcvt_f32_s32_m(__VA_ARGS__) -#define svcvt_f32_s32_x(...) __builtin_sve_svcvt_f32_s32_x(__VA_ARGS__) -#define svcvt_f32_s32_z(...) __builtin_sve_svcvt_f32_s32_z(__VA_ARGS__) -#define svcvt_f32_s64_m(...) __builtin_sve_svcvt_f32_s64_m(__VA_ARGS__) -#define svcvt_f32_s64_x(...) __builtin_sve_svcvt_f32_s64_x(__VA_ARGS__) -#define svcvt_f32_s64_z(...) __builtin_sve_svcvt_f32_s64_z(__VA_ARGS__) -#define svcvt_f32_u32_m(...) __builtin_sve_svcvt_f32_u32_m(__VA_ARGS__) -#define svcvt_f32_u32_x(...) __builtin_sve_svcvt_f32_u32_x(__VA_ARGS__) -#define svcvt_f32_u32_z(...) __builtin_sve_svcvt_f32_u32_z(__VA_ARGS__) -#define svcvt_f32_u64_m(...) __builtin_sve_svcvt_f32_u64_m(__VA_ARGS__) -#define svcvt_f32_u64_x(...) __builtin_sve_svcvt_f32_u64_x(__VA_ARGS__) -#define svcvt_f32_u64_z(...) __builtin_sve_svcvt_f32_u64_z(__VA_ARGS__) -#define svcvt_f64_f16_m(...) __builtin_sve_svcvt_f64_f16_m(__VA_ARGS__) -#define svcvt_f64_f16_x(...) __builtin_sve_svcvt_f64_f16_x(__VA_ARGS__) -#define svcvt_f64_f16_z(...) __builtin_sve_svcvt_f64_f16_z(__VA_ARGS__) -#define svcvt_f64_f32_m(...) __builtin_sve_svcvt_f64_f32_m(__VA_ARGS__) -#define svcvt_f64_f32_x(...) __builtin_sve_svcvt_f64_f32_x(__VA_ARGS__) -#define svcvt_f64_f32_z(...) __builtin_sve_svcvt_f64_f32_z(__VA_ARGS__) -#define svcvt_f64_s32_m(...) __builtin_sve_svcvt_f64_s32_m(__VA_ARGS__) -#define svcvt_f64_s32_x(...) __builtin_sve_svcvt_f64_s32_x(__VA_ARGS__) -#define svcvt_f64_s32_z(...) __builtin_sve_svcvt_f64_s32_z(__VA_ARGS__) -#define svcvt_f64_s64_m(...) __builtin_sve_svcvt_f64_s64_m(__VA_ARGS__) -#define svcvt_f64_s64_x(...) __builtin_sve_svcvt_f64_s64_x(__VA_ARGS__) -#define svcvt_f64_s64_z(...) __builtin_sve_svcvt_f64_s64_z(__VA_ARGS__) -#define svcvt_f64_u32_m(...) __builtin_sve_svcvt_f64_u32_m(__VA_ARGS__) -#define svcvt_f64_u32_x(...) __builtin_sve_svcvt_f64_u32_x(__VA_ARGS__) -#define svcvt_f64_u32_z(...) __builtin_sve_svcvt_f64_u32_z(__VA_ARGS__) -#define svcvt_f64_u64_m(...) __builtin_sve_svcvt_f64_u64_m(__VA_ARGS__) -#define svcvt_f64_u64_x(...) __builtin_sve_svcvt_f64_u64_x(__VA_ARGS__) -#define svcvt_f64_u64_z(...) __builtin_sve_svcvt_f64_u64_z(__VA_ARGS__) -#define svcvt_s16_f16_m(...) __builtin_sve_svcvt_s16_f16_m(__VA_ARGS__) -#define svcvt_s16_f16_x(...) __builtin_sve_svcvt_s16_f16_x(__VA_ARGS__) -#define svcvt_s16_f16_z(...) __builtin_sve_svcvt_s16_f16_z(__VA_ARGS__) -#define svcvt_s32_f16_m(...) __builtin_sve_svcvt_s32_f16_m(__VA_ARGS__) -#define svcvt_s32_f16_x(...) __builtin_sve_svcvt_s32_f16_x(__VA_ARGS__) -#define svcvt_s32_f16_z(...) __builtin_sve_svcvt_s32_f16_z(__VA_ARGS__) -#define svcvt_s32_f32_m(...) __builtin_sve_svcvt_s32_f32_m(__VA_ARGS__) -#define svcvt_s32_f32_x(...) __builtin_sve_svcvt_s32_f32_x(__VA_ARGS__) -#define svcvt_s32_f32_z(...) __builtin_sve_svcvt_s32_f32_z(__VA_ARGS__) -#define svcvt_s32_f64_m(...) __builtin_sve_svcvt_s32_f64_m(__VA_ARGS__) -#define svcvt_s32_f64_x(...) __builtin_sve_svcvt_s32_f64_x(__VA_ARGS__) -#define svcvt_s32_f64_z(...) __builtin_sve_svcvt_s32_f64_z(__VA_ARGS__) -#define svcvt_s64_f16_m(...) __builtin_sve_svcvt_s64_f16_m(__VA_ARGS__) -#define svcvt_s64_f16_x(...) __builtin_sve_svcvt_s64_f16_x(__VA_ARGS__) -#define svcvt_s64_f16_z(...) __builtin_sve_svcvt_s64_f16_z(__VA_ARGS__) -#define svcvt_s64_f32_m(...) __builtin_sve_svcvt_s64_f32_m(__VA_ARGS__) -#define svcvt_s64_f32_x(...) __builtin_sve_svcvt_s64_f32_x(__VA_ARGS__) -#define svcvt_s64_f32_z(...) __builtin_sve_svcvt_s64_f32_z(__VA_ARGS__) -#define svcvt_s64_f64_m(...) __builtin_sve_svcvt_s64_f64_m(__VA_ARGS__) -#define svcvt_s64_f64_x(...) __builtin_sve_svcvt_s64_f64_x(__VA_ARGS__) -#define svcvt_s64_f64_z(...) __builtin_sve_svcvt_s64_f64_z(__VA_ARGS__) -#define svcvt_u16_f16_m(...) __builtin_sve_svcvt_u16_f16_m(__VA_ARGS__) -#define svcvt_u16_f16_x(...) __builtin_sve_svcvt_u16_f16_x(__VA_ARGS__) -#define svcvt_u16_f16_z(...) __builtin_sve_svcvt_u16_f16_z(__VA_ARGS__) -#define svcvt_u32_f16_m(...) __builtin_sve_svcvt_u32_f16_m(__VA_ARGS__) -#define svcvt_u32_f16_x(...) __builtin_sve_svcvt_u32_f16_x(__VA_ARGS__) -#define svcvt_u32_f16_z(...) __builtin_sve_svcvt_u32_f16_z(__VA_ARGS__) -#define svcvt_u32_f32_m(...) __builtin_sve_svcvt_u32_f32_m(__VA_ARGS__) -#define svcvt_u32_f32_x(...) __builtin_sve_svcvt_u32_f32_x(__VA_ARGS__) -#define svcvt_u32_f32_z(...) __builtin_sve_svcvt_u32_f32_z(__VA_ARGS__) -#define svcvt_u32_f64_m(...) __builtin_sve_svcvt_u32_f64_m(__VA_ARGS__) -#define svcvt_u32_f64_x(...) __builtin_sve_svcvt_u32_f64_x(__VA_ARGS__) -#define svcvt_u32_f64_z(...) __builtin_sve_svcvt_u32_f64_z(__VA_ARGS__) -#define svcvt_u64_f16_m(...) __builtin_sve_svcvt_u64_f16_m(__VA_ARGS__) -#define svcvt_u64_f16_x(...) __builtin_sve_svcvt_u64_f16_x(__VA_ARGS__) -#define svcvt_u64_f16_z(...) __builtin_sve_svcvt_u64_f16_z(__VA_ARGS__) -#define svcvt_u64_f32_m(...) __builtin_sve_svcvt_u64_f32_m(__VA_ARGS__) -#define svcvt_u64_f32_x(...) __builtin_sve_svcvt_u64_f32_x(__VA_ARGS__) -#define svcvt_u64_f32_z(...) __builtin_sve_svcvt_u64_f32_z(__VA_ARGS__) -#define svcvt_u64_f64_m(...) __builtin_sve_svcvt_u64_f64_m(__VA_ARGS__) -#define svcvt_u64_f64_x(...) __builtin_sve_svcvt_u64_f64_x(__VA_ARGS__) -#define svcvt_u64_f64_z(...) __builtin_sve_svcvt_u64_f64_z(__VA_ARGS__) -#define svdiv_n_f64_m(...) __builtin_sve_svdiv_n_f64_m(__VA_ARGS__) -#define svdiv_n_f32_m(...) __builtin_sve_svdiv_n_f32_m(__VA_ARGS__) -#define svdiv_n_f16_m(...) __builtin_sve_svdiv_n_f16_m(__VA_ARGS__) -#define svdiv_n_f64_x(...) __builtin_sve_svdiv_n_f64_x(__VA_ARGS__) -#define svdiv_n_f32_x(...) __builtin_sve_svdiv_n_f32_x(__VA_ARGS__) -#define svdiv_n_f16_x(...) __builtin_sve_svdiv_n_f16_x(__VA_ARGS__) -#define svdiv_n_f64_z(...) __builtin_sve_svdiv_n_f64_z(__VA_ARGS__) -#define svdiv_n_f32_z(...) __builtin_sve_svdiv_n_f32_z(__VA_ARGS__) -#define svdiv_n_f16_z(...) __builtin_sve_svdiv_n_f16_z(__VA_ARGS__) -#define svdiv_n_s32_m(...) __builtin_sve_svdiv_n_s32_m(__VA_ARGS__) -#define svdiv_n_s64_m(...) __builtin_sve_svdiv_n_s64_m(__VA_ARGS__) -#define svdiv_n_s32_x(...) __builtin_sve_svdiv_n_s32_x(__VA_ARGS__) -#define svdiv_n_s64_x(...) __builtin_sve_svdiv_n_s64_x(__VA_ARGS__) -#define svdiv_n_s32_z(...) __builtin_sve_svdiv_n_s32_z(__VA_ARGS__) -#define svdiv_n_s64_z(...) __builtin_sve_svdiv_n_s64_z(__VA_ARGS__) -#define svdiv_n_u32_m(...) __builtin_sve_svdiv_n_u32_m(__VA_ARGS__) -#define svdiv_n_u64_m(...) __builtin_sve_svdiv_n_u64_m(__VA_ARGS__) -#define svdiv_n_u32_x(...) __builtin_sve_svdiv_n_u32_x(__VA_ARGS__) -#define svdiv_n_u64_x(...) __builtin_sve_svdiv_n_u64_x(__VA_ARGS__) -#define svdiv_n_u32_z(...) __builtin_sve_svdiv_n_u32_z(__VA_ARGS__) -#define svdiv_n_u64_z(...) __builtin_sve_svdiv_n_u64_z(__VA_ARGS__) -#define svdiv_f64_m(...) __builtin_sve_svdiv_f64_m(__VA_ARGS__) -#define svdiv_f32_m(...) __builtin_sve_svdiv_f32_m(__VA_ARGS__) -#define svdiv_f16_m(...) __builtin_sve_svdiv_f16_m(__VA_ARGS__) -#define svdiv_f64_x(...) __builtin_sve_svdiv_f64_x(__VA_ARGS__) -#define svdiv_f32_x(...) __builtin_sve_svdiv_f32_x(__VA_ARGS__) -#define svdiv_f16_x(...) __builtin_sve_svdiv_f16_x(__VA_ARGS__) -#define svdiv_f64_z(...) __builtin_sve_svdiv_f64_z(__VA_ARGS__) -#define svdiv_f32_z(...) __builtin_sve_svdiv_f32_z(__VA_ARGS__) -#define svdiv_f16_z(...) __builtin_sve_svdiv_f16_z(__VA_ARGS__) -#define svdiv_s32_m(...) __builtin_sve_svdiv_s32_m(__VA_ARGS__) -#define svdiv_s64_m(...) __builtin_sve_svdiv_s64_m(__VA_ARGS__) -#define svdiv_s32_x(...) __builtin_sve_svdiv_s32_x(__VA_ARGS__) -#define svdiv_s64_x(...) __builtin_sve_svdiv_s64_x(__VA_ARGS__) -#define svdiv_s32_z(...) __builtin_sve_svdiv_s32_z(__VA_ARGS__) -#define svdiv_s64_z(...) __builtin_sve_svdiv_s64_z(__VA_ARGS__) -#define svdiv_u32_m(...) __builtin_sve_svdiv_u32_m(__VA_ARGS__) -#define svdiv_u64_m(...) __builtin_sve_svdiv_u64_m(__VA_ARGS__) -#define svdiv_u32_x(...) __builtin_sve_svdiv_u32_x(__VA_ARGS__) -#define svdiv_u64_x(...) __builtin_sve_svdiv_u64_x(__VA_ARGS__) -#define svdiv_u32_z(...) __builtin_sve_svdiv_u32_z(__VA_ARGS__) -#define svdiv_u64_z(...) __builtin_sve_svdiv_u64_z(__VA_ARGS__) -#define svdivr_n_f64_m(...) __builtin_sve_svdivr_n_f64_m(__VA_ARGS__) -#define svdivr_n_f32_m(...) __builtin_sve_svdivr_n_f32_m(__VA_ARGS__) -#define svdivr_n_f16_m(...) __builtin_sve_svdivr_n_f16_m(__VA_ARGS__) -#define svdivr_n_f64_x(...) __builtin_sve_svdivr_n_f64_x(__VA_ARGS__) -#define svdivr_n_f32_x(...) __builtin_sve_svdivr_n_f32_x(__VA_ARGS__) -#define svdivr_n_f16_x(...) __builtin_sve_svdivr_n_f16_x(__VA_ARGS__) -#define svdivr_n_f64_z(...) __builtin_sve_svdivr_n_f64_z(__VA_ARGS__) -#define svdivr_n_f32_z(...) __builtin_sve_svdivr_n_f32_z(__VA_ARGS__) -#define svdivr_n_f16_z(...) __builtin_sve_svdivr_n_f16_z(__VA_ARGS__) -#define svdivr_n_s32_m(...) __builtin_sve_svdivr_n_s32_m(__VA_ARGS__) -#define svdivr_n_s64_m(...) __builtin_sve_svdivr_n_s64_m(__VA_ARGS__) -#define svdivr_n_s32_x(...) __builtin_sve_svdivr_n_s32_x(__VA_ARGS__) -#define svdivr_n_s64_x(...) __builtin_sve_svdivr_n_s64_x(__VA_ARGS__) -#define svdivr_n_s32_z(...) __builtin_sve_svdivr_n_s32_z(__VA_ARGS__) -#define svdivr_n_s64_z(...) __builtin_sve_svdivr_n_s64_z(__VA_ARGS__) -#define svdivr_n_u32_m(...) __builtin_sve_svdivr_n_u32_m(__VA_ARGS__) -#define svdivr_n_u64_m(...) __builtin_sve_svdivr_n_u64_m(__VA_ARGS__) -#define svdivr_n_u32_x(...) __builtin_sve_svdivr_n_u32_x(__VA_ARGS__) -#define svdivr_n_u64_x(...) __builtin_sve_svdivr_n_u64_x(__VA_ARGS__) -#define svdivr_n_u32_z(...) __builtin_sve_svdivr_n_u32_z(__VA_ARGS__) -#define svdivr_n_u64_z(...) __builtin_sve_svdivr_n_u64_z(__VA_ARGS__) -#define svdivr_f64_m(...) __builtin_sve_svdivr_f64_m(__VA_ARGS__) -#define svdivr_f32_m(...) __builtin_sve_svdivr_f32_m(__VA_ARGS__) -#define svdivr_f16_m(...) __builtin_sve_svdivr_f16_m(__VA_ARGS__) -#define svdivr_f64_x(...) __builtin_sve_svdivr_f64_x(__VA_ARGS__) -#define svdivr_f32_x(...) __builtin_sve_svdivr_f32_x(__VA_ARGS__) -#define svdivr_f16_x(...) __builtin_sve_svdivr_f16_x(__VA_ARGS__) -#define svdivr_f64_z(...) __builtin_sve_svdivr_f64_z(__VA_ARGS__) -#define svdivr_f32_z(...) __builtin_sve_svdivr_f32_z(__VA_ARGS__) -#define svdivr_f16_z(...) __builtin_sve_svdivr_f16_z(__VA_ARGS__) -#define svdivr_s32_m(...) __builtin_sve_svdivr_s32_m(__VA_ARGS__) -#define svdivr_s64_m(...) __builtin_sve_svdivr_s64_m(__VA_ARGS__) -#define svdivr_s32_x(...) __builtin_sve_svdivr_s32_x(__VA_ARGS__) -#define svdivr_s64_x(...) __builtin_sve_svdivr_s64_x(__VA_ARGS__) -#define svdivr_s32_z(...) __builtin_sve_svdivr_s32_z(__VA_ARGS__) -#define svdivr_s64_z(...) __builtin_sve_svdivr_s64_z(__VA_ARGS__) -#define svdivr_u32_m(...) __builtin_sve_svdivr_u32_m(__VA_ARGS__) -#define svdivr_u64_m(...) __builtin_sve_svdivr_u64_m(__VA_ARGS__) -#define svdivr_u32_x(...) __builtin_sve_svdivr_u32_x(__VA_ARGS__) -#define svdivr_u64_x(...) __builtin_sve_svdivr_u64_x(__VA_ARGS__) -#define svdivr_u32_z(...) __builtin_sve_svdivr_u32_z(__VA_ARGS__) -#define svdivr_u64_z(...) __builtin_sve_svdivr_u64_z(__VA_ARGS__) -#define svdot_n_s32(...) __builtin_sve_svdot_n_s32(__VA_ARGS__) -#define svdot_n_s64(...) __builtin_sve_svdot_n_s64(__VA_ARGS__) -#define svdot_n_u32(...) __builtin_sve_svdot_n_u32(__VA_ARGS__) -#define svdot_n_u64(...) __builtin_sve_svdot_n_u64(__VA_ARGS__) -#define svdot_s32(...) __builtin_sve_svdot_s32(__VA_ARGS__) -#define svdot_s64(...) __builtin_sve_svdot_s64(__VA_ARGS__) -#define svdot_u32(...) __builtin_sve_svdot_u32(__VA_ARGS__) -#define svdot_u64(...) __builtin_sve_svdot_u64(__VA_ARGS__) -#define svdot_lane_s32(...) __builtin_sve_svdot_lane_s32(__VA_ARGS__) -#define svdot_lane_s64(...) __builtin_sve_svdot_lane_s64(__VA_ARGS__) -#define svdot_lane_u32(...) __builtin_sve_svdot_lane_u32(__VA_ARGS__) -#define svdot_lane_u64(...) __builtin_sve_svdot_lane_u64(__VA_ARGS__) -#define svdup_n_u8(...) __builtin_sve_svdup_n_u8(__VA_ARGS__) -#define svdup_n_u32(...) __builtin_sve_svdup_n_u32(__VA_ARGS__) -#define svdup_n_u64(...) __builtin_sve_svdup_n_u64(__VA_ARGS__) -#define svdup_n_u16(...) __builtin_sve_svdup_n_u16(__VA_ARGS__) -#define svdup_n_s8(...) __builtin_sve_svdup_n_s8(__VA_ARGS__) -#define svdup_n_f64(...) __builtin_sve_svdup_n_f64(__VA_ARGS__) -#define svdup_n_f32(...) __builtin_sve_svdup_n_f32(__VA_ARGS__) -#define svdup_n_f16(...) __builtin_sve_svdup_n_f16(__VA_ARGS__) -#define svdup_n_s32(...) __builtin_sve_svdup_n_s32(__VA_ARGS__) -#define svdup_n_s64(...) __builtin_sve_svdup_n_s64(__VA_ARGS__) -#define svdup_n_s16(...) __builtin_sve_svdup_n_s16(__VA_ARGS__) -#define svdup_n_u8_m(...) __builtin_sve_svdup_n_u8_m(__VA_ARGS__) -#define svdup_n_u32_m(...) __builtin_sve_svdup_n_u32_m(__VA_ARGS__) -#define svdup_n_u64_m(...) __builtin_sve_svdup_n_u64_m(__VA_ARGS__) -#define svdup_n_u16_m(...) __builtin_sve_svdup_n_u16_m(__VA_ARGS__) -#define svdup_n_s8_m(...) __builtin_sve_svdup_n_s8_m(__VA_ARGS__) -#define svdup_n_f64_m(...) __builtin_sve_svdup_n_f64_m(__VA_ARGS__) -#define svdup_n_f32_m(...) __builtin_sve_svdup_n_f32_m(__VA_ARGS__) -#define svdup_n_f16_m(...) __builtin_sve_svdup_n_f16_m(__VA_ARGS__) -#define svdup_n_s32_m(...) __builtin_sve_svdup_n_s32_m(__VA_ARGS__) -#define svdup_n_s64_m(...) __builtin_sve_svdup_n_s64_m(__VA_ARGS__) -#define svdup_n_s16_m(...) __builtin_sve_svdup_n_s16_m(__VA_ARGS__) -#define svdup_n_b8(...) __builtin_sve_svdup_n_b8(__VA_ARGS__) -#define svdup_n_b32(...) __builtin_sve_svdup_n_b32(__VA_ARGS__) -#define svdup_n_b64(...) __builtin_sve_svdup_n_b64(__VA_ARGS__) -#define svdup_n_b16(...) __builtin_sve_svdup_n_b16(__VA_ARGS__) -#define svdup_n_u8_x(...) __builtin_sve_svdup_n_u8_x(__VA_ARGS__) -#define svdup_n_u32_x(...) __builtin_sve_svdup_n_u32_x(__VA_ARGS__) -#define svdup_n_u64_x(...) __builtin_sve_svdup_n_u64_x(__VA_ARGS__) -#define svdup_n_u16_x(...) __builtin_sve_svdup_n_u16_x(__VA_ARGS__) -#define svdup_n_s8_x(...) __builtin_sve_svdup_n_s8_x(__VA_ARGS__) -#define svdup_n_f64_x(...) __builtin_sve_svdup_n_f64_x(__VA_ARGS__) -#define svdup_n_f32_x(...) __builtin_sve_svdup_n_f32_x(__VA_ARGS__) -#define svdup_n_f16_x(...) __builtin_sve_svdup_n_f16_x(__VA_ARGS__) -#define svdup_n_s32_x(...) __builtin_sve_svdup_n_s32_x(__VA_ARGS__) -#define svdup_n_s64_x(...) __builtin_sve_svdup_n_s64_x(__VA_ARGS__) -#define svdup_n_s16_x(...) __builtin_sve_svdup_n_s16_x(__VA_ARGS__) -#define svdup_n_u8_z(...) __builtin_sve_svdup_n_u8_z(__VA_ARGS__) -#define svdup_n_u32_z(...) __builtin_sve_svdup_n_u32_z(__VA_ARGS__) -#define svdup_n_u64_z(...) __builtin_sve_svdup_n_u64_z(__VA_ARGS__) -#define svdup_n_u16_z(...) __builtin_sve_svdup_n_u16_z(__VA_ARGS__) -#define svdup_n_s8_z(...) __builtin_sve_svdup_n_s8_z(__VA_ARGS__) -#define svdup_n_f64_z(...) __builtin_sve_svdup_n_f64_z(__VA_ARGS__) -#define svdup_n_f32_z(...) __builtin_sve_svdup_n_f32_z(__VA_ARGS__) -#define svdup_n_f16_z(...) __builtin_sve_svdup_n_f16_z(__VA_ARGS__) -#define svdup_n_s32_z(...) __builtin_sve_svdup_n_s32_z(__VA_ARGS__) -#define svdup_n_s64_z(...) __builtin_sve_svdup_n_s64_z(__VA_ARGS__) -#define svdup_n_s16_z(...) __builtin_sve_svdup_n_s16_z(__VA_ARGS__) -#define svdup_lane_u8(...) __builtin_sve_svdup_lane_u8(__VA_ARGS__) -#define svdup_lane_u32(...) __builtin_sve_svdup_lane_u32(__VA_ARGS__) -#define svdup_lane_u64(...) __builtin_sve_svdup_lane_u64(__VA_ARGS__) -#define svdup_lane_u16(...) __builtin_sve_svdup_lane_u16(__VA_ARGS__) -#define svdup_lane_s8(...) __builtin_sve_svdup_lane_s8(__VA_ARGS__) -#define svdup_lane_f64(...) __builtin_sve_svdup_lane_f64(__VA_ARGS__) -#define svdup_lane_f32(...) __builtin_sve_svdup_lane_f32(__VA_ARGS__) -#define svdup_lane_f16(...) __builtin_sve_svdup_lane_f16(__VA_ARGS__) -#define svdup_lane_s32(...) __builtin_sve_svdup_lane_s32(__VA_ARGS__) -#define svdup_lane_s64(...) __builtin_sve_svdup_lane_s64(__VA_ARGS__) -#define svdup_lane_s16(...) __builtin_sve_svdup_lane_s16(__VA_ARGS__) -#define svdupq_n_u16(...) __builtin_sve_svdupq_n_u16(__VA_ARGS__) -#define svdupq_n_f16(...) __builtin_sve_svdupq_n_f16(__VA_ARGS__) -#define svdupq_n_s16(...) __builtin_sve_svdupq_n_s16(__VA_ARGS__) -#define svdupq_n_u32(...) __builtin_sve_svdupq_n_u32(__VA_ARGS__) -#define svdupq_n_f32(...) __builtin_sve_svdupq_n_f32(__VA_ARGS__) -#define svdupq_n_s32(...) __builtin_sve_svdupq_n_s32(__VA_ARGS__) -#define svdupq_n_u64(...) __builtin_sve_svdupq_n_u64(__VA_ARGS__) -#define svdupq_n_f64(...) __builtin_sve_svdupq_n_f64(__VA_ARGS__) -#define svdupq_n_s64(...) __builtin_sve_svdupq_n_s64(__VA_ARGS__) -#define svdupq_n_u8(...) __builtin_sve_svdupq_n_u8(__VA_ARGS__) -#define svdupq_n_s8(...) __builtin_sve_svdupq_n_s8(__VA_ARGS__) -#define svdupq_n_b16(...) __builtin_sve_svdupq_n_b16(__VA_ARGS__) -#define svdupq_n_b32(...) __builtin_sve_svdupq_n_b32(__VA_ARGS__) -#define svdupq_n_b64(...) __builtin_sve_svdupq_n_b64(__VA_ARGS__) -#define svdupq_n_b8(...) __builtin_sve_svdupq_n_b8(__VA_ARGS__) -#define svdupq_lane_u8(...) __builtin_sve_svdupq_lane_u8(__VA_ARGS__) -#define svdupq_lane_u32(...) __builtin_sve_svdupq_lane_u32(__VA_ARGS__) -#define svdupq_lane_u64(...) __builtin_sve_svdupq_lane_u64(__VA_ARGS__) -#define svdupq_lane_u16(...) __builtin_sve_svdupq_lane_u16(__VA_ARGS__) -#define svdupq_lane_s8(...) __builtin_sve_svdupq_lane_s8(__VA_ARGS__) -#define svdupq_lane_f64(...) __builtin_sve_svdupq_lane_f64(__VA_ARGS__) -#define svdupq_lane_f32(...) __builtin_sve_svdupq_lane_f32(__VA_ARGS__) -#define svdupq_lane_f16(...) __builtin_sve_svdupq_lane_f16(__VA_ARGS__) -#define svdupq_lane_s32(...) __builtin_sve_svdupq_lane_s32(__VA_ARGS__) -#define svdupq_lane_s64(...) __builtin_sve_svdupq_lane_s64(__VA_ARGS__) -#define svdupq_lane_s16(...) __builtin_sve_svdupq_lane_s16(__VA_ARGS__) -#define sveor_b_z(...) __builtin_sve_sveor_b_z(__VA_ARGS__) -#define sveor_n_u8_m(...) __builtin_sve_sveor_n_u8_m(__VA_ARGS__) -#define sveor_n_u32_m(...) __builtin_sve_sveor_n_u32_m(__VA_ARGS__) -#define sveor_n_u64_m(...) __builtin_sve_sveor_n_u64_m(__VA_ARGS__) -#define sveor_n_u16_m(...) __builtin_sve_sveor_n_u16_m(__VA_ARGS__) -#define sveor_n_s8_m(...) __builtin_sve_sveor_n_s8_m(__VA_ARGS__) -#define sveor_n_s32_m(...) __builtin_sve_sveor_n_s32_m(__VA_ARGS__) -#define sveor_n_s64_m(...) __builtin_sve_sveor_n_s64_m(__VA_ARGS__) -#define sveor_n_s16_m(...) __builtin_sve_sveor_n_s16_m(__VA_ARGS__) -#define sveor_n_u8_x(...) __builtin_sve_sveor_n_u8_x(__VA_ARGS__) -#define sveor_n_u32_x(...) __builtin_sve_sveor_n_u32_x(__VA_ARGS__) -#define sveor_n_u64_x(...) __builtin_sve_sveor_n_u64_x(__VA_ARGS__) -#define sveor_n_u16_x(...) __builtin_sve_sveor_n_u16_x(__VA_ARGS__) -#define sveor_n_s8_x(...) __builtin_sve_sveor_n_s8_x(__VA_ARGS__) -#define sveor_n_s32_x(...) __builtin_sve_sveor_n_s32_x(__VA_ARGS__) -#define sveor_n_s64_x(...) __builtin_sve_sveor_n_s64_x(__VA_ARGS__) -#define sveor_n_s16_x(...) __builtin_sve_sveor_n_s16_x(__VA_ARGS__) -#define sveor_n_u8_z(...) __builtin_sve_sveor_n_u8_z(__VA_ARGS__) -#define sveor_n_u32_z(...) __builtin_sve_sveor_n_u32_z(__VA_ARGS__) -#define sveor_n_u64_z(...) __builtin_sve_sveor_n_u64_z(__VA_ARGS__) -#define sveor_n_u16_z(...) __builtin_sve_sveor_n_u16_z(__VA_ARGS__) -#define sveor_n_s8_z(...) __builtin_sve_sveor_n_s8_z(__VA_ARGS__) -#define sveor_n_s32_z(...) __builtin_sve_sveor_n_s32_z(__VA_ARGS__) -#define sveor_n_s64_z(...) __builtin_sve_sveor_n_s64_z(__VA_ARGS__) -#define sveor_n_s16_z(...) __builtin_sve_sveor_n_s16_z(__VA_ARGS__) -#define sveor_u8_m(...) __builtin_sve_sveor_u8_m(__VA_ARGS__) -#define sveor_u32_m(...) __builtin_sve_sveor_u32_m(__VA_ARGS__) -#define sveor_u64_m(...) __builtin_sve_sveor_u64_m(__VA_ARGS__) -#define sveor_u16_m(...) __builtin_sve_sveor_u16_m(__VA_ARGS__) -#define sveor_s8_m(...) __builtin_sve_sveor_s8_m(__VA_ARGS__) -#define sveor_s32_m(...) __builtin_sve_sveor_s32_m(__VA_ARGS__) -#define sveor_s64_m(...) __builtin_sve_sveor_s64_m(__VA_ARGS__) -#define sveor_s16_m(...) __builtin_sve_sveor_s16_m(__VA_ARGS__) -#define sveor_u8_x(...) __builtin_sve_sveor_u8_x(__VA_ARGS__) -#define sveor_u32_x(...) __builtin_sve_sveor_u32_x(__VA_ARGS__) -#define sveor_u64_x(...) __builtin_sve_sveor_u64_x(__VA_ARGS__) -#define sveor_u16_x(...) __builtin_sve_sveor_u16_x(__VA_ARGS__) -#define sveor_s8_x(...) __builtin_sve_sveor_s8_x(__VA_ARGS__) -#define sveor_s32_x(...) __builtin_sve_sveor_s32_x(__VA_ARGS__) -#define sveor_s64_x(...) __builtin_sve_sveor_s64_x(__VA_ARGS__) -#define sveor_s16_x(...) __builtin_sve_sveor_s16_x(__VA_ARGS__) -#define sveor_u8_z(...) __builtin_sve_sveor_u8_z(__VA_ARGS__) -#define sveor_u32_z(...) __builtin_sve_sveor_u32_z(__VA_ARGS__) -#define sveor_u64_z(...) __builtin_sve_sveor_u64_z(__VA_ARGS__) -#define sveor_u16_z(...) __builtin_sve_sveor_u16_z(__VA_ARGS__) -#define sveor_s8_z(...) __builtin_sve_sveor_s8_z(__VA_ARGS__) -#define sveor_s32_z(...) __builtin_sve_sveor_s32_z(__VA_ARGS__) -#define sveor_s64_z(...) __builtin_sve_sveor_s64_z(__VA_ARGS__) -#define sveor_s16_z(...) __builtin_sve_sveor_s16_z(__VA_ARGS__) -#define sveorv_u8(...) __builtin_sve_sveorv_u8(__VA_ARGS__) -#define sveorv_u32(...) __builtin_sve_sveorv_u32(__VA_ARGS__) -#define sveorv_u64(...) __builtin_sve_sveorv_u64(__VA_ARGS__) -#define sveorv_u16(...) __builtin_sve_sveorv_u16(__VA_ARGS__) -#define sveorv_s8(...) __builtin_sve_sveorv_s8(__VA_ARGS__) -#define sveorv_s32(...) __builtin_sve_sveorv_s32(__VA_ARGS__) -#define sveorv_s64(...) __builtin_sve_sveorv_s64(__VA_ARGS__) -#define sveorv_s16(...) __builtin_sve_sveorv_s16(__VA_ARGS__) -#define svexpa_f64(...) __builtin_sve_svexpa_f64(__VA_ARGS__) -#define svexpa_f32(...) __builtin_sve_svexpa_f32(__VA_ARGS__) -#define svexpa_f16(...) __builtin_sve_svexpa_f16(__VA_ARGS__) -#define svext_u8(...) __builtin_sve_svext_u8(__VA_ARGS__) -#define svext_u32(...) __builtin_sve_svext_u32(__VA_ARGS__) -#define svext_u64(...) __builtin_sve_svext_u64(__VA_ARGS__) -#define svext_u16(...) __builtin_sve_svext_u16(__VA_ARGS__) -#define svext_s8(...) __builtin_sve_svext_s8(__VA_ARGS__) -#define svext_f64(...) __builtin_sve_svext_f64(__VA_ARGS__) -#define svext_f32(...) __builtin_sve_svext_f32(__VA_ARGS__) -#define svext_f16(...) __builtin_sve_svext_f16(__VA_ARGS__) -#define svext_s32(...) __builtin_sve_svext_s32(__VA_ARGS__) -#define svext_s64(...) __builtin_sve_svext_s64(__VA_ARGS__) -#define svext_s16(...) __builtin_sve_svext_s16(__VA_ARGS__) -#define svextb_s32_m(...) __builtin_sve_svextb_s32_m(__VA_ARGS__) -#define svextb_s64_m(...) __builtin_sve_svextb_s64_m(__VA_ARGS__) -#define svextb_s16_m(...) __builtin_sve_svextb_s16_m(__VA_ARGS__) -#define svextb_s32_x(...) __builtin_sve_svextb_s32_x(__VA_ARGS__) -#define svextb_s64_x(...) __builtin_sve_svextb_s64_x(__VA_ARGS__) -#define svextb_s16_x(...) __builtin_sve_svextb_s16_x(__VA_ARGS__) -#define svextb_s32_z(...) __builtin_sve_svextb_s32_z(__VA_ARGS__) -#define svextb_s64_z(...) __builtin_sve_svextb_s64_z(__VA_ARGS__) -#define svextb_s16_z(...) __builtin_sve_svextb_s16_z(__VA_ARGS__) -#define svextb_u32_m(...) __builtin_sve_svextb_u32_m(__VA_ARGS__) -#define svextb_u64_m(...) __builtin_sve_svextb_u64_m(__VA_ARGS__) -#define svextb_u16_m(...) __builtin_sve_svextb_u16_m(__VA_ARGS__) -#define svextb_u32_x(...) __builtin_sve_svextb_u32_x(__VA_ARGS__) -#define svextb_u64_x(...) __builtin_sve_svextb_u64_x(__VA_ARGS__) -#define svextb_u16_x(...) __builtin_sve_svextb_u16_x(__VA_ARGS__) -#define svextb_u32_z(...) __builtin_sve_svextb_u32_z(__VA_ARGS__) -#define svextb_u64_z(...) __builtin_sve_svextb_u64_z(__VA_ARGS__) -#define svextb_u16_z(...) __builtin_sve_svextb_u16_z(__VA_ARGS__) -#define svexth_s32_m(...) __builtin_sve_svexth_s32_m(__VA_ARGS__) -#define svexth_s64_m(...) __builtin_sve_svexth_s64_m(__VA_ARGS__) -#define svexth_s32_x(...) __builtin_sve_svexth_s32_x(__VA_ARGS__) -#define svexth_s64_x(...) __builtin_sve_svexth_s64_x(__VA_ARGS__) -#define svexth_s32_z(...) __builtin_sve_svexth_s32_z(__VA_ARGS__) -#define svexth_s64_z(...) __builtin_sve_svexth_s64_z(__VA_ARGS__) -#define svexth_u32_m(...) __builtin_sve_svexth_u32_m(__VA_ARGS__) -#define svexth_u64_m(...) __builtin_sve_svexth_u64_m(__VA_ARGS__) -#define svexth_u32_x(...) __builtin_sve_svexth_u32_x(__VA_ARGS__) -#define svexth_u64_x(...) __builtin_sve_svexth_u64_x(__VA_ARGS__) -#define svexth_u32_z(...) __builtin_sve_svexth_u32_z(__VA_ARGS__) -#define svexth_u64_z(...) __builtin_sve_svexth_u64_z(__VA_ARGS__) -#define svextw_s64_m(...) __builtin_sve_svextw_s64_m(__VA_ARGS__) -#define svextw_s64_x(...) __builtin_sve_svextw_s64_x(__VA_ARGS__) -#define svextw_s64_z(...) __builtin_sve_svextw_s64_z(__VA_ARGS__) -#define svextw_u64_m(...) __builtin_sve_svextw_u64_m(__VA_ARGS__) -#define svextw_u64_x(...) __builtin_sve_svextw_u64_x(__VA_ARGS__) -#define svextw_u64_z(...) __builtin_sve_svextw_u64_z(__VA_ARGS__) -#define svget2_u8(...) __builtin_sve_svget2_u8(__VA_ARGS__) -#define svget2_u32(...) __builtin_sve_svget2_u32(__VA_ARGS__) -#define svget2_u64(...) __builtin_sve_svget2_u64(__VA_ARGS__) -#define svget2_u16(...) __builtin_sve_svget2_u16(__VA_ARGS__) -#define svget2_s8(...) __builtin_sve_svget2_s8(__VA_ARGS__) -#define svget2_f64(...) __builtin_sve_svget2_f64(__VA_ARGS__) -#define svget2_f32(...) __builtin_sve_svget2_f32(__VA_ARGS__) -#define svget2_f16(...) __builtin_sve_svget2_f16(__VA_ARGS__) -#define svget2_s32(...) __builtin_sve_svget2_s32(__VA_ARGS__) -#define svget2_s64(...) __builtin_sve_svget2_s64(__VA_ARGS__) -#define svget2_s16(...) __builtin_sve_svget2_s16(__VA_ARGS__) -#define svget3_u8(...) __builtin_sve_svget3_u8(__VA_ARGS__) -#define svget3_u32(...) __builtin_sve_svget3_u32(__VA_ARGS__) -#define svget3_u64(...) __builtin_sve_svget3_u64(__VA_ARGS__) -#define svget3_u16(...) __builtin_sve_svget3_u16(__VA_ARGS__) -#define svget3_s8(...) __builtin_sve_svget3_s8(__VA_ARGS__) -#define svget3_f64(...) __builtin_sve_svget3_f64(__VA_ARGS__) -#define svget3_f32(...) __builtin_sve_svget3_f32(__VA_ARGS__) -#define svget3_f16(...) __builtin_sve_svget3_f16(__VA_ARGS__) -#define svget3_s32(...) __builtin_sve_svget3_s32(__VA_ARGS__) -#define svget3_s64(...) __builtin_sve_svget3_s64(__VA_ARGS__) -#define svget3_s16(...) __builtin_sve_svget3_s16(__VA_ARGS__) -#define svget4_u8(...) __builtin_sve_svget4_u8(__VA_ARGS__) -#define svget4_u32(...) __builtin_sve_svget4_u32(__VA_ARGS__) -#define svget4_u64(...) __builtin_sve_svget4_u64(__VA_ARGS__) -#define svget4_u16(...) __builtin_sve_svget4_u16(__VA_ARGS__) -#define svget4_s8(...) __builtin_sve_svget4_s8(__VA_ARGS__) -#define svget4_f64(...) __builtin_sve_svget4_f64(__VA_ARGS__) -#define svget4_f32(...) __builtin_sve_svget4_f32(__VA_ARGS__) -#define svget4_f16(...) __builtin_sve_svget4_f16(__VA_ARGS__) -#define svget4_s32(...) __builtin_sve_svget4_s32(__VA_ARGS__) -#define svget4_s64(...) __builtin_sve_svget4_s64(__VA_ARGS__) -#define svget4_s16(...) __builtin_sve_svget4_s16(__VA_ARGS__) -#define svindex_u8(...) __builtin_sve_svindex_u8(__VA_ARGS__) -#define svindex_u32(...) __builtin_sve_svindex_u32(__VA_ARGS__) -#define svindex_u64(...) __builtin_sve_svindex_u64(__VA_ARGS__) -#define svindex_u16(...) __builtin_sve_svindex_u16(__VA_ARGS__) -#define svindex_s8(...) __builtin_sve_svindex_s8(__VA_ARGS__) -#define svindex_s32(...) __builtin_sve_svindex_s32(__VA_ARGS__) -#define svindex_s64(...) __builtin_sve_svindex_s64(__VA_ARGS__) -#define svindex_s16(...) __builtin_sve_svindex_s16(__VA_ARGS__) -#define svinsr_n_u8(...) __builtin_sve_svinsr_n_u8(__VA_ARGS__) -#define svinsr_n_u32(...) __builtin_sve_svinsr_n_u32(__VA_ARGS__) -#define svinsr_n_u64(...) __builtin_sve_svinsr_n_u64(__VA_ARGS__) -#define svinsr_n_u16(...) __builtin_sve_svinsr_n_u16(__VA_ARGS__) -#define svinsr_n_s8(...) __builtin_sve_svinsr_n_s8(__VA_ARGS__) -#define svinsr_n_f64(...) __builtin_sve_svinsr_n_f64(__VA_ARGS__) -#define svinsr_n_f32(...) __builtin_sve_svinsr_n_f32(__VA_ARGS__) -#define svinsr_n_f16(...) __builtin_sve_svinsr_n_f16(__VA_ARGS__) -#define svinsr_n_s32(...) __builtin_sve_svinsr_n_s32(__VA_ARGS__) -#define svinsr_n_s64(...) __builtin_sve_svinsr_n_s64(__VA_ARGS__) -#define svinsr_n_s16(...) __builtin_sve_svinsr_n_s16(__VA_ARGS__) -#define svlasta_u8(...) __builtin_sve_svlasta_u8(__VA_ARGS__) -#define svlasta_u32(...) __builtin_sve_svlasta_u32(__VA_ARGS__) -#define svlasta_u64(...) __builtin_sve_svlasta_u64(__VA_ARGS__) -#define svlasta_u16(...) __builtin_sve_svlasta_u16(__VA_ARGS__) -#define svlasta_s8(...) __builtin_sve_svlasta_s8(__VA_ARGS__) -#define svlasta_f64(...) __builtin_sve_svlasta_f64(__VA_ARGS__) -#define svlasta_f32(...) __builtin_sve_svlasta_f32(__VA_ARGS__) -#define svlasta_f16(...) __builtin_sve_svlasta_f16(__VA_ARGS__) -#define svlasta_s32(...) __builtin_sve_svlasta_s32(__VA_ARGS__) -#define svlasta_s64(...) __builtin_sve_svlasta_s64(__VA_ARGS__) -#define svlasta_s16(...) __builtin_sve_svlasta_s16(__VA_ARGS__) -#define svlastb_u8(...) __builtin_sve_svlastb_u8(__VA_ARGS__) -#define svlastb_u32(...) __builtin_sve_svlastb_u32(__VA_ARGS__) -#define svlastb_u64(...) __builtin_sve_svlastb_u64(__VA_ARGS__) -#define svlastb_u16(...) __builtin_sve_svlastb_u16(__VA_ARGS__) -#define svlastb_s8(...) __builtin_sve_svlastb_s8(__VA_ARGS__) -#define svlastb_f64(...) __builtin_sve_svlastb_f64(__VA_ARGS__) -#define svlastb_f32(...) __builtin_sve_svlastb_f32(__VA_ARGS__) -#define svlastb_f16(...) __builtin_sve_svlastb_f16(__VA_ARGS__) -#define svlastb_s32(...) __builtin_sve_svlastb_s32(__VA_ARGS__) -#define svlastb_s64(...) __builtin_sve_svlastb_s64(__VA_ARGS__) -#define svlastb_s16(...) __builtin_sve_svlastb_s16(__VA_ARGS__) -#define svld1_u8(...) __builtin_sve_svld1_u8(__VA_ARGS__) -#define svld1_u32(...) __builtin_sve_svld1_u32(__VA_ARGS__) -#define svld1_u64(...) __builtin_sve_svld1_u64(__VA_ARGS__) -#define svld1_u16(...) __builtin_sve_svld1_u16(__VA_ARGS__) -#define svld1_s8(...) __builtin_sve_svld1_s8(__VA_ARGS__) -#define svld1_f64(...) __builtin_sve_svld1_f64(__VA_ARGS__) -#define svld1_f32(...) __builtin_sve_svld1_f32(__VA_ARGS__) -#define svld1_f16(...) __builtin_sve_svld1_f16(__VA_ARGS__) -#define svld1_s32(...) __builtin_sve_svld1_s32(__VA_ARGS__) -#define svld1_s64(...) __builtin_sve_svld1_s64(__VA_ARGS__) -#define svld1_s16(...) __builtin_sve_svld1_s16(__VA_ARGS__) -#define svld1_gather_u32base_index_u32(...) __builtin_sve_svld1_gather_u32base_index_u32(__VA_ARGS__) -#define svld1_gather_u64base_index_u64(...) __builtin_sve_svld1_gather_u64base_index_u64(__VA_ARGS__) -#define svld1_gather_u64base_index_f64(...) __builtin_sve_svld1_gather_u64base_index_f64(__VA_ARGS__) -#define svld1_gather_u32base_index_f32(...) __builtin_sve_svld1_gather_u32base_index_f32(__VA_ARGS__) -#define svld1_gather_u32base_index_s32(...) __builtin_sve_svld1_gather_u32base_index_s32(__VA_ARGS__) -#define svld1_gather_u64base_index_s64(...) __builtin_sve_svld1_gather_u64base_index_s64(__VA_ARGS__) -#define svld1_gather_u32base_offset_u32(...) __builtin_sve_svld1_gather_u32base_offset_u32(__VA_ARGS__) -#define svld1_gather_u64base_offset_u64(...) __builtin_sve_svld1_gather_u64base_offset_u64(__VA_ARGS__) -#define svld1_gather_u64base_offset_f64(...) __builtin_sve_svld1_gather_u64base_offset_f64(__VA_ARGS__) -#define svld1_gather_u32base_offset_f32(...) __builtin_sve_svld1_gather_u32base_offset_f32(__VA_ARGS__) -#define svld1_gather_u32base_offset_s32(...) __builtin_sve_svld1_gather_u32base_offset_s32(__VA_ARGS__) -#define svld1_gather_u64base_offset_s64(...) __builtin_sve_svld1_gather_u64base_offset_s64(__VA_ARGS__) -#define svld1_gather_u32base_u32(...) __builtin_sve_svld1_gather_u32base_u32(__VA_ARGS__) -#define svld1_gather_u64base_u64(...) __builtin_sve_svld1_gather_u64base_u64(__VA_ARGS__) -#define svld1_gather_u64base_f64(...) __builtin_sve_svld1_gather_u64base_f64(__VA_ARGS__) -#define svld1_gather_u32base_f32(...) __builtin_sve_svld1_gather_u32base_f32(__VA_ARGS__) -#define svld1_gather_u32base_s32(...) __builtin_sve_svld1_gather_u32base_s32(__VA_ARGS__) -#define svld1_gather_u64base_s64(...) __builtin_sve_svld1_gather_u64base_s64(__VA_ARGS__) -#define svld1_gather_s32index_u32(...) __builtin_sve_svld1_gather_s32index_u32(__VA_ARGS__) -#define svld1_gather_s32index_f32(...) __builtin_sve_svld1_gather_s32index_f32(__VA_ARGS__) -#define svld1_gather_s32index_s32(...) __builtin_sve_svld1_gather_s32index_s32(__VA_ARGS__) -#define svld1_gather_u32index_u32(...) __builtin_sve_svld1_gather_u32index_u32(__VA_ARGS__) -#define svld1_gather_u32index_f32(...) __builtin_sve_svld1_gather_u32index_f32(__VA_ARGS__) -#define svld1_gather_u32index_s32(...) __builtin_sve_svld1_gather_u32index_s32(__VA_ARGS__) -#define svld1_gather_s64index_u64(...) __builtin_sve_svld1_gather_s64index_u64(__VA_ARGS__) -#define svld1_gather_s64index_f64(...) __builtin_sve_svld1_gather_s64index_f64(__VA_ARGS__) -#define svld1_gather_s64index_s64(...) __builtin_sve_svld1_gather_s64index_s64(__VA_ARGS__) -#define svld1_gather_u64index_u64(...) __builtin_sve_svld1_gather_u64index_u64(__VA_ARGS__) -#define svld1_gather_u64index_f64(...) __builtin_sve_svld1_gather_u64index_f64(__VA_ARGS__) -#define svld1_gather_u64index_s64(...) __builtin_sve_svld1_gather_u64index_s64(__VA_ARGS__) -#define svld1_gather_s32offset_u32(...) __builtin_sve_svld1_gather_s32offset_u32(__VA_ARGS__) -#define svld1_gather_s32offset_f32(...) __builtin_sve_svld1_gather_s32offset_f32(__VA_ARGS__) -#define svld1_gather_s32offset_s32(...) __builtin_sve_svld1_gather_s32offset_s32(__VA_ARGS__) -#define svld1_gather_u32offset_u32(...) __builtin_sve_svld1_gather_u32offset_u32(__VA_ARGS__) -#define svld1_gather_u32offset_f32(...) __builtin_sve_svld1_gather_u32offset_f32(__VA_ARGS__) -#define svld1_gather_u32offset_s32(...) __builtin_sve_svld1_gather_u32offset_s32(__VA_ARGS__) -#define svld1_gather_s64offset_u64(...) __builtin_sve_svld1_gather_s64offset_u64(__VA_ARGS__) -#define svld1_gather_s64offset_f64(...) __builtin_sve_svld1_gather_s64offset_f64(__VA_ARGS__) -#define svld1_gather_s64offset_s64(...) __builtin_sve_svld1_gather_s64offset_s64(__VA_ARGS__) -#define svld1_gather_u64offset_u64(...) __builtin_sve_svld1_gather_u64offset_u64(__VA_ARGS__) -#define svld1_gather_u64offset_f64(...) __builtin_sve_svld1_gather_u64offset_f64(__VA_ARGS__) -#define svld1_gather_u64offset_s64(...) __builtin_sve_svld1_gather_u64offset_s64(__VA_ARGS__) -#define svld1_vnum_u8(...) __builtin_sve_svld1_vnum_u8(__VA_ARGS__) -#define svld1_vnum_u32(...) __builtin_sve_svld1_vnum_u32(__VA_ARGS__) -#define svld1_vnum_u64(...) __builtin_sve_svld1_vnum_u64(__VA_ARGS__) -#define svld1_vnum_u16(...) __builtin_sve_svld1_vnum_u16(__VA_ARGS__) -#define svld1_vnum_s8(...) __builtin_sve_svld1_vnum_s8(__VA_ARGS__) -#define svld1_vnum_f64(...) __builtin_sve_svld1_vnum_f64(__VA_ARGS__) -#define svld1_vnum_f32(...) __builtin_sve_svld1_vnum_f32(__VA_ARGS__) -#define svld1_vnum_f16(...) __builtin_sve_svld1_vnum_f16(__VA_ARGS__) -#define svld1_vnum_s32(...) __builtin_sve_svld1_vnum_s32(__VA_ARGS__) -#define svld1_vnum_s64(...) __builtin_sve_svld1_vnum_s64(__VA_ARGS__) -#define svld1_vnum_s16(...) __builtin_sve_svld1_vnum_s16(__VA_ARGS__) -#define svld1rq_u8(...) __builtin_sve_svld1rq_u8(__VA_ARGS__) -#define svld1rq_u32(...) __builtin_sve_svld1rq_u32(__VA_ARGS__) -#define svld1rq_u64(...) __builtin_sve_svld1rq_u64(__VA_ARGS__) -#define svld1rq_u16(...) __builtin_sve_svld1rq_u16(__VA_ARGS__) -#define svld1rq_s8(...) __builtin_sve_svld1rq_s8(__VA_ARGS__) -#define svld1rq_f64(...) __builtin_sve_svld1rq_f64(__VA_ARGS__) -#define svld1rq_f32(...) __builtin_sve_svld1rq_f32(__VA_ARGS__) -#define svld1rq_f16(...) __builtin_sve_svld1rq_f16(__VA_ARGS__) -#define svld1rq_s32(...) __builtin_sve_svld1rq_s32(__VA_ARGS__) -#define svld1rq_s64(...) __builtin_sve_svld1rq_s64(__VA_ARGS__) -#define svld1rq_s16(...) __builtin_sve_svld1rq_s16(__VA_ARGS__) -#define svld1sb_gather_u32base_offset_u32(...) __builtin_sve_svld1sb_gather_u32base_offset_u32(__VA_ARGS__) -#define svld1sb_gather_u64base_offset_u64(...) __builtin_sve_svld1sb_gather_u64base_offset_u64(__VA_ARGS__) -#define svld1sb_gather_u32base_offset_s32(...) __builtin_sve_svld1sb_gather_u32base_offset_s32(__VA_ARGS__) -#define svld1sb_gather_u64base_offset_s64(...) __builtin_sve_svld1sb_gather_u64base_offset_s64(__VA_ARGS__) -#define svld1sb_gather_u32base_u32(...) __builtin_sve_svld1sb_gather_u32base_u32(__VA_ARGS__) -#define svld1sb_gather_u64base_u64(...) __builtin_sve_svld1sb_gather_u64base_u64(__VA_ARGS__) -#define svld1sb_gather_u32base_s32(...) __builtin_sve_svld1sb_gather_u32base_s32(__VA_ARGS__) -#define svld1sb_gather_u64base_s64(...) __builtin_sve_svld1sb_gather_u64base_s64(__VA_ARGS__) -#define svld1sb_gather_s32offset_u32(...) __builtin_sve_svld1sb_gather_s32offset_u32(__VA_ARGS__) -#define svld1sb_gather_s32offset_s32(...) __builtin_sve_svld1sb_gather_s32offset_s32(__VA_ARGS__) -#define svld1sb_gather_u32offset_u32(...) __builtin_sve_svld1sb_gather_u32offset_u32(__VA_ARGS__) -#define svld1sb_gather_u32offset_s32(...) __builtin_sve_svld1sb_gather_u32offset_s32(__VA_ARGS__) -#define svld1sb_gather_s64offset_u64(...) __builtin_sve_svld1sb_gather_s64offset_u64(__VA_ARGS__) -#define svld1sb_gather_s64offset_s64(...) __builtin_sve_svld1sb_gather_s64offset_s64(__VA_ARGS__) -#define svld1sb_gather_u64offset_u64(...) __builtin_sve_svld1sb_gather_u64offset_u64(__VA_ARGS__) -#define svld1sb_gather_u64offset_s64(...) __builtin_sve_svld1sb_gather_u64offset_s64(__VA_ARGS__) -#define svld1sb_vnum_u32(...) __builtin_sve_svld1sb_vnum_u32(__VA_ARGS__) -#define svld1sb_vnum_u64(...) __builtin_sve_svld1sb_vnum_u64(__VA_ARGS__) -#define svld1sb_vnum_u16(...) __builtin_sve_svld1sb_vnum_u16(__VA_ARGS__) -#define svld1sb_vnum_s32(...) __builtin_sve_svld1sb_vnum_s32(__VA_ARGS__) -#define svld1sb_vnum_s64(...) __builtin_sve_svld1sb_vnum_s64(__VA_ARGS__) -#define svld1sb_vnum_s16(...) __builtin_sve_svld1sb_vnum_s16(__VA_ARGS__) -#define svld1sb_u32(...) __builtin_sve_svld1sb_u32(__VA_ARGS__) -#define svld1sb_u64(...) __builtin_sve_svld1sb_u64(__VA_ARGS__) -#define svld1sb_u16(...) __builtin_sve_svld1sb_u16(__VA_ARGS__) -#define svld1sb_s32(...) __builtin_sve_svld1sb_s32(__VA_ARGS__) -#define svld1sb_s64(...) __builtin_sve_svld1sb_s64(__VA_ARGS__) -#define svld1sb_s16(...) __builtin_sve_svld1sb_s16(__VA_ARGS__) -#define svld1sh_gather_u32base_index_u32(...) __builtin_sve_svld1sh_gather_u32base_index_u32(__VA_ARGS__) -#define svld1sh_gather_u64base_index_u64(...) __builtin_sve_svld1sh_gather_u64base_index_u64(__VA_ARGS__) -#define svld1sh_gather_u32base_index_s32(...) __builtin_sve_svld1sh_gather_u32base_index_s32(__VA_ARGS__) -#define svld1sh_gather_u64base_index_s64(...) __builtin_sve_svld1sh_gather_u64base_index_s64(__VA_ARGS__) -#define svld1sh_gather_u32base_offset_u32(...) __builtin_sve_svld1sh_gather_u32base_offset_u32(__VA_ARGS__) -#define svld1sh_gather_u64base_offset_u64(...) __builtin_sve_svld1sh_gather_u64base_offset_u64(__VA_ARGS__) -#define svld1sh_gather_u32base_offset_s32(...) __builtin_sve_svld1sh_gather_u32base_offset_s32(__VA_ARGS__) -#define svld1sh_gather_u64base_offset_s64(...) __builtin_sve_svld1sh_gather_u64base_offset_s64(__VA_ARGS__) -#define svld1sh_gather_u32base_u32(...) __builtin_sve_svld1sh_gather_u32base_u32(__VA_ARGS__) -#define svld1sh_gather_u64base_u64(...) __builtin_sve_svld1sh_gather_u64base_u64(__VA_ARGS__) -#define svld1sh_gather_u32base_s32(...) __builtin_sve_svld1sh_gather_u32base_s32(__VA_ARGS__) -#define svld1sh_gather_u64base_s64(...) __builtin_sve_svld1sh_gather_u64base_s64(__VA_ARGS__) -#define svld1sh_gather_s32index_u32(...) __builtin_sve_svld1sh_gather_s32index_u32(__VA_ARGS__) -#define svld1sh_gather_s32index_s32(...) __builtin_sve_svld1sh_gather_s32index_s32(__VA_ARGS__) -#define svld1sh_gather_u32index_u32(...) __builtin_sve_svld1sh_gather_u32index_u32(__VA_ARGS__) -#define svld1sh_gather_u32index_s32(...) __builtin_sve_svld1sh_gather_u32index_s32(__VA_ARGS__) -#define svld1sh_gather_s64index_u64(...) __builtin_sve_svld1sh_gather_s64index_u64(__VA_ARGS__) -#define svld1sh_gather_s64index_s64(...) __builtin_sve_svld1sh_gather_s64index_s64(__VA_ARGS__) -#define svld1sh_gather_u64index_u64(...) __builtin_sve_svld1sh_gather_u64index_u64(__VA_ARGS__) -#define svld1sh_gather_u64index_s64(...) __builtin_sve_svld1sh_gather_u64index_s64(__VA_ARGS__) -#define svld1sh_gather_s32offset_u32(...) __builtin_sve_svld1sh_gather_s32offset_u32(__VA_ARGS__) -#define svld1sh_gather_s32offset_s32(...) __builtin_sve_svld1sh_gather_s32offset_s32(__VA_ARGS__) -#define svld1sh_gather_u32offset_u32(...) __builtin_sve_svld1sh_gather_u32offset_u32(__VA_ARGS__) -#define svld1sh_gather_u32offset_s32(...) __builtin_sve_svld1sh_gather_u32offset_s32(__VA_ARGS__) -#define svld1sh_gather_s64offset_u64(...) __builtin_sve_svld1sh_gather_s64offset_u64(__VA_ARGS__) -#define svld1sh_gather_s64offset_s64(...) __builtin_sve_svld1sh_gather_s64offset_s64(__VA_ARGS__) -#define svld1sh_gather_u64offset_u64(...) __builtin_sve_svld1sh_gather_u64offset_u64(__VA_ARGS__) -#define svld1sh_gather_u64offset_s64(...) __builtin_sve_svld1sh_gather_u64offset_s64(__VA_ARGS__) -#define svld1sh_vnum_u32(...) __builtin_sve_svld1sh_vnum_u32(__VA_ARGS__) -#define svld1sh_vnum_u64(...) __builtin_sve_svld1sh_vnum_u64(__VA_ARGS__) -#define svld1sh_vnum_s32(...) __builtin_sve_svld1sh_vnum_s32(__VA_ARGS__) -#define svld1sh_vnum_s64(...) __builtin_sve_svld1sh_vnum_s64(__VA_ARGS__) -#define svld1sh_u32(...) __builtin_sve_svld1sh_u32(__VA_ARGS__) -#define svld1sh_u64(...) __builtin_sve_svld1sh_u64(__VA_ARGS__) -#define svld1sh_s32(...) __builtin_sve_svld1sh_s32(__VA_ARGS__) -#define svld1sh_s64(...) __builtin_sve_svld1sh_s64(__VA_ARGS__) -#define svld1sw_gather_u64base_index_u64(...) __builtin_sve_svld1sw_gather_u64base_index_u64(__VA_ARGS__) -#define svld1sw_gather_u64base_index_s64(...) __builtin_sve_svld1sw_gather_u64base_index_s64(__VA_ARGS__) -#define svld1sw_gather_u64base_offset_u64(...) __builtin_sve_svld1sw_gather_u64base_offset_u64(__VA_ARGS__) -#define svld1sw_gather_u64base_offset_s64(...) __builtin_sve_svld1sw_gather_u64base_offset_s64(__VA_ARGS__) -#define svld1sw_gather_u64base_u64(...) __builtin_sve_svld1sw_gather_u64base_u64(__VA_ARGS__) -#define svld1sw_gather_u64base_s64(...) __builtin_sve_svld1sw_gather_u64base_s64(__VA_ARGS__) -#define svld1sw_gather_s64index_u64(...) __builtin_sve_svld1sw_gather_s64index_u64(__VA_ARGS__) -#define svld1sw_gather_s64index_s64(...) __builtin_sve_svld1sw_gather_s64index_s64(__VA_ARGS__) -#define svld1sw_gather_u64index_u64(...) __builtin_sve_svld1sw_gather_u64index_u64(__VA_ARGS__) -#define svld1sw_gather_u64index_s64(...) __builtin_sve_svld1sw_gather_u64index_s64(__VA_ARGS__) -#define svld1sw_gather_s64offset_u64(...) __builtin_sve_svld1sw_gather_s64offset_u64(__VA_ARGS__) -#define svld1sw_gather_s64offset_s64(...) __builtin_sve_svld1sw_gather_s64offset_s64(__VA_ARGS__) -#define svld1sw_gather_u64offset_u64(...) __builtin_sve_svld1sw_gather_u64offset_u64(__VA_ARGS__) -#define svld1sw_gather_u64offset_s64(...) __builtin_sve_svld1sw_gather_u64offset_s64(__VA_ARGS__) -#define svld1sw_vnum_u64(...) __builtin_sve_svld1sw_vnum_u64(__VA_ARGS__) -#define svld1sw_vnum_s64(...) __builtin_sve_svld1sw_vnum_s64(__VA_ARGS__) -#define svld1sw_u64(...) __builtin_sve_svld1sw_u64(__VA_ARGS__) -#define svld1sw_s64(...) __builtin_sve_svld1sw_s64(__VA_ARGS__) -#define svld1ub_gather_u32base_offset_u32(...) __builtin_sve_svld1ub_gather_u32base_offset_u32(__VA_ARGS__) -#define svld1ub_gather_u64base_offset_u64(...) __builtin_sve_svld1ub_gather_u64base_offset_u64(__VA_ARGS__) -#define svld1ub_gather_u32base_offset_s32(...) __builtin_sve_svld1ub_gather_u32base_offset_s32(__VA_ARGS__) -#define svld1ub_gather_u64base_offset_s64(...) __builtin_sve_svld1ub_gather_u64base_offset_s64(__VA_ARGS__) -#define svld1ub_gather_u32base_u32(...) __builtin_sve_svld1ub_gather_u32base_u32(__VA_ARGS__) -#define svld1ub_gather_u64base_u64(...) __builtin_sve_svld1ub_gather_u64base_u64(__VA_ARGS__) -#define svld1ub_gather_u32base_s32(...) __builtin_sve_svld1ub_gather_u32base_s32(__VA_ARGS__) -#define svld1ub_gather_u64base_s64(...) __builtin_sve_svld1ub_gather_u64base_s64(__VA_ARGS__) -#define svld1ub_gather_s32offset_u32(...) __builtin_sve_svld1ub_gather_s32offset_u32(__VA_ARGS__) -#define svld1ub_gather_s32offset_s32(...) __builtin_sve_svld1ub_gather_s32offset_s32(__VA_ARGS__) -#define svld1ub_gather_u32offset_u32(...) __builtin_sve_svld1ub_gather_u32offset_u32(__VA_ARGS__) -#define svld1ub_gather_u32offset_s32(...) __builtin_sve_svld1ub_gather_u32offset_s32(__VA_ARGS__) -#define svld1ub_gather_s64offset_u64(...) __builtin_sve_svld1ub_gather_s64offset_u64(__VA_ARGS__) -#define svld1ub_gather_s64offset_s64(...) __builtin_sve_svld1ub_gather_s64offset_s64(__VA_ARGS__) -#define svld1ub_gather_u64offset_u64(...) __builtin_sve_svld1ub_gather_u64offset_u64(__VA_ARGS__) -#define svld1ub_gather_u64offset_s64(...) __builtin_sve_svld1ub_gather_u64offset_s64(__VA_ARGS__) -#define svld1ub_vnum_u32(...) __builtin_sve_svld1ub_vnum_u32(__VA_ARGS__) -#define svld1ub_vnum_u64(...) __builtin_sve_svld1ub_vnum_u64(__VA_ARGS__) -#define svld1ub_vnum_u16(...) __builtin_sve_svld1ub_vnum_u16(__VA_ARGS__) -#define svld1ub_vnum_s32(...) __builtin_sve_svld1ub_vnum_s32(__VA_ARGS__) -#define svld1ub_vnum_s64(...) __builtin_sve_svld1ub_vnum_s64(__VA_ARGS__) -#define svld1ub_vnum_s16(...) __builtin_sve_svld1ub_vnum_s16(__VA_ARGS__) -#define svld1ub_u32(...) __builtin_sve_svld1ub_u32(__VA_ARGS__) -#define svld1ub_u64(...) __builtin_sve_svld1ub_u64(__VA_ARGS__) -#define svld1ub_u16(...) __builtin_sve_svld1ub_u16(__VA_ARGS__) -#define svld1ub_s32(...) __builtin_sve_svld1ub_s32(__VA_ARGS__) -#define svld1ub_s64(...) __builtin_sve_svld1ub_s64(__VA_ARGS__) -#define svld1ub_s16(...) __builtin_sve_svld1ub_s16(__VA_ARGS__) -#define svld1uh_gather_u32base_index_u32(...) __builtin_sve_svld1uh_gather_u32base_index_u32(__VA_ARGS__) -#define svld1uh_gather_u64base_index_u64(...) __builtin_sve_svld1uh_gather_u64base_index_u64(__VA_ARGS__) -#define svld1uh_gather_u32base_index_s32(...) __builtin_sve_svld1uh_gather_u32base_index_s32(__VA_ARGS__) -#define svld1uh_gather_u64base_index_s64(...) __builtin_sve_svld1uh_gather_u64base_index_s64(__VA_ARGS__) -#define svld1uh_gather_u32base_offset_u32(...) __builtin_sve_svld1uh_gather_u32base_offset_u32(__VA_ARGS__) -#define svld1uh_gather_u64base_offset_u64(...) __builtin_sve_svld1uh_gather_u64base_offset_u64(__VA_ARGS__) -#define svld1uh_gather_u32base_offset_s32(...) __builtin_sve_svld1uh_gather_u32base_offset_s32(__VA_ARGS__) -#define svld1uh_gather_u64base_offset_s64(...) __builtin_sve_svld1uh_gather_u64base_offset_s64(__VA_ARGS__) -#define svld1uh_gather_u32base_u32(...) __builtin_sve_svld1uh_gather_u32base_u32(__VA_ARGS__) -#define svld1uh_gather_u64base_u64(...) __builtin_sve_svld1uh_gather_u64base_u64(__VA_ARGS__) -#define svld1uh_gather_u32base_s32(...) __builtin_sve_svld1uh_gather_u32base_s32(__VA_ARGS__) -#define svld1uh_gather_u64base_s64(...) __builtin_sve_svld1uh_gather_u64base_s64(__VA_ARGS__) -#define svld1uh_gather_s32index_u32(...) __builtin_sve_svld1uh_gather_s32index_u32(__VA_ARGS__) -#define svld1uh_gather_s32index_s32(...) __builtin_sve_svld1uh_gather_s32index_s32(__VA_ARGS__) -#define svld1uh_gather_u32index_u32(...) __builtin_sve_svld1uh_gather_u32index_u32(__VA_ARGS__) -#define svld1uh_gather_u32index_s32(...) __builtin_sve_svld1uh_gather_u32index_s32(__VA_ARGS__) -#define svld1uh_gather_s64index_u64(...) __builtin_sve_svld1uh_gather_s64index_u64(__VA_ARGS__) -#define svld1uh_gather_s64index_s64(...) __builtin_sve_svld1uh_gather_s64index_s64(__VA_ARGS__) -#define svld1uh_gather_u64index_u64(...) __builtin_sve_svld1uh_gather_u64index_u64(__VA_ARGS__) -#define svld1uh_gather_u64index_s64(...) __builtin_sve_svld1uh_gather_u64index_s64(__VA_ARGS__) -#define svld1uh_gather_s32offset_u32(...) __builtin_sve_svld1uh_gather_s32offset_u32(__VA_ARGS__) -#define svld1uh_gather_s32offset_s32(...) __builtin_sve_svld1uh_gather_s32offset_s32(__VA_ARGS__) -#define svld1uh_gather_u32offset_u32(...) __builtin_sve_svld1uh_gather_u32offset_u32(__VA_ARGS__) -#define svld1uh_gather_u32offset_s32(...) __builtin_sve_svld1uh_gather_u32offset_s32(__VA_ARGS__) -#define svld1uh_gather_s64offset_u64(...) __builtin_sve_svld1uh_gather_s64offset_u64(__VA_ARGS__) -#define svld1uh_gather_s64offset_s64(...) __builtin_sve_svld1uh_gather_s64offset_s64(__VA_ARGS__) -#define svld1uh_gather_u64offset_u64(...) __builtin_sve_svld1uh_gather_u64offset_u64(__VA_ARGS__) -#define svld1uh_gather_u64offset_s64(...) __builtin_sve_svld1uh_gather_u64offset_s64(__VA_ARGS__) -#define svld1uh_vnum_u32(...) __builtin_sve_svld1uh_vnum_u32(__VA_ARGS__) -#define svld1uh_vnum_u64(...) __builtin_sve_svld1uh_vnum_u64(__VA_ARGS__) -#define svld1uh_vnum_s32(...) __builtin_sve_svld1uh_vnum_s32(__VA_ARGS__) -#define svld1uh_vnum_s64(...) __builtin_sve_svld1uh_vnum_s64(__VA_ARGS__) -#define svld1uh_u32(...) __builtin_sve_svld1uh_u32(__VA_ARGS__) -#define svld1uh_u64(...) __builtin_sve_svld1uh_u64(__VA_ARGS__) -#define svld1uh_s32(...) __builtin_sve_svld1uh_s32(__VA_ARGS__) -#define svld1uh_s64(...) __builtin_sve_svld1uh_s64(__VA_ARGS__) -#define svld1uw_gather_u64base_index_u64(...) __builtin_sve_svld1uw_gather_u64base_index_u64(__VA_ARGS__) -#define svld1uw_gather_u64base_index_s64(...) __builtin_sve_svld1uw_gather_u64base_index_s64(__VA_ARGS__) -#define svld1uw_gather_u64base_offset_u64(...) __builtin_sve_svld1uw_gather_u64base_offset_u64(__VA_ARGS__) -#define svld1uw_gather_u64base_offset_s64(...) __builtin_sve_svld1uw_gather_u64base_offset_s64(__VA_ARGS__) -#define svld1uw_gather_u64base_u64(...) __builtin_sve_svld1uw_gather_u64base_u64(__VA_ARGS__) -#define svld1uw_gather_u64base_s64(...) __builtin_sve_svld1uw_gather_u64base_s64(__VA_ARGS__) -#define svld1uw_gather_s64index_u64(...) __builtin_sve_svld1uw_gather_s64index_u64(__VA_ARGS__) -#define svld1uw_gather_s64index_s64(...) __builtin_sve_svld1uw_gather_s64index_s64(__VA_ARGS__) -#define svld1uw_gather_u64index_u64(...) __builtin_sve_svld1uw_gather_u64index_u64(__VA_ARGS__) -#define svld1uw_gather_u64index_s64(...) __builtin_sve_svld1uw_gather_u64index_s64(__VA_ARGS__) -#define svld1uw_gather_s64offset_u64(...) __builtin_sve_svld1uw_gather_s64offset_u64(__VA_ARGS__) -#define svld1uw_gather_s64offset_s64(...) __builtin_sve_svld1uw_gather_s64offset_s64(__VA_ARGS__) -#define svld1uw_gather_u64offset_u64(...) __builtin_sve_svld1uw_gather_u64offset_u64(__VA_ARGS__) -#define svld1uw_gather_u64offset_s64(...) __builtin_sve_svld1uw_gather_u64offset_s64(__VA_ARGS__) -#define svld1uw_vnum_u64(...) __builtin_sve_svld1uw_vnum_u64(__VA_ARGS__) -#define svld1uw_vnum_s64(...) __builtin_sve_svld1uw_vnum_s64(__VA_ARGS__) -#define svld1uw_u64(...) __builtin_sve_svld1uw_u64(__VA_ARGS__) -#define svld1uw_s64(...) __builtin_sve_svld1uw_s64(__VA_ARGS__) -#define svld2_u8(...) __builtin_sve_svld2_u8(__VA_ARGS__) -#define svld2_u32(...) __builtin_sve_svld2_u32(__VA_ARGS__) -#define svld2_u64(...) __builtin_sve_svld2_u64(__VA_ARGS__) -#define svld2_u16(...) __builtin_sve_svld2_u16(__VA_ARGS__) -#define svld2_s8(...) __builtin_sve_svld2_s8(__VA_ARGS__) -#define svld2_f64(...) __builtin_sve_svld2_f64(__VA_ARGS__) -#define svld2_f32(...) __builtin_sve_svld2_f32(__VA_ARGS__) -#define svld2_f16(...) __builtin_sve_svld2_f16(__VA_ARGS__) -#define svld2_s32(...) __builtin_sve_svld2_s32(__VA_ARGS__) -#define svld2_s64(...) __builtin_sve_svld2_s64(__VA_ARGS__) -#define svld2_s16(...) __builtin_sve_svld2_s16(__VA_ARGS__) -#define svld2_vnum_u8(...) __builtin_sve_svld2_vnum_u8(__VA_ARGS__) -#define svld2_vnum_u32(...) __builtin_sve_svld2_vnum_u32(__VA_ARGS__) -#define svld2_vnum_u64(...) __builtin_sve_svld2_vnum_u64(__VA_ARGS__) -#define svld2_vnum_u16(...) __builtin_sve_svld2_vnum_u16(__VA_ARGS__) -#define svld2_vnum_s8(...) __builtin_sve_svld2_vnum_s8(__VA_ARGS__) -#define svld2_vnum_f64(...) __builtin_sve_svld2_vnum_f64(__VA_ARGS__) -#define svld2_vnum_f32(...) __builtin_sve_svld2_vnum_f32(__VA_ARGS__) -#define svld2_vnum_f16(...) __builtin_sve_svld2_vnum_f16(__VA_ARGS__) -#define svld2_vnum_s32(...) __builtin_sve_svld2_vnum_s32(__VA_ARGS__) -#define svld2_vnum_s64(...) __builtin_sve_svld2_vnum_s64(__VA_ARGS__) -#define svld2_vnum_s16(...) __builtin_sve_svld2_vnum_s16(__VA_ARGS__) -#define svld3_u8(...) __builtin_sve_svld3_u8(__VA_ARGS__) -#define svld3_u32(...) __builtin_sve_svld3_u32(__VA_ARGS__) -#define svld3_u64(...) __builtin_sve_svld3_u64(__VA_ARGS__) -#define svld3_u16(...) __builtin_sve_svld3_u16(__VA_ARGS__) -#define svld3_s8(...) __builtin_sve_svld3_s8(__VA_ARGS__) -#define svld3_f64(...) __builtin_sve_svld3_f64(__VA_ARGS__) -#define svld3_f32(...) __builtin_sve_svld3_f32(__VA_ARGS__) -#define svld3_f16(...) __builtin_sve_svld3_f16(__VA_ARGS__) -#define svld3_s32(...) __builtin_sve_svld3_s32(__VA_ARGS__) -#define svld3_s64(...) __builtin_sve_svld3_s64(__VA_ARGS__) -#define svld3_s16(...) __builtin_sve_svld3_s16(__VA_ARGS__) -#define svld3_vnum_u8(...) __builtin_sve_svld3_vnum_u8(__VA_ARGS__) -#define svld3_vnum_u32(...) __builtin_sve_svld3_vnum_u32(__VA_ARGS__) -#define svld3_vnum_u64(...) __builtin_sve_svld3_vnum_u64(__VA_ARGS__) -#define svld3_vnum_u16(...) __builtin_sve_svld3_vnum_u16(__VA_ARGS__) -#define svld3_vnum_s8(...) __builtin_sve_svld3_vnum_s8(__VA_ARGS__) -#define svld3_vnum_f64(...) __builtin_sve_svld3_vnum_f64(__VA_ARGS__) -#define svld3_vnum_f32(...) __builtin_sve_svld3_vnum_f32(__VA_ARGS__) -#define svld3_vnum_f16(...) __builtin_sve_svld3_vnum_f16(__VA_ARGS__) -#define svld3_vnum_s32(...) __builtin_sve_svld3_vnum_s32(__VA_ARGS__) -#define svld3_vnum_s64(...) __builtin_sve_svld3_vnum_s64(__VA_ARGS__) -#define svld3_vnum_s16(...) __builtin_sve_svld3_vnum_s16(__VA_ARGS__) -#define svld4_u8(...) __builtin_sve_svld4_u8(__VA_ARGS__) -#define svld4_u32(...) __builtin_sve_svld4_u32(__VA_ARGS__) -#define svld4_u64(...) __builtin_sve_svld4_u64(__VA_ARGS__) -#define svld4_u16(...) __builtin_sve_svld4_u16(__VA_ARGS__) -#define svld4_s8(...) __builtin_sve_svld4_s8(__VA_ARGS__) -#define svld4_f64(...) __builtin_sve_svld4_f64(__VA_ARGS__) -#define svld4_f32(...) __builtin_sve_svld4_f32(__VA_ARGS__) -#define svld4_f16(...) __builtin_sve_svld4_f16(__VA_ARGS__) -#define svld4_s32(...) __builtin_sve_svld4_s32(__VA_ARGS__) -#define svld4_s64(...) __builtin_sve_svld4_s64(__VA_ARGS__) -#define svld4_s16(...) __builtin_sve_svld4_s16(__VA_ARGS__) -#define svld4_vnum_u8(...) __builtin_sve_svld4_vnum_u8(__VA_ARGS__) -#define svld4_vnum_u32(...) __builtin_sve_svld4_vnum_u32(__VA_ARGS__) -#define svld4_vnum_u64(...) __builtin_sve_svld4_vnum_u64(__VA_ARGS__) -#define svld4_vnum_u16(...) __builtin_sve_svld4_vnum_u16(__VA_ARGS__) -#define svld4_vnum_s8(...) __builtin_sve_svld4_vnum_s8(__VA_ARGS__) -#define svld4_vnum_f64(...) __builtin_sve_svld4_vnum_f64(__VA_ARGS__) -#define svld4_vnum_f32(...) __builtin_sve_svld4_vnum_f32(__VA_ARGS__) -#define svld4_vnum_f16(...) __builtin_sve_svld4_vnum_f16(__VA_ARGS__) -#define svld4_vnum_s32(...) __builtin_sve_svld4_vnum_s32(__VA_ARGS__) -#define svld4_vnum_s64(...) __builtin_sve_svld4_vnum_s64(__VA_ARGS__) -#define svld4_vnum_s16(...) __builtin_sve_svld4_vnum_s16(__VA_ARGS__) -#define svldff1_u8(...) __builtin_sve_svldff1_u8(__VA_ARGS__) -#define svldff1_u32(...) __builtin_sve_svldff1_u32(__VA_ARGS__) -#define svldff1_u64(...) __builtin_sve_svldff1_u64(__VA_ARGS__) -#define svldff1_u16(...) __builtin_sve_svldff1_u16(__VA_ARGS__) -#define svldff1_s8(...) __builtin_sve_svldff1_s8(__VA_ARGS__) -#define svldff1_f64(...) __builtin_sve_svldff1_f64(__VA_ARGS__) -#define svldff1_f32(...) __builtin_sve_svldff1_f32(__VA_ARGS__) -#define svldff1_f16(...) __builtin_sve_svldff1_f16(__VA_ARGS__) -#define svldff1_s32(...) __builtin_sve_svldff1_s32(__VA_ARGS__) -#define svldff1_s64(...) __builtin_sve_svldff1_s64(__VA_ARGS__) -#define svldff1_s16(...) __builtin_sve_svldff1_s16(__VA_ARGS__) -#define svldff1_gather_u32base_index_u32(...) __builtin_sve_svldff1_gather_u32base_index_u32(__VA_ARGS__) -#define svldff1_gather_u64base_index_u64(...) __builtin_sve_svldff1_gather_u64base_index_u64(__VA_ARGS__) -#define svldff1_gather_u64base_index_f64(...) __builtin_sve_svldff1_gather_u64base_index_f64(__VA_ARGS__) -#define svldff1_gather_u32base_index_f32(...) __builtin_sve_svldff1_gather_u32base_index_f32(__VA_ARGS__) -#define svldff1_gather_u32base_index_s32(...) __builtin_sve_svldff1_gather_u32base_index_s32(__VA_ARGS__) -#define svldff1_gather_u64base_index_s64(...) __builtin_sve_svldff1_gather_u64base_index_s64(__VA_ARGS__) -#define svldff1_gather_u32base_offset_u32(...) __builtin_sve_svldff1_gather_u32base_offset_u32(__VA_ARGS__) -#define svldff1_gather_u64base_offset_u64(...) __builtin_sve_svldff1_gather_u64base_offset_u64(__VA_ARGS__) -#define svldff1_gather_u64base_offset_f64(...) __builtin_sve_svldff1_gather_u64base_offset_f64(__VA_ARGS__) -#define svldff1_gather_u32base_offset_f32(...) __builtin_sve_svldff1_gather_u32base_offset_f32(__VA_ARGS__) -#define svldff1_gather_u32base_offset_s32(...) __builtin_sve_svldff1_gather_u32base_offset_s32(__VA_ARGS__) -#define svldff1_gather_u64base_offset_s64(...) __builtin_sve_svldff1_gather_u64base_offset_s64(__VA_ARGS__) -#define svldff1_gather_u32base_u32(...) __builtin_sve_svldff1_gather_u32base_u32(__VA_ARGS__) -#define svldff1_gather_u64base_u64(...) __builtin_sve_svldff1_gather_u64base_u64(__VA_ARGS__) -#define svldff1_gather_u64base_f64(...) __builtin_sve_svldff1_gather_u64base_f64(__VA_ARGS__) -#define svldff1_gather_u32base_f32(...) __builtin_sve_svldff1_gather_u32base_f32(__VA_ARGS__) -#define svldff1_gather_u32base_s32(...) __builtin_sve_svldff1_gather_u32base_s32(__VA_ARGS__) -#define svldff1_gather_u64base_s64(...) __builtin_sve_svldff1_gather_u64base_s64(__VA_ARGS__) -#define svldff1_gather_s32index_u32(...) __builtin_sve_svldff1_gather_s32index_u32(__VA_ARGS__) -#define svldff1_gather_s32index_f32(...) __builtin_sve_svldff1_gather_s32index_f32(__VA_ARGS__) -#define svldff1_gather_s32index_s32(...) __builtin_sve_svldff1_gather_s32index_s32(__VA_ARGS__) -#define svldff1_gather_u32index_u32(...) __builtin_sve_svldff1_gather_u32index_u32(__VA_ARGS__) -#define svldff1_gather_u32index_f32(...) __builtin_sve_svldff1_gather_u32index_f32(__VA_ARGS__) -#define svldff1_gather_u32index_s32(...) __builtin_sve_svldff1_gather_u32index_s32(__VA_ARGS__) -#define svldff1_gather_s64index_u64(...) __builtin_sve_svldff1_gather_s64index_u64(__VA_ARGS__) -#define svldff1_gather_s64index_f64(...) __builtin_sve_svldff1_gather_s64index_f64(__VA_ARGS__) -#define svldff1_gather_s64index_s64(...) __builtin_sve_svldff1_gather_s64index_s64(__VA_ARGS__) -#define svldff1_gather_u64index_u64(...) __builtin_sve_svldff1_gather_u64index_u64(__VA_ARGS__) -#define svldff1_gather_u64index_f64(...) __builtin_sve_svldff1_gather_u64index_f64(__VA_ARGS__) -#define svldff1_gather_u64index_s64(...) __builtin_sve_svldff1_gather_u64index_s64(__VA_ARGS__) -#define svldff1_gather_s32offset_u32(...) __builtin_sve_svldff1_gather_s32offset_u32(__VA_ARGS__) -#define svldff1_gather_s32offset_f32(...) __builtin_sve_svldff1_gather_s32offset_f32(__VA_ARGS__) -#define svldff1_gather_s32offset_s32(...) __builtin_sve_svldff1_gather_s32offset_s32(__VA_ARGS__) -#define svldff1_gather_u32offset_u32(...) __builtin_sve_svldff1_gather_u32offset_u32(__VA_ARGS__) -#define svldff1_gather_u32offset_f32(...) __builtin_sve_svldff1_gather_u32offset_f32(__VA_ARGS__) -#define svldff1_gather_u32offset_s32(...) __builtin_sve_svldff1_gather_u32offset_s32(__VA_ARGS__) -#define svldff1_gather_s64offset_u64(...) __builtin_sve_svldff1_gather_s64offset_u64(__VA_ARGS__) -#define svldff1_gather_s64offset_f64(...) __builtin_sve_svldff1_gather_s64offset_f64(__VA_ARGS__) -#define svldff1_gather_s64offset_s64(...) __builtin_sve_svldff1_gather_s64offset_s64(__VA_ARGS__) -#define svldff1_gather_u64offset_u64(...) __builtin_sve_svldff1_gather_u64offset_u64(__VA_ARGS__) -#define svldff1_gather_u64offset_f64(...) __builtin_sve_svldff1_gather_u64offset_f64(__VA_ARGS__) -#define svldff1_gather_u64offset_s64(...) __builtin_sve_svldff1_gather_u64offset_s64(__VA_ARGS__) -#define svldff1_vnum_u8(...) __builtin_sve_svldff1_vnum_u8(__VA_ARGS__) -#define svldff1_vnum_u32(...) __builtin_sve_svldff1_vnum_u32(__VA_ARGS__) -#define svldff1_vnum_u64(...) __builtin_sve_svldff1_vnum_u64(__VA_ARGS__) -#define svldff1_vnum_u16(...) __builtin_sve_svldff1_vnum_u16(__VA_ARGS__) -#define svldff1_vnum_s8(...) __builtin_sve_svldff1_vnum_s8(__VA_ARGS__) -#define svldff1_vnum_f64(...) __builtin_sve_svldff1_vnum_f64(__VA_ARGS__) -#define svldff1_vnum_f32(...) __builtin_sve_svldff1_vnum_f32(__VA_ARGS__) -#define svldff1_vnum_f16(...) __builtin_sve_svldff1_vnum_f16(__VA_ARGS__) -#define svldff1_vnum_s32(...) __builtin_sve_svldff1_vnum_s32(__VA_ARGS__) -#define svldff1_vnum_s64(...) __builtin_sve_svldff1_vnum_s64(__VA_ARGS__) -#define svldff1_vnum_s16(...) __builtin_sve_svldff1_vnum_s16(__VA_ARGS__) -#define svldff1sb_gather_u32base_offset_u32(...) __builtin_sve_svldff1sb_gather_u32base_offset_u32(__VA_ARGS__) -#define svldff1sb_gather_u64base_offset_u64(...) __builtin_sve_svldff1sb_gather_u64base_offset_u64(__VA_ARGS__) -#define svldff1sb_gather_u32base_offset_s32(...) __builtin_sve_svldff1sb_gather_u32base_offset_s32(__VA_ARGS__) -#define svldff1sb_gather_u64base_offset_s64(...) __builtin_sve_svldff1sb_gather_u64base_offset_s64(__VA_ARGS__) -#define svldff1sb_gather_u32base_u32(...) __builtin_sve_svldff1sb_gather_u32base_u32(__VA_ARGS__) -#define svldff1sb_gather_u64base_u64(...) __builtin_sve_svldff1sb_gather_u64base_u64(__VA_ARGS__) -#define svldff1sb_gather_u32base_s32(...) __builtin_sve_svldff1sb_gather_u32base_s32(__VA_ARGS__) -#define svldff1sb_gather_u64base_s64(...) __builtin_sve_svldff1sb_gather_u64base_s64(__VA_ARGS__) -#define svldff1sb_gather_s32offset_u32(...) __builtin_sve_svldff1sb_gather_s32offset_u32(__VA_ARGS__) -#define svldff1sb_gather_s32offset_s32(...) __builtin_sve_svldff1sb_gather_s32offset_s32(__VA_ARGS__) -#define svldff1sb_gather_u32offset_u32(...) __builtin_sve_svldff1sb_gather_u32offset_u32(__VA_ARGS__) -#define svldff1sb_gather_u32offset_s32(...) __builtin_sve_svldff1sb_gather_u32offset_s32(__VA_ARGS__) -#define svldff1sb_gather_s64offset_u64(...) __builtin_sve_svldff1sb_gather_s64offset_u64(__VA_ARGS__) -#define svldff1sb_gather_s64offset_s64(...) __builtin_sve_svldff1sb_gather_s64offset_s64(__VA_ARGS__) -#define svldff1sb_gather_u64offset_u64(...) __builtin_sve_svldff1sb_gather_u64offset_u64(__VA_ARGS__) -#define svldff1sb_gather_u64offset_s64(...) __builtin_sve_svldff1sb_gather_u64offset_s64(__VA_ARGS__) -#define svldff1sb_vnum_u32(...) __builtin_sve_svldff1sb_vnum_u32(__VA_ARGS__) -#define svldff1sb_vnum_u64(...) __builtin_sve_svldff1sb_vnum_u64(__VA_ARGS__) -#define svldff1sb_vnum_u16(...) __builtin_sve_svldff1sb_vnum_u16(__VA_ARGS__) -#define svldff1sb_vnum_s32(...) __builtin_sve_svldff1sb_vnum_s32(__VA_ARGS__) -#define svldff1sb_vnum_s64(...) __builtin_sve_svldff1sb_vnum_s64(__VA_ARGS__) -#define svldff1sb_vnum_s16(...) __builtin_sve_svldff1sb_vnum_s16(__VA_ARGS__) -#define svldff1sb_u32(...) __builtin_sve_svldff1sb_u32(__VA_ARGS__) -#define svldff1sb_u64(...) __builtin_sve_svldff1sb_u64(__VA_ARGS__) -#define svldff1sb_u16(...) __builtin_sve_svldff1sb_u16(__VA_ARGS__) -#define svldff1sb_s32(...) __builtin_sve_svldff1sb_s32(__VA_ARGS__) -#define svldff1sb_s64(...) __builtin_sve_svldff1sb_s64(__VA_ARGS__) -#define svldff1sb_s16(...) __builtin_sve_svldff1sb_s16(__VA_ARGS__) -#define svldff1sh_gather_u32base_index_u32(...) __builtin_sve_svldff1sh_gather_u32base_index_u32(__VA_ARGS__) -#define svldff1sh_gather_u64base_index_u64(...) __builtin_sve_svldff1sh_gather_u64base_index_u64(__VA_ARGS__) -#define svldff1sh_gather_u32base_index_s32(...) __builtin_sve_svldff1sh_gather_u32base_index_s32(__VA_ARGS__) -#define svldff1sh_gather_u64base_index_s64(...) __builtin_sve_svldff1sh_gather_u64base_index_s64(__VA_ARGS__) -#define svldff1sh_gather_u32base_offset_u32(...) __builtin_sve_svldff1sh_gather_u32base_offset_u32(__VA_ARGS__) -#define svldff1sh_gather_u64base_offset_u64(...) __builtin_sve_svldff1sh_gather_u64base_offset_u64(__VA_ARGS__) -#define svldff1sh_gather_u32base_offset_s32(...) __builtin_sve_svldff1sh_gather_u32base_offset_s32(__VA_ARGS__) -#define svldff1sh_gather_u64base_offset_s64(...) __builtin_sve_svldff1sh_gather_u64base_offset_s64(__VA_ARGS__) -#define svldff1sh_gather_u32base_u32(...) __builtin_sve_svldff1sh_gather_u32base_u32(__VA_ARGS__) -#define svldff1sh_gather_u64base_u64(...) __builtin_sve_svldff1sh_gather_u64base_u64(__VA_ARGS__) -#define svldff1sh_gather_u32base_s32(...) __builtin_sve_svldff1sh_gather_u32base_s32(__VA_ARGS__) -#define svldff1sh_gather_u64base_s64(...) __builtin_sve_svldff1sh_gather_u64base_s64(__VA_ARGS__) -#define svldff1sh_gather_s32index_u32(...) __builtin_sve_svldff1sh_gather_s32index_u32(__VA_ARGS__) -#define svldff1sh_gather_s32index_s32(...) __builtin_sve_svldff1sh_gather_s32index_s32(__VA_ARGS__) -#define svldff1sh_gather_u32index_u32(...) __builtin_sve_svldff1sh_gather_u32index_u32(__VA_ARGS__) -#define svldff1sh_gather_u32index_s32(...) __builtin_sve_svldff1sh_gather_u32index_s32(__VA_ARGS__) -#define svldff1sh_gather_s64index_u64(...) __builtin_sve_svldff1sh_gather_s64index_u64(__VA_ARGS__) -#define svldff1sh_gather_s64index_s64(...) __builtin_sve_svldff1sh_gather_s64index_s64(__VA_ARGS__) -#define svldff1sh_gather_u64index_u64(...) __builtin_sve_svldff1sh_gather_u64index_u64(__VA_ARGS__) -#define svldff1sh_gather_u64index_s64(...) __builtin_sve_svldff1sh_gather_u64index_s64(__VA_ARGS__) -#define svldff1sh_gather_s32offset_u32(...) __builtin_sve_svldff1sh_gather_s32offset_u32(__VA_ARGS__) -#define svldff1sh_gather_s32offset_s32(...) __builtin_sve_svldff1sh_gather_s32offset_s32(__VA_ARGS__) -#define svldff1sh_gather_u32offset_u32(...) __builtin_sve_svldff1sh_gather_u32offset_u32(__VA_ARGS__) -#define svldff1sh_gather_u32offset_s32(...) __builtin_sve_svldff1sh_gather_u32offset_s32(__VA_ARGS__) -#define svldff1sh_gather_s64offset_u64(...) __builtin_sve_svldff1sh_gather_s64offset_u64(__VA_ARGS__) -#define svldff1sh_gather_s64offset_s64(...) __builtin_sve_svldff1sh_gather_s64offset_s64(__VA_ARGS__) -#define svldff1sh_gather_u64offset_u64(...) __builtin_sve_svldff1sh_gather_u64offset_u64(__VA_ARGS__) -#define svldff1sh_gather_u64offset_s64(...) __builtin_sve_svldff1sh_gather_u64offset_s64(__VA_ARGS__) -#define svldff1sh_vnum_u32(...) __builtin_sve_svldff1sh_vnum_u32(__VA_ARGS__) -#define svldff1sh_vnum_u64(...) __builtin_sve_svldff1sh_vnum_u64(__VA_ARGS__) -#define svldff1sh_vnum_s32(...) __builtin_sve_svldff1sh_vnum_s32(__VA_ARGS__) -#define svldff1sh_vnum_s64(...) __builtin_sve_svldff1sh_vnum_s64(__VA_ARGS__) -#define svldff1sh_u32(...) __builtin_sve_svldff1sh_u32(__VA_ARGS__) -#define svldff1sh_u64(...) __builtin_sve_svldff1sh_u64(__VA_ARGS__) -#define svldff1sh_s32(...) __builtin_sve_svldff1sh_s32(__VA_ARGS__) -#define svldff1sh_s64(...) __builtin_sve_svldff1sh_s64(__VA_ARGS__) -#define svldff1sw_gather_u64base_index_u64(...) __builtin_sve_svldff1sw_gather_u64base_index_u64(__VA_ARGS__) -#define svldff1sw_gather_u64base_index_s64(...) __builtin_sve_svldff1sw_gather_u64base_index_s64(__VA_ARGS__) -#define svldff1sw_gather_u64base_offset_u64(...) __builtin_sve_svldff1sw_gather_u64base_offset_u64(__VA_ARGS__) -#define svldff1sw_gather_u64base_offset_s64(...) __builtin_sve_svldff1sw_gather_u64base_offset_s64(__VA_ARGS__) -#define svldff1sw_gather_u64base_u64(...) __builtin_sve_svldff1sw_gather_u64base_u64(__VA_ARGS__) -#define svldff1sw_gather_u64base_s64(...) __builtin_sve_svldff1sw_gather_u64base_s64(__VA_ARGS__) -#define svldff1sw_gather_s64index_u64(...) __builtin_sve_svldff1sw_gather_s64index_u64(__VA_ARGS__) -#define svldff1sw_gather_s64index_s64(...) __builtin_sve_svldff1sw_gather_s64index_s64(__VA_ARGS__) -#define svldff1sw_gather_u64index_u64(...) __builtin_sve_svldff1sw_gather_u64index_u64(__VA_ARGS__) -#define svldff1sw_gather_u64index_s64(...) __builtin_sve_svldff1sw_gather_u64index_s64(__VA_ARGS__) -#define svldff1sw_gather_s64offset_u64(...) __builtin_sve_svldff1sw_gather_s64offset_u64(__VA_ARGS__) -#define svldff1sw_gather_s64offset_s64(...) __builtin_sve_svldff1sw_gather_s64offset_s64(__VA_ARGS__) -#define svldff1sw_gather_u64offset_u64(...) __builtin_sve_svldff1sw_gather_u64offset_u64(__VA_ARGS__) -#define svldff1sw_gather_u64offset_s64(...) __builtin_sve_svldff1sw_gather_u64offset_s64(__VA_ARGS__) -#define svldff1sw_vnum_u64(...) __builtin_sve_svldff1sw_vnum_u64(__VA_ARGS__) -#define svldff1sw_vnum_s64(...) __builtin_sve_svldff1sw_vnum_s64(__VA_ARGS__) -#define svldff1sw_u64(...) __builtin_sve_svldff1sw_u64(__VA_ARGS__) -#define svldff1sw_s64(...) __builtin_sve_svldff1sw_s64(__VA_ARGS__) -#define svldff1ub_gather_u32base_offset_u32(...) __builtin_sve_svldff1ub_gather_u32base_offset_u32(__VA_ARGS__) -#define svldff1ub_gather_u64base_offset_u64(...) __builtin_sve_svldff1ub_gather_u64base_offset_u64(__VA_ARGS__) -#define svldff1ub_gather_u32base_offset_s32(...) __builtin_sve_svldff1ub_gather_u32base_offset_s32(__VA_ARGS__) -#define svldff1ub_gather_u64base_offset_s64(...) __builtin_sve_svldff1ub_gather_u64base_offset_s64(__VA_ARGS__) -#define svldff1ub_gather_u32base_u32(...) __builtin_sve_svldff1ub_gather_u32base_u32(__VA_ARGS__) -#define svldff1ub_gather_u64base_u64(...) __builtin_sve_svldff1ub_gather_u64base_u64(__VA_ARGS__) -#define svldff1ub_gather_u32base_s32(...) __builtin_sve_svldff1ub_gather_u32base_s32(__VA_ARGS__) -#define svldff1ub_gather_u64base_s64(...) __builtin_sve_svldff1ub_gather_u64base_s64(__VA_ARGS__) -#define svldff1ub_gather_s32offset_u32(...) __builtin_sve_svldff1ub_gather_s32offset_u32(__VA_ARGS__) -#define svldff1ub_gather_s32offset_s32(...) __builtin_sve_svldff1ub_gather_s32offset_s32(__VA_ARGS__) -#define svldff1ub_gather_u32offset_u32(...) __builtin_sve_svldff1ub_gather_u32offset_u32(__VA_ARGS__) -#define svldff1ub_gather_u32offset_s32(...) __builtin_sve_svldff1ub_gather_u32offset_s32(__VA_ARGS__) -#define svldff1ub_gather_s64offset_u64(...) __builtin_sve_svldff1ub_gather_s64offset_u64(__VA_ARGS__) -#define svldff1ub_gather_s64offset_s64(...) __builtin_sve_svldff1ub_gather_s64offset_s64(__VA_ARGS__) -#define svldff1ub_gather_u64offset_u64(...) __builtin_sve_svldff1ub_gather_u64offset_u64(__VA_ARGS__) -#define svldff1ub_gather_u64offset_s64(...) __builtin_sve_svldff1ub_gather_u64offset_s64(__VA_ARGS__) -#define svldff1ub_vnum_u32(...) __builtin_sve_svldff1ub_vnum_u32(__VA_ARGS__) -#define svldff1ub_vnum_u64(...) __builtin_sve_svldff1ub_vnum_u64(__VA_ARGS__) -#define svldff1ub_vnum_u16(...) __builtin_sve_svldff1ub_vnum_u16(__VA_ARGS__) -#define svldff1ub_vnum_s32(...) __builtin_sve_svldff1ub_vnum_s32(__VA_ARGS__) -#define svldff1ub_vnum_s64(...) __builtin_sve_svldff1ub_vnum_s64(__VA_ARGS__) -#define svldff1ub_vnum_s16(...) __builtin_sve_svldff1ub_vnum_s16(__VA_ARGS__) -#define svldff1ub_u32(...) __builtin_sve_svldff1ub_u32(__VA_ARGS__) -#define svldff1ub_u64(...) __builtin_sve_svldff1ub_u64(__VA_ARGS__) -#define svldff1ub_u16(...) __builtin_sve_svldff1ub_u16(__VA_ARGS__) -#define svldff1ub_s32(...) __builtin_sve_svldff1ub_s32(__VA_ARGS__) -#define svldff1ub_s64(...) __builtin_sve_svldff1ub_s64(__VA_ARGS__) -#define svldff1ub_s16(...) __builtin_sve_svldff1ub_s16(__VA_ARGS__) -#define svldff1uh_gather_u32base_index_u32(...) __builtin_sve_svldff1uh_gather_u32base_index_u32(__VA_ARGS__) -#define svldff1uh_gather_u64base_index_u64(...) __builtin_sve_svldff1uh_gather_u64base_index_u64(__VA_ARGS__) -#define svldff1uh_gather_u32base_index_s32(...) __builtin_sve_svldff1uh_gather_u32base_index_s32(__VA_ARGS__) -#define svldff1uh_gather_u64base_index_s64(...) __builtin_sve_svldff1uh_gather_u64base_index_s64(__VA_ARGS__) -#define svldff1uh_gather_u32base_offset_u32(...) __builtin_sve_svldff1uh_gather_u32base_offset_u32(__VA_ARGS__) -#define svldff1uh_gather_u64base_offset_u64(...) __builtin_sve_svldff1uh_gather_u64base_offset_u64(__VA_ARGS__) -#define svldff1uh_gather_u32base_offset_s32(...) __builtin_sve_svldff1uh_gather_u32base_offset_s32(__VA_ARGS__) -#define svldff1uh_gather_u64base_offset_s64(...) __builtin_sve_svldff1uh_gather_u64base_offset_s64(__VA_ARGS__) -#define svldff1uh_gather_u32base_u32(...) __builtin_sve_svldff1uh_gather_u32base_u32(__VA_ARGS__) -#define svldff1uh_gather_u64base_u64(...) __builtin_sve_svldff1uh_gather_u64base_u64(__VA_ARGS__) -#define svldff1uh_gather_u32base_s32(...) __builtin_sve_svldff1uh_gather_u32base_s32(__VA_ARGS__) -#define svldff1uh_gather_u64base_s64(...) __builtin_sve_svldff1uh_gather_u64base_s64(__VA_ARGS__) -#define svldff1uh_gather_s32index_u32(...) __builtin_sve_svldff1uh_gather_s32index_u32(__VA_ARGS__) -#define svldff1uh_gather_s32index_s32(...) __builtin_sve_svldff1uh_gather_s32index_s32(__VA_ARGS__) -#define svldff1uh_gather_u32index_u32(...) __builtin_sve_svldff1uh_gather_u32index_u32(__VA_ARGS__) -#define svldff1uh_gather_u32index_s32(...) __builtin_sve_svldff1uh_gather_u32index_s32(__VA_ARGS__) -#define svldff1uh_gather_s64index_u64(...) __builtin_sve_svldff1uh_gather_s64index_u64(__VA_ARGS__) -#define svldff1uh_gather_s64index_s64(...) __builtin_sve_svldff1uh_gather_s64index_s64(__VA_ARGS__) -#define svldff1uh_gather_u64index_u64(...) __builtin_sve_svldff1uh_gather_u64index_u64(__VA_ARGS__) -#define svldff1uh_gather_u64index_s64(...) __builtin_sve_svldff1uh_gather_u64index_s64(__VA_ARGS__) -#define svldff1uh_gather_s32offset_u32(...) __builtin_sve_svldff1uh_gather_s32offset_u32(__VA_ARGS__) -#define svldff1uh_gather_s32offset_s32(...) __builtin_sve_svldff1uh_gather_s32offset_s32(__VA_ARGS__) -#define svldff1uh_gather_u32offset_u32(...) __builtin_sve_svldff1uh_gather_u32offset_u32(__VA_ARGS__) -#define svldff1uh_gather_u32offset_s32(...) __builtin_sve_svldff1uh_gather_u32offset_s32(__VA_ARGS__) -#define svldff1uh_gather_s64offset_u64(...) __builtin_sve_svldff1uh_gather_s64offset_u64(__VA_ARGS__) -#define svldff1uh_gather_s64offset_s64(...) __builtin_sve_svldff1uh_gather_s64offset_s64(__VA_ARGS__) -#define svldff1uh_gather_u64offset_u64(...) __builtin_sve_svldff1uh_gather_u64offset_u64(__VA_ARGS__) -#define svldff1uh_gather_u64offset_s64(...) __builtin_sve_svldff1uh_gather_u64offset_s64(__VA_ARGS__) -#define svldff1uh_vnum_u32(...) __builtin_sve_svldff1uh_vnum_u32(__VA_ARGS__) -#define svldff1uh_vnum_u64(...) __builtin_sve_svldff1uh_vnum_u64(__VA_ARGS__) -#define svldff1uh_vnum_s32(...) __builtin_sve_svldff1uh_vnum_s32(__VA_ARGS__) -#define svldff1uh_vnum_s64(...) __builtin_sve_svldff1uh_vnum_s64(__VA_ARGS__) -#define svldff1uh_u32(...) __builtin_sve_svldff1uh_u32(__VA_ARGS__) -#define svldff1uh_u64(...) __builtin_sve_svldff1uh_u64(__VA_ARGS__) -#define svldff1uh_s32(...) __builtin_sve_svldff1uh_s32(__VA_ARGS__) -#define svldff1uh_s64(...) __builtin_sve_svldff1uh_s64(__VA_ARGS__) -#define svldff1uw_gather_u64base_index_u64(...) __builtin_sve_svldff1uw_gather_u64base_index_u64(__VA_ARGS__) -#define svldff1uw_gather_u64base_index_s64(...) __builtin_sve_svldff1uw_gather_u64base_index_s64(__VA_ARGS__) -#define svldff1uw_gather_u64base_offset_u64(...) __builtin_sve_svldff1uw_gather_u64base_offset_u64(__VA_ARGS__) -#define svldff1uw_gather_u64base_offset_s64(...) __builtin_sve_svldff1uw_gather_u64base_offset_s64(__VA_ARGS__) -#define svldff1uw_gather_u64base_u64(...) __builtin_sve_svldff1uw_gather_u64base_u64(__VA_ARGS__) -#define svldff1uw_gather_u64base_s64(...) __builtin_sve_svldff1uw_gather_u64base_s64(__VA_ARGS__) -#define svldff1uw_gather_s64index_u64(...) __builtin_sve_svldff1uw_gather_s64index_u64(__VA_ARGS__) -#define svldff1uw_gather_s64index_s64(...) __builtin_sve_svldff1uw_gather_s64index_s64(__VA_ARGS__) -#define svldff1uw_gather_u64index_u64(...) __builtin_sve_svldff1uw_gather_u64index_u64(__VA_ARGS__) -#define svldff1uw_gather_u64index_s64(...) __builtin_sve_svldff1uw_gather_u64index_s64(__VA_ARGS__) -#define svldff1uw_gather_s64offset_u64(...) __builtin_sve_svldff1uw_gather_s64offset_u64(__VA_ARGS__) -#define svldff1uw_gather_s64offset_s64(...) __builtin_sve_svldff1uw_gather_s64offset_s64(__VA_ARGS__) -#define svldff1uw_gather_u64offset_u64(...) __builtin_sve_svldff1uw_gather_u64offset_u64(__VA_ARGS__) -#define svldff1uw_gather_u64offset_s64(...) __builtin_sve_svldff1uw_gather_u64offset_s64(__VA_ARGS__) -#define svldff1uw_vnum_u64(...) __builtin_sve_svldff1uw_vnum_u64(__VA_ARGS__) -#define svldff1uw_vnum_s64(...) __builtin_sve_svldff1uw_vnum_s64(__VA_ARGS__) -#define svldff1uw_u64(...) __builtin_sve_svldff1uw_u64(__VA_ARGS__) -#define svldff1uw_s64(...) __builtin_sve_svldff1uw_s64(__VA_ARGS__) -#define svldnf1_u8(...) __builtin_sve_svldnf1_u8(__VA_ARGS__) -#define svldnf1_u32(...) __builtin_sve_svldnf1_u32(__VA_ARGS__) -#define svldnf1_u64(...) __builtin_sve_svldnf1_u64(__VA_ARGS__) -#define svldnf1_u16(...) __builtin_sve_svldnf1_u16(__VA_ARGS__) -#define svldnf1_s8(...) __builtin_sve_svldnf1_s8(__VA_ARGS__) -#define svldnf1_f64(...) __builtin_sve_svldnf1_f64(__VA_ARGS__) -#define svldnf1_f32(...) __builtin_sve_svldnf1_f32(__VA_ARGS__) -#define svldnf1_f16(...) __builtin_sve_svldnf1_f16(__VA_ARGS__) -#define svldnf1_s32(...) __builtin_sve_svldnf1_s32(__VA_ARGS__) -#define svldnf1_s64(...) __builtin_sve_svldnf1_s64(__VA_ARGS__) -#define svldnf1_s16(...) __builtin_sve_svldnf1_s16(__VA_ARGS__) -#define svldnf1_vnum_u8(...) __builtin_sve_svldnf1_vnum_u8(__VA_ARGS__) -#define svldnf1_vnum_u32(...) __builtin_sve_svldnf1_vnum_u32(__VA_ARGS__) -#define svldnf1_vnum_u64(...) __builtin_sve_svldnf1_vnum_u64(__VA_ARGS__) -#define svldnf1_vnum_u16(...) __builtin_sve_svldnf1_vnum_u16(__VA_ARGS__) -#define svldnf1_vnum_s8(...) __builtin_sve_svldnf1_vnum_s8(__VA_ARGS__) -#define svldnf1_vnum_f64(...) __builtin_sve_svldnf1_vnum_f64(__VA_ARGS__) -#define svldnf1_vnum_f32(...) __builtin_sve_svldnf1_vnum_f32(__VA_ARGS__) -#define svldnf1_vnum_f16(...) __builtin_sve_svldnf1_vnum_f16(__VA_ARGS__) -#define svldnf1_vnum_s32(...) __builtin_sve_svldnf1_vnum_s32(__VA_ARGS__) -#define svldnf1_vnum_s64(...) __builtin_sve_svldnf1_vnum_s64(__VA_ARGS__) -#define svldnf1_vnum_s16(...) __builtin_sve_svldnf1_vnum_s16(__VA_ARGS__) -#define svldnf1sb_vnum_u32(...) __builtin_sve_svldnf1sb_vnum_u32(__VA_ARGS__) -#define svldnf1sb_vnum_u64(...) __builtin_sve_svldnf1sb_vnum_u64(__VA_ARGS__) -#define svldnf1sb_vnum_u16(...) __builtin_sve_svldnf1sb_vnum_u16(__VA_ARGS__) -#define svldnf1sb_vnum_s32(...) __builtin_sve_svldnf1sb_vnum_s32(__VA_ARGS__) -#define svldnf1sb_vnum_s64(...) __builtin_sve_svldnf1sb_vnum_s64(__VA_ARGS__) -#define svldnf1sb_vnum_s16(...) __builtin_sve_svldnf1sb_vnum_s16(__VA_ARGS__) -#define svldnf1sb_u32(...) __builtin_sve_svldnf1sb_u32(__VA_ARGS__) -#define svldnf1sb_u64(...) __builtin_sve_svldnf1sb_u64(__VA_ARGS__) -#define svldnf1sb_u16(...) __builtin_sve_svldnf1sb_u16(__VA_ARGS__) -#define svldnf1sb_s32(...) __builtin_sve_svldnf1sb_s32(__VA_ARGS__) -#define svldnf1sb_s64(...) __builtin_sve_svldnf1sb_s64(__VA_ARGS__) -#define svldnf1sb_s16(...) __builtin_sve_svldnf1sb_s16(__VA_ARGS__) -#define svldnf1sh_vnum_u32(...) __builtin_sve_svldnf1sh_vnum_u32(__VA_ARGS__) -#define svldnf1sh_vnum_u64(...) __builtin_sve_svldnf1sh_vnum_u64(__VA_ARGS__) -#define svldnf1sh_vnum_s32(...) __builtin_sve_svldnf1sh_vnum_s32(__VA_ARGS__) -#define svldnf1sh_vnum_s64(...) __builtin_sve_svldnf1sh_vnum_s64(__VA_ARGS__) -#define svldnf1sh_u32(...) __builtin_sve_svldnf1sh_u32(__VA_ARGS__) -#define svldnf1sh_u64(...) __builtin_sve_svldnf1sh_u64(__VA_ARGS__) -#define svldnf1sh_s32(...) __builtin_sve_svldnf1sh_s32(__VA_ARGS__) -#define svldnf1sh_s64(...) __builtin_sve_svldnf1sh_s64(__VA_ARGS__) -#define svldnf1sw_vnum_u64(...) __builtin_sve_svldnf1sw_vnum_u64(__VA_ARGS__) -#define svldnf1sw_vnum_s64(...) __builtin_sve_svldnf1sw_vnum_s64(__VA_ARGS__) -#define svldnf1sw_u64(...) __builtin_sve_svldnf1sw_u64(__VA_ARGS__) -#define svldnf1sw_s64(...) __builtin_sve_svldnf1sw_s64(__VA_ARGS__) -#define svldnf1ub_vnum_u32(...) __builtin_sve_svldnf1ub_vnum_u32(__VA_ARGS__) -#define svldnf1ub_vnum_u64(...) __builtin_sve_svldnf1ub_vnum_u64(__VA_ARGS__) -#define svldnf1ub_vnum_u16(...) __builtin_sve_svldnf1ub_vnum_u16(__VA_ARGS__) -#define svldnf1ub_vnum_s32(...) __builtin_sve_svldnf1ub_vnum_s32(__VA_ARGS__) -#define svldnf1ub_vnum_s64(...) __builtin_sve_svldnf1ub_vnum_s64(__VA_ARGS__) -#define svldnf1ub_vnum_s16(...) __builtin_sve_svldnf1ub_vnum_s16(__VA_ARGS__) -#define svldnf1ub_u32(...) __builtin_sve_svldnf1ub_u32(__VA_ARGS__) -#define svldnf1ub_u64(...) __builtin_sve_svldnf1ub_u64(__VA_ARGS__) -#define svldnf1ub_u16(...) __builtin_sve_svldnf1ub_u16(__VA_ARGS__) -#define svldnf1ub_s32(...) __builtin_sve_svldnf1ub_s32(__VA_ARGS__) -#define svldnf1ub_s64(...) __builtin_sve_svldnf1ub_s64(__VA_ARGS__) -#define svldnf1ub_s16(...) __builtin_sve_svldnf1ub_s16(__VA_ARGS__) -#define svldnf1uh_vnum_u32(...) __builtin_sve_svldnf1uh_vnum_u32(__VA_ARGS__) -#define svldnf1uh_vnum_u64(...) __builtin_sve_svldnf1uh_vnum_u64(__VA_ARGS__) -#define svldnf1uh_vnum_s32(...) __builtin_sve_svldnf1uh_vnum_s32(__VA_ARGS__) -#define svldnf1uh_vnum_s64(...) __builtin_sve_svldnf1uh_vnum_s64(__VA_ARGS__) -#define svldnf1uh_u32(...) __builtin_sve_svldnf1uh_u32(__VA_ARGS__) -#define svldnf1uh_u64(...) __builtin_sve_svldnf1uh_u64(__VA_ARGS__) -#define svldnf1uh_s32(...) __builtin_sve_svldnf1uh_s32(__VA_ARGS__) -#define svldnf1uh_s64(...) __builtin_sve_svldnf1uh_s64(__VA_ARGS__) -#define svldnf1uw_vnum_u64(...) __builtin_sve_svldnf1uw_vnum_u64(__VA_ARGS__) -#define svldnf1uw_vnum_s64(...) __builtin_sve_svldnf1uw_vnum_s64(__VA_ARGS__) -#define svldnf1uw_u64(...) __builtin_sve_svldnf1uw_u64(__VA_ARGS__) -#define svldnf1uw_s64(...) __builtin_sve_svldnf1uw_s64(__VA_ARGS__) -#define svldnt1_u8(...) __builtin_sve_svldnt1_u8(__VA_ARGS__) -#define svldnt1_u32(...) __builtin_sve_svldnt1_u32(__VA_ARGS__) -#define svldnt1_u64(...) __builtin_sve_svldnt1_u64(__VA_ARGS__) -#define svldnt1_u16(...) __builtin_sve_svldnt1_u16(__VA_ARGS__) -#define svldnt1_s8(...) __builtin_sve_svldnt1_s8(__VA_ARGS__) -#define svldnt1_f64(...) __builtin_sve_svldnt1_f64(__VA_ARGS__) -#define svldnt1_f32(...) __builtin_sve_svldnt1_f32(__VA_ARGS__) -#define svldnt1_f16(...) __builtin_sve_svldnt1_f16(__VA_ARGS__) -#define svldnt1_s32(...) __builtin_sve_svldnt1_s32(__VA_ARGS__) -#define svldnt1_s64(...) __builtin_sve_svldnt1_s64(__VA_ARGS__) -#define svldnt1_s16(...) __builtin_sve_svldnt1_s16(__VA_ARGS__) -#define svldnt1_vnum_u8(...) __builtin_sve_svldnt1_vnum_u8(__VA_ARGS__) -#define svldnt1_vnum_u32(...) __builtin_sve_svldnt1_vnum_u32(__VA_ARGS__) -#define svldnt1_vnum_u64(...) __builtin_sve_svldnt1_vnum_u64(__VA_ARGS__) -#define svldnt1_vnum_u16(...) __builtin_sve_svldnt1_vnum_u16(__VA_ARGS__) -#define svldnt1_vnum_s8(...) __builtin_sve_svldnt1_vnum_s8(__VA_ARGS__) -#define svldnt1_vnum_f64(...) __builtin_sve_svldnt1_vnum_f64(__VA_ARGS__) -#define svldnt1_vnum_f32(...) __builtin_sve_svldnt1_vnum_f32(__VA_ARGS__) -#define svldnt1_vnum_f16(...) __builtin_sve_svldnt1_vnum_f16(__VA_ARGS__) -#define svldnt1_vnum_s32(...) __builtin_sve_svldnt1_vnum_s32(__VA_ARGS__) -#define svldnt1_vnum_s64(...) __builtin_sve_svldnt1_vnum_s64(__VA_ARGS__) -#define svldnt1_vnum_s16(...) __builtin_sve_svldnt1_vnum_s16(__VA_ARGS__) -#define svlen_u8(...) __builtin_sve_svlen_u8(__VA_ARGS__) -#define svlen_u32(...) __builtin_sve_svlen_u32(__VA_ARGS__) -#define svlen_u64(...) __builtin_sve_svlen_u64(__VA_ARGS__) -#define svlen_u16(...) __builtin_sve_svlen_u16(__VA_ARGS__) -#define svlen_s8(...) __builtin_sve_svlen_s8(__VA_ARGS__) -#define svlen_f64(...) __builtin_sve_svlen_f64(__VA_ARGS__) -#define svlen_f32(...) __builtin_sve_svlen_f32(__VA_ARGS__) -#define svlen_f16(...) __builtin_sve_svlen_f16(__VA_ARGS__) -#define svlen_s32(...) __builtin_sve_svlen_s32(__VA_ARGS__) -#define svlen_s64(...) __builtin_sve_svlen_s64(__VA_ARGS__) -#define svlen_s16(...) __builtin_sve_svlen_s16(__VA_ARGS__) -#define svlsl_n_u8_m(...) __builtin_sve_svlsl_n_u8_m(__VA_ARGS__) -#define svlsl_n_u32_m(...) __builtin_sve_svlsl_n_u32_m(__VA_ARGS__) -#define svlsl_n_u64_m(...) __builtin_sve_svlsl_n_u64_m(__VA_ARGS__) -#define svlsl_n_u16_m(...) __builtin_sve_svlsl_n_u16_m(__VA_ARGS__) -#define svlsl_n_s8_m(...) __builtin_sve_svlsl_n_s8_m(__VA_ARGS__) -#define svlsl_n_s32_m(...) __builtin_sve_svlsl_n_s32_m(__VA_ARGS__) -#define svlsl_n_s64_m(...) __builtin_sve_svlsl_n_s64_m(__VA_ARGS__) -#define svlsl_n_s16_m(...) __builtin_sve_svlsl_n_s16_m(__VA_ARGS__) -#define svlsl_n_u8_x(...) __builtin_sve_svlsl_n_u8_x(__VA_ARGS__) -#define svlsl_n_u32_x(...) __builtin_sve_svlsl_n_u32_x(__VA_ARGS__) -#define svlsl_n_u64_x(...) __builtin_sve_svlsl_n_u64_x(__VA_ARGS__) -#define svlsl_n_u16_x(...) __builtin_sve_svlsl_n_u16_x(__VA_ARGS__) -#define svlsl_n_s8_x(...) __builtin_sve_svlsl_n_s8_x(__VA_ARGS__) -#define svlsl_n_s32_x(...) __builtin_sve_svlsl_n_s32_x(__VA_ARGS__) -#define svlsl_n_s64_x(...) __builtin_sve_svlsl_n_s64_x(__VA_ARGS__) -#define svlsl_n_s16_x(...) __builtin_sve_svlsl_n_s16_x(__VA_ARGS__) -#define svlsl_n_u8_z(...) __builtin_sve_svlsl_n_u8_z(__VA_ARGS__) -#define svlsl_n_u32_z(...) __builtin_sve_svlsl_n_u32_z(__VA_ARGS__) -#define svlsl_n_u64_z(...) __builtin_sve_svlsl_n_u64_z(__VA_ARGS__) -#define svlsl_n_u16_z(...) __builtin_sve_svlsl_n_u16_z(__VA_ARGS__) -#define svlsl_n_s8_z(...) __builtin_sve_svlsl_n_s8_z(__VA_ARGS__) -#define svlsl_n_s32_z(...) __builtin_sve_svlsl_n_s32_z(__VA_ARGS__) -#define svlsl_n_s64_z(...) __builtin_sve_svlsl_n_s64_z(__VA_ARGS__) -#define svlsl_n_s16_z(...) __builtin_sve_svlsl_n_s16_z(__VA_ARGS__) -#define svlsl_u8_m(...) __builtin_sve_svlsl_u8_m(__VA_ARGS__) -#define svlsl_u32_m(...) __builtin_sve_svlsl_u32_m(__VA_ARGS__) -#define svlsl_u64_m(...) __builtin_sve_svlsl_u64_m(__VA_ARGS__) -#define svlsl_u16_m(...) __builtin_sve_svlsl_u16_m(__VA_ARGS__) -#define svlsl_s8_m(...) __builtin_sve_svlsl_s8_m(__VA_ARGS__) -#define svlsl_s32_m(...) __builtin_sve_svlsl_s32_m(__VA_ARGS__) -#define svlsl_s64_m(...) __builtin_sve_svlsl_s64_m(__VA_ARGS__) -#define svlsl_s16_m(...) __builtin_sve_svlsl_s16_m(__VA_ARGS__) -#define svlsl_u8_x(...) __builtin_sve_svlsl_u8_x(__VA_ARGS__) -#define svlsl_u32_x(...) __builtin_sve_svlsl_u32_x(__VA_ARGS__) -#define svlsl_u64_x(...) __builtin_sve_svlsl_u64_x(__VA_ARGS__) -#define svlsl_u16_x(...) __builtin_sve_svlsl_u16_x(__VA_ARGS__) -#define svlsl_s8_x(...) __builtin_sve_svlsl_s8_x(__VA_ARGS__) -#define svlsl_s32_x(...) __builtin_sve_svlsl_s32_x(__VA_ARGS__) -#define svlsl_s64_x(...) __builtin_sve_svlsl_s64_x(__VA_ARGS__) -#define svlsl_s16_x(...) __builtin_sve_svlsl_s16_x(__VA_ARGS__) -#define svlsl_u8_z(...) __builtin_sve_svlsl_u8_z(__VA_ARGS__) -#define svlsl_u32_z(...) __builtin_sve_svlsl_u32_z(__VA_ARGS__) -#define svlsl_u64_z(...) __builtin_sve_svlsl_u64_z(__VA_ARGS__) -#define svlsl_u16_z(...) __builtin_sve_svlsl_u16_z(__VA_ARGS__) -#define svlsl_s8_z(...) __builtin_sve_svlsl_s8_z(__VA_ARGS__) -#define svlsl_s32_z(...) __builtin_sve_svlsl_s32_z(__VA_ARGS__) -#define svlsl_s64_z(...) __builtin_sve_svlsl_s64_z(__VA_ARGS__) -#define svlsl_s16_z(...) __builtin_sve_svlsl_s16_z(__VA_ARGS__) -#define svlsl_wide_n_u8_m(...) __builtin_sve_svlsl_wide_n_u8_m(__VA_ARGS__) -#define svlsl_wide_n_u32_m(...) __builtin_sve_svlsl_wide_n_u32_m(__VA_ARGS__) -#define svlsl_wide_n_u16_m(...) __builtin_sve_svlsl_wide_n_u16_m(__VA_ARGS__) -#define svlsl_wide_n_s8_m(...) __builtin_sve_svlsl_wide_n_s8_m(__VA_ARGS__) -#define svlsl_wide_n_s32_m(...) __builtin_sve_svlsl_wide_n_s32_m(__VA_ARGS__) -#define svlsl_wide_n_s16_m(...) __builtin_sve_svlsl_wide_n_s16_m(__VA_ARGS__) -#define svlsl_wide_n_u8_x(...) __builtin_sve_svlsl_wide_n_u8_x(__VA_ARGS__) -#define svlsl_wide_n_u32_x(...) __builtin_sve_svlsl_wide_n_u32_x(__VA_ARGS__) -#define svlsl_wide_n_u16_x(...) __builtin_sve_svlsl_wide_n_u16_x(__VA_ARGS__) -#define svlsl_wide_n_s8_x(...) __builtin_sve_svlsl_wide_n_s8_x(__VA_ARGS__) -#define svlsl_wide_n_s32_x(...) __builtin_sve_svlsl_wide_n_s32_x(__VA_ARGS__) -#define svlsl_wide_n_s16_x(...) __builtin_sve_svlsl_wide_n_s16_x(__VA_ARGS__) -#define svlsl_wide_n_u8_z(...) __builtin_sve_svlsl_wide_n_u8_z(__VA_ARGS__) -#define svlsl_wide_n_u32_z(...) __builtin_sve_svlsl_wide_n_u32_z(__VA_ARGS__) -#define svlsl_wide_n_u16_z(...) __builtin_sve_svlsl_wide_n_u16_z(__VA_ARGS__) -#define svlsl_wide_n_s8_z(...) __builtin_sve_svlsl_wide_n_s8_z(__VA_ARGS__) -#define svlsl_wide_n_s32_z(...) __builtin_sve_svlsl_wide_n_s32_z(__VA_ARGS__) -#define svlsl_wide_n_s16_z(...) __builtin_sve_svlsl_wide_n_s16_z(__VA_ARGS__) -#define svlsl_wide_u8_m(...) __builtin_sve_svlsl_wide_u8_m(__VA_ARGS__) -#define svlsl_wide_u32_m(...) __builtin_sve_svlsl_wide_u32_m(__VA_ARGS__) -#define svlsl_wide_u16_m(...) __builtin_sve_svlsl_wide_u16_m(__VA_ARGS__) -#define svlsl_wide_s8_m(...) __builtin_sve_svlsl_wide_s8_m(__VA_ARGS__) -#define svlsl_wide_s32_m(...) __builtin_sve_svlsl_wide_s32_m(__VA_ARGS__) -#define svlsl_wide_s16_m(...) __builtin_sve_svlsl_wide_s16_m(__VA_ARGS__) -#define svlsl_wide_u8_x(...) __builtin_sve_svlsl_wide_u8_x(__VA_ARGS__) -#define svlsl_wide_u32_x(...) __builtin_sve_svlsl_wide_u32_x(__VA_ARGS__) -#define svlsl_wide_u16_x(...) __builtin_sve_svlsl_wide_u16_x(__VA_ARGS__) -#define svlsl_wide_s8_x(...) __builtin_sve_svlsl_wide_s8_x(__VA_ARGS__) -#define svlsl_wide_s32_x(...) __builtin_sve_svlsl_wide_s32_x(__VA_ARGS__) -#define svlsl_wide_s16_x(...) __builtin_sve_svlsl_wide_s16_x(__VA_ARGS__) -#define svlsl_wide_u8_z(...) __builtin_sve_svlsl_wide_u8_z(__VA_ARGS__) -#define svlsl_wide_u32_z(...) __builtin_sve_svlsl_wide_u32_z(__VA_ARGS__) -#define svlsl_wide_u16_z(...) __builtin_sve_svlsl_wide_u16_z(__VA_ARGS__) -#define svlsl_wide_s8_z(...) __builtin_sve_svlsl_wide_s8_z(__VA_ARGS__) -#define svlsl_wide_s32_z(...) __builtin_sve_svlsl_wide_s32_z(__VA_ARGS__) -#define svlsl_wide_s16_z(...) __builtin_sve_svlsl_wide_s16_z(__VA_ARGS__) -#define svlsr_n_u8_m(...) __builtin_sve_svlsr_n_u8_m(__VA_ARGS__) -#define svlsr_n_u32_m(...) __builtin_sve_svlsr_n_u32_m(__VA_ARGS__) -#define svlsr_n_u64_m(...) __builtin_sve_svlsr_n_u64_m(__VA_ARGS__) -#define svlsr_n_u16_m(...) __builtin_sve_svlsr_n_u16_m(__VA_ARGS__) -#define svlsr_n_u8_x(...) __builtin_sve_svlsr_n_u8_x(__VA_ARGS__) -#define svlsr_n_u32_x(...) __builtin_sve_svlsr_n_u32_x(__VA_ARGS__) -#define svlsr_n_u64_x(...) __builtin_sve_svlsr_n_u64_x(__VA_ARGS__) -#define svlsr_n_u16_x(...) __builtin_sve_svlsr_n_u16_x(__VA_ARGS__) -#define svlsr_n_u8_z(...) __builtin_sve_svlsr_n_u8_z(__VA_ARGS__) -#define svlsr_n_u32_z(...) __builtin_sve_svlsr_n_u32_z(__VA_ARGS__) -#define svlsr_n_u64_z(...) __builtin_sve_svlsr_n_u64_z(__VA_ARGS__) -#define svlsr_n_u16_z(...) __builtin_sve_svlsr_n_u16_z(__VA_ARGS__) -#define svlsr_u8_m(...) __builtin_sve_svlsr_u8_m(__VA_ARGS__) -#define svlsr_u32_m(...) __builtin_sve_svlsr_u32_m(__VA_ARGS__) -#define svlsr_u64_m(...) __builtin_sve_svlsr_u64_m(__VA_ARGS__) -#define svlsr_u16_m(...) __builtin_sve_svlsr_u16_m(__VA_ARGS__) -#define svlsr_u8_x(...) __builtin_sve_svlsr_u8_x(__VA_ARGS__) -#define svlsr_u32_x(...) __builtin_sve_svlsr_u32_x(__VA_ARGS__) -#define svlsr_u64_x(...) __builtin_sve_svlsr_u64_x(__VA_ARGS__) -#define svlsr_u16_x(...) __builtin_sve_svlsr_u16_x(__VA_ARGS__) -#define svlsr_u8_z(...) __builtin_sve_svlsr_u8_z(__VA_ARGS__) -#define svlsr_u32_z(...) __builtin_sve_svlsr_u32_z(__VA_ARGS__) -#define svlsr_u64_z(...) __builtin_sve_svlsr_u64_z(__VA_ARGS__) -#define svlsr_u16_z(...) __builtin_sve_svlsr_u16_z(__VA_ARGS__) -#define svlsr_wide_n_u8_m(...) __builtin_sve_svlsr_wide_n_u8_m(__VA_ARGS__) -#define svlsr_wide_n_u32_m(...) __builtin_sve_svlsr_wide_n_u32_m(__VA_ARGS__) -#define svlsr_wide_n_u16_m(...) __builtin_sve_svlsr_wide_n_u16_m(__VA_ARGS__) -#define svlsr_wide_n_u8_x(...) __builtin_sve_svlsr_wide_n_u8_x(__VA_ARGS__) -#define svlsr_wide_n_u32_x(...) __builtin_sve_svlsr_wide_n_u32_x(__VA_ARGS__) -#define svlsr_wide_n_u16_x(...) __builtin_sve_svlsr_wide_n_u16_x(__VA_ARGS__) -#define svlsr_wide_n_u8_z(...) __builtin_sve_svlsr_wide_n_u8_z(__VA_ARGS__) -#define svlsr_wide_n_u32_z(...) __builtin_sve_svlsr_wide_n_u32_z(__VA_ARGS__) -#define svlsr_wide_n_u16_z(...) __builtin_sve_svlsr_wide_n_u16_z(__VA_ARGS__) -#define svlsr_wide_u8_m(...) __builtin_sve_svlsr_wide_u8_m(__VA_ARGS__) -#define svlsr_wide_u32_m(...) __builtin_sve_svlsr_wide_u32_m(__VA_ARGS__) -#define svlsr_wide_u16_m(...) __builtin_sve_svlsr_wide_u16_m(__VA_ARGS__) -#define svlsr_wide_u8_x(...) __builtin_sve_svlsr_wide_u8_x(__VA_ARGS__) -#define svlsr_wide_u32_x(...) __builtin_sve_svlsr_wide_u32_x(__VA_ARGS__) -#define svlsr_wide_u16_x(...) __builtin_sve_svlsr_wide_u16_x(__VA_ARGS__) -#define svlsr_wide_u8_z(...) __builtin_sve_svlsr_wide_u8_z(__VA_ARGS__) -#define svlsr_wide_u32_z(...) __builtin_sve_svlsr_wide_u32_z(__VA_ARGS__) -#define svlsr_wide_u16_z(...) __builtin_sve_svlsr_wide_u16_z(__VA_ARGS__) -#define svmad_n_f64_m(...) __builtin_sve_svmad_n_f64_m(__VA_ARGS__) -#define svmad_n_f32_m(...) __builtin_sve_svmad_n_f32_m(__VA_ARGS__) -#define svmad_n_f16_m(...) __builtin_sve_svmad_n_f16_m(__VA_ARGS__) -#define svmad_n_f64_x(...) __builtin_sve_svmad_n_f64_x(__VA_ARGS__) -#define svmad_n_f32_x(...) __builtin_sve_svmad_n_f32_x(__VA_ARGS__) -#define svmad_n_f16_x(...) __builtin_sve_svmad_n_f16_x(__VA_ARGS__) -#define svmad_n_f64_z(...) __builtin_sve_svmad_n_f64_z(__VA_ARGS__) -#define svmad_n_f32_z(...) __builtin_sve_svmad_n_f32_z(__VA_ARGS__) -#define svmad_n_f16_z(...) __builtin_sve_svmad_n_f16_z(__VA_ARGS__) -#define svmad_n_u8_m(...) __builtin_sve_svmad_n_u8_m(__VA_ARGS__) -#define svmad_n_u32_m(...) __builtin_sve_svmad_n_u32_m(__VA_ARGS__) -#define svmad_n_u64_m(...) __builtin_sve_svmad_n_u64_m(__VA_ARGS__) -#define svmad_n_u16_m(...) __builtin_sve_svmad_n_u16_m(__VA_ARGS__) -#define svmad_n_s8_m(...) __builtin_sve_svmad_n_s8_m(__VA_ARGS__) -#define svmad_n_s32_m(...) __builtin_sve_svmad_n_s32_m(__VA_ARGS__) -#define svmad_n_s64_m(...) __builtin_sve_svmad_n_s64_m(__VA_ARGS__) -#define svmad_n_s16_m(...) __builtin_sve_svmad_n_s16_m(__VA_ARGS__) -#define svmad_n_u8_x(...) __builtin_sve_svmad_n_u8_x(__VA_ARGS__) -#define svmad_n_u32_x(...) __builtin_sve_svmad_n_u32_x(__VA_ARGS__) -#define svmad_n_u64_x(...) __builtin_sve_svmad_n_u64_x(__VA_ARGS__) -#define svmad_n_u16_x(...) __builtin_sve_svmad_n_u16_x(__VA_ARGS__) -#define svmad_n_s8_x(...) __builtin_sve_svmad_n_s8_x(__VA_ARGS__) -#define svmad_n_s32_x(...) __builtin_sve_svmad_n_s32_x(__VA_ARGS__) -#define svmad_n_s64_x(...) __builtin_sve_svmad_n_s64_x(__VA_ARGS__) -#define svmad_n_s16_x(...) __builtin_sve_svmad_n_s16_x(__VA_ARGS__) -#define svmad_n_u8_z(...) __builtin_sve_svmad_n_u8_z(__VA_ARGS__) -#define svmad_n_u32_z(...) __builtin_sve_svmad_n_u32_z(__VA_ARGS__) -#define svmad_n_u64_z(...) __builtin_sve_svmad_n_u64_z(__VA_ARGS__) -#define svmad_n_u16_z(...) __builtin_sve_svmad_n_u16_z(__VA_ARGS__) -#define svmad_n_s8_z(...) __builtin_sve_svmad_n_s8_z(__VA_ARGS__) -#define svmad_n_s32_z(...) __builtin_sve_svmad_n_s32_z(__VA_ARGS__) -#define svmad_n_s64_z(...) __builtin_sve_svmad_n_s64_z(__VA_ARGS__) -#define svmad_n_s16_z(...) __builtin_sve_svmad_n_s16_z(__VA_ARGS__) -#define svmad_f64_m(...) __builtin_sve_svmad_f64_m(__VA_ARGS__) -#define svmad_f32_m(...) __builtin_sve_svmad_f32_m(__VA_ARGS__) -#define svmad_f16_m(...) __builtin_sve_svmad_f16_m(__VA_ARGS__) -#define svmad_f64_x(...) __builtin_sve_svmad_f64_x(__VA_ARGS__) -#define svmad_f32_x(...) __builtin_sve_svmad_f32_x(__VA_ARGS__) -#define svmad_f16_x(...) __builtin_sve_svmad_f16_x(__VA_ARGS__) -#define svmad_f64_z(...) __builtin_sve_svmad_f64_z(__VA_ARGS__) -#define svmad_f32_z(...) __builtin_sve_svmad_f32_z(__VA_ARGS__) -#define svmad_f16_z(...) __builtin_sve_svmad_f16_z(__VA_ARGS__) -#define svmad_u8_m(...) __builtin_sve_svmad_u8_m(__VA_ARGS__) -#define svmad_u32_m(...) __builtin_sve_svmad_u32_m(__VA_ARGS__) -#define svmad_u64_m(...) __builtin_sve_svmad_u64_m(__VA_ARGS__) -#define svmad_u16_m(...) __builtin_sve_svmad_u16_m(__VA_ARGS__) -#define svmad_s8_m(...) __builtin_sve_svmad_s8_m(__VA_ARGS__) -#define svmad_s32_m(...) __builtin_sve_svmad_s32_m(__VA_ARGS__) -#define svmad_s64_m(...) __builtin_sve_svmad_s64_m(__VA_ARGS__) -#define svmad_s16_m(...) __builtin_sve_svmad_s16_m(__VA_ARGS__) -#define svmad_u8_x(...) __builtin_sve_svmad_u8_x(__VA_ARGS__) -#define svmad_u32_x(...) __builtin_sve_svmad_u32_x(__VA_ARGS__) -#define svmad_u64_x(...) __builtin_sve_svmad_u64_x(__VA_ARGS__) -#define svmad_u16_x(...) __builtin_sve_svmad_u16_x(__VA_ARGS__) -#define svmad_s8_x(...) __builtin_sve_svmad_s8_x(__VA_ARGS__) -#define svmad_s32_x(...) __builtin_sve_svmad_s32_x(__VA_ARGS__) -#define svmad_s64_x(...) __builtin_sve_svmad_s64_x(__VA_ARGS__) -#define svmad_s16_x(...) __builtin_sve_svmad_s16_x(__VA_ARGS__) -#define svmad_u8_z(...) __builtin_sve_svmad_u8_z(__VA_ARGS__) -#define svmad_u32_z(...) __builtin_sve_svmad_u32_z(__VA_ARGS__) -#define svmad_u64_z(...) __builtin_sve_svmad_u64_z(__VA_ARGS__) -#define svmad_u16_z(...) __builtin_sve_svmad_u16_z(__VA_ARGS__) -#define svmad_s8_z(...) __builtin_sve_svmad_s8_z(__VA_ARGS__) -#define svmad_s32_z(...) __builtin_sve_svmad_s32_z(__VA_ARGS__) -#define svmad_s64_z(...) __builtin_sve_svmad_s64_z(__VA_ARGS__) -#define svmad_s16_z(...) __builtin_sve_svmad_s16_z(__VA_ARGS__) -#define svmax_n_f64_m(...) __builtin_sve_svmax_n_f64_m(__VA_ARGS__) -#define svmax_n_f32_m(...) __builtin_sve_svmax_n_f32_m(__VA_ARGS__) -#define svmax_n_f16_m(...) __builtin_sve_svmax_n_f16_m(__VA_ARGS__) -#define svmax_n_f64_x(...) __builtin_sve_svmax_n_f64_x(__VA_ARGS__) -#define svmax_n_f32_x(...) __builtin_sve_svmax_n_f32_x(__VA_ARGS__) -#define svmax_n_f16_x(...) __builtin_sve_svmax_n_f16_x(__VA_ARGS__) -#define svmax_n_f64_z(...) __builtin_sve_svmax_n_f64_z(__VA_ARGS__) -#define svmax_n_f32_z(...) __builtin_sve_svmax_n_f32_z(__VA_ARGS__) -#define svmax_n_f16_z(...) __builtin_sve_svmax_n_f16_z(__VA_ARGS__) -#define svmax_n_s8_m(...) __builtin_sve_svmax_n_s8_m(__VA_ARGS__) -#define svmax_n_s32_m(...) __builtin_sve_svmax_n_s32_m(__VA_ARGS__) -#define svmax_n_s64_m(...) __builtin_sve_svmax_n_s64_m(__VA_ARGS__) -#define svmax_n_s16_m(...) __builtin_sve_svmax_n_s16_m(__VA_ARGS__) -#define svmax_n_s8_x(...) __builtin_sve_svmax_n_s8_x(__VA_ARGS__) -#define svmax_n_s32_x(...) __builtin_sve_svmax_n_s32_x(__VA_ARGS__) -#define svmax_n_s64_x(...) __builtin_sve_svmax_n_s64_x(__VA_ARGS__) -#define svmax_n_s16_x(...) __builtin_sve_svmax_n_s16_x(__VA_ARGS__) -#define svmax_n_s8_z(...) __builtin_sve_svmax_n_s8_z(__VA_ARGS__) -#define svmax_n_s32_z(...) __builtin_sve_svmax_n_s32_z(__VA_ARGS__) -#define svmax_n_s64_z(...) __builtin_sve_svmax_n_s64_z(__VA_ARGS__) -#define svmax_n_s16_z(...) __builtin_sve_svmax_n_s16_z(__VA_ARGS__) -#define svmax_n_u8_m(...) __builtin_sve_svmax_n_u8_m(__VA_ARGS__) -#define svmax_n_u32_m(...) __builtin_sve_svmax_n_u32_m(__VA_ARGS__) -#define svmax_n_u64_m(...) __builtin_sve_svmax_n_u64_m(__VA_ARGS__) -#define svmax_n_u16_m(...) __builtin_sve_svmax_n_u16_m(__VA_ARGS__) -#define svmax_n_u8_x(...) __builtin_sve_svmax_n_u8_x(__VA_ARGS__) -#define svmax_n_u32_x(...) __builtin_sve_svmax_n_u32_x(__VA_ARGS__) -#define svmax_n_u64_x(...) __builtin_sve_svmax_n_u64_x(__VA_ARGS__) -#define svmax_n_u16_x(...) __builtin_sve_svmax_n_u16_x(__VA_ARGS__) -#define svmax_n_u8_z(...) __builtin_sve_svmax_n_u8_z(__VA_ARGS__) -#define svmax_n_u32_z(...) __builtin_sve_svmax_n_u32_z(__VA_ARGS__) -#define svmax_n_u64_z(...) __builtin_sve_svmax_n_u64_z(__VA_ARGS__) -#define svmax_n_u16_z(...) __builtin_sve_svmax_n_u16_z(__VA_ARGS__) -#define svmax_f64_m(...) __builtin_sve_svmax_f64_m(__VA_ARGS__) -#define svmax_f32_m(...) __builtin_sve_svmax_f32_m(__VA_ARGS__) -#define svmax_f16_m(...) __builtin_sve_svmax_f16_m(__VA_ARGS__) -#define svmax_f64_x(...) __builtin_sve_svmax_f64_x(__VA_ARGS__) -#define svmax_f32_x(...) __builtin_sve_svmax_f32_x(__VA_ARGS__) -#define svmax_f16_x(...) __builtin_sve_svmax_f16_x(__VA_ARGS__) -#define svmax_f64_z(...) __builtin_sve_svmax_f64_z(__VA_ARGS__) -#define svmax_f32_z(...) __builtin_sve_svmax_f32_z(__VA_ARGS__) -#define svmax_f16_z(...) __builtin_sve_svmax_f16_z(__VA_ARGS__) -#define svmax_s8_m(...) __builtin_sve_svmax_s8_m(__VA_ARGS__) -#define svmax_s32_m(...) __builtin_sve_svmax_s32_m(__VA_ARGS__) -#define svmax_s64_m(...) __builtin_sve_svmax_s64_m(__VA_ARGS__) -#define svmax_s16_m(...) __builtin_sve_svmax_s16_m(__VA_ARGS__) -#define svmax_s8_x(...) __builtin_sve_svmax_s8_x(__VA_ARGS__) -#define svmax_s32_x(...) __builtin_sve_svmax_s32_x(__VA_ARGS__) -#define svmax_s64_x(...) __builtin_sve_svmax_s64_x(__VA_ARGS__) -#define svmax_s16_x(...) __builtin_sve_svmax_s16_x(__VA_ARGS__) -#define svmax_s8_z(...) __builtin_sve_svmax_s8_z(__VA_ARGS__) -#define svmax_s32_z(...) __builtin_sve_svmax_s32_z(__VA_ARGS__) -#define svmax_s64_z(...) __builtin_sve_svmax_s64_z(__VA_ARGS__) -#define svmax_s16_z(...) __builtin_sve_svmax_s16_z(__VA_ARGS__) -#define svmax_u8_m(...) __builtin_sve_svmax_u8_m(__VA_ARGS__) -#define svmax_u32_m(...) __builtin_sve_svmax_u32_m(__VA_ARGS__) -#define svmax_u64_m(...) __builtin_sve_svmax_u64_m(__VA_ARGS__) -#define svmax_u16_m(...) __builtin_sve_svmax_u16_m(__VA_ARGS__) -#define svmax_u8_x(...) __builtin_sve_svmax_u8_x(__VA_ARGS__) -#define svmax_u32_x(...) __builtin_sve_svmax_u32_x(__VA_ARGS__) -#define svmax_u64_x(...) __builtin_sve_svmax_u64_x(__VA_ARGS__) -#define svmax_u16_x(...) __builtin_sve_svmax_u16_x(__VA_ARGS__) -#define svmax_u8_z(...) __builtin_sve_svmax_u8_z(__VA_ARGS__) -#define svmax_u32_z(...) __builtin_sve_svmax_u32_z(__VA_ARGS__) -#define svmax_u64_z(...) __builtin_sve_svmax_u64_z(__VA_ARGS__) -#define svmax_u16_z(...) __builtin_sve_svmax_u16_z(__VA_ARGS__) -#define svmaxnm_n_f64_m(...) __builtin_sve_svmaxnm_n_f64_m(__VA_ARGS__) -#define svmaxnm_n_f32_m(...) __builtin_sve_svmaxnm_n_f32_m(__VA_ARGS__) -#define svmaxnm_n_f16_m(...) __builtin_sve_svmaxnm_n_f16_m(__VA_ARGS__) -#define svmaxnm_n_f64_x(...) __builtin_sve_svmaxnm_n_f64_x(__VA_ARGS__) -#define svmaxnm_n_f32_x(...) __builtin_sve_svmaxnm_n_f32_x(__VA_ARGS__) -#define svmaxnm_n_f16_x(...) __builtin_sve_svmaxnm_n_f16_x(__VA_ARGS__) -#define svmaxnm_n_f64_z(...) __builtin_sve_svmaxnm_n_f64_z(__VA_ARGS__) -#define svmaxnm_n_f32_z(...) __builtin_sve_svmaxnm_n_f32_z(__VA_ARGS__) -#define svmaxnm_n_f16_z(...) __builtin_sve_svmaxnm_n_f16_z(__VA_ARGS__) -#define svmaxnm_f64_m(...) __builtin_sve_svmaxnm_f64_m(__VA_ARGS__) -#define svmaxnm_f32_m(...) __builtin_sve_svmaxnm_f32_m(__VA_ARGS__) -#define svmaxnm_f16_m(...) __builtin_sve_svmaxnm_f16_m(__VA_ARGS__) -#define svmaxnm_f64_x(...) __builtin_sve_svmaxnm_f64_x(__VA_ARGS__) -#define svmaxnm_f32_x(...) __builtin_sve_svmaxnm_f32_x(__VA_ARGS__) -#define svmaxnm_f16_x(...) __builtin_sve_svmaxnm_f16_x(__VA_ARGS__) -#define svmaxnm_f64_z(...) __builtin_sve_svmaxnm_f64_z(__VA_ARGS__) -#define svmaxnm_f32_z(...) __builtin_sve_svmaxnm_f32_z(__VA_ARGS__) -#define svmaxnm_f16_z(...) __builtin_sve_svmaxnm_f16_z(__VA_ARGS__) -#define svmaxnmv_f64(...) __builtin_sve_svmaxnmv_f64(__VA_ARGS__) -#define svmaxnmv_f32(...) __builtin_sve_svmaxnmv_f32(__VA_ARGS__) -#define svmaxnmv_f16(...) __builtin_sve_svmaxnmv_f16(__VA_ARGS__) -#define svmaxv_f64(...) __builtin_sve_svmaxv_f64(__VA_ARGS__) -#define svmaxv_f32(...) __builtin_sve_svmaxv_f32(__VA_ARGS__) -#define svmaxv_f16(...) __builtin_sve_svmaxv_f16(__VA_ARGS__) -#define svmaxv_s8(...) __builtin_sve_svmaxv_s8(__VA_ARGS__) -#define svmaxv_s32(...) __builtin_sve_svmaxv_s32(__VA_ARGS__) -#define svmaxv_s64(...) __builtin_sve_svmaxv_s64(__VA_ARGS__) -#define svmaxv_s16(...) __builtin_sve_svmaxv_s16(__VA_ARGS__) -#define svmaxv_u8(...) __builtin_sve_svmaxv_u8(__VA_ARGS__) -#define svmaxv_u32(...) __builtin_sve_svmaxv_u32(__VA_ARGS__) -#define svmaxv_u64(...) __builtin_sve_svmaxv_u64(__VA_ARGS__) -#define svmaxv_u16(...) __builtin_sve_svmaxv_u16(__VA_ARGS__) -#define svmin_n_f64_m(...) __builtin_sve_svmin_n_f64_m(__VA_ARGS__) -#define svmin_n_f32_m(...) __builtin_sve_svmin_n_f32_m(__VA_ARGS__) -#define svmin_n_f16_m(...) __builtin_sve_svmin_n_f16_m(__VA_ARGS__) -#define svmin_n_f64_x(...) __builtin_sve_svmin_n_f64_x(__VA_ARGS__) -#define svmin_n_f32_x(...) __builtin_sve_svmin_n_f32_x(__VA_ARGS__) -#define svmin_n_f16_x(...) __builtin_sve_svmin_n_f16_x(__VA_ARGS__) -#define svmin_n_f64_z(...) __builtin_sve_svmin_n_f64_z(__VA_ARGS__) -#define svmin_n_f32_z(...) __builtin_sve_svmin_n_f32_z(__VA_ARGS__) -#define svmin_n_f16_z(...) __builtin_sve_svmin_n_f16_z(__VA_ARGS__) -#define svmin_n_s8_m(...) __builtin_sve_svmin_n_s8_m(__VA_ARGS__) -#define svmin_n_s32_m(...) __builtin_sve_svmin_n_s32_m(__VA_ARGS__) -#define svmin_n_s64_m(...) __builtin_sve_svmin_n_s64_m(__VA_ARGS__) -#define svmin_n_s16_m(...) __builtin_sve_svmin_n_s16_m(__VA_ARGS__) -#define svmin_n_s8_x(...) __builtin_sve_svmin_n_s8_x(__VA_ARGS__) -#define svmin_n_s32_x(...) __builtin_sve_svmin_n_s32_x(__VA_ARGS__) -#define svmin_n_s64_x(...) __builtin_sve_svmin_n_s64_x(__VA_ARGS__) -#define svmin_n_s16_x(...) __builtin_sve_svmin_n_s16_x(__VA_ARGS__) -#define svmin_n_s8_z(...) __builtin_sve_svmin_n_s8_z(__VA_ARGS__) -#define svmin_n_s32_z(...) __builtin_sve_svmin_n_s32_z(__VA_ARGS__) -#define svmin_n_s64_z(...) __builtin_sve_svmin_n_s64_z(__VA_ARGS__) -#define svmin_n_s16_z(...) __builtin_sve_svmin_n_s16_z(__VA_ARGS__) -#define svmin_n_u8_m(...) __builtin_sve_svmin_n_u8_m(__VA_ARGS__) -#define svmin_n_u32_m(...) __builtin_sve_svmin_n_u32_m(__VA_ARGS__) -#define svmin_n_u64_m(...) __builtin_sve_svmin_n_u64_m(__VA_ARGS__) -#define svmin_n_u16_m(...) __builtin_sve_svmin_n_u16_m(__VA_ARGS__) -#define svmin_n_u8_x(...) __builtin_sve_svmin_n_u8_x(__VA_ARGS__) -#define svmin_n_u32_x(...) __builtin_sve_svmin_n_u32_x(__VA_ARGS__) -#define svmin_n_u64_x(...) __builtin_sve_svmin_n_u64_x(__VA_ARGS__) -#define svmin_n_u16_x(...) __builtin_sve_svmin_n_u16_x(__VA_ARGS__) -#define svmin_n_u8_z(...) __builtin_sve_svmin_n_u8_z(__VA_ARGS__) -#define svmin_n_u32_z(...) __builtin_sve_svmin_n_u32_z(__VA_ARGS__) -#define svmin_n_u64_z(...) __builtin_sve_svmin_n_u64_z(__VA_ARGS__) -#define svmin_n_u16_z(...) __builtin_sve_svmin_n_u16_z(__VA_ARGS__) -#define svmin_f64_m(...) __builtin_sve_svmin_f64_m(__VA_ARGS__) -#define svmin_f32_m(...) __builtin_sve_svmin_f32_m(__VA_ARGS__) -#define svmin_f16_m(...) __builtin_sve_svmin_f16_m(__VA_ARGS__) -#define svmin_f64_x(...) __builtin_sve_svmin_f64_x(__VA_ARGS__) -#define svmin_f32_x(...) __builtin_sve_svmin_f32_x(__VA_ARGS__) -#define svmin_f16_x(...) __builtin_sve_svmin_f16_x(__VA_ARGS__) -#define svmin_f64_z(...) __builtin_sve_svmin_f64_z(__VA_ARGS__) -#define svmin_f32_z(...) __builtin_sve_svmin_f32_z(__VA_ARGS__) -#define svmin_f16_z(...) __builtin_sve_svmin_f16_z(__VA_ARGS__) -#define svmin_s8_m(...) __builtin_sve_svmin_s8_m(__VA_ARGS__) -#define svmin_s32_m(...) __builtin_sve_svmin_s32_m(__VA_ARGS__) -#define svmin_s64_m(...) __builtin_sve_svmin_s64_m(__VA_ARGS__) -#define svmin_s16_m(...) __builtin_sve_svmin_s16_m(__VA_ARGS__) -#define svmin_s8_x(...) __builtin_sve_svmin_s8_x(__VA_ARGS__) -#define svmin_s32_x(...) __builtin_sve_svmin_s32_x(__VA_ARGS__) -#define svmin_s64_x(...) __builtin_sve_svmin_s64_x(__VA_ARGS__) -#define svmin_s16_x(...) __builtin_sve_svmin_s16_x(__VA_ARGS__) -#define svmin_s8_z(...) __builtin_sve_svmin_s8_z(__VA_ARGS__) -#define svmin_s32_z(...) __builtin_sve_svmin_s32_z(__VA_ARGS__) -#define svmin_s64_z(...) __builtin_sve_svmin_s64_z(__VA_ARGS__) -#define svmin_s16_z(...) __builtin_sve_svmin_s16_z(__VA_ARGS__) -#define svmin_u8_m(...) __builtin_sve_svmin_u8_m(__VA_ARGS__) -#define svmin_u32_m(...) __builtin_sve_svmin_u32_m(__VA_ARGS__) -#define svmin_u64_m(...) __builtin_sve_svmin_u64_m(__VA_ARGS__) -#define svmin_u16_m(...) __builtin_sve_svmin_u16_m(__VA_ARGS__) -#define svmin_u8_x(...) __builtin_sve_svmin_u8_x(__VA_ARGS__) -#define svmin_u32_x(...) __builtin_sve_svmin_u32_x(__VA_ARGS__) -#define svmin_u64_x(...) __builtin_sve_svmin_u64_x(__VA_ARGS__) -#define svmin_u16_x(...) __builtin_sve_svmin_u16_x(__VA_ARGS__) -#define svmin_u8_z(...) __builtin_sve_svmin_u8_z(__VA_ARGS__) -#define svmin_u32_z(...) __builtin_sve_svmin_u32_z(__VA_ARGS__) -#define svmin_u64_z(...) __builtin_sve_svmin_u64_z(__VA_ARGS__) -#define svmin_u16_z(...) __builtin_sve_svmin_u16_z(__VA_ARGS__) -#define svminnm_n_f64_m(...) __builtin_sve_svminnm_n_f64_m(__VA_ARGS__) -#define svminnm_n_f32_m(...) __builtin_sve_svminnm_n_f32_m(__VA_ARGS__) -#define svminnm_n_f16_m(...) __builtin_sve_svminnm_n_f16_m(__VA_ARGS__) -#define svminnm_n_f64_x(...) __builtin_sve_svminnm_n_f64_x(__VA_ARGS__) -#define svminnm_n_f32_x(...) __builtin_sve_svminnm_n_f32_x(__VA_ARGS__) -#define svminnm_n_f16_x(...) __builtin_sve_svminnm_n_f16_x(__VA_ARGS__) -#define svminnm_n_f64_z(...) __builtin_sve_svminnm_n_f64_z(__VA_ARGS__) -#define svminnm_n_f32_z(...) __builtin_sve_svminnm_n_f32_z(__VA_ARGS__) -#define svminnm_n_f16_z(...) __builtin_sve_svminnm_n_f16_z(__VA_ARGS__) -#define svminnm_f64_m(...) __builtin_sve_svminnm_f64_m(__VA_ARGS__) -#define svminnm_f32_m(...) __builtin_sve_svminnm_f32_m(__VA_ARGS__) -#define svminnm_f16_m(...) __builtin_sve_svminnm_f16_m(__VA_ARGS__) -#define svminnm_f64_x(...) __builtin_sve_svminnm_f64_x(__VA_ARGS__) -#define svminnm_f32_x(...) __builtin_sve_svminnm_f32_x(__VA_ARGS__) -#define svminnm_f16_x(...) __builtin_sve_svminnm_f16_x(__VA_ARGS__) -#define svminnm_f64_z(...) __builtin_sve_svminnm_f64_z(__VA_ARGS__) -#define svminnm_f32_z(...) __builtin_sve_svminnm_f32_z(__VA_ARGS__) -#define svminnm_f16_z(...) __builtin_sve_svminnm_f16_z(__VA_ARGS__) -#define svminnmv_f64(...) __builtin_sve_svminnmv_f64(__VA_ARGS__) -#define svminnmv_f32(...) __builtin_sve_svminnmv_f32(__VA_ARGS__) -#define svminnmv_f16(...) __builtin_sve_svminnmv_f16(__VA_ARGS__) -#define svminv_f64(...) __builtin_sve_svminv_f64(__VA_ARGS__) -#define svminv_f32(...) __builtin_sve_svminv_f32(__VA_ARGS__) -#define svminv_f16(...) __builtin_sve_svminv_f16(__VA_ARGS__) -#define svminv_s8(...) __builtin_sve_svminv_s8(__VA_ARGS__) -#define svminv_s32(...) __builtin_sve_svminv_s32(__VA_ARGS__) -#define svminv_s64(...) __builtin_sve_svminv_s64(__VA_ARGS__) -#define svminv_s16(...) __builtin_sve_svminv_s16(__VA_ARGS__) -#define svminv_u8(...) __builtin_sve_svminv_u8(__VA_ARGS__) -#define svminv_u32(...) __builtin_sve_svminv_u32(__VA_ARGS__) -#define svminv_u64(...) __builtin_sve_svminv_u64(__VA_ARGS__) -#define svminv_u16(...) __builtin_sve_svminv_u16(__VA_ARGS__) -#define svmla_n_f64_m(...) __builtin_sve_svmla_n_f64_m(__VA_ARGS__) -#define svmla_n_f32_m(...) __builtin_sve_svmla_n_f32_m(__VA_ARGS__) -#define svmla_n_f16_m(...) __builtin_sve_svmla_n_f16_m(__VA_ARGS__) -#define svmla_n_f64_x(...) __builtin_sve_svmla_n_f64_x(__VA_ARGS__) -#define svmla_n_f32_x(...) __builtin_sve_svmla_n_f32_x(__VA_ARGS__) -#define svmla_n_f16_x(...) __builtin_sve_svmla_n_f16_x(__VA_ARGS__) -#define svmla_n_f64_z(...) __builtin_sve_svmla_n_f64_z(__VA_ARGS__) -#define svmla_n_f32_z(...) __builtin_sve_svmla_n_f32_z(__VA_ARGS__) -#define svmla_n_f16_z(...) __builtin_sve_svmla_n_f16_z(__VA_ARGS__) -#define svmla_n_u8_m(...) __builtin_sve_svmla_n_u8_m(__VA_ARGS__) -#define svmla_n_u32_m(...) __builtin_sve_svmla_n_u32_m(__VA_ARGS__) -#define svmla_n_u64_m(...) __builtin_sve_svmla_n_u64_m(__VA_ARGS__) -#define svmla_n_u16_m(...) __builtin_sve_svmla_n_u16_m(__VA_ARGS__) -#define svmla_n_s8_m(...) __builtin_sve_svmla_n_s8_m(__VA_ARGS__) -#define svmla_n_s32_m(...) __builtin_sve_svmla_n_s32_m(__VA_ARGS__) -#define svmla_n_s64_m(...) __builtin_sve_svmla_n_s64_m(__VA_ARGS__) -#define svmla_n_s16_m(...) __builtin_sve_svmla_n_s16_m(__VA_ARGS__) -#define svmla_n_u8_x(...) __builtin_sve_svmla_n_u8_x(__VA_ARGS__) -#define svmla_n_u32_x(...) __builtin_sve_svmla_n_u32_x(__VA_ARGS__) -#define svmla_n_u64_x(...) __builtin_sve_svmla_n_u64_x(__VA_ARGS__) -#define svmla_n_u16_x(...) __builtin_sve_svmla_n_u16_x(__VA_ARGS__) -#define svmla_n_s8_x(...) __builtin_sve_svmla_n_s8_x(__VA_ARGS__) -#define svmla_n_s32_x(...) __builtin_sve_svmla_n_s32_x(__VA_ARGS__) -#define svmla_n_s64_x(...) __builtin_sve_svmla_n_s64_x(__VA_ARGS__) -#define svmla_n_s16_x(...) __builtin_sve_svmla_n_s16_x(__VA_ARGS__) -#define svmla_n_u8_z(...) __builtin_sve_svmla_n_u8_z(__VA_ARGS__) -#define svmla_n_u32_z(...) __builtin_sve_svmla_n_u32_z(__VA_ARGS__) -#define svmla_n_u64_z(...) __builtin_sve_svmla_n_u64_z(__VA_ARGS__) -#define svmla_n_u16_z(...) __builtin_sve_svmla_n_u16_z(__VA_ARGS__) -#define svmla_n_s8_z(...) __builtin_sve_svmla_n_s8_z(__VA_ARGS__) -#define svmla_n_s32_z(...) __builtin_sve_svmla_n_s32_z(__VA_ARGS__) -#define svmla_n_s64_z(...) __builtin_sve_svmla_n_s64_z(__VA_ARGS__) -#define svmla_n_s16_z(...) __builtin_sve_svmla_n_s16_z(__VA_ARGS__) -#define svmla_f64_m(...) __builtin_sve_svmla_f64_m(__VA_ARGS__) -#define svmla_f32_m(...) __builtin_sve_svmla_f32_m(__VA_ARGS__) -#define svmla_f16_m(...) __builtin_sve_svmla_f16_m(__VA_ARGS__) -#define svmla_f64_x(...) __builtin_sve_svmla_f64_x(__VA_ARGS__) -#define svmla_f32_x(...) __builtin_sve_svmla_f32_x(__VA_ARGS__) -#define svmla_f16_x(...) __builtin_sve_svmla_f16_x(__VA_ARGS__) -#define svmla_f64_z(...) __builtin_sve_svmla_f64_z(__VA_ARGS__) -#define svmla_f32_z(...) __builtin_sve_svmla_f32_z(__VA_ARGS__) -#define svmla_f16_z(...) __builtin_sve_svmla_f16_z(__VA_ARGS__) -#define svmla_u8_m(...) __builtin_sve_svmla_u8_m(__VA_ARGS__) -#define svmla_u32_m(...) __builtin_sve_svmla_u32_m(__VA_ARGS__) -#define svmla_u64_m(...) __builtin_sve_svmla_u64_m(__VA_ARGS__) -#define svmla_u16_m(...) __builtin_sve_svmla_u16_m(__VA_ARGS__) -#define svmla_s8_m(...) __builtin_sve_svmla_s8_m(__VA_ARGS__) -#define svmla_s32_m(...) __builtin_sve_svmla_s32_m(__VA_ARGS__) -#define svmla_s64_m(...) __builtin_sve_svmla_s64_m(__VA_ARGS__) -#define svmla_s16_m(...) __builtin_sve_svmla_s16_m(__VA_ARGS__) -#define svmla_u8_x(...) __builtin_sve_svmla_u8_x(__VA_ARGS__) -#define svmla_u32_x(...) __builtin_sve_svmla_u32_x(__VA_ARGS__) -#define svmla_u64_x(...) __builtin_sve_svmla_u64_x(__VA_ARGS__) -#define svmla_u16_x(...) __builtin_sve_svmla_u16_x(__VA_ARGS__) -#define svmla_s8_x(...) __builtin_sve_svmla_s8_x(__VA_ARGS__) -#define svmla_s32_x(...) __builtin_sve_svmla_s32_x(__VA_ARGS__) -#define svmla_s64_x(...) __builtin_sve_svmla_s64_x(__VA_ARGS__) -#define svmla_s16_x(...) __builtin_sve_svmla_s16_x(__VA_ARGS__) -#define svmla_u8_z(...) __builtin_sve_svmla_u8_z(__VA_ARGS__) -#define svmla_u32_z(...) __builtin_sve_svmla_u32_z(__VA_ARGS__) -#define svmla_u64_z(...) __builtin_sve_svmla_u64_z(__VA_ARGS__) -#define svmla_u16_z(...) __builtin_sve_svmla_u16_z(__VA_ARGS__) -#define svmla_s8_z(...) __builtin_sve_svmla_s8_z(__VA_ARGS__) -#define svmla_s32_z(...) __builtin_sve_svmla_s32_z(__VA_ARGS__) -#define svmla_s64_z(...) __builtin_sve_svmla_s64_z(__VA_ARGS__) -#define svmla_s16_z(...) __builtin_sve_svmla_s16_z(__VA_ARGS__) -#define svmla_lane_f64(...) __builtin_sve_svmla_lane_f64(__VA_ARGS__) -#define svmla_lane_f32(...) __builtin_sve_svmla_lane_f32(__VA_ARGS__) -#define svmla_lane_f16(...) __builtin_sve_svmla_lane_f16(__VA_ARGS__) -#define svmls_n_f64_m(...) __builtin_sve_svmls_n_f64_m(__VA_ARGS__) -#define svmls_n_f32_m(...) __builtin_sve_svmls_n_f32_m(__VA_ARGS__) -#define svmls_n_f16_m(...) __builtin_sve_svmls_n_f16_m(__VA_ARGS__) -#define svmls_n_f64_x(...) __builtin_sve_svmls_n_f64_x(__VA_ARGS__) -#define svmls_n_f32_x(...) __builtin_sve_svmls_n_f32_x(__VA_ARGS__) -#define svmls_n_f16_x(...) __builtin_sve_svmls_n_f16_x(__VA_ARGS__) -#define svmls_n_f64_z(...) __builtin_sve_svmls_n_f64_z(__VA_ARGS__) -#define svmls_n_f32_z(...) __builtin_sve_svmls_n_f32_z(__VA_ARGS__) -#define svmls_n_f16_z(...) __builtin_sve_svmls_n_f16_z(__VA_ARGS__) -#define svmls_n_u8_m(...) __builtin_sve_svmls_n_u8_m(__VA_ARGS__) -#define svmls_n_u32_m(...) __builtin_sve_svmls_n_u32_m(__VA_ARGS__) -#define svmls_n_u64_m(...) __builtin_sve_svmls_n_u64_m(__VA_ARGS__) -#define svmls_n_u16_m(...) __builtin_sve_svmls_n_u16_m(__VA_ARGS__) -#define svmls_n_s8_m(...) __builtin_sve_svmls_n_s8_m(__VA_ARGS__) -#define svmls_n_s32_m(...) __builtin_sve_svmls_n_s32_m(__VA_ARGS__) -#define svmls_n_s64_m(...) __builtin_sve_svmls_n_s64_m(__VA_ARGS__) -#define svmls_n_s16_m(...) __builtin_sve_svmls_n_s16_m(__VA_ARGS__) -#define svmls_n_u8_x(...) __builtin_sve_svmls_n_u8_x(__VA_ARGS__) -#define svmls_n_u32_x(...) __builtin_sve_svmls_n_u32_x(__VA_ARGS__) -#define svmls_n_u64_x(...) __builtin_sve_svmls_n_u64_x(__VA_ARGS__) -#define svmls_n_u16_x(...) __builtin_sve_svmls_n_u16_x(__VA_ARGS__) -#define svmls_n_s8_x(...) __builtin_sve_svmls_n_s8_x(__VA_ARGS__) -#define svmls_n_s32_x(...) __builtin_sve_svmls_n_s32_x(__VA_ARGS__) -#define svmls_n_s64_x(...) __builtin_sve_svmls_n_s64_x(__VA_ARGS__) -#define svmls_n_s16_x(...) __builtin_sve_svmls_n_s16_x(__VA_ARGS__) -#define svmls_n_u8_z(...) __builtin_sve_svmls_n_u8_z(__VA_ARGS__) -#define svmls_n_u32_z(...) __builtin_sve_svmls_n_u32_z(__VA_ARGS__) -#define svmls_n_u64_z(...) __builtin_sve_svmls_n_u64_z(__VA_ARGS__) -#define svmls_n_u16_z(...) __builtin_sve_svmls_n_u16_z(__VA_ARGS__) -#define svmls_n_s8_z(...) __builtin_sve_svmls_n_s8_z(__VA_ARGS__) -#define svmls_n_s32_z(...) __builtin_sve_svmls_n_s32_z(__VA_ARGS__) -#define svmls_n_s64_z(...) __builtin_sve_svmls_n_s64_z(__VA_ARGS__) -#define svmls_n_s16_z(...) __builtin_sve_svmls_n_s16_z(__VA_ARGS__) -#define svmls_f64_m(...) __builtin_sve_svmls_f64_m(__VA_ARGS__) -#define svmls_f32_m(...) __builtin_sve_svmls_f32_m(__VA_ARGS__) -#define svmls_f16_m(...) __builtin_sve_svmls_f16_m(__VA_ARGS__) -#define svmls_f64_x(...) __builtin_sve_svmls_f64_x(__VA_ARGS__) -#define svmls_f32_x(...) __builtin_sve_svmls_f32_x(__VA_ARGS__) -#define svmls_f16_x(...) __builtin_sve_svmls_f16_x(__VA_ARGS__) -#define svmls_f64_z(...) __builtin_sve_svmls_f64_z(__VA_ARGS__) -#define svmls_f32_z(...) __builtin_sve_svmls_f32_z(__VA_ARGS__) -#define svmls_f16_z(...) __builtin_sve_svmls_f16_z(__VA_ARGS__) -#define svmls_u8_m(...) __builtin_sve_svmls_u8_m(__VA_ARGS__) -#define svmls_u32_m(...) __builtin_sve_svmls_u32_m(__VA_ARGS__) -#define svmls_u64_m(...) __builtin_sve_svmls_u64_m(__VA_ARGS__) -#define svmls_u16_m(...) __builtin_sve_svmls_u16_m(__VA_ARGS__) -#define svmls_s8_m(...) __builtin_sve_svmls_s8_m(__VA_ARGS__) -#define svmls_s32_m(...) __builtin_sve_svmls_s32_m(__VA_ARGS__) -#define svmls_s64_m(...) __builtin_sve_svmls_s64_m(__VA_ARGS__) -#define svmls_s16_m(...) __builtin_sve_svmls_s16_m(__VA_ARGS__) -#define svmls_u8_x(...) __builtin_sve_svmls_u8_x(__VA_ARGS__) -#define svmls_u32_x(...) __builtin_sve_svmls_u32_x(__VA_ARGS__) -#define svmls_u64_x(...) __builtin_sve_svmls_u64_x(__VA_ARGS__) -#define svmls_u16_x(...) __builtin_sve_svmls_u16_x(__VA_ARGS__) -#define svmls_s8_x(...) __builtin_sve_svmls_s8_x(__VA_ARGS__) -#define svmls_s32_x(...) __builtin_sve_svmls_s32_x(__VA_ARGS__) -#define svmls_s64_x(...) __builtin_sve_svmls_s64_x(__VA_ARGS__) -#define svmls_s16_x(...) __builtin_sve_svmls_s16_x(__VA_ARGS__) -#define svmls_u8_z(...) __builtin_sve_svmls_u8_z(__VA_ARGS__) -#define svmls_u32_z(...) __builtin_sve_svmls_u32_z(__VA_ARGS__) -#define svmls_u64_z(...) __builtin_sve_svmls_u64_z(__VA_ARGS__) -#define svmls_u16_z(...) __builtin_sve_svmls_u16_z(__VA_ARGS__) -#define svmls_s8_z(...) __builtin_sve_svmls_s8_z(__VA_ARGS__) -#define svmls_s32_z(...) __builtin_sve_svmls_s32_z(__VA_ARGS__) -#define svmls_s64_z(...) __builtin_sve_svmls_s64_z(__VA_ARGS__) -#define svmls_s16_z(...) __builtin_sve_svmls_s16_z(__VA_ARGS__) -#define svmls_lane_f64(...) __builtin_sve_svmls_lane_f64(__VA_ARGS__) -#define svmls_lane_f32(...) __builtin_sve_svmls_lane_f32(__VA_ARGS__) -#define svmls_lane_f16(...) __builtin_sve_svmls_lane_f16(__VA_ARGS__) -#define svmov_b_z(...) __builtin_sve_svmov_b_z(__VA_ARGS__) -#define svmsb_n_f64_m(...) __builtin_sve_svmsb_n_f64_m(__VA_ARGS__) -#define svmsb_n_f32_m(...) __builtin_sve_svmsb_n_f32_m(__VA_ARGS__) -#define svmsb_n_f16_m(...) __builtin_sve_svmsb_n_f16_m(__VA_ARGS__) -#define svmsb_n_f64_x(...) __builtin_sve_svmsb_n_f64_x(__VA_ARGS__) -#define svmsb_n_f32_x(...) __builtin_sve_svmsb_n_f32_x(__VA_ARGS__) -#define svmsb_n_f16_x(...) __builtin_sve_svmsb_n_f16_x(__VA_ARGS__) -#define svmsb_n_f64_z(...) __builtin_sve_svmsb_n_f64_z(__VA_ARGS__) -#define svmsb_n_f32_z(...) __builtin_sve_svmsb_n_f32_z(__VA_ARGS__) -#define svmsb_n_f16_z(...) __builtin_sve_svmsb_n_f16_z(__VA_ARGS__) -#define svmsb_n_u8_m(...) __builtin_sve_svmsb_n_u8_m(__VA_ARGS__) -#define svmsb_n_u32_m(...) __builtin_sve_svmsb_n_u32_m(__VA_ARGS__) -#define svmsb_n_u64_m(...) __builtin_sve_svmsb_n_u64_m(__VA_ARGS__) -#define svmsb_n_u16_m(...) __builtin_sve_svmsb_n_u16_m(__VA_ARGS__) -#define svmsb_n_s8_m(...) __builtin_sve_svmsb_n_s8_m(__VA_ARGS__) -#define svmsb_n_s32_m(...) __builtin_sve_svmsb_n_s32_m(__VA_ARGS__) -#define svmsb_n_s64_m(...) __builtin_sve_svmsb_n_s64_m(__VA_ARGS__) -#define svmsb_n_s16_m(...) __builtin_sve_svmsb_n_s16_m(__VA_ARGS__) -#define svmsb_n_u8_x(...) __builtin_sve_svmsb_n_u8_x(__VA_ARGS__) -#define svmsb_n_u32_x(...) __builtin_sve_svmsb_n_u32_x(__VA_ARGS__) -#define svmsb_n_u64_x(...) __builtin_sve_svmsb_n_u64_x(__VA_ARGS__) -#define svmsb_n_u16_x(...) __builtin_sve_svmsb_n_u16_x(__VA_ARGS__) -#define svmsb_n_s8_x(...) __builtin_sve_svmsb_n_s8_x(__VA_ARGS__) -#define svmsb_n_s32_x(...) __builtin_sve_svmsb_n_s32_x(__VA_ARGS__) -#define svmsb_n_s64_x(...) __builtin_sve_svmsb_n_s64_x(__VA_ARGS__) -#define svmsb_n_s16_x(...) __builtin_sve_svmsb_n_s16_x(__VA_ARGS__) -#define svmsb_n_u8_z(...) __builtin_sve_svmsb_n_u8_z(__VA_ARGS__) -#define svmsb_n_u32_z(...) __builtin_sve_svmsb_n_u32_z(__VA_ARGS__) -#define svmsb_n_u64_z(...) __builtin_sve_svmsb_n_u64_z(__VA_ARGS__) -#define svmsb_n_u16_z(...) __builtin_sve_svmsb_n_u16_z(__VA_ARGS__) -#define svmsb_n_s8_z(...) __builtin_sve_svmsb_n_s8_z(__VA_ARGS__) -#define svmsb_n_s32_z(...) __builtin_sve_svmsb_n_s32_z(__VA_ARGS__) -#define svmsb_n_s64_z(...) __builtin_sve_svmsb_n_s64_z(__VA_ARGS__) -#define svmsb_n_s16_z(...) __builtin_sve_svmsb_n_s16_z(__VA_ARGS__) -#define svmsb_f64_m(...) __builtin_sve_svmsb_f64_m(__VA_ARGS__) -#define svmsb_f32_m(...) __builtin_sve_svmsb_f32_m(__VA_ARGS__) -#define svmsb_f16_m(...) __builtin_sve_svmsb_f16_m(__VA_ARGS__) -#define svmsb_f64_x(...) __builtin_sve_svmsb_f64_x(__VA_ARGS__) -#define svmsb_f32_x(...) __builtin_sve_svmsb_f32_x(__VA_ARGS__) -#define svmsb_f16_x(...) __builtin_sve_svmsb_f16_x(__VA_ARGS__) -#define svmsb_f64_z(...) __builtin_sve_svmsb_f64_z(__VA_ARGS__) -#define svmsb_f32_z(...) __builtin_sve_svmsb_f32_z(__VA_ARGS__) -#define svmsb_f16_z(...) __builtin_sve_svmsb_f16_z(__VA_ARGS__) -#define svmsb_u8_m(...) __builtin_sve_svmsb_u8_m(__VA_ARGS__) -#define svmsb_u32_m(...) __builtin_sve_svmsb_u32_m(__VA_ARGS__) -#define svmsb_u64_m(...) __builtin_sve_svmsb_u64_m(__VA_ARGS__) -#define svmsb_u16_m(...) __builtin_sve_svmsb_u16_m(__VA_ARGS__) -#define svmsb_s8_m(...) __builtin_sve_svmsb_s8_m(__VA_ARGS__) -#define svmsb_s32_m(...) __builtin_sve_svmsb_s32_m(__VA_ARGS__) -#define svmsb_s64_m(...) __builtin_sve_svmsb_s64_m(__VA_ARGS__) -#define svmsb_s16_m(...) __builtin_sve_svmsb_s16_m(__VA_ARGS__) -#define svmsb_u8_x(...) __builtin_sve_svmsb_u8_x(__VA_ARGS__) -#define svmsb_u32_x(...) __builtin_sve_svmsb_u32_x(__VA_ARGS__) -#define svmsb_u64_x(...) __builtin_sve_svmsb_u64_x(__VA_ARGS__) -#define svmsb_u16_x(...) __builtin_sve_svmsb_u16_x(__VA_ARGS__) -#define svmsb_s8_x(...) __builtin_sve_svmsb_s8_x(__VA_ARGS__) -#define svmsb_s32_x(...) __builtin_sve_svmsb_s32_x(__VA_ARGS__) -#define svmsb_s64_x(...) __builtin_sve_svmsb_s64_x(__VA_ARGS__) -#define svmsb_s16_x(...) __builtin_sve_svmsb_s16_x(__VA_ARGS__) -#define svmsb_u8_z(...) __builtin_sve_svmsb_u8_z(__VA_ARGS__) -#define svmsb_u32_z(...) __builtin_sve_svmsb_u32_z(__VA_ARGS__) -#define svmsb_u64_z(...) __builtin_sve_svmsb_u64_z(__VA_ARGS__) -#define svmsb_u16_z(...) __builtin_sve_svmsb_u16_z(__VA_ARGS__) -#define svmsb_s8_z(...) __builtin_sve_svmsb_s8_z(__VA_ARGS__) -#define svmsb_s32_z(...) __builtin_sve_svmsb_s32_z(__VA_ARGS__) -#define svmsb_s64_z(...) __builtin_sve_svmsb_s64_z(__VA_ARGS__) -#define svmsb_s16_z(...) __builtin_sve_svmsb_s16_z(__VA_ARGS__) -#define svmul_n_f64_m(...) __builtin_sve_svmul_n_f64_m(__VA_ARGS__) -#define svmul_n_f32_m(...) __builtin_sve_svmul_n_f32_m(__VA_ARGS__) -#define svmul_n_f16_m(...) __builtin_sve_svmul_n_f16_m(__VA_ARGS__) -#define svmul_n_f64_x(...) __builtin_sve_svmul_n_f64_x(__VA_ARGS__) -#define svmul_n_f32_x(...) __builtin_sve_svmul_n_f32_x(__VA_ARGS__) -#define svmul_n_f16_x(...) __builtin_sve_svmul_n_f16_x(__VA_ARGS__) -#define svmul_n_f64_z(...) __builtin_sve_svmul_n_f64_z(__VA_ARGS__) -#define svmul_n_f32_z(...) __builtin_sve_svmul_n_f32_z(__VA_ARGS__) -#define svmul_n_f16_z(...) __builtin_sve_svmul_n_f16_z(__VA_ARGS__) -#define svmul_n_u8_m(...) __builtin_sve_svmul_n_u8_m(__VA_ARGS__) -#define svmul_n_u32_m(...) __builtin_sve_svmul_n_u32_m(__VA_ARGS__) -#define svmul_n_u64_m(...) __builtin_sve_svmul_n_u64_m(__VA_ARGS__) -#define svmul_n_u16_m(...) __builtin_sve_svmul_n_u16_m(__VA_ARGS__) -#define svmul_n_s8_m(...) __builtin_sve_svmul_n_s8_m(__VA_ARGS__) -#define svmul_n_s32_m(...) __builtin_sve_svmul_n_s32_m(__VA_ARGS__) -#define svmul_n_s64_m(...) __builtin_sve_svmul_n_s64_m(__VA_ARGS__) -#define svmul_n_s16_m(...) __builtin_sve_svmul_n_s16_m(__VA_ARGS__) -#define svmul_n_u8_x(...) __builtin_sve_svmul_n_u8_x(__VA_ARGS__) -#define svmul_n_u32_x(...) __builtin_sve_svmul_n_u32_x(__VA_ARGS__) -#define svmul_n_u64_x(...) __builtin_sve_svmul_n_u64_x(__VA_ARGS__) -#define svmul_n_u16_x(...) __builtin_sve_svmul_n_u16_x(__VA_ARGS__) -#define svmul_n_s8_x(...) __builtin_sve_svmul_n_s8_x(__VA_ARGS__) -#define svmul_n_s32_x(...) __builtin_sve_svmul_n_s32_x(__VA_ARGS__) -#define svmul_n_s64_x(...) __builtin_sve_svmul_n_s64_x(__VA_ARGS__) -#define svmul_n_s16_x(...) __builtin_sve_svmul_n_s16_x(__VA_ARGS__) -#define svmul_n_u8_z(...) __builtin_sve_svmul_n_u8_z(__VA_ARGS__) -#define svmul_n_u32_z(...) __builtin_sve_svmul_n_u32_z(__VA_ARGS__) -#define svmul_n_u64_z(...) __builtin_sve_svmul_n_u64_z(__VA_ARGS__) -#define svmul_n_u16_z(...) __builtin_sve_svmul_n_u16_z(__VA_ARGS__) -#define svmul_n_s8_z(...) __builtin_sve_svmul_n_s8_z(__VA_ARGS__) -#define svmul_n_s32_z(...) __builtin_sve_svmul_n_s32_z(__VA_ARGS__) -#define svmul_n_s64_z(...) __builtin_sve_svmul_n_s64_z(__VA_ARGS__) -#define svmul_n_s16_z(...) __builtin_sve_svmul_n_s16_z(__VA_ARGS__) -#define svmul_f64_m(...) __builtin_sve_svmul_f64_m(__VA_ARGS__) -#define svmul_f32_m(...) __builtin_sve_svmul_f32_m(__VA_ARGS__) -#define svmul_f16_m(...) __builtin_sve_svmul_f16_m(__VA_ARGS__) -#define svmul_f64_x(...) __builtin_sve_svmul_f64_x(__VA_ARGS__) -#define svmul_f32_x(...) __builtin_sve_svmul_f32_x(__VA_ARGS__) -#define svmul_f16_x(...) __builtin_sve_svmul_f16_x(__VA_ARGS__) -#define svmul_f64_z(...) __builtin_sve_svmul_f64_z(__VA_ARGS__) -#define svmul_f32_z(...) __builtin_sve_svmul_f32_z(__VA_ARGS__) -#define svmul_f16_z(...) __builtin_sve_svmul_f16_z(__VA_ARGS__) -#define svmul_u8_m(...) __builtin_sve_svmul_u8_m(__VA_ARGS__) -#define svmul_u32_m(...) __builtin_sve_svmul_u32_m(__VA_ARGS__) -#define svmul_u64_m(...) __builtin_sve_svmul_u64_m(__VA_ARGS__) -#define svmul_u16_m(...) __builtin_sve_svmul_u16_m(__VA_ARGS__) -#define svmul_s8_m(...) __builtin_sve_svmul_s8_m(__VA_ARGS__) -#define svmul_s32_m(...) __builtin_sve_svmul_s32_m(__VA_ARGS__) -#define svmul_s64_m(...) __builtin_sve_svmul_s64_m(__VA_ARGS__) -#define svmul_s16_m(...) __builtin_sve_svmul_s16_m(__VA_ARGS__) -#define svmul_u8_x(...) __builtin_sve_svmul_u8_x(__VA_ARGS__) -#define svmul_u32_x(...) __builtin_sve_svmul_u32_x(__VA_ARGS__) -#define svmul_u64_x(...) __builtin_sve_svmul_u64_x(__VA_ARGS__) -#define svmul_u16_x(...) __builtin_sve_svmul_u16_x(__VA_ARGS__) -#define svmul_s8_x(...) __builtin_sve_svmul_s8_x(__VA_ARGS__) -#define svmul_s32_x(...) __builtin_sve_svmul_s32_x(__VA_ARGS__) -#define svmul_s64_x(...) __builtin_sve_svmul_s64_x(__VA_ARGS__) -#define svmul_s16_x(...) __builtin_sve_svmul_s16_x(__VA_ARGS__) -#define svmul_u8_z(...) __builtin_sve_svmul_u8_z(__VA_ARGS__) -#define svmul_u32_z(...) __builtin_sve_svmul_u32_z(__VA_ARGS__) -#define svmul_u64_z(...) __builtin_sve_svmul_u64_z(__VA_ARGS__) -#define svmul_u16_z(...) __builtin_sve_svmul_u16_z(__VA_ARGS__) -#define svmul_s8_z(...) __builtin_sve_svmul_s8_z(__VA_ARGS__) -#define svmul_s32_z(...) __builtin_sve_svmul_s32_z(__VA_ARGS__) -#define svmul_s64_z(...) __builtin_sve_svmul_s64_z(__VA_ARGS__) -#define svmul_s16_z(...) __builtin_sve_svmul_s16_z(__VA_ARGS__) -#define svmul_lane_f64(...) __builtin_sve_svmul_lane_f64(__VA_ARGS__) -#define svmul_lane_f32(...) __builtin_sve_svmul_lane_f32(__VA_ARGS__) -#define svmul_lane_f16(...) __builtin_sve_svmul_lane_f16(__VA_ARGS__) -#define svmulh_n_s8_m(...) __builtin_sve_svmulh_n_s8_m(__VA_ARGS__) -#define svmulh_n_s32_m(...) __builtin_sve_svmulh_n_s32_m(__VA_ARGS__) -#define svmulh_n_s64_m(...) __builtin_sve_svmulh_n_s64_m(__VA_ARGS__) -#define svmulh_n_s16_m(...) __builtin_sve_svmulh_n_s16_m(__VA_ARGS__) -#define svmulh_n_s8_x(...) __builtin_sve_svmulh_n_s8_x(__VA_ARGS__) -#define svmulh_n_s32_x(...) __builtin_sve_svmulh_n_s32_x(__VA_ARGS__) -#define svmulh_n_s64_x(...) __builtin_sve_svmulh_n_s64_x(__VA_ARGS__) -#define svmulh_n_s16_x(...) __builtin_sve_svmulh_n_s16_x(__VA_ARGS__) -#define svmulh_n_s8_z(...) __builtin_sve_svmulh_n_s8_z(__VA_ARGS__) -#define svmulh_n_s32_z(...) __builtin_sve_svmulh_n_s32_z(__VA_ARGS__) -#define svmulh_n_s64_z(...) __builtin_sve_svmulh_n_s64_z(__VA_ARGS__) -#define svmulh_n_s16_z(...) __builtin_sve_svmulh_n_s16_z(__VA_ARGS__) -#define svmulh_n_u8_m(...) __builtin_sve_svmulh_n_u8_m(__VA_ARGS__) -#define svmulh_n_u32_m(...) __builtin_sve_svmulh_n_u32_m(__VA_ARGS__) -#define svmulh_n_u64_m(...) __builtin_sve_svmulh_n_u64_m(__VA_ARGS__) -#define svmulh_n_u16_m(...) __builtin_sve_svmulh_n_u16_m(__VA_ARGS__) -#define svmulh_n_u8_x(...) __builtin_sve_svmulh_n_u8_x(__VA_ARGS__) -#define svmulh_n_u32_x(...) __builtin_sve_svmulh_n_u32_x(__VA_ARGS__) -#define svmulh_n_u64_x(...) __builtin_sve_svmulh_n_u64_x(__VA_ARGS__) -#define svmulh_n_u16_x(...) __builtin_sve_svmulh_n_u16_x(__VA_ARGS__) -#define svmulh_n_u8_z(...) __builtin_sve_svmulh_n_u8_z(__VA_ARGS__) -#define svmulh_n_u32_z(...) __builtin_sve_svmulh_n_u32_z(__VA_ARGS__) -#define svmulh_n_u64_z(...) __builtin_sve_svmulh_n_u64_z(__VA_ARGS__) -#define svmulh_n_u16_z(...) __builtin_sve_svmulh_n_u16_z(__VA_ARGS__) -#define svmulh_s8_m(...) __builtin_sve_svmulh_s8_m(__VA_ARGS__) -#define svmulh_s32_m(...) __builtin_sve_svmulh_s32_m(__VA_ARGS__) -#define svmulh_s64_m(...) __builtin_sve_svmulh_s64_m(__VA_ARGS__) -#define svmulh_s16_m(...) __builtin_sve_svmulh_s16_m(__VA_ARGS__) -#define svmulh_s8_x(...) __builtin_sve_svmulh_s8_x(__VA_ARGS__) -#define svmulh_s32_x(...) __builtin_sve_svmulh_s32_x(__VA_ARGS__) -#define svmulh_s64_x(...) __builtin_sve_svmulh_s64_x(__VA_ARGS__) -#define svmulh_s16_x(...) __builtin_sve_svmulh_s16_x(__VA_ARGS__) -#define svmulh_s8_z(...) __builtin_sve_svmulh_s8_z(__VA_ARGS__) -#define svmulh_s32_z(...) __builtin_sve_svmulh_s32_z(__VA_ARGS__) -#define svmulh_s64_z(...) __builtin_sve_svmulh_s64_z(__VA_ARGS__) -#define svmulh_s16_z(...) __builtin_sve_svmulh_s16_z(__VA_ARGS__) -#define svmulh_u8_m(...) __builtin_sve_svmulh_u8_m(__VA_ARGS__) -#define svmulh_u32_m(...) __builtin_sve_svmulh_u32_m(__VA_ARGS__) -#define svmulh_u64_m(...) __builtin_sve_svmulh_u64_m(__VA_ARGS__) -#define svmulh_u16_m(...) __builtin_sve_svmulh_u16_m(__VA_ARGS__) -#define svmulh_u8_x(...) __builtin_sve_svmulh_u8_x(__VA_ARGS__) -#define svmulh_u32_x(...) __builtin_sve_svmulh_u32_x(__VA_ARGS__) -#define svmulh_u64_x(...) __builtin_sve_svmulh_u64_x(__VA_ARGS__) -#define svmulh_u16_x(...) __builtin_sve_svmulh_u16_x(__VA_ARGS__) -#define svmulh_u8_z(...) __builtin_sve_svmulh_u8_z(__VA_ARGS__) -#define svmulh_u32_z(...) __builtin_sve_svmulh_u32_z(__VA_ARGS__) -#define svmulh_u64_z(...) __builtin_sve_svmulh_u64_z(__VA_ARGS__) -#define svmulh_u16_z(...) __builtin_sve_svmulh_u16_z(__VA_ARGS__) -#define svmulx_n_f64_m(...) __builtin_sve_svmulx_n_f64_m(__VA_ARGS__) -#define svmulx_n_f32_m(...) __builtin_sve_svmulx_n_f32_m(__VA_ARGS__) -#define svmulx_n_f16_m(...) __builtin_sve_svmulx_n_f16_m(__VA_ARGS__) -#define svmulx_n_f64_x(...) __builtin_sve_svmulx_n_f64_x(__VA_ARGS__) -#define svmulx_n_f32_x(...) __builtin_sve_svmulx_n_f32_x(__VA_ARGS__) -#define svmulx_n_f16_x(...) __builtin_sve_svmulx_n_f16_x(__VA_ARGS__) -#define svmulx_n_f64_z(...) __builtin_sve_svmulx_n_f64_z(__VA_ARGS__) -#define svmulx_n_f32_z(...) __builtin_sve_svmulx_n_f32_z(__VA_ARGS__) -#define svmulx_n_f16_z(...) __builtin_sve_svmulx_n_f16_z(__VA_ARGS__) -#define svmulx_f64_m(...) __builtin_sve_svmulx_f64_m(__VA_ARGS__) -#define svmulx_f32_m(...) __builtin_sve_svmulx_f32_m(__VA_ARGS__) -#define svmulx_f16_m(...) __builtin_sve_svmulx_f16_m(__VA_ARGS__) -#define svmulx_f64_x(...) __builtin_sve_svmulx_f64_x(__VA_ARGS__) -#define svmulx_f32_x(...) __builtin_sve_svmulx_f32_x(__VA_ARGS__) -#define svmulx_f16_x(...) __builtin_sve_svmulx_f16_x(__VA_ARGS__) -#define svmulx_f64_z(...) __builtin_sve_svmulx_f64_z(__VA_ARGS__) -#define svmulx_f32_z(...) __builtin_sve_svmulx_f32_z(__VA_ARGS__) -#define svmulx_f16_z(...) __builtin_sve_svmulx_f16_z(__VA_ARGS__) -#define svnand_b_z(...) __builtin_sve_svnand_b_z(__VA_ARGS__) -#define svneg_f64_m(...) __builtin_sve_svneg_f64_m(__VA_ARGS__) -#define svneg_f32_m(...) __builtin_sve_svneg_f32_m(__VA_ARGS__) -#define svneg_f16_m(...) __builtin_sve_svneg_f16_m(__VA_ARGS__) -#define svneg_f64_x(...) __builtin_sve_svneg_f64_x(__VA_ARGS__) -#define svneg_f32_x(...) __builtin_sve_svneg_f32_x(__VA_ARGS__) -#define svneg_f16_x(...) __builtin_sve_svneg_f16_x(__VA_ARGS__) -#define svneg_f64_z(...) __builtin_sve_svneg_f64_z(__VA_ARGS__) -#define svneg_f32_z(...) __builtin_sve_svneg_f32_z(__VA_ARGS__) -#define svneg_f16_z(...) __builtin_sve_svneg_f16_z(__VA_ARGS__) -#define svneg_s8_m(...) __builtin_sve_svneg_s8_m(__VA_ARGS__) -#define svneg_s32_m(...) __builtin_sve_svneg_s32_m(__VA_ARGS__) -#define svneg_s64_m(...) __builtin_sve_svneg_s64_m(__VA_ARGS__) -#define svneg_s16_m(...) __builtin_sve_svneg_s16_m(__VA_ARGS__) -#define svneg_s8_x(...) __builtin_sve_svneg_s8_x(__VA_ARGS__) -#define svneg_s32_x(...) __builtin_sve_svneg_s32_x(__VA_ARGS__) -#define svneg_s64_x(...) __builtin_sve_svneg_s64_x(__VA_ARGS__) -#define svneg_s16_x(...) __builtin_sve_svneg_s16_x(__VA_ARGS__) -#define svneg_s8_z(...) __builtin_sve_svneg_s8_z(__VA_ARGS__) -#define svneg_s32_z(...) __builtin_sve_svneg_s32_z(__VA_ARGS__) -#define svneg_s64_z(...) __builtin_sve_svneg_s64_z(__VA_ARGS__) -#define svneg_s16_z(...) __builtin_sve_svneg_s16_z(__VA_ARGS__) -#define svnmad_n_f64_m(...) __builtin_sve_svnmad_n_f64_m(__VA_ARGS__) -#define svnmad_n_f32_m(...) __builtin_sve_svnmad_n_f32_m(__VA_ARGS__) -#define svnmad_n_f16_m(...) __builtin_sve_svnmad_n_f16_m(__VA_ARGS__) -#define svnmad_n_f64_x(...) __builtin_sve_svnmad_n_f64_x(__VA_ARGS__) -#define svnmad_n_f32_x(...) __builtin_sve_svnmad_n_f32_x(__VA_ARGS__) -#define svnmad_n_f16_x(...) __builtin_sve_svnmad_n_f16_x(__VA_ARGS__) -#define svnmad_n_f64_z(...) __builtin_sve_svnmad_n_f64_z(__VA_ARGS__) -#define svnmad_n_f32_z(...) __builtin_sve_svnmad_n_f32_z(__VA_ARGS__) -#define svnmad_n_f16_z(...) __builtin_sve_svnmad_n_f16_z(__VA_ARGS__) -#define svnmad_f64_m(...) __builtin_sve_svnmad_f64_m(__VA_ARGS__) -#define svnmad_f32_m(...) __builtin_sve_svnmad_f32_m(__VA_ARGS__) -#define svnmad_f16_m(...) __builtin_sve_svnmad_f16_m(__VA_ARGS__) -#define svnmad_f64_x(...) __builtin_sve_svnmad_f64_x(__VA_ARGS__) -#define svnmad_f32_x(...) __builtin_sve_svnmad_f32_x(__VA_ARGS__) -#define svnmad_f16_x(...) __builtin_sve_svnmad_f16_x(__VA_ARGS__) -#define svnmad_f64_z(...) __builtin_sve_svnmad_f64_z(__VA_ARGS__) -#define svnmad_f32_z(...) __builtin_sve_svnmad_f32_z(__VA_ARGS__) -#define svnmad_f16_z(...) __builtin_sve_svnmad_f16_z(__VA_ARGS__) -#define svnmla_n_f64_m(...) __builtin_sve_svnmla_n_f64_m(__VA_ARGS__) -#define svnmla_n_f32_m(...) __builtin_sve_svnmla_n_f32_m(__VA_ARGS__) -#define svnmla_n_f16_m(...) __builtin_sve_svnmla_n_f16_m(__VA_ARGS__) -#define svnmla_n_f64_x(...) __builtin_sve_svnmla_n_f64_x(__VA_ARGS__) -#define svnmla_n_f32_x(...) __builtin_sve_svnmla_n_f32_x(__VA_ARGS__) -#define svnmla_n_f16_x(...) __builtin_sve_svnmla_n_f16_x(__VA_ARGS__) -#define svnmla_n_f64_z(...) __builtin_sve_svnmla_n_f64_z(__VA_ARGS__) -#define svnmla_n_f32_z(...) __builtin_sve_svnmla_n_f32_z(__VA_ARGS__) -#define svnmla_n_f16_z(...) __builtin_sve_svnmla_n_f16_z(__VA_ARGS__) -#define svnmla_f64_m(...) __builtin_sve_svnmla_f64_m(__VA_ARGS__) -#define svnmla_f32_m(...) __builtin_sve_svnmla_f32_m(__VA_ARGS__) -#define svnmla_f16_m(...) __builtin_sve_svnmla_f16_m(__VA_ARGS__) -#define svnmla_f64_x(...) __builtin_sve_svnmla_f64_x(__VA_ARGS__) -#define svnmla_f32_x(...) __builtin_sve_svnmla_f32_x(__VA_ARGS__) -#define svnmla_f16_x(...) __builtin_sve_svnmla_f16_x(__VA_ARGS__) -#define svnmla_f64_z(...) __builtin_sve_svnmla_f64_z(__VA_ARGS__) -#define svnmla_f32_z(...) __builtin_sve_svnmla_f32_z(__VA_ARGS__) -#define svnmla_f16_z(...) __builtin_sve_svnmla_f16_z(__VA_ARGS__) -#define svnmls_n_f64_m(...) __builtin_sve_svnmls_n_f64_m(__VA_ARGS__) -#define svnmls_n_f32_m(...) __builtin_sve_svnmls_n_f32_m(__VA_ARGS__) -#define svnmls_n_f16_m(...) __builtin_sve_svnmls_n_f16_m(__VA_ARGS__) -#define svnmls_n_f64_x(...) __builtin_sve_svnmls_n_f64_x(__VA_ARGS__) -#define svnmls_n_f32_x(...) __builtin_sve_svnmls_n_f32_x(__VA_ARGS__) -#define svnmls_n_f16_x(...) __builtin_sve_svnmls_n_f16_x(__VA_ARGS__) -#define svnmls_n_f64_z(...) __builtin_sve_svnmls_n_f64_z(__VA_ARGS__) -#define svnmls_n_f32_z(...) __builtin_sve_svnmls_n_f32_z(__VA_ARGS__) -#define svnmls_n_f16_z(...) __builtin_sve_svnmls_n_f16_z(__VA_ARGS__) -#define svnmls_f64_m(...) __builtin_sve_svnmls_f64_m(__VA_ARGS__) -#define svnmls_f32_m(...) __builtin_sve_svnmls_f32_m(__VA_ARGS__) -#define svnmls_f16_m(...) __builtin_sve_svnmls_f16_m(__VA_ARGS__) -#define svnmls_f64_x(...) __builtin_sve_svnmls_f64_x(__VA_ARGS__) -#define svnmls_f32_x(...) __builtin_sve_svnmls_f32_x(__VA_ARGS__) -#define svnmls_f16_x(...) __builtin_sve_svnmls_f16_x(__VA_ARGS__) -#define svnmls_f64_z(...) __builtin_sve_svnmls_f64_z(__VA_ARGS__) -#define svnmls_f32_z(...) __builtin_sve_svnmls_f32_z(__VA_ARGS__) -#define svnmls_f16_z(...) __builtin_sve_svnmls_f16_z(__VA_ARGS__) -#define svnmsb_n_f64_m(...) __builtin_sve_svnmsb_n_f64_m(__VA_ARGS__) -#define svnmsb_n_f32_m(...) __builtin_sve_svnmsb_n_f32_m(__VA_ARGS__) -#define svnmsb_n_f16_m(...) __builtin_sve_svnmsb_n_f16_m(__VA_ARGS__) -#define svnmsb_n_f64_x(...) __builtin_sve_svnmsb_n_f64_x(__VA_ARGS__) -#define svnmsb_n_f32_x(...) __builtin_sve_svnmsb_n_f32_x(__VA_ARGS__) -#define svnmsb_n_f16_x(...) __builtin_sve_svnmsb_n_f16_x(__VA_ARGS__) -#define svnmsb_n_f64_z(...) __builtin_sve_svnmsb_n_f64_z(__VA_ARGS__) -#define svnmsb_n_f32_z(...) __builtin_sve_svnmsb_n_f32_z(__VA_ARGS__) -#define svnmsb_n_f16_z(...) __builtin_sve_svnmsb_n_f16_z(__VA_ARGS__) -#define svnmsb_f64_m(...) __builtin_sve_svnmsb_f64_m(__VA_ARGS__) -#define svnmsb_f32_m(...) __builtin_sve_svnmsb_f32_m(__VA_ARGS__) -#define svnmsb_f16_m(...) __builtin_sve_svnmsb_f16_m(__VA_ARGS__) -#define svnmsb_f64_x(...) __builtin_sve_svnmsb_f64_x(__VA_ARGS__) -#define svnmsb_f32_x(...) __builtin_sve_svnmsb_f32_x(__VA_ARGS__) -#define svnmsb_f16_x(...) __builtin_sve_svnmsb_f16_x(__VA_ARGS__) -#define svnmsb_f64_z(...) __builtin_sve_svnmsb_f64_z(__VA_ARGS__) -#define svnmsb_f32_z(...) __builtin_sve_svnmsb_f32_z(__VA_ARGS__) -#define svnmsb_f16_z(...) __builtin_sve_svnmsb_f16_z(__VA_ARGS__) -#define svnor_b_z(...) __builtin_sve_svnor_b_z(__VA_ARGS__) -#define svnot_b_z(...) __builtin_sve_svnot_b_z(__VA_ARGS__) -#define svnot_u8_m(...) __builtin_sve_svnot_u8_m(__VA_ARGS__) -#define svnot_u32_m(...) __builtin_sve_svnot_u32_m(__VA_ARGS__) -#define svnot_u64_m(...) __builtin_sve_svnot_u64_m(__VA_ARGS__) -#define svnot_u16_m(...) __builtin_sve_svnot_u16_m(__VA_ARGS__) -#define svnot_s8_m(...) __builtin_sve_svnot_s8_m(__VA_ARGS__) -#define svnot_s32_m(...) __builtin_sve_svnot_s32_m(__VA_ARGS__) -#define svnot_s64_m(...) __builtin_sve_svnot_s64_m(__VA_ARGS__) -#define svnot_s16_m(...) __builtin_sve_svnot_s16_m(__VA_ARGS__) -#define svnot_u8_x(...) __builtin_sve_svnot_u8_x(__VA_ARGS__) -#define svnot_u32_x(...) __builtin_sve_svnot_u32_x(__VA_ARGS__) -#define svnot_u64_x(...) __builtin_sve_svnot_u64_x(__VA_ARGS__) -#define svnot_u16_x(...) __builtin_sve_svnot_u16_x(__VA_ARGS__) -#define svnot_s8_x(...) __builtin_sve_svnot_s8_x(__VA_ARGS__) -#define svnot_s32_x(...) __builtin_sve_svnot_s32_x(__VA_ARGS__) -#define svnot_s64_x(...) __builtin_sve_svnot_s64_x(__VA_ARGS__) -#define svnot_s16_x(...) __builtin_sve_svnot_s16_x(__VA_ARGS__) -#define svnot_u8_z(...) __builtin_sve_svnot_u8_z(__VA_ARGS__) -#define svnot_u32_z(...) __builtin_sve_svnot_u32_z(__VA_ARGS__) -#define svnot_u64_z(...) __builtin_sve_svnot_u64_z(__VA_ARGS__) -#define svnot_u16_z(...) __builtin_sve_svnot_u16_z(__VA_ARGS__) -#define svnot_s8_z(...) __builtin_sve_svnot_s8_z(__VA_ARGS__) -#define svnot_s32_z(...) __builtin_sve_svnot_s32_z(__VA_ARGS__) -#define svnot_s64_z(...) __builtin_sve_svnot_s64_z(__VA_ARGS__) -#define svnot_s16_z(...) __builtin_sve_svnot_s16_z(__VA_ARGS__) -#define svorn_b_z(...) __builtin_sve_svorn_b_z(__VA_ARGS__) -#define svorr_b_z(...) __builtin_sve_svorr_b_z(__VA_ARGS__) -#define svorr_n_u8_m(...) __builtin_sve_svorr_n_u8_m(__VA_ARGS__) -#define svorr_n_u32_m(...) __builtin_sve_svorr_n_u32_m(__VA_ARGS__) -#define svorr_n_u64_m(...) __builtin_sve_svorr_n_u64_m(__VA_ARGS__) -#define svorr_n_u16_m(...) __builtin_sve_svorr_n_u16_m(__VA_ARGS__) -#define svorr_n_s8_m(...) __builtin_sve_svorr_n_s8_m(__VA_ARGS__) -#define svorr_n_s32_m(...) __builtin_sve_svorr_n_s32_m(__VA_ARGS__) -#define svorr_n_s64_m(...) __builtin_sve_svorr_n_s64_m(__VA_ARGS__) -#define svorr_n_s16_m(...) __builtin_sve_svorr_n_s16_m(__VA_ARGS__) -#define svorr_n_u8_x(...) __builtin_sve_svorr_n_u8_x(__VA_ARGS__) -#define svorr_n_u32_x(...) __builtin_sve_svorr_n_u32_x(__VA_ARGS__) -#define svorr_n_u64_x(...) __builtin_sve_svorr_n_u64_x(__VA_ARGS__) -#define svorr_n_u16_x(...) __builtin_sve_svorr_n_u16_x(__VA_ARGS__) -#define svorr_n_s8_x(...) __builtin_sve_svorr_n_s8_x(__VA_ARGS__) -#define svorr_n_s32_x(...) __builtin_sve_svorr_n_s32_x(__VA_ARGS__) -#define svorr_n_s64_x(...) __builtin_sve_svorr_n_s64_x(__VA_ARGS__) -#define svorr_n_s16_x(...) __builtin_sve_svorr_n_s16_x(__VA_ARGS__) -#define svorr_n_u8_z(...) __builtin_sve_svorr_n_u8_z(__VA_ARGS__) -#define svorr_n_u32_z(...) __builtin_sve_svorr_n_u32_z(__VA_ARGS__) -#define svorr_n_u64_z(...) __builtin_sve_svorr_n_u64_z(__VA_ARGS__) -#define svorr_n_u16_z(...) __builtin_sve_svorr_n_u16_z(__VA_ARGS__) -#define svorr_n_s8_z(...) __builtin_sve_svorr_n_s8_z(__VA_ARGS__) -#define svorr_n_s32_z(...) __builtin_sve_svorr_n_s32_z(__VA_ARGS__) -#define svorr_n_s64_z(...) __builtin_sve_svorr_n_s64_z(__VA_ARGS__) -#define svorr_n_s16_z(...) __builtin_sve_svorr_n_s16_z(__VA_ARGS__) -#define svorr_u8_m(...) __builtin_sve_svorr_u8_m(__VA_ARGS__) -#define svorr_u32_m(...) __builtin_sve_svorr_u32_m(__VA_ARGS__) -#define svorr_u64_m(...) __builtin_sve_svorr_u64_m(__VA_ARGS__) -#define svorr_u16_m(...) __builtin_sve_svorr_u16_m(__VA_ARGS__) -#define svorr_s8_m(...) __builtin_sve_svorr_s8_m(__VA_ARGS__) -#define svorr_s32_m(...) __builtin_sve_svorr_s32_m(__VA_ARGS__) -#define svorr_s64_m(...) __builtin_sve_svorr_s64_m(__VA_ARGS__) -#define svorr_s16_m(...) __builtin_sve_svorr_s16_m(__VA_ARGS__) -#define svorr_u8_x(...) __builtin_sve_svorr_u8_x(__VA_ARGS__) -#define svorr_u32_x(...) __builtin_sve_svorr_u32_x(__VA_ARGS__) -#define svorr_u64_x(...) __builtin_sve_svorr_u64_x(__VA_ARGS__) -#define svorr_u16_x(...) __builtin_sve_svorr_u16_x(__VA_ARGS__) -#define svorr_s8_x(...) __builtin_sve_svorr_s8_x(__VA_ARGS__) -#define svorr_s32_x(...) __builtin_sve_svorr_s32_x(__VA_ARGS__) -#define svorr_s64_x(...) __builtin_sve_svorr_s64_x(__VA_ARGS__) -#define svorr_s16_x(...) __builtin_sve_svorr_s16_x(__VA_ARGS__) -#define svorr_u8_z(...) __builtin_sve_svorr_u8_z(__VA_ARGS__) -#define svorr_u32_z(...) __builtin_sve_svorr_u32_z(__VA_ARGS__) -#define svorr_u64_z(...) __builtin_sve_svorr_u64_z(__VA_ARGS__) -#define svorr_u16_z(...) __builtin_sve_svorr_u16_z(__VA_ARGS__) -#define svorr_s8_z(...) __builtin_sve_svorr_s8_z(__VA_ARGS__) -#define svorr_s32_z(...) __builtin_sve_svorr_s32_z(__VA_ARGS__) -#define svorr_s64_z(...) __builtin_sve_svorr_s64_z(__VA_ARGS__) -#define svorr_s16_z(...) __builtin_sve_svorr_s16_z(__VA_ARGS__) -#define svorv_u8(...) __builtin_sve_svorv_u8(__VA_ARGS__) -#define svorv_u32(...) __builtin_sve_svorv_u32(__VA_ARGS__) -#define svorv_u64(...) __builtin_sve_svorv_u64(__VA_ARGS__) -#define svorv_u16(...) __builtin_sve_svorv_u16(__VA_ARGS__) -#define svorv_s8(...) __builtin_sve_svorv_s8(__VA_ARGS__) -#define svorv_s32(...) __builtin_sve_svorv_s32(__VA_ARGS__) -#define svorv_s64(...) __builtin_sve_svorv_s64(__VA_ARGS__) -#define svorv_s16(...) __builtin_sve_svorv_s16(__VA_ARGS__) -#define svpfalse_b(...) __builtin_sve_svpfalse_b(__VA_ARGS__) -#define svpfirst_b(...) __builtin_sve_svpfirst_b(__VA_ARGS__) -#define svpnext_b8(...) __builtin_sve_svpnext_b8(__VA_ARGS__) -#define svpnext_b32(...) __builtin_sve_svpnext_b32(__VA_ARGS__) -#define svpnext_b64(...) __builtin_sve_svpnext_b64(__VA_ARGS__) -#define svpnext_b16(...) __builtin_sve_svpnext_b16(__VA_ARGS__) -#define svprfb(...) __builtin_sve_svprfb(__VA_ARGS__) -#define svprfb_gather_u32base(...) __builtin_sve_svprfb_gather_u32base(__VA_ARGS__) -#define svprfb_gather_u64base(...) __builtin_sve_svprfb_gather_u64base(__VA_ARGS__) -#define svprfb_gather_u32base_offset(...) __builtin_sve_svprfb_gather_u32base_offset(__VA_ARGS__) -#define svprfb_gather_u64base_offset(...) __builtin_sve_svprfb_gather_u64base_offset(__VA_ARGS__) -#define svprfb_gather_s32offset(...) __builtin_sve_svprfb_gather_s32offset(__VA_ARGS__) -#define svprfb_gather_u32offset(...) __builtin_sve_svprfb_gather_u32offset(__VA_ARGS__) -#define svprfb_gather_s64offset(...) __builtin_sve_svprfb_gather_s64offset(__VA_ARGS__) -#define svprfb_gather_u64offset(...) __builtin_sve_svprfb_gather_u64offset(__VA_ARGS__) -#define svprfb_vnum(...) __builtin_sve_svprfb_vnum(__VA_ARGS__) -#define svprfd(...) __builtin_sve_svprfd(__VA_ARGS__) -#define svprfd_gather_u32base(...) __builtin_sve_svprfd_gather_u32base(__VA_ARGS__) -#define svprfd_gather_u64base(...) __builtin_sve_svprfd_gather_u64base(__VA_ARGS__) -#define svprfd_gather_u32base_index(...) __builtin_sve_svprfd_gather_u32base_index(__VA_ARGS__) -#define svprfd_gather_u64base_index(...) __builtin_sve_svprfd_gather_u64base_index(__VA_ARGS__) -#define svprfd_gather_s32index(...) __builtin_sve_svprfd_gather_s32index(__VA_ARGS__) -#define svprfd_gather_u32index(...) __builtin_sve_svprfd_gather_u32index(__VA_ARGS__) -#define svprfd_gather_s64index(...) __builtin_sve_svprfd_gather_s64index(__VA_ARGS__) -#define svprfd_gather_u64index(...) __builtin_sve_svprfd_gather_u64index(__VA_ARGS__) -#define svprfd_vnum(...) __builtin_sve_svprfd_vnum(__VA_ARGS__) -#define svprfh(...) __builtin_sve_svprfh(__VA_ARGS__) -#define svprfh_gather_u32base(...) __builtin_sve_svprfh_gather_u32base(__VA_ARGS__) -#define svprfh_gather_u64base(...) __builtin_sve_svprfh_gather_u64base(__VA_ARGS__) -#define svprfh_gather_u32base_index(...) __builtin_sve_svprfh_gather_u32base_index(__VA_ARGS__) -#define svprfh_gather_u64base_index(...) __builtin_sve_svprfh_gather_u64base_index(__VA_ARGS__) -#define svprfh_gather_s32index(...) __builtin_sve_svprfh_gather_s32index(__VA_ARGS__) -#define svprfh_gather_u32index(...) __builtin_sve_svprfh_gather_u32index(__VA_ARGS__) -#define svprfh_gather_s64index(...) __builtin_sve_svprfh_gather_s64index(__VA_ARGS__) -#define svprfh_gather_u64index(...) __builtin_sve_svprfh_gather_u64index(__VA_ARGS__) -#define svprfh_vnum(...) __builtin_sve_svprfh_vnum(__VA_ARGS__) -#define svprfw(...) __builtin_sve_svprfw(__VA_ARGS__) -#define svprfw_gather_u32base(...) __builtin_sve_svprfw_gather_u32base(__VA_ARGS__) -#define svprfw_gather_u64base(...) __builtin_sve_svprfw_gather_u64base(__VA_ARGS__) -#define svprfw_gather_u32base_index(...) __builtin_sve_svprfw_gather_u32base_index(__VA_ARGS__) -#define svprfw_gather_u64base_index(...) __builtin_sve_svprfw_gather_u64base_index(__VA_ARGS__) -#define svprfw_gather_s32index(...) __builtin_sve_svprfw_gather_s32index(__VA_ARGS__) -#define svprfw_gather_u32index(...) __builtin_sve_svprfw_gather_u32index(__VA_ARGS__) -#define svprfw_gather_s64index(...) __builtin_sve_svprfw_gather_s64index(__VA_ARGS__) -#define svprfw_gather_u64index(...) __builtin_sve_svprfw_gather_u64index(__VA_ARGS__) -#define svprfw_vnum(...) __builtin_sve_svprfw_vnum(__VA_ARGS__) -#define svptest_any(...) __builtin_sve_svptest_any(__VA_ARGS__) -#define svptest_first(...) __builtin_sve_svptest_first(__VA_ARGS__) -#define svptest_last(...) __builtin_sve_svptest_last(__VA_ARGS__) -#define svptrue_pat_b8(...) __builtin_sve_svptrue_pat_b8(__VA_ARGS__) -#define svptrue_pat_b32(...) __builtin_sve_svptrue_pat_b32(__VA_ARGS__) -#define svptrue_pat_b64(...) __builtin_sve_svptrue_pat_b64(__VA_ARGS__) -#define svptrue_pat_b16(...) __builtin_sve_svptrue_pat_b16(__VA_ARGS__) -#define svptrue_b8(...) __builtin_sve_svptrue_b8(__VA_ARGS__) -#define svptrue_b32(...) __builtin_sve_svptrue_b32(__VA_ARGS__) -#define svptrue_b64(...) __builtin_sve_svptrue_b64(__VA_ARGS__) -#define svptrue_b16(...) __builtin_sve_svptrue_b16(__VA_ARGS__) -#define svqadd_n_s8(...) __builtin_sve_svqadd_n_s8(__VA_ARGS__) -#define svqadd_n_s32(...) __builtin_sve_svqadd_n_s32(__VA_ARGS__) -#define svqadd_n_s64(...) __builtin_sve_svqadd_n_s64(__VA_ARGS__) -#define svqadd_n_s16(...) __builtin_sve_svqadd_n_s16(__VA_ARGS__) -#define svqadd_n_u8(...) __builtin_sve_svqadd_n_u8(__VA_ARGS__) -#define svqadd_n_u32(...) __builtin_sve_svqadd_n_u32(__VA_ARGS__) -#define svqadd_n_u64(...) __builtin_sve_svqadd_n_u64(__VA_ARGS__) -#define svqadd_n_u16(...) __builtin_sve_svqadd_n_u16(__VA_ARGS__) -#define svqadd_s8(...) __builtin_sve_svqadd_s8(__VA_ARGS__) -#define svqadd_s32(...) __builtin_sve_svqadd_s32(__VA_ARGS__) -#define svqadd_s64(...) __builtin_sve_svqadd_s64(__VA_ARGS__) -#define svqadd_s16(...) __builtin_sve_svqadd_s16(__VA_ARGS__) -#define svqadd_u8(...) __builtin_sve_svqadd_u8(__VA_ARGS__) -#define svqadd_u32(...) __builtin_sve_svqadd_u32(__VA_ARGS__) -#define svqadd_u64(...) __builtin_sve_svqadd_u64(__VA_ARGS__) -#define svqadd_u16(...) __builtin_sve_svqadd_u16(__VA_ARGS__) -#define svqdecb_n_s32(...) __builtin_sve_svqdecb_n_s32(__VA_ARGS__) -#define svqdecb_n_s64(...) __builtin_sve_svqdecb_n_s64(__VA_ARGS__) -#define svqdecb_n_u32(...) __builtin_sve_svqdecb_n_u32(__VA_ARGS__) -#define svqdecb_n_u64(...) __builtin_sve_svqdecb_n_u64(__VA_ARGS__) -#define svqdecb_pat_n_s32(...) __builtin_sve_svqdecb_pat_n_s32(__VA_ARGS__) -#define svqdecb_pat_n_s64(...) __builtin_sve_svqdecb_pat_n_s64(__VA_ARGS__) -#define svqdecb_pat_n_u32(...) __builtin_sve_svqdecb_pat_n_u32(__VA_ARGS__) -#define svqdecb_pat_n_u64(...) __builtin_sve_svqdecb_pat_n_u64(__VA_ARGS__) -#define svqdecd_n_s32(...) __builtin_sve_svqdecd_n_s32(__VA_ARGS__) -#define svqdecd_n_s64(...) __builtin_sve_svqdecd_n_s64(__VA_ARGS__) -#define svqdecd_n_u32(...) __builtin_sve_svqdecd_n_u32(__VA_ARGS__) -#define svqdecd_n_u64(...) __builtin_sve_svqdecd_n_u64(__VA_ARGS__) -#define svqdecd_s64(...) __builtin_sve_svqdecd_s64(__VA_ARGS__) -#define svqdecd_u64(...) __builtin_sve_svqdecd_u64(__VA_ARGS__) -#define svqdecd_pat_n_s32(...) __builtin_sve_svqdecd_pat_n_s32(__VA_ARGS__) -#define svqdecd_pat_n_s64(...) __builtin_sve_svqdecd_pat_n_s64(__VA_ARGS__) -#define svqdecd_pat_n_u32(...) __builtin_sve_svqdecd_pat_n_u32(__VA_ARGS__) -#define svqdecd_pat_n_u64(...) __builtin_sve_svqdecd_pat_n_u64(__VA_ARGS__) -#define svqdecd_pat_s64(...) __builtin_sve_svqdecd_pat_s64(__VA_ARGS__) -#define svqdecd_pat_u64(...) __builtin_sve_svqdecd_pat_u64(__VA_ARGS__) -#define svqdech_n_s32(...) __builtin_sve_svqdech_n_s32(__VA_ARGS__) -#define svqdech_n_s64(...) __builtin_sve_svqdech_n_s64(__VA_ARGS__) -#define svqdech_n_u32(...) __builtin_sve_svqdech_n_u32(__VA_ARGS__) -#define svqdech_n_u64(...) __builtin_sve_svqdech_n_u64(__VA_ARGS__) -#define svqdech_s16(...) __builtin_sve_svqdech_s16(__VA_ARGS__) -#define svqdech_u16(...) __builtin_sve_svqdech_u16(__VA_ARGS__) -#define svqdech_pat_n_s32(...) __builtin_sve_svqdech_pat_n_s32(__VA_ARGS__) -#define svqdech_pat_n_s64(...) __builtin_sve_svqdech_pat_n_s64(__VA_ARGS__) -#define svqdech_pat_n_u32(...) __builtin_sve_svqdech_pat_n_u32(__VA_ARGS__) -#define svqdech_pat_n_u64(...) __builtin_sve_svqdech_pat_n_u64(__VA_ARGS__) -#define svqdech_pat_s16(...) __builtin_sve_svqdech_pat_s16(__VA_ARGS__) -#define svqdech_pat_u16(...) __builtin_sve_svqdech_pat_u16(__VA_ARGS__) -#define svqdecp_n_s32_b8(...) __builtin_sve_svqdecp_n_s32_b8(__VA_ARGS__) -#define svqdecp_n_s32_b32(...) __builtin_sve_svqdecp_n_s32_b32(__VA_ARGS__) -#define svqdecp_n_s32_b64(...) __builtin_sve_svqdecp_n_s32_b64(__VA_ARGS__) -#define svqdecp_n_s32_b16(...) __builtin_sve_svqdecp_n_s32_b16(__VA_ARGS__) -#define svqdecp_n_s64_b8(...) __builtin_sve_svqdecp_n_s64_b8(__VA_ARGS__) -#define svqdecp_n_s64_b32(...) __builtin_sve_svqdecp_n_s64_b32(__VA_ARGS__) -#define svqdecp_n_s64_b64(...) __builtin_sve_svqdecp_n_s64_b64(__VA_ARGS__) -#define svqdecp_n_s64_b16(...) __builtin_sve_svqdecp_n_s64_b16(__VA_ARGS__) -#define svqdecp_n_u32_b8(...) __builtin_sve_svqdecp_n_u32_b8(__VA_ARGS__) -#define svqdecp_n_u32_b32(...) __builtin_sve_svqdecp_n_u32_b32(__VA_ARGS__) -#define svqdecp_n_u32_b64(...) __builtin_sve_svqdecp_n_u32_b64(__VA_ARGS__) -#define svqdecp_n_u32_b16(...) __builtin_sve_svqdecp_n_u32_b16(__VA_ARGS__) -#define svqdecp_n_u64_b8(...) __builtin_sve_svqdecp_n_u64_b8(__VA_ARGS__) -#define svqdecp_n_u64_b32(...) __builtin_sve_svqdecp_n_u64_b32(__VA_ARGS__) -#define svqdecp_n_u64_b64(...) __builtin_sve_svqdecp_n_u64_b64(__VA_ARGS__) -#define svqdecp_n_u64_b16(...) __builtin_sve_svqdecp_n_u64_b16(__VA_ARGS__) -#define svqdecp_s32(...) __builtin_sve_svqdecp_s32(__VA_ARGS__) -#define svqdecp_s64(...) __builtin_sve_svqdecp_s64(__VA_ARGS__) -#define svqdecp_s16(...) __builtin_sve_svqdecp_s16(__VA_ARGS__) -#define svqdecp_u32(...) __builtin_sve_svqdecp_u32(__VA_ARGS__) -#define svqdecp_u64(...) __builtin_sve_svqdecp_u64(__VA_ARGS__) -#define svqdecp_u16(...) __builtin_sve_svqdecp_u16(__VA_ARGS__) -#define svqdecw_n_s32(...) __builtin_sve_svqdecw_n_s32(__VA_ARGS__) -#define svqdecw_n_s64(...) __builtin_sve_svqdecw_n_s64(__VA_ARGS__) -#define svqdecw_n_u32(...) __builtin_sve_svqdecw_n_u32(__VA_ARGS__) -#define svqdecw_n_u64(...) __builtin_sve_svqdecw_n_u64(__VA_ARGS__) -#define svqdecw_s32(...) __builtin_sve_svqdecw_s32(__VA_ARGS__) -#define svqdecw_u32(...) __builtin_sve_svqdecw_u32(__VA_ARGS__) -#define svqdecw_pat_n_s32(...) __builtin_sve_svqdecw_pat_n_s32(__VA_ARGS__) -#define svqdecw_pat_n_s64(...) __builtin_sve_svqdecw_pat_n_s64(__VA_ARGS__) -#define svqdecw_pat_n_u32(...) __builtin_sve_svqdecw_pat_n_u32(__VA_ARGS__) -#define svqdecw_pat_n_u64(...) __builtin_sve_svqdecw_pat_n_u64(__VA_ARGS__) -#define svqdecw_pat_s32(...) __builtin_sve_svqdecw_pat_s32(__VA_ARGS__) -#define svqdecw_pat_u32(...) __builtin_sve_svqdecw_pat_u32(__VA_ARGS__) -#define svqincb_n_s32(...) __builtin_sve_svqincb_n_s32(__VA_ARGS__) -#define svqincb_n_s64(...) __builtin_sve_svqincb_n_s64(__VA_ARGS__) -#define svqincb_n_u32(...) __builtin_sve_svqincb_n_u32(__VA_ARGS__) -#define svqincb_n_u64(...) __builtin_sve_svqincb_n_u64(__VA_ARGS__) -#define svqincb_pat_n_s32(...) __builtin_sve_svqincb_pat_n_s32(__VA_ARGS__) -#define svqincb_pat_n_s64(...) __builtin_sve_svqincb_pat_n_s64(__VA_ARGS__) -#define svqincb_pat_n_u32(...) __builtin_sve_svqincb_pat_n_u32(__VA_ARGS__) -#define svqincb_pat_n_u64(...) __builtin_sve_svqincb_pat_n_u64(__VA_ARGS__) -#define svqincd_n_s32(...) __builtin_sve_svqincd_n_s32(__VA_ARGS__) -#define svqincd_n_s64(...) __builtin_sve_svqincd_n_s64(__VA_ARGS__) -#define svqincd_n_u32(...) __builtin_sve_svqincd_n_u32(__VA_ARGS__) -#define svqincd_n_u64(...) __builtin_sve_svqincd_n_u64(__VA_ARGS__) -#define svqincd_s64(...) __builtin_sve_svqincd_s64(__VA_ARGS__) -#define svqincd_u64(...) __builtin_sve_svqincd_u64(__VA_ARGS__) -#define svqincd_pat_n_s32(...) __builtin_sve_svqincd_pat_n_s32(__VA_ARGS__) -#define svqincd_pat_n_s64(...) __builtin_sve_svqincd_pat_n_s64(__VA_ARGS__) -#define svqincd_pat_n_u32(...) __builtin_sve_svqincd_pat_n_u32(__VA_ARGS__) -#define svqincd_pat_n_u64(...) __builtin_sve_svqincd_pat_n_u64(__VA_ARGS__) -#define svqincd_pat_s64(...) __builtin_sve_svqincd_pat_s64(__VA_ARGS__) -#define svqincd_pat_u64(...) __builtin_sve_svqincd_pat_u64(__VA_ARGS__) -#define svqinch_n_s32(...) __builtin_sve_svqinch_n_s32(__VA_ARGS__) -#define svqinch_n_s64(...) __builtin_sve_svqinch_n_s64(__VA_ARGS__) -#define svqinch_n_u32(...) __builtin_sve_svqinch_n_u32(__VA_ARGS__) -#define svqinch_n_u64(...) __builtin_sve_svqinch_n_u64(__VA_ARGS__) -#define svqinch_s16(...) __builtin_sve_svqinch_s16(__VA_ARGS__) -#define svqinch_u16(...) __builtin_sve_svqinch_u16(__VA_ARGS__) -#define svqinch_pat_n_s32(...) __builtin_sve_svqinch_pat_n_s32(__VA_ARGS__) -#define svqinch_pat_n_s64(...) __builtin_sve_svqinch_pat_n_s64(__VA_ARGS__) -#define svqinch_pat_n_u32(...) __builtin_sve_svqinch_pat_n_u32(__VA_ARGS__) -#define svqinch_pat_n_u64(...) __builtin_sve_svqinch_pat_n_u64(__VA_ARGS__) -#define svqinch_pat_s16(...) __builtin_sve_svqinch_pat_s16(__VA_ARGS__) -#define svqinch_pat_u16(...) __builtin_sve_svqinch_pat_u16(__VA_ARGS__) -#define svqincp_n_s32_b8(...) __builtin_sve_svqincp_n_s32_b8(__VA_ARGS__) -#define svqincp_n_s32_b32(...) __builtin_sve_svqincp_n_s32_b32(__VA_ARGS__) -#define svqincp_n_s32_b64(...) __builtin_sve_svqincp_n_s32_b64(__VA_ARGS__) -#define svqincp_n_s32_b16(...) __builtin_sve_svqincp_n_s32_b16(__VA_ARGS__) -#define svqincp_n_s64_b8(...) __builtin_sve_svqincp_n_s64_b8(__VA_ARGS__) -#define svqincp_n_s64_b32(...) __builtin_sve_svqincp_n_s64_b32(__VA_ARGS__) -#define svqincp_n_s64_b64(...) __builtin_sve_svqincp_n_s64_b64(__VA_ARGS__) -#define svqincp_n_s64_b16(...) __builtin_sve_svqincp_n_s64_b16(__VA_ARGS__) -#define svqincp_n_u32_b8(...) __builtin_sve_svqincp_n_u32_b8(__VA_ARGS__) -#define svqincp_n_u32_b32(...) __builtin_sve_svqincp_n_u32_b32(__VA_ARGS__) -#define svqincp_n_u32_b64(...) __builtin_sve_svqincp_n_u32_b64(__VA_ARGS__) -#define svqincp_n_u32_b16(...) __builtin_sve_svqincp_n_u32_b16(__VA_ARGS__) -#define svqincp_n_u64_b8(...) __builtin_sve_svqincp_n_u64_b8(__VA_ARGS__) -#define svqincp_n_u64_b32(...) __builtin_sve_svqincp_n_u64_b32(__VA_ARGS__) -#define svqincp_n_u64_b64(...) __builtin_sve_svqincp_n_u64_b64(__VA_ARGS__) -#define svqincp_n_u64_b16(...) __builtin_sve_svqincp_n_u64_b16(__VA_ARGS__) -#define svqincp_s32(...) __builtin_sve_svqincp_s32(__VA_ARGS__) -#define svqincp_s64(...) __builtin_sve_svqincp_s64(__VA_ARGS__) -#define svqincp_s16(...) __builtin_sve_svqincp_s16(__VA_ARGS__) -#define svqincp_u32(...) __builtin_sve_svqincp_u32(__VA_ARGS__) -#define svqincp_u64(...) __builtin_sve_svqincp_u64(__VA_ARGS__) -#define svqincp_u16(...) __builtin_sve_svqincp_u16(__VA_ARGS__) -#define svqincw_n_s32(...) __builtin_sve_svqincw_n_s32(__VA_ARGS__) -#define svqincw_n_s64(...) __builtin_sve_svqincw_n_s64(__VA_ARGS__) -#define svqincw_n_u32(...) __builtin_sve_svqincw_n_u32(__VA_ARGS__) -#define svqincw_n_u64(...) __builtin_sve_svqincw_n_u64(__VA_ARGS__) -#define svqincw_s32(...) __builtin_sve_svqincw_s32(__VA_ARGS__) -#define svqincw_u32(...) __builtin_sve_svqincw_u32(__VA_ARGS__) -#define svqincw_pat_n_s32(...) __builtin_sve_svqincw_pat_n_s32(__VA_ARGS__) -#define svqincw_pat_n_s64(...) __builtin_sve_svqincw_pat_n_s64(__VA_ARGS__) -#define svqincw_pat_n_u32(...) __builtin_sve_svqincw_pat_n_u32(__VA_ARGS__) -#define svqincw_pat_n_u64(...) __builtin_sve_svqincw_pat_n_u64(__VA_ARGS__) -#define svqincw_pat_s32(...) __builtin_sve_svqincw_pat_s32(__VA_ARGS__) -#define svqincw_pat_u32(...) __builtin_sve_svqincw_pat_u32(__VA_ARGS__) -#define svqsub_n_s8(...) __builtin_sve_svqsub_n_s8(__VA_ARGS__) -#define svqsub_n_s32(...) __builtin_sve_svqsub_n_s32(__VA_ARGS__) -#define svqsub_n_s64(...) __builtin_sve_svqsub_n_s64(__VA_ARGS__) -#define svqsub_n_s16(...) __builtin_sve_svqsub_n_s16(__VA_ARGS__) -#define svqsub_n_u8(...) __builtin_sve_svqsub_n_u8(__VA_ARGS__) -#define svqsub_n_u32(...) __builtin_sve_svqsub_n_u32(__VA_ARGS__) -#define svqsub_n_u64(...) __builtin_sve_svqsub_n_u64(__VA_ARGS__) -#define svqsub_n_u16(...) __builtin_sve_svqsub_n_u16(__VA_ARGS__) -#define svqsub_s8(...) __builtin_sve_svqsub_s8(__VA_ARGS__) -#define svqsub_s32(...) __builtin_sve_svqsub_s32(__VA_ARGS__) -#define svqsub_s64(...) __builtin_sve_svqsub_s64(__VA_ARGS__) -#define svqsub_s16(...) __builtin_sve_svqsub_s16(__VA_ARGS__) -#define svqsub_u8(...) __builtin_sve_svqsub_u8(__VA_ARGS__) -#define svqsub_u32(...) __builtin_sve_svqsub_u32(__VA_ARGS__) -#define svqsub_u64(...) __builtin_sve_svqsub_u64(__VA_ARGS__) -#define svqsub_u16(...) __builtin_sve_svqsub_u16(__VA_ARGS__) -#define svrbit_u8_m(...) __builtin_sve_svrbit_u8_m(__VA_ARGS__) -#define svrbit_u32_m(...) __builtin_sve_svrbit_u32_m(__VA_ARGS__) -#define svrbit_u64_m(...) __builtin_sve_svrbit_u64_m(__VA_ARGS__) -#define svrbit_u16_m(...) __builtin_sve_svrbit_u16_m(__VA_ARGS__) -#define svrbit_s8_m(...) __builtin_sve_svrbit_s8_m(__VA_ARGS__) -#define svrbit_s32_m(...) __builtin_sve_svrbit_s32_m(__VA_ARGS__) -#define svrbit_s64_m(...) __builtin_sve_svrbit_s64_m(__VA_ARGS__) -#define svrbit_s16_m(...) __builtin_sve_svrbit_s16_m(__VA_ARGS__) -#define svrbit_u8_x(...) __builtin_sve_svrbit_u8_x(__VA_ARGS__) -#define svrbit_u32_x(...) __builtin_sve_svrbit_u32_x(__VA_ARGS__) -#define svrbit_u64_x(...) __builtin_sve_svrbit_u64_x(__VA_ARGS__) -#define svrbit_u16_x(...) __builtin_sve_svrbit_u16_x(__VA_ARGS__) -#define svrbit_s8_x(...) __builtin_sve_svrbit_s8_x(__VA_ARGS__) -#define svrbit_s32_x(...) __builtin_sve_svrbit_s32_x(__VA_ARGS__) -#define svrbit_s64_x(...) __builtin_sve_svrbit_s64_x(__VA_ARGS__) -#define svrbit_s16_x(...) __builtin_sve_svrbit_s16_x(__VA_ARGS__) -#define svrbit_u8_z(...) __builtin_sve_svrbit_u8_z(__VA_ARGS__) -#define svrbit_u32_z(...) __builtin_sve_svrbit_u32_z(__VA_ARGS__) -#define svrbit_u64_z(...) __builtin_sve_svrbit_u64_z(__VA_ARGS__) -#define svrbit_u16_z(...) __builtin_sve_svrbit_u16_z(__VA_ARGS__) -#define svrbit_s8_z(...) __builtin_sve_svrbit_s8_z(__VA_ARGS__) -#define svrbit_s32_z(...) __builtin_sve_svrbit_s32_z(__VA_ARGS__) -#define svrbit_s64_z(...) __builtin_sve_svrbit_s64_z(__VA_ARGS__) -#define svrbit_s16_z(...) __builtin_sve_svrbit_s16_z(__VA_ARGS__) -#define svrdffr(...) __builtin_sve_svrdffr(__VA_ARGS__) -#define svrdffr_z(...) __builtin_sve_svrdffr_z(__VA_ARGS__) -#define svrecpe_f64(...) __builtin_sve_svrecpe_f64(__VA_ARGS__) -#define svrecpe_f32(...) __builtin_sve_svrecpe_f32(__VA_ARGS__) -#define svrecpe_f16(...) __builtin_sve_svrecpe_f16(__VA_ARGS__) -#define svrecps_f64(...) __builtin_sve_svrecps_f64(__VA_ARGS__) -#define svrecps_f32(...) __builtin_sve_svrecps_f32(__VA_ARGS__) -#define svrecps_f16(...) __builtin_sve_svrecps_f16(__VA_ARGS__) -#define svrecpx_f64_m(...) __builtin_sve_svrecpx_f64_m(__VA_ARGS__) -#define svrecpx_f32_m(...) __builtin_sve_svrecpx_f32_m(__VA_ARGS__) -#define svrecpx_f16_m(...) __builtin_sve_svrecpx_f16_m(__VA_ARGS__) -#define svrecpx_f64_x(...) __builtin_sve_svrecpx_f64_x(__VA_ARGS__) -#define svrecpx_f32_x(...) __builtin_sve_svrecpx_f32_x(__VA_ARGS__) -#define svrecpx_f16_x(...) __builtin_sve_svrecpx_f16_x(__VA_ARGS__) -#define svrecpx_f64_z(...) __builtin_sve_svrecpx_f64_z(__VA_ARGS__) -#define svrecpx_f32_z(...) __builtin_sve_svrecpx_f32_z(__VA_ARGS__) -#define svrecpx_f16_z(...) __builtin_sve_svrecpx_f16_z(__VA_ARGS__) -#define svrev_u8(...) __builtin_sve_svrev_u8(__VA_ARGS__) -#define svrev_u32(...) __builtin_sve_svrev_u32(__VA_ARGS__) -#define svrev_u64(...) __builtin_sve_svrev_u64(__VA_ARGS__) -#define svrev_u16(...) __builtin_sve_svrev_u16(__VA_ARGS__) -#define svrev_s8(...) __builtin_sve_svrev_s8(__VA_ARGS__) -#define svrev_f64(...) __builtin_sve_svrev_f64(__VA_ARGS__) -#define svrev_f32(...) __builtin_sve_svrev_f32(__VA_ARGS__) -#define svrev_f16(...) __builtin_sve_svrev_f16(__VA_ARGS__) -#define svrev_s32(...) __builtin_sve_svrev_s32(__VA_ARGS__) -#define svrev_s64(...) __builtin_sve_svrev_s64(__VA_ARGS__) -#define svrev_s16(...) __builtin_sve_svrev_s16(__VA_ARGS__) -#define svrev_b8(...) __builtin_sve_svrev_b8(__VA_ARGS__) -#define svrev_b32(...) __builtin_sve_svrev_b32(__VA_ARGS__) -#define svrev_b64(...) __builtin_sve_svrev_b64(__VA_ARGS__) -#define svrev_b16(...) __builtin_sve_svrev_b16(__VA_ARGS__) -#define svrevb_u32_m(...) __builtin_sve_svrevb_u32_m(__VA_ARGS__) -#define svrevb_u64_m(...) __builtin_sve_svrevb_u64_m(__VA_ARGS__) -#define svrevb_u16_m(...) __builtin_sve_svrevb_u16_m(__VA_ARGS__) -#define svrevb_s32_m(...) __builtin_sve_svrevb_s32_m(__VA_ARGS__) -#define svrevb_s64_m(...) __builtin_sve_svrevb_s64_m(__VA_ARGS__) -#define svrevb_s16_m(...) __builtin_sve_svrevb_s16_m(__VA_ARGS__) -#define svrevb_u32_x(...) __builtin_sve_svrevb_u32_x(__VA_ARGS__) -#define svrevb_u64_x(...) __builtin_sve_svrevb_u64_x(__VA_ARGS__) -#define svrevb_u16_x(...) __builtin_sve_svrevb_u16_x(__VA_ARGS__) -#define svrevb_s32_x(...) __builtin_sve_svrevb_s32_x(__VA_ARGS__) -#define svrevb_s64_x(...) __builtin_sve_svrevb_s64_x(__VA_ARGS__) -#define svrevb_s16_x(...) __builtin_sve_svrevb_s16_x(__VA_ARGS__) -#define svrevb_u32_z(...) __builtin_sve_svrevb_u32_z(__VA_ARGS__) -#define svrevb_u64_z(...) __builtin_sve_svrevb_u64_z(__VA_ARGS__) -#define svrevb_u16_z(...) __builtin_sve_svrevb_u16_z(__VA_ARGS__) -#define svrevb_s32_z(...) __builtin_sve_svrevb_s32_z(__VA_ARGS__) -#define svrevb_s64_z(...) __builtin_sve_svrevb_s64_z(__VA_ARGS__) -#define svrevb_s16_z(...) __builtin_sve_svrevb_s16_z(__VA_ARGS__) -#define svrevh_u32_m(...) __builtin_sve_svrevh_u32_m(__VA_ARGS__) -#define svrevh_u64_m(...) __builtin_sve_svrevh_u64_m(__VA_ARGS__) -#define svrevh_s32_m(...) __builtin_sve_svrevh_s32_m(__VA_ARGS__) -#define svrevh_s64_m(...) __builtin_sve_svrevh_s64_m(__VA_ARGS__) -#define svrevh_u32_x(...) __builtin_sve_svrevh_u32_x(__VA_ARGS__) -#define svrevh_u64_x(...) __builtin_sve_svrevh_u64_x(__VA_ARGS__) -#define svrevh_s32_x(...) __builtin_sve_svrevh_s32_x(__VA_ARGS__) -#define svrevh_s64_x(...) __builtin_sve_svrevh_s64_x(__VA_ARGS__) -#define svrevh_u32_z(...) __builtin_sve_svrevh_u32_z(__VA_ARGS__) -#define svrevh_u64_z(...) __builtin_sve_svrevh_u64_z(__VA_ARGS__) -#define svrevh_s32_z(...) __builtin_sve_svrevh_s32_z(__VA_ARGS__) -#define svrevh_s64_z(...) __builtin_sve_svrevh_s64_z(__VA_ARGS__) -#define svrevw_u64_m(...) __builtin_sve_svrevw_u64_m(__VA_ARGS__) -#define svrevw_s64_m(...) __builtin_sve_svrevw_s64_m(__VA_ARGS__) -#define svrevw_u64_x(...) __builtin_sve_svrevw_u64_x(__VA_ARGS__) -#define svrevw_s64_x(...) __builtin_sve_svrevw_s64_x(__VA_ARGS__) -#define svrevw_u64_z(...) __builtin_sve_svrevw_u64_z(__VA_ARGS__) -#define svrevw_s64_z(...) __builtin_sve_svrevw_s64_z(__VA_ARGS__) -#define svrinta_f64_m(...) __builtin_sve_svrinta_f64_m(__VA_ARGS__) -#define svrinta_f32_m(...) __builtin_sve_svrinta_f32_m(__VA_ARGS__) -#define svrinta_f16_m(...) __builtin_sve_svrinta_f16_m(__VA_ARGS__) -#define svrinta_f64_x(...) __builtin_sve_svrinta_f64_x(__VA_ARGS__) -#define svrinta_f32_x(...) __builtin_sve_svrinta_f32_x(__VA_ARGS__) -#define svrinta_f16_x(...) __builtin_sve_svrinta_f16_x(__VA_ARGS__) -#define svrinta_f64_z(...) __builtin_sve_svrinta_f64_z(__VA_ARGS__) -#define svrinta_f32_z(...) __builtin_sve_svrinta_f32_z(__VA_ARGS__) -#define svrinta_f16_z(...) __builtin_sve_svrinta_f16_z(__VA_ARGS__) -#define svrinti_f64_m(...) __builtin_sve_svrinti_f64_m(__VA_ARGS__) -#define svrinti_f32_m(...) __builtin_sve_svrinti_f32_m(__VA_ARGS__) -#define svrinti_f16_m(...) __builtin_sve_svrinti_f16_m(__VA_ARGS__) -#define svrinti_f64_x(...) __builtin_sve_svrinti_f64_x(__VA_ARGS__) -#define svrinti_f32_x(...) __builtin_sve_svrinti_f32_x(__VA_ARGS__) -#define svrinti_f16_x(...) __builtin_sve_svrinti_f16_x(__VA_ARGS__) -#define svrinti_f64_z(...) __builtin_sve_svrinti_f64_z(__VA_ARGS__) -#define svrinti_f32_z(...) __builtin_sve_svrinti_f32_z(__VA_ARGS__) -#define svrinti_f16_z(...) __builtin_sve_svrinti_f16_z(__VA_ARGS__) -#define svrintm_f64_m(...) __builtin_sve_svrintm_f64_m(__VA_ARGS__) -#define svrintm_f32_m(...) __builtin_sve_svrintm_f32_m(__VA_ARGS__) -#define svrintm_f16_m(...) __builtin_sve_svrintm_f16_m(__VA_ARGS__) -#define svrintm_f64_x(...) __builtin_sve_svrintm_f64_x(__VA_ARGS__) -#define svrintm_f32_x(...) __builtin_sve_svrintm_f32_x(__VA_ARGS__) -#define svrintm_f16_x(...) __builtin_sve_svrintm_f16_x(__VA_ARGS__) -#define svrintm_f64_z(...) __builtin_sve_svrintm_f64_z(__VA_ARGS__) -#define svrintm_f32_z(...) __builtin_sve_svrintm_f32_z(__VA_ARGS__) -#define svrintm_f16_z(...) __builtin_sve_svrintm_f16_z(__VA_ARGS__) -#define svrintn_f64_m(...) __builtin_sve_svrintn_f64_m(__VA_ARGS__) -#define svrintn_f32_m(...) __builtin_sve_svrintn_f32_m(__VA_ARGS__) -#define svrintn_f16_m(...) __builtin_sve_svrintn_f16_m(__VA_ARGS__) -#define svrintn_f64_x(...) __builtin_sve_svrintn_f64_x(__VA_ARGS__) -#define svrintn_f32_x(...) __builtin_sve_svrintn_f32_x(__VA_ARGS__) -#define svrintn_f16_x(...) __builtin_sve_svrintn_f16_x(__VA_ARGS__) -#define svrintn_f64_z(...) __builtin_sve_svrintn_f64_z(__VA_ARGS__) -#define svrintn_f32_z(...) __builtin_sve_svrintn_f32_z(__VA_ARGS__) -#define svrintn_f16_z(...) __builtin_sve_svrintn_f16_z(__VA_ARGS__) -#define svrintp_f64_m(...) __builtin_sve_svrintp_f64_m(__VA_ARGS__) -#define svrintp_f32_m(...) __builtin_sve_svrintp_f32_m(__VA_ARGS__) -#define svrintp_f16_m(...) __builtin_sve_svrintp_f16_m(__VA_ARGS__) -#define svrintp_f64_x(...) __builtin_sve_svrintp_f64_x(__VA_ARGS__) -#define svrintp_f32_x(...) __builtin_sve_svrintp_f32_x(__VA_ARGS__) -#define svrintp_f16_x(...) __builtin_sve_svrintp_f16_x(__VA_ARGS__) -#define svrintp_f64_z(...) __builtin_sve_svrintp_f64_z(__VA_ARGS__) -#define svrintp_f32_z(...) __builtin_sve_svrintp_f32_z(__VA_ARGS__) -#define svrintp_f16_z(...) __builtin_sve_svrintp_f16_z(__VA_ARGS__) -#define svrintx_f64_m(...) __builtin_sve_svrintx_f64_m(__VA_ARGS__) -#define svrintx_f32_m(...) __builtin_sve_svrintx_f32_m(__VA_ARGS__) -#define svrintx_f16_m(...) __builtin_sve_svrintx_f16_m(__VA_ARGS__) -#define svrintx_f64_x(...) __builtin_sve_svrintx_f64_x(__VA_ARGS__) -#define svrintx_f32_x(...) __builtin_sve_svrintx_f32_x(__VA_ARGS__) -#define svrintx_f16_x(...) __builtin_sve_svrintx_f16_x(__VA_ARGS__) -#define svrintx_f64_z(...) __builtin_sve_svrintx_f64_z(__VA_ARGS__) -#define svrintx_f32_z(...) __builtin_sve_svrintx_f32_z(__VA_ARGS__) -#define svrintx_f16_z(...) __builtin_sve_svrintx_f16_z(__VA_ARGS__) -#define svrintz_f64_m(...) __builtin_sve_svrintz_f64_m(__VA_ARGS__) -#define svrintz_f32_m(...) __builtin_sve_svrintz_f32_m(__VA_ARGS__) -#define svrintz_f16_m(...) __builtin_sve_svrintz_f16_m(__VA_ARGS__) -#define svrintz_f64_x(...) __builtin_sve_svrintz_f64_x(__VA_ARGS__) -#define svrintz_f32_x(...) __builtin_sve_svrintz_f32_x(__VA_ARGS__) -#define svrintz_f16_x(...) __builtin_sve_svrintz_f16_x(__VA_ARGS__) -#define svrintz_f64_z(...) __builtin_sve_svrintz_f64_z(__VA_ARGS__) -#define svrintz_f32_z(...) __builtin_sve_svrintz_f32_z(__VA_ARGS__) -#define svrintz_f16_z(...) __builtin_sve_svrintz_f16_z(__VA_ARGS__) -#define svrsqrte_f64(...) __builtin_sve_svrsqrte_f64(__VA_ARGS__) -#define svrsqrte_f32(...) __builtin_sve_svrsqrte_f32(__VA_ARGS__) -#define svrsqrte_f16(...) __builtin_sve_svrsqrte_f16(__VA_ARGS__) -#define svrsqrts_f64(...) __builtin_sve_svrsqrts_f64(__VA_ARGS__) -#define svrsqrts_f32(...) __builtin_sve_svrsqrts_f32(__VA_ARGS__) -#define svrsqrts_f16(...) __builtin_sve_svrsqrts_f16(__VA_ARGS__) -#define svscale_n_f64_m(...) __builtin_sve_svscale_n_f64_m(__VA_ARGS__) -#define svscale_n_f32_m(...) __builtin_sve_svscale_n_f32_m(__VA_ARGS__) -#define svscale_n_f16_m(...) __builtin_sve_svscale_n_f16_m(__VA_ARGS__) -#define svscale_n_f64_x(...) __builtin_sve_svscale_n_f64_x(__VA_ARGS__) -#define svscale_n_f32_x(...) __builtin_sve_svscale_n_f32_x(__VA_ARGS__) -#define svscale_n_f16_x(...) __builtin_sve_svscale_n_f16_x(__VA_ARGS__) -#define svscale_n_f64_z(...) __builtin_sve_svscale_n_f64_z(__VA_ARGS__) -#define svscale_n_f32_z(...) __builtin_sve_svscale_n_f32_z(__VA_ARGS__) -#define svscale_n_f16_z(...) __builtin_sve_svscale_n_f16_z(__VA_ARGS__) -#define svscale_f64_m(...) __builtin_sve_svscale_f64_m(__VA_ARGS__) -#define svscale_f32_m(...) __builtin_sve_svscale_f32_m(__VA_ARGS__) -#define svscale_f16_m(...) __builtin_sve_svscale_f16_m(__VA_ARGS__) -#define svscale_f64_x(...) __builtin_sve_svscale_f64_x(__VA_ARGS__) -#define svscale_f32_x(...) __builtin_sve_svscale_f32_x(__VA_ARGS__) -#define svscale_f16_x(...) __builtin_sve_svscale_f16_x(__VA_ARGS__) -#define svscale_f64_z(...) __builtin_sve_svscale_f64_z(__VA_ARGS__) -#define svscale_f32_z(...) __builtin_sve_svscale_f32_z(__VA_ARGS__) -#define svscale_f16_z(...) __builtin_sve_svscale_f16_z(__VA_ARGS__) -#define svsel_b(...) __builtin_sve_svsel_b(__VA_ARGS__) -#define svsel_u8(...) __builtin_sve_svsel_u8(__VA_ARGS__) -#define svsel_u32(...) __builtin_sve_svsel_u32(__VA_ARGS__) -#define svsel_u64(...) __builtin_sve_svsel_u64(__VA_ARGS__) -#define svsel_u16(...) __builtin_sve_svsel_u16(__VA_ARGS__) -#define svsel_s8(...) __builtin_sve_svsel_s8(__VA_ARGS__) -#define svsel_f64(...) __builtin_sve_svsel_f64(__VA_ARGS__) -#define svsel_f32(...) __builtin_sve_svsel_f32(__VA_ARGS__) -#define svsel_f16(...) __builtin_sve_svsel_f16(__VA_ARGS__) -#define svsel_s32(...) __builtin_sve_svsel_s32(__VA_ARGS__) -#define svsel_s64(...) __builtin_sve_svsel_s64(__VA_ARGS__) -#define svsel_s16(...) __builtin_sve_svsel_s16(__VA_ARGS__) -#define svset2_u8(...) __builtin_sve_svset2_u8(__VA_ARGS__) -#define svset2_u32(...) __builtin_sve_svset2_u32(__VA_ARGS__) -#define svset2_u64(...) __builtin_sve_svset2_u64(__VA_ARGS__) -#define svset2_u16(...) __builtin_sve_svset2_u16(__VA_ARGS__) -#define svset2_s8(...) __builtin_sve_svset2_s8(__VA_ARGS__) -#define svset2_f64(...) __builtin_sve_svset2_f64(__VA_ARGS__) -#define svset2_f32(...) __builtin_sve_svset2_f32(__VA_ARGS__) -#define svset2_f16(...) __builtin_sve_svset2_f16(__VA_ARGS__) -#define svset2_s32(...) __builtin_sve_svset2_s32(__VA_ARGS__) -#define svset2_s64(...) __builtin_sve_svset2_s64(__VA_ARGS__) -#define svset2_s16(...) __builtin_sve_svset2_s16(__VA_ARGS__) -#define svset3_u8(...) __builtin_sve_svset3_u8(__VA_ARGS__) -#define svset3_u32(...) __builtin_sve_svset3_u32(__VA_ARGS__) -#define svset3_u64(...) __builtin_sve_svset3_u64(__VA_ARGS__) -#define svset3_u16(...) __builtin_sve_svset3_u16(__VA_ARGS__) -#define svset3_s8(...) __builtin_sve_svset3_s8(__VA_ARGS__) -#define svset3_f64(...) __builtin_sve_svset3_f64(__VA_ARGS__) -#define svset3_f32(...) __builtin_sve_svset3_f32(__VA_ARGS__) -#define svset3_f16(...) __builtin_sve_svset3_f16(__VA_ARGS__) -#define svset3_s32(...) __builtin_sve_svset3_s32(__VA_ARGS__) -#define svset3_s64(...) __builtin_sve_svset3_s64(__VA_ARGS__) -#define svset3_s16(...) __builtin_sve_svset3_s16(__VA_ARGS__) -#define svset4_u8(...) __builtin_sve_svset4_u8(__VA_ARGS__) -#define svset4_u32(...) __builtin_sve_svset4_u32(__VA_ARGS__) -#define svset4_u64(...) __builtin_sve_svset4_u64(__VA_ARGS__) -#define svset4_u16(...) __builtin_sve_svset4_u16(__VA_ARGS__) -#define svset4_s8(...) __builtin_sve_svset4_s8(__VA_ARGS__) -#define svset4_f64(...) __builtin_sve_svset4_f64(__VA_ARGS__) -#define svset4_f32(...) __builtin_sve_svset4_f32(__VA_ARGS__) -#define svset4_f16(...) __builtin_sve_svset4_f16(__VA_ARGS__) -#define svset4_s32(...) __builtin_sve_svset4_s32(__VA_ARGS__) -#define svset4_s64(...) __builtin_sve_svset4_s64(__VA_ARGS__) -#define svset4_s16(...) __builtin_sve_svset4_s16(__VA_ARGS__) -#define svsetffr(...) __builtin_sve_svsetffr(__VA_ARGS__) -#define svsplice_u8(...) __builtin_sve_svsplice_u8(__VA_ARGS__) -#define svsplice_u32(...) __builtin_sve_svsplice_u32(__VA_ARGS__) -#define svsplice_u64(...) __builtin_sve_svsplice_u64(__VA_ARGS__) -#define svsplice_u16(...) __builtin_sve_svsplice_u16(__VA_ARGS__) -#define svsplice_s8(...) __builtin_sve_svsplice_s8(__VA_ARGS__) -#define svsplice_f64(...) __builtin_sve_svsplice_f64(__VA_ARGS__) -#define svsplice_f32(...) __builtin_sve_svsplice_f32(__VA_ARGS__) -#define svsplice_f16(...) __builtin_sve_svsplice_f16(__VA_ARGS__) -#define svsplice_s32(...) __builtin_sve_svsplice_s32(__VA_ARGS__) -#define svsplice_s64(...) __builtin_sve_svsplice_s64(__VA_ARGS__) -#define svsplice_s16(...) __builtin_sve_svsplice_s16(__VA_ARGS__) -#define svsqrt_f64_m(...) __builtin_sve_svsqrt_f64_m(__VA_ARGS__) -#define svsqrt_f32_m(...) __builtin_sve_svsqrt_f32_m(__VA_ARGS__) -#define svsqrt_f16_m(...) __builtin_sve_svsqrt_f16_m(__VA_ARGS__) -#define svsqrt_f64_x(...) __builtin_sve_svsqrt_f64_x(__VA_ARGS__) -#define svsqrt_f32_x(...) __builtin_sve_svsqrt_f32_x(__VA_ARGS__) -#define svsqrt_f16_x(...) __builtin_sve_svsqrt_f16_x(__VA_ARGS__) -#define svsqrt_f64_z(...) __builtin_sve_svsqrt_f64_z(__VA_ARGS__) -#define svsqrt_f32_z(...) __builtin_sve_svsqrt_f32_z(__VA_ARGS__) -#define svsqrt_f16_z(...) __builtin_sve_svsqrt_f16_z(__VA_ARGS__) -#define svst1_u8(...) __builtin_sve_svst1_u8(__VA_ARGS__) -#define svst1_u32(...) __builtin_sve_svst1_u32(__VA_ARGS__) -#define svst1_u64(...) __builtin_sve_svst1_u64(__VA_ARGS__) -#define svst1_u16(...) __builtin_sve_svst1_u16(__VA_ARGS__) -#define svst1_s8(...) __builtin_sve_svst1_s8(__VA_ARGS__) -#define svst1_f64(...) __builtin_sve_svst1_f64(__VA_ARGS__) -#define svst1_f32(...) __builtin_sve_svst1_f32(__VA_ARGS__) -#define svst1_f16(...) __builtin_sve_svst1_f16(__VA_ARGS__) -#define svst1_s32(...) __builtin_sve_svst1_s32(__VA_ARGS__) -#define svst1_s64(...) __builtin_sve_svst1_s64(__VA_ARGS__) -#define svst1_s16(...) __builtin_sve_svst1_s16(__VA_ARGS__) -#define svst1_scatter_u32base_index_u32(...) __builtin_sve_svst1_scatter_u32base_index_u32(__VA_ARGS__) -#define svst1_scatter_u64base_index_u64(...) __builtin_sve_svst1_scatter_u64base_index_u64(__VA_ARGS__) -#define svst1_scatter_u64base_index_f64(...) __builtin_sve_svst1_scatter_u64base_index_f64(__VA_ARGS__) -#define svst1_scatter_u32base_index_f32(...) __builtin_sve_svst1_scatter_u32base_index_f32(__VA_ARGS__) -#define svst1_scatter_u32base_index_s32(...) __builtin_sve_svst1_scatter_u32base_index_s32(__VA_ARGS__) -#define svst1_scatter_u64base_index_s64(...) __builtin_sve_svst1_scatter_u64base_index_s64(__VA_ARGS__) -#define svst1_scatter_u32base_offset_u32(...) __builtin_sve_svst1_scatter_u32base_offset_u32(__VA_ARGS__) -#define svst1_scatter_u64base_offset_u64(...) __builtin_sve_svst1_scatter_u64base_offset_u64(__VA_ARGS__) -#define svst1_scatter_u64base_offset_f64(...) __builtin_sve_svst1_scatter_u64base_offset_f64(__VA_ARGS__) -#define svst1_scatter_u32base_offset_f32(...) __builtin_sve_svst1_scatter_u32base_offset_f32(__VA_ARGS__) -#define svst1_scatter_u32base_offset_s32(...) __builtin_sve_svst1_scatter_u32base_offset_s32(__VA_ARGS__) -#define svst1_scatter_u64base_offset_s64(...) __builtin_sve_svst1_scatter_u64base_offset_s64(__VA_ARGS__) -#define svst1_scatter_u32base_u32(...) __builtin_sve_svst1_scatter_u32base_u32(__VA_ARGS__) -#define svst1_scatter_u64base_u64(...) __builtin_sve_svst1_scatter_u64base_u64(__VA_ARGS__) -#define svst1_scatter_u64base_f64(...) __builtin_sve_svst1_scatter_u64base_f64(__VA_ARGS__) -#define svst1_scatter_u32base_f32(...) __builtin_sve_svst1_scatter_u32base_f32(__VA_ARGS__) -#define svst1_scatter_u32base_s32(...) __builtin_sve_svst1_scatter_u32base_s32(__VA_ARGS__) -#define svst1_scatter_u64base_s64(...) __builtin_sve_svst1_scatter_u64base_s64(__VA_ARGS__) -#define svst1_scatter_s32index_u32(...) __builtin_sve_svst1_scatter_s32index_u32(__VA_ARGS__) -#define svst1_scatter_s32index_f32(...) __builtin_sve_svst1_scatter_s32index_f32(__VA_ARGS__) -#define svst1_scatter_s32index_s32(...) __builtin_sve_svst1_scatter_s32index_s32(__VA_ARGS__) -#define svst1_scatter_u32index_u32(...) __builtin_sve_svst1_scatter_u32index_u32(__VA_ARGS__) -#define svst1_scatter_u32index_f32(...) __builtin_sve_svst1_scatter_u32index_f32(__VA_ARGS__) -#define svst1_scatter_u32index_s32(...) __builtin_sve_svst1_scatter_u32index_s32(__VA_ARGS__) -#define svst1_scatter_s64index_u64(...) __builtin_sve_svst1_scatter_s64index_u64(__VA_ARGS__) -#define svst1_scatter_s64index_f64(...) __builtin_sve_svst1_scatter_s64index_f64(__VA_ARGS__) -#define svst1_scatter_s64index_s64(...) __builtin_sve_svst1_scatter_s64index_s64(__VA_ARGS__) -#define svst1_scatter_u64index_u64(...) __builtin_sve_svst1_scatter_u64index_u64(__VA_ARGS__) -#define svst1_scatter_u64index_f64(...) __builtin_sve_svst1_scatter_u64index_f64(__VA_ARGS__) -#define svst1_scatter_u64index_s64(...) __builtin_sve_svst1_scatter_u64index_s64(__VA_ARGS__) -#define svst1_scatter_s32offset_u32(...) __builtin_sve_svst1_scatter_s32offset_u32(__VA_ARGS__) -#define svst1_scatter_s32offset_f32(...) __builtin_sve_svst1_scatter_s32offset_f32(__VA_ARGS__) -#define svst1_scatter_s32offset_s32(...) __builtin_sve_svst1_scatter_s32offset_s32(__VA_ARGS__) -#define svst1_scatter_u32offset_u32(...) __builtin_sve_svst1_scatter_u32offset_u32(__VA_ARGS__) -#define svst1_scatter_u32offset_f32(...) __builtin_sve_svst1_scatter_u32offset_f32(__VA_ARGS__) -#define svst1_scatter_u32offset_s32(...) __builtin_sve_svst1_scatter_u32offset_s32(__VA_ARGS__) -#define svst1_scatter_s64offset_u64(...) __builtin_sve_svst1_scatter_s64offset_u64(__VA_ARGS__) -#define svst1_scatter_s64offset_f64(...) __builtin_sve_svst1_scatter_s64offset_f64(__VA_ARGS__) -#define svst1_scatter_s64offset_s64(...) __builtin_sve_svst1_scatter_s64offset_s64(__VA_ARGS__) -#define svst1_scatter_u64offset_u64(...) __builtin_sve_svst1_scatter_u64offset_u64(__VA_ARGS__) -#define svst1_scatter_u64offset_f64(...) __builtin_sve_svst1_scatter_u64offset_f64(__VA_ARGS__) -#define svst1_scatter_u64offset_s64(...) __builtin_sve_svst1_scatter_u64offset_s64(__VA_ARGS__) -#define svst1_vnum_u8(...) __builtin_sve_svst1_vnum_u8(__VA_ARGS__) -#define svst1_vnum_u32(...) __builtin_sve_svst1_vnum_u32(__VA_ARGS__) -#define svst1_vnum_u64(...) __builtin_sve_svst1_vnum_u64(__VA_ARGS__) -#define svst1_vnum_u16(...) __builtin_sve_svst1_vnum_u16(__VA_ARGS__) -#define svst1_vnum_s8(...) __builtin_sve_svst1_vnum_s8(__VA_ARGS__) -#define svst1_vnum_f64(...) __builtin_sve_svst1_vnum_f64(__VA_ARGS__) -#define svst1_vnum_f32(...) __builtin_sve_svst1_vnum_f32(__VA_ARGS__) -#define svst1_vnum_f16(...) __builtin_sve_svst1_vnum_f16(__VA_ARGS__) -#define svst1_vnum_s32(...) __builtin_sve_svst1_vnum_s32(__VA_ARGS__) -#define svst1_vnum_s64(...) __builtin_sve_svst1_vnum_s64(__VA_ARGS__) -#define svst1_vnum_s16(...) __builtin_sve_svst1_vnum_s16(__VA_ARGS__) -#define svst1b_s32(...) __builtin_sve_svst1b_s32(__VA_ARGS__) -#define svst1b_s64(...) __builtin_sve_svst1b_s64(__VA_ARGS__) -#define svst1b_s16(...) __builtin_sve_svst1b_s16(__VA_ARGS__) -#define svst1b_u32(...) __builtin_sve_svst1b_u32(__VA_ARGS__) -#define svst1b_u64(...) __builtin_sve_svst1b_u64(__VA_ARGS__) -#define svst1b_u16(...) __builtin_sve_svst1b_u16(__VA_ARGS__) -#define svst1b_scatter_u32base_offset_u32(...) __builtin_sve_svst1b_scatter_u32base_offset_u32(__VA_ARGS__) -#define svst1b_scatter_u64base_offset_u64(...) __builtin_sve_svst1b_scatter_u64base_offset_u64(__VA_ARGS__) -#define svst1b_scatter_u32base_offset_s32(...) __builtin_sve_svst1b_scatter_u32base_offset_s32(__VA_ARGS__) -#define svst1b_scatter_u64base_offset_s64(...) __builtin_sve_svst1b_scatter_u64base_offset_s64(__VA_ARGS__) -#define svst1b_scatter_u32base_u32(...) __builtin_sve_svst1b_scatter_u32base_u32(__VA_ARGS__) -#define svst1b_scatter_u64base_u64(...) __builtin_sve_svst1b_scatter_u64base_u64(__VA_ARGS__) -#define svst1b_scatter_u32base_s32(...) __builtin_sve_svst1b_scatter_u32base_s32(__VA_ARGS__) -#define svst1b_scatter_u64base_s64(...) __builtin_sve_svst1b_scatter_u64base_s64(__VA_ARGS__) -#define svst1b_scatter_s32offset_s32(...) __builtin_sve_svst1b_scatter_s32offset_s32(__VA_ARGS__) -#define svst1b_scatter_s32offset_u32(...) __builtin_sve_svst1b_scatter_s32offset_u32(__VA_ARGS__) -#define svst1b_scatter_u32offset_s32(...) __builtin_sve_svst1b_scatter_u32offset_s32(__VA_ARGS__) -#define svst1b_scatter_u32offset_u32(...) __builtin_sve_svst1b_scatter_u32offset_u32(__VA_ARGS__) -#define svst1b_scatter_s64offset_s64(...) __builtin_sve_svst1b_scatter_s64offset_s64(__VA_ARGS__) -#define svst1b_scatter_s64offset_u64(...) __builtin_sve_svst1b_scatter_s64offset_u64(__VA_ARGS__) -#define svst1b_scatter_u64offset_s64(...) __builtin_sve_svst1b_scatter_u64offset_s64(__VA_ARGS__) -#define svst1b_scatter_u64offset_u64(...) __builtin_sve_svst1b_scatter_u64offset_u64(__VA_ARGS__) -#define svst1b_vnum_s32(...) __builtin_sve_svst1b_vnum_s32(__VA_ARGS__) -#define svst1b_vnum_s64(...) __builtin_sve_svst1b_vnum_s64(__VA_ARGS__) -#define svst1b_vnum_s16(...) __builtin_sve_svst1b_vnum_s16(__VA_ARGS__) -#define svst1b_vnum_u32(...) __builtin_sve_svst1b_vnum_u32(__VA_ARGS__) -#define svst1b_vnum_u64(...) __builtin_sve_svst1b_vnum_u64(__VA_ARGS__) -#define svst1b_vnum_u16(...) __builtin_sve_svst1b_vnum_u16(__VA_ARGS__) -#define svst1h_s32(...) __builtin_sve_svst1h_s32(__VA_ARGS__) -#define svst1h_s64(...) __builtin_sve_svst1h_s64(__VA_ARGS__) -#define svst1h_u32(...) __builtin_sve_svst1h_u32(__VA_ARGS__) -#define svst1h_u64(...) __builtin_sve_svst1h_u64(__VA_ARGS__) -#define svst1h_scatter_u32base_index_u32(...) __builtin_sve_svst1h_scatter_u32base_index_u32(__VA_ARGS__) -#define svst1h_scatter_u64base_index_u64(...) __builtin_sve_svst1h_scatter_u64base_index_u64(__VA_ARGS__) -#define svst1h_scatter_u32base_index_s32(...) __builtin_sve_svst1h_scatter_u32base_index_s32(__VA_ARGS__) -#define svst1h_scatter_u64base_index_s64(...) __builtin_sve_svst1h_scatter_u64base_index_s64(__VA_ARGS__) -#define svst1h_scatter_u32base_offset_u32(...) __builtin_sve_svst1h_scatter_u32base_offset_u32(__VA_ARGS__) -#define svst1h_scatter_u64base_offset_u64(...) __builtin_sve_svst1h_scatter_u64base_offset_u64(__VA_ARGS__) -#define svst1h_scatter_u32base_offset_s32(...) __builtin_sve_svst1h_scatter_u32base_offset_s32(__VA_ARGS__) -#define svst1h_scatter_u64base_offset_s64(...) __builtin_sve_svst1h_scatter_u64base_offset_s64(__VA_ARGS__) -#define svst1h_scatter_u32base_u32(...) __builtin_sve_svst1h_scatter_u32base_u32(__VA_ARGS__) -#define svst1h_scatter_u64base_u64(...) __builtin_sve_svst1h_scatter_u64base_u64(__VA_ARGS__) -#define svst1h_scatter_u32base_s32(...) __builtin_sve_svst1h_scatter_u32base_s32(__VA_ARGS__) -#define svst1h_scatter_u64base_s64(...) __builtin_sve_svst1h_scatter_u64base_s64(__VA_ARGS__) -#define svst1h_scatter_s32index_s32(...) __builtin_sve_svst1h_scatter_s32index_s32(__VA_ARGS__) -#define svst1h_scatter_s32index_u32(...) __builtin_sve_svst1h_scatter_s32index_u32(__VA_ARGS__) -#define svst1h_scatter_u32index_s32(...) __builtin_sve_svst1h_scatter_u32index_s32(__VA_ARGS__) -#define svst1h_scatter_u32index_u32(...) __builtin_sve_svst1h_scatter_u32index_u32(__VA_ARGS__) -#define svst1h_scatter_s64index_s64(...) __builtin_sve_svst1h_scatter_s64index_s64(__VA_ARGS__) -#define svst1h_scatter_s64index_u64(...) __builtin_sve_svst1h_scatter_s64index_u64(__VA_ARGS__) -#define svst1h_scatter_u64index_s64(...) __builtin_sve_svst1h_scatter_u64index_s64(__VA_ARGS__) -#define svst1h_scatter_u64index_u64(...) __builtin_sve_svst1h_scatter_u64index_u64(__VA_ARGS__) -#define svst1h_scatter_s32offset_s32(...) __builtin_sve_svst1h_scatter_s32offset_s32(__VA_ARGS__) -#define svst1h_scatter_s32offset_u32(...) __builtin_sve_svst1h_scatter_s32offset_u32(__VA_ARGS__) -#define svst1h_scatter_u32offset_s32(...) __builtin_sve_svst1h_scatter_u32offset_s32(__VA_ARGS__) -#define svst1h_scatter_u32offset_u32(...) __builtin_sve_svst1h_scatter_u32offset_u32(__VA_ARGS__) -#define svst1h_scatter_s64offset_s64(...) __builtin_sve_svst1h_scatter_s64offset_s64(__VA_ARGS__) -#define svst1h_scatter_s64offset_u64(...) __builtin_sve_svst1h_scatter_s64offset_u64(__VA_ARGS__) -#define svst1h_scatter_u64offset_s64(...) __builtin_sve_svst1h_scatter_u64offset_s64(__VA_ARGS__) -#define svst1h_scatter_u64offset_u64(...) __builtin_sve_svst1h_scatter_u64offset_u64(__VA_ARGS__) -#define svst1h_vnum_s32(...) __builtin_sve_svst1h_vnum_s32(__VA_ARGS__) -#define svst1h_vnum_s64(...) __builtin_sve_svst1h_vnum_s64(__VA_ARGS__) -#define svst1h_vnum_u32(...) __builtin_sve_svst1h_vnum_u32(__VA_ARGS__) -#define svst1h_vnum_u64(...) __builtin_sve_svst1h_vnum_u64(__VA_ARGS__) -#define svst1w_s64(...) __builtin_sve_svst1w_s64(__VA_ARGS__) -#define svst1w_u64(...) __builtin_sve_svst1w_u64(__VA_ARGS__) -#define svst1w_scatter_u64base_index_u64(...) __builtin_sve_svst1w_scatter_u64base_index_u64(__VA_ARGS__) -#define svst1w_scatter_u64base_index_s64(...) __builtin_sve_svst1w_scatter_u64base_index_s64(__VA_ARGS__) -#define svst1w_scatter_u64base_offset_u64(...) __builtin_sve_svst1w_scatter_u64base_offset_u64(__VA_ARGS__) -#define svst1w_scatter_u64base_offset_s64(...) __builtin_sve_svst1w_scatter_u64base_offset_s64(__VA_ARGS__) -#define svst1w_scatter_u64base_u64(...) __builtin_sve_svst1w_scatter_u64base_u64(__VA_ARGS__) -#define svst1w_scatter_u64base_s64(...) __builtin_sve_svst1w_scatter_u64base_s64(__VA_ARGS__) -#define svst1w_scatter_s64index_s64(...) __builtin_sve_svst1w_scatter_s64index_s64(__VA_ARGS__) -#define svst1w_scatter_s64index_u64(...) __builtin_sve_svst1w_scatter_s64index_u64(__VA_ARGS__) -#define svst1w_scatter_u64index_s64(...) __builtin_sve_svst1w_scatter_u64index_s64(__VA_ARGS__) -#define svst1w_scatter_u64index_u64(...) __builtin_sve_svst1w_scatter_u64index_u64(__VA_ARGS__) -#define svst1w_scatter_s64offset_s64(...) __builtin_sve_svst1w_scatter_s64offset_s64(__VA_ARGS__) -#define svst1w_scatter_s64offset_u64(...) __builtin_sve_svst1w_scatter_s64offset_u64(__VA_ARGS__) -#define svst1w_scatter_u64offset_s64(...) __builtin_sve_svst1w_scatter_u64offset_s64(__VA_ARGS__) -#define svst1w_scatter_u64offset_u64(...) __builtin_sve_svst1w_scatter_u64offset_u64(__VA_ARGS__) -#define svst1w_vnum_s64(...) __builtin_sve_svst1w_vnum_s64(__VA_ARGS__) -#define svst1w_vnum_u64(...) __builtin_sve_svst1w_vnum_u64(__VA_ARGS__) -#define svst2_u8(...) __builtin_sve_svst2_u8(__VA_ARGS__) -#define svst2_u32(...) __builtin_sve_svst2_u32(__VA_ARGS__) -#define svst2_u64(...) __builtin_sve_svst2_u64(__VA_ARGS__) -#define svst2_u16(...) __builtin_sve_svst2_u16(__VA_ARGS__) -#define svst2_s8(...) __builtin_sve_svst2_s8(__VA_ARGS__) -#define svst2_f64(...) __builtin_sve_svst2_f64(__VA_ARGS__) -#define svst2_f32(...) __builtin_sve_svst2_f32(__VA_ARGS__) -#define svst2_f16(...) __builtin_sve_svst2_f16(__VA_ARGS__) -#define svst2_s32(...) __builtin_sve_svst2_s32(__VA_ARGS__) -#define svst2_s64(...) __builtin_sve_svst2_s64(__VA_ARGS__) -#define svst2_s16(...) __builtin_sve_svst2_s16(__VA_ARGS__) -#define svst2_vnum_u8(...) __builtin_sve_svst2_vnum_u8(__VA_ARGS__) -#define svst2_vnum_u32(...) __builtin_sve_svst2_vnum_u32(__VA_ARGS__) -#define svst2_vnum_u64(...) __builtin_sve_svst2_vnum_u64(__VA_ARGS__) -#define svst2_vnum_u16(...) __builtin_sve_svst2_vnum_u16(__VA_ARGS__) -#define svst2_vnum_s8(...) __builtin_sve_svst2_vnum_s8(__VA_ARGS__) -#define svst2_vnum_f64(...) __builtin_sve_svst2_vnum_f64(__VA_ARGS__) -#define svst2_vnum_f32(...) __builtin_sve_svst2_vnum_f32(__VA_ARGS__) -#define svst2_vnum_f16(...) __builtin_sve_svst2_vnum_f16(__VA_ARGS__) -#define svst2_vnum_s32(...) __builtin_sve_svst2_vnum_s32(__VA_ARGS__) -#define svst2_vnum_s64(...) __builtin_sve_svst2_vnum_s64(__VA_ARGS__) -#define svst2_vnum_s16(...) __builtin_sve_svst2_vnum_s16(__VA_ARGS__) -#define svst3_u8(...) __builtin_sve_svst3_u8(__VA_ARGS__) -#define svst3_u32(...) __builtin_sve_svst3_u32(__VA_ARGS__) -#define svst3_u64(...) __builtin_sve_svst3_u64(__VA_ARGS__) -#define svst3_u16(...) __builtin_sve_svst3_u16(__VA_ARGS__) -#define svst3_s8(...) __builtin_sve_svst3_s8(__VA_ARGS__) -#define svst3_f64(...) __builtin_sve_svst3_f64(__VA_ARGS__) -#define svst3_f32(...) __builtin_sve_svst3_f32(__VA_ARGS__) -#define svst3_f16(...) __builtin_sve_svst3_f16(__VA_ARGS__) -#define svst3_s32(...) __builtin_sve_svst3_s32(__VA_ARGS__) -#define svst3_s64(...) __builtin_sve_svst3_s64(__VA_ARGS__) -#define svst3_s16(...) __builtin_sve_svst3_s16(__VA_ARGS__) -#define svst3_vnum_u8(...) __builtin_sve_svst3_vnum_u8(__VA_ARGS__) -#define svst3_vnum_u32(...) __builtin_sve_svst3_vnum_u32(__VA_ARGS__) -#define svst3_vnum_u64(...) __builtin_sve_svst3_vnum_u64(__VA_ARGS__) -#define svst3_vnum_u16(...) __builtin_sve_svst3_vnum_u16(__VA_ARGS__) -#define svst3_vnum_s8(...) __builtin_sve_svst3_vnum_s8(__VA_ARGS__) -#define svst3_vnum_f64(...) __builtin_sve_svst3_vnum_f64(__VA_ARGS__) -#define svst3_vnum_f32(...) __builtin_sve_svst3_vnum_f32(__VA_ARGS__) -#define svst3_vnum_f16(...) __builtin_sve_svst3_vnum_f16(__VA_ARGS__) -#define svst3_vnum_s32(...) __builtin_sve_svst3_vnum_s32(__VA_ARGS__) -#define svst3_vnum_s64(...) __builtin_sve_svst3_vnum_s64(__VA_ARGS__) -#define svst3_vnum_s16(...) __builtin_sve_svst3_vnum_s16(__VA_ARGS__) -#define svst4_u8(...) __builtin_sve_svst4_u8(__VA_ARGS__) -#define svst4_u32(...) __builtin_sve_svst4_u32(__VA_ARGS__) -#define svst4_u64(...) __builtin_sve_svst4_u64(__VA_ARGS__) -#define svst4_u16(...) __builtin_sve_svst4_u16(__VA_ARGS__) -#define svst4_s8(...) __builtin_sve_svst4_s8(__VA_ARGS__) -#define svst4_f64(...) __builtin_sve_svst4_f64(__VA_ARGS__) -#define svst4_f32(...) __builtin_sve_svst4_f32(__VA_ARGS__) -#define svst4_f16(...) __builtin_sve_svst4_f16(__VA_ARGS__) -#define svst4_s32(...) __builtin_sve_svst4_s32(__VA_ARGS__) -#define svst4_s64(...) __builtin_sve_svst4_s64(__VA_ARGS__) -#define svst4_s16(...) __builtin_sve_svst4_s16(__VA_ARGS__) -#define svst4_vnum_u8(...) __builtin_sve_svst4_vnum_u8(__VA_ARGS__) -#define svst4_vnum_u32(...) __builtin_sve_svst4_vnum_u32(__VA_ARGS__) -#define svst4_vnum_u64(...) __builtin_sve_svst4_vnum_u64(__VA_ARGS__) -#define svst4_vnum_u16(...) __builtin_sve_svst4_vnum_u16(__VA_ARGS__) -#define svst4_vnum_s8(...) __builtin_sve_svst4_vnum_s8(__VA_ARGS__) -#define svst4_vnum_f64(...) __builtin_sve_svst4_vnum_f64(__VA_ARGS__) -#define svst4_vnum_f32(...) __builtin_sve_svst4_vnum_f32(__VA_ARGS__) -#define svst4_vnum_f16(...) __builtin_sve_svst4_vnum_f16(__VA_ARGS__) -#define svst4_vnum_s32(...) __builtin_sve_svst4_vnum_s32(__VA_ARGS__) -#define svst4_vnum_s64(...) __builtin_sve_svst4_vnum_s64(__VA_ARGS__) -#define svst4_vnum_s16(...) __builtin_sve_svst4_vnum_s16(__VA_ARGS__) -#define svstnt1_u8(...) __builtin_sve_svstnt1_u8(__VA_ARGS__) -#define svstnt1_u32(...) __builtin_sve_svstnt1_u32(__VA_ARGS__) -#define svstnt1_u64(...) __builtin_sve_svstnt1_u64(__VA_ARGS__) -#define svstnt1_u16(...) __builtin_sve_svstnt1_u16(__VA_ARGS__) -#define svstnt1_s8(...) __builtin_sve_svstnt1_s8(__VA_ARGS__) -#define svstnt1_f64(...) __builtin_sve_svstnt1_f64(__VA_ARGS__) -#define svstnt1_f32(...) __builtin_sve_svstnt1_f32(__VA_ARGS__) -#define svstnt1_f16(...) __builtin_sve_svstnt1_f16(__VA_ARGS__) -#define svstnt1_s32(...) __builtin_sve_svstnt1_s32(__VA_ARGS__) -#define svstnt1_s64(...) __builtin_sve_svstnt1_s64(__VA_ARGS__) -#define svstnt1_s16(...) __builtin_sve_svstnt1_s16(__VA_ARGS__) -#define svstnt1_vnum_u8(...) __builtin_sve_svstnt1_vnum_u8(__VA_ARGS__) -#define svstnt1_vnum_u32(...) __builtin_sve_svstnt1_vnum_u32(__VA_ARGS__) -#define svstnt1_vnum_u64(...) __builtin_sve_svstnt1_vnum_u64(__VA_ARGS__) -#define svstnt1_vnum_u16(...) __builtin_sve_svstnt1_vnum_u16(__VA_ARGS__) -#define svstnt1_vnum_s8(...) __builtin_sve_svstnt1_vnum_s8(__VA_ARGS__) -#define svstnt1_vnum_f64(...) __builtin_sve_svstnt1_vnum_f64(__VA_ARGS__) -#define svstnt1_vnum_f32(...) __builtin_sve_svstnt1_vnum_f32(__VA_ARGS__) -#define svstnt1_vnum_f16(...) __builtin_sve_svstnt1_vnum_f16(__VA_ARGS__) -#define svstnt1_vnum_s32(...) __builtin_sve_svstnt1_vnum_s32(__VA_ARGS__) -#define svstnt1_vnum_s64(...) __builtin_sve_svstnt1_vnum_s64(__VA_ARGS__) -#define svstnt1_vnum_s16(...) __builtin_sve_svstnt1_vnum_s16(__VA_ARGS__) -#define svsub_n_f64_m(...) __builtin_sve_svsub_n_f64_m(__VA_ARGS__) -#define svsub_n_f32_m(...) __builtin_sve_svsub_n_f32_m(__VA_ARGS__) -#define svsub_n_f16_m(...) __builtin_sve_svsub_n_f16_m(__VA_ARGS__) -#define svsub_n_f64_x(...) __builtin_sve_svsub_n_f64_x(__VA_ARGS__) -#define svsub_n_f32_x(...) __builtin_sve_svsub_n_f32_x(__VA_ARGS__) -#define svsub_n_f16_x(...) __builtin_sve_svsub_n_f16_x(__VA_ARGS__) -#define svsub_n_f64_z(...) __builtin_sve_svsub_n_f64_z(__VA_ARGS__) -#define svsub_n_f32_z(...) __builtin_sve_svsub_n_f32_z(__VA_ARGS__) -#define svsub_n_f16_z(...) __builtin_sve_svsub_n_f16_z(__VA_ARGS__) -#define svsub_n_u8_m(...) __builtin_sve_svsub_n_u8_m(__VA_ARGS__) -#define svsub_n_u32_m(...) __builtin_sve_svsub_n_u32_m(__VA_ARGS__) -#define svsub_n_u64_m(...) __builtin_sve_svsub_n_u64_m(__VA_ARGS__) -#define svsub_n_u16_m(...) __builtin_sve_svsub_n_u16_m(__VA_ARGS__) -#define svsub_n_s8_m(...) __builtin_sve_svsub_n_s8_m(__VA_ARGS__) -#define svsub_n_s32_m(...) __builtin_sve_svsub_n_s32_m(__VA_ARGS__) -#define svsub_n_s64_m(...) __builtin_sve_svsub_n_s64_m(__VA_ARGS__) -#define svsub_n_s16_m(...) __builtin_sve_svsub_n_s16_m(__VA_ARGS__) -#define svsub_n_u8_x(...) __builtin_sve_svsub_n_u8_x(__VA_ARGS__) -#define svsub_n_u32_x(...) __builtin_sve_svsub_n_u32_x(__VA_ARGS__) -#define svsub_n_u64_x(...) __builtin_sve_svsub_n_u64_x(__VA_ARGS__) -#define svsub_n_u16_x(...) __builtin_sve_svsub_n_u16_x(__VA_ARGS__) -#define svsub_n_s8_x(...) __builtin_sve_svsub_n_s8_x(__VA_ARGS__) -#define svsub_n_s32_x(...) __builtin_sve_svsub_n_s32_x(__VA_ARGS__) -#define svsub_n_s64_x(...) __builtin_sve_svsub_n_s64_x(__VA_ARGS__) -#define svsub_n_s16_x(...) __builtin_sve_svsub_n_s16_x(__VA_ARGS__) -#define svsub_n_u8_z(...) __builtin_sve_svsub_n_u8_z(__VA_ARGS__) -#define svsub_n_u32_z(...) __builtin_sve_svsub_n_u32_z(__VA_ARGS__) -#define svsub_n_u64_z(...) __builtin_sve_svsub_n_u64_z(__VA_ARGS__) -#define svsub_n_u16_z(...) __builtin_sve_svsub_n_u16_z(__VA_ARGS__) -#define svsub_n_s8_z(...) __builtin_sve_svsub_n_s8_z(__VA_ARGS__) -#define svsub_n_s32_z(...) __builtin_sve_svsub_n_s32_z(__VA_ARGS__) -#define svsub_n_s64_z(...) __builtin_sve_svsub_n_s64_z(__VA_ARGS__) -#define svsub_n_s16_z(...) __builtin_sve_svsub_n_s16_z(__VA_ARGS__) -#define svsub_f64_m(...) __builtin_sve_svsub_f64_m(__VA_ARGS__) -#define svsub_f32_m(...) __builtin_sve_svsub_f32_m(__VA_ARGS__) -#define svsub_f16_m(...) __builtin_sve_svsub_f16_m(__VA_ARGS__) -#define svsub_f64_x(...) __builtin_sve_svsub_f64_x(__VA_ARGS__) -#define svsub_f32_x(...) __builtin_sve_svsub_f32_x(__VA_ARGS__) -#define svsub_f16_x(...) __builtin_sve_svsub_f16_x(__VA_ARGS__) -#define svsub_f64_z(...) __builtin_sve_svsub_f64_z(__VA_ARGS__) -#define svsub_f32_z(...) __builtin_sve_svsub_f32_z(__VA_ARGS__) -#define svsub_f16_z(...) __builtin_sve_svsub_f16_z(__VA_ARGS__) -#define svsub_u8_m(...) __builtin_sve_svsub_u8_m(__VA_ARGS__) -#define svsub_u32_m(...) __builtin_sve_svsub_u32_m(__VA_ARGS__) -#define svsub_u64_m(...) __builtin_sve_svsub_u64_m(__VA_ARGS__) -#define svsub_u16_m(...) __builtin_sve_svsub_u16_m(__VA_ARGS__) -#define svsub_s8_m(...) __builtin_sve_svsub_s8_m(__VA_ARGS__) -#define svsub_s32_m(...) __builtin_sve_svsub_s32_m(__VA_ARGS__) -#define svsub_s64_m(...) __builtin_sve_svsub_s64_m(__VA_ARGS__) -#define svsub_s16_m(...) __builtin_sve_svsub_s16_m(__VA_ARGS__) -#define svsub_u8_x(...) __builtin_sve_svsub_u8_x(__VA_ARGS__) -#define svsub_u32_x(...) __builtin_sve_svsub_u32_x(__VA_ARGS__) -#define svsub_u64_x(...) __builtin_sve_svsub_u64_x(__VA_ARGS__) -#define svsub_u16_x(...) __builtin_sve_svsub_u16_x(__VA_ARGS__) -#define svsub_s8_x(...) __builtin_sve_svsub_s8_x(__VA_ARGS__) -#define svsub_s32_x(...) __builtin_sve_svsub_s32_x(__VA_ARGS__) -#define svsub_s64_x(...) __builtin_sve_svsub_s64_x(__VA_ARGS__) -#define svsub_s16_x(...) __builtin_sve_svsub_s16_x(__VA_ARGS__) -#define svsub_u8_z(...) __builtin_sve_svsub_u8_z(__VA_ARGS__) -#define svsub_u32_z(...) __builtin_sve_svsub_u32_z(__VA_ARGS__) -#define svsub_u64_z(...) __builtin_sve_svsub_u64_z(__VA_ARGS__) -#define svsub_u16_z(...) __builtin_sve_svsub_u16_z(__VA_ARGS__) -#define svsub_s8_z(...) __builtin_sve_svsub_s8_z(__VA_ARGS__) -#define svsub_s32_z(...) __builtin_sve_svsub_s32_z(__VA_ARGS__) -#define svsub_s64_z(...) __builtin_sve_svsub_s64_z(__VA_ARGS__) -#define svsub_s16_z(...) __builtin_sve_svsub_s16_z(__VA_ARGS__) -#define svsubr_n_f64_m(...) __builtin_sve_svsubr_n_f64_m(__VA_ARGS__) -#define svsubr_n_f32_m(...) __builtin_sve_svsubr_n_f32_m(__VA_ARGS__) -#define svsubr_n_f16_m(...) __builtin_sve_svsubr_n_f16_m(__VA_ARGS__) -#define svsubr_n_f64_x(...) __builtin_sve_svsubr_n_f64_x(__VA_ARGS__) -#define svsubr_n_f32_x(...) __builtin_sve_svsubr_n_f32_x(__VA_ARGS__) -#define svsubr_n_f16_x(...) __builtin_sve_svsubr_n_f16_x(__VA_ARGS__) -#define svsubr_n_f64_z(...) __builtin_sve_svsubr_n_f64_z(__VA_ARGS__) -#define svsubr_n_f32_z(...) __builtin_sve_svsubr_n_f32_z(__VA_ARGS__) -#define svsubr_n_f16_z(...) __builtin_sve_svsubr_n_f16_z(__VA_ARGS__) -#define svsubr_n_u8_m(...) __builtin_sve_svsubr_n_u8_m(__VA_ARGS__) -#define svsubr_n_u32_m(...) __builtin_sve_svsubr_n_u32_m(__VA_ARGS__) -#define svsubr_n_u64_m(...) __builtin_sve_svsubr_n_u64_m(__VA_ARGS__) -#define svsubr_n_u16_m(...) __builtin_sve_svsubr_n_u16_m(__VA_ARGS__) -#define svsubr_n_s8_m(...) __builtin_sve_svsubr_n_s8_m(__VA_ARGS__) -#define svsubr_n_s32_m(...) __builtin_sve_svsubr_n_s32_m(__VA_ARGS__) -#define svsubr_n_s64_m(...) __builtin_sve_svsubr_n_s64_m(__VA_ARGS__) -#define svsubr_n_s16_m(...) __builtin_sve_svsubr_n_s16_m(__VA_ARGS__) -#define svsubr_n_u8_x(...) __builtin_sve_svsubr_n_u8_x(__VA_ARGS__) -#define svsubr_n_u32_x(...) __builtin_sve_svsubr_n_u32_x(__VA_ARGS__) -#define svsubr_n_u64_x(...) __builtin_sve_svsubr_n_u64_x(__VA_ARGS__) -#define svsubr_n_u16_x(...) __builtin_sve_svsubr_n_u16_x(__VA_ARGS__) -#define svsubr_n_s8_x(...) __builtin_sve_svsubr_n_s8_x(__VA_ARGS__) -#define svsubr_n_s32_x(...) __builtin_sve_svsubr_n_s32_x(__VA_ARGS__) -#define svsubr_n_s64_x(...) __builtin_sve_svsubr_n_s64_x(__VA_ARGS__) -#define svsubr_n_s16_x(...) __builtin_sve_svsubr_n_s16_x(__VA_ARGS__) -#define svsubr_n_u8_z(...) __builtin_sve_svsubr_n_u8_z(__VA_ARGS__) -#define svsubr_n_u32_z(...) __builtin_sve_svsubr_n_u32_z(__VA_ARGS__) -#define svsubr_n_u64_z(...) __builtin_sve_svsubr_n_u64_z(__VA_ARGS__) -#define svsubr_n_u16_z(...) __builtin_sve_svsubr_n_u16_z(__VA_ARGS__) -#define svsubr_n_s8_z(...) __builtin_sve_svsubr_n_s8_z(__VA_ARGS__) -#define svsubr_n_s32_z(...) __builtin_sve_svsubr_n_s32_z(__VA_ARGS__) -#define svsubr_n_s64_z(...) __builtin_sve_svsubr_n_s64_z(__VA_ARGS__) -#define svsubr_n_s16_z(...) __builtin_sve_svsubr_n_s16_z(__VA_ARGS__) -#define svsubr_f64_m(...) __builtin_sve_svsubr_f64_m(__VA_ARGS__) -#define svsubr_f32_m(...) __builtin_sve_svsubr_f32_m(__VA_ARGS__) -#define svsubr_f16_m(...) __builtin_sve_svsubr_f16_m(__VA_ARGS__) -#define svsubr_f64_x(...) __builtin_sve_svsubr_f64_x(__VA_ARGS__) -#define svsubr_f32_x(...) __builtin_sve_svsubr_f32_x(__VA_ARGS__) -#define svsubr_f16_x(...) __builtin_sve_svsubr_f16_x(__VA_ARGS__) -#define svsubr_f64_z(...) __builtin_sve_svsubr_f64_z(__VA_ARGS__) -#define svsubr_f32_z(...) __builtin_sve_svsubr_f32_z(__VA_ARGS__) -#define svsubr_f16_z(...) __builtin_sve_svsubr_f16_z(__VA_ARGS__) -#define svsubr_u8_m(...) __builtin_sve_svsubr_u8_m(__VA_ARGS__) -#define svsubr_u32_m(...) __builtin_sve_svsubr_u32_m(__VA_ARGS__) -#define svsubr_u64_m(...) __builtin_sve_svsubr_u64_m(__VA_ARGS__) -#define svsubr_u16_m(...) __builtin_sve_svsubr_u16_m(__VA_ARGS__) -#define svsubr_s8_m(...) __builtin_sve_svsubr_s8_m(__VA_ARGS__) -#define svsubr_s32_m(...) __builtin_sve_svsubr_s32_m(__VA_ARGS__) -#define svsubr_s64_m(...) __builtin_sve_svsubr_s64_m(__VA_ARGS__) -#define svsubr_s16_m(...) __builtin_sve_svsubr_s16_m(__VA_ARGS__) -#define svsubr_u8_x(...) __builtin_sve_svsubr_u8_x(__VA_ARGS__) -#define svsubr_u32_x(...) __builtin_sve_svsubr_u32_x(__VA_ARGS__) -#define svsubr_u64_x(...) __builtin_sve_svsubr_u64_x(__VA_ARGS__) -#define svsubr_u16_x(...) __builtin_sve_svsubr_u16_x(__VA_ARGS__) -#define svsubr_s8_x(...) __builtin_sve_svsubr_s8_x(__VA_ARGS__) -#define svsubr_s32_x(...) __builtin_sve_svsubr_s32_x(__VA_ARGS__) -#define svsubr_s64_x(...) __builtin_sve_svsubr_s64_x(__VA_ARGS__) -#define svsubr_s16_x(...) __builtin_sve_svsubr_s16_x(__VA_ARGS__) -#define svsubr_u8_z(...) __builtin_sve_svsubr_u8_z(__VA_ARGS__) -#define svsubr_u32_z(...) __builtin_sve_svsubr_u32_z(__VA_ARGS__) -#define svsubr_u64_z(...) __builtin_sve_svsubr_u64_z(__VA_ARGS__) -#define svsubr_u16_z(...) __builtin_sve_svsubr_u16_z(__VA_ARGS__) -#define svsubr_s8_z(...) __builtin_sve_svsubr_s8_z(__VA_ARGS__) -#define svsubr_s32_z(...) __builtin_sve_svsubr_s32_z(__VA_ARGS__) -#define svsubr_s64_z(...) __builtin_sve_svsubr_s64_z(__VA_ARGS__) -#define svsubr_s16_z(...) __builtin_sve_svsubr_s16_z(__VA_ARGS__) -#define svtbl_u8(...) __builtin_sve_svtbl_u8(__VA_ARGS__) -#define svtbl_u32(...) __builtin_sve_svtbl_u32(__VA_ARGS__) -#define svtbl_u64(...) __builtin_sve_svtbl_u64(__VA_ARGS__) -#define svtbl_u16(...) __builtin_sve_svtbl_u16(__VA_ARGS__) -#define svtbl_s8(...) __builtin_sve_svtbl_s8(__VA_ARGS__) -#define svtbl_f64(...) __builtin_sve_svtbl_f64(__VA_ARGS__) -#define svtbl_f32(...) __builtin_sve_svtbl_f32(__VA_ARGS__) -#define svtbl_f16(...) __builtin_sve_svtbl_f16(__VA_ARGS__) -#define svtbl_s32(...) __builtin_sve_svtbl_s32(__VA_ARGS__) -#define svtbl_s64(...) __builtin_sve_svtbl_s64(__VA_ARGS__) -#define svtbl_s16(...) __builtin_sve_svtbl_s16(__VA_ARGS__) -#define svtmad_f64(...) __builtin_sve_svtmad_f64(__VA_ARGS__) -#define svtmad_f32(...) __builtin_sve_svtmad_f32(__VA_ARGS__) -#define svtmad_f16(...) __builtin_sve_svtmad_f16(__VA_ARGS__) -#define svtrn1_u8(...) __builtin_sve_svtrn1_u8(__VA_ARGS__) -#define svtrn1_u32(...) __builtin_sve_svtrn1_u32(__VA_ARGS__) -#define svtrn1_u64(...) __builtin_sve_svtrn1_u64(__VA_ARGS__) -#define svtrn1_u16(...) __builtin_sve_svtrn1_u16(__VA_ARGS__) -#define svtrn1_s8(...) __builtin_sve_svtrn1_s8(__VA_ARGS__) -#define svtrn1_f64(...) __builtin_sve_svtrn1_f64(__VA_ARGS__) -#define svtrn1_f32(...) __builtin_sve_svtrn1_f32(__VA_ARGS__) -#define svtrn1_f16(...) __builtin_sve_svtrn1_f16(__VA_ARGS__) -#define svtrn1_s32(...) __builtin_sve_svtrn1_s32(__VA_ARGS__) -#define svtrn1_s64(...) __builtin_sve_svtrn1_s64(__VA_ARGS__) -#define svtrn1_s16(...) __builtin_sve_svtrn1_s16(__VA_ARGS__) -#define svtrn1_b8(...) __builtin_sve_svtrn1_b8(__VA_ARGS__) -#define svtrn1_b32(...) __builtin_sve_svtrn1_b32(__VA_ARGS__) -#define svtrn1_b64(...) __builtin_sve_svtrn1_b64(__VA_ARGS__) -#define svtrn1_b16(...) __builtin_sve_svtrn1_b16(__VA_ARGS__) -#define svtrn2_u8(...) __builtin_sve_svtrn2_u8(__VA_ARGS__) -#define svtrn2_u32(...) __builtin_sve_svtrn2_u32(__VA_ARGS__) -#define svtrn2_u64(...) __builtin_sve_svtrn2_u64(__VA_ARGS__) -#define svtrn2_u16(...) __builtin_sve_svtrn2_u16(__VA_ARGS__) -#define svtrn2_s8(...) __builtin_sve_svtrn2_s8(__VA_ARGS__) -#define svtrn2_f64(...) __builtin_sve_svtrn2_f64(__VA_ARGS__) -#define svtrn2_f32(...) __builtin_sve_svtrn2_f32(__VA_ARGS__) -#define svtrn2_f16(...) __builtin_sve_svtrn2_f16(__VA_ARGS__) -#define svtrn2_s32(...) __builtin_sve_svtrn2_s32(__VA_ARGS__) -#define svtrn2_s64(...) __builtin_sve_svtrn2_s64(__VA_ARGS__) -#define svtrn2_s16(...) __builtin_sve_svtrn2_s16(__VA_ARGS__) -#define svtrn2_b8(...) __builtin_sve_svtrn2_b8(__VA_ARGS__) -#define svtrn2_b32(...) __builtin_sve_svtrn2_b32(__VA_ARGS__) -#define svtrn2_b64(...) __builtin_sve_svtrn2_b64(__VA_ARGS__) -#define svtrn2_b16(...) __builtin_sve_svtrn2_b16(__VA_ARGS__) -#define svtsmul_f64(...) __builtin_sve_svtsmul_f64(__VA_ARGS__) -#define svtsmul_f32(...) __builtin_sve_svtsmul_f32(__VA_ARGS__) -#define svtsmul_f16(...) __builtin_sve_svtsmul_f16(__VA_ARGS__) -#define svtssel_f64(...) __builtin_sve_svtssel_f64(__VA_ARGS__) -#define svtssel_f32(...) __builtin_sve_svtssel_f32(__VA_ARGS__) -#define svtssel_f16(...) __builtin_sve_svtssel_f16(__VA_ARGS__) -#define svundef2_u8(...) __builtin_sve_svundef2_u8(__VA_ARGS__) -#define svundef2_u32(...) __builtin_sve_svundef2_u32(__VA_ARGS__) -#define svundef2_u64(...) __builtin_sve_svundef2_u64(__VA_ARGS__) -#define svundef2_u16(...) __builtin_sve_svundef2_u16(__VA_ARGS__) -#define svundef2_s8(...) __builtin_sve_svundef2_s8(__VA_ARGS__) -#define svundef2_f64(...) __builtin_sve_svundef2_f64(__VA_ARGS__) -#define svundef2_f32(...) __builtin_sve_svundef2_f32(__VA_ARGS__) -#define svundef2_f16(...) __builtin_sve_svundef2_f16(__VA_ARGS__) -#define svundef2_s32(...) __builtin_sve_svundef2_s32(__VA_ARGS__) -#define svundef2_s64(...) __builtin_sve_svundef2_s64(__VA_ARGS__) -#define svundef2_s16(...) __builtin_sve_svundef2_s16(__VA_ARGS__) -#define svundef3_u8(...) __builtin_sve_svundef3_u8(__VA_ARGS__) -#define svundef3_u32(...) __builtin_sve_svundef3_u32(__VA_ARGS__) -#define svundef3_u64(...) __builtin_sve_svundef3_u64(__VA_ARGS__) -#define svundef3_u16(...) __builtin_sve_svundef3_u16(__VA_ARGS__) -#define svundef3_s8(...) __builtin_sve_svundef3_s8(__VA_ARGS__) -#define svundef3_f64(...) __builtin_sve_svundef3_f64(__VA_ARGS__) -#define svundef3_f32(...) __builtin_sve_svundef3_f32(__VA_ARGS__) -#define svundef3_f16(...) __builtin_sve_svundef3_f16(__VA_ARGS__) -#define svundef3_s32(...) __builtin_sve_svundef3_s32(__VA_ARGS__) -#define svundef3_s64(...) __builtin_sve_svundef3_s64(__VA_ARGS__) -#define svundef3_s16(...) __builtin_sve_svundef3_s16(__VA_ARGS__) -#define svundef4_u8(...) __builtin_sve_svundef4_u8(__VA_ARGS__) -#define svundef4_u32(...) __builtin_sve_svundef4_u32(__VA_ARGS__) -#define svundef4_u64(...) __builtin_sve_svundef4_u64(__VA_ARGS__) -#define svundef4_u16(...) __builtin_sve_svundef4_u16(__VA_ARGS__) -#define svundef4_s8(...) __builtin_sve_svundef4_s8(__VA_ARGS__) -#define svundef4_f64(...) __builtin_sve_svundef4_f64(__VA_ARGS__) -#define svundef4_f32(...) __builtin_sve_svundef4_f32(__VA_ARGS__) -#define svundef4_f16(...) __builtin_sve_svundef4_f16(__VA_ARGS__) -#define svundef4_s32(...) __builtin_sve_svundef4_s32(__VA_ARGS__) -#define svundef4_s64(...) __builtin_sve_svundef4_s64(__VA_ARGS__) -#define svundef4_s16(...) __builtin_sve_svundef4_s16(__VA_ARGS__) -#define svundef_u8(...) __builtin_sve_svundef_u8(__VA_ARGS__) -#define svundef_u32(...) __builtin_sve_svundef_u32(__VA_ARGS__) -#define svundef_u64(...) __builtin_sve_svundef_u64(__VA_ARGS__) -#define svundef_u16(...) __builtin_sve_svundef_u16(__VA_ARGS__) -#define svundef_s8(...) __builtin_sve_svundef_s8(__VA_ARGS__) -#define svundef_f64(...) __builtin_sve_svundef_f64(__VA_ARGS__) -#define svundef_f32(...) __builtin_sve_svundef_f32(__VA_ARGS__) -#define svundef_f16(...) __builtin_sve_svundef_f16(__VA_ARGS__) -#define svundef_s32(...) __builtin_sve_svundef_s32(__VA_ARGS__) -#define svundef_s64(...) __builtin_sve_svundef_s64(__VA_ARGS__) -#define svundef_s16(...) __builtin_sve_svundef_s16(__VA_ARGS__) -#define svunpkhi_b(...) __builtin_sve_svunpkhi_b(__VA_ARGS__) -#define svunpkhi_s32(...) __builtin_sve_svunpkhi_s32(__VA_ARGS__) -#define svunpkhi_s64(...) __builtin_sve_svunpkhi_s64(__VA_ARGS__) -#define svunpkhi_s16(...) __builtin_sve_svunpkhi_s16(__VA_ARGS__) -#define svunpkhi_u32(...) __builtin_sve_svunpkhi_u32(__VA_ARGS__) -#define svunpkhi_u64(...) __builtin_sve_svunpkhi_u64(__VA_ARGS__) -#define svunpkhi_u16(...) __builtin_sve_svunpkhi_u16(__VA_ARGS__) -#define svunpklo_b(...) __builtin_sve_svunpklo_b(__VA_ARGS__) -#define svunpklo_s32(...) __builtin_sve_svunpklo_s32(__VA_ARGS__) -#define svunpklo_s64(...) __builtin_sve_svunpklo_s64(__VA_ARGS__) -#define svunpklo_s16(...) __builtin_sve_svunpklo_s16(__VA_ARGS__) -#define svunpklo_u32(...) __builtin_sve_svunpklo_u32(__VA_ARGS__) -#define svunpklo_u64(...) __builtin_sve_svunpklo_u64(__VA_ARGS__) -#define svunpklo_u16(...) __builtin_sve_svunpklo_u16(__VA_ARGS__) -#define svuzp1_u8(...) __builtin_sve_svuzp1_u8(__VA_ARGS__) -#define svuzp1_u32(...) __builtin_sve_svuzp1_u32(__VA_ARGS__) -#define svuzp1_u64(...) __builtin_sve_svuzp1_u64(__VA_ARGS__) -#define svuzp1_u16(...) __builtin_sve_svuzp1_u16(__VA_ARGS__) -#define svuzp1_s8(...) __builtin_sve_svuzp1_s8(__VA_ARGS__) -#define svuzp1_f64(...) __builtin_sve_svuzp1_f64(__VA_ARGS__) -#define svuzp1_f32(...) __builtin_sve_svuzp1_f32(__VA_ARGS__) -#define svuzp1_f16(...) __builtin_sve_svuzp1_f16(__VA_ARGS__) -#define svuzp1_s32(...) __builtin_sve_svuzp1_s32(__VA_ARGS__) -#define svuzp1_s64(...) __builtin_sve_svuzp1_s64(__VA_ARGS__) -#define svuzp1_s16(...) __builtin_sve_svuzp1_s16(__VA_ARGS__) -#define svuzp1_b8(...) __builtin_sve_svuzp1_b8(__VA_ARGS__) -#define svuzp1_b32(...) __builtin_sve_svuzp1_b32(__VA_ARGS__) -#define svuzp1_b64(...) __builtin_sve_svuzp1_b64(__VA_ARGS__) -#define svuzp1_b16(...) __builtin_sve_svuzp1_b16(__VA_ARGS__) -#define svuzp2_u8(...) __builtin_sve_svuzp2_u8(__VA_ARGS__) -#define svuzp2_u32(...) __builtin_sve_svuzp2_u32(__VA_ARGS__) -#define svuzp2_u64(...) __builtin_sve_svuzp2_u64(__VA_ARGS__) -#define svuzp2_u16(...) __builtin_sve_svuzp2_u16(__VA_ARGS__) -#define svuzp2_s8(...) __builtin_sve_svuzp2_s8(__VA_ARGS__) -#define svuzp2_f64(...) __builtin_sve_svuzp2_f64(__VA_ARGS__) -#define svuzp2_f32(...) __builtin_sve_svuzp2_f32(__VA_ARGS__) -#define svuzp2_f16(...) __builtin_sve_svuzp2_f16(__VA_ARGS__) -#define svuzp2_s32(...) __builtin_sve_svuzp2_s32(__VA_ARGS__) -#define svuzp2_s64(...) __builtin_sve_svuzp2_s64(__VA_ARGS__) -#define svuzp2_s16(...) __builtin_sve_svuzp2_s16(__VA_ARGS__) -#define svuzp2_b8(...) __builtin_sve_svuzp2_b8(__VA_ARGS__) -#define svuzp2_b32(...) __builtin_sve_svuzp2_b32(__VA_ARGS__) -#define svuzp2_b64(...) __builtin_sve_svuzp2_b64(__VA_ARGS__) -#define svuzp2_b16(...) __builtin_sve_svuzp2_b16(__VA_ARGS__) -#define svwhilele_b8_s32(...) __builtin_sve_svwhilele_b8_s32(__VA_ARGS__) -#define svwhilele_b32_s32(...) __builtin_sve_svwhilele_b32_s32(__VA_ARGS__) -#define svwhilele_b64_s32(...) __builtin_sve_svwhilele_b64_s32(__VA_ARGS__) -#define svwhilele_b16_s32(...) __builtin_sve_svwhilele_b16_s32(__VA_ARGS__) -#define svwhilele_b8_s64(...) __builtin_sve_svwhilele_b8_s64(__VA_ARGS__) -#define svwhilele_b32_s64(...) __builtin_sve_svwhilele_b32_s64(__VA_ARGS__) -#define svwhilele_b64_s64(...) __builtin_sve_svwhilele_b64_s64(__VA_ARGS__) -#define svwhilele_b16_s64(...) __builtin_sve_svwhilele_b16_s64(__VA_ARGS__) -#define svwhilele_b8_u32(...) __builtin_sve_svwhilele_b8_u32(__VA_ARGS__) -#define svwhilele_b32_u32(...) __builtin_sve_svwhilele_b32_u32(__VA_ARGS__) -#define svwhilele_b64_u32(...) __builtin_sve_svwhilele_b64_u32(__VA_ARGS__) -#define svwhilele_b16_u32(...) __builtin_sve_svwhilele_b16_u32(__VA_ARGS__) -#define svwhilele_b8_u64(...) __builtin_sve_svwhilele_b8_u64(__VA_ARGS__) -#define svwhilele_b32_u64(...) __builtin_sve_svwhilele_b32_u64(__VA_ARGS__) -#define svwhilele_b64_u64(...) __builtin_sve_svwhilele_b64_u64(__VA_ARGS__) -#define svwhilele_b16_u64(...) __builtin_sve_svwhilele_b16_u64(__VA_ARGS__) -#define svwhilelt_b8_u32(...) __builtin_sve_svwhilelt_b8_u32(__VA_ARGS__) -#define svwhilelt_b32_u32(...) __builtin_sve_svwhilelt_b32_u32(__VA_ARGS__) -#define svwhilelt_b64_u32(...) __builtin_sve_svwhilelt_b64_u32(__VA_ARGS__) -#define svwhilelt_b16_u32(...) __builtin_sve_svwhilelt_b16_u32(__VA_ARGS__) -#define svwhilelt_b8_u64(...) __builtin_sve_svwhilelt_b8_u64(__VA_ARGS__) -#define svwhilelt_b32_u64(...) __builtin_sve_svwhilelt_b32_u64(__VA_ARGS__) -#define svwhilelt_b64_u64(...) __builtin_sve_svwhilelt_b64_u64(__VA_ARGS__) -#define svwhilelt_b16_u64(...) __builtin_sve_svwhilelt_b16_u64(__VA_ARGS__) -#define svwhilelt_b8_s32(...) __builtin_sve_svwhilelt_b8_s32(__VA_ARGS__) -#define svwhilelt_b32_s32(...) __builtin_sve_svwhilelt_b32_s32(__VA_ARGS__) -#define svwhilelt_b64_s32(...) __builtin_sve_svwhilelt_b64_s32(__VA_ARGS__) -#define svwhilelt_b16_s32(...) __builtin_sve_svwhilelt_b16_s32(__VA_ARGS__) -#define svwhilelt_b8_s64(...) __builtin_sve_svwhilelt_b8_s64(__VA_ARGS__) -#define svwhilelt_b32_s64(...) __builtin_sve_svwhilelt_b32_s64(__VA_ARGS__) -#define svwhilelt_b64_s64(...) __builtin_sve_svwhilelt_b64_s64(__VA_ARGS__) -#define svwhilelt_b16_s64(...) __builtin_sve_svwhilelt_b16_s64(__VA_ARGS__) -#define svwrffr(...) __builtin_sve_svwrffr(__VA_ARGS__) -#define svzip1_u8(...) __builtin_sve_svzip1_u8(__VA_ARGS__) -#define svzip1_u32(...) __builtin_sve_svzip1_u32(__VA_ARGS__) -#define svzip1_u64(...) __builtin_sve_svzip1_u64(__VA_ARGS__) -#define svzip1_u16(...) __builtin_sve_svzip1_u16(__VA_ARGS__) -#define svzip1_s8(...) __builtin_sve_svzip1_s8(__VA_ARGS__) -#define svzip1_f64(...) __builtin_sve_svzip1_f64(__VA_ARGS__) -#define svzip1_f32(...) __builtin_sve_svzip1_f32(__VA_ARGS__) -#define svzip1_f16(...) __builtin_sve_svzip1_f16(__VA_ARGS__) -#define svzip1_s32(...) __builtin_sve_svzip1_s32(__VA_ARGS__) -#define svzip1_s64(...) __builtin_sve_svzip1_s64(__VA_ARGS__) -#define svzip1_s16(...) __builtin_sve_svzip1_s16(__VA_ARGS__) -#define svzip1_b8(...) __builtin_sve_svzip1_b8(__VA_ARGS__) -#define svzip1_b32(...) __builtin_sve_svzip1_b32(__VA_ARGS__) -#define svzip1_b64(...) __builtin_sve_svzip1_b64(__VA_ARGS__) -#define svzip1_b16(...) __builtin_sve_svzip1_b16(__VA_ARGS__) -#define svzip2_u8(...) __builtin_sve_svzip2_u8(__VA_ARGS__) -#define svzip2_u32(...) __builtin_sve_svzip2_u32(__VA_ARGS__) -#define svzip2_u64(...) __builtin_sve_svzip2_u64(__VA_ARGS__) -#define svzip2_u16(...) __builtin_sve_svzip2_u16(__VA_ARGS__) -#define svzip2_s8(...) __builtin_sve_svzip2_s8(__VA_ARGS__) -#define svzip2_f64(...) __builtin_sve_svzip2_f64(__VA_ARGS__) -#define svzip2_f32(...) __builtin_sve_svzip2_f32(__VA_ARGS__) -#define svzip2_f16(...) __builtin_sve_svzip2_f16(__VA_ARGS__) -#define svzip2_s32(...) __builtin_sve_svzip2_s32(__VA_ARGS__) -#define svzip2_s64(...) __builtin_sve_svzip2_s64(__VA_ARGS__) -#define svzip2_s16(...) __builtin_sve_svzip2_s16(__VA_ARGS__) -#define svzip2_b8(...) __builtin_sve_svzip2_b8(__VA_ARGS__) -#define svzip2_b32(...) __builtin_sve_svzip2_b32(__VA_ARGS__) -#define svzip2_b64(...) __builtin_sve_svzip2_b64(__VA_ARGS__) -#define svzip2_b16(...) __builtin_sve_svzip2_b16(__VA_ARGS__) +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_n_f64_m))) +svfloat64_t svabd_n_f64_m(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_n_f32_m))) +svfloat32_t svabd_n_f32_m(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_n_f16_m))) +svfloat16_t svabd_n_f16_m(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_n_f64_x))) +svfloat64_t svabd_n_f64_x(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_n_f32_x))) +svfloat32_t svabd_n_f32_x(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_n_f16_x))) +svfloat16_t svabd_n_f16_x(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_n_f64_z))) +svfloat64_t svabd_n_f64_z(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_n_f32_z))) +svfloat32_t svabd_n_f32_z(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_n_f16_z))) +svfloat16_t svabd_n_f16_z(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_n_s8_m))) +svint8_t svabd_n_s8_m(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_n_s32_m))) +svint32_t svabd_n_s32_m(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_n_s64_m))) +svint64_t svabd_n_s64_m(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_n_s16_m))) +svint16_t svabd_n_s16_m(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_n_s8_x))) +svint8_t svabd_n_s8_x(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_n_s32_x))) +svint32_t svabd_n_s32_x(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_n_s64_x))) +svint64_t svabd_n_s64_x(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_n_s16_x))) +svint16_t svabd_n_s16_x(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_n_s8_z))) +svint8_t svabd_n_s8_z(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_n_s32_z))) +svint32_t svabd_n_s32_z(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_n_s64_z))) +svint64_t svabd_n_s64_z(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_n_s16_z))) +svint16_t svabd_n_s16_z(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_n_u8_m))) +svuint8_t svabd_n_u8_m(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_n_u32_m))) +svuint32_t svabd_n_u32_m(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_n_u64_m))) +svuint64_t svabd_n_u64_m(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_n_u16_m))) +svuint16_t svabd_n_u16_m(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_n_u8_x))) +svuint8_t svabd_n_u8_x(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_n_u32_x))) +svuint32_t svabd_n_u32_x(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_n_u64_x))) +svuint64_t svabd_n_u64_x(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_n_u16_x))) +svuint16_t svabd_n_u16_x(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_n_u8_z))) +svuint8_t svabd_n_u8_z(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_n_u32_z))) +svuint32_t svabd_n_u32_z(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_n_u64_z))) +svuint64_t svabd_n_u64_z(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_n_u16_z))) +svuint16_t svabd_n_u16_z(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_f64_m))) +svfloat64_t svabd_f64_m(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_f32_m))) +svfloat32_t svabd_f32_m(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_f16_m))) +svfloat16_t svabd_f16_m(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_f64_x))) +svfloat64_t svabd_f64_x(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_f32_x))) +svfloat32_t svabd_f32_x(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_f16_x))) +svfloat16_t svabd_f16_x(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_f64_z))) +svfloat64_t svabd_f64_z(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_f32_z))) +svfloat32_t svabd_f32_z(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_f16_z))) +svfloat16_t svabd_f16_z(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_s8_m))) +svint8_t svabd_s8_m(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_s32_m))) +svint32_t svabd_s32_m(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_s64_m))) +svint64_t svabd_s64_m(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_s16_m))) +svint16_t svabd_s16_m(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_s8_x))) +svint8_t svabd_s8_x(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_s32_x))) +svint32_t svabd_s32_x(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_s64_x))) +svint64_t svabd_s64_x(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_s16_x))) +svint16_t svabd_s16_x(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_s8_z))) +svint8_t svabd_s8_z(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_s32_z))) +svint32_t svabd_s32_z(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_s64_z))) +svint64_t svabd_s64_z(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_s16_z))) +svint16_t svabd_s16_z(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_u8_m))) +svuint8_t svabd_u8_m(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_u32_m))) +svuint32_t svabd_u32_m(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_u64_m))) +svuint64_t svabd_u64_m(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_u16_m))) +svuint16_t svabd_u16_m(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_u8_x))) +svuint8_t svabd_u8_x(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_u32_x))) +svuint32_t svabd_u32_x(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_u64_x))) +svuint64_t svabd_u64_x(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_u16_x))) +svuint16_t svabd_u16_x(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_u8_z))) +svuint8_t svabd_u8_z(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_u32_z))) +svuint32_t svabd_u32_z(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_u64_z))) +svuint64_t svabd_u64_z(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_u16_z))) +svuint16_t svabd_u16_z(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabs_f64_m))) +svfloat64_t svabs_f64_m(svfloat64_t, svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabs_f32_m))) +svfloat32_t svabs_f32_m(svfloat32_t, svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabs_f16_m))) +svfloat16_t svabs_f16_m(svfloat16_t, svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabs_f64_x))) +svfloat64_t svabs_f64_x(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabs_f32_x))) +svfloat32_t svabs_f32_x(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabs_f16_x))) +svfloat16_t svabs_f16_x(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabs_f64_z))) +svfloat64_t svabs_f64_z(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabs_f32_z))) +svfloat32_t svabs_f32_z(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabs_f16_z))) +svfloat16_t svabs_f16_z(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabs_s8_m))) +svint8_t svabs_s8_m(svint8_t, svbool_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabs_s32_m))) +svint32_t svabs_s32_m(svint32_t, svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabs_s64_m))) +svint64_t svabs_s64_m(svint64_t, svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabs_s16_m))) +svint16_t svabs_s16_m(svint16_t, svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabs_s8_x))) +svint8_t svabs_s8_x(svbool_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabs_s32_x))) +svint32_t svabs_s32_x(svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabs_s64_x))) +svint64_t svabs_s64_x(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabs_s16_x))) +svint16_t svabs_s16_x(svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabs_s8_z))) +svint8_t svabs_s8_z(svbool_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabs_s32_z))) +svint32_t svabs_s32_z(svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabs_s64_z))) +svint64_t svabs_s64_z(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabs_s16_z))) +svint16_t svabs_s16_z(svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svacge_n_f64))) +svbool_t svacge_n_f64(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svacge_n_f32))) +svbool_t svacge_n_f32(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svacge_n_f16))) +svbool_t svacge_n_f16(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svacge_f64))) +svbool_t svacge_f64(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svacge_f32))) +svbool_t svacge_f32(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svacge_f16))) +svbool_t svacge_f16(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svacgt_n_f64))) +svbool_t svacgt_n_f64(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svacgt_n_f32))) +svbool_t svacgt_n_f32(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svacgt_n_f16))) +svbool_t svacgt_n_f16(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svacgt_f64))) +svbool_t svacgt_f64(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svacgt_f32))) +svbool_t svacgt_f32(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svacgt_f16))) +svbool_t svacgt_f16(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svacle_n_f64))) +svbool_t svacle_n_f64(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svacle_n_f32))) +svbool_t svacle_n_f32(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svacle_n_f16))) +svbool_t svacle_n_f16(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svacle_f64))) +svbool_t svacle_f64(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svacle_f32))) +svbool_t svacle_f32(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svacle_f16))) +svbool_t svacle_f16(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaclt_n_f64))) +svbool_t svaclt_n_f64(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaclt_n_f32))) +svbool_t svaclt_n_f32(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaclt_n_f16))) +svbool_t svaclt_n_f16(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaclt_f64))) +svbool_t svaclt_f64(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaclt_f32))) +svbool_t svaclt_f32(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaclt_f16))) +svbool_t svaclt_f16(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_n_f64_m))) +svfloat64_t svadd_n_f64_m(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_n_f32_m))) +svfloat32_t svadd_n_f32_m(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_n_f16_m))) +svfloat16_t svadd_n_f16_m(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_n_f64_x))) +svfloat64_t svadd_n_f64_x(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_n_f32_x))) +svfloat32_t svadd_n_f32_x(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_n_f16_x))) +svfloat16_t svadd_n_f16_x(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_n_f64_z))) +svfloat64_t svadd_n_f64_z(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_n_f32_z))) +svfloat32_t svadd_n_f32_z(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_n_f16_z))) +svfloat16_t svadd_n_f16_z(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_n_u8_m))) +svuint8_t svadd_n_u8_m(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_n_u32_m))) +svuint32_t svadd_n_u32_m(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_n_u64_m))) +svuint64_t svadd_n_u64_m(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_n_u16_m))) +svuint16_t svadd_n_u16_m(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_n_s8_m))) +svint8_t svadd_n_s8_m(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_n_s32_m))) +svint32_t svadd_n_s32_m(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_n_s64_m))) +svint64_t svadd_n_s64_m(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_n_s16_m))) +svint16_t svadd_n_s16_m(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_n_u8_x))) +svuint8_t svadd_n_u8_x(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_n_u32_x))) +svuint32_t svadd_n_u32_x(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_n_u64_x))) +svuint64_t svadd_n_u64_x(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_n_u16_x))) +svuint16_t svadd_n_u16_x(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_n_s8_x))) +svint8_t svadd_n_s8_x(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_n_s32_x))) +svint32_t svadd_n_s32_x(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_n_s64_x))) +svint64_t svadd_n_s64_x(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_n_s16_x))) +svint16_t svadd_n_s16_x(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_n_u8_z))) +svuint8_t svadd_n_u8_z(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_n_u32_z))) +svuint32_t svadd_n_u32_z(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_n_u64_z))) +svuint64_t svadd_n_u64_z(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_n_u16_z))) +svuint16_t svadd_n_u16_z(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_n_s8_z))) +svint8_t svadd_n_s8_z(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_n_s32_z))) +svint32_t svadd_n_s32_z(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_n_s64_z))) +svint64_t svadd_n_s64_z(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_n_s16_z))) +svint16_t svadd_n_s16_z(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_f64_m))) +svfloat64_t svadd_f64_m(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_f32_m))) +svfloat32_t svadd_f32_m(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_f16_m))) +svfloat16_t svadd_f16_m(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_f64_x))) +svfloat64_t svadd_f64_x(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_f32_x))) +svfloat32_t svadd_f32_x(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_f16_x))) +svfloat16_t svadd_f16_x(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_f64_z))) +svfloat64_t svadd_f64_z(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_f32_z))) +svfloat32_t svadd_f32_z(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_f16_z))) +svfloat16_t svadd_f16_z(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_u8_m))) +svuint8_t svadd_u8_m(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_u32_m))) +svuint32_t svadd_u32_m(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_u64_m))) +svuint64_t svadd_u64_m(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_u16_m))) +svuint16_t svadd_u16_m(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_s8_m))) +svint8_t svadd_s8_m(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_s32_m))) +svint32_t svadd_s32_m(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_s64_m))) +svint64_t svadd_s64_m(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_s16_m))) +svint16_t svadd_s16_m(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_u8_x))) +svuint8_t svadd_u8_x(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_u32_x))) +svuint32_t svadd_u32_x(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_u64_x))) +svuint64_t svadd_u64_x(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_u16_x))) +svuint16_t svadd_u16_x(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_s8_x))) +svint8_t svadd_s8_x(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_s32_x))) +svint32_t svadd_s32_x(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_s64_x))) +svint64_t svadd_s64_x(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_s16_x))) +svint16_t svadd_s16_x(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_u8_z))) +svuint8_t svadd_u8_z(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_u32_z))) +svuint32_t svadd_u32_z(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_u64_z))) +svuint64_t svadd_u64_z(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_u16_z))) +svuint16_t svadd_u16_z(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_s8_z))) +svint8_t svadd_s8_z(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_s32_z))) +svint32_t svadd_s32_z(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_s64_z))) +svint64_t svadd_s64_z(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadd_s16_z))) +svint16_t svadd_s16_z(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadda_f64))) +float64_t svadda_f64(svbool_t, float64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadda_f32))) +float32_t svadda_f32(svbool_t, float32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadda_f16))) +float16_t svadda_f16(svbool_t, float16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddv_s8))) +int64_t svaddv_s8(svbool_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddv_s32))) +int64_t svaddv_s32(svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddv_s64))) +int64_t svaddv_s64(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddv_s16))) +int64_t svaddv_s16(svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddv_u8))) +uint64_t svaddv_u8(svbool_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddv_u32))) +uint64_t svaddv_u32(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddv_u64))) +uint64_t svaddv_u64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddv_u16))) +uint64_t svaddv_u16(svbool_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddv_f64))) +float64_t svaddv_f64(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddv_f32))) +float32_t svaddv_f32(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddv_f16))) +float16_t svaddv_f16(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadrb_u32base_u32offset))) +svuint32_t svadrb_u32base_u32offset(svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadrb_u64base_u64offset))) +svuint64_t svadrb_u64base_u64offset(svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadrb_u32base_s32offset))) +svuint32_t svadrb_u32base_s32offset(svuint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadrb_u64base_s64offset))) +svuint64_t svadrb_u64base_s64offset(svuint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadrd_u32base_u32index))) +svuint32_t svadrd_u32base_u32index(svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadrd_u64base_u64index))) +svuint64_t svadrd_u64base_u64index(svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadrd_u32base_s32index))) +svuint32_t svadrd_u32base_s32index(svuint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadrd_u64base_s64index))) +svuint64_t svadrd_u64base_s64index(svuint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadrh_u32base_u32index))) +svuint32_t svadrh_u32base_u32index(svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadrh_u64base_u64index))) +svuint64_t svadrh_u64base_u64index(svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadrh_u32base_s32index))) +svuint32_t svadrh_u32base_s32index(svuint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadrh_u64base_s64index))) +svuint64_t svadrh_u64base_s64index(svuint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadrw_u32base_u32index))) +svuint32_t svadrw_u32base_u32index(svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadrw_u64base_u64index))) +svuint64_t svadrw_u64base_u64index(svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadrw_u32base_s32index))) +svuint32_t svadrw_u32base_s32index(svuint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadrw_u64base_s64index))) +svuint64_t svadrw_u64base_s64index(svuint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_b_z))) +svbool_t svand_b_z(svbool_t, svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_n_u8_m))) +svuint8_t svand_n_u8_m(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_n_u32_m))) +svuint32_t svand_n_u32_m(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_n_u64_m))) +svuint64_t svand_n_u64_m(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_n_u16_m))) +svuint16_t svand_n_u16_m(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_n_s8_m))) +svint8_t svand_n_s8_m(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_n_s32_m))) +svint32_t svand_n_s32_m(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_n_s64_m))) +svint64_t svand_n_s64_m(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_n_s16_m))) +svint16_t svand_n_s16_m(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_n_u8_x))) +svuint8_t svand_n_u8_x(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_n_u32_x))) +svuint32_t svand_n_u32_x(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_n_u64_x))) +svuint64_t svand_n_u64_x(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_n_u16_x))) +svuint16_t svand_n_u16_x(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_n_s8_x))) +svint8_t svand_n_s8_x(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_n_s32_x))) +svint32_t svand_n_s32_x(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_n_s64_x))) +svint64_t svand_n_s64_x(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_n_s16_x))) +svint16_t svand_n_s16_x(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_n_u8_z))) +svuint8_t svand_n_u8_z(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_n_u32_z))) +svuint32_t svand_n_u32_z(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_n_u64_z))) +svuint64_t svand_n_u64_z(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_n_u16_z))) +svuint16_t svand_n_u16_z(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_n_s8_z))) +svint8_t svand_n_s8_z(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_n_s32_z))) +svint32_t svand_n_s32_z(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_n_s64_z))) +svint64_t svand_n_s64_z(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_n_s16_z))) +svint16_t svand_n_s16_z(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_u8_m))) +svuint8_t svand_u8_m(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_u32_m))) +svuint32_t svand_u32_m(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_u64_m))) +svuint64_t svand_u64_m(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_u16_m))) +svuint16_t svand_u16_m(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_s8_m))) +svint8_t svand_s8_m(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_s32_m))) +svint32_t svand_s32_m(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_s64_m))) +svint64_t svand_s64_m(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_s16_m))) +svint16_t svand_s16_m(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_u8_x))) +svuint8_t svand_u8_x(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_u32_x))) +svuint32_t svand_u32_x(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_u64_x))) +svuint64_t svand_u64_x(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_u16_x))) +svuint16_t svand_u16_x(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_s8_x))) +svint8_t svand_s8_x(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_s32_x))) +svint32_t svand_s32_x(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_s64_x))) +svint64_t svand_s64_x(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_s16_x))) +svint16_t svand_s16_x(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_u8_z))) +svuint8_t svand_u8_z(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_u32_z))) +svuint32_t svand_u32_z(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_u64_z))) +svuint64_t svand_u64_z(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_u16_z))) +svuint16_t svand_u16_z(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_s8_z))) +svint8_t svand_s8_z(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_s32_z))) +svint32_t svand_s32_z(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_s64_z))) +svint64_t svand_s64_z(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svand_s16_z))) +svint16_t svand_s16_z(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svandv_u8))) +uint8_t svandv_u8(svbool_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svandv_u32))) +uint32_t svandv_u32(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svandv_u64))) +uint64_t svandv_u64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svandv_u16))) +uint16_t svandv_u16(svbool_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svandv_s8))) +int8_t svandv_s8(svbool_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svandv_s32))) +int32_t svandv_s32(svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svandv_s64))) +int64_t svandv_s64(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svandv_s16))) +int16_t svandv_s16(svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasr_n_s8_m))) +svint8_t svasr_n_s8_m(svbool_t, svint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasr_n_s32_m))) +svint32_t svasr_n_s32_m(svbool_t, svint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasr_n_s64_m))) +svint64_t svasr_n_s64_m(svbool_t, svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasr_n_s16_m))) +svint16_t svasr_n_s16_m(svbool_t, svint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasr_n_s8_x))) +svint8_t svasr_n_s8_x(svbool_t, svint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasr_n_s32_x))) +svint32_t svasr_n_s32_x(svbool_t, svint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasr_n_s64_x))) +svint64_t svasr_n_s64_x(svbool_t, svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasr_n_s16_x))) +svint16_t svasr_n_s16_x(svbool_t, svint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasr_n_s8_z))) +svint8_t svasr_n_s8_z(svbool_t, svint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasr_n_s32_z))) +svint32_t svasr_n_s32_z(svbool_t, svint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasr_n_s64_z))) +svint64_t svasr_n_s64_z(svbool_t, svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasr_n_s16_z))) +svint16_t svasr_n_s16_z(svbool_t, svint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasr_s8_m))) +svint8_t svasr_s8_m(svbool_t, svint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasr_s32_m))) +svint32_t svasr_s32_m(svbool_t, svint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasr_s64_m))) +svint64_t svasr_s64_m(svbool_t, svint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasr_s16_m))) +svint16_t svasr_s16_m(svbool_t, svint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasr_s8_x))) +svint8_t svasr_s8_x(svbool_t, svint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasr_s32_x))) +svint32_t svasr_s32_x(svbool_t, svint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasr_s64_x))) +svint64_t svasr_s64_x(svbool_t, svint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasr_s16_x))) +svint16_t svasr_s16_x(svbool_t, svint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasr_s8_z))) +svint8_t svasr_s8_z(svbool_t, svint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasr_s32_z))) +svint32_t svasr_s32_z(svbool_t, svint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasr_s64_z))) +svint64_t svasr_s64_z(svbool_t, svint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasr_s16_z))) +svint16_t svasr_s16_z(svbool_t, svint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasr_wide_n_s8_m))) +svint8_t svasr_wide_n_s8_m(svbool_t, svint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasr_wide_n_s32_m))) +svint32_t svasr_wide_n_s32_m(svbool_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasr_wide_n_s16_m))) +svint16_t svasr_wide_n_s16_m(svbool_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasr_wide_n_s8_x))) +svint8_t svasr_wide_n_s8_x(svbool_t, svint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasr_wide_n_s32_x))) +svint32_t svasr_wide_n_s32_x(svbool_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasr_wide_n_s16_x))) +svint16_t svasr_wide_n_s16_x(svbool_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasr_wide_n_s8_z))) +svint8_t svasr_wide_n_s8_z(svbool_t, svint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasr_wide_n_s32_z))) +svint32_t svasr_wide_n_s32_z(svbool_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasr_wide_n_s16_z))) +svint16_t svasr_wide_n_s16_z(svbool_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasr_wide_s8_m))) +svint8_t svasr_wide_s8_m(svbool_t, svint8_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasr_wide_s32_m))) +svint32_t svasr_wide_s32_m(svbool_t, svint32_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasr_wide_s16_m))) +svint16_t svasr_wide_s16_m(svbool_t, svint16_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasr_wide_s8_x))) +svint8_t svasr_wide_s8_x(svbool_t, svint8_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasr_wide_s32_x))) +svint32_t svasr_wide_s32_x(svbool_t, svint32_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasr_wide_s16_x))) +svint16_t svasr_wide_s16_x(svbool_t, svint16_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasr_wide_s8_z))) +svint8_t svasr_wide_s8_z(svbool_t, svint8_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasr_wide_s32_z))) +svint32_t svasr_wide_s32_z(svbool_t, svint32_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasr_wide_s16_z))) +svint16_t svasr_wide_s16_z(svbool_t, svint16_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasrd_n_s8_m))) +svint8_t svasrd_n_s8_m(svbool_t, svint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasrd_n_s32_m))) +svint32_t svasrd_n_s32_m(svbool_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasrd_n_s64_m))) +svint64_t svasrd_n_s64_m(svbool_t, svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasrd_n_s16_m))) +svint16_t svasrd_n_s16_m(svbool_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasrd_n_s8_x))) +svint8_t svasrd_n_s8_x(svbool_t, svint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasrd_n_s32_x))) +svint32_t svasrd_n_s32_x(svbool_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasrd_n_s64_x))) +svint64_t svasrd_n_s64_x(svbool_t, svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasrd_n_s16_x))) +svint16_t svasrd_n_s16_x(svbool_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasrd_n_s8_z))) +svint8_t svasrd_n_s8_z(svbool_t, svint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasrd_n_s32_z))) +svint32_t svasrd_n_s32_z(svbool_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasrd_n_s64_z))) +svint64_t svasrd_n_s64_z(svbool_t, svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svasrd_n_s16_z))) +svint16_t svasrd_n_s16_z(svbool_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_b_z))) +svbool_t svbic_b_z(svbool_t, svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_n_u8_m))) +svuint8_t svbic_n_u8_m(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_n_u32_m))) +svuint32_t svbic_n_u32_m(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_n_u64_m))) +svuint64_t svbic_n_u64_m(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_n_u16_m))) +svuint16_t svbic_n_u16_m(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_n_s8_m))) +svint8_t svbic_n_s8_m(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_n_s32_m))) +svint32_t svbic_n_s32_m(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_n_s64_m))) +svint64_t svbic_n_s64_m(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_n_s16_m))) +svint16_t svbic_n_s16_m(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_n_u8_x))) +svuint8_t svbic_n_u8_x(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_n_u32_x))) +svuint32_t svbic_n_u32_x(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_n_u64_x))) +svuint64_t svbic_n_u64_x(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_n_u16_x))) +svuint16_t svbic_n_u16_x(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_n_s8_x))) +svint8_t svbic_n_s8_x(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_n_s32_x))) +svint32_t svbic_n_s32_x(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_n_s64_x))) +svint64_t svbic_n_s64_x(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_n_s16_x))) +svint16_t svbic_n_s16_x(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_n_u8_z))) +svuint8_t svbic_n_u8_z(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_n_u32_z))) +svuint32_t svbic_n_u32_z(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_n_u64_z))) +svuint64_t svbic_n_u64_z(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_n_u16_z))) +svuint16_t svbic_n_u16_z(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_n_s8_z))) +svint8_t svbic_n_s8_z(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_n_s32_z))) +svint32_t svbic_n_s32_z(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_n_s64_z))) +svint64_t svbic_n_s64_z(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_n_s16_z))) +svint16_t svbic_n_s16_z(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_u8_m))) +svuint8_t svbic_u8_m(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_u32_m))) +svuint32_t svbic_u32_m(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_u64_m))) +svuint64_t svbic_u64_m(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_u16_m))) +svuint16_t svbic_u16_m(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_s8_m))) +svint8_t svbic_s8_m(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_s32_m))) +svint32_t svbic_s32_m(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_s64_m))) +svint64_t svbic_s64_m(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_s16_m))) +svint16_t svbic_s16_m(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_u8_x))) +svuint8_t svbic_u8_x(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_u32_x))) +svuint32_t svbic_u32_x(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_u64_x))) +svuint64_t svbic_u64_x(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_u16_x))) +svuint16_t svbic_u16_x(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_s8_x))) +svint8_t svbic_s8_x(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_s32_x))) +svint32_t svbic_s32_x(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_s64_x))) +svint64_t svbic_s64_x(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_s16_x))) +svint16_t svbic_s16_x(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_u8_z))) +svuint8_t svbic_u8_z(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_u32_z))) +svuint32_t svbic_u32_z(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_u64_z))) +svuint64_t svbic_u64_z(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_u16_z))) +svuint16_t svbic_u16_z(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_s8_z))) +svint8_t svbic_s8_z(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_s32_z))) +svint32_t svbic_s32_z(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_s64_z))) +svint64_t svbic_s64_z(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbic_s16_z))) +svint16_t svbic_s16_z(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbrka_b_m))) +svbool_t svbrka_b_m(svbool_t, svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbrka_b_z))) +svbool_t svbrka_b_z(svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbrkb_b_m))) +svbool_t svbrkb_b_m(svbool_t, svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbrkb_b_z))) +svbool_t svbrkb_b_z(svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbrkn_b_z))) +svbool_t svbrkn_b_z(svbool_t, svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbrkpa_b_z))) +svbool_t svbrkpa_b_z(svbool_t, svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbrkpb_b_z))) +svbool_t svbrkpb_b_z(svbool_t, svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcadd_f64_m))) +svfloat64_t svcadd_f64_m(svbool_t, svfloat64_t, svfloat64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcadd_f32_m))) +svfloat32_t svcadd_f32_m(svbool_t, svfloat32_t, svfloat32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcadd_f16_m))) +svfloat16_t svcadd_f16_m(svbool_t, svfloat16_t, svfloat16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcadd_f64_x))) +svfloat64_t svcadd_f64_x(svbool_t, svfloat64_t, svfloat64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcadd_f32_x))) +svfloat32_t svcadd_f32_x(svbool_t, svfloat32_t, svfloat32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcadd_f16_x))) +svfloat16_t svcadd_f16_x(svbool_t, svfloat16_t, svfloat16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcadd_f64_z))) +svfloat64_t svcadd_f64_z(svbool_t, svfloat64_t, svfloat64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcadd_f32_z))) +svfloat32_t svcadd_f32_z(svbool_t, svfloat32_t, svfloat32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcadd_f16_z))) +svfloat16_t svcadd_f16_z(svbool_t, svfloat16_t, svfloat16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclasta_n_u8))) +uint8_t svclasta_n_u8(svbool_t, uint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclasta_n_u32))) +uint32_t svclasta_n_u32(svbool_t, uint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclasta_n_u64))) +uint64_t svclasta_n_u64(svbool_t, uint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclasta_n_u16))) +uint16_t svclasta_n_u16(svbool_t, uint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclasta_n_s8))) +int8_t svclasta_n_s8(svbool_t, int8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclasta_n_f64))) +float64_t svclasta_n_f64(svbool_t, float64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclasta_n_f32))) +float32_t svclasta_n_f32(svbool_t, float32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclasta_n_f16))) +float16_t svclasta_n_f16(svbool_t, float16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclasta_n_s32))) +int32_t svclasta_n_s32(svbool_t, int32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclasta_n_s64))) +int64_t svclasta_n_s64(svbool_t, int64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclasta_n_s16))) +int16_t svclasta_n_s16(svbool_t, int16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclasta_u8))) +svuint8_t svclasta_u8(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclasta_u32))) +svuint32_t svclasta_u32(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclasta_u64))) +svuint64_t svclasta_u64(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclasta_u16))) +svuint16_t svclasta_u16(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclasta_s8))) +svint8_t svclasta_s8(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclasta_f64))) +svfloat64_t svclasta_f64(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclasta_f32))) +svfloat32_t svclasta_f32(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclasta_f16))) +svfloat16_t svclasta_f16(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclasta_s32))) +svint32_t svclasta_s32(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclasta_s64))) +svint64_t svclasta_s64(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclasta_s16))) +svint16_t svclasta_s16(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclastb_n_u8))) +uint8_t svclastb_n_u8(svbool_t, uint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclastb_n_u32))) +uint32_t svclastb_n_u32(svbool_t, uint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclastb_n_u64))) +uint64_t svclastb_n_u64(svbool_t, uint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclastb_n_u16))) +uint16_t svclastb_n_u16(svbool_t, uint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclastb_n_s8))) +int8_t svclastb_n_s8(svbool_t, int8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclastb_n_f64))) +float64_t svclastb_n_f64(svbool_t, float64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclastb_n_f32))) +float32_t svclastb_n_f32(svbool_t, float32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclastb_n_f16))) +float16_t svclastb_n_f16(svbool_t, float16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclastb_n_s32))) +int32_t svclastb_n_s32(svbool_t, int32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclastb_n_s64))) +int64_t svclastb_n_s64(svbool_t, int64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclastb_n_s16))) +int16_t svclastb_n_s16(svbool_t, int16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclastb_u8))) +svuint8_t svclastb_u8(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclastb_u32))) +svuint32_t svclastb_u32(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclastb_u64))) +svuint64_t svclastb_u64(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclastb_u16))) +svuint16_t svclastb_u16(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclastb_s8))) +svint8_t svclastb_s8(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclastb_f64))) +svfloat64_t svclastb_f64(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclastb_f32))) +svfloat32_t svclastb_f32(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclastb_f16))) +svfloat16_t svclastb_f16(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclastb_s32))) +svint32_t svclastb_s32(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclastb_s64))) +svint64_t svclastb_s64(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclastb_s16))) +svint16_t svclastb_s16(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcls_s8_m))) +svuint8_t svcls_s8_m(svuint8_t, svbool_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcls_s32_m))) +svuint32_t svcls_s32_m(svuint32_t, svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcls_s64_m))) +svuint64_t svcls_s64_m(svuint64_t, svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcls_s16_m))) +svuint16_t svcls_s16_m(svuint16_t, svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcls_s8_x))) +svuint8_t svcls_s8_x(svbool_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcls_s32_x))) +svuint32_t svcls_s32_x(svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcls_s64_x))) +svuint64_t svcls_s64_x(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcls_s16_x))) +svuint16_t svcls_s16_x(svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcls_s8_z))) +svuint8_t svcls_s8_z(svbool_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcls_s32_z))) +svuint32_t svcls_s32_z(svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcls_s64_z))) +svuint64_t svcls_s64_z(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcls_s16_z))) +svuint16_t svcls_s16_z(svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclz_u8_m))) +svuint8_t svclz_u8_m(svuint8_t, svbool_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclz_u32_m))) +svuint32_t svclz_u32_m(svuint32_t, svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclz_u64_m))) +svuint64_t svclz_u64_m(svuint64_t, svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclz_u16_m))) +svuint16_t svclz_u16_m(svuint16_t, svbool_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclz_s8_m))) +svuint8_t svclz_s8_m(svuint8_t, svbool_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclz_s32_m))) +svuint32_t svclz_s32_m(svuint32_t, svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclz_s64_m))) +svuint64_t svclz_s64_m(svuint64_t, svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclz_s16_m))) +svuint16_t svclz_s16_m(svuint16_t, svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclz_u8_x))) +svuint8_t svclz_u8_x(svbool_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclz_u32_x))) +svuint32_t svclz_u32_x(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclz_u64_x))) +svuint64_t svclz_u64_x(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclz_u16_x))) +svuint16_t svclz_u16_x(svbool_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclz_s8_x))) +svuint8_t svclz_s8_x(svbool_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclz_s32_x))) +svuint32_t svclz_s32_x(svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclz_s64_x))) +svuint64_t svclz_s64_x(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclz_s16_x))) +svuint16_t svclz_s16_x(svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclz_u8_z))) +svuint8_t svclz_u8_z(svbool_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclz_u32_z))) +svuint32_t svclz_u32_z(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclz_u64_z))) +svuint64_t svclz_u64_z(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclz_u16_z))) +svuint16_t svclz_u16_z(svbool_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclz_s8_z))) +svuint8_t svclz_s8_z(svbool_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclz_s32_z))) +svuint32_t svclz_s32_z(svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclz_s64_z))) +svuint64_t svclz_s64_z(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclz_s16_z))) +svuint16_t svclz_s16_z(svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmla_f64_m))) +svfloat64_t svcmla_f64_m(svbool_t, svfloat64_t, svfloat64_t, svfloat64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmla_f32_m))) +svfloat32_t svcmla_f32_m(svbool_t, svfloat32_t, svfloat32_t, svfloat32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmla_f16_m))) +svfloat16_t svcmla_f16_m(svbool_t, svfloat16_t, svfloat16_t, svfloat16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmla_f64_x))) +svfloat64_t svcmla_f64_x(svbool_t, svfloat64_t, svfloat64_t, svfloat64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmla_f32_x))) +svfloat32_t svcmla_f32_x(svbool_t, svfloat32_t, svfloat32_t, svfloat32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmla_f16_x))) +svfloat16_t svcmla_f16_x(svbool_t, svfloat16_t, svfloat16_t, svfloat16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmla_f64_z))) +svfloat64_t svcmla_f64_z(svbool_t, svfloat64_t, svfloat64_t, svfloat64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmla_f32_z))) +svfloat32_t svcmla_f32_z(svbool_t, svfloat32_t, svfloat32_t, svfloat32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmla_f16_z))) +svfloat16_t svcmla_f16_z(svbool_t, svfloat16_t, svfloat16_t, svfloat16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmla_lane_f32))) +svfloat32_t svcmla_lane_f32(svfloat32_t, svfloat32_t, svfloat32_t, uint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmla_lane_f16))) +svfloat16_t svcmla_lane_f16(svfloat16_t, svfloat16_t, svfloat16_t, uint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpeq_n_f64))) +svbool_t svcmpeq_n_f64(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpeq_n_f32))) +svbool_t svcmpeq_n_f32(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpeq_n_f16))) +svbool_t svcmpeq_n_f16(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpeq_n_u8))) +svbool_t svcmpeq_n_u8(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpeq_n_u32))) +svbool_t svcmpeq_n_u32(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpeq_n_u64))) +svbool_t svcmpeq_n_u64(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpeq_n_u16))) +svbool_t svcmpeq_n_u16(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpeq_n_s8))) +svbool_t svcmpeq_n_s8(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpeq_n_s32))) +svbool_t svcmpeq_n_s32(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpeq_n_s64))) +svbool_t svcmpeq_n_s64(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpeq_n_s16))) +svbool_t svcmpeq_n_s16(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpeq_u8))) +svbool_t svcmpeq_u8(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpeq_u32))) +svbool_t svcmpeq_u32(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpeq_u64))) +svbool_t svcmpeq_u64(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpeq_u16))) +svbool_t svcmpeq_u16(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpeq_s8))) +svbool_t svcmpeq_s8(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpeq_s32))) +svbool_t svcmpeq_s32(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpeq_s64))) +svbool_t svcmpeq_s64(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpeq_s16))) +svbool_t svcmpeq_s16(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpeq_f64))) +svbool_t svcmpeq_f64(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpeq_f32))) +svbool_t svcmpeq_f32(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpeq_f16))) +svbool_t svcmpeq_f16(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpeq_wide_n_s8))) +svbool_t svcmpeq_wide_n_s8(svbool_t, svint8_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpeq_wide_n_s32))) +svbool_t svcmpeq_wide_n_s32(svbool_t, svint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpeq_wide_n_s16))) +svbool_t svcmpeq_wide_n_s16(svbool_t, svint16_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpeq_wide_s8))) +svbool_t svcmpeq_wide_s8(svbool_t, svint8_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpeq_wide_s32))) +svbool_t svcmpeq_wide_s32(svbool_t, svint32_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpeq_wide_s16))) +svbool_t svcmpeq_wide_s16(svbool_t, svint16_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpge_n_f64))) +svbool_t svcmpge_n_f64(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpge_n_f32))) +svbool_t svcmpge_n_f32(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpge_n_f16))) +svbool_t svcmpge_n_f16(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpge_n_s8))) +svbool_t svcmpge_n_s8(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpge_n_s32))) +svbool_t svcmpge_n_s32(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpge_n_s64))) +svbool_t svcmpge_n_s64(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpge_n_s16))) +svbool_t svcmpge_n_s16(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpge_n_u8))) +svbool_t svcmpge_n_u8(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpge_n_u32))) +svbool_t svcmpge_n_u32(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpge_n_u64))) +svbool_t svcmpge_n_u64(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpge_n_u16))) +svbool_t svcmpge_n_u16(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpge_s8))) +svbool_t svcmpge_s8(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpge_s32))) +svbool_t svcmpge_s32(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpge_s64))) +svbool_t svcmpge_s64(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpge_s16))) +svbool_t svcmpge_s16(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpge_f64))) +svbool_t svcmpge_f64(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpge_f32))) +svbool_t svcmpge_f32(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpge_f16))) +svbool_t svcmpge_f16(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpge_u8))) +svbool_t svcmpge_u8(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpge_u32))) +svbool_t svcmpge_u32(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpge_u64))) +svbool_t svcmpge_u64(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpge_u16))) +svbool_t svcmpge_u16(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpge_wide_n_s8))) +svbool_t svcmpge_wide_n_s8(svbool_t, svint8_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpge_wide_n_s32))) +svbool_t svcmpge_wide_n_s32(svbool_t, svint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpge_wide_n_s16))) +svbool_t svcmpge_wide_n_s16(svbool_t, svint16_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpge_wide_n_u8))) +svbool_t svcmpge_wide_n_u8(svbool_t, svuint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpge_wide_n_u32))) +svbool_t svcmpge_wide_n_u32(svbool_t, svuint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpge_wide_n_u16))) +svbool_t svcmpge_wide_n_u16(svbool_t, svuint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpge_wide_s8))) +svbool_t svcmpge_wide_s8(svbool_t, svint8_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpge_wide_s32))) +svbool_t svcmpge_wide_s32(svbool_t, svint32_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpge_wide_s16))) +svbool_t svcmpge_wide_s16(svbool_t, svint16_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpge_wide_u8))) +svbool_t svcmpge_wide_u8(svbool_t, svuint8_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpge_wide_u32))) +svbool_t svcmpge_wide_u32(svbool_t, svuint32_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpge_wide_u16))) +svbool_t svcmpge_wide_u16(svbool_t, svuint16_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpgt_n_f64))) +svbool_t svcmpgt_n_f64(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpgt_n_f32))) +svbool_t svcmpgt_n_f32(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpgt_n_f16))) +svbool_t svcmpgt_n_f16(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpgt_n_s8))) +svbool_t svcmpgt_n_s8(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpgt_n_s32))) +svbool_t svcmpgt_n_s32(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpgt_n_s64))) +svbool_t svcmpgt_n_s64(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpgt_n_s16))) +svbool_t svcmpgt_n_s16(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpgt_n_u8))) +svbool_t svcmpgt_n_u8(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpgt_n_u32))) +svbool_t svcmpgt_n_u32(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpgt_n_u64))) +svbool_t svcmpgt_n_u64(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpgt_n_u16))) +svbool_t svcmpgt_n_u16(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpgt_s8))) +svbool_t svcmpgt_s8(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpgt_s32))) +svbool_t svcmpgt_s32(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpgt_s64))) +svbool_t svcmpgt_s64(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpgt_s16))) +svbool_t svcmpgt_s16(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpgt_f64))) +svbool_t svcmpgt_f64(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpgt_f32))) +svbool_t svcmpgt_f32(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpgt_f16))) +svbool_t svcmpgt_f16(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpgt_u8))) +svbool_t svcmpgt_u8(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpgt_u32))) +svbool_t svcmpgt_u32(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpgt_u64))) +svbool_t svcmpgt_u64(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpgt_u16))) +svbool_t svcmpgt_u16(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpgt_wide_n_s8))) +svbool_t svcmpgt_wide_n_s8(svbool_t, svint8_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpgt_wide_n_s32))) +svbool_t svcmpgt_wide_n_s32(svbool_t, svint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpgt_wide_n_s16))) +svbool_t svcmpgt_wide_n_s16(svbool_t, svint16_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpgt_wide_n_u8))) +svbool_t svcmpgt_wide_n_u8(svbool_t, svuint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpgt_wide_n_u32))) +svbool_t svcmpgt_wide_n_u32(svbool_t, svuint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpgt_wide_n_u16))) +svbool_t svcmpgt_wide_n_u16(svbool_t, svuint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpgt_wide_s8))) +svbool_t svcmpgt_wide_s8(svbool_t, svint8_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpgt_wide_s32))) +svbool_t svcmpgt_wide_s32(svbool_t, svint32_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpgt_wide_s16))) +svbool_t svcmpgt_wide_s16(svbool_t, svint16_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpgt_wide_u8))) +svbool_t svcmpgt_wide_u8(svbool_t, svuint8_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpgt_wide_u32))) +svbool_t svcmpgt_wide_u32(svbool_t, svuint32_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpgt_wide_u16))) +svbool_t svcmpgt_wide_u16(svbool_t, svuint16_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmple_n_f64))) +svbool_t svcmple_n_f64(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmple_n_f32))) +svbool_t svcmple_n_f32(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmple_n_f16))) +svbool_t svcmple_n_f16(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmple_n_s8))) +svbool_t svcmple_n_s8(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmple_n_s32))) +svbool_t svcmple_n_s32(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmple_n_s64))) +svbool_t svcmple_n_s64(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmple_n_s16))) +svbool_t svcmple_n_s16(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmple_n_u8))) +svbool_t svcmple_n_u8(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmple_n_u32))) +svbool_t svcmple_n_u32(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmple_n_u64))) +svbool_t svcmple_n_u64(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmple_n_u16))) +svbool_t svcmple_n_u16(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmple_s8))) +svbool_t svcmple_s8(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmple_s32))) +svbool_t svcmple_s32(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmple_s64))) +svbool_t svcmple_s64(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmple_s16))) +svbool_t svcmple_s16(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmple_f64))) +svbool_t svcmple_f64(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmple_f32))) +svbool_t svcmple_f32(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmple_f16))) +svbool_t svcmple_f16(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmple_u8))) +svbool_t svcmple_u8(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmple_u32))) +svbool_t svcmple_u32(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmple_u64))) +svbool_t svcmple_u64(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmple_u16))) +svbool_t svcmple_u16(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmple_wide_n_s8))) +svbool_t svcmple_wide_n_s8(svbool_t, svint8_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmple_wide_n_s32))) +svbool_t svcmple_wide_n_s32(svbool_t, svint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmple_wide_n_s16))) +svbool_t svcmple_wide_n_s16(svbool_t, svint16_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmple_wide_n_u8))) +svbool_t svcmple_wide_n_u8(svbool_t, svuint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmple_wide_n_u32))) +svbool_t svcmple_wide_n_u32(svbool_t, svuint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmple_wide_n_u16))) +svbool_t svcmple_wide_n_u16(svbool_t, svuint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmple_wide_s8))) +svbool_t svcmple_wide_s8(svbool_t, svint8_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmple_wide_s32))) +svbool_t svcmple_wide_s32(svbool_t, svint32_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmple_wide_s16))) +svbool_t svcmple_wide_s16(svbool_t, svint16_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmple_wide_u8))) +svbool_t svcmple_wide_u8(svbool_t, svuint8_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmple_wide_u32))) +svbool_t svcmple_wide_u32(svbool_t, svuint32_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmple_wide_u16))) +svbool_t svcmple_wide_u16(svbool_t, svuint16_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmplt_n_u8))) +svbool_t svcmplt_n_u8(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmplt_n_u32))) +svbool_t svcmplt_n_u32(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmplt_n_u64))) +svbool_t svcmplt_n_u64(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmplt_n_u16))) +svbool_t svcmplt_n_u16(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmplt_n_f64))) +svbool_t svcmplt_n_f64(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmplt_n_f32))) +svbool_t svcmplt_n_f32(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmplt_n_f16))) +svbool_t svcmplt_n_f16(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmplt_n_s8))) +svbool_t svcmplt_n_s8(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmplt_n_s32))) +svbool_t svcmplt_n_s32(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmplt_n_s64))) +svbool_t svcmplt_n_s64(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmplt_n_s16))) +svbool_t svcmplt_n_s16(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmplt_u8))) +svbool_t svcmplt_u8(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmplt_u32))) +svbool_t svcmplt_u32(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmplt_u64))) +svbool_t svcmplt_u64(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmplt_u16))) +svbool_t svcmplt_u16(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmplt_s8))) +svbool_t svcmplt_s8(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmplt_s32))) +svbool_t svcmplt_s32(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmplt_s64))) +svbool_t svcmplt_s64(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmplt_s16))) +svbool_t svcmplt_s16(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmplt_f64))) +svbool_t svcmplt_f64(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmplt_f32))) +svbool_t svcmplt_f32(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmplt_f16))) +svbool_t svcmplt_f16(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmplt_wide_n_u8))) +svbool_t svcmplt_wide_n_u8(svbool_t, svuint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmplt_wide_n_u32))) +svbool_t svcmplt_wide_n_u32(svbool_t, svuint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmplt_wide_n_u16))) +svbool_t svcmplt_wide_n_u16(svbool_t, svuint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmplt_wide_n_s8))) +svbool_t svcmplt_wide_n_s8(svbool_t, svint8_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmplt_wide_n_s32))) +svbool_t svcmplt_wide_n_s32(svbool_t, svint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmplt_wide_n_s16))) +svbool_t svcmplt_wide_n_s16(svbool_t, svint16_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmplt_wide_u8))) +svbool_t svcmplt_wide_u8(svbool_t, svuint8_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmplt_wide_u32))) +svbool_t svcmplt_wide_u32(svbool_t, svuint32_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmplt_wide_u16))) +svbool_t svcmplt_wide_u16(svbool_t, svuint16_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmplt_wide_s8))) +svbool_t svcmplt_wide_s8(svbool_t, svint8_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmplt_wide_s32))) +svbool_t svcmplt_wide_s32(svbool_t, svint32_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmplt_wide_s16))) +svbool_t svcmplt_wide_s16(svbool_t, svint16_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpne_n_f64))) +svbool_t svcmpne_n_f64(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpne_n_f32))) +svbool_t svcmpne_n_f32(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpne_n_f16))) +svbool_t svcmpne_n_f16(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpne_n_u8))) +svbool_t svcmpne_n_u8(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpne_n_u32))) +svbool_t svcmpne_n_u32(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpne_n_u64))) +svbool_t svcmpne_n_u64(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpne_n_u16))) +svbool_t svcmpne_n_u16(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpne_n_s8))) +svbool_t svcmpne_n_s8(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpne_n_s32))) +svbool_t svcmpne_n_s32(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpne_n_s64))) +svbool_t svcmpne_n_s64(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpne_n_s16))) +svbool_t svcmpne_n_s16(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpne_u8))) +svbool_t svcmpne_u8(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpne_u32))) +svbool_t svcmpne_u32(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpne_u64))) +svbool_t svcmpne_u64(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpne_u16))) +svbool_t svcmpne_u16(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpne_s8))) +svbool_t svcmpne_s8(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpne_s32))) +svbool_t svcmpne_s32(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpne_s64))) +svbool_t svcmpne_s64(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpne_s16))) +svbool_t svcmpne_s16(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpne_f64))) +svbool_t svcmpne_f64(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpne_f32))) +svbool_t svcmpne_f32(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpne_f16))) +svbool_t svcmpne_f16(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpne_wide_n_s8))) +svbool_t svcmpne_wide_n_s8(svbool_t, svint8_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpne_wide_n_s32))) +svbool_t svcmpne_wide_n_s32(svbool_t, svint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpne_wide_n_s16))) +svbool_t svcmpne_wide_n_s16(svbool_t, svint16_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpne_wide_s8))) +svbool_t svcmpne_wide_s8(svbool_t, svint8_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpne_wide_s32))) +svbool_t svcmpne_wide_s32(svbool_t, svint32_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpne_wide_s16))) +svbool_t svcmpne_wide_s16(svbool_t, svint16_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpuo_n_f64))) +svbool_t svcmpuo_n_f64(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpuo_n_f32))) +svbool_t svcmpuo_n_f32(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpuo_n_f16))) +svbool_t svcmpuo_n_f16(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpuo_f64))) +svbool_t svcmpuo_f64(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpuo_f32))) +svbool_t svcmpuo_f32(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmpuo_f16))) +svbool_t svcmpuo_f16(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnot_u8_m))) +svuint8_t svcnot_u8_m(svuint8_t, svbool_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnot_u32_m))) +svuint32_t svcnot_u32_m(svuint32_t, svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnot_u64_m))) +svuint64_t svcnot_u64_m(svuint64_t, svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnot_u16_m))) +svuint16_t svcnot_u16_m(svuint16_t, svbool_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnot_s8_m))) +svint8_t svcnot_s8_m(svint8_t, svbool_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnot_s32_m))) +svint32_t svcnot_s32_m(svint32_t, svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnot_s64_m))) +svint64_t svcnot_s64_m(svint64_t, svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnot_s16_m))) +svint16_t svcnot_s16_m(svint16_t, svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnot_u8_x))) +svuint8_t svcnot_u8_x(svbool_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnot_u32_x))) +svuint32_t svcnot_u32_x(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnot_u64_x))) +svuint64_t svcnot_u64_x(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnot_u16_x))) +svuint16_t svcnot_u16_x(svbool_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnot_s8_x))) +svint8_t svcnot_s8_x(svbool_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnot_s32_x))) +svint32_t svcnot_s32_x(svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnot_s64_x))) +svint64_t svcnot_s64_x(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnot_s16_x))) +svint16_t svcnot_s16_x(svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnot_u8_z))) +svuint8_t svcnot_u8_z(svbool_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnot_u32_z))) +svuint32_t svcnot_u32_z(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnot_u64_z))) +svuint64_t svcnot_u64_z(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnot_u16_z))) +svuint16_t svcnot_u16_z(svbool_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnot_s8_z))) +svint8_t svcnot_s8_z(svbool_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnot_s32_z))) +svint32_t svcnot_s32_z(svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnot_s64_z))) +svint64_t svcnot_s64_z(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnot_s16_z))) +svint16_t svcnot_s16_z(svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnt_u8_m))) +svuint8_t svcnt_u8_m(svuint8_t, svbool_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnt_u32_m))) +svuint32_t svcnt_u32_m(svuint32_t, svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnt_u64_m))) +svuint64_t svcnt_u64_m(svuint64_t, svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnt_u16_m))) +svuint16_t svcnt_u16_m(svuint16_t, svbool_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnt_s8_m))) +svuint8_t svcnt_s8_m(svuint8_t, svbool_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnt_f64_m))) +svuint64_t svcnt_f64_m(svuint64_t, svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnt_f32_m))) +svuint32_t svcnt_f32_m(svuint32_t, svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnt_f16_m))) +svuint16_t svcnt_f16_m(svuint16_t, svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnt_s32_m))) +svuint32_t svcnt_s32_m(svuint32_t, svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnt_s64_m))) +svuint64_t svcnt_s64_m(svuint64_t, svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnt_s16_m))) +svuint16_t svcnt_s16_m(svuint16_t, svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnt_u8_x))) +svuint8_t svcnt_u8_x(svbool_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnt_u32_x))) +svuint32_t svcnt_u32_x(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnt_u64_x))) +svuint64_t svcnt_u64_x(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnt_u16_x))) +svuint16_t svcnt_u16_x(svbool_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnt_s8_x))) +svuint8_t svcnt_s8_x(svbool_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnt_f64_x))) +svuint64_t svcnt_f64_x(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnt_f32_x))) +svuint32_t svcnt_f32_x(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnt_f16_x))) +svuint16_t svcnt_f16_x(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnt_s32_x))) +svuint32_t svcnt_s32_x(svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnt_s64_x))) +svuint64_t svcnt_s64_x(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnt_s16_x))) +svuint16_t svcnt_s16_x(svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnt_u8_z))) +svuint8_t svcnt_u8_z(svbool_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnt_u32_z))) +svuint32_t svcnt_u32_z(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnt_u64_z))) +svuint64_t svcnt_u64_z(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnt_u16_z))) +svuint16_t svcnt_u16_z(svbool_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnt_s8_z))) +svuint8_t svcnt_s8_z(svbool_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnt_f64_z))) +svuint64_t svcnt_f64_z(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnt_f32_z))) +svuint32_t svcnt_f32_z(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnt_f16_z))) +svuint16_t svcnt_f16_z(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnt_s32_z))) +svuint32_t svcnt_s32_z(svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnt_s64_z))) +svuint64_t svcnt_s64_z(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnt_s16_z))) +svuint16_t svcnt_s16_z(svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcntb))) +uint64_t svcntb(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcntb_pat))) +uint64_t svcntb_pat(enum svpattern); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcntd))) +uint64_t svcntd(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcntd_pat))) +uint64_t svcntd_pat(enum svpattern); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnth))) +uint64_t svcnth(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnth_pat))) +uint64_t svcnth_pat(enum svpattern); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcntp_b8))) +uint64_t svcntp_b8(svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcntp_b32))) +uint64_t svcntp_b32(svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcntp_b64))) +uint64_t svcntp_b64(svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcntp_b16))) +uint64_t svcntp_b16(svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcntw))) +uint64_t svcntw(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcntw_pat))) +uint64_t svcntw_pat(enum svpattern); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcompact_u32))) +svuint32_t svcompact_u32(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcompact_u64))) +svuint64_t svcompact_u64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcompact_f64))) +svfloat64_t svcompact_f64(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcompact_f32))) +svfloat32_t svcompact_f32(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcompact_s32))) +svint32_t svcompact_s32(svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcompact_s64))) +svint64_t svcompact_s64(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcreate2_u8))) +svuint8x2_t svcreate2_u8(svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcreate2_u32))) +svuint32x2_t svcreate2_u32(svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcreate2_u64))) +svuint64x2_t svcreate2_u64(svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcreate2_u16))) +svuint16x2_t svcreate2_u16(svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcreate2_s8))) +svint8x2_t svcreate2_s8(svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcreate2_f64))) +svfloat64x2_t svcreate2_f64(svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcreate2_f32))) +svfloat32x2_t svcreate2_f32(svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcreate2_f16))) +svfloat16x2_t svcreate2_f16(svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcreate2_s32))) +svint32x2_t svcreate2_s32(svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcreate2_s64))) +svint64x2_t svcreate2_s64(svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcreate2_s16))) +svint16x2_t svcreate2_s16(svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcreate3_u8))) +svuint8x3_t svcreate3_u8(svuint8_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcreate3_u32))) +svuint32x3_t svcreate3_u32(svuint32_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcreate3_u64))) +svuint64x3_t svcreate3_u64(svuint64_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcreate3_u16))) +svuint16x3_t svcreate3_u16(svuint16_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcreate3_s8))) +svint8x3_t svcreate3_s8(svint8_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcreate3_f64))) +svfloat64x3_t svcreate3_f64(svfloat64_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcreate3_f32))) +svfloat32x3_t svcreate3_f32(svfloat32_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcreate3_f16))) +svfloat16x3_t svcreate3_f16(svfloat16_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcreate3_s32))) +svint32x3_t svcreate3_s32(svint32_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcreate3_s64))) +svint64x3_t svcreate3_s64(svint64_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcreate3_s16))) +svint16x3_t svcreate3_s16(svint16_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcreate4_u8))) +svuint8x4_t svcreate4_u8(svuint8_t, svuint8_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcreate4_u32))) +svuint32x4_t svcreate4_u32(svuint32_t, svuint32_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcreate4_u64))) +svuint64x4_t svcreate4_u64(svuint64_t, svuint64_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcreate4_u16))) +svuint16x4_t svcreate4_u16(svuint16_t, svuint16_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcreate4_s8))) +svint8x4_t svcreate4_s8(svint8_t, svint8_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcreate4_f64))) +svfloat64x4_t svcreate4_f64(svfloat64_t, svfloat64_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcreate4_f32))) +svfloat32x4_t svcreate4_f32(svfloat32_t, svfloat32_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcreate4_f16))) +svfloat16x4_t svcreate4_f16(svfloat16_t, svfloat16_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcreate4_s32))) +svint32x4_t svcreate4_s32(svint32_t, svint32_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcreate4_s64))) +svint64x4_t svcreate4_s64(svint64_t, svint64_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcreate4_s16))) +svint16x4_t svcreate4_s16(svint16_t, svint16_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f16_f32_m))) +svfloat16_t svcvt_f16_f32_m(svfloat16_t, svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f16_f32_x))) +svfloat16_t svcvt_f16_f32_x(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f16_f32_z))) +svfloat16_t svcvt_f16_f32_z(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f16_f64_m))) +svfloat16_t svcvt_f16_f64_m(svfloat16_t, svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f16_f64_x))) +svfloat16_t svcvt_f16_f64_x(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f16_f64_z))) +svfloat16_t svcvt_f16_f64_z(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f16_s16_m))) +svfloat16_t svcvt_f16_s16_m(svfloat16_t, svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f16_s16_x))) +svfloat16_t svcvt_f16_s16_x(svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f16_s16_z))) +svfloat16_t svcvt_f16_s16_z(svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f16_s32_m))) +svfloat16_t svcvt_f16_s32_m(svfloat16_t, svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f16_s32_x))) +svfloat16_t svcvt_f16_s32_x(svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f16_s32_z))) +svfloat16_t svcvt_f16_s32_z(svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f16_s64_m))) +svfloat16_t svcvt_f16_s64_m(svfloat16_t, svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f16_s64_x))) +svfloat16_t svcvt_f16_s64_x(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f16_s64_z))) +svfloat16_t svcvt_f16_s64_z(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f16_u16_m))) +svfloat16_t svcvt_f16_u16_m(svfloat16_t, svbool_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f16_u16_x))) +svfloat16_t svcvt_f16_u16_x(svbool_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f16_u16_z))) +svfloat16_t svcvt_f16_u16_z(svbool_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f16_u32_m))) +svfloat16_t svcvt_f16_u32_m(svfloat16_t, svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f16_u32_x))) +svfloat16_t svcvt_f16_u32_x(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f16_u32_z))) +svfloat16_t svcvt_f16_u32_z(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f16_u64_m))) +svfloat16_t svcvt_f16_u64_m(svfloat16_t, svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f16_u64_x))) +svfloat16_t svcvt_f16_u64_x(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f16_u64_z))) +svfloat16_t svcvt_f16_u64_z(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f32_f16_m))) +svfloat32_t svcvt_f32_f16_m(svfloat32_t, svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f32_f16_x))) +svfloat32_t svcvt_f32_f16_x(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f32_f16_z))) +svfloat32_t svcvt_f32_f16_z(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f32_f64_m))) +svfloat32_t svcvt_f32_f64_m(svfloat32_t, svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f32_f64_x))) +svfloat32_t svcvt_f32_f64_x(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f32_f64_z))) +svfloat32_t svcvt_f32_f64_z(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f32_s32_m))) +svfloat32_t svcvt_f32_s32_m(svfloat32_t, svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f32_s32_x))) +svfloat32_t svcvt_f32_s32_x(svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f32_s32_z))) +svfloat32_t svcvt_f32_s32_z(svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f32_s64_m))) +svfloat32_t svcvt_f32_s64_m(svfloat32_t, svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f32_s64_x))) +svfloat32_t svcvt_f32_s64_x(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f32_s64_z))) +svfloat32_t svcvt_f32_s64_z(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f32_u32_m))) +svfloat32_t svcvt_f32_u32_m(svfloat32_t, svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f32_u32_x))) +svfloat32_t svcvt_f32_u32_x(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f32_u32_z))) +svfloat32_t svcvt_f32_u32_z(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f32_u64_m))) +svfloat32_t svcvt_f32_u64_m(svfloat32_t, svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f32_u64_x))) +svfloat32_t svcvt_f32_u64_x(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f32_u64_z))) +svfloat32_t svcvt_f32_u64_z(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f64_f16_m))) +svfloat64_t svcvt_f64_f16_m(svfloat64_t, svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f64_f16_x))) +svfloat64_t svcvt_f64_f16_x(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f64_f16_z))) +svfloat64_t svcvt_f64_f16_z(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f64_f32_m))) +svfloat64_t svcvt_f64_f32_m(svfloat64_t, svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f64_f32_x))) +svfloat64_t svcvt_f64_f32_x(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f64_f32_z))) +svfloat64_t svcvt_f64_f32_z(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f64_s32_m))) +svfloat64_t svcvt_f64_s32_m(svfloat64_t, svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f64_s32_x))) +svfloat64_t svcvt_f64_s32_x(svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f64_s32_z))) +svfloat64_t svcvt_f64_s32_z(svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f64_s64_m))) +svfloat64_t svcvt_f64_s64_m(svfloat64_t, svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f64_s64_x))) +svfloat64_t svcvt_f64_s64_x(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f64_s64_z))) +svfloat64_t svcvt_f64_s64_z(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f64_u32_m))) +svfloat64_t svcvt_f64_u32_m(svfloat64_t, svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f64_u32_x))) +svfloat64_t svcvt_f64_u32_x(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f64_u32_z))) +svfloat64_t svcvt_f64_u32_z(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f64_u64_m))) +svfloat64_t svcvt_f64_u64_m(svfloat64_t, svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f64_u64_x))) +svfloat64_t svcvt_f64_u64_x(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_f64_u64_z))) +svfloat64_t svcvt_f64_u64_z(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_s16_f16_m))) +svint16_t svcvt_s16_f16_m(svint16_t, svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_s16_f16_x))) +svint16_t svcvt_s16_f16_x(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_s16_f16_z))) +svint16_t svcvt_s16_f16_z(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_s32_f16_m))) +svint32_t svcvt_s32_f16_m(svint32_t, svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_s32_f16_x))) +svint32_t svcvt_s32_f16_x(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_s32_f16_z))) +svint32_t svcvt_s32_f16_z(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_s32_f32_m))) +svint32_t svcvt_s32_f32_m(svint32_t, svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_s32_f32_x))) +svint32_t svcvt_s32_f32_x(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_s32_f32_z))) +svint32_t svcvt_s32_f32_z(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_s32_f64_m))) +svint32_t svcvt_s32_f64_m(svint32_t, svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_s32_f64_x))) +svint32_t svcvt_s32_f64_x(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_s32_f64_z))) +svint32_t svcvt_s32_f64_z(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_s64_f16_m))) +svint64_t svcvt_s64_f16_m(svint64_t, svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_s64_f16_x))) +svint64_t svcvt_s64_f16_x(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_s64_f16_z))) +svint64_t svcvt_s64_f16_z(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_s64_f32_m))) +svint64_t svcvt_s64_f32_m(svint64_t, svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_s64_f32_x))) +svint64_t svcvt_s64_f32_x(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_s64_f32_z))) +svint64_t svcvt_s64_f32_z(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_s64_f64_m))) +svint64_t svcvt_s64_f64_m(svint64_t, svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_s64_f64_x))) +svint64_t svcvt_s64_f64_x(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_s64_f64_z))) +svint64_t svcvt_s64_f64_z(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_u16_f16_m))) +svuint16_t svcvt_u16_f16_m(svuint16_t, svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_u16_f16_x))) +svuint16_t svcvt_u16_f16_x(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_u16_f16_z))) +svuint16_t svcvt_u16_f16_z(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_u32_f16_m))) +svuint32_t svcvt_u32_f16_m(svuint32_t, svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_u32_f16_x))) +svuint32_t svcvt_u32_f16_x(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_u32_f16_z))) +svuint32_t svcvt_u32_f16_z(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_u32_f32_m))) +svuint32_t svcvt_u32_f32_m(svuint32_t, svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_u32_f32_x))) +svuint32_t svcvt_u32_f32_x(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_u32_f32_z))) +svuint32_t svcvt_u32_f32_z(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_u32_f64_m))) +svuint32_t svcvt_u32_f64_m(svuint32_t, svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_u32_f64_x))) +svuint32_t svcvt_u32_f64_x(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_u32_f64_z))) +svuint32_t svcvt_u32_f64_z(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_u64_f16_m))) +svuint64_t svcvt_u64_f16_m(svuint64_t, svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_u64_f16_x))) +svuint64_t svcvt_u64_f16_x(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_u64_f16_z))) +svuint64_t svcvt_u64_f16_z(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_u64_f32_m))) +svuint64_t svcvt_u64_f32_m(svuint64_t, svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_u64_f32_x))) +svuint64_t svcvt_u64_f32_x(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_u64_f32_z))) +svuint64_t svcvt_u64_f32_z(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_u64_f64_m))) +svuint64_t svcvt_u64_f64_m(svuint64_t, svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_u64_f64_x))) +svuint64_t svcvt_u64_f64_x(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_u64_f64_z))) +svuint64_t svcvt_u64_f64_z(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdiv_n_f64_m))) +svfloat64_t svdiv_n_f64_m(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdiv_n_f32_m))) +svfloat32_t svdiv_n_f32_m(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdiv_n_f16_m))) +svfloat16_t svdiv_n_f16_m(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdiv_n_f64_x))) +svfloat64_t svdiv_n_f64_x(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdiv_n_f32_x))) +svfloat32_t svdiv_n_f32_x(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdiv_n_f16_x))) +svfloat16_t svdiv_n_f16_x(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdiv_n_f64_z))) +svfloat64_t svdiv_n_f64_z(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdiv_n_f32_z))) +svfloat32_t svdiv_n_f32_z(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdiv_n_f16_z))) +svfloat16_t svdiv_n_f16_z(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdiv_n_s32_m))) +svint32_t svdiv_n_s32_m(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdiv_n_s64_m))) +svint64_t svdiv_n_s64_m(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdiv_n_s32_x))) +svint32_t svdiv_n_s32_x(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdiv_n_s64_x))) +svint64_t svdiv_n_s64_x(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdiv_n_s32_z))) +svint32_t svdiv_n_s32_z(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdiv_n_s64_z))) +svint64_t svdiv_n_s64_z(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdiv_n_u32_m))) +svuint32_t svdiv_n_u32_m(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdiv_n_u64_m))) +svuint64_t svdiv_n_u64_m(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdiv_n_u32_x))) +svuint32_t svdiv_n_u32_x(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdiv_n_u64_x))) +svuint64_t svdiv_n_u64_x(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdiv_n_u32_z))) +svuint32_t svdiv_n_u32_z(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdiv_n_u64_z))) +svuint64_t svdiv_n_u64_z(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdiv_f64_m))) +svfloat64_t svdiv_f64_m(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdiv_f32_m))) +svfloat32_t svdiv_f32_m(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdiv_f16_m))) +svfloat16_t svdiv_f16_m(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdiv_f64_x))) +svfloat64_t svdiv_f64_x(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdiv_f32_x))) +svfloat32_t svdiv_f32_x(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdiv_f16_x))) +svfloat16_t svdiv_f16_x(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdiv_f64_z))) +svfloat64_t svdiv_f64_z(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdiv_f32_z))) +svfloat32_t svdiv_f32_z(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdiv_f16_z))) +svfloat16_t svdiv_f16_z(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdiv_s32_m))) +svint32_t svdiv_s32_m(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdiv_s64_m))) +svint64_t svdiv_s64_m(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdiv_s32_x))) +svint32_t svdiv_s32_x(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdiv_s64_x))) +svint64_t svdiv_s64_x(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdiv_s32_z))) +svint32_t svdiv_s32_z(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdiv_s64_z))) +svint64_t svdiv_s64_z(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdiv_u32_m))) +svuint32_t svdiv_u32_m(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdiv_u64_m))) +svuint64_t svdiv_u64_m(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdiv_u32_x))) +svuint32_t svdiv_u32_x(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdiv_u64_x))) +svuint64_t svdiv_u64_x(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdiv_u32_z))) +svuint32_t svdiv_u32_z(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdiv_u64_z))) +svuint64_t svdiv_u64_z(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdivr_n_f64_m))) +svfloat64_t svdivr_n_f64_m(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdivr_n_f32_m))) +svfloat32_t svdivr_n_f32_m(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdivr_n_f16_m))) +svfloat16_t svdivr_n_f16_m(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdivr_n_f64_x))) +svfloat64_t svdivr_n_f64_x(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdivr_n_f32_x))) +svfloat32_t svdivr_n_f32_x(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdivr_n_f16_x))) +svfloat16_t svdivr_n_f16_x(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdivr_n_f64_z))) +svfloat64_t svdivr_n_f64_z(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdivr_n_f32_z))) +svfloat32_t svdivr_n_f32_z(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdivr_n_f16_z))) +svfloat16_t svdivr_n_f16_z(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdivr_n_s32_m))) +svint32_t svdivr_n_s32_m(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdivr_n_s64_m))) +svint64_t svdivr_n_s64_m(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdivr_n_s32_x))) +svint32_t svdivr_n_s32_x(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdivr_n_s64_x))) +svint64_t svdivr_n_s64_x(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdivr_n_s32_z))) +svint32_t svdivr_n_s32_z(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdivr_n_s64_z))) +svint64_t svdivr_n_s64_z(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdivr_n_u32_m))) +svuint32_t svdivr_n_u32_m(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdivr_n_u64_m))) +svuint64_t svdivr_n_u64_m(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdivr_n_u32_x))) +svuint32_t svdivr_n_u32_x(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdivr_n_u64_x))) +svuint64_t svdivr_n_u64_x(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdivr_n_u32_z))) +svuint32_t svdivr_n_u32_z(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdivr_n_u64_z))) +svuint64_t svdivr_n_u64_z(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdivr_f64_m))) +svfloat64_t svdivr_f64_m(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdivr_f32_m))) +svfloat32_t svdivr_f32_m(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdivr_f16_m))) +svfloat16_t svdivr_f16_m(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdivr_f64_x))) +svfloat64_t svdivr_f64_x(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdivr_f32_x))) +svfloat32_t svdivr_f32_x(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdivr_f16_x))) +svfloat16_t svdivr_f16_x(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdivr_f64_z))) +svfloat64_t svdivr_f64_z(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdivr_f32_z))) +svfloat32_t svdivr_f32_z(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdivr_f16_z))) +svfloat16_t svdivr_f16_z(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdivr_s32_m))) +svint32_t svdivr_s32_m(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdivr_s64_m))) +svint64_t svdivr_s64_m(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdivr_s32_x))) +svint32_t svdivr_s32_x(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdivr_s64_x))) +svint64_t svdivr_s64_x(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdivr_s32_z))) +svint32_t svdivr_s32_z(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdivr_s64_z))) +svint64_t svdivr_s64_z(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdivr_u32_m))) +svuint32_t svdivr_u32_m(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdivr_u64_m))) +svuint64_t svdivr_u64_m(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdivr_u32_x))) +svuint32_t svdivr_u32_x(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdivr_u64_x))) +svuint64_t svdivr_u64_x(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdivr_u32_z))) +svuint32_t svdivr_u32_z(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdivr_u64_z))) +svuint64_t svdivr_u64_z(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdot_n_s32))) +svint32_t svdot_n_s32(svint32_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdot_n_s64))) +svint64_t svdot_n_s64(svint64_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdot_n_u32))) +svuint32_t svdot_n_u32(svuint32_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdot_n_u64))) +svuint64_t svdot_n_u64(svuint64_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdot_s32))) +svint32_t svdot_s32(svint32_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdot_s64))) +svint64_t svdot_s64(svint64_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdot_u32))) +svuint32_t svdot_u32(svuint32_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdot_u64))) +svuint64_t svdot_u64(svuint64_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdot_lane_s32))) +svint32_t svdot_lane_s32(svint32_t, svint8_t, svint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdot_lane_s64))) +svint64_t svdot_lane_s64(svint64_t, svint16_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdot_lane_u32))) +svuint32_t svdot_lane_u32(svuint32_t, svuint8_t, svuint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdot_lane_u64))) +svuint64_t svdot_lane_u64(svuint64_t, svuint16_t, svuint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_u8))) +svuint8_t svdup_n_u8(uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_u32))) +svuint32_t svdup_n_u32(uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_u64))) +svuint64_t svdup_n_u64(uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_u16))) +svuint16_t svdup_n_u16(uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_s8))) +svint8_t svdup_n_s8(int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_f64))) +svfloat64_t svdup_n_f64(float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_f32))) +svfloat32_t svdup_n_f32(float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_f16))) +svfloat16_t svdup_n_f16(float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_s32))) +svint32_t svdup_n_s32(int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_s64))) +svint64_t svdup_n_s64(int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_s16))) +svint16_t svdup_n_s16(int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_u8_m))) +svuint8_t svdup_n_u8_m(svuint8_t, svbool_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_u32_m))) +svuint32_t svdup_n_u32_m(svuint32_t, svbool_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_u64_m))) +svuint64_t svdup_n_u64_m(svuint64_t, svbool_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_u16_m))) +svuint16_t svdup_n_u16_m(svuint16_t, svbool_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_s8_m))) +svint8_t svdup_n_s8_m(svint8_t, svbool_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_f64_m))) +svfloat64_t svdup_n_f64_m(svfloat64_t, svbool_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_f32_m))) +svfloat32_t svdup_n_f32_m(svfloat32_t, svbool_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_f16_m))) +svfloat16_t svdup_n_f16_m(svfloat16_t, svbool_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_s32_m))) +svint32_t svdup_n_s32_m(svint32_t, svbool_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_s64_m))) +svint64_t svdup_n_s64_m(svint64_t, svbool_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_s16_m))) +svint16_t svdup_n_s16_m(svint16_t, svbool_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_b8))) +svbool_t svdup_n_b8(bool); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_b32))) +svbool_t svdup_n_b32(bool); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_b64))) +svbool_t svdup_n_b64(bool); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_b16))) +svbool_t svdup_n_b16(bool); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_u8_x))) +svuint8_t svdup_n_u8_x(svbool_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_u32_x))) +svuint32_t svdup_n_u32_x(svbool_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_u64_x))) +svuint64_t svdup_n_u64_x(svbool_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_u16_x))) +svuint16_t svdup_n_u16_x(svbool_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_s8_x))) +svint8_t svdup_n_s8_x(svbool_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_f64_x))) +svfloat64_t svdup_n_f64_x(svbool_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_f32_x))) +svfloat32_t svdup_n_f32_x(svbool_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_f16_x))) +svfloat16_t svdup_n_f16_x(svbool_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_s32_x))) +svint32_t svdup_n_s32_x(svbool_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_s64_x))) +svint64_t svdup_n_s64_x(svbool_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_s16_x))) +svint16_t svdup_n_s16_x(svbool_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_u8_z))) +svuint8_t svdup_n_u8_z(svbool_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_u32_z))) +svuint32_t svdup_n_u32_z(svbool_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_u64_z))) +svuint64_t svdup_n_u64_z(svbool_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_u16_z))) +svuint16_t svdup_n_u16_z(svbool_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_s8_z))) +svint8_t svdup_n_s8_z(svbool_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_f64_z))) +svfloat64_t svdup_n_f64_z(svbool_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_f32_z))) +svfloat32_t svdup_n_f32_z(svbool_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_f16_z))) +svfloat16_t svdup_n_f16_z(svbool_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_s32_z))) +svint32_t svdup_n_s32_z(svbool_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_s64_z))) +svint64_t svdup_n_s64_z(svbool_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_s16_z))) +svint16_t svdup_n_s16_z(svbool_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_lane_u8))) +svuint8_t svdup_lane_u8(svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_lane_u32))) +svuint32_t svdup_lane_u32(svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_lane_u64))) +svuint64_t svdup_lane_u64(svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_lane_u16))) +svuint16_t svdup_lane_u16(svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_lane_s8))) +svint8_t svdup_lane_s8(svint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_lane_f64))) +svfloat64_t svdup_lane_f64(svfloat64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_lane_f32))) +svfloat32_t svdup_lane_f32(svfloat32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_lane_f16))) +svfloat16_t svdup_lane_f16(svfloat16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_lane_s32))) +svint32_t svdup_lane_s32(svint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_lane_s64))) +svint64_t svdup_lane_s64(svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_lane_s16))) +svint16_t svdup_lane_s16(svint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdupq_n_u16))) +svuint16_t svdupq_n_u16(uint16_t, uint16_t, uint16_t, uint16_t, uint16_t, uint16_t, uint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdupq_n_f16))) +svfloat16_t svdupq_n_f16(float16_t, float16_t, float16_t, float16_t, float16_t, float16_t, float16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdupq_n_s16))) +svint16_t svdupq_n_s16(int16_t, int16_t, int16_t, int16_t, int16_t, int16_t, int16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdupq_n_u32))) +svuint32_t svdupq_n_u32(uint32_t, uint32_t, uint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdupq_n_f32))) +svfloat32_t svdupq_n_f32(float32_t, float32_t, float32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdupq_n_s32))) +svint32_t svdupq_n_s32(int32_t, int32_t, int32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdupq_n_u64))) +svuint64_t svdupq_n_u64(uint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdupq_n_f64))) +svfloat64_t svdupq_n_f64(float64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdupq_n_s64))) +svint64_t svdupq_n_s64(int64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdupq_n_u8))) +svuint8_t svdupq_n_u8(uint8_t, uint8_t, uint8_t, uint8_t, uint8_t, uint8_t, uint8_t, uint8_t, uint8_t, uint8_t, uint8_t, uint8_t, uint8_t, uint8_t, uint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdupq_n_s8))) +svint8_t svdupq_n_s8(int8_t, int8_t, int8_t, int8_t, int8_t, int8_t, int8_t, int8_t, int8_t, int8_t, int8_t, int8_t, int8_t, int8_t, int8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdupq_n_b16))) +svbool_t svdupq_n_b16(bool, bool, bool, bool, bool, bool, bool, bool); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdupq_n_b32))) +svbool_t svdupq_n_b32(bool, bool, bool, bool); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdupq_n_b64))) +svbool_t svdupq_n_b64(bool, bool); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdupq_n_b8))) +svbool_t svdupq_n_b8(bool, bool, bool, bool, bool, bool, bool, bool, bool, bool, bool, bool, bool, bool, bool, bool); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdupq_lane_u8))) +svuint8_t svdupq_lane_u8(svuint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdupq_lane_u32))) +svuint32_t svdupq_lane_u32(svuint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdupq_lane_u64))) +svuint64_t svdupq_lane_u64(svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdupq_lane_u16))) +svuint16_t svdupq_lane_u16(svuint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdupq_lane_s8))) +svint8_t svdupq_lane_s8(svint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdupq_lane_f64))) +svfloat64_t svdupq_lane_f64(svfloat64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdupq_lane_f32))) +svfloat32_t svdupq_lane_f32(svfloat32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdupq_lane_f16))) +svfloat16_t svdupq_lane_f16(svfloat16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdupq_lane_s32))) +svint32_t svdupq_lane_s32(svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdupq_lane_s64))) +svint64_t svdupq_lane_s64(svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdupq_lane_s16))) +svint16_t svdupq_lane_s16(svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_b_z))) +svbool_t sveor_b_z(svbool_t, svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_n_u8_m))) +svuint8_t sveor_n_u8_m(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_n_u32_m))) +svuint32_t sveor_n_u32_m(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_n_u64_m))) +svuint64_t sveor_n_u64_m(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_n_u16_m))) +svuint16_t sveor_n_u16_m(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_n_s8_m))) +svint8_t sveor_n_s8_m(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_n_s32_m))) +svint32_t sveor_n_s32_m(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_n_s64_m))) +svint64_t sveor_n_s64_m(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_n_s16_m))) +svint16_t sveor_n_s16_m(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_n_u8_x))) +svuint8_t sveor_n_u8_x(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_n_u32_x))) +svuint32_t sveor_n_u32_x(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_n_u64_x))) +svuint64_t sveor_n_u64_x(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_n_u16_x))) +svuint16_t sveor_n_u16_x(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_n_s8_x))) +svint8_t sveor_n_s8_x(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_n_s32_x))) +svint32_t sveor_n_s32_x(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_n_s64_x))) +svint64_t sveor_n_s64_x(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_n_s16_x))) +svint16_t sveor_n_s16_x(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_n_u8_z))) +svuint8_t sveor_n_u8_z(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_n_u32_z))) +svuint32_t sveor_n_u32_z(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_n_u64_z))) +svuint64_t sveor_n_u64_z(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_n_u16_z))) +svuint16_t sveor_n_u16_z(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_n_s8_z))) +svint8_t sveor_n_s8_z(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_n_s32_z))) +svint32_t sveor_n_s32_z(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_n_s64_z))) +svint64_t sveor_n_s64_z(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_n_s16_z))) +svint16_t sveor_n_s16_z(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_u8_m))) +svuint8_t sveor_u8_m(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_u32_m))) +svuint32_t sveor_u32_m(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_u64_m))) +svuint64_t sveor_u64_m(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_u16_m))) +svuint16_t sveor_u16_m(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_s8_m))) +svint8_t sveor_s8_m(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_s32_m))) +svint32_t sveor_s32_m(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_s64_m))) +svint64_t sveor_s64_m(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_s16_m))) +svint16_t sveor_s16_m(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_u8_x))) +svuint8_t sveor_u8_x(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_u32_x))) +svuint32_t sveor_u32_x(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_u64_x))) +svuint64_t sveor_u64_x(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_u16_x))) +svuint16_t sveor_u16_x(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_s8_x))) +svint8_t sveor_s8_x(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_s32_x))) +svint32_t sveor_s32_x(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_s64_x))) +svint64_t sveor_s64_x(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_s16_x))) +svint16_t sveor_s16_x(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_u8_z))) +svuint8_t sveor_u8_z(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_u32_z))) +svuint32_t sveor_u32_z(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_u64_z))) +svuint64_t sveor_u64_z(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_u16_z))) +svuint16_t sveor_u16_z(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_s8_z))) +svint8_t sveor_s8_z(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_s32_z))) +svint32_t sveor_s32_z(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_s64_z))) +svint64_t sveor_s64_z(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor_s16_z))) +svint16_t sveor_s16_z(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveorv_u8))) +uint8_t sveorv_u8(svbool_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveorv_u32))) +uint32_t sveorv_u32(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveorv_u64))) +uint64_t sveorv_u64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveorv_u16))) +uint16_t sveorv_u16(svbool_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveorv_s8))) +int8_t sveorv_s8(svbool_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveorv_s32))) +int32_t sveorv_s32(svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveorv_s64))) +int64_t sveorv_s64(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveorv_s16))) +int16_t sveorv_s16(svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svexpa_f64))) +svfloat64_t svexpa_f64(svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svexpa_f32))) +svfloat32_t svexpa_f32(svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svexpa_f16))) +svfloat16_t svexpa_f16(svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svext_u8))) +svuint8_t svext_u8(svuint8_t, svuint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svext_u32))) +svuint32_t svext_u32(svuint32_t, svuint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svext_u64))) +svuint64_t svext_u64(svuint64_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svext_u16))) +svuint16_t svext_u16(svuint16_t, svuint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svext_s8))) +svint8_t svext_s8(svint8_t, svint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svext_f64))) +svfloat64_t svext_f64(svfloat64_t, svfloat64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svext_f32))) +svfloat32_t svext_f32(svfloat32_t, svfloat32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svext_f16))) +svfloat16_t svext_f16(svfloat16_t, svfloat16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svext_s32))) +svint32_t svext_s32(svint32_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svext_s64))) +svint64_t svext_s64(svint64_t, svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svext_s16))) +svint16_t svext_s16(svint16_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svextb_s32_m))) +svint32_t svextb_s32_m(svint32_t, svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svextb_s64_m))) +svint64_t svextb_s64_m(svint64_t, svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svextb_s16_m))) +svint16_t svextb_s16_m(svint16_t, svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svextb_s32_x))) +svint32_t svextb_s32_x(svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svextb_s64_x))) +svint64_t svextb_s64_x(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svextb_s16_x))) +svint16_t svextb_s16_x(svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svextb_s32_z))) +svint32_t svextb_s32_z(svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svextb_s64_z))) +svint64_t svextb_s64_z(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svextb_s16_z))) +svint16_t svextb_s16_z(svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svextb_u32_m))) +svuint32_t svextb_u32_m(svuint32_t, svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svextb_u64_m))) +svuint64_t svextb_u64_m(svuint64_t, svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svextb_u16_m))) +svuint16_t svextb_u16_m(svuint16_t, svbool_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svextb_u32_x))) +svuint32_t svextb_u32_x(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svextb_u64_x))) +svuint64_t svextb_u64_x(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svextb_u16_x))) +svuint16_t svextb_u16_x(svbool_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svextb_u32_z))) +svuint32_t svextb_u32_z(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svextb_u64_z))) +svuint64_t svextb_u64_z(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svextb_u16_z))) +svuint16_t svextb_u16_z(svbool_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svexth_s32_m))) +svint32_t svexth_s32_m(svint32_t, svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svexth_s64_m))) +svint64_t svexth_s64_m(svint64_t, svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svexth_s32_x))) +svint32_t svexth_s32_x(svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svexth_s64_x))) +svint64_t svexth_s64_x(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svexth_s32_z))) +svint32_t svexth_s32_z(svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svexth_s64_z))) +svint64_t svexth_s64_z(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svexth_u32_m))) +svuint32_t svexth_u32_m(svuint32_t, svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svexth_u64_m))) +svuint64_t svexth_u64_m(svuint64_t, svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svexth_u32_x))) +svuint32_t svexth_u32_x(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svexth_u64_x))) +svuint64_t svexth_u64_x(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svexth_u32_z))) +svuint32_t svexth_u32_z(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svexth_u64_z))) +svuint64_t svexth_u64_z(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svextw_s64_m))) +svint64_t svextw_s64_m(svint64_t, svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svextw_s64_x))) +svint64_t svextw_s64_x(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svextw_s64_z))) +svint64_t svextw_s64_z(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svextw_u64_m))) +svuint64_t svextw_u64_m(svuint64_t, svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svextw_u64_x))) +svuint64_t svextw_u64_x(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svextw_u64_z))) +svuint64_t svextw_u64_z(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svget2_u8))) +svuint8_t svget2_u8(svuint8x2_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svget2_u32))) +svuint32_t svget2_u32(svuint32x2_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svget2_u64))) +svuint64_t svget2_u64(svuint64x2_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svget2_u16))) +svuint16_t svget2_u16(svuint16x2_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svget2_s8))) +svint8_t svget2_s8(svint8x2_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svget2_f64))) +svfloat64_t svget2_f64(svfloat64x2_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svget2_f32))) +svfloat32_t svget2_f32(svfloat32x2_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svget2_f16))) +svfloat16_t svget2_f16(svfloat16x2_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svget2_s32))) +svint32_t svget2_s32(svint32x2_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svget2_s64))) +svint64_t svget2_s64(svint64x2_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svget2_s16))) +svint16_t svget2_s16(svint16x2_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svget3_u8))) +svuint8_t svget3_u8(svuint8x3_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svget3_u32))) +svuint32_t svget3_u32(svuint32x3_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svget3_u64))) +svuint64_t svget3_u64(svuint64x3_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svget3_u16))) +svuint16_t svget3_u16(svuint16x3_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svget3_s8))) +svint8_t svget3_s8(svint8x3_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svget3_f64))) +svfloat64_t svget3_f64(svfloat64x3_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svget3_f32))) +svfloat32_t svget3_f32(svfloat32x3_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svget3_f16))) +svfloat16_t svget3_f16(svfloat16x3_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svget3_s32))) +svint32_t svget3_s32(svint32x3_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svget3_s64))) +svint64_t svget3_s64(svint64x3_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svget3_s16))) +svint16_t svget3_s16(svint16x3_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svget4_u8))) +svuint8_t svget4_u8(svuint8x4_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svget4_u32))) +svuint32_t svget4_u32(svuint32x4_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svget4_u64))) +svuint64_t svget4_u64(svuint64x4_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svget4_u16))) +svuint16_t svget4_u16(svuint16x4_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svget4_s8))) +svint8_t svget4_s8(svint8x4_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svget4_f64))) +svfloat64_t svget4_f64(svfloat64x4_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svget4_f32))) +svfloat32_t svget4_f32(svfloat32x4_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svget4_f16))) +svfloat16_t svget4_f16(svfloat16x4_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svget4_s32))) +svint32_t svget4_s32(svint32x4_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svget4_s64))) +svint64_t svget4_s64(svint64x4_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svget4_s16))) +svint16_t svget4_s16(svint16x4_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svindex_u8))) +svuint8_t svindex_u8(uint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svindex_u32))) +svuint32_t svindex_u32(uint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svindex_u64))) +svuint64_t svindex_u64(uint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svindex_u16))) +svuint16_t svindex_u16(uint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svindex_s8))) +svint8_t svindex_s8(int8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svindex_s32))) +svint32_t svindex_s32(int32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svindex_s64))) +svint64_t svindex_s64(int64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svindex_s16))) +svint16_t svindex_s16(int16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svinsr_n_u8))) +svuint8_t svinsr_n_u8(svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svinsr_n_u32))) +svuint32_t svinsr_n_u32(svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svinsr_n_u64))) +svuint64_t svinsr_n_u64(svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svinsr_n_u16))) +svuint16_t svinsr_n_u16(svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svinsr_n_s8))) +svint8_t svinsr_n_s8(svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svinsr_n_f64))) +svfloat64_t svinsr_n_f64(svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svinsr_n_f32))) +svfloat32_t svinsr_n_f32(svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svinsr_n_f16))) +svfloat16_t svinsr_n_f16(svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svinsr_n_s32))) +svint32_t svinsr_n_s32(svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svinsr_n_s64))) +svint64_t svinsr_n_s64(svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svinsr_n_s16))) +svint16_t svinsr_n_s16(svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlasta_u8))) +uint8_t svlasta_u8(svbool_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlasta_u32))) +uint32_t svlasta_u32(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlasta_u64))) +uint64_t svlasta_u64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlasta_u16))) +uint16_t svlasta_u16(svbool_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlasta_s8))) +int8_t svlasta_s8(svbool_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlasta_f64))) +float64_t svlasta_f64(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlasta_f32))) +float32_t svlasta_f32(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlasta_f16))) +float16_t svlasta_f16(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlasta_s32))) +int32_t svlasta_s32(svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlasta_s64))) +int64_t svlasta_s64(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlasta_s16))) +int16_t svlasta_s16(svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlastb_u8))) +uint8_t svlastb_u8(svbool_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlastb_u32))) +uint32_t svlastb_u32(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlastb_u64))) +uint64_t svlastb_u64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlastb_u16))) +uint16_t svlastb_u16(svbool_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlastb_s8))) +int8_t svlastb_s8(svbool_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlastb_f64))) +float64_t svlastb_f64(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlastb_f32))) +float32_t svlastb_f32(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlastb_f16))) +float16_t svlastb_f16(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlastb_s32))) +int32_t svlastb_s32(svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlastb_s64))) +int64_t svlastb_s64(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlastb_s16))) +int16_t svlastb_s16(svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_u8))) +svuint8_t svld1_u8(svbool_t, uint8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_u32))) +svuint32_t svld1_u32(svbool_t, uint32_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_u64))) +svuint64_t svld1_u64(svbool_t, uint64_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_u16))) +svuint16_t svld1_u16(svbool_t, uint16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_s8))) +svint8_t svld1_s8(svbool_t, int8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_f64))) +svfloat64_t svld1_f64(svbool_t, float64_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_f32))) +svfloat32_t svld1_f32(svbool_t, float32_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_f16))) +svfloat16_t svld1_f16(svbool_t, float16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_s32))) +svint32_t svld1_s32(svbool_t, int32_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_s64))) +svint64_t svld1_s64(svbool_t, int64_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_s16))) +svint16_t svld1_s16(svbool_t, int16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_gather_u32base_index_u32))) +svuint32_t svld1_gather_u32base_index_u32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_gather_u64base_index_u64))) +svuint64_t svld1_gather_u64base_index_u64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_gather_u64base_index_f64))) +svfloat64_t svld1_gather_u64base_index_f64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_gather_u32base_index_f32))) +svfloat32_t svld1_gather_u32base_index_f32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_gather_u32base_index_s32))) +svint32_t svld1_gather_u32base_index_s32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_gather_u64base_index_s64))) +svint64_t svld1_gather_u64base_index_s64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_gather_u32base_offset_u32))) +svuint32_t svld1_gather_u32base_offset_u32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_gather_u64base_offset_u64))) +svuint64_t svld1_gather_u64base_offset_u64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_gather_u64base_offset_f64))) +svfloat64_t svld1_gather_u64base_offset_f64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_gather_u32base_offset_f32))) +svfloat32_t svld1_gather_u32base_offset_f32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_gather_u32base_offset_s32))) +svint32_t svld1_gather_u32base_offset_s32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_gather_u64base_offset_s64))) +svint64_t svld1_gather_u64base_offset_s64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_gather_u32base_u32))) +svuint32_t svld1_gather_u32base_u32(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_gather_u64base_u64))) +svuint64_t svld1_gather_u64base_u64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_gather_u64base_f64))) +svfloat64_t svld1_gather_u64base_f64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_gather_u32base_f32))) +svfloat32_t svld1_gather_u32base_f32(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_gather_u32base_s32))) +svint32_t svld1_gather_u32base_s32(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_gather_u64base_s64))) +svint64_t svld1_gather_u64base_s64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_gather_s32index_u32))) +svuint32_t svld1_gather_s32index_u32(svbool_t, uint32_t const *, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_gather_s32index_f32))) +svfloat32_t svld1_gather_s32index_f32(svbool_t, float32_t const *, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_gather_s32index_s32))) +svint32_t svld1_gather_s32index_s32(svbool_t, int32_t const *, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_gather_u32index_u32))) +svuint32_t svld1_gather_u32index_u32(svbool_t, uint32_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_gather_u32index_f32))) +svfloat32_t svld1_gather_u32index_f32(svbool_t, float32_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_gather_u32index_s32))) +svint32_t svld1_gather_u32index_s32(svbool_t, int32_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_gather_s64index_u64))) +svuint64_t svld1_gather_s64index_u64(svbool_t, uint64_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_gather_s64index_f64))) +svfloat64_t svld1_gather_s64index_f64(svbool_t, float64_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_gather_s64index_s64))) +svint64_t svld1_gather_s64index_s64(svbool_t, int64_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_gather_u64index_u64))) +svuint64_t svld1_gather_u64index_u64(svbool_t, uint64_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_gather_u64index_f64))) +svfloat64_t svld1_gather_u64index_f64(svbool_t, float64_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_gather_u64index_s64))) +svint64_t svld1_gather_u64index_s64(svbool_t, int64_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_gather_s32offset_u32))) +svuint32_t svld1_gather_s32offset_u32(svbool_t, uint32_t const *, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_gather_s32offset_f32))) +svfloat32_t svld1_gather_s32offset_f32(svbool_t, float32_t const *, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_gather_s32offset_s32))) +svint32_t svld1_gather_s32offset_s32(svbool_t, int32_t const *, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_gather_u32offset_u32))) +svuint32_t svld1_gather_u32offset_u32(svbool_t, uint32_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_gather_u32offset_f32))) +svfloat32_t svld1_gather_u32offset_f32(svbool_t, float32_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_gather_u32offset_s32))) +svint32_t svld1_gather_u32offset_s32(svbool_t, int32_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_gather_s64offset_u64))) +svuint64_t svld1_gather_s64offset_u64(svbool_t, uint64_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_gather_s64offset_f64))) +svfloat64_t svld1_gather_s64offset_f64(svbool_t, float64_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_gather_s64offset_s64))) +svint64_t svld1_gather_s64offset_s64(svbool_t, int64_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_gather_u64offset_u64))) +svuint64_t svld1_gather_u64offset_u64(svbool_t, uint64_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_gather_u64offset_f64))) +svfloat64_t svld1_gather_u64offset_f64(svbool_t, float64_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_gather_u64offset_s64))) +svint64_t svld1_gather_u64offset_s64(svbool_t, int64_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_vnum_u8))) +svuint8_t svld1_vnum_u8(svbool_t, uint8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_vnum_u32))) +svuint32_t svld1_vnum_u32(svbool_t, uint32_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_vnum_u64))) +svuint64_t svld1_vnum_u64(svbool_t, uint64_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_vnum_u16))) +svuint16_t svld1_vnum_u16(svbool_t, uint16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_vnum_s8))) +svint8_t svld1_vnum_s8(svbool_t, int8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_vnum_f64))) +svfloat64_t svld1_vnum_f64(svbool_t, float64_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_vnum_f32))) +svfloat32_t svld1_vnum_f32(svbool_t, float32_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_vnum_f16))) +svfloat16_t svld1_vnum_f16(svbool_t, float16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_vnum_s32))) +svint32_t svld1_vnum_s32(svbool_t, int32_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_vnum_s64))) +svint64_t svld1_vnum_s64(svbool_t, int64_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_vnum_s16))) +svint16_t svld1_vnum_s16(svbool_t, int16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1rq_u8))) +svuint8_t svld1rq_u8(svbool_t, uint8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1rq_u32))) +svuint32_t svld1rq_u32(svbool_t, uint32_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1rq_u64))) +svuint64_t svld1rq_u64(svbool_t, uint64_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1rq_u16))) +svuint16_t svld1rq_u16(svbool_t, uint16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1rq_s8))) +svint8_t svld1rq_s8(svbool_t, int8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1rq_f64))) +svfloat64_t svld1rq_f64(svbool_t, float64_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1rq_f32))) +svfloat32_t svld1rq_f32(svbool_t, float32_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1rq_f16))) +svfloat16_t svld1rq_f16(svbool_t, float16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1rq_s32))) +svint32_t svld1rq_s32(svbool_t, int32_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1rq_s64))) +svint64_t svld1rq_s64(svbool_t, int64_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1rq_s16))) +svint16_t svld1rq_s16(svbool_t, int16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sb_gather_u32base_offset_u32))) +svuint32_t svld1sb_gather_u32base_offset_u32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sb_gather_u64base_offset_u64))) +svuint64_t svld1sb_gather_u64base_offset_u64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sb_gather_u32base_offset_s32))) +svint32_t svld1sb_gather_u32base_offset_s32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sb_gather_u64base_offset_s64))) +svint64_t svld1sb_gather_u64base_offset_s64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sb_gather_u32base_u32))) +svuint32_t svld1sb_gather_u32base_u32(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sb_gather_u64base_u64))) +svuint64_t svld1sb_gather_u64base_u64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sb_gather_u32base_s32))) +svint32_t svld1sb_gather_u32base_s32(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sb_gather_u64base_s64))) +svint64_t svld1sb_gather_u64base_s64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sb_gather_s32offset_u32))) +svuint32_t svld1sb_gather_s32offset_u32(svbool_t, int8_t const *, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sb_gather_s32offset_s32))) +svint32_t svld1sb_gather_s32offset_s32(svbool_t, int8_t const *, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sb_gather_u32offset_u32))) +svuint32_t svld1sb_gather_u32offset_u32(svbool_t, int8_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sb_gather_u32offset_s32))) +svint32_t svld1sb_gather_u32offset_s32(svbool_t, int8_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sb_gather_s64offset_u64))) +svuint64_t svld1sb_gather_s64offset_u64(svbool_t, int8_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sb_gather_s64offset_s64))) +svint64_t svld1sb_gather_s64offset_s64(svbool_t, int8_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sb_gather_u64offset_u64))) +svuint64_t svld1sb_gather_u64offset_u64(svbool_t, int8_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sb_gather_u64offset_s64))) +svint64_t svld1sb_gather_u64offset_s64(svbool_t, int8_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sb_vnum_u32))) +svuint32_t svld1sb_vnum_u32(svbool_t, int8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sb_vnum_u64))) +svuint64_t svld1sb_vnum_u64(svbool_t, int8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sb_vnum_u16))) +svuint16_t svld1sb_vnum_u16(svbool_t, int8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sb_vnum_s32))) +svint32_t svld1sb_vnum_s32(svbool_t, int8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sb_vnum_s64))) +svint64_t svld1sb_vnum_s64(svbool_t, int8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sb_vnum_s16))) +svint16_t svld1sb_vnum_s16(svbool_t, int8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sb_u32))) +svuint32_t svld1sb_u32(svbool_t, int8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sb_u64))) +svuint64_t svld1sb_u64(svbool_t, int8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sb_u16))) +svuint16_t svld1sb_u16(svbool_t, int8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sb_s32))) +svint32_t svld1sb_s32(svbool_t, int8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sb_s64))) +svint64_t svld1sb_s64(svbool_t, int8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sb_s16))) +svint16_t svld1sb_s16(svbool_t, int8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sh_gather_u32base_index_u32))) +svuint32_t svld1sh_gather_u32base_index_u32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sh_gather_u64base_index_u64))) +svuint64_t svld1sh_gather_u64base_index_u64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sh_gather_u32base_index_s32))) +svint32_t svld1sh_gather_u32base_index_s32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sh_gather_u64base_index_s64))) +svint64_t svld1sh_gather_u64base_index_s64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sh_gather_u32base_offset_u32))) +svuint32_t svld1sh_gather_u32base_offset_u32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sh_gather_u64base_offset_u64))) +svuint64_t svld1sh_gather_u64base_offset_u64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sh_gather_u32base_offset_s32))) +svint32_t svld1sh_gather_u32base_offset_s32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sh_gather_u64base_offset_s64))) +svint64_t svld1sh_gather_u64base_offset_s64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sh_gather_u32base_u32))) +svuint32_t svld1sh_gather_u32base_u32(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sh_gather_u64base_u64))) +svuint64_t svld1sh_gather_u64base_u64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sh_gather_u32base_s32))) +svint32_t svld1sh_gather_u32base_s32(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sh_gather_u64base_s64))) +svint64_t svld1sh_gather_u64base_s64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sh_gather_s32index_u32))) +svuint32_t svld1sh_gather_s32index_u32(svbool_t, int16_t const *, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sh_gather_s32index_s32))) +svint32_t svld1sh_gather_s32index_s32(svbool_t, int16_t const *, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sh_gather_u32index_u32))) +svuint32_t svld1sh_gather_u32index_u32(svbool_t, int16_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sh_gather_u32index_s32))) +svint32_t svld1sh_gather_u32index_s32(svbool_t, int16_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sh_gather_s64index_u64))) +svuint64_t svld1sh_gather_s64index_u64(svbool_t, int16_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sh_gather_s64index_s64))) +svint64_t svld1sh_gather_s64index_s64(svbool_t, int16_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sh_gather_u64index_u64))) +svuint64_t svld1sh_gather_u64index_u64(svbool_t, int16_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sh_gather_u64index_s64))) +svint64_t svld1sh_gather_u64index_s64(svbool_t, int16_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sh_gather_s32offset_u32))) +svuint32_t svld1sh_gather_s32offset_u32(svbool_t, int16_t const *, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sh_gather_s32offset_s32))) +svint32_t svld1sh_gather_s32offset_s32(svbool_t, int16_t const *, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sh_gather_u32offset_u32))) +svuint32_t svld1sh_gather_u32offset_u32(svbool_t, int16_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sh_gather_u32offset_s32))) +svint32_t svld1sh_gather_u32offset_s32(svbool_t, int16_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sh_gather_s64offset_u64))) +svuint64_t svld1sh_gather_s64offset_u64(svbool_t, int16_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sh_gather_s64offset_s64))) +svint64_t svld1sh_gather_s64offset_s64(svbool_t, int16_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sh_gather_u64offset_u64))) +svuint64_t svld1sh_gather_u64offset_u64(svbool_t, int16_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sh_gather_u64offset_s64))) +svint64_t svld1sh_gather_u64offset_s64(svbool_t, int16_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sh_vnum_u32))) +svuint32_t svld1sh_vnum_u32(svbool_t, int16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sh_vnum_u64))) +svuint64_t svld1sh_vnum_u64(svbool_t, int16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sh_vnum_s32))) +svint32_t svld1sh_vnum_s32(svbool_t, int16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sh_vnum_s64))) +svint64_t svld1sh_vnum_s64(svbool_t, int16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sh_u32))) +svuint32_t svld1sh_u32(svbool_t, int16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sh_u64))) +svuint64_t svld1sh_u64(svbool_t, int16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sh_s32))) +svint32_t svld1sh_s32(svbool_t, int16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sh_s64))) +svint64_t svld1sh_s64(svbool_t, int16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sw_gather_u64base_index_u64))) +svuint64_t svld1sw_gather_u64base_index_u64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sw_gather_u64base_index_s64))) +svint64_t svld1sw_gather_u64base_index_s64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sw_gather_u64base_offset_u64))) +svuint64_t svld1sw_gather_u64base_offset_u64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sw_gather_u64base_offset_s64))) +svint64_t svld1sw_gather_u64base_offset_s64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sw_gather_u64base_u64))) +svuint64_t svld1sw_gather_u64base_u64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sw_gather_u64base_s64))) +svint64_t svld1sw_gather_u64base_s64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sw_gather_s64index_u64))) +svuint64_t svld1sw_gather_s64index_u64(svbool_t, int32_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sw_gather_s64index_s64))) +svint64_t svld1sw_gather_s64index_s64(svbool_t, int32_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sw_gather_u64index_u64))) +svuint64_t svld1sw_gather_u64index_u64(svbool_t, int32_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sw_gather_u64index_s64))) +svint64_t svld1sw_gather_u64index_s64(svbool_t, int32_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sw_gather_s64offset_u64))) +svuint64_t svld1sw_gather_s64offset_u64(svbool_t, int32_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sw_gather_s64offset_s64))) +svint64_t svld1sw_gather_s64offset_s64(svbool_t, int32_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sw_gather_u64offset_u64))) +svuint64_t svld1sw_gather_u64offset_u64(svbool_t, int32_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sw_gather_u64offset_s64))) +svint64_t svld1sw_gather_u64offset_s64(svbool_t, int32_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sw_vnum_u64))) +svuint64_t svld1sw_vnum_u64(svbool_t, int32_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sw_vnum_s64))) +svint64_t svld1sw_vnum_s64(svbool_t, int32_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sw_u64))) +svuint64_t svld1sw_u64(svbool_t, int32_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1sw_s64))) +svint64_t svld1sw_s64(svbool_t, int32_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1ub_gather_u32base_offset_u32))) +svuint32_t svld1ub_gather_u32base_offset_u32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1ub_gather_u64base_offset_u64))) +svuint64_t svld1ub_gather_u64base_offset_u64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1ub_gather_u32base_offset_s32))) +svint32_t svld1ub_gather_u32base_offset_s32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1ub_gather_u64base_offset_s64))) +svint64_t svld1ub_gather_u64base_offset_s64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1ub_gather_u32base_u32))) +svuint32_t svld1ub_gather_u32base_u32(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1ub_gather_u64base_u64))) +svuint64_t svld1ub_gather_u64base_u64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1ub_gather_u32base_s32))) +svint32_t svld1ub_gather_u32base_s32(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1ub_gather_u64base_s64))) +svint64_t svld1ub_gather_u64base_s64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1ub_gather_s32offset_u32))) +svuint32_t svld1ub_gather_s32offset_u32(svbool_t, uint8_t const *, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1ub_gather_s32offset_s32))) +svint32_t svld1ub_gather_s32offset_s32(svbool_t, uint8_t const *, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1ub_gather_u32offset_u32))) +svuint32_t svld1ub_gather_u32offset_u32(svbool_t, uint8_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1ub_gather_u32offset_s32))) +svint32_t svld1ub_gather_u32offset_s32(svbool_t, uint8_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1ub_gather_s64offset_u64))) +svuint64_t svld1ub_gather_s64offset_u64(svbool_t, uint8_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1ub_gather_s64offset_s64))) +svint64_t svld1ub_gather_s64offset_s64(svbool_t, uint8_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1ub_gather_u64offset_u64))) +svuint64_t svld1ub_gather_u64offset_u64(svbool_t, uint8_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1ub_gather_u64offset_s64))) +svint64_t svld1ub_gather_u64offset_s64(svbool_t, uint8_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1ub_vnum_u32))) +svuint32_t svld1ub_vnum_u32(svbool_t, uint8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1ub_vnum_u64))) +svuint64_t svld1ub_vnum_u64(svbool_t, uint8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1ub_vnum_u16))) +svuint16_t svld1ub_vnum_u16(svbool_t, uint8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1ub_vnum_s32))) +svint32_t svld1ub_vnum_s32(svbool_t, uint8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1ub_vnum_s64))) +svint64_t svld1ub_vnum_s64(svbool_t, uint8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1ub_vnum_s16))) +svint16_t svld1ub_vnum_s16(svbool_t, uint8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1ub_u32))) +svuint32_t svld1ub_u32(svbool_t, uint8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1ub_u64))) +svuint64_t svld1ub_u64(svbool_t, uint8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1ub_u16))) +svuint16_t svld1ub_u16(svbool_t, uint8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1ub_s32))) +svint32_t svld1ub_s32(svbool_t, uint8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1ub_s64))) +svint64_t svld1ub_s64(svbool_t, uint8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1ub_s16))) +svint16_t svld1ub_s16(svbool_t, uint8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uh_gather_u32base_index_u32))) +svuint32_t svld1uh_gather_u32base_index_u32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uh_gather_u64base_index_u64))) +svuint64_t svld1uh_gather_u64base_index_u64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uh_gather_u32base_index_s32))) +svint32_t svld1uh_gather_u32base_index_s32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uh_gather_u64base_index_s64))) +svint64_t svld1uh_gather_u64base_index_s64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uh_gather_u32base_offset_u32))) +svuint32_t svld1uh_gather_u32base_offset_u32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uh_gather_u64base_offset_u64))) +svuint64_t svld1uh_gather_u64base_offset_u64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uh_gather_u32base_offset_s32))) +svint32_t svld1uh_gather_u32base_offset_s32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uh_gather_u64base_offset_s64))) +svint64_t svld1uh_gather_u64base_offset_s64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uh_gather_u32base_u32))) +svuint32_t svld1uh_gather_u32base_u32(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uh_gather_u64base_u64))) +svuint64_t svld1uh_gather_u64base_u64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uh_gather_u32base_s32))) +svint32_t svld1uh_gather_u32base_s32(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uh_gather_u64base_s64))) +svint64_t svld1uh_gather_u64base_s64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uh_gather_s32index_u32))) +svuint32_t svld1uh_gather_s32index_u32(svbool_t, uint16_t const *, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uh_gather_s32index_s32))) +svint32_t svld1uh_gather_s32index_s32(svbool_t, uint16_t const *, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uh_gather_u32index_u32))) +svuint32_t svld1uh_gather_u32index_u32(svbool_t, uint16_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uh_gather_u32index_s32))) +svint32_t svld1uh_gather_u32index_s32(svbool_t, uint16_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uh_gather_s64index_u64))) +svuint64_t svld1uh_gather_s64index_u64(svbool_t, uint16_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uh_gather_s64index_s64))) +svint64_t svld1uh_gather_s64index_s64(svbool_t, uint16_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uh_gather_u64index_u64))) +svuint64_t svld1uh_gather_u64index_u64(svbool_t, uint16_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uh_gather_u64index_s64))) +svint64_t svld1uh_gather_u64index_s64(svbool_t, uint16_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uh_gather_s32offset_u32))) +svuint32_t svld1uh_gather_s32offset_u32(svbool_t, uint16_t const *, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uh_gather_s32offset_s32))) +svint32_t svld1uh_gather_s32offset_s32(svbool_t, uint16_t const *, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uh_gather_u32offset_u32))) +svuint32_t svld1uh_gather_u32offset_u32(svbool_t, uint16_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uh_gather_u32offset_s32))) +svint32_t svld1uh_gather_u32offset_s32(svbool_t, uint16_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uh_gather_s64offset_u64))) +svuint64_t svld1uh_gather_s64offset_u64(svbool_t, uint16_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uh_gather_s64offset_s64))) +svint64_t svld1uh_gather_s64offset_s64(svbool_t, uint16_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uh_gather_u64offset_u64))) +svuint64_t svld1uh_gather_u64offset_u64(svbool_t, uint16_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uh_gather_u64offset_s64))) +svint64_t svld1uh_gather_u64offset_s64(svbool_t, uint16_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uh_vnum_u32))) +svuint32_t svld1uh_vnum_u32(svbool_t, uint16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uh_vnum_u64))) +svuint64_t svld1uh_vnum_u64(svbool_t, uint16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uh_vnum_s32))) +svint32_t svld1uh_vnum_s32(svbool_t, uint16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uh_vnum_s64))) +svint64_t svld1uh_vnum_s64(svbool_t, uint16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uh_u32))) +svuint32_t svld1uh_u32(svbool_t, uint16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uh_u64))) +svuint64_t svld1uh_u64(svbool_t, uint16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uh_s32))) +svint32_t svld1uh_s32(svbool_t, uint16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uh_s64))) +svint64_t svld1uh_s64(svbool_t, uint16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uw_gather_u64base_index_u64))) +svuint64_t svld1uw_gather_u64base_index_u64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uw_gather_u64base_index_s64))) +svint64_t svld1uw_gather_u64base_index_s64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uw_gather_u64base_offset_u64))) +svuint64_t svld1uw_gather_u64base_offset_u64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uw_gather_u64base_offset_s64))) +svint64_t svld1uw_gather_u64base_offset_s64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uw_gather_u64base_u64))) +svuint64_t svld1uw_gather_u64base_u64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uw_gather_u64base_s64))) +svint64_t svld1uw_gather_u64base_s64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uw_gather_s64index_u64))) +svuint64_t svld1uw_gather_s64index_u64(svbool_t, uint32_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uw_gather_s64index_s64))) +svint64_t svld1uw_gather_s64index_s64(svbool_t, uint32_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uw_gather_u64index_u64))) +svuint64_t svld1uw_gather_u64index_u64(svbool_t, uint32_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uw_gather_u64index_s64))) +svint64_t svld1uw_gather_u64index_s64(svbool_t, uint32_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uw_gather_s64offset_u64))) +svuint64_t svld1uw_gather_s64offset_u64(svbool_t, uint32_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uw_gather_s64offset_s64))) +svint64_t svld1uw_gather_s64offset_s64(svbool_t, uint32_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uw_gather_u64offset_u64))) +svuint64_t svld1uw_gather_u64offset_u64(svbool_t, uint32_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uw_gather_u64offset_s64))) +svint64_t svld1uw_gather_u64offset_s64(svbool_t, uint32_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uw_vnum_u64))) +svuint64_t svld1uw_vnum_u64(svbool_t, uint32_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uw_vnum_s64))) +svint64_t svld1uw_vnum_s64(svbool_t, uint32_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uw_u64))) +svuint64_t svld1uw_u64(svbool_t, uint32_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1uw_s64))) +svint64_t svld1uw_s64(svbool_t, uint32_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld2_u8))) +svuint8x2_t svld2_u8(svbool_t, uint8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld2_u32))) +svuint32x2_t svld2_u32(svbool_t, uint32_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld2_u64))) +svuint64x2_t svld2_u64(svbool_t, uint64_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld2_u16))) +svuint16x2_t svld2_u16(svbool_t, uint16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld2_s8))) +svint8x2_t svld2_s8(svbool_t, int8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld2_f64))) +svfloat64x2_t svld2_f64(svbool_t, float64_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld2_f32))) +svfloat32x2_t svld2_f32(svbool_t, float32_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld2_f16))) +svfloat16x2_t svld2_f16(svbool_t, float16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld2_s32))) +svint32x2_t svld2_s32(svbool_t, int32_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld2_s64))) +svint64x2_t svld2_s64(svbool_t, int64_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld2_s16))) +svint16x2_t svld2_s16(svbool_t, int16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld2_vnum_u8))) +svuint8x2_t svld2_vnum_u8(svbool_t, uint8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld2_vnum_u32))) +svuint32x2_t svld2_vnum_u32(svbool_t, uint32_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld2_vnum_u64))) +svuint64x2_t svld2_vnum_u64(svbool_t, uint64_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld2_vnum_u16))) +svuint16x2_t svld2_vnum_u16(svbool_t, uint16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld2_vnum_s8))) +svint8x2_t svld2_vnum_s8(svbool_t, int8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld2_vnum_f64))) +svfloat64x2_t svld2_vnum_f64(svbool_t, float64_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld2_vnum_f32))) +svfloat32x2_t svld2_vnum_f32(svbool_t, float32_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld2_vnum_f16))) +svfloat16x2_t svld2_vnum_f16(svbool_t, float16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld2_vnum_s32))) +svint32x2_t svld2_vnum_s32(svbool_t, int32_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld2_vnum_s64))) +svint64x2_t svld2_vnum_s64(svbool_t, int64_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld2_vnum_s16))) +svint16x2_t svld2_vnum_s16(svbool_t, int16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld3_u8))) +svuint8x3_t svld3_u8(svbool_t, uint8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld3_u32))) +svuint32x3_t svld3_u32(svbool_t, uint32_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld3_u64))) +svuint64x3_t svld3_u64(svbool_t, uint64_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld3_u16))) +svuint16x3_t svld3_u16(svbool_t, uint16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld3_s8))) +svint8x3_t svld3_s8(svbool_t, int8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld3_f64))) +svfloat64x3_t svld3_f64(svbool_t, float64_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld3_f32))) +svfloat32x3_t svld3_f32(svbool_t, float32_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld3_f16))) +svfloat16x3_t svld3_f16(svbool_t, float16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld3_s32))) +svint32x3_t svld3_s32(svbool_t, int32_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld3_s64))) +svint64x3_t svld3_s64(svbool_t, int64_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld3_s16))) +svint16x3_t svld3_s16(svbool_t, int16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld3_vnum_u8))) +svuint8x3_t svld3_vnum_u8(svbool_t, uint8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld3_vnum_u32))) +svuint32x3_t svld3_vnum_u32(svbool_t, uint32_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld3_vnum_u64))) +svuint64x3_t svld3_vnum_u64(svbool_t, uint64_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld3_vnum_u16))) +svuint16x3_t svld3_vnum_u16(svbool_t, uint16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld3_vnum_s8))) +svint8x3_t svld3_vnum_s8(svbool_t, int8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld3_vnum_f64))) +svfloat64x3_t svld3_vnum_f64(svbool_t, float64_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld3_vnum_f32))) +svfloat32x3_t svld3_vnum_f32(svbool_t, float32_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld3_vnum_f16))) +svfloat16x3_t svld3_vnum_f16(svbool_t, float16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld3_vnum_s32))) +svint32x3_t svld3_vnum_s32(svbool_t, int32_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld3_vnum_s64))) +svint64x3_t svld3_vnum_s64(svbool_t, int64_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld3_vnum_s16))) +svint16x3_t svld3_vnum_s16(svbool_t, int16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld4_u8))) +svuint8x4_t svld4_u8(svbool_t, uint8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld4_u32))) +svuint32x4_t svld4_u32(svbool_t, uint32_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld4_u64))) +svuint64x4_t svld4_u64(svbool_t, uint64_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld4_u16))) +svuint16x4_t svld4_u16(svbool_t, uint16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld4_s8))) +svint8x4_t svld4_s8(svbool_t, int8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld4_f64))) +svfloat64x4_t svld4_f64(svbool_t, float64_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld4_f32))) +svfloat32x4_t svld4_f32(svbool_t, float32_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld4_f16))) +svfloat16x4_t svld4_f16(svbool_t, float16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld4_s32))) +svint32x4_t svld4_s32(svbool_t, int32_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld4_s64))) +svint64x4_t svld4_s64(svbool_t, int64_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld4_s16))) +svint16x4_t svld4_s16(svbool_t, int16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld4_vnum_u8))) +svuint8x4_t svld4_vnum_u8(svbool_t, uint8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld4_vnum_u32))) +svuint32x4_t svld4_vnum_u32(svbool_t, uint32_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld4_vnum_u64))) +svuint64x4_t svld4_vnum_u64(svbool_t, uint64_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld4_vnum_u16))) +svuint16x4_t svld4_vnum_u16(svbool_t, uint16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld4_vnum_s8))) +svint8x4_t svld4_vnum_s8(svbool_t, int8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld4_vnum_f64))) +svfloat64x4_t svld4_vnum_f64(svbool_t, float64_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld4_vnum_f32))) +svfloat32x4_t svld4_vnum_f32(svbool_t, float32_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld4_vnum_f16))) +svfloat16x4_t svld4_vnum_f16(svbool_t, float16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld4_vnum_s32))) +svint32x4_t svld4_vnum_s32(svbool_t, int32_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld4_vnum_s64))) +svint64x4_t svld4_vnum_s64(svbool_t, int64_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld4_vnum_s16))) +svint16x4_t svld4_vnum_s16(svbool_t, int16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_u8))) +svuint8_t svldff1_u8(svbool_t, uint8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_u32))) +svuint32_t svldff1_u32(svbool_t, uint32_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_u64))) +svuint64_t svldff1_u64(svbool_t, uint64_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_u16))) +svuint16_t svldff1_u16(svbool_t, uint16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_s8))) +svint8_t svldff1_s8(svbool_t, int8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_f64))) +svfloat64_t svldff1_f64(svbool_t, float64_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_f32))) +svfloat32_t svldff1_f32(svbool_t, float32_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_f16))) +svfloat16_t svldff1_f16(svbool_t, float16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_s32))) +svint32_t svldff1_s32(svbool_t, int32_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_s64))) +svint64_t svldff1_s64(svbool_t, int64_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_s16))) +svint16_t svldff1_s16(svbool_t, int16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_gather_u32base_index_u32))) +svuint32_t svldff1_gather_u32base_index_u32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_gather_u64base_index_u64))) +svuint64_t svldff1_gather_u64base_index_u64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_gather_u64base_index_f64))) +svfloat64_t svldff1_gather_u64base_index_f64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_gather_u32base_index_f32))) +svfloat32_t svldff1_gather_u32base_index_f32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_gather_u32base_index_s32))) +svint32_t svldff1_gather_u32base_index_s32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_gather_u64base_index_s64))) +svint64_t svldff1_gather_u64base_index_s64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_gather_u32base_offset_u32))) +svuint32_t svldff1_gather_u32base_offset_u32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_gather_u64base_offset_u64))) +svuint64_t svldff1_gather_u64base_offset_u64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_gather_u64base_offset_f64))) +svfloat64_t svldff1_gather_u64base_offset_f64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_gather_u32base_offset_f32))) +svfloat32_t svldff1_gather_u32base_offset_f32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_gather_u32base_offset_s32))) +svint32_t svldff1_gather_u32base_offset_s32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_gather_u64base_offset_s64))) +svint64_t svldff1_gather_u64base_offset_s64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_gather_u32base_u32))) +svuint32_t svldff1_gather_u32base_u32(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_gather_u64base_u64))) +svuint64_t svldff1_gather_u64base_u64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_gather_u64base_f64))) +svfloat64_t svldff1_gather_u64base_f64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_gather_u32base_f32))) +svfloat32_t svldff1_gather_u32base_f32(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_gather_u32base_s32))) +svint32_t svldff1_gather_u32base_s32(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_gather_u64base_s64))) +svint64_t svldff1_gather_u64base_s64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_gather_s32index_u32))) +svuint32_t svldff1_gather_s32index_u32(svbool_t, uint32_t const *, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_gather_s32index_f32))) +svfloat32_t svldff1_gather_s32index_f32(svbool_t, float32_t const *, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_gather_s32index_s32))) +svint32_t svldff1_gather_s32index_s32(svbool_t, int32_t const *, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_gather_u32index_u32))) +svuint32_t svldff1_gather_u32index_u32(svbool_t, uint32_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_gather_u32index_f32))) +svfloat32_t svldff1_gather_u32index_f32(svbool_t, float32_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_gather_u32index_s32))) +svint32_t svldff1_gather_u32index_s32(svbool_t, int32_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_gather_s64index_u64))) +svuint64_t svldff1_gather_s64index_u64(svbool_t, uint64_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_gather_s64index_f64))) +svfloat64_t svldff1_gather_s64index_f64(svbool_t, float64_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_gather_s64index_s64))) +svint64_t svldff1_gather_s64index_s64(svbool_t, int64_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_gather_u64index_u64))) +svuint64_t svldff1_gather_u64index_u64(svbool_t, uint64_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_gather_u64index_f64))) +svfloat64_t svldff1_gather_u64index_f64(svbool_t, float64_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_gather_u64index_s64))) +svint64_t svldff1_gather_u64index_s64(svbool_t, int64_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_gather_s32offset_u32))) +svuint32_t svldff1_gather_s32offset_u32(svbool_t, uint32_t const *, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_gather_s32offset_f32))) +svfloat32_t svldff1_gather_s32offset_f32(svbool_t, float32_t const *, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_gather_s32offset_s32))) +svint32_t svldff1_gather_s32offset_s32(svbool_t, int32_t const *, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_gather_u32offset_u32))) +svuint32_t svldff1_gather_u32offset_u32(svbool_t, uint32_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_gather_u32offset_f32))) +svfloat32_t svldff1_gather_u32offset_f32(svbool_t, float32_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_gather_u32offset_s32))) +svint32_t svldff1_gather_u32offset_s32(svbool_t, int32_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_gather_s64offset_u64))) +svuint64_t svldff1_gather_s64offset_u64(svbool_t, uint64_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_gather_s64offset_f64))) +svfloat64_t svldff1_gather_s64offset_f64(svbool_t, float64_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_gather_s64offset_s64))) +svint64_t svldff1_gather_s64offset_s64(svbool_t, int64_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_gather_u64offset_u64))) +svuint64_t svldff1_gather_u64offset_u64(svbool_t, uint64_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_gather_u64offset_f64))) +svfloat64_t svldff1_gather_u64offset_f64(svbool_t, float64_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_gather_u64offset_s64))) +svint64_t svldff1_gather_u64offset_s64(svbool_t, int64_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_vnum_u8))) +svuint8_t svldff1_vnum_u8(svbool_t, uint8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_vnum_u32))) +svuint32_t svldff1_vnum_u32(svbool_t, uint32_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_vnum_u64))) +svuint64_t svldff1_vnum_u64(svbool_t, uint64_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_vnum_u16))) +svuint16_t svldff1_vnum_u16(svbool_t, uint16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_vnum_s8))) +svint8_t svldff1_vnum_s8(svbool_t, int8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_vnum_f64))) +svfloat64_t svldff1_vnum_f64(svbool_t, float64_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_vnum_f32))) +svfloat32_t svldff1_vnum_f32(svbool_t, float32_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_vnum_f16))) +svfloat16_t svldff1_vnum_f16(svbool_t, float16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_vnum_s32))) +svint32_t svldff1_vnum_s32(svbool_t, int32_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_vnum_s64))) +svint64_t svldff1_vnum_s64(svbool_t, int64_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_vnum_s16))) +svint16_t svldff1_vnum_s16(svbool_t, int16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sb_gather_u32base_offset_u32))) +svuint32_t svldff1sb_gather_u32base_offset_u32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sb_gather_u64base_offset_u64))) +svuint64_t svldff1sb_gather_u64base_offset_u64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sb_gather_u32base_offset_s32))) +svint32_t svldff1sb_gather_u32base_offset_s32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sb_gather_u64base_offset_s64))) +svint64_t svldff1sb_gather_u64base_offset_s64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sb_gather_u32base_u32))) +svuint32_t svldff1sb_gather_u32base_u32(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sb_gather_u64base_u64))) +svuint64_t svldff1sb_gather_u64base_u64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sb_gather_u32base_s32))) +svint32_t svldff1sb_gather_u32base_s32(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sb_gather_u64base_s64))) +svint64_t svldff1sb_gather_u64base_s64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sb_gather_s32offset_u32))) +svuint32_t svldff1sb_gather_s32offset_u32(svbool_t, int8_t const *, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sb_gather_s32offset_s32))) +svint32_t svldff1sb_gather_s32offset_s32(svbool_t, int8_t const *, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sb_gather_u32offset_u32))) +svuint32_t svldff1sb_gather_u32offset_u32(svbool_t, int8_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sb_gather_u32offset_s32))) +svint32_t svldff1sb_gather_u32offset_s32(svbool_t, int8_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sb_gather_s64offset_u64))) +svuint64_t svldff1sb_gather_s64offset_u64(svbool_t, int8_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sb_gather_s64offset_s64))) +svint64_t svldff1sb_gather_s64offset_s64(svbool_t, int8_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sb_gather_u64offset_u64))) +svuint64_t svldff1sb_gather_u64offset_u64(svbool_t, int8_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sb_gather_u64offset_s64))) +svint64_t svldff1sb_gather_u64offset_s64(svbool_t, int8_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sb_vnum_u32))) +svuint32_t svldff1sb_vnum_u32(svbool_t, int8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sb_vnum_u64))) +svuint64_t svldff1sb_vnum_u64(svbool_t, int8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sb_vnum_u16))) +svuint16_t svldff1sb_vnum_u16(svbool_t, int8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sb_vnum_s32))) +svint32_t svldff1sb_vnum_s32(svbool_t, int8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sb_vnum_s64))) +svint64_t svldff1sb_vnum_s64(svbool_t, int8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sb_vnum_s16))) +svint16_t svldff1sb_vnum_s16(svbool_t, int8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sb_u32))) +svuint32_t svldff1sb_u32(svbool_t, int8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sb_u64))) +svuint64_t svldff1sb_u64(svbool_t, int8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sb_u16))) +svuint16_t svldff1sb_u16(svbool_t, int8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sb_s32))) +svint32_t svldff1sb_s32(svbool_t, int8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sb_s64))) +svint64_t svldff1sb_s64(svbool_t, int8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sb_s16))) +svint16_t svldff1sb_s16(svbool_t, int8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sh_gather_u32base_index_u32))) +svuint32_t svldff1sh_gather_u32base_index_u32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sh_gather_u64base_index_u64))) +svuint64_t svldff1sh_gather_u64base_index_u64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sh_gather_u32base_index_s32))) +svint32_t svldff1sh_gather_u32base_index_s32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sh_gather_u64base_index_s64))) +svint64_t svldff1sh_gather_u64base_index_s64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sh_gather_u32base_offset_u32))) +svuint32_t svldff1sh_gather_u32base_offset_u32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sh_gather_u64base_offset_u64))) +svuint64_t svldff1sh_gather_u64base_offset_u64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sh_gather_u32base_offset_s32))) +svint32_t svldff1sh_gather_u32base_offset_s32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sh_gather_u64base_offset_s64))) +svint64_t svldff1sh_gather_u64base_offset_s64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sh_gather_u32base_u32))) +svuint32_t svldff1sh_gather_u32base_u32(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sh_gather_u64base_u64))) +svuint64_t svldff1sh_gather_u64base_u64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sh_gather_u32base_s32))) +svint32_t svldff1sh_gather_u32base_s32(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sh_gather_u64base_s64))) +svint64_t svldff1sh_gather_u64base_s64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sh_gather_s32index_u32))) +svuint32_t svldff1sh_gather_s32index_u32(svbool_t, int16_t const *, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sh_gather_s32index_s32))) +svint32_t svldff1sh_gather_s32index_s32(svbool_t, int16_t const *, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sh_gather_u32index_u32))) +svuint32_t svldff1sh_gather_u32index_u32(svbool_t, int16_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sh_gather_u32index_s32))) +svint32_t svldff1sh_gather_u32index_s32(svbool_t, int16_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sh_gather_s64index_u64))) +svuint64_t svldff1sh_gather_s64index_u64(svbool_t, int16_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sh_gather_s64index_s64))) +svint64_t svldff1sh_gather_s64index_s64(svbool_t, int16_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sh_gather_u64index_u64))) +svuint64_t svldff1sh_gather_u64index_u64(svbool_t, int16_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sh_gather_u64index_s64))) +svint64_t svldff1sh_gather_u64index_s64(svbool_t, int16_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sh_gather_s32offset_u32))) +svuint32_t svldff1sh_gather_s32offset_u32(svbool_t, int16_t const *, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sh_gather_s32offset_s32))) +svint32_t svldff1sh_gather_s32offset_s32(svbool_t, int16_t const *, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sh_gather_u32offset_u32))) +svuint32_t svldff1sh_gather_u32offset_u32(svbool_t, int16_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sh_gather_u32offset_s32))) +svint32_t svldff1sh_gather_u32offset_s32(svbool_t, int16_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sh_gather_s64offset_u64))) +svuint64_t svldff1sh_gather_s64offset_u64(svbool_t, int16_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sh_gather_s64offset_s64))) +svint64_t svldff1sh_gather_s64offset_s64(svbool_t, int16_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sh_gather_u64offset_u64))) +svuint64_t svldff1sh_gather_u64offset_u64(svbool_t, int16_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sh_gather_u64offset_s64))) +svint64_t svldff1sh_gather_u64offset_s64(svbool_t, int16_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sh_vnum_u32))) +svuint32_t svldff1sh_vnum_u32(svbool_t, int16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sh_vnum_u64))) +svuint64_t svldff1sh_vnum_u64(svbool_t, int16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sh_vnum_s32))) +svint32_t svldff1sh_vnum_s32(svbool_t, int16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sh_vnum_s64))) +svint64_t svldff1sh_vnum_s64(svbool_t, int16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sh_u32))) +svuint32_t svldff1sh_u32(svbool_t, int16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sh_u64))) +svuint64_t svldff1sh_u64(svbool_t, int16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sh_s32))) +svint32_t svldff1sh_s32(svbool_t, int16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sh_s64))) +svint64_t svldff1sh_s64(svbool_t, int16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sw_gather_u64base_index_u64))) +svuint64_t svldff1sw_gather_u64base_index_u64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sw_gather_u64base_index_s64))) +svint64_t svldff1sw_gather_u64base_index_s64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sw_gather_u64base_offset_u64))) +svuint64_t svldff1sw_gather_u64base_offset_u64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sw_gather_u64base_offset_s64))) +svint64_t svldff1sw_gather_u64base_offset_s64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sw_gather_u64base_u64))) +svuint64_t svldff1sw_gather_u64base_u64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sw_gather_u64base_s64))) +svint64_t svldff1sw_gather_u64base_s64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sw_gather_s64index_u64))) +svuint64_t svldff1sw_gather_s64index_u64(svbool_t, int32_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sw_gather_s64index_s64))) +svint64_t svldff1sw_gather_s64index_s64(svbool_t, int32_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sw_gather_u64index_u64))) +svuint64_t svldff1sw_gather_u64index_u64(svbool_t, int32_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sw_gather_u64index_s64))) +svint64_t svldff1sw_gather_u64index_s64(svbool_t, int32_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sw_gather_s64offset_u64))) +svuint64_t svldff1sw_gather_s64offset_u64(svbool_t, int32_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sw_gather_s64offset_s64))) +svint64_t svldff1sw_gather_s64offset_s64(svbool_t, int32_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sw_gather_u64offset_u64))) +svuint64_t svldff1sw_gather_u64offset_u64(svbool_t, int32_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sw_gather_u64offset_s64))) +svint64_t svldff1sw_gather_u64offset_s64(svbool_t, int32_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sw_vnum_u64))) +svuint64_t svldff1sw_vnum_u64(svbool_t, int32_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sw_vnum_s64))) +svint64_t svldff1sw_vnum_s64(svbool_t, int32_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sw_u64))) +svuint64_t svldff1sw_u64(svbool_t, int32_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1sw_s64))) +svint64_t svldff1sw_s64(svbool_t, int32_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1ub_gather_u32base_offset_u32))) +svuint32_t svldff1ub_gather_u32base_offset_u32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1ub_gather_u64base_offset_u64))) +svuint64_t svldff1ub_gather_u64base_offset_u64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1ub_gather_u32base_offset_s32))) +svint32_t svldff1ub_gather_u32base_offset_s32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1ub_gather_u64base_offset_s64))) +svint64_t svldff1ub_gather_u64base_offset_s64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1ub_gather_u32base_u32))) +svuint32_t svldff1ub_gather_u32base_u32(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1ub_gather_u64base_u64))) +svuint64_t svldff1ub_gather_u64base_u64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1ub_gather_u32base_s32))) +svint32_t svldff1ub_gather_u32base_s32(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1ub_gather_u64base_s64))) +svint64_t svldff1ub_gather_u64base_s64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1ub_gather_s32offset_u32))) +svuint32_t svldff1ub_gather_s32offset_u32(svbool_t, uint8_t const *, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1ub_gather_s32offset_s32))) +svint32_t svldff1ub_gather_s32offset_s32(svbool_t, uint8_t const *, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1ub_gather_u32offset_u32))) +svuint32_t svldff1ub_gather_u32offset_u32(svbool_t, uint8_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1ub_gather_u32offset_s32))) +svint32_t svldff1ub_gather_u32offset_s32(svbool_t, uint8_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1ub_gather_s64offset_u64))) +svuint64_t svldff1ub_gather_s64offset_u64(svbool_t, uint8_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1ub_gather_s64offset_s64))) +svint64_t svldff1ub_gather_s64offset_s64(svbool_t, uint8_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1ub_gather_u64offset_u64))) +svuint64_t svldff1ub_gather_u64offset_u64(svbool_t, uint8_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1ub_gather_u64offset_s64))) +svint64_t svldff1ub_gather_u64offset_s64(svbool_t, uint8_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1ub_vnum_u32))) +svuint32_t svldff1ub_vnum_u32(svbool_t, uint8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1ub_vnum_u64))) +svuint64_t svldff1ub_vnum_u64(svbool_t, uint8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1ub_vnum_u16))) +svuint16_t svldff1ub_vnum_u16(svbool_t, uint8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1ub_vnum_s32))) +svint32_t svldff1ub_vnum_s32(svbool_t, uint8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1ub_vnum_s64))) +svint64_t svldff1ub_vnum_s64(svbool_t, uint8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1ub_vnum_s16))) +svint16_t svldff1ub_vnum_s16(svbool_t, uint8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1ub_u32))) +svuint32_t svldff1ub_u32(svbool_t, uint8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1ub_u64))) +svuint64_t svldff1ub_u64(svbool_t, uint8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1ub_u16))) +svuint16_t svldff1ub_u16(svbool_t, uint8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1ub_s32))) +svint32_t svldff1ub_s32(svbool_t, uint8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1ub_s64))) +svint64_t svldff1ub_s64(svbool_t, uint8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1ub_s16))) +svint16_t svldff1ub_s16(svbool_t, uint8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uh_gather_u32base_index_u32))) +svuint32_t svldff1uh_gather_u32base_index_u32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uh_gather_u64base_index_u64))) +svuint64_t svldff1uh_gather_u64base_index_u64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uh_gather_u32base_index_s32))) +svint32_t svldff1uh_gather_u32base_index_s32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uh_gather_u64base_index_s64))) +svint64_t svldff1uh_gather_u64base_index_s64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uh_gather_u32base_offset_u32))) +svuint32_t svldff1uh_gather_u32base_offset_u32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uh_gather_u64base_offset_u64))) +svuint64_t svldff1uh_gather_u64base_offset_u64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uh_gather_u32base_offset_s32))) +svint32_t svldff1uh_gather_u32base_offset_s32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uh_gather_u64base_offset_s64))) +svint64_t svldff1uh_gather_u64base_offset_s64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uh_gather_u32base_u32))) +svuint32_t svldff1uh_gather_u32base_u32(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uh_gather_u64base_u64))) +svuint64_t svldff1uh_gather_u64base_u64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uh_gather_u32base_s32))) +svint32_t svldff1uh_gather_u32base_s32(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uh_gather_u64base_s64))) +svint64_t svldff1uh_gather_u64base_s64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uh_gather_s32index_u32))) +svuint32_t svldff1uh_gather_s32index_u32(svbool_t, uint16_t const *, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uh_gather_s32index_s32))) +svint32_t svldff1uh_gather_s32index_s32(svbool_t, uint16_t const *, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uh_gather_u32index_u32))) +svuint32_t svldff1uh_gather_u32index_u32(svbool_t, uint16_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uh_gather_u32index_s32))) +svint32_t svldff1uh_gather_u32index_s32(svbool_t, uint16_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uh_gather_s64index_u64))) +svuint64_t svldff1uh_gather_s64index_u64(svbool_t, uint16_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uh_gather_s64index_s64))) +svint64_t svldff1uh_gather_s64index_s64(svbool_t, uint16_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uh_gather_u64index_u64))) +svuint64_t svldff1uh_gather_u64index_u64(svbool_t, uint16_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uh_gather_u64index_s64))) +svint64_t svldff1uh_gather_u64index_s64(svbool_t, uint16_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uh_gather_s32offset_u32))) +svuint32_t svldff1uh_gather_s32offset_u32(svbool_t, uint16_t const *, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uh_gather_s32offset_s32))) +svint32_t svldff1uh_gather_s32offset_s32(svbool_t, uint16_t const *, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uh_gather_u32offset_u32))) +svuint32_t svldff1uh_gather_u32offset_u32(svbool_t, uint16_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uh_gather_u32offset_s32))) +svint32_t svldff1uh_gather_u32offset_s32(svbool_t, uint16_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uh_gather_s64offset_u64))) +svuint64_t svldff1uh_gather_s64offset_u64(svbool_t, uint16_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uh_gather_s64offset_s64))) +svint64_t svldff1uh_gather_s64offset_s64(svbool_t, uint16_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uh_gather_u64offset_u64))) +svuint64_t svldff1uh_gather_u64offset_u64(svbool_t, uint16_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uh_gather_u64offset_s64))) +svint64_t svldff1uh_gather_u64offset_s64(svbool_t, uint16_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uh_vnum_u32))) +svuint32_t svldff1uh_vnum_u32(svbool_t, uint16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uh_vnum_u64))) +svuint64_t svldff1uh_vnum_u64(svbool_t, uint16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uh_vnum_s32))) +svint32_t svldff1uh_vnum_s32(svbool_t, uint16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uh_vnum_s64))) +svint64_t svldff1uh_vnum_s64(svbool_t, uint16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uh_u32))) +svuint32_t svldff1uh_u32(svbool_t, uint16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uh_u64))) +svuint64_t svldff1uh_u64(svbool_t, uint16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uh_s32))) +svint32_t svldff1uh_s32(svbool_t, uint16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uh_s64))) +svint64_t svldff1uh_s64(svbool_t, uint16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uw_gather_u64base_index_u64))) +svuint64_t svldff1uw_gather_u64base_index_u64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uw_gather_u64base_index_s64))) +svint64_t svldff1uw_gather_u64base_index_s64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uw_gather_u64base_offset_u64))) +svuint64_t svldff1uw_gather_u64base_offset_u64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uw_gather_u64base_offset_s64))) +svint64_t svldff1uw_gather_u64base_offset_s64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uw_gather_u64base_u64))) +svuint64_t svldff1uw_gather_u64base_u64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uw_gather_u64base_s64))) +svint64_t svldff1uw_gather_u64base_s64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uw_gather_s64index_u64))) +svuint64_t svldff1uw_gather_s64index_u64(svbool_t, uint32_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uw_gather_s64index_s64))) +svint64_t svldff1uw_gather_s64index_s64(svbool_t, uint32_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uw_gather_u64index_u64))) +svuint64_t svldff1uw_gather_u64index_u64(svbool_t, uint32_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uw_gather_u64index_s64))) +svint64_t svldff1uw_gather_u64index_s64(svbool_t, uint32_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uw_gather_s64offset_u64))) +svuint64_t svldff1uw_gather_s64offset_u64(svbool_t, uint32_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uw_gather_s64offset_s64))) +svint64_t svldff1uw_gather_s64offset_s64(svbool_t, uint32_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uw_gather_u64offset_u64))) +svuint64_t svldff1uw_gather_u64offset_u64(svbool_t, uint32_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uw_gather_u64offset_s64))) +svint64_t svldff1uw_gather_u64offset_s64(svbool_t, uint32_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uw_vnum_u64))) +svuint64_t svldff1uw_vnum_u64(svbool_t, uint32_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uw_vnum_s64))) +svint64_t svldff1uw_vnum_s64(svbool_t, uint32_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uw_u64))) +svuint64_t svldff1uw_u64(svbool_t, uint32_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1uw_s64))) +svint64_t svldff1uw_s64(svbool_t, uint32_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1_u8))) +svuint8_t svldnf1_u8(svbool_t, uint8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1_u32))) +svuint32_t svldnf1_u32(svbool_t, uint32_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1_u64))) +svuint64_t svldnf1_u64(svbool_t, uint64_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1_u16))) +svuint16_t svldnf1_u16(svbool_t, uint16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1_s8))) +svint8_t svldnf1_s8(svbool_t, int8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1_f64))) +svfloat64_t svldnf1_f64(svbool_t, float64_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1_f32))) +svfloat32_t svldnf1_f32(svbool_t, float32_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1_f16))) +svfloat16_t svldnf1_f16(svbool_t, float16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1_s32))) +svint32_t svldnf1_s32(svbool_t, int32_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1_s64))) +svint64_t svldnf1_s64(svbool_t, int64_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1_s16))) +svint16_t svldnf1_s16(svbool_t, int16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1_vnum_u8))) +svuint8_t svldnf1_vnum_u8(svbool_t, uint8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1_vnum_u32))) +svuint32_t svldnf1_vnum_u32(svbool_t, uint32_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1_vnum_u64))) +svuint64_t svldnf1_vnum_u64(svbool_t, uint64_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1_vnum_u16))) +svuint16_t svldnf1_vnum_u16(svbool_t, uint16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1_vnum_s8))) +svint8_t svldnf1_vnum_s8(svbool_t, int8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1_vnum_f64))) +svfloat64_t svldnf1_vnum_f64(svbool_t, float64_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1_vnum_f32))) +svfloat32_t svldnf1_vnum_f32(svbool_t, float32_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1_vnum_f16))) +svfloat16_t svldnf1_vnum_f16(svbool_t, float16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1_vnum_s32))) +svint32_t svldnf1_vnum_s32(svbool_t, int32_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1_vnum_s64))) +svint64_t svldnf1_vnum_s64(svbool_t, int64_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1_vnum_s16))) +svint16_t svldnf1_vnum_s16(svbool_t, int16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1sb_vnum_u32))) +svuint32_t svldnf1sb_vnum_u32(svbool_t, int8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1sb_vnum_u64))) +svuint64_t svldnf1sb_vnum_u64(svbool_t, int8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1sb_vnum_u16))) +svuint16_t svldnf1sb_vnum_u16(svbool_t, int8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1sb_vnum_s32))) +svint32_t svldnf1sb_vnum_s32(svbool_t, int8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1sb_vnum_s64))) +svint64_t svldnf1sb_vnum_s64(svbool_t, int8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1sb_vnum_s16))) +svint16_t svldnf1sb_vnum_s16(svbool_t, int8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1sb_u32))) +svuint32_t svldnf1sb_u32(svbool_t, int8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1sb_u64))) +svuint64_t svldnf1sb_u64(svbool_t, int8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1sb_u16))) +svuint16_t svldnf1sb_u16(svbool_t, int8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1sb_s32))) +svint32_t svldnf1sb_s32(svbool_t, int8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1sb_s64))) +svint64_t svldnf1sb_s64(svbool_t, int8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1sb_s16))) +svint16_t svldnf1sb_s16(svbool_t, int8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1sh_vnum_u32))) +svuint32_t svldnf1sh_vnum_u32(svbool_t, int16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1sh_vnum_u64))) +svuint64_t svldnf1sh_vnum_u64(svbool_t, int16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1sh_vnum_s32))) +svint32_t svldnf1sh_vnum_s32(svbool_t, int16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1sh_vnum_s64))) +svint64_t svldnf1sh_vnum_s64(svbool_t, int16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1sh_u32))) +svuint32_t svldnf1sh_u32(svbool_t, int16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1sh_u64))) +svuint64_t svldnf1sh_u64(svbool_t, int16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1sh_s32))) +svint32_t svldnf1sh_s32(svbool_t, int16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1sh_s64))) +svint64_t svldnf1sh_s64(svbool_t, int16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1sw_vnum_u64))) +svuint64_t svldnf1sw_vnum_u64(svbool_t, int32_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1sw_vnum_s64))) +svint64_t svldnf1sw_vnum_s64(svbool_t, int32_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1sw_u64))) +svuint64_t svldnf1sw_u64(svbool_t, int32_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1sw_s64))) +svint64_t svldnf1sw_s64(svbool_t, int32_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1ub_vnum_u32))) +svuint32_t svldnf1ub_vnum_u32(svbool_t, uint8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1ub_vnum_u64))) +svuint64_t svldnf1ub_vnum_u64(svbool_t, uint8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1ub_vnum_u16))) +svuint16_t svldnf1ub_vnum_u16(svbool_t, uint8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1ub_vnum_s32))) +svint32_t svldnf1ub_vnum_s32(svbool_t, uint8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1ub_vnum_s64))) +svint64_t svldnf1ub_vnum_s64(svbool_t, uint8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1ub_vnum_s16))) +svint16_t svldnf1ub_vnum_s16(svbool_t, uint8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1ub_u32))) +svuint32_t svldnf1ub_u32(svbool_t, uint8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1ub_u64))) +svuint64_t svldnf1ub_u64(svbool_t, uint8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1ub_u16))) +svuint16_t svldnf1ub_u16(svbool_t, uint8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1ub_s32))) +svint32_t svldnf1ub_s32(svbool_t, uint8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1ub_s64))) +svint64_t svldnf1ub_s64(svbool_t, uint8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1ub_s16))) +svint16_t svldnf1ub_s16(svbool_t, uint8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1uh_vnum_u32))) +svuint32_t svldnf1uh_vnum_u32(svbool_t, uint16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1uh_vnum_u64))) +svuint64_t svldnf1uh_vnum_u64(svbool_t, uint16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1uh_vnum_s32))) +svint32_t svldnf1uh_vnum_s32(svbool_t, uint16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1uh_vnum_s64))) +svint64_t svldnf1uh_vnum_s64(svbool_t, uint16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1uh_u32))) +svuint32_t svldnf1uh_u32(svbool_t, uint16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1uh_u64))) +svuint64_t svldnf1uh_u64(svbool_t, uint16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1uh_s32))) +svint32_t svldnf1uh_s32(svbool_t, uint16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1uh_s64))) +svint64_t svldnf1uh_s64(svbool_t, uint16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1uw_vnum_u64))) +svuint64_t svldnf1uw_vnum_u64(svbool_t, uint32_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1uw_vnum_s64))) +svint64_t svldnf1uw_vnum_s64(svbool_t, uint32_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1uw_u64))) +svuint64_t svldnf1uw_u64(svbool_t, uint32_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1uw_s64))) +svint64_t svldnf1uw_s64(svbool_t, uint32_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_u8))) +svuint8_t svldnt1_u8(svbool_t, uint8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_u32))) +svuint32_t svldnt1_u32(svbool_t, uint32_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_u64))) +svuint64_t svldnt1_u64(svbool_t, uint64_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_u16))) +svuint16_t svldnt1_u16(svbool_t, uint16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_s8))) +svint8_t svldnt1_s8(svbool_t, int8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_f64))) +svfloat64_t svldnt1_f64(svbool_t, float64_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_f32))) +svfloat32_t svldnt1_f32(svbool_t, float32_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_f16))) +svfloat16_t svldnt1_f16(svbool_t, float16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_s32))) +svint32_t svldnt1_s32(svbool_t, int32_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_s64))) +svint64_t svldnt1_s64(svbool_t, int64_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_s16))) +svint16_t svldnt1_s16(svbool_t, int16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_vnum_u8))) +svuint8_t svldnt1_vnum_u8(svbool_t, uint8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_vnum_u32))) +svuint32_t svldnt1_vnum_u32(svbool_t, uint32_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_vnum_u64))) +svuint64_t svldnt1_vnum_u64(svbool_t, uint64_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_vnum_u16))) +svuint16_t svldnt1_vnum_u16(svbool_t, uint16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_vnum_s8))) +svint8_t svldnt1_vnum_s8(svbool_t, int8_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_vnum_f64))) +svfloat64_t svldnt1_vnum_f64(svbool_t, float64_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_vnum_f32))) +svfloat32_t svldnt1_vnum_f32(svbool_t, float32_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_vnum_f16))) +svfloat16_t svldnt1_vnum_f16(svbool_t, float16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_vnum_s32))) +svint32_t svldnt1_vnum_s32(svbool_t, int32_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_vnum_s64))) +svint64_t svldnt1_vnum_s64(svbool_t, int64_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_vnum_s16))) +svint16_t svldnt1_vnum_s16(svbool_t, int16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlen_u8))) +uint64_t svlen_u8(svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlen_u32))) +uint64_t svlen_u32(svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlen_u64))) +uint64_t svlen_u64(svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlen_u16))) +uint64_t svlen_u16(svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlen_s8))) +uint64_t svlen_s8(svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlen_f64))) +uint64_t svlen_f64(svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlen_f32))) +uint64_t svlen_f32(svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlen_f16))) +uint64_t svlen_f16(svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlen_s32))) +uint64_t svlen_s32(svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlen_s64))) +uint64_t svlen_s64(svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlen_s16))) +uint64_t svlen_s16(svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_n_u8_m))) +svuint8_t svlsl_n_u8_m(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_n_u32_m))) +svuint32_t svlsl_n_u32_m(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_n_u64_m))) +svuint64_t svlsl_n_u64_m(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_n_u16_m))) +svuint16_t svlsl_n_u16_m(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_n_s8_m))) +svint8_t svlsl_n_s8_m(svbool_t, svint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_n_s32_m))) +svint32_t svlsl_n_s32_m(svbool_t, svint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_n_s64_m))) +svint64_t svlsl_n_s64_m(svbool_t, svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_n_s16_m))) +svint16_t svlsl_n_s16_m(svbool_t, svint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_n_u8_x))) +svuint8_t svlsl_n_u8_x(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_n_u32_x))) +svuint32_t svlsl_n_u32_x(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_n_u64_x))) +svuint64_t svlsl_n_u64_x(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_n_u16_x))) +svuint16_t svlsl_n_u16_x(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_n_s8_x))) +svint8_t svlsl_n_s8_x(svbool_t, svint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_n_s32_x))) +svint32_t svlsl_n_s32_x(svbool_t, svint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_n_s64_x))) +svint64_t svlsl_n_s64_x(svbool_t, svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_n_s16_x))) +svint16_t svlsl_n_s16_x(svbool_t, svint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_n_u8_z))) +svuint8_t svlsl_n_u8_z(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_n_u32_z))) +svuint32_t svlsl_n_u32_z(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_n_u64_z))) +svuint64_t svlsl_n_u64_z(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_n_u16_z))) +svuint16_t svlsl_n_u16_z(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_n_s8_z))) +svint8_t svlsl_n_s8_z(svbool_t, svint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_n_s32_z))) +svint32_t svlsl_n_s32_z(svbool_t, svint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_n_s64_z))) +svint64_t svlsl_n_s64_z(svbool_t, svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_n_s16_z))) +svint16_t svlsl_n_s16_z(svbool_t, svint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_u8_m))) +svuint8_t svlsl_u8_m(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_u32_m))) +svuint32_t svlsl_u32_m(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_u64_m))) +svuint64_t svlsl_u64_m(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_u16_m))) +svuint16_t svlsl_u16_m(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_s8_m))) +svint8_t svlsl_s8_m(svbool_t, svint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_s32_m))) +svint32_t svlsl_s32_m(svbool_t, svint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_s64_m))) +svint64_t svlsl_s64_m(svbool_t, svint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_s16_m))) +svint16_t svlsl_s16_m(svbool_t, svint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_u8_x))) +svuint8_t svlsl_u8_x(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_u32_x))) +svuint32_t svlsl_u32_x(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_u64_x))) +svuint64_t svlsl_u64_x(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_u16_x))) +svuint16_t svlsl_u16_x(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_s8_x))) +svint8_t svlsl_s8_x(svbool_t, svint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_s32_x))) +svint32_t svlsl_s32_x(svbool_t, svint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_s64_x))) +svint64_t svlsl_s64_x(svbool_t, svint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_s16_x))) +svint16_t svlsl_s16_x(svbool_t, svint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_u8_z))) +svuint8_t svlsl_u8_z(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_u32_z))) +svuint32_t svlsl_u32_z(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_u64_z))) +svuint64_t svlsl_u64_z(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_u16_z))) +svuint16_t svlsl_u16_z(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_s8_z))) +svint8_t svlsl_s8_z(svbool_t, svint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_s32_z))) +svint32_t svlsl_s32_z(svbool_t, svint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_s64_z))) +svint64_t svlsl_s64_z(svbool_t, svint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_s16_z))) +svint16_t svlsl_s16_z(svbool_t, svint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_wide_n_u8_m))) +svuint8_t svlsl_wide_n_u8_m(svbool_t, svuint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_wide_n_u32_m))) +svuint32_t svlsl_wide_n_u32_m(svbool_t, svuint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_wide_n_u16_m))) +svuint16_t svlsl_wide_n_u16_m(svbool_t, svuint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_wide_n_s8_m))) +svint8_t svlsl_wide_n_s8_m(svbool_t, svint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_wide_n_s32_m))) +svint32_t svlsl_wide_n_s32_m(svbool_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_wide_n_s16_m))) +svint16_t svlsl_wide_n_s16_m(svbool_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_wide_n_u8_x))) +svuint8_t svlsl_wide_n_u8_x(svbool_t, svuint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_wide_n_u32_x))) +svuint32_t svlsl_wide_n_u32_x(svbool_t, svuint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_wide_n_u16_x))) +svuint16_t svlsl_wide_n_u16_x(svbool_t, svuint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_wide_n_s8_x))) +svint8_t svlsl_wide_n_s8_x(svbool_t, svint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_wide_n_s32_x))) +svint32_t svlsl_wide_n_s32_x(svbool_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_wide_n_s16_x))) +svint16_t svlsl_wide_n_s16_x(svbool_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_wide_n_u8_z))) +svuint8_t svlsl_wide_n_u8_z(svbool_t, svuint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_wide_n_u32_z))) +svuint32_t svlsl_wide_n_u32_z(svbool_t, svuint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_wide_n_u16_z))) +svuint16_t svlsl_wide_n_u16_z(svbool_t, svuint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_wide_n_s8_z))) +svint8_t svlsl_wide_n_s8_z(svbool_t, svint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_wide_n_s32_z))) +svint32_t svlsl_wide_n_s32_z(svbool_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_wide_n_s16_z))) +svint16_t svlsl_wide_n_s16_z(svbool_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_wide_u8_m))) +svuint8_t svlsl_wide_u8_m(svbool_t, svuint8_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_wide_u32_m))) +svuint32_t svlsl_wide_u32_m(svbool_t, svuint32_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_wide_u16_m))) +svuint16_t svlsl_wide_u16_m(svbool_t, svuint16_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_wide_s8_m))) +svint8_t svlsl_wide_s8_m(svbool_t, svint8_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_wide_s32_m))) +svint32_t svlsl_wide_s32_m(svbool_t, svint32_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_wide_s16_m))) +svint16_t svlsl_wide_s16_m(svbool_t, svint16_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_wide_u8_x))) +svuint8_t svlsl_wide_u8_x(svbool_t, svuint8_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_wide_u32_x))) +svuint32_t svlsl_wide_u32_x(svbool_t, svuint32_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_wide_u16_x))) +svuint16_t svlsl_wide_u16_x(svbool_t, svuint16_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_wide_s8_x))) +svint8_t svlsl_wide_s8_x(svbool_t, svint8_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_wide_s32_x))) +svint32_t svlsl_wide_s32_x(svbool_t, svint32_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_wide_s16_x))) +svint16_t svlsl_wide_s16_x(svbool_t, svint16_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_wide_u8_z))) +svuint8_t svlsl_wide_u8_z(svbool_t, svuint8_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_wide_u32_z))) +svuint32_t svlsl_wide_u32_z(svbool_t, svuint32_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_wide_u16_z))) +svuint16_t svlsl_wide_u16_z(svbool_t, svuint16_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_wide_s8_z))) +svint8_t svlsl_wide_s8_z(svbool_t, svint8_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_wide_s32_z))) +svint32_t svlsl_wide_s32_z(svbool_t, svint32_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsl_wide_s16_z))) +svint16_t svlsl_wide_s16_z(svbool_t, svint16_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsr_n_u8_m))) +svuint8_t svlsr_n_u8_m(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsr_n_u32_m))) +svuint32_t svlsr_n_u32_m(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsr_n_u64_m))) +svuint64_t svlsr_n_u64_m(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsr_n_u16_m))) +svuint16_t svlsr_n_u16_m(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsr_n_u8_x))) +svuint8_t svlsr_n_u8_x(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsr_n_u32_x))) +svuint32_t svlsr_n_u32_x(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsr_n_u64_x))) +svuint64_t svlsr_n_u64_x(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsr_n_u16_x))) +svuint16_t svlsr_n_u16_x(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsr_n_u8_z))) +svuint8_t svlsr_n_u8_z(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsr_n_u32_z))) +svuint32_t svlsr_n_u32_z(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsr_n_u64_z))) +svuint64_t svlsr_n_u64_z(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsr_n_u16_z))) +svuint16_t svlsr_n_u16_z(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsr_u8_m))) +svuint8_t svlsr_u8_m(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsr_u32_m))) +svuint32_t svlsr_u32_m(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsr_u64_m))) +svuint64_t svlsr_u64_m(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsr_u16_m))) +svuint16_t svlsr_u16_m(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsr_u8_x))) +svuint8_t svlsr_u8_x(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsr_u32_x))) +svuint32_t svlsr_u32_x(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsr_u64_x))) +svuint64_t svlsr_u64_x(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsr_u16_x))) +svuint16_t svlsr_u16_x(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsr_u8_z))) +svuint8_t svlsr_u8_z(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsr_u32_z))) +svuint32_t svlsr_u32_z(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsr_u64_z))) +svuint64_t svlsr_u64_z(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsr_u16_z))) +svuint16_t svlsr_u16_z(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsr_wide_n_u8_m))) +svuint8_t svlsr_wide_n_u8_m(svbool_t, svuint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsr_wide_n_u32_m))) +svuint32_t svlsr_wide_n_u32_m(svbool_t, svuint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsr_wide_n_u16_m))) +svuint16_t svlsr_wide_n_u16_m(svbool_t, svuint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsr_wide_n_u8_x))) +svuint8_t svlsr_wide_n_u8_x(svbool_t, svuint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsr_wide_n_u32_x))) +svuint32_t svlsr_wide_n_u32_x(svbool_t, svuint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsr_wide_n_u16_x))) +svuint16_t svlsr_wide_n_u16_x(svbool_t, svuint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsr_wide_n_u8_z))) +svuint8_t svlsr_wide_n_u8_z(svbool_t, svuint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsr_wide_n_u32_z))) +svuint32_t svlsr_wide_n_u32_z(svbool_t, svuint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsr_wide_n_u16_z))) +svuint16_t svlsr_wide_n_u16_z(svbool_t, svuint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsr_wide_u8_m))) +svuint8_t svlsr_wide_u8_m(svbool_t, svuint8_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsr_wide_u32_m))) +svuint32_t svlsr_wide_u32_m(svbool_t, svuint32_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsr_wide_u16_m))) +svuint16_t svlsr_wide_u16_m(svbool_t, svuint16_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsr_wide_u8_x))) +svuint8_t svlsr_wide_u8_x(svbool_t, svuint8_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsr_wide_u32_x))) +svuint32_t svlsr_wide_u32_x(svbool_t, svuint32_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsr_wide_u16_x))) +svuint16_t svlsr_wide_u16_x(svbool_t, svuint16_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsr_wide_u8_z))) +svuint8_t svlsr_wide_u8_z(svbool_t, svuint8_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsr_wide_u32_z))) +svuint32_t svlsr_wide_u32_z(svbool_t, svuint32_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlsr_wide_u16_z))) +svuint16_t svlsr_wide_u16_z(svbool_t, svuint16_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_n_f64_m))) +svfloat64_t svmad_n_f64_m(svbool_t, svfloat64_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_n_f32_m))) +svfloat32_t svmad_n_f32_m(svbool_t, svfloat32_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_n_f16_m))) +svfloat16_t svmad_n_f16_m(svbool_t, svfloat16_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_n_f64_x))) +svfloat64_t svmad_n_f64_x(svbool_t, svfloat64_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_n_f32_x))) +svfloat32_t svmad_n_f32_x(svbool_t, svfloat32_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_n_f16_x))) +svfloat16_t svmad_n_f16_x(svbool_t, svfloat16_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_n_f64_z))) +svfloat64_t svmad_n_f64_z(svbool_t, svfloat64_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_n_f32_z))) +svfloat32_t svmad_n_f32_z(svbool_t, svfloat32_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_n_f16_z))) +svfloat16_t svmad_n_f16_z(svbool_t, svfloat16_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_n_u8_m))) +svuint8_t svmad_n_u8_m(svbool_t, svuint8_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_n_u32_m))) +svuint32_t svmad_n_u32_m(svbool_t, svuint32_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_n_u64_m))) +svuint64_t svmad_n_u64_m(svbool_t, svuint64_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_n_u16_m))) +svuint16_t svmad_n_u16_m(svbool_t, svuint16_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_n_s8_m))) +svint8_t svmad_n_s8_m(svbool_t, svint8_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_n_s32_m))) +svint32_t svmad_n_s32_m(svbool_t, svint32_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_n_s64_m))) +svint64_t svmad_n_s64_m(svbool_t, svint64_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_n_s16_m))) +svint16_t svmad_n_s16_m(svbool_t, svint16_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_n_u8_x))) +svuint8_t svmad_n_u8_x(svbool_t, svuint8_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_n_u32_x))) +svuint32_t svmad_n_u32_x(svbool_t, svuint32_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_n_u64_x))) +svuint64_t svmad_n_u64_x(svbool_t, svuint64_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_n_u16_x))) +svuint16_t svmad_n_u16_x(svbool_t, svuint16_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_n_s8_x))) +svint8_t svmad_n_s8_x(svbool_t, svint8_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_n_s32_x))) +svint32_t svmad_n_s32_x(svbool_t, svint32_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_n_s64_x))) +svint64_t svmad_n_s64_x(svbool_t, svint64_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_n_s16_x))) +svint16_t svmad_n_s16_x(svbool_t, svint16_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_n_u8_z))) +svuint8_t svmad_n_u8_z(svbool_t, svuint8_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_n_u32_z))) +svuint32_t svmad_n_u32_z(svbool_t, svuint32_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_n_u64_z))) +svuint64_t svmad_n_u64_z(svbool_t, svuint64_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_n_u16_z))) +svuint16_t svmad_n_u16_z(svbool_t, svuint16_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_n_s8_z))) +svint8_t svmad_n_s8_z(svbool_t, svint8_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_n_s32_z))) +svint32_t svmad_n_s32_z(svbool_t, svint32_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_n_s64_z))) +svint64_t svmad_n_s64_z(svbool_t, svint64_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_n_s16_z))) +svint16_t svmad_n_s16_z(svbool_t, svint16_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_f64_m))) +svfloat64_t svmad_f64_m(svbool_t, svfloat64_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_f32_m))) +svfloat32_t svmad_f32_m(svbool_t, svfloat32_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_f16_m))) +svfloat16_t svmad_f16_m(svbool_t, svfloat16_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_f64_x))) +svfloat64_t svmad_f64_x(svbool_t, svfloat64_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_f32_x))) +svfloat32_t svmad_f32_x(svbool_t, svfloat32_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_f16_x))) +svfloat16_t svmad_f16_x(svbool_t, svfloat16_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_f64_z))) +svfloat64_t svmad_f64_z(svbool_t, svfloat64_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_f32_z))) +svfloat32_t svmad_f32_z(svbool_t, svfloat32_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_f16_z))) +svfloat16_t svmad_f16_z(svbool_t, svfloat16_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_u8_m))) +svuint8_t svmad_u8_m(svbool_t, svuint8_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_u32_m))) +svuint32_t svmad_u32_m(svbool_t, svuint32_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_u64_m))) +svuint64_t svmad_u64_m(svbool_t, svuint64_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_u16_m))) +svuint16_t svmad_u16_m(svbool_t, svuint16_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_s8_m))) +svint8_t svmad_s8_m(svbool_t, svint8_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_s32_m))) +svint32_t svmad_s32_m(svbool_t, svint32_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_s64_m))) +svint64_t svmad_s64_m(svbool_t, svint64_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_s16_m))) +svint16_t svmad_s16_m(svbool_t, svint16_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_u8_x))) +svuint8_t svmad_u8_x(svbool_t, svuint8_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_u32_x))) +svuint32_t svmad_u32_x(svbool_t, svuint32_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_u64_x))) +svuint64_t svmad_u64_x(svbool_t, svuint64_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_u16_x))) +svuint16_t svmad_u16_x(svbool_t, svuint16_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_s8_x))) +svint8_t svmad_s8_x(svbool_t, svint8_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_s32_x))) +svint32_t svmad_s32_x(svbool_t, svint32_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_s64_x))) +svint64_t svmad_s64_x(svbool_t, svint64_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_s16_x))) +svint16_t svmad_s16_x(svbool_t, svint16_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_u8_z))) +svuint8_t svmad_u8_z(svbool_t, svuint8_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_u32_z))) +svuint32_t svmad_u32_z(svbool_t, svuint32_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_u64_z))) +svuint64_t svmad_u64_z(svbool_t, svuint64_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_u16_z))) +svuint16_t svmad_u16_z(svbool_t, svuint16_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_s8_z))) +svint8_t svmad_s8_z(svbool_t, svint8_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_s32_z))) +svint32_t svmad_s32_z(svbool_t, svint32_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_s64_z))) +svint64_t svmad_s64_z(svbool_t, svint64_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmad_s16_z))) +svint16_t svmad_s16_z(svbool_t, svint16_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_n_f64_m))) +svfloat64_t svmax_n_f64_m(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_n_f32_m))) +svfloat32_t svmax_n_f32_m(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_n_f16_m))) +svfloat16_t svmax_n_f16_m(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_n_f64_x))) +svfloat64_t svmax_n_f64_x(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_n_f32_x))) +svfloat32_t svmax_n_f32_x(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_n_f16_x))) +svfloat16_t svmax_n_f16_x(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_n_f64_z))) +svfloat64_t svmax_n_f64_z(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_n_f32_z))) +svfloat32_t svmax_n_f32_z(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_n_f16_z))) +svfloat16_t svmax_n_f16_z(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_n_s8_m))) +svint8_t svmax_n_s8_m(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_n_s32_m))) +svint32_t svmax_n_s32_m(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_n_s64_m))) +svint64_t svmax_n_s64_m(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_n_s16_m))) +svint16_t svmax_n_s16_m(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_n_s8_x))) +svint8_t svmax_n_s8_x(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_n_s32_x))) +svint32_t svmax_n_s32_x(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_n_s64_x))) +svint64_t svmax_n_s64_x(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_n_s16_x))) +svint16_t svmax_n_s16_x(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_n_s8_z))) +svint8_t svmax_n_s8_z(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_n_s32_z))) +svint32_t svmax_n_s32_z(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_n_s64_z))) +svint64_t svmax_n_s64_z(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_n_s16_z))) +svint16_t svmax_n_s16_z(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_n_u8_m))) +svuint8_t svmax_n_u8_m(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_n_u32_m))) +svuint32_t svmax_n_u32_m(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_n_u64_m))) +svuint64_t svmax_n_u64_m(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_n_u16_m))) +svuint16_t svmax_n_u16_m(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_n_u8_x))) +svuint8_t svmax_n_u8_x(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_n_u32_x))) +svuint32_t svmax_n_u32_x(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_n_u64_x))) +svuint64_t svmax_n_u64_x(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_n_u16_x))) +svuint16_t svmax_n_u16_x(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_n_u8_z))) +svuint8_t svmax_n_u8_z(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_n_u32_z))) +svuint32_t svmax_n_u32_z(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_n_u64_z))) +svuint64_t svmax_n_u64_z(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_n_u16_z))) +svuint16_t svmax_n_u16_z(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_f64_m))) +svfloat64_t svmax_f64_m(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_f32_m))) +svfloat32_t svmax_f32_m(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_f16_m))) +svfloat16_t svmax_f16_m(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_f64_x))) +svfloat64_t svmax_f64_x(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_f32_x))) +svfloat32_t svmax_f32_x(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_f16_x))) +svfloat16_t svmax_f16_x(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_f64_z))) +svfloat64_t svmax_f64_z(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_f32_z))) +svfloat32_t svmax_f32_z(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_f16_z))) +svfloat16_t svmax_f16_z(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_s8_m))) +svint8_t svmax_s8_m(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_s32_m))) +svint32_t svmax_s32_m(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_s64_m))) +svint64_t svmax_s64_m(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_s16_m))) +svint16_t svmax_s16_m(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_s8_x))) +svint8_t svmax_s8_x(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_s32_x))) +svint32_t svmax_s32_x(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_s64_x))) +svint64_t svmax_s64_x(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_s16_x))) +svint16_t svmax_s16_x(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_s8_z))) +svint8_t svmax_s8_z(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_s32_z))) +svint32_t svmax_s32_z(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_s64_z))) +svint64_t svmax_s64_z(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_s16_z))) +svint16_t svmax_s16_z(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_u8_m))) +svuint8_t svmax_u8_m(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_u32_m))) +svuint32_t svmax_u32_m(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_u64_m))) +svuint64_t svmax_u64_m(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_u16_m))) +svuint16_t svmax_u16_m(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_u8_x))) +svuint8_t svmax_u8_x(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_u32_x))) +svuint32_t svmax_u32_x(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_u64_x))) +svuint64_t svmax_u64_x(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_u16_x))) +svuint16_t svmax_u16_x(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_u8_z))) +svuint8_t svmax_u8_z(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_u32_z))) +svuint32_t svmax_u32_z(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_u64_z))) +svuint64_t svmax_u64_z(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmax_u16_z))) +svuint16_t svmax_u16_z(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxnm_n_f64_m))) +svfloat64_t svmaxnm_n_f64_m(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxnm_n_f32_m))) +svfloat32_t svmaxnm_n_f32_m(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxnm_n_f16_m))) +svfloat16_t svmaxnm_n_f16_m(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxnm_n_f64_x))) +svfloat64_t svmaxnm_n_f64_x(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxnm_n_f32_x))) +svfloat32_t svmaxnm_n_f32_x(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxnm_n_f16_x))) +svfloat16_t svmaxnm_n_f16_x(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxnm_n_f64_z))) +svfloat64_t svmaxnm_n_f64_z(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxnm_n_f32_z))) +svfloat32_t svmaxnm_n_f32_z(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxnm_n_f16_z))) +svfloat16_t svmaxnm_n_f16_z(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxnm_f64_m))) +svfloat64_t svmaxnm_f64_m(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxnm_f32_m))) +svfloat32_t svmaxnm_f32_m(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxnm_f16_m))) +svfloat16_t svmaxnm_f16_m(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxnm_f64_x))) +svfloat64_t svmaxnm_f64_x(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxnm_f32_x))) +svfloat32_t svmaxnm_f32_x(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxnm_f16_x))) +svfloat16_t svmaxnm_f16_x(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxnm_f64_z))) +svfloat64_t svmaxnm_f64_z(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxnm_f32_z))) +svfloat32_t svmaxnm_f32_z(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxnm_f16_z))) +svfloat16_t svmaxnm_f16_z(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxnmv_f64))) +float64_t svmaxnmv_f64(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxnmv_f32))) +float32_t svmaxnmv_f32(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxnmv_f16))) +float16_t svmaxnmv_f16(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxv_f64))) +float64_t svmaxv_f64(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxv_f32))) +float32_t svmaxv_f32(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxv_f16))) +float16_t svmaxv_f16(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxv_s8))) +int8_t svmaxv_s8(svbool_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxv_s32))) +int32_t svmaxv_s32(svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxv_s64))) +int64_t svmaxv_s64(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxv_s16))) +int16_t svmaxv_s16(svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxv_u8))) +uint8_t svmaxv_u8(svbool_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxv_u32))) +uint32_t svmaxv_u32(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxv_u64))) +uint64_t svmaxv_u64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxv_u16))) +uint16_t svmaxv_u16(svbool_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_n_f64_m))) +svfloat64_t svmin_n_f64_m(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_n_f32_m))) +svfloat32_t svmin_n_f32_m(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_n_f16_m))) +svfloat16_t svmin_n_f16_m(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_n_f64_x))) +svfloat64_t svmin_n_f64_x(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_n_f32_x))) +svfloat32_t svmin_n_f32_x(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_n_f16_x))) +svfloat16_t svmin_n_f16_x(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_n_f64_z))) +svfloat64_t svmin_n_f64_z(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_n_f32_z))) +svfloat32_t svmin_n_f32_z(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_n_f16_z))) +svfloat16_t svmin_n_f16_z(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_n_s8_m))) +svint8_t svmin_n_s8_m(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_n_s32_m))) +svint32_t svmin_n_s32_m(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_n_s64_m))) +svint64_t svmin_n_s64_m(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_n_s16_m))) +svint16_t svmin_n_s16_m(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_n_s8_x))) +svint8_t svmin_n_s8_x(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_n_s32_x))) +svint32_t svmin_n_s32_x(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_n_s64_x))) +svint64_t svmin_n_s64_x(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_n_s16_x))) +svint16_t svmin_n_s16_x(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_n_s8_z))) +svint8_t svmin_n_s8_z(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_n_s32_z))) +svint32_t svmin_n_s32_z(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_n_s64_z))) +svint64_t svmin_n_s64_z(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_n_s16_z))) +svint16_t svmin_n_s16_z(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_n_u8_m))) +svuint8_t svmin_n_u8_m(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_n_u32_m))) +svuint32_t svmin_n_u32_m(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_n_u64_m))) +svuint64_t svmin_n_u64_m(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_n_u16_m))) +svuint16_t svmin_n_u16_m(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_n_u8_x))) +svuint8_t svmin_n_u8_x(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_n_u32_x))) +svuint32_t svmin_n_u32_x(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_n_u64_x))) +svuint64_t svmin_n_u64_x(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_n_u16_x))) +svuint16_t svmin_n_u16_x(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_n_u8_z))) +svuint8_t svmin_n_u8_z(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_n_u32_z))) +svuint32_t svmin_n_u32_z(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_n_u64_z))) +svuint64_t svmin_n_u64_z(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_n_u16_z))) +svuint16_t svmin_n_u16_z(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_f64_m))) +svfloat64_t svmin_f64_m(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_f32_m))) +svfloat32_t svmin_f32_m(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_f16_m))) +svfloat16_t svmin_f16_m(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_f64_x))) +svfloat64_t svmin_f64_x(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_f32_x))) +svfloat32_t svmin_f32_x(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_f16_x))) +svfloat16_t svmin_f16_x(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_f64_z))) +svfloat64_t svmin_f64_z(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_f32_z))) +svfloat32_t svmin_f32_z(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_f16_z))) +svfloat16_t svmin_f16_z(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_s8_m))) +svint8_t svmin_s8_m(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_s32_m))) +svint32_t svmin_s32_m(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_s64_m))) +svint64_t svmin_s64_m(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_s16_m))) +svint16_t svmin_s16_m(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_s8_x))) +svint8_t svmin_s8_x(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_s32_x))) +svint32_t svmin_s32_x(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_s64_x))) +svint64_t svmin_s64_x(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_s16_x))) +svint16_t svmin_s16_x(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_s8_z))) +svint8_t svmin_s8_z(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_s32_z))) +svint32_t svmin_s32_z(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_s64_z))) +svint64_t svmin_s64_z(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_s16_z))) +svint16_t svmin_s16_z(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_u8_m))) +svuint8_t svmin_u8_m(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_u32_m))) +svuint32_t svmin_u32_m(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_u64_m))) +svuint64_t svmin_u64_m(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_u16_m))) +svuint16_t svmin_u16_m(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_u8_x))) +svuint8_t svmin_u8_x(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_u32_x))) +svuint32_t svmin_u32_x(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_u64_x))) +svuint64_t svmin_u64_x(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_u16_x))) +svuint16_t svmin_u16_x(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_u8_z))) +svuint8_t svmin_u8_z(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_u32_z))) +svuint32_t svmin_u32_z(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_u64_z))) +svuint64_t svmin_u64_z(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmin_u16_z))) +svuint16_t svmin_u16_z(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminnm_n_f64_m))) +svfloat64_t svminnm_n_f64_m(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminnm_n_f32_m))) +svfloat32_t svminnm_n_f32_m(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminnm_n_f16_m))) +svfloat16_t svminnm_n_f16_m(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminnm_n_f64_x))) +svfloat64_t svminnm_n_f64_x(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminnm_n_f32_x))) +svfloat32_t svminnm_n_f32_x(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminnm_n_f16_x))) +svfloat16_t svminnm_n_f16_x(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminnm_n_f64_z))) +svfloat64_t svminnm_n_f64_z(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminnm_n_f32_z))) +svfloat32_t svminnm_n_f32_z(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminnm_n_f16_z))) +svfloat16_t svminnm_n_f16_z(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminnm_f64_m))) +svfloat64_t svminnm_f64_m(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminnm_f32_m))) +svfloat32_t svminnm_f32_m(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminnm_f16_m))) +svfloat16_t svminnm_f16_m(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminnm_f64_x))) +svfloat64_t svminnm_f64_x(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminnm_f32_x))) +svfloat32_t svminnm_f32_x(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminnm_f16_x))) +svfloat16_t svminnm_f16_x(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminnm_f64_z))) +svfloat64_t svminnm_f64_z(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminnm_f32_z))) +svfloat32_t svminnm_f32_z(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminnm_f16_z))) +svfloat16_t svminnm_f16_z(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminnmv_f64))) +float64_t svminnmv_f64(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminnmv_f32))) +float32_t svminnmv_f32(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminnmv_f16))) +float16_t svminnmv_f16(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminv_f64))) +float64_t svminv_f64(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminv_f32))) +float32_t svminv_f32(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminv_f16))) +float16_t svminv_f16(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminv_s8))) +int8_t svminv_s8(svbool_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminv_s32))) +int32_t svminv_s32(svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminv_s64))) +int64_t svminv_s64(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminv_s16))) +int16_t svminv_s16(svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminv_u8))) +uint8_t svminv_u8(svbool_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminv_u32))) +uint32_t svminv_u32(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminv_u64))) +uint64_t svminv_u64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminv_u16))) +uint16_t svminv_u16(svbool_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_n_f64_m))) +svfloat64_t svmla_n_f64_m(svbool_t, svfloat64_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_n_f32_m))) +svfloat32_t svmla_n_f32_m(svbool_t, svfloat32_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_n_f16_m))) +svfloat16_t svmla_n_f16_m(svbool_t, svfloat16_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_n_f64_x))) +svfloat64_t svmla_n_f64_x(svbool_t, svfloat64_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_n_f32_x))) +svfloat32_t svmla_n_f32_x(svbool_t, svfloat32_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_n_f16_x))) +svfloat16_t svmla_n_f16_x(svbool_t, svfloat16_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_n_f64_z))) +svfloat64_t svmla_n_f64_z(svbool_t, svfloat64_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_n_f32_z))) +svfloat32_t svmla_n_f32_z(svbool_t, svfloat32_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_n_f16_z))) +svfloat16_t svmla_n_f16_z(svbool_t, svfloat16_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_n_u8_m))) +svuint8_t svmla_n_u8_m(svbool_t, svuint8_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_n_u32_m))) +svuint32_t svmla_n_u32_m(svbool_t, svuint32_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_n_u64_m))) +svuint64_t svmla_n_u64_m(svbool_t, svuint64_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_n_u16_m))) +svuint16_t svmla_n_u16_m(svbool_t, svuint16_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_n_s8_m))) +svint8_t svmla_n_s8_m(svbool_t, svint8_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_n_s32_m))) +svint32_t svmla_n_s32_m(svbool_t, svint32_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_n_s64_m))) +svint64_t svmla_n_s64_m(svbool_t, svint64_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_n_s16_m))) +svint16_t svmla_n_s16_m(svbool_t, svint16_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_n_u8_x))) +svuint8_t svmla_n_u8_x(svbool_t, svuint8_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_n_u32_x))) +svuint32_t svmla_n_u32_x(svbool_t, svuint32_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_n_u64_x))) +svuint64_t svmla_n_u64_x(svbool_t, svuint64_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_n_u16_x))) +svuint16_t svmla_n_u16_x(svbool_t, svuint16_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_n_s8_x))) +svint8_t svmla_n_s8_x(svbool_t, svint8_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_n_s32_x))) +svint32_t svmla_n_s32_x(svbool_t, svint32_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_n_s64_x))) +svint64_t svmla_n_s64_x(svbool_t, svint64_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_n_s16_x))) +svint16_t svmla_n_s16_x(svbool_t, svint16_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_n_u8_z))) +svuint8_t svmla_n_u8_z(svbool_t, svuint8_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_n_u32_z))) +svuint32_t svmla_n_u32_z(svbool_t, svuint32_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_n_u64_z))) +svuint64_t svmla_n_u64_z(svbool_t, svuint64_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_n_u16_z))) +svuint16_t svmla_n_u16_z(svbool_t, svuint16_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_n_s8_z))) +svint8_t svmla_n_s8_z(svbool_t, svint8_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_n_s32_z))) +svint32_t svmla_n_s32_z(svbool_t, svint32_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_n_s64_z))) +svint64_t svmla_n_s64_z(svbool_t, svint64_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_n_s16_z))) +svint16_t svmla_n_s16_z(svbool_t, svint16_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_f64_m))) +svfloat64_t svmla_f64_m(svbool_t, svfloat64_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_f32_m))) +svfloat32_t svmla_f32_m(svbool_t, svfloat32_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_f16_m))) +svfloat16_t svmla_f16_m(svbool_t, svfloat16_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_f64_x))) +svfloat64_t svmla_f64_x(svbool_t, svfloat64_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_f32_x))) +svfloat32_t svmla_f32_x(svbool_t, svfloat32_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_f16_x))) +svfloat16_t svmla_f16_x(svbool_t, svfloat16_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_f64_z))) +svfloat64_t svmla_f64_z(svbool_t, svfloat64_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_f32_z))) +svfloat32_t svmla_f32_z(svbool_t, svfloat32_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_f16_z))) +svfloat16_t svmla_f16_z(svbool_t, svfloat16_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_u8_m))) +svuint8_t svmla_u8_m(svbool_t, svuint8_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_u32_m))) +svuint32_t svmla_u32_m(svbool_t, svuint32_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_u64_m))) +svuint64_t svmla_u64_m(svbool_t, svuint64_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_u16_m))) +svuint16_t svmla_u16_m(svbool_t, svuint16_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_s8_m))) +svint8_t svmla_s8_m(svbool_t, svint8_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_s32_m))) +svint32_t svmla_s32_m(svbool_t, svint32_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_s64_m))) +svint64_t svmla_s64_m(svbool_t, svint64_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_s16_m))) +svint16_t svmla_s16_m(svbool_t, svint16_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_u8_x))) +svuint8_t svmla_u8_x(svbool_t, svuint8_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_u32_x))) +svuint32_t svmla_u32_x(svbool_t, svuint32_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_u64_x))) +svuint64_t svmla_u64_x(svbool_t, svuint64_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_u16_x))) +svuint16_t svmla_u16_x(svbool_t, svuint16_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_s8_x))) +svint8_t svmla_s8_x(svbool_t, svint8_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_s32_x))) +svint32_t svmla_s32_x(svbool_t, svint32_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_s64_x))) +svint64_t svmla_s64_x(svbool_t, svint64_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_s16_x))) +svint16_t svmla_s16_x(svbool_t, svint16_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_u8_z))) +svuint8_t svmla_u8_z(svbool_t, svuint8_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_u32_z))) +svuint32_t svmla_u32_z(svbool_t, svuint32_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_u64_z))) +svuint64_t svmla_u64_z(svbool_t, svuint64_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_u16_z))) +svuint16_t svmla_u16_z(svbool_t, svuint16_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_s8_z))) +svint8_t svmla_s8_z(svbool_t, svint8_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_s32_z))) +svint32_t svmla_s32_z(svbool_t, svint32_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_s64_z))) +svint64_t svmla_s64_z(svbool_t, svint64_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_s16_z))) +svint16_t svmla_s16_z(svbool_t, svint16_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_lane_f64))) +svfloat64_t svmla_lane_f64(svfloat64_t, svfloat64_t, svfloat64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_lane_f32))) +svfloat32_t svmla_lane_f32(svfloat32_t, svfloat32_t, svfloat32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_lane_f16))) +svfloat16_t svmla_lane_f16(svfloat16_t, svfloat16_t, svfloat16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_n_f64_m))) +svfloat64_t svmls_n_f64_m(svbool_t, svfloat64_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_n_f32_m))) +svfloat32_t svmls_n_f32_m(svbool_t, svfloat32_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_n_f16_m))) +svfloat16_t svmls_n_f16_m(svbool_t, svfloat16_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_n_f64_x))) +svfloat64_t svmls_n_f64_x(svbool_t, svfloat64_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_n_f32_x))) +svfloat32_t svmls_n_f32_x(svbool_t, svfloat32_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_n_f16_x))) +svfloat16_t svmls_n_f16_x(svbool_t, svfloat16_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_n_f64_z))) +svfloat64_t svmls_n_f64_z(svbool_t, svfloat64_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_n_f32_z))) +svfloat32_t svmls_n_f32_z(svbool_t, svfloat32_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_n_f16_z))) +svfloat16_t svmls_n_f16_z(svbool_t, svfloat16_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_n_u8_m))) +svuint8_t svmls_n_u8_m(svbool_t, svuint8_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_n_u32_m))) +svuint32_t svmls_n_u32_m(svbool_t, svuint32_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_n_u64_m))) +svuint64_t svmls_n_u64_m(svbool_t, svuint64_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_n_u16_m))) +svuint16_t svmls_n_u16_m(svbool_t, svuint16_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_n_s8_m))) +svint8_t svmls_n_s8_m(svbool_t, svint8_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_n_s32_m))) +svint32_t svmls_n_s32_m(svbool_t, svint32_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_n_s64_m))) +svint64_t svmls_n_s64_m(svbool_t, svint64_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_n_s16_m))) +svint16_t svmls_n_s16_m(svbool_t, svint16_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_n_u8_x))) +svuint8_t svmls_n_u8_x(svbool_t, svuint8_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_n_u32_x))) +svuint32_t svmls_n_u32_x(svbool_t, svuint32_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_n_u64_x))) +svuint64_t svmls_n_u64_x(svbool_t, svuint64_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_n_u16_x))) +svuint16_t svmls_n_u16_x(svbool_t, svuint16_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_n_s8_x))) +svint8_t svmls_n_s8_x(svbool_t, svint8_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_n_s32_x))) +svint32_t svmls_n_s32_x(svbool_t, svint32_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_n_s64_x))) +svint64_t svmls_n_s64_x(svbool_t, svint64_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_n_s16_x))) +svint16_t svmls_n_s16_x(svbool_t, svint16_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_n_u8_z))) +svuint8_t svmls_n_u8_z(svbool_t, svuint8_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_n_u32_z))) +svuint32_t svmls_n_u32_z(svbool_t, svuint32_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_n_u64_z))) +svuint64_t svmls_n_u64_z(svbool_t, svuint64_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_n_u16_z))) +svuint16_t svmls_n_u16_z(svbool_t, svuint16_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_n_s8_z))) +svint8_t svmls_n_s8_z(svbool_t, svint8_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_n_s32_z))) +svint32_t svmls_n_s32_z(svbool_t, svint32_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_n_s64_z))) +svint64_t svmls_n_s64_z(svbool_t, svint64_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_n_s16_z))) +svint16_t svmls_n_s16_z(svbool_t, svint16_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_f64_m))) +svfloat64_t svmls_f64_m(svbool_t, svfloat64_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_f32_m))) +svfloat32_t svmls_f32_m(svbool_t, svfloat32_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_f16_m))) +svfloat16_t svmls_f16_m(svbool_t, svfloat16_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_f64_x))) +svfloat64_t svmls_f64_x(svbool_t, svfloat64_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_f32_x))) +svfloat32_t svmls_f32_x(svbool_t, svfloat32_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_f16_x))) +svfloat16_t svmls_f16_x(svbool_t, svfloat16_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_f64_z))) +svfloat64_t svmls_f64_z(svbool_t, svfloat64_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_f32_z))) +svfloat32_t svmls_f32_z(svbool_t, svfloat32_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_f16_z))) +svfloat16_t svmls_f16_z(svbool_t, svfloat16_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_u8_m))) +svuint8_t svmls_u8_m(svbool_t, svuint8_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_u32_m))) +svuint32_t svmls_u32_m(svbool_t, svuint32_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_u64_m))) +svuint64_t svmls_u64_m(svbool_t, svuint64_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_u16_m))) +svuint16_t svmls_u16_m(svbool_t, svuint16_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_s8_m))) +svint8_t svmls_s8_m(svbool_t, svint8_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_s32_m))) +svint32_t svmls_s32_m(svbool_t, svint32_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_s64_m))) +svint64_t svmls_s64_m(svbool_t, svint64_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_s16_m))) +svint16_t svmls_s16_m(svbool_t, svint16_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_u8_x))) +svuint8_t svmls_u8_x(svbool_t, svuint8_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_u32_x))) +svuint32_t svmls_u32_x(svbool_t, svuint32_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_u64_x))) +svuint64_t svmls_u64_x(svbool_t, svuint64_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_u16_x))) +svuint16_t svmls_u16_x(svbool_t, svuint16_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_s8_x))) +svint8_t svmls_s8_x(svbool_t, svint8_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_s32_x))) +svint32_t svmls_s32_x(svbool_t, svint32_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_s64_x))) +svint64_t svmls_s64_x(svbool_t, svint64_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_s16_x))) +svint16_t svmls_s16_x(svbool_t, svint16_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_u8_z))) +svuint8_t svmls_u8_z(svbool_t, svuint8_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_u32_z))) +svuint32_t svmls_u32_z(svbool_t, svuint32_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_u64_z))) +svuint64_t svmls_u64_z(svbool_t, svuint64_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_u16_z))) +svuint16_t svmls_u16_z(svbool_t, svuint16_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_s8_z))) +svint8_t svmls_s8_z(svbool_t, svint8_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_s32_z))) +svint32_t svmls_s32_z(svbool_t, svint32_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_s64_z))) +svint64_t svmls_s64_z(svbool_t, svint64_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_s16_z))) +svint16_t svmls_s16_z(svbool_t, svint16_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_lane_f64))) +svfloat64_t svmls_lane_f64(svfloat64_t, svfloat64_t, svfloat64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_lane_f32))) +svfloat32_t svmls_lane_f32(svfloat32_t, svfloat32_t, svfloat32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_lane_f16))) +svfloat16_t svmls_lane_f16(svfloat16_t, svfloat16_t, svfloat16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmov_b_z))) +svbool_t svmov_b_z(svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_n_f64_m))) +svfloat64_t svmsb_n_f64_m(svbool_t, svfloat64_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_n_f32_m))) +svfloat32_t svmsb_n_f32_m(svbool_t, svfloat32_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_n_f16_m))) +svfloat16_t svmsb_n_f16_m(svbool_t, svfloat16_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_n_f64_x))) +svfloat64_t svmsb_n_f64_x(svbool_t, svfloat64_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_n_f32_x))) +svfloat32_t svmsb_n_f32_x(svbool_t, svfloat32_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_n_f16_x))) +svfloat16_t svmsb_n_f16_x(svbool_t, svfloat16_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_n_f64_z))) +svfloat64_t svmsb_n_f64_z(svbool_t, svfloat64_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_n_f32_z))) +svfloat32_t svmsb_n_f32_z(svbool_t, svfloat32_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_n_f16_z))) +svfloat16_t svmsb_n_f16_z(svbool_t, svfloat16_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_n_u8_m))) +svuint8_t svmsb_n_u8_m(svbool_t, svuint8_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_n_u32_m))) +svuint32_t svmsb_n_u32_m(svbool_t, svuint32_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_n_u64_m))) +svuint64_t svmsb_n_u64_m(svbool_t, svuint64_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_n_u16_m))) +svuint16_t svmsb_n_u16_m(svbool_t, svuint16_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_n_s8_m))) +svint8_t svmsb_n_s8_m(svbool_t, svint8_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_n_s32_m))) +svint32_t svmsb_n_s32_m(svbool_t, svint32_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_n_s64_m))) +svint64_t svmsb_n_s64_m(svbool_t, svint64_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_n_s16_m))) +svint16_t svmsb_n_s16_m(svbool_t, svint16_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_n_u8_x))) +svuint8_t svmsb_n_u8_x(svbool_t, svuint8_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_n_u32_x))) +svuint32_t svmsb_n_u32_x(svbool_t, svuint32_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_n_u64_x))) +svuint64_t svmsb_n_u64_x(svbool_t, svuint64_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_n_u16_x))) +svuint16_t svmsb_n_u16_x(svbool_t, svuint16_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_n_s8_x))) +svint8_t svmsb_n_s8_x(svbool_t, svint8_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_n_s32_x))) +svint32_t svmsb_n_s32_x(svbool_t, svint32_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_n_s64_x))) +svint64_t svmsb_n_s64_x(svbool_t, svint64_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_n_s16_x))) +svint16_t svmsb_n_s16_x(svbool_t, svint16_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_n_u8_z))) +svuint8_t svmsb_n_u8_z(svbool_t, svuint8_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_n_u32_z))) +svuint32_t svmsb_n_u32_z(svbool_t, svuint32_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_n_u64_z))) +svuint64_t svmsb_n_u64_z(svbool_t, svuint64_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_n_u16_z))) +svuint16_t svmsb_n_u16_z(svbool_t, svuint16_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_n_s8_z))) +svint8_t svmsb_n_s8_z(svbool_t, svint8_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_n_s32_z))) +svint32_t svmsb_n_s32_z(svbool_t, svint32_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_n_s64_z))) +svint64_t svmsb_n_s64_z(svbool_t, svint64_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_n_s16_z))) +svint16_t svmsb_n_s16_z(svbool_t, svint16_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_f64_m))) +svfloat64_t svmsb_f64_m(svbool_t, svfloat64_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_f32_m))) +svfloat32_t svmsb_f32_m(svbool_t, svfloat32_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_f16_m))) +svfloat16_t svmsb_f16_m(svbool_t, svfloat16_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_f64_x))) +svfloat64_t svmsb_f64_x(svbool_t, svfloat64_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_f32_x))) +svfloat32_t svmsb_f32_x(svbool_t, svfloat32_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_f16_x))) +svfloat16_t svmsb_f16_x(svbool_t, svfloat16_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_f64_z))) +svfloat64_t svmsb_f64_z(svbool_t, svfloat64_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_f32_z))) +svfloat32_t svmsb_f32_z(svbool_t, svfloat32_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_f16_z))) +svfloat16_t svmsb_f16_z(svbool_t, svfloat16_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_u8_m))) +svuint8_t svmsb_u8_m(svbool_t, svuint8_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_u32_m))) +svuint32_t svmsb_u32_m(svbool_t, svuint32_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_u64_m))) +svuint64_t svmsb_u64_m(svbool_t, svuint64_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_u16_m))) +svuint16_t svmsb_u16_m(svbool_t, svuint16_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_s8_m))) +svint8_t svmsb_s8_m(svbool_t, svint8_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_s32_m))) +svint32_t svmsb_s32_m(svbool_t, svint32_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_s64_m))) +svint64_t svmsb_s64_m(svbool_t, svint64_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_s16_m))) +svint16_t svmsb_s16_m(svbool_t, svint16_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_u8_x))) +svuint8_t svmsb_u8_x(svbool_t, svuint8_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_u32_x))) +svuint32_t svmsb_u32_x(svbool_t, svuint32_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_u64_x))) +svuint64_t svmsb_u64_x(svbool_t, svuint64_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_u16_x))) +svuint16_t svmsb_u16_x(svbool_t, svuint16_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_s8_x))) +svint8_t svmsb_s8_x(svbool_t, svint8_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_s32_x))) +svint32_t svmsb_s32_x(svbool_t, svint32_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_s64_x))) +svint64_t svmsb_s64_x(svbool_t, svint64_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_s16_x))) +svint16_t svmsb_s16_x(svbool_t, svint16_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_u8_z))) +svuint8_t svmsb_u8_z(svbool_t, svuint8_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_u32_z))) +svuint32_t svmsb_u32_z(svbool_t, svuint32_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_u64_z))) +svuint64_t svmsb_u64_z(svbool_t, svuint64_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_u16_z))) +svuint16_t svmsb_u16_z(svbool_t, svuint16_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_s8_z))) +svint8_t svmsb_s8_z(svbool_t, svint8_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_s32_z))) +svint32_t svmsb_s32_z(svbool_t, svint32_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_s64_z))) +svint64_t svmsb_s64_z(svbool_t, svint64_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmsb_s16_z))) +svint16_t svmsb_s16_z(svbool_t, svint16_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_n_f64_m))) +svfloat64_t svmul_n_f64_m(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_n_f32_m))) +svfloat32_t svmul_n_f32_m(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_n_f16_m))) +svfloat16_t svmul_n_f16_m(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_n_f64_x))) +svfloat64_t svmul_n_f64_x(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_n_f32_x))) +svfloat32_t svmul_n_f32_x(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_n_f16_x))) +svfloat16_t svmul_n_f16_x(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_n_f64_z))) +svfloat64_t svmul_n_f64_z(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_n_f32_z))) +svfloat32_t svmul_n_f32_z(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_n_f16_z))) +svfloat16_t svmul_n_f16_z(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_n_u8_m))) +svuint8_t svmul_n_u8_m(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_n_u32_m))) +svuint32_t svmul_n_u32_m(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_n_u64_m))) +svuint64_t svmul_n_u64_m(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_n_u16_m))) +svuint16_t svmul_n_u16_m(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_n_s8_m))) +svint8_t svmul_n_s8_m(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_n_s32_m))) +svint32_t svmul_n_s32_m(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_n_s64_m))) +svint64_t svmul_n_s64_m(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_n_s16_m))) +svint16_t svmul_n_s16_m(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_n_u8_x))) +svuint8_t svmul_n_u8_x(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_n_u32_x))) +svuint32_t svmul_n_u32_x(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_n_u64_x))) +svuint64_t svmul_n_u64_x(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_n_u16_x))) +svuint16_t svmul_n_u16_x(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_n_s8_x))) +svint8_t svmul_n_s8_x(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_n_s32_x))) +svint32_t svmul_n_s32_x(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_n_s64_x))) +svint64_t svmul_n_s64_x(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_n_s16_x))) +svint16_t svmul_n_s16_x(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_n_u8_z))) +svuint8_t svmul_n_u8_z(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_n_u32_z))) +svuint32_t svmul_n_u32_z(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_n_u64_z))) +svuint64_t svmul_n_u64_z(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_n_u16_z))) +svuint16_t svmul_n_u16_z(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_n_s8_z))) +svint8_t svmul_n_s8_z(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_n_s32_z))) +svint32_t svmul_n_s32_z(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_n_s64_z))) +svint64_t svmul_n_s64_z(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_n_s16_z))) +svint16_t svmul_n_s16_z(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_f64_m))) +svfloat64_t svmul_f64_m(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_f32_m))) +svfloat32_t svmul_f32_m(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_f16_m))) +svfloat16_t svmul_f16_m(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_f64_x))) +svfloat64_t svmul_f64_x(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_f32_x))) +svfloat32_t svmul_f32_x(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_f16_x))) +svfloat16_t svmul_f16_x(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_f64_z))) +svfloat64_t svmul_f64_z(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_f32_z))) +svfloat32_t svmul_f32_z(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_f16_z))) +svfloat16_t svmul_f16_z(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_u8_m))) +svuint8_t svmul_u8_m(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_u32_m))) +svuint32_t svmul_u32_m(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_u64_m))) +svuint64_t svmul_u64_m(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_u16_m))) +svuint16_t svmul_u16_m(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_s8_m))) +svint8_t svmul_s8_m(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_s32_m))) +svint32_t svmul_s32_m(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_s64_m))) +svint64_t svmul_s64_m(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_s16_m))) +svint16_t svmul_s16_m(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_u8_x))) +svuint8_t svmul_u8_x(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_u32_x))) +svuint32_t svmul_u32_x(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_u64_x))) +svuint64_t svmul_u64_x(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_u16_x))) +svuint16_t svmul_u16_x(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_s8_x))) +svint8_t svmul_s8_x(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_s32_x))) +svint32_t svmul_s32_x(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_s64_x))) +svint64_t svmul_s64_x(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_s16_x))) +svint16_t svmul_s16_x(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_u8_z))) +svuint8_t svmul_u8_z(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_u32_z))) +svuint32_t svmul_u32_z(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_u64_z))) +svuint64_t svmul_u64_z(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_u16_z))) +svuint16_t svmul_u16_z(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_s8_z))) +svint8_t svmul_s8_z(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_s32_z))) +svint32_t svmul_s32_z(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_s64_z))) +svint64_t svmul_s64_z(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_s16_z))) +svint16_t svmul_s16_z(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_lane_f64))) +svfloat64_t svmul_lane_f64(svfloat64_t, svfloat64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_lane_f32))) +svfloat32_t svmul_lane_f32(svfloat32_t, svfloat32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_lane_f16))) +svfloat16_t svmul_lane_f16(svfloat16_t, svfloat16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_n_s8_m))) +svint8_t svmulh_n_s8_m(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_n_s32_m))) +svint32_t svmulh_n_s32_m(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_n_s64_m))) +svint64_t svmulh_n_s64_m(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_n_s16_m))) +svint16_t svmulh_n_s16_m(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_n_s8_x))) +svint8_t svmulh_n_s8_x(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_n_s32_x))) +svint32_t svmulh_n_s32_x(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_n_s64_x))) +svint64_t svmulh_n_s64_x(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_n_s16_x))) +svint16_t svmulh_n_s16_x(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_n_s8_z))) +svint8_t svmulh_n_s8_z(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_n_s32_z))) +svint32_t svmulh_n_s32_z(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_n_s64_z))) +svint64_t svmulh_n_s64_z(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_n_s16_z))) +svint16_t svmulh_n_s16_z(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_n_u8_m))) +svuint8_t svmulh_n_u8_m(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_n_u32_m))) +svuint32_t svmulh_n_u32_m(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_n_u64_m))) +svuint64_t svmulh_n_u64_m(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_n_u16_m))) +svuint16_t svmulh_n_u16_m(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_n_u8_x))) +svuint8_t svmulh_n_u8_x(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_n_u32_x))) +svuint32_t svmulh_n_u32_x(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_n_u64_x))) +svuint64_t svmulh_n_u64_x(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_n_u16_x))) +svuint16_t svmulh_n_u16_x(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_n_u8_z))) +svuint8_t svmulh_n_u8_z(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_n_u32_z))) +svuint32_t svmulh_n_u32_z(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_n_u64_z))) +svuint64_t svmulh_n_u64_z(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_n_u16_z))) +svuint16_t svmulh_n_u16_z(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_s8_m))) +svint8_t svmulh_s8_m(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_s32_m))) +svint32_t svmulh_s32_m(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_s64_m))) +svint64_t svmulh_s64_m(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_s16_m))) +svint16_t svmulh_s16_m(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_s8_x))) +svint8_t svmulh_s8_x(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_s32_x))) +svint32_t svmulh_s32_x(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_s64_x))) +svint64_t svmulh_s64_x(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_s16_x))) +svint16_t svmulh_s16_x(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_s8_z))) +svint8_t svmulh_s8_z(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_s32_z))) +svint32_t svmulh_s32_z(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_s64_z))) +svint64_t svmulh_s64_z(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_s16_z))) +svint16_t svmulh_s16_z(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_u8_m))) +svuint8_t svmulh_u8_m(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_u32_m))) +svuint32_t svmulh_u32_m(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_u64_m))) +svuint64_t svmulh_u64_m(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_u16_m))) +svuint16_t svmulh_u16_m(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_u8_x))) +svuint8_t svmulh_u8_x(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_u32_x))) +svuint32_t svmulh_u32_x(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_u64_x))) +svuint64_t svmulh_u64_x(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_u16_x))) +svuint16_t svmulh_u16_x(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_u8_z))) +svuint8_t svmulh_u8_z(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_u32_z))) +svuint32_t svmulh_u32_z(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_u64_z))) +svuint64_t svmulh_u64_z(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulh_u16_z))) +svuint16_t svmulh_u16_z(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulx_n_f64_m))) +svfloat64_t svmulx_n_f64_m(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulx_n_f32_m))) +svfloat32_t svmulx_n_f32_m(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulx_n_f16_m))) +svfloat16_t svmulx_n_f16_m(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulx_n_f64_x))) +svfloat64_t svmulx_n_f64_x(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulx_n_f32_x))) +svfloat32_t svmulx_n_f32_x(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulx_n_f16_x))) +svfloat16_t svmulx_n_f16_x(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulx_n_f64_z))) +svfloat64_t svmulx_n_f64_z(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulx_n_f32_z))) +svfloat32_t svmulx_n_f32_z(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulx_n_f16_z))) +svfloat16_t svmulx_n_f16_z(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulx_f64_m))) +svfloat64_t svmulx_f64_m(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulx_f32_m))) +svfloat32_t svmulx_f32_m(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulx_f16_m))) +svfloat16_t svmulx_f16_m(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulx_f64_x))) +svfloat64_t svmulx_f64_x(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulx_f32_x))) +svfloat32_t svmulx_f32_x(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulx_f16_x))) +svfloat16_t svmulx_f16_x(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulx_f64_z))) +svfloat64_t svmulx_f64_z(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulx_f32_z))) +svfloat32_t svmulx_f32_z(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmulx_f16_z))) +svfloat16_t svmulx_f16_z(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnand_b_z))) +svbool_t svnand_b_z(svbool_t, svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svneg_f64_m))) +svfloat64_t svneg_f64_m(svfloat64_t, svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svneg_f32_m))) +svfloat32_t svneg_f32_m(svfloat32_t, svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svneg_f16_m))) +svfloat16_t svneg_f16_m(svfloat16_t, svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svneg_f64_x))) +svfloat64_t svneg_f64_x(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svneg_f32_x))) +svfloat32_t svneg_f32_x(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svneg_f16_x))) +svfloat16_t svneg_f16_x(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svneg_f64_z))) +svfloat64_t svneg_f64_z(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svneg_f32_z))) +svfloat32_t svneg_f32_z(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svneg_f16_z))) +svfloat16_t svneg_f16_z(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svneg_s8_m))) +svint8_t svneg_s8_m(svint8_t, svbool_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svneg_s32_m))) +svint32_t svneg_s32_m(svint32_t, svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svneg_s64_m))) +svint64_t svneg_s64_m(svint64_t, svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svneg_s16_m))) +svint16_t svneg_s16_m(svint16_t, svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svneg_s8_x))) +svint8_t svneg_s8_x(svbool_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svneg_s32_x))) +svint32_t svneg_s32_x(svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svneg_s64_x))) +svint64_t svneg_s64_x(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svneg_s16_x))) +svint16_t svneg_s16_x(svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svneg_s8_z))) +svint8_t svneg_s8_z(svbool_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svneg_s32_z))) +svint32_t svneg_s32_z(svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svneg_s64_z))) +svint64_t svneg_s64_z(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svneg_s16_z))) +svint16_t svneg_s16_z(svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmad_n_f64_m))) +svfloat64_t svnmad_n_f64_m(svbool_t, svfloat64_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmad_n_f32_m))) +svfloat32_t svnmad_n_f32_m(svbool_t, svfloat32_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmad_n_f16_m))) +svfloat16_t svnmad_n_f16_m(svbool_t, svfloat16_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmad_n_f64_x))) +svfloat64_t svnmad_n_f64_x(svbool_t, svfloat64_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmad_n_f32_x))) +svfloat32_t svnmad_n_f32_x(svbool_t, svfloat32_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmad_n_f16_x))) +svfloat16_t svnmad_n_f16_x(svbool_t, svfloat16_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmad_n_f64_z))) +svfloat64_t svnmad_n_f64_z(svbool_t, svfloat64_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmad_n_f32_z))) +svfloat32_t svnmad_n_f32_z(svbool_t, svfloat32_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmad_n_f16_z))) +svfloat16_t svnmad_n_f16_z(svbool_t, svfloat16_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmad_f64_m))) +svfloat64_t svnmad_f64_m(svbool_t, svfloat64_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmad_f32_m))) +svfloat32_t svnmad_f32_m(svbool_t, svfloat32_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmad_f16_m))) +svfloat16_t svnmad_f16_m(svbool_t, svfloat16_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmad_f64_x))) +svfloat64_t svnmad_f64_x(svbool_t, svfloat64_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmad_f32_x))) +svfloat32_t svnmad_f32_x(svbool_t, svfloat32_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmad_f16_x))) +svfloat16_t svnmad_f16_x(svbool_t, svfloat16_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmad_f64_z))) +svfloat64_t svnmad_f64_z(svbool_t, svfloat64_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmad_f32_z))) +svfloat32_t svnmad_f32_z(svbool_t, svfloat32_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmad_f16_z))) +svfloat16_t svnmad_f16_z(svbool_t, svfloat16_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmla_n_f64_m))) +svfloat64_t svnmla_n_f64_m(svbool_t, svfloat64_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmla_n_f32_m))) +svfloat32_t svnmla_n_f32_m(svbool_t, svfloat32_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmla_n_f16_m))) +svfloat16_t svnmla_n_f16_m(svbool_t, svfloat16_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmla_n_f64_x))) +svfloat64_t svnmla_n_f64_x(svbool_t, svfloat64_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmla_n_f32_x))) +svfloat32_t svnmla_n_f32_x(svbool_t, svfloat32_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmla_n_f16_x))) +svfloat16_t svnmla_n_f16_x(svbool_t, svfloat16_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmla_n_f64_z))) +svfloat64_t svnmla_n_f64_z(svbool_t, svfloat64_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmla_n_f32_z))) +svfloat32_t svnmla_n_f32_z(svbool_t, svfloat32_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmla_n_f16_z))) +svfloat16_t svnmla_n_f16_z(svbool_t, svfloat16_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmla_f64_m))) +svfloat64_t svnmla_f64_m(svbool_t, svfloat64_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmla_f32_m))) +svfloat32_t svnmla_f32_m(svbool_t, svfloat32_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmla_f16_m))) +svfloat16_t svnmla_f16_m(svbool_t, svfloat16_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmla_f64_x))) +svfloat64_t svnmla_f64_x(svbool_t, svfloat64_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmla_f32_x))) +svfloat32_t svnmla_f32_x(svbool_t, svfloat32_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmla_f16_x))) +svfloat16_t svnmla_f16_x(svbool_t, svfloat16_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmla_f64_z))) +svfloat64_t svnmla_f64_z(svbool_t, svfloat64_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmla_f32_z))) +svfloat32_t svnmla_f32_z(svbool_t, svfloat32_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmla_f16_z))) +svfloat16_t svnmla_f16_z(svbool_t, svfloat16_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmls_n_f64_m))) +svfloat64_t svnmls_n_f64_m(svbool_t, svfloat64_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmls_n_f32_m))) +svfloat32_t svnmls_n_f32_m(svbool_t, svfloat32_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmls_n_f16_m))) +svfloat16_t svnmls_n_f16_m(svbool_t, svfloat16_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmls_n_f64_x))) +svfloat64_t svnmls_n_f64_x(svbool_t, svfloat64_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmls_n_f32_x))) +svfloat32_t svnmls_n_f32_x(svbool_t, svfloat32_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmls_n_f16_x))) +svfloat16_t svnmls_n_f16_x(svbool_t, svfloat16_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmls_n_f64_z))) +svfloat64_t svnmls_n_f64_z(svbool_t, svfloat64_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmls_n_f32_z))) +svfloat32_t svnmls_n_f32_z(svbool_t, svfloat32_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmls_n_f16_z))) +svfloat16_t svnmls_n_f16_z(svbool_t, svfloat16_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmls_f64_m))) +svfloat64_t svnmls_f64_m(svbool_t, svfloat64_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmls_f32_m))) +svfloat32_t svnmls_f32_m(svbool_t, svfloat32_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmls_f16_m))) +svfloat16_t svnmls_f16_m(svbool_t, svfloat16_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmls_f64_x))) +svfloat64_t svnmls_f64_x(svbool_t, svfloat64_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmls_f32_x))) +svfloat32_t svnmls_f32_x(svbool_t, svfloat32_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmls_f16_x))) +svfloat16_t svnmls_f16_x(svbool_t, svfloat16_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmls_f64_z))) +svfloat64_t svnmls_f64_z(svbool_t, svfloat64_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmls_f32_z))) +svfloat32_t svnmls_f32_z(svbool_t, svfloat32_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmls_f16_z))) +svfloat16_t svnmls_f16_z(svbool_t, svfloat16_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmsb_n_f64_m))) +svfloat64_t svnmsb_n_f64_m(svbool_t, svfloat64_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmsb_n_f32_m))) +svfloat32_t svnmsb_n_f32_m(svbool_t, svfloat32_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmsb_n_f16_m))) +svfloat16_t svnmsb_n_f16_m(svbool_t, svfloat16_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmsb_n_f64_x))) +svfloat64_t svnmsb_n_f64_x(svbool_t, svfloat64_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmsb_n_f32_x))) +svfloat32_t svnmsb_n_f32_x(svbool_t, svfloat32_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmsb_n_f16_x))) +svfloat16_t svnmsb_n_f16_x(svbool_t, svfloat16_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmsb_n_f64_z))) +svfloat64_t svnmsb_n_f64_z(svbool_t, svfloat64_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmsb_n_f32_z))) +svfloat32_t svnmsb_n_f32_z(svbool_t, svfloat32_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmsb_n_f16_z))) +svfloat16_t svnmsb_n_f16_z(svbool_t, svfloat16_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmsb_f64_m))) +svfloat64_t svnmsb_f64_m(svbool_t, svfloat64_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmsb_f32_m))) +svfloat32_t svnmsb_f32_m(svbool_t, svfloat32_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmsb_f16_m))) +svfloat16_t svnmsb_f16_m(svbool_t, svfloat16_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmsb_f64_x))) +svfloat64_t svnmsb_f64_x(svbool_t, svfloat64_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmsb_f32_x))) +svfloat32_t svnmsb_f32_x(svbool_t, svfloat32_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmsb_f16_x))) +svfloat16_t svnmsb_f16_x(svbool_t, svfloat16_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmsb_f64_z))) +svfloat64_t svnmsb_f64_z(svbool_t, svfloat64_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmsb_f32_z))) +svfloat32_t svnmsb_f32_z(svbool_t, svfloat32_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmsb_f16_z))) +svfloat16_t svnmsb_f16_z(svbool_t, svfloat16_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnor_b_z))) +svbool_t svnor_b_z(svbool_t, svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnot_b_z))) +svbool_t svnot_b_z(svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnot_u8_m))) +svuint8_t svnot_u8_m(svuint8_t, svbool_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnot_u32_m))) +svuint32_t svnot_u32_m(svuint32_t, svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnot_u64_m))) +svuint64_t svnot_u64_m(svuint64_t, svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnot_u16_m))) +svuint16_t svnot_u16_m(svuint16_t, svbool_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnot_s8_m))) +svint8_t svnot_s8_m(svint8_t, svbool_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnot_s32_m))) +svint32_t svnot_s32_m(svint32_t, svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnot_s64_m))) +svint64_t svnot_s64_m(svint64_t, svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnot_s16_m))) +svint16_t svnot_s16_m(svint16_t, svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnot_u8_x))) +svuint8_t svnot_u8_x(svbool_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnot_u32_x))) +svuint32_t svnot_u32_x(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnot_u64_x))) +svuint64_t svnot_u64_x(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnot_u16_x))) +svuint16_t svnot_u16_x(svbool_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnot_s8_x))) +svint8_t svnot_s8_x(svbool_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnot_s32_x))) +svint32_t svnot_s32_x(svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnot_s64_x))) +svint64_t svnot_s64_x(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnot_s16_x))) +svint16_t svnot_s16_x(svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnot_u8_z))) +svuint8_t svnot_u8_z(svbool_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnot_u32_z))) +svuint32_t svnot_u32_z(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnot_u64_z))) +svuint64_t svnot_u64_z(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnot_u16_z))) +svuint16_t svnot_u16_z(svbool_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnot_s8_z))) +svint8_t svnot_s8_z(svbool_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnot_s32_z))) +svint32_t svnot_s32_z(svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnot_s64_z))) +svint64_t svnot_s64_z(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnot_s16_z))) +svint16_t svnot_s16_z(svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorn_b_z))) +svbool_t svorn_b_z(svbool_t, svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_b_z))) +svbool_t svorr_b_z(svbool_t, svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_n_u8_m))) +svuint8_t svorr_n_u8_m(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_n_u32_m))) +svuint32_t svorr_n_u32_m(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_n_u64_m))) +svuint64_t svorr_n_u64_m(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_n_u16_m))) +svuint16_t svorr_n_u16_m(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_n_s8_m))) +svint8_t svorr_n_s8_m(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_n_s32_m))) +svint32_t svorr_n_s32_m(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_n_s64_m))) +svint64_t svorr_n_s64_m(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_n_s16_m))) +svint16_t svorr_n_s16_m(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_n_u8_x))) +svuint8_t svorr_n_u8_x(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_n_u32_x))) +svuint32_t svorr_n_u32_x(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_n_u64_x))) +svuint64_t svorr_n_u64_x(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_n_u16_x))) +svuint16_t svorr_n_u16_x(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_n_s8_x))) +svint8_t svorr_n_s8_x(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_n_s32_x))) +svint32_t svorr_n_s32_x(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_n_s64_x))) +svint64_t svorr_n_s64_x(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_n_s16_x))) +svint16_t svorr_n_s16_x(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_n_u8_z))) +svuint8_t svorr_n_u8_z(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_n_u32_z))) +svuint32_t svorr_n_u32_z(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_n_u64_z))) +svuint64_t svorr_n_u64_z(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_n_u16_z))) +svuint16_t svorr_n_u16_z(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_n_s8_z))) +svint8_t svorr_n_s8_z(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_n_s32_z))) +svint32_t svorr_n_s32_z(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_n_s64_z))) +svint64_t svorr_n_s64_z(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_n_s16_z))) +svint16_t svorr_n_s16_z(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_u8_m))) +svuint8_t svorr_u8_m(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_u32_m))) +svuint32_t svorr_u32_m(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_u64_m))) +svuint64_t svorr_u64_m(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_u16_m))) +svuint16_t svorr_u16_m(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_s8_m))) +svint8_t svorr_s8_m(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_s32_m))) +svint32_t svorr_s32_m(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_s64_m))) +svint64_t svorr_s64_m(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_s16_m))) +svint16_t svorr_s16_m(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_u8_x))) +svuint8_t svorr_u8_x(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_u32_x))) +svuint32_t svorr_u32_x(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_u64_x))) +svuint64_t svorr_u64_x(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_u16_x))) +svuint16_t svorr_u16_x(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_s8_x))) +svint8_t svorr_s8_x(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_s32_x))) +svint32_t svorr_s32_x(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_s64_x))) +svint64_t svorr_s64_x(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_s16_x))) +svint16_t svorr_s16_x(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_u8_z))) +svuint8_t svorr_u8_z(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_u32_z))) +svuint32_t svorr_u32_z(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_u64_z))) +svuint64_t svorr_u64_z(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_u16_z))) +svuint16_t svorr_u16_z(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_s8_z))) +svint8_t svorr_s8_z(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_s32_z))) +svint32_t svorr_s32_z(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_s64_z))) +svint64_t svorr_s64_z(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorr_s16_z))) +svint16_t svorr_s16_z(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorv_u8))) +uint8_t svorv_u8(svbool_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorv_u32))) +uint32_t svorv_u32(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorv_u64))) +uint64_t svorv_u64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorv_u16))) +uint16_t svorv_u16(svbool_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorv_s8))) +int8_t svorv_s8(svbool_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorv_s32))) +int32_t svorv_s32(svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorv_s64))) +int64_t svorv_s64(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorv_s16))) +int16_t svorv_s16(svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svpfalse_b))) +svbool_t svpfalse_b(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svpfirst_b))) +svbool_t svpfirst_b(svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svpnext_b8))) +svbool_t svpnext_b8(svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svpnext_b32))) +svbool_t svpnext_b32(svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svpnext_b64))) +svbool_t svpnext_b64(svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svpnext_b16))) +svbool_t svpnext_b16(svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svprfb))) +void svprfb(svbool_t, void const *, enum svprfop); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svprfb_gather_u32base))) +void svprfb_gather_u32base(svbool_t, svuint32_t, enum svprfop); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svprfb_gather_u64base))) +void svprfb_gather_u64base(svbool_t, svuint64_t, enum svprfop); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svprfb_gather_u32base_offset))) +void svprfb_gather_u32base_offset(svbool_t, svuint32_t, int64_t, enum svprfop); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svprfb_gather_u64base_offset))) +void svprfb_gather_u64base_offset(svbool_t, svuint64_t, int64_t, enum svprfop); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svprfb_gather_s32offset))) +void svprfb_gather_s32offset(svbool_t, void const *, svint32_t, enum svprfop); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svprfb_gather_u32offset))) +void svprfb_gather_u32offset(svbool_t, void const *, svuint32_t, enum svprfop); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svprfb_gather_s64offset))) +void svprfb_gather_s64offset(svbool_t, void const *, svint64_t, enum svprfop); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svprfb_gather_u64offset))) +void svprfb_gather_u64offset(svbool_t, void const *, svuint64_t, enum svprfop); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svprfb_vnum))) +void svprfb_vnum(svbool_t, void const *, int64_t, enum svprfop); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svprfd))) +void svprfd(svbool_t, void const *, enum svprfop); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svprfd_gather_u32base))) +void svprfd_gather_u32base(svbool_t, svuint32_t, enum svprfop); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svprfd_gather_u64base))) +void svprfd_gather_u64base(svbool_t, svuint64_t, enum svprfop); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svprfd_gather_u32base_index))) +void svprfd_gather_u32base_index(svbool_t, svuint32_t, int64_t, enum svprfop); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svprfd_gather_u64base_index))) +void svprfd_gather_u64base_index(svbool_t, svuint64_t, int64_t, enum svprfop); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svprfd_gather_s32index))) +void svprfd_gather_s32index(svbool_t, void const *, svint32_t, enum svprfop); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svprfd_gather_u32index))) +void svprfd_gather_u32index(svbool_t, void const *, svuint32_t, enum svprfop); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svprfd_gather_s64index))) +void svprfd_gather_s64index(svbool_t, void const *, svint64_t, enum svprfop); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svprfd_gather_u64index))) +void svprfd_gather_u64index(svbool_t, void const *, svuint64_t, enum svprfop); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svprfd_vnum))) +void svprfd_vnum(svbool_t, void const *, int64_t, enum svprfop); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svprfh))) +void svprfh(svbool_t, void const *, enum svprfop); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svprfh_gather_u32base))) +void svprfh_gather_u32base(svbool_t, svuint32_t, enum svprfop); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svprfh_gather_u64base))) +void svprfh_gather_u64base(svbool_t, svuint64_t, enum svprfop); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svprfh_gather_u32base_index))) +void svprfh_gather_u32base_index(svbool_t, svuint32_t, int64_t, enum svprfop); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svprfh_gather_u64base_index))) +void svprfh_gather_u64base_index(svbool_t, svuint64_t, int64_t, enum svprfop); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svprfh_gather_s32index))) +void svprfh_gather_s32index(svbool_t, void const *, svint32_t, enum svprfop); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svprfh_gather_u32index))) +void svprfh_gather_u32index(svbool_t, void const *, svuint32_t, enum svprfop); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svprfh_gather_s64index))) +void svprfh_gather_s64index(svbool_t, void const *, svint64_t, enum svprfop); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svprfh_gather_u64index))) +void svprfh_gather_u64index(svbool_t, void const *, svuint64_t, enum svprfop); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svprfh_vnum))) +void svprfh_vnum(svbool_t, void const *, int64_t, enum svprfop); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svprfw))) +void svprfw(svbool_t, void const *, enum svprfop); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svprfw_gather_u32base))) +void svprfw_gather_u32base(svbool_t, svuint32_t, enum svprfop); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svprfw_gather_u64base))) +void svprfw_gather_u64base(svbool_t, svuint64_t, enum svprfop); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svprfw_gather_u32base_index))) +void svprfw_gather_u32base_index(svbool_t, svuint32_t, int64_t, enum svprfop); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svprfw_gather_u64base_index))) +void svprfw_gather_u64base_index(svbool_t, svuint64_t, int64_t, enum svprfop); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svprfw_gather_s32index))) +void svprfw_gather_s32index(svbool_t, void const *, svint32_t, enum svprfop); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svprfw_gather_u32index))) +void svprfw_gather_u32index(svbool_t, void const *, svuint32_t, enum svprfop); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svprfw_gather_s64index))) +void svprfw_gather_s64index(svbool_t, void const *, svint64_t, enum svprfop); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svprfw_gather_u64index))) +void svprfw_gather_u64index(svbool_t, void const *, svuint64_t, enum svprfop); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svprfw_vnum))) +void svprfw_vnum(svbool_t, void const *, int64_t, enum svprfop); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svptest_any))) +bool svptest_any(svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svptest_first))) +bool svptest_first(svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svptest_last))) +bool svptest_last(svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svptrue_pat_b8))) +svbool_t svptrue_pat_b8(enum svpattern); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svptrue_pat_b32))) +svbool_t svptrue_pat_b32(enum svpattern); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svptrue_pat_b64))) +svbool_t svptrue_pat_b64(enum svpattern); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svptrue_pat_b16))) +svbool_t svptrue_pat_b16(enum svpattern); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svptrue_b8))) +svbool_t svptrue_b8(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svptrue_b32))) +svbool_t svptrue_b32(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svptrue_b64))) +svbool_t svptrue_b64(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svptrue_b16))) +svbool_t svptrue_b16(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_n_s8))) +svint8_t svqadd_n_s8(svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_n_s32))) +svint32_t svqadd_n_s32(svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_n_s64))) +svint64_t svqadd_n_s64(svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_n_s16))) +svint16_t svqadd_n_s16(svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_n_u8))) +svuint8_t svqadd_n_u8(svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_n_u32))) +svuint32_t svqadd_n_u32(svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_n_u64))) +svuint64_t svqadd_n_u64(svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_n_u16))) +svuint16_t svqadd_n_u16(svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_s8))) +svint8_t svqadd_s8(svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_s32))) +svint32_t svqadd_s32(svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_s64))) +svint64_t svqadd_s64(svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_s16))) +svint16_t svqadd_s16(svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_u8))) +svuint8_t svqadd_u8(svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_u32))) +svuint32_t svqadd_u32(svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_u64))) +svuint64_t svqadd_u64(svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_u16))) +svuint16_t svqadd_u16(svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecb_n_s32))) +int32_t svqdecb_n_s32(int32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecb_n_s64))) +int64_t svqdecb_n_s64(int64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecb_n_u32))) +uint32_t svqdecb_n_u32(uint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecb_n_u64))) +uint64_t svqdecb_n_u64(uint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecb_pat_n_s32))) +int32_t svqdecb_pat_n_s32(int32_t, enum svpattern, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecb_pat_n_s64))) +int64_t svqdecb_pat_n_s64(int64_t, enum svpattern, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecb_pat_n_u32))) +uint32_t svqdecb_pat_n_u32(uint32_t, enum svpattern, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecb_pat_n_u64))) +uint64_t svqdecb_pat_n_u64(uint64_t, enum svpattern, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecd_n_s32))) +int32_t svqdecd_n_s32(int32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecd_n_s64))) +int64_t svqdecd_n_s64(int64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecd_n_u32))) +uint32_t svqdecd_n_u32(uint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecd_n_u64))) +uint64_t svqdecd_n_u64(uint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecd_s64))) +svint64_t svqdecd_s64(svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecd_u64))) +svuint64_t svqdecd_u64(svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecd_pat_n_s32))) +int32_t svqdecd_pat_n_s32(int32_t, enum svpattern, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecd_pat_n_s64))) +int64_t svqdecd_pat_n_s64(int64_t, enum svpattern, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecd_pat_n_u32))) +uint32_t svqdecd_pat_n_u32(uint32_t, enum svpattern, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecd_pat_n_u64))) +uint64_t svqdecd_pat_n_u64(uint64_t, enum svpattern, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecd_pat_s64))) +svint64_t svqdecd_pat_s64(svint64_t, enum svpattern, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecd_pat_u64))) +svuint64_t svqdecd_pat_u64(svuint64_t, enum svpattern, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdech_n_s32))) +int32_t svqdech_n_s32(int32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdech_n_s64))) +int64_t svqdech_n_s64(int64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdech_n_u32))) +uint32_t svqdech_n_u32(uint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdech_n_u64))) +uint64_t svqdech_n_u64(uint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdech_s16))) +svint16_t svqdech_s16(svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdech_u16))) +svuint16_t svqdech_u16(svuint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdech_pat_n_s32))) +int32_t svqdech_pat_n_s32(int32_t, enum svpattern, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdech_pat_n_s64))) +int64_t svqdech_pat_n_s64(int64_t, enum svpattern, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdech_pat_n_u32))) +uint32_t svqdech_pat_n_u32(uint32_t, enum svpattern, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdech_pat_n_u64))) +uint64_t svqdech_pat_n_u64(uint64_t, enum svpattern, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdech_pat_s16))) +svint16_t svqdech_pat_s16(svint16_t, enum svpattern, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdech_pat_u16))) +svuint16_t svqdech_pat_u16(svuint16_t, enum svpattern, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecp_n_s32_b8))) +int32_t svqdecp_n_s32_b8(int32_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecp_n_s32_b32))) +int32_t svqdecp_n_s32_b32(int32_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecp_n_s32_b64))) +int32_t svqdecp_n_s32_b64(int32_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecp_n_s32_b16))) +int32_t svqdecp_n_s32_b16(int32_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecp_n_s64_b8))) +int64_t svqdecp_n_s64_b8(int64_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecp_n_s64_b32))) +int64_t svqdecp_n_s64_b32(int64_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecp_n_s64_b64))) +int64_t svqdecp_n_s64_b64(int64_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecp_n_s64_b16))) +int64_t svqdecp_n_s64_b16(int64_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecp_n_u32_b8))) +uint32_t svqdecp_n_u32_b8(uint32_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecp_n_u32_b32))) +uint32_t svqdecp_n_u32_b32(uint32_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecp_n_u32_b64))) +uint32_t svqdecp_n_u32_b64(uint32_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecp_n_u32_b16))) +uint32_t svqdecp_n_u32_b16(uint32_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecp_n_u64_b8))) +uint64_t svqdecp_n_u64_b8(uint64_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecp_n_u64_b32))) +uint64_t svqdecp_n_u64_b32(uint64_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecp_n_u64_b64))) +uint64_t svqdecp_n_u64_b64(uint64_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecp_n_u64_b16))) +uint64_t svqdecp_n_u64_b16(uint64_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecp_s32))) +svint32_t svqdecp_s32(svint32_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecp_s64))) +svint64_t svqdecp_s64(svint64_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecp_s16))) +svint16_t svqdecp_s16(svint16_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecp_u32))) +svuint32_t svqdecp_u32(svuint32_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecp_u64))) +svuint64_t svqdecp_u64(svuint64_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecp_u16))) +svuint16_t svqdecp_u16(svuint16_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecw_n_s32))) +int32_t svqdecw_n_s32(int32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecw_n_s64))) +int64_t svqdecw_n_s64(int64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecw_n_u32))) +uint32_t svqdecw_n_u32(uint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecw_n_u64))) +uint64_t svqdecw_n_u64(uint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecw_s32))) +svint32_t svqdecw_s32(svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecw_u32))) +svuint32_t svqdecw_u32(svuint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecw_pat_n_s32))) +int32_t svqdecw_pat_n_s32(int32_t, enum svpattern, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecw_pat_n_s64))) +int64_t svqdecw_pat_n_s64(int64_t, enum svpattern, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecw_pat_n_u32))) +uint32_t svqdecw_pat_n_u32(uint32_t, enum svpattern, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecw_pat_n_u64))) +uint64_t svqdecw_pat_n_u64(uint64_t, enum svpattern, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecw_pat_s32))) +svint32_t svqdecw_pat_s32(svint32_t, enum svpattern, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdecw_pat_u32))) +svuint32_t svqdecw_pat_u32(svuint32_t, enum svpattern, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincb_n_s32))) +int32_t svqincb_n_s32(int32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincb_n_s64))) +int64_t svqincb_n_s64(int64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincb_n_u32))) +uint32_t svqincb_n_u32(uint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincb_n_u64))) +uint64_t svqincb_n_u64(uint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincb_pat_n_s32))) +int32_t svqincb_pat_n_s32(int32_t, enum svpattern, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincb_pat_n_s64))) +int64_t svqincb_pat_n_s64(int64_t, enum svpattern, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincb_pat_n_u32))) +uint32_t svqincb_pat_n_u32(uint32_t, enum svpattern, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincb_pat_n_u64))) +uint64_t svqincb_pat_n_u64(uint64_t, enum svpattern, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincd_n_s32))) +int32_t svqincd_n_s32(int32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincd_n_s64))) +int64_t svqincd_n_s64(int64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincd_n_u32))) +uint32_t svqincd_n_u32(uint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincd_n_u64))) +uint64_t svqincd_n_u64(uint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincd_s64))) +svint64_t svqincd_s64(svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincd_u64))) +svuint64_t svqincd_u64(svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincd_pat_n_s32))) +int32_t svqincd_pat_n_s32(int32_t, enum svpattern, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincd_pat_n_s64))) +int64_t svqincd_pat_n_s64(int64_t, enum svpattern, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincd_pat_n_u32))) +uint32_t svqincd_pat_n_u32(uint32_t, enum svpattern, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincd_pat_n_u64))) +uint64_t svqincd_pat_n_u64(uint64_t, enum svpattern, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincd_pat_s64))) +svint64_t svqincd_pat_s64(svint64_t, enum svpattern, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincd_pat_u64))) +svuint64_t svqincd_pat_u64(svuint64_t, enum svpattern, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqinch_n_s32))) +int32_t svqinch_n_s32(int32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqinch_n_s64))) +int64_t svqinch_n_s64(int64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqinch_n_u32))) +uint32_t svqinch_n_u32(uint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqinch_n_u64))) +uint64_t svqinch_n_u64(uint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqinch_s16))) +svint16_t svqinch_s16(svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqinch_u16))) +svuint16_t svqinch_u16(svuint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqinch_pat_n_s32))) +int32_t svqinch_pat_n_s32(int32_t, enum svpattern, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqinch_pat_n_s64))) +int64_t svqinch_pat_n_s64(int64_t, enum svpattern, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqinch_pat_n_u32))) +uint32_t svqinch_pat_n_u32(uint32_t, enum svpattern, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqinch_pat_n_u64))) +uint64_t svqinch_pat_n_u64(uint64_t, enum svpattern, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqinch_pat_s16))) +svint16_t svqinch_pat_s16(svint16_t, enum svpattern, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqinch_pat_u16))) +svuint16_t svqinch_pat_u16(svuint16_t, enum svpattern, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincp_n_s32_b8))) +int32_t svqincp_n_s32_b8(int32_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincp_n_s32_b32))) +int32_t svqincp_n_s32_b32(int32_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincp_n_s32_b64))) +int32_t svqincp_n_s32_b64(int32_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincp_n_s32_b16))) +int32_t svqincp_n_s32_b16(int32_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincp_n_s64_b8))) +int64_t svqincp_n_s64_b8(int64_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincp_n_s64_b32))) +int64_t svqincp_n_s64_b32(int64_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincp_n_s64_b64))) +int64_t svqincp_n_s64_b64(int64_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincp_n_s64_b16))) +int64_t svqincp_n_s64_b16(int64_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincp_n_u32_b8))) +uint32_t svqincp_n_u32_b8(uint32_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincp_n_u32_b32))) +uint32_t svqincp_n_u32_b32(uint32_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincp_n_u32_b64))) +uint32_t svqincp_n_u32_b64(uint32_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincp_n_u32_b16))) +uint32_t svqincp_n_u32_b16(uint32_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincp_n_u64_b8))) +uint64_t svqincp_n_u64_b8(uint64_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincp_n_u64_b32))) +uint64_t svqincp_n_u64_b32(uint64_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincp_n_u64_b64))) +uint64_t svqincp_n_u64_b64(uint64_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincp_n_u64_b16))) +uint64_t svqincp_n_u64_b16(uint64_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincp_s32))) +svint32_t svqincp_s32(svint32_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincp_s64))) +svint64_t svqincp_s64(svint64_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincp_s16))) +svint16_t svqincp_s16(svint16_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincp_u32))) +svuint32_t svqincp_u32(svuint32_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincp_u64))) +svuint64_t svqincp_u64(svuint64_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincp_u16))) +svuint16_t svqincp_u16(svuint16_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincw_n_s32))) +int32_t svqincw_n_s32(int32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincw_n_s64))) +int64_t svqincw_n_s64(int64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincw_n_u32))) +uint32_t svqincw_n_u32(uint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincw_n_u64))) +uint64_t svqincw_n_u64(uint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincw_s32))) +svint32_t svqincw_s32(svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincw_u32))) +svuint32_t svqincw_u32(svuint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincw_pat_n_s32))) +int32_t svqincw_pat_n_s32(int32_t, enum svpattern, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincw_pat_n_s64))) +int64_t svqincw_pat_n_s64(int64_t, enum svpattern, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincw_pat_n_u32))) +uint32_t svqincw_pat_n_u32(uint32_t, enum svpattern, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincw_pat_n_u64))) +uint64_t svqincw_pat_n_u64(uint64_t, enum svpattern, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincw_pat_s32))) +svint32_t svqincw_pat_s32(svint32_t, enum svpattern, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqincw_pat_u32))) +svuint32_t svqincw_pat_u32(svuint32_t, enum svpattern, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_n_s8))) +svint8_t svqsub_n_s8(svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_n_s32))) +svint32_t svqsub_n_s32(svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_n_s64))) +svint64_t svqsub_n_s64(svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_n_s16))) +svint16_t svqsub_n_s16(svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_n_u8))) +svuint8_t svqsub_n_u8(svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_n_u32))) +svuint32_t svqsub_n_u32(svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_n_u64))) +svuint64_t svqsub_n_u64(svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_n_u16))) +svuint16_t svqsub_n_u16(svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_s8))) +svint8_t svqsub_s8(svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_s32))) +svint32_t svqsub_s32(svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_s64))) +svint64_t svqsub_s64(svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_s16))) +svint16_t svqsub_s16(svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_u8))) +svuint8_t svqsub_u8(svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_u32))) +svuint32_t svqsub_u32(svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_u64))) +svuint64_t svqsub_u64(svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_u16))) +svuint16_t svqsub_u16(svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrbit_u8_m))) +svuint8_t svrbit_u8_m(svuint8_t, svbool_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrbit_u32_m))) +svuint32_t svrbit_u32_m(svuint32_t, svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrbit_u64_m))) +svuint64_t svrbit_u64_m(svuint64_t, svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrbit_u16_m))) +svuint16_t svrbit_u16_m(svuint16_t, svbool_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrbit_s8_m))) +svint8_t svrbit_s8_m(svint8_t, svbool_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrbit_s32_m))) +svint32_t svrbit_s32_m(svint32_t, svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrbit_s64_m))) +svint64_t svrbit_s64_m(svint64_t, svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrbit_s16_m))) +svint16_t svrbit_s16_m(svint16_t, svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrbit_u8_x))) +svuint8_t svrbit_u8_x(svbool_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrbit_u32_x))) +svuint32_t svrbit_u32_x(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrbit_u64_x))) +svuint64_t svrbit_u64_x(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrbit_u16_x))) +svuint16_t svrbit_u16_x(svbool_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrbit_s8_x))) +svint8_t svrbit_s8_x(svbool_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrbit_s32_x))) +svint32_t svrbit_s32_x(svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrbit_s64_x))) +svint64_t svrbit_s64_x(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrbit_s16_x))) +svint16_t svrbit_s16_x(svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrbit_u8_z))) +svuint8_t svrbit_u8_z(svbool_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrbit_u32_z))) +svuint32_t svrbit_u32_z(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrbit_u64_z))) +svuint64_t svrbit_u64_z(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrbit_u16_z))) +svuint16_t svrbit_u16_z(svbool_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrbit_s8_z))) +svint8_t svrbit_s8_z(svbool_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrbit_s32_z))) +svint32_t svrbit_s32_z(svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrbit_s64_z))) +svint64_t svrbit_s64_z(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrbit_s16_z))) +svint16_t svrbit_s16_z(svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrdffr))) +svbool_t svrdffr(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrdffr_z))) +svbool_t svrdffr_z(svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrecpe_f64))) +svfloat64_t svrecpe_f64(svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrecpe_f32))) +svfloat32_t svrecpe_f32(svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrecpe_f16))) +svfloat16_t svrecpe_f16(svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrecps_f64))) +svfloat64_t svrecps_f64(svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrecps_f32))) +svfloat32_t svrecps_f32(svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrecps_f16))) +svfloat16_t svrecps_f16(svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrecpx_f64_m))) +svfloat64_t svrecpx_f64_m(svfloat64_t, svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrecpx_f32_m))) +svfloat32_t svrecpx_f32_m(svfloat32_t, svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrecpx_f16_m))) +svfloat16_t svrecpx_f16_m(svfloat16_t, svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrecpx_f64_x))) +svfloat64_t svrecpx_f64_x(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrecpx_f32_x))) +svfloat32_t svrecpx_f32_x(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrecpx_f16_x))) +svfloat16_t svrecpx_f16_x(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrecpx_f64_z))) +svfloat64_t svrecpx_f64_z(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrecpx_f32_z))) +svfloat32_t svrecpx_f32_z(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrecpx_f16_z))) +svfloat16_t svrecpx_f16_z(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrev_u8))) +svuint8_t svrev_u8(svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrev_u32))) +svuint32_t svrev_u32(svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrev_u64))) +svuint64_t svrev_u64(svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrev_u16))) +svuint16_t svrev_u16(svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrev_s8))) +svint8_t svrev_s8(svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrev_f64))) +svfloat64_t svrev_f64(svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrev_f32))) +svfloat32_t svrev_f32(svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrev_f16))) +svfloat16_t svrev_f16(svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrev_s32))) +svint32_t svrev_s32(svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrev_s64))) +svint64_t svrev_s64(svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrev_s16))) +svint16_t svrev_s16(svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrev_b8))) +svbool_t svrev_b8(svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrev_b32))) +svbool_t svrev_b32(svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrev_b64))) +svbool_t svrev_b64(svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrev_b16))) +svbool_t svrev_b16(svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrevb_u32_m))) +svuint32_t svrevb_u32_m(svuint32_t, svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrevb_u64_m))) +svuint64_t svrevb_u64_m(svuint64_t, svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrevb_u16_m))) +svuint16_t svrevb_u16_m(svuint16_t, svbool_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrevb_s32_m))) +svint32_t svrevb_s32_m(svint32_t, svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrevb_s64_m))) +svint64_t svrevb_s64_m(svint64_t, svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrevb_s16_m))) +svint16_t svrevb_s16_m(svint16_t, svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrevb_u32_x))) +svuint32_t svrevb_u32_x(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrevb_u64_x))) +svuint64_t svrevb_u64_x(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrevb_u16_x))) +svuint16_t svrevb_u16_x(svbool_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrevb_s32_x))) +svint32_t svrevb_s32_x(svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrevb_s64_x))) +svint64_t svrevb_s64_x(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrevb_s16_x))) +svint16_t svrevb_s16_x(svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrevb_u32_z))) +svuint32_t svrevb_u32_z(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrevb_u64_z))) +svuint64_t svrevb_u64_z(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrevb_u16_z))) +svuint16_t svrevb_u16_z(svbool_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrevb_s32_z))) +svint32_t svrevb_s32_z(svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrevb_s64_z))) +svint64_t svrevb_s64_z(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrevb_s16_z))) +svint16_t svrevb_s16_z(svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrevh_u32_m))) +svuint32_t svrevh_u32_m(svuint32_t, svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrevh_u64_m))) +svuint64_t svrevh_u64_m(svuint64_t, svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrevh_s32_m))) +svint32_t svrevh_s32_m(svint32_t, svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrevh_s64_m))) +svint64_t svrevh_s64_m(svint64_t, svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrevh_u32_x))) +svuint32_t svrevh_u32_x(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrevh_u64_x))) +svuint64_t svrevh_u64_x(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrevh_s32_x))) +svint32_t svrevh_s32_x(svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrevh_s64_x))) +svint64_t svrevh_s64_x(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrevh_u32_z))) +svuint32_t svrevh_u32_z(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrevh_u64_z))) +svuint64_t svrevh_u64_z(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrevh_s32_z))) +svint32_t svrevh_s32_z(svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrevh_s64_z))) +svint64_t svrevh_s64_z(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrevw_u64_m))) +svuint64_t svrevw_u64_m(svuint64_t, svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrevw_s64_m))) +svint64_t svrevw_s64_m(svint64_t, svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrevw_u64_x))) +svuint64_t svrevw_u64_x(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrevw_s64_x))) +svint64_t svrevw_s64_x(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrevw_u64_z))) +svuint64_t svrevw_u64_z(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrevw_s64_z))) +svint64_t svrevw_s64_z(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrinta_f64_m))) +svfloat64_t svrinta_f64_m(svfloat64_t, svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrinta_f32_m))) +svfloat32_t svrinta_f32_m(svfloat32_t, svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrinta_f16_m))) +svfloat16_t svrinta_f16_m(svfloat16_t, svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrinta_f64_x))) +svfloat64_t svrinta_f64_x(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrinta_f32_x))) +svfloat32_t svrinta_f32_x(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrinta_f16_x))) +svfloat16_t svrinta_f16_x(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrinta_f64_z))) +svfloat64_t svrinta_f64_z(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrinta_f32_z))) +svfloat32_t svrinta_f32_z(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrinta_f16_z))) +svfloat16_t svrinta_f16_z(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrinti_f64_m))) +svfloat64_t svrinti_f64_m(svfloat64_t, svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrinti_f32_m))) +svfloat32_t svrinti_f32_m(svfloat32_t, svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrinti_f16_m))) +svfloat16_t svrinti_f16_m(svfloat16_t, svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrinti_f64_x))) +svfloat64_t svrinti_f64_x(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrinti_f32_x))) +svfloat32_t svrinti_f32_x(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrinti_f16_x))) +svfloat16_t svrinti_f16_x(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrinti_f64_z))) +svfloat64_t svrinti_f64_z(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrinti_f32_z))) +svfloat32_t svrinti_f32_z(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrinti_f16_z))) +svfloat16_t svrinti_f16_z(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrintm_f64_m))) +svfloat64_t svrintm_f64_m(svfloat64_t, svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrintm_f32_m))) +svfloat32_t svrintm_f32_m(svfloat32_t, svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrintm_f16_m))) +svfloat16_t svrintm_f16_m(svfloat16_t, svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrintm_f64_x))) +svfloat64_t svrintm_f64_x(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrintm_f32_x))) +svfloat32_t svrintm_f32_x(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrintm_f16_x))) +svfloat16_t svrintm_f16_x(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrintm_f64_z))) +svfloat64_t svrintm_f64_z(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrintm_f32_z))) +svfloat32_t svrintm_f32_z(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrintm_f16_z))) +svfloat16_t svrintm_f16_z(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrintn_f64_m))) +svfloat64_t svrintn_f64_m(svfloat64_t, svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrintn_f32_m))) +svfloat32_t svrintn_f32_m(svfloat32_t, svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrintn_f16_m))) +svfloat16_t svrintn_f16_m(svfloat16_t, svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrintn_f64_x))) +svfloat64_t svrintn_f64_x(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrintn_f32_x))) +svfloat32_t svrintn_f32_x(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrintn_f16_x))) +svfloat16_t svrintn_f16_x(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrintn_f64_z))) +svfloat64_t svrintn_f64_z(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrintn_f32_z))) +svfloat32_t svrintn_f32_z(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrintn_f16_z))) +svfloat16_t svrintn_f16_z(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrintp_f64_m))) +svfloat64_t svrintp_f64_m(svfloat64_t, svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrintp_f32_m))) +svfloat32_t svrintp_f32_m(svfloat32_t, svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrintp_f16_m))) +svfloat16_t svrintp_f16_m(svfloat16_t, svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrintp_f64_x))) +svfloat64_t svrintp_f64_x(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrintp_f32_x))) +svfloat32_t svrintp_f32_x(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrintp_f16_x))) +svfloat16_t svrintp_f16_x(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrintp_f64_z))) +svfloat64_t svrintp_f64_z(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrintp_f32_z))) +svfloat32_t svrintp_f32_z(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrintp_f16_z))) +svfloat16_t svrintp_f16_z(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrintx_f64_m))) +svfloat64_t svrintx_f64_m(svfloat64_t, svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrintx_f32_m))) +svfloat32_t svrintx_f32_m(svfloat32_t, svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrintx_f16_m))) +svfloat16_t svrintx_f16_m(svfloat16_t, svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrintx_f64_x))) +svfloat64_t svrintx_f64_x(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrintx_f32_x))) +svfloat32_t svrintx_f32_x(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrintx_f16_x))) +svfloat16_t svrintx_f16_x(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrintx_f64_z))) +svfloat64_t svrintx_f64_z(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrintx_f32_z))) +svfloat32_t svrintx_f32_z(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrintx_f16_z))) +svfloat16_t svrintx_f16_z(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrintz_f64_m))) +svfloat64_t svrintz_f64_m(svfloat64_t, svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrintz_f32_m))) +svfloat32_t svrintz_f32_m(svfloat32_t, svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrintz_f16_m))) +svfloat16_t svrintz_f16_m(svfloat16_t, svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrintz_f64_x))) +svfloat64_t svrintz_f64_x(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrintz_f32_x))) +svfloat32_t svrintz_f32_x(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrintz_f16_x))) +svfloat16_t svrintz_f16_x(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrintz_f64_z))) +svfloat64_t svrintz_f64_z(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrintz_f32_z))) +svfloat32_t svrintz_f32_z(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrintz_f16_z))) +svfloat16_t svrintz_f16_z(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrsqrte_f64))) +svfloat64_t svrsqrte_f64(svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrsqrte_f32))) +svfloat32_t svrsqrte_f32(svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrsqrte_f16))) +svfloat16_t svrsqrte_f16(svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrsqrts_f64))) +svfloat64_t svrsqrts_f64(svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrsqrts_f32))) +svfloat32_t svrsqrts_f32(svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrsqrts_f16))) +svfloat16_t svrsqrts_f16(svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svscale_n_f64_m))) +svfloat64_t svscale_n_f64_m(svbool_t, svfloat64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svscale_n_f32_m))) +svfloat32_t svscale_n_f32_m(svbool_t, svfloat32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svscale_n_f16_m))) +svfloat16_t svscale_n_f16_m(svbool_t, svfloat16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svscale_n_f64_x))) +svfloat64_t svscale_n_f64_x(svbool_t, svfloat64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svscale_n_f32_x))) +svfloat32_t svscale_n_f32_x(svbool_t, svfloat32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svscale_n_f16_x))) +svfloat16_t svscale_n_f16_x(svbool_t, svfloat16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svscale_n_f64_z))) +svfloat64_t svscale_n_f64_z(svbool_t, svfloat64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svscale_n_f32_z))) +svfloat32_t svscale_n_f32_z(svbool_t, svfloat32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svscale_n_f16_z))) +svfloat16_t svscale_n_f16_z(svbool_t, svfloat16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svscale_f64_m))) +svfloat64_t svscale_f64_m(svbool_t, svfloat64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svscale_f32_m))) +svfloat32_t svscale_f32_m(svbool_t, svfloat32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svscale_f16_m))) +svfloat16_t svscale_f16_m(svbool_t, svfloat16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svscale_f64_x))) +svfloat64_t svscale_f64_x(svbool_t, svfloat64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svscale_f32_x))) +svfloat32_t svscale_f32_x(svbool_t, svfloat32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svscale_f16_x))) +svfloat16_t svscale_f16_x(svbool_t, svfloat16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svscale_f64_z))) +svfloat64_t svscale_f64_z(svbool_t, svfloat64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svscale_f32_z))) +svfloat32_t svscale_f32_z(svbool_t, svfloat32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svscale_f16_z))) +svfloat16_t svscale_f16_z(svbool_t, svfloat16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsel_b))) +svbool_t svsel_b(svbool_t, svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsel_u8))) +svuint8_t svsel_u8(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsel_u32))) +svuint32_t svsel_u32(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsel_u64))) +svuint64_t svsel_u64(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsel_u16))) +svuint16_t svsel_u16(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsel_s8))) +svint8_t svsel_s8(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsel_f64))) +svfloat64_t svsel_f64(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsel_f32))) +svfloat32_t svsel_f32(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsel_f16))) +svfloat16_t svsel_f16(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsel_s32))) +svint32_t svsel_s32(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsel_s64))) +svint64_t svsel_s64(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsel_s16))) +svint16_t svsel_s16(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svset2_u8))) +svuint8x2_t svset2_u8(svuint8x2_t, uint64_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svset2_u32))) +svuint32x2_t svset2_u32(svuint32x2_t, uint64_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svset2_u64))) +svuint64x2_t svset2_u64(svuint64x2_t, uint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svset2_u16))) +svuint16x2_t svset2_u16(svuint16x2_t, uint64_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svset2_s8))) +svint8x2_t svset2_s8(svint8x2_t, uint64_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svset2_f64))) +svfloat64x2_t svset2_f64(svfloat64x2_t, uint64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svset2_f32))) +svfloat32x2_t svset2_f32(svfloat32x2_t, uint64_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svset2_f16))) +svfloat16x2_t svset2_f16(svfloat16x2_t, uint64_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svset2_s32))) +svint32x2_t svset2_s32(svint32x2_t, uint64_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svset2_s64))) +svint64x2_t svset2_s64(svint64x2_t, uint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svset2_s16))) +svint16x2_t svset2_s16(svint16x2_t, uint64_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svset3_u8))) +svuint8x3_t svset3_u8(svuint8x3_t, uint64_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svset3_u32))) +svuint32x3_t svset3_u32(svuint32x3_t, uint64_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svset3_u64))) +svuint64x3_t svset3_u64(svuint64x3_t, uint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svset3_u16))) +svuint16x3_t svset3_u16(svuint16x3_t, uint64_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svset3_s8))) +svint8x3_t svset3_s8(svint8x3_t, uint64_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svset3_f64))) +svfloat64x3_t svset3_f64(svfloat64x3_t, uint64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svset3_f32))) +svfloat32x3_t svset3_f32(svfloat32x3_t, uint64_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svset3_f16))) +svfloat16x3_t svset3_f16(svfloat16x3_t, uint64_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svset3_s32))) +svint32x3_t svset3_s32(svint32x3_t, uint64_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svset3_s64))) +svint64x3_t svset3_s64(svint64x3_t, uint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svset3_s16))) +svint16x3_t svset3_s16(svint16x3_t, uint64_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svset4_u8))) +svuint8x4_t svset4_u8(svuint8x4_t, uint64_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svset4_u32))) +svuint32x4_t svset4_u32(svuint32x4_t, uint64_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svset4_u64))) +svuint64x4_t svset4_u64(svuint64x4_t, uint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svset4_u16))) +svuint16x4_t svset4_u16(svuint16x4_t, uint64_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svset4_s8))) +svint8x4_t svset4_s8(svint8x4_t, uint64_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svset4_f64))) +svfloat64x4_t svset4_f64(svfloat64x4_t, uint64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svset4_f32))) +svfloat32x4_t svset4_f32(svfloat32x4_t, uint64_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svset4_f16))) +svfloat16x4_t svset4_f16(svfloat16x4_t, uint64_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svset4_s32))) +svint32x4_t svset4_s32(svint32x4_t, uint64_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svset4_s64))) +svint64x4_t svset4_s64(svint64x4_t, uint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svset4_s16))) +svint16x4_t svset4_s16(svint16x4_t, uint64_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsetffr))) +void svsetffr(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsplice_u8))) +svuint8_t svsplice_u8(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsplice_u32))) +svuint32_t svsplice_u32(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsplice_u64))) +svuint64_t svsplice_u64(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsplice_u16))) +svuint16_t svsplice_u16(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsplice_s8))) +svint8_t svsplice_s8(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsplice_f64))) +svfloat64_t svsplice_f64(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsplice_f32))) +svfloat32_t svsplice_f32(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsplice_f16))) +svfloat16_t svsplice_f16(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsplice_s32))) +svint32_t svsplice_s32(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsplice_s64))) +svint64_t svsplice_s64(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsplice_s16))) +svint16_t svsplice_s16(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsqrt_f64_m))) +svfloat64_t svsqrt_f64_m(svfloat64_t, svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsqrt_f32_m))) +svfloat32_t svsqrt_f32_m(svfloat32_t, svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsqrt_f16_m))) +svfloat16_t svsqrt_f16_m(svfloat16_t, svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsqrt_f64_x))) +svfloat64_t svsqrt_f64_x(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsqrt_f32_x))) +svfloat32_t svsqrt_f32_x(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsqrt_f16_x))) +svfloat16_t svsqrt_f16_x(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsqrt_f64_z))) +svfloat64_t svsqrt_f64_z(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsqrt_f32_z))) +svfloat32_t svsqrt_f32_z(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsqrt_f16_z))) +svfloat16_t svsqrt_f16_z(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_u8))) +void svst1_u8(svbool_t, uint8_t *, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_u32))) +void svst1_u32(svbool_t, uint32_t *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_u64))) +void svst1_u64(svbool_t, uint64_t *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_u16))) +void svst1_u16(svbool_t, uint16_t *, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_s8))) +void svst1_s8(svbool_t, int8_t *, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_f64))) +void svst1_f64(svbool_t, float64_t *, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_f32))) +void svst1_f32(svbool_t, float32_t *, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_f16))) +void svst1_f16(svbool_t, float16_t *, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_s32))) +void svst1_s32(svbool_t, int32_t *, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_s64))) +void svst1_s64(svbool_t, int64_t *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_s16))) +void svst1_s16(svbool_t, int16_t *, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_scatter_u32base_index_u32))) +void svst1_scatter_u32base_index_u32(svbool_t, svuint32_t, int64_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_scatter_u64base_index_u64))) +void svst1_scatter_u64base_index_u64(svbool_t, svuint64_t, int64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_scatter_u64base_index_f64))) +void svst1_scatter_u64base_index_f64(svbool_t, svuint64_t, int64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_scatter_u32base_index_f32))) +void svst1_scatter_u32base_index_f32(svbool_t, svuint32_t, int64_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_scatter_u32base_index_s32))) +void svst1_scatter_u32base_index_s32(svbool_t, svuint32_t, int64_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_scatter_u64base_index_s64))) +void svst1_scatter_u64base_index_s64(svbool_t, svuint64_t, int64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_scatter_u32base_offset_u32))) +void svst1_scatter_u32base_offset_u32(svbool_t, svuint32_t, int64_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_scatter_u64base_offset_u64))) +void svst1_scatter_u64base_offset_u64(svbool_t, svuint64_t, int64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_scatter_u64base_offset_f64))) +void svst1_scatter_u64base_offset_f64(svbool_t, svuint64_t, int64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_scatter_u32base_offset_f32))) +void svst1_scatter_u32base_offset_f32(svbool_t, svuint32_t, int64_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_scatter_u32base_offset_s32))) +void svst1_scatter_u32base_offset_s32(svbool_t, svuint32_t, int64_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_scatter_u64base_offset_s64))) +void svst1_scatter_u64base_offset_s64(svbool_t, svuint64_t, int64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_scatter_u32base_u32))) +void svst1_scatter_u32base_u32(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_scatter_u64base_u64))) +void svst1_scatter_u64base_u64(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_scatter_u64base_f64))) +void svst1_scatter_u64base_f64(svbool_t, svuint64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_scatter_u32base_f32))) +void svst1_scatter_u32base_f32(svbool_t, svuint32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_scatter_u32base_s32))) +void svst1_scatter_u32base_s32(svbool_t, svuint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_scatter_u64base_s64))) +void svst1_scatter_u64base_s64(svbool_t, svuint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_scatter_s32index_u32))) +void svst1_scatter_s32index_u32(svbool_t, uint32_t *, svint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_scatter_s32index_f32))) +void svst1_scatter_s32index_f32(svbool_t, float32_t *, svint32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_scatter_s32index_s32))) +void svst1_scatter_s32index_s32(svbool_t, int32_t *, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_scatter_u32index_u32))) +void svst1_scatter_u32index_u32(svbool_t, uint32_t *, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_scatter_u32index_f32))) +void svst1_scatter_u32index_f32(svbool_t, float32_t *, svuint32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_scatter_u32index_s32))) +void svst1_scatter_u32index_s32(svbool_t, int32_t *, svuint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_scatter_s64index_u64))) +void svst1_scatter_s64index_u64(svbool_t, uint64_t *, svint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_scatter_s64index_f64))) +void svst1_scatter_s64index_f64(svbool_t, float64_t *, svint64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_scatter_s64index_s64))) +void svst1_scatter_s64index_s64(svbool_t, int64_t *, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_scatter_u64index_u64))) +void svst1_scatter_u64index_u64(svbool_t, uint64_t *, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_scatter_u64index_f64))) +void svst1_scatter_u64index_f64(svbool_t, float64_t *, svuint64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_scatter_u64index_s64))) +void svst1_scatter_u64index_s64(svbool_t, int64_t *, svuint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_scatter_s32offset_u32))) +void svst1_scatter_s32offset_u32(svbool_t, uint32_t *, svint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_scatter_s32offset_f32))) +void svst1_scatter_s32offset_f32(svbool_t, float32_t *, svint32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_scatter_s32offset_s32))) +void svst1_scatter_s32offset_s32(svbool_t, int32_t *, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_scatter_u32offset_u32))) +void svst1_scatter_u32offset_u32(svbool_t, uint32_t *, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_scatter_u32offset_f32))) +void svst1_scatter_u32offset_f32(svbool_t, float32_t *, svuint32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_scatter_u32offset_s32))) +void svst1_scatter_u32offset_s32(svbool_t, int32_t *, svuint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_scatter_s64offset_u64))) +void svst1_scatter_s64offset_u64(svbool_t, uint64_t *, svint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_scatter_s64offset_f64))) +void svst1_scatter_s64offset_f64(svbool_t, float64_t *, svint64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_scatter_s64offset_s64))) +void svst1_scatter_s64offset_s64(svbool_t, int64_t *, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_scatter_u64offset_u64))) +void svst1_scatter_u64offset_u64(svbool_t, uint64_t *, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_scatter_u64offset_f64))) +void svst1_scatter_u64offset_f64(svbool_t, float64_t *, svuint64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_scatter_u64offset_s64))) +void svst1_scatter_u64offset_s64(svbool_t, int64_t *, svuint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_vnum_u8))) +void svst1_vnum_u8(svbool_t, uint8_t *, int64_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_vnum_u32))) +void svst1_vnum_u32(svbool_t, uint32_t *, int64_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_vnum_u64))) +void svst1_vnum_u64(svbool_t, uint64_t *, int64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_vnum_u16))) +void svst1_vnum_u16(svbool_t, uint16_t *, int64_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_vnum_s8))) +void svst1_vnum_s8(svbool_t, int8_t *, int64_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_vnum_f64))) +void svst1_vnum_f64(svbool_t, float64_t *, int64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_vnum_f32))) +void svst1_vnum_f32(svbool_t, float32_t *, int64_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_vnum_f16))) +void svst1_vnum_f16(svbool_t, float16_t *, int64_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_vnum_s32))) +void svst1_vnum_s32(svbool_t, int32_t *, int64_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_vnum_s64))) +void svst1_vnum_s64(svbool_t, int64_t *, int64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_vnum_s16))) +void svst1_vnum_s16(svbool_t, int16_t *, int64_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1b_s32))) +void svst1b_s32(svbool_t, int8_t *, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1b_s64))) +void svst1b_s64(svbool_t, int8_t *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1b_s16))) +void svst1b_s16(svbool_t, int8_t *, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1b_u32))) +void svst1b_u32(svbool_t, uint8_t *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1b_u64))) +void svst1b_u64(svbool_t, uint8_t *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1b_u16))) +void svst1b_u16(svbool_t, uint8_t *, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1b_scatter_u32base_offset_u32))) +void svst1b_scatter_u32base_offset_u32(svbool_t, svuint32_t, int64_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1b_scatter_u64base_offset_u64))) +void svst1b_scatter_u64base_offset_u64(svbool_t, svuint64_t, int64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1b_scatter_u32base_offset_s32))) +void svst1b_scatter_u32base_offset_s32(svbool_t, svuint32_t, int64_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1b_scatter_u64base_offset_s64))) +void svst1b_scatter_u64base_offset_s64(svbool_t, svuint64_t, int64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1b_scatter_u32base_u32))) +void svst1b_scatter_u32base_u32(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1b_scatter_u64base_u64))) +void svst1b_scatter_u64base_u64(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1b_scatter_u32base_s32))) +void svst1b_scatter_u32base_s32(svbool_t, svuint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1b_scatter_u64base_s64))) +void svst1b_scatter_u64base_s64(svbool_t, svuint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1b_scatter_s32offset_s32))) +void svst1b_scatter_s32offset_s32(svbool_t, int8_t *, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1b_scatter_s32offset_u32))) +void svst1b_scatter_s32offset_u32(svbool_t, uint8_t *, svint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1b_scatter_u32offset_s32))) +void svst1b_scatter_u32offset_s32(svbool_t, int8_t *, svuint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1b_scatter_u32offset_u32))) +void svst1b_scatter_u32offset_u32(svbool_t, uint8_t *, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1b_scatter_s64offset_s64))) +void svst1b_scatter_s64offset_s64(svbool_t, int8_t *, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1b_scatter_s64offset_u64))) +void svst1b_scatter_s64offset_u64(svbool_t, uint8_t *, svint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1b_scatter_u64offset_s64))) +void svst1b_scatter_u64offset_s64(svbool_t, int8_t *, svuint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1b_scatter_u64offset_u64))) +void svst1b_scatter_u64offset_u64(svbool_t, uint8_t *, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1b_vnum_s32))) +void svst1b_vnum_s32(svbool_t, int8_t *, int64_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1b_vnum_s64))) +void svst1b_vnum_s64(svbool_t, int8_t *, int64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1b_vnum_s16))) +void svst1b_vnum_s16(svbool_t, int8_t *, int64_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1b_vnum_u32))) +void svst1b_vnum_u32(svbool_t, uint8_t *, int64_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1b_vnum_u64))) +void svst1b_vnum_u64(svbool_t, uint8_t *, int64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1b_vnum_u16))) +void svst1b_vnum_u16(svbool_t, uint8_t *, int64_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1h_s32))) +void svst1h_s32(svbool_t, int16_t *, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1h_s64))) +void svst1h_s64(svbool_t, int16_t *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1h_u32))) +void svst1h_u32(svbool_t, uint16_t *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1h_u64))) +void svst1h_u64(svbool_t, uint16_t *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1h_scatter_u32base_index_u32))) +void svst1h_scatter_u32base_index_u32(svbool_t, svuint32_t, int64_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1h_scatter_u64base_index_u64))) +void svst1h_scatter_u64base_index_u64(svbool_t, svuint64_t, int64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1h_scatter_u32base_index_s32))) +void svst1h_scatter_u32base_index_s32(svbool_t, svuint32_t, int64_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1h_scatter_u64base_index_s64))) +void svst1h_scatter_u64base_index_s64(svbool_t, svuint64_t, int64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1h_scatter_u32base_offset_u32))) +void svst1h_scatter_u32base_offset_u32(svbool_t, svuint32_t, int64_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1h_scatter_u64base_offset_u64))) +void svst1h_scatter_u64base_offset_u64(svbool_t, svuint64_t, int64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1h_scatter_u32base_offset_s32))) +void svst1h_scatter_u32base_offset_s32(svbool_t, svuint32_t, int64_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1h_scatter_u64base_offset_s64))) +void svst1h_scatter_u64base_offset_s64(svbool_t, svuint64_t, int64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1h_scatter_u32base_u32))) +void svst1h_scatter_u32base_u32(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1h_scatter_u64base_u64))) +void svst1h_scatter_u64base_u64(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1h_scatter_u32base_s32))) +void svst1h_scatter_u32base_s32(svbool_t, svuint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1h_scatter_u64base_s64))) +void svst1h_scatter_u64base_s64(svbool_t, svuint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1h_scatter_s32index_s32))) +void svst1h_scatter_s32index_s32(svbool_t, int16_t *, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1h_scatter_s32index_u32))) +void svst1h_scatter_s32index_u32(svbool_t, uint16_t *, svint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1h_scatter_u32index_s32))) +void svst1h_scatter_u32index_s32(svbool_t, int16_t *, svuint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1h_scatter_u32index_u32))) +void svst1h_scatter_u32index_u32(svbool_t, uint16_t *, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1h_scatter_s64index_s64))) +void svst1h_scatter_s64index_s64(svbool_t, int16_t *, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1h_scatter_s64index_u64))) +void svst1h_scatter_s64index_u64(svbool_t, uint16_t *, svint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1h_scatter_u64index_s64))) +void svst1h_scatter_u64index_s64(svbool_t, int16_t *, svuint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1h_scatter_u64index_u64))) +void svst1h_scatter_u64index_u64(svbool_t, uint16_t *, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1h_scatter_s32offset_s32))) +void svst1h_scatter_s32offset_s32(svbool_t, int16_t *, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1h_scatter_s32offset_u32))) +void svst1h_scatter_s32offset_u32(svbool_t, uint16_t *, svint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1h_scatter_u32offset_s32))) +void svst1h_scatter_u32offset_s32(svbool_t, int16_t *, svuint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1h_scatter_u32offset_u32))) +void svst1h_scatter_u32offset_u32(svbool_t, uint16_t *, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1h_scatter_s64offset_s64))) +void svst1h_scatter_s64offset_s64(svbool_t, int16_t *, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1h_scatter_s64offset_u64))) +void svst1h_scatter_s64offset_u64(svbool_t, uint16_t *, svint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1h_scatter_u64offset_s64))) +void svst1h_scatter_u64offset_s64(svbool_t, int16_t *, svuint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1h_scatter_u64offset_u64))) +void svst1h_scatter_u64offset_u64(svbool_t, uint16_t *, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1h_vnum_s32))) +void svst1h_vnum_s32(svbool_t, int16_t *, int64_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1h_vnum_s64))) +void svst1h_vnum_s64(svbool_t, int16_t *, int64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1h_vnum_u32))) +void svst1h_vnum_u32(svbool_t, uint16_t *, int64_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1h_vnum_u64))) +void svst1h_vnum_u64(svbool_t, uint16_t *, int64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1w_s64))) +void svst1w_s64(svbool_t, int32_t *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1w_u64))) +void svst1w_u64(svbool_t, uint32_t *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1w_scatter_u64base_index_u64))) +void svst1w_scatter_u64base_index_u64(svbool_t, svuint64_t, int64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1w_scatter_u64base_index_s64))) +void svst1w_scatter_u64base_index_s64(svbool_t, svuint64_t, int64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1w_scatter_u64base_offset_u64))) +void svst1w_scatter_u64base_offset_u64(svbool_t, svuint64_t, int64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1w_scatter_u64base_offset_s64))) +void svst1w_scatter_u64base_offset_s64(svbool_t, svuint64_t, int64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1w_scatter_u64base_u64))) +void svst1w_scatter_u64base_u64(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1w_scatter_u64base_s64))) +void svst1w_scatter_u64base_s64(svbool_t, svuint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1w_scatter_s64index_s64))) +void svst1w_scatter_s64index_s64(svbool_t, int32_t *, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1w_scatter_s64index_u64))) +void svst1w_scatter_s64index_u64(svbool_t, uint32_t *, svint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1w_scatter_u64index_s64))) +void svst1w_scatter_u64index_s64(svbool_t, int32_t *, svuint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1w_scatter_u64index_u64))) +void svst1w_scatter_u64index_u64(svbool_t, uint32_t *, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1w_scatter_s64offset_s64))) +void svst1w_scatter_s64offset_s64(svbool_t, int32_t *, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1w_scatter_s64offset_u64))) +void svst1w_scatter_s64offset_u64(svbool_t, uint32_t *, svint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1w_scatter_u64offset_s64))) +void svst1w_scatter_u64offset_s64(svbool_t, int32_t *, svuint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1w_scatter_u64offset_u64))) +void svst1w_scatter_u64offset_u64(svbool_t, uint32_t *, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1w_vnum_s64))) +void svst1w_vnum_s64(svbool_t, int32_t *, int64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1w_vnum_u64))) +void svst1w_vnum_u64(svbool_t, uint32_t *, int64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst2_u8))) +void svst2_u8(svbool_t, uint8_t *, svuint8x2_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst2_u32))) +void svst2_u32(svbool_t, uint32_t *, svuint32x2_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst2_u64))) +void svst2_u64(svbool_t, uint64_t *, svuint64x2_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst2_u16))) +void svst2_u16(svbool_t, uint16_t *, svuint16x2_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst2_s8))) +void svst2_s8(svbool_t, int8_t *, svint8x2_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst2_f64))) +void svst2_f64(svbool_t, float64_t *, svfloat64x2_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst2_f32))) +void svst2_f32(svbool_t, float32_t *, svfloat32x2_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst2_f16))) +void svst2_f16(svbool_t, float16_t *, svfloat16x2_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst2_s32))) +void svst2_s32(svbool_t, int32_t *, svint32x2_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst2_s64))) +void svst2_s64(svbool_t, int64_t *, svint64x2_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst2_s16))) +void svst2_s16(svbool_t, int16_t *, svint16x2_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst2_vnum_u8))) +void svst2_vnum_u8(svbool_t, uint8_t *, int64_t, svuint8x2_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst2_vnum_u32))) +void svst2_vnum_u32(svbool_t, uint32_t *, int64_t, svuint32x2_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst2_vnum_u64))) +void svst2_vnum_u64(svbool_t, uint64_t *, int64_t, svuint64x2_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst2_vnum_u16))) +void svst2_vnum_u16(svbool_t, uint16_t *, int64_t, svuint16x2_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst2_vnum_s8))) +void svst2_vnum_s8(svbool_t, int8_t *, int64_t, svint8x2_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst2_vnum_f64))) +void svst2_vnum_f64(svbool_t, float64_t *, int64_t, svfloat64x2_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst2_vnum_f32))) +void svst2_vnum_f32(svbool_t, float32_t *, int64_t, svfloat32x2_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst2_vnum_f16))) +void svst2_vnum_f16(svbool_t, float16_t *, int64_t, svfloat16x2_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst2_vnum_s32))) +void svst2_vnum_s32(svbool_t, int32_t *, int64_t, svint32x2_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst2_vnum_s64))) +void svst2_vnum_s64(svbool_t, int64_t *, int64_t, svint64x2_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst2_vnum_s16))) +void svst2_vnum_s16(svbool_t, int16_t *, int64_t, svint16x2_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst3_u8))) +void svst3_u8(svbool_t, uint8_t *, svuint8x3_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst3_u32))) +void svst3_u32(svbool_t, uint32_t *, svuint32x3_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst3_u64))) +void svst3_u64(svbool_t, uint64_t *, svuint64x3_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst3_u16))) +void svst3_u16(svbool_t, uint16_t *, svuint16x3_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst3_s8))) +void svst3_s8(svbool_t, int8_t *, svint8x3_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst3_f64))) +void svst3_f64(svbool_t, float64_t *, svfloat64x3_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst3_f32))) +void svst3_f32(svbool_t, float32_t *, svfloat32x3_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst3_f16))) +void svst3_f16(svbool_t, float16_t *, svfloat16x3_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst3_s32))) +void svst3_s32(svbool_t, int32_t *, svint32x3_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst3_s64))) +void svst3_s64(svbool_t, int64_t *, svint64x3_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst3_s16))) +void svst3_s16(svbool_t, int16_t *, svint16x3_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst3_vnum_u8))) +void svst3_vnum_u8(svbool_t, uint8_t *, int64_t, svuint8x3_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst3_vnum_u32))) +void svst3_vnum_u32(svbool_t, uint32_t *, int64_t, svuint32x3_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst3_vnum_u64))) +void svst3_vnum_u64(svbool_t, uint64_t *, int64_t, svuint64x3_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst3_vnum_u16))) +void svst3_vnum_u16(svbool_t, uint16_t *, int64_t, svuint16x3_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst3_vnum_s8))) +void svst3_vnum_s8(svbool_t, int8_t *, int64_t, svint8x3_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst3_vnum_f64))) +void svst3_vnum_f64(svbool_t, float64_t *, int64_t, svfloat64x3_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst3_vnum_f32))) +void svst3_vnum_f32(svbool_t, float32_t *, int64_t, svfloat32x3_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst3_vnum_f16))) +void svst3_vnum_f16(svbool_t, float16_t *, int64_t, svfloat16x3_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst3_vnum_s32))) +void svst3_vnum_s32(svbool_t, int32_t *, int64_t, svint32x3_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst3_vnum_s64))) +void svst3_vnum_s64(svbool_t, int64_t *, int64_t, svint64x3_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst3_vnum_s16))) +void svst3_vnum_s16(svbool_t, int16_t *, int64_t, svint16x3_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst4_u8))) +void svst4_u8(svbool_t, uint8_t *, svuint8x4_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst4_u32))) +void svst4_u32(svbool_t, uint32_t *, svuint32x4_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst4_u64))) +void svst4_u64(svbool_t, uint64_t *, svuint64x4_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst4_u16))) +void svst4_u16(svbool_t, uint16_t *, svuint16x4_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst4_s8))) +void svst4_s8(svbool_t, int8_t *, svint8x4_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst4_f64))) +void svst4_f64(svbool_t, float64_t *, svfloat64x4_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst4_f32))) +void svst4_f32(svbool_t, float32_t *, svfloat32x4_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst4_f16))) +void svst4_f16(svbool_t, float16_t *, svfloat16x4_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst4_s32))) +void svst4_s32(svbool_t, int32_t *, svint32x4_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst4_s64))) +void svst4_s64(svbool_t, int64_t *, svint64x4_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst4_s16))) +void svst4_s16(svbool_t, int16_t *, svint16x4_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst4_vnum_u8))) +void svst4_vnum_u8(svbool_t, uint8_t *, int64_t, svuint8x4_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst4_vnum_u32))) +void svst4_vnum_u32(svbool_t, uint32_t *, int64_t, svuint32x4_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst4_vnum_u64))) +void svst4_vnum_u64(svbool_t, uint64_t *, int64_t, svuint64x4_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst4_vnum_u16))) +void svst4_vnum_u16(svbool_t, uint16_t *, int64_t, svuint16x4_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst4_vnum_s8))) +void svst4_vnum_s8(svbool_t, int8_t *, int64_t, svint8x4_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst4_vnum_f64))) +void svst4_vnum_f64(svbool_t, float64_t *, int64_t, svfloat64x4_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst4_vnum_f32))) +void svst4_vnum_f32(svbool_t, float32_t *, int64_t, svfloat32x4_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst4_vnum_f16))) +void svst4_vnum_f16(svbool_t, float16_t *, int64_t, svfloat16x4_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst4_vnum_s32))) +void svst4_vnum_s32(svbool_t, int32_t *, int64_t, svint32x4_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst4_vnum_s64))) +void svst4_vnum_s64(svbool_t, int64_t *, int64_t, svint64x4_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst4_vnum_s16))) +void svst4_vnum_s16(svbool_t, int16_t *, int64_t, svint16x4_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_u8))) +void svstnt1_u8(svbool_t, uint8_t *, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_u32))) +void svstnt1_u32(svbool_t, uint32_t *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_u64))) +void svstnt1_u64(svbool_t, uint64_t *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_u16))) +void svstnt1_u16(svbool_t, uint16_t *, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_s8))) +void svstnt1_s8(svbool_t, int8_t *, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_f64))) +void svstnt1_f64(svbool_t, float64_t *, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_f32))) +void svstnt1_f32(svbool_t, float32_t *, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_f16))) +void svstnt1_f16(svbool_t, float16_t *, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_s32))) +void svstnt1_s32(svbool_t, int32_t *, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_s64))) +void svstnt1_s64(svbool_t, int64_t *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_s16))) +void svstnt1_s16(svbool_t, int16_t *, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_vnum_u8))) +void svstnt1_vnum_u8(svbool_t, uint8_t *, int64_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_vnum_u32))) +void svstnt1_vnum_u32(svbool_t, uint32_t *, int64_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_vnum_u64))) +void svstnt1_vnum_u64(svbool_t, uint64_t *, int64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_vnum_u16))) +void svstnt1_vnum_u16(svbool_t, uint16_t *, int64_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_vnum_s8))) +void svstnt1_vnum_s8(svbool_t, int8_t *, int64_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_vnum_f64))) +void svstnt1_vnum_f64(svbool_t, float64_t *, int64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_vnum_f32))) +void svstnt1_vnum_f32(svbool_t, float32_t *, int64_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_vnum_f16))) +void svstnt1_vnum_f16(svbool_t, float16_t *, int64_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_vnum_s32))) +void svstnt1_vnum_s32(svbool_t, int32_t *, int64_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_vnum_s64))) +void svstnt1_vnum_s64(svbool_t, int64_t *, int64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_vnum_s16))) +void svstnt1_vnum_s16(svbool_t, int16_t *, int64_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_n_f64_m))) +svfloat64_t svsub_n_f64_m(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_n_f32_m))) +svfloat32_t svsub_n_f32_m(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_n_f16_m))) +svfloat16_t svsub_n_f16_m(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_n_f64_x))) +svfloat64_t svsub_n_f64_x(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_n_f32_x))) +svfloat32_t svsub_n_f32_x(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_n_f16_x))) +svfloat16_t svsub_n_f16_x(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_n_f64_z))) +svfloat64_t svsub_n_f64_z(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_n_f32_z))) +svfloat32_t svsub_n_f32_z(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_n_f16_z))) +svfloat16_t svsub_n_f16_z(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_n_u8_m))) +svuint8_t svsub_n_u8_m(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_n_u32_m))) +svuint32_t svsub_n_u32_m(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_n_u64_m))) +svuint64_t svsub_n_u64_m(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_n_u16_m))) +svuint16_t svsub_n_u16_m(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_n_s8_m))) +svint8_t svsub_n_s8_m(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_n_s32_m))) +svint32_t svsub_n_s32_m(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_n_s64_m))) +svint64_t svsub_n_s64_m(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_n_s16_m))) +svint16_t svsub_n_s16_m(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_n_u8_x))) +svuint8_t svsub_n_u8_x(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_n_u32_x))) +svuint32_t svsub_n_u32_x(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_n_u64_x))) +svuint64_t svsub_n_u64_x(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_n_u16_x))) +svuint16_t svsub_n_u16_x(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_n_s8_x))) +svint8_t svsub_n_s8_x(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_n_s32_x))) +svint32_t svsub_n_s32_x(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_n_s64_x))) +svint64_t svsub_n_s64_x(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_n_s16_x))) +svint16_t svsub_n_s16_x(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_n_u8_z))) +svuint8_t svsub_n_u8_z(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_n_u32_z))) +svuint32_t svsub_n_u32_z(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_n_u64_z))) +svuint64_t svsub_n_u64_z(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_n_u16_z))) +svuint16_t svsub_n_u16_z(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_n_s8_z))) +svint8_t svsub_n_s8_z(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_n_s32_z))) +svint32_t svsub_n_s32_z(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_n_s64_z))) +svint64_t svsub_n_s64_z(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_n_s16_z))) +svint16_t svsub_n_s16_z(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_f64_m))) +svfloat64_t svsub_f64_m(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_f32_m))) +svfloat32_t svsub_f32_m(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_f16_m))) +svfloat16_t svsub_f16_m(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_f64_x))) +svfloat64_t svsub_f64_x(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_f32_x))) +svfloat32_t svsub_f32_x(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_f16_x))) +svfloat16_t svsub_f16_x(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_f64_z))) +svfloat64_t svsub_f64_z(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_f32_z))) +svfloat32_t svsub_f32_z(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_f16_z))) +svfloat16_t svsub_f16_z(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_u8_m))) +svuint8_t svsub_u8_m(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_u32_m))) +svuint32_t svsub_u32_m(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_u64_m))) +svuint64_t svsub_u64_m(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_u16_m))) +svuint16_t svsub_u16_m(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_s8_m))) +svint8_t svsub_s8_m(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_s32_m))) +svint32_t svsub_s32_m(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_s64_m))) +svint64_t svsub_s64_m(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_s16_m))) +svint16_t svsub_s16_m(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_u8_x))) +svuint8_t svsub_u8_x(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_u32_x))) +svuint32_t svsub_u32_x(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_u64_x))) +svuint64_t svsub_u64_x(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_u16_x))) +svuint16_t svsub_u16_x(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_s8_x))) +svint8_t svsub_s8_x(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_s32_x))) +svint32_t svsub_s32_x(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_s64_x))) +svint64_t svsub_s64_x(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_s16_x))) +svint16_t svsub_s16_x(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_u8_z))) +svuint8_t svsub_u8_z(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_u32_z))) +svuint32_t svsub_u32_z(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_u64_z))) +svuint64_t svsub_u64_z(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_u16_z))) +svuint16_t svsub_u16_z(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_s8_z))) +svint8_t svsub_s8_z(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_s32_z))) +svint32_t svsub_s32_z(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_s64_z))) +svint64_t svsub_s64_z(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsub_s16_z))) +svint16_t svsub_s16_z(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_n_f64_m))) +svfloat64_t svsubr_n_f64_m(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_n_f32_m))) +svfloat32_t svsubr_n_f32_m(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_n_f16_m))) +svfloat16_t svsubr_n_f16_m(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_n_f64_x))) +svfloat64_t svsubr_n_f64_x(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_n_f32_x))) +svfloat32_t svsubr_n_f32_x(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_n_f16_x))) +svfloat16_t svsubr_n_f16_x(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_n_f64_z))) +svfloat64_t svsubr_n_f64_z(svbool_t, svfloat64_t, float64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_n_f32_z))) +svfloat32_t svsubr_n_f32_z(svbool_t, svfloat32_t, float32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_n_f16_z))) +svfloat16_t svsubr_n_f16_z(svbool_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_n_u8_m))) +svuint8_t svsubr_n_u8_m(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_n_u32_m))) +svuint32_t svsubr_n_u32_m(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_n_u64_m))) +svuint64_t svsubr_n_u64_m(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_n_u16_m))) +svuint16_t svsubr_n_u16_m(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_n_s8_m))) +svint8_t svsubr_n_s8_m(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_n_s32_m))) +svint32_t svsubr_n_s32_m(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_n_s64_m))) +svint64_t svsubr_n_s64_m(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_n_s16_m))) +svint16_t svsubr_n_s16_m(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_n_u8_x))) +svuint8_t svsubr_n_u8_x(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_n_u32_x))) +svuint32_t svsubr_n_u32_x(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_n_u64_x))) +svuint64_t svsubr_n_u64_x(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_n_u16_x))) +svuint16_t svsubr_n_u16_x(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_n_s8_x))) +svint8_t svsubr_n_s8_x(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_n_s32_x))) +svint32_t svsubr_n_s32_x(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_n_s64_x))) +svint64_t svsubr_n_s64_x(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_n_s16_x))) +svint16_t svsubr_n_s16_x(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_n_u8_z))) +svuint8_t svsubr_n_u8_z(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_n_u32_z))) +svuint32_t svsubr_n_u32_z(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_n_u64_z))) +svuint64_t svsubr_n_u64_z(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_n_u16_z))) +svuint16_t svsubr_n_u16_z(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_n_s8_z))) +svint8_t svsubr_n_s8_z(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_n_s32_z))) +svint32_t svsubr_n_s32_z(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_n_s64_z))) +svint64_t svsubr_n_s64_z(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_n_s16_z))) +svint16_t svsubr_n_s16_z(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_f64_m))) +svfloat64_t svsubr_f64_m(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_f32_m))) +svfloat32_t svsubr_f32_m(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_f16_m))) +svfloat16_t svsubr_f16_m(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_f64_x))) +svfloat64_t svsubr_f64_x(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_f32_x))) +svfloat32_t svsubr_f32_x(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_f16_x))) +svfloat16_t svsubr_f16_x(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_f64_z))) +svfloat64_t svsubr_f64_z(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_f32_z))) +svfloat32_t svsubr_f32_z(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_f16_z))) +svfloat16_t svsubr_f16_z(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_u8_m))) +svuint8_t svsubr_u8_m(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_u32_m))) +svuint32_t svsubr_u32_m(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_u64_m))) +svuint64_t svsubr_u64_m(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_u16_m))) +svuint16_t svsubr_u16_m(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_s8_m))) +svint8_t svsubr_s8_m(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_s32_m))) +svint32_t svsubr_s32_m(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_s64_m))) +svint64_t svsubr_s64_m(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_s16_m))) +svint16_t svsubr_s16_m(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_u8_x))) +svuint8_t svsubr_u8_x(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_u32_x))) +svuint32_t svsubr_u32_x(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_u64_x))) +svuint64_t svsubr_u64_x(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_u16_x))) +svuint16_t svsubr_u16_x(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_s8_x))) +svint8_t svsubr_s8_x(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_s32_x))) +svint32_t svsubr_s32_x(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_s64_x))) +svint64_t svsubr_s64_x(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_s16_x))) +svint16_t svsubr_s16_x(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_u8_z))) +svuint8_t svsubr_u8_z(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_u32_z))) +svuint32_t svsubr_u32_z(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_u64_z))) +svuint64_t svsubr_u64_z(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_u16_z))) +svuint16_t svsubr_u16_z(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_s8_z))) +svint8_t svsubr_s8_z(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_s32_z))) +svint32_t svsubr_s32_z(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_s64_z))) +svint64_t svsubr_s64_z(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubr_s16_z))) +svint16_t svsubr_s16_z(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtbl_u8))) +svuint8_t svtbl_u8(svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtbl_u32))) +svuint32_t svtbl_u32(svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtbl_u64))) +svuint64_t svtbl_u64(svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtbl_u16))) +svuint16_t svtbl_u16(svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtbl_s8))) +svint8_t svtbl_s8(svint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtbl_f64))) +svfloat64_t svtbl_f64(svfloat64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtbl_f32))) +svfloat32_t svtbl_f32(svfloat32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtbl_f16))) +svfloat16_t svtbl_f16(svfloat16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtbl_s32))) +svint32_t svtbl_s32(svint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtbl_s64))) +svint64_t svtbl_s64(svint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtbl_s16))) +svint16_t svtbl_s16(svint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtmad_f64))) +svfloat64_t svtmad_f64(svfloat64_t, svfloat64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtmad_f32))) +svfloat32_t svtmad_f32(svfloat32_t, svfloat32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtmad_f16))) +svfloat16_t svtmad_f16(svfloat16_t, svfloat16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn1_u8))) +svuint8_t svtrn1_u8(svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn1_u32))) +svuint32_t svtrn1_u32(svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn1_u64))) +svuint64_t svtrn1_u64(svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn1_u16))) +svuint16_t svtrn1_u16(svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn1_s8))) +svint8_t svtrn1_s8(svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn1_f64))) +svfloat64_t svtrn1_f64(svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn1_f32))) +svfloat32_t svtrn1_f32(svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn1_f16))) +svfloat16_t svtrn1_f16(svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn1_s32))) +svint32_t svtrn1_s32(svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn1_s64))) +svint64_t svtrn1_s64(svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn1_s16))) +svint16_t svtrn1_s16(svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn1_b8))) +svbool_t svtrn1_b8(svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn1_b32))) +svbool_t svtrn1_b32(svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn1_b64))) +svbool_t svtrn1_b64(svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn1_b16))) +svbool_t svtrn1_b16(svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn2_u8))) +svuint8_t svtrn2_u8(svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn2_u32))) +svuint32_t svtrn2_u32(svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn2_u64))) +svuint64_t svtrn2_u64(svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn2_u16))) +svuint16_t svtrn2_u16(svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn2_s8))) +svint8_t svtrn2_s8(svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn2_f64))) +svfloat64_t svtrn2_f64(svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn2_f32))) +svfloat32_t svtrn2_f32(svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn2_f16))) +svfloat16_t svtrn2_f16(svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn2_s32))) +svint32_t svtrn2_s32(svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn2_s64))) +svint64_t svtrn2_s64(svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn2_s16))) +svint16_t svtrn2_s16(svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn2_b8))) +svbool_t svtrn2_b8(svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn2_b32))) +svbool_t svtrn2_b32(svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn2_b64))) +svbool_t svtrn2_b64(svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn2_b16))) +svbool_t svtrn2_b16(svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtsmul_f64))) +svfloat64_t svtsmul_f64(svfloat64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtsmul_f32))) +svfloat32_t svtsmul_f32(svfloat32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtsmul_f16))) +svfloat16_t svtsmul_f16(svfloat16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtssel_f64))) +svfloat64_t svtssel_f64(svfloat64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtssel_f32))) +svfloat32_t svtssel_f32(svfloat32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtssel_f16))) +svfloat16_t svtssel_f16(svfloat16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef2_u8))) +svuint8x2_t svundef2_u8(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef2_u32))) +svuint32x2_t svundef2_u32(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef2_u64))) +svuint64x2_t svundef2_u64(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef2_u16))) +svuint16x2_t svundef2_u16(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef2_s8))) +svint8x2_t svundef2_s8(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef2_f64))) +svfloat64x2_t svundef2_f64(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef2_f32))) +svfloat32x2_t svundef2_f32(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef2_f16))) +svfloat16x2_t svundef2_f16(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef2_s32))) +svint32x2_t svundef2_s32(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef2_s64))) +svint64x2_t svundef2_s64(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef2_s16))) +svint16x2_t svundef2_s16(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef3_u8))) +svuint8x3_t svundef3_u8(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef3_u32))) +svuint32x3_t svundef3_u32(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef3_u64))) +svuint64x3_t svundef3_u64(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef3_u16))) +svuint16x3_t svundef3_u16(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef3_s8))) +svint8x3_t svundef3_s8(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef3_f64))) +svfloat64x3_t svundef3_f64(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef3_f32))) +svfloat32x3_t svundef3_f32(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef3_f16))) +svfloat16x3_t svundef3_f16(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef3_s32))) +svint32x3_t svundef3_s32(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef3_s64))) +svint64x3_t svundef3_s64(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef3_s16))) +svint16x3_t svundef3_s16(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef4_u8))) +svuint8x4_t svundef4_u8(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef4_u32))) +svuint32x4_t svundef4_u32(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef4_u64))) +svuint64x4_t svundef4_u64(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef4_u16))) +svuint16x4_t svundef4_u16(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef4_s8))) +svint8x4_t svundef4_s8(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef4_f64))) +svfloat64x4_t svundef4_f64(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef4_f32))) +svfloat32x4_t svundef4_f32(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef4_f16))) +svfloat16x4_t svundef4_f16(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef4_s32))) +svint32x4_t svundef4_s32(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef4_s64))) +svint64x4_t svundef4_s64(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef4_s16))) +svint16x4_t svundef4_s16(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef_u8))) +svuint8_t svundef_u8(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef_u32))) +svuint32_t svundef_u32(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef_u64))) +svuint64_t svundef_u64(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef_u16))) +svuint16_t svundef_u16(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef_s8))) +svint8_t svundef_s8(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef_f64))) +svfloat64_t svundef_f64(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef_f32))) +svfloat32_t svundef_f32(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef_f16))) +svfloat16_t svundef_f16(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef_s32))) +svint32_t svundef_s32(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef_s64))) +svint64_t svundef_s64(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef_s16))) +svint16_t svundef_s16(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svunpkhi_b))) +svbool_t svunpkhi_b(svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svunpkhi_s32))) +svint32_t svunpkhi_s32(svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svunpkhi_s64))) +svint64_t svunpkhi_s64(svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svunpkhi_s16))) +svint16_t svunpkhi_s16(svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svunpkhi_u32))) +svuint32_t svunpkhi_u32(svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svunpkhi_u64))) +svuint64_t svunpkhi_u64(svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svunpkhi_u16))) +svuint16_t svunpkhi_u16(svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svunpklo_b))) +svbool_t svunpklo_b(svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svunpklo_s32))) +svint32_t svunpklo_s32(svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svunpklo_s64))) +svint64_t svunpklo_s64(svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svunpklo_s16))) +svint16_t svunpklo_s16(svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svunpklo_u32))) +svuint32_t svunpklo_u32(svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svunpklo_u64))) +svuint64_t svunpklo_u64(svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svunpklo_u16))) +svuint16_t svunpklo_u16(svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp1_u8))) +svuint8_t svuzp1_u8(svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp1_u32))) +svuint32_t svuzp1_u32(svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp1_u64))) +svuint64_t svuzp1_u64(svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp1_u16))) +svuint16_t svuzp1_u16(svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp1_s8))) +svint8_t svuzp1_s8(svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp1_f64))) +svfloat64_t svuzp1_f64(svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp1_f32))) +svfloat32_t svuzp1_f32(svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp1_f16))) +svfloat16_t svuzp1_f16(svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp1_s32))) +svint32_t svuzp1_s32(svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp1_s64))) +svint64_t svuzp1_s64(svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp1_s16))) +svint16_t svuzp1_s16(svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp1_b8))) +svbool_t svuzp1_b8(svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp1_b32))) +svbool_t svuzp1_b32(svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp1_b64))) +svbool_t svuzp1_b64(svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp1_b16))) +svbool_t svuzp1_b16(svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp2_u8))) +svuint8_t svuzp2_u8(svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp2_u32))) +svuint32_t svuzp2_u32(svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp2_u64))) +svuint64_t svuzp2_u64(svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp2_u16))) +svuint16_t svuzp2_u16(svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp2_s8))) +svint8_t svuzp2_s8(svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp2_f64))) +svfloat64_t svuzp2_f64(svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp2_f32))) +svfloat32_t svuzp2_f32(svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp2_f16))) +svfloat16_t svuzp2_f16(svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp2_s32))) +svint32_t svuzp2_s32(svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp2_s64))) +svint64_t svuzp2_s64(svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp2_s16))) +svint16_t svuzp2_s16(svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp2_b8))) +svbool_t svuzp2_b8(svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp2_b32))) +svbool_t svuzp2_b32(svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp2_b64))) +svbool_t svuzp2_b64(svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp2_b16))) +svbool_t svuzp2_b16(svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilele_b8_s32))) +svbool_t svwhilele_b8_s32(int32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilele_b32_s32))) +svbool_t svwhilele_b32_s32(int32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilele_b64_s32))) +svbool_t svwhilele_b64_s32(int32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilele_b16_s32))) +svbool_t svwhilele_b16_s32(int32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilele_b8_s64))) +svbool_t svwhilele_b8_s64(int64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilele_b32_s64))) +svbool_t svwhilele_b32_s64(int64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilele_b64_s64))) +svbool_t svwhilele_b64_s64(int64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilele_b16_s64))) +svbool_t svwhilele_b16_s64(int64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilele_b8_u32))) +svbool_t svwhilele_b8_u32(uint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilele_b32_u32))) +svbool_t svwhilele_b32_u32(uint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilele_b64_u32))) +svbool_t svwhilele_b64_u32(uint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilele_b16_u32))) +svbool_t svwhilele_b16_u32(uint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilele_b8_u64))) +svbool_t svwhilele_b8_u64(uint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilele_b32_u64))) +svbool_t svwhilele_b32_u64(uint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilele_b64_u64))) +svbool_t svwhilele_b64_u64(uint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilele_b16_u64))) +svbool_t svwhilele_b16_u64(uint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilelt_b8_u32))) +svbool_t svwhilelt_b8_u32(uint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilelt_b32_u32))) +svbool_t svwhilelt_b32_u32(uint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilelt_b64_u32))) +svbool_t svwhilelt_b64_u32(uint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilelt_b16_u32))) +svbool_t svwhilelt_b16_u32(uint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilelt_b8_u64))) +svbool_t svwhilelt_b8_u64(uint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilelt_b32_u64))) +svbool_t svwhilelt_b32_u64(uint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilelt_b64_u64))) +svbool_t svwhilelt_b64_u64(uint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilelt_b16_u64))) +svbool_t svwhilelt_b16_u64(uint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilelt_b8_s32))) +svbool_t svwhilelt_b8_s32(int32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilelt_b32_s32))) +svbool_t svwhilelt_b32_s32(int32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilelt_b64_s32))) +svbool_t svwhilelt_b64_s32(int32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilelt_b16_s32))) +svbool_t svwhilelt_b16_s32(int32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilelt_b8_s64))) +svbool_t svwhilelt_b8_s64(int64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilelt_b32_s64))) +svbool_t svwhilelt_b32_s64(int64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilelt_b64_s64))) +svbool_t svwhilelt_b64_s64(int64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilelt_b16_s64))) +svbool_t svwhilelt_b16_s64(int64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwrffr))) +void svwrffr(svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip1_u8))) +svuint8_t svzip1_u8(svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip1_u32))) +svuint32_t svzip1_u32(svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip1_u64))) +svuint64_t svzip1_u64(svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip1_u16))) +svuint16_t svzip1_u16(svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip1_s8))) +svint8_t svzip1_s8(svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip1_f64))) +svfloat64_t svzip1_f64(svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip1_f32))) +svfloat32_t svzip1_f32(svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip1_f16))) +svfloat16_t svzip1_f16(svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip1_s32))) +svint32_t svzip1_s32(svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip1_s64))) +svint64_t svzip1_s64(svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip1_s16))) +svint16_t svzip1_s16(svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip1_b8))) +svbool_t svzip1_b8(svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip1_b32))) +svbool_t svzip1_b32(svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip1_b64))) +svbool_t svzip1_b64(svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip1_b16))) +svbool_t svzip1_b16(svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip2_u8))) +svuint8_t svzip2_u8(svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip2_u32))) +svuint32_t svzip2_u32(svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip2_u64))) +svuint64_t svzip2_u64(svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip2_u16))) +svuint16_t svzip2_u16(svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip2_s8))) +svint8_t svzip2_s8(svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip2_f64))) +svfloat64_t svzip2_f64(svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip2_f32))) +svfloat32_t svzip2_f32(svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip2_f16))) +svfloat16_t svzip2_f16(svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip2_s32))) +svint32_t svzip2_s32(svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip2_s64))) +svint64_t svzip2_s64(svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip2_s16))) +svint16_t svzip2_s16(svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip2_b8))) +svbool_t svzip2_b8(svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip2_b32))) +svbool_t svzip2_b32(svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip2_b64))) +svbool_t svzip2_b64(svbool_t, svbool_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip2_b16))) +svbool_t svzip2_b16(svbool_t, svbool_t); __aio __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_n_f64_m))) svfloat64_t svabd_m(svbool_t, svfloat64_t, float64_t); __aio __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabd_n_f32_m))) @@ -10009,7 +13830,8 @@ __aio __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorv_s64))) int64_t svorv(svbool_t, svint64_t); __aio __attribute__((__clang_arm_builtin_alias(__builtin_sve_svorv_s16))) int16_t svorv(svbool_t, svint16_t); -#define svpfalse(...) __builtin_sve_svpfalse_b(__VA_ARGS__) +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svpfalse_b))) +svbool_t svpfalse(); __aio __attribute__((__clang_arm_builtin_alias(__builtin_sve_svpfirst_b))) svbool_t svpfirst(svbool_t, svbool_t); __aio __attribute__((__clang_arm_builtin_alias(__builtin_sve_svprfb_gather_u32base))) @@ -11877,31 +15699,55 @@ svint64_t svzip2(svint64_t, svint64_t); __aio __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip2_s16))) svint16_t svzip2(svint16_t, svint16_t); -#if __ARM_FEATURE_SVE2_BITPERM -#define svbdep_n_u8(...) __builtin_sve_svbdep_n_u8(__VA_ARGS__) -#define svbdep_n_u32(...) __builtin_sve_svbdep_n_u32(__VA_ARGS__) -#define svbdep_n_u64(...) __builtin_sve_svbdep_n_u64(__VA_ARGS__) -#define svbdep_n_u16(...) __builtin_sve_svbdep_n_u16(__VA_ARGS__) -#define svbdep_u8(...) __builtin_sve_svbdep_u8(__VA_ARGS__) -#define svbdep_u32(...) __builtin_sve_svbdep_u32(__VA_ARGS__) -#define svbdep_u64(...) __builtin_sve_svbdep_u64(__VA_ARGS__) -#define svbdep_u16(...) __builtin_sve_svbdep_u16(__VA_ARGS__) -#define svbext_n_u8(...) __builtin_sve_svbext_n_u8(__VA_ARGS__) -#define svbext_n_u32(...) __builtin_sve_svbext_n_u32(__VA_ARGS__) -#define svbext_n_u64(...) __builtin_sve_svbext_n_u64(__VA_ARGS__) -#define svbext_n_u16(...) __builtin_sve_svbext_n_u16(__VA_ARGS__) -#define svbext_u8(...) __builtin_sve_svbext_u8(__VA_ARGS__) -#define svbext_u32(...) __builtin_sve_svbext_u32(__VA_ARGS__) -#define svbext_u64(...) __builtin_sve_svbext_u64(__VA_ARGS__) -#define svbext_u16(...) __builtin_sve_svbext_u16(__VA_ARGS__) -#define svbgrp_n_u8(...) __builtin_sve_svbgrp_n_u8(__VA_ARGS__) -#define svbgrp_n_u32(...) __builtin_sve_svbgrp_n_u32(__VA_ARGS__) -#define svbgrp_n_u64(...) __builtin_sve_svbgrp_n_u64(__VA_ARGS__) -#define svbgrp_n_u16(...) __builtin_sve_svbgrp_n_u16(__VA_ARGS__) -#define svbgrp_u8(...) __builtin_sve_svbgrp_u8(__VA_ARGS__) -#define svbgrp_u32(...) __builtin_sve_svbgrp_u32(__VA_ARGS__) -#define svbgrp_u64(...) __builtin_sve_svbgrp_u64(__VA_ARGS__) -#define svbgrp_u16(...) __builtin_sve_svbgrp_u16(__VA_ARGS__) +#if defined (__ARM_FEATURE_SVE2_BITPERM) +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbdep_n_u8))) +svuint8_t svbdep_n_u8(svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbdep_n_u32))) +svuint32_t svbdep_n_u32(svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbdep_n_u64))) +svuint64_t svbdep_n_u64(svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbdep_n_u16))) +svuint16_t svbdep_n_u16(svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbdep_u8))) +svuint8_t svbdep_u8(svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbdep_u32))) +svuint32_t svbdep_u32(svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbdep_u64))) +svuint64_t svbdep_u64(svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbdep_u16))) +svuint16_t svbdep_u16(svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbext_n_u8))) +svuint8_t svbext_n_u8(svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbext_n_u32))) +svuint32_t svbext_n_u32(svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbext_n_u64))) +svuint64_t svbext_n_u64(svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbext_n_u16))) +svuint16_t svbext_n_u16(svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbext_u8))) +svuint8_t svbext_u8(svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbext_u32))) +svuint32_t svbext_u32(svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbext_u64))) +svuint64_t svbext_u64(svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbext_u16))) +svuint16_t svbext_u16(svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbgrp_n_u8))) +svuint8_t svbgrp_n_u8(svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbgrp_n_u32))) +svuint32_t svbgrp_n_u32(svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbgrp_n_u64))) +svuint64_t svbgrp_n_u64(svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbgrp_n_u16))) +svuint16_t svbgrp_n_u16(svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbgrp_u8))) +svuint8_t svbgrp_u8(svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbgrp_u32))) +svuint32_t svbgrp_u32(svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbgrp_u64))) +svuint64_t svbgrp_u64(svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbgrp_u16))) +svuint16_t svbgrp_u16(svuint16_t, svuint16_t); __aio __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbdep_n_u8))) svuint8_t svbdep(svuint8_t, uint8_t); __aio __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbdep_n_u32))) @@ -11950,1866 +15796,3723 @@ __aio __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbgrp_u64))) svuint64_t svbgrp(svuint64_t, svuint64_t); __aio __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbgrp_u16))) svuint16_t svbgrp(svuint16_t, svuint16_t); -#endif //__ARM_FEATURE_SVE2_BITPERM +#endif //defined (__ARM_FEATURE_SVE2_BITPERM) #if defined(__ARM_FEATURE_SVE2) -#define svaba_n_s8(...) __builtin_sve_svaba_n_s8(__VA_ARGS__) -#define svaba_n_s32(...) __builtin_sve_svaba_n_s32(__VA_ARGS__) -#define svaba_n_s64(...) __builtin_sve_svaba_n_s64(__VA_ARGS__) -#define svaba_n_s16(...) __builtin_sve_svaba_n_s16(__VA_ARGS__) -#define svaba_n_u8(...) __builtin_sve_svaba_n_u8(__VA_ARGS__) -#define svaba_n_u32(...) __builtin_sve_svaba_n_u32(__VA_ARGS__) -#define svaba_n_u64(...) __builtin_sve_svaba_n_u64(__VA_ARGS__) -#define svaba_n_u16(...) __builtin_sve_svaba_n_u16(__VA_ARGS__) -#define svaba_s8(...) __builtin_sve_svaba_s8(__VA_ARGS__) -#define svaba_s32(...) __builtin_sve_svaba_s32(__VA_ARGS__) -#define svaba_s64(...) __builtin_sve_svaba_s64(__VA_ARGS__) -#define svaba_s16(...) __builtin_sve_svaba_s16(__VA_ARGS__) -#define svaba_u8(...) __builtin_sve_svaba_u8(__VA_ARGS__) -#define svaba_u32(...) __builtin_sve_svaba_u32(__VA_ARGS__) -#define svaba_u64(...) __builtin_sve_svaba_u64(__VA_ARGS__) -#define svaba_u16(...) __builtin_sve_svaba_u16(__VA_ARGS__) -#define svabalb_n_s32(...) __builtin_sve_svabalb_n_s32(__VA_ARGS__) -#define svabalb_n_s64(...) __builtin_sve_svabalb_n_s64(__VA_ARGS__) -#define svabalb_n_s16(...) __builtin_sve_svabalb_n_s16(__VA_ARGS__) -#define svabalb_n_u32(...) __builtin_sve_svabalb_n_u32(__VA_ARGS__) -#define svabalb_n_u64(...) __builtin_sve_svabalb_n_u64(__VA_ARGS__) -#define svabalb_n_u16(...) __builtin_sve_svabalb_n_u16(__VA_ARGS__) -#define svabalb_s32(...) __builtin_sve_svabalb_s32(__VA_ARGS__) -#define svabalb_s64(...) __builtin_sve_svabalb_s64(__VA_ARGS__) -#define svabalb_s16(...) __builtin_sve_svabalb_s16(__VA_ARGS__) -#define svabalb_u32(...) __builtin_sve_svabalb_u32(__VA_ARGS__) -#define svabalb_u64(...) __builtin_sve_svabalb_u64(__VA_ARGS__) -#define svabalb_u16(...) __builtin_sve_svabalb_u16(__VA_ARGS__) -#define svabalt_n_s32(...) __builtin_sve_svabalt_n_s32(__VA_ARGS__) -#define svabalt_n_s64(...) __builtin_sve_svabalt_n_s64(__VA_ARGS__) -#define svabalt_n_s16(...) __builtin_sve_svabalt_n_s16(__VA_ARGS__) -#define svabalt_n_u32(...) __builtin_sve_svabalt_n_u32(__VA_ARGS__) -#define svabalt_n_u64(...) __builtin_sve_svabalt_n_u64(__VA_ARGS__) -#define svabalt_n_u16(...) __builtin_sve_svabalt_n_u16(__VA_ARGS__) -#define svabalt_s32(...) __builtin_sve_svabalt_s32(__VA_ARGS__) -#define svabalt_s64(...) __builtin_sve_svabalt_s64(__VA_ARGS__) -#define svabalt_s16(...) __builtin_sve_svabalt_s16(__VA_ARGS__) -#define svabalt_u32(...) __builtin_sve_svabalt_u32(__VA_ARGS__) -#define svabalt_u64(...) __builtin_sve_svabalt_u64(__VA_ARGS__) -#define svabalt_u16(...) __builtin_sve_svabalt_u16(__VA_ARGS__) -#define svabdlb_n_s32(...) __builtin_sve_svabdlb_n_s32(__VA_ARGS__) -#define svabdlb_n_s64(...) __builtin_sve_svabdlb_n_s64(__VA_ARGS__) -#define svabdlb_n_s16(...) __builtin_sve_svabdlb_n_s16(__VA_ARGS__) -#define svabdlb_n_u32(...) __builtin_sve_svabdlb_n_u32(__VA_ARGS__) -#define svabdlb_n_u64(...) __builtin_sve_svabdlb_n_u64(__VA_ARGS__) -#define svabdlb_n_u16(...) __builtin_sve_svabdlb_n_u16(__VA_ARGS__) -#define svabdlb_s32(...) __builtin_sve_svabdlb_s32(__VA_ARGS__) -#define svabdlb_s64(...) __builtin_sve_svabdlb_s64(__VA_ARGS__) -#define svabdlb_s16(...) __builtin_sve_svabdlb_s16(__VA_ARGS__) -#define svabdlb_u32(...) __builtin_sve_svabdlb_u32(__VA_ARGS__) -#define svabdlb_u64(...) __builtin_sve_svabdlb_u64(__VA_ARGS__) -#define svabdlb_u16(...) __builtin_sve_svabdlb_u16(__VA_ARGS__) -#define svabdlt_n_s32(...) __builtin_sve_svabdlt_n_s32(__VA_ARGS__) -#define svabdlt_n_s64(...) __builtin_sve_svabdlt_n_s64(__VA_ARGS__) -#define svabdlt_n_s16(...) __builtin_sve_svabdlt_n_s16(__VA_ARGS__) -#define svabdlt_n_u32(...) __builtin_sve_svabdlt_n_u32(__VA_ARGS__) -#define svabdlt_n_u64(...) __builtin_sve_svabdlt_n_u64(__VA_ARGS__) -#define svabdlt_n_u16(...) __builtin_sve_svabdlt_n_u16(__VA_ARGS__) -#define svabdlt_s32(...) __builtin_sve_svabdlt_s32(__VA_ARGS__) -#define svabdlt_s64(...) __builtin_sve_svabdlt_s64(__VA_ARGS__) -#define svabdlt_s16(...) __builtin_sve_svabdlt_s16(__VA_ARGS__) -#define svabdlt_u32(...) __builtin_sve_svabdlt_u32(__VA_ARGS__) -#define svabdlt_u64(...) __builtin_sve_svabdlt_u64(__VA_ARGS__) -#define svabdlt_u16(...) __builtin_sve_svabdlt_u16(__VA_ARGS__) -#define svadalp_s32_m(...) __builtin_sve_svadalp_s32_m(__VA_ARGS__) -#define svadalp_s64_m(...) __builtin_sve_svadalp_s64_m(__VA_ARGS__) -#define svadalp_s16_m(...) __builtin_sve_svadalp_s16_m(__VA_ARGS__) -#define svadalp_s32_x(...) __builtin_sve_svadalp_s32_x(__VA_ARGS__) -#define svadalp_s64_x(...) __builtin_sve_svadalp_s64_x(__VA_ARGS__) -#define svadalp_s16_x(...) __builtin_sve_svadalp_s16_x(__VA_ARGS__) -#define svadalp_s32_z(...) __builtin_sve_svadalp_s32_z(__VA_ARGS__) -#define svadalp_s64_z(...) __builtin_sve_svadalp_s64_z(__VA_ARGS__) -#define svadalp_s16_z(...) __builtin_sve_svadalp_s16_z(__VA_ARGS__) -#define svadalp_u32_m(...) __builtin_sve_svadalp_u32_m(__VA_ARGS__) -#define svadalp_u64_m(...) __builtin_sve_svadalp_u64_m(__VA_ARGS__) -#define svadalp_u16_m(...) __builtin_sve_svadalp_u16_m(__VA_ARGS__) -#define svadalp_u32_x(...) __builtin_sve_svadalp_u32_x(__VA_ARGS__) -#define svadalp_u64_x(...) __builtin_sve_svadalp_u64_x(__VA_ARGS__) -#define svadalp_u16_x(...) __builtin_sve_svadalp_u16_x(__VA_ARGS__) -#define svadalp_u32_z(...) __builtin_sve_svadalp_u32_z(__VA_ARGS__) -#define svadalp_u64_z(...) __builtin_sve_svadalp_u64_z(__VA_ARGS__) -#define svadalp_u16_z(...) __builtin_sve_svadalp_u16_z(__VA_ARGS__) -#define svadclb_n_u32(...) __builtin_sve_svadclb_n_u32(__VA_ARGS__) -#define svadclb_n_u64(...) __builtin_sve_svadclb_n_u64(__VA_ARGS__) -#define svadclb_u32(...) __builtin_sve_svadclb_u32(__VA_ARGS__) -#define svadclb_u64(...) __builtin_sve_svadclb_u64(__VA_ARGS__) -#define svadclt_n_u32(...) __builtin_sve_svadclt_n_u32(__VA_ARGS__) -#define svadclt_n_u64(...) __builtin_sve_svadclt_n_u64(__VA_ARGS__) -#define svadclt_u32(...) __builtin_sve_svadclt_u32(__VA_ARGS__) -#define svadclt_u64(...) __builtin_sve_svadclt_u64(__VA_ARGS__) -#define svaddhnb_n_u32(...) __builtin_sve_svaddhnb_n_u32(__VA_ARGS__) -#define svaddhnb_n_u64(...) __builtin_sve_svaddhnb_n_u64(__VA_ARGS__) -#define svaddhnb_n_u16(...) __builtin_sve_svaddhnb_n_u16(__VA_ARGS__) -#define svaddhnb_n_s32(...) __builtin_sve_svaddhnb_n_s32(__VA_ARGS__) -#define svaddhnb_n_s64(...) __builtin_sve_svaddhnb_n_s64(__VA_ARGS__) -#define svaddhnb_n_s16(...) __builtin_sve_svaddhnb_n_s16(__VA_ARGS__) -#define svaddhnb_u32(...) __builtin_sve_svaddhnb_u32(__VA_ARGS__) -#define svaddhnb_u64(...) __builtin_sve_svaddhnb_u64(__VA_ARGS__) -#define svaddhnb_u16(...) __builtin_sve_svaddhnb_u16(__VA_ARGS__) -#define svaddhnb_s32(...) __builtin_sve_svaddhnb_s32(__VA_ARGS__) -#define svaddhnb_s64(...) __builtin_sve_svaddhnb_s64(__VA_ARGS__) -#define svaddhnb_s16(...) __builtin_sve_svaddhnb_s16(__VA_ARGS__) -#define svaddhnt_n_u32(...) __builtin_sve_svaddhnt_n_u32(__VA_ARGS__) -#define svaddhnt_n_u64(...) __builtin_sve_svaddhnt_n_u64(__VA_ARGS__) -#define svaddhnt_n_u16(...) __builtin_sve_svaddhnt_n_u16(__VA_ARGS__) -#define svaddhnt_n_s32(...) __builtin_sve_svaddhnt_n_s32(__VA_ARGS__) -#define svaddhnt_n_s64(...) __builtin_sve_svaddhnt_n_s64(__VA_ARGS__) -#define svaddhnt_n_s16(...) __builtin_sve_svaddhnt_n_s16(__VA_ARGS__) -#define svaddhnt_u32(...) __builtin_sve_svaddhnt_u32(__VA_ARGS__) -#define svaddhnt_u64(...) __builtin_sve_svaddhnt_u64(__VA_ARGS__) -#define svaddhnt_u16(...) __builtin_sve_svaddhnt_u16(__VA_ARGS__) -#define svaddhnt_s32(...) __builtin_sve_svaddhnt_s32(__VA_ARGS__) -#define svaddhnt_s64(...) __builtin_sve_svaddhnt_s64(__VA_ARGS__) -#define svaddhnt_s16(...) __builtin_sve_svaddhnt_s16(__VA_ARGS__) -#define svaddlb_n_s32(...) __builtin_sve_svaddlb_n_s32(__VA_ARGS__) -#define svaddlb_n_s64(...) __builtin_sve_svaddlb_n_s64(__VA_ARGS__) -#define svaddlb_n_s16(...) __builtin_sve_svaddlb_n_s16(__VA_ARGS__) -#define svaddlb_n_u32(...) __builtin_sve_svaddlb_n_u32(__VA_ARGS__) -#define svaddlb_n_u64(...) __builtin_sve_svaddlb_n_u64(__VA_ARGS__) -#define svaddlb_n_u16(...) __builtin_sve_svaddlb_n_u16(__VA_ARGS__) -#define svaddlb_s32(...) __builtin_sve_svaddlb_s32(__VA_ARGS__) -#define svaddlb_s64(...) __builtin_sve_svaddlb_s64(__VA_ARGS__) -#define svaddlb_s16(...) __builtin_sve_svaddlb_s16(__VA_ARGS__) -#define svaddlb_u32(...) __builtin_sve_svaddlb_u32(__VA_ARGS__) -#define svaddlb_u64(...) __builtin_sve_svaddlb_u64(__VA_ARGS__) -#define svaddlb_u16(...) __builtin_sve_svaddlb_u16(__VA_ARGS__) -#define svaddlbt_n_s32(...) __builtin_sve_svaddlbt_n_s32(__VA_ARGS__) -#define svaddlbt_n_s64(...) __builtin_sve_svaddlbt_n_s64(__VA_ARGS__) -#define svaddlbt_n_s16(...) __builtin_sve_svaddlbt_n_s16(__VA_ARGS__) -#define svaddlbt_s32(...) __builtin_sve_svaddlbt_s32(__VA_ARGS__) -#define svaddlbt_s64(...) __builtin_sve_svaddlbt_s64(__VA_ARGS__) -#define svaddlbt_s16(...) __builtin_sve_svaddlbt_s16(__VA_ARGS__) -#define svaddlt_n_s32(...) __builtin_sve_svaddlt_n_s32(__VA_ARGS__) -#define svaddlt_n_s64(...) __builtin_sve_svaddlt_n_s64(__VA_ARGS__) -#define svaddlt_n_s16(...) __builtin_sve_svaddlt_n_s16(__VA_ARGS__) -#define svaddlt_n_u32(...) __builtin_sve_svaddlt_n_u32(__VA_ARGS__) -#define svaddlt_n_u64(...) __builtin_sve_svaddlt_n_u64(__VA_ARGS__) -#define svaddlt_n_u16(...) __builtin_sve_svaddlt_n_u16(__VA_ARGS__) -#define svaddlt_s32(...) __builtin_sve_svaddlt_s32(__VA_ARGS__) -#define svaddlt_s64(...) __builtin_sve_svaddlt_s64(__VA_ARGS__) -#define svaddlt_s16(...) __builtin_sve_svaddlt_s16(__VA_ARGS__) -#define svaddlt_u32(...) __builtin_sve_svaddlt_u32(__VA_ARGS__) -#define svaddlt_u64(...) __builtin_sve_svaddlt_u64(__VA_ARGS__) -#define svaddlt_u16(...) __builtin_sve_svaddlt_u16(__VA_ARGS__) -#define svaddp_f64_m(...) __builtin_sve_svaddp_f64_m(__VA_ARGS__) -#define svaddp_f32_m(...) __builtin_sve_svaddp_f32_m(__VA_ARGS__) -#define svaddp_f16_m(...) __builtin_sve_svaddp_f16_m(__VA_ARGS__) -#define svaddp_f64_x(...) __builtin_sve_svaddp_f64_x(__VA_ARGS__) -#define svaddp_f32_x(...) __builtin_sve_svaddp_f32_x(__VA_ARGS__) -#define svaddp_f16_x(...) __builtin_sve_svaddp_f16_x(__VA_ARGS__) -#define svaddp_u8_m(...) __builtin_sve_svaddp_u8_m(__VA_ARGS__) -#define svaddp_u32_m(...) __builtin_sve_svaddp_u32_m(__VA_ARGS__) -#define svaddp_u64_m(...) __builtin_sve_svaddp_u64_m(__VA_ARGS__) -#define svaddp_u16_m(...) __builtin_sve_svaddp_u16_m(__VA_ARGS__) -#define svaddp_s8_m(...) __builtin_sve_svaddp_s8_m(__VA_ARGS__) -#define svaddp_s32_m(...) __builtin_sve_svaddp_s32_m(__VA_ARGS__) -#define svaddp_s64_m(...) __builtin_sve_svaddp_s64_m(__VA_ARGS__) -#define svaddp_s16_m(...) __builtin_sve_svaddp_s16_m(__VA_ARGS__) -#define svaddp_u8_x(...) __builtin_sve_svaddp_u8_x(__VA_ARGS__) -#define svaddp_u32_x(...) __builtin_sve_svaddp_u32_x(__VA_ARGS__) -#define svaddp_u64_x(...) __builtin_sve_svaddp_u64_x(__VA_ARGS__) -#define svaddp_u16_x(...) __builtin_sve_svaddp_u16_x(__VA_ARGS__) -#define svaddp_s8_x(...) __builtin_sve_svaddp_s8_x(__VA_ARGS__) -#define svaddp_s32_x(...) __builtin_sve_svaddp_s32_x(__VA_ARGS__) -#define svaddp_s64_x(...) __builtin_sve_svaddp_s64_x(__VA_ARGS__) -#define svaddp_s16_x(...) __builtin_sve_svaddp_s16_x(__VA_ARGS__) -#define svaddwb_n_s32(...) __builtin_sve_svaddwb_n_s32(__VA_ARGS__) -#define svaddwb_n_s64(...) __builtin_sve_svaddwb_n_s64(__VA_ARGS__) -#define svaddwb_n_s16(...) __builtin_sve_svaddwb_n_s16(__VA_ARGS__) -#define svaddwb_n_u32(...) __builtin_sve_svaddwb_n_u32(__VA_ARGS__) -#define svaddwb_n_u64(...) __builtin_sve_svaddwb_n_u64(__VA_ARGS__) -#define svaddwb_n_u16(...) __builtin_sve_svaddwb_n_u16(__VA_ARGS__) -#define svaddwb_s32(...) __builtin_sve_svaddwb_s32(__VA_ARGS__) -#define svaddwb_s64(...) __builtin_sve_svaddwb_s64(__VA_ARGS__) -#define svaddwb_s16(...) __builtin_sve_svaddwb_s16(__VA_ARGS__) -#define svaddwb_u32(...) __builtin_sve_svaddwb_u32(__VA_ARGS__) -#define svaddwb_u64(...) __builtin_sve_svaddwb_u64(__VA_ARGS__) -#define svaddwb_u16(...) __builtin_sve_svaddwb_u16(__VA_ARGS__) -#define svaddwt_n_s32(...) __builtin_sve_svaddwt_n_s32(__VA_ARGS__) -#define svaddwt_n_s64(...) __builtin_sve_svaddwt_n_s64(__VA_ARGS__) -#define svaddwt_n_s16(...) __builtin_sve_svaddwt_n_s16(__VA_ARGS__) -#define svaddwt_n_u32(...) __builtin_sve_svaddwt_n_u32(__VA_ARGS__) -#define svaddwt_n_u64(...) __builtin_sve_svaddwt_n_u64(__VA_ARGS__) -#define svaddwt_n_u16(...) __builtin_sve_svaddwt_n_u16(__VA_ARGS__) -#define svaddwt_s32(...) __builtin_sve_svaddwt_s32(__VA_ARGS__) -#define svaddwt_s64(...) __builtin_sve_svaddwt_s64(__VA_ARGS__) -#define svaddwt_s16(...) __builtin_sve_svaddwt_s16(__VA_ARGS__) -#define svaddwt_u32(...) __builtin_sve_svaddwt_u32(__VA_ARGS__) -#define svaddwt_u64(...) __builtin_sve_svaddwt_u64(__VA_ARGS__) -#define svaddwt_u16(...) __builtin_sve_svaddwt_u16(__VA_ARGS__) -#define svbcax_n_u8(...) __builtin_sve_svbcax_n_u8(__VA_ARGS__) -#define svbcax_n_u32(...) __builtin_sve_svbcax_n_u32(__VA_ARGS__) -#define svbcax_n_u64(...) __builtin_sve_svbcax_n_u64(__VA_ARGS__) -#define svbcax_n_u16(...) __builtin_sve_svbcax_n_u16(__VA_ARGS__) -#define svbcax_n_s8(...) __builtin_sve_svbcax_n_s8(__VA_ARGS__) -#define svbcax_n_s32(...) __builtin_sve_svbcax_n_s32(__VA_ARGS__) -#define svbcax_n_s64(...) __builtin_sve_svbcax_n_s64(__VA_ARGS__) -#define svbcax_n_s16(...) __builtin_sve_svbcax_n_s16(__VA_ARGS__) -#define svbcax_u8(...) __builtin_sve_svbcax_u8(__VA_ARGS__) -#define svbcax_u32(...) __builtin_sve_svbcax_u32(__VA_ARGS__) -#define svbcax_u64(...) __builtin_sve_svbcax_u64(__VA_ARGS__) -#define svbcax_u16(...) __builtin_sve_svbcax_u16(__VA_ARGS__) -#define svbcax_s8(...) __builtin_sve_svbcax_s8(__VA_ARGS__) -#define svbcax_s32(...) __builtin_sve_svbcax_s32(__VA_ARGS__) -#define svbcax_s64(...) __builtin_sve_svbcax_s64(__VA_ARGS__) -#define svbcax_s16(...) __builtin_sve_svbcax_s16(__VA_ARGS__) -#define svbsl1n_n_u8(...) __builtin_sve_svbsl1n_n_u8(__VA_ARGS__) -#define svbsl1n_n_u32(...) __builtin_sve_svbsl1n_n_u32(__VA_ARGS__) -#define svbsl1n_n_u64(...) __builtin_sve_svbsl1n_n_u64(__VA_ARGS__) -#define svbsl1n_n_u16(...) __builtin_sve_svbsl1n_n_u16(__VA_ARGS__) -#define svbsl1n_n_s8(...) __builtin_sve_svbsl1n_n_s8(__VA_ARGS__) -#define svbsl1n_n_s32(...) __builtin_sve_svbsl1n_n_s32(__VA_ARGS__) -#define svbsl1n_n_s64(...) __builtin_sve_svbsl1n_n_s64(__VA_ARGS__) -#define svbsl1n_n_s16(...) __builtin_sve_svbsl1n_n_s16(__VA_ARGS__) -#define svbsl1n_u8(...) __builtin_sve_svbsl1n_u8(__VA_ARGS__) -#define svbsl1n_u32(...) __builtin_sve_svbsl1n_u32(__VA_ARGS__) -#define svbsl1n_u64(...) __builtin_sve_svbsl1n_u64(__VA_ARGS__) -#define svbsl1n_u16(...) __builtin_sve_svbsl1n_u16(__VA_ARGS__) -#define svbsl1n_s8(...) __builtin_sve_svbsl1n_s8(__VA_ARGS__) -#define svbsl1n_s32(...) __builtin_sve_svbsl1n_s32(__VA_ARGS__) -#define svbsl1n_s64(...) __builtin_sve_svbsl1n_s64(__VA_ARGS__) -#define svbsl1n_s16(...) __builtin_sve_svbsl1n_s16(__VA_ARGS__) -#define svbsl2n_n_u8(...) __builtin_sve_svbsl2n_n_u8(__VA_ARGS__) -#define svbsl2n_n_u32(...) __builtin_sve_svbsl2n_n_u32(__VA_ARGS__) -#define svbsl2n_n_u64(...) __builtin_sve_svbsl2n_n_u64(__VA_ARGS__) -#define svbsl2n_n_u16(...) __builtin_sve_svbsl2n_n_u16(__VA_ARGS__) -#define svbsl2n_n_s8(...) __builtin_sve_svbsl2n_n_s8(__VA_ARGS__) -#define svbsl2n_n_s32(...) __builtin_sve_svbsl2n_n_s32(__VA_ARGS__) -#define svbsl2n_n_s64(...) __builtin_sve_svbsl2n_n_s64(__VA_ARGS__) -#define svbsl2n_n_s16(...) __builtin_sve_svbsl2n_n_s16(__VA_ARGS__) -#define svbsl2n_u8(...) __builtin_sve_svbsl2n_u8(__VA_ARGS__) -#define svbsl2n_u32(...) __builtin_sve_svbsl2n_u32(__VA_ARGS__) -#define svbsl2n_u64(...) __builtin_sve_svbsl2n_u64(__VA_ARGS__) -#define svbsl2n_u16(...) __builtin_sve_svbsl2n_u16(__VA_ARGS__) -#define svbsl2n_s8(...) __builtin_sve_svbsl2n_s8(__VA_ARGS__) -#define svbsl2n_s32(...) __builtin_sve_svbsl2n_s32(__VA_ARGS__) -#define svbsl2n_s64(...) __builtin_sve_svbsl2n_s64(__VA_ARGS__) -#define svbsl2n_s16(...) __builtin_sve_svbsl2n_s16(__VA_ARGS__) -#define svbsl_n_u8(...) __builtin_sve_svbsl_n_u8(__VA_ARGS__) -#define svbsl_n_u32(...) __builtin_sve_svbsl_n_u32(__VA_ARGS__) -#define svbsl_n_u64(...) __builtin_sve_svbsl_n_u64(__VA_ARGS__) -#define svbsl_n_u16(...) __builtin_sve_svbsl_n_u16(__VA_ARGS__) -#define svbsl_n_s8(...) __builtin_sve_svbsl_n_s8(__VA_ARGS__) -#define svbsl_n_s32(...) __builtin_sve_svbsl_n_s32(__VA_ARGS__) -#define svbsl_n_s64(...) __builtin_sve_svbsl_n_s64(__VA_ARGS__) -#define svbsl_n_s16(...) __builtin_sve_svbsl_n_s16(__VA_ARGS__) -#define svbsl_u8(...) __builtin_sve_svbsl_u8(__VA_ARGS__) -#define svbsl_u32(...) __builtin_sve_svbsl_u32(__VA_ARGS__) -#define svbsl_u64(...) __builtin_sve_svbsl_u64(__VA_ARGS__) -#define svbsl_u16(...) __builtin_sve_svbsl_u16(__VA_ARGS__) -#define svbsl_s8(...) __builtin_sve_svbsl_s8(__VA_ARGS__) -#define svbsl_s32(...) __builtin_sve_svbsl_s32(__VA_ARGS__) -#define svbsl_s64(...) __builtin_sve_svbsl_s64(__VA_ARGS__) -#define svbsl_s16(...) __builtin_sve_svbsl_s16(__VA_ARGS__) -#define svcadd_u8(...) __builtin_sve_svcadd_u8(__VA_ARGS__) -#define svcadd_u32(...) __builtin_sve_svcadd_u32(__VA_ARGS__) -#define svcadd_u64(...) __builtin_sve_svcadd_u64(__VA_ARGS__) -#define svcadd_u16(...) __builtin_sve_svcadd_u16(__VA_ARGS__) -#define svcadd_s8(...) __builtin_sve_svcadd_s8(__VA_ARGS__) -#define svcadd_s32(...) __builtin_sve_svcadd_s32(__VA_ARGS__) -#define svcadd_s64(...) __builtin_sve_svcadd_s64(__VA_ARGS__) -#define svcadd_s16(...) __builtin_sve_svcadd_s16(__VA_ARGS__) -#define svcdot_s32(...) __builtin_sve_svcdot_s32(__VA_ARGS__) -#define svcdot_s64(...) __builtin_sve_svcdot_s64(__VA_ARGS__) -#define svcdot_lane_s32(...) __builtin_sve_svcdot_lane_s32(__VA_ARGS__) -#define svcdot_lane_s64(...) __builtin_sve_svcdot_lane_s64(__VA_ARGS__) -#define svcmla_u8(...) __builtin_sve_svcmla_u8(__VA_ARGS__) -#define svcmla_u32(...) __builtin_sve_svcmla_u32(__VA_ARGS__) -#define svcmla_u64(...) __builtin_sve_svcmla_u64(__VA_ARGS__) -#define svcmla_u16(...) __builtin_sve_svcmla_u16(__VA_ARGS__) -#define svcmla_s8(...) __builtin_sve_svcmla_s8(__VA_ARGS__) -#define svcmla_s32(...) __builtin_sve_svcmla_s32(__VA_ARGS__) -#define svcmla_s64(...) __builtin_sve_svcmla_s64(__VA_ARGS__) -#define svcmla_s16(...) __builtin_sve_svcmla_s16(__VA_ARGS__) -#define svcmla_lane_u32(...) __builtin_sve_svcmla_lane_u32(__VA_ARGS__) -#define svcmla_lane_u16(...) __builtin_sve_svcmla_lane_u16(__VA_ARGS__) -#define svcmla_lane_s32(...) __builtin_sve_svcmla_lane_s32(__VA_ARGS__) -#define svcmla_lane_s16(...) __builtin_sve_svcmla_lane_s16(__VA_ARGS__) -#define svcvtlt_f32_f16_m(...) __builtin_sve_svcvtlt_f32_f16_m(__VA_ARGS__) -#define svcvtlt_f32_f16_x(...) __builtin_sve_svcvtlt_f32_f16_x(__VA_ARGS__) -#define svcvtlt_f64_f32_m(...) __builtin_sve_svcvtlt_f64_f32_m(__VA_ARGS__) -#define svcvtlt_f64_f32_x(...) __builtin_sve_svcvtlt_f64_f32_x(__VA_ARGS__) -#define svcvtnt_f16_f32_m(...) __builtin_sve_svcvtnt_f16_f32_m(__VA_ARGS__) -#define svcvtnt_f32_f64_m(...) __builtin_sve_svcvtnt_f32_f64_m(__VA_ARGS__) -#define svcvtx_f32_f64_m(...) __builtin_sve_svcvtx_f32_f64_m(__VA_ARGS__) -#define svcvtx_f32_f64_x(...) __builtin_sve_svcvtx_f32_f64_x(__VA_ARGS__) -#define svcvtx_f32_f64_z(...) __builtin_sve_svcvtx_f32_f64_z(__VA_ARGS__) -#define svcvtxnt_f32_f64_m(...) __builtin_sve_svcvtxnt_f32_f64_m(__VA_ARGS__) -#define sveor3_n_u8(...) __builtin_sve_sveor3_n_u8(__VA_ARGS__) -#define sveor3_n_u32(...) __builtin_sve_sveor3_n_u32(__VA_ARGS__) -#define sveor3_n_u64(...) __builtin_sve_sveor3_n_u64(__VA_ARGS__) -#define sveor3_n_u16(...) __builtin_sve_sveor3_n_u16(__VA_ARGS__) -#define sveor3_n_s8(...) __builtin_sve_sveor3_n_s8(__VA_ARGS__) -#define sveor3_n_s32(...) __builtin_sve_sveor3_n_s32(__VA_ARGS__) -#define sveor3_n_s64(...) __builtin_sve_sveor3_n_s64(__VA_ARGS__) -#define sveor3_n_s16(...) __builtin_sve_sveor3_n_s16(__VA_ARGS__) -#define sveor3_u8(...) __builtin_sve_sveor3_u8(__VA_ARGS__) -#define sveor3_u32(...) __builtin_sve_sveor3_u32(__VA_ARGS__) -#define sveor3_u64(...) __builtin_sve_sveor3_u64(__VA_ARGS__) -#define sveor3_u16(...) __builtin_sve_sveor3_u16(__VA_ARGS__) -#define sveor3_s8(...) __builtin_sve_sveor3_s8(__VA_ARGS__) -#define sveor3_s32(...) __builtin_sve_sveor3_s32(__VA_ARGS__) -#define sveor3_s64(...) __builtin_sve_sveor3_s64(__VA_ARGS__) -#define sveor3_s16(...) __builtin_sve_sveor3_s16(__VA_ARGS__) -#define sveorbt_n_u8(...) __builtin_sve_sveorbt_n_u8(__VA_ARGS__) -#define sveorbt_n_u32(...) __builtin_sve_sveorbt_n_u32(__VA_ARGS__) -#define sveorbt_n_u64(...) __builtin_sve_sveorbt_n_u64(__VA_ARGS__) -#define sveorbt_n_u16(...) __builtin_sve_sveorbt_n_u16(__VA_ARGS__) -#define sveorbt_n_s8(...) __builtin_sve_sveorbt_n_s8(__VA_ARGS__) -#define sveorbt_n_s32(...) __builtin_sve_sveorbt_n_s32(__VA_ARGS__) -#define sveorbt_n_s64(...) __builtin_sve_sveorbt_n_s64(__VA_ARGS__) -#define sveorbt_n_s16(...) __builtin_sve_sveorbt_n_s16(__VA_ARGS__) -#define sveorbt_u8(...) __builtin_sve_sveorbt_u8(__VA_ARGS__) -#define sveorbt_u32(...) __builtin_sve_sveorbt_u32(__VA_ARGS__) -#define sveorbt_u64(...) __builtin_sve_sveorbt_u64(__VA_ARGS__) -#define sveorbt_u16(...) __builtin_sve_sveorbt_u16(__VA_ARGS__) -#define sveorbt_s8(...) __builtin_sve_sveorbt_s8(__VA_ARGS__) -#define sveorbt_s32(...) __builtin_sve_sveorbt_s32(__VA_ARGS__) -#define sveorbt_s64(...) __builtin_sve_sveorbt_s64(__VA_ARGS__) -#define sveorbt_s16(...) __builtin_sve_sveorbt_s16(__VA_ARGS__) -#define sveortb_n_u8(...) __builtin_sve_sveortb_n_u8(__VA_ARGS__) -#define sveortb_n_u32(...) __builtin_sve_sveortb_n_u32(__VA_ARGS__) -#define sveortb_n_u64(...) __builtin_sve_sveortb_n_u64(__VA_ARGS__) -#define sveortb_n_u16(...) __builtin_sve_sveortb_n_u16(__VA_ARGS__) -#define sveortb_n_s8(...) __builtin_sve_sveortb_n_s8(__VA_ARGS__) -#define sveortb_n_s32(...) __builtin_sve_sveortb_n_s32(__VA_ARGS__) -#define sveortb_n_s64(...) __builtin_sve_sveortb_n_s64(__VA_ARGS__) -#define sveortb_n_s16(...) __builtin_sve_sveortb_n_s16(__VA_ARGS__) -#define sveortb_u8(...) __builtin_sve_sveortb_u8(__VA_ARGS__) -#define sveortb_u32(...) __builtin_sve_sveortb_u32(__VA_ARGS__) -#define sveortb_u64(...) __builtin_sve_sveortb_u64(__VA_ARGS__) -#define sveortb_u16(...) __builtin_sve_sveortb_u16(__VA_ARGS__) -#define sveortb_s8(...) __builtin_sve_sveortb_s8(__VA_ARGS__) -#define sveortb_s32(...) __builtin_sve_sveortb_s32(__VA_ARGS__) -#define sveortb_s64(...) __builtin_sve_sveortb_s64(__VA_ARGS__) -#define sveortb_s16(...) __builtin_sve_sveortb_s16(__VA_ARGS__) -#define svhadd_n_s8_m(...) __builtin_sve_svhadd_n_s8_m(__VA_ARGS__) -#define svhadd_n_s32_m(...) __builtin_sve_svhadd_n_s32_m(__VA_ARGS__) -#define svhadd_n_s64_m(...) __builtin_sve_svhadd_n_s64_m(__VA_ARGS__) -#define svhadd_n_s16_m(...) __builtin_sve_svhadd_n_s16_m(__VA_ARGS__) -#define svhadd_n_s8_x(...) __builtin_sve_svhadd_n_s8_x(__VA_ARGS__) -#define svhadd_n_s32_x(...) __builtin_sve_svhadd_n_s32_x(__VA_ARGS__) -#define svhadd_n_s64_x(...) __builtin_sve_svhadd_n_s64_x(__VA_ARGS__) -#define svhadd_n_s16_x(...) __builtin_sve_svhadd_n_s16_x(__VA_ARGS__) -#define svhadd_n_s8_z(...) __builtin_sve_svhadd_n_s8_z(__VA_ARGS__) -#define svhadd_n_s32_z(...) __builtin_sve_svhadd_n_s32_z(__VA_ARGS__) -#define svhadd_n_s64_z(...) __builtin_sve_svhadd_n_s64_z(__VA_ARGS__) -#define svhadd_n_s16_z(...) __builtin_sve_svhadd_n_s16_z(__VA_ARGS__) -#define svhadd_n_u8_m(...) __builtin_sve_svhadd_n_u8_m(__VA_ARGS__) -#define svhadd_n_u32_m(...) __builtin_sve_svhadd_n_u32_m(__VA_ARGS__) -#define svhadd_n_u64_m(...) __builtin_sve_svhadd_n_u64_m(__VA_ARGS__) -#define svhadd_n_u16_m(...) __builtin_sve_svhadd_n_u16_m(__VA_ARGS__) -#define svhadd_n_u8_x(...) __builtin_sve_svhadd_n_u8_x(__VA_ARGS__) -#define svhadd_n_u32_x(...) __builtin_sve_svhadd_n_u32_x(__VA_ARGS__) -#define svhadd_n_u64_x(...) __builtin_sve_svhadd_n_u64_x(__VA_ARGS__) -#define svhadd_n_u16_x(...) __builtin_sve_svhadd_n_u16_x(__VA_ARGS__) -#define svhadd_n_u8_z(...) __builtin_sve_svhadd_n_u8_z(__VA_ARGS__) -#define svhadd_n_u32_z(...) __builtin_sve_svhadd_n_u32_z(__VA_ARGS__) -#define svhadd_n_u64_z(...) __builtin_sve_svhadd_n_u64_z(__VA_ARGS__) -#define svhadd_n_u16_z(...) __builtin_sve_svhadd_n_u16_z(__VA_ARGS__) -#define svhadd_s8_m(...) __builtin_sve_svhadd_s8_m(__VA_ARGS__) -#define svhadd_s32_m(...) __builtin_sve_svhadd_s32_m(__VA_ARGS__) -#define svhadd_s64_m(...) __builtin_sve_svhadd_s64_m(__VA_ARGS__) -#define svhadd_s16_m(...) __builtin_sve_svhadd_s16_m(__VA_ARGS__) -#define svhadd_s8_x(...) __builtin_sve_svhadd_s8_x(__VA_ARGS__) -#define svhadd_s32_x(...) __builtin_sve_svhadd_s32_x(__VA_ARGS__) -#define svhadd_s64_x(...) __builtin_sve_svhadd_s64_x(__VA_ARGS__) -#define svhadd_s16_x(...) __builtin_sve_svhadd_s16_x(__VA_ARGS__) -#define svhadd_s8_z(...) __builtin_sve_svhadd_s8_z(__VA_ARGS__) -#define svhadd_s32_z(...) __builtin_sve_svhadd_s32_z(__VA_ARGS__) -#define svhadd_s64_z(...) __builtin_sve_svhadd_s64_z(__VA_ARGS__) -#define svhadd_s16_z(...) __builtin_sve_svhadd_s16_z(__VA_ARGS__) -#define svhadd_u8_m(...) __builtin_sve_svhadd_u8_m(__VA_ARGS__) -#define svhadd_u32_m(...) __builtin_sve_svhadd_u32_m(__VA_ARGS__) -#define svhadd_u64_m(...) __builtin_sve_svhadd_u64_m(__VA_ARGS__) -#define svhadd_u16_m(...) __builtin_sve_svhadd_u16_m(__VA_ARGS__) -#define svhadd_u8_x(...) __builtin_sve_svhadd_u8_x(__VA_ARGS__) -#define svhadd_u32_x(...) __builtin_sve_svhadd_u32_x(__VA_ARGS__) -#define svhadd_u64_x(...) __builtin_sve_svhadd_u64_x(__VA_ARGS__) -#define svhadd_u16_x(...) __builtin_sve_svhadd_u16_x(__VA_ARGS__) -#define svhadd_u8_z(...) __builtin_sve_svhadd_u8_z(__VA_ARGS__) -#define svhadd_u32_z(...) __builtin_sve_svhadd_u32_z(__VA_ARGS__) -#define svhadd_u64_z(...) __builtin_sve_svhadd_u64_z(__VA_ARGS__) -#define svhadd_u16_z(...) __builtin_sve_svhadd_u16_z(__VA_ARGS__) -#define svhistcnt_u32_z(...) __builtin_sve_svhistcnt_u32_z(__VA_ARGS__) -#define svhistcnt_u64_z(...) __builtin_sve_svhistcnt_u64_z(__VA_ARGS__) -#define svhistcnt_s32_z(...) __builtin_sve_svhistcnt_s32_z(__VA_ARGS__) -#define svhistcnt_s64_z(...) __builtin_sve_svhistcnt_s64_z(__VA_ARGS__) -#define svhistseg_u8(...) __builtin_sve_svhistseg_u8(__VA_ARGS__) -#define svhistseg_s8(...) __builtin_sve_svhistseg_s8(__VA_ARGS__) -#define svhsub_n_s8_m(...) __builtin_sve_svhsub_n_s8_m(__VA_ARGS__) -#define svhsub_n_s32_m(...) __builtin_sve_svhsub_n_s32_m(__VA_ARGS__) -#define svhsub_n_s64_m(...) __builtin_sve_svhsub_n_s64_m(__VA_ARGS__) -#define svhsub_n_s16_m(...) __builtin_sve_svhsub_n_s16_m(__VA_ARGS__) -#define svhsub_n_s8_x(...) __builtin_sve_svhsub_n_s8_x(__VA_ARGS__) -#define svhsub_n_s32_x(...) __builtin_sve_svhsub_n_s32_x(__VA_ARGS__) -#define svhsub_n_s64_x(...) __builtin_sve_svhsub_n_s64_x(__VA_ARGS__) -#define svhsub_n_s16_x(...) __builtin_sve_svhsub_n_s16_x(__VA_ARGS__) -#define svhsub_n_s8_z(...) __builtin_sve_svhsub_n_s8_z(__VA_ARGS__) -#define svhsub_n_s32_z(...) __builtin_sve_svhsub_n_s32_z(__VA_ARGS__) -#define svhsub_n_s64_z(...) __builtin_sve_svhsub_n_s64_z(__VA_ARGS__) -#define svhsub_n_s16_z(...) __builtin_sve_svhsub_n_s16_z(__VA_ARGS__) -#define svhsub_n_u8_m(...) __builtin_sve_svhsub_n_u8_m(__VA_ARGS__) -#define svhsub_n_u32_m(...) __builtin_sve_svhsub_n_u32_m(__VA_ARGS__) -#define svhsub_n_u64_m(...) __builtin_sve_svhsub_n_u64_m(__VA_ARGS__) -#define svhsub_n_u16_m(...) __builtin_sve_svhsub_n_u16_m(__VA_ARGS__) -#define svhsub_n_u8_x(...) __builtin_sve_svhsub_n_u8_x(__VA_ARGS__) -#define svhsub_n_u32_x(...) __builtin_sve_svhsub_n_u32_x(__VA_ARGS__) -#define svhsub_n_u64_x(...) __builtin_sve_svhsub_n_u64_x(__VA_ARGS__) -#define svhsub_n_u16_x(...) __builtin_sve_svhsub_n_u16_x(__VA_ARGS__) -#define svhsub_n_u8_z(...) __builtin_sve_svhsub_n_u8_z(__VA_ARGS__) -#define svhsub_n_u32_z(...) __builtin_sve_svhsub_n_u32_z(__VA_ARGS__) -#define svhsub_n_u64_z(...) __builtin_sve_svhsub_n_u64_z(__VA_ARGS__) -#define svhsub_n_u16_z(...) __builtin_sve_svhsub_n_u16_z(__VA_ARGS__) -#define svhsub_s8_m(...) __builtin_sve_svhsub_s8_m(__VA_ARGS__) -#define svhsub_s32_m(...) __builtin_sve_svhsub_s32_m(__VA_ARGS__) -#define svhsub_s64_m(...) __builtin_sve_svhsub_s64_m(__VA_ARGS__) -#define svhsub_s16_m(...) __builtin_sve_svhsub_s16_m(__VA_ARGS__) -#define svhsub_s8_x(...) __builtin_sve_svhsub_s8_x(__VA_ARGS__) -#define svhsub_s32_x(...) __builtin_sve_svhsub_s32_x(__VA_ARGS__) -#define svhsub_s64_x(...) __builtin_sve_svhsub_s64_x(__VA_ARGS__) -#define svhsub_s16_x(...) __builtin_sve_svhsub_s16_x(__VA_ARGS__) -#define svhsub_s8_z(...) __builtin_sve_svhsub_s8_z(__VA_ARGS__) -#define svhsub_s32_z(...) __builtin_sve_svhsub_s32_z(__VA_ARGS__) -#define svhsub_s64_z(...) __builtin_sve_svhsub_s64_z(__VA_ARGS__) -#define svhsub_s16_z(...) __builtin_sve_svhsub_s16_z(__VA_ARGS__) -#define svhsub_u8_m(...) __builtin_sve_svhsub_u8_m(__VA_ARGS__) -#define svhsub_u32_m(...) __builtin_sve_svhsub_u32_m(__VA_ARGS__) -#define svhsub_u64_m(...) __builtin_sve_svhsub_u64_m(__VA_ARGS__) -#define svhsub_u16_m(...) __builtin_sve_svhsub_u16_m(__VA_ARGS__) -#define svhsub_u8_x(...) __builtin_sve_svhsub_u8_x(__VA_ARGS__) -#define svhsub_u32_x(...) __builtin_sve_svhsub_u32_x(__VA_ARGS__) -#define svhsub_u64_x(...) __builtin_sve_svhsub_u64_x(__VA_ARGS__) -#define svhsub_u16_x(...) __builtin_sve_svhsub_u16_x(__VA_ARGS__) -#define svhsub_u8_z(...) __builtin_sve_svhsub_u8_z(__VA_ARGS__) -#define svhsub_u32_z(...) __builtin_sve_svhsub_u32_z(__VA_ARGS__) -#define svhsub_u64_z(...) __builtin_sve_svhsub_u64_z(__VA_ARGS__) -#define svhsub_u16_z(...) __builtin_sve_svhsub_u16_z(__VA_ARGS__) -#define svhsubr_n_s8_m(...) __builtin_sve_svhsubr_n_s8_m(__VA_ARGS__) -#define svhsubr_n_s32_m(...) __builtin_sve_svhsubr_n_s32_m(__VA_ARGS__) -#define svhsubr_n_s64_m(...) __builtin_sve_svhsubr_n_s64_m(__VA_ARGS__) -#define svhsubr_n_s16_m(...) __builtin_sve_svhsubr_n_s16_m(__VA_ARGS__) -#define svhsubr_n_s8_x(...) __builtin_sve_svhsubr_n_s8_x(__VA_ARGS__) -#define svhsubr_n_s32_x(...) __builtin_sve_svhsubr_n_s32_x(__VA_ARGS__) -#define svhsubr_n_s64_x(...) __builtin_sve_svhsubr_n_s64_x(__VA_ARGS__) -#define svhsubr_n_s16_x(...) __builtin_sve_svhsubr_n_s16_x(__VA_ARGS__) -#define svhsubr_n_s8_z(...) __builtin_sve_svhsubr_n_s8_z(__VA_ARGS__) -#define svhsubr_n_s32_z(...) __builtin_sve_svhsubr_n_s32_z(__VA_ARGS__) -#define svhsubr_n_s64_z(...) __builtin_sve_svhsubr_n_s64_z(__VA_ARGS__) -#define svhsubr_n_s16_z(...) __builtin_sve_svhsubr_n_s16_z(__VA_ARGS__) -#define svhsubr_n_u8_m(...) __builtin_sve_svhsubr_n_u8_m(__VA_ARGS__) -#define svhsubr_n_u32_m(...) __builtin_sve_svhsubr_n_u32_m(__VA_ARGS__) -#define svhsubr_n_u64_m(...) __builtin_sve_svhsubr_n_u64_m(__VA_ARGS__) -#define svhsubr_n_u16_m(...) __builtin_sve_svhsubr_n_u16_m(__VA_ARGS__) -#define svhsubr_n_u8_x(...) __builtin_sve_svhsubr_n_u8_x(__VA_ARGS__) -#define svhsubr_n_u32_x(...) __builtin_sve_svhsubr_n_u32_x(__VA_ARGS__) -#define svhsubr_n_u64_x(...) __builtin_sve_svhsubr_n_u64_x(__VA_ARGS__) -#define svhsubr_n_u16_x(...) __builtin_sve_svhsubr_n_u16_x(__VA_ARGS__) -#define svhsubr_n_u8_z(...) __builtin_sve_svhsubr_n_u8_z(__VA_ARGS__) -#define svhsubr_n_u32_z(...) __builtin_sve_svhsubr_n_u32_z(__VA_ARGS__) -#define svhsubr_n_u64_z(...) __builtin_sve_svhsubr_n_u64_z(__VA_ARGS__) -#define svhsubr_n_u16_z(...) __builtin_sve_svhsubr_n_u16_z(__VA_ARGS__) -#define svhsubr_s8_m(...) __builtin_sve_svhsubr_s8_m(__VA_ARGS__) -#define svhsubr_s32_m(...) __builtin_sve_svhsubr_s32_m(__VA_ARGS__) -#define svhsubr_s64_m(...) __builtin_sve_svhsubr_s64_m(__VA_ARGS__) -#define svhsubr_s16_m(...) __builtin_sve_svhsubr_s16_m(__VA_ARGS__) -#define svhsubr_s8_x(...) __builtin_sve_svhsubr_s8_x(__VA_ARGS__) -#define svhsubr_s32_x(...) __builtin_sve_svhsubr_s32_x(__VA_ARGS__) -#define svhsubr_s64_x(...) __builtin_sve_svhsubr_s64_x(__VA_ARGS__) -#define svhsubr_s16_x(...) __builtin_sve_svhsubr_s16_x(__VA_ARGS__) -#define svhsubr_s8_z(...) __builtin_sve_svhsubr_s8_z(__VA_ARGS__) -#define svhsubr_s32_z(...) __builtin_sve_svhsubr_s32_z(__VA_ARGS__) -#define svhsubr_s64_z(...) __builtin_sve_svhsubr_s64_z(__VA_ARGS__) -#define svhsubr_s16_z(...) __builtin_sve_svhsubr_s16_z(__VA_ARGS__) -#define svhsubr_u8_m(...) __builtin_sve_svhsubr_u8_m(__VA_ARGS__) -#define svhsubr_u32_m(...) __builtin_sve_svhsubr_u32_m(__VA_ARGS__) -#define svhsubr_u64_m(...) __builtin_sve_svhsubr_u64_m(__VA_ARGS__) -#define svhsubr_u16_m(...) __builtin_sve_svhsubr_u16_m(__VA_ARGS__) -#define svhsubr_u8_x(...) __builtin_sve_svhsubr_u8_x(__VA_ARGS__) -#define svhsubr_u32_x(...) __builtin_sve_svhsubr_u32_x(__VA_ARGS__) -#define svhsubr_u64_x(...) __builtin_sve_svhsubr_u64_x(__VA_ARGS__) -#define svhsubr_u16_x(...) __builtin_sve_svhsubr_u16_x(__VA_ARGS__) -#define svhsubr_u8_z(...) __builtin_sve_svhsubr_u8_z(__VA_ARGS__) -#define svhsubr_u32_z(...) __builtin_sve_svhsubr_u32_z(__VA_ARGS__) -#define svhsubr_u64_z(...) __builtin_sve_svhsubr_u64_z(__VA_ARGS__) -#define svhsubr_u16_z(...) __builtin_sve_svhsubr_u16_z(__VA_ARGS__) -#define svldnt1_gather_u32base_index_u32(...) __builtin_sve_svldnt1_gather_u32base_index_u32(__VA_ARGS__) -#define svldnt1_gather_u64base_index_u64(...) __builtin_sve_svldnt1_gather_u64base_index_u64(__VA_ARGS__) -#define svldnt1_gather_u64base_index_f64(...) __builtin_sve_svldnt1_gather_u64base_index_f64(__VA_ARGS__) -#define svldnt1_gather_u32base_index_f32(...) __builtin_sve_svldnt1_gather_u32base_index_f32(__VA_ARGS__) -#define svldnt1_gather_u32base_index_s32(...) __builtin_sve_svldnt1_gather_u32base_index_s32(__VA_ARGS__) -#define svldnt1_gather_u64base_index_s64(...) __builtin_sve_svldnt1_gather_u64base_index_s64(__VA_ARGS__) -#define svldnt1_gather_u32base_offset_u32(...) __builtin_sve_svldnt1_gather_u32base_offset_u32(__VA_ARGS__) -#define svldnt1_gather_u64base_offset_u64(...) __builtin_sve_svldnt1_gather_u64base_offset_u64(__VA_ARGS__) -#define svldnt1_gather_u64base_offset_f64(...) __builtin_sve_svldnt1_gather_u64base_offset_f64(__VA_ARGS__) -#define svldnt1_gather_u32base_offset_f32(...) __builtin_sve_svldnt1_gather_u32base_offset_f32(__VA_ARGS__) -#define svldnt1_gather_u32base_offset_s32(...) __builtin_sve_svldnt1_gather_u32base_offset_s32(__VA_ARGS__) -#define svldnt1_gather_u64base_offset_s64(...) __builtin_sve_svldnt1_gather_u64base_offset_s64(__VA_ARGS__) -#define svldnt1_gather_u32base_u32(...) __builtin_sve_svldnt1_gather_u32base_u32(__VA_ARGS__) -#define svldnt1_gather_u64base_u64(...) __builtin_sve_svldnt1_gather_u64base_u64(__VA_ARGS__) -#define svldnt1_gather_u64base_f64(...) __builtin_sve_svldnt1_gather_u64base_f64(__VA_ARGS__) -#define svldnt1_gather_u32base_f32(...) __builtin_sve_svldnt1_gather_u32base_f32(__VA_ARGS__) -#define svldnt1_gather_u32base_s32(...) __builtin_sve_svldnt1_gather_u32base_s32(__VA_ARGS__) -#define svldnt1_gather_u64base_s64(...) __builtin_sve_svldnt1_gather_u64base_s64(__VA_ARGS__) -#define svldnt1_gather_s64index_u64(...) __builtin_sve_svldnt1_gather_s64index_u64(__VA_ARGS__) -#define svldnt1_gather_s64index_f64(...) __builtin_sve_svldnt1_gather_s64index_f64(__VA_ARGS__) -#define svldnt1_gather_s64index_s64(...) __builtin_sve_svldnt1_gather_s64index_s64(__VA_ARGS__) -#define svldnt1_gather_u64index_u64(...) __builtin_sve_svldnt1_gather_u64index_u64(__VA_ARGS__) -#define svldnt1_gather_u64index_f64(...) __builtin_sve_svldnt1_gather_u64index_f64(__VA_ARGS__) -#define svldnt1_gather_u64index_s64(...) __builtin_sve_svldnt1_gather_u64index_s64(__VA_ARGS__) -#define svldnt1_gather_u32offset_u32(...) __builtin_sve_svldnt1_gather_u32offset_u32(__VA_ARGS__) -#define svldnt1_gather_u32offset_f32(...) __builtin_sve_svldnt1_gather_u32offset_f32(__VA_ARGS__) -#define svldnt1_gather_u32offset_s32(...) __builtin_sve_svldnt1_gather_u32offset_s32(__VA_ARGS__) -#define svldnt1_gather_s64offset_u64(...) __builtin_sve_svldnt1_gather_s64offset_u64(__VA_ARGS__) -#define svldnt1_gather_s64offset_f64(...) __builtin_sve_svldnt1_gather_s64offset_f64(__VA_ARGS__) -#define svldnt1_gather_s64offset_s64(...) __builtin_sve_svldnt1_gather_s64offset_s64(__VA_ARGS__) -#define svldnt1_gather_u64offset_u64(...) __builtin_sve_svldnt1_gather_u64offset_u64(__VA_ARGS__) -#define svldnt1_gather_u64offset_f64(...) __builtin_sve_svldnt1_gather_u64offset_f64(__VA_ARGS__) -#define svldnt1_gather_u64offset_s64(...) __builtin_sve_svldnt1_gather_u64offset_s64(__VA_ARGS__) -#define svldnt1sb_gather_u32base_offset_u32(...) __builtin_sve_svldnt1sb_gather_u32base_offset_u32(__VA_ARGS__) -#define svldnt1sb_gather_u64base_offset_u64(...) __builtin_sve_svldnt1sb_gather_u64base_offset_u64(__VA_ARGS__) -#define svldnt1sb_gather_u32base_offset_s32(...) __builtin_sve_svldnt1sb_gather_u32base_offset_s32(__VA_ARGS__) -#define svldnt1sb_gather_u64base_offset_s64(...) __builtin_sve_svldnt1sb_gather_u64base_offset_s64(__VA_ARGS__) -#define svldnt1sb_gather_u32base_u32(...) __builtin_sve_svldnt1sb_gather_u32base_u32(__VA_ARGS__) -#define svldnt1sb_gather_u64base_u64(...) __builtin_sve_svldnt1sb_gather_u64base_u64(__VA_ARGS__) -#define svldnt1sb_gather_u32base_s32(...) __builtin_sve_svldnt1sb_gather_u32base_s32(__VA_ARGS__) -#define svldnt1sb_gather_u64base_s64(...) __builtin_sve_svldnt1sb_gather_u64base_s64(__VA_ARGS__) -#define svldnt1sb_gather_u32offset_u32(...) __builtin_sve_svldnt1sb_gather_u32offset_u32(__VA_ARGS__) -#define svldnt1sb_gather_u32offset_s32(...) __builtin_sve_svldnt1sb_gather_u32offset_s32(__VA_ARGS__) -#define svldnt1sb_gather_s64offset_u64(...) __builtin_sve_svldnt1sb_gather_s64offset_u64(__VA_ARGS__) -#define svldnt1sb_gather_s64offset_s64(...) __builtin_sve_svldnt1sb_gather_s64offset_s64(__VA_ARGS__) -#define svldnt1sb_gather_u64offset_u64(...) __builtin_sve_svldnt1sb_gather_u64offset_u64(__VA_ARGS__) -#define svldnt1sb_gather_u64offset_s64(...) __builtin_sve_svldnt1sb_gather_u64offset_s64(__VA_ARGS__) -#define svldnt1sh_gather_u32base_index_u32(...) __builtin_sve_svldnt1sh_gather_u32base_index_u32(__VA_ARGS__) -#define svldnt1sh_gather_u64base_index_u64(...) __builtin_sve_svldnt1sh_gather_u64base_index_u64(__VA_ARGS__) -#define svldnt1sh_gather_u32base_index_s32(...) __builtin_sve_svldnt1sh_gather_u32base_index_s32(__VA_ARGS__) -#define svldnt1sh_gather_u64base_index_s64(...) __builtin_sve_svldnt1sh_gather_u64base_index_s64(__VA_ARGS__) -#define svldnt1sh_gather_u32base_offset_u32(...) __builtin_sve_svldnt1sh_gather_u32base_offset_u32(__VA_ARGS__) -#define svldnt1sh_gather_u64base_offset_u64(...) __builtin_sve_svldnt1sh_gather_u64base_offset_u64(__VA_ARGS__) -#define svldnt1sh_gather_u32base_offset_s32(...) __builtin_sve_svldnt1sh_gather_u32base_offset_s32(__VA_ARGS__) -#define svldnt1sh_gather_u64base_offset_s64(...) __builtin_sve_svldnt1sh_gather_u64base_offset_s64(__VA_ARGS__) -#define svldnt1sh_gather_u32base_u32(...) __builtin_sve_svldnt1sh_gather_u32base_u32(__VA_ARGS__) -#define svldnt1sh_gather_u64base_u64(...) __builtin_sve_svldnt1sh_gather_u64base_u64(__VA_ARGS__) -#define svldnt1sh_gather_u32base_s32(...) __builtin_sve_svldnt1sh_gather_u32base_s32(__VA_ARGS__) -#define svldnt1sh_gather_u64base_s64(...) __builtin_sve_svldnt1sh_gather_u64base_s64(__VA_ARGS__) -#define svldnt1sh_gather_s64index_u64(...) __builtin_sve_svldnt1sh_gather_s64index_u64(__VA_ARGS__) -#define svldnt1sh_gather_s64index_s64(...) __builtin_sve_svldnt1sh_gather_s64index_s64(__VA_ARGS__) -#define svldnt1sh_gather_u64index_u64(...) __builtin_sve_svldnt1sh_gather_u64index_u64(__VA_ARGS__) -#define svldnt1sh_gather_u64index_s64(...) __builtin_sve_svldnt1sh_gather_u64index_s64(__VA_ARGS__) -#define svldnt1sh_gather_u32offset_u32(...) __builtin_sve_svldnt1sh_gather_u32offset_u32(__VA_ARGS__) -#define svldnt1sh_gather_u32offset_s32(...) __builtin_sve_svldnt1sh_gather_u32offset_s32(__VA_ARGS__) -#define svldnt1sh_gather_s64offset_u64(...) __builtin_sve_svldnt1sh_gather_s64offset_u64(__VA_ARGS__) -#define svldnt1sh_gather_s64offset_s64(...) __builtin_sve_svldnt1sh_gather_s64offset_s64(__VA_ARGS__) -#define svldnt1sh_gather_u64offset_u64(...) __builtin_sve_svldnt1sh_gather_u64offset_u64(__VA_ARGS__) -#define svldnt1sh_gather_u64offset_s64(...) __builtin_sve_svldnt1sh_gather_u64offset_s64(__VA_ARGS__) -#define svldnt1sw_gather_u64base_index_u64(...) __builtin_sve_svldnt1sw_gather_u64base_index_u64(__VA_ARGS__) -#define svldnt1sw_gather_u64base_index_s64(...) __builtin_sve_svldnt1sw_gather_u64base_index_s64(__VA_ARGS__) -#define svldnt1sw_gather_u64base_offset_u64(...) __builtin_sve_svldnt1sw_gather_u64base_offset_u64(__VA_ARGS__) -#define svldnt1sw_gather_u64base_offset_s64(...) __builtin_sve_svldnt1sw_gather_u64base_offset_s64(__VA_ARGS__) -#define svldnt1sw_gather_u64base_u64(...) __builtin_sve_svldnt1sw_gather_u64base_u64(__VA_ARGS__) -#define svldnt1sw_gather_u64base_s64(...) __builtin_sve_svldnt1sw_gather_u64base_s64(__VA_ARGS__) -#define svldnt1sw_gather_s64index_u64(...) __builtin_sve_svldnt1sw_gather_s64index_u64(__VA_ARGS__) -#define svldnt1sw_gather_s64index_s64(...) __builtin_sve_svldnt1sw_gather_s64index_s64(__VA_ARGS__) -#define svldnt1sw_gather_u64index_u64(...) __builtin_sve_svldnt1sw_gather_u64index_u64(__VA_ARGS__) -#define svldnt1sw_gather_u64index_s64(...) __builtin_sve_svldnt1sw_gather_u64index_s64(__VA_ARGS__) -#define svldnt1sw_gather_s64offset_u64(...) __builtin_sve_svldnt1sw_gather_s64offset_u64(__VA_ARGS__) -#define svldnt1sw_gather_s64offset_s64(...) __builtin_sve_svldnt1sw_gather_s64offset_s64(__VA_ARGS__) -#define svldnt1sw_gather_u64offset_u64(...) __builtin_sve_svldnt1sw_gather_u64offset_u64(__VA_ARGS__) -#define svldnt1sw_gather_u64offset_s64(...) __builtin_sve_svldnt1sw_gather_u64offset_s64(__VA_ARGS__) -#define svldnt1ub_gather_u32base_offset_u32(...) __builtin_sve_svldnt1ub_gather_u32base_offset_u32(__VA_ARGS__) -#define svldnt1ub_gather_u64base_offset_u64(...) __builtin_sve_svldnt1ub_gather_u64base_offset_u64(__VA_ARGS__) -#define svldnt1ub_gather_u32base_offset_s32(...) __builtin_sve_svldnt1ub_gather_u32base_offset_s32(__VA_ARGS__) -#define svldnt1ub_gather_u64base_offset_s64(...) __builtin_sve_svldnt1ub_gather_u64base_offset_s64(__VA_ARGS__) -#define svldnt1ub_gather_u32base_u32(...) __builtin_sve_svldnt1ub_gather_u32base_u32(__VA_ARGS__) -#define svldnt1ub_gather_u64base_u64(...) __builtin_sve_svldnt1ub_gather_u64base_u64(__VA_ARGS__) -#define svldnt1ub_gather_u32base_s32(...) __builtin_sve_svldnt1ub_gather_u32base_s32(__VA_ARGS__) -#define svldnt1ub_gather_u64base_s64(...) __builtin_sve_svldnt1ub_gather_u64base_s64(__VA_ARGS__) -#define svldnt1ub_gather_u32offset_u32(...) __builtin_sve_svldnt1ub_gather_u32offset_u32(__VA_ARGS__) -#define svldnt1ub_gather_u32offset_s32(...) __builtin_sve_svldnt1ub_gather_u32offset_s32(__VA_ARGS__) -#define svldnt1ub_gather_s64offset_u64(...) __builtin_sve_svldnt1ub_gather_s64offset_u64(__VA_ARGS__) -#define svldnt1ub_gather_s64offset_s64(...) __builtin_sve_svldnt1ub_gather_s64offset_s64(__VA_ARGS__) -#define svldnt1ub_gather_u64offset_u64(...) __builtin_sve_svldnt1ub_gather_u64offset_u64(__VA_ARGS__) -#define svldnt1ub_gather_u64offset_s64(...) __builtin_sve_svldnt1ub_gather_u64offset_s64(__VA_ARGS__) -#define svldnt1uh_gather_u32base_index_u32(...) __builtin_sve_svldnt1uh_gather_u32base_index_u32(__VA_ARGS__) -#define svldnt1uh_gather_u64base_index_u64(...) __builtin_sve_svldnt1uh_gather_u64base_index_u64(__VA_ARGS__) -#define svldnt1uh_gather_u32base_index_s32(...) __builtin_sve_svldnt1uh_gather_u32base_index_s32(__VA_ARGS__) -#define svldnt1uh_gather_u64base_index_s64(...) __builtin_sve_svldnt1uh_gather_u64base_index_s64(__VA_ARGS__) -#define svldnt1uh_gather_u32base_offset_u32(...) __builtin_sve_svldnt1uh_gather_u32base_offset_u32(__VA_ARGS__) -#define svldnt1uh_gather_u64base_offset_u64(...) __builtin_sve_svldnt1uh_gather_u64base_offset_u64(__VA_ARGS__) -#define svldnt1uh_gather_u32base_offset_s32(...) __builtin_sve_svldnt1uh_gather_u32base_offset_s32(__VA_ARGS__) -#define svldnt1uh_gather_u64base_offset_s64(...) __builtin_sve_svldnt1uh_gather_u64base_offset_s64(__VA_ARGS__) -#define svldnt1uh_gather_u32base_u32(...) __builtin_sve_svldnt1uh_gather_u32base_u32(__VA_ARGS__) -#define svldnt1uh_gather_u64base_u64(...) __builtin_sve_svldnt1uh_gather_u64base_u64(__VA_ARGS__) -#define svldnt1uh_gather_u32base_s32(...) __builtin_sve_svldnt1uh_gather_u32base_s32(__VA_ARGS__) -#define svldnt1uh_gather_u64base_s64(...) __builtin_sve_svldnt1uh_gather_u64base_s64(__VA_ARGS__) -#define svldnt1uh_gather_s64index_u64(...) __builtin_sve_svldnt1uh_gather_s64index_u64(__VA_ARGS__) -#define svldnt1uh_gather_s64index_s64(...) __builtin_sve_svldnt1uh_gather_s64index_s64(__VA_ARGS__) -#define svldnt1uh_gather_u64index_u64(...) __builtin_sve_svldnt1uh_gather_u64index_u64(__VA_ARGS__) -#define svldnt1uh_gather_u64index_s64(...) __builtin_sve_svldnt1uh_gather_u64index_s64(__VA_ARGS__) -#define svldnt1uh_gather_u32offset_u32(...) __builtin_sve_svldnt1uh_gather_u32offset_u32(__VA_ARGS__) -#define svldnt1uh_gather_u32offset_s32(...) __builtin_sve_svldnt1uh_gather_u32offset_s32(__VA_ARGS__) -#define svldnt1uh_gather_s64offset_u64(...) __builtin_sve_svldnt1uh_gather_s64offset_u64(__VA_ARGS__) -#define svldnt1uh_gather_s64offset_s64(...) __builtin_sve_svldnt1uh_gather_s64offset_s64(__VA_ARGS__) -#define svldnt1uh_gather_u64offset_u64(...) __builtin_sve_svldnt1uh_gather_u64offset_u64(__VA_ARGS__) -#define svldnt1uh_gather_u64offset_s64(...) __builtin_sve_svldnt1uh_gather_u64offset_s64(__VA_ARGS__) -#define svldnt1uw_gather_u64base_index_u64(...) __builtin_sve_svldnt1uw_gather_u64base_index_u64(__VA_ARGS__) -#define svldnt1uw_gather_u64base_index_s64(...) __builtin_sve_svldnt1uw_gather_u64base_index_s64(__VA_ARGS__) -#define svldnt1uw_gather_u64base_offset_u64(...) __builtin_sve_svldnt1uw_gather_u64base_offset_u64(__VA_ARGS__) -#define svldnt1uw_gather_u64base_offset_s64(...) __builtin_sve_svldnt1uw_gather_u64base_offset_s64(__VA_ARGS__) -#define svldnt1uw_gather_u64base_u64(...) __builtin_sve_svldnt1uw_gather_u64base_u64(__VA_ARGS__) -#define svldnt1uw_gather_u64base_s64(...) __builtin_sve_svldnt1uw_gather_u64base_s64(__VA_ARGS__) -#define svldnt1uw_gather_s64index_u64(...) __builtin_sve_svldnt1uw_gather_s64index_u64(__VA_ARGS__) -#define svldnt1uw_gather_s64index_s64(...) __builtin_sve_svldnt1uw_gather_s64index_s64(__VA_ARGS__) -#define svldnt1uw_gather_u64index_u64(...) __builtin_sve_svldnt1uw_gather_u64index_u64(__VA_ARGS__) -#define svldnt1uw_gather_u64index_s64(...) __builtin_sve_svldnt1uw_gather_u64index_s64(__VA_ARGS__) -#define svldnt1uw_gather_s64offset_u64(...) __builtin_sve_svldnt1uw_gather_s64offset_u64(__VA_ARGS__) -#define svldnt1uw_gather_s64offset_s64(...) __builtin_sve_svldnt1uw_gather_s64offset_s64(__VA_ARGS__) -#define svldnt1uw_gather_u64offset_u64(...) __builtin_sve_svldnt1uw_gather_u64offset_u64(__VA_ARGS__) -#define svldnt1uw_gather_u64offset_s64(...) __builtin_sve_svldnt1uw_gather_u64offset_s64(__VA_ARGS__) -#define svlogb_f64_m(...) __builtin_sve_svlogb_f64_m(__VA_ARGS__) -#define svlogb_f32_m(...) __builtin_sve_svlogb_f32_m(__VA_ARGS__) -#define svlogb_f16_m(...) __builtin_sve_svlogb_f16_m(__VA_ARGS__) -#define svlogb_f64_x(...) __builtin_sve_svlogb_f64_x(__VA_ARGS__) -#define svlogb_f32_x(...) __builtin_sve_svlogb_f32_x(__VA_ARGS__) -#define svlogb_f16_x(...) __builtin_sve_svlogb_f16_x(__VA_ARGS__) -#define svlogb_f64_z(...) __builtin_sve_svlogb_f64_z(__VA_ARGS__) -#define svlogb_f32_z(...) __builtin_sve_svlogb_f32_z(__VA_ARGS__) -#define svlogb_f16_z(...) __builtin_sve_svlogb_f16_z(__VA_ARGS__) -#define svmatch_u8(...) __builtin_sve_svmatch_u8(__VA_ARGS__) -#define svmatch_u16(...) __builtin_sve_svmatch_u16(__VA_ARGS__) -#define svmatch_s8(...) __builtin_sve_svmatch_s8(__VA_ARGS__) -#define svmatch_s16(...) __builtin_sve_svmatch_s16(__VA_ARGS__) -#define svmaxnmp_f64_m(...) __builtin_sve_svmaxnmp_f64_m(__VA_ARGS__) -#define svmaxnmp_f32_m(...) __builtin_sve_svmaxnmp_f32_m(__VA_ARGS__) -#define svmaxnmp_f16_m(...) __builtin_sve_svmaxnmp_f16_m(__VA_ARGS__) -#define svmaxnmp_f64_x(...) __builtin_sve_svmaxnmp_f64_x(__VA_ARGS__) -#define svmaxnmp_f32_x(...) __builtin_sve_svmaxnmp_f32_x(__VA_ARGS__) -#define svmaxnmp_f16_x(...) __builtin_sve_svmaxnmp_f16_x(__VA_ARGS__) -#define svmaxp_f64_m(...) __builtin_sve_svmaxp_f64_m(__VA_ARGS__) -#define svmaxp_f32_m(...) __builtin_sve_svmaxp_f32_m(__VA_ARGS__) -#define svmaxp_f16_m(...) __builtin_sve_svmaxp_f16_m(__VA_ARGS__) -#define svmaxp_f64_x(...) __builtin_sve_svmaxp_f64_x(__VA_ARGS__) -#define svmaxp_f32_x(...) __builtin_sve_svmaxp_f32_x(__VA_ARGS__) -#define svmaxp_f16_x(...) __builtin_sve_svmaxp_f16_x(__VA_ARGS__) -#define svmaxp_s8_m(...) __builtin_sve_svmaxp_s8_m(__VA_ARGS__) -#define svmaxp_s32_m(...) __builtin_sve_svmaxp_s32_m(__VA_ARGS__) -#define svmaxp_s64_m(...) __builtin_sve_svmaxp_s64_m(__VA_ARGS__) -#define svmaxp_s16_m(...) __builtin_sve_svmaxp_s16_m(__VA_ARGS__) -#define svmaxp_s8_x(...) __builtin_sve_svmaxp_s8_x(__VA_ARGS__) -#define svmaxp_s32_x(...) __builtin_sve_svmaxp_s32_x(__VA_ARGS__) -#define svmaxp_s64_x(...) __builtin_sve_svmaxp_s64_x(__VA_ARGS__) -#define svmaxp_s16_x(...) __builtin_sve_svmaxp_s16_x(__VA_ARGS__) -#define svmaxp_u8_m(...) __builtin_sve_svmaxp_u8_m(__VA_ARGS__) -#define svmaxp_u32_m(...) __builtin_sve_svmaxp_u32_m(__VA_ARGS__) -#define svmaxp_u64_m(...) __builtin_sve_svmaxp_u64_m(__VA_ARGS__) -#define svmaxp_u16_m(...) __builtin_sve_svmaxp_u16_m(__VA_ARGS__) -#define svmaxp_u8_x(...) __builtin_sve_svmaxp_u8_x(__VA_ARGS__) -#define svmaxp_u32_x(...) __builtin_sve_svmaxp_u32_x(__VA_ARGS__) -#define svmaxp_u64_x(...) __builtin_sve_svmaxp_u64_x(__VA_ARGS__) -#define svmaxp_u16_x(...) __builtin_sve_svmaxp_u16_x(__VA_ARGS__) -#define svminnmp_f64_m(...) __builtin_sve_svminnmp_f64_m(__VA_ARGS__) -#define svminnmp_f32_m(...) __builtin_sve_svminnmp_f32_m(__VA_ARGS__) -#define svminnmp_f16_m(...) __builtin_sve_svminnmp_f16_m(__VA_ARGS__) -#define svminnmp_f64_x(...) __builtin_sve_svminnmp_f64_x(__VA_ARGS__) -#define svminnmp_f32_x(...) __builtin_sve_svminnmp_f32_x(__VA_ARGS__) -#define svminnmp_f16_x(...) __builtin_sve_svminnmp_f16_x(__VA_ARGS__) -#define svminp_f64_m(...) __builtin_sve_svminp_f64_m(__VA_ARGS__) -#define svminp_f32_m(...) __builtin_sve_svminp_f32_m(__VA_ARGS__) -#define svminp_f16_m(...) __builtin_sve_svminp_f16_m(__VA_ARGS__) -#define svminp_f64_x(...) __builtin_sve_svminp_f64_x(__VA_ARGS__) -#define svminp_f32_x(...) __builtin_sve_svminp_f32_x(__VA_ARGS__) -#define svminp_f16_x(...) __builtin_sve_svminp_f16_x(__VA_ARGS__) -#define svminp_s8_m(...) __builtin_sve_svminp_s8_m(__VA_ARGS__) -#define svminp_s32_m(...) __builtin_sve_svminp_s32_m(__VA_ARGS__) -#define svminp_s64_m(...) __builtin_sve_svminp_s64_m(__VA_ARGS__) -#define svminp_s16_m(...) __builtin_sve_svminp_s16_m(__VA_ARGS__) -#define svminp_s8_x(...) __builtin_sve_svminp_s8_x(__VA_ARGS__) -#define svminp_s32_x(...) __builtin_sve_svminp_s32_x(__VA_ARGS__) -#define svminp_s64_x(...) __builtin_sve_svminp_s64_x(__VA_ARGS__) -#define svminp_s16_x(...) __builtin_sve_svminp_s16_x(__VA_ARGS__) -#define svminp_u8_m(...) __builtin_sve_svminp_u8_m(__VA_ARGS__) -#define svminp_u32_m(...) __builtin_sve_svminp_u32_m(__VA_ARGS__) -#define svminp_u64_m(...) __builtin_sve_svminp_u64_m(__VA_ARGS__) -#define svminp_u16_m(...) __builtin_sve_svminp_u16_m(__VA_ARGS__) -#define svminp_u8_x(...) __builtin_sve_svminp_u8_x(__VA_ARGS__) -#define svminp_u32_x(...) __builtin_sve_svminp_u32_x(__VA_ARGS__) -#define svminp_u64_x(...) __builtin_sve_svminp_u64_x(__VA_ARGS__) -#define svminp_u16_x(...) __builtin_sve_svminp_u16_x(__VA_ARGS__) -#define svmla_lane_u32(...) __builtin_sve_svmla_lane_u32(__VA_ARGS__) -#define svmla_lane_u64(...) __builtin_sve_svmla_lane_u64(__VA_ARGS__) -#define svmla_lane_u16(...) __builtin_sve_svmla_lane_u16(__VA_ARGS__) -#define svmla_lane_s32(...) __builtin_sve_svmla_lane_s32(__VA_ARGS__) -#define svmla_lane_s64(...) __builtin_sve_svmla_lane_s64(__VA_ARGS__) -#define svmla_lane_s16(...) __builtin_sve_svmla_lane_s16(__VA_ARGS__) -#define svmlalb_n_f32(...) __builtin_sve_svmlalb_n_f32(__VA_ARGS__) -#define svmlalb_n_s32(...) __builtin_sve_svmlalb_n_s32(__VA_ARGS__) -#define svmlalb_n_s64(...) __builtin_sve_svmlalb_n_s64(__VA_ARGS__) -#define svmlalb_n_s16(...) __builtin_sve_svmlalb_n_s16(__VA_ARGS__) -#define svmlalb_n_u32(...) __builtin_sve_svmlalb_n_u32(__VA_ARGS__) -#define svmlalb_n_u64(...) __builtin_sve_svmlalb_n_u64(__VA_ARGS__) -#define svmlalb_n_u16(...) __builtin_sve_svmlalb_n_u16(__VA_ARGS__) -#define svmlalb_f32(...) __builtin_sve_svmlalb_f32(__VA_ARGS__) -#define svmlalb_s32(...) __builtin_sve_svmlalb_s32(__VA_ARGS__) -#define svmlalb_s64(...) __builtin_sve_svmlalb_s64(__VA_ARGS__) -#define svmlalb_s16(...) __builtin_sve_svmlalb_s16(__VA_ARGS__) -#define svmlalb_u32(...) __builtin_sve_svmlalb_u32(__VA_ARGS__) -#define svmlalb_u64(...) __builtin_sve_svmlalb_u64(__VA_ARGS__) -#define svmlalb_u16(...) __builtin_sve_svmlalb_u16(__VA_ARGS__) -#define svmlalb_lane_f32(...) __builtin_sve_svmlalb_lane_f32(__VA_ARGS__) -#define svmlalb_lane_s32(...) __builtin_sve_svmlalb_lane_s32(__VA_ARGS__) -#define svmlalb_lane_s64(...) __builtin_sve_svmlalb_lane_s64(__VA_ARGS__) -#define svmlalb_lane_u32(...) __builtin_sve_svmlalb_lane_u32(__VA_ARGS__) -#define svmlalb_lane_u64(...) __builtin_sve_svmlalb_lane_u64(__VA_ARGS__) -#define svmlalt_n_f32(...) __builtin_sve_svmlalt_n_f32(__VA_ARGS__) -#define svmlalt_n_s32(...) __builtin_sve_svmlalt_n_s32(__VA_ARGS__) -#define svmlalt_n_s64(...) __builtin_sve_svmlalt_n_s64(__VA_ARGS__) -#define svmlalt_n_s16(...) __builtin_sve_svmlalt_n_s16(__VA_ARGS__) -#define svmlalt_n_u32(...) __builtin_sve_svmlalt_n_u32(__VA_ARGS__) -#define svmlalt_n_u64(...) __builtin_sve_svmlalt_n_u64(__VA_ARGS__) -#define svmlalt_n_u16(...) __builtin_sve_svmlalt_n_u16(__VA_ARGS__) -#define svmlalt_f32(...) __builtin_sve_svmlalt_f32(__VA_ARGS__) -#define svmlalt_s32(...) __builtin_sve_svmlalt_s32(__VA_ARGS__) -#define svmlalt_s64(...) __builtin_sve_svmlalt_s64(__VA_ARGS__) -#define svmlalt_s16(...) __builtin_sve_svmlalt_s16(__VA_ARGS__) -#define svmlalt_u32(...) __builtin_sve_svmlalt_u32(__VA_ARGS__) -#define svmlalt_u64(...) __builtin_sve_svmlalt_u64(__VA_ARGS__) -#define svmlalt_u16(...) __builtin_sve_svmlalt_u16(__VA_ARGS__) -#define svmlalt_lane_f32(...) __builtin_sve_svmlalt_lane_f32(__VA_ARGS__) -#define svmlalt_lane_s32(...) __builtin_sve_svmlalt_lane_s32(__VA_ARGS__) -#define svmlalt_lane_s64(...) __builtin_sve_svmlalt_lane_s64(__VA_ARGS__) -#define svmlalt_lane_u32(...) __builtin_sve_svmlalt_lane_u32(__VA_ARGS__) -#define svmlalt_lane_u64(...) __builtin_sve_svmlalt_lane_u64(__VA_ARGS__) -#define svmls_lane_u32(...) __builtin_sve_svmls_lane_u32(__VA_ARGS__) -#define svmls_lane_u64(...) __builtin_sve_svmls_lane_u64(__VA_ARGS__) -#define svmls_lane_u16(...) __builtin_sve_svmls_lane_u16(__VA_ARGS__) -#define svmls_lane_s32(...) __builtin_sve_svmls_lane_s32(__VA_ARGS__) -#define svmls_lane_s64(...) __builtin_sve_svmls_lane_s64(__VA_ARGS__) -#define svmls_lane_s16(...) __builtin_sve_svmls_lane_s16(__VA_ARGS__) -#define svmlslb_n_f32(...) __builtin_sve_svmlslb_n_f32(__VA_ARGS__) -#define svmlslb_n_s32(...) __builtin_sve_svmlslb_n_s32(__VA_ARGS__) -#define svmlslb_n_s64(...) __builtin_sve_svmlslb_n_s64(__VA_ARGS__) -#define svmlslb_n_s16(...) __builtin_sve_svmlslb_n_s16(__VA_ARGS__) -#define svmlslb_n_u32(...) __builtin_sve_svmlslb_n_u32(__VA_ARGS__) -#define svmlslb_n_u64(...) __builtin_sve_svmlslb_n_u64(__VA_ARGS__) -#define svmlslb_n_u16(...) __builtin_sve_svmlslb_n_u16(__VA_ARGS__) -#define svmlslb_f32(...) __builtin_sve_svmlslb_f32(__VA_ARGS__) -#define svmlslb_s32(...) __builtin_sve_svmlslb_s32(__VA_ARGS__) -#define svmlslb_s64(...) __builtin_sve_svmlslb_s64(__VA_ARGS__) -#define svmlslb_s16(...) __builtin_sve_svmlslb_s16(__VA_ARGS__) -#define svmlslb_u32(...) __builtin_sve_svmlslb_u32(__VA_ARGS__) -#define svmlslb_u64(...) __builtin_sve_svmlslb_u64(__VA_ARGS__) -#define svmlslb_u16(...) __builtin_sve_svmlslb_u16(__VA_ARGS__) -#define svmlslb_lane_f32(...) __builtin_sve_svmlslb_lane_f32(__VA_ARGS__) -#define svmlslb_lane_s32(...) __builtin_sve_svmlslb_lane_s32(__VA_ARGS__) -#define svmlslb_lane_s64(...) __builtin_sve_svmlslb_lane_s64(__VA_ARGS__) -#define svmlslb_lane_u32(...) __builtin_sve_svmlslb_lane_u32(__VA_ARGS__) -#define svmlslb_lane_u64(...) __builtin_sve_svmlslb_lane_u64(__VA_ARGS__) -#define svmlslt_n_f32(...) __builtin_sve_svmlslt_n_f32(__VA_ARGS__) -#define svmlslt_n_s32(...) __builtin_sve_svmlslt_n_s32(__VA_ARGS__) -#define svmlslt_n_s64(...) __builtin_sve_svmlslt_n_s64(__VA_ARGS__) -#define svmlslt_n_s16(...) __builtin_sve_svmlslt_n_s16(__VA_ARGS__) -#define svmlslt_n_u32(...) __builtin_sve_svmlslt_n_u32(__VA_ARGS__) -#define svmlslt_n_u64(...) __builtin_sve_svmlslt_n_u64(__VA_ARGS__) -#define svmlslt_n_u16(...) __builtin_sve_svmlslt_n_u16(__VA_ARGS__) -#define svmlslt_f32(...) __builtin_sve_svmlslt_f32(__VA_ARGS__) -#define svmlslt_s32(...) __builtin_sve_svmlslt_s32(__VA_ARGS__) -#define svmlslt_s64(...) __builtin_sve_svmlslt_s64(__VA_ARGS__) -#define svmlslt_s16(...) __builtin_sve_svmlslt_s16(__VA_ARGS__) -#define svmlslt_u32(...) __builtin_sve_svmlslt_u32(__VA_ARGS__) -#define svmlslt_u64(...) __builtin_sve_svmlslt_u64(__VA_ARGS__) -#define svmlslt_u16(...) __builtin_sve_svmlslt_u16(__VA_ARGS__) -#define svmlslt_lane_f32(...) __builtin_sve_svmlslt_lane_f32(__VA_ARGS__) -#define svmlslt_lane_s32(...) __builtin_sve_svmlslt_lane_s32(__VA_ARGS__) -#define svmlslt_lane_s64(...) __builtin_sve_svmlslt_lane_s64(__VA_ARGS__) -#define svmlslt_lane_u32(...) __builtin_sve_svmlslt_lane_u32(__VA_ARGS__) -#define svmlslt_lane_u64(...) __builtin_sve_svmlslt_lane_u64(__VA_ARGS__) -#define svmovlb_s32(...) __builtin_sve_svmovlb_s32(__VA_ARGS__) -#define svmovlb_s64(...) __builtin_sve_svmovlb_s64(__VA_ARGS__) -#define svmovlb_s16(...) __builtin_sve_svmovlb_s16(__VA_ARGS__) -#define svmovlb_u32(...) __builtin_sve_svmovlb_u32(__VA_ARGS__) -#define svmovlb_u64(...) __builtin_sve_svmovlb_u64(__VA_ARGS__) -#define svmovlb_u16(...) __builtin_sve_svmovlb_u16(__VA_ARGS__) -#define svmovlt_s32(...) __builtin_sve_svmovlt_s32(__VA_ARGS__) -#define svmovlt_s64(...) __builtin_sve_svmovlt_s64(__VA_ARGS__) -#define svmovlt_s16(...) __builtin_sve_svmovlt_s16(__VA_ARGS__) -#define svmovlt_u32(...) __builtin_sve_svmovlt_u32(__VA_ARGS__) -#define svmovlt_u64(...) __builtin_sve_svmovlt_u64(__VA_ARGS__) -#define svmovlt_u16(...) __builtin_sve_svmovlt_u16(__VA_ARGS__) -#define svmul_lane_u32(...) __builtin_sve_svmul_lane_u32(__VA_ARGS__) -#define svmul_lane_u64(...) __builtin_sve_svmul_lane_u64(__VA_ARGS__) -#define svmul_lane_u16(...) __builtin_sve_svmul_lane_u16(__VA_ARGS__) -#define svmul_lane_s32(...) __builtin_sve_svmul_lane_s32(__VA_ARGS__) -#define svmul_lane_s64(...) __builtin_sve_svmul_lane_s64(__VA_ARGS__) -#define svmul_lane_s16(...) __builtin_sve_svmul_lane_s16(__VA_ARGS__) -#define svmullb_n_s32(...) __builtin_sve_svmullb_n_s32(__VA_ARGS__) -#define svmullb_n_s64(...) __builtin_sve_svmullb_n_s64(__VA_ARGS__) -#define svmullb_n_s16(...) __builtin_sve_svmullb_n_s16(__VA_ARGS__) -#define svmullb_n_u32(...) __builtin_sve_svmullb_n_u32(__VA_ARGS__) -#define svmullb_n_u64(...) __builtin_sve_svmullb_n_u64(__VA_ARGS__) -#define svmullb_n_u16(...) __builtin_sve_svmullb_n_u16(__VA_ARGS__) -#define svmullb_s32(...) __builtin_sve_svmullb_s32(__VA_ARGS__) -#define svmullb_s64(...) __builtin_sve_svmullb_s64(__VA_ARGS__) -#define svmullb_s16(...) __builtin_sve_svmullb_s16(__VA_ARGS__) -#define svmullb_u32(...) __builtin_sve_svmullb_u32(__VA_ARGS__) -#define svmullb_u64(...) __builtin_sve_svmullb_u64(__VA_ARGS__) -#define svmullb_u16(...) __builtin_sve_svmullb_u16(__VA_ARGS__) -#define svmullb_lane_s32(...) __builtin_sve_svmullb_lane_s32(__VA_ARGS__) -#define svmullb_lane_s64(...) __builtin_sve_svmullb_lane_s64(__VA_ARGS__) -#define svmullb_lane_u32(...) __builtin_sve_svmullb_lane_u32(__VA_ARGS__) -#define svmullb_lane_u64(...) __builtin_sve_svmullb_lane_u64(__VA_ARGS__) -#define svmullt_n_s32(...) __builtin_sve_svmullt_n_s32(__VA_ARGS__) -#define svmullt_n_s64(...) __builtin_sve_svmullt_n_s64(__VA_ARGS__) -#define svmullt_n_s16(...) __builtin_sve_svmullt_n_s16(__VA_ARGS__) -#define svmullt_n_u32(...) __builtin_sve_svmullt_n_u32(__VA_ARGS__) -#define svmullt_n_u64(...) __builtin_sve_svmullt_n_u64(__VA_ARGS__) -#define svmullt_n_u16(...) __builtin_sve_svmullt_n_u16(__VA_ARGS__) -#define svmullt_s32(...) __builtin_sve_svmullt_s32(__VA_ARGS__) -#define svmullt_s64(...) __builtin_sve_svmullt_s64(__VA_ARGS__) -#define svmullt_s16(...) __builtin_sve_svmullt_s16(__VA_ARGS__) -#define svmullt_u32(...) __builtin_sve_svmullt_u32(__VA_ARGS__) -#define svmullt_u64(...) __builtin_sve_svmullt_u64(__VA_ARGS__) -#define svmullt_u16(...) __builtin_sve_svmullt_u16(__VA_ARGS__) -#define svmullt_lane_s32(...) __builtin_sve_svmullt_lane_s32(__VA_ARGS__) -#define svmullt_lane_s64(...) __builtin_sve_svmullt_lane_s64(__VA_ARGS__) -#define svmullt_lane_u32(...) __builtin_sve_svmullt_lane_u32(__VA_ARGS__) -#define svmullt_lane_u64(...) __builtin_sve_svmullt_lane_u64(__VA_ARGS__) -#define svnbsl_n_u8(...) __builtin_sve_svnbsl_n_u8(__VA_ARGS__) -#define svnbsl_n_u32(...) __builtin_sve_svnbsl_n_u32(__VA_ARGS__) -#define svnbsl_n_u64(...) __builtin_sve_svnbsl_n_u64(__VA_ARGS__) -#define svnbsl_n_u16(...) __builtin_sve_svnbsl_n_u16(__VA_ARGS__) -#define svnbsl_n_s8(...) __builtin_sve_svnbsl_n_s8(__VA_ARGS__) -#define svnbsl_n_s32(...) __builtin_sve_svnbsl_n_s32(__VA_ARGS__) -#define svnbsl_n_s64(...) __builtin_sve_svnbsl_n_s64(__VA_ARGS__) -#define svnbsl_n_s16(...) __builtin_sve_svnbsl_n_s16(__VA_ARGS__) -#define svnbsl_u8(...) __builtin_sve_svnbsl_u8(__VA_ARGS__) -#define svnbsl_u32(...) __builtin_sve_svnbsl_u32(__VA_ARGS__) -#define svnbsl_u64(...) __builtin_sve_svnbsl_u64(__VA_ARGS__) -#define svnbsl_u16(...) __builtin_sve_svnbsl_u16(__VA_ARGS__) -#define svnbsl_s8(...) __builtin_sve_svnbsl_s8(__VA_ARGS__) -#define svnbsl_s32(...) __builtin_sve_svnbsl_s32(__VA_ARGS__) -#define svnbsl_s64(...) __builtin_sve_svnbsl_s64(__VA_ARGS__) -#define svnbsl_s16(...) __builtin_sve_svnbsl_s16(__VA_ARGS__) -#define svnmatch_u8(...) __builtin_sve_svnmatch_u8(__VA_ARGS__) -#define svnmatch_u16(...) __builtin_sve_svnmatch_u16(__VA_ARGS__) -#define svnmatch_s8(...) __builtin_sve_svnmatch_s8(__VA_ARGS__) -#define svnmatch_s16(...) __builtin_sve_svnmatch_s16(__VA_ARGS__) -#define svpmul_n_u8(...) __builtin_sve_svpmul_n_u8(__VA_ARGS__) -#define svpmul_u8(...) __builtin_sve_svpmul_u8(__VA_ARGS__) -#define svpmullb_n_u64(...) __builtin_sve_svpmullb_n_u64(__VA_ARGS__) -#define svpmullb_n_u16(...) __builtin_sve_svpmullb_n_u16(__VA_ARGS__) -#define svpmullb_u64(...) __builtin_sve_svpmullb_u64(__VA_ARGS__) -#define svpmullb_u16(...) __builtin_sve_svpmullb_u16(__VA_ARGS__) -#define svpmullb_pair_n_u8(...) __builtin_sve_svpmullb_pair_n_u8(__VA_ARGS__) -#define svpmullb_pair_n_u32(...) __builtin_sve_svpmullb_pair_n_u32(__VA_ARGS__) -#define svpmullb_pair_u8(...) __builtin_sve_svpmullb_pair_u8(__VA_ARGS__) -#define svpmullb_pair_u32(...) __builtin_sve_svpmullb_pair_u32(__VA_ARGS__) -#define svpmullt_n_u64(...) __builtin_sve_svpmullt_n_u64(__VA_ARGS__) -#define svpmullt_n_u16(...) __builtin_sve_svpmullt_n_u16(__VA_ARGS__) -#define svpmullt_u64(...) __builtin_sve_svpmullt_u64(__VA_ARGS__) -#define svpmullt_u16(...) __builtin_sve_svpmullt_u16(__VA_ARGS__) -#define svpmullt_pair_n_u8(...) __builtin_sve_svpmullt_pair_n_u8(__VA_ARGS__) -#define svpmullt_pair_n_u32(...) __builtin_sve_svpmullt_pair_n_u32(__VA_ARGS__) -#define svpmullt_pair_u8(...) __builtin_sve_svpmullt_pair_u8(__VA_ARGS__) -#define svpmullt_pair_u32(...) __builtin_sve_svpmullt_pair_u32(__VA_ARGS__) -#define svqabs_s8_m(...) __builtin_sve_svqabs_s8_m(__VA_ARGS__) -#define svqabs_s32_m(...) __builtin_sve_svqabs_s32_m(__VA_ARGS__) -#define svqabs_s64_m(...) __builtin_sve_svqabs_s64_m(__VA_ARGS__) -#define svqabs_s16_m(...) __builtin_sve_svqabs_s16_m(__VA_ARGS__) -#define svqabs_s8_x(...) __builtin_sve_svqabs_s8_x(__VA_ARGS__) -#define svqabs_s32_x(...) __builtin_sve_svqabs_s32_x(__VA_ARGS__) -#define svqabs_s64_x(...) __builtin_sve_svqabs_s64_x(__VA_ARGS__) -#define svqabs_s16_x(...) __builtin_sve_svqabs_s16_x(__VA_ARGS__) -#define svqabs_s8_z(...) __builtin_sve_svqabs_s8_z(__VA_ARGS__) -#define svqabs_s32_z(...) __builtin_sve_svqabs_s32_z(__VA_ARGS__) -#define svqabs_s64_z(...) __builtin_sve_svqabs_s64_z(__VA_ARGS__) -#define svqabs_s16_z(...) __builtin_sve_svqabs_s16_z(__VA_ARGS__) -#define svqadd_n_s8_m(...) __builtin_sve_svqadd_n_s8_m(__VA_ARGS__) -#define svqadd_n_s32_m(...) __builtin_sve_svqadd_n_s32_m(__VA_ARGS__) -#define svqadd_n_s64_m(...) __builtin_sve_svqadd_n_s64_m(__VA_ARGS__) -#define svqadd_n_s16_m(...) __builtin_sve_svqadd_n_s16_m(__VA_ARGS__) -#define svqadd_n_s8_x(...) __builtin_sve_svqadd_n_s8_x(__VA_ARGS__) -#define svqadd_n_s32_x(...) __builtin_sve_svqadd_n_s32_x(__VA_ARGS__) -#define svqadd_n_s64_x(...) __builtin_sve_svqadd_n_s64_x(__VA_ARGS__) -#define svqadd_n_s16_x(...) __builtin_sve_svqadd_n_s16_x(__VA_ARGS__) -#define svqadd_n_s8_z(...) __builtin_sve_svqadd_n_s8_z(__VA_ARGS__) -#define svqadd_n_s32_z(...) __builtin_sve_svqadd_n_s32_z(__VA_ARGS__) -#define svqadd_n_s64_z(...) __builtin_sve_svqadd_n_s64_z(__VA_ARGS__) -#define svqadd_n_s16_z(...) __builtin_sve_svqadd_n_s16_z(__VA_ARGS__) -#define svqadd_n_u8_m(...) __builtin_sve_svqadd_n_u8_m(__VA_ARGS__) -#define svqadd_n_u32_m(...) __builtin_sve_svqadd_n_u32_m(__VA_ARGS__) -#define svqadd_n_u64_m(...) __builtin_sve_svqadd_n_u64_m(__VA_ARGS__) -#define svqadd_n_u16_m(...) __builtin_sve_svqadd_n_u16_m(__VA_ARGS__) -#define svqadd_n_u8_x(...) __builtin_sve_svqadd_n_u8_x(__VA_ARGS__) -#define svqadd_n_u32_x(...) __builtin_sve_svqadd_n_u32_x(__VA_ARGS__) -#define svqadd_n_u64_x(...) __builtin_sve_svqadd_n_u64_x(__VA_ARGS__) -#define svqadd_n_u16_x(...) __builtin_sve_svqadd_n_u16_x(__VA_ARGS__) -#define svqadd_n_u8_z(...) __builtin_sve_svqadd_n_u8_z(__VA_ARGS__) -#define svqadd_n_u32_z(...) __builtin_sve_svqadd_n_u32_z(__VA_ARGS__) -#define svqadd_n_u64_z(...) __builtin_sve_svqadd_n_u64_z(__VA_ARGS__) -#define svqadd_n_u16_z(...) __builtin_sve_svqadd_n_u16_z(__VA_ARGS__) -#define svqadd_s8_m(...) __builtin_sve_svqadd_s8_m(__VA_ARGS__) -#define svqadd_s32_m(...) __builtin_sve_svqadd_s32_m(__VA_ARGS__) -#define svqadd_s64_m(...) __builtin_sve_svqadd_s64_m(__VA_ARGS__) -#define svqadd_s16_m(...) __builtin_sve_svqadd_s16_m(__VA_ARGS__) -#define svqadd_s8_x(...) __builtin_sve_svqadd_s8_x(__VA_ARGS__) -#define svqadd_s32_x(...) __builtin_sve_svqadd_s32_x(__VA_ARGS__) -#define svqadd_s64_x(...) __builtin_sve_svqadd_s64_x(__VA_ARGS__) -#define svqadd_s16_x(...) __builtin_sve_svqadd_s16_x(__VA_ARGS__) -#define svqadd_s8_z(...) __builtin_sve_svqadd_s8_z(__VA_ARGS__) -#define svqadd_s32_z(...) __builtin_sve_svqadd_s32_z(__VA_ARGS__) -#define svqadd_s64_z(...) __builtin_sve_svqadd_s64_z(__VA_ARGS__) -#define svqadd_s16_z(...) __builtin_sve_svqadd_s16_z(__VA_ARGS__) -#define svqadd_u8_m(...) __builtin_sve_svqadd_u8_m(__VA_ARGS__) -#define svqadd_u32_m(...) __builtin_sve_svqadd_u32_m(__VA_ARGS__) -#define svqadd_u64_m(...) __builtin_sve_svqadd_u64_m(__VA_ARGS__) -#define svqadd_u16_m(...) __builtin_sve_svqadd_u16_m(__VA_ARGS__) -#define svqadd_u8_x(...) __builtin_sve_svqadd_u8_x(__VA_ARGS__) -#define svqadd_u32_x(...) __builtin_sve_svqadd_u32_x(__VA_ARGS__) -#define svqadd_u64_x(...) __builtin_sve_svqadd_u64_x(__VA_ARGS__) -#define svqadd_u16_x(...) __builtin_sve_svqadd_u16_x(__VA_ARGS__) -#define svqadd_u8_z(...) __builtin_sve_svqadd_u8_z(__VA_ARGS__) -#define svqadd_u32_z(...) __builtin_sve_svqadd_u32_z(__VA_ARGS__) -#define svqadd_u64_z(...) __builtin_sve_svqadd_u64_z(__VA_ARGS__) -#define svqadd_u16_z(...) __builtin_sve_svqadd_u16_z(__VA_ARGS__) -#define svqcadd_s8(...) __builtin_sve_svqcadd_s8(__VA_ARGS__) -#define svqcadd_s32(...) __builtin_sve_svqcadd_s32(__VA_ARGS__) -#define svqcadd_s64(...) __builtin_sve_svqcadd_s64(__VA_ARGS__) -#define svqcadd_s16(...) __builtin_sve_svqcadd_s16(__VA_ARGS__) -#define svqdmlalb_n_s32(...) __builtin_sve_svqdmlalb_n_s32(__VA_ARGS__) -#define svqdmlalb_n_s64(...) __builtin_sve_svqdmlalb_n_s64(__VA_ARGS__) -#define svqdmlalb_n_s16(...) __builtin_sve_svqdmlalb_n_s16(__VA_ARGS__) -#define svqdmlalb_s32(...) __builtin_sve_svqdmlalb_s32(__VA_ARGS__) -#define svqdmlalb_s64(...) __builtin_sve_svqdmlalb_s64(__VA_ARGS__) -#define svqdmlalb_s16(...) __builtin_sve_svqdmlalb_s16(__VA_ARGS__) -#define svqdmlalb_lane_s32(...) __builtin_sve_svqdmlalb_lane_s32(__VA_ARGS__) -#define svqdmlalb_lane_s64(...) __builtin_sve_svqdmlalb_lane_s64(__VA_ARGS__) -#define svqdmlalbt_n_s32(...) __builtin_sve_svqdmlalbt_n_s32(__VA_ARGS__) -#define svqdmlalbt_n_s64(...) __builtin_sve_svqdmlalbt_n_s64(__VA_ARGS__) -#define svqdmlalbt_n_s16(...) __builtin_sve_svqdmlalbt_n_s16(__VA_ARGS__) -#define svqdmlalbt_s32(...) __builtin_sve_svqdmlalbt_s32(__VA_ARGS__) -#define svqdmlalbt_s64(...) __builtin_sve_svqdmlalbt_s64(__VA_ARGS__) -#define svqdmlalbt_s16(...) __builtin_sve_svqdmlalbt_s16(__VA_ARGS__) -#define svqdmlalt_n_s32(...) __builtin_sve_svqdmlalt_n_s32(__VA_ARGS__) -#define svqdmlalt_n_s64(...) __builtin_sve_svqdmlalt_n_s64(__VA_ARGS__) -#define svqdmlalt_n_s16(...) __builtin_sve_svqdmlalt_n_s16(__VA_ARGS__) -#define svqdmlalt_s32(...) __builtin_sve_svqdmlalt_s32(__VA_ARGS__) -#define svqdmlalt_s64(...) __builtin_sve_svqdmlalt_s64(__VA_ARGS__) -#define svqdmlalt_s16(...) __builtin_sve_svqdmlalt_s16(__VA_ARGS__) -#define svqdmlalt_lane_s32(...) __builtin_sve_svqdmlalt_lane_s32(__VA_ARGS__) -#define svqdmlalt_lane_s64(...) __builtin_sve_svqdmlalt_lane_s64(__VA_ARGS__) -#define svqdmlslb_n_s32(...) __builtin_sve_svqdmlslb_n_s32(__VA_ARGS__) -#define svqdmlslb_n_s64(...) __builtin_sve_svqdmlslb_n_s64(__VA_ARGS__) -#define svqdmlslb_n_s16(...) __builtin_sve_svqdmlslb_n_s16(__VA_ARGS__) -#define svqdmlslb_s32(...) __builtin_sve_svqdmlslb_s32(__VA_ARGS__) -#define svqdmlslb_s64(...) __builtin_sve_svqdmlslb_s64(__VA_ARGS__) -#define svqdmlslb_s16(...) __builtin_sve_svqdmlslb_s16(__VA_ARGS__) -#define svqdmlslb_lane_s32(...) __builtin_sve_svqdmlslb_lane_s32(__VA_ARGS__) -#define svqdmlslb_lane_s64(...) __builtin_sve_svqdmlslb_lane_s64(__VA_ARGS__) -#define svqdmlslbt_n_s32(...) __builtin_sve_svqdmlslbt_n_s32(__VA_ARGS__) -#define svqdmlslbt_n_s64(...) __builtin_sve_svqdmlslbt_n_s64(__VA_ARGS__) -#define svqdmlslbt_n_s16(...) __builtin_sve_svqdmlslbt_n_s16(__VA_ARGS__) -#define svqdmlslbt_s32(...) __builtin_sve_svqdmlslbt_s32(__VA_ARGS__) -#define svqdmlslbt_s64(...) __builtin_sve_svqdmlslbt_s64(__VA_ARGS__) -#define svqdmlslbt_s16(...) __builtin_sve_svqdmlslbt_s16(__VA_ARGS__) -#define svqdmlslt_n_s32(...) __builtin_sve_svqdmlslt_n_s32(__VA_ARGS__) -#define svqdmlslt_n_s64(...) __builtin_sve_svqdmlslt_n_s64(__VA_ARGS__) -#define svqdmlslt_n_s16(...) __builtin_sve_svqdmlslt_n_s16(__VA_ARGS__) -#define svqdmlslt_s32(...) __builtin_sve_svqdmlslt_s32(__VA_ARGS__) -#define svqdmlslt_s64(...) __builtin_sve_svqdmlslt_s64(__VA_ARGS__) -#define svqdmlslt_s16(...) __builtin_sve_svqdmlslt_s16(__VA_ARGS__) -#define svqdmlslt_lane_s32(...) __builtin_sve_svqdmlslt_lane_s32(__VA_ARGS__) -#define svqdmlslt_lane_s64(...) __builtin_sve_svqdmlslt_lane_s64(__VA_ARGS__) -#define svqdmulh_n_s8(...) __builtin_sve_svqdmulh_n_s8(__VA_ARGS__) -#define svqdmulh_n_s32(...) __builtin_sve_svqdmulh_n_s32(__VA_ARGS__) -#define svqdmulh_n_s64(...) __builtin_sve_svqdmulh_n_s64(__VA_ARGS__) -#define svqdmulh_n_s16(...) __builtin_sve_svqdmulh_n_s16(__VA_ARGS__) -#define svqdmulh_s8(...) __builtin_sve_svqdmulh_s8(__VA_ARGS__) -#define svqdmulh_s32(...) __builtin_sve_svqdmulh_s32(__VA_ARGS__) -#define svqdmulh_s64(...) __builtin_sve_svqdmulh_s64(__VA_ARGS__) -#define svqdmulh_s16(...) __builtin_sve_svqdmulh_s16(__VA_ARGS__) -#define svqdmulh_lane_s32(...) __builtin_sve_svqdmulh_lane_s32(__VA_ARGS__) -#define svqdmulh_lane_s64(...) __builtin_sve_svqdmulh_lane_s64(__VA_ARGS__) -#define svqdmulh_lane_s16(...) __builtin_sve_svqdmulh_lane_s16(__VA_ARGS__) -#define svqdmullb_n_s32(...) __builtin_sve_svqdmullb_n_s32(__VA_ARGS__) -#define svqdmullb_n_s64(...) __builtin_sve_svqdmullb_n_s64(__VA_ARGS__) -#define svqdmullb_n_s16(...) __builtin_sve_svqdmullb_n_s16(__VA_ARGS__) -#define svqdmullb_s32(...) __builtin_sve_svqdmullb_s32(__VA_ARGS__) -#define svqdmullb_s64(...) __builtin_sve_svqdmullb_s64(__VA_ARGS__) -#define svqdmullb_s16(...) __builtin_sve_svqdmullb_s16(__VA_ARGS__) -#define svqdmullb_lane_s32(...) __builtin_sve_svqdmullb_lane_s32(__VA_ARGS__) -#define svqdmullb_lane_s64(...) __builtin_sve_svqdmullb_lane_s64(__VA_ARGS__) -#define svqdmullt_n_s32(...) __builtin_sve_svqdmullt_n_s32(__VA_ARGS__) -#define svqdmullt_n_s64(...) __builtin_sve_svqdmullt_n_s64(__VA_ARGS__) -#define svqdmullt_n_s16(...) __builtin_sve_svqdmullt_n_s16(__VA_ARGS__) -#define svqdmullt_s32(...) __builtin_sve_svqdmullt_s32(__VA_ARGS__) -#define svqdmullt_s64(...) __builtin_sve_svqdmullt_s64(__VA_ARGS__) -#define svqdmullt_s16(...) __builtin_sve_svqdmullt_s16(__VA_ARGS__) -#define svqdmullt_lane_s32(...) __builtin_sve_svqdmullt_lane_s32(__VA_ARGS__) -#define svqdmullt_lane_s64(...) __builtin_sve_svqdmullt_lane_s64(__VA_ARGS__) -#define svqneg_s8_m(...) __builtin_sve_svqneg_s8_m(__VA_ARGS__) -#define svqneg_s32_m(...) __builtin_sve_svqneg_s32_m(__VA_ARGS__) -#define svqneg_s64_m(...) __builtin_sve_svqneg_s64_m(__VA_ARGS__) -#define svqneg_s16_m(...) __builtin_sve_svqneg_s16_m(__VA_ARGS__) -#define svqneg_s8_x(...) __builtin_sve_svqneg_s8_x(__VA_ARGS__) -#define svqneg_s32_x(...) __builtin_sve_svqneg_s32_x(__VA_ARGS__) -#define svqneg_s64_x(...) __builtin_sve_svqneg_s64_x(__VA_ARGS__) -#define svqneg_s16_x(...) __builtin_sve_svqneg_s16_x(__VA_ARGS__) -#define svqneg_s8_z(...) __builtin_sve_svqneg_s8_z(__VA_ARGS__) -#define svqneg_s32_z(...) __builtin_sve_svqneg_s32_z(__VA_ARGS__) -#define svqneg_s64_z(...) __builtin_sve_svqneg_s64_z(__VA_ARGS__) -#define svqneg_s16_z(...) __builtin_sve_svqneg_s16_z(__VA_ARGS__) -#define svqrdcmlah_s8(...) __builtin_sve_svqrdcmlah_s8(__VA_ARGS__) -#define svqrdcmlah_s32(...) __builtin_sve_svqrdcmlah_s32(__VA_ARGS__) -#define svqrdcmlah_s64(...) __builtin_sve_svqrdcmlah_s64(__VA_ARGS__) -#define svqrdcmlah_s16(...) __builtin_sve_svqrdcmlah_s16(__VA_ARGS__) -#define svqrdcmlah_lane_s32(...) __builtin_sve_svqrdcmlah_lane_s32(__VA_ARGS__) -#define svqrdcmlah_lane_s16(...) __builtin_sve_svqrdcmlah_lane_s16(__VA_ARGS__) -#define svqrdmlah_n_s8(...) __builtin_sve_svqrdmlah_n_s8(__VA_ARGS__) -#define svqrdmlah_n_s32(...) __builtin_sve_svqrdmlah_n_s32(__VA_ARGS__) -#define svqrdmlah_n_s64(...) __builtin_sve_svqrdmlah_n_s64(__VA_ARGS__) -#define svqrdmlah_n_s16(...) __builtin_sve_svqrdmlah_n_s16(__VA_ARGS__) -#define svqrdmlah_s8(...) __builtin_sve_svqrdmlah_s8(__VA_ARGS__) -#define svqrdmlah_s32(...) __builtin_sve_svqrdmlah_s32(__VA_ARGS__) -#define svqrdmlah_s64(...) __builtin_sve_svqrdmlah_s64(__VA_ARGS__) -#define svqrdmlah_s16(...) __builtin_sve_svqrdmlah_s16(__VA_ARGS__) -#define svqrdmlah_lane_s32(...) __builtin_sve_svqrdmlah_lane_s32(__VA_ARGS__) -#define svqrdmlah_lane_s64(...) __builtin_sve_svqrdmlah_lane_s64(__VA_ARGS__) -#define svqrdmlah_lane_s16(...) __builtin_sve_svqrdmlah_lane_s16(__VA_ARGS__) -#define svqrdmlsh_n_s8(...) __builtin_sve_svqrdmlsh_n_s8(__VA_ARGS__) -#define svqrdmlsh_n_s32(...) __builtin_sve_svqrdmlsh_n_s32(__VA_ARGS__) -#define svqrdmlsh_n_s64(...) __builtin_sve_svqrdmlsh_n_s64(__VA_ARGS__) -#define svqrdmlsh_n_s16(...) __builtin_sve_svqrdmlsh_n_s16(__VA_ARGS__) -#define svqrdmlsh_s8(...) __builtin_sve_svqrdmlsh_s8(__VA_ARGS__) -#define svqrdmlsh_s32(...) __builtin_sve_svqrdmlsh_s32(__VA_ARGS__) -#define svqrdmlsh_s64(...) __builtin_sve_svqrdmlsh_s64(__VA_ARGS__) -#define svqrdmlsh_s16(...) __builtin_sve_svqrdmlsh_s16(__VA_ARGS__) -#define svqrdmlsh_lane_s32(...) __builtin_sve_svqrdmlsh_lane_s32(__VA_ARGS__) -#define svqrdmlsh_lane_s64(...) __builtin_sve_svqrdmlsh_lane_s64(__VA_ARGS__) -#define svqrdmlsh_lane_s16(...) __builtin_sve_svqrdmlsh_lane_s16(__VA_ARGS__) -#define svqrdmulh_n_s8(...) __builtin_sve_svqrdmulh_n_s8(__VA_ARGS__) -#define svqrdmulh_n_s32(...) __builtin_sve_svqrdmulh_n_s32(__VA_ARGS__) -#define svqrdmulh_n_s64(...) __builtin_sve_svqrdmulh_n_s64(__VA_ARGS__) -#define svqrdmulh_n_s16(...) __builtin_sve_svqrdmulh_n_s16(__VA_ARGS__) -#define svqrdmulh_s8(...) __builtin_sve_svqrdmulh_s8(__VA_ARGS__) -#define svqrdmulh_s32(...) __builtin_sve_svqrdmulh_s32(__VA_ARGS__) -#define svqrdmulh_s64(...) __builtin_sve_svqrdmulh_s64(__VA_ARGS__) -#define svqrdmulh_s16(...) __builtin_sve_svqrdmulh_s16(__VA_ARGS__) -#define svqrdmulh_lane_s32(...) __builtin_sve_svqrdmulh_lane_s32(__VA_ARGS__) -#define svqrdmulh_lane_s64(...) __builtin_sve_svqrdmulh_lane_s64(__VA_ARGS__) -#define svqrdmulh_lane_s16(...) __builtin_sve_svqrdmulh_lane_s16(__VA_ARGS__) -#define svqrshl_n_s8_m(...) __builtin_sve_svqrshl_n_s8_m(__VA_ARGS__) -#define svqrshl_n_s32_m(...) __builtin_sve_svqrshl_n_s32_m(__VA_ARGS__) -#define svqrshl_n_s64_m(...) __builtin_sve_svqrshl_n_s64_m(__VA_ARGS__) -#define svqrshl_n_s16_m(...) __builtin_sve_svqrshl_n_s16_m(__VA_ARGS__) -#define svqrshl_n_s8_x(...) __builtin_sve_svqrshl_n_s8_x(__VA_ARGS__) -#define svqrshl_n_s32_x(...) __builtin_sve_svqrshl_n_s32_x(__VA_ARGS__) -#define svqrshl_n_s64_x(...) __builtin_sve_svqrshl_n_s64_x(__VA_ARGS__) -#define svqrshl_n_s16_x(...) __builtin_sve_svqrshl_n_s16_x(__VA_ARGS__) -#define svqrshl_n_s8_z(...) __builtin_sve_svqrshl_n_s8_z(__VA_ARGS__) -#define svqrshl_n_s32_z(...) __builtin_sve_svqrshl_n_s32_z(__VA_ARGS__) -#define svqrshl_n_s64_z(...) __builtin_sve_svqrshl_n_s64_z(__VA_ARGS__) -#define svqrshl_n_s16_z(...) __builtin_sve_svqrshl_n_s16_z(__VA_ARGS__) -#define svqrshl_n_u8_m(...) __builtin_sve_svqrshl_n_u8_m(__VA_ARGS__) -#define svqrshl_n_u32_m(...) __builtin_sve_svqrshl_n_u32_m(__VA_ARGS__) -#define svqrshl_n_u64_m(...) __builtin_sve_svqrshl_n_u64_m(__VA_ARGS__) -#define svqrshl_n_u16_m(...) __builtin_sve_svqrshl_n_u16_m(__VA_ARGS__) -#define svqrshl_n_u8_x(...) __builtin_sve_svqrshl_n_u8_x(__VA_ARGS__) -#define svqrshl_n_u32_x(...) __builtin_sve_svqrshl_n_u32_x(__VA_ARGS__) -#define svqrshl_n_u64_x(...) __builtin_sve_svqrshl_n_u64_x(__VA_ARGS__) -#define svqrshl_n_u16_x(...) __builtin_sve_svqrshl_n_u16_x(__VA_ARGS__) -#define svqrshl_n_u8_z(...) __builtin_sve_svqrshl_n_u8_z(__VA_ARGS__) -#define svqrshl_n_u32_z(...) __builtin_sve_svqrshl_n_u32_z(__VA_ARGS__) -#define svqrshl_n_u64_z(...) __builtin_sve_svqrshl_n_u64_z(__VA_ARGS__) -#define svqrshl_n_u16_z(...) __builtin_sve_svqrshl_n_u16_z(__VA_ARGS__) -#define svqrshl_s8_m(...) __builtin_sve_svqrshl_s8_m(__VA_ARGS__) -#define svqrshl_s32_m(...) __builtin_sve_svqrshl_s32_m(__VA_ARGS__) -#define svqrshl_s64_m(...) __builtin_sve_svqrshl_s64_m(__VA_ARGS__) -#define svqrshl_s16_m(...) __builtin_sve_svqrshl_s16_m(__VA_ARGS__) -#define svqrshl_s8_x(...) __builtin_sve_svqrshl_s8_x(__VA_ARGS__) -#define svqrshl_s32_x(...) __builtin_sve_svqrshl_s32_x(__VA_ARGS__) -#define svqrshl_s64_x(...) __builtin_sve_svqrshl_s64_x(__VA_ARGS__) -#define svqrshl_s16_x(...) __builtin_sve_svqrshl_s16_x(__VA_ARGS__) -#define svqrshl_s8_z(...) __builtin_sve_svqrshl_s8_z(__VA_ARGS__) -#define svqrshl_s32_z(...) __builtin_sve_svqrshl_s32_z(__VA_ARGS__) -#define svqrshl_s64_z(...) __builtin_sve_svqrshl_s64_z(__VA_ARGS__) -#define svqrshl_s16_z(...) __builtin_sve_svqrshl_s16_z(__VA_ARGS__) -#define svqrshl_u8_m(...) __builtin_sve_svqrshl_u8_m(__VA_ARGS__) -#define svqrshl_u32_m(...) __builtin_sve_svqrshl_u32_m(__VA_ARGS__) -#define svqrshl_u64_m(...) __builtin_sve_svqrshl_u64_m(__VA_ARGS__) -#define svqrshl_u16_m(...) __builtin_sve_svqrshl_u16_m(__VA_ARGS__) -#define svqrshl_u8_x(...) __builtin_sve_svqrshl_u8_x(__VA_ARGS__) -#define svqrshl_u32_x(...) __builtin_sve_svqrshl_u32_x(__VA_ARGS__) -#define svqrshl_u64_x(...) __builtin_sve_svqrshl_u64_x(__VA_ARGS__) -#define svqrshl_u16_x(...) __builtin_sve_svqrshl_u16_x(__VA_ARGS__) -#define svqrshl_u8_z(...) __builtin_sve_svqrshl_u8_z(__VA_ARGS__) -#define svqrshl_u32_z(...) __builtin_sve_svqrshl_u32_z(__VA_ARGS__) -#define svqrshl_u64_z(...) __builtin_sve_svqrshl_u64_z(__VA_ARGS__) -#define svqrshl_u16_z(...) __builtin_sve_svqrshl_u16_z(__VA_ARGS__) -#define svqrshrnb_n_s32(...) __builtin_sve_svqrshrnb_n_s32(__VA_ARGS__) -#define svqrshrnb_n_s64(...) __builtin_sve_svqrshrnb_n_s64(__VA_ARGS__) -#define svqrshrnb_n_s16(...) __builtin_sve_svqrshrnb_n_s16(__VA_ARGS__) -#define svqrshrnb_n_u32(...) __builtin_sve_svqrshrnb_n_u32(__VA_ARGS__) -#define svqrshrnb_n_u64(...) __builtin_sve_svqrshrnb_n_u64(__VA_ARGS__) -#define svqrshrnb_n_u16(...) __builtin_sve_svqrshrnb_n_u16(__VA_ARGS__) -#define svqrshrnt_n_s32(...) __builtin_sve_svqrshrnt_n_s32(__VA_ARGS__) -#define svqrshrnt_n_s64(...) __builtin_sve_svqrshrnt_n_s64(__VA_ARGS__) -#define svqrshrnt_n_s16(...) __builtin_sve_svqrshrnt_n_s16(__VA_ARGS__) -#define svqrshrnt_n_u32(...) __builtin_sve_svqrshrnt_n_u32(__VA_ARGS__) -#define svqrshrnt_n_u64(...) __builtin_sve_svqrshrnt_n_u64(__VA_ARGS__) -#define svqrshrnt_n_u16(...) __builtin_sve_svqrshrnt_n_u16(__VA_ARGS__) -#define svqrshrunb_n_s32(...) __builtin_sve_svqrshrunb_n_s32(__VA_ARGS__) -#define svqrshrunb_n_s64(...) __builtin_sve_svqrshrunb_n_s64(__VA_ARGS__) -#define svqrshrunb_n_s16(...) __builtin_sve_svqrshrunb_n_s16(__VA_ARGS__) -#define svqrshrunt_n_s32(...) __builtin_sve_svqrshrunt_n_s32(__VA_ARGS__) -#define svqrshrunt_n_s64(...) __builtin_sve_svqrshrunt_n_s64(__VA_ARGS__) -#define svqrshrunt_n_s16(...) __builtin_sve_svqrshrunt_n_s16(__VA_ARGS__) -#define svqshl_n_s8_m(...) __builtin_sve_svqshl_n_s8_m(__VA_ARGS__) -#define svqshl_n_s32_m(...) __builtin_sve_svqshl_n_s32_m(__VA_ARGS__) -#define svqshl_n_s64_m(...) __builtin_sve_svqshl_n_s64_m(__VA_ARGS__) -#define svqshl_n_s16_m(...) __builtin_sve_svqshl_n_s16_m(__VA_ARGS__) -#define svqshl_n_s8_x(...) __builtin_sve_svqshl_n_s8_x(__VA_ARGS__) -#define svqshl_n_s32_x(...) __builtin_sve_svqshl_n_s32_x(__VA_ARGS__) -#define svqshl_n_s64_x(...) __builtin_sve_svqshl_n_s64_x(__VA_ARGS__) -#define svqshl_n_s16_x(...) __builtin_sve_svqshl_n_s16_x(__VA_ARGS__) -#define svqshl_n_s8_z(...) __builtin_sve_svqshl_n_s8_z(__VA_ARGS__) -#define svqshl_n_s32_z(...) __builtin_sve_svqshl_n_s32_z(__VA_ARGS__) -#define svqshl_n_s64_z(...) __builtin_sve_svqshl_n_s64_z(__VA_ARGS__) -#define svqshl_n_s16_z(...) __builtin_sve_svqshl_n_s16_z(__VA_ARGS__) -#define svqshl_n_u8_m(...) __builtin_sve_svqshl_n_u8_m(__VA_ARGS__) -#define svqshl_n_u32_m(...) __builtin_sve_svqshl_n_u32_m(__VA_ARGS__) -#define svqshl_n_u64_m(...) __builtin_sve_svqshl_n_u64_m(__VA_ARGS__) -#define svqshl_n_u16_m(...) __builtin_sve_svqshl_n_u16_m(__VA_ARGS__) -#define svqshl_n_u8_x(...) __builtin_sve_svqshl_n_u8_x(__VA_ARGS__) -#define svqshl_n_u32_x(...) __builtin_sve_svqshl_n_u32_x(__VA_ARGS__) -#define svqshl_n_u64_x(...) __builtin_sve_svqshl_n_u64_x(__VA_ARGS__) -#define svqshl_n_u16_x(...) __builtin_sve_svqshl_n_u16_x(__VA_ARGS__) -#define svqshl_n_u8_z(...) __builtin_sve_svqshl_n_u8_z(__VA_ARGS__) -#define svqshl_n_u32_z(...) __builtin_sve_svqshl_n_u32_z(__VA_ARGS__) -#define svqshl_n_u64_z(...) __builtin_sve_svqshl_n_u64_z(__VA_ARGS__) -#define svqshl_n_u16_z(...) __builtin_sve_svqshl_n_u16_z(__VA_ARGS__) -#define svqshl_s8_m(...) __builtin_sve_svqshl_s8_m(__VA_ARGS__) -#define svqshl_s32_m(...) __builtin_sve_svqshl_s32_m(__VA_ARGS__) -#define svqshl_s64_m(...) __builtin_sve_svqshl_s64_m(__VA_ARGS__) -#define svqshl_s16_m(...) __builtin_sve_svqshl_s16_m(__VA_ARGS__) -#define svqshl_s8_x(...) __builtin_sve_svqshl_s8_x(__VA_ARGS__) -#define svqshl_s32_x(...) __builtin_sve_svqshl_s32_x(__VA_ARGS__) -#define svqshl_s64_x(...) __builtin_sve_svqshl_s64_x(__VA_ARGS__) -#define svqshl_s16_x(...) __builtin_sve_svqshl_s16_x(__VA_ARGS__) -#define svqshl_s8_z(...) __builtin_sve_svqshl_s8_z(__VA_ARGS__) -#define svqshl_s32_z(...) __builtin_sve_svqshl_s32_z(__VA_ARGS__) -#define svqshl_s64_z(...) __builtin_sve_svqshl_s64_z(__VA_ARGS__) -#define svqshl_s16_z(...) __builtin_sve_svqshl_s16_z(__VA_ARGS__) -#define svqshl_u8_m(...) __builtin_sve_svqshl_u8_m(__VA_ARGS__) -#define svqshl_u32_m(...) __builtin_sve_svqshl_u32_m(__VA_ARGS__) -#define svqshl_u64_m(...) __builtin_sve_svqshl_u64_m(__VA_ARGS__) -#define svqshl_u16_m(...) __builtin_sve_svqshl_u16_m(__VA_ARGS__) -#define svqshl_u8_x(...) __builtin_sve_svqshl_u8_x(__VA_ARGS__) -#define svqshl_u32_x(...) __builtin_sve_svqshl_u32_x(__VA_ARGS__) -#define svqshl_u64_x(...) __builtin_sve_svqshl_u64_x(__VA_ARGS__) -#define svqshl_u16_x(...) __builtin_sve_svqshl_u16_x(__VA_ARGS__) -#define svqshl_u8_z(...) __builtin_sve_svqshl_u8_z(__VA_ARGS__) -#define svqshl_u32_z(...) __builtin_sve_svqshl_u32_z(__VA_ARGS__) -#define svqshl_u64_z(...) __builtin_sve_svqshl_u64_z(__VA_ARGS__) -#define svqshl_u16_z(...) __builtin_sve_svqshl_u16_z(__VA_ARGS__) -#define svqshlu_n_s8_m(...) __builtin_sve_svqshlu_n_s8_m(__VA_ARGS__) -#define svqshlu_n_s32_m(...) __builtin_sve_svqshlu_n_s32_m(__VA_ARGS__) -#define svqshlu_n_s64_m(...) __builtin_sve_svqshlu_n_s64_m(__VA_ARGS__) -#define svqshlu_n_s16_m(...) __builtin_sve_svqshlu_n_s16_m(__VA_ARGS__) -#define svqshlu_n_s8_x(...) __builtin_sve_svqshlu_n_s8_x(__VA_ARGS__) -#define svqshlu_n_s32_x(...) __builtin_sve_svqshlu_n_s32_x(__VA_ARGS__) -#define svqshlu_n_s64_x(...) __builtin_sve_svqshlu_n_s64_x(__VA_ARGS__) -#define svqshlu_n_s16_x(...) __builtin_sve_svqshlu_n_s16_x(__VA_ARGS__) -#define svqshlu_n_s8_z(...) __builtin_sve_svqshlu_n_s8_z(__VA_ARGS__) -#define svqshlu_n_s32_z(...) __builtin_sve_svqshlu_n_s32_z(__VA_ARGS__) -#define svqshlu_n_s64_z(...) __builtin_sve_svqshlu_n_s64_z(__VA_ARGS__) -#define svqshlu_n_s16_z(...) __builtin_sve_svqshlu_n_s16_z(__VA_ARGS__) -#define svqshrnb_n_s32(...) __builtin_sve_svqshrnb_n_s32(__VA_ARGS__) -#define svqshrnb_n_s64(...) __builtin_sve_svqshrnb_n_s64(__VA_ARGS__) -#define svqshrnb_n_s16(...) __builtin_sve_svqshrnb_n_s16(__VA_ARGS__) -#define svqshrnb_n_u32(...) __builtin_sve_svqshrnb_n_u32(__VA_ARGS__) -#define svqshrnb_n_u64(...) __builtin_sve_svqshrnb_n_u64(__VA_ARGS__) -#define svqshrnb_n_u16(...) __builtin_sve_svqshrnb_n_u16(__VA_ARGS__) -#define svqshrnt_n_s32(...) __builtin_sve_svqshrnt_n_s32(__VA_ARGS__) -#define svqshrnt_n_s64(...) __builtin_sve_svqshrnt_n_s64(__VA_ARGS__) -#define svqshrnt_n_s16(...) __builtin_sve_svqshrnt_n_s16(__VA_ARGS__) -#define svqshrnt_n_u32(...) __builtin_sve_svqshrnt_n_u32(__VA_ARGS__) -#define svqshrnt_n_u64(...) __builtin_sve_svqshrnt_n_u64(__VA_ARGS__) -#define svqshrnt_n_u16(...) __builtin_sve_svqshrnt_n_u16(__VA_ARGS__) -#define svqshrunb_n_s32(...) __builtin_sve_svqshrunb_n_s32(__VA_ARGS__) -#define svqshrunb_n_s64(...) __builtin_sve_svqshrunb_n_s64(__VA_ARGS__) -#define svqshrunb_n_s16(...) __builtin_sve_svqshrunb_n_s16(__VA_ARGS__) -#define svqshrunt_n_s32(...) __builtin_sve_svqshrunt_n_s32(__VA_ARGS__) -#define svqshrunt_n_s64(...) __builtin_sve_svqshrunt_n_s64(__VA_ARGS__) -#define svqshrunt_n_s16(...) __builtin_sve_svqshrunt_n_s16(__VA_ARGS__) -#define svqsub_n_s8_m(...) __builtin_sve_svqsub_n_s8_m(__VA_ARGS__) -#define svqsub_n_s32_m(...) __builtin_sve_svqsub_n_s32_m(__VA_ARGS__) -#define svqsub_n_s64_m(...) __builtin_sve_svqsub_n_s64_m(__VA_ARGS__) -#define svqsub_n_s16_m(...) __builtin_sve_svqsub_n_s16_m(__VA_ARGS__) -#define svqsub_n_s8_x(...) __builtin_sve_svqsub_n_s8_x(__VA_ARGS__) -#define svqsub_n_s32_x(...) __builtin_sve_svqsub_n_s32_x(__VA_ARGS__) -#define svqsub_n_s64_x(...) __builtin_sve_svqsub_n_s64_x(__VA_ARGS__) -#define svqsub_n_s16_x(...) __builtin_sve_svqsub_n_s16_x(__VA_ARGS__) -#define svqsub_n_s8_z(...) __builtin_sve_svqsub_n_s8_z(__VA_ARGS__) -#define svqsub_n_s32_z(...) __builtin_sve_svqsub_n_s32_z(__VA_ARGS__) -#define svqsub_n_s64_z(...) __builtin_sve_svqsub_n_s64_z(__VA_ARGS__) -#define svqsub_n_s16_z(...) __builtin_sve_svqsub_n_s16_z(__VA_ARGS__) -#define svqsub_n_u8_m(...) __builtin_sve_svqsub_n_u8_m(__VA_ARGS__) -#define svqsub_n_u32_m(...) __builtin_sve_svqsub_n_u32_m(__VA_ARGS__) -#define svqsub_n_u64_m(...) __builtin_sve_svqsub_n_u64_m(__VA_ARGS__) -#define svqsub_n_u16_m(...) __builtin_sve_svqsub_n_u16_m(__VA_ARGS__) -#define svqsub_n_u8_x(...) __builtin_sve_svqsub_n_u8_x(__VA_ARGS__) -#define svqsub_n_u32_x(...) __builtin_sve_svqsub_n_u32_x(__VA_ARGS__) -#define svqsub_n_u64_x(...) __builtin_sve_svqsub_n_u64_x(__VA_ARGS__) -#define svqsub_n_u16_x(...) __builtin_sve_svqsub_n_u16_x(__VA_ARGS__) -#define svqsub_n_u8_z(...) __builtin_sve_svqsub_n_u8_z(__VA_ARGS__) -#define svqsub_n_u32_z(...) __builtin_sve_svqsub_n_u32_z(__VA_ARGS__) -#define svqsub_n_u64_z(...) __builtin_sve_svqsub_n_u64_z(__VA_ARGS__) -#define svqsub_n_u16_z(...) __builtin_sve_svqsub_n_u16_z(__VA_ARGS__) -#define svqsub_s8_m(...) __builtin_sve_svqsub_s8_m(__VA_ARGS__) -#define svqsub_s32_m(...) __builtin_sve_svqsub_s32_m(__VA_ARGS__) -#define svqsub_s64_m(...) __builtin_sve_svqsub_s64_m(__VA_ARGS__) -#define svqsub_s16_m(...) __builtin_sve_svqsub_s16_m(__VA_ARGS__) -#define svqsub_s8_x(...) __builtin_sve_svqsub_s8_x(__VA_ARGS__) -#define svqsub_s32_x(...) __builtin_sve_svqsub_s32_x(__VA_ARGS__) -#define svqsub_s64_x(...) __builtin_sve_svqsub_s64_x(__VA_ARGS__) -#define svqsub_s16_x(...) __builtin_sve_svqsub_s16_x(__VA_ARGS__) -#define svqsub_s8_z(...) __builtin_sve_svqsub_s8_z(__VA_ARGS__) -#define svqsub_s32_z(...) __builtin_sve_svqsub_s32_z(__VA_ARGS__) -#define svqsub_s64_z(...) __builtin_sve_svqsub_s64_z(__VA_ARGS__) -#define svqsub_s16_z(...) __builtin_sve_svqsub_s16_z(__VA_ARGS__) -#define svqsub_u8_m(...) __builtin_sve_svqsub_u8_m(__VA_ARGS__) -#define svqsub_u32_m(...) __builtin_sve_svqsub_u32_m(__VA_ARGS__) -#define svqsub_u64_m(...) __builtin_sve_svqsub_u64_m(__VA_ARGS__) -#define svqsub_u16_m(...) __builtin_sve_svqsub_u16_m(__VA_ARGS__) -#define svqsub_u8_x(...) __builtin_sve_svqsub_u8_x(__VA_ARGS__) -#define svqsub_u32_x(...) __builtin_sve_svqsub_u32_x(__VA_ARGS__) -#define svqsub_u64_x(...) __builtin_sve_svqsub_u64_x(__VA_ARGS__) -#define svqsub_u16_x(...) __builtin_sve_svqsub_u16_x(__VA_ARGS__) -#define svqsub_u8_z(...) __builtin_sve_svqsub_u8_z(__VA_ARGS__) -#define svqsub_u32_z(...) __builtin_sve_svqsub_u32_z(__VA_ARGS__) -#define svqsub_u64_z(...) __builtin_sve_svqsub_u64_z(__VA_ARGS__) -#define svqsub_u16_z(...) __builtin_sve_svqsub_u16_z(__VA_ARGS__) -#define svqsubr_n_s8_m(...) __builtin_sve_svqsubr_n_s8_m(__VA_ARGS__) -#define svqsubr_n_s32_m(...) __builtin_sve_svqsubr_n_s32_m(__VA_ARGS__) -#define svqsubr_n_s64_m(...) __builtin_sve_svqsubr_n_s64_m(__VA_ARGS__) -#define svqsubr_n_s16_m(...) __builtin_sve_svqsubr_n_s16_m(__VA_ARGS__) -#define svqsubr_n_s8_x(...) __builtin_sve_svqsubr_n_s8_x(__VA_ARGS__) -#define svqsubr_n_s32_x(...) __builtin_sve_svqsubr_n_s32_x(__VA_ARGS__) -#define svqsubr_n_s64_x(...) __builtin_sve_svqsubr_n_s64_x(__VA_ARGS__) -#define svqsubr_n_s16_x(...) __builtin_sve_svqsubr_n_s16_x(__VA_ARGS__) -#define svqsubr_n_s8_z(...) __builtin_sve_svqsubr_n_s8_z(__VA_ARGS__) -#define svqsubr_n_s32_z(...) __builtin_sve_svqsubr_n_s32_z(__VA_ARGS__) -#define svqsubr_n_s64_z(...) __builtin_sve_svqsubr_n_s64_z(__VA_ARGS__) -#define svqsubr_n_s16_z(...) __builtin_sve_svqsubr_n_s16_z(__VA_ARGS__) -#define svqsubr_n_u8_m(...) __builtin_sve_svqsubr_n_u8_m(__VA_ARGS__) -#define svqsubr_n_u32_m(...) __builtin_sve_svqsubr_n_u32_m(__VA_ARGS__) -#define svqsubr_n_u64_m(...) __builtin_sve_svqsubr_n_u64_m(__VA_ARGS__) -#define svqsubr_n_u16_m(...) __builtin_sve_svqsubr_n_u16_m(__VA_ARGS__) -#define svqsubr_n_u8_x(...) __builtin_sve_svqsubr_n_u8_x(__VA_ARGS__) -#define svqsubr_n_u32_x(...) __builtin_sve_svqsubr_n_u32_x(__VA_ARGS__) -#define svqsubr_n_u64_x(...) __builtin_sve_svqsubr_n_u64_x(__VA_ARGS__) -#define svqsubr_n_u16_x(...) __builtin_sve_svqsubr_n_u16_x(__VA_ARGS__) -#define svqsubr_n_u8_z(...) __builtin_sve_svqsubr_n_u8_z(__VA_ARGS__) -#define svqsubr_n_u32_z(...) __builtin_sve_svqsubr_n_u32_z(__VA_ARGS__) -#define svqsubr_n_u64_z(...) __builtin_sve_svqsubr_n_u64_z(__VA_ARGS__) -#define svqsubr_n_u16_z(...) __builtin_sve_svqsubr_n_u16_z(__VA_ARGS__) -#define svqsubr_s8_m(...) __builtin_sve_svqsubr_s8_m(__VA_ARGS__) -#define svqsubr_s32_m(...) __builtin_sve_svqsubr_s32_m(__VA_ARGS__) -#define svqsubr_s64_m(...) __builtin_sve_svqsubr_s64_m(__VA_ARGS__) -#define svqsubr_s16_m(...) __builtin_sve_svqsubr_s16_m(__VA_ARGS__) -#define svqsubr_s8_x(...) __builtin_sve_svqsubr_s8_x(__VA_ARGS__) -#define svqsubr_s32_x(...) __builtin_sve_svqsubr_s32_x(__VA_ARGS__) -#define svqsubr_s64_x(...) __builtin_sve_svqsubr_s64_x(__VA_ARGS__) -#define svqsubr_s16_x(...) __builtin_sve_svqsubr_s16_x(__VA_ARGS__) -#define svqsubr_s8_z(...) __builtin_sve_svqsubr_s8_z(__VA_ARGS__) -#define svqsubr_s32_z(...) __builtin_sve_svqsubr_s32_z(__VA_ARGS__) -#define svqsubr_s64_z(...) __builtin_sve_svqsubr_s64_z(__VA_ARGS__) -#define svqsubr_s16_z(...) __builtin_sve_svqsubr_s16_z(__VA_ARGS__) -#define svqsubr_u8_m(...) __builtin_sve_svqsubr_u8_m(__VA_ARGS__) -#define svqsubr_u32_m(...) __builtin_sve_svqsubr_u32_m(__VA_ARGS__) -#define svqsubr_u64_m(...) __builtin_sve_svqsubr_u64_m(__VA_ARGS__) -#define svqsubr_u16_m(...) __builtin_sve_svqsubr_u16_m(__VA_ARGS__) -#define svqsubr_u8_x(...) __builtin_sve_svqsubr_u8_x(__VA_ARGS__) -#define svqsubr_u32_x(...) __builtin_sve_svqsubr_u32_x(__VA_ARGS__) -#define svqsubr_u64_x(...) __builtin_sve_svqsubr_u64_x(__VA_ARGS__) -#define svqsubr_u16_x(...) __builtin_sve_svqsubr_u16_x(__VA_ARGS__) -#define svqsubr_u8_z(...) __builtin_sve_svqsubr_u8_z(__VA_ARGS__) -#define svqsubr_u32_z(...) __builtin_sve_svqsubr_u32_z(__VA_ARGS__) -#define svqsubr_u64_z(...) __builtin_sve_svqsubr_u64_z(__VA_ARGS__) -#define svqsubr_u16_z(...) __builtin_sve_svqsubr_u16_z(__VA_ARGS__) -#define svqxtnb_s32(...) __builtin_sve_svqxtnb_s32(__VA_ARGS__) -#define svqxtnb_s64(...) __builtin_sve_svqxtnb_s64(__VA_ARGS__) -#define svqxtnb_s16(...) __builtin_sve_svqxtnb_s16(__VA_ARGS__) -#define svqxtnb_u32(...) __builtin_sve_svqxtnb_u32(__VA_ARGS__) -#define svqxtnb_u64(...) __builtin_sve_svqxtnb_u64(__VA_ARGS__) -#define svqxtnb_u16(...) __builtin_sve_svqxtnb_u16(__VA_ARGS__) -#define svqxtnt_s32(...) __builtin_sve_svqxtnt_s32(__VA_ARGS__) -#define svqxtnt_s64(...) __builtin_sve_svqxtnt_s64(__VA_ARGS__) -#define svqxtnt_s16(...) __builtin_sve_svqxtnt_s16(__VA_ARGS__) -#define svqxtnt_u32(...) __builtin_sve_svqxtnt_u32(__VA_ARGS__) -#define svqxtnt_u64(...) __builtin_sve_svqxtnt_u64(__VA_ARGS__) -#define svqxtnt_u16(...) __builtin_sve_svqxtnt_u16(__VA_ARGS__) -#define svqxtunb_s32(...) __builtin_sve_svqxtunb_s32(__VA_ARGS__) -#define svqxtunb_s64(...) __builtin_sve_svqxtunb_s64(__VA_ARGS__) -#define svqxtunb_s16(...) __builtin_sve_svqxtunb_s16(__VA_ARGS__) -#define svqxtunt_s32(...) __builtin_sve_svqxtunt_s32(__VA_ARGS__) -#define svqxtunt_s64(...) __builtin_sve_svqxtunt_s64(__VA_ARGS__) -#define svqxtunt_s16(...) __builtin_sve_svqxtunt_s16(__VA_ARGS__) -#define svraddhnb_n_u32(...) __builtin_sve_svraddhnb_n_u32(__VA_ARGS__) -#define svraddhnb_n_u64(...) __builtin_sve_svraddhnb_n_u64(__VA_ARGS__) -#define svraddhnb_n_u16(...) __builtin_sve_svraddhnb_n_u16(__VA_ARGS__) -#define svraddhnb_n_s32(...) __builtin_sve_svraddhnb_n_s32(__VA_ARGS__) -#define svraddhnb_n_s64(...) __builtin_sve_svraddhnb_n_s64(__VA_ARGS__) -#define svraddhnb_n_s16(...) __builtin_sve_svraddhnb_n_s16(__VA_ARGS__) -#define svraddhnb_u32(...) __builtin_sve_svraddhnb_u32(__VA_ARGS__) -#define svraddhnb_u64(...) __builtin_sve_svraddhnb_u64(__VA_ARGS__) -#define svraddhnb_u16(...) __builtin_sve_svraddhnb_u16(__VA_ARGS__) -#define svraddhnb_s32(...) __builtin_sve_svraddhnb_s32(__VA_ARGS__) -#define svraddhnb_s64(...) __builtin_sve_svraddhnb_s64(__VA_ARGS__) -#define svraddhnb_s16(...) __builtin_sve_svraddhnb_s16(__VA_ARGS__) -#define svraddhnt_n_u32(...) __builtin_sve_svraddhnt_n_u32(__VA_ARGS__) -#define svraddhnt_n_u64(...) __builtin_sve_svraddhnt_n_u64(__VA_ARGS__) -#define svraddhnt_n_u16(...) __builtin_sve_svraddhnt_n_u16(__VA_ARGS__) -#define svraddhnt_n_s32(...) __builtin_sve_svraddhnt_n_s32(__VA_ARGS__) -#define svraddhnt_n_s64(...) __builtin_sve_svraddhnt_n_s64(__VA_ARGS__) -#define svraddhnt_n_s16(...) __builtin_sve_svraddhnt_n_s16(__VA_ARGS__) -#define svraddhnt_u32(...) __builtin_sve_svraddhnt_u32(__VA_ARGS__) -#define svraddhnt_u64(...) __builtin_sve_svraddhnt_u64(__VA_ARGS__) -#define svraddhnt_u16(...) __builtin_sve_svraddhnt_u16(__VA_ARGS__) -#define svraddhnt_s32(...) __builtin_sve_svraddhnt_s32(__VA_ARGS__) -#define svraddhnt_s64(...) __builtin_sve_svraddhnt_s64(__VA_ARGS__) -#define svraddhnt_s16(...) __builtin_sve_svraddhnt_s16(__VA_ARGS__) -#define svrecpe_u32_m(...) __builtin_sve_svrecpe_u32_m(__VA_ARGS__) -#define svrecpe_u32_x(...) __builtin_sve_svrecpe_u32_x(__VA_ARGS__) -#define svrecpe_u32_z(...) __builtin_sve_svrecpe_u32_z(__VA_ARGS__) -#define svrhadd_n_s8_m(...) __builtin_sve_svrhadd_n_s8_m(__VA_ARGS__) -#define svrhadd_n_s32_m(...) __builtin_sve_svrhadd_n_s32_m(__VA_ARGS__) -#define svrhadd_n_s64_m(...) __builtin_sve_svrhadd_n_s64_m(__VA_ARGS__) -#define svrhadd_n_s16_m(...) __builtin_sve_svrhadd_n_s16_m(__VA_ARGS__) -#define svrhadd_n_s8_x(...) __builtin_sve_svrhadd_n_s8_x(__VA_ARGS__) -#define svrhadd_n_s32_x(...) __builtin_sve_svrhadd_n_s32_x(__VA_ARGS__) -#define svrhadd_n_s64_x(...) __builtin_sve_svrhadd_n_s64_x(__VA_ARGS__) -#define svrhadd_n_s16_x(...) __builtin_sve_svrhadd_n_s16_x(__VA_ARGS__) -#define svrhadd_n_s8_z(...) __builtin_sve_svrhadd_n_s8_z(__VA_ARGS__) -#define svrhadd_n_s32_z(...) __builtin_sve_svrhadd_n_s32_z(__VA_ARGS__) -#define svrhadd_n_s64_z(...) __builtin_sve_svrhadd_n_s64_z(__VA_ARGS__) -#define svrhadd_n_s16_z(...) __builtin_sve_svrhadd_n_s16_z(__VA_ARGS__) -#define svrhadd_n_u8_m(...) __builtin_sve_svrhadd_n_u8_m(__VA_ARGS__) -#define svrhadd_n_u32_m(...) __builtin_sve_svrhadd_n_u32_m(__VA_ARGS__) -#define svrhadd_n_u64_m(...) __builtin_sve_svrhadd_n_u64_m(__VA_ARGS__) -#define svrhadd_n_u16_m(...) __builtin_sve_svrhadd_n_u16_m(__VA_ARGS__) -#define svrhadd_n_u8_x(...) __builtin_sve_svrhadd_n_u8_x(__VA_ARGS__) -#define svrhadd_n_u32_x(...) __builtin_sve_svrhadd_n_u32_x(__VA_ARGS__) -#define svrhadd_n_u64_x(...) __builtin_sve_svrhadd_n_u64_x(__VA_ARGS__) -#define svrhadd_n_u16_x(...) __builtin_sve_svrhadd_n_u16_x(__VA_ARGS__) -#define svrhadd_n_u8_z(...) __builtin_sve_svrhadd_n_u8_z(__VA_ARGS__) -#define svrhadd_n_u32_z(...) __builtin_sve_svrhadd_n_u32_z(__VA_ARGS__) -#define svrhadd_n_u64_z(...) __builtin_sve_svrhadd_n_u64_z(__VA_ARGS__) -#define svrhadd_n_u16_z(...) __builtin_sve_svrhadd_n_u16_z(__VA_ARGS__) -#define svrhadd_s8_m(...) __builtin_sve_svrhadd_s8_m(__VA_ARGS__) -#define svrhadd_s32_m(...) __builtin_sve_svrhadd_s32_m(__VA_ARGS__) -#define svrhadd_s64_m(...) __builtin_sve_svrhadd_s64_m(__VA_ARGS__) -#define svrhadd_s16_m(...) __builtin_sve_svrhadd_s16_m(__VA_ARGS__) -#define svrhadd_s8_x(...) __builtin_sve_svrhadd_s8_x(__VA_ARGS__) -#define svrhadd_s32_x(...) __builtin_sve_svrhadd_s32_x(__VA_ARGS__) -#define svrhadd_s64_x(...) __builtin_sve_svrhadd_s64_x(__VA_ARGS__) -#define svrhadd_s16_x(...) __builtin_sve_svrhadd_s16_x(__VA_ARGS__) -#define svrhadd_s8_z(...) __builtin_sve_svrhadd_s8_z(__VA_ARGS__) -#define svrhadd_s32_z(...) __builtin_sve_svrhadd_s32_z(__VA_ARGS__) -#define svrhadd_s64_z(...) __builtin_sve_svrhadd_s64_z(__VA_ARGS__) -#define svrhadd_s16_z(...) __builtin_sve_svrhadd_s16_z(__VA_ARGS__) -#define svrhadd_u8_m(...) __builtin_sve_svrhadd_u8_m(__VA_ARGS__) -#define svrhadd_u32_m(...) __builtin_sve_svrhadd_u32_m(__VA_ARGS__) -#define svrhadd_u64_m(...) __builtin_sve_svrhadd_u64_m(__VA_ARGS__) -#define svrhadd_u16_m(...) __builtin_sve_svrhadd_u16_m(__VA_ARGS__) -#define svrhadd_u8_x(...) __builtin_sve_svrhadd_u8_x(__VA_ARGS__) -#define svrhadd_u32_x(...) __builtin_sve_svrhadd_u32_x(__VA_ARGS__) -#define svrhadd_u64_x(...) __builtin_sve_svrhadd_u64_x(__VA_ARGS__) -#define svrhadd_u16_x(...) __builtin_sve_svrhadd_u16_x(__VA_ARGS__) -#define svrhadd_u8_z(...) __builtin_sve_svrhadd_u8_z(__VA_ARGS__) -#define svrhadd_u32_z(...) __builtin_sve_svrhadd_u32_z(__VA_ARGS__) -#define svrhadd_u64_z(...) __builtin_sve_svrhadd_u64_z(__VA_ARGS__) -#define svrhadd_u16_z(...) __builtin_sve_svrhadd_u16_z(__VA_ARGS__) -#define svrshl_n_s8_m(...) __builtin_sve_svrshl_n_s8_m(__VA_ARGS__) -#define svrshl_n_s32_m(...) __builtin_sve_svrshl_n_s32_m(__VA_ARGS__) -#define svrshl_n_s64_m(...) __builtin_sve_svrshl_n_s64_m(__VA_ARGS__) -#define svrshl_n_s16_m(...) __builtin_sve_svrshl_n_s16_m(__VA_ARGS__) -#define svrshl_n_s8_x(...) __builtin_sve_svrshl_n_s8_x(__VA_ARGS__) -#define svrshl_n_s32_x(...) __builtin_sve_svrshl_n_s32_x(__VA_ARGS__) -#define svrshl_n_s64_x(...) __builtin_sve_svrshl_n_s64_x(__VA_ARGS__) -#define svrshl_n_s16_x(...) __builtin_sve_svrshl_n_s16_x(__VA_ARGS__) -#define svrshl_n_s8_z(...) __builtin_sve_svrshl_n_s8_z(__VA_ARGS__) -#define svrshl_n_s32_z(...) __builtin_sve_svrshl_n_s32_z(__VA_ARGS__) -#define svrshl_n_s64_z(...) __builtin_sve_svrshl_n_s64_z(__VA_ARGS__) -#define svrshl_n_s16_z(...) __builtin_sve_svrshl_n_s16_z(__VA_ARGS__) -#define svrshl_n_u8_m(...) __builtin_sve_svrshl_n_u8_m(__VA_ARGS__) -#define svrshl_n_u32_m(...) __builtin_sve_svrshl_n_u32_m(__VA_ARGS__) -#define svrshl_n_u64_m(...) __builtin_sve_svrshl_n_u64_m(__VA_ARGS__) -#define svrshl_n_u16_m(...) __builtin_sve_svrshl_n_u16_m(__VA_ARGS__) -#define svrshl_n_u8_x(...) __builtin_sve_svrshl_n_u8_x(__VA_ARGS__) -#define svrshl_n_u32_x(...) __builtin_sve_svrshl_n_u32_x(__VA_ARGS__) -#define svrshl_n_u64_x(...) __builtin_sve_svrshl_n_u64_x(__VA_ARGS__) -#define svrshl_n_u16_x(...) __builtin_sve_svrshl_n_u16_x(__VA_ARGS__) -#define svrshl_n_u8_z(...) __builtin_sve_svrshl_n_u8_z(__VA_ARGS__) -#define svrshl_n_u32_z(...) __builtin_sve_svrshl_n_u32_z(__VA_ARGS__) -#define svrshl_n_u64_z(...) __builtin_sve_svrshl_n_u64_z(__VA_ARGS__) -#define svrshl_n_u16_z(...) __builtin_sve_svrshl_n_u16_z(__VA_ARGS__) -#define svrshl_s8_m(...) __builtin_sve_svrshl_s8_m(__VA_ARGS__) -#define svrshl_s32_m(...) __builtin_sve_svrshl_s32_m(__VA_ARGS__) -#define svrshl_s64_m(...) __builtin_sve_svrshl_s64_m(__VA_ARGS__) -#define svrshl_s16_m(...) __builtin_sve_svrshl_s16_m(__VA_ARGS__) -#define svrshl_s8_x(...) __builtin_sve_svrshl_s8_x(__VA_ARGS__) -#define svrshl_s32_x(...) __builtin_sve_svrshl_s32_x(__VA_ARGS__) -#define svrshl_s64_x(...) __builtin_sve_svrshl_s64_x(__VA_ARGS__) -#define svrshl_s16_x(...) __builtin_sve_svrshl_s16_x(__VA_ARGS__) -#define svrshl_s8_z(...) __builtin_sve_svrshl_s8_z(__VA_ARGS__) -#define svrshl_s32_z(...) __builtin_sve_svrshl_s32_z(__VA_ARGS__) -#define svrshl_s64_z(...) __builtin_sve_svrshl_s64_z(__VA_ARGS__) -#define svrshl_s16_z(...) __builtin_sve_svrshl_s16_z(__VA_ARGS__) -#define svrshl_u8_m(...) __builtin_sve_svrshl_u8_m(__VA_ARGS__) -#define svrshl_u32_m(...) __builtin_sve_svrshl_u32_m(__VA_ARGS__) -#define svrshl_u64_m(...) __builtin_sve_svrshl_u64_m(__VA_ARGS__) -#define svrshl_u16_m(...) __builtin_sve_svrshl_u16_m(__VA_ARGS__) -#define svrshl_u8_x(...) __builtin_sve_svrshl_u8_x(__VA_ARGS__) -#define svrshl_u32_x(...) __builtin_sve_svrshl_u32_x(__VA_ARGS__) -#define svrshl_u64_x(...) __builtin_sve_svrshl_u64_x(__VA_ARGS__) -#define svrshl_u16_x(...) __builtin_sve_svrshl_u16_x(__VA_ARGS__) -#define svrshl_u8_z(...) __builtin_sve_svrshl_u8_z(__VA_ARGS__) -#define svrshl_u32_z(...) __builtin_sve_svrshl_u32_z(__VA_ARGS__) -#define svrshl_u64_z(...) __builtin_sve_svrshl_u64_z(__VA_ARGS__) -#define svrshl_u16_z(...) __builtin_sve_svrshl_u16_z(__VA_ARGS__) -#define svrshr_n_s8_m(...) __builtin_sve_svrshr_n_s8_m(__VA_ARGS__) -#define svrshr_n_s32_m(...) __builtin_sve_svrshr_n_s32_m(__VA_ARGS__) -#define svrshr_n_s64_m(...) __builtin_sve_svrshr_n_s64_m(__VA_ARGS__) -#define svrshr_n_s16_m(...) __builtin_sve_svrshr_n_s16_m(__VA_ARGS__) -#define svrshr_n_u8_m(...) __builtin_sve_svrshr_n_u8_m(__VA_ARGS__) -#define svrshr_n_u32_m(...) __builtin_sve_svrshr_n_u32_m(__VA_ARGS__) -#define svrshr_n_u64_m(...) __builtin_sve_svrshr_n_u64_m(__VA_ARGS__) -#define svrshr_n_u16_m(...) __builtin_sve_svrshr_n_u16_m(__VA_ARGS__) -#define svrshr_n_s8_x(...) __builtin_sve_svrshr_n_s8_x(__VA_ARGS__) -#define svrshr_n_s32_x(...) __builtin_sve_svrshr_n_s32_x(__VA_ARGS__) -#define svrshr_n_s64_x(...) __builtin_sve_svrshr_n_s64_x(__VA_ARGS__) -#define svrshr_n_s16_x(...) __builtin_sve_svrshr_n_s16_x(__VA_ARGS__) -#define svrshr_n_u8_x(...) __builtin_sve_svrshr_n_u8_x(__VA_ARGS__) -#define svrshr_n_u32_x(...) __builtin_sve_svrshr_n_u32_x(__VA_ARGS__) -#define svrshr_n_u64_x(...) __builtin_sve_svrshr_n_u64_x(__VA_ARGS__) -#define svrshr_n_u16_x(...) __builtin_sve_svrshr_n_u16_x(__VA_ARGS__) -#define svrshr_n_s8_z(...) __builtin_sve_svrshr_n_s8_z(__VA_ARGS__) -#define svrshr_n_s32_z(...) __builtin_sve_svrshr_n_s32_z(__VA_ARGS__) -#define svrshr_n_s64_z(...) __builtin_sve_svrshr_n_s64_z(__VA_ARGS__) -#define svrshr_n_s16_z(...) __builtin_sve_svrshr_n_s16_z(__VA_ARGS__) -#define svrshr_n_u8_z(...) __builtin_sve_svrshr_n_u8_z(__VA_ARGS__) -#define svrshr_n_u32_z(...) __builtin_sve_svrshr_n_u32_z(__VA_ARGS__) -#define svrshr_n_u64_z(...) __builtin_sve_svrshr_n_u64_z(__VA_ARGS__) -#define svrshr_n_u16_z(...) __builtin_sve_svrshr_n_u16_z(__VA_ARGS__) -#define svrshrnb_n_u32(...) __builtin_sve_svrshrnb_n_u32(__VA_ARGS__) -#define svrshrnb_n_u64(...) __builtin_sve_svrshrnb_n_u64(__VA_ARGS__) -#define svrshrnb_n_u16(...) __builtin_sve_svrshrnb_n_u16(__VA_ARGS__) -#define svrshrnb_n_s32(...) __builtin_sve_svrshrnb_n_s32(__VA_ARGS__) -#define svrshrnb_n_s64(...) __builtin_sve_svrshrnb_n_s64(__VA_ARGS__) -#define svrshrnb_n_s16(...) __builtin_sve_svrshrnb_n_s16(__VA_ARGS__) -#define svrshrnt_n_u32(...) __builtin_sve_svrshrnt_n_u32(__VA_ARGS__) -#define svrshrnt_n_u64(...) __builtin_sve_svrshrnt_n_u64(__VA_ARGS__) -#define svrshrnt_n_u16(...) __builtin_sve_svrshrnt_n_u16(__VA_ARGS__) -#define svrshrnt_n_s32(...) __builtin_sve_svrshrnt_n_s32(__VA_ARGS__) -#define svrshrnt_n_s64(...) __builtin_sve_svrshrnt_n_s64(__VA_ARGS__) -#define svrshrnt_n_s16(...) __builtin_sve_svrshrnt_n_s16(__VA_ARGS__) -#define svrsqrte_u32_m(...) __builtin_sve_svrsqrte_u32_m(__VA_ARGS__) -#define svrsqrte_u32_x(...) __builtin_sve_svrsqrte_u32_x(__VA_ARGS__) -#define svrsqrte_u32_z(...) __builtin_sve_svrsqrte_u32_z(__VA_ARGS__) -#define svrsra_n_s8(...) __builtin_sve_svrsra_n_s8(__VA_ARGS__) -#define svrsra_n_s32(...) __builtin_sve_svrsra_n_s32(__VA_ARGS__) -#define svrsra_n_s64(...) __builtin_sve_svrsra_n_s64(__VA_ARGS__) -#define svrsra_n_s16(...) __builtin_sve_svrsra_n_s16(__VA_ARGS__) -#define svrsra_n_u8(...) __builtin_sve_svrsra_n_u8(__VA_ARGS__) -#define svrsra_n_u32(...) __builtin_sve_svrsra_n_u32(__VA_ARGS__) -#define svrsra_n_u64(...) __builtin_sve_svrsra_n_u64(__VA_ARGS__) -#define svrsra_n_u16(...) __builtin_sve_svrsra_n_u16(__VA_ARGS__) -#define svrsubhnb_n_u32(...) __builtin_sve_svrsubhnb_n_u32(__VA_ARGS__) -#define svrsubhnb_n_u64(...) __builtin_sve_svrsubhnb_n_u64(__VA_ARGS__) -#define svrsubhnb_n_u16(...) __builtin_sve_svrsubhnb_n_u16(__VA_ARGS__) -#define svrsubhnb_n_s32(...) __builtin_sve_svrsubhnb_n_s32(__VA_ARGS__) -#define svrsubhnb_n_s64(...) __builtin_sve_svrsubhnb_n_s64(__VA_ARGS__) -#define svrsubhnb_n_s16(...) __builtin_sve_svrsubhnb_n_s16(__VA_ARGS__) -#define svrsubhnb_u32(...) __builtin_sve_svrsubhnb_u32(__VA_ARGS__) -#define svrsubhnb_u64(...) __builtin_sve_svrsubhnb_u64(__VA_ARGS__) -#define svrsubhnb_u16(...) __builtin_sve_svrsubhnb_u16(__VA_ARGS__) -#define svrsubhnb_s32(...) __builtin_sve_svrsubhnb_s32(__VA_ARGS__) -#define svrsubhnb_s64(...) __builtin_sve_svrsubhnb_s64(__VA_ARGS__) -#define svrsubhnb_s16(...) __builtin_sve_svrsubhnb_s16(__VA_ARGS__) -#define svrsubhnt_n_u32(...) __builtin_sve_svrsubhnt_n_u32(__VA_ARGS__) -#define svrsubhnt_n_u64(...) __builtin_sve_svrsubhnt_n_u64(__VA_ARGS__) -#define svrsubhnt_n_u16(...) __builtin_sve_svrsubhnt_n_u16(__VA_ARGS__) -#define svrsubhnt_n_s32(...) __builtin_sve_svrsubhnt_n_s32(__VA_ARGS__) -#define svrsubhnt_n_s64(...) __builtin_sve_svrsubhnt_n_s64(__VA_ARGS__) -#define svrsubhnt_n_s16(...) __builtin_sve_svrsubhnt_n_s16(__VA_ARGS__) -#define svrsubhnt_u32(...) __builtin_sve_svrsubhnt_u32(__VA_ARGS__) -#define svrsubhnt_u64(...) __builtin_sve_svrsubhnt_u64(__VA_ARGS__) -#define svrsubhnt_u16(...) __builtin_sve_svrsubhnt_u16(__VA_ARGS__) -#define svrsubhnt_s32(...) __builtin_sve_svrsubhnt_s32(__VA_ARGS__) -#define svrsubhnt_s64(...) __builtin_sve_svrsubhnt_s64(__VA_ARGS__) -#define svrsubhnt_s16(...) __builtin_sve_svrsubhnt_s16(__VA_ARGS__) -#define svsbclb_n_u32(...) __builtin_sve_svsbclb_n_u32(__VA_ARGS__) -#define svsbclb_n_u64(...) __builtin_sve_svsbclb_n_u64(__VA_ARGS__) -#define svsbclb_u32(...) __builtin_sve_svsbclb_u32(__VA_ARGS__) -#define svsbclb_u64(...) __builtin_sve_svsbclb_u64(__VA_ARGS__) -#define svsbclt_n_u32(...) __builtin_sve_svsbclt_n_u32(__VA_ARGS__) -#define svsbclt_n_u64(...) __builtin_sve_svsbclt_n_u64(__VA_ARGS__) -#define svsbclt_u32(...) __builtin_sve_svsbclt_u32(__VA_ARGS__) -#define svsbclt_u64(...) __builtin_sve_svsbclt_u64(__VA_ARGS__) -#define svshllb_n_s32(...) __builtin_sve_svshllb_n_s32(__VA_ARGS__) -#define svshllb_n_s64(...) __builtin_sve_svshllb_n_s64(__VA_ARGS__) -#define svshllb_n_s16(...) __builtin_sve_svshllb_n_s16(__VA_ARGS__) -#define svshllb_n_u32(...) __builtin_sve_svshllb_n_u32(__VA_ARGS__) -#define svshllb_n_u64(...) __builtin_sve_svshllb_n_u64(__VA_ARGS__) -#define svshllb_n_u16(...) __builtin_sve_svshllb_n_u16(__VA_ARGS__) -#define svshllt_n_s32(...) __builtin_sve_svshllt_n_s32(__VA_ARGS__) -#define svshllt_n_s64(...) __builtin_sve_svshllt_n_s64(__VA_ARGS__) -#define svshllt_n_s16(...) __builtin_sve_svshllt_n_s16(__VA_ARGS__) -#define svshllt_n_u32(...) __builtin_sve_svshllt_n_u32(__VA_ARGS__) -#define svshllt_n_u64(...) __builtin_sve_svshllt_n_u64(__VA_ARGS__) -#define svshllt_n_u16(...) __builtin_sve_svshllt_n_u16(__VA_ARGS__) -#define svshrnb_n_u32(...) __builtin_sve_svshrnb_n_u32(__VA_ARGS__) -#define svshrnb_n_u64(...) __builtin_sve_svshrnb_n_u64(__VA_ARGS__) -#define svshrnb_n_u16(...) __builtin_sve_svshrnb_n_u16(__VA_ARGS__) -#define svshrnb_n_s32(...) __builtin_sve_svshrnb_n_s32(__VA_ARGS__) -#define svshrnb_n_s64(...) __builtin_sve_svshrnb_n_s64(__VA_ARGS__) -#define svshrnb_n_s16(...) __builtin_sve_svshrnb_n_s16(__VA_ARGS__) -#define svshrnt_n_u32(...) __builtin_sve_svshrnt_n_u32(__VA_ARGS__) -#define svshrnt_n_u64(...) __builtin_sve_svshrnt_n_u64(__VA_ARGS__) -#define svshrnt_n_u16(...) __builtin_sve_svshrnt_n_u16(__VA_ARGS__) -#define svshrnt_n_s32(...) __builtin_sve_svshrnt_n_s32(__VA_ARGS__) -#define svshrnt_n_s64(...) __builtin_sve_svshrnt_n_s64(__VA_ARGS__) -#define svshrnt_n_s16(...) __builtin_sve_svshrnt_n_s16(__VA_ARGS__) -#define svsli_n_u8(...) __builtin_sve_svsli_n_u8(__VA_ARGS__) -#define svsli_n_u32(...) __builtin_sve_svsli_n_u32(__VA_ARGS__) -#define svsli_n_u64(...) __builtin_sve_svsli_n_u64(__VA_ARGS__) -#define svsli_n_u16(...) __builtin_sve_svsli_n_u16(__VA_ARGS__) -#define svsli_n_s8(...) __builtin_sve_svsli_n_s8(__VA_ARGS__) -#define svsli_n_s32(...) __builtin_sve_svsli_n_s32(__VA_ARGS__) -#define svsli_n_s64(...) __builtin_sve_svsli_n_s64(__VA_ARGS__) -#define svsli_n_s16(...) __builtin_sve_svsli_n_s16(__VA_ARGS__) -#define svsqadd_n_u8_m(...) __builtin_sve_svsqadd_n_u8_m(__VA_ARGS__) -#define svsqadd_n_u32_m(...) __builtin_sve_svsqadd_n_u32_m(__VA_ARGS__) -#define svsqadd_n_u64_m(...) __builtin_sve_svsqadd_n_u64_m(__VA_ARGS__) -#define svsqadd_n_u16_m(...) __builtin_sve_svsqadd_n_u16_m(__VA_ARGS__) -#define svsqadd_n_u8_x(...) __builtin_sve_svsqadd_n_u8_x(__VA_ARGS__) -#define svsqadd_n_u32_x(...) __builtin_sve_svsqadd_n_u32_x(__VA_ARGS__) -#define svsqadd_n_u64_x(...) __builtin_sve_svsqadd_n_u64_x(__VA_ARGS__) -#define svsqadd_n_u16_x(...) __builtin_sve_svsqadd_n_u16_x(__VA_ARGS__) -#define svsqadd_n_u8_z(...) __builtin_sve_svsqadd_n_u8_z(__VA_ARGS__) -#define svsqadd_n_u32_z(...) __builtin_sve_svsqadd_n_u32_z(__VA_ARGS__) -#define svsqadd_n_u64_z(...) __builtin_sve_svsqadd_n_u64_z(__VA_ARGS__) -#define svsqadd_n_u16_z(...) __builtin_sve_svsqadd_n_u16_z(__VA_ARGS__) -#define svsqadd_u8_m(...) __builtin_sve_svsqadd_u8_m(__VA_ARGS__) -#define svsqadd_u32_m(...) __builtin_sve_svsqadd_u32_m(__VA_ARGS__) -#define svsqadd_u64_m(...) __builtin_sve_svsqadd_u64_m(__VA_ARGS__) -#define svsqadd_u16_m(...) __builtin_sve_svsqadd_u16_m(__VA_ARGS__) -#define svsqadd_u8_x(...) __builtin_sve_svsqadd_u8_x(__VA_ARGS__) -#define svsqadd_u32_x(...) __builtin_sve_svsqadd_u32_x(__VA_ARGS__) -#define svsqadd_u64_x(...) __builtin_sve_svsqadd_u64_x(__VA_ARGS__) -#define svsqadd_u16_x(...) __builtin_sve_svsqadd_u16_x(__VA_ARGS__) -#define svsqadd_u8_z(...) __builtin_sve_svsqadd_u8_z(__VA_ARGS__) -#define svsqadd_u32_z(...) __builtin_sve_svsqadd_u32_z(__VA_ARGS__) -#define svsqadd_u64_z(...) __builtin_sve_svsqadd_u64_z(__VA_ARGS__) -#define svsqadd_u16_z(...) __builtin_sve_svsqadd_u16_z(__VA_ARGS__) -#define svsra_n_s8(...) __builtin_sve_svsra_n_s8(__VA_ARGS__) -#define svsra_n_s32(...) __builtin_sve_svsra_n_s32(__VA_ARGS__) -#define svsra_n_s64(...) __builtin_sve_svsra_n_s64(__VA_ARGS__) -#define svsra_n_s16(...) __builtin_sve_svsra_n_s16(__VA_ARGS__) -#define svsra_n_u8(...) __builtin_sve_svsra_n_u8(__VA_ARGS__) -#define svsra_n_u32(...) __builtin_sve_svsra_n_u32(__VA_ARGS__) -#define svsra_n_u64(...) __builtin_sve_svsra_n_u64(__VA_ARGS__) -#define svsra_n_u16(...) __builtin_sve_svsra_n_u16(__VA_ARGS__) -#define svsri_n_u8(...) __builtin_sve_svsri_n_u8(__VA_ARGS__) -#define svsri_n_u32(...) __builtin_sve_svsri_n_u32(__VA_ARGS__) -#define svsri_n_u64(...) __builtin_sve_svsri_n_u64(__VA_ARGS__) -#define svsri_n_u16(...) __builtin_sve_svsri_n_u16(__VA_ARGS__) -#define svsri_n_s8(...) __builtin_sve_svsri_n_s8(__VA_ARGS__) -#define svsri_n_s32(...) __builtin_sve_svsri_n_s32(__VA_ARGS__) -#define svsri_n_s64(...) __builtin_sve_svsri_n_s64(__VA_ARGS__) -#define svsri_n_s16(...) __builtin_sve_svsri_n_s16(__VA_ARGS__) -#define svstnt1_scatter_u32base_index_u32(...) __builtin_sve_svstnt1_scatter_u32base_index_u32(__VA_ARGS__) -#define svstnt1_scatter_u64base_index_u64(...) __builtin_sve_svstnt1_scatter_u64base_index_u64(__VA_ARGS__) -#define svstnt1_scatter_u64base_index_f64(...) __builtin_sve_svstnt1_scatter_u64base_index_f64(__VA_ARGS__) -#define svstnt1_scatter_u32base_index_f32(...) __builtin_sve_svstnt1_scatter_u32base_index_f32(__VA_ARGS__) -#define svstnt1_scatter_u32base_index_s32(...) __builtin_sve_svstnt1_scatter_u32base_index_s32(__VA_ARGS__) -#define svstnt1_scatter_u64base_index_s64(...) __builtin_sve_svstnt1_scatter_u64base_index_s64(__VA_ARGS__) -#define svstnt1_scatter_u32base_offset_u32(...) __builtin_sve_svstnt1_scatter_u32base_offset_u32(__VA_ARGS__) -#define svstnt1_scatter_u64base_offset_u64(...) __builtin_sve_svstnt1_scatter_u64base_offset_u64(__VA_ARGS__) -#define svstnt1_scatter_u64base_offset_f64(...) __builtin_sve_svstnt1_scatter_u64base_offset_f64(__VA_ARGS__) -#define svstnt1_scatter_u32base_offset_f32(...) __builtin_sve_svstnt1_scatter_u32base_offset_f32(__VA_ARGS__) -#define svstnt1_scatter_u32base_offset_s32(...) __builtin_sve_svstnt1_scatter_u32base_offset_s32(__VA_ARGS__) -#define svstnt1_scatter_u64base_offset_s64(...) __builtin_sve_svstnt1_scatter_u64base_offset_s64(__VA_ARGS__) -#define svstnt1_scatter_u32base_u32(...) __builtin_sve_svstnt1_scatter_u32base_u32(__VA_ARGS__) -#define svstnt1_scatter_u64base_u64(...) __builtin_sve_svstnt1_scatter_u64base_u64(__VA_ARGS__) -#define svstnt1_scatter_u64base_f64(...) __builtin_sve_svstnt1_scatter_u64base_f64(__VA_ARGS__) -#define svstnt1_scatter_u32base_f32(...) __builtin_sve_svstnt1_scatter_u32base_f32(__VA_ARGS__) -#define svstnt1_scatter_u32base_s32(...) __builtin_sve_svstnt1_scatter_u32base_s32(__VA_ARGS__) -#define svstnt1_scatter_u64base_s64(...) __builtin_sve_svstnt1_scatter_u64base_s64(__VA_ARGS__) -#define svstnt1_scatter_s64index_u64(...) __builtin_sve_svstnt1_scatter_s64index_u64(__VA_ARGS__) -#define svstnt1_scatter_s64index_f64(...) __builtin_sve_svstnt1_scatter_s64index_f64(__VA_ARGS__) -#define svstnt1_scatter_s64index_s64(...) __builtin_sve_svstnt1_scatter_s64index_s64(__VA_ARGS__) -#define svstnt1_scatter_u64index_u64(...) __builtin_sve_svstnt1_scatter_u64index_u64(__VA_ARGS__) -#define svstnt1_scatter_u64index_f64(...) __builtin_sve_svstnt1_scatter_u64index_f64(__VA_ARGS__) -#define svstnt1_scatter_u64index_s64(...) __builtin_sve_svstnt1_scatter_u64index_s64(__VA_ARGS__) -#define svstnt1_scatter_u32offset_u32(...) __builtin_sve_svstnt1_scatter_u32offset_u32(__VA_ARGS__) -#define svstnt1_scatter_u32offset_f32(...) __builtin_sve_svstnt1_scatter_u32offset_f32(__VA_ARGS__) -#define svstnt1_scatter_u32offset_s32(...) __builtin_sve_svstnt1_scatter_u32offset_s32(__VA_ARGS__) -#define svstnt1_scatter_s64offset_u64(...) __builtin_sve_svstnt1_scatter_s64offset_u64(__VA_ARGS__) -#define svstnt1_scatter_s64offset_f64(...) __builtin_sve_svstnt1_scatter_s64offset_f64(__VA_ARGS__) -#define svstnt1_scatter_s64offset_s64(...) __builtin_sve_svstnt1_scatter_s64offset_s64(__VA_ARGS__) -#define svstnt1_scatter_u64offset_u64(...) __builtin_sve_svstnt1_scatter_u64offset_u64(__VA_ARGS__) -#define svstnt1_scatter_u64offset_f64(...) __builtin_sve_svstnt1_scatter_u64offset_f64(__VA_ARGS__) -#define svstnt1_scatter_u64offset_s64(...) __builtin_sve_svstnt1_scatter_u64offset_s64(__VA_ARGS__) -#define svstnt1b_scatter_u32base_offset_u32(...) __builtin_sve_svstnt1b_scatter_u32base_offset_u32(__VA_ARGS__) -#define svstnt1b_scatter_u64base_offset_u64(...) __builtin_sve_svstnt1b_scatter_u64base_offset_u64(__VA_ARGS__) -#define svstnt1b_scatter_u32base_offset_s32(...) __builtin_sve_svstnt1b_scatter_u32base_offset_s32(__VA_ARGS__) -#define svstnt1b_scatter_u64base_offset_s64(...) __builtin_sve_svstnt1b_scatter_u64base_offset_s64(__VA_ARGS__) -#define svstnt1b_scatter_u32base_u32(...) __builtin_sve_svstnt1b_scatter_u32base_u32(__VA_ARGS__) -#define svstnt1b_scatter_u64base_u64(...) __builtin_sve_svstnt1b_scatter_u64base_u64(__VA_ARGS__) -#define svstnt1b_scatter_u32base_s32(...) __builtin_sve_svstnt1b_scatter_u32base_s32(__VA_ARGS__) -#define svstnt1b_scatter_u64base_s64(...) __builtin_sve_svstnt1b_scatter_u64base_s64(__VA_ARGS__) -#define svstnt1b_scatter_u32offset_s32(...) __builtin_sve_svstnt1b_scatter_u32offset_s32(__VA_ARGS__) -#define svstnt1b_scatter_u32offset_u32(...) __builtin_sve_svstnt1b_scatter_u32offset_u32(__VA_ARGS__) -#define svstnt1b_scatter_s64offset_s64(...) __builtin_sve_svstnt1b_scatter_s64offset_s64(__VA_ARGS__) -#define svstnt1b_scatter_s64offset_u64(...) __builtin_sve_svstnt1b_scatter_s64offset_u64(__VA_ARGS__) -#define svstnt1b_scatter_u64offset_s64(...) __builtin_sve_svstnt1b_scatter_u64offset_s64(__VA_ARGS__) -#define svstnt1b_scatter_u64offset_u64(...) __builtin_sve_svstnt1b_scatter_u64offset_u64(__VA_ARGS__) -#define svstnt1h_scatter_u32base_index_u32(...) __builtin_sve_svstnt1h_scatter_u32base_index_u32(__VA_ARGS__) -#define svstnt1h_scatter_u64base_index_u64(...) __builtin_sve_svstnt1h_scatter_u64base_index_u64(__VA_ARGS__) -#define svstnt1h_scatter_u32base_index_s32(...) __builtin_sve_svstnt1h_scatter_u32base_index_s32(__VA_ARGS__) -#define svstnt1h_scatter_u64base_index_s64(...) __builtin_sve_svstnt1h_scatter_u64base_index_s64(__VA_ARGS__) -#define svstnt1h_scatter_u32base_offset_u32(...) __builtin_sve_svstnt1h_scatter_u32base_offset_u32(__VA_ARGS__) -#define svstnt1h_scatter_u64base_offset_u64(...) __builtin_sve_svstnt1h_scatter_u64base_offset_u64(__VA_ARGS__) -#define svstnt1h_scatter_u32base_offset_s32(...) __builtin_sve_svstnt1h_scatter_u32base_offset_s32(__VA_ARGS__) -#define svstnt1h_scatter_u64base_offset_s64(...) __builtin_sve_svstnt1h_scatter_u64base_offset_s64(__VA_ARGS__) -#define svstnt1h_scatter_u32base_u32(...) __builtin_sve_svstnt1h_scatter_u32base_u32(__VA_ARGS__) -#define svstnt1h_scatter_u64base_u64(...) __builtin_sve_svstnt1h_scatter_u64base_u64(__VA_ARGS__) -#define svstnt1h_scatter_u32base_s32(...) __builtin_sve_svstnt1h_scatter_u32base_s32(__VA_ARGS__) -#define svstnt1h_scatter_u64base_s64(...) __builtin_sve_svstnt1h_scatter_u64base_s64(__VA_ARGS__) -#define svstnt1h_scatter_s64index_s64(...) __builtin_sve_svstnt1h_scatter_s64index_s64(__VA_ARGS__) -#define svstnt1h_scatter_s64index_u64(...) __builtin_sve_svstnt1h_scatter_s64index_u64(__VA_ARGS__) -#define svstnt1h_scatter_u64index_s64(...) __builtin_sve_svstnt1h_scatter_u64index_s64(__VA_ARGS__) -#define svstnt1h_scatter_u64index_u64(...) __builtin_sve_svstnt1h_scatter_u64index_u64(__VA_ARGS__) -#define svstnt1h_scatter_u32offset_s32(...) __builtin_sve_svstnt1h_scatter_u32offset_s32(__VA_ARGS__) -#define svstnt1h_scatter_u32offset_u32(...) __builtin_sve_svstnt1h_scatter_u32offset_u32(__VA_ARGS__) -#define svstnt1h_scatter_s64offset_s64(...) __builtin_sve_svstnt1h_scatter_s64offset_s64(__VA_ARGS__) -#define svstnt1h_scatter_s64offset_u64(...) __builtin_sve_svstnt1h_scatter_s64offset_u64(__VA_ARGS__) -#define svstnt1h_scatter_u64offset_s64(...) __builtin_sve_svstnt1h_scatter_u64offset_s64(__VA_ARGS__) -#define svstnt1h_scatter_u64offset_u64(...) __builtin_sve_svstnt1h_scatter_u64offset_u64(__VA_ARGS__) -#define svstnt1w_scatter_u64base_index_u64(...) __builtin_sve_svstnt1w_scatter_u64base_index_u64(__VA_ARGS__) -#define svstnt1w_scatter_u64base_index_s64(...) __builtin_sve_svstnt1w_scatter_u64base_index_s64(__VA_ARGS__) -#define svstnt1w_scatter_u64base_offset_u64(...) __builtin_sve_svstnt1w_scatter_u64base_offset_u64(__VA_ARGS__) -#define svstnt1w_scatter_u64base_offset_s64(...) __builtin_sve_svstnt1w_scatter_u64base_offset_s64(__VA_ARGS__) -#define svstnt1w_scatter_u64base_u64(...) __builtin_sve_svstnt1w_scatter_u64base_u64(__VA_ARGS__) -#define svstnt1w_scatter_u64base_s64(...) __builtin_sve_svstnt1w_scatter_u64base_s64(__VA_ARGS__) -#define svstnt1w_scatter_s64index_s64(...) __builtin_sve_svstnt1w_scatter_s64index_s64(__VA_ARGS__) -#define svstnt1w_scatter_s64index_u64(...) __builtin_sve_svstnt1w_scatter_s64index_u64(__VA_ARGS__) -#define svstnt1w_scatter_u64index_s64(...) __builtin_sve_svstnt1w_scatter_u64index_s64(__VA_ARGS__) -#define svstnt1w_scatter_u64index_u64(...) __builtin_sve_svstnt1w_scatter_u64index_u64(__VA_ARGS__) -#define svstnt1w_scatter_s64offset_s64(...) __builtin_sve_svstnt1w_scatter_s64offset_s64(__VA_ARGS__) -#define svstnt1w_scatter_s64offset_u64(...) __builtin_sve_svstnt1w_scatter_s64offset_u64(__VA_ARGS__) -#define svstnt1w_scatter_u64offset_s64(...) __builtin_sve_svstnt1w_scatter_u64offset_s64(__VA_ARGS__) -#define svstnt1w_scatter_u64offset_u64(...) __builtin_sve_svstnt1w_scatter_u64offset_u64(__VA_ARGS__) -#define svsubhnb_n_u32(...) __builtin_sve_svsubhnb_n_u32(__VA_ARGS__) -#define svsubhnb_n_u64(...) __builtin_sve_svsubhnb_n_u64(__VA_ARGS__) -#define svsubhnb_n_u16(...) __builtin_sve_svsubhnb_n_u16(__VA_ARGS__) -#define svsubhnb_n_s32(...) __builtin_sve_svsubhnb_n_s32(__VA_ARGS__) -#define svsubhnb_n_s64(...) __builtin_sve_svsubhnb_n_s64(__VA_ARGS__) -#define svsubhnb_n_s16(...) __builtin_sve_svsubhnb_n_s16(__VA_ARGS__) -#define svsubhnb_u32(...) __builtin_sve_svsubhnb_u32(__VA_ARGS__) -#define svsubhnb_u64(...) __builtin_sve_svsubhnb_u64(__VA_ARGS__) -#define svsubhnb_u16(...) __builtin_sve_svsubhnb_u16(__VA_ARGS__) -#define svsubhnb_s32(...) __builtin_sve_svsubhnb_s32(__VA_ARGS__) -#define svsubhnb_s64(...) __builtin_sve_svsubhnb_s64(__VA_ARGS__) -#define svsubhnb_s16(...) __builtin_sve_svsubhnb_s16(__VA_ARGS__) -#define svsubhnt_n_u32(...) __builtin_sve_svsubhnt_n_u32(__VA_ARGS__) -#define svsubhnt_n_u64(...) __builtin_sve_svsubhnt_n_u64(__VA_ARGS__) -#define svsubhnt_n_u16(...) __builtin_sve_svsubhnt_n_u16(__VA_ARGS__) -#define svsubhnt_n_s32(...) __builtin_sve_svsubhnt_n_s32(__VA_ARGS__) -#define svsubhnt_n_s64(...) __builtin_sve_svsubhnt_n_s64(__VA_ARGS__) -#define svsubhnt_n_s16(...) __builtin_sve_svsubhnt_n_s16(__VA_ARGS__) -#define svsubhnt_u32(...) __builtin_sve_svsubhnt_u32(__VA_ARGS__) -#define svsubhnt_u64(...) __builtin_sve_svsubhnt_u64(__VA_ARGS__) -#define svsubhnt_u16(...) __builtin_sve_svsubhnt_u16(__VA_ARGS__) -#define svsubhnt_s32(...) __builtin_sve_svsubhnt_s32(__VA_ARGS__) -#define svsubhnt_s64(...) __builtin_sve_svsubhnt_s64(__VA_ARGS__) -#define svsubhnt_s16(...) __builtin_sve_svsubhnt_s16(__VA_ARGS__) -#define svsublb_n_s32(...) __builtin_sve_svsublb_n_s32(__VA_ARGS__) -#define svsublb_n_s64(...) __builtin_sve_svsublb_n_s64(__VA_ARGS__) -#define svsublb_n_s16(...) __builtin_sve_svsublb_n_s16(__VA_ARGS__) -#define svsublb_n_u32(...) __builtin_sve_svsublb_n_u32(__VA_ARGS__) -#define svsublb_n_u64(...) __builtin_sve_svsublb_n_u64(__VA_ARGS__) -#define svsublb_n_u16(...) __builtin_sve_svsublb_n_u16(__VA_ARGS__) -#define svsublb_s32(...) __builtin_sve_svsublb_s32(__VA_ARGS__) -#define svsublb_s64(...) __builtin_sve_svsublb_s64(__VA_ARGS__) -#define svsublb_s16(...) __builtin_sve_svsublb_s16(__VA_ARGS__) -#define svsublb_u32(...) __builtin_sve_svsublb_u32(__VA_ARGS__) -#define svsublb_u64(...) __builtin_sve_svsublb_u64(__VA_ARGS__) -#define svsublb_u16(...) __builtin_sve_svsublb_u16(__VA_ARGS__) -#define svsublbt_n_s32(...) __builtin_sve_svsublbt_n_s32(__VA_ARGS__) -#define svsublbt_n_s64(...) __builtin_sve_svsublbt_n_s64(__VA_ARGS__) -#define svsublbt_n_s16(...) __builtin_sve_svsublbt_n_s16(__VA_ARGS__) -#define svsublbt_s32(...) __builtin_sve_svsublbt_s32(__VA_ARGS__) -#define svsublbt_s64(...) __builtin_sve_svsublbt_s64(__VA_ARGS__) -#define svsublbt_s16(...) __builtin_sve_svsublbt_s16(__VA_ARGS__) -#define svsublt_n_s32(...) __builtin_sve_svsublt_n_s32(__VA_ARGS__) -#define svsublt_n_s64(...) __builtin_sve_svsublt_n_s64(__VA_ARGS__) -#define svsublt_n_s16(...) __builtin_sve_svsublt_n_s16(__VA_ARGS__) -#define svsublt_n_u32(...) __builtin_sve_svsublt_n_u32(__VA_ARGS__) -#define svsublt_n_u64(...) __builtin_sve_svsublt_n_u64(__VA_ARGS__) -#define svsublt_n_u16(...) __builtin_sve_svsublt_n_u16(__VA_ARGS__) -#define svsublt_s32(...) __builtin_sve_svsublt_s32(__VA_ARGS__) -#define svsublt_s64(...) __builtin_sve_svsublt_s64(__VA_ARGS__) -#define svsublt_s16(...) __builtin_sve_svsublt_s16(__VA_ARGS__) -#define svsublt_u32(...) __builtin_sve_svsublt_u32(__VA_ARGS__) -#define svsublt_u64(...) __builtin_sve_svsublt_u64(__VA_ARGS__) -#define svsublt_u16(...) __builtin_sve_svsublt_u16(__VA_ARGS__) -#define svsubltb_n_s32(...) __builtin_sve_svsubltb_n_s32(__VA_ARGS__) -#define svsubltb_n_s64(...) __builtin_sve_svsubltb_n_s64(__VA_ARGS__) -#define svsubltb_n_s16(...) __builtin_sve_svsubltb_n_s16(__VA_ARGS__) -#define svsubltb_s32(...) __builtin_sve_svsubltb_s32(__VA_ARGS__) -#define svsubltb_s64(...) __builtin_sve_svsubltb_s64(__VA_ARGS__) -#define svsubltb_s16(...) __builtin_sve_svsubltb_s16(__VA_ARGS__) -#define svsubwb_n_s32(...) __builtin_sve_svsubwb_n_s32(__VA_ARGS__) -#define svsubwb_n_s64(...) __builtin_sve_svsubwb_n_s64(__VA_ARGS__) -#define svsubwb_n_s16(...) __builtin_sve_svsubwb_n_s16(__VA_ARGS__) -#define svsubwb_n_u32(...) __builtin_sve_svsubwb_n_u32(__VA_ARGS__) -#define svsubwb_n_u64(...) __builtin_sve_svsubwb_n_u64(__VA_ARGS__) -#define svsubwb_n_u16(...) __builtin_sve_svsubwb_n_u16(__VA_ARGS__) -#define svsubwb_s32(...) __builtin_sve_svsubwb_s32(__VA_ARGS__) -#define svsubwb_s64(...) __builtin_sve_svsubwb_s64(__VA_ARGS__) -#define svsubwb_s16(...) __builtin_sve_svsubwb_s16(__VA_ARGS__) -#define svsubwb_u32(...) __builtin_sve_svsubwb_u32(__VA_ARGS__) -#define svsubwb_u64(...) __builtin_sve_svsubwb_u64(__VA_ARGS__) -#define svsubwb_u16(...) __builtin_sve_svsubwb_u16(__VA_ARGS__) -#define svsubwt_n_s32(...) __builtin_sve_svsubwt_n_s32(__VA_ARGS__) -#define svsubwt_n_s64(...) __builtin_sve_svsubwt_n_s64(__VA_ARGS__) -#define svsubwt_n_s16(...) __builtin_sve_svsubwt_n_s16(__VA_ARGS__) -#define svsubwt_n_u32(...) __builtin_sve_svsubwt_n_u32(__VA_ARGS__) -#define svsubwt_n_u64(...) __builtin_sve_svsubwt_n_u64(__VA_ARGS__) -#define svsubwt_n_u16(...) __builtin_sve_svsubwt_n_u16(__VA_ARGS__) -#define svsubwt_s32(...) __builtin_sve_svsubwt_s32(__VA_ARGS__) -#define svsubwt_s64(...) __builtin_sve_svsubwt_s64(__VA_ARGS__) -#define svsubwt_s16(...) __builtin_sve_svsubwt_s16(__VA_ARGS__) -#define svsubwt_u32(...) __builtin_sve_svsubwt_u32(__VA_ARGS__) -#define svsubwt_u64(...) __builtin_sve_svsubwt_u64(__VA_ARGS__) -#define svsubwt_u16(...) __builtin_sve_svsubwt_u16(__VA_ARGS__) -#define svtbl2_u8(...) __builtin_sve_svtbl2_u8(__VA_ARGS__) -#define svtbl2_u32(...) __builtin_sve_svtbl2_u32(__VA_ARGS__) -#define svtbl2_u64(...) __builtin_sve_svtbl2_u64(__VA_ARGS__) -#define svtbl2_u16(...) __builtin_sve_svtbl2_u16(__VA_ARGS__) -#define svtbl2_s8(...) __builtin_sve_svtbl2_s8(__VA_ARGS__) -#define svtbl2_f64(...) __builtin_sve_svtbl2_f64(__VA_ARGS__) -#define svtbl2_f32(...) __builtin_sve_svtbl2_f32(__VA_ARGS__) -#define svtbl2_f16(...) __builtin_sve_svtbl2_f16(__VA_ARGS__) -#define svtbl2_s32(...) __builtin_sve_svtbl2_s32(__VA_ARGS__) -#define svtbl2_s64(...) __builtin_sve_svtbl2_s64(__VA_ARGS__) -#define svtbl2_s16(...) __builtin_sve_svtbl2_s16(__VA_ARGS__) -#define svtbx_u8(...) __builtin_sve_svtbx_u8(__VA_ARGS__) -#define svtbx_u32(...) __builtin_sve_svtbx_u32(__VA_ARGS__) -#define svtbx_u64(...) __builtin_sve_svtbx_u64(__VA_ARGS__) -#define svtbx_u16(...) __builtin_sve_svtbx_u16(__VA_ARGS__) -#define svtbx_s8(...) __builtin_sve_svtbx_s8(__VA_ARGS__) -#define svtbx_f64(...) __builtin_sve_svtbx_f64(__VA_ARGS__) -#define svtbx_f32(...) __builtin_sve_svtbx_f32(__VA_ARGS__) -#define svtbx_f16(...) __builtin_sve_svtbx_f16(__VA_ARGS__) -#define svtbx_s32(...) __builtin_sve_svtbx_s32(__VA_ARGS__) -#define svtbx_s64(...) __builtin_sve_svtbx_s64(__VA_ARGS__) -#define svtbx_s16(...) __builtin_sve_svtbx_s16(__VA_ARGS__) -#define svuqadd_n_s8_m(...) __builtin_sve_svuqadd_n_s8_m(__VA_ARGS__) -#define svuqadd_n_s32_m(...) __builtin_sve_svuqadd_n_s32_m(__VA_ARGS__) -#define svuqadd_n_s64_m(...) __builtin_sve_svuqadd_n_s64_m(__VA_ARGS__) -#define svuqadd_n_s16_m(...) __builtin_sve_svuqadd_n_s16_m(__VA_ARGS__) -#define svuqadd_n_s8_x(...) __builtin_sve_svuqadd_n_s8_x(__VA_ARGS__) -#define svuqadd_n_s32_x(...) __builtin_sve_svuqadd_n_s32_x(__VA_ARGS__) -#define svuqadd_n_s64_x(...) __builtin_sve_svuqadd_n_s64_x(__VA_ARGS__) -#define svuqadd_n_s16_x(...) __builtin_sve_svuqadd_n_s16_x(__VA_ARGS__) -#define svuqadd_n_s8_z(...) __builtin_sve_svuqadd_n_s8_z(__VA_ARGS__) -#define svuqadd_n_s32_z(...) __builtin_sve_svuqadd_n_s32_z(__VA_ARGS__) -#define svuqadd_n_s64_z(...) __builtin_sve_svuqadd_n_s64_z(__VA_ARGS__) -#define svuqadd_n_s16_z(...) __builtin_sve_svuqadd_n_s16_z(__VA_ARGS__) -#define svuqadd_s8_m(...) __builtin_sve_svuqadd_s8_m(__VA_ARGS__) -#define svuqadd_s32_m(...) __builtin_sve_svuqadd_s32_m(__VA_ARGS__) -#define svuqadd_s64_m(...) __builtin_sve_svuqadd_s64_m(__VA_ARGS__) -#define svuqadd_s16_m(...) __builtin_sve_svuqadd_s16_m(__VA_ARGS__) -#define svuqadd_s8_x(...) __builtin_sve_svuqadd_s8_x(__VA_ARGS__) -#define svuqadd_s32_x(...) __builtin_sve_svuqadd_s32_x(__VA_ARGS__) -#define svuqadd_s64_x(...) __builtin_sve_svuqadd_s64_x(__VA_ARGS__) -#define svuqadd_s16_x(...) __builtin_sve_svuqadd_s16_x(__VA_ARGS__) -#define svuqadd_s8_z(...) __builtin_sve_svuqadd_s8_z(__VA_ARGS__) -#define svuqadd_s32_z(...) __builtin_sve_svuqadd_s32_z(__VA_ARGS__) -#define svuqadd_s64_z(...) __builtin_sve_svuqadd_s64_z(__VA_ARGS__) -#define svuqadd_s16_z(...) __builtin_sve_svuqadd_s16_z(__VA_ARGS__) -#define svwhilege_b8_s32(...) __builtin_sve_svwhilege_b8_s32(__VA_ARGS__) -#define svwhilege_b32_s32(...) __builtin_sve_svwhilege_b32_s32(__VA_ARGS__) -#define svwhilege_b64_s32(...) __builtin_sve_svwhilege_b64_s32(__VA_ARGS__) -#define svwhilege_b16_s32(...) __builtin_sve_svwhilege_b16_s32(__VA_ARGS__) -#define svwhilege_b8_s64(...) __builtin_sve_svwhilege_b8_s64(__VA_ARGS__) -#define svwhilege_b32_s64(...) __builtin_sve_svwhilege_b32_s64(__VA_ARGS__) -#define svwhilege_b64_s64(...) __builtin_sve_svwhilege_b64_s64(__VA_ARGS__) -#define svwhilege_b16_s64(...) __builtin_sve_svwhilege_b16_s64(__VA_ARGS__) -#define svwhilege_b8_u32(...) __builtin_sve_svwhilege_b8_u32(__VA_ARGS__) -#define svwhilege_b32_u32(...) __builtin_sve_svwhilege_b32_u32(__VA_ARGS__) -#define svwhilege_b64_u32(...) __builtin_sve_svwhilege_b64_u32(__VA_ARGS__) -#define svwhilege_b16_u32(...) __builtin_sve_svwhilege_b16_u32(__VA_ARGS__) -#define svwhilege_b8_u64(...) __builtin_sve_svwhilege_b8_u64(__VA_ARGS__) -#define svwhilege_b32_u64(...) __builtin_sve_svwhilege_b32_u64(__VA_ARGS__) -#define svwhilege_b64_u64(...) __builtin_sve_svwhilege_b64_u64(__VA_ARGS__) -#define svwhilege_b16_u64(...) __builtin_sve_svwhilege_b16_u64(__VA_ARGS__) -#define svwhilegt_b8_s32(...) __builtin_sve_svwhilegt_b8_s32(__VA_ARGS__) -#define svwhilegt_b32_s32(...) __builtin_sve_svwhilegt_b32_s32(__VA_ARGS__) -#define svwhilegt_b64_s32(...) __builtin_sve_svwhilegt_b64_s32(__VA_ARGS__) -#define svwhilegt_b16_s32(...) __builtin_sve_svwhilegt_b16_s32(__VA_ARGS__) -#define svwhilegt_b8_s64(...) __builtin_sve_svwhilegt_b8_s64(__VA_ARGS__) -#define svwhilegt_b32_s64(...) __builtin_sve_svwhilegt_b32_s64(__VA_ARGS__) -#define svwhilegt_b64_s64(...) __builtin_sve_svwhilegt_b64_s64(__VA_ARGS__) -#define svwhilegt_b16_s64(...) __builtin_sve_svwhilegt_b16_s64(__VA_ARGS__) -#define svwhilegt_b8_u32(...) __builtin_sve_svwhilegt_b8_u32(__VA_ARGS__) -#define svwhilegt_b32_u32(...) __builtin_sve_svwhilegt_b32_u32(__VA_ARGS__) -#define svwhilegt_b64_u32(...) __builtin_sve_svwhilegt_b64_u32(__VA_ARGS__) -#define svwhilegt_b16_u32(...) __builtin_sve_svwhilegt_b16_u32(__VA_ARGS__) -#define svwhilegt_b8_u64(...) __builtin_sve_svwhilegt_b8_u64(__VA_ARGS__) -#define svwhilegt_b32_u64(...) __builtin_sve_svwhilegt_b32_u64(__VA_ARGS__) -#define svwhilegt_b64_u64(...) __builtin_sve_svwhilegt_b64_u64(__VA_ARGS__) -#define svwhilegt_b16_u64(...) __builtin_sve_svwhilegt_b16_u64(__VA_ARGS__) -#define svwhilerw_u8(...) __builtin_sve_svwhilerw_u8(__VA_ARGS__) -#define svwhilerw_s8(...) __builtin_sve_svwhilerw_s8(__VA_ARGS__) -#define svwhilerw_u64(...) __builtin_sve_svwhilerw_u64(__VA_ARGS__) -#define svwhilerw_f64(...) __builtin_sve_svwhilerw_f64(__VA_ARGS__) -#define svwhilerw_s64(...) __builtin_sve_svwhilerw_s64(__VA_ARGS__) -#define svwhilerw_u16(...) __builtin_sve_svwhilerw_u16(__VA_ARGS__) -#define svwhilerw_f16(...) __builtin_sve_svwhilerw_f16(__VA_ARGS__) -#define svwhilerw_s16(...) __builtin_sve_svwhilerw_s16(__VA_ARGS__) -#define svwhilerw_u32(...) __builtin_sve_svwhilerw_u32(__VA_ARGS__) -#define svwhilerw_f32(...) __builtin_sve_svwhilerw_f32(__VA_ARGS__) -#define svwhilerw_s32(...) __builtin_sve_svwhilerw_s32(__VA_ARGS__) -#define svwhilewr_u8(...) __builtin_sve_svwhilewr_u8(__VA_ARGS__) -#define svwhilewr_s8(...) __builtin_sve_svwhilewr_s8(__VA_ARGS__) -#define svwhilewr_u64(...) __builtin_sve_svwhilewr_u64(__VA_ARGS__) -#define svwhilewr_f64(...) __builtin_sve_svwhilewr_f64(__VA_ARGS__) -#define svwhilewr_s64(...) __builtin_sve_svwhilewr_s64(__VA_ARGS__) -#define svwhilewr_u16(...) __builtin_sve_svwhilewr_u16(__VA_ARGS__) -#define svwhilewr_f16(...) __builtin_sve_svwhilewr_f16(__VA_ARGS__) -#define svwhilewr_s16(...) __builtin_sve_svwhilewr_s16(__VA_ARGS__) -#define svwhilewr_u32(...) __builtin_sve_svwhilewr_u32(__VA_ARGS__) -#define svwhilewr_f32(...) __builtin_sve_svwhilewr_f32(__VA_ARGS__) -#define svwhilewr_s32(...) __builtin_sve_svwhilewr_s32(__VA_ARGS__) -#define svxar_n_u8(...) __builtin_sve_svxar_n_u8(__VA_ARGS__) -#define svxar_n_u32(...) __builtin_sve_svxar_n_u32(__VA_ARGS__) -#define svxar_n_u64(...) __builtin_sve_svxar_n_u64(__VA_ARGS__) -#define svxar_n_u16(...) __builtin_sve_svxar_n_u16(__VA_ARGS__) -#define svxar_n_s8(...) __builtin_sve_svxar_n_s8(__VA_ARGS__) -#define svxar_n_s32(...) __builtin_sve_svxar_n_s32(__VA_ARGS__) -#define svxar_n_s64(...) __builtin_sve_svxar_n_s64(__VA_ARGS__) -#define svxar_n_s16(...) __builtin_sve_svxar_n_s16(__VA_ARGS__) +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaba_n_s8))) +svint8_t svaba_n_s8(svint8_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaba_n_s32))) +svint32_t svaba_n_s32(svint32_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaba_n_s64))) +svint64_t svaba_n_s64(svint64_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaba_n_s16))) +svint16_t svaba_n_s16(svint16_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaba_n_u8))) +svuint8_t svaba_n_u8(svuint8_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaba_n_u32))) +svuint32_t svaba_n_u32(svuint32_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaba_n_u64))) +svuint64_t svaba_n_u64(svuint64_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaba_n_u16))) +svuint16_t svaba_n_u16(svuint16_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaba_s8))) +svint8_t svaba_s8(svint8_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaba_s32))) +svint32_t svaba_s32(svint32_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaba_s64))) +svint64_t svaba_s64(svint64_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaba_s16))) +svint16_t svaba_s16(svint16_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaba_u8))) +svuint8_t svaba_u8(svuint8_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaba_u32))) +svuint32_t svaba_u32(svuint32_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaba_u64))) +svuint64_t svaba_u64(svuint64_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaba_u16))) +svuint16_t svaba_u16(svuint16_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabalb_n_s32))) +svint32_t svabalb_n_s32(svint32_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabalb_n_s64))) +svint64_t svabalb_n_s64(svint64_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabalb_n_s16))) +svint16_t svabalb_n_s16(svint16_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabalb_n_u32))) +svuint32_t svabalb_n_u32(svuint32_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabalb_n_u64))) +svuint64_t svabalb_n_u64(svuint64_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabalb_n_u16))) +svuint16_t svabalb_n_u16(svuint16_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabalb_s32))) +svint32_t svabalb_s32(svint32_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabalb_s64))) +svint64_t svabalb_s64(svint64_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabalb_s16))) +svint16_t svabalb_s16(svint16_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabalb_u32))) +svuint32_t svabalb_u32(svuint32_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabalb_u64))) +svuint64_t svabalb_u64(svuint64_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabalb_u16))) +svuint16_t svabalb_u16(svuint16_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabalt_n_s32))) +svint32_t svabalt_n_s32(svint32_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabalt_n_s64))) +svint64_t svabalt_n_s64(svint64_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabalt_n_s16))) +svint16_t svabalt_n_s16(svint16_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabalt_n_u32))) +svuint32_t svabalt_n_u32(svuint32_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabalt_n_u64))) +svuint64_t svabalt_n_u64(svuint64_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabalt_n_u16))) +svuint16_t svabalt_n_u16(svuint16_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabalt_s32))) +svint32_t svabalt_s32(svint32_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabalt_s64))) +svint64_t svabalt_s64(svint64_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabalt_s16))) +svint16_t svabalt_s16(svint16_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabalt_u32))) +svuint32_t svabalt_u32(svuint32_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabalt_u64))) +svuint64_t svabalt_u64(svuint64_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabalt_u16))) +svuint16_t svabalt_u16(svuint16_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabdlb_n_s32))) +svint32_t svabdlb_n_s32(svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabdlb_n_s64))) +svint64_t svabdlb_n_s64(svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabdlb_n_s16))) +svint16_t svabdlb_n_s16(svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabdlb_n_u32))) +svuint32_t svabdlb_n_u32(svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabdlb_n_u64))) +svuint64_t svabdlb_n_u64(svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabdlb_n_u16))) +svuint16_t svabdlb_n_u16(svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabdlb_s32))) +svint32_t svabdlb_s32(svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabdlb_s64))) +svint64_t svabdlb_s64(svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabdlb_s16))) +svint16_t svabdlb_s16(svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabdlb_u32))) +svuint32_t svabdlb_u32(svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabdlb_u64))) +svuint64_t svabdlb_u64(svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabdlb_u16))) +svuint16_t svabdlb_u16(svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabdlt_n_s32))) +svint32_t svabdlt_n_s32(svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabdlt_n_s64))) +svint64_t svabdlt_n_s64(svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabdlt_n_s16))) +svint16_t svabdlt_n_s16(svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabdlt_n_u32))) +svuint32_t svabdlt_n_u32(svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabdlt_n_u64))) +svuint64_t svabdlt_n_u64(svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabdlt_n_u16))) +svuint16_t svabdlt_n_u16(svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabdlt_s32))) +svint32_t svabdlt_s32(svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabdlt_s64))) +svint64_t svabdlt_s64(svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabdlt_s16))) +svint16_t svabdlt_s16(svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabdlt_u32))) +svuint32_t svabdlt_u32(svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabdlt_u64))) +svuint64_t svabdlt_u64(svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svabdlt_u16))) +svuint16_t svabdlt_u16(svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadalp_s32_m))) +svint32_t svadalp_s32_m(svbool_t, svint32_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadalp_s64_m))) +svint64_t svadalp_s64_m(svbool_t, svint64_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadalp_s16_m))) +svint16_t svadalp_s16_m(svbool_t, svint16_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadalp_s32_x))) +svint32_t svadalp_s32_x(svbool_t, svint32_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadalp_s64_x))) +svint64_t svadalp_s64_x(svbool_t, svint64_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadalp_s16_x))) +svint16_t svadalp_s16_x(svbool_t, svint16_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadalp_s32_z))) +svint32_t svadalp_s32_z(svbool_t, svint32_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadalp_s64_z))) +svint64_t svadalp_s64_z(svbool_t, svint64_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadalp_s16_z))) +svint16_t svadalp_s16_z(svbool_t, svint16_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadalp_u32_m))) +svuint32_t svadalp_u32_m(svbool_t, svuint32_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadalp_u64_m))) +svuint64_t svadalp_u64_m(svbool_t, svuint64_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadalp_u16_m))) +svuint16_t svadalp_u16_m(svbool_t, svuint16_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadalp_u32_x))) +svuint32_t svadalp_u32_x(svbool_t, svuint32_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadalp_u64_x))) +svuint64_t svadalp_u64_x(svbool_t, svuint64_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadalp_u16_x))) +svuint16_t svadalp_u16_x(svbool_t, svuint16_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadalp_u32_z))) +svuint32_t svadalp_u32_z(svbool_t, svuint32_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadalp_u64_z))) +svuint64_t svadalp_u64_z(svbool_t, svuint64_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadalp_u16_z))) +svuint16_t svadalp_u16_z(svbool_t, svuint16_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadclb_n_u32))) +svuint32_t svadclb_n_u32(svuint32_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadclb_n_u64))) +svuint64_t svadclb_n_u64(svuint64_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadclb_u32))) +svuint32_t svadclb_u32(svuint32_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadclb_u64))) +svuint64_t svadclb_u64(svuint64_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadclt_n_u32))) +svuint32_t svadclt_n_u32(svuint32_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadclt_n_u64))) +svuint64_t svadclt_n_u64(svuint64_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadclt_u32))) +svuint32_t svadclt_u32(svuint32_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svadclt_u64))) +svuint64_t svadclt_u64(svuint64_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddhnb_n_u32))) +svuint16_t svaddhnb_n_u32(svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddhnb_n_u64))) +svuint32_t svaddhnb_n_u64(svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddhnb_n_u16))) +svuint8_t svaddhnb_n_u16(svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddhnb_n_s32))) +svint16_t svaddhnb_n_s32(svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddhnb_n_s64))) +svint32_t svaddhnb_n_s64(svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddhnb_n_s16))) +svint8_t svaddhnb_n_s16(svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddhnb_u32))) +svuint16_t svaddhnb_u32(svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddhnb_u64))) +svuint32_t svaddhnb_u64(svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddhnb_u16))) +svuint8_t svaddhnb_u16(svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddhnb_s32))) +svint16_t svaddhnb_s32(svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddhnb_s64))) +svint32_t svaddhnb_s64(svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddhnb_s16))) +svint8_t svaddhnb_s16(svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddhnt_n_u32))) +svuint16_t svaddhnt_n_u32(svuint16_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddhnt_n_u64))) +svuint32_t svaddhnt_n_u64(svuint32_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddhnt_n_u16))) +svuint8_t svaddhnt_n_u16(svuint8_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddhnt_n_s32))) +svint16_t svaddhnt_n_s32(svint16_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddhnt_n_s64))) +svint32_t svaddhnt_n_s64(svint32_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddhnt_n_s16))) +svint8_t svaddhnt_n_s16(svint8_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddhnt_u32))) +svuint16_t svaddhnt_u32(svuint16_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddhnt_u64))) +svuint32_t svaddhnt_u64(svuint32_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddhnt_u16))) +svuint8_t svaddhnt_u16(svuint8_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddhnt_s32))) +svint16_t svaddhnt_s32(svint16_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddhnt_s64))) +svint32_t svaddhnt_s64(svint32_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddhnt_s16))) +svint8_t svaddhnt_s16(svint8_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddlb_n_s32))) +svint32_t svaddlb_n_s32(svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddlb_n_s64))) +svint64_t svaddlb_n_s64(svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddlb_n_s16))) +svint16_t svaddlb_n_s16(svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddlb_n_u32))) +svuint32_t svaddlb_n_u32(svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddlb_n_u64))) +svuint64_t svaddlb_n_u64(svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddlb_n_u16))) +svuint16_t svaddlb_n_u16(svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddlb_s32))) +svint32_t svaddlb_s32(svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddlb_s64))) +svint64_t svaddlb_s64(svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddlb_s16))) +svint16_t svaddlb_s16(svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddlb_u32))) +svuint32_t svaddlb_u32(svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddlb_u64))) +svuint64_t svaddlb_u64(svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddlb_u16))) +svuint16_t svaddlb_u16(svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddlbt_n_s32))) +svint32_t svaddlbt_n_s32(svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddlbt_n_s64))) +svint64_t svaddlbt_n_s64(svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddlbt_n_s16))) +svint16_t svaddlbt_n_s16(svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddlbt_s32))) +svint32_t svaddlbt_s32(svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddlbt_s64))) +svint64_t svaddlbt_s64(svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddlbt_s16))) +svint16_t svaddlbt_s16(svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddlt_n_s32))) +svint32_t svaddlt_n_s32(svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddlt_n_s64))) +svint64_t svaddlt_n_s64(svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddlt_n_s16))) +svint16_t svaddlt_n_s16(svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddlt_n_u32))) +svuint32_t svaddlt_n_u32(svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddlt_n_u64))) +svuint64_t svaddlt_n_u64(svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddlt_n_u16))) +svuint16_t svaddlt_n_u16(svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddlt_s32))) +svint32_t svaddlt_s32(svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddlt_s64))) +svint64_t svaddlt_s64(svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddlt_s16))) +svint16_t svaddlt_s16(svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddlt_u32))) +svuint32_t svaddlt_u32(svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddlt_u64))) +svuint64_t svaddlt_u64(svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddlt_u16))) +svuint16_t svaddlt_u16(svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddp_f64_m))) +svfloat64_t svaddp_f64_m(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddp_f32_m))) +svfloat32_t svaddp_f32_m(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddp_f16_m))) +svfloat16_t svaddp_f16_m(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddp_f64_x))) +svfloat64_t svaddp_f64_x(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddp_f32_x))) +svfloat32_t svaddp_f32_x(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddp_f16_x))) +svfloat16_t svaddp_f16_x(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddp_u8_m))) +svuint8_t svaddp_u8_m(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddp_u32_m))) +svuint32_t svaddp_u32_m(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddp_u64_m))) +svuint64_t svaddp_u64_m(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddp_u16_m))) +svuint16_t svaddp_u16_m(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddp_s8_m))) +svint8_t svaddp_s8_m(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddp_s32_m))) +svint32_t svaddp_s32_m(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddp_s64_m))) +svint64_t svaddp_s64_m(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddp_s16_m))) +svint16_t svaddp_s16_m(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddp_u8_x))) +svuint8_t svaddp_u8_x(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddp_u32_x))) +svuint32_t svaddp_u32_x(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddp_u64_x))) +svuint64_t svaddp_u64_x(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddp_u16_x))) +svuint16_t svaddp_u16_x(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddp_s8_x))) +svint8_t svaddp_s8_x(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddp_s32_x))) +svint32_t svaddp_s32_x(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddp_s64_x))) +svint64_t svaddp_s64_x(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddp_s16_x))) +svint16_t svaddp_s16_x(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddwb_n_s32))) +svint32_t svaddwb_n_s32(svint32_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddwb_n_s64))) +svint64_t svaddwb_n_s64(svint64_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddwb_n_s16))) +svint16_t svaddwb_n_s16(svint16_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddwb_n_u32))) +svuint32_t svaddwb_n_u32(svuint32_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddwb_n_u64))) +svuint64_t svaddwb_n_u64(svuint64_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddwb_n_u16))) +svuint16_t svaddwb_n_u16(svuint16_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddwb_s32))) +svint32_t svaddwb_s32(svint32_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddwb_s64))) +svint64_t svaddwb_s64(svint64_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddwb_s16))) +svint16_t svaddwb_s16(svint16_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddwb_u32))) +svuint32_t svaddwb_u32(svuint32_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddwb_u64))) +svuint64_t svaddwb_u64(svuint64_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddwb_u16))) +svuint16_t svaddwb_u16(svuint16_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddwt_n_s32))) +svint32_t svaddwt_n_s32(svint32_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddwt_n_s64))) +svint64_t svaddwt_n_s64(svint64_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddwt_n_s16))) +svint16_t svaddwt_n_s16(svint16_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddwt_n_u32))) +svuint32_t svaddwt_n_u32(svuint32_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddwt_n_u64))) +svuint64_t svaddwt_n_u64(svuint64_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddwt_n_u16))) +svuint16_t svaddwt_n_u16(svuint16_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddwt_s32))) +svint32_t svaddwt_s32(svint32_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddwt_s64))) +svint64_t svaddwt_s64(svint64_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddwt_s16))) +svint16_t svaddwt_s16(svint16_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddwt_u32))) +svuint32_t svaddwt_u32(svuint32_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddwt_u64))) +svuint64_t svaddwt_u64(svuint64_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaddwt_u16))) +svuint16_t svaddwt_u16(svuint16_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbcax_n_u8))) +svuint8_t svbcax_n_u8(svuint8_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbcax_n_u32))) +svuint32_t svbcax_n_u32(svuint32_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbcax_n_u64))) +svuint64_t svbcax_n_u64(svuint64_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbcax_n_u16))) +svuint16_t svbcax_n_u16(svuint16_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbcax_n_s8))) +svint8_t svbcax_n_s8(svint8_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbcax_n_s32))) +svint32_t svbcax_n_s32(svint32_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbcax_n_s64))) +svint64_t svbcax_n_s64(svint64_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbcax_n_s16))) +svint16_t svbcax_n_s16(svint16_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbcax_u8))) +svuint8_t svbcax_u8(svuint8_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbcax_u32))) +svuint32_t svbcax_u32(svuint32_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbcax_u64))) +svuint64_t svbcax_u64(svuint64_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbcax_u16))) +svuint16_t svbcax_u16(svuint16_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbcax_s8))) +svint8_t svbcax_s8(svint8_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbcax_s32))) +svint32_t svbcax_s32(svint32_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbcax_s64))) +svint64_t svbcax_s64(svint64_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbcax_s16))) +svint16_t svbcax_s16(svint16_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl1n_n_u8))) +svuint8_t svbsl1n_n_u8(svuint8_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl1n_n_u32))) +svuint32_t svbsl1n_n_u32(svuint32_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl1n_n_u64))) +svuint64_t svbsl1n_n_u64(svuint64_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl1n_n_u16))) +svuint16_t svbsl1n_n_u16(svuint16_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl1n_n_s8))) +svint8_t svbsl1n_n_s8(svint8_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl1n_n_s32))) +svint32_t svbsl1n_n_s32(svint32_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl1n_n_s64))) +svint64_t svbsl1n_n_s64(svint64_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl1n_n_s16))) +svint16_t svbsl1n_n_s16(svint16_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl1n_u8))) +svuint8_t svbsl1n_u8(svuint8_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl1n_u32))) +svuint32_t svbsl1n_u32(svuint32_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl1n_u64))) +svuint64_t svbsl1n_u64(svuint64_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl1n_u16))) +svuint16_t svbsl1n_u16(svuint16_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl1n_s8))) +svint8_t svbsl1n_s8(svint8_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl1n_s32))) +svint32_t svbsl1n_s32(svint32_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl1n_s64))) +svint64_t svbsl1n_s64(svint64_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl1n_s16))) +svint16_t svbsl1n_s16(svint16_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl2n_n_u8))) +svuint8_t svbsl2n_n_u8(svuint8_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl2n_n_u32))) +svuint32_t svbsl2n_n_u32(svuint32_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl2n_n_u64))) +svuint64_t svbsl2n_n_u64(svuint64_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl2n_n_u16))) +svuint16_t svbsl2n_n_u16(svuint16_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl2n_n_s8))) +svint8_t svbsl2n_n_s8(svint8_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl2n_n_s32))) +svint32_t svbsl2n_n_s32(svint32_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl2n_n_s64))) +svint64_t svbsl2n_n_s64(svint64_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl2n_n_s16))) +svint16_t svbsl2n_n_s16(svint16_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl2n_u8))) +svuint8_t svbsl2n_u8(svuint8_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl2n_u32))) +svuint32_t svbsl2n_u32(svuint32_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl2n_u64))) +svuint64_t svbsl2n_u64(svuint64_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl2n_u16))) +svuint16_t svbsl2n_u16(svuint16_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl2n_s8))) +svint8_t svbsl2n_s8(svint8_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl2n_s32))) +svint32_t svbsl2n_s32(svint32_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl2n_s64))) +svint64_t svbsl2n_s64(svint64_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl2n_s16))) +svint16_t svbsl2n_s16(svint16_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl_n_u8))) +svuint8_t svbsl_n_u8(svuint8_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl_n_u32))) +svuint32_t svbsl_n_u32(svuint32_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl_n_u64))) +svuint64_t svbsl_n_u64(svuint64_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl_n_u16))) +svuint16_t svbsl_n_u16(svuint16_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl_n_s8))) +svint8_t svbsl_n_s8(svint8_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl_n_s32))) +svint32_t svbsl_n_s32(svint32_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl_n_s64))) +svint64_t svbsl_n_s64(svint64_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl_n_s16))) +svint16_t svbsl_n_s16(svint16_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl_u8))) +svuint8_t svbsl_u8(svuint8_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl_u32))) +svuint32_t svbsl_u32(svuint32_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl_u64))) +svuint64_t svbsl_u64(svuint64_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl_u16))) +svuint16_t svbsl_u16(svuint16_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl_s8))) +svint8_t svbsl_s8(svint8_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl_s32))) +svint32_t svbsl_s32(svint32_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl_s64))) +svint64_t svbsl_s64(svint64_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbsl_s16))) +svint16_t svbsl_s16(svint16_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcadd_u8))) +svuint8_t svcadd_u8(svuint8_t, svuint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcadd_u32))) +svuint32_t svcadd_u32(svuint32_t, svuint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcadd_u64))) +svuint64_t svcadd_u64(svuint64_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcadd_u16))) +svuint16_t svcadd_u16(svuint16_t, svuint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcadd_s8))) +svint8_t svcadd_s8(svint8_t, svint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcadd_s32))) +svint32_t svcadd_s32(svint32_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcadd_s64))) +svint64_t svcadd_s64(svint64_t, svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcadd_s16))) +svint16_t svcadd_s16(svint16_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcdot_s32))) +svint32_t svcdot_s32(svint32_t, svint8_t, svint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcdot_s64))) +svint64_t svcdot_s64(svint64_t, svint16_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcdot_lane_s32))) +svint32_t svcdot_lane_s32(svint32_t, svint8_t, svint8_t, uint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcdot_lane_s64))) +svint64_t svcdot_lane_s64(svint64_t, svint16_t, svint16_t, uint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmla_u8))) +svuint8_t svcmla_u8(svuint8_t, svuint8_t, svuint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmla_u32))) +svuint32_t svcmla_u32(svuint32_t, svuint32_t, svuint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmla_u64))) +svuint64_t svcmla_u64(svuint64_t, svuint64_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmla_u16))) +svuint16_t svcmla_u16(svuint16_t, svuint16_t, svuint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmla_s8))) +svint8_t svcmla_s8(svint8_t, svint8_t, svint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmla_s32))) +svint32_t svcmla_s32(svint32_t, svint32_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmla_s64))) +svint64_t svcmla_s64(svint64_t, svint64_t, svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmla_s16))) +svint16_t svcmla_s16(svint16_t, svint16_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmla_lane_u32))) +svuint32_t svcmla_lane_u32(svuint32_t, svuint32_t, svuint32_t, uint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmla_lane_u16))) +svuint16_t svcmla_lane_u16(svuint16_t, svuint16_t, svuint16_t, uint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmla_lane_s32))) +svint32_t svcmla_lane_s32(svint32_t, svint32_t, svint32_t, uint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcmla_lane_s16))) +svint16_t svcmla_lane_s16(svint16_t, svint16_t, svint16_t, uint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvtlt_f32_f16_m))) +svfloat32_t svcvtlt_f32_f16_m(svfloat32_t, svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvtlt_f32_f16_x))) +svfloat32_t svcvtlt_f32_f16_x(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvtlt_f64_f32_m))) +svfloat64_t svcvtlt_f64_f32_m(svfloat64_t, svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvtlt_f64_f32_x))) +svfloat64_t svcvtlt_f64_f32_x(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvtnt_f16_f32_m))) +svfloat16_t svcvtnt_f16_f32_m(svfloat16_t, svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvtnt_f32_f64_m))) +svfloat32_t svcvtnt_f32_f64_m(svfloat32_t, svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvtx_f32_f64_m))) +svfloat32_t svcvtx_f32_f64_m(svfloat32_t, svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvtx_f32_f64_x))) +svfloat32_t svcvtx_f32_f64_x(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvtx_f32_f64_z))) +svfloat32_t svcvtx_f32_f64_z(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvtxnt_f32_f64_m))) +svfloat32_t svcvtxnt_f32_f64_m(svfloat32_t, svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor3_n_u8))) +svuint8_t sveor3_n_u8(svuint8_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor3_n_u32))) +svuint32_t sveor3_n_u32(svuint32_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor3_n_u64))) +svuint64_t sveor3_n_u64(svuint64_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor3_n_u16))) +svuint16_t sveor3_n_u16(svuint16_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor3_n_s8))) +svint8_t sveor3_n_s8(svint8_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor3_n_s32))) +svint32_t sveor3_n_s32(svint32_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor3_n_s64))) +svint64_t sveor3_n_s64(svint64_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor3_n_s16))) +svint16_t sveor3_n_s16(svint16_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor3_u8))) +svuint8_t sveor3_u8(svuint8_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor3_u32))) +svuint32_t sveor3_u32(svuint32_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor3_u64))) +svuint64_t sveor3_u64(svuint64_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor3_u16))) +svuint16_t sveor3_u16(svuint16_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor3_s8))) +svint8_t sveor3_s8(svint8_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor3_s32))) +svint32_t sveor3_s32(svint32_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor3_s64))) +svint64_t sveor3_s64(svint64_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveor3_s16))) +svint16_t sveor3_s16(svint16_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveorbt_n_u8))) +svuint8_t sveorbt_n_u8(svuint8_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveorbt_n_u32))) +svuint32_t sveorbt_n_u32(svuint32_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveorbt_n_u64))) +svuint64_t sveorbt_n_u64(svuint64_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveorbt_n_u16))) +svuint16_t sveorbt_n_u16(svuint16_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveorbt_n_s8))) +svint8_t sveorbt_n_s8(svint8_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveorbt_n_s32))) +svint32_t sveorbt_n_s32(svint32_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveorbt_n_s64))) +svint64_t sveorbt_n_s64(svint64_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveorbt_n_s16))) +svint16_t sveorbt_n_s16(svint16_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveorbt_u8))) +svuint8_t sveorbt_u8(svuint8_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveorbt_u32))) +svuint32_t sveorbt_u32(svuint32_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveorbt_u64))) +svuint64_t sveorbt_u64(svuint64_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveorbt_u16))) +svuint16_t sveorbt_u16(svuint16_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveorbt_s8))) +svint8_t sveorbt_s8(svint8_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveorbt_s32))) +svint32_t sveorbt_s32(svint32_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveorbt_s64))) +svint64_t sveorbt_s64(svint64_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveorbt_s16))) +svint16_t sveorbt_s16(svint16_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveortb_n_u8))) +svuint8_t sveortb_n_u8(svuint8_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveortb_n_u32))) +svuint32_t sveortb_n_u32(svuint32_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveortb_n_u64))) +svuint64_t sveortb_n_u64(svuint64_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveortb_n_u16))) +svuint16_t sveortb_n_u16(svuint16_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveortb_n_s8))) +svint8_t sveortb_n_s8(svint8_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveortb_n_s32))) +svint32_t sveortb_n_s32(svint32_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveortb_n_s64))) +svint64_t sveortb_n_s64(svint64_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveortb_n_s16))) +svint16_t sveortb_n_s16(svint16_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveortb_u8))) +svuint8_t sveortb_u8(svuint8_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveortb_u32))) +svuint32_t sveortb_u32(svuint32_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveortb_u64))) +svuint64_t sveortb_u64(svuint64_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveortb_u16))) +svuint16_t sveortb_u16(svuint16_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveortb_s8))) +svint8_t sveortb_s8(svint8_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveortb_s32))) +svint32_t sveortb_s32(svint32_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveortb_s64))) +svint64_t sveortb_s64(svint64_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_sveortb_s16))) +svint16_t sveortb_s16(svint16_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_n_s8_m))) +svint8_t svhadd_n_s8_m(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_n_s32_m))) +svint32_t svhadd_n_s32_m(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_n_s64_m))) +svint64_t svhadd_n_s64_m(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_n_s16_m))) +svint16_t svhadd_n_s16_m(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_n_s8_x))) +svint8_t svhadd_n_s8_x(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_n_s32_x))) +svint32_t svhadd_n_s32_x(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_n_s64_x))) +svint64_t svhadd_n_s64_x(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_n_s16_x))) +svint16_t svhadd_n_s16_x(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_n_s8_z))) +svint8_t svhadd_n_s8_z(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_n_s32_z))) +svint32_t svhadd_n_s32_z(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_n_s64_z))) +svint64_t svhadd_n_s64_z(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_n_s16_z))) +svint16_t svhadd_n_s16_z(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_n_u8_m))) +svuint8_t svhadd_n_u8_m(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_n_u32_m))) +svuint32_t svhadd_n_u32_m(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_n_u64_m))) +svuint64_t svhadd_n_u64_m(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_n_u16_m))) +svuint16_t svhadd_n_u16_m(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_n_u8_x))) +svuint8_t svhadd_n_u8_x(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_n_u32_x))) +svuint32_t svhadd_n_u32_x(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_n_u64_x))) +svuint64_t svhadd_n_u64_x(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_n_u16_x))) +svuint16_t svhadd_n_u16_x(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_n_u8_z))) +svuint8_t svhadd_n_u8_z(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_n_u32_z))) +svuint32_t svhadd_n_u32_z(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_n_u64_z))) +svuint64_t svhadd_n_u64_z(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_n_u16_z))) +svuint16_t svhadd_n_u16_z(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_s8_m))) +svint8_t svhadd_s8_m(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_s32_m))) +svint32_t svhadd_s32_m(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_s64_m))) +svint64_t svhadd_s64_m(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_s16_m))) +svint16_t svhadd_s16_m(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_s8_x))) +svint8_t svhadd_s8_x(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_s32_x))) +svint32_t svhadd_s32_x(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_s64_x))) +svint64_t svhadd_s64_x(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_s16_x))) +svint16_t svhadd_s16_x(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_s8_z))) +svint8_t svhadd_s8_z(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_s32_z))) +svint32_t svhadd_s32_z(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_s64_z))) +svint64_t svhadd_s64_z(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_s16_z))) +svint16_t svhadd_s16_z(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_u8_m))) +svuint8_t svhadd_u8_m(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_u32_m))) +svuint32_t svhadd_u32_m(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_u64_m))) +svuint64_t svhadd_u64_m(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_u16_m))) +svuint16_t svhadd_u16_m(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_u8_x))) +svuint8_t svhadd_u8_x(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_u32_x))) +svuint32_t svhadd_u32_x(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_u64_x))) +svuint64_t svhadd_u64_x(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_u16_x))) +svuint16_t svhadd_u16_x(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_u8_z))) +svuint8_t svhadd_u8_z(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_u32_z))) +svuint32_t svhadd_u32_z(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_u64_z))) +svuint64_t svhadd_u64_z(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhadd_u16_z))) +svuint16_t svhadd_u16_z(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhistcnt_u32_z))) +svuint32_t svhistcnt_u32_z(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhistcnt_u64_z))) +svuint64_t svhistcnt_u64_z(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhistcnt_s32_z))) +svuint32_t svhistcnt_s32_z(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhistcnt_s64_z))) +svuint64_t svhistcnt_s64_z(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhistseg_u8))) +svuint8_t svhistseg_u8(svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhistseg_s8))) +svuint8_t svhistseg_s8(svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_n_s8_m))) +svint8_t svhsub_n_s8_m(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_n_s32_m))) +svint32_t svhsub_n_s32_m(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_n_s64_m))) +svint64_t svhsub_n_s64_m(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_n_s16_m))) +svint16_t svhsub_n_s16_m(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_n_s8_x))) +svint8_t svhsub_n_s8_x(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_n_s32_x))) +svint32_t svhsub_n_s32_x(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_n_s64_x))) +svint64_t svhsub_n_s64_x(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_n_s16_x))) +svint16_t svhsub_n_s16_x(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_n_s8_z))) +svint8_t svhsub_n_s8_z(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_n_s32_z))) +svint32_t svhsub_n_s32_z(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_n_s64_z))) +svint64_t svhsub_n_s64_z(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_n_s16_z))) +svint16_t svhsub_n_s16_z(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_n_u8_m))) +svuint8_t svhsub_n_u8_m(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_n_u32_m))) +svuint32_t svhsub_n_u32_m(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_n_u64_m))) +svuint64_t svhsub_n_u64_m(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_n_u16_m))) +svuint16_t svhsub_n_u16_m(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_n_u8_x))) +svuint8_t svhsub_n_u8_x(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_n_u32_x))) +svuint32_t svhsub_n_u32_x(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_n_u64_x))) +svuint64_t svhsub_n_u64_x(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_n_u16_x))) +svuint16_t svhsub_n_u16_x(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_n_u8_z))) +svuint8_t svhsub_n_u8_z(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_n_u32_z))) +svuint32_t svhsub_n_u32_z(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_n_u64_z))) +svuint64_t svhsub_n_u64_z(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_n_u16_z))) +svuint16_t svhsub_n_u16_z(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_s8_m))) +svint8_t svhsub_s8_m(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_s32_m))) +svint32_t svhsub_s32_m(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_s64_m))) +svint64_t svhsub_s64_m(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_s16_m))) +svint16_t svhsub_s16_m(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_s8_x))) +svint8_t svhsub_s8_x(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_s32_x))) +svint32_t svhsub_s32_x(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_s64_x))) +svint64_t svhsub_s64_x(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_s16_x))) +svint16_t svhsub_s16_x(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_s8_z))) +svint8_t svhsub_s8_z(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_s32_z))) +svint32_t svhsub_s32_z(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_s64_z))) +svint64_t svhsub_s64_z(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_s16_z))) +svint16_t svhsub_s16_z(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_u8_m))) +svuint8_t svhsub_u8_m(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_u32_m))) +svuint32_t svhsub_u32_m(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_u64_m))) +svuint64_t svhsub_u64_m(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_u16_m))) +svuint16_t svhsub_u16_m(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_u8_x))) +svuint8_t svhsub_u8_x(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_u32_x))) +svuint32_t svhsub_u32_x(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_u64_x))) +svuint64_t svhsub_u64_x(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_u16_x))) +svuint16_t svhsub_u16_x(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_u8_z))) +svuint8_t svhsub_u8_z(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_u32_z))) +svuint32_t svhsub_u32_z(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_u64_z))) +svuint64_t svhsub_u64_z(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsub_u16_z))) +svuint16_t svhsub_u16_z(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_n_s8_m))) +svint8_t svhsubr_n_s8_m(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_n_s32_m))) +svint32_t svhsubr_n_s32_m(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_n_s64_m))) +svint64_t svhsubr_n_s64_m(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_n_s16_m))) +svint16_t svhsubr_n_s16_m(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_n_s8_x))) +svint8_t svhsubr_n_s8_x(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_n_s32_x))) +svint32_t svhsubr_n_s32_x(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_n_s64_x))) +svint64_t svhsubr_n_s64_x(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_n_s16_x))) +svint16_t svhsubr_n_s16_x(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_n_s8_z))) +svint8_t svhsubr_n_s8_z(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_n_s32_z))) +svint32_t svhsubr_n_s32_z(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_n_s64_z))) +svint64_t svhsubr_n_s64_z(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_n_s16_z))) +svint16_t svhsubr_n_s16_z(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_n_u8_m))) +svuint8_t svhsubr_n_u8_m(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_n_u32_m))) +svuint32_t svhsubr_n_u32_m(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_n_u64_m))) +svuint64_t svhsubr_n_u64_m(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_n_u16_m))) +svuint16_t svhsubr_n_u16_m(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_n_u8_x))) +svuint8_t svhsubr_n_u8_x(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_n_u32_x))) +svuint32_t svhsubr_n_u32_x(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_n_u64_x))) +svuint64_t svhsubr_n_u64_x(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_n_u16_x))) +svuint16_t svhsubr_n_u16_x(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_n_u8_z))) +svuint8_t svhsubr_n_u8_z(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_n_u32_z))) +svuint32_t svhsubr_n_u32_z(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_n_u64_z))) +svuint64_t svhsubr_n_u64_z(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_n_u16_z))) +svuint16_t svhsubr_n_u16_z(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_s8_m))) +svint8_t svhsubr_s8_m(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_s32_m))) +svint32_t svhsubr_s32_m(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_s64_m))) +svint64_t svhsubr_s64_m(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_s16_m))) +svint16_t svhsubr_s16_m(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_s8_x))) +svint8_t svhsubr_s8_x(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_s32_x))) +svint32_t svhsubr_s32_x(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_s64_x))) +svint64_t svhsubr_s64_x(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_s16_x))) +svint16_t svhsubr_s16_x(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_s8_z))) +svint8_t svhsubr_s8_z(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_s32_z))) +svint32_t svhsubr_s32_z(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_s64_z))) +svint64_t svhsubr_s64_z(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_s16_z))) +svint16_t svhsubr_s16_z(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_u8_m))) +svuint8_t svhsubr_u8_m(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_u32_m))) +svuint32_t svhsubr_u32_m(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_u64_m))) +svuint64_t svhsubr_u64_m(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_u16_m))) +svuint16_t svhsubr_u16_m(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_u8_x))) +svuint8_t svhsubr_u8_x(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_u32_x))) +svuint32_t svhsubr_u32_x(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_u64_x))) +svuint64_t svhsubr_u64_x(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_u16_x))) +svuint16_t svhsubr_u16_x(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_u8_z))) +svuint8_t svhsubr_u8_z(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_u32_z))) +svuint32_t svhsubr_u32_z(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_u64_z))) +svuint64_t svhsubr_u64_z(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svhsubr_u16_z))) +svuint16_t svhsubr_u16_z(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_gather_u32base_index_u32))) +svuint32_t svldnt1_gather_u32base_index_u32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_gather_u64base_index_u64))) +svuint64_t svldnt1_gather_u64base_index_u64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_gather_u64base_index_f64))) +svfloat64_t svldnt1_gather_u64base_index_f64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_gather_u32base_index_f32))) +svfloat32_t svldnt1_gather_u32base_index_f32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_gather_u32base_index_s32))) +svint32_t svldnt1_gather_u32base_index_s32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_gather_u64base_index_s64))) +svint64_t svldnt1_gather_u64base_index_s64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_gather_u32base_offset_u32))) +svuint32_t svldnt1_gather_u32base_offset_u32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_gather_u64base_offset_u64))) +svuint64_t svldnt1_gather_u64base_offset_u64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_gather_u64base_offset_f64))) +svfloat64_t svldnt1_gather_u64base_offset_f64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_gather_u32base_offset_f32))) +svfloat32_t svldnt1_gather_u32base_offset_f32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_gather_u32base_offset_s32))) +svint32_t svldnt1_gather_u32base_offset_s32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_gather_u64base_offset_s64))) +svint64_t svldnt1_gather_u64base_offset_s64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_gather_u32base_u32))) +svuint32_t svldnt1_gather_u32base_u32(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_gather_u64base_u64))) +svuint64_t svldnt1_gather_u64base_u64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_gather_u64base_f64))) +svfloat64_t svldnt1_gather_u64base_f64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_gather_u32base_f32))) +svfloat32_t svldnt1_gather_u32base_f32(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_gather_u32base_s32))) +svint32_t svldnt1_gather_u32base_s32(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_gather_u64base_s64))) +svint64_t svldnt1_gather_u64base_s64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_gather_s64index_u64))) +svuint64_t svldnt1_gather_s64index_u64(svbool_t, uint64_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_gather_s64index_f64))) +svfloat64_t svldnt1_gather_s64index_f64(svbool_t, float64_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_gather_s64index_s64))) +svint64_t svldnt1_gather_s64index_s64(svbool_t, int64_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_gather_u64index_u64))) +svuint64_t svldnt1_gather_u64index_u64(svbool_t, uint64_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_gather_u64index_f64))) +svfloat64_t svldnt1_gather_u64index_f64(svbool_t, float64_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_gather_u64index_s64))) +svint64_t svldnt1_gather_u64index_s64(svbool_t, int64_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_gather_u32offset_u32))) +svuint32_t svldnt1_gather_u32offset_u32(svbool_t, uint32_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_gather_u32offset_f32))) +svfloat32_t svldnt1_gather_u32offset_f32(svbool_t, float32_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_gather_u32offset_s32))) +svint32_t svldnt1_gather_u32offset_s32(svbool_t, int32_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_gather_s64offset_u64))) +svuint64_t svldnt1_gather_s64offset_u64(svbool_t, uint64_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_gather_s64offset_f64))) +svfloat64_t svldnt1_gather_s64offset_f64(svbool_t, float64_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_gather_s64offset_s64))) +svint64_t svldnt1_gather_s64offset_s64(svbool_t, int64_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_gather_u64offset_u64))) +svuint64_t svldnt1_gather_u64offset_u64(svbool_t, uint64_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_gather_u64offset_f64))) +svfloat64_t svldnt1_gather_u64offset_f64(svbool_t, float64_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_gather_u64offset_s64))) +svint64_t svldnt1_gather_u64offset_s64(svbool_t, int64_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sb_gather_u32base_offset_u32))) +svuint32_t svldnt1sb_gather_u32base_offset_u32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sb_gather_u64base_offset_u64))) +svuint64_t svldnt1sb_gather_u64base_offset_u64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sb_gather_u32base_offset_s32))) +svint32_t svldnt1sb_gather_u32base_offset_s32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sb_gather_u64base_offset_s64))) +svint64_t svldnt1sb_gather_u64base_offset_s64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sb_gather_u32base_u32))) +svuint32_t svldnt1sb_gather_u32base_u32(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sb_gather_u64base_u64))) +svuint64_t svldnt1sb_gather_u64base_u64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sb_gather_u32base_s32))) +svint32_t svldnt1sb_gather_u32base_s32(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sb_gather_u64base_s64))) +svint64_t svldnt1sb_gather_u64base_s64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sb_gather_u32offset_u32))) +svuint32_t svldnt1sb_gather_u32offset_u32(svbool_t, int8_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sb_gather_u32offset_s32))) +svint32_t svldnt1sb_gather_u32offset_s32(svbool_t, int8_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sb_gather_s64offset_u64))) +svuint64_t svldnt1sb_gather_s64offset_u64(svbool_t, int8_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sb_gather_s64offset_s64))) +svint64_t svldnt1sb_gather_s64offset_s64(svbool_t, int8_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sb_gather_u64offset_u64))) +svuint64_t svldnt1sb_gather_u64offset_u64(svbool_t, int8_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sb_gather_u64offset_s64))) +svint64_t svldnt1sb_gather_u64offset_s64(svbool_t, int8_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sh_gather_u32base_index_u32))) +svuint32_t svldnt1sh_gather_u32base_index_u32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sh_gather_u64base_index_u64))) +svuint64_t svldnt1sh_gather_u64base_index_u64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sh_gather_u32base_index_s32))) +svint32_t svldnt1sh_gather_u32base_index_s32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sh_gather_u64base_index_s64))) +svint64_t svldnt1sh_gather_u64base_index_s64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sh_gather_u32base_offset_u32))) +svuint32_t svldnt1sh_gather_u32base_offset_u32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sh_gather_u64base_offset_u64))) +svuint64_t svldnt1sh_gather_u64base_offset_u64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sh_gather_u32base_offset_s32))) +svint32_t svldnt1sh_gather_u32base_offset_s32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sh_gather_u64base_offset_s64))) +svint64_t svldnt1sh_gather_u64base_offset_s64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sh_gather_u32base_u32))) +svuint32_t svldnt1sh_gather_u32base_u32(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sh_gather_u64base_u64))) +svuint64_t svldnt1sh_gather_u64base_u64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sh_gather_u32base_s32))) +svint32_t svldnt1sh_gather_u32base_s32(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sh_gather_u64base_s64))) +svint64_t svldnt1sh_gather_u64base_s64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sh_gather_s64index_u64))) +svuint64_t svldnt1sh_gather_s64index_u64(svbool_t, int16_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sh_gather_s64index_s64))) +svint64_t svldnt1sh_gather_s64index_s64(svbool_t, int16_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sh_gather_u64index_u64))) +svuint64_t svldnt1sh_gather_u64index_u64(svbool_t, int16_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sh_gather_u64index_s64))) +svint64_t svldnt1sh_gather_u64index_s64(svbool_t, int16_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sh_gather_u32offset_u32))) +svuint32_t svldnt1sh_gather_u32offset_u32(svbool_t, int16_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sh_gather_u32offset_s32))) +svint32_t svldnt1sh_gather_u32offset_s32(svbool_t, int16_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sh_gather_s64offset_u64))) +svuint64_t svldnt1sh_gather_s64offset_u64(svbool_t, int16_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sh_gather_s64offset_s64))) +svint64_t svldnt1sh_gather_s64offset_s64(svbool_t, int16_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sh_gather_u64offset_u64))) +svuint64_t svldnt1sh_gather_u64offset_u64(svbool_t, int16_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sh_gather_u64offset_s64))) +svint64_t svldnt1sh_gather_u64offset_s64(svbool_t, int16_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sw_gather_u64base_index_u64))) +svuint64_t svldnt1sw_gather_u64base_index_u64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sw_gather_u64base_index_s64))) +svint64_t svldnt1sw_gather_u64base_index_s64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sw_gather_u64base_offset_u64))) +svuint64_t svldnt1sw_gather_u64base_offset_u64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sw_gather_u64base_offset_s64))) +svint64_t svldnt1sw_gather_u64base_offset_s64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sw_gather_u64base_u64))) +svuint64_t svldnt1sw_gather_u64base_u64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sw_gather_u64base_s64))) +svint64_t svldnt1sw_gather_u64base_s64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sw_gather_s64index_u64))) +svuint64_t svldnt1sw_gather_s64index_u64(svbool_t, int32_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sw_gather_s64index_s64))) +svint64_t svldnt1sw_gather_s64index_s64(svbool_t, int32_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sw_gather_u64index_u64))) +svuint64_t svldnt1sw_gather_u64index_u64(svbool_t, int32_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sw_gather_u64index_s64))) +svint64_t svldnt1sw_gather_u64index_s64(svbool_t, int32_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sw_gather_s64offset_u64))) +svuint64_t svldnt1sw_gather_s64offset_u64(svbool_t, int32_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sw_gather_s64offset_s64))) +svint64_t svldnt1sw_gather_s64offset_s64(svbool_t, int32_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sw_gather_u64offset_u64))) +svuint64_t svldnt1sw_gather_u64offset_u64(svbool_t, int32_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1sw_gather_u64offset_s64))) +svint64_t svldnt1sw_gather_u64offset_s64(svbool_t, int32_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1ub_gather_u32base_offset_u32))) +svuint32_t svldnt1ub_gather_u32base_offset_u32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1ub_gather_u64base_offset_u64))) +svuint64_t svldnt1ub_gather_u64base_offset_u64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1ub_gather_u32base_offset_s32))) +svint32_t svldnt1ub_gather_u32base_offset_s32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1ub_gather_u64base_offset_s64))) +svint64_t svldnt1ub_gather_u64base_offset_s64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1ub_gather_u32base_u32))) +svuint32_t svldnt1ub_gather_u32base_u32(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1ub_gather_u64base_u64))) +svuint64_t svldnt1ub_gather_u64base_u64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1ub_gather_u32base_s32))) +svint32_t svldnt1ub_gather_u32base_s32(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1ub_gather_u64base_s64))) +svint64_t svldnt1ub_gather_u64base_s64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1ub_gather_u32offset_u32))) +svuint32_t svldnt1ub_gather_u32offset_u32(svbool_t, uint8_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1ub_gather_u32offset_s32))) +svint32_t svldnt1ub_gather_u32offset_s32(svbool_t, uint8_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1ub_gather_s64offset_u64))) +svuint64_t svldnt1ub_gather_s64offset_u64(svbool_t, uint8_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1ub_gather_s64offset_s64))) +svint64_t svldnt1ub_gather_s64offset_s64(svbool_t, uint8_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1ub_gather_u64offset_u64))) +svuint64_t svldnt1ub_gather_u64offset_u64(svbool_t, uint8_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1ub_gather_u64offset_s64))) +svint64_t svldnt1ub_gather_u64offset_s64(svbool_t, uint8_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1uh_gather_u32base_index_u32))) +svuint32_t svldnt1uh_gather_u32base_index_u32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1uh_gather_u64base_index_u64))) +svuint64_t svldnt1uh_gather_u64base_index_u64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1uh_gather_u32base_index_s32))) +svint32_t svldnt1uh_gather_u32base_index_s32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1uh_gather_u64base_index_s64))) +svint64_t svldnt1uh_gather_u64base_index_s64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1uh_gather_u32base_offset_u32))) +svuint32_t svldnt1uh_gather_u32base_offset_u32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1uh_gather_u64base_offset_u64))) +svuint64_t svldnt1uh_gather_u64base_offset_u64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1uh_gather_u32base_offset_s32))) +svint32_t svldnt1uh_gather_u32base_offset_s32(svbool_t, svuint32_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1uh_gather_u64base_offset_s64))) +svint64_t svldnt1uh_gather_u64base_offset_s64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1uh_gather_u32base_u32))) +svuint32_t svldnt1uh_gather_u32base_u32(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1uh_gather_u64base_u64))) +svuint64_t svldnt1uh_gather_u64base_u64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1uh_gather_u32base_s32))) +svint32_t svldnt1uh_gather_u32base_s32(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1uh_gather_u64base_s64))) +svint64_t svldnt1uh_gather_u64base_s64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1uh_gather_s64index_u64))) +svuint64_t svldnt1uh_gather_s64index_u64(svbool_t, uint16_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1uh_gather_s64index_s64))) +svint64_t svldnt1uh_gather_s64index_s64(svbool_t, uint16_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1uh_gather_u64index_u64))) +svuint64_t svldnt1uh_gather_u64index_u64(svbool_t, uint16_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1uh_gather_u64index_s64))) +svint64_t svldnt1uh_gather_u64index_s64(svbool_t, uint16_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1uh_gather_u32offset_u32))) +svuint32_t svldnt1uh_gather_u32offset_u32(svbool_t, uint16_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1uh_gather_u32offset_s32))) +svint32_t svldnt1uh_gather_u32offset_s32(svbool_t, uint16_t const *, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1uh_gather_s64offset_u64))) +svuint64_t svldnt1uh_gather_s64offset_u64(svbool_t, uint16_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1uh_gather_s64offset_s64))) +svint64_t svldnt1uh_gather_s64offset_s64(svbool_t, uint16_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1uh_gather_u64offset_u64))) +svuint64_t svldnt1uh_gather_u64offset_u64(svbool_t, uint16_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1uh_gather_u64offset_s64))) +svint64_t svldnt1uh_gather_u64offset_s64(svbool_t, uint16_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1uw_gather_u64base_index_u64))) +svuint64_t svldnt1uw_gather_u64base_index_u64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1uw_gather_u64base_index_s64))) +svint64_t svldnt1uw_gather_u64base_index_s64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1uw_gather_u64base_offset_u64))) +svuint64_t svldnt1uw_gather_u64base_offset_u64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1uw_gather_u64base_offset_s64))) +svint64_t svldnt1uw_gather_u64base_offset_s64(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1uw_gather_u64base_u64))) +svuint64_t svldnt1uw_gather_u64base_u64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1uw_gather_u64base_s64))) +svint64_t svldnt1uw_gather_u64base_s64(svbool_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1uw_gather_s64index_u64))) +svuint64_t svldnt1uw_gather_s64index_u64(svbool_t, uint32_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1uw_gather_s64index_s64))) +svint64_t svldnt1uw_gather_s64index_s64(svbool_t, uint32_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1uw_gather_u64index_u64))) +svuint64_t svldnt1uw_gather_u64index_u64(svbool_t, uint32_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1uw_gather_u64index_s64))) +svint64_t svldnt1uw_gather_u64index_s64(svbool_t, uint32_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1uw_gather_s64offset_u64))) +svuint64_t svldnt1uw_gather_s64offset_u64(svbool_t, uint32_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1uw_gather_s64offset_s64))) +svint64_t svldnt1uw_gather_s64offset_s64(svbool_t, uint32_t const *, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1uw_gather_u64offset_u64))) +svuint64_t svldnt1uw_gather_u64offset_u64(svbool_t, uint32_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1uw_gather_u64offset_s64))) +svint64_t svldnt1uw_gather_u64offset_s64(svbool_t, uint32_t const *, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlogb_f64_m))) +svint64_t svlogb_f64_m(svint64_t, svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlogb_f32_m))) +svint32_t svlogb_f32_m(svint32_t, svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlogb_f16_m))) +svint16_t svlogb_f16_m(svint16_t, svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlogb_f64_x))) +svint64_t svlogb_f64_x(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlogb_f32_x))) +svint32_t svlogb_f32_x(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlogb_f16_x))) +svint16_t svlogb_f16_x(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlogb_f64_z))) +svint64_t svlogb_f64_z(svbool_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlogb_f32_z))) +svint32_t svlogb_f32_z(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlogb_f16_z))) +svint16_t svlogb_f16_z(svbool_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmatch_u8))) +svbool_t svmatch_u8(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmatch_u16))) +svbool_t svmatch_u16(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmatch_s8))) +svbool_t svmatch_s8(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmatch_s16))) +svbool_t svmatch_s16(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxnmp_f64_m))) +svfloat64_t svmaxnmp_f64_m(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxnmp_f32_m))) +svfloat32_t svmaxnmp_f32_m(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxnmp_f16_m))) +svfloat16_t svmaxnmp_f16_m(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxnmp_f64_x))) +svfloat64_t svmaxnmp_f64_x(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxnmp_f32_x))) +svfloat32_t svmaxnmp_f32_x(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxnmp_f16_x))) +svfloat16_t svmaxnmp_f16_x(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxp_f64_m))) +svfloat64_t svmaxp_f64_m(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxp_f32_m))) +svfloat32_t svmaxp_f32_m(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxp_f16_m))) +svfloat16_t svmaxp_f16_m(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxp_f64_x))) +svfloat64_t svmaxp_f64_x(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxp_f32_x))) +svfloat32_t svmaxp_f32_x(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxp_f16_x))) +svfloat16_t svmaxp_f16_x(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxp_s8_m))) +svint8_t svmaxp_s8_m(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxp_s32_m))) +svint32_t svmaxp_s32_m(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxp_s64_m))) +svint64_t svmaxp_s64_m(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxp_s16_m))) +svint16_t svmaxp_s16_m(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxp_s8_x))) +svint8_t svmaxp_s8_x(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxp_s32_x))) +svint32_t svmaxp_s32_x(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxp_s64_x))) +svint64_t svmaxp_s64_x(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxp_s16_x))) +svint16_t svmaxp_s16_x(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxp_u8_m))) +svuint8_t svmaxp_u8_m(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxp_u32_m))) +svuint32_t svmaxp_u32_m(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxp_u64_m))) +svuint64_t svmaxp_u64_m(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxp_u16_m))) +svuint16_t svmaxp_u16_m(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxp_u8_x))) +svuint8_t svmaxp_u8_x(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxp_u32_x))) +svuint32_t svmaxp_u32_x(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxp_u64_x))) +svuint64_t svmaxp_u64_x(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmaxp_u16_x))) +svuint16_t svmaxp_u16_x(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminnmp_f64_m))) +svfloat64_t svminnmp_f64_m(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminnmp_f32_m))) +svfloat32_t svminnmp_f32_m(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminnmp_f16_m))) +svfloat16_t svminnmp_f16_m(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminnmp_f64_x))) +svfloat64_t svminnmp_f64_x(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminnmp_f32_x))) +svfloat32_t svminnmp_f32_x(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminnmp_f16_x))) +svfloat16_t svminnmp_f16_x(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminp_f64_m))) +svfloat64_t svminp_f64_m(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminp_f32_m))) +svfloat32_t svminp_f32_m(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminp_f16_m))) +svfloat16_t svminp_f16_m(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminp_f64_x))) +svfloat64_t svminp_f64_x(svbool_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminp_f32_x))) +svfloat32_t svminp_f32_x(svbool_t, svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminp_f16_x))) +svfloat16_t svminp_f16_x(svbool_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminp_s8_m))) +svint8_t svminp_s8_m(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminp_s32_m))) +svint32_t svminp_s32_m(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminp_s64_m))) +svint64_t svminp_s64_m(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminp_s16_m))) +svint16_t svminp_s16_m(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminp_s8_x))) +svint8_t svminp_s8_x(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminp_s32_x))) +svint32_t svminp_s32_x(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminp_s64_x))) +svint64_t svminp_s64_x(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminp_s16_x))) +svint16_t svminp_s16_x(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminp_u8_m))) +svuint8_t svminp_u8_m(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminp_u32_m))) +svuint32_t svminp_u32_m(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminp_u64_m))) +svuint64_t svminp_u64_m(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminp_u16_m))) +svuint16_t svminp_u16_m(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminp_u8_x))) +svuint8_t svminp_u8_x(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminp_u32_x))) +svuint32_t svminp_u32_x(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminp_u64_x))) +svuint64_t svminp_u64_x(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svminp_u16_x))) +svuint16_t svminp_u16_x(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_lane_u32))) +svuint32_t svmla_lane_u32(svuint32_t, svuint32_t, svuint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_lane_u64))) +svuint64_t svmla_lane_u64(svuint64_t, svuint64_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_lane_u16))) +svuint16_t svmla_lane_u16(svuint16_t, svuint16_t, svuint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_lane_s32))) +svint32_t svmla_lane_s32(svint32_t, svint32_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_lane_s64))) +svint64_t svmla_lane_s64(svint64_t, svint64_t, svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmla_lane_s16))) +svint16_t svmla_lane_s16(svint16_t, svint16_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlalb_n_f32))) +svfloat32_t svmlalb_n_f32(svfloat32_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlalb_n_s32))) +svint32_t svmlalb_n_s32(svint32_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlalb_n_s64))) +svint64_t svmlalb_n_s64(svint64_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlalb_n_s16))) +svint16_t svmlalb_n_s16(svint16_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlalb_n_u32))) +svuint32_t svmlalb_n_u32(svuint32_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlalb_n_u64))) +svuint64_t svmlalb_n_u64(svuint64_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlalb_n_u16))) +svuint16_t svmlalb_n_u16(svuint16_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlalb_f32))) +svfloat32_t svmlalb_f32(svfloat32_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlalb_s32))) +svint32_t svmlalb_s32(svint32_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlalb_s64))) +svint64_t svmlalb_s64(svint64_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlalb_s16))) +svint16_t svmlalb_s16(svint16_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlalb_u32))) +svuint32_t svmlalb_u32(svuint32_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlalb_u64))) +svuint64_t svmlalb_u64(svuint64_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlalb_u16))) +svuint16_t svmlalb_u16(svuint16_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlalb_lane_f32))) +svfloat32_t svmlalb_lane_f32(svfloat32_t, svfloat16_t, svfloat16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlalb_lane_s32))) +svint32_t svmlalb_lane_s32(svint32_t, svint16_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlalb_lane_s64))) +svint64_t svmlalb_lane_s64(svint64_t, svint32_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlalb_lane_u32))) +svuint32_t svmlalb_lane_u32(svuint32_t, svuint16_t, svuint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlalb_lane_u64))) +svuint64_t svmlalb_lane_u64(svuint64_t, svuint32_t, svuint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlalt_n_f32))) +svfloat32_t svmlalt_n_f32(svfloat32_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlalt_n_s32))) +svint32_t svmlalt_n_s32(svint32_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlalt_n_s64))) +svint64_t svmlalt_n_s64(svint64_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlalt_n_s16))) +svint16_t svmlalt_n_s16(svint16_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlalt_n_u32))) +svuint32_t svmlalt_n_u32(svuint32_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlalt_n_u64))) +svuint64_t svmlalt_n_u64(svuint64_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlalt_n_u16))) +svuint16_t svmlalt_n_u16(svuint16_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlalt_f32))) +svfloat32_t svmlalt_f32(svfloat32_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlalt_s32))) +svint32_t svmlalt_s32(svint32_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlalt_s64))) +svint64_t svmlalt_s64(svint64_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlalt_s16))) +svint16_t svmlalt_s16(svint16_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlalt_u32))) +svuint32_t svmlalt_u32(svuint32_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlalt_u64))) +svuint64_t svmlalt_u64(svuint64_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlalt_u16))) +svuint16_t svmlalt_u16(svuint16_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlalt_lane_f32))) +svfloat32_t svmlalt_lane_f32(svfloat32_t, svfloat16_t, svfloat16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlalt_lane_s32))) +svint32_t svmlalt_lane_s32(svint32_t, svint16_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlalt_lane_s64))) +svint64_t svmlalt_lane_s64(svint64_t, svint32_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlalt_lane_u32))) +svuint32_t svmlalt_lane_u32(svuint32_t, svuint16_t, svuint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlalt_lane_u64))) +svuint64_t svmlalt_lane_u64(svuint64_t, svuint32_t, svuint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_lane_u32))) +svuint32_t svmls_lane_u32(svuint32_t, svuint32_t, svuint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_lane_u64))) +svuint64_t svmls_lane_u64(svuint64_t, svuint64_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_lane_u16))) +svuint16_t svmls_lane_u16(svuint16_t, svuint16_t, svuint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_lane_s32))) +svint32_t svmls_lane_s32(svint32_t, svint32_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_lane_s64))) +svint64_t svmls_lane_s64(svint64_t, svint64_t, svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmls_lane_s16))) +svint16_t svmls_lane_s16(svint16_t, svint16_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlslb_n_f32))) +svfloat32_t svmlslb_n_f32(svfloat32_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlslb_n_s32))) +svint32_t svmlslb_n_s32(svint32_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlslb_n_s64))) +svint64_t svmlslb_n_s64(svint64_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlslb_n_s16))) +svint16_t svmlslb_n_s16(svint16_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlslb_n_u32))) +svuint32_t svmlslb_n_u32(svuint32_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlslb_n_u64))) +svuint64_t svmlslb_n_u64(svuint64_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlslb_n_u16))) +svuint16_t svmlslb_n_u16(svuint16_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlslb_f32))) +svfloat32_t svmlslb_f32(svfloat32_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlslb_s32))) +svint32_t svmlslb_s32(svint32_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlslb_s64))) +svint64_t svmlslb_s64(svint64_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlslb_s16))) +svint16_t svmlslb_s16(svint16_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlslb_u32))) +svuint32_t svmlslb_u32(svuint32_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlslb_u64))) +svuint64_t svmlslb_u64(svuint64_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlslb_u16))) +svuint16_t svmlslb_u16(svuint16_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlslb_lane_f32))) +svfloat32_t svmlslb_lane_f32(svfloat32_t, svfloat16_t, svfloat16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlslb_lane_s32))) +svint32_t svmlslb_lane_s32(svint32_t, svint16_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlslb_lane_s64))) +svint64_t svmlslb_lane_s64(svint64_t, svint32_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlslb_lane_u32))) +svuint32_t svmlslb_lane_u32(svuint32_t, svuint16_t, svuint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlslb_lane_u64))) +svuint64_t svmlslb_lane_u64(svuint64_t, svuint32_t, svuint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlslt_n_f32))) +svfloat32_t svmlslt_n_f32(svfloat32_t, svfloat16_t, float16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlslt_n_s32))) +svint32_t svmlslt_n_s32(svint32_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlslt_n_s64))) +svint64_t svmlslt_n_s64(svint64_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlslt_n_s16))) +svint16_t svmlslt_n_s16(svint16_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlslt_n_u32))) +svuint32_t svmlslt_n_u32(svuint32_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlslt_n_u64))) +svuint64_t svmlslt_n_u64(svuint64_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlslt_n_u16))) +svuint16_t svmlslt_n_u16(svuint16_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlslt_f32))) +svfloat32_t svmlslt_f32(svfloat32_t, svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlslt_s32))) +svint32_t svmlslt_s32(svint32_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlslt_s64))) +svint64_t svmlslt_s64(svint64_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlslt_s16))) +svint16_t svmlslt_s16(svint16_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlslt_u32))) +svuint32_t svmlslt_u32(svuint32_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlslt_u64))) +svuint64_t svmlslt_u64(svuint64_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlslt_u16))) +svuint16_t svmlslt_u16(svuint16_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlslt_lane_f32))) +svfloat32_t svmlslt_lane_f32(svfloat32_t, svfloat16_t, svfloat16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlslt_lane_s32))) +svint32_t svmlslt_lane_s32(svint32_t, svint16_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlslt_lane_s64))) +svint64_t svmlslt_lane_s64(svint64_t, svint32_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlslt_lane_u32))) +svuint32_t svmlslt_lane_u32(svuint32_t, svuint16_t, svuint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmlslt_lane_u64))) +svuint64_t svmlslt_lane_u64(svuint64_t, svuint32_t, svuint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmovlb_s32))) +svint32_t svmovlb_s32(svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmovlb_s64))) +svint64_t svmovlb_s64(svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmovlb_s16))) +svint16_t svmovlb_s16(svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmovlb_u32))) +svuint32_t svmovlb_u32(svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmovlb_u64))) +svuint64_t svmovlb_u64(svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmovlb_u16))) +svuint16_t svmovlb_u16(svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmovlt_s32))) +svint32_t svmovlt_s32(svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmovlt_s64))) +svint64_t svmovlt_s64(svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmovlt_s16))) +svint16_t svmovlt_s16(svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmovlt_u32))) +svuint32_t svmovlt_u32(svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmovlt_u64))) +svuint64_t svmovlt_u64(svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmovlt_u16))) +svuint16_t svmovlt_u16(svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_lane_u32))) +svuint32_t svmul_lane_u32(svuint32_t, svuint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_lane_u64))) +svuint64_t svmul_lane_u64(svuint64_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_lane_u16))) +svuint16_t svmul_lane_u16(svuint16_t, svuint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_lane_s32))) +svint32_t svmul_lane_s32(svint32_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_lane_s64))) +svint64_t svmul_lane_s64(svint64_t, svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmul_lane_s16))) +svint16_t svmul_lane_s16(svint16_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmullb_n_s32))) +svint32_t svmullb_n_s32(svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmullb_n_s64))) +svint64_t svmullb_n_s64(svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmullb_n_s16))) +svint16_t svmullb_n_s16(svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmullb_n_u32))) +svuint32_t svmullb_n_u32(svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmullb_n_u64))) +svuint64_t svmullb_n_u64(svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmullb_n_u16))) +svuint16_t svmullb_n_u16(svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmullb_s32))) +svint32_t svmullb_s32(svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmullb_s64))) +svint64_t svmullb_s64(svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmullb_s16))) +svint16_t svmullb_s16(svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmullb_u32))) +svuint32_t svmullb_u32(svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmullb_u64))) +svuint64_t svmullb_u64(svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmullb_u16))) +svuint16_t svmullb_u16(svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmullb_lane_s32))) +svint32_t svmullb_lane_s32(svint16_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmullb_lane_s64))) +svint64_t svmullb_lane_s64(svint32_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmullb_lane_u32))) +svuint32_t svmullb_lane_u32(svuint16_t, svuint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmullb_lane_u64))) +svuint64_t svmullb_lane_u64(svuint32_t, svuint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmullt_n_s32))) +svint32_t svmullt_n_s32(svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmullt_n_s64))) +svint64_t svmullt_n_s64(svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmullt_n_s16))) +svint16_t svmullt_n_s16(svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmullt_n_u32))) +svuint32_t svmullt_n_u32(svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmullt_n_u64))) +svuint64_t svmullt_n_u64(svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmullt_n_u16))) +svuint16_t svmullt_n_u16(svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmullt_s32))) +svint32_t svmullt_s32(svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmullt_s64))) +svint64_t svmullt_s64(svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmullt_s16))) +svint16_t svmullt_s16(svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmullt_u32))) +svuint32_t svmullt_u32(svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmullt_u64))) +svuint64_t svmullt_u64(svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmullt_u16))) +svuint16_t svmullt_u16(svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmullt_lane_s32))) +svint32_t svmullt_lane_s32(svint16_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmullt_lane_s64))) +svint64_t svmullt_lane_s64(svint32_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmullt_lane_u32))) +svuint32_t svmullt_lane_u32(svuint16_t, svuint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmullt_lane_u64))) +svuint64_t svmullt_lane_u64(svuint32_t, svuint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnbsl_n_u8))) +svuint8_t svnbsl_n_u8(svuint8_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnbsl_n_u32))) +svuint32_t svnbsl_n_u32(svuint32_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnbsl_n_u64))) +svuint64_t svnbsl_n_u64(svuint64_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnbsl_n_u16))) +svuint16_t svnbsl_n_u16(svuint16_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnbsl_n_s8))) +svint8_t svnbsl_n_s8(svint8_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnbsl_n_s32))) +svint32_t svnbsl_n_s32(svint32_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnbsl_n_s64))) +svint64_t svnbsl_n_s64(svint64_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnbsl_n_s16))) +svint16_t svnbsl_n_s16(svint16_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnbsl_u8))) +svuint8_t svnbsl_u8(svuint8_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnbsl_u32))) +svuint32_t svnbsl_u32(svuint32_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnbsl_u64))) +svuint64_t svnbsl_u64(svuint64_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnbsl_u16))) +svuint16_t svnbsl_u16(svuint16_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnbsl_s8))) +svint8_t svnbsl_s8(svint8_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnbsl_s32))) +svint32_t svnbsl_s32(svint32_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnbsl_s64))) +svint64_t svnbsl_s64(svint64_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnbsl_s16))) +svint16_t svnbsl_s16(svint16_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmatch_u8))) +svbool_t svnmatch_u8(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmatch_u16))) +svbool_t svnmatch_u16(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmatch_s8))) +svbool_t svnmatch_s8(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svnmatch_s16))) +svbool_t svnmatch_s16(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svpmul_n_u8))) +svuint8_t svpmul_n_u8(svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svpmul_u8))) +svuint8_t svpmul_u8(svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svpmullb_n_u64))) +svuint64_t svpmullb_n_u64(svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svpmullb_n_u16))) +svuint16_t svpmullb_n_u16(svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svpmullb_u64))) +svuint64_t svpmullb_u64(svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svpmullb_u16))) +svuint16_t svpmullb_u16(svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svpmullb_pair_n_u8))) +svuint8_t svpmullb_pair_n_u8(svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svpmullb_pair_n_u32))) +svuint32_t svpmullb_pair_n_u32(svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svpmullb_pair_u8))) +svuint8_t svpmullb_pair_u8(svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svpmullb_pair_u32))) +svuint32_t svpmullb_pair_u32(svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svpmullt_n_u64))) +svuint64_t svpmullt_n_u64(svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svpmullt_n_u16))) +svuint16_t svpmullt_n_u16(svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svpmullt_u64))) +svuint64_t svpmullt_u64(svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svpmullt_u16))) +svuint16_t svpmullt_u16(svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svpmullt_pair_n_u8))) +svuint8_t svpmullt_pair_n_u8(svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svpmullt_pair_n_u32))) +svuint32_t svpmullt_pair_n_u32(svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svpmullt_pair_u8))) +svuint8_t svpmullt_pair_u8(svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svpmullt_pair_u32))) +svuint32_t svpmullt_pair_u32(svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqabs_s8_m))) +svint8_t svqabs_s8_m(svint8_t, svbool_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqabs_s32_m))) +svint32_t svqabs_s32_m(svint32_t, svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqabs_s64_m))) +svint64_t svqabs_s64_m(svint64_t, svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqabs_s16_m))) +svint16_t svqabs_s16_m(svint16_t, svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqabs_s8_x))) +svint8_t svqabs_s8_x(svbool_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqabs_s32_x))) +svint32_t svqabs_s32_x(svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqabs_s64_x))) +svint64_t svqabs_s64_x(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqabs_s16_x))) +svint16_t svqabs_s16_x(svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqabs_s8_z))) +svint8_t svqabs_s8_z(svbool_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqabs_s32_z))) +svint32_t svqabs_s32_z(svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqabs_s64_z))) +svint64_t svqabs_s64_z(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqabs_s16_z))) +svint16_t svqabs_s16_z(svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_n_s8_m))) +svint8_t svqadd_n_s8_m(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_n_s32_m))) +svint32_t svqadd_n_s32_m(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_n_s64_m))) +svint64_t svqadd_n_s64_m(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_n_s16_m))) +svint16_t svqadd_n_s16_m(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_n_s8_x))) +svint8_t svqadd_n_s8_x(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_n_s32_x))) +svint32_t svqadd_n_s32_x(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_n_s64_x))) +svint64_t svqadd_n_s64_x(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_n_s16_x))) +svint16_t svqadd_n_s16_x(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_n_s8_z))) +svint8_t svqadd_n_s8_z(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_n_s32_z))) +svint32_t svqadd_n_s32_z(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_n_s64_z))) +svint64_t svqadd_n_s64_z(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_n_s16_z))) +svint16_t svqadd_n_s16_z(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_n_u8_m))) +svuint8_t svqadd_n_u8_m(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_n_u32_m))) +svuint32_t svqadd_n_u32_m(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_n_u64_m))) +svuint64_t svqadd_n_u64_m(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_n_u16_m))) +svuint16_t svqadd_n_u16_m(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_n_u8_x))) +svuint8_t svqadd_n_u8_x(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_n_u32_x))) +svuint32_t svqadd_n_u32_x(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_n_u64_x))) +svuint64_t svqadd_n_u64_x(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_n_u16_x))) +svuint16_t svqadd_n_u16_x(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_n_u8_z))) +svuint8_t svqadd_n_u8_z(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_n_u32_z))) +svuint32_t svqadd_n_u32_z(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_n_u64_z))) +svuint64_t svqadd_n_u64_z(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_n_u16_z))) +svuint16_t svqadd_n_u16_z(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_s8_m))) +svint8_t svqadd_s8_m(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_s32_m))) +svint32_t svqadd_s32_m(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_s64_m))) +svint64_t svqadd_s64_m(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_s16_m))) +svint16_t svqadd_s16_m(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_s8_x))) +svint8_t svqadd_s8_x(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_s32_x))) +svint32_t svqadd_s32_x(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_s64_x))) +svint64_t svqadd_s64_x(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_s16_x))) +svint16_t svqadd_s16_x(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_s8_z))) +svint8_t svqadd_s8_z(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_s32_z))) +svint32_t svqadd_s32_z(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_s64_z))) +svint64_t svqadd_s64_z(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_s16_z))) +svint16_t svqadd_s16_z(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_u8_m))) +svuint8_t svqadd_u8_m(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_u32_m))) +svuint32_t svqadd_u32_m(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_u64_m))) +svuint64_t svqadd_u64_m(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_u16_m))) +svuint16_t svqadd_u16_m(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_u8_x))) +svuint8_t svqadd_u8_x(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_u32_x))) +svuint32_t svqadd_u32_x(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_u64_x))) +svuint64_t svqadd_u64_x(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_u16_x))) +svuint16_t svqadd_u16_x(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_u8_z))) +svuint8_t svqadd_u8_z(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_u32_z))) +svuint32_t svqadd_u32_z(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_u64_z))) +svuint64_t svqadd_u64_z(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqadd_u16_z))) +svuint16_t svqadd_u16_z(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqcadd_s8))) +svint8_t svqcadd_s8(svint8_t, svint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqcadd_s32))) +svint32_t svqcadd_s32(svint32_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqcadd_s64))) +svint64_t svqcadd_s64(svint64_t, svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqcadd_s16))) +svint16_t svqcadd_s16(svint16_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmlalb_n_s32))) +svint32_t svqdmlalb_n_s32(svint32_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmlalb_n_s64))) +svint64_t svqdmlalb_n_s64(svint64_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmlalb_n_s16))) +svint16_t svqdmlalb_n_s16(svint16_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmlalb_s32))) +svint32_t svqdmlalb_s32(svint32_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmlalb_s64))) +svint64_t svqdmlalb_s64(svint64_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmlalb_s16))) +svint16_t svqdmlalb_s16(svint16_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmlalb_lane_s32))) +svint32_t svqdmlalb_lane_s32(svint32_t, svint16_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmlalb_lane_s64))) +svint64_t svqdmlalb_lane_s64(svint64_t, svint32_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmlalbt_n_s32))) +svint32_t svqdmlalbt_n_s32(svint32_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmlalbt_n_s64))) +svint64_t svqdmlalbt_n_s64(svint64_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmlalbt_n_s16))) +svint16_t svqdmlalbt_n_s16(svint16_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmlalbt_s32))) +svint32_t svqdmlalbt_s32(svint32_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmlalbt_s64))) +svint64_t svqdmlalbt_s64(svint64_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmlalbt_s16))) +svint16_t svqdmlalbt_s16(svint16_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmlalt_n_s32))) +svint32_t svqdmlalt_n_s32(svint32_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmlalt_n_s64))) +svint64_t svqdmlalt_n_s64(svint64_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmlalt_n_s16))) +svint16_t svqdmlalt_n_s16(svint16_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmlalt_s32))) +svint32_t svqdmlalt_s32(svint32_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmlalt_s64))) +svint64_t svqdmlalt_s64(svint64_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmlalt_s16))) +svint16_t svqdmlalt_s16(svint16_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmlalt_lane_s32))) +svint32_t svqdmlalt_lane_s32(svint32_t, svint16_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmlalt_lane_s64))) +svint64_t svqdmlalt_lane_s64(svint64_t, svint32_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmlslb_n_s32))) +svint32_t svqdmlslb_n_s32(svint32_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmlslb_n_s64))) +svint64_t svqdmlslb_n_s64(svint64_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmlslb_n_s16))) +svint16_t svqdmlslb_n_s16(svint16_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmlslb_s32))) +svint32_t svqdmlslb_s32(svint32_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmlslb_s64))) +svint64_t svqdmlslb_s64(svint64_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmlslb_s16))) +svint16_t svqdmlslb_s16(svint16_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmlslb_lane_s32))) +svint32_t svqdmlslb_lane_s32(svint32_t, svint16_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmlslb_lane_s64))) +svint64_t svqdmlslb_lane_s64(svint64_t, svint32_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmlslbt_n_s32))) +svint32_t svqdmlslbt_n_s32(svint32_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmlslbt_n_s64))) +svint64_t svqdmlslbt_n_s64(svint64_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmlslbt_n_s16))) +svint16_t svqdmlslbt_n_s16(svint16_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmlslbt_s32))) +svint32_t svqdmlslbt_s32(svint32_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmlslbt_s64))) +svint64_t svqdmlslbt_s64(svint64_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmlslbt_s16))) +svint16_t svqdmlslbt_s16(svint16_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmlslt_n_s32))) +svint32_t svqdmlslt_n_s32(svint32_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmlslt_n_s64))) +svint64_t svqdmlslt_n_s64(svint64_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmlslt_n_s16))) +svint16_t svqdmlslt_n_s16(svint16_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmlslt_s32))) +svint32_t svqdmlslt_s32(svint32_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmlslt_s64))) +svint64_t svqdmlslt_s64(svint64_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmlslt_s16))) +svint16_t svqdmlslt_s16(svint16_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmlslt_lane_s32))) +svint32_t svqdmlslt_lane_s32(svint32_t, svint16_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmlslt_lane_s64))) +svint64_t svqdmlslt_lane_s64(svint64_t, svint32_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmulh_n_s8))) +svint8_t svqdmulh_n_s8(svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmulh_n_s32))) +svint32_t svqdmulh_n_s32(svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmulh_n_s64))) +svint64_t svqdmulh_n_s64(svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmulh_n_s16))) +svint16_t svqdmulh_n_s16(svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmulh_s8))) +svint8_t svqdmulh_s8(svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmulh_s32))) +svint32_t svqdmulh_s32(svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmulh_s64))) +svint64_t svqdmulh_s64(svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmulh_s16))) +svint16_t svqdmulh_s16(svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmulh_lane_s32))) +svint32_t svqdmulh_lane_s32(svint32_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmulh_lane_s64))) +svint64_t svqdmulh_lane_s64(svint64_t, svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmulh_lane_s16))) +svint16_t svqdmulh_lane_s16(svint16_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmullb_n_s32))) +svint32_t svqdmullb_n_s32(svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmullb_n_s64))) +svint64_t svqdmullb_n_s64(svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmullb_n_s16))) +svint16_t svqdmullb_n_s16(svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmullb_s32))) +svint32_t svqdmullb_s32(svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmullb_s64))) +svint64_t svqdmullb_s64(svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmullb_s16))) +svint16_t svqdmullb_s16(svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmullb_lane_s32))) +svint32_t svqdmullb_lane_s32(svint16_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmullb_lane_s64))) +svint64_t svqdmullb_lane_s64(svint32_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmullt_n_s32))) +svint32_t svqdmullt_n_s32(svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmullt_n_s64))) +svint64_t svqdmullt_n_s64(svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmullt_n_s16))) +svint16_t svqdmullt_n_s16(svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmullt_s32))) +svint32_t svqdmullt_s32(svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmullt_s64))) +svint64_t svqdmullt_s64(svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmullt_s16))) +svint16_t svqdmullt_s16(svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmullt_lane_s32))) +svint32_t svqdmullt_lane_s32(svint16_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqdmullt_lane_s64))) +svint64_t svqdmullt_lane_s64(svint32_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqneg_s8_m))) +svint8_t svqneg_s8_m(svint8_t, svbool_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqneg_s32_m))) +svint32_t svqneg_s32_m(svint32_t, svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqneg_s64_m))) +svint64_t svqneg_s64_m(svint64_t, svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqneg_s16_m))) +svint16_t svqneg_s16_m(svint16_t, svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqneg_s8_x))) +svint8_t svqneg_s8_x(svbool_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqneg_s32_x))) +svint32_t svqneg_s32_x(svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqneg_s64_x))) +svint64_t svqneg_s64_x(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqneg_s16_x))) +svint16_t svqneg_s16_x(svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqneg_s8_z))) +svint8_t svqneg_s8_z(svbool_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqneg_s32_z))) +svint32_t svqneg_s32_z(svbool_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqneg_s64_z))) +svint64_t svqneg_s64_z(svbool_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqneg_s16_z))) +svint16_t svqneg_s16_z(svbool_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrdcmlah_s8))) +svint8_t svqrdcmlah_s8(svint8_t, svint8_t, svint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrdcmlah_s32))) +svint32_t svqrdcmlah_s32(svint32_t, svint32_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrdcmlah_s64))) +svint64_t svqrdcmlah_s64(svint64_t, svint64_t, svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrdcmlah_s16))) +svint16_t svqrdcmlah_s16(svint16_t, svint16_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrdcmlah_lane_s32))) +svint32_t svqrdcmlah_lane_s32(svint32_t, svint32_t, svint32_t, uint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrdcmlah_lane_s16))) +svint16_t svqrdcmlah_lane_s16(svint16_t, svint16_t, svint16_t, uint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrdmlah_n_s8))) +svint8_t svqrdmlah_n_s8(svint8_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrdmlah_n_s32))) +svint32_t svqrdmlah_n_s32(svint32_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrdmlah_n_s64))) +svint64_t svqrdmlah_n_s64(svint64_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrdmlah_n_s16))) +svint16_t svqrdmlah_n_s16(svint16_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrdmlah_s8))) +svint8_t svqrdmlah_s8(svint8_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrdmlah_s32))) +svint32_t svqrdmlah_s32(svint32_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrdmlah_s64))) +svint64_t svqrdmlah_s64(svint64_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrdmlah_s16))) +svint16_t svqrdmlah_s16(svint16_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrdmlah_lane_s32))) +svint32_t svqrdmlah_lane_s32(svint32_t, svint32_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrdmlah_lane_s64))) +svint64_t svqrdmlah_lane_s64(svint64_t, svint64_t, svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrdmlah_lane_s16))) +svint16_t svqrdmlah_lane_s16(svint16_t, svint16_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrdmlsh_n_s8))) +svint8_t svqrdmlsh_n_s8(svint8_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrdmlsh_n_s32))) +svint32_t svqrdmlsh_n_s32(svint32_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrdmlsh_n_s64))) +svint64_t svqrdmlsh_n_s64(svint64_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrdmlsh_n_s16))) +svint16_t svqrdmlsh_n_s16(svint16_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrdmlsh_s8))) +svint8_t svqrdmlsh_s8(svint8_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrdmlsh_s32))) +svint32_t svqrdmlsh_s32(svint32_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrdmlsh_s64))) +svint64_t svqrdmlsh_s64(svint64_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrdmlsh_s16))) +svint16_t svqrdmlsh_s16(svint16_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrdmlsh_lane_s32))) +svint32_t svqrdmlsh_lane_s32(svint32_t, svint32_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrdmlsh_lane_s64))) +svint64_t svqrdmlsh_lane_s64(svint64_t, svint64_t, svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrdmlsh_lane_s16))) +svint16_t svqrdmlsh_lane_s16(svint16_t, svint16_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrdmulh_n_s8))) +svint8_t svqrdmulh_n_s8(svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrdmulh_n_s32))) +svint32_t svqrdmulh_n_s32(svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrdmulh_n_s64))) +svint64_t svqrdmulh_n_s64(svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrdmulh_n_s16))) +svint16_t svqrdmulh_n_s16(svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrdmulh_s8))) +svint8_t svqrdmulh_s8(svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrdmulh_s32))) +svint32_t svqrdmulh_s32(svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrdmulh_s64))) +svint64_t svqrdmulh_s64(svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrdmulh_s16))) +svint16_t svqrdmulh_s16(svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrdmulh_lane_s32))) +svint32_t svqrdmulh_lane_s32(svint32_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrdmulh_lane_s64))) +svint64_t svqrdmulh_lane_s64(svint64_t, svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrdmulh_lane_s16))) +svint16_t svqrdmulh_lane_s16(svint16_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_n_s8_m))) +svint8_t svqrshl_n_s8_m(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_n_s32_m))) +svint32_t svqrshl_n_s32_m(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_n_s64_m))) +svint64_t svqrshl_n_s64_m(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_n_s16_m))) +svint16_t svqrshl_n_s16_m(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_n_s8_x))) +svint8_t svqrshl_n_s8_x(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_n_s32_x))) +svint32_t svqrshl_n_s32_x(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_n_s64_x))) +svint64_t svqrshl_n_s64_x(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_n_s16_x))) +svint16_t svqrshl_n_s16_x(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_n_s8_z))) +svint8_t svqrshl_n_s8_z(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_n_s32_z))) +svint32_t svqrshl_n_s32_z(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_n_s64_z))) +svint64_t svqrshl_n_s64_z(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_n_s16_z))) +svint16_t svqrshl_n_s16_z(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_n_u8_m))) +svuint8_t svqrshl_n_u8_m(svbool_t, svuint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_n_u32_m))) +svuint32_t svqrshl_n_u32_m(svbool_t, svuint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_n_u64_m))) +svuint64_t svqrshl_n_u64_m(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_n_u16_m))) +svuint16_t svqrshl_n_u16_m(svbool_t, svuint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_n_u8_x))) +svuint8_t svqrshl_n_u8_x(svbool_t, svuint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_n_u32_x))) +svuint32_t svqrshl_n_u32_x(svbool_t, svuint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_n_u64_x))) +svuint64_t svqrshl_n_u64_x(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_n_u16_x))) +svuint16_t svqrshl_n_u16_x(svbool_t, svuint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_n_u8_z))) +svuint8_t svqrshl_n_u8_z(svbool_t, svuint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_n_u32_z))) +svuint32_t svqrshl_n_u32_z(svbool_t, svuint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_n_u64_z))) +svuint64_t svqrshl_n_u64_z(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_n_u16_z))) +svuint16_t svqrshl_n_u16_z(svbool_t, svuint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_s8_m))) +svint8_t svqrshl_s8_m(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_s32_m))) +svint32_t svqrshl_s32_m(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_s64_m))) +svint64_t svqrshl_s64_m(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_s16_m))) +svint16_t svqrshl_s16_m(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_s8_x))) +svint8_t svqrshl_s8_x(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_s32_x))) +svint32_t svqrshl_s32_x(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_s64_x))) +svint64_t svqrshl_s64_x(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_s16_x))) +svint16_t svqrshl_s16_x(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_s8_z))) +svint8_t svqrshl_s8_z(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_s32_z))) +svint32_t svqrshl_s32_z(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_s64_z))) +svint64_t svqrshl_s64_z(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_s16_z))) +svint16_t svqrshl_s16_z(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_u8_m))) +svuint8_t svqrshl_u8_m(svbool_t, svuint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_u32_m))) +svuint32_t svqrshl_u32_m(svbool_t, svuint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_u64_m))) +svuint64_t svqrshl_u64_m(svbool_t, svuint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_u16_m))) +svuint16_t svqrshl_u16_m(svbool_t, svuint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_u8_x))) +svuint8_t svqrshl_u8_x(svbool_t, svuint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_u32_x))) +svuint32_t svqrshl_u32_x(svbool_t, svuint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_u64_x))) +svuint64_t svqrshl_u64_x(svbool_t, svuint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_u16_x))) +svuint16_t svqrshl_u16_x(svbool_t, svuint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_u8_z))) +svuint8_t svqrshl_u8_z(svbool_t, svuint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_u32_z))) +svuint32_t svqrshl_u32_z(svbool_t, svuint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_u64_z))) +svuint64_t svqrshl_u64_z(svbool_t, svuint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshl_u16_z))) +svuint16_t svqrshl_u16_z(svbool_t, svuint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshrnb_n_s32))) +svint16_t svqrshrnb_n_s32(svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshrnb_n_s64))) +svint32_t svqrshrnb_n_s64(svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshrnb_n_s16))) +svint8_t svqrshrnb_n_s16(svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshrnb_n_u32))) +svuint16_t svqrshrnb_n_u32(svuint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshrnb_n_u64))) +svuint32_t svqrshrnb_n_u64(svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshrnb_n_u16))) +svuint8_t svqrshrnb_n_u16(svuint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshrnt_n_s32))) +svint16_t svqrshrnt_n_s32(svint16_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshrnt_n_s64))) +svint32_t svqrshrnt_n_s64(svint32_t, svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshrnt_n_s16))) +svint8_t svqrshrnt_n_s16(svint8_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshrnt_n_u32))) +svuint16_t svqrshrnt_n_u32(svuint16_t, svuint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshrnt_n_u64))) +svuint32_t svqrshrnt_n_u64(svuint32_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshrnt_n_u16))) +svuint8_t svqrshrnt_n_u16(svuint8_t, svuint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshrunb_n_s32))) +svuint16_t svqrshrunb_n_s32(svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshrunb_n_s64))) +svuint32_t svqrshrunb_n_s64(svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshrunb_n_s16))) +svuint8_t svqrshrunb_n_s16(svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshrunt_n_s32))) +svuint16_t svqrshrunt_n_s32(svuint16_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshrunt_n_s64))) +svuint32_t svqrshrunt_n_s64(svuint32_t, svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqrshrunt_n_s16))) +svuint8_t svqrshrunt_n_s16(svuint8_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_n_s8_m))) +svint8_t svqshl_n_s8_m(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_n_s32_m))) +svint32_t svqshl_n_s32_m(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_n_s64_m))) +svint64_t svqshl_n_s64_m(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_n_s16_m))) +svint16_t svqshl_n_s16_m(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_n_s8_x))) +svint8_t svqshl_n_s8_x(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_n_s32_x))) +svint32_t svqshl_n_s32_x(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_n_s64_x))) +svint64_t svqshl_n_s64_x(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_n_s16_x))) +svint16_t svqshl_n_s16_x(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_n_s8_z))) +svint8_t svqshl_n_s8_z(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_n_s32_z))) +svint32_t svqshl_n_s32_z(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_n_s64_z))) +svint64_t svqshl_n_s64_z(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_n_s16_z))) +svint16_t svqshl_n_s16_z(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_n_u8_m))) +svuint8_t svqshl_n_u8_m(svbool_t, svuint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_n_u32_m))) +svuint32_t svqshl_n_u32_m(svbool_t, svuint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_n_u64_m))) +svuint64_t svqshl_n_u64_m(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_n_u16_m))) +svuint16_t svqshl_n_u16_m(svbool_t, svuint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_n_u8_x))) +svuint8_t svqshl_n_u8_x(svbool_t, svuint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_n_u32_x))) +svuint32_t svqshl_n_u32_x(svbool_t, svuint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_n_u64_x))) +svuint64_t svqshl_n_u64_x(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_n_u16_x))) +svuint16_t svqshl_n_u16_x(svbool_t, svuint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_n_u8_z))) +svuint8_t svqshl_n_u8_z(svbool_t, svuint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_n_u32_z))) +svuint32_t svqshl_n_u32_z(svbool_t, svuint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_n_u64_z))) +svuint64_t svqshl_n_u64_z(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_n_u16_z))) +svuint16_t svqshl_n_u16_z(svbool_t, svuint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_s8_m))) +svint8_t svqshl_s8_m(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_s32_m))) +svint32_t svqshl_s32_m(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_s64_m))) +svint64_t svqshl_s64_m(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_s16_m))) +svint16_t svqshl_s16_m(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_s8_x))) +svint8_t svqshl_s8_x(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_s32_x))) +svint32_t svqshl_s32_x(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_s64_x))) +svint64_t svqshl_s64_x(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_s16_x))) +svint16_t svqshl_s16_x(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_s8_z))) +svint8_t svqshl_s8_z(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_s32_z))) +svint32_t svqshl_s32_z(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_s64_z))) +svint64_t svqshl_s64_z(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_s16_z))) +svint16_t svqshl_s16_z(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_u8_m))) +svuint8_t svqshl_u8_m(svbool_t, svuint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_u32_m))) +svuint32_t svqshl_u32_m(svbool_t, svuint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_u64_m))) +svuint64_t svqshl_u64_m(svbool_t, svuint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_u16_m))) +svuint16_t svqshl_u16_m(svbool_t, svuint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_u8_x))) +svuint8_t svqshl_u8_x(svbool_t, svuint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_u32_x))) +svuint32_t svqshl_u32_x(svbool_t, svuint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_u64_x))) +svuint64_t svqshl_u64_x(svbool_t, svuint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_u16_x))) +svuint16_t svqshl_u16_x(svbool_t, svuint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_u8_z))) +svuint8_t svqshl_u8_z(svbool_t, svuint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_u32_z))) +svuint32_t svqshl_u32_z(svbool_t, svuint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_u64_z))) +svuint64_t svqshl_u64_z(svbool_t, svuint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshl_u16_z))) +svuint16_t svqshl_u16_z(svbool_t, svuint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshlu_n_s8_m))) +svuint8_t svqshlu_n_s8_m(svbool_t, svint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshlu_n_s32_m))) +svuint32_t svqshlu_n_s32_m(svbool_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshlu_n_s64_m))) +svuint64_t svqshlu_n_s64_m(svbool_t, svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshlu_n_s16_m))) +svuint16_t svqshlu_n_s16_m(svbool_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshlu_n_s8_x))) +svuint8_t svqshlu_n_s8_x(svbool_t, svint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshlu_n_s32_x))) +svuint32_t svqshlu_n_s32_x(svbool_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshlu_n_s64_x))) +svuint64_t svqshlu_n_s64_x(svbool_t, svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshlu_n_s16_x))) +svuint16_t svqshlu_n_s16_x(svbool_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshlu_n_s8_z))) +svuint8_t svqshlu_n_s8_z(svbool_t, svint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshlu_n_s32_z))) +svuint32_t svqshlu_n_s32_z(svbool_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshlu_n_s64_z))) +svuint64_t svqshlu_n_s64_z(svbool_t, svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshlu_n_s16_z))) +svuint16_t svqshlu_n_s16_z(svbool_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshrnb_n_s32))) +svint16_t svqshrnb_n_s32(svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshrnb_n_s64))) +svint32_t svqshrnb_n_s64(svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshrnb_n_s16))) +svint8_t svqshrnb_n_s16(svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshrnb_n_u32))) +svuint16_t svqshrnb_n_u32(svuint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshrnb_n_u64))) +svuint32_t svqshrnb_n_u64(svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshrnb_n_u16))) +svuint8_t svqshrnb_n_u16(svuint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshrnt_n_s32))) +svint16_t svqshrnt_n_s32(svint16_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshrnt_n_s64))) +svint32_t svqshrnt_n_s64(svint32_t, svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshrnt_n_s16))) +svint8_t svqshrnt_n_s16(svint8_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshrnt_n_u32))) +svuint16_t svqshrnt_n_u32(svuint16_t, svuint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshrnt_n_u64))) +svuint32_t svqshrnt_n_u64(svuint32_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshrnt_n_u16))) +svuint8_t svqshrnt_n_u16(svuint8_t, svuint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshrunb_n_s32))) +svuint16_t svqshrunb_n_s32(svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshrunb_n_s64))) +svuint32_t svqshrunb_n_s64(svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshrunb_n_s16))) +svuint8_t svqshrunb_n_s16(svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshrunt_n_s32))) +svuint16_t svqshrunt_n_s32(svuint16_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshrunt_n_s64))) +svuint32_t svqshrunt_n_s64(svuint32_t, svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqshrunt_n_s16))) +svuint8_t svqshrunt_n_s16(svuint8_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_n_s8_m))) +svint8_t svqsub_n_s8_m(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_n_s32_m))) +svint32_t svqsub_n_s32_m(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_n_s64_m))) +svint64_t svqsub_n_s64_m(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_n_s16_m))) +svint16_t svqsub_n_s16_m(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_n_s8_x))) +svint8_t svqsub_n_s8_x(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_n_s32_x))) +svint32_t svqsub_n_s32_x(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_n_s64_x))) +svint64_t svqsub_n_s64_x(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_n_s16_x))) +svint16_t svqsub_n_s16_x(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_n_s8_z))) +svint8_t svqsub_n_s8_z(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_n_s32_z))) +svint32_t svqsub_n_s32_z(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_n_s64_z))) +svint64_t svqsub_n_s64_z(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_n_s16_z))) +svint16_t svqsub_n_s16_z(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_n_u8_m))) +svuint8_t svqsub_n_u8_m(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_n_u32_m))) +svuint32_t svqsub_n_u32_m(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_n_u64_m))) +svuint64_t svqsub_n_u64_m(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_n_u16_m))) +svuint16_t svqsub_n_u16_m(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_n_u8_x))) +svuint8_t svqsub_n_u8_x(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_n_u32_x))) +svuint32_t svqsub_n_u32_x(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_n_u64_x))) +svuint64_t svqsub_n_u64_x(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_n_u16_x))) +svuint16_t svqsub_n_u16_x(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_n_u8_z))) +svuint8_t svqsub_n_u8_z(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_n_u32_z))) +svuint32_t svqsub_n_u32_z(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_n_u64_z))) +svuint64_t svqsub_n_u64_z(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_n_u16_z))) +svuint16_t svqsub_n_u16_z(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_s8_m))) +svint8_t svqsub_s8_m(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_s32_m))) +svint32_t svqsub_s32_m(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_s64_m))) +svint64_t svqsub_s64_m(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_s16_m))) +svint16_t svqsub_s16_m(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_s8_x))) +svint8_t svqsub_s8_x(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_s32_x))) +svint32_t svqsub_s32_x(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_s64_x))) +svint64_t svqsub_s64_x(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_s16_x))) +svint16_t svqsub_s16_x(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_s8_z))) +svint8_t svqsub_s8_z(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_s32_z))) +svint32_t svqsub_s32_z(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_s64_z))) +svint64_t svqsub_s64_z(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_s16_z))) +svint16_t svqsub_s16_z(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_u8_m))) +svuint8_t svqsub_u8_m(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_u32_m))) +svuint32_t svqsub_u32_m(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_u64_m))) +svuint64_t svqsub_u64_m(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_u16_m))) +svuint16_t svqsub_u16_m(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_u8_x))) +svuint8_t svqsub_u8_x(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_u32_x))) +svuint32_t svqsub_u32_x(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_u64_x))) +svuint64_t svqsub_u64_x(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_u16_x))) +svuint16_t svqsub_u16_x(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_u8_z))) +svuint8_t svqsub_u8_z(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_u32_z))) +svuint32_t svqsub_u32_z(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_u64_z))) +svuint64_t svqsub_u64_z(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsub_u16_z))) +svuint16_t svqsub_u16_z(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_n_s8_m))) +svint8_t svqsubr_n_s8_m(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_n_s32_m))) +svint32_t svqsubr_n_s32_m(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_n_s64_m))) +svint64_t svqsubr_n_s64_m(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_n_s16_m))) +svint16_t svqsubr_n_s16_m(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_n_s8_x))) +svint8_t svqsubr_n_s8_x(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_n_s32_x))) +svint32_t svqsubr_n_s32_x(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_n_s64_x))) +svint64_t svqsubr_n_s64_x(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_n_s16_x))) +svint16_t svqsubr_n_s16_x(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_n_s8_z))) +svint8_t svqsubr_n_s8_z(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_n_s32_z))) +svint32_t svqsubr_n_s32_z(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_n_s64_z))) +svint64_t svqsubr_n_s64_z(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_n_s16_z))) +svint16_t svqsubr_n_s16_z(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_n_u8_m))) +svuint8_t svqsubr_n_u8_m(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_n_u32_m))) +svuint32_t svqsubr_n_u32_m(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_n_u64_m))) +svuint64_t svqsubr_n_u64_m(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_n_u16_m))) +svuint16_t svqsubr_n_u16_m(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_n_u8_x))) +svuint8_t svqsubr_n_u8_x(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_n_u32_x))) +svuint32_t svqsubr_n_u32_x(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_n_u64_x))) +svuint64_t svqsubr_n_u64_x(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_n_u16_x))) +svuint16_t svqsubr_n_u16_x(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_n_u8_z))) +svuint8_t svqsubr_n_u8_z(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_n_u32_z))) +svuint32_t svqsubr_n_u32_z(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_n_u64_z))) +svuint64_t svqsubr_n_u64_z(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_n_u16_z))) +svuint16_t svqsubr_n_u16_z(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_s8_m))) +svint8_t svqsubr_s8_m(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_s32_m))) +svint32_t svqsubr_s32_m(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_s64_m))) +svint64_t svqsubr_s64_m(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_s16_m))) +svint16_t svqsubr_s16_m(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_s8_x))) +svint8_t svqsubr_s8_x(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_s32_x))) +svint32_t svqsubr_s32_x(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_s64_x))) +svint64_t svqsubr_s64_x(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_s16_x))) +svint16_t svqsubr_s16_x(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_s8_z))) +svint8_t svqsubr_s8_z(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_s32_z))) +svint32_t svqsubr_s32_z(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_s64_z))) +svint64_t svqsubr_s64_z(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_s16_z))) +svint16_t svqsubr_s16_z(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_u8_m))) +svuint8_t svqsubr_u8_m(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_u32_m))) +svuint32_t svqsubr_u32_m(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_u64_m))) +svuint64_t svqsubr_u64_m(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_u16_m))) +svuint16_t svqsubr_u16_m(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_u8_x))) +svuint8_t svqsubr_u8_x(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_u32_x))) +svuint32_t svqsubr_u32_x(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_u64_x))) +svuint64_t svqsubr_u64_x(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_u16_x))) +svuint16_t svqsubr_u16_x(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_u8_z))) +svuint8_t svqsubr_u8_z(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_u32_z))) +svuint32_t svqsubr_u32_z(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_u64_z))) +svuint64_t svqsubr_u64_z(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqsubr_u16_z))) +svuint16_t svqsubr_u16_z(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqxtnb_s32))) +svint16_t svqxtnb_s32(svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqxtnb_s64))) +svint32_t svqxtnb_s64(svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqxtnb_s16))) +svint8_t svqxtnb_s16(svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqxtnb_u32))) +svuint16_t svqxtnb_u32(svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqxtnb_u64))) +svuint32_t svqxtnb_u64(svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqxtnb_u16))) +svuint8_t svqxtnb_u16(svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqxtnt_s32))) +svint16_t svqxtnt_s32(svint16_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqxtnt_s64))) +svint32_t svqxtnt_s64(svint32_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqxtnt_s16))) +svint8_t svqxtnt_s16(svint8_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqxtnt_u32))) +svuint16_t svqxtnt_u32(svuint16_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqxtnt_u64))) +svuint32_t svqxtnt_u64(svuint32_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqxtnt_u16))) +svuint8_t svqxtnt_u16(svuint8_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqxtunb_s32))) +svuint16_t svqxtunb_s32(svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqxtunb_s64))) +svuint32_t svqxtunb_s64(svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqxtunb_s16))) +svuint8_t svqxtunb_s16(svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqxtunt_s32))) +svuint16_t svqxtunt_s32(svuint16_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqxtunt_s64))) +svuint32_t svqxtunt_s64(svuint32_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svqxtunt_s16))) +svuint8_t svqxtunt_s16(svuint8_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svraddhnb_n_u32))) +svuint16_t svraddhnb_n_u32(svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svraddhnb_n_u64))) +svuint32_t svraddhnb_n_u64(svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svraddhnb_n_u16))) +svuint8_t svraddhnb_n_u16(svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svraddhnb_n_s32))) +svint16_t svraddhnb_n_s32(svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svraddhnb_n_s64))) +svint32_t svraddhnb_n_s64(svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svraddhnb_n_s16))) +svint8_t svraddhnb_n_s16(svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svraddhnb_u32))) +svuint16_t svraddhnb_u32(svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svraddhnb_u64))) +svuint32_t svraddhnb_u64(svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svraddhnb_u16))) +svuint8_t svraddhnb_u16(svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svraddhnb_s32))) +svint16_t svraddhnb_s32(svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svraddhnb_s64))) +svint32_t svraddhnb_s64(svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svraddhnb_s16))) +svint8_t svraddhnb_s16(svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svraddhnt_n_u32))) +svuint16_t svraddhnt_n_u32(svuint16_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svraddhnt_n_u64))) +svuint32_t svraddhnt_n_u64(svuint32_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svraddhnt_n_u16))) +svuint8_t svraddhnt_n_u16(svuint8_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svraddhnt_n_s32))) +svint16_t svraddhnt_n_s32(svint16_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svraddhnt_n_s64))) +svint32_t svraddhnt_n_s64(svint32_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svraddhnt_n_s16))) +svint8_t svraddhnt_n_s16(svint8_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svraddhnt_u32))) +svuint16_t svraddhnt_u32(svuint16_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svraddhnt_u64))) +svuint32_t svraddhnt_u64(svuint32_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svraddhnt_u16))) +svuint8_t svraddhnt_u16(svuint8_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svraddhnt_s32))) +svint16_t svraddhnt_s32(svint16_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svraddhnt_s64))) +svint32_t svraddhnt_s64(svint32_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svraddhnt_s16))) +svint8_t svraddhnt_s16(svint8_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrecpe_u32_m))) +svuint32_t svrecpe_u32_m(svuint32_t, svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrecpe_u32_x))) +svuint32_t svrecpe_u32_x(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrecpe_u32_z))) +svuint32_t svrecpe_u32_z(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_n_s8_m))) +svint8_t svrhadd_n_s8_m(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_n_s32_m))) +svint32_t svrhadd_n_s32_m(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_n_s64_m))) +svint64_t svrhadd_n_s64_m(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_n_s16_m))) +svint16_t svrhadd_n_s16_m(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_n_s8_x))) +svint8_t svrhadd_n_s8_x(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_n_s32_x))) +svint32_t svrhadd_n_s32_x(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_n_s64_x))) +svint64_t svrhadd_n_s64_x(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_n_s16_x))) +svint16_t svrhadd_n_s16_x(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_n_s8_z))) +svint8_t svrhadd_n_s8_z(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_n_s32_z))) +svint32_t svrhadd_n_s32_z(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_n_s64_z))) +svint64_t svrhadd_n_s64_z(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_n_s16_z))) +svint16_t svrhadd_n_s16_z(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_n_u8_m))) +svuint8_t svrhadd_n_u8_m(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_n_u32_m))) +svuint32_t svrhadd_n_u32_m(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_n_u64_m))) +svuint64_t svrhadd_n_u64_m(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_n_u16_m))) +svuint16_t svrhadd_n_u16_m(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_n_u8_x))) +svuint8_t svrhadd_n_u8_x(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_n_u32_x))) +svuint32_t svrhadd_n_u32_x(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_n_u64_x))) +svuint64_t svrhadd_n_u64_x(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_n_u16_x))) +svuint16_t svrhadd_n_u16_x(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_n_u8_z))) +svuint8_t svrhadd_n_u8_z(svbool_t, svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_n_u32_z))) +svuint32_t svrhadd_n_u32_z(svbool_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_n_u64_z))) +svuint64_t svrhadd_n_u64_z(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_n_u16_z))) +svuint16_t svrhadd_n_u16_z(svbool_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_s8_m))) +svint8_t svrhadd_s8_m(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_s32_m))) +svint32_t svrhadd_s32_m(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_s64_m))) +svint64_t svrhadd_s64_m(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_s16_m))) +svint16_t svrhadd_s16_m(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_s8_x))) +svint8_t svrhadd_s8_x(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_s32_x))) +svint32_t svrhadd_s32_x(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_s64_x))) +svint64_t svrhadd_s64_x(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_s16_x))) +svint16_t svrhadd_s16_x(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_s8_z))) +svint8_t svrhadd_s8_z(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_s32_z))) +svint32_t svrhadd_s32_z(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_s64_z))) +svint64_t svrhadd_s64_z(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_s16_z))) +svint16_t svrhadd_s16_z(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_u8_m))) +svuint8_t svrhadd_u8_m(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_u32_m))) +svuint32_t svrhadd_u32_m(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_u64_m))) +svuint64_t svrhadd_u64_m(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_u16_m))) +svuint16_t svrhadd_u16_m(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_u8_x))) +svuint8_t svrhadd_u8_x(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_u32_x))) +svuint32_t svrhadd_u32_x(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_u64_x))) +svuint64_t svrhadd_u64_x(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_u16_x))) +svuint16_t svrhadd_u16_x(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_u8_z))) +svuint8_t svrhadd_u8_z(svbool_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_u32_z))) +svuint32_t svrhadd_u32_z(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_u64_z))) +svuint64_t svrhadd_u64_z(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrhadd_u16_z))) +svuint16_t svrhadd_u16_z(svbool_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_n_s8_m))) +svint8_t svrshl_n_s8_m(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_n_s32_m))) +svint32_t svrshl_n_s32_m(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_n_s64_m))) +svint64_t svrshl_n_s64_m(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_n_s16_m))) +svint16_t svrshl_n_s16_m(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_n_s8_x))) +svint8_t svrshl_n_s8_x(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_n_s32_x))) +svint32_t svrshl_n_s32_x(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_n_s64_x))) +svint64_t svrshl_n_s64_x(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_n_s16_x))) +svint16_t svrshl_n_s16_x(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_n_s8_z))) +svint8_t svrshl_n_s8_z(svbool_t, svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_n_s32_z))) +svint32_t svrshl_n_s32_z(svbool_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_n_s64_z))) +svint64_t svrshl_n_s64_z(svbool_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_n_s16_z))) +svint16_t svrshl_n_s16_z(svbool_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_n_u8_m))) +svuint8_t svrshl_n_u8_m(svbool_t, svuint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_n_u32_m))) +svuint32_t svrshl_n_u32_m(svbool_t, svuint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_n_u64_m))) +svuint64_t svrshl_n_u64_m(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_n_u16_m))) +svuint16_t svrshl_n_u16_m(svbool_t, svuint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_n_u8_x))) +svuint8_t svrshl_n_u8_x(svbool_t, svuint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_n_u32_x))) +svuint32_t svrshl_n_u32_x(svbool_t, svuint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_n_u64_x))) +svuint64_t svrshl_n_u64_x(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_n_u16_x))) +svuint16_t svrshl_n_u16_x(svbool_t, svuint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_n_u8_z))) +svuint8_t svrshl_n_u8_z(svbool_t, svuint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_n_u32_z))) +svuint32_t svrshl_n_u32_z(svbool_t, svuint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_n_u64_z))) +svuint64_t svrshl_n_u64_z(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_n_u16_z))) +svuint16_t svrshl_n_u16_z(svbool_t, svuint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_s8_m))) +svint8_t svrshl_s8_m(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_s32_m))) +svint32_t svrshl_s32_m(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_s64_m))) +svint64_t svrshl_s64_m(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_s16_m))) +svint16_t svrshl_s16_m(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_s8_x))) +svint8_t svrshl_s8_x(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_s32_x))) +svint32_t svrshl_s32_x(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_s64_x))) +svint64_t svrshl_s64_x(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_s16_x))) +svint16_t svrshl_s16_x(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_s8_z))) +svint8_t svrshl_s8_z(svbool_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_s32_z))) +svint32_t svrshl_s32_z(svbool_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_s64_z))) +svint64_t svrshl_s64_z(svbool_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_s16_z))) +svint16_t svrshl_s16_z(svbool_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_u8_m))) +svuint8_t svrshl_u8_m(svbool_t, svuint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_u32_m))) +svuint32_t svrshl_u32_m(svbool_t, svuint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_u64_m))) +svuint64_t svrshl_u64_m(svbool_t, svuint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_u16_m))) +svuint16_t svrshl_u16_m(svbool_t, svuint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_u8_x))) +svuint8_t svrshl_u8_x(svbool_t, svuint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_u32_x))) +svuint32_t svrshl_u32_x(svbool_t, svuint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_u64_x))) +svuint64_t svrshl_u64_x(svbool_t, svuint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_u16_x))) +svuint16_t svrshl_u16_x(svbool_t, svuint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_u8_z))) +svuint8_t svrshl_u8_z(svbool_t, svuint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_u32_z))) +svuint32_t svrshl_u32_z(svbool_t, svuint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_u64_z))) +svuint64_t svrshl_u64_z(svbool_t, svuint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshl_u16_z))) +svuint16_t svrshl_u16_z(svbool_t, svuint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshr_n_s8_m))) +svint8_t svrshr_n_s8_m(svbool_t, svint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshr_n_s32_m))) +svint32_t svrshr_n_s32_m(svbool_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshr_n_s64_m))) +svint64_t svrshr_n_s64_m(svbool_t, svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshr_n_s16_m))) +svint16_t svrshr_n_s16_m(svbool_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshr_n_u8_m))) +svuint8_t svrshr_n_u8_m(svbool_t, svuint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshr_n_u32_m))) +svuint32_t svrshr_n_u32_m(svbool_t, svuint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshr_n_u64_m))) +svuint64_t svrshr_n_u64_m(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshr_n_u16_m))) +svuint16_t svrshr_n_u16_m(svbool_t, svuint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshr_n_s8_x))) +svint8_t svrshr_n_s8_x(svbool_t, svint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshr_n_s32_x))) +svint32_t svrshr_n_s32_x(svbool_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshr_n_s64_x))) +svint64_t svrshr_n_s64_x(svbool_t, svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshr_n_s16_x))) +svint16_t svrshr_n_s16_x(svbool_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshr_n_u8_x))) +svuint8_t svrshr_n_u8_x(svbool_t, svuint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshr_n_u32_x))) +svuint32_t svrshr_n_u32_x(svbool_t, svuint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshr_n_u64_x))) +svuint64_t svrshr_n_u64_x(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshr_n_u16_x))) +svuint16_t svrshr_n_u16_x(svbool_t, svuint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshr_n_s8_z))) +svint8_t svrshr_n_s8_z(svbool_t, svint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshr_n_s32_z))) +svint32_t svrshr_n_s32_z(svbool_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshr_n_s64_z))) +svint64_t svrshr_n_s64_z(svbool_t, svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshr_n_s16_z))) +svint16_t svrshr_n_s16_z(svbool_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshr_n_u8_z))) +svuint8_t svrshr_n_u8_z(svbool_t, svuint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshr_n_u32_z))) +svuint32_t svrshr_n_u32_z(svbool_t, svuint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshr_n_u64_z))) +svuint64_t svrshr_n_u64_z(svbool_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshr_n_u16_z))) +svuint16_t svrshr_n_u16_z(svbool_t, svuint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshrnb_n_u32))) +svuint16_t svrshrnb_n_u32(svuint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshrnb_n_u64))) +svuint32_t svrshrnb_n_u64(svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshrnb_n_u16))) +svuint8_t svrshrnb_n_u16(svuint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshrnb_n_s32))) +svint16_t svrshrnb_n_s32(svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshrnb_n_s64))) +svint32_t svrshrnb_n_s64(svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshrnb_n_s16))) +svint8_t svrshrnb_n_s16(svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshrnt_n_u32))) +svuint16_t svrshrnt_n_u32(svuint16_t, svuint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshrnt_n_u64))) +svuint32_t svrshrnt_n_u64(svuint32_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshrnt_n_u16))) +svuint8_t svrshrnt_n_u16(svuint8_t, svuint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshrnt_n_s32))) +svint16_t svrshrnt_n_s32(svint16_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshrnt_n_s64))) +svint32_t svrshrnt_n_s64(svint32_t, svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrshrnt_n_s16))) +svint8_t svrshrnt_n_s16(svint8_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrsqrte_u32_m))) +svuint32_t svrsqrte_u32_m(svuint32_t, svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrsqrte_u32_x))) +svuint32_t svrsqrte_u32_x(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrsqrte_u32_z))) +svuint32_t svrsqrte_u32_z(svbool_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrsra_n_s8))) +svint8_t svrsra_n_s8(svint8_t, svint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrsra_n_s32))) +svint32_t svrsra_n_s32(svint32_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrsra_n_s64))) +svint64_t svrsra_n_s64(svint64_t, svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrsra_n_s16))) +svint16_t svrsra_n_s16(svint16_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrsra_n_u8))) +svuint8_t svrsra_n_u8(svuint8_t, svuint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrsra_n_u32))) +svuint32_t svrsra_n_u32(svuint32_t, svuint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrsra_n_u64))) +svuint64_t svrsra_n_u64(svuint64_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrsra_n_u16))) +svuint16_t svrsra_n_u16(svuint16_t, svuint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrsubhnb_n_u32))) +svuint16_t svrsubhnb_n_u32(svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrsubhnb_n_u64))) +svuint32_t svrsubhnb_n_u64(svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrsubhnb_n_u16))) +svuint8_t svrsubhnb_n_u16(svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrsubhnb_n_s32))) +svint16_t svrsubhnb_n_s32(svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrsubhnb_n_s64))) +svint32_t svrsubhnb_n_s64(svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrsubhnb_n_s16))) +svint8_t svrsubhnb_n_s16(svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrsubhnb_u32))) +svuint16_t svrsubhnb_u32(svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrsubhnb_u64))) +svuint32_t svrsubhnb_u64(svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrsubhnb_u16))) +svuint8_t svrsubhnb_u16(svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrsubhnb_s32))) +svint16_t svrsubhnb_s32(svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrsubhnb_s64))) +svint32_t svrsubhnb_s64(svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrsubhnb_s16))) +svint8_t svrsubhnb_s16(svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrsubhnt_n_u32))) +svuint16_t svrsubhnt_n_u32(svuint16_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrsubhnt_n_u64))) +svuint32_t svrsubhnt_n_u64(svuint32_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrsubhnt_n_u16))) +svuint8_t svrsubhnt_n_u16(svuint8_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrsubhnt_n_s32))) +svint16_t svrsubhnt_n_s32(svint16_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrsubhnt_n_s64))) +svint32_t svrsubhnt_n_s64(svint32_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrsubhnt_n_s16))) +svint8_t svrsubhnt_n_s16(svint8_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrsubhnt_u32))) +svuint16_t svrsubhnt_u32(svuint16_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrsubhnt_u64))) +svuint32_t svrsubhnt_u64(svuint32_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrsubhnt_u16))) +svuint8_t svrsubhnt_u16(svuint8_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrsubhnt_s32))) +svint16_t svrsubhnt_s32(svint16_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrsubhnt_s64))) +svint32_t svrsubhnt_s64(svint32_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrsubhnt_s16))) +svint8_t svrsubhnt_s16(svint8_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsbclb_n_u32))) +svuint32_t svsbclb_n_u32(svuint32_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsbclb_n_u64))) +svuint64_t svsbclb_n_u64(svuint64_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsbclb_u32))) +svuint32_t svsbclb_u32(svuint32_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsbclb_u64))) +svuint64_t svsbclb_u64(svuint64_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsbclt_n_u32))) +svuint32_t svsbclt_n_u32(svuint32_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsbclt_n_u64))) +svuint64_t svsbclt_n_u64(svuint64_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsbclt_u32))) +svuint32_t svsbclt_u32(svuint32_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsbclt_u64))) +svuint64_t svsbclt_u64(svuint64_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svshllb_n_s32))) +svint32_t svshllb_n_s32(svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svshllb_n_s64))) +svint64_t svshllb_n_s64(svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svshllb_n_s16))) +svint16_t svshllb_n_s16(svint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svshllb_n_u32))) +svuint32_t svshllb_n_u32(svuint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svshllb_n_u64))) +svuint64_t svshllb_n_u64(svuint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svshllb_n_u16))) +svuint16_t svshllb_n_u16(svuint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svshllt_n_s32))) +svint32_t svshllt_n_s32(svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svshllt_n_s64))) +svint64_t svshllt_n_s64(svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svshllt_n_s16))) +svint16_t svshllt_n_s16(svint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svshllt_n_u32))) +svuint32_t svshllt_n_u32(svuint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svshllt_n_u64))) +svuint64_t svshllt_n_u64(svuint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svshllt_n_u16))) +svuint16_t svshllt_n_u16(svuint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svshrnb_n_u32))) +svuint16_t svshrnb_n_u32(svuint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svshrnb_n_u64))) +svuint32_t svshrnb_n_u64(svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svshrnb_n_u16))) +svuint8_t svshrnb_n_u16(svuint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svshrnb_n_s32))) +svint16_t svshrnb_n_s32(svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svshrnb_n_s64))) +svint32_t svshrnb_n_s64(svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svshrnb_n_s16))) +svint8_t svshrnb_n_s16(svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svshrnt_n_u32))) +svuint16_t svshrnt_n_u32(svuint16_t, svuint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svshrnt_n_u64))) +svuint32_t svshrnt_n_u64(svuint32_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svshrnt_n_u16))) +svuint8_t svshrnt_n_u16(svuint8_t, svuint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svshrnt_n_s32))) +svint16_t svshrnt_n_s32(svint16_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svshrnt_n_s64))) +svint32_t svshrnt_n_s64(svint32_t, svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svshrnt_n_s16))) +svint8_t svshrnt_n_s16(svint8_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsli_n_u8))) +svuint8_t svsli_n_u8(svuint8_t, svuint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsli_n_u32))) +svuint32_t svsli_n_u32(svuint32_t, svuint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsli_n_u64))) +svuint64_t svsli_n_u64(svuint64_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsli_n_u16))) +svuint16_t svsli_n_u16(svuint16_t, svuint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsli_n_s8))) +svint8_t svsli_n_s8(svint8_t, svint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsli_n_s32))) +svint32_t svsli_n_s32(svint32_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsli_n_s64))) +svint64_t svsli_n_s64(svint64_t, svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsli_n_s16))) +svint16_t svsli_n_s16(svint16_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsqadd_n_u8_m))) +svuint8_t svsqadd_n_u8_m(svbool_t, svuint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsqadd_n_u32_m))) +svuint32_t svsqadd_n_u32_m(svbool_t, svuint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsqadd_n_u64_m))) +svuint64_t svsqadd_n_u64_m(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsqadd_n_u16_m))) +svuint16_t svsqadd_n_u16_m(svbool_t, svuint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsqadd_n_u8_x))) +svuint8_t svsqadd_n_u8_x(svbool_t, svuint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsqadd_n_u32_x))) +svuint32_t svsqadd_n_u32_x(svbool_t, svuint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsqadd_n_u64_x))) +svuint64_t svsqadd_n_u64_x(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsqadd_n_u16_x))) +svuint16_t svsqadd_n_u16_x(svbool_t, svuint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsqadd_n_u8_z))) +svuint8_t svsqadd_n_u8_z(svbool_t, svuint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsqadd_n_u32_z))) +svuint32_t svsqadd_n_u32_z(svbool_t, svuint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsqadd_n_u64_z))) +svuint64_t svsqadd_n_u64_z(svbool_t, svuint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsqadd_n_u16_z))) +svuint16_t svsqadd_n_u16_z(svbool_t, svuint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsqadd_u8_m))) +svuint8_t svsqadd_u8_m(svbool_t, svuint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsqadd_u32_m))) +svuint32_t svsqadd_u32_m(svbool_t, svuint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsqadd_u64_m))) +svuint64_t svsqadd_u64_m(svbool_t, svuint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsqadd_u16_m))) +svuint16_t svsqadd_u16_m(svbool_t, svuint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsqadd_u8_x))) +svuint8_t svsqadd_u8_x(svbool_t, svuint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsqadd_u32_x))) +svuint32_t svsqadd_u32_x(svbool_t, svuint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsqadd_u64_x))) +svuint64_t svsqadd_u64_x(svbool_t, svuint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsqadd_u16_x))) +svuint16_t svsqadd_u16_x(svbool_t, svuint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsqadd_u8_z))) +svuint8_t svsqadd_u8_z(svbool_t, svuint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsqadd_u32_z))) +svuint32_t svsqadd_u32_z(svbool_t, svuint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsqadd_u64_z))) +svuint64_t svsqadd_u64_z(svbool_t, svuint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsqadd_u16_z))) +svuint16_t svsqadd_u16_z(svbool_t, svuint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsra_n_s8))) +svint8_t svsra_n_s8(svint8_t, svint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsra_n_s32))) +svint32_t svsra_n_s32(svint32_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsra_n_s64))) +svint64_t svsra_n_s64(svint64_t, svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsra_n_s16))) +svint16_t svsra_n_s16(svint16_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsra_n_u8))) +svuint8_t svsra_n_u8(svuint8_t, svuint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsra_n_u32))) +svuint32_t svsra_n_u32(svuint32_t, svuint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsra_n_u64))) +svuint64_t svsra_n_u64(svuint64_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsra_n_u16))) +svuint16_t svsra_n_u16(svuint16_t, svuint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsri_n_u8))) +svuint8_t svsri_n_u8(svuint8_t, svuint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsri_n_u32))) +svuint32_t svsri_n_u32(svuint32_t, svuint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsri_n_u64))) +svuint64_t svsri_n_u64(svuint64_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsri_n_u16))) +svuint16_t svsri_n_u16(svuint16_t, svuint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsri_n_s8))) +svint8_t svsri_n_s8(svint8_t, svint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsri_n_s32))) +svint32_t svsri_n_s32(svint32_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsri_n_s64))) +svint64_t svsri_n_s64(svint64_t, svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsri_n_s16))) +svint16_t svsri_n_s16(svint16_t, svint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_scatter_u32base_index_u32))) +void svstnt1_scatter_u32base_index_u32(svbool_t, svuint32_t, int64_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_scatter_u64base_index_u64))) +void svstnt1_scatter_u64base_index_u64(svbool_t, svuint64_t, int64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_scatter_u64base_index_f64))) +void svstnt1_scatter_u64base_index_f64(svbool_t, svuint64_t, int64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_scatter_u32base_index_f32))) +void svstnt1_scatter_u32base_index_f32(svbool_t, svuint32_t, int64_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_scatter_u32base_index_s32))) +void svstnt1_scatter_u32base_index_s32(svbool_t, svuint32_t, int64_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_scatter_u64base_index_s64))) +void svstnt1_scatter_u64base_index_s64(svbool_t, svuint64_t, int64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_scatter_u32base_offset_u32))) +void svstnt1_scatter_u32base_offset_u32(svbool_t, svuint32_t, int64_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_scatter_u64base_offset_u64))) +void svstnt1_scatter_u64base_offset_u64(svbool_t, svuint64_t, int64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_scatter_u64base_offset_f64))) +void svstnt1_scatter_u64base_offset_f64(svbool_t, svuint64_t, int64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_scatter_u32base_offset_f32))) +void svstnt1_scatter_u32base_offset_f32(svbool_t, svuint32_t, int64_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_scatter_u32base_offset_s32))) +void svstnt1_scatter_u32base_offset_s32(svbool_t, svuint32_t, int64_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_scatter_u64base_offset_s64))) +void svstnt1_scatter_u64base_offset_s64(svbool_t, svuint64_t, int64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_scatter_u32base_u32))) +void svstnt1_scatter_u32base_u32(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_scatter_u64base_u64))) +void svstnt1_scatter_u64base_u64(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_scatter_u64base_f64))) +void svstnt1_scatter_u64base_f64(svbool_t, svuint64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_scatter_u32base_f32))) +void svstnt1_scatter_u32base_f32(svbool_t, svuint32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_scatter_u32base_s32))) +void svstnt1_scatter_u32base_s32(svbool_t, svuint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_scatter_u64base_s64))) +void svstnt1_scatter_u64base_s64(svbool_t, svuint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_scatter_s64index_u64))) +void svstnt1_scatter_s64index_u64(svbool_t, uint64_t *, svint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_scatter_s64index_f64))) +void svstnt1_scatter_s64index_f64(svbool_t, float64_t *, svint64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_scatter_s64index_s64))) +void svstnt1_scatter_s64index_s64(svbool_t, int64_t *, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_scatter_u64index_u64))) +void svstnt1_scatter_u64index_u64(svbool_t, uint64_t *, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_scatter_u64index_f64))) +void svstnt1_scatter_u64index_f64(svbool_t, float64_t *, svuint64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_scatter_u64index_s64))) +void svstnt1_scatter_u64index_s64(svbool_t, int64_t *, svuint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_scatter_u32offset_u32))) +void svstnt1_scatter_u32offset_u32(svbool_t, uint32_t *, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_scatter_u32offset_f32))) +void svstnt1_scatter_u32offset_f32(svbool_t, float32_t *, svuint32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_scatter_u32offset_s32))) +void svstnt1_scatter_u32offset_s32(svbool_t, int32_t *, svuint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_scatter_s64offset_u64))) +void svstnt1_scatter_s64offset_u64(svbool_t, uint64_t *, svint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_scatter_s64offset_f64))) +void svstnt1_scatter_s64offset_f64(svbool_t, float64_t *, svint64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_scatter_s64offset_s64))) +void svstnt1_scatter_s64offset_s64(svbool_t, int64_t *, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_scatter_u64offset_u64))) +void svstnt1_scatter_u64offset_u64(svbool_t, uint64_t *, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_scatter_u64offset_f64))) +void svstnt1_scatter_u64offset_f64(svbool_t, float64_t *, svuint64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_scatter_u64offset_s64))) +void svstnt1_scatter_u64offset_s64(svbool_t, int64_t *, svuint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1b_scatter_u32base_offset_u32))) +void svstnt1b_scatter_u32base_offset_u32(svbool_t, svuint32_t, int64_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1b_scatter_u64base_offset_u64))) +void svstnt1b_scatter_u64base_offset_u64(svbool_t, svuint64_t, int64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1b_scatter_u32base_offset_s32))) +void svstnt1b_scatter_u32base_offset_s32(svbool_t, svuint32_t, int64_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1b_scatter_u64base_offset_s64))) +void svstnt1b_scatter_u64base_offset_s64(svbool_t, svuint64_t, int64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1b_scatter_u32base_u32))) +void svstnt1b_scatter_u32base_u32(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1b_scatter_u64base_u64))) +void svstnt1b_scatter_u64base_u64(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1b_scatter_u32base_s32))) +void svstnt1b_scatter_u32base_s32(svbool_t, svuint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1b_scatter_u64base_s64))) +void svstnt1b_scatter_u64base_s64(svbool_t, svuint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1b_scatter_u32offset_s32))) +void svstnt1b_scatter_u32offset_s32(svbool_t, int8_t *, svuint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1b_scatter_u32offset_u32))) +void svstnt1b_scatter_u32offset_u32(svbool_t, uint8_t *, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1b_scatter_s64offset_s64))) +void svstnt1b_scatter_s64offset_s64(svbool_t, int8_t *, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1b_scatter_s64offset_u64))) +void svstnt1b_scatter_s64offset_u64(svbool_t, uint8_t *, svint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1b_scatter_u64offset_s64))) +void svstnt1b_scatter_u64offset_s64(svbool_t, int8_t *, svuint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1b_scatter_u64offset_u64))) +void svstnt1b_scatter_u64offset_u64(svbool_t, uint8_t *, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1h_scatter_u32base_index_u32))) +void svstnt1h_scatter_u32base_index_u32(svbool_t, svuint32_t, int64_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1h_scatter_u64base_index_u64))) +void svstnt1h_scatter_u64base_index_u64(svbool_t, svuint64_t, int64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1h_scatter_u32base_index_s32))) +void svstnt1h_scatter_u32base_index_s32(svbool_t, svuint32_t, int64_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1h_scatter_u64base_index_s64))) +void svstnt1h_scatter_u64base_index_s64(svbool_t, svuint64_t, int64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1h_scatter_u32base_offset_u32))) +void svstnt1h_scatter_u32base_offset_u32(svbool_t, svuint32_t, int64_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1h_scatter_u64base_offset_u64))) +void svstnt1h_scatter_u64base_offset_u64(svbool_t, svuint64_t, int64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1h_scatter_u32base_offset_s32))) +void svstnt1h_scatter_u32base_offset_s32(svbool_t, svuint32_t, int64_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1h_scatter_u64base_offset_s64))) +void svstnt1h_scatter_u64base_offset_s64(svbool_t, svuint64_t, int64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1h_scatter_u32base_u32))) +void svstnt1h_scatter_u32base_u32(svbool_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1h_scatter_u64base_u64))) +void svstnt1h_scatter_u64base_u64(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1h_scatter_u32base_s32))) +void svstnt1h_scatter_u32base_s32(svbool_t, svuint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1h_scatter_u64base_s64))) +void svstnt1h_scatter_u64base_s64(svbool_t, svuint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1h_scatter_s64index_s64))) +void svstnt1h_scatter_s64index_s64(svbool_t, int16_t *, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1h_scatter_s64index_u64))) +void svstnt1h_scatter_s64index_u64(svbool_t, uint16_t *, svint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1h_scatter_u64index_s64))) +void svstnt1h_scatter_u64index_s64(svbool_t, int16_t *, svuint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1h_scatter_u64index_u64))) +void svstnt1h_scatter_u64index_u64(svbool_t, uint16_t *, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1h_scatter_u32offset_s32))) +void svstnt1h_scatter_u32offset_s32(svbool_t, int16_t *, svuint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1h_scatter_u32offset_u32))) +void svstnt1h_scatter_u32offset_u32(svbool_t, uint16_t *, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1h_scatter_s64offset_s64))) +void svstnt1h_scatter_s64offset_s64(svbool_t, int16_t *, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1h_scatter_s64offset_u64))) +void svstnt1h_scatter_s64offset_u64(svbool_t, uint16_t *, svint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1h_scatter_u64offset_s64))) +void svstnt1h_scatter_u64offset_s64(svbool_t, int16_t *, svuint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1h_scatter_u64offset_u64))) +void svstnt1h_scatter_u64offset_u64(svbool_t, uint16_t *, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1w_scatter_u64base_index_u64))) +void svstnt1w_scatter_u64base_index_u64(svbool_t, svuint64_t, int64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1w_scatter_u64base_index_s64))) +void svstnt1w_scatter_u64base_index_s64(svbool_t, svuint64_t, int64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1w_scatter_u64base_offset_u64))) +void svstnt1w_scatter_u64base_offset_u64(svbool_t, svuint64_t, int64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1w_scatter_u64base_offset_s64))) +void svstnt1w_scatter_u64base_offset_s64(svbool_t, svuint64_t, int64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1w_scatter_u64base_u64))) +void svstnt1w_scatter_u64base_u64(svbool_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1w_scatter_u64base_s64))) +void svstnt1w_scatter_u64base_s64(svbool_t, svuint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1w_scatter_s64index_s64))) +void svstnt1w_scatter_s64index_s64(svbool_t, int32_t *, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1w_scatter_s64index_u64))) +void svstnt1w_scatter_s64index_u64(svbool_t, uint32_t *, svint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1w_scatter_u64index_s64))) +void svstnt1w_scatter_u64index_s64(svbool_t, int32_t *, svuint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1w_scatter_u64index_u64))) +void svstnt1w_scatter_u64index_u64(svbool_t, uint32_t *, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1w_scatter_s64offset_s64))) +void svstnt1w_scatter_s64offset_s64(svbool_t, int32_t *, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1w_scatter_s64offset_u64))) +void svstnt1w_scatter_s64offset_u64(svbool_t, uint32_t *, svint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1w_scatter_u64offset_s64))) +void svstnt1w_scatter_u64offset_s64(svbool_t, int32_t *, svuint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1w_scatter_u64offset_u64))) +void svstnt1w_scatter_u64offset_u64(svbool_t, uint32_t *, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubhnb_n_u32))) +svuint16_t svsubhnb_n_u32(svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubhnb_n_u64))) +svuint32_t svsubhnb_n_u64(svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubhnb_n_u16))) +svuint8_t svsubhnb_n_u16(svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubhnb_n_s32))) +svint16_t svsubhnb_n_s32(svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubhnb_n_s64))) +svint32_t svsubhnb_n_s64(svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubhnb_n_s16))) +svint8_t svsubhnb_n_s16(svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubhnb_u32))) +svuint16_t svsubhnb_u32(svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubhnb_u64))) +svuint32_t svsubhnb_u64(svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubhnb_u16))) +svuint8_t svsubhnb_u16(svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubhnb_s32))) +svint16_t svsubhnb_s32(svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubhnb_s64))) +svint32_t svsubhnb_s64(svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubhnb_s16))) +svint8_t svsubhnb_s16(svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubhnt_n_u32))) +svuint16_t svsubhnt_n_u32(svuint16_t, svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubhnt_n_u64))) +svuint32_t svsubhnt_n_u64(svuint32_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubhnt_n_u16))) +svuint8_t svsubhnt_n_u16(svuint8_t, svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubhnt_n_s32))) +svint16_t svsubhnt_n_s32(svint16_t, svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubhnt_n_s64))) +svint32_t svsubhnt_n_s64(svint32_t, svint64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubhnt_n_s16))) +svint8_t svsubhnt_n_s16(svint8_t, svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubhnt_u32))) +svuint16_t svsubhnt_u32(svuint16_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubhnt_u64))) +svuint32_t svsubhnt_u64(svuint32_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubhnt_u16))) +svuint8_t svsubhnt_u16(svuint8_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubhnt_s32))) +svint16_t svsubhnt_s32(svint16_t, svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubhnt_s64))) +svint32_t svsubhnt_s64(svint32_t, svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubhnt_s16))) +svint8_t svsubhnt_s16(svint8_t, svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsublb_n_s32))) +svint32_t svsublb_n_s32(svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsublb_n_s64))) +svint64_t svsublb_n_s64(svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsublb_n_s16))) +svint16_t svsublb_n_s16(svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsublb_n_u32))) +svuint32_t svsublb_n_u32(svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsublb_n_u64))) +svuint64_t svsublb_n_u64(svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsublb_n_u16))) +svuint16_t svsublb_n_u16(svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsublb_s32))) +svint32_t svsublb_s32(svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsublb_s64))) +svint64_t svsublb_s64(svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsublb_s16))) +svint16_t svsublb_s16(svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsublb_u32))) +svuint32_t svsublb_u32(svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsublb_u64))) +svuint64_t svsublb_u64(svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsublb_u16))) +svuint16_t svsublb_u16(svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsublbt_n_s32))) +svint32_t svsublbt_n_s32(svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsublbt_n_s64))) +svint64_t svsublbt_n_s64(svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsublbt_n_s16))) +svint16_t svsublbt_n_s16(svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsublbt_s32))) +svint32_t svsublbt_s32(svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsublbt_s64))) +svint64_t svsublbt_s64(svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsublbt_s16))) +svint16_t svsublbt_s16(svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsublt_n_s32))) +svint32_t svsublt_n_s32(svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsublt_n_s64))) +svint64_t svsublt_n_s64(svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsublt_n_s16))) +svint16_t svsublt_n_s16(svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsublt_n_u32))) +svuint32_t svsublt_n_u32(svuint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsublt_n_u64))) +svuint64_t svsublt_n_u64(svuint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsublt_n_u16))) +svuint16_t svsublt_n_u16(svuint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsublt_s32))) +svint32_t svsublt_s32(svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsublt_s64))) +svint64_t svsublt_s64(svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsublt_s16))) +svint16_t svsublt_s16(svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsublt_u32))) +svuint32_t svsublt_u32(svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsublt_u64))) +svuint64_t svsublt_u64(svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsublt_u16))) +svuint16_t svsublt_u16(svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubltb_n_s32))) +svint32_t svsubltb_n_s32(svint16_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubltb_n_s64))) +svint64_t svsubltb_n_s64(svint32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubltb_n_s16))) +svint16_t svsubltb_n_s16(svint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubltb_s32))) +svint32_t svsubltb_s32(svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubltb_s64))) +svint64_t svsubltb_s64(svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubltb_s16))) +svint16_t svsubltb_s16(svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubwb_n_s32))) +svint32_t svsubwb_n_s32(svint32_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubwb_n_s64))) +svint64_t svsubwb_n_s64(svint64_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubwb_n_s16))) +svint16_t svsubwb_n_s16(svint16_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubwb_n_u32))) +svuint32_t svsubwb_n_u32(svuint32_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubwb_n_u64))) +svuint64_t svsubwb_n_u64(svuint64_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubwb_n_u16))) +svuint16_t svsubwb_n_u16(svuint16_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubwb_s32))) +svint32_t svsubwb_s32(svint32_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubwb_s64))) +svint64_t svsubwb_s64(svint64_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubwb_s16))) +svint16_t svsubwb_s16(svint16_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubwb_u32))) +svuint32_t svsubwb_u32(svuint32_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubwb_u64))) +svuint64_t svsubwb_u64(svuint64_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubwb_u16))) +svuint16_t svsubwb_u16(svuint16_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubwt_n_s32))) +svint32_t svsubwt_n_s32(svint32_t, int16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubwt_n_s64))) +svint64_t svsubwt_n_s64(svint64_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubwt_n_s16))) +svint16_t svsubwt_n_s16(svint16_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubwt_n_u32))) +svuint32_t svsubwt_n_u32(svuint32_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubwt_n_u64))) +svuint64_t svsubwt_n_u64(svuint64_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubwt_n_u16))) +svuint16_t svsubwt_n_u16(svuint16_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubwt_s32))) +svint32_t svsubwt_s32(svint32_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubwt_s64))) +svint64_t svsubwt_s64(svint64_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubwt_s16))) +svint16_t svsubwt_s16(svint16_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubwt_u32))) +svuint32_t svsubwt_u32(svuint32_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubwt_u64))) +svuint64_t svsubwt_u64(svuint64_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsubwt_u16))) +svuint16_t svsubwt_u16(svuint16_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtbl2_u8))) +svuint8_t svtbl2_u8(svuint8x2_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtbl2_u32))) +svuint32_t svtbl2_u32(svuint32x2_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtbl2_u64))) +svuint64_t svtbl2_u64(svuint64x2_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtbl2_u16))) +svuint16_t svtbl2_u16(svuint16x2_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtbl2_s8))) +svint8_t svtbl2_s8(svint8x2_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtbl2_f64))) +svfloat64_t svtbl2_f64(svfloat64x2_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtbl2_f32))) +svfloat32_t svtbl2_f32(svfloat32x2_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtbl2_f16))) +svfloat16_t svtbl2_f16(svfloat16x2_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtbl2_s32))) +svint32_t svtbl2_s32(svint32x2_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtbl2_s64))) +svint64_t svtbl2_s64(svint64x2_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtbl2_s16))) +svint16_t svtbl2_s16(svint16x2_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtbx_u8))) +svuint8_t svtbx_u8(svuint8_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtbx_u32))) +svuint32_t svtbx_u32(svuint32_t, svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtbx_u64))) +svuint64_t svtbx_u64(svuint64_t, svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtbx_u16))) +svuint16_t svtbx_u16(svuint16_t, svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtbx_s8))) +svint8_t svtbx_s8(svint8_t, svint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtbx_f64))) +svfloat64_t svtbx_f64(svfloat64_t, svfloat64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtbx_f32))) +svfloat32_t svtbx_f32(svfloat32_t, svfloat32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtbx_f16))) +svfloat16_t svtbx_f16(svfloat16_t, svfloat16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtbx_s32))) +svint32_t svtbx_s32(svint32_t, svint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtbx_s64))) +svint64_t svtbx_s64(svint64_t, svint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtbx_s16))) +svint16_t svtbx_s16(svint16_t, svint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuqadd_n_s8_m))) +svint8_t svuqadd_n_s8_m(svbool_t, svint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuqadd_n_s32_m))) +svint32_t svuqadd_n_s32_m(svbool_t, svint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuqadd_n_s64_m))) +svint64_t svuqadd_n_s64_m(svbool_t, svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuqadd_n_s16_m))) +svint16_t svuqadd_n_s16_m(svbool_t, svint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuqadd_n_s8_x))) +svint8_t svuqadd_n_s8_x(svbool_t, svint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuqadd_n_s32_x))) +svint32_t svuqadd_n_s32_x(svbool_t, svint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuqadd_n_s64_x))) +svint64_t svuqadd_n_s64_x(svbool_t, svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuqadd_n_s16_x))) +svint16_t svuqadd_n_s16_x(svbool_t, svint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuqadd_n_s8_z))) +svint8_t svuqadd_n_s8_z(svbool_t, svint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuqadd_n_s32_z))) +svint32_t svuqadd_n_s32_z(svbool_t, svint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuqadd_n_s64_z))) +svint64_t svuqadd_n_s64_z(svbool_t, svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuqadd_n_s16_z))) +svint16_t svuqadd_n_s16_z(svbool_t, svint16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuqadd_s8_m))) +svint8_t svuqadd_s8_m(svbool_t, svint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuqadd_s32_m))) +svint32_t svuqadd_s32_m(svbool_t, svint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuqadd_s64_m))) +svint64_t svuqadd_s64_m(svbool_t, svint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuqadd_s16_m))) +svint16_t svuqadd_s16_m(svbool_t, svint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuqadd_s8_x))) +svint8_t svuqadd_s8_x(svbool_t, svint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuqadd_s32_x))) +svint32_t svuqadd_s32_x(svbool_t, svint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuqadd_s64_x))) +svint64_t svuqadd_s64_x(svbool_t, svint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuqadd_s16_x))) +svint16_t svuqadd_s16_x(svbool_t, svint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuqadd_s8_z))) +svint8_t svuqadd_s8_z(svbool_t, svint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuqadd_s32_z))) +svint32_t svuqadd_s32_z(svbool_t, svint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuqadd_s64_z))) +svint64_t svuqadd_s64_z(svbool_t, svint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuqadd_s16_z))) +svint16_t svuqadd_s16_z(svbool_t, svint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilege_b8_s32))) +svbool_t svwhilege_b8_s32(int32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilege_b32_s32))) +svbool_t svwhilege_b32_s32(int32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilege_b64_s32))) +svbool_t svwhilege_b64_s32(int32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilege_b16_s32))) +svbool_t svwhilege_b16_s32(int32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilege_b8_s64))) +svbool_t svwhilege_b8_s64(int64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilege_b32_s64))) +svbool_t svwhilege_b32_s64(int64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilege_b64_s64))) +svbool_t svwhilege_b64_s64(int64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilege_b16_s64))) +svbool_t svwhilege_b16_s64(int64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilege_b8_u32))) +svbool_t svwhilege_b8_u32(uint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilege_b32_u32))) +svbool_t svwhilege_b32_u32(uint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilege_b64_u32))) +svbool_t svwhilege_b64_u32(uint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilege_b16_u32))) +svbool_t svwhilege_b16_u32(uint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilege_b8_u64))) +svbool_t svwhilege_b8_u64(uint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilege_b32_u64))) +svbool_t svwhilege_b32_u64(uint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilege_b64_u64))) +svbool_t svwhilege_b64_u64(uint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilege_b16_u64))) +svbool_t svwhilege_b16_u64(uint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilegt_b8_s32))) +svbool_t svwhilegt_b8_s32(int32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilegt_b32_s32))) +svbool_t svwhilegt_b32_s32(int32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilegt_b64_s32))) +svbool_t svwhilegt_b64_s32(int32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilegt_b16_s32))) +svbool_t svwhilegt_b16_s32(int32_t, int32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilegt_b8_s64))) +svbool_t svwhilegt_b8_s64(int64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilegt_b32_s64))) +svbool_t svwhilegt_b32_s64(int64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilegt_b64_s64))) +svbool_t svwhilegt_b64_s64(int64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilegt_b16_s64))) +svbool_t svwhilegt_b16_s64(int64_t, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilegt_b8_u32))) +svbool_t svwhilegt_b8_u32(uint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilegt_b32_u32))) +svbool_t svwhilegt_b32_u32(uint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilegt_b64_u32))) +svbool_t svwhilegt_b64_u32(uint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilegt_b16_u32))) +svbool_t svwhilegt_b16_u32(uint32_t, uint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilegt_b8_u64))) +svbool_t svwhilegt_b8_u64(uint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilegt_b32_u64))) +svbool_t svwhilegt_b32_u64(uint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilegt_b64_u64))) +svbool_t svwhilegt_b64_u64(uint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilegt_b16_u64))) +svbool_t svwhilegt_b16_u64(uint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilerw_u8))) +svbool_t svwhilerw_u8(uint8_t const *, uint8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilerw_s8))) +svbool_t svwhilerw_s8(int8_t const *, int8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilerw_u64))) +svbool_t svwhilerw_u64(uint64_t const *, uint64_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilerw_f64))) +svbool_t svwhilerw_f64(float64_t const *, float64_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilerw_s64))) +svbool_t svwhilerw_s64(int64_t const *, int64_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilerw_u16))) +svbool_t svwhilerw_u16(uint16_t const *, uint16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilerw_f16))) +svbool_t svwhilerw_f16(float16_t const *, float16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilerw_s16))) +svbool_t svwhilerw_s16(int16_t const *, int16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilerw_u32))) +svbool_t svwhilerw_u32(uint32_t const *, uint32_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilerw_f32))) +svbool_t svwhilerw_f32(float32_t const *, float32_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilerw_s32))) +svbool_t svwhilerw_s32(int32_t const *, int32_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilewr_u8))) +svbool_t svwhilewr_u8(uint8_t const *, uint8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilewr_s8))) +svbool_t svwhilewr_s8(int8_t const *, int8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilewr_u64))) +svbool_t svwhilewr_u64(uint64_t const *, uint64_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilewr_f64))) +svbool_t svwhilewr_f64(float64_t const *, float64_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilewr_s64))) +svbool_t svwhilewr_s64(int64_t const *, int64_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilewr_u16))) +svbool_t svwhilewr_u16(uint16_t const *, uint16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilewr_f16))) +svbool_t svwhilewr_f16(float16_t const *, float16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilewr_s16))) +svbool_t svwhilewr_s16(int16_t const *, int16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilewr_u32))) +svbool_t svwhilewr_u32(uint32_t const *, uint32_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilewr_f32))) +svbool_t svwhilewr_f32(float32_t const *, float32_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilewr_s32))) +svbool_t svwhilewr_s32(int32_t const *, int32_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svxar_n_u8))) +svuint8_t svxar_n_u8(svuint8_t, svuint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svxar_n_u32))) +svuint32_t svxar_n_u32(svuint32_t, svuint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svxar_n_u64))) +svuint64_t svxar_n_u64(svuint64_t, svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svxar_n_u16))) +svuint16_t svxar_n_u16(svuint16_t, svuint16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svxar_n_s8))) +svint8_t svxar_n_s8(svint8_t, svint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svxar_n_s32))) +svint32_t svxar_n_s32(svint32_t, svint32_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svxar_n_s64))) +svint64_t svxar_n_s64(svint64_t, svint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svxar_n_s16))) +svint16_t svxar_n_s16(svint16_t, svint16_t, uint64_t); __aio __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaba_n_s8))) svint8_t svaba(svint8_t, svint8_t, int8_t); __aio __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaba_n_s32))) @@ -17527,8 +23230,10 @@ svint16_t svxar(svint16_t, svint16_t, uint64_t); #endif //defined(__ARM_FEATURE_SVE2) #if defined(__ARM_FEATURE_SVE2) && defined(__ARM_FEATURE_BF16_SCALAR_ARITHMETIC) -#define svwhilerw_bf16(...) __builtin_sve_svwhilerw_bf16(__VA_ARGS__) -#define svwhilewr_bf16(...) __builtin_sve_svwhilewr_bf16(__VA_ARGS__) +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilerw_bf16))) +svbool_t svwhilerw_bf16(bfloat16_t const *, bfloat16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilewr_bf16))) +svbool_t svwhilewr_bf16(bfloat16_t const *, bfloat16_t const *); __aio __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilerw_bf16))) svbool_t svwhilerw(bfloat16_t const *, bfloat16_t const *); __aio __attribute__((__clang_arm_builtin_alias(__builtin_sve_svwhilewr_bf16))) @@ -17536,8 +23241,10 @@ svbool_t svwhilewr(bfloat16_t const *, bfloat16_t const *); #endif //defined(__ARM_FEATURE_SVE2) && defined(__ARM_FEATURE_BF16_SCALAR_ARITHMETIC) #if defined(__ARM_FEATURE_SVE2) && defined(__ARM_FEATURE_SVE_BF16) -#define svtbl2_bf16(...) __builtin_sve_svtbl2_bf16(__VA_ARGS__) -#define svtbx_bf16(...) __builtin_sve_svtbx_bf16(__VA_ARGS__) +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtbl2_bf16))) +svbfloat16_t svtbl2_bf16(svbfloat16x2_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtbx_bf16))) +svbfloat16_t svtbx_bf16(svbfloat16_t, svbfloat16_t, svuint16_t); __aio __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtbl2_bf16))) svbfloat16_t svtbl2(svbfloat16x2_t, svuint16_t); __aio __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtbx_bf16))) @@ -17545,14 +23252,22 @@ svbfloat16_t svtbx(svbfloat16_t, svbfloat16_t, svuint16_t); #endif //defined(__ARM_FEATURE_SVE2) && defined(__ARM_FEATURE_SVE_BF16) #if defined(__ARM_FEATURE_SVE2_AES) -#define svaesd_u8(...) __builtin_sve_svaesd_u8(__VA_ARGS__) -#define svaese_u8(...) __builtin_sve_svaese_u8(__VA_ARGS__) -#define svaesimc_u8(...) __builtin_sve_svaesimc_u8(__VA_ARGS__) -#define svaesmc_u8(...) __builtin_sve_svaesmc_u8(__VA_ARGS__) -#define svpmullb_pair_n_u64(...) __builtin_sve_svpmullb_pair_n_u64(__VA_ARGS__) -#define svpmullb_pair_u64(...) __builtin_sve_svpmullb_pair_u64(__VA_ARGS__) -#define svpmullt_pair_n_u64(...) __builtin_sve_svpmullt_pair_n_u64(__VA_ARGS__) -#define svpmullt_pair_u64(...) __builtin_sve_svpmullt_pair_u64(__VA_ARGS__) +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaesd_u8))) +svuint8_t svaesd_u8(svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaese_u8))) +svuint8_t svaese_u8(svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaesimc_u8))) +svuint8_t svaesimc_u8(svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaesmc_u8))) +svuint8_t svaesmc_u8(svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svpmullb_pair_n_u64))) +svuint64_t svpmullb_pair_n_u64(svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svpmullb_pair_u64))) +svuint64_t svpmullb_pair_u64(svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svpmullt_pair_n_u64))) +svuint64_t svpmullt_pair_n_u64(svuint64_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svpmullt_pair_u64))) +svuint64_t svpmullt_pair_u64(svuint64_t, svuint64_t); __aio __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaesd_u8))) svuint8_t svaesd(svuint8_t, svuint8_t); __aio __attribute__((__clang_arm_builtin_alias(__builtin_sve_svaese_u8))) @@ -17572,8 +23287,10 @@ svuint64_t svpmullt_pair(svuint64_t, svuint64_t); #endif //defined(__ARM_FEATURE_SVE2_AES) #if defined(__ARM_FEATURE_SVE2_SHA3) -#define svrax1_u64(...) __builtin_sve_svrax1_u64(__VA_ARGS__) -#define svrax1_s64(...) __builtin_sve_svrax1_s64(__VA_ARGS__) +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrax1_u64))) +svuint64_t svrax1_u64(svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrax1_s64))) +svint64_t svrax1_s64(svint64_t, svint64_t); __aio __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrax1_u64))) svuint64_t svrax1(svuint64_t, svuint64_t); __aio __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrax1_s64))) @@ -17581,8 +23298,10 @@ svint64_t svrax1(svint64_t, svint64_t); #endif //defined(__ARM_FEATURE_SVE2_SHA3) #if defined(__ARM_FEATURE_SVE2_SM4) -#define svsm4e_u32(...) __builtin_sve_svsm4e_u32(__VA_ARGS__) -#define svsm4ekey_u32(...) __builtin_sve_svsm4ekey_u32(__VA_ARGS__) +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsm4e_u32))) +svuint32_t svsm4e_u32(svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsm4ekey_u32))) +svuint32_t svsm4ekey_u32(svuint32_t, svuint32_t); __aio __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsm4e_u32))) svuint32_t svsm4e(svuint32_t, svuint32_t); __aio __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsm4ekey_u32))) @@ -17590,87 +23309,168 @@ svuint32_t svsm4ekey(svuint32_t, svuint32_t); #endif //defined(__ARM_FEATURE_SVE2_SM4) #if defined(__ARM_FEATURE_SVE_BF16) -#define svbfdot_n_f32(...) __builtin_sve_svbfdot_n_f32(__VA_ARGS__) -#define svbfdot_f32(...) __builtin_sve_svbfdot_f32(__VA_ARGS__) -#define svbfdot_lane_f32(...) __builtin_sve_svbfdot_lane_f32(__VA_ARGS__) -#define svbfmlalb_n_f32(...) __builtin_sve_svbfmlalb_n_f32(__VA_ARGS__) -#define svbfmlalb_f32(...) __builtin_sve_svbfmlalb_f32(__VA_ARGS__) -#define svbfmlalb_lane_f32(...) __builtin_sve_svbfmlalb_lane_f32(__VA_ARGS__) -#define svbfmlalt_n_f32(...) __builtin_sve_svbfmlalt_n_f32(__VA_ARGS__) -#define svbfmlalt_f32(...) __builtin_sve_svbfmlalt_f32(__VA_ARGS__) -#define svbfmlalt_lane_f32(...) __builtin_sve_svbfmlalt_lane_f32(__VA_ARGS__) -#define svbfmmla_f32(...) __builtin_sve_svbfmmla_f32(__VA_ARGS__) -#define svclasta_n_bf16(...) __builtin_sve_svclasta_n_bf16(__VA_ARGS__) -#define svclasta_bf16(...) __builtin_sve_svclasta_bf16(__VA_ARGS__) -#define svclastb_n_bf16(...) __builtin_sve_svclastb_n_bf16(__VA_ARGS__) -#define svclastb_bf16(...) __builtin_sve_svclastb_bf16(__VA_ARGS__) -#define svcnt_bf16_m(...) __builtin_sve_svcnt_bf16_m(__VA_ARGS__) -#define svcnt_bf16_x(...) __builtin_sve_svcnt_bf16_x(__VA_ARGS__) -#define svcnt_bf16_z(...) __builtin_sve_svcnt_bf16_z(__VA_ARGS__) -#define svcreate2_bf16(...) __builtin_sve_svcreate2_bf16(__VA_ARGS__) -#define svcreate3_bf16(...) __builtin_sve_svcreate3_bf16(__VA_ARGS__) -#define svcreate4_bf16(...) __builtin_sve_svcreate4_bf16(__VA_ARGS__) -#define svcvt_bf16_f32_m(...) __builtin_sve_svcvt_bf16_f32_m(__VA_ARGS__) -#define svcvt_bf16_f32_x(...) __builtin_sve_svcvt_bf16_f32_x(__VA_ARGS__) -#define svcvt_bf16_f32_z(...) __builtin_sve_svcvt_bf16_f32_z(__VA_ARGS__) -#define svcvtnt_bf16_f32_m(...) __builtin_sve_svcvtnt_bf16_f32_m(__VA_ARGS__) -#define svdup_n_bf16(...) __builtin_sve_svdup_n_bf16(__VA_ARGS__) -#define svdup_n_bf16_m(...) __builtin_sve_svdup_n_bf16_m(__VA_ARGS__) -#define svdup_n_bf16_x(...) __builtin_sve_svdup_n_bf16_x(__VA_ARGS__) -#define svdup_n_bf16_z(...) __builtin_sve_svdup_n_bf16_z(__VA_ARGS__) -#define svdup_lane_bf16(...) __builtin_sve_svdup_lane_bf16(__VA_ARGS__) -#define svdupq_n_bf16(...) __builtin_sve_svdupq_n_bf16(__VA_ARGS__) -#define svdupq_lane_bf16(...) __builtin_sve_svdupq_lane_bf16(__VA_ARGS__) -#define svext_bf16(...) __builtin_sve_svext_bf16(__VA_ARGS__) -#define svget2_bf16(...) __builtin_sve_svget2_bf16(__VA_ARGS__) -#define svget3_bf16(...) __builtin_sve_svget3_bf16(__VA_ARGS__) -#define svget4_bf16(...) __builtin_sve_svget4_bf16(__VA_ARGS__) -#define svinsr_n_bf16(...) __builtin_sve_svinsr_n_bf16(__VA_ARGS__) -#define svlasta_bf16(...) __builtin_sve_svlasta_bf16(__VA_ARGS__) -#define svlastb_bf16(...) __builtin_sve_svlastb_bf16(__VA_ARGS__) -#define svld1_bf16(...) __builtin_sve_svld1_bf16(__VA_ARGS__) -#define svld1_vnum_bf16(...) __builtin_sve_svld1_vnum_bf16(__VA_ARGS__) -#define svld1rq_bf16(...) __builtin_sve_svld1rq_bf16(__VA_ARGS__) -#define svld2_bf16(...) __builtin_sve_svld2_bf16(__VA_ARGS__) -#define svld2_vnum_bf16(...) __builtin_sve_svld2_vnum_bf16(__VA_ARGS__) -#define svld3_bf16(...) __builtin_sve_svld3_bf16(__VA_ARGS__) -#define svld3_vnum_bf16(...) __builtin_sve_svld3_vnum_bf16(__VA_ARGS__) -#define svld4_bf16(...) __builtin_sve_svld4_bf16(__VA_ARGS__) -#define svld4_vnum_bf16(...) __builtin_sve_svld4_vnum_bf16(__VA_ARGS__) -#define svldff1_bf16(...) __builtin_sve_svldff1_bf16(__VA_ARGS__) -#define svldff1_vnum_bf16(...) __builtin_sve_svldff1_vnum_bf16(__VA_ARGS__) -#define svldnf1_bf16(...) __builtin_sve_svldnf1_bf16(__VA_ARGS__) -#define svldnf1_vnum_bf16(...) __builtin_sve_svldnf1_vnum_bf16(__VA_ARGS__) -#define svldnt1_bf16(...) __builtin_sve_svldnt1_bf16(__VA_ARGS__) -#define svldnt1_vnum_bf16(...) __builtin_sve_svldnt1_vnum_bf16(__VA_ARGS__) -#define svlen_bf16(...) __builtin_sve_svlen_bf16(__VA_ARGS__) -#define svrev_bf16(...) __builtin_sve_svrev_bf16(__VA_ARGS__) -#define svsel_bf16(...) __builtin_sve_svsel_bf16(__VA_ARGS__) -#define svset2_bf16(...) __builtin_sve_svset2_bf16(__VA_ARGS__) -#define svset3_bf16(...) __builtin_sve_svset3_bf16(__VA_ARGS__) -#define svset4_bf16(...) __builtin_sve_svset4_bf16(__VA_ARGS__) -#define svsplice_bf16(...) __builtin_sve_svsplice_bf16(__VA_ARGS__) -#define svst1_bf16(...) __builtin_sve_svst1_bf16(__VA_ARGS__) -#define svst1_vnum_bf16(...) __builtin_sve_svst1_vnum_bf16(__VA_ARGS__) -#define svst2_bf16(...) __builtin_sve_svst2_bf16(__VA_ARGS__) -#define svst2_vnum_bf16(...) __builtin_sve_svst2_vnum_bf16(__VA_ARGS__) -#define svst3_bf16(...) __builtin_sve_svst3_bf16(__VA_ARGS__) -#define svst3_vnum_bf16(...) __builtin_sve_svst3_vnum_bf16(__VA_ARGS__) -#define svst4_bf16(...) __builtin_sve_svst4_bf16(__VA_ARGS__) -#define svst4_vnum_bf16(...) __builtin_sve_svst4_vnum_bf16(__VA_ARGS__) -#define svstnt1_bf16(...) __builtin_sve_svstnt1_bf16(__VA_ARGS__) -#define svstnt1_vnum_bf16(...) __builtin_sve_svstnt1_vnum_bf16(__VA_ARGS__) -#define svtbl_bf16(...) __builtin_sve_svtbl_bf16(__VA_ARGS__) -#define svtrn1_bf16(...) __builtin_sve_svtrn1_bf16(__VA_ARGS__) -#define svtrn2_bf16(...) __builtin_sve_svtrn2_bf16(__VA_ARGS__) -#define svundef2_bf16(...) __builtin_sve_svundef2_bf16(__VA_ARGS__) -#define svundef3_bf16(...) __builtin_sve_svundef3_bf16(__VA_ARGS__) -#define svundef4_bf16(...) __builtin_sve_svundef4_bf16(__VA_ARGS__) -#define svundef_bf16(...) __builtin_sve_svundef_bf16(__VA_ARGS__) -#define svuzp1_bf16(...) __builtin_sve_svuzp1_bf16(__VA_ARGS__) -#define svuzp2_bf16(...) __builtin_sve_svuzp2_bf16(__VA_ARGS__) -#define svzip1_bf16(...) __builtin_sve_svzip1_bf16(__VA_ARGS__) -#define svzip2_bf16(...) __builtin_sve_svzip2_bf16(__VA_ARGS__) +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbfdot_n_f32))) +svfloat32_t svbfdot_n_f32(svfloat32_t, svbfloat16_t, bfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbfdot_f32))) +svfloat32_t svbfdot_f32(svfloat32_t, svbfloat16_t, svbfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbfdot_lane_f32))) +svfloat32_t svbfdot_lane_f32(svfloat32_t, svbfloat16_t, svbfloat16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbfmlalb_n_f32))) +svfloat32_t svbfmlalb_n_f32(svfloat32_t, svbfloat16_t, bfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbfmlalb_f32))) +svfloat32_t svbfmlalb_f32(svfloat32_t, svbfloat16_t, svbfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbfmlalb_lane_f32))) +svfloat32_t svbfmlalb_lane_f32(svfloat32_t, svbfloat16_t, svbfloat16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbfmlalt_n_f32))) +svfloat32_t svbfmlalt_n_f32(svfloat32_t, svbfloat16_t, bfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbfmlalt_f32))) +svfloat32_t svbfmlalt_f32(svfloat32_t, svbfloat16_t, svbfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbfmlalt_lane_f32))) +svfloat32_t svbfmlalt_lane_f32(svfloat32_t, svbfloat16_t, svbfloat16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbfmmla_f32))) +svfloat32_t svbfmmla_f32(svfloat32_t, svbfloat16_t, svbfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclasta_n_bf16))) +bfloat16_t svclasta_n_bf16(svbool_t, bfloat16_t, svbfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclasta_bf16))) +svbfloat16_t svclasta_bf16(svbool_t, svbfloat16_t, svbfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclastb_n_bf16))) +bfloat16_t svclastb_n_bf16(svbool_t, bfloat16_t, svbfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svclastb_bf16))) +svbfloat16_t svclastb_bf16(svbool_t, svbfloat16_t, svbfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnt_bf16_m))) +svuint16_t svcnt_bf16_m(svuint16_t, svbool_t, svbfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnt_bf16_x))) +svuint16_t svcnt_bf16_x(svbool_t, svbfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcnt_bf16_z))) +svuint16_t svcnt_bf16_z(svbool_t, svbfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcreate2_bf16))) +svbfloat16x2_t svcreate2_bf16(svbfloat16_t, svbfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcreate3_bf16))) +svbfloat16x3_t svcreate3_bf16(svbfloat16_t, svbfloat16_t, svbfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcreate4_bf16))) +svbfloat16x4_t svcreate4_bf16(svbfloat16_t, svbfloat16_t, svbfloat16_t, svbfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_bf16_f32_m))) +svbfloat16_t svcvt_bf16_f32_m(svbfloat16_t, svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_bf16_f32_x))) +svbfloat16_t svcvt_bf16_f32_x(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvt_bf16_f32_z))) +svbfloat16_t svcvt_bf16_f32_z(svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svcvtnt_bf16_f32_m))) +svbfloat16_t svcvtnt_bf16_f32_m(svbfloat16_t, svbool_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_bf16))) +svbfloat16_t svdup_n_bf16(bfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_bf16_m))) +svbfloat16_t svdup_n_bf16_m(svbfloat16_t, svbool_t, bfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_bf16_x))) +svbfloat16_t svdup_n_bf16_x(svbool_t, bfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_n_bf16_z))) +svbfloat16_t svdup_n_bf16_z(svbool_t, bfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdup_lane_bf16))) +svbfloat16_t svdup_lane_bf16(svbfloat16_t, uint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdupq_n_bf16))) +svbfloat16_t svdupq_n_bf16(bfloat16_t, bfloat16_t, bfloat16_t, bfloat16_t, bfloat16_t, bfloat16_t, bfloat16_t, bfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svdupq_lane_bf16))) +svbfloat16_t svdupq_lane_bf16(svbfloat16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svext_bf16))) +svbfloat16_t svext_bf16(svbfloat16_t, svbfloat16_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svget2_bf16))) +svbfloat16_t svget2_bf16(svbfloat16x2_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svget3_bf16))) +svbfloat16_t svget3_bf16(svbfloat16x3_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svget4_bf16))) +svbfloat16_t svget4_bf16(svbfloat16x4_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svinsr_n_bf16))) +svbfloat16_t svinsr_n_bf16(svbfloat16_t, bfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlasta_bf16))) +bfloat16_t svlasta_bf16(svbool_t, svbfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlastb_bf16))) +bfloat16_t svlastb_bf16(svbool_t, svbfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_bf16))) +svbfloat16_t svld1_bf16(svbool_t, bfloat16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1_vnum_bf16))) +svbfloat16_t svld1_vnum_bf16(svbool_t, bfloat16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1rq_bf16))) +svbfloat16_t svld1rq_bf16(svbool_t, bfloat16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld2_bf16))) +svbfloat16x2_t svld2_bf16(svbool_t, bfloat16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld2_vnum_bf16))) +svbfloat16x2_t svld2_vnum_bf16(svbool_t, bfloat16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld3_bf16))) +svbfloat16x3_t svld3_bf16(svbool_t, bfloat16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld3_vnum_bf16))) +svbfloat16x3_t svld3_vnum_bf16(svbool_t, bfloat16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld4_bf16))) +svbfloat16x4_t svld4_bf16(svbool_t, bfloat16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld4_vnum_bf16))) +svbfloat16x4_t svld4_vnum_bf16(svbool_t, bfloat16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_bf16))) +svbfloat16_t svldff1_bf16(svbool_t, bfloat16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldff1_vnum_bf16))) +svbfloat16_t svldff1_vnum_bf16(svbool_t, bfloat16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1_bf16))) +svbfloat16_t svldnf1_bf16(svbool_t, bfloat16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnf1_vnum_bf16))) +svbfloat16_t svldnf1_vnum_bf16(svbool_t, bfloat16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_bf16))) +svbfloat16_t svldnt1_bf16(svbool_t, bfloat16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svldnt1_vnum_bf16))) +svbfloat16_t svldnt1_vnum_bf16(svbool_t, bfloat16_t const *, int64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svlen_bf16))) +uint64_t svlen_bf16(svbfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svrev_bf16))) +svbfloat16_t svrev_bf16(svbfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsel_bf16))) +svbfloat16_t svsel_bf16(svbool_t, svbfloat16_t, svbfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svset2_bf16))) +svbfloat16x2_t svset2_bf16(svbfloat16x2_t, uint64_t, svbfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svset3_bf16))) +svbfloat16x3_t svset3_bf16(svbfloat16x3_t, uint64_t, svbfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svset4_bf16))) +svbfloat16x4_t svset4_bf16(svbfloat16x4_t, uint64_t, svbfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsplice_bf16))) +svbfloat16_t svsplice_bf16(svbool_t, svbfloat16_t, svbfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_bf16))) +void svst1_bf16(svbool_t, bfloat16_t *, svbfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst1_vnum_bf16))) +void svst1_vnum_bf16(svbool_t, bfloat16_t *, int64_t, svbfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst2_bf16))) +void svst2_bf16(svbool_t, bfloat16_t *, svbfloat16x2_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst2_vnum_bf16))) +void svst2_vnum_bf16(svbool_t, bfloat16_t *, int64_t, svbfloat16x2_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst3_bf16))) +void svst3_bf16(svbool_t, bfloat16_t *, svbfloat16x3_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst3_vnum_bf16))) +void svst3_vnum_bf16(svbool_t, bfloat16_t *, int64_t, svbfloat16x3_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst4_bf16))) +void svst4_bf16(svbool_t, bfloat16_t *, svbfloat16x4_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svst4_vnum_bf16))) +void svst4_vnum_bf16(svbool_t, bfloat16_t *, int64_t, svbfloat16x4_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_bf16))) +void svstnt1_bf16(svbool_t, bfloat16_t *, svbfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svstnt1_vnum_bf16))) +void svstnt1_vnum_bf16(svbool_t, bfloat16_t *, int64_t, svbfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtbl_bf16))) +svbfloat16_t svtbl_bf16(svbfloat16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn1_bf16))) +svbfloat16_t svtrn1_bf16(svbfloat16_t, svbfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn2_bf16))) +svbfloat16_t svtrn2_bf16(svbfloat16_t, svbfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef2_bf16))) +svbfloat16x2_t svundef2_bf16(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef3_bf16))) +svbfloat16x3_t svundef3_bf16(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef4_bf16))) +svbfloat16x4_t svundef4_bf16(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svundef_bf16))) +svbfloat16_t svundef_bf16(); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp1_bf16))) +svbfloat16_t svuzp1_bf16(svbfloat16_t, svbfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp2_bf16))) +svbfloat16_t svuzp2_bf16(svbfloat16_t, svbfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip1_bf16))) +svbfloat16_t svzip1_bf16(svbfloat16_t, svbfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip2_bf16))) +svbfloat16_t svzip2_bf16(svbfloat16_t, svbfloat16_t); __aio __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbfdot_n_f32))) svfloat32_t svbfdot(svfloat32_t, svbfloat16_t, bfloat16_t); __aio __attribute__((__clang_arm_builtin_alias(__builtin_sve_svbfdot_f32))) @@ -17828,90 +23628,169 @@ svbfloat16_t svzip2(svbfloat16_t, svbfloat16_t); #endif //defined(__ARM_FEATURE_SVE_BF16) #if defined(__ARM_FEATURE_SVE_MATMUL_FP32) -#define svmmla_f32(...) __builtin_sve_svmmla_f32(__VA_ARGS__) +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmmla_f32))) +svfloat32_t svmmla_f32(svfloat32_t, svfloat32_t, svfloat32_t); __aio __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmmla_f32))) svfloat32_t svmmla(svfloat32_t, svfloat32_t, svfloat32_t); #endif //defined(__ARM_FEATURE_SVE_MATMUL_FP32) #if defined(__ARM_FEATURE_SVE_MATMUL_FP64) -#define svld1ro_u8(...) __builtin_sve_svld1ro_u8(__VA_ARGS__) -#define svld1ro_u32(...) __builtin_sve_svld1ro_u32(__VA_ARGS__) -#define svld1ro_u64(...) __builtin_sve_svld1ro_u64(__VA_ARGS__) -#define svld1ro_u16(...) __builtin_sve_svld1ro_u16(__VA_ARGS__) -#define svld1ro_s8(...) __builtin_sve_svld1ro_s8(__VA_ARGS__) -#define svld1ro_f64(...) __builtin_sve_svld1ro_f64(__VA_ARGS__) -#define svld1ro_f32(...) __builtin_sve_svld1ro_f32(__VA_ARGS__) -#define svld1ro_f16(...) __builtin_sve_svld1ro_f16(__VA_ARGS__) -#define svld1ro_s32(...) __builtin_sve_svld1ro_s32(__VA_ARGS__) -#define svld1ro_s64(...) __builtin_sve_svld1ro_s64(__VA_ARGS__) -#define svld1ro_s16(...) __builtin_sve_svld1ro_s16(__VA_ARGS__) -#define svmmla_f64(...) __builtin_sve_svmmla_f64(__VA_ARGS__) -#define svtrn1q_u8(...) __builtin_sve_svtrn1q_u8(__VA_ARGS__) -#define svtrn1q_u32(...) __builtin_sve_svtrn1q_u32(__VA_ARGS__) -#define svtrn1q_u64(...) __builtin_sve_svtrn1q_u64(__VA_ARGS__) -#define svtrn1q_u16(...) __builtin_sve_svtrn1q_u16(__VA_ARGS__) -#define svtrn1q_s8(...) __builtin_sve_svtrn1q_s8(__VA_ARGS__) -#define svtrn1q_f64(...) __builtin_sve_svtrn1q_f64(__VA_ARGS__) -#define svtrn1q_f32(...) __builtin_sve_svtrn1q_f32(__VA_ARGS__) -#define svtrn1q_f16(...) __builtin_sve_svtrn1q_f16(__VA_ARGS__) -#define svtrn1q_s32(...) __builtin_sve_svtrn1q_s32(__VA_ARGS__) -#define svtrn1q_s64(...) __builtin_sve_svtrn1q_s64(__VA_ARGS__) -#define svtrn1q_s16(...) __builtin_sve_svtrn1q_s16(__VA_ARGS__) -#define svtrn2q_u8(...) __builtin_sve_svtrn2q_u8(__VA_ARGS__) -#define svtrn2q_u32(...) __builtin_sve_svtrn2q_u32(__VA_ARGS__) -#define svtrn2q_u64(...) __builtin_sve_svtrn2q_u64(__VA_ARGS__) -#define svtrn2q_u16(...) __builtin_sve_svtrn2q_u16(__VA_ARGS__) -#define svtrn2q_s8(...) __builtin_sve_svtrn2q_s8(__VA_ARGS__) -#define svtrn2q_f64(...) __builtin_sve_svtrn2q_f64(__VA_ARGS__) -#define svtrn2q_f32(...) __builtin_sve_svtrn2q_f32(__VA_ARGS__) -#define svtrn2q_f16(...) __builtin_sve_svtrn2q_f16(__VA_ARGS__) -#define svtrn2q_s32(...) __builtin_sve_svtrn2q_s32(__VA_ARGS__) -#define svtrn2q_s64(...) __builtin_sve_svtrn2q_s64(__VA_ARGS__) -#define svtrn2q_s16(...) __builtin_sve_svtrn2q_s16(__VA_ARGS__) -#define svuzp1q_u8(...) __builtin_sve_svuzp1q_u8(__VA_ARGS__) -#define svuzp1q_u32(...) __builtin_sve_svuzp1q_u32(__VA_ARGS__) -#define svuzp1q_u64(...) __builtin_sve_svuzp1q_u64(__VA_ARGS__) -#define svuzp1q_u16(...) __builtin_sve_svuzp1q_u16(__VA_ARGS__) -#define svuzp1q_s8(...) __builtin_sve_svuzp1q_s8(__VA_ARGS__) -#define svuzp1q_f64(...) __builtin_sve_svuzp1q_f64(__VA_ARGS__) -#define svuzp1q_f32(...) __builtin_sve_svuzp1q_f32(__VA_ARGS__) -#define svuzp1q_f16(...) __builtin_sve_svuzp1q_f16(__VA_ARGS__) -#define svuzp1q_s32(...) __builtin_sve_svuzp1q_s32(__VA_ARGS__) -#define svuzp1q_s64(...) __builtin_sve_svuzp1q_s64(__VA_ARGS__) -#define svuzp1q_s16(...) __builtin_sve_svuzp1q_s16(__VA_ARGS__) -#define svuzp2q_u8(...) __builtin_sve_svuzp2q_u8(__VA_ARGS__) -#define svuzp2q_u32(...) __builtin_sve_svuzp2q_u32(__VA_ARGS__) -#define svuzp2q_u64(...) __builtin_sve_svuzp2q_u64(__VA_ARGS__) -#define svuzp2q_u16(...) __builtin_sve_svuzp2q_u16(__VA_ARGS__) -#define svuzp2q_s8(...) __builtin_sve_svuzp2q_s8(__VA_ARGS__) -#define svuzp2q_f64(...) __builtin_sve_svuzp2q_f64(__VA_ARGS__) -#define svuzp2q_f32(...) __builtin_sve_svuzp2q_f32(__VA_ARGS__) -#define svuzp2q_f16(...) __builtin_sve_svuzp2q_f16(__VA_ARGS__) -#define svuzp2q_s32(...) __builtin_sve_svuzp2q_s32(__VA_ARGS__) -#define svuzp2q_s64(...) __builtin_sve_svuzp2q_s64(__VA_ARGS__) -#define svuzp2q_s16(...) __builtin_sve_svuzp2q_s16(__VA_ARGS__) -#define svzip1q_u8(...) __builtin_sve_svzip1q_u8(__VA_ARGS__) -#define svzip1q_u32(...) __builtin_sve_svzip1q_u32(__VA_ARGS__) -#define svzip1q_u64(...) __builtin_sve_svzip1q_u64(__VA_ARGS__) -#define svzip1q_u16(...) __builtin_sve_svzip1q_u16(__VA_ARGS__) -#define svzip1q_s8(...) __builtin_sve_svzip1q_s8(__VA_ARGS__) -#define svzip1q_f64(...) __builtin_sve_svzip1q_f64(__VA_ARGS__) -#define svzip1q_f32(...) __builtin_sve_svzip1q_f32(__VA_ARGS__) -#define svzip1q_f16(...) __builtin_sve_svzip1q_f16(__VA_ARGS__) -#define svzip1q_s32(...) __builtin_sve_svzip1q_s32(__VA_ARGS__) -#define svzip1q_s64(...) __builtin_sve_svzip1q_s64(__VA_ARGS__) -#define svzip1q_s16(...) __builtin_sve_svzip1q_s16(__VA_ARGS__) -#define svzip2q_u8(...) __builtin_sve_svzip2q_u8(__VA_ARGS__) -#define svzip2q_u32(...) __builtin_sve_svzip2q_u32(__VA_ARGS__) -#define svzip2q_u64(...) __builtin_sve_svzip2q_u64(__VA_ARGS__) -#define svzip2q_u16(...) __builtin_sve_svzip2q_u16(__VA_ARGS__) -#define svzip2q_s8(...) __builtin_sve_svzip2q_s8(__VA_ARGS__) -#define svzip2q_f64(...) __builtin_sve_svzip2q_f64(__VA_ARGS__) -#define svzip2q_f32(...) __builtin_sve_svzip2q_f32(__VA_ARGS__) -#define svzip2q_f16(...) __builtin_sve_svzip2q_f16(__VA_ARGS__) -#define svzip2q_s32(...) __builtin_sve_svzip2q_s32(__VA_ARGS__) -#define svzip2q_s64(...) __builtin_sve_svzip2q_s64(__VA_ARGS__) -#define svzip2q_s16(...) __builtin_sve_svzip2q_s16(__VA_ARGS__) +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1ro_u8))) +svuint8_t svld1ro_u8(svbool_t, uint8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1ro_u32))) +svuint32_t svld1ro_u32(svbool_t, uint32_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1ro_u64))) +svuint64_t svld1ro_u64(svbool_t, uint64_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1ro_u16))) +svuint16_t svld1ro_u16(svbool_t, uint16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1ro_s8))) +svint8_t svld1ro_s8(svbool_t, int8_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1ro_f64))) +svfloat64_t svld1ro_f64(svbool_t, float64_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1ro_f32))) +svfloat32_t svld1ro_f32(svbool_t, float32_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1ro_f16))) +svfloat16_t svld1ro_f16(svbool_t, float16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1ro_s32))) +svint32_t svld1ro_s32(svbool_t, int32_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1ro_s64))) +svint64_t svld1ro_s64(svbool_t, int64_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1ro_s16))) +svint16_t svld1ro_s16(svbool_t, int16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmmla_f64))) +svfloat64_t svmmla_f64(svfloat64_t, svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn1q_u8))) +svuint8_t svtrn1q_u8(svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn1q_u32))) +svuint32_t svtrn1q_u32(svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn1q_u64))) +svuint64_t svtrn1q_u64(svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn1q_u16))) +svuint16_t svtrn1q_u16(svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn1q_s8))) +svint8_t svtrn1q_s8(svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn1q_f64))) +svfloat64_t svtrn1q_f64(svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn1q_f32))) +svfloat32_t svtrn1q_f32(svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn1q_f16))) +svfloat16_t svtrn1q_f16(svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn1q_s32))) +svint32_t svtrn1q_s32(svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn1q_s64))) +svint64_t svtrn1q_s64(svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn1q_s16))) +svint16_t svtrn1q_s16(svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn2q_u8))) +svuint8_t svtrn2q_u8(svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn2q_u32))) +svuint32_t svtrn2q_u32(svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn2q_u64))) +svuint64_t svtrn2q_u64(svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn2q_u16))) +svuint16_t svtrn2q_u16(svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn2q_s8))) +svint8_t svtrn2q_s8(svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn2q_f64))) +svfloat64_t svtrn2q_f64(svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn2q_f32))) +svfloat32_t svtrn2q_f32(svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn2q_f16))) +svfloat16_t svtrn2q_f16(svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn2q_s32))) +svint32_t svtrn2q_s32(svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn2q_s64))) +svint64_t svtrn2q_s64(svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn2q_s16))) +svint16_t svtrn2q_s16(svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp1q_u8))) +svuint8_t svuzp1q_u8(svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp1q_u32))) +svuint32_t svuzp1q_u32(svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp1q_u64))) +svuint64_t svuzp1q_u64(svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp1q_u16))) +svuint16_t svuzp1q_u16(svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp1q_s8))) +svint8_t svuzp1q_s8(svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp1q_f64))) +svfloat64_t svuzp1q_f64(svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp1q_f32))) +svfloat32_t svuzp1q_f32(svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp1q_f16))) +svfloat16_t svuzp1q_f16(svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp1q_s32))) +svint32_t svuzp1q_s32(svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp1q_s64))) +svint64_t svuzp1q_s64(svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp1q_s16))) +svint16_t svuzp1q_s16(svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp2q_u8))) +svuint8_t svuzp2q_u8(svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp2q_u32))) +svuint32_t svuzp2q_u32(svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp2q_u64))) +svuint64_t svuzp2q_u64(svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp2q_u16))) +svuint16_t svuzp2q_u16(svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp2q_s8))) +svint8_t svuzp2q_s8(svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp2q_f64))) +svfloat64_t svuzp2q_f64(svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp2q_f32))) +svfloat32_t svuzp2q_f32(svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp2q_f16))) +svfloat16_t svuzp2q_f16(svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp2q_s32))) +svint32_t svuzp2q_s32(svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp2q_s64))) +svint64_t svuzp2q_s64(svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp2q_s16))) +svint16_t svuzp2q_s16(svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip1q_u8))) +svuint8_t svzip1q_u8(svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip1q_u32))) +svuint32_t svzip1q_u32(svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip1q_u64))) +svuint64_t svzip1q_u64(svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip1q_u16))) +svuint16_t svzip1q_u16(svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip1q_s8))) +svint8_t svzip1q_s8(svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip1q_f64))) +svfloat64_t svzip1q_f64(svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip1q_f32))) +svfloat32_t svzip1q_f32(svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip1q_f16))) +svfloat16_t svzip1q_f16(svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip1q_s32))) +svint32_t svzip1q_s32(svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip1q_s64))) +svint64_t svzip1q_s64(svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip1q_s16))) +svint16_t svzip1q_s16(svint16_t, svint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip2q_u8))) +svuint8_t svzip2q_u8(svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip2q_u32))) +svuint32_t svzip2q_u32(svuint32_t, svuint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip2q_u64))) +svuint64_t svzip2q_u64(svuint64_t, svuint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip2q_u16))) +svuint16_t svzip2q_u16(svuint16_t, svuint16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip2q_s8))) +svint8_t svzip2q_s8(svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip2q_f64))) +svfloat64_t svzip2q_f64(svfloat64_t, svfloat64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip2q_f32))) +svfloat32_t svzip2q_f32(svfloat32_t, svfloat32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip2q_f16))) +svfloat16_t svzip2q_f16(svfloat16_t, svfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip2q_s32))) +svint32_t svzip2q_s32(svint32_t, svint32_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip2q_s64))) +svint64_t svzip2q_s64(svint64_t, svint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip2q_s16))) +svint16_t svzip2q_s16(svint16_t, svint16_t); __aio __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1ro_u8))) svuint8_t svld1ro(svbool_t, uint8_t const *); __aio __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1ro_u32))) @@ -18071,13 +23950,20 @@ svint16_t svzip2q(svint16_t, svint16_t); #endif //defined(__ARM_FEATURE_SVE_MATMUL_FP64) #if defined(__ARM_FEATURE_SVE_MATMUL_FP64) && defined(__ARM_FEATURE_SVE_BF16) -#define svld1ro_bf16(...) __builtin_sve_svld1ro_bf16(__VA_ARGS__) -#define svtrn1q_bf16(...) __builtin_sve_svtrn1q_bf16(__VA_ARGS__) -#define svtrn2q_bf16(...) __builtin_sve_svtrn2q_bf16(__VA_ARGS__) -#define svuzp1q_bf16(...) __builtin_sve_svuzp1q_bf16(__VA_ARGS__) -#define svuzp2q_bf16(...) __builtin_sve_svuzp2q_bf16(__VA_ARGS__) -#define svzip1q_bf16(...) __builtin_sve_svzip1q_bf16(__VA_ARGS__) -#define svzip2q_bf16(...) __builtin_sve_svzip2q_bf16(__VA_ARGS__) +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1ro_bf16))) +svbfloat16_t svld1ro_bf16(svbool_t, bfloat16_t const *); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn1q_bf16))) +svbfloat16_t svtrn1q_bf16(svbfloat16_t, svbfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn2q_bf16))) +svbfloat16_t svtrn2q_bf16(svbfloat16_t, svbfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp1q_bf16))) +svbfloat16_t svuzp1q_bf16(svbfloat16_t, svbfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svuzp2q_bf16))) +svbfloat16_t svuzp2q_bf16(svbfloat16_t, svbfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip1q_bf16))) +svbfloat16_t svzip1q_bf16(svbfloat16_t, svbfloat16_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svzip2q_bf16))) +svbfloat16_t svzip2q_bf16(svbfloat16_t, svbfloat16_t); __aio __attribute__((__clang_arm_builtin_alias(__builtin_sve_svld1ro_bf16))) svbfloat16_t svld1ro(svbool_t, bfloat16_t const *); __aio __attribute__((__clang_arm_builtin_alias(__builtin_sve_svtrn1q_bf16))) @@ -18095,15 +23981,24 @@ svbfloat16_t svzip2q(svbfloat16_t, svbfloat16_t); #endif //defined(__ARM_FEATURE_SVE_MATMUL_FP64) && defined(__ARM_FEATURE_SVE_BF16) #if defined(__ARM_FEATURE_SVE_MATMUL_INT8) -#define svmmla_s32(...) __builtin_sve_svmmla_s32(__VA_ARGS__) -#define svmmla_u32(...) __builtin_sve_svmmla_u32(__VA_ARGS__) -#define svsudot_n_s32(...) __builtin_sve_svsudot_n_s32(__VA_ARGS__) -#define svsudot_s32(...) __builtin_sve_svsudot_s32(__VA_ARGS__) -#define svsudot_lane_s32(...) __builtin_sve_svsudot_lane_s32(__VA_ARGS__) -#define svusdot_n_s32(...) __builtin_sve_svusdot_n_s32(__VA_ARGS__) -#define svusdot_s32(...) __builtin_sve_svusdot_s32(__VA_ARGS__) -#define svusdot_lane_s32(...) __builtin_sve_svusdot_lane_s32(__VA_ARGS__) -#define svusmmla_s32(...) __builtin_sve_svusmmla_s32(__VA_ARGS__) +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmmla_s32))) +svint32_t svmmla_s32(svint32_t, svint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmmla_u32))) +svuint32_t svmmla_u32(svuint32_t, svuint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsudot_n_s32))) +svint32_t svsudot_n_s32(svint32_t, svint8_t, uint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsudot_s32))) +svint32_t svsudot_s32(svint32_t, svint8_t, svuint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svsudot_lane_s32))) +svint32_t svsudot_lane_s32(svint32_t, svint8_t, svuint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svusdot_n_s32))) +svint32_t svusdot_n_s32(svint32_t, svuint8_t, int8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svusdot_s32))) +svint32_t svusdot_s32(svint32_t, svuint8_t, svint8_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svusdot_lane_s32))) +svint32_t svusdot_lane_s32(svint32_t, svuint8_t, svint8_t, uint64_t); +__ai __attribute__((__clang_arm_builtin_alias(__builtin_sve_svusmmla_s32))) +svint32_t svusmmla_s32(svint32_t, svuint8_t, svint8_t); __aio __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmmla_s32))) svint32_t svmmla(svint32_t, svint8_t, svint8_t); __aio __attribute__((__clang_arm_builtin_alias(__builtin_sve_svmmla_u32))) diff --git a/lib/include/avx512fintrin.h b/lib/include/avx512fintrin.h index f226382cb..010bcadab 100644 --- a/lib/include/avx512fintrin.h +++ b/lib/include/avx512fintrin.h @@ -9300,8 +9300,11 @@ _mm512_mask_abs_pd(__m512d __W, __mmask8 __K, __m512d __A) * computations. In vector-reduction arithmetic, the evaluation order is * independent of the order of the input elements of V. - * For floating point types, we always assume the elements are reassociable even - * if -fast-math is off. + * For floating-point intrinsics: + * 1. When using fadd/fmul intrinsics, the order of operations within the + * vector is unspecified (associative math). + * 2. When using fmin/fmax intrinsics, NaN or -0.0 elements within the vector + * produce unspecified results. * Used bisection method. At each step, we partition the vector with previous * step in half, and the operation is performed on its two halves. @@ -9524,75 +9527,49 @@ _mm512_mask_reduce_min_epu32(__mmask16 __M, __m512i __V) { return __builtin_ia32_reduce_umin_d512((__v16si)__V); } -#define _mm512_mask_reduce_operator(op) \ - __m256d __t1 = _mm512_extractf64x4_pd(__V, 0); \ - __m256d __t2 = _mm512_extractf64x4_pd(__V, 1); \ - __m256d __t3 = _mm256_##op(__t1, __t2); \ - __m128d __t4 = _mm256_extractf128_pd(__t3, 0); \ - __m128d __t5 = _mm256_extractf128_pd(__t3, 1); \ - __m128d __t6 = _mm_##op(__t4, __t5); \ - __m128d __t7 = __builtin_shufflevector(__t6, __t6, 1, 0); \ - __m128d __t8 = _mm_##op(__t6, __t7); \ - return __t8[0] - static __inline__ double __DEFAULT_FN_ATTRS512 _mm512_reduce_max_pd(__m512d __V) { - _mm512_mask_reduce_operator(max_pd); + return __builtin_ia32_reduce_fmax_pd512(__V); } static __inline__ double __DEFAULT_FN_ATTRS512 _mm512_reduce_min_pd(__m512d __V) { - _mm512_mask_reduce_operator(min_pd); + return __builtin_ia32_reduce_fmin_pd512(__V); } static __inline__ double __DEFAULT_FN_ATTRS512 _mm512_mask_reduce_max_pd(__mmask8 __M, __m512d __V) { __V = _mm512_mask_mov_pd(_mm512_set1_pd(-__builtin_inf()), __M, __V); - _mm512_mask_reduce_operator(max_pd); + return __builtin_ia32_reduce_fmax_pd512(__V); } static __inline__ double __DEFAULT_FN_ATTRS512 _mm512_mask_reduce_min_pd(__mmask8 __M, __m512d __V) { __V = _mm512_mask_mov_pd(_mm512_set1_pd(__builtin_inf()), __M, __V); - _mm512_mask_reduce_operator(min_pd); -} -#undef _mm512_mask_reduce_operator - -#define _mm512_mask_reduce_operator(op) \ - __m256 __t1 = (__m256)_mm512_extractf64x4_pd((__m512d)__V, 0); \ - __m256 __t2 = (__m256)_mm512_extractf64x4_pd((__m512d)__V, 1); \ - __m256 __t3 = _mm256_##op(__t1, __t2); \ - __m128 __t4 = _mm256_extractf128_ps(__t3, 0); \ - __m128 __t5 = _mm256_extractf128_ps(__t3, 1); \ - __m128 __t6 = _mm_##op(__t4, __t5); \ - __m128 __t7 = __builtin_shufflevector(__t6, __t6, 2, 3, 0, 1); \ - __m128 __t8 = _mm_##op(__t6, __t7); \ - __m128 __t9 = __builtin_shufflevector(__t8, __t8, 1, 0, 3, 2); \ - __m128 __t10 = _mm_##op(__t8, __t9); \ - return __t10[0] + return __builtin_ia32_reduce_fmin_pd512(__V); +} static __inline__ float __DEFAULT_FN_ATTRS512 _mm512_reduce_max_ps(__m512 __V) { - _mm512_mask_reduce_operator(max_ps); + return __builtin_ia32_reduce_fmax_ps512(__V); } static __inline__ float __DEFAULT_FN_ATTRS512 _mm512_reduce_min_ps(__m512 __V) { - _mm512_mask_reduce_operator(min_ps); + return __builtin_ia32_reduce_fmin_ps512(__V); } static __inline__ float __DEFAULT_FN_ATTRS512 _mm512_mask_reduce_max_ps(__mmask16 __M, __m512 __V) { __V = _mm512_mask_mov_ps(_mm512_set1_ps(-__builtin_inff()), __M, __V); - _mm512_mask_reduce_operator(max_ps); + return __builtin_ia32_reduce_fmax_ps512(__V); } static __inline__ float __DEFAULT_FN_ATTRS512 _mm512_mask_reduce_min_ps(__mmask16 __M, __m512 __V) { __V = _mm512_mask_mov_ps(_mm512_set1_ps(__builtin_inff()), __M, __V); - _mm512_mask_reduce_operator(min_ps); + return __builtin_ia32_reduce_fmin_ps512(__V); } -#undef _mm512_mask_reduce_operator /// Moves the least significant 32 bits of a vector of [16 x i32] to a /// 32-bit signed integer value. @@ -9611,6 +9588,169 @@ _mm512_cvtsi512_si32(__m512i __A) { return __b[0]; } +/// Loads 8 double-precision (64-bit) floating-point elements stored at memory +/// locations starting at location \a base_addr at packed 32-bit integer indices +/// stored in the lower half of \a vindex scaled by \a scale them in dst. +/// +/// This intrinsic corresponds to the VGATHERDPD instructions. +/// +/// \operation +/// FOR j := 0 to 7 +/// i := j*64 +/// m := j*32 +/// addr := base_addr + SignExtend64(vindex[m+31:m]) * ZeroExtend64(scale) * 8 +/// dst[i+63:i] := MEM[addr+63:addr] +/// ENDFOR +/// dst[MAX:512] := 0 +/// \endoperation +#define _mm512_i32logather_pd(vindex, base_addr, scale) \ + _mm512_i32gather_pd(_mm512_castsi512_si256(vindex), (base_addr), (scale)) + +/// Loads 8 double-precision (64-bit) floating-point elements from memory +/// starting at location \a base_addr at packed 32-bit integer indices stored in +/// the lower half of \a vindex scaled by \a scale into dst using writemask +/// \a mask (elements are copied from \a src when the corresponding mask bit is +/// not set). +/// +/// This intrinsic corresponds to the VGATHERDPD instructions. +/// +/// \operation +/// FOR j := 0 to 7 +/// i := j*64 +/// m := j*32 +/// IF mask[j] +/// addr := base_addr + SignExtend64(vindex[m+31:m]) * ZeroExtend64(scale) * 8 +/// dst[i+63:i] := MEM[addr+63:addr] +/// ELSE +/// dst[i+63:i] := src[i+63:i] +/// FI +/// ENDFOR +/// dst[MAX:512] := 0 +/// \endoperation +#define _mm512_mask_i32logather_pd(src, mask, vindex, base_addr, scale) \ + _mm512_mask_i32gather_pd((src), (mask), _mm512_castsi512_si256(vindex), \ + (base_addr), (scale)) + +/// Loads 8 64-bit integer elements from memory starting at location \a base_addr +/// at packed 32-bit integer indices stored in the lower half of \a vindex +/// scaled by \a scale and stores them in dst. +/// +/// This intrinsic corresponds to the VPGATHERDQ instructions. +/// +/// \operation +/// FOR j := 0 to 7 +/// i := j*64 +/// m := j*32 +/// addr := base_addr + SignExtend64(vindex[m+31:m]) * ZeroExtend64(scale) * 8 +/// dst[i+63:i] := MEM[addr+63:addr] +/// ENDFOR +/// dst[MAX:512] := 0 +/// \endoperation +#define _mm512_i32logather_epi64(vindex, base_addr, scale) \ + _mm512_i32gather_epi64(_mm512_castsi512_si256(vindex), (base_addr), (scale)) + +/// Loads 8 64-bit integer elements from memory starting at location \a base_addr +/// at packed 32-bit integer indices stored in the lower half of \a vindex +/// scaled by \a scale and stores them in dst using writemask \a mask (elements +/// are copied from \a src when the corresponding mask bit is not set). +/// +/// This intrinsic corresponds to the VPGATHERDQ instructions. +/// +/// \operation +/// FOR j := 0 to 7 +/// i := j*64 +/// m := j*32 +/// IF mask[j] +/// addr := base_addr + SignExtend64(vindex[m+31:m]) * ZeroExtend64(scale) * 8 +/// dst[i+63:i] := MEM[addr+63:addr] +/// ELSE +/// dst[i+63:i] := src[i+63:i] +/// FI +/// ENDFOR +/// dst[MAX:512] := 0 +/// \endoperation +#define _mm512_mask_i32logather_epi64(src, mask, vindex, base_addr, scale) \ + _mm512_mask_i32gather_epi64((src), (mask), _mm512_castsi512_si256(vindex), \ + (base_addr), (scale)) + +/// Stores 8 packed double-precision (64-bit) floating-point elements in \a v1 +/// and to memory locations starting at location \a base_addr at packed 32-bit +/// integer indices stored in \a vindex scaled by \a scale. +/// +/// This intrinsic corresponds to the VSCATTERDPD instructions. +/// +/// \operation +/// FOR j := 0 to 7 +/// i := j*64 +/// m := j*32 +/// addr := base_addr + SignExtend64(vindex[m+31:m]) * ZeroExtend64(scale) * 8 +/// MEM[addr+63:addr] := v1[i+63:i] +/// ENDFOR +/// \endoperation +#define _mm512_i32loscatter_pd(base_addr, vindex, v1, scale) \ + _mm512_i32scatter_pd((base_addr), _mm512_castsi512_si256(vindex), (v1), (scale)) + +/// Stores 8 packed double-precision (64-bit) floating-point elements in \a v1 +/// to memory locations starting at location \a base_addr at packed 32-bit +/// integer indices stored in \a vindex scaled by \a scale. Only those elements +/// whose corresponding mask bit is set in writemask \a mask are written to +/// memory. +/// +/// This intrinsic corresponds to the VSCATTERDPD instructions. +/// +/// \operation +/// FOR j := 0 to 7 +/// i := j*64 +/// m := j*32 +/// IF mask[j] +/// addr := base_addr + SignExtend64(vindex[m+31:m]) * ZeroExtend64(scale) * 8 +/// MEM[addr+63:addr] := a[i+63:i] +/// FI +/// ENDFOR +/// \endoperation +#define _mm512_mask_i32loscatter_pd(base_addr, mask, vindex, v1, scale) \ + _mm512_mask_i32scatter_pd((base_addr), (mask), \ + _mm512_castsi512_si256(vindex), (v1), (scale)) + +/// Stores 8 packed 64-bit integer elements located in \a v1 and stores them in +/// memory locations starting at location \a base_addr at packed 32-bit integer +/// indices stored in \a vindex scaled by \a scale. +/// +/// This intrinsic corresponds to the VPSCATTERDQ instructions. +/// +/// \operation +/// FOR j := 0 to 7 +/// i := j*64 +/// m := j*32 +/// addr := base_addr + SignExtend64(vindex[m+31:m]) * ZeroExtend64(scale) * 8 +/// MEM[addr+63:addr] := a[i+63:i] +/// ENDFOR +/// \endoperation +#define _mm512_i32loscatter_epi64(base_addr, vindex, v1, scale) \ + _mm512_i32scatter_epi64((base_addr), \ + _mm512_castsi512_si256(vindex), (v1), (scale)) + +/// Stores 8 packed 64-bit integer elements located in a and stores them in +/// memory locations starting at location \a base_addr at packed 32-bit integer +/// indices stored in \a vindex scaled by scale using writemask \a mask (elements +/// whose corresponding mask bit is not set are not written to memory). +/// +/// This intrinsic corresponds to the VPSCATTERDQ instructions. +/// +/// \operation +/// FOR j := 0 to 7 +/// i := j*64 +/// m := j*32 +/// IF mask[j] +/// addr := base_addr + SignExtend64(vindex[m+31:m]) * ZeroExtend64(scale) * 8 +/// MEM[addr+63:addr] := a[i+63:i] +/// FI +/// ENDFOR +/// \endoperation +#define _mm512_mask_i32loscatter_epi64(base_addr, mask, vindex, v1, scale) \ + _mm512_mask_i32scatter_epi64((base_addr), (mask), \ + _mm512_castsi512_si256(vindex), (v1), (scale)) + #undef __DEFAULT_FN_ATTRS512 #undef __DEFAULT_FN_ATTRS128 #undef __DEFAULT_FN_ATTRS diff --git a/lib/include/builtins.h b/lib/include/builtins.h new file mode 100644 index 000000000..65095861c --- /dev/null +++ b/lib/include/builtins.h @@ -0,0 +1,16 @@ +/*===---- builtins.h - Standard header for extra builtins -----------------===*\ + * + * Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. + * See https://llvm.org/LICENSE.txt for license information. + * SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception + * +\*===----------------------------------------------------------------------===*/ + +/// Some legacy compilers have builtin definitions in a file named builtins.h. +/// This header file has been added to allow compatibility with code that was +/// written for those compilers. Code may have an include line for this file +/// and to avoid an error an empty file with this name is provided. +#ifndef __BUILTINS_H +#define __BUILTINS_H + +#endif /* __BUILTINS_H */ diff --git a/lib/include/cuda_wrappers/complex b/lib/include/cuda_wrappers/complex index 11d40a82a..e6805b604 100644 --- a/lib/include/cuda_wrappers/complex +++ b/lib/include/cuda_wrappers/complex @@ -72,8 +72,16 @@ #define _GLIBCXX_USE_C99_COMPLEX 0 #define _GLIBCXX_USE_C99_COMPLEX_TR1 0 +// Work around a compatibility issue with libstdc++ 11.1.0 +// https://bugs.llvm.org/show_bug.cgi?id=50383 +#pragma push_macro("__failed_assertion") +#if _GLIBCXX_RELEASE == 11 +#define __failed_assertion __cuda_failed_assertion +#endif + #include_next +#pragma pop_macro("__failed_assertion") #pragma pop_macro("_GLIBCXX_USE_C99_COMPLEX_TR1") #pragma pop_macro("_GLIBCXX_USE_C99_COMPLEX") diff --git a/lib/include/hexagon_circ_brev_intrinsics.h b/lib/include/hexagon_circ_brev_intrinsics.h new file mode 100644 index 000000000..c53786d3c --- /dev/null +++ b/lib/include/hexagon_circ_brev_intrinsics.h @@ -0,0 +1,298 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _HEXAGON_CIRC_BREV_INTRINSICS_H_ +#define _HEXAGON_CIRC_BREV_INTRINSICS_H_ 1 + +#include +#include + +/* Circular Load */ +/* ========================================================================== + Assembly Syntax: Return=instruction() + C Intrinsic Prototype: void Q6_circ_load_update_D(Word64 dst, Word64 *ptr, UWord32 incr, UWord32 bufsize, UWord32 K) + Instruction Type: InstructionType + Execution Slots: SLOT0123 + ========================================================================== */ +#define Q6_circ_load_update_D(dest,ptr,incr,bufsize,K) \ + { ptr = (int64_t *) HEXAGON_circ_ldd (ptr, &(dest), ((((K)+1)<<24)|((bufsize)<<3)), ((incr)*8)); } + +/* ========================================================================== + Assembly Syntax: Return=instruction() + C Intrinsic Prototype: void Q6_circ_load_update_W(Word32 dst, Word32 *ptr, UWord32 incr, UWord32 bufsize, UWord32 K) + Instruction Type: InstructionType + Execution Slots: SLOT0123 + ========================================================================== */ +#define Q6_circ_load_update_W(dest,ptr,incr,bufsize,K) \ + { ptr = (int *) HEXAGON_circ_ldw (ptr, &(dest), (((K)<<24)|((bufsize)<<2)), ((incr)*4)); } + +/* ========================================================================== + Assembly Syntax: Return=instruction() + C Intrinsic Prototype: void Q6_circ_load_update_H(Word16 dst, Word16 *ptr, UWord32 incr, UWord32 bufsize, UWord32 K) + Instruction Type: InstructionType + Execution Slots: SLOT0123 + ========================================================================== */ +#define Q6_circ_load_update_H(dest,ptr,incr,bufsize,K) \ + { ptr = (int16_t *) HEXAGON_circ_ldh (ptr, &(dest), ((((K)-1)<<24)|((bufsize)<<1)), ((incr)*2)); } + +/* ========================================================================== + Assembly Syntax: Return=instruction() + C Intrinsic Prototype: void Q6_circ_load_update_UH( UWord16 dst, UWord16 *ptr, UWord32 incr, UWord32 bufsize, UWord32 K) + Instruction Type: InstructionType + Execution Slots: SLOT0123 + ========================================================================== */ +#define Q6_circ_load_update_UH(dest,ptr,incr,bufsize,K) \ + { ptr = (uint16_t *) HEXAGON_circ_lduh (ptr, &(dest), ((((K)-1)<<24)|((bufsize)<<1)), ((incr)*2)); } + +/* ========================================================================== + Assembly Syntax: Return=instruction() + C Intrinsic Prototype: void Q6_circ_load_update_B(Word8 dst, Word8 *ptr, UWord32 incr, UWord32 bufsize, UWord32 K) + Instruction Type: InstructionType + Execution Slots: SLOT0123 + ========================================================================== */ +#define Q6_circ_load_update_B(dest,ptr,incr,bufsize,K) \ + { ptr = (int8_t *) HEXAGON_circ_ldb (ptr, &(dest), ((((K)-2)<<24)|(bufsize)), incr); } + +/* ========================================================================== + Assembly Syntax: Return=instruction() + C Intrinsic Prototype: void Q6_circ_load_update_UB(UWord8 dst, UWord8 *ptr, UWord32 incr, UWord32 bufsize, UWord32 K) + Instruction Type: InstructionType + Execution Slots: SLOT0123 + ========================================================================== */ +#define Q6_circ_load_update_UB(dest,ptr,incr,bufsize,K) \ + { ptr = (uint8_t *) HEXAGON_circ_ldub (ptr, &(dest), ((((K)-2)<<24)|(bufsize)), incr); } + +/* Circular Store */ +/* ========================================================================== + Assembly Syntax: Return=instruction() + C Intrinsic Prototype: void Q6_circ_store_update_D(Word64 *src, Word64 *ptr, UWord32 incr, UWord32 bufsize, UWord32 K) + Instruction Type: InstructionType + Execution Slots: SLOT0123 + ========================================================================== */ +#define Q6_circ_store_update_D(src,ptr,incr,bufsize,K) \ + { ptr = (int64_t *) HEXAGON_circ_std (ptr, src, ((((K)+1)<<24)|((bufsize)<<3)), ((incr)*8)); } + +/* ========================================================================== + Assembly Syntax: Return=instruction() + C Intrinsic Prototype: void Q6_circ_store_update_W(Word32 *src, Word32 *ptr, UWord32 incr, UWord32 bufsize, UWord32 K) + Instruction Type: InstructionType + Execution Slots: SLOT0123 + ========================================================================== */ +#define Q6_circ_store_update_W(src,ptr,incr,bufsize,K) \ + { ptr = (int *) HEXAGON_circ_stw (ptr, src, (((K)<<24)|((bufsize)<<2)), ((incr)*4)); } + +/* ========================================================================== + Assembly Syntax: Return=instruction() + C Intrinsic Prototype: void Q6_circ_store_update_HL(Word16 *src, Word16 *ptr, UWord32 incr, UWord32 bufsize, UWord32 K) + Instruction Type: InstructionType + Execution Slots: SLOT0123 + ========================================================================== */ +#define Q6_circ_store_update_HL(src,ptr,incr,bufsize,K) \ + { ptr = (int16_t *) HEXAGON_circ_sth (ptr, src, ((((K)-1)<<24)|((bufsize)<<1)), ((incr)*2)); } + +/* ========================================================================== + Assembly Syntax: Return=instruction() + C Intrinsic Prototype: void Q6_circ_store_update_HH(Word16 *src, Word16 *ptr, UWord32 incr, UWord32 bufsize, UWord32 K) + Instruction Type: InstructionType + Execution Slots: SLOT0123 + ========================================================================== */ +#define Q6_circ_store_update_HH(src,ptr,incr,bufsize,K) \ + { ptr = (int16_t *) HEXAGON_circ_sthhi (ptr, src, ((((K)-1)<<24)|((bufsize)<<1)), ((incr)*2)); } + +/* ========================================================================== + Assembly Syntax: Return=instruction() + C Intrinsic Prototype: void Q6_circ_store_update_B(Word8 *src, Word8 *ptr, UWord32 I4, UWord32 bufsize, UWord64 K) + Instruction Type: InstructionType + Execution Slots: SLOT0123 + ========================================================================== */ +#define Q6_circ_store_update_B(src,ptr,incr,bufsize,K) \ + { ptr = (int8_t *) HEXAGON_circ_stb (ptr, src, ((((K)-2)<<24)|(bufsize)), incr); } + + +/* Bit Reverse Load */ +/* ========================================================================== + Assembly Syntax: Return=instruction() + C Intrinsic Prototype: void Q6_bitrev_load_update_D(Word64 dst, Word64 *ptr, UWord32 Iu4) + Instruction Type: InstructionType + Execution Slots: SLOT0123 + ========================================================================== */ +#define Q6_bitrev_load_update_D(dest,ptr,log2bufsize) \ + { ptr = (int64_t *) HEXAGON_brev_ldd (ptr, &(dest), (1<<(16-((log2bufsize) + 3)))); } + +/* ========================================================================== + Assembly Syntax: Return=instruction() + C Intrinsic Prototype: void Q6_bitrev_load_update_W(Word32 dst, Word32 *ptr, UWord32 Iu4) + Instruction Type: InstructionType + Execution Slots: SLOT0123 + ========================================================================== */ +#define Q6_bitrev_load_update_W(dest,ptr,log2bufsize) \ + { ptr = (int *) HEXAGON_brev_ldw (ptr, &(dest), (1<<(16-((log2bufsize) + 2)))); } + +/* ========================================================================== + Assembly Syntax: Return=instruction() + C Intrinsic Prototype: void Q6_bitrev_load_update_H(Word16 dst, Word16 *ptr, UWord32 Iu4) + Instruction Type: InstructionType + Execution Slots: SLOT0123 + ========================================================================== */ +#define Q6_bitrev_load_update_H(dest,ptr,log2bufsize) \ + { ptr = (int16_t *) HEXAGON_brev_ldh (ptr, &(dest), (1<<(16-((log2bufsize) + 1)))); } + +/* ========================================================================== + Assembly Syntax: Return=instruction() + C Intrinsic Prototype: void Q6_bitrev_load_update_UH(UWord16 dst, UWord16 *ptr, UWord32 Iu4) + Instruction Type: InstructionType + Execution Slots: SLOT0123 + ========================================================================== */ +#define Q6_bitrev_load_update_UH(dest,ptr,log2bufsize) \ + { ptr = (uint16_t *) HEXAGON_brev_lduh (ptr, &(dest), (1<<(16-((log2bufsize) + 1)))); } + +/* ========================================================================== + Assembly Syntax: Return=instruction() + C Intrinsic Prototype: void Q6_bitrev_load_update_B(Word8 dst, Word8 *ptr, UWord32 Iu4) + Instruction Type: InstructionType + Execution Slots: SLOT0123 + ========================================================================== */ +#define Q6_bitrev_load_update_B(dest,ptr,log2bufsize) \ + { ptr = (int8_t *) HEXAGON_brev_ldb (ptr, &(dest), (1<<(16-((log2bufsize))))); } + +/* ========================================================================== + Assembly Syntax: Return=instruction() + C Intrinsic Prototype: void Q6_bitrev_load_update_UB(UWord8 dst, UWord8 *ptr, UWord32 Iu4) + Instruction Type: InstructionType + Execution Slots: SLOT0123 + ========================================================================== */ +#define Q6_bitrev_load_update_UB(dest,ptr,log2bufsize) \ + { ptr = (uint8_t *) HEXAGON_brev_ldub (ptr, &(dest), (1<<(16-((log2bufsize))))); } + +/* Bit Reverse Store */ + +/* ========================================================================== + Assembly Syntax: Return=instruction() + C Intrinsic Prototype: void Q6_bitrev_store_update_D(Word64 *src, Word64 *ptr, UWord32 Iu4) + Instruction Type: InstructionType + Execution Slots: SLOT0123 + ========================================================================== */ +#define Q6_bitrev_store_update_D(src,ptr,log2bufsize) \ + { ptr = (int64_t *) HEXAGON_brev_std (ptr, src, (1<<(16-((log2bufsize) + 3)))); } + +/* ========================================================================== + Assembly Syntax: Return=instruction() + C Intrinsic Prototype: void Q6_bitrev_store_update_W(Word32 *src, Word32 *ptr, UWord32 Iu4) + Instruction Type: InstructionType + Execution Slots: SLOT0123 + ========================================================================== */ +#define Q6_bitrev_store_update_W(src,ptr,log2bufsize) \ + { ptr = (int *) HEXAGON_brev_stw (ptr, src, (1<<(16-((log2bufsize) + 2)))); } + +/* ========================================================================== + Assembly Syntax: Return=instruction() + C Intrinsic Prototype: void Q6_bitrev_store_update_HL(Word16 *src, Word16 *ptr, Word32 Iu4) + Instruction Type: InstructionType + Execution Slots: SLOT0123 + ========================================================================== */ +#define Q6_bitrev_store_update_HL(src,ptr,log2bufsize) \ + { ptr = (int16_t *) HEXAGON_brev_sth (ptr, src, (1<<(16-((log2bufsize) + 1)))); } + +/* ========================================================================== + Assembly Syntax: Return=instruction() + C Intrinsic Prototype: void Q6_bitrev_store_update_HH(Word16 *src, Word16 *ptr, UWord32 Iu4) + Instruction Type: InstructionType + Execution Slots: SLOT0123 + ========================================================================== */ +#define Q6_bitrev_store_update_HH(src,ptr,log2bufsize) \ + { ptr = (int16_t *) HEXAGON_brev_sthhi (ptr, src, (1<<(16-((log2bufsize) + 1)))); } + +/* ========================================================================== + Assembly Syntax: Return=instruction() + C Intrinsic Prototype: void Q6_bitrev_store_update_B(Word8 *src, Word8 *ptr, UWord32 Iu4) + Instruction Type: InstructionType + Execution Slots: SLOT0123 + ========================================================================== */ +#define Q6_bitrev_store_update_B(src,ptr,log2bufsize) \ + { ptr = (int8_t *) HEXAGON_brev_stb (ptr, src, (1<<(16-((log2bufsize))))); } + + +#define HEXAGON_circ_ldd __builtin_circ_ldd +#define HEXAGON_circ_ldw __builtin_circ_ldw +#define HEXAGON_circ_ldh __builtin_circ_ldh +#define HEXAGON_circ_lduh __builtin_circ_lduh +#define HEXAGON_circ_ldb __builtin_circ_ldb +#define HEXAGON_circ_ldub __builtin_circ_ldub + + +#define HEXAGON_circ_std __builtin_circ_std +#define HEXAGON_circ_stw __builtin_circ_stw +#define HEXAGON_circ_sth __builtin_circ_sth +#define HEXAGON_circ_sthhi __builtin_circ_sthhi +#define HEXAGON_circ_stb __builtin_circ_stb + + +#define HEXAGON_brev_ldd __builtin_brev_ldd +#define HEXAGON_brev_ldw __builtin_brev_ldw +#define HEXAGON_brev_ldh __builtin_brev_ldh +#define HEXAGON_brev_lduh __builtin_brev_lduh +#define HEXAGON_brev_ldb __builtin_brev_ldb +#define HEXAGON_brev_ldub __builtin_brev_ldub + +#define HEXAGON_brev_std __builtin_brev_std +#define HEXAGON_brev_stw __builtin_brev_stw +#define HEXAGON_brev_sth __builtin_brev_sth +#define HEXAGON_brev_sthhi __builtin_brev_sthhi +#define HEXAGON_brev_stb __builtin_brev_stb + +#ifdef __HVX__ +/* ========================================================================== + Assembly Syntax: if (Qt) vmem(Rt+#0) = Vs + C Intrinsic Prototype: void Q6_vmaskedstoreq_QAV(HVX_VectorPred Qt, HVX_VectorAddress A, HVX_Vector Vs) + Instruction Type: COPROC_VMEM + Execution Slots: SLOT0 + ========================================================================== */ + +#define Q6_vmaskedstoreq_QAV __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmaskedstoreq) + +/* ========================================================================== + Assembly Syntax: if (!Qt) vmem(Rt+#0) = Vs + C Intrinsic Prototype: void Q6_vmaskedstorenq_QAV(HVX_VectorPred Qt, HVX_VectorAddress A, HVX_Vector Vs) + Instruction Type: COPROC_VMEM + Execution Slots: SLOT0 + ========================================================================== */ + +#define Q6_vmaskedstorenq_QAV __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmaskedstorenq) + +/* ========================================================================== + Assembly Syntax: if (Qt) vmem(Rt+#0):nt = Vs + C Intrinsic Prototype: void Q6_vmaskedstorentq_QAV(HVX_VectorPred Qt, HVX_VectorAddress A, HVX_Vector Vs) + Instruction Type: COPROC_VMEM + Execution Slots: SLOT0 + ========================================================================== */ + +#define Q6_vmaskedstorentq_QAV __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmaskedstorentq) + +/* ========================================================================== + Assembly Syntax: if (!Qt) vmem(Rt+#0):nt = Vs + C Intrinsic Prototype: void Q6_vmaskedstorentnq_QAV(HVX_VectorPred Qt, HVX_VectorAddress A, HVX_Vector Vs) + Instruction Type: COPROC_VMEM + Execution Slots: SLOT0 + ========================================================================== */ + +#define Q6_vmaskedstorentnq_QAV __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmaskedstorentnq) + +#endif + + +#endif /* #ifndef _HEXAGON_CIRC_BREV_INTRINSICS_H_ */ + +#ifdef __NOT_DEFINED__ +/*** comment block template ***/ +/* ========================================================================== + Assembly Syntax: Return=instruction() + C Intrinsic Prototype: ReturnType Intrinsic(ParamType Rs, ParamType Rt) + Instruction Type: InstructionType + Execution Slots: SLOT0123 + ========================================================================== */ +#endif /*** __NOT_DEFINED__ ***/ diff --git a/lib/include/hexagon_protos.h b/lib/include/hexagon_protos.h new file mode 100644 index 000000000..cdffd93bb --- /dev/null +++ b/lib/include/hexagon_protos.h @@ -0,0 +1,8450 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// +// Automatically generated file, do not edit! +//===----------------------------------------------------------------------===// + + + +#ifndef __HEXAGON_PROTOS_H_ +#define __HEXAGON_PROTOS_H_ 1 + +/* ========================================================================== + Assembly Syntax: Rd32=abs(Rs32) + C Intrinsic Prototype: Word32 Q6_R_abs_R(Word32 Rs) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_abs_R __builtin_HEXAGON_A2_abs + +/* ========================================================================== + Assembly Syntax: Rdd32=abs(Rss32) + C Intrinsic Prototype: Word64 Q6_P_abs_P(Word64 Rss) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_abs_P __builtin_HEXAGON_A2_absp + +/* ========================================================================== + Assembly Syntax: Rd32=abs(Rs32):sat + C Intrinsic Prototype: Word32 Q6_R_abs_R_sat(Word32 Rs) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_abs_R_sat __builtin_HEXAGON_A2_abssat + +/* ========================================================================== + Assembly Syntax: Rd32=add(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_add_RR(Word32 Rs, Word32 Rt) + Instruction Type: ALU32_3op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_add_RR __builtin_HEXAGON_A2_add + +/* ========================================================================== + Assembly Syntax: Rd32=add(Rt32.h,Rs32.h):<<16 + C Intrinsic Prototype: Word32 Q6_R_add_RhRh_s16(Word32 Rt, Word32 Rs) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_add_RhRh_s16 __builtin_HEXAGON_A2_addh_h16_hh + +/* ========================================================================== + Assembly Syntax: Rd32=add(Rt32.h,Rs32.l):<<16 + C Intrinsic Prototype: Word32 Q6_R_add_RhRl_s16(Word32 Rt, Word32 Rs) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_add_RhRl_s16 __builtin_HEXAGON_A2_addh_h16_hl + +/* ========================================================================== + Assembly Syntax: Rd32=add(Rt32.l,Rs32.h):<<16 + C Intrinsic Prototype: Word32 Q6_R_add_RlRh_s16(Word32 Rt, Word32 Rs) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_add_RlRh_s16 __builtin_HEXAGON_A2_addh_h16_lh + +/* ========================================================================== + Assembly Syntax: Rd32=add(Rt32.l,Rs32.l):<<16 + C Intrinsic Prototype: Word32 Q6_R_add_RlRl_s16(Word32 Rt, Word32 Rs) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_add_RlRl_s16 __builtin_HEXAGON_A2_addh_h16_ll + +/* ========================================================================== + Assembly Syntax: Rd32=add(Rt32.h,Rs32.h):sat:<<16 + C Intrinsic Prototype: Word32 Q6_R_add_RhRh_sat_s16(Word32 Rt, Word32 Rs) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_add_RhRh_sat_s16 __builtin_HEXAGON_A2_addh_h16_sat_hh + +/* ========================================================================== + Assembly Syntax: Rd32=add(Rt32.h,Rs32.l):sat:<<16 + C Intrinsic Prototype: Word32 Q6_R_add_RhRl_sat_s16(Word32 Rt, Word32 Rs) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_add_RhRl_sat_s16 __builtin_HEXAGON_A2_addh_h16_sat_hl + +/* ========================================================================== + Assembly Syntax: Rd32=add(Rt32.l,Rs32.h):sat:<<16 + C Intrinsic Prototype: Word32 Q6_R_add_RlRh_sat_s16(Word32 Rt, Word32 Rs) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_add_RlRh_sat_s16 __builtin_HEXAGON_A2_addh_h16_sat_lh + +/* ========================================================================== + Assembly Syntax: Rd32=add(Rt32.l,Rs32.l):sat:<<16 + C Intrinsic Prototype: Word32 Q6_R_add_RlRl_sat_s16(Word32 Rt, Word32 Rs) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_add_RlRl_sat_s16 __builtin_HEXAGON_A2_addh_h16_sat_ll + +/* ========================================================================== + Assembly Syntax: Rd32=add(Rt32.l,Rs32.h) + C Intrinsic Prototype: Word32 Q6_R_add_RlRh(Word32 Rt, Word32 Rs) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_add_RlRh __builtin_HEXAGON_A2_addh_l16_hl + +/* ========================================================================== + Assembly Syntax: Rd32=add(Rt32.l,Rs32.l) + C Intrinsic Prototype: Word32 Q6_R_add_RlRl(Word32 Rt, Word32 Rs) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_add_RlRl __builtin_HEXAGON_A2_addh_l16_ll + +/* ========================================================================== + Assembly Syntax: Rd32=add(Rt32.l,Rs32.h):sat + C Intrinsic Prototype: Word32 Q6_R_add_RlRh_sat(Word32 Rt, Word32 Rs) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_add_RlRh_sat __builtin_HEXAGON_A2_addh_l16_sat_hl + +/* ========================================================================== + Assembly Syntax: Rd32=add(Rt32.l,Rs32.l):sat + C Intrinsic Prototype: Word32 Q6_R_add_RlRl_sat(Word32 Rt, Word32 Rs) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_add_RlRl_sat __builtin_HEXAGON_A2_addh_l16_sat_ll + +/* ========================================================================== + Assembly Syntax: Rd32=add(Rs32,#s16) + C Intrinsic Prototype: Word32 Q6_R_add_RI(Word32 Rs, Word32 Is16) + Instruction Type: ALU32_ADDI + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_add_RI __builtin_HEXAGON_A2_addi + +/* ========================================================================== + Assembly Syntax: Rdd32=add(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_add_PP(Word64 Rss, Word64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_add_PP __builtin_HEXAGON_A2_addp + +/* ========================================================================== + Assembly Syntax: Rdd32=add(Rss32,Rtt32):sat + C Intrinsic Prototype: Word64 Q6_P_add_PP_sat(Word64 Rss, Word64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_add_PP_sat __builtin_HEXAGON_A2_addpsat + +/* ========================================================================== + Assembly Syntax: Rd32=add(Rs32,Rt32):sat + C Intrinsic Prototype: Word32 Q6_R_add_RR_sat(Word32 Rs, Word32 Rt) + Instruction Type: ALU32_3op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_add_RR_sat __builtin_HEXAGON_A2_addsat + +/* ========================================================================== + Assembly Syntax: Rdd32=add(Rs32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_add_RP(Word32 Rs, Word64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_P_add_RP __builtin_HEXAGON_A2_addsp + +/* ========================================================================== + Assembly Syntax: Rd32=and(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_and_RR(Word32 Rs, Word32 Rt) + Instruction Type: ALU32_3op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_and_RR __builtin_HEXAGON_A2_and + +/* ========================================================================== + Assembly Syntax: Rd32=and(Rs32,#s10) + C Intrinsic Prototype: Word32 Q6_R_and_RI(Word32 Rs, Word32 Is10) + Instruction Type: ALU32_2op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_and_RI __builtin_HEXAGON_A2_andir + +/* ========================================================================== + Assembly Syntax: Rdd32=and(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_and_PP(Word64 Rss, Word64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_and_PP __builtin_HEXAGON_A2_andp + +/* ========================================================================== + Assembly Syntax: Rd32=aslh(Rs32) + C Intrinsic Prototype: Word32 Q6_R_aslh_R(Word32 Rs) + Instruction Type: ALU32_2op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_aslh_R __builtin_HEXAGON_A2_aslh + +/* ========================================================================== + Assembly Syntax: Rd32=asrh(Rs32) + C Intrinsic Prototype: Word32 Q6_R_asrh_R(Word32 Rs) + Instruction Type: ALU32_2op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_asrh_R __builtin_HEXAGON_A2_asrh + +/* ========================================================================== + Assembly Syntax: Rd32=combine(Rt32.h,Rs32.h) + C Intrinsic Prototype: Word32 Q6_R_combine_RhRh(Word32 Rt, Word32 Rs) + Instruction Type: ALU32_3op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_combine_RhRh __builtin_HEXAGON_A2_combine_hh + +/* ========================================================================== + Assembly Syntax: Rd32=combine(Rt32.h,Rs32.l) + C Intrinsic Prototype: Word32 Q6_R_combine_RhRl(Word32 Rt, Word32 Rs) + Instruction Type: ALU32_3op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_combine_RhRl __builtin_HEXAGON_A2_combine_hl + +/* ========================================================================== + Assembly Syntax: Rd32=combine(Rt32.l,Rs32.h) + C Intrinsic Prototype: Word32 Q6_R_combine_RlRh(Word32 Rt, Word32 Rs) + Instruction Type: ALU32_3op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_combine_RlRh __builtin_HEXAGON_A2_combine_lh + +/* ========================================================================== + Assembly Syntax: Rd32=combine(Rt32.l,Rs32.l) + C Intrinsic Prototype: Word32 Q6_R_combine_RlRl(Word32 Rt, Word32 Rs) + Instruction Type: ALU32_3op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_combine_RlRl __builtin_HEXAGON_A2_combine_ll + +/* ========================================================================== + Assembly Syntax: Rdd32=combine(#s8,#S8) + C Intrinsic Prototype: Word64 Q6_P_combine_II(Word32 Is8, Word32 IS8) + Instruction Type: ALU32_2op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_P_combine_II __builtin_HEXAGON_A2_combineii + +/* ========================================================================== + Assembly Syntax: Rdd32=combine(Rs32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_combine_RR(Word32 Rs, Word32 Rt) + Instruction Type: ALU32_3op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_P_combine_RR __builtin_HEXAGON_A2_combinew + +/* ========================================================================== + Assembly Syntax: Rd32=max(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_max_RR(Word32 Rs, Word32 Rt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_max_RR __builtin_HEXAGON_A2_max + +/* ========================================================================== + Assembly Syntax: Rdd32=max(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_max_PP(Word64 Rss, Word64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_max_PP __builtin_HEXAGON_A2_maxp + +/* ========================================================================== + Assembly Syntax: Rd32=maxu(Rs32,Rt32) + C Intrinsic Prototype: UWord32 Q6_R_maxu_RR(Word32 Rs, Word32 Rt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_maxu_RR __builtin_HEXAGON_A2_maxu + +/* ========================================================================== + Assembly Syntax: Rdd32=maxu(Rss32,Rtt32) + C Intrinsic Prototype: UWord64 Q6_P_maxu_PP(Word64 Rss, Word64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_maxu_PP __builtin_HEXAGON_A2_maxup + +/* ========================================================================== + Assembly Syntax: Rd32=min(Rt32,Rs32) + C Intrinsic Prototype: Word32 Q6_R_min_RR(Word32 Rt, Word32 Rs) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_min_RR __builtin_HEXAGON_A2_min + +/* ========================================================================== + Assembly Syntax: Rdd32=min(Rtt32,Rss32) + C Intrinsic Prototype: Word64 Q6_P_min_PP(Word64 Rtt, Word64 Rss) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_min_PP __builtin_HEXAGON_A2_minp + +/* ========================================================================== + Assembly Syntax: Rd32=minu(Rt32,Rs32) + C Intrinsic Prototype: UWord32 Q6_R_minu_RR(Word32 Rt, Word32 Rs) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_minu_RR __builtin_HEXAGON_A2_minu + +/* ========================================================================== + Assembly Syntax: Rdd32=minu(Rtt32,Rss32) + C Intrinsic Prototype: UWord64 Q6_P_minu_PP(Word64 Rtt, Word64 Rss) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_minu_PP __builtin_HEXAGON_A2_minup + +/* ========================================================================== + Assembly Syntax: Rd32=neg(Rs32) + C Intrinsic Prototype: Word32 Q6_R_neg_R(Word32 Rs) + Instruction Type: ALU32_2op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_neg_R __builtin_HEXAGON_A2_neg + +/* ========================================================================== + Assembly Syntax: Rdd32=neg(Rss32) + C Intrinsic Prototype: Word64 Q6_P_neg_P(Word64 Rss) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_neg_P __builtin_HEXAGON_A2_negp + +/* ========================================================================== + Assembly Syntax: Rd32=neg(Rs32):sat + C Intrinsic Prototype: Word32 Q6_R_neg_R_sat(Word32 Rs) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_neg_R_sat __builtin_HEXAGON_A2_negsat + +/* ========================================================================== + Assembly Syntax: Rd32=not(Rs32) + C Intrinsic Prototype: Word32 Q6_R_not_R(Word32 Rs) + Instruction Type: ALU32_2op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_not_R __builtin_HEXAGON_A2_not + +/* ========================================================================== + Assembly Syntax: Rdd32=not(Rss32) + C Intrinsic Prototype: Word64 Q6_P_not_P(Word64 Rss) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_not_P __builtin_HEXAGON_A2_notp + +/* ========================================================================== + Assembly Syntax: Rd32=or(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_or_RR(Word32 Rs, Word32 Rt) + Instruction Type: ALU32_3op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_or_RR __builtin_HEXAGON_A2_or + +/* ========================================================================== + Assembly Syntax: Rd32=or(Rs32,#s10) + C Intrinsic Prototype: Word32 Q6_R_or_RI(Word32 Rs, Word32 Is10) + Instruction Type: ALU32_2op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_or_RI __builtin_HEXAGON_A2_orir + +/* ========================================================================== + Assembly Syntax: Rdd32=or(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_or_PP(Word64 Rss, Word64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_or_PP __builtin_HEXAGON_A2_orp + +/* ========================================================================== + Assembly Syntax: Rd32=round(Rss32):sat + C Intrinsic Prototype: Word32 Q6_R_round_P_sat(Word64 Rss) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_round_P_sat __builtin_HEXAGON_A2_roundsat + +/* ========================================================================== + Assembly Syntax: Rd32=sat(Rss32) + C Intrinsic Prototype: Word32 Q6_R_sat_P(Word64 Rss) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_sat_P __builtin_HEXAGON_A2_sat + +/* ========================================================================== + Assembly Syntax: Rd32=satb(Rs32) + C Intrinsic Prototype: Word32 Q6_R_satb_R(Word32 Rs) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_satb_R __builtin_HEXAGON_A2_satb + +/* ========================================================================== + Assembly Syntax: Rd32=sath(Rs32) + C Intrinsic Prototype: Word32 Q6_R_sath_R(Word32 Rs) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_sath_R __builtin_HEXAGON_A2_sath + +/* ========================================================================== + Assembly Syntax: Rd32=satub(Rs32) + C Intrinsic Prototype: Word32 Q6_R_satub_R(Word32 Rs) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_satub_R __builtin_HEXAGON_A2_satub + +/* ========================================================================== + Assembly Syntax: Rd32=satuh(Rs32) + C Intrinsic Prototype: Word32 Q6_R_satuh_R(Word32 Rs) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_satuh_R __builtin_HEXAGON_A2_satuh + +/* ========================================================================== + Assembly Syntax: Rd32=sub(Rt32,Rs32) + C Intrinsic Prototype: Word32 Q6_R_sub_RR(Word32 Rt, Word32 Rs) + Instruction Type: ALU32_3op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_sub_RR __builtin_HEXAGON_A2_sub + +/* ========================================================================== + Assembly Syntax: Rd32=sub(Rt32.h,Rs32.h):<<16 + C Intrinsic Prototype: Word32 Q6_R_sub_RhRh_s16(Word32 Rt, Word32 Rs) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_sub_RhRh_s16 __builtin_HEXAGON_A2_subh_h16_hh + +/* ========================================================================== + Assembly Syntax: Rd32=sub(Rt32.h,Rs32.l):<<16 + C Intrinsic Prototype: Word32 Q6_R_sub_RhRl_s16(Word32 Rt, Word32 Rs) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_sub_RhRl_s16 __builtin_HEXAGON_A2_subh_h16_hl + +/* ========================================================================== + Assembly Syntax: Rd32=sub(Rt32.l,Rs32.h):<<16 + C Intrinsic Prototype: Word32 Q6_R_sub_RlRh_s16(Word32 Rt, Word32 Rs) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_sub_RlRh_s16 __builtin_HEXAGON_A2_subh_h16_lh + +/* ========================================================================== + Assembly Syntax: Rd32=sub(Rt32.l,Rs32.l):<<16 + C Intrinsic Prototype: Word32 Q6_R_sub_RlRl_s16(Word32 Rt, Word32 Rs) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_sub_RlRl_s16 __builtin_HEXAGON_A2_subh_h16_ll + +/* ========================================================================== + Assembly Syntax: Rd32=sub(Rt32.h,Rs32.h):sat:<<16 + C Intrinsic Prototype: Word32 Q6_R_sub_RhRh_sat_s16(Word32 Rt, Word32 Rs) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_sub_RhRh_sat_s16 __builtin_HEXAGON_A2_subh_h16_sat_hh + +/* ========================================================================== + Assembly Syntax: Rd32=sub(Rt32.h,Rs32.l):sat:<<16 + C Intrinsic Prototype: Word32 Q6_R_sub_RhRl_sat_s16(Word32 Rt, Word32 Rs) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_sub_RhRl_sat_s16 __builtin_HEXAGON_A2_subh_h16_sat_hl + +/* ========================================================================== + Assembly Syntax: Rd32=sub(Rt32.l,Rs32.h):sat:<<16 + C Intrinsic Prototype: Word32 Q6_R_sub_RlRh_sat_s16(Word32 Rt, Word32 Rs) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_sub_RlRh_sat_s16 __builtin_HEXAGON_A2_subh_h16_sat_lh + +/* ========================================================================== + Assembly Syntax: Rd32=sub(Rt32.l,Rs32.l):sat:<<16 + C Intrinsic Prototype: Word32 Q6_R_sub_RlRl_sat_s16(Word32 Rt, Word32 Rs) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_sub_RlRl_sat_s16 __builtin_HEXAGON_A2_subh_h16_sat_ll + +/* ========================================================================== + Assembly Syntax: Rd32=sub(Rt32.l,Rs32.h) + C Intrinsic Prototype: Word32 Q6_R_sub_RlRh(Word32 Rt, Word32 Rs) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_sub_RlRh __builtin_HEXAGON_A2_subh_l16_hl + +/* ========================================================================== + Assembly Syntax: Rd32=sub(Rt32.l,Rs32.l) + C Intrinsic Prototype: Word32 Q6_R_sub_RlRl(Word32 Rt, Word32 Rs) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_sub_RlRl __builtin_HEXAGON_A2_subh_l16_ll + +/* ========================================================================== + Assembly Syntax: Rd32=sub(Rt32.l,Rs32.h):sat + C Intrinsic Prototype: Word32 Q6_R_sub_RlRh_sat(Word32 Rt, Word32 Rs) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_sub_RlRh_sat __builtin_HEXAGON_A2_subh_l16_sat_hl + +/* ========================================================================== + Assembly Syntax: Rd32=sub(Rt32.l,Rs32.l):sat + C Intrinsic Prototype: Word32 Q6_R_sub_RlRl_sat(Word32 Rt, Word32 Rs) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_sub_RlRl_sat __builtin_HEXAGON_A2_subh_l16_sat_ll + +/* ========================================================================== + Assembly Syntax: Rdd32=sub(Rtt32,Rss32) + C Intrinsic Prototype: Word64 Q6_P_sub_PP(Word64 Rtt, Word64 Rss) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_sub_PP __builtin_HEXAGON_A2_subp + +/* ========================================================================== + Assembly Syntax: Rd32=sub(#s10,Rs32) + C Intrinsic Prototype: Word32 Q6_R_sub_IR(Word32 Is10, Word32 Rs) + Instruction Type: ALU32_2op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_sub_IR __builtin_HEXAGON_A2_subri + +/* ========================================================================== + Assembly Syntax: Rd32=sub(Rt32,Rs32):sat + C Intrinsic Prototype: Word32 Q6_R_sub_RR_sat(Word32 Rt, Word32 Rs) + Instruction Type: ALU32_3op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_sub_RR_sat __builtin_HEXAGON_A2_subsat + +/* ========================================================================== + Assembly Syntax: Rd32=vaddh(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_vaddh_RR(Word32 Rs, Word32 Rt) + Instruction Type: ALU32_3op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_vaddh_RR __builtin_HEXAGON_A2_svaddh + +/* ========================================================================== + Assembly Syntax: Rd32=vaddh(Rs32,Rt32):sat + C Intrinsic Prototype: Word32 Q6_R_vaddh_RR_sat(Word32 Rs, Word32 Rt) + Instruction Type: ALU32_3op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_vaddh_RR_sat __builtin_HEXAGON_A2_svaddhs + +/* ========================================================================== + Assembly Syntax: Rd32=vadduh(Rs32,Rt32):sat + C Intrinsic Prototype: Word32 Q6_R_vadduh_RR_sat(Word32 Rs, Word32 Rt) + Instruction Type: ALU32_3op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_vadduh_RR_sat __builtin_HEXAGON_A2_svadduhs + +/* ========================================================================== + Assembly Syntax: Rd32=vavgh(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_vavgh_RR(Word32 Rs, Word32 Rt) + Instruction Type: ALU32_3op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_vavgh_RR __builtin_HEXAGON_A2_svavgh + +/* ========================================================================== + Assembly Syntax: Rd32=vavgh(Rs32,Rt32):rnd + C Intrinsic Prototype: Word32 Q6_R_vavgh_RR_rnd(Word32 Rs, Word32 Rt) + Instruction Type: ALU32_3op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_vavgh_RR_rnd __builtin_HEXAGON_A2_svavghs + +/* ========================================================================== + Assembly Syntax: Rd32=vnavgh(Rt32,Rs32) + C Intrinsic Prototype: Word32 Q6_R_vnavgh_RR(Word32 Rt, Word32 Rs) + Instruction Type: ALU32_3op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_vnavgh_RR __builtin_HEXAGON_A2_svnavgh + +/* ========================================================================== + Assembly Syntax: Rd32=vsubh(Rt32,Rs32) + C Intrinsic Prototype: Word32 Q6_R_vsubh_RR(Word32 Rt, Word32 Rs) + Instruction Type: ALU32_3op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_vsubh_RR __builtin_HEXAGON_A2_svsubh + +/* ========================================================================== + Assembly Syntax: Rd32=vsubh(Rt32,Rs32):sat + C Intrinsic Prototype: Word32 Q6_R_vsubh_RR_sat(Word32 Rt, Word32 Rs) + Instruction Type: ALU32_3op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_vsubh_RR_sat __builtin_HEXAGON_A2_svsubhs + +/* ========================================================================== + Assembly Syntax: Rd32=vsubuh(Rt32,Rs32):sat + C Intrinsic Prototype: Word32 Q6_R_vsubuh_RR_sat(Word32 Rt, Word32 Rs) + Instruction Type: ALU32_3op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_vsubuh_RR_sat __builtin_HEXAGON_A2_svsubuhs + +/* ========================================================================== + Assembly Syntax: Rd32=swiz(Rs32) + C Intrinsic Prototype: Word32 Q6_R_swiz_R(Word32 Rs) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_swiz_R __builtin_HEXAGON_A2_swiz + +/* ========================================================================== + Assembly Syntax: Rd32=sxtb(Rs32) + C Intrinsic Prototype: Word32 Q6_R_sxtb_R(Word32 Rs) + Instruction Type: ALU32_2op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_sxtb_R __builtin_HEXAGON_A2_sxtb + +/* ========================================================================== + Assembly Syntax: Rd32=sxth(Rs32) + C Intrinsic Prototype: Word32 Q6_R_sxth_R(Word32 Rs) + Instruction Type: ALU32_2op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_sxth_R __builtin_HEXAGON_A2_sxth + +/* ========================================================================== + Assembly Syntax: Rdd32=sxtw(Rs32) + C Intrinsic Prototype: Word64 Q6_P_sxtw_R(Word32 Rs) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_sxtw_R __builtin_HEXAGON_A2_sxtw + +/* ========================================================================== + Assembly Syntax: Rd32=Rs32 + C Intrinsic Prototype: Word32 Q6_R_equals_R(Word32 Rs) + Instruction Type: ALU32_2op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_equals_R __builtin_HEXAGON_A2_tfr + +/* ========================================================================== + Assembly Syntax: Rx32.h=#u16 + C Intrinsic Prototype: Word32 Q6_Rh_equals_I(Word32 Rx, Word32 Iu16) + Instruction Type: ALU32_2op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Rh_equals_I __builtin_HEXAGON_A2_tfrih + +/* ========================================================================== + Assembly Syntax: Rx32.l=#u16 + C Intrinsic Prototype: Word32 Q6_Rl_equals_I(Word32 Rx, Word32 Iu16) + Instruction Type: ALU32_2op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Rl_equals_I __builtin_HEXAGON_A2_tfril + +/* ========================================================================== + Assembly Syntax: Rdd32=Rss32 + C Intrinsic Prototype: Word64 Q6_P_equals_P(Word64 Rss) + Instruction Type: ALU32_2op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_P_equals_P __builtin_HEXAGON_A2_tfrp + +/* ========================================================================== + Assembly Syntax: Rdd32=#s8 + C Intrinsic Prototype: Word64 Q6_P_equals_I(Word32 Is8) + Instruction Type: ALU64 + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_P_equals_I __builtin_HEXAGON_A2_tfrpi + +/* ========================================================================== + Assembly Syntax: Rd32=#s16 + C Intrinsic Prototype: Word32 Q6_R_equals_I(Word32 Is16) + Instruction Type: ALU32_2op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_equals_I __builtin_HEXAGON_A2_tfrsi + +/* ========================================================================== + Assembly Syntax: Rdd32=vabsh(Rss32) + C Intrinsic Prototype: Word64 Q6_P_vabsh_P(Word64 Rss) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vabsh_P __builtin_HEXAGON_A2_vabsh + +/* ========================================================================== + Assembly Syntax: Rdd32=vabsh(Rss32):sat + C Intrinsic Prototype: Word64 Q6_P_vabsh_P_sat(Word64 Rss) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vabsh_P_sat __builtin_HEXAGON_A2_vabshsat + +/* ========================================================================== + Assembly Syntax: Rdd32=vabsw(Rss32) + C Intrinsic Prototype: Word64 Q6_P_vabsw_P(Word64 Rss) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vabsw_P __builtin_HEXAGON_A2_vabsw + +/* ========================================================================== + Assembly Syntax: Rdd32=vabsw(Rss32):sat + C Intrinsic Prototype: Word64 Q6_P_vabsw_P_sat(Word64 Rss) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vabsw_P_sat __builtin_HEXAGON_A2_vabswsat + +/* ========================================================================== + Assembly Syntax: Rdd32=vaddb(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_vaddb_PP(Word64 Rss, Word64 Rtt) + Instruction Type: MAPPING + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_P_vaddb_PP __builtin_HEXAGON_A2_vaddb_map + +/* ========================================================================== + Assembly Syntax: Rdd32=vaddh(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_vaddh_PP(Word64 Rss, Word64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vaddh_PP __builtin_HEXAGON_A2_vaddh + +/* ========================================================================== + Assembly Syntax: Rdd32=vaddh(Rss32,Rtt32):sat + C Intrinsic Prototype: Word64 Q6_P_vaddh_PP_sat(Word64 Rss, Word64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vaddh_PP_sat __builtin_HEXAGON_A2_vaddhs + +/* ========================================================================== + Assembly Syntax: Rdd32=vaddub(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_vaddub_PP(Word64 Rss, Word64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vaddub_PP __builtin_HEXAGON_A2_vaddub + +/* ========================================================================== + Assembly Syntax: Rdd32=vaddub(Rss32,Rtt32):sat + C Intrinsic Prototype: Word64 Q6_P_vaddub_PP_sat(Word64 Rss, Word64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vaddub_PP_sat __builtin_HEXAGON_A2_vaddubs + +/* ========================================================================== + Assembly Syntax: Rdd32=vadduh(Rss32,Rtt32):sat + C Intrinsic Prototype: Word64 Q6_P_vadduh_PP_sat(Word64 Rss, Word64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vadduh_PP_sat __builtin_HEXAGON_A2_vadduhs + +/* ========================================================================== + Assembly Syntax: Rdd32=vaddw(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_vaddw_PP(Word64 Rss, Word64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vaddw_PP __builtin_HEXAGON_A2_vaddw + +/* ========================================================================== + Assembly Syntax: Rdd32=vaddw(Rss32,Rtt32):sat + C Intrinsic Prototype: Word64 Q6_P_vaddw_PP_sat(Word64 Rss, Word64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vaddw_PP_sat __builtin_HEXAGON_A2_vaddws + +/* ========================================================================== + Assembly Syntax: Rdd32=vavgh(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_vavgh_PP(Word64 Rss, Word64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vavgh_PP __builtin_HEXAGON_A2_vavgh + +/* ========================================================================== + Assembly Syntax: Rdd32=vavgh(Rss32,Rtt32):crnd + C Intrinsic Prototype: Word64 Q6_P_vavgh_PP_crnd(Word64 Rss, Word64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vavgh_PP_crnd __builtin_HEXAGON_A2_vavghcr + +/* ========================================================================== + Assembly Syntax: Rdd32=vavgh(Rss32,Rtt32):rnd + C Intrinsic Prototype: Word64 Q6_P_vavgh_PP_rnd(Word64 Rss, Word64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vavgh_PP_rnd __builtin_HEXAGON_A2_vavghr + +/* ========================================================================== + Assembly Syntax: Rdd32=vavgub(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_vavgub_PP(Word64 Rss, Word64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vavgub_PP __builtin_HEXAGON_A2_vavgub + +/* ========================================================================== + Assembly Syntax: Rdd32=vavgub(Rss32,Rtt32):rnd + C Intrinsic Prototype: Word64 Q6_P_vavgub_PP_rnd(Word64 Rss, Word64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vavgub_PP_rnd __builtin_HEXAGON_A2_vavgubr + +/* ========================================================================== + Assembly Syntax: Rdd32=vavguh(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_vavguh_PP(Word64 Rss, Word64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vavguh_PP __builtin_HEXAGON_A2_vavguh + +/* ========================================================================== + Assembly Syntax: Rdd32=vavguh(Rss32,Rtt32):rnd + C Intrinsic Prototype: Word64 Q6_P_vavguh_PP_rnd(Word64 Rss, Word64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vavguh_PP_rnd __builtin_HEXAGON_A2_vavguhr + +/* ========================================================================== + Assembly Syntax: Rdd32=vavguw(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_vavguw_PP(Word64 Rss, Word64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vavguw_PP __builtin_HEXAGON_A2_vavguw + +/* ========================================================================== + Assembly Syntax: Rdd32=vavguw(Rss32,Rtt32):rnd + C Intrinsic Prototype: Word64 Q6_P_vavguw_PP_rnd(Word64 Rss, Word64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vavguw_PP_rnd __builtin_HEXAGON_A2_vavguwr + +/* ========================================================================== + Assembly Syntax: Rdd32=vavgw(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_vavgw_PP(Word64 Rss, Word64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vavgw_PP __builtin_HEXAGON_A2_vavgw + +/* ========================================================================== + Assembly Syntax: Rdd32=vavgw(Rss32,Rtt32):crnd + C Intrinsic Prototype: Word64 Q6_P_vavgw_PP_crnd(Word64 Rss, Word64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vavgw_PP_crnd __builtin_HEXAGON_A2_vavgwcr + +/* ========================================================================== + Assembly Syntax: Rdd32=vavgw(Rss32,Rtt32):rnd + C Intrinsic Prototype: Word64 Q6_P_vavgw_PP_rnd(Word64 Rss, Word64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vavgw_PP_rnd __builtin_HEXAGON_A2_vavgwr + +/* ========================================================================== + Assembly Syntax: Pd4=vcmpb.eq(Rss32,Rtt32) + C Intrinsic Prototype: Byte Q6_p_vcmpb_eq_PP(Word64 Rss, Word64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_vcmpb_eq_PP __builtin_HEXAGON_A2_vcmpbeq + +/* ========================================================================== + Assembly Syntax: Pd4=vcmpb.gtu(Rss32,Rtt32) + C Intrinsic Prototype: Byte Q6_p_vcmpb_gtu_PP(Word64 Rss, Word64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_vcmpb_gtu_PP __builtin_HEXAGON_A2_vcmpbgtu + +/* ========================================================================== + Assembly Syntax: Pd4=vcmph.eq(Rss32,Rtt32) + C Intrinsic Prototype: Byte Q6_p_vcmph_eq_PP(Word64 Rss, Word64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_vcmph_eq_PP __builtin_HEXAGON_A2_vcmpheq + +/* ========================================================================== + Assembly Syntax: Pd4=vcmph.gt(Rss32,Rtt32) + C Intrinsic Prototype: Byte Q6_p_vcmph_gt_PP(Word64 Rss, Word64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_vcmph_gt_PP __builtin_HEXAGON_A2_vcmphgt + +/* ========================================================================== + Assembly Syntax: Pd4=vcmph.gtu(Rss32,Rtt32) + C Intrinsic Prototype: Byte Q6_p_vcmph_gtu_PP(Word64 Rss, Word64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_vcmph_gtu_PP __builtin_HEXAGON_A2_vcmphgtu + +/* ========================================================================== + Assembly Syntax: Pd4=vcmpw.eq(Rss32,Rtt32) + C Intrinsic Prototype: Byte Q6_p_vcmpw_eq_PP(Word64 Rss, Word64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_vcmpw_eq_PP __builtin_HEXAGON_A2_vcmpweq + +/* ========================================================================== + Assembly Syntax: Pd4=vcmpw.gt(Rss32,Rtt32) + C Intrinsic Prototype: Byte Q6_p_vcmpw_gt_PP(Word64 Rss, Word64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_vcmpw_gt_PP __builtin_HEXAGON_A2_vcmpwgt + +/* ========================================================================== + Assembly Syntax: Pd4=vcmpw.gtu(Rss32,Rtt32) + C Intrinsic Prototype: Byte Q6_p_vcmpw_gtu_PP(Word64 Rss, Word64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_vcmpw_gtu_PP __builtin_HEXAGON_A2_vcmpwgtu + +/* ========================================================================== + Assembly Syntax: Rdd32=vconj(Rss32):sat + C Intrinsic Prototype: Word64 Q6_P_vconj_P_sat(Word64 Rss) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vconj_P_sat __builtin_HEXAGON_A2_vconj + +/* ========================================================================== + Assembly Syntax: Rdd32=vmaxb(Rtt32,Rss32) + C Intrinsic Prototype: Word64 Q6_P_vmaxb_PP(Word64 Rtt, Word64 Rss) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmaxb_PP __builtin_HEXAGON_A2_vmaxb + +/* ========================================================================== + Assembly Syntax: Rdd32=vmaxh(Rtt32,Rss32) + C Intrinsic Prototype: Word64 Q6_P_vmaxh_PP(Word64 Rtt, Word64 Rss) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmaxh_PP __builtin_HEXAGON_A2_vmaxh + +/* ========================================================================== + Assembly Syntax: Rdd32=vmaxub(Rtt32,Rss32) + C Intrinsic Prototype: Word64 Q6_P_vmaxub_PP(Word64 Rtt, Word64 Rss) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmaxub_PP __builtin_HEXAGON_A2_vmaxub + +/* ========================================================================== + Assembly Syntax: Rdd32=vmaxuh(Rtt32,Rss32) + C Intrinsic Prototype: Word64 Q6_P_vmaxuh_PP(Word64 Rtt, Word64 Rss) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmaxuh_PP __builtin_HEXAGON_A2_vmaxuh + +/* ========================================================================== + Assembly Syntax: Rdd32=vmaxuw(Rtt32,Rss32) + C Intrinsic Prototype: Word64 Q6_P_vmaxuw_PP(Word64 Rtt, Word64 Rss) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmaxuw_PP __builtin_HEXAGON_A2_vmaxuw + +/* ========================================================================== + Assembly Syntax: Rdd32=vmaxw(Rtt32,Rss32) + C Intrinsic Prototype: Word64 Q6_P_vmaxw_PP(Word64 Rtt, Word64 Rss) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmaxw_PP __builtin_HEXAGON_A2_vmaxw + +/* ========================================================================== + Assembly Syntax: Rdd32=vminb(Rtt32,Rss32) + C Intrinsic Prototype: Word64 Q6_P_vminb_PP(Word64 Rtt, Word64 Rss) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vminb_PP __builtin_HEXAGON_A2_vminb + +/* ========================================================================== + Assembly Syntax: Rdd32=vminh(Rtt32,Rss32) + C Intrinsic Prototype: Word64 Q6_P_vminh_PP(Word64 Rtt, Word64 Rss) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vminh_PP __builtin_HEXAGON_A2_vminh + +/* ========================================================================== + Assembly Syntax: Rdd32=vminub(Rtt32,Rss32) + C Intrinsic Prototype: Word64 Q6_P_vminub_PP(Word64 Rtt, Word64 Rss) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vminub_PP __builtin_HEXAGON_A2_vminub + +/* ========================================================================== + Assembly Syntax: Rdd32=vminuh(Rtt32,Rss32) + C Intrinsic Prototype: Word64 Q6_P_vminuh_PP(Word64 Rtt, Word64 Rss) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vminuh_PP __builtin_HEXAGON_A2_vminuh + +/* ========================================================================== + Assembly Syntax: Rdd32=vminuw(Rtt32,Rss32) + C Intrinsic Prototype: Word64 Q6_P_vminuw_PP(Word64 Rtt, Word64 Rss) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vminuw_PP __builtin_HEXAGON_A2_vminuw + +/* ========================================================================== + Assembly Syntax: Rdd32=vminw(Rtt32,Rss32) + C Intrinsic Prototype: Word64 Q6_P_vminw_PP(Word64 Rtt, Word64 Rss) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vminw_PP __builtin_HEXAGON_A2_vminw + +/* ========================================================================== + Assembly Syntax: Rdd32=vnavgh(Rtt32,Rss32) + C Intrinsic Prototype: Word64 Q6_P_vnavgh_PP(Word64 Rtt, Word64 Rss) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vnavgh_PP __builtin_HEXAGON_A2_vnavgh + +/* ========================================================================== + Assembly Syntax: Rdd32=vnavgh(Rtt32,Rss32):crnd:sat + C Intrinsic Prototype: Word64 Q6_P_vnavgh_PP_crnd_sat(Word64 Rtt, Word64 Rss) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vnavgh_PP_crnd_sat __builtin_HEXAGON_A2_vnavghcr + +/* ========================================================================== + Assembly Syntax: Rdd32=vnavgh(Rtt32,Rss32):rnd:sat + C Intrinsic Prototype: Word64 Q6_P_vnavgh_PP_rnd_sat(Word64 Rtt, Word64 Rss) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vnavgh_PP_rnd_sat __builtin_HEXAGON_A2_vnavghr + +/* ========================================================================== + Assembly Syntax: Rdd32=vnavgw(Rtt32,Rss32) + C Intrinsic Prototype: Word64 Q6_P_vnavgw_PP(Word64 Rtt, Word64 Rss) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vnavgw_PP __builtin_HEXAGON_A2_vnavgw + +/* ========================================================================== + Assembly Syntax: Rdd32=vnavgw(Rtt32,Rss32):crnd:sat + C Intrinsic Prototype: Word64 Q6_P_vnavgw_PP_crnd_sat(Word64 Rtt, Word64 Rss) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vnavgw_PP_crnd_sat __builtin_HEXAGON_A2_vnavgwcr + +/* ========================================================================== + Assembly Syntax: Rdd32=vnavgw(Rtt32,Rss32):rnd:sat + C Intrinsic Prototype: Word64 Q6_P_vnavgw_PP_rnd_sat(Word64 Rtt, Word64 Rss) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vnavgw_PP_rnd_sat __builtin_HEXAGON_A2_vnavgwr + +/* ========================================================================== + Assembly Syntax: Rdd32=vraddub(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_vraddub_PP(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vraddub_PP __builtin_HEXAGON_A2_vraddub + +/* ========================================================================== + Assembly Syntax: Rxx32+=vraddub(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_vraddubacc_PP(Word64 Rxx, Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vraddubacc_PP __builtin_HEXAGON_A2_vraddub_acc + +/* ========================================================================== + Assembly Syntax: Rdd32=vrsadub(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_vrsadub_PP(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vrsadub_PP __builtin_HEXAGON_A2_vrsadub + +/* ========================================================================== + Assembly Syntax: Rxx32+=vrsadub(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_vrsadubacc_PP(Word64 Rxx, Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vrsadubacc_PP __builtin_HEXAGON_A2_vrsadub_acc + +/* ========================================================================== + Assembly Syntax: Rdd32=vsubb(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_vsubb_PP(Word64 Rss, Word64 Rtt) + Instruction Type: MAPPING + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_P_vsubb_PP __builtin_HEXAGON_A2_vsubb_map + +/* ========================================================================== + Assembly Syntax: Rdd32=vsubh(Rtt32,Rss32) + C Intrinsic Prototype: Word64 Q6_P_vsubh_PP(Word64 Rtt, Word64 Rss) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vsubh_PP __builtin_HEXAGON_A2_vsubh + +/* ========================================================================== + Assembly Syntax: Rdd32=vsubh(Rtt32,Rss32):sat + C Intrinsic Prototype: Word64 Q6_P_vsubh_PP_sat(Word64 Rtt, Word64 Rss) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vsubh_PP_sat __builtin_HEXAGON_A2_vsubhs + +/* ========================================================================== + Assembly Syntax: Rdd32=vsubub(Rtt32,Rss32) + C Intrinsic Prototype: Word64 Q6_P_vsubub_PP(Word64 Rtt, Word64 Rss) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vsubub_PP __builtin_HEXAGON_A2_vsubub + +/* ========================================================================== + Assembly Syntax: Rdd32=vsubub(Rtt32,Rss32):sat + C Intrinsic Prototype: Word64 Q6_P_vsubub_PP_sat(Word64 Rtt, Word64 Rss) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vsubub_PP_sat __builtin_HEXAGON_A2_vsububs + +/* ========================================================================== + Assembly Syntax: Rdd32=vsubuh(Rtt32,Rss32):sat + C Intrinsic Prototype: Word64 Q6_P_vsubuh_PP_sat(Word64 Rtt, Word64 Rss) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vsubuh_PP_sat __builtin_HEXAGON_A2_vsubuhs + +/* ========================================================================== + Assembly Syntax: Rdd32=vsubw(Rtt32,Rss32) + C Intrinsic Prototype: Word64 Q6_P_vsubw_PP(Word64 Rtt, Word64 Rss) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vsubw_PP __builtin_HEXAGON_A2_vsubw + +/* ========================================================================== + Assembly Syntax: Rdd32=vsubw(Rtt32,Rss32):sat + C Intrinsic Prototype: Word64 Q6_P_vsubw_PP_sat(Word64 Rtt, Word64 Rss) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vsubw_PP_sat __builtin_HEXAGON_A2_vsubws + +/* ========================================================================== + Assembly Syntax: Rd32=xor(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_xor_RR(Word32 Rs, Word32 Rt) + Instruction Type: ALU32_3op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_xor_RR __builtin_HEXAGON_A2_xor + +/* ========================================================================== + Assembly Syntax: Rdd32=xor(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_xor_PP(Word64 Rss, Word64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_xor_PP __builtin_HEXAGON_A2_xorp + +/* ========================================================================== + Assembly Syntax: Rd32=zxtb(Rs32) + C Intrinsic Prototype: Word32 Q6_R_zxtb_R(Word32 Rs) + Instruction Type: ALU32_2op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_zxtb_R __builtin_HEXAGON_A2_zxtb + +/* ========================================================================== + Assembly Syntax: Rd32=zxth(Rs32) + C Intrinsic Prototype: Word32 Q6_R_zxth_R(Word32 Rs) + Instruction Type: ALU32_2op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_zxth_R __builtin_HEXAGON_A2_zxth + +/* ========================================================================== + Assembly Syntax: Rd32=and(Rt32,~Rs32) + C Intrinsic Prototype: Word32 Q6_R_and_RnR(Word32 Rt, Word32 Rs) + Instruction Type: ALU32_3op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_and_RnR __builtin_HEXAGON_A4_andn + +/* ========================================================================== + Assembly Syntax: Rdd32=and(Rtt32,~Rss32) + C Intrinsic Prototype: Word64 Q6_P_and_PnP(Word64 Rtt, Word64 Rss) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_and_PnP __builtin_HEXAGON_A4_andnp + +/* ========================================================================== + Assembly Syntax: Rdd32=bitsplit(Rs32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_bitsplit_RR(Word32 Rs, Word32 Rt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_bitsplit_RR __builtin_HEXAGON_A4_bitsplit + +/* ========================================================================== + Assembly Syntax: Rdd32=bitsplit(Rs32,#u5) + C Intrinsic Prototype: Word64 Q6_P_bitsplit_RI(Word32 Rs, Word32 Iu5) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_bitsplit_RI __builtin_HEXAGON_A4_bitspliti + +/* ========================================================================== + Assembly Syntax: Pd4=boundscheck(Rs32,Rtt32) + C Intrinsic Prototype: Byte Q6_p_boundscheck_RP(Word32 Rs, Word64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_p_boundscheck_RP __builtin_HEXAGON_A4_boundscheck + +/* ========================================================================== + Assembly Syntax: Pd4=cmpb.eq(Rs32,Rt32) + C Intrinsic Prototype: Byte Q6_p_cmpb_eq_RR(Word32 Rs, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_cmpb_eq_RR __builtin_HEXAGON_A4_cmpbeq + +/* ========================================================================== + Assembly Syntax: Pd4=cmpb.eq(Rs32,#u8) + C Intrinsic Prototype: Byte Q6_p_cmpb_eq_RI(Word32 Rs, Word32 Iu8) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_cmpb_eq_RI __builtin_HEXAGON_A4_cmpbeqi + +/* ========================================================================== + Assembly Syntax: Pd4=cmpb.gt(Rs32,Rt32) + C Intrinsic Prototype: Byte Q6_p_cmpb_gt_RR(Word32 Rs, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_cmpb_gt_RR __builtin_HEXAGON_A4_cmpbgt + +/* ========================================================================== + Assembly Syntax: Pd4=cmpb.gt(Rs32,#s8) + C Intrinsic Prototype: Byte Q6_p_cmpb_gt_RI(Word32 Rs, Word32 Is8) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_cmpb_gt_RI __builtin_HEXAGON_A4_cmpbgti + +/* ========================================================================== + Assembly Syntax: Pd4=cmpb.gtu(Rs32,Rt32) + C Intrinsic Prototype: Byte Q6_p_cmpb_gtu_RR(Word32 Rs, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_cmpb_gtu_RR __builtin_HEXAGON_A4_cmpbgtu + +/* ========================================================================== + Assembly Syntax: Pd4=cmpb.gtu(Rs32,#u7) + C Intrinsic Prototype: Byte Q6_p_cmpb_gtu_RI(Word32 Rs, Word32 Iu7) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_cmpb_gtu_RI __builtin_HEXAGON_A4_cmpbgtui + +/* ========================================================================== + Assembly Syntax: Pd4=cmph.eq(Rs32,Rt32) + C Intrinsic Prototype: Byte Q6_p_cmph_eq_RR(Word32 Rs, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_cmph_eq_RR __builtin_HEXAGON_A4_cmpheq + +/* ========================================================================== + Assembly Syntax: Pd4=cmph.eq(Rs32,#s8) + C Intrinsic Prototype: Byte Q6_p_cmph_eq_RI(Word32 Rs, Word32 Is8) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_cmph_eq_RI __builtin_HEXAGON_A4_cmpheqi + +/* ========================================================================== + Assembly Syntax: Pd4=cmph.gt(Rs32,Rt32) + C Intrinsic Prototype: Byte Q6_p_cmph_gt_RR(Word32 Rs, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_cmph_gt_RR __builtin_HEXAGON_A4_cmphgt + +/* ========================================================================== + Assembly Syntax: Pd4=cmph.gt(Rs32,#s8) + C Intrinsic Prototype: Byte Q6_p_cmph_gt_RI(Word32 Rs, Word32 Is8) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_cmph_gt_RI __builtin_HEXAGON_A4_cmphgti + +/* ========================================================================== + Assembly Syntax: Pd4=cmph.gtu(Rs32,Rt32) + C Intrinsic Prototype: Byte Q6_p_cmph_gtu_RR(Word32 Rs, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_cmph_gtu_RR __builtin_HEXAGON_A4_cmphgtu + +/* ========================================================================== + Assembly Syntax: Pd4=cmph.gtu(Rs32,#u7) + C Intrinsic Prototype: Byte Q6_p_cmph_gtu_RI(Word32 Rs, Word32 Iu7) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_cmph_gtu_RI __builtin_HEXAGON_A4_cmphgtui + +/* ========================================================================== + Assembly Syntax: Rdd32=combine(#s8,Rs32) + C Intrinsic Prototype: Word64 Q6_P_combine_IR(Word32 Is8, Word32 Rs) + Instruction Type: ALU32_2op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_P_combine_IR __builtin_HEXAGON_A4_combineir + +/* ========================================================================== + Assembly Syntax: Rdd32=combine(Rs32,#s8) + C Intrinsic Prototype: Word64 Q6_P_combine_RI(Word32 Rs, Word32 Is8) + Instruction Type: ALU32_2op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_P_combine_RI __builtin_HEXAGON_A4_combineri + +/* ========================================================================== + Assembly Syntax: Rd32=cround(Rs32,#u5) + C Intrinsic Prototype: Word32 Q6_R_cround_RI(Word32 Rs, Word32 Iu5) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_cround_RI __builtin_HEXAGON_A4_cround_ri + +/* ========================================================================== + Assembly Syntax: Rd32=cround(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_cround_RR(Word32 Rs, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_cround_RR __builtin_HEXAGON_A4_cround_rr + +/* ========================================================================== + Assembly Syntax: Rd32=modwrap(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_modwrap_RR(Word32 Rs, Word32 Rt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_modwrap_RR __builtin_HEXAGON_A4_modwrapu + +/* ========================================================================== + Assembly Syntax: Rd32=or(Rt32,~Rs32) + C Intrinsic Prototype: Word32 Q6_R_or_RnR(Word32 Rt, Word32 Rs) + Instruction Type: ALU32_3op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_or_RnR __builtin_HEXAGON_A4_orn + +/* ========================================================================== + Assembly Syntax: Rdd32=or(Rtt32,~Rss32) + C Intrinsic Prototype: Word64 Q6_P_or_PnP(Word64 Rtt, Word64 Rss) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_or_PnP __builtin_HEXAGON_A4_ornp + +/* ========================================================================== + Assembly Syntax: Rd32=cmp.eq(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_cmp_eq_RR(Word32 Rs, Word32 Rt) + Instruction Type: ALU32_3op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_cmp_eq_RR __builtin_HEXAGON_A4_rcmpeq + +/* ========================================================================== + Assembly Syntax: Rd32=cmp.eq(Rs32,#s8) + C Intrinsic Prototype: Word32 Q6_R_cmp_eq_RI(Word32 Rs, Word32 Is8) + Instruction Type: ALU32_2op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_cmp_eq_RI __builtin_HEXAGON_A4_rcmpeqi + +/* ========================================================================== + Assembly Syntax: Rd32=!cmp.eq(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_not_cmp_eq_RR(Word32 Rs, Word32 Rt) + Instruction Type: ALU32_3op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_not_cmp_eq_RR __builtin_HEXAGON_A4_rcmpneq + +/* ========================================================================== + Assembly Syntax: Rd32=!cmp.eq(Rs32,#s8) + C Intrinsic Prototype: Word32 Q6_R_not_cmp_eq_RI(Word32 Rs, Word32 Is8) + Instruction Type: ALU32_2op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_not_cmp_eq_RI __builtin_HEXAGON_A4_rcmpneqi + +/* ========================================================================== + Assembly Syntax: Rd32=round(Rs32,#u5) + C Intrinsic Prototype: Word32 Q6_R_round_RI(Word32 Rs, Word32 Iu5) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_round_RI __builtin_HEXAGON_A4_round_ri + +/* ========================================================================== + Assembly Syntax: Rd32=round(Rs32,#u5):sat + C Intrinsic Prototype: Word32 Q6_R_round_RI_sat(Word32 Rs, Word32 Iu5) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_round_RI_sat __builtin_HEXAGON_A4_round_ri_sat + +/* ========================================================================== + Assembly Syntax: Rd32=round(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_round_RR(Word32 Rs, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_round_RR __builtin_HEXAGON_A4_round_rr + +/* ========================================================================== + Assembly Syntax: Rd32=round(Rs32,Rt32):sat + C Intrinsic Prototype: Word32 Q6_R_round_RR_sat(Word32 Rs, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_round_RR_sat __builtin_HEXAGON_A4_round_rr_sat + +/* ========================================================================== + Assembly Syntax: Pd4=tlbmatch(Rss32,Rt32) + C Intrinsic Prototype: Byte Q6_p_tlbmatch_PR(Word64 Rss, Word32 Rt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_tlbmatch_PR __builtin_HEXAGON_A4_tlbmatch + +/* ========================================================================== + Assembly Syntax: Pd4=any8(vcmpb.eq(Rss32,Rtt32)) + C Intrinsic Prototype: Byte Q6_p_any8_vcmpb_eq_PP(Word64 Rss, Word64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_any8_vcmpb_eq_PP __builtin_HEXAGON_A4_vcmpbeq_any + +/* ========================================================================== + Assembly Syntax: Pd4=vcmpb.eq(Rss32,#u8) + C Intrinsic Prototype: Byte Q6_p_vcmpb_eq_PI(Word64 Rss, Word32 Iu8) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_vcmpb_eq_PI __builtin_HEXAGON_A4_vcmpbeqi + +/* ========================================================================== + Assembly Syntax: Pd4=vcmpb.gt(Rss32,Rtt32) + C Intrinsic Prototype: Byte Q6_p_vcmpb_gt_PP(Word64 Rss, Word64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_vcmpb_gt_PP __builtin_HEXAGON_A4_vcmpbgt + +/* ========================================================================== + Assembly Syntax: Pd4=vcmpb.gt(Rss32,#s8) + C Intrinsic Prototype: Byte Q6_p_vcmpb_gt_PI(Word64 Rss, Word32 Is8) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_vcmpb_gt_PI __builtin_HEXAGON_A4_vcmpbgti + +/* ========================================================================== + Assembly Syntax: Pd4=vcmpb.gtu(Rss32,#u7) + C Intrinsic Prototype: Byte Q6_p_vcmpb_gtu_PI(Word64 Rss, Word32 Iu7) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_vcmpb_gtu_PI __builtin_HEXAGON_A4_vcmpbgtui + +/* ========================================================================== + Assembly Syntax: Pd4=vcmph.eq(Rss32,#s8) + C Intrinsic Prototype: Byte Q6_p_vcmph_eq_PI(Word64 Rss, Word32 Is8) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_vcmph_eq_PI __builtin_HEXAGON_A4_vcmpheqi + +/* ========================================================================== + Assembly Syntax: Pd4=vcmph.gt(Rss32,#s8) + C Intrinsic Prototype: Byte Q6_p_vcmph_gt_PI(Word64 Rss, Word32 Is8) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_vcmph_gt_PI __builtin_HEXAGON_A4_vcmphgti + +/* ========================================================================== + Assembly Syntax: Pd4=vcmph.gtu(Rss32,#u7) + C Intrinsic Prototype: Byte Q6_p_vcmph_gtu_PI(Word64 Rss, Word32 Iu7) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_vcmph_gtu_PI __builtin_HEXAGON_A4_vcmphgtui + +/* ========================================================================== + Assembly Syntax: Pd4=vcmpw.eq(Rss32,#s8) + C Intrinsic Prototype: Byte Q6_p_vcmpw_eq_PI(Word64 Rss, Word32 Is8) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_vcmpw_eq_PI __builtin_HEXAGON_A4_vcmpweqi + +/* ========================================================================== + Assembly Syntax: Pd4=vcmpw.gt(Rss32,#s8) + C Intrinsic Prototype: Byte Q6_p_vcmpw_gt_PI(Word64 Rss, Word32 Is8) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_vcmpw_gt_PI __builtin_HEXAGON_A4_vcmpwgti + +/* ========================================================================== + Assembly Syntax: Pd4=vcmpw.gtu(Rss32,#u7) + C Intrinsic Prototype: Byte Q6_p_vcmpw_gtu_PI(Word64 Rss, Word32 Iu7) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_vcmpw_gtu_PI __builtin_HEXAGON_A4_vcmpwgtui + +/* ========================================================================== + Assembly Syntax: Rxx32=vrmaxh(Rss32,Ru32) + C Intrinsic Prototype: Word64 Q6_P_vrmaxh_PR(Word64 Rxx, Word64 Rss, Word32 Ru) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vrmaxh_PR __builtin_HEXAGON_A4_vrmaxh + +/* ========================================================================== + Assembly Syntax: Rxx32=vrmaxuh(Rss32,Ru32) + C Intrinsic Prototype: Word64 Q6_P_vrmaxuh_PR(Word64 Rxx, Word64 Rss, Word32 Ru) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vrmaxuh_PR __builtin_HEXAGON_A4_vrmaxuh + +/* ========================================================================== + Assembly Syntax: Rxx32=vrmaxuw(Rss32,Ru32) + C Intrinsic Prototype: Word64 Q6_P_vrmaxuw_PR(Word64 Rxx, Word64 Rss, Word32 Ru) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vrmaxuw_PR __builtin_HEXAGON_A4_vrmaxuw + +/* ========================================================================== + Assembly Syntax: Rxx32=vrmaxw(Rss32,Ru32) + C Intrinsic Prototype: Word64 Q6_P_vrmaxw_PR(Word64 Rxx, Word64 Rss, Word32 Ru) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vrmaxw_PR __builtin_HEXAGON_A4_vrmaxw + +/* ========================================================================== + Assembly Syntax: Rxx32=vrminh(Rss32,Ru32) + C Intrinsic Prototype: Word64 Q6_P_vrminh_PR(Word64 Rxx, Word64 Rss, Word32 Ru) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vrminh_PR __builtin_HEXAGON_A4_vrminh + +/* ========================================================================== + Assembly Syntax: Rxx32=vrminuh(Rss32,Ru32) + C Intrinsic Prototype: Word64 Q6_P_vrminuh_PR(Word64 Rxx, Word64 Rss, Word32 Ru) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vrminuh_PR __builtin_HEXAGON_A4_vrminuh + +/* ========================================================================== + Assembly Syntax: Rxx32=vrminuw(Rss32,Ru32) + C Intrinsic Prototype: Word64 Q6_P_vrminuw_PR(Word64 Rxx, Word64 Rss, Word32 Ru) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vrminuw_PR __builtin_HEXAGON_A4_vrminuw + +/* ========================================================================== + Assembly Syntax: Rxx32=vrminw(Rss32,Ru32) + C Intrinsic Prototype: Word64 Q6_P_vrminw_PR(Word64 Rxx, Word64 Rss, Word32 Ru) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vrminw_PR __builtin_HEXAGON_A4_vrminw + +/* ========================================================================== + Assembly Syntax: Rd32=vaddhub(Rss32,Rtt32):sat + C Intrinsic Prototype: Word32 Q6_R_vaddhub_PP_sat(Word64 Rss, Word64 Rtt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_vaddhub_PP_sat __builtin_HEXAGON_A5_vaddhubs + +/* ========================================================================== + Assembly Syntax: Pd4=all8(Ps4) + C Intrinsic Prototype: Byte Q6_p_all8_p(Byte Ps) + Instruction Type: CR + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_all8_p __builtin_HEXAGON_C2_all8 + +/* ========================================================================== + Assembly Syntax: Pd4=and(Pt4,Ps4) + C Intrinsic Prototype: Byte Q6_p_and_pp(Byte Pt, Byte Ps) + Instruction Type: CR + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_and_pp __builtin_HEXAGON_C2_and + +/* ========================================================================== + Assembly Syntax: Pd4=and(Pt4,!Ps4) + C Intrinsic Prototype: Byte Q6_p_and_pnp(Byte Pt, Byte Ps) + Instruction Type: CR + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_and_pnp __builtin_HEXAGON_C2_andn + +/* ========================================================================== + Assembly Syntax: Pd4=any8(Ps4) + C Intrinsic Prototype: Byte Q6_p_any8_p(Byte Ps) + Instruction Type: CR + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_any8_p __builtin_HEXAGON_C2_any8 + +/* ========================================================================== + Assembly Syntax: Pd4=bitsclr(Rs32,Rt32) + C Intrinsic Prototype: Byte Q6_p_bitsclr_RR(Word32 Rs, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_bitsclr_RR __builtin_HEXAGON_C2_bitsclr + +/* ========================================================================== + Assembly Syntax: Pd4=bitsclr(Rs32,#u6) + C Intrinsic Prototype: Byte Q6_p_bitsclr_RI(Word32 Rs, Word32 Iu6) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_bitsclr_RI __builtin_HEXAGON_C2_bitsclri + +/* ========================================================================== + Assembly Syntax: Pd4=bitsset(Rs32,Rt32) + C Intrinsic Prototype: Byte Q6_p_bitsset_RR(Word32 Rs, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_bitsset_RR __builtin_HEXAGON_C2_bitsset + +/* ========================================================================== + Assembly Syntax: Pd4=cmp.eq(Rs32,Rt32) + C Intrinsic Prototype: Byte Q6_p_cmp_eq_RR(Word32 Rs, Word32 Rt) + Instruction Type: ALU32_3op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_p_cmp_eq_RR __builtin_HEXAGON_C2_cmpeq + +/* ========================================================================== + Assembly Syntax: Pd4=cmp.eq(Rs32,#s10) + C Intrinsic Prototype: Byte Q6_p_cmp_eq_RI(Word32 Rs, Word32 Is10) + Instruction Type: ALU32_2op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_p_cmp_eq_RI __builtin_HEXAGON_C2_cmpeqi + +/* ========================================================================== + Assembly Syntax: Pd4=cmp.eq(Rss32,Rtt32) + C Intrinsic Prototype: Byte Q6_p_cmp_eq_PP(Word64 Rss, Word64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_cmp_eq_PP __builtin_HEXAGON_C2_cmpeqp + +/* ========================================================================== + Assembly Syntax: Pd4=cmp.ge(Rs32,#s8) + C Intrinsic Prototype: Byte Q6_p_cmp_ge_RI(Word32 Rs, Word32 Is8) + Instruction Type: ALU32_2op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_p_cmp_ge_RI __builtin_HEXAGON_C2_cmpgei + +/* ========================================================================== + Assembly Syntax: Pd4=cmp.geu(Rs32,#u8) + C Intrinsic Prototype: Byte Q6_p_cmp_geu_RI(Word32 Rs, Word32 Iu8) + Instruction Type: ALU32_2op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_p_cmp_geu_RI __builtin_HEXAGON_C2_cmpgeui + +/* ========================================================================== + Assembly Syntax: Pd4=cmp.gt(Rs32,Rt32) + C Intrinsic Prototype: Byte Q6_p_cmp_gt_RR(Word32 Rs, Word32 Rt) + Instruction Type: ALU32_3op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_p_cmp_gt_RR __builtin_HEXAGON_C2_cmpgt + +/* ========================================================================== + Assembly Syntax: Pd4=cmp.gt(Rs32,#s10) + C Intrinsic Prototype: Byte Q6_p_cmp_gt_RI(Word32 Rs, Word32 Is10) + Instruction Type: ALU32_2op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_p_cmp_gt_RI __builtin_HEXAGON_C2_cmpgti + +/* ========================================================================== + Assembly Syntax: Pd4=cmp.gt(Rss32,Rtt32) + C Intrinsic Prototype: Byte Q6_p_cmp_gt_PP(Word64 Rss, Word64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_cmp_gt_PP __builtin_HEXAGON_C2_cmpgtp + +/* ========================================================================== + Assembly Syntax: Pd4=cmp.gtu(Rs32,Rt32) + C Intrinsic Prototype: Byte Q6_p_cmp_gtu_RR(Word32 Rs, Word32 Rt) + Instruction Type: ALU32_3op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_p_cmp_gtu_RR __builtin_HEXAGON_C2_cmpgtu + +/* ========================================================================== + Assembly Syntax: Pd4=cmp.gtu(Rs32,#u9) + C Intrinsic Prototype: Byte Q6_p_cmp_gtu_RI(Word32 Rs, Word32 Iu9) + Instruction Type: ALU32_2op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_p_cmp_gtu_RI __builtin_HEXAGON_C2_cmpgtui + +/* ========================================================================== + Assembly Syntax: Pd4=cmp.gtu(Rss32,Rtt32) + C Intrinsic Prototype: Byte Q6_p_cmp_gtu_PP(Word64 Rss, Word64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_cmp_gtu_PP __builtin_HEXAGON_C2_cmpgtup + +/* ========================================================================== + Assembly Syntax: Pd4=cmp.lt(Rs32,Rt32) + C Intrinsic Prototype: Byte Q6_p_cmp_lt_RR(Word32 Rs, Word32 Rt) + Instruction Type: ALU32_3op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_p_cmp_lt_RR __builtin_HEXAGON_C2_cmplt + +/* ========================================================================== + Assembly Syntax: Pd4=cmp.ltu(Rs32,Rt32) + C Intrinsic Prototype: Byte Q6_p_cmp_ltu_RR(Word32 Rs, Word32 Rt) + Instruction Type: ALU32_3op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_p_cmp_ltu_RR __builtin_HEXAGON_C2_cmpltu + +/* ========================================================================== + Assembly Syntax: Rdd32=mask(Pt4) + C Intrinsic Prototype: Word64 Q6_P_mask_p(Byte Pt) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mask_p __builtin_HEXAGON_C2_mask + +/* ========================================================================== + Assembly Syntax: Rd32=mux(Pu4,Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_mux_pRR(Byte Pu, Word32 Rs, Word32 Rt) + Instruction Type: ALU32_3op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_mux_pRR __builtin_HEXAGON_C2_mux + +/* ========================================================================== + Assembly Syntax: Rd32=mux(Pu4,#s8,#S8) + C Intrinsic Prototype: Word32 Q6_R_mux_pII(Byte Pu, Word32 Is8, Word32 IS8) + Instruction Type: ALU32_2op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_mux_pII __builtin_HEXAGON_C2_muxii + +/* ========================================================================== + Assembly Syntax: Rd32=mux(Pu4,Rs32,#s8) + C Intrinsic Prototype: Word32 Q6_R_mux_pRI(Byte Pu, Word32 Rs, Word32 Is8) + Instruction Type: ALU32_2op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_mux_pRI __builtin_HEXAGON_C2_muxir + +/* ========================================================================== + Assembly Syntax: Rd32=mux(Pu4,#s8,Rs32) + C Intrinsic Prototype: Word32 Q6_R_mux_pIR(Byte Pu, Word32 Is8, Word32 Rs) + Instruction Type: ALU32_2op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_mux_pIR __builtin_HEXAGON_C2_muxri + +/* ========================================================================== + Assembly Syntax: Pd4=not(Ps4) + C Intrinsic Prototype: Byte Q6_p_not_p(Byte Ps) + Instruction Type: CR + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_not_p __builtin_HEXAGON_C2_not + +/* ========================================================================== + Assembly Syntax: Pd4=or(Pt4,Ps4) + C Intrinsic Prototype: Byte Q6_p_or_pp(Byte Pt, Byte Ps) + Instruction Type: CR + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_or_pp __builtin_HEXAGON_C2_or + +/* ========================================================================== + Assembly Syntax: Pd4=or(Pt4,!Ps4) + C Intrinsic Prototype: Byte Q6_p_or_pnp(Byte Pt, Byte Ps) + Instruction Type: CR + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_or_pnp __builtin_HEXAGON_C2_orn + +/* ========================================================================== + Assembly Syntax: Pd4=Ps4 + C Intrinsic Prototype: Byte Q6_p_equals_p(Byte Ps) + Instruction Type: MAPPING + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_p_equals_p __builtin_HEXAGON_C2_pxfer_map + +/* ========================================================================== + Assembly Syntax: Rd32=Ps4 + C Intrinsic Prototype: Word32 Q6_R_equals_p(Byte Ps) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_equals_p __builtin_HEXAGON_C2_tfrpr + +/* ========================================================================== + Assembly Syntax: Pd4=Rs32 + C Intrinsic Prototype: Byte Q6_p_equals_R(Word32 Rs) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_equals_R __builtin_HEXAGON_C2_tfrrp + +/* ========================================================================== + Assembly Syntax: Rd32=vitpack(Ps4,Pt4) + C Intrinsic Prototype: Word32 Q6_R_vitpack_pp(Byte Ps, Byte Pt) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_vitpack_pp __builtin_HEXAGON_C2_vitpack + +/* ========================================================================== + Assembly Syntax: Rdd32=vmux(Pu4,Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_vmux_pPP(Byte Pu, Word64 Rss, Word64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmux_pPP __builtin_HEXAGON_C2_vmux + +/* ========================================================================== + Assembly Syntax: Pd4=xor(Ps4,Pt4) + C Intrinsic Prototype: Byte Q6_p_xor_pp(Byte Ps, Byte Pt) + Instruction Type: CR + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_xor_pp __builtin_HEXAGON_C2_xor + +/* ========================================================================== + Assembly Syntax: Pd4=and(Ps4,and(Pt4,Pu4)) + C Intrinsic Prototype: Byte Q6_p_and_and_ppp(Byte Ps, Byte Pt, Byte Pu) + Instruction Type: CR + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_and_and_ppp __builtin_HEXAGON_C4_and_and + +/* ========================================================================== + Assembly Syntax: Pd4=and(Ps4,and(Pt4,!Pu4)) + C Intrinsic Prototype: Byte Q6_p_and_and_ppnp(Byte Ps, Byte Pt, Byte Pu) + Instruction Type: CR + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_and_and_ppnp __builtin_HEXAGON_C4_and_andn + +/* ========================================================================== + Assembly Syntax: Pd4=and(Ps4,or(Pt4,Pu4)) + C Intrinsic Prototype: Byte Q6_p_and_or_ppp(Byte Ps, Byte Pt, Byte Pu) + Instruction Type: CR + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_and_or_ppp __builtin_HEXAGON_C4_and_or + +/* ========================================================================== + Assembly Syntax: Pd4=and(Ps4,or(Pt4,!Pu4)) + C Intrinsic Prototype: Byte Q6_p_and_or_ppnp(Byte Ps, Byte Pt, Byte Pu) + Instruction Type: CR + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_and_or_ppnp __builtin_HEXAGON_C4_and_orn + +/* ========================================================================== + Assembly Syntax: Pd4=!cmp.gt(Rs32,Rt32) + C Intrinsic Prototype: Byte Q6_p_not_cmp_gt_RR(Word32 Rs, Word32 Rt) + Instruction Type: ALU32_3op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_p_not_cmp_gt_RR __builtin_HEXAGON_C4_cmplte + +/* ========================================================================== + Assembly Syntax: Pd4=!cmp.gt(Rs32,#s10) + C Intrinsic Prototype: Byte Q6_p_not_cmp_gt_RI(Word32 Rs, Word32 Is10) + Instruction Type: ALU32_2op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_p_not_cmp_gt_RI __builtin_HEXAGON_C4_cmpltei + +/* ========================================================================== + Assembly Syntax: Pd4=!cmp.gtu(Rs32,Rt32) + C Intrinsic Prototype: Byte Q6_p_not_cmp_gtu_RR(Word32 Rs, Word32 Rt) + Instruction Type: ALU32_3op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_p_not_cmp_gtu_RR __builtin_HEXAGON_C4_cmplteu + +/* ========================================================================== + Assembly Syntax: Pd4=!cmp.gtu(Rs32,#u9) + C Intrinsic Prototype: Byte Q6_p_not_cmp_gtu_RI(Word32 Rs, Word32 Iu9) + Instruction Type: ALU32_2op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_p_not_cmp_gtu_RI __builtin_HEXAGON_C4_cmplteui + +/* ========================================================================== + Assembly Syntax: Pd4=!cmp.eq(Rs32,Rt32) + C Intrinsic Prototype: Byte Q6_p_not_cmp_eq_RR(Word32 Rs, Word32 Rt) + Instruction Type: ALU32_3op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_p_not_cmp_eq_RR __builtin_HEXAGON_C4_cmpneq + +/* ========================================================================== + Assembly Syntax: Pd4=!cmp.eq(Rs32,#s10) + C Intrinsic Prototype: Byte Q6_p_not_cmp_eq_RI(Word32 Rs, Word32 Is10) + Instruction Type: ALU32_2op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_p_not_cmp_eq_RI __builtin_HEXAGON_C4_cmpneqi + +/* ========================================================================== + Assembly Syntax: Pd4=fastcorner9(Ps4,Pt4) + C Intrinsic Prototype: Byte Q6_p_fastcorner9_pp(Byte Ps, Byte Pt) + Instruction Type: CR + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_fastcorner9_pp __builtin_HEXAGON_C4_fastcorner9 + +/* ========================================================================== + Assembly Syntax: Pd4=!fastcorner9(Ps4,Pt4) + C Intrinsic Prototype: Byte Q6_p_not_fastcorner9_pp(Byte Ps, Byte Pt) + Instruction Type: CR + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_not_fastcorner9_pp __builtin_HEXAGON_C4_fastcorner9_not + +/* ========================================================================== + Assembly Syntax: Pd4=!bitsclr(Rs32,Rt32) + C Intrinsic Prototype: Byte Q6_p_not_bitsclr_RR(Word32 Rs, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_not_bitsclr_RR __builtin_HEXAGON_C4_nbitsclr + +/* ========================================================================== + Assembly Syntax: Pd4=!bitsclr(Rs32,#u6) + C Intrinsic Prototype: Byte Q6_p_not_bitsclr_RI(Word32 Rs, Word32 Iu6) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_not_bitsclr_RI __builtin_HEXAGON_C4_nbitsclri + +/* ========================================================================== + Assembly Syntax: Pd4=!bitsset(Rs32,Rt32) + C Intrinsic Prototype: Byte Q6_p_not_bitsset_RR(Word32 Rs, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_not_bitsset_RR __builtin_HEXAGON_C4_nbitsset + +/* ========================================================================== + Assembly Syntax: Pd4=or(Ps4,and(Pt4,Pu4)) + C Intrinsic Prototype: Byte Q6_p_or_and_ppp(Byte Ps, Byte Pt, Byte Pu) + Instruction Type: CR + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_or_and_ppp __builtin_HEXAGON_C4_or_and + +/* ========================================================================== + Assembly Syntax: Pd4=or(Ps4,and(Pt4,!Pu4)) + C Intrinsic Prototype: Byte Q6_p_or_and_ppnp(Byte Ps, Byte Pt, Byte Pu) + Instruction Type: CR + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_or_and_ppnp __builtin_HEXAGON_C4_or_andn + +/* ========================================================================== + Assembly Syntax: Pd4=or(Ps4,or(Pt4,Pu4)) + C Intrinsic Prototype: Byte Q6_p_or_or_ppp(Byte Ps, Byte Pt, Byte Pu) + Instruction Type: CR + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_or_or_ppp __builtin_HEXAGON_C4_or_or + +/* ========================================================================== + Assembly Syntax: Pd4=or(Ps4,or(Pt4,!Pu4)) + C Intrinsic Prototype: Byte Q6_p_or_or_ppnp(Byte Ps, Byte Pt, Byte Pu) + Instruction Type: CR + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_or_or_ppnp __builtin_HEXAGON_C4_or_orn + +/* ========================================================================== + Assembly Syntax: Rdd32=convert_d2df(Rss32) + C Intrinsic Prototype: Float64 Q6_P_convert_d2df_P(Word64 Rss) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_convert_d2df_P __builtin_HEXAGON_F2_conv_d2df + +/* ========================================================================== + Assembly Syntax: Rd32=convert_d2sf(Rss32) + C Intrinsic Prototype: Float32 Q6_R_convert_d2sf_P(Word64 Rss) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_convert_d2sf_P __builtin_HEXAGON_F2_conv_d2sf + +/* ========================================================================== + Assembly Syntax: Rdd32=convert_df2d(Rss32) + C Intrinsic Prototype: Word64 Q6_P_convert_df2d_P(Float64 Rss) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_convert_df2d_P __builtin_HEXAGON_F2_conv_df2d + +/* ========================================================================== + Assembly Syntax: Rdd32=convert_df2d(Rss32):chop + C Intrinsic Prototype: Word64 Q6_P_convert_df2d_P_chop(Float64 Rss) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_convert_df2d_P_chop __builtin_HEXAGON_F2_conv_df2d_chop + +/* ========================================================================== + Assembly Syntax: Rd32=convert_df2sf(Rss32) + C Intrinsic Prototype: Float32 Q6_R_convert_df2sf_P(Float64 Rss) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_convert_df2sf_P __builtin_HEXAGON_F2_conv_df2sf + +/* ========================================================================== + Assembly Syntax: Rdd32=convert_df2ud(Rss32) + C Intrinsic Prototype: Word64 Q6_P_convert_df2ud_P(Float64 Rss) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_convert_df2ud_P __builtin_HEXAGON_F2_conv_df2ud + +/* ========================================================================== + Assembly Syntax: Rdd32=convert_df2ud(Rss32):chop + C Intrinsic Prototype: Word64 Q6_P_convert_df2ud_P_chop(Float64 Rss) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_convert_df2ud_P_chop __builtin_HEXAGON_F2_conv_df2ud_chop + +/* ========================================================================== + Assembly Syntax: Rd32=convert_df2uw(Rss32) + C Intrinsic Prototype: Word32 Q6_R_convert_df2uw_P(Float64 Rss) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_convert_df2uw_P __builtin_HEXAGON_F2_conv_df2uw + +/* ========================================================================== + Assembly Syntax: Rd32=convert_df2uw(Rss32):chop + C Intrinsic Prototype: Word32 Q6_R_convert_df2uw_P_chop(Float64 Rss) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_convert_df2uw_P_chop __builtin_HEXAGON_F2_conv_df2uw_chop + +/* ========================================================================== + Assembly Syntax: Rd32=convert_df2w(Rss32) + C Intrinsic Prototype: Word32 Q6_R_convert_df2w_P(Float64 Rss) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_convert_df2w_P __builtin_HEXAGON_F2_conv_df2w + +/* ========================================================================== + Assembly Syntax: Rd32=convert_df2w(Rss32):chop + C Intrinsic Prototype: Word32 Q6_R_convert_df2w_P_chop(Float64 Rss) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_convert_df2w_P_chop __builtin_HEXAGON_F2_conv_df2w_chop + +/* ========================================================================== + Assembly Syntax: Rdd32=convert_sf2d(Rs32) + C Intrinsic Prototype: Word64 Q6_P_convert_sf2d_R(Float32 Rs) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_convert_sf2d_R __builtin_HEXAGON_F2_conv_sf2d + +/* ========================================================================== + Assembly Syntax: Rdd32=convert_sf2d(Rs32):chop + C Intrinsic Prototype: Word64 Q6_P_convert_sf2d_R_chop(Float32 Rs) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_convert_sf2d_R_chop __builtin_HEXAGON_F2_conv_sf2d_chop + +/* ========================================================================== + Assembly Syntax: Rdd32=convert_sf2df(Rs32) + C Intrinsic Prototype: Float64 Q6_P_convert_sf2df_R(Float32 Rs) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_convert_sf2df_R __builtin_HEXAGON_F2_conv_sf2df + +/* ========================================================================== + Assembly Syntax: Rdd32=convert_sf2ud(Rs32) + C Intrinsic Prototype: Word64 Q6_P_convert_sf2ud_R(Float32 Rs) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_convert_sf2ud_R __builtin_HEXAGON_F2_conv_sf2ud + +/* ========================================================================== + Assembly Syntax: Rdd32=convert_sf2ud(Rs32):chop + C Intrinsic Prototype: Word64 Q6_P_convert_sf2ud_R_chop(Float32 Rs) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_convert_sf2ud_R_chop __builtin_HEXAGON_F2_conv_sf2ud_chop + +/* ========================================================================== + Assembly Syntax: Rd32=convert_sf2uw(Rs32) + C Intrinsic Prototype: Word32 Q6_R_convert_sf2uw_R(Float32 Rs) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_convert_sf2uw_R __builtin_HEXAGON_F2_conv_sf2uw + +/* ========================================================================== + Assembly Syntax: Rd32=convert_sf2uw(Rs32):chop + C Intrinsic Prototype: Word32 Q6_R_convert_sf2uw_R_chop(Float32 Rs) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_convert_sf2uw_R_chop __builtin_HEXAGON_F2_conv_sf2uw_chop + +/* ========================================================================== + Assembly Syntax: Rd32=convert_sf2w(Rs32) + C Intrinsic Prototype: Word32 Q6_R_convert_sf2w_R(Float32 Rs) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_convert_sf2w_R __builtin_HEXAGON_F2_conv_sf2w + +/* ========================================================================== + Assembly Syntax: Rd32=convert_sf2w(Rs32):chop + C Intrinsic Prototype: Word32 Q6_R_convert_sf2w_R_chop(Float32 Rs) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_convert_sf2w_R_chop __builtin_HEXAGON_F2_conv_sf2w_chop + +/* ========================================================================== + Assembly Syntax: Rdd32=convert_ud2df(Rss32) + C Intrinsic Prototype: Float64 Q6_P_convert_ud2df_P(Word64 Rss) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_convert_ud2df_P __builtin_HEXAGON_F2_conv_ud2df + +/* ========================================================================== + Assembly Syntax: Rd32=convert_ud2sf(Rss32) + C Intrinsic Prototype: Float32 Q6_R_convert_ud2sf_P(Word64 Rss) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_convert_ud2sf_P __builtin_HEXAGON_F2_conv_ud2sf + +/* ========================================================================== + Assembly Syntax: Rdd32=convert_uw2df(Rs32) + C Intrinsic Prototype: Float64 Q6_P_convert_uw2df_R(Word32 Rs) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_convert_uw2df_R __builtin_HEXAGON_F2_conv_uw2df + +/* ========================================================================== + Assembly Syntax: Rd32=convert_uw2sf(Rs32) + C Intrinsic Prototype: Float32 Q6_R_convert_uw2sf_R(Word32 Rs) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_convert_uw2sf_R __builtin_HEXAGON_F2_conv_uw2sf + +/* ========================================================================== + Assembly Syntax: Rdd32=convert_w2df(Rs32) + C Intrinsic Prototype: Float64 Q6_P_convert_w2df_R(Word32 Rs) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_convert_w2df_R __builtin_HEXAGON_F2_conv_w2df + +/* ========================================================================== + Assembly Syntax: Rd32=convert_w2sf(Rs32) + C Intrinsic Prototype: Float32 Q6_R_convert_w2sf_R(Word32 Rs) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_convert_w2sf_R __builtin_HEXAGON_F2_conv_w2sf + +/* ========================================================================== + Assembly Syntax: Pd4=dfclass(Rss32,#u5) + C Intrinsic Prototype: Byte Q6_p_dfclass_PI(Float64 Rss, Word32 Iu5) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_dfclass_PI __builtin_HEXAGON_F2_dfclass + +/* ========================================================================== + Assembly Syntax: Pd4=dfcmp.eq(Rss32,Rtt32) + C Intrinsic Prototype: Byte Q6_p_dfcmp_eq_PP(Float64 Rss, Float64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_dfcmp_eq_PP __builtin_HEXAGON_F2_dfcmpeq + +/* ========================================================================== + Assembly Syntax: Pd4=dfcmp.ge(Rss32,Rtt32) + C Intrinsic Prototype: Byte Q6_p_dfcmp_ge_PP(Float64 Rss, Float64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_dfcmp_ge_PP __builtin_HEXAGON_F2_dfcmpge + +/* ========================================================================== + Assembly Syntax: Pd4=dfcmp.gt(Rss32,Rtt32) + C Intrinsic Prototype: Byte Q6_p_dfcmp_gt_PP(Float64 Rss, Float64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_dfcmp_gt_PP __builtin_HEXAGON_F2_dfcmpgt + +/* ========================================================================== + Assembly Syntax: Pd4=dfcmp.uo(Rss32,Rtt32) + C Intrinsic Prototype: Byte Q6_p_dfcmp_uo_PP(Float64 Rss, Float64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_dfcmp_uo_PP __builtin_HEXAGON_F2_dfcmpuo + +/* ========================================================================== + Assembly Syntax: Rdd32=dfmake(#u10):neg + C Intrinsic Prototype: Float64 Q6_P_dfmake_I_neg(Word32 Iu10) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_dfmake_I_neg __builtin_HEXAGON_F2_dfimm_n + +/* ========================================================================== + Assembly Syntax: Rdd32=dfmake(#u10):pos + C Intrinsic Prototype: Float64 Q6_P_dfmake_I_pos(Word32 Iu10) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_dfmake_I_pos __builtin_HEXAGON_F2_dfimm_p + +/* ========================================================================== + Assembly Syntax: Rd32=sfadd(Rs32,Rt32) + C Intrinsic Prototype: Float32 Q6_R_sfadd_RR(Float32 Rs, Float32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_sfadd_RR __builtin_HEXAGON_F2_sfadd + +/* ========================================================================== + Assembly Syntax: Pd4=sfclass(Rs32,#u5) + C Intrinsic Prototype: Byte Q6_p_sfclass_RI(Float32 Rs, Word32 Iu5) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_sfclass_RI __builtin_HEXAGON_F2_sfclass + +/* ========================================================================== + Assembly Syntax: Pd4=sfcmp.eq(Rs32,Rt32) + C Intrinsic Prototype: Byte Q6_p_sfcmp_eq_RR(Float32 Rs, Float32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_sfcmp_eq_RR __builtin_HEXAGON_F2_sfcmpeq + +/* ========================================================================== + Assembly Syntax: Pd4=sfcmp.ge(Rs32,Rt32) + C Intrinsic Prototype: Byte Q6_p_sfcmp_ge_RR(Float32 Rs, Float32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_sfcmp_ge_RR __builtin_HEXAGON_F2_sfcmpge + +/* ========================================================================== + Assembly Syntax: Pd4=sfcmp.gt(Rs32,Rt32) + C Intrinsic Prototype: Byte Q6_p_sfcmp_gt_RR(Float32 Rs, Float32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_sfcmp_gt_RR __builtin_HEXAGON_F2_sfcmpgt + +/* ========================================================================== + Assembly Syntax: Pd4=sfcmp.uo(Rs32,Rt32) + C Intrinsic Prototype: Byte Q6_p_sfcmp_uo_RR(Float32 Rs, Float32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_sfcmp_uo_RR __builtin_HEXAGON_F2_sfcmpuo + +/* ========================================================================== + Assembly Syntax: Rd32=sffixupd(Rs32,Rt32) + C Intrinsic Prototype: Float32 Q6_R_sffixupd_RR(Float32 Rs, Float32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_sffixupd_RR __builtin_HEXAGON_F2_sffixupd + +/* ========================================================================== + Assembly Syntax: Rd32=sffixupn(Rs32,Rt32) + C Intrinsic Prototype: Float32 Q6_R_sffixupn_RR(Float32 Rs, Float32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_sffixupn_RR __builtin_HEXAGON_F2_sffixupn + +/* ========================================================================== + Assembly Syntax: Rd32=sffixupr(Rs32) + C Intrinsic Prototype: Float32 Q6_R_sffixupr_R(Float32 Rs) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_sffixupr_R __builtin_HEXAGON_F2_sffixupr + +/* ========================================================================== + Assembly Syntax: Rx32+=sfmpy(Rs32,Rt32) + C Intrinsic Prototype: Float32 Q6_R_sfmpyacc_RR(Float32 Rx, Float32 Rs, Float32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_sfmpyacc_RR __builtin_HEXAGON_F2_sffma + +/* ========================================================================== + Assembly Syntax: Rx32+=sfmpy(Rs32,Rt32):lib + C Intrinsic Prototype: Float32 Q6_R_sfmpyacc_RR_lib(Float32 Rx, Float32 Rs, Float32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_sfmpyacc_RR_lib __builtin_HEXAGON_F2_sffma_lib + +/* ========================================================================== + Assembly Syntax: Rx32+=sfmpy(Rs32,Rt32,Pu4):scale + C Intrinsic Prototype: Float32 Q6_R_sfmpyacc_RRp_scale(Float32 Rx, Float32 Rs, Float32 Rt, Byte Pu) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_sfmpyacc_RRp_scale __builtin_HEXAGON_F2_sffma_sc + +/* ========================================================================== + Assembly Syntax: Rx32-=sfmpy(Rs32,Rt32) + C Intrinsic Prototype: Float32 Q6_R_sfmpynac_RR(Float32 Rx, Float32 Rs, Float32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_sfmpynac_RR __builtin_HEXAGON_F2_sffms + +/* ========================================================================== + Assembly Syntax: Rx32-=sfmpy(Rs32,Rt32):lib + C Intrinsic Prototype: Float32 Q6_R_sfmpynac_RR_lib(Float32 Rx, Float32 Rs, Float32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_sfmpynac_RR_lib __builtin_HEXAGON_F2_sffms_lib + +/* ========================================================================== + Assembly Syntax: Rd32=sfmake(#u10):neg + C Intrinsic Prototype: Float32 Q6_R_sfmake_I_neg(Word32 Iu10) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_sfmake_I_neg __builtin_HEXAGON_F2_sfimm_n + +/* ========================================================================== + Assembly Syntax: Rd32=sfmake(#u10):pos + C Intrinsic Prototype: Float32 Q6_R_sfmake_I_pos(Word32 Iu10) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_sfmake_I_pos __builtin_HEXAGON_F2_sfimm_p + +/* ========================================================================== + Assembly Syntax: Rd32=sfmax(Rs32,Rt32) + C Intrinsic Prototype: Float32 Q6_R_sfmax_RR(Float32 Rs, Float32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_sfmax_RR __builtin_HEXAGON_F2_sfmax + +/* ========================================================================== + Assembly Syntax: Rd32=sfmin(Rs32,Rt32) + C Intrinsic Prototype: Float32 Q6_R_sfmin_RR(Float32 Rs, Float32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_sfmin_RR __builtin_HEXAGON_F2_sfmin + +/* ========================================================================== + Assembly Syntax: Rd32=sfmpy(Rs32,Rt32) + C Intrinsic Prototype: Float32 Q6_R_sfmpy_RR(Float32 Rs, Float32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_sfmpy_RR __builtin_HEXAGON_F2_sfmpy + +/* ========================================================================== + Assembly Syntax: Rd32=sfsub(Rs32,Rt32) + C Intrinsic Prototype: Float32 Q6_R_sfsub_RR(Float32 Rs, Float32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_sfsub_RR __builtin_HEXAGON_F2_sfsub + +/* ========================================================================== + Assembly Syntax: Rd32=memb(Rx32++#s4:0:circ(Mu2)) + C Intrinsic Prototype: Word32 Q6_R_memb_IM_circ(void** Rx, Word32 Is4_0, Word32 Mu, void* BaseAddress) + Instruction Type: LD + Execution Slots: SLOT01 + ========================================================================== */ + +#define Q6_R_memb_IM_circ __builtin_HEXAGON_L2_loadrb_pci + +/* ========================================================================== + Assembly Syntax: Rd32=memb(Rx32++I:circ(Mu2)) + C Intrinsic Prototype: Word32 Q6_R_memb_M_circ(void** Rx, Word32 Mu, void* BaseAddress) + Instruction Type: LD + Execution Slots: SLOT01 + ========================================================================== */ + +#define Q6_R_memb_M_circ __builtin_HEXAGON_L2_loadrb_pcr + +/* ========================================================================== + Assembly Syntax: Rdd32=memd(Rx32++#s4:3:circ(Mu2)) + C Intrinsic Prototype: Word64 Q6_P_memd_IM_circ(void** Rx, Word32 Is4_3, Word32 Mu, void* BaseAddress) + Instruction Type: LD + Execution Slots: SLOT01 + ========================================================================== */ + +#define Q6_P_memd_IM_circ __builtin_HEXAGON_L2_loadrd_pci + +/* ========================================================================== + Assembly Syntax: Rdd32=memd(Rx32++I:circ(Mu2)) + C Intrinsic Prototype: Word64 Q6_P_memd_M_circ(void** Rx, Word32 Mu, void* BaseAddress) + Instruction Type: LD + Execution Slots: SLOT01 + ========================================================================== */ + +#define Q6_P_memd_M_circ __builtin_HEXAGON_L2_loadrd_pcr + +/* ========================================================================== + Assembly Syntax: Rd32=memh(Rx32++#s4:1:circ(Mu2)) + C Intrinsic Prototype: Word32 Q6_R_memh_IM_circ(void** Rx, Word32 Is4_1, Word32 Mu, void* BaseAddress) + Instruction Type: LD + Execution Slots: SLOT01 + ========================================================================== */ + +#define Q6_R_memh_IM_circ __builtin_HEXAGON_L2_loadrh_pci + +/* ========================================================================== + Assembly Syntax: Rd32=memh(Rx32++I:circ(Mu2)) + C Intrinsic Prototype: Word32 Q6_R_memh_M_circ(void** Rx, Word32 Mu, void* BaseAddress) + Instruction Type: LD + Execution Slots: SLOT01 + ========================================================================== */ + +#define Q6_R_memh_M_circ __builtin_HEXAGON_L2_loadrh_pcr + +/* ========================================================================== + Assembly Syntax: Rd32=memw(Rx32++#s4:2:circ(Mu2)) + C Intrinsic Prototype: Word32 Q6_R_memw_IM_circ(void** Rx, Word32 Is4_2, Word32 Mu, void* BaseAddress) + Instruction Type: LD + Execution Slots: SLOT01 + ========================================================================== */ + +#define Q6_R_memw_IM_circ __builtin_HEXAGON_L2_loadri_pci + +/* ========================================================================== + Assembly Syntax: Rd32=memw(Rx32++I:circ(Mu2)) + C Intrinsic Prototype: Word32 Q6_R_memw_M_circ(void** Rx, Word32 Mu, void* BaseAddress) + Instruction Type: LD + Execution Slots: SLOT01 + ========================================================================== */ + +#define Q6_R_memw_M_circ __builtin_HEXAGON_L2_loadri_pcr + +/* ========================================================================== + Assembly Syntax: Rd32=memub(Rx32++#s4:0:circ(Mu2)) + C Intrinsic Prototype: Word32 Q6_R_memub_IM_circ(void** Rx, Word32 Is4_0, Word32 Mu, void* BaseAddress) + Instruction Type: LD + Execution Slots: SLOT01 + ========================================================================== */ + +#define Q6_R_memub_IM_circ __builtin_HEXAGON_L2_loadrub_pci + +/* ========================================================================== + Assembly Syntax: Rd32=memub(Rx32++I:circ(Mu2)) + C Intrinsic Prototype: Word32 Q6_R_memub_M_circ(void** Rx, Word32 Mu, void* BaseAddress) + Instruction Type: LD + Execution Slots: SLOT01 + ========================================================================== */ + +#define Q6_R_memub_M_circ __builtin_HEXAGON_L2_loadrub_pcr + +/* ========================================================================== + Assembly Syntax: Rd32=memuh(Rx32++#s4:1:circ(Mu2)) + C Intrinsic Prototype: Word32 Q6_R_memuh_IM_circ(void** Rx, Word32 Is4_1, Word32 Mu, void* BaseAddress) + Instruction Type: LD + Execution Slots: SLOT01 + ========================================================================== */ + +#define Q6_R_memuh_IM_circ __builtin_HEXAGON_L2_loadruh_pci + +/* ========================================================================== + Assembly Syntax: Rd32=memuh(Rx32++I:circ(Mu2)) + C Intrinsic Prototype: Word32 Q6_R_memuh_M_circ(void** Rx, Word32 Mu, void* BaseAddress) + Instruction Type: LD + Execution Slots: SLOT01 + ========================================================================== */ + +#define Q6_R_memuh_M_circ __builtin_HEXAGON_L2_loadruh_pcr + +/* ========================================================================== + Assembly Syntax: Rx32+=add(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_addacc_RR(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_addacc_RR __builtin_HEXAGON_M2_acci + +/* ========================================================================== + Assembly Syntax: Rx32+=add(Rs32,#s8) + C Intrinsic Prototype: Word32 Q6_R_addacc_RI(Word32 Rx, Word32 Rs, Word32 Is8) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_addacc_RI __builtin_HEXAGON_M2_accii + +/* ========================================================================== + Assembly Syntax: Rxx32+=cmpyi(Rs32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_cmpyiacc_RR(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_cmpyiacc_RR __builtin_HEXAGON_M2_cmaci_s0 + +/* ========================================================================== + Assembly Syntax: Rxx32+=cmpyr(Rs32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_cmpyracc_RR(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_cmpyracc_RR __builtin_HEXAGON_M2_cmacr_s0 + +/* ========================================================================== + Assembly Syntax: Rxx32+=cmpy(Rs32,Rt32):sat + C Intrinsic Prototype: Word64 Q6_P_cmpyacc_RR_sat(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_cmpyacc_RR_sat __builtin_HEXAGON_M2_cmacs_s0 + +/* ========================================================================== + Assembly Syntax: Rxx32+=cmpy(Rs32,Rt32):<<1:sat + C Intrinsic Prototype: Word64 Q6_P_cmpyacc_RR_s1_sat(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_cmpyacc_RR_s1_sat __builtin_HEXAGON_M2_cmacs_s1 + +/* ========================================================================== + Assembly Syntax: Rxx32+=cmpy(Rs32,Rt32*):sat + C Intrinsic Prototype: Word64 Q6_P_cmpyacc_RR_conj_sat(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_cmpyacc_RR_conj_sat __builtin_HEXAGON_M2_cmacsc_s0 + +/* ========================================================================== + Assembly Syntax: Rxx32+=cmpy(Rs32,Rt32*):<<1:sat + C Intrinsic Prototype: Word64 Q6_P_cmpyacc_RR_conj_s1_sat(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_cmpyacc_RR_conj_s1_sat __builtin_HEXAGON_M2_cmacsc_s1 + +/* ========================================================================== + Assembly Syntax: Rdd32=cmpyi(Rs32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_cmpyi_RR(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_cmpyi_RR __builtin_HEXAGON_M2_cmpyi_s0 + +/* ========================================================================== + Assembly Syntax: Rdd32=cmpyr(Rs32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_cmpyr_RR(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_cmpyr_RR __builtin_HEXAGON_M2_cmpyr_s0 + +/* ========================================================================== + Assembly Syntax: Rd32=cmpy(Rs32,Rt32):rnd:sat + C Intrinsic Prototype: Word32 Q6_R_cmpy_RR_rnd_sat(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_cmpy_RR_rnd_sat __builtin_HEXAGON_M2_cmpyrs_s0 + +/* ========================================================================== + Assembly Syntax: Rd32=cmpy(Rs32,Rt32):<<1:rnd:sat + C Intrinsic Prototype: Word32 Q6_R_cmpy_RR_s1_rnd_sat(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_cmpy_RR_s1_rnd_sat __builtin_HEXAGON_M2_cmpyrs_s1 + +/* ========================================================================== + Assembly Syntax: Rd32=cmpy(Rs32,Rt32*):rnd:sat + C Intrinsic Prototype: Word32 Q6_R_cmpy_RR_conj_rnd_sat(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_cmpy_RR_conj_rnd_sat __builtin_HEXAGON_M2_cmpyrsc_s0 + +/* ========================================================================== + Assembly Syntax: Rd32=cmpy(Rs32,Rt32*):<<1:rnd:sat + C Intrinsic Prototype: Word32 Q6_R_cmpy_RR_conj_s1_rnd_sat(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_cmpy_RR_conj_s1_rnd_sat __builtin_HEXAGON_M2_cmpyrsc_s1 + +/* ========================================================================== + Assembly Syntax: Rdd32=cmpy(Rs32,Rt32):sat + C Intrinsic Prototype: Word64 Q6_P_cmpy_RR_sat(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_cmpy_RR_sat __builtin_HEXAGON_M2_cmpys_s0 + +/* ========================================================================== + Assembly Syntax: Rdd32=cmpy(Rs32,Rt32):<<1:sat + C Intrinsic Prototype: Word64 Q6_P_cmpy_RR_s1_sat(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_cmpy_RR_s1_sat __builtin_HEXAGON_M2_cmpys_s1 + +/* ========================================================================== + Assembly Syntax: Rdd32=cmpy(Rs32,Rt32*):sat + C Intrinsic Prototype: Word64 Q6_P_cmpy_RR_conj_sat(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_cmpy_RR_conj_sat __builtin_HEXAGON_M2_cmpysc_s0 + +/* ========================================================================== + Assembly Syntax: Rdd32=cmpy(Rs32,Rt32*):<<1:sat + C Intrinsic Prototype: Word64 Q6_P_cmpy_RR_conj_s1_sat(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_cmpy_RR_conj_s1_sat __builtin_HEXAGON_M2_cmpysc_s1 + +/* ========================================================================== + Assembly Syntax: Rxx32-=cmpy(Rs32,Rt32):sat + C Intrinsic Prototype: Word64 Q6_P_cmpynac_RR_sat(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_cmpynac_RR_sat __builtin_HEXAGON_M2_cnacs_s0 + +/* ========================================================================== + Assembly Syntax: Rxx32-=cmpy(Rs32,Rt32):<<1:sat + C Intrinsic Prototype: Word64 Q6_P_cmpynac_RR_s1_sat(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_cmpynac_RR_s1_sat __builtin_HEXAGON_M2_cnacs_s1 + +/* ========================================================================== + Assembly Syntax: Rxx32-=cmpy(Rs32,Rt32*):sat + C Intrinsic Prototype: Word64 Q6_P_cmpynac_RR_conj_sat(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_cmpynac_RR_conj_sat __builtin_HEXAGON_M2_cnacsc_s0 + +/* ========================================================================== + Assembly Syntax: Rxx32-=cmpy(Rs32,Rt32*):<<1:sat + C Intrinsic Prototype: Word64 Q6_P_cmpynac_RR_conj_s1_sat(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_cmpynac_RR_conj_s1_sat __builtin_HEXAGON_M2_cnacsc_s1 + +/* ========================================================================== + Assembly Syntax: Rxx32+=mpy(Rs32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_mpyacc_RR(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpyacc_RR __builtin_HEXAGON_M2_dpmpyss_acc_s0 + +/* ========================================================================== + Assembly Syntax: Rxx32-=mpy(Rs32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_mpynac_RR(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpynac_RR __builtin_HEXAGON_M2_dpmpyss_nac_s0 + +/* ========================================================================== + Assembly Syntax: Rd32=mpy(Rs32,Rt32):rnd + C Intrinsic Prototype: Word32 Q6_R_mpy_RR_rnd(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpy_RR_rnd __builtin_HEXAGON_M2_dpmpyss_rnd_s0 + +/* ========================================================================== + Assembly Syntax: Rdd32=mpy(Rs32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_mpy_RR(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpy_RR __builtin_HEXAGON_M2_dpmpyss_s0 + +/* ========================================================================== + Assembly Syntax: Rxx32+=mpyu(Rs32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_mpyuacc_RR(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpyuacc_RR __builtin_HEXAGON_M2_dpmpyuu_acc_s0 + +/* ========================================================================== + Assembly Syntax: Rxx32-=mpyu(Rs32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_mpyunac_RR(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpyunac_RR __builtin_HEXAGON_M2_dpmpyuu_nac_s0 + +/* ========================================================================== + Assembly Syntax: Rdd32=mpyu(Rs32,Rt32) + C Intrinsic Prototype: UWord64 Q6_P_mpyu_RR(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpyu_RR __builtin_HEXAGON_M2_dpmpyuu_s0 + +/* ========================================================================== + Assembly Syntax: Rd32=mpy(Rs32,Rt32.h):<<1:rnd:sat + C Intrinsic Prototype: Word32 Q6_R_mpy_RRh_s1_rnd_sat(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpy_RRh_s1_rnd_sat __builtin_HEXAGON_M2_hmmpyh_rs1 + +/* ========================================================================== + Assembly Syntax: Rd32=mpy(Rs32,Rt32.h):<<1:sat + C Intrinsic Prototype: Word32 Q6_R_mpy_RRh_s1_sat(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpy_RRh_s1_sat __builtin_HEXAGON_M2_hmmpyh_s1 + +/* ========================================================================== + Assembly Syntax: Rd32=mpy(Rs32,Rt32.l):<<1:rnd:sat + C Intrinsic Prototype: Word32 Q6_R_mpy_RRl_s1_rnd_sat(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpy_RRl_s1_rnd_sat __builtin_HEXAGON_M2_hmmpyl_rs1 + +/* ========================================================================== + Assembly Syntax: Rd32=mpy(Rs32,Rt32.l):<<1:sat + C Intrinsic Prototype: Word32 Q6_R_mpy_RRl_s1_sat(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpy_RRl_s1_sat __builtin_HEXAGON_M2_hmmpyl_s1 + +/* ========================================================================== + Assembly Syntax: Rx32+=mpyi(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_mpyiacc_RR(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyiacc_RR __builtin_HEXAGON_M2_maci + +/* ========================================================================== + Assembly Syntax: Rx32-=mpyi(Rs32,#u8) + C Intrinsic Prototype: Word32 Q6_R_mpyinac_RI(Word32 Rx, Word32 Rs, Word32 Iu8) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyinac_RI __builtin_HEXAGON_M2_macsin + +/* ========================================================================== + Assembly Syntax: Rx32+=mpyi(Rs32,#u8) + C Intrinsic Prototype: Word32 Q6_R_mpyiacc_RI(Word32 Rx, Word32 Rs, Word32 Iu8) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyiacc_RI __builtin_HEXAGON_M2_macsip + +/* ========================================================================== + Assembly Syntax: Rxx32+=vmpywoh(Rss32,Rtt32):rnd:sat + C Intrinsic Prototype: Word64 Q6_P_vmpywohacc_PP_rnd_sat(Word64 Rxx, Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpywohacc_PP_rnd_sat __builtin_HEXAGON_M2_mmachs_rs0 + +/* ========================================================================== + Assembly Syntax: Rxx32+=vmpywoh(Rss32,Rtt32):<<1:rnd:sat + C Intrinsic Prototype: Word64 Q6_P_vmpywohacc_PP_s1_rnd_sat(Word64 Rxx, Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpywohacc_PP_s1_rnd_sat __builtin_HEXAGON_M2_mmachs_rs1 + +/* ========================================================================== + Assembly Syntax: Rxx32+=vmpywoh(Rss32,Rtt32):sat + C Intrinsic Prototype: Word64 Q6_P_vmpywohacc_PP_sat(Word64 Rxx, Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpywohacc_PP_sat __builtin_HEXAGON_M2_mmachs_s0 + +/* ========================================================================== + Assembly Syntax: Rxx32+=vmpywoh(Rss32,Rtt32):<<1:sat + C Intrinsic Prototype: Word64 Q6_P_vmpywohacc_PP_s1_sat(Word64 Rxx, Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpywohacc_PP_s1_sat __builtin_HEXAGON_M2_mmachs_s1 + +/* ========================================================================== + Assembly Syntax: Rxx32+=vmpyweh(Rss32,Rtt32):rnd:sat + C Intrinsic Prototype: Word64 Q6_P_vmpywehacc_PP_rnd_sat(Word64 Rxx, Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpywehacc_PP_rnd_sat __builtin_HEXAGON_M2_mmacls_rs0 + +/* ========================================================================== + Assembly Syntax: Rxx32+=vmpyweh(Rss32,Rtt32):<<1:rnd:sat + C Intrinsic Prototype: Word64 Q6_P_vmpywehacc_PP_s1_rnd_sat(Word64 Rxx, Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpywehacc_PP_s1_rnd_sat __builtin_HEXAGON_M2_mmacls_rs1 + +/* ========================================================================== + Assembly Syntax: Rxx32+=vmpyweh(Rss32,Rtt32):sat + C Intrinsic Prototype: Word64 Q6_P_vmpywehacc_PP_sat(Word64 Rxx, Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpywehacc_PP_sat __builtin_HEXAGON_M2_mmacls_s0 + +/* ========================================================================== + Assembly Syntax: Rxx32+=vmpyweh(Rss32,Rtt32):<<1:sat + C Intrinsic Prototype: Word64 Q6_P_vmpywehacc_PP_s1_sat(Word64 Rxx, Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpywehacc_PP_s1_sat __builtin_HEXAGON_M2_mmacls_s1 + +/* ========================================================================== + Assembly Syntax: Rxx32+=vmpywouh(Rss32,Rtt32):rnd:sat + C Intrinsic Prototype: Word64 Q6_P_vmpywouhacc_PP_rnd_sat(Word64 Rxx, Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpywouhacc_PP_rnd_sat __builtin_HEXAGON_M2_mmacuhs_rs0 + +/* ========================================================================== + Assembly Syntax: Rxx32+=vmpywouh(Rss32,Rtt32):<<1:rnd:sat + C Intrinsic Prototype: Word64 Q6_P_vmpywouhacc_PP_s1_rnd_sat(Word64 Rxx, Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpywouhacc_PP_s1_rnd_sat __builtin_HEXAGON_M2_mmacuhs_rs1 + +/* ========================================================================== + Assembly Syntax: Rxx32+=vmpywouh(Rss32,Rtt32):sat + C Intrinsic Prototype: Word64 Q6_P_vmpywouhacc_PP_sat(Word64 Rxx, Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpywouhacc_PP_sat __builtin_HEXAGON_M2_mmacuhs_s0 + +/* ========================================================================== + Assembly Syntax: Rxx32+=vmpywouh(Rss32,Rtt32):<<1:sat + C Intrinsic Prototype: Word64 Q6_P_vmpywouhacc_PP_s1_sat(Word64 Rxx, Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpywouhacc_PP_s1_sat __builtin_HEXAGON_M2_mmacuhs_s1 + +/* ========================================================================== + Assembly Syntax: Rxx32+=vmpyweuh(Rss32,Rtt32):rnd:sat + C Intrinsic Prototype: Word64 Q6_P_vmpyweuhacc_PP_rnd_sat(Word64 Rxx, Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpyweuhacc_PP_rnd_sat __builtin_HEXAGON_M2_mmaculs_rs0 + +/* ========================================================================== + Assembly Syntax: Rxx32+=vmpyweuh(Rss32,Rtt32):<<1:rnd:sat + C Intrinsic Prototype: Word64 Q6_P_vmpyweuhacc_PP_s1_rnd_sat(Word64 Rxx, Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpyweuhacc_PP_s1_rnd_sat __builtin_HEXAGON_M2_mmaculs_rs1 + +/* ========================================================================== + Assembly Syntax: Rxx32+=vmpyweuh(Rss32,Rtt32):sat + C Intrinsic Prototype: Word64 Q6_P_vmpyweuhacc_PP_sat(Word64 Rxx, Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpyweuhacc_PP_sat __builtin_HEXAGON_M2_mmaculs_s0 + +/* ========================================================================== + Assembly Syntax: Rxx32+=vmpyweuh(Rss32,Rtt32):<<1:sat + C Intrinsic Prototype: Word64 Q6_P_vmpyweuhacc_PP_s1_sat(Word64 Rxx, Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpyweuhacc_PP_s1_sat __builtin_HEXAGON_M2_mmaculs_s1 + +/* ========================================================================== + Assembly Syntax: Rdd32=vmpywoh(Rss32,Rtt32):rnd:sat + C Intrinsic Prototype: Word64 Q6_P_vmpywoh_PP_rnd_sat(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpywoh_PP_rnd_sat __builtin_HEXAGON_M2_mmpyh_rs0 + +/* ========================================================================== + Assembly Syntax: Rdd32=vmpywoh(Rss32,Rtt32):<<1:rnd:sat + C Intrinsic Prototype: Word64 Q6_P_vmpywoh_PP_s1_rnd_sat(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpywoh_PP_s1_rnd_sat __builtin_HEXAGON_M2_mmpyh_rs1 + +/* ========================================================================== + Assembly Syntax: Rdd32=vmpywoh(Rss32,Rtt32):sat + C Intrinsic Prototype: Word64 Q6_P_vmpywoh_PP_sat(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpywoh_PP_sat __builtin_HEXAGON_M2_mmpyh_s0 + +/* ========================================================================== + Assembly Syntax: Rdd32=vmpywoh(Rss32,Rtt32):<<1:sat + C Intrinsic Prototype: Word64 Q6_P_vmpywoh_PP_s1_sat(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpywoh_PP_s1_sat __builtin_HEXAGON_M2_mmpyh_s1 + +/* ========================================================================== + Assembly Syntax: Rdd32=vmpyweh(Rss32,Rtt32):rnd:sat + C Intrinsic Prototype: Word64 Q6_P_vmpyweh_PP_rnd_sat(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpyweh_PP_rnd_sat __builtin_HEXAGON_M2_mmpyl_rs0 + +/* ========================================================================== + Assembly Syntax: Rdd32=vmpyweh(Rss32,Rtt32):<<1:rnd:sat + C Intrinsic Prototype: Word64 Q6_P_vmpyweh_PP_s1_rnd_sat(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpyweh_PP_s1_rnd_sat __builtin_HEXAGON_M2_mmpyl_rs1 + +/* ========================================================================== + Assembly Syntax: Rdd32=vmpyweh(Rss32,Rtt32):sat + C Intrinsic Prototype: Word64 Q6_P_vmpyweh_PP_sat(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpyweh_PP_sat __builtin_HEXAGON_M2_mmpyl_s0 + +/* ========================================================================== + Assembly Syntax: Rdd32=vmpyweh(Rss32,Rtt32):<<1:sat + C Intrinsic Prototype: Word64 Q6_P_vmpyweh_PP_s1_sat(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpyweh_PP_s1_sat __builtin_HEXAGON_M2_mmpyl_s1 + +/* ========================================================================== + Assembly Syntax: Rdd32=vmpywouh(Rss32,Rtt32):rnd:sat + C Intrinsic Prototype: Word64 Q6_P_vmpywouh_PP_rnd_sat(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpywouh_PP_rnd_sat __builtin_HEXAGON_M2_mmpyuh_rs0 + +/* ========================================================================== + Assembly Syntax: Rdd32=vmpywouh(Rss32,Rtt32):<<1:rnd:sat + C Intrinsic Prototype: Word64 Q6_P_vmpywouh_PP_s1_rnd_sat(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpywouh_PP_s1_rnd_sat __builtin_HEXAGON_M2_mmpyuh_rs1 + +/* ========================================================================== + Assembly Syntax: Rdd32=vmpywouh(Rss32,Rtt32):sat + C Intrinsic Prototype: Word64 Q6_P_vmpywouh_PP_sat(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpywouh_PP_sat __builtin_HEXAGON_M2_mmpyuh_s0 + +/* ========================================================================== + Assembly Syntax: Rdd32=vmpywouh(Rss32,Rtt32):<<1:sat + C Intrinsic Prototype: Word64 Q6_P_vmpywouh_PP_s1_sat(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpywouh_PP_s1_sat __builtin_HEXAGON_M2_mmpyuh_s1 + +/* ========================================================================== + Assembly Syntax: Rdd32=vmpyweuh(Rss32,Rtt32):rnd:sat + C Intrinsic Prototype: Word64 Q6_P_vmpyweuh_PP_rnd_sat(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpyweuh_PP_rnd_sat __builtin_HEXAGON_M2_mmpyul_rs0 + +/* ========================================================================== + Assembly Syntax: Rdd32=vmpyweuh(Rss32,Rtt32):<<1:rnd:sat + C Intrinsic Prototype: Word64 Q6_P_vmpyweuh_PP_s1_rnd_sat(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpyweuh_PP_s1_rnd_sat __builtin_HEXAGON_M2_mmpyul_rs1 + +/* ========================================================================== + Assembly Syntax: Rdd32=vmpyweuh(Rss32,Rtt32):sat + C Intrinsic Prototype: Word64 Q6_P_vmpyweuh_PP_sat(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpyweuh_PP_sat __builtin_HEXAGON_M2_mmpyul_s0 + +/* ========================================================================== + Assembly Syntax: Rdd32=vmpyweuh(Rss32,Rtt32):<<1:sat + C Intrinsic Prototype: Word64 Q6_P_vmpyweuh_PP_s1_sat(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpyweuh_PP_s1_sat __builtin_HEXAGON_M2_mmpyul_s1 + +/* ========================================================================== + Assembly Syntax: Rx32+=mpy(Rs32.h,Rt32.h) + C Intrinsic Prototype: Word32 Q6_R_mpyacc_RhRh(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyacc_RhRh __builtin_HEXAGON_M2_mpy_acc_hh_s0 + +/* ========================================================================== + Assembly Syntax: Rx32+=mpy(Rs32.h,Rt32.h):<<1 + C Intrinsic Prototype: Word32 Q6_R_mpyacc_RhRh_s1(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyacc_RhRh_s1 __builtin_HEXAGON_M2_mpy_acc_hh_s1 + +/* ========================================================================== + Assembly Syntax: Rx32+=mpy(Rs32.h,Rt32.l) + C Intrinsic Prototype: Word32 Q6_R_mpyacc_RhRl(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyacc_RhRl __builtin_HEXAGON_M2_mpy_acc_hl_s0 + +/* ========================================================================== + Assembly Syntax: Rx32+=mpy(Rs32.h,Rt32.l):<<1 + C Intrinsic Prototype: Word32 Q6_R_mpyacc_RhRl_s1(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyacc_RhRl_s1 __builtin_HEXAGON_M2_mpy_acc_hl_s1 + +/* ========================================================================== + Assembly Syntax: Rx32+=mpy(Rs32.l,Rt32.h) + C Intrinsic Prototype: Word32 Q6_R_mpyacc_RlRh(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyacc_RlRh __builtin_HEXAGON_M2_mpy_acc_lh_s0 + +/* ========================================================================== + Assembly Syntax: Rx32+=mpy(Rs32.l,Rt32.h):<<1 + C Intrinsic Prototype: Word32 Q6_R_mpyacc_RlRh_s1(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyacc_RlRh_s1 __builtin_HEXAGON_M2_mpy_acc_lh_s1 + +/* ========================================================================== + Assembly Syntax: Rx32+=mpy(Rs32.l,Rt32.l) + C Intrinsic Prototype: Word32 Q6_R_mpyacc_RlRl(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyacc_RlRl __builtin_HEXAGON_M2_mpy_acc_ll_s0 + +/* ========================================================================== + Assembly Syntax: Rx32+=mpy(Rs32.l,Rt32.l):<<1 + C Intrinsic Prototype: Word32 Q6_R_mpyacc_RlRl_s1(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyacc_RlRl_s1 __builtin_HEXAGON_M2_mpy_acc_ll_s1 + +/* ========================================================================== + Assembly Syntax: Rx32+=mpy(Rs32.h,Rt32.h):sat + C Intrinsic Prototype: Word32 Q6_R_mpyacc_RhRh_sat(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyacc_RhRh_sat __builtin_HEXAGON_M2_mpy_acc_sat_hh_s0 + +/* ========================================================================== + Assembly Syntax: Rx32+=mpy(Rs32.h,Rt32.h):<<1:sat + C Intrinsic Prototype: Word32 Q6_R_mpyacc_RhRh_s1_sat(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyacc_RhRh_s1_sat __builtin_HEXAGON_M2_mpy_acc_sat_hh_s1 + +/* ========================================================================== + Assembly Syntax: Rx32+=mpy(Rs32.h,Rt32.l):sat + C Intrinsic Prototype: Word32 Q6_R_mpyacc_RhRl_sat(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyacc_RhRl_sat __builtin_HEXAGON_M2_mpy_acc_sat_hl_s0 + +/* ========================================================================== + Assembly Syntax: Rx32+=mpy(Rs32.h,Rt32.l):<<1:sat + C Intrinsic Prototype: Word32 Q6_R_mpyacc_RhRl_s1_sat(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyacc_RhRl_s1_sat __builtin_HEXAGON_M2_mpy_acc_sat_hl_s1 + +/* ========================================================================== + Assembly Syntax: Rx32+=mpy(Rs32.l,Rt32.h):sat + C Intrinsic Prototype: Word32 Q6_R_mpyacc_RlRh_sat(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyacc_RlRh_sat __builtin_HEXAGON_M2_mpy_acc_sat_lh_s0 + +/* ========================================================================== + Assembly Syntax: Rx32+=mpy(Rs32.l,Rt32.h):<<1:sat + C Intrinsic Prototype: Word32 Q6_R_mpyacc_RlRh_s1_sat(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyacc_RlRh_s1_sat __builtin_HEXAGON_M2_mpy_acc_sat_lh_s1 + +/* ========================================================================== + Assembly Syntax: Rx32+=mpy(Rs32.l,Rt32.l):sat + C Intrinsic Prototype: Word32 Q6_R_mpyacc_RlRl_sat(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyacc_RlRl_sat __builtin_HEXAGON_M2_mpy_acc_sat_ll_s0 + +/* ========================================================================== + Assembly Syntax: Rx32+=mpy(Rs32.l,Rt32.l):<<1:sat + C Intrinsic Prototype: Word32 Q6_R_mpyacc_RlRl_s1_sat(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyacc_RlRl_s1_sat __builtin_HEXAGON_M2_mpy_acc_sat_ll_s1 + +/* ========================================================================== + Assembly Syntax: Rd32=mpy(Rs32.h,Rt32.h) + C Intrinsic Prototype: Word32 Q6_R_mpy_RhRh(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpy_RhRh __builtin_HEXAGON_M2_mpy_hh_s0 + +/* ========================================================================== + Assembly Syntax: Rd32=mpy(Rs32.h,Rt32.h):<<1 + C Intrinsic Prototype: Word32 Q6_R_mpy_RhRh_s1(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpy_RhRh_s1 __builtin_HEXAGON_M2_mpy_hh_s1 + +/* ========================================================================== + Assembly Syntax: Rd32=mpy(Rs32.h,Rt32.l) + C Intrinsic Prototype: Word32 Q6_R_mpy_RhRl(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpy_RhRl __builtin_HEXAGON_M2_mpy_hl_s0 + +/* ========================================================================== + Assembly Syntax: Rd32=mpy(Rs32.h,Rt32.l):<<1 + C Intrinsic Prototype: Word32 Q6_R_mpy_RhRl_s1(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpy_RhRl_s1 __builtin_HEXAGON_M2_mpy_hl_s1 + +/* ========================================================================== + Assembly Syntax: Rd32=mpy(Rs32.l,Rt32.h) + C Intrinsic Prototype: Word32 Q6_R_mpy_RlRh(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpy_RlRh __builtin_HEXAGON_M2_mpy_lh_s0 + +/* ========================================================================== + Assembly Syntax: Rd32=mpy(Rs32.l,Rt32.h):<<1 + C Intrinsic Prototype: Word32 Q6_R_mpy_RlRh_s1(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpy_RlRh_s1 __builtin_HEXAGON_M2_mpy_lh_s1 + +/* ========================================================================== + Assembly Syntax: Rd32=mpy(Rs32.l,Rt32.l) + C Intrinsic Prototype: Word32 Q6_R_mpy_RlRl(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpy_RlRl __builtin_HEXAGON_M2_mpy_ll_s0 + +/* ========================================================================== + Assembly Syntax: Rd32=mpy(Rs32.l,Rt32.l):<<1 + C Intrinsic Prototype: Word32 Q6_R_mpy_RlRl_s1(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpy_RlRl_s1 __builtin_HEXAGON_M2_mpy_ll_s1 + +/* ========================================================================== + Assembly Syntax: Rx32-=mpy(Rs32.h,Rt32.h) + C Intrinsic Prototype: Word32 Q6_R_mpynac_RhRh(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpynac_RhRh __builtin_HEXAGON_M2_mpy_nac_hh_s0 + +/* ========================================================================== + Assembly Syntax: Rx32-=mpy(Rs32.h,Rt32.h):<<1 + C Intrinsic Prototype: Word32 Q6_R_mpynac_RhRh_s1(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpynac_RhRh_s1 __builtin_HEXAGON_M2_mpy_nac_hh_s1 + +/* ========================================================================== + Assembly Syntax: Rx32-=mpy(Rs32.h,Rt32.l) + C Intrinsic Prototype: Word32 Q6_R_mpynac_RhRl(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpynac_RhRl __builtin_HEXAGON_M2_mpy_nac_hl_s0 + +/* ========================================================================== + Assembly Syntax: Rx32-=mpy(Rs32.h,Rt32.l):<<1 + C Intrinsic Prototype: Word32 Q6_R_mpynac_RhRl_s1(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpynac_RhRl_s1 __builtin_HEXAGON_M2_mpy_nac_hl_s1 + +/* ========================================================================== + Assembly Syntax: Rx32-=mpy(Rs32.l,Rt32.h) + C Intrinsic Prototype: Word32 Q6_R_mpynac_RlRh(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpynac_RlRh __builtin_HEXAGON_M2_mpy_nac_lh_s0 + +/* ========================================================================== + Assembly Syntax: Rx32-=mpy(Rs32.l,Rt32.h):<<1 + C Intrinsic Prototype: Word32 Q6_R_mpynac_RlRh_s1(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpynac_RlRh_s1 __builtin_HEXAGON_M2_mpy_nac_lh_s1 + +/* ========================================================================== + Assembly Syntax: Rx32-=mpy(Rs32.l,Rt32.l) + C Intrinsic Prototype: Word32 Q6_R_mpynac_RlRl(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpynac_RlRl __builtin_HEXAGON_M2_mpy_nac_ll_s0 + +/* ========================================================================== + Assembly Syntax: Rx32-=mpy(Rs32.l,Rt32.l):<<1 + C Intrinsic Prototype: Word32 Q6_R_mpynac_RlRl_s1(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpynac_RlRl_s1 __builtin_HEXAGON_M2_mpy_nac_ll_s1 + +/* ========================================================================== + Assembly Syntax: Rx32-=mpy(Rs32.h,Rt32.h):sat + C Intrinsic Prototype: Word32 Q6_R_mpynac_RhRh_sat(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpynac_RhRh_sat __builtin_HEXAGON_M2_mpy_nac_sat_hh_s0 + +/* ========================================================================== + Assembly Syntax: Rx32-=mpy(Rs32.h,Rt32.h):<<1:sat + C Intrinsic Prototype: Word32 Q6_R_mpynac_RhRh_s1_sat(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpynac_RhRh_s1_sat __builtin_HEXAGON_M2_mpy_nac_sat_hh_s1 + +/* ========================================================================== + Assembly Syntax: Rx32-=mpy(Rs32.h,Rt32.l):sat + C Intrinsic Prototype: Word32 Q6_R_mpynac_RhRl_sat(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpynac_RhRl_sat __builtin_HEXAGON_M2_mpy_nac_sat_hl_s0 + +/* ========================================================================== + Assembly Syntax: Rx32-=mpy(Rs32.h,Rt32.l):<<1:sat + C Intrinsic Prototype: Word32 Q6_R_mpynac_RhRl_s1_sat(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpynac_RhRl_s1_sat __builtin_HEXAGON_M2_mpy_nac_sat_hl_s1 + +/* ========================================================================== + Assembly Syntax: Rx32-=mpy(Rs32.l,Rt32.h):sat + C Intrinsic Prototype: Word32 Q6_R_mpynac_RlRh_sat(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpynac_RlRh_sat __builtin_HEXAGON_M2_mpy_nac_sat_lh_s0 + +/* ========================================================================== + Assembly Syntax: Rx32-=mpy(Rs32.l,Rt32.h):<<1:sat + C Intrinsic Prototype: Word32 Q6_R_mpynac_RlRh_s1_sat(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpynac_RlRh_s1_sat __builtin_HEXAGON_M2_mpy_nac_sat_lh_s1 + +/* ========================================================================== + Assembly Syntax: Rx32-=mpy(Rs32.l,Rt32.l):sat + C Intrinsic Prototype: Word32 Q6_R_mpynac_RlRl_sat(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpynac_RlRl_sat __builtin_HEXAGON_M2_mpy_nac_sat_ll_s0 + +/* ========================================================================== + Assembly Syntax: Rx32-=mpy(Rs32.l,Rt32.l):<<1:sat + C Intrinsic Prototype: Word32 Q6_R_mpynac_RlRl_s1_sat(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpynac_RlRl_s1_sat __builtin_HEXAGON_M2_mpy_nac_sat_ll_s1 + +/* ========================================================================== + Assembly Syntax: Rd32=mpy(Rs32.h,Rt32.h):rnd + C Intrinsic Prototype: Word32 Q6_R_mpy_RhRh_rnd(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpy_RhRh_rnd __builtin_HEXAGON_M2_mpy_rnd_hh_s0 + +/* ========================================================================== + Assembly Syntax: Rd32=mpy(Rs32.h,Rt32.h):<<1:rnd + C Intrinsic Prototype: Word32 Q6_R_mpy_RhRh_s1_rnd(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpy_RhRh_s1_rnd __builtin_HEXAGON_M2_mpy_rnd_hh_s1 + +/* ========================================================================== + Assembly Syntax: Rd32=mpy(Rs32.h,Rt32.l):rnd + C Intrinsic Prototype: Word32 Q6_R_mpy_RhRl_rnd(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpy_RhRl_rnd __builtin_HEXAGON_M2_mpy_rnd_hl_s0 + +/* ========================================================================== + Assembly Syntax: Rd32=mpy(Rs32.h,Rt32.l):<<1:rnd + C Intrinsic Prototype: Word32 Q6_R_mpy_RhRl_s1_rnd(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpy_RhRl_s1_rnd __builtin_HEXAGON_M2_mpy_rnd_hl_s1 + +/* ========================================================================== + Assembly Syntax: Rd32=mpy(Rs32.l,Rt32.h):rnd + C Intrinsic Prototype: Word32 Q6_R_mpy_RlRh_rnd(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpy_RlRh_rnd __builtin_HEXAGON_M2_mpy_rnd_lh_s0 + +/* ========================================================================== + Assembly Syntax: Rd32=mpy(Rs32.l,Rt32.h):<<1:rnd + C Intrinsic Prototype: Word32 Q6_R_mpy_RlRh_s1_rnd(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpy_RlRh_s1_rnd __builtin_HEXAGON_M2_mpy_rnd_lh_s1 + +/* ========================================================================== + Assembly Syntax: Rd32=mpy(Rs32.l,Rt32.l):rnd + C Intrinsic Prototype: Word32 Q6_R_mpy_RlRl_rnd(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpy_RlRl_rnd __builtin_HEXAGON_M2_mpy_rnd_ll_s0 + +/* ========================================================================== + Assembly Syntax: Rd32=mpy(Rs32.l,Rt32.l):<<1:rnd + C Intrinsic Prototype: Word32 Q6_R_mpy_RlRl_s1_rnd(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpy_RlRl_s1_rnd __builtin_HEXAGON_M2_mpy_rnd_ll_s1 + +/* ========================================================================== + Assembly Syntax: Rd32=mpy(Rs32.h,Rt32.h):sat + C Intrinsic Prototype: Word32 Q6_R_mpy_RhRh_sat(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpy_RhRh_sat __builtin_HEXAGON_M2_mpy_sat_hh_s0 + +/* ========================================================================== + Assembly Syntax: Rd32=mpy(Rs32.h,Rt32.h):<<1:sat + C Intrinsic Prototype: Word32 Q6_R_mpy_RhRh_s1_sat(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpy_RhRh_s1_sat __builtin_HEXAGON_M2_mpy_sat_hh_s1 + +/* ========================================================================== + Assembly Syntax: Rd32=mpy(Rs32.h,Rt32.l):sat + C Intrinsic Prototype: Word32 Q6_R_mpy_RhRl_sat(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpy_RhRl_sat __builtin_HEXAGON_M2_mpy_sat_hl_s0 + +/* ========================================================================== + Assembly Syntax: Rd32=mpy(Rs32.h,Rt32.l):<<1:sat + C Intrinsic Prototype: Word32 Q6_R_mpy_RhRl_s1_sat(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpy_RhRl_s1_sat __builtin_HEXAGON_M2_mpy_sat_hl_s1 + +/* ========================================================================== + Assembly Syntax: Rd32=mpy(Rs32.l,Rt32.h):sat + C Intrinsic Prototype: Word32 Q6_R_mpy_RlRh_sat(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpy_RlRh_sat __builtin_HEXAGON_M2_mpy_sat_lh_s0 + +/* ========================================================================== + Assembly Syntax: Rd32=mpy(Rs32.l,Rt32.h):<<1:sat + C Intrinsic Prototype: Word32 Q6_R_mpy_RlRh_s1_sat(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpy_RlRh_s1_sat __builtin_HEXAGON_M2_mpy_sat_lh_s1 + +/* ========================================================================== + Assembly Syntax: Rd32=mpy(Rs32.l,Rt32.l):sat + C Intrinsic Prototype: Word32 Q6_R_mpy_RlRl_sat(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpy_RlRl_sat __builtin_HEXAGON_M2_mpy_sat_ll_s0 + +/* ========================================================================== + Assembly Syntax: Rd32=mpy(Rs32.l,Rt32.l):<<1:sat + C Intrinsic Prototype: Word32 Q6_R_mpy_RlRl_s1_sat(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpy_RlRl_s1_sat __builtin_HEXAGON_M2_mpy_sat_ll_s1 + +/* ========================================================================== + Assembly Syntax: Rd32=mpy(Rs32.h,Rt32.h):rnd:sat + C Intrinsic Prototype: Word32 Q6_R_mpy_RhRh_rnd_sat(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpy_RhRh_rnd_sat __builtin_HEXAGON_M2_mpy_sat_rnd_hh_s0 + +/* ========================================================================== + Assembly Syntax: Rd32=mpy(Rs32.h,Rt32.h):<<1:rnd:sat + C Intrinsic Prototype: Word32 Q6_R_mpy_RhRh_s1_rnd_sat(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpy_RhRh_s1_rnd_sat __builtin_HEXAGON_M2_mpy_sat_rnd_hh_s1 + +/* ========================================================================== + Assembly Syntax: Rd32=mpy(Rs32.h,Rt32.l):rnd:sat + C Intrinsic Prototype: Word32 Q6_R_mpy_RhRl_rnd_sat(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpy_RhRl_rnd_sat __builtin_HEXAGON_M2_mpy_sat_rnd_hl_s0 + +/* ========================================================================== + Assembly Syntax: Rd32=mpy(Rs32.h,Rt32.l):<<1:rnd:sat + C Intrinsic Prototype: Word32 Q6_R_mpy_RhRl_s1_rnd_sat(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpy_RhRl_s1_rnd_sat __builtin_HEXAGON_M2_mpy_sat_rnd_hl_s1 + +/* ========================================================================== + Assembly Syntax: Rd32=mpy(Rs32.l,Rt32.h):rnd:sat + C Intrinsic Prototype: Word32 Q6_R_mpy_RlRh_rnd_sat(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpy_RlRh_rnd_sat __builtin_HEXAGON_M2_mpy_sat_rnd_lh_s0 + +/* ========================================================================== + Assembly Syntax: Rd32=mpy(Rs32.l,Rt32.h):<<1:rnd:sat + C Intrinsic Prototype: Word32 Q6_R_mpy_RlRh_s1_rnd_sat(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpy_RlRh_s1_rnd_sat __builtin_HEXAGON_M2_mpy_sat_rnd_lh_s1 + +/* ========================================================================== + Assembly Syntax: Rd32=mpy(Rs32.l,Rt32.l):rnd:sat + C Intrinsic Prototype: Word32 Q6_R_mpy_RlRl_rnd_sat(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpy_RlRl_rnd_sat __builtin_HEXAGON_M2_mpy_sat_rnd_ll_s0 + +/* ========================================================================== + Assembly Syntax: Rd32=mpy(Rs32.l,Rt32.l):<<1:rnd:sat + C Intrinsic Prototype: Word32 Q6_R_mpy_RlRl_s1_rnd_sat(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpy_RlRl_s1_rnd_sat __builtin_HEXAGON_M2_mpy_sat_rnd_ll_s1 + +/* ========================================================================== + Assembly Syntax: Rd32=mpy(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_mpy_RR(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpy_RR __builtin_HEXAGON_M2_mpy_up + +/* ========================================================================== + Assembly Syntax: Rd32=mpy(Rs32,Rt32):<<1 + C Intrinsic Prototype: Word32 Q6_R_mpy_RR_s1(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpy_RR_s1 __builtin_HEXAGON_M2_mpy_up_s1 + +/* ========================================================================== + Assembly Syntax: Rd32=mpy(Rs32,Rt32):<<1:sat + C Intrinsic Prototype: Word32 Q6_R_mpy_RR_s1_sat(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpy_RR_s1_sat __builtin_HEXAGON_M2_mpy_up_s1_sat + +/* ========================================================================== + Assembly Syntax: Rxx32+=mpy(Rs32.h,Rt32.h) + C Intrinsic Prototype: Word64 Q6_P_mpyacc_RhRh(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpyacc_RhRh __builtin_HEXAGON_M2_mpyd_acc_hh_s0 + +/* ========================================================================== + Assembly Syntax: Rxx32+=mpy(Rs32.h,Rt32.h):<<1 + C Intrinsic Prototype: Word64 Q6_P_mpyacc_RhRh_s1(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpyacc_RhRh_s1 __builtin_HEXAGON_M2_mpyd_acc_hh_s1 + +/* ========================================================================== + Assembly Syntax: Rxx32+=mpy(Rs32.h,Rt32.l) + C Intrinsic Prototype: Word64 Q6_P_mpyacc_RhRl(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpyacc_RhRl __builtin_HEXAGON_M2_mpyd_acc_hl_s0 + +/* ========================================================================== + Assembly Syntax: Rxx32+=mpy(Rs32.h,Rt32.l):<<1 + C Intrinsic Prototype: Word64 Q6_P_mpyacc_RhRl_s1(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpyacc_RhRl_s1 __builtin_HEXAGON_M2_mpyd_acc_hl_s1 + +/* ========================================================================== + Assembly Syntax: Rxx32+=mpy(Rs32.l,Rt32.h) + C Intrinsic Prototype: Word64 Q6_P_mpyacc_RlRh(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpyacc_RlRh __builtin_HEXAGON_M2_mpyd_acc_lh_s0 + +/* ========================================================================== + Assembly Syntax: Rxx32+=mpy(Rs32.l,Rt32.h):<<1 + C Intrinsic Prototype: Word64 Q6_P_mpyacc_RlRh_s1(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpyacc_RlRh_s1 __builtin_HEXAGON_M2_mpyd_acc_lh_s1 + +/* ========================================================================== + Assembly Syntax: Rxx32+=mpy(Rs32.l,Rt32.l) + C Intrinsic Prototype: Word64 Q6_P_mpyacc_RlRl(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpyacc_RlRl __builtin_HEXAGON_M2_mpyd_acc_ll_s0 + +/* ========================================================================== + Assembly Syntax: Rxx32+=mpy(Rs32.l,Rt32.l):<<1 + C Intrinsic Prototype: Word64 Q6_P_mpyacc_RlRl_s1(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpyacc_RlRl_s1 __builtin_HEXAGON_M2_mpyd_acc_ll_s1 + +/* ========================================================================== + Assembly Syntax: Rdd32=mpy(Rs32.h,Rt32.h) + C Intrinsic Prototype: Word64 Q6_P_mpy_RhRh(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpy_RhRh __builtin_HEXAGON_M2_mpyd_hh_s0 + +/* ========================================================================== + Assembly Syntax: Rdd32=mpy(Rs32.h,Rt32.h):<<1 + C Intrinsic Prototype: Word64 Q6_P_mpy_RhRh_s1(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpy_RhRh_s1 __builtin_HEXAGON_M2_mpyd_hh_s1 + +/* ========================================================================== + Assembly Syntax: Rdd32=mpy(Rs32.h,Rt32.l) + C Intrinsic Prototype: Word64 Q6_P_mpy_RhRl(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpy_RhRl __builtin_HEXAGON_M2_mpyd_hl_s0 + +/* ========================================================================== + Assembly Syntax: Rdd32=mpy(Rs32.h,Rt32.l):<<1 + C Intrinsic Prototype: Word64 Q6_P_mpy_RhRl_s1(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpy_RhRl_s1 __builtin_HEXAGON_M2_mpyd_hl_s1 + +/* ========================================================================== + Assembly Syntax: Rdd32=mpy(Rs32.l,Rt32.h) + C Intrinsic Prototype: Word64 Q6_P_mpy_RlRh(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpy_RlRh __builtin_HEXAGON_M2_mpyd_lh_s0 + +/* ========================================================================== + Assembly Syntax: Rdd32=mpy(Rs32.l,Rt32.h):<<1 + C Intrinsic Prototype: Word64 Q6_P_mpy_RlRh_s1(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpy_RlRh_s1 __builtin_HEXAGON_M2_mpyd_lh_s1 + +/* ========================================================================== + Assembly Syntax: Rdd32=mpy(Rs32.l,Rt32.l) + C Intrinsic Prototype: Word64 Q6_P_mpy_RlRl(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpy_RlRl __builtin_HEXAGON_M2_mpyd_ll_s0 + +/* ========================================================================== + Assembly Syntax: Rdd32=mpy(Rs32.l,Rt32.l):<<1 + C Intrinsic Prototype: Word64 Q6_P_mpy_RlRl_s1(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpy_RlRl_s1 __builtin_HEXAGON_M2_mpyd_ll_s1 + +/* ========================================================================== + Assembly Syntax: Rxx32-=mpy(Rs32.h,Rt32.h) + C Intrinsic Prototype: Word64 Q6_P_mpynac_RhRh(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpynac_RhRh __builtin_HEXAGON_M2_mpyd_nac_hh_s0 + +/* ========================================================================== + Assembly Syntax: Rxx32-=mpy(Rs32.h,Rt32.h):<<1 + C Intrinsic Prototype: Word64 Q6_P_mpynac_RhRh_s1(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpynac_RhRh_s1 __builtin_HEXAGON_M2_mpyd_nac_hh_s1 + +/* ========================================================================== + Assembly Syntax: Rxx32-=mpy(Rs32.h,Rt32.l) + C Intrinsic Prototype: Word64 Q6_P_mpynac_RhRl(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpynac_RhRl __builtin_HEXAGON_M2_mpyd_nac_hl_s0 + +/* ========================================================================== + Assembly Syntax: Rxx32-=mpy(Rs32.h,Rt32.l):<<1 + C Intrinsic Prototype: Word64 Q6_P_mpynac_RhRl_s1(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpynac_RhRl_s1 __builtin_HEXAGON_M2_mpyd_nac_hl_s1 + +/* ========================================================================== + Assembly Syntax: Rxx32-=mpy(Rs32.l,Rt32.h) + C Intrinsic Prototype: Word64 Q6_P_mpynac_RlRh(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpynac_RlRh __builtin_HEXAGON_M2_mpyd_nac_lh_s0 + +/* ========================================================================== + Assembly Syntax: Rxx32-=mpy(Rs32.l,Rt32.h):<<1 + C Intrinsic Prototype: Word64 Q6_P_mpynac_RlRh_s1(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpynac_RlRh_s1 __builtin_HEXAGON_M2_mpyd_nac_lh_s1 + +/* ========================================================================== + Assembly Syntax: Rxx32-=mpy(Rs32.l,Rt32.l) + C Intrinsic Prototype: Word64 Q6_P_mpynac_RlRl(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpynac_RlRl __builtin_HEXAGON_M2_mpyd_nac_ll_s0 + +/* ========================================================================== + Assembly Syntax: Rxx32-=mpy(Rs32.l,Rt32.l):<<1 + C Intrinsic Prototype: Word64 Q6_P_mpynac_RlRl_s1(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpynac_RlRl_s1 __builtin_HEXAGON_M2_mpyd_nac_ll_s1 + +/* ========================================================================== + Assembly Syntax: Rdd32=mpy(Rs32.h,Rt32.h):rnd + C Intrinsic Prototype: Word64 Q6_P_mpy_RhRh_rnd(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpy_RhRh_rnd __builtin_HEXAGON_M2_mpyd_rnd_hh_s0 + +/* ========================================================================== + Assembly Syntax: Rdd32=mpy(Rs32.h,Rt32.h):<<1:rnd + C Intrinsic Prototype: Word64 Q6_P_mpy_RhRh_s1_rnd(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpy_RhRh_s1_rnd __builtin_HEXAGON_M2_mpyd_rnd_hh_s1 + +/* ========================================================================== + Assembly Syntax: Rdd32=mpy(Rs32.h,Rt32.l):rnd + C Intrinsic Prototype: Word64 Q6_P_mpy_RhRl_rnd(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpy_RhRl_rnd __builtin_HEXAGON_M2_mpyd_rnd_hl_s0 + +/* ========================================================================== + Assembly Syntax: Rdd32=mpy(Rs32.h,Rt32.l):<<1:rnd + C Intrinsic Prototype: Word64 Q6_P_mpy_RhRl_s1_rnd(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpy_RhRl_s1_rnd __builtin_HEXAGON_M2_mpyd_rnd_hl_s1 + +/* ========================================================================== + Assembly Syntax: Rdd32=mpy(Rs32.l,Rt32.h):rnd + C Intrinsic Prototype: Word64 Q6_P_mpy_RlRh_rnd(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpy_RlRh_rnd __builtin_HEXAGON_M2_mpyd_rnd_lh_s0 + +/* ========================================================================== + Assembly Syntax: Rdd32=mpy(Rs32.l,Rt32.h):<<1:rnd + C Intrinsic Prototype: Word64 Q6_P_mpy_RlRh_s1_rnd(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpy_RlRh_s1_rnd __builtin_HEXAGON_M2_mpyd_rnd_lh_s1 + +/* ========================================================================== + Assembly Syntax: Rdd32=mpy(Rs32.l,Rt32.l):rnd + C Intrinsic Prototype: Word64 Q6_P_mpy_RlRl_rnd(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpy_RlRl_rnd __builtin_HEXAGON_M2_mpyd_rnd_ll_s0 + +/* ========================================================================== + Assembly Syntax: Rdd32=mpy(Rs32.l,Rt32.l):<<1:rnd + C Intrinsic Prototype: Word64 Q6_P_mpy_RlRl_s1_rnd(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpy_RlRl_s1_rnd __builtin_HEXAGON_M2_mpyd_rnd_ll_s1 + +/* ========================================================================== + Assembly Syntax: Rd32=mpyi(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_mpyi_RR(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyi_RR __builtin_HEXAGON_M2_mpyi + +/* ========================================================================== + Assembly Syntax: Rd32=mpyi(Rs32,#m9) + C Intrinsic Prototype: Word32 Q6_R_mpyi_RI(Word32 Rs, Word32 Im9) + Instruction Type: M + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_mpyi_RI __builtin_HEXAGON_M2_mpysmi + +/* ========================================================================== + Assembly Syntax: Rd32=mpysu(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_mpysu_RR(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpysu_RR __builtin_HEXAGON_M2_mpysu_up + +/* ========================================================================== + Assembly Syntax: Rx32+=mpyu(Rs32.h,Rt32.h) + C Intrinsic Prototype: Word32 Q6_R_mpyuacc_RhRh(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyuacc_RhRh __builtin_HEXAGON_M2_mpyu_acc_hh_s0 + +/* ========================================================================== + Assembly Syntax: Rx32+=mpyu(Rs32.h,Rt32.h):<<1 + C Intrinsic Prototype: Word32 Q6_R_mpyuacc_RhRh_s1(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyuacc_RhRh_s1 __builtin_HEXAGON_M2_mpyu_acc_hh_s1 + +/* ========================================================================== + Assembly Syntax: Rx32+=mpyu(Rs32.h,Rt32.l) + C Intrinsic Prototype: Word32 Q6_R_mpyuacc_RhRl(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyuacc_RhRl __builtin_HEXAGON_M2_mpyu_acc_hl_s0 + +/* ========================================================================== + Assembly Syntax: Rx32+=mpyu(Rs32.h,Rt32.l):<<1 + C Intrinsic Prototype: Word32 Q6_R_mpyuacc_RhRl_s1(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyuacc_RhRl_s1 __builtin_HEXAGON_M2_mpyu_acc_hl_s1 + +/* ========================================================================== + Assembly Syntax: Rx32+=mpyu(Rs32.l,Rt32.h) + C Intrinsic Prototype: Word32 Q6_R_mpyuacc_RlRh(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyuacc_RlRh __builtin_HEXAGON_M2_mpyu_acc_lh_s0 + +/* ========================================================================== + Assembly Syntax: Rx32+=mpyu(Rs32.l,Rt32.h):<<1 + C Intrinsic Prototype: Word32 Q6_R_mpyuacc_RlRh_s1(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyuacc_RlRh_s1 __builtin_HEXAGON_M2_mpyu_acc_lh_s1 + +/* ========================================================================== + Assembly Syntax: Rx32+=mpyu(Rs32.l,Rt32.l) + C Intrinsic Prototype: Word32 Q6_R_mpyuacc_RlRl(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyuacc_RlRl __builtin_HEXAGON_M2_mpyu_acc_ll_s0 + +/* ========================================================================== + Assembly Syntax: Rx32+=mpyu(Rs32.l,Rt32.l):<<1 + C Intrinsic Prototype: Word32 Q6_R_mpyuacc_RlRl_s1(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyuacc_RlRl_s1 __builtin_HEXAGON_M2_mpyu_acc_ll_s1 + +/* ========================================================================== + Assembly Syntax: Rd32=mpyu(Rs32.h,Rt32.h) + C Intrinsic Prototype: UWord32 Q6_R_mpyu_RhRh(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyu_RhRh __builtin_HEXAGON_M2_mpyu_hh_s0 + +/* ========================================================================== + Assembly Syntax: Rd32=mpyu(Rs32.h,Rt32.h):<<1 + C Intrinsic Prototype: UWord32 Q6_R_mpyu_RhRh_s1(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyu_RhRh_s1 __builtin_HEXAGON_M2_mpyu_hh_s1 + +/* ========================================================================== + Assembly Syntax: Rd32=mpyu(Rs32.h,Rt32.l) + C Intrinsic Prototype: UWord32 Q6_R_mpyu_RhRl(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyu_RhRl __builtin_HEXAGON_M2_mpyu_hl_s0 + +/* ========================================================================== + Assembly Syntax: Rd32=mpyu(Rs32.h,Rt32.l):<<1 + C Intrinsic Prototype: UWord32 Q6_R_mpyu_RhRl_s1(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyu_RhRl_s1 __builtin_HEXAGON_M2_mpyu_hl_s1 + +/* ========================================================================== + Assembly Syntax: Rd32=mpyu(Rs32.l,Rt32.h) + C Intrinsic Prototype: UWord32 Q6_R_mpyu_RlRh(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyu_RlRh __builtin_HEXAGON_M2_mpyu_lh_s0 + +/* ========================================================================== + Assembly Syntax: Rd32=mpyu(Rs32.l,Rt32.h):<<1 + C Intrinsic Prototype: UWord32 Q6_R_mpyu_RlRh_s1(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyu_RlRh_s1 __builtin_HEXAGON_M2_mpyu_lh_s1 + +/* ========================================================================== + Assembly Syntax: Rd32=mpyu(Rs32.l,Rt32.l) + C Intrinsic Prototype: UWord32 Q6_R_mpyu_RlRl(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyu_RlRl __builtin_HEXAGON_M2_mpyu_ll_s0 + +/* ========================================================================== + Assembly Syntax: Rd32=mpyu(Rs32.l,Rt32.l):<<1 + C Intrinsic Prototype: UWord32 Q6_R_mpyu_RlRl_s1(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyu_RlRl_s1 __builtin_HEXAGON_M2_mpyu_ll_s1 + +/* ========================================================================== + Assembly Syntax: Rx32-=mpyu(Rs32.h,Rt32.h) + C Intrinsic Prototype: Word32 Q6_R_mpyunac_RhRh(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyunac_RhRh __builtin_HEXAGON_M2_mpyu_nac_hh_s0 + +/* ========================================================================== + Assembly Syntax: Rx32-=mpyu(Rs32.h,Rt32.h):<<1 + C Intrinsic Prototype: Word32 Q6_R_mpyunac_RhRh_s1(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyunac_RhRh_s1 __builtin_HEXAGON_M2_mpyu_nac_hh_s1 + +/* ========================================================================== + Assembly Syntax: Rx32-=mpyu(Rs32.h,Rt32.l) + C Intrinsic Prototype: Word32 Q6_R_mpyunac_RhRl(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyunac_RhRl __builtin_HEXAGON_M2_mpyu_nac_hl_s0 + +/* ========================================================================== + Assembly Syntax: Rx32-=mpyu(Rs32.h,Rt32.l):<<1 + C Intrinsic Prototype: Word32 Q6_R_mpyunac_RhRl_s1(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyunac_RhRl_s1 __builtin_HEXAGON_M2_mpyu_nac_hl_s1 + +/* ========================================================================== + Assembly Syntax: Rx32-=mpyu(Rs32.l,Rt32.h) + C Intrinsic Prototype: Word32 Q6_R_mpyunac_RlRh(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyunac_RlRh __builtin_HEXAGON_M2_mpyu_nac_lh_s0 + +/* ========================================================================== + Assembly Syntax: Rx32-=mpyu(Rs32.l,Rt32.h):<<1 + C Intrinsic Prototype: Word32 Q6_R_mpyunac_RlRh_s1(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyunac_RlRh_s1 __builtin_HEXAGON_M2_mpyu_nac_lh_s1 + +/* ========================================================================== + Assembly Syntax: Rx32-=mpyu(Rs32.l,Rt32.l) + C Intrinsic Prototype: Word32 Q6_R_mpyunac_RlRl(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyunac_RlRl __builtin_HEXAGON_M2_mpyu_nac_ll_s0 + +/* ========================================================================== + Assembly Syntax: Rx32-=mpyu(Rs32.l,Rt32.l):<<1 + C Intrinsic Prototype: Word32 Q6_R_mpyunac_RlRl_s1(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyunac_RlRl_s1 __builtin_HEXAGON_M2_mpyu_nac_ll_s1 + +/* ========================================================================== + Assembly Syntax: Rd32=mpyu(Rs32,Rt32) + C Intrinsic Prototype: UWord32 Q6_R_mpyu_RR(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyu_RR __builtin_HEXAGON_M2_mpyu_up + +/* ========================================================================== + Assembly Syntax: Rxx32+=mpyu(Rs32.h,Rt32.h) + C Intrinsic Prototype: Word64 Q6_P_mpyuacc_RhRh(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpyuacc_RhRh __builtin_HEXAGON_M2_mpyud_acc_hh_s0 + +/* ========================================================================== + Assembly Syntax: Rxx32+=mpyu(Rs32.h,Rt32.h):<<1 + C Intrinsic Prototype: Word64 Q6_P_mpyuacc_RhRh_s1(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpyuacc_RhRh_s1 __builtin_HEXAGON_M2_mpyud_acc_hh_s1 + +/* ========================================================================== + Assembly Syntax: Rxx32+=mpyu(Rs32.h,Rt32.l) + C Intrinsic Prototype: Word64 Q6_P_mpyuacc_RhRl(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpyuacc_RhRl __builtin_HEXAGON_M2_mpyud_acc_hl_s0 + +/* ========================================================================== + Assembly Syntax: Rxx32+=mpyu(Rs32.h,Rt32.l):<<1 + C Intrinsic Prototype: Word64 Q6_P_mpyuacc_RhRl_s1(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpyuacc_RhRl_s1 __builtin_HEXAGON_M2_mpyud_acc_hl_s1 + +/* ========================================================================== + Assembly Syntax: Rxx32+=mpyu(Rs32.l,Rt32.h) + C Intrinsic Prototype: Word64 Q6_P_mpyuacc_RlRh(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpyuacc_RlRh __builtin_HEXAGON_M2_mpyud_acc_lh_s0 + +/* ========================================================================== + Assembly Syntax: Rxx32+=mpyu(Rs32.l,Rt32.h):<<1 + C Intrinsic Prototype: Word64 Q6_P_mpyuacc_RlRh_s1(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpyuacc_RlRh_s1 __builtin_HEXAGON_M2_mpyud_acc_lh_s1 + +/* ========================================================================== + Assembly Syntax: Rxx32+=mpyu(Rs32.l,Rt32.l) + C Intrinsic Prototype: Word64 Q6_P_mpyuacc_RlRl(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpyuacc_RlRl __builtin_HEXAGON_M2_mpyud_acc_ll_s0 + +/* ========================================================================== + Assembly Syntax: Rxx32+=mpyu(Rs32.l,Rt32.l):<<1 + C Intrinsic Prototype: Word64 Q6_P_mpyuacc_RlRl_s1(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpyuacc_RlRl_s1 __builtin_HEXAGON_M2_mpyud_acc_ll_s1 + +/* ========================================================================== + Assembly Syntax: Rdd32=mpyu(Rs32.h,Rt32.h) + C Intrinsic Prototype: UWord64 Q6_P_mpyu_RhRh(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpyu_RhRh __builtin_HEXAGON_M2_mpyud_hh_s0 + +/* ========================================================================== + Assembly Syntax: Rdd32=mpyu(Rs32.h,Rt32.h):<<1 + C Intrinsic Prototype: UWord64 Q6_P_mpyu_RhRh_s1(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpyu_RhRh_s1 __builtin_HEXAGON_M2_mpyud_hh_s1 + +/* ========================================================================== + Assembly Syntax: Rdd32=mpyu(Rs32.h,Rt32.l) + C Intrinsic Prototype: UWord64 Q6_P_mpyu_RhRl(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpyu_RhRl __builtin_HEXAGON_M2_mpyud_hl_s0 + +/* ========================================================================== + Assembly Syntax: Rdd32=mpyu(Rs32.h,Rt32.l):<<1 + C Intrinsic Prototype: UWord64 Q6_P_mpyu_RhRl_s1(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpyu_RhRl_s1 __builtin_HEXAGON_M2_mpyud_hl_s1 + +/* ========================================================================== + Assembly Syntax: Rdd32=mpyu(Rs32.l,Rt32.h) + C Intrinsic Prototype: UWord64 Q6_P_mpyu_RlRh(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpyu_RlRh __builtin_HEXAGON_M2_mpyud_lh_s0 + +/* ========================================================================== + Assembly Syntax: Rdd32=mpyu(Rs32.l,Rt32.h):<<1 + C Intrinsic Prototype: UWord64 Q6_P_mpyu_RlRh_s1(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpyu_RlRh_s1 __builtin_HEXAGON_M2_mpyud_lh_s1 + +/* ========================================================================== + Assembly Syntax: Rdd32=mpyu(Rs32.l,Rt32.l) + C Intrinsic Prototype: UWord64 Q6_P_mpyu_RlRl(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpyu_RlRl __builtin_HEXAGON_M2_mpyud_ll_s0 + +/* ========================================================================== + Assembly Syntax: Rdd32=mpyu(Rs32.l,Rt32.l):<<1 + C Intrinsic Prototype: UWord64 Q6_P_mpyu_RlRl_s1(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpyu_RlRl_s1 __builtin_HEXAGON_M2_mpyud_ll_s1 + +/* ========================================================================== + Assembly Syntax: Rxx32-=mpyu(Rs32.h,Rt32.h) + C Intrinsic Prototype: Word64 Q6_P_mpyunac_RhRh(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpyunac_RhRh __builtin_HEXAGON_M2_mpyud_nac_hh_s0 + +/* ========================================================================== + Assembly Syntax: Rxx32-=mpyu(Rs32.h,Rt32.h):<<1 + C Intrinsic Prototype: Word64 Q6_P_mpyunac_RhRh_s1(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpyunac_RhRh_s1 __builtin_HEXAGON_M2_mpyud_nac_hh_s1 + +/* ========================================================================== + Assembly Syntax: Rxx32-=mpyu(Rs32.h,Rt32.l) + C Intrinsic Prototype: Word64 Q6_P_mpyunac_RhRl(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpyunac_RhRl __builtin_HEXAGON_M2_mpyud_nac_hl_s0 + +/* ========================================================================== + Assembly Syntax: Rxx32-=mpyu(Rs32.h,Rt32.l):<<1 + C Intrinsic Prototype: Word64 Q6_P_mpyunac_RhRl_s1(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpyunac_RhRl_s1 __builtin_HEXAGON_M2_mpyud_nac_hl_s1 + +/* ========================================================================== + Assembly Syntax: Rxx32-=mpyu(Rs32.l,Rt32.h) + C Intrinsic Prototype: Word64 Q6_P_mpyunac_RlRh(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpyunac_RlRh __builtin_HEXAGON_M2_mpyud_nac_lh_s0 + +/* ========================================================================== + Assembly Syntax: Rxx32-=mpyu(Rs32.l,Rt32.h):<<1 + C Intrinsic Prototype: Word64 Q6_P_mpyunac_RlRh_s1(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpyunac_RlRh_s1 __builtin_HEXAGON_M2_mpyud_nac_lh_s1 + +/* ========================================================================== + Assembly Syntax: Rxx32-=mpyu(Rs32.l,Rt32.l) + C Intrinsic Prototype: Word64 Q6_P_mpyunac_RlRl(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpyunac_RlRl __builtin_HEXAGON_M2_mpyud_nac_ll_s0 + +/* ========================================================================== + Assembly Syntax: Rxx32-=mpyu(Rs32.l,Rt32.l):<<1 + C Intrinsic Prototype: Word64 Q6_P_mpyunac_RlRl_s1(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_mpyunac_RlRl_s1 __builtin_HEXAGON_M2_mpyud_nac_ll_s1 + +/* ========================================================================== + Assembly Syntax: Rd32=mpyui(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_mpyui_RR(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_mpyui_RR __builtin_HEXAGON_M2_mpyui + +/* ========================================================================== + Assembly Syntax: Rx32-=add(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_addnac_RR(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_addnac_RR __builtin_HEXAGON_M2_nacci + +/* ========================================================================== + Assembly Syntax: Rx32-=add(Rs32,#s8) + C Intrinsic Prototype: Word32 Q6_R_addnac_RI(Word32 Rx, Word32 Rs, Word32 Is8) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_addnac_RI __builtin_HEXAGON_M2_naccii + +/* ========================================================================== + Assembly Syntax: Rx32+=sub(Rt32,Rs32) + C Intrinsic Prototype: Word32 Q6_R_subacc_RR(Word32 Rx, Word32 Rt, Word32 Rs) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_subacc_RR __builtin_HEXAGON_M2_subacc + +/* ========================================================================== + Assembly Syntax: Rdd32=vabsdiffh(Rtt32,Rss32) + C Intrinsic Prototype: Word64 Q6_P_vabsdiffh_PP(Word64 Rtt, Word64 Rss) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vabsdiffh_PP __builtin_HEXAGON_M2_vabsdiffh + +/* ========================================================================== + Assembly Syntax: Rdd32=vabsdiffw(Rtt32,Rss32) + C Intrinsic Prototype: Word64 Q6_P_vabsdiffw_PP(Word64 Rtt, Word64 Rss) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vabsdiffw_PP __builtin_HEXAGON_M2_vabsdiffw + +/* ========================================================================== + Assembly Syntax: Rxx32+=vcmpyi(Rss32,Rtt32):sat + C Intrinsic Prototype: Word64 Q6_P_vcmpyiacc_PP_sat(Word64 Rxx, Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vcmpyiacc_PP_sat __builtin_HEXAGON_M2_vcmac_s0_sat_i + +/* ========================================================================== + Assembly Syntax: Rxx32+=vcmpyr(Rss32,Rtt32):sat + C Intrinsic Prototype: Word64 Q6_P_vcmpyracc_PP_sat(Word64 Rxx, Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vcmpyracc_PP_sat __builtin_HEXAGON_M2_vcmac_s0_sat_r + +/* ========================================================================== + Assembly Syntax: Rdd32=vcmpyi(Rss32,Rtt32):sat + C Intrinsic Prototype: Word64 Q6_P_vcmpyi_PP_sat(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vcmpyi_PP_sat __builtin_HEXAGON_M2_vcmpy_s0_sat_i + +/* ========================================================================== + Assembly Syntax: Rdd32=vcmpyr(Rss32,Rtt32):sat + C Intrinsic Prototype: Word64 Q6_P_vcmpyr_PP_sat(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vcmpyr_PP_sat __builtin_HEXAGON_M2_vcmpy_s0_sat_r + +/* ========================================================================== + Assembly Syntax: Rdd32=vcmpyi(Rss32,Rtt32):<<1:sat + C Intrinsic Prototype: Word64 Q6_P_vcmpyi_PP_s1_sat(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vcmpyi_PP_s1_sat __builtin_HEXAGON_M2_vcmpy_s1_sat_i + +/* ========================================================================== + Assembly Syntax: Rdd32=vcmpyr(Rss32,Rtt32):<<1:sat + C Intrinsic Prototype: Word64 Q6_P_vcmpyr_PP_s1_sat(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vcmpyr_PP_s1_sat __builtin_HEXAGON_M2_vcmpy_s1_sat_r + +/* ========================================================================== + Assembly Syntax: Rxx32+=vdmpy(Rss32,Rtt32):sat + C Intrinsic Prototype: Word64 Q6_P_vdmpyacc_PP_sat(Word64 Rxx, Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vdmpyacc_PP_sat __builtin_HEXAGON_M2_vdmacs_s0 + +/* ========================================================================== + Assembly Syntax: Rxx32+=vdmpy(Rss32,Rtt32):<<1:sat + C Intrinsic Prototype: Word64 Q6_P_vdmpyacc_PP_s1_sat(Word64 Rxx, Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vdmpyacc_PP_s1_sat __builtin_HEXAGON_M2_vdmacs_s1 + +/* ========================================================================== + Assembly Syntax: Rd32=vdmpy(Rss32,Rtt32):rnd:sat + C Intrinsic Prototype: Word32 Q6_R_vdmpy_PP_rnd_sat(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_vdmpy_PP_rnd_sat __builtin_HEXAGON_M2_vdmpyrs_s0 + +/* ========================================================================== + Assembly Syntax: Rd32=vdmpy(Rss32,Rtt32):<<1:rnd:sat + C Intrinsic Prototype: Word32 Q6_R_vdmpy_PP_s1_rnd_sat(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_vdmpy_PP_s1_rnd_sat __builtin_HEXAGON_M2_vdmpyrs_s1 + +/* ========================================================================== + Assembly Syntax: Rdd32=vdmpy(Rss32,Rtt32):sat + C Intrinsic Prototype: Word64 Q6_P_vdmpy_PP_sat(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vdmpy_PP_sat __builtin_HEXAGON_M2_vdmpys_s0 + +/* ========================================================================== + Assembly Syntax: Rdd32=vdmpy(Rss32,Rtt32):<<1:sat + C Intrinsic Prototype: Word64 Q6_P_vdmpy_PP_s1_sat(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vdmpy_PP_s1_sat __builtin_HEXAGON_M2_vdmpys_s1 + +/* ========================================================================== + Assembly Syntax: Rxx32+=vmpyh(Rs32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_vmpyhacc_RR(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpyhacc_RR __builtin_HEXAGON_M2_vmac2 + +/* ========================================================================== + Assembly Syntax: Rxx32+=vmpyeh(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_vmpyehacc_PP(Word64 Rxx, Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpyehacc_PP __builtin_HEXAGON_M2_vmac2es + +/* ========================================================================== + Assembly Syntax: Rxx32+=vmpyeh(Rss32,Rtt32):sat + C Intrinsic Prototype: Word64 Q6_P_vmpyehacc_PP_sat(Word64 Rxx, Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpyehacc_PP_sat __builtin_HEXAGON_M2_vmac2es_s0 + +/* ========================================================================== + Assembly Syntax: Rxx32+=vmpyeh(Rss32,Rtt32):<<1:sat + C Intrinsic Prototype: Word64 Q6_P_vmpyehacc_PP_s1_sat(Word64 Rxx, Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpyehacc_PP_s1_sat __builtin_HEXAGON_M2_vmac2es_s1 + +/* ========================================================================== + Assembly Syntax: Rxx32+=vmpyh(Rs32,Rt32):sat + C Intrinsic Prototype: Word64 Q6_P_vmpyhacc_RR_sat(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpyhacc_RR_sat __builtin_HEXAGON_M2_vmac2s_s0 + +/* ========================================================================== + Assembly Syntax: Rxx32+=vmpyh(Rs32,Rt32):<<1:sat + C Intrinsic Prototype: Word64 Q6_P_vmpyhacc_RR_s1_sat(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpyhacc_RR_s1_sat __builtin_HEXAGON_M2_vmac2s_s1 + +/* ========================================================================== + Assembly Syntax: Rxx32+=vmpyhsu(Rs32,Rt32):sat + C Intrinsic Prototype: Word64 Q6_P_vmpyhsuacc_RR_sat(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpyhsuacc_RR_sat __builtin_HEXAGON_M2_vmac2su_s0 + +/* ========================================================================== + Assembly Syntax: Rxx32+=vmpyhsu(Rs32,Rt32):<<1:sat + C Intrinsic Prototype: Word64 Q6_P_vmpyhsuacc_RR_s1_sat(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpyhsuacc_RR_s1_sat __builtin_HEXAGON_M2_vmac2su_s1 + +/* ========================================================================== + Assembly Syntax: Rdd32=vmpyeh(Rss32,Rtt32):sat + C Intrinsic Prototype: Word64 Q6_P_vmpyeh_PP_sat(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpyeh_PP_sat __builtin_HEXAGON_M2_vmpy2es_s0 + +/* ========================================================================== + Assembly Syntax: Rdd32=vmpyeh(Rss32,Rtt32):<<1:sat + C Intrinsic Prototype: Word64 Q6_P_vmpyeh_PP_s1_sat(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpyeh_PP_s1_sat __builtin_HEXAGON_M2_vmpy2es_s1 + +/* ========================================================================== + Assembly Syntax: Rdd32=vmpyh(Rs32,Rt32):sat + C Intrinsic Prototype: Word64 Q6_P_vmpyh_RR_sat(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpyh_RR_sat __builtin_HEXAGON_M2_vmpy2s_s0 + +/* ========================================================================== + Assembly Syntax: Rd32=vmpyh(Rs32,Rt32):rnd:sat + C Intrinsic Prototype: Word32 Q6_R_vmpyh_RR_rnd_sat(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_vmpyh_RR_rnd_sat __builtin_HEXAGON_M2_vmpy2s_s0pack + +/* ========================================================================== + Assembly Syntax: Rdd32=vmpyh(Rs32,Rt32):<<1:sat + C Intrinsic Prototype: Word64 Q6_P_vmpyh_RR_s1_sat(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpyh_RR_s1_sat __builtin_HEXAGON_M2_vmpy2s_s1 + +/* ========================================================================== + Assembly Syntax: Rd32=vmpyh(Rs32,Rt32):<<1:rnd:sat + C Intrinsic Prototype: Word32 Q6_R_vmpyh_RR_s1_rnd_sat(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_vmpyh_RR_s1_rnd_sat __builtin_HEXAGON_M2_vmpy2s_s1pack + +/* ========================================================================== + Assembly Syntax: Rdd32=vmpyhsu(Rs32,Rt32):sat + C Intrinsic Prototype: Word64 Q6_P_vmpyhsu_RR_sat(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpyhsu_RR_sat __builtin_HEXAGON_M2_vmpy2su_s0 + +/* ========================================================================== + Assembly Syntax: Rdd32=vmpyhsu(Rs32,Rt32):<<1:sat + C Intrinsic Prototype: Word64 Q6_P_vmpyhsu_RR_s1_sat(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpyhsu_RR_s1_sat __builtin_HEXAGON_M2_vmpy2su_s1 + +/* ========================================================================== + Assembly Syntax: Rd32=vraddh(Rss32,Rtt32) + C Intrinsic Prototype: Word32 Q6_R_vraddh_PP(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_vraddh_PP __builtin_HEXAGON_M2_vraddh + +/* ========================================================================== + Assembly Syntax: Rd32=vradduh(Rss32,Rtt32) + C Intrinsic Prototype: Word32 Q6_R_vradduh_PP(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_vradduh_PP __builtin_HEXAGON_M2_vradduh + +/* ========================================================================== + Assembly Syntax: Rxx32+=vrcmpyi(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_vrcmpyiacc_PP(Word64 Rxx, Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vrcmpyiacc_PP __builtin_HEXAGON_M2_vrcmaci_s0 + +/* ========================================================================== + Assembly Syntax: Rxx32+=vrcmpyi(Rss32,Rtt32*) + C Intrinsic Prototype: Word64 Q6_P_vrcmpyiacc_PP_conj(Word64 Rxx, Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vrcmpyiacc_PP_conj __builtin_HEXAGON_M2_vrcmaci_s0c + +/* ========================================================================== + Assembly Syntax: Rxx32+=vrcmpyr(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_vrcmpyracc_PP(Word64 Rxx, Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vrcmpyracc_PP __builtin_HEXAGON_M2_vrcmacr_s0 + +/* ========================================================================== + Assembly Syntax: Rxx32+=vrcmpyr(Rss32,Rtt32*) + C Intrinsic Prototype: Word64 Q6_P_vrcmpyracc_PP_conj(Word64 Rxx, Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vrcmpyracc_PP_conj __builtin_HEXAGON_M2_vrcmacr_s0c + +/* ========================================================================== + Assembly Syntax: Rdd32=vrcmpyi(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_vrcmpyi_PP(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vrcmpyi_PP __builtin_HEXAGON_M2_vrcmpyi_s0 + +/* ========================================================================== + Assembly Syntax: Rdd32=vrcmpyi(Rss32,Rtt32*) + C Intrinsic Prototype: Word64 Q6_P_vrcmpyi_PP_conj(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vrcmpyi_PP_conj __builtin_HEXAGON_M2_vrcmpyi_s0c + +/* ========================================================================== + Assembly Syntax: Rdd32=vrcmpyr(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_vrcmpyr_PP(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vrcmpyr_PP __builtin_HEXAGON_M2_vrcmpyr_s0 + +/* ========================================================================== + Assembly Syntax: Rdd32=vrcmpyr(Rss32,Rtt32*) + C Intrinsic Prototype: Word64 Q6_P_vrcmpyr_PP_conj(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vrcmpyr_PP_conj __builtin_HEXAGON_M2_vrcmpyr_s0c + +/* ========================================================================== + Assembly Syntax: Rxx32+=vrcmpys(Rss32,Rt32):<<1:sat + C Intrinsic Prototype: Word64 Q6_P_vrcmpysacc_PR_s1_sat(Word64 Rxx, Word64 Rss, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_P_vrcmpysacc_PR_s1_sat __builtin_HEXAGON_M2_vrcmpys_acc_s1 + +/* ========================================================================== + Assembly Syntax: Rdd32=vrcmpys(Rss32,Rt32):<<1:sat + C Intrinsic Prototype: Word64 Q6_P_vrcmpys_PR_s1_sat(Word64 Rss, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_P_vrcmpys_PR_s1_sat __builtin_HEXAGON_M2_vrcmpys_s1 + +/* ========================================================================== + Assembly Syntax: Rd32=vrcmpys(Rss32,Rt32):<<1:rnd:sat + C Intrinsic Prototype: Word32 Q6_R_vrcmpys_PR_s1_rnd_sat(Word64 Rss, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_vrcmpys_PR_s1_rnd_sat __builtin_HEXAGON_M2_vrcmpys_s1rp + +/* ========================================================================== + Assembly Syntax: Rxx32+=vrmpyh(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_vrmpyhacc_PP(Word64 Rxx, Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vrmpyhacc_PP __builtin_HEXAGON_M2_vrmac_s0 + +/* ========================================================================== + Assembly Syntax: Rdd32=vrmpyh(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_vrmpyh_PP(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vrmpyh_PP __builtin_HEXAGON_M2_vrmpy_s0 + +/* ========================================================================== + Assembly Syntax: Rx32^=xor(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_xorxacc_RR(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_xorxacc_RR __builtin_HEXAGON_M2_xor_xacc + +/* ========================================================================== + Assembly Syntax: Rx32&=and(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_andand_RR(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_andand_RR __builtin_HEXAGON_M4_and_and + +/* ========================================================================== + Assembly Syntax: Rx32&=and(Rs32,~Rt32) + C Intrinsic Prototype: Word32 Q6_R_andand_RnR(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_andand_RnR __builtin_HEXAGON_M4_and_andn + +/* ========================================================================== + Assembly Syntax: Rx32&=or(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_orand_RR(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_orand_RR __builtin_HEXAGON_M4_and_or + +/* ========================================================================== + Assembly Syntax: Rx32&=xor(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_xorand_RR(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_xorand_RR __builtin_HEXAGON_M4_and_xor + +/* ========================================================================== + Assembly Syntax: Rd32=cmpyiwh(Rss32,Rt32):<<1:rnd:sat + C Intrinsic Prototype: Word32 Q6_R_cmpyiwh_PR_s1_rnd_sat(Word64 Rss, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_cmpyiwh_PR_s1_rnd_sat __builtin_HEXAGON_M4_cmpyi_wh + +/* ========================================================================== + Assembly Syntax: Rd32=cmpyiwh(Rss32,Rt32*):<<1:rnd:sat + C Intrinsic Prototype: Word32 Q6_R_cmpyiwh_PR_conj_s1_rnd_sat(Word64 Rss, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_cmpyiwh_PR_conj_s1_rnd_sat __builtin_HEXAGON_M4_cmpyi_whc + +/* ========================================================================== + Assembly Syntax: Rd32=cmpyrwh(Rss32,Rt32):<<1:rnd:sat + C Intrinsic Prototype: Word32 Q6_R_cmpyrwh_PR_s1_rnd_sat(Word64 Rss, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_cmpyrwh_PR_s1_rnd_sat __builtin_HEXAGON_M4_cmpyr_wh + +/* ========================================================================== + Assembly Syntax: Rd32=cmpyrwh(Rss32,Rt32*):<<1:rnd:sat + C Intrinsic Prototype: Word32 Q6_R_cmpyrwh_PR_conj_s1_rnd_sat(Word64 Rss, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_cmpyrwh_PR_conj_s1_rnd_sat __builtin_HEXAGON_M4_cmpyr_whc + +/* ========================================================================== + Assembly Syntax: Rx32+=mpy(Rs32,Rt32):<<1:sat + C Intrinsic Prototype: Word32 Q6_R_mpyacc_RR_s1_sat(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyacc_RR_s1_sat __builtin_HEXAGON_M4_mac_up_s1_sat + +/* ========================================================================== + Assembly Syntax: Rd32=add(#u6,mpyi(Rs32,#U6)) + C Intrinsic Prototype: Word32 Q6_R_add_mpyi_IRI(Word32 Iu6, Word32 Rs, Word32 IU6) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_add_mpyi_IRI __builtin_HEXAGON_M4_mpyri_addi + +/* ========================================================================== + Assembly Syntax: Rd32=add(Ru32,mpyi(Rs32,#u6)) + C Intrinsic Prototype: Word32 Q6_R_add_mpyi_RRI(Word32 Ru, Word32 Rs, Word32 Iu6) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_add_mpyi_RRI __builtin_HEXAGON_M4_mpyri_addr + +/* ========================================================================== + Assembly Syntax: Rd32=add(Ru32,mpyi(#u6:2,Rs32)) + C Intrinsic Prototype: Word32 Q6_R_add_mpyi_RIR(Word32 Ru, Word32 Iu6_2, Word32 Rs) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_add_mpyi_RIR __builtin_HEXAGON_M4_mpyri_addr_u2 + +/* ========================================================================== + Assembly Syntax: Rd32=add(#u6,mpyi(Rs32,Rt32)) + C Intrinsic Prototype: Word32 Q6_R_add_mpyi_IRR(Word32 Iu6, Word32 Rs, Word32 Rt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_add_mpyi_IRR __builtin_HEXAGON_M4_mpyrr_addi + +/* ========================================================================== + Assembly Syntax: Ry32=add(Ru32,mpyi(Ry32,Rs32)) + C Intrinsic Prototype: Word32 Q6_R_add_mpyi_RRR(Word32 Ru, Word32 Ry, Word32 Rs) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_add_mpyi_RRR __builtin_HEXAGON_M4_mpyrr_addr + +/* ========================================================================== + Assembly Syntax: Rx32-=mpy(Rs32,Rt32):<<1:sat + C Intrinsic Prototype: Word32 Q6_R_mpynac_RR_s1_sat(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpynac_RR_s1_sat __builtin_HEXAGON_M4_nac_up_s1_sat + +/* ========================================================================== + Assembly Syntax: Rx32|=and(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_andor_RR(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_andor_RR __builtin_HEXAGON_M4_or_and + +/* ========================================================================== + Assembly Syntax: Rx32|=and(Rs32,~Rt32) + C Intrinsic Prototype: Word32 Q6_R_andor_RnR(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_andor_RnR __builtin_HEXAGON_M4_or_andn + +/* ========================================================================== + Assembly Syntax: Rx32|=or(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_oror_RR(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_oror_RR __builtin_HEXAGON_M4_or_or + +/* ========================================================================== + Assembly Syntax: Rx32|=xor(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_xoror_RR(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_xoror_RR __builtin_HEXAGON_M4_or_xor + +/* ========================================================================== + Assembly Syntax: Rdd32=pmpyw(Rs32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_pmpyw_RR(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_pmpyw_RR __builtin_HEXAGON_M4_pmpyw + +/* ========================================================================== + Assembly Syntax: Rxx32^=pmpyw(Rs32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_pmpywxacc_RR(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_pmpywxacc_RR __builtin_HEXAGON_M4_pmpyw_acc + +/* ========================================================================== + Assembly Syntax: Rdd32=vpmpyh(Rs32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_vpmpyh_RR(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vpmpyh_RR __builtin_HEXAGON_M4_vpmpyh + +/* ========================================================================== + Assembly Syntax: Rxx32^=vpmpyh(Rs32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_vpmpyhxacc_RR(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vpmpyhxacc_RR __builtin_HEXAGON_M4_vpmpyh_acc + +/* ========================================================================== + Assembly Syntax: Rxx32+=vrmpyweh(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_vrmpywehacc_PP(Word64 Rxx, Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vrmpywehacc_PP __builtin_HEXAGON_M4_vrmpyeh_acc_s0 + +/* ========================================================================== + Assembly Syntax: Rxx32+=vrmpyweh(Rss32,Rtt32):<<1 + C Intrinsic Prototype: Word64 Q6_P_vrmpywehacc_PP_s1(Word64 Rxx, Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vrmpywehacc_PP_s1 __builtin_HEXAGON_M4_vrmpyeh_acc_s1 + +/* ========================================================================== + Assembly Syntax: Rdd32=vrmpyweh(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_vrmpyweh_PP(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vrmpyweh_PP __builtin_HEXAGON_M4_vrmpyeh_s0 + +/* ========================================================================== + Assembly Syntax: Rdd32=vrmpyweh(Rss32,Rtt32):<<1 + C Intrinsic Prototype: Word64 Q6_P_vrmpyweh_PP_s1(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vrmpyweh_PP_s1 __builtin_HEXAGON_M4_vrmpyeh_s1 + +/* ========================================================================== + Assembly Syntax: Rxx32+=vrmpywoh(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_vrmpywohacc_PP(Word64 Rxx, Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vrmpywohacc_PP __builtin_HEXAGON_M4_vrmpyoh_acc_s0 + +/* ========================================================================== + Assembly Syntax: Rxx32+=vrmpywoh(Rss32,Rtt32):<<1 + C Intrinsic Prototype: Word64 Q6_P_vrmpywohacc_PP_s1(Word64 Rxx, Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vrmpywohacc_PP_s1 __builtin_HEXAGON_M4_vrmpyoh_acc_s1 + +/* ========================================================================== + Assembly Syntax: Rdd32=vrmpywoh(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_vrmpywoh_PP(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vrmpywoh_PP __builtin_HEXAGON_M4_vrmpyoh_s0 + +/* ========================================================================== + Assembly Syntax: Rdd32=vrmpywoh(Rss32,Rtt32):<<1 + C Intrinsic Prototype: Word64 Q6_P_vrmpywoh_PP_s1(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vrmpywoh_PP_s1 __builtin_HEXAGON_M4_vrmpyoh_s1 + +/* ========================================================================== + Assembly Syntax: Rx32^=and(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_andxacc_RR(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_andxacc_RR __builtin_HEXAGON_M4_xor_and + +/* ========================================================================== + Assembly Syntax: Rx32^=and(Rs32,~Rt32) + C Intrinsic Prototype: Word32 Q6_R_andxacc_RnR(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_andxacc_RnR __builtin_HEXAGON_M4_xor_andn + +/* ========================================================================== + Assembly Syntax: Rx32^=or(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_orxacc_RR(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_orxacc_RR __builtin_HEXAGON_M4_xor_or + +/* ========================================================================== + Assembly Syntax: Rxx32^=xor(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_xorxacc_PP(Word64 Rxx, Word64 Rss, Word64 Rtt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_xorxacc_PP __builtin_HEXAGON_M4_xor_xacc + +/* ========================================================================== + Assembly Syntax: Rxx32+=vdmpybsu(Rss32,Rtt32):sat + C Intrinsic Prototype: Word64 Q6_P_vdmpybsuacc_PP_sat(Word64 Rxx, Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vdmpybsuacc_PP_sat __builtin_HEXAGON_M5_vdmacbsu + +/* ========================================================================== + Assembly Syntax: Rdd32=vdmpybsu(Rss32,Rtt32):sat + C Intrinsic Prototype: Word64 Q6_P_vdmpybsu_PP_sat(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vdmpybsu_PP_sat __builtin_HEXAGON_M5_vdmpybsu + +/* ========================================================================== + Assembly Syntax: Rxx32+=vmpybsu(Rs32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_vmpybsuacc_RR(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpybsuacc_RR __builtin_HEXAGON_M5_vmacbsu + +/* ========================================================================== + Assembly Syntax: Rxx32+=vmpybu(Rs32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_vmpybuacc_RR(Word64 Rxx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpybuacc_RR __builtin_HEXAGON_M5_vmacbuu + +/* ========================================================================== + Assembly Syntax: Rdd32=vmpybsu(Rs32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_vmpybsu_RR(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpybsu_RR __builtin_HEXAGON_M5_vmpybsu + +/* ========================================================================== + Assembly Syntax: Rdd32=vmpybu(Rs32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_vmpybu_RR(Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vmpybu_RR __builtin_HEXAGON_M5_vmpybuu + +/* ========================================================================== + Assembly Syntax: Rxx32+=vrmpybsu(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_vrmpybsuacc_PP(Word64 Rxx, Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vrmpybsuacc_PP __builtin_HEXAGON_M5_vrmacbsu + +/* ========================================================================== + Assembly Syntax: Rxx32+=vrmpybu(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_vrmpybuacc_PP(Word64 Rxx, Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vrmpybuacc_PP __builtin_HEXAGON_M5_vrmacbuu + +/* ========================================================================== + Assembly Syntax: Rdd32=vrmpybsu(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_vrmpybsu_PP(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vrmpybsu_PP __builtin_HEXAGON_M5_vrmpybsu + +/* ========================================================================== + Assembly Syntax: Rdd32=vrmpybu(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_vrmpybu_PP(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vrmpybu_PP __builtin_HEXAGON_M5_vrmpybuu + +/* ========================================================================== + Assembly Syntax: Rd32=addasl(Rt32,Rs32,#u3) + C Intrinsic Prototype: Word32 Q6_R_addasl_RRI(Word32 Rt, Word32 Rs, Word32 Iu3) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_addasl_RRI __builtin_HEXAGON_S2_addasl_rrri + +/* ========================================================================== + Assembly Syntax: Rdd32=asl(Rss32,#u6) + C Intrinsic Prototype: Word64 Q6_P_asl_PI(Word64 Rss, Word32 Iu6) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_asl_PI __builtin_HEXAGON_S2_asl_i_p + +/* ========================================================================== + Assembly Syntax: Rxx32+=asl(Rss32,#u6) + C Intrinsic Prototype: Word64 Q6_P_aslacc_PI(Word64 Rxx, Word64 Rss, Word32 Iu6) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_aslacc_PI __builtin_HEXAGON_S2_asl_i_p_acc + +/* ========================================================================== + Assembly Syntax: Rxx32&=asl(Rss32,#u6) + C Intrinsic Prototype: Word64 Q6_P_asland_PI(Word64 Rxx, Word64 Rss, Word32 Iu6) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_asland_PI __builtin_HEXAGON_S2_asl_i_p_and + +/* ========================================================================== + Assembly Syntax: Rxx32-=asl(Rss32,#u6) + C Intrinsic Prototype: Word64 Q6_P_aslnac_PI(Word64 Rxx, Word64 Rss, Word32 Iu6) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_aslnac_PI __builtin_HEXAGON_S2_asl_i_p_nac + +/* ========================================================================== + Assembly Syntax: Rxx32|=asl(Rss32,#u6) + C Intrinsic Prototype: Word64 Q6_P_aslor_PI(Word64 Rxx, Word64 Rss, Word32 Iu6) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_aslor_PI __builtin_HEXAGON_S2_asl_i_p_or + +/* ========================================================================== + Assembly Syntax: Rxx32^=asl(Rss32,#u6) + C Intrinsic Prototype: Word64 Q6_P_aslxacc_PI(Word64 Rxx, Word64 Rss, Word32 Iu6) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_aslxacc_PI __builtin_HEXAGON_S2_asl_i_p_xacc + +/* ========================================================================== + Assembly Syntax: Rd32=asl(Rs32,#u5) + C Intrinsic Prototype: Word32 Q6_R_asl_RI(Word32 Rs, Word32 Iu5) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_asl_RI __builtin_HEXAGON_S2_asl_i_r + +/* ========================================================================== + Assembly Syntax: Rx32+=asl(Rs32,#u5) + C Intrinsic Prototype: Word32 Q6_R_aslacc_RI(Word32 Rx, Word32 Rs, Word32 Iu5) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_aslacc_RI __builtin_HEXAGON_S2_asl_i_r_acc + +/* ========================================================================== + Assembly Syntax: Rx32&=asl(Rs32,#u5) + C Intrinsic Prototype: Word32 Q6_R_asland_RI(Word32 Rx, Word32 Rs, Word32 Iu5) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_asland_RI __builtin_HEXAGON_S2_asl_i_r_and + +/* ========================================================================== + Assembly Syntax: Rx32-=asl(Rs32,#u5) + C Intrinsic Prototype: Word32 Q6_R_aslnac_RI(Word32 Rx, Word32 Rs, Word32 Iu5) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_aslnac_RI __builtin_HEXAGON_S2_asl_i_r_nac + +/* ========================================================================== + Assembly Syntax: Rx32|=asl(Rs32,#u5) + C Intrinsic Prototype: Word32 Q6_R_aslor_RI(Word32 Rx, Word32 Rs, Word32 Iu5) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_aslor_RI __builtin_HEXAGON_S2_asl_i_r_or + +/* ========================================================================== + Assembly Syntax: Rd32=asl(Rs32,#u5):sat + C Intrinsic Prototype: Word32 Q6_R_asl_RI_sat(Word32 Rs, Word32 Iu5) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_asl_RI_sat __builtin_HEXAGON_S2_asl_i_r_sat + +/* ========================================================================== + Assembly Syntax: Rx32^=asl(Rs32,#u5) + C Intrinsic Prototype: Word32 Q6_R_aslxacc_RI(Word32 Rx, Word32 Rs, Word32 Iu5) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_aslxacc_RI __builtin_HEXAGON_S2_asl_i_r_xacc + +/* ========================================================================== + Assembly Syntax: Rdd32=vaslh(Rss32,#u4) + C Intrinsic Prototype: Word64 Q6_P_vaslh_PI(Word64 Rss, Word32 Iu4) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vaslh_PI __builtin_HEXAGON_S2_asl_i_vh + +/* ========================================================================== + Assembly Syntax: Rdd32=vaslw(Rss32,#u5) + C Intrinsic Prototype: Word64 Q6_P_vaslw_PI(Word64 Rss, Word32 Iu5) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vaslw_PI __builtin_HEXAGON_S2_asl_i_vw + +/* ========================================================================== + Assembly Syntax: Rdd32=asl(Rss32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_asl_PR(Word64 Rss, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_asl_PR __builtin_HEXAGON_S2_asl_r_p + +/* ========================================================================== + Assembly Syntax: Rxx32+=asl(Rss32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_aslacc_PR(Word64 Rxx, Word64 Rss, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_aslacc_PR __builtin_HEXAGON_S2_asl_r_p_acc + +/* ========================================================================== + Assembly Syntax: Rxx32&=asl(Rss32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_asland_PR(Word64 Rxx, Word64 Rss, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_asland_PR __builtin_HEXAGON_S2_asl_r_p_and + +/* ========================================================================== + Assembly Syntax: Rxx32-=asl(Rss32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_aslnac_PR(Word64 Rxx, Word64 Rss, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_aslnac_PR __builtin_HEXAGON_S2_asl_r_p_nac + +/* ========================================================================== + Assembly Syntax: Rxx32|=asl(Rss32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_aslor_PR(Word64 Rxx, Word64 Rss, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_aslor_PR __builtin_HEXAGON_S2_asl_r_p_or + +/* ========================================================================== + Assembly Syntax: Rxx32^=asl(Rss32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_aslxacc_PR(Word64 Rxx, Word64 Rss, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_aslxacc_PR __builtin_HEXAGON_S2_asl_r_p_xor + +/* ========================================================================== + Assembly Syntax: Rd32=asl(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_asl_RR(Word32 Rs, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_asl_RR __builtin_HEXAGON_S2_asl_r_r + +/* ========================================================================== + Assembly Syntax: Rx32+=asl(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_aslacc_RR(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_aslacc_RR __builtin_HEXAGON_S2_asl_r_r_acc + +/* ========================================================================== + Assembly Syntax: Rx32&=asl(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_asland_RR(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_asland_RR __builtin_HEXAGON_S2_asl_r_r_and + +/* ========================================================================== + Assembly Syntax: Rx32-=asl(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_aslnac_RR(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_aslnac_RR __builtin_HEXAGON_S2_asl_r_r_nac + +/* ========================================================================== + Assembly Syntax: Rx32|=asl(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_aslor_RR(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_aslor_RR __builtin_HEXAGON_S2_asl_r_r_or + +/* ========================================================================== + Assembly Syntax: Rd32=asl(Rs32,Rt32):sat + C Intrinsic Prototype: Word32 Q6_R_asl_RR_sat(Word32 Rs, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_asl_RR_sat __builtin_HEXAGON_S2_asl_r_r_sat + +/* ========================================================================== + Assembly Syntax: Rdd32=vaslh(Rss32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_vaslh_PR(Word64 Rss, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vaslh_PR __builtin_HEXAGON_S2_asl_r_vh + +/* ========================================================================== + Assembly Syntax: Rdd32=vaslw(Rss32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_vaslw_PR(Word64 Rss, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vaslw_PR __builtin_HEXAGON_S2_asl_r_vw + +/* ========================================================================== + Assembly Syntax: Rdd32=asr(Rss32,#u6) + C Intrinsic Prototype: Word64 Q6_P_asr_PI(Word64 Rss, Word32 Iu6) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_asr_PI __builtin_HEXAGON_S2_asr_i_p + +/* ========================================================================== + Assembly Syntax: Rxx32+=asr(Rss32,#u6) + C Intrinsic Prototype: Word64 Q6_P_asracc_PI(Word64 Rxx, Word64 Rss, Word32 Iu6) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_asracc_PI __builtin_HEXAGON_S2_asr_i_p_acc + +/* ========================================================================== + Assembly Syntax: Rxx32&=asr(Rss32,#u6) + C Intrinsic Prototype: Word64 Q6_P_asrand_PI(Word64 Rxx, Word64 Rss, Word32 Iu6) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_asrand_PI __builtin_HEXAGON_S2_asr_i_p_and + +/* ========================================================================== + Assembly Syntax: Rxx32-=asr(Rss32,#u6) + C Intrinsic Prototype: Word64 Q6_P_asrnac_PI(Word64 Rxx, Word64 Rss, Word32 Iu6) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_asrnac_PI __builtin_HEXAGON_S2_asr_i_p_nac + +/* ========================================================================== + Assembly Syntax: Rxx32|=asr(Rss32,#u6) + C Intrinsic Prototype: Word64 Q6_P_asror_PI(Word64 Rxx, Word64 Rss, Word32 Iu6) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_asror_PI __builtin_HEXAGON_S2_asr_i_p_or + +/* ========================================================================== + Assembly Syntax: Rdd32=asr(Rss32,#u6):rnd + C Intrinsic Prototype: Word64 Q6_P_asr_PI_rnd(Word64 Rss, Word32 Iu6) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_asr_PI_rnd __builtin_HEXAGON_S2_asr_i_p_rnd + +/* ========================================================================== + Assembly Syntax: Rdd32=asrrnd(Rss32,#u6) + C Intrinsic Prototype: Word64 Q6_P_asrrnd_PI(Word64 Rss, Word32 Iu6) + Instruction Type: S_2op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_P_asrrnd_PI __builtin_HEXAGON_S2_asr_i_p_rnd_goodsyntax + +/* ========================================================================== + Assembly Syntax: Rd32=asr(Rs32,#u5) + C Intrinsic Prototype: Word32 Q6_R_asr_RI(Word32 Rs, Word32 Iu5) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_asr_RI __builtin_HEXAGON_S2_asr_i_r + +/* ========================================================================== + Assembly Syntax: Rx32+=asr(Rs32,#u5) + C Intrinsic Prototype: Word32 Q6_R_asracc_RI(Word32 Rx, Word32 Rs, Word32 Iu5) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_asracc_RI __builtin_HEXAGON_S2_asr_i_r_acc + +/* ========================================================================== + Assembly Syntax: Rx32&=asr(Rs32,#u5) + C Intrinsic Prototype: Word32 Q6_R_asrand_RI(Word32 Rx, Word32 Rs, Word32 Iu5) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_asrand_RI __builtin_HEXAGON_S2_asr_i_r_and + +/* ========================================================================== + Assembly Syntax: Rx32-=asr(Rs32,#u5) + C Intrinsic Prototype: Word32 Q6_R_asrnac_RI(Word32 Rx, Word32 Rs, Word32 Iu5) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_asrnac_RI __builtin_HEXAGON_S2_asr_i_r_nac + +/* ========================================================================== + Assembly Syntax: Rx32|=asr(Rs32,#u5) + C Intrinsic Prototype: Word32 Q6_R_asror_RI(Word32 Rx, Word32 Rs, Word32 Iu5) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_asror_RI __builtin_HEXAGON_S2_asr_i_r_or + +/* ========================================================================== + Assembly Syntax: Rd32=asr(Rs32,#u5):rnd + C Intrinsic Prototype: Word32 Q6_R_asr_RI_rnd(Word32 Rs, Word32 Iu5) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_asr_RI_rnd __builtin_HEXAGON_S2_asr_i_r_rnd + +/* ========================================================================== + Assembly Syntax: Rd32=asrrnd(Rs32,#u5) + C Intrinsic Prototype: Word32 Q6_R_asrrnd_RI(Word32 Rs, Word32 Iu5) + Instruction Type: S_2op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_asrrnd_RI __builtin_HEXAGON_S2_asr_i_r_rnd_goodsyntax + +/* ========================================================================== + Assembly Syntax: Rd32=vasrw(Rss32,#u5) + C Intrinsic Prototype: Word32 Q6_R_vasrw_PI(Word64 Rss, Word32 Iu5) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_vasrw_PI __builtin_HEXAGON_S2_asr_i_svw_trun + +/* ========================================================================== + Assembly Syntax: Rdd32=vasrh(Rss32,#u4) + C Intrinsic Prototype: Word64 Q6_P_vasrh_PI(Word64 Rss, Word32 Iu4) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vasrh_PI __builtin_HEXAGON_S2_asr_i_vh + +/* ========================================================================== + Assembly Syntax: Rdd32=vasrw(Rss32,#u5) + C Intrinsic Prototype: Word64 Q6_P_vasrw_PI(Word64 Rss, Word32 Iu5) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vasrw_PI __builtin_HEXAGON_S2_asr_i_vw + +/* ========================================================================== + Assembly Syntax: Rdd32=asr(Rss32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_asr_PR(Word64 Rss, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_asr_PR __builtin_HEXAGON_S2_asr_r_p + +/* ========================================================================== + Assembly Syntax: Rxx32+=asr(Rss32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_asracc_PR(Word64 Rxx, Word64 Rss, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_asracc_PR __builtin_HEXAGON_S2_asr_r_p_acc + +/* ========================================================================== + Assembly Syntax: Rxx32&=asr(Rss32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_asrand_PR(Word64 Rxx, Word64 Rss, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_asrand_PR __builtin_HEXAGON_S2_asr_r_p_and + +/* ========================================================================== + Assembly Syntax: Rxx32-=asr(Rss32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_asrnac_PR(Word64 Rxx, Word64 Rss, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_asrnac_PR __builtin_HEXAGON_S2_asr_r_p_nac + +/* ========================================================================== + Assembly Syntax: Rxx32|=asr(Rss32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_asror_PR(Word64 Rxx, Word64 Rss, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_asror_PR __builtin_HEXAGON_S2_asr_r_p_or + +/* ========================================================================== + Assembly Syntax: Rxx32^=asr(Rss32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_asrxacc_PR(Word64 Rxx, Word64 Rss, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_asrxacc_PR __builtin_HEXAGON_S2_asr_r_p_xor + +/* ========================================================================== + Assembly Syntax: Rd32=asr(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_asr_RR(Word32 Rs, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_asr_RR __builtin_HEXAGON_S2_asr_r_r + +/* ========================================================================== + Assembly Syntax: Rx32+=asr(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_asracc_RR(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_asracc_RR __builtin_HEXAGON_S2_asr_r_r_acc + +/* ========================================================================== + Assembly Syntax: Rx32&=asr(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_asrand_RR(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_asrand_RR __builtin_HEXAGON_S2_asr_r_r_and + +/* ========================================================================== + Assembly Syntax: Rx32-=asr(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_asrnac_RR(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_asrnac_RR __builtin_HEXAGON_S2_asr_r_r_nac + +/* ========================================================================== + Assembly Syntax: Rx32|=asr(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_asror_RR(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_asror_RR __builtin_HEXAGON_S2_asr_r_r_or + +/* ========================================================================== + Assembly Syntax: Rd32=asr(Rs32,Rt32):sat + C Intrinsic Prototype: Word32 Q6_R_asr_RR_sat(Word32 Rs, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_asr_RR_sat __builtin_HEXAGON_S2_asr_r_r_sat + +/* ========================================================================== + Assembly Syntax: Rd32=vasrw(Rss32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_vasrw_PR(Word64 Rss, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_vasrw_PR __builtin_HEXAGON_S2_asr_r_svw_trun + +/* ========================================================================== + Assembly Syntax: Rdd32=vasrh(Rss32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_vasrh_PR(Word64 Rss, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vasrh_PR __builtin_HEXAGON_S2_asr_r_vh + +/* ========================================================================== + Assembly Syntax: Rdd32=vasrw(Rss32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_vasrw_PR(Word64 Rss, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vasrw_PR __builtin_HEXAGON_S2_asr_r_vw + +/* ========================================================================== + Assembly Syntax: Rd32=brev(Rs32) + C Intrinsic Prototype: Word32 Q6_R_brev_R(Word32 Rs) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_brev_R __builtin_HEXAGON_S2_brev + +/* ========================================================================== + Assembly Syntax: Rdd32=brev(Rss32) + C Intrinsic Prototype: Word64 Q6_P_brev_P(Word64 Rss) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_brev_P __builtin_HEXAGON_S2_brevp + +/* ========================================================================== + Assembly Syntax: Rd32=cl0(Rs32) + C Intrinsic Prototype: Word32 Q6_R_cl0_R(Word32 Rs) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_cl0_R __builtin_HEXAGON_S2_cl0 + +/* ========================================================================== + Assembly Syntax: Rd32=cl0(Rss32) + C Intrinsic Prototype: Word32 Q6_R_cl0_P(Word64 Rss) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_cl0_P __builtin_HEXAGON_S2_cl0p + +/* ========================================================================== + Assembly Syntax: Rd32=cl1(Rs32) + C Intrinsic Prototype: Word32 Q6_R_cl1_R(Word32 Rs) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_cl1_R __builtin_HEXAGON_S2_cl1 + +/* ========================================================================== + Assembly Syntax: Rd32=cl1(Rss32) + C Intrinsic Prototype: Word32 Q6_R_cl1_P(Word64 Rss) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_cl1_P __builtin_HEXAGON_S2_cl1p + +/* ========================================================================== + Assembly Syntax: Rd32=clb(Rs32) + C Intrinsic Prototype: Word32 Q6_R_clb_R(Word32 Rs) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_clb_R __builtin_HEXAGON_S2_clb + +/* ========================================================================== + Assembly Syntax: Rd32=normamt(Rs32) + C Intrinsic Prototype: Word32 Q6_R_normamt_R(Word32 Rs) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_normamt_R __builtin_HEXAGON_S2_clbnorm + +/* ========================================================================== + Assembly Syntax: Rd32=clb(Rss32) + C Intrinsic Prototype: Word32 Q6_R_clb_P(Word64 Rss) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_clb_P __builtin_HEXAGON_S2_clbp + +/* ========================================================================== + Assembly Syntax: Rd32=clrbit(Rs32,#u5) + C Intrinsic Prototype: Word32 Q6_R_clrbit_RI(Word32 Rs, Word32 Iu5) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_clrbit_RI __builtin_HEXAGON_S2_clrbit_i + +/* ========================================================================== + Assembly Syntax: Rd32=clrbit(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_clrbit_RR(Word32 Rs, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_clrbit_RR __builtin_HEXAGON_S2_clrbit_r + +/* ========================================================================== + Assembly Syntax: Rd32=ct0(Rs32) + C Intrinsic Prototype: Word32 Q6_R_ct0_R(Word32 Rs) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_ct0_R __builtin_HEXAGON_S2_ct0 + +/* ========================================================================== + Assembly Syntax: Rd32=ct0(Rss32) + C Intrinsic Prototype: Word32 Q6_R_ct0_P(Word64 Rss) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_ct0_P __builtin_HEXAGON_S2_ct0p + +/* ========================================================================== + Assembly Syntax: Rd32=ct1(Rs32) + C Intrinsic Prototype: Word32 Q6_R_ct1_R(Word32 Rs) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_ct1_R __builtin_HEXAGON_S2_ct1 + +/* ========================================================================== + Assembly Syntax: Rd32=ct1(Rss32) + C Intrinsic Prototype: Word32 Q6_R_ct1_P(Word64 Rss) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_ct1_P __builtin_HEXAGON_S2_ct1p + +/* ========================================================================== + Assembly Syntax: Rdd32=deinterleave(Rss32) + C Intrinsic Prototype: Word64 Q6_P_deinterleave_P(Word64 Rss) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_deinterleave_P __builtin_HEXAGON_S2_deinterleave + +/* ========================================================================== + Assembly Syntax: Rd32=extractu(Rs32,#u5,#U5) + C Intrinsic Prototype: Word32 Q6_R_extractu_RII(Word32 Rs, Word32 Iu5, Word32 IU5) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_extractu_RII __builtin_HEXAGON_S2_extractu + +/* ========================================================================== + Assembly Syntax: Rd32=extractu(Rs32,Rtt32) + C Intrinsic Prototype: Word32 Q6_R_extractu_RP(Word32 Rs, Word64 Rtt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_extractu_RP __builtin_HEXAGON_S2_extractu_rp + +/* ========================================================================== + Assembly Syntax: Rdd32=extractu(Rss32,#u6,#U6) + C Intrinsic Prototype: Word64 Q6_P_extractu_PII(Word64 Rss, Word32 Iu6, Word32 IU6) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_extractu_PII __builtin_HEXAGON_S2_extractup + +/* ========================================================================== + Assembly Syntax: Rdd32=extractu(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_extractu_PP(Word64 Rss, Word64 Rtt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_extractu_PP __builtin_HEXAGON_S2_extractup_rp + +/* ========================================================================== + Assembly Syntax: Rx32=insert(Rs32,#u5,#U5) + C Intrinsic Prototype: Word32 Q6_R_insert_RII(Word32 Rx, Word32 Rs, Word32 Iu5, Word32 IU5) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_insert_RII __builtin_HEXAGON_S2_insert + +/* ========================================================================== + Assembly Syntax: Rx32=insert(Rs32,Rtt32) + C Intrinsic Prototype: Word32 Q6_R_insert_RP(Word32 Rx, Word32 Rs, Word64 Rtt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_insert_RP __builtin_HEXAGON_S2_insert_rp + +/* ========================================================================== + Assembly Syntax: Rxx32=insert(Rss32,#u6,#U6) + C Intrinsic Prototype: Word64 Q6_P_insert_PII(Word64 Rxx, Word64 Rss, Word32 Iu6, Word32 IU6) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_insert_PII __builtin_HEXAGON_S2_insertp + +/* ========================================================================== + Assembly Syntax: Rxx32=insert(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_insert_PP(Word64 Rxx, Word64 Rss, Word64 Rtt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_insert_PP __builtin_HEXAGON_S2_insertp_rp + +/* ========================================================================== + Assembly Syntax: Rdd32=interleave(Rss32) + C Intrinsic Prototype: Word64 Q6_P_interleave_P(Word64 Rss) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_interleave_P __builtin_HEXAGON_S2_interleave + +/* ========================================================================== + Assembly Syntax: Rdd32=lfs(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_lfs_PP(Word64 Rss, Word64 Rtt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_lfs_PP __builtin_HEXAGON_S2_lfsp + +/* ========================================================================== + Assembly Syntax: Rdd32=lsl(Rss32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_lsl_PR(Word64 Rss, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_lsl_PR __builtin_HEXAGON_S2_lsl_r_p + +/* ========================================================================== + Assembly Syntax: Rxx32+=lsl(Rss32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_lslacc_PR(Word64 Rxx, Word64 Rss, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_lslacc_PR __builtin_HEXAGON_S2_lsl_r_p_acc + +/* ========================================================================== + Assembly Syntax: Rxx32&=lsl(Rss32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_lsland_PR(Word64 Rxx, Word64 Rss, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_lsland_PR __builtin_HEXAGON_S2_lsl_r_p_and + +/* ========================================================================== + Assembly Syntax: Rxx32-=lsl(Rss32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_lslnac_PR(Word64 Rxx, Word64 Rss, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_lslnac_PR __builtin_HEXAGON_S2_lsl_r_p_nac + +/* ========================================================================== + Assembly Syntax: Rxx32|=lsl(Rss32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_lslor_PR(Word64 Rxx, Word64 Rss, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_lslor_PR __builtin_HEXAGON_S2_lsl_r_p_or + +/* ========================================================================== + Assembly Syntax: Rxx32^=lsl(Rss32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_lslxacc_PR(Word64 Rxx, Word64 Rss, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_lslxacc_PR __builtin_HEXAGON_S2_lsl_r_p_xor + +/* ========================================================================== + Assembly Syntax: Rd32=lsl(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_lsl_RR(Word32 Rs, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_lsl_RR __builtin_HEXAGON_S2_lsl_r_r + +/* ========================================================================== + Assembly Syntax: Rx32+=lsl(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_lslacc_RR(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_lslacc_RR __builtin_HEXAGON_S2_lsl_r_r_acc + +/* ========================================================================== + Assembly Syntax: Rx32&=lsl(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_lsland_RR(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_lsland_RR __builtin_HEXAGON_S2_lsl_r_r_and + +/* ========================================================================== + Assembly Syntax: Rx32-=lsl(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_lslnac_RR(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_lslnac_RR __builtin_HEXAGON_S2_lsl_r_r_nac + +/* ========================================================================== + Assembly Syntax: Rx32|=lsl(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_lslor_RR(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_lslor_RR __builtin_HEXAGON_S2_lsl_r_r_or + +/* ========================================================================== + Assembly Syntax: Rdd32=vlslh(Rss32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_vlslh_PR(Word64 Rss, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vlslh_PR __builtin_HEXAGON_S2_lsl_r_vh + +/* ========================================================================== + Assembly Syntax: Rdd32=vlslw(Rss32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_vlslw_PR(Word64 Rss, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vlslw_PR __builtin_HEXAGON_S2_lsl_r_vw + +/* ========================================================================== + Assembly Syntax: Rdd32=lsr(Rss32,#u6) + C Intrinsic Prototype: Word64 Q6_P_lsr_PI(Word64 Rss, Word32 Iu6) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_lsr_PI __builtin_HEXAGON_S2_lsr_i_p + +/* ========================================================================== + Assembly Syntax: Rxx32+=lsr(Rss32,#u6) + C Intrinsic Prototype: Word64 Q6_P_lsracc_PI(Word64 Rxx, Word64 Rss, Word32 Iu6) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_lsracc_PI __builtin_HEXAGON_S2_lsr_i_p_acc + +/* ========================================================================== + Assembly Syntax: Rxx32&=lsr(Rss32,#u6) + C Intrinsic Prototype: Word64 Q6_P_lsrand_PI(Word64 Rxx, Word64 Rss, Word32 Iu6) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_lsrand_PI __builtin_HEXAGON_S2_lsr_i_p_and + +/* ========================================================================== + Assembly Syntax: Rxx32-=lsr(Rss32,#u6) + C Intrinsic Prototype: Word64 Q6_P_lsrnac_PI(Word64 Rxx, Word64 Rss, Word32 Iu6) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_lsrnac_PI __builtin_HEXAGON_S2_lsr_i_p_nac + +/* ========================================================================== + Assembly Syntax: Rxx32|=lsr(Rss32,#u6) + C Intrinsic Prototype: Word64 Q6_P_lsror_PI(Word64 Rxx, Word64 Rss, Word32 Iu6) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_lsror_PI __builtin_HEXAGON_S2_lsr_i_p_or + +/* ========================================================================== + Assembly Syntax: Rxx32^=lsr(Rss32,#u6) + C Intrinsic Prototype: Word64 Q6_P_lsrxacc_PI(Word64 Rxx, Word64 Rss, Word32 Iu6) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_lsrxacc_PI __builtin_HEXAGON_S2_lsr_i_p_xacc + +/* ========================================================================== + Assembly Syntax: Rd32=lsr(Rs32,#u5) + C Intrinsic Prototype: Word32 Q6_R_lsr_RI(Word32 Rs, Word32 Iu5) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_lsr_RI __builtin_HEXAGON_S2_lsr_i_r + +/* ========================================================================== + Assembly Syntax: Rx32+=lsr(Rs32,#u5) + C Intrinsic Prototype: Word32 Q6_R_lsracc_RI(Word32 Rx, Word32 Rs, Word32 Iu5) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_lsracc_RI __builtin_HEXAGON_S2_lsr_i_r_acc + +/* ========================================================================== + Assembly Syntax: Rx32&=lsr(Rs32,#u5) + C Intrinsic Prototype: Word32 Q6_R_lsrand_RI(Word32 Rx, Word32 Rs, Word32 Iu5) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_lsrand_RI __builtin_HEXAGON_S2_lsr_i_r_and + +/* ========================================================================== + Assembly Syntax: Rx32-=lsr(Rs32,#u5) + C Intrinsic Prototype: Word32 Q6_R_lsrnac_RI(Word32 Rx, Word32 Rs, Word32 Iu5) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_lsrnac_RI __builtin_HEXAGON_S2_lsr_i_r_nac + +/* ========================================================================== + Assembly Syntax: Rx32|=lsr(Rs32,#u5) + C Intrinsic Prototype: Word32 Q6_R_lsror_RI(Word32 Rx, Word32 Rs, Word32 Iu5) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_lsror_RI __builtin_HEXAGON_S2_lsr_i_r_or + +/* ========================================================================== + Assembly Syntax: Rx32^=lsr(Rs32,#u5) + C Intrinsic Prototype: Word32 Q6_R_lsrxacc_RI(Word32 Rx, Word32 Rs, Word32 Iu5) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_lsrxacc_RI __builtin_HEXAGON_S2_lsr_i_r_xacc + +/* ========================================================================== + Assembly Syntax: Rdd32=vlsrh(Rss32,#u4) + C Intrinsic Prototype: Word64 Q6_P_vlsrh_PI(Word64 Rss, Word32 Iu4) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vlsrh_PI __builtin_HEXAGON_S2_lsr_i_vh + +/* ========================================================================== + Assembly Syntax: Rdd32=vlsrw(Rss32,#u5) + C Intrinsic Prototype: Word64 Q6_P_vlsrw_PI(Word64 Rss, Word32 Iu5) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vlsrw_PI __builtin_HEXAGON_S2_lsr_i_vw + +/* ========================================================================== + Assembly Syntax: Rdd32=lsr(Rss32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_lsr_PR(Word64 Rss, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_lsr_PR __builtin_HEXAGON_S2_lsr_r_p + +/* ========================================================================== + Assembly Syntax: Rxx32+=lsr(Rss32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_lsracc_PR(Word64 Rxx, Word64 Rss, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_lsracc_PR __builtin_HEXAGON_S2_lsr_r_p_acc + +/* ========================================================================== + Assembly Syntax: Rxx32&=lsr(Rss32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_lsrand_PR(Word64 Rxx, Word64 Rss, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_lsrand_PR __builtin_HEXAGON_S2_lsr_r_p_and + +/* ========================================================================== + Assembly Syntax: Rxx32-=lsr(Rss32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_lsrnac_PR(Word64 Rxx, Word64 Rss, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_lsrnac_PR __builtin_HEXAGON_S2_lsr_r_p_nac + +/* ========================================================================== + Assembly Syntax: Rxx32|=lsr(Rss32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_lsror_PR(Word64 Rxx, Word64 Rss, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_lsror_PR __builtin_HEXAGON_S2_lsr_r_p_or + +/* ========================================================================== + Assembly Syntax: Rxx32^=lsr(Rss32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_lsrxacc_PR(Word64 Rxx, Word64 Rss, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_lsrxacc_PR __builtin_HEXAGON_S2_lsr_r_p_xor + +/* ========================================================================== + Assembly Syntax: Rd32=lsr(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_lsr_RR(Word32 Rs, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_lsr_RR __builtin_HEXAGON_S2_lsr_r_r + +/* ========================================================================== + Assembly Syntax: Rx32+=lsr(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_lsracc_RR(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_lsracc_RR __builtin_HEXAGON_S2_lsr_r_r_acc + +/* ========================================================================== + Assembly Syntax: Rx32&=lsr(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_lsrand_RR(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_lsrand_RR __builtin_HEXAGON_S2_lsr_r_r_and + +/* ========================================================================== + Assembly Syntax: Rx32-=lsr(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_lsrnac_RR(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_lsrnac_RR __builtin_HEXAGON_S2_lsr_r_r_nac + +/* ========================================================================== + Assembly Syntax: Rx32|=lsr(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_lsror_RR(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_lsror_RR __builtin_HEXAGON_S2_lsr_r_r_or + +/* ========================================================================== + Assembly Syntax: Rdd32=vlsrh(Rss32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_vlsrh_PR(Word64 Rss, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vlsrh_PR __builtin_HEXAGON_S2_lsr_r_vh + +/* ========================================================================== + Assembly Syntax: Rdd32=vlsrw(Rss32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_vlsrw_PR(Word64 Rss, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vlsrw_PR __builtin_HEXAGON_S2_lsr_r_vw + +/* ========================================================================== + Assembly Syntax: Rdd32=packhl(Rs32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_packhl_RR(Word32 Rs, Word32 Rt) + Instruction Type: ALU32_3op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_P_packhl_RR __builtin_HEXAGON_S2_packhl + +/* ========================================================================== + Assembly Syntax: Rd32=parity(Rss32,Rtt32) + C Intrinsic Prototype: Word32 Q6_R_parity_PP(Word64 Rss, Word64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_parity_PP __builtin_HEXAGON_S2_parityp + +/* ========================================================================== + Assembly Syntax: Rd32=setbit(Rs32,#u5) + C Intrinsic Prototype: Word32 Q6_R_setbit_RI(Word32 Rs, Word32 Iu5) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_setbit_RI __builtin_HEXAGON_S2_setbit_i + +/* ========================================================================== + Assembly Syntax: Rd32=setbit(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_setbit_RR(Word32 Rs, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_setbit_RR __builtin_HEXAGON_S2_setbit_r + +/* ========================================================================== + Assembly Syntax: Rdd32=shuffeb(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_shuffeb_PP(Word64 Rss, Word64 Rtt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_shuffeb_PP __builtin_HEXAGON_S2_shuffeb + +/* ========================================================================== + Assembly Syntax: Rdd32=shuffeh(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_shuffeh_PP(Word64 Rss, Word64 Rtt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_shuffeh_PP __builtin_HEXAGON_S2_shuffeh + +/* ========================================================================== + Assembly Syntax: Rdd32=shuffob(Rtt32,Rss32) + C Intrinsic Prototype: Word64 Q6_P_shuffob_PP(Word64 Rtt, Word64 Rss) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_shuffob_PP __builtin_HEXAGON_S2_shuffob + +/* ========================================================================== + Assembly Syntax: Rdd32=shuffoh(Rtt32,Rss32) + C Intrinsic Prototype: Word64 Q6_P_shuffoh_PP(Word64 Rtt, Word64 Rss) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_shuffoh_PP __builtin_HEXAGON_S2_shuffoh + +/* ========================================================================== + Assembly Syntax: memb(Rx32++#s4:0:circ(Mu2))=Rt32 + C Intrinsic Prototype: void Q6_memb_IMR_circ(void** Rx, Word32 Is4_0, Word32 Mu, Word32 Rt, void* BaseAddress) + Instruction Type: ST + Execution Slots: SLOT01 + ========================================================================== */ + +#define Q6_memb_IMR_circ __builtin_HEXAGON_S2_storerb_pci + +/* ========================================================================== + Assembly Syntax: memb(Rx32++I:circ(Mu2))=Rt32 + C Intrinsic Prototype: void Q6_memb_MR_circ(void** Rx, Word32 Mu, Word32 Rt, void* BaseAddress) + Instruction Type: ST + Execution Slots: SLOT01 + ========================================================================== */ + +#define Q6_memb_MR_circ __builtin_HEXAGON_S2_storerb_pcr + +/* ========================================================================== + Assembly Syntax: memd(Rx32++#s4:3:circ(Mu2))=Rtt32 + C Intrinsic Prototype: void Q6_memd_IMP_circ(void** Rx, Word32 Is4_3, Word32 Mu, Word64 Rtt, void* BaseAddress) + Instruction Type: ST + Execution Slots: SLOT01 + ========================================================================== */ + +#define Q6_memd_IMP_circ __builtin_HEXAGON_S2_storerd_pci + +/* ========================================================================== + Assembly Syntax: memd(Rx32++I:circ(Mu2))=Rtt32 + C Intrinsic Prototype: void Q6_memd_MP_circ(void** Rx, Word32 Mu, Word64 Rtt, void* BaseAddress) + Instruction Type: ST + Execution Slots: SLOT01 + ========================================================================== */ + +#define Q6_memd_MP_circ __builtin_HEXAGON_S2_storerd_pcr + +/* ========================================================================== + Assembly Syntax: memh(Rx32++#s4:1:circ(Mu2))=Rt32.h + C Intrinsic Prototype: void Q6_memh_IMRh_circ(void** Rx, Word32 Is4_1, Word32 Mu, Word32 Rt, void* BaseAddress) + Instruction Type: ST + Execution Slots: SLOT01 + ========================================================================== */ + +#define Q6_memh_IMRh_circ __builtin_HEXAGON_S2_storerf_pci + +/* ========================================================================== + Assembly Syntax: memh(Rx32++I:circ(Mu2))=Rt32.h + C Intrinsic Prototype: void Q6_memh_MRh_circ(void** Rx, Word32 Mu, Word32 Rt, void* BaseAddress) + Instruction Type: ST + Execution Slots: SLOT01 + ========================================================================== */ + +#define Q6_memh_MRh_circ __builtin_HEXAGON_S2_storerf_pcr + +/* ========================================================================== + Assembly Syntax: memh(Rx32++#s4:1:circ(Mu2))=Rt32 + C Intrinsic Prototype: void Q6_memh_IMR_circ(void** Rx, Word32 Is4_1, Word32 Mu, Word32 Rt, void* BaseAddress) + Instruction Type: ST + Execution Slots: SLOT01 + ========================================================================== */ + +#define Q6_memh_IMR_circ __builtin_HEXAGON_S2_storerh_pci + +/* ========================================================================== + Assembly Syntax: memh(Rx32++I:circ(Mu2))=Rt32 + C Intrinsic Prototype: void Q6_memh_MR_circ(void** Rx, Word32 Mu, Word32 Rt, void* BaseAddress) + Instruction Type: ST + Execution Slots: SLOT01 + ========================================================================== */ + +#define Q6_memh_MR_circ __builtin_HEXAGON_S2_storerh_pcr + +/* ========================================================================== + Assembly Syntax: memw(Rx32++#s4:2:circ(Mu2))=Rt32 + C Intrinsic Prototype: void Q6_memw_IMR_circ(void** Rx, Word32 Is4_2, Word32 Mu, Word32 Rt, void* BaseAddress) + Instruction Type: ST + Execution Slots: SLOT01 + ========================================================================== */ + +#define Q6_memw_IMR_circ __builtin_HEXAGON_S2_storeri_pci + +/* ========================================================================== + Assembly Syntax: memw(Rx32++I:circ(Mu2))=Rt32 + C Intrinsic Prototype: void Q6_memw_MR_circ(void** Rx, Word32 Mu, Word32 Rt, void* BaseAddress) + Instruction Type: ST + Execution Slots: SLOT01 + ========================================================================== */ + +#define Q6_memw_MR_circ __builtin_HEXAGON_S2_storeri_pcr + +/* ========================================================================== + Assembly Syntax: Rd32=vsathb(Rs32) + C Intrinsic Prototype: Word32 Q6_R_vsathb_R(Word32 Rs) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_vsathb_R __builtin_HEXAGON_S2_svsathb + +/* ========================================================================== + Assembly Syntax: Rd32=vsathub(Rs32) + C Intrinsic Prototype: Word32 Q6_R_vsathub_R(Word32 Rs) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_vsathub_R __builtin_HEXAGON_S2_svsathub + +/* ========================================================================== + Assembly Syntax: Rx32=tableidxb(Rs32,#u4,#U5) + C Intrinsic Prototype: Word32 Q6_R_tableidxb_RII(Word32 Rx, Word32 Rs, Word32 Iu4, Word32 IU5) + Instruction Type: S_2op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_tableidxb_RII __builtin_HEXAGON_S2_tableidxb_goodsyntax + +/* ========================================================================== + Assembly Syntax: Rx32=tableidxd(Rs32,#u4,#U5) + C Intrinsic Prototype: Word32 Q6_R_tableidxd_RII(Word32 Rx, Word32 Rs, Word32 Iu4, Word32 IU5) + Instruction Type: S_2op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_tableidxd_RII __builtin_HEXAGON_S2_tableidxd_goodsyntax + +/* ========================================================================== + Assembly Syntax: Rx32=tableidxh(Rs32,#u4,#U5) + C Intrinsic Prototype: Word32 Q6_R_tableidxh_RII(Word32 Rx, Word32 Rs, Word32 Iu4, Word32 IU5) + Instruction Type: S_2op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_tableidxh_RII __builtin_HEXAGON_S2_tableidxh_goodsyntax + +/* ========================================================================== + Assembly Syntax: Rx32=tableidxw(Rs32,#u4,#U5) + C Intrinsic Prototype: Word32 Q6_R_tableidxw_RII(Word32 Rx, Word32 Rs, Word32 Iu4, Word32 IU5) + Instruction Type: S_2op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_tableidxw_RII __builtin_HEXAGON_S2_tableidxw_goodsyntax + +/* ========================================================================== + Assembly Syntax: Rd32=togglebit(Rs32,#u5) + C Intrinsic Prototype: Word32 Q6_R_togglebit_RI(Word32 Rs, Word32 Iu5) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_togglebit_RI __builtin_HEXAGON_S2_togglebit_i + +/* ========================================================================== + Assembly Syntax: Rd32=togglebit(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_togglebit_RR(Word32 Rs, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_togglebit_RR __builtin_HEXAGON_S2_togglebit_r + +/* ========================================================================== + Assembly Syntax: Pd4=tstbit(Rs32,#u5) + C Intrinsic Prototype: Byte Q6_p_tstbit_RI(Word32 Rs, Word32 Iu5) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_tstbit_RI __builtin_HEXAGON_S2_tstbit_i + +/* ========================================================================== + Assembly Syntax: Pd4=tstbit(Rs32,Rt32) + C Intrinsic Prototype: Byte Q6_p_tstbit_RR(Word32 Rs, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_tstbit_RR __builtin_HEXAGON_S2_tstbit_r + +/* ========================================================================== + Assembly Syntax: Rdd32=valignb(Rtt32,Rss32,#u3) + C Intrinsic Prototype: Word64 Q6_P_valignb_PPI(Word64 Rtt, Word64 Rss, Word32 Iu3) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_valignb_PPI __builtin_HEXAGON_S2_valignib + +/* ========================================================================== + Assembly Syntax: Rdd32=valignb(Rtt32,Rss32,Pu4) + C Intrinsic Prototype: Word64 Q6_P_valignb_PPp(Word64 Rtt, Word64 Rss, Byte Pu) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_valignb_PPp __builtin_HEXAGON_S2_valignrb + +/* ========================================================================== + Assembly Syntax: Rdd32=vcnegh(Rss32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_vcnegh_PR(Word64 Rss, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vcnegh_PR __builtin_HEXAGON_S2_vcnegh + +/* ========================================================================== + Assembly Syntax: Rdd32=vcrotate(Rss32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_vcrotate_PR(Word64 Rss, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vcrotate_PR __builtin_HEXAGON_S2_vcrotate + +/* ========================================================================== + Assembly Syntax: Rxx32+=vrcnegh(Rss32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_vrcneghacc_PR(Word64 Rxx, Word64 Rss, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vrcneghacc_PR __builtin_HEXAGON_S2_vrcnegh + +/* ========================================================================== + Assembly Syntax: Rd32=vrndwh(Rss32) + C Intrinsic Prototype: Word32 Q6_R_vrndwh_P(Word64 Rss) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_vrndwh_P __builtin_HEXAGON_S2_vrndpackwh + +/* ========================================================================== + Assembly Syntax: Rd32=vrndwh(Rss32):sat + C Intrinsic Prototype: Word32 Q6_R_vrndwh_P_sat(Word64 Rss) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_vrndwh_P_sat __builtin_HEXAGON_S2_vrndpackwhs + +/* ========================================================================== + Assembly Syntax: Rd32=vsathb(Rss32) + C Intrinsic Prototype: Word32 Q6_R_vsathb_P(Word64 Rss) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_vsathb_P __builtin_HEXAGON_S2_vsathb + +/* ========================================================================== + Assembly Syntax: Rdd32=vsathb(Rss32) + C Intrinsic Prototype: Word64 Q6_P_vsathb_P(Word64 Rss) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vsathb_P __builtin_HEXAGON_S2_vsathb_nopack + +/* ========================================================================== + Assembly Syntax: Rd32=vsathub(Rss32) + C Intrinsic Prototype: Word32 Q6_R_vsathub_P(Word64 Rss) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_vsathub_P __builtin_HEXAGON_S2_vsathub + +/* ========================================================================== + Assembly Syntax: Rdd32=vsathub(Rss32) + C Intrinsic Prototype: Word64 Q6_P_vsathub_P(Word64 Rss) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vsathub_P __builtin_HEXAGON_S2_vsathub_nopack + +/* ========================================================================== + Assembly Syntax: Rd32=vsatwh(Rss32) + C Intrinsic Prototype: Word32 Q6_R_vsatwh_P(Word64 Rss) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_vsatwh_P __builtin_HEXAGON_S2_vsatwh + +/* ========================================================================== + Assembly Syntax: Rdd32=vsatwh(Rss32) + C Intrinsic Prototype: Word64 Q6_P_vsatwh_P(Word64 Rss) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vsatwh_P __builtin_HEXAGON_S2_vsatwh_nopack + +/* ========================================================================== + Assembly Syntax: Rd32=vsatwuh(Rss32) + C Intrinsic Prototype: Word32 Q6_R_vsatwuh_P(Word64 Rss) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_vsatwuh_P __builtin_HEXAGON_S2_vsatwuh + +/* ========================================================================== + Assembly Syntax: Rdd32=vsatwuh(Rss32) + C Intrinsic Prototype: Word64 Q6_P_vsatwuh_P(Word64 Rss) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vsatwuh_P __builtin_HEXAGON_S2_vsatwuh_nopack + +/* ========================================================================== + Assembly Syntax: Rd32=vsplatb(Rs32) + C Intrinsic Prototype: Word32 Q6_R_vsplatb_R(Word32 Rs) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_vsplatb_R __builtin_HEXAGON_S2_vsplatrb + +/* ========================================================================== + Assembly Syntax: Rdd32=vsplath(Rs32) + C Intrinsic Prototype: Word64 Q6_P_vsplath_R(Word32 Rs) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vsplath_R __builtin_HEXAGON_S2_vsplatrh + +/* ========================================================================== + Assembly Syntax: Rdd32=vspliceb(Rss32,Rtt32,#u3) + C Intrinsic Prototype: Word64 Q6_P_vspliceb_PPI(Word64 Rss, Word64 Rtt, Word32 Iu3) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vspliceb_PPI __builtin_HEXAGON_S2_vspliceib + +/* ========================================================================== + Assembly Syntax: Rdd32=vspliceb(Rss32,Rtt32,Pu4) + C Intrinsic Prototype: Word64 Q6_P_vspliceb_PPp(Word64 Rss, Word64 Rtt, Byte Pu) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vspliceb_PPp __builtin_HEXAGON_S2_vsplicerb + +/* ========================================================================== + Assembly Syntax: Rdd32=vsxtbh(Rs32) + C Intrinsic Prototype: Word64 Q6_P_vsxtbh_R(Word32 Rs) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vsxtbh_R __builtin_HEXAGON_S2_vsxtbh + +/* ========================================================================== + Assembly Syntax: Rdd32=vsxthw(Rs32) + C Intrinsic Prototype: Word64 Q6_P_vsxthw_R(Word32 Rs) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vsxthw_R __builtin_HEXAGON_S2_vsxthw + +/* ========================================================================== + Assembly Syntax: Rd32=vtrunehb(Rss32) + C Intrinsic Prototype: Word32 Q6_R_vtrunehb_P(Word64 Rss) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_vtrunehb_P __builtin_HEXAGON_S2_vtrunehb + +/* ========================================================================== + Assembly Syntax: Rdd32=vtrunewh(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_vtrunewh_PP(Word64 Rss, Word64 Rtt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vtrunewh_PP __builtin_HEXAGON_S2_vtrunewh + +/* ========================================================================== + Assembly Syntax: Rd32=vtrunohb(Rss32) + C Intrinsic Prototype: Word32 Q6_R_vtrunohb_P(Word64 Rss) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_vtrunohb_P __builtin_HEXAGON_S2_vtrunohb + +/* ========================================================================== + Assembly Syntax: Rdd32=vtrunowh(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_vtrunowh_PP(Word64 Rss, Word64 Rtt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vtrunowh_PP __builtin_HEXAGON_S2_vtrunowh + +/* ========================================================================== + Assembly Syntax: Rdd32=vzxtbh(Rs32) + C Intrinsic Prototype: Word64 Q6_P_vzxtbh_R(Word32 Rs) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vzxtbh_R __builtin_HEXAGON_S2_vzxtbh + +/* ========================================================================== + Assembly Syntax: Rdd32=vzxthw(Rs32) + C Intrinsic Prototype: Word64 Q6_P_vzxthw_R(Word32 Rs) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vzxthw_R __builtin_HEXAGON_S2_vzxthw + +/* ========================================================================== + Assembly Syntax: Rd32=add(Rs32,add(Ru32,#s6)) + C Intrinsic Prototype: Word32 Q6_R_add_add_RRI(Word32 Rs, Word32 Ru, Word32 Is6) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_add_add_RRI __builtin_HEXAGON_S4_addaddi + +/* ========================================================================== + Assembly Syntax: Rx32=add(#u8,asl(Rx32,#U5)) + C Intrinsic Prototype: Word32 Q6_R_add_asl_IRI(Word32 Iu8, Word32 Rx, Word32 IU5) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_add_asl_IRI __builtin_HEXAGON_S4_addi_asl_ri + +/* ========================================================================== + Assembly Syntax: Rx32=add(#u8,lsr(Rx32,#U5)) + C Intrinsic Prototype: Word32 Q6_R_add_lsr_IRI(Word32 Iu8, Word32 Rx, Word32 IU5) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_add_lsr_IRI __builtin_HEXAGON_S4_addi_lsr_ri + +/* ========================================================================== + Assembly Syntax: Rx32=and(#u8,asl(Rx32,#U5)) + C Intrinsic Prototype: Word32 Q6_R_and_asl_IRI(Word32 Iu8, Word32 Rx, Word32 IU5) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_and_asl_IRI __builtin_HEXAGON_S4_andi_asl_ri + +/* ========================================================================== + Assembly Syntax: Rx32=and(#u8,lsr(Rx32,#U5)) + C Intrinsic Prototype: Word32 Q6_R_and_lsr_IRI(Word32 Iu8, Word32 Rx, Word32 IU5) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_and_lsr_IRI __builtin_HEXAGON_S4_andi_lsr_ri + +/* ========================================================================== + Assembly Syntax: Rd32=add(clb(Rs32),#s6) + C Intrinsic Prototype: Word32 Q6_R_add_clb_RI(Word32 Rs, Word32 Is6) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_add_clb_RI __builtin_HEXAGON_S4_clbaddi + +/* ========================================================================== + Assembly Syntax: Rd32=add(clb(Rss32),#s6) + C Intrinsic Prototype: Word32 Q6_R_add_clb_PI(Word64 Rss, Word32 Is6) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_add_clb_PI __builtin_HEXAGON_S4_clbpaddi + +/* ========================================================================== + Assembly Syntax: Rd32=normamt(Rss32) + C Intrinsic Prototype: Word32 Q6_R_normamt_P(Word64 Rss) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_normamt_P __builtin_HEXAGON_S4_clbpnorm + +/* ========================================================================== + Assembly Syntax: Rd32=extract(Rs32,#u5,#U5) + C Intrinsic Prototype: Word32 Q6_R_extract_RII(Word32 Rs, Word32 Iu5, Word32 IU5) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_extract_RII __builtin_HEXAGON_S4_extract + +/* ========================================================================== + Assembly Syntax: Rd32=extract(Rs32,Rtt32) + C Intrinsic Prototype: Word32 Q6_R_extract_RP(Word32 Rs, Word64 Rtt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_extract_RP __builtin_HEXAGON_S4_extract_rp + +/* ========================================================================== + Assembly Syntax: Rdd32=extract(Rss32,#u6,#U6) + C Intrinsic Prototype: Word64 Q6_P_extract_PII(Word64 Rss, Word32 Iu6, Word32 IU6) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_extract_PII __builtin_HEXAGON_S4_extractp + +/* ========================================================================== + Assembly Syntax: Rdd32=extract(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_extract_PP(Word64 Rss, Word64 Rtt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_extract_PP __builtin_HEXAGON_S4_extractp_rp + +/* ========================================================================== + Assembly Syntax: Rd32=lsl(#s6,Rt32) + C Intrinsic Prototype: Word32 Q6_R_lsl_IR(Word32 Is6, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_lsl_IR __builtin_HEXAGON_S4_lsli + +/* ========================================================================== + Assembly Syntax: Pd4=!tstbit(Rs32,#u5) + C Intrinsic Prototype: Byte Q6_p_not_tstbit_RI(Word32 Rs, Word32 Iu5) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_not_tstbit_RI __builtin_HEXAGON_S4_ntstbit_i + +/* ========================================================================== + Assembly Syntax: Pd4=!tstbit(Rs32,Rt32) + C Intrinsic Prototype: Byte Q6_p_not_tstbit_RR(Word32 Rs, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_not_tstbit_RR __builtin_HEXAGON_S4_ntstbit_r + +/* ========================================================================== + Assembly Syntax: Rx32|=and(Rs32,#s10) + C Intrinsic Prototype: Word32 Q6_R_andor_RI(Word32 Rx, Word32 Rs, Word32 Is10) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_andor_RI __builtin_HEXAGON_S4_or_andi + +/* ========================================================================== + Assembly Syntax: Rx32=or(Ru32,and(Rx32,#s10)) + C Intrinsic Prototype: Word32 Q6_R_or_and_RRI(Word32 Ru, Word32 Rx, Word32 Is10) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_or_and_RRI __builtin_HEXAGON_S4_or_andix + +/* ========================================================================== + Assembly Syntax: Rx32|=or(Rs32,#s10) + C Intrinsic Prototype: Word32 Q6_R_oror_RI(Word32 Rx, Word32 Rs, Word32 Is10) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_oror_RI __builtin_HEXAGON_S4_or_ori + +/* ========================================================================== + Assembly Syntax: Rx32=or(#u8,asl(Rx32,#U5)) + C Intrinsic Prototype: Word32 Q6_R_or_asl_IRI(Word32 Iu8, Word32 Rx, Word32 IU5) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_or_asl_IRI __builtin_HEXAGON_S4_ori_asl_ri + +/* ========================================================================== + Assembly Syntax: Rx32=or(#u8,lsr(Rx32,#U5)) + C Intrinsic Prototype: Word32 Q6_R_or_lsr_IRI(Word32 Iu8, Word32 Rx, Word32 IU5) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_or_lsr_IRI __builtin_HEXAGON_S4_ori_lsr_ri + +/* ========================================================================== + Assembly Syntax: Rd32=parity(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_parity_RR(Word32 Rs, Word32 Rt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_parity_RR __builtin_HEXAGON_S4_parity + +/* ========================================================================== + Assembly Syntax: Rd32=add(Rs32,sub(#s6,Ru32)) + C Intrinsic Prototype: Word32 Q6_R_add_sub_RIR(Word32 Rs, Word32 Is6, Word32 Ru) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_add_sub_RIR __builtin_HEXAGON_S4_subaddi + +/* ========================================================================== + Assembly Syntax: Rx32=sub(#u8,asl(Rx32,#U5)) + C Intrinsic Prototype: Word32 Q6_R_sub_asl_IRI(Word32 Iu8, Word32 Rx, Word32 IU5) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_sub_asl_IRI __builtin_HEXAGON_S4_subi_asl_ri + +/* ========================================================================== + Assembly Syntax: Rx32=sub(#u8,lsr(Rx32,#U5)) + C Intrinsic Prototype: Word32 Q6_R_sub_lsr_IRI(Word32 Iu8, Word32 Rx, Word32 IU5) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_sub_lsr_IRI __builtin_HEXAGON_S4_subi_lsr_ri + +/* ========================================================================== + Assembly Syntax: Rdd32=vrcrotate(Rss32,Rt32,#u2) + C Intrinsic Prototype: Word64 Q6_P_vrcrotate_PRI(Word64 Rss, Word32 Rt, Word32 Iu2) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vrcrotate_PRI __builtin_HEXAGON_S4_vrcrotate + +/* ========================================================================== + Assembly Syntax: Rxx32+=vrcrotate(Rss32,Rt32,#u2) + C Intrinsic Prototype: Word64 Q6_P_vrcrotateacc_PRI(Word64 Rxx, Word64 Rss, Word32 Rt, Word32 Iu2) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vrcrotateacc_PRI __builtin_HEXAGON_S4_vrcrotate_acc + +/* ========================================================================== + Assembly Syntax: Rdd32=vxaddsubh(Rss32,Rtt32):sat + C Intrinsic Prototype: Word64 Q6_P_vxaddsubh_PP_sat(Word64 Rss, Word64 Rtt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vxaddsubh_PP_sat __builtin_HEXAGON_S4_vxaddsubh + +/* ========================================================================== + Assembly Syntax: Rdd32=vxaddsubh(Rss32,Rtt32):rnd:>>1:sat + C Intrinsic Prototype: Word64 Q6_P_vxaddsubh_PP_rnd_rs1_sat(Word64 Rss, Word64 Rtt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vxaddsubh_PP_rnd_rs1_sat __builtin_HEXAGON_S4_vxaddsubhr + +/* ========================================================================== + Assembly Syntax: Rdd32=vxaddsubw(Rss32,Rtt32):sat + C Intrinsic Prototype: Word64 Q6_P_vxaddsubw_PP_sat(Word64 Rss, Word64 Rtt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vxaddsubw_PP_sat __builtin_HEXAGON_S4_vxaddsubw + +/* ========================================================================== + Assembly Syntax: Rdd32=vxsubaddh(Rss32,Rtt32):sat + C Intrinsic Prototype: Word64 Q6_P_vxsubaddh_PP_sat(Word64 Rss, Word64 Rtt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vxsubaddh_PP_sat __builtin_HEXAGON_S4_vxsubaddh + +/* ========================================================================== + Assembly Syntax: Rdd32=vxsubaddh(Rss32,Rtt32):rnd:>>1:sat + C Intrinsic Prototype: Word64 Q6_P_vxsubaddh_PP_rnd_rs1_sat(Word64 Rss, Word64 Rtt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vxsubaddh_PP_rnd_rs1_sat __builtin_HEXAGON_S4_vxsubaddhr + +/* ========================================================================== + Assembly Syntax: Rdd32=vxsubaddw(Rss32,Rtt32):sat + C Intrinsic Prototype: Word64 Q6_P_vxsubaddw_PP_sat(Word64 Rss, Word64 Rtt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vxsubaddw_PP_sat __builtin_HEXAGON_S4_vxsubaddw + +/* ========================================================================== + Assembly Syntax: Rd32=vasrhub(Rss32,#u4):rnd:sat + C Intrinsic Prototype: Word32 Q6_R_vasrhub_PI_rnd_sat(Word64 Rss, Word32 Iu4) + Instruction Type: S_2op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_R_vasrhub_PI_rnd_sat __builtin_HEXAGON_S5_asrhub_rnd_sat_goodsyntax + +/* ========================================================================== + Assembly Syntax: Rd32=vasrhub(Rss32,#u4):sat + C Intrinsic Prototype: Word32 Q6_R_vasrhub_PI_sat(Word64 Rss, Word32 Iu4) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_vasrhub_PI_sat __builtin_HEXAGON_S5_asrhub_sat + +/* ========================================================================== + Assembly Syntax: Rd32=popcount(Rss32) + C Intrinsic Prototype: Word32 Q6_R_popcount_P(Word64 Rss) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_popcount_P __builtin_HEXAGON_S5_popcountp + +/* ========================================================================== + Assembly Syntax: Rdd32=vasrh(Rss32,#u4):rnd + C Intrinsic Prototype: Word64 Q6_P_vasrh_PI_rnd(Word64 Rss, Word32 Iu4) + Instruction Type: S_2op + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_P_vasrh_PI_rnd __builtin_HEXAGON_S5_vasrhrnd_goodsyntax + +/* ========================================================================== + Assembly Syntax: dccleana(Rs32) + C Intrinsic Prototype: void Q6_dccleana_A(Address Rs) + Instruction Type: ST + Execution Slots: SLOT0 + ========================================================================== */ + +#define Q6_dccleana_A __builtin_HEXAGON_Y2_dccleana + +/* ========================================================================== + Assembly Syntax: dccleaninva(Rs32) + C Intrinsic Prototype: void Q6_dccleaninva_A(Address Rs) + Instruction Type: ST + Execution Slots: SLOT0 + ========================================================================== */ + +#define Q6_dccleaninva_A __builtin_HEXAGON_Y2_dccleaninva + +/* ========================================================================== + Assembly Syntax: dcfetch(Rs32) + C Intrinsic Prototype: void Q6_dcfetch_A(Address Rs) + Instruction Type: MAPPING + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_dcfetch_A __builtin_HEXAGON_Y2_dcfetch + +/* ========================================================================== + Assembly Syntax: dcinva(Rs32) + C Intrinsic Prototype: void Q6_dcinva_A(Address Rs) + Instruction Type: ST + Execution Slots: SLOT0 + ========================================================================== */ + +#define Q6_dcinva_A __builtin_HEXAGON_Y2_dcinva + +/* ========================================================================== + Assembly Syntax: dczeroa(Rs32) + C Intrinsic Prototype: void Q6_dczeroa_A(Address Rs) + Instruction Type: ST + Execution Slots: SLOT0 + ========================================================================== */ + +#define Q6_dczeroa_A __builtin_HEXAGON_Y2_dczeroa + +/* ========================================================================== + Assembly Syntax: l2fetch(Rs32,Rt32) + C Intrinsic Prototype: void Q6_l2fetch_AR(Address Rs, Word32 Rt) + Instruction Type: ST + Execution Slots: SLOT0 + ========================================================================== */ + +#define Q6_l2fetch_AR __builtin_HEXAGON_Y4_l2fetch + +/* ========================================================================== + Assembly Syntax: l2fetch(Rs32,Rtt32) + C Intrinsic Prototype: void Q6_l2fetch_AP(Address Rs, Word64 Rtt) + Instruction Type: ST + Execution Slots: SLOT0 + ========================================================================== */ + +#define Q6_l2fetch_AP __builtin_HEXAGON_Y5_l2fetch + +#if __HEXAGON_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Rdd32=rol(Rss32,#u6) + C Intrinsic Prototype: Word64 Q6_P_rol_PI(Word64 Rss, Word32 Iu6) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_rol_PI __builtin_HEXAGON_S6_rol_i_p +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HEXAGON_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Rxx32+=rol(Rss32,#u6) + C Intrinsic Prototype: Word64 Q6_P_rolacc_PI(Word64 Rxx, Word64 Rss, Word32 Iu6) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_rolacc_PI __builtin_HEXAGON_S6_rol_i_p_acc +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HEXAGON_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Rxx32&=rol(Rss32,#u6) + C Intrinsic Prototype: Word64 Q6_P_roland_PI(Word64 Rxx, Word64 Rss, Word32 Iu6) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_roland_PI __builtin_HEXAGON_S6_rol_i_p_and +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HEXAGON_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Rxx32-=rol(Rss32,#u6) + C Intrinsic Prototype: Word64 Q6_P_rolnac_PI(Word64 Rxx, Word64 Rss, Word32 Iu6) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_rolnac_PI __builtin_HEXAGON_S6_rol_i_p_nac +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HEXAGON_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Rxx32|=rol(Rss32,#u6) + C Intrinsic Prototype: Word64 Q6_P_rolor_PI(Word64 Rxx, Word64 Rss, Word32 Iu6) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_rolor_PI __builtin_HEXAGON_S6_rol_i_p_or +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HEXAGON_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Rxx32^=rol(Rss32,#u6) + C Intrinsic Prototype: Word64 Q6_P_rolxacc_PI(Word64 Rxx, Word64 Rss, Word32 Iu6) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_rolxacc_PI __builtin_HEXAGON_S6_rol_i_p_xacc +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HEXAGON_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Rd32=rol(Rs32,#u5) + C Intrinsic Prototype: Word32 Q6_R_rol_RI(Word32 Rs, Word32 Iu5) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_rol_RI __builtin_HEXAGON_S6_rol_i_r +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HEXAGON_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Rx32+=rol(Rs32,#u5) + C Intrinsic Prototype: Word32 Q6_R_rolacc_RI(Word32 Rx, Word32 Rs, Word32 Iu5) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_rolacc_RI __builtin_HEXAGON_S6_rol_i_r_acc +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HEXAGON_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Rx32&=rol(Rs32,#u5) + C Intrinsic Prototype: Word32 Q6_R_roland_RI(Word32 Rx, Word32 Rs, Word32 Iu5) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_roland_RI __builtin_HEXAGON_S6_rol_i_r_and +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HEXAGON_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Rx32-=rol(Rs32,#u5) + C Intrinsic Prototype: Word32 Q6_R_rolnac_RI(Word32 Rx, Word32 Rs, Word32 Iu5) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_rolnac_RI __builtin_HEXAGON_S6_rol_i_r_nac +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HEXAGON_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Rx32|=rol(Rs32,#u5) + C Intrinsic Prototype: Word32 Q6_R_rolor_RI(Word32 Rx, Word32 Rs, Word32 Iu5) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_rolor_RI __builtin_HEXAGON_S6_rol_i_r_or +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HEXAGON_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Rx32^=rol(Rs32,#u5) + C Intrinsic Prototype: Word32 Q6_R_rolxacc_RI(Word32 Rx, Word32 Rs, Word32 Iu5) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_rolxacc_RI __builtin_HEXAGON_S6_rol_i_r_xacc +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HEXAGON_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Rdd32=vabsdiffb(Rtt32,Rss32) + C Intrinsic Prototype: Word64 Q6_P_vabsdiffb_PP(Word64 Rtt, Word64 Rss) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vabsdiffb_PP __builtin_HEXAGON_M6_vabsdiffb +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HEXAGON_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Rdd32=vabsdiffub(Rtt32,Rss32) + C Intrinsic Prototype: Word64 Q6_P_vabsdiffub_PP(Word64 Rtt, Word64 Rss) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vabsdiffub_PP __builtin_HEXAGON_M6_vabsdiffub +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HEXAGON_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Rdd32=vsplatb(Rs32) + C Intrinsic Prototype: Word64 Q6_P_vsplatb_R(Word32 Rs) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vsplatb_R __builtin_HEXAGON_S6_vsplatrbp +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HEXAGON_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Rdd32=vtrunehb(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_vtrunehb_PP(Word64 Rss, Word64 Rtt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vtrunehb_PP __builtin_HEXAGON_S6_vtrunehb_ppp +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HEXAGON_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Rdd32=vtrunohb(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_vtrunohb_PP(Word64 Rss, Word64 Rtt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vtrunohb_PP __builtin_HEXAGON_S6_vtrunohb_ppp +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HEXAGON_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Vd32=vmem(Rt32):nt + C Intrinsic Prototype: HVX_Vector Q6_V_vmem_R_nt(Word32 Rt) + Instruction Type: MAPPING + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_V_vmem_R_nt __builtin_HEXAGON_V6_ldntnt0 +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HEXAGON_ARCH__ >= 65 +/* ========================================================================== + Assembly Syntax: Pd4=!any8(vcmpb.eq(Rss32,Rtt32)) + C Intrinsic Prototype: Byte Q6_p_not_any8_vcmpb_eq_PP(Word64 Rss, Word64 Rtt) + Instruction Type: ALU64 + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_p_not_any8_vcmpb_eq_PP __builtin_HEXAGON_A6_vcmpbeq_notany +#endif /* __HEXAGON_ARCH___ >= 65 */ + +#if __HEXAGON_ARCH__ >= 66 +/* ========================================================================== + Assembly Syntax: Rdd32=dfadd(Rss32,Rtt32) + C Intrinsic Prototype: Float64 Q6_P_dfadd_PP(Float64 Rss, Float64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_dfadd_PP __builtin_HEXAGON_F2_dfadd +#endif /* __HEXAGON_ARCH___ >= 66 */ + +#if __HEXAGON_ARCH__ >= 66 +/* ========================================================================== + Assembly Syntax: Rdd32=dfsub(Rss32,Rtt32) + C Intrinsic Prototype: Float64 Q6_P_dfsub_PP(Float64 Rss, Float64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_dfsub_PP __builtin_HEXAGON_F2_dfsub +#endif /* __HEXAGON_ARCH___ >= 66 */ + +#if __HEXAGON_ARCH__ >= 66 +/* ========================================================================== + Assembly Syntax: Rx32-=mpyi(Rs32,Rt32) + C Intrinsic Prototype: Word32 Q6_R_mpyinac_RR(Word32 Rx, Word32 Rs, Word32 Rt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mpyinac_RR __builtin_HEXAGON_M2_mnaci +#endif /* __HEXAGON_ARCH___ >= 66 */ + +#if __HEXAGON_ARCH__ >= 66 +/* ========================================================================== + Assembly Syntax: Rd32=mask(#u5,#U5) + C Intrinsic Prototype: Word32 Q6_R_mask_II(Word32 Iu5, Word32 IU5) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_mask_II __builtin_HEXAGON_S2_mask +#endif /* __HEXAGON_ARCH___ >= 66 */ + +#if __HEXAGON_ARCH__ >= 67 && defined __HEXAGON_AUDIO__ +/* ========================================================================== + Assembly Syntax: Rd32=clip(Rs32,#u5) + C Intrinsic Prototype: Word32 Q6_R_clip_RI(Word32 Rs, Word32 Iu5) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_R_clip_RI __builtin_HEXAGON_A7_clip +#endif /* __HEXAGON_ARCH___ >= 67 && defined __HEXAGON_AUDIO__*/ + +#if __HEXAGON_ARCH__ >= 67 && defined __HEXAGON_AUDIO__ +/* ========================================================================== + Assembly Syntax: Rdd32=cround(Rss32,#u6) + C Intrinsic Prototype: Word64 Q6_P_cround_PI(Word64 Rss, Word32 Iu6) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_cround_PI __builtin_HEXAGON_A7_croundd_ri +#endif /* __HEXAGON_ARCH___ >= 67 && defined __HEXAGON_AUDIO__*/ + +#if __HEXAGON_ARCH__ >= 67 && defined __HEXAGON_AUDIO__ +/* ========================================================================== + Assembly Syntax: Rdd32=cround(Rss32,Rt32) + C Intrinsic Prototype: Word64 Q6_P_cround_PR(Word64 Rss, Word32 Rt) + Instruction Type: S_3op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_cround_PR __builtin_HEXAGON_A7_croundd_rr +#endif /* __HEXAGON_ARCH___ >= 67 && defined __HEXAGON_AUDIO__*/ + +#if __HEXAGON_ARCH__ >= 67 && defined __HEXAGON_AUDIO__ +/* ========================================================================== + Assembly Syntax: Rdd32=vclip(Rss32,#u5) + C Intrinsic Prototype: Word64 Q6_P_vclip_PI(Word64 Rss, Word32 Iu5) + Instruction Type: S_2op + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_vclip_PI __builtin_HEXAGON_A7_vclip +#endif /* __HEXAGON_ARCH___ >= 67 && defined __HEXAGON_AUDIO__*/ + +#if __HEXAGON_ARCH__ >= 67 +/* ========================================================================== + Assembly Syntax: Rdd32=dfmax(Rss32,Rtt32) + C Intrinsic Prototype: Float64 Q6_P_dfmax_PP(Float64 Rss, Float64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_dfmax_PP __builtin_HEXAGON_F2_dfmax +#endif /* __HEXAGON_ARCH___ >= 67 */ + +#if __HEXAGON_ARCH__ >= 67 +/* ========================================================================== + Assembly Syntax: Rdd32=dfmin(Rss32,Rtt32) + C Intrinsic Prototype: Float64 Q6_P_dfmin_PP(Float64 Rss, Float64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_dfmin_PP __builtin_HEXAGON_F2_dfmin +#endif /* __HEXAGON_ARCH___ >= 67 */ + +#if __HEXAGON_ARCH__ >= 67 +/* ========================================================================== + Assembly Syntax: Rdd32=dfmpyfix(Rss32,Rtt32) + C Intrinsic Prototype: Float64 Q6_P_dfmpyfix_PP(Float64 Rss, Float64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_dfmpyfix_PP __builtin_HEXAGON_F2_dfmpyfix +#endif /* __HEXAGON_ARCH___ >= 67 */ + +#if __HEXAGON_ARCH__ >= 67 +/* ========================================================================== + Assembly Syntax: Rxx32+=dfmpyhh(Rss32,Rtt32) + C Intrinsic Prototype: Float64 Q6_P_dfmpyhhacc_PP(Float64 Rxx, Float64 Rss, Float64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_dfmpyhhacc_PP __builtin_HEXAGON_F2_dfmpyhh +#endif /* __HEXAGON_ARCH___ >= 67 */ + +#if __HEXAGON_ARCH__ >= 67 +/* ========================================================================== + Assembly Syntax: Rxx32+=dfmpylh(Rss32,Rtt32) + C Intrinsic Prototype: Float64 Q6_P_dfmpylhacc_PP(Float64 Rxx, Float64 Rss, Float64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_dfmpylhacc_PP __builtin_HEXAGON_F2_dfmpylh +#endif /* __HEXAGON_ARCH___ >= 67 */ + +#if __HEXAGON_ARCH__ >= 67 +/* ========================================================================== + Assembly Syntax: Rdd32=dfmpyll(Rss32,Rtt32) + C Intrinsic Prototype: Float64 Q6_P_dfmpyll_PP(Float64 Rss, Float64 Rtt) + Instruction Type: M + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_P_dfmpyll_PP __builtin_HEXAGON_F2_dfmpyll +#endif /* __HEXAGON_ARCH___ >= 67 */ + +#if __HEXAGON_ARCH__ >= 67 && defined __HEXAGON_AUDIO__ +/* ========================================================================== + Assembly Syntax: Rdd32=cmpyiw(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_cmpyiw_PP(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT3 + ========================================================================== */ + +#define Q6_P_cmpyiw_PP __builtin_HEXAGON_M7_dcmpyiw +#endif /* __HEXAGON_ARCH___ >= 67 && defined __HEXAGON_AUDIO__*/ + +#if __HEXAGON_ARCH__ >= 67 && defined __HEXAGON_AUDIO__ +/* ========================================================================== + Assembly Syntax: Rxx32+=cmpyiw(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_cmpyiwacc_PP(Word64 Rxx, Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT3 + ========================================================================== */ + +#define Q6_P_cmpyiwacc_PP __builtin_HEXAGON_M7_dcmpyiw_acc +#endif /* __HEXAGON_ARCH___ >= 67 && defined __HEXAGON_AUDIO__*/ + +#if __HEXAGON_ARCH__ >= 67 && defined __HEXAGON_AUDIO__ +/* ========================================================================== + Assembly Syntax: Rdd32=cmpyiw(Rss32,Rtt32*) + C Intrinsic Prototype: Word64 Q6_P_cmpyiw_PP_conj(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT3 + ========================================================================== */ + +#define Q6_P_cmpyiw_PP_conj __builtin_HEXAGON_M7_dcmpyiwc +#endif /* __HEXAGON_ARCH___ >= 67 && defined __HEXAGON_AUDIO__*/ + +#if __HEXAGON_ARCH__ >= 67 && defined __HEXAGON_AUDIO__ +/* ========================================================================== + Assembly Syntax: Rxx32+=cmpyiw(Rss32,Rtt32*) + C Intrinsic Prototype: Word64 Q6_P_cmpyiwacc_PP_conj(Word64 Rxx, Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT3 + ========================================================================== */ + +#define Q6_P_cmpyiwacc_PP_conj __builtin_HEXAGON_M7_dcmpyiwc_acc +#endif /* __HEXAGON_ARCH___ >= 67 && defined __HEXAGON_AUDIO__*/ + +#if __HEXAGON_ARCH__ >= 67 && defined __HEXAGON_AUDIO__ +/* ========================================================================== + Assembly Syntax: Rdd32=cmpyrw(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_cmpyrw_PP(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT3 + ========================================================================== */ + +#define Q6_P_cmpyrw_PP __builtin_HEXAGON_M7_dcmpyrw +#endif /* __HEXAGON_ARCH___ >= 67 && defined __HEXAGON_AUDIO__*/ + +#if __HEXAGON_ARCH__ >= 67 && defined __HEXAGON_AUDIO__ +/* ========================================================================== + Assembly Syntax: Rxx32+=cmpyrw(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_cmpyrwacc_PP(Word64 Rxx, Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT3 + ========================================================================== */ + +#define Q6_P_cmpyrwacc_PP __builtin_HEXAGON_M7_dcmpyrw_acc +#endif /* __HEXAGON_ARCH___ >= 67 && defined __HEXAGON_AUDIO__*/ + +#if __HEXAGON_ARCH__ >= 67 && defined __HEXAGON_AUDIO__ +/* ========================================================================== + Assembly Syntax: Rdd32=cmpyrw(Rss32,Rtt32*) + C Intrinsic Prototype: Word64 Q6_P_cmpyrw_PP_conj(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT3 + ========================================================================== */ + +#define Q6_P_cmpyrw_PP_conj __builtin_HEXAGON_M7_dcmpyrwc +#endif /* __HEXAGON_ARCH___ >= 67 && defined __HEXAGON_AUDIO__*/ + +#if __HEXAGON_ARCH__ >= 67 && defined __HEXAGON_AUDIO__ +/* ========================================================================== + Assembly Syntax: Rxx32+=cmpyrw(Rss32,Rtt32*) + C Intrinsic Prototype: Word64 Q6_P_cmpyrwacc_PP_conj(Word64 Rxx, Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT3 + ========================================================================== */ + +#define Q6_P_cmpyrwacc_PP_conj __builtin_HEXAGON_M7_dcmpyrwc_acc +#endif /* __HEXAGON_ARCH___ >= 67 && defined __HEXAGON_AUDIO__*/ + +#if __HEXAGON_ARCH__ >= 67 && defined __HEXAGON_AUDIO__ +/* ========================================================================== + Assembly Syntax: Rdd32=vdmpyw(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_vdmpyw_PP(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT3 + ========================================================================== */ + +#define Q6_P_vdmpyw_PP __builtin_HEXAGON_M7_vdmpy +#endif /* __HEXAGON_ARCH___ >= 67 && defined __HEXAGON_AUDIO__*/ + +#if __HEXAGON_ARCH__ >= 67 && defined __HEXAGON_AUDIO__ +/* ========================================================================== + Assembly Syntax: Rxx32+=vdmpyw(Rss32,Rtt32) + C Intrinsic Prototype: Word64 Q6_P_vdmpywacc_PP(Word64 Rxx, Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT3 + ========================================================================== */ + +#define Q6_P_vdmpywacc_PP __builtin_HEXAGON_M7_vdmpy_acc +#endif /* __HEXAGON_ARCH___ >= 67 && defined __HEXAGON_AUDIO__*/ + +#if __HEXAGON_ARCH__ >= 67 && defined __HEXAGON_AUDIO__ +/* ========================================================================== + Assembly Syntax: Rd32=cmpyiw(Rss32,Rtt32):<<1:sat + C Intrinsic Prototype: Word32 Q6_R_cmpyiw_PP_s1_sat(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT3 + ========================================================================== */ + +#define Q6_R_cmpyiw_PP_s1_sat __builtin_HEXAGON_M7_wcmpyiw +#endif /* __HEXAGON_ARCH___ >= 67 && defined __HEXAGON_AUDIO__*/ + +#if __HEXAGON_ARCH__ >= 67 && defined __HEXAGON_AUDIO__ +/* ========================================================================== + Assembly Syntax: Rd32=cmpyiw(Rss32,Rtt32):<<1:rnd:sat + C Intrinsic Prototype: Word32 Q6_R_cmpyiw_PP_s1_rnd_sat(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT3 + ========================================================================== */ + +#define Q6_R_cmpyiw_PP_s1_rnd_sat __builtin_HEXAGON_M7_wcmpyiw_rnd +#endif /* __HEXAGON_ARCH___ >= 67 && defined __HEXAGON_AUDIO__*/ + +#if __HEXAGON_ARCH__ >= 67 && defined __HEXAGON_AUDIO__ +/* ========================================================================== + Assembly Syntax: Rd32=cmpyiw(Rss32,Rtt32*):<<1:sat + C Intrinsic Prototype: Word32 Q6_R_cmpyiw_PP_conj_s1_sat(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT3 + ========================================================================== */ + +#define Q6_R_cmpyiw_PP_conj_s1_sat __builtin_HEXAGON_M7_wcmpyiwc +#endif /* __HEXAGON_ARCH___ >= 67 && defined __HEXAGON_AUDIO__*/ + +#if __HEXAGON_ARCH__ >= 67 && defined __HEXAGON_AUDIO__ +/* ========================================================================== + Assembly Syntax: Rd32=cmpyiw(Rss32,Rtt32*):<<1:rnd:sat + C Intrinsic Prototype: Word32 Q6_R_cmpyiw_PP_conj_s1_rnd_sat(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT3 + ========================================================================== */ + +#define Q6_R_cmpyiw_PP_conj_s1_rnd_sat __builtin_HEXAGON_M7_wcmpyiwc_rnd +#endif /* __HEXAGON_ARCH___ >= 67 && defined __HEXAGON_AUDIO__*/ + +#if __HEXAGON_ARCH__ >= 67 && defined __HEXAGON_AUDIO__ +/* ========================================================================== + Assembly Syntax: Rd32=cmpyrw(Rss32,Rtt32):<<1:sat + C Intrinsic Prototype: Word32 Q6_R_cmpyrw_PP_s1_sat(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT3 + ========================================================================== */ + +#define Q6_R_cmpyrw_PP_s1_sat __builtin_HEXAGON_M7_wcmpyrw +#endif /* __HEXAGON_ARCH___ >= 67 && defined __HEXAGON_AUDIO__*/ + +#if __HEXAGON_ARCH__ >= 67 && defined __HEXAGON_AUDIO__ +/* ========================================================================== + Assembly Syntax: Rd32=cmpyrw(Rss32,Rtt32):<<1:rnd:sat + C Intrinsic Prototype: Word32 Q6_R_cmpyrw_PP_s1_rnd_sat(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT3 + ========================================================================== */ + +#define Q6_R_cmpyrw_PP_s1_rnd_sat __builtin_HEXAGON_M7_wcmpyrw_rnd +#endif /* __HEXAGON_ARCH___ >= 67 && defined __HEXAGON_AUDIO__*/ + +#if __HEXAGON_ARCH__ >= 67 && defined __HEXAGON_AUDIO__ +/* ========================================================================== + Assembly Syntax: Rd32=cmpyrw(Rss32,Rtt32*):<<1:sat + C Intrinsic Prototype: Word32 Q6_R_cmpyrw_PP_conj_s1_sat(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT3 + ========================================================================== */ + +#define Q6_R_cmpyrw_PP_conj_s1_sat __builtin_HEXAGON_M7_wcmpyrwc +#endif /* __HEXAGON_ARCH___ >= 67 && defined __HEXAGON_AUDIO__*/ + +#if __HEXAGON_ARCH__ >= 67 && defined __HEXAGON_AUDIO__ +/* ========================================================================== + Assembly Syntax: Rd32=cmpyrw(Rss32,Rtt32*):<<1:rnd:sat + C Intrinsic Prototype: Word32 Q6_R_cmpyrw_PP_conj_s1_rnd_sat(Word64 Rss, Word64 Rtt) + Instruction Type: M + Execution Slots: SLOT3 + ========================================================================== */ + +#define Q6_R_cmpyrw_PP_conj_s1_rnd_sat __builtin_HEXAGON_M7_wcmpyrwc_rnd +#endif /* __HEXAGON_ARCH___ >= 67 && defined __HEXAGON_AUDIO__*/ + +#if __HEXAGON_ARCH__ >= 68 +/* ========================================================================== + Assembly Syntax: dmlink(Rs32,Rt32) + C Intrinsic Prototype: void Q6_dmlink_AA(Address Rs, Address Rt) + Instruction Type: ST + Execution Slots: SLOT0 + ========================================================================== */ + +#define Q6_dmlink_AA __builtin_HEXAGON_Y6_dmlink +#endif /* __HEXAGON_ARCH___ >= 68 */ + +#if __HEXAGON_ARCH__ >= 68 +/* ========================================================================== + Assembly Syntax: Rd32=dmpause + C Intrinsic Prototype: Word32 Q6_R_dmpause() + Instruction Type: ST + Execution Slots: SLOT0 + ========================================================================== */ + +#define Q6_R_dmpause __builtin_HEXAGON_Y6_dmpause +#endif /* __HEXAGON_ARCH___ >= 68 */ + +#if __HEXAGON_ARCH__ >= 68 +/* ========================================================================== + Assembly Syntax: Rd32=dmpoll + C Intrinsic Prototype: Word32 Q6_R_dmpoll() + Instruction Type: ST + Execution Slots: SLOT0 + ========================================================================== */ + +#define Q6_R_dmpoll __builtin_HEXAGON_Y6_dmpoll +#endif /* __HEXAGON_ARCH___ >= 68 */ + +#if __HEXAGON_ARCH__ >= 68 +/* ========================================================================== + Assembly Syntax: dmresume(Rs32) + C Intrinsic Prototype: void Q6_dmresume_A(Address Rs) + Instruction Type: ST + Execution Slots: SLOT0 + ========================================================================== */ + +#define Q6_dmresume_A __builtin_HEXAGON_Y6_dmresume +#endif /* __HEXAGON_ARCH___ >= 68 */ + +#if __HEXAGON_ARCH__ >= 68 +/* ========================================================================== + Assembly Syntax: dmstart(Rs32) + C Intrinsic Prototype: void Q6_dmstart_A(Address Rs) + Instruction Type: ST + Execution Slots: SLOT0 + ========================================================================== */ + +#define Q6_dmstart_A __builtin_HEXAGON_Y6_dmstart +#endif /* __HEXAGON_ARCH___ >= 68 */ + +#if __HEXAGON_ARCH__ >= 68 +/* ========================================================================== + Assembly Syntax: Rd32=dmwait + C Intrinsic Prototype: Word32 Q6_R_dmwait() + Instruction Type: ST + Execution Slots: SLOT0 + ========================================================================== */ + +#define Q6_R_dmwait __builtin_HEXAGON_Y6_dmwait +#endif /* __HEXAGON_ARCH___ >= 68 */ + +#include +#ifdef __HVX__ +#include +#endif /* __HVX__ */ +#endif diff --git a/lib/include/hexagon_types.h b/lib/include/hexagon_types.h new file mode 100644 index 000000000..695880941 --- /dev/null +++ b/lib/include/hexagon_types.h @@ -0,0 +1,2653 @@ +/******************************************************************************/ +/* (c) 2020 Qualcomm Innovation Center, Inc. All rights reserved. */ +/* */ +/******************************************************************************/ +#ifndef HEXAGON_TYPES_H +#define HEXAGON_TYPES_H + +#include + +/* Hexagon names */ +#define HEXAGON_Vect HEXAGON_Vect64 +#define HEXAGON_V_GET_D HEXAGON_V64_GET_D +#define HEXAGON_V_GET_UD HEXAGON_V64_GET_UD +#define HEXAGON_V_GET_W0 HEXAGON_V64_GET_W0 +#define HEXAGON_V_GET_W1 HEXAGON_V64_GET_W1 +#define HEXAGON_V_GET_UW0 HEXAGON_V64_GET_UW0 +#define HEXAGON_V_GET_UW1 HEXAGON_V64_GET_UW1 +#define HEXAGON_V_GET_H0 HEXAGON_V64_GET_H0 +#define HEXAGON_V_GET_H1 HEXAGON_V64_GET_H1 +#define HEXAGON_V_GET_H2 HEXAGON_V64_GET_H2 +#define HEXAGON_V_GET_H3 HEXAGON_V64_GET_H3 +#define HEXAGON_V_GET_UH0 HEXAGON_V64_GET_UH0 +#define HEXAGON_V_GET_UH1 HEXAGON_V64_GET_UH1 +#define HEXAGON_V_GET_UH2 HEXAGON_V64_GET_UH2 +#define HEXAGON_V_GET_UH3 HEXAGON_V64_GET_UH3 +#define HEXAGON_V_GET_B0 HEXAGON_V64_GET_B0 +#define HEXAGON_V_GET_B1 HEXAGON_V64_GET_B1 +#define HEXAGON_V_GET_B2 HEXAGON_V64_GET_B2 +#define HEXAGON_V_GET_B3 HEXAGON_V64_GET_B3 +#define HEXAGON_V_GET_B4 HEXAGON_V64_GET_B4 +#define HEXAGON_V_GET_B5 HEXAGON_V64_GET_B5 +#define HEXAGON_V_GET_B6 HEXAGON_V64_GET_B6 +#define HEXAGON_V_GET_B7 HEXAGON_V64_GET_B7 +#define HEXAGON_V_GET_UB0 HEXAGON_V64_GET_UB0 +#define HEXAGON_V_GET_UB1 HEXAGON_V64_GET_UB1 +#define HEXAGON_V_GET_UB2 HEXAGON_V64_GET_UB2 +#define HEXAGON_V_GET_UB3 HEXAGON_V64_GET_UB3 +#define HEXAGON_V_GET_UB4 HEXAGON_V64_GET_UB4 +#define HEXAGON_V_GET_UB5 HEXAGON_V64_GET_UB5 +#define HEXAGON_V_GET_UB6 HEXAGON_V64_GET_UB6 +#define HEXAGON_V_GET_UB7 HEXAGON_V64_GET_UB7 +#define HEXAGON_V_PUT_D HEXAGON_V64_PUT_D +#define HEXAGON_V_PUT_W0 HEXAGON_V64_PUT_W0 +#define HEXAGON_V_PUT_W1 HEXAGON_V64_PUT_W1 +#define HEXAGON_V_PUT_H0 HEXAGON_V64_PUT_H0 +#define HEXAGON_V_PUT_H1 HEXAGON_V64_PUT_H1 +#define HEXAGON_V_PUT_H2 HEXAGON_V64_PUT_H2 +#define HEXAGON_V_PUT_H3 HEXAGON_V64_PUT_H3 +#define HEXAGON_V_PUT_B0 HEXAGON_V64_PUT_B0 +#define HEXAGON_V_PUT_B1 HEXAGON_V64_PUT_B1 +#define HEXAGON_V_PUT_B2 HEXAGON_V64_PUT_B2 +#define HEXAGON_V_PUT_B3 HEXAGON_V64_PUT_B3 +#define HEXAGON_V_PUT_B4 HEXAGON_V64_PUT_B4 +#define HEXAGON_V_PUT_B5 HEXAGON_V64_PUT_B5 +#define HEXAGON_V_PUT_B6 HEXAGON_V64_PUT_B6 +#define HEXAGON_V_PUT_B7 HEXAGON_V64_PUT_B7 +#define HEXAGON_V_CREATE_D HEXAGON_V64_CREATE_D +#define HEXAGON_V_CREATE_W HEXAGON_V64_CREATE_W +#define HEXAGON_V_CREATE_H HEXAGON_V64_CREATE_H +#define HEXAGON_V_CREATE_B HEXAGON_V64_CREATE_B + +#ifdef __cplusplus +#define HEXAGON_VectC HEXAGON_Vect64C +#endif /* __cplusplus */ + +/* 64 Bit Vectors */ + +typedef long long __attribute__((__may_alias__)) HEXAGON_Vect64; + +/* Extract doubleword macros */ + +#define HEXAGON_V64_GET_D(v) (v) +#define HEXAGON_V64_GET_UD(v) ((unsigned long long)(v)) + +/* Extract word macros */ + +#define HEXAGON_V64_GET_W0(v) \ + __extension__({ \ + union { \ + long long d; \ + int w[2]; \ + } _HEXAGON_V64_internal_union; \ + _HEXAGON_V64_internal_union.d = (v); \ + _HEXAGON_V64_internal_union.w[0]; \ + }) +#define HEXAGON_V64_GET_W1(v) \ + __extension__({ \ + union { \ + long long d; \ + int w[2]; \ + } _HEXAGON_V64_internal_union; \ + _HEXAGON_V64_internal_union.d = (v); \ + _HEXAGON_V64_internal_union.w[1]; \ + }) +#define HEXAGON_V64_GET_UW0(v) \ + __extension__({ \ + union { \ + long long d; \ + unsigned int uw[2]; \ + } _HEXAGON_V64_internal_union; \ + _HEXAGON_V64_internal_union.d = (v); \ + _HEXAGON_V64_internal_union.uw[0]; \ + }) +#define HEXAGON_V64_GET_UW1(v) \ + __extension__({ \ + union { \ + long long d; \ + unsigned int uw[2]; \ + } _HEXAGON_V64_internal_union; \ + _HEXAGON_V64_internal_union.d = (v); \ + _HEXAGON_V64_internal_union.uw[1]; \ + }) + +/* Extract half word macros */ + +#define HEXAGON_V64_GET_H0(v) \ + __extension__({ \ + union { \ + long long d; \ + short h[4]; \ + } _HEXAGON_V64_internal_union; \ + _HEXAGON_V64_internal_union.d = (v); \ + _HEXAGON_V64_internal_union.h[0]; \ + }) +#define HEXAGON_V64_GET_H1(v) \ + __extension__({ \ + union { \ + long long d; \ + short h[4]; \ + } _HEXAGON_V64_internal_union; \ + _HEXAGON_V64_internal_union.d = (v); \ + _HEXAGON_V64_internal_union.h[1]; \ + }) +#define HEXAGON_V64_GET_H2(v) \ + __extension__({ \ + union { \ + long long d; \ + short h[4]; \ + } _HEXAGON_V64_internal_union; \ + _HEXAGON_V64_internal_union.d = (v); \ + _HEXAGON_V64_internal_union.h[2]; \ + }) +#define HEXAGON_V64_GET_H3(v) \ + __extension__({ \ + union { \ + long long d; \ + short h[4]; \ + } _HEXAGON_V64_internal_union; \ + _HEXAGON_V64_internal_union.d = (v); \ + _HEXAGON_V64_internal_union.h[3]; \ + }) +#define HEXAGON_V64_GET_UH0(v) \ + __extension__({ \ + union { \ + long long d; \ + unsigned short uh[4]; \ + } _HEXAGON_V64_internal_union; \ + _HEXAGON_V64_internal_union.d = (v); \ + _HEXAGON_V64_internal_union.uh[0]; \ + }) +#define HEXAGON_V64_GET_UH1(v) \ + __extension__({ \ + union { \ + long long d; \ + unsigned short uh[4]; \ + } _HEXAGON_V64_internal_union; \ + _HEXAGON_V64_internal_union.d = (v); \ + _HEXAGON_V64_internal_union.uh[1]; \ + }) +#define HEXAGON_V64_GET_UH2(v) \ + __extension__({ \ + union { \ + long long d; \ + unsigned short uh[4]; \ + } _HEXAGON_V64_internal_union; \ + _HEXAGON_V64_internal_union.d = (v); \ + _HEXAGON_V64_internal_union.uh[2]; \ + }) +#define HEXAGON_V64_GET_UH3(v) \ + __extension__({ \ + union { \ + long long d; \ + unsigned short uh[4]; \ + } _HEXAGON_V64_internal_union; \ + _HEXAGON_V64_internal_union.d = (v); \ + _HEXAGON_V64_internal_union.uh[3]; \ + }) + +/* Extract byte macros */ + +#define HEXAGON_V64_GET_B0(v) \ + __extension__({ \ + union { \ + long long d; \ + signed char b[8]; \ + } _HEXAGON_V64_internal_union; \ + _HEXAGON_V64_internal_union.d = (v); \ + _HEXAGON_V64_internal_union.b[0]; \ + }) +#define HEXAGON_V64_GET_B1(v) \ + __extension__({ \ + union { \ + long long d; \ + signed char b[8]; \ + } _HEXAGON_V64_internal_union; \ + _HEXAGON_V64_internal_union.d = (v); \ + _HEXAGON_V64_internal_union.b[1]; \ + }) +#define HEXAGON_V64_GET_B2(v) \ + __extension__({ \ + union { \ + long long d; \ + signed char b[8]; \ + } _HEXAGON_V64_internal_union; \ + _HEXAGON_V64_internal_union.d = (v); \ + _HEXAGON_V64_internal_union.b[2]; \ + }) +#define HEXAGON_V64_GET_B3(v) \ + __extension__({ \ + union { \ + long long d; \ + signed char b[8]; \ + } _HEXAGON_V64_internal_union; \ + _HEXAGON_V64_internal_union.d = (v); \ + _HEXAGON_V64_internal_union.b[3]; \ + }) +#define HEXAGON_V64_GET_B4(v) \ + __extension__({ \ + union { \ + long long d; \ + signed char b[8]; \ + } _HEXAGON_V64_internal_union; \ + _HEXAGON_V64_internal_union.d = (v); \ + _HEXAGON_V64_internal_union.b[4]; \ + }) +#define HEXAGON_V64_GET_B5(v) \ + __extension__({ \ + union { \ + long long d; \ + signed char b[8]; \ + } _HEXAGON_V64_internal_union; \ + _HEXAGON_V64_internal_union.d = (v); \ + _HEXAGON_V64_internal_union.b[5]; \ + }) +#define HEXAGON_V64_GET_B6(v) \ + __extension__({ \ + union { \ + long long d; \ + signed char b[8]; \ + } _HEXAGON_V64_internal_union; \ + _HEXAGON_V64_internal_union.d = (v); \ + _HEXAGON_V64_internal_union.b[6]; \ + }) +#define HEXAGON_V64_GET_B7(v) \ + __extension__({ \ + union { \ + long long d; \ + signed char b[8]; \ + } _HEXAGON_V64_internal_union; \ + _HEXAGON_V64_internal_union.d = (v); \ + _HEXAGON_V64_internal_union.b[7]; \ + }) +#define HEXAGON_V64_GET_UB0(v) \ + __extension__({ \ + union { \ + long long d; \ + unsigned char ub[8]; \ + } _HEXAGON_V64_internal_union; \ + _HEXAGON_V64_internal_union.d = (v); \ + _HEXAGON_V64_internal_union.ub[0]; \ + }) +#define HEXAGON_V64_GET_UB1(v) \ + __extension__({ \ + union { \ + long long d; \ + unsigned char ub[8]; \ + } _HEXAGON_V64_internal_union; \ + _HEXAGON_V64_internal_union.d = (v); \ + _HEXAGON_V64_internal_union.ub[1]; \ + }) +#define HEXAGON_V64_GET_UB2(v) \ + __extension__({ \ + union { \ + long long d; \ + unsigned char ub[8]; \ + } _HEXAGON_V64_internal_union; \ + _HEXAGON_V64_internal_union.d = (v); \ + _HEXAGON_V64_internal_union.ub[2]; \ + }) +#define HEXAGON_V64_GET_UB3(v) \ + __extension__({ \ + union { \ + long long d; \ + unsigned char ub[8]; \ + } _HEXAGON_V64_internal_union; \ + _HEXAGON_V64_internal_union.d = (v); \ + _HEXAGON_V64_internal_union.ub[3]; \ + }) +#define HEXAGON_V64_GET_UB4(v) \ + __extension__({ \ + union { \ + long long d; \ + unsigned char ub[8]; \ + } _HEXAGON_V64_internal_union; \ + _HEXAGON_V64_internal_union.d = (v); \ + _HEXAGON_V64_internal_union.ub[4]; \ + }) +#define HEXAGON_V64_GET_UB5(v) \ + __extension__({ \ + union { \ + long long d; \ + unsigned char ub[8]; \ + } _HEXAGON_V64_internal_union; \ + _HEXAGON_V64_internal_union.d = (v); \ + _HEXAGON_V64_internal_union.ub[5]; \ + }) +#define HEXAGON_V64_GET_UB6(v) \ + __extension__({ \ + union { \ + long long d; \ + unsigned char ub[8]; \ + } _HEXAGON_V64_internal_union; \ + _HEXAGON_V64_internal_union.d = (v); \ + _HEXAGON_V64_internal_union.ub[6]; \ + }) +#define HEXAGON_V64_GET_UB7(v) \ + __extension__({ \ + union { \ + long long d; \ + unsigned char ub[8]; \ + } _HEXAGON_V64_internal_union; \ + _HEXAGON_V64_internal_union.d = (v); \ + _HEXAGON_V64_internal_union.ub[7]; \ + }) + +/* NOTE: All set macros return a HEXAGON_Vect64 type */ + +/* Set doubleword macro */ + +#define HEXAGON_V64_PUT_D(v, new) (new) + +/* Set word macros */ + +#ifdef __hexagon__ + +#define HEXAGON_V64_PUT_W0(v, new) \ + __extension__({ \ + union { \ + long long d; \ + int w[2]; \ + } _HEXAGON_V64_internal_union; \ + _HEXAGON_V64_internal_union.d = (v); \ + _HEXAGON_V64_internal_union.w[0] = (new); \ + _HEXAGON_V64_internal_union.d; \ + }) +#define HEXAGON_V64_PUT_W1(v, new) \ + __extension__({ \ + union { \ + long long d; \ + int w[2]; \ + } _HEXAGON_V64_internal_union; \ + _HEXAGON_V64_internal_union.d = (v); \ + _HEXAGON_V64_internal_union.w[1] = (new); \ + _HEXAGON_V64_internal_union.d; \ + }) + +#else /* !__hexagon__ */ + +#define HEXAGON_V64_PUT_W0(v, new) \ + (((v) & 0xffffffff00000000LL) | ((HEXAGON_Vect64)((unsigned int)(new)))) +#define HEXAGON_V64_PUT_W1(v, new) \ + (((v) & 0x00000000ffffffffLL) | (((HEXAGON_Vect64)(new)) << 32LL)) + +#endif /* !__hexagon__ */ + +/* Set half word macros */ + +#ifdef __hexagon__ + +#define HEXAGON_V64_PUT_H0(v, new) \ + __extension__({ \ + union { \ + long long d; \ + short h[4]; \ + } _HEXAGON_V64_internal_union; \ + _HEXAGON_V64_internal_union.d = (v); \ + _HEXAGON_V64_internal_union.h[0] = (new); \ + _HEXAGON_V64_internal_union.d; \ + }) +#define HEXAGON_V64_PUT_H1(v, new) \ + __extension__({ \ + union { \ + long long d; \ + short h[4]; \ + } _HEXAGON_V64_internal_union; \ + _HEXAGON_V64_internal_union.d = (v); \ + _HEXAGON_V64_internal_union.h[1] = (new); \ + _HEXAGON_V64_internal_union.d; \ + }) +#define HEXAGON_V64_PUT_H2(v, new) \ + __extension__({ \ + union { \ + long long d; \ + short h[4]; \ + } _HEXAGON_V64_internal_union; \ + _HEXAGON_V64_internal_union.d = (v); \ + _HEXAGON_V64_internal_union.h[2] = (new); \ + _HEXAGON_V64_internal_union.d; \ + }) +#define HEXAGON_V64_PUT_H3(v, new) \ + __extension__({ \ + union { \ + long long d; \ + short h[4]; \ + } _HEXAGON_V64_internal_union; \ + _HEXAGON_V64_internal_union.d = (v); \ + _HEXAGON_V64_internal_union.h[3] = (new); \ + _HEXAGON_V64_internal_union.d; \ + }) + +#else /* !__hexagon__ */ + +#define HEXAGON_V64_PUT_H0(v, new) \ + (((v) & 0xffffffffffff0000LL) | ((HEXAGON_Vect64)((unsigned short)(new)))) +#define HEXAGON_V64_PUT_H1(v, new) \ + (((v) & 0xffffffff0000ffffLL) | (((HEXAGON_Vect64)((unsigned short)(new))) << 16LL)) +#define HEXAGON_V64_PUT_H2(v, new) \ + (((v) & 0xffff0000ffffffffLL) | (((HEXAGON_Vect64)((unsigned short)(new))) << 32LL)) +#define HEXAGON_V64_PUT_H3(v, new) \ + (((v) & 0x0000ffffffffffffLL) | (((HEXAGON_Vect64)(new)) << 48LL)) + +#endif /* !__hexagon__ */ + +/* Set byte macros */ + +#ifdef __hexagon__ + +#define HEXAGON_V64_PUT_B0(v, new) \ + __extension__({ \ + union { \ + long long d; \ + char b[8]; \ + } _HEXAGON_V64_internal_union; \ + _HEXAGON_V64_internal_union.d = (v); \ + _HEXAGON_V64_internal_union.b[0] = (new); \ + _HEXAGON_V64_internal_union.d; \ + }) +#define HEXAGON_V64_PUT_B1(v, new) \ + __extension__({ \ + union { \ + long long d; \ + char b[8]; \ + } _HEXAGON_V64_internal_union; \ + _HEXAGON_V64_internal_union.d = (v); \ + _HEXAGON_V64_internal_union.b[1] = (new); \ + _HEXAGON_V64_internal_union.d; \ + }) +#define HEXAGON_V64_PUT_B2(v, new) \ + __extension__({ \ + union { \ + long long d; \ + char b[8]; \ + } _HEXAGON_V64_internal_union; \ + _HEXAGON_V64_internal_union.d = (v); \ + _HEXAGON_V64_internal_union.b[2] = (new); \ + _HEXAGON_V64_internal_union.d; \ + }) +#define HEXAGON_V64_PUT_B3(v, new) \ + __extension__({ \ + union { \ + long long d; \ + char b[8]; \ + } _HEXAGON_V64_internal_union; \ + _HEXAGON_V64_internal_union.d = (v); \ + _HEXAGON_V64_internal_union.b[3] = (new); \ + _HEXAGON_V64_internal_union.d; \ + }) +#define HEXAGON_V64_PUT_B4(v, new) \ + __extension__({ \ + union { \ + long long d; \ + char b[8]; \ + } _HEXAGON_V64_internal_union; \ + _HEXAGON_V64_internal_union.d = (v); \ + _HEXAGON_V64_internal_union.b[4] = (new); \ + _HEXAGON_V64_internal_union.d; \ + }) +#define HEXAGON_V64_PUT_B5(v, new) \ + __extension__({ \ + union { \ + long long d; \ + char b[8]; \ + } _HEXAGON_V64_internal_union; \ + _HEXAGON_V64_internal_union.d = (v); \ + _HEXAGON_V64_internal_union.b[5] = (new); \ + _HEXAGON_V64_internal_union.d; \ + }) +#define HEXAGON_V64_PUT_B6(v, new) \ + __extension__({ \ + union { \ + long long d; \ + char b[8]; \ + } _HEXAGON_V64_internal_union; \ + _HEXAGON_V64_internal_union.d = (v); \ + _HEXAGON_V64_internal_union.b[6] = (new); \ + _HEXAGON_V64_internal_union.d; \ + }) +#define HEXAGON_V64_PUT_B7(v, new) \ + __extension__({ \ + union { \ + long long d; \ + char b[8]; \ + } _HEXAGON_V64_internal_union; \ + _HEXAGON_V64_internal_union.d = (v); \ + _HEXAGON_V64_internal_union.b[7] = (new); \ + _HEXAGON_V64_internal_union.d; \ + }) + +#else /* !__hexagon__ */ + +#define HEXAGON_V64_PUT_B0(v, new) \ + (((v) & 0xffffffffffffff00LL) | ((HEXAGON_Vect64)((unsigned char)(new)))) +#define HEXAGON_V64_PUT_B1(v, new) \ + (((v) & 0xffffffffffff00ffLL) | (((HEXAGON_Vect64)((unsigned char)(new))) << 8LL)) +#define HEXAGON_V64_PUT_B2(v, new) \ + (((v) & 0xffffffffff00ffffLL) | (((HEXAGON_Vect64)((unsigned char)(new))) << 16LL)) +#define HEXAGON_V64_PUT_B3(v, new) \ + (((v) & 0xffffffff00ffffffLL) | (((HEXAGON_Vect64)((unsigned char)(new))) << 24LL)) +#define HEXAGON_V64_PUT_B4(v, new) \ + (((v) & 0xffffff00ffffffffLL) | (((HEXAGON_Vect64)((unsigned char)(new))) << 32LL)) +#define HEXAGON_V64_PUT_B5(v, new) \ + (((v) & 0xffff00ffffffffffLL) | (((HEXAGON_Vect64)((unsigned char)(new))) << 40LL)) +#define HEXAGON_V64_PUT_B6(v, new) \ + (((v) & 0xff00ffffffffffffLL) | (((HEXAGON_Vect64)((unsigned char)(new))) << 48LL)) +#define HEXAGON_V64_PUT_B7(v, new) \ + (((v) & 0x00ffffffffffffffLL) | (((HEXAGON_Vect64)(new)) << 56LL)) + +#endif /* !__hexagon__ */ + +/* NOTE: All create macros return a HEXAGON_Vect64 type */ + +/* Create from a doubleword */ + +#define HEXAGON_V64_CREATE_D(d) (d) + +/* Create from words */ + +#ifdef __hexagon__ + +#define HEXAGON_V64_CREATE_W(w1, w0) \ + __extension__({ \ + union { \ + long long d; \ + int w[2]; \ + } _HEXAGON_V64_internal_union; \ + _HEXAGON_V64_internal_union.w[0] = (w0); \ + _HEXAGON_V64_internal_union.w[1] = (w1); \ + _HEXAGON_V64_internal_union.d; \ + }) + +#else /* !__hexagon__ */ + +#define HEXAGON_V64_CREATE_W(w1, w0) \ + ((((HEXAGON_Vect64)(w1)) << 32LL) | ((HEXAGON_Vect64)((w0) & 0xffffffff))) + +#endif /* !__hexagon__ */ + +/* Create from half words */ + +#ifdef __hexagon__ + +#define HEXAGON_V64_CREATE_H(h3, h2, h1, h0) \ + __extension__({ \ + union { \ + long long d; \ + short h[4]; \ + } _HEXAGON_V64_internal_union; \ + _HEXAGON_V64_internal_union.h[0] = (h0); \ + _HEXAGON_V64_internal_union.h[1] = (h1); \ + _HEXAGON_V64_internal_union.h[2] = (h2); \ + _HEXAGON_V64_internal_union.h[3] = (h3); \ + _HEXAGON_V64_internal_union.d; \ + }) + +#else /* !__hexagon__ */ + +#define HEXAGON_V64_CREATE_H(h3, h2, h1, h0) \ + ((((HEXAGON_Vect64)(h3)) << 48LL) | (((HEXAGON_Vect64)((h2) & 0xffff)) << 32LL) | \ + (((HEXAGON_Vect64)((h1) & 0xffff)) << 16LL) | ((HEXAGON_Vect64)((h0) & 0xffff))) + +#endif /* !__hexagon__ */ + +/* Create from bytes */ + +#ifdef __hexagon__ + +#define HEXAGON_V64_CREATE_B(b7, b6, b5, b4, b3, b2, b1, b0) \ + __extension__({ \ + union { \ + long long d; \ + char b[8]; \ + } _HEXAGON_V64_internal_union; \ + _HEXAGON_V64_internal_union.b[0] = (b0); \ + _HEXAGON_V64_internal_union.b[1] = (b1); \ + _HEXAGON_V64_internal_union.b[2] = (b2); \ + _HEXAGON_V64_internal_union.b[3] = (b3); \ + _HEXAGON_V64_internal_union.b[4] = (b4); \ + _HEXAGON_V64_internal_union.b[5] = (b5); \ + _HEXAGON_V64_internal_union.b[6] = (b6); \ + _HEXAGON_V64_internal_union.b[7] = (b7); \ + _HEXAGON_V64_internal_union.d; \ + }) + +#else /* !__hexagon__ */ + +#define HEXAGON_V64_CREATE_B(b7, b6, b5, b4, b3, b2, b1, b0) \ + ((((HEXAGON_Vect64)(b7)) << 56LL) | (((HEXAGON_Vect64)((b6) & 0xff)) << 48LL) | \ + (((HEXAGON_Vect64)((b5) & 0xff)) << 40LL) | (((HEXAGON_Vect64)((b4) & 0xff)) << 32LL) | \ + (((HEXAGON_Vect64)((b3) & 0xff)) << 24LL) | (((HEXAGON_Vect64)((b2) & 0xff)) << 16LL) | \ + (((HEXAGON_Vect64)((b1) & 0xff)) << 8LL) | ((HEXAGON_Vect64)((b0) & 0xff))) + +#endif /* !__hexagon__ */ + +#ifdef __cplusplus + +class HEXAGON_Vect64C { +public: + // Constructors + HEXAGON_Vect64C(long long d = 0) : data(d) {}; + HEXAGON_Vect64C(int w1, int w0) : data(HEXAGON_V64_CREATE_W(w1, w0)) {}; + HEXAGON_Vect64C(short h3, short h2, short h1, short h0) + : data(HEXAGON_V64_CREATE_H(h3, h2, h1, h0)) {}; + HEXAGON_Vect64C(signed char b7, signed char b6, signed char b5, signed char b4, + signed char b3, signed char b2, signed char b1, signed char b0) + : data(HEXAGON_V64_CREATE_B(b7, b6, b5, b4, b3, b2, b1, b0)) {}; + HEXAGON_Vect64C(const HEXAGON_Vect64C &v) : data(v.data) {}; + + HEXAGON_Vect64C &operator=(const HEXAGON_Vect64C &v) { + data = v.data; + return *this; + }; + + operator long long() { + return data; + }; + + // Extract doubleword methods + long long D(void) { + return HEXAGON_V64_GET_D(data); + }; + unsigned long long UD(void) { + return HEXAGON_V64_GET_UD(data); + }; + + // Extract word methods + int W0(void) { + return HEXAGON_V64_GET_W0(data); + }; + int W1(void) { + return HEXAGON_V64_GET_W1(data); + }; + unsigned int UW0(void) { + return HEXAGON_V64_GET_UW0(data); + }; + unsigned int UW1(void) { + return HEXAGON_V64_GET_UW1(data); + }; + + // Extract half word methods + short H0(void) { + return HEXAGON_V64_GET_H0(data); + }; + short H1(void) { + return HEXAGON_V64_GET_H1(data); + }; + short H2(void) { + return HEXAGON_V64_GET_H2(data); + }; + short H3(void) { + return HEXAGON_V64_GET_H3(data); + }; + unsigned short UH0(void) { + return HEXAGON_V64_GET_UH0(data); + }; + unsigned short UH1(void) { + return HEXAGON_V64_GET_UH1(data); + }; + unsigned short UH2(void) { + return HEXAGON_V64_GET_UH2(data); + }; + unsigned short UH3(void) { + return HEXAGON_V64_GET_UH3(data); + }; + + // Extract byte methods + signed char B0(void) { + return HEXAGON_V64_GET_B0(data); + }; + signed char B1(void) { + return HEXAGON_V64_GET_B1(data); + }; + signed char B2(void) { + return HEXAGON_V64_GET_B2(data); + }; + signed char B3(void) { + return HEXAGON_V64_GET_B3(data); + }; + signed char B4(void) { + return HEXAGON_V64_GET_B4(data); + }; + signed char B5(void) { + return HEXAGON_V64_GET_B5(data); + }; + signed char B6(void) { + return HEXAGON_V64_GET_B6(data); + }; + signed char B7(void) { + return HEXAGON_V64_GET_B7(data); + }; + unsigned char UB0(void) { + return HEXAGON_V64_GET_UB0(data); + }; + unsigned char UB1(void) { + return HEXAGON_V64_GET_UB1(data); + }; + unsigned char UB2(void) { + return HEXAGON_V64_GET_UB2(data); + }; + unsigned char UB3(void) { + return HEXAGON_V64_GET_UB3(data); + }; + unsigned char UB4(void) { + return HEXAGON_V64_GET_UB4(data); + }; + unsigned char UB5(void) { + return HEXAGON_V64_GET_UB5(data); + }; + unsigned char UB6(void) { + return HEXAGON_V64_GET_UB6(data); + }; + unsigned char UB7(void) { + return HEXAGON_V64_GET_UB7(data); + }; + + // NOTE: All set methods return a HEXAGON_Vect64C type + + // Set doubleword method + HEXAGON_Vect64C D(long long d) { + return HEXAGON_Vect64C(HEXAGON_V64_PUT_D(data, d)); + }; + + // Set word methods + HEXAGON_Vect64C W0(int w) { + return HEXAGON_Vect64C(HEXAGON_V64_PUT_W0(data, w)); + }; + HEXAGON_Vect64C W1(int w) { + return HEXAGON_Vect64C(HEXAGON_V64_PUT_W1(data, w)); + }; + + // Set half word methods + HEXAGON_Vect64C H0(short h) { + return HEXAGON_Vect64C(HEXAGON_V64_PUT_H0(data, h)); + }; + HEXAGON_Vect64C H1(short h) { + return HEXAGON_Vect64C(HEXAGON_V64_PUT_H1(data, h)); + }; + HEXAGON_Vect64C H2(short h) { + return HEXAGON_Vect64C(HEXAGON_V64_PUT_H2(data, h)); + }; + HEXAGON_Vect64C H3(short h) { + return HEXAGON_Vect64C(HEXAGON_V64_PUT_H3(data, h)); + }; + + // Set byte methods + HEXAGON_Vect64C B0(signed char b) { + return HEXAGON_Vect64C(HEXAGON_V64_PUT_B0(data, b)); + }; + HEXAGON_Vect64C B1(signed char b) { + return HEXAGON_Vect64C(HEXAGON_V64_PUT_B1(data, b)); + }; + HEXAGON_Vect64C B2(signed char b) { + return HEXAGON_Vect64C(HEXAGON_V64_PUT_B2(data, b)); + }; + HEXAGON_Vect64C B3(signed char b) { + return HEXAGON_Vect64C(HEXAGON_V64_PUT_B3(data, b)); + }; + HEXAGON_Vect64C B4(signed char b) { + return HEXAGON_Vect64C(HEXAGON_V64_PUT_B4(data, b)); + }; + HEXAGON_Vect64C B5(signed char b) { + return HEXAGON_Vect64C(HEXAGON_V64_PUT_B5(data, b)); + }; + HEXAGON_Vect64C B6(signed char b) { + return HEXAGON_Vect64C(HEXAGON_V64_PUT_B6(data, b)); + }; + HEXAGON_Vect64C B7(signed char b) { + return HEXAGON_Vect64C(HEXAGON_V64_PUT_B7(data, b)); + }; + +private: + long long data; +}; + +#endif /* __cplusplus */ + +/* 32 Bit Vectors */ + +typedef int HEXAGON_Vect32; + +/* Extract word macros */ + +#define HEXAGON_V32_GET_W(v) (v) +#define HEXAGON_V32_GET_UW(v) ((unsigned int)(v)) + +/* Extract half word macros */ + +#define HEXAGON_V32_GET_H0(v) \ + __extension__({ \ + union { \ + int w; \ + short h[2]; \ + } _HEXAGON_V32_internal_union; \ + _HEXAGON_V32_internal_union.w = (v); \ + _HEXAGON_V32_internal_union.h[0]; \ + }) +#define HEXAGON_V32_GET_H1(v) \ + __extension__({ \ + union { \ + int w; \ + short h[2]; \ + } _HEXAGON_V32_internal_union; \ + _HEXAGON_V32_internal_union.w = (v); \ + _HEXAGON_V32_internal_union.h[1]; \ + }) +#define HEXAGON_V32_GET_UH0(v) \ + __extension__({ \ + union { \ + int w; \ + unsigned short uh[2]; \ + } _HEXAGON_V32_internal_union; \ + _HEXAGON_V32_internal_union.w = (v); \ + _HEXAGON_V32_internal_union.uh[0]; \ + }) +#define HEXAGON_V32_GET_UH1(v) \ + __extension__({ \ + union { \ + int w; \ + unsigned short uh[2]; \ + } _HEXAGON_V32_internal_union; \ + _HEXAGON_V32_internal_union.w = (v); \ + _HEXAGON_V32_internal_union.uh[1]; \ + }) + +/* Extract byte macros */ + +#define HEXAGON_V32_GET_B0(v) \ + __extension__({ \ + union { \ + int w; \ + signed char b[4]; \ + } _HEXAGON_V32_internal_union; \ + _HEXAGON_V32_internal_union.w = (v); \ + _HEXAGON_V32_internal_union.b[0]; \ + }) +#define HEXAGON_V32_GET_B1(v) \ + __extension__({ \ + union { \ + int w; \ + signed char b[4]; \ + } _HEXAGON_V32_internal_union; \ + _HEXAGON_V32_internal_union.w = (v); \ + _HEXAGON_V32_internal_union.b[1]; \ + }) +#define HEXAGON_V32_GET_B2(v) \ + __extension__({ \ + union { \ + int w; \ + signed char b[4]; \ + } _HEXAGON_V32_internal_union; \ + _HEXAGON_V32_internal_union.w = (v); \ + _HEXAGON_V32_internal_union.b[2]; \ + }) +#define HEXAGON_V32_GET_B3(v) \ + __extension__({ \ + union { \ + int w; \ + signed char b[4]; \ + } _HEXAGON_V32_internal_union; \ + _HEXAGON_V32_internal_union.w = (v); \ + _HEXAGON_V32_internal_union.b[3]; \ + }) +#define HEXAGON_V32_GET_UB0(v) \ + __extension__({ \ + union { \ + int w; \ + unsigned char ub[4]; \ + } _HEXAGON_V32_internal_union; \ + _HEXAGON_V32_internal_union.w = (v); \ + _HEXAGON_V32_internal_union.ub[0]; \ + }) +#define HEXAGON_V32_GET_UB1(v) \ + __extension__({ \ + union { \ + int w; \ + unsigned char ub[4]; \ + } _HEXAGON_V32_internal_union; \ + _HEXAGON_V32_internal_union.w = (v); \ + _HEXAGON_V32_internal_union.ub[1]; \ + }) +#define HEXAGON_V32_GET_UB2(v) \ + __extension__({ \ + union { \ + int w; \ + unsigned char ub[4]; \ + } _HEXAGON_V32_internal_union; \ + _HEXAGON_V32_internal_union.w = (v); \ + _HEXAGON_V32_internal_union.ub[2]; \ + }) +#define HEXAGON_V32_GET_UB3(v) \ + __extension__({ \ + union { \ + int w; \ + unsigned char ub[4]; \ + } _HEXAGON_V32_internal_union; \ + _HEXAGON_V32_internal_union.w = (v); \ + _HEXAGON_V32_internal_union.ub[3]; \ + }) + +/* NOTE: All set macros return a HEXAGON_Vect32 type */ + +/* Set word macro */ + +#define HEXAGON_V32_PUT_W(v, new) (new) + +/* Set half word macros */ + +#ifdef __hexagon__ + +#define HEXAGON_V32_PUT_H0(v, new) \ + __extension__({ \ + union { \ + int w; \ + short h[2]; \ + } _HEXAGON_V32_internal_union; \ + _HEXAGON_V32_internal_union.w = (v); \ + _HEXAGON_V32_internal_union.h[0] = (new); \ + _HEXAGON_V32_internal_union.w; \ + }) +#define HEXAGON_V32_PUT_H1(v, new) \ + __extension__({ \ + union { \ + int w; \ + short h[2]; \ + } _HEXAGON_V32_internal_union; \ + _HEXAGON_V32_internal_union.w = (v); \ + _HEXAGON_V32_internal_union.h[1] = (new); \ + _HEXAGON_V32_internal_union.w; \ + }) + +#else /* !__hexagon__ */ + +#define HEXAGON_V32_PUT_H0(v, new) \ + (((v) & 0xffff0000) | ((HEXAGON_Vect32)((unsigned short)(new)))) +#define HEXAGON_V32_PUT_H1(v, new) (((v) & 0x0000ffff) | (((HEXAGON_Vect32)(new)) << 16)) + +#endif /* !__hexagon__ */ + +/* Set byte macros */ + +#ifdef __hexagon__ + +#define HEXAGON_V32_PUT_B0(v, new) \ + __extension__({ \ + union { \ + int w; \ + char b[4]; \ + } _HEXAGON_V32_internal_union; \ + _HEXAGON_V32_internal_union.w = (v); \ + _HEXAGON_V32_internal_union.b[0] = (new); \ + _HEXAGON_V32_internal_union.w; \ + }) +#define HEXAGON_V32_PUT_B1(v, new) \ + __extension__({ \ + union { \ + int w; \ + char b[4]; \ + } _HEXAGON_V32_internal_union; \ + _HEXAGON_V32_internal_union.w = (v); \ + _HEXAGON_V32_internal_union.b[1] = (new); \ + _HEXAGON_V32_internal_union.w; \ + }) +#define HEXAGON_V32_PUT_B2(v, new) \ + __extension__({ \ + union { \ + int w; \ + char b[4]; \ + } _HEXAGON_V32_internal_union; \ + _HEXAGON_V32_internal_union.w = (v); \ + _HEXAGON_V32_internal_union.b[2] = (new); \ + _HEXAGON_V32_internal_union.w; \ + }) +#define HEXAGON_V32_PUT_B3(v, new) \ + __extension__({ \ + union { \ + int w; \ + char b[4]; \ + } _HEXAGON_V32_internal_union; \ + _HEXAGON_V32_internal_union.w = (v); \ + _HEXAGON_V32_internal_union.b[3] = (new); \ + _HEXAGON_V32_internal_union.w; \ + }) + +#else /* !__hexagon__ */ + +#define HEXAGON_V32_PUT_B0(v, new) \ + (((v) & 0xffffff00) | ((HEXAGON_Vect32)((unsigned char)(new)))) +#define HEXAGON_V32_PUT_B1(v, new) \ + (((v) & 0xffff00ff) | (((HEXAGON_Vect32)((unsigned char)(new))) << 8)) +#define HEXAGON_V32_PUT_B2(v, new) \ + (((v) & 0xff00ffff) | (((HEXAGON_Vect32)((unsigned char)(new))) << 16)) +#define HEXAGON_V32_PUT_B3(v, new) (((v) & 0x00ffffff) | (((HEXAGON_Vect32)(new)) << 24)) + +#endif /* !__hexagon__ */ + +/* NOTE: All create macros return a HEXAGON_Vect32 type */ + +/* Create from a word */ + +#define HEXAGON_V32_CREATE_W(w) (w) + +/* Create from half words */ + +#ifdef __hexagon__ + +#define HEXAGON_V32_CREATE_H(h1, h0) \ + __extension__({ \ + union { \ + long long d; \ + short h[2]; \ + } _HEXAGON_V32_internal_union; \ + _HEXAGON_V32_internal_union.h[0] = (h0); \ + _HEXAGON_V32_internal_union.h[1] = (h1); \ + _HEXAGON_V32_internal_union.d; \ + }) + +#else /* !__hexagon__ */ + +#define HEXAGON_V32_CREATE_H(h1, h0) \ + ((((HEXAGON_Vect32)(h1)) << 16) | ((HEXAGON_Vect32)((h0) & 0xffff))) + +#endif /* !__hexagon__ */ + +/* Create from bytes */ +#ifdef __hexagon__ + +#define HEXAGON_V32_CREATE_B(b3, b2, b1, b0) \ + __extension__({ \ + union { \ + long long d; \ + char b[4]; \ + } _HEXAGON_V32_internal_union; \ + _HEXAGON_V32_internal_union.b[0] = (b0); \ + _HEXAGON_V32_internal_union.b[1] = (b1); \ + _HEXAGON_V32_internal_union.b[2] = (b2); \ + _HEXAGON_V32_internal_union.b[3] = (b3); \ + _HEXAGON_V32_internal_union.d; \ + }) + +#else /* !__hexagon__ */ + +#define HEXAGON_V32_CREATE_B(b3, b2, b1, b0) \ + ((((HEXAGON_Vect32)(b3)) << 24) | (((HEXAGON_Vect32)((b2) & 0xff)) << 16) | \ + (((HEXAGON_Vect32)((b1) & 0xff)) << 8) | ((HEXAGON_Vect32)((b0) & 0xff))) + +#endif /* !__hexagon__ */ + +#ifdef __cplusplus + +class HEXAGON_Vect32C { +public: + // Constructors + HEXAGON_Vect32C(int w = 0) : data(w) {}; + HEXAGON_Vect32C(short h1, short h0) : data(HEXAGON_V32_CREATE_H(h1, h0)) {}; + HEXAGON_Vect32C(signed char b3, signed char b2, signed char b1, signed char b0) + : data(HEXAGON_V32_CREATE_B(b3, b2, b1, b0)) {}; + HEXAGON_Vect32C(const HEXAGON_Vect32C &v) : data(v.data) {}; + + HEXAGON_Vect32C &operator=(const HEXAGON_Vect32C &v) { + data = v.data; + return *this; + }; + + operator int() { + return data; + }; + + // Extract word methods + int W(void) { + return HEXAGON_V32_GET_W(data); + }; + unsigned int UW(void) { + return HEXAGON_V32_GET_UW(data); + }; + + // Extract half word methods + short H0(void) { + return HEXAGON_V32_GET_H0(data); + }; + short H1(void) { + return HEXAGON_V32_GET_H1(data); + }; + unsigned short UH0(void) { + return HEXAGON_V32_GET_UH0(data); + }; + unsigned short UH1(void) { + return HEXAGON_V32_GET_UH1(data); + }; + + // Extract byte methods + signed char B0(void) { + return HEXAGON_V32_GET_B0(data); + }; + signed char B1(void) { + return HEXAGON_V32_GET_B1(data); + }; + signed char B2(void) { + return HEXAGON_V32_GET_B2(data); + }; + signed char B3(void) { + return HEXAGON_V32_GET_B3(data); + }; + unsigned char UB0(void) { + return HEXAGON_V32_GET_UB0(data); + }; + unsigned char UB1(void) { + return HEXAGON_V32_GET_UB1(data); + }; + unsigned char UB2(void) { + return HEXAGON_V32_GET_UB2(data); + }; + unsigned char UB3(void) { + return HEXAGON_V32_GET_UB3(data); + }; + + // NOTE: All set methods return a HEXAGON_Vect32C type + + // Set word method + HEXAGON_Vect32C W(int w) { + return HEXAGON_Vect32C(HEXAGON_V32_PUT_W(data, w)); + }; + + // Set half word methods + HEXAGON_Vect32C H0(short h) { + return HEXAGON_Vect32C(HEXAGON_V32_PUT_H0(data, h)); + }; + HEXAGON_Vect32C H1(short h) { + return HEXAGON_Vect32C(HEXAGON_V32_PUT_H1(data, h)); + }; + + // Set byte methods + HEXAGON_Vect32C B0(signed char b) { + return HEXAGON_Vect32C(HEXAGON_V32_PUT_B0(data, b)); + }; + HEXAGON_Vect32C B1(signed char b) { + return HEXAGON_Vect32C(HEXAGON_V32_PUT_B1(data, b)); + }; + HEXAGON_Vect32C B2(signed char b) { + return HEXAGON_Vect32C(HEXAGON_V32_PUT_B2(data, b)); + }; + HEXAGON_Vect32C B3(signed char b) { + return HEXAGON_Vect32C(HEXAGON_V32_PUT_B3(data, b)); + }; + +private: + int data; +}; + +#endif /* __cplusplus */ + +// V65 Silver types +#if __Q6S_ARCH__ >= 65 + // Silver vector types are 128 bytes, and pairs are 256. The vector predicate + // types are 16 bytes and 32 bytes for pairs. + typedef long HEXAGON_VecPred128 __attribute__((__vector_size__(16))) + __attribute__((aligned(128))); + + typedef long HEXAGON_VecPred256 __attribute__((__vector_size__(32))) + __attribute__((aligned(128))); + + typedef long HEXAGON_Vect1024 __attribute__((__vector_size__(128))) + __attribute__((aligned(128))); + + typedef long HEXAGON_Vect2048 __attribute__((__vector_size__(256))) + __attribute__((aligned(256))); + + typedef long HEXAGON_UVect1024 __attribute__((__vector_size__(128))) + __attribute__((aligned(4))); + + typedef long HEXAGON_UVect2048 __attribute__((__vector_size__(256))) + __attribute__((aligned(4))); + + #define Q6S_VectorPredPair HEXAGON_VecPred256 + #define Q6S_VectorPred HEXAGON_VecPred128 + #define Q6S_Vector HEXAGON_Vect1024 + #define Q6S_VectorPair HEXAGON_Vect2048 + #define Q6S_UVector HEXAGON_UVect1024 + #define Q6S_UVectorPair HEXAGON_UVect2048 + +#else /* __Q6S_ARCH__ >= 65 */ + +// V65 Vector types +#if __HVX_ARCH__ >= 65 +#if defined __HVX__ && (__HVX_LENGTH__ == 128) + typedef long HEXAGON_VecPred128 __attribute__((__vector_size__(128))) + __attribute__((aligned(128))); + + typedef long HEXAGON_Vect1024 __attribute__((__vector_size__(128))) + __attribute__((aligned(128))); + + typedef long HEXAGON_Vect2048 __attribute__((__vector_size__(256))) + __attribute__((aligned(256))); + + typedef long HEXAGON_UVect1024 __attribute__((__vector_size__(128))) + __attribute__((aligned(4))); + + typedef long HEXAGON_UVect2048 __attribute__((__vector_size__(256))) + __attribute__((aligned(4))); + + #define HVX_VectorPred HEXAGON_VecPred128 + #define HVX_Vector HEXAGON_Vect1024 + #define HVX_VectorPair HEXAGON_Vect2048 + #define HVX_UVector HEXAGON_UVect1024 + #define HVX_UVectorPair HEXAGON_UVect2048 +#else /* defined __HVX__ && (__HVX_LENGTH__ == 128) */ +#if defined __HVX__ && (__HVX_LENGTH__ == 64) + typedef long HEXAGON_VecPred64 __attribute__((__vector_size__(64))) + __attribute__((aligned(64))); + + typedef long HEXAGON_Vect512 __attribute__((__vector_size__(64))) + __attribute__((aligned(64))); + + typedef long HEXAGON_Vect1024 __attribute__((__vector_size__(128))) + __attribute__((aligned(128))); + + typedef long HEXAGON_UVect512 __attribute__((__vector_size__(64))) + __attribute__((aligned(4))); + + typedef long HEXAGON_UVect1024 __attribute__((__vector_size__(128))) + __attribute__((aligned(4))); + + #define HVX_VectorPred HEXAGON_VecPred64 + #define HVX_Vector HEXAGON_Vect512 + #define HVX_VectorPair HEXAGON_Vect1024 + #define HVX_UVector HEXAGON_UVect512 + #define HVX_UVectorPair HEXAGON_UVect1024 +#endif /* defined __HVX__ && (__HVX_LENGTH__ == 64) */ +#endif /* defined __HVX__ && (__HVX_LENGTH__ == 128) */ +#endif /* __HVX_ARCH__ >= 65 */ +#endif /* __Q6S_ARCH__ >= 65 */ + +/* Predicates */ + +typedef int HEXAGON_Pred; + +/*** + *** backward compatibility aliases + ***/ + +/* Old names */ +#define Q6Vect Q6Vect64 +#define Q6V_GET_D Q6V64_GET_D +#define Q6V_GET_UD Q6V64_GET_UD +#define Q6V_GET_W0 Q6V64_GET_W0 +#define Q6V_GET_W1 Q6V64_GET_W1 +#define Q6V_GET_UW0 Q6V64_GET_UW0 +#define Q6V_GET_UW1 Q6V64_GET_UW1 +#define Q6V_GET_H0 Q6V64_GET_H0 +#define Q6V_GET_H1 Q6V64_GET_H1 +#define Q6V_GET_H2 Q6V64_GET_H2 +#define Q6V_GET_H3 Q6V64_GET_H3 +#define Q6V_GET_UH0 Q6V64_GET_UH0 +#define Q6V_GET_UH1 Q6V64_GET_UH1 +#define Q6V_GET_UH2 Q6V64_GET_UH2 +#define Q6V_GET_UH3 Q6V64_GET_UH3 +#define Q6V_GET_B0 Q6V64_GET_B0 +#define Q6V_GET_B1 Q6V64_GET_B1 +#define Q6V_GET_B2 Q6V64_GET_B2 +#define Q6V_GET_B3 Q6V64_GET_B3 +#define Q6V_GET_B4 Q6V64_GET_B4 +#define Q6V_GET_B5 Q6V64_GET_B5 +#define Q6V_GET_B6 Q6V64_GET_B6 +#define Q6V_GET_B7 Q6V64_GET_B7 +#define Q6V_GET_UB0 Q6V64_GET_UB0 +#define Q6V_GET_UB1 Q6V64_GET_UB1 +#define Q6V_GET_UB2 Q6V64_GET_UB2 +#define Q6V_GET_UB3 Q6V64_GET_UB3 +#define Q6V_GET_UB4 Q6V64_GET_UB4 +#define Q6V_GET_UB5 Q6V64_GET_UB5 +#define Q6V_GET_UB6 Q6V64_GET_UB6 +#define Q6V_GET_UB7 Q6V64_GET_UB7 +#define Q6V_PUT_D Q6V64_PUT_D +#define Q6V_PUT_W0 Q6V64_PUT_W0 +#define Q6V_PUT_W1 Q6V64_PUT_W1 +#define Q6V_PUT_H0 Q6V64_PUT_H0 +#define Q6V_PUT_H1 Q6V64_PUT_H1 +#define Q6V_PUT_H2 Q6V64_PUT_H2 +#define Q6V_PUT_H3 Q6V64_PUT_H3 +#define Q6V_PUT_B0 Q6V64_PUT_B0 +#define Q6V_PUT_B1 Q6V64_PUT_B1 +#define Q6V_PUT_B2 Q6V64_PUT_B2 +#define Q6V_PUT_B3 Q6V64_PUT_B3 +#define Q6V_PUT_B4 Q6V64_PUT_B4 +#define Q6V_PUT_B5 Q6V64_PUT_B5 +#define Q6V_PUT_B6 Q6V64_PUT_B6 +#define Q6V_PUT_B7 Q6V64_PUT_B7 +#define Q6V_CREATE_D Q6V64_CREATE_D +#define Q6V_CREATE_W Q6V64_CREATE_W +#define Q6V_CREATE_H Q6V64_CREATE_H +#define Q6V_CREATE_B Q6V64_CREATE_B + +#ifdef __cplusplus +#define Q6VectC Q6Vect64C +#endif /* __cplusplus */ + +/* 64 Bit Vectors */ + +typedef long long __attribute__((__may_alias__)) Q6Vect64; + +/* Extract doubleword macros */ + +#define Q6V64_GET_D(v) (v) +#define Q6V64_GET_UD(v) ((unsigned long long)(v)) + +/* Extract word macros */ + +#define Q6V64_GET_W0(v) \ + __extension__({ \ + union { \ + long long d; \ + int w[2]; \ + } _Q6V64_internal_union; \ + _Q6V64_internal_union.d = (v); \ + _Q6V64_internal_union.w[0]; \ + }) +#define Q6V64_GET_W1(v) \ + __extension__({ \ + union { \ + long long d; \ + int w[2]; \ + } _Q6V64_internal_union; \ + _Q6V64_internal_union.d = (v); \ + _Q6V64_internal_union.w[1]; \ + }) +#define Q6V64_GET_UW0(v) \ + __extension__({ \ + union { \ + long long d; \ + unsigned int uw[2]; \ + } _Q6V64_internal_union; \ + _Q6V64_internal_union.d = (v); \ + _Q6V64_internal_union.uw[0]; \ + }) +#define Q6V64_GET_UW1(v) \ + __extension__({ \ + union { \ + long long d; \ + unsigned int uw[2]; \ + } _Q6V64_internal_union; \ + _Q6V64_internal_union.d = (v); \ + _Q6V64_internal_union.uw[1]; \ + }) + +/* Extract half word macros */ + +#define Q6V64_GET_H0(v) \ + __extension__({ \ + union { \ + long long d; \ + short h[4]; \ + } _Q6V64_internal_union; \ + _Q6V64_internal_union.d = (v); \ + _Q6V64_internal_union.h[0]; \ + }) +#define Q6V64_GET_H1(v) \ + __extension__({ \ + union { \ + long long d; \ + short h[4]; \ + } _Q6V64_internal_union; \ + _Q6V64_internal_union.d = (v); \ + _Q6V64_internal_union.h[1]; \ + }) +#define Q6V64_GET_H2(v) \ + __extension__({ \ + union { \ + long long d; \ + short h[4]; \ + } _Q6V64_internal_union; \ + _Q6V64_internal_union.d = (v); \ + _Q6V64_internal_union.h[2]; \ + }) +#define Q6V64_GET_H3(v) \ + __extension__({ \ + union { \ + long long d; \ + short h[4]; \ + } _Q6V64_internal_union; \ + _Q6V64_internal_union.d = (v); \ + _Q6V64_internal_union.h[3]; \ + }) +#define Q6V64_GET_UH0(v) \ + __extension__({ \ + union { \ + long long d; \ + unsigned short uh[4]; \ + } _Q6V64_internal_union; \ + _Q6V64_internal_union.d = (v); \ + _Q6V64_internal_union.uh[0]; \ + }) +#define Q6V64_GET_UH1(v) \ + __extension__({ \ + union { \ + long long d; \ + unsigned short uh[4]; \ + } _Q6V64_internal_union; \ + _Q6V64_internal_union.d = (v); \ + _Q6V64_internal_union.uh[1]; \ + }) +#define Q6V64_GET_UH2(v) \ + __extension__({ \ + union { \ + long long d; \ + unsigned short uh[4]; \ + } _Q6V64_internal_union; \ + _Q6V64_internal_union.d = (v); \ + _Q6V64_internal_union.uh[2]; \ + }) +#define Q6V64_GET_UH3(v) \ + __extension__({ \ + union { \ + long long d; \ + unsigned short uh[4]; \ + } _Q6V64_internal_union; \ + _Q6V64_internal_union.d = (v); \ + _Q6V64_internal_union.uh[3]; \ + }) + +/* Extract byte macros */ + +#define Q6V64_GET_B0(v) \ + __extension__({ \ + union { \ + long long d; \ + signed char b[8]; \ + } _Q6V64_internal_union; \ + _Q6V64_internal_union.d = (v); \ + _Q6V64_internal_union.b[0]; \ + }) +#define Q6V64_GET_B1(v) \ + __extension__({ \ + union { \ + long long d; \ + signed char b[8]; \ + } _Q6V64_internal_union; \ + _Q6V64_internal_union.d = (v); \ + _Q6V64_internal_union.b[1]; \ + }) +#define Q6V64_GET_B2(v) \ + __extension__({ \ + union { \ + long long d; \ + signed char b[8]; \ + } _Q6V64_internal_union; \ + _Q6V64_internal_union.d = (v); \ + _Q6V64_internal_union.b[2]; \ + }) +#define Q6V64_GET_B3(v) \ + __extension__({ \ + union { \ + long long d; \ + signed char b[8]; \ + } _Q6V64_internal_union; \ + _Q6V64_internal_union.d = (v); \ + _Q6V64_internal_union.b[3]; \ + }) +#define Q6V64_GET_B4(v) \ + __extension__({ \ + union { \ + long long d; \ + signed char b[8]; \ + } _Q6V64_internal_union; \ + _Q6V64_internal_union.d = (v); \ + _Q6V64_internal_union.b[4]; \ + }) +#define Q6V64_GET_B5(v) \ + __extension__({ \ + union { \ + long long d; \ + signed char b[8]; \ + } _Q6V64_internal_union; \ + _Q6V64_internal_union.d = (v); \ + _Q6V64_internal_union.b[5]; \ + }) +#define Q6V64_GET_B6(v) \ + __extension__({ \ + union { \ + long long d; \ + signed char b[8]; \ + } _Q6V64_internal_union; \ + _Q6V64_internal_union.d = (v); \ + _Q6V64_internal_union.b[6]; \ + }) +#define Q6V64_GET_B7(v) \ + __extension__({ \ + union { \ + long long d; \ + signed char b[8]; \ + } _Q6V64_internal_union; \ + _Q6V64_internal_union.d = (v); \ + _Q6V64_internal_union.b[7]; \ + }) +#define Q6V64_GET_UB0(v) \ + __extension__({ \ + union { \ + long long d; \ + unsigned char ub[8]; \ + } _Q6V64_internal_union; \ + _Q6V64_internal_union.d = (v); \ + _Q6V64_internal_union.ub[0]; \ + }) +#define Q6V64_GET_UB1(v) \ + __extension__({ \ + union { \ + long long d; \ + unsigned char ub[8]; \ + } _Q6V64_internal_union; \ + _Q6V64_internal_union.d = (v); \ + _Q6V64_internal_union.ub[1]; \ + }) +#define Q6V64_GET_UB2(v) \ + __extension__({ \ + union { \ + long long d; \ + unsigned char ub[8]; \ + } _Q6V64_internal_union; \ + _Q6V64_internal_union.d = (v); \ + _Q6V64_internal_union.ub[2]; \ + }) +#define Q6V64_GET_UB3(v) \ + __extension__({ \ + union { \ + long long d; \ + unsigned char ub[8]; \ + } _Q6V64_internal_union; \ + _Q6V64_internal_union.d = (v); \ + _Q6V64_internal_union.ub[3]; \ + }) +#define Q6V64_GET_UB4(v) \ + __extension__({ \ + union { \ + long long d; \ + unsigned char ub[8]; \ + } _Q6V64_internal_union; \ + _Q6V64_internal_union.d = (v); \ + _Q6V64_internal_union.ub[4]; \ + }) +#define Q6V64_GET_UB5(v) \ + __extension__({ \ + union { \ + long long d; \ + unsigned char ub[8]; \ + } _Q6V64_internal_union; \ + _Q6V64_internal_union.d = (v); \ + _Q6V64_internal_union.ub[5]; \ + }) +#define Q6V64_GET_UB6(v) \ + __extension__({ \ + union { \ + long long d; \ + unsigned char ub[8]; \ + } _Q6V64_internal_union; \ + _Q6V64_internal_union.d = (v); \ + _Q6V64_internal_union.ub[6]; \ + }) +#define Q6V64_GET_UB7(v) \ + __extension__({ \ + union { \ + long long d; \ + unsigned char ub[8]; \ + } _Q6V64_internal_union; \ + _Q6V64_internal_union.d = (v); \ + _Q6V64_internal_union.ub[7]; \ + }) + +/* NOTE: All set macros return a Q6Vect64 type */ + +/* Set doubleword macro */ + +#define Q6V64_PUT_D(v, new) (new) + +/* Set word macros */ + +#ifdef __qdsp6__ + +#define Q6V64_PUT_W0(v, new) \ + __extension__({ \ + union { \ + long long d; \ + int w[2]; \ + } _Q6V64_internal_union; \ + _Q6V64_internal_union.d = (v); \ + _Q6V64_internal_union.w[0] = (new); \ + _Q6V64_internal_union.d; \ + }) +#define Q6V64_PUT_W1(v, new) \ + __extension__({ \ + union { \ + long long d; \ + int w[2]; \ + } _Q6V64_internal_union; \ + _Q6V64_internal_union.d = (v); \ + _Q6V64_internal_union.w[1] = (new); \ + _Q6V64_internal_union.d; \ + }) + +#else /* !__qdsp6__ */ + +#define Q6V64_PUT_W0(v, new) \ + (((v) & 0xffffffff00000000LL) | ((Q6Vect64)((unsigned int)(new)))) +#define Q6V64_PUT_W1(v, new) \ + (((v) & 0x00000000ffffffffLL) | (((Q6Vect64)(new)) << 32LL)) + +#endif /* !__qdsp6__ */ + +/* Set half word macros */ + +#ifdef __qdsp6__ + +#define Q6V64_PUT_H0(v, new) \ + __extension__({ \ + union { \ + long long d; \ + short h[4]; \ + } _Q6V64_internal_union; \ + _Q6V64_internal_union.d = (v); \ + _Q6V64_internal_union.h[0] = (new); \ + _Q6V64_internal_union.d; \ + }) +#define Q6V64_PUT_H1(v, new) \ + __extension__({ \ + union { \ + long long d; \ + short h[4]; \ + } _Q6V64_internal_union; \ + _Q6V64_internal_union.d = (v); \ + _Q6V64_internal_union.h[1] = (new); \ + _Q6V64_internal_union.d; \ + }) +#define Q6V64_PUT_H2(v, new) \ + __extension__({ \ + union { \ + long long d; \ + short h[4]; \ + } _Q6V64_internal_union; \ + _Q6V64_internal_union.d = (v); \ + _Q6V64_internal_union.h[2] = (new); \ + _Q6V64_internal_union.d; \ + }) +#define Q6V64_PUT_H3(v, new) \ + __extension__({ \ + union { \ + long long d; \ + short h[4]; \ + } _Q6V64_internal_union; \ + _Q6V64_internal_union.d = (v); \ + _Q6V64_internal_union.h[3] = (new); \ + _Q6V64_internal_union.d; \ + }) + +#else /* !__qdsp6__ */ + +#define Q6V64_PUT_H0(v, new) \ + (((v) & 0xffffffffffff0000LL) | ((Q6Vect64)((unsigned short)(new)))) +#define Q6V64_PUT_H1(v, new) \ + (((v) & 0xffffffff0000ffffLL) | (((Q6Vect64)((unsigned short)(new))) << 16LL)) +#define Q6V64_PUT_H2(v, new) \ + (((v) & 0xffff0000ffffffffLL) | (((Q6Vect64)((unsigned short)(new))) << 32LL)) +#define Q6V64_PUT_H3(v, new) \ + (((v) & 0x0000ffffffffffffLL) | (((Q6Vect64)(new)) << 48LL)) + +#endif /* !__qdsp6__ */ + +/* Set byte macros */ + +#ifdef __qdsp6__ + +#define Q6V64_PUT_B0(v, new) \ + __extension__({ \ + union { \ + long long d; \ + char b[8]; \ + } _Q6V64_internal_union; \ + _Q6V64_internal_union.d = (v); \ + _Q6V64_internal_union.b[0] = (new); \ + _Q6V64_internal_union.d; \ + }) +#define Q6V64_PUT_B1(v, new) \ + __extension__({ \ + union { \ + long long d; \ + char b[8]; \ + } _Q6V64_internal_union; \ + _Q6V64_internal_union.d = (v); \ + _Q6V64_internal_union.b[1] = (new); \ + _Q6V64_internal_union.d; \ + }) +#define Q6V64_PUT_B2(v, new) \ + __extension__({ \ + union { \ + long long d; \ + char b[8]; \ + } _Q6V64_internal_union; \ + _Q6V64_internal_union.d = (v); \ + _Q6V64_internal_union.b[2] = (new); \ + _Q6V64_internal_union.d; \ + }) +#define Q6V64_PUT_B3(v, new) \ + __extension__({ \ + union { \ + long long d; \ + char b[8]; \ + } _Q6V64_internal_union; \ + _Q6V64_internal_union.d = (v); \ + _Q6V64_internal_union.b[3] = (new); \ + _Q6V64_internal_union.d; \ + }) +#define Q6V64_PUT_B4(v, new) \ + __extension__({ \ + union { \ + long long d; \ + char b[8]; \ + } _Q6V64_internal_union; \ + _Q6V64_internal_union.d = (v); \ + _Q6V64_internal_union.b[4] = (new); \ + _Q6V64_internal_union.d; \ + }) +#define Q6V64_PUT_B5(v, new) \ + __extension__({ \ + union { \ + long long d; \ + char b[8]; \ + } _Q6V64_internal_union; \ + _Q6V64_internal_union.d = (v); \ + _Q6V64_internal_union.b[5] = (new); \ + _Q6V64_internal_union.d; \ + }) +#define Q6V64_PUT_B6(v, new) \ + __extension__({ \ + union { \ + long long d; \ + char b[8]; \ + } _Q6V64_internal_union; \ + _Q6V64_internal_union.d = (v); \ + _Q6V64_internal_union.b[6] = (new); \ + _Q6V64_internal_union.d; \ + }) +#define Q6V64_PUT_B7(v, new) \ + __extension__({ \ + union { \ + long long d; \ + char b[8]; \ + } _Q6V64_internal_union; \ + _Q6V64_internal_union.d = (v); \ + _Q6V64_internal_union.b[7] = (new); \ + _Q6V64_internal_union.d; \ + }) + +#else /* !__qdsp6__ */ + +#define Q6V64_PUT_B0(v, new) \ + (((v) & 0xffffffffffffff00LL) | ((Q6Vect64)((unsigned char)(new)))) +#define Q6V64_PUT_B1(v, new) \ + (((v) & 0xffffffffffff00ffLL) | (((Q6Vect64)((unsigned char)(new))) << 8LL)) +#define Q6V64_PUT_B2(v, new) \ + (((v) & 0xffffffffff00ffffLL) | (((Q6Vect64)((unsigned char)(new))) << 16LL)) +#define Q6V64_PUT_B3(v, new) \ + (((v) & 0xffffffff00ffffffLL) | (((Q6Vect64)((unsigned char)(new))) << 24LL)) +#define Q6V64_PUT_B4(v, new) \ + (((v) & 0xffffff00ffffffffLL) | (((Q6Vect64)((unsigned char)(new))) << 32LL)) +#define Q6V64_PUT_B5(v, new) \ + (((v) & 0xffff00ffffffffffLL) | (((Q6Vect64)((unsigned char)(new))) << 40LL)) +#define Q6V64_PUT_B6(v, new) \ + (((v) & 0xff00ffffffffffffLL) | (((Q6Vect64)((unsigned char)(new))) << 48LL)) +#define Q6V64_PUT_B7(v, new) \ + (((v) & 0x00ffffffffffffffLL) | (((Q6Vect64)(new)) << 56LL)) + +#endif /* !__qdsp6__ */ + +/* NOTE: All create macros return a Q6Vect64 type */ + +/* Create from a doubleword */ + +#define Q6V64_CREATE_D(d) (d) + +/* Create from words */ + +#ifdef __qdsp6__ + +#define Q6V64_CREATE_W(w1, w0) \ + __extension__({ \ + union { \ + long long d; \ + int w[2]; \ + } _Q6V64_internal_union; \ + _Q6V64_internal_union.w[0] = (w0); \ + _Q6V64_internal_union.w[1] = (w1); \ + _Q6V64_internal_union.d; \ + }) + +#else /* !__qdsp6__ */ + +#define Q6V64_CREATE_W(w1, w0) \ + ((((Q6Vect64)(w1)) << 32LL) | ((Q6Vect64)((w0) & 0xffffffff))) + +#endif /* !__qdsp6__ */ + +/* Create from half words */ + +#ifdef __qdsp6__ + +#define Q6V64_CREATE_H(h3, h2, h1, h0) \ + __extension__({ \ + union { \ + long long d; \ + short h[4]; \ + } _Q6V64_internal_union; \ + _Q6V64_internal_union.h[0] = (h0); \ + _Q6V64_internal_union.h[1] = (h1); \ + _Q6V64_internal_union.h[2] = (h2); \ + _Q6V64_internal_union.h[3] = (h3); \ + _Q6V64_internal_union.d; \ + }) + +#else /* !__qdsp6__ */ + +#define Q6V64_CREATE_H(h3, h2, h1, h0) \ + ((((Q6Vect64)(h3)) << 48LL) | (((Q6Vect64)((h2) & 0xffff)) << 32LL) | \ + (((Q6Vect64)((h1) & 0xffff)) << 16LL) | ((Q6Vect64)((h0) & 0xffff))) + +#endif /* !__qdsp6__ */ + +/* Create from bytes */ + +#ifdef __qdsp6__ + +#define Q6V64_CREATE_B(b7, b6, b5, b4, b3, b2, b1, b0) \ + __extension__({ \ + union { \ + long long d; \ + char b[8]; \ + } _Q6V64_internal_union; \ + _Q6V64_internal_union.b[0] = (b0); \ + _Q6V64_internal_union.b[1] = (b1); \ + _Q6V64_internal_union.b[2] = (b2); \ + _Q6V64_internal_union.b[3] = (b3); \ + _Q6V64_internal_union.b[4] = (b4); \ + _Q6V64_internal_union.b[5] = (b5); \ + _Q6V64_internal_union.b[6] = (b6); \ + _Q6V64_internal_union.b[7] = (b7); \ + _Q6V64_internal_union.d; \ + }) + +#else /* !__qdsp6__ */ + +#define Q6V64_CREATE_B(b7, b6, b5, b4, b3, b2, b1, b0) \ + ((((Q6Vect64)(b7)) << 56LL) | (((Q6Vect64)((b6) & 0xff)) << 48LL) | \ + (((Q6Vect64)((b5) & 0xff)) << 40LL) | (((Q6Vect64)((b4) & 0xff)) << 32LL) | \ + (((Q6Vect64)((b3) & 0xff)) << 24LL) | (((Q6Vect64)((b2) & 0xff)) << 16LL) | \ + (((Q6Vect64)((b1) & 0xff)) << 8LL) | ((Q6Vect64)((b0) & 0xff))) + +#endif /* !__qdsp6__ */ + +#ifdef __cplusplus + +class Q6Vect64C { +public: + // Constructors + Q6Vect64C(long long d = 0) : data(d) {}; + Q6Vect64C(int w1, int w0) : data(Q6V64_CREATE_W(w1, w0)) {}; + Q6Vect64C(short h3, short h2, short h1, short h0) + : data(Q6V64_CREATE_H(h3, h2, h1, h0)) {}; + Q6Vect64C(signed char b7, signed char b6, signed char b5, signed char b4, + signed char b3, signed char b2, signed char b1, signed char b0) + : data(Q6V64_CREATE_B(b7, b6, b5, b4, b3, b2, b1, b0)) {}; + Q6Vect64C(const Q6Vect64C &v) : data(v.data) {}; + + Q6Vect64C &operator=(const Q6Vect64C &v) { + data = v.data; + return *this; + }; + + operator long long() { + return data; + }; + + // Extract doubleword methods + long long D(void) { + return Q6V64_GET_D(data); + }; + unsigned long long UD(void) { + return Q6V64_GET_UD(data); + }; + + // Extract word methods + int W0(void) { + return Q6V64_GET_W0(data); + }; + int W1(void) { + return Q6V64_GET_W1(data); + }; + unsigned int UW0(void) { + return Q6V64_GET_UW0(data); + }; + unsigned int UW1(void) { + return Q6V64_GET_UW1(data); + }; + + // Extract half word methods + short H0(void) { + return Q6V64_GET_H0(data); + }; + short H1(void) { + return Q6V64_GET_H1(data); + }; + short H2(void) { + return Q6V64_GET_H2(data); + }; + short H3(void) { + return Q6V64_GET_H3(data); + }; + unsigned short UH0(void) { + return Q6V64_GET_UH0(data); + }; + unsigned short UH1(void) { + return Q6V64_GET_UH1(data); + }; + unsigned short UH2(void) { + return Q6V64_GET_UH2(data); + }; + unsigned short UH3(void) { + return Q6V64_GET_UH3(data); + }; + + // Extract byte methods + signed char B0(void) { + return Q6V64_GET_B0(data); + }; + signed char B1(void) { + return Q6V64_GET_B1(data); + }; + signed char B2(void) { + return Q6V64_GET_B2(data); + }; + signed char B3(void) { + return Q6V64_GET_B3(data); + }; + signed char B4(void) { + return Q6V64_GET_B4(data); + }; + signed char B5(void) { + return Q6V64_GET_B5(data); + }; + signed char B6(void) { + return Q6V64_GET_B6(data); + }; + signed char B7(void) { + return Q6V64_GET_B7(data); + }; + unsigned char UB0(void) { + return Q6V64_GET_UB0(data); + }; + unsigned char UB1(void) { + return Q6V64_GET_UB1(data); + }; + unsigned char UB2(void) { + return Q6V64_GET_UB2(data); + }; + unsigned char UB3(void) { + return Q6V64_GET_UB3(data); + }; + unsigned char UB4(void) { + return Q6V64_GET_UB4(data); + }; + unsigned char UB5(void) { + return Q6V64_GET_UB5(data); + }; + unsigned char UB6(void) { + return Q6V64_GET_UB6(data); + }; + unsigned char UB7(void) { + return Q6V64_GET_UB7(data); + }; + + // NOTE: All set methods return a Q6Vect64C type + + // Set doubleword method + Q6Vect64C D(long long d) { + return Q6Vect64C(Q6V64_PUT_D(data, d)); + }; + + // Set word methods + Q6Vect64C W0(int w) { + return Q6Vect64C(Q6V64_PUT_W0(data, w)); + }; + Q6Vect64C W1(int w) { + return Q6Vect64C(Q6V64_PUT_W1(data, w)); + }; + + // Set half word methods + Q6Vect64C H0(short h) { + return Q6Vect64C(Q6V64_PUT_H0(data, h)); + }; + Q6Vect64C H1(short h) { + return Q6Vect64C(Q6V64_PUT_H1(data, h)); + }; + Q6Vect64C H2(short h) { + return Q6Vect64C(Q6V64_PUT_H2(data, h)); + }; + Q6Vect64C H3(short h) { + return Q6Vect64C(Q6V64_PUT_H3(data, h)); + }; + + // Set byte methods + Q6Vect64C B0(signed char b) { + return Q6Vect64C(Q6V64_PUT_B0(data, b)); + }; + Q6Vect64C B1(signed char b) { + return Q6Vect64C(Q6V64_PUT_B1(data, b)); + }; + Q6Vect64C B2(signed char b) { + return Q6Vect64C(Q6V64_PUT_B2(data, b)); + }; + Q6Vect64C B3(signed char b) { + return Q6Vect64C(Q6V64_PUT_B3(data, b)); + }; + Q6Vect64C B4(signed char b) { + return Q6Vect64C(Q6V64_PUT_B4(data, b)); + }; + Q6Vect64C B5(signed char b) { + return Q6Vect64C(Q6V64_PUT_B5(data, b)); + }; + Q6Vect64C B6(signed char b) { + return Q6Vect64C(Q6V64_PUT_B6(data, b)); + }; + Q6Vect64C B7(signed char b) { + return Q6Vect64C(Q6V64_PUT_B7(data, b)); + }; + +private: + long long data; +}; + +#endif /* __cplusplus */ + +/* 32 Bit Vectors */ + +typedef int Q6Vect32; + +/* Extract word macros */ + +#define Q6V32_GET_W(v) (v) +#define Q6V32_GET_UW(v) ((unsigned int)(v)) + +/* Extract half word macros */ + +#define Q6V32_GET_H0(v) \ + __extension__({ \ + union { \ + int w; \ + short h[2]; \ + } _Q6V32_internal_union; \ + _Q6V32_internal_union.w = (v); \ + _Q6V32_internal_union.h[0]; \ + }) +#define Q6V32_GET_H1(v) \ + __extension__({ \ + union { \ + int w; \ + short h[2]; \ + } _Q6V32_internal_union; \ + _Q6V32_internal_union.w = (v); \ + _Q6V32_internal_union.h[1]; \ + }) +#define Q6V32_GET_UH0(v) \ + __extension__({ \ + union { \ + int w; \ + unsigned short uh[2]; \ + } _Q6V32_internal_union; \ + _Q6V32_internal_union.w = (v); \ + _Q6V32_internal_union.uh[0]; \ + }) +#define Q6V32_GET_UH1(v) \ + __extension__({ \ + union { \ + int w; \ + unsigned short uh[2]; \ + } _Q6V32_internal_union; \ + _Q6V32_internal_union.w = (v); \ + _Q6V32_internal_union.uh[1]; \ + }) + +/* Extract byte macros */ + +#define Q6V32_GET_B0(v) \ + __extension__({ \ + union { \ + int w; \ + signed char b[4]; \ + } _Q6V32_internal_union; \ + _Q6V32_internal_union.w = (v); \ + _Q6V32_internal_union.b[0]; \ + }) +#define Q6V32_GET_B1(v) \ + __extension__({ \ + union { \ + int w; \ + signed char b[4]; \ + } _Q6V32_internal_union; \ + _Q6V32_internal_union.w = (v); \ + _Q6V32_internal_union.b[1]; \ + }) +#define Q6V32_GET_B2(v) \ + __extension__({ \ + union { \ + int w; \ + signed char b[4]; \ + } _Q6V32_internal_union; \ + _Q6V32_internal_union.w = (v); \ + _Q6V32_internal_union.b[2]; \ + }) +#define Q6V32_GET_B3(v) \ + __extension__({ \ + union { \ + int w; \ + signed char b[4]; \ + } _Q6V32_internal_union; \ + _Q6V32_internal_union.w = (v); \ + _Q6V32_internal_union.b[3]; \ + }) +#define Q6V32_GET_UB0(v) \ + __extension__({ \ + union { \ + int w; \ + unsigned char ub[4]; \ + } _Q6V32_internal_union; \ + _Q6V32_internal_union.w = (v); \ + _Q6V32_internal_union.ub[0]; \ + }) +#define Q6V32_GET_UB1(v) \ + __extension__({ \ + union { \ + int w; \ + unsigned char ub[4]; \ + } _Q6V32_internal_union; \ + _Q6V32_internal_union.w = (v); \ + _Q6V32_internal_union.ub[1]; \ + }) +#define Q6V32_GET_UB2(v) \ + __extension__({ \ + union { \ + int w; \ + unsigned char ub[4]; \ + } _Q6V32_internal_union; \ + _Q6V32_internal_union.w = (v); \ + _Q6V32_internal_union.ub[2]; \ + }) +#define Q6V32_GET_UB3(v) \ + __extension__({ \ + union { \ + int w; \ + unsigned char ub[4]; \ + } _Q6V32_internal_union; \ + _Q6V32_internal_union.w = (v); \ + _Q6V32_internal_union.ub[3]; \ + }) + +/* NOTE: All set macros return a Q6Vect32 type */ + +/* Set word macro */ + +#define Q6V32_PUT_W(v, new) (new) + +/* Set half word macros */ + +#ifdef __qdsp6__ + +#define Q6V32_PUT_H0(v, new) \ + __extension__({ \ + union { \ + int w; \ + short h[2]; \ + } _Q6V32_internal_union; \ + _Q6V32_internal_union.w = (v); \ + _Q6V32_internal_union.h[0] = (new); \ + _Q6V32_internal_union.w; \ + }) +#define Q6V32_PUT_H1(v, new) \ + __extension__({ \ + union { \ + int w; \ + short h[2]; \ + } _Q6V32_internal_union; \ + _Q6V32_internal_union.w = (v); \ + _Q6V32_internal_union.h[1] = (new); \ + _Q6V32_internal_union.w; \ + }) + +#else /* !__qdsp6__ */ + +#define Q6V32_PUT_H0(v, new) \ + (((v) & 0xffff0000) | ((Q6Vect32)((unsigned short)(new)))) +#define Q6V32_PUT_H1(v, new) (((v) & 0x0000ffff) | (((Q6Vect32)(new)) << 16)) + +#endif /* !__qdsp6__ */ + +/* Set byte macros */ + +#ifdef __qdsp6__ + +#define Q6V32_PUT_B0(v, new) \ + __extension__({ \ + union { \ + int w; \ + char b[4]; \ + } _Q6V32_internal_union; \ + _Q6V32_internal_union.w = (v); \ + _Q6V32_internal_union.b[0] = (new); \ + _Q6V32_internal_union.w; \ + }) +#define Q6V32_PUT_B1(v, new) \ + __extension__({ \ + union { \ + int w; \ + char b[4]; \ + } _Q6V32_internal_union; \ + _Q6V32_internal_union.w = (v); \ + _Q6V32_internal_union.b[1] = (new); \ + _Q6V32_internal_union.w; \ + }) +#define Q6V32_PUT_B2(v, new) \ + __extension__({ \ + union { \ + int w; \ + char b[4]; \ + } _Q6V32_internal_union; \ + _Q6V32_internal_union.w = (v); \ + _Q6V32_internal_union.b[2] = (new); \ + _Q6V32_internal_union.w; \ + }) +#define Q6V32_PUT_B3(v, new) \ + __extension__({ \ + union { \ + int w; \ + char b[4]; \ + } _Q6V32_internal_union; \ + _Q6V32_internal_union.w = (v); \ + _Q6V32_internal_union.b[3] = (new); \ + _Q6V32_internal_union.w; \ + }) + +#else /* !__qdsp6__ */ + +#define Q6V32_PUT_B0(v, new) \ + (((v) & 0xffffff00) | ((Q6Vect32)((unsigned char)(new)))) +#define Q6V32_PUT_B1(v, new) \ + (((v) & 0xffff00ff) | (((Q6Vect32)((unsigned char)(new))) << 8)) +#define Q6V32_PUT_B2(v, new) \ + (((v) & 0xff00ffff) | (((Q6Vect32)((unsigned char)(new))) << 16)) +#define Q6V32_PUT_B3(v, new) (((v) & 0x00ffffff) | (((Q6Vect32)(new)) << 24)) + +#endif /* !__qdsp6__ */ + +/* NOTE: All create macros return a Q6Vect32 type */ + +/* Create from a word */ + +#define Q6V32_CREATE_W(w) (w) + +/* Create from half words */ + +#ifdef __qdsp6__ + +#define Q6V32_CREATE_H(h1, h0) \ + __extension__({ \ + union { \ + long long d; \ + short h[2]; \ + } _Q6V32_internal_union; \ + _Q6V32_internal_union.h[0] = (h0); \ + _Q6V32_internal_union.h[1] = (h1); \ + _Q6V32_internal_union.d; \ + }) + +#else /* !__qdsp6__ */ + +#define Q6V32_CREATE_H(h1, h0) \ + ((((Q6Vect32)(h1)) << 16) | ((Q6Vect32)((h0) & 0xffff))) + +#endif /* !__qdsp6__ */ + +/* Create from bytes */ +#ifdef __qdsp6__ + +#define Q6V32_CREATE_B(b3, b2, b1, b0) \ + __extension__({ \ + union { \ + long long d; \ + char b[4]; \ + } _Q6V32_internal_union; \ + _Q6V32_internal_union.b[0] = (b0); \ + _Q6V32_internal_union.b[1] = (b1); \ + _Q6V32_internal_union.b[2] = (b2); \ + _Q6V32_internal_union.b[3] = (b3); \ + _Q6V32_internal_union.d; \ + }) + +#else /* !__qdsp6__ */ + +#define Q6V32_CREATE_B(b3, b2, b1, b0) \ + ((((Q6Vect32)(b3)) << 24) | (((Q6Vect32)((b2) & 0xff)) << 16) | \ + (((Q6Vect32)((b1) & 0xff)) << 8) | ((Q6Vect32)((b0) & 0xff))) + +#endif /* !__qdsp6__ */ + +#ifdef __cplusplus + +class Q6Vect32C { +public: + // Constructors + Q6Vect32C(int w = 0) : data(w) {}; + Q6Vect32C(short h1, short h0) : data(Q6V32_CREATE_H(h1, h0)) {}; + Q6Vect32C(signed char b3, signed char b2, signed char b1, signed char b0) + : data(Q6V32_CREATE_B(b3, b2, b1, b0)) {}; + Q6Vect32C(const Q6Vect32C &v) : data(v.data) {}; + + Q6Vect32C &operator=(const Q6Vect32C &v) { + data = v.data; + return *this; + }; + + operator int() { + return data; + }; + + // Extract word methods + int W(void) { + return Q6V32_GET_W(data); + }; + unsigned int UW(void) { + return Q6V32_GET_UW(data); + }; + + // Extract half word methods + short H0(void) { + return Q6V32_GET_H0(data); + }; + short H1(void) { + return Q6V32_GET_H1(data); + }; + unsigned short UH0(void) { + return Q6V32_GET_UH0(data); + }; + unsigned short UH1(void) { + return Q6V32_GET_UH1(data); + }; + + // Extract byte methods + signed char B0(void) { + return Q6V32_GET_B0(data); + }; + signed char B1(void) { + return Q6V32_GET_B1(data); + }; + signed char B2(void) { + return Q6V32_GET_B2(data); + }; + signed char B3(void) { + return Q6V32_GET_B3(data); + }; + unsigned char UB0(void) { + return Q6V32_GET_UB0(data); + }; + unsigned char UB1(void) { + return Q6V32_GET_UB1(data); + }; + unsigned char UB2(void) { + return Q6V32_GET_UB2(data); + }; + unsigned char UB3(void) { + return Q6V32_GET_UB3(data); + }; + + // NOTE: All set methods return a Q6Vect32C type + + // Set word method + Q6Vect32C W(int w) { + return Q6Vect32C(Q6V32_PUT_W(data, w)); + }; + + // Set half word methods + Q6Vect32C H0(short h) { + return Q6Vect32C(Q6V32_PUT_H0(data, h)); + }; + Q6Vect32C H1(short h) { + return Q6Vect32C(Q6V32_PUT_H1(data, h)); + }; + + // Set byte methods + Q6Vect32C B0(signed char b) { + return Q6Vect32C(Q6V32_PUT_B0(data, b)); + }; + Q6Vect32C B1(signed char b) { + return Q6Vect32C(Q6V32_PUT_B1(data, b)); + }; + Q6Vect32C B2(signed char b) { + return Q6Vect32C(Q6V32_PUT_B2(data, b)); + }; + Q6Vect32C B3(signed char b) { + return Q6Vect32C(Q6V32_PUT_B3(data, b)); + }; + +private: + int data; +}; + +#endif /* __cplusplus */ + +// V65 Vector types +#if __HVX_ARCH__ >= 65 +#if defined __HVX__ && (__HVX_LENGTH__ == 128) +typedef long Q6VecPred128 __attribute__((__vector_size__(128))) + __attribute__((aligned(128))); + +typedef long Q6Vect1024 __attribute__((__vector_size__(128))) + __attribute__((aligned(128))); + +typedef long Q6Vect2048 __attribute__((__vector_size__(256))) + __attribute__((aligned(256))); + +#else /* defined __HVX__ && (__HVX_LENGTH__ == 128) */ +#if defined __HVX__ && (__HVX_LENGTH__ == 64) +typedef long Q6VecPred64 __attribute__((__vector_size__(64))) + __attribute__((aligned(64))); + +typedef long Q6Vect512 __attribute__((__vector_size__(64))) + __attribute__((aligned(64))); + +typedef long Q6Vect1024 __attribute__((__vector_size__(128))) + __attribute__((aligned(128))); + +#endif /* defined __HVX__ && (__HVX_LENGTH__ == 64) */ +#endif /* defined __HVX__ && (__HVX_LENGTH__ == 128) */ +#endif /* __HVX_ARCH__ >= 65 */ + +/* Predicates */ + +typedef int Q6Pred; + + +#ifdef __HVX__ + +// Extract HVX VectorPair macro. +#define HEXAGON_HVX_GET_W(v) (v) + +// Extract HVX Vector macros. +#define HEXAGON_HVX_GET_V0(v) \ + __extension__({ \ + union { \ + HVX_VectorPair W; \ + HVX_Vector V[2]; \ + } _HEXAGON_HVX_internal_union; \ + _HEXAGON_HVX_internal_union.W = (v); \ + _HEXAGON_HVX_internal_union.V[0]; \ + }) +#define HEXAGON_HVX_GET_V1(v) \ + __extension__({ \ + union { \ + HVX_VectorPair W; \ + HVX_Vector V[2]; \ + } _HEXAGON_HVX_internal_union; \ + _HEXAGON_HVX_internal_union.W = (v); \ + _HEXAGON_HVX_internal_union.V[1]; \ + }) +#define HEXAGON_HVX_GET_P(v) \ + __extension__({ \ + union { \ + HVX_VectorPair W; \ + HVX_VectorPred P[2]; \ + } _HEXAGON_HVX_internal_union; \ + _HEXAGON_HVX_internal_union.W = (v); \ + _HEXAGON_HVX_internal_union.P[0]; \ + }) + +// Set HVX VectorPair macro. +#define HEXAGON_HVX_PUT_W(v, new) (new) + +// Set HVX Vector macros. +#define HEXAGON_HVX_PUT_V0(v, new) \ + __extension__({ \ + union { \ + HVX_VectorPair W; \ + HVX_Vector V[2]; \ + } _HEXAGON_HVX_internal_union; \ + _HEXAGON_HVX_internal_union.W = (v); \ + _HEXAGON_HVX_internal_union.V[0] = (new); \ + _HEXAGON_HVX_internal_union.W; \ + }) + +#define HEXAGON_HVX_PUT_V1(v, new) \ + __extension__({ \ + union { \ + HVX_VectorPair W; \ + HVX_Vector V[2]; \ + } _HEXAGON_HVX_internal_union; \ + _HEXAGON_HVX_internal_union.W = (v); \ + _HEXAGON_HVX_internal_union.V[1] = (new); \ + _HEXAGON_HVX_internal_union.W; \ + }) + +#define HEXAGON_HVX_PUT_P(v, new) \ + __extension__({ \ + union { \ + HVX_VectorPair W; \ + HVX_VectorPred P[2]; \ + } _HEXAGON_HVX_internal_union; \ + _HEXAGON_HVX_internal_union.W = (v); \ + _HEXAGON_HVX_internal_union.P[0] = (new); \ + _HEXAGON_HVX_internal_union.W; \ + }) + + +#define HEXAGON_HVX_CREATE_W(v1, v0) \ + __extension__({ \ + union { \ + HVX_VectorPair W; \ + HVX_Vector V[2]; \ + } _HEXAGON_HVX_internal_union; \ + _HEXAGON_HVX_internal_union.V[0] = (v0); \ + _HEXAGON_HVX_internal_union.V[1] = (v1); \ + _HEXAGON_HVX_internal_union.W; \ + }) + +#ifdef __cplusplus + +class HVX_Vect { +public: + // Constructors. + // Default. + HVX_Vect() : data(Q6_W_vcombine_VV(Q6_V_vzero(), Q6_V_vzero())){}; + + // Custom constructors. + HVX_Vect(HVX_VectorPair W) : data(W){}; + HVX_Vect(HVX_Vector v1, HVX_Vector v0) : data(HEXAGON_HVX_CREATE_W(v1, v0)){}; + + // Copy constructor. + HVX_Vect(const HVX_Vect &W) = default; + + // Move constructor. + HVX_Vect(HVX_Vect &&W) = default; + + // Assignment operator. + HVX_Vect &operator=(const HVX_Vect &W) = default; + + operator HVX_VectorPair() { return data; }; + + // Extract VectorPair method. + HVX_VectorPair W(void) { return HEXAGON_HVX_GET_W(data); }; + + // Extract Vector methods. + HVX_Vector V0(void) { return HEXAGON_HVX_GET_V0(data); }; + HVX_Vector V1(void) { return HEXAGON_HVX_GET_V1(data); }; + HVX_VectorPred P(void) { return HEXAGON_HVX_GET_P(data); }; + + // NOTE: All set methods return a HVX_Vect type. + // Set HVX VectorPair method. + HVX_Vect W(HVX_VectorPair w) { return HVX_Vect(HEXAGON_HVX_PUT_W(data, w)); }; + + // Set HVX Vector methods. + HVX_Vect V0(HVX_Vector v) { return HVX_Vect(HEXAGON_HVX_PUT_V0(data, v)); }; + HVX_Vect V1(HVX_Vector v) { return HVX_Vect(HEXAGON_HVX_PUT_V1(data, v)); }; + HVX_Vect P(HVX_VectorPred p) { return HVX_Vect(HEXAGON_HVX_PUT_P(data, p)); }; + +private: + HVX_VectorPair data; +}; + +#endif /* __cplusplus */ +#endif /* __HVX__ */ + +#define HEXAGON_UDMA_DM0_STATUS_IDLE 0x00000000 +#define HEXAGON_UDMA_DM0_STATUS_RUN 0x00000001 +#define HEXAGON_UDMA_DM0_STATUS_ERROR 0x00000002 +#define HEXAGON_UDMA_DESC_DSTATE_INCOMPLETE 0 +#define HEXAGON_UDMA_DESC_DSTATE_COMPLETE 1 +#define HEXAGON_UDMA_DESC_ORDER_NOORDER 0 +#define HEXAGON_UDMA_DESC_ORDER_ORDER 1 +#define HEXAGON_UDMA_DESC_BYPASS_OFF 0 +#define HEXAGON_UDMA_DESC_BYPASS_ON 1 +#define HEXAGON_UDMA_DESC_COMP_NONE 0 +#define HEXAGON_UDMA_DESC_COMP_DLBC 1 +#define HEXAGON_UDMA_DESC_DESCTYPE_TYPE0 0 +#define HEXAGON_UDMA_DESC_DESCTYPE_TYPE1 1 + +typedef struct hexagon_udma_descriptor_type0_s +{ + void *next; + unsigned int length:24; + unsigned int desctype:2; + unsigned int dstcomp:1; + unsigned int srccomp:1; + unsigned int dstbypass:1; + unsigned int srcbypass:1; + unsigned int order:1; + unsigned int dstate:1; + void *src; + void *dst; +} hexagon_udma_descriptor_type0_t; + +typedef struct hexagon_udma_descriptor_type1_s +{ + void *next; + unsigned int length:24; + unsigned int desctype:2; + unsigned int dstcomp:1; + unsigned int srccomp:1; + unsigned int dstbypass:1; + unsigned int srcbypass:1; + unsigned int order:1; + unsigned int dstate:1; + void *src; + void *dst; + unsigned int allocation:28; + unsigned int padding:4; + unsigned int roiwidth:16; + unsigned int roiheight:16; + unsigned int srcstride:16; + unsigned int dststride:16; + unsigned int srcwidthoffset:16; + unsigned int dstwidthoffset:16; +} hexagon_udma_descriptor_type1_t; + +#endif /* !HEXAGON_TYPES_H */ diff --git a/lib/include/hvx_hexagon_protos.h b/lib/include/hvx_hexagon_protos.h new file mode 100644 index 000000000..41ce7a6b9 --- /dev/null +++ b/lib/include/hvx_hexagon_protos.h @@ -0,0 +1,4392 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// +// Automatically generated file, do not edit! +//===----------------------------------------------------------------------===// + + + +#ifndef _HVX_HEXAGON_PROTOS_H_ +#define _HVX_HEXAGON_PROTOS_H_ 1 + +#ifdef __HVX__ +#if __HVX_LENGTH__ == 128 +#define __BUILTIN_VECTOR_WRAP(a) a ## _128B +#else +#define __BUILTIN_VECTOR_WRAP(a) a +#endif + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Rd32=vextract(Vu32,Rs32) + C Intrinsic Prototype: Word32 Q6_R_vextract_VR(HVX_Vector Vu, Word32 Rs) + Instruction Type: LD + Execution Slots: SLOT0 + ========================================================================== */ + +#define Q6_R_vextract_VR __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_extractw) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32=hi(Vss32) + C Intrinsic Prototype: HVX_Vector Q6_V_hi_W(HVX_VectorPair Vss) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_V_hi_W __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_hi) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32=lo(Vss32) + C Intrinsic Prototype: HVX_Vector Q6_V_lo_W(HVX_VectorPair Vss) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_V_lo_W __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_lo) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32=vsplat(Rt32) + C Intrinsic Prototype: HVX_Vector Q6_V_vsplat_R(Word32 Rt) + Instruction Type: CVI_VX_LATE + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_V_vsplat_R __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_lvsplatw) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Qd4=and(Qs4,Qt4) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_and_QQ(HVX_VectorPred Qs, HVX_VectorPred Qt) + Instruction Type: CVI_VA_DV + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Q_and_QQ __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_pred_and) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Qd4=and(Qs4,!Qt4) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_and_QQn(HVX_VectorPred Qs, HVX_VectorPred Qt) + Instruction Type: CVI_VA_DV + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Q_and_QQn __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_pred_and_n) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Qd4=not(Qs4) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_not_Q(HVX_VectorPred Qs) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Q_not_Q __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_pred_not) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Qd4=or(Qs4,Qt4) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_or_QQ(HVX_VectorPred Qs, HVX_VectorPred Qt) + Instruction Type: CVI_VA_DV + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Q_or_QQ __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_pred_or) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Qd4=or(Qs4,!Qt4) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_or_QQn(HVX_VectorPred Qs, HVX_VectorPred Qt) + Instruction Type: CVI_VA_DV + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Q_or_QQn __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_pred_or_n) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Qd4=vsetq(Rt32) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_vsetq_R(Word32 Rt) + Instruction Type: CVI_VP + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Q_vsetq_R __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_pred_scalar2) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Qd4=xor(Qs4,Qt4) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_xor_QQ(HVX_VectorPred Qs, HVX_VectorPred Qt) + Instruction Type: CVI_VA_DV + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Q_xor_QQ __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_pred_xor) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: if (!Qv4) vmem(Rt32+#s4)=Vs32 + C Intrinsic Prototype: void Q6_vmem_QnRIV(HVX_VectorPred Qv, HVX_Vector* Rt, HVX_Vector Vs) + Instruction Type: CVI_VM_ST + Execution Slots: SLOT0 + ========================================================================== */ + +#define Q6_vmem_QnRIV __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vS32b_nqpred_ai) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: if (!Qv4) vmem(Rt32+#s4):nt=Vs32 + C Intrinsic Prototype: void Q6_vmem_QnRIV_nt(HVX_VectorPred Qv, HVX_Vector* Rt, HVX_Vector Vs) + Instruction Type: CVI_VM_ST + Execution Slots: SLOT0 + ========================================================================== */ + +#define Q6_vmem_QnRIV_nt __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vS32b_nt_nqpred_ai) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: if (Qv4) vmem(Rt32+#s4):nt=Vs32 + C Intrinsic Prototype: void Q6_vmem_QRIV_nt(HVX_VectorPred Qv, HVX_Vector* Rt, HVX_Vector Vs) + Instruction Type: CVI_VM_ST + Execution Slots: SLOT0 + ========================================================================== */ + +#define Q6_vmem_QRIV_nt __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vS32b_nt_qpred_ai) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: if (Qv4) vmem(Rt32+#s4)=Vs32 + C Intrinsic Prototype: void Q6_vmem_QRIV(HVX_VectorPred Qv, HVX_Vector* Rt, HVX_Vector Vs) + Instruction Type: CVI_VM_ST + Execution Slots: SLOT0 + ========================================================================== */ + +#define Q6_vmem_QRIV __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vS32b_qpred_ai) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.uh=vabsdiff(Vu32.h,Vv32.h) + C Intrinsic Prototype: HVX_Vector Q6_Vuh_vabsdiff_VhVh(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vuh_vabsdiff_VhVh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vabsdiffh) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.ub=vabsdiff(Vu32.ub,Vv32.ub) + C Intrinsic Prototype: HVX_Vector Q6_Vub_vabsdiff_VubVub(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vub_vabsdiff_VubVub __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vabsdiffub) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.uh=vabsdiff(Vu32.uh,Vv32.uh) + C Intrinsic Prototype: HVX_Vector Q6_Vuh_vabsdiff_VuhVuh(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vuh_vabsdiff_VuhVuh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vabsdiffuh) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.uw=vabsdiff(Vu32.w,Vv32.w) + C Intrinsic Prototype: HVX_Vector Q6_Vuw_vabsdiff_VwVw(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vuw_vabsdiff_VwVw __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vabsdiffw) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.h=vabs(Vu32.h) + C Intrinsic Prototype: HVX_Vector Q6_Vh_vabs_Vh(HVX_Vector Vu) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vh_vabs_Vh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vabsh) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.h=vabs(Vu32.h):sat + C Intrinsic Prototype: HVX_Vector Q6_Vh_vabs_Vh_sat(HVX_Vector Vu) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vh_vabs_Vh_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vabsh_sat) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.w=vabs(Vu32.w) + C Intrinsic Prototype: HVX_Vector Q6_Vw_vabs_Vw(HVX_Vector Vu) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vw_vabs_Vw __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vabsw) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.w=vabs(Vu32.w):sat + C Intrinsic Prototype: HVX_Vector Q6_Vw_vabs_Vw_sat(HVX_Vector Vu) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vw_vabs_Vw_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vabsw_sat) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.b=vadd(Vu32.b,Vv32.b) + C Intrinsic Prototype: HVX_Vector Q6_Vb_vadd_VbVb(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vb_vadd_VbVb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vaddb) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.b=vadd(Vuu32.b,Vvv32.b) + C Intrinsic Prototype: HVX_VectorPair Q6_Wb_vadd_WbWb(HVX_VectorPair Vuu, HVX_VectorPair Vvv) + Instruction Type: CVI_VA_DV + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Wb_vadd_WbWb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vaddb_dv) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: if (!Qv4) Vx32.b+=Vu32.b + C Intrinsic Prototype: HVX_Vector Q6_Vb_condacc_QnVbVb(HVX_VectorPred Qv, HVX_Vector Vx, HVX_Vector Vu) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vb_condacc_QnVbVb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vaddbnq) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: if (Qv4) Vx32.b+=Vu32.b + C Intrinsic Prototype: HVX_Vector Q6_Vb_condacc_QVbVb(HVX_VectorPred Qv, HVX_Vector Vx, HVX_Vector Vu) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vb_condacc_QVbVb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vaddbq) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.h=vadd(Vu32.h,Vv32.h) + C Intrinsic Prototype: HVX_Vector Q6_Vh_vadd_VhVh(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vh_vadd_VhVh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vaddh) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.h=vadd(Vuu32.h,Vvv32.h) + C Intrinsic Prototype: HVX_VectorPair Q6_Wh_vadd_WhWh(HVX_VectorPair Vuu, HVX_VectorPair Vvv) + Instruction Type: CVI_VA_DV + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Wh_vadd_WhWh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vaddh_dv) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: if (!Qv4) Vx32.h+=Vu32.h + C Intrinsic Prototype: HVX_Vector Q6_Vh_condacc_QnVhVh(HVX_VectorPred Qv, HVX_Vector Vx, HVX_Vector Vu) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vh_condacc_QnVhVh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vaddhnq) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: if (Qv4) Vx32.h+=Vu32.h + C Intrinsic Prototype: HVX_Vector Q6_Vh_condacc_QVhVh(HVX_VectorPred Qv, HVX_Vector Vx, HVX_Vector Vu) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vh_condacc_QVhVh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vaddhq) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.h=vadd(Vu32.h,Vv32.h):sat + C Intrinsic Prototype: HVX_Vector Q6_Vh_vadd_VhVh_sat(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vh_vadd_VhVh_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vaddhsat) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.h=vadd(Vuu32.h,Vvv32.h):sat + C Intrinsic Prototype: HVX_VectorPair Q6_Wh_vadd_WhWh_sat(HVX_VectorPair Vuu, HVX_VectorPair Vvv) + Instruction Type: CVI_VA_DV + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Wh_vadd_WhWh_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vaddhsat_dv) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.w=vadd(Vu32.h,Vv32.h) + C Intrinsic Prototype: HVX_VectorPair Q6_Ww_vadd_VhVh(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Ww_vadd_VhVh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vaddhw) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.h=vadd(Vu32.ub,Vv32.ub) + C Intrinsic Prototype: HVX_VectorPair Q6_Wh_vadd_VubVub(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Wh_vadd_VubVub __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vaddubh) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.ub=vadd(Vu32.ub,Vv32.ub):sat + C Intrinsic Prototype: HVX_Vector Q6_Vub_vadd_VubVub_sat(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vub_vadd_VubVub_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vaddubsat) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.ub=vadd(Vuu32.ub,Vvv32.ub):sat + C Intrinsic Prototype: HVX_VectorPair Q6_Wub_vadd_WubWub_sat(HVX_VectorPair Vuu, HVX_VectorPair Vvv) + Instruction Type: CVI_VA_DV + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Wub_vadd_WubWub_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vaddubsat_dv) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.uh=vadd(Vu32.uh,Vv32.uh):sat + C Intrinsic Prototype: HVX_Vector Q6_Vuh_vadd_VuhVuh_sat(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vuh_vadd_VuhVuh_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vadduhsat) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.uh=vadd(Vuu32.uh,Vvv32.uh):sat + C Intrinsic Prototype: HVX_VectorPair Q6_Wuh_vadd_WuhWuh_sat(HVX_VectorPair Vuu, HVX_VectorPair Vvv) + Instruction Type: CVI_VA_DV + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Wuh_vadd_WuhWuh_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vadduhsat_dv) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.w=vadd(Vu32.uh,Vv32.uh) + C Intrinsic Prototype: HVX_VectorPair Q6_Ww_vadd_VuhVuh(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Ww_vadd_VuhVuh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vadduhw) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.w=vadd(Vu32.w,Vv32.w) + C Intrinsic Prototype: HVX_Vector Q6_Vw_vadd_VwVw(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vw_vadd_VwVw __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vaddw) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.w=vadd(Vuu32.w,Vvv32.w) + C Intrinsic Prototype: HVX_VectorPair Q6_Ww_vadd_WwWw(HVX_VectorPair Vuu, HVX_VectorPair Vvv) + Instruction Type: CVI_VA_DV + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Ww_vadd_WwWw __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vaddw_dv) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: if (!Qv4) Vx32.w+=Vu32.w + C Intrinsic Prototype: HVX_Vector Q6_Vw_condacc_QnVwVw(HVX_VectorPred Qv, HVX_Vector Vx, HVX_Vector Vu) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vw_condacc_QnVwVw __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vaddwnq) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: if (Qv4) Vx32.w+=Vu32.w + C Intrinsic Prototype: HVX_Vector Q6_Vw_condacc_QVwVw(HVX_VectorPred Qv, HVX_Vector Vx, HVX_Vector Vu) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vw_condacc_QVwVw __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vaddwq) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.w=vadd(Vu32.w,Vv32.w):sat + C Intrinsic Prototype: HVX_Vector Q6_Vw_vadd_VwVw_sat(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vw_vadd_VwVw_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vaddwsat) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.w=vadd(Vuu32.w,Vvv32.w):sat + C Intrinsic Prototype: HVX_VectorPair Q6_Ww_vadd_WwWw_sat(HVX_VectorPair Vuu, HVX_VectorPair Vvv) + Instruction Type: CVI_VA_DV + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Ww_vadd_WwWw_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vaddwsat_dv) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32=valign(Vu32,Vv32,Rt8) + C Intrinsic Prototype: HVX_Vector Q6_V_valign_VVR(HVX_Vector Vu, HVX_Vector Vv, Word32 Rt) + Instruction Type: CVI_VP + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_V_valign_VVR __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_valignb) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32=valign(Vu32,Vv32,#u3) + C Intrinsic Prototype: HVX_Vector Q6_V_valign_VVI(HVX_Vector Vu, HVX_Vector Vv, Word32 Iu3) + Instruction Type: CVI_VP + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_V_valign_VVI __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_valignbi) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32=vand(Vu32,Vv32) + C Intrinsic Prototype: HVX_Vector Q6_V_vand_VV(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_V_vand_VV __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vand) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32=vand(Qu4,Rt32) + C Intrinsic Prototype: HVX_Vector Q6_V_vand_QR(HVX_VectorPred Qu, Word32 Rt) + Instruction Type: CVI_VX_LATE + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_V_vand_QR __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vandqrt) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vx32|=vand(Qu4,Rt32) + C Intrinsic Prototype: HVX_Vector Q6_V_vandor_VQR(HVX_Vector Vx, HVX_VectorPred Qu, Word32 Rt) + Instruction Type: CVI_VX_LATE + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_V_vandor_VQR __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vandqrt_acc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Qd4=vand(Vu32,Rt32) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_vand_VR(HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VX_LATE + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Q_vand_VR __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vandvrt) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Qx4|=vand(Vu32,Rt32) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_vandor_QVR(HVX_VectorPred Qx, HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VX_LATE + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Q_vandor_QVR __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vandvrt_acc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.h=vasl(Vu32.h,Rt32) + C Intrinsic Prototype: HVX_Vector Q6_Vh_vasl_VhR(HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vh_vasl_VhR __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vaslh) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.h=vasl(Vu32.h,Vv32.h) + C Intrinsic Prototype: HVX_Vector Q6_Vh_vasl_VhVh(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vh_vasl_VhVh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vaslhv) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.w=vasl(Vu32.w,Rt32) + C Intrinsic Prototype: HVX_Vector Q6_Vw_vasl_VwR(HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vw_vasl_VwR __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vaslw) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vx32.w+=vasl(Vu32.w,Rt32) + C Intrinsic Prototype: HVX_Vector Q6_Vw_vaslacc_VwVwR(HVX_Vector Vx, HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vw_vaslacc_VwVwR __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vaslw_acc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.w=vasl(Vu32.w,Vv32.w) + C Intrinsic Prototype: HVX_Vector Q6_Vw_vasl_VwVw(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vw_vasl_VwVw __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vaslwv) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.h=vasr(Vu32.h,Rt32) + C Intrinsic Prototype: HVX_Vector Q6_Vh_vasr_VhR(HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vh_vasr_VhR __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vasrh) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.b=vasr(Vu32.h,Vv32.h,Rt8):rnd:sat + C Intrinsic Prototype: HVX_Vector Q6_Vb_vasr_VhVhR_rnd_sat(HVX_Vector Vu, HVX_Vector Vv, Word32 Rt) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vb_vasr_VhVhR_rnd_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vasrhbrndsat) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.ub=vasr(Vu32.h,Vv32.h,Rt8):rnd:sat + C Intrinsic Prototype: HVX_Vector Q6_Vub_vasr_VhVhR_rnd_sat(HVX_Vector Vu, HVX_Vector Vv, Word32 Rt) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vub_vasr_VhVhR_rnd_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vasrhubrndsat) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.ub=vasr(Vu32.h,Vv32.h,Rt8):sat + C Intrinsic Prototype: HVX_Vector Q6_Vub_vasr_VhVhR_sat(HVX_Vector Vu, HVX_Vector Vv, Word32 Rt) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vub_vasr_VhVhR_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vasrhubsat) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.h=vasr(Vu32.h,Vv32.h) + C Intrinsic Prototype: HVX_Vector Q6_Vh_vasr_VhVh(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vh_vasr_VhVh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vasrhv) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.w=vasr(Vu32.w,Rt32) + C Intrinsic Prototype: HVX_Vector Q6_Vw_vasr_VwR(HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vw_vasr_VwR __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vasrw) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vx32.w+=vasr(Vu32.w,Rt32) + C Intrinsic Prototype: HVX_Vector Q6_Vw_vasracc_VwVwR(HVX_Vector Vx, HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vw_vasracc_VwVwR __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vasrw_acc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.h=vasr(Vu32.w,Vv32.w,Rt8) + C Intrinsic Prototype: HVX_Vector Q6_Vh_vasr_VwVwR(HVX_Vector Vu, HVX_Vector Vv, Word32 Rt) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vh_vasr_VwVwR __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vasrwh) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.h=vasr(Vu32.w,Vv32.w,Rt8):rnd:sat + C Intrinsic Prototype: HVX_Vector Q6_Vh_vasr_VwVwR_rnd_sat(HVX_Vector Vu, HVX_Vector Vv, Word32 Rt) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vh_vasr_VwVwR_rnd_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vasrwhrndsat) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.h=vasr(Vu32.w,Vv32.w,Rt8):sat + C Intrinsic Prototype: HVX_Vector Q6_Vh_vasr_VwVwR_sat(HVX_Vector Vu, HVX_Vector Vv, Word32 Rt) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vh_vasr_VwVwR_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vasrwhsat) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.uh=vasr(Vu32.w,Vv32.w,Rt8):sat + C Intrinsic Prototype: HVX_Vector Q6_Vuh_vasr_VwVwR_sat(HVX_Vector Vu, HVX_Vector Vv, Word32 Rt) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vuh_vasr_VwVwR_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vasrwuhsat) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.w=vasr(Vu32.w,Vv32.w) + C Intrinsic Prototype: HVX_Vector Q6_Vw_vasr_VwVw(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vw_vasr_VwVw __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vasrwv) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32=Vu32 + C Intrinsic Prototype: HVX_Vector Q6_V_equals_V(HVX_Vector Vu) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_V_equals_V __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vassign) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32=Vuu32 + C Intrinsic Prototype: HVX_VectorPair Q6_W_equals_W(HVX_VectorPair Vuu) + Instruction Type: CVI_VA_DV + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_W_equals_W __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vassignp) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.h=vavg(Vu32.h,Vv32.h) + C Intrinsic Prototype: HVX_Vector Q6_Vh_vavg_VhVh(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vh_vavg_VhVh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vavgh) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.h=vavg(Vu32.h,Vv32.h):rnd + C Intrinsic Prototype: HVX_Vector Q6_Vh_vavg_VhVh_rnd(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vh_vavg_VhVh_rnd __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vavghrnd) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.ub=vavg(Vu32.ub,Vv32.ub) + C Intrinsic Prototype: HVX_Vector Q6_Vub_vavg_VubVub(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vub_vavg_VubVub __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vavgub) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.ub=vavg(Vu32.ub,Vv32.ub):rnd + C Intrinsic Prototype: HVX_Vector Q6_Vub_vavg_VubVub_rnd(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vub_vavg_VubVub_rnd __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vavgubrnd) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.uh=vavg(Vu32.uh,Vv32.uh) + C Intrinsic Prototype: HVX_Vector Q6_Vuh_vavg_VuhVuh(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vuh_vavg_VuhVuh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vavguh) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.uh=vavg(Vu32.uh,Vv32.uh):rnd + C Intrinsic Prototype: HVX_Vector Q6_Vuh_vavg_VuhVuh_rnd(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vuh_vavg_VuhVuh_rnd __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vavguhrnd) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.w=vavg(Vu32.w,Vv32.w) + C Intrinsic Prototype: HVX_Vector Q6_Vw_vavg_VwVw(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vw_vavg_VwVw __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vavgw) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.w=vavg(Vu32.w,Vv32.w):rnd + C Intrinsic Prototype: HVX_Vector Q6_Vw_vavg_VwVw_rnd(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vw_vavg_VwVw_rnd __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vavgwrnd) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.uh=vcl0(Vu32.uh) + C Intrinsic Prototype: HVX_Vector Q6_Vuh_vcl0_Vuh(HVX_Vector Vu) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vuh_vcl0_Vuh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vcl0h) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.uw=vcl0(Vu32.uw) + C Intrinsic Prototype: HVX_Vector Q6_Vuw_vcl0_Vuw(HVX_Vector Vu) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vuw_vcl0_Vuw __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vcl0w) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32=vcombine(Vu32,Vv32) + C Intrinsic Prototype: HVX_VectorPair Q6_W_vcombine_VV(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA_DV + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_W_vcombine_VV __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vcombine) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32=#0 + C Intrinsic Prototype: HVX_Vector Q6_V_vzero() + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_V_vzero __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vd0) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.b=vdeal(Vu32.b) + C Intrinsic Prototype: HVX_Vector Q6_Vb_vdeal_Vb(HVX_Vector Vu) + Instruction Type: CVI_VP + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vb_vdeal_Vb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vdealb) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.b=vdeale(Vu32.b,Vv32.b) + C Intrinsic Prototype: HVX_Vector Q6_Vb_vdeale_VbVb(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VP + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vb_vdeale_VbVb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vdealb4w) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.h=vdeal(Vu32.h) + C Intrinsic Prototype: HVX_Vector Q6_Vh_vdeal_Vh(HVX_Vector Vu) + Instruction Type: CVI_VP + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vh_vdeal_Vh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vdealh) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32=vdeal(Vu32,Vv32,Rt8) + C Intrinsic Prototype: HVX_VectorPair Q6_W_vdeal_VVR(HVX_Vector Vu, HVX_Vector Vv, Word32 Rt) + Instruction Type: CVI_VP_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_W_vdeal_VVR __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vdealvdd) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32=vdelta(Vu32,Vv32) + C Intrinsic Prototype: HVX_Vector Q6_V_vdelta_VV(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VP + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_V_vdelta_VV __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vdelta) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.h=vdmpy(Vu32.ub,Rt32.b) + C Intrinsic Prototype: HVX_Vector Q6_Vh_vdmpy_VubRb(HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VX + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vh_vdmpy_VubRb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vdmpybus) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vx32.h+=vdmpy(Vu32.ub,Rt32.b) + C Intrinsic Prototype: HVX_Vector Q6_Vh_vdmpyacc_VhVubRb(HVX_Vector Vx, HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VX + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vh_vdmpyacc_VhVubRb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vdmpybus_acc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.h=vdmpy(Vuu32.ub,Rt32.b) + C Intrinsic Prototype: HVX_VectorPair Q6_Wh_vdmpy_WubRb(HVX_VectorPair Vuu, Word32 Rt) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Wh_vdmpy_WubRb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vdmpybus_dv) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vxx32.h+=vdmpy(Vuu32.ub,Rt32.b) + C Intrinsic Prototype: HVX_VectorPair Q6_Wh_vdmpyacc_WhWubRb(HVX_VectorPair Vxx, HVX_VectorPair Vuu, Word32 Rt) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Wh_vdmpyacc_WhWubRb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vdmpybus_dv_acc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.w=vdmpy(Vu32.h,Rt32.b) + C Intrinsic Prototype: HVX_Vector Q6_Vw_vdmpy_VhRb(HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VX + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vw_vdmpy_VhRb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vdmpyhb) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vx32.w+=vdmpy(Vu32.h,Rt32.b) + C Intrinsic Prototype: HVX_Vector Q6_Vw_vdmpyacc_VwVhRb(HVX_Vector Vx, HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VX + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vw_vdmpyacc_VwVhRb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vdmpyhb_acc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.w=vdmpy(Vuu32.h,Rt32.b) + C Intrinsic Prototype: HVX_VectorPair Q6_Ww_vdmpy_WhRb(HVX_VectorPair Vuu, Word32 Rt) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Ww_vdmpy_WhRb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vdmpyhb_dv) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vxx32.w+=vdmpy(Vuu32.h,Rt32.b) + C Intrinsic Prototype: HVX_VectorPair Q6_Ww_vdmpyacc_WwWhRb(HVX_VectorPair Vxx, HVX_VectorPair Vuu, Word32 Rt) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Ww_vdmpyacc_WwWhRb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vdmpyhb_dv_acc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.w=vdmpy(Vuu32.h,Rt32.h):sat + C Intrinsic Prototype: HVX_Vector Q6_Vw_vdmpy_WhRh_sat(HVX_VectorPair Vuu, Word32 Rt) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vw_vdmpy_WhRh_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vdmpyhisat) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vx32.w+=vdmpy(Vuu32.h,Rt32.h):sat + C Intrinsic Prototype: HVX_Vector Q6_Vw_vdmpyacc_VwWhRh_sat(HVX_Vector Vx, HVX_VectorPair Vuu, Word32 Rt) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vw_vdmpyacc_VwWhRh_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vdmpyhisat_acc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.w=vdmpy(Vu32.h,Rt32.h):sat + C Intrinsic Prototype: HVX_Vector Q6_Vw_vdmpy_VhRh_sat(HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vw_vdmpy_VhRh_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vdmpyhsat) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vx32.w+=vdmpy(Vu32.h,Rt32.h):sat + C Intrinsic Prototype: HVX_Vector Q6_Vw_vdmpyacc_VwVhRh_sat(HVX_Vector Vx, HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vw_vdmpyacc_VwVhRh_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vdmpyhsat_acc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.w=vdmpy(Vuu32.h,Rt32.uh,#1):sat + C Intrinsic Prototype: HVX_Vector Q6_Vw_vdmpy_WhRuh_sat(HVX_VectorPair Vuu, Word32 Rt) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vw_vdmpy_WhRuh_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vdmpyhsuisat) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vx32.w+=vdmpy(Vuu32.h,Rt32.uh,#1):sat + C Intrinsic Prototype: HVX_Vector Q6_Vw_vdmpyacc_VwWhRuh_sat(HVX_Vector Vx, HVX_VectorPair Vuu, Word32 Rt) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vw_vdmpyacc_VwWhRuh_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vdmpyhsuisat_acc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.w=vdmpy(Vu32.h,Rt32.uh):sat + C Intrinsic Prototype: HVX_Vector Q6_Vw_vdmpy_VhRuh_sat(HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vw_vdmpy_VhRuh_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vdmpyhsusat) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vx32.w+=vdmpy(Vu32.h,Rt32.uh):sat + C Intrinsic Prototype: HVX_Vector Q6_Vw_vdmpyacc_VwVhRuh_sat(HVX_Vector Vx, HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vw_vdmpyacc_VwVhRuh_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vdmpyhsusat_acc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.w=vdmpy(Vu32.h,Vv32.h):sat + C Intrinsic Prototype: HVX_Vector Q6_Vw_vdmpy_VhVh_sat(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vw_vdmpy_VhVh_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vdmpyhvsat) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vx32.w+=vdmpy(Vu32.h,Vv32.h):sat + C Intrinsic Prototype: HVX_Vector Q6_Vw_vdmpyacc_VwVhVh_sat(HVX_Vector Vx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vw_vdmpyacc_VwVhVh_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vdmpyhvsat_acc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.uw=vdsad(Vuu32.uh,Rt32.uh) + C Intrinsic Prototype: HVX_VectorPair Q6_Wuw_vdsad_WuhRuh(HVX_VectorPair Vuu, Word32 Rt) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Wuw_vdsad_WuhRuh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vdsaduh) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vxx32.uw+=vdsad(Vuu32.uh,Rt32.uh) + C Intrinsic Prototype: HVX_VectorPair Q6_Wuw_vdsadacc_WuwWuhRuh(HVX_VectorPair Vxx, HVX_VectorPair Vuu, Word32 Rt) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Wuw_vdsadacc_WuwWuhRuh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vdsaduh_acc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Qd4=vcmp.eq(Vu32.b,Vv32.b) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_vcmp_eq_VbVb(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Q_vcmp_eq_VbVb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_veqb) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Qx4&=vcmp.eq(Vu32.b,Vv32.b) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_vcmp_eqand_QVbVb(HVX_VectorPred Qx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Q_vcmp_eqand_QVbVb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_veqb_and) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Qx4|=vcmp.eq(Vu32.b,Vv32.b) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_vcmp_eqor_QVbVb(HVX_VectorPred Qx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Q_vcmp_eqor_QVbVb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_veqb_or) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Qx4^=vcmp.eq(Vu32.b,Vv32.b) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_vcmp_eqxacc_QVbVb(HVX_VectorPred Qx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Q_vcmp_eqxacc_QVbVb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_veqb_xor) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Qd4=vcmp.eq(Vu32.h,Vv32.h) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_vcmp_eq_VhVh(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Q_vcmp_eq_VhVh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_veqh) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Qx4&=vcmp.eq(Vu32.h,Vv32.h) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_vcmp_eqand_QVhVh(HVX_VectorPred Qx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Q_vcmp_eqand_QVhVh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_veqh_and) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Qx4|=vcmp.eq(Vu32.h,Vv32.h) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_vcmp_eqor_QVhVh(HVX_VectorPred Qx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Q_vcmp_eqor_QVhVh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_veqh_or) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Qx4^=vcmp.eq(Vu32.h,Vv32.h) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_vcmp_eqxacc_QVhVh(HVX_VectorPred Qx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Q_vcmp_eqxacc_QVhVh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_veqh_xor) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Qd4=vcmp.eq(Vu32.w,Vv32.w) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_vcmp_eq_VwVw(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Q_vcmp_eq_VwVw __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_veqw) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Qx4&=vcmp.eq(Vu32.w,Vv32.w) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_vcmp_eqand_QVwVw(HVX_VectorPred Qx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Q_vcmp_eqand_QVwVw __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_veqw_and) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Qx4|=vcmp.eq(Vu32.w,Vv32.w) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_vcmp_eqor_QVwVw(HVX_VectorPred Qx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Q_vcmp_eqor_QVwVw __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_veqw_or) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Qx4^=vcmp.eq(Vu32.w,Vv32.w) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_vcmp_eqxacc_QVwVw(HVX_VectorPred Qx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Q_vcmp_eqxacc_QVwVw __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_veqw_xor) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Qd4=vcmp.gt(Vu32.b,Vv32.b) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_vcmp_gt_VbVb(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Q_vcmp_gt_VbVb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vgtb) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Qx4&=vcmp.gt(Vu32.b,Vv32.b) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_vcmp_gtand_QVbVb(HVX_VectorPred Qx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Q_vcmp_gtand_QVbVb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vgtb_and) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Qx4|=vcmp.gt(Vu32.b,Vv32.b) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_vcmp_gtor_QVbVb(HVX_VectorPred Qx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Q_vcmp_gtor_QVbVb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vgtb_or) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Qx4^=vcmp.gt(Vu32.b,Vv32.b) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_vcmp_gtxacc_QVbVb(HVX_VectorPred Qx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Q_vcmp_gtxacc_QVbVb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vgtb_xor) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Qd4=vcmp.gt(Vu32.h,Vv32.h) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_vcmp_gt_VhVh(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Q_vcmp_gt_VhVh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vgth) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Qx4&=vcmp.gt(Vu32.h,Vv32.h) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_vcmp_gtand_QVhVh(HVX_VectorPred Qx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Q_vcmp_gtand_QVhVh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vgth_and) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Qx4|=vcmp.gt(Vu32.h,Vv32.h) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_vcmp_gtor_QVhVh(HVX_VectorPred Qx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Q_vcmp_gtor_QVhVh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vgth_or) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Qx4^=vcmp.gt(Vu32.h,Vv32.h) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_vcmp_gtxacc_QVhVh(HVX_VectorPred Qx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Q_vcmp_gtxacc_QVhVh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vgth_xor) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Qd4=vcmp.gt(Vu32.ub,Vv32.ub) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_vcmp_gt_VubVub(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Q_vcmp_gt_VubVub __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vgtub) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Qx4&=vcmp.gt(Vu32.ub,Vv32.ub) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_vcmp_gtand_QVubVub(HVX_VectorPred Qx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Q_vcmp_gtand_QVubVub __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vgtub_and) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Qx4|=vcmp.gt(Vu32.ub,Vv32.ub) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_vcmp_gtor_QVubVub(HVX_VectorPred Qx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Q_vcmp_gtor_QVubVub __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vgtub_or) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Qx4^=vcmp.gt(Vu32.ub,Vv32.ub) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_vcmp_gtxacc_QVubVub(HVX_VectorPred Qx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Q_vcmp_gtxacc_QVubVub __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vgtub_xor) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Qd4=vcmp.gt(Vu32.uh,Vv32.uh) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_vcmp_gt_VuhVuh(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Q_vcmp_gt_VuhVuh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vgtuh) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Qx4&=vcmp.gt(Vu32.uh,Vv32.uh) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_vcmp_gtand_QVuhVuh(HVX_VectorPred Qx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Q_vcmp_gtand_QVuhVuh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vgtuh_and) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Qx4|=vcmp.gt(Vu32.uh,Vv32.uh) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_vcmp_gtor_QVuhVuh(HVX_VectorPred Qx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Q_vcmp_gtor_QVuhVuh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vgtuh_or) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Qx4^=vcmp.gt(Vu32.uh,Vv32.uh) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_vcmp_gtxacc_QVuhVuh(HVX_VectorPred Qx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Q_vcmp_gtxacc_QVuhVuh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vgtuh_xor) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Qd4=vcmp.gt(Vu32.uw,Vv32.uw) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_vcmp_gt_VuwVuw(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Q_vcmp_gt_VuwVuw __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vgtuw) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Qx4&=vcmp.gt(Vu32.uw,Vv32.uw) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_vcmp_gtand_QVuwVuw(HVX_VectorPred Qx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Q_vcmp_gtand_QVuwVuw __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vgtuw_and) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Qx4|=vcmp.gt(Vu32.uw,Vv32.uw) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_vcmp_gtor_QVuwVuw(HVX_VectorPred Qx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Q_vcmp_gtor_QVuwVuw __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vgtuw_or) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Qx4^=vcmp.gt(Vu32.uw,Vv32.uw) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_vcmp_gtxacc_QVuwVuw(HVX_VectorPred Qx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Q_vcmp_gtxacc_QVuwVuw __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vgtuw_xor) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Qd4=vcmp.gt(Vu32.w,Vv32.w) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_vcmp_gt_VwVw(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Q_vcmp_gt_VwVw __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vgtw) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Qx4&=vcmp.gt(Vu32.w,Vv32.w) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_vcmp_gtand_QVwVw(HVX_VectorPred Qx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Q_vcmp_gtand_QVwVw __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vgtw_and) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Qx4|=vcmp.gt(Vu32.w,Vv32.w) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_vcmp_gtor_QVwVw(HVX_VectorPred Qx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Q_vcmp_gtor_QVwVw __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vgtw_or) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Qx4^=vcmp.gt(Vu32.w,Vv32.w) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_vcmp_gtxacc_QVwVw(HVX_VectorPred Qx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Q_vcmp_gtxacc_QVwVw __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vgtw_xor) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vx32.w=vinsert(Rt32) + C Intrinsic Prototype: HVX_Vector Q6_Vw_vinsert_VwR(HVX_Vector Vx, Word32 Rt) + Instruction Type: CVI_VX_LATE + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vw_vinsert_VwR __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vinsertwr) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32=vlalign(Vu32,Vv32,Rt8) + C Intrinsic Prototype: HVX_Vector Q6_V_vlalign_VVR(HVX_Vector Vu, HVX_Vector Vv, Word32 Rt) + Instruction Type: CVI_VP + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_V_vlalign_VVR __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vlalignb) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32=vlalign(Vu32,Vv32,#u3) + C Intrinsic Prototype: HVX_Vector Q6_V_vlalign_VVI(HVX_Vector Vu, HVX_Vector Vv, Word32 Iu3) + Instruction Type: CVI_VP + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_V_vlalign_VVI __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vlalignbi) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.uh=vlsr(Vu32.uh,Rt32) + C Intrinsic Prototype: HVX_Vector Q6_Vuh_vlsr_VuhR(HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vuh_vlsr_VuhR __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vlsrh) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.h=vlsr(Vu32.h,Vv32.h) + C Intrinsic Prototype: HVX_Vector Q6_Vh_vlsr_VhVh(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vh_vlsr_VhVh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vlsrhv) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.uw=vlsr(Vu32.uw,Rt32) + C Intrinsic Prototype: HVX_Vector Q6_Vuw_vlsr_VuwR(HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vuw_vlsr_VuwR __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vlsrw) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.w=vlsr(Vu32.w,Vv32.w) + C Intrinsic Prototype: HVX_Vector Q6_Vw_vlsr_VwVw(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vw_vlsr_VwVw __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vlsrwv) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.b=vlut32(Vu32.b,Vv32.b,Rt8) + C Intrinsic Prototype: HVX_Vector Q6_Vb_vlut32_VbVbR(HVX_Vector Vu, HVX_Vector Vv, Word32 Rt) + Instruction Type: CVI_VP + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vb_vlut32_VbVbR __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vlutvvb) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vx32.b|=vlut32(Vu32.b,Vv32.b,Rt8) + C Intrinsic Prototype: HVX_Vector Q6_Vb_vlut32or_VbVbVbR(HVX_Vector Vx, HVX_Vector Vu, HVX_Vector Vv, Word32 Rt) + Instruction Type: CVI_VP_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vb_vlut32or_VbVbVbR __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vlutvvb_oracc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.h=vlut16(Vu32.b,Vv32.h,Rt8) + C Intrinsic Prototype: HVX_VectorPair Q6_Wh_vlut16_VbVhR(HVX_Vector Vu, HVX_Vector Vv, Word32 Rt) + Instruction Type: CVI_VP_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Wh_vlut16_VbVhR __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vlutvwh) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vxx32.h|=vlut16(Vu32.b,Vv32.h,Rt8) + C Intrinsic Prototype: HVX_VectorPair Q6_Wh_vlut16or_WhVbVhR(HVX_VectorPair Vxx, HVX_Vector Vu, HVX_Vector Vv, Word32 Rt) + Instruction Type: CVI_VP_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Wh_vlut16or_WhVbVhR __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vlutvwh_oracc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.h=vmax(Vu32.h,Vv32.h) + C Intrinsic Prototype: HVX_Vector Q6_Vh_vmax_VhVh(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vh_vmax_VhVh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmaxh) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.ub=vmax(Vu32.ub,Vv32.ub) + C Intrinsic Prototype: HVX_Vector Q6_Vub_vmax_VubVub(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vub_vmax_VubVub __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmaxub) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.uh=vmax(Vu32.uh,Vv32.uh) + C Intrinsic Prototype: HVX_Vector Q6_Vuh_vmax_VuhVuh(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vuh_vmax_VuhVuh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmaxuh) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.w=vmax(Vu32.w,Vv32.w) + C Intrinsic Prototype: HVX_Vector Q6_Vw_vmax_VwVw(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vw_vmax_VwVw __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmaxw) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.h=vmin(Vu32.h,Vv32.h) + C Intrinsic Prototype: HVX_Vector Q6_Vh_vmin_VhVh(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vh_vmin_VhVh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vminh) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.ub=vmin(Vu32.ub,Vv32.ub) + C Intrinsic Prototype: HVX_Vector Q6_Vub_vmin_VubVub(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vub_vmin_VubVub __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vminub) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.uh=vmin(Vu32.uh,Vv32.uh) + C Intrinsic Prototype: HVX_Vector Q6_Vuh_vmin_VuhVuh(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vuh_vmin_VuhVuh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vminuh) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.w=vmin(Vu32.w,Vv32.w) + C Intrinsic Prototype: HVX_Vector Q6_Vw_vmin_VwVw(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vw_vmin_VwVw __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vminw) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.h=vmpa(Vuu32.ub,Rt32.b) + C Intrinsic Prototype: HVX_VectorPair Q6_Wh_vmpa_WubRb(HVX_VectorPair Vuu, Word32 Rt) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Wh_vmpa_WubRb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpabus) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vxx32.h+=vmpa(Vuu32.ub,Rt32.b) + C Intrinsic Prototype: HVX_VectorPair Q6_Wh_vmpaacc_WhWubRb(HVX_VectorPair Vxx, HVX_VectorPair Vuu, Word32 Rt) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Wh_vmpaacc_WhWubRb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpabus_acc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.h=vmpa(Vuu32.ub,Vvv32.b) + C Intrinsic Prototype: HVX_VectorPair Q6_Wh_vmpa_WubWb(HVX_VectorPair Vuu, HVX_VectorPair Vvv) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Wh_vmpa_WubWb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpabusv) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.h=vmpa(Vuu32.ub,Vvv32.ub) + C Intrinsic Prototype: HVX_VectorPair Q6_Wh_vmpa_WubWub(HVX_VectorPair Vuu, HVX_VectorPair Vvv) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Wh_vmpa_WubWub __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpabuuv) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.w=vmpa(Vuu32.h,Rt32.b) + C Intrinsic Prototype: HVX_VectorPair Q6_Ww_vmpa_WhRb(HVX_VectorPair Vuu, Word32 Rt) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Ww_vmpa_WhRb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpahb) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vxx32.w+=vmpa(Vuu32.h,Rt32.b) + C Intrinsic Prototype: HVX_VectorPair Q6_Ww_vmpaacc_WwWhRb(HVX_VectorPair Vxx, HVX_VectorPair Vuu, Word32 Rt) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Ww_vmpaacc_WwWhRb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpahb_acc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.h=vmpy(Vu32.ub,Rt32.b) + C Intrinsic Prototype: HVX_VectorPair Q6_Wh_vmpy_VubRb(HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Wh_vmpy_VubRb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpybus) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vxx32.h+=vmpy(Vu32.ub,Rt32.b) + C Intrinsic Prototype: HVX_VectorPair Q6_Wh_vmpyacc_WhVubRb(HVX_VectorPair Vxx, HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Wh_vmpyacc_WhVubRb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpybus_acc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.h=vmpy(Vu32.ub,Vv32.b) + C Intrinsic Prototype: HVX_VectorPair Q6_Wh_vmpy_VubVb(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Wh_vmpy_VubVb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpybusv) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vxx32.h+=vmpy(Vu32.ub,Vv32.b) + C Intrinsic Prototype: HVX_VectorPair Q6_Wh_vmpyacc_WhVubVb(HVX_VectorPair Vxx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Wh_vmpyacc_WhVubVb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpybusv_acc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.h=vmpy(Vu32.b,Vv32.b) + C Intrinsic Prototype: HVX_VectorPair Q6_Wh_vmpy_VbVb(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Wh_vmpy_VbVb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpybv) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vxx32.h+=vmpy(Vu32.b,Vv32.b) + C Intrinsic Prototype: HVX_VectorPair Q6_Wh_vmpyacc_WhVbVb(HVX_VectorPair Vxx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Wh_vmpyacc_WhVbVb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpybv_acc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.w=vmpye(Vu32.w,Vv32.uh) + C Intrinsic Prototype: HVX_Vector Q6_Vw_vmpye_VwVuh(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vw_vmpye_VwVuh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpyewuh) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.w=vmpy(Vu32.h,Rt32.h) + C Intrinsic Prototype: HVX_VectorPair Q6_Ww_vmpy_VhRh(HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Ww_vmpy_VhRh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpyh) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vxx32.w+=vmpy(Vu32.h,Rt32.h):sat + C Intrinsic Prototype: HVX_VectorPair Q6_Ww_vmpyacc_WwVhRh_sat(HVX_VectorPair Vxx, HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Ww_vmpyacc_WwVhRh_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpyhsat_acc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.h=vmpy(Vu32.h,Rt32.h):<<1:rnd:sat + C Intrinsic Prototype: HVX_Vector Q6_Vh_vmpy_VhRh_s1_rnd_sat(HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vh_vmpy_VhRh_s1_rnd_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpyhsrs) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.h=vmpy(Vu32.h,Rt32.h):<<1:sat + C Intrinsic Prototype: HVX_Vector Q6_Vh_vmpy_VhRh_s1_sat(HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vh_vmpy_VhRh_s1_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpyhss) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.w=vmpy(Vu32.h,Vv32.uh) + C Intrinsic Prototype: HVX_VectorPair Q6_Ww_vmpy_VhVuh(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Ww_vmpy_VhVuh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpyhus) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vxx32.w+=vmpy(Vu32.h,Vv32.uh) + C Intrinsic Prototype: HVX_VectorPair Q6_Ww_vmpyacc_WwVhVuh(HVX_VectorPair Vxx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Ww_vmpyacc_WwVhVuh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpyhus_acc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.w=vmpy(Vu32.h,Vv32.h) + C Intrinsic Prototype: HVX_VectorPair Q6_Ww_vmpy_VhVh(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Ww_vmpy_VhVh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpyhv) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vxx32.w+=vmpy(Vu32.h,Vv32.h) + C Intrinsic Prototype: HVX_VectorPair Q6_Ww_vmpyacc_WwVhVh(HVX_VectorPair Vxx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Ww_vmpyacc_WwVhVh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpyhv_acc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.h=vmpy(Vu32.h,Vv32.h):<<1:rnd:sat + C Intrinsic Prototype: HVX_Vector Q6_Vh_vmpy_VhVh_s1_rnd_sat(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vh_vmpy_VhVh_s1_rnd_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpyhvsrs) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.w=vmpyieo(Vu32.h,Vv32.h) + C Intrinsic Prototype: HVX_Vector Q6_Vw_vmpyieo_VhVh(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vw_vmpyieo_VhVh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpyieoh) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vx32.w+=vmpyie(Vu32.w,Vv32.h) + C Intrinsic Prototype: HVX_Vector Q6_Vw_vmpyieacc_VwVwVh(HVX_Vector Vx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vw_vmpyieacc_VwVwVh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpyiewh_acc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.w=vmpyie(Vu32.w,Vv32.uh) + C Intrinsic Prototype: HVX_Vector Q6_Vw_vmpyie_VwVuh(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vw_vmpyie_VwVuh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpyiewuh) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vx32.w+=vmpyie(Vu32.w,Vv32.uh) + C Intrinsic Prototype: HVX_Vector Q6_Vw_vmpyieacc_VwVwVuh(HVX_Vector Vx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vw_vmpyieacc_VwVwVuh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpyiewuh_acc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.h=vmpyi(Vu32.h,Vv32.h) + C Intrinsic Prototype: HVX_Vector Q6_Vh_vmpyi_VhVh(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vh_vmpyi_VhVh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpyih) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vx32.h+=vmpyi(Vu32.h,Vv32.h) + C Intrinsic Prototype: HVX_Vector Q6_Vh_vmpyiacc_VhVhVh(HVX_Vector Vx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vh_vmpyiacc_VhVhVh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpyih_acc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.h=vmpyi(Vu32.h,Rt32.b) + C Intrinsic Prototype: HVX_Vector Q6_Vh_vmpyi_VhRb(HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VX + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vh_vmpyi_VhRb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpyihb) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vx32.h+=vmpyi(Vu32.h,Rt32.b) + C Intrinsic Prototype: HVX_Vector Q6_Vh_vmpyiacc_VhVhRb(HVX_Vector Vx, HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VX + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vh_vmpyiacc_VhVhRb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpyihb_acc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.w=vmpyio(Vu32.w,Vv32.h) + C Intrinsic Prototype: HVX_Vector Q6_Vw_vmpyio_VwVh(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vw_vmpyio_VwVh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpyiowh) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.w=vmpyi(Vu32.w,Rt32.b) + C Intrinsic Prototype: HVX_Vector Q6_Vw_vmpyi_VwRb(HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VX + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vw_vmpyi_VwRb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpyiwb) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vx32.w+=vmpyi(Vu32.w,Rt32.b) + C Intrinsic Prototype: HVX_Vector Q6_Vw_vmpyiacc_VwVwRb(HVX_Vector Vx, HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VX + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vw_vmpyiacc_VwVwRb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpyiwb_acc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.w=vmpyi(Vu32.w,Rt32.h) + C Intrinsic Prototype: HVX_Vector Q6_Vw_vmpyi_VwRh(HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vw_vmpyi_VwRh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpyiwh) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vx32.w+=vmpyi(Vu32.w,Rt32.h) + C Intrinsic Prototype: HVX_Vector Q6_Vw_vmpyiacc_VwVwRh(HVX_Vector Vx, HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vw_vmpyiacc_VwVwRh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpyiwh_acc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.w=vmpyo(Vu32.w,Vv32.h):<<1:sat + C Intrinsic Prototype: HVX_Vector Q6_Vw_vmpyo_VwVh_s1_sat(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vw_vmpyo_VwVh_s1_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpyowh) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.w=vmpyo(Vu32.w,Vv32.h):<<1:rnd:sat + C Intrinsic Prototype: HVX_Vector Q6_Vw_vmpyo_VwVh_s1_rnd_sat(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vw_vmpyo_VwVh_s1_rnd_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpyowh_rnd) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vx32.w+=vmpyo(Vu32.w,Vv32.h):<<1:rnd:sat:shift + C Intrinsic Prototype: HVX_Vector Q6_Vw_vmpyoacc_VwVwVh_s1_rnd_sat_shift(HVX_Vector Vx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vw_vmpyoacc_VwVwVh_s1_rnd_sat_shift __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpyowh_rnd_sacc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vx32.w+=vmpyo(Vu32.w,Vv32.h):<<1:sat:shift + C Intrinsic Prototype: HVX_Vector Q6_Vw_vmpyoacc_VwVwVh_s1_sat_shift(HVX_Vector Vx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vw_vmpyoacc_VwVwVh_s1_sat_shift __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpyowh_sacc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.uh=vmpy(Vu32.ub,Rt32.ub) + C Intrinsic Prototype: HVX_VectorPair Q6_Wuh_vmpy_VubRub(HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Wuh_vmpy_VubRub __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpyub) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vxx32.uh+=vmpy(Vu32.ub,Rt32.ub) + C Intrinsic Prototype: HVX_VectorPair Q6_Wuh_vmpyacc_WuhVubRub(HVX_VectorPair Vxx, HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Wuh_vmpyacc_WuhVubRub __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpyub_acc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.uh=vmpy(Vu32.ub,Vv32.ub) + C Intrinsic Prototype: HVX_VectorPair Q6_Wuh_vmpy_VubVub(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Wuh_vmpy_VubVub __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpyubv) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vxx32.uh+=vmpy(Vu32.ub,Vv32.ub) + C Intrinsic Prototype: HVX_VectorPair Q6_Wuh_vmpyacc_WuhVubVub(HVX_VectorPair Vxx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Wuh_vmpyacc_WuhVubVub __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpyubv_acc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.uw=vmpy(Vu32.uh,Rt32.uh) + C Intrinsic Prototype: HVX_VectorPair Q6_Wuw_vmpy_VuhRuh(HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Wuw_vmpy_VuhRuh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpyuh) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vxx32.uw+=vmpy(Vu32.uh,Rt32.uh) + C Intrinsic Prototype: HVX_VectorPair Q6_Wuw_vmpyacc_WuwVuhRuh(HVX_VectorPair Vxx, HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Wuw_vmpyacc_WuwVuhRuh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpyuh_acc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.uw=vmpy(Vu32.uh,Vv32.uh) + C Intrinsic Prototype: HVX_VectorPair Q6_Wuw_vmpy_VuhVuh(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Wuw_vmpy_VuhVuh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpyuhv) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vxx32.uw+=vmpy(Vu32.uh,Vv32.uh) + C Intrinsic Prototype: HVX_VectorPair Q6_Wuw_vmpyacc_WuwVuhVuh(HVX_VectorPair Vxx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Wuw_vmpyacc_WuwVuhVuh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpyuhv_acc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32=vmux(Qt4,Vu32,Vv32) + C Intrinsic Prototype: HVX_Vector Q6_V_vmux_QVV(HVX_VectorPred Qt, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_V_vmux_QVV __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmux) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.h=vnavg(Vu32.h,Vv32.h) + C Intrinsic Prototype: HVX_Vector Q6_Vh_vnavg_VhVh(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vh_vnavg_VhVh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vnavgh) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.b=vnavg(Vu32.ub,Vv32.ub) + C Intrinsic Prototype: HVX_Vector Q6_Vb_vnavg_VubVub(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vb_vnavg_VubVub __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vnavgub) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.w=vnavg(Vu32.w,Vv32.w) + C Intrinsic Prototype: HVX_Vector Q6_Vw_vnavg_VwVw(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vw_vnavg_VwVw __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vnavgw) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.h=vnormamt(Vu32.h) + C Intrinsic Prototype: HVX_Vector Q6_Vh_vnormamt_Vh(HVX_Vector Vu) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vh_vnormamt_Vh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vnormamth) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.w=vnormamt(Vu32.w) + C Intrinsic Prototype: HVX_Vector Q6_Vw_vnormamt_Vw(HVX_Vector Vu) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vw_vnormamt_Vw __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vnormamtw) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32=vnot(Vu32) + C Intrinsic Prototype: HVX_Vector Q6_V_vnot_V(HVX_Vector Vu) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_V_vnot_V __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vnot) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32=vor(Vu32,Vv32) + C Intrinsic Prototype: HVX_Vector Q6_V_vor_VV(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_V_vor_VV __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vor) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.b=vpacke(Vu32.h,Vv32.h) + C Intrinsic Prototype: HVX_Vector Q6_Vb_vpacke_VhVh(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VP + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vb_vpacke_VhVh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vpackeb) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.h=vpacke(Vu32.w,Vv32.w) + C Intrinsic Prototype: HVX_Vector Q6_Vh_vpacke_VwVw(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VP + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vh_vpacke_VwVw __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vpackeh) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.b=vpack(Vu32.h,Vv32.h):sat + C Intrinsic Prototype: HVX_Vector Q6_Vb_vpack_VhVh_sat(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VP + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vb_vpack_VhVh_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vpackhb_sat) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.ub=vpack(Vu32.h,Vv32.h):sat + C Intrinsic Prototype: HVX_Vector Q6_Vub_vpack_VhVh_sat(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VP + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vub_vpack_VhVh_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vpackhub_sat) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.b=vpacko(Vu32.h,Vv32.h) + C Intrinsic Prototype: HVX_Vector Q6_Vb_vpacko_VhVh(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VP + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vb_vpacko_VhVh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vpackob) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.h=vpacko(Vu32.w,Vv32.w) + C Intrinsic Prototype: HVX_Vector Q6_Vh_vpacko_VwVw(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VP + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vh_vpacko_VwVw __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vpackoh) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.h=vpack(Vu32.w,Vv32.w):sat + C Intrinsic Prototype: HVX_Vector Q6_Vh_vpack_VwVw_sat(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VP + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vh_vpack_VwVw_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vpackwh_sat) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.uh=vpack(Vu32.w,Vv32.w):sat + C Intrinsic Prototype: HVX_Vector Q6_Vuh_vpack_VwVw_sat(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VP + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vuh_vpack_VwVw_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vpackwuh_sat) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.h=vpopcount(Vu32.h) + C Intrinsic Prototype: HVX_Vector Q6_Vh_vpopcount_Vh(HVX_Vector Vu) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vh_vpopcount_Vh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vpopcounth) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32=vrdelta(Vu32,Vv32) + C Intrinsic Prototype: HVX_Vector Q6_V_vrdelta_VV(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VP + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_V_vrdelta_VV __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vrdelta) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.w=vrmpy(Vu32.ub,Rt32.b) + C Intrinsic Prototype: HVX_Vector Q6_Vw_vrmpy_VubRb(HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VX + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vw_vrmpy_VubRb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vrmpybus) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vx32.w+=vrmpy(Vu32.ub,Rt32.b) + C Intrinsic Prototype: HVX_Vector Q6_Vw_vrmpyacc_VwVubRb(HVX_Vector Vx, HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VX + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vw_vrmpyacc_VwVubRb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vrmpybus_acc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.w=vrmpy(Vuu32.ub,Rt32.b,#u1) + C Intrinsic Prototype: HVX_VectorPair Q6_Ww_vrmpy_WubRbI(HVX_VectorPair Vuu, Word32 Rt, Word32 Iu1) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Ww_vrmpy_WubRbI __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vrmpybusi) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vxx32.w+=vrmpy(Vuu32.ub,Rt32.b,#u1) + C Intrinsic Prototype: HVX_VectorPair Q6_Ww_vrmpyacc_WwWubRbI(HVX_VectorPair Vxx, HVX_VectorPair Vuu, Word32 Rt, Word32 Iu1) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Ww_vrmpyacc_WwWubRbI __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vrmpybusi_acc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.w=vrmpy(Vu32.ub,Vv32.b) + C Intrinsic Prototype: HVX_Vector Q6_Vw_vrmpy_VubVb(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vw_vrmpy_VubVb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vrmpybusv) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vx32.w+=vrmpy(Vu32.ub,Vv32.b) + C Intrinsic Prototype: HVX_Vector Q6_Vw_vrmpyacc_VwVubVb(HVX_Vector Vx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vw_vrmpyacc_VwVubVb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vrmpybusv_acc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.w=vrmpy(Vu32.b,Vv32.b) + C Intrinsic Prototype: HVX_Vector Q6_Vw_vrmpy_VbVb(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vw_vrmpy_VbVb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vrmpybv) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vx32.w+=vrmpy(Vu32.b,Vv32.b) + C Intrinsic Prototype: HVX_Vector Q6_Vw_vrmpyacc_VwVbVb(HVX_Vector Vx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vw_vrmpyacc_VwVbVb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vrmpybv_acc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.uw=vrmpy(Vu32.ub,Rt32.ub) + C Intrinsic Prototype: HVX_Vector Q6_Vuw_vrmpy_VubRub(HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VX + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vuw_vrmpy_VubRub __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vrmpyub) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vx32.uw+=vrmpy(Vu32.ub,Rt32.ub) + C Intrinsic Prototype: HVX_Vector Q6_Vuw_vrmpyacc_VuwVubRub(HVX_Vector Vx, HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VX + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vuw_vrmpyacc_VuwVubRub __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vrmpyub_acc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.uw=vrmpy(Vuu32.ub,Rt32.ub,#u1) + C Intrinsic Prototype: HVX_VectorPair Q6_Wuw_vrmpy_WubRubI(HVX_VectorPair Vuu, Word32 Rt, Word32 Iu1) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Wuw_vrmpy_WubRubI __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vrmpyubi) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vxx32.uw+=vrmpy(Vuu32.ub,Rt32.ub,#u1) + C Intrinsic Prototype: HVX_VectorPair Q6_Wuw_vrmpyacc_WuwWubRubI(HVX_VectorPair Vxx, HVX_VectorPair Vuu, Word32 Rt, Word32 Iu1) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Wuw_vrmpyacc_WuwWubRubI __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vrmpyubi_acc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.uw=vrmpy(Vu32.ub,Vv32.ub) + C Intrinsic Prototype: HVX_Vector Q6_Vuw_vrmpy_VubVub(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vuw_vrmpy_VubVub __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vrmpyubv) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vx32.uw+=vrmpy(Vu32.ub,Vv32.ub) + C Intrinsic Prototype: HVX_Vector Q6_Vuw_vrmpyacc_VuwVubVub(HVX_Vector Vx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vuw_vrmpyacc_VuwVubVub __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vrmpyubv_acc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32=vror(Vu32,Rt32) + C Intrinsic Prototype: HVX_Vector Q6_V_vror_VR(HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VP + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_V_vror_VR __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vror) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.b=vround(Vu32.h,Vv32.h):sat + C Intrinsic Prototype: HVX_Vector Q6_Vb_vround_VhVh_sat(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vb_vround_VhVh_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vroundhb) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.ub=vround(Vu32.h,Vv32.h):sat + C Intrinsic Prototype: HVX_Vector Q6_Vub_vround_VhVh_sat(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vub_vround_VhVh_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vroundhub) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.h=vround(Vu32.w,Vv32.w):sat + C Intrinsic Prototype: HVX_Vector Q6_Vh_vround_VwVw_sat(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vh_vround_VwVw_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vroundwh) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.uh=vround(Vu32.w,Vv32.w):sat + C Intrinsic Prototype: HVX_Vector Q6_Vuh_vround_VwVw_sat(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vuh_vround_VwVw_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vroundwuh) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.uw=vrsad(Vuu32.ub,Rt32.ub,#u1) + C Intrinsic Prototype: HVX_VectorPair Q6_Wuw_vrsad_WubRubI(HVX_VectorPair Vuu, Word32 Rt, Word32 Iu1) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Wuw_vrsad_WubRubI __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vrsadubi) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vxx32.uw+=vrsad(Vuu32.ub,Rt32.ub,#u1) + C Intrinsic Prototype: HVX_VectorPair Q6_Wuw_vrsadacc_WuwWubRubI(HVX_VectorPair Vxx, HVX_VectorPair Vuu, Word32 Rt, Word32 Iu1) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Wuw_vrsadacc_WuwWubRubI __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vrsadubi_acc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.ub=vsat(Vu32.h,Vv32.h) + C Intrinsic Prototype: HVX_Vector Q6_Vub_vsat_VhVh(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vub_vsat_VhVh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vsathub) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.h=vsat(Vu32.w,Vv32.w) + C Intrinsic Prototype: HVX_Vector Q6_Vh_vsat_VwVw(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vh_vsat_VwVw __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vsatwh) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.h=vsxt(Vu32.b) + C Intrinsic Prototype: HVX_VectorPair Q6_Wh_vsxt_Vb(HVX_Vector Vu) + Instruction Type: CVI_VA_DV + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Wh_vsxt_Vb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vsb) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.w=vsxt(Vu32.h) + C Intrinsic Prototype: HVX_VectorPair Q6_Ww_vsxt_Vh(HVX_Vector Vu) + Instruction Type: CVI_VA_DV + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Ww_vsxt_Vh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vsh) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.h=vshuffe(Vu32.h,Vv32.h) + C Intrinsic Prototype: HVX_Vector Q6_Vh_vshuffe_VhVh(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vh_vshuffe_VhVh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vshufeh) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.b=vshuff(Vu32.b) + C Intrinsic Prototype: HVX_Vector Q6_Vb_vshuff_Vb(HVX_Vector Vu) + Instruction Type: CVI_VP + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vb_vshuff_Vb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vshuffb) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.b=vshuffe(Vu32.b,Vv32.b) + C Intrinsic Prototype: HVX_Vector Q6_Vb_vshuffe_VbVb(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vb_vshuffe_VbVb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vshuffeb) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.h=vshuff(Vu32.h) + C Intrinsic Prototype: HVX_Vector Q6_Vh_vshuff_Vh(HVX_Vector Vu) + Instruction Type: CVI_VP + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vh_vshuff_Vh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vshuffh) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.b=vshuffo(Vu32.b,Vv32.b) + C Intrinsic Prototype: HVX_Vector Q6_Vb_vshuffo_VbVb(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vb_vshuffo_VbVb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vshuffob) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32=vshuff(Vu32,Vv32,Rt8) + C Intrinsic Prototype: HVX_VectorPair Q6_W_vshuff_VVR(HVX_Vector Vu, HVX_Vector Vv, Word32 Rt) + Instruction Type: CVI_VP_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_W_vshuff_VVR __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vshuffvdd) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.b=vshuffoe(Vu32.b,Vv32.b) + C Intrinsic Prototype: HVX_VectorPair Q6_Wb_vshuffoe_VbVb(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA_DV + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Wb_vshuffoe_VbVb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vshufoeb) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.h=vshuffoe(Vu32.h,Vv32.h) + C Intrinsic Prototype: HVX_VectorPair Q6_Wh_vshuffoe_VhVh(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA_DV + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Wh_vshuffoe_VhVh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vshufoeh) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.h=vshuffo(Vu32.h,Vv32.h) + C Intrinsic Prototype: HVX_Vector Q6_Vh_vshuffo_VhVh(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vh_vshuffo_VhVh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vshufoh) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.b=vsub(Vu32.b,Vv32.b) + C Intrinsic Prototype: HVX_Vector Q6_Vb_vsub_VbVb(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vb_vsub_VbVb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vsubb) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.b=vsub(Vuu32.b,Vvv32.b) + C Intrinsic Prototype: HVX_VectorPair Q6_Wb_vsub_WbWb(HVX_VectorPair Vuu, HVX_VectorPair Vvv) + Instruction Type: CVI_VA_DV + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Wb_vsub_WbWb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vsubb_dv) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: if (!Qv4) Vx32.b-=Vu32.b + C Intrinsic Prototype: HVX_Vector Q6_Vb_condnac_QnVbVb(HVX_VectorPred Qv, HVX_Vector Vx, HVX_Vector Vu) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vb_condnac_QnVbVb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vsubbnq) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: if (Qv4) Vx32.b-=Vu32.b + C Intrinsic Prototype: HVX_Vector Q6_Vb_condnac_QVbVb(HVX_VectorPred Qv, HVX_Vector Vx, HVX_Vector Vu) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vb_condnac_QVbVb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vsubbq) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.h=vsub(Vu32.h,Vv32.h) + C Intrinsic Prototype: HVX_Vector Q6_Vh_vsub_VhVh(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vh_vsub_VhVh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vsubh) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.h=vsub(Vuu32.h,Vvv32.h) + C Intrinsic Prototype: HVX_VectorPair Q6_Wh_vsub_WhWh(HVX_VectorPair Vuu, HVX_VectorPair Vvv) + Instruction Type: CVI_VA_DV + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Wh_vsub_WhWh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vsubh_dv) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: if (!Qv4) Vx32.h-=Vu32.h + C Intrinsic Prototype: HVX_Vector Q6_Vh_condnac_QnVhVh(HVX_VectorPred Qv, HVX_Vector Vx, HVX_Vector Vu) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vh_condnac_QnVhVh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vsubhnq) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: if (Qv4) Vx32.h-=Vu32.h + C Intrinsic Prototype: HVX_Vector Q6_Vh_condnac_QVhVh(HVX_VectorPred Qv, HVX_Vector Vx, HVX_Vector Vu) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vh_condnac_QVhVh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vsubhq) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.h=vsub(Vu32.h,Vv32.h):sat + C Intrinsic Prototype: HVX_Vector Q6_Vh_vsub_VhVh_sat(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vh_vsub_VhVh_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vsubhsat) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.h=vsub(Vuu32.h,Vvv32.h):sat + C Intrinsic Prototype: HVX_VectorPair Q6_Wh_vsub_WhWh_sat(HVX_VectorPair Vuu, HVX_VectorPair Vvv) + Instruction Type: CVI_VA_DV + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Wh_vsub_WhWh_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vsubhsat_dv) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.w=vsub(Vu32.h,Vv32.h) + C Intrinsic Prototype: HVX_VectorPair Q6_Ww_vsub_VhVh(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Ww_vsub_VhVh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vsubhw) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.h=vsub(Vu32.ub,Vv32.ub) + C Intrinsic Prototype: HVX_VectorPair Q6_Wh_vsub_VubVub(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Wh_vsub_VubVub __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vsububh) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.ub=vsub(Vu32.ub,Vv32.ub):sat + C Intrinsic Prototype: HVX_Vector Q6_Vub_vsub_VubVub_sat(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vub_vsub_VubVub_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vsububsat) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.ub=vsub(Vuu32.ub,Vvv32.ub):sat + C Intrinsic Prototype: HVX_VectorPair Q6_Wub_vsub_WubWub_sat(HVX_VectorPair Vuu, HVX_VectorPair Vvv) + Instruction Type: CVI_VA_DV + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Wub_vsub_WubWub_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vsububsat_dv) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.uh=vsub(Vu32.uh,Vv32.uh):sat + C Intrinsic Prototype: HVX_Vector Q6_Vuh_vsub_VuhVuh_sat(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vuh_vsub_VuhVuh_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vsubuhsat) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.uh=vsub(Vuu32.uh,Vvv32.uh):sat + C Intrinsic Prototype: HVX_VectorPair Q6_Wuh_vsub_WuhWuh_sat(HVX_VectorPair Vuu, HVX_VectorPair Vvv) + Instruction Type: CVI_VA_DV + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Wuh_vsub_WuhWuh_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vsubuhsat_dv) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.w=vsub(Vu32.uh,Vv32.uh) + C Intrinsic Prototype: HVX_VectorPair Q6_Ww_vsub_VuhVuh(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Ww_vsub_VuhVuh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vsubuhw) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.w=vsub(Vu32.w,Vv32.w) + C Intrinsic Prototype: HVX_Vector Q6_Vw_vsub_VwVw(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vw_vsub_VwVw __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vsubw) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.w=vsub(Vuu32.w,Vvv32.w) + C Intrinsic Prototype: HVX_VectorPair Q6_Ww_vsub_WwWw(HVX_VectorPair Vuu, HVX_VectorPair Vvv) + Instruction Type: CVI_VA_DV + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Ww_vsub_WwWw __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vsubw_dv) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: if (!Qv4) Vx32.w-=Vu32.w + C Intrinsic Prototype: HVX_Vector Q6_Vw_condnac_QnVwVw(HVX_VectorPred Qv, HVX_Vector Vx, HVX_Vector Vu) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vw_condnac_QnVwVw __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vsubwnq) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: if (Qv4) Vx32.w-=Vu32.w + C Intrinsic Prototype: HVX_Vector Q6_Vw_condnac_QVwVw(HVX_VectorPred Qv, HVX_Vector Vx, HVX_Vector Vu) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vw_condnac_QVwVw __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vsubwq) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32.w=vsub(Vu32.w,Vv32.w):sat + C Intrinsic Prototype: HVX_Vector Q6_Vw_vsub_VwVw_sat(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vw_vsub_VwVw_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vsubwsat) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.w=vsub(Vuu32.w,Vvv32.w):sat + C Intrinsic Prototype: HVX_VectorPair Q6_Ww_vsub_WwWw_sat(HVX_VectorPair Vuu, HVX_VectorPair Vvv) + Instruction Type: CVI_VA_DV + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Ww_vsub_WwWw_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vsubwsat_dv) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32=vswap(Qt4,Vu32,Vv32) + C Intrinsic Prototype: HVX_VectorPair Q6_W_vswap_QVV(HVX_VectorPred Qt, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA_DV + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_W_vswap_QVV __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vswap) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.h=vtmpy(Vuu32.b,Rt32.b) + C Intrinsic Prototype: HVX_VectorPair Q6_Wh_vtmpy_WbRb(HVX_VectorPair Vuu, Word32 Rt) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Wh_vtmpy_WbRb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vtmpyb) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vxx32.h+=vtmpy(Vuu32.b,Rt32.b) + C Intrinsic Prototype: HVX_VectorPair Q6_Wh_vtmpyacc_WhWbRb(HVX_VectorPair Vxx, HVX_VectorPair Vuu, Word32 Rt) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Wh_vtmpyacc_WhWbRb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vtmpyb_acc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.h=vtmpy(Vuu32.ub,Rt32.b) + C Intrinsic Prototype: HVX_VectorPair Q6_Wh_vtmpy_WubRb(HVX_VectorPair Vuu, Word32 Rt) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Wh_vtmpy_WubRb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vtmpybus) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vxx32.h+=vtmpy(Vuu32.ub,Rt32.b) + C Intrinsic Prototype: HVX_VectorPair Q6_Wh_vtmpyacc_WhWubRb(HVX_VectorPair Vxx, HVX_VectorPair Vuu, Word32 Rt) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Wh_vtmpyacc_WhWubRb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vtmpybus_acc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.w=vtmpy(Vuu32.h,Rt32.b) + C Intrinsic Prototype: HVX_VectorPair Q6_Ww_vtmpy_WhRb(HVX_VectorPair Vuu, Word32 Rt) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Ww_vtmpy_WhRb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vtmpyhb) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vxx32.w+=vtmpy(Vuu32.h,Rt32.b) + C Intrinsic Prototype: HVX_VectorPair Q6_Ww_vtmpyacc_WwWhRb(HVX_VectorPair Vxx, HVX_VectorPair Vuu, Word32 Rt) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Ww_vtmpyacc_WwWhRb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vtmpyhb_acc) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.h=vunpack(Vu32.b) + C Intrinsic Prototype: HVX_VectorPair Q6_Wh_vunpack_Vb(HVX_Vector Vu) + Instruction Type: CVI_VP_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Wh_vunpack_Vb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vunpackb) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.w=vunpack(Vu32.h) + C Intrinsic Prototype: HVX_VectorPair Q6_Ww_vunpack_Vh(HVX_Vector Vu) + Instruction Type: CVI_VP_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Ww_vunpack_Vh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vunpackh) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vxx32.h|=vunpacko(Vu32.b) + C Intrinsic Prototype: HVX_VectorPair Q6_Wh_vunpackoor_WhVb(HVX_VectorPair Vxx, HVX_Vector Vu) + Instruction Type: CVI_VP_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Wh_vunpackoor_WhVb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vunpackob) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vxx32.w|=vunpacko(Vu32.h) + C Intrinsic Prototype: HVX_VectorPair Q6_Ww_vunpackoor_WwVh(HVX_VectorPair Vxx, HVX_Vector Vu) + Instruction Type: CVI_VP_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Ww_vunpackoor_WwVh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vunpackoh) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.uh=vunpack(Vu32.ub) + C Intrinsic Prototype: HVX_VectorPair Q6_Wuh_vunpack_Vub(HVX_Vector Vu) + Instruction Type: CVI_VP_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Wuh_vunpack_Vub __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vunpackub) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.uw=vunpack(Vu32.uh) + C Intrinsic Prototype: HVX_VectorPair Q6_Wuw_vunpack_Vuh(HVX_Vector Vu) + Instruction Type: CVI_VP_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Wuw_vunpack_Vuh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vunpackuh) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vd32=vxor(Vu32,Vv32) + C Intrinsic Prototype: HVX_Vector Q6_V_vxor_VV(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_V_vxor_VV __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vxor) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.uh=vzxt(Vu32.ub) + C Intrinsic Prototype: HVX_VectorPair Q6_Wuh_vzxt_Vub(HVX_Vector Vu) + Instruction Type: CVI_VA_DV + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Wuh_vzxt_Vub __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vzb) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 60 +/* ========================================================================== + Assembly Syntax: Vdd32.uw=vzxt(Vu32.uh) + C Intrinsic Prototype: HVX_VectorPair Q6_Wuw_vzxt_Vuh(HVX_Vector Vu) + Instruction Type: CVI_VA_DV + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Wuw_vzxt_Vuh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vzh) +#endif /* __HEXAGON_ARCH___ >= 60 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Vd32.b=vsplat(Rt32) + C Intrinsic Prototype: HVX_Vector Q6_Vb_vsplat_R(Word32 Rt) + Instruction Type: CVI_VX_LATE + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vb_vsplat_R __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_lvsplatb) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Vd32.h=vsplat(Rt32) + C Intrinsic Prototype: HVX_Vector Q6_Vh_vsplat_R(Word32 Rt) + Instruction Type: CVI_VX_LATE + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vh_vsplat_R __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_lvsplath) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Qd4=vsetq2(Rt32) + C Intrinsic Prototype: HVX_VectorPred Q6_Q_vsetq2_R(Word32 Rt) + Instruction Type: CVI_VP + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Q_vsetq2_R __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_pred_scalar2v2) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Qd4.b=vshuffe(Qs4.h,Qt4.h) + C Intrinsic Prototype: HVX_VectorPred Q6_Qb_vshuffe_QhQh(HVX_VectorPred Qs, HVX_VectorPred Qt) + Instruction Type: CVI_VA_DV + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Qb_vshuffe_QhQh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_shuffeqh) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Qd4.h=vshuffe(Qs4.w,Qt4.w) + C Intrinsic Prototype: HVX_VectorPred Q6_Qh_vshuffe_QwQw(HVX_VectorPred Qs, HVX_VectorPred Qt) + Instruction Type: CVI_VA_DV + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Qh_vshuffe_QwQw __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_shuffeqw) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Vd32.b=vadd(Vu32.b,Vv32.b):sat + C Intrinsic Prototype: HVX_Vector Q6_Vb_vadd_VbVb_sat(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vb_vadd_VbVb_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vaddbsat) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Vdd32.b=vadd(Vuu32.b,Vvv32.b):sat + C Intrinsic Prototype: HVX_VectorPair Q6_Wb_vadd_WbWb_sat(HVX_VectorPair Vuu, HVX_VectorPair Vvv) + Instruction Type: CVI_VA_DV + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Wb_vadd_WbWb_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vaddbsat_dv) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Vd32.w=vadd(Vu32.w,Vv32.w,Qx4):carry + C Intrinsic Prototype: HVX_Vector Q6_Vw_vadd_VwVwQ_carry(HVX_Vector Vu, HVX_Vector Vv, HVX_VectorPred* Qx) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vw_vadd_VwVwQ_carry __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vaddcarry) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Vd32.h=vadd(vclb(Vu32.h),Vv32.h) + C Intrinsic Prototype: HVX_Vector Q6_Vh_vadd_vclb_VhVh(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vh_vadd_vclb_VhVh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vaddclbh) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Vd32.w=vadd(vclb(Vu32.w),Vv32.w) + C Intrinsic Prototype: HVX_Vector Q6_Vw_vadd_vclb_VwVw(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vw_vadd_vclb_VwVw __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vaddclbw) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Vxx32.w+=vadd(Vu32.h,Vv32.h) + C Intrinsic Prototype: HVX_VectorPair Q6_Ww_vaddacc_WwVhVh(HVX_VectorPair Vxx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Ww_vaddacc_WwVhVh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vaddhw_acc) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Vxx32.h+=vadd(Vu32.ub,Vv32.ub) + C Intrinsic Prototype: HVX_VectorPair Q6_Wh_vaddacc_WhVubVub(HVX_VectorPair Vxx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Wh_vaddacc_WhVubVub __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vaddubh_acc) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Vd32.ub=vadd(Vu32.ub,Vv32.b):sat + C Intrinsic Prototype: HVX_Vector Q6_Vub_vadd_VubVb_sat(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vub_vadd_VubVb_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vaddububb_sat) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Vxx32.w+=vadd(Vu32.uh,Vv32.uh) + C Intrinsic Prototype: HVX_VectorPair Q6_Ww_vaddacc_WwVuhVuh(HVX_VectorPair Vxx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Ww_vaddacc_WwVuhVuh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vadduhw_acc) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Vd32.uw=vadd(Vu32.uw,Vv32.uw):sat + C Intrinsic Prototype: HVX_Vector Q6_Vuw_vadd_VuwVuw_sat(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vuw_vadd_VuwVuw_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vadduwsat) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Vdd32.uw=vadd(Vuu32.uw,Vvv32.uw):sat + C Intrinsic Prototype: HVX_VectorPair Q6_Wuw_vadd_WuwWuw_sat(HVX_VectorPair Vuu, HVX_VectorPair Vvv) + Instruction Type: CVI_VA_DV + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Wuw_vadd_WuwWuw_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vadduwsat_dv) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Vd32=vand(!Qu4,Rt32) + C Intrinsic Prototype: HVX_Vector Q6_V_vand_QnR(HVX_VectorPred Qu, Word32 Rt) + Instruction Type: CVI_VX_LATE + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_V_vand_QnR __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vandnqrt) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Vx32|=vand(!Qu4,Rt32) + C Intrinsic Prototype: HVX_Vector Q6_V_vandor_VQnR(HVX_Vector Vx, HVX_VectorPred Qu, Word32 Rt) + Instruction Type: CVI_VX_LATE + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_V_vandor_VQnR __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vandnqrt_acc) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Vd32=vand(!Qv4,Vu32) + C Intrinsic Prototype: HVX_Vector Q6_V_vand_QnV(HVX_VectorPred Qv, HVX_Vector Vu) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_V_vand_QnV __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vandvnqv) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Vd32=vand(Qv4,Vu32) + C Intrinsic Prototype: HVX_Vector Q6_V_vand_QV(HVX_VectorPred Qv, HVX_Vector Vu) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_V_vand_QV __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vandvqv) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Vd32.b=vasr(Vu32.h,Vv32.h,Rt8):sat + C Intrinsic Prototype: HVX_Vector Q6_Vb_vasr_VhVhR_sat(HVX_Vector Vu, HVX_Vector Vv, Word32 Rt) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vb_vasr_VhVhR_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vasrhbsat) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Vd32.uh=vasr(Vu32.uw,Vv32.uw,Rt8):rnd:sat + C Intrinsic Prototype: HVX_Vector Q6_Vuh_vasr_VuwVuwR_rnd_sat(HVX_Vector Vu, HVX_Vector Vv, Word32 Rt) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vuh_vasr_VuwVuwR_rnd_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vasruwuhrndsat) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Vd32.uh=vasr(Vu32.w,Vv32.w,Rt8):rnd:sat + C Intrinsic Prototype: HVX_Vector Q6_Vuh_vasr_VwVwR_rnd_sat(HVX_Vector Vu, HVX_Vector Vv, Word32 Rt) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vuh_vasr_VwVwR_rnd_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vasrwuhrndsat) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Vd32.ub=vlsr(Vu32.ub,Rt32) + C Intrinsic Prototype: HVX_Vector Q6_Vub_vlsr_VubR(HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vub_vlsr_VubR __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vlsrb) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Vd32.b=vlut32(Vu32.b,Vv32.b,Rt8):nomatch + C Intrinsic Prototype: HVX_Vector Q6_Vb_vlut32_VbVbR_nomatch(HVX_Vector Vu, HVX_Vector Vv, Word32 Rt) + Instruction Type: CVI_VP + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vb_vlut32_VbVbR_nomatch __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vlutvvb_nm) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Vx32.b|=vlut32(Vu32.b,Vv32.b,#u3) + C Intrinsic Prototype: HVX_Vector Q6_Vb_vlut32or_VbVbVbI(HVX_Vector Vx, HVX_Vector Vu, HVX_Vector Vv, Word32 Iu3) + Instruction Type: CVI_VP_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vb_vlut32or_VbVbVbI __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vlutvvb_oracci) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Vd32.b=vlut32(Vu32.b,Vv32.b,#u3) + C Intrinsic Prototype: HVX_Vector Q6_Vb_vlut32_VbVbI(HVX_Vector Vu, HVX_Vector Vv, Word32 Iu3) + Instruction Type: CVI_VP + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vb_vlut32_VbVbI __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vlutvvbi) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Vdd32.h=vlut16(Vu32.b,Vv32.h,Rt8):nomatch + C Intrinsic Prototype: HVX_VectorPair Q6_Wh_vlut16_VbVhR_nomatch(HVX_Vector Vu, HVX_Vector Vv, Word32 Rt) + Instruction Type: CVI_VP_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Wh_vlut16_VbVhR_nomatch __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vlutvwh_nm) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Vxx32.h|=vlut16(Vu32.b,Vv32.h,#u3) + C Intrinsic Prototype: HVX_VectorPair Q6_Wh_vlut16or_WhVbVhI(HVX_VectorPair Vxx, HVX_Vector Vu, HVX_Vector Vv, Word32 Iu3) + Instruction Type: CVI_VP_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Wh_vlut16or_WhVbVhI __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vlutvwh_oracci) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Vdd32.h=vlut16(Vu32.b,Vv32.h,#u3) + C Intrinsic Prototype: HVX_VectorPair Q6_Wh_vlut16_VbVhI(HVX_Vector Vu, HVX_Vector Vv, Word32 Iu3) + Instruction Type: CVI_VP_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Wh_vlut16_VbVhI __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vlutvwhi) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Vd32.b=vmax(Vu32.b,Vv32.b) + C Intrinsic Prototype: HVX_Vector Q6_Vb_vmax_VbVb(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vb_vmax_VbVb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmaxb) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Vd32.b=vmin(Vu32.b,Vv32.b) + C Intrinsic Prototype: HVX_Vector Q6_Vb_vmin_VbVb(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vb_vmin_VbVb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vminb) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Vdd32.w=vmpa(Vuu32.uh,Rt32.b) + C Intrinsic Prototype: HVX_VectorPair Q6_Ww_vmpa_WuhRb(HVX_VectorPair Vuu, Word32 Rt) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Ww_vmpa_WuhRb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpauhb) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Vxx32.w+=vmpa(Vuu32.uh,Rt32.b) + C Intrinsic Prototype: HVX_VectorPair Q6_Ww_vmpaacc_WwWuhRb(HVX_VectorPair Vxx, HVX_VectorPair Vuu, Word32 Rt) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Ww_vmpaacc_WwWuhRb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpauhb_acc) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Vdd32=vmpye(Vu32.w,Vv32.uh) + C Intrinsic Prototype: HVX_VectorPair Q6_W_vmpye_VwVuh(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_W_vmpye_VwVuh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpyewuh_64) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Vd32.w=vmpyi(Vu32.w,Rt32.ub) + C Intrinsic Prototype: HVX_Vector Q6_Vw_vmpyi_VwRub(HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VX + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vw_vmpyi_VwRub __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpyiwub) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Vx32.w+=vmpyi(Vu32.w,Rt32.ub) + C Intrinsic Prototype: HVX_Vector Q6_Vw_vmpyiacc_VwVwRub(HVX_Vector Vx, HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VX + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vw_vmpyiacc_VwVwRub __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpyiwub_acc) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Vxx32+=vmpyo(Vu32.w,Vv32.h) + C Intrinsic Prototype: HVX_VectorPair Q6_W_vmpyoacc_WVwVh(HVX_VectorPair Vxx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_W_vmpyoacc_WVwVh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpyowh_64_acc) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Vd32.ub=vround(Vu32.uh,Vv32.uh):sat + C Intrinsic Prototype: HVX_Vector Q6_Vub_vround_VuhVuh_sat(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vub_vround_VuhVuh_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vrounduhub) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Vd32.uh=vround(Vu32.uw,Vv32.uw):sat + C Intrinsic Prototype: HVX_Vector Q6_Vuh_vround_VuwVuw_sat(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vuh_vround_VuwVuw_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vrounduwuh) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Vd32.uh=vsat(Vu32.uw,Vv32.uw) + C Intrinsic Prototype: HVX_Vector Q6_Vuh_vsat_VuwVuw(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vuh_vsat_VuwVuw __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vsatuwuh) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Vd32.b=vsub(Vu32.b,Vv32.b):sat + C Intrinsic Prototype: HVX_Vector Q6_Vb_vsub_VbVb_sat(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vb_vsub_VbVb_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vsubbsat) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Vdd32.b=vsub(Vuu32.b,Vvv32.b):sat + C Intrinsic Prototype: HVX_VectorPair Q6_Wb_vsub_WbWb_sat(HVX_VectorPair Vuu, HVX_VectorPair Vvv) + Instruction Type: CVI_VA_DV + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Wb_vsub_WbWb_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vsubbsat_dv) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Vd32.w=vsub(Vu32.w,Vv32.w,Qx4):carry + C Intrinsic Prototype: HVX_Vector Q6_Vw_vsub_VwVwQ_carry(HVX_Vector Vu, HVX_Vector Vv, HVX_VectorPred* Qx) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vw_vsub_VwVwQ_carry __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vsubcarry) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Vd32.ub=vsub(Vu32.ub,Vv32.b):sat + C Intrinsic Prototype: HVX_Vector Q6_Vub_vsub_VubVb_sat(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vub_vsub_VubVb_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vsubububb_sat) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Vd32.uw=vsub(Vu32.uw,Vv32.uw):sat + C Intrinsic Prototype: HVX_Vector Q6_Vuw_vsub_VuwVuw_sat(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vuw_vsub_VuwVuw_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vsubuwsat) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 62 +/* ========================================================================== + Assembly Syntax: Vdd32.uw=vsub(Vuu32.uw,Vvv32.uw):sat + C Intrinsic Prototype: HVX_VectorPair Q6_Wuw_vsub_WuwWuw_sat(HVX_VectorPair Vuu, HVX_VectorPair Vvv) + Instruction Type: CVI_VA_DV + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Wuw_vsub_WuwWuw_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vsubuwsat_dv) +#endif /* __HEXAGON_ARCH___ >= 62 */ + +#if __HVX_ARCH__ >= 65 +/* ========================================================================== + Assembly Syntax: Vd32.b=vabs(Vu32.b) + C Intrinsic Prototype: HVX_Vector Q6_Vb_vabs_Vb(HVX_Vector Vu) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vb_vabs_Vb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vabsb) +#endif /* __HEXAGON_ARCH___ >= 65 */ + +#if __HVX_ARCH__ >= 65 +/* ========================================================================== + Assembly Syntax: Vd32.b=vabs(Vu32.b):sat + C Intrinsic Prototype: HVX_Vector Q6_Vb_vabs_Vb_sat(HVX_Vector Vu) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vb_vabs_Vb_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vabsb_sat) +#endif /* __HEXAGON_ARCH___ >= 65 */ + +#if __HVX_ARCH__ >= 65 +/* ========================================================================== + Assembly Syntax: Vx32.h+=vasl(Vu32.h,Rt32) + C Intrinsic Prototype: HVX_Vector Q6_Vh_vaslacc_VhVhR(HVX_Vector Vx, HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vh_vaslacc_VhVhR __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vaslh_acc) +#endif /* __HEXAGON_ARCH___ >= 65 */ + +#if __HVX_ARCH__ >= 65 +/* ========================================================================== + Assembly Syntax: Vx32.h+=vasr(Vu32.h,Rt32) + C Intrinsic Prototype: HVX_Vector Q6_Vh_vasracc_VhVhR(HVX_Vector Vx, HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vh_vasracc_VhVhR __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vasrh_acc) +#endif /* __HEXAGON_ARCH___ >= 65 */ + +#if __HVX_ARCH__ >= 65 +/* ========================================================================== + Assembly Syntax: Vd32.ub=vasr(Vu32.uh,Vv32.uh,Rt8):rnd:sat + C Intrinsic Prototype: HVX_Vector Q6_Vub_vasr_VuhVuhR_rnd_sat(HVX_Vector Vu, HVX_Vector Vv, Word32 Rt) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vub_vasr_VuhVuhR_rnd_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vasruhubrndsat) +#endif /* __HEXAGON_ARCH___ >= 65 */ + +#if __HVX_ARCH__ >= 65 +/* ========================================================================== + Assembly Syntax: Vd32.ub=vasr(Vu32.uh,Vv32.uh,Rt8):sat + C Intrinsic Prototype: HVX_Vector Q6_Vub_vasr_VuhVuhR_sat(HVX_Vector Vu, HVX_Vector Vv, Word32 Rt) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vub_vasr_VuhVuhR_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vasruhubsat) +#endif /* __HEXAGON_ARCH___ >= 65 */ + +#if __HVX_ARCH__ >= 65 +/* ========================================================================== + Assembly Syntax: Vd32.uh=vasr(Vu32.uw,Vv32.uw,Rt8):sat + C Intrinsic Prototype: HVX_Vector Q6_Vuh_vasr_VuwVuwR_sat(HVX_Vector Vu, HVX_Vector Vv, Word32 Rt) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vuh_vasr_VuwVuwR_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vasruwuhsat) +#endif /* __HEXAGON_ARCH___ >= 65 */ + +#if __HVX_ARCH__ >= 65 +/* ========================================================================== + Assembly Syntax: Vd32.b=vavg(Vu32.b,Vv32.b) + C Intrinsic Prototype: HVX_Vector Q6_Vb_vavg_VbVb(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vb_vavg_VbVb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vavgb) +#endif /* __HEXAGON_ARCH___ >= 65 */ + +#if __HVX_ARCH__ >= 65 +/* ========================================================================== + Assembly Syntax: Vd32.b=vavg(Vu32.b,Vv32.b):rnd + C Intrinsic Prototype: HVX_Vector Q6_Vb_vavg_VbVb_rnd(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vb_vavg_VbVb_rnd __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vavgbrnd) +#endif /* __HEXAGON_ARCH___ >= 65 */ + +#if __HVX_ARCH__ >= 65 +/* ========================================================================== + Assembly Syntax: Vd32.uw=vavg(Vu32.uw,Vv32.uw) + C Intrinsic Prototype: HVX_Vector Q6_Vuw_vavg_VuwVuw(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vuw_vavg_VuwVuw __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vavguw) +#endif /* __HEXAGON_ARCH___ >= 65 */ + +#if __HVX_ARCH__ >= 65 +/* ========================================================================== + Assembly Syntax: Vd32.uw=vavg(Vu32.uw,Vv32.uw):rnd + C Intrinsic Prototype: HVX_Vector Q6_Vuw_vavg_VuwVuw_rnd(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vuw_vavg_VuwVuw_rnd __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vavguwrnd) +#endif /* __HEXAGON_ARCH___ >= 65 */ + +#if __HVX_ARCH__ >= 65 +/* ========================================================================== + Assembly Syntax: Vdd32=#0 + C Intrinsic Prototype: HVX_VectorPair Q6_W_vzero() + Instruction Type: MAPPING + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_W_vzero __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vdd0) +#endif /* __HEXAGON_ARCH___ >= 65 */ + +#if __HVX_ARCH__ >= 65 +/* ========================================================================== + Assembly Syntax: vtmp.h=vgather(Rt32,Mu2,Vv32.h).h + C Intrinsic Prototype: void Q6_vgather_ARMVh(HVX_Vector* Rs, Word32 Rt, Word32 Mu, HVX_Vector Vv) + Instruction Type: CVI_GATHER + Execution Slots: SLOT01 + ========================================================================== */ + +#define Q6_vgather_ARMVh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vgathermh) +#endif /* __HEXAGON_ARCH___ >= 65 */ + +#if __HVX_ARCH__ >= 65 +/* ========================================================================== + Assembly Syntax: if (Qs4) vtmp.h=vgather(Rt32,Mu2,Vv32.h).h + C Intrinsic Prototype: void Q6_vgather_AQRMVh(HVX_Vector* Rs, HVX_VectorPred Qs, Word32 Rt, Word32 Mu, HVX_Vector Vv) + Instruction Type: CVI_GATHER + Execution Slots: SLOT01 + ========================================================================== */ + +#define Q6_vgather_AQRMVh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vgathermhq) +#endif /* __HEXAGON_ARCH___ >= 65 */ + +#if __HVX_ARCH__ >= 65 +/* ========================================================================== + Assembly Syntax: vtmp.h=vgather(Rt32,Mu2,Vvv32.w).h + C Intrinsic Prototype: void Q6_vgather_ARMWw(HVX_Vector* Rs, Word32 Rt, Word32 Mu, HVX_VectorPair Vvv) + Instruction Type: CVI_GATHER_DV + Execution Slots: SLOT01 + ========================================================================== */ + +#define Q6_vgather_ARMWw __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vgathermhw) +#endif /* __HEXAGON_ARCH___ >= 65 */ + +#if __HVX_ARCH__ >= 65 +/* ========================================================================== + Assembly Syntax: if (Qs4) vtmp.h=vgather(Rt32,Mu2,Vvv32.w).h + C Intrinsic Prototype: void Q6_vgather_AQRMWw(HVX_Vector* Rs, HVX_VectorPred Qs, Word32 Rt, Word32 Mu, HVX_VectorPair Vvv) + Instruction Type: CVI_GATHER_DV + Execution Slots: SLOT01 + ========================================================================== */ + +#define Q6_vgather_AQRMWw __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vgathermhwq) +#endif /* __HEXAGON_ARCH___ >= 65 */ + +#if __HVX_ARCH__ >= 65 +/* ========================================================================== + Assembly Syntax: vtmp.w=vgather(Rt32,Mu2,Vv32.w).w + C Intrinsic Prototype: void Q6_vgather_ARMVw(HVX_Vector* Rs, Word32 Rt, Word32 Mu, HVX_Vector Vv) + Instruction Type: CVI_GATHER + Execution Slots: SLOT01 + ========================================================================== */ + +#define Q6_vgather_ARMVw __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vgathermw) +#endif /* __HEXAGON_ARCH___ >= 65 */ + +#if __HVX_ARCH__ >= 65 +/* ========================================================================== + Assembly Syntax: if (Qs4) vtmp.w=vgather(Rt32,Mu2,Vv32.w).w + C Intrinsic Prototype: void Q6_vgather_AQRMVw(HVX_Vector* Rs, HVX_VectorPred Qs, Word32 Rt, Word32 Mu, HVX_Vector Vv) + Instruction Type: CVI_GATHER + Execution Slots: SLOT01 + ========================================================================== */ + +#define Q6_vgather_AQRMVw __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vgathermwq) +#endif /* __HEXAGON_ARCH___ >= 65 */ + +#if __HVX_ARCH__ >= 65 +/* ========================================================================== + Assembly Syntax: Vd32.h=vlut4(Vu32.uh,Rtt32.h) + C Intrinsic Prototype: HVX_Vector Q6_Vh_vlut4_VuhPh(HVX_Vector Vu, Word64 Rtt) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT2 + ========================================================================== */ + +#define Q6_Vh_vlut4_VuhPh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vlut4) +#endif /* __HEXAGON_ARCH___ >= 65 */ + +#if __HVX_ARCH__ >= 65 +/* ========================================================================== + Assembly Syntax: Vdd32.h=vmpa(Vuu32.ub,Rt32.ub) + C Intrinsic Prototype: HVX_VectorPair Q6_Wh_vmpa_WubRub(HVX_VectorPair Vuu, Word32 Rt) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Wh_vmpa_WubRub __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpabuu) +#endif /* __HEXAGON_ARCH___ >= 65 */ + +#if __HVX_ARCH__ >= 65 +/* ========================================================================== + Assembly Syntax: Vxx32.h+=vmpa(Vuu32.ub,Rt32.ub) + C Intrinsic Prototype: HVX_VectorPair Q6_Wh_vmpaacc_WhWubRub(HVX_VectorPair Vxx, HVX_VectorPair Vuu, Word32 Rt) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Wh_vmpaacc_WhWubRub __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpabuu_acc) +#endif /* __HEXAGON_ARCH___ >= 65 */ + +#if __HVX_ARCH__ >= 65 +/* ========================================================================== + Assembly Syntax: Vx32.h=vmpa(Vx32.h,Vu32.h,Rtt32.h):sat + C Intrinsic Prototype: HVX_Vector Q6_Vh_vmpa_VhVhVhPh_sat(HVX_Vector Vx, HVX_Vector Vu, Word64 Rtt) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT2 + ========================================================================== */ + +#define Q6_Vh_vmpa_VhVhVhPh_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpahhsat) +#endif /* __HEXAGON_ARCH___ >= 65 */ + +#if __HVX_ARCH__ >= 65 +/* ========================================================================== + Assembly Syntax: Vx32.h=vmpa(Vx32.h,Vu32.uh,Rtt32.uh):sat + C Intrinsic Prototype: HVX_Vector Q6_Vh_vmpa_VhVhVuhPuh_sat(HVX_Vector Vx, HVX_Vector Vu, Word64 Rtt) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT2 + ========================================================================== */ + +#define Q6_Vh_vmpa_VhVhVuhPuh_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpauhuhsat) +#endif /* __HEXAGON_ARCH___ >= 65 */ + +#if __HVX_ARCH__ >= 65 +/* ========================================================================== + Assembly Syntax: Vx32.h=vmps(Vx32.h,Vu32.uh,Rtt32.uh):sat + C Intrinsic Prototype: HVX_Vector Q6_Vh_vmps_VhVhVuhPuh_sat(HVX_Vector Vx, HVX_Vector Vu, Word64 Rtt) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT2 + ========================================================================== */ + +#define Q6_Vh_vmps_VhVhVuhPuh_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpsuhuhsat) +#endif /* __HEXAGON_ARCH___ >= 65 */ + +#if __HVX_ARCH__ >= 65 +/* ========================================================================== + Assembly Syntax: Vxx32.w+=vmpy(Vu32.h,Rt32.h) + C Intrinsic Prototype: HVX_VectorPair Q6_Ww_vmpyacc_WwVhRh(HVX_VectorPair Vxx, HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Ww_vmpyacc_WwVhRh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpyh_acc) +#endif /* __HEXAGON_ARCH___ >= 65 */ + +#if __HVX_ARCH__ >= 65 +/* ========================================================================== + Assembly Syntax: Vd32.uw=vmpye(Vu32.uh,Rt32.uh) + C Intrinsic Prototype: HVX_Vector Q6_Vuw_vmpye_VuhRuh(HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VX + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vuw_vmpye_VuhRuh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpyuhe) +#endif /* __HEXAGON_ARCH___ >= 65 */ + +#if __HVX_ARCH__ >= 65 +/* ========================================================================== + Assembly Syntax: Vx32.uw+=vmpye(Vu32.uh,Rt32.uh) + C Intrinsic Prototype: HVX_Vector Q6_Vuw_vmpyeacc_VuwVuhRuh(HVX_Vector Vx, HVX_Vector Vu, Word32 Rt) + Instruction Type: CVI_VX + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Vuw_vmpyeacc_VuwVuhRuh __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vmpyuhe_acc) +#endif /* __HEXAGON_ARCH___ >= 65 */ + +#if __HVX_ARCH__ >= 65 +/* ========================================================================== + Assembly Syntax: Vd32.b=vnavg(Vu32.b,Vv32.b) + C Intrinsic Prototype: HVX_Vector Q6_Vb_vnavg_VbVb(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vb_vnavg_VbVb __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vnavgb) +#endif /* __HEXAGON_ARCH___ >= 65 */ + +#if __HVX_ARCH__ >= 65 +/* ========================================================================== + Assembly Syntax: Vd32.b=prefixsum(Qv4) + C Intrinsic Prototype: HVX_Vector Q6_Vb_prefixsum_Q(HVX_VectorPred Qv) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vb_prefixsum_Q __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vprefixqb) +#endif /* __HEXAGON_ARCH___ >= 65 */ + +#if __HVX_ARCH__ >= 65 +/* ========================================================================== + Assembly Syntax: Vd32.h=prefixsum(Qv4) + C Intrinsic Prototype: HVX_Vector Q6_Vh_prefixsum_Q(HVX_VectorPred Qv) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vh_prefixsum_Q __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vprefixqh) +#endif /* __HEXAGON_ARCH___ >= 65 */ + +#if __HVX_ARCH__ >= 65 +/* ========================================================================== + Assembly Syntax: Vd32.w=prefixsum(Qv4) + C Intrinsic Prototype: HVX_Vector Q6_Vw_prefixsum_Q(HVX_VectorPred Qv) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vw_prefixsum_Q __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vprefixqw) +#endif /* __HEXAGON_ARCH___ >= 65 */ + +#if __HVX_ARCH__ >= 65 +/* ========================================================================== + Assembly Syntax: vscatter(Rt32,Mu2,Vv32.h).h=Vw32 + C Intrinsic Prototype: void Q6_vscatter_RMVhV(Word32 Rt, Word32 Mu, HVX_Vector Vv, HVX_Vector Vw) + Instruction Type: CVI_SCATTER + Execution Slots: SLOT0 + ========================================================================== */ + +#define Q6_vscatter_RMVhV __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vscattermh) +#endif /* __HEXAGON_ARCH___ >= 65 */ + +#if __HVX_ARCH__ >= 65 +/* ========================================================================== + Assembly Syntax: vscatter(Rt32,Mu2,Vv32.h).h+=Vw32 + C Intrinsic Prototype: void Q6_vscatteracc_RMVhV(Word32 Rt, Word32 Mu, HVX_Vector Vv, HVX_Vector Vw) + Instruction Type: CVI_SCATTER + Execution Slots: SLOT0 + ========================================================================== */ + +#define Q6_vscatteracc_RMVhV __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vscattermh_add) +#endif /* __HEXAGON_ARCH___ >= 65 */ + +#if __HVX_ARCH__ >= 65 +/* ========================================================================== + Assembly Syntax: if (Qs4) vscatter(Rt32,Mu2,Vv32.h).h=Vw32 + C Intrinsic Prototype: void Q6_vscatter_QRMVhV(HVX_VectorPred Qs, Word32 Rt, Word32 Mu, HVX_Vector Vv, HVX_Vector Vw) + Instruction Type: CVI_SCATTER + Execution Slots: SLOT0 + ========================================================================== */ + +#define Q6_vscatter_QRMVhV __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vscattermhq) +#endif /* __HEXAGON_ARCH___ >= 65 */ + +#if __HVX_ARCH__ >= 65 +/* ========================================================================== + Assembly Syntax: vscatter(Rt32,Mu2,Vvv32.w).h=Vw32 + C Intrinsic Prototype: void Q6_vscatter_RMWwV(Word32 Rt, Word32 Mu, HVX_VectorPair Vvv, HVX_Vector Vw) + Instruction Type: CVI_SCATTER_DV + Execution Slots: SLOT0 + ========================================================================== */ + +#define Q6_vscatter_RMWwV __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vscattermhw) +#endif /* __HEXAGON_ARCH___ >= 65 */ + +#if __HVX_ARCH__ >= 65 +/* ========================================================================== + Assembly Syntax: vscatter(Rt32,Mu2,Vvv32.w).h+=Vw32 + C Intrinsic Prototype: void Q6_vscatteracc_RMWwV(Word32 Rt, Word32 Mu, HVX_VectorPair Vvv, HVX_Vector Vw) + Instruction Type: CVI_SCATTER_DV + Execution Slots: SLOT0 + ========================================================================== */ + +#define Q6_vscatteracc_RMWwV __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vscattermhw_add) +#endif /* __HEXAGON_ARCH___ >= 65 */ + +#if __HVX_ARCH__ >= 65 +/* ========================================================================== + Assembly Syntax: if (Qs4) vscatter(Rt32,Mu2,Vvv32.w).h=Vw32 + C Intrinsic Prototype: void Q6_vscatter_QRMWwV(HVX_VectorPred Qs, Word32 Rt, Word32 Mu, HVX_VectorPair Vvv, HVX_Vector Vw) + Instruction Type: CVI_SCATTER_DV + Execution Slots: SLOT0 + ========================================================================== */ + +#define Q6_vscatter_QRMWwV __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vscattermhwq) +#endif /* __HEXAGON_ARCH___ >= 65 */ + +#if __HVX_ARCH__ >= 65 +/* ========================================================================== + Assembly Syntax: vscatter(Rt32,Mu2,Vv32.w).w=Vw32 + C Intrinsic Prototype: void Q6_vscatter_RMVwV(Word32 Rt, Word32 Mu, HVX_Vector Vv, HVX_Vector Vw) + Instruction Type: CVI_SCATTER + Execution Slots: SLOT0 + ========================================================================== */ + +#define Q6_vscatter_RMVwV __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vscattermw) +#endif /* __HEXAGON_ARCH___ >= 65 */ + +#if __HVX_ARCH__ >= 65 +/* ========================================================================== + Assembly Syntax: vscatter(Rt32,Mu2,Vv32.w).w+=Vw32 + C Intrinsic Prototype: void Q6_vscatteracc_RMVwV(Word32 Rt, Word32 Mu, HVX_Vector Vv, HVX_Vector Vw) + Instruction Type: CVI_SCATTER + Execution Slots: SLOT0 + ========================================================================== */ + +#define Q6_vscatteracc_RMVwV __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vscattermw_add) +#endif /* __HEXAGON_ARCH___ >= 65 */ + +#if __HVX_ARCH__ >= 65 +/* ========================================================================== + Assembly Syntax: if (Qs4) vscatter(Rt32,Mu2,Vv32.w).w=Vw32 + C Intrinsic Prototype: void Q6_vscatter_QRMVwV(HVX_VectorPred Qs, Word32 Rt, Word32 Mu, HVX_Vector Vv, HVX_Vector Vw) + Instruction Type: CVI_SCATTER + Execution Slots: SLOT0 + ========================================================================== */ + +#define Q6_vscatter_QRMVwV __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vscattermwq) +#endif /* __HEXAGON_ARCH___ >= 65 */ + +#if __HVX_ARCH__ >= 66 +/* ========================================================================== + Assembly Syntax: Vd32.w=vadd(Vu32.w,Vv32.w,Qs4):carry:sat + C Intrinsic Prototype: HVX_Vector Q6_Vw_vadd_VwVwQ_carry_sat(HVX_Vector Vu, HVX_Vector Vv, HVX_VectorPred Qs) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vw_vadd_VwVwQ_carry_sat __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vaddcarrysat) +#endif /* __HEXAGON_ARCH___ >= 66 */ + +#if __HVX_ARCH__ >= 66 +/* ========================================================================== + Assembly Syntax: Vxx32.w=vasrinto(Vu32.w,Vv32.w) + C Intrinsic Prototype: HVX_VectorPair Q6_Ww_vasrinto_WwVwVw(HVX_VectorPair Vxx, HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VP_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Ww_vasrinto_WwVwVw __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vasr_into) +#endif /* __HEXAGON_ARCH___ >= 66 */ + +#if __HVX_ARCH__ >= 66 +/* ========================================================================== + Assembly Syntax: Vd32.uw=vrotr(Vu32.uw,Vv32.uw) + C Intrinsic Prototype: HVX_Vector Q6_Vuw_vrotr_VuwVuw(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VS + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vuw_vrotr_VuwVuw __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vrotr) +#endif /* __HEXAGON_ARCH___ >= 66 */ + +#if __HVX_ARCH__ >= 66 +/* ========================================================================== + Assembly Syntax: Vd32.w=vsatdw(Vu32.w,Vv32.w) + C Intrinsic Prototype: HVX_Vector Q6_Vw_vsatdw_VwVw(HVX_Vector Vu, HVX_Vector Vv) + Instruction Type: CVI_VA + Execution Slots: SLOT0123 + ========================================================================== */ + +#define Q6_Vw_vsatdw_VwVw __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_vsatdw) +#endif /* __HEXAGON_ARCH___ >= 66 */ + +#if __HVX_ARCH__ >= 68 +/* ========================================================================== + Assembly Syntax: Vdd32.w=v6mpy(Vuu32.ub,Vvv32.b,#u2):h + C Intrinsic Prototype: HVX_VectorPair Q6_Ww_v6mpy_WubWbI_h(HVX_VectorPair Vuu, HVX_VectorPair Vvv, Word32 Iu2) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Ww_v6mpy_WubWbI_h __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_v6mpyhubs10) +#endif /* __HEXAGON_ARCH___ >= 68 */ + +#if __HVX_ARCH__ >= 68 +/* ========================================================================== + Assembly Syntax: Vxx32.w+=v6mpy(Vuu32.ub,Vvv32.b,#u2):h + C Intrinsic Prototype: HVX_VectorPair Q6_Ww_v6mpyacc_WwWubWbI_h(HVX_VectorPair Vxx, HVX_VectorPair Vuu, HVX_VectorPair Vvv, Word32 Iu2) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Ww_v6mpyacc_WwWubWbI_h __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_v6mpyhubs10_vxx) +#endif /* __HEXAGON_ARCH___ >= 68 */ + +#if __HVX_ARCH__ >= 68 +/* ========================================================================== + Assembly Syntax: Vdd32.w=v6mpy(Vuu32.ub,Vvv32.b,#u2):v + C Intrinsic Prototype: HVX_VectorPair Q6_Ww_v6mpy_WubWbI_v(HVX_VectorPair Vuu, HVX_VectorPair Vvv, Word32 Iu2) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Ww_v6mpy_WubWbI_v __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_v6mpyvubs10) +#endif /* __HEXAGON_ARCH___ >= 68 */ + +#if __HVX_ARCH__ >= 68 +/* ========================================================================== + Assembly Syntax: Vxx32.w+=v6mpy(Vuu32.ub,Vvv32.b,#u2):v + C Intrinsic Prototype: HVX_VectorPair Q6_Ww_v6mpyacc_WwWubWbI_v(HVX_VectorPair Vxx, HVX_VectorPair Vuu, HVX_VectorPair Vvv, Word32 Iu2) + Instruction Type: CVI_VX_DV + Execution Slots: SLOT23 + ========================================================================== */ + +#define Q6_Ww_v6mpyacc_WwWubWbI_v __BUILTIN_VECTOR_WRAP(__builtin_HEXAGON_V6_v6mpyvubs10_vxx) +#endif /* __HEXAGON_ARCH___ >= 68 */ + +#endif /* __HVX__ */ + +#endif diff --git a/lib/include/immintrin.h b/lib/include/immintrin.h index 22f7a520c..56d3dadf6 100644 --- a/lib/include/immintrin.h +++ b/lib/include/immintrin.h @@ -72,11 +72,6 @@ #include #endif -#if !(defined(_MSC_VER) || defined(__SCE__)) || __has_feature(modules) || \ - defined(__VPCLMULQDQ__) -#include -#endif - /* No feature check desired due to internal checks */ #include @@ -230,6 +225,11 @@ #include #endif +#if !(defined(_MSC_VER) || defined(__SCE__)) || __has_feature(modules) || \ + defined(__VPCLMULQDQ__) +#include +#endif + #if !(defined(_MSC_VER) || defined(__SCE__)) || __has_feature(modules) || \ defined(__VAES__) #include diff --git a/lib/include/intrin.h b/lib/include/intrin.h index a78b96997..34ec79d6a 100644 --- a/lib/include/intrin.h +++ b/lib/include/intrin.h @@ -451,24 +451,47 @@ unsigned char _InterlockedCompareExchange128_rel(__int64 volatile *_Destination, static __inline__ void __DEFAULT_FN_ATTRS __movsb(unsigned char *__dst, unsigned char const *__src, size_t __n) { - __asm__ __volatile__("rep movsb" : "+D"(__dst), "+S"(__src), "+c"(__n) - : : "memory"); +#if defined(__x86_64__) + __asm__ __volatile__("rep movsb" + : "+D"(__dst), "+S"(__src), "+c"(__n) + : + : "memory"); +#else + __asm__ __volatile__("xchg %%esi, %1\nrep movsb\nxchg %%esi, %1" + : "+D"(__dst), "+r"(__src), "+c"(__n) + : + : "memory"); +#endif } static __inline__ void __DEFAULT_FN_ATTRS __movsd(unsigned long *__dst, unsigned long const *__src, size_t __n) { +#if defined(__x86_64__) __asm__ __volatile__("rep movsl" : "+D"(__dst), "+S"(__src), "+c"(__n) : : "memory"); +#else + __asm__ __volatile__("xchg %%esi, %1\nrep movsl\nxchg %%esi, %1" + : "+D"(__dst), "+r"(__src), "+c"(__n) + : + : "memory"); +#endif } static __inline__ void __DEFAULT_FN_ATTRS __movsw(unsigned short *__dst, unsigned short const *__src, size_t __n) { +#if defined(__x86_64__) __asm__ __volatile__("rep movsw" : "+D"(__dst), "+S"(__src), "+c"(__n) : : "memory"); +#else + __asm__ __volatile__("xchg %%esi, %1\nrep movsw\nxchg %%esi, %1" + : "+D"(__dst), "+r"(__src), "+c"(__n) + : + : "memory"); +#endif } static __inline__ void __DEFAULT_FN_ATTRS __stosd(unsigned long *__dst, unsigned long __x, @@ -507,16 +530,26 @@ static __inline__ void __DEFAULT_FN_ATTRS __stosq(unsigned __int64 *__dst, |* Misc \*----------------------------------------------------------------------------*/ #if defined(__i386__) || defined(__x86_64__) +#if defined(__i386__) +#define __cpuid_count(__leaf, __count, __eax, __ebx, __ecx, __edx) \ + __asm("cpuid" \ + : "=a"(__eax), "=b"(__ebx), "=c"(__ecx), "=d"(__edx) \ + : "0"(__leaf), "2"(__count)) +#else +/* x86-64 uses %rbx as the base register, so preserve it. */ +#define __cpuid_count(__leaf, __count, __eax, __ebx, __ecx, __edx) \ + __asm("xchgq %%rbx,%q1\n" \ + "cpuid\n" \ + "xchgq %%rbx,%q1" \ + : "=a"(__eax), "=r"(__ebx), "=c"(__ecx), "=d"(__edx) \ + : "0"(__leaf), "2"(__count)) +#endif static __inline__ void __DEFAULT_FN_ATTRS __cpuid(int __info[4], int __level) { - __asm__("cpuid" - : "=a"(__info[0]), "=b"(__info[1]), "=c"(__info[2]), "=d"(__info[3]) - : "a"(__level), "c"(0)); + __cpuid_count(__level, 0, __info[0], __info[1], __info[2], __info[3]); } static __inline__ void __DEFAULT_FN_ATTRS __cpuidex(int __info[4], int __level, int __ecx) { - __asm__("cpuid" - : "=a"(__info[0]), "=b"(__info[1]), "=c"(__info[2]), "=d"(__info[3]) - : "a"(__level), "c"(__ecx)); + __cpuid_count(__level, __ecx, __info[0], __info[1], __info[2], __info[3]); } static __inline__ void __DEFAULT_FN_ATTRS __halt(void) { __asm__ volatile("hlt"); @@ -541,6 +574,9 @@ void _WriteStatusReg(int, __int64); unsigned short __cdecl _byteswap_ushort(unsigned short val); unsigned long __cdecl _byteswap_ulong (unsigned long val); unsigned __int64 __cdecl _byteswap_uint64(unsigned __int64 val); + +__int64 __mulh(__int64 __a, __int64 __b); +unsigned __int64 __umulh(unsigned __int64 __a, unsigned __int64 __b); #endif /*----------------------------------------------------------------------------*\ diff --git a/lib/include/keylockerintrin.h b/lib/include/keylockerintrin.h index c15d39c8e..68b0a5689 100644 --- a/lib/include/keylockerintrin.h +++ b/lib/include/keylockerintrin.h @@ -230,10 +230,12 @@ _mm_aesenc128kl_u8(__m128i* __odata, __m128i __idata, const void *__h) { /// HandleKeyType (Handle[511:0]) != HANDLE_KEY_TYPE_AES256 ) /// IF (IllegalHandle) /// ZF := 1 +/// MEM[__odata+127:__odata] := 0 /// ELSE /// (UnwrappedKey, Authentic) := UnwrapKeyAndAuthenticate512 (Handle[511:0], IWKey) /// IF (Authentic == 0) /// ZF := 1 +/// MEM[__odata+127:__odata] := 0 /// ELSE /// MEM[__odata+127:__odata] := AES256Encrypt (__idata[127:0], UnwrappedKey) /// ZF := 0 @@ -267,10 +269,12 @@ _mm_aesenc256kl_u8(__m128i* __odata, __m128i __idata, const void *__h) { /// HandleKeyType (Handle[383:0]) != HANDLE_KEY_TYPE_AES128) /// IF (IllegalHandle) /// ZF := 1 +/// MEM[__odata+127:__odata] := 0 /// ELSE /// (UnwrappedKey, Authentic) := UnwrapKeyAndAuthenticate384 (Handle[383:0], IWKey) /// IF (Authentic == 0) /// ZF := 1 +/// MEM[__odata+127:__odata] := 0 /// ELSE /// MEM[__odata+127:__odata] := AES128Decrypt (__idata[127:0], UnwrappedKey) /// ZF := 0 @@ -304,10 +308,12 @@ _mm_aesdec128kl_u8(__m128i* __odata, __m128i __idata, const void *__h) { /// HandleKeyType (Handle[511:0]) != HANDLE_KEY_TYPE_AES256) /// IF (IllegalHandle) /// ZF := 1 +/// MEM[__odata+127:__odata] := 0 /// ELSE /// (UnwrappedKey, Authentic) := UnwrapKeyAndAuthenticate512 (Handle[511:0], IWKey) /// IF (Authentic == 0) /// ZF := 1 +/// MEM[__odata+127:__odata] := 0 /// ELSE /// MEM[__odata+127:__odata] := AES256Decrypt (__idata[127:0], UnwrappedKey) /// ZF := 0 @@ -354,10 +360,16 @@ _mm_aesdec256kl_u8(__m128i* __odata, __m128i __idata, const void *__h) { /// HandleKeyType (Handle[383:0]) != HANDLE_KEY_TYPE_AES128 ) /// IF (IllegalHandle) /// ZF := 1 +/// FOR i := 0 to 7 +/// __odata[i] := 0 +/// ENDFOR /// ELSE /// (UnwrappedKey, Authentic) := UnwrapKeyAndAuthenticate384 (Handle[383:0], IWKey) /// IF Authentic == 0 /// ZF := 1 +/// FOR i := 0 to 7 +/// __odata[i] := 0 +/// ENDFOR /// ELSE /// FOR i := 0 to 7 /// __odata[i] := AES128Encrypt (__idata[i], UnwrappedKey) @@ -394,10 +406,16 @@ _mm_aesencwide128kl_u8(__m128i __odata[8], const __m128i __idata[8], const void* /// HandleKeyType (Handle[511:0]) != HANDLE_KEY_TYPE_AES512 ) /// IF (IllegalHandle) /// ZF := 1 +/// FOR i := 0 to 7 +/// __odata[i] := 0 +/// ENDFOR /// ELSE /// (UnwrappedKey, Authentic) := UnwrapKeyAndAuthenticate512 (Handle[511:0], IWKey) /// IF Authentic == 0 /// ZF := 1 +/// FOR i := 0 to 7 +/// __odata[i] := 0 +/// ENDFOR /// ELSE /// FOR i := 0 to 7 /// __odata[i] := AES256Encrypt (__idata[i], UnwrappedKey) @@ -434,10 +452,16 @@ _mm_aesencwide256kl_u8(__m128i __odata[8], const __m128i __idata[8], const void* /// HandleKeyType (Handle) != HANDLE_KEY_TYPE_AES128 ) /// IF (IllegalHandle) /// ZF := 1 +/// FOR i := 0 to 7 +/// __odata[i] := 0 +/// ENDFOR /// ELSE /// (UnwrappedKey, Authentic) := UnwrapKeyAndAuthenticate384 (Handle[383:0], IWKey) /// IF Authentic == 0 /// ZF := 1 +/// FOR i := 0 to 7 +/// __odata[i] := 0 +/// ENDFOR /// ELSE /// FOR i := 0 to 7 /// __odata[i] := AES128Decrypt (__idata[i], UnwrappedKey) @@ -474,10 +498,16 @@ _mm_aesdecwide128kl_u8(__m128i __odata[8], const __m128i __idata[8], const void* /// HandleKeyType (Handle) != HANDLE_KEY_TYPE_AES512 ) /// If (IllegalHandle) /// ZF := 1 +/// FOR i := 0 to 7 +/// __odata[i] := 0 +/// ENDFOR /// ELSE /// (UnwrappedKey, Authentic) := UnwrapKeyAndAuthenticate512 (Handle[511:0], IWKey) /// IF Authentic == 0 /// ZF := 1 +/// FOR i := 0 to 7 +/// __odata[i] := 0 +/// ENDFOR /// ELSE /// FOR i := 0 to 7 /// __odata[i] := AES256Decrypt (__idata[i], UnwrappedKey) diff --git a/lib/include/opencl-c-base.h b/lib/include/opencl-c-base.h index e8dcd7037..3c5e2c973 100644 --- a/lib/include/opencl-c-base.h +++ b/lib/include/opencl-c-base.h @@ -21,9 +21,37 @@ #define cl_khr_subgroup_shuffle 1 #define cl_khr_subgroup_shuffle_relative 1 #define cl_khr_subgroup_clustered_reduce 1 +#define cl_khr_extended_bit_ops 1 +#define cl_khr_integer_dot_product 1 +#define __opencl_c_integer_dot_product_input_4x8bit 1 +#define __opencl_c_integer_dot_product_input_4x8bit_packed 1 + #endif // defined(__SPIR__) #endif // (defined(__OPENCL_CPP_VERSION__) || __OPENCL_C_VERSION__ >= 200) +// Define feature macros for OpenCL C 2.0 +#if (defined(__OPENCL_CPP_VERSION__) || __OPENCL_C_VERSION__ == 200) +#define __opencl_c_pipes 1 +#define __opencl_c_generic_address_space 1 +#define __opencl_c_work_group_collective_functions 1 +#define __opencl_c_atomic_order_acq_rel 1 +#define __opencl_c_atomic_order_seq_cst 1 +#define __opencl_c_atomic_scope_device 1 +#define __opencl_c_atomic_scope_all_devices 1 +#define __opencl_c_device_enqueue 1 +#define __opencl_c_read_write_images 1 +#define __opencl_c_program_scope_global_variables 1 +#define __opencl_c_images 1 +#endif + +// Define header-only feature macros for OpenCL C 3.0. +#if (__OPENCL_C_VERSION__ == 300) +// For the SPIR target all features are supported. +#if defined(__SPIR__) +#define __opencl_c_atomic_scope_all_devices 1 +#endif // defined(__SPIR__) +#endif // (__OPENCL_C_VERSION__ == 300) + // built-in scalar data types: /** @@ -141,7 +169,9 @@ typedef double double8 __attribute__((ext_vector_type(8))); typedef double double16 __attribute__((ext_vector_type(16))); #endif -#if defined(__OPENCL_CPP_VERSION__) || (__OPENCL_C_VERSION__ >= CL_VERSION_2_0) +#if defined(__OPENCL_CPP_VERSION__) +#define NULL nullptr +#elif defined(__OPENCL_C_VERSION__) #define NULL ((void*)0) #endif @@ -297,7 +327,12 @@ typedef enum memory_scope { memory_scope_work_item = __OPENCL_MEMORY_SCOPE_WORK_ITEM, memory_scope_work_group = __OPENCL_MEMORY_SCOPE_WORK_GROUP, memory_scope_device = __OPENCL_MEMORY_SCOPE_DEVICE, +#if defined(__opencl_c_atomic_scope_all_devices) memory_scope_all_svm_devices = __OPENCL_MEMORY_SCOPE_ALL_SVM_DEVICES, +#if (__OPENCL_C_VERSION__ >= CL_VERSION_3_0) + memory_scope_all_devices = memory_scope_all_svm_devices, +#endif // __OPENCL_C_VERSION__ >= CL_VERSION_3_0 +#endif // defined(__opencl_c_atomic_scope_all_devices) #if defined(cl_intel_subgroups) || defined(cl_khr_subgroups) memory_scope_sub_group = __OPENCL_MEMORY_SCOPE_SUB_GROUP #endif @@ -322,7 +357,9 @@ typedef enum memory_order memory_order_acquire = __ATOMIC_ACQUIRE, memory_order_release = __ATOMIC_RELEASE, memory_order_acq_rel = __ATOMIC_ACQ_REL, +#if defined(__opencl_c_atomic_order_seq_cst) memory_order_seq_cst = __ATOMIC_SEQ_CST +#endif } memory_order; #endif // defined(__OPENCL_CPP_VERSION__) || (__OPENCL_C_VERSION__ >= CL_VERSION_2_0) @@ -445,8 +482,113 @@ typedef struct { #endif // defined(__OPENCL_CPP_VERSION__) || (__OPENCL_C_VERSION__ >= CL_VERSION_2_0) +/** + * OpenCL v1.1/1.2/2.0 s6.2.4.2 - as_type operators + * Reinterprets a data type as another data type of the same size + */ +#define as_char(x) __builtin_astype((x), char) +#define as_char2(x) __builtin_astype((x), char2) +#define as_char3(x) __builtin_astype((x), char3) +#define as_char4(x) __builtin_astype((x), char4) +#define as_char8(x) __builtin_astype((x), char8) +#define as_char16(x) __builtin_astype((x), char16) + +#define as_uchar(x) __builtin_astype((x), uchar) +#define as_uchar2(x) __builtin_astype((x), uchar2) +#define as_uchar3(x) __builtin_astype((x), uchar3) +#define as_uchar4(x) __builtin_astype((x), uchar4) +#define as_uchar8(x) __builtin_astype((x), uchar8) +#define as_uchar16(x) __builtin_astype((x), uchar16) + +#define as_short(x) __builtin_astype((x), short) +#define as_short2(x) __builtin_astype((x), short2) +#define as_short3(x) __builtin_astype((x), short3) +#define as_short4(x) __builtin_astype((x), short4) +#define as_short8(x) __builtin_astype((x), short8) +#define as_short16(x) __builtin_astype((x), short16) + +#define as_ushort(x) __builtin_astype((x), ushort) +#define as_ushort2(x) __builtin_astype((x), ushort2) +#define as_ushort3(x) __builtin_astype((x), ushort3) +#define as_ushort4(x) __builtin_astype((x), ushort4) +#define as_ushort8(x) __builtin_astype((x), ushort8) +#define as_ushort16(x) __builtin_astype((x), ushort16) + +#define as_int(x) __builtin_astype((x), int) +#define as_int2(x) __builtin_astype((x), int2) +#define as_int3(x) __builtin_astype((x), int3) +#define as_int4(x) __builtin_astype((x), int4) +#define as_int8(x) __builtin_astype((x), int8) +#define as_int16(x) __builtin_astype((x), int16) + +#define as_uint(x) __builtin_astype((x), uint) +#define as_uint2(x) __builtin_astype((x), uint2) +#define as_uint3(x) __builtin_astype((x), uint3) +#define as_uint4(x) __builtin_astype((x), uint4) +#define as_uint8(x) __builtin_astype((x), uint8) +#define as_uint16(x) __builtin_astype((x), uint16) + +#define as_long(x) __builtin_astype((x), long) +#define as_long2(x) __builtin_astype((x), long2) +#define as_long3(x) __builtin_astype((x), long3) +#define as_long4(x) __builtin_astype((x), long4) +#define as_long8(x) __builtin_astype((x), long8) +#define as_long16(x) __builtin_astype((x), long16) + +#define as_ulong(x) __builtin_astype((x), ulong) +#define as_ulong2(x) __builtin_astype((x), ulong2) +#define as_ulong3(x) __builtin_astype((x), ulong3) +#define as_ulong4(x) __builtin_astype((x), ulong4) +#define as_ulong8(x) __builtin_astype((x), ulong8) +#define as_ulong16(x) __builtin_astype((x), ulong16) + +#define as_float(x) __builtin_astype((x), float) +#define as_float2(x) __builtin_astype((x), float2) +#define as_float3(x) __builtin_astype((x), float3) +#define as_float4(x) __builtin_astype((x), float4) +#define as_float8(x) __builtin_astype((x), float8) +#define as_float16(x) __builtin_astype((x), float16) + +#ifdef cl_khr_fp64 +#define as_double(x) __builtin_astype((x), double) +#define as_double2(x) __builtin_astype((x), double2) +#define as_double3(x) __builtin_astype((x), double3) +#define as_double4(x) __builtin_astype((x), double4) +#define as_double8(x) __builtin_astype((x), double8) +#define as_double16(x) __builtin_astype((x), double16) +#endif // cl_khr_fp64 + +#ifdef cl_khr_fp16 +#define as_half(x) __builtin_astype((x), half) +#define as_half2(x) __builtin_astype((x), half2) +#define as_half3(x) __builtin_astype((x), half3) +#define as_half4(x) __builtin_astype((x), half4) +#define as_half8(x) __builtin_astype((x), half8) +#define as_half16(x) __builtin_astype((x), half16) +#endif // cl_khr_fp16 + +#define as_size_t(x) __builtin_astype((x), size_t) +#define as_ptrdiff_t(x) __builtin_astype((x), ptrdiff_t) +#define as_intptr_t(x) __builtin_astype((x), intptr_t) +#define as_uintptr_t(x) __builtin_astype((x), uintptr_t) + +// OpenCL v1.1 s6.9, v1.2/2.0 s6.10 - Function qualifiers + +#define __kernel_exec(X, typen) __kernel \ + __attribute__((work_group_size_hint(X, 1, 1))) \ + __attribute__((vec_type_hint(typen))) + +#define kernel_exec(X, typen) __kernel \ + __attribute__((work_group_size_hint(X, 1, 1))) \ + __attribute__((vec_type_hint(typen))) + +#if defined(__OPENCL_CPP_VERSION__) || (__OPENCL_C_VERSION__ >= CL_VERSION_1_2) +// OpenCL v1.2 s6.12.13, v2.0 s6.13.13 - printf + +int printf(__constant const char* st, ...) __attribute__((format(printf, 1, 2))); +#endif + #ifdef cl_intel_device_side_avc_motion_estimation -#pragma OPENCL EXTENSION cl_intel_device_side_avc_motion_estimation : begin #define CLK_AVC_ME_MAJOR_16x16_INTEL 0x0 #define CLK_AVC_ME_MAJOR_16x8_INTEL 0x1 @@ -580,7 +722,6 @@ typedef struct { #define CLK_AVC_IME_RESULT_DUAL_REFERENCE_STREAMOUT_INITIALIZE_INTEL 0x0 #define CLK_AVC_IME_RESULT_DUAL_REFERENCE_STREAMIN_INITIALIZE_INTEL 0x0 -#pragma OPENCL EXTENSION cl_intel_device_side_avc_motion_estimation : end #endif // cl_intel_device_side_avc_motion_estimation // Disable any extensions we may have enabled previously. diff --git a/lib/include/opencl-c.h b/lib/include/opencl-c.h index ab665628c..fc50dd718 100644 --- a/lib/include/opencl-c.h +++ b/lib/include/opencl-c.h @@ -23,10 +23,11 @@ #endif //cl_khr_3d_image_writes #endif //__OPENCL_C_VERSION__ < CL_VERSION_2_0 -#if defined(__OPENCL_CPP_VERSION__) || (__OPENCL_C_VERSION__ >= CL_VERSION_1_2) + +#if (defined(__OPENCL_CPP_VERSION__) || (__OPENCL_C_VERSION__ >= CL_VERSION_1_2)) && defined(__SPIR__) #pragma OPENCL EXTENSION cl_intel_planar_yuv : begin #pragma OPENCL EXTENSION cl_intel_planar_yuv : end -#endif // defined(__OPENCL_CPP_VERSION__) || (__OPENCL_C_VERSION__ >= CL_VERSION_1_2) +#endif // (defined(__OPENCL_CPP_VERSION__) || (__OPENCL_C_VERSION__ >= CL_VERSION_1_2)) && defined(__SPIR__) #define __ovld __attribute__((overloadable)) #define __conv __attribute__((convergent)) @@ -6339,101 +6340,6 @@ half16 __ovld __cnfn convert_half16_rtz(double16); #endif // cl_khr_fp16 -/** - * OpenCL v1.1/1.2/2.0 s6.2.4.2 - as_type operators - * Reinterprets a data type as another data type of the same size - */ -#define as_char(x) __builtin_astype((x), char) -#define as_char2(x) __builtin_astype((x), char2) -#define as_char3(x) __builtin_astype((x), char3) -#define as_char4(x) __builtin_astype((x), char4) -#define as_char8(x) __builtin_astype((x), char8) -#define as_char16(x) __builtin_astype((x), char16) - -#define as_uchar(x) __builtin_astype((x), uchar) -#define as_uchar2(x) __builtin_astype((x), uchar2) -#define as_uchar3(x) __builtin_astype((x), uchar3) -#define as_uchar4(x) __builtin_astype((x), uchar4) -#define as_uchar8(x) __builtin_astype((x), uchar8) -#define as_uchar16(x) __builtin_astype((x), uchar16) - -#define as_short(x) __builtin_astype((x), short) -#define as_short2(x) __builtin_astype((x), short2) -#define as_short3(x) __builtin_astype((x), short3) -#define as_short4(x) __builtin_astype((x), short4) -#define as_short8(x) __builtin_astype((x), short8) -#define as_short16(x) __builtin_astype((x), short16) - -#define as_ushort(x) __builtin_astype((x), ushort) -#define as_ushort2(x) __builtin_astype((x), ushort2) -#define as_ushort3(x) __builtin_astype((x), ushort3) -#define as_ushort4(x) __builtin_astype((x), ushort4) -#define as_ushort8(x) __builtin_astype((x), ushort8) -#define as_ushort16(x) __builtin_astype((x), ushort16) - -#define as_int(x) __builtin_astype((x), int) -#define as_int2(x) __builtin_astype((x), int2) -#define as_int3(x) __builtin_astype((x), int3) -#define as_int4(x) __builtin_astype((x), int4) -#define as_int8(x) __builtin_astype((x), int8) -#define as_int16(x) __builtin_astype((x), int16) - -#define as_uint(x) __builtin_astype((x), uint) -#define as_uint2(x) __builtin_astype((x), uint2) -#define as_uint3(x) __builtin_astype((x), uint3) -#define as_uint4(x) __builtin_astype((x), uint4) -#define as_uint8(x) __builtin_astype((x), uint8) -#define as_uint16(x) __builtin_astype((x), uint16) - -#define as_long(x) __builtin_astype((x), long) -#define as_long2(x) __builtin_astype((x), long2) -#define as_long3(x) __builtin_astype((x), long3) -#define as_long4(x) __builtin_astype((x), long4) -#define as_long8(x) __builtin_astype((x), long8) -#define as_long16(x) __builtin_astype((x), long16) - -#define as_ulong(x) __builtin_astype((x), ulong) -#define as_ulong2(x) __builtin_astype((x), ulong2) -#define as_ulong3(x) __builtin_astype((x), ulong3) -#define as_ulong4(x) __builtin_astype((x), ulong4) -#define as_ulong8(x) __builtin_astype((x), ulong8) -#define as_ulong16(x) __builtin_astype((x), ulong16) - -#define as_float(x) __builtin_astype((x), float) -#define as_float2(x) __builtin_astype((x), float2) -#define as_float3(x) __builtin_astype((x), float3) -#define as_float4(x) __builtin_astype((x), float4) -#define as_float8(x) __builtin_astype((x), float8) -#define as_float16(x) __builtin_astype((x), float16) - -#ifdef cl_khr_fp64 -#define as_double(x) __builtin_astype((x), double) -#define as_double2(x) __builtin_astype((x), double2) -#define as_double3(x) __builtin_astype((x), double3) -#define as_double4(x) __builtin_astype((x), double4) -#define as_double8(x) __builtin_astype((x), double8) -#define as_double16(x) __builtin_astype((x), double16) -#endif //cl_khr_fp64 - -#ifdef cl_khr_fp16 -#define as_half(x) __builtin_astype((x), half) -#define as_half2(x) __builtin_astype((x), half2) -#define as_half3(x) __builtin_astype((x), half3) -#define as_half4(x) __builtin_astype((x), half4) -#define as_half8(x) __builtin_astype((x), half8) -#define as_half16(x) __builtin_astype((x), half16) -#endif //cl_khr_fp16 - -// OpenCL v1.1 s6.9, v1.2/2.0 s6.10 - Function qualifiers - -#define __kernel_exec(X, typen) __kernel \ - __attribute__((work_group_size_hint(X, 1, 1))) \ - __attribute__((vec_type_hint(typen))) - -#define kernel_exec(X, typen) __kernel \ - __attribute__((work_group_size_hint(X, 1, 1))) \ - __attribute__((vec_type_hint(typen))) - // OpenCL v1.1 s6.11.1, v1.2 s6.12.1, v2.0 s6.13.1 - Work-item Functions /** @@ -6494,8 +6400,7 @@ size_t __ovld __cnfn get_local_id(uint dimindx); * Returns the number of work-groups that will execute a * kernel for dimension identified by dimindx. * Valid values of dimindx are 0 to get_work_dim() - 1. - * For other values of dimindx, get_num_groups () returns - * 1. + * For other values of dimindx, get_num_groups() returns 1. * For clEnqueueTask, this always returns 1. */ size_t __ovld __cnfn get_num_groups(uint dimindx); @@ -7354,7 +7259,7 @@ half16 __ovld __cnfn fmod(half16 x, half16 y); * Returns fmin(x - floor (x), 0x1.fffffep-1f ). * floor(x) is returned in iptr. */ -#if defined(__OPENCL_CPP_VERSION__) || (__OPENCL_C_VERSION__ >= CL_VERSION_2_0) +#if defined(__opencl_c_generic_address_space) float __ovld fract(float x, float *iptr); float2 __ovld fract(float2 x, float2 *iptr); float3 __ovld fract(float3 x, float3 *iptr); @@ -7436,7 +7341,7 @@ half4 __ovld fract(half4 x, __private half4 *iptr); half8 __ovld fract(half8 x, __private half8 *iptr); half16 __ovld fract(half16 x, __private half16 *iptr); #endif //cl_khr_fp16 -#endif //defined(__OPENCL_CPP_VERSION__) || (__OPENCL_C_VERSION__ >= CL_VERSION_2_0) +#endif //defined(__opencl_c_generic_address_space) /** * Extract mantissa and exponent from x. For each @@ -7444,7 +7349,7 @@ half16 __ovld fract(half16 x, __private half16 *iptr); * magnitude in the interval [1/2, 1) or 0. Each * component of x equals mantissa returned * 2^exp. */ -#if defined(__OPENCL_CPP_VERSION__) || (__OPENCL_C_VERSION__ >= CL_VERSION_2_0) +#if defined(__opencl_c_generic_address_space) float __ovld frexp(float x, int *exp); float2 __ovld frexp(float2 x, int2 *exp); float3 __ovld frexp(float3 x, int3 *exp); @@ -7526,7 +7431,7 @@ half4 __ovld frexp(half4 x, __private int4 *exp); half8 __ovld frexp(half8 x, __private int8 *exp); half16 __ovld frexp(half16 x, __private int16 *exp); #endif //cl_khr_fp16 -#endif //defined(__OPENCL_CPP_VERSION__) || (__OPENCL_C_VERSION__ >= CL_VERSION_2_0) +#endif //defined(__opencl_c_generic_address_space) /** * Compute the value of the square root of x^2 + y^2 @@ -7651,7 +7556,7 @@ half8 __ovld __cnfn lgamma(half8 x); half16 __ovld __cnfn lgamma(half16 x); #endif //cl_khr_fp16 -#if defined(__OPENCL_CPP_VERSION__) || (__OPENCL_C_VERSION__ >= CL_VERSION_2_0) +#if defined(__opencl_c_generic_address_space) float __ovld lgamma_r(float x, int *signp); float2 __ovld lgamma_r(float2 x, int2 *signp); float3 __ovld lgamma_r(float3 x, int3 *signp); @@ -7733,7 +7638,7 @@ half4 __ovld lgamma_r(half4 x, __private int4 *signp); half8 __ovld lgamma_r(half8 x, __private int8 *signp); half16 __ovld lgamma_r(half16 x, __private int16 *signp); #endif //cl_khr_fp16 -#endif //defined(__OPENCL_CPP_VERSION__) || (__OPENCL_C_VERSION__ >= CL_VERSION_2_0) +#endif //defined(__opencl_c_generic_address_space) /** * Compute natural logarithm. @@ -7957,7 +7862,7 @@ half16 __ovld __cnfn minmag(half16 x, half16 y); * the argument. It stores the integral part in the object * pointed to by iptr. */ -#if defined(__OPENCL_CPP_VERSION__) || (__OPENCL_C_VERSION__ >= CL_VERSION_2_0) +#if defined(__opencl_c_generic_address_space) float __ovld modf(float x, float *iptr); float2 __ovld modf(float2 x, float2 *iptr); float3 __ovld modf(float3 x, float3 *iptr); @@ -8039,7 +7944,7 @@ half4 __ovld modf(half4 x, __private half4 *iptr); half8 __ovld modf(half8 x, __private half8 *iptr); half16 __ovld modf(half16 x, __private half16 *iptr); #endif //cl_khr_fp16 -#endif //defined(__OPENCL_CPP_VERSION__) || (__OPENCL_C_VERSION__ >= CL_VERSION_2_0) +#endif //defined(__opencl_c_generic_address_space) /** * Returns a quiet NaN. The nancode may be placed @@ -8217,7 +8122,7 @@ half16 __ovld __cnfn remainder(half16 x, half16 y); * sign as x/y. It stores this signed value in the object * pointed to by quo. */ -#if defined(__OPENCL_CPP_VERSION__) || (__OPENCL_C_VERSION__ >= CL_VERSION_2_0) +#if defined(__opencl_c_generic_address_space) float __ovld remquo(float x, float y, int *quo); float2 __ovld remquo(float2 x, float2 y, int2 *quo); float3 __ovld remquo(float3 x, float3 y, int3 *quo); @@ -8300,7 +8205,7 @@ half4 __ovld remquo(half4 x, half4 y, __private int4 *quo); half8 __ovld remquo(half8 x, half8 y, __private int8 *quo); half16 __ovld remquo(half16 x, half16 y, __private int16 *quo); #endif //cl_khr_fp16 -#endif //defined(__OPENCL_CPP_VERSION__) || (__OPENCL_C_VERSION__ >= CL_VERSION_2_0) +#endif //defined(__opencl_c_generic_address_space) /** * Round to integral value (using round to nearest * even rounding mode) in floating-point format. @@ -8441,7 +8346,7 @@ half16 __ovld __cnfn sin(half16); * is the return value and computed cosine is returned * in cosval. */ -#if defined(__OPENCL_CPP_VERSION__) || (__OPENCL_C_VERSION__ >= CL_VERSION_2_0) +#if defined(__opencl_c_generic_address_space) float __ovld sincos(float x, float *cosval); float2 __ovld sincos(float2 x, float2 *cosval); float3 __ovld sincos(float3 x, float3 *cosval); @@ -8523,7 +8428,7 @@ half4 __ovld sincos(half4 x, __private half4 *cosval); half8 __ovld sincos(half8 x, __private half8 *cosval); half16 __ovld sincos(half16 x, __private half16 *cosval); #endif //cl_khr_fp16 -#endif //defined(__OPENCL_CPP_VERSION__) || (__OPENCL_C_VERSION__ >= CL_VERSION_2_0) +#endif //defined(__opencl_c_generic_address_space) /** * Compute hyperbolic sine. @@ -9449,54 +9354,54 @@ ulong16 __ovld __cnfn clz(ulong16 x); * component type of x, if x is a vector. */ #if defined(__OPENCL_CPP_VERSION__) || (__OPENCL_C_VERSION__ >= CL_VERSION_2_0) -char __ovld ctz(char x); -uchar __ovld ctz(uchar x); -char2 __ovld ctz(char2 x); -uchar2 __ovld ctz(uchar2 x); -char3 __ovld ctz(char3 x); -uchar3 __ovld ctz(uchar3 x); -char4 __ovld ctz(char4 x); -uchar4 __ovld ctz(uchar4 x); -char8 __ovld ctz(char8 x); -uchar8 __ovld ctz(uchar8 x); -char16 __ovld ctz(char16 x); -uchar16 __ovld ctz(uchar16 x); -short __ovld ctz(short x); -ushort __ovld ctz(ushort x); -short2 __ovld ctz(short2 x); -ushort2 __ovld ctz(ushort2 x); -short3 __ovld ctz(short3 x); -ushort3 __ovld ctz(ushort3 x); -short4 __ovld ctz(short4 x); -ushort4 __ovld ctz(ushort4 x); -short8 __ovld ctz(short8 x); -ushort8 __ovld ctz(ushort8 x); -short16 __ovld ctz(short16 x); -ushort16 __ovld ctz(ushort16 x); -int __ovld ctz(int x); -uint __ovld ctz(uint x); -int2 __ovld ctz(int2 x); -uint2 __ovld ctz(uint2 x); -int3 __ovld ctz(int3 x); -uint3 __ovld ctz(uint3 x); -int4 __ovld ctz(int4 x); -uint4 __ovld ctz(uint4 x); -int8 __ovld ctz(int8 x); -uint8 __ovld ctz(uint8 x); -int16 __ovld ctz(int16 x); -uint16 __ovld ctz(uint16 x); -long __ovld ctz(long x); -ulong __ovld ctz(ulong x); -long2 __ovld ctz(long2 x); -ulong2 __ovld ctz(ulong2 x); -long3 __ovld ctz(long3 x); -ulong3 __ovld ctz(ulong3 x); -long4 __ovld ctz(long4 x); -ulong4 __ovld ctz(ulong4 x); -long8 __ovld ctz(long8 x); -ulong8 __ovld ctz(ulong8 x); -long16 __ovld ctz(long16 x); -ulong16 __ovld ctz(ulong16 x); +char __ovld __cnfn ctz(char x); +uchar __ovld __cnfn ctz(uchar x); +char2 __ovld __cnfn ctz(char2 x); +uchar2 __ovld __cnfn ctz(uchar2 x); +char3 __ovld __cnfn ctz(char3 x); +uchar3 __ovld __cnfn ctz(uchar3 x); +char4 __ovld __cnfn ctz(char4 x); +uchar4 __ovld __cnfn ctz(uchar4 x); +char8 __ovld __cnfn ctz(char8 x); +uchar8 __ovld __cnfn ctz(uchar8 x); +char16 __ovld __cnfn ctz(char16 x); +uchar16 __ovld __cnfn ctz(uchar16 x); +short __ovld __cnfn ctz(short x); +ushort __ovld __cnfn ctz(ushort x); +short2 __ovld __cnfn ctz(short2 x); +ushort2 __ovld __cnfn ctz(ushort2 x); +short3 __ovld __cnfn ctz(short3 x); +ushort3 __ovld __cnfn ctz(ushort3 x); +short4 __ovld __cnfn ctz(short4 x); +ushort4 __ovld __cnfn ctz(ushort4 x); +short8 __ovld __cnfn ctz(short8 x); +ushort8 __ovld __cnfn ctz(ushort8 x); +short16 __ovld __cnfn ctz(short16 x); +ushort16 __ovld __cnfn ctz(ushort16 x); +int __ovld __cnfn ctz(int x); +uint __ovld __cnfn ctz(uint x); +int2 __ovld __cnfn ctz(int2 x); +uint2 __ovld __cnfn ctz(uint2 x); +int3 __ovld __cnfn ctz(int3 x); +uint3 __ovld __cnfn ctz(uint3 x); +int4 __ovld __cnfn ctz(int4 x); +uint4 __ovld __cnfn ctz(uint4 x); +int8 __ovld __cnfn ctz(int8 x); +uint8 __ovld __cnfn ctz(uint8 x); +int16 __ovld __cnfn ctz(int16 x); +uint16 __ovld __cnfn ctz(uint16 x); +long __ovld __cnfn ctz(long x); +ulong __ovld __cnfn ctz(ulong x); +long2 __ovld __cnfn ctz(long2 x); +ulong2 __ovld __cnfn ctz(ulong2 x); +long3 __ovld __cnfn ctz(long3 x); +ulong3 __ovld __cnfn ctz(ulong3 x); +long4 __ovld __cnfn ctz(long4 x); +ulong4 __ovld __cnfn ctz(ulong4 x); +long8 __ovld __cnfn ctz(long8 x); +ulong8 __ovld __cnfn ctz(ulong8 x); +long16 __ovld __cnfn ctz(long16 x); +ulong16 __ovld __cnfn ctz(ulong16 x); #endif //defined(__OPENCL_CPP_VERSION__) || (__OPENCL_C_VERSION__ >= CL_VERSION_2_0) /** @@ -10002,6 +9907,7 @@ ulong16 __ovld __cnfn upsample(uint16 hi, uint16 lo); /* * popcount(x): returns the number of set bit in x */ +#if defined(__OPENCL_CPP_VERSION__) || (__OPENCL_C_VERSION__ >= CL_VERSION_1_2) char __ovld __cnfn popcount(char x); uchar __ovld __cnfn popcount(uchar x); char2 __ovld __cnfn popcount(char2 x); @@ -10050,6 +9956,7 @@ long8 __ovld __cnfn popcount(long8 x); ulong8 __ovld __cnfn popcount(ulong8 x); long16 __ovld __cnfn popcount(long16 x); ulong16 __ovld __cnfn popcount(ulong16 x); +#endif // defined(__OPENCL_CPP_VERSION__) || (__OPENCL_C_VERSION__ >= CL_VERSION_1_2) /** * Multiply two 24-bit integer values x and y and add @@ -11342,7 +11249,7 @@ half8 __ovld vload8(size_t offset, const __constant half *p); half16 __ovld vload16(size_t offset, const __constant half *p); #endif //cl_khr_fp16 -#if defined(__OPENCL_CPP_VERSION__) || (__OPENCL_C_VERSION__ >= CL_VERSION_2_0) +#if defined(__opencl_c_generic_address_space) char2 __ovld vload2(size_t offset, const char *p); uchar2 __ovld vload2(size_t offset, const uchar *p); short2 __ovld vload2(size_t offset, const short *p); @@ -11580,9 +11487,9 @@ half4 __ovld vload4(size_t offset, const __private half *p); half8 __ovld vload8(size_t offset, const __private half *p); half16 __ovld vload16(size_t offset, const __private half *p); #endif //cl_khr_fp16 -#endif //defined(__OPENCL_CPP_VERSION__) || (__OPENCL_C_VERSION__ >= CL_VERSION_2_0) +#endif //defined(__opencl_c_generic_address_space) -#if defined(__OPENCL_CPP_VERSION__) || (__OPENCL_C_VERSION__ >= CL_VERSION_2_0) +#if defined(__opencl_c_generic_address_space) void __ovld vstore2(char2 data, size_t offset, char *p); void __ovld vstore2(uchar2 data, size_t offset, uchar *p); void __ovld vstore2(short2 data, size_t offset, short *p); @@ -11816,7 +11723,7 @@ void __ovld vstore4(half4 data, size_t offset, __private half *p); void __ovld vstore8(half8 data, size_t offset, __private half *p); void __ovld vstore16(half16 data, size_t offset, __private half *p); #endif //cl_khr_fp16 -#endif //defined(__OPENCL_CPP_VERSION__) || (__OPENCL_C_VERSION__ >= CL_VERSION_2_0) +#endif //defined(__opencl_c_generic_address_space) /** * Read sizeof (half) bytes of data from address @@ -11827,13 +11734,13 @@ void __ovld vstore16(half16 data, size_t offset, __private half *p); * must be 16-bit aligned. */ float __ovld vload_half(size_t offset, const __constant half *p); -#if defined(__OPENCL_CPP_VERSION__) || (__OPENCL_C_VERSION__ >= CL_VERSION_2_0) +#if defined(__opencl_c_generic_address_space) float __ovld vload_half(size_t offset, const half *p); #else float __ovld vload_half(size_t offset, const __global half *p); float __ovld vload_half(size_t offset, const __local half *p); float __ovld vload_half(size_t offset, const __private half *p); -#endif //defined(__OPENCL_CPP_VERSION__) || (__OPENCL_C_VERSION__ >= CL_VERSION_2_0) +#endif //defined(__opencl_c_generic_address_space) /** * Read sizeof (halfn) bytes of data from address @@ -11848,7 +11755,7 @@ float3 __ovld vload_half3(size_t offset, const __constant half *p); float4 __ovld vload_half4(size_t offset, const __constant half *p); float8 __ovld vload_half8(size_t offset, const __constant half *p); float16 __ovld vload_half16(size_t offset, const __constant half *p); -#if defined(__OPENCL_CPP_VERSION__) || (__OPENCL_C_VERSION__ >= CL_VERSION_2_0) +#if defined(__opencl_c_generic_address_space) float2 __ovld vload_half2(size_t offset, const half *p); float3 __ovld vload_half3(size_t offset, const half *p); float4 __ovld vload_half4(size_t offset, const half *p); @@ -11870,7 +11777,7 @@ float3 __ovld vload_half3(size_t offset, const __private half *p); float4 __ovld vload_half4(size_t offset, const __private half *p); float8 __ovld vload_half8(size_t offset, const __private half *p); float16 __ovld vload_half16(size_t offset, const __private half *p); -#endif //defined(__OPENCL_CPP_VERSION__) || (__OPENCL_C_VERSION__ >= CL_VERSION_2_0) +#endif //defined(__opencl_c_generic_address_space) /** * The float value given by data is first @@ -11883,7 +11790,7 @@ float16 __ovld vload_half16(size_t offset, const __private half *p); * The default current rounding mode is round to * nearest even. */ -#if defined(__OPENCL_CPP_VERSION__) || (__OPENCL_C_VERSION__ >= CL_VERSION_2_0) +#if defined(__opencl_c_generic_address_space) void __ovld vstore_half(float data, size_t offset, half *p); void __ovld vstore_half_rte(float data, size_t offset, half *p); void __ovld vstore_half_rtz(float data, size_t offset, half *p); @@ -11929,7 +11836,7 @@ void __ovld vstore_half_rtz(double data, size_t offset, __private half *p); void __ovld vstore_half_rtp(double data, size_t offset, __private half *p); void __ovld vstore_half_rtn(double data, size_t offset, __private half *p); #endif //cl_khr_fp64 -#endif //defined(__OPENCL_CPP_VERSION__) || (__OPENCL_C_VERSION__ >= CL_VERSION_2_0) +#endif //defined(__opencl_c_generic_address_space) /** * The floatn value given by data is converted to @@ -11942,7 +11849,7 @@ void __ovld vstore_half_rtn(double data, size_t offset, __private half *p); * The default current rounding mode is round to * nearest even. */ -#if defined(__OPENCL_CPP_VERSION__) || (__OPENCL_C_VERSION__ >= CL_VERSION_2_0) +#if defined(__opencl_c_generic_address_space) void __ovld vstore_half2(float2 data, size_t offset, half *p); void __ovld vstore_half3(float3 data, size_t offset, half *p); void __ovld vstore_half4(float4 data, size_t offset, half *p); @@ -12148,7 +12055,7 @@ void __ovld vstore_half4_rtn(double4 data, size_t offset, __private half *p); void __ovld vstore_half8_rtn(double8 data, size_t offset, __private half *p); void __ovld vstore_half16_rtn(double16 data, size_t offset, __private half *p); #endif //cl_khr_fp64 -#endif //defined(__OPENCL_CPP_VERSION__) || (__OPENCL_C_VERSION__ >= CL_VERSION_2_0) +#endif //defined(__opencl_c_generic_address_space) /** * For n = 1, 2, 4, 8 and 16 read sizeof (halfn) @@ -12169,7 +12076,7 @@ float3 __ovld vloada_half3(size_t offset, const __constant half *p); float4 __ovld vloada_half4(size_t offset, const __constant half *p); float8 __ovld vloada_half8(size_t offset, const __constant half *p); float16 __ovld vloada_half16(size_t offset, const __constant half *p); -#if defined(__OPENCL_CPP_VERSION__) || (__OPENCL_C_VERSION__ >= CL_VERSION_2_0) +#if defined(__opencl_c_generic_address_space) float __ovld vloada_half(size_t offset, const half *p); float2 __ovld vloada_half2(size_t offset, const half *p); float3 __ovld vloada_half3(size_t offset, const half *p); @@ -12195,7 +12102,7 @@ float3 __ovld vloada_half3(size_t offset, const __private half *p); float4 __ovld vloada_half4(size_t offset, const __private half *p); float8 __ovld vloada_half8(size_t offset, const __private half *p); float16 __ovld vloada_half16(size_t offset, const __private half *p); -#endif //defined(__OPENCL_CPP_VERSION__) || (__OPENCL_C_VERSION__ >= CL_VERSION_2_0) +#endif //defined(__opencl_c_generic_address_space) /** * The floatn value given by data is converted to @@ -12213,7 +12120,7 @@ float16 __ovld vloada_half16(size_t offset, const __private half *p); * mode. The default current rounding mode is * round to nearest even. */ -#if defined(__OPENCL_CPP_VERSION__) || (__OPENCL_C_VERSION__ >= CL_VERSION_2_0) +#if defined(__opencl_c_generic_address_space) void __ovld vstorea_half(float data, size_t offset, half *p); void __ovld vstorea_half2(float2 data, size_t offset, half *p); void __ovld vstorea_half3(float3 data, size_t offset, half *p); @@ -12498,7 +12405,7 @@ void __ovld vstorea_half4_rtn(double4 data,size_t offset, __private half *p); void __ovld vstorea_half8_rtn(double8 data,size_t offset, __private half *p); void __ovld vstorea_half16_rtn(double16 data,size_t offset, __private half *p); #endif //cl_khr_fp64 -#endif //defined(__OPENCL_CPP_VERSION__) || (__OPENCL_C_VERSION__ >= CL_VERSION_2_0) +#endif //defined(__opencl_c_generic_address_space) // OpenCL v1.1 s6.11.8, v1.2 s6.12.8, v2.0 s6.13.8 - Synchronization Functions @@ -12582,7 +12489,7 @@ void __ovld write_mem_fence(cl_mem_fence_flags flags); // OpenCL v2.0 s6.13.9 - Address Space Qualifier Functions -#if defined(__OPENCL_CPP_VERSION__) || (__OPENCL_C_VERSION__ >= CL_VERSION_2_0) +#if defined(__opencl_c_generic_address_space) cl_mem_fence_flags __ovld get_fence(const void *ptr); cl_mem_fence_flags __ovld get_fence(void *ptr); @@ -12593,7 +12500,7 @@ cl_mem_fence_flags __ovld get_fence(void *ptr); * where gentype is builtin type or user defined type. */ -#endif //defined(__OPENCL_CPP_VERSION__) || (__OPENCL_C_VERSION__ >= CL_VERSION_2_0) +#endif //defined(__opencl_c_generic_address_space) // OpenCL v1.1 s6.11.10, v1.2 s6.12.10, v2.0 s6.13.10 - Async Copies from Global to Local Memory, Local to Global Memory, and Prefetch @@ -13397,291 +13304,324 @@ void __ovld atomic_init(volatile atomic_double *object, double value); void __ovld atomic_work_item_fence(cl_mem_fence_flags flags, memory_order order, memory_scope scope); // atomic_fetch() +// OpenCL v2.0 s6.13.11.7.5: +// add/sub: atomic type argument can be uintptr_t/intptr_t, value type argument can be ptrdiff_t. +#if defined(__opencl_c_atomic_order_seq_cst) && defined(__opencl_c_atomic_scope_device) int __ovld atomic_fetch_add(volatile atomic_int *object, int operand); -int __ovld atomic_fetch_add_explicit(volatile atomic_int *object, int operand, memory_order order); -int __ovld atomic_fetch_add_explicit(volatile atomic_int *object, int operand, memory_order order, memory_scope scope); uint __ovld atomic_fetch_add(volatile atomic_uint *object, uint operand); -uint __ovld atomic_fetch_add_explicit(volatile atomic_uint *object, uint operand, memory_order order); -uint __ovld atomic_fetch_add_explicit(volatile atomic_uint *object, uint operand, memory_order order, memory_scope scope); int __ovld atomic_fetch_sub(volatile atomic_int *object, int operand); -int __ovld atomic_fetch_sub_explicit(volatile atomic_int *object, int operand, memory_order order); -int __ovld atomic_fetch_sub_explicit(volatile atomic_int *object, int operand, memory_order order, memory_scope scope); uint __ovld atomic_fetch_sub(volatile atomic_uint *object, uint operand); -uint __ovld atomic_fetch_sub_explicit(volatile atomic_uint *object, uint operand, memory_order order); -uint __ovld atomic_fetch_sub_explicit(volatile atomic_uint *object, uint operand, memory_order order, memory_scope scope); int __ovld atomic_fetch_or(volatile atomic_int *object, int operand); -int __ovld atomic_fetch_or_explicit(volatile atomic_int *object, int operand, memory_order order); -int __ovld atomic_fetch_or_explicit(volatile atomic_int *object, int operand, memory_order order, memory_scope scope); uint __ovld atomic_fetch_or(volatile atomic_uint *object, uint operand); -uint __ovld atomic_fetch_or_explicit(volatile atomic_uint *object, uint operand, memory_order order); -uint __ovld atomic_fetch_or_explicit(volatile atomic_uint *object, uint operand, memory_order order, memory_scope scope); int __ovld atomic_fetch_xor(volatile atomic_int *object, int operand); -int __ovld atomic_fetch_xor_explicit(volatile atomic_int *object, int operand, memory_order order); -int __ovld atomic_fetch_xor_explicit(volatile atomic_int *object, int operand, memory_order order, memory_scope scope); uint __ovld atomic_fetch_xor(volatile atomic_uint *object, uint operand); -uint __ovld atomic_fetch_xor_explicit(volatile atomic_uint *object, uint operand, memory_order order); -uint __ovld atomic_fetch_xor_explicit(volatile atomic_uint *object, uint operand, memory_order order, memory_scope scope); int __ovld atomic_fetch_and(volatile atomic_int *object, int operand); -int __ovld atomic_fetch_and_explicit(volatile atomic_int *object, int operand, memory_order order); -int __ovld atomic_fetch_and_explicit(volatile atomic_int *object, int operand, memory_order order, memory_scope scope); uint __ovld atomic_fetch_and(volatile atomic_uint *object, uint operand); -uint __ovld atomic_fetch_and_explicit(volatile atomic_uint *object, uint operand, memory_order order); -uint __ovld atomic_fetch_and_explicit(volatile atomic_uint *object, uint operand, memory_order order, memory_scope scope); int __ovld atomic_fetch_min(volatile atomic_int *object, int operand); -int __ovld atomic_fetch_min_explicit(volatile atomic_int *object, int operand, memory_order order); -int __ovld atomic_fetch_min_explicit(volatile atomic_int *object, int operand, memory_order order, memory_scope scope); uint __ovld atomic_fetch_min(volatile atomic_uint *object, uint operand); -uint __ovld atomic_fetch_min_explicit(volatile atomic_uint *object, uint operand, memory_order order); -uint __ovld atomic_fetch_min_explicit(volatile atomic_uint *object, uint operand, memory_order order, memory_scope scope); int __ovld atomic_fetch_max(volatile atomic_int *object, int operand); -int __ovld atomic_fetch_max_explicit(volatile atomic_int *object, int operand, memory_order order); -int __ovld atomic_fetch_max_explicit(volatile atomic_int *object, int operand, memory_order order, memory_scope scope); uint __ovld atomic_fetch_max(volatile atomic_uint *object, uint operand); -uint __ovld atomic_fetch_max_explicit(volatile atomic_uint *object, uint operand, memory_order order); -uint __ovld atomic_fetch_max_explicit(volatile atomic_uint *object, uint operand, memory_order order, memory_scope scope); #if defined(cl_khr_int64_base_atomics) && defined(cl_khr_int64_extended_atomics) long __ovld atomic_fetch_add(volatile atomic_long *object, long operand); -long __ovld atomic_fetch_add_explicit(volatile atomic_long *object, long operand, memory_order order); -long __ovld atomic_fetch_add_explicit(volatile atomic_long *object, long operand, memory_order order, memory_scope scope); ulong __ovld atomic_fetch_add(volatile atomic_ulong *object, ulong operand); -ulong __ovld atomic_fetch_add_explicit(volatile atomic_ulong *object, ulong operand, memory_order order); -ulong __ovld atomic_fetch_add_explicit(volatile atomic_ulong *object, ulong operand, memory_order order, memory_scope scope); long __ovld atomic_fetch_sub(volatile atomic_long *object, long operand); -long __ovld atomic_fetch_sub_explicit(volatile atomic_long *object, long operand, memory_order order); -long __ovld atomic_fetch_sub_explicit(volatile atomic_long *object, long operand, memory_order order, memory_scope scope); ulong __ovld atomic_fetch_sub(volatile atomic_ulong *object, ulong operand); -ulong __ovld atomic_fetch_sub_explicit(volatile atomic_ulong *object, ulong operand, memory_order order); -ulong __ovld atomic_fetch_sub_explicit(volatile atomic_ulong *object, ulong operand, memory_order order, memory_scope scope); long __ovld atomic_fetch_or(volatile atomic_long *object, long operand); -long __ovld atomic_fetch_or_explicit(volatile atomic_long *object, long operand, memory_order order); -long __ovld atomic_fetch_or_explicit(volatile atomic_long *object, long operand, memory_order order, memory_scope scope); ulong __ovld atomic_fetch_or(volatile atomic_ulong *object, ulong operand); -ulong __ovld atomic_fetch_or_explicit(volatile atomic_ulong *object, ulong operand, memory_order order); -ulong __ovld atomic_fetch_or_explicit(volatile atomic_ulong *object, ulong operand, memory_order order, memory_scope scope); long __ovld atomic_fetch_xor(volatile atomic_long *object, long operand); -long __ovld atomic_fetch_xor_explicit(volatile atomic_long *object, long operand, memory_order order); -long __ovld atomic_fetch_xor_explicit(volatile atomic_long *object, long operand, memory_order order, memory_scope scope); ulong __ovld atomic_fetch_xor(volatile atomic_ulong *object, ulong operand); -ulong __ovld atomic_fetch_xor_explicit(volatile atomic_ulong *object, ulong operand, memory_order order); -ulong __ovld atomic_fetch_xor_explicit(volatile atomic_ulong *object, ulong operand, memory_order order, memory_scope scope); long __ovld atomic_fetch_and(volatile atomic_long *object, long operand); -long __ovld atomic_fetch_and_explicit(volatile atomic_long *object, long operand, memory_order order); -long __ovld atomic_fetch_and_explicit(volatile atomic_long *object, long operand, memory_order order, memory_scope scope); ulong __ovld atomic_fetch_and(volatile atomic_ulong *object, ulong operand); -ulong __ovld atomic_fetch_and_explicit(volatile atomic_ulong *object, ulong operand, memory_order order); -ulong __ovld atomic_fetch_and_explicit(volatile atomic_ulong *object, ulong operand, memory_order order, memory_scope scope); long __ovld atomic_fetch_min(volatile atomic_long *object, long operand); -long __ovld atomic_fetch_min_explicit(volatile atomic_long *object, long operand, memory_order order); -long __ovld atomic_fetch_min_explicit(volatile atomic_long *object, long operand, memory_order order, memory_scope scope); ulong __ovld atomic_fetch_min(volatile atomic_ulong *object, ulong operand); -ulong __ovld atomic_fetch_min_explicit(volatile atomic_ulong *object, ulong operand, memory_order order); -ulong __ovld atomic_fetch_min_explicit(volatile atomic_ulong *object, ulong operand, memory_order order, memory_scope scope); long __ovld atomic_fetch_max(volatile atomic_long *object, long operand); -long __ovld atomic_fetch_max_explicit(volatile atomic_long *object, long operand, memory_order order); -long __ovld atomic_fetch_max_explicit(volatile atomic_long *object, long operand, memory_order order, memory_scope scope); ulong __ovld atomic_fetch_max(volatile atomic_ulong *object, ulong operand); -ulong __ovld atomic_fetch_max_explicit(volatile atomic_ulong *object, ulong operand, memory_order order); -ulong __ovld atomic_fetch_max_explicit(volatile atomic_ulong *object, ulong operand, memory_order order, memory_scope scope); +uintptr_t __ovld atomic_fetch_add(volatile atomic_uintptr_t *object, ptrdiff_t operand); +uintptr_t __ovld atomic_fetch_sub(volatile atomic_uintptr_t *object, ptrdiff_t operand); #endif //defined(cl_khr_int64_base_atomics) && defined(cl_khr_int64_extended_atomics) +#endif -// OpenCL v2.0 s6.13.11.7.5: -// add/sub: atomic type argument can be uintptr_t/intptr_t, value type argument can be ptrdiff_t. -// or/xor/and/min/max: atomic type argument can be intptr_t/uintptr_t, value type argument can be intptr_t/uintptr_t. - +#if defined(__opencl_c_atomic_scope_device) +int __ovld atomic_fetch_add_explicit(volatile atomic_int *object, int operand, memory_order order); +uint __ovld atomic_fetch_add_explicit(volatile atomic_uint *object, uint operand, memory_order order); +int __ovld atomic_fetch_sub_explicit(volatile atomic_int *object, int operand, memory_order order); +uint __ovld atomic_fetch_sub_explicit(volatile atomic_uint *object, uint operand, memory_order order); +int __ovld atomic_fetch_or_explicit(volatile atomic_int *object, int operand, memory_order order); +uint __ovld atomic_fetch_or_explicit(volatile atomic_uint *object, uint operand, memory_order order); +int __ovld atomic_fetch_xor_explicit(volatile atomic_int *object, int operand, memory_order order); +uint __ovld atomic_fetch_xor_explicit(volatile atomic_uint *object, uint operand, memory_order order); +int __ovld atomic_fetch_and_explicit(volatile atomic_int *object, int operand, memory_order order); +uint __ovld atomic_fetch_and_explicit(volatile atomic_uint *object, uint operand, memory_order order); +int __ovld atomic_fetch_min_explicit(volatile atomic_int *object, int operand, memory_order order); +uint __ovld atomic_fetch_min_explicit(volatile atomic_uint *object, uint operand, memory_order order); +int __ovld atomic_fetch_max_explicit(volatile atomic_int *object, int operand, memory_order order); +uint __ovld atomic_fetch_max_explicit(volatile atomic_uint *object, uint operand, memory_order order); #if defined(cl_khr_int64_base_atomics) && defined(cl_khr_int64_extended_atomics) -uintptr_t __ovld atomic_fetch_add(volatile atomic_uintptr_t *object, ptrdiff_t operand); +long __ovld atomic_fetch_add_explicit(volatile atomic_long *object, long operand, memory_order order); +ulong __ovld atomic_fetch_add_explicit(volatile atomic_ulong *object, ulong operand, memory_order order); +long __ovld atomic_fetch_sub_explicit(volatile atomic_long *object, long operand, memory_order order); +ulong __ovld atomic_fetch_sub_explicit(volatile atomic_ulong *object, ulong operand, memory_order order); +long __ovld atomic_fetch_or_explicit(volatile atomic_long *object, long operand, memory_order order); +ulong __ovld atomic_fetch_or_explicit(volatile atomic_ulong *object, ulong operand, memory_order order); +long __ovld atomic_fetch_xor_explicit(volatile atomic_long *object, long operand, memory_order order); +ulong __ovld atomic_fetch_xor_explicit(volatile atomic_ulong *object, ulong operand, memory_order order); +long __ovld atomic_fetch_and_explicit(volatile atomic_long *object, long operand, memory_order order); +ulong __ovld atomic_fetch_and_explicit(volatile atomic_ulong *object, ulong operand, memory_order order); +long __ovld atomic_fetch_min_explicit(volatile atomic_long *object, long operand, memory_order order); +ulong __ovld atomic_fetch_min_explicit(volatile atomic_ulong *object, ulong operand, memory_order order); +long __ovld atomic_fetch_max_explicit(volatile atomic_long *object, long operand, memory_order order); +ulong __ovld atomic_fetch_max_explicit(volatile atomic_ulong *object, ulong operand, memory_order order); uintptr_t __ovld atomic_fetch_add_explicit(volatile atomic_uintptr_t *object, ptrdiff_t operand, memory_order order); -uintptr_t __ovld atomic_fetch_add_explicit(volatile atomic_uintptr_t *object, ptrdiff_t operand, memory_order order, memory_scope scope); -uintptr_t __ovld atomic_fetch_sub(volatile atomic_uintptr_t *object, ptrdiff_t operand); uintptr_t __ovld atomic_fetch_sub_explicit(volatile atomic_uintptr_t *object, ptrdiff_t operand, memory_order order); -uintptr_t __ovld atomic_fetch_sub_explicit(volatile atomic_uintptr_t *object, ptrdiff_t operand, memory_order order, memory_scope scope); +#endif //defined(cl_khr_int64_base_atomics) && defined(cl_khr_int64_extended_atomics) +#endif -uintptr_t __ovld atomic_fetch_or(volatile atomic_uintptr_t *object, intptr_t operand); -uintptr_t __ovld atomic_fetch_or_explicit(volatile atomic_uintptr_t *object, intptr_t operand, memory_order order); -uintptr_t __ovld atomic_fetch_or_explicit(volatile atomic_uintptr_t *object, intptr_t operand, memory_order order, memory_scope scope); -uintptr_t __ovld atomic_fetch_xor(volatile atomic_uintptr_t *object, intptr_t operand); -uintptr_t __ovld atomic_fetch_xor_explicit(volatile atomic_uintptr_t *object, intptr_t operand, memory_order order); -uintptr_t __ovld atomic_fetch_xor_explicit(volatile atomic_uintptr_t *object, intptr_t operand, memory_order order, memory_scope scope); -uintptr_t __ovld atomic_fetch_and(volatile atomic_uintptr_t *object, intptr_t operand); -uintptr_t __ovld atomic_fetch_and_explicit(volatile atomic_uintptr_t *object, intptr_t operand, memory_order order); -uintptr_t __ovld atomic_fetch_and_explicit(volatile atomic_uintptr_t *object, intptr_t operand, memory_order order, memory_scope scope); -uintptr_t __ovld atomic_fetch_min(volatile atomic_uintptr_t *object, intptr_t opermax); -uintptr_t __ovld atomic_fetch_min_explicit(volatile atomic_uintptr_t *object, intptr_t opermax, memory_order minder); -uintptr_t __ovld atomic_fetch_min_explicit(volatile atomic_uintptr_t *object, intptr_t opermax, memory_order minder, memory_scope scope); -uintptr_t __ovld atomic_fetch_max(volatile atomic_uintptr_t *object, intptr_t opermax); -uintptr_t __ovld atomic_fetch_max_explicit(volatile atomic_uintptr_t *object, intptr_t opermax, memory_order minder); -uintptr_t __ovld atomic_fetch_max_explicit(volatile atomic_uintptr_t *object, intptr_t opermax, memory_order minder, memory_scope scope); - -intptr_t __ovld atomic_fetch_or(volatile atomic_intptr_t *object, uintptr_t operand); -intptr_t __ovld atomic_fetch_or_explicit(volatile atomic_intptr_t *object, uintptr_t operand, memory_order order); -intptr_t __ovld atomic_fetch_or_explicit(volatile atomic_intptr_t *object, uintptr_t operand, memory_order order, memory_scope scope); -intptr_t __ovld atomic_fetch_xor(volatile atomic_intptr_t *object, uintptr_t operand); -intptr_t __ovld atomic_fetch_xor_explicit(volatile atomic_intptr_t *object, uintptr_t operand, memory_order order); -intptr_t __ovld atomic_fetch_xor_explicit(volatile atomic_intptr_t *object, uintptr_t operand, memory_order order, memory_scope scope); -intptr_t __ovld atomic_fetch_and(volatile atomic_intptr_t *object, uintptr_t operand); -intptr_t __ovld atomic_fetch_and_explicit(volatile atomic_intptr_t *object, uintptr_t operand, memory_order order); -intptr_t __ovld atomic_fetch_and_explicit(volatile atomic_intptr_t *object, uintptr_t operand, memory_order order, memory_scope scope); -intptr_t __ovld atomic_fetch_min(volatile atomic_intptr_t *object, uintptr_t opermax); -intptr_t __ovld atomic_fetch_min_explicit(volatile atomic_intptr_t *object, uintptr_t opermax, memory_order minder); -intptr_t __ovld atomic_fetch_min_explicit(volatile atomic_intptr_t *object, uintptr_t opermax, memory_order minder, memory_scope scope); -intptr_t __ovld atomic_fetch_max(volatile atomic_intptr_t *object, uintptr_t opermax); -intptr_t __ovld atomic_fetch_max_explicit(volatile atomic_intptr_t *object, uintptr_t opermax, memory_order minder); -intptr_t __ovld atomic_fetch_max_explicit(volatile atomic_intptr_t *object, uintptr_t opermax, memory_order minder, memory_scope scope); +int __ovld atomic_fetch_add_explicit(volatile atomic_int *object, int operand, memory_order order, memory_scope scope); +uint __ovld atomic_fetch_add_explicit(volatile atomic_uint *object, uint operand, memory_order order, memory_scope scope); +int __ovld atomic_fetch_sub_explicit(volatile atomic_int *object, int operand, memory_order order, memory_scope scope); +uint __ovld atomic_fetch_sub_explicit(volatile atomic_uint *object, uint operand, memory_order order, memory_scope scope); +int __ovld atomic_fetch_or_explicit(volatile atomic_int *object, int operand, memory_order order, memory_scope scope); +uint __ovld atomic_fetch_or_explicit(volatile atomic_uint *object, uint operand, memory_order order, memory_scope scope); +int __ovld atomic_fetch_xor_explicit(volatile atomic_int *object, int operand, memory_order order, memory_scope scope); +uint __ovld atomic_fetch_xor_explicit(volatile atomic_uint *object, uint operand, memory_order order, memory_scope scope); +int __ovld atomic_fetch_and_explicit(volatile atomic_int *object, int operand, memory_order order, memory_scope scope); +uint __ovld atomic_fetch_and_explicit(volatile atomic_uint *object, uint operand, memory_order order, memory_scope scope); +int __ovld atomic_fetch_min_explicit(volatile atomic_int *object, int operand, memory_order order, memory_scope scope); +uint __ovld atomic_fetch_min_explicit(volatile atomic_uint *object, uint operand, memory_order order, memory_scope scope); +int __ovld atomic_fetch_max_explicit(volatile atomic_int *object, int operand, memory_order order, memory_scope scope); +uint __ovld atomic_fetch_max_explicit(volatile atomic_uint *object, uint operand, memory_order order, memory_scope scope); +#if defined(cl_khr_int64_base_atomics) && defined(cl_khr_int64_extended_atomics) +long __ovld atomic_fetch_add_explicit(volatile atomic_long *object, long operand, memory_order order, memory_scope scope); +ulong __ovld atomic_fetch_add_explicit(volatile atomic_ulong *object, ulong operand, memory_order order, memory_scope scope); +long __ovld atomic_fetch_sub_explicit(volatile atomic_long *object, long operand, memory_order order, memory_scope scope); +ulong __ovld atomic_fetch_sub_explicit(volatile atomic_ulong *object, ulong operand, memory_order order, memory_scope scope); +long __ovld atomic_fetch_or_explicit(volatile atomic_long *object, long operand, memory_order order, memory_scope scope); +ulong __ovld atomic_fetch_or_explicit(volatile atomic_ulong *object, ulong operand, memory_order order, memory_scope scope); +long __ovld atomic_fetch_xor_explicit(volatile atomic_long *object, long operand, memory_order order, memory_scope scope); +ulong __ovld atomic_fetch_xor_explicit(volatile atomic_ulong *object, ulong operand, memory_order order, memory_scope scope); +long __ovld atomic_fetch_and_explicit(volatile atomic_long *object, long operand, memory_order order, memory_scope scope); +ulong __ovld atomic_fetch_and_explicit(volatile atomic_ulong *object, ulong operand, memory_order order, memory_scope scope); +long __ovld atomic_fetch_min_explicit(volatile atomic_long *object, long operand, memory_order order, memory_scope scope); +ulong __ovld atomic_fetch_min_explicit(volatile atomic_ulong *object, ulong operand, memory_order order, memory_scope scope); +long __ovld atomic_fetch_max_explicit(volatile atomic_long *object, long operand, memory_order order, memory_scope scope); +ulong __ovld atomic_fetch_max_explicit(volatile atomic_ulong *object, ulong operand, memory_order order, memory_scope scope); +#endif //defined(cl_khr_int64_base_atomics) && defined(cl_khr_int64_extended_atomics) +#if defined(cl_khr_int64_base_atomics) && defined(cl_khr_int64_extended_atomics) +uintptr_t __ovld atomic_fetch_add_explicit(volatile atomic_uintptr_t *object, ptrdiff_t operand, memory_order order, memory_scope scope); +uintptr_t __ovld atomic_fetch_sub_explicit(volatile atomic_uintptr_t *object, ptrdiff_t operand, memory_order order, memory_scope scope); #endif // atomic_store() +#if defined(__opencl_c_atomic_order_seq_cst) && defined(__opencl_c_atomic_scope_device) void __ovld atomic_store(volatile atomic_int *object, int desired); -void __ovld atomic_store_explicit(volatile atomic_int *object, int desired, memory_order order); -void __ovld atomic_store_explicit(volatile atomic_int *object, int desired, memory_order order, memory_scope scope); void __ovld atomic_store(volatile atomic_uint *object, uint desired); -void __ovld atomic_store_explicit(volatile atomic_uint *object, uint desired, memory_order order); -void __ovld atomic_store_explicit(volatile atomic_uint *object, uint desired, memory_order order, memory_scope scope); void __ovld atomic_store(volatile atomic_float *object, float desired); -void __ovld atomic_store_explicit(volatile atomic_float *object, float desired, memory_order order); -void __ovld atomic_store_explicit(volatile atomic_float *object, float desired, memory_order order, memory_scope scope); + #if defined(cl_khr_int64_base_atomics) && defined(cl_khr_int64_extended_atomics) #ifdef cl_khr_fp64 void __ovld atomic_store(volatile atomic_double *object, double desired); -void __ovld atomic_store_explicit(volatile atomic_double *object, double desired, memory_order order); -void __ovld atomic_store_explicit(volatile atomic_double *object, double desired, memory_order order, memory_scope scope); #endif //cl_khr_fp64 void __ovld atomic_store(volatile atomic_long *object, long desired); -void __ovld atomic_store_explicit(volatile atomic_long *object, long desired, memory_order order); -void __ovld atomic_store_explicit(volatile atomic_long *object, long desired, memory_order order, memory_scope scope); void __ovld atomic_store(volatile atomic_ulong *object, ulong desired); +#endif +#endif + +#if defined(__opencl_c_atomic_scope_device) +void __ovld atomic_store_explicit(volatile atomic_int *object, int desired, memory_order order); +void __ovld atomic_store_explicit(volatile atomic_uint *object, uint desired, memory_order order); +void __ovld atomic_store_explicit(volatile atomic_float *object, float desired, memory_order order); +#if defined(cl_khr_int64_base_atomics) && defined(cl_khr_int64_extended_atomics) +#ifdef cl_khr_fp64 +void __ovld atomic_store_explicit(volatile atomic_double *object, double desired, memory_order order); +#endif //cl_khr_fp64 +void __ovld atomic_store_explicit(volatile atomic_long *object, long desired, memory_order order); void __ovld atomic_store_explicit(volatile atomic_ulong *object, ulong desired, memory_order order); +#endif +#endif + +void __ovld atomic_store_explicit(volatile atomic_int *object, int desired, memory_order order, memory_scope scope); +void __ovld atomic_store_explicit(volatile atomic_uint *object, uint desired, memory_order order, memory_scope scope); +void __ovld atomic_store_explicit(volatile atomic_float *object, float desired, memory_order order, memory_scope scope); +#if defined(cl_khr_int64_base_atomics) && defined(cl_khr_int64_extended_atomics) +#ifdef cl_khr_fp64 +void __ovld atomic_store_explicit(volatile atomic_double *object, double desired, memory_order order, memory_scope scope); +#endif //cl_khr_fp64 +void __ovld atomic_store_explicit(volatile atomic_long *object, long desired, memory_order order, memory_scope scope); void __ovld atomic_store_explicit(volatile atomic_ulong *object, ulong desired, memory_order order, memory_scope scope); #endif // atomic_load() - +#if defined(__opencl_c_atomic_order_seq_cst) && defined(__opencl_c_atomic_scope_device) int __ovld atomic_load(volatile atomic_int *object); -int __ovld atomic_load_explicit(volatile atomic_int *object, memory_order order); -int __ovld atomic_load_explicit(volatile atomic_int *object, memory_order order, memory_scope scope); uint __ovld atomic_load(volatile atomic_uint *object); -uint __ovld atomic_load_explicit(volatile atomic_uint *object, memory_order order); -uint __ovld atomic_load_explicit(volatile atomic_uint *object, memory_order order, memory_scope scope); float __ovld atomic_load(volatile atomic_float *object); -float __ovld atomic_load_explicit(volatile atomic_float *object, memory_order order); -float __ovld atomic_load_explicit(volatile atomic_float *object, memory_order order, memory_scope scope); #if defined(cl_khr_int64_base_atomics) && defined(cl_khr_int64_extended_atomics) #ifdef cl_khr_fp64 double __ovld atomic_load(volatile atomic_double *object); -double __ovld atomic_load_explicit(volatile atomic_double *object, memory_order order); -double __ovld atomic_load_explicit(volatile atomic_double *object, memory_order order, memory_scope scope); #endif //cl_khr_fp64 long __ovld atomic_load(volatile atomic_long *object); -long __ovld atomic_load_explicit(volatile atomic_long *object, memory_order order); -long __ovld atomic_load_explicit(volatile atomic_long *object, memory_order order, memory_scope scope); ulong __ovld atomic_load(volatile atomic_ulong *object); +#endif +#endif + +#if defined(__opencl_c_atomic_scope_device) +int __ovld atomic_load_explicit(volatile atomic_int *object, memory_order order); +uint __ovld atomic_load_explicit(volatile atomic_uint *object, memory_order order); +float __ovld atomic_load_explicit(volatile atomic_float *object, memory_order order); +#if defined(cl_khr_int64_base_atomics) && defined(cl_khr_int64_extended_atomics) +#ifdef cl_khr_fp64 +double __ovld atomic_load_explicit(volatile atomic_double *object, memory_order order); +#endif //cl_khr_fp64 +long __ovld atomic_load_explicit(volatile atomic_long *object, memory_order order); ulong __ovld atomic_load_explicit(volatile atomic_ulong *object, memory_order order); +#endif +#endif + +int __ovld atomic_load_explicit(volatile atomic_int *object, memory_order order, memory_scope scope); +uint __ovld atomic_load_explicit(volatile atomic_uint *object, memory_order order, memory_scope scope); +float __ovld atomic_load_explicit(volatile atomic_float *object, memory_order order, memory_scope scope); +#if defined(cl_khr_int64_base_atomics) && defined(cl_khr_int64_extended_atomics) +#ifdef cl_khr_fp64 +double __ovld atomic_load_explicit(volatile atomic_double *object, memory_order order, memory_scope scope); +#endif //cl_khr_fp64 +long __ovld atomic_load_explicit(volatile atomic_long *object, memory_order order, memory_scope scope); ulong __ovld atomic_load_explicit(volatile atomic_ulong *object, memory_order order, memory_scope scope); #endif // atomic_exchange() +#if defined(__opencl_c_atomic_order_seq_cst) && defined(__opencl_c_atomic_scope_device) int __ovld atomic_exchange(volatile atomic_int *object, int desired); -int __ovld atomic_exchange_explicit(volatile atomic_int *object, int desired, memory_order order); -int __ovld atomic_exchange_explicit(volatile atomic_int *object, int desired, memory_order order, memory_scope scope); uint __ovld atomic_exchange(volatile atomic_uint *object, uint desired); -uint __ovld atomic_exchange_explicit(volatile atomic_uint *object, uint desired, memory_order order); -uint __ovld atomic_exchange_explicit(volatile atomic_uint *object, uint desired, memory_order order, memory_scope scope); float __ovld atomic_exchange(volatile atomic_float *object, float desired); -float __ovld atomic_exchange_explicit(volatile atomic_float *object, float desired, memory_order order); -float __ovld atomic_exchange_explicit(volatile atomic_float *object, float desired, memory_order order, memory_scope scope); #if defined(cl_khr_int64_base_atomics) && defined(cl_khr_int64_extended_atomics) #ifdef cl_khr_fp64 double __ovld atomic_exchange(volatile atomic_double *object, double desired); -double __ovld atomic_exchange_explicit(volatile atomic_double *object, double desired, memory_order order); -double __ovld atomic_exchange_explicit(volatile atomic_double *object, double desired, memory_order order, memory_scope scope); #endif //cl_khr_fp64 long __ovld atomic_exchange(volatile atomic_long *object, long desired); -long __ovld atomic_exchange_explicit(volatile atomic_long *object, long desired, memory_order order); -long __ovld atomic_exchange_explicit(volatile atomic_long *object, long desired, memory_order order, memory_scope scope); ulong __ovld atomic_exchange(volatile atomic_ulong *object, ulong desired); +#endif +#endif + +#if defined(__opencl_c_atomic_scope_device) +int __ovld atomic_exchange_explicit(volatile atomic_int *object, int desired, memory_order order); +uint __ovld atomic_exchange_explicit(volatile atomic_uint *object, uint desired, memory_order order); +float __ovld atomic_exchange_explicit(volatile atomic_float *object, float desired, memory_order order); +#if defined(cl_khr_int64_base_atomics) && defined(cl_khr_int64_extended_atomics) +#ifdef cl_khr_fp64 +double __ovld atomic_exchange_explicit(volatile atomic_double *object, double desired, memory_order order); +#endif //cl_khr_fp64 +long __ovld atomic_exchange_explicit(volatile atomic_long *object, long desired, memory_order order); ulong __ovld atomic_exchange_explicit(volatile atomic_ulong *object, ulong desired, memory_order order); +#endif +#endif + +int __ovld atomic_exchange_explicit(volatile atomic_int *object, int desired, memory_order order, memory_scope scope); +uint __ovld atomic_exchange_explicit(volatile atomic_uint *object, uint desired, memory_order order, memory_scope scope); +float __ovld atomic_exchange_explicit(volatile atomic_float *object, float desired, memory_order order, memory_scope scope); +#if defined(cl_khr_int64_base_atomics) && defined(cl_khr_int64_extended_atomics) +#ifdef cl_khr_fp64 +double __ovld atomic_exchange_explicit(volatile atomic_double *object, double desired, memory_order order, memory_scope scope); +#endif //cl_khr_fp64 +long __ovld atomic_exchange_explicit(volatile atomic_long *object, long desired, memory_order order, memory_scope scope); ulong __ovld atomic_exchange_explicit(volatile atomic_ulong *object, ulong desired, memory_order order, memory_scope scope); #endif // atomic_compare_exchange_strong() and atomic_compare_exchange_weak() - +#if defined(__opencl_c_atomic_order_seq_cst) && defined(__opencl_c_atomic_scope_device) bool __ovld atomic_compare_exchange_strong(volatile atomic_int *object, int *expected, int desired); +bool __ovld atomic_compare_exchange_strong(volatile atomic_uint *object, uint *expected, uint desired); +bool __ovld atomic_compare_exchange_weak(volatile atomic_int *object, int *expected, int desired); +bool __ovld atomic_compare_exchange_weak(volatile atomic_uint *object, uint *expected, uint desired); +bool __ovld atomic_compare_exchange_strong(volatile atomic_float *object, float *expected, float desired); +bool __ovld atomic_compare_exchange_weak(volatile atomic_float *object, float *expected, float desired); + +#if defined(cl_khr_int64_base_atomics) && defined(cl_khr_int64_extended_atomics) +#ifdef cl_khr_fp64 +bool __ovld atomic_compare_exchange_strong(volatile atomic_double *object, double *expected, double desired); +bool __ovld atomic_compare_exchange_weak(volatile atomic_double *object, double *expected, double desired); +#endif //cl_khr_fp64 +bool __ovld atomic_compare_exchange_strong(volatile atomic_long *object, long *expected, long desired); +bool __ovld atomic_compare_exchange_weak(volatile atomic_long *object, long *expected, long desired); +bool __ovld atomic_compare_exchange_strong(volatile atomic_ulong *object, ulong *expected, ulong desired); +bool __ovld atomic_compare_exchange_weak(volatile atomic_ulong *object, ulong *expected, ulong desired); +#endif +#endif + bool __ovld atomic_compare_exchange_strong_explicit(volatile atomic_int *object, int *expected, int desired, memory_order success, memory_order failure); -bool __ovld atomic_compare_exchange_strong_explicit(volatile atomic_int *object, int *expected, - int desired, memory_order success, memory_order failure, memory_scope scope); -bool __ovld atomic_compare_exchange_strong(volatile atomic_uint *object, uint *expected, uint desired); bool __ovld atomic_compare_exchange_strong_explicit(volatile atomic_uint *object, uint *expected, uint desired, memory_order success, memory_order failure); -bool __ovld atomic_compare_exchange_strong_explicit(volatile atomic_uint *object, uint *expected, - uint desired, memory_order success, memory_order failure, memory_scope scope); -bool __ovld atomic_compare_exchange_weak(volatile atomic_int *object, int *expected, int desired); bool __ovld atomic_compare_exchange_weak_explicit(volatile atomic_int *object, int *expected, int desired, memory_order success, memory_order failure); -bool __ovld atomic_compare_exchange_weak_explicit(volatile atomic_int *object, int *expected, - int desired, memory_order success, memory_order failure, memory_scope scope); -bool __ovld atomic_compare_exchange_weak(volatile atomic_uint *object, uint *expected, uint desired); bool __ovld atomic_compare_exchange_weak_explicit(volatile atomic_uint *object, uint *expected, uint desired, memory_order success, memory_order failure); -bool __ovld atomic_compare_exchange_weak_explicit(volatile atomic_uint *object, uint *expected, - uint desired, memory_order success, memory_order failure, memory_scope scope); -bool __ovld atomic_compare_exchange_strong(volatile atomic_float *object, float *expected, float desired); bool __ovld atomic_compare_exchange_strong_explicit(volatile atomic_float *object, float *expected, float desired, memory_order success, memory_order failure); -bool __ovld atomic_compare_exchange_strong_explicit(volatile atomic_float *object, float *expected, - float desired, memory_order success, memory_order failure, memory_scope scope); -bool __ovld atomic_compare_exchange_weak(volatile atomic_float *object, float *expected, float desired); bool __ovld atomic_compare_exchange_weak_explicit(volatile atomic_float *object, float *expected, float desired, memory_order success, memory_order failure); -bool __ovld atomic_compare_exchange_weak_explicit(volatile atomic_float *object, float *expected, - float desired, memory_order success, memory_order failure, memory_scope scope); #if defined(cl_khr_int64_base_atomics) && defined(cl_khr_int64_extended_atomics) #ifdef cl_khr_fp64 -bool __ovld atomic_compare_exchange_strong(volatile atomic_double *object, double *expected, double desired); bool __ovld atomic_compare_exchange_strong_explicit(volatile atomic_double *object, double *expected, double desired, memory_order success, memory_order failure); -bool __ovld atomic_compare_exchange_strong_explicit(volatile atomic_double *object, double *expected, - double desired, memory_order success, memory_order failure, memory_scope scope); -bool __ovld atomic_compare_exchange_weak(volatile atomic_double *object, double *expected, double desired); bool __ovld atomic_compare_exchange_weak_explicit(volatile atomic_double *object, double *expected, double desired, memory_order success, memory_order failure); -bool __ovld atomic_compare_exchange_weak_explicit(volatile atomic_double *object, double *expected, - double desired, memory_order success, memory_order failure, memory_scope scope); #endif //cl_khr_fp64 -bool __ovld atomic_compare_exchange_strong(volatile atomic_long *object, long *expected, long desired); bool __ovld atomic_compare_exchange_strong_explicit(volatile atomic_long *object, long *expected, long desired, memory_order success, memory_order failure); -bool __ovld atomic_compare_exchange_strong_explicit(volatile atomic_long *object, long *expected, - long desired, memory_order success, memory_order failure, memory_scope scope); -bool __ovld atomic_compare_exchange_weak(volatile atomic_long *object, long *expected, long desired); bool __ovld atomic_compare_exchange_weak_explicit(volatile atomic_long *object, long *expected, long desired, memory_order success, memory_order failure); -bool __ovld atomic_compare_exchange_weak_explicit(volatile atomic_long *object, long *expected, - long desired, memory_order success, memory_order failure, memory_scope scope); -bool __ovld atomic_compare_exchange_strong(volatile atomic_ulong *object, ulong *expected, ulong desired); bool __ovld atomic_compare_exchange_strong_explicit(volatile atomic_ulong *object, ulong *expected, ulong desired, memory_order success, memory_order failure); -bool __ovld atomic_compare_exchange_strong_explicit(volatile atomic_ulong *object, ulong *expected, - ulong desired, memory_order success, memory_order failure, memory_scope scope); -bool __ovld atomic_compare_exchange_weak(volatile atomic_ulong *object, ulong *expected, ulong desired); bool __ovld atomic_compare_exchange_weak_explicit(volatile atomic_ulong *object, ulong *expected, ulong desired, memory_order success, memory_order failure); +#endif + +bool __ovld atomic_compare_exchange_strong_explicit(volatile atomic_int *object, int *expected, + int desired, memory_order success, memory_order failure, memory_scope scope); +bool __ovld atomic_compare_exchange_strong_explicit(volatile atomic_uint *object, uint *expected, + uint desired, memory_order success, memory_order failure, memory_scope scope); +bool __ovld atomic_compare_exchange_weak_explicit(volatile atomic_int *object, int *expected, + int desired, memory_order success, memory_order failure, memory_scope scope); +bool __ovld atomic_compare_exchange_weak_explicit(volatile atomic_uint *object, uint *expected, + uint desired, memory_order success, memory_order failure, memory_scope scope); +bool __ovld atomic_compare_exchange_strong_explicit(volatile atomic_float *object, float *expected, + float desired, memory_order success, memory_order failure, memory_scope scope); +bool __ovld atomic_compare_exchange_weak_explicit(volatile atomic_float *object, float *expected, + float desired, memory_order success, memory_order failure, memory_scope scope); +#if defined(cl_khr_int64_base_atomics) && defined(cl_khr_int64_extended_atomics) +#ifdef cl_khr_fp64 +bool __ovld atomic_compare_exchange_strong_explicit(volatile atomic_double *object, double *expected, + double desired, memory_order success, memory_order failure, memory_scope scope); +bool __ovld atomic_compare_exchange_weak_explicit(volatile atomic_double *object, double *expected, + double desired, memory_order success, memory_order failure, memory_scope scope); +#endif //cl_khr_fp64 +bool __ovld atomic_compare_exchange_strong_explicit(volatile atomic_long *object, long *expected, + long desired, memory_order success, memory_order failure, memory_scope scope); +bool __ovld atomic_compare_exchange_weak_explicit(volatile atomic_long *object, long *expected, + long desired, memory_order success, memory_order failure, memory_scope scope); +bool __ovld atomic_compare_exchange_strong_explicit(volatile atomic_ulong *object, ulong *expected, + ulong desired, memory_order success, memory_order failure, memory_scope scope); bool __ovld atomic_compare_exchange_weak_explicit(volatile atomic_ulong *object, ulong *expected, ulong desired, memory_order success, memory_order failure, memory_scope scope); #endif // atomic_flag_test_and_set() and atomic_flag_clear() - +#if defined(__opencl_c_atomic_order_seq_cst) && defined(__opencl_c_atomic_scope_device) bool __ovld atomic_flag_test_and_set(volatile atomic_flag *object); -bool __ovld atomic_flag_test_and_set_explicit(volatile atomic_flag *object, memory_order order); -bool __ovld atomic_flag_test_and_set_explicit(volatile atomic_flag *object, memory_order order, memory_scope scope); void __ovld atomic_flag_clear(volatile atomic_flag *object); +#endif + +#if defined(__opencl_c_atomic_scope_device) +bool __ovld atomic_flag_test_and_set_explicit(volatile atomic_flag *object, memory_order order); void __ovld atomic_flag_clear_explicit(volatile atomic_flag *object, memory_order order); -void __ovld atomic_flag_clear_explicit(volatile atomic_flag *object, memory_order order, memory_scope scope); +#endif +bool __ovld atomic_flag_test_and_set_explicit(volatile atomic_flag *object, memory_order order, memory_scope scope); +void __ovld atomic_flag_clear_explicit(volatile atomic_flag *object, memory_order order, memory_scope scope); #endif //defined(__OPENCL_CPP_VERSION__) || (__OPENCL_C_VERSION__ >= CL_VERSION_2_0) // OpenCL v1.1 s6.11.12, v1.2 s6.12.12, v2.0 s6.13.12 - Miscellaneous Vector Functions @@ -14176,12 +14116,6 @@ half16 __ovld __cnfn shuffle2(half8 x, half8 y, ushort16 mask); half16 __ovld __cnfn shuffle2(half16 x, half16 y, ushort16 mask); #endif //cl_khr_fp16 -#if defined(__OPENCL_CPP_VERSION__) || (__OPENCL_C_VERSION__ >= CL_VERSION_1_2) -// OpenCL v1.2 s6.12.13, v2.0 s6.13.13 - printf - -int printf(__constant const char* st, ...) __attribute__((format(printf, 1, 2))); -#endif - // OpenCL v1.1 s6.11.3, v1.2 s6.12.14, v2.0 s6.13.14 - Image Read and Write Functions #ifdef cl_khr_gl_msaa_sharing @@ -16130,6 +16064,230 @@ double __ovld sub_group_clustered_reduce_max( double value, uint clustersize ); #endif // cl_khr_subgroup_clustered_reduce +#if defined(cl_khr_extended_bit_ops) +char __ovld __cnfn bitfield_insert(char, char, uint, uint); +uchar __ovld __cnfn bitfield_insert(uchar, uchar, uint, uint); +short __ovld __cnfn bitfield_insert(short, short, uint, uint); +ushort __ovld __cnfn bitfield_insert(ushort, ushort, uint, uint); +int __ovld __cnfn bitfield_insert(int, int, uint, uint); +uint __ovld __cnfn bitfield_insert(uint, uint, uint, uint); +long __ovld __cnfn bitfield_insert(long, long, uint, uint); +ulong __ovld __cnfn bitfield_insert(ulong, ulong, uint, uint); +char2 __ovld __cnfn bitfield_insert(char2, char2, uint, uint); +uchar2 __ovld __cnfn bitfield_insert(uchar2, uchar2, uint, uint); +short2 __ovld __cnfn bitfield_insert(short2, short2, uint, uint); +ushort2 __ovld __cnfn bitfield_insert(ushort2, ushort2, uint, uint); +int2 __ovld __cnfn bitfield_insert(int2, int2, uint, uint); +uint2 __ovld __cnfn bitfield_insert(uint2, uint2, uint, uint); +long2 __ovld __cnfn bitfield_insert(long2, long2, uint, uint); +ulong2 __ovld __cnfn bitfield_insert(ulong2, ulong2, uint, uint); +char3 __ovld __cnfn bitfield_insert(char3, char3, uint, uint); +uchar3 __ovld __cnfn bitfield_insert(uchar3, uchar3, uint, uint); +short3 __ovld __cnfn bitfield_insert(short3, short3, uint, uint); +ushort3 __ovld __cnfn bitfield_insert(ushort3, ushort3, uint, uint); +int3 __ovld __cnfn bitfield_insert(int3, int3, uint, uint); +uint3 __ovld __cnfn bitfield_insert(uint3, uint3, uint, uint); +long3 __ovld __cnfn bitfield_insert(long3, long3, uint, uint); +ulong3 __ovld __cnfn bitfield_insert(ulong3, ulong3, uint, uint); +char4 __ovld __cnfn bitfield_insert(char4, char4, uint, uint); +uchar4 __ovld __cnfn bitfield_insert(uchar4, uchar4, uint, uint); +short4 __ovld __cnfn bitfield_insert(short4, short4, uint, uint); +ushort4 __ovld __cnfn bitfield_insert(ushort4, ushort4, uint, uint); +int4 __ovld __cnfn bitfield_insert(int4, int4, uint, uint); +uint4 __ovld __cnfn bitfield_insert(uint4, uint4, uint, uint); +long4 __ovld __cnfn bitfield_insert(long4, long4, uint, uint); +ulong4 __ovld __cnfn bitfield_insert(ulong4, ulong4, uint, uint); +char8 __ovld __cnfn bitfield_insert(char8, char8, uint, uint); +uchar8 __ovld __cnfn bitfield_insert(uchar8, uchar8, uint, uint); +short8 __ovld __cnfn bitfield_insert(short8, short8, uint, uint); +ushort8 __ovld __cnfn bitfield_insert(ushort8, ushort8, uint, uint); +int8 __ovld __cnfn bitfield_insert(int8, int8, uint, uint); +uint8 __ovld __cnfn bitfield_insert(uint8, uint8, uint, uint); +long8 __ovld __cnfn bitfield_insert(long8, long8, uint, uint); +ulong8 __ovld __cnfn bitfield_insert(ulong8, ulong8, uint, uint); +char16 __ovld __cnfn bitfield_insert(char16, char16, uint, uint); +uchar16 __ovld __cnfn bitfield_insert(uchar16, uchar16, uint, uint); +short16 __ovld __cnfn bitfield_insert(short16, short16, uint, uint); +ushort16 __ovld __cnfn bitfield_insert(ushort16, ushort16, uint, uint); +int16 __ovld __cnfn bitfield_insert(int16, int16, uint, uint); +uint16 __ovld __cnfn bitfield_insert(uint16, uint16, uint, uint); +long16 __ovld __cnfn bitfield_insert(long16, long16, uint, uint); +ulong16 __ovld __cnfn bitfield_insert(ulong16, ulong16, uint, uint); + +char __ovld __cnfn bitfield_extract_signed(char, uint, uint); +short __ovld __cnfn bitfield_extract_signed(short, uint, uint); +int __ovld __cnfn bitfield_extract_signed(int, uint, uint); +long __ovld __cnfn bitfield_extract_signed(long, uint, uint); +char2 __ovld __cnfn bitfield_extract_signed(char2, uint, uint); +short2 __ovld __cnfn bitfield_extract_signed(short2, uint, uint); +int2 __ovld __cnfn bitfield_extract_signed(int2, uint, uint); +long2 __ovld __cnfn bitfield_extract_signed(long2, uint, uint); +char3 __ovld __cnfn bitfield_extract_signed(char3, uint, uint); +short3 __ovld __cnfn bitfield_extract_signed(short3, uint, uint); +int3 __ovld __cnfn bitfield_extract_signed(int3, uint, uint); +long3 __ovld __cnfn bitfield_extract_signed(long3, uint, uint); +char4 __ovld __cnfn bitfield_extract_signed(char4, uint, uint); +short4 __ovld __cnfn bitfield_extract_signed(short4, uint, uint); +int4 __ovld __cnfn bitfield_extract_signed(int4, uint, uint); +long4 __ovld __cnfn bitfield_extract_signed(long4, uint, uint); +char8 __ovld __cnfn bitfield_extract_signed(char8, uint, uint); +short8 __ovld __cnfn bitfield_extract_signed(short8, uint, uint); +int8 __ovld __cnfn bitfield_extract_signed(int8, uint, uint); +long8 __ovld __cnfn bitfield_extract_signed(long8, uint, uint); +char16 __ovld __cnfn bitfield_extract_signed(char16, uint, uint); +short16 __ovld __cnfn bitfield_extract_signed(short16, uint, uint); +int16 __ovld __cnfn bitfield_extract_signed(int16, uint, uint); +long16 __ovld __cnfn bitfield_extract_signed(long16, uint, uint); + +char __ovld __cnfn bitfield_extract_signed(uchar, uint, uint); +short __ovld __cnfn bitfield_extract_signed(ushort, uint, uint); +int __ovld __cnfn bitfield_extract_signed(uint, uint, uint); +long __ovld __cnfn bitfield_extract_signed(ulong, uint, uint); +char2 __ovld __cnfn bitfield_extract_signed(uchar2, uint, uint); +short2 __ovld __cnfn bitfield_extract_signed(ushort2, uint, uint); +int2 __ovld __cnfn bitfield_extract_signed(uint2, uint, uint); +long2 __ovld __cnfn bitfield_extract_signed(ulong2, uint, uint); +char3 __ovld __cnfn bitfield_extract_signed(uchar3, uint, uint); +short3 __ovld __cnfn bitfield_extract_signed(ushort3, uint, uint); +int3 __ovld __cnfn bitfield_extract_signed(uint3, uint, uint); +long3 __ovld __cnfn bitfield_extract_signed(ulong3, uint, uint); +char4 __ovld __cnfn bitfield_extract_signed(uchar4, uint, uint); +short4 __ovld __cnfn bitfield_extract_signed(ushort4, uint, uint); +int4 __ovld __cnfn bitfield_extract_signed(uint4, uint, uint); +long4 __ovld __cnfn bitfield_extract_signed(ulong4, uint, uint); +char8 __ovld __cnfn bitfield_extract_signed(uchar8, uint, uint); +short8 __ovld __cnfn bitfield_extract_signed(ushort8, uint, uint); +int8 __ovld __cnfn bitfield_extract_signed(uint8, uint, uint); +long8 __ovld __cnfn bitfield_extract_signed(ulong8, uint, uint); +char16 __ovld __cnfn bitfield_extract_signed(uchar16, uint, uint); +short16 __ovld __cnfn bitfield_extract_signed(ushort16, uint, uint); +int16 __ovld __cnfn bitfield_extract_signed(uint16, uint, uint); +long16 __ovld __cnfn bitfield_extract_signed(ulong16, uint, uint); + +uchar __ovld __cnfn bitfield_extract_unsigned(char, uint, uint); +ushort __ovld __cnfn bitfield_extract_unsigned(short, uint, uint); +uint __ovld __cnfn bitfield_extract_unsigned(int, uint, uint); +ulong __ovld __cnfn bitfield_extract_unsigned(long, uint, uint); +uchar2 __ovld __cnfn bitfield_extract_unsigned(char2, uint, uint); +ushort2 __ovld __cnfn bitfield_extract_unsigned(short2, uint, uint); +uint2 __ovld __cnfn bitfield_extract_unsigned(int2, uint, uint); +ulong2 __ovld __cnfn bitfield_extract_unsigned(long2, uint, uint); +uchar3 __ovld __cnfn bitfield_extract_unsigned(char3, uint, uint); +ushort3 __ovld __cnfn bitfield_extract_unsigned(short3, uint, uint); +uint3 __ovld __cnfn bitfield_extract_unsigned(int3, uint, uint); +ulong3 __ovld __cnfn bitfield_extract_unsigned(long3, uint, uint); +uchar4 __ovld __cnfn bitfield_extract_unsigned(char4, uint, uint); +ushort4 __ovld __cnfn bitfield_extract_unsigned(short4, uint, uint); +uint4 __ovld __cnfn bitfield_extract_unsigned(int4, uint, uint); +ulong4 __ovld __cnfn bitfield_extract_unsigned(long4, uint, uint); +uchar8 __ovld __cnfn bitfield_extract_unsigned(char8, uint, uint); +ushort8 __ovld __cnfn bitfield_extract_unsigned(short8, uint, uint); +uint8 __ovld __cnfn bitfield_extract_unsigned(int8, uint, uint); +ulong8 __ovld __cnfn bitfield_extract_unsigned(long8, uint, uint); +uchar16 __ovld __cnfn bitfield_extract_unsigned(char16, uint, uint); +ushort16 __ovld __cnfn bitfield_extract_unsigned(short16, uint, uint); +uint16 __ovld __cnfn bitfield_extract_unsigned(int16, uint, uint); +ulong16 __ovld __cnfn bitfield_extract_unsigned(long16, uint, uint); + +uchar __ovld __cnfn bitfield_extract_unsigned(uchar, uint, uint); +ushort __ovld __cnfn bitfield_extract_unsigned(ushort, uint, uint); +uint __ovld __cnfn bitfield_extract_unsigned(uint, uint, uint); +ulong __ovld __cnfn bitfield_extract_unsigned(ulong, uint, uint); +uchar2 __ovld __cnfn bitfield_extract_unsigned(uchar2, uint, uint); +ushort2 __ovld __cnfn bitfield_extract_unsigned(ushort2, uint, uint); +uint2 __ovld __cnfn bitfield_extract_unsigned(uint2, uint, uint); +ulong2 __ovld __cnfn bitfield_extract_unsigned(ulong2, uint, uint); +uchar3 __ovld __cnfn bitfield_extract_unsigned(uchar3, uint, uint); +ushort3 __ovld __cnfn bitfield_extract_unsigned(ushort3, uint, uint); +uint3 __ovld __cnfn bitfield_extract_unsigned(uint3, uint, uint); +ulong3 __ovld __cnfn bitfield_extract_unsigned(ulong3, uint, uint); +uchar4 __ovld __cnfn bitfield_extract_unsigned(uchar4, uint, uint); +ushort4 __ovld __cnfn bitfield_extract_unsigned(ushort4, uint, uint); +uint4 __ovld __cnfn bitfield_extract_unsigned(uint4, uint, uint); +ulong4 __ovld __cnfn bitfield_extract_unsigned(ulong4, uint, uint); +uchar8 __ovld __cnfn bitfield_extract_unsigned(uchar8, uint, uint); +ushort8 __ovld __cnfn bitfield_extract_unsigned(ushort8, uint, uint); +uint8 __ovld __cnfn bitfield_extract_unsigned(uint8, uint, uint); +ulong8 __ovld __cnfn bitfield_extract_unsigned(ulong8, uint, uint); +uchar16 __ovld __cnfn bitfield_extract_unsigned(uchar16, uint, uint); +ushort16 __ovld __cnfn bitfield_extract_unsigned(ushort16, uint, uint); +uint16 __ovld __cnfn bitfield_extract_unsigned(uint16, uint, uint); +ulong16 __ovld __cnfn bitfield_extract_unsigned(ulong16, uint, uint); + +char __ovld __cnfn bit_reverse(char); +uchar __ovld __cnfn bit_reverse(uchar); +short __ovld __cnfn bit_reverse(short); +ushort __ovld __cnfn bit_reverse(ushort); +int __ovld __cnfn bit_reverse(int); +uint __ovld __cnfn bit_reverse(uint); +long __ovld __cnfn bit_reverse(long); +ulong __ovld __cnfn bit_reverse(ulong); +char2 __ovld __cnfn bit_reverse(char2); +uchar2 __ovld __cnfn bit_reverse(uchar2); +short2 __ovld __cnfn bit_reverse(short2); +ushort2 __ovld __cnfn bit_reverse(ushort2); +int2 __ovld __cnfn bit_reverse(int2); +uint2 __ovld __cnfn bit_reverse(uint2); +long2 __ovld __cnfn bit_reverse(long2); +ulong2 __ovld __cnfn bit_reverse(ulong2); +char3 __ovld __cnfn bit_reverse(char3); +uchar3 __ovld __cnfn bit_reverse(uchar3); +short3 __ovld __cnfn bit_reverse(short3); +ushort3 __ovld __cnfn bit_reverse(ushort3); +int3 __ovld __cnfn bit_reverse(int3); +uint3 __ovld __cnfn bit_reverse(uint3); +long3 __ovld __cnfn bit_reverse(long3); +ulong3 __ovld __cnfn bit_reverse(ulong3); +char4 __ovld __cnfn bit_reverse(char4); +uchar4 __ovld __cnfn bit_reverse(uchar4); +short4 __ovld __cnfn bit_reverse(short4); +ushort4 __ovld __cnfn bit_reverse(ushort4); +int4 __ovld __cnfn bit_reverse(int4); +uint4 __ovld __cnfn bit_reverse(uint4); +long4 __ovld __cnfn bit_reverse(long4); +ulong4 __ovld __cnfn bit_reverse(ulong4); +char8 __ovld __cnfn bit_reverse(char8); +uchar8 __ovld __cnfn bit_reverse(uchar8); +short8 __ovld __cnfn bit_reverse(short8); +ushort8 __ovld __cnfn bit_reverse(ushort8); +int8 __ovld __cnfn bit_reverse(int8); +uint8 __ovld __cnfn bit_reverse(uint8); +long8 __ovld __cnfn bit_reverse(long8); +ulong8 __ovld __cnfn bit_reverse(ulong8); +char16 __ovld __cnfn bit_reverse(char16); +uchar16 __ovld __cnfn bit_reverse(uchar16); +short16 __ovld __cnfn bit_reverse(short16); +ushort16 __ovld __cnfn bit_reverse(ushort16); +int16 __ovld __cnfn bit_reverse(int16); +uint16 __ovld __cnfn bit_reverse(uint16); +long16 __ovld __cnfn bit_reverse(long16); +ulong16 __ovld __cnfn bit_reverse(ulong16); +#endif // cl_khr_extended_bit_ops + +#if defined(__opencl_c_integer_dot_product_input_4x8bit) +uint __ovld __cnfn dot(uchar4, uchar4); +int __ovld __cnfn dot(char4, char4); +int __ovld __cnfn dot(uchar4, char4); +int __ovld __cnfn dot(char4, uchar4); + +uint __ovld __cnfn dot_acc_sat(uchar4, uchar4, uint); +int __ovld __cnfn dot_acc_sat(char4, char4, int); +int __ovld __cnfn dot_acc_sat(uchar4, char4, int); +int __ovld __cnfn dot_acc_sat(char4, uchar4, int); +#endif // __opencl_c_integer_dot_product_input_4x8bit + +#if defined(__opencl_c_integer_dot_product_input_4x8bit_packed) +uint __ovld __cnfn dot_4x8packed_uu_uint(uint, uint); +int __ovld __cnfn dot_4x8packed_ss_int(uint, uint); +int __ovld __cnfn dot_4x8packed_us_int(uint, uint); +int __ovld __cnfn dot_4x8packed_su_int(uint, uint); + +uint __ovld __cnfn dot_acc_sat_4x8packed_uu_uint(uint, uint, uint); +int __ovld __cnfn dot_acc_sat_4x8packed_ss_int(uint, uint, int); +int __ovld __cnfn dot_acc_sat_4x8packed_us_int(uint, uint, int); +int __ovld __cnfn dot_acc_sat_4x8packed_su_int(uint, uint, int); +#endif // __opencl_c_integer_dot_product_input_4x8bit_packed + #if defined(cl_intel_subgroups) // Intel-Specific Sub Group Functions float __ovld __conv intel_sub_group_shuffle( float x, uint c ); @@ -17127,31 +17285,23 @@ uint16 __ovld amd_sadw(uint16 src0, uint16 src1, uint16 src2); #endif // cl_amd_media_ops2 #if defined(cl_arm_integer_dot_product_int8) -#pragma OPENCL EXTENSION cl_arm_integer_dot_product_int8 : begin uint __ovld arm_dot(uchar4 a, uchar4 b); int __ovld arm_dot(char4 a, char4 b); -#pragma OPENCL EXTENSION cl_arm_integer_dot_product_int8 : end #endif // defined(cl_arm_integer_dot_product_int8) #if defined(cl_arm_integer_dot_product_accumulate_int8) -#pragma OPENCL EXTENSION cl_arm_integer_dot_product_accumulate_int8 : begin uint __ovld arm_dot_acc(uchar4 a, uchar4 b, uint c); int __ovld arm_dot_acc(char4 a, char4 b, int c); -#pragma OPENCL EXTENSION cl_arm_integer_dot_product_accumulate_int8 : end #endif // defined(cl_arm_integer_dot_product_accumulate_int8) #if defined(cl_arm_integer_dot_product_accumulate_int16) -#pragma OPENCL EXTENSION cl_arm_integer_dot_product_accumulate_int16 : begin uint __ovld arm_dot_acc(ushort2 a, ushort2 b, uint c); int __ovld arm_dot_acc(short2 a, short2 b, int c); -#pragma OPENCL EXTENSION cl_arm_integer_dot_product_accumulate_int16 : end #endif // defined(cl_arm_integer_dot_product_accumulate_int16) #if defined(cl_arm_integer_dot_product_accumulate_saturate_int8) -#pragma OPENCL EXTENSION cl_arm_integer_dot_product_accumulate_saturate_int8 : begin uint __ovld arm_dot_acc_sat(uchar4 a, uchar4 b, uint c); int __ovld arm_dot_acc_sat(char4 a, char4 b, int c); -#pragma OPENCL EXTENSION cl_arm_integer_dot_product_accumulate_saturate_int8 : end #endif // defined(cl_arm_integer_dot_product_accumulate_saturate_int8) // Disable any extensions we may have enabled previously. diff --git a/lib/include/openmp_wrappers/__clang_openmp_device_functions.h b/lib/include/openmp_wrappers/__clang_openmp_device_functions.h index 406c9748e..279fb26fb 100644 --- a/lib/include/openmp_wrappers/__clang_openmp_device_functions.h +++ b/lib/include/openmp_wrappers/__clang_openmp_device_functions.h @@ -14,13 +14,13 @@ #error "This file is for OpenMP compilation only." #endif -#pragma omp begin declare variant match( \ - device = {arch(nvptx, nvptx64)}, implementation = {extension(match_any)}) - #ifdef __cplusplus extern "C" { #endif +#pragma omp begin declare variant match( \ + device = {arch(nvptx, nvptx64)}, implementation = {extension(match_any)}) + #define __CUDA__ #define __OPENMP_NVPTX__ @@ -33,10 +33,74 @@ extern "C" { #undef __OPENMP_NVPTX__ #undef __CUDA__ +#pragma omp end declare variant + +#ifdef __AMDGCN__ +#pragma omp begin declare variant match(device = {arch(amdgcn)}) + +// Import types which will be used by __clang_hip_libdevice_declares.h +#ifndef __cplusplus +#include +#include +#endif + +#define __OPENMP_AMDGCN__ +#pragma push_macro("__device__") +#define __device__ + +/// Include declarations for libdevice functions. +#include <__clang_hip_libdevice_declares.h> + +#pragma pop_macro("__device__") +#undef __OPENMP_AMDGCN__ + +#pragma omp end declare variant +#endif + #ifdef __cplusplus } // extern "C" #endif -#pragma omp end declare variant +// Ensure we make `_ZdlPv`, aka. `operator delete(void*)` available without the +// need to `include ` in C++ mode. +#ifdef __cplusplus + +// We require malloc/free. +#include + +#pragma push_macro("OPENMP_NOEXCEPT") +#if __cplusplus >= 201103L +#define OPENMP_NOEXCEPT noexcept +#else +#define OPENMP_NOEXCEPT +#endif + +// Device overrides for non-placement new and delete. +inline void *operator new(__SIZE_TYPE__ size) { + if (size == 0) + size = 1; + return ::malloc(size); +} + +inline void *operator new[](__SIZE_TYPE__ size) { return ::operator new(size); } + +inline void operator delete(void *ptr)OPENMP_NOEXCEPT { ::free(ptr); } + +inline void operator delete[](void *ptr) OPENMP_NOEXCEPT { + ::operator delete(ptr); +} + +// Sized delete, C++14 only. +#if __cplusplus >= 201402L +inline void operator delete(void *ptr, __SIZE_TYPE__ size)OPENMP_NOEXCEPT { + ::operator delete(ptr); +} +inline void operator delete[](void *ptr, __SIZE_TYPE__ size) OPENMP_NOEXCEPT { + ::operator delete(ptr); +} +#endif + +#pragma pop_macro("OPENMP_NOEXCEPT") +#endif #endif diff --git a/lib/include/openmp_wrappers/cmath b/lib/include/openmp_wrappers/cmath index 1aff66af7..22a720aca 100644 --- a/lib/include/openmp_wrappers/cmath +++ b/lib/include/openmp_wrappers/cmath @@ -75,4 +75,58 @@ __DEVICE__ float tgamma(float __x) { return ::tgammaf(__x); } #pragma omp end declare variant +#ifdef __AMDGCN__ +#pragma omp begin declare variant match(device = {arch(amdgcn)}) + +#pragma push_macro("__constant__") +#define __constant__ __attribute__((constant)) +#define __OPENMP_AMDGCN__ + +#include <__clang_hip_cmath.h> + +#pragma pop_macro("__constant__") +#undef __OPENMP_AMDGCN__ + +// Define overloads otherwise which are absent +#define __DEVICE__ static constexpr __attribute__((always_inline, nothrow)) + +__DEVICE__ float acos(float __x) { return ::acosf(__x); } +__DEVICE__ float acosh(float __x) { return ::acoshf(__x); } +__DEVICE__ float asin(float __x) { return ::asinf(__x); } +__DEVICE__ float asinh(float __x) { return ::asinhf(__x); } +__DEVICE__ float atan(float __x) { return ::atanf(__x); } +__DEVICE__ float atan2(float __x, float __y) { return ::atan2f(__x, __y); } +__DEVICE__ float atanh(float __x) { return ::atanhf(__x); } +__DEVICE__ float cbrt(float __x) { return ::cbrtf(__x); } +__DEVICE__ float cosh(float __x) { return ::coshf(__x); } +__DEVICE__ float erf(float __x) { return ::erff(__x); } +__DEVICE__ float erfc(float __x) { return ::erfcf(__x); } +__DEVICE__ float exp2(float __x) { return ::exp2f(__x); } +__DEVICE__ float expm1(float __x) { return ::expm1f(__x); } +__DEVICE__ float fdim(float __x, float __y) { return ::fdimf(__x, __y); } +__DEVICE__ float hypot(float __x, float __y) { return ::hypotf(__x, __y); } +__DEVICE__ int ilogb(float __x) { return ::ilogbf(__x); } +__DEVICE__ float ldexp(float __arg, int __exp) { + return ::ldexpf(__arg, __exp); +} +__DEVICE__ float lgamma(float __x) { return ::lgammaf(__x); } +__DEVICE__ float log1p(float __x) { return ::log1pf(__x); } +__DEVICE__ float logb(float __x) { return ::logbf(__x); } +__DEVICE__ float nextafter(float __x, float __y) { + return ::nextafterf(__x, __y); +} +__DEVICE__ float remainder(float __x, float __y) { + return ::remainderf(__x, __y); +} +__DEVICE__ float scalbn(float __x, int __y) { return ::scalbnf(__x, __y); } +__DEVICE__ float sinh(float __x) { return ::sinhf(__x); } +__DEVICE__ float tan(float __x) { return ::tanf(__x); } +__DEVICE__ float tanh(float __x) { return ::tanhf(__x); } +__DEVICE__ float tgamma(float __x) { return ::tgammaf(__x); } + +#undef __DEVICE__ + +#pragma omp end declare variant +#endif // __AMDGCN__ + #endif diff --git a/lib/include/openmp_wrappers/complex b/lib/include/openmp_wrappers/complex index 142e526b8..dfd6193c9 100644 --- a/lib/include/openmp_wrappers/complex +++ b/lib/include/openmp_wrappers/complex @@ -17,7 +17,6 @@ // We require std::math functions in the complex builtins below. #include -#define __CUDA__ #define __OPENMP_NVPTX__ #include <__clang_cuda_complex_builtins.h> #undef __OPENMP_NVPTX__ @@ -26,9 +25,6 @@ // Grab the host header too. #include_next - -#ifdef __cplusplus - // If we are compiling against libc++, the macro _LIBCPP_STD_VER should be set // after including above. Since the complex header we use is a // simplified version of the libc++, we don't need it in this case. If we @@ -48,5 +44,3 @@ #pragma omp end declare variant #endif - -#endif diff --git a/lib/include/openmp_wrappers/complex.h b/lib/include/openmp_wrappers/complex.h index 00d278548..15dc415b8 100644 --- a/lib/include/openmp_wrappers/complex.h +++ b/lib/include/openmp_wrappers/complex.h @@ -17,7 +17,6 @@ // We require math functions in the complex builtins below. #include -#define __CUDA__ #define __OPENMP_NVPTX__ #include <__clang_cuda_complex_builtins.h> #undef __OPENMP_NVPTX__ diff --git a/lib/include/openmp_wrappers/math.h b/lib/include/openmp_wrappers/math.h index c64af8b13..1e3c07cfd 100644 --- a/lib/include/openmp_wrappers/math.h +++ b/lib/include/openmp_wrappers/math.h @@ -48,4 +48,14 @@ #pragma omp end declare variant +#ifdef __AMDGCN__ +#pragma omp begin declare variant match(device = {arch(amdgcn)}) + +#define __OPENMP_AMDGCN__ +#include <__clang_hip_math.h> +#undef __OPENMP_AMDGCN__ + +#pragma omp end declare variant +#endif + #endif diff --git a/lib/include/openmp_wrappers/new b/lib/include/openmp_wrappers/new index 1387d925b..985ddc567 100644 --- a/lib/include/openmp_wrappers/new +++ b/lib/include/openmp_wrappers/new @@ -9,6 +9,8 @@ #ifndef __CLANG_OPENMP_WRAPPERS_NEW #define __CLANG_OPENMP_WRAPPERS_NEW +// We need the system for the std::nothrow_t. The new/delete operators +// which do not use nothrow_t are provided without the header. #include_next #if defined(__NVPTX__) && defined(_OPENMP) @@ -22,48 +24,24 @@ #define OPENMP_NOEXCEPT #endif -// Device overrides for non-placement new and delete. -inline void *operator new(__SIZE_TYPE__ size) { - if (size == 0) - size = 1; - return ::malloc(size); -} inline void *operator new(__SIZE_TYPE__ size, const std::nothrow_t &) OPENMP_NOEXCEPT { return ::operator new(size); } -inline void *operator new[](__SIZE_TYPE__ size) { return ::operator new(size); } inline void *operator new[](__SIZE_TYPE__ size, const std::nothrow_t &) { return ::operator new(size); } -inline void operator delete(void *ptr)OPENMP_NOEXCEPT { - if (ptr) - ::free(ptr); -} inline void operator delete(void *ptr, const std::nothrow_t &)OPENMP_NOEXCEPT { ::operator delete(ptr); } -inline void operator delete[](void *ptr) OPENMP_NOEXCEPT { - ::operator delete(ptr); -} inline void operator delete[](void *ptr, const std::nothrow_t &) OPENMP_NOEXCEPT { ::operator delete(ptr); } -// Sized delete, C++14 only. -#if __cplusplus >= 201402L -inline void operator delete(void *ptr, __SIZE_TYPE__ size)OPENMP_NOEXCEPT { - ::operator delete(ptr); -} -inline void operator delete[](void *ptr, __SIZE_TYPE__ size) OPENMP_NOEXCEPT { - ::operator delete(ptr); -} -#endif - #pragma pop_macro("OPENMP_NOEXCEPT") #endif diff --git a/lib/include/ppc_wrappers/xmmintrin.h b/lib/include/ppc_wrappers/xmmintrin.h index 0f429fa04..0e45b9676 100644 --- a/lib/include/ppc_wrappers/xmmintrin.h +++ b/lib/include/ppc_wrappers/xmmintrin.h @@ -28,7 +28,7 @@ Most SSE scalar float intrinsic operations can be performed more efficiently as C language float scalar operations or optimized to use vector SIMD operations. We recommend this for new applications. */ -#error "Please read comment above. Use -DNO_WARN_X86_INTRINSICS to disable this error." +#error "Please read comment above. Use -DNO_WARN_X86_INTRINSICS to disable this error." #endif #ifndef _XMMINTRIN_H_INCLUDED @@ -62,14 +62,13 @@ /* The Intel API is flexible enough that we must allow aliasing with other vector types, and their scalar components. */ -typedef float __m128 __attribute__ ((__vector_size__ (16), __may_alias__)); +typedef vector float __m128 __attribute__((__may_alias__)); /* Unaligned version of the same type. */ -typedef float __m128_u __attribute__ ((__vector_size__ (16), __may_alias__, - __aligned__ (1))); +typedef vector float __m128_u __attribute__((__may_alias__, __aligned__(1))); /* Internal data types for implementing the intrinsics. */ -typedef float __v4sf __attribute__ ((__vector_size__ (16))); +typedef vector float __v4sf; /* Create an undefined vector. */ extern __inline __m128 __attribute__((__gnu_inline__, __always_inline__, __artificial__)) diff --git a/lib/include/riscv_vector.h b/lib/include/riscv_vector.h new file mode 100644 index 000000000..7fc360f9d --- /dev/null +++ b/lib/include/riscv_vector.h @@ -0,0 +1,123865 @@ +/*===---- riscv_vector.h - RISC-V V-extension RVVIntrinsics -------------------=== + * + * + * Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. + * See https://llvm.org/LICENSE.txt for license information. + * SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception + * + *===-----------------------------------------------------------------------=== + */ + +#ifndef __RISCV_VECTOR_H +#define __RISCV_VECTOR_H + +#include +#include + +#ifndef __riscv_vector +#error "Vector intrinsics require the vector extension." +#endif + +#ifdef __cplusplus +extern "C" { +#endif + + +#define vsetvl_e8mf8(avl) __builtin_rvv_vsetvli((size_t)(avl), 0, 5) +#define vsetvl_e8mf4(avl) __builtin_rvv_vsetvli((size_t)(avl), 0, 6) +#define vsetvl_e8mf2(avl) __builtin_rvv_vsetvli((size_t)(avl), 0, 7) +#define vsetvl_e8m1(avl) __builtin_rvv_vsetvli((size_t)(avl), 0, 0) +#define vsetvl_e8m2(avl) __builtin_rvv_vsetvli((size_t)(avl), 0, 1) +#define vsetvl_e8m4(avl) __builtin_rvv_vsetvli((size_t)(avl), 0, 2) +#define vsetvl_e8m8(avl) __builtin_rvv_vsetvli((size_t)(avl), 0, 3) + +#define vsetvl_e16mf4(avl) __builtin_rvv_vsetvli((size_t)(avl), 1, 6) +#define vsetvl_e16mf2(avl) __builtin_rvv_vsetvli((size_t)(avl), 1, 7) +#define vsetvl_e16m1(avl) __builtin_rvv_vsetvli((size_t)(avl), 1, 0) +#define vsetvl_e16m2(avl) __builtin_rvv_vsetvli((size_t)(avl), 1, 1) +#define vsetvl_e16m4(avl) __builtin_rvv_vsetvli((size_t)(avl), 1, 2) +#define vsetvl_e16m8(avl) __builtin_rvv_vsetvli((size_t)(avl), 1, 3) + +#define vsetvl_e32mf2(avl) __builtin_rvv_vsetvli((size_t)(avl), 2, 7) +#define vsetvl_e32m1(avl) __builtin_rvv_vsetvli((size_t)(avl), 2, 0) +#define vsetvl_e32m2(avl) __builtin_rvv_vsetvli((size_t)(avl), 2, 1) +#define vsetvl_e32m4(avl) __builtin_rvv_vsetvli((size_t)(avl), 2, 2) +#define vsetvl_e32m8(avl) __builtin_rvv_vsetvli((size_t)(avl), 2, 3) + +#define vsetvl_e64m1(avl) __builtin_rvv_vsetvli((size_t)(avl), 3, 0) +#define vsetvl_e64m2(avl) __builtin_rvv_vsetvli((size_t)(avl), 3, 1) +#define vsetvl_e64m4(avl) __builtin_rvv_vsetvli((size_t)(avl), 3, 2) +#define vsetvl_e64m8(avl) __builtin_rvv_vsetvli((size_t)(avl), 3, 3) + + +#define vsetvlmax_e8mf8() __builtin_rvv_vsetvlimax(0, 5) +#define vsetvlmax_e8mf4() __builtin_rvv_vsetvlimax(0, 6) +#define vsetvlmax_e8mf2() __builtin_rvv_vsetvlimax(0, 7) +#define vsetvlmax_e8m1() __builtin_rvv_vsetvlimax(0, 0) +#define vsetvlmax_e8m2() __builtin_rvv_vsetvlimax(0, 1) +#define vsetvlmax_e8m4() __builtin_rvv_vsetvlimax(0, 2) +#define vsetvlmax_e8m8() __builtin_rvv_vsetvlimax(0, 3) + +#define vsetvlmax_e16mf4() __builtin_rvv_vsetvlimax(1, 6) +#define vsetvlmax_e16mf2() __builtin_rvv_vsetvlimax(1, 7) +#define vsetvlmax_e16m1() __builtin_rvv_vsetvlimax(1, 0) +#define vsetvlmax_e16m2() __builtin_rvv_vsetvlimax(1, 1) +#define vsetvlmax_e16m4() __builtin_rvv_vsetvlimax(1, 2) +#define vsetvlmax_e16m8() __builtin_rvv_vsetvlimax(1, 3) + +#define vsetvlmax_e32mf2() __builtin_rvv_vsetvlimax(2, 7) +#define vsetvlmax_e32m1() __builtin_rvv_vsetvlimax(2, 0) +#define vsetvlmax_e32m2() __builtin_rvv_vsetvlimax(2, 1) +#define vsetvlmax_e32m4() __builtin_rvv_vsetvlimax(2, 2) +#define vsetvlmax_e32m8() __builtin_rvv_vsetvlimax(2, 3) + +#define vsetvlmax_e64m1() __builtin_rvv_vsetvlimax(3, 0) +#define vsetvlmax_e64m2() __builtin_rvv_vsetvlimax(3, 1) +#define vsetvlmax_e64m4() __builtin_rvv_vsetvlimax(3, 2) +#define vsetvlmax_e64m8() __builtin_rvv_vsetvlimax(3, 3) + +typedef __rvv_bool64_t vbool64_t; +typedef __rvv_bool32_t vbool32_t; +typedef __rvv_bool16_t vbool16_t; +typedef __rvv_bool8_t vbool8_t; +typedef __rvv_bool4_t vbool4_t; +typedef __rvv_bool2_t vbool2_t; +typedef __rvv_bool1_t vbool1_t; +typedef __rvv_int8mf8_t vint8mf8_t; +typedef __rvv_uint8mf8_t vuint8mf8_t; +typedef __rvv_int8mf4_t vint8mf4_t; +typedef __rvv_uint8mf4_t vuint8mf4_t; +typedef __rvv_int8mf2_t vint8mf2_t; +typedef __rvv_uint8mf2_t vuint8mf2_t; +typedef __rvv_int8m1_t vint8m1_t; +typedef __rvv_uint8m1_t vuint8m1_t; +typedef __rvv_int8m2_t vint8m2_t; +typedef __rvv_uint8m2_t vuint8m2_t; +typedef __rvv_int8m4_t vint8m4_t; +typedef __rvv_uint8m4_t vuint8m4_t; +typedef __rvv_int8m8_t vint8m8_t; +typedef __rvv_uint8m8_t vuint8m8_t; +typedef __rvv_int16mf4_t vint16mf4_t; +typedef __rvv_uint16mf4_t vuint16mf4_t; +typedef __rvv_int16mf2_t vint16mf2_t; +typedef __rvv_uint16mf2_t vuint16mf2_t; +typedef __rvv_int16m1_t vint16m1_t; +typedef __rvv_uint16m1_t vuint16m1_t; +typedef __rvv_int16m2_t vint16m2_t; +typedef __rvv_uint16m2_t vuint16m2_t; +typedef __rvv_int16m4_t vint16m4_t; +typedef __rvv_uint16m4_t vuint16m4_t; +typedef __rvv_int16m8_t vint16m8_t; +typedef __rvv_uint16m8_t vuint16m8_t; +typedef __rvv_int32mf2_t vint32mf2_t; +typedef __rvv_uint32mf2_t vuint32mf2_t; +typedef __rvv_int32m1_t vint32m1_t; +typedef __rvv_uint32m1_t vuint32m1_t; +typedef __rvv_int32m2_t vint32m2_t; +typedef __rvv_uint32m2_t vuint32m2_t; +typedef __rvv_int32m4_t vint32m4_t; +typedef __rvv_uint32m4_t vuint32m4_t; +typedef __rvv_int32m8_t vint32m8_t; +typedef __rvv_uint32m8_t vuint32m8_t; +typedef __rvv_int64m1_t vint64m1_t; +typedef __rvv_uint64m1_t vuint64m1_t; +typedef __rvv_int64m2_t vint64m2_t; +typedef __rvv_uint64m2_t vuint64m2_t; +typedef __rvv_int64m4_t vint64m4_t; +typedef __rvv_uint64m4_t vuint64m4_t; +typedef __rvv_int64m8_t vint64m8_t; +typedef __rvv_uint64m8_t vuint64m8_t; +#if defined(__riscv_zfh) +typedef __rvv_float16mf4_t vfloat16mf4_t; +typedef __rvv_float16mf2_t vfloat16mf2_t; +typedef __rvv_float16m1_t vfloat16m1_t; +typedef __rvv_float16m2_t vfloat16m2_t; +typedef __rvv_float16m4_t vfloat16m4_t; +typedef __rvv_float16m8_t vfloat16m8_t; +#endif +#if defined(__riscv_f) +typedef __rvv_float32mf2_t vfloat32mf2_t; +typedef __rvv_float32m1_t vfloat32m1_t; +typedef __rvv_float32m2_t vfloat32m2_t; +typedef __rvv_float32m4_t vfloat32m4_t; +typedef __rvv_float32m8_t vfloat32m8_t; +#endif +#if defined(__riscv_d) +typedef __rvv_float64m1_t vfloat64m1_t; +typedef __rvv_float64m2_t vfloat64m2_t; +typedef __rvv_float64m4_t vfloat64m4_t; +typedef __rvv_float64m8_t vfloat64m8_t; +#endif + +#define vadd_vv_i8m1(op0, op1, op2) \ +__builtin_rvv_vadd_vv_i8m1((vint8m1_t)(op0), (vint8m1_t)(op1), (size_t)(op2)) +#define vadd_vv_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vv_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vadd_vv_i8m2(op0, op1, op2) \ +__builtin_rvv_vadd_vv_i8m2((vint8m2_t)(op0), (vint8m2_t)(op1), (size_t)(op2)) +#define vadd_vv_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vv_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vadd_vv_i8m4(op0, op1, op2) \ +__builtin_rvv_vadd_vv_i8m4((vint8m4_t)(op0), (vint8m4_t)(op1), (size_t)(op2)) +#define vadd_vv_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vv_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vadd_vv_i8m8(op0, op1, op2) \ +__builtin_rvv_vadd_vv_i8m8((vint8m8_t)(op0), (vint8m8_t)(op1), (size_t)(op2)) +#define vadd_vv_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vv_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (vint8m8_t)(op3), (size_t)(op4)) +#define vadd_vv_i8mf2(op0, op1, op2) \ +__builtin_rvv_vadd_vv_i8mf2((vint8mf2_t)(op0), (vint8mf2_t)(op1), (size_t)(op2)) +#define vadd_vv_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vv_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vadd_vv_i8mf4(op0, op1, op2) \ +__builtin_rvv_vadd_vv_i8mf4((vint8mf4_t)(op0), (vint8mf4_t)(op1), (size_t)(op2)) +#define vadd_vv_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vv_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vadd_vv_i8mf8(op0, op1, op2) \ +__builtin_rvv_vadd_vv_i8mf8((vint8mf8_t)(op0), (vint8mf8_t)(op1), (size_t)(op2)) +#define vadd_vv_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vv_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vadd_vv_i16m1(op0, op1, op2) \ +__builtin_rvv_vadd_vv_i16m1((vint16m1_t)(op0), (vint16m1_t)(op1), (size_t)(op2)) +#define vadd_vv_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vv_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vadd_vv_i16m2(op0, op1, op2) \ +__builtin_rvv_vadd_vv_i16m2((vint16m2_t)(op0), (vint16m2_t)(op1), (size_t)(op2)) +#define vadd_vv_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vv_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vadd_vv_i16m4(op0, op1, op2) \ +__builtin_rvv_vadd_vv_i16m4((vint16m4_t)(op0), (vint16m4_t)(op1), (size_t)(op2)) +#define vadd_vv_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vv_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vadd_vv_i16m8(op0, op1, op2) \ +__builtin_rvv_vadd_vv_i16m8((vint16m8_t)(op0), (vint16m8_t)(op1), (size_t)(op2)) +#define vadd_vv_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vv_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (vint16m8_t)(op3), (size_t)(op4)) +#define vadd_vv_i16mf2(op0, op1, op2) \ +__builtin_rvv_vadd_vv_i16mf2((vint16mf2_t)(op0), (vint16mf2_t)(op1), (size_t)(op2)) +#define vadd_vv_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vv_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vadd_vv_i16mf4(op0, op1, op2) \ +__builtin_rvv_vadd_vv_i16mf4((vint16mf4_t)(op0), (vint16mf4_t)(op1), (size_t)(op2)) +#define vadd_vv_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vv_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vadd_vv_i32m1(op0, op1, op2) \ +__builtin_rvv_vadd_vv_i32m1((vint32m1_t)(op0), (vint32m1_t)(op1), (size_t)(op2)) +#define vadd_vv_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vv_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vadd_vv_i32m2(op0, op1, op2) \ +__builtin_rvv_vadd_vv_i32m2((vint32m2_t)(op0), (vint32m2_t)(op1), (size_t)(op2)) +#define vadd_vv_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vv_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vadd_vv_i32m4(op0, op1, op2) \ +__builtin_rvv_vadd_vv_i32m4((vint32m4_t)(op0), (vint32m4_t)(op1), (size_t)(op2)) +#define vadd_vv_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vv_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vadd_vv_i32m8(op0, op1, op2) \ +__builtin_rvv_vadd_vv_i32m8((vint32m8_t)(op0), (vint32m8_t)(op1), (size_t)(op2)) +#define vadd_vv_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vv_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vadd_vv_i32mf2(op0, op1, op2) \ +__builtin_rvv_vadd_vv_i32mf2((vint32mf2_t)(op0), (vint32mf2_t)(op1), (size_t)(op2)) +#define vadd_vv_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vv_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vadd_vv_i64m1(op0, op1, op2) \ +__builtin_rvv_vadd_vv_i64m1((vint64m1_t)(op0), (vint64m1_t)(op1), (size_t)(op2)) +#define vadd_vv_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vv_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vadd_vv_i64m2(op0, op1, op2) \ +__builtin_rvv_vadd_vv_i64m2((vint64m2_t)(op0), (vint64m2_t)(op1), (size_t)(op2)) +#define vadd_vv_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vv_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vadd_vv_i64m4(op0, op1, op2) \ +__builtin_rvv_vadd_vv_i64m4((vint64m4_t)(op0), (vint64m4_t)(op1), (size_t)(op2)) +#define vadd_vv_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vv_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vadd_vv_i64m8(op0, op1, op2) \ +__builtin_rvv_vadd_vv_i64m8((vint64m8_t)(op0), (vint64m8_t)(op1), (size_t)(op2)) +#define vadd_vv_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vv_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vwaddu_vv_u16mf4(op0, op1, op2) \ +__builtin_rvv_vwaddu_vv_u16mf4((vuint8mf8_t)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vwaddu_vv_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_vv_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vwaddu_vv_u16mf2(op0, op1, op2) \ +__builtin_rvv_vwaddu_vv_u16mf2((vuint8mf4_t)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vwaddu_vv_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_vv_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vwaddu_vv_u16m1(op0, op1, op2) \ +__builtin_rvv_vwaddu_vv_u16m1((vuint8mf2_t)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vwaddu_vv_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_vv_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vwaddu_vv_u16m2(op0, op1, op2) \ +__builtin_rvv_vwaddu_vv_u16m2((vuint8m1_t)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vwaddu_vv_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_vv_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vwaddu_vv_u16m4(op0, op1, op2) \ +__builtin_rvv_vwaddu_vv_u16m4((vuint8m2_t)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vwaddu_vv_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_vv_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vwaddu_vv_u16m8(op0, op1, op2) \ +__builtin_rvv_vwaddu_vv_u16m8((vuint8m4_t)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vwaddu_vv_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_vv_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vwaddu_vv_u32mf2(op0, op1, op2) \ +__builtin_rvv_vwaddu_vv_u32mf2((vuint16mf4_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vwaddu_vv_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_vv_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vwaddu_vv_u32m1(op0, op1, op2) \ +__builtin_rvv_vwaddu_vv_u32m1((vuint16mf2_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vwaddu_vv_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_vv_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vwaddu_vv_u32m2(op0, op1, op2) \ +__builtin_rvv_vwaddu_vv_u32m2((vuint16m1_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vwaddu_vv_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_vv_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vwaddu_vv_u32m4(op0, op1, op2) \ +__builtin_rvv_vwaddu_vv_u32m4((vuint16m2_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vwaddu_vv_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_vv_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vwaddu_vv_u32m8(op0, op1, op2) \ +__builtin_rvv_vwaddu_vv_u32m8((vuint16m4_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vwaddu_vv_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_vv_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vwaddu_vv_u64m1(op0, op1, op2) \ +__builtin_rvv_vwaddu_vv_u64m1((vuint32mf2_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vwaddu_vv_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_vv_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vwaddu_vv_u64m2(op0, op1, op2) \ +__builtin_rvv_vwaddu_vv_u64m2((vuint32m1_t)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vwaddu_vv_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_vv_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vwaddu_vv_u64m4(op0, op1, op2) \ +__builtin_rvv_vwaddu_vv_u64m4((vuint32m2_t)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vwaddu_vv_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_vv_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vwaddu_vv_u64m8(op0, op1, op2) \ +__builtin_rvv_vwaddu_vv_u64m8((vuint32m4_t)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vwaddu_vv_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_vv_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vluxei8_v_u8m1(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_u8m1((const uint8_t *)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vluxei8_v_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (const uint8_t *)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vluxei8_v_u8m2(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_u8m2((const uint8_t *)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vluxei8_v_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (const uint8_t *)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vluxei8_v_u8m4(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_u8m4((const uint8_t *)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vluxei8_v_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (const uint8_t *)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vluxei8_v_u8m8(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_u8m8((const uint8_t *)(op0), (vuint8m8_t)(op1), (size_t)(op2)) +#define vluxei8_v_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (const uint8_t *)(op2), (vuint8m8_t)(op3), (size_t)(op4)) +#define vluxei8_v_u8mf2(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_u8mf2((const uint8_t *)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vluxei8_v_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (const uint8_t *)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vluxei8_v_u8mf4(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_u8mf4((const uint8_t *)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vluxei8_v_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (const uint8_t *)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vluxei8_v_u8mf8(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_u8mf8((const uint8_t *)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vluxei8_v_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (const uint8_t *)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vlse64_v_u64m1(op0, op1, op2) \ +__builtin_rvv_vlse64_v_u64m1((const uint64_t *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse64_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse64_v_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (const uint64_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse64_v_u64m2(op0, op1, op2) \ +__builtin_rvv_vlse64_v_u64m2((const uint64_t *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse64_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse64_v_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (const uint64_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse64_v_u64m4(op0, op1, op2) \ +__builtin_rvv_vlse64_v_u64m4((const uint64_t *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse64_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse64_v_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (const uint64_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse64_v_u64m8(op0, op1, op2) \ +__builtin_rvv_vlse64_v_u64m8((const uint64_t *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse64_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse64_v_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (const uint64_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vsse16_v_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vsse16_v_i16m1((int16_t *)(op0), (ptrdiff_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vsse16_v_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse16_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (ptrdiff_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vsse16_v_i16m2(op0, op1, op2, op3) \ +__builtin_rvv_vsse16_v_i16m2((int16_t *)(op0), (ptrdiff_t)(op1), (vint16m2_t)(op2), (size_t)(op3)) +#define vsse16_v_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse16_v_i16m2_m((vbool8_t)(op0), (int16_t *)(op1), (ptrdiff_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vsse16_v_i16m4(op0, op1, op2, op3) \ +__builtin_rvv_vsse16_v_i16m4((int16_t *)(op0), (ptrdiff_t)(op1), (vint16m4_t)(op2), (size_t)(op3)) +#define vsse16_v_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse16_v_i16m4_m((vbool4_t)(op0), (int16_t *)(op1), (ptrdiff_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vsse16_v_i16m8(op0, op1, op2, op3) \ +__builtin_rvv_vsse16_v_i16m8((int16_t *)(op0), (ptrdiff_t)(op1), (vint16m8_t)(op2), (size_t)(op3)) +#define vsse16_v_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse16_v_i16m8_m((vbool2_t)(op0), (int16_t *)(op1), (ptrdiff_t)(op2), (vint16m8_t)(op3), (size_t)(op4)) +#define vsse16_v_i16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsse16_v_i16mf2((int16_t *)(op0), (ptrdiff_t)(op1), (vint16mf2_t)(op2), (size_t)(op3)) +#define vsse16_v_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse16_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (ptrdiff_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vsse16_v_i16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsse16_v_i16mf4((int16_t *)(op0), (ptrdiff_t)(op1), (vint16mf4_t)(op2), (size_t)(op3)) +#define vsse16_v_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse16_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (ptrdiff_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vsse16_v_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vsse16_v_u16m1((uint16_t *)(op0), (ptrdiff_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vsse16_v_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse16_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (ptrdiff_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vsse16_v_u16m2(op0, op1, op2, op3) \ +__builtin_rvv_vsse16_v_u16m2((uint16_t *)(op0), (ptrdiff_t)(op1), (vuint16m2_t)(op2), (size_t)(op3)) +#define vsse16_v_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse16_v_u16m2_m((vbool8_t)(op0), (uint16_t *)(op1), (ptrdiff_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vsse16_v_u16m4(op0, op1, op2, op3) \ +__builtin_rvv_vsse16_v_u16m4((uint16_t *)(op0), (ptrdiff_t)(op1), (vuint16m4_t)(op2), (size_t)(op3)) +#define vsse16_v_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse16_v_u16m4_m((vbool4_t)(op0), (uint16_t *)(op1), (ptrdiff_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vsse16_v_u16m8(op0, op1, op2, op3) \ +__builtin_rvv_vsse16_v_u16m8((uint16_t *)(op0), (ptrdiff_t)(op1), (vuint16m8_t)(op2), (size_t)(op3)) +#define vsse16_v_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse16_v_u16m8_m((vbool2_t)(op0), (uint16_t *)(op1), (ptrdiff_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vsse16_v_u16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsse16_v_u16mf2((uint16_t *)(op0), (ptrdiff_t)(op1), (vuint16mf2_t)(op2), (size_t)(op3)) +#define vsse16_v_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse16_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (ptrdiff_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vsse16_v_u16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsse16_v_u16mf4((uint16_t *)(op0), (ptrdiff_t)(op1), (vuint16mf4_t)(op2), (size_t)(op3)) +#define vsse16_v_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse16_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (ptrdiff_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vsse32_v_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vsse32_v_i32m1((int32_t *)(op0), (ptrdiff_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vsse32_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse32_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (ptrdiff_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vsse32_v_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vsse32_v_i32m2((int32_t *)(op0), (ptrdiff_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vsse32_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse32_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (ptrdiff_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vsse32_v_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vsse32_v_i32m4((int32_t *)(op0), (ptrdiff_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vsse32_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse32_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (ptrdiff_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vsse32_v_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vsse32_v_i32m8((int32_t *)(op0), (ptrdiff_t)(op1), (vint32m8_t)(op2), (size_t)(op3)) +#define vsse32_v_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse32_v_i32m8_m((vbool4_t)(op0), (int32_t *)(op1), (ptrdiff_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vsse32_v_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsse32_v_i32mf2((int32_t *)(op0), (ptrdiff_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vsse32_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse32_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (ptrdiff_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vsse32_v_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vsse32_v_u32m1((uint32_t *)(op0), (ptrdiff_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vsse32_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse32_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (ptrdiff_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vsse32_v_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vsse32_v_u32m2((uint32_t *)(op0), (ptrdiff_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vsse32_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse32_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (ptrdiff_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vsse32_v_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vsse32_v_u32m4((uint32_t *)(op0), (ptrdiff_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vsse32_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse32_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (ptrdiff_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vsse32_v_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vsse32_v_u32m8((uint32_t *)(op0), (ptrdiff_t)(op1), (vuint32m8_t)(op2), (size_t)(op3)) +#define vsse32_v_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse32_v_u32m8_m((vbool4_t)(op0), (uint32_t *)(op1), (ptrdiff_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vsse32_v_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsse32_v_u32mf2((uint32_t *)(op0), (ptrdiff_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vsse32_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse32_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (ptrdiff_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vsse64_v_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vsse64_v_i64m1((int64_t *)(op0), (ptrdiff_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vsse64_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse64_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (ptrdiff_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vsse64_v_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vsse64_v_i64m2((int64_t *)(op0), (ptrdiff_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vsse64_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse64_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (ptrdiff_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vsse64_v_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vsse64_v_i64m4((int64_t *)(op0), (ptrdiff_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vsse64_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse64_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (ptrdiff_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vsse64_v_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vsse64_v_i64m8((int64_t *)(op0), (ptrdiff_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vsse64_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse64_v_i64m8_m((vbool8_t)(op0), (int64_t *)(op1), (ptrdiff_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vsse64_v_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vsse64_v_u64m1((uint64_t *)(op0), (ptrdiff_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vsse64_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse64_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (ptrdiff_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vsse64_v_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vsse64_v_u64m2((uint64_t *)(op0), (ptrdiff_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vsse64_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse64_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (ptrdiff_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vsse64_v_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vsse64_v_u64m4((uint64_t *)(op0), (ptrdiff_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vsse64_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse64_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (ptrdiff_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vsse64_v_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vsse64_v_u64m8((uint64_t *)(op0), (ptrdiff_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vsse64_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse64_v_u64m8_m((vbool8_t)(op0), (uint64_t *)(op1), (ptrdiff_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vluxei16_v_i8m1(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_i8m1((const int8_t *)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vluxei16_v_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (const int8_t *)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vluxei16_v_i8m2(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_i8m2((const int8_t *)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vluxei16_v_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (const int8_t *)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vluxei16_v_i8m4(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_i8m4((const int8_t *)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vluxei16_v_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (const int8_t *)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vluxei16_v_i8mf2(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_i8mf2((const int8_t *)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vluxei16_v_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (const int8_t *)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vluxei16_v_i8mf4(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_i8mf4((const int8_t *)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vluxei16_v_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (const int8_t *)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vluxei16_v_i8mf8(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_i8mf8((const int8_t *)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vluxei16_v_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (const int8_t *)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vluxei16_v_u8m1(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_u8m1((const uint8_t *)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vluxei16_v_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (const uint8_t *)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vluxei16_v_u8m2(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_u8m2((const uint8_t *)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vluxei16_v_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (const uint8_t *)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vluxei16_v_u8m4(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_u8m4((const uint8_t *)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vluxei16_v_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (const uint8_t *)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vluxei16_v_u8mf2(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_u8mf2((const uint8_t *)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vluxei16_v_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (const uint8_t *)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vluxei16_v_u8mf4(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_u8mf4((const uint8_t *)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vluxei16_v_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (const uint8_t *)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vluxei16_v_u8mf8(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_u8mf8((const uint8_t *)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vluxei16_v_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (const uint8_t *)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vluxei32_v_i8m1(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_i8m1((const int8_t *)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vluxei32_v_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (const int8_t *)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vluxei32_v_i8m2(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_i8m2((const int8_t *)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vluxei32_v_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (const int8_t *)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vluxei32_v_i8mf2(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_i8mf2((const int8_t *)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vluxei32_v_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (const int8_t *)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vluxei32_v_i8mf4(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_i8mf4((const int8_t *)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vluxei32_v_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (const int8_t *)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vluxei32_v_i8mf8(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_i8mf8((const int8_t *)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vluxei32_v_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (const int8_t *)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vluxei32_v_u8m1(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_u8m1((const uint8_t *)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vluxei32_v_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (const uint8_t *)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vluxei32_v_u8m2(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_u8m2((const uint8_t *)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vluxei32_v_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (const uint8_t *)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vluxei32_v_u8mf2(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_u8mf2((const uint8_t *)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vluxei32_v_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (const uint8_t *)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vluxei32_v_u8mf4(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_u8mf4((const uint8_t *)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vluxei32_v_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (const uint8_t *)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vluxei32_v_u8mf8(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_u8mf8((const uint8_t *)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vluxei32_v_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (const uint8_t *)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vluxei64_v_i8m1(op0, op1, op2) \ +__builtin_rvv_vluxei64_v_i8m1((const int8_t *)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vluxei64_v_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei64_v_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (const int8_t *)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vluxei64_v_i8mf2(op0, op1, op2) \ +__builtin_rvv_vluxei64_v_i8mf2((const int8_t *)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vluxei64_v_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei64_v_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (const int8_t *)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vluxei64_v_i8mf4(op0, op1, op2) \ +__builtin_rvv_vluxei64_v_i8mf4((const int8_t *)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vluxei64_v_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei64_v_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (const int8_t *)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vluxei64_v_i8mf8(op0, op1, op2) \ +__builtin_rvv_vluxei64_v_i8mf8((const int8_t *)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vluxei64_v_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei64_v_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (const int8_t *)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vluxei64_v_u8m1(op0, op1, op2) \ +__builtin_rvv_vluxei64_v_u8m1((const uint8_t *)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vluxei64_v_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei64_v_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (const uint8_t *)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vluxei64_v_u8mf2(op0, op1, op2) \ +__builtin_rvv_vluxei64_v_u8mf2((const uint8_t *)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vluxei64_v_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei64_v_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (const uint8_t *)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vluxei64_v_u8mf4(op0, op1, op2) \ +__builtin_rvv_vluxei64_v_u8mf4((const uint8_t *)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vluxei64_v_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei64_v_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (const uint8_t *)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vluxei64_v_u8mf8(op0, op1, op2) \ +__builtin_rvv_vluxei64_v_u8mf8((const uint8_t *)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vluxei64_v_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei64_v_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (const uint8_t *)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vluxei8_v_i16m1(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_i16m1((const int16_t *)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vluxei8_v_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (const int16_t *)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vluxei8_v_i16m2(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_i16m2((const int16_t *)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vluxei8_v_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (const int16_t *)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vluxei8_v_i16m4(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_i16m4((const int16_t *)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vluxei8_v_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (const int16_t *)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vluxei8_v_i16m8(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_i16m8((const int16_t *)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vluxei8_v_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (const int16_t *)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vluxei8_v_i16mf2(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_i16mf2((const int16_t *)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vluxei8_v_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (const int16_t *)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vluxei8_v_i16mf4(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_i16mf4((const int16_t *)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vluxei8_v_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (const int16_t *)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vluxei8_v_u16m1(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_u16m1((const uint16_t *)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vluxei8_v_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (const uint16_t *)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vluxei8_v_u16m2(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_u16m2((const uint16_t *)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vluxei8_v_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (const uint16_t *)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vluxei8_v_u16m4(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_u16m4((const uint16_t *)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vluxei8_v_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (const uint16_t *)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vluxei8_v_u16m8(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_u16m8((const uint16_t *)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vluxei8_v_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (const uint16_t *)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vluxei8_v_u16mf2(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_u16mf2((const uint16_t *)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vluxei8_v_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (const uint16_t *)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vluxei8_v_u16mf4(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_u16mf4((const uint16_t *)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vluxei8_v_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (const uint16_t *)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vluxei16_v_i16m1(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_i16m1((const int16_t *)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vluxei16_v_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (const int16_t *)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vluxei16_v_i16m2(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_i16m2((const int16_t *)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vluxei16_v_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (const int16_t *)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vluxei16_v_i16m4(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_i16m4((const int16_t *)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vluxei16_v_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (const int16_t *)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vluxei16_v_i16m8(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_i16m8((const int16_t *)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vluxei16_v_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (const int16_t *)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vluxei16_v_i16mf2(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_i16mf2((const int16_t *)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vluxei16_v_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (const int16_t *)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vluxei16_v_i16mf4(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_i16mf4((const int16_t *)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vluxei16_v_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (const int16_t *)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vluxei16_v_u16m1(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_u16m1((const uint16_t *)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vluxei16_v_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (const uint16_t *)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vluxei16_v_u16m2(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_u16m2((const uint16_t *)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vluxei16_v_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (const uint16_t *)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vluxei16_v_u16m4(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_u16m4((const uint16_t *)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vluxei16_v_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (const uint16_t *)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vluxei16_v_u16m8(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_u16m8((const uint16_t *)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vluxei16_v_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (const uint16_t *)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vluxei16_v_u16mf2(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_u16mf2((const uint16_t *)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vluxei16_v_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (const uint16_t *)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vluxei16_v_u16mf4(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_u16mf4((const uint16_t *)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vluxei16_v_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (const uint16_t *)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vluxei32_v_i16m1(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_i16m1((const int16_t *)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vluxei32_v_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (const int16_t *)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vluxei32_v_i16m2(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_i16m2((const int16_t *)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vluxei32_v_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (const int16_t *)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vluxei32_v_i16m4(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_i16m4((const int16_t *)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vluxei32_v_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (const int16_t *)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vluxei32_v_i16mf2(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_i16mf2((const int16_t *)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vluxei32_v_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (const int16_t *)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vluxei32_v_i16mf4(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_i16mf4((const int16_t *)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vluxei32_v_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (const int16_t *)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vluxei32_v_u16m1(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_u16m1((const uint16_t *)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vluxei32_v_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (const uint16_t *)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vluxei32_v_u16m2(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_u16m2((const uint16_t *)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vluxei32_v_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (const uint16_t *)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vluxei32_v_u16m4(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_u16m4((const uint16_t *)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vluxei32_v_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (const uint16_t *)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vluxei32_v_u16mf2(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_u16mf2((const uint16_t *)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vluxei32_v_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (const uint16_t *)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vluxei32_v_u16mf4(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_u16mf4((const uint16_t *)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vluxei32_v_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (const uint16_t *)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vluxei64_v_i16m1(op0, op1, op2) \ +__builtin_rvv_vluxei64_v_i16m1((const int16_t *)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vluxei64_v_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei64_v_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (const int16_t *)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vluxei64_v_i16m2(op0, op1, op2) \ +__builtin_rvv_vluxei64_v_i16m2((const int16_t *)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vluxei64_v_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei64_v_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (const int16_t *)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vluxei64_v_i16mf2(op0, op1, op2) \ +__builtin_rvv_vluxei64_v_i16mf2((const int16_t *)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vluxei64_v_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei64_v_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (const int16_t *)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vluxei64_v_i16mf4(op0, op1, op2) \ +__builtin_rvv_vluxei64_v_i16mf4((const int16_t *)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vluxei64_v_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei64_v_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (const int16_t *)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vluxei64_v_u16m1(op0, op1, op2) \ +__builtin_rvv_vluxei64_v_u16m1((const uint16_t *)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vluxei64_v_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei64_v_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (const uint16_t *)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vluxei64_v_u16m2(op0, op1, op2) \ +__builtin_rvv_vluxei64_v_u16m2((const uint16_t *)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vluxei64_v_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei64_v_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (const uint16_t *)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vluxei64_v_u16mf2(op0, op1, op2) \ +__builtin_rvv_vluxei64_v_u16mf2((const uint16_t *)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vluxei64_v_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei64_v_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (const uint16_t *)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vluxei64_v_u16mf4(op0, op1, op2) \ +__builtin_rvv_vluxei64_v_u16mf4((const uint16_t *)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vluxei64_v_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei64_v_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (const uint16_t *)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vluxei8_v_i32m1(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_i32m1((const int32_t *)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vluxei8_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (const int32_t *)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vluxei8_v_i32m2(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_i32m2((const int32_t *)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vluxei8_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (const int32_t *)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vluxei8_v_i32m4(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_i32m4((const int32_t *)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vluxei8_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (const int32_t *)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vluxei8_v_i32m8(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_i32m8((const int32_t *)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vluxei8_v_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (const int32_t *)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vluxei8_v_i32mf2(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_i32mf2((const int32_t *)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vluxei8_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (const int32_t *)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vluxei8_v_u32m1(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_u32m1((const uint32_t *)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vluxei8_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (const uint32_t *)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vluxei8_v_u32m2(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_u32m2((const uint32_t *)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vluxei8_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (const uint32_t *)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vluxei8_v_u32m4(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_u32m4((const uint32_t *)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vluxei8_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (const uint32_t *)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vluxei8_v_u32m8(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_u32m8((const uint32_t *)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vluxei8_v_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (const uint32_t *)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vluxei8_v_u32mf2(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_u32mf2((const uint32_t *)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vluxei8_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (const uint32_t *)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vluxei16_v_i32m1(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_i32m1((const int32_t *)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vluxei16_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (const int32_t *)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vluxei16_v_i32m2(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_i32m2((const int32_t *)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vluxei16_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (const int32_t *)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vluxei16_v_i32m4(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_i32m4((const int32_t *)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vluxei16_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (const int32_t *)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vluxei16_v_i32m8(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_i32m8((const int32_t *)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vluxei16_v_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (const int32_t *)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vluxei16_v_i32mf2(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_i32mf2((const int32_t *)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vluxei16_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (const int32_t *)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vluxei16_v_u32m1(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_u32m1((const uint32_t *)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vluxei16_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (const uint32_t *)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vluxei16_v_u32m2(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_u32m2((const uint32_t *)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vluxei16_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (const uint32_t *)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vluxei16_v_u32m4(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_u32m4((const uint32_t *)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vluxei16_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (const uint32_t *)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vluxei16_v_u32m8(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_u32m8((const uint32_t *)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vluxei16_v_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (const uint32_t *)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vluxei16_v_u32mf2(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_u32mf2((const uint32_t *)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vluxei16_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (const uint32_t *)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vluxei32_v_i32m1(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_i32m1((const int32_t *)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vluxei32_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (const int32_t *)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vluxei32_v_i32m2(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_i32m2((const int32_t *)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vluxei32_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (const int32_t *)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vluxei32_v_i32m4(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_i32m4((const int32_t *)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vluxei32_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (const int32_t *)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vluxei32_v_i32m8(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_i32m8((const int32_t *)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vluxei32_v_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (const int32_t *)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vluxei32_v_i32mf2(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_i32mf2((const int32_t *)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vluxei32_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (const int32_t *)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vluxei32_v_u32m1(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_u32m1((const uint32_t *)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vluxei32_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (const uint32_t *)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vluxei32_v_u32m2(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_u32m2((const uint32_t *)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vluxei32_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (const uint32_t *)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vluxei32_v_u32m4(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_u32m4((const uint32_t *)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vluxei32_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (const uint32_t *)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vluxei32_v_u32m8(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_u32m8((const uint32_t *)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vluxei32_v_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (const uint32_t *)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vluxei32_v_u32mf2(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_u32mf2((const uint32_t *)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vluxei32_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (const uint32_t *)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vluxei64_v_i32m1(op0, op1, op2) \ +__builtin_rvv_vluxei64_v_i32m1((const int32_t *)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vluxei64_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei64_v_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (const int32_t *)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vluxei64_v_i32m2(op0, op1, op2) \ +__builtin_rvv_vluxei64_v_i32m2((const int32_t *)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vluxei64_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei64_v_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (const int32_t *)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vluxei64_v_i32m4(op0, op1, op2) \ +__builtin_rvv_vluxei64_v_i32m4((const int32_t *)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vluxei64_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei64_v_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (const int32_t *)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vluxei64_v_i32mf2(op0, op1, op2) \ +__builtin_rvv_vluxei64_v_i32mf2((const int32_t *)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vluxei64_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei64_v_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (const int32_t *)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vluxei64_v_u32m1(op0, op1, op2) \ +__builtin_rvv_vluxei64_v_u32m1((const uint32_t *)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vluxei64_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei64_v_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (const uint32_t *)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vluxei64_v_u32m2(op0, op1, op2) \ +__builtin_rvv_vluxei64_v_u32m2((const uint32_t *)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vluxei64_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei64_v_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (const uint32_t *)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vluxei64_v_u32m4(op0, op1, op2) \ +__builtin_rvv_vluxei64_v_u32m4((const uint32_t *)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vluxei64_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei64_v_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (const uint32_t *)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vluxei64_v_u32mf2(op0, op1, op2) \ +__builtin_rvv_vluxei64_v_u32mf2((const uint32_t *)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vluxei64_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei64_v_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (const uint32_t *)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vluxei8_v_i64m1(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_i64m1((const int64_t *)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vluxei8_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (const int64_t *)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vluxei8_v_i64m2(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_i64m2((const int64_t *)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vluxei8_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (const int64_t *)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vluxei8_v_i64m4(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_i64m4((const int64_t *)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vluxei8_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (const int64_t *)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vluxei8_v_i64m8(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_i64m8((const int64_t *)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vluxei8_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (const int64_t *)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vluxei8_v_u64m1(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_u64m1((const uint64_t *)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vluxei8_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (const uint64_t *)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vluxei8_v_u64m2(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_u64m2((const uint64_t *)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vluxei8_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (const uint64_t *)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vluxei8_v_u64m4(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_u64m4((const uint64_t *)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vluxei8_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (const uint64_t *)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vluxei8_v_u64m8(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_u64m8((const uint64_t *)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vluxei8_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (const uint64_t *)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vluxei16_v_i64m1(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_i64m1((const int64_t *)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vluxei16_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (const int64_t *)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vluxei16_v_i64m2(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_i64m2((const int64_t *)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vluxei16_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (const int64_t *)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vluxei16_v_i64m4(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_i64m4((const int64_t *)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vluxei16_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (const int64_t *)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vluxei16_v_i64m8(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_i64m8((const int64_t *)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vluxei16_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (const int64_t *)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vluxei16_v_u64m1(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_u64m1((const uint64_t *)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vluxei16_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (const uint64_t *)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vluxei16_v_u64m2(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_u64m2((const uint64_t *)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vluxei16_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (const uint64_t *)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vluxei16_v_u64m4(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_u64m4((const uint64_t *)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vluxei16_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (const uint64_t *)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vluxei16_v_u64m8(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_u64m8((const uint64_t *)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vluxei16_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (const uint64_t *)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vluxei32_v_i64m1(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_i64m1((const int64_t *)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vluxei32_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (const int64_t *)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vluxei32_v_i64m2(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_i64m2((const int64_t *)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vluxei32_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (const int64_t *)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vluxei32_v_i64m4(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_i64m4((const int64_t *)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vluxei32_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (const int64_t *)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vluxei32_v_i64m8(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_i64m8((const int64_t *)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vluxei32_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (const int64_t *)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vluxei32_v_u64m1(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_u64m1((const uint64_t *)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vluxei32_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (const uint64_t *)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vluxei32_v_u64m2(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_u64m2((const uint64_t *)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vluxei32_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (const uint64_t *)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vluxei32_v_u64m4(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_u64m4((const uint64_t *)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vluxei32_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (const uint64_t *)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vluxei32_v_u64m8(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_u64m8((const uint64_t *)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vluxei32_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (const uint64_t *)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vluxei64_v_i64m1(op0, op1, op2) \ +__builtin_rvv_vluxei64_v_i64m1((const int64_t *)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vluxei64_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei64_v_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (const int64_t *)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vluxei64_v_i64m2(op0, op1, op2) \ +__builtin_rvv_vluxei64_v_i64m2((const int64_t *)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vluxei64_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei64_v_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (const int64_t *)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vluxei64_v_i64m4(op0, op1, op2) \ +__builtin_rvv_vluxei64_v_i64m4((const int64_t *)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vluxei64_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei64_v_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (const int64_t *)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vluxei64_v_i64m8(op0, op1, op2) \ +__builtin_rvv_vluxei64_v_i64m8((const int64_t *)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vluxei64_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei64_v_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (const int64_t *)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vluxei64_v_u64m1(op0, op1, op2) \ +__builtin_rvv_vluxei64_v_u64m1((const uint64_t *)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vluxei64_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei64_v_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (const uint64_t *)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vluxei64_v_u64m2(op0, op1, op2) \ +__builtin_rvv_vluxei64_v_u64m2((const uint64_t *)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vluxei64_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei64_v_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (const uint64_t *)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vluxei64_v_u64m4(op0, op1, op2) \ +__builtin_rvv_vluxei64_v_u64m4((const uint64_t *)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vluxei64_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei64_v_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (const uint64_t *)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vluxei64_v_u64m8(op0, op1, op2) \ +__builtin_rvv_vluxei64_v_u64m8((const uint64_t *)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vluxei64_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei64_v_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (const uint64_t *)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vloxei8_v_i8m1(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_i8m1((const int8_t *)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vloxei8_v_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (const int8_t *)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vloxei8_v_i8m2(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_i8m2((const int8_t *)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vloxei8_v_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (const int8_t *)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vloxei8_v_i8m4(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_i8m4((const int8_t *)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vloxei8_v_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (const int8_t *)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vloxei8_v_i8m8(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_i8m8((const int8_t *)(op0), (vuint8m8_t)(op1), (size_t)(op2)) +#define vloxei8_v_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (const int8_t *)(op2), (vuint8m8_t)(op3), (size_t)(op4)) +#define vloxei8_v_i8mf2(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_i8mf2((const int8_t *)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vloxei8_v_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (const int8_t *)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vloxei8_v_i8mf4(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_i8mf4((const int8_t *)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vloxei8_v_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (const int8_t *)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vloxei8_v_i8mf8(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_i8mf8((const int8_t *)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vloxei8_v_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (const int8_t *)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vloxei8_v_u8m1(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_u8m1((const uint8_t *)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vloxei8_v_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (const uint8_t *)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vloxei8_v_u8m2(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_u8m2((const uint8_t *)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vloxei8_v_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (const uint8_t *)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vloxei8_v_u8m4(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_u8m4((const uint8_t *)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vloxei8_v_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (const uint8_t *)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vloxei8_v_u8m8(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_u8m8((const uint8_t *)(op0), (vuint8m8_t)(op1), (size_t)(op2)) +#define vloxei8_v_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (const uint8_t *)(op2), (vuint8m8_t)(op3), (size_t)(op4)) +#define vloxei8_v_u8mf2(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_u8mf2((const uint8_t *)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vloxei8_v_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (const uint8_t *)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vloxei8_v_u8mf4(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_u8mf4((const uint8_t *)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vloxei8_v_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (const uint8_t *)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vloxei8_v_u8mf8(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_u8mf8((const uint8_t *)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vloxei8_v_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (const uint8_t *)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vloxei16_v_i8m1(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_i8m1((const int8_t *)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vloxei16_v_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (const int8_t *)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vloxei16_v_i8m2(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_i8m2((const int8_t *)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vloxei16_v_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (const int8_t *)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vloxei16_v_i8m4(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_i8m4((const int8_t *)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vloxei16_v_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (const int8_t *)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vloxei16_v_i8mf2(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_i8mf2((const int8_t *)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vloxei16_v_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (const int8_t *)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vloxei16_v_i8mf4(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_i8mf4((const int8_t *)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vloxei16_v_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (const int8_t *)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vloxei16_v_i8mf8(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_i8mf8((const int8_t *)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vloxei16_v_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (const int8_t *)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vloxei16_v_u8m1(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_u8m1((const uint8_t *)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vloxei16_v_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (const uint8_t *)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vloxei16_v_u8m2(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_u8m2((const uint8_t *)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vloxei16_v_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (const uint8_t *)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vloxei16_v_u8m4(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_u8m4((const uint8_t *)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vloxei16_v_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (const uint8_t *)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vloxei16_v_u8mf2(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_u8mf2((const uint8_t *)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vloxei16_v_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (const uint8_t *)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vloxei16_v_u8mf4(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_u8mf4((const uint8_t *)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vloxei16_v_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (const uint8_t *)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vloxei16_v_u8mf8(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_u8mf8((const uint8_t *)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vloxei16_v_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (const uint8_t *)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vloxei32_v_i8m1(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_i8m1((const int8_t *)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vloxei32_v_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (const int8_t *)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vloxei32_v_i8m2(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_i8m2((const int8_t *)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vloxei32_v_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (const int8_t *)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vloxei32_v_i8mf2(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_i8mf2((const int8_t *)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vloxei32_v_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (const int8_t *)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vloxei32_v_i8mf4(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_i8mf4((const int8_t *)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vloxei32_v_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (const int8_t *)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vloxei32_v_i8mf8(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_i8mf8((const int8_t *)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vloxei32_v_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (const int8_t *)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vloxei32_v_u8m1(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_u8m1((const uint8_t *)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vloxei32_v_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (const uint8_t *)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vloxei32_v_u8m2(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_u8m2((const uint8_t *)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vloxei32_v_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (const uint8_t *)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vloxei32_v_u8mf2(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_u8mf2((const uint8_t *)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vloxei32_v_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (const uint8_t *)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vloxei32_v_u8mf4(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_u8mf4((const uint8_t *)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vloxei32_v_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (const uint8_t *)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vloxei32_v_u8mf8(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_u8mf8((const uint8_t *)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vloxei32_v_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (const uint8_t *)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vloxei64_v_i8m1(op0, op1, op2) \ +__builtin_rvv_vloxei64_v_i8m1((const int8_t *)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vloxei64_v_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei64_v_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (const int8_t *)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vloxei64_v_i8mf2(op0, op1, op2) \ +__builtin_rvv_vloxei64_v_i8mf2((const int8_t *)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vloxei64_v_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei64_v_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (const int8_t *)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vloxei64_v_i8mf4(op0, op1, op2) \ +__builtin_rvv_vloxei64_v_i8mf4((const int8_t *)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vloxei64_v_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei64_v_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (const int8_t *)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vloxei64_v_i8mf8(op0, op1, op2) \ +__builtin_rvv_vloxei64_v_i8mf8((const int8_t *)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vloxei64_v_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei64_v_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (const int8_t *)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vloxei64_v_u8m1(op0, op1, op2) \ +__builtin_rvv_vloxei64_v_u8m1((const uint8_t *)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vloxei64_v_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei64_v_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (const uint8_t *)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vloxei64_v_u8mf2(op0, op1, op2) \ +__builtin_rvv_vloxei64_v_u8mf2((const uint8_t *)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vloxei64_v_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei64_v_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (const uint8_t *)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vloxei64_v_u8mf4(op0, op1, op2) \ +__builtin_rvv_vloxei64_v_u8mf4((const uint8_t *)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vloxei64_v_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei64_v_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (const uint8_t *)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vloxei64_v_u8mf8(op0, op1, op2) \ +__builtin_rvv_vloxei64_v_u8mf8((const uint8_t *)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vloxei64_v_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei64_v_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (const uint8_t *)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vloxei8_v_i16m1(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_i16m1((const int16_t *)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vloxei8_v_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (const int16_t *)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vloxei8_v_i16m2(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_i16m2((const int16_t *)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vloxei8_v_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (const int16_t *)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vloxei8_v_i16m4(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_i16m4((const int16_t *)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vloxei8_v_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (const int16_t *)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vloxei8_v_i16m8(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_i16m8((const int16_t *)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vloxei8_v_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (const int16_t *)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vloxei8_v_i16mf2(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_i16mf2((const int16_t *)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vloxei8_v_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (const int16_t *)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vloxei8_v_i16mf4(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_i16mf4((const int16_t *)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vloxei8_v_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (const int16_t *)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vloxei8_v_u16m1(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_u16m1((const uint16_t *)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vloxei8_v_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (const uint16_t *)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vloxei8_v_u16m2(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_u16m2((const uint16_t *)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vloxei8_v_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (const uint16_t *)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vloxei8_v_u16m4(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_u16m4((const uint16_t *)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vloxei8_v_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (const uint16_t *)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vloxei8_v_u16m8(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_u16m8((const uint16_t *)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vloxei8_v_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (const uint16_t *)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vloxei8_v_u16mf2(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_u16mf2((const uint16_t *)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vloxei8_v_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (const uint16_t *)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vloxei8_v_u16mf4(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_u16mf4((const uint16_t *)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vloxei8_v_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (const uint16_t *)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vloxei16_v_i16m1(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_i16m1((const int16_t *)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vloxei16_v_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (const int16_t *)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vloxei16_v_i16m2(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_i16m2((const int16_t *)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vloxei16_v_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (const int16_t *)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vloxei16_v_i16m4(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_i16m4((const int16_t *)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vloxei16_v_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (const int16_t *)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vloxei16_v_i16m8(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_i16m8((const int16_t *)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vloxei16_v_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (const int16_t *)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vloxei16_v_i16mf2(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_i16mf2((const int16_t *)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vloxei16_v_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (const int16_t *)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vloxei16_v_i16mf4(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_i16mf4((const int16_t *)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vloxei16_v_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (const int16_t *)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vloxei16_v_u16m1(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_u16m1((const uint16_t *)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vloxei16_v_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (const uint16_t *)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vloxei16_v_u16m2(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_u16m2((const uint16_t *)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vloxei16_v_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (const uint16_t *)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vloxei16_v_u16m4(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_u16m4((const uint16_t *)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vloxei16_v_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (const uint16_t *)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vloxei16_v_u16m8(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_u16m8((const uint16_t *)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vloxei16_v_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (const uint16_t *)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vloxei16_v_u16mf2(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_u16mf2((const uint16_t *)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vloxei16_v_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (const uint16_t *)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vloxei16_v_u16mf4(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_u16mf4((const uint16_t *)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vloxei16_v_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (const uint16_t *)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vloxei32_v_i16m1(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_i16m1((const int16_t *)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vloxei32_v_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (const int16_t *)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vloxei32_v_i16m2(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_i16m2((const int16_t *)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vloxei32_v_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (const int16_t *)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vloxei32_v_i16m4(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_i16m4((const int16_t *)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vloxei32_v_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (const int16_t *)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vloxei32_v_i16mf2(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_i16mf2((const int16_t *)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vloxei32_v_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (const int16_t *)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vloxei32_v_i16mf4(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_i16mf4((const int16_t *)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vloxei32_v_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (const int16_t *)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vloxei32_v_u16m1(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_u16m1((const uint16_t *)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vloxei32_v_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (const uint16_t *)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vloxei32_v_u16m2(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_u16m2((const uint16_t *)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vloxei32_v_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (const uint16_t *)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vloxei32_v_u16m4(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_u16m4((const uint16_t *)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vloxei32_v_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (const uint16_t *)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vloxei32_v_u16mf2(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_u16mf2((const uint16_t *)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vloxei32_v_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (const uint16_t *)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vloxei32_v_u16mf4(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_u16mf4((const uint16_t *)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vloxei32_v_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (const uint16_t *)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vloxei64_v_i16m1(op0, op1, op2) \ +__builtin_rvv_vloxei64_v_i16m1((const int16_t *)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vloxei64_v_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei64_v_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (const int16_t *)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vloxei64_v_i16m2(op0, op1, op2) \ +__builtin_rvv_vloxei64_v_i16m2((const int16_t *)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vloxei64_v_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei64_v_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (const int16_t *)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vloxei64_v_i16mf2(op0, op1, op2) \ +__builtin_rvv_vloxei64_v_i16mf2((const int16_t *)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vloxei64_v_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei64_v_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (const int16_t *)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vloxei64_v_i16mf4(op0, op1, op2) \ +__builtin_rvv_vloxei64_v_i16mf4((const int16_t *)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vloxei64_v_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei64_v_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (const int16_t *)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vloxei64_v_u16m1(op0, op1, op2) \ +__builtin_rvv_vloxei64_v_u16m1((const uint16_t *)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vloxei64_v_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei64_v_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (const uint16_t *)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vloxei64_v_u16m2(op0, op1, op2) \ +__builtin_rvv_vloxei64_v_u16m2((const uint16_t *)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vloxei64_v_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei64_v_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (const uint16_t *)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vloxei64_v_u16mf2(op0, op1, op2) \ +__builtin_rvv_vloxei64_v_u16mf2((const uint16_t *)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vloxei64_v_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei64_v_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (const uint16_t *)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vloxei64_v_u16mf4(op0, op1, op2) \ +__builtin_rvv_vloxei64_v_u16mf4((const uint16_t *)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vloxei64_v_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei64_v_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (const uint16_t *)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vloxei8_v_i32m1(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_i32m1((const int32_t *)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vloxei8_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (const int32_t *)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vloxei8_v_i32m2(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_i32m2((const int32_t *)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vloxei8_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (const int32_t *)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vloxei8_v_i32m4(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_i32m4((const int32_t *)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vloxei8_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (const int32_t *)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vloxei8_v_i32m8(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_i32m8((const int32_t *)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vloxei8_v_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (const int32_t *)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vloxei8_v_i32mf2(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_i32mf2((const int32_t *)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vloxei8_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (const int32_t *)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vloxei8_v_u32m1(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_u32m1((const uint32_t *)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vloxei8_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (const uint32_t *)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vloxei8_v_u32m2(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_u32m2((const uint32_t *)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vloxei8_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (const uint32_t *)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vloxei8_v_u32m4(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_u32m4((const uint32_t *)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vloxei8_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (const uint32_t *)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vloxei8_v_u32m8(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_u32m8((const uint32_t *)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vloxei8_v_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (const uint32_t *)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vloxei8_v_u32mf2(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_u32mf2((const uint32_t *)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vloxei8_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (const uint32_t *)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vloxei16_v_i32m1(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_i32m1((const int32_t *)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vloxei16_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (const int32_t *)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vloxei16_v_i32m2(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_i32m2((const int32_t *)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vloxei16_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (const int32_t *)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vloxei16_v_i32m4(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_i32m4((const int32_t *)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vloxei16_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (const int32_t *)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vloxei16_v_i32m8(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_i32m8((const int32_t *)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vloxei16_v_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (const int32_t *)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vloxei16_v_i32mf2(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_i32mf2((const int32_t *)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vloxei16_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (const int32_t *)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vloxei16_v_u32m1(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_u32m1((const uint32_t *)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vloxei16_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (const uint32_t *)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vloxei16_v_u32m2(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_u32m2((const uint32_t *)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vloxei16_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (const uint32_t *)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vloxei16_v_u32m4(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_u32m4((const uint32_t *)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vloxei16_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (const uint32_t *)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vloxei16_v_u32m8(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_u32m8((const uint32_t *)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vloxei16_v_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (const uint32_t *)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vloxei16_v_u32mf2(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_u32mf2((const uint32_t *)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vloxei16_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (const uint32_t *)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vloxei32_v_i32m1(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_i32m1((const int32_t *)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vloxei32_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (const int32_t *)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vloxei32_v_i32m2(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_i32m2((const int32_t *)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vloxei32_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (const int32_t *)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vloxei32_v_i32m4(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_i32m4((const int32_t *)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vloxei32_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (const int32_t *)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vloxei32_v_i32m8(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_i32m8((const int32_t *)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vloxei32_v_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (const int32_t *)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vloxei32_v_i32mf2(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_i32mf2((const int32_t *)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vloxei32_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (const int32_t *)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vloxei32_v_u32m1(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_u32m1((const uint32_t *)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vloxei32_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (const uint32_t *)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vloxei32_v_u32m2(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_u32m2((const uint32_t *)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vloxei32_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (const uint32_t *)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vloxei32_v_u32m4(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_u32m4((const uint32_t *)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vloxei32_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (const uint32_t *)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vloxei32_v_u32m8(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_u32m8((const uint32_t *)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vloxei32_v_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (const uint32_t *)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vloxei32_v_u32mf2(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_u32mf2((const uint32_t *)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vloxei32_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (const uint32_t *)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vloxei64_v_i32m1(op0, op1, op2) \ +__builtin_rvv_vloxei64_v_i32m1((const int32_t *)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vloxei64_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei64_v_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (const int32_t *)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vloxei64_v_i32m2(op0, op1, op2) \ +__builtin_rvv_vloxei64_v_i32m2((const int32_t *)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vloxei64_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei64_v_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (const int32_t *)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vloxei64_v_i32m4(op0, op1, op2) \ +__builtin_rvv_vloxei64_v_i32m4((const int32_t *)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vloxei64_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei64_v_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (const int32_t *)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vloxei64_v_i32mf2(op0, op1, op2) \ +__builtin_rvv_vloxei64_v_i32mf2((const int32_t *)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vloxei64_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei64_v_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (const int32_t *)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vloxei64_v_u32m1(op0, op1, op2) \ +__builtin_rvv_vloxei64_v_u32m1((const uint32_t *)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vloxei64_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei64_v_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (const uint32_t *)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vloxei64_v_u32m2(op0, op1, op2) \ +__builtin_rvv_vloxei64_v_u32m2((const uint32_t *)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vloxei64_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei64_v_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (const uint32_t *)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vloxei64_v_u32m4(op0, op1, op2) \ +__builtin_rvv_vloxei64_v_u32m4((const uint32_t *)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vloxei64_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei64_v_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (const uint32_t *)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vloxei64_v_u32mf2(op0, op1, op2) \ +__builtin_rvv_vloxei64_v_u32mf2((const uint32_t *)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vloxei64_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei64_v_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (const uint32_t *)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vloxei8_v_i64m1(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_i64m1((const int64_t *)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vloxei8_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (const int64_t *)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vloxei8_v_i64m2(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_i64m2((const int64_t *)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vloxei8_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (const int64_t *)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vloxei8_v_i64m4(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_i64m4((const int64_t *)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vloxei8_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (const int64_t *)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vloxei8_v_i64m8(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_i64m8((const int64_t *)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vloxei8_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (const int64_t *)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vse8_v_i8m1(op0, op1, op2) \ +__builtin_rvv_vse8_v_i8m1((int8_t *)(op0), (vint8m1_t)(op1), (size_t)(op2)) +#define vse8_v_i8m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vse8_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vse8_v_i8m2(op0, op1, op2) \ +__builtin_rvv_vse8_v_i8m2((int8_t *)(op0), (vint8m2_t)(op1), (size_t)(op2)) +#define vse8_v_i8m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vse8_v_i8m2_m((vbool4_t)(op0), (int8_t *)(op1), (vint8m2_t)(op2), (size_t)(op3)) +#define vse8_v_i8m4(op0, op1, op2) \ +__builtin_rvv_vse8_v_i8m4((int8_t *)(op0), (vint8m4_t)(op1), (size_t)(op2)) +#define vse8_v_i8m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vse8_v_i8m4_m((vbool2_t)(op0), (int8_t *)(op1), (vint8m4_t)(op2), (size_t)(op3)) +#define vse8_v_i8m8(op0, op1, op2) \ +__builtin_rvv_vse8_v_i8m8((int8_t *)(op0), (vint8m8_t)(op1), (size_t)(op2)) +#define vse8_v_i8m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vse8_v_i8m8_m((vbool1_t)(op0), (int8_t *)(op1), (vint8m8_t)(op2), (size_t)(op3)) +#define vse8_v_i8mf2(op0, op1, op2) \ +__builtin_rvv_vse8_v_i8mf2((int8_t *)(op0), (vint8mf2_t)(op1), (size_t)(op2)) +#define vse8_v_i8mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vse8_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vint8mf2_t)(op2), (size_t)(op3)) +#define vse8_v_i8mf4(op0, op1, op2) \ +__builtin_rvv_vse8_v_i8mf4((int8_t *)(op0), (vint8mf4_t)(op1), (size_t)(op2)) +#define vse8_v_i8mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vse8_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vint8mf4_t)(op2), (size_t)(op3)) +#define vse8_v_i8mf8(op0, op1, op2) \ +__builtin_rvv_vse8_v_i8mf8((int8_t *)(op0), (vint8mf8_t)(op1), (size_t)(op2)) +#define vse8_v_i8mf8_m(op0, op1, op2, op3) \ +__builtin_rvv_vse8_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vint8mf8_t)(op2), (size_t)(op3)) +#define vloxei8_v_u64m1(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_u64m1((const uint64_t *)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vloxei8_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (const uint64_t *)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vloxei8_v_u64m2(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_u64m2((const uint64_t *)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vloxei8_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (const uint64_t *)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vloxei8_v_u64m4(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_u64m4((const uint64_t *)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vloxei8_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (const uint64_t *)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vloxei8_v_u64m8(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_u64m8((const uint64_t *)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vloxei8_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (const uint64_t *)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vloxei16_v_i64m1(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_i64m1((const int64_t *)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vloxei16_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (const int64_t *)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vloxei16_v_i64m2(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_i64m2((const int64_t *)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vloxei16_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (const int64_t *)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vloxei16_v_i64m4(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_i64m4((const int64_t *)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vloxei16_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (const int64_t *)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vloxei16_v_i64m8(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_i64m8((const int64_t *)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vloxei16_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (const int64_t *)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vloxei16_v_u64m1(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_u64m1((const uint64_t *)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vloxei16_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (const uint64_t *)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vloxei16_v_u64m2(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_u64m2((const uint64_t *)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vloxei16_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (const uint64_t *)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vloxei16_v_u64m4(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_u64m4((const uint64_t *)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vloxei16_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (const uint64_t *)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vloxei16_v_u64m8(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_u64m8((const uint64_t *)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vloxei16_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (const uint64_t *)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vloxei32_v_i64m1(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_i64m1((const int64_t *)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vloxei32_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (const int64_t *)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vloxei32_v_i64m2(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_i64m2((const int64_t *)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vloxei32_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (const int64_t *)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vloxei32_v_i64m4(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_i64m4((const int64_t *)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vloxei32_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (const int64_t *)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vloxei32_v_i64m8(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_i64m8((const int64_t *)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vloxei32_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (const int64_t *)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vloxei32_v_u64m1(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_u64m1((const uint64_t *)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vloxei32_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (const uint64_t *)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vloxei32_v_u64m2(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_u64m2((const uint64_t *)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vloxei32_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (const uint64_t *)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vloxei32_v_u64m4(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_u64m4((const uint64_t *)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vloxei32_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (const uint64_t *)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vloxei32_v_u64m8(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_u64m8((const uint64_t *)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vloxei32_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (const uint64_t *)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vloxei64_v_i64m1(op0, op1, op2) \ +__builtin_rvv_vloxei64_v_i64m1((const int64_t *)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vloxei64_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei64_v_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (const int64_t *)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vloxei64_v_i64m2(op0, op1, op2) \ +__builtin_rvv_vloxei64_v_i64m2((const int64_t *)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vloxei64_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei64_v_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (const int64_t *)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vloxei64_v_i64m4(op0, op1, op2) \ +__builtin_rvv_vloxei64_v_i64m4((const int64_t *)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vloxei64_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei64_v_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (const int64_t *)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vloxei64_v_i64m8(op0, op1, op2) \ +__builtin_rvv_vloxei64_v_i64m8((const int64_t *)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vloxei64_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei64_v_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (const int64_t *)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vloxei64_v_u64m1(op0, op1, op2) \ +__builtin_rvv_vloxei64_v_u64m1((const uint64_t *)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vloxei64_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei64_v_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (const uint64_t *)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vloxei64_v_u64m2(op0, op1, op2) \ +__builtin_rvv_vloxei64_v_u64m2((const uint64_t *)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vloxei64_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei64_v_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (const uint64_t *)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vloxei64_v_u64m4(op0, op1, op2) \ +__builtin_rvv_vloxei64_v_u64m4((const uint64_t *)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vloxei64_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei64_v_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (const uint64_t *)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vloxei64_v_u64m8(op0, op1, op2) \ +__builtin_rvv_vloxei64_v_u64m8((const uint64_t *)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vloxei64_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei64_v_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (const uint64_t *)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vse8_v_u8m1(op0, op1, op2) \ +__builtin_rvv_vse8_v_u8m1((uint8_t *)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vse8_v_u8m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vse8_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vse8_v_u8m2(op0, op1, op2) \ +__builtin_rvv_vse8_v_u8m2((uint8_t *)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vse8_v_u8m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vse8_v_u8m2_m((vbool4_t)(op0), (uint8_t *)(op1), (vuint8m2_t)(op2), (size_t)(op3)) +#define vse8_v_u8m4(op0, op1, op2) \ +__builtin_rvv_vse8_v_u8m4((uint8_t *)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vse8_v_u8m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vse8_v_u8m4_m((vbool2_t)(op0), (uint8_t *)(op1), (vuint8m4_t)(op2), (size_t)(op3)) +#define vse8_v_u8m8(op0, op1, op2) \ +__builtin_rvv_vse8_v_u8m8((uint8_t *)(op0), (vuint8m8_t)(op1), (size_t)(op2)) +#define vse8_v_u8m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vse8_v_u8m8_m((vbool1_t)(op0), (uint8_t *)(op1), (vuint8m8_t)(op2), (size_t)(op3)) +#define vse8_v_u8mf2(op0, op1, op2) \ +__builtin_rvv_vse8_v_u8mf2((uint8_t *)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vse8_v_u8mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vse8_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint8mf2_t)(op2), (size_t)(op3)) +#define vse8_v_u8mf4(op0, op1, op2) \ +__builtin_rvv_vse8_v_u8mf4((uint8_t *)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vse8_v_u8mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vse8_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint8mf4_t)(op2), (size_t)(op3)) +#define vse8_v_u8mf8(op0, op1, op2) \ +__builtin_rvv_vse8_v_u8mf8((uint8_t *)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vse8_v_u8mf8_m(op0, op1, op2, op3) \ +__builtin_rvv_vse8_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint8mf8_t)(op2), (size_t)(op3)) +#define vwaddu_wv_u16mf4(op0, op1, op2) \ +__builtin_rvv_vwaddu_wv_u16mf4((vuint16mf4_t)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vwaddu_wv_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_wv_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vwaddu_wv_u16mf2(op0, op1, op2) \ +__builtin_rvv_vwaddu_wv_u16mf2((vuint16mf2_t)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vwaddu_wv_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_wv_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vwaddu_wv_u16m1(op0, op1, op2) \ +__builtin_rvv_vwaddu_wv_u16m1((vuint16m1_t)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vwaddu_wv_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_wv_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vwaddu_wv_u16m2(op0, op1, op2) \ +__builtin_rvv_vwaddu_wv_u16m2((vuint16m2_t)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vwaddu_wv_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_wv_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vwaddu_wv_u16m4(op0, op1, op2) \ +__builtin_rvv_vwaddu_wv_u16m4((vuint16m4_t)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vwaddu_wv_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_wv_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vwaddu_wv_u16m8(op0, op1, op2) \ +__builtin_rvv_vwaddu_wv_u16m8((vuint16m8_t)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vwaddu_wv_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_wv_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vwaddu_wv_u32mf2(op0, op1, op2) \ +__builtin_rvv_vwaddu_wv_u32mf2((vuint32mf2_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vwaddu_wv_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_wv_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vwaddu_wv_u32m1(op0, op1, op2) \ +__builtin_rvv_vwaddu_wv_u32m1((vuint32m1_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vwaddu_wv_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_wv_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vwaddu_wv_u32m2(op0, op1, op2) \ +__builtin_rvv_vwaddu_wv_u32m2((vuint32m2_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vwaddu_wv_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_wv_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vwaddu_wv_u32m4(op0, op1, op2) \ +__builtin_rvv_vwaddu_wv_u32m4((vuint32m4_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vwaddu_wv_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_wv_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vwaddu_wv_u32m8(op0, op1, op2) \ +__builtin_rvv_vwaddu_wv_u32m8((vuint32m8_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vwaddu_wv_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_wv_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vwaddu_wv_u64m1(op0, op1, op2) \ +__builtin_rvv_vwaddu_wv_u64m1((vuint64m1_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vwaddu_wv_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_wv_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vwaddu_wv_u64m2(op0, op1, op2) \ +__builtin_rvv_vwaddu_wv_u64m2((vuint64m2_t)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vwaddu_wv_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_wv_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vwaddu_wv_u64m4(op0, op1, op2) \ +__builtin_rvv_vwaddu_wv_u64m4((vuint64m4_t)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vwaddu_wv_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_wv_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vwaddu_wv_u64m8(op0, op1, op2) \ +__builtin_rvv_vwaddu_wv_u64m8((vuint64m8_t)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vwaddu_wv_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_wv_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vsse8_v_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vsse8_v_i8m1((int8_t *)(op0), (ptrdiff_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vsse8_v_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse8_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (ptrdiff_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vsse8_v_i8m2(op0, op1, op2, op3) \ +__builtin_rvv_vsse8_v_i8m2((int8_t *)(op0), (ptrdiff_t)(op1), (vint8m2_t)(op2), (size_t)(op3)) +#define vsse8_v_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse8_v_i8m2_m((vbool4_t)(op0), (int8_t *)(op1), (ptrdiff_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vsse8_v_i8m4(op0, op1, op2, op3) \ +__builtin_rvv_vsse8_v_i8m4((int8_t *)(op0), (ptrdiff_t)(op1), (vint8m4_t)(op2), (size_t)(op3)) +#define vsse8_v_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse8_v_i8m4_m((vbool2_t)(op0), (int8_t *)(op1), (ptrdiff_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vsse8_v_i8m8(op0, op1, op2, op3) \ +__builtin_rvv_vsse8_v_i8m8((int8_t *)(op0), (ptrdiff_t)(op1), (vint8m8_t)(op2), (size_t)(op3)) +#define vsse8_v_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse8_v_i8m8_m((vbool1_t)(op0), (int8_t *)(op1), (ptrdiff_t)(op2), (vint8m8_t)(op3), (size_t)(op4)) +#define vsse8_v_i8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsse8_v_i8mf2((int8_t *)(op0), (ptrdiff_t)(op1), (vint8mf2_t)(op2), (size_t)(op3)) +#define vsse8_v_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse8_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (ptrdiff_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vsse8_v_i8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsse8_v_i8mf4((int8_t *)(op0), (ptrdiff_t)(op1), (vint8mf4_t)(op2), (size_t)(op3)) +#define vsse8_v_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse8_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (ptrdiff_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vsse8_v_i8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vsse8_v_i8mf8((int8_t *)(op0), (ptrdiff_t)(op1), (vint8mf8_t)(op2), (size_t)(op3)) +#define vsse8_v_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse8_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (ptrdiff_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vsuxei16_v_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_i8m1((int8_t *)(op0), (vuint16m2_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vsuxei16_v_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint16m2_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vsuxei16_v_i8m2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_i8m2((int8_t *)(op0), (vuint16m4_t)(op1), (vint8m2_t)(op2), (size_t)(op3)) +#define vsuxei16_v_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_i8m2_m((vbool4_t)(op0), (int8_t *)(op1), (vuint16m4_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vsuxei16_v_i8m4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_i8m4((int8_t *)(op0), (vuint16m8_t)(op1), (vint8m4_t)(op2), (size_t)(op3)) +#define vsuxei16_v_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_i8m4_m((vbool2_t)(op0), (int8_t *)(op1), (vuint16m8_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vsuxei16_v_i8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_i8mf2((int8_t *)(op0), (vuint16m1_t)(op1), (vint8mf2_t)(op2), (size_t)(op3)) +#define vsuxei16_v_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint16m1_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vsuxei16_v_i8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_i8mf4((int8_t *)(op0), (vuint16mf2_t)(op1), (vint8mf4_t)(op2), (size_t)(op3)) +#define vsuxei16_v_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint16mf2_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vsuxei16_v_i8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_i8mf8((int8_t *)(op0), (vuint16mf4_t)(op1), (vint8mf8_t)(op2), (size_t)(op3)) +#define vsuxei16_v_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint16mf4_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vsuxei16_v_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_u8m1((uint8_t *)(op0), (vuint16m2_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vsuxei16_v_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint16m2_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vsuxei16_v_u8m2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_u8m2((uint8_t *)(op0), (vuint16m4_t)(op1), (vuint8m2_t)(op2), (size_t)(op3)) +#define vsuxei16_v_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_u8m2_m((vbool4_t)(op0), (uint8_t *)(op1), (vuint16m4_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vsuxei16_v_u8m4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_u8m4((uint8_t *)(op0), (vuint16m8_t)(op1), (vuint8m4_t)(op2), (size_t)(op3)) +#define vsuxei16_v_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_u8m4_m((vbool2_t)(op0), (uint8_t *)(op1), (vuint16m8_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vsuxei16_v_u8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_u8mf2((uint8_t *)(op0), (vuint16m1_t)(op1), (vuint8mf2_t)(op2), (size_t)(op3)) +#define vsuxei16_v_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint16m1_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vsuxei16_v_u8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_u8mf4((uint8_t *)(op0), (vuint16mf2_t)(op1), (vuint8mf4_t)(op2), (size_t)(op3)) +#define vsuxei16_v_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint16mf2_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vsuxei16_v_u8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_u8mf8((uint8_t *)(op0), (vuint16mf4_t)(op1), (vuint8mf8_t)(op2), (size_t)(op3)) +#define vsuxei16_v_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint16mf4_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vsuxei32_v_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_i8m1((int8_t *)(op0), (vuint32m4_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vsuxei32_v_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint32m4_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vsuxei32_v_i8m2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_i8m2((int8_t *)(op0), (vuint32m8_t)(op1), (vint8m2_t)(op2), (size_t)(op3)) +#define vsuxei32_v_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_i8m2_m((vbool4_t)(op0), (int8_t *)(op1), (vuint32m8_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vsuxei32_v_i8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_i8mf2((int8_t *)(op0), (vuint32m2_t)(op1), (vint8mf2_t)(op2), (size_t)(op3)) +#define vsuxei32_v_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint32m2_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vsuxei32_v_i8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_i8mf4((int8_t *)(op0), (vuint32m1_t)(op1), (vint8mf4_t)(op2), (size_t)(op3)) +#define vsuxei32_v_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint32m1_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vsuxei32_v_i8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_i8mf8((int8_t *)(op0), (vuint32mf2_t)(op1), (vint8mf8_t)(op2), (size_t)(op3)) +#define vsuxei32_v_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint32mf2_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vadd_vx_i8m1(op0, op1, op2) \ +__builtin_rvv_vadd_vx_i8m1((vint8m1_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vadd_vx_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vx_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vadd_vx_i8m2(op0, op1, op2) \ +__builtin_rvv_vadd_vx_i8m2((vint8m2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vadd_vx_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vx_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vadd_vx_i8m4(op0, op1, op2) \ +__builtin_rvv_vadd_vx_i8m4((vint8m4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vadd_vx_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vx_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vadd_vx_i8m8(op0, op1, op2) \ +__builtin_rvv_vadd_vx_i8m8((vint8m8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vadd_vx_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vx_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vadd_vx_i8mf2(op0, op1, op2) \ +__builtin_rvv_vadd_vx_i8mf2((vint8mf2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vadd_vx_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vx_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vadd_vx_i8mf4(op0, op1, op2) \ +__builtin_rvv_vadd_vx_i8mf4((vint8mf4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vadd_vx_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vx_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vadd_vx_i8mf8(op0, op1, op2) \ +__builtin_rvv_vadd_vx_i8mf8((vint8mf8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vadd_vx_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vx_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vadd_vx_i16m1(op0, op1, op2) \ +__builtin_rvv_vadd_vx_i16m1((vint16m1_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vadd_vx_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vx_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vadd_vx_i16m2(op0, op1, op2) \ +__builtin_rvv_vadd_vx_i16m2((vint16m2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vadd_vx_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vx_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vadd_vx_i16m4(op0, op1, op2) \ +__builtin_rvv_vadd_vx_i16m4((vint16m4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vadd_vx_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vx_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vadd_vx_i16m8(op0, op1, op2) \ +__builtin_rvv_vadd_vx_i16m8((vint16m8_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vadd_vx_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vx_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vadd_vx_i16mf2(op0, op1, op2) \ +__builtin_rvv_vadd_vx_i16mf2((vint16mf2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vadd_vx_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vx_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vadd_vx_i16mf4(op0, op1, op2) \ +__builtin_rvv_vadd_vx_i16mf4((vint16mf4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vadd_vx_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vx_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vadd_vx_i32m1(op0, op1, op2) \ +__builtin_rvv_vadd_vx_i32m1((vint32m1_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vadd_vx_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vx_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vadd_vx_i32m2(op0, op1, op2) \ +__builtin_rvv_vadd_vx_i32m2((vint32m2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vadd_vx_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vx_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vadd_vx_i32m4(op0, op1, op2) \ +__builtin_rvv_vadd_vx_i32m4((vint32m4_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vadd_vx_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vx_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vadd_vx_i32m8(op0, op1, op2) \ +__builtin_rvv_vadd_vx_i32m8((vint32m8_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vadd_vx_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vx_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vadd_vx_i32mf2(op0, op1, op2) \ +__builtin_rvv_vadd_vx_i32mf2((vint32mf2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vadd_vx_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vx_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vadd_vx_i64m1(op0, op1, op2) \ +__builtin_rvv_vadd_vx_i64m1((vint64m1_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vadd_vx_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vx_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vadd_vx_i64m2(op0, op1, op2) \ +__builtin_rvv_vadd_vx_i64m2((vint64m2_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vadd_vx_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vx_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vadd_vx_i64m4(op0, op1, op2) \ +__builtin_rvv_vadd_vx_i64m4((vint64m4_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vadd_vx_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vx_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vadd_vx_i64m8(op0, op1, op2) \ +__builtin_rvv_vadd_vx_i64m8((vint64m8_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vadd_vx_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vx_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vadd_vv_u8m1(op0, op1, op2) \ +__builtin_rvv_vadd_vv_u8m1((vuint8m1_t)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vadd_vv_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vv_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vadd_vv_u8m2(op0, op1, op2) \ +__builtin_rvv_vadd_vv_u8m2((vuint8m2_t)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vadd_vv_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vv_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vadd_vv_u8m4(op0, op1, op2) \ +__builtin_rvv_vadd_vv_u8m4((vuint8m4_t)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vadd_vv_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vv_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vadd_vv_u8m8(op0, op1, op2) \ +__builtin_rvv_vadd_vv_u8m8((vuint8m8_t)(op0), (vuint8m8_t)(op1), (size_t)(op2)) +#define vadd_vv_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vv_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (vuint8m8_t)(op3), (size_t)(op4)) +#define vadd_vv_u8mf2(op0, op1, op2) \ +__builtin_rvv_vadd_vv_u8mf2((vuint8mf2_t)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vadd_vv_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vv_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vadd_vv_u8mf4(op0, op1, op2) \ +__builtin_rvv_vadd_vv_u8mf4((vuint8mf4_t)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vadd_vv_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vv_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vadd_vv_u8mf8(op0, op1, op2) \ +__builtin_rvv_vadd_vv_u8mf8((vuint8mf8_t)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vadd_vv_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vv_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vadd_vv_u16m1(op0, op1, op2) \ +__builtin_rvv_vadd_vv_u16m1((vuint16m1_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vadd_vv_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vv_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vadd_vv_u16m2(op0, op1, op2) \ +__builtin_rvv_vadd_vv_u16m2((vuint16m2_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vadd_vv_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vv_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vadd_vv_u16m4(op0, op1, op2) \ +__builtin_rvv_vadd_vv_u16m4((vuint16m4_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vadd_vv_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vv_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vadd_vv_u16m8(op0, op1, op2) \ +__builtin_rvv_vadd_vv_u16m8((vuint16m8_t)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vadd_vv_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vv_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vadd_vv_u16mf2(op0, op1, op2) \ +__builtin_rvv_vadd_vv_u16mf2((vuint16mf2_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vadd_vv_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vv_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vadd_vv_u16mf4(op0, op1, op2) \ +__builtin_rvv_vadd_vv_u16mf4((vuint16mf4_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vadd_vv_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vv_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vadd_vv_u32m1(op0, op1, op2) \ +__builtin_rvv_vadd_vv_u32m1((vuint32m1_t)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vadd_vv_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vv_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vadd_vv_u32m2(op0, op1, op2) \ +__builtin_rvv_vadd_vv_u32m2((vuint32m2_t)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vadd_vv_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vv_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vadd_vv_u32m4(op0, op1, op2) \ +__builtin_rvv_vadd_vv_u32m4((vuint32m4_t)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vadd_vv_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vv_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vadd_vv_u32m8(op0, op1, op2) \ +__builtin_rvv_vadd_vv_u32m8((vuint32m8_t)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vadd_vv_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vv_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vadd_vv_u32mf2(op0, op1, op2) \ +__builtin_rvv_vadd_vv_u32mf2((vuint32mf2_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vadd_vv_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vv_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vadd_vv_u64m1(op0, op1, op2) \ +__builtin_rvv_vadd_vv_u64m1((vuint64m1_t)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vadd_vv_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vv_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vadd_vv_u64m2(op0, op1, op2) \ +__builtin_rvv_vadd_vv_u64m2((vuint64m2_t)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vadd_vv_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vv_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vadd_vv_u64m4(op0, op1, op2) \ +__builtin_rvv_vadd_vv_u64m4((vuint64m4_t)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vadd_vv_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vv_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vadd_vv_u64m8(op0, op1, op2) \ +__builtin_rvv_vadd_vv_u64m8((vuint64m8_t)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vadd_vv_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vv_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vadd_vx_u8m1(op0, op1, op2) \ +__builtin_rvv_vadd_vx_u8m1((vuint8m1_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vadd_vx_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vx_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vadd_vx_u8m2(op0, op1, op2) \ +__builtin_rvv_vadd_vx_u8m2((vuint8m2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vadd_vx_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vx_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vadd_vx_u8m4(op0, op1, op2) \ +__builtin_rvv_vadd_vx_u8m4((vuint8m4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vadd_vx_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vx_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vadd_vx_u8m8(op0, op1, op2) \ +__builtin_rvv_vadd_vx_u8m8((vuint8m8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vadd_vx_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vx_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vadd_vx_u8mf2(op0, op1, op2) \ +__builtin_rvv_vadd_vx_u8mf2((vuint8mf2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vadd_vx_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vx_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vadd_vx_u8mf4(op0, op1, op2) \ +__builtin_rvv_vadd_vx_u8mf4((vuint8mf4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vadd_vx_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vx_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vadd_vx_u8mf8(op0, op1, op2) \ +__builtin_rvv_vadd_vx_u8mf8((vuint8mf8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vadd_vx_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vx_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vadd_vx_u16m1(op0, op1, op2) \ +__builtin_rvv_vadd_vx_u16m1((vuint16m1_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vadd_vx_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vx_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vadd_vx_u16m2(op0, op1, op2) \ +__builtin_rvv_vadd_vx_u16m2((vuint16m2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vadd_vx_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vx_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vadd_vx_u16m4(op0, op1, op2) \ +__builtin_rvv_vadd_vx_u16m4((vuint16m4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vadd_vx_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vx_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vadd_vx_u16m8(op0, op1, op2) \ +__builtin_rvv_vadd_vx_u16m8((vuint16m8_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vadd_vx_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vx_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vadd_vx_u16mf2(op0, op1, op2) \ +__builtin_rvv_vadd_vx_u16mf2((vuint16mf2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vadd_vx_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vx_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vadd_vx_u16mf4(op0, op1, op2) \ +__builtin_rvv_vadd_vx_u16mf4((vuint16mf4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vadd_vx_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vx_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vadd_vx_u32m1(op0, op1, op2) \ +__builtin_rvv_vadd_vx_u32m1((vuint32m1_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vadd_vx_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vx_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vadd_vx_u32m2(op0, op1, op2) \ +__builtin_rvv_vadd_vx_u32m2((vuint32m2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vadd_vx_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vx_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vadd_vx_u32m4(op0, op1, op2) \ +__builtin_rvv_vadd_vx_u32m4((vuint32m4_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vadd_vx_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vx_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vadd_vx_u32m8(op0, op1, op2) \ +__builtin_rvv_vadd_vx_u32m8((vuint32m8_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vadd_vx_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vx_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vadd_vx_u32mf2(op0, op1, op2) \ +__builtin_rvv_vadd_vx_u32mf2((vuint32mf2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vadd_vx_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vx_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vadd_vx_u64m1(op0, op1, op2) \ +__builtin_rvv_vadd_vx_u64m1((vuint64m1_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vadd_vx_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vx_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vadd_vx_u64m2(op0, op1, op2) \ +__builtin_rvv_vadd_vx_u64m2((vuint64m2_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vadd_vx_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vx_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vadd_vx_u64m4(op0, op1, op2) \ +__builtin_rvv_vadd_vx_u64m4((vuint64m4_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vadd_vx_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vx_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vadd_vx_u64m8(op0, op1, op2) \ +__builtin_rvv_vadd_vx_u64m8((vuint64m8_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vadd_vx_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vadd_vx_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vsub_vv_i8m1(op0, op1, op2) \ +__builtin_rvv_vsub_vv_i8m1((vint8m1_t)(op0), (vint8m1_t)(op1), (size_t)(op2)) +#define vsub_vv_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vv_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vsub_vv_i8m2(op0, op1, op2) \ +__builtin_rvv_vsub_vv_i8m2((vint8m2_t)(op0), (vint8m2_t)(op1), (size_t)(op2)) +#define vsub_vv_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vv_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vsub_vv_i8m4(op0, op1, op2) \ +__builtin_rvv_vsub_vv_i8m4((vint8m4_t)(op0), (vint8m4_t)(op1), (size_t)(op2)) +#define vsub_vv_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vv_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vsub_vv_i8m8(op0, op1, op2) \ +__builtin_rvv_vsub_vv_i8m8((vint8m8_t)(op0), (vint8m8_t)(op1), (size_t)(op2)) +#define vsub_vv_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vv_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (vint8m8_t)(op3), (size_t)(op4)) +#define vsub_vv_i8mf2(op0, op1, op2) \ +__builtin_rvv_vsub_vv_i8mf2((vint8mf2_t)(op0), (vint8mf2_t)(op1), (size_t)(op2)) +#define vsub_vv_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vv_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vsub_vv_i8mf4(op0, op1, op2) \ +__builtin_rvv_vsub_vv_i8mf4((vint8mf4_t)(op0), (vint8mf4_t)(op1), (size_t)(op2)) +#define vsub_vv_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vv_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vsub_vv_i8mf8(op0, op1, op2) \ +__builtin_rvv_vsub_vv_i8mf8((vint8mf8_t)(op0), (vint8mf8_t)(op1), (size_t)(op2)) +#define vsub_vv_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vv_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vsub_vv_i16m1(op0, op1, op2) \ +__builtin_rvv_vsub_vv_i16m1((vint16m1_t)(op0), (vint16m1_t)(op1), (size_t)(op2)) +#define vsub_vv_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vv_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vsub_vv_i16m2(op0, op1, op2) \ +__builtin_rvv_vsub_vv_i16m2((vint16m2_t)(op0), (vint16m2_t)(op1), (size_t)(op2)) +#define vsub_vv_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vv_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vsub_vv_i16m4(op0, op1, op2) \ +__builtin_rvv_vsub_vv_i16m4((vint16m4_t)(op0), (vint16m4_t)(op1), (size_t)(op2)) +#define vsub_vv_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vv_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vsub_vv_i16m8(op0, op1, op2) \ +__builtin_rvv_vsub_vv_i16m8((vint16m8_t)(op0), (vint16m8_t)(op1), (size_t)(op2)) +#define vsub_vv_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vv_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (vint16m8_t)(op3), (size_t)(op4)) +#define vsub_vv_i16mf2(op0, op1, op2) \ +__builtin_rvv_vsub_vv_i16mf2((vint16mf2_t)(op0), (vint16mf2_t)(op1), (size_t)(op2)) +#define vsub_vv_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vv_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vsub_vv_i16mf4(op0, op1, op2) \ +__builtin_rvv_vsub_vv_i16mf4((vint16mf4_t)(op0), (vint16mf4_t)(op1), (size_t)(op2)) +#define vsub_vv_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vv_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vsub_vv_i32m1(op0, op1, op2) \ +__builtin_rvv_vsub_vv_i32m1((vint32m1_t)(op0), (vint32m1_t)(op1), (size_t)(op2)) +#define vsub_vv_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vv_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vsub_vv_i32m2(op0, op1, op2) \ +__builtin_rvv_vsub_vv_i32m2((vint32m2_t)(op0), (vint32m2_t)(op1), (size_t)(op2)) +#define vsub_vv_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vv_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vsub_vv_i32m4(op0, op1, op2) \ +__builtin_rvv_vsub_vv_i32m4((vint32m4_t)(op0), (vint32m4_t)(op1), (size_t)(op2)) +#define vsub_vv_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vv_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vsub_vv_i32m8(op0, op1, op2) \ +__builtin_rvv_vsub_vv_i32m8((vint32m8_t)(op0), (vint32m8_t)(op1), (size_t)(op2)) +#define vsub_vv_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vv_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vsub_vv_i32mf2(op0, op1, op2) \ +__builtin_rvv_vsub_vv_i32mf2((vint32mf2_t)(op0), (vint32mf2_t)(op1), (size_t)(op2)) +#define vsub_vv_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vv_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vsub_vv_i64m1(op0, op1, op2) \ +__builtin_rvv_vsub_vv_i64m1((vint64m1_t)(op0), (vint64m1_t)(op1), (size_t)(op2)) +#define vsub_vv_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vv_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vsub_vv_i64m2(op0, op1, op2) \ +__builtin_rvv_vsub_vv_i64m2((vint64m2_t)(op0), (vint64m2_t)(op1), (size_t)(op2)) +#define vsub_vv_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vv_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vsub_vv_i64m4(op0, op1, op2) \ +__builtin_rvv_vsub_vv_i64m4((vint64m4_t)(op0), (vint64m4_t)(op1), (size_t)(op2)) +#define vsub_vv_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vv_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vsub_vv_i64m8(op0, op1, op2) \ +__builtin_rvv_vsub_vv_i64m8((vint64m8_t)(op0), (vint64m8_t)(op1), (size_t)(op2)) +#define vsub_vv_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vv_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vsuxei32_v_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_u8m1((uint8_t *)(op0), (vuint32m4_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vsuxei32_v_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint32m4_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vsuxei32_v_u8m2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_u8m2((uint8_t *)(op0), (vuint32m8_t)(op1), (vuint8m2_t)(op2), (size_t)(op3)) +#define vsuxei32_v_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_u8m2_m((vbool4_t)(op0), (uint8_t *)(op1), (vuint32m8_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vsuxei32_v_u8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_u8mf2((uint8_t *)(op0), (vuint32m2_t)(op1), (vuint8mf2_t)(op2), (size_t)(op3)) +#define vsuxei32_v_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint32m2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vsuxei32_v_u8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_u8mf4((uint8_t *)(op0), (vuint32m1_t)(op1), (vuint8mf4_t)(op2), (size_t)(op3)) +#define vsuxei32_v_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint32m1_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vsuxei32_v_u8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_u8mf8((uint8_t *)(op0), (vuint32mf2_t)(op1), (vuint8mf8_t)(op2), (size_t)(op3)) +#define vsuxei32_v_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint32mf2_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vsub_vx_i8m1(op0, op1, op2) \ +__builtin_rvv_vsub_vx_i8m1((vint8m1_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vsub_vx_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vx_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vsub_vx_i8m2(op0, op1, op2) \ +__builtin_rvv_vsub_vx_i8m2((vint8m2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vsub_vx_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vx_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vsub_vx_i8m4(op0, op1, op2) \ +__builtin_rvv_vsub_vx_i8m4((vint8m4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vsub_vx_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vx_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vsub_vx_i8m8(op0, op1, op2) \ +__builtin_rvv_vsub_vx_i8m8((vint8m8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vsub_vx_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vx_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vsub_vx_i8mf2(op0, op1, op2) \ +__builtin_rvv_vsub_vx_i8mf2((vint8mf2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vsub_vx_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vx_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vsub_vx_i8mf4(op0, op1, op2) \ +__builtin_rvv_vsub_vx_i8mf4((vint8mf4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vsub_vx_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vx_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vsub_vx_i8mf8(op0, op1, op2) \ +__builtin_rvv_vsub_vx_i8mf8((vint8mf8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vsub_vx_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vx_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vsub_vx_i16m1(op0, op1, op2) \ +__builtin_rvv_vsub_vx_i16m1((vint16m1_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vsub_vx_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vx_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vsub_vx_i16m2(op0, op1, op2) \ +__builtin_rvv_vsub_vx_i16m2((vint16m2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vsub_vx_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vx_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vsub_vx_i16m4(op0, op1, op2) \ +__builtin_rvv_vsub_vx_i16m4((vint16m4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vsub_vx_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vx_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vsub_vx_i16m8(op0, op1, op2) \ +__builtin_rvv_vsub_vx_i16m8((vint16m8_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vsub_vx_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vx_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vsub_vx_i16mf2(op0, op1, op2) \ +__builtin_rvv_vsub_vx_i16mf2((vint16mf2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vsub_vx_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vx_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vsub_vx_i16mf4(op0, op1, op2) \ +__builtin_rvv_vsub_vx_i16mf4((vint16mf4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vsub_vx_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vx_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vsub_vx_i32m1(op0, op1, op2) \ +__builtin_rvv_vsub_vx_i32m1((vint32m1_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vsub_vx_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vx_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vsub_vx_i32m2(op0, op1, op2) \ +__builtin_rvv_vsub_vx_i32m2((vint32m2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vsub_vx_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vx_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vsub_vx_i32m4(op0, op1, op2) \ +__builtin_rvv_vsub_vx_i32m4((vint32m4_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vsub_vx_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vx_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vsub_vx_i32m8(op0, op1, op2) \ +__builtin_rvv_vsub_vx_i32m8((vint32m8_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vsub_vx_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vx_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vsub_vx_i32mf2(op0, op1, op2) \ +__builtin_rvv_vsub_vx_i32mf2((vint32mf2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vsub_vx_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vx_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vsub_vx_i64m1(op0, op1, op2) \ +__builtin_rvv_vsub_vx_i64m1((vint64m1_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vsub_vx_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vx_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vsub_vx_i64m2(op0, op1, op2) \ +__builtin_rvv_vsub_vx_i64m2((vint64m2_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vsub_vx_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vx_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vsub_vx_i64m4(op0, op1, op2) \ +__builtin_rvv_vsub_vx_i64m4((vint64m4_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vsub_vx_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vx_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vsub_vx_i64m8(op0, op1, op2) \ +__builtin_rvv_vsub_vx_i64m8((vint64m8_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vsub_vx_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vx_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vsub_vv_u8m1(op0, op1, op2) \ +__builtin_rvv_vsub_vv_u8m1((vuint8m1_t)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vsub_vv_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vv_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vsub_vv_u8m2(op0, op1, op2) \ +__builtin_rvv_vsub_vv_u8m2((vuint8m2_t)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vsub_vv_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vv_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vsub_vv_u8m4(op0, op1, op2) \ +__builtin_rvv_vsub_vv_u8m4((vuint8m4_t)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vsub_vv_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vv_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vsub_vv_u8m8(op0, op1, op2) \ +__builtin_rvv_vsub_vv_u8m8((vuint8m8_t)(op0), (vuint8m8_t)(op1), (size_t)(op2)) +#define vsub_vv_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vv_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (vuint8m8_t)(op3), (size_t)(op4)) +#define vsub_vv_u8mf2(op0, op1, op2) \ +__builtin_rvv_vsub_vv_u8mf2((vuint8mf2_t)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vsub_vv_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vv_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vsub_vv_u8mf4(op0, op1, op2) \ +__builtin_rvv_vsub_vv_u8mf4((vuint8mf4_t)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vsub_vv_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vv_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vsub_vv_u8mf8(op0, op1, op2) \ +__builtin_rvv_vsub_vv_u8mf8((vuint8mf8_t)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vsub_vv_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vv_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vsub_vv_u16m1(op0, op1, op2) \ +__builtin_rvv_vsub_vv_u16m1((vuint16m1_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vsub_vv_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vv_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vsub_vv_u16m2(op0, op1, op2) \ +__builtin_rvv_vsub_vv_u16m2((vuint16m2_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vsub_vv_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vv_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vsub_vv_u16m4(op0, op1, op2) \ +__builtin_rvv_vsub_vv_u16m4((vuint16m4_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vsub_vv_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vv_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vsub_vv_u16m8(op0, op1, op2) \ +__builtin_rvv_vsub_vv_u16m8((vuint16m8_t)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vsub_vv_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vv_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vsub_vv_u16mf2(op0, op1, op2) \ +__builtin_rvv_vsub_vv_u16mf2((vuint16mf2_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vsub_vv_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vv_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vsub_vv_u16mf4(op0, op1, op2) \ +__builtin_rvv_vsub_vv_u16mf4((vuint16mf4_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vsub_vv_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vv_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vsub_vv_u32m1(op0, op1, op2) \ +__builtin_rvv_vsub_vv_u32m1((vuint32m1_t)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vsub_vv_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vv_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vsub_vv_u32m2(op0, op1, op2) \ +__builtin_rvv_vsub_vv_u32m2((vuint32m2_t)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vsub_vv_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vv_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vsub_vv_u32m4(op0, op1, op2) \ +__builtin_rvv_vsub_vv_u32m4((vuint32m4_t)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vsub_vv_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vv_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vsub_vv_u32m8(op0, op1, op2) \ +__builtin_rvv_vsub_vv_u32m8((vuint32m8_t)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vsub_vv_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vv_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vsub_vv_u32mf2(op0, op1, op2) \ +__builtin_rvv_vsub_vv_u32mf2((vuint32mf2_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vsub_vv_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vv_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vsub_vv_u64m1(op0, op1, op2) \ +__builtin_rvv_vsub_vv_u64m1((vuint64m1_t)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vsub_vv_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vv_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vsub_vv_u64m2(op0, op1, op2) \ +__builtin_rvv_vsub_vv_u64m2((vuint64m2_t)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vsub_vv_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vv_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vsub_vv_u64m4(op0, op1, op2) \ +__builtin_rvv_vsub_vv_u64m4((vuint64m4_t)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vsub_vv_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vv_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vsub_vv_u64m8(op0, op1, op2) \ +__builtin_rvv_vsub_vv_u64m8((vuint64m8_t)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vsub_vv_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vv_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vsub_vx_u8m1(op0, op1, op2) \ +__builtin_rvv_vsub_vx_u8m1((vuint8m1_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vsub_vx_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vx_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vsub_vx_u8m2(op0, op1, op2) \ +__builtin_rvv_vsub_vx_u8m2((vuint8m2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vsub_vx_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vx_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vsub_vx_u8m4(op0, op1, op2) \ +__builtin_rvv_vsub_vx_u8m4((vuint8m4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vsub_vx_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vx_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vsub_vx_u8m8(op0, op1, op2) \ +__builtin_rvv_vsub_vx_u8m8((vuint8m8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vsub_vx_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vx_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vsub_vx_u8mf2(op0, op1, op2) \ +__builtin_rvv_vsub_vx_u8mf2((vuint8mf2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vsub_vx_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vx_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vsub_vx_u8mf4(op0, op1, op2) \ +__builtin_rvv_vsub_vx_u8mf4((vuint8mf4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vsub_vx_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vx_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vsub_vx_u8mf8(op0, op1, op2) \ +__builtin_rvv_vsub_vx_u8mf8((vuint8mf8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vsub_vx_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vx_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vsub_vx_u16m1(op0, op1, op2) \ +__builtin_rvv_vsub_vx_u16m1((vuint16m1_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vsub_vx_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vx_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vsub_vx_u16m2(op0, op1, op2) \ +__builtin_rvv_vsub_vx_u16m2((vuint16m2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vsub_vx_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vx_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vsub_vx_u16m4(op0, op1, op2) \ +__builtin_rvv_vsub_vx_u16m4((vuint16m4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vsub_vx_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vx_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vsub_vx_u16m8(op0, op1, op2) \ +__builtin_rvv_vsub_vx_u16m8((vuint16m8_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vsub_vx_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vx_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vsub_vx_u16mf2(op0, op1, op2) \ +__builtin_rvv_vsub_vx_u16mf2((vuint16mf2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vsub_vx_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vx_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vsub_vx_u16mf4(op0, op1, op2) \ +__builtin_rvv_vsub_vx_u16mf4((vuint16mf4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vsub_vx_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vx_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vsub_vx_u32m1(op0, op1, op2) \ +__builtin_rvv_vsub_vx_u32m1((vuint32m1_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vsub_vx_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vx_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vsub_vx_u32m2(op0, op1, op2) \ +__builtin_rvv_vsub_vx_u32m2((vuint32m2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vsub_vx_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vx_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vsub_vx_u32m4(op0, op1, op2) \ +__builtin_rvv_vsub_vx_u32m4((vuint32m4_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vsub_vx_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vx_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vsub_vx_u32m8(op0, op1, op2) \ +__builtin_rvv_vsub_vx_u32m8((vuint32m8_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vsub_vx_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vx_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vsub_vx_u32mf2(op0, op1, op2) \ +__builtin_rvv_vsub_vx_u32mf2((vuint32mf2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vsub_vx_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vx_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vsub_vx_u64m1(op0, op1, op2) \ +__builtin_rvv_vsub_vx_u64m1((vuint64m1_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vsub_vx_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vx_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vsub_vx_u64m2(op0, op1, op2) \ +__builtin_rvv_vsub_vx_u64m2((vuint64m2_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vsub_vx_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vx_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vsub_vx_u64m4(op0, op1, op2) \ +__builtin_rvv_vsub_vx_u64m4((vuint64m4_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vsub_vx_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vx_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vsub_vx_u64m8(op0, op1, op2) \ +__builtin_rvv_vsub_vx_u64m8((vuint64m8_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vsub_vx_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsub_vx_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vrsub_vx_i8m1(op0, op1, op2) \ +__builtin_rvv_vrsub_vx_i8m1((vint8m1_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vrsub_vx_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrsub_vx_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vrsub_vx_i8m2(op0, op1, op2) \ +__builtin_rvv_vrsub_vx_i8m2((vint8m2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vrsub_vx_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrsub_vx_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vrsub_vx_i8m4(op0, op1, op2) \ +__builtin_rvv_vrsub_vx_i8m4((vint8m4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vrsub_vx_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrsub_vx_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vrsub_vx_i8m8(op0, op1, op2) \ +__builtin_rvv_vrsub_vx_i8m8((vint8m8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vrsub_vx_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrsub_vx_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vrsub_vx_i8mf2(op0, op1, op2) \ +__builtin_rvv_vrsub_vx_i8mf2((vint8mf2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vrsub_vx_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrsub_vx_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vrsub_vx_i8mf4(op0, op1, op2) \ +__builtin_rvv_vrsub_vx_i8mf4((vint8mf4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vrsub_vx_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrsub_vx_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vrsub_vx_i8mf8(op0, op1, op2) \ +__builtin_rvv_vrsub_vx_i8mf8((vint8mf8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vrsub_vx_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrsub_vx_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vrsub_vx_i16m1(op0, op1, op2) \ +__builtin_rvv_vrsub_vx_i16m1((vint16m1_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vrsub_vx_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrsub_vx_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vrsub_vx_i16m2(op0, op1, op2) \ +__builtin_rvv_vrsub_vx_i16m2((vint16m2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vrsub_vx_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrsub_vx_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vrsub_vx_i16m4(op0, op1, op2) \ +__builtin_rvv_vrsub_vx_i16m4((vint16m4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vrsub_vx_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrsub_vx_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vrsub_vx_i16m8(op0, op1, op2) \ +__builtin_rvv_vrsub_vx_i16m8((vint16m8_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vrsub_vx_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrsub_vx_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vrsub_vx_i16mf2(op0, op1, op2) \ +__builtin_rvv_vrsub_vx_i16mf2((vint16mf2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vrsub_vx_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrsub_vx_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vrsub_vx_i16mf4(op0, op1, op2) \ +__builtin_rvv_vrsub_vx_i16mf4((vint16mf4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vrsub_vx_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrsub_vx_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vrsub_vx_i32m1(op0, op1, op2) \ +__builtin_rvv_vrsub_vx_i32m1((vint32m1_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vrsub_vx_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrsub_vx_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vrsub_vx_i32m2(op0, op1, op2) \ +__builtin_rvv_vrsub_vx_i32m2((vint32m2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vrsub_vx_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrsub_vx_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vrsub_vx_i32m4(op0, op1, op2) \ +__builtin_rvv_vrsub_vx_i32m4((vint32m4_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vrsub_vx_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrsub_vx_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vrsub_vx_i32m8(op0, op1, op2) \ +__builtin_rvv_vrsub_vx_i32m8((vint32m8_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vrsub_vx_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrsub_vx_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vrsub_vx_i32mf2(op0, op1, op2) \ +__builtin_rvv_vrsub_vx_i32mf2((vint32mf2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vrsub_vx_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrsub_vx_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vrsub_vx_i64m1(op0, op1, op2) \ +__builtin_rvv_vrsub_vx_i64m1((vint64m1_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vrsub_vx_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrsub_vx_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vrsub_vx_i64m2(op0, op1, op2) \ +__builtin_rvv_vrsub_vx_i64m2((vint64m2_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vrsub_vx_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrsub_vx_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vrsub_vx_i64m4(op0, op1, op2) \ +__builtin_rvv_vrsub_vx_i64m4((vint64m4_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vrsub_vx_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrsub_vx_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vrsub_vx_i64m8(op0, op1, op2) \ +__builtin_rvv_vrsub_vx_i64m8((vint64m8_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vrsub_vx_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrsub_vx_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vrsub_vx_u8m1(op0, op1, op2) \ +__builtin_rvv_vrsub_vx_u8m1((vuint8m1_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vrsub_vx_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrsub_vx_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vrsub_vx_u8m2(op0, op1, op2) \ +__builtin_rvv_vrsub_vx_u8m2((vuint8m2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vrsub_vx_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrsub_vx_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vrsub_vx_u8m4(op0, op1, op2) \ +__builtin_rvv_vrsub_vx_u8m4((vuint8m4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vrsub_vx_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrsub_vx_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vrsub_vx_u8m8(op0, op1, op2) \ +__builtin_rvv_vrsub_vx_u8m8((vuint8m8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vrsub_vx_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrsub_vx_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vrsub_vx_u8mf2(op0, op1, op2) \ +__builtin_rvv_vrsub_vx_u8mf2((vuint8mf2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vrsub_vx_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrsub_vx_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vrsub_vx_u8mf4(op0, op1, op2) \ +__builtin_rvv_vrsub_vx_u8mf4((vuint8mf4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vrsub_vx_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrsub_vx_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vrsub_vx_u8mf8(op0, op1, op2) \ +__builtin_rvv_vrsub_vx_u8mf8((vuint8mf8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vrsub_vx_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrsub_vx_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vrsub_vx_u16m1(op0, op1, op2) \ +__builtin_rvv_vrsub_vx_u16m1((vuint16m1_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vrsub_vx_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrsub_vx_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vrsub_vx_u16m2(op0, op1, op2) \ +__builtin_rvv_vrsub_vx_u16m2((vuint16m2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vrsub_vx_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrsub_vx_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vrsub_vx_u16m4(op0, op1, op2) \ +__builtin_rvv_vrsub_vx_u16m4((vuint16m4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vrsub_vx_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrsub_vx_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vrsub_vx_u16m8(op0, op1, op2) \ +__builtin_rvv_vrsub_vx_u16m8((vuint16m8_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vrsub_vx_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrsub_vx_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vrsub_vx_u16mf2(op0, op1, op2) \ +__builtin_rvv_vrsub_vx_u16mf2((vuint16mf2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vrsub_vx_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrsub_vx_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vrsub_vx_u16mf4(op0, op1, op2) \ +__builtin_rvv_vrsub_vx_u16mf4((vuint16mf4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vrsub_vx_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrsub_vx_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vrsub_vx_u32m1(op0, op1, op2) \ +__builtin_rvv_vrsub_vx_u32m1((vuint32m1_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vrsub_vx_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrsub_vx_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vrsub_vx_u32m2(op0, op1, op2) \ +__builtin_rvv_vrsub_vx_u32m2((vuint32m2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vrsub_vx_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrsub_vx_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vrsub_vx_u32m4(op0, op1, op2) \ +__builtin_rvv_vrsub_vx_u32m4((vuint32m4_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vrsub_vx_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrsub_vx_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vrsub_vx_u32m8(op0, op1, op2) \ +__builtin_rvv_vrsub_vx_u32m8((vuint32m8_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vrsub_vx_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrsub_vx_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vrsub_vx_u32mf2(op0, op1, op2) \ +__builtin_rvv_vrsub_vx_u32mf2((vuint32mf2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vrsub_vx_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrsub_vx_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vrsub_vx_u64m1(op0, op1, op2) \ +__builtin_rvv_vrsub_vx_u64m1((vuint64m1_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vrsub_vx_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrsub_vx_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vrsub_vx_u64m2(op0, op1, op2) \ +__builtin_rvv_vrsub_vx_u64m2((vuint64m2_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vrsub_vx_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrsub_vx_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vrsub_vx_u64m4(op0, op1, op2) \ +__builtin_rvv_vrsub_vx_u64m4((vuint64m4_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vrsub_vx_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrsub_vx_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vrsub_vx_u64m8(op0, op1, op2) \ +__builtin_rvv_vrsub_vx_u64m8((vuint64m8_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vrsub_vx_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrsub_vx_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vwaddu_vx_u16mf4(op0, op1, op2) \ +__builtin_rvv_vwaddu_vx_u16mf4((vuint8mf8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vwaddu_vx_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_vx_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint8mf8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vwaddu_vx_u16mf2(op0, op1, op2) \ +__builtin_rvv_vwaddu_vx_u16mf2((vuint8mf4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vwaddu_vx_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_vx_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint8mf4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vwaddu_vx_u16m1(op0, op1, op2) \ +__builtin_rvv_vwaddu_vx_u16m1((vuint8mf2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vwaddu_vx_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_vx_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint8mf2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vwaddu_vx_u16m2(op0, op1, op2) \ +__builtin_rvv_vwaddu_vx_u16m2((vuint8m1_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vwaddu_vx_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_vx_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint8m1_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vwaddu_vx_u16m4(op0, op1, op2) \ +__builtin_rvv_vwaddu_vx_u16m4((vuint8m2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vwaddu_vx_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_vx_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint8m2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vwaddu_vx_u16m8(op0, op1, op2) \ +__builtin_rvv_vwaddu_vx_u16m8((vuint8m4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vwaddu_vx_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_vx_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint8m4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vwaddu_vx_u32mf2(op0, op1, op2) \ +__builtin_rvv_vwaddu_vx_u32mf2((vuint16mf4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vwaddu_vx_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_vx_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint16mf4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vwaddu_vx_u32m1(op0, op1, op2) \ +__builtin_rvv_vwaddu_vx_u32m1((vuint16mf2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vwaddu_vx_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_vx_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint16mf2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vwaddu_vx_u32m2(op0, op1, op2) \ +__builtin_rvv_vwaddu_vx_u32m2((vuint16m1_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vwaddu_vx_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_vx_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint16m1_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vwaddu_vx_u32m4(op0, op1, op2) \ +__builtin_rvv_vwaddu_vx_u32m4((vuint16m2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vwaddu_vx_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_vx_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint16m2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vwaddu_vx_u32m8(op0, op1, op2) \ +__builtin_rvv_vwaddu_vx_u32m8((vuint16m4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vwaddu_vx_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_vx_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint16m4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vwaddu_vx_u64m1(op0, op1, op2) \ +__builtin_rvv_vwaddu_vx_u64m1((vuint32mf2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vwaddu_vx_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_vx_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint32mf2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vwaddu_vx_u64m2(op0, op1, op2) \ +__builtin_rvv_vwaddu_vx_u64m2((vuint32m1_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vwaddu_vx_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_vx_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint32m1_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vwaddu_vx_u64m4(op0, op1, op2) \ +__builtin_rvv_vwaddu_vx_u64m4((vuint32m2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vwaddu_vx_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_vx_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint32m2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vwaddu_vx_u64m8(op0, op1, op2) \ +__builtin_rvv_vwaddu_vx_u64m8((vuint32m4_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vwaddu_vx_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_vx_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint32m4_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vwsubu_vv_u16mf4(op0, op1, op2) \ +__builtin_rvv_vwsubu_vv_u16mf4((vuint8mf8_t)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vwsubu_vv_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_vv_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vwsubu_vv_u16mf2(op0, op1, op2) \ +__builtin_rvv_vwsubu_vv_u16mf2((vuint8mf4_t)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vwsubu_vv_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_vv_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vwsubu_vv_u16m1(op0, op1, op2) \ +__builtin_rvv_vwsubu_vv_u16m1((vuint8mf2_t)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vwsubu_vv_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_vv_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vwsubu_vv_u16m2(op0, op1, op2) \ +__builtin_rvv_vwsubu_vv_u16m2((vuint8m1_t)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vwsubu_vv_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_vv_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vwsubu_vv_u16m4(op0, op1, op2) \ +__builtin_rvv_vwsubu_vv_u16m4((vuint8m2_t)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vwsubu_vv_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_vv_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vwsubu_vv_u16m8(op0, op1, op2) \ +__builtin_rvv_vwsubu_vv_u16m8((vuint8m4_t)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vwsubu_vv_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_vv_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vwsubu_vv_u32mf2(op0, op1, op2) \ +__builtin_rvv_vwsubu_vv_u32mf2((vuint16mf4_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vwsubu_vv_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_vv_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vwsubu_vv_u32m1(op0, op1, op2) \ +__builtin_rvv_vwsubu_vv_u32m1((vuint16mf2_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vwsubu_vv_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_vv_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vwsubu_vv_u32m2(op0, op1, op2) \ +__builtin_rvv_vwsubu_vv_u32m2((vuint16m1_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vwsubu_vv_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_vv_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vwsubu_vv_u32m4(op0, op1, op2) \ +__builtin_rvv_vwsubu_vv_u32m4((vuint16m2_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vwsubu_vv_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_vv_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vwsubu_vv_u32m8(op0, op1, op2) \ +__builtin_rvv_vwsubu_vv_u32m8((vuint16m4_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vwsubu_vv_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_vv_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vwsubu_vv_u64m1(op0, op1, op2) \ +__builtin_rvv_vwsubu_vv_u64m1((vuint32mf2_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vwsubu_vv_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_vv_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vwsubu_vv_u64m2(op0, op1, op2) \ +__builtin_rvv_vwsubu_vv_u64m2((vuint32m1_t)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vwsubu_vv_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_vv_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vwsubu_vv_u64m4(op0, op1, op2) \ +__builtin_rvv_vwsubu_vv_u64m4((vuint32m2_t)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vwsubu_vv_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_vv_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vwsubu_vv_u64m8(op0, op1, op2) \ +__builtin_rvv_vwsubu_vv_u64m8((vuint32m4_t)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vwsubu_vv_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_vv_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vwsubu_vx_u16mf4(op0, op1, op2) \ +__builtin_rvv_vwsubu_vx_u16mf4((vuint8mf8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vwsubu_vx_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_vx_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint8mf8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vwsubu_vx_u16mf2(op0, op1, op2) \ +__builtin_rvv_vwsubu_vx_u16mf2((vuint8mf4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vwsubu_vx_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_vx_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint8mf4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vwsubu_vx_u16m1(op0, op1, op2) \ +__builtin_rvv_vwsubu_vx_u16m1((vuint8mf2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vwsubu_vx_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_vx_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint8mf2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vwsubu_vx_u16m2(op0, op1, op2) \ +__builtin_rvv_vwsubu_vx_u16m2((vuint8m1_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vwsubu_vx_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_vx_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint8m1_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vwsubu_vx_u16m4(op0, op1, op2) \ +__builtin_rvv_vwsubu_vx_u16m4((vuint8m2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vwsubu_vx_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_vx_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint8m2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vwsubu_vx_u16m8(op0, op1, op2) \ +__builtin_rvv_vwsubu_vx_u16m8((vuint8m4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vwsubu_vx_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_vx_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint8m4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vwsubu_vx_u32mf2(op0, op1, op2) \ +__builtin_rvv_vwsubu_vx_u32mf2((vuint16mf4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vwsubu_vx_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_vx_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint16mf4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vwsubu_vx_u32m1(op0, op1, op2) \ +__builtin_rvv_vwsubu_vx_u32m1((vuint16mf2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vwsubu_vx_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_vx_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint16mf2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vwsubu_vx_u32m2(op0, op1, op2) \ +__builtin_rvv_vwsubu_vx_u32m2((vuint16m1_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vwsubu_vx_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_vx_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint16m1_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vwsubu_vx_u32m4(op0, op1, op2) \ +__builtin_rvv_vwsubu_vx_u32m4((vuint16m2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vwsubu_vx_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_vx_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint16m2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vwsubu_vx_u32m8(op0, op1, op2) \ +__builtin_rvv_vwsubu_vx_u32m8((vuint16m4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vwsubu_vx_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_vx_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint16m4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vwsubu_vx_u64m1(op0, op1, op2) \ +__builtin_rvv_vwsubu_vx_u64m1((vuint32mf2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vwsubu_vx_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_vx_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint32mf2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vwsubu_vx_u64m2(op0, op1, op2) \ +__builtin_rvv_vwsubu_vx_u64m2((vuint32m1_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vwsubu_vx_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_vx_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint32m1_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vwsubu_vx_u64m4(op0, op1, op2) \ +__builtin_rvv_vwsubu_vx_u64m4((vuint32m2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vwsubu_vx_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_vx_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint32m2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vwsubu_vx_u64m8(op0, op1, op2) \ +__builtin_rvv_vwsubu_vx_u64m8((vuint32m4_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vwsubu_vx_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_vx_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint32m4_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vwadd_vv_i16mf4(op0, op1, op2) \ +__builtin_rvv_vwadd_vv_i16mf4((vint8mf8_t)(op0), (vint8mf8_t)(op1), (size_t)(op2)) +#define vwadd_vv_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_vv_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vwadd_vv_i16mf2(op0, op1, op2) \ +__builtin_rvv_vwadd_vv_i16mf2((vint8mf4_t)(op0), (vint8mf4_t)(op1), (size_t)(op2)) +#define vwadd_vv_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_vv_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vwadd_vv_i16m1(op0, op1, op2) \ +__builtin_rvv_vwadd_vv_i16m1((vint8mf2_t)(op0), (vint8mf2_t)(op1), (size_t)(op2)) +#define vwadd_vv_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_vv_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vwadd_vv_i16m2(op0, op1, op2) \ +__builtin_rvv_vwadd_vv_i16m2((vint8m1_t)(op0), (vint8m1_t)(op1), (size_t)(op2)) +#define vwadd_vv_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_vv_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vwadd_vv_i16m4(op0, op1, op2) \ +__builtin_rvv_vwadd_vv_i16m4((vint8m2_t)(op0), (vint8m2_t)(op1), (size_t)(op2)) +#define vwadd_vv_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_vv_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vwadd_vv_i16m8(op0, op1, op2) \ +__builtin_rvv_vwadd_vv_i16m8((vint8m4_t)(op0), (vint8m4_t)(op1), (size_t)(op2)) +#define vwadd_vv_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_vv_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint8m4_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vwadd_vv_i32mf2(op0, op1, op2) \ +__builtin_rvv_vwadd_vv_i32mf2((vint16mf4_t)(op0), (vint16mf4_t)(op1), (size_t)(op2)) +#define vwadd_vv_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_vv_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vwadd_vv_i32m1(op0, op1, op2) \ +__builtin_rvv_vwadd_vv_i32m1((vint16mf2_t)(op0), (vint16mf2_t)(op1), (size_t)(op2)) +#define vwadd_vv_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_vv_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vwadd_vv_i32m2(op0, op1, op2) \ +__builtin_rvv_vwadd_vv_i32m2((vint16m1_t)(op0), (vint16m1_t)(op1), (size_t)(op2)) +#define vwadd_vv_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_vv_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vwadd_vv_i32m4(op0, op1, op2) \ +__builtin_rvv_vwadd_vv_i32m4((vint16m2_t)(op0), (vint16m2_t)(op1), (size_t)(op2)) +#define vwadd_vv_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_vv_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vwadd_vv_i32m8(op0, op1, op2) \ +__builtin_rvv_vwadd_vv_i32m8((vint16m4_t)(op0), (vint16m4_t)(op1), (size_t)(op2)) +#define vwadd_vv_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_vv_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint16m4_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vwadd_vv_i64m1(op0, op1, op2) \ +__builtin_rvv_vwadd_vv_i64m1((vint32mf2_t)(op0), (vint32mf2_t)(op1), (size_t)(op2)) +#define vwadd_vv_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_vv_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vwadd_vv_i64m2(op0, op1, op2) \ +__builtin_rvv_vwadd_vv_i64m2((vint32m1_t)(op0), (vint32m1_t)(op1), (size_t)(op2)) +#define vwadd_vv_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_vv_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vwadd_vv_i64m4(op0, op1, op2) \ +__builtin_rvv_vwadd_vv_i64m4((vint32m2_t)(op0), (vint32m2_t)(op1), (size_t)(op2)) +#define vwadd_vv_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_vv_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vwadd_vv_i64m8(op0, op1, op2) \ +__builtin_rvv_vwadd_vv_i64m8((vint32m4_t)(op0), (vint32m4_t)(op1), (size_t)(op2)) +#define vwadd_vv_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_vv_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vwadd_vx_i16mf4(op0, op1, op2) \ +__builtin_rvv_vwadd_vx_i16mf4((vint8mf8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vwadd_vx_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_vx_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint8mf8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vwadd_vx_i16mf2(op0, op1, op2) \ +__builtin_rvv_vwadd_vx_i16mf2((vint8mf4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vwadd_vx_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_vx_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint8mf4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vwadd_vx_i16m1(op0, op1, op2) \ +__builtin_rvv_vwadd_vx_i16m1((vint8mf2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vwadd_vx_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_vx_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint8mf2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vwadd_vx_i16m2(op0, op1, op2) \ +__builtin_rvv_vwadd_vx_i16m2((vint8m1_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vwadd_vx_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_vx_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint8m1_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vwadd_vx_i16m4(op0, op1, op2) \ +__builtin_rvv_vwadd_vx_i16m4((vint8m2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vwadd_vx_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_vx_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint8m2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vwadd_vx_i16m8(op0, op1, op2) \ +__builtin_rvv_vwadd_vx_i16m8((vint8m4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vwadd_vx_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_vx_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint8m4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vwadd_vx_i32mf2(op0, op1, op2) \ +__builtin_rvv_vwadd_vx_i32mf2((vint16mf4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vwadd_vx_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_vx_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint16mf4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vwadd_vx_i32m1(op0, op1, op2) \ +__builtin_rvv_vwadd_vx_i32m1((vint16mf2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vwadd_vx_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_vx_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint16mf2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vwadd_vx_i32m2(op0, op1, op2) \ +__builtin_rvv_vwadd_vx_i32m2((vint16m1_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vwadd_vx_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_vx_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint16m1_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vwadd_vx_i32m4(op0, op1, op2) \ +__builtin_rvv_vwadd_vx_i32m4((vint16m2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vwadd_vx_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_vx_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint16m2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vwadd_vx_i32m8(op0, op1, op2) \ +__builtin_rvv_vwadd_vx_i32m8((vint16m4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vwadd_vx_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_vx_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint16m4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vwadd_vx_i64m1(op0, op1, op2) \ +__builtin_rvv_vwadd_vx_i64m1((vint32mf2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vwadd_vx_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_vx_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint32mf2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vwadd_vx_i64m2(op0, op1, op2) \ +__builtin_rvv_vwadd_vx_i64m2((vint32m1_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vwadd_vx_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_vx_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint32m1_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vwadd_vx_i64m4(op0, op1, op2) \ +__builtin_rvv_vwadd_vx_i64m4((vint32m2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vwadd_vx_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_vx_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint32m2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vwadd_vx_i64m8(op0, op1, op2) \ +__builtin_rvv_vwadd_vx_i64m8((vint32m4_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vwadd_vx_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_vx_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint32m4_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vsuxei64_v_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei64_v_i8m1((int8_t *)(op0), (vuint64m8_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vsuxei64_v_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei64_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint64m8_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vsuxei64_v_i8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei64_v_i8mf2((int8_t *)(op0), (vuint64m4_t)(op1), (vint8mf2_t)(op2), (size_t)(op3)) +#define vsuxei64_v_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei64_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint64m4_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vsuxei64_v_i8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei64_v_i8mf4((int8_t *)(op0), (vuint64m2_t)(op1), (vint8mf4_t)(op2), (size_t)(op3)) +#define vsuxei64_v_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei64_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint64m2_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vsuxei64_v_i8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei64_v_i8mf8((int8_t *)(op0), (vuint64m1_t)(op1), (vint8mf8_t)(op2), (size_t)(op3)) +#define vsuxei64_v_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei64_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint64m1_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vwsub_vv_i16mf4(op0, op1, op2) \ +__builtin_rvv_vwsub_vv_i16mf4((vint8mf8_t)(op0), (vint8mf8_t)(op1), (size_t)(op2)) +#define vwsub_vv_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_vv_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vwsub_vv_i16mf2(op0, op1, op2) \ +__builtin_rvv_vwsub_vv_i16mf2((vint8mf4_t)(op0), (vint8mf4_t)(op1), (size_t)(op2)) +#define vwsub_vv_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_vv_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vwsub_vv_i16m1(op0, op1, op2) \ +__builtin_rvv_vwsub_vv_i16m1((vint8mf2_t)(op0), (vint8mf2_t)(op1), (size_t)(op2)) +#define vwsub_vv_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_vv_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vwsub_vv_i16m2(op0, op1, op2) \ +__builtin_rvv_vwsub_vv_i16m2((vint8m1_t)(op0), (vint8m1_t)(op1), (size_t)(op2)) +#define vwsub_vv_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_vv_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vwsub_vv_i16m4(op0, op1, op2) \ +__builtin_rvv_vwsub_vv_i16m4((vint8m2_t)(op0), (vint8m2_t)(op1), (size_t)(op2)) +#define vwsub_vv_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_vv_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vwsub_vv_i16m8(op0, op1, op2) \ +__builtin_rvv_vwsub_vv_i16m8((vint8m4_t)(op0), (vint8m4_t)(op1), (size_t)(op2)) +#define vwsub_vv_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_vv_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint8m4_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vwsub_vv_i32mf2(op0, op1, op2) \ +__builtin_rvv_vwsub_vv_i32mf2((vint16mf4_t)(op0), (vint16mf4_t)(op1), (size_t)(op2)) +#define vwsub_vv_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_vv_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vwsub_vv_i32m1(op0, op1, op2) \ +__builtin_rvv_vwsub_vv_i32m1((vint16mf2_t)(op0), (vint16mf2_t)(op1), (size_t)(op2)) +#define vwsub_vv_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_vv_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vwsub_vv_i32m2(op0, op1, op2) \ +__builtin_rvv_vwsub_vv_i32m2((vint16m1_t)(op0), (vint16m1_t)(op1), (size_t)(op2)) +#define vwsub_vv_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_vv_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vwsub_vv_i32m4(op0, op1, op2) \ +__builtin_rvv_vwsub_vv_i32m4((vint16m2_t)(op0), (vint16m2_t)(op1), (size_t)(op2)) +#define vwsub_vv_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_vv_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vwsub_vv_i32m8(op0, op1, op2) \ +__builtin_rvv_vwsub_vv_i32m8((vint16m4_t)(op0), (vint16m4_t)(op1), (size_t)(op2)) +#define vwsub_vv_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_vv_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint16m4_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vwsub_vv_i64m1(op0, op1, op2) \ +__builtin_rvv_vwsub_vv_i64m1((vint32mf2_t)(op0), (vint32mf2_t)(op1), (size_t)(op2)) +#define vwsub_vv_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_vv_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vwsub_vv_i64m2(op0, op1, op2) \ +__builtin_rvv_vwsub_vv_i64m2((vint32m1_t)(op0), (vint32m1_t)(op1), (size_t)(op2)) +#define vwsub_vv_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_vv_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vwsub_vv_i64m4(op0, op1, op2) \ +__builtin_rvv_vwsub_vv_i64m4((vint32m2_t)(op0), (vint32m2_t)(op1), (size_t)(op2)) +#define vwsub_vv_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_vv_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vwsub_vv_i64m8(op0, op1, op2) \ +__builtin_rvv_vwsub_vv_i64m8((vint32m4_t)(op0), (vint32m4_t)(op1), (size_t)(op2)) +#define vwsub_vv_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_vv_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vwsub_vx_i16mf4(op0, op1, op2) \ +__builtin_rvv_vwsub_vx_i16mf4((vint8mf8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vwsub_vx_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_vx_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint8mf8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vwsub_vx_i16mf2(op0, op1, op2) \ +__builtin_rvv_vwsub_vx_i16mf2((vint8mf4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vwsub_vx_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_vx_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint8mf4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vwsub_vx_i16m1(op0, op1, op2) \ +__builtin_rvv_vwsub_vx_i16m1((vint8mf2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vwsub_vx_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_vx_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint8mf2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vwsub_vx_i16m2(op0, op1, op2) \ +__builtin_rvv_vwsub_vx_i16m2((vint8m1_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vwsub_vx_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_vx_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint8m1_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vwsub_vx_i16m4(op0, op1, op2) \ +__builtin_rvv_vwsub_vx_i16m4((vint8m2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vwsub_vx_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_vx_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint8m2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vwsub_vx_i16m8(op0, op1, op2) \ +__builtin_rvv_vwsub_vx_i16m8((vint8m4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vwsub_vx_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_vx_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint8m4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vwsub_vx_i32mf2(op0, op1, op2) \ +__builtin_rvv_vwsub_vx_i32mf2((vint16mf4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vwsub_vx_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_vx_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint16mf4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vwsub_vx_i32m1(op0, op1, op2) \ +__builtin_rvv_vwsub_vx_i32m1((vint16mf2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vwsub_vx_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_vx_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint16mf2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vwsub_vx_i32m2(op0, op1, op2) \ +__builtin_rvv_vwsub_vx_i32m2((vint16m1_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vwsub_vx_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_vx_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint16m1_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vwsub_vx_i32m4(op0, op1, op2) \ +__builtin_rvv_vwsub_vx_i32m4((vint16m2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vwsub_vx_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_vx_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint16m2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vwsub_vx_i32m8(op0, op1, op2) \ +__builtin_rvv_vwsub_vx_i32m8((vint16m4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vwsub_vx_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_vx_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint16m4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vwsub_vx_i64m1(op0, op1, op2) \ +__builtin_rvv_vwsub_vx_i64m1((vint32mf2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vwsub_vx_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_vx_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint32mf2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vwsub_vx_i64m2(op0, op1, op2) \ +__builtin_rvv_vwsub_vx_i64m2((vint32m1_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vwsub_vx_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_vx_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint32m1_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vwsub_vx_i64m4(op0, op1, op2) \ +__builtin_rvv_vwsub_vx_i64m4((vint32m2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vwsub_vx_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_vx_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint32m2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vwsub_vx_i64m8(op0, op1, op2) \ +__builtin_rvv_vwsub_vx_i64m8((vint32m4_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vwsub_vx_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_vx_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint32m4_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vwaddu_wx_u16mf4(op0, op1, op2) \ +__builtin_rvv_vwaddu_wx_u16mf4((vuint16mf4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vwaddu_wx_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_wx_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vwaddu_wx_u16mf2(op0, op1, op2) \ +__builtin_rvv_vwaddu_wx_u16mf2((vuint16mf2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vwaddu_wx_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_wx_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vwaddu_wx_u16m1(op0, op1, op2) \ +__builtin_rvv_vwaddu_wx_u16m1((vuint16m1_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vwaddu_wx_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_wx_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vwaddu_wx_u16m2(op0, op1, op2) \ +__builtin_rvv_vwaddu_wx_u16m2((vuint16m2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vwaddu_wx_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_wx_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vwaddu_wx_u16m4(op0, op1, op2) \ +__builtin_rvv_vwaddu_wx_u16m4((vuint16m4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vwaddu_wx_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_wx_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vwaddu_wx_u16m8(op0, op1, op2) \ +__builtin_rvv_vwaddu_wx_u16m8((vuint16m8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vwaddu_wx_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_wx_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vwaddu_wx_u32mf2(op0, op1, op2) \ +__builtin_rvv_vwaddu_wx_u32mf2((vuint32mf2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vwaddu_wx_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_wx_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vwaddu_wx_u32m1(op0, op1, op2) \ +__builtin_rvv_vwaddu_wx_u32m1((vuint32m1_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vwaddu_wx_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_wx_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vwaddu_wx_u32m2(op0, op1, op2) \ +__builtin_rvv_vwaddu_wx_u32m2((vuint32m2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vwaddu_wx_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_wx_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vwaddu_wx_u32m4(op0, op1, op2) \ +__builtin_rvv_vwaddu_wx_u32m4((vuint32m4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vwaddu_wx_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_wx_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vwaddu_wx_u32m8(op0, op1, op2) \ +__builtin_rvv_vwaddu_wx_u32m8((vuint32m8_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vwaddu_wx_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_wx_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vwaddu_wx_u64m1(op0, op1, op2) \ +__builtin_rvv_vwaddu_wx_u64m1((vuint64m1_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vwaddu_wx_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_wx_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vwaddu_wx_u64m2(op0, op1, op2) \ +__builtin_rvv_vwaddu_wx_u64m2((vuint64m2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vwaddu_wx_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_wx_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vwaddu_wx_u64m4(op0, op1, op2) \ +__builtin_rvv_vwaddu_wx_u64m4((vuint64m4_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vwaddu_wx_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_wx_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vwaddu_wx_u64m8(op0, op1, op2) \ +__builtin_rvv_vwaddu_wx_u64m8((vuint64m8_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vwaddu_wx_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwaddu_wx_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vwsubu_wv_u16mf4(op0, op1, op2) \ +__builtin_rvv_vwsubu_wv_u16mf4((vuint16mf4_t)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vwsubu_wv_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_wv_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vwsubu_wv_u16mf2(op0, op1, op2) \ +__builtin_rvv_vwsubu_wv_u16mf2((vuint16mf2_t)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vwsubu_wv_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_wv_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vwsubu_wv_u16m1(op0, op1, op2) \ +__builtin_rvv_vwsubu_wv_u16m1((vuint16m1_t)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vwsubu_wv_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_wv_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vwsubu_wv_u16m2(op0, op1, op2) \ +__builtin_rvv_vwsubu_wv_u16m2((vuint16m2_t)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vwsubu_wv_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_wv_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vwsubu_wv_u16m4(op0, op1, op2) \ +__builtin_rvv_vwsubu_wv_u16m4((vuint16m4_t)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vwsubu_wv_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_wv_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vwsubu_wv_u16m8(op0, op1, op2) \ +__builtin_rvv_vwsubu_wv_u16m8((vuint16m8_t)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vwsubu_wv_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_wv_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vwsubu_wv_u32mf2(op0, op1, op2) \ +__builtin_rvv_vwsubu_wv_u32mf2((vuint32mf2_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vwsubu_wv_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_wv_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vwsubu_wv_u32m1(op0, op1, op2) \ +__builtin_rvv_vwsubu_wv_u32m1((vuint32m1_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vwsubu_wv_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_wv_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vwsubu_wv_u32m2(op0, op1, op2) \ +__builtin_rvv_vwsubu_wv_u32m2((vuint32m2_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vwsubu_wv_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_wv_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vwsubu_wv_u32m4(op0, op1, op2) \ +__builtin_rvv_vwsubu_wv_u32m4((vuint32m4_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vwsubu_wv_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_wv_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vwsubu_wv_u32m8(op0, op1, op2) \ +__builtin_rvv_vwsubu_wv_u32m8((vuint32m8_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vwsubu_wv_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_wv_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vwsubu_wv_u64m1(op0, op1, op2) \ +__builtin_rvv_vwsubu_wv_u64m1((vuint64m1_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vwsubu_wv_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_wv_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vwsubu_wv_u64m2(op0, op1, op2) \ +__builtin_rvv_vwsubu_wv_u64m2((vuint64m2_t)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vwsubu_wv_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_wv_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vwsubu_wv_u64m4(op0, op1, op2) \ +__builtin_rvv_vwsubu_wv_u64m4((vuint64m4_t)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vwsubu_wv_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_wv_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vwsubu_wv_u64m8(op0, op1, op2) \ +__builtin_rvv_vwsubu_wv_u64m8((vuint64m8_t)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vwsubu_wv_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_wv_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vwsubu_wx_u16mf4(op0, op1, op2) \ +__builtin_rvv_vwsubu_wx_u16mf4((vuint16mf4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vwsubu_wx_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_wx_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vwsubu_wx_u16mf2(op0, op1, op2) \ +__builtin_rvv_vwsubu_wx_u16mf2((vuint16mf2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vwsubu_wx_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_wx_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vwsubu_wx_u16m1(op0, op1, op2) \ +__builtin_rvv_vwsubu_wx_u16m1((vuint16m1_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vwsubu_wx_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_wx_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vwsubu_wx_u16m2(op0, op1, op2) \ +__builtin_rvv_vwsubu_wx_u16m2((vuint16m2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vwsubu_wx_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_wx_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vwsubu_wx_u16m4(op0, op1, op2) \ +__builtin_rvv_vwsubu_wx_u16m4((vuint16m4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vwsubu_wx_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_wx_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vwsubu_wx_u16m8(op0, op1, op2) \ +__builtin_rvv_vwsubu_wx_u16m8((vuint16m8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vwsubu_wx_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_wx_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vwsubu_wx_u32mf2(op0, op1, op2) \ +__builtin_rvv_vwsubu_wx_u32mf2((vuint32mf2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vwsubu_wx_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_wx_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vwsubu_wx_u32m1(op0, op1, op2) \ +__builtin_rvv_vwsubu_wx_u32m1((vuint32m1_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vwsubu_wx_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_wx_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vwsubu_wx_u32m2(op0, op1, op2) \ +__builtin_rvv_vwsubu_wx_u32m2((vuint32m2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vwsubu_wx_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_wx_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vwsubu_wx_u32m4(op0, op1, op2) \ +__builtin_rvv_vwsubu_wx_u32m4((vuint32m4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vwsubu_wx_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_wx_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vwsubu_wx_u32m8(op0, op1, op2) \ +__builtin_rvv_vwsubu_wx_u32m8((vuint32m8_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vwsubu_wx_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_wx_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vwsubu_wx_u64m1(op0, op1, op2) \ +__builtin_rvv_vwsubu_wx_u64m1((vuint64m1_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vwsubu_wx_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_wx_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vwsubu_wx_u64m2(op0, op1, op2) \ +__builtin_rvv_vwsubu_wx_u64m2((vuint64m2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vwsubu_wx_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_wx_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vwsubu_wx_u64m4(op0, op1, op2) \ +__builtin_rvv_vwsubu_wx_u64m4((vuint64m4_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vwsubu_wx_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_wx_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vwsubu_wx_u64m8(op0, op1, op2) \ +__builtin_rvv_vwsubu_wx_u64m8((vuint64m8_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vwsubu_wx_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsubu_wx_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vwadd_wv_i16mf4(op0, op1, op2) \ +__builtin_rvv_vwadd_wv_i16mf4((vint16mf4_t)(op0), (vint8mf8_t)(op1), (size_t)(op2)) +#define vwadd_wv_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_wv_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vwadd_wv_i16mf2(op0, op1, op2) \ +__builtin_rvv_vwadd_wv_i16mf2((vint16mf2_t)(op0), (vint8mf4_t)(op1), (size_t)(op2)) +#define vwadd_wv_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_wv_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vwadd_wv_i16m1(op0, op1, op2) \ +__builtin_rvv_vwadd_wv_i16m1((vint16m1_t)(op0), (vint8mf2_t)(op1), (size_t)(op2)) +#define vwadd_wv_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_wv_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vwadd_wv_i16m2(op0, op1, op2) \ +__builtin_rvv_vwadd_wv_i16m2((vint16m2_t)(op0), (vint8m1_t)(op1), (size_t)(op2)) +#define vwadd_wv_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_wv_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vwadd_wv_i16m4(op0, op1, op2) \ +__builtin_rvv_vwadd_wv_i16m4((vint16m4_t)(op0), (vint8m2_t)(op1), (size_t)(op2)) +#define vwadd_wv_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_wv_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vwadd_wv_i16m8(op0, op1, op2) \ +__builtin_rvv_vwadd_wv_i16m8((vint16m8_t)(op0), (vint8m4_t)(op1), (size_t)(op2)) +#define vwadd_wv_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_wv_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vwadd_wv_i32mf2(op0, op1, op2) \ +__builtin_rvv_vwadd_wv_i32mf2((vint32mf2_t)(op0), (vint16mf4_t)(op1), (size_t)(op2)) +#define vwadd_wv_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_wv_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vwadd_wv_i32m1(op0, op1, op2) \ +__builtin_rvv_vwadd_wv_i32m1((vint32m1_t)(op0), (vint16mf2_t)(op1), (size_t)(op2)) +#define vwadd_wv_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_wv_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vwadd_wv_i32m2(op0, op1, op2) \ +__builtin_rvv_vwadd_wv_i32m2((vint32m2_t)(op0), (vint16m1_t)(op1), (size_t)(op2)) +#define vwadd_wv_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_wv_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vwadd_wv_i32m4(op0, op1, op2) \ +__builtin_rvv_vwadd_wv_i32m4((vint32m4_t)(op0), (vint16m2_t)(op1), (size_t)(op2)) +#define vwadd_wv_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_wv_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vwadd_wv_i32m8(op0, op1, op2) \ +__builtin_rvv_vwadd_wv_i32m8((vint32m8_t)(op0), (vint16m4_t)(op1), (size_t)(op2)) +#define vwadd_wv_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_wv_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vwadd_wv_i64m1(op0, op1, op2) \ +__builtin_rvv_vwadd_wv_i64m1((vint64m1_t)(op0), (vint32mf2_t)(op1), (size_t)(op2)) +#define vwadd_wv_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_wv_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vwadd_wv_i64m2(op0, op1, op2) \ +__builtin_rvv_vwadd_wv_i64m2((vint64m2_t)(op0), (vint32m1_t)(op1), (size_t)(op2)) +#define vwadd_wv_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_wv_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vwadd_wv_i64m4(op0, op1, op2) \ +__builtin_rvv_vwadd_wv_i64m4((vint64m4_t)(op0), (vint32m2_t)(op1), (size_t)(op2)) +#define vwadd_wv_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_wv_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vwadd_wv_i64m8(op0, op1, op2) \ +__builtin_rvv_vwadd_wv_i64m8((vint64m8_t)(op0), (vint32m4_t)(op1), (size_t)(op2)) +#define vwadd_wv_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_wv_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vwadd_wx_i16mf4(op0, op1, op2) \ +__builtin_rvv_vwadd_wx_i16mf4((vint16mf4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vwadd_wx_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_wx_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vwadd_wx_i16mf2(op0, op1, op2) \ +__builtin_rvv_vwadd_wx_i16mf2((vint16mf2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vwadd_wx_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_wx_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vwadd_wx_i16m1(op0, op1, op2) \ +__builtin_rvv_vwadd_wx_i16m1((vint16m1_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vwadd_wx_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_wx_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vwadd_wx_i16m2(op0, op1, op2) \ +__builtin_rvv_vwadd_wx_i16m2((vint16m2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vwadd_wx_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_wx_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vwadd_wx_i16m4(op0, op1, op2) \ +__builtin_rvv_vwadd_wx_i16m4((vint16m4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vwadd_wx_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_wx_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vwadd_wx_i16m8(op0, op1, op2) \ +__builtin_rvv_vwadd_wx_i16m8((vint16m8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vwadd_wx_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_wx_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vwadd_wx_i32mf2(op0, op1, op2) \ +__builtin_rvv_vwadd_wx_i32mf2((vint32mf2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vwadd_wx_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_wx_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vwadd_wx_i32m1(op0, op1, op2) \ +__builtin_rvv_vwadd_wx_i32m1((vint32m1_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vwadd_wx_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_wx_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vwadd_wx_i32m2(op0, op1, op2) \ +__builtin_rvv_vwadd_wx_i32m2((vint32m2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vwadd_wx_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_wx_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vwadd_wx_i32m4(op0, op1, op2) \ +__builtin_rvv_vwadd_wx_i32m4((vint32m4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vwadd_wx_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_wx_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vwadd_wx_i32m8(op0, op1, op2) \ +__builtin_rvv_vwadd_wx_i32m8((vint32m8_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vwadd_wx_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_wx_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vwadd_wx_i64m1(op0, op1, op2) \ +__builtin_rvv_vwadd_wx_i64m1((vint64m1_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vwadd_wx_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_wx_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vwadd_wx_i64m2(op0, op1, op2) \ +__builtin_rvv_vwadd_wx_i64m2((vint64m2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vwadd_wx_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_wx_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vwadd_wx_i64m4(op0, op1, op2) \ +__builtin_rvv_vwadd_wx_i64m4((vint64m4_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vwadd_wx_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_wx_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vwadd_wx_i64m8(op0, op1, op2) \ +__builtin_rvv_vwadd_wx_i64m8((vint64m8_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vwadd_wx_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwadd_wx_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vwsub_wv_i16mf4(op0, op1, op2) \ +__builtin_rvv_vwsub_wv_i16mf4((vint16mf4_t)(op0), (vint8mf8_t)(op1), (size_t)(op2)) +#define vwsub_wv_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_wv_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vwsub_wv_i16mf2(op0, op1, op2) \ +__builtin_rvv_vwsub_wv_i16mf2((vint16mf2_t)(op0), (vint8mf4_t)(op1), (size_t)(op2)) +#define vwsub_wv_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_wv_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vwsub_wv_i16m1(op0, op1, op2) \ +__builtin_rvv_vwsub_wv_i16m1((vint16m1_t)(op0), (vint8mf2_t)(op1), (size_t)(op2)) +#define vwsub_wv_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_wv_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vwsub_wv_i16m2(op0, op1, op2) \ +__builtin_rvv_vwsub_wv_i16m2((vint16m2_t)(op0), (vint8m1_t)(op1), (size_t)(op2)) +#define vwsub_wv_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_wv_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vwsub_wv_i16m4(op0, op1, op2) \ +__builtin_rvv_vwsub_wv_i16m4((vint16m4_t)(op0), (vint8m2_t)(op1), (size_t)(op2)) +#define vwsub_wv_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_wv_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vwsub_wv_i16m8(op0, op1, op2) \ +__builtin_rvv_vwsub_wv_i16m8((vint16m8_t)(op0), (vint8m4_t)(op1), (size_t)(op2)) +#define vwsub_wv_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_wv_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vwsub_wv_i32mf2(op0, op1, op2) \ +__builtin_rvv_vwsub_wv_i32mf2((vint32mf2_t)(op0), (vint16mf4_t)(op1), (size_t)(op2)) +#define vwsub_wv_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_wv_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vwsub_wv_i32m1(op0, op1, op2) \ +__builtin_rvv_vwsub_wv_i32m1((vint32m1_t)(op0), (vint16mf2_t)(op1), (size_t)(op2)) +#define vwsub_wv_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_wv_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vwsub_wv_i32m2(op0, op1, op2) \ +__builtin_rvv_vwsub_wv_i32m2((vint32m2_t)(op0), (vint16m1_t)(op1), (size_t)(op2)) +#define vwsub_wv_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_wv_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vwsub_wv_i32m4(op0, op1, op2) \ +__builtin_rvv_vwsub_wv_i32m4((vint32m4_t)(op0), (vint16m2_t)(op1), (size_t)(op2)) +#define vwsub_wv_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_wv_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vwsub_wv_i32m8(op0, op1, op2) \ +__builtin_rvv_vwsub_wv_i32m8((vint32m8_t)(op0), (vint16m4_t)(op1), (size_t)(op2)) +#define vwsub_wv_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_wv_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vwsub_wv_i64m1(op0, op1, op2) \ +__builtin_rvv_vwsub_wv_i64m1((vint64m1_t)(op0), (vint32mf2_t)(op1), (size_t)(op2)) +#define vwsub_wv_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_wv_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vwsub_wv_i64m2(op0, op1, op2) \ +__builtin_rvv_vwsub_wv_i64m2((vint64m2_t)(op0), (vint32m1_t)(op1), (size_t)(op2)) +#define vwsub_wv_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_wv_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vwsub_wv_i64m4(op0, op1, op2) \ +__builtin_rvv_vwsub_wv_i64m4((vint64m4_t)(op0), (vint32m2_t)(op1), (size_t)(op2)) +#define vwsub_wv_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_wv_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vwsub_wv_i64m8(op0, op1, op2) \ +__builtin_rvv_vwsub_wv_i64m8((vint64m8_t)(op0), (vint32m4_t)(op1), (size_t)(op2)) +#define vwsub_wv_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_wv_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vwsub_wx_i16mf4(op0, op1, op2) \ +__builtin_rvv_vwsub_wx_i16mf4((vint16mf4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vwsub_wx_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_wx_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vwsub_wx_i16mf2(op0, op1, op2) \ +__builtin_rvv_vwsub_wx_i16mf2((vint16mf2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vwsub_wx_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_wx_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vwsub_wx_i16m1(op0, op1, op2) \ +__builtin_rvv_vwsub_wx_i16m1((vint16m1_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vwsub_wx_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_wx_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vwsub_wx_i16m2(op0, op1, op2) \ +__builtin_rvv_vwsub_wx_i16m2((vint16m2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vwsub_wx_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_wx_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vwsub_wx_i16m4(op0, op1, op2) \ +__builtin_rvv_vwsub_wx_i16m4((vint16m4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vwsub_wx_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_wx_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vwsub_wx_i16m8(op0, op1, op2) \ +__builtin_rvv_vwsub_wx_i16m8((vint16m8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vwsub_wx_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_wx_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vwsub_wx_i32mf2(op0, op1, op2) \ +__builtin_rvv_vwsub_wx_i32mf2((vint32mf2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vwsub_wx_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_wx_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vwsub_wx_i32m1(op0, op1, op2) \ +__builtin_rvv_vwsub_wx_i32m1((vint32m1_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vwsub_wx_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_wx_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vwsub_wx_i32m2(op0, op1, op2) \ +__builtin_rvv_vwsub_wx_i32m2((vint32m2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vwsub_wx_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_wx_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vwsub_wx_i32m4(op0, op1, op2) \ +__builtin_rvv_vwsub_wx_i32m4((vint32m4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vwsub_wx_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_wx_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vwsub_wx_i32m8(op0, op1, op2) \ +__builtin_rvv_vwsub_wx_i32m8((vint32m8_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vwsub_wx_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_wx_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vwsub_wx_i64m1(op0, op1, op2) \ +__builtin_rvv_vwsub_wx_i64m1((vint64m1_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vwsub_wx_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_wx_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vwsub_wx_i64m2(op0, op1, op2) \ +__builtin_rvv_vwsub_wx_i64m2((vint64m2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vwsub_wx_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_wx_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vwsub_wx_i64m4(op0, op1, op2) \ +__builtin_rvv_vwsub_wx_i64m4((vint64m4_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vwsub_wx_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_wx_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vwsub_wx_i64m8(op0, op1, op2) \ +__builtin_rvv_vwsub_wx_i64m8((vint64m8_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vwsub_wx_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwsub_wx_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vwcvt_x_x_v_i16m2(op0, op1) \ +__builtin_rvv_vwcvt_x_x_v_i16m2((vint8m1_t)(op0), (size_t)(op1)) +#define vwcvt_x_x_v_i16m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vwcvt_x_x_v_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vwcvt_x_x_v_i16m4(op0, op1) \ +__builtin_rvv_vwcvt_x_x_v_i16m4((vint8m2_t)(op0), (size_t)(op1)) +#define vwcvt_x_x_v_i16m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vwcvt_x_x_v_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint8m2_t)(op2), (size_t)(op3)) +#define vwcvt_x_x_v_i16m8(op0, op1) \ +__builtin_rvv_vwcvt_x_x_v_i16m8((vint8m4_t)(op0), (size_t)(op1)) +#define vwcvt_x_x_v_i16m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vwcvt_x_x_v_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint8m4_t)(op2), (size_t)(op3)) +#define vwcvt_x_x_v_i16m1(op0, op1) \ +__builtin_rvv_vwcvt_x_x_v_i16m1((vint8mf2_t)(op0), (size_t)(op1)) +#define vwcvt_x_x_v_i16m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vwcvt_x_x_v_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint8mf2_t)(op2), (size_t)(op3)) +#define vwcvt_x_x_v_i16mf2(op0, op1) \ +__builtin_rvv_vwcvt_x_x_v_i16mf2((vint8mf4_t)(op0), (size_t)(op1)) +#define vwcvt_x_x_v_i16mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vwcvt_x_x_v_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint8mf4_t)(op2), (size_t)(op3)) +#define vwcvt_x_x_v_i16mf4(op0, op1) \ +__builtin_rvv_vwcvt_x_x_v_i16mf4((vint8mf8_t)(op0), (size_t)(op1)) +#define vwcvt_x_x_v_i16mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vwcvt_x_x_v_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint8mf8_t)(op2), (size_t)(op3)) +#define vwcvt_x_x_v_i32m2(op0, op1) \ +__builtin_rvv_vwcvt_x_x_v_i32m2((vint16m1_t)(op0), (size_t)(op1)) +#define vwcvt_x_x_v_i32m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vwcvt_x_x_v_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vwcvt_x_x_v_i32m4(op0, op1) \ +__builtin_rvv_vwcvt_x_x_v_i32m4((vint16m2_t)(op0), (size_t)(op1)) +#define vwcvt_x_x_v_i32m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vwcvt_x_x_v_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint16m2_t)(op2), (size_t)(op3)) +#define vwcvt_x_x_v_i32m8(op0, op1) \ +__builtin_rvv_vwcvt_x_x_v_i32m8((vint16m4_t)(op0), (size_t)(op1)) +#define vwcvt_x_x_v_i32m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vwcvt_x_x_v_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint16m4_t)(op2), (size_t)(op3)) +#define vwcvt_x_x_v_i32m1(op0, op1) \ +__builtin_rvv_vwcvt_x_x_v_i32m1((vint16mf2_t)(op0), (size_t)(op1)) +#define vwcvt_x_x_v_i32m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vwcvt_x_x_v_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint16mf2_t)(op2), (size_t)(op3)) +#define vwcvt_x_x_v_i32mf2(op0, op1) \ +__builtin_rvv_vwcvt_x_x_v_i32mf2((vint16mf4_t)(op0), (size_t)(op1)) +#define vwcvt_x_x_v_i32mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vwcvt_x_x_v_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint16mf4_t)(op2), (size_t)(op3)) +#define vwcvt_x_x_v_i64m2(op0, op1) \ +__builtin_rvv_vwcvt_x_x_v_i64m2((vint32m1_t)(op0), (size_t)(op1)) +#define vwcvt_x_x_v_i64m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vwcvt_x_x_v_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vwcvt_x_x_v_i64m4(op0, op1) \ +__builtin_rvv_vwcvt_x_x_v_i64m4((vint32m2_t)(op0), (size_t)(op1)) +#define vwcvt_x_x_v_i64m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vwcvt_x_x_v_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vwcvt_x_x_v_i64m8(op0, op1) \ +__builtin_rvv_vwcvt_x_x_v_i64m8((vint32m4_t)(op0), (size_t)(op1)) +#define vwcvt_x_x_v_i64m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vwcvt_x_x_v_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vwcvt_x_x_v_i64m1(op0, op1) \ +__builtin_rvv_vwcvt_x_x_v_i64m1((vint32mf2_t)(op0), (size_t)(op1)) +#define vwcvt_x_x_v_i64m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vwcvt_x_x_v_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vsuxei64_v_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei64_v_u8m1((uint8_t *)(op0), (vuint64m8_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vsuxei64_v_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei64_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint64m8_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vsuxei64_v_u8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei64_v_u8mf2((uint8_t *)(op0), (vuint64m4_t)(op1), (vuint8mf2_t)(op2), (size_t)(op3)) +#define vsuxei64_v_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei64_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint64m4_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vsuxei64_v_u8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei64_v_u8mf4((uint8_t *)(op0), (vuint64m2_t)(op1), (vuint8mf4_t)(op2), (size_t)(op3)) +#define vsuxei64_v_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei64_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint64m2_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vsuxei64_v_u8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei64_v_u8mf8((uint8_t *)(op0), (vuint64m1_t)(op1), (vuint8mf8_t)(op2), (size_t)(op3)) +#define vsuxei64_v_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei64_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint64m1_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vadc_vvm_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vvm_i8m1((vint8m1_t)(op0), (vint8m1_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vadc_vvm_i8m2(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vvm_i8m2((vint8m2_t)(op0), (vint8m2_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vadc_vvm_i8m4(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vvm_i8m4((vint8m4_t)(op0), (vint8m4_t)(op1), (vbool2_t)(op2), (size_t)(op3)) +#define vadc_vvm_i8m8(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vvm_i8m8((vint8m8_t)(op0), (vint8m8_t)(op1), (vbool1_t)(op2), (size_t)(op3)) +#define vadc_vvm_i8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vvm_i8mf2((vint8mf2_t)(op0), (vint8mf2_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vadc_vvm_i8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vvm_i8mf4((vint8mf4_t)(op0), (vint8mf4_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vadc_vvm_i8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vvm_i8mf8((vint8mf8_t)(op0), (vint8mf8_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vadc_vvm_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vvm_i16m1((vint16m1_t)(op0), (vint16m1_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vadc_vvm_i16m2(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vvm_i16m2((vint16m2_t)(op0), (vint16m2_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vadc_vvm_i16m4(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vvm_i16m4((vint16m4_t)(op0), (vint16m4_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vadc_vvm_i16m8(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vvm_i16m8((vint16m8_t)(op0), (vint16m8_t)(op1), (vbool2_t)(op2), (size_t)(op3)) +#define vadc_vvm_i16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vvm_i16mf2((vint16mf2_t)(op0), (vint16mf2_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vadc_vvm_i16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vvm_i16mf4((vint16mf4_t)(op0), (vint16mf4_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vadc_vvm_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vvm_i32m1((vint32m1_t)(op0), (vint32m1_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vadc_vvm_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vvm_i32m2((vint32m2_t)(op0), (vint32m2_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vadc_vvm_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vvm_i32m4((vint32m4_t)(op0), (vint32m4_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vadc_vvm_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vvm_i32m8((vint32m8_t)(op0), (vint32m8_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vadc_vvm_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vvm_i32mf2((vint32mf2_t)(op0), (vint32mf2_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vadc_vvm_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vvm_i64m1((vint64m1_t)(op0), (vint64m1_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vadc_vvm_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vvm_i64m2((vint64m2_t)(op0), (vint64m2_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vadc_vvm_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vvm_i64m4((vint64m4_t)(op0), (vint64m4_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vadc_vvm_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vvm_i64m8((vint64m8_t)(op0), (vint64m8_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vadc_vxm_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vxm_i8m1((vint8m1_t)(op0), (int8_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vadc_vxm_i8m2(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vxm_i8m2((vint8m2_t)(op0), (int8_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vadc_vxm_i8m4(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vxm_i8m4((vint8m4_t)(op0), (int8_t)(op1), (vbool2_t)(op2), (size_t)(op3)) +#define vadc_vxm_i8m8(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vxm_i8m8((vint8m8_t)(op0), (int8_t)(op1), (vbool1_t)(op2), (size_t)(op3)) +#define vadc_vxm_i8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vxm_i8mf2((vint8mf2_t)(op0), (int8_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vadc_vxm_i8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vxm_i8mf4((vint8mf4_t)(op0), (int8_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vadc_vxm_i8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vxm_i8mf8((vint8mf8_t)(op0), (int8_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vadc_vxm_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vxm_i16m1((vint16m1_t)(op0), (int16_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vadc_vxm_i16m2(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vxm_i16m2((vint16m2_t)(op0), (int16_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vadc_vxm_i16m4(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vxm_i16m4((vint16m4_t)(op0), (int16_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vadc_vxm_i16m8(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vxm_i16m8((vint16m8_t)(op0), (int16_t)(op1), (vbool2_t)(op2), (size_t)(op3)) +#define vadc_vxm_i16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vxm_i16mf2((vint16mf2_t)(op0), (int16_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vadc_vxm_i16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vxm_i16mf4((vint16mf4_t)(op0), (int16_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vadc_vxm_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vxm_i32m1((vint32m1_t)(op0), (int32_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vadc_vxm_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vxm_i32m2((vint32m2_t)(op0), (int32_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vadc_vxm_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vxm_i32m4((vint32m4_t)(op0), (int32_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vadc_vxm_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vxm_i32m8((vint32m8_t)(op0), (int32_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vadc_vxm_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vxm_i32mf2((vint32mf2_t)(op0), (int32_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vadc_vxm_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vxm_i64m1((vint64m1_t)(op0), (int64_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vadc_vxm_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vxm_i64m2((vint64m2_t)(op0), (int64_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vadc_vxm_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vxm_i64m4((vint64m4_t)(op0), (int64_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vadc_vxm_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vxm_i64m8((vint64m8_t)(op0), (int64_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vadc_vvm_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vvm_u8m1((vuint8m1_t)(op0), (vuint8m1_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vadc_vvm_u8m2(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vvm_u8m2((vuint8m2_t)(op0), (vuint8m2_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vadc_vvm_u8m4(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vvm_u8m4((vuint8m4_t)(op0), (vuint8m4_t)(op1), (vbool2_t)(op2), (size_t)(op3)) +#define vadc_vvm_u8m8(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vvm_u8m8((vuint8m8_t)(op0), (vuint8m8_t)(op1), (vbool1_t)(op2), (size_t)(op3)) +#define vadc_vvm_u8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vvm_u8mf2((vuint8mf2_t)(op0), (vuint8mf2_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vadc_vvm_u8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vvm_u8mf4((vuint8mf4_t)(op0), (vuint8mf4_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vadc_vvm_u8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vvm_u8mf8((vuint8mf8_t)(op0), (vuint8mf8_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vadc_vvm_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vvm_u16m1((vuint16m1_t)(op0), (vuint16m1_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vadc_vvm_u16m2(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vvm_u16m2((vuint16m2_t)(op0), (vuint16m2_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vadc_vvm_u16m4(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vvm_u16m4((vuint16m4_t)(op0), (vuint16m4_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vadc_vvm_u16m8(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vvm_u16m8((vuint16m8_t)(op0), (vuint16m8_t)(op1), (vbool2_t)(op2), (size_t)(op3)) +#define vadc_vvm_u16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vvm_u16mf2((vuint16mf2_t)(op0), (vuint16mf2_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vadc_vvm_u16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vvm_u16mf4((vuint16mf4_t)(op0), (vuint16mf4_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vadc_vvm_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vvm_u32m1((vuint32m1_t)(op0), (vuint32m1_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vadc_vvm_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vvm_u32m2((vuint32m2_t)(op0), (vuint32m2_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vadc_vvm_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vvm_u32m4((vuint32m4_t)(op0), (vuint32m4_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vadc_vvm_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vvm_u32m8((vuint32m8_t)(op0), (vuint32m8_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vadc_vvm_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vvm_u32mf2((vuint32mf2_t)(op0), (vuint32mf2_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vadc_vvm_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vvm_u64m1((vuint64m1_t)(op0), (vuint64m1_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vadc_vvm_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vvm_u64m2((vuint64m2_t)(op0), (vuint64m2_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vadc_vvm_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vvm_u64m4((vuint64m4_t)(op0), (vuint64m4_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vadc_vvm_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vvm_u64m8((vuint64m8_t)(op0), (vuint64m8_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vadc_vxm_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vxm_u8m1((vuint8m1_t)(op0), (uint8_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vadc_vxm_u8m2(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vxm_u8m2((vuint8m2_t)(op0), (uint8_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vadc_vxm_u8m4(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vxm_u8m4((vuint8m4_t)(op0), (uint8_t)(op1), (vbool2_t)(op2), (size_t)(op3)) +#define vadc_vxm_u8m8(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vxm_u8m8((vuint8m8_t)(op0), (uint8_t)(op1), (vbool1_t)(op2), (size_t)(op3)) +#define vadc_vxm_u8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vxm_u8mf2((vuint8mf2_t)(op0), (uint8_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vadc_vxm_u8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vxm_u8mf4((vuint8mf4_t)(op0), (uint8_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vadc_vxm_u8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vxm_u8mf8((vuint8mf8_t)(op0), (uint8_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vadc_vxm_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vxm_u16m1((vuint16m1_t)(op0), (uint16_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vadc_vxm_u16m2(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vxm_u16m2((vuint16m2_t)(op0), (uint16_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vadc_vxm_u16m4(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vxm_u16m4((vuint16m4_t)(op0), (uint16_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vadc_vxm_u16m8(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vxm_u16m8((vuint16m8_t)(op0), (uint16_t)(op1), (vbool2_t)(op2), (size_t)(op3)) +#define vadc_vxm_u16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vxm_u16mf2((vuint16mf2_t)(op0), (uint16_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vadc_vxm_u16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vxm_u16mf4((vuint16mf4_t)(op0), (uint16_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vadc_vxm_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vxm_u32m1((vuint32m1_t)(op0), (uint32_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vadc_vxm_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vxm_u32m2((vuint32m2_t)(op0), (uint32_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vadc_vxm_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vxm_u32m4((vuint32m4_t)(op0), (uint32_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vadc_vxm_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vxm_u32m8((vuint32m8_t)(op0), (uint32_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vadc_vxm_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vxm_u32mf2((vuint32mf2_t)(op0), (uint32_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vadc_vxm_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vxm_u64m1((vuint64m1_t)(op0), (uint64_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vadc_vxm_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vxm_u64m2((vuint64m2_t)(op0), (uint64_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vadc_vxm_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vxm_u64m4((vuint64m4_t)(op0), (uint64_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vadc_vxm_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vadc_vxm_u64m8((vuint64m8_t)(op0), (uint64_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vmadc_vvm_i8m1_b8(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vvm_i8m1_b8((vint8m1_t)(op0), (vint8m1_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vmadc_vvm_i8m2_b4(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vvm_i8m2_b4((vint8m2_t)(op0), (vint8m2_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vmadc_vvm_i8m4_b2(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vvm_i8m4_b2((vint8m4_t)(op0), (vint8m4_t)(op1), (vbool2_t)(op2), (size_t)(op3)) +#define vmadc_vvm_i8m8_b1(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vvm_i8m8_b1((vint8m8_t)(op0), (vint8m8_t)(op1), (vbool1_t)(op2), (size_t)(op3)) +#define vmadc_vvm_i8mf2_b16(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vvm_i8mf2_b16((vint8mf2_t)(op0), (vint8mf2_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vmadc_vvm_i8mf4_b32(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vvm_i8mf4_b32((vint8mf4_t)(op0), (vint8mf4_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vmadc_vvm_i8mf8_b64(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vvm_i8mf8_b64((vint8mf8_t)(op0), (vint8mf8_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vmadc_vvm_i16m1_b16(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vvm_i16m1_b16((vint16m1_t)(op0), (vint16m1_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vmadc_vvm_i16m2_b8(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vvm_i16m2_b8((vint16m2_t)(op0), (vint16m2_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vmadc_vvm_i16m4_b4(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vvm_i16m4_b4((vint16m4_t)(op0), (vint16m4_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vmadc_vvm_i16m8_b2(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vvm_i16m8_b2((vint16m8_t)(op0), (vint16m8_t)(op1), (vbool2_t)(op2), (size_t)(op3)) +#define vmadc_vvm_i16mf2_b32(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vvm_i16mf2_b32((vint16mf2_t)(op0), (vint16mf2_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vmadc_vvm_i16mf4_b64(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vvm_i16mf4_b64((vint16mf4_t)(op0), (vint16mf4_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vmadc_vvm_i32m1_b32(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vvm_i32m1_b32((vint32m1_t)(op0), (vint32m1_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vmadc_vvm_i32m2_b16(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vvm_i32m2_b16((vint32m2_t)(op0), (vint32m2_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vmadc_vvm_i32m4_b8(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vvm_i32m4_b8((vint32m4_t)(op0), (vint32m4_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vmadc_vvm_i32m8_b4(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vvm_i32m8_b4((vint32m8_t)(op0), (vint32m8_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vmadc_vvm_i32mf2_b64(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vvm_i32mf2_b64((vint32mf2_t)(op0), (vint32mf2_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vmadc_vvm_i64m1_b64(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vvm_i64m1_b64((vint64m1_t)(op0), (vint64m1_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vmadc_vvm_i64m2_b32(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vvm_i64m2_b32((vint64m2_t)(op0), (vint64m2_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vmadc_vvm_i64m4_b16(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vvm_i64m4_b16((vint64m4_t)(op0), (vint64m4_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vmadc_vvm_i64m8_b8(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vvm_i64m8_b8((vint64m8_t)(op0), (vint64m8_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vmadc_vxm_i8m1_b8(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vxm_i8m1_b8((vint8m1_t)(op0), (int8_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vmadc_vxm_i8m2_b4(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vxm_i8m2_b4((vint8m2_t)(op0), (int8_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vmadc_vxm_i8m4_b2(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vxm_i8m4_b2((vint8m4_t)(op0), (int8_t)(op1), (vbool2_t)(op2), (size_t)(op3)) +#define vmadc_vxm_i8m8_b1(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vxm_i8m8_b1((vint8m8_t)(op0), (int8_t)(op1), (vbool1_t)(op2), (size_t)(op3)) +#define vmadc_vxm_i8mf2_b16(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vxm_i8mf2_b16((vint8mf2_t)(op0), (int8_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vmadc_vxm_i8mf4_b32(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vxm_i8mf4_b32((vint8mf4_t)(op0), (int8_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vmadc_vxm_i8mf8_b64(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vxm_i8mf8_b64((vint8mf8_t)(op0), (int8_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vmadc_vxm_i16m1_b16(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vxm_i16m1_b16((vint16m1_t)(op0), (int16_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vmadc_vxm_i16m2_b8(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vxm_i16m2_b8((vint16m2_t)(op0), (int16_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vmadc_vxm_i16m4_b4(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vxm_i16m4_b4((vint16m4_t)(op0), (int16_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vmadc_vxm_i16m8_b2(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vxm_i16m8_b2((vint16m8_t)(op0), (int16_t)(op1), (vbool2_t)(op2), (size_t)(op3)) +#define vmadc_vxm_i16mf2_b32(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vxm_i16mf2_b32((vint16mf2_t)(op0), (int16_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vmadc_vxm_i16mf4_b64(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vxm_i16mf4_b64((vint16mf4_t)(op0), (int16_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vmadc_vxm_i32m1_b32(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vxm_i32m1_b32((vint32m1_t)(op0), (int32_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vmadc_vxm_i32m2_b16(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vxm_i32m2_b16((vint32m2_t)(op0), (int32_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vmadc_vxm_i32m4_b8(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vxm_i32m4_b8((vint32m4_t)(op0), (int32_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vmadc_vxm_i32m8_b4(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vxm_i32m8_b4((vint32m8_t)(op0), (int32_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vmadc_vxm_i32mf2_b64(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vxm_i32mf2_b64((vint32mf2_t)(op0), (int32_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vmadc_vxm_i64m1_b64(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vxm_i64m1_b64((vint64m1_t)(op0), (int64_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vmadc_vxm_i64m2_b32(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vxm_i64m2_b32((vint64m2_t)(op0), (int64_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vmadc_vxm_i64m4_b16(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vxm_i64m4_b16((vint64m4_t)(op0), (int64_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vmadc_vxm_i64m8_b8(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vxm_i64m8_b8((vint64m8_t)(op0), (int64_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vmadc_vvm_u8m1_b8(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vvm_u8m1_b8((vuint8m1_t)(op0), (vuint8m1_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vmadc_vvm_u8m2_b4(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vvm_u8m2_b4((vuint8m2_t)(op0), (vuint8m2_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vmadc_vvm_u8m4_b2(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vvm_u8m4_b2((vuint8m4_t)(op0), (vuint8m4_t)(op1), (vbool2_t)(op2), (size_t)(op3)) +#define vmadc_vvm_u8m8_b1(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vvm_u8m8_b1((vuint8m8_t)(op0), (vuint8m8_t)(op1), (vbool1_t)(op2), (size_t)(op3)) +#define vmadc_vvm_u8mf2_b16(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vvm_u8mf2_b16((vuint8mf2_t)(op0), (vuint8mf2_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vmadc_vvm_u8mf4_b32(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vvm_u8mf4_b32((vuint8mf4_t)(op0), (vuint8mf4_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vmadc_vvm_u8mf8_b64(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vvm_u8mf8_b64((vuint8mf8_t)(op0), (vuint8mf8_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vmadc_vvm_u16m1_b16(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vvm_u16m1_b16((vuint16m1_t)(op0), (vuint16m1_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vmadc_vvm_u16m2_b8(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vvm_u16m2_b8((vuint16m2_t)(op0), (vuint16m2_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vmadc_vvm_u16m4_b4(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vvm_u16m4_b4((vuint16m4_t)(op0), (vuint16m4_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vmadc_vvm_u16m8_b2(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vvm_u16m8_b2((vuint16m8_t)(op0), (vuint16m8_t)(op1), (vbool2_t)(op2), (size_t)(op3)) +#define vmadc_vvm_u16mf2_b32(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vvm_u16mf2_b32((vuint16mf2_t)(op0), (vuint16mf2_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vmadc_vvm_u16mf4_b64(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vvm_u16mf4_b64((vuint16mf4_t)(op0), (vuint16mf4_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vmadc_vvm_u32m1_b32(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vvm_u32m1_b32((vuint32m1_t)(op0), (vuint32m1_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vmadc_vvm_u32m2_b16(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vvm_u32m2_b16((vuint32m2_t)(op0), (vuint32m2_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vmadc_vvm_u32m4_b8(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vvm_u32m4_b8((vuint32m4_t)(op0), (vuint32m4_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vmadc_vvm_u32m8_b4(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vvm_u32m8_b4((vuint32m8_t)(op0), (vuint32m8_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vmadc_vvm_u32mf2_b64(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vvm_u32mf2_b64((vuint32mf2_t)(op0), (vuint32mf2_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vmadc_vvm_u64m1_b64(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vvm_u64m1_b64((vuint64m1_t)(op0), (vuint64m1_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vmadc_vvm_u64m2_b32(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vvm_u64m2_b32((vuint64m2_t)(op0), (vuint64m2_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vmadc_vvm_u64m4_b16(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vvm_u64m4_b16((vuint64m4_t)(op0), (vuint64m4_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vmadc_vvm_u64m8_b8(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vvm_u64m8_b8((vuint64m8_t)(op0), (vuint64m8_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vmadc_vxm_u8m1_b8(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vxm_u8m1_b8((vuint8m1_t)(op0), (uint8_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vmadc_vxm_u8m2_b4(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vxm_u8m2_b4((vuint8m2_t)(op0), (uint8_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vmadc_vxm_u8m4_b2(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vxm_u8m4_b2((vuint8m4_t)(op0), (uint8_t)(op1), (vbool2_t)(op2), (size_t)(op3)) +#define vmadc_vxm_u8m8_b1(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vxm_u8m8_b1((vuint8m8_t)(op0), (uint8_t)(op1), (vbool1_t)(op2), (size_t)(op3)) +#define vmadc_vxm_u8mf2_b16(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vxm_u8mf2_b16((vuint8mf2_t)(op0), (uint8_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vmadc_vxm_u8mf4_b32(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vxm_u8mf4_b32((vuint8mf4_t)(op0), (uint8_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vmadc_vxm_u8mf8_b64(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vxm_u8mf8_b64((vuint8mf8_t)(op0), (uint8_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vmadc_vxm_u16m1_b16(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vxm_u16m1_b16((vuint16m1_t)(op0), (uint16_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vmadc_vxm_u16m2_b8(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vxm_u16m2_b8((vuint16m2_t)(op0), (uint16_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vmadc_vxm_u16m4_b4(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vxm_u16m4_b4((vuint16m4_t)(op0), (uint16_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vmadc_vxm_u16m8_b2(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vxm_u16m8_b2((vuint16m8_t)(op0), (uint16_t)(op1), (vbool2_t)(op2), (size_t)(op3)) +#define vmadc_vxm_u16mf2_b32(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vxm_u16mf2_b32((vuint16mf2_t)(op0), (uint16_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vmadc_vxm_u16mf4_b64(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vxm_u16mf4_b64((vuint16mf4_t)(op0), (uint16_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vmadc_vxm_u32m1_b32(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vxm_u32m1_b32((vuint32m1_t)(op0), (uint32_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vmadc_vxm_u32m2_b16(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vxm_u32m2_b16((vuint32m2_t)(op0), (uint32_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vmadc_vxm_u32m4_b8(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vxm_u32m4_b8((vuint32m4_t)(op0), (uint32_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vmadc_vxm_u32m8_b4(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vxm_u32m8_b4((vuint32m8_t)(op0), (uint32_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vmadc_vxm_u32mf2_b64(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vxm_u32mf2_b64((vuint32mf2_t)(op0), (uint32_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vmadc_vxm_u64m1_b64(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vxm_u64m1_b64((vuint64m1_t)(op0), (uint64_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vmadc_vxm_u64m2_b32(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vxm_u64m2_b32((vuint64m2_t)(op0), (uint64_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vmadc_vxm_u64m4_b16(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vxm_u64m4_b16((vuint64m4_t)(op0), (uint64_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vmadc_vxm_u64m8_b8(op0, op1, op2, op3) \ +__builtin_rvv_vmadc_vxm_u64m8_b8((vuint64m8_t)(op0), (uint64_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vmadc_vv_i8m1_b8(op0, op1, op2) \ +__builtin_rvv_vmadc_vv_i8m1_b8((vint8m1_t)(op0), (vint8m1_t)(op1), (size_t)(op2)) +#define vmadc_vv_i8m2_b4(op0, op1, op2) \ +__builtin_rvv_vmadc_vv_i8m2_b4((vint8m2_t)(op0), (vint8m2_t)(op1), (size_t)(op2)) +#define vmadc_vv_i8m4_b2(op0, op1, op2) \ +__builtin_rvv_vmadc_vv_i8m4_b2((vint8m4_t)(op0), (vint8m4_t)(op1), (size_t)(op2)) +#define vmadc_vv_i8m8_b1(op0, op1, op2) \ +__builtin_rvv_vmadc_vv_i8m8_b1((vint8m8_t)(op0), (vint8m8_t)(op1), (size_t)(op2)) +#define vmadc_vv_i8mf2_b16(op0, op1, op2) \ +__builtin_rvv_vmadc_vv_i8mf2_b16((vint8mf2_t)(op0), (vint8mf2_t)(op1), (size_t)(op2)) +#define vmadc_vv_i8mf4_b32(op0, op1, op2) \ +__builtin_rvv_vmadc_vv_i8mf4_b32((vint8mf4_t)(op0), (vint8mf4_t)(op1), (size_t)(op2)) +#define vmadc_vv_i8mf8_b64(op0, op1, op2) \ +__builtin_rvv_vmadc_vv_i8mf8_b64((vint8mf8_t)(op0), (vint8mf8_t)(op1), (size_t)(op2)) +#define vmadc_vv_i16m1_b16(op0, op1, op2) \ +__builtin_rvv_vmadc_vv_i16m1_b16((vint16m1_t)(op0), (vint16m1_t)(op1), (size_t)(op2)) +#define vmadc_vv_i16m2_b8(op0, op1, op2) \ +__builtin_rvv_vmadc_vv_i16m2_b8((vint16m2_t)(op0), (vint16m2_t)(op1), (size_t)(op2)) +#define vmadc_vv_i16m4_b4(op0, op1, op2) \ +__builtin_rvv_vmadc_vv_i16m4_b4((vint16m4_t)(op0), (vint16m4_t)(op1), (size_t)(op2)) +#define vmadc_vv_i16m8_b2(op0, op1, op2) \ +__builtin_rvv_vmadc_vv_i16m8_b2((vint16m8_t)(op0), (vint16m8_t)(op1), (size_t)(op2)) +#define vmadc_vv_i16mf2_b32(op0, op1, op2) \ +__builtin_rvv_vmadc_vv_i16mf2_b32((vint16mf2_t)(op0), (vint16mf2_t)(op1), (size_t)(op2)) +#define vmadc_vv_i16mf4_b64(op0, op1, op2) \ +__builtin_rvv_vmadc_vv_i16mf4_b64((vint16mf4_t)(op0), (vint16mf4_t)(op1), (size_t)(op2)) +#define vmadc_vv_i32m1_b32(op0, op1, op2) \ +__builtin_rvv_vmadc_vv_i32m1_b32((vint32m1_t)(op0), (vint32m1_t)(op1), (size_t)(op2)) +#define vmadc_vv_i32m2_b16(op0, op1, op2) \ +__builtin_rvv_vmadc_vv_i32m2_b16((vint32m2_t)(op0), (vint32m2_t)(op1), (size_t)(op2)) +#define vmadc_vv_i32m4_b8(op0, op1, op2) \ +__builtin_rvv_vmadc_vv_i32m4_b8((vint32m4_t)(op0), (vint32m4_t)(op1), (size_t)(op2)) +#define vmadc_vv_i32m8_b4(op0, op1, op2) \ +__builtin_rvv_vmadc_vv_i32m8_b4((vint32m8_t)(op0), (vint32m8_t)(op1), (size_t)(op2)) +#define vmadc_vv_i32mf2_b64(op0, op1, op2) \ +__builtin_rvv_vmadc_vv_i32mf2_b64((vint32mf2_t)(op0), (vint32mf2_t)(op1), (size_t)(op2)) +#define vmadc_vv_i64m1_b64(op0, op1, op2) \ +__builtin_rvv_vmadc_vv_i64m1_b64((vint64m1_t)(op0), (vint64m1_t)(op1), (size_t)(op2)) +#define vmadc_vv_i64m2_b32(op0, op1, op2) \ +__builtin_rvv_vmadc_vv_i64m2_b32((vint64m2_t)(op0), (vint64m2_t)(op1), (size_t)(op2)) +#define vmadc_vv_i64m4_b16(op0, op1, op2) \ +__builtin_rvv_vmadc_vv_i64m4_b16((vint64m4_t)(op0), (vint64m4_t)(op1), (size_t)(op2)) +#define vmadc_vv_i64m8_b8(op0, op1, op2) \ +__builtin_rvv_vmadc_vv_i64m8_b8((vint64m8_t)(op0), (vint64m8_t)(op1), (size_t)(op2)) +#define vmadc_vx_i8m1_b8(op0, op1, op2) \ +__builtin_rvv_vmadc_vx_i8m1_b8((vint8m1_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmadc_vx_i8m2_b4(op0, op1, op2) \ +__builtin_rvv_vmadc_vx_i8m2_b4((vint8m2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmadc_vx_i8m4_b2(op0, op1, op2) \ +__builtin_rvv_vmadc_vx_i8m4_b2((vint8m4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmadc_vx_i8m8_b1(op0, op1, op2) \ +__builtin_rvv_vmadc_vx_i8m8_b1((vint8m8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmadc_vx_i8mf2_b16(op0, op1, op2) \ +__builtin_rvv_vmadc_vx_i8mf2_b16((vint8mf2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmadc_vx_i8mf4_b32(op0, op1, op2) \ +__builtin_rvv_vmadc_vx_i8mf4_b32((vint8mf4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmadc_vx_i8mf8_b64(op0, op1, op2) \ +__builtin_rvv_vmadc_vx_i8mf8_b64((vint8mf8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmadc_vx_i16m1_b16(op0, op1, op2) \ +__builtin_rvv_vmadc_vx_i16m1_b16((vint16m1_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmadc_vx_i16m2_b8(op0, op1, op2) \ +__builtin_rvv_vmadc_vx_i16m2_b8((vint16m2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmadc_vx_i16m4_b4(op0, op1, op2) \ +__builtin_rvv_vmadc_vx_i16m4_b4((vint16m4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmadc_vx_i16m8_b2(op0, op1, op2) \ +__builtin_rvv_vmadc_vx_i16m8_b2((vint16m8_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmadc_vx_i16mf2_b32(op0, op1, op2) \ +__builtin_rvv_vmadc_vx_i16mf2_b32((vint16mf2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmadc_vx_i16mf4_b64(op0, op1, op2) \ +__builtin_rvv_vmadc_vx_i16mf4_b64((vint16mf4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmadc_vx_i32m1_b32(op0, op1, op2) \ +__builtin_rvv_vmadc_vx_i32m1_b32((vint32m1_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmadc_vx_i32m2_b16(op0, op1, op2) \ +__builtin_rvv_vmadc_vx_i32m2_b16((vint32m2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmadc_vx_i32m4_b8(op0, op1, op2) \ +__builtin_rvv_vmadc_vx_i32m4_b8((vint32m4_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmadc_vx_i32m8_b4(op0, op1, op2) \ +__builtin_rvv_vmadc_vx_i32m8_b4((vint32m8_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmadc_vx_i32mf2_b64(op0, op1, op2) \ +__builtin_rvv_vmadc_vx_i32mf2_b64((vint32mf2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmadc_vx_i64m1_b64(op0, op1, op2) \ +__builtin_rvv_vmadc_vx_i64m1_b64((vint64m1_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmadc_vx_i64m2_b32(op0, op1, op2) \ +__builtin_rvv_vmadc_vx_i64m2_b32((vint64m2_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmadc_vx_i64m4_b16(op0, op1, op2) \ +__builtin_rvv_vmadc_vx_i64m4_b16((vint64m4_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmadc_vx_i64m8_b8(op0, op1, op2) \ +__builtin_rvv_vmadc_vx_i64m8_b8((vint64m8_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vsuxei8_v_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_i16m1((int16_t *)(op0), (vuint8mf2_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vsuxei8_v_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint8mf2_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vsuxei8_v_i16m2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_i16m2((int16_t *)(op0), (vuint8m1_t)(op1), (vint16m2_t)(op2), (size_t)(op3)) +#define vsuxei8_v_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_i16m2_m((vbool8_t)(op0), (int16_t *)(op1), (vuint8m1_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vsuxei8_v_i16m4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_i16m4((int16_t *)(op0), (vuint8m2_t)(op1), (vint16m4_t)(op2), (size_t)(op3)) +#define vsuxei8_v_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_i16m4_m((vbool4_t)(op0), (int16_t *)(op1), (vuint8m2_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vsuxei8_v_i16m8(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_i16m8((int16_t *)(op0), (vuint8m4_t)(op1), (vint16m8_t)(op2), (size_t)(op3)) +#define vsuxei8_v_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_i16m8_m((vbool2_t)(op0), (int16_t *)(op1), (vuint8m4_t)(op2), (vint16m8_t)(op3), (size_t)(op4)) +#define vsuxei8_v_i16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_i16mf2((int16_t *)(op0), (vuint8mf4_t)(op1), (vint16mf2_t)(op2), (size_t)(op3)) +#define vsuxei8_v_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint8mf4_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vsuxei8_v_i16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_i16mf4((int16_t *)(op0), (vuint8mf8_t)(op1), (vint16mf4_t)(op2), (size_t)(op3)) +#define vsuxei8_v_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint8mf8_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vmadc_vv_u8m1_b8(op0, op1, op2) \ +__builtin_rvv_vmadc_vv_u8m1_b8((vuint8m1_t)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vmadc_vv_u8m2_b4(op0, op1, op2) \ +__builtin_rvv_vmadc_vv_u8m2_b4((vuint8m2_t)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vmadc_vv_u8m4_b2(op0, op1, op2) \ +__builtin_rvv_vmadc_vv_u8m4_b2((vuint8m4_t)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vmadc_vv_u8m8_b1(op0, op1, op2) \ +__builtin_rvv_vmadc_vv_u8m8_b1((vuint8m8_t)(op0), (vuint8m8_t)(op1), (size_t)(op2)) +#define vmadc_vv_u8mf2_b16(op0, op1, op2) \ +__builtin_rvv_vmadc_vv_u8mf2_b16((vuint8mf2_t)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vmadc_vv_u8mf4_b32(op0, op1, op2) \ +__builtin_rvv_vmadc_vv_u8mf4_b32((vuint8mf4_t)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vmadc_vv_u8mf8_b64(op0, op1, op2) \ +__builtin_rvv_vmadc_vv_u8mf8_b64((vuint8mf8_t)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vmadc_vv_u16m1_b16(op0, op1, op2) \ +__builtin_rvv_vmadc_vv_u16m1_b16((vuint16m1_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vmadc_vv_u16m2_b8(op0, op1, op2) \ +__builtin_rvv_vmadc_vv_u16m2_b8((vuint16m2_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vmadc_vv_u16m4_b4(op0, op1, op2) \ +__builtin_rvv_vmadc_vv_u16m4_b4((vuint16m4_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vmadc_vv_u16m8_b2(op0, op1, op2) \ +__builtin_rvv_vmadc_vv_u16m8_b2((vuint16m8_t)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vmadc_vv_u16mf2_b32(op0, op1, op2) \ +__builtin_rvv_vmadc_vv_u16mf2_b32((vuint16mf2_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vmadc_vv_u16mf4_b64(op0, op1, op2) \ +__builtin_rvv_vmadc_vv_u16mf4_b64((vuint16mf4_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vmadc_vv_u32m1_b32(op0, op1, op2) \ +__builtin_rvv_vmadc_vv_u32m1_b32((vuint32m1_t)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vmadc_vv_u32m2_b16(op0, op1, op2) \ +__builtin_rvv_vmadc_vv_u32m2_b16((vuint32m2_t)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vmadc_vv_u32m4_b8(op0, op1, op2) \ +__builtin_rvv_vmadc_vv_u32m4_b8((vuint32m4_t)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vmadc_vv_u32m8_b4(op0, op1, op2) \ +__builtin_rvv_vmadc_vv_u32m8_b4((vuint32m8_t)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vmadc_vv_u32mf2_b64(op0, op1, op2) \ +__builtin_rvv_vmadc_vv_u32mf2_b64((vuint32mf2_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vmadc_vv_u64m1_b64(op0, op1, op2) \ +__builtin_rvv_vmadc_vv_u64m1_b64((vuint64m1_t)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vmadc_vv_u64m2_b32(op0, op1, op2) \ +__builtin_rvv_vmadc_vv_u64m2_b32((vuint64m2_t)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vmadc_vv_u64m4_b16(op0, op1, op2) \ +__builtin_rvv_vmadc_vv_u64m4_b16((vuint64m4_t)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vmadc_vv_u64m8_b8(op0, op1, op2) \ +__builtin_rvv_vmadc_vv_u64m8_b8((vuint64m8_t)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vmadc_vx_u8m1_b8(op0, op1, op2) \ +__builtin_rvv_vmadc_vx_u8m1_b8((vuint8m1_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmadc_vx_u8m2_b4(op0, op1, op2) \ +__builtin_rvv_vmadc_vx_u8m2_b4((vuint8m2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmadc_vx_u8m4_b2(op0, op1, op2) \ +__builtin_rvv_vmadc_vx_u8m4_b2((vuint8m4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmadc_vx_u8m8_b1(op0, op1, op2) \ +__builtin_rvv_vmadc_vx_u8m8_b1((vuint8m8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmadc_vx_u8mf2_b16(op0, op1, op2) \ +__builtin_rvv_vmadc_vx_u8mf2_b16((vuint8mf2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmadc_vx_u8mf4_b32(op0, op1, op2) \ +__builtin_rvv_vmadc_vx_u8mf4_b32((vuint8mf4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmadc_vx_u8mf8_b64(op0, op1, op2) \ +__builtin_rvv_vmadc_vx_u8mf8_b64((vuint8mf8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmadc_vx_u16m1_b16(op0, op1, op2) \ +__builtin_rvv_vmadc_vx_u16m1_b16((vuint16m1_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmadc_vx_u16m2_b8(op0, op1, op2) \ +__builtin_rvv_vmadc_vx_u16m2_b8((vuint16m2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmadc_vx_u16m4_b4(op0, op1, op2) \ +__builtin_rvv_vmadc_vx_u16m4_b4((vuint16m4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmadc_vx_u16m8_b2(op0, op1, op2) \ +__builtin_rvv_vmadc_vx_u16m8_b2((vuint16m8_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmadc_vx_u16mf2_b32(op0, op1, op2) \ +__builtin_rvv_vmadc_vx_u16mf2_b32((vuint16mf2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmadc_vx_u16mf4_b64(op0, op1, op2) \ +__builtin_rvv_vmadc_vx_u16mf4_b64((vuint16mf4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmadc_vx_u32m1_b32(op0, op1, op2) \ +__builtin_rvv_vmadc_vx_u32m1_b32((vuint32m1_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmadc_vx_u32m2_b16(op0, op1, op2) \ +__builtin_rvv_vmadc_vx_u32m2_b16((vuint32m2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmadc_vx_u32m4_b8(op0, op1, op2) \ +__builtin_rvv_vmadc_vx_u32m4_b8((vuint32m4_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmadc_vx_u32m8_b4(op0, op1, op2) \ +__builtin_rvv_vmadc_vx_u32m8_b4((vuint32m8_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmadc_vx_u32mf2_b64(op0, op1, op2) \ +__builtin_rvv_vmadc_vx_u32mf2_b64((vuint32mf2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmadc_vx_u64m1_b64(op0, op1, op2) \ +__builtin_rvv_vmadc_vx_u64m1_b64((vuint64m1_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmadc_vx_u64m2_b32(op0, op1, op2) \ +__builtin_rvv_vmadc_vx_u64m2_b32((vuint64m2_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmadc_vx_u64m4_b16(op0, op1, op2) \ +__builtin_rvv_vmadc_vx_u64m4_b16((vuint64m4_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmadc_vx_u64m8_b8(op0, op1, op2) \ +__builtin_rvv_vmadc_vx_u64m8_b8((vuint64m8_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vsbc_vvm_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vvm_i8m1((vint8m1_t)(op0), (vint8m1_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vsbc_vvm_i8m2(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vvm_i8m2((vint8m2_t)(op0), (vint8m2_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vsbc_vvm_i8m4(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vvm_i8m4((vint8m4_t)(op0), (vint8m4_t)(op1), (vbool2_t)(op2), (size_t)(op3)) +#define vsbc_vvm_i8m8(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vvm_i8m8((vint8m8_t)(op0), (vint8m8_t)(op1), (vbool1_t)(op2), (size_t)(op3)) +#define vsbc_vvm_i8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vvm_i8mf2((vint8mf2_t)(op0), (vint8mf2_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vsbc_vvm_i8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vvm_i8mf4((vint8mf4_t)(op0), (vint8mf4_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vsbc_vvm_i8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vvm_i8mf8((vint8mf8_t)(op0), (vint8mf8_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vsbc_vvm_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vvm_i16m1((vint16m1_t)(op0), (vint16m1_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vsbc_vvm_i16m2(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vvm_i16m2((vint16m2_t)(op0), (vint16m2_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vsbc_vvm_i16m4(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vvm_i16m4((vint16m4_t)(op0), (vint16m4_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vsbc_vvm_i16m8(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vvm_i16m8((vint16m8_t)(op0), (vint16m8_t)(op1), (vbool2_t)(op2), (size_t)(op3)) +#define vsbc_vvm_i16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vvm_i16mf2((vint16mf2_t)(op0), (vint16mf2_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vsbc_vvm_i16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vvm_i16mf4((vint16mf4_t)(op0), (vint16mf4_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vsbc_vvm_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vvm_i32m1((vint32m1_t)(op0), (vint32m1_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vsbc_vvm_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vvm_i32m2((vint32m2_t)(op0), (vint32m2_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vsbc_vvm_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vvm_i32m4((vint32m4_t)(op0), (vint32m4_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vsbc_vvm_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vvm_i32m8((vint32m8_t)(op0), (vint32m8_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vsbc_vvm_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vvm_i32mf2((vint32mf2_t)(op0), (vint32mf2_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vsbc_vvm_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vvm_i64m1((vint64m1_t)(op0), (vint64m1_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vsbc_vvm_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vvm_i64m2((vint64m2_t)(op0), (vint64m2_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vsbc_vvm_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vvm_i64m4((vint64m4_t)(op0), (vint64m4_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vsbc_vvm_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vvm_i64m8((vint64m8_t)(op0), (vint64m8_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vsbc_vxm_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vxm_i8m1((vint8m1_t)(op0), (int8_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vsbc_vxm_i8m2(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vxm_i8m2((vint8m2_t)(op0), (int8_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vsbc_vxm_i8m4(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vxm_i8m4((vint8m4_t)(op0), (int8_t)(op1), (vbool2_t)(op2), (size_t)(op3)) +#define vsbc_vxm_i8m8(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vxm_i8m8((vint8m8_t)(op0), (int8_t)(op1), (vbool1_t)(op2), (size_t)(op3)) +#define vsbc_vxm_i8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vxm_i8mf2((vint8mf2_t)(op0), (int8_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vsbc_vxm_i8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vxm_i8mf4((vint8mf4_t)(op0), (int8_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vsbc_vxm_i8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vxm_i8mf8((vint8mf8_t)(op0), (int8_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vsbc_vxm_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vxm_i16m1((vint16m1_t)(op0), (int16_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vsbc_vxm_i16m2(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vxm_i16m2((vint16m2_t)(op0), (int16_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vsbc_vxm_i16m4(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vxm_i16m4((vint16m4_t)(op0), (int16_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vsbc_vxm_i16m8(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vxm_i16m8((vint16m8_t)(op0), (int16_t)(op1), (vbool2_t)(op2), (size_t)(op3)) +#define vsbc_vxm_i16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vxm_i16mf2((vint16mf2_t)(op0), (int16_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vsbc_vxm_i16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vxm_i16mf4((vint16mf4_t)(op0), (int16_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vsbc_vxm_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vxm_i32m1((vint32m1_t)(op0), (int32_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vsbc_vxm_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vxm_i32m2((vint32m2_t)(op0), (int32_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vsbc_vxm_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vxm_i32m4((vint32m4_t)(op0), (int32_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vsbc_vxm_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vxm_i32m8((vint32m8_t)(op0), (int32_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vsbc_vxm_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vxm_i32mf2((vint32mf2_t)(op0), (int32_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vsbc_vxm_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vxm_i64m1((vint64m1_t)(op0), (int64_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vsbc_vxm_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vxm_i64m2((vint64m2_t)(op0), (int64_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vsbc_vxm_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vxm_i64m4((vint64m4_t)(op0), (int64_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vsbc_vxm_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vxm_i64m8((vint64m8_t)(op0), (int64_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vsbc_vvm_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vvm_u8m1((vuint8m1_t)(op0), (vuint8m1_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vsbc_vvm_u8m2(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vvm_u8m2((vuint8m2_t)(op0), (vuint8m2_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vsbc_vvm_u8m4(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vvm_u8m4((vuint8m4_t)(op0), (vuint8m4_t)(op1), (vbool2_t)(op2), (size_t)(op3)) +#define vsbc_vvm_u8m8(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vvm_u8m8((vuint8m8_t)(op0), (vuint8m8_t)(op1), (vbool1_t)(op2), (size_t)(op3)) +#define vsbc_vvm_u8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vvm_u8mf2((vuint8mf2_t)(op0), (vuint8mf2_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vsbc_vvm_u8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vvm_u8mf4((vuint8mf4_t)(op0), (vuint8mf4_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vsbc_vvm_u8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vvm_u8mf8((vuint8mf8_t)(op0), (vuint8mf8_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vsbc_vvm_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vvm_u16m1((vuint16m1_t)(op0), (vuint16m1_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vsbc_vvm_u16m2(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vvm_u16m2((vuint16m2_t)(op0), (vuint16m2_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vsbc_vvm_u16m4(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vvm_u16m4((vuint16m4_t)(op0), (vuint16m4_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vsbc_vvm_u16m8(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vvm_u16m8((vuint16m8_t)(op0), (vuint16m8_t)(op1), (vbool2_t)(op2), (size_t)(op3)) +#define vsbc_vvm_u16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vvm_u16mf2((vuint16mf2_t)(op0), (vuint16mf2_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vsbc_vvm_u16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vvm_u16mf4((vuint16mf4_t)(op0), (vuint16mf4_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vsbc_vvm_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vvm_u32m1((vuint32m1_t)(op0), (vuint32m1_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vsbc_vvm_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vvm_u32m2((vuint32m2_t)(op0), (vuint32m2_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vsbc_vvm_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vvm_u32m4((vuint32m4_t)(op0), (vuint32m4_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vsbc_vvm_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vvm_u32m8((vuint32m8_t)(op0), (vuint32m8_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vsbc_vvm_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vvm_u32mf2((vuint32mf2_t)(op0), (vuint32mf2_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vsbc_vvm_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vvm_u64m1((vuint64m1_t)(op0), (vuint64m1_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vsbc_vvm_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vvm_u64m2((vuint64m2_t)(op0), (vuint64m2_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vsbc_vvm_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vvm_u64m4((vuint64m4_t)(op0), (vuint64m4_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vsbc_vvm_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vvm_u64m8((vuint64m8_t)(op0), (vuint64m8_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vsbc_vxm_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vxm_u8m1((vuint8m1_t)(op0), (uint8_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vsbc_vxm_u8m2(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vxm_u8m2((vuint8m2_t)(op0), (uint8_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vsbc_vxm_u8m4(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vxm_u8m4((vuint8m4_t)(op0), (uint8_t)(op1), (vbool2_t)(op2), (size_t)(op3)) +#define vsbc_vxm_u8m8(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vxm_u8m8((vuint8m8_t)(op0), (uint8_t)(op1), (vbool1_t)(op2), (size_t)(op3)) +#define vsbc_vxm_u8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vxm_u8mf2((vuint8mf2_t)(op0), (uint8_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vsbc_vxm_u8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vxm_u8mf4((vuint8mf4_t)(op0), (uint8_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vsbc_vxm_u8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vxm_u8mf8((vuint8mf8_t)(op0), (uint8_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vsbc_vxm_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vxm_u16m1((vuint16m1_t)(op0), (uint16_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vsbc_vxm_u16m2(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vxm_u16m2((vuint16m2_t)(op0), (uint16_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vsbc_vxm_u16m4(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vxm_u16m4((vuint16m4_t)(op0), (uint16_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vsbc_vxm_u16m8(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vxm_u16m8((vuint16m8_t)(op0), (uint16_t)(op1), (vbool2_t)(op2), (size_t)(op3)) +#define vsbc_vxm_u16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vxm_u16mf2((vuint16mf2_t)(op0), (uint16_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vsbc_vxm_u16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vxm_u16mf4((vuint16mf4_t)(op0), (uint16_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vsbc_vxm_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vxm_u32m1((vuint32m1_t)(op0), (uint32_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vsbc_vxm_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vxm_u32m2((vuint32m2_t)(op0), (uint32_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vsbc_vxm_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vxm_u32m4((vuint32m4_t)(op0), (uint32_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vsbc_vxm_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vxm_u32m8((vuint32m8_t)(op0), (uint32_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vsbc_vxm_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vxm_u32mf2((vuint32mf2_t)(op0), (uint32_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vsbc_vxm_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vxm_u64m1((vuint64m1_t)(op0), (uint64_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vsbc_vxm_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vxm_u64m2((vuint64m2_t)(op0), (uint64_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vsbc_vxm_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vxm_u64m4((vuint64m4_t)(op0), (uint64_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vsbc_vxm_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vsbc_vxm_u64m8((vuint64m8_t)(op0), (uint64_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vmsbc_vvm_i8m1_b8(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vvm_i8m1_b8((vint8m1_t)(op0), (vint8m1_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vmsbc_vvm_i8m2_b4(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vvm_i8m2_b4((vint8m2_t)(op0), (vint8m2_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vmsbc_vvm_i8m4_b2(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vvm_i8m4_b2((vint8m4_t)(op0), (vint8m4_t)(op1), (vbool2_t)(op2), (size_t)(op3)) +#define vmsbc_vvm_i8m8_b1(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vvm_i8m8_b1((vint8m8_t)(op0), (vint8m8_t)(op1), (vbool1_t)(op2), (size_t)(op3)) +#define vmsbc_vvm_i8mf2_b16(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vvm_i8mf2_b16((vint8mf2_t)(op0), (vint8mf2_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vmsbc_vvm_i8mf4_b32(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vvm_i8mf4_b32((vint8mf4_t)(op0), (vint8mf4_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vmsbc_vvm_i8mf8_b64(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vvm_i8mf8_b64((vint8mf8_t)(op0), (vint8mf8_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vmsbc_vvm_i16m1_b16(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vvm_i16m1_b16((vint16m1_t)(op0), (vint16m1_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vmsbc_vvm_i16m2_b8(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vvm_i16m2_b8((vint16m2_t)(op0), (vint16m2_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vmsbc_vvm_i16m4_b4(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vvm_i16m4_b4((vint16m4_t)(op0), (vint16m4_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vmsbc_vvm_i16m8_b2(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vvm_i16m8_b2((vint16m8_t)(op0), (vint16m8_t)(op1), (vbool2_t)(op2), (size_t)(op3)) +#define vmsbc_vvm_i16mf2_b32(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vvm_i16mf2_b32((vint16mf2_t)(op0), (vint16mf2_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vmsbc_vvm_i16mf4_b64(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vvm_i16mf4_b64((vint16mf4_t)(op0), (vint16mf4_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vmsbc_vvm_i32m1_b32(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vvm_i32m1_b32((vint32m1_t)(op0), (vint32m1_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vmsbc_vvm_i32m2_b16(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vvm_i32m2_b16((vint32m2_t)(op0), (vint32m2_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vmsbc_vvm_i32m4_b8(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vvm_i32m4_b8((vint32m4_t)(op0), (vint32m4_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vmsbc_vvm_i32m8_b4(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vvm_i32m8_b4((vint32m8_t)(op0), (vint32m8_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vmsbc_vvm_i32mf2_b64(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vvm_i32mf2_b64((vint32mf2_t)(op0), (vint32mf2_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vmsbc_vvm_i64m1_b64(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vvm_i64m1_b64((vint64m1_t)(op0), (vint64m1_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vmsbc_vvm_i64m2_b32(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vvm_i64m2_b32((vint64m2_t)(op0), (vint64m2_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vmsbc_vvm_i64m4_b16(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vvm_i64m4_b16((vint64m4_t)(op0), (vint64m4_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vmsbc_vvm_i64m8_b8(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vvm_i64m8_b8((vint64m8_t)(op0), (vint64m8_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vmsbc_vxm_i8m1_b8(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vxm_i8m1_b8((vint8m1_t)(op0), (int8_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vmsbc_vxm_i8m2_b4(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vxm_i8m2_b4((vint8m2_t)(op0), (int8_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vmsbc_vxm_i8m4_b2(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vxm_i8m4_b2((vint8m4_t)(op0), (int8_t)(op1), (vbool2_t)(op2), (size_t)(op3)) +#define vmsbc_vxm_i8m8_b1(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vxm_i8m8_b1((vint8m8_t)(op0), (int8_t)(op1), (vbool1_t)(op2), (size_t)(op3)) +#define vmsbc_vxm_i8mf2_b16(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vxm_i8mf2_b16((vint8mf2_t)(op0), (int8_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vmsbc_vxm_i8mf4_b32(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vxm_i8mf4_b32((vint8mf4_t)(op0), (int8_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vmsbc_vxm_i8mf8_b64(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vxm_i8mf8_b64((vint8mf8_t)(op0), (int8_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vmsbc_vxm_i16m1_b16(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vxm_i16m1_b16((vint16m1_t)(op0), (int16_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vmsbc_vxm_i16m2_b8(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vxm_i16m2_b8((vint16m2_t)(op0), (int16_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vmsbc_vxm_i16m4_b4(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vxm_i16m4_b4((vint16m4_t)(op0), (int16_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vmsbc_vxm_i16m8_b2(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vxm_i16m8_b2((vint16m8_t)(op0), (int16_t)(op1), (vbool2_t)(op2), (size_t)(op3)) +#define vmsbc_vxm_i16mf2_b32(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vxm_i16mf2_b32((vint16mf2_t)(op0), (int16_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vmsbc_vxm_i16mf4_b64(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vxm_i16mf4_b64((vint16mf4_t)(op0), (int16_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vmsbc_vxm_i32m1_b32(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vxm_i32m1_b32((vint32m1_t)(op0), (int32_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vmsbc_vxm_i32m2_b16(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vxm_i32m2_b16((vint32m2_t)(op0), (int32_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vmsbc_vxm_i32m4_b8(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vxm_i32m4_b8((vint32m4_t)(op0), (int32_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vmsbc_vxm_i32m8_b4(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vxm_i32m8_b4((vint32m8_t)(op0), (int32_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vmsbc_vxm_i32mf2_b64(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vxm_i32mf2_b64((vint32mf2_t)(op0), (int32_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vmsbc_vxm_i64m1_b64(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vxm_i64m1_b64((vint64m1_t)(op0), (int64_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vmsbc_vxm_i64m2_b32(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vxm_i64m2_b32((vint64m2_t)(op0), (int64_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vmsbc_vxm_i64m4_b16(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vxm_i64m4_b16((vint64m4_t)(op0), (int64_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vmsbc_vxm_i64m8_b8(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vxm_i64m8_b8((vint64m8_t)(op0), (int64_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vmsbc_vvm_u8m1_b8(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vvm_u8m1_b8((vuint8m1_t)(op0), (vuint8m1_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vmsbc_vvm_u8m2_b4(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vvm_u8m2_b4((vuint8m2_t)(op0), (vuint8m2_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vmsbc_vvm_u8m4_b2(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vvm_u8m4_b2((vuint8m4_t)(op0), (vuint8m4_t)(op1), (vbool2_t)(op2), (size_t)(op3)) +#define vmsbc_vvm_u8m8_b1(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vvm_u8m8_b1((vuint8m8_t)(op0), (vuint8m8_t)(op1), (vbool1_t)(op2), (size_t)(op3)) +#define vmsbc_vvm_u8mf2_b16(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vvm_u8mf2_b16((vuint8mf2_t)(op0), (vuint8mf2_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vmsbc_vvm_u8mf4_b32(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vvm_u8mf4_b32((vuint8mf4_t)(op0), (vuint8mf4_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vmsbc_vvm_u8mf8_b64(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vvm_u8mf8_b64((vuint8mf8_t)(op0), (vuint8mf8_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vmsbc_vvm_u16m1_b16(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vvm_u16m1_b16((vuint16m1_t)(op0), (vuint16m1_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vmsbc_vvm_u16m2_b8(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vvm_u16m2_b8((vuint16m2_t)(op0), (vuint16m2_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vmsbc_vvm_u16m4_b4(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vvm_u16m4_b4((vuint16m4_t)(op0), (vuint16m4_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vmsbc_vvm_u16m8_b2(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vvm_u16m8_b2((vuint16m8_t)(op0), (vuint16m8_t)(op1), (vbool2_t)(op2), (size_t)(op3)) +#define vmsbc_vvm_u16mf2_b32(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vvm_u16mf2_b32((vuint16mf2_t)(op0), (vuint16mf2_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vmsbc_vvm_u16mf4_b64(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vvm_u16mf4_b64((vuint16mf4_t)(op0), (vuint16mf4_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vmsbc_vvm_u32m1_b32(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vvm_u32m1_b32((vuint32m1_t)(op0), (vuint32m1_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vmsbc_vvm_u32m2_b16(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vvm_u32m2_b16((vuint32m2_t)(op0), (vuint32m2_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vmsbc_vvm_u32m4_b8(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vvm_u32m4_b8((vuint32m4_t)(op0), (vuint32m4_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vmsbc_vvm_u32m8_b4(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vvm_u32m8_b4((vuint32m8_t)(op0), (vuint32m8_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vmsbc_vvm_u32mf2_b64(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vvm_u32mf2_b64((vuint32mf2_t)(op0), (vuint32mf2_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vmsbc_vvm_u64m1_b64(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vvm_u64m1_b64((vuint64m1_t)(op0), (vuint64m1_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vmsbc_vvm_u64m2_b32(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vvm_u64m2_b32((vuint64m2_t)(op0), (vuint64m2_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vmsbc_vvm_u64m4_b16(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vvm_u64m4_b16((vuint64m4_t)(op0), (vuint64m4_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vmsbc_vvm_u64m8_b8(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vvm_u64m8_b8((vuint64m8_t)(op0), (vuint64m8_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vmsbc_vxm_u8m1_b8(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vxm_u8m1_b8((vuint8m1_t)(op0), (uint8_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vmsbc_vxm_u8m2_b4(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vxm_u8m2_b4((vuint8m2_t)(op0), (uint8_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vmsbc_vxm_u8m4_b2(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vxm_u8m4_b2((vuint8m4_t)(op0), (uint8_t)(op1), (vbool2_t)(op2), (size_t)(op3)) +#define vmsbc_vxm_u8m8_b1(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vxm_u8m8_b1((vuint8m8_t)(op0), (uint8_t)(op1), (vbool1_t)(op2), (size_t)(op3)) +#define vmsbc_vxm_u8mf2_b16(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vxm_u8mf2_b16((vuint8mf2_t)(op0), (uint8_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vmsbc_vxm_u8mf4_b32(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vxm_u8mf4_b32((vuint8mf4_t)(op0), (uint8_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vmsbc_vxm_u8mf8_b64(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vxm_u8mf8_b64((vuint8mf8_t)(op0), (uint8_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vmsbc_vxm_u16m1_b16(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vxm_u16m1_b16((vuint16m1_t)(op0), (uint16_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vmsbc_vxm_u16m2_b8(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vxm_u16m2_b8((vuint16m2_t)(op0), (uint16_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vmsbc_vxm_u16m4_b4(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vxm_u16m4_b4((vuint16m4_t)(op0), (uint16_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vmsbc_vxm_u16m8_b2(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vxm_u16m8_b2((vuint16m8_t)(op0), (uint16_t)(op1), (vbool2_t)(op2), (size_t)(op3)) +#define vmsbc_vxm_u16mf2_b32(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vxm_u16mf2_b32((vuint16mf2_t)(op0), (uint16_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vmsbc_vxm_u16mf4_b64(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vxm_u16mf4_b64((vuint16mf4_t)(op0), (uint16_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vmsbc_vxm_u32m1_b32(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vxm_u32m1_b32((vuint32m1_t)(op0), (uint32_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vmsbc_vxm_u32m2_b16(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vxm_u32m2_b16((vuint32m2_t)(op0), (uint32_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vmsbc_vxm_u32m4_b8(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vxm_u32m4_b8((vuint32m4_t)(op0), (uint32_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vmsbc_vxm_u32m8_b4(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vxm_u32m8_b4((vuint32m8_t)(op0), (uint32_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vmsbc_vxm_u32mf2_b64(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vxm_u32mf2_b64((vuint32mf2_t)(op0), (uint32_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vmsbc_vxm_u64m1_b64(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vxm_u64m1_b64((vuint64m1_t)(op0), (uint64_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vmsbc_vxm_u64m2_b32(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vxm_u64m2_b32((vuint64m2_t)(op0), (uint64_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vmsbc_vxm_u64m4_b16(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vxm_u64m4_b16((vuint64m4_t)(op0), (uint64_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vmsbc_vxm_u64m8_b8(op0, op1, op2, op3) \ +__builtin_rvv_vmsbc_vxm_u64m8_b8((vuint64m8_t)(op0), (uint64_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vsuxei8_v_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_u16m1((uint16_t *)(op0), (vuint8mf2_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vsuxei8_v_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint8mf2_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vsuxei8_v_u16m2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_u16m2((uint16_t *)(op0), (vuint8m1_t)(op1), (vuint16m2_t)(op2), (size_t)(op3)) +#define vsuxei8_v_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_u16m2_m((vbool8_t)(op0), (uint16_t *)(op1), (vuint8m1_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vsuxei8_v_u16m4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_u16m4((uint16_t *)(op0), (vuint8m2_t)(op1), (vuint16m4_t)(op2), (size_t)(op3)) +#define vsuxei8_v_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_u16m4_m((vbool4_t)(op0), (uint16_t *)(op1), (vuint8m2_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vsuxei8_v_u16m8(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_u16m8((uint16_t *)(op0), (vuint8m4_t)(op1), (vuint16m8_t)(op2), (size_t)(op3)) +#define vsuxei8_v_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_u16m8_m((vbool2_t)(op0), (uint16_t *)(op1), (vuint8m4_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vsuxei8_v_u16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_u16mf2((uint16_t *)(op0), (vuint8mf4_t)(op1), (vuint16mf2_t)(op2), (size_t)(op3)) +#define vsuxei8_v_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint8mf4_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vsuxei8_v_u16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_u16mf4((uint16_t *)(op0), (vuint8mf8_t)(op1), (vuint16mf4_t)(op2), (size_t)(op3)) +#define vsuxei8_v_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint8mf8_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vmsbc_vv_i8m1_b8(op0, op1, op2) \ +__builtin_rvv_vmsbc_vv_i8m1_b8((vint8m1_t)(op0), (vint8m1_t)(op1), (size_t)(op2)) +#define vmsbc_vv_i8m2_b4(op0, op1, op2) \ +__builtin_rvv_vmsbc_vv_i8m2_b4((vint8m2_t)(op0), (vint8m2_t)(op1), (size_t)(op2)) +#define vmsbc_vv_i8m4_b2(op0, op1, op2) \ +__builtin_rvv_vmsbc_vv_i8m4_b2((vint8m4_t)(op0), (vint8m4_t)(op1), (size_t)(op2)) +#define vmsbc_vv_i8m8_b1(op0, op1, op2) \ +__builtin_rvv_vmsbc_vv_i8m8_b1((vint8m8_t)(op0), (vint8m8_t)(op1), (size_t)(op2)) +#define vmsbc_vv_i8mf2_b16(op0, op1, op2) \ +__builtin_rvv_vmsbc_vv_i8mf2_b16((vint8mf2_t)(op0), (vint8mf2_t)(op1), (size_t)(op2)) +#define vmsbc_vv_i8mf4_b32(op0, op1, op2) \ +__builtin_rvv_vmsbc_vv_i8mf4_b32((vint8mf4_t)(op0), (vint8mf4_t)(op1), (size_t)(op2)) +#define vmsbc_vv_i8mf8_b64(op0, op1, op2) \ +__builtin_rvv_vmsbc_vv_i8mf8_b64((vint8mf8_t)(op0), (vint8mf8_t)(op1), (size_t)(op2)) +#define vmsbc_vv_i16m1_b16(op0, op1, op2) \ +__builtin_rvv_vmsbc_vv_i16m1_b16((vint16m1_t)(op0), (vint16m1_t)(op1), (size_t)(op2)) +#define vmsbc_vv_i16m2_b8(op0, op1, op2) \ +__builtin_rvv_vmsbc_vv_i16m2_b8((vint16m2_t)(op0), (vint16m2_t)(op1), (size_t)(op2)) +#define vmsbc_vv_i16m4_b4(op0, op1, op2) \ +__builtin_rvv_vmsbc_vv_i16m4_b4((vint16m4_t)(op0), (vint16m4_t)(op1), (size_t)(op2)) +#define vmsbc_vv_i16m8_b2(op0, op1, op2) \ +__builtin_rvv_vmsbc_vv_i16m8_b2((vint16m8_t)(op0), (vint16m8_t)(op1), (size_t)(op2)) +#define vmsbc_vv_i16mf2_b32(op0, op1, op2) \ +__builtin_rvv_vmsbc_vv_i16mf2_b32((vint16mf2_t)(op0), (vint16mf2_t)(op1), (size_t)(op2)) +#define vmsbc_vv_i16mf4_b64(op0, op1, op2) \ +__builtin_rvv_vmsbc_vv_i16mf4_b64((vint16mf4_t)(op0), (vint16mf4_t)(op1), (size_t)(op2)) +#define vmsbc_vv_i32m1_b32(op0, op1, op2) \ +__builtin_rvv_vmsbc_vv_i32m1_b32((vint32m1_t)(op0), (vint32m1_t)(op1), (size_t)(op2)) +#define vmsbc_vv_i32m2_b16(op0, op1, op2) \ +__builtin_rvv_vmsbc_vv_i32m2_b16((vint32m2_t)(op0), (vint32m2_t)(op1), (size_t)(op2)) +#define vmsbc_vv_i32m4_b8(op0, op1, op2) \ +__builtin_rvv_vmsbc_vv_i32m4_b8((vint32m4_t)(op0), (vint32m4_t)(op1), (size_t)(op2)) +#define vmsbc_vv_i32m8_b4(op0, op1, op2) \ +__builtin_rvv_vmsbc_vv_i32m8_b4((vint32m8_t)(op0), (vint32m8_t)(op1), (size_t)(op2)) +#define vmsbc_vv_i32mf2_b64(op0, op1, op2) \ +__builtin_rvv_vmsbc_vv_i32mf2_b64((vint32mf2_t)(op0), (vint32mf2_t)(op1), (size_t)(op2)) +#define vmsbc_vv_i64m1_b64(op0, op1, op2) \ +__builtin_rvv_vmsbc_vv_i64m1_b64((vint64m1_t)(op0), (vint64m1_t)(op1), (size_t)(op2)) +#define vmsbc_vv_i64m2_b32(op0, op1, op2) \ +__builtin_rvv_vmsbc_vv_i64m2_b32((vint64m2_t)(op0), (vint64m2_t)(op1), (size_t)(op2)) +#define vmsbc_vv_i64m4_b16(op0, op1, op2) \ +__builtin_rvv_vmsbc_vv_i64m4_b16((vint64m4_t)(op0), (vint64m4_t)(op1), (size_t)(op2)) +#define vmsbc_vv_i64m8_b8(op0, op1, op2) \ +__builtin_rvv_vmsbc_vv_i64m8_b8((vint64m8_t)(op0), (vint64m8_t)(op1), (size_t)(op2)) +#define vmsbc_vx_i8m1_b8(op0, op1, op2) \ +__builtin_rvv_vmsbc_vx_i8m1_b8((vint8m1_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmsbc_vx_i8m2_b4(op0, op1, op2) \ +__builtin_rvv_vmsbc_vx_i8m2_b4((vint8m2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmsbc_vx_i8m4_b2(op0, op1, op2) \ +__builtin_rvv_vmsbc_vx_i8m4_b2((vint8m4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmsbc_vx_i8m8_b1(op0, op1, op2) \ +__builtin_rvv_vmsbc_vx_i8m8_b1((vint8m8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmsbc_vx_i8mf2_b16(op0, op1, op2) \ +__builtin_rvv_vmsbc_vx_i8mf2_b16((vint8mf2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmsbc_vx_i8mf4_b32(op0, op1, op2) \ +__builtin_rvv_vmsbc_vx_i8mf4_b32((vint8mf4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmsbc_vx_i8mf8_b64(op0, op1, op2) \ +__builtin_rvv_vmsbc_vx_i8mf8_b64((vint8mf8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmsbc_vx_i16m1_b16(op0, op1, op2) \ +__builtin_rvv_vmsbc_vx_i16m1_b16((vint16m1_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmsbc_vx_i16m2_b8(op0, op1, op2) \ +__builtin_rvv_vmsbc_vx_i16m2_b8((vint16m2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmsbc_vx_i16m4_b4(op0, op1, op2) \ +__builtin_rvv_vmsbc_vx_i16m4_b4((vint16m4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmsbc_vx_i16m8_b2(op0, op1, op2) \ +__builtin_rvv_vmsbc_vx_i16m8_b2((vint16m8_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmsbc_vx_i16mf2_b32(op0, op1, op2) \ +__builtin_rvv_vmsbc_vx_i16mf2_b32((vint16mf2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmsbc_vx_i16mf4_b64(op0, op1, op2) \ +__builtin_rvv_vmsbc_vx_i16mf4_b64((vint16mf4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmsbc_vx_i32m1_b32(op0, op1, op2) \ +__builtin_rvv_vmsbc_vx_i32m1_b32((vint32m1_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmsbc_vx_i32m2_b16(op0, op1, op2) \ +__builtin_rvv_vmsbc_vx_i32m2_b16((vint32m2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmsbc_vx_i32m4_b8(op0, op1, op2) \ +__builtin_rvv_vmsbc_vx_i32m4_b8((vint32m4_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmsbc_vx_i32m8_b4(op0, op1, op2) \ +__builtin_rvv_vmsbc_vx_i32m8_b4((vint32m8_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmsbc_vx_i32mf2_b64(op0, op1, op2) \ +__builtin_rvv_vmsbc_vx_i32mf2_b64((vint32mf2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmsbc_vx_i64m1_b64(op0, op1, op2) \ +__builtin_rvv_vmsbc_vx_i64m1_b64((vint64m1_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmsbc_vx_i64m2_b32(op0, op1, op2) \ +__builtin_rvv_vmsbc_vx_i64m2_b32((vint64m2_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmsbc_vx_i64m4_b16(op0, op1, op2) \ +__builtin_rvv_vmsbc_vx_i64m4_b16((vint64m4_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmsbc_vx_i64m8_b8(op0, op1, op2) \ +__builtin_rvv_vmsbc_vx_i64m8_b8((vint64m8_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmsbc_vv_u8m1_b8(op0, op1, op2) \ +__builtin_rvv_vmsbc_vv_u8m1_b8((vuint8m1_t)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vmsbc_vv_u8m2_b4(op0, op1, op2) \ +__builtin_rvv_vmsbc_vv_u8m2_b4((vuint8m2_t)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vmsbc_vv_u8m4_b2(op0, op1, op2) \ +__builtin_rvv_vmsbc_vv_u8m4_b2((vuint8m4_t)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vmsbc_vv_u8m8_b1(op0, op1, op2) \ +__builtin_rvv_vmsbc_vv_u8m8_b1((vuint8m8_t)(op0), (vuint8m8_t)(op1), (size_t)(op2)) +#define vmsbc_vv_u8mf2_b16(op0, op1, op2) \ +__builtin_rvv_vmsbc_vv_u8mf2_b16((vuint8mf2_t)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vmsbc_vv_u8mf4_b32(op0, op1, op2) \ +__builtin_rvv_vmsbc_vv_u8mf4_b32((vuint8mf4_t)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vmsbc_vv_u8mf8_b64(op0, op1, op2) \ +__builtin_rvv_vmsbc_vv_u8mf8_b64((vuint8mf8_t)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vmsbc_vv_u16m1_b16(op0, op1, op2) \ +__builtin_rvv_vmsbc_vv_u16m1_b16((vuint16m1_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vmsbc_vv_u16m2_b8(op0, op1, op2) \ +__builtin_rvv_vmsbc_vv_u16m2_b8((vuint16m2_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vmsbc_vv_u16m4_b4(op0, op1, op2) \ +__builtin_rvv_vmsbc_vv_u16m4_b4((vuint16m4_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vmsbc_vv_u16m8_b2(op0, op1, op2) \ +__builtin_rvv_vmsbc_vv_u16m8_b2((vuint16m8_t)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vmsbc_vv_u16mf2_b32(op0, op1, op2) \ +__builtin_rvv_vmsbc_vv_u16mf2_b32((vuint16mf2_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vmsbc_vv_u16mf4_b64(op0, op1, op2) \ +__builtin_rvv_vmsbc_vv_u16mf4_b64((vuint16mf4_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vmsbc_vv_u32m1_b32(op0, op1, op2) \ +__builtin_rvv_vmsbc_vv_u32m1_b32((vuint32m1_t)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vmsbc_vv_u32m2_b16(op0, op1, op2) \ +__builtin_rvv_vmsbc_vv_u32m2_b16((vuint32m2_t)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vmsbc_vv_u32m4_b8(op0, op1, op2) \ +__builtin_rvv_vmsbc_vv_u32m4_b8((vuint32m4_t)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vmsbc_vv_u32m8_b4(op0, op1, op2) \ +__builtin_rvv_vmsbc_vv_u32m8_b4((vuint32m8_t)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vmsbc_vv_u32mf2_b64(op0, op1, op2) \ +__builtin_rvv_vmsbc_vv_u32mf2_b64((vuint32mf2_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vmsbc_vv_u64m1_b64(op0, op1, op2) \ +__builtin_rvv_vmsbc_vv_u64m1_b64((vuint64m1_t)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vmsbc_vv_u64m2_b32(op0, op1, op2) \ +__builtin_rvv_vmsbc_vv_u64m2_b32((vuint64m2_t)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vmsbc_vv_u64m4_b16(op0, op1, op2) \ +__builtin_rvv_vmsbc_vv_u64m4_b16((vuint64m4_t)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vmsbc_vv_u64m8_b8(op0, op1, op2) \ +__builtin_rvv_vmsbc_vv_u64m8_b8((vuint64m8_t)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vmsbc_vx_u8m1_b8(op0, op1, op2) \ +__builtin_rvv_vmsbc_vx_u8m1_b8((vuint8m1_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmsbc_vx_u8m2_b4(op0, op1, op2) \ +__builtin_rvv_vmsbc_vx_u8m2_b4((vuint8m2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmsbc_vx_u8m4_b2(op0, op1, op2) \ +__builtin_rvv_vmsbc_vx_u8m4_b2((vuint8m4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmsbc_vx_u8m8_b1(op0, op1, op2) \ +__builtin_rvv_vmsbc_vx_u8m8_b1((vuint8m8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmsbc_vx_u8mf2_b16(op0, op1, op2) \ +__builtin_rvv_vmsbc_vx_u8mf2_b16((vuint8mf2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmsbc_vx_u8mf4_b32(op0, op1, op2) \ +__builtin_rvv_vmsbc_vx_u8mf4_b32((vuint8mf4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmsbc_vx_u8mf8_b64(op0, op1, op2) \ +__builtin_rvv_vmsbc_vx_u8mf8_b64((vuint8mf8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmsbc_vx_u16m1_b16(op0, op1, op2) \ +__builtin_rvv_vmsbc_vx_u16m1_b16((vuint16m1_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmsbc_vx_u16m2_b8(op0, op1, op2) \ +__builtin_rvv_vmsbc_vx_u16m2_b8((vuint16m2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmsbc_vx_u16m4_b4(op0, op1, op2) \ +__builtin_rvv_vmsbc_vx_u16m4_b4((vuint16m4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmsbc_vx_u16m8_b2(op0, op1, op2) \ +__builtin_rvv_vmsbc_vx_u16m8_b2((vuint16m8_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmsbc_vx_u16mf2_b32(op0, op1, op2) \ +__builtin_rvv_vmsbc_vx_u16mf2_b32((vuint16mf2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmsbc_vx_u16mf4_b64(op0, op1, op2) \ +__builtin_rvv_vmsbc_vx_u16mf4_b64((vuint16mf4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmsbc_vx_u32m1_b32(op0, op1, op2) \ +__builtin_rvv_vmsbc_vx_u32m1_b32((vuint32m1_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmsbc_vx_u32m2_b16(op0, op1, op2) \ +__builtin_rvv_vmsbc_vx_u32m2_b16((vuint32m2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmsbc_vx_u32m4_b8(op0, op1, op2) \ +__builtin_rvv_vmsbc_vx_u32m4_b8((vuint32m4_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmsbc_vx_u32m8_b4(op0, op1, op2) \ +__builtin_rvv_vmsbc_vx_u32m8_b4((vuint32m8_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmsbc_vx_u32mf2_b64(op0, op1, op2) \ +__builtin_rvv_vmsbc_vx_u32mf2_b64((vuint32mf2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmsbc_vx_u64m1_b64(op0, op1, op2) \ +__builtin_rvv_vmsbc_vx_u64m1_b64((vuint64m1_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmsbc_vx_u64m2_b32(op0, op1, op2) \ +__builtin_rvv_vmsbc_vx_u64m2_b32((vuint64m2_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmsbc_vx_u64m4_b16(op0, op1, op2) \ +__builtin_rvv_vmsbc_vx_u64m4_b16((vuint64m4_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmsbc_vx_u64m8_b8(op0, op1, op2) \ +__builtin_rvv_vmsbc_vx_u64m8_b8((vuint64m8_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vand_vv_i8m1(op0, op1, op2) \ +__builtin_rvv_vand_vv_i8m1((vint8m1_t)(op0), (vint8m1_t)(op1), (size_t)(op2)) +#define vand_vv_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vv_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vand_vv_i8m2(op0, op1, op2) \ +__builtin_rvv_vand_vv_i8m2((vint8m2_t)(op0), (vint8m2_t)(op1), (size_t)(op2)) +#define vand_vv_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vv_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vand_vv_i8m4(op0, op1, op2) \ +__builtin_rvv_vand_vv_i8m4((vint8m4_t)(op0), (vint8m4_t)(op1), (size_t)(op2)) +#define vand_vv_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vv_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vand_vv_i8m8(op0, op1, op2) \ +__builtin_rvv_vand_vv_i8m8((vint8m8_t)(op0), (vint8m8_t)(op1), (size_t)(op2)) +#define vand_vv_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vv_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (vint8m8_t)(op3), (size_t)(op4)) +#define vand_vv_i8mf2(op0, op1, op2) \ +__builtin_rvv_vand_vv_i8mf2((vint8mf2_t)(op0), (vint8mf2_t)(op1), (size_t)(op2)) +#define vand_vv_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vv_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vand_vv_i8mf4(op0, op1, op2) \ +__builtin_rvv_vand_vv_i8mf4((vint8mf4_t)(op0), (vint8mf4_t)(op1), (size_t)(op2)) +#define vand_vv_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vv_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vand_vv_i8mf8(op0, op1, op2) \ +__builtin_rvv_vand_vv_i8mf8((vint8mf8_t)(op0), (vint8mf8_t)(op1), (size_t)(op2)) +#define vand_vv_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vv_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vand_vv_i16m1(op0, op1, op2) \ +__builtin_rvv_vand_vv_i16m1((vint16m1_t)(op0), (vint16m1_t)(op1), (size_t)(op2)) +#define vand_vv_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vv_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vand_vv_i16m2(op0, op1, op2) \ +__builtin_rvv_vand_vv_i16m2((vint16m2_t)(op0), (vint16m2_t)(op1), (size_t)(op2)) +#define vand_vv_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vv_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vand_vv_i16m4(op0, op1, op2) \ +__builtin_rvv_vand_vv_i16m4((vint16m4_t)(op0), (vint16m4_t)(op1), (size_t)(op2)) +#define vand_vv_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vv_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vand_vv_i16m8(op0, op1, op2) \ +__builtin_rvv_vand_vv_i16m8((vint16m8_t)(op0), (vint16m8_t)(op1), (size_t)(op2)) +#define vand_vv_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vv_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (vint16m8_t)(op3), (size_t)(op4)) +#define vand_vv_i16mf2(op0, op1, op2) \ +__builtin_rvv_vand_vv_i16mf2((vint16mf2_t)(op0), (vint16mf2_t)(op1), (size_t)(op2)) +#define vand_vv_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vv_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vand_vv_i16mf4(op0, op1, op2) \ +__builtin_rvv_vand_vv_i16mf4((vint16mf4_t)(op0), (vint16mf4_t)(op1), (size_t)(op2)) +#define vand_vv_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vv_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vand_vv_i32m1(op0, op1, op2) \ +__builtin_rvv_vand_vv_i32m1((vint32m1_t)(op0), (vint32m1_t)(op1), (size_t)(op2)) +#define vand_vv_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vv_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vand_vv_i32m2(op0, op1, op2) \ +__builtin_rvv_vand_vv_i32m2((vint32m2_t)(op0), (vint32m2_t)(op1), (size_t)(op2)) +#define vand_vv_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vv_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vand_vv_i32m4(op0, op1, op2) \ +__builtin_rvv_vand_vv_i32m4((vint32m4_t)(op0), (vint32m4_t)(op1), (size_t)(op2)) +#define vand_vv_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vv_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vand_vv_i32m8(op0, op1, op2) \ +__builtin_rvv_vand_vv_i32m8((vint32m8_t)(op0), (vint32m8_t)(op1), (size_t)(op2)) +#define vand_vv_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vv_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vand_vv_i32mf2(op0, op1, op2) \ +__builtin_rvv_vand_vv_i32mf2((vint32mf2_t)(op0), (vint32mf2_t)(op1), (size_t)(op2)) +#define vand_vv_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vv_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vand_vv_i64m1(op0, op1, op2) \ +__builtin_rvv_vand_vv_i64m1((vint64m1_t)(op0), (vint64m1_t)(op1), (size_t)(op2)) +#define vand_vv_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vv_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vand_vv_i64m2(op0, op1, op2) \ +__builtin_rvv_vand_vv_i64m2((vint64m2_t)(op0), (vint64m2_t)(op1), (size_t)(op2)) +#define vand_vv_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vv_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vand_vv_i64m4(op0, op1, op2) \ +__builtin_rvv_vand_vv_i64m4((vint64m4_t)(op0), (vint64m4_t)(op1), (size_t)(op2)) +#define vand_vv_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vv_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vand_vv_i64m8(op0, op1, op2) \ +__builtin_rvv_vand_vv_i64m8((vint64m8_t)(op0), (vint64m8_t)(op1), (size_t)(op2)) +#define vand_vv_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vv_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vand_vx_i8m1(op0, op1, op2) \ +__builtin_rvv_vand_vx_i8m1((vint8m1_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vand_vx_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vx_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vand_vx_i8m2(op0, op1, op2) \ +__builtin_rvv_vand_vx_i8m2((vint8m2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vand_vx_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vx_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vand_vx_i8m4(op0, op1, op2) \ +__builtin_rvv_vand_vx_i8m4((vint8m4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vand_vx_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vx_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vand_vx_i8m8(op0, op1, op2) \ +__builtin_rvv_vand_vx_i8m8((vint8m8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vand_vx_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vx_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vand_vx_i8mf2(op0, op1, op2) \ +__builtin_rvv_vand_vx_i8mf2((vint8mf2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vand_vx_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vx_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vand_vx_i8mf4(op0, op1, op2) \ +__builtin_rvv_vand_vx_i8mf4((vint8mf4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vand_vx_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vx_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vand_vx_i8mf8(op0, op1, op2) \ +__builtin_rvv_vand_vx_i8mf8((vint8mf8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vand_vx_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vx_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vand_vx_i16m1(op0, op1, op2) \ +__builtin_rvv_vand_vx_i16m1((vint16m1_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vand_vx_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vx_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vand_vx_i16m2(op0, op1, op2) \ +__builtin_rvv_vand_vx_i16m2((vint16m2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vand_vx_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vx_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vand_vx_i16m4(op0, op1, op2) \ +__builtin_rvv_vand_vx_i16m4((vint16m4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vand_vx_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vx_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vand_vx_i16m8(op0, op1, op2) \ +__builtin_rvv_vand_vx_i16m8((vint16m8_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vand_vx_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vx_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vand_vx_i16mf2(op0, op1, op2) \ +__builtin_rvv_vand_vx_i16mf2((vint16mf2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vand_vx_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vx_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vand_vx_i16mf4(op0, op1, op2) \ +__builtin_rvv_vand_vx_i16mf4((vint16mf4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vand_vx_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vx_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vand_vx_i32m1(op0, op1, op2) \ +__builtin_rvv_vand_vx_i32m1((vint32m1_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vand_vx_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vx_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vand_vx_i32m2(op0, op1, op2) \ +__builtin_rvv_vand_vx_i32m2((vint32m2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vand_vx_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vx_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vand_vx_i32m4(op0, op1, op2) \ +__builtin_rvv_vand_vx_i32m4((vint32m4_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vand_vx_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vx_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vand_vx_i32m8(op0, op1, op2) \ +__builtin_rvv_vand_vx_i32m8((vint32m8_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vand_vx_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vx_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vand_vx_i32mf2(op0, op1, op2) \ +__builtin_rvv_vand_vx_i32mf2((vint32mf2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vand_vx_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vx_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vand_vx_i64m1(op0, op1, op2) \ +__builtin_rvv_vand_vx_i64m1((vint64m1_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vand_vx_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vx_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vand_vx_i64m2(op0, op1, op2) \ +__builtin_rvv_vand_vx_i64m2((vint64m2_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vand_vx_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vx_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vand_vx_i64m4(op0, op1, op2) \ +__builtin_rvv_vand_vx_i64m4((vint64m4_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vand_vx_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vx_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vand_vx_i64m8(op0, op1, op2) \ +__builtin_rvv_vand_vx_i64m8((vint64m8_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vand_vx_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vx_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vand_vv_u8m1(op0, op1, op2) \ +__builtin_rvv_vand_vv_u8m1((vuint8m1_t)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vand_vv_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vv_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vand_vv_u8m2(op0, op1, op2) \ +__builtin_rvv_vand_vv_u8m2((vuint8m2_t)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vand_vv_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vv_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vand_vv_u8m4(op0, op1, op2) \ +__builtin_rvv_vand_vv_u8m4((vuint8m4_t)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vand_vv_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vv_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vand_vv_u8m8(op0, op1, op2) \ +__builtin_rvv_vand_vv_u8m8((vuint8m8_t)(op0), (vuint8m8_t)(op1), (size_t)(op2)) +#define vand_vv_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vv_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (vuint8m8_t)(op3), (size_t)(op4)) +#define vand_vv_u8mf2(op0, op1, op2) \ +__builtin_rvv_vand_vv_u8mf2((vuint8mf2_t)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vand_vv_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vv_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vand_vv_u8mf4(op0, op1, op2) \ +__builtin_rvv_vand_vv_u8mf4((vuint8mf4_t)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vand_vv_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vv_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vand_vv_u8mf8(op0, op1, op2) \ +__builtin_rvv_vand_vv_u8mf8((vuint8mf8_t)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vand_vv_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vv_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vand_vv_u16m1(op0, op1, op2) \ +__builtin_rvv_vand_vv_u16m1((vuint16m1_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vand_vv_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vv_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vand_vv_u16m2(op0, op1, op2) \ +__builtin_rvv_vand_vv_u16m2((vuint16m2_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vand_vv_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vv_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vand_vv_u16m4(op0, op1, op2) \ +__builtin_rvv_vand_vv_u16m4((vuint16m4_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vand_vv_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vv_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vand_vv_u16m8(op0, op1, op2) \ +__builtin_rvv_vand_vv_u16m8((vuint16m8_t)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vand_vv_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vv_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vand_vv_u16mf2(op0, op1, op2) \ +__builtin_rvv_vand_vv_u16mf2((vuint16mf2_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vand_vv_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vv_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vand_vv_u16mf4(op0, op1, op2) \ +__builtin_rvv_vand_vv_u16mf4((vuint16mf4_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vand_vv_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vv_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vand_vv_u32m1(op0, op1, op2) \ +__builtin_rvv_vand_vv_u32m1((vuint32m1_t)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vand_vv_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vv_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vand_vv_u32m2(op0, op1, op2) \ +__builtin_rvv_vand_vv_u32m2((vuint32m2_t)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vand_vv_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vv_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vand_vv_u32m4(op0, op1, op2) \ +__builtin_rvv_vand_vv_u32m4((vuint32m4_t)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vand_vv_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vv_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vand_vv_u32m8(op0, op1, op2) \ +__builtin_rvv_vand_vv_u32m8((vuint32m8_t)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vand_vv_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vv_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vand_vv_u32mf2(op0, op1, op2) \ +__builtin_rvv_vand_vv_u32mf2((vuint32mf2_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vand_vv_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vv_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vand_vv_u64m1(op0, op1, op2) \ +__builtin_rvv_vand_vv_u64m1((vuint64m1_t)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vand_vv_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vv_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vand_vv_u64m2(op0, op1, op2) \ +__builtin_rvv_vand_vv_u64m2((vuint64m2_t)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vand_vv_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vv_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vand_vv_u64m4(op0, op1, op2) \ +__builtin_rvv_vand_vv_u64m4((vuint64m4_t)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vand_vv_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vv_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vand_vv_u64m8(op0, op1, op2) \ +__builtin_rvv_vand_vv_u64m8((vuint64m8_t)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vand_vv_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vv_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vand_vx_u8m1(op0, op1, op2) \ +__builtin_rvv_vand_vx_u8m1((vuint8m1_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vand_vx_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vx_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vand_vx_u8m2(op0, op1, op2) \ +__builtin_rvv_vand_vx_u8m2((vuint8m2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vand_vx_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vx_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vand_vx_u8m4(op0, op1, op2) \ +__builtin_rvv_vand_vx_u8m4((vuint8m4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vand_vx_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vx_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vand_vx_u8m8(op0, op1, op2) \ +__builtin_rvv_vand_vx_u8m8((vuint8m8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vand_vx_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vx_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vand_vx_u8mf2(op0, op1, op2) \ +__builtin_rvv_vand_vx_u8mf2((vuint8mf2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vand_vx_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vx_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vand_vx_u8mf4(op0, op1, op2) \ +__builtin_rvv_vand_vx_u8mf4((vuint8mf4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vand_vx_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vx_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vand_vx_u8mf8(op0, op1, op2) \ +__builtin_rvv_vand_vx_u8mf8((vuint8mf8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vand_vx_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vx_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vand_vx_u16m1(op0, op1, op2) \ +__builtin_rvv_vand_vx_u16m1((vuint16m1_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vand_vx_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vx_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vand_vx_u16m2(op0, op1, op2) \ +__builtin_rvv_vand_vx_u16m2((vuint16m2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vand_vx_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vx_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vand_vx_u16m4(op0, op1, op2) \ +__builtin_rvv_vand_vx_u16m4((vuint16m4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vand_vx_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vx_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vand_vx_u16m8(op0, op1, op2) \ +__builtin_rvv_vand_vx_u16m8((vuint16m8_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vand_vx_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vx_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vand_vx_u16mf2(op0, op1, op2) \ +__builtin_rvv_vand_vx_u16mf2((vuint16mf2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vand_vx_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vx_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vand_vx_u16mf4(op0, op1, op2) \ +__builtin_rvv_vand_vx_u16mf4((vuint16mf4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vand_vx_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vx_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vand_vx_u32m1(op0, op1, op2) \ +__builtin_rvv_vand_vx_u32m1((vuint32m1_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vand_vx_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vx_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vand_vx_u32m2(op0, op1, op2) \ +__builtin_rvv_vand_vx_u32m2((vuint32m2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vand_vx_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vx_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vand_vx_u32m4(op0, op1, op2) \ +__builtin_rvv_vand_vx_u32m4((vuint32m4_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vand_vx_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vx_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vand_vx_u32m8(op0, op1, op2) \ +__builtin_rvv_vand_vx_u32m8((vuint32m8_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vand_vx_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vx_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vand_vx_u32mf2(op0, op1, op2) \ +__builtin_rvv_vand_vx_u32mf2((vuint32mf2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vand_vx_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vx_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vand_vx_u64m1(op0, op1, op2) \ +__builtin_rvv_vand_vx_u64m1((vuint64m1_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vand_vx_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vx_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vand_vx_u64m2(op0, op1, op2) \ +__builtin_rvv_vand_vx_u64m2((vuint64m2_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vand_vx_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vx_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vand_vx_u64m4(op0, op1, op2) \ +__builtin_rvv_vand_vx_u64m4((vuint64m4_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vand_vx_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vx_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vand_vx_u64m8(op0, op1, op2) \ +__builtin_rvv_vand_vx_u64m8((vuint64m8_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vand_vx_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vand_vx_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vxor_vv_i8m1(op0, op1, op2) \ +__builtin_rvv_vxor_vv_i8m1((vint8m1_t)(op0), (vint8m1_t)(op1), (size_t)(op2)) +#define vxor_vv_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vv_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vxor_vv_i8m2(op0, op1, op2) \ +__builtin_rvv_vxor_vv_i8m2((vint8m2_t)(op0), (vint8m2_t)(op1), (size_t)(op2)) +#define vxor_vv_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vv_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vxor_vv_i8m4(op0, op1, op2) \ +__builtin_rvv_vxor_vv_i8m4((vint8m4_t)(op0), (vint8m4_t)(op1), (size_t)(op2)) +#define vxor_vv_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vv_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vxor_vv_i8m8(op0, op1, op2) \ +__builtin_rvv_vxor_vv_i8m8((vint8m8_t)(op0), (vint8m8_t)(op1), (size_t)(op2)) +#define vxor_vv_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vv_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (vint8m8_t)(op3), (size_t)(op4)) +#define vxor_vv_i8mf2(op0, op1, op2) \ +__builtin_rvv_vxor_vv_i8mf2((vint8mf2_t)(op0), (vint8mf2_t)(op1), (size_t)(op2)) +#define vxor_vv_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vv_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vxor_vv_i8mf4(op0, op1, op2) \ +__builtin_rvv_vxor_vv_i8mf4((vint8mf4_t)(op0), (vint8mf4_t)(op1), (size_t)(op2)) +#define vxor_vv_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vv_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vxor_vv_i8mf8(op0, op1, op2) \ +__builtin_rvv_vxor_vv_i8mf8((vint8mf8_t)(op0), (vint8mf8_t)(op1), (size_t)(op2)) +#define vxor_vv_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vv_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vxor_vv_i16m1(op0, op1, op2) \ +__builtin_rvv_vxor_vv_i16m1((vint16m1_t)(op0), (vint16m1_t)(op1), (size_t)(op2)) +#define vxor_vv_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vv_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vxor_vv_i16m2(op0, op1, op2) \ +__builtin_rvv_vxor_vv_i16m2((vint16m2_t)(op0), (vint16m2_t)(op1), (size_t)(op2)) +#define vxor_vv_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vv_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vxor_vv_i16m4(op0, op1, op2) \ +__builtin_rvv_vxor_vv_i16m4((vint16m4_t)(op0), (vint16m4_t)(op1), (size_t)(op2)) +#define vxor_vv_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vv_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vxor_vv_i16m8(op0, op1, op2) \ +__builtin_rvv_vxor_vv_i16m8((vint16m8_t)(op0), (vint16m8_t)(op1), (size_t)(op2)) +#define vxor_vv_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vv_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (vint16m8_t)(op3), (size_t)(op4)) +#define vxor_vv_i16mf2(op0, op1, op2) \ +__builtin_rvv_vxor_vv_i16mf2((vint16mf2_t)(op0), (vint16mf2_t)(op1), (size_t)(op2)) +#define vxor_vv_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vv_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vxor_vv_i16mf4(op0, op1, op2) \ +__builtin_rvv_vxor_vv_i16mf4((vint16mf4_t)(op0), (vint16mf4_t)(op1), (size_t)(op2)) +#define vxor_vv_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vv_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vxor_vv_i32m1(op0, op1, op2) \ +__builtin_rvv_vxor_vv_i32m1((vint32m1_t)(op0), (vint32m1_t)(op1), (size_t)(op2)) +#define vxor_vv_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vv_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vxor_vv_i32m2(op0, op1, op2) \ +__builtin_rvv_vxor_vv_i32m2((vint32m2_t)(op0), (vint32m2_t)(op1), (size_t)(op2)) +#define vxor_vv_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vv_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vxor_vv_i32m4(op0, op1, op2) \ +__builtin_rvv_vxor_vv_i32m4((vint32m4_t)(op0), (vint32m4_t)(op1), (size_t)(op2)) +#define vxor_vv_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vv_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vxor_vv_i32m8(op0, op1, op2) \ +__builtin_rvv_vxor_vv_i32m8((vint32m8_t)(op0), (vint32m8_t)(op1), (size_t)(op2)) +#define vxor_vv_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vv_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vxor_vv_i32mf2(op0, op1, op2) \ +__builtin_rvv_vxor_vv_i32mf2((vint32mf2_t)(op0), (vint32mf2_t)(op1), (size_t)(op2)) +#define vxor_vv_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vv_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vxor_vv_i64m1(op0, op1, op2) \ +__builtin_rvv_vxor_vv_i64m1((vint64m1_t)(op0), (vint64m1_t)(op1), (size_t)(op2)) +#define vxor_vv_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vv_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vxor_vv_i64m2(op0, op1, op2) \ +__builtin_rvv_vxor_vv_i64m2((vint64m2_t)(op0), (vint64m2_t)(op1), (size_t)(op2)) +#define vxor_vv_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vv_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vxor_vv_i64m4(op0, op1, op2) \ +__builtin_rvv_vxor_vv_i64m4((vint64m4_t)(op0), (vint64m4_t)(op1), (size_t)(op2)) +#define vxor_vv_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vv_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vxor_vv_i64m8(op0, op1, op2) \ +__builtin_rvv_vxor_vv_i64m8((vint64m8_t)(op0), (vint64m8_t)(op1), (size_t)(op2)) +#define vxor_vv_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vv_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vxor_vx_i8m1(op0, op1, op2) \ +__builtin_rvv_vxor_vx_i8m1((vint8m1_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vxor_vx_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vx_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vxor_vx_i8m2(op0, op1, op2) \ +__builtin_rvv_vxor_vx_i8m2((vint8m2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vxor_vx_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vx_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vxor_vx_i8m4(op0, op1, op2) \ +__builtin_rvv_vxor_vx_i8m4((vint8m4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vxor_vx_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vx_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vxor_vx_i8m8(op0, op1, op2) \ +__builtin_rvv_vxor_vx_i8m8((vint8m8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vxor_vx_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vx_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vxor_vx_i8mf2(op0, op1, op2) \ +__builtin_rvv_vxor_vx_i8mf2((vint8mf2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vxor_vx_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vx_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vxor_vx_i8mf4(op0, op1, op2) \ +__builtin_rvv_vxor_vx_i8mf4((vint8mf4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vxor_vx_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vx_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vxor_vx_i8mf8(op0, op1, op2) \ +__builtin_rvv_vxor_vx_i8mf8((vint8mf8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vxor_vx_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vx_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vxor_vx_i16m1(op0, op1, op2) \ +__builtin_rvv_vxor_vx_i16m1((vint16m1_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vxor_vx_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vx_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vxor_vx_i16m2(op0, op1, op2) \ +__builtin_rvv_vxor_vx_i16m2((vint16m2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vxor_vx_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vx_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vxor_vx_i16m4(op0, op1, op2) \ +__builtin_rvv_vxor_vx_i16m4((vint16m4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vxor_vx_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vx_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vxor_vx_i16m8(op0, op1, op2) \ +__builtin_rvv_vxor_vx_i16m8((vint16m8_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vxor_vx_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vx_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vxor_vx_i16mf2(op0, op1, op2) \ +__builtin_rvv_vxor_vx_i16mf2((vint16mf2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vxor_vx_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vx_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vxor_vx_i16mf4(op0, op1, op2) \ +__builtin_rvv_vxor_vx_i16mf4((vint16mf4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vxor_vx_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vx_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vxor_vx_i32m1(op0, op1, op2) \ +__builtin_rvv_vxor_vx_i32m1((vint32m1_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vxor_vx_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vx_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vxor_vx_i32m2(op0, op1, op2) \ +__builtin_rvv_vxor_vx_i32m2((vint32m2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vxor_vx_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vx_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vxor_vx_i32m4(op0, op1, op2) \ +__builtin_rvv_vxor_vx_i32m4((vint32m4_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vxor_vx_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vx_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vxor_vx_i32m8(op0, op1, op2) \ +__builtin_rvv_vxor_vx_i32m8((vint32m8_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vxor_vx_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vx_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vxor_vx_i32mf2(op0, op1, op2) \ +__builtin_rvv_vxor_vx_i32mf2((vint32mf2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vxor_vx_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vx_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vxor_vx_i64m1(op0, op1, op2) \ +__builtin_rvv_vxor_vx_i64m1((vint64m1_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vxor_vx_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vx_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vxor_vx_i64m2(op0, op1, op2) \ +__builtin_rvv_vxor_vx_i64m2((vint64m2_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vxor_vx_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vx_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vxor_vx_i64m4(op0, op1, op2) \ +__builtin_rvv_vxor_vx_i64m4((vint64m4_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vxor_vx_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vx_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vxor_vx_i64m8(op0, op1, op2) \ +__builtin_rvv_vxor_vx_i64m8((vint64m8_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vxor_vx_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vx_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vsuxei16_v_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_i16m1((int16_t *)(op0), (vuint16m1_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vsuxei16_v_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vsuxei16_v_i16m2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_i16m2((int16_t *)(op0), (vuint16m2_t)(op1), (vint16m2_t)(op2), (size_t)(op3)) +#define vsuxei16_v_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_i16m2_m((vbool8_t)(op0), (int16_t *)(op1), (vuint16m2_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vsuxei16_v_i16m4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_i16m4((int16_t *)(op0), (vuint16m4_t)(op1), (vint16m4_t)(op2), (size_t)(op3)) +#define vsuxei16_v_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_i16m4_m((vbool4_t)(op0), (int16_t *)(op1), (vuint16m4_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vsuxei16_v_i16m8(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_i16m8((int16_t *)(op0), (vuint16m8_t)(op1), (vint16m8_t)(op2), (size_t)(op3)) +#define vsuxei16_v_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_i16m8_m((vbool2_t)(op0), (int16_t *)(op1), (vuint16m8_t)(op2), (vint16m8_t)(op3), (size_t)(op4)) +#define vsuxei16_v_i16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_i16mf2((int16_t *)(op0), (vuint16mf2_t)(op1), (vint16mf2_t)(op2), (size_t)(op3)) +#define vsuxei16_v_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint16mf2_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vsuxei16_v_i16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_i16mf4((int16_t *)(op0), (vuint16mf4_t)(op1), (vint16mf4_t)(op2), (size_t)(op3)) +#define vsuxei16_v_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint16mf4_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vxor_vv_u8m1(op0, op1, op2) \ +__builtin_rvv_vxor_vv_u8m1((vuint8m1_t)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vxor_vv_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vv_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vxor_vv_u8m2(op0, op1, op2) \ +__builtin_rvv_vxor_vv_u8m2((vuint8m2_t)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vxor_vv_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vv_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vxor_vv_u8m4(op0, op1, op2) \ +__builtin_rvv_vxor_vv_u8m4((vuint8m4_t)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vxor_vv_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vv_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vxor_vv_u8m8(op0, op1, op2) \ +__builtin_rvv_vxor_vv_u8m8((vuint8m8_t)(op0), (vuint8m8_t)(op1), (size_t)(op2)) +#define vxor_vv_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vv_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (vuint8m8_t)(op3), (size_t)(op4)) +#define vxor_vv_u8mf2(op0, op1, op2) \ +__builtin_rvv_vxor_vv_u8mf2((vuint8mf2_t)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vxor_vv_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vv_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vxor_vv_u8mf4(op0, op1, op2) \ +__builtin_rvv_vxor_vv_u8mf4((vuint8mf4_t)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vxor_vv_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vv_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vxor_vv_u8mf8(op0, op1, op2) \ +__builtin_rvv_vxor_vv_u8mf8((vuint8mf8_t)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vxor_vv_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vv_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vxor_vv_u16m1(op0, op1, op2) \ +__builtin_rvv_vxor_vv_u16m1((vuint16m1_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vxor_vv_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vv_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vxor_vv_u16m2(op0, op1, op2) \ +__builtin_rvv_vxor_vv_u16m2((vuint16m2_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vxor_vv_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vv_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vxor_vv_u16m4(op0, op1, op2) \ +__builtin_rvv_vxor_vv_u16m4((vuint16m4_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vxor_vv_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vv_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vxor_vv_u16m8(op0, op1, op2) \ +__builtin_rvv_vxor_vv_u16m8((vuint16m8_t)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vxor_vv_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vv_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vxor_vv_u16mf2(op0, op1, op2) \ +__builtin_rvv_vxor_vv_u16mf2((vuint16mf2_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vxor_vv_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vv_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vxor_vv_u16mf4(op0, op1, op2) \ +__builtin_rvv_vxor_vv_u16mf4((vuint16mf4_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vxor_vv_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vv_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vxor_vv_u32m1(op0, op1, op2) \ +__builtin_rvv_vxor_vv_u32m1((vuint32m1_t)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vxor_vv_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vv_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vxor_vv_u32m2(op0, op1, op2) \ +__builtin_rvv_vxor_vv_u32m2((vuint32m2_t)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vxor_vv_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vv_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vxor_vv_u32m4(op0, op1, op2) \ +__builtin_rvv_vxor_vv_u32m4((vuint32m4_t)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vxor_vv_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vv_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vxor_vv_u32m8(op0, op1, op2) \ +__builtin_rvv_vxor_vv_u32m8((vuint32m8_t)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vxor_vv_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vv_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vxor_vv_u32mf2(op0, op1, op2) \ +__builtin_rvv_vxor_vv_u32mf2((vuint32mf2_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vxor_vv_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vv_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vxor_vv_u64m1(op0, op1, op2) \ +__builtin_rvv_vxor_vv_u64m1((vuint64m1_t)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vxor_vv_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vv_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vxor_vv_u64m2(op0, op1, op2) \ +__builtin_rvv_vxor_vv_u64m2((vuint64m2_t)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vxor_vv_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vv_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vxor_vv_u64m4(op0, op1, op2) \ +__builtin_rvv_vxor_vv_u64m4((vuint64m4_t)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vxor_vv_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vv_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vxor_vv_u64m8(op0, op1, op2) \ +__builtin_rvv_vxor_vv_u64m8((vuint64m8_t)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vxor_vv_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vv_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vxor_vx_u8m1(op0, op1, op2) \ +__builtin_rvv_vxor_vx_u8m1((vuint8m1_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vxor_vx_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vx_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vxor_vx_u8m2(op0, op1, op2) \ +__builtin_rvv_vxor_vx_u8m2((vuint8m2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vxor_vx_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vx_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vxor_vx_u8m4(op0, op1, op2) \ +__builtin_rvv_vxor_vx_u8m4((vuint8m4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vxor_vx_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vx_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vxor_vx_u8m8(op0, op1, op2) \ +__builtin_rvv_vxor_vx_u8m8((vuint8m8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vxor_vx_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vx_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vxor_vx_u8mf2(op0, op1, op2) \ +__builtin_rvv_vxor_vx_u8mf2((vuint8mf2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vxor_vx_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vx_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vxor_vx_u8mf4(op0, op1, op2) \ +__builtin_rvv_vxor_vx_u8mf4((vuint8mf4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vxor_vx_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vx_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vxor_vx_u8mf8(op0, op1, op2) \ +__builtin_rvv_vxor_vx_u8mf8((vuint8mf8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vxor_vx_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vx_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vxor_vx_u16m1(op0, op1, op2) \ +__builtin_rvv_vxor_vx_u16m1((vuint16m1_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vxor_vx_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vx_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vxor_vx_u16m2(op0, op1, op2) \ +__builtin_rvv_vxor_vx_u16m2((vuint16m2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vxor_vx_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vx_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vxor_vx_u16m4(op0, op1, op2) \ +__builtin_rvv_vxor_vx_u16m4((vuint16m4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vxor_vx_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vx_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vxor_vx_u16m8(op0, op1, op2) \ +__builtin_rvv_vxor_vx_u16m8((vuint16m8_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vxor_vx_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vx_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vxor_vx_u16mf2(op0, op1, op2) \ +__builtin_rvv_vxor_vx_u16mf2((vuint16mf2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vxor_vx_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vx_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vxor_vx_u16mf4(op0, op1, op2) \ +__builtin_rvv_vxor_vx_u16mf4((vuint16mf4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vxor_vx_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vx_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vxor_vx_u32m1(op0, op1, op2) \ +__builtin_rvv_vxor_vx_u32m1((vuint32m1_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vxor_vx_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vx_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vxor_vx_u32m2(op0, op1, op2) \ +__builtin_rvv_vxor_vx_u32m2((vuint32m2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vxor_vx_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vx_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vxor_vx_u32m4(op0, op1, op2) \ +__builtin_rvv_vxor_vx_u32m4((vuint32m4_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vxor_vx_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vx_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vxor_vx_u32m8(op0, op1, op2) \ +__builtin_rvv_vxor_vx_u32m8((vuint32m8_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vxor_vx_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vx_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vxor_vx_u32mf2(op0, op1, op2) \ +__builtin_rvv_vxor_vx_u32mf2((vuint32mf2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vxor_vx_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vx_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vxor_vx_u64m1(op0, op1, op2) \ +__builtin_rvv_vxor_vx_u64m1((vuint64m1_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vxor_vx_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vx_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vxor_vx_u64m2(op0, op1, op2) \ +__builtin_rvv_vxor_vx_u64m2((vuint64m2_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vxor_vx_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vx_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vxor_vx_u64m4(op0, op1, op2) \ +__builtin_rvv_vxor_vx_u64m4((vuint64m4_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vxor_vx_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vx_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vxor_vx_u64m8(op0, op1, op2) \ +__builtin_rvv_vxor_vx_u64m8((vuint64m8_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vxor_vx_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vxor_vx_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vor_vv_i8m1(op0, op1, op2) \ +__builtin_rvv_vor_vv_i8m1((vint8m1_t)(op0), (vint8m1_t)(op1), (size_t)(op2)) +#define vor_vv_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vv_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vor_vv_i8m2(op0, op1, op2) \ +__builtin_rvv_vor_vv_i8m2((vint8m2_t)(op0), (vint8m2_t)(op1), (size_t)(op2)) +#define vor_vv_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vv_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vor_vv_i8m4(op0, op1, op2) \ +__builtin_rvv_vor_vv_i8m4((vint8m4_t)(op0), (vint8m4_t)(op1), (size_t)(op2)) +#define vor_vv_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vv_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vor_vv_i8m8(op0, op1, op2) \ +__builtin_rvv_vor_vv_i8m8((vint8m8_t)(op0), (vint8m8_t)(op1), (size_t)(op2)) +#define vor_vv_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vv_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (vint8m8_t)(op3), (size_t)(op4)) +#define vor_vv_i8mf2(op0, op1, op2) \ +__builtin_rvv_vor_vv_i8mf2((vint8mf2_t)(op0), (vint8mf2_t)(op1), (size_t)(op2)) +#define vor_vv_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vv_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vor_vv_i8mf4(op0, op1, op2) \ +__builtin_rvv_vor_vv_i8mf4((vint8mf4_t)(op0), (vint8mf4_t)(op1), (size_t)(op2)) +#define vor_vv_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vv_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vor_vv_i8mf8(op0, op1, op2) \ +__builtin_rvv_vor_vv_i8mf8((vint8mf8_t)(op0), (vint8mf8_t)(op1), (size_t)(op2)) +#define vor_vv_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vv_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vor_vv_i16m1(op0, op1, op2) \ +__builtin_rvv_vor_vv_i16m1((vint16m1_t)(op0), (vint16m1_t)(op1), (size_t)(op2)) +#define vor_vv_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vv_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vor_vv_i16m2(op0, op1, op2) \ +__builtin_rvv_vor_vv_i16m2((vint16m2_t)(op0), (vint16m2_t)(op1), (size_t)(op2)) +#define vor_vv_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vv_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vor_vv_i16m4(op0, op1, op2) \ +__builtin_rvv_vor_vv_i16m4((vint16m4_t)(op0), (vint16m4_t)(op1), (size_t)(op2)) +#define vor_vv_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vv_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vor_vv_i16m8(op0, op1, op2) \ +__builtin_rvv_vor_vv_i16m8((vint16m8_t)(op0), (vint16m8_t)(op1), (size_t)(op2)) +#define vor_vv_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vv_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (vint16m8_t)(op3), (size_t)(op4)) +#define vor_vv_i16mf2(op0, op1, op2) \ +__builtin_rvv_vor_vv_i16mf2((vint16mf2_t)(op0), (vint16mf2_t)(op1), (size_t)(op2)) +#define vor_vv_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vv_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vor_vv_i16mf4(op0, op1, op2) \ +__builtin_rvv_vor_vv_i16mf4((vint16mf4_t)(op0), (vint16mf4_t)(op1), (size_t)(op2)) +#define vor_vv_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vv_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vor_vv_i32m1(op0, op1, op2) \ +__builtin_rvv_vor_vv_i32m1((vint32m1_t)(op0), (vint32m1_t)(op1), (size_t)(op2)) +#define vor_vv_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vv_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vor_vv_i32m2(op0, op1, op2) \ +__builtin_rvv_vor_vv_i32m2((vint32m2_t)(op0), (vint32m2_t)(op1), (size_t)(op2)) +#define vor_vv_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vv_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vor_vv_i32m4(op0, op1, op2) \ +__builtin_rvv_vor_vv_i32m4((vint32m4_t)(op0), (vint32m4_t)(op1), (size_t)(op2)) +#define vor_vv_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vv_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vor_vv_i32m8(op0, op1, op2) \ +__builtin_rvv_vor_vv_i32m8((vint32m8_t)(op0), (vint32m8_t)(op1), (size_t)(op2)) +#define vor_vv_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vv_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vor_vv_i32mf2(op0, op1, op2) \ +__builtin_rvv_vor_vv_i32mf2((vint32mf2_t)(op0), (vint32mf2_t)(op1), (size_t)(op2)) +#define vor_vv_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vv_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vor_vv_i64m1(op0, op1, op2) \ +__builtin_rvv_vor_vv_i64m1((vint64m1_t)(op0), (vint64m1_t)(op1), (size_t)(op2)) +#define vor_vv_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vv_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vor_vv_i64m2(op0, op1, op2) \ +__builtin_rvv_vor_vv_i64m2((vint64m2_t)(op0), (vint64m2_t)(op1), (size_t)(op2)) +#define vor_vv_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vv_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vor_vv_i64m4(op0, op1, op2) \ +__builtin_rvv_vor_vv_i64m4((vint64m4_t)(op0), (vint64m4_t)(op1), (size_t)(op2)) +#define vor_vv_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vv_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vor_vv_i64m8(op0, op1, op2) \ +__builtin_rvv_vor_vv_i64m8((vint64m8_t)(op0), (vint64m8_t)(op1), (size_t)(op2)) +#define vor_vv_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vv_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vor_vx_i8m1(op0, op1, op2) \ +__builtin_rvv_vor_vx_i8m1((vint8m1_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vor_vx_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vx_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vor_vx_i8m2(op0, op1, op2) \ +__builtin_rvv_vor_vx_i8m2((vint8m2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vor_vx_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vx_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vor_vx_i8m4(op0, op1, op2) \ +__builtin_rvv_vor_vx_i8m4((vint8m4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vor_vx_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vx_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vor_vx_i8m8(op0, op1, op2) \ +__builtin_rvv_vor_vx_i8m8((vint8m8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vor_vx_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vx_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vor_vx_i8mf2(op0, op1, op2) \ +__builtin_rvv_vor_vx_i8mf2((vint8mf2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vor_vx_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vx_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vor_vx_i8mf4(op0, op1, op2) \ +__builtin_rvv_vor_vx_i8mf4((vint8mf4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vor_vx_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vx_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vor_vx_i8mf8(op0, op1, op2) \ +__builtin_rvv_vor_vx_i8mf8((vint8mf8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vor_vx_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vx_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vor_vx_i16m1(op0, op1, op2) \ +__builtin_rvv_vor_vx_i16m1((vint16m1_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vor_vx_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vx_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vor_vx_i16m2(op0, op1, op2) \ +__builtin_rvv_vor_vx_i16m2((vint16m2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vor_vx_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vx_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vor_vx_i16m4(op0, op1, op2) \ +__builtin_rvv_vor_vx_i16m4((vint16m4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vor_vx_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vx_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vor_vx_i16m8(op0, op1, op2) \ +__builtin_rvv_vor_vx_i16m8((vint16m8_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vor_vx_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vx_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vor_vx_i16mf2(op0, op1, op2) \ +__builtin_rvv_vor_vx_i16mf2((vint16mf2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vor_vx_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vx_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vor_vx_i16mf4(op0, op1, op2) \ +__builtin_rvv_vor_vx_i16mf4((vint16mf4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vor_vx_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vx_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vor_vx_i32m1(op0, op1, op2) \ +__builtin_rvv_vor_vx_i32m1((vint32m1_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vor_vx_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vx_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vor_vx_i32m2(op0, op1, op2) \ +__builtin_rvv_vor_vx_i32m2((vint32m2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vor_vx_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vx_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vor_vx_i32m4(op0, op1, op2) \ +__builtin_rvv_vor_vx_i32m4((vint32m4_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vor_vx_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vx_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vor_vx_i32m8(op0, op1, op2) \ +__builtin_rvv_vor_vx_i32m8((vint32m8_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vor_vx_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vx_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vor_vx_i32mf2(op0, op1, op2) \ +__builtin_rvv_vor_vx_i32mf2((vint32mf2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vor_vx_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vx_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vor_vx_i64m1(op0, op1, op2) \ +__builtin_rvv_vor_vx_i64m1((vint64m1_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vor_vx_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vx_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vor_vx_i64m2(op0, op1, op2) \ +__builtin_rvv_vor_vx_i64m2((vint64m2_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vor_vx_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vx_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vor_vx_i64m4(op0, op1, op2) \ +__builtin_rvv_vor_vx_i64m4((vint64m4_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vor_vx_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vx_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vor_vx_i64m8(op0, op1, op2) \ +__builtin_rvv_vor_vx_i64m8((vint64m8_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vor_vx_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vx_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vor_vv_u8m1(op0, op1, op2) \ +__builtin_rvv_vor_vv_u8m1((vuint8m1_t)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vor_vv_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vv_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vor_vv_u8m2(op0, op1, op2) \ +__builtin_rvv_vor_vv_u8m2((vuint8m2_t)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vor_vv_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vv_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vor_vv_u8m4(op0, op1, op2) \ +__builtin_rvv_vor_vv_u8m4((vuint8m4_t)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vor_vv_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vv_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vor_vv_u8m8(op0, op1, op2) \ +__builtin_rvv_vor_vv_u8m8((vuint8m8_t)(op0), (vuint8m8_t)(op1), (size_t)(op2)) +#define vor_vv_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vv_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (vuint8m8_t)(op3), (size_t)(op4)) +#define vor_vv_u8mf2(op0, op1, op2) \ +__builtin_rvv_vor_vv_u8mf2((vuint8mf2_t)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vor_vv_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vv_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vor_vv_u8mf4(op0, op1, op2) \ +__builtin_rvv_vor_vv_u8mf4((vuint8mf4_t)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vor_vv_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vv_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vor_vv_u8mf8(op0, op1, op2) \ +__builtin_rvv_vor_vv_u8mf8((vuint8mf8_t)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vor_vv_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vv_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vor_vv_u16m1(op0, op1, op2) \ +__builtin_rvv_vor_vv_u16m1((vuint16m1_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vor_vv_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vv_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vor_vv_u16m2(op0, op1, op2) \ +__builtin_rvv_vor_vv_u16m2((vuint16m2_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vor_vv_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vv_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vor_vv_u16m4(op0, op1, op2) \ +__builtin_rvv_vor_vv_u16m4((vuint16m4_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vor_vv_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vv_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vor_vv_u16m8(op0, op1, op2) \ +__builtin_rvv_vor_vv_u16m8((vuint16m8_t)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vor_vv_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vv_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vor_vv_u16mf2(op0, op1, op2) \ +__builtin_rvv_vor_vv_u16mf2((vuint16mf2_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vor_vv_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vv_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vor_vv_u16mf4(op0, op1, op2) \ +__builtin_rvv_vor_vv_u16mf4((vuint16mf4_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vor_vv_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vv_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vor_vv_u32m1(op0, op1, op2) \ +__builtin_rvv_vor_vv_u32m1((vuint32m1_t)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vor_vv_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vv_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vor_vv_u32m2(op0, op1, op2) \ +__builtin_rvv_vor_vv_u32m2((vuint32m2_t)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vor_vv_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vv_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vor_vv_u32m4(op0, op1, op2) \ +__builtin_rvv_vor_vv_u32m4((vuint32m4_t)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vor_vv_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vv_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vor_vv_u32m8(op0, op1, op2) \ +__builtin_rvv_vor_vv_u32m8((vuint32m8_t)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vor_vv_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vv_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vor_vv_u32mf2(op0, op1, op2) \ +__builtin_rvv_vor_vv_u32mf2((vuint32mf2_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vor_vv_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vv_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vor_vv_u64m1(op0, op1, op2) \ +__builtin_rvv_vor_vv_u64m1((vuint64m1_t)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vor_vv_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vv_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vor_vv_u64m2(op0, op1, op2) \ +__builtin_rvv_vor_vv_u64m2((vuint64m2_t)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vor_vv_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vv_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vor_vv_u64m4(op0, op1, op2) \ +__builtin_rvv_vor_vv_u64m4((vuint64m4_t)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vor_vv_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vv_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vor_vv_u64m8(op0, op1, op2) \ +__builtin_rvv_vor_vv_u64m8((vuint64m8_t)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vor_vv_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vv_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vor_vx_u8m1(op0, op1, op2) \ +__builtin_rvv_vor_vx_u8m1((vuint8m1_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vor_vx_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vx_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vor_vx_u8m2(op0, op1, op2) \ +__builtin_rvv_vor_vx_u8m2((vuint8m2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vor_vx_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vx_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vor_vx_u8m4(op0, op1, op2) \ +__builtin_rvv_vor_vx_u8m4((vuint8m4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vor_vx_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vx_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vor_vx_u8m8(op0, op1, op2) \ +__builtin_rvv_vor_vx_u8m8((vuint8m8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vor_vx_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vx_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vor_vx_u8mf2(op0, op1, op2) \ +__builtin_rvv_vor_vx_u8mf2((vuint8mf2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vor_vx_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vx_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vor_vx_u8mf4(op0, op1, op2) \ +__builtin_rvv_vor_vx_u8mf4((vuint8mf4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vor_vx_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vx_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vor_vx_u8mf8(op0, op1, op2) \ +__builtin_rvv_vor_vx_u8mf8((vuint8mf8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vor_vx_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vx_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vor_vx_u16m1(op0, op1, op2) \ +__builtin_rvv_vor_vx_u16m1((vuint16m1_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vor_vx_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vx_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vor_vx_u16m2(op0, op1, op2) \ +__builtin_rvv_vor_vx_u16m2((vuint16m2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vor_vx_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vx_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vor_vx_u16m4(op0, op1, op2) \ +__builtin_rvv_vor_vx_u16m4((vuint16m4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vor_vx_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vx_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vor_vx_u16m8(op0, op1, op2) \ +__builtin_rvv_vor_vx_u16m8((vuint16m8_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vor_vx_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vx_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vor_vx_u16mf2(op0, op1, op2) \ +__builtin_rvv_vor_vx_u16mf2((vuint16mf2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vor_vx_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vx_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vor_vx_u16mf4(op0, op1, op2) \ +__builtin_rvv_vor_vx_u16mf4((vuint16mf4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vor_vx_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vx_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vor_vx_u32m1(op0, op1, op2) \ +__builtin_rvv_vor_vx_u32m1((vuint32m1_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vor_vx_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vx_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vor_vx_u32m2(op0, op1, op2) \ +__builtin_rvv_vor_vx_u32m2((vuint32m2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vor_vx_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vx_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vor_vx_u32m4(op0, op1, op2) \ +__builtin_rvv_vor_vx_u32m4((vuint32m4_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vor_vx_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vx_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vor_vx_u32m8(op0, op1, op2) \ +__builtin_rvv_vor_vx_u32m8((vuint32m8_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vor_vx_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vx_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vor_vx_u32mf2(op0, op1, op2) \ +__builtin_rvv_vor_vx_u32mf2((vuint32mf2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vor_vx_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vx_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vor_vx_u64m1(op0, op1, op2) \ +__builtin_rvv_vor_vx_u64m1((vuint64m1_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vor_vx_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vx_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vor_vx_u64m2(op0, op1, op2) \ +__builtin_rvv_vor_vx_u64m2((vuint64m2_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vor_vx_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vx_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vor_vx_u64m4(op0, op1, op2) \ +__builtin_rvv_vor_vx_u64m4((vuint64m4_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vor_vx_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vx_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vor_vx_u64m8(op0, op1, op2) \ +__builtin_rvv_vor_vx_u64m8((vuint64m8_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vor_vx_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vor_vx_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vsll_vv_i8m1(op0, op1, op2) \ +__builtin_rvv_vsll_vv_i8m1((vint8m1_t)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vsll_vv_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vv_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vsll_vv_i8m2(op0, op1, op2) \ +__builtin_rvv_vsll_vv_i8m2((vint8m2_t)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vsll_vv_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vv_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vsll_vv_i8m4(op0, op1, op2) \ +__builtin_rvv_vsll_vv_i8m4((vint8m4_t)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vsll_vv_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vv_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vsll_vv_i8m8(op0, op1, op2) \ +__builtin_rvv_vsll_vv_i8m8((vint8m8_t)(op0), (vuint8m8_t)(op1), (size_t)(op2)) +#define vsll_vv_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vv_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (vuint8m8_t)(op3), (size_t)(op4)) +#define vsll_vv_i8mf2(op0, op1, op2) \ +__builtin_rvv_vsll_vv_i8mf2((vint8mf2_t)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vsll_vv_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vv_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vsll_vv_i8mf4(op0, op1, op2) \ +__builtin_rvv_vsll_vv_i8mf4((vint8mf4_t)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vsll_vv_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vv_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vsll_vv_i8mf8(op0, op1, op2) \ +__builtin_rvv_vsll_vv_i8mf8((vint8mf8_t)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vsll_vv_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vv_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vsll_vv_i16m1(op0, op1, op2) \ +__builtin_rvv_vsll_vv_i16m1((vint16m1_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vsll_vv_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vv_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vsll_vv_i16m2(op0, op1, op2) \ +__builtin_rvv_vsll_vv_i16m2((vint16m2_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vsll_vv_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vv_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vsll_vv_i16m4(op0, op1, op2) \ +__builtin_rvv_vsll_vv_i16m4((vint16m4_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vsll_vv_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vv_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vsll_vv_i16m8(op0, op1, op2) \ +__builtin_rvv_vsll_vv_i16m8((vint16m8_t)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vsll_vv_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vv_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vsll_vv_i16mf2(op0, op1, op2) \ +__builtin_rvv_vsll_vv_i16mf2((vint16mf2_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vsll_vv_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vv_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vsll_vv_i16mf4(op0, op1, op2) \ +__builtin_rvv_vsll_vv_i16mf4((vint16mf4_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vsll_vv_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vv_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vsll_vv_i32m1(op0, op1, op2) \ +__builtin_rvv_vsll_vv_i32m1((vint32m1_t)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vsll_vv_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vv_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vsll_vv_i32m2(op0, op1, op2) \ +__builtin_rvv_vsll_vv_i32m2((vint32m2_t)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vsll_vv_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vv_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vsll_vv_i32m4(op0, op1, op2) \ +__builtin_rvv_vsll_vv_i32m4((vint32m4_t)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vsll_vv_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vv_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vsll_vv_i32m8(op0, op1, op2) \ +__builtin_rvv_vsll_vv_i32m8((vint32m8_t)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vsll_vv_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vv_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vsll_vv_i32mf2(op0, op1, op2) \ +__builtin_rvv_vsll_vv_i32mf2((vint32mf2_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vsll_vv_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vv_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vsll_vv_i64m1(op0, op1, op2) \ +__builtin_rvv_vsll_vv_i64m1((vint64m1_t)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vsll_vv_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vv_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vsll_vv_i64m2(op0, op1, op2) \ +__builtin_rvv_vsll_vv_i64m2((vint64m2_t)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vsll_vv_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vv_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vsll_vv_i64m4(op0, op1, op2) \ +__builtin_rvv_vsll_vv_i64m4((vint64m4_t)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vsll_vv_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vv_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vsll_vv_i64m8(op0, op1, op2) \ +__builtin_rvv_vsll_vv_i64m8((vint64m8_t)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vsll_vv_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vv_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vsll_vx_i8m1(op0, op1, op2) \ +__builtin_rvv_vsll_vx_i8m1((vint8m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsll_vx_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vx_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsll_vx_i8m2(op0, op1, op2) \ +__builtin_rvv_vsll_vx_i8m2((vint8m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsll_vx_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vx_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsll_vx_i8m4(op0, op1, op2) \ +__builtin_rvv_vsll_vx_i8m4((vint8m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsll_vx_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vx_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsll_vx_i8m8(op0, op1, op2) \ +__builtin_rvv_vsll_vx_i8m8((vint8m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsll_vx_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vx_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsll_vx_i8mf2(op0, op1, op2) \ +__builtin_rvv_vsll_vx_i8mf2((vint8mf2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsll_vx_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vx_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsll_vx_i8mf4(op0, op1, op2) \ +__builtin_rvv_vsll_vx_i8mf4((vint8mf4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsll_vx_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vx_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsll_vx_i8mf8(op0, op1, op2) \ +__builtin_rvv_vsll_vx_i8mf8((vint8mf8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsll_vx_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vx_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsll_vx_i16m1(op0, op1, op2) \ +__builtin_rvv_vsll_vx_i16m1((vint16m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsll_vx_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vx_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsll_vx_i16m2(op0, op1, op2) \ +__builtin_rvv_vsll_vx_i16m2((vint16m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsll_vx_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vx_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsll_vx_i16m4(op0, op1, op2) \ +__builtin_rvv_vsll_vx_i16m4((vint16m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsll_vx_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vx_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsll_vx_i16m8(op0, op1, op2) \ +__builtin_rvv_vsll_vx_i16m8((vint16m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsll_vx_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vx_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsll_vx_i16mf2(op0, op1, op2) \ +__builtin_rvv_vsll_vx_i16mf2((vint16mf2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsll_vx_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vx_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsll_vx_i16mf4(op0, op1, op2) \ +__builtin_rvv_vsll_vx_i16mf4((vint16mf4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsll_vx_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vx_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsll_vx_i32m1(op0, op1, op2) \ +__builtin_rvv_vsll_vx_i32m1((vint32m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsll_vx_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vx_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsll_vx_i32m2(op0, op1, op2) \ +__builtin_rvv_vsll_vx_i32m2((vint32m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsll_vx_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vx_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsll_vx_i32m4(op0, op1, op2) \ +__builtin_rvv_vsll_vx_i32m4((vint32m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsll_vx_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vx_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsll_vx_i32m8(op0, op1, op2) \ +__builtin_rvv_vsll_vx_i32m8((vint32m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsll_vx_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vx_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsll_vx_i32mf2(op0, op1, op2) \ +__builtin_rvv_vsll_vx_i32mf2((vint32mf2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsll_vx_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vx_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsll_vx_i64m1(op0, op1, op2) \ +__builtin_rvv_vsll_vx_i64m1((vint64m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsll_vx_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vx_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsll_vx_i64m2(op0, op1, op2) \ +__builtin_rvv_vsll_vx_i64m2((vint64m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsll_vx_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vx_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsll_vx_i64m4(op0, op1, op2) \ +__builtin_rvv_vsll_vx_i64m4((vint64m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsll_vx_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vx_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsll_vx_i64m8(op0, op1, op2) \ +__builtin_rvv_vsll_vx_i64m8((vint64m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsll_vx_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vx_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsll_vv_u8m1(op0, op1, op2) \ +__builtin_rvv_vsll_vv_u8m1((vuint8m1_t)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vsll_vv_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vv_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vsll_vv_u8m2(op0, op1, op2) \ +__builtin_rvv_vsll_vv_u8m2((vuint8m2_t)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vsll_vv_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vv_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vsll_vv_u8m4(op0, op1, op2) \ +__builtin_rvv_vsll_vv_u8m4((vuint8m4_t)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vsll_vv_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vv_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vsll_vv_u8m8(op0, op1, op2) \ +__builtin_rvv_vsll_vv_u8m8((vuint8m8_t)(op0), (vuint8m8_t)(op1), (size_t)(op2)) +#define vsll_vv_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vv_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (vuint8m8_t)(op3), (size_t)(op4)) +#define vsll_vv_u8mf2(op0, op1, op2) \ +__builtin_rvv_vsll_vv_u8mf2((vuint8mf2_t)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vsll_vv_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vv_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vsll_vv_u8mf4(op0, op1, op2) \ +__builtin_rvv_vsll_vv_u8mf4((vuint8mf4_t)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vsll_vv_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vv_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vsll_vv_u8mf8(op0, op1, op2) \ +__builtin_rvv_vsll_vv_u8mf8((vuint8mf8_t)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vsll_vv_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vv_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vsll_vv_u16m1(op0, op1, op2) \ +__builtin_rvv_vsll_vv_u16m1((vuint16m1_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vsll_vv_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vv_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vsll_vv_u16m2(op0, op1, op2) \ +__builtin_rvv_vsll_vv_u16m2((vuint16m2_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vsll_vv_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vv_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vsll_vv_u16m4(op0, op1, op2) \ +__builtin_rvv_vsll_vv_u16m4((vuint16m4_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vsll_vv_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vv_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vsll_vv_u16m8(op0, op1, op2) \ +__builtin_rvv_vsll_vv_u16m8((vuint16m8_t)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vsll_vv_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vv_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vsll_vv_u16mf2(op0, op1, op2) \ +__builtin_rvv_vsll_vv_u16mf2((vuint16mf2_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vsll_vv_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vv_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vsll_vv_u16mf4(op0, op1, op2) \ +__builtin_rvv_vsll_vv_u16mf4((vuint16mf4_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vsll_vv_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vv_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vsll_vv_u32m1(op0, op1, op2) \ +__builtin_rvv_vsll_vv_u32m1((vuint32m1_t)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vsll_vv_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vv_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vsll_vv_u32m2(op0, op1, op2) \ +__builtin_rvv_vsll_vv_u32m2((vuint32m2_t)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vsll_vv_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vv_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vsll_vv_u32m4(op0, op1, op2) \ +__builtin_rvv_vsll_vv_u32m4((vuint32m4_t)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vsll_vv_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vv_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vsll_vv_u32m8(op0, op1, op2) \ +__builtin_rvv_vsll_vv_u32m8((vuint32m8_t)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vsll_vv_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vv_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vsll_vv_u32mf2(op0, op1, op2) \ +__builtin_rvv_vsll_vv_u32mf2((vuint32mf2_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vsll_vv_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vv_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vsll_vv_u64m1(op0, op1, op2) \ +__builtin_rvv_vsll_vv_u64m1((vuint64m1_t)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vsll_vv_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vv_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vsll_vv_u64m2(op0, op1, op2) \ +__builtin_rvv_vsll_vv_u64m2((vuint64m2_t)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vsll_vv_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vv_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vsll_vv_u64m4(op0, op1, op2) \ +__builtin_rvv_vsll_vv_u64m4((vuint64m4_t)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vsll_vv_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vv_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vsll_vv_u64m8(op0, op1, op2) \ +__builtin_rvv_vsll_vv_u64m8((vuint64m8_t)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vsll_vv_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vv_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vsll_vx_u8m1(op0, op1, op2) \ +__builtin_rvv_vsll_vx_u8m1((vuint8m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsll_vx_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vx_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsll_vx_u8m2(op0, op1, op2) \ +__builtin_rvv_vsll_vx_u8m2((vuint8m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsll_vx_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vx_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsll_vx_u8m4(op0, op1, op2) \ +__builtin_rvv_vsll_vx_u8m4((vuint8m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsll_vx_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vx_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsll_vx_u8m8(op0, op1, op2) \ +__builtin_rvv_vsll_vx_u8m8((vuint8m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsll_vx_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vx_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsll_vx_u8mf2(op0, op1, op2) \ +__builtin_rvv_vsll_vx_u8mf2((vuint8mf2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsll_vx_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vx_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsll_vx_u8mf4(op0, op1, op2) \ +__builtin_rvv_vsll_vx_u8mf4((vuint8mf4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsll_vx_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vx_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsll_vx_u8mf8(op0, op1, op2) \ +__builtin_rvv_vsll_vx_u8mf8((vuint8mf8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsll_vx_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vx_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsll_vx_u16m1(op0, op1, op2) \ +__builtin_rvv_vsll_vx_u16m1((vuint16m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsll_vx_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vx_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsll_vx_u16m2(op0, op1, op2) \ +__builtin_rvv_vsll_vx_u16m2((vuint16m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsll_vx_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vx_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsll_vx_u16m4(op0, op1, op2) \ +__builtin_rvv_vsll_vx_u16m4((vuint16m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsll_vx_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vx_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsll_vx_u16m8(op0, op1, op2) \ +__builtin_rvv_vsll_vx_u16m8((vuint16m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsll_vx_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vx_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsll_vx_u16mf2(op0, op1, op2) \ +__builtin_rvv_vsll_vx_u16mf2((vuint16mf2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsll_vx_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vx_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsll_vx_u16mf4(op0, op1, op2) \ +__builtin_rvv_vsll_vx_u16mf4((vuint16mf4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsll_vx_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vx_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsll_vx_u32m1(op0, op1, op2) \ +__builtin_rvv_vsll_vx_u32m1((vuint32m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsll_vx_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vx_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsll_vx_u32m2(op0, op1, op2) \ +__builtin_rvv_vsll_vx_u32m2((vuint32m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsll_vx_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vx_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsll_vx_u32m4(op0, op1, op2) \ +__builtin_rvv_vsll_vx_u32m4((vuint32m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsll_vx_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vx_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsll_vx_u32m8(op0, op1, op2) \ +__builtin_rvv_vsll_vx_u32m8((vuint32m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsll_vx_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vx_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsll_vx_u32mf2(op0, op1, op2) \ +__builtin_rvv_vsll_vx_u32mf2((vuint32mf2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsll_vx_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vx_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsll_vx_u64m1(op0, op1, op2) \ +__builtin_rvv_vsll_vx_u64m1((vuint64m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsll_vx_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vx_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsll_vx_u64m2(op0, op1, op2) \ +__builtin_rvv_vsll_vx_u64m2((vuint64m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsll_vx_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vx_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsll_vx_u64m4(op0, op1, op2) \ +__builtin_rvv_vsll_vx_u64m4((vuint64m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsll_vx_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vx_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsll_vx_u64m8(op0, op1, op2) \ +__builtin_rvv_vsll_vx_u64m8((vuint64m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsll_vx_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsll_vx_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsuxei16_v_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_u16m1((uint16_t *)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vsuxei16_v_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vsuxei16_v_u16m2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_u16m2((uint16_t *)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (size_t)(op3)) +#define vsuxei16_v_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_u16m2_m((vbool8_t)(op0), (uint16_t *)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vsuxei16_v_u16m4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_u16m4((uint16_t *)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (size_t)(op3)) +#define vsuxei16_v_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_u16m4_m((vbool4_t)(op0), (uint16_t *)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vsuxei16_v_u16m8(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_u16m8((uint16_t *)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (size_t)(op3)) +#define vsuxei16_v_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_u16m8_m((vbool2_t)(op0), (uint16_t *)(op1), (vuint16m8_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vsuxei16_v_u16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_u16mf2((uint16_t *)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (size_t)(op3)) +#define vsuxei16_v_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vsuxei16_v_u16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_u16mf4((uint16_t *)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (size_t)(op3)) +#define vsuxei16_v_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vsrl_vv_u8m1(op0, op1, op2) \ +__builtin_rvv_vsrl_vv_u8m1((vuint8m1_t)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vsrl_vv_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsrl_vv_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vsrl_vv_u8m2(op0, op1, op2) \ +__builtin_rvv_vsrl_vv_u8m2((vuint8m2_t)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vsrl_vv_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsrl_vv_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vsrl_vv_u8m4(op0, op1, op2) \ +__builtin_rvv_vsrl_vv_u8m4((vuint8m4_t)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vsrl_vv_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsrl_vv_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vsrl_vv_u8m8(op0, op1, op2) \ +__builtin_rvv_vsrl_vv_u8m8((vuint8m8_t)(op0), (vuint8m8_t)(op1), (size_t)(op2)) +#define vsrl_vv_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsrl_vv_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (vuint8m8_t)(op3), (size_t)(op4)) +#define vsrl_vv_u8mf2(op0, op1, op2) \ +__builtin_rvv_vsrl_vv_u8mf2((vuint8mf2_t)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vsrl_vv_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsrl_vv_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vsrl_vv_u8mf4(op0, op1, op2) \ +__builtin_rvv_vsrl_vv_u8mf4((vuint8mf4_t)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vsrl_vv_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsrl_vv_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vsrl_vv_u8mf8(op0, op1, op2) \ +__builtin_rvv_vsrl_vv_u8mf8((vuint8mf8_t)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vsrl_vv_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsrl_vv_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vsrl_vv_u16m1(op0, op1, op2) \ +__builtin_rvv_vsrl_vv_u16m1((vuint16m1_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vsrl_vv_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsrl_vv_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vsrl_vv_u16m2(op0, op1, op2) \ +__builtin_rvv_vsrl_vv_u16m2((vuint16m2_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vsrl_vv_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsrl_vv_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vsrl_vv_u16m4(op0, op1, op2) \ +__builtin_rvv_vsrl_vv_u16m4((vuint16m4_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vsrl_vv_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsrl_vv_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vsrl_vv_u16m8(op0, op1, op2) \ +__builtin_rvv_vsrl_vv_u16m8((vuint16m8_t)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vsrl_vv_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsrl_vv_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vsrl_vv_u16mf2(op0, op1, op2) \ +__builtin_rvv_vsrl_vv_u16mf2((vuint16mf2_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vsrl_vv_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsrl_vv_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vsrl_vv_u16mf4(op0, op1, op2) \ +__builtin_rvv_vsrl_vv_u16mf4((vuint16mf4_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vsrl_vv_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsrl_vv_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vsrl_vv_u32m1(op0, op1, op2) \ +__builtin_rvv_vsrl_vv_u32m1((vuint32m1_t)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vsrl_vv_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsrl_vv_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vsrl_vv_u32m2(op0, op1, op2) \ +__builtin_rvv_vsrl_vv_u32m2((vuint32m2_t)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vsrl_vv_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsrl_vv_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vsrl_vv_u32m4(op0, op1, op2) \ +__builtin_rvv_vsrl_vv_u32m4((vuint32m4_t)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vsrl_vv_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsrl_vv_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vsrl_vv_u32m8(op0, op1, op2) \ +__builtin_rvv_vsrl_vv_u32m8((vuint32m8_t)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vsrl_vv_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsrl_vv_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vsrl_vv_u32mf2(op0, op1, op2) \ +__builtin_rvv_vsrl_vv_u32mf2((vuint32mf2_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vsrl_vv_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsrl_vv_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vsrl_vv_u64m1(op0, op1, op2) \ +__builtin_rvv_vsrl_vv_u64m1((vuint64m1_t)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vsrl_vv_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsrl_vv_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vsrl_vv_u64m2(op0, op1, op2) \ +__builtin_rvv_vsrl_vv_u64m2((vuint64m2_t)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vsrl_vv_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsrl_vv_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vsrl_vv_u64m4(op0, op1, op2) \ +__builtin_rvv_vsrl_vv_u64m4((vuint64m4_t)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vsrl_vv_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsrl_vv_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vsrl_vv_u64m8(op0, op1, op2) \ +__builtin_rvv_vsrl_vv_u64m8((vuint64m8_t)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vsrl_vv_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsrl_vv_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vsrl_vx_u8m1(op0, op1, op2) \ +__builtin_rvv_vsrl_vx_u8m1((vuint8m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsrl_vx_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsrl_vx_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsrl_vx_u8m2(op0, op1, op2) \ +__builtin_rvv_vsrl_vx_u8m2((vuint8m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsrl_vx_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsrl_vx_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsrl_vx_u8m4(op0, op1, op2) \ +__builtin_rvv_vsrl_vx_u8m4((vuint8m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsrl_vx_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsrl_vx_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsrl_vx_u8m8(op0, op1, op2) \ +__builtin_rvv_vsrl_vx_u8m8((vuint8m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsrl_vx_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsrl_vx_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsrl_vx_u8mf2(op0, op1, op2) \ +__builtin_rvv_vsrl_vx_u8mf2((vuint8mf2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsrl_vx_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsrl_vx_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsrl_vx_u8mf4(op0, op1, op2) \ +__builtin_rvv_vsrl_vx_u8mf4((vuint8mf4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsrl_vx_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsrl_vx_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsrl_vx_u8mf8(op0, op1, op2) \ +__builtin_rvv_vsrl_vx_u8mf8((vuint8mf8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsrl_vx_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsrl_vx_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsrl_vx_u16m1(op0, op1, op2) \ +__builtin_rvv_vsrl_vx_u16m1((vuint16m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsrl_vx_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsrl_vx_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsrl_vx_u16m2(op0, op1, op2) \ +__builtin_rvv_vsrl_vx_u16m2((vuint16m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsrl_vx_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsrl_vx_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsrl_vx_u16m4(op0, op1, op2) \ +__builtin_rvv_vsrl_vx_u16m4((vuint16m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsrl_vx_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsrl_vx_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsrl_vx_u16m8(op0, op1, op2) \ +__builtin_rvv_vsrl_vx_u16m8((vuint16m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsrl_vx_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsrl_vx_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsrl_vx_u16mf2(op0, op1, op2) \ +__builtin_rvv_vsrl_vx_u16mf2((vuint16mf2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsrl_vx_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsrl_vx_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsrl_vx_u16mf4(op0, op1, op2) \ +__builtin_rvv_vsrl_vx_u16mf4((vuint16mf4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsrl_vx_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsrl_vx_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsrl_vx_u32m1(op0, op1, op2) \ +__builtin_rvv_vsrl_vx_u32m1((vuint32m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsrl_vx_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsrl_vx_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsrl_vx_u32m2(op0, op1, op2) \ +__builtin_rvv_vsrl_vx_u32m2((vuint32m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsrl_vx_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsrl_vx_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsrl_vx_u32m4(op0, op1, op2) \ +__builtin_rvv_vsrl_vx_u32m4((vuint32m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsrl_vx_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsrl_vx_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsrl_vx_u32m8(op0, op1, op2) \ +__builtin_rvv_vsrl_vx_u32m8((vuint32m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsrl_vx_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsrl_vx_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsrl_vx_u32mf2(op0, op1, op2) \ +__builtin_rvv_vsrl_vx_u32mf2((vuint32mf2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsrl_vx_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsrl_vx_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsrl_vx_u64m1(op0, op1, op2) \ +__builtin_rvv_vsrl_vx_u64m1((vuint64m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsrl_vx_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsrl_vx_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsrl_vx_u64m2(op0, op1, op2) \ +__builtin_rvv_vsrl_vx_u64m2((vuint64m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsrl_vx_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsrl_vx_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsrl_vx_u64m4(op0, op1, op2) \ +__builtin_rvv_vsrl_vx_u64m4((vuint64m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsrl_vx_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsrl_vx_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsrl_vx_u64m8(op0, op1, op2) \ +__builtin_rvv_vsrl_vx_u64m8((vuint64m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsrl_vx_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsrl_vx_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsra_vv_i8m1(op0, op1, op2) \ +__builtin_rvv_vsra_vv_i8m1((vint8m1_t)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vsra_vv_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsra_vv_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vsra_vv_i8m2(op0, op1, op2) \ +__builtin_rvv_vsra_vv_i8m2((vint8m2_t)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vsra_vv_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsra_vv_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vsra_vv_i8m4(op0, op1, op2) \ +__builtin_rvv_vsra_vv_i8m4((vint8m4_t)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vsra_vv_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsra_vv_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vsra_vv_i8m8(op0, op1, op2) \ +__builtin_rvv_vsra_vv_i8m8((vint8m8_t)(op0), (vuint8m8_t)(op1), (size_t)(op2)) +#define vsra_vv_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsra_vv_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (vuint8m8_t)(op3), (size_t)(op4)) +#define vsra_vv_i8mf2(op0, op1, op2) \ +__builtin_rvv_vsra_vv_i8mf2((vint8mf2_t)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vsra_vv_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsra_vv_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vsra_vv_i8mf4(op0, op1, op2) \ +__builtin_rvv_vsra_vv_i8mf4((vint8mf4_t)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vsra_vv_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsra_vv_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vsra_vv_i8mf8(op0, op1, op2) \ +__builtin_rvv_vsra_vv_i8mf8((vint8mf8_t)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vsra_vv_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsra_vv_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vsra_vv_i16m1(op0, op1, op2) \ +__builtin_rvv_vsra_vv_i16m1((vint16m1_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vsra_vv_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsra_vv_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vsra_vv_i16m2(op0, op1, op2) \ +__builtin_rvv_vsra_vv_i16m2((vint16m2_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vsra_vv_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsra_vv_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vsra_vv_i16m4(op0, op1, op2) \ +__builtin_rvv_vsra_vv_i16m4((vint16m4_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vsra_vv_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsra_vv_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vsra_vv_i16m8(op0, op1, op2) \ +__builtin_rvv_vsra_vv_i16m8((vint16m8_t)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vsra_vv_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsra_vv_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vsra_vv_i16mf2(op0, op1, op2) \ +__builtin_rvv_vsra_vv_i16mf2((vint16mf2_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vsra_vv_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsra_vv_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vsra_vv_i16mf4(op0, op1, op2) \ +__builtin_rvv_vsra_vv_i16mf4((vint16mf4_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vsra_vv_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsra_vv_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vsra_vv_i32m1(op0, op1, op2) \ +__builtin_rvv_vsra_vv_i32m1((vint32m1_t)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vsra_vv_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsra_vv_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vsra_vv_i32m2(op0, op1, op2) \ +__builtin_rvv_vsra_vv_i32m2((vint32m2_t)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vsra_vv_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsra_vv_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vsra_vv_i32m4(op0, op1, op2) \ +__builtin_rvv_vsra_vv_i32m4((vint32m4_t)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vsra_vv_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsra_vv_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vsra_vv_i32m8(op0, op1, op2) \ +__builtin_rvv_vsra_vv_i32m8((vint32m8_t)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vsra_vv_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsra_vv_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vsra_vv_i32mf2(op0, op1, op2) \ +__builtin_rvv_vsra_vv_i32mf2((vint32mf2_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vsra_vv_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsra_vv_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vsra_vv_i64m1(op0, op1, op2) \ +__builtin_rvv_vsra_vv_i64m1((vint64m1_t)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vsra_vv_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsra_vv_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vsra_vv_i64m2(op0, op1, op2) \ +__builtin_rvv_vsra_vv_i64m2((vint64m2_t)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vsra_vv_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsra_vv_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vsra_vv_i64m4(op0, op1, op2) \ +__builtin_rvv_vsra_vv_i64m4((vint64m4_t)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vsra_vv_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsra_vv_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vsra_vv_i64m8(op0, op1, op2) \ +__builtin_rvv_vsra_vv_i64m8((vint64m8_t)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vsra_vv_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsra_vv_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vsra_vx_i8m1(op0, op1, op2) \ +__builtin_rvv_vsra_vx_i8m1((vint8m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsra_vx_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsra_vx_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsra_vx_i8m2(op0, op1, op2) \ +__builtin_rvv_vsra_vx_i8m2((vint8m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsra_vx_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsra_vx_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsra_vx_i8m4(op0, op1, op2) \ +__builtin_rvv_vsra_vx_i8m4((vint8m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsra_vx_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsra_vx_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsra_vx_i8m8(op0, op1, op2) \ +__builtin_rvv_vsra_vx_i8m8((vint8m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsra_vx_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsra_vx_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsra_vx_i8mf2(op0, op1, op2) \ +__builtin_rvv_vsra_vx_i8mf2((vint8mf2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsra_vx_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsra_vx_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsra_vx_i8mf4(op0, op1, op2) \ +__builtin_rvv_vsra_vx_i8mf4((vint8mf4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsra_vx_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsra_vx_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsra_vx_i8mf8(op0, op1, op2) \ +__builtin_rvv_vsra_vx_i8mf8((vint8mf8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsra_vx_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsra_vx_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsra_vx_i16m1(op0, op1, op2) \ +__builtin_rvv_vsra_vx_i16m1((vint16m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsra_vx_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsra_vx_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsra_vx_i16m2(op0, op1, op2) \ +__builtin_rvv_vsra_vx_i16m2((vint16m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsra_vx_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsra_vx_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsra_vx_i16m4(op0, op1, op2) \ +__builtin_rvv_vsra_vx_i16m4((vint16m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsra_vx_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsra_vx_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsra_vx_i16m8(op0, op1, op2) \ +__builtin_rvv_vsra_vx_i16m8((vint16m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsra_vx_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsra_vx_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsra_vx_i16mf2(op0, op1, op2) \ +__builtin_rvv_vsra_vx_i16mf2((vint16mf2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsra_vx_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsra_vx_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsra_vx_i16mf4(op0, op1, op2) \ +__builtin_rvv_vsra_vx_i16mf4((vint16mf4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsra_vx_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsra_vx_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsra_vx_i32m1(op0, op1, op2) \ +__builtin_rvv_vsra_vx_i32m1((vint32m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsra_vx_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsra_vx_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsra_vx_i32m2(op0, op1, op2) \ +__builtin_rvv_vsra_vx_i32m2((vint32m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsra_vx_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsra_vx_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsra_vx_i32m4(op0, op1, op2) \ +__builtin_rvv_vsra_vx_i32m4((vint32m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsra_vx_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsra_vx_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsra_vx_i32m8(op0, op1, op2) \ +__builtin_rvv_vsra_vx_i32m8((vint32m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsra_vx_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsra_vx_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsra_vx_i32mf2(op0, op1, op2) \ +__builtin_rvv_vsra_vx_i32mf2((vint32mf2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsra_vx_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsra_vx_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsra_vx_i64m1(op0, op1, op2) \ +__builtin_rvv_vsra_vx_i64m1((vint64m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsra_vx_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsra_vx_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsra_vx_i64m2(op0, op1, op2) \ +__builtin_rvv_vsra_vx_i64m2((vint64m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsra_vx_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsra_vx_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsra_vx_i64m4(op0, op1, op2) \ +__builtin_rvv_vsra_vx_i64m4((vint64m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsra_vx_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsra_vx_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsra_vx_i64m8(op0, op1, op2) \ +__builtin_rvv_vsra_vx_i64m8((vint64m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vsra_vx_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsra_vx_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnsrl_wv_u8m1(op0, op1, op2) \ +__builtin_rvv_vnsrl_wv_u8m1((vuint16m2_t)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vnsrl_wv_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsrl_wv_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint16m2_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vnsrl_wv_u8m2(op0, op1, op2) \ +__builtin_rvv_vnsrl_wv_u8m2((vuint16m4_t)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vnsrl_wv_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsrl_wv_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint16m4_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vnsrl_wv_u8m4(op0, op1, op2) \ +__builtin_rvv_vnsrl_wv_u8m4((vuint16m8_t)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vnsrl_wv_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsrl_wv_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint16m8_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vnsrl_wv_u8mf2(op0, op1, op2) \ +__builtin_rvv_vnsrl_wv_u8mf2((vuint16m1_t)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vnsrl_wv_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsrl_wv_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint16m1_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vnsrl_wv_u8mf4(op0, op1, op2) \ +__builtin_rvv_vnsrl_wv_u8mf4((vuint16mf2_t)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vnsrl_wv_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsrl_wv_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint16mf2_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vnsrl_wv_u8mf8(op0, op1, op2) \ +__builtin_rvv_vnsrl_wv_u8mf8((vuint16mf4_t)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vnsrl_wv_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsrl_wv_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint16mf4_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vnsrl_wv_u16m1(op0, op1, op2) \ +__builtin_rvv_vnsrl_wv_u16m1((vuint32m2_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vnsrl_wv_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsrl_wv_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint32m2_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vnsrl_wv_u16m2(op0, op1, op2) \ +__builtin_rvv_vnsrl_wv_u16m2((vuint32m4_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vnsrl_wv_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsrl_wv_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint32m4_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vnsrl_wv_u16m4(op0, op1, op2) \ +__builtin_rvv_vnsrl_wv_u16m4((vuint32m8_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vnsrl_wv_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsrl_wv_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint32m8_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vnsrl_wv_u16mf2(op0, op1, op2) \ +__builtin_rvv_vnsrl_wv_u16mf2((vuint32m1_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vnsrl_wv_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsrl_wv_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint32m1_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vnsrl_wv_u16mf4(op0, op1, op2) \ +__builtin_rvv_vnsrl_wv_u16mf4((vuint32mf2_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vnsrl_wv_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsrl_wv_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint32mf2_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vnsrl_wv_u32m1(op0, op1, op2) \ +__builtin_rvv_vnsrl_wv_u32m1((vuint64m2_t)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vnsrl_wv_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsrl_wv_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint64m2_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vnsrl_wv_u32m2(op0, op1, op2) \ +__builtin_rvv_vnsrl_wv_u32m2((vuint64m4_t)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vnsrl_wv_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsrl_wv_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint64m4_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vnsrl_wv_u32m4(op0, op1, op2) \ +__builtin_rvv_vnsrl_wv_u32m4((vuint64m8_t)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vnsrl_wv_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsrl_wv_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint64m8_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vnsrl_wv_u32mf2(op0, op1, op2) \ +__builtin_rvv_vnsrl_wv_u32mf2((vuint64m1_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vnsrl_wv_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsrl_wv_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint64m1_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vnsrl_wx_u8m1(op0, op1, op2) \ +__builtin_rvv_vnsrl_wx_u8m1((vuint16m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnsrl_wx_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsrl_wx_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint16m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnsrl_wx_u8m2(op0, op1, op2) \ +__builtin_rvv_vnsrl_wx_u8m2((vuint16m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnsrl_wx_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsrl_wx_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint16m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnsrl_wx_u8m4(op0, op1, op2) \ +__builtin_rvv_vnsrl_wx_u8m4((vuint16m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnsrl_wx_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsrl_wx_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint16m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnsrl_wx_u8mf2(op0, op1, op2) \ +__builtin_rvv_vnsrl_wx_u8mf2((vuint16m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnsrl_wx_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsrl_wx_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint16m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnsrl_wx_u8mf4(op0, op1, op2) \ +__builtin_rvv_vnsrl_wx_u8mf4((vuint16mf2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnsrl_wx_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsrl_wx_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint16mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnsrl_wx_u8mf8(op0, op1, op2) \ +__builtin_rvv_vnsrl_wx_u8mf8((vuint16mf4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnsrl_wx_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsrl_wx_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint16mf4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnsrl_wx_u16m1(op0, op1, op2) \ +__builtin_rvv_vnsrl_wx_u16m1((vuint32m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnsrl_wx_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsrl_wx_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint32m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnsrl_wx_u16m2(op0, op1, op2) \ +__builtin_rvv_vnsrl_wx_u16m2((vuint32m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnsrl_wx_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsrl_wx_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint32m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnsrl_wx_u16m4(op0, op1, op2) \ +__builtin_rvv_vnsrl_wx_u16m4((vuint32m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnsrl_wx_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsrl_wx_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint32m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnsrl_wx_u16mf2(op0, op1, op2) \ +__builtin_rvv_vnsrl_wx_u16mf2((vuint32m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnsrl_wx_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsrl_wx_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint32m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnsrl_wx_u16mf4(op0, op1, op2) \ +__builtin_rvv_vnsrl_wx_u16mf4((vuint32mf2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnsrl_wx_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsrl_wx_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnsrl_wx_u32m1(op0, op1, op2) \ +__builtin_rvv_vnsrl_wx_u32m1((vuint64m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnsrl_wx_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsrl_wx_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint64m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnsrl_wx_u32m2(op0, op1, op2) \ +__builtin_rvv_vnsrl_wx_u32m2((vuint64m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnsrl_wx_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsrl_wx_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint64m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnsrl_wx_u32m4(op0, op1, op2) \ +__builtin_rvv_vnsrl_wx_u32m4((vuint64m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnsrl_wx_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsrl_wx_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint64m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnsrl_wx_u32mf2(op0, op1, op2) \ +__builtin_rvv_vnsrl_wx_u32mf2((vuint64m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnsrl_wx_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsrl_wx_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint64m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnsra_wv_i8m1(op0, op1, op2) \ +__builtin_rvv_vnsra_wv_i8m1((vint16m2_t)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vnsra_wv_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsra_wv_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint16m2_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vnsra_wv_i8m2(op0, op1, op2) \ +__builtin_rvv_vnsra_wv_i8m2((vint16m4_t)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vnsra_wv_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsra_wv_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint16m4_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vnsra_wv_i8m4(op0, op1, op2) \ +__builtin_rvv_vnsra_wv_i8m4((vint16m8_t)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vnsra_wv_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsra_wv_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint16m8_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vnsra_wv_i8mf2(op0, op1, op2) \ +__builtin_rvv_vnsra_wv_i8mf2((vint16m1_t)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vnsra_wv_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsra_wv_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint16m1_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vnsra_wv_i8mf4(op0, op1, op2) \ +__builtin_rvv_vnsra_wv_i8mf4((vint16mf2_t)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vnsra_wv_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsra_wv_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint16mf2_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vnsra_wv_i8mf8(op0, op1, op2) \ +__builtin_rvv_vnsra_wv_i8mf8((vint16mf4_t)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vnsra_wv_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsra_wv_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint16mf4_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vnsra_wv_i16m1(op0, op1, op2) \ +__builtin_rvv_vnsra_wv_i16m1((vint32m2_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vnsra_wv_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsra_wv_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint32m2_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vnsra_wv_i16m2(op0, op1, op2) \ +__builtin_rvv_vnsra_wv_i16m2((vint32m4_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vnsra_wv_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsra_wv_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint32m4_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vnsra_wv_i16m4(op0, op1, op2) \ +__builtin_rvv_vnsra_wv_i16m4((vint32m8_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vnsra_wv_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsra_wv_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint32m8_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vnsra_wv_i16mf2(op0, op1, op2) \ +__builtin_rvv_vnsra_wv_i16mf2((vint32m1_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vnsra_wv_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsra_wv_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint32m1_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vnsra_wv_i16mf4(op0, op1, op2) \ +__builtin_rvv_vnsra_wv_i16mf4((vint32mf2_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vnsra_wv_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsra_wv_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint32mf2_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vnsra_wv_i32m1(op0, op1, op2) \ +__builtin_rvv_vnsra_wv_i32m1((vint64m2_t)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vnsra_wv_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsra_wv_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint64m2_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vnsra_wv_i32m2(op0, op1, op2) \ +__builtin_rvv_vnsra_wv_i32m2((vint64m4_t)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vnsra_wv_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsra_wv_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint64m4_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vnsra_wv_i32m4(op0, op1, op2) \ +__builtin_rvv_vnsra_wv_i32m4((vint64m8_t)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vnsra_wv_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsra_wv_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint64m8_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vnsra_wv_i32mf2(op0, op1, op2) \ +__builtin_rvv_vnsra_wv_i32mf2((vint64m1_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vnsra_wv_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsra_wv_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint64m1_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vnsra_wx_i8m1(op0, op1, op2) \ +__builtin_rvv_vnsra_wx_i8m1((vint16m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnsra_wx_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsra_wx_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint16m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnsra_wx_i8m2(op0, op1, op2) \ +__builtin_rvv_vnsra_wx_i8m2((vint16m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnsra_wx_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsra_wx_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint16m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnsra_wx_i8m4(op0, op1, op2) \ +__builtin_rvv_vnsra_wx_i8m4((vint16m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnsra_wx_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsra_wx_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint16m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnsra_wx_i8mf2(op0, op1, op2) \ +__builtin_rvv_vnsra_wx_i8mf2((vint16m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnsra_wx_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsra_wx_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint16m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnsra_wx_i8mf4(op0, op1, op2) \ +__builtin_rvv_vnsra_wx_i8mf4((vint16mf2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnsra_wx_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsra_wx_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint16mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnsra_wx_i8mf8(op0, op1, op2) \ +__builtin_rvv_vnsra_wx_i8mf8((vint16mf4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnsra_wx_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsra_wx_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint16mf4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnsra_wx_i16m1(op0, op1, op2) \ +__builtin_rvv_vnsra_wx_i16m1((vint32m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnsra_wx_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsra_wx_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint32m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnsra_wx_i16m2(op0, op1, op2) \ +__builtin_rvv_vnsra_wx_i16m2((vint32m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnsra_wx_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsra_wx_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint32m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnsra_wx_i16m4(op0, op1, op2) \ +__builtin_rvv_vnsra_wx_i16m4((vint32m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnsra_wx_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsra_wx_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint32m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnsra_wx_i16mf2(op0, op1, op2) \ +__builtin_rvv_vnsra_wx_i16mf2((vint32m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnsra_wx_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsra_wx_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint32m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnsra_wx_i16mf4(op0, op1, op2) \ +__builtin_rvv_vnsra_wx_i16mf4((vint32mf2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnsra_wx_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsra_wx_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint32mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnsra_wx_i32m1(op0, op1, op2) \ +__builtin_rvv_vnsra_wx_i32m1((vint64m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnsra_wx_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsra_wx_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint64m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnsra_wx_i32m2(op0, op1, op2) \ +__builtin_rvv_vnsra_wx_i32m2((vint64m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnsra_wx_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsra_wx_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint64m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnsra_wx_i32m4(op0, op1, op2) \ +__builtin_rvv_vnsra_wx_i32m4((vint64m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnsra_wx_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsra_wx_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint64m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnsra_wx_i32mf2(op0, op1, op2) \ +__builtin_rvv_vnsra_wx_i32mf2((vint64m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnsra_wx_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnsra_wx_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint64m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vncvt_x_x_w_u8m1(op0, op1) \ +__builtin_rvv_vncvt_x_x_w_u8m1((vuint16m2_t)(op0), (size_t)(op1)) +#define vncvt_x_x_w_u8m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vncvt_x_x_w_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint16m2_t)(op2), (size_t)(op3)) +#define vncvt_x_x_w_u8m2(op0, op1) \ +__builtin_rvv_vncvt_x_x_w_u8m2((vuint16m4_t)(op0), (size_t)(op1)) +#define vncvt_x_x_w_u8m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vncvt_x_x_w_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint16m4_t)(op2), (size_t)(op3)) +#define vncvt_x_x_w_u8m4(op0, op1) \ +__builtin_rvv_vncvt_x_x_w_u8m4((vuint16m8_t)(op0), (size_t)(op1)) +#define vncvt_x_x_w_u8m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vncvt_x_x_w_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint16m8_t)(op2), (size_t)(op3)) +#define vncvt_x_x_w_u8mf2(op0, op1) \ +__builtin_rvv_vncvt_x_x_w_u8mf2((vuint16m1_t)(op0), (size_t)(op1)) +#define vncvt_x_x_w_u8mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vncvt_x_x_w_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vncvt_x_x_w_u8mf4(op0, op1) \ +__builtin_rvv_vncvt_x_x_w_u8mf4((vuint16mf2_t)(op0), (size_t)(op1)) +#define vncvt_x_x_w_u8mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vncvt_x_x_w_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint16mf2_t)(op2), (size_t)(op3)) +#define vncvt_x_x_w_u8mf8(op0, op1) \ +__builtin_rvv_vncvt_x_x_w_u8mf8((vuint16mf4_t)(op0), (size_t)(op1)) +#define vncvt_x_x_w_u8mf8_m(op0, op1, op2, op3) \ +__builtin_rvv_vncvt_x_x_w_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint16mf4_t)(op2), (size_t)(op3)) +#define vncvt_x_x_w_u16m1(op0, op1) \ +__builtin_rvv_vncvt_x_x_w_u16m1((vuint32m2_t)(op0), (size_t)(op1)) +#define vncvt_x_x_w_u16m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vncvt_x_x_w_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vncvt_x_x_w_u16m2(op0, op1) \ +__builtin_rvv_vncvt_x_x_w_u16m2((vuint32m4_t)(op0), (size_t)(op1)) +#define vncvt_x_x_w_u16m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vncvt_x_x_w_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vncvt_x_x_w_u16m4(op0, op1) \ +__builtin_rvv_vncvt_x_x_w_u16m4((vuint32m8_t)(op0), (size_t)(op1)) +#define vncvt_x_x_w_u16m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vncvt_x_x_w_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint32m8_t)(op2), (size_t)(op3)) +#define vncvt_x_x_w_u16mf2(op0, op1) \ +__builtin_rvv_vncvt_x_x_w_u16mf2((vuint32m1_t)(op0), (size_t)(op1)) +#define vncvt_x_x_w_u16mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vncvt_x_x_w_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vncvt_x_x_w_u16mf4(op0, op1) \ +__builtin_rvv_vncvt_x_x_w_u16mf4((vuint32mf2_t)(op0), (size_t)(op1)) +#define vncvt_x_x_w_u16mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vncvt_x_x_w_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vncvt_x_x_w_u32m1(op0, op1) \ +__builtin_rvv_vncvt_x_x_w_u32m1((vuint64m2_t)(op0), (size_t)(op1)) +#define vncvt_x_x_w_u32m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vncvt_x_x_w_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vncvt_x_x_w_u32m2(op0, op1) \ +__builtin_rvv_vncvt_x_x_w_u32m2((vuint64m4_t)(op0), (size_t)(op1)) +#define vncvt_x_x_w_u32m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vncvt_x_x_w_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vncvt_x_x_w_u32m4(op0, op1) \ +__builtin_rvv_vncvt_x_x_w_u32m4((vuint64m8_t)(op0), (size_t)(op1)) +#define vncvt_x_x_w_u32m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vncvt_x_x_w_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vncvt_x_x_w_u32mf2(op0, op1) \ +__builtin_rvv_vncvt_x_x_w_u32mf2((vuint64m1_t)(op0), (size_t)(op1)) +#define vncvt_x_x_w_u32mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vncvt_x_x_w_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vmseq_vv_i8m1_b8(op0, op1, op2) \ +__builtin_rvv_vmseq_vv_i8m1_b8((vint8m1_t)(op0), (vint8m1_t)(op1), (size_t)(op2)) +#define vmseq_vv_i8m1_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vv_i8m1_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vmseq_vv_i8m2_b4(op0, op1, op2) \ +__builtin_rvv_vmseq_vv_i8m2_b4((vint8m2_t)(op0), (vint8m2_t)(op1), (size_t)(op2)) +#define vmseq_vv_i8m2_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vv_i8m2_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vmseq_vv_i8m4_b2(op0, op1, op2) \ +__builtin_rvv_vmseq_vv_i8m4_b2((vint8m4_t)(op0), (vint8m4_t)(op1), (size_t)(op2)) +#define vmseq_vv_i8m4_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vv_i8m4_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vint8m4_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vmseq_vv_i8m8_b1(op0, op1, op2) \ +__builtin_rvv_vmseq_vv_i8m8_b1((vint8m8_t)(op0), (vint8m8_t)(op1), (size_t)(op2)) +#define vmseq_vv_i8m8_b1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vv_i8m8_b1_m((vbool1_t)(op0), (vbool1_t)(op1), (vint8m8_t)(op2), (vint8m8_t)(op3), (size_t)(op4)) +#define vmseq_vv_i8mf2_b16(op0, op1, op2) \ +__builtin_rvv_vmseq_vv_i8mf2_b16((vint8mf2_t)(op0), (vint8mf2_t)(op1), (size_t)(op2)) +#define vmseq_vv_i8mf2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vv_i8mf2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vmseq_vv_i8mf4_b32(op0, op1, op2) \ +__builtin_rvv_vmseq_vv_i8mf4_b32((vint8mf4_t)(op0), (vint8mf4_t)(op1), (size_t)(op2)) +#define vmseq_vv_i8mf4_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vv_i8mf4_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vmseq_vv_i8mf8_b64(op0, op1, op2) \ +__builtin_rvv_vmseq_vv_i8mf8_b64((vint8mf8_t)(op0), (vint8mf8_t)(op1), (size_t)(op2)) +#define vmseq_vv_i8mf8_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vv_i8mf8_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vmseq_vv_i16m1_b16(op0, op1, op2) \ +__builtin_rvv_vmseq_vv_i16m1_b16((vint16m1_t)(op0), (vint16m1_t)(op1), (size_t)(op2)) +#define vmseq_vv_i16m1_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vv_i16m1_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vmseq_vv_i16m2_b8(op0, op1, op2) \ +__builtin_rvv_vmseq_vv_i16m2_b8((vint16m2_t)(op0), (vint16m2_t)(op1), (size_t)(op2)) +#define vmseq_vv_i16m2_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vv_i16m2_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vmseq_vv_i16m4_b4(op0, op1, op2) \ +__builtin_rvv_vmseq_vv_i16m4_b4((vint16m4_t)(op0), (vint16m4_t)(op1), (size_t)(op2)) +#define vmseq_vv_i16m4_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vv_i16m4_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vint16m4_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vmseq_vv_i16m8_b2(op0, op1, op2) \ +__builtin_rvv_vmseq_vv_i16m8_b2((vint16m8_t)(op0), (vint16m8_t)(op1), (size_t)(op2)) +#define vmseq_vv_i16m8_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vv_i16m8_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vint16m8_t)(op2), (vint16m8_t)(op3), (size_t)(op4)) +#define vmseq_vv_i16mf2_b32(op0, op1, op2) \ +__builtin_rvv_vmseq_vv_i16mf2_b32((vint16mf2_t)(op0), (vint16mf2_t)(op1), (size_t)(op2)) +#define vmseq_vv_i16mf2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vv_i16mf2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vmseq_vv_i16mf4_b64(op0, op1, op2) \ +__builtin_rvv_vmseq_vv_i16mf4_b64((vint16mf4_t)(op0), (vint16mf4_t)(op1), (size_t)(op2)) +#define vmseq_vv_i16mf4_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vv_i16mf4_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vmseq_vv_i32m1_b32(op0, op1, op2) \ +__builtin_rvv_vmseq_vv_i32m1_b32((vint32m1_t)(op0), (vint32m1_t)(op1), (size_t)(op2)) +#define vmseq_vv_i32m1_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vv_i32m1_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vmseq_vv_i32m2_b16(op0, op1, op2) \ +__builtin_rvv_vmseq_vv_i32m2_b16((vint32m2_t)(op0), (vint32m2_t)(op1), (size_t)(op2)) +#define vmseq_vv_i32m2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vv_i32m2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vmseq_vv_i32m4_b8(op0, op1, op2) \ +__builtin_rvv_vmseq_vv_i32m4_b8((vint32m4_t)(op0), (vint32m4_t)(op1), (size_t)(op2)) +#define vmseq_vv_i32m4_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vv_i32m4_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vmseq_vv_i32m8_b4(op0, op1, op2) \ +__builtin_rvv_vmseq_vv_i32m8_b4((vint32m8_t)(op0), (vint32m8_t)(op1), (size_t)(op2)) +#define vmseq_vv_i32m8_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vv_i32m8_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vint32m8_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vmseq_vv_i32mf2_b64(op0, op1, op2) \ +__builtin_rvv_vmseq_vv_i32mf2_b64((vint32mf2_t)(op0), (vint32mf2_t)(op1), (size_t)(op2)) +#define vmseq_vv_i32mf2_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vv_i32mf2_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vmseq_vv_i64m1_b64(op0, op1, op2) \ +__builtin_rvv_vmseq_vv_i64m1_b64((vint64m1_t)(op0), (vint64m1_t)(op1), (size_t)(op2)) +#define vmseq_vv_i64m1_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vv_i64m1_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vmseq_vv_i64m2_b32(op0, op1, op2) \ +__builtin_rvv_vmseq_vv_i64m2_b32((vint64m2_t)(op0), (vint64m2_t)(op1), (size_t)(op2)) +#define vmseq_vv_i64m2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vv_i64m2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vmseq_vv_i64m4_b16(op0, op1, op2) \ +__builtin_rvv_vmseq_vv_i64m4_b16((vint64m4_t)(op0), (vint64m4_t)(op1), (size_t)(op2)) +#define vmseq_vv_i64m4_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vv_i64m4_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint64m4_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vmseq_vv_i64m8_b8(op0, op1, op2) \ +__builtin_rvv_vmseq_vv_i64m8_b8((vint64m8_t)(op0), (vint64m8_t)(op1), (size_t)(op2)) +#define vmseq_vv_i64m8_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vv_i64m8_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint64m8_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vsse8_v_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vsse8_v_u8m1((uint8_t *)(op0), (ptrdiff_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vsse8_v_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse8_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (ptrdiff_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vsse8_v_u8m2(op0, op1, op2, op3) \ +__builtin_rvv_vsse8_v_u8m2((uint8_t *)(op0), (ptrdiff_t)(op1), (vuint8m2_t)(op2), (size_t)(op3)) +#define vsse8_v_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse8_v_u8m2_m((vbool4_t)(op0), (uint8_t *)(op1), (ptrdiff_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vsse8_v_u8m4(op0, op1, op2, op3) \ +__builtin_rvv_vsse8_v_u8m4((uint8_t *)(op0), (ptrdiff_t)(op1), (vuint8m4_t)(op2), (size_t)(op3)) +#define vsse8_v_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse8_v_u8m4_m((vbool2_t)(op0), (uint8_t *)(op1), (ptrdiff_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vsse8_v_u8m8(op0, op1, op2, op3) \ +__builtin_rvv_vsse8_v_u8m8((uint8_t *)(op0), (ptrdiff_t)(op1), (vuint8m8_t)(op2), (size_t)(op3)) +#define vsse8_v_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse8_v_u8m8_m((vbool1_t)(op0), (uint8_t *)(op1), (ptrdiff_t)(op2), (vuint8m8_t)(op3), (size_t)(op4)) +#define vsse8_v_u8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsse8_v_u8mf2((uint8_t *)(op0), (ptrdiff_t)(op1), (vuint8mf2_t)(op2), (size_t)(op3)) +#define vsse8_v_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse8_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (ptrdiff_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vsse8_v_u8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsse8_v_u8mf4((uint8_t *)(op0), (ptrdiff_t)(op1), (vuint8mf4_t)(op2), (size_t)(op3)) +#define vsse8_v_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse8_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (ptrdiff_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vsse8_v_u8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vsse8_v_u8mf8((uint8_t *)(op0), (ptrdiff_t)(op1), (vuint8mf8_t)(op2), (size_t)(op3)) +#define vsse8_v_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse8_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (ptrdiff_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vsuxei32_v_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_i16m1((int16_t *)(op0), (vuint32m2_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vsuxei32_v_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint32m2_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vsuxei32_v_i16m2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_i16m2((int16_t *)(op0), (vuint32m4_t)(op1), (vint16m2_t)(op2), (size_t)(op3)) +#define vsuxei32_v_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_i16m2_m((vbool8_t)(op0), (int16_t *)(op1), (vuint32m4_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vsuxei32_v_i16m4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_i16m4((int16_t *)(op0), (vuint32m8_t)(op1), (vint16m4_t)(op2), (size_t)(op3)) +#define vsuxei32_v_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_i16m4_m((vbool4_t)(op0), (int16_t *)(op1), (vuint32m8_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vsuxei32_v_i16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_i16mf2((int16_t *)(op0), (vuint32m1_t)(op1), (vint16mf2_t)(op2), (size_t)(op3)) +#define vsuxei32_v_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint32m1_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vsuxei32_v_i16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_i16mf4((int16_t *)(op0), (vuint32mf2_t)(op1), (vint16mf4_t)(op2), (size_t)(op3)) +#define vsuxei32_v_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint32mf2_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vmseq_vx_i8m1_b8(op0, op1, op2) \ +__builtin_rvv_vmseq_vx_i8m1_b8((vint8m1_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmseq_vx_i8m1_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vx_i8m1_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint8m1_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmseq_vx_i8m2_b4(op0, op1, op2) \ +__builtin_rvv_vmseq_vx_i8m2_b4((vint8m2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmseq_vx_i8m2_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vx_i8m2_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vint8m2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmseq_vx_i8m4_b2(op0, op1, op2) \ +__builtin_rvv_vmseq_vx_i8m4_b2((vint8m4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmseq_vx_i8m4_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vx_i8m4_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vint8m4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmseq_vx_i8m8_b1(op0, op1, op2) \ +__builtin_rvv_vmseq_vx_i8m8_b1((vint8m8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmseq_vx_i8m8_b1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vx_i8m8_b1_m((vbool1_t)(op0), (vbool1_t)(op1), (vint8m8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmseq_vx_i8mf2_b16(op0, op1, op2) \ +__builtin_rvv_vmseq_vx_i8mf2_b16((vint8mf2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmseq_vx_i8mf2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vx_i8mf2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint8mf2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmseq_vx_i8mf4_b32(op0, op1, op2) \ +__builtin_rvv_vmseq_vx_i8mf4_b32((vint8mf4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmseq_vx_i8mf4_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vx_i8mf4_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint8mf4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmseq_vx_i8mf8_b64(op0, op1, op2) \ +__builtin_rvv_vmseq_vx_i8mf8_b64((vint8mf8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmseq_vx_i8mf8_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vx_i8mf8_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint8mf8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmseq_vx_i16m1_b16(op0, op1, op2) \ +__builtin_rvv_vmseq_vx_i16m1_b16((vint16m1_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmseq_vx_i16m1_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vx_i16m1_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint16m1_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmseq_vx_i16m2_b8(op0, op1, op2) \ +__builtin_rvv_vmseq_vx_i16m2_b8((vint16m2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmseq_vx_i16m2_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vx_i16m2_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint16m2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmseq_vx_i16m4_b4(op0, op1, op2) \ +__builtin_rvv_vmseq_vx_i16m4_b4((vint16m4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmseq_vx_i16m4_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vx_i16m4_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vint16m4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmseq_vx_i16m8_b2(op0, op1, op2) \ +__builtin_rvv_vmseq_vx_i16m8_b2((vint16m8_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmseq_vx_i16m8_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vx_i16m8_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vint16m8_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmseq_vx_i16mf2_b32(op0, op1, op2) \ +__builtin_rvv_vmseq_vx_i16mf2_b32((vint16mf2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmseq_vx_i16mf2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vx_i16mf2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint16mf2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmseq_vx_i16mf4_b64(op0, op1, op2) \ +__builtin_rvv_vmseq_vx_i16mf4_b64((vint16mf4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmseq_vx_i16mf4_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vx_i16mf4_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint16mf4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmseq_vx_i32m1_b32(op0, op1, op2) \ +__builtin_rvv_vmseq_vx_i32m1_b32((vint32m1_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmseq_vx_i32m1_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vx_i32m1_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint32m1_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmseq_vx_i32m2_b16(op0, op1, op2) \ +__builtin_rvv_vmseq_vx_i32m2_b16((vint32m2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmseq_vx_i32m2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vx_i32m2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint32m2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmseq_vx_i32m4_b8(op0, op1, op2) \ +__builtin_rvv_vmseq_vx_i32m4_b8((vint32m4_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmseq_vx_i32m4_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vx_i32m4_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint32m4_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmseq_vx_i32m8_b4(op0, op1, op2) \ +__builtin_rvv_vmseq_vx_i32m8_b4((vint32m8_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmseq_vx_i32m8_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vx_i32m8_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vint32m8_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmseq_vx_i32mf2_b64(op0, op1, op2) \ +__builtin_rvv_vmseq_vx_i32mf2_b64((vint32mf2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmseq_vx_i32mf2_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vx_i32mf2_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint32mf2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmseq_vx_i64m1_b64(op0, op1, op2) \ +__builtin_rvv_vmseq_vx_i64m1_b64((vint64m1_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmseq_vx_i64m1_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vx_i64m1_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint64m1_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vmseq_vx_i64m2_b32(op0, op1, op2) \ +__builtin_rvv_vmseq_vx_i64m2_b32((vint64m2_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmseq_vx_i64m2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vx_i64m2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint64m2_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vmseq_vx_i64m4_b16(op0, op1, op2) \ +__builtin_rvv_vmseq_vx_i64m4_b16((vint64m4_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmseq_vx_i64m4_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vx_i64m4_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint64m4_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vmseq_vx_i64m8_b8(op0, op1, op2) \ +__builtin_rvv_vmseq_vx_i64m8_b8((vint64m8_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmseq_vx_i64m8_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vx_i64m8_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint64m8_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vmseq_vv_u8m1_b8(op0, op1, op2) \ +__builtin_rvv_vmseq_vv_u8m1_b8((vuint8m1_t)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vmseq_vv_u8m1_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vv_u8m1_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vmseq_vv_u8m2_b4(op0, op1, op2) \ +__builtin_rvv_vmseq_vv_u8m2_b4((vuint8m2_t)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vmseq_vv_u8m2_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vv_u8m2_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vmseq_vv_u8m4_b2(op0, op1, op2) \ +__builtin_rvv_vmseq_vv_u8m4_b2((vuint8m4_t)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vmseq_vv_u8m4_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vv_u8m4_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vuint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vmseq_vv_u8m8_b1(op0, op1, op2) \ +__builtin_rvv_vmseq_vv_u8m8_b1((vuint8m8_t)(op0), (vuint8m8_t)(op1), (size_t)(op2)) +#define vmseq_vv_u8m8_b1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vv_u8m8_b1_m((vbool1_t)(op0), (vbool1_t)(op1), (vuint8m8_t)(op2), (vuint8m8_t)(op3), (size_t)(op4)) +#define vmseq_vv_u8mf2_b16(op0, op1, op2) \ +__builtin_rvv_vmseq_vv_u8mf2_b16((vuint8mf2_t)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vmseq_vv_u8mf2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vv_u8mf2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vmseq_vv_u8mf4_b32(op0, op1, op2) \ +__builtin_rvv_vmseq_vv_u8mf4_b32((vuint8mf4_t)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vmseq_vv_u8mf4_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vv_u8mf4_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vmseq_vv_u8mf8_b64(op0, op1, op2) \ +__builtin_rvv_vmseq_vv_u8mf8_b64((vuint8mf8_t)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vmseq_vv_u8mf8_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vv_u8mf8_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vmseq_vv_u16m1_b16(op0, op1, op2) \ +__builtin_rvv_vmseq_vv_u16m1_b16((vuint16m1_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vmseq_vv_u16m1_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vv_u16m1_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vmseq_vv_u16m2_b8(op0, op1, op2) \ +__builtin_rvv_vmseq_vv_u16m2_b8((vuint16m2_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vmseq_vv_u16m2_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vv_u16m2_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vmseq_vv_u16m4_b4(op0, op1, op2) \ +__builtin_rvv_vmseq_vv_u16m4_b4((vuint16m4_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vmseq_vv_u16m4_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vv_u16m4_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vmseq_vv_u16m8_b2(op0, op1, op2) \ +__builtin_rvv_vmseq_vv_u16m8_b2((vuint16m8_t)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vmseq_vv_u16m8_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vv_u16m8_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vuint16m8_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vmseq_vv_u16mf2_b32(op0, op1, op2) \ +__builtin_rvv_vmseq_vv_u16mf2_b32((vuint16mf2_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vmseq_vv_u16mf2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vv_u16mf2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vmseq_vv_u16mf4_b64(op0, op1, op2) \ +__builtin_rvv_vmseq_vv_u16mf4_b64((vuint16mf4_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vmseq_vv_u16mf4_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vv_u16mf4_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vmseq_vv_u32m1_b32(op0, op1, op2) \ +__builtin_rvv_vmseq_vv_u32m1_b32((vuint32m1_t)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vmseq_vv_u32m1_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vv_u32m1_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vmseq_vv_u32m2_b16(op0, op1, op2) \ +__builtin_rvv_vmseq_vv_u32m2_b16((vuint32m2_t)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vmseq_vv_u32m2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vv_u32m2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vmseq_vv_u32m4_b8(op0, op1, op2) \ +__builtin_rvv_vmseq_vv_u32m4_b8((vuint32m4_t)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vmseq_vv_u32m4_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vv_u32m4_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vmseq_vv_u32m8_b4(op0, op1, op2) \ +__builtin_rvv_vmseq_vv_u32m8_b4((vuint32m8_t)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vmseq_vv_u32m8_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vv_u32m8_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vuint32m8_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vmseq_vv_u32mf2_b64(op0, op1, op2) \ +__builtin_rvv_vmseq_vv_u32mf2_b64((vuint32mf2_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vmseq_vv_u32mf2_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vv_u32mf2_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vmseq_vv_u64m1_b64(op0, op1, op2) \ +__builtin_rvv_vmseq_vv_u64m1_b64((vuint64m1_t)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vmseq_vv_u64m1_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vv_u64m1_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vmseq_vv_u64m2_b32(op0, op1, op2) \ +__builtin_rvv_vmseq_vv_u64m2_b32((vuint64m2_t)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vmseq_vv_u64m2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vv_u64m2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vmseq_vv_u64m4_b16(op0, op1, op2) \ +__builtin_rvv_vmseq_vv_u64m4_b16((vuint64m4_t)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vmseq_vv_u64m4_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vv_u64m4_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vmseq_vv_u64m8_b8(op0, op1, op2) \ +__builtin_rvv_vmseq_vv_u64m8_b8((vuint64m8_t)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vmseq_vv_u64m8_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vv_u64m8_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint64m8_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vmseq_vx_u8m1_b8(op0, op1, op2) \ +__builtin_rvv_vmseq_vx_u8m1_b8((vuint8m1_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmseq_vx_u8m1_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vx_u8m1_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint8m1_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmseq_vx_u8m2_b4(op0, op1, op2) \ +__builtin_rvv_vmseq_vx_u8m2_b4((vuint8m2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmseq_vx_u8m2_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vx_u8m2_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vuint8m2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmseq_vx_u8m4_b2(op0, op1, op2) \ +__builtin_rvv_vmseq_vx_u8m4_b2((vuint8m4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmseq_vx_u8m4_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vx_u8m4_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vuint8m4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmseq_vx_u8m8_b1(op0, op1, op2) \ +__builtin_rvv_vmseq_vx_u8m8_b1((vuint8m8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmseq_vx_u8m8_b1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vx_u8m8_b1_m((vbool1_t)(op0), (vbool1_t)(op1), (vuint8m8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmseq_vx_u8mf2_b16(op0, op1, op2) \ +__builtin_rvv_vmseq_vx_u8mf2_b16((vuint8mf2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmseq_vx_u8mf2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vx_u8mf2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint8mf2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmseq_vx_u8mf4_b32(op0, op1, op2) \ +__builtin_rvv_vmseq_vx_u8mf4_b32((vuint8mf4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmseq_vx_u8mf4_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vx_u8mf4_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint8mf4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmseq_vx_u8mf8_b64(op0, op1, op2) \ +__builtin_rvv_vmseq_vx_u8mf8_b64((vuint8mf8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmseq_vx_u8mf8_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vx_u8mf8_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint8mf8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmseq_vx_u16m1_b16(op0, op1, op2) \ +__builtin_rvv_vmseq_vx_u16m1_b16((vuint16m1_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmseq_vx_u16m1_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vx_u16m1_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint16m1_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmseq_vx_u16m2_b8(op0, op1, op2) \ +__builtin_rvv_vmseq_vx_u16m2_b8((vuint16m2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmseq_vx_u16m2_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vx_u16m2_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint16m2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmseq_vx_u16m4_b4(op0, op1, op2) \ +__builtin_rvv_vmseq_vx_u16m4_b4((vuint16m4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmseq_vx_u16m4_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vx_u16m4_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vuint16m4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmseq_vx_u16m8_b2(op0, op1, op2) \ +__builtin_rvv_vmseq_vx_u16m8_b2((vuint16m8_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmseq_vx_u16m8_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vx_u16m8_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vuint16m8_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmseq_vx_u16mf2_b32(op0, op1, op2) \ +__builtin_rvv_vmseq_vx_u16mf2_b32((vuint16mf2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmseq_vx_u16mf2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vx_u16mf2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint16mf2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmseq_vx_u16mf4_b64(op0, op1, op2) \ +__builtin_rvv_vmseq_vx_u16mf4_b64((vuint16mf4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmseq_vx_u16mf4_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vx_u16mf4_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint16mf4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmseq_vx_u32m1_b32(op0, op1, op2) \ +__builtin_rvv_vmseq_vx_u32m1_b32((vuint32m1_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmseq_vx_u32m1_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vx_u32m1_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint32m1_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmseq_vx_u32m2_b16(op0, op1, op2) \ +__builtin_rvv_vmseq_vx_u32m2_b16((vuint32m2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmseq_vx_u32m2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vx_u32m2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint32m2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmseq_vx_u32m4_b8(op0, op1, op2) \ +__builtin_rvv_vmseq_vx_u32m4_b8((vuint32m4_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmseq_vx_u32m4_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vx_u32m4_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint32m4_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmseq_vx_u32m8_b4(op0, op1, op2) \ +__builtin_rvv_vmseq_vx_u32m8_b4((vuint32m8_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmseq_vx_u32m8_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vx_u32m8_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vuint32m8_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmseq_vx_u32mf2_b64(op0, op1, op2) \ +__builtin_rvv_vmseq_vx_u32mf2_b64((vuint32mf2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmseq_vx_u32mf2_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vx_u32mf2_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint32mf2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmseq_vx_u64m1_b64(op0, op1, op2) \ +__builtin_rvv_vmseq_vx_u64m1_b64((vuint64m1_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmseq_vx_u64m1_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vx_u64m1_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint64m1_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vmseq_vx_u64m2_b32(op0, op1, op2) \ +__builtin_rvv_vmseq_vx_u64m2_b32((vuint64m2_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmseq_vx_u64m2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vx_u64m2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint64m2_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vmseq_vx_u64m4_b16(op0, op1, op2) \ +__builtin_rvv_vmseq_vx_u64m4_b16((vuint64m4_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmseq_vx_u64m4_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vx_u64m4_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint64m4_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vmseq_vx_u64m8_b8(op0, op1, op2) \ +__builtin_rvv_vmseq_vx_u64m8_b8((vuint64m8_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmseq_vx_u64m8_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmseq_vx_u64m8_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint64m8_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vmsne_vv_i8m1_b8(op0, op1, op2) \ +__builtin_rvv_vmsne_vv_i8m1_b8((vint8m1_t)(op0), (vint8m1_t)(op1), (size_t)(op2)) +#define vmsne_vv_i8m1_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vv_i8m1_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vmsne_vv_i8m2_b4(op0, op1, op2) \ +__builtin_rvv_vmsne_vv_i8m2_b4((vint8m2_t)(op0), (vint8m2_t)(op1), (size_t)(op2)) +#define vmsne_vv_i8m2_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vv_i8m2_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vmsne_vv_i8m4_b2(op0, op1, op2) \ +__builtin_rvv_vmsne_vv_i8m4_b2((vint8m4_t)(op0), (vint8m4_t)(op1), (size_t)(op2)) +#define vmsne_vv_i8m4_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vv_i8m4_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vint8m4_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vmsne_vv_i8m8_b1(op0, op1, op2) \ +__builtin_rvv_vmsne_vv_i8m8_b1((vint8m8_t)(op0), (vint8m8_t)(op1), (size_t)(op2)) +#define vmsne_vv_i8m8_b1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vv_i8m8_b1_m((vbool1_t)(op0), (vbool1_t)(op1), (vint8m8_t)(op2), (vint8m8_t)(op3), (size_t)(op4)) +#define vmsne_vv_i8mf2_b16(op0, op1, op2) \ +__builtin_rvv_vmsne_vv_i8mf2_b16((vint8mf2_t)(op0), (vint8mf2_t)(op1), (size_t)(op2)) +#define vmsne_vv_i8mf2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vv_i8mf2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vmsne_vv_i8mf4_b32(op0, op1, op2) \ +__builtin_rvv_vmsne_vv_i8mf4_b32((vint8mf4_t)(op0), (vint8mf4_t)(op1), (size_t)(op2)) +#define vmsne_vv_i8mf4_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vv_i8mf4_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vmsne_vv_i8mf8_b64(op0, op1, op2) \ +__builtin_rvv_vmsne_vv_i8mf8_b64((vint8mf8_t)(op0), (vint8mf8_t)(op1), (size_t)(op2)) +#define vmsne_vv_i8mf8_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vv_i8mf8_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vmsne_vv_i16m1_b16(op0, op1, op2) \ +__builtin_rvv_vmsne_vv_i16m1_b16((vint16m1_t)(op0), (vint16m1_t)(op1), (size_t)(op2)) +#define vmsne_vv_i16m1_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vv_i16m1_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vmsne_vv_i16m2_b8(op0, op1, op2) \ +__builtin_rvv_vmsne_vv_i16m2_b8((vint16m2_t)(op0), (vint16m2_t)(op1), (size_t)(op2)) +#define vmsne_vv_i16m2_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vv_i16m2_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vmsne_vv_i16m4_b4(op0, op1, op2) \ +__builtin_rvv_vmsne_vv_i16m4_b4((vint16m4_t)(op0), (vint16m4_t)(op1), (size_t)(op2)) +#define vmsne_vv_i16m4_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vv_i16m4_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vint16m4_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vmsne_vv_i16m8_b2(op0, op1, op2) \ +__builtin_rvv_vmsne_vv_i16m8_b2((vint16m8_t)(op0), (vint16m8_t)(op1), (size_t)(op2)) +#define vmsne_vv_i16m8_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vv_i16m8_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vint16m8_t)(op2), (vint16m8_t)(op3), (size_t)(op4)) +#define vmsne_vv_i16mf2_b32(op0, op1, op2) \ +__builtin_rvv_vmsne_vv_i16mf2_b32((vint16mf2_t)(op0), (vint16mf2_t)(op1), (size_t)(op2)) +#define vmsne_vv_i16mf2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vv_i16mf2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vmsne_vv_i16mf4_b64(op0, op1, op2) \ +__builtin_rvv_vmsne_vv_i16mf4_b64((vint16mf4_t)(op0), (vint16mf4_t)(op1), (size_t)(op2)) +#define vmsne_vv_i16mf4_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vv_i16mf4_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vmsne_vv_i32m1_b32(op0, op1, op2) \ +__builtin_rvv_vmsne_vv_i32m1_b32((vint32m1_t)(op0), (vint32m1_t)(op1), (size_t)(op2)) +#define vmsne_vv_i32m1_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vv_i32m1_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vmsne_vv_i32m2_b16(op0, op1, op2) \ +__builtin_rvv_vmsne_vv_i32m2_b16((vint32m2_t)(op0), (vint32m2_t)(op1), (size_t)(op2)) +#define vmsne_vv_i32m2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vv_i32m2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vmsne_vv_i32m4_b8(op0, op1, op2) \ +__builtin_rvv_vmsne_vv_i32m4_b8((vint32m4_t)(op0), (vint32m4_t)(op1), (size_t)(op2)) +#define vmsne_vv_i32m4_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vv_i32m4_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vmsne_vv_i32m8_b4(op0, op1, op2) \ +__builtin_rvv_vmsne_vv_i32m8_b4((vint32m8_t)(op0), (vint32m8_t)(op1), (size_t)(op2)) +#define vmsne_vv_i32m8_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vv_i32m8_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vint32m8_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vmsne_vv_i32mf2_b64(op0, op1, op2) \ +__builtin_rvv_vmsne_vv_i32mf2_b64((vint32mf2_t)(op0), (vint32mf2_t)(op1), (size_t)(op2)) +#define vmsne_vv_i32mf2_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vv_i32mf2_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vmsne_vv_i64m1_b64(op0, op1, op2) \ +__builtin_rvv_vmsne_vv_i64m1_b64((vint64m1_t)(op0), (vint64m1_t)(op1), (size_t)(op2)) +#define vmsne_vv_i64m1_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vv_i64m1_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vmsne_vv_i64m2_b32(op0, op1, op2) \ +__builtin_rvv_vmsne_vv_i64m2_b32((vint64m2_t)(op0), (vint64m2_t)(op1), (size_t)(op2)) +#define vmsne_vv_i64m2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vv_i64m2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vmsne_vv_i64m4_b16(op0, op1, op2) \ +__builtin_rvv_vmsne_vv_i64m4_b16((vint64m4_t)(op0), (vint64m4_t)(op1), (size_t)(op2)) +#define vmsne_vv_i64m4_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vv_i64m4_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint64m4_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vmsne_vv_i64m8_b8(op0, op1, op2) \ +__builtin_rvv_vmsne_vv_i64m8_b8((vint64m8_t)(op0), (vint64m8_t)(op1), (size_t)(op2)) +#define vmsne_vv_i64m8_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vv_i64m8_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint64m8_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vmsne_vx_i8m1_b8(op0, op1, op2) \ +__builtin_rvv_vmsne_vx_i8m1_b8((vint8m1_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmsne_vx_i8m1_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vx_i8m1_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint8m1_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmsne_vx_i8m2_b4(op0, op1, op2) \ +__builtin_rvv_vmsne_vx_i8m2_b4((vint8m2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmsne_vx_i8m2_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vx_i8m2_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vint8m2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmsne_vx_i8m4_b2(op0, op1, op2) \ +__builtin_rvv_vmsne_vx_i8m4_b2((vint8m4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmsne_vx_i8m4_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vx_i8m4_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vint8m4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmsne_vx_i8m8_b1(op0, op1, op2) \ +__builtin_rvv_vmsne_vx_i8m8_b1((vint8m8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmsne_vx_i8m8_b1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vx_i8m8_b1_m((vbool1_t)(op0), (vbool1_t)(op1), (vint8m8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmsne_vx_i8mf2_b16(op0, op1, op2) \ +__builtin_rvv_vmsne_vx_i8mf2_b16((vint8mf2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmsne_vx_i8mf2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vx_i8mf2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint8mf2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmsne_vx_i8mf4_b32(op0, op1, op2) \ +__builtin_rvv_vmsne_vx_i8mf4_b32((vint8mf4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmsne_vx_i8mf4_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vx_i8mf4_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint8mf4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmsne_vx_i8mf8_b64(op0, op1, op2) \ +__builtin_rvv_vmsne_vx_i8mf8_b64((vint8mf8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmsne_vx_i8mf8_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vx_i8mf8_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint8mf8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmsne_vx_i16m1_b16(op0, op1, op2) \ +__builtin_rvv_vmsne_vx_i16m1_b16((vint16m1_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmsne_vx_i16m1_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vx_i16m1_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint16m1_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmsne_vx_i16m2_b8(op0, op1, op2) \ +__builtin_rvv_vmsne_vx_i16m2_b8((vint16m2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmsne_vx_i16m2_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vx_i16m2_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint16m2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmsne_vx_i16m4_b4(op0, op1, op2) \ +__builtin_rvv_vmsne_vx_i16m4_b4((vint16m4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmsne_vx_i16m4_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vx_i16m4_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vint16m4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmsne_vx_i16m8_b2(op0, op1, op2) \ +__builtin_rvv_vmsne_vx_i16m8_b2((vint16m8_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmsne_vx_i16m8_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vx_i16m8_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vint16m8_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmsne_vx_i16mf2_b32(op0, op1, op2) \ +__builtin_rvv_vmsne_vx_i16mf2_b32((vint16mf2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmsne_vx_i16mf2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vx_i16mf2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint16mf2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmsne_vx_i16mf4_b64(op0, op1, op2) \ +__builtin_rvv_vmsne_vx_i16mf4_b64((vint16mf4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmsne_vx_i16mf4_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vx_i16mf4_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint16mf4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmsne_vx_i32m1_b32(op0, op1, op2) \ +__builtin_rvv_vmsne_vx_i32m1_b32((vint32m1_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmsne_vx_i32m1_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vx_i32m1_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint32m1_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmsne_vx_i32m2_b16(op0, op1, op2) \ +__builtin_rvv_vmsne_vx_i32m2_b16((vint32m2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmsne_vx_i32m2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vx_i32m2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint32m2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmsne_vx_i32m4_b8(op0, op1, op2) \ +__builtin_rvv_vmsne_vx_i32m4_b8((vint32m4_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmsne_vx_i32m4_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vx_i32m4_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint32m4_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmsne_vx_i32m8_b4(op0, op1, op2) \ +__builtin_rvv_vmsne_vx_i32m8_b4((vint32m8_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmsne_vx_i32m8_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vx_i32m8_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vint32m8_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmsne_vx_i32mf2_b64(op0, op1, op2) \ +__builtin_rvv_vmsne_vx_i32mf2_b64((vint32mf2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmsne_vx_i32mf2_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vx_i32mf2_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint32mf2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmsne_vx_i64m1_b64(op0, op1, op2) \ +__builtin_rvv_vmsne_vx_i64m1_b64((vint64m1_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmsne_vx_i64m1_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vx_i64m1_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint64m1_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vmsne_vx_i64m2_b32(op0, op1, op2) \ +__builtin_rvv_vmsne_vx_i64m2_b32((vint64m2_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmsne_vx_i64m2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vx_i64m2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint64m2_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vmsne_vx_i64m4_b16(op0, op1, op2) \ +__builtin_rvv_vmsne_vx_i64m4_b16((vint64m4_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmsne_vx_i64m4_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vx_i64m4_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint64m4_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vmsne_vx_i64m8_b8(op0, op1, op2) \ +__builtin_rvv_vmsne_vx_i64m8_b8((vint64m8_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmsne_vx_i64m8_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vx_i64m8_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint64m8_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vmsne_vv_u8m1_b8(op0, op1, op2) \ +__builtin_rvv_vmsne_vv_u8m1_b8((vuint8m1_t)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vmsne_vv_u8m1_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vv_u8m1_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vmsne_vv_u8m2_b4(op0, op1, op2) \ +__builtin_rvv_vmsne_vv_u8m2_b4((vuint8m2_t)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vmsne_vv_u8m2_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vv_u8m2_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vmsne_vv_u8m4_b2(op0, op1, op2) \ +__builtin_rvv_vmsne_vv_u8m4_b2((vuint8m4_t)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vmsne_vv_u8m4_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vv_u8m4_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vuint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vmsne_vv_u8m8_b1(op0, op1, op2) \ +__builtin_rvv_vmsne_vv_u8m8_b1((vuint8m8_t)(op0), (vuint8m8_t)(op1), (size_t)(op2)) +#define vmsne_vv_u8m8_b1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vv_u8m8_b1_m((vbool1_t)(op0), (vbool1_t)(op1), (vuint8m8_t)(op2), (vuint8m8_t)(op3), (size_t)(op4)) +#define vmsne_vv_u8mf2_b16(op0, op1, op2) \ +__builtin_rvv_vmsne_vv_u8mf2_b16((vuint8mf2_t)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vmsne_vv_u8mf2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vv_u8mf2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vmsne_vv_u8mf4_b32(op0, op1, op2) \ +__builtin_rvv_vmsne_vv_u8mf4_b32((vuint8mf4_t)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vmsne_vv_u8mf4_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vv_u8mf4_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vmsne_vv_u8mf8_b64(op0, op1, op2) \ +__builtin_rvv_vmsne_vv_u8mf8_b64((vuint8mf8_t)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vmsne_vv_u8mf8_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vv_u8mf8_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vmsne_vv_u16m1_b16(op0, op1, op2) \ +__builtin_rvv_vmsne_vv_u16m1_b16((vuint16m1_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vmsne_vv_u16m1_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vv_u16m1_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vmsne_vv_u16m2_b8(op0, op1, op2) \ +__builtin_rvv_vmsne_vv_u16m2_b8((vuint16m2_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vmsne_vv_u16m2_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vv_u16m2_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vmsne_vv_u16m4_b4(op0, op1, op2) \ +__builtin_rvv_vmsne_vv_u16m4_b4((vuint16m4_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vmsne_vv_u16m4_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vv_u16m4_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vmsne_vv_u16m8_b2(op0, op1, op2) \ +__builtin_rvv_vmsne_vv_u16m8_b2((vuint16m8_t)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vmsne_vv_u16m8_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vv_u16m8_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vuint16m8_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vmsne_vv_u16mf2_b32(op0, op1, op2) \ +__builtin_rvv_vmsne_vv_u16mf2_b32((vuint16mf2_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vmsne_vv_u16mf2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vv_u16mf2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vmsne_vv_u16mf4_b64(op0, op1, op2) \ +__builtin_rvv_vmsne_vv_u16mf4_b64((vuint16mf4_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vmsne_vv_u16mf4_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vv_u16mf4_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vmsne_vv_u32m1_b32(op0, op1, op2) \ +__builtin_rvv_vmsne_vv_u32m1_b32((vuint32m1_t)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vmsne_vv_u32m1_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vv_u32m1_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vmsne_vv_u32m2_b16(op0, op1, op2) \ +__builtin_rvv_vmsne_vv_u32m2_b16((vuint32m2_t)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vmsne_vv_u32m2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vv_u32m2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vmsne_vv_u32m4_b8(op0, op1, op2) \ +__builtin_rvv_vmsne_vv_u32m4_b8((vuint32m4_t)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vmsne_vv_u32m4_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vv_u32m4_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vmsne_vv_u32m8_b4(op0, op1, op2) \ +__builtin_rvv_vmsne_vv_u32m8_b4((vuint32m8_t)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vmsne_vv_u32m8_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vv_u32m8_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vuint32m8_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vmsne_vv_u32mf2_b64(op0, op1, op2) \ +__builtin_rvv_vmsne_vv_u32mf2_b64((vuint32mf2_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vmsne_vv_u32mf2_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vv_u32mf2_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vmsne_vv_u64m1_b64(op0, op1, op2) \ +__builtin_rvv_vmsne_vv_u64m1_b64((vuint64m1_t)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vmsne_vv_u64m1_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vv_u64m1_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vmsne_vv_u64m2_b32(op0, op1, op2) \ +__builtin_rvv_vmsne_vv_u64m2_b32((vuint64m2_t)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vmsne_vv_u64m2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vv_u64m2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vmsne_vv_u64m4_b16(op0, op1, op2) \ +__builtin_rvv_vmsne_vv_u64m4_b16((vuint64m4_t)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vmsne_vv_u64m4_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vv_u64m4_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vmsne_vv_u64m8_b8(op0, op1, op2) \ +__builtin_rvv_vmsne_vv_u64m8_b8((vuint64m8_t)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vmsne_vv_u64m8_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vv_u64m8_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint64m8_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vmsne_vx_u8m1_b8(op0, op1, op2) \ +__builtin_rvv_vmsne_vx_u8m1_b8((vuint8m1_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmsne_vx_u8m1_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vx_u8m1_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint8m1_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmsne_vx_u8m2_b4(op0, op1, op2) \ +__builtin_rvv_vmsne_vx_u8m2_b4((vuint8m2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmsne_vx_u8m2_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vx_u8m2_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vuint8m2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmsne_vx_u8m4_b2(op0, op1, op2) \ +__builtin_rvv_vmsne_vx_u8m4_b2((vuint8m4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmsne_vx_u8m4_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vx_u8m4_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vuint8m4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmsne_vx_u8m8_b1(op0, op1, op2) \ +__builtin_rvv_vmsne_vx_u8m8_b1((vuint8m8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmsne_vx_u8m8_b1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vx_u8m8_b1_m((vbool1_t)(op0), (vbool1_t)(op1), (vuint8m8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmsne_vx_u8mf2_b16(op0, op1, op2) \ +__builtin_rvv_vmsne_vx_u8mf2_b16((vuint8mf2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmsne_vx_u8mf2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vx_u8mf2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint8mf2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmsne_vx_u8mf4_b32(op0, op1, op2) \ +__builtin_rvv_vmsne_vx_u8mf4_b32((vuint8mf4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmsne_vx_u8mf4_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vx_u8mf4_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint8mf4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmsne_vx_u8mf8_b64(op0, op1, op2) \ +__builtin_rvv_vmsne_vx_u8mf8_b64((vuint8mf8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmsne_vx_u8mf8_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vx_u8mf8_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint8mf8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmsne_vx_u16m1_b16(op0, op1, op2) \ +__builtin_rvv_vmsne_vx_u16m1_b16((vuint16m1_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmsne_vx_u16m1_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vx_u16m1_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint16m1_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmsne_vx_u16m2_b8(op0, op1, op2) \ +__builtin_rvv_vmsne_vx_u16m2_b8((vuint16m2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmsne_vx_u16m2_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vx_u16m2_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint16m2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmsne_vx_u16m4_b4(op0, op1, op2) \ +__builtin_rvv_vmsne_vx_u16m4_b4((vuint16m4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmsne_vx_u16m4_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vx_u16m4_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vuint16m4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmsne_vx_u16m8_b2(op0, op1, op2) \ +__builtin_rvv_vmsne_vx_u16m8_b2((vuint16m8_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmsne_vx_u16m8_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vx_u16m8_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vuint16m8_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmsne_vx_u16mf2_b32(op0, op1, op2) \ +__builtin_rvv_vmsne_vx_u16mf2_b32((vuint16mf2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmsne_vx_u16mf2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vx_u16mf2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint16mf2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmsne_vx_u16mf4_b64(op0, op1, op2) \ +__builtin_rvv_vmsne_vx_u16mf4_b64((vuint16mf4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmsne_vx_u16mf4_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vx_u16mf4_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint16mf4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmsne_vx_u32m1_b32(op0, op1, op2) \ +__builtin_rvv_vmsne_vx_u32m1_b32((vuint32m1_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmsne_vx_u32m1_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vx_u32m1_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint32m1_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmsne_vx_u32m2_b16(op0, op1, op2) \ +__builtin_rvv_vmsne_vx_u32m2_b16((vuint32m2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmsne_vx_u32m2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vx_u32m2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint32m2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmsne_vx_u32m4_b8(op0, op1, op2) \ +__builtin_rvv_vmsne_vx_u32m4_b8((vuint32m4_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmsne_vx_u32m4_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vx_u32m4_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint32m4_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmsne_vx_u32m8_b4(op0, op1, op2) \ +__builtin_rvv_vmsne_vx_u32m8_b4((vuint32m8_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmsne_vx_u32m8_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vx_u32m8_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vuint32m8_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmsne_vx_u32mf2_b64(op0, op1, op2) \ +__builtin_rvv_vmsne_vx_u32mf2_b64((vuint32mf2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmsne_vx_u32mf2_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vx_u32mf2_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint32mf2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmsne_vx_u64m1_b64(op0, op1, op2) \ +__builtin_rvv_vmsne_vx_u64m1_b64((vuint64m1_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmsne_vx_u64m1_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vx_u64m1_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint64m1_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vmsne_vx_u64m2_b32(op0, op1, op2) \ +__builtin_rvv_vmsne_vx_u64m2_b32((vuint64m2_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmsne_vx_u64m2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vx_u64m2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint64m2_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vmsne_vx_u64m4_b16(op0, op1, op2) \ +__builtin_rvv_vmsne_vx_u64m4_b16((vuint64m4_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmsne_vx_u64m4_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vx_u64m4_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint64m4_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vmsne_vx_u64m8_b8(op0, op1, op2) \ +__builtin_rvv_vmsne_vx_u64m8_b8((vuint64m8_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmsne_vx_u64m8_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsne_vx_u64m8_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint64m8_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vmsltu_vv_u8m1_b8(op0, op1, op2) \ +__builtin_rvv_vmsltu_vv_u8m1_b8((vuint8m1_t)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vmsltu_vv_u8m1_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsltu_vv_u8m1_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vmsltu_vv_u8m2_b4(op0, op1, op2) \ +__builtin_rvv_vmsltu_vv_u8m2_b4((vuint8m2_t)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vmsltu_vv_u8m2_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsltu_vv_u8m2_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vmsltu_vv_u8m4_b2(op0, op1, op2) \ +__builtin_rvv_vmsltu_vv_u8m4_b2((vuint8m4_t)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vmsltu_vv_u8m4_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsltu_vv_u8m4_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vuint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vmsltu_vv_u8m8_b1(op0, op1, op2) \ +__builtin_rvv_vmsltu_vv_u8m8_b1((vuint8m8_t)(op0), (vuint8m8_t)(op1), (size_t)(op2)) +#define vmsltu_vv_u8m8_b1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsltu_vv_u8m8_b1_m((vbool1_t)(op0), (vbool1_t)(op1), (vuint8m8_t)(op2), (vuint8m8_t)(op3), (size_t)(op4)) +#define vmsltu_vv_u8mf2_b16(op0, op1, op2) \ +__builtin_rvv_vmsltu_vv_u8mf2_b16((vuint8mf2_t)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vmsltu_vv_u8mf2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsltu_vv_u8mf2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vmsltu_vv_u8mf4_b32(op0, op1, op2) \ +__builtin_rvv_vmsltu_vv_u8mf4_b32((vuint8mf4_t)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vmsltu_vv_u8mf4_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsltu_vv_u8mf4_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vmsltu_vv_u8mf8_b64(op0, op1, op2) \ +__builtin_rvv_vmsltu_vv_u8mf8_b64((vuint8mf8_t)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vmsltu_vv_u8mf8_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsltu_vv_u8mf8_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vmsltu_vv_u16m1_b16(op0, op1, op2) \ +__builtin_rvv_vmsltu_vv_u16m1_b16((vuint16m1_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vmsltu_vv_u16m1_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsltu_vv_u16m1_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vmsltu_vv_u16m2_b8(op0, op1, op2) \ +__builtin_rvv_vmsltu_vv_u16m2_b8((vuint16m2_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vmsltu_vv_u16m2_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsltu_vv_u16m2_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vmsltu_vv_u16m4_b4(op0, op1, op2) \ +__builtin_rvv_vmsltu_vv_u16m4_b4((vuint16m4_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vmsltu_vv_u16m4_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsltu_vv_u16m4_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vmsltu_vv_u16m8_b2(op0, op1, op2) \ +__builtin_rvv_vmsltu_vv_u16m8_b2((vuint16m8_t)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vmsltu_vv_u16m8_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsltu_vv_u16m8_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vuint16m8_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vmsltu_vv_u16mf2_b32(op0, op1, op2) \ +__builtin_rvv_vmsltu_vv_u16mf2_b32((vuint16mf2_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vmsltu_vv_u16mf2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsltu_vv_u16mf2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vmsltu_vv_u16mf4_b64(op0, op1, op2) \ +__builtin_rvv_vmsltu_vv_u16mf4_b64((vuint16mf4_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vmsltu_vv_u16mf4_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsltu_vv_u16mf4_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vmsltu_vv_u32m1_b32(op0, op1, op2) \ +__builtin_rvv_vmsltu_vv_u32m1_b32((vuint32m1_t)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vmsltu_vv_u32m1_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsltu_vv_u32m1_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vmsltu_vv_u32m2_b16(op0, op1, op2) \ +__builtin_rvv_vmsltu_vv_u32m2_b16((vuint32m2_t)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vmsltu_vv_u32m2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsltu_vv_u32m2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vmsltu_vv_u32m4_b8(op0, op1, op2) \ +__builtin_rvv_vmsltu_vv_u32m4_b8((vuint32m4_t)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vmsltu_vv_u32m4_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsltu_vv_u32m4_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vmsltu_vv_u32m8_b4(op0, op1, op2) \ +__builtin_rvv_vmsltu_vv_u32m8_b4((vuint32m8_t)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vmsltu_vv_u32m8_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsltu_vv_u32m8_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vuint32m8_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vmsltu_vv_u32mf2_b64(op0, op1, op2) \ +__builtin_rvv_vmsltu_vv_u32mf2_b64((vuint32mf2_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vmsltu_vv_u32mf2_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsltu_vv_u32mf2_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vmsltu_vv_u64m1_b64(op0, op1, op2) \ +__builtin_rvv_vmsltu_vv_u64m1_b64((vuint64m1_t)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vmsltu_vv_u64m1_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsltu_vv_u64m1_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vmsltu_vv_u64m2_b32(op0, op1, op2) \ +__builtin_rvv_vmsltu_vv_u64m2_b32((vuint64m2_t)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vmsltu_vv_u64m2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsltu_vv_u64m2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vmsltu_vv_u64m4_b16(op0, op1, op2) \ +__builtin_rvv_vmsltu_vv_u64m4_b16((vuint64m4_t)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vmsltu_vv_u64m4_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsltu_vv_u64m4_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vmsltu_vv_u64m8_b8(op0, op1, op2) \ +__builtin_rvv_vmsltu_vv_u64m8_b8((vuint64m8_t)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vmsltu_vv_u64m8_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsltu_vv_u64m8_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint64m8_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vmsltu_vx_u8m1_b8(op0, op1, op2) \ +__builtin_rvv_vmsltu_vx_u8m1_b8((vuint8m1_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmsltu_vx_u8m1_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsltu_vx_u8m1_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint8m1_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmsltu_vx_u8m2_b4(op0, op1, op2) \ +__builtin_rvv_vmsltu_vx_u8m2_b4((vuint8m2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmsltu_vx_u8m2_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsltu_vx_u8m2_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vuint8m2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmsltu_vx_u8m4_b2(op0, op1, op2) \ +__builtin_rvv_vmsltu_vx_u8m4_b2((vuint8m4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmsltu_vx_u8m4_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsltu_vx_u8m4_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vuint8m4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmsltu_vx_u8m8_b1(op0, op1, op2) \ +__builtin_rvv_vmsltu_vx_u8m8_b1((vuint8m8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmsltu_vx_u8m8_b1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsltu_vx_u8m8_b1_m((vbool1_t)(op0), (vbool1_t)(op1), (vuint8m8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmsltu_vx_u8mf2_b16(op0, op1, op2) \ +__builtin_rvv_vmsltu_vx_u8mf2_b16((vuint8mf2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmsltu_vx_u8mf2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsltu_vx_u8mf2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint8mf2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmsltu_vx_u8mf4_b32(op0, op1, op2) \ +__builtin_rvv_vmsltu_vx_u8mf4_b32((vuint8mf4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmsltu_vx_u8mf4_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsltu_vx_u8mf4_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint8mf4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmsltu_vx_u8mf8_b64(op0, op1, op2) \ +__builtin_rvv_vmsltu_vx_u8mf8_b64((vuint8mf8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmsltu_vx_u8mf8_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsltu_vx_u8mf8_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint8mf8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmsltu_vx_u16m1_b16(op0, op1, op2) \ +__builtin_rvv_vmsltu_vx_u16m1_b16((vuint16m1_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmsltu_vx_u16m1_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsltu_vx_u16m1_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint16m1_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmsltu_vx_u16m2_b8(op0, op1, op2) \ +__builtin_rvv_vmsltu_vx_u16m2_b8((vuint16m2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmsltu_vx_u16m2_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsltu_vx_u16m2_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint16m2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmsltu_vx_u16m4_b4(op0, op1, op2) \ +__builtin_rvv_vmsltu_vx_u16m4_b4((vuint16m4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmsltu_vx_u16m4_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsltu_vx_u16m4_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vuint16m4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmsltu_vx_u16m8_b2(op0, op1, op2) \ +__builtin_rvv_vmsltu_vx_u16m8_b2((vuint16m8_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmsltu_vx_u16m8_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsltu_vx_u16m8_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vuint16m8_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmsltu_vx_u16mf2_b32(op0, op1, op2) \ +__builtin_rvv_vmsltu_vx_u16mf2_b32((vuint16mf2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmsltu_vx_u16mf2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsltu_vx_u16mf2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint16mf2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmsltu_vx_u16mf4_b64(op0, op1, op2) \ +__builtin_rvv_vmsltu_vx_u16mf4_b64((vuint16mf4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmsltu_vx_u16mf4_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsltu_vx_u16mf4_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint16mf4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmsltu_vx_u32m1_b32(op0, op1, op2) \ +__builtin_rvv_vmsltu_vx_u32m1_b32((vuint32m1_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmsltu_vx_u32m1_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsltu_vx_u32m1_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint32m1_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmsltu_vx_u32m2_b16(op0, op1, op2) \ +__builtin_rvv_vmsltu_vx_u32m2_b16((vuint32m2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmsltu_vx_u32m2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsltu_vx_u32m2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint32m2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmsltu_vx_u32m4_b8(op0, op1, op2) \ +__builtin_rvv_vmsltu_vx_u32m4_b8((vuint32m4_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmsltu_vx_u32m4_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsltu_vx_u32m4_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint32m4_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmsltu_vx_u32m8_b4(op0, op1, op2) \ +__builtin_rvv_vmsltu_vx_u32m8_b4((vuint32m8_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmsltu_vx_u32m8_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsltu_vx_u32m8_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vuint32m8_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmsltu_vx_u32mf2_b64(op0, op1, op2) \ +__builtin_rvv_vmsltu_vx_u32mf2_b64((vuint32mf2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmsltu_vx_u32mf2_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsltu_vx_u32mf2_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint32mf2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmsltu_vx_u64m1_b64(op0, op1, op2) \ +__builtin_rvv_vmsltu_vx_u64m1_b64((vuint64m1_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmsltu_vx_u64m1_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsltu_vx_u64m1_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint64m1_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vmsltu_vx_u64m2_b32(op0, op1, op2) \ +__builtin_rvv_vmsltu_vx_u64m2_b32((vuint64m2_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmsltu_vx_u64m2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsltu_vx_u64m2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint64m2_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vmsltu_vx_u64m4_b16(op0, op1, op2) \ +__builtin_rvv_vmsltu_vx_u64m4_b16((vuint64m4_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmsltu_vx_u64m4_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsltu_vx_u64m4_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint64m4_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vmsltu_vx_u64m8_b8(op0, op1, op2) \ +__builtin_rvv_vmsltu_vx_u64m8_b8((vuint64m8_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmsltu_vx_u64m8_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsltu_vx_u64m8_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint64m8_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vmslt_vv_i8m1_b8(op0, op1, op2) \ +__builtin_rvv_vmslt_vv_i8m1_b8((vint8m1_t)(op0), (vint8m1_t)(op1), (size_t)(op2)) +#define vmslt_vv_i8m1_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmslt_vv_i8m1_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vmslt_vv_i8m2_b4(op0, op1, op2) \ +__builtin_rvv_vmslt_vv_i8m2_b4((vint8m2_t)(op0), (vint8m2_t)(op1), (size_t)(op2)) +#define vmslt_vv_i8m2_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmslt_vv_i8m2_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vmslt_vv_i8m4_b2(op0, op1, op2) \ +__builtin_rvv_vmslt_vv_i8m4_b2((vint8m4_t)(op0), (vint8m4_t)(op1), (size_t)(op2)) +#define vmslt_vv_i8m4_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmslt_vv_i8m4_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vint8m4_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vmslt_vv_i8m8_b1(op0, op1, op2) \ +__builtin_rvv_vmslt_vv_i8m8_b1((vint8m8_t)(op0), (vint8m8_t)(op1), (size_t)(op2)) +#define vmslt_vv_i8m8_b1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmslt_vv_i8m8_b1_m((vbool1_t)(op0), (vbool1_t)(op1), (vint8m8_t)(op2), (vint8m8_t)(op3), (size_t)(op4)) +#define vmslt_vv_i8mf2_b16(op0, op1, op2) \ +__builtin_rvv_vmslt_vv_i8mf2_b16((vint8mf2_t)(op0), (vint8mf2_t)(op1), (size_t)(op2)) +#define vmslt_vv_i8mf2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmslt_vv_i8mf2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vmslt_vv_i8mf4_b32(op0, op1, op2) \ +__builtin_rvv_vmslt_vv_i8mf4_b32((vint8mf4_t)(op0), (vint8mf4_t)(op1), (size_t)(op2)) +#define vmslt_vv_i8mf4_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmslt_vv_i8mf4_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vmslt_vv_i8mf8_b64(op0, op1, op2) \ +__builtin_rvv_vmslt_vv_i8mf8_b64((vint8mf8_t)(op0), (vint8mf8_t)(op1), (size_t)(op2)) +#define vmslt_vv_i8mf8_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmslt_vv_i8mf8_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vmslt_vv_i16m1_b16(op0, op1, op2) \ +__builtin_rvv_vmslt_vv_i16m1_b16((vint16m1_t)(op0), (vint16m1_t)(op1), (size_t)(op2)) +#define vmslt_vv_i16m1_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmslt_vv_i16m1_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vmslt_vv_i16m2_b8(op0, op1, op2) \ +__builtin_rvv_vmslt_vv_i16m2_b8((vint16m2_t)(op0), (vint16m2_t)(op1), (size_t)(op2)) +#define vmslt_vv_i16m2_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmslt_vv_i16m2_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vmslt_vv_i16m4_b4(op0, op1, op2) \ +__builtin_rvv_vmslt_vv_i16m4_b4((vint16m4_t)(op0), (vint16m4_t)(op1), (size_t)(op2)) +#define vmslt_vv_i16m4_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmslt_vv_i16m4_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vint16m4_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vmslt_vv_i16m8_b2(op0, op1, op2) \ +__builtin_rvv_vmslt_vv_i16m8_b2((vint16m8_t)(op0), (vint16m8_t)(op1), (size_t)(op2)) +#define vmslt_vv_i16m8_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmslt_vv_i16m8_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vint16m8_t)(op2), (vint16m8_t)(op3), (size_t)(op4)) +#define vmslt_vv_i16mf2_b32(op0, op1, op2) \ +__builtin_rvv_vmslt_vv_i16mf2_b32((vint16mf2_t)(op0), (vint16mf2_t)(op1), (size_t)(op2)) +#define vmslt_vv_i16mf2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmslt_vv_i16mf2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vmslt_vv_i16mf4_b64(op0, op1, op2) \ +__builtin_rvv_vmslt_vv_i16mf4_b64((vint16mf4_t)(op0), (vint16mf4_t)(op1), (size_t)(op2)) +#define vmslt_vv_i16mf4_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmslt_vv_i16mf4_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vmslt_vv_i32m1_b32(op0, op1, op2) \ +__builtin_rvv_vmslt_vv_i32m1_b32((vint32m1_t)(op0), (vint32m1_t)(op1), (size_t)(op2)) +#define vmslt_vv_i32m1_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmslt_vv_i32m1_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vmslt_vv_i32m2_b16(op0, op1, op2) \ +__builtin_rvv_vmslt_vv_i32m2_b16((vint32m2_t)(op0), (vint32m2_t)(op1), (size_t)(op2)) +#define vmslt_vv_i32m2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmslt_vv_i32m2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vmslt_vv_i32m4_b8(op0, op1, op2) \ +__builtin_rvv_vmslt_vv_i32m4_b8((vint32m4_t)(op0), (vint32m4_t)(op1), (size_t)(op2)) +#define vmslt_vv_i32m4_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmslt_vv_i32m4_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vmslt_vv_i32m8_b4(op0, op1, op2) \ +__builtin_rvv_vmslt_vv_i32m8_b4((vint32m8_t)(op0), (vint32m8_t)(op1), (size_t)(op2)) +#define vmslt_vv_i32m8_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmslt_vv_i32m8_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vint32m8_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vmslt_vv_i32mf2_b64(op0, op1, op2) \ +__builtin_rvv_vmslt_vv_i32mf2_b64((vint32mf2_t)(op0), (vint32mf2_t)(op1), (size_t)(op2)) +#define vmslt_vv_i32mf2_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmslt_vv_i32mf2_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vmslt_vv_i64m1_b64(op0, op1, op2) \ +__builtin_rvv_vmslt_vv_i64m1_b64((vint64m1_t)(op0), (vint64m1_t)(op1), (size_t)(op2)) +#define vmslt_vv_i64m1_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmslt_vv_i64m1_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vmslt_vv_i64m2_b32(op0, op1, op2) \ +__builtin_rvv_vmslt_vv_i64m2_b32((vint64m2_t)(op0), (vint64m2_t)(op1), (size_t)(op2)) +#define vmslt_vv_i64m2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmslt_vv_i64m2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vmslt_vv_i64m4_b16(op0, op1, op2) \ +__builtin_rvv_vmslt_vv_i64m4_b16((vint64m4_t)(op0), (vint64m4_t)(op1), (size_t)(op2)) +#define vmslt_vv_i64m4_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmslt_vv_i64m4_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint64m4_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vmslt_vv_i64m8_b8(op0, op1, op2) \ +__builtin_rvv_vmslt_vv_i64m8_b8((vint64m8_t)(op0), (vint64m8_t)(op1), (size_t)(op2)) +#define vmslt_vv_i64m8_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmslt_vv_i64m8_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint64m8_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vsuxei32_v_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_u16m1((uint16_t *)(op0), (vuint32m2_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vsuxei32_v_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint32m2_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vsuxei32_v_u16m2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_u16m2((uint16_t *)(op0), (vuint32m4_t)(op1), (vuint16m2_t)(op2), (size_t)(op3)) +#define vsuxei32_v_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_u16m2_m((vbool8_t)(op0), (uint16_t *)(op1), (vuint32m4_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vsuxei32_v_u16m4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_u16m4((uint16_t *)(op0), (vuint32m8_t)(op1), (vuint16m4_t)(op2), (size_t)(op3)) +#define vsuxei32_v_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_u16m4_m((vbool4_t)(op0), (uint16_t *)(op1), (vuint32m8_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vsuxei32_v_u16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_u16mf2((uint16_t *)(op0), (vuint32m1_t)(op1), (vuint16mf2_t)(op2), (size_t)(op3)) +#define vsuxei32_v_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint32m1_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vsuxei32_v_u16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_u16mf4((uint16_t *)(op0), (vuint32mf2_t)(op1), (vuint16mf4_t)(op2), (size_t)(op3)) +#define vsuxei32_v_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint32mf2_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vmslt_vx_i8m1_b8(op0, op1, op2) \ +__builtin_rvv_vmslt_vx_i8m1_b8((vint8m1_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmslt_vx_i8m1_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmslt_vx_i8m1_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint8m1_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmslt_vx_i8m2_b4(op0, op1, op2) \ +__builtin_rvv_vmslt_vx_i8m2_b4((vint8m2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmslt_vx_i8m2_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmslt_vx_i8m2_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vint8m2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmslt_vx_i8m4_b2(op0, op1, op2) \ +__builtin_rvv_vmslt_vx_i8m4_b2((vint8m4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmslt_vx_i8m4_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmslt_vx_i8m4_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vint8m4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmslt_vx_i8m8_b1(op0, op1, op2) \ +__builtin_rvv_vmslt_vx_i8m8_b1((vint8m8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmslt_vx_i8m8_b1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmslt_vx_i8m8_b1_m((vbool1_t)(op0), (vbool1_t)(op1), (vint8m8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmslt_vx_i8mf2_b16(op0, op1, op2) \ +__builtin_rvv_vmslt_vx_i8mf2_b16((vint8mf2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmslt_vx_i8mf2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmslt_vx_i8mf2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint8mf2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmslt_vx_i8mf4_b32(op0, op1, op2) \ +__builtin_rvv_vmslt_vx_i8mf4_b32((vint8mf4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmslt_vx_i8mf4_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmslt_vx_i8mf4_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint8mf4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmslt_vx_i8mf8_b64(op0, op1, op2) \ +__builtin_rvv_vmslt_vx_i8mf8_b64((vint8mf8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmslt_vx_i8mf8_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmslt_vx_i8mf8_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint8mf8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmslt_vx_i16m1_b16(op0, op1, op2) \ +__builtin_rvv_vmslt_vx_i16m1_b16((vint16m1_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmslt_vx_i16m1_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmslt_vx_i16m1_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint16m1_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmslt_vx_i16m2_b8(op0, op1, op2) \ +__builtin_rvv_vmslt_vx_i16m2_b8((vint16m2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmslt_vx_i16m2_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmslt_vx_i16m2_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint16m2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmslt_vx_i16m4_b4(op0, op1, op2) \ +__builtin_rvv_vmslt_vx_i16m4_b4((vint16m4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmslt_vx_i16m4_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmslt_vx_i16m4_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vint16m4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmslt_vx_i16m8_b2(op0, op1, op2) \ +__builtin_rvv_vmslt_vx_i16m8_b2((vint16m8_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmslt_vx_i16m8_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmslt_vx_i16m8_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vint16m8_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmslt_vx_i16mf2_b32(op0, op1, op2) \ +__builtin_rvv_vmslt_vx_i16mf2_b32((vint16mf2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmslt_vx_i16mf2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmslt_vx_i16mf2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint16mf2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmslt_vx_i16mf4_b64(op0, op1, op2) \ +__builtin_rvv_vmslt_vx_i16mf4_b64((vint16mf4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmslt_vx_i16mf4_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmslt_vx_i16mf4_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint16mf4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmslt_vx_i32m1_b32(op0, op1, op2) \ +__builtin_rvv_vmslt_vx_i32m1_b32((vint32m1_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmslt_vx_i32m1_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmslt_vx_i32m1_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint32m1_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmslt_vx_i32m2_b16(op0, op1, op2) \ +__builtin_rvv_vmslt_vx_i32m2_b16((vint32m2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmslt_vx_i32m2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmslt_vx_i32m2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint32m2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmslt_vx_i32m4_b8(op0, op1, op2) \ +__builtin_rvv_vmslt_vx_i32m4_b8((vint32m4_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmslt_vx_i32m4_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmslt_vx_i32m4_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint32m4_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmslt_vx_i32m8_b4(op0, op1, op2) \ +__builtin_rvv_vmslt_vx_i32m8_b4((vint32m8_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmslt_vx_i32m8_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmslt_vx_i32m8_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vint32m8_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmslt_vx_i32mf2_b64(op0, op1, op2) \ +__builtin_rvv_vmslt_vx_i32mf2_b64((vint32mf2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmslt_vx_i32mf2_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmslt_vx_i32mf2_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint32mf2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmslt_vx_i64m1_b64(op0, op1, op2) \ +__builtin_rvv_vmslt_vx_i64m1_b64((vint64m1_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmslt_vx_i64m1_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmslt_vx_i64m1_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint64m1_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vmslt_vx_i64m2_b32(op0, op1, op2) \ +__builtin_rvv_vmslt_vx_i64m2_b32((vint64m2_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmslt_vx_i64m2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmslt_vx_i64m2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint64m2_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vmslt_vx_i64m4_b16(op0, op1, op2) \ +__builtin_rvv_vmslt_vx_i64m4_b16((vint64m4_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmslt_vx_i64m4_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmslt_vx_i64m4_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint64m4_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vmslt_vx_i64m8_b8(op0, op1, op2) \ +__builtin_rvv_vmslt_vx_i64m8_b8((vint64m8_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmslt_vx_i64m8_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmslt_vx_i64m8_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint64m8_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vmsleu_vv_u8m1_b8(op0, op1, op2) \ +__builtin_rvv_vmsleu_vv_u8m1_b8((vuint8m1_t)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vmsleu_vv_u8m1_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsleu_vv_u8m1_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vmsleu_vv_u8m2_b4(op0, op1, op2) \ +__builtin_rvv_vmsleu_vv_u8m2_b4((vuint8m2_t)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vmsleu_vv_u8m2_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsleu_vv_u8m2_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vmsleu_vv_u8m4_b2(op0, op1, op2) \ +__builtin_rvv_vmsleu_vv_u8m4_b2((vuint8m4_t)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vmsleu_vv_u8m4_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsleu_vv_u8m4_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vuint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vmsleu_vv_u8m8_b1(op0, op1, op2) \ +__builtin_rvv_vmsleu_vv_u8m8_b1((vuint8m8_t)(op0), (vuint8m8_t)(op1), (size_t)(op2)) +#define vmsleu_vv_u8m8_b1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsleu_vv_u8m8_b1_m((vbool1_t)(op0), (vbool1_t)(op1), (vuint8m8_t)(op2), (vuint8m8_t)(op3), (size_t)(op4)) +#define vmsleu_vv_u8mf2_b16(op0, op1, op2) \ +__builtin_rvv_vmsleu_vv_u8mf2_b16((vuint8mf2_t)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vmsleu_vv_u8mf2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsleu_vv_u8mf2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vmsleu_vv_u8mf4_b32(op0, op1, op2) \ +__builtin_rvv_vmsleu_vv_u8mf4_b32((vuint8mf4_t)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vmsleu_vv_u8mf4_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsleu_vv_u8mf4_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vmsleu_vv_u8mf8_b64(op0, op1, op2) \ +__builtin_rvv_vmsleu_vv_u8mf8_b64((vuint8mf8_t)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vmsleu_vv_u8mf8_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsleu_vv_u8mf8_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vmsleu_vv_u16m1_b16(op0, op1, op2) \ +__builtin_rvv_vmsleu_vv_u16m1_b16((vuint16m1_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vmsleu_vv_u16m1_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsleu_vv_u16m1_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vmsleu_vv_u16m2_b8(op0, op1, op2) \ +__builtin_rvv_vmsleu_vv_u16m2_b8((vuint16m2_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vmsleu_vv_u16m2_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsleu_vv_u16m2_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vmsleu_vv_u16m4_b4(op0, op1, op2) \ +__builtin_rvv_vmsleu_vv_u16m4_b4((vuint16m4_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vmsleu_vv_u16m4_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsleu_vv_u16m4_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vmsleu_vv_u16m8_b2(op0, op1, op2) \ +__builtin_rvv_vmsleu_vv_u16m8_b2((vuint16m8_t)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vmsleu_vv_u16m8_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsleu_vv_u16m8_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vuint16m8_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vmsleu_vv_u16mf2_b32(op0, op1, op2) \ +__builtin_rvv_vmsleu_vv_u16mf2_b32((vuint16mf2_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vmsleu_vv_u16mf2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsleu_vv_u16mf2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vmsleu_vv_u16mf4_b64(op0, op1, op2) \ +__builtin_rvv_vmsleu_vv_u16mf4_b64((vuint16mf4_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vmsleu_vv_u16mf4_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsleu_vv_u16mf4_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vmsleu_vv_u32m1_b32(op0, op1, op2) \ +__builtin_rvv_vmsleu_vv_u32m1_b32((vuint32m1_t)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vmsleu_vv_u32m1_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsleu_vv_u32m1_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vmsleu_vv_u32m2_b16(op0, op1, op2) \ +__builtin_rvv_vmsleu_vv_u32m2_b16((vuint32m2_t)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vmsleu_vv_u32m2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsleu_vv_u32m2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vmsleu_vv_u32m4_b8(op0, op1, op2) \ +__builtin_rvv_vmsleu_vv_u32m4_b8((vuint32m4_t)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vmsleu_vv_u32m4_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsleu_vv_u32m4_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vmsleu_vv_u32m8_b4(op0, op1, op2) \ +__builtin_rvv_vmsleu_vv_u32m8_b4((vuint32m8_t)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vmsleu_vv_u32m8_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsleu_vv_u32m8_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vuint32m8_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vmsleu_vv_u32mf2_b64(op0, op1, op2) \ +__builtin_rvv_vmsleu_vv_u32mf2_b64((vuint32mf2_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vmsleu_vv_u32mf2_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsleu_vv_u32mf2_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vmsleu_vv_u64m1_b64(op0, op1, op2) \ +__builtin_rvv_vmsleu_vv_u64m1_b64((vuint64m1_t)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vmsleu_vv_u64m1_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsleu_vv_u64m1_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vmsleu_vv_u64m2_b32(op0, op1, op2) \ +__builtin_rvv_vmsleu_vv_u64m2_b32((vuint64m2_t)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vmsleu_vv_u64m2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsleu_vv_u64m2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vmsleu_vv_u64m4_b16(op0, op1, op2) \ +__builtin_rvv_vmsleu_vv_u64m4_b16((vuint64m4_t)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vmsleu_vv_u64m4_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsleu_vv_u64m4_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vmsleu_vv_u64m8_b8(op0, op1, op2) \ +__builtin_rvv_vmsleu_vv_u64m8_b8((vuint64m8_t)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vmsleu_vv_u64m8_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsleu_vv_u64m8_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint64m8_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vmsleu_vx_u8m1_b8(op0, op1, op2) \ +__builtin_rvv_vmsleu_vx_u8m1_b8((vuint8m1_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmsleu_vx_u8m1_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsleu_vx_u8m1_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint8m1_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmsleu_vx_u8m2_b4(op0, op1, op2) \ +__builtin_rvv_vmsleu_vx_u8m2_b4((vuint8m2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmsleu_vx_u8m2_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsleu_vx_u8m2_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vuint8m2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmsleu_vx_u8m4_b2(op0, op1, op2) \ +__builtin_rvv_vmsleu_vx_u8m4_b2((vuint8m4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmsleu_vx_u8m4_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsleu_vx_u8m4_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vuint8m4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmsleu_vx_u8m8_b1(op0, op1, op2) \ +__builtin_rvv_vmsleu_vx_u8m8_b1((vuint8m8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmsleu_vx_u8m8_b1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsleu_vx_u8m8_b1_m((vbool1_t)(op0), (vbool1_t)(op1), (vuint8m8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmsleu_vx_u8mf2_b16(op0, op1, op2) \ +__builtin_rvv_vmsleu_vx_u8mf2_b16((vuint8mf2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmsleu_vx_u8mf2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsleu_vx_u8mf2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint8mf2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmsleu_vx_u8mf4_b32(op0, op1, op2) \ +__builtin_rvv_vmsleu_vx_u8mf4_b32((vuint8mf4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmsleu_vx_u8mf4_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsleu_vx_u8mf4_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint8mf4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmsleu_vx_u8mf8_b64(op0, op1, op2) \ +__builtin_rvv_vmsleu_vx_u8mf8_b64((vuint8mf8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmsleu_vx_u8mf8_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsleu_vx_u8mf8_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint8mf8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmsleu_vx_u16m1_b16(op0, op1, op2) \ +__builtin_rvv_vmsleu_vx_u16m1_b16((vuint16m1_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmsleu_vx_u16m1_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsleu_vx_u16m1_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint16m1_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmsleu_vx_u16m2_b8(op0, op1, op2) \ +__builtin_rvv_vmsleu_vx_u16m2_b8((vuint16m2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmsleu_vx_u16m2_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsleu_vx_u16m2_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint16m2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmsleu_vx_u16m4_b4(op0, op1, op2) \ +__builtin_rvv_vmsleu_vx_u16m4_b4((vuint16m4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmsleu_vx_u16m4_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsleu_vx_u16m4_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vuint16m4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmsleu_vx_u16m8_b2(op0, op1, op2) \ +__builtin_rvv_vmsleu_vx_u16m8_b2((vuint16m8_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmsleu_vx_u16m8_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsleu_vx_u16m8_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vuint16m8_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmsleu_vx_u16mf2_b32(op0, op1, op2) \ +__builtin_rvv_vmsleu_vx_u16mf2_b32((vuint16mf2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmsleu_vx_u16mf2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsleu_vx_u16mf2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint16mf2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmsleu_vx_u16mf4_b64(op0, op1, op2) \ +__builtin_rvv_vmsleu_vx_u16mf4_b64((vuint16mf4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmsleu_vx_u16mf4_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsleu_vx_u16mf4_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint16mf4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmsleu_vx_u32m1_b32(op0, op1, op2) \ +__builtin_rvv_vmsleu_vx_u32m1_b32((vuint32m1_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmsleu_vx_u32m1_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsleu_vx_u32m1_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint32m1_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmsleu_vx_u32m2_b16(op0, op1, op2) \ +__builtin_rvv_vmsleu_vx_u32m2_b16((vuint32m2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmsleu_vx_u32m2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsleu_vx_u32m2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint32m2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmsleu_vx_u32m4_b8(op0, op1, op2) \ +__builtin_rvv_vmsleu_vx_u32m4_b8((vuint32m4_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmsleu_vx_u32m4_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsleu_vx_u32m4_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint32m4_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmsleu_vx_u32m8_b4(op0, op1, op2) \ +__builtin_rvv_vmsleu_vx_u32m8_b4((vuint32m8_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmsleu_vx_u32m8_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsleu_vx_u32m8_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vuint32m8_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmsleu_vx_u32mf2_b64(op0, op1, op2) \ +__builtin_rvv_vmsleu_vx_u32mf2_b64((vuint32mf2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmsleu_vx_u32mf2_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsleu_vx_u32mf2_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint32mf2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmsleu_vx_u64m1_b64(op0, op1, op2) \ +__builtin_rvv_vmsleu_vx_u64m1_b64((vuint64m1_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmsleu_vx_u64m1_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsleu_vx_u64m1_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint64m1_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vmsleu_vx_u64m2_b32(op0, op1, op2) \ +__builtin_rvv_vmsleu_vx_u64m2_b32((vuint64m2_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmsleu_vx_u64m2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsleu_vx_u64m2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint64m2_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vmsleu_vx_u64m4_b16(op0, op1, op2) \ +__builtin_rvv_vmsleu_vx_u64m4_b16((vuint64m4_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmsleu_vx_u64m4_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsleu_vx_u64m4_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint64m4_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vmsleu_vx_u64m8_b8(op0, op1, op2) \ +__builtin_rvv_vmsleu_vx_u64m8_b8((vuint64m8_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmsleu_vx_u64m8_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsleu_vx_u64m8_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint64m8_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vmsle_vv_i8m1_b8(op0, op1, op2) \ +__builtin_rvv_vmsle_vv_i8m1_b8((vint8m1_t)(op0), (vint8m1_t)(op1), (size_t)(op2)) +#define vmsle_vv_i8m1_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsle_vv_i8m1_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vmsle_vv_i8m2_b4(op0, op1, op2) \ +__builtin_rvv_vmsle_vv_i8m2_b4((vint8m2_t)(op0), (vint8m2_t)(op1), (size_t)(op2)) +#define vmsle_vv_i8m2_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsle_vv_i8m2_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vmsle_vv_i8m4_b2(op0, op1, op2) \ +__builtin_rvv_vmsle_vv_i8m4_b2((vint8m4_t)(op0), (vint8m4_t)(op1), (size_t)(op2)) +#define vmsle_vv_i8m4_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsle_vv_i8m4_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vint8m4_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vmsle_vv_i8m8_b1(op0, op1, op2) \ +__builtin_rvv_vmsle_vv_i8m8_b1((vint8m8_t)(op0), (vint8m8_t)(op1), (size_t)(op2)) +#define vmsle_vv_i8m8_b1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsle_vv_i8m8_b1_m((vbool1_t)(op0), (vbool1_t)(op1), (vint8m8_t)(op2), (vint8m8_t)(op3), (size_t)(op4)) +#define vmsle_vv_i8mf2_b16(op0, op1, op2) \ +__builtin_rvv_vmsle_vv_i8mf2_b16((vint8mf2_t)(op0), (vint8mf2_t)(op1), (size_t)(op2)) +#define vmsle_vv_i8mf2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsle_vv_i8mf2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vmsle_vv_i8mf4_b32(op0, op1, op2) \ +__builtin_rvv_vmsle_vv_i8mf4_b32((vint8mf4_t)(op0), (vint8mf4_t)(op1), (size_t)(op2)) +#define vmsle_vv_i8mf4_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsle_vv_i8mf4_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vmsle_vv_i8mf8_b64(op0, op1, op2) \ +__builtin_rvv_vmsle_vv_i8mf8_b64((vint8mf8_t)(op0), (vint8mf8_t)(op1), (size_t)(op2)) +#define vmsle_vv_i8mf8_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsle_vv_i8mf8_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vmsle_vv_i16m1_b16(op0, op1, op2) \ +__builtin_rvv_vmsle_vv_i16m1_b16((vint16m1_t)(op0), (vint16m1_t)(op1), (size_t)(op2)) +#define vmsle_vv_i16m1_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsle_vv_i16m1_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vmsle_vv_i16m2_b8(op0, op1, op2) \ +__builtin_rvv_vmsle_vv_i16m2_b8((vint16m2_t)(op0), (vint16m2_t)(op1), (size_t)(op2)) +#define vmsle_vv_i16m2_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsle_vv_i16m2_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vmsle_vv_i16m4_b4(op0, op1, op2) \ +__builtin_rvv_vmsle_vv_i16m4_b4((vint16m4_t)(op0), (vint16m4_t)(op1), (size_t)(op2)) +#define vmsle_vv_i16m4_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsle_vv_i16m4_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vint16m4_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vmsle_vv_i16m8_b2(op0, op1, op2) \ +__builtin_rvv_vmsle_vv_i16m8_b2((vint16m8_t)(op0), (vint16m8_t)(op1), (size_t)(op2)) +#define vmsle_vv_i16m8_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsle_vv_i16m8_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vint16m8_t)(op2), (vint16m8_t)(op3), (size_t)(op4)) +#define vmsle_vv_i16mf2_b32(op0, op1, op2) \ +__builtin_rvv_vmsle_vv_i16mf2_b32((vint16mf2_t)(op0), (vint16mf2_t)(op1), (size_t)(op2)) +#define vmsle_vv_i16mf2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsle_vv_i16mf2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vmsle_vv_i16mf4_b64(op0, op1, op2) \ +__builtin_rvv_vmsle_vv_i16mf4_b64((vint16mf4_t)(op0), (vint16mf4_t)(op1), (size_t)(op2)) +#define vmsle_vv_i16mf4_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsle_vv_i16mf4_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vmsle_vv_i32m1_b32(op0, op1, op2) \ +__builtin_rvv_vmsle_vv_i32m1_b32((vint32m1_t)(op0), (vint32m1_t)(op1), (size_t)(op2)) +#define vmsle_vv_i32m1_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsle_vv_i32m1_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vmsle_vv_i32m2_b16(op0, op1, op2) \ +__builtin_rvv_vmsle_vv_i32m2_b16((vint32m2_t)(op0), (vint32m2_t)(op1), (size_t)(op2)) +#define vmsle_vv_i32m2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsle_vv_i32m2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vmsle_vv_i32m4_b8(op0, op1, op2) \ +__builtin_rvv_vmsle_vv_i32m4_b8((vint32m4_t)(op0), (vint32m4_t)(op1), (size_t)(op2)) +#define vmsle_vv_i32m4_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsle_vv_i32m4_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vmsle_vv_i32m8_b4(op0, op1, op2) \ +__builtin_rvv_vmsle_vv_i32m8_b4((vint32m8_t)(op0), (vint32m8_t)(op1), (size_t)(op2)) +#define vmsle_vv_i32m8_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsle_vv_i32m8_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vint32m8_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vmsle_vv_i32mf2_b64(op0, op1, op2) \ +__builtin_rvv_vmsle_vv_i32mf2_b64((vint32mf2_t)(op0), (vint32mf2_t)(op1), (size_t)(op2)) +#define vmsle_vv_i32mf2_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsle_vv_i32mf2_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vmsle_vv_i64m1_b64(op0, op1, op2) \ +__builtin_rvv_vmsle_vv_i64m1_b64((vint64m1_t)(op0), (vint64m1_t)(op1), (size_t)(op2)) +#define vmsle_vv_i64m1_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsle_vv_i64m1_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vmsle_vv_i64m2_b32(op0, op1, op2) \ +__builtin_rvv_vmsle_vv_i64m2_b32((vint64m2_t)(op0), (vint64m2_t)(op1), (size_t)(op2)) +#define vmsle_vv_i64m2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsle_vv_i64m2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vmsle_vv_i64m4_b16(op0, op1, op2) \ +__builtin_rvv_vmsle_vv_i64m4_b16((vint64m4_t)(op0), (vint64m4_t)(op1), (size_t)(op2)) +#define vmsle_vv_i64m4_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsle_vv_i64m4_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint64m4_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vmsle_vv_i64m8_b8(op0, op1, op2) \ +__builtin_rvv_vmsle_vv_i64m8_b8((vint64m8_t)(op0), (vint64m8_t)(op1), (size_t)(op2)) +#define vmsle_vv_i64m8_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsle_vv_i64m8_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint64m8_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vmsle_vx_i8m1_b8(op0, op1, op2) \ +__builtin_rvv_vmsle_vx_i8m1_b8((vint8m1_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmsle_vx_i8m1_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsle_vx_i8m1_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint8m1_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmsle_vx_i8m2_b4(op0, op1, op2) \ +__builtin_rvv_vmsle_vx_i8m2_b4((vint8m2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmsle_vx_i8m2_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsle_vx_i8m2_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vint8m2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmsle_vx_i8m4_b2(op0, op1, op2) \ +__builtin_rvv_vmsle_vx_i8m4_b2((vint8m4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmsle_vx_i8m4_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsle_vx_i8m4_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vint8m4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmsle_vx_i8m8_b1(op0, op1, op2) \ +__builtin_rvv_vmsle_vx_i8m8_b1((vint8m8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmsle_vx_i8m8_b1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsle_vx_i8m8_b1_m((vbool1_t)(op0), (vbool1_t)(op1), (vint8m8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmsle_vx_i8mf2_b16(op0, op1, op2) \ +__builtin_rvv_vmsle_vx_i8mf2_b16((vint8mf2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmsle_vx_i8mf2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsle_vx_i8mf2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint8mf2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmsle_vx_i8mf4_b32(op0, op1, op2) \ +__builtin_rvv_vmsle_vx_i8mf4_b32((vint8mf4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmsle_vx_i8mf4_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsle_vx_i8mf4_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint8mf4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmsle_vx_i8mf8_b64(op0, op1, op2) \ +__builtin_rvv_vmsle_vx_i8mf8_b64((vint8mf8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmsle_vx_i8mf8_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsle_vx_i8mf8_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint8mf8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmsle_vx_i16m1_b16(op0, op1, op2) \ +__builtin_rvv_vmsle_vx_i16m1_b16((vint16m1_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmsle_vx_i16m1_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsle_vx_i16m1_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint16m1_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmsle_vx_i16m2_b8(op0, op1, op2) \ +__builtin_rvv_vmsle_vx_i16m2_b8((vint16m2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmsle_vx_i16m2_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsle_vx_i16m2_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint16m2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmsle_vx_i16m4_b4(op0, op1, op2) \ +__builtin_rvv_vmsle_vx_i16m4_b4((vint16m4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmsle_vx_i16m4_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsle_vx_i16m4_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vint16m4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmsle_vx_i16m8_b2(op0, op1, op2) \ +__builtin_rvv_vmsle_vx_i16m8_b2((vint16m8_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmsle_vx_i16m8_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsle_vx_i16m8_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vint16m8_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmsle_vx_i16mf2_b32(op0, op1, op2) \ +__builtin_rvv_vmsle_vx_i16mf2_b32((vint16mf2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmsle_vx_i16mf2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsle_vx_i16mf2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint16mf2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmsle_vx_i16mf4_b64(op0, op1, op2) \ +__builtin_rvv_vmsle_vx_i16mf4_b64((vint16mf4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmsle_vx_i16mf4_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsle_vx_i16mf4_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint16mf4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmsle_vx_i32m1_b32(op0, op1, op2) \ +__builtin_rvv_vmsle_vx_i32m1_b32((vint32m1_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmsle_vx_i32m1_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsle_vx_i32m1_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint32m1_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmsle_vx_i32m2_b16(op0, op1, op2) \ +__builtin_rvv_vmsle_vx_i32m2_b16((vint32m2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmsle_vx_i32m2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsle_vx_i32m2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint32m2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmsle_vx_i32m4_b8(op0, op1, op2) \ +__builtin_rvv_vmsle_vx_i32m4_b8((vint32m4_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmsle_vx_i32m4_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsle_vx_i32m4_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint32m4_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmsle_vx_i32m8_b4(op0, op1, op2) \ +__builtin_rvv_vmsle_vx_i32m8_b4((vint32m8_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmsle_vx_i32m8_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsle_vx_i32m8_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vint32m8_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmsle_vx_i32mf2_b64(op0, op1, op2) \ +__builtin_rvv_vmsle_vx_i32mf2_b64((vint32mf2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmsle_vx_i32mf2_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsle_vx_i32mf2_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint32mf2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmsle_vx_i64m1_b64(op0, op1, op2) \ +__builtin_rvv_vmsle_vx_i64m1_b64((vint64m1_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmsle_vx_i64m1_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsle_vx_i64m1_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint64m1_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vmsle_vx_i64m2_b32(op0, op1, op2) \ +__builtin_rvv_vmsle_vx_i64m2_b32((vint64m2_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmsle_vx_i64m2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsle_vx_i64m2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint64m2_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vmsle_vx_i64m4_b16(op0, op1, op2) \ +__builtin_rvv_vmsle_vx_i64m4_b16((vint64m4_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmsle_vx_i64m4_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsle_vx_i64m4_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint64m4_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vmsle_vx_i64m8_b8(op0, op1, op2) \ +__builtin_rvv_vmsle_vx_i64m8_b8((vint64m8_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmsle_vx_i64m8_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsle_vx_i64m8_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint64m8_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vmsgtu_vv_u8m1_b8(op0, op1, op2) \ +__builtin_rvv_vmsgtu_vv_u8m1_b8((vuint8m1_t)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vmsgtu_vv_u8m1_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgtu_vv_u8m1_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vmsgtu_vv_u8m2_b4(op0, op1, op2) \ +__builtin_rvv_vmsgtu_vv_u8m2_b4((vuint8m2_t)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vmsgtu_vv_u8m2_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgtu_vv_u8m2_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vmsgtu_vv_u8m4_b2(op0, op1, op2) \ +__builtin_rvv_vmsgtu_vv_u8m4_b2((vuint8m4_t)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vmsgtu_vv_u8m4_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgtu_vv_u8m4_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vuint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vmsgtu_vv_u8m8_b1(op0, op1, op2) \ +__builtin_rvv_vmsgtu_vv_u8m8_b1((vuint8m8_t)(op0), (vuint8m8_t)(op1), (size_t)(op2)) +#define vmsgtu_vv_u8m8_b1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgtu_vv_u8m8_b1_m((vbool1_t)(op0), (vbool1_t)(op1), (vuint8m8_t)(op2), (vuint8m8_t)(op3), (size_t)(op4)) +#define vmsgtu_vv_u8mf2_b16(op0, op1, op2) \ +__builtin_rvv_vmsgtu_vv_u8mf2_b16((vuint8mf2_t)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vmsgtu_vv_u8mf2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgtu_vv_u8mf2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vmsgtu_vv_u8mf4_b32(op0, op1, op2) \ +__builtin_rvv_vmsgtu_vv_u8mf4_b32((vuint8mf4_t)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vmsgtu_vv_u8mf4_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgtu_vv_u8mf4_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vmsgtu_vv_u8mf8_b64(op0, op1, op2) \ +__builtin_rvv_vmsgtu_vv_u8mf8_b64((vuint8mf8_t)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vmsgtu_vv_u8mf8_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgtu_vv_u8mf8_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vmsgtu_vv_u16m1_b16(op0, op1, op2) \ +__builtin_rvv_vmsgtu_vv_u16m1_b16((vuint16m1_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vmsgtu_vv_u16m1_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgtu_vv_u16m1_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vmsgtu_vv_u16m2_b8(op0, op1, op2) \ +__builtin_rvv_vmsgtu_vv_u16m2_b8((vuint16m2_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vmsgtu_vv_u16m2_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgtu_vv_u16m2_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vmsgtu_vv_u16m4_b4(op0, op1, op2) \ +__builtin_rvv_vmsgtu_vv_u16m4_b4((vuint16m4_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vmsgtu_vv_u16m4_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgtu_vv_u16m4_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vmsgtu_vv_u16m8_b2(op0, op1, op2) \ +__builtin_rvv_vmsgtu_vv_u16m8_b2((vuint16m8_t)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vmsgtu_vv_u16m8_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgtu_vv_u16m8_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vuint16m8_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vmsgtu_vv_u16mf2_b32(op0, op1, op2) \ +__builtin_rvv_vmsgtu_vv_u16mf2_b32((vuint16mf2_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vmsgtu_vv_u16mf2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgtu_vv_u16mf2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vmsgtu_vv_u16mf4_b64(op0, op1, op2) \ +__builtin_rvv_vmsgtu_vv_u16mf4_b64((vuint16mf4_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vmsgtu_vv_u16mf4_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgtu_vv_u16mf4_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vmsgtu_vv_u32m1_b32(op0, op1, op2) \ +__builtin_rvv_vmsgtu_vv_u32m1_b32((vuint32m1_t)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vmsgtu_vv_u32m1_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgtu_vv_u32m1_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vmsgtu_vv_u32m2_b16(op0, op1, op2) \ +__builtin_rvv_vmsgtu_vv_u32m2_b16((vuint32m2_t)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vmsgtu_vv_u32m2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgtu_vv_u32m2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vmsgtu_vv_u32m4_b8(op0, op1, op2) \ +__builtin_rvv_vmsgtu_vv_u32m4_b8((vuint32m4_t)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vmsgtu_vv_u32m4_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgtu_vv_u32m4_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vmsgtu_vv_u32m8_b4(op0, op1, op2) \ +__builtin_rvv_vmsgtu_vv_u32m8_b4((vuint32m8_t)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vmsgtu_vv_u32m8_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgtu_vv_u32m8_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vuint32m8_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vmsgtu_vv_u32mf2_b64(op0, op1, op2) \ +__builtin_rvv_vmsgtu_vv_u32mf2_b64((vuint32mf2_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vmsgtu_vv_u32mf2_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgtu_vv_u32mf2_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vmsgtu_vv_u64m1_b64(op0, op1, op2) \ +__builtin_rvv_vmsgtu_vv_u64m1_b64((vuint64m1_t)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vmsgtu_vv_u64m1_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgtu_vv_u64m1_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vmsgtu_vv_u64m2_b32(op0, op1, op2) \ +__builtin_rvv_vmsgtu_vv_u64m2_b32((vuint64m2_t)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vmsgtu_vv_u64m2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgtu_vv_u64m2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vmsgtu_vv_u64m4_b16(op0, op1, op2) \ +__builtin_rvv_vmsgtu_vv_u64m4_b16((vuint64m4_t)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vmsgtu_vv_u64m4_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgtu_vv_u64m4_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vmsgtu_vv_u64m8_b8(op0, op1, op2) \ +__builtin_rvv_vmsgtu_vv_u64m8_b8((vuint64m8_t)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vmsgtu_vv_u64m8_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgtu_vv_u64m8_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint64m8_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vmsgtu_vx_u8m1_b8(op0, op1, op2) \ +__builtin_rvv_vmsgtu_vx_u8m1_b8((vuint8m1_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmsgtu_vx_u8m1_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgtu_vx_u8m1_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint8m1_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmsgtu_vx_u8m2_b4(op0, op1, op2) \ +__builtin_rvv_vmsgtu_vx_u8m2_b4((vuint8m2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmsgtu_vx_u8m2_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgtu_vx_u8m2_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vuint8m2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmsgtu_vx_u8m4_b2(op0, op1, op2) \ +__builtin_rvv_vmsgtu_vx_u8m4_b2((vuint8m4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmsgtu_vx_u8m4_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgtu_vx_u8m4_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vuint8m4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmsgtu_vx_u8m8_b1(op0, op1, op2) \ +__builtin_rvv_vmsgtu_vx_u8m8_b1((vuint8m8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmsgtu_vx_u8m8_b1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgtu_vx_u8m8_b1_m((vbool1_t)(op0), (vbool1_t)(op1), (vuint8m8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmsgtu_vx_u8mf2_b16(op0, op1, op2) \ +__builtin_rvv_vmsgtu_vx_u8mf2_b16((vuint8mf2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmsgtu_vx_u8mf2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgtu_vx_u8mf2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint8mf2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmsgtu_vx_u8mf4_b32(op0, op1, op2) \ +__builtin_rvv_vmsgtu_vx_u8mf4_b32((vuint8mf4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmsgtu_vx_u8mf4_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgtu_vx_u8mf4_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint8mf4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmsgtu_vx_u8mf8_b64(op0, op1, op2) \ +__builtin_rvv_vmsgtu_vx_u8mf8_b64((vuint8mf8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmsgtu_vx_u8mf8_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgtu_vx_u8mf8_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint8mf8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmsgtu_vx_u16m1_b16(op0, op1, op2) \ +__builtin_rvv_vmsgtu_vx_u16m1_b16((vuint16m1_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmsgtu_vx_u16m1_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgtu_vx_u16m1_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint16m1_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmsgtu_vx_u16m2_b8(op0, op1, op2) \ +__builtin_rvv_vmsgtu_vx_u16m2_b8((vuint16m2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmsgtu_vx_u16m2_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgtu_vx_u16m2_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint16m2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmsgtu_vx_u16m4_b4(op0, op1, op2) \ +__builtin_rvv_vmsgtu_vx_u16m4_b4((vuint16m4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmsgtu_vx_u16m4_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgtu_vx_u16m4_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vuint16m4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmsgtu_vx_u16m8_b2(op0, op1, op2) \ +__builtin_rvv_vmsgtu_vx_u16m8_b2((vuint16m8_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmsgtu_vx_u16m8_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgtu_vx_u16m8_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vuint16m8_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmsgtu_vx_u16mf2_b32(op0, op1, op2) \ +__builtin_rvv_vmsgtu_vx_u16mf2_b32((vuint16mf2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmsgtu_vx_u16mf2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgtu_vx_u16mf2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint16mf2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmsgtu_vx_u16mf4_b64(op0, op1, op2) \ +__builtin_rvv_vmsgtu_vx_u16mf4_b64((vuint16mf4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmsgtu_vx_u16mf4_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgtu_vx_u16mf4_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint16mf4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmsgtu_vx_u32m1_b32(op0, op1, op2) \ +__builtin_rvv_vmsgtu_vx_u32m1_b32((vuint32m1_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmsgtu_vx_u32m1_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgtu_vx_u32m1_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint32m1_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmsgtu_vx_u32m2_b16(op0, op1, op2) \ +__builtin_rvv_vmsgtu_vx_u32m2_b16((vuint32m2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmsgtu_vx_u32m2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgtu_vx_u32m2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint32m2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmsgtu_vx_u32m4_b8(op0, op1, op2) \ +__builtin_rvv_vmsgtu_vx_u32m4_b8((vuint32m4_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmsgtu_vx_u32m4_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgtu_vx_u32m4_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint32m4_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmsgtu_vx_u32m8_b4(op0, op1, op2) \ +__builtin_rvv_vmsgtu_vx_u32m8_b4((vuint32m8_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmsgtu_vx_u32m8_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgtu_vx_u32m8_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vuint32m8_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmsgtu_vx_u32mf2_b64(op0, op1, op2) \ +__builtin_rvv_vmsgtu_vx_u32mf2_b64((vuint32mf2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmsgtu_vx_u32mf2_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgtu_vx_u32mf2_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint32mf2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmsgtu_vx_u64m1_b64(op0, op1, op2) \ +__builtin_rvv_vmsgtu_vx_u64m1_b64((vuint64m1_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmsgtu_vx_u64m1_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgtu_vx_u64m1_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint64m1_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vmsgtu_vx_u64m2_b32(op0, op1, op2) \ +__builtin_rvv_vmsgtu_vx_u64m2_b32((vuint64m2_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmsgtu_vx_u64m2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgtu_vx_u64m2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint64m2_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vmsgtu_vx_u64m4_b16(op0, op1, op2) \ +__builtin_rvv_vmsgtu_vx_u64m4_b16((vuint64m4_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmsgtu_vx_u64m4_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgtu_vx_u64m4_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint64m4_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vmsgtu_vx_u64m8_b8(op0, op1, op2) \ +__builtin_rvv_vmsgtu_vx_u64m8_b8((vuint64m8_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmsgtu_vx_u64m8_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgtu_vx_u64m8_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint64m8_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vmsgt_vv_i8m1_b8(op0, op1, op2) \ +__builtin_rvv_vmsgt_vv_i8m1_b8((vint8m1_t)(op0), (vint8m1_t)(op1), (size_t)(op2)) +#define vmsgt_vv_i8m1_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgt_vv_i8m1_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vmsgt_vv_i8m2_b4(op0, op1, op2) \ +__builtin_rvv_vmsgt_vv_i8m2_b4((vint8m2_t)(op0), (vint8m2_t)(op1), (size_t)(op2)) +#define vmsgt_vv_i8m2_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgt_vv_i8m2_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vmsgt_vv_i8m4_b2(op0, op1, op2) \ +__builtin_rvv_vmsgt_vv_i8m4_b2((vint8m4_t)(op0), (vint8m4_t)(op1), (size_t)(op2)) +#define vmsgt_vv_i8m4_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgt_vv_i8m4_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vint8m4_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vmsgt_vv_i8m8_b1(op0, op1, op2) \ +__builtin_rvv_vmsgt_vv_i8m8_b1((vint8m8_t)(op0), (vint8m8_t)(op1), (size_t)(op2)) +#define vmsgt_vv_i8m8_b1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgt_vv_i8m8_b1_m((vbool1_t)(op0), (vbool1_t)(op1), (vint8m8_t)(op2), (vint8m8_t)(op3), (size_t)(op4)) +#define vmsgt_vv_i8mf2_b16(op0, op1, op2) \ +__builtin_rvv_vmsgt_vv_i8mf2_b16((vint8mf2_t)(op0), (vint8mf2_t)(op1), (size_t)(op2)) +#define vmsgt_vv_i8mf2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgt_vv_i8mf2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vmsgt_vv_i8mf4_b32(op0, op1, op2) \ +__builtin_rvv_vmsgt_vv_i8mf4_b32((vint8mf4_t)(op0), (vint8mf4_t)(op1), (size_t)(op2)) +#define vmsgt_vv_i8mf4_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgt_vv_i8mf4_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vmsgt_vv_i8mf8_b64(op0, op1, op2) \ +__builtin_rvv_vmsgt_vv_i8mf8_b64((vint8mf8_t)(op0), (vint8mf8_t)(op1), (size_t)(op2)) +#define vmsgt_vv_i8mf8_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgt_vv_i8mf8_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vmsgt_vv_i16m1_b16(op0, op1, op2) \ +__builtin_rvv_vmsgt_vv_i16m1_b16((vint16m1_t)(op0), (vint16m1_t)(op1), (size_t)(op2)) +#define vmsgt_vv_i16m1_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgt_vv_i16m1_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vmsgt_vv_i16m2_b8(op0, op1, op2) \ +__builtin_rvv_vmsgt_vv_i16m2_b8((vint16m2_t)(op0), (vint16m2_t)(op1), (size_t)(op2)) +#define vmsgt_vv_i16m2_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgt_vv_i16m2_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vmsgt_vv_i16m4_b4(op0, op1, op2) \ +__builtin_rvv_vmsgt_vv_i16m4_b4((vint16m4_t)(op0), (vint16m4_t)(op1), (size_t)(op2)) +#define vmsgt_vv_i16m4_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgt_vv_i16m4_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vint16m4_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vmsgt_vv_i16m8_b2(op0, op1, op2) \ +__builtin_rvv_vmsgt_vv_i16m8_b2((vint16m8_t)(op0), (vint16m8_t)(op1), (size_t)(op2)) +#define vmsgt_vv_i16m8_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgt_vv_i16m8_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vint16m8_t)(op2), (vint16m8_t)(op3), (size_t)(op4)) +#define vmsgt_vv_i16mf2_b32(op0, op1, op2) \ +__builtin_rvv_vmsgt_vv_i16mf2_b32((vint16mf2_t)(op0), (vint16mf2_t)(op1), (size_t)(op2)) +#define vmsgt_vv_i16mf2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgt_vv_i16mf2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vmsgt_vv_i16mf4_b64(op0, op1, op2) \ +__builtin_rvv_vmsgt_vv_i16mf4_b64((vint16mf4_t)(op0), (vint16mf4_t)(op1), (size_t)(op2)) +#define vmsgt_vv_i16mf4_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgt_vv_i16mf4_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vmsgt_vv_i32m1_b32(op0, op1, op2) \ +__builtin_rvv_vmsgt_vv_i32m1_b32((vint32m1_t)(op0), (vint32m1_t)(op1), (size_t)(op2)) +#define vmsgt_vv_i32m1_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgt_vv_i32m1_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vmsgt_vv_i32m2_b16(op0, op1, op2) \ +__builtin_rvv_vmsgt_vv_i32m2_b16((vint32m2_t)(op0), (vint32m2_t)(op1), (size_t)(op2)) +#define vmsgt_vv_i32m2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgt_vv_i32m2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vmsgt_vv_i32m4_b8(op0, op1, op2) \ +__builtin_rvv_vmsgt_vv_i32m4_b8((vint32m4_t)(op0), (vint32m4_t)(op1), (size_t)(op2)) +#define vmsgt_vv_i32m4_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgt_vv_i32m4_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vmsgt_vv_i32m8_b4(op0, op1, op2) \ +__builtin_rvv_vmsgt_vv_i32m8_b4((vint32m8_t)(op0), (vint32m8_t)(op1), (size_t)(op2)) +#define vmsgt_vv_i32m8_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgt_vv_i32m8_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vint32m8_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vmsgt_vv_i32mf2_b64(op0, op1, op2) \ +__builtin_rvv_vmsgt_vv_i32mf2_b64((vint32mf2_t)(op0), (vint32mf2_t)(op1), (size_t)(op2)) +#define vmsgt_vv_i32mf2_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgt_vv_i32mf2_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vmsgt_vv_i64m1_b64(op0, op1, op2) \ +__builtin_rvv_vmsgt_vv_i64m1_b64((vint64m1_t)(op0), (vint64m1_t)(op1), (size_t)(op2)) +#define vmsgt_vv_i64m1_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgt_vv_i64m1_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vmsgt_vv_i64m2_b32(op0, op1, op2) \ +__builtin_rvv_vmsgt_vv_i64m2_b32((vint64m2_t)(op0), (vint64m2_t)(op1), (size_t)(op2)) +#define vmsgt_vv_i64m2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgt_vv_i64m2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vmsgt_vv_i64m4_b16(op0, op1, op2) \ +__builtin_rvv_vmsgt_vv_i64m4_b16((vint64m4_t)(op0), (vint64m4_t)(op1), (size_t)(op2)) +#define vmsgt_vv_i64m4_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgt_vv_i64m4_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint64m4_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vmsgt_vv_i64m8_b8(op0, op1, op2) \ +__builtin_rvv_vmsgt_vv_i64m8_b8((vint64m8_t)(op0), (vint64m8_t)(op1), (size_t)(op2)) +#define vmsgt_vv_i64m8_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgt_vv_i64m8_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint64m8_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vmsgt_vx_i8m1_b8(op0, op1, op2) \ +__builtin_rvv_vmsgt_vx_i8m1_b8((vint8m1_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmsgt_vx_i8m1_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgt_vx_i8m1_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint8m1_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmsgt_vx_i8m2_b4(op0, op1, op2) \ +__builtin_rvv_vmsgt_vx_i8m2_b4((vint8m2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmsgt_vx_i8m2_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgt_vx_i8m2_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vint8m2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmsgt_vx_i8m4_b2(op0, op1, op2) \ +__builtin_rvv_vmsgt_vx_i8m4_b2((vint8m4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmsgt_vx_i8m4_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgt_vx_i8m4_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vint8m4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmsgt_vx_i8m8_b1(op0, op1, op2) \ +__builtin_rvv_vmsgt_vx_i8m8_b1((vint8m8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmsgt_vx_i8m8_b1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgt_vx_i8m8_b1_m((vbool1_t)(op0), (vbool1_t)(op1), (vint8m8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmsgt_vx_i8mf2_b16(op0, op1, op2) \ +__builtin_rvv_vmsgt_vx_i8mf2_b16((vint8mf2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmsgt_vx_i8mf2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgt_vx_i8mf2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint8mf2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmsgt_vx_i8mf4_b32(op0, op1, op2) \ +__builtin_rvv_vmsgt_vx_i8mf4_b32((vint8mf4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmsgt_vx_i8mf4_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgt_vx_i8mf4_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint8mf4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmsgt_vx_i8mf8_b64(op0, op1, op2) \ +__builtin_rvv_vmsgt_vx_i8mf8_b64((vint8mf8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmsgt_vx_i8mf8_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgt_vx_i8mf8_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint8mf8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmsgt_vx_i16m1_b16(op0, op1, op2) \ +__builtin_rvv_vmsgt_vx_i16m1_b16((vint16m1_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmsgt_vx_i16m1_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgt_vx_i16m1_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint16m1_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmsgt_vx_i16m2_b8(op0, op1, op2) \ +__builtin_rvv_vmsgt_vx_i16m2_b8((vint16m2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmsgt_vx_i16m2_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgt_vx_i16m2_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint16m2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmsgt_vx_i16m4_b4(op0, op1, op2) \ +__builtin_rvv_vmsgt_vx_i16m4_b4((vint16m4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmsgt_vx_i16m4_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgt_vx_i16m4_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vint16m4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmsgt_vx_i16m8_b2(op0, op1, op2) \ +__builtin_rvv_vmsgt_vx_i16m8_b2((vint16m8_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmsgt_vx_i16m8_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgt_vx_i16m8_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vint16m8_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmsgt_vx_i16mf2_b32(op0, op1, op2) \ +__builtin_rvv_vmsgt_vx_i16mf2_b32((vint16mf2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmsgt_vx_i16mf2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgt_vx_i16mf2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint16mf2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmsgt_vx_i16mf4_b64(op0, op1, op2) \ +__builtin_rvv_vmsgt_vx_i16mf4_b64((vint16mf4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmsgt_vx_i16mf4_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgt_vx_i16mf4_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint16mf4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmsgt_vx_i32m1_b32(op0, op1, op2) \ +__builtin_rvv_vmsgt_vx_i32m1_b32((vint32m1_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmsgt_vx_i32m1_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgt_vx_i32m1_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint32m1_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmsgt_vx_i32m2_b16(op0, op1, op2) \ +__builtin_rvv_vmsgt_vx_i32m2_b16((vint32m2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmsgt_vx_i32m2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgt_vx_i32m2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint32m2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmsgt_vx_i32m4_b8(op0, op1, op2) \ +__builtin_rvv_vmsgt_vx_i32m4_b8((vint32m4_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmsgt_vx_i32m4_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgt_vx_i32m4_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint32m4_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmsgt_vx_i32m8_b4(op0, op1, op2) \ +__builtin_rvv_vmsgt_vx_i32m8_b4((vint32m8_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmsgt_vx_i32m8_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgt_vx_i32m8_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vint32m8_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmsgt_vx_i32mf2_b64(op0, op1, op2) \ +__builtin_rvv_vmsgt_vx_i32mf2_b64((vint32mf2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmsgt_vx_i32mf2_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgt_vx_i32mf2_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint32mf2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmsgt_vx_i64m1_b64(op0, op1, op2) \ +__builtin_rvv_vmsgt_vx_i64m1_b64((vint64m1_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmsgt_vx_i64m1_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgt_vx_i64m1_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint64m1_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vmsgt_vx_i64m2_b32(op0, op1, op2) \ +__builtin_rvv_vmsgt_vx_i64m2_b32((vint64m2_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmsgt_vx_i64m2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgt_vx_i64m2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint64m2_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vmsgt_vx_i64m4_b16(op0, op1, op2) \ +__builtin_rvv_vmsgt_vx_i64m4_b16((vint64m4_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmsgt_vx_i64m4_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgt_vx_i64m4_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint64m4_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vmsgt_vx_i64m8_b8(op0, op1, op2) \ +__builtin_rvv_vmsgt_vx_i64m8_b8((vint64m8_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmsgt_vx_i64m8_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgt_vx_i64m8_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint64m8_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vmsgeu_vv_u8m1_b8(op0, op1, op2) \ +__builtin_rvv_vmsgeu_vv_u8m1_b8((vuint8m1_t)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vmsgeu_vv_u8m1_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgeu_vv_u8m1_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vmsgeu_vv_u8m2_b4(op0, op1, op2) \ +__builtin_rvv_vmsgeu_vv_u8m2_b4((vuint8m2_t)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vmsgeu_vv_u8m2_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgeu_vv_u8m2_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vmsgeu_vv_u8m4_b2(op0, op1, op2) \ +__builtin_rvv_vmsgeu_vv_u8m4_b2((vuint8m4_t)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vmsgeu_vv_u8m4_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgeu_vv_u8m4_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vuint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vmsgeu_vv_u8m8_b1(op0, op1, op2) \ +__builtin_rvv_vmsgeu_vv_u8m8_b1((vuint8m8_t)(op0), (vuint8m8_t)(op1), (size_t)(op2)) +#define vmsgeu_vv_u8m8_b1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgeu_vv_u8m8_b1_m((vbool1_t)(op0), (vbool1_t)(op1), (vuint8m8_t)(op2), (vuint8m8_t)(op3), (size_t)(op4)) +#define vmsgeu_vv_u8mf2_b16(op0, op1, op2) \ +__builtin_rvv_vmsgeu_vv_u8mf2_b16((vuint8mf2_t)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vmsgeu_vv_u8mf2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgeu_vv_u8mf2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vmsgeu_vv_u8mf4_b32(op0, op1, op2) \ +__builtin_rvv_vmsgeu_vv_u8mf4_b32((vuint8mf4_t)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vmsgeu_vv_u8mf4_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgeu_vv_u8mf4_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vmsgeu_vv_u8mf8_b64(op0, op1, op2) \ +__builtin_rvv_vmsgeu_vv_u8mf8_b64((vuint8mf8_t)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vmsgeu_vv_u8mf8_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgeu_vv_u8mf8_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vmsgeu_vv_u16m1_b16(op0, op1, op2) \ +__builtin_rvv_vmsgeu_vv_u16m1_b16((vuint16m1_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vmsgeu_vv_u16m1_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgeu_vv_u16m1_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vmsgeu_vv_u16m2_b8(op0, op1, op2) \ +__builtin_rvv_vmsgeu_vv_u16m2_b8((vuint16m2_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vmsgeu_vv_u16m2_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgeu_vv_u16m2_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vmsgeu_vv_u16m4_b4(op0, op1, op2) \ +__builtin_rvv_vmsgeu_vv_u16m4_b4((vuint16m4_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vmsgeu_vv_u16m4_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgeu_vv_u16m4_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vmsgeu_vv_u16m8_b2(op0, op1, op2) \ +__builtin_rvv_vmsgeu_vv_u16m8_b2((vuint16m8_t)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vmsgeu_vv_u16m8_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgeu_vv_u16m8_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vuint16m8_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vmsgeu_vv_u16mf2_b32(op0, op1, op2) \ +__builtin_rvv_vmsgeu_vv_u16mf2_b32((vuint16mf2_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vmsgeu_vv_u16mf2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgeu_vv_u16mf2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vmsgeu_vv_u16mf4_b64(op0, op1, op2) \ +__builtin_rvv_vmsgeu_vv_u16mf4_b64((vuint16mf4_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vmsgeu_vv_u16mf4_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgeu_vv_u16mf4_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vmsgeu_vv_u32m1_b32(op0, op1, op2) \ +__builtin_rvv_vmsgeu_vv_u32m1_b32((vuint32m1_t)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vmsgeu_vv_u32m1_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgeu_vv_u32m1_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vmsgeu_vv_u32m2_b16(op0, op1, op2) \ +__builtin_rvv_vmsgeu_vv_u32m2_b16((vuint32m2_t)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vmsgeu_vv_u32m2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgeu_vv_u32m2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vmsgeu_vv_u32m4_b8(op0, op1, op2) \ +__builtin_rvv_vmsgeu_vv_u32m4_b8((vuint32m4_t)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vmsgeu_vv_u32m4_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgeu_vv_u32m4_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vmsgeu_vv_u32m8_b4(op0, op1, op2) \ +__builtin_rvv_vmsgeu_vv_u32m8_b4((vuint32m8_t)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vmsgeu_vv_u32m8_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgeu_vv_u32m8_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vuint32m8_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vmsgeu_vv_u32mf2_b64(op0, op1, op2) \ +__builtin_rvv_vmsgeu_vv_u32mf2_b64((vuint32mf2_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vmsgeu_vv_u32mf2_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgeu_vv_u32mf2_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vmsgeu_vv_u64m1_b64(op0, op1, op2) \ +__builtin_rvv_vmsgeu_vv_u64m1_b64((vuint64m1_t)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vmsgeu_vv_u64m1_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgeu_vv_u64m1_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vmsgeu_vv_u64m2_b32(op0, op1, op2) \ +__builtin_rvv_vmsgeu_vv_u64m2_b32((vuint64m2_t)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vmsgeu_vv_u64m2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgeu_vv_u64m2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vmsgeu_vv_u64m4_b16(op0, op1, op2) \ +__builtin_rvv_vmsgeu_vv_u64m4_b16((vuint64m4_t)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vmsgeu_vv_u64m4_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgeu_vv_u64m4_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vmsgeu_vv_u64m8_b8(op0, op1, op2) \ +__builtin_rvv_vmsgeu_vv_u64m8_b8((vuint64m8_t)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vmsgeu_vv_u64m8_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgeu_vv_u64m8_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint64m8_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vsuxei64_v_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei64_v_i16m1((int16_t *)(op0), (vuint64m4_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vsuxei64_v_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei64_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint64m4_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vsuxei64_v_i16m2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei64_v_i16m2((int16_t *)(op0), (vuint64m8_t)(op1), (vint16m2_t)(op2), (size_t)(op3)) +#define vsuxei64_v_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei64_v_i16m2_m((vbool8_t)(op0), (int16_t *)(op1), (vuint64m8_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vsuxei64_v_i16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei64_v_i16mf2((int16_t *)(op0), (vuint64m2_t)(op1), (vint16mf2_t)(op2), (size_t)(op3)) +#define vsuxei64_v_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei64_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint64m2_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vsuxei64_v_i16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei64_v_i16mf4((int16_t *)(op0), (vuint64m1_t)(op1), (vint16mf4_t)(op2), (size_t)(op3)) +#define vsuxei64_v_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei64_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint64m1_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vmsgeu_vx_u8m1_b8(op0, op1, op2) \ +__builtin_rvv_vmsgeu_vx_u8m1_b8((vuint8m1_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmsgeu_vx_u8m1_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgeu_vx_u8m1_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint8m1_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmsgeu_vx_u8m2_b4(op0, op1, op2) \ +__builtin_rvv_vmsgeu_vx_u8m2_b4((vuint8m2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmsgeu_vx_u8m2_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgeu_vx_u8m2_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vuint8m2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmsgeu_vx_u8m4_b2(op0, op1, op2) \ +__builtin_rvv_vmsgeu_vx_u8m4_b2((vuint8m4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmsgeu_vx_u8m4_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgeu_vx_u8m4_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vuint8m4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmsgeu_vx_u8m8_b1(op0, op1, op2) \ +__builtin_rvv_vmsgeu_vx_u8m8_b1((vuint8m8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmsgeu_vx_u8m8_b1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgeu_vx_u8m8_b1_m((vbool1_t)(op0), (vbool1_t)(op1), (vuint8m8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmsgeu_vx_u8mf2_b16(op0, op1, op2) \ +__builtin_rvv_vmsgeu_vx_u8mf2_b16((vuint8mf2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmsgeu_vx_u8mf2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgeu_vx_u8mf2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint8mf2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmsgeu_vx_u8mf4_b32(op0, op1, op2) \ +__builtin_rvv_vmsgeu_vx_u8mf4_b32((vuint8mf4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmsgeu_vx_u8mf4_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgeu_vx_u8mf4_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint8mf4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmsgeu_vx_u8mf8_b64(op0, op1, op2) \ +__builtin_rvv_vmsgeu_vx_u8mf8_b64((vuint8mf8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmsgeu_vx_u8mf8_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgeu_vx_u8mf8_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint8mf8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmsgeu_vx_u16m1_b16(op0, op1, op2) \ +__builtin_rvv_vmsgeu_vx_u16m1_b16((vuint16m1_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmsgeu_vx_u16m1_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgeu_vx_u16m1_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint16m1_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmsgeu_vx_u16m2_b8(op0, op1, op2) \ +__builtin_rvv_vmsgeu_vx_u16m2_b8((vuint16m2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmsgeu_vx_u16m2_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgeu_vx_u16m2_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint16m2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmsgeu_vx_u16m4_b4(op0, op1, op2) \ +__builtin_rvv_vmsgeu_vx_u16m4_b4((vuint16m4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmsgeu_vx_u16m4_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgeu_vx_u16m4_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vuint16m4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmsgeu_vx_u16m8_b2(op0, op1, op2) \ +__builtin_rvv_vmsgeu_vx_u16m8_b2((vuint16m8_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmsgeu_vx_u16m8_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgeu_vx_u16m8_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vuint16m8_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmsgeu_vx_u16mf2_b32(op0, op1, op2) \ +__builtin_rvv_vmsgeu_vx_u16mf2_b32((vuint16mf2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmsgeu_vx_u16mf2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgeu_vx_u16mf2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint16mf2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmsgeu_vx_u16mf4_b64(op0, op1, op2) \ +__builtin_rvv_vmsgeu_vx_u16mf4_b64((vuint16mf4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmsgeu_vx_u16mf4_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgeu_vx_u16mf4_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint16mf4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmsgeu_vx_u32m1_b32(op0, op1, op2) \ +__builtin_rvv_vmsgeu_vx_u32m1_b32((vuint32m1_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmsgeu_vx_u32m1_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgeu_vx_u32m1_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint32m1_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmsgeu_vx_u32m2_b16(op0, op1, op2) \ +__builtin_rvv_vmsgeu_vx_u32m2_b16((vuint32m2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmsgeu_vx_u32m2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgeu_vx_u32m2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint32m2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmsgeu_vx_u32m4_b8(op0, op1, op2) \ +__builtin_rvv_vmsgeu_vx_u32m4_b8((vuint32m4_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmsgeu_vx_u32m4_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgeu_vx_u32m4_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint32m4_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmsgeu_vx_u32m8_b4(op0, op1, op2) \ +__builtin_rvv_vmsgeu_vx_u32m8_b4((vuint32m8_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmsgeu_vx_u32m8_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgeu_vx_u32m8_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vuint32m8_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmsgeu_vx_u32mf2_b64(op0, op1, op2) \ +__builtin_rvv_vmsgeu_vx_u32mf2_b64((vuint32mf2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmsgeu_vx_u32mf2_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgeu_vx_u32mf2_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint32mf2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmsgeu_vx_u64m1_b64(op0, op1, op2) \ +__builtin_rvv_vmsgeu_vx_u64m1_b64((vuint64m1_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmsgeu_vx_u64m1_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgeu_vx_u64m1_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vuint64m1_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vmsgeu_vx_u64m2_b32(op0, op1, op2) \ +__builtin_rvv_vmsgeu_vx_u64m2_b32((vuint64m2_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmsgeu_vx_u64m2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgeu_vx_u64m2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vuint64m2_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vmsgeu_vx_u64m4_b16(op0, op1, op2) \ +__builtin_rvv_vmsgeu_vx_u64m4_b16((vuint64m4_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmsgeu_vx_u64m4_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgeu_vx_u64m4_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vuint64m4_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vmsgeu_vx_u64m8_b8(op0, op1, op2) \ +__builtin_rvv_vmsgeu_vx_u64m8_b8((vuint64m8_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmsgeu_vx_u64m8_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsgeu_vx_u64m8_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vuint64m8_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vmsge_vv_i8m1_b8(op0, op1, op2) \ +__builtin_rvv_vmsge_vv_i8m1_b8((vint8m1_t)(op0), (vint8m1_t)(op1), (size_t)(op2)) +#define vmsge_vv_i8m1_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsge_vv_i8m1_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vmsge_vv_i8m2_b4(op0, op1, op2) \ +__builtin_rvv_vmsge_vv_i8m2_b4((vint8m2_t)(op0), (vint8m2_t)(op1), (size_t)(op2)) +#define vmsge_vv_i8m2_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsge_vv_i8m2_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vmsge_vv_i8m4_b2(op0, op1, op2) \ +__builtin_rvv_vmsge_vv_i8m4_b2((vint8m4_t)(op0), (vint8m4_t)(op1), (size_t)(op2)) +#define vmsge_vv_i8m4_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsge_vv_i8m4_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vint8m4_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vmsge_vv_i8m8_b1(op0, op1, op2) \ +__builtin_rvv_vmsge_vv_i8m8_b1((vint8m8_t)(op0), (vint8m8_t)(op1), (size_t)(op2)) +#define vmsge_vv_i8m8_b1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsge_vv_i8m8_b1_m((vbool1_t)(op0), (vbool1_t)(op1), (vint8m8_t)(op2), (vint8m8_t)(op3), (size_t)(op4)) +#define vmsge_vv_i8mf2_b16(op0, op1, op2) \ +__builtin_rvv_vmsge_vv_i8mf2_b16((vint8mf2_t)(op0), (vint8mf2_t)(op1), (size_t)(op2)) +#define vmsge_vv_i8mf2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsge_vv_i8mf2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vmsge_vv_i8mf4_b32(op0, op1, op2) \ +__builtin_rvv_vmsge_vv_i8mf4_b32((vint8mf4_t)(op0), (vint8mf4_t)(op1), (size_t)(op2)) +#define vmsge_vv_i8mf4_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsge_vv_i8mf4_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vmsge_vv_i8mf8_b64(op0, op1, op2) \ +__builtin_rvv_vmsge_vv_i8mf8_b64((vint8mf8_t)(op0), (vint8mf8_t)(op1), (size_t)(op2)) +#define vmsge_vv_i8mf8_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsge_vv_i8mf8_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vmsge_vv_i16m1_b16(op0, op1, op2) \ +__builtin_rvv_vmsge_vv_i16m1_b16((vint16m1_t)(op0), (vint16m1_t)(op1), (size_t)(op2)) +#define vmsge_vv_i16m1_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsge_vv_i16m1_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vmsge_vv_i16m2_b8(op0, op1, op2) \ +__builtin_rvv_vmsge_vv_i16m2_b8((vint16m2_t)(op0), (vint16m2_t)(op1), (size_t)(op2)) +#define vmsge_vv_i16m2_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsge_vv_i16m2_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vmsge_vv_i16m4_b4(op0, op1, op2) \ +__builtin_rvv_vmsge_vv_i16m4_b4((vint16m4_t)(op0), (vint16m4_t)(op1), (size_t)(op2)) +#define vmsge_vv_i16m4_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsge_vv_i16m4_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vint16m4_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vmsge_vv_i16m8_b2(op0, op1, op2) \ +__builtin_rvv_vmsge_vv_i16m8_b2((vint16m8_t)(op0), (vint16m8_t)(op1), (size_t)(op2)) +#define vmsge_vv_i16m8_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsge_vv_i16m8_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vint16m8_t)(op2), (vint16m8_t)(op3), (size_t)(op4)) +#define vmsge_vv_i16mf2_b32(op0, op1, op2) \ +__builtin_rvv_vmsge_vv_i16mf2_b32((vint16mf2_t)(op0), (vint16mf2_t)(op1), (size_t)(op2)) +#define vmsge_vv_i16mf2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsge_vv_i16mf2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vmsge_vv_i16mf4_b64(op0, op1, op2) \ +__builtin_rvv_vmsge_vv_i16mf4_b64((vint16mf4_t)(op0), (vint16mf4_t)(op1), (size_t)(op2)) +#define vmsge_vv_i16mf4_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsge_vv_i16mf4_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vmsge_vv_i32m1_b32(op0, op1, op2) \ +__builtin_rvv_vmsge_vv_i32m1_b32((vint32m1_t)(op0), (vint32m1_t)(op1), (size_t)(op2)) +#define vmsge_vv_i32m1_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsge_vv_i32m1_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vmsge_vv_i32m2_b16(op0, op1, op2) \ +__builtin_rvv_vmsge_vv_i32m2_b16((vint32m2_t)(op0), (vint32m2_t)(op1), (size_t)(op2)) +#define vmsge_vv_i32m2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsge_vv_i32m2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vmsge_vv_i32m4_b8(op0, op1, op2) \ +__builtin_rvv_vmsge_vv_i32m4_b8((vint32m4_t)(op0), (vint32m4_t)(op1), (size_t)(op2)) +#define vmsge_vv_i32m4_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsge_vv_i32m4_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vmsge_vv_i32m8_b4(op0, op1, op2) \ +__builtin_rvv_vmsge_vv_i32m8_b4((vint32m8_t)(op0), (vint32m8_t)(op1), (size_t)(op2)) +#define vmsge_vv_i32m8_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsge_vv_i32m8_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vint32m8_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vmsge_vv_i32mf2_b64(op0, op1, op2) \ +__builtin_rvv_vmsge_vv_i32mf2_b64((vint32mf2_t)(op0), (vint32mf2_t)(op1), (size_t)(op2)) +#define vmsge_vv_i32mf2_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsge_vv_i32mf2_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vmsge_vv_i64m1_b64(op0, op1, op2) \ +__builtin_rvv_vmsge_vv_i64m1_b64((vint64m1_t)(op0), (vint64m1_t)(op1), (size_t)(op2)) +#define vmsge_vv_i64m1_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsge_vv_i64m1_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vmsge_vv_i64m2_b32(op0, op1, op2) \ +__builtin_rvv_vmsge_vv_i64m2_b32((vint64m2_t)(op0), (vint64m2_t)(op1), (size_t)(op2)) +#define vmsge_vv_i64m2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsge_vv_i64m2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vmsge_vv_i64m4_b16(op0, op1, op2) \ +__builtin_rvv_vmsge_vv_i64m4_b16((vint64m4_t)(op0), (vint64m4_t)(op1), (size_t)(op2)) +#define vmsge_vv_i64m4_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsge_vv_i64m4_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint64m4_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vmsge_vv_i64m8_b8(op0, op1, op2) \ +__builtin_rvv_vmsge_vv_i64m8_b8((vint64m8_t)(op0), (vint64m8_t)(op1), (size_t)(op2)) +#define vmsge_vv_i64m8_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsge_vv_i64m8_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint64m8_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vmsge_vx_i8m1_b8(op0, op1, op2) \ +__builtin_rvv_vmsge_vx_i8m1_b8((vint8m1_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmsge_vx_i8m1_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsge_vx_i8m1_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint8m1_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmsge_vx_i8m2_b4(op0, op1, op2) \ +__builtin_rvv_vmsge_vx_i8m2_b4((vint8m2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmsge_vx_i8m2_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsge_vx_i8m2_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vint8m2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmsge_vx_i8m4_b2(op0, op1, op2) \ +__builtin_rvv_vmsge_vx_i8m4_b2((vint8m4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmsge_vx_i8m4_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsge_vx_i8m4_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vint8m4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmsge_vx_i8m8_b1(op0, op1, op2) \ +__builtin_rvv_vmsge_vx_i8m8_b1((vint8m8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmsge_vx_i8m8_b1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsge_vx_i8m8_b1_m((vbool1_t)(op0), (vbool1_t)(op1), (vint8m8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmsge_vx_i8mf2_b16(op0, op1, op2) \ +__builtin_rvv_vmsge_vx_i8mf2_b16((vint8mf2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmsge_vx_i8mf2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsge_vx_i8mf2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint8mf2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmsge_vx_i8mf4_b32(op0, op1, op2) \ +__builtin_rvv_vmsge_vx_i8mf4_b32((vint8mf4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmsge_vx_i8mf4_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsge_vx_i8mf4_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint8mf4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmsge_vx_i8mf8_b64(op0, op1, op2) \ +__builtin_rvv_vmsge_vx_i8mf8_b64((vint8mf8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmsge_vx_i8mf8_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsge_vx_i8mf8_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint8mf8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmsge_vx_i16m1_b16(op0, op1, op2) \ +__builtin_rvv_vmsge_vx_i16m1_b16((vint16m1_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmsge_vx_i16m1_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsge_vx_i16m1_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint16m1_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmsge_vx_i16m2_b8(op0, op1, op2) \ +__builtin_rvv_vmsge_vx_i16m2_b8((vint16m2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmsge_vx_i16m2_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsge_vx_i16m2_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint16m2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmsge_vx_i16m4_b4(op0, op1, op2) \ +__builtin_rvv_vmsge_vx_i16m4_b4((vint16m4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmsge_vx_i16m4_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsge_vx_i16m4_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vint16m4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmsge_vx_i16m8_b2(op0, op1, op2) \ +__builtin_rvv_vmsge_vx_i16m8_b2((vint16m8_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmsge_vx_i16m8_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsge_vx_i16m8_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vint16m8_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmsge_vx_i16mf2_b32(op0, op1, op2) \ +__builtin_rvv_vmsge_vx_i16mf2_b32((vint16mf2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmsge_vx_i16mf2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsge_vx_i16mf2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint16mf2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmsge_vx_i16mf4_b64(op0, op1, op2) \ +__builtin_rvv_vmsge_vx_i16mf4_b64((vint16mf4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmsge_vx_i16mf4_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsge_vx_i16mf4_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint16mf4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmsge_vx_i32m1_b32(op0, op1, op2) \ +__builtin_rvv_vmsge_vx_i32m1_b32((vint32m1_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmsge_vx_i32m1_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsge_vx_i32m1_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint32m1_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmsge_vx_i32m2_b16(op0, op1, op2) \ +__builtin_rvv_vmsge_vx_i32m2_b16((vint32m2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmsge_vx_i32m2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsge_vx_i32m2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint32m2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmsge_vx_i32m4_b8(op0, op1, op2) \ +__builtin_rvv_vmsge_vx_i32m4_b8((vint32m4_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmsge_vx_i32m4_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsge_vx_i32m4_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint32m4_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmsge_vx_i32m8_b4(op0, op1, op2) \ +__builtin_rvv_vmsge_vx_i32m8_b4((vint32m8_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmsge_vx_i32m8_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsge_vx_i32m8_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vint32m8_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmsge_vx_i32mf2_b64(op0, op1, op2) \ +__builtin_rvv_vmsge_vx_i32mf2_b64((vint32mf2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmsge_vx_i32mf2_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsge_vx_i32mf2_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint32mf2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmsge_vx_i64m1_b64(op0, op1, op2) \ +__builtin_rvv_vmsge_vx_i64m1_b64((vint64m1_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmsge_vx_i64m1_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsge_vx_i64m1_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vint64m1_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vmsge_vx_i64m2_b32(op0, op1, op2) \ +__builtin_rvv_vmsge_vx_i64m2_b32((vint64m2_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmsge_vx_i64m2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsge_vx_i64m2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vint64m2_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vmsge_vx_i64m4_b16(op0, op1, op2) \ +__builtin_rvv_vmsge_vx_i64m4_b16((vint64m4_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmsge_vx_i64m4_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsge_vx_i64m4_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vint64m4_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vmsge_vx_i64m8_b8(op0, op1, op2) \ +__builtin_rvv_vmsge_vx_i64m8_b8((vint64m8_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmsge_vx_i64m8_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmsge_vx_i64m8_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vint64m8_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vminu_vv_u8m1(op0, op1, op2) \ +__builtin_rvv_vminu_vv_u8m1((vuint8m1_t)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vminu_vv_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vminu_vv_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vminu_vv_u8m2(op0, op1, op2) \ +__builtin_rvv_vminu_vv_u8m2((vuint8m2_t)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vminu_vv_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vminu_vv_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vminu_vv_u8m4(op0, op1, op2) \ +__builtin_rvv_vminu_vv_u8m4((vuint8m4_t)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vminu_vv_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vminu_vv_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vminu_vv_u8m8(op0, op1, op2) \ +__builtin_rvv_vminu_vv_u8m8((vuint8m8_t)(op0), (vuint8m8_t)(op1), (size_t)(op2)) +#define vminu_vv_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vminu_vv_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (vuint8m8_t)(op3), (size_t)(op4)) +#define vminu_vv_u8mf2(op0, op1, op2) \ +__builtin_rvv_vminu_vv_u8mf2((vuint8mf2_t)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vminu_vv_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vminu_vv_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vminu_vv_u8mf4(op0, op1, op2) \ +__builtin_rvv_vminu_vv_u8mf4((vuint8mf4_t)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vminu_vv_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vminu_vv_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vminu_vv_u8mf8(op0, op1, op2) \ +__builtin_rvv_vminu_vv_u8mf8((vuint8mf8_t)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vminu_vv_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vminu_vv_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vminu_vv_u16m1(op0, op1, op2) \ +__builtin_rvv_vminu_vv_u16m1((vuint16m1_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vminu_vv_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vminu_vv_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vminu_vv_u16m2(op0, op1, op2) \ +__builtin_rvv_vminu_vv_u16m2((vuint16m2_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vminu_vv_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vminu_vv_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vminu_vv_u16m4(op0, op1, op2) \ +__builtin_rvv_vminu_vv_u16m4((vuint16m4_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vminu_vv_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vminu_vv_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vminu_vv_u16m8(op0, op1, op2) \ +__builtin_rvv_vminu_vv_u16m8((vuint16m8_t)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vminu_vv_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vminu_vv_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vminu_vv_u16mf2(op0, op1, op2) \ +__builtin_rvv_vminu_vv_u16mf2((vuint16mf2_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vminu_vv_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vminu_vv_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vminu_vv_u16mf4(op0, op1, op2) \ +__builtin_rvv_vminu_vv_u16mf4((vuint16mf4_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vminu_vv_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vminu_vv_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vminu_vv_u32m1(op0, op1, op2) \ +__builtin_rvv_vminu_vv_u32m1((vuint32m1_t)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vminu_vv_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vminu_vv_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vminu_vv_u32m2(op0, op1, op2) \ +__builtin_rvv_vminu_vv_u32m2((vuint32m2_t)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vminu_vv_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vminu_vv_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vminu_vv_u32m4(op0, op1, op2) \ +__builtin_rvv_vminu_vv_u32m4((vuint32m4_t)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vminu_vv_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vminu_vv_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vminu_vv_u32m8(op0, op1, op2) \ +__builtin_rvv_vminu_vv_u32m8((vuint32m8_t)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vminu_vv_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vminu_vv_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vminu_vv_u32mf2(op0, op1, op2) \ +__builtin_rvv_vminu_vv_u32mf2((vuint32mf2_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vminu_vv_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vminu_vv_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vminu_vv_u64m1(op0, op1, op2) \ +__builtin_rvv_vminu_vv_u64m1((vuint64m1_t)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vminu_vv_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vminu_vv_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vminu_vv_u64m2(op0, op1, op2) \ +__builtin_rvv_vminu_vv_u64m2((vuint64m2_t)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vminu_vv_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vminu_vv_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vminu_vv_u64m4(op0, op1, op2) \ +__builtin_rvv_vminu_vv_u64m4((vuint64m4_t)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vminu_vv_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vminu_vv_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vminu_vv_u64m8(op0, op1, op2) \ +__builtin_rvv_vminu_vv_u64m8((vuint64m8_t)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vminu_vv_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vminu_vv_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vminu_vx_u8m1(op0, op1, op2) \ +__builtin_rvv_vminu_vx_u8m1((vuint8m1_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vminu_vx_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vminu_vx_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vminu_vx_u8m2(op0, op1, op2) \ +__builtin_rvv_vminu_vx_u8m2((vuint8m2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vminu_vx_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vminu_vx_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vminu_vx_u8m4(op0, op1, op2) \ +__builtin_rvv_vminu_vx_u8m4((vuint8m4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vminu_vx_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vminu_vx_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vminu_vx_u8m8(op0, op1, op2) \ +__builtin_rvv_vminu_vx_u8m8((vuint8m8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vminu_vx_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vminu_vx_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vminu_vx_u8mf2(op0, op1, op2) \ +__builtin_rvv_vminu_vx_u8mf2((vuint8mf2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vminu_vx_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vminu_vx_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vminu_vx_u8mf4(op0, op1, op2) \ +__builtin_rvv_vminu_vx_u8mf4((vuint8mf4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vminu_vx_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vminu_vx_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vminu_vx_u8mf8(op0, op1, op2) \ +__builtin_rvv_vminu_vx_u8mf8((vuint8mf8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vminu_vx_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vminu_vx_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vminu_vx_u16m1(op0, op1, op2) \ +__builtin_rvv_vminu_vx_u16m1((vuint16m1_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vminu_vx_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vminu_vx_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vminu_vx_u16m2(op0, op1, op2) \ +__builtin_rvv_vminu_vx_u16m2((vuint16m2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vminu_vx_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vminu_vx_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vminu_vx_u16m4(op0, op1, op2) \ +__builtin_rvv_vminu_vx_u16m4((vuint16m4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vminu_vx_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vminu_vx_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vminu_vx_u16m8(op0, op1, op2) \ +__builtin_rvv_vminu_vx_u16m8((vuint16m8_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vminu_vx_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vminu_vx_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vminu_vx_u16mf2(op0, op1, op2) \ +__builtin_rvv_vminu_vx_u16mf2((vuint16mf2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vminu_vx_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vminu_vx_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vminu_vx_u16mf4(op0, op1, op2) \ +__builtin_rvv_vminu_vx_u16mf4((vuint16mf4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vminu_vx_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vminu_vx_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vminu_vx_u32m1(op0, op1, op2) \ +__builtin_rvv_vminu_vx_u32m1((vuint32m1_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vminu_vx_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vminu_vx_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vminu_vx_u32m2(op0, op1, op2) \ +__builtin_rvv_vminu_vx_u32m2((vuint32m2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vminu_vx_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vminu_vx_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vminu_vx_u32m4(op0, op1, op2) \ +__builtin_rvv_vminu_vx_u32m4((vuint32m4_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vminu_vx_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vminu_vx_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vminu_vx_u32m8(op0, op1, op2) \ +__builtin_rvv_vminu_vx_u32m8((vuint32m8_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vminu_vx_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vminu_vx_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vminu_vx_u32mf2(op0, op1, op2) \ +__builtin_rvv_vminu_vx_u32mf2((vuint32mf2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vminu_vx_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vminu_vx_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vminu_vx_u64m1(op0, op1, op2) \ +__builtin_rvv_vminu_vx_u64m1((vuint64m1_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vminu_vx_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vminu_vx_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vminu_vx_u64m2(op0, op1, op2) \ +__builtin_rvv_vminu_vx_u64m2((vuint64m2_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vminu_vx_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vminu_vx_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vminu_vx_u64m4(op0, op1, op2) \ +__builtin_rvv_vminu_vx_u64m4((vuint64m4_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vminu_vx_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vminu_vx_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vminu_vx_u64m8(op0, op1, op2) \ +__builtin_rvv_vminu_vx_u64m8((vuint64m8_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vminu_vx_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vminu_vx_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vmin_vv_i8m1(op0, op1, op2) \ +__builtin_rvv_vmin_vv_i8m1((vint8m1_t)(op0), (vint8m1_t)(op1), (size_t)(op2)) +#define vmin_vv_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmin_vv_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vmin_vv_i8m2(op0, op1, op2) \ +__builtin_rvv_vmin_vv_i8m2((vint8m2_t)(op0), (vint8m2_t)(op1), (size_t)(op2)) +#define vmin_vv_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmin_vv_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vmin_vv_i8m4(op0, op1, op2) \ +__builtin_rvv_vmin_vv_i8m4((vint8m4_t)(op0), (vint8m4_t)(op1), (size_t)(op2)) +#define vmin_vv_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmin_vv_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vmin_vv_i8m8(op0, op1, op2) \ +__builtin_rvv_vmin_vv_i8m8((vint8m8_t)(op0), (vint8m8_t)(op1), (size_t)(op2)) +#define vmin_vv_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmin_vv_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (vint8m8_t)(op3), (size_t)(op4)) +#define vmin_vv_i8mf2(op0, op1, op2) \ +__builtin_rvv_vmin_vv_i8mf2((vint8mf2_t)(op0), (vint8mf2_t)(op1), (size_t)(op2)) +#define vmin_vv_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmin_vv_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vmin_vv_i8mf4(op0, op1, op2) \ +__builtin_rvv_vmin_vv_i8mf4((vint8mf4_t)(op0), (vint8mf4_t)(op1), (size_t)(op2)) +#define vmin_vv_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmin_vv_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vmin_vv_i8mf8(op0, op1, op2) \ +__builtin_rvv_vmin_vv_i8mf8((vint8mf8_t)(op0), (vint8mf8_t)(op1), (size_t)(op2)) +#define vmin_vv_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmin_vv_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vmin_vv_i16m1(op0, op1, op2) \ +__builtin_rvv_vmin_vv_i16m1((vint16m1_t)(op0), (vint16m1_t)(op1), (size_t)(op2)) +#define vmin_vv_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmin_vv_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vmin_vv_i16m2(op0, op1, op2) \ +__builtin_rvv_vmin_vv_i16m2((vint16m2_t)(op0), (vint16m2_t)(op1), (size_t)(op2)) +#define vmin_vv_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmin_vv_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vmin_vv_i16m4(op0, op1, op2) \ +__builtin_rvv_vmin_vv_i16m4((vint16m4_t)(op0), (vint16m4_t)(op1), (size_t)(op2)) +#define vmin_vv_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmin_vv_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vmin_vv_i16m8(op0, op1, op2) \ +__builtin_rvv_vmin_vv_i16m8((vint16m8_t)(op0), (vint16m8_t)(op1), (size_t)(op2)) +#define vmin_vv_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmin_vv_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (vint16m8_t)(op3), (size_t)(op4)) +#define vmin_vv_i16mf2(op0, op1, op2) \ +__builtin_rvv_vmin_vv_i16mf2((vint16mf2_t)(op0), (vint16mf2_t)(op1), (size_t)(op2)) +#define vmin_vv_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmin_vv_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vmin_vv_i16mf4(op0, op1, op2) \ +__builtin_rvv_vmin_vv_i16mf4((vint16mf4_t)(op0), (vint16mf4_t)(op1), (size_t)(op2)) +#define vmin_vv_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmin_vv_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vmin_vv_i32m1(op0, op1, op2) \ +__builtin_rvv_vmin_vv_i32m1((vint32m1_t)(op0), (vint32m1_t)(op1), (size_t)(op2)) +#define vmin_vv_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmin_vv_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vmin_vv_i32m2(op0, op1, op2) \ +__builtin_rvv_vmin_vv_i32m2((vint32m2_t)(op0), (vint32m2_t)(op1), (size_t)(op2)) +#define vmin_vv_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmin_vv_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vmin_vv_i32m4(op0, op1, op2) \ +__builtin_rvv_vmin_vv_i32m4((vint32m4_t)(op0), (vint32m4_t)(op1), (size_t)(op2)) +#define vmin_vv_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmin_vv_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vmin_vv_i32m8(op0, op1, op2) \ +__builtin_rvv_vmin_vv_i32m8((vint32m8_t)(op0), (vint32m8_t)(op1), (size_t)(op2)) +#define vmin_vv_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmin_vv_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vmin_vv_i32mf2(op0, op1, op2) \ +__builtin_rvv_vmin_vv_i32mf2((vint32mf2_t)(op0), (vint32mf2_t)(op1), (size_t)(op2)) +#define vmin_vv_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmin_vv_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vmin_vv_i64m1(op0, op1, op2) \ +__builtin_rvv_vmin_vv_i64m1((vint64m1_t)(op0), (vint64m1_t)(op1), (size_t)(op2)) +#define vmin_vv_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmin_vv_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vmin_vv_i64m2(op0, op1, op2) \ +__builtin_rvv_vmin_vv_i64m2((vint64m2_t)(op0), (vint64m2_t)(op1), (size_t)(op2)) +#define vmin_vv_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmin_vv_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vmin_vv_i64m4(op0, op1, op2) \ +__builtin_rvv_vmin_vv_i64m4((vint64m4_t)(op0), (vint64m4_t)(op1), (size_t)(op2)) +#define vmin_vv_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmin_vv_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vmin_vv_i64m8(op0, op1, op2) \ +__builtin_rvv_vmin_vv_i64m8((vint64m8_t)(op0), (vint64m8_t)(op1), (size_t)(op2)) +#define vmin_vv_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmin_vv_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vmin_vx_i8m1(op0, op1, op2) \ +__builtin_rvv_vmin_vx_i8m1((vint8m1_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmin_vx_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmin_vx_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmin_vx_i8m2(op0, op1, op2) \ +__builtin_rvv_vmin_vx_i8m2((vint8m2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmin_vx_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmin_vx_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmin_vx_i8m4(op0, op1, op2) \ +__builtin_rvv_vmin_vx_i8m4((vint8m4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmin_vx_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmin_vx_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmin_vx_i8m8(op0, op1, op2) \ +__builtin_rvv_vmin_vx_i8m8((vint8m8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmin_vx_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmin_vx_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmin_vx_i8mf2(op0, op1, op2) \ +__builtin_rvv_vmin_vx_i8mf2((vint8mf2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmin_vx_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmin_vx_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmin_vx_i8mf4(op0, op1, op2) \ +__builtin_rvv_vmin_vx_i8mf4((vint8mf4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmin_vx_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmin_vx_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmin_vx_i8mf8(op0, op1, op2) \ +__builtin_rvv_vmin_vx_i8mf8((vint8mf8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmin_vx_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmin_vx_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmin_vx_i16m1(op0, op1, op2) \ +__builtin_rvv_vmin_vx_i16m1((vint16m1_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmin_vx_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmin_vx_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmin_vx_i16m2(op0, op1, op2) \ +__builtin_rvv_vmin_vx_i16m2((vint16m2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmin_vx_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmin_vx_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmin_vx_i16m4(op0, op1, op2) \ +__builtin_rvv_vmin_vx_i16m4((vint16m4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmin_vx_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmin_vx_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmin_vx_i16m8(op0, op1, op2) \ +__builtin_rvv_vmin_vx_i16m8((vint16m8_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmin_vx_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmin_vx_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmin_vx_i16mf2(op0, op1, op2) \ +__builtin_rvv_vmin_vx_i16mf2((vint16mf2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmin_vx_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmin_vx_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmin_vx_i16mf4(op0, op1, op2) \ +__builtin_rvv_vmin_vx_i16mf4((vint16mf4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmin_vx_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmin_vx_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmin_vx_i32m1(op0, op1, op2) \ +__builtin_rvv_vmin_vx_i32m1((vint32m1_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmin_vx_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmin_vx_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmin_vx_i32m2(op0, op1, op2) \ +__builtin_rvv_vmin_vx_i32m2((vint32m2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmin_vx_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmin_vx_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmin_vx_i32m4(op0, op1, op2) \ +__builtin_rvv_vmin_vx_i32m4((vint32m4_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmin_vx_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmin_vx_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmin_vx_i32m8(op0, op1, op2) \ +__builtin_rvv_vmin_vx_i32m8((vint32m8_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmin_vx_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmin_vx_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmin_vx_i32mf2(op0, op1, op2) \ +__builtin_rvv_vmin_vx_i32mf2((vint32mf2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmin_vx_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmin_vx_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmin_vx_i64m1(op0, op1, op2) \ +__builtin_rvv_vmin_vx_i64m1((vint64m1_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmin_vx_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmin_vx_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vmin_vx_i64m2(op0, op1, op2) \ +__builtin_rvv_vmin_vx_i64m2((vint64m2_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmin_vx_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmin_vx_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vmin_vx_i64m4(op0, op1, op2) \ +__builtin_rvv_vmin_vx_i64m4((vint64m4_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmin_vx_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmin_vx_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vmin_vx_i64m8(op0, op1, op2) \ +__builtin_rvv_vmin_vx_i64m8((vint64m8_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmin_vx_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmin_vx_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vmaxu_vv_u8m1(op0, op1, op2) \ +__builtin_rvv_vmaxu_vv_u8m1((vuint8m1_t)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vmaxu_vv_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmaxu_vv_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vmaxu_vv_u8m2(op0, op1, op2) \ +__builtin_rvv_vmaxu_vv_u8m2((vuint8m2_t)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vmaxu_vv_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmaxu_vv_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vmaxu_vv_u8m4(op0, op1, op2) \ +__builtin_rvv_vmaxu_vv_u8m4((vuint8m4_t)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vmaxu_vv_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmaxu_vv_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vmaxu_vv_u8m8(op0, op1, op2) \ +__builtin_rvv_vmaxu_vv_u8m8((vuint8m8_t)(op0), (vuint8m8_t)(op1), (size_t)(op2)) +#define vmaxu_vv_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmaxu_vv_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (vuint8m8_t)(op3), (size_t)(op4)) +#define vmaxu_vv_u8mf2(op0, op1, op2) \ +__builtin_rvv_vmaxu_vv_u8mf2((vuint8mf2_t)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vmaxu_vv_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmaxu_vv_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vmaxu_vv_u8mf4(op0, op1, op2) \ +__builtin_rvv_vmaxu_vv_u8mf4((vuint8mf4_t)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vmaxu_vv_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmaxu_vv_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vmaxu_vv_u8mf8(op0, op1, op2) \ +__builtin_rvv_vmaxu_vv_u8mf8((vuint8mf8_t)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vmaxu_vv_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmaxu_vv_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vmaxu_vv_u16m1(op0, op1, op2) \ +__builtin_rvv_vmaxu_vv_u16m1((vuint16m1_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vmaxu_vv_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmaxu_vv_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vmaxu_vv_u16m2(op0, op1, op2) \ +__builtin_rvv_vmaxu_vv_u16m2((vuint16m2_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vmaxu_vv_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmaxu_vv_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vmaxu_vv_u16m4(op0, op1, op2) \ +__builtin_rvv_vmaxu_vv_u16m4((vuint16m4_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vmaxu_vv_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmaxu_vv_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vmaxu_vv_u16m8(op0, op1, op2) \ +__builtin_rvv_vmaxu_vv_u16m8((vuint16m8_t)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vmaxu_vv_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmaxu_vv_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vmaxu_vv_u16mf2(op0, op1, op2) \ +__builtin_rvv_vmaxu_vv_u16mf2((vuint16mf2_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vmaxu_vv_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmaxu_vv_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vmaxu_vv_u16mf4(op0, op1, op2) \ +__builtin_rvv_vmaxu_vv_u16mf4((vuint16mf4_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vmaxu_vv_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmaxu_vv_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vmaxu_vv_u32m1(op0, op1, op2) \ +__builtin_rvv_vmaxu_vv_u32m1((vuint32m1_t)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vmaxu_vv_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmaxu_vv_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vmaxu_vv_u32m2(op0, op1, op2) \ +__builtin_rvv_vmaxu_vv_u32m2((vuint32m2_t)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vmaxu_vv_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmaxu_vv_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vmaxu_vv_u32m4(op0, op1, op2) \ +__builtin_rvv_vmaxu_vv_u32m4((vuint32m4_t)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vmaxu_vv_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmaxu_vv_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vmaxu_vv_u32m8(op0, op1, op2) \ +__builtin_rvv_vmaxu_vv_u32m8((vuint32m8_t)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vmaxu_vv_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmaxu_vv_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vmaxu_vv_u32mf2(op0, op1, op2) \ +__builtin_rvv_vmaxu_vv_u32mf2((vuint32mf2_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vmaxu_vv_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmaxu_vv_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vmaxu_vv_u64m1(op0, op1, op2) \ +__builtin_rvv_vmaxu_vv_u64m1((vuint64m1_t)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vmaxu_vv_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmaxu_vv_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vmaxu_vv_u64m2(op0, op1, op2) \ +__builtin_rvv_vmaxu_vv_u64m2((vuint64m2_t)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vmaxu_vv_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmaxu_vv_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vmaxu_vv_u64m4(op0, op1, op2) \ +__builtin_rvv_vmaxu_vv_u64m4((vuint64m4_t)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vmaxu_vv_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmaxu_vv_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vmaxu_vv_u64m8(op0, op1, op2) \ +__builtin_rvv_vmaxu_vv_u64m8((vuint64m8_t)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vmaxu_vv_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmaxu_vv_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vmaxu_vx_u8m1(op0, op1, op2) \ +__builtin_rvv_vmaxu_vx_u8m1((vuint8m1_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmaxu_vx_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmaxu_vx_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmaxu_vx_u8m2(op0, op1, op2) \ +__builtin_rvv_vmaxu_vx_u8m2((vuint8m2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmaxu_vx_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmaxu_vx_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmaxu_vx_u8m4(op0, op1, op2) \ +__builtin_rvv_vmaxu_vx_u8m4((vuint8m4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmaxu_vx_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmaxu_vx_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmaxu_vx_u8m8(op0, op1, op2) \ +__builtin_rvv_vmaxu_vx_u8m8((vuint8m8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmaxu_vx_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmaxu_vx_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmaxu_vx_u8mf2(op0, op1, op2) \ +__builtin_rvv_vmaxu_vx_u8mf2((vuint8mf2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmaxu_vx_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmaxu_vx_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmaxu_vx_u8mf4(op0, op1, op2) \ +__builtin_rvv_vmaxu_vx_u8mf4((vuint8mf4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmaxu_vx_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmaxu_vx_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmaxu_vx_u8mf8(op0, op1, op2) \ +__builtin_rvv_vmaxu_vx_u8mf8((vuint8mf8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmaxu_vx_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmaxu_vx_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmaxu_vx_u16m1(op0, op1, op2) \ +__builtin_rvv_vmaxu_vx_u16m1((vuint16m1_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmaxu_vx_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmaxu_vx_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmaxu_vx_u16m2(op0, op1, op2) \ +__builtin_rvv_vmaxu_vx_u16m2((vuint16m2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmaxu_vx_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmaxu_vx_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmaxu_vx_u16m4(op0, op1, op2) \ +__builtin_rvv_vmaxu_vx_u16m4((vuint16m4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmaxu_vx_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmaxu_vx_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmaxu_vx_u16m8(op0, op1, op2) \ +__builtin_rvv_vmaxu_vx_u16m8((vuint16m8_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmaxu_vx_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmaxu_vx_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmaxu_vx_u16mf2(op0, op1, op2) \ +__builtin_rvv_vmaxu_vx_u16mf2((vuint16mf2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmaxu_vx_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmaxu_vx_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmaxu_vx_u16mf4(op0, op1, op2) \ +__builtin_rvv_vmaxu_vx_u16mf4((vuint16mf4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmaxu_vx_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmaxu_vx_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmaxu_vx_u32m1(op0, op1, op2) \ +__builtin_rvv_vmaxu_vx_u32m1((vuint32m1_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmaxu_vx_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmaxu_vx_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmaxu_vx_u32m2(op0, op1, op2) \ +__builtin_rvv_vmaxu_vx_u32m2((vuint32m2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmaxu_vx_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmaxu_vx_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmaxu_vx_u32m4(op0, op1, op2) \ +__builtin_rvv_vmaxu_vx_u32m4((vuint32m4_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmaxu_vx_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmaxu_vx_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmaxu_vx_u32m8(op0, op1, op2) \ +__builtin_rvv_vmaxu_vx_u32m8((vuint32m8_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmaxu_vx_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmaxu_vx_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmaxu_vx_u32mf2(op0, op1, op2) \ +__builtin_rvv_vmaxu_vx_u32mf2((vuint32mf2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmaxu_vx_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmaxu_vx_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmaxu_vx_u64m1(op0, op1, op2) \ +__builtin_rvv_vmaxu_vx_u64m1((vuint64m1_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmaxu_vx_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmaxu_vx_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vmaxu_vx_u64m2(op0, op1, op2) \ +__builtin_rvv_vmaxu_vx_u64m2((vuint64m2_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmaxu_vx_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmaxu_vx_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vmaxu_vx_u64m4(op0, op1, op2) \ +__builtin_rvv_vmaxu_vx_u64m4((vuint64m4_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmaxu_vx_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmaxu_vx_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vmaxu_vx_u64m8(op0, op1, op2) \ +__builtin_rvv_vmaxu_vx_u64m8((vuint64m8_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmaxu_vx_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmaxu_vx_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vmax_vv_i8m1(op0, op1, op2) \ +__builtin_rvv_vmax_vv_i8m1((vint8m1_t)(op0), (vint8m1_t)(op1), (size_t)(op2)) +#define vmax_vv_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmax_vv_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vmax_vv_i8m2(op0, op1, op2) \ +__builtin_rvv_vmax_vv_i8m2((vint8m2_t)(op0), (vint8m2_t)(op1), (size_t)(op2)) +#define vmax_vv_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmax_vv_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vmax_vv_i8m4(op0, op1, op2) \ +__builtin_rvv_vmax_vv_i8m4((vint8m4_t)(op0), (vint8m4_t)(op1), (size_t)(op2)) +#define vmax_vv_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmax_vv_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vmax_vv_i8m8(op0, op1, op2) \ +__builtin_rvv_vmax_vv_i8m8((vint8m8_t)(op0), (vint8m8_t)(op1), (size_t)(op2)) +#define vmax_vv_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmax_vv_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (vint8m8_t)(op3), (size_t)(op4)) +#define vmax_vv_i8mf2(op0, op1, op2) \ +__builtin_rvv_vmax_vv_i8mf2((vint8mf2_t)(op0), (vint8mf2_t)(op1), (size_t)(op2)) +#define vmax_vv_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmax_vv_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vmax_vv_i8mf4(op0, op1, op2) \ +__builtin_rvv_vmax_vv_i8mf4((vint8mf4_t)(op0), (vint8mf4_t)(op1), (size_t)(op2)) +#define vmax_vv_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmax_vv_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vmax_vv_i8mf8(op0, op1, op2) \ +__builtin_rvv_vmax_vv_i8mf8((vint8mf8_t)(op0), (vint8mf8_t)(op1), (size_t)(op2)) +#define vmax_vv_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmax_vv_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vmax_vv_i16m1(op0, op1, op2) \ +__builtin_rvv_vmax_vv_i16m1((vint16m1_t)(op0), (vint16m1_t)(op1), (size_t)(op2)) +#define vmax_vv_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmax_vv_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vmax_vv_i16m2(op0, op1, op2) \ +__builtin_rvv_vmax_vv_i16m2((vint16m2_t)(op0), (vint16m2_t)(op1), (size_t)(op2)) +#define vmax_vv_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmax_vv_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vmax_vv_i16m4(op0, op1, op2) \ +__builtin_rvv_vmax_vv_i16m4((vint16m4_t)(op0), (vint16m4_t)(op1), (size_t)(op2)) +#define vmax_vv_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmax_vv_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vmax_vv_i16m8(op0, op1, op2) \ +__builtin_rvv_vmax_vv_i16m8((vint16m8_t)(op0), (vint16m8_t)(op1), (size_t)(op2)) +#define vmax_vv_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmax_vv_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (vint16m8_t)(op3), (size_t)(op4)) +#define vmax_vv_i16mf2(op0, op1, op2) \ +__builtin_rvv_vmax_vv_i16mf2((vint16mf2_t)(op0), (vint16mf2_t)(op1), (size_t)(op2)) +#define vmax_vv_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmax_vv_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vmax_vv_i16mf4(op0, op1, op2) \ +__builtin_rvv_vmax_vv_i16mf4((vint16mf4_t)(op0), (vint16mf4_t)(op1), (size_t)(op2)) +#define vmax_vv_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmax_vv_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vmax_vv_i32m1(op0, op1, op2) \ +__builtin_rvv_vmax_vv_i32m1((vint32m1_t)(op0), (vint32m1_t)(op1), (size_t)(op2)) +#define vmax_vv_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmax_vv_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vmax_vv_i32m2(op0, op1, op2) \ +__builtin_rvv_vmax_vv_i32m2((vint32m2_t)(op0), (vint32m2_t)(op1), (size_t)(op2)) +#define vmax_vv_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmax_vv_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vmax_vv_i32m4(op0, op1, op2) \ +__builtin_rvv_vmax_vv_i32m4((vint32m4_t)(op0), (vint32m4_t)(op1), (size_t)(op2)) +#define vmax_vv_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmax_vv_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vmax_vv_i32m8(op0, op1, op2) \ +__builtin_rvv_vmax_vv_i32m8((vint32m8_t)(op0), (vint32m8_t)(op1), (size_t)(op2)) +#define vmax_vv_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmax_vv_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vmax_vv_i32mf2(op0, op1, op2) \ +__builtin_rvv_vmax_vv_i32mf2((vint32mf2_t)(op0), (vint32mf2_t)(op1), (size_t)(op2)) +#define vmax_vv_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmax_vv_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vmax_vv_i64m1(op0, op1, op2) \ +__builtin_rvv_vmax_vv_i64m1((vint64m1_t)(op0), (vint64m1_t)(op1), (size_t)(op2)) +#define vmax_vv_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmax_vv_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vmax_vv_i64m2(op0, op1, op2) \ +__builtin_rvv_vmax_vv_i64m2((vint64m2_t)(op0), (vint64m2_t)(op1), (size_t)(op2)) +#define vmax_vv_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmax_vv_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vmax_vv_i64m4(op0, op1, op2) \ +__builtin_rvv_vmax_vv_i64m4((vint64m4_t)(op0), (vint64m4_t)(op1), (size_t)(op2)) +#define vmax_vv_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmax_vv_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vmax_vv_i64m8(op0, op1, op2) \ +__builtin_rvv_vmax_vv_i64m8((vint64m8_t)(op0), (vint64m8_t)(op1), (size_t)(op2)) +#define vmax_vv_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmax_vv_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vsuxei64_v_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei64_v_u16m1((uint16_t *)(op0), (vuint64m4_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vsuxei64_v_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei64_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint64m4_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vsuxei64_v_u16m2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei64_v_u16m2((uint16_t *)(op0), (vuint64m8_t)(op1), (vuint16m2_t)(op2), (size_t)(op3)) +#define vsuxei64_v_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei64_v_u16m2_m((vbool8_t)(op0), (uint16_t *)(op1), (vuint64m8_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vsuxei64_v_u16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei64_v_u16mf2((uint16_t *)(op0), (vuint64m2_t)(op1), (vuint16mf2_t)(op2), (size_t)(op3)) +#define vsuxei64_v_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei64_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint64m2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vsuxei64_v_u16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei64_v_u16mf4((uint16_t *)(op0), (vuint64m1_t)(op1), (vuint16mf4_t)(op2), (size_t)(op3)) +#define vsuxei64_v_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei64_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint64m1_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vmax_vx_i8m1(op0, op1, op2) \ +__builtin_rvv_vmax_vx_i8m1((vint8m1_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmax_vx_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmax_vx_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmax_vx_i8m2(op0, op1, op2) \ +__builtin_rvv_vmax_vx_i8m2((vint8m2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmax_vx_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmax_vx_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmax_vx_i8m4(op0, op1, op2) \ +__builtin_rvv_vmax_vx_i8m4((vint8m4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmax_vx_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmax_vx_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmax_vx_i8m8(op0, op1, op2) \ +__builtin_rvv_vmax_vx_i8m8((vint8m8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmax_vx_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmax_vx_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmax_vx_i8mf2(op0, op1, op2) \ +__builtin_rvv_vmax_vx_i8mf2((vint8mf2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmax_vx_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmax_vx_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmax_vx_i8mf4(op0, op1, op2) \ +__builtin_rvv_vmax_vx_i8mf4((vint8mf4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmax_vx_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmax_vx_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmax_vx_i8mf8(op0, op1, op2) \ +__builtin_rvv_vmax_vx_i8mf8((vint8mf8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmax_vx_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmax_vx_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmax_vx_i16m1(op0, op1, op2) \ +__builtin_rvv_vmax_vx_i16m1((vint16m1_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmax_vx_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmax_vx_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmax_vx_i16m2(op0, op1, op2) \ +__builtin_rvv_vmax_vx_i16m2((vint16m2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmax_vx_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmax_vx_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmax_vx_i16m4(op0, op1, op2) \ +__builtin_rvv_vmax_vx_i16m4((vint16m4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmax_vx_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmax_vx_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmax_vx_i16m8(op0, op1, op2) \ +__builtin_rvv_vmax_vx_i16m8((vint16m8_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmax_vx_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmax_vx_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmax_vx_i16mf2(op0, op1, op2) \ +__builtin_rvv_vmax_vx_i16mf2((vint16mf2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmax_vx_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmax_vx_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmax_vx_i16mf4(op0, op1, op2) \ +__builtin_rvv_vmax_vx_i16mf4((vint16mf4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmax_vx_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmax_vx_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmax_vx_i32m1(op0, op1, op2) \ +__builtin_rvv_vmax_vx_i32m1((vint32m1_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmax_vx_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmax_vx_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmax_vx_i32m2(op0, op1, op2) \ +__builtin_rvv_vmax_vx_i32m2((vint32m2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmax_vx_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmax_vx_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmax_vx_i32m4(op0, op1, op2) \ +__builtin_rvv_vmax_vx_i32m4((vint32m4_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmax_vx_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmax_vx_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmax_vx_i32m8(op0, op1, op2) \ +__builtin_rvv_vmax_vx_i32m8((vint32m8_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmax_vx_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmax_vx_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmax_vx_i32mf2(op0, op1, op2) \ +__builtin_rvv_vmax_vx_i32mf2((vint32mf2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmax_vx_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmax_vx_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmax_vx_i64m1(op0, op1, op2) \ +__builtin_rvv_vmax_vx_i64m1((vint64m1_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmax_vx_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmax_vx_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vmax_vx_i64m2(op0, op1, op2) \ +__builtin_rvv_vmax_vx_i64m2((vint64m2_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmax_vx_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmax_vx_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vmax_vx_i64m4(op0, op1, op2) \ +__builtin_rvv_vmax_vx_i64m4((vint64m4_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmax_vx_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmax_vx_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vmax_vx_i64m8(op0, op1, op2) \ +__builtin_rvv_vmax_vx_i64m8((vint64m8_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmax_vx_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmax_vx_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vmul_vv_i8m1(op0, op1, op2) \ +__builtin_rvv_vmul_vv_i8m1((vint8m1_t)(op0), (vint8m1_t)(op1), (size_t)(op2)) +#define vmul_vv_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vv_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vmul_vv_i8m2(op0, op1, op2) \ +__builtin_rvv_vmul_vv_i8m2((vint8m2_t)(op0), (vint8m2_t)(op1), (size_t)(op2)) +#define vmul_vv_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vv_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vmul_vv_i8m4(op0, op1, op2) \ +__builtin_rvv_vmul_vv_i8m4((vint8m4_t)(op0), (vint8m4_t)(op1), (size_t)(op2)) +#define vmul_vv_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vv_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vmul_vv_i8m8(op0, op1, op2) \ +__builtin_rvv_vmul_vv_i8m8((vint8m8_t)(op0), (vint8m8_t)(op1), (size_t)(op2)) +#define vmul_vv_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vv_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (vint8m8_t)(op3), (size_t)(op4)) +#define vmul_vv_i8mf2(op0, op1, op2) \ +__builtin_rvv_vmul_vv_i8mf2((vint8mf2_t)(op0), (vint8mf2_t)(op1), (size_t)(op2)) +#define vmul_vv_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vv_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vmul_vv_i8mf4(op0, op1, op2) \ +__builtin_rvv_vmul_vv_i8mf4((vint8mf4_t)(op0), (vint8mf4_t)(op1), (size_t)(op2)) +#define vmul_vv_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vv_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vmul_vv_i8mf8(op0, op1, op2) \ +__builtin_rvv_vmul_vv_i8mf8((vint8mf8_t)(op0), (vint8mf8_t)(op1), (size_t)(op2)) +#define vmul_vv_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vv_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vmul_vv_i16m1(op0, op1, op2) \ +__builtin_rvv_vmul_vv_i16m1((vint16m1_t)(op0), (vint16m1_t)(op1), (size_t)(op2)) +#define vmul_vv_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vv_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vmul_vv_i16m2(op0, op1, op2) \ +__builtin_rvv_vmul_vv_i16m2((vint16m2_t)(op0), (vint16m2_t)(op1), (size_t)(op2)) +#define vmul_vv_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vv_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vmul_vv_i16m4(op0, op1, op2) \ +__builtin_rvv_vmul_vv_i16m4((vint16m4_t)(op0), (vint16m4_t)(op1), (size_t)(op2)) +#define vmul_vv_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vv_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vmul_vv_i16m8(op0, op1, op2) \ +__builtin_rvv_vmul_vv_i16m8((vint16m8_t)(op0), (vint16m8_t)(op1), (size_t)(op2)) +#define vmul_vv_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vv_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (vint16m8_t)(op3), (size_t)(op4)) +#define vmul_vv_i16mf2(op0, op1, op2) \ +__builtin_rvv_vmul_vv_i16mf2((vint16mf2_t)(op0), (vint16mf2_t)(op1), (size_t)(op2)) +#define vmul_vv_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vv_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vmul_vv_i16mf4(op0, op1, op2) \ +__builtin_rvv_vmul_vv_i16mf4((vint16mf4_t)(op0), (vint16mf4_t)(op1), (size_t)(op2)) +#define vmul_vv_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vv_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vmul_vv_i32m1(op0, op1, op2) \ +__builtin_rvv_vmul_vv_i32m1((vint32m1_t)(op0), (vint32m1_t)(op1), (size_t)(op2)) +#define vmul_vv_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vv_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vmul_vv_i32m2(op0, op1, op2) \ +__builtin_rvv_vmul_vv_i32m2((vint32m2_t)(op0), (vint32m2_t)(op1), (size_t)(op2)) +#define vmul_vv_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vv_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vmul_vv_i32m4(op0, op1, op2) \ +__builtin_rvv_vmul_vv_i32m4((vint32m4_t)(op0), (vint32m4_t)(op1), (size_t)(op2)) +#define vmul_vv_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vv_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vmul_vv_i32m8(op0, op1, op2) \ +__builtin_rvv_vmul_vv_i32m8((vint32m8_t)(op0), (vint32m8_t)(op1), (size_t)(op2)) +#define vmul_vv_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vv_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vmul_vv_i32mf2(op0, op1, op2) \ +__builtin_rvv_vmul_vv_i32mf2((vint32mf2_t)(op0), (vint32mf2_t)(op1), (size_t)(op2)) +#define vmul_vv_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vv_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vmul_vv_i64m1(op0, op1, op2) \ +__builtin_rvv_vmul_vv_i64m1((vint64m1_t)(op0), (vint64m1_t)(op1), (size_t)(op2)) +#define vmul_vv_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vv_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vmul_vv_i64m2(op0, op1, op2) \ +__builtin_rvv_vmul_vv_i64m2((vint64m2_t)(op0), (vint64m2_t)(op1), (size_t)(op2)) +#define vmul_vv_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vv_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vmul_vv_i64m4(op0, op1, op2) \ +__builtin_rvv_vmul_vv_i64m4((vint64m4_t)(op0), (vint64m4_t)(op1), (size_t)(op2)) +#define vmul_vv_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vv_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vmul_vv_i64m8(op0, op1, op2) \ +__builtin_rvv_vmul_vv_i64m8((vint64m8_t)(op0), (vint64m8_t)(op1), (size_t)(op2)) +#define vmul_vv_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vv_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vmul_vx_i8m1(op0, op1, op2) \ +__builtin_rvv_vmul_vx_i8m1((vint8m1_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmul_vx_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vx_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmul_vx_i8m2(op0, op1, op2) \ +__builtin_rvv_vmul_vx_i8m2((vint8m2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmul_vx_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vx_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmul_vx_i8m4(op0, op1, op2) \ +__builtin_rvv_vmul_vx_i8m4((vint8m4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmul_vx_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vx_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmul_vx_i8m8(op0, op1, op2) \ +__builtin_rvv_vmul_vx_i8m8((vint8m8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmul_vx_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vx_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmul_vx_i8mf2(op0, op1, op2) \ +__builtin_rvv_vmul_vx_i8mf2((vint8mf2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmul_vx_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vx_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmul_vx_i8mf4(op0, op1, op2) \ +__builtin_rvv_vmul_vx_i8mf4((vint8mf4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmul_vx_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vx_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmul_vx_i8mf8(op0, op1, op2) \ +__builtin_rvv_vmul_vx_i8mf8((vint8mf8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmul_vx_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vx_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmul_vx_i16m1(op0, op1, op2) \ +__builtin_rvv_vmul_vx_i16m1((vint16m1_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmul_vx_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vx_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmul_vx_i16m2(op0, op1, op2) \ +__builtin_rvv_vmul_vx_i16m2((vint16m2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmul_vx_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vx_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmul_vx_i16m4(op0, op1, op2) \ +__builtin_rvv_vmul_vx_i16m4((vint16m4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmul_vx_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vx_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmul_vx_i16m8(op0, op1, op2) \ +__builtin_rvv_vmul_vx_i16m8((vint16m8_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmul_vx_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vx_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmul_vx_i16mf2(op0, op1, op2) \ +__builtin_rvv_vmul_vx_i16mf2((vint16mf2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmul_vx_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vx_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmul_vx_i16mf4(op0, op1, op2) \ +__builtin_rvv_vmul_vx_i16mf4((vint16mf4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmul_vx_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vx_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmul_vx_i32m1(op0, op1, op2) \ +__builtin_rvv_vmul_vx_i32m1((vint32m1_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmul_vx_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vx_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmul_vx_i32m2(op0, op1, op2) \ +__builtin_rvv_vmul_vx_i32m2((vint32m2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmul_vx_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vx_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmul_vx_i32m4(op0, op1, op2) \ +__builtin_rvv_vmul_vx_i32m4((vint32m4_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmul_vx_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vx_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmul_vx_i32m8(op0, op1, op2) \ +__builtin_rvv_vmul_vx_i32m8((vint32m8_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmul_vx_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vx_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmul_vx_i32mf2(op0, op1, op2) \ +__builtin_rvv_vmul_vx_i32mf2((vint32mf2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmul_vx_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vx_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmul_vx_i64m1(op0, op1, op2) \ +__builtin_rvv_vmul_vx_i64m1((vint64m1_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmul_vx_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vx_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vmul_vx_i64m2(op0, op1, op2) \ +__builtin_rvv_vmul_vx_i64m2((vint64m2_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmul_vx_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vx_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vmul_vx_i64m4(op0, op1, op2) \ +__builtin_rvv_vmul_vx_i64m4((vint64m4_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmul_vx_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vx_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vmul_vx_i64m8(op0, op1, op2) \ +__builtin_rvv_vmul_vx_i64m8((vint64m8_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmul_vx_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vx_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vmul_vv_u8m1(op0, op1, op2) \ +__builtin_rvv_vmul_vv_u8m1((vuint8m1_t)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vmul_vv_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vv_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vmul_vv_u8m2(op0, op1, op2) \ +__builtin_rvv_vmul_vv_u8m2((vuint8m2_t)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vmul_vv_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vv_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vmul_vv_u8m4(op0, op1, op2) \ +__builtin_rvv_vmul_vv_u8m4((vuint8m4_t)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vmul_vv_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vv_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vmul_vv_u8m8(op0, op1, op2) \ +__builtin_rvv_vmul_vv_u8m8((vuint8m8_t)(op0), (vuint8m8_t)(op1), (size_t)(op2)) +#define vmul_vv_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vv_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (vuint8m8_t)(op3), (size_t)(op4)) +#define vmul_vv_u8mf2(op0, op1, op2) \ +__builtin_rvv_vmul_vv_u8mf2((vuint8mf2_t)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vmul_vv_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vv_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vmul_vv_u8mf4(op0, op1, op2) \ +__builtin_rvv_vmul_vv_u8mf4((vuint8mf4_t)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vmul_vv_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vv_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vmul_vv_u8mf8(op0, op1, op2) \ +__builtin_rvv_vmul_vv_u8mf8((vuint8mf8_t)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vmul_vv_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vv_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vmul_vv_u16m1(op0, op1, op2) \ +__builtin_rvv_vmul_vv_u16m1((vuint16m1_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vmul_vv_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vv_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vmul_vv_u16m2(op0, op1, op2) \ +__builtin_rvv_vmul_vv_u16m2((vuint16m2_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vmul_vv_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vv_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vmul_vv_u16m4(op0, op1, op2) \ +__builtin_rvv_vmul_vv_u16m4((vuint16m4_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vmul_vv_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vv_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vmul_vv_u16m8(op0, op1, op2) \ +__builtin_rvv_vmul_vv_u16m8((vuint16m8_t)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vmul_vv_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vv_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vmul_vv_u16mf2(op0, op1, op2) \ +__builtin_rvv_vmul_vv_u16mf2((vuint16mf2_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vmul_vv_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vv_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vmul_vv_u16mf4(op0, op1, op2) \ +__builtin_rvv_vmul_vv_u16mf4((vuint16mf4_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vmul_vv_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vv_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vmul_vv_u32m1(op0, op1, op2) \ +__builtin_rvv_vmul_vv_u32m1((vuint32m1_t)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vmul_vv_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vv_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vmul_vv_u32m2(op0, op1, op2) \ +__builtin_rvv_vmul_vv_u32m2((vuint32m2_t)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vmul_vv_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vv_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vmul_vv_u32m4(op0, op1, op2) \ +__builtin_rvv_vmul_vv_u32m4((vuint32m4_t)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vmul_vv_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vv_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vmul_vv_u32m8(op0, op1, op2) \ +__builtin_rvv_vmul_vv_u32m8((vuint32m8_t)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vmul_vv_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vv_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vmul_vv_u32mf2(op0, op1, op2) \ +__builtin_rvv_vmul_vv_u32mf2((vuint32mf2_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vmul_vv_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vv_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vmul_vv_u64m1(op0, op1, op2) \ +__builtin_rvv_vmul_vv_u64m1((vuint64m1_t)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vmul_vv_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vv_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vmul_vv_u64m2(op0, op1, op2) \ +__builtin_rvv_vmul_vv_u64m2((vuint64m2_t)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vmul_vv_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vv_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vmul_vv_u64m4(op0, op1, op2) \ +__builtin_rvv_vmul_vv_u64m4((vuint64m4_t)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vmul_vv_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vv_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vmul_vv_u64m8(op0, op1, op2) \ +__builtin_rvv_vmul_vv_u64m8((vuint64m8_t)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vmul_vv_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vv_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vmul_vx_u8m1(op0, op1, op2) \ +__builtin_rvv_vmul_vx_u8m1((vuint8m1_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmul_vx_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vx_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmul_vx_u8m2(op0, op1, op2) \ +__builtin_rvv_vmul_vx_u8m2((vuint8m2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmul_vx_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vx_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmul_vx_u8m4(op0, op1, op2) \ +__builtin_rvv_vmul_vx_u8m4((vuint8m4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmul_vx_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vx_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmul_vx_u8m8(op0, op1, op2) \ +__builtin_rvv_vmul_vx_u8m8((vuint8m8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmul_vx_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vx_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmul_vx_u8mf2(op0, op1, op2) \ +__builtin_rvv_vmul_vx_u8mf2((vuint8mf2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmul_vx_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vx_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmul_vx_u8mf4(op0, op1, op2) \ +__builtin_rvv_vmul_vx_u8mf4((vuint8mf4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmul_vx_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vx_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmul_vx_u8mf8(op0, op1, op2) \ +__builtin_rvv_vmul_vx_u8mf8((vuint8mf8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmul_vx_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vx_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmul_vx_u16m1(op0, op1, op2) \ +__builtin_rvv_vmul_vx_u16m1((vuint16m1_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmul_vx_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vx_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmul_vx_u16m2(op0, op1, op2) \ +__builtin_rvv_vmul_vx_u16m2((vuint16m2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmul_vx_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vx_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmul_vx_u16m4(op0, op1, op2) \ +__builtin_rvv_vmul_vx_u16m4((vuint16m4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmul_vx_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vx_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmul_vx_u16m8(op0, op1, op2) \ +__builtin_rvv_vmul_vx_u16m8((vuint16m8_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmul_vx_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vx_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmul_vx_u16mf2(op0, op1, op2) \ +__builtin_rvv_vmul_vx_u16mf2((vuint16mf2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmul_vx_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vx_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmul_vx_u16mf4(op0, op1, op2) \ +__builtin_rvv_vmul_vx_u16mf4((vuint16mf4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmul_vx_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vx_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmul_vx_u32m1(op0, op1, op2) \ +__builtin_rvv_vmul_vx_u32m1((vuint32m1_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmul_vx_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vx_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmul_vx_u32m2(op0, op1, op2) \ +__builtin_rvv_vmul_vx_u32m2((vuint32m2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmul_vx_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vx_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmul_vx_u32m4(op0, op1, op2) \ +__builtin_rvv_vmul_vx_u32m4((vuint32m4_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmul_vx_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vx_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmul_vx_u32m8(op0, op1, op2) \ +__builtin_rvv_vmul_vx_u32m8((vuint32m8_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmul_vx_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vx_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmul_vx_u32mf2(op0, op1, op2) \ +__builtin_rvv_vmul_vx_u32mf2((vuint32mf2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmul_vx_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vx_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmul_vx_u64m1(op0, op1, op2) \ +__builtin_rvv_vmul_vx_u64m1((vuint64m1_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmul_vx_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vx_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vmul_vx_u64m2(op0, op1, op2) \ +__builtin_rvv_vmul_vx_u64m2((vuint64m2_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmul_vx_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vx_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vmul_vx_u64m4(op0, op1, op2) \ +__builtin_rvv_vmul_vx_u64m4((vuint64m4_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmul_vx_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vx_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vmul_vx_u64m8(op0, op1, op2) \ +__builtin_rvv_vmul_vx_u64m8((vuint64m8_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmul_vx_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmul_vx_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vmulh_vv_i8m1(op0, op1, op2) \ +__builtin_rvv_vmulh_vv_i8m1((vint8m1_t)(op0), (vint8m1_t)(op1), (size_t)(op2)) +#define vmulh_vv_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulh_vv_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vmulh_vv_i8m2(op0, op1, op2) \ +__builtin_rvv_vmulh_vv_i8m2((vint8m2_t)(op0), (vint8m2_t)(op1), (size_t)(op2)) +#define vmulh_vv_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulh_vv_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vmulh_vv_i8m4(op0, op1, op2) \ +__builtin_rvv_vmulh_vv_i8m4((vint8m4_t)(op0), (vint8m4_t)(op1), (size_t)(op2)) +#define vmulh_vv_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulh_vv_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vmulh_vv_i8m8(op0, op1, op2) \ +__builtin_rvv_vmulh_vv_i8m8((vint8m8_t)(op0), (vint8m8_t)(op1), (size_t)(op2)) +#define vmulh_vv_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulh_vv_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (vint8m8_t)(op3), (size_t)(op4)) +#define vmulh_vv_i8mf2(op0, op1, op2) \ +__builtin_rvv_vmulh_vv_i8mf2((vint8mf2_t)(op0), (vint8mf2_t)(op1), (size_t)(op2)) +#define vmulh_vv_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulh_vv_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vmulh_vv_i8mf4(op0, op1, op2) \ +__builtin_rvv_vmulh_vv_i8mf4((vint8mf4_t)(op0), (vint8mf4_t)(op1), (size_t)(op2)) +#define vmulh_vv_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulh_vv_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vmulh_vv_i8mf8(op0, op1, op2) \ +__builtin_rvv_vmulh_vv_i8mf8((vint8mf8_t)(op0), (vint8mf8_t)(op1), (size_t)(op2)) +#define vmulh_vv_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulh_vv_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vmulh_vv_i16m1(op0, op1, op2) \ +__builtin_rvv_vmulh_vv_i16m1((vint16m1_t)(op0), (vint16m1_t)(op1), (size_t)(op2)) +#define vmulh_vv_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulh_vv_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vmulh_vv_i16m2(op0, op1, op2) \ +__builtin_rvv_vmulh_vv_i16m2((vint16m2_t)(op0), (vint16m2_t)(op1), (size_t)(op2)) +#define vmulh_vv_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulh_vv_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vmulh_vv_i16m4(op0, op1, op2) \ +__builtin_rvv_vmulh_vv_i16m4((vint16m4_t)(op0), (vint16m4_t)(op1), (size_t)(op2)) +#define vmulh_vv_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulh_vv_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vmulh_vv_i16m8(op0, op1, op2) \ +__builtin_rvv_vmulh_vv_i16m8((vint16m8_t)(op0), (vint16m8_t)(op1), (size_t)(op2)) +#define vmulh_vv_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulh_vv_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (vint16m8_t)(op3), (size_t)(op4)) +#define vmulh_vv_i16mf2(op0, op1, op2) \ +__builtin_rvv_vmulh_vv_i16mf2((vint16mf2_t)(op0), (vint16mf2_t)(op1), (size_t)(op2)) +#define vmulh_vv_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulh_vv_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vmulh_vv_i16mf4(op0, op1, op2) \ +__builtin_rvv_vmulh_vv_i16mf4((vint16mf4_t)(op0), (vint16mf4_t)(op1), (size_t)(op2)) +#define vmulh_vv_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulh_vv_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vmulh_vv_i32m1(op0, op1, op2) \ +__builtin_rvv_vmulh_vv_i32m1((vint32m1_t)(op0), (vint32m1_t)(op1), (size_t)(op2)) +#define vmulh_vv_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulh_vv_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vmulh_vv_i32m2(op0, op1, op2) \ +__builtin_rvv_vmulh_vv_i32m2((vint32m2_t)(op0), (vint32m2_t)(op1), (size_t)(op2)) +#define vmulh_vv_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulh_vv_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vmulh_vv_i32m4(op0, op1, op2) \ +__builtin_rvv_vmulh_vv_i32m4((vint32m4_t)(op0), (vint32m4_t)(op1), (size_t)(op2)) +#define vmulh_vv_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulh_vv_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vmulh_vv_i32m8(op0, op1, op2) \ +__builtin_rvv_vmulh_vv_i32m8((vint32m8_t)(op0), (vint32m8_t)(op1), (size_t)(op2)) +#define vmulh_vv_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulh_vv_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vmulh_vv_i32mf2(op0, op1, op2) \ +__builtin_rvv_vmulh_vv_i32mf2((vint32mf2_t)(op0), (vint32mf2_t)(op1), (size_t)(op2)) +#define vmulh_vv_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulh_vv_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vmulh_vv_i64m1(op0, op1, op2) \ +__builtin_rvv_vmulh_vv_i64m1((vint64m1_t)(op0), (vint64m1_t)(op1), (size_t)(op2)) +#define vmulh_vv_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulh_vv_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vmulh_vv_i64m2(op0, op1, op2) \ +__builtin_rvv_vmulh_vv_i64m2((vint64m2_t)(op0), (vint64m2_t)(op1), (size_t)(op2)) +#define vmulh_vv_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulh_vv_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vmulh_vv_i64m4(op0, op1, op2) \ +__builtin_rvv_vmulh_vv_i64m4((vint64m4_t)(op0), (vint64m4_t)(op1), (size_t)(op2)) +#define vmulh_vv_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulh_vv_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vmulh_vv_i64m8(op0, op1, op2) \ +__builtin_rvv_vmulh_vv_i64m8((vint64m8_t)(op0), (vint64m8_t)(op1), (size_t)(op2)) +#define vmulh_vv_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulh_vv_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vmulh_vx_i8m1(op0, op1, op2) \ +__builtin_rvv_vmulh_vx_i8m1((vint8m1_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmulh_vx_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulh_vx_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmulh_vx_i8m2(op0, op1, op2) \ +__builtin_rvv_vmulh_vx_i8m2((vint8m2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmulh_vx_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulh_vx_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmulh_vx_i8m4(op0, op1, op2) \ +__builtin_rvv_vmulh_vx_i8m4((vint8m4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmulh_vx_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulh_vx_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmulh_vx_i8m8(op0, op1, op2) \ +__builtin_rvv_vmulh_vx_i8m8((vint8m8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmulh_vx_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulh_vx_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmulh_vx_i8mf2(op0, op1, op2) \ +__builtin_rvv_vmulh_vx_i8mf2((vint8mf2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmulh_vx_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulh_vx_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmulh_vx_i8mf4(op0, op1, op2) \ +__builtin_rvv_vmulh_vx_i8mf4((vint8mf4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmulh_vx_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulh_vx_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmulh_vx_i8mf8(op0, op1, op2) \ +__builtin_rvv_vmulh_vx_i8mf8((vint8mf8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmulh_vx_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulh_vx_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vmulh_vx_i16m1(op0, op1, op2) \ +__builtin_rvv_vmulh_vx_i16m1((vint16m1_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmulh_vx_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulh_vx_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmulh_vx_i16m2(op0, op1, op2) \ +__builtin_rvv_vmulh_vx_i16m2((vint16m2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmulh_vx_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulh_vx_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmulh_vx_i16m4(op0, op1, op2) \ +__builtin_rvv_vmulh_vx_i16m4((vint16m4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmulh_vx_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulh_vx_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmulh_vx_i16m8(op0, op1, op2) \ +__builtin_rvv_vmulh_vx_i16m8((vint16m8_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmulh_vx_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulh_vx_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmulh_vx_i16mf2(op0, op1, op2) \ +__builtin_rvv_vmulh_vx_i16mf2((vint16mf2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmulh_vx_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulh_vx_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmulh_vx_i16mf4(op0, op1, op2) \ +__builtin_rvv_vmulh_vx_i16mf4((vint16mf4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmulh_vx_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulh_vx_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vmulh_vx_i32m1(op0, op1, op2) \ +__builtin_rvv_vmulh_vx_i32m1((vint32m1_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmulh_vx_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulh_vx_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmulh_vx_i32m2(op0, op1, op2) \ +__builtin_rvv_vmulh_vx_i32m2((vint32m2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmulh_vx_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulh_vx_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmulh_vx_i32m4(op0, op1, op2) \ +__builtin_rvv_vmulh_vx_i32m4((vint32m4_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmulh_vx_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulh_vx_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmulh_vx_i32m8(op0, op1, op2) \ +__builtin_rvv_vmulh_vx_i32m8((vint32m8_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmulh_vx_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulh_vx_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmulh_vx_i32mf2(op0, op1, op2) \ +__builtin_rvv_vmulh_vx_i32mf2((vint32mf2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmulh_vx_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulh_vx_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vmulh_vx_i64m1(op0, op1, op2) \ +__builtin_rvv_vmulh_vx_i64m1((vint64m1_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmulh_vx_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulh_vx_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vmulh_vx_i64m2(op0, op1, op2) \ +__builtin_rvv_vmulh_vx_i64m2((vint64m2_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmulh_vx_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulh_vx_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vmulh_vx_i64m4(op0, op1, op2) \ +__builtin_rvv_vmulh_vx_i64m4((vint64m4_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmulh_vx_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulh_vx_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vmulh_vx_i64m8(op0, op1, op2) \ +__builtin_rvv_vmulh_vx_i64m8((vint64m8_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmulh_vx_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulh_vx_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vmulhu_vv_u8m1(op0, op1, op2) \ +__builtin_rvv_vmulhu_vv_u8m1((vuint8m1_t)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vmulhu_vv_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhu_vv_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vmulhu_vv_u8m2(op0, op1, op2) \ +__builtin_rvv_vmulhu_vv_u8m2((vuint8m2_t)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vmulhu_vv_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhu_vv_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vmulhu_vv_u8m4(op0, op1, op2) \ +__builtin_rvv_vmulhu_vv_u8m4((vuint8m4_t)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vmulhu_vv_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhu_vv_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vmulhu_vv_u8m8(op0, op1, op2) \ +__builtin_rvv_vmulhu_vv_u8m8((vuint8m8_t)(op0), (vuint8m8_t)(op1), (size_t)(op2)) +#define vmulhu_vv_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhu_vv_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (vuint8m8_t)(op3), (size_t)(op4)) +#define vmulhu_vv_u8mf2(op0, op1, op2) \ +__builtin_rvv_vmulhu_vv_u8mf2((vuint8mf2_t)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vmulhu_vv_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhu_vv_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vmulhu_vv_u8mf4(op0, op1, op2) \ +__builtin_rvv_vmulhu_vv_u8mf4((vuint8mf4_t)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vmulhu_vv_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhu_vv_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vmulhu_vv_u8mf8(op0, op1, op2) \ +__builtin_rvv_vmulhu_vv_u8mf8((vuint8mf8_t)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vmulhu_vv_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhu_vv_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vmulhu_vv_u16m1(op0, op1, op2) \ +__builtin_rvv_vmulhu_vv_u16m1((vuint16m1_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vmulhu_vv_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhu_vv_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vmulhu_vv_u16m2(op0, op1, op2) \ +__builtin_rvv_vmulhu_vv_u16m2((vuint16m2_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vmulhu_vv_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhu_vv_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vmulhu_vv_u16m4(op0, op1, op2) \ +__builtin_rvv_vmulhu_vv_u16m4((vuint16m4_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vmulhu_vv_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhu_vv_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vmulhu_vv_u16m8(op0, op1, op2) \ +__builtin_rvv_vmulhu_vv_u16m8((vuint16m8_t)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vmulhu_vv_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhu_vv_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vmulhu_vv_u16mf2(op0, op1, op2) \ +__builtin_rvv_vmulhu_vv_u16mf2((vuint16mf2_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vmulhu_vv_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhu_vv_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vmulhu_vv_u16mf4(op0, op1, op2) \ +__builtin_rvv_vmulhu_vv_u16mf4((vuint16mf4_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vmulhu_vv_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhu_vv_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vmulhu_vv_u32m1(op0, op1, op2) \ +__builtin_rvv_vmulhu_vv_u32m1((vuint32m1_t)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vmulhu_vv_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhu_vv_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vmulhu_vv_u32m2(op0, op1, op2) \ +__builtin_rvv_vmulhu_vv_u32m2((vuint32m2_t)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vmulhu_vv_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhu_vv_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vmulhu_vv_u32m4(op0, op1, op2) \ +__builtin_rvv_vmulhu_vv_u32m4((vuint32m4_t)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vmulhu_vv_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhu_vv_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vmulhu_vv_u32m8(op0, op1, op2) \ +__builtin_rvv_vmulhu_vv_u32m8((vuint32m8_t)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vmulhu_vv_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhu_vv_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vmulhu_vv_u32mf2(op0, op1, op2) \ +__builtin_rvv_vmulhu_vv_u32mf2((vuint32mf2_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vmulhu_vv_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhu_vv_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vmulhu_vv_u64m1(op0, op1, op2) \ +__builtin_rvv_vmulhu_vv_u64m1((vuint64m1_t)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vmulhu_vv_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhu_vv_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vmulhu_vv_u64m2(op0, op1, op2) \ +__builtin_rvv_vmulhu_vv_u64m2((vuint64m2_t)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vmulhu_vv_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhu_vv_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vmulhu_vv_u64m4(op0, op1, op2) \ +__builtin_rvv_vmulhu_vv_u64m4((vuint64m4_t)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vmulhu_vv_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhu_vv_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vmulhu_vv_u64m8(op0, op1, op2) \ +__builtin_rvv_vmulhu_vv_u64m8((vuint64m8_t)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vmulhu_vv_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhu_vv_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vmulhu_vx_u8m1(op0, op1, op2) \ +__builtin_rvv_vmulhu_vx_u8m1((vuint8m1_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmulhu_vx_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhu_vx_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmulhu_vx_u8m2(op0, op1, op2) \ +__builtin_rvv_vmulhu_vx_u8m2((vuint8m2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmulhu_vx_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhu_vx_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmulhu_vx_u8m4(op0, op1, op2) \ +__builtin_rvv_vmulhu_vx_u8m4((vuint8m4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmulhu_vx_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhu_vx_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmulhu_vx_u8m8(op0, op1, op2) \ +__builtin_rvv_vmulhu_vx_u8m8((vuint8m8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmulhu_vx_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhu_vx_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmulhu_vx_u8mf2(op0, op1, op2) \ +__builtin_rvv_vmulhu_vx_u8mf2((vuint8mf2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmulhu_vx_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhu_vx_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmulhu_vx_u8mf4(op0, op1, op2) \ +__builtin_rvv_vmulhu_vx_u8mf4((vuint8mf4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmulhu_vx_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhu_vx_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmulhu_vx_u8mf8(op0, op1, op2) \ +__builtin_rvv_vmulhu_vx_u8mf8((vuint8mf8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmulhu_vx_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhu_vx_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmulhu_vx_u16m1(op0, op1, op2) \ +__builtin_rvv_vmulhu_vx_u16m1((vuint16m1_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmulhu_vx_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhu_vx_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmulhu_vx_u16m2(op0, op1, op2) \ +__builtin_rvv_vmulhu_vx_u16m2((vuint16m2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmulhu_vx_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhu_vx_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmulhu_vx_u16m4(op0, op1, op2) \ +__builtin_rvv_vmulhu_vx_u16m4((vuint16m4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmulhu_vx_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhu_vx_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmulhu_vx_u16m8(op0, op1, op2) \ +__builtin_rvv_vmulhu_vx_u16m8((vuint16m8_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmulhu_vx_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhu_vx_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmulhu_vx_u16mf2(op0, op1, op2) \ +__builtin_rvv_vmulhu_vx_u16mf2((vuint16mf2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmulhu_vx_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhu_vx_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmulhu_vx_u16mf4(op0, op1, op2) \ +__builtin_rvv_vmulhu_vx_u16mf4((vuint16mf4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmulhu_vx_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhu_vx_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmulhu_vx_u32m1(op0, op1, op2) \ +__builtin_rvv_vmulhu_vx_u32m1((vuint32m1_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmulhu_vx_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhu_vx_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmulhu_vx_u32m2(op0, op1, op2) \ +__builtin_rvv_vmulhu_vx_u32m2((vuint32m2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmulhu_vx_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhu_vx_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmulhu_vx_u32m4(op0, op1, op2) \ +__builtin_rvv_vmulhu_vx_u32m4((vuint32m4_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmulhu_vx_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhu_vx_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmulhu_vx_u32m8(op0, op1, op2) \ +__builtin_rvv_vmulhu_vx_u32m8((vuint32m8_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmulhu_vx_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhu_vx_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmulhu_vx_u32mf2(op0, op1, op2) \ +__builtin_rvv_vmulhu_vx_u32mf2((vuint32mf2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmulhu_vx_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhu_vx_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmulhu_vx_u64m1(op0, op1, op2) \ +__builtin_rvv_vmulhu_vx_u64m1((vuint64m1_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmulhu_vx_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhu_vx_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vmulhu_vx_u64m2(op0, op1, op2) \ +__builtin_rvv_vmulhu_vx_u64m2((vuint64m2_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmulhu_vx_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhu_vx_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vmulhu_vx_u64m4(op0, op1, op2) \ +__builtin_rvv_vmulhu_vx_u64m4((vuint64m4_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmulhu_vx_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhu_vx_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vmulhu_vx_u64m8(op0, op1, op2) \ +__builtin_rvv_vmulhu_vx_u64m8((vuint64m8_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmulhu_vx_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhu_vx_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vmulhsu_vv_i8m1(op0, op1, op2) \ +__builtin_rvv_vmulhsu_vv_i8m1((vint8m1_t)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vmulhsu_vv_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhsu_vv_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vmulhsu_vv_i8m2(op0, op1, op2) \ +__builtin_rvv_vmulhsu_vv_i8m2((vint8m2_t)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vmulhsu_vv_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhsu_vv_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vmulhsu_vv_i8m4(op0, op1, op2) \ +__builtin_rvv_vmulhsu_vv_i8m4((vint8m4_t)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vmulhsu_vv_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhsu_vv_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vmulhsu_vv_i8m8(op0, op1, op2) \ +__builtin_rvv_vmulhsu_vv_i8m8((vint8m8_t)(op0), (vuint8m8_t)(op1), (size_t)(op2)) +#define vmulhsu_vv_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhsu_vv_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (vuint8m8_t)(op3), (size_t)(op4)) +#define vmulhsu_vv_i8mf2(op0, op1, op2) \ +__builtin_rvv_vmulhsu_vv_i8mf2((vint8mf2_t)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vmulhsu_vv_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhsu_vv_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vmulhsu_vv_i8mf4(op0, op1, op2) \ +__builtin_rvv_vmulhsu_vv_i8mf4((vint8mf4_t)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vmulhsu_vv_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhsu_vv_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vmulhsu_vv_i8mf8(op0, op1, op2) \ +__builtin_rvv_vmulhsu_vv_i8mf8((vint8mf8_t)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vmulhsu_vv_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhsu_vv_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vmulhsu_vv_i16m1(op0, op1, op2) \ +__builtin_rvv_vmulhsu_vv_i16m1((vint16m1_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vmulhsu_vv_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhsu_vv_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vmulhsu_vv_i16m2(op0, op1, op2) \ +__builtin_rvv_vmulhsu_vv_i16m2((vint16m2_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vmulhsu_vv_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhsu_vv_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vmulhsu_vv_i16m4(op0, op1, op2) \ +__builtin_rvv_vmulhsu_vv_i16m4((vint16m4_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vmulhsu_vv_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhsu_vv_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vmulhsu_vv_i16m8(op0, op1, op2) \ +__builtin_rvv_vmulhsu_vv_i16m8((vint16m8_t)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vmulhsu_vv_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhsu_vv_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vmulhsu_vv_i16mf2(op0, op1, op2) \ +__builtin_rvv_vmulhsu_vv_i16mf2((vint16mf2_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vmulhsu_vv_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhsu_vv_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vmulhsu_vv_i16mf4(op0, op1, op2) \ +__builtin_rvv_vmulhsu_vv_i16mf4((vint16mf4_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vmulhsu_vv_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhsu_vv_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vmulhsu_vv_i32m1(op0, op1, op2) \ +__builtin_rvv_vmulhsu_vv_i32m1((vint32m1_t)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vmulhsu_vv_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhsu_vv_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vmulhsu_vv_i32m2(op0, op1, op2) \ +__builtin_rvv_vmulhsu_vv_i32m2((vint32m2_t)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vmulhsu_vv_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhsu_vv_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vmulhsu_vv_i32m4(op0, op1, op2) \ +__builtin_rvv_vmulhsu_vv_i32m4((vint32m4_t)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vmulhsu_vv_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhsu_vv_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vmulhsu_vv_i32m8(op0, op1, op2) \ +__builtin_rvv_vmulhsu_vv_i32m8((vint32m8_t)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vmulhsu_vv_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhsu_vv_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vmulhsu_vv_i32mf2(op0, op1, op2) \ +__builtin_rvv_vmulhsu_vv_i32mf2((vint32mf2_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vmulhsu_vv_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhsu_vv_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vmulhsu_vv_i64m1(op0, op1, op2) \ +__builtin_rvv_vmulhsu_vv_i64m1((vint64m1_t)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vmulhsu_vv_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhsu_vv_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vmulhsu_vv_i64m2(op0, op1, op2) \ +__builtin_rvv_vmulhsu_vv_i64m2((vint64m2_t)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vmulhsu_vv_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhsu_vv_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vmulhsu_vv_i64m4(op0, op1, op2) \ +__builtin_rvv_vmulhsu_vv_i64m4((vint64m4_t)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vmulhsu_vv_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhsu_vv_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vmulhsu_vv_i64m8(op0, op1, op2) \ +__builtin_rvv_vmulhsu_vv_i64m8((vint64m8_t)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vmulhsu_vv_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhsu_vv_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vsuxei8_v_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_i32m1((int32_t *)(op0), (vuint8mf4_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vsuxei8_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint8mf4_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vsuxei8_v_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_i32m2((int32_t *)(op0), (vuint8mf2_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vsuxei8_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint8mf2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vsuxei8_v_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_i32m4((int32_t *)(op0), (vuint8m1_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vsuxei8_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vuint8m1_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vsuxei8_v_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_i32m8((int32_t *)(op0), (vuint8m2_t)(op1), (vint32m8_t)(op2), (size_t)(op3)) +#define vsuxei8_v_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_i32m8_m((vbool4_t)(op0), (int32_t *)(op1), (vuint8m2_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vsuxei8_v_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_i32mf2((int32_t *)(op0), (vuint8mf8_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vsuxei8_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint8mf8_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vmulhsu_vx_i8m1(op0, op1, op2) \ +__builtin_rvv_vmulhsu_vx_i8m1((vint8m1_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmulhsu_vx_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhsu_vx_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmulhsu_vx_i8m2(op0, op1, op2) \ +__builtin_rvv_vmulhsu_vx_i8m2((vint8m2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmulhsu_vx_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhsu_vx_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmulhsu_vx_i8m4(op0, op1, op2) \ +__builtin_rvv_vmulhsu_vx_i8m4((vint8m4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmulhsu_vx_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhsu_vx_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmulhsu_vx_i8m8(op0, op1, op2) \ +__builtin_rvv_vmulhsu_vx_i8m8((vint8m8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmulhsu_vx_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhsu_vx_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmulhsu_vx_i8mf2(op0, op1, op2) \ +__builtin_rvv_vmulhsu_vx_i8mf2((vint8mf2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmulhsu_vx_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhsu_vx_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmulhsu_vx_i8mf4(op0, op1, op2) \ +__builtin_rvv_vmulhsu_vx_i8mf4((vint8mf4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmulhsu_vx_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhsu_vx_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmulhsu_vx_i8mf8(op0, op1, op2) \ +__builtin_rvv_vmulhsu_vx_i8mf8((vint8mf8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmulhsu_vx_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhsu_vx_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vmulhsu_vx_i16m1(op0, op1, op2) \ +__builtin_rvv_vmulhsu_vx_i16m1((vint16m1_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmulhsu_vx_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhsu_vx_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmulhsu_vx_i16m2(op0, op1, op2) \ +__builtin_rvv_vmulhsu_vx_i16m2((vint16m2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmulhsu_vx_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhsu_vx_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmulhsu_vx_i16m4(op0, op1, op2) \ +__builtin_rvv_vmulhsu_vx_i16m4((vint16m4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmulhsu_vx_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhsu_vx_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmulhsu_vx_i16m8(op0, op1, op2) \ +__builtin_rvv_vmulhsu_vx_i16m8((vint16m8_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmulhsu_vx_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhsu_vx_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmulhsu_vx_i16mf2(op0, op1, op2) \ +__builtin_rvv_vmulhsu_vx_i16mf2((vint16mf2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmulhsu_vx_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhsu_vx_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmulhsu_vx_i16mf4(op0, op1, op2) \ +__builtin_rvv_vmulhsu_vx_i16mf4((vint16mf4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmulhsu_vx_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhsu_vx_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vmulhsu_vx_i32m1(op0, op1, op2) \ +__builtin_rvv_vmulhsu_vx_i32m1((vint32m1_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmulhsu_vx_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhsu_vx_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmulhsu_vx_i32m2(op0, op1, op2) \ +__builtin_rvv_vmulhsu_vx_i32m2((vint32m2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmulhsu_vx_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhsu_vx_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmulhsu_vx_i32m4(op0, op1, op2) \ +__builtin_rvv_vmulhsu_vx_i32m4((vint32m4_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmulhsu_vx_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhsu_vx_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmulhsu_vx_i32m8(op0, op1, op2) \ +__builtin_rvv_vmulhsu_vx_i32m8((vint32m8_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmulhsu_vx_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhsu_vx_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmulhsu_vx_i32mf2(op0, op1, op2) \ +__builtin_rvv_vmulhsu_vx_i32mf2((vint32mf2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmulhsu_vx_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhsu_vx_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmulhsu_vx_i64m1(op0, op1, op2) \ +__builtin_rvv_vmulhsu_vx_i64m1((vint64m1_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmulhsu_vx_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhsu_vx_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vmulhsu_vx_i64m2(op0, op1, op2) \ +__builtin_rvv_vmulhsu_vx_i64m2((vint64m2_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmulhsu_vx_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhsu_vx_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vmulhsu_vx_i64m4(op0, op1, op2) \ +__builtin_rvv_vmulhsu_vx_i64m4((vint64m4_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmulhsu_vx_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhsu_vx_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vmulhsu_vx_i64m8(op0, op1, op2) \ +__builtin_rvv_vmulhsu_vx_i64m8((vint64m8_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmulhsu_vx_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmulhsu_vx_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vdivu_vv_u8m1(op0, op1, op2) \ +__builtin_rvv_vdivu_vv_u8m1((vuint8m1_t)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vdivu_vv_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdivu_vv_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vdivu_vv_u8m2(op0, op1, op2) \ +__builtin_rvv_vdivu_vv_u8m2((vuint8m2_t)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vdivu_vv_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdivu_vv_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vdivu_vv_u8m4(op0, op1, op2) \ +__builtin_rvv_vdivu_vv_u8m4((vuint8m4_t)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vdivu_vv_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdivu_vv_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vdivu_vv_u8m8(op0, op1, op2) \ +__builtin_rvv_vdivu_vv_u8m8((vuint8m8_t)(op0), (vuint8m8_t)(op1), (size_t)(op2)) +#define vdivu_vv_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdivu_vv_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (vuint8m8_t)(op3), (size_t)(op4)) +#define vdivu_vv_u8mf2(op0, op1, op2) \ +__builtin_rvv_vdivu_vv_u8mf2((vuint8mf2_t)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vdivu_vv_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdivu_vv_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vdivu_vv_u8mf4(op0, op1, op2) \ +__builtin_rvv_vdivu_vv_u8mf4((vuint8mf4_t)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vdivu_vv_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdivu_vv_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vdivu_vv_u8mf8(op0, op1, op2) \ +__builtin_rvv_vdivu_vv_u8mf8((vuint8mf8_t)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vdivu_vv_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdivu_vv_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vdivu_vv_u16m1(op0, op1, op2) \ +__builtin_rvv_vdivu_vv_u16m1((vuint16m1_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vdivu_vv_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdivu_vv_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vdivu_vv_u16m2(op0, op1, op2) \ +__builtin_rvv_vdivu_vv_u16m2((vuint16m2_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vdivu_vv_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdivu_vv_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vdivu_vv_u16m4(op0, op1, op2) \ +__builtin_rvv_vdivu_vv_u16m4((vuint16m4_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vdivu_vv_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdivu_vv_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vdivu_vv_u16m8(op0, op1, op2) \ +__builtin_rvv_vdivu_vv_u16m8((vuint16m8_t)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vdivu_vv_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdivu_vv_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vdivu_vv_u16mf2(op0, op1, op2) \ +__builtin_rvv_vdivu_vv_u16mf2((vuint16mf2_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vdivu_vv_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdivu_vv_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vdivu_vv_u16mf4(op0, op1, op2) \ +__builtin_rvv_vdivu_vv_u16mf4((vuint16mf4_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vdivu_vv_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdivu_vv_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vdivu_vv_u32m1(op0, op1, op2) \ +__builtin_rvv_vdivu_vv_u32m1((vuint32m1_t)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vdivu_vv_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdivu_vv_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vdivu_vv_u32m2(op0, op1, op2) \ +__builtin_rvv_vdivu_vv_u32m2((vuint32m2_t)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vdivu_vv_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdivu_vv_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vdivu_vv_u32m4(op0, op1, op2) \ +__builtin_rvv_vdivu_vv_u32m4((vuint32m4_t)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vdivu_vv_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdivu_vv_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vdivu_vv_u32m8(op0, op1, op2) \ +__builtin_rvv_vdivu_vv_u32m8((vuint32m8_t)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vdivu_vv_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdivu_vv_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vdivu_vv_u32mf2(op0, op1, op2) \ +__builtin_rvv_vdivu_vv_u32mf2((vuint32mf2_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vdivu_vv_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdivu_vv_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vdivu_vv_u64m1(op0, op1, op2) \ +__builtin_rvv_vdivu_vv_u64m1((vuint64m1_t)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vdivu_vv_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdivu_vv_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vdivu_vv_u64m2(op0, op1, op2) \ +__builtin_rvv_vdivu_vv_u64m2((vuint64m2_t)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vdivu_vv_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdivu_vv_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vdivu_vv_u64m4(op0, op1, op2) \ +__builtin_rvv_vdivu_vv_u64m4((vuint64m4_t)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vdivu_vv_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdivu_vv_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vdivu_vv_u64m8(op0, op1, op2) \ +__builtin_rvv_vdivu_vv_u64m8((vuint64m8_t)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vdivu_vv_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdivu_vv_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vdivu_vx_u8m1(op0, op1, op2) \ +__builtin_rvv_vdivu_vx_u8m1((vuint8m1_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vdivu_vx_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdivu_vx_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vdivu_vx_u8m2(op0, op1, op2) \ +__builtin_rvv_vdivu_vx_u8m2((vuint8m2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vdivu_vx_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdivu_vx_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vdivu_vx_u8m4(op0, op1, op2) \ +__builtin_rvv_vdivu_vx_u8m4((vuint8m4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vdivu_vx_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdivu_vx_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vdivu_vx_u8m8(op0, op1, op2) \ +__builtin_rvv_vdivu_vx_u8m8((vuint8m8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vdivu_vx_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdivu_vx_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vdivu_vx_u8mf2(op0, op1, op2) \ +__builtin_rvv_vdivu_vx_u8mf2((vuint8mf2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vdivu_vx_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdivu_vx_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vdivu_vx_u8mf4(op0, op1, op2) \ +__builtin_rvv_vdivu_vx_u8mf4((vuint8mf4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vdivu_vx_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdivu_vx_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vdivu_vx_u8mf8(op0, op1, op2) \ +__builtin_rvv_vdivu_vx_u8mf8((vuint8mf8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vdivu_vx_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdivu_vx_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vdivu_vx_u16m1(op0, op1, op2) \ +__builtin_rvv_vdivu_vx_u16m1((vuint16m1_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vdivu_vx_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdivu_vx_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vdivu_vx_u16m2(op0, op1, op2) \ +__builtin_rvv_vdivu_vx_u16m2((vuint16m2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vdivu_vx_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdivu_vx_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vdivu_vx_u16m4(op0, op1, op2) \ +__builtin_rvv_vdivu_vx_u16m4((vuint16m4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vdivu_vx_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdivu_vx_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vdivu_vx_u16m8(op0, op1, op2) \ +__builtin_rvv_vdivu_vx_u16m8((vuint16m8_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vdivu_vx_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdivu_vx_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vdivu_vx_u16mf2(op0, op1, op2) \ +__builtin_rvv_vdivu_vx_u16mf2((vuint16mf2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vdivu_vx_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdivu_vx_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vdivu_vx_u16mf4(op0, op1, op2) \ +__builtin_rvv_vdivu_vx_u16mf4((vuint16mf4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vdivu_vx_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdivu_vx_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vdivu_vx_u32m1(op0, op1, op2) \ +__builtin_rvv_vdivu_vx_u32m1((vuint32m1_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vdivu_vx_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdivu_vx_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vdivu_vx_u32m2(op0, op1, op2) \ +__builtin_rvv_vdivu_vx_u32m2((vuint32m2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vdivu_vx_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdivu_vx_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vdivu_vx_u32m4(op0, op1, op2) \ +__builtin_rvv_vdivu_vx_u32m4((vuint32m4_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vdivu_vx_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdivu_vx_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vdivu_vx_u32m8(op0, op1, op2) \ +__builtin_rvv_vdivu_vx_u32m8((vuint32m8_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vdivu_vx_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdivu_vx_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vdivu_vx_u32mf2(op0, op1, op2) \ +__builtin_rvv_vdivu_vx_u32mf2((vuint32mf2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vdivu_vx_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdivu_vx_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vdivu_vx_u64m1(op0, op1, op2) \ +__builtin_rvv_vdivu_vx_u64m1((vuint64m1_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vdivu_vx_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdivu_vx_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vdivu_vx_u64m2(op0, op1, op2) \ +__builtin_rvv_vdivu_vx_u64m2((vuint64m2_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vdivu_vx_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdivu_vx_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vdivu_vx_u64m4(op0, op1, op2) \ +__builtin_rvv_vdivu_vx_u64m4((vuint64m4_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vdivu_vx_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdivu_vx_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vdivu_vx_u64m8(op0, op1, op2) \ +__builtin_rvv_vdivu_vx_u64m8((vuint64m8_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vdivu_vx_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdivu_vx_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vdiv_vv_i8m1(op0, op1, op2) \ +__builtin_rvv_vdiv_vv_i8m1((vint8m1_t)(op0), (vint8m1_t)(op1), (size_t)(op2)) +#define vdiv_vv_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdiv_vv_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vdiv_vv_i8m2(op0, op1, op2) \ +__builtin_rvv_vdiv_vv_i8m2((vint8m2_t)(op0), (vint8m2_t)(op1), (size_t)(op2)) +#define vdiv_vv_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdiv_vv_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vdiv_vv_i8m4(op0, op1, op2) \ +__builtin_rvv_vdiv_vv_i8m4((vint8m4_t)(op0), (vint8m4_t)(op1), (size_t)(op2)) +#define vdiv_vv_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdiv_vv_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vdiv_vv_i8m8(op0, op1, op2) \ +__builtin_rvv_vdiv_vv_i8m8((vint8m8_t)(op0), (vint8m8_t)(op1), (size_t)(op2)) +#define vdiv_vv_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdiv_vv_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (vint8m8_t)(op3), (size_t)(op4)) +#define vdiv_vv_i8mf2(op0, op1, op2) \ +__builtin_rvv_vdiv_vv_i8mf2((vint8mf2_t)(op0), (vint8mf2_t)(op1), (size_t)(op2)) +#define vdiv_vv_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdiv_vv_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vdiv_vv_i8mf4(op0, op1, op2) \ +__builtin_rvv_vdiv_vv_i8mf4((vint8mf4_t)(op0), (vint8mf4_t)(op1), (size_t)(op2)) +#define vdiv_vv_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdiv_vv_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vdiv_vv_i8mf8(op0, op1, op2) \ +__builtin_rvv_vdiv_vv_i8mf8((vint8mf8_t)(op0), (vint8mf8_t)(op1), (size_t)(op2)) +#define vdiv_vv_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdiv_vv_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vdiv_vv_i16m1(op0, op1, op2) \ +__builtin_rvv_vdiv_vv_i16m1((vint16m1_t)(op0), (vint16m1_t)(op1), (size_t)(op2)) +#define vdiv_vv_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdiv_vv_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vdiv_vv_i16m2(op0, op1, op2) \ +__builtin_rvv_vdiv_vv_i16m2((vint16m2_t)(op0), (vint16m2_t)(op1), (size_t)(op2)) +#define vdiv_vv_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdiv_vv_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vdiv_vv_i16m4(op0, op1, op2) \ +__builtin_rvv_vdiv_vv_i16m4((vint16m4_t)(op0), (vint16m4_t)(op1), (size_t)(op2)) +#define vdiv_vv_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdiv_vv_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vdiv_vv_i16m8(op0, op1, op2) \ +__builtin_rvv_vdiv_vv_i16m8((vint16m8_t)(op0), (vint16m8_t)(op1), (size_t)(op2)) +#define vdiv_vv_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdiv_vv_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (vint16m8_t)(op3), (size_t)(op4)) +#define vdiv_vv_i16mf2(op0, op1, op2) \ +__builtin_rvv_vdiv_vv_i16mf2((vint16mf2_t)(op0), (vint16mf2_t)(op1), (size_t)(op2)) +#define vdiv_vv_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdiv_vv_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vdiv_vv_i16mf4(op0, op1, op2) \ +__builtin_rvv_vdiv_vv_i16mf4((vint16mf4_t)(op0), (vint16mf4_t)(op1), (size_t)(op2)) +#define vdiv_vv_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdiv_vv_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vdiv_vv_i32m1(op0, op1, op2) \ +__builtin_rvv_vdiv_vv_i32m1((vint32m1_t)(op0), (vint32m1_t)(op1), (size_t)(op2)) +#define vdiv_vv_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdiv_vv_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vdiv_vv_i32m2(op0, op1, op2) \ +__builtin_rvv_vdiv_vv_i32m2((vint32m2_t)(op0), (vint32m2_t)(op1), (size_t)(op2)) +#define vdiv_vv_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdiv_vv_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vdiv_vv_i32m4(op0, op1, op2) \ +__builtin_rvv_vdiv_vv_i32m4((vint32m4_t)(op0), (vint32m4_t)(op1), (size_t)(op2)) +#define vdiv_vv_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdiv_vv_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vdiv_vv_i32m8(op0, op1, op2) \ +__builtin_rvv_vdiv_vv_i32m8((vint32m8_t)(op0), (vint32m8_t)(op1), (size_t)(op2)) +#define vdiv_vv_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdiv_vv_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vdiv_vv_i32mf2(op0, op1, op2) \ +__builtin_rvv_vdiv_vv_i32mf2((vint32mf2_t)(op0), (vint32mf2_t)(op1), (size_t)(op2)) +#define vdiv_vv_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdiv_vv_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vdiv_vv_i64m1(op0, op1, op2) \ +__builtin_rvv_vdiv_vv_i64m1((vint64m1_t)(op0), (vint64m1_t)(op1), (size_t)(op2)) +#define vdiv_vv_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdiv_vv_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vdiv_vv_i64m2(op0, op1, op2) \ +__builtin_rvv_vdiv_vv_i64m2((vint64m2_t)(op0), (vint64m2_t)(op1), (size_t)(op2)) +#define vdiv_vv_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdiv_vv_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vdiv_vv_i64m4(op0, op1, op2) \ +__builtin_rvv_vdiv_vv_i64m4((vint64m4_t)(op0), (vint64m4_t)(op1), (size_t)(op2)) +#define vdiv_vv_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdiv_vv_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vdiv_vv_i64m8(op0, op1, op2) \ +__builtin_rvv_vdiv_vv_i64m8((vint64m8_t)(op0), (vint64m8_t)(op1), (size_t)(op2)) +#define vdiv_vv_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdiv_vv_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vdiv_vx_i8m1(op0, op1, op2) \ +__builtin_rvv_vdiv_vx_i8m1((vint8m1_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vdiv_vx_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdiv_vx_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vdiv_vx_i8m2(op0, op1, op2) \ +__builtin_rvv_vdiv_vx_i8m2((vint8m2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vdiv_vx_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdiv_vx_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vdiv_vx_i8m4(op0, op1, op2) \ +__builtin_rvv_vdiv_vx_i8m4((vint8m4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vdiv_vx_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdiv_vx_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vdiv_vx_i8m8(op0, op1, op2) \ +__builtin_rvv_vdiv_vx_i8m8((vint8m8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vdiv_vx_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdiv_vx_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vdiv_vx_i8mf2(op0, op1, op2) \ +__builtin_rvv_vdiv_vx_i8mf2((vint8mf2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vdiv_vx_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdiv_vx_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vdiv_vx_i8mf4(op0, op1, op2) \ +__builtin_rvv_vdiv_vx_i8mf4((vint8mf4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vdiv_vx_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdiv_vx_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vdiv_vx_i8mf8(op0, op1, op2) \ +__builtin_rvv_vdiv_vx_i8mf8((vint8mf8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vdiv_vx_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdiv_vx_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vdiv_vx_i16m1(op0, op1, op2) \ +__builtin_rvv_vdiv_vx_i16m1((vint16m1_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vdiv_vx_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdiv_vx_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vdiv_vx_i16m2(op0, op1, op2) \ +__builtin_rvv_vdiv_vx_i16m2((vint16m2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vdiv_vx_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdiv_vx_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vdiv_vx_i16m4(op0, op1, op2) \ +__builtin_rvv_vdiv_vx_i16m4((vint16m4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vdiv_vx_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdiv_vx_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vdiv_vx_i16m8(op0, op1, op2) \ +__builtin_rvv_vdiv_vx_i16m8((vint16m8_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vdiv_vx_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdiv_vx_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vdiv_vx_i16mf2(op0, op1, op2) \ +__builtin_rvv_vdiv_vx_i16mf2((vint16mf2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vdiv_vx_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdiv_vx_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vdiv_vx_i16mf4(op0, op1, op2) \ +__builtin_rvv_vdiv_vx_i16mf4((vint16mf4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vdiv_vx_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdiv_vx_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vdiv_vx_i32m1(op0, op1, op2) \ +__builtin_rvv_vdiv_vx_i32m1((vint32m1_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vdiv_vx_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdiv_vx_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vdiv_vx_i32m2(op0, op1, op2) \ +__builtin_rvv_vdiv_vx_i32m2((vint32m2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vdiv_vx_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdiv_vx_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vdiv_vx_i32m4(op0, op1, op2) \ +__builtin_rvv_vdiv_vx_i32m4((vint32m4_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vdiv_vx_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdiv_vx_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vdiv_vx_i32m8(op0, op1, op2) \ +__builtin_rvv_vdiv_vx_i32m8((vint32m8_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vdiv_vx_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdiv_vx_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vdiv_vx_i32mf2(op0, op1, op2) \ +__builtin_rvv_vdiv_vx_i32mf2((vint32mf2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vdiv_vx_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdiv_vx_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vdiv_vx_i64m1(op0, op1, op2) \ +__builtin_rvv_vdiv_vx_i64m1((vint64m1_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vdiv_vx_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdiv_vx_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vdiv_vx_i64m2(op0, op1, op2) \ +__builtin_rvv_vdiv_vx_i64m2((vint64m2_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vdiv_vx_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdiv_vx_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vdiv_vx_i64m4(op0, op1, op2) \ +__builtin_rvv_vdiv_vx_i64m4((vint64m4_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vdiv_vx_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdiv_vx_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vdiv_vx_i64m8(op0, op1, op2) \ +__builtin_rvv_vdiv_vx_i64m8((vint64m8_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vdiv_vx_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vdiv_vx_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vremu_vv_u8m1(op0, op1, op2) \ +__builtin_rvv_vremu_vv_u8m1((vuint8m1_t)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vremu_vv_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vremu_vv_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vremu_vv_u8m2(op0, op1, op2) \ +__builtin_rvv_vremu_vv_u8m2((vuint8m2_t)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vremu_vv_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vremu_vv_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vremu_vv_u8m4(op0, op1, op2) \ +__builtin_rvv_vremu_vv_u8m4((vuint8m4_t)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vremu_vv_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vremu_vv_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vremu_vv_u8m8(op0, op1, op2) \ +__builtin_rvv_vremu_vv_u8m8((vuint8m8_t)(op0), (vuint8m8_t)(op1), (size_t)(op2)) +#define vremu_vv_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vremu_vv_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (vuint8m8_t)(op3), (size_t)(op4)) +#define vremu_vv_u8mf2(op0, op1, op2) \ +__builtin_rvv_vremu_vv_u8mf2((vuint8mf2_t)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vremu_vv_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vremu_vv_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vremu_vv_u8mf4(op0, op1, op2) \ +__builtin_rvv_vremu_vv_u8mf4((vuint8mf4_t)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vremu_vv_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vremu_vv_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vremu_vv_u8mf8(op0, op1, op2) \ +__builtin_rvv_vremu_vv_u8mf8((vuint8mf8_t)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vremu_vv_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vremu_vv_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vremu_vv_u16m1(op0, op1, op2) \ +__builtin_rvv_vremu_vv_u16m1((vuint16m1_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vremu_vv_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vremu_vv_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vremu_vv_u16m2(op0, op1, op2) \ +__builtin_rvv_vremu_vv_u16m2((vuint16m2_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vremu_vv_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vremu_vv_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vremu_vv_u16m4(op0, op1, op2) \ +__builtin_rvv_vremu_vv_u16m4((vuint16m4_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vremu_vv_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vremu_vv_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vremu_vv_u16m8(op0, op1, op2) \ +__builtin_rvv_vremu_vv_u16m8((vuint16m8_t)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vremu_vv_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vremu_vv_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vremu_vv_u16mf2(op0, op1, op2) \ +__builtin_rvv_vremu_vv_u16mf2((vuint16mf2_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vremu_vv_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vremu_vv_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vremu_vv_u16mf4(op0, op1, op2) \ +__builtin_rvv_vremu_vv_u16mf4((vuint16mf4_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vremu_vv_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vremu_vv_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vremu_vv_u32m1(op0, op1, op2) \ +__builtin_rvv_vremu_vv_u32m1((vuint32m1_t)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vremu_vv_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vremu_vv_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vremu_vv_u32m2(op0, op1, op2) \ +__builtin_rvv_vremu_vv_u32m2((vuint32m2_t)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vremu_vv_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vremu_vv_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vremu_vv_u32m4(op0, op1, op2) \ +__builtin_rvv_vremu_vv_u32m4((vuint32m4_t)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vremu_vv_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vremu_vv_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vremu_vv_u32m8(op0, op1, op2) \ +__builtin_rvv_vremu_vv_u32m8((vuint32m8_t)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vremu_vv_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vremu_vv_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vremu_vv_u32mf2(op0, op1, op2) \ +__builtin_rvv_vremu_vv_u32mf2((vuint32mf2_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vremu_vv_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vremu_vv_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vremu_vv_u64m1(op0, op1, op2) \ +__builtin_rvv_vremu_vv_u64m1((vuint64m1_t)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vremu_vv_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vremu_vv_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vremu_vv_u64m2(op0, op1, op2) \ +__builtin_rvv_vremu_vv_u64m2((vuint64m2_t)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vremu_vv_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vremu_vv_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vremu_vv_u64m4(op0, op1, op2) \ +__builtin_rvv_vremu_vv_u64m4((vuint64m4_t)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vremu_vv_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vremu_vv_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vremu_vv_u64m8(op0, op1, op2) \ +__builtin_rvv_vremu_vv_u64m8((vuint64m8_t)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vremu_vv_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vremu_vv_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vremu_vx_u8m1(op0, op1, op2) \ +__builtin_rvv_vremu_vx_u8m1((vuint8m1_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vremu_vx_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vremu_vx_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vremu_vx_u8m2(op0, op1, op2) \ +__builtin_rvv_vremu_vx_u8m2((vuint8m2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vremu_vx_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vremu_vx_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vremu_vx_u8m4(op0, op1, op2) \ +__builtin_rvv_vremu_vx_u8m4((vuint8m4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vremu_vx_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vremu_vx_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vremu_vx_u8m8(op0, op1, op2) \ +__builtin_rvv_vremu_vx_u8m8((vuint8m8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vremu_vx_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vremu_vx_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vremu_vx_u8mf2(op0, op1, op2) \ +__builtin_rvv_vremu_vx_u8mf2((vuint8mf2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vremu_vx_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vremu_vx_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vremu_vx_u8mf4(op0, op1, op2) \ +__builtin_rvv_vremu_vx_u8mf4((vuint8mf4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vremu_vx_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vremu_vx_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vremu_vx_u8mf8(op0, op1, op2) \ +__builtin_rvv_vremu_vx_u8mf8((vuint8mf8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vremu_vx_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vremu_vx_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vremu_vx_u16m1(op0, op1, op2) \ +__builtin_rvv_vremu_vx_u16m1((vuint16m1_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vremu_vx_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vremu_vx_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vremu_vx_u16m2(op0, op1, op2) \ +__builtin_rvv_vremu_vx_u16m2((vuint16m2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vremu_vx_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vremu_vx_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vremu_vx_u16m4(op0, op1, op2) \ +__builtin_rvv_vremu_vx_u16m4((vuint16m4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vremu_vx_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vremu_vx_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vremu_vx_u16m8(op0, op1, op2) \ +__builtin_rvv_vremu_vx_u16m8((vuint16m8_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vremu_vx_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vremu_vx_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vremu_vx_u16mf2(op0, op1, op2) \ +__builtin_rvv_vremu_vx_u16mf2((vuint16mf2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vremu_vx_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vremu_vx_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vremu_vx_u16mf4(op0, op1, op2) \ +__builtin_rvv_vremu_vx_u16mf4((vuint16mf4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vremu_vx_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vremu_vx_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vremu_vx_u32m1(op0, op1, op2) \ +__builtin_rvv_vremu_vx_u32m1((vuint32m1_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vremu_vx_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vremu_vx_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vremu_vx_u32m2(op0, op1, op2) \ +__builtin_rvv_vremu_vx_u32m2((vuint32m2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vremu_vx_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vremu_vx_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vremu_vx_u32m4(op0, op1, op2) \ +__builtin_rvv_vremu_vx_u32m4((vuint32m4_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vremu_vx_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vremu_vx_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vremu_vx_u32m8(op0, op1, op2) \ +__builtin_rvv_vremu_vx_u32m8((vuint32m8_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vremu_vx_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vremu_vx_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vremu_vx_u32mf2(op0, op1, op2) \ +__builtin_rvv_vremu_vx_u32mf2((vuint32mf2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vremu_vx_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vremu_vx_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vremu_vx_u64m1(op0, op1, op2) \ +__builtin_rvv_vremu_vx_u64m1((vuint64m1_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vremu_vx_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vremu_vx_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vremu_vx_u64m2(op0, op1, op2) \ +__builtin_rvv_vremu_vx_u64m2((vuint64m2_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vremu_vx_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vremu_vx_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vremu_vx_u64m4(op0, op1, op2) \ +__builtin_rvv_vremu_vx_u64m4((vuint64m4_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vremu_vx_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vremu_vx_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vremu_vx_u64m8(op0, op1, op2) \ +__builtin_rvv_vremu_vx_u64m8((vuint64m8_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vremu_vx_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vremu_vx_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vrem_vv_i8m1(op0, op1, op2) \ +__builtin_rvv_vrem_vv_i8m1((vint8m1_t)(op0), (vint8m1_t)(op1), (size_t)(op2)) +#define vrem_vv_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrem_vv_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vrem_vv_i8m2(op0, op1, op2) \ +__builtin_rvv_vrem_vv_i8m2((vint8m2_t)(op0), (vint8m2_t)(op1), (size_t)(op2)) +#define vrem_vv_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrem_vv_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vrem_vv_i8m4(op0, op1, op2) \ +__builtin_rvv_vrem_vv_i8m4((vint8m4_t)(op0), (vint8m4_t)(op1), (size_t)(op2)) +#define vrem_vv_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrem_vv_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vrem_vv_i8m8(op0, op1, op2) \ +__builtin_rvv_vrem_vv_i8m8((vint8m8_t)(op0), (vint8m8_t)(op1), (size_t)(op2)) +#define vrem_vv_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrem_vv_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (vint8m8_t)(op3), (size_t)(op4)) +#define vrem_vv_i8mf2(op0, op1, op2) \ +__builtin_rvv_vrem_vv_i8mf2((vint8mf2_t)(op0), (vint8mf2_t)(op1), (size_t)(op2)) +#define vrem_vv_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrem_vv_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vrem_vv_i8mf4(op0, op1, op2) \ +__builtin_rvv_vrem_vv_i8mf4((vint8mf4_t)(op0), (vint8mf4_t)(op1), (size_t)(op2)) +#define vrem_vv_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrem_vv_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vrem_vv_i8mf8(op0, op1, op2) \ +__builtin_rvv_vrem_vv_i8mf8((vint8mf8_t)(op0), (vint8mf8_t)(op1), (size_t)(op2)) +#define vrem_vv_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrem_vv_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vrem_vv_i16m1(op0, op1, op2) \ +__builtin_rvv_vrem_vv_i16m1((vint16m1_t)(op0), (vint16m1_t)(op1), (size_t)(op2)) +#define vrem_vv_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrem_vv_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vrem_vv_i16m2(op0, op1, op2) \ +__builtin_rvv_vrem_vv_i16m2((vint16m2_t)(op0), (vint16m2_t)(op1), (size_t)(op2)) +#define vrem_vv_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrem_vv_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vrem_vv_i16m4(op0, op1, op2) \ +__builtin_rvv_vrem_vv_i16m4((vint16m4_t)(op0), (vint16m4_t)(op1), (size_t)(op2)) +#define vrem_vv_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrem_vv_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vrem_vv_i16m8(op0, op1, op2) \ +__builtin_rvv_vrem_vv_i16m8((vint16m8_t)(op0), (vint16m8_t)(op1), (size_t)(op2)) +#define vrem_vv_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrem_vv_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (vint16m8_t)(op3), (size_t)(op4)) +#define vrem_vv_i16mf2(op0, op1, op2) \ +__builtin_rvv_vrem_vv_i16mf2((vint16mf2_t)(op0), (vint16mf2_t)(op1), (size_t)(op2)) +#define vrem_vv_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrem_vv_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vrem_vv_i16mf4(op0, op1, op2) \ +__builtin_rvv_vrem_vv_i16mf4((vint16mf4_t)(op0), (vint16mf4_t)(op1), (size_t)(op2)) +#define vrem_vv_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrem_vv_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vrem_vv_i32m1(op0, op1, op2) \ +__builtin_rvv_vrem_vv_i32m1((vint32m1_t)(op0), (vint32m1_t)(op1), (size_t)(op2)) +#define vrem_vv_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrem_vv_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vrem_vv_i32m2(op0, op1, op2) \ +__builtin_rvv_vrem_vv_i32m2((vint32m2_t)(op0), (vint32m2_t)(op1), (size_t)(op2)) +#define vrem_vv_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrem_vv_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vrem_vv_i32m4(op0, op1, op2) \ +__builtin_rvv_vrem_vv_i32m4((vint32m4_t)(op0), (vint32m4_t)(op1), (size_t)(op2)) +#define vrem_vv_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrem_vv_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vrem_vv_i32m8(op0, op1, op2) \ +__builtin_rvv_vrem_vv_i32m8((vint32m8_t)(op0), (vint32m8_t)(op1), (size_t)(op2)) +#define vrem_vv_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrem_vv_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vrem_vv_i32mf2(op0, op1, op2) \ +__builtin_rvv_vrem_vv_i32mf2((vint32mf2_t)(op0), (vint32mf2_t)(op1), (size_t)(op2)) +#define vrem_vv_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrem_vv_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vrem_vv_i64m1(op0, op1, op2) \ +__builtin_rvv_vrem_vv_i64m1((vint64m1_t)(op0), (vint64m1_t)(op1), (size_t)(op2)) +#define vrem_vv_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrem_vv_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vrem_vv_i64m2(op0, op1, op2) \ +__builtin_rvv_vrem_vv_i64m2((vint64m2_t)(op0), (vint64m2_t)(op1), (size_t)(op2)) +#define vrem_vv_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrem_vv_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vrem_vv_i64m4(op0, op1, op2) \ +__builtin_rvv_vrem_vv_i64m4((vint64m4_t)(op0), (vint64m4_t)(op1), (size_t)(op2)) +#define vrem_vv_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrem_vv_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vrem_vv_i64m8(op0, op1, op2) \ +__builtin_rvv_vrem_vv_i64m8((vint64m8_t)(op0), (vint64m8_t)(op1), (size_t)(op2)) +#define vrem_vv_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrem_vv_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vrem_vx_i8m1(op0, op1, op2) \ +__builtin_rvv_vrem_vx_i8m1((vint8m1_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vrem_vx_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrem_vx_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vrem_vx_i8m2(op0, op1, op2) \ +__builtin_rvv_vrem_vx_i8m2((vint8m2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vrem_vx_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrem_vx_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vrem_vx_i8m4(op0, op1, op2) \ +__builtin_rvv_vrem_vx_i8m4((vint8m4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vrem_vx_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrem_vx_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vrem_vx_i8m8(op0, op1, op2) \ +__builtin_rvv_vrem_vx_i8m8((vint8m8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vrem_vx_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrem_vx_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vrem_vx_i8mf2(op0, op1, op2) \ +__builtin_rvv_vrem_vx_i8mf2((vint8mf2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vrem_vx_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrem_vx_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vrem_vx_i8mf4(op0, op1, op2) \ +__builtin_rvv_vrem_vx_i8mf4((vint8mf4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vrem_vx_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrem_vx_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vrem_vx_i8mf8(op0, op1, op2) \ +__builtin_rvv_vrem_vx_i8mf8((vint8mf8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vrem_vx_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrem_vx_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vrem_vx_i16m1(op0, op1, op2) \ +__builtin_rvv_vrem_vx_i16m1((vint16m1_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vrem_vx_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrem_vx_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vrem_vx_i16m2(op0, op1, op2) \ +__builtin_rvv_vrem_vx_i16m2((vint16m2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vrem_vx_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrem_vx_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vrem_vx_i16m4(op0, op1, op2) \ +__builtin_rvv_vrem_vx_i16m4((vint16m4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vrem_vx_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrem_vx_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vrem_vx_i16m8(op0, op1, op2) \ +__builtin_rvv_vrem_vx_i16m8((vint16m8_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vrem_vx_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrem_vx_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vrem_vx_i16mf2(op0, op1, op2) \ +__builtin_rvv_vrem_vx_i16mf2((vint16mf2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vrem_vx_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrem_vx_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vrem_vx_i16mf4(op0, op1, op2) \ +__builtin_rvv_vrem_vx_i16mf4((vint16mf4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vrem_vx_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrem_vx_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vrem_vx_i32m1(op0, op1, op2) \ +__builtin_rvv_vrem_vx_i32m1((vint32m1_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vrem_vx_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrem_vx_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vrem_vx_i32m2(op0, op1, op2) \ +__builtin_rvv_vrem_vx_i32m2((vint32m2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vrem_vx_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrem_vx_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vrem_vx_i32m4(op0, op1, op2) \ +__builtin_rvv_vrem_vx_i32m4((vint32m4_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vrem_vx_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrem_vx_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vrem_vx_i32m8(op0, op1, op2) \ +__builtin_rvv_vrem_vx_i32m8((vint32m8_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vrem_vx_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrem_vx_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vrem_vx_i32mf2(op0, op1, op2) \ +__builtin_rvv_vrem_vx_i32mf2((vint32mf2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vrem_vx_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrem_vx_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vrem_vx_i64m1(op0, op1, op2) \ +__builtin_rvv_vrem_vx_i64m1((vint64m1_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vrem_vx_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrem_vx_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vrem_vx_i64m2(op0, op1, op2) \ +__builtin_rvv_vrem_vx_i64m2((vint64m2_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vrem_vx_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrem_vx_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vrem_vx_i64m4(op0, op1, op2) \ +__builtin_rvv_vrem_vx_i64m4((vint64m4_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vrem_vx_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrem_vx_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vrem_vx_i64m8(op0, op1, op2) \ +__builtin_rvv_vrem_vx_i64m8((vint64m8_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vrem_vx_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrem_vx_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vwmul_vv_i16mf4(op0, op1, op2) \ +__builtin_rvv_vwmul_vv_i16mf4((vint8mf8_t)(op0), (vint8mf8_t)(op1), (size_t)(op2)) +#define vwmul_vv_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmul_vv_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vwmul_vv_i16mf2(op0, op1, op2) \ +__builtin_rvv_vwmul_vv_i16mf2((vint8mf4_t)(op0), (vint8mf4_t)(op1), (size_t)(op2)) +#define vwmul_vv_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmul_vv_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vwmul_vv_i16m1(op0, op1, op2) \ +__builtin_rvv_vwmul_vv_i16m1((vint8mf2_t)(op0), (vint8mf2_t)(op1), (size_t)(op2)) +#define vwmul_vv_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmul_vv_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vwmul_vv_i16m2(op0, op1, op2) \ +__builtin_rvv_vwmul_vv_i16m2((vint8m1_t)(op0), (vint8m1_t)(op1), (size_t)(op2)) +#define vwmul_vv_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmul_vv_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vwmul_vv_i16m4(op0, op1, op2) \ +__builtin_rvv_vwmul_vv_i16m4((vint8m2_t)(op0), (vint8m2_t)(op1), (size_t)(op2)) +#define vwmul_vv_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmul_vv_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vwmul_vv_i16m8(op0, op1, op2) \ +__builtin_rvv_vwmul_vv_i16m8((vint8m4_t)(op0), (vint8m4_t)(op1), (size_t)(op2)) +#define vwmul_vv_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmul_vv_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint8m4_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vwmul_vv_i32mf2(op0, op1, op2) \ +__builtin_rvv_vwmul_vv_i32mf2((vint16mf4_t)(op0), (vint16mf4_t)(op1), (size_t)(op2)) +#define vwmul_vv_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmul_vv_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vwmul_vv_i32m1(op0, op1, op2) \ +__builtin_rvv_vwmul_vv_i32m1((vint16mf2_t)(op0), (vint16mf2_t)(op1), (size_t)(op2)) +#define vwmul_vv_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmul_vv_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vwmul_vv_i32m2(op0, op1, op2) \ +__builtin_rvv_vwmul_vv_i32m2((vint16m1_t)(op0), (vint16m1_t)(op1), (size_t)(op2)) +#define vwmul_vv_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmul_vv_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vwmul_vv_i32m4(op0, op1, op2) \ +__builtin_rvv_vwmul_vv_i32m4((vint16m2_t)(op0), (vint16m2_t)(op1), (size_t)(op2)) +#define vwmul_vv_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmul_vv_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vwmul_vv_i32m8(op0, op1, op2) \ +__builtin_rvv_vwmul_vv_i32m8((vint16m4_t)(op0), (vint16m4_t)(op1), (size_t)(op2)) +#define vwmul_vv_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmul_vv_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint16m4_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vwmul_vv_i64m1(op0, op1, op2) \ +__builtin_rvv_vwmul_vv_i64m1((vint32mf2_t)(op0), (vint32mf2_t)(op1), (size_t)(op2)) +#define vwmul_vv_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmul_vv_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vwmul_vv_i64m2(op0, op1, op2) \ +__builtin_rvv_vwmul_vv_i64m2((vint32m1_t)(op0), (vint32m1_t)(op1), (size_t)(op2)) +#define vwmul_vv_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmul_vv_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vwmul_vv_i64m4(op0, op1, op2) \ +__builtin_rvv_vwmul_vv_i64m4((vint32m2_t)(op0), (vint32m2_t)(op1), (size_t)(op2)) +#define vwmul_vv_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmul_vv_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vwmul_vv_i64m8(op0, op1, op2) \ +__builtin_rvv_vwmul_vv_i64m8((vint32m4_t)(op0), (vint32m4_t)(op1), (size_t)(op2)) +#define vwmul_vv_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmul_vv_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vsuxei8_v_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_u32m1((uint32_t *)(op0), (vuint8mf4_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vsuxei8_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint8mf4_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vsuxei8_v_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_u32m2((uint32_t *)(op0), (vuint8mf2_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vsuxei8_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint8mf2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vsuxei8_v_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_u32m4((uint32_t *)(op0), (vuint8m1_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vsuxei8_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint8m1_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vsuxei8_v_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_u32m8((uint32_t *)(op0), (vuint8m2_t)(op1), (vuint32m8_t)(op2), (size_t)(op3)) +#define vsuxei8_v_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_u32m8_m((vbool4_t)(op0), (uint32_t *)(op1), (vuint8m2_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vsuxei8_v_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_u32mf2((uint32_t *)(op0), (vuint8mf8_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vsuxei8_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint8mf8_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vwmul_vx_i16mf4(op0, op1, op2) \ +__builtin_rvv_vwmul_vx_i16mf4((vint8mf8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vwmul_vx_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmul_vx_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint8mf8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vwmul_vx_i16mf2(op0, op1, op2) \ +__builtin_rvv_vwmul_vx_i16mf2((vint8mf4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vwmul_vx_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmul_vx_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint8mf4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vwmul_vx_i16m1(op0, op1, op2) \ +__builtin_rvv_vwmul_vx_i16m1((vint8mf2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vwmul_vx_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmul_vx_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint8mf2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vwmul_vx_i16m2(op0, op1, op2) \ +__builtin_rvv_vwmul_vx_i16m2((vint8m1_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vwmul_vx_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmul_vx_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint8m1_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vwmul_vx_i16m4(op0, op1, op2) \ +__builtin_rvv_vwmul_vx_i16m4((vint8m2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vwmul_vx_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmul_vx_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint8m2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vwmul_vx_i16m8(op0, op1, op2) \ +__builtin_rvv_vwmul_vx_i16m8((vint8m4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vwmul_vx_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmul_vx_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint8m4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vwmul_vx_i32mf2(op0, op1, op2) \ +__builtin_rvv_vwmul_vx_i32mf2((vint16mf4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vwmul_vx_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmul_vx_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint16mf4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vwmul_vx_i32m1(op0, op1, op2) \ +__builtin_rvv_vwmul_vx_i32m1((vint16mf2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vwmul_vx_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmul_vx_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint16mf2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vwmul_vx_i32m2(op0, op1, op2) \ +__builtin_rvv_vwmul_vx_i32m2((vint16m1_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vwmul_vx_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmul_vx_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint16m1_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vwmul_vx_i32m4(op0, op1, op2) \ +__builtin_rvv_vwmul_vx_i32m4((vint16m2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vwmul_vx_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmul_vx_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint16m2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vwmul_vx_i32m8(op0, op1, op2) \ +__builtin_rvv_vwmul_vx_i32m8((vint16m4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vwmul_vx_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmul_vx_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint16m4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vwmul_vx_i64m1(op0, op1, op2) \ +__builtin_rvv_vwmul_vx_i64m1((vint32mf2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vwmul_vx_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmul_vx_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint32mf2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vwmul_vx_i64m2(op0, op1, op2) \ +__builtin_rvv_vwmul_vx_i64m2((vint32m1_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vwmul_vx_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmul_vx_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint32m1_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vwmul_vx_i64m4(op0, op1, op2) \ +__builtin_rvv_vwmul_vx_i64m4((vint32m2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vwmul_vx_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmul_vx_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint32m2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vwmul_vx_i64m8(op0, op1, op2) \ +__builtin_rvv_vwmul_vx_i64m8((vint32m4_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vwmul_vx_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmul_vx_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint32m4_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vwmulu_vv_u16mf4(op0, op1, op2) \ +__builtin_rvv_vwmulu_vv_u16mf4((vuint8mf8_t)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vwmulu_vv_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulu_vv_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vwmulu_vv_u16mf2(op0, op1, op2) \ +__builtin_rvv_vwmulu_vv_u16mf2((vuint8mf4_t)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vwmulu_vv_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulu_vv_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vwmulu_vv_u16m1(op0, op1, op2) \ +__builtin_rvv_vwmulu_vv_u16m1((vuint8mf2_t)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vwmulu_vv_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulu_vv_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vwmulu_vv_u16m2(op0, op1, op2) \ +__builtin_rvv_vwmulu_vv_u16m2((vuint8m1_t)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vwmulu_vv_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulu_vv_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vwmulu_vv_u16m4(op0, op1, op2) \ +__builtin_rvv_vwmulu_vv_u16m4((vuint8m2_t)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vwmulu_vv_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulu_vv_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vwmulu_vv_u16m8(op0, op1, op2) \ +__builtin_rvv_vwmulu_vv_u16m8((vuint8m4_t)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vwmulu_vv_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulu_vv_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vwmulu_vv_u32mf2(op0, op1, op2) \ +__builtin_rvv_vwmulu_vv_u32mf2((vuint16mf4_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vwmulu_vv_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulu_vv_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vwmulu_vv_u32m1(op0, op1, op2) \ +__builtin_rvv_vwmulu_vv_u32m1((vuint16mf2_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vwmulu_vv_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulu_vv_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vwmulu_vv_u32m2(op0, op1, op2) \ +__builtin_rvv_vwmulu_vv_u32m2((vuint16m1_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vwmulu_vv_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulu_vv_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vwmulu_vv_u32m4(op0, op1, op2) \ +__builtin_rvv_vwmulu_vv_u32m4((vuint16m2_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vwmulu_vv_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulu_vv_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vwmulu_vv_u32m8(op0, op1, op2) \ +__builtin_rvv_vwmulu_vv_u32m8((vuint16m4_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vwmulu_vv_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulu_vv_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vwmulu_vv_u64m1(op0, op1, op2) \ +__builtin_rvv_vwmulu_vv_u64m1((vuint32mf2_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vwmulu_vv_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulu_vv_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vwmulu_vv_u64m2(op0, op1, op2) \ +__builtin_rvv_vwmulu_vv_u64m2((vuint32m1_t)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vwmulu_vv_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulu_vv_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vwmulu_vv_u64m4(op0, op1, op2) \ +__builtin_rvv_vwmulu_vv_u64m4((vuint32m2_t)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vwmulu_vv_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulu_vv_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vwmulu_vv_u64m8(op0, op1, op2) \ +__builtin_rvv_vwmulu_vv_u64m8((vuint32m4_t)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vwmulu_vv_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulu_vv_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vwmulu_vx_u16mf4(op0, op1, op2) \ +__builtin_rvv_vwmulu_vx_u16mf4((vuint8mf8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vwmulu_vx_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulu_vx_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint8mf8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vwmulu_vx_u16mf2(op0, op1, op2) \ +__builtin_rvv_vwmulu_vx_u16mf2((vuint8mf4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vwmulu_vx_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulu_vx_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint8mf4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vwmulu_vx_u16m1(op0, op1, op2) \ +__builtin_rvv_vwmulu_vx_u16m1((vuint8mf2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vwmulu_vx_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulu_vx_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint8mf2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vwmulu_vx_u16m2(op0, op1, op2) \ +__builtin_rvv_vwmulu_vx_u16m2((vuint8m1_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vwmulu_vx_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulu_vx_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint8m1_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vwmulu_vx_u16m4(op0, op1, op2) \ +__builtin_rvv_vwmulu_vx_u16m4((vuint8m2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vwmulu_vx_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulu_vx_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint8m2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vwmulu_vx_u16m8(op0, op1, op2) \ +__builtin_rvv_vwmulu_vx_u16m8((vuint8m4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vwmulu_vx_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulu_vx_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint8m4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vwmulu_vx_u32mf2(op0, op1, op2) \ +__builtin_rvv_vwmulu_vx_u32mf2((vuint16mf4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vwmulu_vx_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulu_vx_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint16mf4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vwmulu_vx_u32m1(op0, op1, op2) \ +__builtin_rvv_vwmulu_vx_u32m1((vuint16mf2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vwmulu_vx_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulu_vx_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint16mf2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vwmulu_vx_u32m2(op0, op1, op2) \ +__builtin_rvv_vwmulu_vx_u32m2((vuint16m1_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vwmulu_vx_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulu_vx_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint16m1_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vwmulu_vx_u32m4(op0, op1, op2) \ +__builtin_rvv_vwmulu_vx_u32m4((vuint16m2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vwmulu_vx_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulu_vx_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint16m2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vwmulu_vx_u32m8(op0, op1, op2) \ +__builtin_rvv_vwmulu_vx_u32m8((vuint16m4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vwmulu_vx_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulu_vx_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint16m4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vwmulu_vx_u64m1(op0, op1, op2) \ +__builtin_rvv_vwmulu_vx_u64m1((vuint32mf2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vwmulu_vx_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulu_vx_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint32mf2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vwmulu_vx_u64m2(op0, op1, op2) \ +__builtin_rvv_vwmulu_vx_u64m2((vuint32m1_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vwmulu_vx_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulu_vx_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint32m1_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vwmulu_vx_u64m4(op0, op1, op2) \ +__builtin_rvv_vwmulu_vx_u64m4((vuint32m2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vwmulu_vx_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulu_vx_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint32m2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vwmulu_vx_u64m8(op0, op1, op2) \ +__builtin_rvv_vwmulu_vx_u64m8((vuint32m4_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vwmulu_vx_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulu_vx_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint32m4_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vwmulsu_vv_i16mf4(op0, op1, op2) \ +__builtin_rvv_vwmulsu_vv_i16mf4((vint8mf8_t)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vwmulsu_vv_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulsu_vv_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vwmulsu_vv_i16mf2(op0, op1, op2) \ +__builtin_rvv_vwmulsu_vv_i16mf2((vint8mf4_t)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vwmulsu_vv_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulsu_vv_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vwmulsu_vv_i16m1(op0, op1, op2) \ +__builtin_rvv_vwmulsu_vv_i16m1((vint8mf2_t)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vwmulsu_vv_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulsu_vv_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vwmulsu_vv_i16m2(op0, op1, op2) \ +__builtin_rvv_vwmulsu_vv_i16m2((vint8m1_t)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vwmulsu_vv_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulsu_vv_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vwmulsu_vv_i16m4(op0, op1, op2) \ +__builtin_rvv_vwmulsu_vv_i16m4((vint8m2_t)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vwmulsu_vv_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulsu_vv_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vwmulsu_vv_i16m8(op0, op1, op2) \ +__builtin_rvv_vwmulsu_vv_i16m8((vint8m4_t)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vwmulsu_vv_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulsu_vv_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vwmulsu_vv_i32mf2(op0, op1, op2) \ +__builtin_rvv_vwmulsu_vv_i32mf2((vint16mf4_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vwmulsu_vv_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulsu_vv_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vwmulsu_vv_i32m1(op0, op1, op2) \ +__builtin_rvv_vwmulsu_vv_i32m1((vint16mf2_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vwmulsu_vv_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulsu_vv_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vwmulsu_vv_i32m2(op0, op1, op2) \ +__builtin_rvv_vwmulsu_vv_i32m2((vint16m1_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vwmulsu_vv_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulsu_vv_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vwmulsu_vv_i32m4(op0, op1, op2) \ +__builtin_rvv_vwmulsu_vv_i32m4((vint16m2_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vwmulsu_vv_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulsu_vv_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vwmulsu_vv_i32m8(op0, op1, op2) \ +__builtin_rvv_vwmulsu_vv_i32m8((vint16m4_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vwmulsu_vv_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulsu_vv_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vwmulsu_vv_i64m1(op0, op1, op2) \ +__builtin_rvv_vwmulsu_vv_i64m1((vint32mf2_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vwmulsu_vv_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulsu_vv_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vwmulsu_vv_i64m2(op0, op1, op2) \ +__builtin_rvv_vwmulsu_vv_i64m2((vint32m1_t)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vwmulsu_vv_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulsu_vv_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vwmulsu_vv_i64m4(op0, op1, op2) \ +__builtin_rvv_vwmulsu_vv_i64m4((vint32m2_t)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vwmulsu_vv_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulsu_vv_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vwmulsu_vv_i64m8(op0, op1, op2) \ +__builtin_rvv_vwmulsu_vv_i64m8((vint32m4_t)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vwmulsu_vv_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulsu_vv_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vwmulsu_vx_i16mf4(op0, op1, op2) \ +__builtin_rvv_vwmulsu_vx_i16mf4((vint8mf8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vwmulsu_vx_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulsu_vx_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint8mf8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vwmulsu_vx_i16mf2(op0, op1, op2) \ +__builtin_rvv_vwmulsu_vx_i16mf2((vint8mf4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vwmulsu_vx_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulsu_vx_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint8mf4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vwmulsu_vx_i16m1(op0, op1, op2) \ +__builtin_rvv_vwmulsu_vx_i16m1((vint8mf2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vwmulsu_vx_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulsu_vx_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint8mf2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vwmulsu_vx_i16m2(op0, op1, op2) \ +__builtin_rvv_vwmulsu_vx_i16m2((vint8m1_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vwmulsu_vx_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulsu_vx_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint8m1_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vwmulsu_vx_i16m4(op0, op1, op2) \ +__builtin_rvv_vwmulsu_vx_i16m4((vint8m2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vwmulsu_vx_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulsu_vx_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint8m2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vwmulsu_vx_i16m8(op0, op1, op2) \ +__builtin_rvv_vwmulsu_vx_i16m8((vint8m4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vwmulsu_vx_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulsu_vx_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint8m4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vwmulsu_vx_i32mf2(op0, op1, op2) \ +__builtin_rvv_vwmulsu_vx_i32mf2((vint16mf4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vwmulsu_vx_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulsu_vx_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint16mf4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vwmulsu_vx_i32m1(op0, op1, op2) \ +__builtin_rvv_vwmulsu_vx_i32m1((vint16mf2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vwmulsu_vx_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulsu_vx_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint16mf2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vwmulsu_vx_i32m2(op0, op1, op2) \ +__builtin_rvv_vwmulsu_vx_i32m2((vint16m1_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vwmulsu_vx_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulsu_vx_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint16m1_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vwmulsu_vx_i32m4(op0, op1, op2) \ +__builtin_rvv_vwmulsu_vx_i32m4((vint16m2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vwmulsu_vx_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulsu_vx_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint16m2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vwmulsu_vx_i32m8(op0, op1, op2) \ +__builtin_rvv_vwmulsu_vx_i32m8((vint16m4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vwmulsu_vx_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulsu_vx_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint16m4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vwmulsu_vx_i64m1(op0, op1, op2) \ +__builtin_rvv_vwmulsu_vx_i64m1((vint32mf2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vwmulsu_vx_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulsu_vx_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint32mf2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vwmulsu_vx_i64m2(op0, op1, op2) \ +__builtin_rvv_vwmulsu_vx_i64m2((vint32m1_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vwmulsu_vx_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulsu_vx_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint32m1_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vwmulsu_vx_i64m4(op0, op1, op2) \ +__builtin_rvv_vwmulsu_vx_i64m4((vint32m2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vwmulsu_vx_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulsu_vx_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint32m2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vwmulsu_vx_i64m8(op0, op1, op2) \ +__builtin_rvv_vwmulsu_vx_i64m8((vint32m4_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vwmulsu_vx_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmulsu_vx_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint32m4_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vmacc_vv_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vv_i8m1((vint8m1_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vmacc_vv_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vv_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vmacc_vv_i8m2(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vv_i8m2((vint8m2_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (size_t)(op3)) +#define vmacc_vv_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vv_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vmacc_vv_i8m4(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vv_i8m4((vint8m4_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (size_t)(op3)) +#define vmacc_vv_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vv_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vmacc_vv_i8m8(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vv_i8m8((vint8m8_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (size_t)(op3)) +#define vmacc_vv_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vv_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (vint8m8_t)(op3), (size_t)(op4)) +#define vmacc_vv_i8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vv_i8mf2((vint8mf2_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (size_t)(op3)) +#define vmacc_vv_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vv_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vmacc_vv_i8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vv_i8mf4((vint8mf4_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (size_t)(op3)) +#define vmacc_vv_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vv_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vmacc_vv_i8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vv_i8mf8((vint8mf8_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (size_t)(op3)) +#define vmacc_vv_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vv_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vmacc_vv_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vv_i16m1((vint16m1_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vmacc_vv_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vv_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vmacc_vv_i16m2(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vv_i16m2((vint16m2_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (size_t)(op3)) +#define vmacc_vv_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vv_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vmacc_vv_i16m4(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vv_i16m4((vint16m4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (size_t)(op3)) +#define vmacc_vv_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vv_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vmacc_vv_i16m8(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vv_i16m8((vint16m8_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (size_t)(op3)) +#define vmacc_vv_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vv_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (vint16m8_t)(op3), (size_t)(op4)) +#define vmacc_vv_i16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vv_i16mf2((vint16mf2_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (size_t)(op3)) +#define vmacc_vv_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vv_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vmacc_vv_i16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vv_i16mf4((vint16mf4_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (size_t)(op3)) +#define vmacc_vv_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vv_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vmacc_vv_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vv_i32m1((vint32m1_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vmacc_vv_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vv_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vmacc_vv_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vv_i32m2((vint32m2_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vmacc_vv_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vv_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vmacc_vv_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vv_i32m4((vint32m4_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vmacc_vv_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vv_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vmacc_vv_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vv_i32m8((vint32m8_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (size_t)(op3)) +#define vmacc_vv_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vv_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vmacc_vv_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vv_i32mf2((vint32mf2_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vmacc_vv_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vv_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vmacc_vv_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vv_i64m1((vint64m1_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vmacc_vv_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vv_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vmacc_vv_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vv_i64m2((vint64m2_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vmacc_vv_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vv_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vmacc_vv_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vv_i64m4((vint64m4_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vmacc_vv_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vv_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vmacc_vv_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vv_i64m8((vint64m8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vmacc_vv_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vv_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vmacc_vx_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vx_i8m1((vint8m1_t)(op0), (int8_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vmacc_vx_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vx_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (int8_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vmacc_vx_i8m2(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vx_i8m2((vint8m2_t)(op0), (int8_t)(op1), (vint8m2_t)(op2), (size_t)(op3)) +#define vmacc_vx_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vx_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (int8_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vmacc_vx_i8m4(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vx_i8m4((vint8m4_t)(op0), (int8_t)(op1), (vint8m4_t)(op2), (size_t)(op3)) +#define vmacc_vx_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vx_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (int8_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vmacc_vx_i8m8(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vx_i8m8((vint8m8_t)(op0), (int8_t)(op1), (vint8m8_t)(op2), (size_t)(op3)) +#define vmacc_vx_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vx_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (int8_t)(op2), (vint8m8_t)(op3), (size_t)(op4)) +#define vmacc_vx_i8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vx_i8mf2((vint8mf2_t)(op0), (int8_t)(op1), (vint8mf2_t)(op2), (size_t)(op3)) +#define vmacc_vx_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vx_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (int8_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vmacc_vx_i8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vx_i8mf4((vint8mf4_t)(op0), (int8_t)(op1), (vint8mf4_t)(op2), (size_t)(op3)) +#define vmacc_vx_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vx_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (int8_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vmacc_vx_i8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vx_i8mf8((vint8mf8_t)(op0), (int8_t)(op1), (vint8mf8_t)(op2), (size_t)(op3)) +#define vmacc_vx_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vx_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (int8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vmacc_vx_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vx_i16m1((vint16m1_t)(op0), (int16_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vmacc_vx_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vx_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (int16_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vmacc_vx_i16m2(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vx_i16m2((vint16m2_t)(op0), (int16_t)(op1), (vint16m2_t)(op2), (size_t)(op3)) +#define vmacc_vx_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vx_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (int16_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vmacc_vx_i16m4(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vx_i16m4((vint16m4_t)(op0), (int16_t)(op1), (vint16m4_t)(op2), (size_t)(op3)) +#define vmacc_vx_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vx_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (int16_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vmacc_vx_i16m8(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vx_i16m8((vint16m8_t)(op0), (int16_t)(op1), (vint16m8_t)(op2), (size_t)(op3)) +#define vmacc_vx_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vx_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (int16_t)(op2), (vint16m8_t)(op3), (size_t)(op4)) +#define vmacc_vx_i16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vx_i16mf2((vint16mf2_t)(op0), (int16_t)(op1), (vint16mf2_t)(op2), (size_t)(op3)) +#define vmacc_vx_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vx_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (int16_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vmacc_vx_i16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vx_i16mf4((vint16mf4_t)(op0), (int16_t)(op1), (vint16mf4_t)(op2), (size_t)(op3)) +#define vmacc_vx_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vx_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (int16_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vmacc_vx_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vx_i32m1((vint32m1_t)(op0), (int32_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vmacc_vx_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vx_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (int32_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vmacc_vx_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vx_i32m2((vint32m2_t)(op0), (int32_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vmacc_vx_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vx_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (int32_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vmacc_vx_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vx_i32m4((vint32m4_t)(op0), (int32_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vmacc_vx_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vx_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (int32_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vmacc_vx_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vx_i32m8((vint32m8_t)(op0), (int32_t)(op1), (vint32m8_t)(op2), (size_t)(op3)) +#define vmacc_vx_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vx_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (int32_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vmacc_vx_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vx_i32mf2((vint32mf2_t)(op0), (int32_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vmacc_vx_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vx_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (int32_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vmacc_vx_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vx_i64m1((vint64m1_t)(op0), (int64_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vmacc_vx_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vx_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (int64_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vmacc_vx_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vx_i64m2((vint64m2_t)(op0), (int64_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vmacc_vx_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vx_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (int64_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vmacc_vx_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vx_i64m4((vint64m4_t)(op0), (int64_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vmacc_vx_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vx_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (int64_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vmacc_vx_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vx_i64m8((vint64m8_t)(op0), (int64_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vmacc_vx_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vx_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (int64_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vmacc_vv_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vv_u8m1((vuint8m1_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vmacc_vv_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vv_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vmacc_vv_u8m2(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vv_u8m2((vuint8m2_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (size_t)(op3)) +#define vmacc_vv_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vv_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vmacc_vv_u8m4(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vv_u8m4((vuint8m4_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (size_t)(op3)) +#define vmacc_vv_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vv_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vmacc_vv_u8m8(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vv_u8m8((vuint8m8_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (size_t)(op3)) +#define vmacc_vv_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vv_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (vuint8m8_t)(op3), (size_t)(op4)) +#define vmacc_vv_u8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vv_u8mf2((vuint8mf2_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (size_t)(op3)) +#define vmacc_vv_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vv_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vmacc_vv_u8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vv_u8mf4((vuint8mf4_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (size_t)(op3)) +#define vmacc_vv_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vv_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vmacc_vv_u8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vv_u8mf8((vuint8mf8_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (size_t)(op3)) +#define vmacc_vv_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vv_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vmacc_vv_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vv_u16m1((vuint16m1_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vmacc_vv_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vv_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vmacc_vv_u16m2(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vv_u16m2((vuint16m2_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (size_t)(op3)) +#define vmacc_vv_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vv_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vmacc_vv_u16m4(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vv_u16m4((vuint16m4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (size_t)(op3)) +#define vmacc_vv_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vv_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vmacc_vv_u16m8(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vv_u16m8((vuint16m8_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (size_t)(op3)) +#define vmacc_vv_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vv_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vmacc_vv_u16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vv_u16mf2((vuint16mf2_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (size_t)(op3)) +#define vmacc_vv_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vv_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vmacc_vv_u16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vv_u16mf4((vuint16mf4_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (size_t)(op3)) +#define vmacc_vv_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vv_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vmacc_vv_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vv_u32m1((vuint32m1_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vmacc_vv_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vv_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vmacc_vv_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vv_u32m2((vuint32m2_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vmacc_vv_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vv_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vmacc_vv_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vv_u32m4((vuint32m4_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vmacc_vv_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vv_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vmacc_vv_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vv_u32m8((vuint32m8_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (size_t)(op3)) +#define vmacc_vv_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vv_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vmacc_vv_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vv_u32mf2((vuint32mf2_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vmacc_vv_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vv_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vmacc_vv_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vv_u64m1((vuint64m1_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vmacc_vv_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vv_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vmacc_vv_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vv_u64m2((vuint64m2_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vmacc_vv_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vv_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vmacc_vv_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vv_u64m4((vuint64m4_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vmacc_vv_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vv_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vmacc_vv_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vv_u64m8((vuint64m8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vmacc_vv_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vv_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vmacc_vx_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vx_u8m1((vuint8m1_t)(op0), (uint8_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vmacc_vx_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vx_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (uint8_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vmacc_vx_u8m2(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vx_u8m2((vuint8m2_t)(op0), (uint8_t)(op1), (vuint8m2_t)(op2), (size_t)(op3)) +#define vmacc_vx_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vx_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (uint8_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vmacc_vx_u8m4(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vx_u8m4((vuint8m4_t)(op0), (uint8_t)(op1), (vuint8m4_t)(op2), (size_t)(op3)) +#define vmacc_vx_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vx_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (uint8_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vmacc_vx_u8m8(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vx_u8m8((vuint8m8_t)(op0), (uint8_t)(op1), (vuint8m8_t)(op2), (size_t)(op3)) +#define vmacc_vx_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vx_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (uint8_t)(op2), (vuint8m8_t)(op3), (size_t)(op4)) +#define vmacc_vx_u8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vx_u8mf2((vuint8mf2_t)(op0), (uint8_t)(op1), (vuint8mf2_t)(op2), (size_t)(op3)) +#define vmacc_vx_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vx_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (uint8_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vmacc_vx_u8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vx_u8mf4((vuint8mf4_t)(op0), (uint8_t)(op1), (vuint8mf4_t)(op2), (size_t)(op3)) +#define vmacc_vx_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vx_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (uint8_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vmacc_vx_u8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vx_u8mf8((vuint8mf8_t)(op0), (uint8_t)(op1), (vuint8mf8_t)(op2), (size_t)(op3)) +#define vmacc_vx_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vx_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (uint8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vmacc_vx_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vx_u16m1((vuint16m1_t)(op0), (uint16_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vmacc_vx_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vx_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (uint16_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vmacc_vx_u16m2(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vx_u16m2((vuint16m2_t)(op0), (uint16_t)(op1), (vuint16m2_t)(op2), (size_t)(op3)) +#define vmacc_vx_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vx_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (uint16_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vmacc_vx_u16m4(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vx_u16m4((vuint16m4_t)(op0), (uint16_t)(op1), (vuint16m4_t)(op2), (size_t)(op3)) +#define vmacc_vx_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vx_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (uint16_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vmacc_vx_u16m8(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vx_u16m8((vuint16m8_t)(op0), (uint16_t)(op1), (vuint16m8_t)(op2), (size_t)(op3)) +#define vmacc_vx_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vx_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (uint16_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vmacc_vx_u16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vx_u16mf2((vuint16mf2_t)(op0), (uint16_t)(op1), (vuint16mf2_t)(op2), (size_t)(op3)) +#define vmacc_vx_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vx_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (uint16_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vmacc_vx_u16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vx_u16mf4((vuint16mf4_t)(op0), (uint16_t)(op1), (vuint16mf4_t)(op2), (size_t)(op3)) +#define vmacc_vx_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vx_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (uint16_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vmacc_vx_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vx_u32m1((vuint32m1_t)(op0), (uint32_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vmacc_vx_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vx_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (uint32_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vmacc_vx_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vx_u32m2((vuint32m2_t)(op0), (uint32_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vmacc_vx_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vx_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (uint32_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vmacc_vx_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vx_u32m4((vuint32m4_t)(op0), (uint32_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vmacc_vx_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vx_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (uint32_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vmacc_vx_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vx_u32m8((vuint32m8_t)(op0), (uint32_t)(op1), (vuint32m8_t)(op2), (size_t)(op3)) +#define vmacc_vx_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vx_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (uint32_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vmacc_vx_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vx_u32mf2((vuint32mf2_t)(op0), (uint32_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vmacc_vx_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vx_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (uint32_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vmacc_vx_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vx_u64m1((vuint64m1_t)(op0), (uint64_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vmacc_vx_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vx_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (uint64_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vmacc_vx_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vx_u64m2((vuint64m2_t)(op0), (uint64_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vmacc_vx_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vx_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (uint64_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vmacc_vx_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vx_u64m4((vuint64m4_t)(op0), (uint64_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vmacc_vx_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vx_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (uint64_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vmacc_vx_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vmacc_vx_u64m8((vuint64m8_t)(op0), (uint64_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vmacc_vx_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmacc_vx_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (uint64_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vnmsac_vv_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vv_i8m1((vint8m1_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vnmsac_vv_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vv_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vnmsac_vv_i8m2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vv_i8m2((vint8m2_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (size_t)(op3)) +#define vnmsac_vv_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vv_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vnmsac_vv_i8m4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vv_i8m4((vint8m4_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (size_t)(op3)) +#define vnmsac_vv_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vv_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vnmsac_vv_i8m8(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vv_i8m8((vint8m8_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (size_t)(op3)) +#define vnmsac_vv_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vv_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (vint8m8_t)(op3), (size_t)(op4)) +#define vnmsac_vv_i8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vv_i8mf2((vint8mf2_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (size_t)(op3)) +#define vnmsac_vv_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vv_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vnmsac_vv_i8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vv_i8mf4((vint8mf4_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (size_t)(op3)) +#define vnmsac_vv_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vv_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vnmsac_vv_i8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vv_i8mf8((vint8mf8_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (size_t)(op3)) +#define vnmsac_vv_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vv_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vnmsac_vv_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vv_i16m1((vint16m1_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vnmsac_vv_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vv_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vnmsac_vv_i16m2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vv_i16m2((vint16m2_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (size_t)(op3)) +#define vnmsac_vv_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vv_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vnmsac_vv_i16m4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vv_i16m4((vint16m4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (size_t)(op3)) +#define vnmsac_vv_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vv_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vnmsac_vv_i16m8(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vv_i16m8((vint16m8_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (size_t)(op3)) +#define vnmsac_vv_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vv_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (vint16m8_t)(op3), (size_t)(op4)) +#define vnmsac_vv_i16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vv_i16mf2((vint16mf2_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (size_t)(op3)) +#define vnmsac_vv_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vv_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vnmsac_vv_i16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vv_i16mf4((vint16mf4_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (size_t)(op3)) +#define vnmsac_vv_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vv_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vnmsac_vv_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vv_i32m1((vint32m1_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vnmsac_vv_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vv_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vnmsac_vv_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vv_i32m2((vint32m2_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vnmsac_vv_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vv_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vnmsac_vv_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vv_i32m4((vint32m4_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vnmsac_vv_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vv_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vnmsac_vv_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vv_i32m8((vint32m8_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (size_t)(op3)) +#define vnmsac_vv_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vv_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vnmsac_vv_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vv_i32mf2((vint32mf2_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vnmsac_vv_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vv_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vnmsac_vv_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vv_i64m1((vint64m1_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vnmsac_vv_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vv_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vnmsac_vv_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vv_i64m2((vint64m2_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vnmsac_vv_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vv_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vnmsac_vv_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vv_i64m4((vint64m4_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vnmsac_vv_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vv_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vnmsac_vv_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vv_i64m8((vint64m8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vnmsac_vv_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vv_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vsuxei16_v_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_i32m1((int32_t *)(op0), (vuint16mf2_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vsuxei16_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint16mf2_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vsuxei16_v_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_i32m2((int32_t *)(op0), (vuint16m1_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vsuxei16_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint16m1_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vsuxei16_v_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_i32m4((int32_t *)(op0), (vuint16m2_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vsuxei16_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vuint16m2_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vsuxei16_v_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_i32m8((int32_t *)(op0), (vuint16m4_t)(op1), (vint32m8_t)(op2), (size_t)(op3)) +#define vsuxei16_v_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_i32m8_m((vbool4_t)(op0), (int32_t *)(op1), (vuint16m4_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vsuxei16_v_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_i32mf2((int32_t *)(op0), (vuint16mf4_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vsuxei16_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint16mf4_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vnmsac_vx_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vx_i8m1((vint8m1_t)(op0), (int8_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vnmsac_vx_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vx_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (int8_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vnmsac_vx_i8m2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vx_i8m2((vint8m2_t)(op0), (int8_t)(op1), (vint8m2_t)(op2), (size_t)(op3)) +#define vnmsac_vx_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vx_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (int8_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vnmsac_vx_i8m4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vx_i8m4((vint8m4_t)(op0), (int8_t)(op1), (vint8m4_t)(op2), (size_t)(op3)) +#define vnmsac_vx_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vx_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (int8_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vnmsac_vx_i8m8(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vx_i8m8((vint8m8_t)(op0), (int8_t)(op1), (vint8m8_t)(op2), (size_t)(op3)) +#define vnmsac_vx_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vx_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (int8_t)(op2), (vint8m8_t)(op3), (size_t)(op4)) +#define vnmsac_vx_i8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vx_i8mf2((vint8mf2_t)(op0), (int8_t)(op1), (vint8mf2_t)(op2), (size_t)(op3)) +#define vnmsac_vx_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vx_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (int8_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vnmsac_vx_i8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vx_i8mf4((vint8mf4_t)(op0), (int8_t)(op1), (vint8mf4_t)(op2), (size_t)(op3)) +#define vnmsac_vx_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vx_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (int8_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vnmsac_vx_i8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vx_i8mf8((vint8mf8_t)(op0), (int8_t)(op1), (vint8mf8_t)(op2), (size_t)(op3)) +#define vnmsac_vx_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vx_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (int8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vnmsac_vx_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vx_i16m1((vint16m1_t)(op0), (int16_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vnmsac_vx_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vx_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (int16_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vnmsac_vx_i16m2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vx_i16m2((vint16m2_t)(op0), (int16_t)(op1), (vint16m2_t)(op2), (size_t)(op3)) +#define vnmsac_vx_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vx_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (int16_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vnmsac_vx_i16m4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vx_i16m4((vint16m4_t)(op0), (int16_t)(op1), (vint16m4_t)(op2), (size_t)(op3)) +#define vnmsac_vx_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vx_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (int16_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vnmsac_vx_i16m8(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vx_i16m8((vint16m8_t)(op0), (int16_t)(op1), (vint16m8_t)(op2), (size_t)(op3)) +#define vnmsac_vx_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vx_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (int16_t)(op2), (vint16m8_t)(op3), (size_t)(op4)) +#define vnmsac_vx_i16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vx_i16mf2((vint16mf2_t)(op0), (int16_t)(op1), (vint16mf2_t)(op2), (size_t)(op3)) +#define vnmsac_vx_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vx_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (int16_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vnmsac_vx_i16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vx_i16mf4((vint16mf4_t)(op0), (int16_t)(op1), (vint16mf4_t)(op2), (size_t)(op3)) +#define vnmsac_vx_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vx_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (int16_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vnmsac_vx_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vx_i32m1((vint32m1_t)(op0), (int32_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vnmsac_vx_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vx_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (int32_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vnmsac_vx_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vx_i32m2((vint32m2_t)(op0), (int32_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vnmsac_vx_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vx_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (int32_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vnmsac_vx_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vx_i32m4((vint32m4_t)(op0), (int32_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vnmsac_vx_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vx_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (int32_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vnmsac_vx_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vx_i32m8((vint32m8_t)(op0), (int32_t)(op1), (vint32m8_t)(op2), (size_t)(op3)) +#define vnmsac_vx_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vx_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (int32_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vnmsac_vx_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vx_i32mf2((vint32mf2_t)(op0), (int32_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vnmsac_vx_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vx_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (int32_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vnmsac_vx_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vx_i64m1((vint64m1_t)(op0), (int64_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vnmsac_vx_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vx_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (int64_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vnmsac_vx_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vx_i64m2((vint64m2_t)(op0), (int64_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vnmsac_vx_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vx_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (int64_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vnmsac_vx_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vx_i64m4((vint64m4_t)(op0), (int64_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vnmsac_vx_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vx_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (int64_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vnmsac_vx_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vx_i64m8((vint64m8_t)(op0), (int64_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vnmsac_vx_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vx_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (int64_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vnmsac_vv_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vv_u8m1((vuint8m1_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vnmsac_vv_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vv_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vnmsac_vv_u8m2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vv_u8m2((vuint8m2_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (size_t)(op3)) +#define vnmsac_vv_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vv_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vnmsac_vv_u8m4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vv_u8m4((vuint8m4_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (size_t)(op3)) +#define vnmsac_vv_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vv_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vnmsac_vv_u8m8(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vv_u8m8((vuint8m8_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (size_t)(op3)) +#define vnmsac_vv_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vv_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (vuint8m8_t)(op3), (size_t)(op4)) +#define vnmsac_vv_u8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vv_u8mf2((vuint8mf2_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (size_t)(op3)) +#define vnmsac_vv_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vv_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vnmsac_vv_u8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vv_u8mf4((vuint8mf4_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (size_t)(op3)) +#define vnmsac_vv_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vv_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vnmsac_vv_u8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vv_u8mf8((vuint8mf8_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (size_t)(op3)) +#define vnmsac_vv_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vv_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vnmsac_vv_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vv_u16m1((vuint16m1_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vnmsac_vv_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vv_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vnmsac_vv_u16m2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vv_u16m2((vuint16m2_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (size_t)(op3)) +#define vnmsac_vv_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vv_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vnmsac_vv_u16m4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vv_u16m4((vuint16m4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (size_t)(op3)) +#define vnmsac_vv_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vv_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vnmsac_vv_u16m8(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vv_u16m8((vuint16m8_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (size_t)(op3)) +#define vnmsac_vv_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vv_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vnmsac_vv_u16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vv_u16mf2((vuint16mf2_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (size_t)(op3)) +#define vnmsac_vv_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vv_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vnmsac_vv_u16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vv_u16mf4((vuint16mf4_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (size_t)(op3)) +#define vnmsac_vv_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vv_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vnmsac_vv_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vv_u32m1((vuint32m1_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vnmsac_vv_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vv_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vnmsac_vv_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vv_u32m2((vuint32m2_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vnmsac_vv_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vv_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vnmsac_vv_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vv_u32m4((vuint32m4_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vnmsac_vv_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vv_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vnmsac_vv_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vv_u32m8((vuint32m8_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (size_t)(op3)) +#define vnmsac_vv_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vv_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vnmsac_vv_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vv_u32mf2((vuint32mf2_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vnmsac_vv_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vv_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vnmsac_vv_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vv_u64m1((vuint64m1_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vnmsac_vv_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vv_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vnmsac_vv_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vv_u64m2((vuint64m2_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vnmsac_vv_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vv_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vnmsac_vv_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vv_u64m4((vuint64m4_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vnmsac_vv_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vv_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vnmsac_vv_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vv_u64m8((vuint64m8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vnmsac_vv_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vv_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vnmsac_vx_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vx_u8m1((vuint8m1_t)(op0), (uint8_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vnmsac_vx_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vx_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (uint8_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vnmsac_vx_u8m2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vx_u8m2((vuint8m2_t)(op0), (uint8_t)(op1), (vuint8m2_t)(op2), (size_t)(op3)) +#define vnmsac_vx_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vx_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (uint8_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vnmsac_vx_u8m4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vx_u8m4((vuint8m4_t)(op0), (uint8_t)(op1), (vuint8m4_t)(op2), (size_t)(op3)) +#define vnmsac_vx_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vx_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (uint8_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vnmsac_vx_u8m8(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vx_u8m8((vuint8m8_t)(op0), (uint8_t)(op1), (vuint8m8_t)(op2), (size_t)(op3)) +#define vnmsac_vx_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vx_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (uint8_t)(op2), (vuint8m8_t)(op3), (size_t)(op4)) +#define vnmsac_vx_u8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vx_u8mf2((vuint8mf2_t)(op0), (uint8_t)(op1), (vuint8mf2_t)(op2), (size_t)(op3)) +#define vnmsac_vx_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vx_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (uint8_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vnmsac_vx_u8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vx_u8mf4((vuint8mf4_t)(op0), (uint8_t)(op1), (vuint8mf4_t)(op2), (size_t)(op3)) +#define vnmsac_vx_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vx_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (uint8_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vnmsac_vx_u8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vx_u8mf8((vuint8mf8_t)(op0), (uint8_t)(op1), (vuint8mf8_t)(op2), (size_t)(op3)) +#define vnmsac_vx_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vx_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (uint8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vnmsac_vx_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vx_u16m1((vuint16m1_t)(op0), (uint16_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vnmsac_vx_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vx_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (uint16_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vnmsac_vx_u16m2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vx_u16m2((vuint16m2_t)(op0), (uint16_t)(op1), (vuint16m2_t)(op2), (size_t)(op3)) +#define vnmsac_vx_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vx_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (uint16_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vnmsac_vx_u16m4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vx_u16m4((vuint16m4_t)(op0), (uint16_t)(op1), (vuint16m4_t)(op2), (size_t)(op3)) +#define vnmsac_vx_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vx_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (uint16_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vnmsac_vx_u16m8(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vx_u16m8((vuint16m8_t)(op0), (uint16_t)(op1), (vuint16m8_t)(op2), (size_t)(op3)) +#define vnmsac_vx_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vx_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (uint16_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vnmsac_vx_u16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vx_u16mf2((vuint16mf2_t)(op0), (uint16_t)(op1), (vuint16mf2_t)(op2), (size_t)(op3)) +#define vnmsac_vx_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vx_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (uint16_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vnmsac_vx_u16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vx_u16mf4((vuint16mf4_t)(op0), (uint16_t)(op1), (vuint16mf4_t)(op2), (size_t)(op3)) +#define vnmsac_vx_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vx_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (uint16_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vnmsac_vx_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vx_u32m1((vuint32m1_t)(op0), (uint32_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vnmsac_vx_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vx_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (uint32_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vnmsac_vx_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vx_u32m2((vuint32m2_t)(op0), (uint32_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vnmsac_vx_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vx_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (uint32_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vnmsac_vx_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vx_u32m4((vuint32m4_t)(op0), (uint32_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vnmsac_vx_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vx_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (uint32_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vnmsac_vx_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vx_u32m8((vuint32m8_t)(op0), (uint32_t)(op1), (vuint32m8_t)(op2), (size_t)(op3)) +#define vnmsac_vx_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vx_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (uint32_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vnmsac_vx_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vx_u32mf2((vuint32mf2_t)(op0), (uint32_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vnmsac_vx_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vx_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (uint32_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vnmsac_vx_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vx_u64m1((vuint64m1_t)(op0), (uint64_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vnmsac_vx_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vx_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (uint64_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vnmsac_vx_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vx_u64m2((vuint64m2_t)(op0), (uint64_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vnmsac_vx_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vx_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (uint64_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vnmsac_vx_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vx_u64m4((vuint64m4_t)(op0), (uint64_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vnmsac_vx_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vx_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (uint64_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vnmsac_vx_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vnmsac_vx_u64m8((vuint64m8_t)(op0), (uint64_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vnmsac_vx_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsac_vx_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (uint64_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vmadd_vv_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vv_i8m1((vint8m1_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vmadd_vv_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vv_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vmadd_vv_i8m2(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vv_i8m2((vint8m2_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (size_t)(op3)) +#define vmadd_vv_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vv_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vmadd_vv_i8m4(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vv_i8m4((vint8m4_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (size_t)(op3)) +#define vmadd_vv_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vv_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vmadd_vv_i8m8(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vv_i8m8((vint8m8_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (size_t)(op3)) +#define vmadd_vv_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vv_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (vint8m8_t)(op3), (size_t)(op4)) +#define vmadd_vv_i8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vv_i8mf2((vint8mf2_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (size_t)(op3)) +#define vmadd_vv_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vv_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vmadd_vv_i8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vv_i8mf4((vint8mf4_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (size_t)(op3)) +#define vmadd_vv_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vv_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vmadd_vv_i8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vv_i8mf8((vint8mf8_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (size_t)(op3)) +#define vmadd_vv_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vv_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vmadd_vv_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vv_i16m1((vint16m1_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vmadd_vv_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vv_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vmadd_vv_i16m2(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vv_i16m2((vint16m2_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (size_t)(op3)) +#define vmadd_vv_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vv_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vmadd_vv_i16m4(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vv_i16m4((vint16m4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (size_t)(op3)) +#define vmadd_vv_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vv_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vmadd_vv_i16m8(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vv_i16m8((vint16m8_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (size_t)(op3)) +#define vmadd_vv_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vv_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (vint16m8_t)(op3), (size_t)(op4)) +#define vmadd_vv_i16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vv_i16mf2((vint16mf2_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (size_t)(op3)) +#define vmadd_vv_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vv_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vmadd_vv_i16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vv_i16mf4((vint16mf4_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (size_t)(op3)) +#define vmadd_vv_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vv_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vmadd_vv_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vv_i32m1((vint32m1_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vmadd_vv_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vv_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vmadd_vv_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vv_i32m2((vint32m2_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vmadd_vv_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vv_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vmadd_vv_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vv_i32m4((vint32m4_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vmadd_vv_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vv_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vmadd_vv_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vv_i32m8((vint32m8_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (size_t)(op3)) +#define vmadd_vv_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vv_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vmadd_vv_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vv_i32mf2((vint32mf2_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vmadd_vv_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vv_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vmadd_vv_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vv_i64m1((vint64m1_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vmadd_vv_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vv_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vmadd_vv_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vv_i64m2((vint64m2_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vmadd_vv_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vv_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vmadd_vv_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vv_i64m4((vint64m4_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vmadd_vv_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vv_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vmadd_vv_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vv_i64m8((vint64m8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vmadd_vv_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vv_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vmadd_vx_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vx_i8m1((vint8m1_t)(op0), (int8_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vmadd_vx_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vx_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (int8_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vmadd_vx_i8m2(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vx_i8m2((vint8m2_t)(op0), (int8_t)(op1), (vint8m2_t)(op2), (size_t)(op3)) +#define vmadd_vx_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vx_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (int8_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vmadd_vx_i8m4(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vx_i8m4((vint8m4_t)(op0), (int8_t)(op1), (vint8m4_t)(op2), (size_t)(op3)) +#define vmadd_vx_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vx_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (int8_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vmadd_vx_i8m8(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vx_i8m8((vint8m8_t)(op0), (int8_t)(op1), (vint8m8_t)(op2), (size_t)(op3)) +#define vmadd_vx_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vx_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (int8_t)(op2), (vint8m8_t)(op3), (size_t)(op4)) +#define vmadd_vx_i8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vx_i8mf2((vint8mf2_t)(op0), (int8_t)(op1), (vint8mf2_t)(op2), (size_t)(op3)) +#define vmadd_vx_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vx_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (int8_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vmadd_vx_i8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vx_i8mf4((vint8mf4_t)(op0), (int8_t)(op1), (vint8mf4_t)(op2), (size_t)(op3)) +#define vmadd_vx_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vx_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (int8_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vmadd_vx_i8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vx_i8mf8((vint8mf8_t)(op0), (int8_t)(op1), (vint8mf8_t)(op2), (size_t)(op3)) +#define vmadd_vx_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vx_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (int8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vmadd_vx_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vx_i16m1((vint16m1_t)(op0), (int16_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vmadd_vx_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vx_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (int16_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vmadd_vx_i16m2(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vx_i16m2((vint16m2_t)(op0), (int16_t)(op1), (vint16m2_t)(op2), (size_t)(op3)) +#define vmadd_vx_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vx_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (int16_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vmadd_vx_i16m4(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vx_i16m4((vint16m4_t)(op0), (int16_t)(op1), (vint16m4_t)(op2), (size_t)(op3)) +#define vmadd_vx_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vx_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (int16_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vmadd_vx_i16m8(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vx_i16m8((vint16m8_t)(op0), (int16_t)(op1), (vint16m8_t)(op2), (size_t)(op3)) +#define vmadd_vx_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vx_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (int16_t)(op2), (vint16m8_t)(op3), (size_t)(op4)) +#define vmadd_vx_i16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vx_i16mf2((vint16mf2_t)(op0), (int16_t)(op1), (vint16mf2_t)(op2), (size_t)(op3)) +#define vmadd_vx_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vx_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (int16_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vmadd_vx_i16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vx_i16mf4((vint16mf4_t)(op0), (int16_t)(op1), (vint16mf4_t)(op2), (size_t)(op3)) +#define vmadd_vx_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vx_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (int16_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vmadd_vx_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vx_i32m1((vint32m1_t)(op0), (int32_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vmadd_vx_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vx_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (int32_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vmadd_vx_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vx_i32m2((vint32m2_t)(op0), (int32_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vmadd_vx_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vx_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (int32_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vmadd_vx_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vx_i32m4((vint32m4_t)(op0), (int32_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vmadd_vx_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vx_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (int32_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vmadd_vx_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vx_i32m8((vint32m8_t)(op0), (int32_t)(op1), (vint32m8_t)(op2), (size_t)(op3)) +#define vmadd_vx_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vx_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (int32_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vmadd_vx_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vx_i32mf2((vint32mf2_t)(op0), (int32_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vmadd_vx_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vx_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (int32_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vmadd_vx_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vx_i64m1((vint64m1_t)(op0), (int64_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vmadd_vx_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vx_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (int64_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vmadd_vx_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vx_i64m2((vint64m2_t)(op0), (int64_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vmadd_vx_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vx_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (int64_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vmadd_vx_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vx_i64m4((vint64m4_t)(op0), (int64_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vmadd_vx_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vx_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (int64_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vmadd_vx_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vx_i64m8((vint64m8_t)(op0), (int64_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vmadd_vx_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vx_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (int64_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vmadd_vv_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vv_u8m1((vuint8m1_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vmadd_vv_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vv_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vmadd_vv_u8m2(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vv_u8m2((vuint8m2_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (size_t)(op3)) +#define vmadd_vv_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vv_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vmadd_vv_u8m4(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vv_u8m4((vuint8m4_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (size_t)(op3)) +#define vmadd_vv_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vv_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vmadd_vv_u8m8(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vv_u8m8((vuint8m8_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (size_t)(op3)) +#define vmadd_vv_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vv_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (vuint8m8_t)(op3), (size_t)(op4)) +#define vmadd_vv_u8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vv_u8mf2((vuint8mf2_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (size_t)(op3)) +#define vmadd_vv_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vv_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vmadd_vv_u8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vv_u8mf4((vuint8mf4_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (size_t)(op3)) +#define vmadd_vv_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vv_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vmadd_vv_u8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vv_u8mf8((vuint8mf8_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (size_t)(op3)) +#define vmadd_vv_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vv_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vmadd_vv_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vv_u16m1((vuint16m1_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vmadd_vv_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vv_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vmadd_vv_u16m2(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vv_u16m2((vuint16m2_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (size_t)(op3)) +#define vmadd_vv_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vv_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vmadd_vv_u16m4(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vv_u16m4((vuint16m4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (size_t)(op3)) +#define vmadd_vv_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vv_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vmadd_vv_u16m8(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vv_u16m8((vuint16m8_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (size_t)(op3)) +#define vmadd_vv_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vv_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vmadd_vv_u16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vv_u16mf2((vuint16mf2_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (size_t)(op3)) +#define vmadd_vv_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vv_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vmadd_vv_u16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vv_u16mf4((vuint16mf4_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (size_t)(op3)) +#define vmadd_vv_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vv_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vmadd_vv_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vv_u32m1((vuint32m1_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vmadd_vv_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vv_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vmadd_vv_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vv_u32m2((vuint32m2_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vmadd_vv_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vv_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vmadd_vv_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vv_u32m4((vuint32m4_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vmadd_vv_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vv_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vmadd_vv_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vv_u32m8((vuint32m8_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (size_t)(op3)) +#define vmadd_vv_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vv_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vmadd_vv_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vv_u32mf2((vuint32mf2_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vmadd_vv_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vv_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vmadd_vv_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vv_u64m1((vuint64m1_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vmadd_vv_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vv_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vmadd_vv_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vv_u64m2((vuint64m2_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vmadd_vv_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vv_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vmadd_vv_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vv_u64m4((vuint64m4_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vmadd_vv_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vv_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vmadd_vv_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vv_u64m8((vuint64m8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vmadd_vv_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vv_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vmadd_vx_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vx_u8m1((vuint8m1_t)(op0), (uint8_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vmadd_vx_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vx_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (uint8_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vmadd_vx_u8m2(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vx_u8m2((vuint8m2_t)(op0), (uint8_t)(op1), (vuint8m2_t)(op2), (size_t)(op3)) +#define vmadd_vx_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vx_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (uint8_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vmadd_vx_u8m4(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vx_u8m4((vuint8m4_t)(op0), (uint8_t)(op1), (vuint8m4_t)(op2), (size_t)(op3)) +#define vmadd_vx_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vx_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (uint8_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vmadd_vx_u8m8(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vx_u8m8((vuint8m8_t)(op0), (uint8_t)(op1), (vuint8m8_t)(op2), (size_t)(op3)) +#define vmadd_vx_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vx_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (uint8_t)(op2), (vuint8m8_t)(op3), (size_t)(op4)) +#define vmadd_vx_u8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vx_u8mf2((vuint8mf2_t)(op0), (uint8_t)(op1), (vuint8mf2_t)(op2), (size_t)(op3)) +#define vmadd_vx_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vx_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (uint8_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vmadd_vx_u8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vx_u8mf4((vuint8mf4_t)(op0), (uint8_t)(op1), (vuint8mf4_t)(op2), (size_t)(op3)) +#define vmadd_vx_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vx_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (uint8_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vmadd_vx_u8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vx_u8mf8((vuint8mf8_t)(op0), (uint8_t)(op1), (vuint8mf8_t)(op2), (size_t)(op3)) +#define vmadd_vx_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vx_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (uint8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vmadd_vx_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vx_u16m1((vuint16m1_t)(op0), (uint16_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vmadd_vx_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vx_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (uint16_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vmadd_vx_u16m2(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vx_u16m2((vuint16m2_t)(op0), (uint16_t)(op1), (vuint16m2_t)(op2), (size_t)(op3)) +#define vmadd_vx_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vx_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (uint16_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vmadd_vx_u16m4(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vx_u16m4((vuint16m4_t)(op0), (uint16_t)(op1), (vuint16m4_t)(op2), (size_t)(op3)) +#define vmadd_vx_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vx_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (uint16_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vmadd_vx_u16m8(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vx_u16m8((vuint16m8_t)(op0), (uint16_t)(op1), (vuint16m8_t)(op2), (size_t)(op3)) +#define vmadd_vx_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vx_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (uint16_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vmadd_vx_u16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vx_u16mf2((vuint16mf2_t)(op0), (uint16_t)(op1), (vuint16mf2_t)(op2), (size_t)(op3)) +#define vmadd_vx_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vx_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (uint16_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vmadd_vx_u16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vx_u16mf4((vuint16mf4_t)(op0), (uint16_t)(op1), (vuint16mf4_t)(op2), (size_t)(op3)) +#define vmadd_vx_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vx_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (uint16_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vmadd_vx_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vx_u32m1((vuint32m1_t)(op0), (uint32_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vmadd_vx_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vx_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (uint32_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vmadd_vx_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vx_u32m2((vuint32m2_t)(op0), (uint32_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vmadd_vx_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vx_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (uint32_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vmadd_vx_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vx_u32m4((vuint32m4_t)(op0), (uint32_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vmadd_vx_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vx_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (uint32_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vmadd_vx_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vx_u32m8((vuint32m8_t)(op0), (uint32_t)(op1), (vuint32m8_t)(op2), (size_t)(op3)) +#define vmadd_vx_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vx_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (uint32_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vmadd_vx_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vx_u32mf2((vuint32mf2_t)(op0), (uint32_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vmadd_vx_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vx_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (uint32_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vmadd_vx_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vx_u64m1((vuint64m1_t)(op0), (uint64_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vmadd_vx_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vx_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (uint64_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vmadd_vx_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vx_u64m2((vuint64m2_t)(op0), (uint64_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vmadd_vx_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vx_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (uint64_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vmadd_vx_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vx_u64m4((vuint64m4_t)(op0), (uint64_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vmadd_vx_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vx_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (uint64_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vmadd_vx_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vmadd_vx_u64m8((vuint64m8_t)(op0), (uint64_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vmadd_vx_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmadd_vx_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (uint64_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vnmsub_vv_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vv_i8m1((vint8m1_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vnmsub_vv_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vv_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vnmsub_vv_i8m2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vv_i8m2((vint8m2_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (size_t)(op3)) +#define vnmsub_vv_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vv_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vnmsub_vv_i8m4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vv_i8m4((vint8m4_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (size_t)(op3)) +#define vnmsub_vv_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vv_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vnmsub_vv_i8m8(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vv_i8m8((vint8m8_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (size_t)(op3)) +#define vnmsub_vv_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vv_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (vint8m8_t)(op3), (size_t)(op4)) +#define vnmsub_vv_i8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vv_i8mf2((vint8mf2_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (size_t)(op3)) +#define vnmsub_vv_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vv_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vnmsub_vv_i8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vv_i8mf4((vint8mf4_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (size_t)(op3)) +#define vnmsub_vv_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vv_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vnmsub_vv_i8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vv_i8mf8((vint8mf8_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (size_t)(op3)) +#define vnmsub_vv_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vv_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vnmsub_vv_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vv_i16m1((vint16m1_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vnmsub_vv_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vv_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vnmsub_vv_i16m2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vv_i16m2((vint16m2_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (size_t)(op3)) +#define vnmsub_vv_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vv_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vnmsub_vv_i16m4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vv_i16m4((vint16m4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (size_t)(op3)) +#define vnmsub_vv_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vv_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vnmsub_vv_i16m8(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vv_i16m8((vint16m8_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (size_t)(op3)) +#define vnmsub_vv_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vv_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (vint16m8_t)(op3), (size_t)(op4)) +#define vnmsub_vv_i16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vv_i16mf2((vint16mf2_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (size_t)(op3)) +#define vnmsub_vv_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vv_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vnmsub_vv_i16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vv_i16mf4((vint16mf4_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (size_t)(op3)) +#define vnmsub_vv_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vv_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vnmsub_vv_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vv_i32m1((vint32m1_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vnmsub_vv_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vv_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vnmsub_vv_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vv_i32m2((vint32m2_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vnmsub_vv_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vv_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vnmsub_vv_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vv_i32m4((vint32m4_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vnmsub_vv_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vv_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vnmsub_vv_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vv_i32m8((vint32m8_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (size_t)(op3)) +#define vnmsub_vv_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vv_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vnmsub_vv_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vv_i32mf2((vint32mf2_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vnmsub_vv_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vv_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vnmsub_vv_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vv_i64m1((vint64m1_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vnmsub_vv_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vv_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vnmsub_vv_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vv_i64m2((vint64m2_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vnmsub_vv_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vv_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vnmsub_vv_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vv_i64m4((vint64m4_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vnmsub_vv_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vv_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vnmsub_vv_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vv_i64m8((vint64m8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vnmsub_vv_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vv_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vnmsub_vx_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vx_i8m1((vint8m1_t)(op0), (int8_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vnmsub_vx_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vx_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (int8_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vnmsub_vx_i8m2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vx_i8m2((vint8m2_t)(op0), (int8_t)(op1), (vint8m2_t)(op2), (size_t)(op3)) +#define vnmsub_vx_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vx_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (int8_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vnmsub_vx_i8m4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vx_i8m4((vint8m4_t)(op0), (int8_t)(op1), (vint8m4_t)(op2), (size_t)(op3)) +#define vnmsub_vx_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vx_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (int8_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vnmsub_vx_i8m8(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vx_i8m8((vint8m8_t)(op0), (int8_t)(op1), (vint8m8_t)(op2), (size_t)(op3)) +#define vnmsub_vx_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vx_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (int8_t)(op2), (vint8m8_t)(op3), (size_t)(op4)) +#define vnmsub_vx_i8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vx_i8mf2((vint8mf2_t)(op0), (int8_t)(op1), (vint8mf2_t)(op2), (size_t)(op3)) +#define vnmsub_vx_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vx_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (int8_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vnmsub_vx_i8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vx_i8mf4((vint8mf4_t)(op0), (int8_t)(op1), (vint8mf4_t)(op2), (size_t)(op3)) +#define vnmsub_vx_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vx_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (int8_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vnmsub_vx_i8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vx_i8mf8((vint8mf8_t)(op0), (int8_t)(op1), (vint8mf8_t)(op2), (size_t)(op3)) +#define vnmsub_vx_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vx_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (int8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vnmsub_vx_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vx_i16m1((vint16m1_t)(op0), (int16_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vnmsub_vx_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vx_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (int16_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vnmsub_vx_i16m2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vx_i16m2((vint16m2_t)(op0), (int16_t)(op1), (vint16m2_t)(op2), (size_t)(op3)) +#define vnmsub_vx_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vx_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (int16_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vnmsub_vx_i16m4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vx_i16m4((vint16m4_t)(op0), (int16_t)(op1), (vint16m4_t)(op2), (size_t)(op3)) +#define vnmsub_vx_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vx_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (int16_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vnmsub_vx_i16m8(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vx_i16m8((vint16m8_t)(op0), (int16_t)(op1), (vint16m8_t)(op2), (size_t)(op3)) +#define vnmsub_vx_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vx_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (int16_t)(op2), (vint16m8_t)(op3), (size_t)(op4)) +#define vnmsub_vx_i16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vx_i16mf2((vint16mf2_t)(op0), (int16_t)(op1), (vint16mf2_t)(op2), (size_t)(op3)) +#define vnmsub_vx_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vx_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (int16_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vnmsub_vx_i16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vx_i16mf4((vint16mf4_t)(op0), (int16_t)(op1), (vint16mf4_t)(op2), (size_t)(op3)) +#define vnmsub_vx_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vx_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (int16_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vnmsub_vx_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vx_i32m1((vint32m1_t)(op0), (int32_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vnmsub_vx_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vx_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (int32_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vnmsub_vx_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vx_i32m2((vint32m2_t)(op0), (int32_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vnmsub_vx_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vx_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (int32_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vnmsub_vx_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vx_i32m4((vint32m4_t)(op0), (int32_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vnmsub_vx_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vx_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (int32_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vnmsub_vx_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vx_i32m8((vint32m8_t)(op0), (int32_t)(op1), (vint32m8_t)(op2), (size_t)(op3)) +#define vnmsub_vx_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vx_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (int32_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vnmsub_vx_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vx_i32mf2((vint32mf2_t)(op0), (int32_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vnmsub_vx_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vx_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (int32_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vnmsub_vx_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vx_i64m1((vint64m1_t)(op0), (int64_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vnmsub_vx_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vx_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (int64_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vnmsub_vx_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vx_i64m2((vint64m2_t)(op0), (int64_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vnmsub_vx_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vx_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (int64_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vnmsub_vx_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vx_i64m4((vint64m4_t)(op0), (int64_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vnmsub_vx_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vx_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (int64_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vnmsub_vx_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vx_i64m8((vint64m8_t)(op0), (int64_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vnmsub_vx_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vx_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (int64_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vnmsub_vv_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vv_u8m1((vuint8m1_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vnmsub_vv_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vv_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vnmsub_vv_u8m2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vv_u8m2((vuint8m2_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (size_t)(op3)) +#define vnmsub_vv_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vv_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vnmsub_vv_u8m4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vv_u8m4((vuint8m4_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (size_t)(op3)) +#define vnmsub_vv_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vv_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vnmsub_vv_u8m8(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vv_u8m8((vuint8m8_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (size_t)(op3)) +#define vnmsub_vv_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vv_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (vuint8m8_t)(op3), (size_t)(op4)) +#define vnmsub_vv_u8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vv_u8mf2((vuint8mf2_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (size_t)(op3)) +#define vnmsub_vv_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vv_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vnmsub_vv_u8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vv_u8mf4((vuint8mf4_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (size_t)(op3)) +#define vnmsub_vv_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vv_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vnmsub_vv_u8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vv_u8mf8((vuint8mf8_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (size_t)(op3)) +#define vnmsub_vv_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vv_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vnmsub_vv_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vv_u16m1((vuint16m1_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vnmsub_vv_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vv_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vnmsub_vv_u16m2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vv_u16m2((vuint16m2_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (size_t)(op3)) +#define vnmsub_vv_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vv_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vnmsub_vv_u16m4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vv_u16m4((vuint16m4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (size_t)(op3)) +#define vnmsub_vv_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vv_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vnmsub_vv_u16m8(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vv_u16m8((vuint16m8_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (size_t)(op3)) +#define vnmsub_vv_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vv_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vnmsub_vv_u16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vv_u16mf2((vuint16mf2_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (size_t)(op3)) +#define vnmsub_vv_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vv_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vnmsub_vv_u16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vv_u16mf4((vuint16mf4_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (size_t)(op3)) +#define vnmsub_vv_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vv_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vnmsub_vv_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vv_u32m1((vuint32m1_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vnmsub_vv_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vv_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vnmsub_vv_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vv_u32m2((vuint32m2_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vnmsub_vv_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vv_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vnmsub_vv_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vv_u32m4((vuint32m4_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vnmsub_vv_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vv_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vnmsub_vv_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vv_u32m8((vuint32m8_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (size_t)(op3)) +#define vnmsub_vv_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vv_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vnmsub_vv_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vv_u32mf2((vuint32mf2_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vnmsub_vv_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vv_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vnmsub_vv_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vv_u64m1((vuint64m1_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vnmsub_vv_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vv_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vnmsub_vv_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vv_u64m2((vuint64m2_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vnmsub_vv_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vv_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vnmsub_vv_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vv_u64m4((vuint64m4_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vnmsub_vv_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vv_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vnmsub_vv_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vv_u64m8((vuint64m8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vnmsub_vv_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vv_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vsuxei16_v_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_u32m1((uint32_t *)(op0), (vuint16mf2_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vsuxei16_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint16mf2_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vsuxei16_v_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_u32m2((uint32_t *)(op0), (vuint16m1_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vsuxei16_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint16m1_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vsuxei16_v_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_u32m4((uint32_t *)(op0), (vuint16m2_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vsuxei16_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint16m2_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vsuxei16_v_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_u32m8((uint32_t *)(op0), (vuint16m4_t)(op1), (vuint32m8_t)(op2), (size_t)(op3)) +#define vsuxei16_v_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_u32m8_m((vbool4_t)(op0), (uint32_t *)(op1), (vuint16m4_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vsuxei16_v_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_u32mf2((uint32_t *)(op0), (vuint16mf4_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vsuxei16_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint16mf4_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vnmsub_vx_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vx_u8m1((vuint8m1_t)(op0), (uint8_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vnmsub_vx_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vx_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (uint8_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vnmsub_vx_u8m2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vx_u8m2((vuint8m2_t)(op0), (uint8_t)(op1), (vuint8m2_t)(op2), (size_t)(op3)) +#define vnmsub_vx_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vx_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (uint8_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vnmsub_vx_u8m4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vx_u8m4((vuint8m4_t)(op0), (uint8_t)(op1), (vuint8m4_t)(op2), (size_t)(op3)) +#define vnmsub_vx_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vx_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (uint8_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vnmsub_vx_u8m8(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vx_u8m8((vuint8m8_t)(op0), (uint8_t)(op1), (vuint8m8_t)(op2), (size_t)(op3)) +#define vnmsub_vx_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vx_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (uint8_t)(op2), (vuint8m8_t)(op3), (size_t)(op4)) +#define vnmsub_vx_u8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vx_u8mf2((vuint8mf2_t)(op0), (uint8_t)(op1), (vuint8mf2_t)(op2), (size_t)(op3)) +#define vnmsub_vx_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vx_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (uint8_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vnmsub_vx_u8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vx_u8mf4((vuint8mf4_t)(op0), (uint8_t)(op1), (vuint8mf4_t)(op2), (size_t)(op3)) +#define vnmsub_vx_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vx_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (uint8_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vnmsub_vx_u8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vx_u8mf8((vuint8mf8_t)(op0), (uint8_t)(op1), (vuint8mf8_t)(op2), (size_t)(op3)) +#define vnmsub_vx_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vx_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (uint8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vnmsub_vx_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vx_u16m1((vuint16m1_t)(op0), (uint16_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vnmsub_vx_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vx_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (uint16_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vnmsub_vx_u16m2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vx_u16m2((vuint16m2_t)(op0), (uint16_t)(op1), (vuint16m2_t)(op2), (size_t)(op3)) +#define vnmsub_vx_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vx_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (uint16_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vnmsub_vx_u16m4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vx_u16m4((vuint16m4_t)(op0), (uint16_t)(op1), (vuint16m4_t)(op2), (size_t)(op3)) +#define vnmsub_vx_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vx_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (uint16_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vnmsub_vx_u16m8(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vx_u16m8((vuint16m8_t)(op0), (uint16_t)(op1), (vuint16m8_t)(op2), (size_t)(op3)) +#define vnmsub_vx_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vx_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (uint16_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vnmsub_vx_u16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vx_u16mf2((vuint16mf2_t)(op0), (uint16_t)(op1), (vuint16mf2_t)(op2), (size_t)(op3)) +#define vnmsub_vx_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vx_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (uint16_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vnmsub_vx_u16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vx_u16mf4((vuint16mf4_t)(op0), (uint16_t)(op1), (vuint16mf4_t)(op2), (size_t)(op3)) +#define vnmsub_vx_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vx_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (uint16_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vnmsub_vx_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vx_u32m1((vuint32m1_t)(op0), (uint32_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vnmsub_vx_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vx_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (uint32_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vnmsub_vx_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vx_u32m2((vuint32m2_t)(op0), (uint32_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vnmsub_vx_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vx_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (uint32_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vnmsub_vx_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vx_u32m4((vuint32m4_t)(op0), (uint32_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vnmsub_vx_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vx_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (uint32_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vnmsub_vx_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vx_u32m8((vuint32m8_t)(op0), (uint32_t)(op1), (vuint32m8_t)(op2), (size_t)(op3)) +#define vnmsub_vx_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vx_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (uint32_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vnmsub_vx_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vx_u32mf2((vuint32mf2_t)(op0), (uint32_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vnmsub_vx_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vx_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (uint32_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vnmsub_vx_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vx_u64m1((vuint64m1_t)(op0), (uint64_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vnmsub_vx_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vx_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (uint64_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vnmsub_vx_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vx_u64m2((vuint64m2_t)(op0), (uint64_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vnmsub_vx_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vx_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (uint64_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vnmsub_vx_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vx_u64m4((vuint64m4_t)(op0), (uint64_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vnmsub_vx_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vx_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (uint64_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vnmsub_vx_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vnmsub_vx_u64m8((vuint64m8_t)(op0), (uint64_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vnmsub_vx_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnmsub_vx_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (uint64_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vwmaccu_vv_u16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccu_vv_u16mf4((vuint16mf4_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (size_t)(op3)) +#define vwmaccu_vv_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccu_vv_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vwmaccu_vv_u16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccu_vv_u16mf2((vuint16mf2_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (size_t)(op3)) +#define vwmaccu_vv_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccu_vv_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vwmaccu_vv_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccu_vv_u16m1((vuint16m1_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (size_t)(op3)) +#define vwmaccu_vv_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccu_vv_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vwmaccu_vv_u16m2(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccu_vv_u16m2((vuint16m2_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vwmaccu_vv_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccu_vv_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vwmaccu_vv_u16m4(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccu_vv_u16m4((vuint16m4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (size_t)(op3)) +#define vwmaccu_vv_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccu_vv_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vwmaccu_vv_u16m8(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccu_vv_u16m8((vuint16m8_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (size_t)(op3)) +#define vwmaccu_vv_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccu_vv_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vwmaccu_vv_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccu_vv_u32mf2((vuint32mf2_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (size_t)(op3)) +#define vwmaccu_vv_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccu_vv_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vwmaccu_vv_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccu_vv_u32m1((vuint32m1_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (size_t)(op3)) +#define vwmaccu_vv_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccu_vv_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vwmaccu_vv_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccu_vv_u32m2((vuint32m2_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vwmaccu_vv_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccu_vv_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vwmaccu_vv_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccu_vv_u32m4((vuint32m4_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (size_t)(op3)) +#define vwmaccu_vv_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccu_vv_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vwmaccu_vv_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccu_vv_u32m8((vuint32m8_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (size_t)(op3)) +#define vwmaccu_vv_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccu_vv_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vwmaccu_vv_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccu_vv_u64m1((vuint64m1_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vwmaccu_vv_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccu_vv_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vwmaccu_vv_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccu_vv_u64m2((vuint64m2_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vwmaccu_vv_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccu_vv_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vwmaccu_vv_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccu_vv_u64m4((vuint64m4_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vwmaccu_vv_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccu_vv_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vwmaccu_vv_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccu_vv_u64m8((vuint64m8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vwmaccu_vv_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccu_vv_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vwmaccu_vx_u16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccu_vx_u16mf4((vuint16mf4_t)(op0), (uint8_t)(op1), (vuint8mf8_t)(op2), (size_t)(op3)) +#define vwmaccu_vx_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccu_vx_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (uint8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vwmaccu_vx_u16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccu_vx_u16mf2((vuint16mf2_t)(op0), (uint8_t)(op1), (vuint8mf4_t)(op2), (size_t)(op3)) +#define vwmaccu_vx_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccu_vx_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (uint8_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vwmaccu_vx_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccu_vx_u16m1((vuint16m1_t)(op0), (uint8_t)(op1), (vuint8mf2_t)(op2), (size_t)(op3)) +#define vwmaccu_vx_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccu_vx_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (uint8_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vwmaccu_vx_u16m2(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccu_vx_u16m2((vuint16m2_t)(op0), (uint8_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vwmaccu_vx_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccu_vx_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (uint8_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vwmaccu_vx_u16m4(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccu_vx_u16m4((vuint16m4_t)(op0), (uint8_t)(op1), (vuint8m2_t)(op2), (size_t)(op3)) +#define vwmaccu_vx_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccu_vx_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (uint8_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vwmaccu_vx_u16m8(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccu_vx_u16m8((vuint16m8_t)(op0), (uint8_t)(op1), (vuint8m4_t)(op2), (size_t)(op3)) +#define vwmaccu_vx_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccu_vx_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (uint8_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vwmaccu_vx_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccu_vx_u32mf2((vuint32mf2_t)(op0), (uint16_t)(op1), (vuint16mf4_t)(op2), (size_t)(op3)) +#define vwmaccu_vx_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccu_vx_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (uint16_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vwmaccu_vx_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccu_vx_u32m1((vuint32m1_t)(op0), (uint16_t)(op1), (vuint16mf2_t)(op2), (size_t)(op3)) +#define vwmaccu_vx_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccu_vx_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (uint16_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vwmaccu_vx_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccu_vx_u32m2((vuint32m2_t)(op0), (uint16_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vwmaccu_vx_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccu_vx_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (uint16_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vwmaccu_vx_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccu_vx_u32m4((vuint32m4_t)(op0), (uint16_t)(op1), (vuint16m2_t)(op2), (size_t)(op3)) +#define vwmaccu_vx_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccu_vx_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (uint16_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vwmaccu_vx_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccu_vx_u32m8((vuint32m8_t)(op0), (uint16_t)(op1), (vuint16m4_t)(op2), (size_t)(op3)) +#define vwmaccu_vx_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccu_vx_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (uint16_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vwmaccu_vx_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccu_vx_u64m1((vuint64m1_t)(op0), (uint32_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vwmaccu_vx_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccu_vx_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (uint32_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vwmaccu_vx_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccu_vx_u64m2((vuint64m2_t)(op0), (uint32_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vwmaccu_vx_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccu_vx_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (uint32_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vwmaccu_vx_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccu_vx_u64m4((vuint64m4_t)(op0), (uint32_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vwmaccu_vx_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccu_vx_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (uint32_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vwmaccu_vx_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccu_vx_u64m8((vuint64m8_t)(op0), (uint32_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vwmaccu_vx_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccu_vx_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (uint32_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vwmacc_vv_i16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vwmacc_vv_i16mf4((vint16mf4_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (size_t)(op3)) +#define vwmacc_vv_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmacc_vv_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vwmacc_vv_i16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vwmacc_vv_i16mf2((vint16mf2_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (size_t)(op3)) +#define vwmacc_vv_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmacc_vv_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vwmacc_vv_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vwmacc_vv_i16m1((vint16m1_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (size_t)(op3)) +#define vwmacc_vv_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmacc_vv_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vwmacc_vv_i16m2(op0, op1, op2, op3) \ +__builtin_rvv_vwmacc_vv_i16m2((vint16m2_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vwmacc_vv_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmacc_vv_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vwmacc_vv_i16m4(op0, op1, op2, op3) \ +__builtin_rvv_vwmacc_vv_i16m4((vint16m4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (size_t)(op3)) +#define vwmacc_vv_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmacc_vv_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vwmacc_vv_i16m8(op0, op1, op2, op3) \ +__builtin_rvv_vwmacc_vv_i16m8((vint16m8_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (size_t)(op3)) +#define vwmacc_vv_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmacc_vv_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint8m4_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vwmacc_vv_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vwmacc_vv_i32mf2((vint32mf2_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (size_t)(op3)) +#define vwmacc_vv_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmacc_vv_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vwmacc_vv_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vwmacc_vv_i32m1((vint32m1_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (size_t)(op3)) +#define vwmacc_vv_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmacc_vv_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vwmacc_vv_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vwmacc_vv_i32m2((vint32m2_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vwmacc_vv_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmacc_vv_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vwmacc_vv_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vwmacc_vv_i32m4((vint32m4_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (size_t)(op3)) +#define vwmacc_vv_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmacc_vv_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vwmacc_vv_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vwmacc_vv_i32m8((vint32m8_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (size_t)(op3)) +#define vwmacc_vv_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmacc_vv_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint16m4_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vwmacc_vv_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vwmacc_vv_i64m1((vint64m1_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vwmacc_vv_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmacc_vv_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vwmacc_vv_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vwmacc_vv_i64m2((vint64m2_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vwmacc_vv_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmacc_vv_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vwmacc_vv_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vwmacc_vv_i64m4((vint64m4_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vwmacc_vv_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmacc_vv_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vwmacc_vv_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vwmacc_vv_i64m8((vint64m8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vwmacc_vv_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmacc_vv_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vwmacc_vx_i16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vwmacc_vx_i16mf4((vint16mf4_t)(op0), (int8_t)(op1), (vint8mf8_t)(op2), (size_t)(op3)) +#define vwmacc_vx_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmacc_vx_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (int8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vwmacc_vx_i16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vwmacc_vx_i16mf2((vint16mf2_t)(op0), (int8_t)(op1), (vint8mf4_t)(op2), (size_t)(op3)) +#define vwmacc_vx_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmacc_vx_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (int8_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vwmacc_vx_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vwmacc_vx_i16m1((vint16m1_t)(op0), (int8_t)(op1), (vint8mf2_t)(op2), (size_t)(op3)) +#define vwmacc_vx_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmacc_vx_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (int8_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vwmacc_vx_i16m2(op0, op1, op2, op3) \ +__builtin_rvv_vwmacc_vx_i16m2((vint16m2_t)(op0), (int8_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vwmacc_vx_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmacc_vx_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (int8_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vwmacc_vx_i16m4(op0, op1, op2, op3) \ +__builtin_rvv_vwmacc_vx_i16m4((vint16m4_t)(op0), (int8_t)(op1), (vint8m2_t)(op2), (size_t)(op3)) +#define vwmacc_vx_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmacc_vx_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (int8_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vwmacc_vx_i16m8(op0, op1, op2, op3) \ +__builtin_rvv_vwmacc_vx_i16m8((vint16m8_t)(op0), (int8_t)(op1), (vint8m4_t)(op2), (size_t)(op3)) +#define vwmacc_vx_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmacc_vx_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (int8_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vwmacc_vx_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vwmacc_vx_i32mf2((vint32mf2_t)(op0), (int16_t)(op1), (vint16mf4_t)(op2), (size_t)(op3)) +#define vwmacc_vx_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmacc_vx_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (int16_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vwmacc_vx_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vwmacc_vx_i32m1((vint32m1_t)(op0), (int16_t)(op1), (vint16mf2_t)(op2), (size_t)(op3)) +#define vwmacc_vx_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmacc_vx_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (int16_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vwmacc_vx_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vwmacc_vx_i32m2((vint32m2_t)(op0), (int16_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vwmacc_vx_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmacc_vx_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (int16_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vwmacc_vx_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vwmacc_vx_i32m4((vint32m4_t)(op0), (int16_t)(op1), (vint16m2_t)(op2), (size_t)(op3)) +#define vwmacc_vx_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmacc_vx_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (int16_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vwmacc_vx_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vwmacc_vx_i32m8((vint32m8_t)(op0), (int16_t)(op1), (vint16m4_t)(op2), (size_t)(op3)) +#define vwmacc_vx_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmacc_vx_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (int16_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vwmacc_vx_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vwmacc_vx_i64m1((vint64m1_t)(op0), (int32_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vwmacc_vx_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmacc_vx_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (int32_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vwmacc_vx_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vwmacc_vx_i64m2((vint64m2_t)(op0), (int32_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vwmacc_vx_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmacc_vx_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (int32_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vwmacc_vx_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vwmacc_vx_i64m4((vint64m4_t)(op0), (int32_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vwmacc_vx_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmacc_vx_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (int32_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vwmacc_vx_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vwmacc_vx_i64m8((vint64m8_t)(op0), (int32_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vwmacc_vx_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmacc_vx_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (int32_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vwmaccsu_vv_i16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccsu_vv_i16mf4((vint16mf4_t)(op0), (vint8mf8_t)(op1), (vuint8mf8_t)(op2), (size_t)(op3)) +#define vwmaccsu_vv_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccsu_vv_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vwmaccsu_vv_i16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccsu_vv_i16mf2((vint16mf2_t)(op0), (vint8mf4_t)(op1), (vuint8mf4_t)(op2), (size_t)(op3)) +#define vwmaccsu_vv_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccsu_vv_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vwmaccsu_vv_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccsu_vv_i16m1((vint16m1_t)(op0), (vint8mf2_t)(op1), (vuint8mf2_t)(op2), (size_t)(op3)) +#define vwmaccsu_vv_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccsu_vv_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vwmaccsu_vv_i16m2(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccsu_vv_i16m2((vint16m2_t)(op0), (vint8m1_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vwmaccsu_vv_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccsu_vv_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vwmaccsu_vv_i16m4(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccsu_vv_i16m4((vint16m4_t)(op0), (vint8m2_t)(op1), (vuint8m2_t)(op2), (size_t)(op3)) +#define vwmaccsu_vv_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccsu_vv_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vwmaccsu_vv_i16m8(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccsu_vv_i16m8((vint16m8_t)(op0), (vint8m4_t)(op1), (vuint8m4_t)(op2), (size_t)(op3)) +#define vwmaccsu_vv_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccsu_vv_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vwmaccsu_vv_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccsu_vv_i32mf2((vint32mf2_t)(op0), (vint16mf4_t)(op1), (vuint16mf4_t)(op2), (size_t)(op3)) +#define vwmaccsu_vv_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccsu_vv_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vwmaccsu_vv_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccsu_vv_i32m1((vint32m1_t)(op0), (vint16mf2_t)(op1), (vuint16mf2_t)(op2), (size_t)(op3)) +#define vwmaccsu_vv_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccsu_vv_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vwmaccsu_vv_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccsu_vv_i32m2((vint32m2_t)(op0), (vint16m1_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vwmaccsu_vv_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccsu_vv_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vwmaccsu_vv_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccsu_vv_i32m4((vint32m4_t)(op0), (vint16m2_t)(op1), (vuint16m2_t)(op2), (size_t)(op3)) +#define vwmaccsu_vv_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccsu_vv_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vwmaccsu_vv_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccsu_vv_i32m8((vint32m8_t)(op0), (vint16m4_t)(op1), (vuint16m4_t)(op2), (size_t)(op3)) +#define vwmaccsu_vv_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccsu_vv_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vwmaccsu_vv_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccsu_vv_i64m1((vint64m1_t)(op0), (vint32mf2_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vwmaccsu_vv_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccsu_vv_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vwmaccsu_vv_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccsu_vv_i64m2((vint64m2_t)(op0), (vint32m1_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vwmaccsu_vv_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccsu_vv_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vwmaccsu_vv_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccsu_vv_i64m4((vint64m4_t)(op0), (vint32m2_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vwmaccsu_vv_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccsu_vv_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vwmaccsu_vv_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccsu_vv_i64m8((vint64m8_t)(op0), (vint32m4_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vwmaccsu_vv_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccsu_vv_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vwmaccsu_vx_i16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccsu_vx_i16mf4((vint16mf4_t)(op0), (int8_t)(op1), (vuint8mf8_t)(op2), (size_t)(op3)) +#define vwmaccsu_vx_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccsu_vx_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (int8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vwmaccsu_vx_i16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccsu_vx_i16mf2((vint16mf2_t)(op0), (int8_t)(op1), (vuint8mf4_t)(op2), (size_t)(op3)) +#define vwmaccsu_vx_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccsu_vx_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (int8_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vwmaccsu_vx_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccsu_vx_i16m1((vint16m1_t)(op0), (int8_t)(op1), (vuint8mf2_t)(op2), (size_t)(op3)) +#define vwmaccsu_vx_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccsu_vx_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (int8_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vwmaccsu_vx_i16m2(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccsu_vx_i16m2((vint16m2_t)(op0), (int8_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vwmaccsu_vx_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccsu_vx_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (int8_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vwmaccsu_vx_i16m4(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccsu_vx_i16m4((vint16m4_t)(op0), (int8_t)(op1), (vuint8m2_t)(op2), (size_t)(op3)) +#define vwmaccsu_vx_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccsu_vx_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (int8_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vwmaccsu_vx_i16m8(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccsu_vx_i16m8((vint16m8_t)(op0), (int8_t)(op1), (vuint8m4_t)(op2), (size_t)(op3)) +#define vwmaccsu_vx_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccsu_vx_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (int8_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vwmaccsu_vx_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccsu_vx_i32mf2((vint32mf2_t)(op0), (int16_t)(op1), (vuint16mf4_t)(op2), (size_t)(op3)) +#define vwmaccsu_vx_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccsu_vx_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (int16_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vwmaccsu_vx_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccsu_vx_i32m1((vint32m1_t)(op0), (int16_t)(op1), (vuint16mf2_t)(op2), (size_t)(op3)) +#define vwmaccsu_vx_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccsu_vx_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (int16_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vwmaccsu_vx_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccsu_vx_i32m2((vint32m2_t)(op0), (int16_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vwmaccsu_vx_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccsu_vx_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (int16_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vwmaccsu_vx_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccsu_vx_i32m4((vint32m4_t)(op0), (int16_t)(op1), (vuint16m2_t)(op2), (size_t)(op3)) +#define vwmaccsu_vx_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccsu_vx_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (int16_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vwmaccsu_vx_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccsu_vx_i32m8((vint32m8_t)(op0), (int16_t)(op1), (vuint16m4_t)(op2), (size_t)(op3)) +#define vwmaccsu_vx_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccsu_vx_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (int16_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vwmaccsu_vx_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccsu_vx_i64m1((vint64m1_t)(op0), (int32_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vwmaccsu_vx_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccsu_vx_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (int32_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vwmaccsu_vx_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccsu_vx_i64m2((vint64m2_t)(op0), (int32_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vwmaccsu_vx_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccsu_vx_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (int32_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vwmaccsu_vx_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccsu_vx_i64m4((vint64m4_t)(op0), (int32_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vwmaccsu_vx_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccsu_vx_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (int32_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vwmaccsu_vx_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccsu_vx_i64m8((vint64m8_t)(op0), (int32_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vwmaccsu_vx_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccsu_vx_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (int32_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vwmaccus_vx_i16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccus_vx_i16mf4((vint16mf4_t)(op0), (uint8_t)(op1), (vint8mf8_t)(op2), (size_t)(op3)) +#define vwmaccus_vx_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccus_vx_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (uint8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vwmaccus_vx_i16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccus_vx_i16mf2((vint16mf2_t)(op0), (uint8_t)(op1), (vint8mf4_t)(op2), (size_t)(op3)) +#define vwmaccus_vx_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccus_vx_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (uint8_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vwmaccus_vx_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccus_vx_i16m1((vint16m1_t)(op0), (uint8_t)(op1), (vint8mf2_t)(op2), (size_t)(op3)) +#define vwmaccus_vx_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccus_vx_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (uint8_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vwmaccus_vx_i16m2(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccus_vx_i16m2((vint16m2_t)(op0), (uint8_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vwmaccus_vx_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccus_vx_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (uint8_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vwmaccus_vx_i16m4(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccus_vx_i16m4((vint16m4_t)(op0), (uint8_t)(op1), (vint8m2_t)(op2), (size_t)(op3)) +#define vwmaccus_vx_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccus_vx_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (uint8_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vwmaccus_vx_i16m8(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccus_vx_i16m8((vint16m8_t)(op0), (uint8_t)(op1), (vint8m4_t)(op2), (size_t)(op3)) +#define vwmaccus_vx_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccus_vx_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (uint8_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vwmaccus_vx_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccus_vx_i32mf2((vint32mf2_t)(op0), (uint16_t)(op1), (vint16mf4_t)(op2), (size_t)(op3)) +#define vwmaccus_vx_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccus_vx_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (uint16_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vwmaccus_vx_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccus_vx_i32m1((vint32m1_t)(op0), (uint16_t)(op1), (vint16mf2_t)(op2), (size_t)(op3)) +#define vwmaccus_vx_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccus_vx_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (uint16_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vwmaccus_vx_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccus_vx_i32m2((vint32m2_t)(op0), (uint16_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vwmaccus_vx_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccus_vx_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (uint16_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vwmaccus_vx_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccus_vx_i32m4((vint32m4_t)(op0), (uint16_t)(op1), (vint16m2_t)(op2), (size_t)(op3)) +#define vwmaccus_vx_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccus_vx_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (uint16_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vwmaccus_vx_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccus_vx_i32m8((vint32m8_t)(op0), (uint16_t)(op1), (vint16m4_t)(op2), (size_t)(op3)) +#define vwmaccus_vx_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccus_vx_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (uint16_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vwmaccus_vx_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccus_vx_i64m1((vint64m1_t)(op0), (uint32_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vwmaccus_vx_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccus_vx_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (uint32_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vwmaccus_vx_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccus_vx_i64m2((vint64m2_t)(op0), (uint32_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vwmaccus_vx_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccus_vx_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (uint32_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vwmaccus_vx_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccus_vx_i64m4((vint64m4_t)(op0), (uint32_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vwmaccus_vx_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccus_vx_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (uint32_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vwmaccus_vx_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vwmaccus_vx_i64m8((vint64m8_t)(op0), (uint32_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vwmaccus_vx_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwmaccus_vx_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (uint32_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vmerge_vvm_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_i8m1((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vmerge_vvm_i8m2(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_i8m2((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (size_t)(op3)) +#define vmerge_vvm_i8m4(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_i8m4((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (size_t)(op3)) +#define vmerge_vvm_i8m8(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_i8m8((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (size_t)(op3)) +#define vmerge_vvm_i8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_i8mf2((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (size_t)(op3)) +#define vmerge_vvm_i8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_i8mf4((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (size_t)(op3)) +#define vmerge_vvm_i8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_i8mf8((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (size_t)(op3)) +#define vmerge_vvm_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_i16m1((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vmerge_vvm_i16m2(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_i16m2((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (size_t)(op3)) +#define vmerge_vvm_i16m4(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_i16m4((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (size_t)(op3)) +#define vmerge_vvm_i16m8(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_i16m8((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (size_t)(op3)) +#define vmerge_vvm_i16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_i16mf2((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (size_t)(op3)) +#define vmerge_vvm_i16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_i16mf4((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (size_t)(op3)) +#define vmerge_vvm_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_i32m1((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vmerge_vvm_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_i32m2((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vmerge_vvm_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_i32m4((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vmerge_vvm_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_i32m8((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (size_t)(op3)) +#define vmerge_vvm_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_i32mf2((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vmerge_vvm_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_i64m1((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vmerge_vvm_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_i64m2((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vmerge_vvm_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_i64m4((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vmerge_vvm_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_i64m8((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vmerge_vxm_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vxm_i8m1((vbool8_t)(op0), (vint8m1_t)(op1), (int8_t)(op2), (size_t)(op3)) +#define vmerge_vxm_i8m2(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vxm_i8m2((vbool4_t)(op0), (vint8m2_t)(op1), (int8_t)(op2), (size_t)(op3)) +#define vmerge_vxm_i8m4(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vxm_i8m4((vbool2_t)(op0), (vint8m4_t)(op1), (int8_t)(op2), (size_t)(op3)) +#define vmerge_vxm_i8m8(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vxm_i8m8((vbool1_t)(op0), (vint8m8_t)(op1), (int8_t)(op2), (size_t)(op3)) +#define vmerge_vxm_i8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vxm_i8mf2((vbool16_t)(op0), (vint8mf2_t)(op1), (int8_t)(op2), (size_t)(op3)) +#define vmerge_vxm_i8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vxm_i8mf4((vbool32_t)(op0), (vint8mf4_t)(op1), (int8_t)(op2), (size_t)(op3)) +#define vmerge_vxm_i8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vxm_i8mf8((vbool64_t)(op0), (vint8mf8_t)(op1), (int8_t)(op2), (size_t)(op3)) +#define vmerge_vxm_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vxm_i16m1((vbool16_t)(op0), (vint16m1_t)(op1), (int16_t)(op2), (size_t)(op3)) +#define vmerge_vxm_i16m2(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vxm_i16m2((vbool8_t)(op0), (vint16m2_t)(op1), (int16_t)(op2), (size_t)(op3)) +#define vmerge_vxm_i16m4(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vxm_i16m4((vbool4_t)(op0), (vint16m4_t)(op1), (int16_t)(op2), (size_t)(op3)) +#define vmerge_vxm_i16m8(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vxm_i16m8((vbool2_t)(op0), (vint16m8_t)(op1), (int16_t)(op2), (size_t)(op3)) +#define vmerge_vxm_i16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vxm_i16mf2((vbool32_t)(op0), (vint16mf2_t)(op1), (int16_t)(op2), (size_t)(op3)) +#define vmerge_vxm_i16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vxm_i16mf4((vbool64_t)(op0), (vint16mf4_t)(op1), (int16_t)(op2), (size_t)(op3)) +#define vmerge_vxm_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vxm_i32m1((vbool32_t)(op0), (vint32m1_t)(op1), (int32_t)(op2), (size_t)(op3)) +#define vmerge_vxm_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vxm_i32m2((vbool16_t)(op0), (vint32m2_t)(op1), (int32_t)(op2), (size_t)(op3)) +#define vmerge_vxm_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vxm_i32m4((vbool8_t)(op0), (vint32m4_t)(op1), (int32_t)(op2), (size_t)(op3)) +#define vmerge_vxm_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vxm_i32m8((vbool4_t)(op0), (vint32m8_t)(op1), (int32_t)(op2), (size_t)(op3)) +#define vmerge_vxm_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vxm_i32mf2((vbool64_t)(op0), (vint32mf2_t)(op1), (int32_t)(op2), (size_t)(op3)) +#define vmerge_vxm_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vxm_i64m1((vbool64_t)(op0), (vint64m1_t)(op1), (int64_t)(op2), (size_t)(op3)) +#define vmerge_vxm_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vxm_i64m2((vbool32_t)(op0), (vint64m2_t)(op1), (int64_t)(op2), (size_t)(op3)) +#define vmerge_vxm_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vxm_i64m4((vbool16_t)(op0), (vint64m4_t)(op1), (int64_t)(op2), (size_t)(op3)) +#define vmerge_vxm_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vxm_i64m8((vbool8_t)(op0), (vint64m8_t)(op1), (int64_t)(op2), (size_t)(op3)) +#define vsuxei32_v_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_i32m1((int32_t *)(op0), (vuint32m1_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vsuxei32_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vsuxei32_v_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_i32m2((int32_t *)(op0), (vuint32m2_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vsuxei32_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vsuxei32_v_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_i32m4((int32_t *)(op0), (vuint32m4_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vsuxei32_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vuint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vsuxei32_v_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_i32m8((int32_t *)(op0), (vuint32m8_t)(op1), (vint32m8_t)(op2), (size_t)(op3)) +#define vsuxei32_v_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_i32m8_m((vbool4_t)(op0), (int32_t *)(op1), (vuint32m8_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vsuxei32_v_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_i32mf2((int32_t *)(op0), (vuint32mf2_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vsuxei32_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vmerge_vvm_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_u8m1((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vmerge_vvm_u8m2(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_u8m2((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (size_t)(op3)) +#define vmerge_vvm_u8m4(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_u8m4((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (size_t)(op3)) +#define vmerge_vvm_u8m8(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_u8m8((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (size_t)(op3)) +#define vmerge_vvm_u8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_u8mf2((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (size_t)(op3)) +#define vmerge_vvm_u8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_u8mf4((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (size_t)(op3)) +#define vmerge_vvm_u8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_u8mf8((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (size_t)(op3)) +#define vmerge_vvm_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_u16m1((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vmerge_vvm_u16m2(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_u16m2((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (size_t)(op3)) +#define vmerge_vvm_u16m4(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_u16m4((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (size_t)(op3)) +#define vmerge_vvm_u16m8(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_u16m8((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (size_t)(op3)) +#define vmerge_vvm_u16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_u16mf2((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (size_t)(op3)) +#define vmerge_vvm_u16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_u16mf4((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (size_t)(op3)) +#define vmerge_vvm_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_u32m1((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vmerge_vvm_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_u32m2((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vmerge_vvm_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_u32m4((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vmerge_vvm_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_u32m8((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (size_t)(op3)) +#define vmerge_vvm_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_u32mf2((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vmerge_vvm_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_u64m1((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vmerge_vvm_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_u64m2((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vmerge_vvm_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_u64m4((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vmerge_vvm_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_u64m8((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vmerge_vxm_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vxm_u8m1((vbool8_t)(op0), (vuint8m1_t)(op1), (uint8_t)(op2), (size_t)(op3)) +#define vmerge_vxm_u8m2(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vxm_u8m2((vbool4_t)(op0), (vuint8m2_t)(op1), (uint8_t)(op2), (size_t)(op3)) +#define vmerge_vxm_u8m4(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vxm_u8m4((vbool2_t)(op0), (vuint8m4_t)(op1), (uint8_t)(op2), (size_t)(op3)) +#define vmerge_vxm_u8m8(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vxm_u8m8((vbool1_t)(op0), (vuint8m8_t)(op1), (uint8_t)(op2), (size_t)(op3)) +#define vmerge_vxm_u8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vxm_u8mf2((vbool16_t)(op0), (vuint8mf2_t)(op1), (uint8_t)(op2), (size_t)(op3)) +#define vmerge_vxm_u8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vxm_u8mf4((vbool32_t)(op0), (vuint8mf4_t)(op1), (uint8_t)(op2), (size_t)(op3)) +#define vmerge_vxm_u8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vxm_u8mf8((vbool64_t)(op0), (vuint8mf8_t)(op1), (uint8_t)(op2), (size_t)(op3)) +#define vmerge_vxm_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vxm_u16m1((vbool16_t)(op0), (vuint16m1_t)(op1), (uint16_t)(op2), (size_t)(op3)) +#define vmerge_vxm_u16m2(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vxm_u16m2((vbool8_t)(op0), (vuint16m2_t)(op1), (uint16_t)(op2), (size_t)(op3)) +#define vmerge_vxm_u16m4(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vxm_u16m4((vbool4_t)(op0), (vuint16m4_t)(op1), (uint16_t)(op2), (size_t)(op3)) +#define vmerge_vxm_u16m8(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vxm_u16m8((vbool2_t)(op0), (vuint16m8_t)(op1), (uint16_t)(op2), (size_t)(op3)) +#define vmerge_vxm_u16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vxm_u16mf2((vbool32_t)(op0), (vuint16mf2_t)(op1), (uint16_t)(op2), (size_t)(op3)) +#define vmerge_vxm_u16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vxm_u16mf4((vbool64_t)(op0), (vuint16mf4_t)(op1), (uint16_t)(op2), (size_t)(op3)) +#define vmerge_vxm_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vxm_u32m1((vbool32_t)(op0), (vuint32m1_t)(op1), (uint32_t)(op2), (size_t)(op3)) +#define vmerge_vxm_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vxm_u32m2((vbool16_t)(op0), (vuint32m2_t)(op1), (uint32_t)(op2), (size_t)(op3)) +#define vmerge_vxm_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vxm_u32m4((vbool8_t)(op0), (vuint32m4_t)(op1), (uint32_t)(op2), (size_t)(op3)) +#define vmerge_vxm_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vxm_u32m8((vbool4_t)(op0), (vuint32m8_t)(op1), (uint32_t)(op2), (size_t)(op3)) +#define vmerge_vxm_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vxm_u32mf2((vbool64_t)(op0), (vuint32mf2_t)(op1), (uint32_t)(op2), (size_t)(op3)) +#define vmerge_vxm_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vxm_u64m1((vbool64_t)(op0), (vuint64m1_t)(op1), (uint64_t)(op2), (size_t)(op3)) +#define vmerge_vxm_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vxm_u64m2((vbool32_t)(op0), (vuint64m2_t)(op1), (uint64_t)(op2), (size_t)(op3)) +#define vmerge_vxm_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vxm_u64m4((vbool16_t)(op0), (vuint64m4_t)(op1), (uint64_t)(op2), (size_t)(op3)) +#define vmerge_vxm_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vxm_u64m8((vbool8_t)(op0), (vuint64m8_t)(op1), (uint64_t)(op2), (size_t)(op3)) +#define vmv_v_v_u8m1(op0, op1) \ +__builtin_rvv_vmv_v_v_u8m1((vuint8m1_t)(op0), (size_t)(op1)) +#define vmv_v_v_u8m2(op0, op1) \ +__builtin_rvv_vmv_v_v_u8m2((vuint8m2_t)(op0), (size_t)(op1)) +#define vmv_v_v_u8m4(op0, op1) \ +__builtin_rvv_vmv_v_v_u8m4((vuint8m4_t)(op0), (size_t)(op1)) +#define vmv_v_v_u8m8(op0, op1) \ +__builtin_rvv_vmv_v_v_u8m8((vuint8m8_t)(op0), (size_t)(op1)) +#define vmv_v_v_u8mf2(op0, op1) \ +__builtin_rvv_vmv_v_v_u8mf2((vuint8mf2_t)(op0), (size_t)(op1)) +#define vmv_v_v_u8mf4(op0, op1) \ +__builtin_rvv_vmv_v_v_u8mf4((vuint8mf4_t)(op0), (size_t)(op1)) +#define vmv_v_v_u8mf8(op0, op1) \ +__builtin_rvv_vmv_v_v_u8mf8((vuint8mf8_t)(op0), (size_t)(op1)) +#define vmv_v_v_u16m1(op0, op1) \ +__builtin_rvv_vmv_v_v_u16m1((vuint16m1_t)(op0), (size_t)(op1)) +#define vmv_v_v_u16m2(op0, op1) \ +__builtin_rvv_vmv_v_v_u16m2((vuint16m2_t)(op0), (size_t)(op1)) +#define vmv_v_v_u16m4(op0, op1) \ +__builtin_rvv_vmv_v_v_u16m4((vuint16m4_t)(op0), (size_t)(op1)) +#define vmv_v_v_u16m8(op0, op1) \ +__builtin_rvv_vmv_v_v_u16m8((vuint16m8_t)(op0), (size_t)(op1)) +#define vmv_v_v_u16mf2(op0, op1) \ +__builtin_rvv_vmv_v_v_u16mf2((vuint16mf2_t)(op0), (size_t)(op1)) +#define vmv_v_v_u16mf4(op0, op1) \ +__builtin_rvv_vmv_v_v_u16mf4((vuint16mf4_t)(op0), (size_t)(op1)) +#define vmv_v_v_u32m1(op0, op1) \ +__builtin_rvv_vmv_v_v_u32m1((vuint32m1_t)(op0), (size_t)(op1)) +#define vmv_v_v_u32m2(op0, op1) \ +__builtin_rvv_vmv_v_v_u32m2((vuint32m2_t)(op0), (size_t)(op1)) +#define vmv_v_v_u32m4(op0, op1) \ +__builtin_rvv_vmv_v_v_u32m4((vuint32m4_t)(op0), (size_t)(op1)) +#define vmv_v_v_u32m8(op0, op1) \ +__builtin_rvv_vmv_v_v_u32m8((vuint32m8_t)(op0), (size_t)(op1)) +#define vmv_v_v_u32mf2(op0, op1) \ +__builtin_rvv_vmv_v_v_u32mf2((vuint32mf2_t)(op0), (size_t)(op1)) +#define vmv_v_v_u64m1(op0, op1) \ +__builtin_rvv_vmv_v_v_u64m1((vuint64m1_t)(op0), (size_t)(op1)) +#define vmv_v_v_u64m2(op0, op1) \ +__builtin_rvv_vmv_v_v_u64m2((vuint64m2_t)(op0), (size_t)(op1)) +#define vmv_v_v_u64m4(op0, op1) \ +__builtin_rvv_vmv_v_v_u64m4((vuint64m4_t)(op0), (size_t)(op1)) +#define vmv_v_v_u64m8(op0, op1) \ +__builtin_rvv_vmv_v_v_u64m8((vuint64m8_t)(op0), (size_t)(op1)) +#define vmv_v_v_i8m1(op0, op1) \ +__builtin_rvv_vmv_v_v_i8m1((vint8m1_t)(op0), (size_t)(op1)) +#define vmv_v_v_i8m2(op0, op1) \ +__builtin_rvv_vmv_v_v_i8m2((vint8m2_t)(op0), (size_t)(op1)) +#define vmv_v_v_i8m4(op0, op1) \ +__builtin_rvv_vmv_v_v_i8m4((vint8m4_t)(op0), (size_t)(op1)) +#define vmv_v_v_i8m8(op0, op1) \ +__builtin_rvv_vmv_v_v_i8m8((vint8m8_t)(op0), (size_t)(op1)) +#define vmv_v_v_i8mf2(op0, op1) \ +__builtin_rvv_vmv_v_v_i8mf2((vint8mf2_t)(op0), (size_t)(op1)) +#define vmv_v_v_i8mf4(op0, op1) \ +__builtin_rvv_vmv_v_v_i8mf4((vint8mf4_t)(op0), (size_t)(op1)) +#define vmv_v_v_i8mf8(op0, op1) \ +__builtin_rvv_vmv_v_v_i8mf8((vint8mf8_t)(op0), (size_t)(op1)) +#define vmv_v_v_i16m1(op0, op1) \ +__builtin_rvv_vmv_v_v_i16m1((vint16m1_t)(op0), (size_t)(op1)) +#define vmv_v_v_i16m2(op0, op1) \ +__builtin_rvv_vmv_v_v_i16m2((vint16m2_t)(op0), (size_t)(op1)) +#define vmv_v_v_i16m4(op0, op1) \ +__builtin_rvv_vmv_v_v_i16m4((vint16m4_t)(op0), (size_t)(op1)) +#define vmv_v_v_i16m8(op0, op1) \ +__builtin_rvv_vmv_v_v_i16m8((vint16m8_t)(op0), (size_t)(op1)) +#define vmv_v_v_i16mf2(op0, op1) \ +__builtin_rvv_vmv_v_v_i16mf2((vint16mf2_t)(op0), (size_t)(op1)) +#define vmv_v_v_i16mf4(op0, op1) \ +__builtin_rvv_vmv_v_v_i16mf4((vint16mf4_t)(op0), (size_t)(op1)) +#define vmv_v_v_i32m1(op0, op1) \ +__builtin_rvv_vmv_v_v_i32m1((vint32m1_t)(op0), (size_t)(op1)) +#define vmv_v_v_i32m2(op0, op1) \ +__builtin_rvv_vmv_v_v_i32m2((vint32m2_t)(op0), (size_t)(op1)) +#define vmv_v_v_i32m4(op0, op1) \ +__builtin_rvv_vmv_v_v_i32m4((vint32m4_t)(op0), (size_t)(op1)) +#define vmv_v_v_i32m8(op0, op1) \ +__builtin_rvv_vmv_v_v_i32m8((vint32m8_t)(op0), (size_t)(op1)) +#define vmv_v_v_i32mf2(op0, op1) \ +__builtin_rvv_vmv_v_v_i32mf2((vint32mf2_t)(op0), (size_t)(op1)) +#define vmv_v_v_i64m1(op0, op1) \ +__builtin_rvv_vmv_v_v_i64m1((vint64m1_t)(op0), (size_t)(op1)) +#define vmv_v_v_i64m2(op0, op1) \ +__builtin_rvv_vmv_v_v_i64m2((vint64m2_t)(op0), (size_t)(op1)) +#define vmv_v_v_i64m4(op0, op1) \ +__builtin_rvv_vmv_v_v_i64m4((vint64m4_t)(op0), (size_t)(op1)) +#define vmv_v_v_i64m8(op0, op1) \ +__builtin_rvv_vmv_v_v_i64m8((vint64m8_t)(op0), (size_t)(op1)) +#define vmv_v_x_i8m1(op0, op1) \ +__builtin_rvv_vmv_v_x_i8m1((int8_t)(op0), (size_t)(op1)) +#define vmv_v_x_i8m2(op0, op1) \ +__builtin_rvv_vmv_v_x_i8m2((int8_t)(op0), (size_t)(op1)) +#define vmv_v_x_i8m4(op0, op1) \ +__builtin_rvv_vmv_v_x_i8m4((int8_t)(op0), (size_t)(op1)) +#define vmv_v_x_i8m8(op0, op1) \ +__builtin_rvv_vmv_v_x_i8m8((int8_t)(op0), (size_t)(op1)) +#define vmv_v_x_i8mf2(op0, op1) \ +__builtin_rvv_vmv_v_x_i8mf2((int8_t)(op0), (size_t)(op1)) +#define vmv_v_x_i8mf4(op0, op1) \ +__builtin_rvv_vmv_v_x_i8mf4((int8_t)(op0), (size_t)(op1)) +#define vmv_v_x_i8mf8(op0, op1) \ +__builtin_rvv_vmv_v_x_i8mf8((int8_t)(op0), (size_t)(op1)) +#define vmv_v_x_i16m1(op0, op1) \ +__builtin_rvv_vmv_v_x_i16m1((int16_t)(op0), (size_t)(op1)) +#define vmv_v_x_i16m2(op0, op1) \ +__builtin_rvv_vmv_v_x_i16m2((int16_t)(op0), (size_t)(op1)) +#define vmv_v_x_i16m4(op0, op1) \ +__builtin_rvv_vmv_v_x_i16m4((int16_t)(op0), (size_t)(op1)) +#define vmv_v_x_i16m8(op0, op1) \ +__builtin_rvv_vmv_v_x_i16m8((int16_t)(op0), (size_t)(op1)) +#define vmv_v_x_i16mf2(op0, op1) \ +__builtin_rvv_vmv_v_x_i16mf2((int16_t)(op0), (size_t)(op1)) +#define vmv_v_x_i16mf4(op0, op1) \ +__builtin_rvv_vmv_v_x_i16mf4((int16_t)(op0), (size_t)(op1)) +#define vmv_v_x_i32m1(op0, op1) \ +__builtin_rvv_vmv_v_x_i32m1((int32_t)(op0), (size_t)(op1)) +#define vmv_v_x_i32m2(op0, op1) \ +__builtin_rvv_vmv_v_x_i32m2((int32_t)(op0), (size_t)(op1)) +#define vmv_v_x_i32m4(op0, op1) \ +__builtin_rvv_vmv_v_x_i32m4((int32_t)(op0), (size_t)(op1)) +#define vmv_v_x_i32m8(op0, op1) \ +__builtin_rvv_vmv_v_x_i32m8((int32_t)(op0), (size_t)(op1)) +#define vmv_v_x_i32mf2(op0, op1) \ +__builtin_rvv_vmv_v_x_i32mf2((int32_t)(op0), (size_t)(op1)) +#define vmv_v_x_i64m1(op0, op1) \ +__builtin_rvv_vmv_v_x_i64m1((int64_t)(op0), (size_t)(op1)) +#define vmv_v_x_i64m2(op0, op1) \ +__builtin_rvv_vmv_v_x_i64m2((int64_t)(op0), (size_t)(op1)) +#define vmv_v_x_i64m4(op0, op1) \ +__builtin_rvv_vmv_v_x_i64m4((int64_t)(op0), (size_t)(op1)) +#define vmv_v_x_i64m8(op0, op1) \ +__builtin_rvv_vmv_v_x_i64m8((int64_t)(op0), (size_t)(op1)) +#define vmv_v_x_u8m1(op0, op1) \ +__builtin_rvv_vmv_v_x_u8m1((uint8_t)(op0), (size_t)(op1)) +#define vmv_v_x_u8m2(op0, op1) \ +__builtin_rvv_vmv_v_x_u8m2((uint8_t)(op0), (size_t)(op1)) +#define vmv_v_x_u8m4(op0, op1) \ +__builtin_rvv_vmv_v_x_u8m4((uint8_t)(op0), (size_t)(op1)) +#define vmv_v_x_u8m8(op0, op1) \ +__builtin_rvv_vmv_v_x_u8m8((uint8_t)(op0), (size_t)(op1)) +#define vmv_v_x_u8mf2(op0, op1) \ +__builtin_rvv_vmv_v_x_u8mf2((uint8_t)(op0), (size_t)(op1)) +#define vmv_v_x_u8mf4(op0, op1) \ +__builtin_rvv_vmv_v_x_u8mf4((uint8_t)(op0), (size_t)(op1)) +#define vmv_v_x_u8mf8(op0, op1) \ +__builtin_rvv_vmv_v_x_u8mf8((uint8_t)(op0), (size_t)(op1)) +#define vmv_v_x_u16m1(op0, op1) \ +__builtin_rvv_vmv_v_x_u16m1((uint16_t)(op0), (size_t)(op1)) +#define vmv_v_x_u16m2(op0, op1) \ +__builtin_rvv_vmv_v_x_u16m2((uint16_t)(op0), (size_t)(op1)) +#define vmv_v_x_u16m4(op0, op1) \ +__builtin_rvv_vmv_v_x_u16m4((uint16_t)(op0), (size_t)(op1)) +#define vmv_v_x_u16m8(op0, op1) \ +__builtin_rvv_vmv_v_x_u16m8((uint16_t)(op0), (size_t)(op1)) +#define vmv_v_x_u16mf2(op0, op1) \ +__builtin_rvv_vmv_v_x_u16mf2((uint16_t)(op0), (size_t)(op1)) +#define vmv_v_x_u16mf4(op0, op1) \ +__builtin_rvv_vmv_v_x_u16mf4((uint16_t)(op0), (size_t)(op1)) +#define vmv_v_x_u32m1(op0, op1) \ +__builtin_rvv_vmv_v_x_u32m1((uint32_t)(op0), (size_t)(op1)) +#define vmv_v_x_u32m2(op0, op1) \ +__builtin_rvv_vmv_v_x_u32m2((uint32_t)(op0), (size_t)(op1)) +#define vmv_v_x_u32m4(op0, op1) \ +__builtin_rvv_vmv_v_x_u32m4((uint32_t)(op0), (size_t)(op1)) +#define vmv_v_x_u32m8(op0, op1) \ +__builtin_rvv_vmv_v_x_u32m8((uint32_t)(op0), (size_t)(op1)) +#define vmv_v_x_u32mf2(op0, op1) \ +__builtin_rvv_vmv_v_x_u32mf2((uint32_t)(op0), (size_t)(op1)) +#define vmv_v_x_u64m1(op0, op1) \ +__builtin_rvv_vmv_v_x_u64m1((uint64_t)(op0), (size_t)(op1)) +#define vmv_v_x_u64m2(op0, op1) \ +__builtin_rvv_vmv_v_x_u64m2((uint64_t)(op0), (size_t)(op1)) +#define vmv_v_x_u64m4(op0, op1) \ +__builtin_rvv_vmv_v_x_u64m4((uint64_t)(op0), (size_t)(op1)) +#define vmv_v_x_u64m8(op0, op1) \ +__builtin_rvv_vmv_v_x_u64m8((uint64_t)(op0), (size_t)(op1)) +#define vsaddu_vv_u8m1(op0, op1, op2) \ +__builtin_rvv_vsaddu_vv_u8m1((vuint8m1_t)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vsaddu_vv_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsaddu_vv_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vsaddu_vv_u8m2(op0, op1, op2) \ +__builtin_rvv_vsaddu_vv_u8m2((vuint8m2_t)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vsaddu_vv_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsaddu_vv_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vsaddu_vv_u8m4(op0, op1, op2) \ +__builtin_rvv_vsaddu_vv_u8m4((vuint8m4_t)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vsaddu_vv_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsaddu_vv_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vsaddu_vv_u8m8(op0, op1, op2) \ +__builtin_rvv_vsaddu_vv_u8m8((vuint8m8_t)(op0), (vuint8m8_t)(op1), (size_t)(op2)) +#define vsaddu_vv_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsaddu_vv_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (vuint8m8_t)(op3), (size_t)(op4)) +#define vsaddu_vv_u8mf2(op0, op1, op2) \ +__builtin_rvv_vsaddu_vv_u8mf2((vuint8mf2_t)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vsaddu_vv_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsaddu_vv_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vsaddu_vv_u8mf4(op0, op1, op2) \ +__builtin_rvv_vsaddu_vv_u8mf4((vuint8mf4_t)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vsaddu_vv_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsaddu_vv_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vsaddu_vv_u8mf8(op0, op1, op2) \ +__builtin_rvv_vsaddu_vv_u8mf8((vuint8mf8_t)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vsaddu_vv_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsaddu_vv_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vsaddu_vv_u16m1(op0, op1, op2) \ +__builtin_rvv_vsaddu_vv_u16m1((vuint16m1_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vsaddu_vv_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsaddu_vv_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vsaddu_vv_u16m2(op0, op1, op2) \ +__builtin_rvv_vsaddu_vv_u16m2((vuint16m2_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vsaddu_vv_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsaddu_vv_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vsaddu_vv_u16m4(op0, op1, op2) \ +__builtin_rvv_vsaddu_vv_u16m4((vuint16m4_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vsaddu_vv_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsaddu_vv_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vsaddu_vv_u16m8(op0, op1, op2) \ +__builtin_rvv_vsaddu_vv_u16m8((vuint16m8_t)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vsaddu_vv_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsaddu_vv_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vsaddu_vv_u16mf2(op0, op1, op2) \ +__builtin_rvv_vsaddu_vv_u16mf2((vuint16mf2_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vsaddu_vv_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsaddu_vv_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vsaddu_vv_u16mf4(op0, op1, op2) \ +__builtin_rvv_vsaddu_vv_u16mf4((vuint16mf4_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vsaddu_vv_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsaddu_vv_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vsaddu_vv_u32m1(op0, op1, op2) \ +__builtin_rvv_vsaddu_vv_u32m1((vuint32m1_t)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vsaddu_vv_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsaddu_vv_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vsaddu_vv_u32m2(op0, op1, op2) \ +__builtin_rvv_vsaddu_vv_u32m2((vuint32m2_t)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vsaddu_vv_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsaddu_vv_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vsaddu_vv_u32m4(op0, op1, op2) \ +__builtin_rvv_vsaddu_vv_u32m4((vuint32m4_t)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vsaddu_vv_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsaddu_vv_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vsaddu_vv_u32m8(op0, op1, op2) \ +__builtin_rvv_vsaddu_vv_u32m8((vuint32m8_t)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vsaddu_vv_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsaddu_vv_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vsaddu_vv_u32mf2(op0, op1, op2) \ +__builtin_rvv_vsaddu_vv_u32mf2((vuint32mf2_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vsaddu_vv_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsaddu_vv_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vsaddu_vv_u64m1(op0, op1, op2) \ +__builtin_rvv_vsaddu_vv_u64m1((vuint64m1_t)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vsaddu_vv_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsaddu_vv_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vsaddu_vv_u64m2(op0, op1, op2) \ +__builtin_rvv_vsaddu_vv_u64m2((vuint64m2_t)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vsaddu_vv_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsaddu_vv_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vsaddu_vv_u64m4(op0, op1, op2) \ +__builtin_rvv_vsaddu_vv_u64m4((vuint64m4_t)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vsaddu_vv_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsaddu_vv_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vsaddu_vv_u64m8(op0, op1, op2) \ +__builtin_rvv_vsaddu_vv_u64m8((vuint64m8_t)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vsaddu_vv_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsaddu_vv_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vsaddu_vx_u8m1(op0, op1, op2) \ +__builtin_rvv_vsaddu_vx_u8m1((vuint8m1_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vsaddu_vx_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsaddu_vx_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vsaddu_vx_u8m2(op0, op1, op2) \ +__builtin_rvv_vsaddu_vx_u8m2((vuint8m2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vsaddu_vx_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsaddu_vx_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vsaddu_vx_u8m4(op0, op1, op2) \ +__builtin_rvv_vsaddu_vx_u8m4((vuint8m4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vsaddu_vx_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsaddu_vx_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vsaddu_vx_u8m8(op0, op1, op2) \ +__builtin_rvv_vsaddu_vx_u8m8((vuint8m8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vsaddu_vx_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsaddu_vx_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vsaddu_vx_u8mf2(op0, op1, op2) \ +__builtin_rvv_vsaddu_vx_u8mf2((vuint8mf2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vsaddu_vx_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsaddu_vx_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vsaddu_vx_u8mf4(op0, op1, op2) \ +__builtin_rvv_vsaddu_vx_u8mf4((vuint8mf4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vsaddu_vx_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsaddu_vx_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vsaddu_vx_u8mf8(op0, op1, op2) \ +__builtin_rvv_vsaddu_vx_u8mf8((vuint8mf8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vsaddu_vx_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsaddu_vx_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vsaddu_vx_u16m1(op0, op1, op2) \ +__builtin_rvv_vsaddu_vx_u16m1((vuint16m1_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vsaddu_vx_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsaddu_vx_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vsaddu_vx_u16m2(op0, op1, op2) \ +__builtin_rvv_vsaddu_vx_u16m2((vuint16m2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vsaddu_vx_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsaddu_vx_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vsaddu_vx_u16m4(op0, op1, op2) \ +__builtin_rvv_vsaddu_vx_u16m4((vuint16m4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vsaddu_vx_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsaddu_vx_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vsaddu_vx_u16m8(op0, op1, op2) \ +__builtin_rvv_vsaddu_vx_u16m8((vuint16m8_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vsaddu_vx_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsaddu_vx_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vsaddu_vx_u16mf2(op0, op1, op2) \ +__builtin_rvv_vsaddu_vx_u16mf2((vuint16mf2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vsaddu_vx_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsaddu_vx_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vsaddu_vx_u16mf4(op0, op1, op2) \ +__builtin_rvv_vsaddu_vx_u16mf4((vuint16mf4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vsaddu_vx_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsaddu_vx_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vsaddu_vx_u32m1(op0, op1, op2) \ +__builtin_rvv_vsaddu_vx_u32m1((vuint32m1_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vsaddu_vx_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsaddu_vx_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vsaddu_vx_u32m2(op0, op1, op2) \ +__builtin_rvv_vsaddu_vx_u32m2((vuint32m2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vsaddu_vx_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsaddu_vx_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vsaddu_vx_u32m4(op0, op1, op2) \ +__builtin_rvv_vsaddu_vx_u32m4((vuint32m4_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vsaddu_vx_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsaddu_vx_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vsaddu_vx_u32m8(op0, op1, op2) \ +__builtin_rvv_vsaddu_vx_u32m8((vuint32m8_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vsaddu_vx_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsaddu_vx_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vsaddu_vx_u32mf2(op0, op1, op2) \ +__builtin_rvv_vsaddu_vx_u32mf2((vuint32mf2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vsaddu_vx_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsaddu_vx_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vsaddu_vx_u64m1(op0, op1, op2) \ +__builtin_rvv_vsaddu_vx_u64m1((vuint64m1_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vsaddu_vx_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsaddu_vx_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vsaddu_vx_u64m2(op0, op1, op2) \ +__builtin_rvv_vsaddu_vx_u64m2((vuint64m2_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vsaddu_vx_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsaddu_vx_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vsaddu_vx_u64m4(op0, op1, op2) \ +__builtin_rvv_vsaddu_vx_u64m4((vuint64m4_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vsaddu_vx_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsaddu_vx_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vsaddu_vx_u64m8(op0, op1, op2) \ +__builtin_rvv_vsaddu_vx_u64m8((vuint64m8_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vsaddu_vx_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsaddu_vx_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vsadd_vv_i8m1(op0, op1, op2) \ +__builtin_rvv_vsadd_vv_i8m1((vint8m1_t)(op0), (vint8m1_t)(op1), (size_t)(op2)) +#define vsadd_vv_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsadd_vv_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vsadd_vv_i8m2(op0, op1, op2) \ +__builtin_rvv_vsadd_vv_i8m2((vint8m2_t)(op0), (vint8m2_t)(op1), (size_t)(op2)) +#define vsadd_vv_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsadd_vv_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vsadd_vv_i8m4(op0, op1, op2) \ +__builtin_rvv_vsadd_vv_i8m4((vint8m4_t)(op0), (vint8m4_t)(op1), (size_t)(op2)) +#define vsadd_vv_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsadd_vv_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vsadd_vv_i8m8(op0, op1, op2) \ +__builtin_rvv_vsadd_vv_i8m8((vint8m8_t)(op0), (vint8m8_t)(op1), (size_t)(op2)) +#define vsadd_vv_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsadd_vv_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (vint8m8_t)(op3), (size_t)(op4)) +#define vsadd_vv_i8mf2(op0, op1, op2) \ +__builtin_rvv_vsadd_vv_i8mf2((vint8mf2_t)(op0), (vint8mf2_t)(op1), (size_t)(op2)) +#define vsadd_vv_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsadd_vv_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vsadd_vv_i8mf4(op0, op1, op2) \ +__builtin_rvv_vsadd_vv_i8mf4((vint8mf4_t)(op0), (vint8mf4_t)(op1), (size_t)(op2)) +#define vsadd_vv_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsadd_vv_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vsadd_vv_i8mf8(op0, op1, op2) \ +__builtin_rvv_vsadd_vv_i8mf8((vint8mf8_t)(op0), (vint8mf8_t)(op1), (size_t)(op2)) +#define vsadd_vv_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsadd_vv_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vsadd_vv_i16m1(op0, op1, op2) \ +__builtin_rvv_vsadd_vv_i16m1((vint16m1_t)(op0), (vint16m1_t)(op1), (size_t)(op2)) +#define vsadd_vv_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsadd_vv_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vsadd_vv_i16m2(op0, op1, op2) \ +__builtin_rvv_vsadd_vv_i16m2((vint16m2_t)(op0), (vint16m2_t)(op1), (size_t)(op2)) +#define vsadd_vv_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsadd_vv_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vsadd_vv_i16m4(op0, op1, op2) \ +__builtin_rvv_vsadd_vv_i16m4((vint16m4_t)(op0), (vint16m4_t)(op1), (size_t)(op2)) +#define vsadd_vv_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsadd_vv_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vsadd_vv_i16m8(op0, op1, op2) \ +__builtin_rvv_vsadd_vv_i16m8((vint16m8_t)(op0), (vint16m8_t)(op1), (size_t)(op2)) +#define vsadd_vv_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsadd_vv_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (vint16m8_t)(op3), (size_t)(op4)) +#define vsadd_vv_i16mf2(op0, op1, op2) \ +__builtin_rvv_vsadd_vv_i16mf2((vint16mf2_t)(op0), (vint16mf2_t)(op1), (size_t)(op2)) +#define vsadd_vv_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsadd_vv_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vsadd_vv_i16mf4(op0, op1, op2) \ +__builtin_rvv_vsadd_vv_i16mf4((vint16mf4_t)(op0), (vint16mf4_t)(op1), (size_t)(op2)) +#define vsadd_vv_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsadd_vv_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vsadd_vv_i32m1(op0, op1, op2) \ +__builtin_rvv_vsadd_vv_i32m1((vint32m1_t)(op0), (vint32m1_t)(op1), (size_t)(op2)) +#define vsadd_vv_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsadd_vv_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vsadd_vv_i32m2(op0, op1, op2) \ +__builtin_rvv_vsadd_vv_i32m2((vint32m2_t)(op0), (vint32m2_t)(op1), (size_t)(op2)) +#define vsadd_vv_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsadd_vv_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vsadd_vv_i32m4(op0, op1, op2) \ +__builtin_rvv_vsadd_vv_i32m4((vint32m4_t)(op0), (vint32m4_t)(op1), (size_t)(op2)) +#define vsadd_vv_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsadd_vv_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vsadd_vv_i32m8(op0, op1, op2) \ +__builtin_rvv_vsadd_vv_i32m8((vint32m8_t)(op0), (vint32m8_t)(op1), (size_t)(op2)) +#define vsadd_vv_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsadd_vv_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vsadd_vv_i32mf2(op0, op1, op2) \ +__builtin_rvv_vsadd_vv_i32mf2((vint32mf2_t)(op0), (vint32mf2_t)(op1), (size_t)(op2)) +#define vsadd_vv_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsadd_vv_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vsadd_vv_i64m1(op0, op1, op2) \ +__builtin_rvv_vsadd_vv_i64m1((vint64m1_t)(op0), (vint64m1_t)(op1), (size_t)(op2)) +#define vsadd_vv_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsadd_vv_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vsadd_vv_i64m2(op0, op1, op2) \ +__builtin_rvv_vsadd_vv_i64m2((vint64m2_t)(op0), (vint64m2_t)(op1), (size_t)(op2)) +#define vsadd_vv_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsadd_vv_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vsadd_vv_i64m4(op0, op1, op2) \ +__builtin_rvv_vsadd_vv_i64m4((vint64m4_t)(op0), (vint64m4_t)(op1), (size_t)(op2)) +#define vsadd_vv_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsadd_vv_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vsadd_vv_i64m8(op0, op1, op2) \ +__builtin_rvv_vsadd_vv_i64m8((vint64m8_t)(op0), (vint64m8_t)(op1), (size_t)(op2)) +#define vsadd_vv_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsadd_vv_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vsadd_vx_i8m1(op0, op1, op2) \ +__builtin_rvv_vsadd_vx_i8m1((vint8m1_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vsadd_vx_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsadd_vx_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vsadd_vx_i8m2(op0, op1, op2) \ +__builtin_rvv_vsadd_vx_i8m2((vint8m2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vsadd_vx_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsadd_vx_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vsadd_vx_i8m4(op0, op1, op2) \ +__builtin_rvv_vsadd_vx_i8m4((vint8m4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vsadd_vx_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsadd_vx_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vsadd_vx_i8m8(op0, op1, op2) \ +__builtin_rvv_vsadd_vx_i8m8((vint8m8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vsadd_vx_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsadd_vx_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vsadd_vx_i8mf2(op0, op1, op2) \ +__builtin_rvv_vsadd_vx_i8mf2((vint8mf2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vsadd_vx_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsadd_vx_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vsadd_vx_i8mf4(op0, op1, op2) \ +__builtin_rvv_vsadd_vx_i8mf4((vint8mf4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vsadd_vx_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsadd_vx_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vsadd_vx_i8mf8(op0, op1, op2) \ +__builtin_rvv_vsadd_vx_i8mf8((vint8mf8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vsadd_vx_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsadd_vx_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vsadd_vx_i16m1(op0, op1, op2) \ +__builtin_rvv_vsadd_vx_i16m1((vint16m1_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vsadd_vx_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsadd_vx_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vsadd_vx_i16m2(op0, op1, op2) \ +__builtin_rvv_vsadd_vx_i16m2((vint16m2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vsadd_vx_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsadd_vx_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vsadd_vx_i16m4(op0, op1, op2) \ +__builtin_rvv_vsadd_vx_i16m4((vint16m4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vsadd_vx_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsadd_vx_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vsadd_vx_i16m8(op0, op1, op2) \ +__builtin_rvv_vsadd_vx_i16m8((vint16m8_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vsadd_vx_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsadd_vx_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vsadd_vx_i16mf2(op0, op1, op2) \ +__builtin_rvv_vsadd_vx_i16mf2((vint16mf2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vsadd_vx_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsadd_vx_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vsadd_vx_i16mf4(op0, op1, op2) \ +__builtin_rvv_vsadd_vx_i16mf4((vint16mf4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vsadd_vx_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsadd_vx_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vsadd_vx_i32m1(op0, op1, op2) \ +__builtin_rvv_vsadd_vx_i32m1((vint32m1_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vsadd_vx_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsadd_vx_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vsadd_vx_i32m2(op0, op1, op2) \ +__builtin_rvv_vsadd_vx_i32m2((vint32m2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vsadd_vx_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsadd_vx_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vsadd_vx_i32m4(op0, op1, op2) \ +__builtin_rvv_vsadd_vx_i32m4((vint32m4_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vsadd_vx_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsadd_vx_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vsadd_vx_i32m8(op0, op1, op2) \ +__builtin_rvv_vsadd_vx_i32m8((vint32m8_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vsadd_vx_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsadd_vx_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vsadd_vx_i32mf2(op0, op1, op2) \ +__builtin_rvv_vsadd_vx_i32mf2((vint32mf2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vsadd_vx_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsadd_vx_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vsadd_vx_i64m1(op0, op1, op2) \ +__builtin_rvv_vsadd_vx_i64m1((vint64m1_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vsadd_vx_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsadd_vx_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vsadd_vx_i64m2(op0, op1, op2) \ +__builtin_rvv_vsadd_vx_i64m2((vint64m2_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vsadd_vx_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsadd_vx_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vsadd_vx_i64m4(op0, op1, op2) \ +__builtin_rvv_vsadd_vx_i64m4((vint64m4_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vsadd_vx_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsadd_vx_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vsadd_vx_i64m8(op0, op1, op2) \ +__builtin_rvv_vsadd_vx_i64m8((vint64m8_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vsadd_vx_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsadd_vx_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vsuxei32_v_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_u32m1((uint32_t *)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vsuxei32_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vsuxei32_v_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_u32m2((uint32_t *)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vsuxei32_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vsuxei32_v_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_u32m4((uint32_t *)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vsuxei32_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vsuxei32_v_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_u32m8((uint32_t *)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (size_t)(op3)) +#define vsuxei32_v_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_u32m8_m((vbool4_t)(op0), (uint32_t *)(op1), (vuint32m8_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vsuxei32_v_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_u32mf2((uint32_t *)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vsuxei32_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vssubu_vv_u8m1(op0, op1, op2) \ +__builtin_rvv_vssubu_vv_u8m1((vuint8m1_t)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vssubu_vv_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssubu_vv_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vssubu_vv_u8m2(op0, op1, op2) \ +__builtin_rvv_vssubu_vv_u8m2((vuint8m2_t)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vssubu_vv_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssubu_vv_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vssubu_vv_u8m4(op0, op1, op2) \ +__builtin_rvv_vssubu_vv_u8m4((vuint8m4_t)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vssubu_vv_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssubu_vv_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vssubu_vv_u8m8(op0, op1, op2) \ +__builtin_rvv_vssubu_vv_u8m8((vuint8m8_t)(op0), (vuint8m8_t)(op1), (size_t)(op2)) +#define vssubu_vv_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssubu_vv_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (vuint8m8_t)(op3), (size_t)(op4)) +#define vssubu_vv_u8mf2(op0, op1, op2) \ +__builtin_rvv_vssubu_vv_u8mf2((vuint8mf2_t)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vssubu_vv_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssubu_vv_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vssubu_vv_u8mf4(op0, op1, op2) \ +__builtin_rvv_vssubu_vv_u8mf4((vuint8mf4_t)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vssubu_vv_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssubu_vv_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vssubu_vv_u8mf8(op0, op1, op2) \ +__builtin_rvv_vssubu_vv_u8mf8((vuint8mf8_t)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vssubu_vv_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssubu_vv_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vssubu_vv_u16m1(op0, op1, op2) \ +__builtin_rvv_vssubu_vv_u16m1((vuint16m1_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vssubu_vv_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssubu_vv_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vssubu_vv_u16m2(op0, op1, op2) \ +__builtin_rvv_vssubu_vv_u16m2((vuint16m2_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vssubu_vv_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssubu_vv_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vssubu_vv_u16m4(op0, op1, op2) \ +__builtin_rvv_vssubu_vv_u16m4((vuint16m4_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vssubu_vv_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssubu_vv_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vssubu_vv_u16m8(op0, op1, op2) \ +__builtin_rvv_vssubu_vv_u16m8((vuint16m8_t)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vssubu_vv_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssubu_vv_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vssubu_vv_u16mf2(op0, op1, op2) \ +__builtin_rvv_vssubu_vv_u16mf2((vuint16mf2_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vssubu_vv_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssubu_vv_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vssubu_vv_u16mf4(op0, op1, op2) \ +__builtin_rvv_vssubu_vv_u16mf4((vuint16mf4_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vssubu_vv_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssubu_vv_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vssubu_vv_u32m1(op0, op1, op2) \ +__builtin_rvv_vssubu_vv_u32m1((vuint32m1_t)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vssubu_vv_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssubu_vv_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vssubu_vv_u32m2(op0, op1, op2) \ +__builtin_rvv_vssubu_vv_u32m2((vuint32m2_t)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vssubu_vv_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssubu_vv_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vssubu_vv_u32m4(op0, op1, op2) \ +__builtin_rvv_vssubu_vv_u32m4((vuint32m4_t)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vssubu_vv_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssubu_vv_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vssubu_vv_u32m8(op0, op1, op2) \ +__builtin_rvv_vssubu_vv_u32m8((vuint32m8_t)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vssubu_vv_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssubu_vv_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vssubu_vv_u32mf2(op0, op1, op2) \ +__builtin_rvv_vssubu_vv_u32mf2((vuint32mf2_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vssubu_vv_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssubu_vv_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vssubu_vv_u64m1(op0, op1, op2) \ +__builtin_rvv_vssubu_vv_u64m1((vuint64m1_t)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vssubu_vv_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssubu_vv_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vssubu_vv_u64m2(op0, op1, op2) \ +__builtin_rvv_vssubu_vv_u64m2((vuint64m2_t)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vssubu_vv_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssubu_vv_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vssubu_vv_u64m4(op0, op1, op2) \ +__builtin_rvv_vssubu_vv_u64m4((vuint64m4_t)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vssubu_vv_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssubu_vv_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vssubu_vv_u64m8(op0, op1, op2) \ +__builtin_rvv_vssubu_vv_u64m8((vuint64m8_t)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vssubu_vv_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssubu_vv_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vssubu_vx_u8m1(op0, op1, op2) \ +__builtin_rvv_vssubu_vx_u8m1((vuint8m1_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vssubu_vx_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssubu_vx_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vssubu_vx_u8m2(op0, op1, op2) \ +__builtin_rvv_vssubu_vx_u8m2((vuint8m2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vssubu_vx_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssubu_vx_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vssubu_vx_u8m4(op0, op1, op2) \ +__builtin_rvv_vssubu_vx_u8m4((vuint8m4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vssubu_vx_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssubu_vx_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vssubu_vx_u8m8(op0, op1, op2) \ +__builtin_rvv_vssubu_vx_u8m8((vuint8m8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vssubu_vx_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssubu_vx_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vssubu_vx_u8mf2(op0, op1, op2) \ +__builtin_rvv_vssubu_vx_u8mf2((vuint8mf2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vssubu_vx_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssubu_vx_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vssubu_vx_u8mf4(op0, op1, op2) \ +__builtin_rvv_vssubu_vx_u8mf4((vuint8mf4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vssubu_vx_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssubu_vx_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vssubu_vx_u8mf8(op0, op1, op2) \ +__builtin_rvv_vssubu_vx_u8mf8((vuint8mf8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vssubu_vx_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssubu_vx_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vssubu_vx_u16m1(op0, op1, op2) \ +__builtin_rvv_vssubu_vx_u16m1((vuint16m1_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vssubu_vx_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssubu_vx_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vssubu_vx_u16m2(op0, op1, op2) \ +__builtin_rvv_vssubu_vx_u16m2((vuint16m2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vssubu_vx_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssubu_vx_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vssubu_vx_u16m4(op0, op1, op2) \ +__builtin_rvv_vssubu_vx_u16m4((vuint16m4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vssubu_vx_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssubu_vx_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vssubu_vx_u16m8(op0, op1, op2) \ +__builtin_rvv_vssubu_vx_u16m8((vuint16m8_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vssubu_vx_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssubu_vx_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vssubu_vx_u16mf2(op0, op1, op2) \ +__builtin_rvv_vssubu_vx_u16mf2((vuint16mf2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vssubu_vx_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssubu_vx_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vssubu_vx_u16mf4(op0, op1, op2) \ +__builtin_rvv_vssubu_vx_u16mf4((vuint16mf4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vssubu_vx_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssubu_vx_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vssubu_vx_u32m1(op0, op1, op2) \ +__builtin_rvv_vssubu_vx_u32m1((vuint32m1_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vssubu_vx_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssubu_vx_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vssubu_vx_u32m2(op0, op1, op2) \ +__builtin_rvv_vssubu_vx_u32m2((vuint32m2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vssubu_vx_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssubu_vx_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vssubu_vx_u32m4(op0, op1, op2) \ +__builtin_rvv_vssubu_vx_u32m4((vuint32m4_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vssubu_vx_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssubu_vx_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vssubu_vx_u32m8(op0, op1, op2) \ +__builtin_rvv_vssubu_vx_u32m8((vuint32m8_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vssubu_vx_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssubu_vx_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vssubu_vx_u32mf2(op0, op1, op2) \ +__builtin_rvv_vssubu_vx_u32mf2((vuint32mf2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vssubu_vx_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssubu_vx_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vssubu_vx_u64m1(op0, op1, op2) \ +__builtin_rvv_vssubu_vx_u64m1((vuint64m1_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vssubu_vx_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssubu_vx_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vssubu_vx_u64m2(op0, op1, op2) \ +__builtin_rvv_vssubu_vx_u64m2((vuint64m2_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vssubu_vx_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssubu_vx_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vssubu_vx_u64m4(op0, op1, op2) \ +__builtin_rvv_vssubu_vx_u64m4((vuint64m4_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vssubu_vx_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssubu_vx_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vssubu_vx_u64m8(op0, op1, op2) \ +__builtin_rvv_vssubu_vx_u64m8((vuint64m8_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vssubu_vx_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssubu_vx_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vssub_vv_i8m1(op0, op1, op2) \ +__builtin_rvv_vssub_vv_i8m1((vint8m1_t)(op0), (vint8m1_t)(op1), (size_t)(op2)) +#define vssub_vv_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssub_vv_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vssub_vv_i8m2(op0, op1, op2) \ +__builtin_rvv_vssub_vv_i8m2((vint8m2_t)(op0), (vint8m2_t)(op1), (size_t)(op2)) +#define vssub_vv_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssub_vv_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vssub_vv_i8m4(op0, op1, op2) \ +__builtin_rvv_vssub_vv_i8m4((vint8m4_t)(op0), (vint8m4_t)(op1), (size_t)(op2)) +#define vssub_vv_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssub_vv_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vssub_vv_i8m8(op0, op1, op2) \ +__builtin_rvv_vssub_vv_i8m8((vint8m8_t)(op0), (vint8m8_t)(op1), (size_t)(op2)) +#define vssub_vv_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssub_vv_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (vint8m8_t)(op3), (size_t)(op4)) +#define vssub_vv_i8mf2(op0, op1, op2) \ +__builtin_rvv_vssub_vv_i8mf2((vint8mf2_t)(op0), (vint8mf2_t)(op1), (size_t)(op2)) +#define vssub_vv_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssub_vv_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vssub_vv_i8mf4(op0, op1, op2) \ +__builtin_rvv_vssub_vv_i8mf4((vint8mf4_t)(op0), (vint8mf4_t)(op1), (size_t)(op2)) +#define vssub_vv_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssub_vv_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vssub_vv_i8mf8(op0, op1, op2) \ +__builtin_rvv_vssub_vv_i8mf8((vint8mf8_t)(op0), (vint8mf8_t)(op1), (size_t)(op2)) +#define vssub_vv_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssub_vv_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vssub_vv_i16m1(op0, op1, op2) \ +__builtin_rvv_vssub_vv_i16m1((vint16m1_t)(op0), (vint16m1_t)(op1), (size_t)(op2)) +#define vssub_vv_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssub_vv_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vssub_vv_i16m2(op0, op1, op2) \ +__builtin_rvv_vssub_vv_i16m2((vint16m2_t)(op0), (vint16m2_t)(op1), (size_t)(op2)) +#define vssub_vv_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssub_vv_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vssub_vv_i16m4(op0, op1, op2) \ +__builtin_rvv_vssub_vv_i16m4((vint16m4_t)(op0), (vint16m4_t)(op1), (size_t)(op2)) +#define vssub_vv_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssub_vv_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vssub_vv_i16m8(op0, op1, op2) \ +__builtin_rvv_vssub_vv_i16m8((vint16m8_t)(op0), (vint16m8_t)(op1), (size_t)(op2)) +#define vssub_vv_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssub_vv_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (vint16m8_t)(op3), (size_t)(op4)) +#define vssub_vv_i16mf2(op0, op1, op2) \ +__builtin_rvv_vssub_vv_i16mf2((vint16mf2_t)(op0), (vint16mf2_t)(op1), (size_t)(op2)) +#define vssub_vv_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssub_vv_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vssub_vv_i16mf4(op0, op1, op2) \ +__builtin_rvv_vssub_vv_i16mf4((vint16mf4_t)(op0), (vint16mf4_t)(op1), (size_t)(op2)) +#define vssub_vv_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssub_vv_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vssub_vv_i32m1(op0, op1, op2) \ +__builtin_rvv_vssub_vv_i32m1((vint32m1_t)(op0), (vint32m1_t)(op1), (size_t)(op2)) +#define vssub_vv_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssub_vv_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vssub_vv_i32m2(op0, op1, op2) \ +__builtin_rvv_vssub_vv_i32m2((vint32m2_t)(op0), (vint32m2_t)(op1), (size_t)(op2)) +#define vssub_vv_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssub_vv_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vssub_vv_i32m4(op0, op1, op2) \ +__builtin_rvv_vssub_vv_i32m4((vint32m4_t)(op0), (vint32m4_t)(op1), (size_t)(op2)) +#define vssub_vv_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssub_vv_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vssub_vv_i32m8(op0, op1, op2) \ +__builtin_rvv_vssub_vv_i32m8((vint32m8_t)(op0), (vint32m8_t)(op1), (size_t)(op2)) +#define vssub_vv_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssub_vv_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vssub_vv_i32mf2(op0, op1, op2) \ +__builtin_rvv_vssub_vv_i32mf2((vint32mf2_t)(op0), (vint32mf2_t)(op1), (size_t)(op2)) +#define vssub_vv_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssub_vv_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vssub_vv_i64m1(op0, op1, op2) \ +__builtin_rvv_vssub_vv_i64m1((vint64m1_t)(op0), (vint64m1_t)(op1), (size_t)(op2)) +#define vssub_vv_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssub_vv_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vssub_vv_i64m2(op0, op1, op2) \ +__builtin_rvv_vssub_vv_i64m2((vint64m2_t)(op0), (vint64m2_t)(op1), (size_t)(op2)) +#define vssub_vv_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssub_vv_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vssub_vv_i64m4(op0, op1, op2) \ +__builtin_rvv_vssub_vv_i64m4((vint64m4_t)(op0), (vint64m4_t)(op1), (size_t)(op2)) +#define vssub_vv_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssub_vv_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vssub_vv_i64m8(op0, op1, op2) \ +__builtin_rvv_vssub_vv_i64m8((vint64m8_t)(op0), (vint64m8_t)(op1), (size_t)(op2)) +#define vssub_vv_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssub_vv_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vssub_vx_i8m1(op0, op1, op2) \ +__builtin_rvv_vssub_vx_i8m1((vint8m1_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vssub_vx_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssub_vx_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vssub_vx_i8m2(op0, op1, op2) \ +__builtin_rvv_vssub_vx_i8m2((vint8m2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vssub_vx_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssub_vx_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vssub_vx_i8m4(op0, op1, op2) \ +__builtin_rvv_vssub_vx_i8m4((vint8m4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vssub_vx_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssub_vx_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vssub_vx_i8m8(op0, op1, op2) \ +__builtin_rvv_vssub_vx_i8m8((vint8m8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vssub_vx_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssub_vx_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vssub_vx_i8mf2(op0, op1, op2) \ +__builtin_rvv_vssub_vx_i8mf2((vint8mf2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vssub_vx_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssub_vx_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vssub_vx_i8mf4(op0, op1, op2) \ +__builtin_rvv_vssub_vx_i8mf4((vint8mf4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vssub_vx_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssub_vx_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vssub_vx_i8mf8(op0, op1, op2) \ +__builtin_rvv_vssub_vx_i8mf8((vint8mf8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vssub_vx_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssub_vx_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vssub_vx_i16m1(op0, op1, op2) \ +__builtin_rvv_vssub_vx_i16m1((vint16m1_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vssub_vx_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssub_vx_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vssub_vx_i16m2(op0, op1, op2) \ +__builtin_rvv_vssub_vx_i16m2((vint16m2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vssub_vx_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssub_vx_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vssub_vx_i16m4(op0, op1, op2) \ +__builtin_rvv_vssub_vx_i16m4((vint16m4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vssub_vx_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssub_vx_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vssub_vx_i16m8(op0, op1, op2) \ +__builtin_rvv_vssub_vx_i16m8((vint16m8_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vssub_vx_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssub_vx_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vssub_vx_i16mf2(op0, op1, op2) \ +__builtin_rvv_vssub_vx_i16mf2((vint16mf2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vssub_vx_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssub_vx_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vssub_vx_i16mf4(op0, op1, op2) \ +__builtin_rvv_vssub_vx_i16mf4((vint16mf4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vssub_vx_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssub_vx_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vssub_vx_i32m1(op0, op1, op2) \ +__builtin_rvv_vssub_vx_i32m1((vint32m1_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vssub_vx_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssub_vx_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vssub_vx_i32m2(op0, op1, op2) \ +__builtin_rvv_vssub_vx_i32m2((vint32m2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vssub_vx_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssub_vx_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vssub_vx_i32m4(op0, op1, op2) \ +__builtin_rvv_vssub_vx_i32m4((vint32m4_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vssub_vx_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssub_vx_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vssub_vx_i32m8(op0, op1, op2) \ +__builtin_rvv_vssub_vx_i32m8((vint32m8_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vssub_vx_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssub_vx_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vssub_vx_i32mf2(op0, op1, op2) \ +__builtin_rvv_vssub_vx_i32mf2((vint32mf2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vssub_vx_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssub_vx_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vssub_vx_i64m1(op0, op1, op2) \ +__builtin_rvv_vssub_vx_i64m1((vint64m1_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vssub_vx_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssub_vx_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vssub_vx_i64m2(op0, op1, op2) \ +__builtin_rvv_vssub_vx_i64m2((vint64m2_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vssub_vx_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssub_vx_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vssub_vx_i64m4(op0, op1, op2) \ +__builtin_rvv_vssub_vx_i64m4((vint64m4_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vssub_vx_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssub_vx_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vssub_vx_i64m8(op0, op1, op2) \ +__builtin_rvv_vssub_vx_i64m8((vint64m8_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vssub_vx_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssub_vx_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vaaddu_vv_u8m1(op0, op1, op2) \ +__builtin_rvv_vaaddu_vv_u8m1((vuint8m1_t)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vaaddu_vv_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaaddu_vv_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vaaddu_vv_u8m2(op0, op1, op2) \ +__builtin_rvv_vaaddu_vv_u8m2((vuint8m2_t)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vaaddu_vv_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaaddu_vv_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vaaddu_vv_u8m4(op0, op1, op2) \ +__builtin_rvv_vaaddu_vv_u8m4((vuint8m4_t)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vaaddu_vv_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaaddu_vv_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vaaddu_vv_u8m8(op0, op1, op2) \ +__builtin_rvv_vaaddu_vv_u8m8((vuint8m8_t)(op0), (vuint8m8_t)(op1), (size_t)(op2)) +#define vaaddu_vv_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaaddu_vv_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (vuint8m8_t)(op3), (size_t)(op4)) +#define vaaddu_vv_u8mf2(op0, op1, op2) \ +__builtin_rvv_vaaddu_vv_u8mf2((vuint8mf2_t)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vaaddu_vv_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaaddu_vv_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vaaddu_vv_u8mf4(op0, op1, op2) \ +__builtin_rvv_vaaddu_vv_u8mf4((vuint8mf4_t)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vaaddu_vv_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaaddu_vv_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vaaddu_vv_u8mf8(op0, op1, op2) \ +__builtin_rvv_vaaddu_vv_u8mf8((vuint8mf8_t)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vaaddu_vv_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaaddu_vv_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vaaddu_vv_u16m1(op0, op1, op2) \ +__builtin_rvv_vaaddu_vv_u16m1((vuint16m1_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vaaddu_vv_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaaddu_vv_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vaaddu_vv_u16m2(op0, op1, op2) \ +__builtin_rvv_vaaddu_vv_u16m2((vuint16m2_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vaaddu_vv_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaaddu_vv_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vaaddu_vv_u16m4(op0, op1, op2) \ +__builtin_rvv_vaaddu_vv_u16m4((vuint16m4_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vaaddu_vv_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaaddu_vv_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vaaddu_vv_u16m8(op0, op1, op2) \ +__builtin_rvv_vaaddu_vv_u16m8((vuint16m8_t)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vaaddu_vv_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaaddu_vv_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vaaddu_vv_u16mf2(op0, op1, op2) \ +__builtin_rvv_vaaddu_vv_u16mf2((vuint16mf2_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vaaddu_vv_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaaddu_vv_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vaaddu_vv_u16mf4(op0, op1, op2) \ +__builtin_rvv_vaaddu_vv_u16mf4((vuint16mf4_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vaaddu_vv_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaaddu_vv_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vaaddu_vv_u32m1(op0, op1, op2) \ +__builtin_rvv_vaaddu_vv_u32m1((vuint32m1_t)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vaaddu_vv_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaaddu_vv_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vaaddu_vv_u32m2(op0, op1, op2) \ +__builtin_rvv_vaaddu_vv_u32m2((vuint32m2_t)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vaaddu_vv_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaaddu_vv_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vaaddu_vv_u32m4(op0, op1, op2) \ +__builtin_rvv_vaaddu_vv_u32m4((vuint32m4_t)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vaaddu_vv_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaaddu_vv_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vaaddu_vv_u32m8(op0, op1, op2) \ +__builtin_rvv_vaaddu_vv_u32m8((vuint32m8_t)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vaaddu_vv_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaaddu_vv_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vaaddu_vv_u32mf2(op0, op1, op2) \ +__builtin_rvv_vaaddu_vv_u32mf2((vuint32mf2_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vaaddu_vv_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaaddu_vv_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vaaddu_vv_u64m1(op0, op1, op2) \ +__builtin_rvv_vaaddu_vv_u64m1((vuint64m1_t)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vaaddu_vv_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaaddu_vv_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vaaddu_vv_u64m2(op0, op1, op2) \ +__builtin_rvv_vaaddu_vv_u64m2((vuint64m2_t)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vaaddu_vv_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaaddu_vv_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vaaddu_vv_u64m4(op0, op1, op2) \ +__builtin_rvv_vaaddu_vv_u64m4((vuint64m4_t)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vaaddu_vv_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaaddu_vv_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vaaddu_vv_u64m8(op0, op1, op2) \ +__builtin_rvv_vaaddu_vv_u64m8((vuint64m8_t)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vaaddu_vv_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaaddu_vv_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vaaddu_vx_u8m1(op0, op1, op2) \ +__builtin_rvv_vaaddu_vx_u8m1((vuint8m1_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vaaddu_vx_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaaddu_vx_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vaaddu_vx_u8m2(op0, op1, op2) \ +__builtin_rvv_vaaddu_vx_u8m2((vuint8m2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vaaddu_vx_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaaddu_vx_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vaaddu_vx_u8m4(op0, op1, op2) \ +__builtin_rvv_vaaddu_vx_u8m4((vuint8m4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vaaddu_vx_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaaddu_vx_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vaaddu_vx_u8m8(op0, op1, op2) \ +__builtin_rvv_vaaddu_vx_u8m8((vuint8m8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vaaddu_vx_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaaddu_vx_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vaaddu_vx_u8mf2(op0, op1, op2) \ +__builtin_rvv_vaaddu_vx_u8mf2((vuint8mf2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vaaddu_vx_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaaddu_vx_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vaaddu_vx_u8mf4(op0, op1, op2) \ +__builtin_rvv_vaaddu_vx_u8mf4((vuint8mf4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vaaddu_vx_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaaddu_vx_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vaaddu_vx_u8mf8(op0, op1, op2) \ +__builtin_rvv_vaaddu_vx_u8mf8((vuint8mf8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vaaddu_vx_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaaddu_vx_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vaaddu_vx_u16m1(op0, op1, op2) \ +__builtin_rvv_vaaddu_vx_u16m1((vuint16m1_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vaaddu_vx_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaaddu_vx_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vaaddu_vx_u16m2(op0, op1, op2) \ +__builtin_rvv_vaaddu_vx_u16m2((vuint16m2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vaaddu_vx_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaaddu_vx_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vaaddu_vx_u16m4(op0, op1, op2) \ +__builtin_rvv_vaaddu_vx_u16m4((vuint16m4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vaaddu_vx_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaaddu_vx_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vaaddu_vx_u16m8(op0, op1, op2) \ +__builtin_rvv_vaaddu_vx_u16m8((vuint16m8_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vaaddu_vx_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaaddu_vx_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vaaddu_vx_u16mf2(op0, op1, op2) \ +__builtin_rvv_vaaddu_vx_u16mf2((vuint16mf2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vaaddu_vx_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaaddu_vx_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vaaddu_vx_u16mf4(op0, op1, op2) \ +__builtin_rvv_vaaddu_vx_u16mf4((vuint16mf4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vaaddu_vx_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaaddu_vx_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vaaddu_vx_u32m1(op0, op1, op2) \ +__builtin_rvv_vaaddu_vx_u32m1((vuint32m1_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vaaddu_vx_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaaddu_vx_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vaaddu_vx_u32m2(op0, op1, op2) \ +__builtin_rvv_vaaddu_vx_u32m2((vuint32m2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vaaddu_vx_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaaddu_vx_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vaaddu_vx_u32m4(op0, op1, op2) \ +__builtin_rvv_vaaddu_vx_u32m4((vuint32m4_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vaaddu_vx_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaaddu_vx_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vaaddu_vx_u32m8(op0, op1, op2) \ +__builtin_rvv_vaaddu_vx_u32m8((vuint32m8_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vaaddu_vx_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaaddu_vx_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vaaddu_vx_u32mf2(op0, op1, op2) \ +__builtin_rvv_vaaddu_vx_u32mf2((vuint32mf2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vaaddu_vx_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaaddu_vx_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vaaddu_vx_u64m1(op0, op1, op2) \ +__builtin_rvv_vaaddu_vx_u64m1((vuint64m1_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vaaddu_vx_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaaddu_vx_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vaaddu_vx_u64m2(op0, op1, op2) \ +__builtin_rvv_vaaddu_vx_u64m2((vuint64m2_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vaaddu_vx_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaaddu_vx_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vaaddu_vx_u64m4(op0, op1, op2) \ +__builtin_rvv_vaaddu_vx_u64m4((vuint64m4_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vaaddu_vx_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaaddu_vx_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vaaddu_vx_u64m8(op0, op1, op2) \ +__builtin_rvv_vaaddu_vx_u64m8((vuint64m8_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vaaddu_vx_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaaddu_vx_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vaadd_vv_i8m1(op0, op1, op2) \ +__builtin_rvv_vaadd_vv_i8m1((vint8m1_t)(op0), (vint8m1_t)(op1), (size_t)(op2)) +#define vaadd_vv_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaadd_vv_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vaadd_vv_i8m2(op0, op1, op2) \ +__builtin_rvv_vaadd_vv_i8m2((vint8m2_t)(op0), (vint8m2_t)(op1), (size_t)(op2)) +#define vaadd_vv_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaadd_vv_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vaadd_vv_i8m4(op0, op1, op2) \ +__builtin_rvv_vaadd_vv_i8m4((vint8m4_t)(op0), (vint8m4_t)(op1), (size_t)(op2)) +#define vaadd_vv_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaadd_vv_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vaadd_vv_i8m8(op0, op1, op2) \ +__builtin_rvv_vaadd_vv_i8m8((vint8m8_t)(op0), (vint8m8_t)(op1), (size_t)(op2)) +#define vaadd_vv_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaadd_vv_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (vint8m8_t)(op3), (size_t)(op4)) +#define vaadd_vv_i8mf2(op0, op1, op2) \ +__builtin_rvv_vaadd_vv_i8mf2((vint8mf2_t)(op0), (vint8mf2_t)(op1), (size_t)(op2)) +#define vaadd_vv_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaadd_vv_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vaadd_vv_i8mf4(op0, op1, op2) \ +__builtin_rvv_vaadd_vv_i8mf4((vint8mf4_t)(op0), (vint8mf4_t)(op1), (size_t)(op2)) +#define vaadd_vv_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaadd_vv_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vaadd_vv_i8mf8(op0, op1, op2) \ +__builtin_rvv_vaadd_vv_i8mf8((vint8mf8_t)(op0), (vint8mf8_t)(op1), (size_t)(op2)) +#define vaadd_vv_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaadd_vv_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vaadd_vv_i16m1(op0, op1, op2) \ +__builtin_rvv_vaadd_vv_i16m1((vint16m1_t)(op0), (vint16m1_t)(op1), (size_t)(op2)) +#define vaadd_vv_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaadd_vv_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vaadd_vv_i16m2(op0, op1, op2) \ +__builtin_rvv_vaadd_vv_i16m2((vint16m2_t)(op0), (vint16m2_t)(op1), (size_t)(op2)) +#define vaadd_vv_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaadd_vv_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vaadd_vv_i16m4(op0, op1, op2) \ +__builtin_rvv_vaadd_vv_i16m4((vint16m4_t)(op0), (vint16m4_t)(op1), (size_t)(op2)) +#define vaadd_vv_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaadd_vv_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vaadd_vv_i16m8(op0, op1, op2) \ +__builtin_rvv_vaadd_vv_i16m8((vint16m8_t)(op0), (vint16m8_t)(op1), (size_t)(op2)) +#define vaadd_vv_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaadd_vv_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (vint16m8_t)(op3), (size_t)(op4)) +#define vaadd_vv_i16mf2(op0, op1, op2) \ +__builtin_rvv_vaadd_vv_i16mf2((vint16mf2_t)(op0), (vint16mf2_t)(op1), (size_t)(op2)) +#define vaadd_vv_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaadd_vv_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vaadd_vv_i16mf4(op0, op1, op2) \ +__builtin_rvv_vaadd_vv_i16mf4((vint16mf4_t)(op0), (vint16mf4_t)(op1), (size_t)(op2)) +#define vaadd_vv_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaadd_vv_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vaadd_vv_i32m1(op0, op1, op2) \ +__builtin_rvv_vaadd_vv_i32m1((vint32m1_t)(op0), (vint32m1_t)(op1), (size_t)(op2)) +#define vaadd_vv_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaadd_vv_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vaadd_vv_i32m2(op0, op1, op2) \ +__builtin_rvv_vaadd_vv_i32m2((vint32m2_t)(op0), (vint32m2_t)(op1), (size_t)(op2)) +#define vaadd_vv_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaadd_vv_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vaadd_vv_i32m4(op0, op1, op2) \ +__builtin_rvv_vaadd_vv_i32m4((vint32m4_t)(op0), (vint32m4_t)(op1), (size_t)(op2)) +#define vaadd_vv_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaadd_vv_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vaadd_vv_i32m8(op0, op1, op2) \ +__builtin_rvv_vaadd_vv_i32m8((vint32m8_t)(op0), (vint32m8_t)(op1), (size_t)(op2)) +#define vaadd_vv_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaadd_vv_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vaadd_vv_i32mf2(op0, op1, op2) \ +__builtin_rvv_vaadd_vv_i32mf2((vint32mf2_t)(op0), (vint32mf2_t)(op1), (size_t)(op2)) +#define vaadd_vv_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaadd_vv_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vaadd_vv_i64m1(op0, op1, op2) \ +__builtin_rvv_vaadd_vv_i64m1((vint64m1_t)(op0), (vint64m1_t)(op1), (size_t)(op2)) +#define vaadd_vv_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaadd_vv_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vaadd_vv_i64m2(op0, op1, op2) \ +__builtin_rvv_vaadd_vv_i64m2((vint64m2_t)(op0), (vint64m2_t)(op1), (size_t)(op2)) +#define vaadd_vv_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaadd_vv_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vaadd_vv_i64m4(op0, op1, op2) \ +__builtin_rvv_vaadd_vv_i64m4((vint64m4_t)(op0), (vint64m4_t)(op1), (size_t)(op2)) +#define vaadd_vv_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaadd_vv_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vaadd_vv_i64m8(op0, op1, op2) \ +__builtin_rvv_vaadd_vv_i64m8((vint64m8_t)(op0), (vint64m8_t)(op1), (size_t)(op2)) +#define vaadd_vv_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaadd_vv_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vaadd_vx_i8m1(op0, op1, op2) \ +__builtin_rvv_vaadd_vx_i8m1((vint8m1_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vaadd_vx_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaadd_vx_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vaadd_vx_i8m2(op0, op1, op2) \ +__builtin_rvv_vaadd_vx_i8m2((vint8m2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vaadd_vx_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaadd_vx_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vaadd_vx_i8m4(op0, op1, op2) \ +__builtin_rvv_vaadd_vx_i8m4((vint8m4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vaadd_vx_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaadd_vx_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vaadd_vx_i8m8(op0, op1, op2) \ +__builtin_rvv_vaadd_vx_i8m8((vint8m8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vaadd_vx_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaadd_vx_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vaadd_vx_i8mf2(op0, op1, op2) \ +__builtin_rvv_vaadd_vx_i8mf2((vint8mf2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vaadd_vx_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaadd_vx_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vaadd_vx_i8mf4(op0, op1, op2) \ +__builtin_rvv_vaadd_vx_i8mf4((vint8mf4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vaadd_vx_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaadd_vx_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vaadd_vx_i8mf8(op0, op1, op2) \ +__builtin_rvv_vaadd_vx_i8mf8((vint8mf8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vaadd_vx_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaadd_vx_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vaadd_vx_i16m1(op0, op1, op2) \ +__builtin_rvv_vaadd_vx_i16m1((vint16m1_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vaadd_vx_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaadd_vx_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vaadd_vx_i16m2(op0, op1, op2) \ +__builtin_rvv_vaadd_vx_i16m2((vint16m2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vaadd_vx_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaadd_vx_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vaadd_vx_i16m4(op0, op1, op2) \ +__builtin_rvv_vaadd_vx_i16m4((vint16m4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vaadd_vx_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaadd_vx_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vaadd_vx_i16m8(op0, op1, op2) \ +__builtin_rvv_vaadd_vx_i16m8((vint16m8_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vaadd_vx_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaadd_vx_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vaadd_vx_i16mf2(op0, op1, op2) \ +__builtin_rvv_vaadd_vx_i16mf2((vint16mf2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vaadd_vx_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaadd_vx_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vaadd_vx_i16mf4(op0, op1, op2) \ +__builtin_rvv_vaadd_vx_i16mf4((vint16mf4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vaadd_vx_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaadd_vx_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vaadd_vx_i32m1(op0, op1, op2) \ +__builtin_rvv_vaadd_vx_i32m1((vint32m1_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vaadd_vx_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaadd_vx_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vaadd_vx_i32m2(op0, op1, op2) \ +__builtin_rvv_vaadd_vx_i32m2((vint32m2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vaadd_vx_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaadd_vx_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vaadd_vx_i32m4(op0, op1, op2) \ +__builtin_rvv_vaadd_vx_i32m4((vint32m4_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vaadd_vx_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaadd_vx_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vaadd_vx_i32m8(op0, op1, op2) \ +__builtin_rvv_vaadd_vx_i32m8((vint32m8_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vaadd_vx_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaadd_vx_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vaadd_vx_i32mf2(op0, op1, op2) \ +__builtin_rvv_vaadd_vx_i32mf2((vint32mf2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vaadd_vx_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaadd_vx_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vaadd_vx_i64m1(op0, op1, op2) \ +__builtin_rvv_vaadd_vx_i64m1((vint64m1_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vaadd_vx_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaadd_vx_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vaadd_vx_i64m2(op0, op1, op2) \ +__builtin_rvv_vaadd_vx_i64m2((vint64m2_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vaadd_vx_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaadd_vx_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vaadd_vx_i64m4(op0, op1, op2) \ +__builtin_rvv_vaadd_vx_i64m4((vint64m4_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vaadd_vx_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaadd_vx_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vaadd_vx_i64m8(op0, op1, op2) \ +__builtin_rvv_vaadd_vx_i64m8((vint64m8_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vaadd_vx_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vaadd_vx_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vasubu_vv_u8m1(op0, op1, op2) \ +__builtin_rvv_vasubu_vv_u8m1((vuint8m1_t)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vasubu_vv_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasubu_vv_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vasubu_vv_u8m2(op0, op1, op2) \ +__builtin_rvv_vasubu_vv_u8m2((vuint8m2_t)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vasubu_vv_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasubu_vv_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vasubu_vv_u8m4(op0, op1, op2) \ +__builtin_rvv_vasubu_vv_u8m4((vuint8m4_t)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vasubu_vv_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasubu_vv_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vasubu_vv_u8m8(op0, op1, op2) \ +__builtin_rvv_vasubu_vv_u8m8((vuint8m8_t)(op0), (vuint8m8_t)(op1), (size_t)(op2)) +#define vasubu_vv_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasubu_vv_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (vuint8m8_t)(op3), (size_t)(op4)) +#define vasubu_vv_u8mf2(op0, op1, op2) \ +__builtin_rvv_vasubu_vv_u8mf2((vuint8mf2_t)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vasubu_vv_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasubu_vv_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vasubu_vv_u8mf4(op0, op1, op2) \ +__builtin_rvv_vasubu_vv_u8mf4((vuint8mf4_t)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vasubu_vv_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasubu_vv_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vasubu_vv_u8mf8(op0, op1, op2) \ +__builtin_rvv_vasubu_vv_u8mf8((vuint8mf8_t)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vasubu_vv_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasubu_vv_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vasubu_vv_u16m1(op0, op1, op2) \ +__builtin_rvv_vasubu_vv_u16m1((vuint16m1_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vasubu_vv_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasubu_vv_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vasubu_vv_u16m2(op0, op1, op2) \ +__builtin_rvv_vasubu_vv_u16m2((vuint16m2_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vasubu_vv_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasubu_vv_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vasubu_vv_u16m4(op0, op1, op2) \ +__builtin_rvv_vasubu_vv_u16m4((vuint16m4_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vasubu_vv_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasubu_vv_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vasubu_vv_u16m8(op0, op1, op2) \ +__builtin_rvv_vasubu_vv_u16m8((vuint16m8_t)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vasubu_vv_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasubu_vv_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vasubu_vv_u16mf2(op0, op1, op2) \ +__builtin_rvv_vasubu_vv_u16mf2((vuint16mf2_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vasubu_vv_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasubu_vv_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vasubu_vv_u16mf4(op0, op1, op2) \ +__builtin_rvv_vasubu_vv_u16mf4((vuint16mf4_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vasubu_vv_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasubu_vv_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vasubu_vv_u32m1(op0, op1, op2) \ +__builtin_rvv_vasubu_vv_u32m1((vuint32m1_t)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vasubu_vv_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasubu_vv_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vasubu_vv_u32m2(op0, op1, op2) \ +__builtin_rvv_vasubu_vv_u32m2((vuint32m2_t)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vasubu_vv_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasubu_vv_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vasubu_vv_u32m4(op0, op1, op2) \ +__builtin_rvv_vasubu_vv_u32m4((vuint32m4_t)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vasubu_vv_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasubu_vv_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vasubu_vv_u32m8(op0, op1, op2) \ +__builtin_rvv_vasubu_vv_u32m8((vuint32m8_t)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vasubu_vv_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasubu_vv_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vasubu_vv_u32mf2(op0, op1, op2) \ +__builtin_rvv_vasubu_vv_u32mf2((vuint32mf2_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vasubu_vv_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasubu_vv_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vasubu_vv_u64m1(op0, op1, op2) \ +__builtin_rvv_vasubu_vv_u64m1((vuint64m1_t)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vasubu_vv_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasubu_vv_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vasubu_vv_u64m2(op0, op1, op2) \ +__builtin_rvv_vasubu_vv_u64m2((vuint64m2_t)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vasubu_vv_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasubu_vv_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vasubu_vv_u64m4(op0, op1, op2) \ +__builtin_rvv_vasubu_vv_u64m4((vuint64m4_t)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vasubu_vv_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasubu_vv_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vasubu_vv_u64m8(op0, op1, op2) \ +__builtin_rvv_vasubu_vv_u64m8((vuint64m8_t)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vasubu_vv_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasubu_vv_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vasubu_vx_u8m1(op0, op1, op2) \ +__builtin_rvv_vasubu_vx_u8m1((vuint8m1_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vasubu_vx_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasubu_vx_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vasubu_vx_u8m2(op0, op1, op2) \ +__builtin_rvv_vasubu_vx_u8m2((vuint8m2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vasubu_vx_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasubu_vx_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vasubu_vx_u8m4(op0, op1, op2) \ +__builtin_rvv_vasubu_vx_u8m4((vuint8m4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vasubu_vx_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasubu_vx_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vasubu_vx_u8m8(op0, op1, op2) \ +__builtin_rvv_vasubu_vx_u8m8((vuint8m8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vasubu_vx_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasubu_vx_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vasubu_vx_u8mf2(op0, op1, op2) \ +__builtin_rvv_vasubu_vx_u8mf2((vuint8mf2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vasubu_vx_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasubu_vx_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vasubu_vx_u8mf4(op0, op1, op2) \ +__builtin_rvv_vasubu_vx_u8mf4((vuint8mf4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vasubu_vx_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasubu_vx_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vasubu_vx_u8mf8(op0, op1, op2) \ +__builtin_rvv_vasubu_vx_u8mf8((vuint8mf8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vasubu_vx_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasubu_vx_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (uint8_t)(op3), (size_t)(op4)) +#define vasubu_vx_u16m1(op0, op1, op2) \ +__builtin_rvv_vasubu_vx_u16m1((vuint16m1_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vasubu_vx_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasubu_vx_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vasubu_vx_u16m2(op0, op1, op2) \ +__builtin_rvv_vasubu_vx_u16m2((vuint16m2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vasubu_vx_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasubu_vx_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vasubu_vx_u16m4(op0, op1, op2) \ +__builtin_rvv_vasubu_vx_u16m4((vuint16m4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vasubu_vx_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasubu_vx_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vasubu_vx_u16m8(op0, op1, op2) \ +__builtin_rvv_vasubu_vx_u16m8((vuint16m8_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vasubu_vx_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasubu_vx_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vasubu_vx_u16mf2(op0, op1, op2) \ +__builtin_rvv_vasubu_vx_u16mf2((vuint16mf2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vasubu_vx_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasubu_vx_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vasubu_vx_u16mf4(op0, op1, op2) \ +__builtin_rvv_vasubu_vx_u16mf4((vuint16mf4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vasubu_vx_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasubu_vx_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (uint16_t)(op3), (size_t)(op4)) +#define vasubu_vx_u32m1(op0, op1, op2) \ +__builtin_rvv_vasubu_vx_u32m1((vuint32m1_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vasubu_vx_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasubu_vx_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vasubu_vx_u32m2(op0, op1, op2) \ +__builtin_rvv_vasubu_vx_u32m2((vuint32m2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vasubu_vx_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasubu_vx_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vasubu_vx_u32m4(op0, op1, op2) \ +__builtin_rvv_vasubu_vx_u32m4((vuint32m4_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vasubu_vx_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasubu_vx_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vasubu_vx_u32m8(op0, op1, op2) \ +__builtin_rvv_vasubu_vx_u32m8((vuint32m8_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vasubu_vx_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasubu_vx_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vasubu_vx_u32mf2(op0, op1, op2) \ +__builtin_rvv_vasubu_vx_u32mf2((vuint32mf2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vasubu_vx_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasubu_vx_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (uint32_t)(op3), (size_t)(op4)) +#define vasubu_vx_u64m1(op0, op1, op2) \ +__builtin_rvv_vasubu_vx_u64m1((vuint64m1_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vasubu_vx_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasubu_vx_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vasubu_vx_u64m2(op0, op1, op2) \ +__builtin_rvv_vasubu_vx_u64m2((vuint64m2_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vasubu_vx_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasubu_vx_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vasubu_vx_u64m4(op0, op1, op2) \ +__builtin_rvv_vasubu_vx_u64m4((vuint64m4_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vasubu_vx_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasubu_vx_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vasubu_vx_u64m8(op0, op1, op2) \ +__builtin_rvv_vasubu_vx_u64m8((vuint64m8_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vasubu_vx_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasubu_vx_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (uint64_t)(op3), (size_t)(op4)) +#define vsuxei8_v_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_i8m1((int8_t *)(op0), (vuint8m1_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vsuxei8_v_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vsuxei8_v_i8m2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_i8m2((int8_t *)(op0), (vuint8m2_t)(op1), (vint8m2_t)(op2), (size_t)(op3)) +#define vsuxei8_v_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_i8m2_m((vbool4_t)(op0), (int8_t *)(op1), (vuint8m2_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vsuxei8_v_i8m4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_i8m4((int8_t *)(op0), (vuint8m4_t)(op1), (vint8m4_t)(op2), (size_t)(op3)) +#define vsuxei8_v_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_i8m4_m((vbool2_t)(op0), (int8_t *)(op1), (vuint8m4_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vsuxei8_v_i8m8(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_i8m8((int8_t *)(op0), (vuint8m8_t)(op1), (vint8m8_t)(op2), (size_t)(op3)) +#define vsuxei8_v_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_i8m8_m((vbool1_t)(op0), (int8_t *)(op1), (vuint8m8_t)(op2), (vint8m8_t)(op3), (size_t)(op4)) +#define vsuxei8_v_i8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_i8mf2((int8_t *)(op0), (vuint8mf2_t)(op1), (vint8mf2_t)(op2), (size_t)(op3)) +#define vsuxei8_v_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint8mf2_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vsuxei8_v_i8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_i8mf4((int8_t *)(op0), (vuint8mf4_t)(op1), (vint8mf4_t)(op2), (size_t)(op3)) +#define vsuxei8_v_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint8mf4_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vsuxei8_v_i8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_i8mf8((int8_t *)(op0), (vuint8mf8_t)(op1), (vint8mf8_t)(op2), (size_t)(op3)) +#define vsuxei8_v_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint8mf8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vsuxei64_v_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei64_v_i32m1((int32_t *)(op0), (vuint64m2_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vsuxei64_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei64_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint64m2_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vsuxei64_v_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei64_v_i32m2((int32_t *)(op0), (vuint64m4_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vsuxei64_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei64_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint64m4_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vsuxei64_v_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei64_v_i32m4((int32_t *)(op0), (vuint64m8_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vsuxei64_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei64_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vuint64m8_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vsuxei64_v_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei64_v_i32mf2((int32_t *)(op0), (vuint64m1_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vsuxei64_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei64_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint64m1_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vasub_vv_i8m1(op0, op1, op2) \ +__builtin_rvv_vasub_vv_i8m1((vint8m1_t)(op0), (vint8m1_t)(op1), (size_t)(op2)) +#define vasub_vv_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasub_vv_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vasub_vv_i8m2(op0, op1, op2) \ +__builtin_rvv_vasub_vv_i8m2((vint8m2_t)(op0), (vint8m2_t)(op1), (size_t)(op2)) +#define vasub_vv_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasub_vv_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vasub_vv_i8m4(op0, op1, op2) \ +__builtin_rvv_vasub_vv_i8m4((vint8m4_t)(op0), (vint8m4_t)(op1), (size_t)(op2)) +#define vasub_vv_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasub_vv_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vasub_vv_i8m8(op0, op1, op2) \ +__builtin_rvv_vasub_vv_i8m8((vint8m8_t)(op0), (vint8m8_t)(op1), (size_t)(op2)) +#define vasub_vv_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasub_vv_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (vint8m8_t)(op3), (size_t)(op4)) +#define vasub_vv_i8mf2(op0, op1, op2) \ +__builtin_rvv_vasub_vv_i8mf2((vint8mf2_t)(op0), (vint8mf2_t)(op1), (size_t)(op2)) +#define vasub_vv_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasub_vv_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vasub_vv_i8mf4(op0, op1, op2) \ +__builtin_rvv_vasub_vv_i8mf4((vint8mf4_t)(op0), (vint8mf4_t)(op1), (size_t)(op2)) +#define vasub_vv_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasub_vv_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vasub_vv_i8mf8(op0, op1, op2) \ +__builtin_rvv_vasub_vv_i8mf8((vint8mf8_t)(op0), (vint8mf8_t)(op1), (size_t)(op2)) +#define vasub_vv_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasub_vv_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vasub_vv_i16m1(op0, op1, op2) \ +__builtin_rvv_vasub_vv_i16m1((vint16m1_t)(op0), (vint16m1_t)(op1), (size_t)(op2)) +#define vasub_vv_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasub_vv_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vasub_vv_i16m2(op0, op1, op2) \ +__builtin_rvv_vasub_vv_i16m2((vint16m2_t)(op0), (vint16m2_t)(op1), (size_t)(op2)) +#define vasub_vv_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasub_vv_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vasub_vv_i16m4(op0, op1, op2) \ +__builtin_rvv_vasub_vv_i16m4((vint16m4_t)(op0), (vint16m4_t)(op1), (size_t)(op2)) +#define vasub_vv_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasub_vv_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vasub_vv_i16m8(op0, op1, op2) \ +__builtin_rvv_vasub_vv_i16m8((vint16m8_t)(op0), (vint16m8_t)(op1), (size_t)(op2)) +#define vasub_vv_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasub_vv_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (vint16m8_t)(op3), (size_t)(op4)) +#define vasub_vv_i16mf2(op0, op1, op2) \ +__builtin_rvv_vasub_vv_i16mf2((vint16mf2_t)(op0), (vint16mf2_t)(op1), (size_t)(op2)) +#define vasub_vv_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasub_vv_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vasub_vv_i16mf4(op0, op1, op2) \ +__builtin_rvv_vasub_vv_i16mf4((vint16mf4_t)(op0), (vint16mf4_t)(op1), (size_t)(op2)) +#define vasub_vv_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasub_vv_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vasub_vv_i32m1(op0, op1, op2) \ +__builtin_rvv_vasub_vv_i32m1((vint32m1_t)(op0), (vint32m1_t)(op1), (size_t)(op2)) +#define vasub_vv_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasub_vv_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vasub_vv_i32m2(op0, op1, op2) \ +__builtin_rvv_vasub_vv_i32m2((vint32m2_t)(op0), (vint32m2_t)(op1), (size_t)(op2)) +#define vasub_vv_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasub_vv_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vasub_vv_i32m4(op0, op1, op2) \ +__builtin_rvv_vasub_vv_i32m4((vint32m4_t)(op0), (vint32m4_t)(op1), (size_t)(op2)) +#define vasub_vv_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasub_vv_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vasub_vv_i32m8(op0, op1, op2) \ +__builtin_rvv_vasub_vv_i32m8((vint32m8_t)(op0), (vint32m8_t)(op1), (size_t)(op2)) +#define vasub_vv_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasub_vv_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vasub_vv_i32mf2(op0, op1, op2) \ +__builtin_rvv_vasub_vv_i32mf2((vint32mf2_t)(op0), (vint32mf2_t)(op1), (size_t)(op2)) +#define vasub_vv_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasub_vv_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vasub_vv_i64m1(op0, op1, op2) \ +__builtin_rvv_vasub_vv_i64m1((vint64m1_t)(op0), (vint64m1_t)(op1), (size_t)(op2)) +#define vasub_vv_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasub_vv_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vasub_vv_i64m2(op0, op1, op2) \ +__builtin_rvv_vasub_vv_i64m2((vint64m2_t)(op0), (vint64m2_t)(op1), (size_t)(op2)) +#define vasub_vv_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasub_vv_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vasub_vv_i64m4(op0, op1, op2) \ +__builtin_rvv_vasub_vv_i64m4((vint64m4_t)(op0), (vint64m4_t)(op1), (size_t)(op2)) +#define vasub_vv_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasub_vv_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vasub_vv_i64m8(op0, op1, op2) \ +__builtin_rvv_vasub_vv_i64m8((vint64m8_t)(op0), (vint64m8_t)(op1), (size_t)(op2)) +#define vasub_vv_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasub_vv_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vasub_vx_i8m1(op0, op1, op2) \ +__builtin_rvv_vasub_vx_i8m1((vint8m1_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vasub_vx_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasub_vx_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vasub_vx_i8m2(op0, op1, op2) \ +__builtin_rvv_vasub_vx_i8m2((vint8m2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vasub_vx_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasub_vx_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vasub_vx_i8m4(op0, op1, op2) \ +__builtin_rvv_vasub_vx_i8m4((vint8m4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vasub_vx_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasub_vx_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vasub_vx_i8m8(op0, op1, op2) \ +__builtin_rvv_vasub_vx_i8m8((vint8m8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vasub_vx_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasub_vx_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vasub_vx_i8mf2(op0, op1, op2) \ +__builtin_rvv_vasub_vx_i8mf2((vint8mf2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vasub_vx_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasub_vx_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vasub_vx_i8mf4(op0, op1, op2) \ +__builtin_rvv_vasub_vx_i8mf4((vint8mf4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vasub_vx_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasub_vx_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vasub_vx_i8mf8(op0, op1, op2) \ +__builtin_rvv_vasub_vx_i8mf8((vint8mf8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vasub_vx_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasub_vx_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vasub_vx_i16m1(op0, op1, op2) \ +__builtin_rvv_vasub_vx_i16m1((vint16m1_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vasub_vx_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasub_vx_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vasub_vx_i16m2(op0, op1, op2) \ +__builtin_rvv_vasub_vx_i16m2((vint16m2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vasub_vx_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasub_vx_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vasub_vx_i16m4(op0, op1, op2) \ +__builtin_rvv_vasub_vx_i16m4((vint16m4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vasub_vx_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasub_vx_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vasub_vx_i16m8(op0, op1, op2) \ +__builtin_rvv_vasub_vx_i16m8((vint16m8_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vasub_vx_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasub_vx_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vasub_vx_i16mf2(op0, op1, op2) \ +__builtin_rvv_vasub_vx_i16mf2((vint16mf2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vasub_vx_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasub_vx_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vasub_vx_i16mf4(op0, op1, op2) \ +__builtin_rvv_vasub_vx_i16mf4((vint16mf4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vasub_vx_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasub_vx_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vasub_vx_i32m1(op0, op1, op2) \ +__builtin_rvv_vasub_vx_i32m1((vint32m1_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vasub_vx_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasub_vx_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vasub_vx_i32m2(op0, op1, op2) \ +__builtin_rvv_vasub_vx_i32m2((vint32m2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vasub_vx_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasub_vx_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vasub_vx_i32m4(op0, op1, op2) \ +__builtin_rvv_vasub_vx_i32m4((vint32m4_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vasub_vx_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasub_vx_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vasub_vx_i32m8(op0, op1, op2) \ +__builtin_rvv_vasub_vx_i32m8((vint32m8_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vasub_vx_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasub_vx_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vasub_vx_i32mf2(op0, op1, op2) \ +__builtin_rvv_vasub_vx_i32mf2((vint32mf2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vasub_vx_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasub_vx_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vasub_vx_i64m1(op0, op1, op2) \ +__builtin_rvv_vasub_vx_i64m1((vint64m1_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vasub_vx_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasub_vx_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vasub_vx_i64m2(op0, op1, op2) \ +__builtin_rvv_vasub_vx_i64m2((vint64m2_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vasub_vx_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasub_vx_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vasub_vx_i64m4(op0, op1, op2) \ +__builtin_rvv_vasub_vx_i64m4((vint64m4_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vasub_vx_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasub_vx_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vasub_vx_i64m8(op0, op1, op2) \ +__builtin_rvv_vasub_vx_i64m8((vint64m8_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vasub_vx_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vasub_vx_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vsmul_vv_i8m1(op0, op1, op2) \ +__builtin_rvv_vsmul_vv_i8m1((vint8m1_t)(op0), (vint8m1_t)(op1), (size_t)(op2)) +#define vsmul_vv_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsmul_vv_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vsmul_vv_i8m2(op0, op1, op2) \ +__builtin_rvv_vsmul_vv_i8m2((vint8m2_t)(op0), (vint8m2_t)(op1), (size_t)(op2)) +#define vsmul_vv_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsmul_vv_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vsmul_vv_i8m4(op0, op1, op2) \ +__builtin_rvv_vsmul_vv_i8m4((vint8m4_t)(op0), (vint8m4_t)(op1), (size_t)(op2)) +#define vsmul_vv_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsmul_vv_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vsmul_vv_i8m8(op0, op1, op2) \ +__builtin_rvv_vsmul_vv_i8m8((vint8m8_t)(op0), (vint8m8_t)(op1), (size_t)(op2)) +#define vsmul_vv_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsmul_vv_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (vint8m8_t)(op3), (size_t)(op4)) +#define vsmul_vv_i8mf2(op0, op1, op2) \ +__builtin_rvv_vsmul_vv_i8mf2((vint8mf2_t)(op0), (vint8mf2_t)(op1), (size_t)(op2)) +#define vsmul_vv_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsmul_vv_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vsmul_vv_i8mf4(op0, op1, op2) \ +__builtin_rvv_vsmul_vv_i8mf4((vint8mf4_t)(op0), (vint8mf4_t)(op1), (size_t)(op2)) +#define vsmul_vv_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsmul_vv_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vsmul_vv_i8mf8(op0, op1, op2) \ +__builtin_rvv_vsmul_vv_i8mf8((vint8mf8_t)(op0), (vint8mf8_t)(op1), (size_t)(op2)) +#define vsmul_vv_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsmul_vv_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vsmul_vv_i16m1(op0, op1, op2) \ +__builtin_rvv_vsmul_vv_i16m1((vint16m1_t)(op0), (vint16m1_t)(op1), (size_t)(op2)) +#define vsmul_vv_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsmul_vv_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vsmul_vv_i16m2(op0, op1, op2) \ +__builtin_rvv_vsmul_vv_i16m2((vint16m2_t)(op0), (vint16m2_t)(op1), (size_t)(op2)) +#define vsmul_vv_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsmul_vv_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vsmul_vv_i16m4(op0, op1, op2) \ +__builtin_rvv_vsmul_vv_i16m4((vint16m4_t)(op0), (vint16m4_t)(op1), (size_t)(op2)) +#define vsmul_vv_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsmul_vv_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vsmul_vv_i16m8(op0, op1, op2) \ +__builtin_rvv_vsmul_vv_i16m8((vint16m8_t)(op0), (vint16m8_t)(op1), (size_t)(op2)) +#define vsmul_vv_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsmul_vv_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (vint16m8_t)(op3), (size_t)(op4)) +#define vsmul_vv_i16mf2(op0, op1, op2) \ +__builtin_rvv_vsmul_vv_i16mf2((vint16mf2_t)(op0), (vint16mf2_t)(op1), (size_t)(op2)) +#define vsmul_vv_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsmul_vv_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vsmul_vv_i16mf4(op0, op1, op2) \ +__builtin_rvv_vsmul_vv_i16mf4((vint16mf4_t)(op0), (vint16mf4_t)(op1), (size_t)(op2)) +#define vsmul_vv_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsmul_vv_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vsmul_vv_i32m1(op0, op1, op2) \ +__builtin_rvv_vsmul_vv_i32m1((vint32m1_t)(op0), (vint32m1_t)(op1), (size_t)(op2)) +#define vsmul_vv_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsmul_vv_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vsmul_vv_i32m2(op0, op1, op2) \ +__builtin_rvv_vsmul_vv_i32m2((vint32m2_t)(op0), (vint32m2_t)(op1), (size_t)(op2)) +#define vsmul_vv_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsmul_vv_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vsmul_vv_i32m4(op0, op1, op2) \ +__builtin_rvv_vsmul_vv_i32m4((vint32m4_t)(op0), (vint32m4_t)(op1), (size_t)(op2)) +#define vsmul_vv_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsmul_vv_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vsmul_vv_i32m8(op0, op1, op2) \ +__builtin_rvv_vsmul_vv_i32m8((vint32m8_t)(op0), (vint32m8_t)(op1), (size_t)(op2)) +#define vsmul_vv_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsmul_vv_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vsmul_vv_i32mf2(op0, op1, op2) \ +__builtin_rvv_vsmul_vv_i32mf2((vint32mf2_t)(op0), (vint32mf2_t)(op1), (size_t)(op2)) +#define vsmul_vv_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsmul_vv_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vsmul_vv_i64m1(op0, op1, op2) \ +__builtin_rvv_vsmul_vv_i64m1((vint64m1_t)(op0), (vint64m1_t)(op1), (size_t)(op2)) +#define vsmul_vv_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsmul_vv_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vsmul_vv_i64m2(op0, op1, op2) \ +__builtin_rvv_vsmul_vv_i64m2((vint64m2_t)(op0), (vint64m2_t)(op1), (size_t)(op2)) +#define vsmul_vv_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsmul_vv_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vsmul_vv_i64m4(op0, op1, op2) \ +__builtin_rvv_vsmul_vv_i64m4((vint64m4_t)(op0), (vint64m4_t)(op1), (size_t)(op2)) +#define vsmul_vv_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsmul_vv_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vsmul_vv_i64m8(op0, op1, op2) \ +__builtin_rvv_vsmul_vv_i64m8((vint64m8_t)(op0), (vint64m8_t)(op1), (size_t)(op2)) +#define vsmul_vv_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsmul_vv_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vsmul_vx_i8m1(op0, op1, op2) \ +__builtin_rvv_vsmul_vx_i8m1((vint8m1_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vsmul_vx_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsmul_vx_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vsmul_vx_i8m2(op0, op1, op2) \ +__builtin_rvv_vsmul_vx_i8m2((vint8m2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vsmul_vx_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsmul_vx_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vsmul_vx_i8m4(op0, op1, op2) \ +__builtin_rvv_vsmul_vx_i8m4((vint8m4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vsmul_vx_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsmul_vx_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vsmul_vx_i8m8(op0, op1, op2) \ +__builtin_rvv_vsmul_vx_i8m8((vint8m8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vsmul_vx_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsmul_vx_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vsmul_vx_i8mf2(op0, op1, op2) \ +__builtin_rvv_vsmul_vx_i8mf2((vint8mf2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vsmul_vx_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsmul_vx_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vsmul_vx_i8mf4(op0, op1, op2) \ +__builtin_rvv_vsmul_vx_i8mf4((vint8mf4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vsmul_vx_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsmul_vx_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vsmul_vx_i8mf8(op0, op1, op2) \ +__builtin_rvv_vsmul_vx_i8mf8((vint8mf8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vsmul_vx_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsmul_vx_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vsmul_vx_i16m1(op0, op1, op2) \ +__builtin_rvv_vsmul_vx_i16m1((vint16m1_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vsmul_vx_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsmul_vx_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vsmul_vx_i16m2(op0, op1, op2) \ +__builtin_rvv_vsmul_vx_i16m2((vint16m2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vsmul_vx_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsmul_vx_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vsmul_vx_i16m4(op0, op1, op2) \ +__builtin_rvv_vsmul_vx_i16m4((vint16m4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vsmul_vx_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsmul_vx_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vsmul_vx_i16m8(op0, op1, op2) \ +__builtin_rvv_vsmul_vx_i16m8((vint16m8_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vsmul_vx_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsmul_vx_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vsmul_vx_i16mf2(op0, op1, op2) \ +__builtin_rvv_vsmul_vx_i16mf2((vint16mf2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vsmul_vx_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsmul_vx_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vsmul_vx_i16mf4(op0, op1, op2) \ +__builtin_rvv_vsmul_vx_i16mf4((vint16mf4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vsmul_vx_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsmul_vx_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vsmul_vx_i32m1(op0, op1, op2) \ +__builtin_rvv_vsmul_vx_i32m1((vint32m1_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vsmul_vx_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsmul_vx_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vsmul_vx_i32m2(op0, op1, op2) \ +__builtin_rvv_vsmul_vx_i32m2((vint32m2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vsmul_vx_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsmul_vx_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vsmul_vx_i32m4(op0, op1, op2) \ +__builtin_rvv_vsmul_vx_i32m4((vint32m4_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vsmul_vx_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsmul_vx_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vsmul_vx_i32m8(op0, op1, op2) \ +__builtin_rvv_vsmul_vx_i32m8((vint32m8_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vsmul_vx_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsmul_vx_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vsmul_vx_i32mf2(op0, op1, op2) \ +__builtin_rvv_vsmul_vx_i32mf2((vint32mf2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vsmul_vx_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsmul_vx_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vsmul_vx_i64m1(op0, op1, op2) \ +__builtin_rvv_vsmul_vx_i64m1((vint64m1_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vsmul_vx_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsmul_vx_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vsmul_vx_i64m2(op0, op1, op2) \ +__builtin_rvv_vsmul_vx_i64m2((vint64m2_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vsmul_vx_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsmul_vx_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vsmul_vx_i64m4(op0, op1, op2) \ +__builtin_rvv_vsmul_vx_i64m4((vint64m4_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vsmul_vx_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsmul_vx_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vsmul_vx_i64m8(op0, op1, op2) \ +__builtin_rvv_vsmul_vx_i64m8((vint64m8_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vsmul_vx_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsmul_vx_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vssrl_vv_u8m1(op0, op1, op2) \ +__builtin_rvv_vssrl_vv_u8m1((vuint8m1_t)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vssrl_vv_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssrl_vv_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vssrl_vv_u8m2(op0, op1, op2) \ +__builtin_rvv_vssrl_vv_u8m2((vuint8m2_t)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vssrl_vv_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssrl_vv_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vssrl_vv_u8m4(op0, op1, op2) \ +__builtin_rvv_vssrl_vv_u8m4((vuint8m4_t)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vssrl_vv_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssrl_vv_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vssrl_vv_u8m8(op0, op1, op2) \ +__builtin_rvv_vssrl_vv_u8m8((vuint8m8_t)(op0), (vuint8m8_t)(op1), (size_t)(op2)) +#define vssrl_vv_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssrl_vv_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (vuint8m8_t)(op3), (size_t)(op4)) +#define vssrl_vv_u8mf2(op0, op1, op2) \ +__builtin_rvv_vssrl_vv_u8mf2((vuint8mf2_t)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vssrl_vv_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssrl_vv_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vssrl_vv_u8mf4(op0, op1, op2) \ +__builtin_rvv_vssrl_vv_u8mf4((vuint8mf4_t)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vssrl_vv_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssrl_vv_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vssrl_vv_u8mf8(op0, op1, op2) \ +__builtin_rvv_vssrl_vv_u8mf8((vuint8mf8_t)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vssrl_vv_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssrl_vv_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vssrl_vv_u16m1(op0, op1, op2) \ +__builtin_rvv_vssrl_vv_u16m1((vuint16m1_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vssrl_vv_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssrl_vv_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vssrl_vv_u16m2(op0, op1, op2) \ +__builtin_rvv_vssrl_vv_u16m2((vuint16m2_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vssrl_vv_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssrl_vv_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vssrl_vv_u16m4(op0, op1, op2) \ +__builtin_rvv_vssrl_vv_u16m4((vuint16m4_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vssrl_vv_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssrl_vv_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vssrl_vv_u16m8(op0, op1, op2) \ +__builtin_rvv_vssrl_vv_u16m8((vuint16m8_t)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vssrl_vv_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssrl_vv_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vssrl_vv_u16mf2(op0, op1, op2) \ +__builtin_rvv_vssrl_vv_u16mf2((vuint16mf2_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vssrl_vv_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssrl_vv_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vssrl_vv_u16mf4(op0, op1, op2) \ +__builtin_rvv_vssrl_vv_u16mf4((vuint16mf4_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vssrl_vv_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssrl_vv_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vssrl_vv_u32m1(op0, op1, op2) \ +__builtin_rvv_vssrl_vv_u32m1((vuint32m1_t)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vssrl_vv_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssrl_vv_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vssrl_vv_u32m2(op0, op1, op2) \ +__builtin_rvv_vssrl_vv_u32m2((vuint32m2_t)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vssrl_vv_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssrl_vv_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vssrl_vv_u32m4(op0, op1, op2) \ +__builtin_rvv_vssrl_vv_u32m4((vuint32m4_t)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vssrl_vv_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssrl_vv_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vssrl_vv_u32m8(op0, op1, op2) \ +__builtin_rvv_vssrl_vv_u32m8((vuint32m8_t)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vssrl_vv_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssrl_vv_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vssrl_vv_u32mf2(op0, op1, op2) \ +__builtin_rvv_vssrl_vv_u32mf2((vuint32mf2_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vssrl_vv_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssrl_vv_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vssrl_vv_u64m1(op0, op1, op2) \ +__builtin_rvv_vssrl_vv_u64m1((vuint64m1_t)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vssrl_vv_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssrl_vv_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vssrl_vv_u64m2(op0, op1, op2) \ +__builtin_rvv_vssrl_vv_u64m2((vuint64m2_t)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vssrl_vv_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssrl_vv_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vssrl_vv_u64m4(op0, op1, op2) \ +__builtin_rvv_vssrl_vv_u64m4((vuint64m4_t)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vssrl_vv_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssrl_vv_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vssrl_vv_u64m8(op0, op1, op2) \ +__builtin_rvv_vssrl_vv_u64m8((vuint64m8_t)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vssrl_vv_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssrl_vv_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vssrl_vx_u8m1(op0, op1, op2) \ +__builtin_rvv_vssrl_vx_u8m1((vuint8m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vssrl_vx_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssrl_vx_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vssrl_vx_u8m2(op0, op1, op2) \ +__builtin_rvv_vssrl_vx_u8m2((vuint8m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vssrl_vx_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssrl_vx_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vssrl_vx_u8m4(op0, op1, op2) \ +__builtin_rvv_vssrl_vx_u8m4((vuint8m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vssrl_vx_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssrl_vx_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vssrl_vx_u8m8(op0, op1, op2) \ +__builtin_rvv_vssrl_vx_u8m8((vuint8m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vssrl_vx_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssrl_vx_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vssrl_vx_u8mf2(op0, op1, op2) \ +__builtin_rvv_vssrl_vx_u8mf2((vuint8mf2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vssrl_vx_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssrl_vx_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vssrl_vx_u8mf4(op0, op1, op2) \ +__builtin_rvv_vssrl_vx_u8mf4((vuint8mf4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vssrl_vx_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssrl_vx_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vssrl_vx_u8mf8(op0, op1, op2) \ +__builtin_rvv_vssrl_vx_u8mf8((vuint8mf8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vssrl_vx_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssrl_vx_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vssrl_vx_u16m1(op0, op1, op2) \ +__builtin_rvv_vssrl_vx_u16m1((vuint16m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vssrl_vx_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssrl_vx_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vssrl_vx_u16m2(op0, op1, op2) \ +__builtin_rvv_vssrl_vx_u16m2((vuint16m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vssrl_vx_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssrl_vx_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vssrl_vx_u16m4(op0, op1, op2) \ +__builtin_rvv_vssrl_vx_u16m4((vuint16m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vssrl_vx_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssrl_vx_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vssrl_vx_u16m8(op0, op1, op2) \ +__builtin_rvv_vssrl_vx_u16m8((vuint16m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vssrl_vx_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssrl_vx_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vssrl_vx_u16mf2(op0, op1, op2) \ +__builtin_rvv_vssrl_vx_u16mf2((vuint16mf2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vssrl_vx_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssrl_vx_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vssrl_vx_u16mf4(op0, op1, op2) \ +__builtin_rvv_vssrl_vx_u16mf4((vuint16mf4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vssrl_vx_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssrl_vx_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vssrl_vx_u32m1(op0, op1, op2) \ +__builtin_rvv_vssrl_vx_u32m1((vuint32m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vssrl_vx_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssrl_vx_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vssrl_vx_u32m2(op0, op1, op2) \ +__builtin_rvv_vssrl_vx_u32m2((vuint32m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vssrl_vx_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssrl_vx_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vssrl_vx_u32m4(op0, op1, op2) \ +__builtin_rvv_vssrl_vx_u32m4((vuint32m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vssrl_vx_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssrl_vx_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vssrl_vx_u32m8(op0, op1, op2) \ +__builtin_rvv_vssrl_vx_u32m8((vuint32m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vssrl_vx_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssrl_vx_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vssrl_vx_u32mf2(op0, op1, op2) \ +__builtin_rvv_vssrl_vx_u32mf2((vuint32mf2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vssrl_vx_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssrl_vx_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vssrl_vx_u64m1(op0, op1, op2) \ +__builtin_rvv_vssrl_vx_u64m1((vuint64m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vssrl_vx_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssrl_vx_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vssrl_vx_u64m2(op0, op1, op2) \ +__builtin_rvv_vssrl_vx_u64m2((vuint64m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vssrl_vx_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssrl_vx_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vssrl_vx_u64m4(op0, op1, op2) \ +__builtin_rvv_vssrl_vx_u64m4((vuint64m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vssrl_vx_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssrl_vx_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vssrl_vx_u64m8(op0, op1, op2) \ +__builtin_rvv_vssrl_vx_u64m8((vuint64m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vssrl_vx_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssrl_vx_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vssra_vv_i8m1(op0, op1, op2) \ +__builtin_rvv_vssra_vv_i8m1((vint8m1_t)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vssra_vv_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssra_vv_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vssra_vv_i8m2(op0, op1, op2) \ +__builtin_rvv_vssra_vv_i8m2((vint8m2_t)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vssra_vv_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssra_vv_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vssra_vv_i8m4(op0, op1, op2) \ +__builtin_rvv_vssra_vv_i8m4((vint8m4_t)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vssra_vv_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssra_vv_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vssra_vv_i8m8(op0, op1, op2) \ +__builtin_rvv_vssra_vv_i8m8((vint8m8_t)(op0), (vuint8m8_t)(op1), (size_t)(op2)) +#define vssra_vv_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssra_vv_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (vuint8m8_t)(op3), (size_t)(op4)) +#define vssra_vv_i8mf2(op0, op1, op2) \ +__builtin_rvv_vssra_vv_i8mf2((vint8mf2_t)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vssra_vv_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssra_vv_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vssra_vv_i8mf4(op0, op1, op2) \ +__builtin_rvv_vssra_vv_i8mf4((vint8mf4_t)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vssra_vv_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssra_vv_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vssra_vv_i8mf8(op0, op1, op2) \ +__builtin_rvv_vssra_vv_i8mf8((vint8mf8_t)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vssra_vv_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssra_vv_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vssra_vv_i16m1(op0, op1, op2) \ +__builtin_rvv_vssra_vv_i16m1((vint16m1_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vssra_vv_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssra_vv_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vssra_vv_i16m2(op0, op1, op2) \ +__builtin_rvv_vssra_vv_i16m2((vint16m2_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vssra_vv_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssra_vv_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vssra_vv_i16m4(op0, op1, op2) \ +__builtin_rvv_vssra_vv_i16m4((vint16m4_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vssra_vv_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssra_vv_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vssra_vv_i16m8(op0, op1, op2) \ +__builtin_rvv_vssra_vv_i16m8((vint16m8_t)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vssra_vv_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssra_vv_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vssra_vv_i16mf2(op0, op1, op2) \ +__builtin_rvv_vssra_vv_i16mf2((vint16mf2_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vssra_vv_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssra_vv_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vssra_vv_i16mf4(op0, op1, op2) \ +__builtin_rvv_vssra_vv_i16mf4((vint16mf4_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vssra_vv_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssra_vv_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vssra_vv_i32m1(op0, op1, op2) \ +__builtin_rvv_vssra_vv_i32m1((vint32m1_t)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vssra_vv_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssra_vv_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vssra_vv_i32m2(op0, op1, op2) \ +__builtin_rvv_vssra_vv_i32m2((vint32m2_t)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vssra_vv_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssra_vv_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vssra_vv_i32m4(op0, op1, op2) \ +__builtin_rvv_vssra_vv_i32m4((vint32m4_t)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vssra_vv_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssra_vv_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vssra_vv_i32m8(op0, op1, op2) \ +__builtin_rvv_vssra_vv_i32m8((vint32m8_t)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vssra_vv_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssra_vv_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vssra_vv_i32mf2(op0, op1, op2) \ +__builtin_rvv_vssra_vv_i32mf2((vint32mf2_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vssra_vv_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssra_vv_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vssra_vv_i64m1(op0, op1, op2) \ +__builtin_rvv_vssra_vv_i64m1((vint64m1_t)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vssra_vv_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssra_vv_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vssra_vv_i64m2(op0, op1, op2) \ +__builtin_rvv_vssra_vv_i64m2((vint64m2_t)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vssra_vv_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssra_vv_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vssra_vv_i64m4(op0, op1, op2) \ +__builtin_rvv_vssra_vv_i64m4((vint64m4_t)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vssra_vv_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssra_vv_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vssra_vv_i64m8(op0, op1, op2) \ +__builtin_rvv_vssra_vv_i64m8((vint64m8_t)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vssra_vv_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssra_vv_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vssra_vx_i8m1(op0, op1, op2) \ +__builtin_rvv_vssra_vx_i8m1((vint8m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vssra_vx_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssra_vx_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vssra_vx_i8m2(op0, op1, op2) \ +__builtin_rvv_vssra_vx_i8m2((vint8m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vssra_vx_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssra_vx_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vssra_vx_i8m4(op0, op1, op2) \ +__builtin_rvv_vssra_vx_i8m4((vint8m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vssra_vx_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssra_vx_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vssra_vx_i8m8(op0, op1, op2) \ +__builtin_rvv_vssra_vx_i8m8((vint8m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vssra_vx_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssra_vx_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vssra_vx_i8mf2(op0, op1, op2) \ +__builtin_rvv_vssra_vx_i8mf2((vint8mf2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vssra_vx_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssra_vx_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vssra_vx_i8mf4(op0, op1, op2) \ +__builtin_rvv_vssra_vx_i8mf4((vint8mf4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vssra_vx_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssra_vx_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vssra_vx_i8mf8(op0, op1, op2) \ +__builtin_rvv_vssra_vx_i8mf8((vint8mf8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vssra_vx_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssra_vx_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vssra_vx_i16m1(op0, op1, op2) \ +__builtin_rvv_vssra_vx_i16m1((vint16m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vssra_vx_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssra_vx_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vssra_vx_i16m2(op0, op1, op2) \ +__builtin_rvv_vssra_vx_i16m2((vint16m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vssra_vx_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssra_vx_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vssra_vx_i16m4(op0, op1, op2) \ +__builtin_rvv_vssra_vx_i16m4((vint16m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vssra_vx_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssra_vx_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vssra_vx_i16m8(op0, op1, op2) \ +__builtin_rvv_vssra_vx_i16m8((vint16m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vssra_vx_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssra_vx_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vssra_vx_i16mf2(op0, op1, op2) \ +__builtin_rvv_vssra_vx_i16mf2((vint16mf2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vssra_vx_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssra_vx_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vssra_vx_i16mf4(op0, op1, op2) \ +__builtin_rvv_vssra_vx_i16mf4((vint16mf4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vssra_vx_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssra_vx_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vssra_vx_i32m1(op0, op1, op2) \ +__builtin_rvv_vssra_vx_i32m1((vint32m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vssra_vx_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssra_vx_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vssra_vx_i32m2(op0, op1, op2) \ +__builtin_rvv_vssra_vx_i32m2((vint32m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vssra_vx_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssra_vx_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vssra_vx_i32m4(op0, op1, op2) \ +__builtin_rvv_vssra_vx_i32m4((vint32m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vssra_vx_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssra_vx_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vssra_vx_i32m8(op0, op1, op2) \ +__builtin_rvv_vssra_vx_i32m8((vint32m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vssra_vx_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssra_vx_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vssra_vx_i32mf2(op0, op1, op2) \ +__builtin_rvv_vssra_vx_i32mf2((vint32mf2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vssra_vx_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssra_vx_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vssra_vx_i64m1(op0, op1, op2) \ +__builtin_rvv_vssra_vx_i64m1((vint64m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vssra_vx_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssra_vx_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vssra_vx_i64m2(op0, op1, op2) \ +__builtin_rvv_vssra_vx_i64m2((vint64m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vssra_vx_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssra_vx_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vssra_vx_i64m4(op0, op1, op2) \ +__builtin_rvv_vssra_vx_i64m4((vint64m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vssra_vx_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssra_vx_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vssra_vx_i64m8(op0, op1, op2) \ +__builtin_rvv_vssra_vx_i64m8((vint64m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vssra_vx_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssra_vx_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnclipu_wv_u8m1(op0, op1, op2) \ +__builtin_rvv_vnclipu_wv_u8m1((vuint16m2_t)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vnclipu_wv_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclipu_wv_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint16m2_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vnclipu_wv_u8m2(op0, op1, op2) \ +__builtin_rvv_vnclipu_wv_u8m2((vuint16m4_t)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vnclipu_wv_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclipu_wv_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint16m4_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vnclipu_wv_u8m4(op0, op1, op2) \ +__builtin_rvv_vnclipu_wv_u8m4((vuint16m8_t)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vnclipu_wv_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclipu_wv_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint16m8_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vnclipu_wv_u8mf2(op0, op1, op2) \ +__builtin_rvv_vnclipu_wv_u8mf2((vuint16m1_t)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vnclipu_wv_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclipu_wv_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint16m1_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vnclipu_wv_u8mf4(op0, op1, op2) \ +__builtin_rvv_vnclipu_wv_u8mf4((vuint16mf2_t)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vnclipu_wv_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclipu_wv_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint16mf2_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vnclipu_wv_u8mf8(op0, op1, op2) \ +__builtin_rvv_vnclipu_wv_u8mf8((vuint16mf4_t)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vnclipu_wv_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclipu_wv_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint16mf4_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vnclipu_wv_u16m1(op0, op1, op2) \ +__builtin_rvv_vnclipu_wv_u16m1((vuint32m2_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vnclipu_wv_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclipu_wv_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint32m2_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vnclipu_wv_u16m2(op0, op1, op2) \ +__builtin_rvv_vnclipu_wv_u16m2((vuint32m4_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vnclipu_wv_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclipu_wv_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint32m4_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vnclipu_wv_u16m4(op0, op1, op2) \ +__builtin_rvv_vnclipu_wv_u16m4((vuint32m8_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vnclipu_wv_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclipu_wv_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint32m8_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vnclipu_wv_u16mf2(op0, op1, op2) \ +__builtin_rvv_vnclipu_wv_u16mf2((vuint32m1_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vnclipu_wv_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclipu_wv_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint32m1_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vnclipu_wv_u16mf4(op0, op1, op2) \ +__builtin_rvv_vnclipu_wv_u16mf4((vuint32mf2_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vnclipu_wv_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclipu_wv_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint32mf2_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vnclipu_wv_u32m1(op0, op1, op2) \ +__builtin_rvv_vnclipu_wv_u32m1((vuint64m2_t)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vnclipu_wv_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclipu_wv_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint64m2_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vnclipu_wv_u32m2(op0, op1, op2) \ +__builtin_rvv_vnclipu_wv_u32m2((vuint64m4_t)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vnclipu_wv_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclipu_wv_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint64m4_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vnclipu_wv_u32m4(op0, op1, op2) \ +__builtin_rvv_vnclipu_wv_u32m4((vuint64m8_t)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vnclipu_wv_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclipu_wv_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint64m8_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vnclipu_wv_u32mf2(op0, op1, op2) \ +__builtin_rvv_vnclipu_wv_u32mf2((vuint64m1_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vnclipu_wv_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclipu_wv_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint64m1_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vnclipu_wx_u8m1(op0, op1, op2) \ +__builtin_rvv_vnclipu_wx_u8m1((vuint16m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnclipu_wx_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclipu_wx_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint16m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnclipu_wx_u8m2(op0, op1, op2) \ +__builtin_rvv_vnclipu_wx_u8m2((vuint16m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnclipu_wx_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclipu_wx_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint16m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnclipu_wx_u8m4(op0, op1, op2) \ +__builtin_rvv_vnclipu_wx_u8m4((vuint16m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnclipu_wx_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclipu_wx_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint16m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnclipu_wx_u8mf2(op0, op1, op2) \ +__builtin_rvv_vnclipu_wx_u8mf2((vuint16m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnclipu_wx_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclipu_wx_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint16m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnclipu_wx_u8mf4(op0, op1, op2) \ +__builtin_rvv_vnclipu_wx_u8mf4((vuint16mf2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnclipu_wx_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclipu_wx_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint16mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnclipu_wx_u8mf8(op0, op1, op2) \ +__builtin_rvv_vnclipu_wx_u8mf8((vuint16mf4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnclipu_wx_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclipu_wx_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint16mf4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnclipu_wx_u16m1(op0, op1, op2) \ +__builtin_rvv_vnclipu_wx_u16m1((vuint32m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnclipu_wx_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclipu_wx_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint32m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnclipu_wx_u16m2(op0, op1, op2) \ +__builtin_rvv_vnclipu_wx_u16m2((vuint32m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnclipu_wx_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclipu_wx_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint32m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnclipu_wx_u16m4(op0, op1, op2) \ +__builtin_rvv_vnclipu_wx_u16m4((vuint32m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnclipu_wx_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclipu_wx_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint32m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnclipu_wx_u16mf2(op0, op1, op2) \ +__builtin_rvv_vnclipu_wx_u16mf2((vuint32m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnclipu_wx_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclipu_wx_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint32m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnclipu_wx_u16mf4(op0, op1, op2) \ +__builtin_rvv_vnclipu_wx_u16mf4((vuint32mf2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnclipu_wx_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclipu_wx_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnclipu_wx_u32m1(op0, op1, op2) \ +__builtin_rvv_vnclipu_wx_u32m1((vuint64m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnclipu_wx_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclipu_wx_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint64m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnclipu_wx_u32m2(op0, op1, op2) \ +__builtin_rvv_vnclipu_wx_u32m2((vuint64m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnclipu_wx_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclipu_wx_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint64m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnclipu_wx_u32m4(op0, op1, op2) \ +__builtin_rvv_vnclipu_wx_u32m4((vuint64m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnclipu_wx_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclipu_wx_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint64m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnclipu_wx_u32mf2(op0, op1, op2) \ +__builtin_rvv_vnclipu_wx_u32mf2((vuint64m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnclipu_wx_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclipu_wx_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint64m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsuxei64_v_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei64_v_u32m1((uint32_t *)(op0), (vuint64m2_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vsuxei64_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei64_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint64m2_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vsuxei64_v_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei64_v_u32m2((uint32_t *)(op0), (vuint64m4_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vsuxei64_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei64_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint64m4_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vsuxei64_v_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei64_v_u32m4((uint32_t *)(op0), (vuint64m8_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vsuxei64_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei64_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint64m8_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vsuxei64_v_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei64_v_u32mf2((uint32_t *)(op0), (vuint64m1_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vsuxei64_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei64_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint64m1_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vnclip_wv_i8m1(op0, op1, op2) \ +__builtin_rvv_vnclip_wv_i8m1((vint16m2_t)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vnclip_wv_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclip_wv_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint16m2_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vnclip_wv_i8m2(op0, op1, op2) \ +__builtin_rvv_vnclip_wv_i8m2((vint16m4_t)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vnclip_wv_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclip_wv_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint16m4_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vnclip_wv_i8m4(op0, op1, op2) \ +__builtin_rvv_vnclip_wv_i8m4((vint16m8_t)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vnclip_wv_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclip_wv_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint16m8_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vnclip_wv_i8mf2(op0, op1, op2) \ +__builtin_rvv_vnclip_wv_i8mf2((vint16m1_t)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vnclip_wv_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclip_wv_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint16m1_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vnclip_wv_i8mf4(op0, op1, op2) \ +__builtin_rvv_vnclip_wv_i8mf4((vint16mf2_t)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vnclip_wv_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclip_wv_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint16mf2_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vnclip_wv_i8mf8(op0, op1, op2) \ +__builtin_rvv_vnclip_wv_i8mf8((vint16mf4_t)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vnclip_wv_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclip_wv_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint16mf4_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vnclip_wv_i16m1(op0, op1, op2) \ +__builtin_rvv_vnclip_wv_i16m1((vint32m2_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vnclip_wv_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclip_wv_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint32m2_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vnclip_wv_i16m2(op0, op1, op2) \ +__builtin_rvv_vnclip_wv_i16m2((vint32m4_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vnclip_wv_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclip_wv_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint32m4_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vnclip_wv_i16m4(op0, op1, op2) \ +__builtin_rvv_vnclip_wv_i16m4((vint32m8_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vnclip_wv_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclip_wv_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint32m8_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vnclip_wv_i16mf2(op0, op1, op2) \ +__builtin_rvv_vnclip_wv_i16mf2((vint32m1_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vnclip_wv_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclip_wv_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint32m1_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vnclip_wv_i16mf4(op0, op1, op2) \ +__builtin_rvv_vnclip_wv_i16mf4((vint32mf2_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vnclip_wv_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclip_wv_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint32mf2_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vnclip_wv_i32m1(op0, op1, op2) \ +__builtin_rvv_vnclip_wv_i32m1((vint64m2_t)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vnclip_wv_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclip_wv_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint64m2_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vnclip_wv_i32m2(op0, op1, op2) \ +__builtin_rvv_vnclip_wv_i32m2((vint64m4_t)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vnclip_wv_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclip_wv_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint64m4_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vnclip_wv_i32m4(op0, op1, op2) \ +__builtin_rvv_vnclip_wv_i32m4((vint64m8_t)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vnclip_wv_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclip_wv_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint64m8_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vnclip_wv_i32mf2(op0, op1, op2) \ +__builtin_rvv_vnclip_wv_i32mf2((vint64m1_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vnclip_wv_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclip_wv_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint64m1_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vnclip_wx_i8m1(op0, op1, op2) \ +__builtin_rvv_vnclip_wx_i8m1((vint16m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnclip_wx_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclip_wx_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint16m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnclip_wx_i8m2(op0, op1, op2) \ +__builtin_rvv_vnclip_wx_i8m2((vint16m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnclip_wx_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclip_wx_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint16m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnclip_wx_i8m4(op0, op1, op2) \ +__builtin_rvv_vnclip_wx_i8m4((vint16m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnclip_wx_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclip_wx_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint16m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnclip_wx_i8mf2(op0, op1, op2) \ +__builtin_rvv_vnclip_wx_i8mf2((vint16m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnclip_wx_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclip_wx_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint16m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnclip_wx_i8mf4(op0, op1, op2) \ +__builtin_rvv_vnclip_wx_i8mf4((vint16mf2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnclip_wx_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclip_wx_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint16mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnclip_wx_i8mf8(op0, op1, op2) \ +__builtin_rvv_vnclip_wx_i8mf8((vint16mf4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnclip_wx_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclip_wx_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint16mf4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnclip_wx_i16m1(op0, op1, op2) \ +__builtin_rvv_vnclip_wx_i16m1((vint32m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnclip_wx_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclip_wx_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint32m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnclip_wx_i16m2(op0, op1, op2) \ +__builtin_rvv_vnclip_wx_i16m2((vint32m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnclip_wx_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclip_wx_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint32m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnclip_wx_i16m4(op0, op1, op2) \ +__builtin_rvv_vnclip_wx_i16m4((vint32m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnclip_wx_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclip_wx_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint32m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnclip_wx_i16mf2(op0, op1, op2) \ +__builtin_rvv_vnclip_wx_i16mf2((vint32m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnclip_wx_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclip_wx_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint32m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnclip_wx_i16mf4(op0, op1, op2) \ +__builtin_rvv_vnclip_wx_i16mf4((vint32mf2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnclip_wx_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclip_wx_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint32mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnclip_wx_i32m1(op0, op1, op2) \ +__builtin_rvv_vnclip_wx_i32m1((vint64m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnclip_wx_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclip_wx_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint64m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnclip_wx_i32m2(op0, op1, op2) \ +__builtin_rvv_vnclip_wx_i32m2((vint64m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnclip_wx_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclip_wx_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint64m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnclip_wx_i32m4(op0, op1, op2) \ +__builtin_rvv_vnclip_wx_i32m4((vint64m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnclip_wx_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclip_wx_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint64m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vnclip_wx_i32mf2(op0, op1, op2) \ +__builtin_rvv_vnclip_wx_i32mf2((vint64m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vnclip_wx_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vnclip_wx_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint64m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vsuxei8_v_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_i64m1((int64_t *)(op0), (vuint8mf8_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vsuxei8_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint8mf8_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vsuxei8_v_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_i64m2((int64_t *)(op0), (vuint8mf4_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vsuxei8_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint8mf4_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vsuxei8_v_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_i64m4((int64_t *)(op0), (vuint8mf2_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vsuxei8_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vuint8mf2_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vsuxei8_v_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_i64m8((int64_t *)(op0), (vuint8m1_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vsuxei8_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_i64m8_m((vbool8_t)(op0), (int64_t *)(op1), (vuint8m1_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vsuxei8_v_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_u64m1((uint64_t *)(op0), (vuint8mf8_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vsuxei8_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint8mf8_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vsuxei8_v_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_u64m2((uint64_t *)(op0), (vuint8mf4_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vsuxei8_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint8mf4_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vsuxei8_v_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_u64m4((uint64_t *)(op0), (vuint8mf2_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vsuxei8_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint8mf2_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vsuxei8_v_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_u64m8((uint64_t *)(op0), (vuint8m1_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vsuxei8_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_u64m8_m((vbool8_t)(op0), (uint64_t *)(op1), (vuint8m1_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vsuxei16_v_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_i64m1((int64_t *)(op0), (vuint16mf4_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vsuxei16_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint16mf4_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vsuxei16_v_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_i64m2((int64_t *)(op0), (vuint16mf2_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vsuxei16_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint16mf2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vsuxei16_v_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_i64m4((int64_t *)(op0), (vuint16m1_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vsuxei16_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vuint16m1_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vsuxei16_v_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_i64m8((int64_t *)(op0), (vuint16m2_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vsuxei16_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_i64m8_m((vbool8_t)(op0), (int64_t *)(op1), (vuint16m2_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vsuxei16_v_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_u64m1((uint64_t *)(op0), (vuint16mf4_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vsuxei16_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint16mf4_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vsuxei16_v_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_u64m2((uint64_t *)(op0), (vuint16mf2_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vsuxei16_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint16mf2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vsuxei16_v_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_u64m4((uint64_t *)(op0), (vuint16m1_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vsuxei16_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint16m1_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vsuxei16_v_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_u64m8((uint64_t *)(op0), (vuint16m2_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vsuxei16_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_u64m8_m((vbool8_t)(op0), (uint64_t *)(op1), (vuint16m2_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vsuxei32_v_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_i64m1((int64_t *)(op0), (vuint32mf2_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vsuxei32_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint32mf2_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vsuxei32_v_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_i64m2((int64_t *)(op0), (vuint32m1_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vsuxei32_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint32m1_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vsuxei32_v_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_i64m4((int64_t *)(op0), (vuint32m2_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vsuxei32_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vuint32m2_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vsuxei32_v_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_i64m8((int64_t *)(op0), (vuint32m4_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vsuxei32_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_i64m8_m((vbool8_t)(op0), (int64_t *)(op1), (vuint32m4_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vsuxei32_v_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_u64m1((uint64_t *)(op0), (vuint32mf2_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vsuxei32_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint32mf2_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vsuxei32_v_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_u64m2((uint64_t *)(op0), (vuint32m1_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vsuxei32_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint32m1_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vsuxei32_v_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_u64m4((uint64_t *)(op0), (vuint32m2_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vsuxei32_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint32m2_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vsuxei32_v_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_u64m8((uint64_t *)(op0), (vuint32m4_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vsuxei32_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_u64m8_m((vbool8_t)(op0), (uint64_t *)(op1), (vuint32m4_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vsuxei64_v_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei64_v_i64m1((int64_t *)(op0), (vuint64m1_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vsuxei64_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei64_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vsuxei64_v_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei64_v_i64m2((int64_t *)(op0), (vuint64m2_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vsuxei64_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei64_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vsuxei64_v_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei64_v_i64m4((int64_t *)(op0), (vuint64m4_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vsuxei64_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei64_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vuint64m4_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vsuxei64_v_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei64_v_i64m8((int64_t *)(op0), (vuint64m8_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vsuxei64_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei64_v_i64m8_m((vbool8_t)(op0), (int64_t *)(op1), (vuint64m8_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vredsum_vs_i8m1_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredsum_vs_i8m1_i8m1((vint8m1_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vredsum_vs_i8m1_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredsum_vs_i8m1_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vredsum_vs_i8m2_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredsum_vs_i8m2_i8m1((vint8m1_t)(op0), (vint8m2_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vredsum_vs_i8m2_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredsum_vs_i8m2_i8m1_m((vbool4_t)(op0), (vint8m1_t)(op1), (vint8m2_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vredsum_vs_i8m4_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredsum_vs_i8m4_i8m1((vint8m1_t)(op0), (vint8m4_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vredsum_vs_i8m4_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredsum_vs_i8m4_i8m1_m((vbool2_t)(op0), (vint8m1_t)(op1), (vint8m4_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vredsum_vs_i8m8_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredsum_vs_i8m8_i8m1((vint8m1_t)(op0), (vint8m8_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vredsum_vs_i8m8_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredsum_vs_i8m8_i8m1_m((vbool1_t)(op0), (vint8m1_t)(op1), (vint8m8_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vredsum_vs_i8mf2_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredsum_vs_i8mf2_i8m1((vint8m1_t)(op0), (vint8mf2_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vredsum_vs_i8mf2_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredsum_vs_i8mf2_i8m1_m((vbool16_t)(op0), (vint8m1_t)(op1), (vint8mf2_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vredsum_vs_i8mf4_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredsum_vs_i8mf4_i8m1((vint8m1_t)(op0), (vint8mf4_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vredsum_vs_i8mf4_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredsum_vs_i8mf4_i8m1_m((vbool32_t)(op0), (vint8m1_t)(op1), (vint8mf4_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vredsum_vs_i8mf8_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredsum_vs_i8mf8_i8m1((vint8m1_t)(op0), (vint8mf8_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vredsum_vs_i8mf8_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredsum_vs_i8mf8_i8m1_m((vbool64_t)(op0), (vint8m1_t)(op1), (vint8mf8_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vredsum_vs_i16m1_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredsum_vs_i16m1_i16m1((vint16m1_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vredsum_vs_i16m1_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredsum_vs_i16m1_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vredsum_vs_i16m2_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredsum_vs_i16m2_i16m1((vint16m1_t)(op0), (vint16m2_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vredsum_vs_i16m2_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredsum_vs_i16m2_i16m1_m((vbool8_t)(op0), (vint16m1_t)(op1), (vint16m2_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vredsum_vs_i16m4_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredsum_vs_i16m4_i16m1((vint16m1_t)(op0), (vint16m4_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vredsum_vs_i16m4_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredsum_vs_i16m4_i16m1_m((vbool4_t)(op0), (vint16m1_t)(op1), (vint16m4_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vredsum_vs_i16m8_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredsum_vs_i16m8_i16m1((vint16m1_t)(op0), (vint16m8_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vredsum_vs_i16m8_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredsum_vs_i16m8_i16m1_m((vbool2_t)(op0), (vint16m1_t)(op1), (vint16m8_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vredsum_vs_i16mf2_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredsum_vs_i16mf2_i16m1((vint16m1_t)(op0), (vint16mf2_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vredsum_vs_i16mf2_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredsum_vs_i16mf2_i16m1_m((vbool32_t)(op0), (vint16m1_t)(op1), (vint16mf2_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vredsum_vs_i16mf4_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredsum_vs_i16mf4_i16m1((vint16m1_t)(op0), (vint16mf4_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vredsum_vs_i16mf4_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredsum_vs_i16mf4_i16m1_m((vbool64_t)(op0), (vint16m1_t)(op1), (vint16mf4_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vredsum_vs_i32m1_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredsum_vs_i32m1_i32m1((vint32m1_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vredsum_vs_i32m1_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredsum_vs_i32m1_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vredsum_vs_i32m2_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredsum_vs_i32m2_i32m1((vint32m1_t)(op0), (vint32m2_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vredsum_vs_i32m2_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredsum_vs_i32m2_i32m1_m((vbool16_t)(op0), (vint32m1_t)(op1), (vint32m2_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vredsum_vs_i32m4_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredsum_vs_i32m4_i32m1((vint32m1_t)(op0), (vint32m4_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vredsum_vs_i32m4_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredsum_vs_i32m4_i32m1_m((vbool8_t)(op0), (vint32m1_t)(op1), (vint32m4_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vredsum_vs_i32m8_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredsum_vs_i32m8_i32m1((vint32m1_t)(op0), (vint32m8_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vredsum_vs_i32m8_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredsum_vs_i32m8_i32m1_m((vbool4_t)(op0), (vint32m1_t)(op1), (vint32m8_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vredsum_vs_i32mf2_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredsum_vs_i32mf2_i32m1((vint32m1_t)(op0), (vint32mf2_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vredsum_vs_i32mf2_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredsum_vs_i32mf2_i32m1_m((vbool64_t)(op0), (vint32m1_t)(op1), (vint32mf2_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vredsum_vs_i64m1_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredsum_vs_i64m1_i64m1((vint64m1_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vredsum_vs_i64m1_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredsum_vs_i64m1_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vredsum_vs_i64m2_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredsum_vs_i64m2_i64m1((vint64m1_t)(op0), (vint64m2_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vredsum_vs_i64m2_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredsum_vs_i64m2_i64m1_m((vbool32_t)(op0), (vint64m1_t)(op1), (vint64m2_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vredsum_vs_i64m4_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredsum_vs_i64m4_i64m1((vint64m1_t)(op0), (vint64m4_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vredsum_vs_i64m4_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredsum_vs_i64m4_i64m1_m((vbool16_t)(op0), (vint64m1_t)(op1), (vint64m4_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vredsum_vs_i64m8_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredsum_vs_i64m8_i64m1((vint64m1_t)(op0), (vint64m8_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vredsum_vs_i64m8_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredsum_vs_i64m8_i64m1_m((vbool8_t)(op0), (vint64m1_t)(op1), (vint64m8_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vredsum_vs_u8m1_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredsum_vs_u8m1_u8m1((vuint8m1_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vredsum_vs_u8m1_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredsum_vs_u8m1_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vredsum_vs_u8m2_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredsum_vs_u8m2_u8m1((vuint8m1_t)(op0), (vuint8m2_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vredsum_vs_u8m2_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredsum_vs_u8m2_u8m1_m((vbool4_t)(op0), (vuint8m1_t)(op1), (vuint8m2_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vredsum_vs_u8m4_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredsum_vs_u8m4_u8m1((vuint8m1_t)(op0), (vuint8m4_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vredsum_vs_u8m4_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredsum_vs_u8m4_u8m1_m((vbool2_t)(op0), (vuint8m1_t)(op1), (vuint8m4_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vredsum_vs_u8m8_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredsum_vs_u8m8_u8m1((vuint8m1_t)(op0), (vuint8m8_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vredsum_vs_u8m8_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredsum_vs_u8m8_u8m1_m((vbool1_t)(op0), (vuint8m1_t)(op1), (vuint8m8_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vredsum_vs_u8mf2_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredsum_vs_u8mf2_u8m1((vuint8m1_t)(op0), (vuint8mf2_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vredsum_vs_u8mf2_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredsum_vs_u8mf2_u8m1_m((vbool16_t)(op0), (vuint8m1_t)(op1), (vuint8mf2_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vredsum_vs_u8mf4_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredsum_vs_u8mf4_u8m1((vuint8m1_t)(op0), (vuint8mf4_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vredsum_vs_u8mf4_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredsum_vs_u8mf4_u8m1_m((vbool32_t)(op0), (vuint8m1_t)(op1), (vuint8mf4_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vredsum_vs_u8mf8_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredsum_vs_u8mf8_u8m1((vuint8m1_t)(op0), (vuint8mf8_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vredsum_vs_u8mf8_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredsum_vs_u8mf8_u8m1_m((vbool64_t)(op0), (vuint8m1_t)(op1), (vuint8mf8_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vredsum_vs_u16m1_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredsum_vs_u16m1_u16m1((vuint16m1_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vredsum_vs_u16m1_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredsum_vs_u16m1_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vredsum_vs_u16m2_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredsum_vs_u16m2_u16m1((vuint16m1_t)(op0), (vuint16m2_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vredsum_vs_u16m2_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredsum_vs_u16m2_u16m1_m((vbool8_t)(op0), (vuint16m1_t)(op1), (vuint16m2_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vredsum_vs_u16m4_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredsum_vs_u16m4_u16m1((vuint16m1_t)(op0), (vuint16m4_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vredsum_vs_u16m4_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredsum_vs_u16m4_u16m1_m((vbool4_t)(op0), (vuint16m1_t)(op1), (vuint16m4_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vredsum_vs_u16m8_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredsum_vs_u16m8_u16m1((vuint16m1_t)(op0), (vuint16m8_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vredsum_vs_u16m8_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredsum_vs_u16m8_u16m1_m((vbool2_t)(op0), (vuint16m1_t)(op1), (vuint16m8_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vredsum_vs_u16mf2_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredsum_vs_u16mf2_u16m1((vuint16m1_t)(op0), (vuint16mf2_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vredsum_vs_u16mf2_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredsum_vs_u16mf2_u16m1_m((vbool32_t)(op0), (vuint16m1_t)(op1), (vuint16mf2_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vredsum_vs_u16mf4_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredsum_vs_u16mf4_u16m1((vuint16m1_t)(op0), (vuint16mf4_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vredsum_vs_u16mf4_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredsum_vs_u16mf4_u16m1_m((vbool64_t)(op0), (vuint16m1_t)(op1), (vuint16mf4_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vredsum_vs_u32m1_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredsum_vs_u32m1_u32m1((vuint32m1_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vredsum_vs_u32m1_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredsum_vs_u32m1_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vredsum_vs_u32m2_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredsum_vs_u32m2_u32m1((vuint32m1_t)(op0), (vuint32m2_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vredsum_vs_u32m2_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredsum_vs_u32m2_u32m1_m((vbool16_t)(op0), (vuint32m1_t)(op1), (vuint32m2_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vredsum_vs_u32m4_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredsum_vs_u32m4_u32m1((vuint32m1_t)(op0), (vuint32m4_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vredsum_vs_u32m4_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredsum_vs_u32m4_u32m1_m((vbool8_t)(op0), (vuint32m1_t)(op1), (vuint32m4_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vredsum_vs_u32m8_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredsum_vs_u32m8_u32m1((vuint32m1_t)(op0), (vuint32m8_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vredsum_vs_u32m8_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredsum_vs_u32m8_u32m1_m((vbool4_t)(op0), (vuint32m1_t)(op1), (vuint32m8_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vredsum_vs_u32mf2_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredsum_vs_u32mf2_u32m1((vuint32m1_t)(op0), (vuint32mf2_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vredsum_vs_u32mf2_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredsum_vs_u32mf2_u32m1_m((vbool64_t)(op0), (vuint32m1_t)(op1), (vuint32mf2_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vredsum_vs_u64m1_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredsum_vs_u64m1_u64m1((vuint64m1_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vredsum_vs_u64m1_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredsum_vs_u64m1_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vredsum_vs_u64m2_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredsum_vs_u64m2_u64m1((vuint64m1_t)(op0), (vuint64m2_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vredsum_vs_u64m2_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredsum_vs_u64m2_u64m1_m((vbool32_t)(op0), (vuint64m1_t)(op1), (vuint64m2_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vredsum_vs_u64m4_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredsum_vs_u64m4_u64m1((vuint64m1_t)(op0), (vuint64m4_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vredsum_vs_u64m4_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredsum_vs_u64m4_u64m1_m((vbool16_t)(op0), (vuint64m1_t)(op1), (vuint64m4_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vredsum_vs_u64m8_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredsum_vs_u64m8_u64m1((vuint64m1_t)(op0), (vuint64m8_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vredsum_vs_u64m8_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredsum_vs_u64m8_u64m1_m((vbool8_t)(op0), (vuint64m1_t)(op1), (vuint64m8_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vredmaxu_vs_u8m1_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmaxu_vs_u8m1_u8m1((vuint8m1_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vredmaxu_vs_u8m1_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmaxu_vs_u8m1_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vredmaxu_vs_u8m2_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmaxu_vs_u8m2_u8m1((vuint8m1_t)(op0), (vuint8m2_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vredmaxu_vs_u8m2_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmaxu_vs_u8m2_u8m1_m((vbool4_t)(op0), (vuint8m1_t)(op1), (vuint8m2_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vredmaxu_vs_u8m4_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmaxu_vs_u8m4_u8m1((vuint8m1_t)(op0), (vuint8m4_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vredmaxu_vs_u8m4_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmaxu_vs_u8m4_u8m1_m((vbool2_t)(op0), (vuint8m1_t)(op1), (vuint8m4_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vredmaxu_vs_u8m8_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmaxu_vs_u8m8_u8m1((vuint8m1_t)(op0), (vuint8m8_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vredmaxu_vs_u8m8_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmaxu_vs_u8m8_u8m1_m((vbool1_t)(op0), (vuint8m1_t)(op1), (vuint8m8_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vredmaxu_vs_u8mf2_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmaxu_vs_u8mf2_u8m1((vuint8m1_t)(op0), (vuint8mf2_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vredmaxu_vs_u8mf2_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmaxu_vs_u8mf2_u8m1_m((vbool16_t)(op0), (vuint8m1_t)(op1), (vuint8mf2_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vredmaxu_vs_u8mf4_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmaxu_vs_u8mf4_u8m1((vuint8m1_t)(op0), (vuint8mf4_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vredmaxu_vs_u8mf4_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmaxu_vs_u8mf4_u8m1_m((vbool32_t)(op0), (vuint8m1_t)(op1), (vuint8mf4_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vredmaxu_vs_u8mf8_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmaxu_vs_u8mf8_u8m1((vuint8m1_t)(op0), (vuint8mf8_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vredmaxu_vs_u8mf8_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmaxu_vs_u8mf8_u8m1_m((vbool64_t)(op0), (vuint8m1_t)(op1), (vuint8mf8_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vredmaxu_vs_u16m1_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmaxu_vs_u16m1_u16m1((vuint16m1_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vredmaxu_vs_u16m1_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmaxu_vs_u16m1_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vredmaxu_vs_u16m2_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmaxu_vs_u16m2_u16m1((vuint16m1_t)(op0), (vuint16m2_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vredmaxu_vs_u16m2_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmaxu_vs_u16m2_u16m1_m((vbool8_t)(op0), (vuint16m1_t)(op1), (vuint16m2_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vredmaxu_vs_u16m4_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmaxu_vs_u16m4_u16m1((vuint16m1_t)(op0), (vuint16m4_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vredmaxu_vs_u16m4_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmaxu_vs_u16m4_u16m1_m((vbool4_t)(op0), (vuint16m1_t)(op1), (vuint16m4_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vredmaxu_vs_u16m8_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmaxu_vs_u16m8_u16m1((vuint16m1_t)(op0), (vuint16m8_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vredmaxu_vs_u16m8_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmaxu_vs_u16m8_u16m1_m((vbool2_t)(op0), (vuint16m1_t)(op1), (vuint16m8_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vredmaxu_vs_u16mf2_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmaxu_vs_u16mf2_u16m1((vuint16m1_t)(op0), (vuint16mf2_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vredmaxu_vs_u16mf2_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmaxu_vs_u16mf2_u16m1_m((vbool32_t)(op0), (vuint16m1_t)(op1), (vuint16mf2_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vredmaxu_vs_u16mf4_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmaxu_vs_u16mf4_u16m1((vuint16m1_t)(op0), (vuint16mf4_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vredmaxu_vs_u16mf4_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmaxu_vs_u16mf4_u16m1_m((vbool64_t)(op0), (vuint16m1_t)(op1), (vuint16mf4_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vredmaxu_vs_u32m1_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmaxu_vs_u32m1_u32m1((vuint32m1_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vredmaxu_vs_u32m1_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmaxu_vs_u32m1_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vredmaxu_vs_u32m2_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmaxu_vs_u32m2_u32m1((vuint32m1_t)(op0), (vuint32m2_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vredmaxu_vs_u32m2_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmaxu_vs_u32m2_u32m1_m((vbool16_t)(op0), (vuint32m1_t)(op1), (vuint32m2_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vredmaxu_vs_u32m4_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmaxu_vs_u32m4_u32m1((vuint32m1_t)(op0), (vuint32m4_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vredmaxu_vs_u32m4_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmaxu_vs_u32m4_u32m1_m((vbool8_t)(op0), (vuint32m1_t)(op1), (vuint32m4_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vredmaxu_vs_u32m8_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmaxu_vs_u32m8_u32m1((vuint32m1_t)(op0), (vuint32m8_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vredmaxu_vs_u32m8_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmaxu_vs_u32m8_u32m1_m((vbool4_t)(op0), (vuint32m1_t)(op1), (vuint32m8_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vredmaxu_vs_u32mf2_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmaxu_vs_u32mf2_u32m1((vuint32m1_t)(op0), (vuint32mf2_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vredmaxu_vs_u32mf2_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmaxu_vs_u32mf2_u32m1_m((vbool64_t)(op0), (vuint32m1_t)(op1), (vuint32mf2_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vredmaxu_vs_u64m1_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmaxu_vs_u64m1_u64m1((vuint64m1_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vredmaxu_vs_u64m1_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmaxu_vs_u64m1_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vredmaxu_vs_u64m2_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmaxu_vs_u64m2_u64m1((vuint64m1_t)(op0), (vuint64m2_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vredmaxu_vs_u64m2_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmaxu_vs_u64m2_u64m1_m((vbool32_t)(op0), (vuint64m1_t)(op1), (vuint64m2_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vredmaxu_vs_u64m4_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmaxu_vs_u64m4_u64m1((vuint64m1_t)(op0), (vuint64m4_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vredmaxu_vs_u64m4_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmaxu_vs_u64m4_u64m1_m((vbool16_t)(op0), (vuint64m1_t)(op1), (vuint64m4_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vredmaxu_vs_u64m8_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmaxu_vs_u64m8_u64m1((vuint64m1_t)(op0), (vuint64m8_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vredmaxu_vs_u64m8_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmaxu_vs_u64m8_u64m1_m((vbool8_t)(op0), (vuint64m1_t)(op1), (vuint64m8_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vredmax_vs_i8m1_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmax_vs_i8m1_i8m1((vint8m1_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vredmax_vs_i8m1_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmax_vs_i8m1_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vredmax_vs_i8m2_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmax_vs_i8m2_i8m1((vint8m1_t)(op0), (vint8m2_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vredmax_vs_i8m2_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmax_vs_i8m2_i8m1_m((vbool4_t)(op0), (vint8m1_t)(op1), (vint8m2_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vredmax_vs_i8m4_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmax_vs_i8m4_i8m1((vint8m1_t)(op0), (vint8m4_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vredmax_vs_i8m4_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmax_vs_i8m4_i8m1_m((vbool2_t)(op0), (vint8m1_t)(op1), (vint8m4_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vredmax_vs_i8m8_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmax_vs_i8m8_i8m1((vint8m1_t)(op0), (vint8m8_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vredmax_vs_i8m8_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmax_vs_i8m8_i8m1_m((vbool1_t)(op0), (vint8m1_t)(op1), (vint8m8_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vredmax_vs_i8mf2_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmax_vs_i8mf2_i8m1((vint8m1_t)(op0), (vint8mf2_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vredmax_vs_i8mf2_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmax_vs_i8mf2_i8m1_m((vbool16_t)(op0), (vint8m1_t)(op1), (vint8mf2_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vredmax_vs_i8mf4_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmax_vs_i8mf4_i8m1((vint8m1_t)(op0), (vint8mf4_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vredmax_vs_i8mf4_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmax_vs_i8mf4_i8m1_m((vbool32_t)(op0), (vint8m1_t)(op1), (vint8mf4_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vredmax_vs_i8mf8_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmax_vs_i8mf8_i8m1((vint8m1_t)(op0), (vint8mf8_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vredmax_vs_i8mf8_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmax_vs_i8mf8_i8m1_m((vbool64_t)(op0), (vint8m1_t)(op1), (vint8mf8_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vredmax_vs_i16m1_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmax_vs_i16m1_i16m1((vint16m1_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vredmax_vs_i16m1_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmax_vs_i16m1_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vredmax_vs_i16m2_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmax_vs_i16m2_i16m1((vint16m1_t)(op0), (vint16m2_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vredmax_vs_i16m2_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmax_vs_i16m2_i16m1_m((vbool8_t)(op0), (vint16m1_t)(op1), (vint16m2_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vredmax_vs_i16m4_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmax_vs_i16m4_i16m1((vint16m1_t)(op0), (vint16m4_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vredmax_vs_i16m4_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmax_vs_i16m4_i16m1_m((vbool4_t)(op0), (vint16m1_t)(op1), (vint16m4_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vredmax_vs_i16m8_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmax_vs_i16m8_i16m1((vint16m1_t)(op0), (vint16m8_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vredmax_vs_i16m8_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmax_vs_i16m8_i16m1_m((vbool2_t)(op0), (vint16m1_t)(op1), (vint16m8_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vredmax_vs_i16mf2_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmax_vs_i16mf2_i16m1((vint16m1_t)(op0), (vint16mf2_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vredmax_vs_i16mf2_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmax_vs_i16mf2_i16m1_m((vbool32_t)(op0), (vint16m1_t)(op1), (vint16mf2_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vredmax_vs_i16mf4_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmax_vs_i16mf4_i16m1((vint16m1_t)(op0), (vint16mf4_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vredmax_vs_i16mf4_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmax_vs_i16mf4_i16m1_m((vbool64_t)(op0), (vint16m1_t)(op1), (vint16mf4_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vredmax_vs_i32m1_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmax_vs_i32m1_i32m1((vint32m1_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vredmax_vs_i32m1_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmax_vs_i32m1_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vredmax_vs_i32m2_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmax_vs_i32m2_i32m1((vint32m1_t)(op0), (vint32m2_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vredmax_vs_i32m2_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmax_vs_i32m2_i32m1_m((vbool16_t)(op0), (vint32m1_t)(op1), (vint32m2_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vredmax_vs_i32m4_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmax_vs_i32m4_i32m1((vint32m1_t)(op0), (vint32m4_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vredmax_vs_i32m4_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmax_vs_i32m4_i32m1_m((vbool8_t)(op0), (vint32m1_t)(op1), (vint32m4_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vredmax_vs_i32m8_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmax_vs_i32m8_i32m1((vint32m1_t)(op0), (vint32m8_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vredmax_vs_i32m8_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmax_vs_i32m8_i32m1_m((vbool4_t)(op0), (vint32m1_t)(op1), (vint32m8_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vredmax_vs_i32mf2_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmax_vs_i32mf2_i32m1((vint32m1_t)(op0), (vint32mf2_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vredmax_vs_i32mf2_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmax_vs_i32mf2_i32m1_m((vbool64_t)(op0), (vint32m1_t)(op1), (vint32mf2_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vredmax_vs_i64m1_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmax_vs_i64m1_i64m1((vint64m1_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vredmax_vs_i64m1_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmax_vs_i64m1_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vredmax_vs_i64m2_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmax_vs_i64m2_i64m1((vint64m1_t)(op0), (vint64m2_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vredmax_vs_i64m2_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmax_vs_i64m2_i64m1_m((vbool32_t)(op0), (vint64m1_t)(op1), (vint64m2_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vredmax_vs_i64m4_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmax_vs_i64m4_i64m1((vint64m1_t)(op0), (vint64m4_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vredmax_vs_i64m4_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmax_vs_i64m4_i64m1_m((vbool16_t)(op0), (vint64m1_t)(op1), (vint64m4_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vredmax_vs_i64m8_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmax_vs_i64m8_i64m1((vint64m1_t)(op0), (vint64m8_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vredmax_vs_i64m8_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmax_vs_i64m8_i64m1_m((vbool8_t)(op0), (vint64m1_t)(op1), (vint64m8_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vredminu_vs_u8m1_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredminu_vs_u8m1_u8m1((vuint8m1_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vredminu_vs_u8m1_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredminu_vs_u8m1_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vredminu_vs_u8m2_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredminu_vs_u8m2_u8m1((vuint8m1_t)(op0), (vuint8m2_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vredminu_vs_u8m2_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredminu_vs_u8m2_u8m1_m((vbool4_t)(op0), (vuint8m1_t)(op1), (vuint8m2_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vredminu_vs_u8m4_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredminu_vs_u8m4_u8m1((vuint8m1_t)(op0), (vuint8m4_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vredminu_vs_u8m4_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredminu_vs_u8m4_u8m1_m((vbool2_t)(op0), (vuint8m1_t)(op1), (vuint8m4_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vredminu_vs_u8m8_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredminu_vs_u8m8_u8m1((vuint8m1_t)(op0), (vuint8m8_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vredminu_vs_u8m8_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredminu_vs_u8m8_u8m1_m((vbool1_t)(op0), (vuint8m1_t)(op1), (vuint8m8_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vredminu_vs_u8mf2_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredminu_vs_u8mf2_u8m1((vuint8m1_t)(op0), (vuint8mf2_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vredminu_vs_u8mf2_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredminu_vs_u8mf2_u8m1_m((vbool16_t)(op0), (vuint8m1_t)(op1), (vuint8mf2_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vredminu_vs_u8mf4_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredminu_vs_u8mf4_u8m1((vuint8m1_t)(op0), (vuint8mf4_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vredminu_vs_u8mf4_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredminu_vs_u8mf4_u8m1_m((vbool32_t)(op0), (vuint8m1_t)(op1), (vuint8mf4_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vredminu_vs_u8mf8_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredminu_vs_u8mf8_u8m1((vuint8m1_t)(op0), (vuint8mf8_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vredminu_vs_u8mf8_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredminu_vs_u8mf8_u8m1_m((vbool64_t)(op0), (vuint8m1_t)(op1), (vuint8mf8_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vredminu_vs_u16m1_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredminu_vs_u16m1_u16m1((vuint16m1_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vredminu_vs_u16m1_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredminu_vs_u16m1_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vredminu_vs_u16m2_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredminu_vs_u16m2_u16m1((vuint16m1_t)(op0), (vuint16m2_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vredminu_vs_u16m2_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredminu_vs_u16m2_u16m1_m((vbool8_t)(op0), (vuint16m1_t)(op1), (vuint16m2_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vredminu_vs_u16m4_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredminu_vs_u16m4_u16m1((vuint16m1_t)(op0), (vuint16m4_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vredminu_vs_u16m4_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredminu_vs_u16m4_u16m1_m((vbool4_t)(op0), (vuint16m1_t)(op1), (vuint16m4_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vredminu_vs_u16m8_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredminu_vs_u16m8_u16m1((vuint16m1_t)(op0), (vuint16m8_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vredminu_vs_u16m8_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredminu_vs_u16m8_u16m1_m((vbool2_t)(op0), (vuint16m1_t)(op1), (vuint16m8_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vredminu_vs_u16mf2_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredminu_vs_u16mf2_u16m1((vuint16m1_t)(op0), (vuint16mf2_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vredminu_vs_u16mf2_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredminu_vs_u16mf2_u16m1_m((vbool32_t)(op0), (vuint16m1_t)(op1), (vuint16mf2_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vredminu_vs_u16mf4_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredminu_vs_u16mf4_u16m1((vuint16m1_t)(op0), (vuint16mf4_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vredminu_vs_u16mf4_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredminu_vs_u16mf4_u16m1_m((vbool64_t)(op0), (vuint16m1_t)(op1), (vuint16mf4_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vredminu_vs_u32m1_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredminu_vs_u32m1_u32m1((vuint32m1_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vredminu_vs_u32m1_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredminu_vs_u32m1_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vredminu_vs_u32m2_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredminu_vs_u32m2_u32m1((vuint32m1_t)(op0), (vuint32m2_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vredminu_vs_u32m2_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredminu_vs_u32m2_u32m1_m((vbool16_t)(op0), (vuint32m1_t)(op1), (vuint32m2_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vredminu_vs_u32m4_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredminu_vs_u32m4_u32m1((vuint32m1_t)(op0), (vuint32m4_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vredminu_vs_u32m4_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredminu_vs_u32m4_u32m1_m((vbool8_t)(op0), (vuint32m1_t)(op1), (vuint32m4_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vredminu_vs_u32m8_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredminu_vs_u32m8_u32m1((vuint32m1_t)(op0), (vuint32m8_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vredminu_vs_u32m8_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredminu_vs_u32m8_u32m1_m((vbool4_t)(op0), (vuint32m1_t)(op1), (vuint32m8_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vredminu_vs_u32mf2_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredminu_vs_u32mf2_u32m1((vuint32m1_t)(op0), (vuint32mf2_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vredminu_vs_u32mf2_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredminu_vs_u32mf2_u32m1_m((vbool64_t)(op0), (vuint32m1_t)(op1), (vuint32mf2_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vredminu_vs_u64m1_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredminu_vs_u64m1_u64m1((vuint64m1_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vredminu_vs_u64m1_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredminu_vs_u64m1_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vredminu_vs_u64m2_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredminu_vs_u64m2_u64m1((vuint64m1_t)(op0), (vuint64m2_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vredminu_vs_u64m2_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredminu_vs_u64m2_u64m1_m((vbool32_t)(op0), (vuint64m1_t)(op1), (vuint64m2_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vredminu_vs_u64m4_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredminu_vs_u64m4_u64m1((vuint64m1_t)(op0), (vuint64m4_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vredminu_vs_u64m4_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredminu_vs_u64m4_u64m1_m((vbool16_t)(op0), (vuint64m1_t)(op1), (vuint64m4_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vredminu_vs_u64m8_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredminu_vs_u64m8_u64m1((vuint64m1_t)(op0), (vuint64m8_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vredminu_vs_u64m8_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredminu_vs_u64m8_u64m1_m((vbool8_t)(op0), (vuint64m1_t)(op1), (vuint64m8_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vredmin_vs_i8m1_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmin_vs_i8m1_i8m1((vint8m1_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vredmin_vs_i8m1_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmin_vs_i8m1_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vredmin_vs_i8m2_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmin_vs_i8m2_i8m1((vint8m1_t)(op0), (vint8m2_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vredmin_vs_i8m2_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmin_vs_i8m2_i8m1_m((vbool4_t)(op0), (vint8m1_t)(op1), (vint8m2_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vredmin_vs_i8m4_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmin_vs_i8m4_i8m1((vint8m1_t)(op0), (vint8m4_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vredmin_vs_i8m4_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmin_vs_i8m4_i8m1_m((vbool2_t)(op0), (vint8m1_t)(op1), (vint8m4_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vredmin_vs_i8m8_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmin_vs_i8m8_i8m1((vint8m1_t)(op0), (vint8m8_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vredmin_vs_i8m8_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmin_vs_i8m8_i8m1_m((vbool1_t)(op0), (vint8m1_t)(op1), (vint8m8_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vredmin_vs_i8mf2_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmin_vs_i8mf2_i8m1((vint8m1_t)(op0), (vint8mf2_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vredmin_vs_i8mf2_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmin_vs_i8mf2_i8m1_m((vbool16_t)(op0), (vint8m1_t)(op1), (vint8mf2_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vredmin_vs_i8mf4_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmin_vs_i8mf4_i8m1((vint8m1_t)(op0), (vint8mf4_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vredmin_vs_i8mf4_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmin_vs_i8mf4_i8m1_m((vbool32_t)(op0), (vint8m1_t)(op1), (vint8mf4_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vredmin_vs_i8mf8_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmin_vs_i8mf8_i8m1((vint8m1_t)(op0), (vint8mf8_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vredmin_vs_i8mf8_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmin_vs_i8mf8_i8m1_m((vbool64_t)(op0), (vint8m1_t)(op1), (vint8mf8_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vredmin_vs_i16m1_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmin_vs_i16m1_i16m1((vint16m1_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vredmin_vs_i16m1_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmin_vs_i16m1_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vredmin_vs_i16m2_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmin_vs_i16m2_i16m1((vint16m1_t)(op0), (vint16m2_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vredmin_vs_i16m2_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmin_vs_i16m2_i16m1_m((vbool8_t)(op0), (vint16m1_t)(op1), (vint16m2_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vredmin_vs_i16m4_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmin_vs_i16m4_i16m1((vint16m1_t)(op0), (vint16m4_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vredmin_vs_i16m4_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmin_vs_i16m4_i16m1_m((vbool4_t)(op0), (vint16m1_t)(op1), (vint16m4_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vredmin_vs_i16m8_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmin_vs_i16m8_i16m1((vint16m1_t)(op0), (vint16m8_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vredmin_vs_i16m8_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmin_vs_i16m8_i16m1_m((vbool2_t)(op0), (vint16m1_t)(op1), (vint16m8_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vredmin_vs_i16mf2_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmin_vs_i16mf2_i16m1((vint16m1_t)(op0), (vint16mf2_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vredmin_vs_i16mf2_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmin_vs_i16mf2_i16m1_m((vbool32_t)(op0), (vint16m1_t)(op1), (vint16mf2_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vredmin_vs_i16mf4_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmin_vs_i16mf4_i16m1((vint16m1_t)(op0), (vint16mf4_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vredmin_vs_i16mf4_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmin_vs_i16mf4_i16m1_m((vbool64_t)(op0), (vint16m1_t)(op1), (vint16mf4_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vredmin_vs_i32m1_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmin_vs_i32m1_i32m1((vint32m1_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vredmin_vs_i32m1_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmin_vs_i32m1_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vredmin_vs_i32m2_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmin_vs_i32m2_i32m1((vint32m1_t)(op0), (vint32m2_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vredmin_vs_i32m2_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmin_vs_i32m2_i32m1_m((vbool16_t)(op0), (vint32m1_t)(op1), (vint32m2_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vredmin_vs_i32m4_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmin_vs_i32m4_i32m1((vint32m1_t)(op0), (vint32m4_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vredmin_vs_i32m4_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmin_vs_i32m4_i32m1_m((vbool8_t)(op0), (vint32m1_t)(op1), (vint32m4_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vredmin_vs_i32m8_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmin_vs_i32m8_i32m1((vint32m1_t)(op0), (vint32m8_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vredmin_vs_i32m8_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmin_vs_i32m8_i32m1_m((vbool4_t)(op0), (vint32m1_t)(op1), (vint32m8_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vredmin_vs_i32mf2_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmin_vs_i32mf2_i32m1((vint32m1_t)(op0), (vint32mf2_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vredmin_vs_i32mf2_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmin_vs_i32mf2_i32m1_m((vbool64_t)(op0), (vint32m1_t)(op1), (vint32mf2_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vredmin_vs_i64m1_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmin_vs_i64m1_i64m1((vint64m1_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vredmin_vs_i64m1_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmin_vs_i64m1_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vredmin_vs_i64m2_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmin_vs_i64m2_i64m1((vint64m1_t)(op0), (vint64m2_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vredmin_vs_i64m2_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmin_vs_i64m2_i64m1_m((vbool32_t)(op0), (vint64m1_t)(op1), (vint64m2_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vredmin_vs_i64m4_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmin_vs_i64m4_i64m1((vint64m1_t)(op0), (vint64m4_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vredmin_vs_i64m4_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmin_vs_i64m4_i64m1_m((vbool16_t)(op0), (vint64m1_t)(op1), (vint64m4_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vredmin_vs_i64m8_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredmin_vs_i64m8_i64m1((vint64m1_t)(op0), (vint64m8_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vredmin_vs_i64m8_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredmin_vs_i64m8_i64m1_m((vbool8_t)(op0), (vint64m1_t)(op1), (vint64m8_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vredand_vs_i8m1_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredand_vs_i8m1_i8m1((vint8m1_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vredand_vs_i8m1_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredand_vs_i8m1_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vredand_vs_i8m2_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredand_vs_i8m2_i8m1((vint8m1_t)(op0), (vint8m2_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vredand_vs_i8m2_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredand_vs_i8m2_i8m1_m((vbool4_t)(op0), (vint8m1_t)(op1), (vint8m2_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vredand_vs_i8m4_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredand_vs_i8m4_i8m1((vint8m1_t)(op0), (vint8m4_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vredand_vs_i8m4_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredand_vs_i8m4_i8m1_m((vbool2_t)(op0), (vint8m1_t)(op1), (vint8m4_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vredand_vs_i8m8_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredand_vs_i8m8_i8m1((vint8m1_t)(op0), (vint8m8_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vredand_vs_i8m8_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredand_vs_i8m8_i8m1_m((vbool1_t)(op0), (vint8m1_t)(op1), (vint8m8_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vredand_vs_i8mf2_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredand_vs_i8mf2_i8m1((vint8m1_t)(op0), (vint8mf2_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vredand_vs_i8mf2_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredand_vs_i8mf2_i8m1_m((vbool16_t)(op0), (vint8m1_t)(op1), (vint8mf2_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vredand_vs_i8mf4_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredand_vs_i8mf4_i8m1((vint8m1_t)(op0), (vint8mf4_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vredand_vs_i8mf4_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredand_vs_i8mf4_i8m1_m((vbool32_t)(op0), (vint8m1_t)(op1), (vint8mf4_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vredand_vs_i8mf8_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredand_vs_i8mf8_i8m1((vint8m1_t)(op0), (vint8mf8_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vredand_vs_i8mf8_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredand_vs_i8mf8_i8m1_m((vbool64_t)(op0), (vint8m1_t)(op1), (vint8mf8_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vredand_vs_i16m1_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredand_vs_i16m1_i16m1((vint16m1_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vredand_vs_i16m1_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredand_vs_i16m1_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vredand_vs_i16m2_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredand_vs_i16m2_i16m1((vint16m1_t)(op0), (vint16m2_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vredand_vs_i16m2_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredand_vs_i16m2_i16m1_m((vbool8_t)(op0), (vint16m1_t)(op1), (vint16m2_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vredand_vs_i16m4_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredand_vs_i16m4_i16m1((vint16m1_t)(op0), (vint16m4_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vredand_vs_i16m4_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredand_vs_i16m4_i16m1_m((vbool4_t)(op0), (vint16m1_t)(op1), (vint16m4_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vredand_vs_i16m8_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredand_vs_i16m8_i16m1((vint16m1_t)(op0), (vint16m8_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vredand_vs_i16m8_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredand_vs_i16m8_i16m1_m((vbool2_t)(op0), (vint16m1_t)(op1), (vint16m8_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vredand_vs_i16mf2_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredand_vs_i16mf2_i16m1((vint16m1_t)(op0), (vint16mf2_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vredand_vs_i16mf2_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredand_vs_i16mf2_i16m1_m((vbool32_t)(op0), (vint16m1_t)(op1), (vint16mf2_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vredand_vs_i16mf4_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredand_vs_i16mf4_i16m1((vint16m1_t)(op0), (vint16mf4_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vredand_vs_i16mf4_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredand_vs_i16mf4_i16m1_m((vbool64_t)(op0), (vint16m1_t)(op1), (vint16mf4_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vredand_vs_i32m1_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredand_vs_i32m1_i32m1((vint32m1_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vredand_vs_i32m1_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredand_vs_i32m1_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vredand_vs_i32m2_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredand_vs_i32m2_i32m1((vint32m1_t)(op0), (vint32m2_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vredand_vs_i32m2_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredand_vs_i32m2_i32m1_m((vbool16_t)(op0), (vint32m1_t)(op1), (vint32m2_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vredand_vs_i32m4_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredand_vs_i32m4_i32m1((vint32m1_t)(op0), (vint32m4_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vredand_vs_i32m4_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredand_vs_i32m4_i32m1_m((vbool8_t)(op0), (vint32m1_t)(op1), (vint32m4_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vredand_vs_i32m8_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredand_vs_i32m8_i32m1((vint32m1_t)(op0), (vint32m8_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vredand_vs_i32m8_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredand_vs_i32m8_i32m1_m((vbool4_t)(op0), (vint32m1_t)(op1), (vint32m8_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vredand_vs_i32mf2_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredand_vs_i32mf2_i32m1((vint32m1_t)(op0), (vint32mf2_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vredand_vs_i32mf2_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredand_vs_i32mf2_i32m1_m((vbool64_t)(op0), (vint32m1_t)(op1), (vint32mf2_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vredand_vs_i64m1_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredand_vs_i64m1_i64m1((vint64m1_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vredand_vs_i64m1_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredand_vs_i64m1_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vredand_vs_i64m2_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredand_vs_i64m2_i64m1((vint64m1_t)(op0), (vint64m2_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vredand_vs_i64m2_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredand_vs_i64m2_i64m1_m((vbool32_t)(op0), (vint64m1_t)(op1), (vint64m2_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vredand_vs_i64m4_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredand_vs_i64m4_i64m1((vint64m1_t)(op0), (vint64m4_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vredand_vs_i64m4_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredand_vs_i64m4_i64m1_m((vbool16_t)(op0), (vint64m1_t)(op1), (vint64m4_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vredand_vs_i64m8_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredand_vs_i64m8_i64m1((vint64m1_t)(op0), (vint64m8_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vredand_vs_i64m8_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredand_vs_i64m8_i64m1_m((vbool8_t)(op0), (vint64m1_t)(op1), (vint64m8_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vredand_vs_u8m1_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredand_vs_u8m1_u8m1((vuint8m1_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vredand_vs_u8m1_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredand_vs_u8m1_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vredand_vs_u8m2_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredand_vs_u8m2_u8m1((vuint8m1_t)(op0), (vuint8m2_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vredand_vs_u8m2_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredand_vs_u8m2_u8m1_m((vbool4_t)(op0), (vuint8m1_t)(op1), (vuint8m2_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vredand_vs_u8m4_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredand_vs_u8m4_u8m1((vuint8m1_t)(op0), (vuint8m4_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vredand_vs_u8m4_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredand_vs_u8m4_u8m1_m((vbool2_t)(op0), (vuint8m1_t)(op1), (vuint8m4_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vredand_vs_u8m8_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredand_vs_u8m8_u8m1((vuint8m1_t)(op0), (vuint8m8_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vredand_vs_u8m8_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredand_vs_u8m8_u8m1_m((vbool1_t)(op0), (vuint8m1_t)(op1), (vuint8m8_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vredand_vs_u8mf2_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredand_vs_u8mf2_u8m1((vuint8m1_t)(op0), (vuint8mf2_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vredand_vs_u8mf2_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredand_vs_u8mf2_u8m1_m((vbool16_t)(op0), (vuint8m1_t)(op1), (vuint8mf2_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vredand_vs_u8mf4_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredand_vs_u8mf4_u8m1((vuint8m1_t)(op0), (vuint8mf4_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vredand_vs_u8mf4_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredand_vs_u8mf4_u8m1_m((vbool32_t)(op0), (vuint8m1_t)(op1), (vuint8mf4_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vredand_vs_u8mf8_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredand_vs_u8mf8_u8m1((vuint8m1_t)(op0), (vuint8mf8_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vredand_vs_u8mf8_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredand_vs_u8mf8_u8m1_m((vbool64_t)(op0), (vuint8m1_t)(op1), (vuint8mf8_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vredand_vs_u16m1_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredand_vs_u16m1_u16m1((vuint16m1_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vredand_vs_u16m1_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredand_vs_u16m1_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vredand_vs_u16m2_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredand_vs_u16m2_u16m1((vuint16m1_t)(op0), (vuint16m2_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vredand_vs_u16m2_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredand_vs_u16m2_u16m1_m((vbool8_t)(op0), (vuint16m1_t)(op1), (vuint16m2_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vredand_vs_u16m4_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredand_vs_u16m4_u16m1((vuint16m1_t)(op0), (vuint16m4_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vredand_vs_u16m4_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredand_vs_u16m4_u16m1_m((vbool4_t)(op0), (vuint16m1_t)(op1), (vuint16m4_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vredand_vs_u16m8_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredand_vs_u16m8_u16m1((vuint16m1_t)(op0), (vuint16m8_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vredand_vs_u16m8_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredand_vs_u16m8_u16m1_m((vbool2_t)(op0), (vuint16m1_t)(op1), (vuint16m8_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vredand_vs_u16mf2_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredand_vs_u16mf2_u16m1((vuint16m1_t)(op0), (vuint16mf2_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vredand_vs_u16mf2_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredand_vs_u16mf2_u16m1_m((vbool32_t)(op0), (vuint16m1_t)(op1), (vuint16mf2_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vredand_vs_u16mf4_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredand_vs_u16mf4_u16m1((vuint16m1_t)(op0), (vuint16mf4_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vredand_vs_u16mf4_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredand_vs_u16mf4_u16m1_m((vbool64_t)(op0), (vuint16m1_t)(op1), (vuint16mf4_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vredand_vs_u32m1_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredand_vs_u32m1_u32m1((vuint32m1_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vredand_vs_u32m1_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredand_vs_u32m1_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vredand_vs_u32m2_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredand_vs_u32m2_u32m1((vuint32m1_t)(op0), (vuint32m2_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vredand_vs_u32m2_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredand_vs_u32m2_u32m1_m((vbool16_t)(op0), (vuint32m1_t)(op1), (vuint32m2_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vredand_vs_u32m4_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredand_vs_u32m4_u32m1((vuint32m1_t)(op0), (vuint32m4_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vredand_vs_u32m4_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredand_vs_u32m4_u32m1_m((vbool8_t)(op0), (vuint32m1_t)(op1), (vuint32m4_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vredand_vs_u32m8_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredand_vs_u32m8_u32m1((vuint32m1_t)(op0), (vuint32m8_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vredand_vs_u32m8_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredand_vs_u32m8_u32m1_m((vbool4_t)(op0), (vuint32m1_t)(op1), (vuint32m8_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vredand_vs_u32mf2_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredand_vs_u32mf2_u32m1((vuint32m1_t)(op0), (vuint32mf2_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vredand_vs_u32mf2_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredand_vs_u32mf2_u32m1_m((vbool64_t)(op0), (vuint32m1_t)(op1), (vuint32mf2_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vredand_vs_u64m1_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredand_vs_u64m1_u64m1((vuint64m1_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vredand_vs_u64m1_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredand_vs_u64m1_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vredand_vs_u64m2_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredand_vs_u64m2_u64m1((vuint64m1_t)(op0), (vuint64m2_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vredand_vs_u64m2_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredand_vs_u64m2_u64m1_m((vbool32_t)(op0), (vuint64m1_t)(op1), (vuint64m2_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vredand_vs_u64m4_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredand_vs_u64m4_u64m1((vuint64m1_t)(op0), (vuint64m4_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vredand_vs_u64m4_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredand_vs_u64m4_u64m1_m((vbool16_t)(op0), (vuint64m1_t)(op1), (vuint64m4_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vredand_vs_u64m8_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredand_vs_u64m8_u64m1((vuint64m1_t)(op0), (vuint64m8_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vredand_vs_u64m8_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredand_vs_u64m8_u64m1_m((vbool8_t)(op0), (vuint64m1_t)(op1), (vuint64m8_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vsuxei64_v_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei64_v_u64m1((uint64_t *)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vsuxei64_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei64_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vsuxei64_v_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei64_v_u64m2((uint64_t *)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vsuxei64_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei64_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vsuxei64_v_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei64_v_u64m4((uint64_t *)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vsuxei64_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei64_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vsuxei64_v_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei64_v_u64m8((uint64_t *)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vsuxei64_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei64_v_u64m8_m((vbool8_t)(op0), (uint64_t *)(op1), (vuint64m8_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vredor_vs_i8m1_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredor_vs_i8m1_i8m1((vint8m1_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vredor_vs_i8m1_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredor_vs_i8m1_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vredor_vs_i8m2_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredor_vs_i8m2_i8m1((vint8m1_t)(op0), (vint8m2_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vredor_vs_i8m2_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredor_vs_i8m2_i8m1_m((vbool4_t)(op0), (vint8m1_t)(op1), (vint8m2_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vredor_vs_i8m4_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredor_vs_i8m4_i8m1((vint8m1_t)(op0), (vint8m4_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vredor_vs_i8m4_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredor_vs_i8m4_i8m1_m((vbool2_t)(op0), (vint8m1_t)(op1), (vint8m4_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vredor_vs_i8m8_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredor_vs_i8m8_i8m1((vint8m1_t)(op0), (vint8m8_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vredor_vs_i8m8_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredor_vs_i8m8_i8m1_m((vbool1_t)(op0), (vint8m1_t)(op1), (vint8m8_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vredor_vs_i8mf2_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredor_vs_i8mf2_i8m1((vint8m1_t)(op0), (vint8mf2_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vredor_vs_i8mf2_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredor_vs_i8mf2_i8m1_m((vbool16_t)(op0), (vint8m1_t)(op1), (vint8mf2_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vredor_vs_i8mf4_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredor_vs_i8mf4_i8m1((vint8m1_t)(op0), (vint8mf4_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vredor_vs_i8mf4_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredor_vs_i8mf4_i8m1_m((vbool32_t)(op0), (vint8m1_t)(op1), (vint8mf4_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vredor_vs_i8mf8_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredor_vs_i8mf8_i8m1((vint8m1_t)(op0), (vint8mf8_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vredor_vs_i8mf8_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredor_vs_i8mf8_i8m1_m((vbool64_t)(op0), (vint8m1_t)(op1), (vint8mf8_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vredor_vs_i16m1_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredor_vs_i16m1_i16m1((vint16m1_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vredor_vs_i16m1_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredor_vs_i16m1_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vredor_vs_i16m2_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredor_vs_i16m2_i16m1((vint16m1_t)(op0), (vint16m2_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vredor_vs_i16m2_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredor_vs_i16m2_i16m1_m((vbool8_t)(op0), (vint16m1_t)(op1), (vint16m2_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vredor_vs_i16m4_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredor_vs_i16m4_i16m1((vint16m1_t)(op0), (vint16m4_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vredor_vs_i16m4_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredor_vs_i16m4_i16m1_m((vbool4_t)(op0), (vint16m1_t)(op1), (vint16m4_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vredor_vs_i16m8_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredor_vs_i16m8_i16m1((vint16m1_t)(op0), (vint16m8_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vredor_vs_i16m8_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredor_vs_i16m8_i16m1_m((vbool2_t)(op0), (vint16m1_t)(op1), (vint16m8_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vredor_vs_i16mf2_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredor_vs_i16mf2_i16m1((vint16m1_t)(op0), (vint16mf2_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vredor_vs_i16mf2_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredor_vs_i16mf2_i16m1_m((vbool32_t)(op0), (vint16m1_t)(op1), (vint16mf2_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vredor_vs_i16mf4_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredor_vs_i16mf4_i16m1((vint16m1_t)(op0), (vint16mf4_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vredor_vs_i16mf4_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredor_vs_i16mf4_i16m1_m((vbool64_t)(op0), (vint16m1_t)(op1), (vint16mf4_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vredor_vs_i32m1_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredor_vs_i32m1_i32m1((vint32m1_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vredor_vs_i32m1_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredor_vs_i32m1_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vredor_vs_i32m2_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredor_vs_i32m2_i32m1((vint32m1_t)(op0), (vint32m2_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vredor_vs_i32m2_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredor_vs_i32m2_i32m1_m((vbool16_t)(op0), (vint32m1_t)(op1), (vint32m2_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vredor_vs_i32m4_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredor_vs_i32m4_i32m1((vint32m1_t)(op0), (vint32m4_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vredor_vs_i32m4_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredor_vs_i32m4_i32m1_m((vbool8_t)(op0), (vint32m1_t)(op1), (vint32m4_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vredor_vs_i32m8_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredor_vs_i32m8_i32m1((vint32m1_t)(op0), (vint32m8_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vredor_vs_i32m8_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredor_vs_i32m8_i32m1_m((vbool4_t)(op0), (vint32m1_t)(op1), (vint32m8_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vredor_vs_i32mf2_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredor_vs_i32mf2_i32m1((vint32m1_t)(op0), (vint32mf2_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vredor_vs_i32mf2_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredor_vs_i32mf2_i32m1_m((vbool64_t)(op0), (vint32m1_t)(op1), (vint32mf2_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vredor_vs_i64m1_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredor_vs_i64m1_i64m1((vint64m1_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vredor_vs_i64m1_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredor_vs_i64m1_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vredor_vs_i64m2_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredor_vs_i64m2_i64m1((vint64m1_t)(op0), (vint64m2_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vredor_vs_i64m2_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredor_vs_i64m2_i64m1_m((vbool32_t)(op0), (vint64m1_t)(op1), (vint64m2_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vredor_vs_i64m4_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredor_vs_i64m4_i64m1((vint64m1_t)(op0), (vint64m4_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vredor_vs_i64m4_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredor_vs_i64m4_i64m1_m((vbool16_t)(op0), (vint64m1_t)(op1), (vint64m4_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vredor_vs_i64m8_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredor_vs_i64m8_i64m1((vint64m1_t)(op0), (vint64m8_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vredor_vs_i64m8_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredor_vs_i64m8_i64m1_m((vbool8_t)(op0), (vint64m1_t)(op1), (vint64m8_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vredor_vs_u8m1_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredor_vs_u8m1_u8m1((vuint8m1_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vredor_vs_u8m1_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredor_vs_u8m1_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vredor_vs_u8m2_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredor_vs_u8m2_u8m1((vuint8m1_t)(op0), (vuint8m2_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vredor_vs_u8m2_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredor_vs_u8m2_u8m1_m((vbool4_t)(op0), (vuint8m1_t)(op1), (vuint8m2_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vredor_vs_u8m4_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredor_vs_u8m4_u8m1((vuint8m1_t)(op0), (vuint8m4_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vredor_vs_u8m4_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredor_vs_u8m4_u8m1_m((vbool2_t)(op0), (vuint8m1_t)(op1), (vuint8m4_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vredor_vs_u8m8_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredor_vs_u8m8_u8m1((vuint8m1_t)(op0), (vuint8m8_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vredor_vs_u8m8_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredor_vs_u8m8_u8m1_m((vbool1_t)(op0), (vuint8m1_t)(op1), (vuint8m8_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vredor_vs_u8mf2_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredor_vs_u8mf2_u8m1((vuint8m1_t)(op0), (vuint8mf2_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vredor_vs_u8mf2_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredor_vs_u8mf2_u8m1_m((vbool16_t)(op0), (vuint8m1_t)(op1), (vuint8mf2_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vredor_vs_u8mf4_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredor_vs_u8mf4_u8m1((vuint8m1_t)(op0), (vuint8mf4_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vredor_vs_u8mf4_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredor_vs_u8mf4_u8m1_m((vbool32_t)(op0), (vuint8m1_t)(op1), (vuint8mf4_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vredor_vs_u8mf8_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredor_vs_u8mf8_u8m1((vuint8m1_t)(op0), (vuint8mf8_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vredor_vs_u8mf8_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredor_vs_u8mf8_u8m1_m((vbool64_t)(op0), (vuint8m1_t)(op1), (vuint8mf8_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vredor_vs_u16m1_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredor_vs_u16m1_u16m1((vuint16m1_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vredor_vs_u16m1_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredor_vs_u16m1_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vredor_vs_u16m2_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredor_vs_u16m2_u16m1((vuint16m1_t)(op0), (vuint16m2_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vredor_vs_u16m2_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredor_vs_u16m2_u16m1_m((vbool8_t)(op0), (vuint16m1_t)(op1), (vuint16m2_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vredor_vs_u16m4_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredor_vs_u16m4_u16m1((vuint16m1_t)(op0), (vuint16m4_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vredor_vs_u16m4_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredor_vs_u16m4_u16m1_m((vbool4_t)(op0), (vuint16m1_t)(op1), (vuint16m4_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vredor_vs_u16m8_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredor_vs_u16m8_u16m1((vuint16m1_t)(op0), (vuint16m8_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vredor_vs_u16m8_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredor_vs_u16m8_u16m1_m((vbool2_t)(op0), (vuint16m1_t)(op1), (vuint16m8_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vredor_vs_u16mf2_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredor_vs_u16mf2_u16m1((vuint16m1_t)(op0), (vuint16mf2_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vredor_vs_u16mf2_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredor_vs_u16mf2_u16m1_m((vbool32_t)(op0), (vuint16m1_t)(op1), (vuint16mf2_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vredor_vs_u16mf4_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredor_vs_u16mf4_u16m1((vuint16m1_t)(op0), (vuint16mf4_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vredor_vs_u16mf4_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredor_vs_u16mf4_u16m1_m((vbool64_t)(op0), (vuint16m1_t)(op1), (vuint16mf4_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vredor_vs_u32m1_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredor_vs_u32m1_u32m1((vuint32m1_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vredor_vs_u32m1_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredor_vs_u32m1_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vredor_vs_u32m2_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredor_vs_u32m2_u32m1((vuint32m1_t)(op0), (vuint32m2_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vredor_vs_u32m2_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredor_vs_u32m2_u32m1_m((vbool16_t)(op0), (vuint32m1_t)(op1), (vuint32m2_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vredor_vs_u32m4_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredor_vs_u32m4_u32m1((vuint32m1_t)(op0), (vuint32m4_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vredor_vs_u32m4_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredor_vs_u32m4_u32m1_m((vbool8_t)(op0), (vuint32m1_t)(op1), (vuint32m4_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vredor_vs_u32m8_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredor_vs_u32m8_u32m1((vuint32m1_t)(op0), (vuint32m8_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vredor_vs_u32m8_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredor_vs_u32m8_u32m1_m((vbool4_t)(op0), (vuint32m1_t)(op1), (vuint32m8_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vredor_vs_u32mf2_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredor_vs_u32mf2_u32m1((vuint32m1_t)(op0), (vuint32mf2_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vredor_vs_u32mf2_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredor_vs_u32mf2_u32m1_m((vbool64_t)(op0), (vuint32m1_t)(op1), (vuint32mf2_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vredor_vs_u64m1_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredor_vs_u64m1_u64m1((vuint64m1_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vredor_vs_u64m1_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredor_vs_u64m1_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vredor_vs_u64m2_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredor_vs_u64m2_u64m1((vuint64m1_t)(op0), (vuint64m2_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vredor_vs_u64m2_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredor_vs_u64m2_u64m1_m((vbool32_t)(op0), (vuint64m1_t)(op1), (vuint64m2_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vredor_vs_u64m4_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredor_vs_u64m4_u64m1((vuint64m1_t)(op0), (vuint64m4_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vredor_vs_u64m4_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredor_vs_u64m4_u64m1_m((vbool16_t)(op0), (vuint64m1_t)(op1), (vuint64m4_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vredor_vs_u64m8_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredor_vs_u64m8_u64m1((vuint64m1_t)(op0), (vuint64m8_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vredor_vs_u64m8_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredor_vs_u64m8_u64m1_m((vbool8_t)(op0), (vuint64m1_t)(op1), (vuint64m8_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vredxor_vs_i8m1_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredxor_vs_i8m1_i8m1((vint8m1_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vredxor_vs_i8m1_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredxor_vs_i8m1_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vredxor_vs_i8m2_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredxor_vs_i8m2_i8m1((vint8m1_t)(op0), (vint8m2_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vredxor_vs_i8m2_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredxor_vs_i8m2_i8m1_m((vbool4_t)(op0), (vint8m1_t)(op1), (vint8m2_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vredxor_vs_i8m4_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredxor_vs_i8m4_i8m1((vint8m1_t)(op0), (vint8m4_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vredxor_vs_i8m4_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredxor_vs_i8m4_i8m1_m((vbool2_t)(op0), (vint8m1_t)(op1), (vint8m4_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vredxor_vs_i8m8_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredxor_vs_i8m8_i8m1((vint8m1_t)(op0), (vint8m8_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vredxor_vs_i8m8_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredxor_vs_i8m8_i8m1_m((vbool1_t)(op0), (vint8m1_t)(op1), (vint8m8_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vredxor_vs_i8mf2_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredxor_vs_i8mf2_i8m1((vint8m1_t)(op0), (vint8mf2_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vredxor_vs_i8mf2_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredxor_vs_i8mf2_i8m1_m((vbool16_t)(op0), (vint8m1_t)(op1), (vint8mf2_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vredxor_vs_i8mf4_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredxor_vs_i8mf4_i8m1((vint8m1_t)(op0), (vint8mf4_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vredxor_vs_i8mf4_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredxor_vs_i8mf4_i8m1_m((vbool32_t)(op0), (vint8m1_t)(op1), (vint8mf4_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vredxor_vs_i8mf8_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredxor_vs_i8mf8_i8m1((vint8m1_t)(op0), (vint8mf8_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vredxor_vs_i8mf8_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredxor_vs_i8mf8_i8m1_m((vbool64_t)(op0), (vint8m1_t)(op1), (vint8mf8_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vredxor_vs_i16m1_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredxor_vs_i16m1_i16m1((vint16m1_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vredxor_vs_i16m1_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredxor_vs_i16m1_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vredxor_vs_i16m2_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredxor_vs_i16m2_i16m1((vint16m1_t)(op0), (vint16m2_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vredxor_vs_i16m2_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredxor_vs_i16m2_i16m1_m((vbool8_t)(op0), (vint16m1_t)(op1), (vint16m2_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vredxor_vs_i16m4_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredxor_vs_i16m4_i16m1((vint16m1_t)(op0), (vint16m4_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vredxor_vs_i16m4_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredxor_vs_i16m4_i16m1_m((vbool4_t)(op0), (vint16m1_t)(op1), (vint16m4_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vredxor_vs_i16m8_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredxor_vs_i16m8_i16m1((vint16m1_t)(op0), (vint16m8_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vredxor_vs_i16m8_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredxor_vs_i16m8_i16m1_m((vbool2_t)(op0), (vint16m1_t)(op1), (vint16m8_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vredxor_vs_i16mf2_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredxor_vs_i16mf2_i16m1((vint16m1_t)(op0), (vint16mf2_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vredxor_vs_i16mf2_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredxor_vs_i16mf2_i16m1_m((vbool32_t)(op0), (vint16m1_t)(op1), (vint16mf2_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vredxor_vs_i16mf4_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredxor_vs_i16mf4_i16m1((vint16m1_t)(op0), (vint16mf4_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vredxor_vs_i16mf4_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredxor_vs_i16mf4_i16m1_m((vbool64_t)(op0), (vint16m1_t)(op1), (vint16mf4_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vredxor_vs_i32m1_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredxor_vs_i32m1_i32m1((vint32m1_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vredxor_vs_i32m1_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredxor_vs_i32m1_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vredxor_vs_i32m2_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredxor_vs_i32m2_i32m1((vint32m1_t)(op0), (vint32m2_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vredxor_vs_i32m2_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredxor_vs_i32m2_i32m1_m((vbool16_t)(op0), (vint32m1_t)(op1), (vint32m2_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vredxor_vs_i32m4_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredxor_vs_i32m4_i32m1((vint32m1_t)(op0), (vint32m4_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vredxor_vs_i32m4_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredxor_vs_i32m4_i32m1_m((vbool8_t)(op0), (vint32m1_t)(op1), (vint32m4_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vredxor_vs_i32m8_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredxor_vs_i32m8_i32m1((vint32m1_t)(op0), (vint32m8_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vredxor_vs_i32m8_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredxor_vs_i32m8_i32m1_m((vbool4_t)(op0), (vint32m1_t)(op1), (vint32m8_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vredxor_vs_i32mf2_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredxor_vs_i32mf2_i32m1((vint32m1_t)(op0), (vint32mf2_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vredxor_vs_i32mf2_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredxor_vs_i32mf2_i32m1_m((vbool64_t)(op0), (vint32m1_t)(op1), (vint32mf2_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vredxor_vs_i64m1_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredxor_vs_i64m1_i64m1((vint64m1_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vredxor_vs_i64m1_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredxor_vs_i64m1_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vredxor_vs_i64m2_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredxor_vs_i64m2_i64m1((vint64m1_t)(op0), (vint64m2_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vredxor_vs_i64m2_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredxor_vs_i64m2_i64m1_m((vbool32_t)(op0), (vint64m1_t)(op1), (vint64m2_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vredxor_vs_i64m4_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredxor_vs_i64m4_i64m1((vint64m1_t)(op0), (vint64m4_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vredxor_vs_i64m4_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredxor_vs_i64m4_i64m1_m((vbool16_t)(op0), (vint64m1_t)(op1), (vint64m4_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vredxor_vs_i64m8_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredxor_vs_i64m8_i64m1((vint64m1_t)(op0), (vint64m8_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vredxor_vs_i64m8_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredxor_vs_i64m8_i64m1_m((vbool8_t)(op0), (vint64m1_t)(op1), (vint64m8_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vredxor_vs_u8m1_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredxor_vs_u8m1_u8m1((vuint8m1_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vredxor_vs_u8m1_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredxor_vs_u8m1_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vredxor_vs_u8m2_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredxor_vs_u8m2_u8m1((vuint8m1_t)(op0), (vuint8m2_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vredxor_vs_u8m2_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredxor_vs_u8m2_u8m1_m((vbool4_t)(op0), (vuint8m1_t)(op1), (vuint8m2_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vredxor_vs_u8m4_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredxor_vs_u8m4_u8m1((vuint8m1_t)(op0), (vuint8m4_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vredxor_vs_u8m4_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredxor_vs_u8m4_u8m1_m((vbool2_t)(op0), (vuint8m1_t)(op1), (vuint8m4_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vredxor_vs_u8m8_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredxor_vs_u8m8_u8m1((vuint8m1_t)(op0), (vuint8m8_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vredxor_vs_u8m8_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredxor_vs_u8m8_u8m1_m((vbool1_t)(op0), (vuint8m1_t)(op1), (vuint8m8_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vredxor_vs_u8mf2_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredxor_vs_u8mf2_u8m1((vuint8m1_t)(op0), (vuint8mf2_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vredxor_vs_u8mf2_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredxor_vs_u8mf2_u8m1_m((vbool16_t)(op0), (vuint8m1_t)(op1), (vuint8mf2_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vredxor_vs_u8mf4_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredxor_vs_u8mf4_u8m1((vuint8m1_t)(op0), (vuint8mf4_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vredxor_vs_u8mf4_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredxor_vs_u8mf4_u8m1_m((vbool32_t)(op0), (vuint8m1_t)(op1), (vuint8mf4_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vredxor_vs_u8mf8_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vredxor_vs_u8mf8_u8m1((vuint8m1_t)(op0), (vuint8mf8_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vredxor_vs_u8mf8_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredxor_vs_u8mf8_u8m1_m((vbool64_t)(op0), (vuint8m1_t)(op1), (vuint8mf8_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vredxor_vs_u16m1_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredxor_vs_u16m1_u16m1((vuint16m1_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vredxor_vs_u16m1_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredxor_vs_u16m1_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vredxor_vs_u16m2_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredxor_vs_u16m2_u16m1((vuint16m1_t)(op0), (vuint16m2_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vredxor_vs_u16m2_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredxor_vs_u16m2_u16m1_m((vbool8_t)(op0), (vuint16m1_t)(op1), (vuint16m2_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vredxor_vs_u16m4_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredxor_vs_u16m4_u16m1((vuint16m1_t)(op0), (vuint16m4_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vredxor_vs_u16m4_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredxor_vs_u16m4_u16m1_m((vbool4_t)(op0), (vuint16m1_t)(op1), (vuint16m4_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vredxor_vs_u16m8_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredxor_vs_u16m8_u16m1((vuint16m1_t)(op0), (vuint16m8_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vredxor_vs_u16m8_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredxor_vs_u16m8_u16m1_m((vbool2_t)(op0), (vuint16m1_t)(op1), (vuint16m8_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vredxor_vs_u16mf2_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredxor_vs_u16mf2_u16m1((vuint16m1_t)(op0), (vuint16mf2_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vredxor_vs_u16mf2_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredxor_vs_u16mf2_u16m1_m((vbool32_t)(op0), (vuint16m1_t)(op1), (vuint16mf2_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vredxor_vs_u16mf4_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vredxor_vs_u16mf4_u16m1((vuint16m1_t)(op0), (vuint16mf4_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vredxor_vs_u16mf4_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredxor_vs_u16mf4_u16m1_m((vbool64_t)(op0), (vuint16m1_t)(op1), (vuint16mf4_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vredxor_vs_u32m1_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredxor_vs_u32m1_u32m1((vuint32m1_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vredxor_vs_u32m1_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredxor_vs_u32m1_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vredxor_vs_u32m2_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredxor_vs_u32m2_u32m1((vuint32m1_t)(op0), (vuint32m2_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vredxor_vs_u32m2_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredxor_vs_u32m2_u32m1_m((vbool16_t)(op0), (vuint32m1_t)(op1), (vuint32m2_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vredxor_vs_u32m4_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredxor_vs_u32m4_u32m1((vuint32m1_t)(op0), (vuint32m4_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vredxor_vs_u32m4_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredxor_vs_u32m4_u32m1_m((vbool8_t)(op0), (vuint32m1_t)(op1), (vuint32m4_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vredxor_vs_u32m8_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredxor_vs_u32m8_u32m1((vuint32m1_t)(op0), (vuint32m8_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vredxor_vs_u32m8_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredxor_vs_u32m8_u32m1_m((vbool4_t)(op0), (vuint32m1_t)(op1), (vuint32m8_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vredxor_vs_u32mf2_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vredxor_vs_u32mf2_u32m1((vuint32m1_t)(op0), (vuint32mf2_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vredxor_vs_u32mf2_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredxor_vs_u32mf2_u32m1_m((vbool64_t)(op0), (vuint32m1_t)(op1), (vuint32mf2_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vredxor_vs_u64m1_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredxor_vs_u64m1_u64m1((vuint64m1_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vredxor_vs_u64m1_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredxor_vs_u64m1_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vredxor_vs_u64m2_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredxor_vs_u64m2_u64m1((vuint64m1_t)(op0), (vuint64m2_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vredxor_vs_u64m2_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredxor_vs_u64m2_u64m1_m((vbool32_t)(op0), (vuint64m1_t)(op1), (vuint64m2_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vredxor_vs_u64m4_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredxor_vs_u64m4_u64m1((vuint64m1_t)(op0), (vuint64m4_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vredxor_vs_u64m4_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredxor_vs_u64m4_u64m1_m((vbool16_t)(op0), (vuint64m1_t)(op1), (vuint64m4_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vredxor_vs_u64m8_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vredxor_vs_u64m8_u64m1((vuint64m1_t)(op0), (vuint64m8_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vredxor_vs_u64m8_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vredxor_vs_u64m8_u64m1_m((vbool8_t)(op0), (vuint64m1_t)(op1), (vuint64m8_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vwredsum_vs_i8m1_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vwredsum_vs_i8m1_i16m1((vint16m1_t)(op0), (vint8m1_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vwredsum_vs_i8m1_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwredsum_vs_i8m1_i16m1_m((vbool8_t)(op0), (vint16m1_t)(op1), (vint8m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vwredsum_vs_i8m2_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vwredsum_vs_i8m2_i16m1((vint16m1_t)(op0), (vint8m2_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vwredsum_vs_i8m2_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwredsum_vs_i8m2_i16m1_m((vbool4_t)(op0), (vint16m1_t)(op1), (vint8m2_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vwredsum_vs_i8m4_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vwredsum_vs_i8m4_i16m1((vint16m1_t)(op0), (vint8m4_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vwredsum_vs_i8m4_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwredsum_vs_i8m4_i16m1_m((vbool2_t)(op0), (vint16m1_t)(op1), (vint8m4_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vwredsum_vs_i8m8_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vwredsum_vs_i8m8_i16m1((vint16m1_t)(op0), (vint8m8_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vwredsum_vs_i8m8_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwredsum_vs_i8m8_i16m1_m((vbool1_t)(op0), (vint16m1_t)(op1), (vint8m8_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vwredsum_vs_i8mf2_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vwredsum_vs_i8mf2_i16m1((vint16m1_t)(op0), (vint8mf2_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vwredsum_vs_i8mf2_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwredsum_vs_i8mf2_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint8mf2_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vwredsum_vs_i8mf4_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vwredsum_vs_i8mf4_i16m1((vint16m1_t)(op0), (vint8mf4_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vwredsum_vs_i8mf4_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwredsum_vs_i8mf4_i16m1_m((vbool32_t)(op0), (vint16m1_t)(op1), (vint8mf4_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vwredsum_vs_i8mf8_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vwredsum_vs_i8mf8_i16m1((vint16m1_t)(op0), (vint8mf8_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vwredsum_vs_i8mf8_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwredsum_vs_i8mf8_i16m1_m((vbool64_t)(op0), (vint16m1_t)(op1), (vint8mf8_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vwredsum_vs_i16m1_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vwredsum_vs_i16m1_i32m1((vint32m1_t)(op0), (vint16m1_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vwredsum_vs_i16m1_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwredsum_vs_i16m1_i32m1_m((vbool16_t)(op0), (vint32m1_t)(op1), (vint16m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vwredsum_vs_i16m2_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vwredsum_vs_i16m2_i32m1((vint32m1_t)(op0), (vint16m2_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vwredsum_vs_i16m2_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwredsum_vs_i16m2_i32m1_m((vbool8_t)(op0), (vint32m1_t)(op1), (vint16m2_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vwredsum_vs_i16m4_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vwredsum_vs_i16m4_i32m1((vint32m1_t)(op0), (vint16m4_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vwredsum_vs_i16m4_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwredsum_vs_i16m4_i32m1_m((vbool4_t)(op0), (vint32m1_t)(op1), (vint16m4_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vwredsum_vs_i16m8_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vwredsum_vs_i16m8_i32m1((vint32m1_t)(op0), (vint16m8_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vwredsum_vs_i16m8_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwredsum_vs_i16m8_i32m1_m((vbool2_t)(op0), (vint32m1_t)(op1), (vint16m8_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vwredsum_vs_i16mf2_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vwredsum_vs_i16mf2_i32m1((vint32m1_t)(op0), (vint16mf2_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vwredsum_vs_i16mf2_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwredsum_vs_i16mf2_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint16mf2_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vwredsum_vs_i16mf4_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vwredsum_vs_i16mf4_i32m1((vint32m1_t)(op0), (vint16mf4_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vwredsum_vs_i16mf4_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwredsum_vs_i16mf4_i32m1_m((vbool64_t)(op0), (vint32m1_t)(op1), (vint16mf4_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vwredsum_vs_i32m1_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vwredsum_vs_i32m1_i64m1((vint64m1_t)(op0), (vint32m1_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vwredsum_vs_i32m1_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwredsum_vs_i32m1_i64m1_m((vbool32_t)(op0), (vint64m1_t)(op1), (vint32m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vwredsum_vs_i32m2_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vwredsum_vs_i32m2_i64m1((vint64m1_t)(op0), (vint32m2_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vwredsum_vs_i32m2_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwredsum_vs_i32m2_i64m1_m((vbool16_t)(op0), (vint64m1_t)(op1), (vint32m2_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vwredsum_vs_i32m4_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vwredsum_vs_i32m4_i64m1((vint64m1_t)(op0), (vint32m4_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vwredsum_vs_i32m4_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwredsum_vs_i32m4_i64m1_m((vbool8_t)(op0), (vint64m1_t)(op1), (vint32m4_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vwredsum_vs_i32m8_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vwredsum_vs_i32m8_i64m1((vint64m1_t)(op0), (vint32m8_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vwredsum_vs_i32m8_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwredsum_vs_i32m8_i64m1_m((vbool4_t)(op0), (vint64m1_t)(op1), (vint32m8_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vwredsum_vs_i32mf2_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vwredsum_vs_i32mf2_i64m1((vint64m1_t)(op0), (vint32mf2_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vwredsum_vs_i32mf2_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwredsum_vs_i32mf2_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint32mf2_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vwredsumu_vs_u8m1_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vwredsumu_vs_u8m1_u16m1((vuint16m1_t)(op0), (vuint8m1_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vwredsumu_vs_u8m1_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwredsumu_vs_u8m1_u16m1_m((vbool8_t)(op0), (vuint16m1_t)(op1), (vuint8m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vwredsumu_vs_u8m2_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vwredsumu_vs_u8m2_u16m1((vuint16m1_t)(op0), (vuint8m2_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vwredsumu_vs_u8m2_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwredsumu_vs_u8m2_u16m1_m((vbool4_t)(op0), (vuint16m1_t)(op1), (vuint8m2_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vwredsumu_vs_u8m4_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vwredsumu_vs_u8m4_u16m1((vuint16m1_t)(op0), (vuint8m4_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vwredsumu_vs_u8m4_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwredsumu_vs_u8m4_u16m1_m((vbool2_t)(op0), (vuint16m1_t)(op1), (vuint8m4_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vwredsumu_vs_u8m8_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vwredsumu_vs_u8m8_u16m1((vuint16m1_t)(op0), (vuint8m8_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vwredsumu_vs_u8m8_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwredsumu_vs_u8m8_u16m1_m((vbool1_t)(op0), (vuint16m1_t)(op1), (vuint8m8_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vwredsumu_vs_u8mf2_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vwredsumu_vs_u8mf2_u16m1((vuint16m1_t)(op0), (vuint8mf2_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vwredsumu_vs_u8mf2_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwredsumu_vs_u8mf2_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint8mf2_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vwredsumu_vs_u8mf4_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vwredsumu_vs_u8mf4_u16m1((vuint16m1_t)(op0), (vuint8mf4_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vwredsumu_vs_u8mf4_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwredsumu_vs_u8mf4_u16m1_m((vbool32_t)(op0), (vuint16m1_t)(op1), (vuint8mf4_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vwredsumu_vs_u8mf8_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vwredsumu_vs_u8mf8_u16m1((vuint16m1_t)(op0), (vuint8mf8_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vwredsumu_vs_u8mf8_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwredsumu_vs_u8mf8_u16m1_m((vbool64_t)(op0), (vuint16m1_t)(op1), (vuint8mf8_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vwredsumu_vs_u16m1_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vwredsumu_vs_u16m1_u32m1((vuint32m1_t)(op0), (vuint16m1_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vwredsumu_vs_u16m1_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwredsumu_vs_u16m1_u32m1_m((vbool16_t)(op0), (vuint32m1_t)(op1), (vuint16m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vwredsumu_vs_u16m2_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vwredsumu_vs_u16m2_u32m1((vuint32m1_t)(op0), (vuint16m2_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vwredsumu_vs_u16m2_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwredsumu_vs_u16m2_u32m1_m((vbool8_t)(op0), (vuint32m1_t)(op1), (vuint16m2_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vwredsumu_vs_u16m4_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vwredsumu_vs_u16m4_u32m1((vuint32m1_t)(op0), (vuint16m4_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vwredsumu_vs_u16m4_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwredsumu_vs_u16m4_u32m1_m((vbool4_t)(op0), (vuint32m1_t)(op1), (vuint16m4_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vwredsumu_vs_u16m8_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vwredsumu_vs_u16m8_u32m1((vuint32m1_t)(op0), (vuint16m8_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vwredsumu_vs_u16m8_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwredsumu_vs_u16m8_u32m1_m((vbool2_t)(op0), (vuint32m1_t)(op1), (vuint16m8_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vwredsumu_vs_u16mf2_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vwredsumu_vs_u16mf2_u32m1((vuint32m1_t)(op0), (vuint16mf2_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vwredsumu_vs_u16mf2_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwredsumu_vs_u16mf2_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint16mf2_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vwredsumu_vs_u16mf4_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vwredsumu_vs_u16mf4_u32m1((vuint32m1_t)(op0), (vuint16mf4_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vwredsumu_vs_u16mf4_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwredsumu_vs_u16mf4_u32m1_m((vbool64_t)(op0), (vuint32m1_t)(op1), (vuint16mf4_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vwredsumu_vs_u32m1_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vwredsumu_vs_u32m1_u64m1((vuint64m1_t)(op0), (vuint32m1_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vwredsumu_vs_u32m1_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwredsumu_vs_u32m1_u64m1_m((vbool32_t)(op0), (vuint64m1_t)(op1), (vuint32m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vwredsumu_vs_u32m2_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vwredsumu_vs_u32m2_u64m1((vuint64m1_t)(op0), (vuint32m2_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vwredsumu_vs_u32m2_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwredsumu_vs_u32m2_u64m1_m((vbool16_t)(op0), (vuint64m1_t)(op1), (vuint32m2_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vwredsumu_vs_u32m4_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vwredsumu_vs_u32m4_u64m1((vuint64m1_t)(op0), (vuint32m4_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vwredsumu_vs_u32m4_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwredsumu_vs_u32m4_u64m1_m((vbool8_t)(op0), (vuint64m1_t)(op1), (vuint32m4_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vwredsumu_vs_u32m8_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vwredsumu_vs_u32m8_u64m1((vuint64m1_t)(op0), (vuint32m8_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vwredsumu_vs_u32m8_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwredsumu_vs_u32m8_u64m1_m((vbool4_t)(op0), (vuint64m1_t)(op1), (vuint32m8_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vwredsumu_vs_u32mf2_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vwredsumu_vs_u32mf2_u64m1((vuint64m1_t)(op0), (vuint32mf2_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vwredsumu_vs_u32mf2_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vwredsumu_vs_u32mf2_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint32mf2_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vsuxei8_v_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_u8m1((uint8_t *)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vsuxei8_v_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vsuxei8_v_u8m2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_u8m2((uint8_t *)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (size_t)(op3)) +#define vsuxei8_v_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_u8m2_m((vbool4_t)(op0), (uint8_t *)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vsuxei8_v_u8m4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_u8m4((uint8_t *)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (size_t)(op3)) +#define vsuxei8_v_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_u8m4_m((vbool2_t)(op0), (uint8_t *)(op1), (vuint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vsuxei8_v_u8m8(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_u8m8((uint8_t *)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (size_t)(op3)) +#define vsuxei8_v_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_u8m8_m((vbool1_t)(op0), (uint8_t *)(op1), (vuint8m8_t)(op2), (vuint8m8_t)(op3), (size_t)(op4)) +#define vsuxei8_v_u8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_u8mf2((uint8_t *)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (size_t)(op3)) +#define vsuxei8_v_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vsuxei8_v_u8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_u8mf4((uint8_t *)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (size_t)(op3)) +#define vsuxei8_v_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vsuxei8_v_u8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_u8mf8((uint8_t *)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (size_t)(op3)) +#define vsuxei8_v_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vmnot_m_b8(op0, op1) \ +__builtin_rvv_vmnot_m_b8((vbool8_t)(op0), (size_t)(op1)) +#define vmnot_m_b4(op0, op1) \ +__builtin_rvv_vmnot_m_b4((vbool4_t)(op0), (size_t)(op1)) +#define vmnot_m_b2(op0, op1) \ +__builtin_rvv_vmnot_m_b2((vbool2_t)(op0), (size_t)(op1)) +#define vmnot_m_b1(op0, op1) \ +__builtin_rvv_vmnot_m_b1((vbool1_t)(op0), (size_t)(op1)) +#define vmnot_m_b16(op0, op1) \ +__builtin_rvv_vmnot_m_b16((vbool16_t)(op0), (size_t)(op1)) +#define vmnot_m_b32(op0, op1) \ +__builtin_rvv_vmnot_m_b32((vbool32_t)(op0), (size_t)(op1)) +#define vmnot_m_b64(op0, op1) \ +__builtin_rvv_vmnot_m_b64((vbool64_t)(op0), (size_t)(op1)) +#define viota_m_u8m1(op0, op1) \ +__builtin_rvv_viota_m_u8m1((vbool8_t)(op0), (size_t)(op1)) +#define viota_m_u8m1_m(op0, op1, op2, op3) \ +__builtin_rvv_viota_m_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define viota_m_u8m2(op0, op1) \ +__builtin_rvv_viota_m_u8m2((vbool4_t)(op0), (size_t)(op1)) +#define viota_m_u8m2_m(op0, op1, op2, op3) \ +__builtin_rvv_viota_m_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define viota_m_u8m4(op0, op1) \ +__builtin_rvv_viota_m_u8m4((vbool2_t)(op0), (size_t)(op1)) +#define viota_m_u8m4_m(op0, op1, op2, op3) \ +__builtin_rvv_viota_m_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vbool2_t)(op2), (size_t)(op3)) +#define viota_m_u8m8(op0, op1) \ +__builtin_rvv_viota_m_u8m8((vbool1_t)(op0), (size_t)(op1)) +#define viota_m_u8m8_m(op0, op1, op2, op3) \ +__builtin_rvv_viota_m_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vbool1_t)(op2), (size_t)(op3)) +#define viota_m_u8mf2(op0, op1) \ +__builtin_rvv_viota_m_u8mf2((vbool16_t)(op0), (size_t)(op1)) +#define viota_m_u8mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_viota_m_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define viota_m_u8mf4(op0, op1) \ +__builtin_rvv_viota_m_u8mf4((vbool32_t)(op0), (size_t)(op1)) +#define viota_m_u8mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_viota_m_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define viota_m_u8mf8(op0, op1) \ +__builtin_rvv_viota_m_u8mf8((vbool64_t)(op0), (size_t)(op1)) +#define viota_m_u8mf8_m(op0, op1, op2, op3) \ +__builtin_rvv_viota_m_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define viota_m_u16m1(op0, op1) \ +__builtin_rvv_viota_m_u16m1((vbool16_t)(op0), (size_t)(op1)) +#define viota_m_u16m1_m(op0, op1, op2, op3) \ +__builtin_rvv_viota_m_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define viota_m_u16m2(op0, op1) \ +__builtin_rvv_viota_m_u16m2((vbool8_t)(op0), (size_t)(op1)) +#define viota_m_u16m2_m(op0, op1, op2, op3) \ +__builtin_rvv_viota_m_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define viota_m_u16m4(op0, op1) \ +__builtin_rvv_viota_m_u16m4((vbool4_t)(op0), (size_t)(op1)) +#define viota_m_u16m4_m(op0, op1, op2, op3) \ +__builtin_rvv_viota_m_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define viota_m_u16m8(op0, op1) \ +__builtin_rvv_viota_m_u16m8((vbool2_t)(op0), (size_t)(op1)) +#define viota_m_u16m8_m(op0, op1, op2, op3) \ +__builtin_rvv_viota_m_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vbool2_t)(op2), (size_t)(op3)) +#define viota_m_u16mf2(op0, op1) \ +__builtin_rvv_viota_m_u16mf2((vbool32_t)(op0), (size_t)(op1)) +#define viota_m_u16mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_viota_m_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define viota_m_u16mf4(op0, op1) \ +__builtin_rvv_viota_m_u16mf4((vbool64_t)(op0), (size_t)(op1)) +#define viota_m_u16mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_viota_m_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define viota_m_u32m1(op0, op1) \ +__builtin_rvv_viota_m_u32m1((vbool32_t)(op0), (size_t)(op1)) +#define viota_m_u32m1_m(op0, op1, op2, op3) \ +__builtin_rvv_viota_m_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define viota_m_u32m2(op0, op1) \ +__builtin_rvv_viota_m_u32m2((vbool16_t)(op0), (size_t)(op1)) +#define viota_m_u32m2_m(op0, op1, op2, op3) \ +__builtin_rvv_viota_m_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define viota_m_u32m4(op0, op1) \ +__builtin_rvv_viota_m_u32m4((vbool8_t)(op0), (size_t)(op1)) +#define viota_m_u32m4_m(op0, op1, op2, op3) \ +__builtin_rvv_viota_m_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define viota_m_u32m8(op0, op1) \ +__builtin_rvv_viota_m_u32m8((vbool4_t)(op0), (size_t)(op1)) +#define viota_m_u32m8_m(op0, op1, op2, op3) \ +__builtin_rvv_viota_m_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define viota_m_u32mf2(op0, op1) \ +__builtin_rvv_viota_m_u32mf2((vbool64_t)(op0), (size_t)(op1)) +#define viota_m_u32mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_viota_m_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define viota_m_u64m1(op0, op1) \ +__builtin_rvv_viota_m_u64m1((vbool64_t)(op0), (size_t)(op1)) +#define viota_m_u64m1_m(op0, op1, op2, op3) \ +__builtin_rvv_viota_m_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define viota_m_u64m2(op0, op1) \ +__builtin_rvv_viota_m_u64m2((vbool32_t)(op0), (size_t)(op1)) +#define viota_m_u64m2_m(op0, op1, op2, op3) \ +__builtin_rvv_viota_m_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define viota_m_u64m4(op0, op1) \ +__builtin_rvv_viota_m_u64m4((vbool16_t)(op0), (size_t)(op1)) +#define viota_m_u64m4_m(op0, op1, op2, op3) \ +__builtin_rvv_viota_m_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define viota_m_u64m8(op0, op1) \ +__builtin_rvv_viota_m_u64m8((vbool8_t)(op0), (size_t)(op1)) +#define viota_m_u64m8_m(op0, op1, op2, op3) \ +__builtin_rvv_viota_m_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vid_v_i8m1(op0) \ +__builtin_rvv_vid_v_i8m1((size_t)(op0)) +#define vid_v_i8m1_m(op0, op1, op2) \ +__builtin_rvv_vid_v_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (size_t)(op2)) +#define vid_v_i8m2(op0) \ +__builtin_rvv_vid_v_i8m2((size_t)(op0)) +#define vid_v_i8m2_m(op0, op1, op2) \ +__builtin_rvv_vid_v_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (size_t)(op2)) +#define vid_v_i8m4(op0) \ +__builtin_rvv_vid_v_i8m4((size_t)(op0)) +#define vid_v_i8m4_m(op0, op1, op2) \ +__builtin_rvv_vid_v_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (size_t)(op2)) +#define vid_v_i8m8(op0) \ +__builtin_rvv_vid_v_i8m8((size_t)(op0)) +#define vid_v_i8m8_m(op0, op1, op2) \ +__builtin_rvv_vid_v_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (size_t)(op2)) +#define vid_v_i8mf2(op0) \ +__builtin_rvv_vid_v_i8mf2((size_t)(op0)) +#define vid_v_i8mf2_m(op0, op1, op2) \ +__builtin_rvv_vid_v_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (size_t)(op2)) +#define vid_v_i8mf4(op0) \ +__builtin_rvv_vid_v_i8mf4((size_t)(op0)) +#define vid_v_i8mf4_m(op0, op1, op2) \ +__builtin_rvv_vid_v_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (size_t)(op2)) +#define vid_v_i8mf8(op0) \ +__builtin_rvv_vid_v_i8mf8((size_t)(op0)) +#define vid_v_i8mf8_m(op0, op1, op2) \ +__builtin_rvv_vid_v_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (size_t)(op2)) +#define vid_v_i16m1(op0) \ +__builtin_rvv_vid_v_i16m1((size_t)(op0)) +#define vid_v_i16m1_m(op0, op1, op2) \ +__builtin_rvv_vid_v_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (size_t)(op2)) +#define vid_v_i16m2(op0) \ +__builtin_rvv_vid_v_i16m2((size_t)(op0)) +#define vid_v_i16m2_m(op0, op1, op2) \ +__builtin_rvv_vid_v_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (size_t)(op2)) +#define vid_v_i16m4(op0) \ +__builtin_rvv_vid_v_i16m4((size_t)(op0)) +#define vid_v_i16m4_m(op0, op1, op2) \ +__builtin_rvv_vid_v_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (size_t)(op2)) +#define vid_v_i16m8(op0) \ +__builtin_rvv_vid_v_i16m8((size_t)(op0)) +#define vid_v_i16m8_m(op0, op1, op2) \ +__builtin_rvv_vid_v_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (size_t)(op2)) +#define vid_v_i16mf2(op0) \ +__builtin_rvv_vid_v_i16mf2((size_t)(op0)) +#define vid_v_i16mf2_m(op0, op1, op2) \ +__builtin_rvv_vid_v_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (size_t)(op2)) +#define vid_v_i16mf4(op0) \ +__builtin_rvv_vid_v_i16mf4((size_t)(op0)) +#define vid_v_i16mf4_m(op0, op1, op2) \ +__builtin_rvv_vid_v_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (size_t)(op2)) +#define vid_v_i32m1(op0) \ +__builtin_rvv_vid_v_i32m1((size_t)(op0)) +#define vid_v_i32m1_m(op0, op1, op2) \ +__builtin_rvv_vid_v_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (size_t)(op2)) +#define vid_v_i32m2(op0) \ +__builtin_rvv_vid_v_i32m2((size_t)(op0)) +#define vid_v_i32m2_m(op0, op1, op2) \ +__builtin_rvv_vid_v_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (size_t)(op2)) +#define vid_v_i32m4(op0) \ +__builtin_rvv_vid_v_i32m4((size_t)(op0)) +#define vid_v_i32m4_m(op0, op1, op2) \ +__builtin_rvv_vid_v_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (size_t)(op2)) +#define vid_v_i32m8(op0) \ +__builtin_rvv_vid_v_i32m8((size_t)(op0)) +#define vid_v_i32m8_m(op0, op1, op2) \ +__builtin_rvv_vid_v_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (size_t)(op2)) +#define vid_v_i32mf2(op0) \ +__builtin_rvv_vid_v_i32mf2((size_t)(op0)) +#define vid_v_i32mf2_m(op0, op1, op2) \ +__builtin_rvv_vid_v_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (size_t)(op2)) +#define vid_v_i64m1(op0) \ +__builtin_rvv_vid_v_i64m1((size_t)(op0)) +#define vid_v_i64m1_m(op0, op1, op2) \ +__builtin_rvv_vid_v_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (size_t)(op2)) +#define vid_v_i64m2(op0) \ +__builtin_rvv_vid_v_i64m2((size_t)(op0)) +#define vid_v_i64m2_m(op0, op1, op2) \ +__builtin_rvv_vid_v_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (size_t)(op2)) +#define vid_v_i64m4(op0) \ +__builtin_rvv_vid_v_i64m4((size_t)(op0)) +#define vid_v_i64m4_m(op0, op1, op2) \ +__builtin_rvv_vid_v_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (size_t)(op2)) +#define vid_v_i64m8(op0) \ +__builtin_rvv_vid_v_i64m8((size_t)(op0)) +#define vid_v_i64m8_m(op0, op1, op2) \ +__builtin_rvv_vid_v_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (size_t)(op2)) +#define vid_v_u8m1(op0) \ +__builtin_rvv_vid_v_u8m1((size_t)(op0)) +#define vid_v_u8m1_m(op0, op1, op2) \ +__builtin_rvv_vid_v_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vid_v_u8m2(op0) \ +__builtin_rvv_vid_v_u8m2((size_t)(op0)) +#define vid_v_u8m2_m(op0, op1, op2) \ +__builtin_rvv_vid_v_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vid_v_u8m4(op0) \ +__builtin_rvv_vid_v_u8m4((size_t)(op0)) +#define vid_v_u8m4_m(op0, op1, op2) \ +__builtin_rvv_vid_v_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vid_v_u8m8(op0) \ +__builtin_rvv_vid_v_u8m8((size_t)(op0)) +#define vid_v_u8m8_m(op0, op1, op2) \ +__builtin_rvv_vid_v_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (size_t)(op2)) +#define vid_v_u8mf2(op0) \ +__builtin_rvv_vid_v_u8mf2((size_t)(op0)) +#define vid_v_u8mf2_m(op0, op1, op2) \ +__builtin_rvv_vid_v_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vid_v_u8mf4(op0) \ +__builtin_rvv_vid_v_u8mf4((size_t)(op0)) +#define vid_v_u8mf4_m(op0, op1, op2) \ +__builtin_rvv_vid_v_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vid_v_u8mf8(op0) \ +__builtin_rvv_vid_v_u8mf8((size_t)(op0)) +#define vid_v_u8mf8_m(op0, op1, op2) \ +__builtin_rvv_vid_v_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vid_v_u16m1(op0) \ +__builtin_rvv_vid_v_u16m1((size_t)(op0)) +#define vid_v_u16m1_m(op0, op1, op2) \ +__builtin_rvv_vid_v_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vid_v_u16m2(op0) \ +__builtin_rvv_vid_v_u16m2((size_t)(op0)) +#define vid_v_u16m2_m(op0, op1, op2) \ +__builtin_rvv_vid_v_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vid_v_u16m4(op0) \ +__builtin_rvv_vid_v_u16m4((size_t)(op0)) +#define vid_v_u16m4_m(op0, op1, op2) \ +__builtin_rvv_vid_v_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vid_v_u16m8(op0) \ +__builtin_rvv_vid_v_u16m8((size_t)(op0)) +#define vid_v_u16m8_m(op0, op1, op2) \ +__builtin_rvv_vid_v_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vid_v_u16mf2(op0) \ +__builtin_rvv_vid_v_u16mf2((size_t)(op0)) +#define vid_v_u16mf2_m(op0, op1, op2) \ +__builtin_rvv_vid_v_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vid_v_u16mf4(op0) \ +__builtin_rvv_vid_v_u16mf4((size_t)(op0)) +#define vid_v_u16mf4_m(op0, op1, op2) \ +__builtin_rvv_vid_v_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vid_v_u32m1(op0) \ +__builtin_rvv_vid_v_u32m1((size_t)(op0)) +#define vid_v_u32m1_m(op0, op1, op2) \ +__builtin_rvv_vid_v_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vid_v_u32m2(op0) \ +__builtin_rvv_vid_v_u32m2((size_t)(op0)) +#define vid_v_u32m2_m(op0, op1, op2) \ +__builtin_rvv_vid_v_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vid_v_u32m4(op0) \ +__builtin_rvv_vid_v_u32m4((size_t)(op0)) +#define vid_v_u32m4_m(op0, op1, op2) \ +__builtin_rvv_vid_v_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vid_v_u32m8(op0) \ +__builtin_rvv_vid_v_u32m8((size_t)(op0)) +#define vid_v_u32m8_m(op0, op1, op2) \ +__builtin_rvv_vid_v_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vid_v_u32mf2(op0) \ +__builtin_rvv_vid_v_u32mf2((size_t)(op0)) +#define vid_v_u32mf2_m(op0, op1, op2) \ +__builtin_rvv_vid_v_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vid_v_u64m1(op0) \ +__builtin_rvv_vid_v_u64m1((size_t)(op0)) +#define vid_v_u64m1_m(op0, op1, op2) \ +__builtin_rvv_vid_v_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vid_v_u64m2(op0) \ +__builtin_rvv_vid_v_u64m2((size_t)(op0)) +#define vid_v_u64m2_m(op0, op1, op2) \ +__builtin_rvv_vid_v_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vid_v_u64m4(op0) \ +__builtin_rvv_vid_v_u64m4((size_t)(op0)) +#define vid_v_u64m4_m(op0, op1, op2) \ +__builtin_rvv_vid_v_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vid_v_u64m8(op0) \ +__builtin_rvv_vid_v_u64m8((size_t)(op0)) +#define vid_v_u64m8_m(op0, op1, op2) \ +__builtin_rvv_vid_v_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vmv_x_s_i8m1_i8(op0) \ +__builtin_rvv_vmv_x_s_i8m1_i8((vint8m1_t)(op0)) +#define vmv_x_s_i8m2_i8(op0) \ +__builtin_rvv_vmv_x_s_i8m2_i8((vint8m2_t)(op0)) +#define vmv_x_s_i8m4_i8(op0) \ +__builtin_rvv_vmv_x_s_i8m4_i8((vint8m4_t)(op0)) +#define vmv_x_s_i8m8_i8(op0) \ +__builtin_rvv_vmv_x_s_i8m8_i8((vint8m8_t)(op0)) +#define vmv_x_s_i8mf2_i8(op0) \ +__builtin_rvv_vmv_x_s_i8mf2_i8((vint8mf2_t)(op0)) +#define vmv_x_s_i8mf4_i8(op0) \ +__builtin_rvv_vmv_x_s_i8mf4_i8((vint8mf4_t)(op0)) +#define vmv_x_s_i8mf8_i8(op0) \ +__builtin_rvv_vmv_x_s_i8mf8_i8((vint8mf8_t)(op0)) +#define vmv_x_s_i16m1_i16(op0) \ +__builtin_rvv_vmv_x_s_i16m1_i16((vint16m1_t)(op0)) +#define vmv_x_s_i16m2_i16(op0) \ +__builtin_rvv_vmv_x_s_i16m2_i16((vint16m2_t)(op0)) +#define vmv_x_s_i16m4_i16(op0) \ +__builtin_rvv_vmv_x_s_i16m4_i16((vint16m4_t)(op0)) +#define vmv_x_s_i16m8_i16(op0) \ +__builtin_rvv_vmv_x_s_i16m8_i16((vint16m8_t)(op0)) +#define vmv_x_s_i16mf2_i16(op0) \ +__builtin_rvv_vmv_x_s_i16mf2_i16((vint16mf2_t)(op0)) +#define vmv_x_s_i16mf4_i16(op0) \ +__builtin_rvv_vmv_x_s_i16mf4_i16((vint16mf4_t)(op0)) +#define vmv_x_s_i32m1_i32(op0) \ +__builtin_rvv_vmv_x_s_i32m1_i32((vint32m1_t)(op0)) +#define vmv_x_s_i32m2_i32(op0) \ +__builtin_rvv_vmv_x_s_i32m2_i32((vint32m2_t)(op0)) +#define vmv_x_s_i32m4_i32(op0) \ +__builtin_rvv_vmv_x_s_i32m4_i32((vint32m4_t)(op0)) +#define vmv_x_s_i32m8_i32(op0) \ +__builtin_rvv_vmv_x_s_i32m8_i32((vint32m8_t)(op0)) +#define vmv_x_s_i32mf2_i32(op0) \ +__builtin_rvv_vmv_x_s_i32mf2_i32((vint32mf2_t)(op0)) +#define vmv_x_s_i64m1_i64(op0) \ +__builtin_rvv_vmv_x_s_i64m1_i64((vint64m1_t)(op0)) +#define vmv_x_s_i64m2_i64(op0) \ +__builtin_rvv_vmv_x_s_i64m2_i64((vint64m2_t)(op0)) +#define vmv_x_s_i64m4_i64(op0) \ +__builtin_rvv_vmv_x_s_i64m4_i64((vint64m4_t)(op0)) +#define vmv_x_s_i64m8_i64(op0) \ +__builtin_rvv_vmv_x_s_i64m8_i64((vint64m8_t)(op0)) +#define vmv_x_s_u8m1_u8(op0) \ +__builtin_rvv_vmv_x_s_u8m1_u8((vuint8m1_t)(op0)) +#define vmv_x_s_u8m2_u8(op0) \ +__builtin_rvv_vmv_x_s_u8m2_u8((vuint8m2_t)(op0)) +#define vmv_x_s_u8m4_u8(op0) \ +__builtin_rvv_vmv_x_s_u8m4_u8((vuint8m4_t)(op0)) +#define vmv_x_s_u8m8_u8(op0) \ +__builtin_rvv_vmv_x_s_u8m8_u8((vuint8m8_t)(op0)) +#define vmv_x_s_u8mf2_u8(op0) \ +__builtin_rvv_vmv_x_s_u8mf2_u8((vuint8mf2_t)(op0)) +#define vmv_x_s_u8mf4_u8(op0) \ +__builtin_rvv_vmv_x_s_u8mf4_u8((vuint8mf4_t)(op0)) +#define vmv_x_s_u8mf8_u8(op0) \ +__builtin_rvv_vmv_x_s_u8mf8_u8((vuint8mf8_t)(op0)) +#define vmv_x_s_u16m1_u16(op0) \ +__builtin_rvv_vmv_x_s_u16m1_u16((vuint16m1_t)(op0)) +#define vmv_x_s_u16m2_u16(op0) \ +__builtin_rvv_vmv_x_s_u16m2_u16((vuint16m2_t)(op0)) +#define vmv_x_s_u16m4_u16(op0) \ +__builtin_rvv_vmv_x_s_u16m4_u16((vuint16m4_t)(op0)) +#define vmv_x_s_u16m8_u16(op0) \ +__builtin_rvv_vmv_x_s_u16m8_u16((vuint16m8_t)(op0)) +#define vmv_x_s_u16mf2_u16(op0) \ +__builtin_rvv_vmv_x_s_u16mf2_u16((vuint16mf2_t)(op0)) +#define vmv_x_s_u16mf4_u16(op0) \ +__builtin_rvv_vmv_x_s_u16mf4_u16((vuint16mf4_t)(op0)) +#define vmv_x_s_u32m1_u32(op0) \ +__builtin_rvv_vmv_x_s_u32m1_u32((vuint32m1_t)(op0)) +#define vmv_x_s_u32m2_u32(op0) \ +__builtin_rvv_vmv_x_s_u32m2_u32((vuint32m2_t)(op0)) +#define vmv_x_s_u32m4_u32(op0) \ +__builtin_rvv_vmv_x_s_u32m4_u32((vuint32m4_t)(op0)) +#define vmv_x_s_u32m8_u32(op0) \ +__builtin_rvv_vmv_x_s_u32m8_u32((vuint32m8_t)(op0)) +#define vmv_x_s_u32mf2_u32(op0) \ +__builtin_rvv_vmv_x_s_u32mf2_u32((vuint32mf2_t)(op0)) +#define vmv_x_s_u64m1_u64(op0) \ +__builtin_rvv_vmv_x_s_u64m1_u64((vuint64m1_t)(op0)) +#define vmv_x_s_u64m2_u64(op0) \ +__builtin_rvv_vmv_x_s_u64m2_u64((vuint64m2_t)(op0)) +#define vmv_x_s_u64m4_u64(op0) \ +__builtin_rvv_vmv_x_s_u64m4_u64((vuint64m4_t)(op0)) +#define vmv_x_s_u64m8_u64(op0) \ +__builtin_rvv_vmv_x_s_u64m8_u64((vuint64m8_t)(op0)) +#define vmv_s_x_i8m1(op0, op1, op2) \ +__builtin_rvv_vmv_s_x_i8m1((vint8m1_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmv_s_x_i8m2(op0, op1, op2) \ +__builtin_rvv_vmv_s_x_i8m2((vint8m2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmv_s_x_i8m4(op0, op1, op2) \ +__builtin_rvv_vmv_s_x_i8m4((vint8m4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmv_s_x_i8m8(op0, op1, op2) \ +__builtin_rvv_vmv_s_x_i8m8((vint8m8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmv_s_x_i8mf2(op0, op1, op2) \ +__builtin_rvv_vmv_s_x_i8mf2((vint8mf2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmv_s_x_i8mf4(op0, op1, op2) \ +__builtin_rvv_vmv_s_x_i8mf4((vint8mf4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmv_s_x_i8mf8(op0, op1, op2) \ +__builtin_rvv_vmv_s_x_i8mf8((vint8mf8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vmv_s_x_i16m1(op0, op1, op2) \ +__builtin_rvv_vmv_s_x_i16m1((vint16m1_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmv_s_x_i16m2(op0, op1, op2) \ +__builtin_rvv_vmv_s_x_i16m2((vint16m2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmv_s_x_i16m4(op0, op1, op2) \ +__builtin_rvv_vmv_s_x_i16m4((vint16m4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmv_s_x_i16m8(op0, op1, op2) \ +__builtin_rvv_vmv_s_x_i16m8((vint16m8_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmv_s_x_i16mf2(op0, op1, op2) \ +__builtin_rvv_vmv_s_x_i16mf2((vint16mf2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmv_s_x_i16mf4(op0, op1, op2) \ +__builtin_rvv_vmv_s_x_i16mf4((vint16mf4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vmv_s_x_i32m1(op0, op1, op2) \ +__builtin_rvv_vmv_s_x_i32m1((vint32m1_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmv_s_x_i32m2(op0, op1, op2) \ +__builtin_rvv_vmv_s_x_i32m2((vint32m2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmv_s_x_i32m4(op0, op1, op2) \ +__builtin_rvv_vmv_s_x_i32m4((vint32m4_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmv_s_x_i32m8(op0, op1, op2) \ +__builtin_rvv_vmv_s_x_i32m8((vint32m8_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmv_s_x_i32mf2(op0, op1, op2) \ +__builtin_rvv_vmv_s_x_i32mf2((vint32mf2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vmv_s_x_i64m1(op0, op1, op2) \ +__builtin_rvv_vmv_s_x_i64m1((vint64m1_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmv_s_x_i64m2(op0, op1, op2) \ +__builtin_rvv_vmv_s_x_i64m2((vint64m2_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmv_s_x_i64m4(op0, op1, op2) \ +__builtin_rvv_vmv_s_x_i64m4((vint64m4_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmv_s_x_i64m8(op0, op1, op2) \ +__builtin_rvv_vmv_s_x_i64m8((vint64m8_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vmv_s_x_u8m1(op0, op1, op2) \ +__builtin_rvv_vmv_s_x_u8m1((vuint8m1_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmv_s_x_u8m2(op0, op1, op2) \ +__builtin_rvv_vmv_s_x_u8m2((vuint8m2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmv_s_x_u8m4(op0, op1, op2) \ +__builtin_rvv_vmv_s_x_u8m4((vuint8m4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmv_s_x_u8m8(op0, op1, op2) \ +__builtin_rvv_vmv_s_x_u8m8((vuint8m8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmv_s_x_u8mf2(op0, op1, op2) \ +__builtin_rvv_vmv_s_x_u8mf2((vuint8mf2_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmv_s_x_u8mf4(op0, op1, op2) \ +__builtin_rvv_vmv_s_x_u8mf4((vuint8mf4_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmv_s_x_u8mf8(op0, op1, op2) \ +__builtin_rvv_vmv_s_x_u8mf8((vuint8mf8_t)(op0), (uint8_t)(op1), (size_t)(op2)) +#define vmv_s_x_u16m1(op0, op1, op2) \ +__builtin_rvv_vmv_s_x_u16m1((vuint16m1_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmv_s_x_u16m2(op0, op1, op2) \ +__builtin_rvv_vmv_s_x_u16m2((vuint16m2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmv_s_x_u16m4(op0, op1, op2) \ +__builtin_rvv_vmv_s_x_u16m4((vuint16m4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmv_s_x_u16m8(op0, op1, op2) \ +__builtin_rvv_vmv_s_x_u16m8((vuint16m8_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmv_s_x_u16mf2(op0, op1, op2) \ +__builtin_rvv_vmv_s_x_u16mf2((vuint16mf2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmv_s_x_u16mf4(op0, op1, op2) \ +__builtin_rvv_vmv_s_x_u16mf4((vuint16mf4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vmv_s_x_u32m1(op0, op1, op2) \ +__builtin_rvv_vmv_s_x_u32m1((vuint32m1_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmv_s_x_u32m2(op0, op1, op2) \ +__builtin_rvv_vmv_s_x_u32m2((vuint32m2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmv_s_x_u32m4(op0, op1, op2) \ +__builtin_rvv_vmv_s_x_u32m4((vuint32m4_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmv_s_x_u32m8(op0, op1, op2) \ +__builtin_rvv_vmv_s_x_u32m8((vuint32m8_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmv_s_x_u32mf2(op0, op1, op2) \ +__builtin_rvv_vmv_s_x_u32mf2((vuint32mf2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vmv_s_x_u64m1(op0, op1, op2) \ +__builtin_rvv_vmv_s_x_u64m1((vuint64m1_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmv_s_x_u64m2(op0, op1, op2) \ +__builtin_rvv_vmv_s_x_u64m2((vuint64m2_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmv_s_x_u64m4(op0, op1, op2) \ +__builtin_rvv_vmv_s_x_u64m4((vuint64m4_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vmv_s_x_u64m8(op0, op1, op2) \ +__builtin_rvv_vmv_s_x_u64m8((vuint64m8_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vfmv_s_x_u16m1(op0, op1, op2) \ +__builtin_rvv_vfmv_s_x_u16m1((vuint16m1_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vfmv_s_x_u16m2(op0, op1, op2) \ +__builtin_rvv_vfmv_s_x_u16m2((vuint16m2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vfmv_s_x_u16m4(op0, op1, op2) \ +__builtin_rvv_vfmv_s_x_u16m4((vuint16m4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vfmv_s_x_u16m8(op0, op1, op2) \ +__builtin_rvv_vfmv_s_x_u16m8((vuint16m8_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vfmv_s_x_u16mf2(op0, op1, op2) \ +__builtin_rvv_vfmv_s_x_u16mf2((vuint16mf2_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vfmv_s_x_u16mf4(op0, op1, op2) \ +__builtin_rvv_vfmv_s_x_u16mf4((vuint16mf4_t)(op0), (uint16_t)(op1), (size_t)(op2)) +#define vfmv_s_x_u32m1(op0, op1, op2) \ +__builtin_rvv_vfmv_s_x_u32m1((vuint32m1_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vfmv_s_x_u32m2(op0, op1, op2) \ +__builtin_rvv_vfmv_s_x_u32m2((vuint32m2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vfmv_s_x_u32m4(op0, op1, op2) \ +__builtin_rvv_vfmv_s_x_u32m4((vuint32m4_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vfmv_s_x_u32m8(op0, op1, op2) \ +__builtin_rvv_vfmv_s_x_u32m8((vuint32m8_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vfmv_s_x_u32mf2(op0, op1, op2) \ +__builtin_rvv_vfmv_s_x_u32mf2((vuint32mf2_t)(op0), (uint32_t)(op1), (size_t)(op2)) +#define vfmv_s_x_u64m1(op0, op1, op2) \ +__builtin_rvv_vfmv_s_x_u64m1((vuint64m1_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vfmv_s_x_u64m2(op0, op1, op2) \ +__builtin_rvv_vfmv_s_x_u64m2((vuint64m2_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vfmv_s_x_u64m4(op0, op1, op2) \ +__builtin_rvv_vfmv_s_x_u64m4((vuint64m4_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vfmv_s_x_u64m8(op0, op1, op2) \ +__builtin_rvv_vfmv_s_x_u64m8((vuint64m8_t)(op0), (uint64_t)(op1), (size_t)(op2)) +#define vslideup_vx_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_i8m1((vint8m1_t)(op0), (vint8m1_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_i8m2(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_i8m2((vint8m2_t)(op0), (vint8m2_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_i8m4(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_i8m4((vint8m4_t)(op0), (vint8m4_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_i8m8(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_i8m8((vint8m8_t)(op0), (vint8m8_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_i8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_i8mf2((vint8mf2_t)(op0), (vint8mf2_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_i8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_i8mf4((vint8mf4_t)(op0), (vint8mf4_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_i8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_i8mf8((vint8mf8_t)(op0), (vint8mf8_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_i16m1((vint16m1_t)(op0), (vint16m1_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_i16m2(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_i16m2((vint16m2_t)(op0), (vint16m2_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_i16m4(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_i16m4((vint16m4_t)(op0), (vint16m4_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_i16m8(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_i16m8((vint16m8_t)(op0), (vint16m8_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_i16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_i16mf2((vint16mf2_t)(op0), (vint16mf2_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_i16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_i16mf4((vint16mf4_t)(op0), (vint16mf4_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_i32m1((vint32m1_t)(op0), (vint32m1_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_i32m2((vint32m2_t)(op0), (vint32m2_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_i32m4((vint32m4_t)(op0), (vint32m4_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_i32m8((vint32m8_t)(op0), (vint32m8_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_i32mf2((vint32mf2_t)(op0), (vint32mf2_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_i64m1((vint64m1_t)(op0), (vint64m1_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_i64m2((vint64m2_t)(op0), (vint64m2_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_i64m4((vint64m4_t)(op0), (vint64m4_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_i64m8((vint64m8_t)(op0), (vint64m8_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_u8m1((vuint8m1_t)(op0), (vuint8m1_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_u8m2(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_u8m2((vuint8m2_t)(op0), (vuint8m2_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_u8m4(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_u8m4((vuint8m4_t)(op0), (vuint8m4_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_u8m8(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_u8m8((vuint8m8_t)(op0), (vuint8m8_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_u8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_u8mf2((vuint8mf2_t)(op0), (vuint8mf2_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_u8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_u8mf4((vuint8mf4_t)(op0), (vuint8mf4_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_u8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_u8mf8((vuint8mf8_t)(op0), (vuint8mf8_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_u16m1((vuint16m1_t)(op0), (vuint16m1_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_u16m2(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_u16m2((vuint16m2_t)(op0), (vuint16m2_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_u16m4(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_u16m4((vuint16m4_t)(op0), (vuint16m4_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_u16m8(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_u16m8((vuint16m8_t)(op0), (vuint16m8_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_u16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_u16mf2((vuint16mf2_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_u16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_u16mf4((vuint16mf4_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_u32m1((vuint32m1_t)(op0), (vuint32m1_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_u32m2((vuint32m2_t)(op0), (vuint32m2_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_u32m4((vuint32m4_t)(op0), (vuint32m4_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_u32m8((vuint32m8_t)(op0), (vuint32m8_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_u32mf2((vuint32mf2_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_u64m1((vuint64m1_t)(op0), (vuint64m1_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_u64m2((vuint64m2_t)(op0), (vuint64m2_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_u64m4((vuint64m4_t)(op0), (vuint64m4_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_u64m8((vuint64m8_t)(op0), (vuint64m8_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_i8m1((vint8m1_t)(op0), (vint8m1_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_i8m2(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_i8m2((vint8m2_t)(op0), (vint8m2_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_i8m4(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_i8m4((vint8m4_t)(op0), (vint8m4_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_i8m8(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_i8m8((vint8m8_t)(op0), (vint8m8_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_i8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_i8mf2((vint8mf2_t)(op0), (vint8mf2_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_i8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_i8mf4((vint8mf4_t)(op0), (vint8mf4_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_i8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_i8mf8((vint8mf8_t)(op0), (vint8mf8_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_i16m1((vint16m1_t)(op0), (vint16m1_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_i16m2(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_i16m2((vint16m2_t)(op0), (vint16m2_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_i16m4(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_i16m4((vint16m4_t)(op0), (vint16m4_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_i16m8(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_i16m8((vint16m8_t)(op0), (vint16m8_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_i16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_i16mf2((vint16mf2_t)(op0), (vint16mf2_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_i16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_i16mf4((vint16mf4_t)(op0), (vint16mf4_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_i32m1((vint32m1_t)(op0), (vint32m1_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_i32m2((vint32m2_t)(op0), (vint32m2_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_i32m4((vint32m4_t)(op0), (vint32m4_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_i32m8((vint32m8_t)(op0), (vint32m8_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_i32mf2((vint32mf2_t)(op0), (vint32mf2_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_i64m1((vint64m1_t)(op0), (vint64m1_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_i64m2((vint64m2_t)(op0), (vint64m2_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_i64m4((vint64m4_t)(op0), (vint64m4_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_i64m8((vint64m8_t)(op0), (vint64m8_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_u8m1((vuint8m1_t)(op0), (vuint8m1_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_u8m2(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_u8m2((vuint8m2_t)(op0), (vuint8m2_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_u8m4(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_u8m4((vuint8m4_t)(op0), (vuint8m4_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_u8m8(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_u8m8((vuint8m8_t)(op0), (vuint8m8_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_u8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_u8mf2((vuint8mf2_t)(op0), (vuint8mf2_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_u8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_u8mf4((vuint8mf4_t)(op0), (vuint8mf4_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_u8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_u8mf8((vuint8mf8_t)(op0), (vuint8mf8_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_u16m1((vuint16m1_t)(op0), (vuint16m1_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_u16m2(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_u16m2((vuint16m2_t)(op0), (vuint16m2_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_u16m4(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_u16m4((vuint16m4_t)(op0), (vuint16m4_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_u16m8(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_u16m8((vuint16m8_t)(op0), (vuint16m8_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_u16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_u16mf2((vuint16mf2_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_u16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_u16mf4((vuint16mf4_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_u32m1((vuint32m1_t)(op0), (vuint32m1_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_u32m2((vuint32m2_t)(op0), (vuint32m2_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_u32m4((vuint32m4_t)(op0), (vuint32m4_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_u32m8((vuint32m8_t)(op0), (vuint32m8_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_u32mf2((vuint32mf2_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_u64m1((vuint64m1_t)(op0), (vuint64m1_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_u64m2((vuint64m2_t)(op0), (vuint64m2_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_u64m4((vuint64m4_t)(op0), (vuint64m4_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_u64m8((vuint64m8_t)(op0), (vuint64m8_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslide1up_vx_i8m1(op0, op1, op2) \ +__builtin_rvv_vslide1up_vx_i8m1((vint8m1_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vslide1up_vx_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1up_vx_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vslide1up_vx_i8m2(op0, op1, op2) \ +__builtin_rvv_vslide1up_vx_i8m2((vint8m2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vslide1up_vx_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1up_vx_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vslide1up_vx_i8m4(op0, op1, op2) \ +__builtin_rvv_vslide1up_vx_i8m4((vint8m4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vslide1up_vx_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1up_vx_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vslide1up_vx_i8m8(op0, op1, op2) \ +__builtin_rvv_vslide1up_vx_i8m8((vint8m8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vslide1up_vx_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1up_vx_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vslide1up_vx_i8mf2(op0, op1, op2) \ +__builtin_rvv_vslide1up_vx_i8mf2((vint8mf2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vslide1up_vx_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1up_vx_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vslide1up_vx_i8mf4(op0, op1, op2) \ +__builtin_rvv_vslide1up_vx_i8mf4((vint8mf4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vslide1up_vx_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1up_vx_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vslide1up_vx_i8mf8(op0, op1, op2) \ +__builtin_rvv_vslide1up_vx_i8mf8((vint8mf8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vslide1up_vx_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1up_vx_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vslide1up_vx_i16m1(op0, op1, op2) \ +__builtin_rvv_vslide1up_vx_i16m1((vint16m1_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vslide1up_vx_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1up_vx_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vslide1up_vx_i16m2(op0, op1, op2) \ +__builtin_rvv_vslide1up_vx_i16m2((vint16m2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vslide1up_vx_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1up_vx_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vslide1up_vx_i16m4(op0, op1, op2) \ +__builtin_rvv_vslide1up_vx_i16m4((vint16m4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vslide1up_vx_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1up_vx_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vslide1up_vx_i16m8(op0, op1, op2) \ +__builtin_rvv_vslide1up_vx_i16m8((vint16m8_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vslide1up_vx_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1up_vx_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vslide1up_vx_i16mf2(op0, op1, op2) \ +__builtin_rvv_vslide1up_vx_i16mf2((vint16mf2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vslide1up_vx_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1up_vx_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vslide1up_vx_i16mf4(op0, op1, op2) \ +__builtin_rvv_vslide1up_vx_i16mf4((vint16mf4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vslide1up_vx_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1up_vx_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vslide1up_vx_i32m1(op0, op1, op2) \ +__builtin_rvv_vslide1up_vx_i32m1((vint32m1_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vslide1up_vx_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1up_vx_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vslide1up_vx_i32m2(op0, op1, op2) \ +__builtin_rvv_vslide1up_vx_i32m2((vint32m2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vslide1up_vx_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1up_vx_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vslide1up_vx_i32m4(op0, op1, op2) \ +__builtin_rvv_vslide1up_vx_i32m4((vint32m4_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vslide1up_vx_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1up_vx_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vslide1up_vx_i32m8(op0, op1, op2) \ +__builtin_rvv_vslide1up_vx_i32m8((vint32m8_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vslide1up_vx_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1up_vx_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vslide1up_vx_i32mf2(op0, op1, op2) \ +__builtin_rvv_vslide1up_vx_i32mf2((vint32mf2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vslide1up_vx_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1up_vx_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vslide1up_vx_i64m1(op0, op1, op2) \ +__builtin_rvv_vslide1up_vx_i64m1((vint64m1_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vslide1up_vx_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1up_vx_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vslide1up_vx_i64m2(op0, op1, op2) \ +__builtin_rvv_vslide1up_vx_i64m2((vint64m2_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vslide1up_vx_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1up_vx_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vslide1up_vx_i64m4(op0, op1, op2) \ +__builtin_rvv_vslide1up_vx_i64m4((vint64m4_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vslide1up_vx_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1up_vx_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vslide1up_vx_i64m8(op0, op1, op2) \ +__builtin_rvv_vslide1up_vx_i64m8((vint64m8_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vslide1up_vx_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1up_vx_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vslide1up_vx_u8m1(op0, op1, op2) \ +__builtin_rvv_vslide1up_vx_u8m1((vuint8m1_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vslide1up_vx_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1up_vx_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vslide1up_vx_u8m2(op0, op1, op2) \ +__builtin_rvv_vslide1up_vx_u8m2((vuint8m2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vslide1up_vx_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1up_vx_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vslide1up_vx_u8m4(op0, op1, op2) \ +__builtin_rvv_vslide1up_vx_u8m4((vuint8m4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vslide1up_vx_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1up_vx_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vslide1up_vx_u8m8(op0, op1, op2) \ +__builtin_rvv_vslide1up_vx_u8m8((vuint8m8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vslide1up_vx_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1up_vx_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vslide1up_vx_u8mf2(op0, op1, op2) \ +__builtin_rvv_vslide1up_vx_u8mf2((vuint8mf2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vslide1up_vx_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1up_vx_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vslide1up_vx_u8mf4(op0, op1, op2) \ +__builtin_rvv_vslide1up_vx_u8mf4((vuint8mf4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vslide1up_vx_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1up_vx_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vslide1up_vx_u8mf8(op0, op1, op2) \ +__builtin_rvv_vslide1up_vx_u8mf8((vuint8mf8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vslide1up_vx_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1up_vx_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vslide1up_vx_u16m1(op0, op1, op2) \ +__builtin_rvv_vslide1up_vx_u16m1((vuint16m1_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vslide1up_vx_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1up_vx_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vslide1up_vx_u16m2(op0, op1, op2) \ +__builtin_rvv_vslide1up_vx_u16m2((vuint16m2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vslide1up_vx_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1up_vx_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vslide1up_vx_u16m4(op0, op1, op2) \ +__builtin_rvv_vslide1up_vx_u16m4((vuint16m4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vslide1up_vx_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1up_vx_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vslide1up_vx_u16m8(op0, op1, op2) \ +__builtin_rvv_vslide1up_vx_u16m8((vuint16m8_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vslide1up_vx_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1up_vx_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vslide1up_vx_u16mf2(op0, op1, op2) \ +__builtin_rvv_vslide1up_vx_u16mf2((vuint16mf2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vslide1up_vx_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1up_vx_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vslide1up_vx_u16mf4(op0, op1, op2) \ +__builtin_rvv_vslide1up_vx_u16mf4((vuint16mf4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vslide1up_vx_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1up_vx_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vslide1up_vx_u32m1(op0, op1, op2) \ +__builtin_rvv_vslide1up_vx_u32m1((vuint32m1_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vslide1up_vx_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1up_vx_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vslide1up_vx_u32m2(op0, op1, op2) \ +__builtin_rvv_vslide1up_vx_u32m2((vuint32m2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vslide1up_vx_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1up_vx_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vslide1up_vx_u32m4(op0, op1, op2) \ +__builtin_rvv_vslide1up_vx_u32m4((vuint32m4_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vslide1up_vx_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1up_vx_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vslide1up_vx_u32m8(op0, op1, op2) \ +__builtin_rvv_vslide1up_vx_u32m8((vuint32m8_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vslide1up_vx_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1up_vx_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vslide1up_vx_u32mf2(op0, op1, op2) \ +__builtin_rvv_vslide1up_vx_u32mf2((vuint32mf2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vslide1up_vx_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1up_vx_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vslide1up_vx_u64m1(op0, op1, op2) \ +__builtin_rvv_vslide1up_vx_u64m1((vuint64m1_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vslide1up_vx_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1up_vx_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vslide1up_vx_u64m2(op0, op1, op2) \ +__builtin_rvv_vslide1up_vx_u64m2((vuint64m2_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vslide1up_vx_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1up_vx_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vslide1up_vx_u64m4(op0, op1, op2) \ +__builtin_rvv_vslide1up_vx_u64m4((vuint64m4_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vslide1up_vx_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1up_vx_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vslide1up_vx_u64m8(op0, op1, op2) \ +__builtin_rvv_vslide1up_vx_u64m8((vuint64m8_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vslide1up_vx_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1up_vx_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vslide1down_vx_i8m1(op0, op1, op2) \ +__builtin_rvv_vslide1down_vx_i8m1((vint8m1_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vslide1down_vx_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1down_vx_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vslide1down_vx_i8m2(op0, op1, op2) \ +__builtin_rvv_vslide1down_vx_i8m2((vint8m2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vslide1down_vx_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1down_vx_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vslide1down_vx_i8m4(op0, op1, op2) \ +__builtin_rvv_vslide1down_vx_i8m4((vint8m4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vslide1down_vx_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1down_vx_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vslide1down_vx_i8m8(op0, op1, op2) \ +__builtin_rvv_vslide1down_vx_i8m8((vint8m8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vslide1down_vx_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1down_vx_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vslide1down_vx_i8mf2(op0, op1, op2) \ +__builtin_rvv_vslide1down_vx_i8mf2((vint8mf2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vslide1down_vx_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1down_vx_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vslide1down_vx_i8mf4(op0, op1, op2) \ +__builtin_rvv_vslide1down_vx_i8mf4((vint8mf4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vslide1down_vx_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1down_vx_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vslide1down_vx_i8mf8(op0, op1, op2) \ +__builtin_rvv_vslide1down_vx_i8mf8((vint8mf8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vslide1down_vx_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1down_vx_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vslide1down_vx_i16m1(op0, op1, op2) \ +__builtin_rvv_vslide1down_vx_i16m1((vint16m1_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vslide1down_vx_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1down_vx_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vslide1down_vx_i16m2(op0, op1, op2) \ +__builtin_rvv_vslide1down_vx_i16m2((vint16m2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vslide1down_vx_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1down_vx_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vslide1down_vx_i16m4(op0, op1, op2) \ +__builtin_rvv_vslide1down_vx_i16m4((vint16m4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vslide1down_vx_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1down_vx_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vslide1down_vx_i16m8(op0, op1, op2) \ +__builtin_rvv_vslide1down_vx_i16m8((vint16m8_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vslide1down_vx_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1down_vx_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vslide1down_vx_i16mf2(op0, op1, op2) \ +__builtin_rvv_vslide1down_vx_i16mf2((vint16mf2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vslide1down_vx_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1down_vx_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vslide1down_vx_i16mf4(op0, op1, op2) \ +__builtin_rvv_vslide1down_vx_i16mf4((vint16mf4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vslide1down_vx_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1down_vx_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vslide1down_vx_i32m1(op0, op1, op2) \ +__builtin_rvv_vslide1down_vx_i32m1((vint32m1_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vslide1down_vx_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1down_vx_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vslide1down_vx_i32m2(op0, op1, op2) \ +__builtin_rvv_vslide1down_vx_i32m2((vint32m2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vslide1down_vx_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1down_vx_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vslide1down_vx_i32m4(op0, op1, op2) \ +__builtin_rvv_vslide1down_vx_i32m4((vint32m4_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vslide1down_vx_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1down_vx_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vslide1down_vx_i32m8(op0, op1, op2) \ +__builtin_rvv_vslide1down_vx_i32m8((vint32m8_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vslide1down_vx_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1down_vx_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vslide1down_vx_i32mf2(op0, op1, op2) \ +__builtin_rvv_vslide1down_vx_i32mf2((vint32mf2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vslide1down_vx_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1down_vx_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vslide1down_vx_i64m1(op0, op1, op2) \ +__builtin_rvv_vslide1down_vx_i64m1((vint64m1_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vslide1down_vx_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1down_vx_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vslide1down_vx_i64m2(op0, op1, op2) \ +__builtin_rvv_vslide1down_vx_i64m2((vint64m2_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vslide1down_vx_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1down_vx_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vslide1down_vx_i64m4(op0, op1, op2) \ +__builtin_rvv_vslide1down_vx_i64m4((vint64m4_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vslide1down_vx_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1down_vx_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vslide1down_vx_i64m8(op0, op1, op2) \ +__builtin_rvv_vslide1down_vx_i64m8((vint64m8_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vslide1down_vx_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1down_vx_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vslide1down_vx_u8m1(op0, op1, op2) \ +__builtin_rvv_vslide1down_vx_u8m1((vuint8m1_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vslide1down_vx_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1down_vx_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vslide1down_vx_u8m2(op0, op1, op2) \ +__builtin_rvv_vslide1down_vx_u8m2((vuint8m2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vslide1down_vx_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1down_vx_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vslide1down_vx_u8m4(op0, op1, op2) \ +__builtin_rvv_vslide1down_vx_u8m4((vuint8m4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vslide1down_vx_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1down_vx_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vslide1down_vx_u8m8(op0, op1, op2) \ +__builtin_rvv_vslide1down_vx_u8m8((vuint8m8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vslide1down_vx_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1down_vx_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vslide1down_vx_u8mf2(op0, op1, op2) \ +__builtin_rvv_vslide1down_vx_u8mf2((vuint8mf2_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vslide1down_vx_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1down_vx_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vslide1down_vx_u8mf4(op0, op1, op2) \ +__builtin_rvv_vslide1down_vx_u8mf4((vuint8mf4_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vslide1down_vx_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1down_vx_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vslide1down_vx_u8mf8(op0, op1, op2) \ +__builtin_rvv_vslide1down_vx_u8mf8((vuint8mf8_t)(op0), (int8_t)(op1), (size_t)(op2)) +#define vslide1down_vx_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1down_vx_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (int8_t)(op3), (size_t)(op4)) +#define vslide1down_vx_u16m1(op0, op1, op2) \ +__builtin_rvv_vslide1down_vx_u16m1((vuint16m1_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vslide1down_vx_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1down_vx_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vslide1down_vx_u16m2(op0, op1, op2) \ +__builtin_rvv_vslide1down_vx_u16m2((vuint16m2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vslide1down_vx_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1down_vx_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vslide1down_vx_u16m4(op0, op1, op2) \ +__builtin_rvv_vslide1down_vx_u16m4((vuint16m4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vslide1down_vx_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1down_vx_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vslide1down_vx_u16m8(op0, op1, op2) \ +__builtin_rvv_vslide1down_vx_u16m8((vuint16m8_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vslide1down_vx_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1down_vx_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vslide1down_vx_u16mf2(op0, op1, op2) \ +__builtin_rvv_vslide1down_vx_u16mf2((vuint16mf2_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vslide1down_vx_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1down_vx_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vslide1down_vx_u16mf4(op0, op1, op2) \ +__builtin_rvv_vslide1down_vx_u16mf4((vuint16mf4_t)(op0), (int16_t)(op1), (size_t)(op2)) +#define vslide1down_vx_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1down_vx_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (int16_t)(op3), (size_t)(op4)) +#define vslide1down_vx_u32m1(op0, op1, op2) \ +__builtin_rvv_vslide1down_vx_u32m1((vuint32m1_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vslide1down_vx_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1down_vx_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vslide1down_vx_u32m2(op0, op1, op2) \ +__builtin_rvv_vslide1down_vx_u32m2((vuint32m2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vslide1down_vx_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1down_vx_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vslide1down_vx_u32m4(op0, op1, op2) \ +__builtin_rvv_vslide1down_vx_u32m4((vuint32m4_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vslide1down_vx_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1down_vx_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vslide1down_vx_u32m8(op0, op1, op2) \ +__builtin_rvv_vslide1down_vx_u32m8((vuint32m8_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vslide1down_vx_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1down_vx_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vslide1down_vx_u32mf2(op0, op1, op2) \ +__builtin_rvv_vslide1down_vx_u32mf2((vuint32mf2_t)(op0), (int32_t)(op1), (size_t)(op2)) +#define vslide1down_vx_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1down_vx_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (int32_t)(op3), (size_t)(op4)) +#define vslide1down_vx_u64m1(op0, op1, op2) \ +__builtin_rvv_vslide1down_vx_u64m1((vuint64m1_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vslide1down_vx_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1down_vx_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vslide1down_vx_u64m2(op0, op1, op2) \ +__builtin_rvv_vslide1down_vx_u64m2((vuint64m2_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vslide1down_vx_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1down_vx_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vslide1down_vx_u64m4(op0, op1, op2) \ +__builtin_rvv_vslide1down_vx_u64m4((vuint64m4_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vslide1down_vx_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1down_vx_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vslide1down_vx_u64m8(op0, op1, op2) \ +__builtin_rvv_vslide1down_vx_u64m8((vuint64m8_t)(op0), (int64_t)(op1), (size_t)(op2)) +#define vslide1down_vx_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslide1down_vx_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (int64_t)(op3), (size_t)(op4)) +#define vrgather_vv_i8m1(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_i8m1((vint8m1_t)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vrgather_vv_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vrgather_vv_i8m2(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_i8m2((vint8m2_t)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vrgather_vv_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vrgather_vv_i8m4(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_i8m4((vint8m4_t)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vrgather_vv_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vrgather_vv_i8m8(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_i8m8((vint8m8_t)(op0), (vuint8m8_t)(op1), (size_t)(op2)) +#define vrgather_vv_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (vuint8m8_t)(op3), (size_t)(op4)) +#define vrgather_vv_i8mf2(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_i8mf2((vint8mf2_t)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vrgather_vv_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vrgather_vv_i8mf4(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_i8mf4((vint8mf4_t)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vrgather_vv_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vrgather_vv_i8mf8(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_i8mf8((vint8mf8_t)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vrgather_vv_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vrgather_vv_i16m1(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_i16m1((vint16m1_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vrgather_vv_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vrgather_vv_i16m2(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_i16m2((vint16m2_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vrgather_vv_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vrgather_vv_i16m4(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_i16m4((vint16m4_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vrgather_vv_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vrgather_vv_i16m8(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_i16m8((vint16m8_t)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vrgather_vv_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vrgather_vv_i16mf2(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_i16mf2((vint16mf2_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vrgather_vv_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vrgather_vv_i16mf4(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_i16mf4((vint16mf4_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vrgather_vv_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vrgather_vv_i32m1(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_i32m1((vint32m1_t)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vrgather_vv_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vrgather_vv_i32m2(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_i32m2((vint32m2_t)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vrgather_vv_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vrgather_vv_i32m4(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_i32m4((vint32m4_t)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vrgather_vv_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vrgather_vv_i32m8(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_i32m8((vint32m8_t)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vrgather_vv_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vrgather_vv_i32mf2(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_i32mf2((vint32mf2_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vrgather_vv_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vrgather_vv_i64m1(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_i64m1((vint64m1_t)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vrgather_vv_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vrgather_vv_i64m2(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_i64m2((vint64m2_t)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vrgather_vv_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vrgather_vv_i64m4(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_i64m4((vint64m4_t)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vrgather_vv_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vrgather_vv_i64m8(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_i64m8((vint64m8_t)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vrgather_vv_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vrgather_vx_i8m1(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_i8m1((vint8m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_i8m2(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_i8m2((vint8m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_i8m4(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_i8m4((vint8m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_i8m8(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_i8m8((vint8m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_i8mf2(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_i8mf2((vint8mf2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_i8mf4(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_i8mf4((vint8mf4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_i8mf8(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_i8mf8((vint8mf8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_i16m1(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_i16m1((vint16m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_i16m2(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_i16m2((vint16m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_i16m4(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_i16m4((vint16m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_i16m8(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_i16m8((vint16m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_i16mf2(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_i16mf2((vint16mf2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_i16mf4(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_i16mf4((vint16mf4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_i32m1(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_i32m1((vint32m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_i32m2(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_i32m2((vint32m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_i32m4(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_i32m4((vint32m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_i32m8(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_i32m8((vint32m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_i32mf2(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_i32mf2((vint32mf2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_i64m1(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_i64m1((vint64m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_i64m2(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_i64m2((vint64m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_i64m4(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_i64m4((vint64m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_i64m8(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_i64m8((vint64m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_i8m1(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_i8m1((vint8m1_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_i8m2(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_i8m2((vint8m2_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_i8m4(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_i8m4((vint8m4_t)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_i8mf2(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_i8mf2((vint8mf2_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_i8mf4(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_i8mf4((vint8mf4_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_i8mf8(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_i8mf8((vint8mf8_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_i16m1(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_i16m1((vint16m1_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_i16m2(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_i16m2((vint16m2_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_i16m4(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_i16m4((vint16m4_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_i16m8(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_i16m8((vint16m8_t)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_i16mf2(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_i16mf2((vint16mf2_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_i16mf4(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_i16mf4((vint16mf4_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_i32m1(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_i32m1((vint32m1_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_i32m2(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_i32m2((vint32m2_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_i32m4(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_i32m4((vint32m4_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_i32m8(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_i32m8((vint32m8_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_i32mf2(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_i32mf2((vint32mf2_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_i64m1(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_i64m1((vint64m1_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_i64m2(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_i64m2((vint64m2_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_i64m4(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_i64m4((vint64m4_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_i64m8(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_i64m8((vint64m8_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vrgather_vv_u8m1(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_u8m1((vuint8m1_t)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vrgather_vv_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vrgather_vv_u8m2(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_u8m2((vuint8m2_t)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vrgather_vv_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vrgather_vv_u8m4(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_u8m4((vuint8m4_t)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vrgather_vv_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vrgather_vv_u8m8(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_u8m8((vuint8m8_t)(op0), (vuint8m8_t)(op1), (size_t)(op2)) +#define vrgather_vv_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (vuint8m8_t)(op3), (size_t)(op4)) +#define vrgather_vv_u8mf2(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_u8mf2((vuint8mf2_t)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vrgather_vv_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vrgather_vv_u8mf4(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_u8mf4((vuint8mf4_t)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vrgather_vv_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vrgather_vv_u8mf8(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_u8mf8((vuint8mf8_t)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vrgather_vv_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vrgather_vv_u16m1(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_u16m1((vuint16m1_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vrgather_vv_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vrgather_vv_u16m2(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_u16m2((vuint16m2_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vrgather_vv_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vrgather_vv_u16m4(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_u16m4((vuint16m4_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vrgather_vv_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vrgather_vv_u16m8(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_u16m8((vuint16m8_t)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vrgather_vv_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vrgather_vv_u16mf2(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_u16mf2((vuint16mf2_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vrgather_vv_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vrgather_vv_u16mf4(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_u16mf4((vuint16mf4_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vrgather_vv_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vrgather_vv_u32m1(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_u32m1((vuint32m1_t)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vrgather_vv_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vrgather_vv_u32m2(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_u32m2((vuint32m2_t)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vrgather_vv_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vrgather_vv_u32m4(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_u32m4((vuint32m4_t)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vrgather_vv_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vrgather_vv_u32m8(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_u32m8((vuint32m8_t)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vrgather_vv_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vrgather_vv_u32mf2(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_u32mf2((vuint32mf2_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vrgather_vv_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vrgather_vv_u64m1(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_u64m1((vuint64m1_t)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vrgather_vv_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vrgather_vv_u64m2(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_u64m2((vuint64m2_t)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vrgather_vv_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vrgather_vv_u64m4(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_u64m4((vuint64m4_t)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vrgather_vv_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vrgather_vv_u64m8(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_u64m8((vuint64m8_t)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vrgather_vv_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vrgather_vx_u8m1(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_u8m1((vuint8m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_u8m2(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_u8m2((vuint8m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_u8m4(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_u8m4((vuint8m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_u8m8(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_u8m8((vuint8m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_u8mf2(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_u8mf2((vuint8mf2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_u8mf4(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_u8mf4((vuint8mf4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_u8mf8(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_u8mf8((vuint8mf8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_u16m1(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_u16m1((vuint16m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_u16m2(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_u16m2((vuint16m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_u16m4(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_u16m4((vuint16m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_u16m8(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_u16m8((vuint16m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_u16mf2(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_u16mf2((vuint16mf2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_u16mf4(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_u16mf4((vuint16mf4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_u32m1(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_u32m1((vuint32m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_u32m2(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_u32m2((vuint32m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_u32m4(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_u32m4((vuint32m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_u32m8(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_u32m8((vuint32m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_u32mf2(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_u32mf2((vuint32mf2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_u64m1(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_u64m1((vuint64m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_u64m2(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_u64m2((vuint64m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_u64m4(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_u64m4((vuint64m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_u64m8(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_u64m8((vuint64m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_u8m1(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_u8m1((vuint8m1_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_u8m2(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_u8m2((vuint8m2_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_u8m4(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_u8m4((vuint8m4_t)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_u8mf2(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_u8mf2((vuint8mf2_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_u8mf4(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_u8mf4((vuint8mf4_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_u8mf8(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_u8mf8((vuint8mf8_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_u16m1(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_u16m1((vuint16m1_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_u16m2(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_u16m2((vuint16m2_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_u16m4(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_u16m4((vuint16m4_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_u16m8(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_u16m8((vuint16m8_t)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_u16mf2(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_u16mf2((vuint16mf2_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_u16mf4(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_u16mf4((vuint16mf4_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_u32m1(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_u32m1((vuint32m1_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_u32m2(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_u32m2((vuint32m2_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_u32m4(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_u32m4((vuint32m4_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_u32m8(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_u32m8((vuint32m8_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_u32mf2(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_u32mf2((vuint32mf2_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_u64m1(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_u64m1((vuint64m1_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_u64m2(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_u64m2((vuint64m2_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_u64m4(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_u64m4((vuint64m4_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_u64m8(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_u64m8((vuint64m8_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vcompress_vm_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_i8m1((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vcompress_vm_i8m2(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_i8m2((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (size_t)(op3)) +#define vcompress_vm_i8m4(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_i8m4((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (size_t)(op3)) +#define vcompress_vm_i8m8(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_i8m8((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (size_t)(op3)) +#define vcompress_vm_i8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_i8mf2((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (size_t)(op3)) +#define vcompress_vm_i8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_i8mf4((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (size_t)(op3)) +#define vcompress_vm_i8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_i8mf8((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (size_t)(op3)) +#define vcompress_vm_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_i16m1((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vcompress_vm_i16m2(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_i16m2((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (size_t)(op3)) +#define vcompress_vm_i16m4(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_i16m4((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (size_t)(op3)) +#define vcompress_vm_i16m8(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_i16m8((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (size_t)(op3)) +#define vcompress_vm_i16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_i16mf2((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (size_t)(op3)) +#define vcompress_vm_i16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_i16mf4((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (size_t)(op3)) +#define vcompress_vm_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_i32m1((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vcompress_vm_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_i32m2((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vcompress_vm_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_i32m4((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vcompress_vm_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_i32m8((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (size_t)(op3)) +#define vcompress_vm_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_i32mf2((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vcompress_vm_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_i64m1((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vcompress_vm_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_i64m2((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vcompress_vm_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_i64m4((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vcompress_vm_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_i64m8((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vcompress_vm_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_u8m1((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vcompress_vm_u8m2(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_u8m2((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (size_t)(op3)) +#define vcompress_vm_u8m4(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_u8m4((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (size_t)(op3)) +#define vcompress_vm_u8m8(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_u8m8((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (size_t)(op3)) +#define vcompress_vm_u8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_u8mf2((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (size_t)(op3)) +#define vcompress_vm_u8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_u8mf4((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (size_t)(op3)) +#define vcompress_vm_u8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_u8mf8((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (size_t)(op3)) +#define vcompress_vm_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_u16m1((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vcompress_vm_u16m2(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_u16m2((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (size_t)(op3)) +#define vcompress_vm_u16m4(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_u16m4((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (size_t)(op3)) +#define vcompress_vm_u16m8(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_u16m8((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (size_t)(op3)) +#define vcompress_vm_u16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_u16mf2((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (size_t)(op3)) +#define vcompress_vm_u16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_u16mf4((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (size_t)(op3)) +#define vcompress_vm_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_u32m1((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vcompress_vm_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_u32m2((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vcompress_vm_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_u32m4((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vcompress_vm_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_u32m8((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (size_t)(op3)) +#define vcompress_vm_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_u32mf2((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vcompress_vm_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_u64m1((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vcompress_vm_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_u64m2((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vcompress_vm_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_u64m4((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vcompress_vm_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_u64m8((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vget_v_i8m2_i8m1(op0, op1) \ +__builtin_rvv_vget_v_i8m2_i8m1((vint8m2_t)(op0), (size_t)(op1)) +#define vget_v_i8m4_i8m1(op0, op1) \ +__builtin_rvv_vget_v_i8m4_i8m1((vint8m4_t)(op0), (size_t)(op1)) +#define vget_v_i8m8_i8m1(op0, op1) \ +__builtin_rvv_vget_v_i8m8_i8m1((vint8m8_t)(op0), (size_t)(op1)) +#define vget_v_i16m2_i16m1(op0, op1) \ +__builtin_rvv_vget_v_i16m2_i16m1((vint16m2_t)(op0), (size_t)(op1)) +#define vget_v_i16m4_i16m1(op0, op1) \ +__builtin_rvv_vget_v_i16m4_i16m1((vint16m4_t)(op0), (size_t)(op1)) +#define vget_v_i16m8_i16m1(op0, op1) \ +__builtin_rvv_vget_v_i16m8_i16m1((vint16m8_t)(op0), (size_t)(op1)) +#define vget_v_i32m2_i32m1(op0, op1) \ +__builtin_rvv_vget_v_i32m2_i32m1((vint32m2_t)(op0), (size_t)(op1)) +#define vget_v_i32m4_i32m1(op0, op1) \ +__builtin_rvv_vget_v_i32m4_i32m1((vint32m4_t)(op0), (size_t)(op1)) +#define vget_v_i32m8_i32m1(op0, op1) \ +__builtin_rvv_vget_v_i32m8_i32m1((vint32m8_t)(op0), (size_t)(op1)) +#define vget_v_i64m2_i64m1(op0, op1) \ +__builtin_rvv_vget_v_i64m2_i64m1((vint64m2_t)(op0), (size_t)(op1)) +#define vget_v_i64m4_i64m1(op0, op1) \ +__builtin_rvv_vget_v_i64m4_i64m1((vint64m4_t)(op0), (size_t)(op1)) +#define vget_v_i64m8_i64m1(op0, op1) \ +__builtin_rvv_vget_v_i64m8_i64m1((vint64m8_t)(op0), (size_t)(op1)) +#define vget_v_u8m2_u8m1(op0, op1) \ +__builtin_rvv_vget_v_u8m2_u8m1((vuint8m2_t)(op0), (size_t)(op1)) +#define vget_v_u8m4_u8m1(op0, op1) \ +__builtin_rvv_vget_v_u8m4_u8m1((vuint8m4_t)(op0), (size_t)(op1)) +#define vget_v_u8m8_u8m1(op0, op1) \ +__builtin_rvv_vget_v_u8m8_u8m1((vuint8m8_t)(op0), (size_t)(op1)) +#define vget_v_u16m2_u16m1(op0, op1) \ +__builtin_rvv_vget_v_u16m2_u16m1((vuint16m2_t)(op0), (size_t)(op1)) +#define vget_v_u16m4_u16m1(op0, op1) \ +__builtin_rvv_vget_v_u16m4_u16m1((vuint16m4_t)(op0), (size_t)(op1)) +#define vget_v_u16m8_u16m1(op0, op1) \ +__builtin_rvv_vget_v_u16m8_u16m1((vuint16m8_t)(op0), (size_t)(op1)) +#define vget_v_u32m2_u32m1(op0, op1) \ +__builtin_rvv_vget_v_u32m2_u32m1((vuint32m2_t)(op0), (size_t)(op1)) +#define vget_v_u32m4_u32m1(op0, op1) \ +__builtin_rvv_vget_v_u32m4_u32m1((vuint32m4_t)(op0), (size_t)(op1)) +#define vget_v_u32m8_u32m1(op0, op1) \ +__builtin_rvv_vget_v_u32m8_u32m1((vuint32m8_t)(op0), (size_t)(op1)) +#define vget_v_u64m2_u64m1(op0, op1) \ +__builtin_rvv_vget_v_u64m2_u64m1((vuint64m2_t)(op0), (size_t)(op1)) +#define vget_v_u64m4_u64m1(op0, op1) \ +__builtin_rvv_vget_v_u64m4_u64m1((vuint64m4_t)(op0), (size_t)(op1)) +#define vget_v_u64m8_u64m1(op0, op1) \ +__builtin_rvv_vget_v_u64m8_u64m1((vuint64m8_t)(op0), (size_t)(op1)) +#define vget_v_i8m4_i8m2(op0, op1) \ +__builtin_rvv_vget_v_i8m4_i8m2((vint8m4_t)(op0), (size_t)(op1)) +#define vget_v_i8m8_i8m2(op0, op1) \ +__builtin_rvv_vget_v_i8m8_i8m2((vint8m8_t)(op0), (size_t)(op1)) +#define vget_v_i16m4_i16m2(op0, op1) \ +__builtin_rvv_vget_v_i16m4_i16m2((vint16m4_t)(op0), (size_t)(op1)) +#define vget_v_i16m8_i16m2(op0, op1) \ +__builtin_rvv_vget_v_i16m8_i16m2((vint16m8_t)(op0), (size_t)(op1)) +#define vget_v_i32m4_i32m2(op0, op1) \ +__builtin_rvv_vget_v_i32m4_i32m2((vint32m4_t)(op0), (size_t)(op1)) +#define vget_v_i32m8_i32m2(op0, op1) \ +__builtin_rvv_vget_v_i32m8_i32m2((vint32m8_t)(op0), (size_t)(op1)) +#define vget_v_i64m4_i64m2(op0, op1) \ +__builtin_rvv_vget_v_i64m4_i64m2((vint64m4_t)(op0), (size_t)(op1)) +#define vget_v_i64m8_i64m2(op0, op1) \ +__builtin_rvv_vget_v_i64m8_i64m2((vint64m8_t)(op0), (size_t)(op1)) +#define vget_v_u8m4_u8m2(op0, op1) \ +__builtin_rvv_vget_v_u8m4_u8m2((vuint8m4_t)(op0), (size_t)(op1)) +#define vget_v_u8m8_u8m2(op0, op1) \ +__builtin_rvv_vget_v_u8m8_u8m2((vuint8m8_t)(op0), (size_t)(op1)) +#define vget_v_u16m4_u16m2(op0, op1) \ +__builtin_rvv_vget_v_u16m4_u16m2((vuint16m4_t)(op0), (size_t)(op1)) +#define vget_v_u16m8_u16m2(op0, op1) \ +__builtin_rvv_vget_v_u16m8_u16m2((vuint16m8_t)(op0), (size_t)(op1)) +#define vget_v_u32m4_u32m2(op0, op1) \ +__builtin_rvv_vget_v_u32m4_u32m2((vuint32m4_t)(op0), (size_t)(op1)) +#define vget_v_u32m8_u32m2(op0, op1) \ +__builtin_rvv_vget_v_u32m8_u32m2((vuint32m8_t)(op0), (size_t)(op1)) +#define vget_v_u64m4_u64m2(op0, op1) \ +__builtin_rvv_vget_v_u64m4_u64m2((vuint64m4_t)(op0), (size_t)(op1)) +#define vget_v_u64m8_u64m2(op0, op1) \ +__builtin_rvv_vget_v_u64m8_u64m2((vuint64m8_t)(op0), (size_t)(op1)) +#define vget_v_i8m8_i8m4(op0, op1) \ +__builtin_rvv_vget_v_i8m8_i8m4((vint8m8_t)(op0), (size_t)(op1)) +#define vget_v_i16m8_i16m4(op0, op1) \ +__builtin_rvv_vget_v_i16m8_i16m4((vint16m8_t)(op0), (size_t)(op1)) +#define vget_v_i32m8_i32m4(op0, op1) \ +__builtin_rvv_vget_v_i32m8_i32m4((vint32m8_t)(op0), (size_t)(op1)) +#define vget_v_i64m8_i64m4(op0, op1) \ +__builtin_rvv_vget_v_i64m8_i64m4((vint64m8_t)(op0), (size_t)(op1)) +#define vget_v_u8m8_u8m4(op0, op1) \ +__builtin_rvv_vget_v_u8m8_u8m4((vuint8m8_t)(op0), (size_t)(op1)) +#define vget_v_u16m8_u16m4(op0, op1) \ +__builtin_rvv_vget_v_u16m8_u16m4((vuint16m8_t)(op0), (size_t)(op1)) +#define vget_v_u32m8_u32m4(op0, op1) \ +__builtin_rvv_vget_v_u32m8_u32m4((vuint32m8_t)(op0), (size_t)(op1)) +#define vget_v_u64m8_u64m4(op0, op1) \ +__builtin_rvv_vget_v_u64m8_u64m4((vuint64m8_t)(op0), (size_t)(op1)) +#define vset_v_i8m1_i8m2(op0, op1, op2) \ +__builtin_rvv_vset_v_i8m1_i8m2((vint8m2_t)(op0), (size_t)(op1), (vint8m1_t)(op2)) +#define vset_v_i16m1_i16m2(op0, op1, op2) \ +__builtin_rvv_vset_v_i16m1_i16m2((vint16m2_t)(op0), (size_t)(op1), (vint16m1_t)(op2)) +#define vset_v_i32m1_i32m2(op0, op1, op2) \ +__builtin_rvv_vset_v_i32m1_i32m2((vint32m2_t)(op0), (size_t)(op1), (vint32m1_t)(op2)) +#define vset_v_i64m1_i64m2(op0, op1, op2) \ +__builtin_rvv_vset_v_i64m1_i64m2((vint64m2_t)(op0), (size_t)(op1), (vint64m1_t)(op2)) +#define vset_v_u8m1_u8m2(op0, op1, op2) \ +__builtin_rvv_vset_v_u8m1_u8m2((vuint8m2_t)(op0), (size_t)(op1), (vuint8m1_t)(op2)) +#define vset_v_u16m1_u16m2(op0, op1, op2) \ +__builtin_rvv_vset_v_u16m1_u16m2((vuint16m2_t)(op0), (size_t)(op1), (vuint16m1_t)(op2)) +#define vset_v_u32m1_u32m2(op0, op1, op2) \ +__builtin_rvv_vset_v_u32m1_u32m2((vuint32m2_t)(op0), (size_t)(op1), (vuint32m1_t)(op2)) +#define vset_v_u64m1_u64m2(op0, op1, op2) \ +__builtin_rvv_vset_v_u64m1_u64m2((vuint64m2_t)(op0), (size_t)(op1), (vuint64m1_t)(op2)) +#define vset_v_i8m1_i8m4(op0, op1, op2) \ +__builtin_rvv_vset_v_i8m1_i8m4((vint8m4_t)(op0), (size_t)(op1), (vint8m1_t)(op2)) +#define vset_v_i8m2_i8m4(op0, op1, op2) \ +__builtin_rvv_vset_v_i8m2_i8m4((vint8m4_t)(op0), (size_t)(op1), (vint8m2_t)(op2)) +#define vset_v_i16m1_i16m4(op0, op1, op2) \ +__builtin_rvv_vset_v_i16m1_i16m4((vint16m4_t)(op0), (size_t)(op1), (vint16m1_t)(op2)) +#define vset_v_i16m2_i16m4(op0, op1, op2) \ +__builtin_rvv_vset_v_i16m2_i16m4((vint16m4_t)(op0), (size_t)(op1), (vint16m2_t)(op2)) +#define vset_v_i32m1_i32m4(op0, op1, op2) \ +__builtin_rvv_vset_v_i32m1_i32m4((vint32m4_t)(op0), (size_t)(op1), (vint32m1_t)(op2)) +#define vset_v_i32m2_i32m4(op0, op1, op2) \ +__builtin_rvv_vset_v_i32m2_i32m4((vint32m4_t)(op0), (size_t)(op1), (vint32m2_t)(op2)) +#define vset_v_i64m1_i64m4(op0, op1, op2) \ +__builtin_rvv_vset_v_i64m1_i64m4((vint64m4_t)(op0), (size_t)(op1), (vint64m1_t)(op2)) +#define vset_v_i64m2_i64m4(op0, op1, op2) \ +__builtin_rvv_vset_v_i64m2_i64m4((vint64m4_t)(op0), (size_t)(op1), (vint64m2_t)(op2)) +#define vset_v_u8m1_u8m4(op0, op1, op2) \ +__builtin_rvv_vset_v_u8m1_u8m4((vuint8m4_t)(op0), (size_t)(op1), (vuint8m1_t)(op2)) +#define vset_v_u8m2_u8m4(op0, op1, op2) \ +__builtin_rvv_vset_v_u8m2_u8m4((vuint8m4_t)(op0), (size_t)(op1), (vuint8m2_t)(op2)) +#define vset_v_u16m1_u16m4(op0, op1, op2) \ +__builtin_rvv_vset_v_u16m1_u16m4((vuint16m4_t)(op0), (size_t)(op1), (vuint16m1_t)(op2)) +#define vset_v_u16m2_u16m4(op0, op1, op2) \ +__builtin_rvv_vset_v_u16m2_u16m4((vuint16m4_t)(op0), (size_t)(op1), (vuint16m2_t)(op2)) +#define vset_v_u32m1_u32m4(op0, op1, op2) \ +__builtin_rvv_vset_v_u32m1_u32m4((vuint32m4_t)(op0), (size_t)(op1), (vuint32m1_t)(op2)) +#define vset_v_u32m2_u32m4(op0, op1, op2) \ +__builtin_rvv_vset_v_u32m2_u32m4((vuint32m4_t)(op0), (size_t)(op1), (vuint32m2_t)(op2)) +#define vset_v_u64m1_u64m4(op0, op1, op2) \ +__builtin_rvv_vset_v_u64m1_u64m4((vuint64m4_t)(op0), (size_t)(op1), (vuint64m1_t)(op2)) +#define vset_v_u64m2_u64m4(op0, op1, op2) \ +__builtin_rvv_vset_v_u64m2_u64m4((vuint64m4_t)(op0), (size_t)(op1), (vuint64m2_t)(op2)) +#define vset_v_i8m1_i8m8(op0, op1, op2) \ +__builtin_rvv_vset_v_i8m1_i8m8((vint8m8_t)(op0), (size_t)(op1), (vint8m1_t)(op2)) +#define vset_v_i8m2_i8m8(op0, op1, op2) \ +__builtin_rvv_vset_v_i8m2_i8m8((vint8m8_t)(op0), (size_t)(op1), (vint8m2_t)(op2)) +#define vset_v_i8m4_i8m8(op0, op1, op2) \ +__builtin_rvv_vset_v_i8m4_i8m8((vint8m8_t)(op0), (size_t)(op1), (vint8m4_t)(op2)) +#define vset_v_i16m1_i16m8(op0, op1, op2) \ +__builtin_rvv_vset_v_i16m1_i16m8((vint16m8_t)(op0), (size_t)(op1), (vint16m1_t)(op2)) +#define vset_v_i16m2_i16m8(op0, op1, op2) \ +__builtin_rvv_vset_v_i16m2_i16m8((vint16m8_t)(op0), (size_t)(op1), (vint16m2_t)(op2)) +#define vset_v_i16m4_i16m8(op0, op1, op2) \ +__builtin_rvv_vset_v_i16m4_i16m8((vint16m8_t)(op0), (size_t)(op1), (vint16m4_t)(op2)) +#define vset_v_i32m1_i32m8(op0, op1, op2) \ +__builtin_rvv_vset_v_i32m1_i32m8((vint32m8_t)(op0), (size_t)(op1), (vint32m1_t)(op2)) +#define vset_v_i32m2_i32m8(op0, op1, op2) \ +__builtin_rvv_vset_v_i32m2_i32m8((vint32m8_t)(op0), (size_t)(op1), (vint32m2_t)(op2)) +#define vset_v_i32m4_i32m8(op0, op1, op2) \ +__builtin_rvv_vset_v_i32m4_i32m8((vint32m8_t)(op0), (size_t)(op1), (vint32m4_t)(op2)) +#define vset_v_i64m1_i64m8(op0, op1, op2) \ +__builtin_rvv_vset_v_i64m1_i64m8((vint64m8_t)(op0), (size_t)(op1), (vint64m1_t)(op2)) +#define vset_v_i64m2_i64m8(op0, op1, op2) \ +__builtin_rvv_vset_v_i64m2_i64m8((vint64m8_t)(op0), (size_t)(op1), (vint64m2_t)(op2)) +#define vset_v_i64m4_i64m8(op0, op1, op2) \ +__builtin_rvv_vset_v_i64m4_i64m8((vint64m8_t)(op0), (size_t)(op1), (vint64m4_t)(op2)) +#define vset_v_u8m1_u8m8(op0, op1, op2) \ +__builtin_rvv_vset_v_u8m1_u8m8((vuint8m8_t)(op0), (size_t)(op1), (vuint8m1_t)(op2)) +#define vset_v_u8m2_u8m8(op0, op1, op2) \ +__builtin_rvv_vset_v_u8m2_u8m8((vuint8m8_t)(op0), (size_t)(op1), (vuint8m2_t)(op2)) +#define vset_v_u8m4_u8m8(op0, op1, op2) \ +__builtin_rvv_vset_v_u8m4_u8m8((vuint8m8_t)(op0), (size_t)(op1), (vuint8m4_t)(op2)) +#define vset_v_u16m1_u16m8(op0, op1, op2) \ +__builtin_rvv_vset_v_u16m1_u16m8((vuint16m8_t)(op0), (size_t)(op1), (vuint16m1_t)(op2)) +#define vset_v_u16m2_u16m8(op0, op1, op2) \ +__builtin_rvv_vset_v_u16m2_u16m8((vuint16m8_t)(op0), (size_t)(op1), (vuint16m2_t)(op2)) +#define vset_v_u16m4_u16m8(op0, op1, op2) \ +__builtin_rvv_vset_v_u16m4_u16m8((vuint16m8_t)(op0), (size_t)(op1), (vuint16m4_t)(op2)) +#define vset_v_u32m1_u32m8(op0, op1, op2) \ +__builtin_rvv_vset_v_u32m1_u32m8((vuint32m8_t)(op0), (size_t)(op1), (vuint32m1_t)(op2)) +#define vset_v_u32m2_u32m8(op0, op1, op2) \ +__builtin_rvv_vset_v_u32m2_u32m8((vuint32m8_t)(op0), (size_t)(op1), (vuint32m2_t)(op2)) +#define vset_v_u32m4_u32m8(op0, op1, op2) \ +__builtin_rvv_vset_v_u32m4_u32m8((vuint32m8_t)(op0), (size_t)(op1), (vuint32m4_t)(op2)) +#define vset_v_u64m1_u64m8(op0, op1, op2) \ +__builtin_rvv_vset_v_u64m1_u64m8((vuint64m8_t)(op0), (size_t)(op1), (vuint64m1_t)(op2)) +#define vset_v_u64m2_u64m8(op0, op1, op2) \ +__builtin_rvv_vset_v_u64m2_u64m8((vuint64m8_t)(op0), (size_t)(op1), (vuint64m2_t)(op2)) +#define vset_v_u64m4_u64m8(op0, op1, op2) \ +__builtin_rvv_vset_v_u64m4_u64m8((vuint64m8_t)(op0), (size_t)(op1), (vuint64m4_t)(op2)) +#define vsoxei8_v_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_i8m1((int8_t *)(op0), (vuint8m1_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vsoxei8_v_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vsoxei8_v_i8m2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_i8m2((int8_t *)(op0), (vuint8m2_t)(op1), (vint8m2_t)(op2), (size_t)(op3)) +#define vsoxei8_v_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_i8m2_m((vbool4_t)(op0), (int8_t *)(op1), (vuint8m2_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vsoxei8_v_i8m4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_i8m4((int8_t *)(op0), (vuint8m4_t)(op1), (vint8m4_t)(op2), (size_t)(op3)) +#define vsoxei8_v_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_i8m4_m((vbool2_t)(op0), (int8_t *)(op1), (vuint8m4_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vsoxei8_v_i8m8(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_i8m8((int8_t *)(op0), (vuint8m8_t)(op1), (vint8m8_t)(op2), (size_t)(op3)) +#define vsoxei8_v_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_i8m8_m((vbool1_t)(op0), (int8_t *)(op1), (vuint8m8_t)(op2), (vint8m8_t)(op3), (size_t)(op4)) +#define vsoxei8_v_i8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_i8mf2((int8_t *)(op0), (vuint8mf2_t)(op1), (vint8mf2_t)(op2), (size_t)(op3)) +#define vsoxei8_v_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint8mf2_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vsoxei8_v_i8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_i8mf4((int8_t *)(op0), (vuint8mf4_t)(op1), (vint8mf4_t)(op2), (size_t)(op3)) +#define vsoxei8_v_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint8mf4_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vsoxei8_v_i8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_i8mf8((int8_t *)(op0), (vuint8mf8_t)(op1), (vint8mf8_t)(op2), (size_t)(op3)) +#define vsoxei8_v_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint8mf8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vsoxei8_v_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_u8m1((uint8_t *)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vsoxei8_v_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vsoxei8_v_u8m2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_u8m2((uint8_t *)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (size_t)(op3)) +#define vsoxei8_v_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_u8m2_m((vbool4_t)(op0), (uint8_t *)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vsoxei8_v_u8m4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_u8m4((uint8_t *)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (size_t)(op3)) +#define vsoxei8_v_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_u8m4_m((vbool2_t)(op0), (uint8_t *)(op1), (vuint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vsoxei8_v_u8m8(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_u8m8((uint8_t *)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (size_t)(op3)) +#define vsoxei8_v_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_u8m8_m((vbool1_t)(op0), (uint8_t *)(op1), (vuint8m8_t)(op2), (vuint8m8_t)(op3), (size_t)(op4)) +#define vsoxei8_v_u8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_u8mf2((uint8_t *)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (size_t)(op3)) +#define vsoxei8_v_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vsoxei8_v_u8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_u8mf4((uint8_t *)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (size_t)(op3)) +#define vsoxei8_v_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vsoxei8_v_u8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_u8mf8((uint8_t *)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (size_t)(op3)) +#define vsoxei8_v_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vsoxei16_v_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_i8m1((int8_t *)(op0), (vuint16m2_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vsoxei16_v_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint16m2_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vsoxei16_v_i8m2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_i8m2((int8_t *)(op0), (vuint16m4_t)(op1), (vint8m2_t)(op2), (size_t)(op3)) +#define vsoxei16_v_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_i8m2_m((vbool4_t)(op0), (int8_t *)(op1), (vuint16m4_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vsoxei16_v_i8m4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_i8m4((int8_t *)(op0), (vuint16m8_t)(op1), (vint8m4_t)(op2), (size_t)(op3)) +#define vsoxei16_v_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_i8m4_m((vbool2_t)(op0), (int8_t *)(op1), (vuint16m8_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vsoxei16_v_i8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_i8mf2((int8_t *)(op0), (vuint16m1_t)(op1), (vint8mf2_t)(op2), (size_t)(op3)) +#define vsoxei16_v_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint16m1_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vsoxei16_v_i8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_i8mf4((int8_t *)(op0), (vuint16mf2_t)(op1), (vint8mf4_t)(op2), (size_t)(op3)) +#define vsoxei16_v_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint16mf2_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vsoxei16_v_i8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_i8mf8((int8_t *)(op0), (vuint16mf4_t)(op1), (vint8mf8_t)(op2), (size_t)(op3)) +#define vsoxei16_v_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint16mf4_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vsoxei16_v_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_u8m1((uint8_t *)(op0), (vuint16m2_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vsoxei16_v_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint16m2_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vsoxei16_v_u8m2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_u8m2((uint8_t *)(op0), (vuint16m4_t)(op1), (vuint8m2_t)(op2), (size_t)(op3)) +#define vsoxei16_v_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_u8m2_m((vbool4_t)(op0), (uint8_t *)(op1), (vuint16m4_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vsoxei16_v_u8m4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_u8m4((uint8_t *)(op0), (vuint16m8_t)(op1), (vuint8m4_t)(op2), (size_t)(op3)) +#define vsoxei16_v_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_u8m4_m((vbool2_t)(op0), (uint8_t *)(op1), (vuint16m8_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vsoxei16_v_u8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_u8mf2((uint8_t *)(op0), (vuint16m1_t)(op1), (vuint8mf2_t)(op2), (size_t)(op3)) +#define vsoxei16_v_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint16m1_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vsoxei16_v_u8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_u8mf4((uint8_t *)(op0), (vuint16mf2_t)(op1), (vuint8mf4_t)(op2), (size_t)(op3)) +#define vsoxei16_v_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint16mf2_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vsoxei16_v_u8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_u8mf8((uint8_t *)(op0), (vuint16mf4_t)(op1), (vuint8mf8_t)(op2), (size_t)(op3)) +#define vsoxei16_v_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint16mf4_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vsoxei32_v_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_i8m1((int8_t *)(op0), (vuint32m4_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vsoxei32_v_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint32m4_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vsoxei32_v_i8m2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_i8m2((int8_t *)(op0), (vuint32m8_t)(op1), (vint8m2_t)(op2), (size_t)(op3)) +#define vsoxei32_v_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_i8m2_m((vbool4_t)(op0), (int8_t *)(op1), (vuint32m8_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vsoxei32_v_i8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_i8mf2((int8_t *)(op0), (vuint32m2_t)(op1), (vint8mf2_t)(op2), (size_t)(op3)) +#define vsoxei32_v_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint32m2_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vsoxei32_v_i8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_i8mf4((int8_t *)(op0), (vuint32m1_t)(op1), (vint8mf4_t)(op2), (size_t)(op3)) +#define vsoxei32_v_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint32m1_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vsoxei32_v_i8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_i8mf8((int8_t *)(op0), (vuint32mf2_t)(op1), (vint8mf8_t)(op2), (size_t)(op3)) +#define vsoxei32_v_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint32mf2_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vsoxei32_v_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_u8m1((uint8_t *)(op0), (vuint32m4_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vsoxei32_v_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint32m4_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vsoxei32_v_u8m2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_u8m2((uint8_t *)(op0), (vuint32m8_t)(op1), (vuint8m2_t)(op2), (size_t)(op3)) +#define vsoxei32_v_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_u8m2_m((vbool4_t)(op0), (uint8_t *)(op1), (vuint32m8_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vsoxei32_v_u8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_u8mf2((uint8_t *)(op0), (vuint32m2_t)(op1), (vuint8mf2_t)(op2), (size_t)(op3)) +#define vsoxei32_v_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint32m2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vsoxei32_v_u8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_u8mf4((uint8_t *)(op0), (vuint32m1_t)(op1), (vuint8mf4_t)(op2), (size_t)(op3)) +#define vsoxei32_v_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint32m1_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vsoxei32_v_u8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_u8mf8((uint8_t *)(op0), (vuint32mf2_t)(op1), (vuint8mf8_t)(op2), (size_t)(op3)) +#define vsoxei32_v_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint32mf2_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vsoxei64_v_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei64_v_i8m1((int8_t *)(op0), (vuint64m8_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vsoxei64_v_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei64_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint64m8_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vsoxei64_v_i8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei64_v_i8mf2((int8_t *)(op0), (vuint64m4_t)(op1), (vint8mf2_t)(op2), (size_t)(op3)) +#define vsoxei64_v_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei64_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint64m4_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vsoxei64_v_i8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei64_v_i8mf4((int8_t *)(op0), (vuint64m2_t)(op1), (vint8mf4_t)(op2), (size_t)(op3)) +#define vsoxei64_v_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei64_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint64m2_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vsoxei64_v_i8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei64_v_i8mf8((int8_t *)(op0), (vuint64m1_t)(op1), (vint8mf8_t)(op2), (size_t)(op3)) +#define vsoxei64_v_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei64_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint64m1_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vsoxei64_v_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei64_v_u8m1((uint8_t *)(op0), (vuint64m8_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vsoxei64_v_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei64_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint64m8_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vsoxei64_v_u8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei64_v_u8mf2((uint8_t *)(op0), (vuint64m4_t)(op1), (vuint8mf2_t)(op2), (size_t)(op3)) +#define vsoxei64_v_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei64_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint64m4_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vsoxei64_v_u8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei64_v_u8mf4((uint8_t *)(op0), (vuint64m2_t)(op1), (vuint8mf4_t)(op2), (size_t)(op3)) +#define vsoxei64_v_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei64_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint64m2_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vsoxei64_v_u8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei64_v_u8mf8((uint8_t *)(op0), (vuint64m1_t)(op1), (vuint8mf8_t)(op2), (size_t)(op3)) +#define vsoxei64_v_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei64_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint64m1_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vsoxei8_v_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_i16m1((int16_t *)(op0), (vuint8mf2_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vsoxei8_v_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint8mf2_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vsoxei8_v_i16m2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_i16m2((int16_t *)(op0), (vuint8m1_t)(op1), (vint16m2_t)(op2), (size_t)(op3)) +#define vsoxei8_v_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_i16m2_m((vbool8_t)(op0), (int16_t *)(op1), (vuint8m1_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vsoxei8_v_i16m4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_i16m4((int16_t *)(op0), (vuint8m2_t)(op1), (vint16m4_t)(op2), (size_t)(op3)) +#define vsoxei8_v_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_i16m4_m((vbool4_t)(op0), (int16_t *)(op1), (vuint8m2_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vsoxei8_v_i16m8(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_i16m8((int16_t *)(op0), (vuint8m4_t)(op1), (vint16m8_t)(op2), (size_t)(op3)) +#define vsoxei8_v_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_i16m8_m((vbool2_t)(op0), (int16_t *)(op1), (vuint8m4_t)(op2), (vint16m8_t)(op3), (size_t)(op4)) +#define vsoxei8_v_i16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_i16mf2((int16_t *)(op0), (vuint8mf4_t)(op1), (vint16mf2_t)(op2), (size_t)(op3)) +#define vsoxei8_v_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint8mf4_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vsoxei8_v_i16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_i16mf4((int16_t *)(op0), (vuint8mf8_t)(op1), (vint16mf4_t)(op2), (size_t)(op3)) +#define vsoxei8_v_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint8mf8_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vsoxei8_v_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_u16m1((uint16_t *)(op0), (vuint8mf2_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vsoxei8_v_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint8mf2_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vsoxei8_v_u16m2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_u16m2((uint16_t *)(op0), (vuint8m1_t)(op1), (vuint16m2_t)(op2), (size_t)(op3)) +#define vsoxei8_v_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_u16m2_m((vbool8_t)(op0), (uint16_t *)(op1), (vuint8m1_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vsoxei8_v_u16m4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_u16m4((uint16_t *)(op0), (vuint8m2_t)(op1), (vuint16m4_t)(op2), (size_t)(op3)) +#define vsoxei8_v_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_u16m4_m((vbool4_t)(op0), (uint16_t *)(op1), (vuint8m2_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vsoxei8_v_u16m8(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_u16m8((uint16_t *)(op0), (vuint8m4_t)(op1), (vuint16m8_t)(op2), (size_t)(op3)) +#define vsoxei8_v_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_u16m8_m((vbool2_t)(op0), (uint16_t *)(op1), (vuint8m4_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vsoxei8_v_u16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_u16mf2((uint16_t *)(op0), (vuint8mf4_t)(op1), (vuint16mf2_t)(op2), (size_t)(op3)) +#define vsoxei8_v_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint8mf4_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vsoxei8_v_u16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_u16mf4((uint16_t *)(op0), (vuint8mf8_t)(op1), (vuint16mf4_t)(op2), (size_t)(op3)) +#define vsoxei8_v_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint8mf8_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vsoxei16_v_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_i16m1((int16_t *)(op0), (vuint16m1_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vsoxei16_v_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vsoxei16_v_i16m2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_i16m2((int16_t *)(op0), (vuint16m2_t)(op1), (vint16m2_t)(op2), (size_t)(op3)) +#define vsoxei16_v_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_i16m2_m((vbool8_t)(op0), (int16_t *)(op1), (vuint16m2_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vsoxei16_v_i16m4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_i16m4((int16_t *)(op0), (vuint16m4_t)(op1), (vint16m4_t)(op2), (size_t)(op3)) +#define vsoxei16_v_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_i16m4_m((vbool4_t)(op0), (int16_t *)(op1), (vuint16m4_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vsoxei16_v_i16m8(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_i16m8((int16_t *)(op0), (vuint16m8_t)(op1), (vint16m8_t)(op2), (size_t)(op3)) +#define vsoxei16_v_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_i16m8_m((vbool2_t)(op0), (int16_t *)(op1), (vuint16m8_t)(op2), (vint16m8_t)(op3), (size_t)(op4)) +#define vsoxei16_v_i16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_i16mf2((int16_t *)(op0), (vuint16mf2_t)(op1), (vint16mf2_t)(op2), (size_t)(op3)) +#define vsoxei16_v_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint16mf2_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vsoxei16_v_i16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_i16mf4((int16_t *)(op0), (vuint16mf4_t)(op1), (vint16mf4_t)(op2), (size_t)(op3)) +#define vsoxei16_v_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint16mf4_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vsoxei16_v_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_u16m1((uint16_t *)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vsoxei16_v_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vsoxei16_v_u16m2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_u16m2((uint16_t *)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (size_t)(op3)) +#define vsoxei16_v_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_u16m2_m((vbool8_t)(op0), (uint16_t *)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vsoxei16_v_u16m4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_u16m4((uint16_t *)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (size_t)(op3)) +#define vsoxei16_v_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_u16m4_m((vbool4_t)(op0), (uint16_t *)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vsoxei16_v_u16m8(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_u16m8((uint16_t *)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (size_t)(op3)) +#define vsoxei16_v_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_u16m8_m((vbool2_t)(op0), (uint16_t *)(op1), (vuint16m8_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vsoxei16_v_u16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_u16mf2((uint16_t *)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (size_t)(op3)) +#define vsoxei16_v_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vsoxei16_v_u16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_u16mf4((uint16_t *)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (size_t)(op3)) +#define vsoxei16_v_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vsoxei32_v_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_i16m1((int16_t *)(op0), (vuint32m2_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vsoxei32_v_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint32m2_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vsoxei32_v_i16m2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_i16m2((int16_t *)(op0), (vuint32m4_t)(op1), (vint16m2_t)(op2), (size_t)(op3)) +#define vsoxei32_v_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_i16m2_m((vbool8_t)(op0), (int16_t *)(op1), (vuint32m4_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vsoxei32_v_i16m4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_i16m4((int16_t *)(op0), (vuint32m8_t)(op1), (vint16m4_t)(op2), (size_t)(op3)) +#define vsoxei32_v_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_i16m4_m((vbool4_t)(op0), (int16_t *)(op1), (vuint32m8_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vsoxei32_v_i16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_i16mf2((int16_t *)(op0), (vuint32m1_t)(op1), (vint16mf2_t)(op2), (size_t)(op3)) +#define vsoxei32_v_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint32m1_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vsoxei32_v_i16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_i16mf4((int16_t *)(op0), (vuint32mf2_t)(op1), (vint16mf4_t)(op2), (size_t)(op3)) +#define vsoxei32_v_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint32mf2_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vsoxei32_v_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_u16m1((uint16_t *)(op0), (vuint32m2_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vsoxei32_v_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint32m2_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vsoxei32_v_u16m2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_u16m2((uint16_t *)(op0), (vuint32m4_t)(op1), (vuint16m2_t)(op2), (size_t)(op3)) +#define vsoxei32_v_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_u16m2_m((vbool8_t)(op0), (uint16_t *)(op1), (vuint32m4_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vsoxei32_v_u16m4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_u16m4((uint16_t *)(op0), (vuint32m8_t)(op1), (vuint16m4_t)(op2), (size_t)(op3)) +#define vsoxei32_v_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_u16m4_m((vbool4_t)(op0), (uint16_t *)(op1), (vuint32m8_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vsoxei32_v_u16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_u16mf2((uint16_t *)(op0), (vuint32m1_t)(op1), (vuint16mf2_t)(op2), (size_t)(op3)) +#define vsoxei32_v_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint32m1_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vsoxei32_v_u16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_u16mf4((uint16_t *)(op0), (vuint32mf2_t)(op1), (vuint16mf4_t)(op2), (size_t)(op3)) +#define vsoxei32_v_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint32mf2_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vsoxei64_v_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei64_v_i16m1((int16_t *)(op0), (vuint64m4_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vsoxei64_v_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei64_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint64m4_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vsoxei64_v_i16m2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei64_v_i16m2((int16_t *)(op0), (vuint64m8_t)(op1), (vint16m2_t)(op2), (size_t)(op3)) +#define vsoxei64_v_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei64_v_i16m2_m((vbool8_t)(op0), (int16_t *)(op1), (vuint64m8_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vsoxei64_v_i16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei64_v_i16mf2((int16_t *)(op0), (vuint64m2_t)(op1), (vint16mf2_t)(op2), (size_t)(op3)) +#define vsoxei64_v_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei64_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint64m2_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vsoxei64_v_i16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei64_v_i16mf4((int16_t *)(op0), (vuint64m1_t)(op1), (vint16mf4_t)(op2), (size_t)(op3)) +#define vsoxei64_v_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei64_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint64m1_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vsoxei64_v_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei64_v_u16m1((uint16_t *)(op0), (vuint64m4_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vsoxei64_v_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei64_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint64m4_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vsoxei64_v_u16m2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei64_v_u16m2((uint16_t *)(op0), (vuint64m8_t)(op1), (vuint16m2_t)(op2), (size_t)(op3)) +#define vsoxei64_v_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei64_v_u16m2_m((vbool8_t)(op0), (uint16_t *)(op1), (vuint64m8_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vsoxei64_v_u16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei64_v_u16mf2((uint16_t *)(op0), (vuint64m2_t)(op1), (vuint16mf2_t)(op2), (size_t)(op3)) +#define vsoxei64_v_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei64_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint64m2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vsoxei64_v_u16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei64_v_u16mf4((uint16_t *)(op0), (vuint64m1_t)(op1), (vuint16mf4_t)(op2), (size_t)(op3)) +#define vsoxei64_v_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei64_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint64m1_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vsoxei8_v_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_i32m1((int32_t *)(op0), (vuint8mf4_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vsoxei8_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint8mf4_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vsoxei8_v_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_i32m2((int32_t *)(op0), (vuint8mf2_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vsoxei8_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint8mf2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vsoxei8_v_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_i32m4((int32_t *)(op0), (vuint8m1_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vsoxei8_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vuint8m1_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vsoxei8_v_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_i32m8((int32_t *)(op0), (vuint8m2_t)(op1), (vint32m8_t)(op2), (size_t)(op3)) +#define vsoxei8_v_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_i32m8_m((vbool4_t)(op0), (int32_t *)(op1), (vuint8m2_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vsoxei8_v_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_i32mf2((int32_t *)(op0), (vuint8mf8_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vsoxei8_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint8mf8_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vsoxei8_v_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_u32m1((uint32_t *)(op0), (vuint8mf4_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vsoxei8_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint8mf4_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vsoxei8_v_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_u32m2((uint32_t *)(op0), (vuint8mf2_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vsoxei8_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint8mf2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vsoxei8_v_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_u32m4((uint32_t *)(op0), (vuint8m1_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vsoxei8_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint8m1_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vsoxei8_v_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_u32m8((uint32_t *)(op0), (vuint8m2_t)(op1), (vuint32m8_t)(op2), (size_t)(op3)) +#define vsoxei8_v_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_u32m8_m((vbool4_t)(op0), (uint32_t *)(op1), (vuint8m2_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vsoxei8_v_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_u32mf2((uint32_t *)(op0), (vuint8mf8_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vsoxei8_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint8mf8_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vsoxei16_v_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_i32m1((int32_t *)(op0), (vuint16mf2_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vsoxei16_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint16mf2_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vsoxei16_v_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_i32m2((int32_t *)(op0), (vuint16m1_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vsoxei16_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint16m1_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vsoxei16_v_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_i32m4((int32_t *)(op0), (vuint16m2_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vsoxei16_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vuint16m2_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vsoxei16_v_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_i32m8((int32_t *)(op0), (vuint16m4_t)(op1), (vint32m8_t)(op2), (size_t)(op3)) +#define vsoxei16_v_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_i32m8_m((vbool4_t)(op0), (int32_t *)(op1), (vuint16m4_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vsoxei16_v_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_i32mf2((int32_t *)(op0), (vuint16mf4_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vsoxei16_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint16mf4_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vsoxei16_v_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_u32m1((uint32_t *)(op0), (vuint16mf2_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vsoxei16_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint16mf2_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vsoxei16_v_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_u32m2((uint32_t *)(op0), (vuint16m1_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vsoxei16_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint16m1_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vsoxei16_v_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_u32m4((uint32_t *)(op0), (vuint16m2_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vsoxei16_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint16m2_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vsoxei16_v_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_u32m8((uint32_t *)(op0), (vuint16m4_t)(op1), (vuint32m8_t)(op2), (size_t)(op3)) +#define vsoxei16_v_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_u32m8_m((vbool4_t)(op0), (uint32_t *)(op1), (vuint16m4_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vsoxei16_v_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_u32mf2((uint32_t *)(op0), (vuint16mf4_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vsoxei16_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint16mf4_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vsoxei32_v_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_i32m1((int32_t *)(op0), (vuint32m1_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vsoxei32_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vsoxei32_v_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_i32m2((int32_t *)(op0), (vuint32m2_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vsoxei32_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vsoxei32_v_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_i32m4((int32_t *)(op0), (vuint32m4_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vsoxei32_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vuint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vsoxei32_v_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_i32m8((int32_t *)(op0), (vuint32m8_t)(op1), (vint32m8_t)(op2), (size_t)(op3)) +#define vsoxei32_v_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_i32m8_m((vbool4_t)(op0), (int32_t *)(op1), (vuint32m8_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vsoxei32_v_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_i32mf2((int32_t *)(op0), (vuint32mf2_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vsoxei32_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vsoxei32_v_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_u32m1((uint32_t *)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vsoxei32_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vsoxei32_v_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_u32m2((uint32_t *)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vsoxei32_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vsoxei32_v_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_u32m4((uint32_t *)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vsoxei32_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vsoxei32_v_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_u32m8((uint32_t *)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (size_t)(op3)) +#define vsoxei32_v_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_u32m8_m((vbool4_t)(op0), (uint32_t *)(op1), (vuint32m8_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vsoxei32_v_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_u32mf2((uint32_t *)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vsoxei32_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vsoxei64_v_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei64_v_i32m1((int32_t *)(op0), (vuint64m2_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vsoxei64_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei64_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint64m2_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vsoxei64_v_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei64_v_i32m2((int32_t *)(op0), (vuint64m4_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vsoxei64_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei64_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint64m4_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vsoxei64_v_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei64_v_i32m4((int32_t *)(op0), (vuint64m8_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vsoxei64_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei64_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vuint64m8_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vsoxei64_v_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei64_v_i32mf2((int32_t *)(op0), (vuint64m1_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vsoxei64_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei64_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint64m1_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vsoxei64_v_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei64_v_u32m1((uint32_t *)(op0), (vuint64m2_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vsoxei64_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei64_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint64m2_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vsoxei64_v_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei64_v_u32m2((uint32_t *)(op0), (vuint64m4_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vsoxei64_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei64_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint64m4_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vsoxei64_v_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei64_v_u32m4((uint32_t *)(op0), (vuint64m8_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vsoxei64_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei64_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint64m8_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vsoxei64_v_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei64_v_u32mf2((uint32_t *)(op0), (vuint64m1_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vsoxei64_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei64_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint64m1_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vsoxei8_v_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_i64m1((int64_t *)(op0), (vuint8mf8_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vsoxei8_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint8mf8_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vsoxei8_v_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_i64m2((int64_t *)(op0), (vuint8mf4_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vsoxei8_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint8mf4_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vsoxei8_v_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_i64m4((int64_t *)(op0), (vuint8mf2_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vsoxei8_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vuint8mf2_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vsoxei8_v_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_i64m8((int64_t *)(op0), (vuint8m1_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vsoxei8_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_i64m8_m((vbool8_t)(op0), (int64_t *)(op1), (vuint8m1_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vsoxei8_v_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_u64m1((uint64_t *)(op0), (vuint8mf8_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vsoxei8_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint8mf8_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vsoxei8_v_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_u64m2((uint64_t *)(op0), (vuint8mf4_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vsoxei8_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint8mf4_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vsoxei8_v_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_u64m4((uint64_t *)(op0), (vuint8mf2_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vsoxei8_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint8mf2_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vsoxei8_v_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_u64m8((uint64_t *)(op0), (vuint8m1_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vsoxei8_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_u64m8_m((vbool8_t)(op0), (uint64_t *)(op1), (vuint8m1_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vsoxei16_v_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_i64m1((int64_t *)(op0), (vuint16mf4_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vsoxei16_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint16mf4_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vsoxei16_v_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_i64m2((int64_t *)(op0), (vuint16mf2_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vsoxei16_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint16mf2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vsoxei16_v_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_i64m4((int64_t *)(op0), (vuint16m1_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vsoxei16_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vuint16m1_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vsoxei16_v_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_i64m8((int64_t *)(op0), (vuint16m2_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vsoxei16_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_i64m8_m((vbool8_t)(op0), (int64_t *)(op1), (vuint16m2_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vsoxei16_v_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_u64m1((uint64_t *)(op0), (vuint16mf4_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vsoxei16_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint16mf4_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vsoxei16_v_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_u64m2((uint64_t *)(op0), (vuint16mf2_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vsoxei16_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint16mf2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vsoxei16_v_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_u64m4((uint64_t *)(op0), (vuint16m1_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vsoxei16_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint16m1_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vsoxei16_v_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_u64m8((uint64_t *)(op0), (vuint16m2_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vsoxei16_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_u64m8_m((vbool8_t)(op0), (uint64_t *)(op1), (vuint16m2_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vsoxei32_v_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_i64m1((int64_t *)(op0), (vuint32mf2_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vsoxei32_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint32mf2_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vsoxei32_v_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_i64m2((int64_t *)(op0), (vuint32m1_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vsoxei32_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint32m1_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vsoxei32_v_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_i64m4((int64_t *)(op0), (vuint32m2_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vsoxei32_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vuint32m2_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vsoxei32_v_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_i64m8((int64_t *)(op0), (vuint32m4_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vsoxei32_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_i64m8_m((vbool8_t)(op0), (int64_t *)(op1), (vuint32m4_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vsoxei32_v_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_u64m1((uint64_t *)(op0), (vuint32mf2_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vsoxei32_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint32mf2_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vsoxei32_v_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_u64m2((uint64_t *)(op0), (vuint32m1_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vsoxei32_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint32m1_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vsoxei32_v_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_u64m4((uint64_t *)(op0), (vuint32m2_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vsoxei32_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint32m2_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vsoxei32_v_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_u64m8((uint64_t *)(op0), (vuint32m4_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vsoxei32_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_u64m8_m((vbool8_t)(op0), (uint64_t *)(op1), (vuint32m4_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vsoxei64_v_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei64_v_i64m1((int64_t *)(op0), (vuint64m1_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vsoxei64_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei64_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vsoxei64_v_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei64_v_i64m2((int64_t *)(op0), (vuint64m2_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vsoxei64_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei64_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vsoxei64_v_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei64_v_i64m4((int64_t *)(op0), (vuint64m4_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vsoxei64_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei64_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vuint64m4_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vsoxei64_v_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei64_v_i64m8((int64_t *)(op0), (vuint64m8_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vsoxei64_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei64_v_i64m8_m((vbool8_t)(op0), (int64_t *)(op1), (vuint64m8_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vsoxei64_v_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei64_v_u64m1((uint64_t *)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vsoxei64_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei64_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vsoxei64_v_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei64_v_u64m2((uint64_t *)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vsoxei64_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei64_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vsoxei64_v_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei64_v_u64m4((uint64_t *)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vsoxei64_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei64_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vsoxei64_v_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei64_v_u64m8((uint64_t *)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vsoxei64_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei64_v_u64m8_m((vbool8_t)(op0), (uint64_t *)(op1), (vuint64m8_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vle16ff_v_i16m1(op0, op1, op2) \ +__builtin_rvv_vle16ff_v_i16m1((const int16_t *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle16ff_v_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle16ff_v_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (const int16_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle16ff_v_i16m2(op0, op1, op2) \ +__builtin_rvv_vle16ff_v_i16m2((const int16_t *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle16ff_v_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle16ff_v_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (const int16_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle16ff_v_i16m4(op0, op1, op2) \ +__builtin_rvv_vle16ff_v_i16m4((const int16_t *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle16ff_v_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle16ff_v_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (const int16_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle16ff_v_i16m8(op0, op1, op2) \ +__builtin_rvv_vle16ff_v_i16m8((const int16_t *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle16ff_v_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle16ff_v_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (const int16_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle16ff_v_i16mf2(op0, op1, op2) \ +__builtin_rvv_vle16ff_v_i16mf2((const int16_t *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle16ff_v_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle16ff_v_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (const int16_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle16ff_v_i16mf4(op0, op1, op2) \ +__builtin_rvv_vle16ff_v_i16mf4((const int16_t *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle16ff_v_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle16ff_v_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (const int16_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle16ff_v_u16m1(op0, op1, op2) \ +__builtin_rvv_vle16ff_v_u16m1((const uint16_t *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle16ff_v_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle16ff_v_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (const uint16_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle16ff_v_u16m2(op0, op1, op2) \ +__builtin_rvv_vle16ff_v_u16m2((const uint16_t *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle16ff_v_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle16ff_v_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (const uint16_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle16ff_v_u16m4(op0, op1, op2) \ +__builtin_rvv_vle16ff_v_u16m4((const uint16_t *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle16ff_v_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle16ff_v_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (const uint16_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle16ff_v_u16m8(op0, op1, op2) \ +__builtin_rvv_vle16ff_v_u16m8((const uint16_t *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle16ff_v_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle16ff_v_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (const uint16_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle16ff_v_u16mf2(op0, op1, op2) \ +__builtin_rvv_vle16ff_v_u16mf2((const uint16_t *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle16ff_v_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle16ff_v_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (const uint16_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle16ff_v_u16mf4(op0, op1, op2) \ +__builtin_rvv_vle16ff_v_u16mf4((const uint16_t *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle16ff_v_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle16ff_v_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (const uint16_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle32ff_v_i32m1(op0, op1, op2) \ +__builtin_rvv_vle32ff_v_i32m1((const int32_t *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle32ff_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle32ff_v_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (const int32_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle32ff_v_i32m2(op0, op1, op2) \ +__builtin_rvv_vle32ff_v_i32m2((const int32_t *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle32ff_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle32ff_v_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (const int32_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle32ff_v_i32m4(op0, op1, op2) \ +__builtin_rvv_vle32ff_v_i32m4((const int32_t *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle32ff_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle32ff_v_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (const int32_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle32ff_v_i32m8(op0, op1, op2) \ +__builtin_rvv_vle32ff_v_i32m8((const int32_t *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle32ff_v_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle32ff_v_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (const int32_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle32ff_v_i32mf2(op0, op1, op2) \ +__builtin_rvv_vle32ff_v_i32mf2((const int32_t *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle32ff_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle32ff_v_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (const int32_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle32ff_v_u32m1(op0, op1, op2) \ +__builtin_rvv_vle32ff_v_u32m1((const uint32_t *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle32ff_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle32ff_v_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (const uint32_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle32ff_v_u32m2(op0, op1, op2) \ +__builtin_rvv_vle32ff_v_u32m2((const uint32_t *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle32ff_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle32ff_v_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (const uint32_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle32ff_v_u32m4(op0, op1, op2) \ +__builtin_rvv_vle32ff_v_u32m4((const uint32_t *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle32ff_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle32ff_v_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (const uint32_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle32ff_v_u32m8(op0, op1, op2) \ +__builtin_rvv_vle32ff_v_u32m8((const uint32_t *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle32ff_v_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle32ff_v_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (const uint32_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle32ff_v_u32mf2(op0, op1, op2) \ +__builtin_rvv_vle32ff_v_u32mf2((const uint32_t *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle32ff_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle32ff_v_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (const uint32_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle64ff_v_i64m1(op0, op1, op2) \ +__builtin_rvv_vle64ff_v_i64m1((const int64_t *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle64ff_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle64ff_v_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (const int64_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle64ff_v_i64m2(op0, op1, op2) \ +__builtin_rvv_vle64ff_v_i64m2((const int64_t *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle64ff_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle64ff_v_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (const int64_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle64ff_v_i64m4(op0, op1, op2) \ +__builtin_rvv_vle64ff_v_i64m4((const int64_t *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle64ff_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle64ff_v_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (const int64_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle64ff_v_i64m8(op0, op1, op2) \ +__builtin_rvv_vle64ff_v_i64m8((const int64_t *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle64ff_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle64ff_v_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (const int64_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle64ff_v_u64m1(op0, op1, op2) \ +__builtin_rvv_vle64ff_v_u64m1((const uint64_t *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle64ff_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle64ff_v_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (const uint64_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle64ff_v_u64m2(op0, op1, op2) \ +__builtin_rvv_vle64ff_v_u64m2((const uint64_t *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle64ff_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle64ff_v_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (const uint64_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle64ff_v_u64m4(op0, op1, op2) \ +__builtin_rvv_vle64ff_v_u64m4((const uint64_t *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle64ff_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle64ff_v_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (const uint64_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle64ff_v_u64m8(op0, op1, op2) \ +__builtin_rvv_vle64ff_v_u64m8((const uint64_t *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle64ff_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle64ff_v_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (const uint64_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle8_v_i8m1(op0, op1) \ +__builtin_rvv_vle8_v_i8m1((const int8_t *)(op0), (size_t)(op1)) +#define vle8_v_i8m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vle8_v_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (const int8_t *)(op2), (size_t)(op3)) +#define vle8_v_i8m2(op0, op1) \ +__builtin_rvv_vle8_v_i8m2((const int8_t *)(op0), (size_t)(op1)) +#define vle8_v_i8m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vle8_v_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (const int8_t *)(op2), (size_t)(op3)) +#define vle8_v_i8m4(op0, op1) \ +__builtin_rvv_vle8_v_i8m4((const int8_t *)(op0), (size_t)(op1)) +#define vle8_v_i8m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vle8_v_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (const int8_t *)(op2), (size_t)(op3)) +#define vle8_v_i8m8(op0, op1) \ +__builtin_rvv_vle8_v_i8m8((const int8_t *)(op0), (size_t)(op1)) +#define vle8_v_i8m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vle8_v_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (const int8_t *)(op2), (size_t)(op3)) +#define vle8_v_i8mf2(op0, op1) \ +__builtin_rvv_vle8_v_i8mf2((const int8_t *)(op0), (size_t)(op1)) +#define vle8_v_i8mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vle8_v_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (const int8_t *)(op2), (size_t)(op3)) +#define vle8_v_i8mf4(op0, op1) \ +__builtin_rvv_vle8_v_i8mf4((const int8_t *)(op0), (size_t)(op1)) +#define vle8_v_i8mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vle8_v_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (const int8_t *)(op2), (size_t)(op3)) +#define vle8_v_i8mf8(op0, op1) \ +__builtin_rvv_vle8_v_i8mf8((const int8_t *)(op0), (size_t)(op1)) +#define vle8_v_i8mf8_m(op0, op1, op2, op3) \ +__builtin_rvv_vle8_v_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (const int8_t *)(op2), (size_t)(op3)) +#define vle8_v_u8m1(op0, op1) \ +__builtin_rvv_vle8_v_u8m1((const uint8_t *)(op0), (size_t)(op1)) +#define vle8_v_u8m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vle8_v_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (const uint8_t *)(op2), (size_t)(op3)) +#define vle8_v_u8m2(op0, op1) \ +__builtin_rvv_vle8_v_u8m2((const uint8_t *)(op0), (size_t)(op1)) +#define vle8_v_u8m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vle8_v_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (const uint8_t *)(op2), (size_t)(op3)) +#define vle8_v_u8m4(op0, op1) \ +__builtin_rvv_vle8_v_u8m4((const uint8_t *)(op0), (size_t)(op1)) +#define vle8_v_u8m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vle8_v_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (const uint8_t *)(op2), (size_t)(op3)) +#define vle8_v_u8m8(op0, op1) \ +__builtin_rvv_vle8_v_u8m8((const uint8_t *)(op0), (size_t)(op1)) +#define vle8_v_u8m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vle8_v_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (const uint8_t *)(op2), (size_t)(op3)) +#define vle8_v_u8mf2(op0, op1) \ +__builtin_rvv_vle8_v_u8mf2((const uint8_t *)(op0), (size_t)(op1)) +#define vle8_v_u8mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vle8_v_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (const uint8_t *)(op2), (size_t)(op3)) +#define vle8_v_u8mf4(op0, op1) \ +__builtin_rvv_vle8_v_u8mf4((const uint8_t *)(op0), (size_t)(op1)) +#define vle8_v_u8mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vle8_v_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (const uint8_t *)(op2), (size_t)(op3)) +#define vle8_v_u8mf8(op0, op1) \ +__builtin_rvv_vle8_v_u8mf8((const uint8_t *)(op0), (size_t)(op1)) +#define vle8_v_u8mf8_m(op0, op1, op2, op3) \ +__builtin_rvv_vle8_v_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (const uint8_t *)(op2), (size_t)(op3)) +#define vle8ff_v_i8m1(op0, op1, op2) \ +__builtin_rvv_vle8ff_v_i8m1((const int8_t *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle8ff_v_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle8ff_v_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (const int8_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle8ff_v_i8m2(op0, op1, op2) \ +__builtin_rvv_vle8ff_v_i8m2((const int8_t *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle8ff_v_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle8ff_v_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (const int8_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle8ff_v_i8m4(op0, op1, op2) \ +__builtin_rvv_vle8ff_v_i8m4((const int8_t *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle8ff_v_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle8ff_v_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (const int8_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle8ff_v_i8m8(op0, op1, op2) \ +__builtin_rvv_vle8ff_v_i8m8((const int8_t *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle8ff_v_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle8ff_v_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (const int8_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle8ff_v_i8mf2(op0, op1, op2) \ +__builtin_rvv_vle8ff_v_i8mf2((const int8_t *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle8ff_v_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle8ff_v_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (const int8_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle8ff_v_i8mf4(op0, op1, op2) \ +__builtin_rvv_vle8ff_v_i8mf4((const int8_t *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle8ff_v_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle8ff_v_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (const int8_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle8ff_v_i8mf8(op0, op1, op2) \ +__builtin_rvv_vle8ff_v_i8mf8((const int8_t *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle8ff_v_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle8ff_v_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (const int8_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle8ff_v_u8m1(op0, op1, op2) \ +__builtin_rvv_vle8ff_v_u8m1((const uint8_t *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle8ff_v_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle8ff_v_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (const uint8_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle8ff_v_u8m2(op0, op1, op2) \ +__builtin_rvv_vle8ff_v_u8m2((const uint8_t *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle8ff_v_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle8ff_v_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (const uint8_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle8ff_v_u8m4(op0, op1, op2) \ +__builtin_rvv_vle8ff_v_u8m4((const uint8_t *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle8ff_v_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle8ff_v_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (const uint8_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle8ff_v_u8m8(op0, op1, op2) \ +__builtin_rvv_vle8ff_v_u8m8((const uint8_t *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle8ff_v_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle8ff_v_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (const uint8_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle8ff_v_u8mf2(op0, op1, op2) \ +__builtin_rvv_vle8ff_v_u8mf2((const uint8_t *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle8ff_v_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle8ff_v_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (const uint8_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle8ff_v_u8mf4(op0, op1, op2) \ +__builtin_rvv_vle8ff_v_u8mf4((const uint8_t *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle8ff_v_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle8ff_v_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (const uint8_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle8ff_v_u8mf8(op0, op1, op2) \ +__builtin_rvv_vle8ff_v_u8mf8((const uint8_t *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle8ff_v_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle8ff_v_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (const uint8_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vneg_v_i8m1(op0, op1) \ +__builtin_rvv_vneg_v_i8m1((vint8m1_t)(op0), (size_t)(op1)) +#define vneg_v_i8m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vneg_v_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vneg_v_i8m2(op0, op1) \ +__builtin_rvv_vneg_v_i8m2((vint8m2_t)(op0), (size_t)(op1)) +#define vneg_v_i8m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vneg_v_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (size_t)(op3)) +#define vneg_v_i8m4(op0, op1) \ +__builtin_rvv_vneg_v_i8m4((vint8m4_t)(op0), (size_t)(op1)) +#define vneg_v_i8m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vneg_v_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (size_t)(op3)) +#define vneg_v_i8m8(op0, op1) \ +__builtin_rvv_vneg_v_i8m8((vint8m8_t)(op0), (size_t)(op1)) +#define vneg_v_i8m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vneg_v_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (size_t)(op3)) +#define vneg_v_i8mf2(op0, op1) \ +__builtin_rvv_vneg_v_i8mf2((vint8mf2_t)(op0), (size_t)(op1)) +#define vneg_v_i8mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vneg_v_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (size_t)(op3)) +#define vneg_v_i8mf4(op0, op1) \ +__builtin_rvv_vneg_v_i8mf4((vint8mf4_t)(op0), (size_t)(op1)) +#define vneg_v_i8mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vneg_v_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (size_t)(op3)) +#define vneg_v_i8mf8(op0, op1) \ +__builtin_rvv_vneg_v_i8mf8((vint8mf8_t)(op0), (size_t)(op1)) +#define vneg_v_i8mf8_m(op0, op1, op2, op3) \ +__builtin_rvv_vneg_v_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (size_t)(op3)) +#define vneg_v_i16m1(op0, op1) \ +__builtin_rvv_vneg_v_i16m1((vint16m1_t)(op0), (size_t)(op1)) +#define vneg_v_i16m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vneg_v_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vneg_v_i16m2(op0, op1) \ +__builtin_rvv_vneg_v_i16m2((vint16m2_t)(op0), (size_t)(op1)) +#define vneg_v_i16m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vneg_v_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (size_t)(op3)) +#define vneg_v_i16m4(op0, op1) \ +__builtin_rvv_vneg_v_i16m4((vint16m4_t)(op0), (size_t)(op1)) +#define vneg_v_i16m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vneg_v_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (size_t)(op3)) +#define vneg_v_i16m8(op0, op1) \ +__builtin_rvv_vneg_v_i16m8((vint16m8_t)(op0), (size_t)(op1)) +#define vneg_v_i16m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vneg_v_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (size_t)(op3)) +#define vneg_v_i16mf2(op0, op1) \ +__builtin_rvv_vneg_v_i16mf2((vint16mf2_t)(op0), (size_t)(op1)) +#define vneg_v_i16mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vneg_v_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (size_t)(op3)) +#define vneg_v_i16mf4(op0, op1) \ +__builtin_rvv_vneg_v_i16mf4((vint16mf4_t)(op0), (size_t)(op1)) +#define vneg_v_i16mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vneg_v_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (size_t)(op3)) +#define vneg_v_i32m1(op0, op1) \ +__builtin_rvv_vneg_v_i32m1((vint32m1_t)(op0), (size_t)(op1)) +#define vneg_v_i32m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vneg_v_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vneg_v_i32m2(op0, op1) \ +__builtin_rvv_vneg_v_i32m2((vint32m2_t)(op0), (size_t)(op1)) +#define vneg_v_i32m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vneg_v_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vneg_v_i32m4(op0, op1) \ +__builtin_rvv_vneg_v_i32m4((vint32m4_t)(op0), (size_t)(op1)) +#define vneg_v_i32m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vneg_v_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vneg_v_i32m8(op0, op1) \ +__builtin_rvv_vneg_v_i32m8((vint32m8_t)(op0), (size_t)(op1)) +#define vneg_v_i32m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vneg_v_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (size_t)(op3)) +#define vneg_v_i32mf2(op0, op1) \ +__builtin_rvv_vneg_v_i32mf2((vint32mf2_t)(op0), (size_t)(op1)) +#define vneg_v_i32mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vneg_v_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vneg_v_i64m1(op0, op1) \ +__builtin_rvv_vneg_v_i64m1((vint64m1_t)(op0), (size_t)(op1)) +#define vneg_v_i64m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vneg_v_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vneg_v_i64m2(op0, op1) \ +__builtin_rvv_vneg_v_i64m2((vint64m2_t)(op0), (size_t)(op1)) +#define vneg_v_i64m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vneg_v_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vneg_v_i64m4(op0, op1) \ +__builtin_rvv_vneg_v_i64m4((vint64m4_t)(op0), (size_t)(op1)) +#define vneg_v_i64m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vneg_v_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vneg_v_i64m8(op0, op1) \ +__builtin_rvv_vneg_v_i64m8((vint64m8_t)(op0), (size_t)(op1)) +#define vneg_v_i64m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vneg_v_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vnot_v_i8m1(op0, op1) \ +__builtin_rvv_vnot_v_i8m1((vint8m1_t)(op0), (size_t)(op1)) +#define vnot_v_i8m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vnot_v_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vnot_v_i8m2(op0, op1) \ +__builtin_rvv_vnot_v_i8m2((vint8m2_t)(op0), (size_t)(op1)) +#define vnot_v_i8m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vnot_v_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (size_t)(op3)) +#define vnot_v_i8m4(op0, op1) \ +__builtin_rvv_vnot_v_i8m4((vint8m4_t)(op0), (size_t)(op1)) +#define vnot_v_i8m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vnot_v_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (size_t)(op3)) +#define vnot_v_i8m8(op0, op1) \ +__builtin_rvv_vnot_v_i8m8((vint8m8_t)(op0), (size_t)(op1)) +#define vnot_v_i8m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vnot_v_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (vint8m8_t)(op2), (size_t)(op3)) +#define vnot_v_i8mf2(op0, op1) \ +__builtin_rvv_vnot_v_i8mf2((vint8mf2_t)(op0), (size_t)(op1)) +#define vnot_v_i8mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vnot_v_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (size_t)(op3)) +#define vnot_v_i8mf4(op0, op1) \ +__builtin_rvv_vnot_v_i8mf4((vint8mf4_t)(op0), (size_t)(op1)) +#define vnot_v_i8mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vnot_v_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (size_t)(op3)) +#define vnot_v_i8mf8(op0, op1) \ +__builtin_rvv_vnot_v_i8mf8((vint8mf8_t)(op0), (size_t)(op1)) +#define vnot_v_i8mf8_m(op0, op1, op2, op3) \ +__builtin_rvv_vnot_v_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (size_t)(op3)) +#define vnot_v_i16m1(op0, op1) \ +__builtin_rvv_vnot_v_i16m1((vint16m1_t)(op0), (size_t)(op1)) +#define vnot_v_i16m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vnot_v_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vnot_v_i16m2(op0, op1) \ +__builtin_rvv_vnot_v_i16m2((vint16m2_t)(op0), (size_t)(op1)) +#define vnot_v_i16m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vnot_v_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (size_t)(op3)) +#define vnot_v_i16m4(op0, op1) \ +__builtin_rvv_vnot_v_i16m4((vint16m4_t)(op0), (size_t)(op1)) +#define vnot_v_i16m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vnot_v_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (size_t)(op3)) +#define vnot_v_i16m8(op0, op1) \ +__builtin_rvv_vnot_v_i16m8((vint16m8_t)(op0), (size_t)(op1)) +#define vnot_v_i16m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vnot_v_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint16m8_t)(op2), (size_t)(op3)) +#define vnot_v_i16mf2(op0, op1) \ +__builtin_rvv_vnot_v_i16mf2((vint16mf2_t)(op0), (size_t)(op1)) +#define vnot_v_i16mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vnot_v_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (size_t)(op3)) +#define vnot_v_i16mf4(op0, op1) \ +__builtin_rvv_vnot_v_i16mf4((vint16mf4_t)(op0), (size_t)(op1)) +#define vnot_v_i16mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vnot_v_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (size_t)(op3)) +#define vnot_v_i32m1(op0, op1) \ +__builtin_rvv_vnot_v_i32m1((vint32m1_t)(op0), (size_t)(op1)) +#define vnot_v_i32m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vnot_v_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vnot_v_i32m2(op0, op1) \ +__builtin_rvv_vnot_v_i32m2((vint32m2_t)(op0), (size_t)(op1)) +#define vnot_v_i32m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vnot_v_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vnot_v_i32m4(op0, op1) \ +__builtin_rvv_vnot_v_i32m4((vint32m4_t)(op0), (size_t)(op1)) +#define vnot_v_i32m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vnot_v_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vnot_v_i32m8(op0, op1) \ +__builtin_rvv_vnot_v_i32m8((vint32m8_t)(op0), (size_t)(op1)) +#define vnot_v_i32m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vnot_v_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint32m8_t)(op2), (size_t)(op3)) +#define vnot_v_i32mf2(op0, op1) \ +__builtin_rvv_vnot_v_i32mf2((vint32mf2_t)(op0), (size_t)(op1)) +#define vnot_v_i32mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vnot_v_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vnot_v_i64m1(op0, op1) \ +__builtin_rvv_vnot_v_i64m1((vint64m1_t)(op0), (size_t)(op1)) +#define vnot_v_i64m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vnot_v_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vnot_v_i64m2(op0, op1) \ +__builtin_rvv_vnot_v_i64m2((vint64m2_t)(op0), (size_t)(op1)) +#define vnot_v_i64m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vnot_v_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vnot_v_i64m4(op0, op1) \ +__builtin_rvv_vnot_v_i64m4((vint64m4_t)(op0), (size_t)(op1)) +#define vnot_v_i64m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vnot_v_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vnot_v_i64m8(op0, op1) \ +__builtin_rvv_vnot_v_i64m8((vint64m8_t)(op0), (size_t)(op1)) +#define vnot_v_i64m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vnot_v_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vlse8_v_i8m1(op0, op1, op2) \ +__builtin_rvv_vlse8_v_i8m1((const int8_t *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse8_v_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse8_v_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (const int8_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse8_v_i8m2(op0, op1, op2) \ +__builtin_rvv_vlse8_v_i8m2((const int8_t *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse8_v_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse8_v_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (const int8_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse8_v_i8m4(op0, op1, op2) \ +__builtin_rvv_vlse8_v_i8m4((const int8_t *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse8_v_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse8_v_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (const int8_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse8_v_i8m8(op0, op1, op2) \ +__builtin_rvv_vlse8_v_i8m8((const int8_t *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse8_v_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse8_v_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (const int8_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse8_v_i8mf2(op0, op1, op2) \ +__builtin_rvv_vlse8_v_i8mf2((const int8_t *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse8_v_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse8_v_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (const int8_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse8_v_i8mf4(op0, op1, op2) \ +__builtin_rvv_vlse8_v_i8mf4((const int8_t *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse8_v_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse8_v_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (const int8_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse8_v_i8mf8(op0, op1, op2) \ +__builtin_rvv_vlse8_v_i8mf8((const int8_t *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse8_v_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse8_v_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (const int8_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vnot_v_u8m1(op0, op1) \ +__builtin_rvv_vnot_v_u8m1((vuint8m1_t)(op0), (size_t)(op1)) +#define vnot_v_u8m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vnot_v_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vnot_v_u8m2(op0, op1) \ +__builtin_rvv_vnot_v_u8m2((vuint8m2_t)(op0), (size_t)(op1)) +#define vnot_v_u8m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vnot_v_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (size_t)(op3)) +#define vnot_v_u8m4(op0, op1) \ +__builtin_rvv_vnot_v_u8m4((vuint8m4_t)(op0), (size_t)(op1)) +#define vnot_v_u8m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vnot_v_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (size_t)(op3)) +#define vnot_v_u8m8(op0, op1) \ +__builtin_rvv_vnot_v_u8m8((vuint8m8_t)(op0), (size_t)(op1)) +#define vnot_v_u8m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vnot_v_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (vuint8m8_t)(op2), (size_t)(op3)) +#define vnot_v_u8mf2(op0, op1) \ +__builtin_rvv_vnot_v_u8mf2((vuint8mf2_t)(op0), (size_t)(op1)) +#define vnot_v_u8mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vnot_v_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (size_t)(op3)) +#define vnot_v_u8mf4(op0, op1) \ +__builtin_rvv_vnot_v_u8mf4((vuint8mf4_t)(op0), (size_t)(op1)) +#define vnot_v_u8mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vnot_v_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (size_t)(op3)) +#define vnot_v_u8mf8(op0, op1) \ +__builtin_rvv_vnot_v_u8mf8((vuint8mf8_t)(op0), (size_t)(op1)) +#define vnot_v_u8mf8_m(op0, op1, op2, op3) \ +__builtin_rvv_vnot_v_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (size_t)(op3)) +#define vnot_v_u16m1(op0, op1) \ +__builtin_rvv_vnot_v_u16m1((vuint16m1_t)(op0), (size_t)(op1)) +#define vnot_v_u16m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vnot_v_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vnot_v_u16m2(op0, op1) \ +__builtin_rvv_vnot_v_u16m2((vuint16m2_t)(op0), (size_t)(op1)) +#define vnot_v_u16m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vnot_v_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (size_t)(op3)) +#define vnot_v_u16m4(op0, op1) \ +__builtin_rvv_vnot_v_u16m4((vuint16m4_t)(op0), (size_t)(op1)) +#define vnot_v_u16m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vnot_v_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (size_t)(op3)) +#define vnot_v_u16m8(op0, op1) \ +__builtin_rvv_vnot_v_u16m8((vuint16m8_t)(op0), (size_t)(op1)) +#define vnot_v_u16m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vnot_v_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint16m8_t)(op2), (size_t)(op3)) +#define vnot_v_u16mf2(op0, op1) \ +__builtin_rvv_vnot_v_u16mf2((vuint16mf2_t)(op0), (size_t)(op1)) +#define vnot_v_u16mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vnot_v_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (size_t)(op3)) +#define vnot_v_u16mf4(op0, op1) \ +__builtin_rvv_vnot_v_u16mf4((vuint16mf4_t)(op0), (size_t)(op1)) +#define vnot_v_u16mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vnot_v_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (size_t)(op3)) +#define vnot_v_u32m1(op0, op1) \ +__builtin_rvv_vnot_v_u32m1((vuint32m1_t)(op0), (size_t)(op1)) +#define vnot_v_u32m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vnot_v_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vnot_v_u32m2(op0, op1) \ +__builtin_rvv_vnot_v_u32m2((vuint32m2_t)(op0), (size_t)(op1)) +#define vnot_v_u32m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vnot_v_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vnot_v_u32m4(op0, op1) \ +__builtin_rvv_vnot_v_u32m4((vuint32m4_t)(op0), (size_t)(op1)) +#define vnot_v_u32m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vnot_v_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vnot_v_u32m8(op0, op1) \ +__builtin_rvv_vnot_v_u32m8((vuint32m8_t)(op0), (size_t)(op1)) +#define vnot_v_u32m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vnot_v_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (size_t)(op3)) +#define vnot_v_u32mf2(op0, op1) \ +__builtin_rvv_vnot_v_u32mf2((vuint32mf2_t)(op0), (size_t)(op1)) +#define vnot_v_u32mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vnot_v_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vnot_v_u64m1(op0, op1) \ +__builtin_rvv_vnot_v_u64m1((vuint64m1_t)(op0), (size_t)(op1)) +#define vnot_v_u64m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vnot_v_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vnot_v_u64m2(op0, op1) \ +__builtin_rvv_vnot_v_u64m2((vuint64m2_t)(op0), (size_t)(op1)) +#define vnot_v_u64m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vnot_v_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vnot_v_u64m4(op0, op1) \ +__builtin_rvv_vnot_v_u64m4((vuint64m4_t)(op0), (size_t)(op1)) +#define vnot_v_u64m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vnot_v_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vnot_v_u64m8(op0, op1) \ +__builtin_rvv_vnot_v_u64m8((vuint64m8_t)(op0), (size_t)(op1)) +#define vnot_v_u64m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vnot_v_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vmmv_m_b8(op0, op1) \ +__builtin_rvv_vmmv_m_b8((vbool8_t)(op0), (size_t)(op1)) +#define vmmv_m_b4(op0, op1) \ +__builtin_rvv_vmmv_m_b4((vbool4_t)(op0), (size_t)(op1)) +#define vmmv_m_b2(op0, op1) \ +__builtin_rvv_vmmv_m_b2((vbool2_t)(op0), (size_t)(op1)) +#define vmmv_m_b1(op0, op1) \ +__builtin_rvv_vmmv_m_b1((vbool1_t)(op0), (size_t)(op1)) +#define vmmv_m_b16(op0, op1) \ +__builtin_rvv_vmmv_m_b16((vbool16_t)(op0), (size_t)(op1)) +#define vmmv_m_b32(op0, op1) \ +__builtin_rvv_vmmv_m_b32((vbool32_t)(op0), (size_t)(op1)) +#define vmmv_m_b64(op0, op1) \ +__builtin_rvv_vmmv_m_b64((vbool64_t)(op0), (size_t)(op1)) +#define vwcvtu_x_x_v_u16m2(op0, op1) \ +__builtin_rvv_vwcvtu_x_x_v_u16m2((vuint8m1_t)(op0), (size_t)(op1)) +#define vwcvtu_x_x_v_u16m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vwcvtu_x_x_v_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vwcvtu_x_x_v_u16m4(op0, op1) \ +__builtin_rvv_vwcvtu_x_x_v_u16m4((vuint8m2_t)(op0), (size_t)(op1)) +#define vwcvtu_x_x_v_u16m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vwcvtu_x_x_v_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint8m2_t)(op2), (size_t)(op3)) +#define vwcvtu_x_x_v_u16m8(op0, op1) \ +__builtin_rvv_vwcvtu_x_x_v_u16m8((vuint8m4_t)(op0), (size_t)(op1)) +#define vwcvtu_x_x_v_u16m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vwcvtu_x_x_v_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint8m4_t)(op2), (size_t)(op3)) +#define vwcvtu_x_x_v_u16m1(op0, op1) \ +__builtin_rvv_vwcvtu_x_x_v_u16m1((vuint8mf2_t)(op0), (size_t)(op1)) +#define vwcvtu_x_x_v_u16m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vwcvtu_x_x_v_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint8mf2_t)(op2), (size_t)(op3)) +#define vwcvtu_x_x_v_u16mf2(op0, op1) \ +__builtin_rvv_vwcvtu_x_x_v_u16mf2((vuint8mf4_t)(op0), (size_t)(op1)) +#define vwcvtu_x_x_v_u16mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vwcvtu_x_x_v_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint8mf4_t)(op2), (size_t)(op3)) +#define vwcvtu_x_x_v_u16mf4(op0, op1) \ +__builtin_rvv_vwcvtu_x_x_v_u16mf4((vuint8mf8_t)(op0), (size_t)(op1)) +#define vwcvtu_x_x_v_u16mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vwcvtu_x_x_v_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint8mf8_t)(op2), (size_t)(op3)) +#define vwcvtu_x_x_v_u32m2(op0, op1) \ +__builtin_rvv_vwcvtu_x_x_v_u32m2((vuint16m1_t)(op0), (size_t)(op1)) +#define vwcvtu_x_x_v_u32m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vwcvtu_x_x_v_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vwcvtu_x_x_v_u32m4(op0, op1) \ +__builtin_rvv_vwcvtu_x_x_v_u32m4((vuint16m2_t)(op0), (size_t)(op1)) +#define vwcvtu_x_x_v_u32m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vwcvtu_x_x_v_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint16m2_t)(op2), (size_t)(op3)) +#define vwcvtu_x_x_v_u32m8(op0, op1) \ +__builtin_rvv_vwcvtu_x_x_v_u32m8((vuint16m4_t)(op0), (size_t)(op1)) +#define vwcvtu_x_x_v_u32m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vwcvtu_x_x_v_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint16m4_t)(op2), (size_t)(op3)) +#define vwcvtu_x_x_v_u32m1(op0, op1) \ +__builtin_rvv_vwcvtu_x_x_v_u32m1((vuint16mf2_t)(op0), (size_t)(op1)) +#define vwcvtu_x_x_v_u32m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vwcvtu_x_x_v_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint16mf2_t)(op2), (size_t)(op3)) +#define vwcvtu_x_x_v_u32mf2(op0, op1) \ +__builtin_rvv_vwcvtu_x_x_v_u32mf2((vuint16mf4_t)(op0), (size_t)(op1)) +#define vwcvtu_x_x_v_u32mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vwcvtu_x_x_v_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint16mf4_t)(op2), (size_t)(op3)) +#define vwcvtu_x_x_v_u64m2(op0, op1) \ +__builtin_rvv_vwcvtu_x_x_v_u64m2((vuint32m1_t)(op0), (size_t)(op1)) +#define vwcvtu_x_x_v_u64m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vwcvtu_x_x_v_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vwcvtu_x_x_v_u64m4(op0, op1) \ +__builtin_rvv_vwcvtu_x_x_v_u64m4((vuint32m2_t)(op0), (size_t)(op1)) +#define vwcvtu_x_x_v_u64m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vwcvtu_x_x_v_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vwcvtu_x_x_v_u64m8(op0, op1) \ +__builtin_rvv_vwcvtu_x_x_v_u64m8((vuint32m4_t)(op0), (size_t)(op1)) +#define vwcvtu_x_x_v_u64m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vwcvtu_x_x_v_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vwcvtu_x_x_v_u64m1(op0, op1) \ +__builtin_rvv_vwcvtu_x_x_v_u64m1((vuint32mf2_t)(op0), (size_t)(op1)) +#define vwcvtu_x_x_v_u64m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vwcvtu_x_x_v_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vncvt_x_x_w_i8m1(op0, op1) \ +__builtin_rvv_vncvt_x_x_w_i8m1((vint16m2_t)(op0), (size_t)(op1)) +#define vncvt_x_x_w_i8m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vncvt_x_x_w_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vint16m2_t)(op2), (size_t)(op3)) +#define vncvt_x_x_w_i8m2(op0, op1) \ +__builtin_rvv_vncvt_x_x_w_i8m2((vint16m4_t)(op0), (size_t)(op1)) +#define vncvt_x_x_w_i8m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vncvt_x_x_w_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vint16m4_t)(op2), (size_t)(op3)) +#define vncvt_x_x_w_i8m4(op0, op1) \ +__builtin_rvv_vncvt_x_x_w_i8m4((vint16m8_t)(op0), (size_t)(op1)) +#define vncvt_x_x_w_i8m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vncvt_x_x_w_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vint16m8_t)(op2), (size_t)(op3)) +#define vncvt_x_x_w_i8mf2(op0, op1) \ +__builtin_rvv_vncvt_x_x_w_i8mf2((vint16m1_t)(op0), (size_t)(op1)) +#define vncvt_x_x_w_i8mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vncvt_x_x_w_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vncvt_x_x_w_i8mf4(op0, op1) \ +__builtin_rvv_vncvt_x_x_w_i8mf4((vint16mf2_t)(op0), (size_t)(op1)) +#define vncvt_x_x_w_i8mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vncvt_x_x_w_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vint16mf2_t)(op2), (size_t)(op3)) +#define vncvt_x_x_w_i8mf8(op0, op1) \ +__builtin_rvv_vncvt_x_x_w_i8mf8((vint16mf4_t)(op0), (size_t)(op1)) +#define vncvt_x_x_w_i8mf8_m(op0, op1, op2, op3) \ +__builtin_rvv_vncvt_x_x_w_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vint16mf4_t)(op2), (size_t)(op3)) +#define vncvt_x_x_w_i16m1(op0, op1) \ +__builtin_rvv_vncvt_x_x_w_i16m1((vint32m2_t)(op0), (size_t)(op1)) +#define vncvt_x_x_w_i16m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vncvt_x_x_w_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vncvt_x_x_w_i16m2(op0, op1) \ +__builtin_rvv_vncvt_x_x_w_i16m2((vint32m4_t)(op0), (size_t)(op1)) +#define vncvt_x_x_w_i16m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vncvt_x_x_w_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vncvt_x_x_w_i16m4(op0, op1) \ +__builtin_rvv_vncvt_x_x_w_i16m4((vint32m8_t)(op0), (size_t)(op1)) +#define vncvt_x_x_w_i16m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vncvt_x_x_w_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint32m8_t)(op2), (size_t)(op3)) +#define vncvt_x_x_w_i16mf2(op0, op1) \ +__builtin_rvv_vncvt_x_x_w_i16mf2((vint32m1_t)(op0), (size_t)(op1)) +#define vncvt_x_x_w_i16mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vncvt_x_x_w_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vncvt_x_x_w_i16mf4(op0, op1) \ +__builtin_rvv_vncvt_x_x_w_i16mf4((vint32mf2_t)(op0), (size_t)(op1)) +#define vncvt_x_x_w_i16mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vncvt_x_x_w_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vncvt_x_x_w_i32m1(op0, op1) \ +__builtin_rvv_vncvt_x_x_w_i32m1((vint64m2_t)(op0), (size_t)(op1)) +#define vncvt_x_x_w_i32m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vncvt_x_x_w_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vncvt_x_x_w_i32m2(op0, op1) \ +__builtin_rvv_vncvt_x_x_w_i32m2((vint64m4_t)(op0), (size_t)(op1)) +#define vncvt_x_x_w_i32m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vncvt_x_x_w_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vncvt_x_x_w_i32m4(op0, op1) \ +__builtin_rvv_vncvt_x_x_w_i32m4((vint64m8_t)(op0), (size_t)(op1)) +#define vncvt_x_x_w_i32m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vncvt_x_x_w_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vncvt_x_x_w_i32mf2(op0, op1) \ +__builtin_rvv_vncvt_x_x_w_i32mf2((vint64m1_t)(op0), (size_t)(op1)) +#define vncvt_x_x_w_i32mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vncvt_x_x_w_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vle16_v_i16m1(op0, op1) \ +__builtin_rvv_vle16_v_i16m1((const int16_t *)(op0), (size_t)(op1)) +#define vle16_v_i16m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vle16_v_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (const int16_t *)(op2), (size_t)(op3)) +#define vle16_v_i16m2(op0, op1) \ +__builtin_rvv_vle16_v_i16m2((const int16_t *)(op0), (size_t)(op1)) +#define vle16_v_i16m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vle16_v_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (const int16_t *)(op2), (size_t)(op3)) +#define vle16_v_i16m4(op0, op1) \ +__builtin_rvv_vle16_v_i16m4((const int16_t *)(op0), (size_t)(op1)) +#define vle16_v_i16m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vle16_v_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (const int16_t *)(op2), (size_t)(op3)) +#define vle16_v_i16m8(op0, op1) \ +__builtin_rvv_vle16_v_i16m8((const int16_t *)(op0), (size_t)(op1)) +#define vle16_v_i16m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vle16_v_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (const int16_t *)(op2), (size_t)(op3)) +#define vle16_v_i16mf2(op0, op1) \ +__builtin_rvv_vle16_v_i16mf2((const int16_t *)(op0), (size_t)(op1)) +#define vle16_v_i16mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vle16_v_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (const int16_t *)(op2), (size_t)(op3)) +#define vle16_v_i16mf4(op0, op1) \ +__builtin_rvv_vle16_v_i16mf4((const int16_t *)(op0), (size_t)(op1)) +#define vle16_v_i16mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vle16_v_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (const int16_t *)(op2), (size_t)(op3)) +#define vle16_v_u16m1(op0, op1) \ +__builtin_rvv_vle16_v_u16m1((const uint16_t *)(op0), (size_t)(op1)) +#define vle16_v_u16m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vle16_v_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (const uint16_t *)(op2), (size_t)(op3)) +#define vle16_v_u16m2(op0, op1) \ +__builtin_rvv_vle16_v_u16m2((const uint16_t *)(op0), (size_t)(op1)) +#define vle16_v_u16m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vle16_v_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (const uint16_t *)(op2), (size_t)(op3)) +#define vle16_v_u16m4(op0, op1) \ +__builtin_rvv_vle16_v_u16m4((const uint16_t *)(op0), (size_t)(op1)) +#define vle16_v_u16m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vle16_v_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (const uint16_t *)(op2), (size_t)(op3)) +#define vle16_v_u16m8(op0, op1) \ +__builtin_rvv_vle16_v_u16m8((const uint16_t *)(op0), (size_t)(op1)) +#define vle16_v_u16m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vle16_v_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (const uint16_t *)(op2), (size_t)(op3)) +#define vle16_v_u16mf2(op0, op1) \ +__builtin_rvv_vle16_v_u16mf2((const uint16_t *)(op0), (size_t)(op1)) +#define vle16_v_u16mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vle16_v_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (const uint16_t *)(op2), (size_t)(op3)) +#define vle16_v_u16mf4(op0, op1) \ +__builtin_rvv_vle16_v_u16mf4((const uint16_t *)(op0), (size_t)(op1)) +#define vle16_v_u16mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vle16_v_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (const uint16_t *)(op2), (size_t)(op3)) +#define vle32_v_i32m1(op0, op1) \ +__builtin_rvv_vle32_v_i32m1((const int32_t *)(op0), (size_t)(op1)) +#define vle32_v_i32m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vle32_v_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (const int32_t *)(op2), (size_t)(op3)) +#define vle32_v_i32m2(op0, op1) \ +__builtin_rvv_vle32_v_i32m2((const int32_t *)(op0), (size_t)(op1)) +#define vle32_v_i32m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vle32_v_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (const int32_t *)(op2), (size_t)(op3)) +#define vle32_v_i32m4(op0, op1) \ +__builtin_rvv_vle32_v_i32m4((const int32_t *)(op0), (size_t)(op1)) +#define vle32_v_i32m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vle32_v_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (const int32_t *)(op2), (size_t)(op3)) +#define vle32_v_i32m8(op0, op1) \ +__builtin_rvv_vle32_v_i32m8((const int32_t *)(op0), (size_t)(op1)) +#define vle32_v_i32m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vle32_v_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (const int32_t *)(op2), (size_t)(op3)) +#define vle32_v_i32mf2(op0, op1) \ +__builtin_rvv_vle32_v_i32mf2((const int32_t *)(op0), (size_t)(op1)) +#define vle32_v_i32mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vle32_v_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (const int32_t *)(op2), (size_t)(op3)) +#define vle32_v_u32m1(op0, op1) \ +__builtin_rvv_vle32_v_u32m1((const uint32_t *)(op0), (size_t)(op1)) +#define vle32_v_u32m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vle32_v_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (const uint32_t *)(op2), (size_t)(op3)) +#define vle32_v_u32m2(op0, op1) \ +__builtin_rvv_vle32_v_u32m2((const uint32_t *)(op0), (size_t)(op1)) +#define vle32_v_u32m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vle32_v_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (const uint32_t *)(op2), (size_t)(op3)) +#define vle32_v_u32m4(op0, op1) \ +__builtin_rvv_vle32_v_u32m4((const uint32_t *)(op0), (size_t)(op1)) +#define vle32_v_u32m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vle32_v_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (const uint32_t *)(op2), (size_t)(op3)) +#define vle32_v_u32m8(op0, op1) \ +__builtin_rvv_vle32_v_u32m8((const uint32_t *)(op0), (size_t)(op1)) +#define vle32_v_u32m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vle32_v_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (const uint32_t *)(op2), (size_t)(op3)) +#define vle32_v_u32mf2(op0, op1) \ +__builtin_rvv_vle32_v_u32mf2((const uint32_t *)(op0), (size_t)(op1)) +#define vle32_v_u32mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vle32_v_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (const uint32_t *)(op2), (size_t)(op3)) +#define vlse8_v_u8m1(op0, op1, op2) \ +__builtin_rvv_vlse8_v_u8m1((const uint8_t *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse8_v_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse8_v_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (const uint8_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse8_v_u8m2(op0, op1, op2) \ +__builtin_rvv_vlse8_v_u8m2((const uint8_t *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse8_v_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse8_v_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (const uint8_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse8_v_u8m4(op0, op1, op2) \ +__builtin_rvv_vlse8_v_u8m4((const uint8_t *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse8_v_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse8_v_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (const uint8_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse8_v_u8m8(op0, op1, op2) \ +__builtin_rvv_vlse8_v_u8m8((const uint8_t *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse8_v_u8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse8_v_u8m8_m((vbool1_t)(op0), (vuint8m8_t)(op1), (const uint8_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse8_v_u8mf2(op0, op1, op2) \ +__builtin_rvv_vlse8_v_u8mf2((const uint8_t *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse8_v_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse8_v_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (const uint8_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse8_v_u8mf4(op0, op1, op2) \ +__builtin_rvv_vlse8_v_u8mf4((const uint8_t *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse8_v_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse8_v_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (const uint8_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse8_v_u8mf8(op0, op1, op2) \ +__builtin_rvv_vlse8_v_u8mf8((const uint8_t *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse8_v_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse8_v_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (const uint8_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vle64_v_i64m1(op0, op1) \ +__builtin_rvv_vle64_v_i64m1((const int64_t *)(op0), (size_t)(op1)) +#define vle64_v_i64m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vle64_v_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (const int64_t *)(op2), (size_t)(op3)) +#define vle64_v_i64m2(op0, op1) \ +__builtin_rvv_vle64_v_i64m2((const int64_t *)(op0), (size_t)(op1)) +#define vle64_v_i64m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vle64_v_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (const int64_t *)(op2), (size_t)(op3)) +#define vle64_v_i64m4(op0, op1) \ +__builtin_rvv_vle64_v_i64m4((const int64_t *)(op0), (size_t)(op1)) +#define vle64_v_i64m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vle64_v_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (const int64_t *)(op2), (size_t)(op3)) +#define vle64_v_i64m8(op0, op1) \ +__builtin_rvv_vle64_v_i64m8((const int64_t *)(op0), (size_t)(op1)) +#define vle64_v_i64m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vle64_v_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (const int64_t *)(op2), (size_t)(op3)) +#define vle64_v_u64m1(op0, op1) \ +__builtin_rvv_vle64_v_u64m1((const uint64_t *)(op0), (size_t)(op1)) +#define vle64_v_u64m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vle64_v_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (const uint64_t *)(op2), (size_t)(op3)) +#define vle64_v_u64m2(op0, op1) \ +__builtin_rvv_vle64_v_u64m2((const uint64_t *)(op0), (size_t)(op1)) +#define vle64_v_u64m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vle64_v_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (const uint64_t *)(op2), (size_t)(op3)) +#define vle64_v_u64m4(op0, op1) \ +__builtin_rvv_vle64_v_u64m4((const uint64_t *)(op0), (size_t)(op1)) +#define vle64_v_u64m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vle64_v_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (const uint64_t *)(op2), (size_t)(op3)) +#define vle64_v_u64m8(op0, op1) \ +__builtin_rvv_vle64_v_u64m8((const uint64_t *)(op0), (size_t)(op1)) +#define vle64_v_u64m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vle64_v_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (const uint64_t *)(op2), (size_t)(op3)) +#define vse16_v_i16m1(op0, op1, op2) \ +__builtin_rvv_vse16_v_i16m1((int16_t *)(op0), (vint16m1_t)(op1), (size_t)(op2)) +#define vse16_v_i16m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vse16_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vse16_v_i16m2(op0, op1, op2) \ +__builtin_rvv_vse16_v_i16m2((int16_t *)(op0), (vint16m2_t)(op1), (size_t)(op2)) +#define vse16_v_i16m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vse16_v_i16m2_m((vbool8_t)(op0), (int16_t *)(op1), (vint16m2_t)(op2), (size_t)(op3)) +#define vse16_v_i16m4(op0, op1, op2) \ +__builtin_rvv_vse16_v_i16m4((int16_t *)(op0), (vint16m4_t)(op1), (size_t)(op2)) +#define vse16_v_i16m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vse16_v_i16m4_m((vbool4_t)(op0), (int16_t *)(op1), (vint16m4_t)(op2), (size_t)(op3)) +#define vse16_v_i16m8(op0, op1, op2) \ +__builtin_rvv_vse16_v_i16m8((int16_t *)(op0), (vint16m8_t)(op1), (size_t)(op2)) +#define vse16_v_i16m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vse16_v_i16m8_m((vbool2_t)(op0), (int16_t *)(op1), (vint16m8_t)(op2), (size_t)(op3)) +#define vse16_v_i16mf2(op0, op1, op2) \ +__builtin_rvv_vse16_v_i16mf2((int16_t *)(op0), (vint16mf2_t)(op1), (size_t)(op2)) +#define vse16_v_i16mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vse16_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vint16mf2_t)(op2), (size_t)(op3)) +#define vse16_v_i16mf4(op0, op1, op2) \ +__builtin_rvv_vse16_v_i16mf4((int16_t *)(op0), (vint16mf4_t)(op1), (size_t)(op2)) +#define vse16_v_i16mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vse16_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vint16mf4_t)(op2), (size_t)(op3)) +#define vse16_v_u16m1(op0, op1, op2) \ +__builtin_rvv_vse16_v_u16m1((uint16_t *)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vse16_v_u16m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vse16_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vse16_v_u16m2(op0, op1, op2) \ +__builtin_rvv_vse16_v_u16m2((uint16_t *)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vse16_v_u16m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vse16_v_u16m2_m((vbool8_t)(op0), (uint16_t *)(op1), (vuint16m2_t)(op2), (size_t)(op3)) +#define vse16_v_u16m4(op0, op1, op2) \ +__builtin_rvv_vse16_v_u16m4((uint16_t *)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vse16_v_u16m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vse16_v_u16m4_m((vbool4_t)(op0), (uint16_t *)(op1), (vuint16m4_t)(op2), (size_t)(op3)) +#define vse16_v_u16m8(op0, op1, op2) \ +__builtin_rvv_vse16_v_u16m8((uint16_t *)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vse16_v_u16m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vse16_v_u16m8_m((vbool2_t)(op0), (uint16_t *)(op1), (vuint16m8_t)(op2), (size_t)(op3)) +#define vse16_v_u16mf2(op0, op1, op2) \ +__builtin_rvv_vse16_v_u16mf2((uint16_t *)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vse16_v_u16mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vse16_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint16mf2_t)(op2), (size_t)(op3)) +#define vse16_v_u16mf4(op0, op1, op2) \ +__builtin_rvv_vse16_v_u16mf4((uint16_t *)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vse16_v_u16mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vse16_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint16mf4_t)(op2), (size_t)(op3)) +#define vse32_v_i32m1(op0, op1, op2) \ +__builtin_rvv_vse32_v_i32m1((int32_t *)(op0), (vint32m1_t)(op1), (size_t)(op2)) +#define vse32_v_i32m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vse32_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vse32_v_i32m2(op0, op1, op2) \ +__builtin_rvv_vse32_v_i32m2((int32_t *)(op0), (vint32m2_t)(op1), (size_t)(op2)) +#define vse32_v_i32m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vse32_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vse32_v_i32m4(op0, op1, op2) \ +__builtin_rvv_vse32_v_i32m4((int32_t *)(op0), (vint32m4_t)(op1), (size_t)(op2)) +#define vse32_v_i32m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vse32_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vse32_v_i32m8(op0, op1, op2) \ +__builtin_rvv_vse32_v_i32m8((int32_t *)(op0), (vint32m8_t)(op1), (size_t)(op2)) +#define vse32_v_i32m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vse32_v_i32m8_m((vbool4_t)(op0), (int32_t *)(op1), (vint32m8_t)(op2), (size_t)(op3)) +#define vse32_v_i32mf2(op0, op1, op2) \ +__builtin_rvv_vse32_v_i32mf2((int32_t *)(op0), (vint32mf2_t)(op1), (size_t)(op2)) +#define vse32_v_i32mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vse32_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vse32_v_u32m1(op0, op1, op2) \ +__builtin_rvv_vse32_v_u32m1((uint32_t *)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vse32_v_u32m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vse32_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vse32_v_u32m2(op0, op1, op2) \ +__builtin_rvv_vse32_v_u32m2((uint32_t *)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vse32_v_u32m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vse32_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vse32_v_u32m4(op0, op1, op2) \ +__builtin_rvv_vse32_v_u32m4((uint32_t *)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vse32_v_u32m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vse32_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vse32_v_u32m8(op0, op1, op2) \ +__builtin_rvv_vse32_v_u32m8((uint32_t *)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vse32_v_u32m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vse32_v_u32m8_m((vbool4_t)(op0), (uint32_t *)(op1), (vuint32m8_t)(op2), (size_t)(op3)) +#define vse32_v_u32mf2(op0, op1, op2) \ +__builtin_rvv_vse32_v_u32mf2((uint32_t *)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vse32_v_u32mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vse32_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vluxei8_v_i8m1(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_i8m1((const int8_t *)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vluxei8_v_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (const int8_t *)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vluxei8_v_i8m2(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_i8m2((const int8_t *)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vluxei8_v_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (const int8_t *)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vluxei8_v_i8m4(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_i8m4((const int8_t *)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vluxei8_v_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (const int8_t *)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vluxei8_v_i8m8(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_i8m8((const int8_t *)(op0), (vuint8m8_t)(op1), (size_t)(op2)) +#define vluxei8_v_i8m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_i8m8_m((vbool1_t)(op0), (vint8m8_t)(op1), (const int8_t *)(op2), (vuint8m8_t)(op3), (size_t)(op4)) +#define vluxei8_v_i8mf2(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_i8mf2((const int8_t *)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vluxei8_v_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (const int8_t *)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vluxei8_v_i8mf4(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_i8mf4((const int8_t *)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vluxei8_v_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (const int8_t *)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vluxei8_v_i8mf8(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_i8mf8((const int8_t *)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vluxei8_v_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (const int8_t *)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vse64_v_i64m1(op0, op1, op2) \ +__builtin_rvv_vse64_v_i64m1((int64_t *)(op0), (vint64m1_t)(op1), (size_t)(op2)) +#define vse64_v_i64m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vse64_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vse64_v_i64m2(op0, op1, op2) \ +__builtin_rvv_vse64_v_i64m2((int64_t *)(op0), (vint64m2_t)(op1), (size_t)(op2)) +#define vse64_v_i64m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vse64_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vse64_v_i64m4(op0, op1, op2) \ +__builtin_rvv_vse64_v_i64m4((int64_t *)(op0), (vint64m4_t)(op1), (size_t)(op2)) +#define vse64_v_i64m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vse64_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vse64_v_i64m8(op0, op1, op2) \ +__builtin_rvv_vse64_v_i64m8((int64_t *)(op0), (vint64m8_t)(op1), (size_t)(op2)) +#define vse64_v_i64m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vse64_v_i64m8_m((vbool8_t)(op0), (int64_t *)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vse64_v_u64m1(op0, op1, op2) \ +__builtin_rvv_vse64_v_u64m1((uint64_t *)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vse64_v_u64m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vse64_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vse64_v_u64m2(op0, op1, op2) \ +__builtin_rvv_vse64_v_u64m2((uint64_t *)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vse64_v_u64m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vse64_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vse64_v_u64m4(op0, op1, op2) \ +__builtin_rvv_vse64_v_u64m4((uint64_t *)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vse64_v_u64m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vse64_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vse64_v_u64m8(op0, op1, op2) \ +__builtin_rvv_vse64_v_u64m8((uint64_t *)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vse64_v_u64m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vse64_v_u64m8_m((vbool8_t)(op0), (uint64_t *)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vlse16_v_i16m1(op0, op1, op2) \ +__builtin_rvv_vlse16_v_i16m1((const int16_t *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse16_v_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse16_v_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (const int16_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse16_v_i16m2(op0, op1, op2) \ +__builtin_rvv_vlse16_v_i16m2((const int16_t *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse16_v_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse16_v_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (const int16_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse16_v_i16m4(op0, op1, op2) \ +__builtin_rvv_vlse16_v_i16m4((const int16_t *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse16_v_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse16_v_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (const int16_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse16_v_i16m8(op0, op1, op2) \ +__builtin_rvv_vlse16_v_i16m8((const int16_t *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse16_v_i16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse16_v_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (const int16_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse16_v_i16mf2(op0, op1, op2) \ +__builtin_rvv_vlse16_v_i16mf2((const int16_t *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse16_v_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse16_v_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (const int16_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse16_v_i16mf4(op0, op1, op2) \ +__builtin_rvv_vlse16_v_i16mf4((const int16_t *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse16_v_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse16_v_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (const int16_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse16_v_u16m1(op0, op1, op2) \ +__builtin_rvv_vlse16_v_u16m1((const uint16_t *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse16_v_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse16_v_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (const uint16_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse16_v_u16m2(op0, op1, op2) \ +__builtin_rvv_vlse16_v_u16m2((const uint16_t *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse16_v_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse16_v_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (const uint16_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse16_v_u16m4(op0, op1, op2) \ +__builtin_rvv_vlse16_v_u16m4((const uint16_t *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse16_v_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse16_v_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (const uint16_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse16_v_u16m8(op0, op1, op2) \ +__builtin_rvv_vlse16_v_u16m8((const uint16_t *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse16_v_u16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse16_v_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (const uint16_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse16_v_u16mf2(op0, op1, op2) \ +__builtin_rvv_vlse16_v_u16mf2((const uint16_t *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse16_v_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse16_v_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (const uint16_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse16_v_u16mf4(op0, op1, op2) \ +__builtin_rvv_vlse16_v_u16mf4((const uint16_t *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse16_v_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse16_v_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (const uint16_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse32_v_i32m1(op0, op1, op2) \ +__builtin_rvv_vlse32_v_i32m1((const int32_t *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse32_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse32_v_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (const int32_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse32_v_i32m2(op0, op1, op2) \ +__builtin_rvv_vlse32_v_i32m2((const int32_t *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse32_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse32_v_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (const int32_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse32_v_i32m4(op0, op1, op2) \ +__builtin_rvv_vlse32_v_i32m4((const int32_t *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse32_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse32_v_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (const int32_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse32_v_i32m8(op0, op1, op2) \ +__builtin_rvv_vlse32_v_i32m8((const int32_t *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse32_v_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse32_v_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (const int32_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse32_v_i32mf2(op0, op1, op2) \ +__builtin_rvv_vlse32_v_i32mf2((const int32_t *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse32_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse32_v_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (const int32_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse32_v_u32m1(op0, op1, op2) \ +__builtin_rvv_vlse32_v_u32m1((const uint32_t *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse32_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse32_v_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (const uint32_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse32_v_u32m2(op0, op1, op2) \ +__builtin_rvv_vlse32_v_u32m2((const uint32_t *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse32_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse32_v_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (const uint32_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse32_v_u32m4(op0, op1, op2) \ +__builtin_rvv_vlse32_v_u32m4((const uint32_t *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse32_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse32_v_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (const uint32_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse32_v_u32m8(op0, op1, op2) \ +__builtin_rvv_vlse32_v_u32m8((const uint32_t *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse32_v_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse32_v_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (const uint32_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse32_v_u32mf2(op0, op1, op2) \ +__builtin_rvv_vlse32_v_u32mf2((const uint32_t *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse32_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse32_v_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (const uint32_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse64_v_i64m1(op0, op1, op2) \ +__builtin_rvv_vlse64_v_i64m1((const int64_t *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse64_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse64_v_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (const int64_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse64_v_i64m2(op0, op1, op2) \ +__builtin_rvv_vlse64_v_i64m2((const int64_t *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse64_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse64_v_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (const int64_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse64_v_i64m4(op0, op1, op2) \ +__builtin_rvv_vlse64_v_i64m4((const int64_t *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse64_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse64_v_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (const int64_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse64_v_i64m8(op0, op1, op2) \ +__builtin_rvv_vlse64_v_i64m8((const int64_t *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse64_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse64_v_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (const int64_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vfirst_m_b8(op0, op1) \ +__builtin_rvv_vfirst_m_b8((vbool8_t)(op0), (size_t)(op1)) +#define vfirst_m_b8_m(op0, op1, op2) \ +__builtin_rvv_vfirst_m_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (size_t)(op2)) +#define vfirst_m_b4(op0, op1) \ +__builtin_rvv_vfirst_m_b4((vbool4_t)(op0), (size_t)(op1)) +#define vfirst_m_b4_m(op0, op1, op2) \ +__builtin_rvv_vfirst_m_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (size_t)(op2)) +#define vfirst_m_b2(op0, op1) \ +__builtin_rvv_vfirst_m_b2((vbool2_t)(op0), (size_t)(op1)) +#define vfirst_m_b2_m(op0, op1, op2) \ +__builtin_rvv_vfirst_m_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (size_t)(op2)) +#define vfirst_m_b1(op0, op1) \ +__builtin_rvv_vfirst_m_b1((vbool1_t)(op0), (size_t)(op1)) +#define vfirst_m_b1_m(op0, op1, op2) \ +__builtin_rvv_vfirst_m_b1_m((vbool1_t)(op0), (vbool1_t)(op1), (size_t)(op2)) +#define vfirst_m_b16(op0, op1) \ +__builtin_rvv_vfirst_m_b16((vbool16_t)(op0), (size_t)(op1)) +#define vfirst_m_b16_m(op0, op1, op2) \ +__builtin_rvv_vfirst_m_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (size_t)(op2)) +#define vfirst_m_b32(op0, op1) \ +__builtin_rvv_vfirst_m_b32((vbool32_t)(op0), (size_t)(op1)) +#define vfirst_m_b32_m(op0, op1, op2) \ +__builtin_rvv_vfirst_m_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (size_t)(op2)) +#define vfirst_m_b64(op0, op1) \ +__builtin_rvv_vfirst_m_b64((vbool64_t)(op0), (size_t)(op1)) +#define vfirst_m_b64_m(op0, op1, op2) \ +__builtin_rvv_vfirst_m_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (size_t)(op2)) +#define vle1_v_b8(op0, op1) \ +__builtin_rvv_vle1_v_b8((const uint8_t *)(op0), (size_t)(op1)) +#define vle1_v_b4(op0, op1) \ +__builtin_rvv_vle1_v_b4((const uint8_t *)(op0), (size_t)(op1)) +#define vle1_v_b2(op0, op1) \ +__builtin_rvv_vle1_v_b2((const uint8_t *)(op0), (size_t)(op1)) +#define vle1_v_b1(op0, op1) \ +__builtin_rvv_vle1_v_b1((const uint8_t *)(op0), (size_t)(op1)) +#define vle1_v_b16(op0, op1) \ +__builtin_rvv_vle1_v_b16((const uint8_t *)(op0), (size_t)(op1)) +#define vle1_v_b32(op0, op1) \ +__builtin_rvv_vle1_v_b32((const uint8_t *)(op0), (size_t)(op1)) +#define vle1_v_b64(op0, op1) \ +__builtin_rvv_vle1_v_b64((const uint8_t *)(op0), (size_t)(op1)) +#define vlmul_ext_v_i8mf2_i8m1(op0) \ +__builtin_rvv_vlmul_ext_v_i8mf2_i8m1((vint8mf2_t)(op0)) +#define vlmul_ext_v_i8mf4_i8m1(op0) \ +__builtin_rvv_vlmul_ext_v_i8mf4_i8m1((vint8mf4_t)(op0)) +#define vlmul_ext_v_i8mf8_i8m1(op0) \ +__builtin_rvv_vlmul_ext_v_i8mf8_i8m1((vint8mf8_t)(op0)) +#define vlmul_ext_v_i16mf2_i16m1(op0) \ +__builtin_rvv_vlmul_ext_v_i16mf2_i16m1((vint16mf2_t)(op0)) +#define vlmul_ext_v_i16mf4_i16m1(op0) \ +__builtin_rvv_vlmul_ext_v_i16mf4_i16m1((vint16mf4_t)(op0)) +#define vlmul_ext_v_i32mf2_i32m1(op0) \ +__builtin_rvv_vlmul_ext_v_i32mf2_i32m1((vint32mf2_t)(op0)) +#define vlmul_ext_v_i8mf4_i8mf2(op0) \ +__builtin_rvv_vlmul_ext_v_i8mf4_i8mf2((vint8mf4_t)(op0)) +#define vlmul_ext_v_i8mf8_i8mf2(op0) \ +__builtin_rvv_vlmul_ext_v_i8mf8_i8mf2((vint8mf8_t)(op0)) +#define vlmul_ext_v_i16mf4_i16mf2(op0) \ +__builtin_rvv_vlmul_ext_v_i16mf4_i16mf2((vint16mf4_t)(op0)) +#define vlmul_ext_v_i8mf8_i8mf4(op0) \ +__builtin_rvv_vlmul_ext_v_i8mf8_i8mf4((vint8mf8_t)(op0)) +#define vlmul_ext_v_i8m1_i8m2(op0) \ +__builtin_rvv_vlmul_ext_v_i8m1_i8m2((vint8m1_t)(op0)) +#define vlmul_ext_v_i8mf2_i8m2(op0) \ +__builtin_rvv_vlmul_ext_v_i8mf2_i8m2((vint8mf2_t)(op0)) +#define vlmul_ext_v_i8mf4_i8m2(op0) \ +__builtin_rvv_vlmul_ext_v_i8mf4_i8m2((vint8mf4_t)(op0)) +#define vlmul_ext_v_i8mf8_i8m2(op0) \ +__builtin_rvv_vlmul_ext_v_i8mf8_i8m2((vint8mf8_t)(op0)) +#define vlmul_ext_v_i16m1_i16m2(op0) \ +__builtin_rvv_vlmul_ext_v_i16m1_i16m2((vint16m1_t)(op0)) +#define vlmul_ext_v_i16mf2_i16m2(op0) \ +__builtin_rvv_vlmul_ext_v_i16mf2_i16m2((vint16mf2_t)(op0)) +#define vlmul_ext_v_i16mf4_i16m2(op0) \ +__builtin_rvv_vlmul_ext_v_i16mf4_i16m2((vint16mf4_t)(op0)) +#define vlmul_ext_v_i32m1_i32m2(op0) \ +__builtin_rvv_vlmul_ext_v_i32m1_i32m2((vint32m1_t)(op0)) +#define vlmul_ext_v_i32mf2_i32m2(op0) \ +__builtin_rvv_vlmul_ext_v_i32mf2_i32m2((vint32mf2_t)(op0)) +#define vlmul_ext_v_i64m1_i64m2(op0) \ +__builtin_rvv_vlmul_ext_v_i64m1_i64m2((vint64m1_t)(op0)) +#define vlmul_ext_v_i8m1_i8m4(op0) \ +__builtin_rvv_vlmul_ext_v_i8m1_i8m4((vint8m1_t)(op0)) +#define vlmul_ext_v_i8m2_i8m4(op0) \ +__builtin_rvv_vlmul_ext_v_i8m2_i8m4((vint8m2_t)(op0)) +#define vlmul_ext_v_i8mf2_i8m4(op0) \ +__builtin_rvv_vlmul_ext_v_i8mf2_i8m4((vint8mf2_t)(op0)) +#define vlmul_ext_v_i8mf4_i8m4(op0) \ +__builtin_rvv_vlmul_ext_v_i8mf4_i8m4((vint8mf4_t)(op0)) +#define vlmul_ext_v_i8mf8_i8m4(op0) \ +__builtin_rvv_vlmul_ext_v_i8mf8_i8m4((vint8mf8_t)(op0)) +#define vlmul_ext_v_i16m1_i16m4(op0) \ +__builtin_rvv_vlmul_ext_v_i16m1_i16m4((vint16m1_t)(op0)) +#define vlmul_ext_v_i16m2_i16m4(op0) \ +__builtin_rvv_vlmul_ext_v_i16m2_i16m4((vint16m2_t)(op0)) +#define vlmul_ext_v_i16mf2_i16m4(op0) \ +__builtin_rvv_vlmul_ext_v_i16mf2_i16m4((vint16mf2_t)(op0)) +#define vlmul_ext_v_i16mf4_i16m4(op0) \ +__builtin_rvv_vlmul_ext_v_i16mf4_i16m4((vint16mf4_t)(op0)) +#define vlmul_ext_v_i32m1_i32m4(op0) \ +__builtin_rvv_vlmul_ext_v_i32m1_i32m4((vint32m1_t)(op0)) +#define vlmul_ext_v_i32m2_i32m4(op0) \ +__builtin_rvv_vlmul_ext_v_i32m2_i32m4((vint32m2_t)(op0)) +#define vlmul_ext_v_i32mf2_i32m4(op0) \ +__builtin_rvv_vlmul_ext_v_i32mf2_i32m4((vint32mf2_t)(op0)) +#define vlmul_ext_v_i64m1_i64m4(op0) \ +__builtin_rvv_vlmul_ext_v_i64m1_i64m4((vint64m1_t)(op0)) +#define vlmul_ext_v_i64m2_i64m4(op0) \ +__builtin_rvv_vlmul_ext_v_i64m2_i64m4((vint64m2_t)(op0)) +#define vlmul_ext_v_i8m1_i8m8(op0) \ +__builtin_rvv_vlmul_ext_v_i8m1_i8m8((vint8m1_t)(op0)) +#define vlmul_ext_v_i8m2_i8m8(op0) \ +__builtin_rvv_vlmul_ext_v_i8m2_i8m8((vint8m2_t)(op0)) +#define vlmul_ext_v_i8m4_i8m8(op0) \ +__builtin_rvv_vlmul_ext_v_i8m4_i8m8((vint8m4_t)(op0)) +#define vlmul_ext_v_i8mf2_i8m8(op0) \ +__builtin_rvv_vlmul_ext_v_i8mf2_i8m8((vint8mf2_t)(op0)) +#define vlmul_ext_v_i8mf4_i8m8(op0) \ +__builtin_rvv_vlmul_ext_v_i8mf4_i8m8((vint8mf4_t)(op0)) +#define vlmul_ext_v_i8mf8_i8m8(op0) \ +__builtin_rvv_vlmul_ext_v_i8mf8_i8m8((vint8mf8_t)(op0)) +#define vlmul_ext_v_i16m1_i16m8(op0) \ +__builtin_rvv_vlmul_ext_v_i16m1_i16m8((vint16m1_t)(op0)) +#define vlmul_ext_v_i16m2_i16m8(op0) \ +__builtin_rvv_vlmul_ext_v_i16m2_i16m8((vint16m2_t)(op0)) +#define vlmul_ext_v_i16m4_i16m8(op0) \ +__builtin_rvv_vlmul_ext_v_i16m4_i16m8((vint16m4_t)(op0)) +#define vlmul_ext_v_i16mf2_i16m8(op0) \ +__builtin_rvv_vlmul_ext_v_i16mf2_i16m8((vint16mf2_t)(op0)) +#define vlmul_ext_v_i16mf4_i16m8(op0) \ +__builtin_rvv_vlmul_ext_v_i16mf4_i16m8((vint16mf4_t)(op0)) +#define vlmul_ext_v_i32m1_i32m8(op0) \ +__builtin_rvv_vlmul_ext_v_i32m1_i32m8((vint32m1_t)(op0)) +#define vlmul_ext_v_i32m2_i32m8(op0) \ +__builtin_rvv_vlmul_ext_v_i32m2_i32m8((vint32m2_t)(op0)) +#define vlmul_ext_v_i32m4_i32m8(op0) \ +__builtin_rvv_vlmul_ext_v_i32m4_i32m8((vint32m4_t)(op0)) +#define vlmul_ext_v_i32mf2_i32m8(op0) \ +__builtin_rvv_vlmul_ext_v_i32mf2_i32m8((vint32mf2_t)(op0)) +#define vlmul_ext_v_i64m1_i64m8(op0) \ +__builtin_rvv_vlmul_ext_v_i64m1_i64m8((vint64m1_t)(op0)) +#define vlmul_ext_v_i64m2_i64m8(op0) \ +__builtin_rvv_vlmul_ext_v_i64m2_i64m8((vint64m2_t)(op0)) +#define vlmul_ext_v_i64m4_i64m8(op0) \ +__builtin_rvv_vlmul_ext_v_i64m4_i64m8((vint64m4_t)(op0)) +#define vlmul_ext_v_u8mf2_u8m1(op0) \ +__builtin_rvv_vlmul_ext_v_u8mf2_u8m1((vuint8mf2_t)(op0)) +#define vlmul_ext_v_u8mf4_u8m1(op0) \ +__builtin_rvv_vlmul_ext_v_u8mf4_u8m1((vuint8mf4_t)(op0)) +#define vlmul_ext_v_u8mf8_u8m1(op0) \ +__builtin_rvv_vlmul_ext_v_u8mf8_u8m1((vuint8mf8_t)(op0)) +#define vlmul_ext_v_u16mf2_u16m1(op0) \ +__builtin_rvv_vlmul_ext_v_u16mf2_u16m1((vuint16mf2_t)(op0)) +#define vlmul_ext_v_u16mf4_u16m1(op0) \ +__builtin_rvv_vlmul_ext_v_u16mf4_u16m1((vuint16mf4_t)(op0)) +#define vlmul_ext_v_u32mf2_u32m1(op0) \ +__builtin_rvv_vlmul_ext_v_u32mf2_u32m1((vuint32mf2_t)(op0)) +#define vlmul_ext_v_u8mf4_u8mf2(op0) \ +__builtin_rvv_vlmul_ext_v_u8mf4_u8mf2((vuint8mf4_t)(op0)) +#define vlmul_ext_v_u8mf8_u8mf2(op0) \ +__builtin_rvv_vlmul_ext_v_u8mf8_u8mf2((vuint8mf8_t)(op0)) +#define vlmul_ext_v_u16mf4_u16mf2(op0) \ +__builtin_rvv_vlmul_ext_v_u16mf4_u16mf2((vuint16mf4_t)(op0)) +#define vlmul_ext_v_u8mf8_u8mf4(op0) \ +__builtin_rvv_vlmul_ext_v_u8mf8_u8mf4((vuint8mf8_t)(op0)) +#define vlmul_ext_v_u8m1_u8m2(op0) \ +__builtin_rvv_vlmul_ext_v_u8m1_u8m2((vuint8m1_t)(op0)) +#define vlmul_ext_v_u8mf2_u8m2(op0) \ +__builtin_rvv_vlmul_ext_v_u8mf2_u8m2((vuint8mf2_t)(op0)) +#define vlmul_ext_v_u8mf4_u8m2(op0) \ +__builtin_rvv_vlmul_ext_v_u8mf4_u8m2((vuint8mf4_t)(op0)) +#define vlmul_ext_v_u8mf8_u8m2(op0) \ +__builtin_rvv_vlmul_ext_v_u8mf8_u8m2((vuint8mf8_t)(op0)) +#define vlmul_ext_v_u16m1_u16m2(op0) \ +__builtin_rvv_vlmul_ext_v_u16m1_u16m2((vuint16m1_t)(op0)) +#define vlmul_ext_v_u16mf2_u16m2(op0) \ +__builtin_rvv_vlmul_ext_v_u16mf2_u16m2((vuint16mf2_t)(op0)) +#define vlmul_ext_v_u16mf4_u16m2(op0) \ +__builtin_rvv_vlmul_ext_v_u16mf4_u16m2((vuint16mf4_t)(op0)) +#define vlmul_ext_v_u32m1_u32m2(op0) \ +__builtin_rvv_vlmul_ext_v_u32m1_u32m2((vuint32m1_t)(op0)) +#define vlmul_ext_v_u32mf2_u32m2(op0) \ +__builtin_rvv_vlmul_ext_v_u32mf2_u32m2((vuint32mf2_t)(op0)) +#define vlmul_ext_v_u64m1_u64m2(op0) \ +__builtin_rvv_vlmul_ext_v_u64m1_u64m2((vuint64m1_t)(op0)) +#define vlmul_ext_v_u8m1_u8m4(op0) \ +__builtin_rvv_vlmul_ext_v_u8m1_u8m4((vuint8m1_t)(op0)) +#define vlmul_ext_v_u8m2_u8m4(op0) \ +__builtin_rvv_vlmul_ext_v_u8m2_u8m4((vuint8m2_t)(op0)) +#define vlmul_ext_v_u8mf2_u8m4(op0) \ +__builtin_rvv_vlmul_ext_v_u8mf2_u8m4((vuint8mf2_t)(op0)) +#define vlmul_ext_v_u8mf4_u8m4(op0) \ +__builtin_rvv_vlmul_ext_v_u8mf4_u8m4((vuint8mf4_t)(op0)) +#define vlmul_ext_v_u8mf8_u8m4(op0) \ +__builtin_rvv_vlmul_ext_v_u8mf8_u8m4((vuint8mf8_t)(op0)) +#define vlmul_ext_v_u16m1_u16m4(op0) \ +__builtin_rvv_vlmul_ext_v_u16m1_u16m4((vuint16m1_t)(op0)) +#define vlmul_ext_v_u16m2_u16m4(op0) \ +__builtin_rvv_vlmul_ext_v_u16m2_u16m4((vuint16m2_t)(op0)) +#define vlmul_ext_v_u16mf2_u16m4(op0) \ +__builtin_rvv_vlmul_ext_v_u16mf2_u16m4((vuint16mf2_t)(op0)) +#define vlmul_ext_v_u16mf4_u16m4(op0) \ +__builtin_rvv_vlmul_ext_v_u16mf4_u16m4((vuint16mf4_t)(op0)) +#define vlmul_ext_v_u32m1_u32m4(op0) \ +__builtin_rvv_vlmul_ext_v_u32m1_u32m4((vuint32m1_t)(op0)) +#define vlmul_ext_v_u32m2_u32m4(op0) \ +__builtin_rvv_vlmul_ext_v_u32m2_u32m4((vuint32m2_t)(op0)) +#define vlmul_ext_v_u32mf2_u32m4(op0) \ +__builtin_rvv_vlmul_ext_v_u32mf2_u32m4((vuint32mf2_t)(op0)) +#define vlmul_ext_v_u64m1_u64m4(op0) \ +__builtin_rvv_vlmul_ext_v_u64m1_u64m4((vuint64m1_t)(op0)) +#define vlmul_ext_v_u64m2_u64m4(op0) \ +__builtin_rvv_vlmul_ext_v_u64m2_u64m4((vuint64m2_t)(op0)) +#define vlmul_ext_v_u8m1_u8m8(op0) \ +__builtin_rvv_vlmul_ext_v_u8m1_u8m8((vuint8m1_t)(op0)) +#define vlmul_ext_v_u8m2_u8m8(op0) \ +__builtin_rvv_vlmul_ext_v_u8m2_u8m8((vuint8m2_t)(op0)) +#define vlmul_ext_v_u8m4_u8m8(op0) \ +__builtin_rvv_vlmul_ext_v_u8m4_u8m8((vuint8m4_t)(op0)) +#define vlmul_ext_v_u8mf2_u8m8(op0) \ +__builtin_rvv_vlmul_ext_v_u8mf2_u8m8((vuint8mf2_t)(op0)) +#define vlmul_ext_v_u8mf4_u8m8(op0) \ +__builtin_rvv_vlmul_ext_v_u8mf4_u8m8((vuint8mf4_t)(op0)) +#define vlmul_ext_v_u8mf8_u8m8(op0) \ +__builtin_rvv_vlmul_ext_v_u8mf8_u8m8((vuint8mf8_t)(op0)) +#define vlmul_ext_v_u16m1_u16m8(op0) \ +__builtin_rvv_vlmul_ext_v_u16m1_u16m8((vuint16m1_t)(op0)) +#define vlmul_ext_v_u16m2_u16m8(op0) \ +__builtin_rvv_vlmul_ext_v_u16m2_u16m8((vuint16m2_t)(op0)) +#define vlmul_ext_v_u16m4_u16m8(op0) \ +__builtin_rvv_vlmul_ext_v_u16m4_u16m8((vuint16m4_t)(op0)) +#define vlmul_ext_v_u16mf2_u16m8(op0) \ +__builtin_rvv_vlmul_ext_v_u16mf2_u16m8((vuint16mf2_t)(op0)) +#define vlmul_ext_v_u16mf4_u16m8(op0) \ +__builtin_rvv_vlmul_ext_v_u16mf4_u16m8((vuint16mf4_t)(op0)) +#define vlmul_ext_v_u32m1_u32m8(op0) \ +__builtin_rvv_vlmul_ext_v_u32m1_u32m8((vuint32m1_t)(op0)) +#define vlmul_ext_v_u32m2_u32m8(op0) \ +__builtin_rvv_vlmul_ext_v_u32m2_u32m8((vuint32m2_t)(op0)) +#define vlmul_ext_v_u32m4_u32m8(op0) \ +__builtin_rvv_vlmul_ext_v_u32m4_u32m8((vuint32m4_t)(op0)) +#define vlmul_ext_v_u32mf2_u32m8(op0) \ +__builtin_rvv_vlmul_ext_v_u32mf2_u32m8((vuint32mf2_t)(op0)) +#define vlmul_ext_v_u64m1_u64m8(op0) \ +__builtin_rvv_vlmul_ext_v_u64m1_u64m8((vuint64m1_t)(op0)) +#define vlmul_ext_v_u64m2_u64m8(op0) \ +__builtin_rvv_vlmul_ext_v_u64m2_u64m8((vuint64m2_t)(op0)) +#define vlmul_ext_v_u64m4_u64m8(op0) \ +__builtin_rvv_vlmul_ext_v_u64m4_u64m8((vuint64m4_t)(op0)) +#define vlmul_trunc_v_i8m1_i8mf2(op0) \ +__builtin_rvv_vlmul_trunc_v_i8m1_i8mf2((vint8m1_t)(op0)) +#define vlmul_trunc_v_i8m2_i8mf2(op0) \ +__builtin_rvv_vlmul_trunc_v_i8m2_i8mf2((vint8m2_t)(op0)) +#define vlmul_trunc_v_i8m4_i8mf2(op0) \ +__builtin_rvv_vlmul_trunc_v_i8m4_i8mf2((vint8m4_t)(op0)) +#define vlmul_trunc_v_i8m8_i8mf2(op0) \ +__builtin_rvv_vlmul_trunc_v_i8m8_i8mf2((vint8m8_t)(op0)) +#define vlmul_trunc_v_i16m1_i16mf2(op0) \ +__builtin_rvv_vlmul_trunc_v_i16m1_i16mf2((vint16m1_t)(op0)) +#define vlmul_trunc_v_i16m2_i16mf2(op0) \ +__builtin_rvv_vlmul_trunc_v_i16m2_i16mf2((vint16m2_t)(op0)) +#define vlmul_trunc_v_i16m4_i16mf2(op0) \ +__builtin_rvv_vlmul_trunc_v_i16m4_i16mf2((vint16m4_t)(op0)) +#define vlmul_trunc_v_i16m8_i16mf2(op0) \ +__builtin_rvv_vlmul_trunc_v_i16m8_i16mf2((vint16m8_t)(op0)) +#define vlmul_trunc_v_i32m1_i32mf2(op0) \ +__builtin_rvv_vlmul_trunc_v_i32m1_i32mf2((vint32m1_t)(op0)) +#define vlmul_trunc_v_i32m2_i32mf2(op0) \ +__builtin_rvv_vlmul_trunc_v_i32m2_i32mf2((vint32m2_t)(op0)) +#define vlmul_trunc_v_i32m4_i32mf2(op0) \ +__builtin_rvv_vlmul_trunc_v_i32m4_i32mf2((vint32m4_t)(op0)) +#define vlmul_trunc_v_i32m8_i32mf2(op0) \ +__builtin_rvv_vlmul_trunc_v_i32m8_i32mf2((vint32m8_t)(op0)) +#define vlmul_trunc_v_i8m1_i8mf4(op0) \ +__builtin_rvv_vlmul_trunc_v_i8m1_i8mf4((vint8m1_t)(op0)) +#define vlmul_trunc_v_i8m2_i8mf4(op0) \ +__builtin_rvv_vlmul_trunc_v_i8m2_i8mf4((vint8m2_t)(op0)) +#define vlmul_trunc_v_i8m4_i8mf4(op0) \ +__builtin_rvv_vlmul_trunc_v_i8m4_i8mf4((vint8m4_t)(op0)) +#define vlmul_trunc_v_i8m8_i8mf4(op0) \ +__builtin_rvv_vlmul_trunc_v_i8m8_i8mf4((vint8m8_t)(op0)) +#define vlmul_trunc_v_i8mf2_i8mf4(op0) \ +__builtin_rvv_vlmul_trunc_v_i8mf2_i8mf4((vint8mf2_t)(op0)) +#define vlmul_trunc_v_i16m1_i16mf4(op0) \ +__builtin_rvv_vlmul_trunc_v_i16m1_i16mf4((vint16m1_t)(op0)) +#define vlmul_trunc_v_i16m2_i16mf4(op0) \ +__builtin_rvv_vlmul_trunc_v_i16m2_i16mf4((vint16m2_t)(op0)) +#define vlmul_trunc_v_i16m4_i16mf4(op0) \ +__builtin_rvv_vlmul_trunc_v_i16m4_i16mf4((vint16m4_t)(op0)) +#define vlmul_trunc_v_i16m8_i16mf4(op0) \ +__builtin_rvv_vlmul_trunc_v_i16m8_i16mf4((vint16m8_t)(op0)) +#define vlmul_trunc_v_i16mf2_i16mf4(op0) \ +__builtin_rvv_vlmul_trunc_v_i16mf2_i16mf4((vint16mf2_t)(op0)) +#define vlmul_trunc_v_i8m1_i8mf8(op0) \ +__builtin_rvv_vlmul_trunc_v_i8m1_i8mf8((vint8m1_t)(op0)) +#define vlmul_trunc_v_i8m2_i8mf8(op0) \ +__builtin_rvv_vlmul_trunc_v_i8m2_i8mf8((vint8m2_t)(op0)) +#define vlmul_trunc_v_i8m4_i8mf8(op0) \ +__builtin_rvv_vlmul_trunc_v_i8m4_i8mf8((vint8m4_t)(op0)) +#define vlmul_trunc_v_i8m8_i8mf8(op0) \ +__builtin_rvv_vlmul_trunc_v_i8m8_i8mf8((vint8m8_t)(op0)) +#define vlmul_trunc_v_i8mf2_i8mf8(op0) \ +__builtin_rvv_vlmul_trunc_v_i8mf2_i8mf8((vint8mf2_t)(op0)) +#define vlmul_trunc_v_i8mf4_i8mf8(op0) \ +__builtin_rvv_vlmul_trunc_v_i8mf4_i8mf8((vint8mf4_t)(op0)) +#define vlmul_trunc_v_i8m2_i8m1(op0) \ +__builtin_rvv_vlmul_trunc_v_i8m2_i8m1((vint8m2_t)(op0)) +#define vlmul_trunc_v_i8m4_i8m1(op0) \ +__builtin_rvv_vlmul_trunc_v_i8m4_i8m1((vint8m4_t)(op0)) +#define vlmul_trunc_v_i8m8_i8m1(op0) \ +__builtin_rvv_vlmul_trunc_v_i8m8_i8m1((vint8m8_t)(op0)) +#define vlmul_trunc_v_i16m2_i16m1(op0) \ +__builtin_rvv_vlmul_trunc_v_i16m2_i16m1((vint16m2_t)(op0)) +#define vlmul_trunc_v_i16m4_i16m1(op0) \ +__builtin_rvv_vlmul_trunc_v_i16m4_i16m1((vint16m4_t)(op0)) +#define vlmul_trunc_v_i16m8_i16m1(op0) \ +__builtin_rvv_vlmul_trunc_v_i16m8_i16m1((vint16m8_t)(op0)) +#define vlmul_trunc_v_i32m2_i32m1(op0) \ +__builtin_rvv_vlmul_trunc_v_i32m2_i32m1((vint32m2_t)(op0)) +#define vlmul_trunc_v_i32m4_i32m1(op0) \ +__builtin_rvv_vlmul_trunc_v_i32m4_i32m1((vint32m4_t)(op0)) +#define vlmul_trunc_v_i32m8_i32m1(op0) \ +__builtin_rvv_vlmul_trunc_v_i32m8_i32m1((vint32m8_t)(op0)) +#define vlmul_trunc_v_i64m2_i64m1(op0) \ +__builtin_rvv_vlmul_trunc_v_i64m2_i64m1((vint64m2_t)(op0)) +#define vlmul_trunc_v_i64m4_i64m1(op0) \ +__builtin_rvv_vlmul_trunc_v_i64m4_i64m1((vint64m4_t)(op0)) +#define vlmul_trunc_v_i64m8_i64m1(op0) \ +__builtin_rvv_vlmul_trunc_v_i64m8_i64m1((vint64m8_t)(op0)) +#define vlmul_trunc_v_i8m4_i8m2(op0) \ +__builtin_rvv_vlmul_trunc_v_i8m4_i8m2((vint8m4_t)(op0)) +#define vlmul_trunc_v_i8m8_i8m2(op0) \ +__builtin_rvv_vlmul_trunc_v_i8m8_i8m2((vint8m8_t)(op0)) +#define vlmul_trunc_v_i16m4_i16m2(op0) \ +__builtin_rvv_vlmul_trunc_v_i16m4_i16m2((vint16m4_t)(op0)) +#define vlmul_trunc_v_i16m8_i16m2(op0) \ +__builtin_rvv_vlmul_trunc_v_i16m8_i16m2((vint16m8_t)(op0)) +#define vlmul_trunc_v_i32m4_i32m2(op0) \ +__builtin_rvv_vlmul_trunc_v_i32m4_i32m2((vint32m4_t)(op0)) +#define vlmul_trunc_v_i32m8_i32m2(op0) \ +__builtin_rvv_vlmul_trunc_v_i32m8_i32m2((vint32m8_t)(op0)) +#define vlmul_trunc_v_i64m4_i64m2(op0) \ +__builtin_rvv_vlmul_trunc_v_i64m4_i64m2((vint64m4_t)(op0)) +#define vlmul_trunc_v_i64m8_i64m2(op0) \ +__builtin_rvv_vlmul_trunc_v_i64m8_i64m2((vint64m8_t)(op0)) +#define vlmul_trunc_v_i8m8_i8m4(op0) \ +__builtin_rvv_vlmul_trunc_v_i8m8_i8m4((vint8m8_t)(op0)) +#define vlmul_trunc_v_i16m8_i16m4(op0) \ +__builtin_rvv_vlmul_trunc_v_i16m8_i16m4((vint16m8_t)(op0)) +#define vlmul_trunc_v_i32m8_i32m4(op0) \ +__builtin_rvv_vlmul_trunc_v_i32m8_i32m4((vint32m8_t)(op0)) +#define vlmul_trunc_v_i64m8_i64m4(op0) \ +__builtin_rvv_vlmul_trunc_v_i64m8_i64m4((vint64m8_t)(op0)) +#define vlmul_trunc_v_u8m1_u8mf2(op0) \ +__builtin_rvv_vlmul_trunc_v_u8m1_u8mf2((vuint8m1_t)(op0)) +#define vlmul_trunc_v_u8m2_u8mf2(op0) \ +__builtin_rvv_vlmul_trunc_v_u8m2_u8mf2((vuint8m2_t)(op0)) +#define vlmul_trunc_v_u8m4_u8mf2(op0) \ +__builtin_rvv_vlmul_trunc_v_u8m4_u8mf2((vuint8m4_t)(op0)) +#define vlmul_trunc_v_u8m8_u8mf2(op0) \ +__builtin_rvv_vlmul_trunc_v_u8m8_u8mf2((vuint8m8_t)(op0)) +#define vlmul_trunc_v_u16m1_u16mf2(op0) \ +__builtin_rvv_vlmul_trunc_v_u16m1_u16mf2((vuint16m1_t)(op0)) +#define vlmul_trunc_v_u16m2_u16mf2(op0) \ +__builtin_rvv_vlmul_trunc_v_u16m2_u16mf2((vuint16m2_t)(op0)) +#define vlmul_trunc_v_u16m4_u16mf2(op0) \ +__builtin_rvv_vlmul_trunc_v_u16m4_u16mf2((vuint16m4_t)(op0)) +#define vlmul_trunc_v_u16m8_u16mf2(op0) \ +__builtin_rvv_vlmul_trunc_v_u16m8_u16mf2((vuint16m8_t)(op0)) +#define vlmul_trunc_v_u32m1_u32mf2(op0) \ +__builtin_rvv_vlmul_trunc_v_u32m1_u32mf2((vuint32m1_t)(op0)) +#define vlmul_trunc_v_u32m2_u32mf2(op0) \ +__builtin_rvv_vlmul_trunc_v_u32m2_u32mf2((vuint32m2_t)(op0)) +#define vlmul_trunc_v_u32m4_u32mf2(op0) \ +__builtin_rvv_vlmul_trunc_v_u32m4_u32mf2((vuint32m4_t)(op0)) +#define vlmul_trunc_v_u32m8_u32mf2(op0) \ +__builtin_rvv_vlmul_trunc_v_u32m8_u32mf2((vuint32m8_t)(op0)) +#define vlmul_trunc_v_u8m1_u8mf4(op0) \ +__builtin_rvv_vlmul_trunc_v_u8m1_u8mf4((vuint8m1_t)(op0)) +#define vlmul_trunc_v_u8m2_u8mf4(op0) \ +__builtin_rvv_vlmul_trunc_v_u8m2_u8mf4((vuint8m2_t)(op0)) +#define vlmul_trunc_v_u8m4_u8mf4(op0) \ +__builtin_rvv_vlmul_trunc_v_u8m4_u8mf4((vuint8m4_t)(op0)) +#define vlmul_trunc_v_u8m8_u8mf4(op0) \ +__builtin_rvv_vlmul_trunc_v_u8m8_u8mf4((vuint8m8_t)(op0)) +#define vlmul_trunc_v_u8mf2_u8mf4(op0) \ +__builtin_rvv_vlmul_trunc_v_u8mf2_u8mf4((vuint8mf2_t)(op0)) +#define vlmul_trunc_v_u16m1_u16mf4(op0) \ +__builtin_rvv_vlmul_trunc_v_u16m1_u16mf4((vuint16m1_t)(op0)) +#define vlmul_trunc_v_u16m2_u16mf4(op0) \ +__builtin_rvv_vlmul_trunc_v_u16m2_u16mf4((vuint16m2_t)(op0)) +#define vlmul_trunc_v_u16m4_u16mf4(op0) \ +__builtin_rvv_vlmul_trunc_v_u16m4_u16mf4((vuint16m4_t)(op0)) +#define vlmul_trunc_v_u16m8_u16mf4(op0) \ +__builtin_rvv_vlmul_trunc_v_u16m8_u16mf4((vuint16m8_t)(op0)) +#define vlmul_trunc_v_u16mf2_u16mf4(op0) \ +__builtin_rvv_vlmul_trunc_v_u16mf2_u16mf4((vuint16mf2_t)(op0)) +#define vlmul_trunc_v_u8m1_u8mf8(op0) \ +__builtin_rvv_vlmul_trunc_v_u8m1_u8mf8((vuint8m1_t)(op0)) +#define vlmul_trunc_v_u8m2_u8mf8(op0) \ +__builtin_rvv_vlmul_trunc_v_u8m2_u8mf8((vuint8m2_t)(op0)) +#define vlmul_trunc_v_u8m4_u8mf8(op0) \ +__builtin_rvv_vlmul_trunc_v_u8m4_u8mf8((vuint8m4_t)(op0)) +#define vlmul_trunc_v_u8m8_u8mf8(op0) \ +__builtin_rvv_vlmul_trunc_v_u8m8_u8mf8((vuint8m8_t)(op0)) +#define vlmul_trunc_v_u8mf2_u8mf8(op0) \ +__builtin_rvv_vlmul_trunc_v_u8mf2_u8mf8((vuint8mf2_t)(op0)) +#define vlmul_trunc_v_u8mf4_u8mf8(op0) \ +__builtin_rvv_vlmul_trunc_v_u8mf4_u8mf8((vuint8mf4_t)(op0)) +#define vlmul_trunc_v_u8m2_u8m1(op0) \ +__builtin_rvv_vlmul_trunc_v_u8m2_u8m1((vuint8m2_t)(op0)) +#define vlmul_trunc_v_u8m4_u8m1(op0) \ +__builtin_rvv_vlmul_trunc_v_u8m4_u8m1((vuint8m4_t)(op0)) +#define vlmul_trunc_v_u8m8_u8m1(op0) \ +__builtin_rvv_vlmul_trunc_v_u8m8_u8m1((vuint8m8_t)(op0)) +#define vlmul_trunc_v_u16m2_u16m1(op0) \ +__builtin_rvv_vlmul_trunc_v_u16m2_u16m1((vuint16m2_t)(op0)) +#define vlmul_trunc_v_u16m4_u16m1(op0) \ +__builtin_rvv_vlmul_trunc_v_u16m4_u16m1((vuint16m4_t)(op0)) +#define vlmul_trunc_v_u16m8_u16m1(op0) \ +__builtin_rvv_vlmul_trunc_v_u16m8_u16m1((vuint16m8_t)(op0)) +#define vlmul_trunc_v_u32m2_u32m1(op0) \ +__builtin_rvv_vlmul_trunc_v_u32m2_u32m1((vuint32m2_t)(op0)) +#define vlmul_trunc_v_u32m4_u32m1(op0) \ +__builtin_rvv_vlmul_trunc_v_u32m4_u32m1((vuint32m4_t)(op0)) +#define vlmul_trunc_v_u32m8_u32m1(op0) \ +__builtin_rvv_vlmul_trunc_v_u32m8_u32m1((vuint32m8_t)(op0)) +#define vlmul_trunc_v_u64m2_u64m1(op0) \ +__builtin_rvv_vlmul_trunc_v_u64m2_u64m1((vuint64m2_t)(op0)) +#define vlmul_trunc_v_u64m4_u64m1(op0) \ +__builtin_rvv_vlmul_trunc_v_u64m4_u64m1((vuint64m4_t)(op0)) +#define vlmul_trunc_v_u64m8_u64m1(op0) \ +__builtin_rvv_vlmul_trunc_v_u64m8_u64m1((vuint64m8_t)(op0)) +#define vlmul_trunc_v_u8m4_u8m2(op0) \ +__builtin_rvv_vlmul_trunc_v_u8m4_u8m2((vuint8m4_t)(op0)) +#define vlmul_trunc_v_u8m8_u8m2(op0) \ +__builtin_rvv_vlmul_trunc_v_u8m8_u8m2((vuint8m8_t)(op0)) +#define vlmul_trunc_v_u16m4_u16m2(op0) \ +__builtin_rvv_vlmul_trunc_v_u16m4_u16m2((vuint16m4_t)(op0)) +#define vlmul_trunc_v_u16m8_u16m2(op0) \ +__builtin_rvv_vlmul_trunc_v_u16m8_u16m2((vuint16m8_t)(op0)) +#define vlmul_trunc_v_u32m4_u32m2(op0) \ +__builtin_rvv_vlmul_trunc_v_u32m4_u32m2((vuint32m4_t)(op0)) +#define vlmul_trunc_v_u32m8_u32m2(op0) \ +__builtin_rvv_vlmul_trunc_v_u32m8_u32m2((vuint32m8_t)(op0)) +#define vlmul_trunc_v_u64m4_u64m2(op0) \ +__builtin_rvv_vlmul_trunc_v_u64m4_u64m2((vuint64m4_t)(op0)) +#define vlmul_trunc_v_u64m8_u64m2(op0) \ +__builtin_rvv_vlmul_trunc_v_u64m8_u64m2((vuint64m8_t)(op0)) +#define vlmul_trunc_v_u8m8_u8m4(op0) \ +__builtin_rvv_vlmul_trunc_v_u8m8_u8m4((vuint8m8_t)(op0)) +#define vlmul_trunc_v_u16m8_u16m4(op0) \ +__builtin_rvv_vlmul_trunc_v_u16m8_u16m4((vuint16m8_t)(op0)) +#define vlmul_trunc_v_u32m8_u32m4(op0) \ +__builtin_rvv_vlmul_trunc_v_u32m8_u32m4((vuint32m8_t)(op0)) +#define vlmul_trunc_v_u64m8_u64m4(op0) \ +__builtin_rvv_vlmul_trunc_v_u64m8_u64m4((vuint64m8_t)(op0)) +#define vmand_mm_b8(op0, op1, op2) \ +__builtin_rvv_vmand_mm_b8((vbool8_t)(op0), (vbool8_t)(op1), (size_t)(op2)) +#define vmand_mm_b4(op0, op1, op2) \ +__builtin_rvv_vmand_mm_b4((vbool4_t)(op0), (vbool4_t)(op1), (size_t)(op2)) +#define vmand_mm_b2(op0, op1, op2) \ +__builtin_rvv_vmand_mm_b2((vbool2_t)(op0), (vbool2_t)(op1), (size_t)(op2)) +#define vmand_mm_b1(op0, op1, op2) \ +__builtin_rvv_vmand_mm_b1((vbool1_t)(op0), (vbool1_t)(op1), (size_t)(op2)) +#define vmand_mm_b16(op0, op1, op2) \ +__builtin_rvv_vmand_mm_b16((vbool16_t)(op0), (vbool16_t)(op1), (size_t)(op2)) +#define vmand_mm_b32(op0, op1, op2) \ +__builtin_rvv_vmand_mm_b32((vbool32_t)(op0), (vbool32_t)(op1), (size_t)(op2)) +#define vmand_mm_b64(op0, op1, op2) \ +__builtin_rvv_vmand_mm_b64((vbool64_t)(op0), (vbool64_t)(op1), (size_t)(op2)) +#define vmandnot_mm_b8(op0, op1, op2) \ +__builtin_rvv_vmandnot_mm_b8((vbool8_t)(op0), (vbool8_t)(op1), (size_t)(op2)) +#define vmandnot_mm_b4(op0, op1, op2) \ +__builtin_rvv_vmandnot_mm_b4((vbool4_t)(op0), (vbool4_t)(op1), (size_t)(op2)) +#define vmandnot_mm_b2(op0, op1, op2) \ +__builtin_rvv_vmandnot_mm_b2((vbool2_t)(op0), (vbool2_t)(op1), (size_t)(op2)) +#define vmandnot_mm_b1(op0, op1, op2) \ +__builtin_rvv_vmandnot_mm_b1((vbool1_t)(op0), (vbool1_t)(op1), (size_t)(op2)) +#define vmandnot_mm_b16(op0, op1, op2) \ +__builtin_rvv_vmandnot_mm_b16((vbool16_t)(op0), (vbool16_t)(op1), (size_t)(op2)) +#define vmandnot_mm_b32(op0, op1, op2) \ +__builtin_rvv_vmandnot_mm_b32((vbool32_t)(op0), (vbool32_t)(op1), (size_t)(op2)) +#define vmandnot_mm_b64(op0, op1, op2) \ +__builtin_rvv_vmandnot_mm_b64((vbool64_t)(op0), (vbool64_t)(op1), (size_t)(op2)) +#define vmclr_m_b8(op0) \ +__builtin_rvv_vmclr_m_b8((size_t)(op0)) +#define vmclr_m_b4(op0) \ +__builtin_rvv_vmclr_m_b4((size_t)(op0)) +#define vmclr_m_b2(op0) \ +__builtin_rvv_vmclr_m_b2((size_t)(op0)) +#define vmclr_m_b1(op0) \ +__builtin_rvv_vmclr_m_b1((size_t)(op0)) +#define vmclr_m_b16(op0) \ +__builtin_rvv_vmclr_m_b16((size_t)(op0)) +#define vmclr_m_b32(op0) \ +__builtin_rvv_vmclr_m_b32((size_t)(op0)) +#define vmclr_m_b64(op0) \ +__builtin_rvv_vmclr_m_b64((size_t)(op0)) +#define vmnand_mm_b8(op0, op1, op2) \ +__builtin_rvv_vmnand_mm_b8((vbool8_t)(op0), (vbool8_t)(op1), (size_t)(op2)) +#define vmnand_mm_b4(op0, op1, op2) \ +__builtin_rvv_vmnand_mm_b4((vbool4_t)(op0), (vbool4_t)(op1), (size_t)(op2)) +#define vmnand_mm_b2(op0, op1, op2) \ +__builtin_rvv_vmnand_mm_b2((vbool2_t)(op0), (vbool2_t)(op1), (size_t)(op2)) +#define vmnand_mm_b1(op0, op1, op2) \ +__builtin_rvv_vmnand_mm_b1((vbool1_t)(op0), (vbool1_t)(op1), (size_t)(op2)) +#define vmnand_mm_b16(op0, op1, op2) \ +__builtin_rvv_vmnand_mm_b16((vbool16_t)(op0), (vbool16_t)(op1), (size_t)(op2)) +#define vmnand_mm_b32(op0, op1, op2) \ +__builtin_rvv_vmnand_mm_b32((vbool32_t)(op0), (vbool32_t)(op1), (size_t)(op2)) +#define vmnand_mm_b64(op0, op1, op2) \ +__builtin_rvv_vmnand_mm_b64((vbool64_t)(op0), (vbool64_t)(op1), (size_t)(op2)) +#define vmnor_mm_b8(op0, op1, op2) \ +__builtin_rvv_vmnor_mm_b8((vbool8_t)(op0), (vbool8_t)(op1), (size_t)(op2)) +#define vmnor_mm_b4(op0, op1, op2) \ +__builtin_rvv_vmnor_mm_b4((vbool4_t)(op0), (vbool4_t)(op1), (size_t)(op2)) +#define vmnor_mm_b2(op0, op1, op2) \ +__builtin_rvv_vmnor_mm_b2((vbool2_t)(op0), (vbool2_t)(op1), (size_t)(op2)) +#define vmnor_mm_b1(op0, op1, op2) \ +__builtin_rvv_vmnor_mm_b1((vbool1_t)(op0), (vbool1_t)(op1), (size_t)(op2)) +#define vmnor_mm_b16(op0, op1, op2) \ +__builtin_rvv_vmnor_mm_b16((vbool16_t)(op0), (vbool16_t)(op1), (size_t)(op2)) +#define vmnor_mm_b32(op0, op1, op2) \ +__builtin_rvv_vmnor_mm_b32((vbool32_t)(op0), (vbool32_t)(op1), (size_t)(op2)) +#define vmnor_mm_b64(op0, op1, op2) \ +__builtin_rvv_vmnor_mm_b64((vbool64_t)(op0), (vbool64_t)(op1), (size_t)(op2)) +#define vmor_mm_b8(op0, op1, op2) \ +__builtin_rvv_vmor_mm_b8((vbool8_t)(op0), (vbool8_t)(op1), (size_t)(op2)) +#define vmor_mm_b4(op0, op1, op2) \ +__builtin_rvv_vmor_mm_b4((vbool4_t)(op0), (vbool4_t)(op1), (size_t)(op2)) +#define vmor_mm_b2(op0, op1, op2) \ +__builtin_rvv_vmor_mm_b2((vbool2_t)(op0), (vbool2_t)(op1), (size_t)(op2)) +#define vmor_mm_b1(op0, op1, op2) \ +__builtin_rvv_vmor_mm_b1((vbool1_t)(op0), (vbool1_t)(op1), (size_t)(op2)) +#define vmor_mm_b16(op0, op1, op2) \ +__builtin_rvv_vmor_mm_b16((vbool16_t)(op0), (vbool16_t)(op1), (size_t)(op2)) +#define vmor_mm_b32(op0, op1, op2) \ +__builtin_rvv_vmor_mm_b32((vbool32_t)(op0), (vbool32_t)(op1), (size_t)(op2)) +#define vmor_mm_b64(op0, op1, op2) \ +__builtin_rvv_vmor_mm_b64((vbool64_t)(op0), (vbool64_t)(op1), (size_t)(op2)) +#define vmornot_mm_b8(op0, op1, op2) \ +__builtin_rvv_vmornot_mm_b8((vbool8_t)(op0), (vbool8_t)(op1), (size_t)(op2)) +#define vmornot_mm_b4(op0, op1, op2) \ +__builtin_rvv_vmornot_mm_b4((vbool4_t)(op0), (vbool4_t)(op1), (size_t)(op2)) +#define vmornot_mm_b2(op0, op1, op2) \ +__builtin_rvv_vmornot_mm_b2((vbool2_t)(op0), (vbool2_t)(op1), (size_t)(op2)) +#define vmornot_mm_b1(op0, op1, op2) \ +__builtin_rvv_vmornot_mm_b1((vbool1_t)(op0), (vbool1_t)(op1), (size_t)(op2)) +#define vmornot_mm_b16(op0, op1, op2) \ +__builtin_rvv_vmornot_mm_b16((vbool16_t)(op0), (vbool16_t)(op1), (size_t)(op2)) +#define vmornot_mm_b32(op0, op1, op2) \ +__builtin_rvv_vmornot_mm_b32((vbool32_t)(op0), (vbool32_t)(op1), (size_t)(op2)) +#define vmornot_mm_b64(op0, op1, op2) \ +__builtin_rvv_vmornot_mm_b64((vbool64_t)(op0), (vbool64_t)(op1), (size_t)(op2)) +#define vmsbf_m_b8(op0, op1) \ +__builtin_rvv_vmsbf_m_b8((vbool8_t)(op0), (size_t)(op1)) +#define vmsbf_m_b8_m(op0, op1, op2, op3) \ +__builtin_rvv_vmsbf_m_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vmsbf_m_b4(op0, op1) \ +__builtin_rvv_vmsbf_m_b4((vbool4_t)(op0), (size_t)(op1)) +#define vmsbf_m_b4_m(op0, op1, op2, op3) \ +__builtin_rvv_vmsbf_m_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vmsbf_m_b2(op0, op1) \ +__builtin_rvv_vmsbf_m_b2((vbool2_t)(op0), (size_t)(op1)) +#define vmsbf_m_b2_m(op0, op1, op2, op3) \ +__builtin_rvv_vmsbf_m_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vbool2_t)(op2), (size_t)(op3)) +#define vmsbf_m_b1(op0, op1) \ +__builtin_rvv_vmsbf_m_b1((vbool1_t)(op0), (size_t)(op1)) +#define vmsbf_m_b1_m(op0, op1, op2, op3) \ +__builtin_rvv_vmsbf_m_b1_m((vbool1_t)(op0), (vbool1_t)(op1), (vbool1_t)(op2), (size_t)(op3)) +#define vmsbf_m_b16(op0, op1) \ +__builtin_rvv_vmsbf_m_b16((vbool16_t)(op0), (size_t)(op1)) +#define vmsbf_m_b16_m(op0, op1, op2, op3) \ +__builtin_rvv_vmsbf_m_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vmsbf_m_b32(op0, op1) \ +__builtin_rvv_vmsbf_m_b32((vbool32_t)(op0), (size_t)(op1)) +#define vmsbf_m_b32_m(op0, op1, op2, op3) \ +__builtin_rvv_vmsbf_m_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vmsbf_m_b64(op0, op1) \ +__builtin_rvv_vmsbf_m_b64((vbool64_t)(op0), (size_t)(op1)) +#define vmsbf_m_b64_m(op0, op1, op2, op3) \ +__builtin_rvv_vmsbf_m_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vmset_m_b8(op0) \ +__builtin_rvv_vmset_m_b8((size_t)(op0)) +#define vmset_m_b4(op0) \ +__builtin_rvv_vmset_m_b4((size_t)(op0)) +#define vmset_m_b2(op0) \ +__builtin_rvv_vmset_m_b2((size_t)(op0)) +#define vmset_m_b1(op0) \ +__builtin_rvv_vmset_m_b1((size_t)(op0)) +#define vmset_m_b16(op0) \ +__builtin_rvv_vmset_m_b16((size_t)(op0)) +#define vmset_m_b32(op0) \ +__builtin_rvv_vmset_m_b32((size_t)(op0)) +#define vmset_m_b64(op0) \ +__builtin_rvv_vmset_m_b64((size_t)(op0)) +#define vmsif_m_b8(op0, op1) \ +__builtin_rvv_vmsif_m_b8((vbool8_t)(op0), (size_t)(op1)) +#define vmsif_m_b8_m(op0, op1, op2, op3) \ +__builtin_rvv_vmsif_m_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vmsif_m_b4(op0, op1) \ +__builtin_rvv_vmsif_m_b4((vbool4_t)(op0), (size_t)(op1)) +#define vmsif_m_b4_m(op0, op1, op2, op3) \ +__builtin_rvv_vmsif_m_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vmsif_m_b2(op0, op1) \ +__builtin_rvv_vmsif_m_b2((vbool2_t)(op0), (size_t)(op1)) +#define vmsif_m_b2_m(op0, op1, op2, op3) \ +__builtin_rvv_vmsif_m_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vbool2_t)(op2), (size_t)(op3)) +#define vmsif_m_b1(op0, op1) \ +__builtin_rvv_vmsif_m_b1((vbool1_t)(op0), (size_t)(op1)) +#define vmsif_m_b1_m(op0, op1, op2, op3) \ +__builtin_rvv_vmsif_m_b1_m((vbool1_t)(op0), (vbool1_t)(op1), (vbool1_t)(op2), (size_t)(op3)) +#define vmsif_m_b16(op0, op1) \ +__builtin_rvv_vmsif_m_b16((vbool16_t)(op0), (size_t)(op1)) +#define vmsif_m_b16_m(op0, op1, op2, op3) \ +__builtin_rvv_vmsif_m_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vmsif_m_b32(op0, op1) \ +__builtin_rvv_vmsif_m_b32((vbool32_t)(op0), (size_t)(op1)) +#define vmsif_m_b32_m(op0, op1, op2, op3) \ +__builtin_rvv_vmsif_m_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vmsif_m_b64(op0, op1) \ +__builtin_rvv_vmsif_m_b64((vbool64_t)(op0), (size_t)(op1)) +#define vmsif_m_b64_m(op0, op1, op2, op3) \ +__builtin_rvv_vmsif_m_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vmsof_m_b8(op0, op1) \ +__builtin_rvv_vmsof_m_b8((vbool8_t)(op0), (size_t)(op1)) +#define vmsof_m_b8_m(op0, op1, op2, op3) \ +__builtin_rvv_vmsof_m_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vbool8_t)(op2), (size_t)(op3)) +#define vmsof_m_b4(op0, op1) \ +__builtin_rvv_vmsof_m_b4((vbool4_t)(op0), (size_t)(op1)) +#define vmsof_m_b4_m(op0, op1, op2, op3) \ +__builtin_rvv_vmsof_m_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vbool4_t)(op2), (size_t)(op3)) +#define vmsof_m_b2(op0, op1) \ +__builtin_rvv_vmsof_m_b2((vbool2_t)(op0), (size_t)(op1)) +#define vmsof_m_b2_m(op0, op1, op2, op3) \ +__builtin_rvv_vmsof_m_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vbool2_t)(op2), (size_t)(op3)) +#define vmsof_m_b1(op0, op1) \ +__builtin_rvv_vmsof_m_b1((vbool1_t)(op0), (size_t)(op1)) +#define vmsof_m_b1_m(op0, op1, op2, op3) \ +__builtin_rvv_vmsof_m_b1_m((vbool1_t)(op0), (vbool1_t)(op1), (vbool1_t)(op2), (size_t)(op3)) +#define vmsof_m_b16(op0, op1) \ +__builtin_rvv_vmsof_m_b16((vbool16_t)(op0), (size_t)(op1)) +#define vmsof_m_b16_m(op0, op1, op2, op3) \ +__builtin_rvv_vmsof_m_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vbool16_t)(op2), (size_t)(op3)) +#define vmsof_m_b32(op0, op1) \ +__builtin_rvv_vmsof_m_b32((vbool32_t)(op0), (size_t)(op1)) +#define vmsof_m_b32_m(op0, op1, op2, op3) \ +__builtin_rvv_vmsof_m_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vbool32_t)(op2), (size_t)(op3)) +#define vmsof_m_b64(op0, op1) \ +__builtin_rvv_vmsof_m_b64((vbool64_t)(op0), (size_t)(op1)) +#define vmsof_m_b64_m(op0, op1, op2, op3) \ +__builtin_rvv_vmsof_m_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vbool64_t)(op2), (size_t)(op3)) +#define vmxnor_mm_b8(op0, op1, op2) \ +__builtin_rvv_vmxnor_mm_b8((vbool8_t)(op0), (vbool8_t)(op1), (size_t)(op2)) +#define vmxnor_mm_b4(op0, op1, op2) \ +__builtin_rvv_vmxnor_mm_b4((vbool4_t)(op0), (vbool4_t)(op1), (size_t)(op2)) +#define vmxnor_mm_b2(op0, op1, op2) \ +__builtin_rvv_vmxnor_mm_b2((vbool2_t)(op0), (vbool2_t)(op1), (size_t)(op2)) +#define vmxnor_mm_b1(op0, op1, op2) \ +__builtin_rvv_vmxnor_mm_b1((vbool1_t)(op0), (vbool1_t)(op1), (size_t)(op2)) +#define vmxnor_mm_b16(op0, op1, op2) \ +__builtin_rvv_vmxnor_mm_b16((vbool16_t)(op0), (vbool16_t)(op1), (size_t)(op2)) +#define vmxnor_mm_b32(op0, op1, op2) \ +__builtin_rvv_vmxnor_mm_b32((vbool32_t)(op0), (vbool32_t)(op1), (size_t)(op2)) +#define vmxnor_mm_b64(op0, op1, op2) \ +__builtin_rvv_vmxnor_mm_b64((vbool64_t)(op0), (vbool64_t)(op1), (size_t)(op2)) +#define vmxor_mm_b8(op0, op1, op2) \ +__builtin_rvv_vmxor_mm_b8((vbool8_t)(op0), (vbool8_t)(op1), (size_t)(op2)) +#define vmxor_mm_b4(op0, op1, op2) \ +__builtin_rvv_vmxor_mm_b4((vbool4_t)(op0), (vbool4_t)(op1), (size_t)(op2)) +#define vmxor_mm_b2(op0, op1, op2) \ +__builtin_rvv_vmxor_mm_b2((vbool2_t)(op0), (vbool2_t)(op1), (size_t)(op2)) +#define vmxor_mm_b1(op0, op1, op2) \ +__builtin_rvv_vmxor_mm_b1((vbool1_t)(op0), (vbool1_t)(op1), (size_t)(op2)) +#define vmxor_mm_b16(op0, op1, op2) \ +__builtin_rvv_vmxor_mm_b16((vbool16_t)(op0), (vbool16_t)(op1), (size_t)(op2)) +#define vmxor_mm_b32(op0, op1, op2) \ +__builtin_rvv_vmxor_mm_b32((vbool32_t)(op0), (vbool32_t)(op1), (size_t)(op2)) +#define vmxor_mm_b64(op0, op1, op2) \ +__builtin_rvv_vmxor_mm_b64((vbool64_t)(op0), (vbool64_t)(op1), (size_t)(op2)) +#define vpopc_m_b8(op0, op1) \ +__builtin_rvv_vpopc_m_b8((vbool8_t)(op0), (size_t)(op1)) +#define vpopc_m_b8_m(op0, op1, op2) \ +__builtin_rvv_vpopc_m_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (size_t)(op2)) +#define vpopc_m_b4(op0, op1) \ +__builtin_rvv_vpopc_m_b4((vbool4_t)(op0), (size_t)(op1)) +#define vpopc_m_b4_m(op0, op1, op2) \ +__builtin_rvv_vpopc_m_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (size_t)(op2)) +#define vpopc_m_b2(op0, op1) \ +__builtin_rvv_vpopc_m_b2((vbool2_t)(op0), (size_t)(op1)) +#define vpopc_m_b2_m(op0, op1, op2) \ +__builtin_rvv_vpopc_m_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (size_t)(op2)) +#define vpopc_m_b1(op0, op1) \ +__builtin_rvv_vpopc_m_b1((vbool1_t)(op0), (size_t)(op1)) +#define vpopc_m_b1_m(op0, op1, op2) \ +__builtin_rvv_vpopc_m_b1_m((vbool1_t)(op0), (vbool1_t)(op1), (size_t)(op2)) +#define vpopc_m_b16(op0, op1) \ +__builtin_rvv_vpopc_m_b16((vbool16_t)(op0), (size_t)(op1)) +#define vpopc_m_b16_m(op0, op1, op2) \ +__builtin_rvv_vpopc_m_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (size_t)(op2)) +#define vpopc_m_b32(op0, op1) \ +__builtin_rvv_vpopc_m_b32((vbool32_t)(op0), (size_t)(op1)) +#define vpopc_m_b32_m(op0, op1, op2) \ +__builtin_rvv_vpopc_m_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (size_t)(op2)) +#define vpopc_m_b64(op0, op1) \ +__builtin_rvv_vpopc_m_b64((vbool64_t)(op0), (size_t)(op1)) +#define vpopc_m_b64_m(op0, op1, op2) \ +__builtin_rvv_vpopc_m_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (size_t)(op2)) +#define vreinterpret_v_i8m1_i16m1(op0) \ +__builtin_rvv_vreinterpret_v_i8m1_i16m1((vint8m1_t)(op0)) +#define vreinterpret_v_i8m2_i16m2(op0) \ +__builtin_rvv_vreinterpret_v_i8m2_i16m2((vint8m2_t)(op0)) +#define vreinterpret_v_i8m4_i16m4(op0) \ +__builtin_rvv_vreinterpret_v_i8m4_i16m4((vint8m4_t)(op0)) +#define vreinterpret_v_i8m8_i16m8(op0) \ +__builtin_rvv_vreinterpret_v_i8m8_i16m8((vint8m8_t)(op0)) +#define vreinterpret_v_i8mf2_i16mf2(op0) \ +__builtin_rvv_vreinterpret_v_i8mf2_i16mf2((vint8mf2_t)(op0)) +#define vreinterpret_v_i8mf4_i16mf4(op0) \ +__builtin_rvv_vreinterpret_v_i8mf4_i16mf4((vint8mf4_t)(op0)) +#define vreinterpret_v_i32m1_i16m1(op0) \ +__builtin_rvv_vreinterpret_v_i32m1_i16m1((vint32m1_t)(op0)) +#define vreinterpret_v_i32m2_i16m2(op0) \ +__builtin_rvv_vreinterpret_v_i32m2_i16m2((vint32m2_t)(op0)) +#define vreinterpret_v_i32m4_i16m4(op0) \ +__builtin_rvv_vreinterpret_v_i32m4_i16m4((vint32m4_t)(op0)) +#define vreinterpret_v_i32m8_i16m8(op0) \ +__builtin_rvv_vreinterpret_v_i32m8_i16m8((vint32m8_t)(op0)) +#define vreinterpret_v_i32mf2_i16mf2(op0) \ +__builtin_rvv_vreinterpret_v_i32mf2_i16mf2((vint32mf2_t)(op0)) +#define vreinterpret_v_i64m1_i16m1(op0) \ +__builtin_rvv_vreinterpret_v_i64m1_i16m1((vint64m1_t)(op0)) +#define vreinterpret_v_i64m2_i16m2(op0) \ +__builtin_rvv_vreinterpret_v_i64m2_i16m2((vint64m2_t)(op0)) +#define vreinterpret_v_i64m4_i16m4(op0) \ +__builtin_rvv_vreinterpret_v_i64m4_i16m4((vint64m4_t)(op0)) +#define vreinterpret_v_i64m8_i16m8(op0) \ +__builtin_rvv_vreinterpret_v_i64m8_i16m8((vint64m8_t)(op0)) +#define vreinterpret_v_i8m1_i32m1(op0) \ +__builtin_rvv_vreinterpret_v_i8m1_i32m1((vint8m1_t)(op0)) +#define vreinterpret_v_i8m2_i32m2(op0) \ +__builtin_rvv_vreinterpret_v_i8m2_i32m2((vint8m2_t)(op0)) +#define vreinterpret_v_i8m4_i32m4(op0) \ +__builtin_rvv_vreinterpret_v_i8m4_i32m4((vint8m4_t)(op0)) +#define vreinterpret_v_i8m8_i32m8(op0) \ +__builtin_rvv_vreinterpret_v_i8m8_i32m8((vint8m8_t)(op0)) +#define vreinterpret_v_i8mf2_i32mf2(op0) \ +__builtin_rvv_vreinterpret_v_i8mf2_i32mf2((vint8mf2_t)(op0)) +#define vreinterpret_v_i16m1_i32m1(op0) \ +__builtin_rvv_vreinterpret_v_i16m1_i32m1((vint16m1_t)(op0)) +#define vreinterpret_v_i16m2_i32m2(op0) \ +__builtin_rvv_vreinterpret_v_i16m2_i32m2((vint16m2_t)(op0)) +#define vreinterpret_v_i16m4_i32m4(op0) \ +__builtin_rvv_vreinterpret_v_i16m4_i32m4((vint16m4_t)(op0)) +#define vreinterpret_v_i16m8_i32m8(op0) \ +__builtin_rvv_vreinterpret_v_i16m8_i32m8((vint16m8_t)(op0)) +#define vreinterpret_v_i16mf2_i32mf2(op0) \ +__builtin_rvv_vreinterpret_v_i16mf2_i32mf2((vint16mf2_t)(op0)) +#define vreinterpret_v_i64m1_i32m1(op0) \ +__builtin_rvv_vreinterpret_v_i64m1_i32m1((vint64m1_t)(op0)) +#define vreinterpret_v_i64m2_i32m2(op0) \ +__builtin_rvv_vreinterpret_v_i64m2_i32m2((vint64m2_t)(op0)) +#define vreinterpret_v_i64m4_i32m4(op0) \ +__builtin_rvv_vreinterpret_v_i64m4_i32m4((vint64m4_t)(op0)) +#define vreinterpret_v_i64m8_i32m8(op0) \ +__builtin_rvv_vreinterpret_v_i64m8_i32m8((vint64m8_t)(op0)) +#define vreinterpret_v_i8m1_i64m1(op0) \ +__builtin_rvv_vreinterpret_v_i8m1_i64m1((vint8m1_t)(op0)) +#define vreinterpret_v_i8m2_i64m2(op0) \ +__builtin_rvv_vreinterpret_v_i8m2_i64m2((vint8m2_t)(op0)) +#define vreinterpret_v_i8m4_i64m4(op0) \ +__builtin_rvv_vreinterpret_v_i8m4_i64m4((vint8m4_t)(op0)) +#define vreinterpret_v_i8m8_i64m8(op0) \ +__builtin_rvv_vreinterpret_v_i8m8_i64m8((vint8m8_t)(op0)) +#define vreinterpret_v_i16m1_i64m1(op0) \ +__builtin_rvv_vreinterpret_v_i16m1_i64m1((vint16m1_t)(op0)) +#define vreinterpret_v_i16m2_i64m2(op0) \ +__builtin_rvv_vreinterpret_v_i16m2_i64m2((vint16m2_t)(op0)) +#define vreinterpret_v_i16m4_i64m4(op0) \ +__builtin_rvv_vreinterpret_v_i16m4_i64m4((vint16m4_t)(op0)) +#define vreinterpret_v_i16m8_i64m8(op0) \ +__builtin_rvv_vreinterpret_v_i16m8_i64m8((vint16m8_t)(op0)) +#define vreinterpret_v_i32m1_i64m1(op0) \ +__builtin_rvv_vreinterpret_v_i32m1_i64m1((vint32m1_t)(op0)) +#define vreinterpret_v_i32m2_i64m2(op0) \ +__builtin_rvv_vreinterpret_v_i32m2_i64m2((vint32m2_t)(op0)) +#define vreinterpret_v_i32m4_i64m4(op0) \ +__builtin_rvv_vreinterpret_v_i32m4_i64m4((vint32m4_t)(op0)) +#define vreinterpret_v_i32m8_i64m8(op0) \ +__builtin_rvv_vreinterpret_v_i32m8_i64m8((vint32m8_t)(op0)) +#define vreinterpret_v_i16m1_i8m1(op0) \ +__builtin_rvv_vreinterpret_v_i16m1_i8m1((vint16m1_t)(op0)) +#define vreinterpret_v_i16m2_i8m2(op0) \ +__builtin_rvv_vreinterpret_v_i16m2_i8m2((vint16m2_t)(op0)) +#define vreinterpret_v_i16m4_i8m4(op0) \ +__builtin_rvv_vreinterpret_v_i16m4_i8m4((vint16m4_t)(op0)) +#define vreinterpret_v_i16m8_i8m8(op0) \ +__builtin_rvv_vreinterpret_v_i16m8_i8m8((vint16m8_t)(op0)) +#define vreinterpret_v_i16mf2_i8mf2(op0) \ +__builtin_rvv_vreinterpret_v_i16mf2_i8mf2((vint16mf2_t)(op0)) +#define vreinterpret_v_i16mf4_i8mf4(op0) \ +__builtin_rvv_vreinterpret_v_i16mf4_i8mf4((vint16mf4_t)(op0)) +#define vreinterpret_v_i32m1_i8m1(op0) \ +__builtin_rvv_vreinterpret_v_i32m1_i8m1((vint32m1_t)(op0)) +#define vreinterpret_v_i32m2_i8m2(op0) \ +__builtin_rvv_vreinterpret_v_i32m2_i8m2((vint32m2_t)(op0)) +#define vreinterpret_v_i32m4_i8m4(op0) \ +__builtin_rvv_vreinterpret_v_i32m4_i8m4((vint32m4_t)(op0)) +#define vreinterpret_v_i32m8_i8m8(op0) \ +__builtin_rvv_vreinterpret_v_i32m8_i8m8((vint32m8_t)(op0)) +#define vreinterpret_v_i32mf2_i8mf2(op0) \ +__builtin_rvv_vreinterpret_v_i32mf2_i8mf2((vint32mf2_t)(op0)) +#define vreinterpret_v_i64m1_i8m1(op0) \ +__builtin_rvv_vreinterpret_v_i64m1_i8m1((vint64m1_t)(op0)) +#define vreinterpret_v_i64m2_i8m2(op0) \ +__builtin_rvv_vreinterpret_v_i64m2_i8m2((vint64m2_t)(op0)) +#define vreinterpret_v_i64m4_i8m4(op0) \ +__builtin_rvv_vreinterpret_v_i64m4_i8m4((vint64m4_t)(op0)) +#define vreinterpret_v_i64m8_i8m8(op0) \ +__builtin_rvv_vreinterpret_v_i64m8_i8m8((vint64m8_t)(op0)) +#define vreinterpret_v_u8m1_i8m1(op0) \ +__builtin_rvv_vreinterpret_v_u8m1_i8m1((vuint8m1_t)(op0)) +#define vreinterpret_v_u8m2_i8m2(op0) \ +__builtin_rvv_vreinterpret_v_u8m2_i8m2((vuint8m2_t)(op0)) +#define vreinterpret_v_u8m4_i8m4(op0) \ +__builtin_rvv_vreinterpret_v_u8m4_i8m4((vuint8m4_t)(op0)) +#define vreinterpret_v_u8m8_i8m8(op0) \ +__builtin_rvv_vreinterpret_v_u8m8_i8m8((vuint8m8_t)(op0)) +#define vreinterpret_v_u8mf2_i8mf2(op0) \ +__builtin_rvv_vreinterpret_v_u8mf2_i8mf2((vuint8mf2_t)(op0)) +#define vreinterpret_v_u8mf4_i8mf4(op0) \ +__builtin_rvv_vreinterpret_v_u8mf4_i8mf4((vuint8mf4_t)(op0)) +#define vreinterpret_v_u8mf8_i8mf8(op0) \ +__builtin_rvv_vreinterpret_v_u8mf8_i8mf8((vuint8mf8_t)(op0)) +#define vreinterpret_v_u16m1_i16m1(op0) \ +__builtin_rvv_vreinterpret_v_u16m1_i16m1((vuint16m1_t)(op0)) +#define vreinterpret_v_u16m2_i16m2(op0) \ +__builtin_rvv_vreinterpret_v_u16m2_i16m2((vuint16m2_t)(op0)) +#define vreinterpret_v_u16m4_i16m4(op0) \ +__builtin_rvv_vreinterpret_v_u16m4_i16m4((vuint16m4_t)(op0)) +#define vreinterpret_v_u16m8_i16m8(op0) \ +__builtin_rvv_vreinterpret_v_u16m8_i16m8((vuint16m8_t)(op0)) +#define vreinterpret_v_u16mf2_i16mf2(op0) \ +__builtin_rvv_vreinterpret_v_u16mf2_i16mf2((vuint16mf2_t)(op0)) +#define vreinterpret_v_u16mf4_i16mf4(op0) \ +__builtin_rvv_vreinterpret_v_u16mf4_i16mf4((vuint16mf4_t)(op0)) +#define vreinterpret_v_u32m1_i32m1(op0) \ +__builtin_rvv_vreinterpret_v_u32m1_i32m1((vuint32m1_t)(op0)) +#define vreinterpret_v_u32m2_i32m2(op0) \ +__builtin_rvv_vreinterpret_v_u32m2_i32m2((vuint32m2_t)(op0)) +#define vreinterpret_v_u32m4_i32m4(op0) \ +__builtin_rvv_vreinterpret_v_u32m4_i32m4((vuint32m4_t)(op0)) +#define vreinterpret_v_u32m8_i32m8(op0) \ +__builtin_rvv_vreinterpret_v_u32m8_i32m8((vuint32m8_t)(op0)) +#define vreinterpret_v_u32mf2_i32mf2(op0) \ +__builtin_rvv_vreinterpret_v_u32mf2_i32mf2((vuint32mf2_t)(op0)) +#define vreinterpret_v_u64m1_i64m1(op0) \ +__builtin_rvv_vreinterpret_v_u64m1_i64m1((vuint64m1_t)(op0)) +#define vreinterpret_v_u64m2_i64m2(op0) \ +__builtin_rvv_vreinterpret_v_u64m2_i64m2((vuint64m2_t)(op0)) +#define vreinterpret_v_u64m4_i64m4(op0) \ +__builtin_rvv_vreinterpret_v_u64m4_i64m4((vuint64m4_t)(op0)) +#define vreinterpret_v_u64m8_i64m8(op0) \ +__builtin_rvv_vreinterpret_v_u64m8_i64m8((vuint64m8_t)(op0)) +#define vreinterpret_v_u8m1_u16m1(op0) \ +__builtin_rvv_vreinterpret_v_u8m1_u16m1((vuint8m1_t)(op0)) +#define vreinterpret_v_u8m2_u16m2(op0) \ +__builtin_rvv_vreinterpret_v_u8m2_u16m2((vuint8m2_t)(op0)) +#define vreinterpret_v_u8m4_u16m4(op0) \ +__builtin_rvv_vreinterpret_v_u8m4_u16m4((vuint8m4_t)(op0)) +#define vreinterpret_v_u8m8_u16m8(op0) \ +__builtin_rvv_vreinterpret_v_u8m8_u16m8((vuint8m8_t)(op0)) +#define vreinterpret_v_u8mf2_u16mf2(op0) \ +__builtin_rvv_vreinterpret_v_u8mf2_u16mf2((vuint8mf2_t)(op0)) +#define vreinterpret_v_u8mf4_u16mf4(op0) \ +__builtin_rvv_vreinterpret_v_u8mf4_u16mf4((vuint8mf4_t)(op0)) +#define vreinterpret_v_u32m1_u16m1(op0) \ +__builtin_rvv_vreinterpret_v_u32m1_u16m1((vuint32m1_t)(op0)) +#define vreinterpret_v_u32m2_u16m2(op0) \ +__builtin_rvv_vreinterpret_v_u32m2_u16m2((vuint32m2_t)(op0)) +#define vreinterpret_v_u32m4_u16m4(op0) \ +__builtin_rvv_vreinterpret_v_u32m4_u16m4((vuint32m4_t)(op0)) +#define vreinterpret_v_u32m8_u16m8(op0) \ +__builtin_rvv_vreinterpret_v_u32m8_u16m8((vuint32m8_t)(op0)) +#define vreinterpret_v_u32mf2_u16mf2(op0) \ +__builtin_rvv_vreinterpret_v_u32mf2_u16mf2((vuint32mf2_t)(op0)) +#define vreinterpret_v_u64m1_u16m1(op0) \ +__builtin_rvv_vreinterpret_v_u64m1_u16m1((vuint64m1_t)(op0)) +#define vreinterpret_v_u64m2_u16m2(op0) \ +__builtin_rvv_vreinterpret_v_u64m2_u16m2((vuint64m2_t)(op0)) +#define vreinterpret_v_u64m4_u16m4(op0) \ +__builtin_rvv_vreinterpret_v_u64m4_u16m4((vuint64m4_t)(op0)) +#define vreinterpret_v_u64m8_u16m8(op0) \ +__builtin_rvv_vreinterpret_v_u64m8_u16m8((vuint64m8_t)(op0)) +#define vreinterpret_v_u8m1_u32m1(op0) \ +__builtin_rvv_vreinterpret_v_u8m1_u32m1((vuint8m1_t)(op0)) +#define vreinterpret_v_u8m2_u32m2(op0) \ +__builtin_rvv_vreinterpret_v_u8m2_u32m2((vuint8m2_t)(op0)) +#define vreinterpret_v_u8m4_u32m4(op0) \ +__builtin_rvv_vreinterpret_v_u8m4_u32m4((vuint8m4_t)(op0)) +#define vreinterpret_v_u8m8_u32m8(op0) \ +__builtin_rvv_vreinterpret_v_u8m8_u32m8((vuint8m8_t)(op0)) +#define vreinterpret_v_u8mf2_u32mf2(op0) \ +__builtin_rvv_vreinterpret_v_u8mf2_u32mf2((vuint8mf2_t)(op0)) +#define vreinterpret_v_u16m1_u32m1(op0) \ +__builtin_rvv_vreinterpret_v_u16m1_u32m1((vuint16m1_t)(op0)) +#define vreinterpret_v_u16m2_u32m2(op0) \ +__builtin_rvv_vreinterpret_v_u16m2_u32m2((vuint16m2_t)(op0)) +#define vreinterpret_v_u16m4_u32m4(op0) \ +__builtin_rvv_vreinterpret_v_u16m4_u32m4((vuint16m4_t)(op0)) +#define vreinterpret_v_u16m8_u32m8(op0) \ +__builtin_rvv_vreinterpret_v_u16m8_u32m8((vuint16m8_t)(op0)) +#define vreinterpret_v_u16mf2_u32mf2(op0) \ +__builtin_rvv_vreinterpret_v_u16mf2_u32mf2((vuint16mf2_t)(op0)) +#define vreinterpret_v_u64m1_u32m1(op0) \ +__builtin_rvv_vreinterpret_v_u64m1_u32m1((vuint64m1_t)(op0)) +#define vreinterpret_v_u64m2_u32m2(op0) \ +__builtin_rvv_vreinterpret_v_u64m2_u32m2((vuint64m2_t)(op0)) +#define vreinterpret_v_u64m4_u32m4(op0) \ +__builtin_rvv_vreinterpret_v_u64m4_u32m4((vuint64m4_t)(op0)) +#define vreinterpret_v_u64m8_u32m8(op0) \ +__builtin_rvv_vreinterpret_v_u64m8_u32m8((vuint64m8_t)(op0)) +#define vreinterpret_v_u8m1_u64m1(op0) \ +__builtin_rvv_vreinterpret_v_u8m1_u64m1((vuint8m1_t)(op0)) +#define vreinterpret_v_u8m2_u64m2(op0) \ +__builtin_rvv_vreinterpret_v_u8m2_u64m2((vuint8m2_t)(op0)) +#define vreinterpret_v_u8m4_u64m4(op0) \ +__builtin_rvv_vreinterpret_v_u8m4_u64m4((vuint8m4_t)(op0)) +#define vreinterpret_v_u8m8_u64m8(op0) \ +__builtin_rvv_vreinterpret_v_u8m8_u64m8((vuint8m8_t)(op0)) +#define vreinterpret_v_u16m1_u64m1(op0) \ +__builtin_rvv_vreinterpret_v_u16m1_u64m1((vuint16m1_t)(op0)) +#define vreinterpret_v_u16m2_u64m2(op0) \ +__builtin_rvv_vreinterpret_v_u16m2_u64m2((vuint16m2_t)(op0)) +#define vreinterpret_v_u16m4_u64m4(op0) \ +__builtin_rvv_vreinterpret_v_u16m4_u64m4((vuint16m4_t)(op0)) +#define vreinterpret_v_u16m8_u64m8(op0) \ +__builtin_rvv_vreinterpret_v_u16m8_u64m8((vuint16m8_t)(op0)) +#define vreinterpret_v_u32m1_u64m1(op0) \ +__builtin_rvv_vreinterpret_v_u32m1_u64m1((vuint32m1_t)(op0)) +#define vreinterpret_v_u32m2_u64m2(op0) \ +__builtin_rvv_vreinterpret_v_u32m2_u64m2((vuint32m2_t)(op0)) +#define vreinterpret_v_u32m4_u64m4(op0) \ +__builtin_rvv_vreinterpret_v_u32m4_u64m4((vuint32m4_t)(op0)) +#define vreinterpret_v_u32m8_u64m8(op0) \ +__builtin_rvv_vreinterpret_v_u32m8_u64m8((vuint32m8_t)(op0)) +#define vreinterpret_v_u16m1_u8m1(op0) \ +__builtin_rvv_vreinterpret_v_u16m1_u8m1((vuint16m1_t)(op0)) +#define vreinterpret_v_u16m2_u8m2(op0) \ +__builtin_rvv_vreinterpret_v_u16m2_u8m2((vuint16m2_t)(op0)) +#define vreinterpret_v_u16m4_u8m4(op0) \ +__builtin_rvv_vreinterpret_v_u16m4_u8m4((vuint16m4_t)(op0)) +#define vreinterpret_v_u16m8_u8m8(op0) \ +__builtin_rvv_vreinterpret_v_u16m8_u8m8((vuint16m8_t)(op0)) +#define vreinterpret_v_u16mf2_u8mf2(op0) \ +__builtin_rvv_vreinterpret_v_u16mf2_u8mf2((vuint16mf2_t)(op0)) +#define vreinterpret_v_u16mf4_u8mf4(op0) \ +__builtin_rvv_vreinterpret_v_u16mf4_u8mf4((vuint16mf4_t)(op0)) +#define vreinterpret_v_u32m1_u8m1(op0) \ +__builtin_rvv_vreinterpret_v_u32m1_u8m1((vuint32m1_t)(op0)) +#define vreinterpret_v_u32m2_u8m2(op0) \ +__builtin_rvv_vreinterpret_v_u32m2_u8m2((vuint32m2_t)(op0)) +#define vreinterpret_v_u32m4_u8m4(op0) \ +__builtin_rvv_vreinterpret_v_u32m4_u8m4((vuint32m4_t)(op0)) +#define vreinterpret_v_u32m8_u8m8(op0) \ +__builtin_rvv_vreinterpret_v_u32m8_u8m8((vuint32m8_t)(op0)) +#define vreinterpret_v_u32mf2_u8mf2(op0) \ +__builtin_rvv_vreinterpret_v_u32mf2_u8mf2((vuint32mf2_t)(op0)) +#define vreinterpret_v_u64m1_u8m1(op0) \ +__builtin_rvv_vreinterpret_v_u64m1_u8m1((vuint64m1_t)(op0)) +#define vreinterpret_v_u64m2_u8m2(op0) \ +__builtin_rvv_vreinterpret_v_u64m2_u8m2((vuint64m2_t)(op0)) +#define vreinterpret_v_u64m4_u8m4(op0) \ +__builtin_rvv_vreinterpret_v_u64m4_u8m4((vuint64m4_t)(op0)) +#define vreinterpret_v_u64m8_u8m8(op0) \ +__builtin_rvv_vreinterpret_v_u64m8_u8m8((vuint64m8_t)(op0)) +#define vreinterpret_v_i8m1_u8m1(op0) \ +__builtin_rvv_vreinterpret_v_i8m1_u8m1((vint8m1_t)(op0)) +#define vreinterpret_v_i8m2_u8m2(op0) \ +__builtin_rvv_vreinterpret_v_i8m2_u8m2((vint8m2_t)(op0)) +#define vreinterpret_v_i8m4_u8m4(op0) \ +__builtin_rvv_vreinterpret_v_i8m4_u8m4((vint8m4_t)(op0)) +#define vreinterpret_v_i8m8_u8m8(op0) \ +__builtin_rvv_vreinterpret_v_i8m8_u8m8((vint8m8_t)(op0)) +#define vreinterpret_v_i8mf2_u8mf2(op0) \ +__builtin_rvv_vreinterpret_v_i8mf2_u8mf2((vint8mf2_t)(op0)) +#define vreinterpret_v_i8mf4_u8mf4(op0) \ +__builtin_rvv_vreinterpret_v_i8mf4_u8mf4((vint8mf4_t)(op0)) +#define vreinterpret_v_i8mf8_u8mf8(op0) \ +__builtin_rvv_vreinterpret_v_i8mf8_u8mf8((vint8mf8_t)(op0)) +#define vreinterpret_v_i16m1_u16m1(op0) \ +__builtin_rvv_vreinterpret_v_i16m1_u16m1((vint16m1_t)(op0)) +#define vreinterpret_v_i16m2_u16m2(op0) \ +__builtin_rvv_vreinterpret_v_i16m2_u16m2((vint16m2_t)(op0)) +#define vreinterpret_v_i16m4_u16m4(op0) \ +__builtin_rvv_vreinterpret_v_i16m4_u16m4((vint16m4_t)(op0)) +#define vreinterpret_v_i16m8_u16m8(op0) \ +__builtin_rvv_vreinterpret_v_i16m8_u16m8((vint16m8_t)(op0)) +#define vreinterpret_v_i16mf2_u16mf2(op0) \ +__builtin_rvv_vreinterpret_v_i16mf2_u16mf2((vint16mf2_t)(op0)) +#define vreinterpret_v_i16mf4_u16mf4(op0) \ +__builtin_rvv_vreinterpret_v_i16mf4_u16mf4((vint16mf4_t)(op0)) +#define vreinterpret_v_i32m1_u32m1(op0) \ +__builtin_rvv_vreinterpret_v_i32m1_u32m1((vint32m1_t)(op0)) +#define vreinterpret_v_i32m2_u32m2(op0) \ +__builtin_rvv_vreinterpret_v_i32m2_u32m2((vint32m2_t)(op0)) +#define vreinterpret_v_i32m4_u32m4(op0) \ +__builtin_rvv_vreinterpret_v_i32m4_u32m4((vint32m4_t)(op0)) +#define vreinterpret_v_i32m8_u32m8(op0) \ +__builtin_rvv_vreinterpret_v_i32m8_u32m8((vint32m8_t)(op0)) +#define vreinterpret_v_i32mf2_u32mf2(op0) \ +__builtin_rvv_vreinterpret_v_i32mf2_u32mf2((vint32mf2_t)(op0)) +#define vreinterpret_v_i64m1_u64m1(op0) \ +__builtin_rvv_vreinterpret_v_i64m1_u64m1((vint64m1_t)(op0)) +#define vreinterpret_v_i64m2_u64m2(op0) \ +__builtin_rvv_vreinterpret_v_i64m2_u64m2((vint64m2_t)(op0)) +#define vreinterpret_v_i64m4_u64m4(op0) \ +__builtin_rvv_vreinterpret_v_i64m4_u64m4((vint64m4_t)(op0)) +#define vreinterpret_v_i64m8_u64m8(op0) \ +__builtin_rvv_vreinterpret_v_i64m8_u64m8((vint64m8_t)(op0)) +#define vse1_v_b8(op0, op1, op2) \ +__builtin_rvv_vse1_v_b8((uint8_t *)(op0), (vbool8_t)(op1), (size_t)(op2)) +#define vse1_v_b4(op0, op1, op2) \ +__builtin_rvv_vse1_v_b4((uint8_t *)(op0), (vbool4_t)(op1), (size_t)(op2)) +#define vse1_v_b2(op0, op1, op2) \ +__builtin_rvv_vse1_v_b2((uint8_t *)(op0), (vbool2_t)(op1), (size_t)(op2)) +#define vse1_v_b1(op0, op1, op2) \ +__builtin_rvv_vse1_v_b1((uint8_t *)(op0), (vbool1_t)(op1), (size_t)(op2)) +#define vse1_v_b16(op0, op1, op2) \ +__builtin_rvv_vse1_v_b16((uint8_t *)(op0), (vbool16_t)(op1), (size_t)(op2)) +#define vse1_v_b32(op0, op1, op2) \ +__builtin_rvv_vse1_v_b32((uint8_t *)(op0), (vbool32_t)(op1), (size_t)(op2)) +#define vse1_v_b64(op0, op1, op2) \ +__builtin_rvv_vse1_v_b64((uint8_t *)(op0), (vbool64_t)(op1), (size_t)(op2)) +#define vsext_vf2_i16mf4(op0, op1) \ +__builtin_rvv_vsext_vf2_i16mf4((vint8mf8_t)(op0), (size_t)(op1)) +#define vsext_vf2_i16mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vsext_vf2_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vint8mf8_t)(op2), (size_t)(op3)) +#define vsext_vf2_i16mf2(op0, op1) \ +__builtin_rvv_vsext_vf2_i16mf2((vint8mf4_t)(op0), (size_t)(op1)) +#define vsext_vf2_i16mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vsext_vf2_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vint8mf4_t)(op2), (size_t)(op3)) +#define vsext_vf2_i16m1(op0, op1) \ +__builtin_rvv_vsext_vf2_i16m1((vint8mf2_t)(op0), (size_t)(op1)) +#define vsext_vf2_i16m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vsext_vf2_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vint8mf2_t)(op2), (size_t)(op3)) +#define vsext_vf2_i16m2(op0, op1) \ +__builtin_rvv_vsext_vf2_i16m2((vint8m1_t)(op0), (size_t)(op1)) +#define vsext_vf2_i16m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vsext_vf2_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vsext_vf2_i16m4(op0, op1) \ +__builtin_rvv_vsext_vf2_i16m4((vint8m2_t)(op0), (size_t)(op1)) +#define vsext_vf2_i16m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vsext_vf2_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vint8m2_t)(op2), (size_t)(op3)) +#define vsext_vf2_i16m8(op0, op1) \ +__builtin_rvv_vsext_vf2_i16m8((vint8m4_t)(op0), (size_t)(op1)) +#define vsext_vf2_i16m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vsext_vf2_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vint8m4_t)(op2), (size_t)(op3)) +#define vsext_vf2_i32mf2(op0, op1) \ +__builtin_rvv_vsext_vf2_i32mf2((vint16mf4_t)(op0), (size_t)(op1)) +#define vsext_vf2_i32mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vsext_vf2_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint16mf4_t)(op2), (size_t)(op3)) +#define vsext_vf2_i32m1(op0, op1) \ +__builtin_rvv_vsext_vf2_i32m1((vint16mf2_t)(op0), (size_t)(op1)) +#define vsext_vf2_i32m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vsext_vf2_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint16mf2_t)(op2), (size_t)(op3)) +#define vsext_vf2_i32m2(op0, op1) \ +__builtin_rvv_vsext_vf2_i32m2((vint16m1_t)(op0), (size_t)(op1)) +#define vsext_vf2_i32m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vsext_vf2_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vsext_vf2_i32m4(op0, op1) \ +__builtin_rvv_vsext_vf2_i32m4((vint16m2_t)(op0), (size_t)(op1)) +#define vsext_vf2_i32m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vsext_vf2_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint16m2_t)(op2), (size_t)(op3)) +#define vsext_vf2_i32m8(op0, op1) \ +__builtin_rvv_vsext_vf2_i32m8((vint16m4_t)(op0), (size_t)(op1)) +#define vsext_vf2_i32m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vsext_vf2_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint16m4_t)(op2), (size_t)(op3)) +#define vsext_vf2_i64m1(op0, op1) \ +__builtin_rvv_vsext_vf2_i64m1((vint32mf2_t)(op0), (size_t)(op1)) +#define vsext_vf2_i64m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vsext_vf2_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vsext_vf2_i64m2(op0, op1) \ +__builtin_rvv_vsext_vf2_i64m2((vint32m1_t)(op0), (size_t)(op1)) +#define vsext_vf2_i64m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vsext_vf2_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vsext_vf2_i64m4(op0, op1) \ +__builtin_rvv_vsext_vf2_i64m4((vint32m2_t)(op0), (size_t)(op1)) +#define vsext_vf2_i64m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vsext_vf2_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vsext_vf2_i64m8(op0, op1) \ +__builtin_rvv_vsext_vf2_i64m8((vint32m4_t)(op0), (size_t)(op1)) +#define vsext_vf2_i64m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vsext_vf2_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vsext_vf4_i32mf2(op0, op1) \ +__builtin_rvv_vsext_vf4_i32mf2((vint8mf8_t)(op0), (size_t)(op1)) +#define vsext_vf4_i32mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vsext_vf4_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vint8mf8_t)(op2), (size_t)(op3)) +#define vsext_vf4_i32m1(op0, op1) \ +__builtin_rvv_vsext_vf4_i32m1((vint8mf4_t)(op0), (size_t)(op1)) +#define vsext_vf4_i32m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vsext_vf4_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vint8mf4_t)(op2), (size_t)(op3)) +#define vsext_vf4_i32m2(op0, op1) \ +__builtin_rvv_vsext_vf4_i32m2((vint8mf2_t)(op0), (size_t)(op1)) +#define vsext_vf4_i32m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vsext_vf4_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vint8mf2_t)(op2), (size_t)(op3)) +#define vsext_vf4_i32m4(op0, op1) \ +__builtin_rvv_vsext_vf4_i32m4((vint8m1_t)(op0), (size_t)(op1)) +#define vsext_vf4_i32m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vsext_vf4_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vsext_vf4_i32m8(op0, op1) \ +__builtin_rvv_vsext_vf4_i32m8((vint8m2_t)(op0), (size_t)(op1)) +#define vsext_vf4_i32m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vsext_vf4_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vint8m2_t)(op2), (size_t)(op3)) +#define vsext_vf4_i64m1(op0, op1) \ +__builtin_rvv_vsext_vf4_i64m1((vint16mf4_t)(op0), (size_t)(op1)) +#define vsext_vf4_i64m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vsext_vf4_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint16mf4_t)(op2), (size_t)(op3)) +#define vsext_vf4_i64m2(op0, op1) \ +__builtin_rvv_vsext_vf4_i64m2((vint16mf2_t)(op0), (size_t)(op1)) +#define vsext_vf4_i64m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vsext_vf4_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint16mf2_t)(op2), (size_t)(op3)) +#define vsext_vf4_i64m4(op0, op1) \ +__builtin_rvv_vsext_vf4_i64m4((vint16m1_t)(op0), (size_t)(op1)) +#define vsext_vf4_i64m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vsext_vf4_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vsext_vf4_i64m8(op0, op1) \ +__builtin_rvv_vsext_vf4_i64m8((vint16m2_t)(op0), (size_t)(op1)) +#define vsext_vf4_i64m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vsext_vf4_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint16m2_t)(op2), (size_t)(op3)) +#define vsext_vf8_i64m1(op0, op1) \ +__builtin_rvv_vsext_vf8_i64m1((vint8mf8_t)(op0), (size_t)(op1)) +#define vsext_vf8_i64m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vsext_vf8_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vint8mf8_t)(op2), (size_t)(op3)) +#define vsext_vf8_i64m2(op0, op1) \ +__builtin_rvv_vsext_vf8_i64m2((vint8mf4_t)(op0), (size_t)(op1)) +#define vsext_vf8_i64m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vsext_vf8_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vint8mf4_t)(op2), (size_t)(op3)) +#define vsext_vf8_i64m4(op0, op1) \ +__builtin_rvv_vsext_vf8_i64m4((vint8mf2_t)(op0), (size_t)(op1)) +#define vsext_vf8_i64m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vsext_vf8_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vint8mf2_t)(op2), (size_t)(op3)) +#define vsext_vf8_i64m8(op0, op1) \ +__builtin_rvv_vsext_vf8_i64m8((vint8m1_t)(op0), (size_t)(op1)) +#define vsext_vf8_i64m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vsext_vf8_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vundefined_i8m1() \ +__builtin_rvv_vundefined_i8m1() +#define vundefined_i8m2() \ +__builtin_rvv_vundefined_i8m2() +#define vundefined_i8m4() \ +__builtin_rvv_vundefined_i8m4() +#define vundefined_i8m8() \ +__builtin_rvv_vundefined_i8m8() +#define vundefined_i8mf2() \ +__builtin_rvv_vundefined_i8mf2() +#define vundefined_i8mf4() \ +__builtin_rvv_vundefined_i8mf4() +#define vundefined_i8mf8() \ +__builtin_rvv_vundefined_i8mf8() +#define vundefined_i16m1() \ +__builtin_rvv_vundefined_i16m1() +#define vundefined_i16m2() \ +__builtin_rvv_vundefined_i16m2() +#define vundefined_i16m4() \ +__builtin_rvv_vundefined_i16m4() +#define vundefined_i16m8() \ +__builtin_rvv_vundefined_i16m8() +#define vundefined_i16mf2() \ +__builtin_rvv_vundefined_i16mf2() +#define vundefined_i16mf4() \ +__builtin_rvv_vundefined_i16mf4() +#define vundefined_i32m1() \ +__builtin_rvv_vundefined_i32m1() +#define vundefined_i32m2() \ +__builtin_rvv_vundefined_i32m2() +#define vundefined_i32m4() \ +__builtin_rvv_vundefined_i32m4() +#define vundefined_i32m8() \ +__builtin_rvv_vundefined_i32m8() +#define vundefined_i32mf2() \ +__builtin_rvv_vundefined_i32mf2() +#define vundefined_i64m1() \ +__builtin_rvv_vundefined_i64m1() +#define vundefined_i64m2() \ +__builtin_rvv_vundefined_i64m2() +#define vundefined_i64m4() \ +__builtin_rvv_vundefined_i64m4() +#define vundefined_i64m8() \ +__builtin_rvv_vundefined_i64m8() +#define vundefined_u8m1() \ +__builtin_rvv_vundefined_u8m1() +#define vundefined_u8m2() \ +__builtin_rvv_vundefined_u8m2() +#define vundefined_u8m4() \ +__builtin_rvv_vundefined_u8m4() +#define vundefined_u8m8() \ +__builtin_rvv_vundefined_u8m8() +#define vundefined_u8mf2() \ +__builtin_rvv_vundefined_u8mf2() +#define vundefined_u8mf4() \ +__builtin_rvv_vundefined_u8mf4() +#define vundefined_u8mf8() \ +__builtin_rvv_vundefined_u8mf8() +#define vundefined_u16m1() \ +__builtin_rvv_vundefined_u16m1() +#define vundefined_u16m2() \ +__builtin_rvv_vundefined_u16m2() +#define vundefined_u16m4() \ +__builtin_rvv_vundefined_u16m4() +#define vundefined_u16m8() \ +__builtin_rvv_vundefined_u16m8() +#define vundefined_u16mf2() \ +__builtin_rvv_vundefined_u16mf2() +#define vundefined_u16mf4() \ +__builtin_rvv_vundefined_u16mf4() +#define vundefined_u32m1() \ +__builtin_rvv_vundefined_u32m1() +#define vundefined_u32m2() \ +__builtin_rvv_vundefined_u32m2() +#define vundefined_u32m4() \ +__builtin_rvv_vundefined_u32m4() +#define vundefined_u32m8() \ +__builtin_rvv_vundefined_u32m8() +#define vundefined_u32mf2() \ +__builtin_rvv_vundefined_u32mf2() +#define vundefined_u64m1() \ +__builtin_rvv_vundefined_u64m1() +#define vundefined_u64m2() \ +__builtin_rvv_vundefined_u64m2() +#define vundefined_u64m4() \ +__builtin_rvv_vundefined_u64m4() +#define vundefined_u64m8() \ +__builtin_rvv_vundefined_u64m8() +#define vzext_vf2_u16mf4(op0, op1) \ +__builtin_rvv_vzext_vf2_u16mf4((vuint8mf8_t)(op0), (size_t)(op1)) +#define vzext_vf2_u16mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vzext_vf2_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vuint8mf8_t)(op2), (size_t)(op3)) +#define vzext_vf2_u16mf2(op0, op1) \ +__builtin_rvv_vzext_vf2_u16mf2((vuint8mf4_t)(op0), (size_t)(op1)) +#define vzext_vf2_u16mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vzext_vf2_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vuint8mf4_t)(op2), (size_t)(op3)) +#define vzext_vf2_u16m1(op0, op1) \ +__builtin_rvv_vzext_vf2_u16m1((vuint8mf2_t)(op0), (size_t)(op1)) +#define vzext_vf2_u16m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vzext_vf2_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vuint8mf2_t)(op2), (size_t)(op3)) +#define vzext_vf2_u16m2(op0, op1) \ +__builtin_rvv_vzext_vf2_u16m2((vuint8m1_t)(op0), (size_t)(op1)) +#define vzext_vf2_u16m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vzext_vf2_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vzext_vf2_u16m4(op0, op1) \ +__builtin_rvv_vzext_vf2_u16m4((vuint8m2_t)(op0), (size_t)(op1)) +#define vzext_vf2_u16m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vzext_vf2_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vuint8m2_t)(op2), (size_t)(op3)) +#define vzext_vf2_u16m8(op0, op1) \ +__builtin_rvv_vzext_vf2_u16m8((vuint8m4_t)(op0), (size_t)(op1)) +#define vzext_vf2_u16m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vzext_vf2_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vuint8m4_t)(op2), (size_t)(op3)) +#define vzext_vf2_u32mf2(op0, op1) \ +__builtin_rvv_vzext_vf2_u32mf2((vuint16mf4_t)(op0), (size_t)(op1)) +#define vzext_vf2_u32mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vzext_vf2_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint16mf4_t)(op2), (size_t)(op3)) +#define vzext_vf2_u32m1(op0, op1) \ +__builtin_rvv_vzext_vf2_u32m1((vuint16mf2_t)(op0), (size_t)(op1)) +#define vzext_vf2_u32m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vzext_vf2_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint16mf2_t)(op2), (size_t)(op3)) +#define vzext_vf2_u32m2(op0, op1) \ +__builtin_rvv_vzext_vf2_u32m2((vuint16m1_t)(op0), (size_t)(op1)) +#define vzext_vf2_u32m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vzext_vf2_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vzext_vf2_u32m4(op0, op1) \ +__builtin_rvv_vzext_vf2_u32m4((vuint16m2_t)(op0), (size_t)(op1)) +#define vzext_vf2_u32m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vzext_vf2_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint16m2_t)(op2), (size_t)(op3)) +#define vzext_vf2_u32m8(op0, op1) \ +__builtin_rvv_vzext_vf2_u32m8((vuint16m4_t)(op0), (size_t)(op1)) +#define vzext_vf2_u32m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vzext_vf2_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint16m4_t)(op2), (size_t)(op3)) +#define vzext_vf2_u64m1(op0, op1) \ +__builtin_rvv_vzext_vf2_u64m1((vuint32mf2_t)(op0), (size_t)(op1)) +#define vzext_vf2_u64m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vzext_vf2_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vzext_vf2_u64m2(op0, op1) \ +__builtin_rvv_vzext_vf2_u64m2((vuint32m1_t)(op0), (size_t)(op1)) +#define vzext_vf2_u64m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vzext_vf2_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vzext_vf2_u64m4(op0, op1) \ +__builtin_rvv_vzext_vf2_u64m4((vuint32m2_t)(op0), (size_t)(op1)) +#define vzext_vf2_u64m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vzext_vf2_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vzext_vf2_u64m8(op0, op1) \ +__builtin_rvv_vzext_vf2_u64m8((vuint32m4_t)(op0), (size_t)(op1)) +#define vzext_vf2_u64m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vzext_vf2_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vzext_vf4_u32mf2(op0, op1) \ +__builtin_rvv_vzext_vf4_u32mf2((vuint8mf8_t)(op0), (size_t)(op1)) +#define vzext_vf4_u32mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vzext_vf4_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vuint8mf8_t)(op2), (size_t)(op3)) +#define vzext_vf4_u32m1(op0, op1) \ +__builtin_rvv_vzext_vf4_u32m1((vuint8mf4_t)(op0), (size_t)(op1)) +#define vzext_vf4_u32m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vzext_vf4_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vuint8mf4_t)(op2), (size_t)(op3)) +#define vzext_vf4_u32m2(op0, op1) \ +__builtin_rvv_vzext_vf4_u32m2((vuint8mf2_t)(op0), (size_t)(op1)) +#define vzext_vf4_u32m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vzext_vf4_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vuint8mf2_t)(op2), (size_t)(op3)) +#define vzext_vf4_u32m4(op0, op1) \ +__builtin_rvv_vzext_vf4_u32m4((vuint8m1_t)(op0), (size_t)(op1)) +#define vzext_vf4_u32m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vzext_vf4_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vzext_vf4_u32m8(op0, op1) \ +__builtin_rvv_vzext_vf4_u32m8((vuint8m2_t)(op0), (size_t)(op1)) +#define vzext_vf4_u32m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vzext_vf4_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vuint8m2_t)(op2), (size_t)(op3)) +#define vzext_vf4_u64m1(op0, op1) \ +__builtin_rvv_vzext_vf4_u64m1((vuint16mf4_t)(op0), (size_t)(op1)) +#define vzext_vf4_u64m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vzext_vf4_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint16mf4_t)(op2), (size_t)(op3)) +#define vzext_vf4_u64m2(op0, op1) \ +__builtin_rvv_vzext_vf4_u64m2((vuint16mf2_t)(op0), (size_t)(op1)) +#define vzext_vf4_u64m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vzext_vf4_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint16mf2_t)(op2), (size_t)(op3)) +#define vzext_vf4_u64m4(op0, op1) \ +__builtin_rvv_vzext_vf4_u64m4((vuint16m1_t)(op0), (size_t)(op1)) +#define vzext_vf4_u64m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vzext_vf4_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vzext_vf4_u64m8(op0, op1) \ +__builtin_rvv_vzext_vf4_u64m8((vuint16m2_t)(op0), (size_t)(op1)) +#define vzext_vf4_u64m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vzext_vf4_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint16m2_t)(op2), (size_t)(op3)) +#define vzext_vf8_u64m1(op0, op1) \ +__builtin_rvv_vzext_vf8_u64m1((vuint8mf8_t)(op0), (size_t)(op1)) +#define vzext_vf8_u64m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vzext_vf8_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vuint8mf8_t)(op2), (size_t)(op3)) +#define vzext_vf8_u64m2(op0, op1) \ +__builtin_rvv_vzext_vf8_u64m2((vuint8mf4_t)(op0), (size_t)(op1)) +#define vzext_vf8_u64m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vzext_vf8_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vuint8mf4_t)(op2), (size_t)(op3)) +#define vzext_vf8_u64m4(op0, op1) \ +__builtin_rvv_vzext_vf8_u64m4((vuint8mf2_t)(op0), (size_t)(op1)) +#define vzext_vf8_u64m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vzext_vf8_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vuint8mf2_t)(op2), (size_t)(op3)) +#define vzext_vf8_u64m8(op0, op1) \ +__builtin_rvv_vzext_vf8_u64m8((vuint8m1_t)(op0), (size_t)(op1)) +#define vzext_vf8_u64m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vzext_vf8_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#if defined(__riscv_f) +#define vsse32_v_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vsse32_v_f32m1((float *)(op0), (ptrdiff_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vsse32_v_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse32_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (ptrdiff_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vsse32_v_f32m2(op0, op1, op2, op3) \ +__builtin_rvv_vsse32_v_f32m2((float *)(op0), (ptrdiff_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vsse32_v_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse32_v_f32m2_m((vbool16_t)(op0), (float *)(op1), (ptrdiff_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vsse32_v_f32m4(op0, op1, op2, op3) \ +__builtin_rvv_vsse32_v_f32m4((float *)(op0), (ptrdiff_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vsse32_v_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse32_v_f32m4_m((vbool8_t)(op0), (float *)(op1), (ptrdiff_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vsse32_v_f32m8(op0, op1, op2, op3) \ +__builtin_rvv_vsse32_v_f32m8((float *)(op0), (ptrdiff_t)(op1), (vfloat32m8_t)(op2), (size_t)(op3)) +#define vsse32_v_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse32_v_f32m8_m((vbool4_t)(op0), (float *)(op1), (ptrdiff_t)(op2), (vfloat32m8_t)(op3), (size_t)(op4)) +#define vsse32_v_f32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsse32_v_f32mf2((float *)(op0), (ptrdiff_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vsse32_v_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse32_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (ptrdiff_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vluxei8_v_f32m1(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_f32m1((const float *)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vluxei8_v_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (const float *)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vluxei8_v_f32m2(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_f32m2((const float *)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vluxei8_v_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (const float *)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vluxei8_v_f32m4(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_f32m4((const float *)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vluxei8_v_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (const float *)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vluxei8_v_f32m8(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_f32m8((const float *)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vluxei8_v_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (const float *)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vluxei8_v_f32mf2(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_f32mf2((const float *)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vluxei8_v_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (const float *)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vluxei16_v_f32m1(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_f32m1((const float *)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vluxei16_v_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (const float *)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vluxei16_v_f32m2(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_f32m2((const float *)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vluxei16_v_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (const float *)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vluxei16_v_f32m4(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_f32m4((const float *)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vluxei16_v_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (const float *)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vluxei16_v_f32m8(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_f32m8((const float *)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vluxei16_v_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (const float *)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vluxei16_v_f32mf2(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_f32mf2((const float *)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vluxei16_v_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (const float *)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vluxei32_v_f32m1(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_f32m1((const float *)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vluxei32_v_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (const float *)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vluxei32_v_f32m2(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_f32m2((const float *)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vluxei32_v_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (const float *)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vluxei32_v_f32m4(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_f32m4((const float *)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vluxei32_v_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (const float *)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vluxei32_v_f32m8(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_f32m8((const float *)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vluxei32_v_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (const float *)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vluxei32_v_f32mf2(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_f32mf2((const float *)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vluxei32_v_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (const float *)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vluxei64_v_f32m1(op0, op1, op2) \ +__builtin_rvv_vluxei64_v_f32m1((const float *)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vluxei64_v_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei64_v_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (const float *)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vluxei64_v_f32m2(op0, op1, op2) \ +__builtin_rvv_vluxei64_v_f32m2((const float *)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vluxei64_v_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei64_v_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (const float *)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vluxei64_v_f32m4(op0, op1, op2) \ +__builtin_rvv_vluxei64_v_f32m4((const float *)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vluxei64_v_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei64_v_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (const float *)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vluxei64_v_f32mf2(op0, op1, op2) \ +__builtin_rvv_vluxei64_v_f32mf2((const float *)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vluxei64_v_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei64_v_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (const float *)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vloxei8_v_f32m1(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_f32m1((const float *)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vloxei8_v_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (const float *)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vloxei8_v_f32m2(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_f32m2((const float *)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vloxei8_v_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (const float *)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vloxei8_v_f32m4(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_f32m4((const float *)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vloxei8_v_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (const float *)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vloxei8_v_f32m8(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_f32m8((const float *)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vloxei8_v_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (const float *)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vloxei8_v_f32mf2(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_f32mf2((const float *)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vloxei8_v_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (const float *)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vloxei16_v_f32m1(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_f32m1((const float *)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vloxei16_v_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (const float *)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vloxei16_v_f32m2(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_f32m2((const float *)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vloxei16_v_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (const float *)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vloxei16_v_f32m4(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_f32m4((const float *)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vloxei16_v_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (const float *)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vloxei16_v_f32m8(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_f32m8((const float *)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vloxei16_v_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (const float *)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vloxei16_v_f32mf2(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_f32mf2((const float *)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vloxei16_v_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (const float *)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vloxei32_v_f32m1(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_f32m1((const float *)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vloxei32_v_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (const float *)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vloxei32_v_f32m2(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_f32m2((const float *)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vloxei32_v_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (const float *)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vloxei32_v_f32m4(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_f32m4((const float *)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vloxei32_v_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (const float *)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vloxei32_v_f32m8(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_f32m8((const float *)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vloxei32_v_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (const float *)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vloxei32_v_f32mf2(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_f32mf2((const float *)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vloxei32_v_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (const float *)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vloxei64_v_f32m1(op0, op1, op2) \ +__builtin_rvv_vloxei64_v_f32m1((const float *)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vloxei64_v_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei64_v_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (const float *)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vloxei64_v_f32m2(op0, op1, op2) \ +__builtin_rvv_vloxei64_v_f32m2((const float *)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vloxei64_v_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei64_v_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (const float *)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vloxei64_v_f32m4(op0, op1, op2) \ +__builtin_rvv_vloxei64_v_f32m4((const float *)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vloxei64_v_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei64_v_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (const float *)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vloxei64_v_f32mf2(op0, op1, op2) \ +__builtin_rvv_vloxei64_v_f32mf2((const float *)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vloxei64_v_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei64_v_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (const float *)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vmv_v_v_f32m1(op0, op1) \ +__builtin_rvv_vmv_v_v_f32m1((vfloat32m1_t)(op0), (size_t)(op1)) +#define vmv_v_v_f32m2(op0, op1) \ +__builtin_rvv_vmv_v_v_f32m2((vfloat32m2_t)(op0), (size_t)(op1)) +#define vmv_v_v_f32m4(op0, op1) \ +__builtin_rvv_vmv_v_v_f32m4((vfloat32m4_t)(op0), (size_t)(op1)) +#define vmv_v_v_f32m8(op0, op1) \ +__builtin_rvv_vmv_v_v_f32m8((vfloat32m8_t)(op0), (size_t)(op1)) +#define vmv_v_v_f32mf2(op0, op1) \ +__builtin_rvv_vmv_v_v_f32mf2((vfloat32mf2_t)(op0), (size_t)(op1)) +#define vfadd_vv_f32m1(op0, op1, op2) \ +__builtin_rvv_vfadd_vv_f32m1((vfloat32m1_t)(op0), (vfloat32m1_t)(op1), (size_t)(op2)) +#define vfadd_vv_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfadd_vv_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfadd_vv_f32m2(op0, op1, op2) \ +__builtin_rvv_vfadd_vv_f32m2((vfloat32m2_t)(op0), (vfloat32m2_t)(op1), (size_t)(op2)) +#define vfadd_vv_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfadd_vv_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vfadd_vv_f32m4(op0, op1, op2) \ +__builtin_rvv_vfadd_vv_f32m4((vfloat32m4_t)(op0), (vfloat32m4_t)(op1), (size_t)(op2)) +#define vfadd_vv_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfadd_vv_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vfadd_vv_f32m8(op0, op1, op2) \ +__builtin_rvv_vfadd_vv_f32m8((vfloat32m8_t)(op0), (vfloat32m8_t)(op1), (size_t)(op2)) +#define vfadd_vv_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfadd_vv_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (vfloat32m8_t)(op3), (size_t)(op4)) +#define vfadd_vv_f32mf2(op0, op1, op2) \ +__builtin_rvv_vfadd_vv_f32mf2((vfloat32mf2_t)(op0), (vfloat32mf2_t)(op1), (size_t)(op2)) +#define vfadd_vv_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfadd_vv_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vfadd_vf_f32m1(op0, op1, op2) \ +__builtin_rvv_vfadd_vf_f32m1((vfloat32m1_t)(op0), (float)(op1), (size_t)(op2)) +#define vfadd_vf_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfadd_vf_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (float)(op3), (size_t)(op4)) +#define vfadd_vf_f32m2(op0, op1, op2) \ +__builtin_rvv_vfadd_vf_f32m2((vfloat32m2_t)(op0), (float)(op1), (size_t)(op2)) +#define vfadd_vf_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfadd_vf_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (float)(op3), (size_t)(op4)) +#define vfadd_vf_f32m4(op0, op1, op2) \ +__builtin_rvv_vfadd_vf_f32m4((vfloat32m4_t)(op0), (float)(op1), (size_t)(op2)) +#define vfadd_vf_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfadd_vf_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (float)(op3), (size_t)(op4)) +#define vfadd_vf_f32m8(op0, op1, op2) \ +__builtin_rvv_vfadd_vf_f32m8((vfloat32m8_t)(op0), (float)(op1), (size_t)(op2)) +#define vfadd_vf_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfadd_vf_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (float)(op3), (size_t)(op4)) +#define vfadd_vf_f32mf2(op0, op1, op2) \ +__builtin_rvv_vfadd_vf_f32mf2((vfloat32mf2_t)(op0), (float)(op1), (size_t)(op2)) +#define vfadd_vf_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfadd_vf_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (float)(op3), (size_t)(op4)) +#define vfsub_vv_f32m1(op0, op1, op2) \ +__builtin_rvv_vfsub_vv_f32m1((vfloat32m1_t)(op0), (vfloat32m1_t)(op1), (size_t)(op2)) +#define vfsub_vv_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsub_vv_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfsub_vv_f32m2(op0, op1, op2) \ +__builtin_rvv_vfsub_vv_f32m2((vfloat32m2_t)(op0), (vfloat32m2_t)(op1), (size_t)(op2)) +#define vfsub_vv_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsub_vv_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vfsub_vv_f32m4(op0, op1, op2) \ +__builtin_rvv_vfsub_vv_f32m4((vfloat32m4_t)(op0), (vfloat32m4_t)(op1), (size_t)(op2)) +#define vfsub_vv_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsub_vv_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vfsub_vv_f32m8(op0, op1, op2) \ +__builtin_rvv_vfsub_vv_f32m8((vfloat32m8_t)(op0), (vfloat32m8_t)(op1), (size_t)(op2)) +#define vfsub_vv_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsub_vv_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (vfloat32m8_t)(op3), (size_t)(op4)) +#define vfsub_vv_f32mf2(op0, op1, op2) \ +__builtin_rvv_vfsub_vv_f32mf2((vfloat32mf2_t)(op0), (vfloat32mf2_t)(op1), (size_t)(op2)) +#define vfsub_vv_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsub_vv_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vfsub_vf_f32m1(op0, op1, op2) \ +__builtin_rvv_vfsub_vf_f32m1((vfloat32m1_t)(op0), (float)(op1), (size_t)(op2)) +#define vfsub_vf_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsub_vf_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (float)(op3), (size_t)(op4)) +#define vfsub_vf_f32m2(op0, op1, op2) \ +__builtin_rvv_vfsub_vf_f32m2((vfloat32m2_t)(op0), (float)(op1), (size_t)(op2)) +#define vfsub_vf_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsub_vf_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (float)(op3), (size_t)(op4)) +#define vfsub_vf_f32m4(op0, op1, op2) \ +__builtin_rvv_vfsub_vf_f32m4((vfloat32m4_t)(op0), (float)(op1), (size_t)(op2)) +#define vfsub_vf_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsub_vf_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (float)(op3), (size_t)(op4)) +#define vfsub_vf_f32m8(op0, op1, op2) \ +__builtin_rvv_vfsub_vf_f32m8((vfloat32m8_t)(op0), (float)(op1), (size_t)(op2)) +#define vfsub_vf_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsub_vf_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (float)(op3), (size_t)(op4)) +#define vfsub_vf_f32mf2(op0, op1, op2) \ +__builtin_rvv_vfsub_vf_f32mf2((vfloat32mf2_t)(op0), (float)(op1), (size_t)(op2)) +#define vfsub_vf_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsub_vf_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (float)(op3), (size_t)(op4)) +#define vfrsub_vf_f32m1(op0, op1, op2) \ +__builtin_rvv_vfrsub_vf_f32m1((vfloat32m1_t)(op0), (float)(op1), (size_t)(op2)) +#define vfrsub_vf_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfrsub_vf_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (float)(op3), (size_t)(op4)) +#define vfrsub_vf_f32m2(op0, op1, op2) \ +__builtin_rvv_vfrsub_vf_f32m2((vfloat32m2_t)(op0), (float)(op1), (size_t)(op2)) +#define vfrsub_vf_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfrsub_vf_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (float)(op3), (size_t)(op4)) +#define vfrsub_vf_f32m4(op0, op1, op2) \ +__builtin_rvv_vfrsub_vf_f32m4((vfloat32m4_t)(op0), (float)(op1), (size_t)(op2)) +#define vfrsub_vf_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfrsub_vf_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (float)(op3), (size_t)(op4)) +#define vfrsub_vf_f32m8(op0, op1, op2) \ +__builtin_rvv_vfrsub_vf_f32m8((vfloat32m8_t)(op0), (float)(op1), (size_t)(op2)) +#define vfrsub_vf_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfrsub_vf_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (float)(op3), (size_t)(op4)) +#define vfrsub_vf_f32mf2(op0, op1, op2) \ +__builtin_rvv_vfrsub_vf_f32mf2((vfloat32mf2_t)(op0), (float)(op1), (size_t)(op2)) +#define vfrsub_vf_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfrsub_vf_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (float)(op3), (size_t)(op4)) +#define vfmul_vv_f32m1(op0, op1, op2) \ +__builtin_rvv_vfmul_vv_f32m1((vfloat32m1_t)(op0), (vfloat32m1_t)(op1), (size_t)(op2)) +#define vfmul_vv_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmul_vv_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfmul_vv_f32m2(op0, op1, op2) \ +__builtin_rvv_vfmul_vv_f32m2((vfloat32m2_t)(op0), (vfloat32m2_t)(op1), (size_t)(op2)) +#define vfmul_vv_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmul_vv_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vfmul_vv_f32m4(op0, op1, op2) \ +__builtin_rvv_vfmul_vv_f32m4((vfloat32m4_t)(op0), (vfloat32m4_t)(op1), (size_t)(op2)) +#define vfmul_vv_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmul_vv_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vfmul_vv_f32m8(op0, op1, op2) \ +__builtin_rvv_vfmul_vv_f32m8((vfloat32m8_t)(op0), (vfloat32m8_t)(op1), (size_t)(op2)) +#define vfmul_vv_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmul_vv_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (vfloat32m8_t)(op3), (size_t)(op4)) +#define vfmul_vv_f32mf2(op0, op1, op2) \ +__builtin_rvv_vfmul_vv_f32mf2((vfloat32mf2_t)(op0), (vfloat32mf2_t)(op1), (size_t)(op2)) +#define vfmul_vv_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmul_vv_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vfmul_vf_f32m1(op0, op1, op2) \ +__builtin_rvv_vfmul_vf_f32m1((vfloat32m1_t)(op0), (float)(op1), (size_t)(op2)) +#define vfmul_vf_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmul_vf_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (float)(op3), (size_t)(op4)) +#define vfmul_vf_f32m2(op0, op1, op2) \ +__builtin_rvv_vfmul_vf_f32m2((vfloat32m2_t)(op0), (float)(op1), (size_t)(op2)) +#define vfmul_vf_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmul_vf_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (float)(op3), (size_t)(op4)) +#define vfmul_vf_f32m4(op0, op1, op2) \ +__builtin_rvv_vfmul_vf_f32m4((vfloat32m4_t)(op0), (float)(op1), (size_t)(op2)) +#define vfmul_vf_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmul_vf_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (float)(op3), (size_t)(op4)) +#define vfmul_vf_f32m8(op0, op1, op2) \ +__builtin_rvv_vfmul_vf_f32m8((vfloat32m8_t)(op0), (float)(op1), (size_t)(op2)) +#define vfmul_vf_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmul_vf_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (float)(op3), (size_t)(op4)) +#define vfmul_vf_f32mf2(op0, op1, op2) \ +__builtin_rvv_vfmul_vf_f32mf2((vfloat32mf2_t)(op0), (float)(op1), (size_t)(op2)) +#define vfmul_vf_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmul_vf_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (float)(op3), (size_t)(op4)) +#define vfdiv_vv_f32m1(op0, op1, op2) \ +__builtin_rvv_vfdiv_vv_f32m1((vfloat32m1_t)(op0), (vfloat32m1_t)(op1), (size_t)(op2)) +#define vfdiv_vv_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfdiv_vv_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfdiv_vv_f32m2(op0, op1, op2) \ +__builtin_rvv_vfdiv_vv_f32m2((vfloat32m2_t)(op0), (vfloat32m2_t)(op1), (size_t)(op2)) +#define vfdiv_vv_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfdiv_vv_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vfdiv_vv_f32m4(op0, op1, op2) \ +__builtin_rvv_vfdiv_vv_f32m4((vfloat32m4_t)(op0), (vfloat32m4_t)(op1), (size_t)(op2)) +#define vfdiv_vv_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfdiv_vv_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vfdiv_vv_f32m8(op0, op1, op2) \ +__builtin_rvv_vfdiv_vv_f32m8((vfloat32m8_t)(op0), (vfloat32m8_t)(op1), (size_t)(op2)) +#define vfdiv_vv_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfdiv_vv_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (vfloat32m8_t)(op3), (size_t)(op4)) +#define vfdiv_vv_f32mf2(op0, op1, op2) \ +__builtin_rvv_vfdiv_vv_f32mf2((vfloat32mf2_t)(op0), (vfloat32mf2_t)(op1), (size_t)(op2)) +#define vfdiv_vv_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfdiv_vv_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vfdiv_vf_f32m1(op0, op1, op2) \ +__builtin_rvv_vfdiv_vf_f32m1((vfloat32m1_t)(op0), (float)(op1), (size_t)(op2)) +#define vfdiv_vf_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfdiv_vf_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (float)(op3), (size_t)(op4)) +#define vfdiv_vf_f32m2(op0, op1, op2) \ +__builtin_rvv_vfdiv_vf_f32m2((vfloat32m2_t)(op0), (float)(op1), (size_t)(op2)) +#define vfdiv_vf_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfdiv_vf_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (float)(op3), (size_t)(op4)) +#define vfdiv_vf_f32m4(op0, op1, op2) \ +__builtin_rvv_vfdiv_vf_f32m4((vfloat32m4_t)(op0), (float)(op1), (size_t)(op2)) +#define vfdiv_vf_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfdiv_vf_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (float)(op3), (size_t)(op4)) +#define vfdiv_vf_f32m8(op0, op1, op2) \ +__builtin_rvv_vfdiv_vf_f32m8((vfloat32m8_t)(op0), (float)(op1), (size_t)(op2)) +#define vfdiv_vf_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfdiv_vf_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (float)(op3), (size_t)(op4)) +#define vfdiv_vf_f32mf2(op0, op1, op2) \ +__builtin_rvv_vfdiv_vf_f32mf2((vfloat32mf2_t)(op0), (float)(op1), (size_t)(op2)) +#define vfdiv_vf_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfdiv_vf_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (float)(op3), (size_t)(op4)) +#define vfrdiv_vf_f32m1(op0, op1, op2) \ +__builtin_rvv_vfrdiv_vf_f32m1((vfloat32m1_t)(op0), (float)(op1), (size_t)(op2)) +#define vfrdiv_vf_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfrdiv_vf_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (float)(op3), (size_t)(op4)) +#define vfrdiv_vf_f32m2(op0, op1, op2) \ +__builtin_rvv_vfrdiv_vf_f32m2((vfloat32m2_t)(op0), (float)(op1), (size_t)(op2)) +#define vfrdiv_vf_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfrdiv_vf_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (float)(op3), (size_t)(op4)) +#define vfrdiv_vf_f32m4(op0, op1, op2) \ +__builtin_rvv_vfrdiv_vf_f32m4((vfloat32m4_t)(op0), (float)(op1), (size_t)(op2)) +#define vfrdiv_vf_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfrdiv_vf_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (float)(op3), (size_t)(op4)) +#define vfrdiv_vf_f32m8(op0, op1, op2) \ +__builtin_rvv_vfrdiv_vf_f32m8((vfloat32m8_t)(op0), (float)(op1), (size_t)(op2)) +#define vfrdiv_vf_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfrdiv_vf_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (float)(op3), (size_t)(op4)) +#define vfrdiv_vf_f32mf2(op0, op1, op2) \ +__builtin_rvv_vfrdiv_vf_f32mf2((vfloat32mf2_t)(op0), (float)(op1), (size_t)(op2)) +#define vfrdiv_vf_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfrdiv_vf_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (float)(op3), (size_t)(op4)) +#define vfmacc_vv_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfmacc_vv_f32m1((vfloat32m1_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfmacc_vv_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmacc_vv_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfmacc_vv_f32m2(op0, op1, op2, op3) \ +__builtin_rvv_vfmacc_vv_f32m2((vfloat32m2_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vfmacc_vv_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmacc_vv_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vfmacc_vv_f32m4(op0, op1, op2, op3) \ +__builtin_rvv_vfmacc_vv_f32m4((vfloat32m4_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vfmacc_vv_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmacc_vv_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vfmacc_vv_f32m8(op0, op1, op2, op3) \ +__builtin_rvv_vfmacc_vv_f32m8((vfloat32m8_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (size_t)(op3)) +#define vfmacc_vv_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmacc_vv_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (vfloat32m8_t)(op3), (size_t)(op4)) +#define vfmacc_vv_f32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vfmacc_vv_f32mf2((vfloat32mf2_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vfmacc_vv_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmacc_vv_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vfmacc_vf_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfmacc_vf_f32m1((vfloat32m1_t)(op0), (float)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfmacc_vf_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmacc_vf_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (float)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfmacc_vf_f32m2(op0, op1, op2, op3) \ +__builtin_rvv_vfmacc_vf_f32m2((vfloat32m2_t)(op0), (float)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vfmacc_vf_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmacc_vf_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (float)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vfmacc_vf_f32m4(op0, op1, op2, op3) \ +__builtin_rvv_vfmacc_vf_f32m4((vfloat32m4_t)(op0), (float)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vfmacc_vf_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmacc_vf_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (float)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vfmacc_vf_f32m8(op0, op1, op2, op3) \ +__builtin_rvv_vfmacc_vf_f32m8((vfloat32m8_t)(op0), (float)(op1), (vfloat32m8_t)(op2), (size_t)(op3)) +#define vfmacc_vf_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmacc_vf_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (float)(op2), (vfloat32m8_t)(op3), (size_t)(op4)) +#define vfmacc_vf_f32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vfmacc_vf_f32mf2((vfloat32mf2_t)(op0), (float)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vfmacc_vf_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmacc_vf_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (float)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vfnmacc_vv_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfnmacc_vv_f32m1((vfloat32m1_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfnmacc_vv_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmacc_vv_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfnmacc_vv_f32m2(op0, op1, op2, op3) \ +__builtin_rvv_vfnmacc_vv_f32m2((vfloat32m2_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vfnmacc_vv_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmacc_vv_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vfnmacc_vv_f32m4(op0, op1, op2, op3) \ +__builtin_rvv_vfnmacc_vv_f32m4((vfloat32m4_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vfnmacc_vv_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmacc_vv_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vfnmacc_vv_f32m8(op0, op1, op2, op3) \ +__builtin_rvv_vfnmacc_vv_f32m8((vfloat32m8_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (size_t)(op3)) +#define vfnmacc_vv_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmacc_vv_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (vfloat32m8_t)(op3), (size_t)(op4)) +#define vfnmacc_vv_f32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vfnmacc_vv_f32mf2((vfloat32mf2_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vfnmacc_vv_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmacc_vv_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vfnmacc_vf_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfnmacc_vf_f32m1((vfloat32m1_t)(op0), (float)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfnmacc_vf_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmacc_vf_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (float)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfnmacc_vf_f32m2(op0, op1, op2, op3) \ +__builtin_rvv_vfnmacc_vf_f32m2((vfloat32m2_t)(op0), (float)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vfnmacc_vf_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmacc_vf_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (float)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vfnmacc_vf_f32m4(op0, op1, op2, op3) \ +__builtin_rvv_vfnmacc_vf_f32m4((vfloat32m4_t)(op0), (float)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vfnmacc_vf_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmacc_vf_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (float)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vfnmacc_vf_f32m8(op0, op1, op2, op3) \ +__builtin_rvv_vfnmacc_vf_f32m8((vfloat32m8_t)(op0), (float)(op1), (vfloat32m8_t)(op2), (size_t)(op3)) +#define vfnmacc_vf_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmacc_vf_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (float)(op2), (vfloat32m8_t)(op3), (size_t)(op4)) +#define vfnmacc_vf_f32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vfnmacc_vf_f32mf2((vfloat32mf2_t)(op0), (float)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vfnmacc_vf_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmacc_vf_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (float)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vfmsac_vv_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfmsac_vv_f32m1((vfloat32m1_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfmsac_vv_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsac_vv_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfmsac_vv_f32m2(op0, op1, op2, op3) \ +__builtin_rvv_vfmsac_vv_f32m2((vfloat32m2_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vfmsac_vv_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsac_vv_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vfmsac_vv_f32m4(op0, op1, op2, op3) \ +__builtin_rvv_vfmsac_vv_f32m4((vfloat32m4_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vfmsac_vv_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsac_vv_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vfmsac_vv_f32m8(op0, op1, op2, op3) \ +__builtin_rvv_vfmsac_vv_f32m8((vfloat32m8_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (size_t)(op3)) +#define vfmsac_vv_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsac_vv_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (vfloat32m8_t)(op3), (size_t)(op4)) +#define vfmsac_vv_f32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vfmsac_vv_f32mf2((vfloat32mf2_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vfmsac_vv_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsac_vv_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vfmsac_vf_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfmsac_vf_f32m1((vfloat32m1_t)(op0), (float)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfmsac_vf_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsac_vf_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (float)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfmsac_vf_f32m2(op0, op1, op2, op3) \ +__builtin_rvv_vfmsac_vf_f32m2((vfloat32m2_t)(op0), (float)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vfmsac_vf_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsac_vf_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (float)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vfmsac_vf_f32m4(op0, op1, op2, op3) \ +__builtin_rvv_vfmsac_vf_f32m4((vfloat32m4_t)(op0), (float)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vfmsac_vf_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsac_vf_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (float)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vfmsac_vf_f32m8(op0, op1, op2, op3) \ +__builtin_rvv_vfmsac_vf_f32m8((vfloat32m8_t)(op0), (float)(op1), (vfloat32m8_t)(op2), (size_t)(op3)) +#define vfmsac_vf_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsac_vf_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (float)(op2), (vfloat32m8_t)(op3), (size_t)(op4)) +#define vfmsac_vf_f32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vfmsac_vf_f32mf2((vfloat32mf2_t)(op0), (float)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vfmsac_vf_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsac_vf_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (float)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vfnmsac_vv_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsac_vv_f32m1((vfloat32m1_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfnmsac_vv_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsac_vv_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfnmsac_vv_f32m2(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsac_vv_f32m2((vfloat32m2_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vfnmsac_vv_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsac_vv_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vfnmsac_vv_f32m4(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsac_vv_f32m4((vfloat32m4_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vfnmsac_vv_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsac_vv_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vfnmsac_vv_f32m8(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsac_vv_f32m8((vfloat32m8_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (size_t)(op3)) +#define vfnmsac_vv_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsac_vv_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (vfloat32m8_t)(op3), (size_t)(op4)) +#define vfnmsac_vv_f32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsac_vv_f32mf2((vfloat32mf2_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vfnmsac_vv_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsac_vv_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vfnmsac_vf_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsac_vf_f32m1((vfloat32m1_t)(op0), (float)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfnmsac_vf_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsac_vf_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (float)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfnmsac_vf_f32m2(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsac_vf_f32m2((vfloat32m2_t)(op0), (float)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vfnmsac_vf_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsac_vf_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (float)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vfnmsac_vf_f32m4(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsac_vf_f32m4((vfloat32m4_t)(op0), (float)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vfnmsac_vf_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsac_vf_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (float)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vfnmsac_vf_f32m8(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsac_vf_f32m8((vfloat32m8_t)(op0), (float)(op1), (vfloat32m8_t)(op2), (size_t)(op3)) +#define vfnmsac_vf_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsac_vf_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (float)(op2), (vfloat32m8_t)(op3), (size_t)(op4)) +#define vfnmsac_vf_f32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsac_vf_f32mf2((vfloat32mf2_t)(op0), (float)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vfnmsac_vf_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsac_vf_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (float)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vfmadd_vv_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfmadd_vv_f32m1((vfloat32m1_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfmadd_vv_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmadd_vv_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfmadd_vv_f32m2(op0, op1, op2, op3) \ +__builtin_rvv_vfmadd_vv_f32m2((vfloat32m2_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vfmadd_vv_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmadd_vv_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vfmadd_vv_f32m4(op0, op1, op2, op3) \ +__builtin_rvv_vfmadd_vv_f32m4((vfloat32m4_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vfmadd_vv_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmadd_vv_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vfmadd_vv_f32m8(op0, op1, op2, op3) \ +__builtin_rvv_vfmadd_vv_f32m8((vfloat32m8_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (size_t)(op3)) +#define vfmadd_vv_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmadd_vv_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (vfloat32m8_t)(op3), (size_t)(op4)) +#define vfmadd_vv_f32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vfmadd_vv_f32mf2((vfloat32mf2_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vfmadd_vv_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmadd_vv_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vfmadd_vf_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfmadd_vf_f32m1((vfloat32m1_t)(op0), (float)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfmadd_vf_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmadd_vf_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (float)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfmadd_vf_f32m2(op0, op1, op2, op3) \ +__builtin_rvv_vfmadd_vf_f32m2((vfloat32m2_t)(op0), (float)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vfmadd_vf_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmadd_vf_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (float)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vfmadd_vf_f32m4(op0, op1, op2, op3) \ +__builtin_rvv_vfmadd_vf_f32m4((vfloat32m4_t)(op0), (float)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vfmadd_vf_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmadd_vf_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (float)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vfmadd_vf_f32m8(op0, op1, op2, op3) \ +__builtin_rvv_vfmadd_vf_f32m8((vfloat32m8_t)(op0), (float)(op1), (vfloat32m8_t)(op2), (size_t)(op3)) +#define vfmadd_vf_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmadd_vf_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (float)(op2), (vfloat32m8_t)(op3), (size_t)(op4)) +#define vfmadd_vf_f32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vfmadd_vf_f32mf2((vfloat32mf2_t)(op0), (float)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vfmadd_vf_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmadd_vf_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (float)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vfnmadd_vv_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfnmadd_vv_f32m1((vfloat32m1_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfnmadd_vv_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmadd_vv_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfnmadd_vv_f32m2(op0, op1, op2, op3) \ +__builtin_rvv_vfnmadd_vv_f32m2((vfloat32m2_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vfnmadd_vv_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmadd_vv_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vfnmadd_vv_f32m4(op0, op1, op2, op3) \ +__builtin_rvv_vfnmadd_vv_f32m4((vfloat32m4_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vfnmadd_vv_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmadd_vv_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vfnmadd_vv_f32m8(op0, op1, op2, op3) \ +__builtin_rvv_vfnmadd_vv_f32m8((vfloat32m8_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (size_t)(op3)) +#define vfnmadd_vv_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmadd_vv_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (vfloat32m8_t)(op3), (size_t)(op4)) +#define vfnmadd_vv_f32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vfnmadd_vv_f32mf2((vfloat32mf2_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vfnmadd_vv_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmadd_vv_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vfnmadd_vf_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfnmadd_vf_f32m1((vfloat32m1_t)(op0), (float)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfnmadd_vf_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmadd_vf_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (float)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfnmadd_vf_f32m2(op0, op1, op2, op3) \ +__builtin_rvv_vfnmadd_vf_f32m2((vfloat32m2_t)(op0), (float)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vfnmadd_vf_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmadd_vf_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (float)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vfnmadd_vf_f32m4(op0, op1, op2, op3) \ +__builtin_rvv_vfnmadd_vf_f32m4((vfloat32m4_t)(op0), (float)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vfnmadd_vf_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmadd_vf_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (float)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vfnmadd_vf_f32m8(op0, op1, op2, op3) \ +__builtin_rvv_vfnmadd_vf_f32m8((vfloat32m8_t)(op0), (float)(op1), (vfloat32m8_t)(op2), (size_t)(op3)) +#define vfnmadd_vf_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmadd_vf_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (float)(op2), (vfloat32m8_t)(op3), (size_t)(op4)) +#define vfnmadd_vf_f32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vfnmadd_vf_f32mf2((vfloat32mf2_t)(op0), (float)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vfnmadd_vf_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmadd_vf_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (float)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vfmsub_vv_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfmsub_vv_f32m1((vfloat32m1_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfmsub_vv_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsub_vv_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfmsub_vv_f32m2(op0, op1, op2, op3) \ +__builtin_rvv_vfmsub_vv_f32m2((vfloat32m2_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vfmsub_vv_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsub_vv_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vfmsub_vv_f32m4(op0, op1, op2, op3) \ +__builtin_rvv_vfmsub_vv_f32m4((vfloat32m4_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vfmsub_vv_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsub_vv_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vfmsub_vv_f32m8(op0, op1, op2, op3) \ +__builtin_rvv_vfmsub_vv_f32m8((vfloat32m8_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (size_t)(op3)) +#define vfmsub_vv_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsub_vv_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (vfloat32m8_t)(op3), (size_t)(op4)) +#define vfmsub_vv_f32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vfmsub_vv_f32mf2((vfloat32mf2_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vfmsub_vv_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsub_vv_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vfmsub_vf_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfmsub_vf_f32m1((vfloat32m1_t)(op0), (float)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfmsub_vf_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsub_vf_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (float)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfmsub_vf_f32m2(op0, op1, op2, op3) \ +__builtin_rvv_vfmsub_vf_f32m2((vfloat32m2_t)(op0), (float)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vfmsub_vf_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsub_vf_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (float)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vfmsub_vf_f32m4(op0, op1, op2, op3) \ +__builtin_rvv_vfmsub_vf_f32m4((vfloat32m4_t)(op0), (float)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vfmsub_vf_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsub_vf_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (float)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vfmsub_vf_f32m8(op0, op1, op2, op3) \ +__builtin_rvv_vfmsub_vf_f32m8((vfloat32m8_t)(op0), (float)(op1), (vfloat32m8_t)(op2), (size_t)(op3)) +#define vfmsub_vf_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsub_vf_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (float)(op2), (vfloat32m8_t)(op3), (size_t)(op4)) +#define vfmsub_vf_f32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vfmsub_vf_f32mf2((vfloat32mf2_t)(op0), (float)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vfmsub_vf_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsub_vf_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (float)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vfnmsub_vv_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsub_vv_f32m1((vfloat32m1_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfnmsub_vv_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsub_vv_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfnmsub_vv_f32m2(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsub_vv_f32m2((vfloat32m2_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vfnmsub_vv_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsub_vv_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vfnmsub_vv_f32m4(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsub_vv_f32m4((vfloat32m4_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vfnmsub_vv_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsub_vv_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vfnmsub_vv_f32m8(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsub_vv_f32m8((vfloat32m8_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (size_t)(op3)) +#define vfnmsub_vv_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsub_vv_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (vfloat32m8_t)(op3), (size_t)(op4)) +#define vfnmsub_vv_f32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsub_vv_f32mf2((vfloat32mf2_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vfnmsub_vv_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsub_vv_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vfnmsub_vf_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsub_vf_f32m1((vfloat32m1_t)(op0), (float)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfnmsub_vf_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsub_vf_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (float)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfnmsub_vf_f32m2(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsub_vf_f32m2((vfloat32m2_t)(op0), (float)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vfnmsub_vf_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsub_vf_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (float)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vfnmsub_vf_f32m4(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsub_vf_f32m4((vfloat32m4_t)(op0), (float)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vfnmsub_vf_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsub_vf_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (float)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vfnmsub_vf_f32m8(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsub_vf_f32m8((vfloat32m8_t)(op0), (float)(op1), (vfloat32m8_t)(op2), (size_t)(op3)) +#define vfnmsub_vf_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsub_vf_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (float)(op2), (vfloat32m8_t)(op3), (size_t)(op4)) +#define vfnmsub_vf_f32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsub_vf_f32mf2((vfloat32mf2_t)(op0), (float)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vfnmsub_vf_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsub_vf_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (float)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vfmin_vv_f32m1(op0, op1, op2) \ +__builtin_rvv_vfmin_vv_f32m1((vfloat32m1_t)(op0), (vfloat32m1_t)(op1), (size_t)(op2)) +#define vfmin_vv_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmin_vv_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfmin_vv_f32m2(op0, op1, op2) \ +__builtin_rvv_vfmin_vv_f32m2((vfloat32m2_t)(op0), (vfloat32m2_t)(op1), (size_t)(op2)) +#define vfmin_vv_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmin_vv_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vfmin_vv_f32m4(op0, op1, op2) \ +__builtin_rvv_vfmin_vv_f32m4((vfloat32m4_t)(op0), (vfloat32m4_t)(op1), (size_t)(op2)) +#define vfmin_vv_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmin_vv_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vfmin_vv_f32m8(op0, op1, op2) \ +__builtin_rvv_vfmin_vv_f32m8((vfloat32m8_t)(op0), (vfloat32m8_t)(op1), (size_t)(op2)) +#define vfmin_vv_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmin_vv_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (vfloat32m8_t)(op3), (size_t)(op4)) +#define vfmin_vv_f32mf2(op0, op1, op2) \ +__builtin_rvv_vfmin_vv_f32mf2((vfloat32mf2_t)(op0), (vfloat32mf2_t)(op1), (size_t)(op2)) +#define vfmin_vv_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmin_vv_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vfmin_vf_f32m1(op0, op1, op2) \ +__builtin_rvv_vfmin_vf_f32m1((vfloat32m1_t)(op0), (float)(op1), (size_t)(op2)) +#define vfmin_vf_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmin_vf_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (float)(op3), (size_t)(op4)) +#define vfmin_vf_f32m2(op0, op1, op2) \ +__builtin_rvv_vfmin_vf_f32m2((vfloat32m2_t)(op0), (float)(op1), (size_t)(op2)) +#define vfmin_vf_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmin_vf_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (float)(op3), (size_t)(op4)) +#define vfmin_vf_f32m4(op0, op1, op2) \ +__builtin_rvv_vfmin_vf_f32m4((vfloat32m4_t)(op0), (float)(op1), (size_t)(op2)) +#define vfmin_vf_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmin_vf_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (float)(op3), (size_t)(op4)) +#define vfmin_vf_f32m8(op0, op1, op2) \ +__builtin_rvv_vfmin_vf_f32m8((vfloat32m8_t)(op0), (float)(op1), (size_t)(op2)) +#define vfmin_vf_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmin_vf_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (float)(op3), (size_t)(op4)) +#define vfmin_vf_f32mf2(op0, op1, op2) \ +__builtin_rvv_vfmin_vf_f32mf2((vfloat32mf2_t)(op0), (float)(op1), (size_t)(op2)) +#define vfmin_vf_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmin_vf_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (float)(op3), (size_t)(op4)) +#define vfmax_vv_f32m1(op0, op1, op2) \ +__builtin_rvv_vfmax_vv_f32m1((vfloat32m1_t)(op0), (vfloat32m1_t)(op1), (size_t)(op2)) +#define vfmax_vv_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmax_vv_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfmax_vv_f32m2(op0, op1, op2) \ +__builtin_rvv_vfmax_vv_f32m2((vfloat32m2_t)(op0), (vfloat32m2_t)(op1), (size_t)(op2)) +#define vfmax_vv_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmax_vv_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vfmax_vv_f32m4(op0, op1, op2) \ +__builtin_rvv_vfmax_vv_f32m4((vfloat32m4_t)(op0), (vfloat32m4_t)(op1), (size_t)(op2)) +#define vfmax_vv_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmax_vv_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vfmax_vv_f32m8(op0, op1, op2) \ +__builtin_rvv_vfmax_vv_f32m8((vfloat32m8_t)(op0), (vfloat32m8_t)(op1), (size_t)(op2)) +#define vfmax_vv_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmax_vv_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (vfloat32m8_t)(op3), (size_t)(op4)) +#define vfmax_vv_f32mf2(op0, op1, op2) \ +__builtin_rvv_vfmax_vv_f32mf2((vfloat32mf2_t)(op0), (vfloat32mf2_t)(op1), (size_t)(op2)) +#define vfmax_vv_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmax_vv_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vfmax_vf_f32m1(op0, op1, op2) \ +__builtin_rvv_vfmax_vf_f32m1((vfloat32m1_t)(op0), (float)(op1), (size_t)(op2)) +#define vfmax_vf_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmax_vf_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (float)(op3), (size_t)(op4)) +#define vfmax_vf_f32m2(op0, op1, op2) \ +__builtin_rvv_vfmax_vf_f32m2((vfloat32m2_t)(op0), (float)(op1), (size_t)(op2)) +#define vfmax_vf_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmax_vf_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (float)(op3), (size_t)(op4)) +#define vfmax_vf_f32m4(op0, op1, op2) \ +__builtin_rvv_vfmax_vf_f32m4((vfloat32m4_t)(op0), (float)(op1), (size_t)(op2)) +#define vfmax_vf_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmax_vf_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (float)(op3), (size_t)(op4)) +#define vfmax_vf_f32m8(op0, op1, op2) \ +__builtin_rvv_vfmax_vf_f32m8((vfloat32m8_t)(op0), (float)(op1), (size_t)(op2)) +#define vfmax_vf_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmax_vf_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (float)(op3), (size_t)(op4)) +#define vfmax_vf_f32mf2(op0, op1, op2) \ +__builtin_rvv_vfmax_vf_f32mf2((vfloat32mf2_t)(op0), (float)(op1), (size_t)(op2)) +#define vfmax_vf_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmax_vf_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (float)(op3), (size_t)(op4)) +#define vfsgnj_vv_f32m1(op0, op1, op2) \ +__builtin_rvv_vfsgnj_vv_f32m1((vfloat32m1_t)(op0), (vfloat32m1_t)(op1), (size_t)(op2)) +#define vfsgnj_vv_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnj_vv_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfsgnj_vv_f32m2(op0, op1, op2) \ +__builtin_rvv_vfsgnj_vv_f32m2((vfloat32m2_t)(op0), (vfloat32m2_t)(op1), (size_t)(op2)) +#define vfsgnj_vv_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnj_vv_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vfsgnj_vv_f32m4(op0, op1, op2) \ +__builtin_rvv_vfsgnj_vv_f32m4((vfloat32m4_t)(op0), (vfloat32m4_t)(op1), (size_t)(op2)) +#define vfsgnj_vv_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnj_vv_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vfsgnj_vv_f32m8(op0, op1, op2) \ +__builtin_rvv_vfsgnj_vv_f32m8((vfloat32m8_t)(op0), (vfloat32m8_t)(op1), (size_t)(op2)) +#define vfsgnj_vv_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnj_vv_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (vfloat32m8_t)(op3), (size_t)(op4)) +#define vfsgnj_vv_f32mf2(op0, op1, op2) \ +__builtin_rvv_vfsgnj_vv_f32mf2((vfloat32mf2_t)(op0), (vfloat32mf2_t)(op1), (size_t)(op2)) +#define vfsgnj_vv_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnj_vv_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vfsgnj_vf_f32m1(op0, op1, op2) \ +__builtin_rvv_vfsgnj_vf_f32m1((vfloat32m1_t)(op0), (float)(op1), (size_t)(op2)) +#define vfsgnj_vf_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnj_vf_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (float)(op3), (size_t)(op4)) +#define vfsgnj_vf_f32m2(op0, op1, op2) \ +__builtin_rvv_vfsgnj_vf_f32m2((vfloat32m2_t)(op0), (float)(op1), (size_t)(op2)) +#define vfsgnj_vf_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnj_vf_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (float)(op3), (size_t)(op4)) +#define vfsgnj_vf_f32m4(op0, op1, op2) \ +__builtin_rvv_vfsgnj_vf_f32m4((vfloat32m4_t)(op0), (float)(op1), (size_t)(op2)) +#define vfsgnj_vf_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnj_vf_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (float)(op3), (size_t)(op4)) +#define vfsgnj_vf_f32m8(op0, op1, op2) \ +__builtin_rvv_vfsgnj_vf_f32m8((vfloat32m8_t)(op0), (float)(op1), (size_t)(op2)) +#define vfsgnj_vf_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnj_vf_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (float)(op3), (size_t)(op4)) +#define vfsgnj_vf_f32mf2(op0, op1, op2) \ +__builtin_rvv_vfsgnj_vf_f32mf2((vfloat32mf2_t)(op0), (float)(op1), (size_t)(op2)) +#define vfsgnj_vf_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnj_vf_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (float)(op3), (size_t)(op4)) +#define vfsgnjn_vv_f32m1(op0, op1, op2) \ +__builtin_rvv_vfsgnjn_vv_f32m1((vfloat32m1_t)(op0), (vfloat32m1_t)(op1), (size_t)(op2)) +#define vfsgnjn_vv_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjn_vv_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfsgnjn_vv_f32m2(op0, op1, op2) \ +__builtin_rvv_vfsgnjn_vv_f32m2((vfloat32m2_t)(op0), (vfloat32m2_t)(op1), (size_t)(op2)) +#define vfsgnjn_vv_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjn_vv_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vfsgnjn_vv_f32m4(op0, op1, op2) \ +__builtin_rvv_vfsgnjn_vv_f32m4((vfloat32m4_t)(op0), (vfloat32m4_t)(op1), (size_t)(op2)) +#define vfsgnjn_vv_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjn_vv_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vfsgnjn_vv_f32m8(op0, op1, op2) \ +__builtin_rvv_vfsgnjn_vv_f32m8((vfloat32m8_t)(op0), (vfloat32m8_t)(op1), (size_t)(op2)) +#define vfsgnjn_vv_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjn_vv_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (vfloat32m8_t)(op3), (size_t)(op4)) +#define vfsgnjn_vv_f32mf2(op0, op1, op2) \ +__builtin_rvv_vfsgnjn_vv_f32mf2((vfloat32mf2_t)(op0), (vfloat32mf2_t)(op1), (size_t)(op2)) +#define vfsgnjn_vv_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjn_vv_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vfsgnjn_vf_f32m1(op0, op1, op2) \ +__builtin_rvv_vfsgnjn_vf_f32m1((vfloat32m1_t)(op0), (float)(op1), (size_t)(op2)) +#define vfsgnjn_vf_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjn_vf_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (float)(op3), (size_t)(op4)) +#define vfsgnjn_vf_f32m2(op0, op1, op2) \ +__builtin_rvv_vfsgnjn_vf_f32m2((vfloat32m2_t)(op0), (float)(op1), (size_t)(op2)) +#define vfsgnjn_vf_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjn_vf_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (float)(op3), (size_t)(op4)) +#define vfsgnjn_vf_f32m4(op0, op1, op2) \ +__builtin_rvv_vfsgnjn_vf_f32m4((vfloat32m4_t)(op0), (float)(op1), (size_t)(op2)) +#define vfsgnjn_vf_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjn_vf_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (float)(op3), (size_t)(op4)) +#define vfsgnjn_vf_f32m8(op0, op1, op2) \ +__builtin_rvv_vfsgnjn_vf_f32m8((vfloat32m8_t)(op0), (float)(op1), (size_t)(op2)) +#define vfsgnjn_vf_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjn_vf_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (float)(op3), (size_t)(op4)) +#define vfsgnjn_vf_f32mf2(op0, op1, op2) \ +__builtin_rvv_vfsgnjn_vf_f32mf2((vfloat32mf2_t)(op0), (float)(op1), (size_t)(op2)) +#define vfsgnjn_vf_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjn_vf_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (float)(op3), (size_t)(op4)) +#define vfsgnjx_vv_f32m1(op0, op1, op2) \ +__builtin_rvv_vfsgnjx_vv_f32m1((vfloat32m1_t)(op0), (vfloat32m1_t)(op1), (size_t)(op2)) +#define vfsgnjx_vv_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjx_vv_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfsgnjx_vv_f32m2(op0, op1, op2) \ +__builtin_rvv_vfsgnjx_vv_f32m2((vfloat32m2_t)(op0), (vfloat32m2_t)(op1), (size_t)(op2)) +#define vfsgnjx_vv_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjx_vv_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vfsgnjx_vv_f32m4(op0, op1, op2) \ +__builtin_rvv_vfsgnjx_vv_f32m4((vfloat32m4_t)(op0), (vfloat32m4_t)(op1), (size_t)(op2)) +#define vfsgnjx_vv_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjx_vv_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vfsgnjx_vv_f32m8(op0, op1, op2) \ +__builtin_rvv_vfsgnjx_vv_f32m8((vfloat32m8_t)(op0), (vfloat32m8_t)(op1), (size_t)(op2)) +#define vfsgnjx_vv_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjx_vv_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (vfloat32m8_t)(op3), (size_t)(op4)) +#define vfsgnjx_vv_f32mf2(op0, op1, op2) \ +__builtin_rvv_vfsgnjx_vv_f32mf2((vfloat32mf2_t)(op0), (vfloat32mf2_t)(op1), (size_t)(op2)) +#define vfsgnjx_vv_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjx_vv_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vfsgnjx_vf_f32m1(op0, op1, op2) \ +__builtin_rvv_vfsgnjx_vf_f32m1((vfloat32m1_t)(op0), (float)(op1), (size_t)(op2)) +#define vfsgnjx_vf_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjx_vf_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (float)(op3), (size_t)(op4)) +#define vfsgnjx_vf_f32m2(op0, op1, op2) \ +__builtin_rvv_vfsgnjx_vf_f32m2((vfloat32m2_t)(op0), (float)(op1), (size_t)(op2)) +#define vfsgnjx_vf_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjx_vf_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (float)(op3), (size_t)(op4)) +#define vfsgnjx_vf_f32m4(op0, op1, op2) \ +__builtin_rvv_vfsgnjx_vf_f32m4((vfloat32m4_t)(op0), (float)(op1), (size_t)(op2)) +#define vfsgnjx_vf_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjx_vf_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (float)(op3), (size_t)(op4)) +#define vfsgnjx_vf_f32m8(op0, op1, op2) \ +__builtin_rvv_vfsgnjx_vf_f32m8((vfloat32m8_t)(op0), (float)(op1), (size_t)(op2)) +#define vfsgnjx_vf_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjx_vf_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (float)(op3), (size_t)(op4)) +#define vfsgnjx_vf_f32mf2(op0, op1, op2) \ +__builtin_rvv_vfsgnjx_vf_f32mf2((vfloat32mf2_t)(op0), (float)(op1), (size_t)(op2)) +#define vfsgnjx_vf_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjx_vf_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (float)(op3), (size_t)(op4)) +#define vfabs_v_f32m1(op0, op1) \ +__builtin_rvv_vfabs_v_f32m1((vfloat32m1_t)(op0), (size_t)(op1)) +#define vfabs_v_f32m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfabs_v_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfabs_v_f32m2(op0, op1) \ +__builtin_rvv_vfabs_v_f32m2((vfloat32m2_t)(op0), (size_t)(op1)) +#define vfabs_v_f32m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfabs_v_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vfabs_v_f32m4(op0, op1) \ +__builtin_rvv_vfabs_v_f32m4((vfloat32m4_t)(op0), (size_t)(op1)) +#define vfabs_v_f32m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfabs_v_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vfabs_v_f32m8(op0, op1) \ +__builtin_rvv_vfabs_v_f32m8((vfloat32m8_t)(op0), (size_t)(op1)) +#define vfabs_v_f32m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfabs_v_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (size_t)(op3)) +#define vfabs_v_f32mf2(op0, op1) \ +__builtin_rvv_vfabs_v_f32mf2((vfloat32mf2_t)(op0), (size_t)(op1)) +#define vfabs_v_f32mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfabs_v_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vmfeq_vv_f32m1_b32(op0, op1, op2) \ +__builtin_rvv_vmfeq_vv_f32m1_b32((vfloat32m1_t)(op0), (vfloat32m1_t)(op1), (size_t)(op2)) +#define vmfeq_vv_f32m1_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfeq_vv_f32m1_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vmfeq_vv_f32m2_b16(op0, op1, op2) \ +__builtin_rvv_vmfeq_vv_f32m2_b16((vfloat32m2_t)(op0), (vfloat32m2_t)(op1), (size_t)(op2)) +#define vmfeq_vv_f32m2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfeq_vv_f32m2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vmfeq_vv_f32m4_b8(op0, op1, op2) \ +__builtin_rvv_vmfeq_vv_f32m4_b8((vfloat32m4_t)(op0), (vfloat32m4_t)(op1), (size_t)(op2)) +#define vmfeq_vv_f32m4_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfeq_vv_f32m4_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vfloat32m4_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vmfeq_vv_f32m8_b4(op0, op1, op2) \ +__builtin_rvv_vmfeq_vv_f32m8_b4((vfloat32m8_t)(op0), (vfloat32m8_t)(op1), (size_t)(op2)) +#define vmfeq_vv_f32m8_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfeq_vv_f32m8_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vfloat32m8_t)(op2), (vfloat32m8_t)(op3), (size_t)(op4)) +#define vmfeq_vv_f32mf2_b64(op0, op1, op2) \ +__builtin_rvv_vmfeq_vv_f32mf2_b64((vfloat32mf2_t)(op0), (vfloat32mf2_t)(op1), (size_t)(op2)) +#define vmfeq_vv_f32mf2_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfeq_vv_f32mf2_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vmfeq_vf_f32m1_b32(op0, op1, op2) \ +__builtin_rvv_vmfeq_vf_f32m1_b32((vfloat32m1_t)(op0), (float)(op1), (size_t)(op2)) +#define vmfeq_vf_f32m1_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfeq_vf_f32m1_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vfloat32m1_t)(op2), (float)(op3), (size_t)(op4)) +#define vmfeq_vf_f32m2_b16(op0, op1, op2) \ +__builtin_rvv_vmfeq_vf_f32m2_b16((vfloat32m2_t)(op0), (float)(op1), (size_t)(op2)) +#define vmfeq_vf_f32m2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfeq_vf_f32m2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vfloat32m2_t)(op2), (float)(op3), (size_t)(op4)) +#define vmfeq_vf_f32m4_b8(op0, op1, op2) \ +__builtin_rvv_vmfeq_vf_f32m4_b8((vfloat32m4_t)(op0), (float)(op1), (size_t)(op2)) +#define vmfeq_vf_f32m4_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfeq_vf_f32m4_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vfloat32m4_t)(op2), (float)(op3), (size_t)(op4)) +#define vmfeq_vf_f32m8_b4(op0, op1, op2) \ +__builtin_rvv_vmfeq_vf_f32m8_b4((vfloat32m8_t)(op0), (float)(op1), (size_t)(op2)) +#define vmfeq_vf_f32m8_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfeq_vf_f32m8_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vfloat32m8_t)(op2), (float)(op3), (size_t)(op4)) +#define vmfeq_vf_f32mf2_b64(op0, op1, op2) \ +__builtin_rvv_vmfeq_vf_f32mf2_b64((vfloat32mf2_t)(op0), (float)(op1), (size_t)(op2)) +#define vmfeq_vf_f32mf2_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfeq_vf_f32mf2_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vfloat32mf2_t)(op2), (float)(op3), (size_t)(op4)) +#define vmfne_vv_f32m1_b32(op0, op1, op2) \ +__builtin_rvv_vmfne_vv_f32m1_b32((vfloat32m1_t)(op0), (vfloat32m1_t)(op1), (size_t)(op2)) +#define vmfne_vv_f32m1_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfne_vv_f32m1_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vmfne_vv_f32m2_b16(op0, op1, op2) \ +__builtin_rvv_vmfne_vv_f32m2_b16((vfloat32m2_t)(op0), (vfloat32m2_t)(op1), (size_t)(op2)) +#define vmfne_vv_f32m2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfne_vv_f32m2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vmfne_vv_f32m4_b8(op0, op1, op2) \ +__builtin_rvv_vmfne_vv_f32m4_b8((vfloat32m4_t)(op0), (vfloat32m4_t)(op1), (size_t)(op2)) +#define vmfne_vv_f32m4_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfne_vv_f32m4_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vfloat32m4_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vmfne_vv_f32m8_b4(op0, op1, op2) \ +__builtin_rvv_vmfne_vv_f32m8_b4((vfloat32m8_t)(op0), (vfloat32m8_t)(op1), (size_t)(op2)) +#define vmfne_vv_f32m8_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfne_vv_f32m8_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vfloat32m8_t)(op2), (vfloat32m8_t)(op3), (size_t)(op4)) +#define vmfne_vv_f32mf2_b64(op0, op1, op2) \ +__builtin_rvv_vmfne_vv_f32mf2_b64((vfloat32mf2_t)(op0), (vfloat32mf2_t)(op1), (size_t)(op2)) +#define vmfne_vv_f32mf2_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfne_vv_f32mf2_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vmfne_vf_f32m1_b32(op0, op1, op2) \ +__builtin_rvv_vmfne_vf_f32m1_b32((vfloat32m1_t)(op0), (float)(op1), (size_t)(op2)) +#define vmfne_vf_f32m1_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfne_vf_f32m1_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vfloat32m1_t)(op2), (float)(op3), (size_t)(op4)) +#define vmfne_vf_f32m2_b16(op0, op1, op2) \ +__builtin_rvv_vmfne_vf_f32m2_b16((vfloat32m2_t)(op0), (float)(op1), (size_t)(op2)) +#define vmfne_vf_f32m2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfne_vf_f32m2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vfloat32m2_t)(op2), (float)(op3), (size_t)(op4)) +#define vmfne_vf_f32m4_b8(op0, op1, op2) \ +__builtin_rvv_vmfne_vf_f32m4_b8((vfloat32m4_t)(op0), (float)(op1), (size_t)(op2)) +#define vmfne_vf_f32m4_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfne_vf_f32m4_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vfloat32m4_t)(op2), (float)(op3), (size_t)(op4)) +#define vmfne_vf_f32m8_b4(op0, op1, op2) \ +__builtin_rvv_vmfne_vf_f32m8_b4((vfloat32m8_t)(op0), (float)(op1), (size_t)(op2)) +#define vmfne_vf_f32m8_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfne_vf_f32m8_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vfloat32m8_t)(op2), (float)(op3), (size_t)(op4)) +#define vmfne_vf_f32mf2_b64(op0, op1, op2) \ +__builtin_rvv_vmfne_vf_f32mf2_b64((vfloat32mf2_t)(op0), (float)(op1), (size_t)(op2)) +#define vmfne_vf_f32mf2_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfne_vf_f32mf2_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vfloat32mf2_t)(op2), (float)(op3), (size_t)(op4)) +#define vmflt_vv_f32m1_b32(op0, op1, op2) \ +__builtin_rvv_vmflt_vv_f32m1_b32((vfloat32m1_t)(op0), (vfloat32m1_t)(op1), (size_t)(op2)) +#define vmflt_vv_f32m1_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmflt_vv_f32m1_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vmflt_vv_f32m2_b16(op0, op1, op2) \ +__builtin_rvv_vmflt_vv_f32m2_b16((vfloat32m2_t)(op0), (vfloat32m2_t)(op1), (size_t)(op2)) +#define vmflt_vv_f32m2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmflt_vv_f32m2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vmflt_vv_f32m4_b8(op0, op1, op2) \ +__builtin_rvv_vmflt_vv_f32m4_b8((vfloat32m4_t)(op0), (vfloat32m4_t)(op1), (size_t)(op2)) +#define vmflt_vv_f32m4_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmflt_vv_f32m4_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vfloat32m4_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vmflt_vv_f32m8_b4(op0, op1, op2) \ +__builtin_rvv_vmflt_vv_f32m8_b4((vfloat32m8_t)(op0), (vfloat32m8_t)(op1), (size_t)(op2)) +#define vmflt_vv_f32m8_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmflt_vv_f32m8_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vfloat32m8_t)(op2), (vfloat32m8_t)(op3), (size_t)(op4)) +#define vmflt_vv_f32mf2_b64(op0, op1, op2) \ +__builtin_rvv_vmflt_vv_f32mf2_b64((vfloat32mf2_t)(op0), (vfloat32mf2_t)(op1), (size_t)(op2)) +#define vmflt_vv_f32mf2_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmflt_vv_f32mf2_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vmflt_vf_f32m1_b32(op0, op1, op2) \ +__builtin_rvv_vmflt_vf_f32m1_b32((vfloat32m1_t)(op0), (float)(op1), (size_t)(op2)) +#define vmflt_vf_f32m1_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmflt_vf_f32m1_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vfloat32m1_t)(op2), (float)(op3), (size_t)(op4)) +#define vmflt_vf_f32m2_b16(op0, op1, op2) \ +__builtin_rvv_vmflt_vf_f32m2_b16((vfloat32m2_t)(op0), (float)(op1), (size_t)(op2)) +#define vmflt_vf_f32m2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmflt_vf_f32m2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vfloat32m2_t)(op2), (float)(op3), (size_t)(op4)) +#define vmflt_vf_f32m4_b8(op0, op1, op2) \ +__builtin_rvv_vmflt_vf_f32m4_b8((vfloat32m4_t)(op0), (float)(op1), (size_t)(op2)) +#define vmflt_vf_f32m4_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmflt_vf_f32m4_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vfloat32m4_t)(op2), (float)(op3), (size_t)(op4)) +#define vmflt_vf_f32m8_b4(op0, op1, op2) \ +__builtin_rvv_vmflt_vf_f32m8_b4((vfloat32m8_t)(op0), (float)(op1), (size_t)(op2)) +#define vmflt_vf_f32m8_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmflt_vf_f32m8_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vfloat32m8_t)(op2), (float)(op3), (size_t)(op4)) +#define vmflt_vf_f32mf2_b64(op0, op1, op2) \ +__builtin_rvv_vmflt_vf_f32mf2_b64((vfloat32mf2_t)(op0), (float)(op1), (size_t)(op2)) +#define vmflt_vf_f32mf2_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmflt_vf_f32mf2_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vfloat32mf2_t)(op2), (float)(op3), (size_t)(op4)) +#define vmfle_vv_f32m1_b32(op0, op1, op2) \ +__builtin_rvv_vmfle_vv_f32m1_b32((vfloat32m1_t)(op0), (vfloat32m1_t)(op1), (size_t)(op2)) +#define vmfle_vv_f32m1_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfle_vv_f32m1_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vmfle_vv_f32m2_b16(op0, op1, op2) \ +__builtin_rvv_vmfle_vv_f32m2_b16((vfloat32m2_t)(op0), (vfloat32m2_t)(op1), (size_t)(op2)) +#define vmfle_vv_f32m2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfle_vv_f32m2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vmfle_vv_f32m4_b8(op0, op1, op2) \ +__builtin_rvv_vmfle_vv_f32m4_b8((vfloat32m4_t)(op0), (vfloat32m4_t)(op1), (size_t)(op2)) +#define vmfle_vv_f32m4_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfle_vv_f32m4_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vfloat32m4_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vmfle_vv_f32m8_b4(op0, op1, op2) \ +__builtin_rvv_vmfle_vv_f32m8_b4((vfloat32m8_t)(op0), (vfloat32m8_t)(op1), (size_t)(op2)) +#define vmfle_vv_f32m8_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfle_vv_f32m8_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vfloat32m8_t)(op2), (vfloat32m8_t)(op3), (size_t)(op4)) +#define vmfle_vv_f32mf2_b64(op0, op1, op2) \ +__builtin_rvv_vmfle_vv_f32mf2_b64((vfloat32mf2_t)(op0), (vfloat32mf2_t)(op1), (size_t)(op2)) +#define vmfle_vv_f32mf2_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfle_vv_f32mf2_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vmfle_vf_f32m1_b32(op0, op1, op2) \ +__builtin_rvv_vmfle_vf_f32m1_b32((vfloat32m1_t)(op0), (float)(op1), (size_t)(op2)) +#define vmfle_vf_f32m1_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfle_vf_f32m1_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vfloat32m1_t)(op2), (float)(op3), (size_t)(op4)) +#define vmfle_vf_f32m2_b16(op0, op1, op2) \ +__builtin_rvv_vmfle_vf_f32m2_b16((vfloat32m2_t)(op0), (float)(op1), (size_t)(op2)) +#define vmfle_vf_f32m2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfle_vf_f32m2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vfloat32m2_t)(op2), (float)(op3), (size_t)(op4)) +#define vmfle_vf_f32m4_b8(op0, op1, op2) \ +__builtin_rvv_vmfle_vf_f32m4_b8((vfloat32m4_t)(op0), (float)(op1), (size_t)(op2)) +#define vmfle_vf_f32m4_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfle_vf_f32m4_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vfloat32m4_t)(op2), (float)(op3), (size_t)(op4)) +#define vmfle_vf_f32m8_b4(op0, op1, op2) \ +__builtin_rvv_vmfle_vf_f32m8_b4((vfloat32m8_t)(op0), (float)(op1), (size_t)(op2)) +#define vmfle_vf_f32m8_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfle_vf_f32m8_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vfloat32m8_t)(op2), (float)(op3), (size_t)(op4)) +#define vmfle_vf_f32mf2_b64(op0, op1, op2) \ +__builtin_rvv_vmfle_vf_f32mf2_b64((vfloat32mf2_t)(op0), (float)(op1), (size_t)(op2)) +#define vmfle_vf_f32mf2_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfle_vf_f32mf2_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vfloat32mf2_t)(op2), (float)(op3), (size_t)(op4)) +#define vmfgt_vv_f32m1_b32(op0, op1, op2) \ +__builtin_rvv_vmfgt_vv_f32m1_b32((vfloat32m1_t)(op0), (vfloat32m1_t)(op1), (size_t)(op2)) +#define vmfgt_vv_f32m1_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfgt_vv_f32m1_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vmfgt_vv_f32m2_b16(op0, op1, op2) \ +__builtin_rvv_vmfgt_vv_f32m2_b16((vfloat32m2_t)(op0), (vfloat32m2_t)(op1), (size_t)(op2)) +#define vmfgt_vv_f32m2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfgt_vv_f32m2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vmfgt_vv_f32m4_b8(op0, op1, op2) \ +__builtin_rvv_vmfgt_vv_f32m4_b8((vfloat32m4_t)(op0), (vfloat32m4_t)(op1), (size_t)(op2)) +#define vmfgt_vv_f32m4_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfgt_vv_f32m4_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vfloat32m4_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vmfgt_vv_f32m8_b4(op0, op1, op2) \ +__builtin_rvv_vmfgt_vv_f32m8_b4((vfloat32m8_t)(op0), (vfloat32m8_t)(op1), (size_t)(op2)) +#define vmfgt_vv_f32m8_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfgt_vv_f32m8_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vfloat32m8_t)(op2), (vfloat32m8_t)(op3), (size_t)(op4)) +#define vmfgt_vv_f32mf2_b64(op0, op1, op2) \ +__builtin_rvv_vmfgt_vv_f32mf2_b64((vfloat32mf2_t)(op0), (vfloat32mf2_t)(op1), (size_t)(op2)) +#define vmfgt_vv_f32mf2_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfgt_vv_f32mf2_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vmfgt_vf_f32m1_b32(op0, op1, op2) \ +__builtin_rvv_vmfgt_vf_f32m1_b32((vfloat32m1_t)(op0), (float)(op1), (size_t)(op2)) +#define vmfgt_vf_f32m1_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfgt_vf_f32m1_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vfloat32m1_t)(op2), (float)(op3), (size_t)(op4)) +#define vmfgt_vf_f32m2_b16(op0, op1, op2) \ +__builtin_rvv_vmfgt_vf_f32m2_b16((vfloat32m2_t)(op0), (float)(op1), (size_t)(op2)) +#define vmfgt_vf_f32m2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfgt_vf_f32m2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vfloat32m2_t)(op2), (float)(op3), (size_t)(op4)) +#define vmfgt_vf_f32m4_b8(op0, op1, op2) \ +__builtin_rvv_vmfgt_vf_f32m4_b8((vfloat32m4_t)(op0), (float)(op1), (size_t)(op2)) +#define vmfgt_vf_f32m4_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfgt_vf_f32m4_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vfloat32m4_t)(op2), (float)(op3), (size_t)(op4)) +#define vmfgt_vf_f32m8_b4(op0, op1, op2) \ +__builtin_rvv_vmfgt_vf_f32m8_b4((vfloat32m8_t)(op0), (float)(op1), (size_t)(op2)) +#define vmfgt_vf_f32m8_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfgt_vf_f32m8_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vfloat32m8_t)(op2), (float)(op3), (size_t)(op4)) +#define vmfgt_vf_f32mf2_b64(op0, op1, op2) \ +__builtin_rvv_vmfgt_vf_f32mf2_b64((vfloat32mf2_t)(op0), (float)(op1), (size_t)(op2)) +#define vmfgt_vf_f32mf2_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfgt_vf_f32mf2_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vfloat32mf2_t)(op2), (float)(op3), (size_t)(op4)) +#define vmfge_vv_f32m1_b32(op0, op1, op2) \ +__builtin_rvv_vmfge_vv_f32m1_b32((vfloat32m1_t)(op0), (vfloat32m1_t)(op1), (size_t)(op2)) +#define vmfge_vv_f32m1_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfge_vv_f32m1_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vmfge_vv_f32m2_b16(op0, op1, op2) \ +__builtin_rvv_vmfge_vv_f32m2_b16((vfloat32m2_t)(op0), (vfloat32m2_t)(op1), (size_t)(op2)) +#define vmfge_vv_f32m2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfge_vv_f32m2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vmfge_vv_f32m4_b8(op0, op1, op2) \ +__builtin_rvv_vmfge_vv_f32m4_b8((vfloat32m4_t)(op0), (vfloat32m4_t)(op1), (size_t)(op2)) +#define vmfge_vv_f32m4_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfge_vv_f32m4_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vfloat32m4_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vmfge_vv_f32m8_b4(op0, op1, op2) \ +__builtin_rvv_vmfge_vv_f32m8_b4((vfloat32m8_t)(op0), (vfloat32m8_t)(op1), (size_t)(op2)) +#define vmfge_vv_f32m8_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfge_vv_f32m8_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vfloat32m8_t)(op2), (vfloat32m8_t)(op3), (size_t)(op4)) +#define vmfge_vv_f32mf2_b64(op0, op1, op2) \ +__builtin_rvv_vmfge_vv_f32mf2_b64((vfloat32mf2_t)(op0), (vfloat32mf2_t)(op1), (size_t)(op2)) +#define vmfge_vv_f32mf2_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfge_vv_f32mf2_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vmfge_vf_f32m1_b32(op0, op1, op2) \ +__builtin_rvv_vmfge_vf_f32m1_b32((vfloat32m1_t)(op0), (float)(op1), (size_t)(op2)) +#define vmfge_vf_f32m1_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfge_vf_f32m1_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vfloat32m1_t)(op2), (float)(op3), (size_t)(op4)) +#define vmfge_vf_f32m2_b16(op0, op1, op2) \ +__builtin_rvv_vmfge_vf_f32m2_b16((vfloat32m2_t)(op0), (float)(op1), (size_t)(op2)) +#define vmfge_vf_f32m2_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfge_vf_f32m2_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vfloat32m2_t)(op2), (float)(op3), (size_t)(op4)) +#define vmfge_vf_f32m4_b8(op0, op1, op2) \ +__builtin_rvv_vmfge_vf_f32m4_b8((vfloat32m4_t)(op0), (float)(op1), (size_t)(op2)) +#define vmfge_vf_f32m4_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfge_vf_f32m4_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vfloat32m4_t)(op2), (float)(op3), (size_t)(op4)) +#define vmfge_vf_f32m8_b4(op0, op1, op2) \ +__builtin_rvv_vmfge_vf_f32m8_b4((vfloat32m8_t)(op0), (float)(op1), (size_t)(op2)) +#define vmfge_vf_f32m8_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfge_vf_f32m8_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vfloat32m8_t)(op2), (float)(op3), (size_t)(op4)) +#define vmfge_vf_f32mf2_b64(op0, op1, op2) \ +__builtin_rvv_vmfge_vf_f32mf2_b64((vfloat32mf2_t)(op0), (float)(op1), (size_t)(op2)) +#define vmfge_vf_f32mf2_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfge_vf_f32mf2_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vfloat32mf2_t)(op2), (float)(op3), (size_t)(op4)) +#define vmerge_vvm_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_f32m1((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vmerge_vvm_f32m2(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_f32m2((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vmerge_vvm_f32m4(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_f32m4((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vmerge_vvm_f32m8(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_f32m8((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (size_t)(op3)) +#define vmerge_vvm_f32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_f32mf2((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vfmerge_vfm_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfmerge_vfm_f32m1((vbool32_t)(op0), (vfloat32m1_t)(op1), (float)(op2), (size_t)(op3)) +#define vfmerge_vfm_f32m2(op0, op1, op2, op3) \ +__builtin_rvv_vfmerge_vfm_f32m2((vbool16_t)(op0), (vfloat32m2_t)(op1), (float)(op2), (size_t)(op3)) +#define vfmerge_vfm_f32m4(op0, op1, op2, op3) \ +__builtin_rvv_vfmerge_vfm_f32m4((vbool8_t)(op0), (vfloat32m4_t)(op1), (float)(op2), (size_t)(op3)) +#define vfmerge_vfm_f32m8(op0, op1, op2, op3) \ +__builtin_rvv_vfmerge_vfm_f32m8((vbool4_t)(op0), (vfloat32m8_t)(op1), (float)(op2), (size_t)(op3)) +#define vfmerge_vfm_f32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vfmerge_vfm_f32mf2((vbool64_t)(op0), (vfloat32mf2_t)(op1), (float)(op2), (size_t)(op3)) +#define vfmv_v_f_f32m1(op0, op1) \ +__builtin_rvv_vfmv_v_f_f32m1((float)(op0), (size_t)(op1)) +#define vfmv_v_f_f32m2(op0, op1) \ +__builtin_rvv_vfmv_v_f_f32m2((float)(op0), (size_t)(op1)) +#define vfmv_v_f_f32m4(op0, op1) \ +__builtin_rvv_vfmv_v_f_f32m4((float)(op0), (size_t)(op1)) +#define vfmv_v_f_f32m8(op0, op1) \ +__builtin_rvv_vfmv_v_f_f32m8((float)(op0), (size_t)(op1)) +#define vfmv_v_f_f32mf2(op0, op1) \ +__builtin_rvv_vfmv_v_f_f32mf2((float)(op0), (size_t)(op1)) +#define vfredmax_vs_f32m1_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredmax_vs_f32m1_f32m1((vfloat32m1_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfredmax_vs_f32m1_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredmax_vs_f32m1_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfredmax_vs_f32m2_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredmax_vs_f32m2_f32m1((vfloat32m1_t)(op0), (vfloat32m2_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfredmax_vs_f32m2_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredmax_vs_f32m2_f32m1_m((vbool16_t)(op0), (vfloat32m1_t)(op1), (vfloat32m2_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfredmax_vs_f32m4_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredmax_vs_f32m4_f32m1((vfloat32m1_t)(op0), (vfloat32m4_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfredmax_vs_f32m4_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredmax_vs_f32m4_f32m1_m((vbool8_t)(op0), (vfloat32m1_t)(op1), (vfloat32m4_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfredmax_vs_f32m8_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredmax_vs_f32m8_f32m1((vfloat32m1_t)(op0), (vfloat32m8_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfredmax_vs_f32m8_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredmax_vs_f32m8_f32m1_m((vbool4_t)(op0), (vfloat32m1_t)(op1), (vfloat32m8_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfredmax_vs_f32mf2_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredmax_vs_f32mf2_f32m1((vfloat32m1_t)(op0), (vfloat32mf2_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfredmax_vs_f32mf2_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredmax_vs_f32mf2_f32m1_m((vbool64_t)(op0), (vfloat32m1_t)(op1), (vfloat32mf2_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfredmin_vs_f32m1_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredmin_vs_f32m1_f32m1((vfloat32m1_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfredmin_vs_f32m1_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredmin_vs_f32m1_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfredmin_vs_f32m2_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredmin_vs_f32m2_f32m1((vfloat32m1_t)(op0), (vfloat32m2_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfredmin_vs_f32m2_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredmin_vs_f32m2_f32m1_m((vbool16_t)(op0), (vfloat32m1_t)(op1), (vfloat32m2_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfredmin_vs_f32m4_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredmin_vs_f32m4_f32m1((vfloat32m1_t)(op0), (vfloat32m4_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfredmin_vs_f32m4_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredmin_vs_f32m4_f32m1_m((vbool8_t)(op0), (vfloat32m1_t)(op1), (vfloat32m4_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfredmin_vs_f32m8_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredmin_vs_f32m8_f32m1((vfloat32m1_t)(op0), (vfloat32m8_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfredmin_vs_f32m8_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredmin_vs_f32m8_f32m1_m((vbool4_t)(op0), (vfloat32m1_t)(op1), (vfloat32m8_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfredmin_vs_f32mf2_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredmin_vs_f32mf2_f32m1((vfloat32m1_t)(op0), (vfloat32mf2_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfredmin_vs_f32mf2_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredmin_vs_f32mf2_f32m1_m((vbool64_t)(op0), (vfloat32m1_t)(op1), (vfloat32mf2_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfredsum_vs_f32m1_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredsum_vs_f32m1_f32m1((vfloat32m1_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfredsum_vs_f32m1_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredsum_vs_f32m1_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfredsum_vs_f32m2_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredsum_vs_f32m2_f32m1((vfloat32m1_t)(op0), (vfloat32m2_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfredsum_vs_f32m2_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredsum_vs_f32m2_f32m1_m((vbool16_t)(op0), (vfloat32m1_t)(op1), (vfloat32m2_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfredsum_vs_f32m4_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredsum_vs_f32m4_f32m1((vfloat32m1_t)(op0), (vfloat32m4_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfredsum_vs_f32m4_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredsum_vs_f32m4_f32m1_m((vbool8_t)(op0), (vfloat32m1_t)(op1), (vfloat32m4_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfredsum_vs_f32m8_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredsum_vs_f32m8_f32m1((vfloat32m1_t)(op0), (vfloat32m8_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfredsum_vs_f32m8_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredsum_vs_f32m8_f32m1_m((vbool4_t)(op0), (vfloat32m1_t)(op1), (vfloat32m8_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfredsum_vs_f32mf2_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredsum_vs_f32mf2_f32m1((vfloat32m1_t)(op0), (vfloat32mf2_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfredsum_vs_f32mf2_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredsum_vs_f32mf2_f32m1_m((vbool64_t)(op0), (vfloat32m1_t)(op1), (vfloat32mf2_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfredosum_vs_f32m1_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredosum_vs_f32m1_f32m1((vfloat32m1_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfredosum_vs_f32m1_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredosum_vs_f32m1_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfredosum_vs_f32m2_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredosum_vs_f32m2_f32m1((vfloat32m1_t)(op0), (vfloat32m2_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfredosum_vs_f32m2_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredosum_vs_f32m2_f32m1_m((vbool16_t)(op0), (vfloat32m1_t)(op1), (vfloat32m2_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfredosum_vs_f32m4_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredosum_vs_f32m4_f32m1((vfloat32m1_t)(op0), (vfloat32m4_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfredosum_vs_f32m4_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredosum_vs_f32m4_f32m1_m((vbool8_t)(op0), (vfloat32m1_t)(op1), (vfloat32m4_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfredosum_vs_f32m8_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredosum_vs_f32m8_f32m1((vfloat32m1_t)(op0), (vfloat32m8_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfredosum_vs_f32m8_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredosum_vs_f32m8_f32m1_m((vbool4_t)(op0), (vfloat32m1_t)(op1), (vfloat32m8_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfredosum_vs_f32mf2_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredosum_vs_f32mf2_f32m1((vfloat32m1_t)(op0), (vfloat32mf2_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfredosum_vs_f32mf2_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredosum_vs_f32mf2_f32m1_m((vbool64_t)(op0), (vfloat32m1_t)(op1), (vfloat32mf2_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfmv_f_s_f32m1_f32(op0) \ +__builtin_rvv_vfmv_f_s_f32m1_f32((vfloat32m1_t)(op0)) +#define vfmv_f_s_f32m2_f32(op0) \ +__builtin_rvv_vfmv_f_s_f32m2_f32((vfloat32m2_t)(op0)) +#define vfmv_f_s_f32m4_f32(op0) \ +__builtin_rvv_vfmv_f_s_f32m4_f32((vfloat32m4_t)(op0)) +#define vfmv_f_s_f32m8_f32(op0) \ +__builtin_rvv_vfmv_f_s_f32m8_f32((vfloat32m8_t)(op0)) +#define vfmv_f_s_f32mf2_f32(op0) \ +__builtin_rvv_vfmv_f_s_f32mf2_f32((vfloat32mf2_t)(op0)) +#define vfmv_s_f_f32m1(op0, op1, op2) \ +__builtin_rvv_vfmv_s_f_f32m1((vfloat32m1_t)(op0), (float)(op1), (size_t)(op2)) +#define vfmv_s_f_f32m2(op0, op1, op2) \ +__builtin_rvv_vfmv_s_f_f32m2((vfloat32m2_t)(op0), (float)(op1), (size_t)(op2)) +#define vfmv_s_f_f32m4(op0, op1, op2) \ +__builtin_rvv_vfmv_s_f_f32m4((vfloat32m4_t)(op0), (float)(op1), (size_t)(op2)) +#define vfmv_s_f_f32m8(op0, op1, op2) \ +__builtin_rvv_vfmv_s_f_f32m8((vfloat32m8_t)(op0), (float)(op1), (size_t)(op2)) +#define vfmv_s_f_f32mf2(op0, op1, op2) \ +__builtin_rvv_vfmv_s_f_f32mf2((vfloat32mf2_t)(op0), (float)(op1), (size_t)(op2)) +#define vslideup_vx_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_f32m1((vfloat32m1_t)(op0), (vfloat32m1_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_f32m2(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_f32m2((vfloat32m2_t)(op0), (vfloat32m2_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_f32m4(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_f32m4((vfloat32m4_t)(op0), (vfloat32m4_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_f32m8(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_f32m8((vfloat32m8_t)(op0), (vfloat32m8_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_f32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_f32mf2((vfloat32mf2_t)(op0), (vfloat32mf2_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_f32m1((vfloat32m1_t)(op0), (vfloat32m1_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_f32m2(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_f32m2((vfloat32m2_t)(op0), (vfloat32m2_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_f32m4(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_f32m4((vfloat32m4_t)(op0), (vfloat32m4_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_f32m8(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_f32m8((vfloat32m8_t)(op0), (vfloat32m8_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_f32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_f32mf2((vfloat32mf2_t)(op0), (vfloat32mf2_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vfslide1up_vf_f32m1(op0, op1, op2) \ +__builtin_rvv_vfslide1up_vf_f32m1((vfloat32m1_t)(op0), (float)(op1), (size_t)(op2)) +#define vfslide1up_vf_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfslide1up_vf_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (float)(op3), (size_t)(op4)) +#define vfslide1up_vf_f32m2(op0, op1, op2) \ +__builtin_rvv_vfslide1up_vf_f32m2((vfloat32m2_t)(op0), (float)(op1), (size_t)(op2)) +#define vfslide1up_vf_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfslide1up_vf_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (float)(op3), (size_t)(op4)) +#define vfslide1up_vf_f32m4(op0, op1, op2) \ +__builtin_rvv_vfslide1up_vf_f32m4((vfloat32m4_t)(op0), (float)(op1), (size_t)(op2)) +#define vfslide1up_vf_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfslide1up_vf_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (float)(op3), (size_t)(op4)) +#define vfslide1up_vf_f32m8(op0, op1, op2) \ +__builtin_rvv_vfslide1up_vf_f32m8((vfloat32m8_t)(op0), (float)(op1), (size_t)(op2)) +#define vfslide1up_vf_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfslide1up_vf_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (float)(op3), (size_t)(op4)) +#define vfslide1up_vf_f32mf2(op0, op1, op2) \ +__builtin_rvv_vfslide1up_vf_f32mf2((vfloat32mf2_t)(op0), (float)(op1), (size_t)(op2)) +#define vfslide1up_vf_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfslide1up_vf_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (float)(op3), (size_t)(op4)) +#define vfslide1down_vf_f32m1(op0, op1, op2) \ +__builtin_rvv_vfslide1down_vf_f32m1((vfloat32m1_t)(op0), (float)(op1), (size_t)(op2)) +#define vfslide1down_vf_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfslide1down_vf_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (float)(op3), (size_t)(op4)) +#define vfslide1down_vf_f32m2(op0, op1, op2) \ +__builtin_rvv_vfslide1down_vf_f32m2((vfloat32m2_t)(op0), (float)(op1), (size_t)(op2)) +#define vfslide1down_vf_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfslide1down_vf_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (float)(op3), (size_t)(op4)) +#define vfslide1down_vf_f32m4(op0, op1, op2) \ +__builtin_rvv_vfslide1down_vf_f32m4((vfloat32m4_t)(op0), (float)(op1), (size_t)(op2)) +#define vfslide1down_vf_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfslide1down_vf_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (float)(op3), (size_t)(op4)) +#define vfslide1down_vf_f32m8(op0, op1, op2) \ +__builtin_rvv_vfslide1down_vf_f32m8((vfloat32m8_t)(op0), (float)(op1), (size_t)(op2)) +#define vfslide1down_vf_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfslide1down_vf_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (float)(op3), (size_t)(op4)) +#define vfslide1down_vf_f32mf2(op0, op1, op2) \ +__builtin_rvv_vfslide1down_vf_f32mf2((vfloat32mf2_t)(op0), (float)(op1), (size_t)(op2)) +#define vfslide1down_vf_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfslide1down_vf_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (float)(op3), (size_t)(op4)) +#define vrgather_vv_f32m1(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_f32m1((vfloat32m1_t)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vrgather_vv_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vrgather_vv_f32m2(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_f32m2((vfloat32m2_t)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vrgather_vv_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vrgather_vv_f32m4(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_f32m4((vfloat32m4_t)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vrgather_vv_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vrgather_vv_f32m8(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_f32m8((vfloat32m8_t)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vrgather_vv_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vrgather_vv_f32mf2(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_f32mf2((vfloat32mf2_t)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vrgather_vv_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vrgather_vx_f32m1(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_f32m1((vfloat32m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_f32m2(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_f32m2((vfloat32m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_f32m4(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_f32m4((vfloat32m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_f32m8(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_f32m8((vfloat32m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_f32mf2(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_f32mf2((vfloat32mf2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_f32m1(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_f32m1((vfloat32m1_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_f32m2(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_f32m2((vfloat32m2_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_f32m4(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_f32m4((vfloat32m4_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_f32m8(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_f32m8((vfloat32m8_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_f32mf2(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_f32mf2((vfloat32mf2_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vcompress_vm_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_f32m1((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vcompress_vm_f32m2(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_f32m2((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vcompress_vm_f32m4(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_f32m4((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vcompress_vm_f32m8(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_f32m8((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (size_t)(op3)) +#define vcompress_vm_f32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_f32mf2((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vget_v_f32m2_f32m1(op0, op1) \ +__builtin_rvv_vget_v_f32m2_f32m1((vfloat32m2_t)(op0), (size_t)(op1)) +#define vget_v_f32m4_f32m1(op0, op1) \ +__builtin_rvv_vget_v_f32m4_f32m1((vfloat32m4_t)(op0), (size_t)(op1)) +#define vget_v_f32m8_f32m1(op0, op1) \ +__builtin_rvv_vget_v_f32m8_f32m1((vfloat32m8_t)(op0), (size_t)(op1)) +#define vget_v_f32m4_f32m2(op0, op1) \ +__builtin_rvv_vget_v_f32m4_f32m2((vfloat32m4_t)(op0), (size_t)(op1)) +#define vget_v_f32m8_f32m2(op0, op1) \ +__builtin_rvv_vget_v_f32m8_f32m2((vfloat32m8_t)(op0), (size_t)(op1)) +#define vget_v_f32m8_f32m4(op0, op1) \ +__builtin_rvv_vget_v_f32m8_f32m4((vfloat32m8_t)(op0), (size_t)(op1)) +#define vset_v_f32m1_f32m2(op0, op1, op2) \ +__builtin_rvv_vset_v_f32m1_f32m2((vfloat32m2_t)(op0), (size_t)(op1), (vfloat32m1_t)(op2)) +#define vset_v_f32m1_f32m4(op0, op1, op2) \ +__builtin_rvv_vset_v_f32m1_f32m4((vfloat32m4_t)(op0), (size_t)(op1), (vfloat32m1_t)(op2)) +#define vset_v_f32m2_f32m4(op0, op1, op2) \ +__builtin_rvv_vset_v_f32m2_f32m4((vfloat32m4_t)(op0), (size_t)(op1), (vfloat32m2_t)(op2)) +#define vsuxei8_v_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_f32m1((float *)(op0), (vuint8mf4_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vsuxei8_v_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint8mf4_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vsuxei8_v_f32m2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_f32m2((float *)(op0), (vuint8mf2_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vsuxei8_v_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_f32m2_m((vbool16_t)(op0), (float *)(op1), (vuint8mf2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vsuxei8_v_f32m4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_f32m4((float *)(op0), (vuint8m1_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vsuxei8_v_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_f32m4_m((vbool8_t)(op0), (float *)(op1), (vuint8m1_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vsuxei8_v_f32m8(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_f32m8((float *)(op0), (vuint8m2_t)(op1), (vfloat32m8_t)(op2), (size_t)(op3)) +#define vsuxei8_v_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_f32m8_m((vbool4_t)(op0), (float *)(op1), (vuint8m2_t)(op2), (vfloat32m8_t)(op3), (size_t)(op4)) +#define vsuxei8_v_f32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_f32mf2((float *)(op0), (vuint8mf8_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vsuxei8_v_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint8mf8_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vset_v_f32m1_f32m8(op0, op1, op2) \ +__builtin_rvv_vset_v_f32m1_f32m8((vfloat32m8_t)(op0), (size_t)(op1), (vfloat32m1_t)(op2)) +#define vset_v_f32m2_f32m8(op0, op1, op2) \ +__builtin_rvv_vset_v_f32m2_f32m8((vfloat32m8_t)(op0), (size_t)(op1), (vfloat32m2_t)(op2)) +#define vset_v_f32m4_f32m8(op0, op1, op2) \ +__builtin_rvv_vset_v_f32m4_f32m8((vfloat32m8_t)(op0), (size_t)(op1), (vfloat32m4_t)(op2)) +#define vsuxei16_v_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_f32m1((float *)(op0), (vuint16mf2_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vsuxei16_v_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint16mf2_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vsuxei16_v_f32m2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_f32m2((float *)(op0), (vuint16m1_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vsuxei16_v_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_f32m2_m((vbool16_t)(op0), (float *)(op1), (vuint16m1_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vsuxei16_v_f32m4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_f32m4((float *)(op0), (vuint16m2_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vsuxei16_v_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_f32m4_m((vbool8_t)(op0), (float *)(op1), (vuint16m2_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vsuxei16_v_f32m8(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_f32m8((float *)(op0), (vuint16m4_t)(op1), (vfloat32m8_t)(op2), (size_t)(op3)) +#define vsuxei16_v_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_f32m8_m((vbool4_t)(op0), (float *)(op1), (vuint16m4_t)(op2), (vfloat32m8_t)(op3), (size_t)(op4)) +#define vsuxei16_v_f32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_f32mf2((float *)(op0), (vuint16mf4_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vsuxei16_v_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint16mf4_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vsuxei32_v_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_f32m1((float *)(op0), (vuint32m1_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vsuxei32_v_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vsuxei32_v_f32m2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_f32m2((float *)(op0), (vuint32m2_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vsuxei32_v_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_f32m2_m((vbool16_t)(op0), (float *)(op1), (vuint32m2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vsuxei32_v_f32m4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_f32m4((float *)(op0), (vuint32m4_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vsuxei32_v_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_f32m4_m((vbool8_t)(op0), (float *)(op1), (vuint32m4_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vsuxei32_v_f32m8(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_f32m8((float *)(op0), (vuint32m8_t)(op1), (vfloat32m8_t)(op2), (size_t)(op3)) +#define vsuxei32_v_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_f32m8_m((vbool4_t)(op0), (float *)(op1), (vuint32m8_t)(op2), (vfloat32m8_t)(op3), (size_t)(op4)) +#define vsuxei32_v_f32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_f32mf2((float *)(op0), (vuint32mf2_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vsuxei32_v_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint32mf2_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vsuxei64_v_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei64_v_f32m1((float *)(op0), (vuint64m2_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vsuxei64_v_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei64_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint64m2_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vsuxei64_v_f32m2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei64_v_f32m2((float *)(op0), (vuint64m4_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vsuxei64_v_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei64_v_f32m2_m((vbool16_t)(op0), (float *)(op1), (vuint64m4_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vsuxei64_v_f32m4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei64_v_f32m4((float *)(op0), (vuint64m8_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vsuxei64_v_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei64_v_f32m4_m((vbool8_t)(op0), (float *)(op1), (vuint64m8_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vsuxei64_v_f32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei64_v_f32mf2((float *)(op0), (vuint64m1_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vsuxei64_v_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei64_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint64m1_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vsoxei8_v_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_f32m1((float *)(op0), (vuint8mf4_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vsoxei8_v_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint8mf4_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vsoxei8_v_f32m2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_f32m2((float *)(op0), (vuint8mf2_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vsoxei8_v_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_f32m2_m((vbool16_t)(op0), (float *)(op1), (vuint8mf2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vsoxei8_v_f32m4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_f32m4((float *)(op0), (vuint8m1_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vsoxei8_v_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_f32m4_m((vbool8_t)(op0), (float *)(op1), (vuint8m1_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vsoxei8_v_f32m8(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_f32m8((float *)(op0), (vuint8m2_t)(op1), (vfloat32m8_t)(op2), (size_t)(op3)) +#define vsoxei8_v_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_f32m8_m((vbool4_t)(op0), (float *)(op1), (vuint8m2_t)(op2), (vfloat32m8_t)(op3), (size_t)(op4)) +#define vsoxei8_v_f32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_f32mf2((float *)(op0), (vuint8mf8_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vsoxei8_v_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint8mf8_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vsoxei16_v_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_f32m1((float *)(op0), (vuint16mf2_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vsoxei16_v_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint16mf2_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vsoxei16_v_f32m2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_f32m2((float *)(op0), (vuint16m1_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vsoxei16_v_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_f32m2_m((vbool16_t)(op0), (float *)(op1), (vuint16m1_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vsoxei16_v_f32m4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_f32m4((float *)(op0), (vuint16m2_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vsoxei16_v_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_f32m4_m((vbool8_t)(op0), (float *)(op1), (vuint16m2_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vsoxei16_v_f32m8(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_f32m8((float *)(op0), (vuint16m4_t)(op1), (vfloat32m8_t)(op2), (size_t)(op3)) +#define vsoxei16_v_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_f32m8_m((vbool4_t)(op0), (float *)(op1), (vuint16m4_t)(op2), (vfloat32m8_t)(op3), (size_t)(op4)) +#define vsoxei16_v_f32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_f32mf2((float *)(op0), (vuint16mf4_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vsoxei16_v_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint16mf4_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vsoxei32_v_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_f32m1((float *)(op0), (vuint32m1_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vsoxei32_v_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vsoxei32_v_f32m2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_f32m2((float *)(op0), (vuint32m2_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vsoxei32_v_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_f32m2_m((vbool16_t)(op0), (float *)(op1), (vuint32m2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vsoxei32_v_f32m4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_f32m4((float *)(op0), (vuint32m4_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vsoxei32_v_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_f32m4_m((vbool8_t)(op0), (float *)(op1), (vuint32m4_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vsoxei32_v_f32m8(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_f32m8((float *)(op0), (vuint32m8_t)(op1), (vfloat32m8_t)(op2), (size_t)(op3)) +#define vsoxei32_v_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_f32m8_m((vbool4_t)(op0), (float *)(op1), (vuint32m8_t)(op2), (vfloat32m8_t)(op3), (size_t)(op4)) +#define vsoxei32_v_f32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_f32mf2((float *)(op0), (vuint32mf2_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vsoxei32_v_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint32mf2_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vsoxei64_v_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei64_v_f32m1((float *)(op0), (vuint64m2_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vsoxei64_v_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei64_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint64m2_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vsoxei64_v_f32m2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei64_v_f32m2((float *)(op0), (vuint64m4_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vsoxei64_v_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei64_v_f32m2_m((vbool16_t)(op0), (float *)(op1), (vuint64m4_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vsoxei64_v_f32m4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei64_v_f32m4((float *)(op0), (vuint64m8_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vsoxei64_v_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei64_v_f32m4_m((vbool8_t)(op0), (float *)(op1), (vuint64m8_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vsoxei64_v_f32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei64_v_f32mf2((float *)(op0), (vuint64m1_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vsoxei64_v_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei64_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint64m1_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vle32ff_v_f32m1(op0, op1, op2) \ +__builtin_rvv_vle32ff_v_f32m1((const float *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle32ff_v_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle32ff_v_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (const float *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle32ff_v_f32m2(op0, op1, op2) \ +__builtin_rvv_vle32ff_v_f32m2((const float *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle32ff_v_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle32ff_v_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (const float *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle32ff_v_f32m4(op0, op1, op2) \ +__builtin_rvv_vle32ff_v_f32m4((const float *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle32ff_v_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle32ff_v_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (const float *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle32ff_v_f32m8(op0, op1, op2) \ +__builtin_rvv_vle32ff_v_f32m8((const float *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle32ff_v_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle32ff_v_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (const float *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle32ff_v_f32mf2(op0, op1, op2) \ +__builtin_rvv_vle32ff_v_f32mf2((const float *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle32ff_v_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle32ff_v_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (const float *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vfneg_v_f32m1(op0, op1) \ +__builtin_rvv_vfneg_v_f32m1((vfloat32m1_t)(op0), (size_t)(op1)) +#define vfneg_v_f32m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfneg_v_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfneg_v_f32m2(op0, op1) \ +__builtin_rvv_vfneg_v_f32m2((vfloat32m2_t)(op0), (size_t)(op1)) +#define vfneg_v_f32m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfneg_v_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vfneg_v_f32m4(op0, op1) \ +__builtin_rvv_vfneg_v_f32m4((vfloat32m4_t)(op0), (size_t)(op1)) +#define vfneg_v_f32m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfneg_v_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vfneg_v_f32m8(op0, op1) \ +__builtin_rvv_vfneg_v_f32m8((vfloat32m8_t)(op0), (size_t)(op1)) +#define vfneg_v_f32m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfneg_v_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (size_t)(op3)) +#define vfneg_v_f32mf2(op0, op1) \ +__builtin_rvv_vfneg_v_f32mf2((vfloat32mf2_t)(op0), (size_t)(op1)) +#define vfneg_v_f32mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfneg_v_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vle32_v_f32m1(op0, op1) \ +__builtin_rvv_vle32_v_f32m1((const float *)(op0), (size_t)(op1)) +#define vle32_v_f32m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vle32_v_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (const float *)(op2), (size_t)(op3)) +#define vle32_v_f32m2(op0, op1) \ +__builtin_rvv_vle32_v_f32m2((const float *)(op0), (size_t)(op1)) +#define vle32_v_f32m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vle32_v_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (const float *)(op2), (size_t)(op3)) +#define vle32_v_f32m4(op0, op1) \ +__builtin_rvv_vle32_v_f32m4((const float *)(op0), (size_t)(op1)) +#define vle32_v_f32m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vle32_v_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (const float *)(op2), (size_t)(op3)) +#define vle32_v_f32m8(op0, op1) \ +__builtin_rvv_vle32_v_f32m8((const float *)(op0), (size_t)(op1)) +#define vle32_v_f32m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vle32_v_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (const float *)(op2), (size_t)(op3)) +#define vle32_v_f32mf2(op0, op1) \ +__builtin_rvv_vle32_v_f32mf2((const float *)(op0), (size_t)(op1)) +#define vle32_v_f32mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vle32_v_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (const float *)(op2), (size_t)(op3)) +#define vse32_v_f32m1(op0, op1, op2) \ +__builtin_rvv_vse32_v_f32m1((float *)(op0), (vfloat32m1_t)(op1), (size_t)(op2)) +#define vse32_v_f32m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vse32_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vse32_v_f32m2(op0, op1, op2) \ +__builtin_rvv_vse32_v_f32m2((float *)(op0), (vfloat32m2_t)(op1), (size_t)(op2)) +#define vse32_v_f32m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vse32_v_f32m2_m((vbool16_t)(op0), (float *)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vse32_v_f32m4(op0, op1, op2) \ +__builtin_rvv_vse32_v_f32m4((float *)(op0), (vfloat32m4_t)(op1), (size_t)(op2)) +#define vse32_v_f32m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vse32_v_f32m4_m((vbool8_t)(op0), (float *)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vse32_v_f32m8(op0, op1, op2) \ +__builtin_rvv_vse32_v_f32m8((float *)(op0), (vfloat32m8_t)(op1), (size_t)(op2)) +#define vse32_v_f32m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vse32_v_f32m8_m((vbool4_t)(op0), (float *)(op1), (vfloat32m8_t)(op2), (size_t)(op3)) +#define vse32_v_f32mf2(op0, op1, op2) \ +__builtin_rvv_vse32_v_f32mf2((float *)(op0), (vfloat32mf2_t)(op1), (size_t)(op2)) +#define vse32_v_f32mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vse32_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vlse32_v_f32m1(op0, op1, op2) \ +__builtin_rvv_vlse32_v_f32m1((const float *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse32_v_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse32_v_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (const float *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse32_v_f32m2(op0, op1, op2) \ +__builtin_rvv_vlse32_v_f32m2((const float *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse32_v_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse32_v_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (const float *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse32_v_f32m4(op0, op1, op2) \ +__builtin_rvv_vlse32_v_f32m4((const float *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse32_v_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse32_v_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (const float *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse32_v_f32m8(op0, op1, op2) \ +__builtin_rvv_vlse32_v_f32m8((const float *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse32_v_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse32_v_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (const float *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse32_v_f32mf2(op0, op1, op2) \ +__builtin_rvv_vlse32_v_f32mf2((const float *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse32_v_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse32_v_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (const float *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vfclass_v_u32m1(op0, op1) \ +__builtin_rvv_vfclass_v_u32m1((vfloat32m1_t)(op0), (size_t)(op1)) +#define vfclass_v_u32m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfclass_v_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfclass_v_u32m2(op0, op1) \ +__builtin_rvv_vfclass_v_u32m2((vfloat32m2_t)(op0), (size_t)(op1)) +#define vfclass_v_u32m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfclass_v_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vfclass_v_u32m4(op0, op1) \ +__builtin_rvv_vfclass_v_u32m4((vfloat32m4_t)(op0), (size_t)(op1)) +#define vfclass_v_u32m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfclass_v_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vfclass_v_u32m8(op0, op1) \ +__builtin_rvv_vfclass_v_u32m8((vfloat32m8_t)(op0), (size_t)(op1)) +#define vfclass_v_u32m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfclass_v_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vfloat32m8_t)(op2), (size_t)(op3)) +#define vfclass_v_u32mf2(op0, op1) \ +__builtin_rvv_vfclass_v_u32mf2((vfloat32mf2_t)(op0), (size_t)(op1)) +#define vfclass_v_u32mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfclass_v_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vfcvt_f_x_v_f32m1(op0, op1) \ +__builtin_rvv_vfcvt_f_x_v_f32m1((vint32m1_t)(op0), (size_t)(op1)) +#define vfcvt_f_x_v_f32m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_f_x_v_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vfcvt_f_x_v_f32m2(op0, op1) \ +__builtin_rvv_vfcvt_f_x_v_f32m2((vint32m2_t)(op0), (size_t)(op1)) +#define vfcvt_f_x_v_f32m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_f_x_v_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vfcvt_f_x_v_f32m4(op0, op1) \ +__builtin_rvv_vfcvt_f_x_v_f32m4((vint32m4_t)(op0), (size_t)(op1)) +#define vfcvt_f_x_v_f32m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_f_x_v_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vfcvt_f_x_v_f32m8(op0, op1) \ +__builtin_rvv_vfcvt_f_x_v_f32m8((vint32m8_t)(op0), (size_t)(op1)) +#define vfcvt_f_x_v_f32m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_f_x_v_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vint32m8_t)(op2), (size_t)(op3)) +#define vfcvt_f_x_v_f32mf2(op0, op1) \ +__builtin_rvv_vfcvt_f_x_v_f32mf2((vint32mf2_t)(op0), (size_t)(op1)) +#define vfcvt_f_x_v_f32mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_f_x_v_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vfcvt_f_xu_v_f32m1(op0, op1) \ +__builtin_rvv_vfcvt_f_xu_v_f32m1((vuint32m1_t)(op0), (size_t)(op1)) +#define vfcvt_f_xu_v_f32m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_f_xu_v_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vfcvt_f_xu_v_f32m2(op0, op1) \ +__builtin_rvv_vfcvt_f_xu_v_f32m2((vuint32m2_t)(op0), (size_t)(op1)) +#define vfcvt_f_xu_v_f32m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_f_xu_v_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vfcvt_f_xu_v_f32m4(op0, op1) \ +__builtin_rvv_vfcvt_f_xu_v_f32m4((vuint32m4_t)(op0), (size_t)(op1)) +#define vfcvt_f_xu_v_f32m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_f_xu_v_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vfcvt_f_xu_v_f32m8(op0, op1) \ +__builtin_rvv_vfcvt_f_xu_v_f32m8((vuint32m8_t)(op0), (size_t)(op1)) +#define vfcvt_f_xu_v_f32m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_f_xu_v_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vuint32m8_t)(op2), (size_t)(op3)) +#define vfcvt_f_xu_v_f32mf2(op0, op1) \ +__builtin_rvv_vfcvt_f_xu_v_f32mf2((vuint32mf2_t)(op0), (size_t)(op1)) +#define vfcvt_f_xu_v_f32mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_f_xu_v_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vfcvt_rtz_x_f_v_i32m1(op0, op1) \ +__builtin_rvv_vfcvt_rtz_x_f_v_i32m1((vfloat32m1_t)(op0), (size_t)(op1)) +#define vfcvt_rtz_x_f_v_i32m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_rtz_x_f_v_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfcvt_rtz_x_f_v_i32m2(op0, op1) \ +__builtin_rvv_vfcvt_rtz_x_f_v_i32m2((vfloat32m2_t)(op0), (size_t)(op1)) +#define vfcvt_rtz_x_f_v_i32m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_rtz_x_f_v_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vfcvt_rtz_x_f_v_i32m4(op0, op1) \ +__builtin_rvv_vfcvt_rtz_x_f_v_i32m4((vfloat32m4_t)(op0), (size_t)(op1)) +#define vfcvt_rtz_x_f_v_i32m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_rtz_x_f_v_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vfcvt_rtz_x_f_v_i32m8(op0, op1) \ +__builtin_rvv_vfcvt_rtz_x_f_v_i32m8((vfloat32m8_t)(op0), (size_t)(op1)) +#define vfcvt_rtz_x_f_v_i32m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_rtz_x_f_v_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vfloat32m8_t)(op2), (size_t)(op3)) +#define vfcvt_rtz_x_f_v_i32mf2(op0, op1) \ +__builtin_rvv_vfcvt_rtz_x_f_v_i32mf2((vfloat32mf2_t)(op0), (size_t)(op1)) +#define vfcvt_rtz_x_f_v_i32mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_rtz_x_f_v_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vfcvt_rtz_xu_f_v_u32m1(op0, op1) \ +__builtin_rvv_vfcvt_rtz_xu_f_v_u32m1((vfloat32m1_t)(op0), (size_t)(op1)) +#define vfcvt_rtz_xu_f_v_u32m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_rtz_xu_f_v_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfcvt_rtz_xu_f_v_u32m2(op0, op1) \ +__builtin_rvv_vfcvt_rtz_xu_f_v_u32m2((vfloat32m2_t)(op0), (size_t)(op1)) +#define vfcvt_rtz_xu_f_v_u32m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_rtz_xu_f_v_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vfcvt_rtz_xu_f_v_u32m4(op0, op1) \ +__builtin_rvv_vfcvt_rtz_xu_f_v_u32m4((vfloat32m4_t)(op0), (size_t)(op1)) +#define vfcvt_rtz_xu_f_v_u32m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_rtz_xu_f_v_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vfcvt_rtz_xu_f_v_u32m8(op0, op1) \ +__builtin_rvv_vfcvt_rtz_xu_f_v_u32m8((vfloat32m8_t)(op0), (size_t)(op1)) +#define vfcvt_rtz_xu_f_v_u32m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_rtz_xu_f_v_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vfloat32m8_t)(op2), (size_t)(op3)) +#define vfcvt_rtz_xu_f_v_u32mf2(op0, op1) \ +__builtin_rvv_vfcvt_rtz_xu_f_v_u32mf2((vfloat32mf2_t)(op0), (size_t)(op1)) +#define vfcvt_rtz_xu_f_v_u32mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_rtz_xu_f_v_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vfcvt_x_f_v_i32m1(op0, op1) \ +__builtin_rvv_vfcvt_x_f_v_i32m1((vfloat32m1_t)(op0), (size_t)(op1)) +#define vfcvt_x_f_v_i32m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_x_f_v_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfcvt_x_f_v_i32m2(op0, op1) \ +__builtin_rvv_vfcvt_x_f_v_i32m2((vfloat32m2_t)(op0), (size_t)(op1)) +#define vfcvt_x_f_v_i32m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_x_f_v_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vfcvt_x_f_v_i32m4(op0, op1) \ +__builtin_rvv_vfcvt_x_f_v_i32m4((vfloat32m4_t)(op0), (size_t)(op1)) +#define vfcvt_x_f_v_i32m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_x_f_v_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vfcvt_x_f_v_i32m8(op0, op1) \ +__builtin_rvv_vfcvt_x_f_v_i32m8((vfloat32m8_t)(op0), (size_t)(op1)) +#define vfcvt_x_f_v_i32m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_x_f_v_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vfloat32m8_t)(op2), (size_t)(op3)) +#define vfcvt_x_f_v_i32mf2(op0, op1) \ +__builtin_rvv_vfcvt_x_f_v_i32mf2((vfloat32mf2_t)(op0), (size_t)(op1)) +#define vfcvt_x_f_v_i32mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_x_f_v_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vfcvt_xu_f_v_u32m1(op0, op1) \ +__builtin_rvv_vfcvt_xu_f_v_u32m1((vfloat32m1_t)(op0), (size_t)(op1)) +#define vfcvt_xu_f_v_u32m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_xu_f_v_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfcvt_xu_f_v_u32m2(op0, op1) \ +__builtin_rvv_vfcvt_xu_f_v_u32m2((vfloat32m2_t)(op0), (size_t)(op1)) +#define vfcvt_xu_f_v_u32m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_xu_f_v_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vfcvt_xu_f_v_u32m4(op0, op1) \ +__builtin_rvv_vfcvt_xu_f_v_u32m4((vfloat32m4_t)(op0), (size_t)(op1)) +#define vfcvt_xu_f_v_u32m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_xu_f_v_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vfcvt_xu_f_v_u32m8(op0, op1) \ +__builtin_rvv_vfcvt_xu_f_v_u32m8((vfloat32m8_t)(op0), (size_t)(op1)) +#define vfcvt_xu_f_v_u32m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_xu_f_v_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vfloat32m8_t)(op2), (size_t)(op3)) +#define vfcvt_xu_f_v_u32mf2(op0, op1) \ +__builtin_rvv_vfcvt_xu_f_v_u32mf2((vfloat32mf2_t)(op0), (size_t)(op1)) +#define vfcvt_xu_f_v_u32mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_xu_f_v_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vfncvt_f_x_w_f32mf2(op0, op1) \ +__builtin_rvv_vfncvt_f_x_w_f32mf2((vint64m1_t)(op0), (size_t)(op1)) +#define vfncvt_f_x_w_f32mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_f_x_w_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vfncvt_f_x_w_f32m1(op0, op1) \ +__builtin_rvv_vfncvt_f_x_w_f32m1((vint64m2_t)(op0), (size_t)(op1)) +#define vfncvt_f_x_w_f32m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_f_x_w_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vfncvt_f_x_w_f32m2(op0, op1) \ +__builtin_rvv_vfncvt_f_x_w_f32m2((vint64m4_t)(op0), (size_t)(op1)) +#define vfncvt_f_x_w_f32m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_f_x_w_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vfncvt_f_x_w_f32m4(op0, op1) \ +__builtin_rvv_vfncvt_f_x_w_f32m4((vint64m8_t)(op0), (size_t)(op1)) +#define vfncvt_f_x_w_f32m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_f_x_w_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vfncvt_f_xu_w_f32mf2(op0, op1) \ +__builtin_rvv_vfncvt_f_xu_w_f32mf2((vuint64m1_t)(op0), (size_t)(op1)) +#define vfncvt_f_xu_w_f32mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_f_xu_w_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vfncvt_f_xu_w_f32m1(op0, op1) \ +__builtin_rvv_vfncvt_f_xu_w_f32m1((vuint64m2_t)(op0), (size_t)(op1)) +#define vfncvt_f_xu_w_f32m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_f_xu_w_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vfncvt_f_xu_w_f32m2(op0, op1) \ +__builtin_rvv_vfncvt_f_xu_w_f32m2((vuint64m4_t)(op0), (size_t)(op1)) +#define vfncvt_f_xu_w_f32m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_f_xu_w_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vfncvt_f_xu_w_f32m4(op0, op1) \ +__builtin_rvv_vfncvt_f_xu_w_f32m4((vuint64m8_t)(op0), (size_t)(op1)) +#define vfncvt_f_xu_w_f32m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_f_xu_w_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vfncvt_rtz_x_f_w_i16mf4(op0, op1) \ +__builtin_rvv_vfncvt_rtz_x_f_w_i16mf4((vfloat32mf2_t)(op0), (size_t)(op1)) +#define vfncvt_rtz_x_f_w_i16mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_rtz_x_f_w_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vfncvt_rtz_x_f_w_i16mf2(op0, op1) \ +__builtin_rvv_vfncvt_rtz_x_f_w_i16mf2((vfloat32m1_t)(op0), (size_t)(op1)) +#define vfncvt_rtz_x_f_w_i16mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_rtz_x_f_w_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfncvt_rtz_x_f_w_i16m1(op0, op1) \ +__builtin_rvv_vfncvt_rtz_x_f_w_i16m1((vfloat32m2_t)(op0), (size_t)(op1)) +#define vfncvt_rtz_x_f_w_i16m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_rtz_x_f_w_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vfncvt_rtz_x_f_w_i16m2(op0, op1) \ +__builtin_rvv_vfncvt_rtz_x_f_w_i16m2((vfloat32m4_t)(op0), (size_t)(op1)) +#define vfncvt_rtz_x_f_w_i16m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_rtz_x_f_w_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vfncvt_rtz_x_f_w_i16m4(op0, op1) \ +__builtin_rvv_vfncvt_rtz_x_f_w_i16m4((vfloat32m8_t)(op0), (size_t)(op1)) +#define vfncvt_rtz_x_f_w_i16m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_rtz_x_f_w_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vfloat32m8_t)(op2), (size_t)(op3)) +#define vfncvt_rtz_xu_f_w_u16mf4(op0, op1) \ +__builtin_rvv_vfncvt_rtz_xu_f_w_u16mf4((vfloat32mf2_t)(op0), (size_t)(op1)) +#define vfncvt_rtz_xu_f_w_u16mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_rtz_xu_f_w_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vfncvt_rtz_xu_f_w_u16mf2(op0, op1) \ +__builtin_rvv_vfncvt_rtz_xu_f_w_u16mf2((vfloat32m1_t)(op0), (size_t)(op1)) +#define vfncvt_rtz_xu_f_w_u16mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_rtz_xu_f_w_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfncvt_rtz_xu_f_w_u16m1(op0, op1) \ +__builtin_rvv_vfncvt_rtz_xu_f_w_u16m1((vfloat32m2_t)(op0), (size_t)(op1)) +#define vfncvt_rtz_xu_f_w_u16m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_rtz_xu_f_w_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vfncvt_rtz_xu_f_w_u16m2(op0, op1) \ +__builtin_rvv_vfncvt_rtz_xu_f_w_u16m2((vfloat32m4_t)(op0), (size_t)(op1)) +#define vfncvt_rtz_xu_f_w_u16m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_rtz_xu_f_w_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vfncvt_rtz_xu_f_w_u16m4(op0, op1) \ +__builtin_rvv_vfncvt_rtz_xu_f_w_u16m4((vfloat32m8_t)(op0), (size_t)(op1)) +#define vfncvt_rtz_xu_f_w_u16m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_rtz_xu_f_w_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vfloat32m8_t)(op2), (size_t)(op3)) +#define vfncvt_x_f_w_i16mf4(op0, op1) \ +__builtin_rvv_vfncvt_x_f_w_i16mf4((vfloat32mf2_t)(op0), (size_t)(op1)) +#define vfncvt_x_f_w_i16mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_x_f_w_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vfncvt_x_f_w_i16mf2(op0, op1) \ +__builtin_rvv_vfncvt_x_f_w_i16mf2((vfloat32m1_t)(op0), (size_t)(op1)) +#define vfncvt_x_f_w_i16mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_x_f_w_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfncvt_x_f_w_i16m1(op0, op1) \ +__builtin_rvv_vfncvt_x_f_w_i16m1((vfloat32m2_t)(op0), (size_t)(op1)) +#define vfncvt_x_f_w_i16m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_x_f_w_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vfncvt_x_f_w_i16m2(op0, op1) \ +__builtin_rvv_vfncvt_x_f_w_i16m2((vfloat32m4_t)(op0), (size_t)(op1)) +#define vfncvt_x_f_w_i16m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_x_f_w_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vfncvt_x_f_w_i16m4(op0, op1) \ +__builtin_rvv_vfncvt_x_f_w_i16m4((vfloat32m8_t)(op0), (size_t)(op1)) +#define vfncvt_x_f_w_i16m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_x_f_w_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vfloat32m8_t)(op2), (size_t)(op3)) +#define vfncvt_xu_f_w_u16mf4(op0, op1) \ +__builtin_rvv_vfncvt_xu_f_w_u16mf4((vfloat32mf2_t)(op0), (size_t)(op1)) +#define vfncvt_xu_f_w_u16mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_xu_f_w_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vfncvt_xu_f_w_u16mf2(op0, op1) \ +__builtin_rvv_vfncvt_xu_f_w_u16mf2((vfloat32m1_t)(op0), (size_t)(op1)) +#define vfncvt_xu_f_w_u16mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_xu_f_w_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfncvt_xu_f_w_u16m1(op0, op1) \ +__builtin_rvv_vfncvt_xu_f_w_u16m1((vfloat32m2_t)(op0), (size_t)(op1)) +#define vfncvt_xu_f_w_u16m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_xu_f_w_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vfncvt_xu_f_w_u16m2(op0, op1) \ +__builtin_rvv_vfncvt_xu_f_w_u16m2((vfloat32m4_t)(op0), (size_t)(op1)) +#define vfncvt_xu_f_w_u16m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_xu_f_w_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vfncvt_xu_f_w_u16m4(op0, op1) \ +__builtin_rvv_vfncvt_xu_f_w_u16m4((vfloat32m8_t)(op0), (size_t)(op1)) +#define vfncvt_xu_f_w_u16m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_xu_f_w_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vfloat32m8_t)(op2), (size_t)(op3)) +#define vfrec7_v_f32m1(op0, op1) \ +__builtin_rvv_vfrec7_v_f32m1((vfloat32m1_t)(op0), (size_t)(op1)) +#define vfrec7_v_f32m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfrec7_v_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfrec7_v_f32m2(op0, op1) \ +__builtin_rvv_vfrec7_v_f32m2((vfloat32m2_t)(op0), (size_t)(op1)) +#define vfrec7_v_f32m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfrec7_v_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vfrec7_v_f32m4(op0, op1) \ +__builtin_rvv_vfrec7_v_f32m4((vfloat32m4_t)(op0), (size_t)(op1)) +#define vfrec7_v_f32m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfrec7_v_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vfrec7_v_f32m8(op0, op1) \ +__builtin_rvv_vfrec7_v_f32m8((vfloat32m8_t)(op0), (size_t)(op1)) +#define vfrec7_v_f32m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfrec7_v_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (size_t)(op3)) +#define vfrec7_v_f32mf2(op0, op1) \ +__builtin_rvv_vfrec7_v_f32mf2((vfloat32mf2_t)(op0), (size_t)(op1)) +#define vfrec7_v_f32mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfrec7_v_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vfrsqrt7_v_f32m1(op0, op1) \ +__builtin_rvv_vfrsqrt7_v_f32m1((vfloat32m1_t)(op0), (size_t)(op1)) +#define vfrsqrt7_v_f32m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfrsqrt7_v_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfrsqrt7_v_f32m2(op0, op1) \ +__builtin_rvv_vfrsqrt7_v_f32m2((vfloat32m2_t)(op0), (size_t)(op1)) +#define vfrsqrt7_v_f32m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfrsqrt7_v_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vfrsqrt7_v_f32m4(op0, op1) \ +__builtin_rvv_vfrsqrt7_v_f32m4((vfloat32m4_t)(op0), (size_t)(op1)) +#define vfrsqrt7_v_f32m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfrsqrt7_v_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vfrsqrt7_v_f32m8(op0, op1) \ +__builtin_rvv_vfrsqrt7_v_f32m8((vfloat32m8_t)(op0), (size_t)(op1)) +#define vfrsqrt7_v_f32m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfrsqrt7_v_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (size_t)(op3)) +#define vfrsqrt7_v_f32mf2(op0, op1) \ +__builtin_rvv_vfrsqrt7_v_f32mf2((vfloat32mf2_t)(op0), (size_t)(op1)) +#define vfrsqrt7_v_f32mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfrsqrt7_v_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vfsqrt_v_f32m1(op0, op1) \ +__builtin_rvv_vfsqrt_v_f32m1((vfloat32m1_t)(op0), (size_t)(op1)) +#define vfsqrt_v_f32m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfsqrt_v_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfsqrt_v_f32m2(op0, op1) \ +__builtin_rvv_vfsqrt_v_f32m2((vfloat32m2_t)(op0), (size_t)(op1)) +#define vfsqrt_v_f32m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfsqrt_v_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vfsqrt_v_f32m4(op0, op1) \ +__builtin_rvv_vfsqrt_v_f32m4((vfloat32m4_t)(op0), (size_t)(op1)) +#define vfsqrt_v_f32m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfsqrt_v_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vfsqrt_v_f32m8(op0, op1) \ +__builtin_rvv_vfsqrt_v_f32m8((vfloat32m8_t)(op0), (size_t)(op1)) +#define vfsqrt_v_f32m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfsqrt_v_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (size_t)(op3)) +#define vfsqrt_v_f32mf2(op0, op1) \ +__builtin_rvv_vfsqrt_v_f32mf2((vfloat32mf2_t)(op0), (size_t)(op1)) +#define vfsqrt_v_f32mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfsqrt_v_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vfwcvt_f_x_v_f32mf2(op0, op1) \ +__builtin_rvv_vfwcvt_f_x_v_f32mf2((vint16mf4_t)(op0), (size_t)(op1)) +#define vfwcvt_f_x_v_f32mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_f_x_v_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vint16mf4_t)(op2), (size_t)(op3)) +#define vfwcvt_f_x_v_f32m1(op0, op1) \ +__builtin_rvv_vfwcvt_f_x_v_f32m1((vint16mf2_t)(op0), (size_t)(op1)) +#define vfwcvt_f_x_v_f32m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_f_x_v_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vint16mf2_t)(op2), (size_t)(op3)) +#define vfwcvt_f_x_v_f32m2(op0, op1) \ +__builtin_rvv_vfwcvt_f_x_v_f32m2((vint16m1_t)(op0), (size_t)(op1)) +#define vfwcvt_f_x_v_f32m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_f_x_v_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vfwcvt_f_x_v_f32m4(op0, op1) \ +__builtin_rvv_vfwcvt_f_x_v_f32m4((vint16m2_t)(op0), (size_t)(op1)) +#define vfwcvt_f_x_v_f32m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_f_x_v_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vint16m2_t)(op2), (size_t)(op3)) +#define vfwcvt_f_x_v_f32m8(op0, op1) \ +__builtin_rvv_vfwcvt_f_x_v_f32m8((vint16m4_t)(op0), (size_t)(op1)) +#define vfwcvt_f_x_v_f32m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_f_x_v_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vint16m4_t)(op2), (size_t)(op3)) +#define vfwcvt_f_xu_v_f32mf2(op0, op1) \ +__builtin_rvv_vfwcvt_f_xu_v_f32mf2((vuint16mf4_t)(op0), (size_t)(op1)) +#define vfwcvt_f_xu_v_f32mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_f_xu_v_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vuint16mf4_t)(op2), (size_t)(op3)) +#define vfwcvt_f_xu_v_f32m1(op0, op1) \ +__builtin_rvv_vfwcvt_f_xu_v_f32m1((vuint16mf2_t)(op0), (size_t)(op1)) +#define vfwcvt_f_xu_v_f32m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_f_xu_v_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vuint16mf2_t)(op2), (size_t)(op3)) +#define vfwcvt_f_xu_v_f32m2(op0, op1) \ +__builtin_rvv_vfwcvt_f_xu_v_f32m2((vuint16m1_t)(op0), (size_t)(op1)) +#define vfwcvt_f_xu_v_f32m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_f_xu_v_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vfwcvt_f_xu_v_f32m4(op0, op1) \ +__builtin_rvv_vfwcvt_f_xu_v_f32m4((vuint16m2_t)(op0), (size_t)(op1)) +#define vfwcvt_f_xu_v_f32m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_f_xu_v_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vuint16m2_t)(op2), (size_t)(op3)) +#define vfwcvt_f_xu_v_f32m8(op0, op1) \ +__builtin_rvv_vfwcvt_f_xu_v_f32m8((vuint16m4_t)(op0), (size_t)(op1)) +#define vfwcvt_f_xu_v_f32m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_f_xu_v_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vuint16m4_t)(op2), (size_t)(op3)) +#define vfwcvt_rtz_x_f_v_i64m1(op0, op1) \ +__builtin_rvv_vfwcvt_rtz_x_f_v_i64m1((vfloat32mf2_t)(op0), (size_t)(op1)) +#define vfwcvt_rtz_x_f_v_i64m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_rtz_x_f_v_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vfwcvt_rtz_x_f_v_i64m2(op0, op1) \ +__builtin_rvv_vfwcvt_rtz_x_f_v_i64m2((vfloat32m1_t)(op0), (size_t)(op1)) +#define vfwcvt_rtz_x_f_v_i64m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_rtz_x_f_v_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfwcvt_rtz_x_f_v_i64m4(op0, op1) \ +__builtin_rvv_vfwcvt_rtz_x_f_v_i64m4((vfloat32m2_t)(op0), (size_t)(op1)) +#define vfwcvt_rtz_x_f_v_i64m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_rtz_x_f_v_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vfwcvt_rtz_x_f_v_i64m8(op0, op1) \ +__builtin_rvv_vfwcvt_rtz_x_f_v_i64m8((vfloat32m4_t)(op0), (size_t)(op1)) +#define vfwcvt_rtz_x_f_v_i64m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_rtz_x_f_v_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vfwcvt_rtz_xu_f_v_u64m1(op0, op1) \ +__builtin_rvv_vfwcvt_rtz_xu_f_v_u64m1((vfloat32mf2_t)(op0), (size_t)(op1)) +#define vfwcvt_rtz_xu_f_v_u64m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_rtz_xu_f_v_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vfwcvt_rtz_xu_f_v_u64m2(op0, op1) \ +__builtin_rvv_vfwcvt_rtz_xu_f_v_u64m2((vfloat32m1_t)(op0), (size_t)(op1)) +#define vfwcvt_rtz_xu_f_v_u64m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_rtz_xu_f_v_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfwcvt_rtz_xu_f_v_u64m4(op0, op1) \ +__builtin_rvv_vfwcvt_rtz_xu_f_v_u64m4((vfloat32m2_t)(op0), (size_t)(op1)) +#define vfwcvt_rtz_xu_f_v_u64m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_rtz_xu_f_v_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vfwcvt_rtz_xu_f_v_u64m8(op0, op1) \ +__builtin_rvv_vfwcvt_rtz_xu_f_v_u64m8((vfloat32m4_t)(op0), (size_t)(op1)) +#define vfwcvt_rtz_xu_f_v_u64m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_rtz_xu_f_v_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vfwcvt_x_f_v_i64m1(op0, op1) \ +__builtin_rvv_vfwcvt_x_f_v_i64m1((vfloat32mf2_t)(op0), (size_t)(op1)) +#define vfwcvt_x_f_v_i64m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_x_f_v_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vfwcvt_x_f_v_i64m2(op0, op1) \ +__builtin_rvv_vfwcvt_x_f_v_i64m2((vfloat32m1_t)(op0), (size_t)(op1)) +#define vfwcvt_x_f_v_i64m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_x_f_v_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfwcvt_x_f_v_i64m4(op0, op1) \ +__builtin_rvv_vfwcvt_x_f_v_i64m4((vfloat32m2_t)(op0), (size_t)(op1)) +#define vfwcvt_x_f_v_i64m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_x_f_v_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vfwcvt_x_f_v_i64m8(op0, op1) \ +__builtin_rvv_vfwcvt_x_f_v_i64m8((vfloat32m4_t)(op0), (size_t)(op1)) +#define vfwcvt_x_f_v_i64m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_x_f_v_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vfwcvt_xu_f_v_u64m1(op0, op1) \ +__builtin_rvv_vfwcvt_xu_f_v_u64m1((vfloat32mf2_t)(op0), (size_t)(op1)) +#define vfwcvt_xu_f_v_u64m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_xu_f_v_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vfwcvt_xu_f_v_u64m2(op0, op1) \ +__builtin_rvv_vfwcvt_xu_f_v_u64m2((vfloat32m1_t)(op0), (size_t)(op1)) +#define vfwcvt_xu_f_v_u64m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_xu_f_v_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfwcvt_xu_f_v_u64m4(op0, op1) \ +__builtin_rvv_vfwcvt_xu_f_v_u64m4((vfloat32m2_t)(op0), (size_t)(op1)) +#define vfwcvt_xu_f_v_u64m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_xu_f_v_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vfwcvt_xu_f_v_u64m8(op0, op1) \ +__builtin_rvv_vfwcvt_xu_f_v_u64m8((vfloat32m4_t)(op0), (size_t)(op1)) +#define vfwcvt_xu_f_v_u64m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_xu_f_v_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vlmul_ext_v_f32mf2_f32m1(op0) \ +__builtin_rvv_vlmul_ext_v_f32mf2_f32m1((vfloat32mf2_t)(op0)) +#define vlmul_ext_v_f32m1_f32m2(op0) \ +__builtin_rvv_vlmul_ext_v_f32m1_f32m2((vfloat32m1_t)(op0)) +#define vlmul_ext_v_f32mf2_f32m2(op0) \ +__builtin_rvv_vlmul_ext_v_f32mf2_f32m2((vfloat32mf2_t)(op0)) +#define vlmul_ext_v_f32m1_f32m4(op0) \ +__builtin_rvv_vlmul_ext_v_f32m1_f32m4((vfloat32m1_t)(op0)) +#define vlmul_ext_v_f32m2_f32m4(op0) \ +__builtin_rvv_vlmul_ext_v_f32m2_f32m4((vfloat32m2_t)(op0)) +#define vlmul_ext_v_f32mf2_f32m4(op0) \ +__builtin_rvv_vlmul_ext_v_f32mf2_f32m4((vfloat32mf2_t)(op0)) +#define vlmul_ext_v_f32m1_f32m8(op0) \ +__builtin_rvv_vlmul_ext_v_f32m1_f32m8((vfloat32m1_t)(op0)) +#define vlmul_ext_v_f32m2_f32m8(op0) \ +__builtin_rvv_vlmul_ext_v_f32m2_f32m8((vfloat32m2_t)(op0)) +#define vlmul_ext_v_f32m4_f32m8(op0) \ +__builtin_rvv_vlmul_ext_v_f32m4_f32m8((vfloat32m4_t)(op0)) +#define vlmul_ext_v_f32mf2_f32m8(op0) \ +__builtin_rvv_vlmul_ext_v_f32mf2_f32m8((vfloat32mf2_t)(op0)) +#define vlmul_trunc_v_f32m1_f32mf2(op0) \ +__builtin_rvv_vlmul_trunc_v_f32m1_f32mf2((vfloat32m1_t)(op0)) +#define vlmul_trunc_v_f32m2_f32mf2(op0) \ +__builtin_rvv_vlmul_trunc_v_f32m2_f32mf2((vfloat32m2_t)(op0)) +#define vlmul_trunc_v_f32m4_f32mf2(op0) \ +__builtin_rvv_vlmul_trunc_v_f32m4_f32mf2((vfloat32m4_t)(op0)) +#define vlmul_trunc_v_f32m8_f32mf2(op0) \ +__builtin_rvv_vlmul_trunc_v_f32m8_f32mf2((vfloat32m8_t)(op0)) +#define vlmul_trunc_v_f32m2_f32m1(op0) \ +__builtin_rvv_vlmul_trunc_v_f32m2_f32m1((vfloat32m2_t)(op0)) +#define vlmul_trunc_v_f32m4_f32m1(op0) \ +__builtin_rvv_vlmul_trunc_v_f32m4_f32m1((vfloat32m4_t)(op0)) +#define vlmul_trunc_v_f32m8_f32m1(op0) \ +__builtin_rvv_vlmul_trunc_v_f32m8_f32m1((vfloat32m8_t)(op0)) +#define vlmul_trunc_v_f32m4_f32m2(op0) \ +__builtin_rvv_vlmul_trunc_v_f32m4_f32m2((vfloat32m4_t)(op0)) +#define vlmul_trunc_v_f32m8_f32m2(op0) \ +__builtin_rvv_vlmul_trunc_v_f32m8_f32m2((vfloat32m8_t)(op0)) +#define vlmul_trunc_v_f32m8_f32m4(op0) \ +__builtin_rvv_vlmul_trunc_v_f32m8_f32m4((vfloat32m8_t)(op0)) +#define vreinterpret_v_i32m1_f32m1(op0) \ +__builtin_rvv_vreinterpret_v_i32m1_f32m1((vint32m1_t)(op0)) +#define vreinterpret_v_i32m2_f32m2(op0) \ +__builtin_rvv_vreinterpret_v_i32m2_f32m2((vint32m2_t)(op0)) +#define vreinterpret_v_i32m4_f32m4(op0) \ +__builtin_rvv_vreinterpret_v_i32m4_f32m4((vint32m4_t)(op0)) +#define vreinterpret_v_i32m8_f32m8(op0) \ +__builtin_rvv_vreinterpret_v_i32m8_f32m8((vint32m8_t)(op0)) +#define vreinterpret_v_i32mf2_f32mf2(op0) \ +__builtin_rvv_vreinterpret_v_i32mf2_f32mf2((vint32mf2_t)(op0)) +#define vreinterpret_v_u32m1_f32m1(op0) \ +__builtin_rvv_vreinterpret_v_u32m1_f32m1((vuint32m1_t)(op0)) +#define vreinterpret_v_u32m2_f32m2(op0) \ +__builtin_rvv_vreinterpret_v_u32m2_f32m2((vuint32m2_t)(op0)) +#define vreinterpret_v_u32m4_f32m4(op0) \ +__builtin_rvv_vreinterpret_v_u32m4_f32m4((vuint32m4_t)(op0)) +#define vreinterpret_v_u32m8_f32m8(op0) \ +__builtin_rvv_vreinterpret_v_u32m8_f32m8((vuint32m8_t)(op0)) +#define vreinterpret_v_u32mf2_f32mf2(op0) \ +__builtin_rvv_vreinterpret_v_u32mf2_f32mf2((vuint32mf2_t)(op0)) +#define vreinterpret_v_f32m1_i32m1(op0) \ +__builtin_rvv_vreinterpret_v_f32m1_i32m1((vfloat32m1_t)(op0)) +#define vreinterpret_v_f32m2_i32m2(op0) \ +__builtin_rvv_vreinterpret_v_f32m2_i32m2((vfloat32m2_t)(op0)) +#define vreinterpret_v_f32m4_i32m4(op0) \ +__builtin_rvv_vreinterpret_v_f32m4_i32m4((vfloat32m4_t)(op0)) +#define vreinterpret_v_f32m8_i32m8(op0) \ +__builtin_rvv_vreinterpret_v_f32m8_i32m8((vfloat32m8_t)(op0)) +#define vreinterpret_v_f32mf2_i32mf2(op0) \ +__builtin_rvv_vreinterpret_v_f32mf2_i32mf2((vfloat32mf2_t)(op0)) +#define vreinterpret_v_f32m1_u32m1(op0) \ +__builtin_rvv_vreinterpret_v_f32m1_u32m1((vfloat32m1_t)(op0)) +#define vreinterpret_v_f32m2_u32m2(op0) \ +__builtin_rvv_vreinterpret_v_f32m2_u32m2((vfloat32m2_t)(op0)) +#define vreinterpret_v_f32m4_u32m4(op0) \ +__builtin_rvv_vreinterpret_v_f32m4_u32m4((vfloat32m4_t)(op0)) +#define vreinterpret_v_f32m8_u32m8(op0) \ +__builtin_rvv_vreinterpret_v_f32m8_u32m8((vfloat32m8_t)(op0)) +#define vreinterpret_v_f32mf2_u32mf2(op0) \ +__builtin_rvv_vreinterpret_v_f32mf2_u32mf2((vfloat32mf2_t)(op0)) +#define vundefined_f32m1() \ +__builtin_rvv_vundefined_f32m1() +#define vundefined_f32m2() \ +__builtin_rvv_vundefined_f32m2() +#define vundefined_f32m4() \ +__builtin_rvv_vundefined_f32m4() +#define vundefined_f32m8() \ +__builtin_rvv_vundefined_f32m8() +#define vundefined_f32mf2() \ +__builtin_rvv_vundefined_f32mf2() +#endif + +#if defined(__riscv_d) +#define vlse64_v_f64m1(op0, op1, op2) \ +__builtin_rvv_vlse64_v_f64m1((const double *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse64_v_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse64_v_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (const double *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse64_v_f64m2(op0, op1, op2) \ +__builtin_rvv_vlse64_v_f64m2((const double *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse64_v_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse64_v_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (const double *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse64_v_f64m4(op0, op1, op2) \ +__builtin_rvv_vlse64_v_f64m4((const double *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse64_v_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse64_v_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (const double *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse64_v_f64m8(op0, op1, op2) \ +__builtin_rvv_vlse64_v_f64m8((const double *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse64_v_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse64_v_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (const double *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vsse64_v_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vsse64_v_f64m1((double *)(op0), (ptrdiff_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vsse64_v_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse64_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (ptrdiff_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vsse64_v_f64m2(op0, op1, op2, op3) \ +__builtin_rvv_vsse64_v_f64m2((double *)(op0), (ptrdiff_t)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vsse64_v_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse64_v_f64m2_m((vbool32_t)(op0), (double *)(op1), (ptrdiff_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vsse64_v_f64m4(op0, op1, op2, op3) \ +__builtin_rvv_vsse64_v_f64m4((double *)(op0), (ptrdiff_t)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vsse64_v_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse64_v_f64m4_m((vbool16_t)(op0), (double *)(op1), (ptrdiff_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vsse64_v_f64m8(op0, op1, op2, op3) \ +__builtin_rvv_vsse64_v_f64m8((double *)(op0), (ptrdiff_t)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vsse64_v_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse64_v_f64m8_m((vbool8_t)(op0), (double *)(op1), (ptrdiff_t)(op2), (vfloat64m8_t)(op3), (size_t)(op4)) +#define vluxei8_v_f64m1(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_f64m1((const double *)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vluxei8_v_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (const double *)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vluxei8_v_f64m2(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_f64m2((const double *)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vluxei8_v_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (const double *)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vluxei8_v_f64m4(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_f64m4((const double *)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vluxei8_v_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (const double *)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vluxei8_v_f64m8(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_f64m8((const double *)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vluxei8_v_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (const double *)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vluxei16_v_f64m1(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_f64m1((const double *)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vluxei16_v_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (const double *)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vluxei16_v_f64m2(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_f64m2((const double *)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vluxei16_v_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (const double *)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vluxei16_v_f64m4(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_f64m4((const double *)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vluxei16_v_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (const double *)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vluxei16_v_f64m8(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_f64m8((const double *)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vluxei16_v_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (const double *)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vluxei32_v_f64m1(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_f64m1((const double *)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vluxei32_v_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (const double *)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vluxei32_v_f64m2(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_f64m2((const double *)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vluxei32_v_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (const double *)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vluxei32_v_f64m4(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_f64m4((const double *)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vluxei32_v_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (const double *)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vluxei32_v_f64m8(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_f64m8((const double *)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vluxei32_v_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (const double *)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vluxei64_v_f64m1(op0, op1, op2) \ +__builtin_rvv_vluxei64_v_f64m1((const double *)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vluxei64_v_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei64_v_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (const double *)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vluxei64_v_f64m2(op0, op1, op2) \ +__builtin_rvv_vluxei64_v_f64m2((const double *)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vluxei64_v_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei64_v_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (const double *)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vluxei64_v_f64m4(op0, op1, op2) \ +__builtin_rvv_vluxei64_v_f64m4((const double *)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vluxei64_v_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei64_v_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (const double *)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vluxei64_v_f64m8(op0, op1, op2) \ +__builtin_rvv_vluxei64_v_f64m8((const double *)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vluxei64_v_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei64_v_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (const double *)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vloxei8_v_f64m1(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_f64m1((const double *)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vloxei8_v_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (const double *)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vloxei8_v_f64m2(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_f64m2((const double *)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vloxei8_v_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (const double *)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vloxei8_v_f64m4(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_f64m4((const double *)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vloxei8_v_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (const double *)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vloxei8_v_f64m8(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_f64m8((const double *)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vloxei8_v_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (const double *)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vloxei16_v_f64m1(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_f64m1((const double *)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vloxei16_v_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (const double *)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vloxei16_v_f64m2(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_f64m2((const double *)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vloxei16_v_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (const double *)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vloxei16_v_f64m4(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_f64m4((const double *)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vloxei16_v_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (const double *)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vloxei16_v_f64m8(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_f64m8((const double *)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vloxei16_v_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (const double *)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vloxei32_v_f64m1(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_f64m1((const double *)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vloxei32_v_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (const double *)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vloxei32_v_f64m2(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_f64m2((const double *)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vloxei32_v_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (const double *)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vloxei32_v_f64m4(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_f64m4((const double *)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vloxei32_v_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (const double *)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vloxei32_v_f64m8(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_f64m8((const double *)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vloxei32_v_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (const double *)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vloxei64_v_f64m1(op0, op1, op2) \ +__builtin_rvv_vloxei64_v_f64m1((const double *)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vloxei64_v_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei64_v_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (const double *)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vloxei64_v_f64m2(op0, op1, op2) \ +__builtin_rvv_vloxei64_v_f64m2((const double *)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vloxei64_v_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei64_v_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (const double *)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vloxei64_v_f64m4(op0, op1, op2) \ +__builtin_rvv_vloxei64_v_f64m4((const double *)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vloxei64_v_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei64_v_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (const double *)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vloxei64_v_f64m8(op0, op1, op2) \ +__builtin_rvv_vloxei64_v_f64m8((const double *)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vloxei64_v_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei64_v_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (const double *)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vmv_v_v_f64m1(op0, op1) \ +__builtin_rvv_vmv_v_v_f64m1((vfloat64m1_t)(op0), (size_t)(op1)) +#define vmv_v_v_f64m2(op0, op1) \ +__builtin_rvv_vmv_v_v_f64m2((vfloat64m2_t)(op0), (size_t)(op1)) +#define vmv_v_v_f64m4(op0, op1) \ +__builtin_rvv_vmv_v_v_f64m4((vfloat64m4_t)(op0), (size_t)(op1)) +#define vmv_v_v_f64m8(op0, op1) \ +__builtin_rvv_vmv_v_v_f64m8((vfloat64m8_t)(op0), (size_t)(op1)) +#define vfadd_vv_f64m1(op0, op1, op2) \ +__builtin_rvv_vfadd_vv_f64m1((vfloat64m1_t)(op0), (vfloat64m1_t)(op1), (size_t)(op2)) +#define vfadd_vv_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfadd_vv_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfadd_vv_f64m2(op0, op1, op2) \ +__builtin_rvv_vfadd_vv_f64m2((vfloat64m2_t)(op0), (vfloat64m2_t)(op1), (size_t)(op2)) +#define vfadd_vv_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfadd_vv_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vfadd_vv_f64m4(op0, op1, op2) \ +__builtin_rvv_vfadd_vv_f64m4((vfloat64m4_t)(op0), (vfloat64m4_t)(op1), (size_t)(op2)) +#define vfadd_vv_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfadd_vv_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vfadd_vv_f64m8(op0, op1, op2) \ +__builtin_rvv_vfadd_vv_f64m8((vfloat64m8_t)(op0), (vfloat64m8_t)(op1), (size_t)(op2)) +#define vfadd_vv_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfadd_vv_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (vfloat64m8_t)(op3), (size_t)(op4)) +#define vfadd_vf_f64m1(op0, op1, op2) \ +__builtin_rvv_vfadd_vf_f64m1((vfloat64m1_t)(op0), (double)(op1), (size_t)(op2)) +#define vfadd_vf_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfadd_vf_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (double)(op3), (size_t)(op4)) +#define vfadd_vf_f64m2(op0, op1, op2) \ +__builtin_rvv_vfadd_vf_f64m2((vfloat64m2_t)(op0), (double)(op1), (size_t)(op2)) +#define vfadd_vf_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfadd_vf_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (double)(op3), (size_t)(op4)) +#define vfadd_vf_f64m4(op0, op1, op2) \ +__builtin_rvv_vfadd_vf_f64m4((vfloat64m4_t)(op0), (double)(op1), (size_t)(op2)) +#define vfadd_vf_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfadd_vf_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (double)(op3), (size_t)(op4)) +#define vfadd_vf_f64m8(op0, op1, op2) \ +__builtin_rvv_vfadd_vf_f64m8((vfloat64m8_t)(op0), (double)(op1), (size_t)(op2)) +#define vfadd_vf_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfadd_vf_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (double)(op3), (size_t)(op4)) +#define vfsub_vv_f64m1(op0, op1, op2) \ +__builtin_rvv_vfsub_vv_f64m1((vfloat64m1_t)(op0), (vfloat64m1_t)(op1), (size_t)(op2)) +#define vfsub_vv_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsub_vv_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfsub_vv_f64m2(op0, op1, op2) \ +__builtin_rvv_vfsub_vv_f64m2((vfloat64m2_t)(op0), (vfloat64m2_t)(op1), (size_t)(op2)) +#define vfsub_vv_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsub_vv_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vfsub_vv_f64m4(op0, op1, op2) \ +__builtin_rvv_vfsub_vv_f64m4((vfloat64m4_t)(op0), (vfloat64m4_t)(op1), (size_t)(op2)) +#define vfsub_vv_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsub_vv_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vfsub_vv_f64m8(op0, op1, op2) \ +__builtin_rvv_vfsub_vv_f64m8((vfloat64m8_t)(op0), (vfloat64m8_t)(op1), (size_t)(op2)) +#define vfsub_vv_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsub_vv_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (vfloat64m8_t)(op3), (size_t)(op4)) +#define vfsub_vf_f64m1(op0, op1, op2) \ +__builtin_rvv_vfsub_vf_f64m1((vfloat64m1_t)(op0), (double)(op1), (size_t)(op2)) +#define vfsub_vf_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsub_vf_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (double)(op3), (size_t)(op4)) +#define vfsub_vf_f64m2(op0, op1, op2) \ +__builtin_rvv_vfsub_vf_f64m2((vfloat64m2_t)(op0), (double)(op1), (size_t)(op2)) +#define vfsub_vf_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsub_vf_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (double)(op3), (size_t)(op4)) +#define vfsub_vf_f64m4(op0, op1, op2) \ +__builtin_rvv_vfsub_vf_f64m4((vfloat64m4_t)(op0), (double)(op1), (size_t)(op2)) +#define vfsub_vf_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsub_vf_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (double)(op3), (size_t)(op4)) +#define vfsub_vf_f64m8(op0, op1, op2) \ +__builtin_rvv_vfsub_vf_f64m8((vfloat64m8_t)(op0), (double)(op1), (size_t)(op2)) +#define vfsub_vf_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsub_vf_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (double)(op3), (size_t)(op4)) +#define vfrsub_vf_f64m1(op0, op1, op2) \ +__builtin_rvv_vfrsub_vf_f64m1((vfloat64m1_t)(op0), (double)(op1), (size_t)(op2)) +#define vfrsub_vf_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfrsub_vf_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (double)(op3), (size_t)(op4)) +#define vfrsub_vf_f64m2(op0, op1, op2) \ +__builtin_rvv_vfrsub_vf_f64m2((vfloat64m2_t)(op0), (double)(op1), (size_t)(op2)) +#define vfrsub_vf_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfrsub_vf_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (double)(op3), (size_t)(op4)) +#define vfrsub_vf_f64m4(op0, op1, op2) \ +__builtin_rvv_vfrsub_vf_f64m4((vfloat64m4_t)(op0), (double)(op1), (size_t)(op2)) +#define vfrsub_vf_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfrsub_vf_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (double)(op3), (size_t)(op4)) +#define vfrsub_vf_f64m8(op0, op1, op2) \ +__builtin_rvv_vfrsub_vf_f64m8((vfloat64m8_t)(op0), (double)(op1), (size_t)(op2)) +#define vfrsub_vf_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfrsub_vf_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (double)(op3), (size_t)(op4)) +#define vfmul_vv_f64m1(op0, op1, op2) \ +__builtin_rvv_vfmul_vv_f64m1((vfloat64m1_t)(op0), (vfloat64m1_t)(op1), (size_t)(op2)) +#define vfmul_vv_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmul_vv_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfmul_vv_f64m2(op0, op1, op2) \ +__builtin_rvv_vfmul_vv_f64m2((vfloat64m2_t)(op0), (vfloat64m2_t)(op1), (size_t)(op2)) +#define vfmul_vv_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmul_vv_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vfmul_vv_f64m4(op0, op1, op2) \ +__builtin_rvv_vfmul_vv_f64m4((vfloat64m4_t)(op0), (vfloat64m4_t)(op1), (size_t)(op2)) +#define vfmul_vv_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmul_vv_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vfmul_vv_f64m8(op0, op1, op2) \ +__builtin_rvv_vfmul_vv_f64m8((vfloat64m8_t)(op0), (vfloat64m8_t)(op1), (size_t)(op2)) +#define vfmul_vv_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmul_vv_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (vfloat64m8_t)(op3), (size_t)(op4)) +#define vfmul_vf_f64m1(op0, op1, op2) \ +__builtin_rvv_vfmul_vf_f64m1((vfloat64m1_t)(op0), (double)(op1), (size_t)(op2)) +#define vfmul_vf_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmul_vf_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (double)(op3), (size_t)(op4)) +#define vfmul_vf_f64m2(op0, op1, op2) \ +__builtin_rvv_vfmul_vf_f64m2((vfloat64m2_t)(op0), (double)(op1), (size_t)(op2)) +#define vfmul_vf_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmul_vf_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (double)(op3), (size_t)(op4)) +#define vfmul_vf_f64m4(op0, op1, op2) \ +__builtin_rvv_vfmul_vf_f64m4((vfloat64m4_t)(op0), (double)(op1), (size_t)(op2)) +#define vfmul_vf_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmul_vf_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (double)(op3), (size_t)(op4)) +#define vfmul_vf_f64m8(op0, op1, op2) \ +__builtin_rvv_vfmul_vf_f64m8((vfloat64m8_t)(op0), (double)(op1), (size_t)(op2)) +#define vfmul_vf_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmul_vf_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (double)(op3), (size_t)(op4)) +#define vfdiv_vv_f64m1(op0, op1, op2) \ +__builtin_rvv_vfdiv_vv_f64m1((vfloat64m1_t)(op0), (vfloat64m1_t)(op1), (size_t)(op2)) +#define vfdiv_vv_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfdiv_vv_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfdiv_vv_f64m2(op0, op1, op2) \ +__builtin_rvv_vfdiv_vv_f64m2((vfloat64m2_t)(op0), (vfloat64m2_t)(op1), (size_t)(op2)) +#define vfdiv_vv_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfdiv_vv_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vfdiv_vv_f64m4(op0, op1, op2) \ +__builtin_rvv_vfdiv_vv_f64m4((vfloat64m4_t)(op0), (vfloat64m4_t)(op1), (size_t)(op2)) +#define vfdiv_vv_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfdiv_vv_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vfdiv_vv_f64m8(op0, op1, op2) \ +__builtin_rvv_vfdiv_vv_f64m8((vfloat64m8_t)(op0), (vfloat64m8_t)(op1), (size_t)(op2)) +#define vfdiv_vv_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfdiv_vv_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (vfloat64m8_t)(op3), (size_t)(op4)) +#define vfdiv_vf_f64m1(op0, op1, op2) \ +__builtin_rvv_vfdiv_vf_f64m1((vfloat64m1_t)(op0), (double)(op1), (size_t)(op2)) +#define vfdiv_vf_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfdiv_vf_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (double)(op3), (size_t)(op4)) +#define vfdiv_vf_f64m2(op0, op1, op2) \ +__builtin_rvv_vfdiv_vf_f64m2((vfloat64m2_t)(op0), (double)(op1), (size_t)(op2)) +#define vfdiv_vf_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfdiv_vf_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (double)(op3), (size_t)(op4)) +#define vfdiv_vf_f64m4(op0, op1, op2) \ +__builtin_rvv_vfdiv_vf_f64m4((vfloat64m4_t)(op0), (double)(op1), (size_t)(op2)) +#define vfdiv_vf_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfdiv_vf_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (double)(op3), (size_t)(op4)) +#define vfdiv_vf_f64m8(op0, op1, op2) \ +__builtin_rvv_vfdiv_vf_f64m8((vfloat64m8_t)(op0), (double)(op1), (size_t)(op2)) +#define vfdiv_vf_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfdiv_vf_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (double)(op3), (size_t)(op4)) +#define vfrdiv_vf_f64m1(op0, op1, op2) \ +__builtin_rvv_vfrdiv_vf_f64m1((vfloat64m1_t)(op0), (double)(op1), (size_t)(op2)) +#define vfrdiv_vf_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfrdiv_vf_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (double)(op3), (size_t)(op4)) +#define vfrdiv_vf_f64m2(op0, op1, op2) \ +__builtin_rvv_vfrdiv_vf_f64m2((vfloat64m2_t)(op0), (double)(op1), (size_t)(op2)) +#define vfrdiv_vf_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfrdiv_vf_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (double)(op3), (size_t)(op4)) +#define vfrdiv_vf_f64m4(op0, op1, op2) \ +__builtin_rvv_vfrdiv_vf_f64m4((vfloat64m4_t)(op0), (double)(op1), (size_t)(op2)) +#define vfrdiv_vf_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfrdiv_vf_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (double)(op3), (size_t)(op4)) +#define vfrdiv_vf_f64m8(op0, op1, op2) \ +__builtin_rvv_vfrdiv_vf_f64m8((vfloat64m8_t)(op0), (double)(op1), (size_t)(op2)) +#define vfrdiv_vf_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfrdiv_vf_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (double)(op3), (size_t)(op4)) +#define vfmacc_vv_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfmacc_vv_f64m1((vfloat64m1_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfmacc_vv_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmacc_vv_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfmacc_vv_f64m2(op0, op1, op2, op3) \ +__builtin_rvv_vfmacc_vv_f64m2((vfloat64m2_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vfmacc_vv_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmacc_vv_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vfmacc_vv_f64m4(op0, op1, op2, op3) \ +__builtin_rvv_vfmacc_vv_f64m4((vfloat64m4_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vfmacc_vv_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmacc_vv_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vfmacc_vv_f64m8(op0, op1, op2, op3) \ +__builtin_rvv_vfmacc_vv_f64m8((vfloat64m8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vfmacc_vv_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmacc_vv_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (vfloat64m8_t)(op3), (size_t)(op4)) +#define vfmacc_vf_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfmacc_vf_f64m1((vfloat64m1_t)(op0), (double)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfmacc_vf_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmacc_vf_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (double)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfmacc_vf_f64m2(op0, op1, op2, op3) \ +__builtin_rvv_vfmacc_vf_f64m2((vfloat64m2_t)(op0), (double)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vfmacc_vf_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmacc_vf_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (double)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vfmacc_vf_f64m4(op0, op1, op2, op3) \ +__builtin_rvv_vfmacc_vf_f64m4((vfloat64m4_t)(op0), (double)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vfmacc_vf_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmacc_vf_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (double)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vfmacc_vf_f64m8(op0, op1, op2, op3) \ +__builtin_rvv_vfmacc_vf_f64m8((vfloat64m8_t)(op0), (double)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vfmacc_vf_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmacc_vf_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (double)(op2), (vfloat64m8_t)(op3), (size_t)(op4)) +#define vfnmacc_vv_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfnmacc_vv_f64m1((vfloat64m1_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfnmacc_vv_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmacc_vv_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfnmacc_vv_f64m2(op0, op1, op2, op3) \ +__builtin_rvv_vfnmacc_vv_f64m2((vfloat64m2_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vfnmacc_vv_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmacc_vv_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vfnmacc_vv_f64m4(op0, op1, op2, op3) \ +__builtin_rvv_vfnmacc_vv_f64m4((vfloat64m4_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vfnmacc_vv_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmacc_vv_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vfnmacc_vv_f64m8(op0, op1, op2, op3) \ +__builtin_rvv_vfnmacc_vv_f64m8((vfloat64m8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vfnmacc_vv_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmacc_vv_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (vfloat64m8_t)(op3), (size_t)(op4)) +#define vfnmacc_vf_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfnmacc_vf_f64m1((vfloat64m1_t)(op0), (double)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfnmacc_vf_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmacc_vf_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (double)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfnmacc_vf_f64m2(op0, op1, op2, op3) \ +__builtin_rvv_vfnmacc_vf_f64m2((vfloat64m2_t)(op0), (double)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vfnmacc_vf_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmacc_vf_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (double)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vfnmacc_vf_f64m4(op0, op1, op2, op3) \ +__builtin_rvv_vfnmacc_vf_f64m4((vfloat64m4_t)(op0), (double)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vfnmacc_vf_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmacc_vf_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (double)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vfnmacc_vf_f64m8(op0, op1, op2, op3) \ +__builtin_rvv_vfnmacc_vf_f64m8((vfloat64m8_t)(op0), (double)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vfnmacc_vf_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmacc_vf_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (double)(op2), (vfloat64m8_t)(op3), (size_t)(op4)) +#define vfmsac_vv_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfmsac_vv_f64m1((vfloat64m1_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfmsac_vv_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsac_vv_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfmsac_vv_f64m2(op0, op1, op2, op3) \ +__builtin_rvv_vfmsac_vv_f64m2((vfloat64m2_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vfmsac_vv_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsac_vv_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vfmsac_vv_f64m4(op0, op1, op2, op3) \ +__builtin_rvv_vfmsac_vv_f64m4((vfloat64m4_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vfmsac_vv_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsac_vv_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vfmsac_vv_f64m8(op0, op1, op2, op3) \ +__builtin_rvv_vfmsac_vv_f64m8((vfloat64m8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vfmsac_vv_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsac_vv_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (vfloat64m8_t)(op3), (size_t)(op4)) +#define vfmsac_vf_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfmsac_vf_f64m1((vfloat64m1_t)(op0), (double)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfmsac_vf_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsac_vf_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (double)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfmsac_vf_f64m2(op0, op1, op2, op3) \ +__builtin_rvv_vfmsac_vf_f64m2((vfloat64m2_t)(op0), (double)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vfmsac_vf_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsac_vf_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (double)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vfmsac_vf_f64m4(op0, op1, op2, op3) \ +__builtin_rvv_vfmsac_vf_f64m4((vfloat64m4_t)(op0), (double)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vfmsac_vf_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsac_vf_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (double)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vfmsac_vf_f64m8(op0, op1, op2, op3) \ +__builtin_rvv_vfmsac_vf_f64m8((vfloat64m8_t)(op0), (double)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vfmsac_vf_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsac_vf_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (double)(op2), (vfloat64m8_t)(op3), (size_t)(op4)) +#define vfnmsac_vv_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsac_vv_f64m1((vfloat64m1_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfnmsac_vv_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsac_vv_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfnmsac_vv_f64m2(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsac_vv_f64m2((vfloat64m2_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vfnmsac_vv_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsac_vv_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vfnmsac_vv_f64m4(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsac_vv_f64m4((vfloat64m4_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vfnmsac_vv_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsac_vv_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vfnmsac_vv_f64m8(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsac_vv_f64m8((vfloat64m8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vfnmsac_vv_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsac_vv_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (vfloat64m8_t)(op3), (size_t)(op4)) +#define vfnmsac_vf_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsac_vf_f64m1((vfloat64m1_t)(op0), (double)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfnmsac_vf_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsac_vf_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (double)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfnmsac_vf_f64m2(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsac_vf_f64m2((vfloat64m2_t)(op0), (double)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vfnmsac_vf_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsac_vf_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (double)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vfnmsac_vf_f64m4(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsac_vf_f64m4((vfloat64m4_t)(op0), (double)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vfnmsac_vf_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsac_vf_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (double)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vfnmsac_vf_f64m8(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsac_vf_f64m8((vfloat64m8_t)(op0), (double)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vfnmsac_vf_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsac_vf_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (double)(op2), (vfloat64m8_t)(op3), (size_t)(op4)) +#define vfmadd_vv_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfmadd_vv_f64m1((vfloat64m1_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfmadd_vv_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmadd_vv_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfmadd_vv_f64m2(op0, op1, op2, op3) \ +__builtin_rvv_vfmadd_vv_f64m2((vfloat64m2_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vfmadd_vv_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmadd_vv_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vfmadd_vv_f64m4(op0, op1, op2, op3) \ +__builtin_rvv_vfmadd_vv_f64m4((vfloat64m4_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vfmadd_vv_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmadd_vv_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vfmadd_vv_f64m8(op0, op1, op2, op3) \ +__builtin_rvv_vfmadd_vv_f64m8((vfloat64m8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vfmadd_vv_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmadd_vv_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (vfloat64m8_t)(op3), (size_t)(op4)) +#define vfmadd_vf_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfmadd_vf_f64m1((vfloat64m1_t)(op0), (double)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfmadd_vf_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmadd_vf_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (double)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfmadd_vf_f64m2(op0, op1, op2, op3) \ +__builtin_rvv_vfmadd_vf_f64m2((vfloat64m2_t)(op0), (double)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vfmadd_vf_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmadd_vf_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (double)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vfmadd_vf_f64m4(op0, op1, op2, op3) \ +__builtin_rvv_vfmadd_vf_f64m4((vfloat64m4_t)(op0), (double)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vfmadd_vf_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmadd_vf_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (double)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vfmadd_vf_f64m8(op0, op1, op2, op3) \ +__builtin_rvv_vfmadd_vf_f64m8((vfloat64m8_t)(op0), (double)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vfmadd_vf_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmadd_vf_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (double)(op2), (vfloat64m8_t)(op3), (size_t)(op4)) +#define vfnmadd_vv_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfnmadd_vv_f64m1((vfloat64m1_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfnmadd_vv_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmadd_vv_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfnmadd_vv_f64m2(op0, op1, op2, op3) \ +__builtin_rvv_vfnmadd_vv_f64m2((vfloat64m2_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vfnmadd_vv_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmadd_vv_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vfnmadd_vv_f64m4(op0, op1, op2, op3) \ +__builtin_rvv_vfnmadd_vv_f64m4((vfloat64m4_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vfnmadd_vv_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmadd_vv_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vfnmadd_vv_f64m8(op0, op1, op2, op3) \ +__builtin_rvv_vfnmadd_vv_f64m8((vfloat64m8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vfnmadd_vv_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmadd_vv_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (vfloat64m8_t)(op3), (size_t)(op4)) +#define vfnmadd_vf_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfnmadd_vf_f64m1((vfloat64m1_t)(op0), (double)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfnmadd_vf_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmadd_vf_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (double)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfnmadd_vf_f64m2(op0, op1, op2, op3) \ +__builtin_rvv_vfnmadd_vf_f64m2((vfloat64m2_t)(op0), (double)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vfnmadd_vf_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmadd_vf_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (double)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vfnmadd_vf_f64m4(op0, op1, op2, op3) \ +__builtin_rvv_vfnmadd_vf_f64m4((vfloat64m4_t)(op0), (double)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vfnmadd_vf_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmadd_vf_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (double)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vfnmadd_vf_f64m8(op0, op1, op2, op3) \ +__builtin_rvv_vfnmadd_vf_f64m8((vfloat64m8_t)(op0), (double)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vfnmadd_vf_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmadd_vf_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (double)(op2), (vfloat64m8_t)(op3), (size_t)(op4)) +#define vfmsub_vv_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfmsub_vv_f64m1((vfloat64m1_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfmsub_vv_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsub_vv_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfmsub_vv_f64m2(op0, op1, op2, op3) \ +__builtin_rvv_vfmsub_vv_f64m2((vfloat64m2_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vfmsub_vv_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsub_vv_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vfmsub_vv_f64m4(op0, op1, op2, op3) \ +__builtin_rvv_vfmsub_vv_f64m4((vfloat64m4_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vfmsub_vv_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsub_vv_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vfmsub_vv_f64m8(op0, op1, op2, op3) \ +__builtin_rvv_vfmsub_vv_f64m8((vfloat64m8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vfmsub_vv_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsub_vv_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (vfloat64m8_t)(op3), (size_t)(op4)) +#define vfmsub_vf_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfmsub_vf_f64m1((vfloat64m1_t)(op0), (double)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfmsub_vf_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsub_vf_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (double)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfmsub_vf_f64m2(op0, op1, op2, op3) \ +__builtin_rvv_vfmsub_vf_f64m2((vfloat64m2_t)(op0), (double)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vfmsub_vf_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsub_vf_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (double)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vfmsub_vf_f64m4(op0, op1, op2, op3) \ +__builtin_rvv_vfmsub_vf_f64m4((vfloat64m4_t)(op0), (double)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vfmsub_vf_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsub_vf_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (double)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vfmsub_vf_f64m8(op0, op1, op2, op3) \ +__builtin_rvv_vfmsub_vf_f64m8((vfloat64m8_t)(op0), (double)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vfmsub_vf_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsub_vf_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (double)(op2), (vfloat64m8_t)(op3), (size_t)(op4)) +#define vfnmsub_vv_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsub_vv_f64m1((vfloat64m1_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfnmsub_vv_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsub_vv_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfnmsub_vv_f64m2(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsub_vv_f64m2((vfloat64m2_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vfnmsub_vv_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsub_vv_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vfnmsub_vv_f64m4(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsub_vv_f64m4((vfloat64m4_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vfnmsub_vv_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsub_vv_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vfnmsub_vv_f64m8(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsub_vv_f64m8((vfloat64m8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vfnmsub_vv_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsub_vv_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (vfloat64m8_t)(op3), (size_t)(op4)) +#define vfnmsub_vf_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsub_vf_f64m1((vfloat64m1_t)(op0), (double)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfnmsub_vf_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsub_vf_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (double)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfnmsub_vf_f64m2(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsub_vf_f64m2((vfloat64m2_t)(op0), (double)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vfnmsub_vf_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsub_vf_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (double)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vfnmsub_vf_f64m4(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsub_vf_f64m4((vfloat64m4_t)(op0), (double)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vfnmsub_vf_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsub_vf_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (double)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vfnmsub_vf_f64m8(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsub_vf_f64m8((vfloat64m8_t)(op0), (double)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vfnmsub_vf_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsub_vf_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (double)(op2), (vfloat64m8_t)(op3), (size_t)(op4)) +#define vfmin_vv_f64m1(op0, op1, op2) \ +__builtin_rvv_vfmin_vv_f64m1((vfloat64m1_t)(op0), (vfloat64m1_t)(op1), (size_t)(op2)) +#define vfmin_vv_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmin_vv_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfmin_vv_f64m2(op0, op1, op2) \ +__builtin_rvv_vfmin_vv_f64m2((vfloat64m2_t)(op0), (vfloat64m2_t)(op1), (size_t)(op2)) +#define vfmin_vv_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmin_vv_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vfmin_vv_f64m4(op0, op1, op2) \ +__builtin_rvv_vfmin_vv_f64m4((vfloat64m4_t)(op0), (vfloat64m4_t)(op1), (size_t)(op2)) +#define vfmin_vv_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmin_vv_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vfmin_vv_f64m8(op0, op1, op2) \ +__builtin_rvv_vfmin_vv_f64m8((vfloat64m8_t)(op0), (vfloat64m8_t)(op1), (size_t)(op2)) +#define vfmin_vv_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmin_vv_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (vfloat64m8_t)(op3), (size_t)(op4)) +#define vfmin_vf_f64m1(op0, op1, op2) \ +__builtin_rvv_vfmin_vf_f64m1((vfloat64m1_t)(op0), (double)(op1), (size_t)(op2)) +#define vfmin_vf_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmin_vf_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (double)(op3), (size_t)(op4)) +#define vfmin_vf_f64m2(op0, op1, op2) \ +__builtin_rvv_vfmin_vf_f64m2((vfloat64m2_t)(op0), (double)(op1), (size_t)(op2)) +#define vfmin_vf_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmin_vf_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (double)(op3), (size_t)(op4)) +#define vfmin_vf_f64m4(op0, op1, op2) \ +__builtin_rvv_vfmin_vf_f64m4((vfloat64m4_t)(op0), (double)(op1), (size_t)(op2)) +#define vfmin_vf_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmin_vf_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (double)(op3), (size_t)(op4)) +#define vfmin_vf_f64m8(op0, op1, op2) \ +__builtin_rvv_vfmin_vf_f64m8((vfloat64m8_t)(op0), (double)(op1), (size_t)(op2)) +#define vfmin_vf_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmin_vf_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (double)(op3), (size_t)(op4)) +#define vfmax_vv_f64m1(op0, op1, op2) \ +__builtin_rvv_vfmax_vv_f64m1((vfloat64m1_t)(op0), (vfloat64m1_t)(op1), (size_t)(op2)) +#define vfmax_vv_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmax_vv_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfmax_vv_f64m2(op0, op1, op2) \ +__builtin_rvv_vfmax_vv_f64m2((vfloat64m2_t)(op0), (vfloat64m2_t)(op1), (size_t)(op2)) +#define vfmax_vv_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmax_vv_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vfmax_vv_f64m4(op0, op1, op2) \ +__builtin_rvv_vfmax_vv_f64m4((vfloat64m4_t)(op0), (vfloat64m4_t)(op1), (size_t)(op2)) +#define vfmax_vv_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmax_vv_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vfmax_vv_f64m8(op0, op1, op2) \ +__builtin_rvv_vfmax_vv_f64m8((vfloat64m8_t)(op0), (vfloat64m8_t)(op1), (size_t)(op2)) +#define vfmax_vv_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmax_vv_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (vfloat64m8_t)(op3), (size_t)(op4)) +#define vfmax_vf_f64m1(op0, op1, op2) \ +__builtin_rvv_vfmax_vf_f64m1((vfloat64m1_t)(op0), (double)(op1), (size_t)(op2)) +#define vfmax_vf_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmax_vf_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (double)(op3), (size_t)(op4)) +#define vfmax_vf_f64m2(op0, op1, op2) \ +__builtin_rvv_vfmax_vf_f64m2((vfloat64m2_t)(op0), (double)(op1), (size_t)(op2)) +#define vfmax_vf_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmax_vf_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (double)(op3), (size_t)(op4)) +#define vfmax_vf_f64m4(op0, op1, op2) \ +__builtin_rvv_vfmax_vf_f64m4((vfloat64m4_t)(op0), (double)(op1), (size_t)(op2)) +#define vfmax_vf_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmax_vf_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (double)(op3), (size_t)(op4)) +#define vfmax_vf_f64m8(op0, op1, op2) \ +__builtin_rvv_vfmax_vf_f64m8((vfloat64m8_t)(op0), (double)(op1), (size_t)(op2)) +#define vfmax_vf_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmax_vf_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (double)(op3), (size_t)(op4)) +#define vfsgnj_vv_f64m1(op0, op1, op2) \ +__builtin_rvv_vfsgnj_vv_f64m1((vfloat64m1_t)(op0), (vfloat64m1_t)(op1), (size_t)(op2)) +#define vfsgnj_vv_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnj_vv_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfsgnj_vv_f64m2(op0, op1, op2) \ +__builtin_rvv_vfsgnj_vv_f64m2((vfloat64m2_t)(op0), (vfloat64m2_t)(op1), (size_t)(op2)) +#define vfsgnj_vv_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnj_vv_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vfsgnj_vv_f64m4(op0, op1, op2) \ +__builtin_rvv_vfsgnj_vv_f64m4((vfloat64m4_t)(op0), (vfloat64m4_t)(op1), (size_t)(op2)) +#define vfsgnj_vv_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnj_vv_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vfsgnj_vv_f64m8(op0, op1, op2) \ +__builtin_rvv_vfsgnj_vv_f64m8((vfloat64m8_t)(op0), (vfloat64m8_t)(op1), (size_t)(op2)) +#define vfsgnj_vv_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnj_vv_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (vfloat64m8_t)(op3), (size_t)(op4)) +#define vfsgnj_vf_f64m1(op0, op1, op2) \ +__builtin_rvv_vfsgnj_vf_f64m1((vfloat64m1_t)(op0), (double)(op1), (size_t)(op2)) +#define vfsgnj_vf_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnj_vf_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (double)(op3), (size_t)(op4)) +#define vfsgnj_vf_f64m2(op0, op1, op2) \ +__builtin_rvv_vfsgnj_vf_f64m2((vfloat64m2_t)(op0), (double)(op1), (size_t)(op2)) +#define vfsgnj_vf_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnj_vf_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (double)(op3), (size_t)(op4)) +#define vfsgnj_vf_f64m4(op0, op1, op2) \ +__builtin_rvv_vfsgnj_vf_f64m4((vfloat64m4_t)(op0), (double)(op1), (size_t)(op2)) +#define vfsgnj_vf_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnj_vf_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (double)(op3), (size_t)(op4)) +#define vfsgnj_vf_f64m8(op0, op1, op2) \ +__builtin_rvv_vfsgnj_vf_f64m8((vfloat64m8_t)(op0), (double)(op1), (size_t)(op2)) +#define vfsgnj_vf_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnj_vf_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (double)(op3), (size_t)(op4)) +#define vfsgnjn_vv_f64m1(op0, op1, op2) \ +__builtin_rvv_vfsgnjn_vv_f64m1((vfloat64m1_t)(op0), (vfloat64m1_t)(op1), (size_t)(op2)) +#define vfsgnjn_vv_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjn_vv_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfsgnjn_vv_f64m2(op0, op1, op2) \ +__builtin_rvv_vfsgnjn_vv_f64m2((vfloat64m2_t)(op0), (vfloat64m2_t)(op1), (size_t)(op2)) +#define vfsgnjn_vv_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjn_vv_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vfsgnjn_vv_f64m4(op0, op1, op2) \ +__builtin_rvv_vfsgnjn_vv_f64m4((vfloat64m4_t)(op0), (vfloat64m4_t)(op1), (size_t)(op2)) +#define vfsgnjn_vv_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjn_vv_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vfsgnjn_vv_f64m8(op0, op1, op2) \ +__builtin_rvv_vfsgnjn_vv_f64m8((vfloat64m8_t)(op0), (vfloat64m8_t)(op1), (size_t)(op2)) +#define vfsgnjn_vv_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjn_vv_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (vfloat64m8_t)(op3), (size_t)(op4)) +#define vfsgnjn_vf_f64m1(op0, op1, op2) \ +__builtin_rvv_vfsgnjn_vf_f64m1((vfloat64m1_t)(op0), (double)(op1), (size_t)(op2)) +#define vfsgnjn_vf_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjn_vf_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (double)(op3), (size_t)(op4)) +#define vfsgnjn_vf_f64m2(op0, op1, op2) \ +__builtin_rvv_vfsgnjn_vf_f64m2((vfloat64m2_t)(op0), (double)(op1), (size_t)(op2)) +#define vfsgnjn_vf_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjn_vf_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (double)(op3), (size_t)(op4)) +#define vfsgnjn_vf_f64m4(op0, op1, op2) \ +__builtin_rvv_vfsgnjn_vf_f64m4((vfloat64m4_t)(op0), (double)(op1), (size_t)(op2)) +#define vfsgnjn_vf_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjn_vf_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (double)(op3), (size_t)(op4)) +#define vfsgnjn_vf_f64m8(op0, op1, op2) \ +__builtin_rvv_vfsgnjn_vf_f64m8((vfloat64m8_t)(op0), (double)(op1), (size_t)(op2)) +#define vfsgnjn_vf_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjn_vf_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (double)(op3), (size_t)(op4)) +#define vfsgnjx_vv_f64m1(op0, op1, op2) \ +__builtin_rvv_vfsgnjx_vv_f64m1((vfloat64m1_t)(op0), (vfloat64m1_t)(op1), (size_t)(op2)) +#define vfsgnjx_vv_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjx_vv_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfsgnjx_vv_f64m2(op0, op1, op2) \ +__builtin_rvv_vfsgnjx_vv_f64m2((vfloat64m2_t)(op0), (vfloat64m2_t)(op1), (size_t)(op2)) +#define vfsgnjx_vv_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjx_vv_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vfsgnjx_vv_f64m4(op0, op1, op2) \ +__builtin_rvv_vfsgnjx_vv_f64m4((vfloat64m4_t)(op0), (vfloat64m4_t)(op1), (size_t)(op2)) +#define vfsgnjx_vv_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjx_vv_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vfsgnjx_vv_f64m8(op0, op1, op2) \ +__builtin_rvv_vfsgnjx_vv_f64m8((vfloat64m8_t)(op0), (vfloat64m8_t)(op1), (size_t)(op2)) +#define vfsgnjx_vv_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjx_vv_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (vfloat64m8_t)(op3), (size_t)(op4)) +#define vfsgnjx_vf_f64m1(op0, op1, op2) \ +__builtin_rvv_vfsgnjx_vf_f64m1((vfloat64m1_t)(op0), (double)(op1), (size_t)(op2)) +#define vfsgnjx_vf_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjx_vf_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (double)(op3), (size_t)(op4)) +#define vfsgnjx_vf_f64m2(op0, op1, op2) \ +__builtin_rvv_vfsgnjx_vf_f64m2((vfloat64m2_t)(op0), (double)(op1), (size_t)(op2)) +#define vfsgnjx_vf_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjx_vf_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (double)(op3), (size_t)(op4)) +#define vfsgnjx_vf_f64m4(op0, op1, op2) \ +__builtin_rvv_vfsgnjx_vf_f64m4((vfloat64m4_t)(op0), (double)(op1), (size_t)(op2)) +#define vfsgnjx_vf_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjx_vf_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (double)(op3), (size_t)(op4)) +#define vfsgnjx_vf_f64m8(op0, op1, op2) \ +__builtin_rvv_vfsgnjx_vf_f64m8((vfloat64m8_t)(op0), (double)(op1), (size_t)(op2)) +#define vfsgnjx_vf_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjx_vf_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (double)(op3), (size_t)(op4)) +#define vfabs_v_f64m1(op0, op1) \ +__builtin_rvv_vfabs_v_f64m1((vfloat64m1_t)(op0), (size_t)(op1)) +#define vfabs_v_f64m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfabs_v_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfabs_v_f64m2(op0, op1) \ +__builtin_rvv_vfabs_v_f64m2((vfloat64m2_t)(op0), (size_t)(op1)) +#define vfabs_v_f64m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfabs_v_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vfabs_v_f64m4(op0, op1) \ +__builtin_rvv_vfabs_v_f64m4((vfloat64m4_t)(op0), (size_t)(op1)) +#define vfabs_v_f64m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfabs_v_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vfabs_v_f64m8(op0, op1) \ +__builtin_rvv_vfabs_v_f64m8((vfloat64m8_t)(op0), (size_t)(op1)) +#define vfabs_v_f64m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfabs_v_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vmfeq_vv_f64m1_b64(op0, op1, op2) \ +__builtin_rvv_vmfeq_vv_f64m1_b64((vfloat64m1_t)(op0), (vfloat64m1_t)(op1), (size_t)(op2)) +#define vmfeq_vv_f64m1_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfeq_vv_f64m1_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vmfeq_vv_f64m2_b32(op0, op1, op2) \ +__builtin_rvv_vmfeq_vv_f64m2_b32((vfloat64m2_t)(op0), (vfloat64m2_t)(op1), (size_t)(op2)) +#define vmfeq_vv_f64m2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfeq_vv_f64m2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vmfeq_vv_f64m4_b16(op0, op1, op2) \ +__builtin_rvv_vmfeq_vv_f64m4_b16((vfloat64m4_t)(op0), (vfloat64m4_t)(op1), (size_t)(op2)) +#define vmfeq_vv_f64m4_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfeq_vv_f64m4_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vfloat64m4_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vmfeq_vv_f64m8_b8(op0, op1, op2) \ +__builtin_rvv_vmfeq_vv_f64m8_b8((vfloat64m8_t)(op0), (vfloat64m8_t)(op1), (size_t)(op2)) +#define vmfeq_vv_f64m8_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfeq_vv_f64m8_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vfloat64m8_t)(op2), (vfloat64m8_t)(op3), (size_t)(op4)) +#define vmfeq_vf_f64m1_b64(op0, op1, op2) \ +__builtin_rvv_vmfeq_vf_f64m1_b64((vfloat64m1_t)(op0), (double)(op1), (size_t)(op2)) +#define vmfeq_vf_f64m1_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfeq_vf_f64m1_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vfloat64m1_t)(op2), (double)(op3), (size_t)(op4)) +#define vmfeq_vf_f64m2_b32(op0, op1, op2) \ +__builtin_rvv_vmfeq_vf_f64m2_b32((vfloat64m2_t)(op0), (double)(op1), (size_t)(op2)) +#define vmfeq_vf_f64m2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfeq_vf_f64m2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vfloat64m2_t)(op2), (double)(op3), (size_t)(op4)) +#define vmfeq_vf_f64m4_b16(op0, op1, op2) \ +__builtin_rvv_vmfeq_vf_f64m4_b16((vfloat64m4_t)(op0), (double)(op1), (size_t)(op2)) +#define vmfeq_vf_f64m4_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfeq_vf_f64m4_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vfloat64m4_t)(op2), (double)(op3), (size_t)(op4)) +#define vmfeq_vf_f64m8_b8(op0, op1, op2) \ +__builtin_rvv_vmfeq_vf_f64m8_b8((vfloat64m8_t)(op0), (double)(op1), (size_t)(op2)) +#define vmfeq_vf_f64m8_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfeq_vf_f64m8_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vfloat64m8_t)(op2), (double)(op3), (size_t)(op4)) +#define vmfne_vv_f64m1_b64(op0, op1, op2) \ +__builtin_rvv_vmfne_vv_f64m1_b64((vfloat64m1_t)(op0), (vfloat64m1_t)(op1), (size_t)(op2)) +#define vmfne_vv_f64m1_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfne_vv_f64m1_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vmfne_vv_f64m2_b32(op0, op1, op2) \ +__builtin_rvv_vmfne_vv_f64m2_b32((vfloat64m2_t)(op0), (vfloat64m2_t)(op1), (size_t)(op2)) +#define vmfne_vv_f64m2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfne_vv_f64m2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vmfne_vv_f64m4_b16(op0, op1, op2) \ +__builtin_rvv_vmfne_vv_f64m4_b16((vfloat64m4_t)(op0), (vfloat64m4_t)(op1), (size_t)(op2)) +#define vmfne_vv_f64m4_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfne_vv_f64m4_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vfloat64m4_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vmfne_vv_f64m8_b8(op0, op1, op2) \ +__builtin_rvv_vmfne_vv_f64m8_b8((vfloat64m8_t)(op0), (vfloat64m8_t)(op1), (size_t)(op2)) +#define vmfne_vv_f64m8_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfne_vv_f64m8_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vfloat64m8_t)(op2), (vfloat64m8_t)(op3), (size_t)(op4)) +#define vmfne_vf_f64m1_b64(op0, op1, op2) \ +__builtin_rvv_vmfne_vf_f64m1_b64((vfloat64m1_t)(op0), (double)(op1), (size_t)(op2)) +#define vmfne_vf_f64m1_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfne_vf_f64m1_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vfloat64m1_t)(op2), (double)(op3), (size_t)(op4)) +#define vmfne_vf_f64m2_b32(op0, op1, op2) \ +__builtin_rvv_vmfne_vf_f64m2_b32((vfloat64m2_t)(op0), (double)(op1), (size_t)(op2)) +#define vmfne_vf_f64m2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfne_vf_f64m2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vfloat64m2_t)(op2), (double)(op3), (size_t)(op4)) +#define vmfne_vf_f64m4_b16(op0, op1, op2) \ +__builtin_rvv_vmfne_vf_f64m4_b16((vfloat64m4_t)(op0), (double)(op1), (size_t)(op2)) +#define vmfne_vf_f64m4_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfne_vf_f64m4_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vfloat64m4_t)(op2), (double)(op3), (size_t)(op4)) +#define vmfne_vf_f64m8_b8(op0, op1, op2) \ +__builtin_rvv_vmfne_vf_f64m8_b8((vfloat64m8_t)(op0), (double)(op1), (size_t)(op2)) +#define vmfne_vf_f64m8_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfne_vf_f64m8_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vfloat64m8_t)(op2), (double)(op3), (size_t)(op4)) +#define vmflt_vv_f64m1_b64(op0, op1, op2) \ +__builtin_rvv_vmflt_vv_f64m1_b64((vfloat64m1_t)(op0), (vfloat64m1_t)(op1), (size_t)(op2)) +#define vmflt_vv_f64m1_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmflt_vv_f64m1_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vmflt_vv_f64m2_b32(op0, op1, op2) \ +__builtin_rvv_vmflt_vv_f64m2_b32((vfloat64m2_t)(op0), (vfloat64m2_t)(op1), (size_t)(op2)) +#define vmflt_vv_f64m2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmflt_vv_f64m2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vmflt_vv_f64m4_b16(op0, op1, op2) \ +__builtin_rvv_vmflt_vv_f64m4_b16((vfloat64m4_t)(op0), (vfloat64m4_t)(op1), (size_t)(op2)) +#define vmflt_vv_f64m4_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmflt_vv_f64m4_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vfloat64m4_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vmflt_vv_f64m8_b8(op0, op1, op2) \ +__builtin_rvv_vmflt_vv_f64m8_b8((vfloat64m8_t)(op0), (vfloat64m8_t)(op1), (size_t)(op2)) +#define vmflt_vv_f64m8_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmflt_vv_f64m8_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vfloat64m8_t)(op2), (vfloat64m8_t)(op3), (size_t)(op4)) +#define vmflt_vf_f64m1_b64(op0, op1, op2) \ +__builtin_rvv_vmflt_vf_f64m1_b64((vfloat64m1_t)(op0), (double)(op1), (size_t)(op2)) +#define vmflt_vf_f64m1_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmflt_vf_f64m1_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vfloat64m1_t)(op2), (double)(op3), (size_t)(op4)) +#define vmflt_vf_f64m2_b32(op0, op1, op2) \ +__builtin_rvv_vmflt_vf_f64m2_b32((vfloat64m2_t)(op0), (double)(op1), (size_t)(op2)) +#define vmflt_vf_f64m2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmflt_vf_f64m2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vfloat64m2_t)(op2), (double)(op3), (size_t)(op4)) +#define vmflt_vf_f64m4_b16(op0, op1, op2) \ +__builtin_rvv_vmflt_vf_f64m4_b16((vfloat64m4_t)(op0), (double)(op1), (size_t)(op2)) +#define vmflt_vf_f64m4_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmflt_vf_f64m4_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vfloat64m4_t)(op2), (double)(op3), (size_t)(op4)) +#define vmflt_vf_f64m8_b8(op0, op1, op2) \ +__builtin_rvv_vmflt_vf_f64m8_b8((vfloat64m8_t)(op0), (double)(op1), (size_t)(op2)) +#define vmflt_vf_f64m8_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmflt_vf_f64m8_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vfloat64m8_t)(op2), (double)(op3), (size_t)(op4)) +#define vmfle_vv_f64m1_b64(op0, op1, op2) \ +__builtin_rvv_vmfle_vv_f64m1_b64((vfloat64m1_t)(op0), (vfloat64m1_t)(op1), (size_t)(op2)) +#define vmfle_vv_f64m1_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfle_vv_f64m1_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vmfle_vv_f64m2_b32(op0, op1, op2) \ +__builtin_rvv_vmfle_vv_f64m2_b32((vfloat64m2_t)(op0), (vfloat64m2_t)(op1), (size_t)(op2)) +#define vmfle_vv_f64m2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfle_vv_f64m2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vmfle_vv_f64m4_b16(op0, op1, op2) \ +__builtin_rvv_vmfle_vv_f64m4_b16((vfloat64m4_t)(op0), (vfloat64m4_t)(op1), (size_t)(op2)) +#define vmfle_vv_f64m4_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfle_vv_f64m4_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vfloat64m4_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vmfle_vv_f64m8_b8(op0, op1, op2) \ +__builtin_rvv_vmfle_vv_f64m8_b8((vfloat64m8_t)(op0), (vfloat64m8_t)(op1), (size_t)(op2)) +#define vmfle_vv_f64m8_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfle_vv_f64m8_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vfloat64m8_t)(op2), (vfloat64m8_t)(op3), (size_t)(op4)) +#define vmfle_vf_f64m1_b64(op0, op1, op2) \ +__builtin_rvv_vmfle_vf_f64m1_b64((vfloat64m1_t)(op0), (double)(op1), (size_t)(op2)) +#define vmfle_vf_f64m1_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfle_vf_f64m1_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vfloat64m1_t)(op2), (double)(op3), (size_t)(op4)) +#define vmfle_vf_f64m2_b32(op0, op1, op2) \ +__builtin_rvv_vmfle_vf_f64m2_b32((vfloat64m2_t)(op0), (double)(op1), (size_t)(op2)) +#define vmfle_vf_f64m2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfle_vf_f64m2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vfloat64m2_t)(op2), (double)(op3), (size_t)(op4)) +#define vmfle_vf_f64m4_b16(op0, op1, op2) \ +__builtin_rvv_vmfle_vf_f64m4_b16((vfloat64m4_t)(op0), (double)(op1), (size_t)(op2)) +#define vmfle_vf_f64m4_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfle_vf_f64m4_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vfloat64m4_t)(op2), (double)(op3), (size_t)(op4)) +#define vmfle_vf_f64m8_b8(op0, op1, op2) \ +__builtin_rvv_vmfle_vf_f64m8_b8((vfloat64m8_t)(op0), (double)(op1), (size_t)(op2)) +#define vmfle_vf_f64m8_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfle_vf_f64m8_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vfloat64m8_t)(op2), (double)(op3), (size_t)(op4)) +#define vmfgt_vv_f64m1_b64(op0, op1, op2) \ +__builtin_rvv_vmfgt_vv_f64m1_b64((vfloat64m1_t)(op0), (vfloat64m1_t)(op1), (size_t)(op2)) +#define vmfgt_vv_f64m1_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfgt_vv_f64m1_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vmfgt_vv_f64m2_b32(op0, op1, op2) \ +__builtin_rvv_vmfgt_vv_f64m2_b32((vfloat64m2_t)(op0), (vfloat64m2_t)(op1), (size_t)(op2)) +#define vmfgt_vv_f64m2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfgt_vv_f64m2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vmfgt_vv_f64m4_b16(op0, op1, op2) \ +__builtin_rvv_vmfgt_vv_f64m4_b16((vfloat64m4_t)(op0), (vfloat64m4_t)(op1), (size_t)(op2)) +#define vmfgt_vv_f64m4_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfgt_vv_f64m4_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vfloat64m4_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vmfgt_vv_f64m8_b8(op0, op1, op2) \ +__builtin_rvv_vmfgt_vv_f64m8_b8((vfloat64m8_t)(op0), (vfloat64m8_t)(op1), (size_t)(op2)) +#define vmfgt_vv_f64m8_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfgt_vv_f64m8_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vfloat64m8_t)(op2), (vfloat64m8_t)(op3), (size_t)(op4)) +#define vmfgt_vf_f64m1_b64(op0, op1, op2) \ +__builtin_rvv_vmfgt_vf_f64m1_b64((vfloat64m1_t)(op0), (double)(op1), (size_t)(op2)) +#define vmfgt_vf_f64m1_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfgt_vf_f64m1_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vfloat64m1_t)(op2), (double)(op3), (size_t)(op4)) +#define vmfgt_vf_f64m2_b32(op0, op1, op2) \ +__builtin_rvv_vmfgt_vf_f64m2_b32((vfloat64m2_t)(op0), (double)(op1), (size_t)(op2)) +#define vmfgt_vf_f64m2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfgt_vf_f64m2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vfloat64m2_t)(op2), (double)(op3), (size_t)(op4)) +#define vmfgt_vf_f64m4_b16(op0, op1, op2) \ +__builtin_rvv_vmfgt_vf_f64m4_b16((vfloat64m4_t)(op0), (double)(op1), (size_t)(op2)) +#define vmfgt_vf_f64m4_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfgt_vf_f64m4_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vfloat64m4_t)(op2), (double)(op3), (size_t)(op4)) +#define vmfgt_vf_f64m8_b8(op0, op1, op2) \ +__builtin_rvv_vmfgt_vf_f64m8_b8((vfloat64m8_t)(op0), (double)(op1), (size_t)(op2)) +#define vmfgt_vf_f64m8_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfgt_vf_f64m8_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vfloat64m8_t)(op2), (double)(op3), (size_t)(op4)) +#define vmfge_vv_f64m1_b64(op0, op1, op2) \ +__builtin_rvv_vmfge_vv_f64m1_b64((vfloat64m1_t)(op0), (vfloat64m1_t)(op1), (size_t)(op2)) +#define vmfge_vv_f64m1_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfge_vv_f64m1_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vmfge_vv_f64m2_b32(op0, op1, op2) \ +__builtin_rvv_vmfge_vv_f64m2_b32((vfloat64m2_t)(op0), (vfloat64m2_t)(op1), (size_t)(op2)) +#define vmfge_vv_f64m2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfge_vv_f64m2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vmfge_vv_f64m4_b16(op0, op1, op2) \ +__builtin_rvv_vmfge_vv_f64m4_b16((vfloat64m4_t)(op0), (vfloat64m4_t)(op1), (size_t)(op2)) +#define vmfge_vv_f64m4_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfge_vv_f64m4_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vfloat64m4_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vmfge_vv_f64m8_b8(op0, op1, op2) \ +__builtin_rvv_vmfge_vv_f64m8_b8((vfloat64m8_t)(op0), (vfloat64m8_t)(op1), (size_t)(op2)) +#define vmfge_vv_f64m8_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfge_vv_f64m8_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vfloat64m8_t)(op2), (vfloat64m8_t)(op3), (size_t)(op4)) +#define vmfge_vf_f64m1_b64(op0, op1, op2) \ +__builtin_rvv_vmfge_vf_f64m1_b64((vfloat64m1_t)(op0), (double)(op1), (size_t)(op2)) +#define vmfge_vf_f64m1_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfge_vf_f64m1_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vfloat64m1_t)(op2), (double)(op3), (size_t)(op4)) +#define vmfge_vf_f64m2_b32(op0, op1, op2) \ +__builtin_rvv_vmfge_vf_f64m2_b32((vfloat64m2_t)(op0), (double)(op1), (size_t)(op2)) +#define vmfge_vf_f64m2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfge_vf_f64m2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vfloat64m2_t)(op2), (double)(op3), (size_t)(op4)) +#define vmfge_vf_f64m4_b16(op0, op1, op2) \ +__builtin_rvv_vmfge_vf_f64m4_b16((vfloat64m4_t)(op0), (double)(op1), (size_t)(op2)) +#define vmfge_vf_f64m4_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfge_vf_f64m4_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vfloat64m4_t)(op2), (double)(op3), (size_t)(op4)) +#define vmfge_vf_f64m8_b8(op0, op1, op2) \ +__builtin_rvv_vmfge_vf_f64m8_b8((vfloat64m8_t)(op0), (double)(op1), (size_t)(op2)) +#define vmfge_vf_f64m8_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfge_vf_f64m8_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vfloat64m8_t)(op2), (double)(op3), (size_t)(op4)) +#define vmerge_vvm_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_f64m1((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vmerge_vvm_f64m2(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_f64m2((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vmerge_vvm_f64m4(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_f64m4((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vmerge_vvm_f64m8(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_f64m8((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vfmerge_vfm_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfmerge_vfm_f64m1((vbool64_t)(op0), (vfloat64m1_t)(op1), (double)(op2), (size_t)(op3)) +#define vfmerge_vfm_f64m2(op0, op1, op2, op3) \ +__builtin_rvv_vfmerge_vfm_f64m2((vbool32_t)(op0), (vfloat64m2_t)(op1), (double)(op2), (size_t)(op3)) +#define vfmerge_vfm_f64m4(op0, op1, op2, op3) \ +__builtin_rvv_vfmerge_vfm_f64m4((vbool16_t)(op0), (vfloat64m4_t)(op1), (double)(op2), (size_t)(op3)) +#define vfmerge_vfm_f64m8(op0, op1, op2, op3) \ +__builtin_rvv_vfmerge_vfm_f64m8((vbool8_t)(op0), (vfloat64m8_t)(op1), (double)(op2), (size_t)(op3)) +#define vfmv_v_f_f64m1(op0, op1) \ +__builtin_rvv_vfmv_v_f_f64m1((double)(op0), (size_t)(op1)) +#define vfmv_v_f_f64m2(op0, op1) \ +__builtin_rvv_vfmv_v_f_f64m2((double)(op0), (size_t)(op1)) +#define vfmv_v_f_f64m4(op0, op1) \ +__builtin_rvv_vfmv_v_f_f64m4((double)(op0), (size_t)(op1)) +#define vfmv_v_f_f64m8(op0, op1) \ +__builtin_rvv_vfmv_v_f_f64m8((double)(op0), (size_t)(op1)) +#define vfredmax_vs_f64m1_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredmax_vs_f64m1_f64m1((vfloat64m1_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfredmax_vs_f64m1_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredmax_vs_f64m1_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfredmax_vs_f64m2_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredmax_vs_f64m2_f64m1((vfloat64m1_t)(op0), (vfloat64m2_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfredmax_vs_f64m2_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredmax_vs_f64m2_f64m1_m((vbool32_t)(op0), (vfloat64m1_t)(op1), (vfloat64m2_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfredmax_vs_f64m4_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredmax_vs_f64m4_f64m1((vfloat64m1_t)(op0), (vfloat64m4_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfredmax_vs_f64m4_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredmax_vs_f64m4_f64m1_m((vbool16_t)(op0), (vfloat64m1_t)(op1), (vfloat64m4_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfredmax_vs_f64m8_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredmax_vs_f64m8_f64m1((vfloat64m1_t)(op0), (vfloat64m8_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfredmax_vs_f64m8_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredmax_vs_f64m8_f64m1_m((vbool8_t)(op0), (vfloat64m1_t)(op1), (vfloat64m8_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfredmin_vs_f64m1_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredmin_vs_f64m1_f64m1((vfloat64m1_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfredmin_vs_f64m1_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredmin_vs_f64m1_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfredmin_vs_f64m2_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredmin_vs_f64m2_f64m1((vfloat64m1_t)(op0), (vfloat64m2_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfredmin_vs_f64m2_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredmin_vs_f64m2_f64m1_m((vbool32_t)(op0), (vfloat64m1_t)(op1), (vfloat64m2_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfredmin_vs_f64m4_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredmin_vs_f64m4_f64m1((vfloat64m1_t)(op0), (vfloat64m4_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfredmin_vs_f64m4_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredmin_vs_f64m4_f64m1_m((vbool16_t)(op0), (vfloat64m1_t)(op1), (vfloat64m4_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfredmin_vs_f64m8_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredmin_vs_f64m8_f64m1((vfloat64m1_t)(op0), (vfloat64m8_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfredmin_vs_f64m8_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredmin_vs_f64m8_f64m1_m((vbool8_t)(op0), (vfloat64m1_t)(op1), (vfloat64m8_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfredsum_vs_f64m1_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredsum_vs_f64m1_f64m1((vfloat64m1_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfredsum_vs_f64m1_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredsum_vs_f64m1_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfredsum_vs_f64m2_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredsum_vs_f64m2_f64m1((vfloat64m1_t)(op0), (vfloat64m2_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfredsum_vs_f64m2_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredsum_vs_f64m2_f64m1_m((vbool32_t)(op0), (vfloat64m1_t)(op1), (vfloat64m2_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfredsum_vs_f64m4_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredsum_vs_f64m4_f64m1((vfloat64m1_t)(op0), (vfloat64m4_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfredsum_vs_f64m4_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredsum_vs_f64m4_f64m1_m((vbool16_t)(op0), (vfloat64m1_t)(op1), (vfloat64m4_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfredsum_vs_f64m8_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredsum_vs_f64m8_f64m1((vfloat64m1_t)(op0), (vfloat64m8_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfredsum_vs_f64m8_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredsum_vs_f64m8_f64m1_m((vbool8_t)(op0), (vfloat64m1_t)(op1), (vfloat64m8_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfredosum_vs_f64m1_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredosum_vs_f64m1_f64m1((vfloat64m1_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfredosum_vs_f64m1_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredosum_vs_f64m1_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfredosum_vs_f64m2_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredosum_vs_f64m2_f64m1((vfloat64m1_t)(op0), (vfloat64m2_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfredosum_vs_f64m2_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredosum_vs_f64m2_f64m1_m((vbool32_t)(op0), (vfloat64m1_t)(op1), (vfloat64m2_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfredosum_vs_f64m4_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredosum_vs_f64m4_f64m1((vfloat64m1_t)(op0), (vfloat64m4_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfredosum_vs_f64m4_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredosum_vs_f64m4_f64m1_m((vbool16_t)(op0), (vfloat64m1_t)(op1), (vfloat64m4_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfredosum_vs_f64m8_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredosum_vs_f64m8_f64m1((vfloat64m1_t)(op0), (vfloat64m8_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfredosum_vs_f64m8_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredosum_vs_f64m8_f64m1_m((vbool8_t)(op0), (vfloat64m1_t)(op1), (vfloat64m8_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfmv_f_s_f64m1_f64(op0) \ +__builtin_rvv_vfmv_f_s_f64m1_f64((vfloat64m1_t)(op0)) +#define vfmv_f_s_f64m2_f64(op0) \ +__builtin_rvv_vfmv_f_s_f64m2_f64((vfloat64m2_t)(op0)) +#define vfmv_f_s_f64m4_f64(op0) \ +__builtin_rvv_vfmv_f_s_f64m4_f64((vfloat64m4_t)(op0)) +#define vfmv_f_s_f64m8_f64(op0) \ +__builtin_rvv_vfmv_f_s_f64m8_f64((vfloat64m8_t)(op0)) +#define vfmv_s_f_f64m1(op0, op1, op2) \ +__builtin_rvv_vfmv_s_f_f64m1((vfloat64m1_t)(op0), (double)(op1), (size_t)(op2)) +#define vfmv_s_f_f64m2(op0, op1, op2) \ +__builtin_rvv_vfmv_s_f_f64m2((vfloat64m2_t)(op0), (double)(op1), (size_t)(op2)) +#define vfmv_s_f_f64m4(op0, op1, op2) \ +__builtin_rvv_vfmv_s_f_f64m4((vfloat64m4_t)(op0), (double)(op1), (size_t)(op2)) +#define vfmv_s_f_f64m8(op0, op1, op2) \ +__builtin_rvv_vfmv_s_f_f64m8((vfloat64m8_t)(op0), (double)(op1), (size_t)(op2)) +#define vslideup_vx_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_f64m1((vfloat64m1_t)(op0), (vfloat64m1_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_f64m2(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_f64m2((vfloat64m2_t)(op0), (vfloat64m2_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_f64m4(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_f64m4((vfloat64m4_t)(op0), (vfloat64m4_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_f64m8(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_f64m8((vfloat64m8_t)(op0), (vfloat64m8_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_f64m1((vfloat64m1_t)(op0), (vfloat64m1_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_f64m2(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_f64m2((vfloat64m2_t)(op0), (vfloat64m2_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_f64m4(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_f64m4((vfloat64m4_t)(op0), (vfloat64m4_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_f64m8(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_f64m8((vfloat64m8_t)(op0), (vfloat64m8_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vfslide1up_vf_f64m1(op0, op1, op2) \ +__builtin_rvv_vfslide1up_vf_f64m1((vfloat64m1_t)(op0), (double)(op1), (size_t)(op2)) +#define vfslide1up_vf_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfslide1up_vf_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (double)(op3), (size_t)(op4)) +#define vfslide1up_vf_f64m2(op0, op1, op2) \ +__builtin_rvv_vfslide1up_vf_f64m2((vfloat64m2_t)(op0), (double)(op1), (size_t)(op2)) +#define vfslide1up_vf_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfslide1up_vf_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (double)(op3), (size_t)(op4)) +#define vfslide1up_vf_f64m4(op0, op1, op2) \ +__builtin_rvv_vfslide1up_vf_f64m4((vfloat64m4_t)(op0), (double)(op1), (size_t)(op2)) +#define vfslide1up_vf_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfslide1up_vf_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (double)(op3), (size_t)(op4)) +#define vfslide1up_vf_f64m8(op0, op1, op2) \ +__builtin_rvv_vfslide1up_vf_f64m8((vfloat64m8_t)(op0), (double)(op1), (size_t)(op2)) +#define vfslide1up_vf_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfslide1up_vf_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (double)(op3), (size_t)(op4)) +#define vfslide1down_vf_f64m1(op0, op1, op2) \ +__builtin_rvv_vfslide1down_vf_f64m1((vfloat64m1_t)(op0), (double)(op1), (size_t)(op2)) +#define vfslide1down_vf_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfslide1down_vf_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (double)(op3), (size_t)(op4)) +#define vfslide1down_vf_f64m2(op0, op1, op2) \ +__builtin_rvv_vfslide1down_vf_f64m2((vfloat64m2_t)(op0), (double)(op1), (size_t)(op2)) +#define vfslide1down_vf_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfslide1down_vf_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (double)(op3), (size_t)(op4)) +#define vfslide1down_vf_f64m4(op0, op1, op2) \ +__builtin_rvv_vfslide1down_vf_f64m4((vfloat64m4_t)(op0), (double)(op1), (size_t)(op2)) +#define vfslide1down_vf_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfslide1down_vf_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (double)(op3), (size_t)(op4)) +#define vfslide1down_vf_f64m8(op0, op1, op2) \ +__builtin_rvv_vfslide1down_vf_f64m8((vfloat64m8_t)(op0), (double)(op1), (size_t)(op2)) +#define vfslide1down_vf_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfslide1down_vf_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (double)(op3), (size_t)(op4)) +#define vrgather_vv_f64m1(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_f64m1((vfloat64m1_t)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vrgather_vv_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vrgather_vv_f64m2(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_f64m2((vfloat64m2_t)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vrgather_vv_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vrgather_vv_f64m4(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_f64m4((vfloat64m4_t)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vrgather_vv_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vrgather_vv_f64m8(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_f64m8((vfloat64m8_t)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vrgather_vv_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vrgather_vx_f64m1(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_f64m1((vfloat64m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_f64m2(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_f64m2((vfloat64m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_f64m4(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_f64m4((vfloat64m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_f64m8(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_f64m8((vfloat64m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_f64m1(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_f64m1((vfloat64m1_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_f64m2(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_f64m2((vfloat64m2_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_f64m4(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_f64m4((vfloat64m4_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_f64m8(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_f64m8((vfloat64m8_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vcompress_vm_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_f64m1((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vcompress_vm_f64m2(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_f64m2((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vcompress_vm_f64m4(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_f64m4((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vcompress_vm_f64m8(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_f64m8((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vget_v_f64m2_f64m1(op0, op1) \ +__builtin_rvv_vget_v_f64m2_f64m1((vfloat64m2_t)(op0), (size_t)(op1)) +#define vget_v_f64m4_f64m1(op0, op1) \ +__builtin_rvv_vget_v_f64m4_f64m1((vfloat64m4_t)(op0), (size_t)(op1)) +#define vget_v_f64m8_f64m1(op0, op1) \ +__builtin_rvv_vget_v_f64m8_f64m1((vfloat64m8_t)(op0), (size_t)(op1)) +#define vget_v_f64m4_f64m2(op0, op1) \ +__builtin_rvv_vget_v_f64m4_f64m2((vfloat64m4_t)(op0), (size_t)(op1)) +#define vget_v_f64m8_f64m2(op0, op1) \ +__builtin_rvv_vget_v_f64m8_f64m2((vfloat64m8_t)(op0), (size_t)(op1)) +#define vget_v_f64m8_f64m4(op0, op1) \ +__builtin_rvv_vget_v_f64m8_f64m4((vfloat64m8_t)(op0), (size_t)(op1)) +#define vset_v_f64m1_f64m2(op0, op1, op2) \ +__builtin_rvv_vset_v_f64m1_f64m2((vfloat64m2_t)(op0), (size_t)(op1), (vfloat64m1_t)(op2)) +#define vset_v_f64m1_f64m4(op0, op1, op2) \ +__builtin_rvv_vset_v_f64m1_f64m4((vfloat64m4_t)(op0), (size_t)(op1), (vfloat64m1_t)(op2)) +#define vset_v_f64m2_f64m4(op0, op1, op2) \ +__builtin_rvv_vset_v_f64m2_f64m4((vfloat64m4_t)(op0), (size_t)(op1), (vfloat64m2_t)(op2)) +#define vset_v_f64m1_f64m8(op0, op1, op2) \ +__builtin_rvv_vset_v_f64m1_f64m8((vfloat64m8_t)(op0), (size_t)(op1), (vfloat64m1_t)(op2)) +#define vset_v_f64m2_f64m8(op0, op1, op2) \ +__builtin_rvv_vset_v_f64m2_f64m8((vfloat64m8_t)(op0), (size_t)(op1), (vfloat64m2_t)(op2)) +#define vset_v_f64m4_f64m8(op0, op1, op2) \ +__builtin_rvv_vset_v_f64m4_f64m8((vfloat64m8_t)(op0), (size_t)(op1), (vfloat64m4_t)(op2)) +#define vsuxei8_v_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_f64m1((double *)(op0), (vuint8mf8_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vsuxei8_v_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint8mf8_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vsuxei8_v_f64m2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_f64m2((double *)(op0), (vuint8mf4_t)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vsuxei8_v_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_f64m2_m((vbool32_t)(op0), (double *)(op1), (vuint8mf4_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vsuxei8_v_f64m4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_f64m4((double *)(op0), (vuint8mf2_t)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vsuxei8_v_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_f64m4_m((vbool16_t)(op0), (double *)(op1), (vuint8mf2_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vsuxei8_v_f64m8(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_f64m8((double *)(op0), (vuint8m1_t)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vsuxei8_v_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_f64m8_m((vbool8_t)(op0), (double *)(op1), (vuint8m1_t)(op2), (vfloat64m8_t)(op3), (size_t)(op4)) +#define vsuxei16_v_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_f64m1((double *)(op0), (vuint16mf4_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vsuxei16_v_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint16mf4_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vsuxei16_v_f64m2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_f64m2((double *)(op0), (vuint16mf2_t)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vsuxei16_v_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_f64m2_m((vbool32_t)(op0), (double *)(op1), (vuint16mf2_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vsuxei16_v_f64m4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_f64m4((double *)(op0), (vuint16m1_t)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vsuxei16_v_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_f64m4_m((vbool16_t)(op0), (double *)(op1), (vuint16m1_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vsuxei16_v_f64m8(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_f64m8((double *)(op0), (vuint16m2_t)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vsuxei16_v_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_f64m8_m((vbool8_t)(op0), (double *)(op1), (vuint16m2_t)(op2), (vfloat64m8_t)(op3), (size_t)(op4)) +#define vsuxei32_v_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_f64m1((double *)(op0), (vuint32mf2_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vsuxei32_v_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint32mf2_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vsuxei32_v_f64m2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_f64m2((double *)(op0), (vuint32m1_t)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vsuxei32_v_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_f64m2_m((vbool32_t)(op0), (double *)(op1), (vuint32m1_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vsuxei32_v_f64m4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_f64m4((double *)(op0), (vuint32m2_t)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vsuxei32_v_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_f64m4_m((vbool16_t)(op0), (double *)(op1), (vuint32m2_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vsuxei32_v_f64m8(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_f64m8((double *)(op0), (vuint32m4_t)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vsuxei32_v_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_f64m8_m((vbool8_t)(op0), (double *)(op1), (vuint32m4_t)(op2), (vfloat64m8_t)(op3), (size_t)(op4)) +#define vsuxei64_v_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei64_v_f64m1((double *)(op0), (vuint64m1_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vsuxei64_v_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei64_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint64m1_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vsuxei64_v_f64m2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei64_v_f64m2((double *)(op0), (vuint64m2_t)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vsuxei64_v_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei64_v_f64m2_m((vbool32_t)(op0), (double *)(op1), (vuint64m2_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vsuxei64_v_f64m4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei64_v_f64m4((double *)(op0), (vuint64m4_t)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vsuxei64_v_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei64_v_f64m4_m((vbool16_t)(op0), (double *)(op1), (vuint64m4_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vsuxei64_v_f64m8(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei64_v_f64m8((double *)(op0), (vuint64m8_t)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vsuxei64_v_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei64_v_f64m8_m((vbool8_t)(op0), (double *)(op1), (vuint64m8_t)(op2), (vfloat64m8_t)(op3), (size_t)(op4)) +#define vsoxei8_v_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_f64m1((double *)(op0), (vuint8mf8_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vsoxei8_v_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint8mf8_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vsoxei8_v_f64m2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_f64m2((double *)(op0), (vuint8mf4_t)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vsoxei8_v_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_f64m2_m((vbool32_t)(op0), (double *)(op1), (vuint8mf4_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vsoxei8_v_f64m4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_f64m4((double *)(op0), (vuint8mf2_t)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vsoxei8_v_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_f64m4_m((vbool16_t)(op0), (double *)(op1), (vuint8mf2_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vsoxei8_v_f64m8(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_f64m8((double *)(op0), (vuint8m1_t)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vsoxei8_v_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_f64m8_m((vbool8_t)(op0), (double *)(op1), (vuint8m1_t)(op2), (vfloat64m8_t)(op3), (size_t)(op4)) +#define vsoxei16_v_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_f64m1((double *)(op0), (vuint16mf4_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vsoxei16_v_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint16mf4_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vsoxei16_v_f64m2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_f64m2((double *)(op0), (vuint16mf2_t)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vsoxei16_v_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_f64m2_m((vbool32_t)(op0), (double *)(op1), (vuint16mf2_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vsoxei16_v_f64m4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_f64m4((double *)(op0), (vuint16m1_t)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vsoxei16_v_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_f64m4_m((vbool16_t)(op0), (double *)(op1), (vuint16m1_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vsoxei16_v_f64m8(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_f64m8((double *)(op0), (vuint16m2_t)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vsoxei16_v_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_f64m8_m((vbool8_t)(op0), (double *)(op1), (vuint16m2_t)(op2), (vfloat64m8_t)(op3), (size_t)(op4)) +#define vsoxei32_v_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_f64m1((double *)(op0), (vuint32mf2_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vsoxei32_v_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint32mf2_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vsoxei32_v_f64m2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_f64m2((double *)(op0), (vuint32m1_t)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vsoxei32_v_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_f64m2_m((vbool32_t)(op0), (double *)(op1), (vuint32m1_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vsoxei32_v_f64m4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_f64m4((double *)(op0), (vuint32m2_t)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vsoxei32_v_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_f64m4_m((vbool16_t)(op0), (double *)(op1), (vuint32m2_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vsoxei32_v_f64m8(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_f64m8((double *)(op0), (vuint32m4_t)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vsoxei32_v_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_f64m8_m((vbool8_t)(op0), (double *)(op1), (vuint32m4_t)(op2), (vfloat64m8_t)(op3), (size_t)(op4)) +#define vsoxei64_v_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei64_v_f64m1((double *)(op0), (vuint64m1_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vsoxei64_v_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei64_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint64m1_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vsoxei64_v_f64m2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei64_v_f64m2((double *)(op0), (vuint64m2_t)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vsoxei64_v_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei64_v_f64m2_m((vbool32_t)(op0), (double *)(op1), (vuint64m2_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vsoxei64_v_f64m4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei64_v_f64m4((double *)(op0), (vuint64m4_t)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vsoxei64_v_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei64_v_f64m4_m((vbool16_t)(op0), (double *)(op1), (vuint64m4_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vsoxei64_v_f64m8(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei64_v_f64m8((double *)(op0), (vuint64m8_t)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vsoxei64_v_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei64_v_f64m8_m((vbool8_t)(op0), (double *)(op1), (vuint64m8_t)(op2), (vfloat64m8_t)(op3), (size_t)(op4)) +#define vle64ff_v_f64m1(op0, op1, op2) \ +__builtin_rvv_vle64ff_v_f64m1((const double *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle64ff_v_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle64ff_v_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (const double *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle64ff_v_f64m2(op0, op1, op2) \ +__builtin_rvv_vle64ff_v_f64m2((const double *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle64ff_v_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle64ff_v_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (const double *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle64ff_v_f64m4(op0, op1, op2) \ +__builtin_rvv_vle64ff_v_f64m4((const double *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle64ff_v_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle64ff_v_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (const double *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle64ff_v_f64m8(op0, op1, op2) \ +__builtin_rvv_vle64ff_v_f64m8((const double *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle64ff_v_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle64ff_v_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (const double *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vfneg_v_f64m1(op0, op1) \ +__builtin_rvv_vfneg_v_f64m1((vfloat64m1_t)(op0), (size_t)(op1)) +#define vfneg_v_f64m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfneg_v_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfneg_v_f64m2(op0, op1) \ +__builtin_rvv_vfneg_v_f64m2((vfloat64m2_t)(op0), (size_t)(op1)) +#define vfneg_v_f64m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfneg_v_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vfneg_v_f64m4(op0, op1) \ +__builtin_rvv_vfneg_v_f64m4((vfloat64m4_t)(op0), (size_t)(op1)) +#define vfneg_v_f64m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfneg_v_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vfneg_v_f64m8(op0, op1) \ +__builtin_rvv_vfneg_v_f64m8((vfloat64m8_t)(op0), (size_t)(op1)) +#define vfneg_v_f64m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfneg_v_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vle64_v_f64m1(op0, op1) \ +__builtin_rvv_vle64_v_f64m1((const double *)(op0), (size_t)(op1)) +#define vle64_v_f64m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vle64_v_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (const double *)(op2), (size_t)(op3)) +#define vle64_v_f64m2(op0, op1) \ +__builtin_rvv_vle64_v_f64m2((const double *)(op0), (size_t)(op1)) +#define vle64_v_f64m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vle64_v_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (const double *)(op2), (size_t)(op3)) +#define vle64_v_f64m4(op0, op1) \ +__builtin_rvv_vle64_v_f64m4((const double *)(op0), (size_t)(op1)) +#define vle64_v_f64m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vle64_v_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (const double *)(op2), (size_t)(op3)) +#define vle64_v_f64m8(op0, op1) \ +__builtin_rvv_vle64_v_f64m8((const double *)(op0), (size_t)(op1)) +#define vle64_v_f64m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vle64_v_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (const double *)(op2), (size_t)(op3)) +#define vse64_v_f64m1(op0, op1, op2) \ +__builtin_rvv_vse64_v_f64m1((double *)(op0), (vfloat64m1_t)(op1), (size_t)(op2)) +#define vse64_v_f64m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vse64_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vse64_v_f64m2(op0, op1, op2) \ +__builtin_rvv_vse64_v_f64m2((double *)(op0), (vfloat64m2_t)(op1), (size_t)(op2)) +#define vse64_v_f64m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vse64_v_f64m2_m((vbool32_t)(op0), (double *)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vse64_v_f64m4(op0, op1, op2) \ +__builtin_rvv_vse64_v_f64m4((double *)(op0), (vfloat64m4_t)(op1), (size_t)(op2)) +#define vse64_v_f64m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vse64_v_f64m4_m((vbool16_t)(op0), (double *)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vse64_v_f64m8(op0, op1, op2) \ +__builtin_rvv_vse64_v_f64m8((double *)(op0), (vfloat64m8_t)(op1), (size_t)(op2)) +#define vse64_v_f64m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vse64_v_f64m8_m((vbool8_t)(op0), (double *)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vfclass_v_u64m1(op0, op1) \ +__builtin_rvv_vfclass_v_u64m1((vfloat64m1_t)(op0), (size_t)(op1)) +#define vfclass_v_u64m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfclass_v_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfclass_v_u64m2(op0, op1) \ +__builtin_rvv_vfclass_v_u64m2((vfloat64m2_t)(op0), (size_t)(op1)) +#define vfclass_v_u64m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfclass_v_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vfclass_v_u64m4(op0, op1) \ +__builtin_rvv_vfclass_v_u64m4((vfloat64m4_t)(op0), (size_t)(op1)) +#define vfclass_v_u64m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfclass_v_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vfclass_v_u64m8(op0, op1) \ +__builtin_rvv_vfclass_v_u64m8((vfloat64m8_t)(op0), (size_t)(op1)) +#define vfclass_v_u64m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfclass_v_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vfcvt_f_x_v_f64m1(op0, op1) \ +__builtin_rvv_vfcvt_f_x_v_f64m1((vint64m1_t)(op0), (size_t)(op1)) +#define vfcvt_f_x_v_f64m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_f_x_v_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vfcvt_f_x_v_f64m2(op0, op1) \ +__builtin_rvv_vfcvt_f_x_v_f64m2((vint64m2_t)(op0), (size_t)(op1)) +#define vfcvt_f_x_v_f64m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_f_x_v_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vfcvt_f_x_v_f64m4(op0, op1) \ +__builtin_rvv_vfcvt_f_x_v_f64m4((vint64m4_t)(op0), (size_t)(op1)) +#define vfcvt_f_x_v_f64m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_f_x_v_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vfcvt_f_x_v_f64m8(op0, op1) \ +__builtin_rvv_vfcvt_f_x_v_f64m8((vint64m8_t)(op0), (size_t)(op1)) +#define vfcvt_f_x_v_f64m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_f_x_v_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vfcvt_f_xu_v_f64m1(op0, op1) \ +__builtin_rvv_vfcvt_f_xu_v_f64m1((vuint64m1_t)(op0), (size_t)(op1)) +#define vfcvt_f_xu_v_f64m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_f_xu_v_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vfcvt_f_xu_v_f64m2(op0, op1) \ +__builtin_rvv_vfcvt_f_xu_v_f64m2((vuint64m2_t)(op0), (size_t)(op1)) +#define vfcvt_f_xu_v_f64m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_f_xu_v_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vfcvt_f_xu_v_f64m4(op0, op1) \ +__builtin_rvv_vfcvt_f_xu_v_f64m4((vuint64m4_t)(op0), (size_t)(op1)) +#define vfcvt_f_xu_v_f64m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_f_xu_v_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vfcvt_f_xu_v_f64m8(op0, op1) \ +__builtin_rvv_vfcvt_f_xu_v_f64m8((vuint64m8_t)(op0), (size_t)(op1)) +#define vfcvt_f_xu_v_f64m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_f_xu_v_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vfcvt_rtz_x_f_v_i64m1(op0, op1) \ +__builtin_rvv_vfcvt_rtz_x_f_v_i64m1((vfloat64m1_t)(op0), (size_t)(op1)) +#define vfcvt_rtz_x_f_v_i64m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_rtz_x_f_v_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfcvt_rtz_x_f_v_i64m2(op0, op1) \ +__builtin_rvv_vfcvt_rtz_x_f_v_i64m2((vfloat64m2_t)(op0), (size_t)(op1)) +#define vfcvt_rtz_x_f_v_i64m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_rtz_x_f_v_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vfcvt_rtz_x_f_v_i64m4(op0, op1) \ +__builtin_rvv_vfcvt_rtz_x_f_v_i64m4((vfloat64m4_t)(op0), (size_t)(op1)) +#define vfcvt_rtz_x_f_v_i64m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_rtz_x_f_v_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vfcvt_rtz_x_f_v_i64m8(op0, op1) \ +__builtin_rvv_vfcvt_rtz_x_f_v_i64m8((vfloat64m8_t)(op0), (size_t)(op1)) +#define vfcvt_rtz_x_f_v_i64m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_rtz_x_f_v_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vfcvt_rtz_xu_f_v_u64m1(op0, op1) \ +__builtin_rvv_vfcvt_rtz_xu_f_v_u64m1((vfloat64m1_t)(op0), (size_t)(op1)) +#define vfcvt_rtz_xu_f_v_u64m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_rtz_xu_f_v_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfcvt_rtz_xu_f_v_u64m2(op0, op1) \ +__builtin_rvv_vfcvt_rtz_xu_f_v_u64m2((vfloat64m2_t)(op0), (size_t)(op1)) +#define vfcvt_rtz_xu_f_v_u64m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_rtz_xu_f_v_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vfcvt_rtz_xu_f_v_u64m4(op0, op1) \ +__builtin_rvv_vfcvt_rtz_xu_f_v_u64m4((vfloat64m4_t)(op0), (size_t)(op1)) +#define vfcvt_rtz_xu_f_v_u64m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_rtz_xu_f_v_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vfcvt_rtz_xu_f_v_u64m8(op0, op1) \ +__builtin_rvv_vfcvt_rtz_xu_f_v_u64m8((vfloat64m8_t)(op0), (size_t)(op1)) +#define vfcvt_rtz_xu_f_v_u64m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_rtz_xu_f_v_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vfcvt_x_f_v_i64m1(op0, op1) \ +__builtin_rvv_vfcvt_x_f_v_i64m1((vfloat64m1_t)(op0), (size_t)(op1)) +#define vfcvt_x_f_v_i64m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_x_f_v_i64m1_m((vbool64_t)(op0), (vint64m1_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfcvt_x_f_v_i64m2(op0, op1) \ +__builtin_rvv_vfcvt_x_f_v_i64m2((vfloat64m2_t)(op0), (size_t)(op1)) +#define vfcvt_x_f_v_i64m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_x_f_v_i64m2_m((vbool32_t)(op0), (vint64m2_t)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vfcvt_x_f_v_i64m4(op0, op1) \ +__builtin_rvv_vfcvt_x_f_v_i64m4((vfloat64m4_t)(op0), (size_t)(op1)) +#define vfcvt_x_f_v_i64m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_x_f_v_i64m4_m((vbool16_t)(op0), (vint64m4_t)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vfcvt_x_f_v_i64m8(op0, op1) \ +__builtin_rvv_vfcvt_x_f_v_i64m8((vfloat64m8_t)(op0), (size_t)(op1)) +#define vfcvt_x_f_v_i64m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_x_f_v_i64m8_m((vbool8_t)(op0), (vint64m8_t)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vfcvt_xu_f_v_u64m1(op0, op1) \ +__builtin_rvv_vfcvt_xu_f_v_u64m1((vfloat64m1_t)(op0), (size_t)(op1)) +#define vfcvt_xu_f_v_u64m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_xu_f_v_u64m1_m((vbool64_t)(op0), (vuint64m1_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfcvt_xu_f_v_u64m2(op0, op1) \ +__builtin_rvv_vfcvt_xu_f_v_u64m2((vfloat64m2_t)(op0), (size_t)(op1)) +#define vfcvt_xu_f_v_u64m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_xu_f_v_u64m2_m((vbool32_t)(op0), (vuint64m2_t)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vfcvt_xu_f_v_u64m4(op0, op1) \ +__builtin_rvv_vfcvt_xu_f_v_u64m4((vfloat64m4_t)(op0), (size_t)(op1)) +#define vfcvt_xu_f_v_u64m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_xu_f_v_u64m4_m((vbool16_t)(op0), (vuint64m4_t)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vfcvt_xu_f_v_u64m8(op0, op1) \ +__builtin_rvv_vfcvt_xu_f_v_u64m8((vfloat64m8_t)(op0), (size_t)(op1)) +#define vfcvt_xu_f_v_u64m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_xu_f_v_u64m8_m((vbool8_t)(op0), (vuint64m8_t)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vfncvt_rtz_x_f_w_i32mf2(op0, op1) \ +__builtin_rvv_vfncvt_rtz_x_f_w_i32mf2((vfloat64m1_t)(op0), (size_t)(op1)) +#define vfncvt_rtz_x_f_w_i32mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_rtz_x_f_w_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfncvt_rtz_x_f_w_i32m1(op0, op1) \ +__builtin_rvv_vfncvt_rtz_x_f_w_i32m1((vfloat64m2_t)(op0), (size_t)(op1)) +#define vfncvt_rtz_x_f_w_i32m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_rtz_x_f_w_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vfncvt_rtz_x_f_w_i32m2(op0, op1) \ +__builtin_rvv_vfncvt_rtz_x_f_w_i32m2((vfloat64m4_t)(op0), (size_t)(op1)) +#define vfncvt_rtz_x_f_w_i32m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_rtz_x_f_w_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vfncvt_rtz_x_f_w_i32m4(op0, op1) \ +__builtin_rvv_vfncvt_rtz_x_f_w_i32m4((vfloat64m8_t)(op0), (size_t)(op1)) +#define vfncvt_rtz_x_f_w_i32m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_rtz_x_f_w_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vfncvt_rtz_xu_f_w_u32mf2(op0, op1) \ +__builtin_rvv_vfncvt_rtz_xu_f_w_u32mf2((vfloat64m1_t)(op0), (size_t)(op1)) +#define vfncvt_rtz_xu_f_w_u32mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_rtz_xu_f_w_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfncvt_rtz_xu_f_w_u32m1(op0, op1) \ +__builtin_rvv_vfncvt_rtz_xu_f_w_u32m1((vfloat64m2_t)(op0), (size_t)(op1)) +#define vfncvt_rtz_xu_f_w_u32m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_rtz_xu_f_w_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vfncvt_rtz_xu_f_w_u32m2(op0, op1) \ +__builtin_rvv_vfncvt_rtz_xu_f_w_u32m2((vfloat64m4_t)(op0), (size_t)(op1)) +#define vfncvt_rtz_xu_f_w_u32m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_rtz_xu_f_w_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vfncvt_rtz_xu_f_w_u32m4(op0, op1) \ +__builtin_rvv_vfncvt_rtz_xu_f_w_u32m4((vfloat64m8_t)(op0), (size_t)(op1)) +#define vfncvt_rtz_xu_f_w_u32m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_rtz_xu_f_w_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vfncvt_x_f_w_i32mf2(op0, op1) \ +__builtin_rvv_vfncvt_x_f_w_i32mf2((vfloat64m1_t)(op0), (size_t)(op1)) +#define vfncvt_x_f_w_i32mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_x_f_w_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfncvt_x_f_w_i32m1(op0, op1) \ +__builtin_rvv_vfncvt_x_f_w_i32m1((vfloat64m2_t)(op0), (size_t)(op1)) +#define vfncvt_x_f_w_i32m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_x_f_w_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vfncvt_x_f_w_i32m2(op0, op1) \ +__builtin_rvv_vfncvt_x_f_w_i32m2((vfloat64m4_t)(op0), (size_t)(op1)) +#define vfncvt_x_f_w_i32m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_x_f_w_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vfncvt_x_f_w_i32m4(op0, op1) \ +__builtin_rvv_vfncvt_x_f_w_i32m4((vfloat64m8_t)(op0), (size_t)(op1)) +#define vfncvt_x_f_w_i32m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_x_f_w_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vfncvt_xu_f_w_u32mf2(op0, op1) \ +__builtin_rvv_vfncvt_xu_f_w_u32mf2((vfloat64m1_t)(op0), (size_t)(op1)) +#define vfncvt_xu_f_w_u32mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_xu_f_w_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfncvt_xu_f_w_u32m1(op0, op1) \ +__builtin_rvv_vfncvt_xu_f_w_u32m1((vfloat64m2_t)(op0), (size_t)(op1)) +#define vfncvt_xu_f_w_u32m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_xu_f_w_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vfncvt_xu_f_w_u32m2(op0, op1) \ +__builtin_rvv_vfncvt_xu_f_w_u32m2((vfloat64m4_t)(op0), (size_t)(op1)) +#define vfncvt_xu_f_w_u32m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_xu_f_w_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vfncvt_xu_f_w_u32m4(op0, op1) \ +__builtin_rvv_vfncvt_xu_f_w_u32m4((vfloat64m8_t)(op0), (size_t)(op1)) +#define vfncvt_xu_f_w_u32m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_xu_f_w_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vfrec7_v_f64m1(op0, op1) \ +__builtin_rvv_vfrec7_v_f64m1((vfloat64m1_t)(op0), (size_t)(op1)) +#define vfrec7_v_f64m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfrec7_v_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfrec7_v_f64m2(op0, op1) \ +__builtin_rvv_vfrec7_v_f64m2((vfloat64m2_t)(op0), (size_t)(op1)) +#define vfrec7_v_f64m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfrec7_v_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vfrec7_v_f64m4(op0, op1) \ +__builtin_rvv_vfrec7_v_f64m4((vfloat64m4_t)(op0), (size_t)(op1)) +#define vfrec7_v_f64m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfrec7_v_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vfrec7_v_f64m8(op0, op1) \ +__builtin_rvv_vfrec7_v_f64m8((vfloat64m8_t)(op0), (size_t)(op1)) +#define vfrec7_v_f64m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfrec7_v_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vfrsqrt7_v_f64m1(op0, op1) \ +__builtin_rvv_vfrsqrt7_v_f64m1((vfloat64m1_t)(op0), (size_t)(op1)) +#define vfrsqrt7_v_f64m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfrsqrt7_v_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfrsqrt7_v_f64m2(op0, op1) \ +__builtin_rvv_vfrsqrt7_v_f64m2((vfloat64m2_t)(op0), (size_t)(op1)) +#define vfrsqrt7_v_f64m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfrsqrt7_v_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vfrsqrt7_v_f64m4(op0, op1) \ +__builtin_rvv_vfrsqrt7_v_f64m4((vfloat64m4_t)(op0), (size_t)(op1)) +#define vfrsqrt7_v_f64m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfrsqrt7_v_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vfrsqrt7_v_f64m8(op0, op1) \ +__builtin_rvv_vfrsqrt7_v_f64m8((vfloat64m8_t)(op0), (size_t)(op1)) +#define vfrsqrt7_v_f64m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfrsqrt7_v_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vfsqrt_v_f64m1(op0, op1) \ +__builtin_rvv_vfsqrt_v_f64m1((vfloat64m1_t)(op0), (size_t)(op1)) +#define vfsqrt_v_f64m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfsqrt_v_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfsqrt_v_f64m2(op0, op1) \ +__builtin_rvv_vfsqrt_v_f64m2((vfloat64m2_t)(op0), (size_t)(op1)) +#define vfsqrt_v_f64m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfsqrt_v_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vfsqrt_v_f64m4(op0, op1) \ +__builtin_rvv_vfsqrt_v_f64m4((vfloat64m4_t)(op0), (size_t)(op1)) +#define vfsqrt_v_f64m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfsqrt_v_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vfsqrt_v_f64m8(op0, op1) \ +__builtin_rvv_vfsqrt_v_f64m8((vfloat64m8_t)(op0), (size_t)(op1)) +#define vfsqrt_v_f64m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfsqrt_v_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vfwcvt_f_x_v_f64m1(op0, op1) \ +__builtin_rvv_vfwcvt_f_x_v_f64m1((vint32mf2_t)(op0), (size_t)(op1)) +#define vfwcvt_f_x_v_f64m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_f_x_v_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vfwcvt_f_x_v_f64m2(op0, op1) \ +__builtin_rvv_vfwcvt_f_x_v_f64m2((vint32m1_t)(op0), (size_t)(op1)) +#define vfwcvt_f_x_v_f64m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_f_x_v_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vfwcvt_f_x_v_f64m4(op0, op1) \ +__builtin_rvv_vfwcvt_f_x_v_f64m4((vint32m2_t)(op0), (size_t)(op1)) +#define vfwcvt_f_x_v_f64m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_f_x_v_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vfwcvt_f_x_v_f64m8(op0, op1) \ +__builtin_rvv_vfwcvt_f_x_v_f64m8((vint32m4_t)(op0), (size_t)(op1)) +#define vfwcvt_f_x_v_f64m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_f_x_v_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vfwcvt_f_xu_v_f64m1(op0, op1) \ +__builtin_rvv_vfwcvt_f_xu_v_f64m1((vuint32mf2_t)(op0), (size_t)(op1)) +#define vfwcvt_f_xu_v_f64m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_f_xu_v_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vfwcvt_f_xu_v_f64m2(op0, op1) \ +__builtin_rvv_vfwcvt_f_xu_v_f64m2((vuint32m1_t)(op0), (size_t)(op1)) +#define vfwcvt_f_xu_v_f64m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_f_xu_v_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vfwcvt_f_xu_v_f64m4(op0, op1) \ +__builtin_rvv_vfwcvt_f_xu_v_f64m4((vuint32m2_t)(op0), (size_t)(op1)) +#define vfwcvt_f_xu_v_f64m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_f_xu_v_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vfwcvt_f_xu_v_f64m8(op0, op1) \ +__builtin_rvv_vfwcvt_f_xu_v_f64m8((vuint32m4_t)(op0), (size_t)(op1)) +#define vfwcvt_f_xu_v_f64m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_f_xu_v_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vlmul_ext_v_f64m1_f64m2(op0) \ +__builtin_rvv_vlmul_ext_v_f64m1_f64m2((vfloat64m1_t)(op0)) +#define vlmul_ext_v_f64m1_f64m4(op0) \ +__builtin_rvv_vlmul_ext_v_f64m1_f64m4((vfloat64m1_t)(op0)) +#define vlmul_ext_v_f64m2_f64m4(op0) \ +__builtin_rvv_vlmul_ext_v_f64m2_f64m4((vfloat64m2_t)(op0)) +#define vlmul_ext_v_f64m1_f64m8(op0) \ +__builtin_rvv_vlmul_ext_v_f64m1_f64m8((vfloat64m1_t)(op0)) +#define vlmul_ext_v_f64m2_f64m8(op0) \ +__builtin_rvv_vlmul_ext_v_f64m2_f64m8((vfloat64m2_t)(op0)) +#define vlmul_ext_v_f64m4_f64m8(op0) \ +__builtin_rvv_vlmul_ext_v_f64m4_f64m8((vfloat64m4_t)(op0)) +#define vlmul_trunc_v_f64m2_f64m1(op0) \ +__builtin_rvv_vlmul_trunc_v_f64m2_f64m1((vfloat64m2_t)(op0)) +#define vlmul_trunc_v_f64m4_f64m1(op0) \ +__builtin_rvv_vlmul_trunc_v_f64m4_f64m1((vfloat64m4_t)(op0)) +#define vlmul_trunc_v_f64m8_f64m1(op0) \ +__builtin_rvv_vlmul_trunc_v_f64m8_f64m1((vfloat64m8_t)(op0)) +#define vlmul_trunc_v_f64m4_f64m2(op0) \ +__builtin_rvv_vlmul_trunc_v_f64m4_f64m2((vfloat64m4_t)(op0)) +#define vlmul_trunc_v_f64m8_f64m2(op0) \ +__builtin_rvv_vlmul_trunc_v_f64m8_f64m2((vfloat64m8_t)(op0)) +#define vlmul_trunc_v_f64m8_f64m4(op0) \ +__builtin_rvv_vlmul_trunc_v_f64m8_f64m4((vfloat64m8_t)(op0)) +#define vreinterpret_v_i64m1_f64m1(op0) \ +__builtin_rvv_vreinterpret_v_i64m1_f64m1((vint64m1_t)(op0)) +#define vreinterpret_v_i64m2_f64m2(op0) \ +__builtin_rvv_vreinterpret_v_i64m2_f64m2((vint64m2_t)(op0)) +#define vreinterpret_v_i64m4_f64m4(op0) \ +__builtin_rvv_vreinterpret_v_i64m4_f64m4((vint64m4_t)(op0)) +#define vreinterpret_v_i64m8_f64m8(op0) \ +__builtin_rvv_vreinterpret_v_i64m8_f64m8((vint64m8_t)(op0)) +#define vreinterpret_v_u64m1_f64m1(op0) \ +__builtin_rvv_vreinterpret_v_u64m1_f64m1((vuint64m1_t)(op0)) +#define vreinterpret_v_u64m2_f64m2(op0) \ +__builtin_rvv_vreinterpret_v_u64m2_f64m2((vuint64m2_t)(op0)) +#define vreinterpret_v_u64m4_f64m4(op0) \ +__builtin_rvv_vreinterpret_v_u64m4_f64m4((vuint64m4_t)(op0)) +#define vreinterpret_v_u64m8_f64m8(op0) \ +__builtin_rvv_vreinterpret_v_u64m8_f64m8((vuint64m8_t)(op0)) +#define vreinterpret_v_f64m1_i64m1(op0) \ +__builtin_rvv_vreinterpret_v_f64m1_i64m1((vfloat64m1_t)(op0)) +#define vreinterpret_v_f64m2_i64m2(op0) \ +__builtin_rvv_vreinterpret_v_f64m2_i64m2((vfloat64m2_t)(op0)) +#define vreinterpret_v_f64m4_i64m4(op0) \ +__builtin_rvv_vreinterpret_v_f64m4_i64m4((vfloat64m4_t)(op0)) +#define vreinterpret_v_f64m8_i64m8(op0) \ +__builtin_rvv_vreinterpret_v_f64m8_i64m8((vfloat64m8_t)(op0)) +#define vreinterpret_v_f64m1_u64m1(op0) \ +__builtin_rvv_vreinterpret_v_f64m1_u64m1((vfloat64m1_t)(op0)) +#define vreinterpret_v_f64m2_u64m2(op0) \ +__builtin_rvv_vreinterpret_v_f64m2_u64m2((vfloat64m2_t)(op0)) +#define vreinterpret_v_f64m4_u64m4(op0) \ +__builtin_rvv_vreinterpret_v_f64m4_u64m4((vfloat64m4_t)(op0)) +#define vreinterpret_v_f64m8_u64m8(op0) \ +__builtin_rvv_vreinterpret_v_f64m8_u64m8((vfloat64m8_t)(op0)) +#define vundefined_f64m1() \ +__builtin_rvv_vundefined_f64m1() +#define vundefined_f64m2() \ +__builtin_rvv_vundefined_f64m2() +#define vundefined_f64m4() \ +__builtin_rvv_vundefined_f64m4() +#define vundefined_f64m8() \ +__builtin_rvv_vundefined_f64m8() +#endif + +#if defined(__riscv_f) && defined(__riscv_d) +#define vfwadd_vv_f64m1(op0, op1, op2) \ +__builtin_rvv_vfwadd_vv_f64m1((vfloat32mf2_t)(op0), (vfloat32mf2_t)(op1), (size_t)(op2)) +#define vfwadd_vv_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwadd_vv_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vfwadd_vv_f64m2(op0, op1, op2) \ +__builtin_rvv_vfwadd_vv_f64m2((vfloat32m1_t)(op0), (vfloat32m1_t)(op1), (size_t)(op2)) +#define vfwadd_vv_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwadd_vv_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfwadd_vv_f64m4(op0, op1, op2) \ +__builtin_rvv_vfwadd_vv_f64m4((vfloat32m2_t)(op0), (vfloat32m2_t)(op1), (size_t)(op2)) +#define vfwadd_vv_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwadd_vv_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vfwadd_vv_f64m8(op0, op1, op2) \ +__builtin_rvv_vfwadd_vv_f64m8((vfloat32m4_t)(op0), (vfloat32m4_t)(op1), (size_t)(op2)) +#define vfwadd_vv_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwadd_vv_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat32m4_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vfwadd_vf_f64m1(op0, op1, op2) \ +__builtin_rvv_vfwadd_vf_f64m1((vfloat32mf2_t)(op0), (float)(op1), (size_t)(op2)) +#define vfwadd_vf_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwadd_vf_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat32mf2_t)(op2), (float)(op3), (size_t)(op4)) +#define vfwadd_vf_f64m2(op0, op1, op2) \ +__builtin_rvv_vfwadd_vf_f64m2((vfloat32m1_t)(op0), (float)(op1), (size_t)(op2)) +#define vfwadd_vf_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwadd_vf_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat32m1_t)(op2), (float)(op3), (size_t)(op4)) +#define vfwadd_vf_f64m4(op0, op1, op2) \ +__builtin_rvv_vfwadd_vf_f64m4((vfloat32m2_t)(op0), (float)(op1), (size_t)(op2)) +#define vfwadd_vf_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwadd_vf_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat32m2_t)(op2), (float)(op3), (size_t)(op4)) +#define vfwadd_vf_f64m8(op0, op1, op2) \ +__builtin_rvv_vfwadd_vf_f64m8((vfloat32m4_t)(op0), (float)(op1), (size_t)(op2)) +#define vfwadd_vf_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwadd_vf_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat32m4_t)(op2), (float)(op3), (size_t)(op4)) +#define vfwsub_vv_f64m1(op0, op1, op2) \ +__builtin_rvv_vfwsub_vv_f64m1((vfloat32mf2_t)(op0), (vfloat32mf2_t)(op1), (size_t)(op2)) +#define vfwsub_vv_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwsub_vv_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vfwsub_vv_f64m2(op0, op1, op2) \ +__builtin_rvv_vfwsub_vv_f64m2((vfloat32m1_t)(op0), (vfloat32m1_t)(op1), (size_t)(op2)) +#define vfwsub_vv_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwsub_vv_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfwsub_vv_f64m4(op0, op1, op2) \ +__builtin_rvv_vfwsub_vv_f64m4((vfloat32m2_t)(op0), (vfloat32m2_t)(op1), (size_t)(op2)) +#define vfwsub_vv_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwsub_vv_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vfwsub_vv_f64m8(op0, op1, op2) \ +__builtin_rvv_vfwsub_vv_f64m8((vfloat32m4_t)(op0), (vfloat32m4_t)(op1), (size_t)(op2)) +#define vfwsub_vv_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwsub_vv_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat32m4_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vfwsub_vf_f64m1(op0, op1, op2) \ +__builtin_rvv_vfwsub_vf_f64m1((vfloat32mf2_t)(op0), (float)(op1), (size_t)(op2)) +#define vfwsub_vf_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwsub_vf_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat32mf2_t)(op2), (float)(op3), (size_t)(op4)) +#define vfwsub_vf_f64m2(op0, op1, op2) \ +__builtin_rvv_vfwsub_vf_f64m2((vfloat32m1_t)(op0), (float)(op1), (size_t)(op2)) +#define vfwsub_vf_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwsub_vf_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat32m1_t)(op2), (float)(op3), (size_t)(op4)) +#define vfwsub_vf_f64m4(op0, op1, op2) \ +__builtin_rvv_vfwsub_vf_f64m4((vfloat32m2_t)(op0), (float)(op1), (size_t)(op2)) +#define vfwsub_vf_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwsub_vf_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat32m2_t)(op2), (float)(op3), (size_t)(op4)) +#define vfwsub_vf_f64m8(op0, op1, op2) \ +__builtin_rvv_vfwsub_vf_f64m8((vfloat32m4_t)(op0), (float)(op1), (size_t)(op2)) +#define vfwsub_vf_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwsub_vf_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat32m4_t)(op2), (float)(op3), (size_t)(op4)) +#define vfwadd_wv_f64m1(op0, op1, op2) \ +__builtin_rvv_vfwadd_wv_f64m1((vfloat64m1_t)(op0), (vfloat32mf2_t)(op1), (size_t)(op2)) +#define vfwadd_wv_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwadd_wv_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vfwadd_wv_f64m2(op0, op1, op2) \ +__builtin_rvv_vfwadd_wv_f64m2((vfloat64m2_t)(op0), (vfloat32m1_t)(op1), (size_t)(op2)) +#define vfwadd_wv_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwadd_wv_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfwadd_wv_f64m4(op0, op1, op2) \ +__builtin_rvv_vfwadd_wv_f64m4((vfloat64m4_t)(op0), (vfloat32m2_t)(op1), (size_t)(op2)) +#define vfwadd_wv_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwadd_wv_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vfwadd_wv_f64m8(op0, op1, op2) \ +__builtin_rvv_vfwadd_wv_f64m8((vfloat64m8_t)(op0), (vfloat32m4_t)(op1), (size_t)(op2)) +#define vfwadd_wv_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwadd_wv_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vfwadd_wf_f64m1(op0, op1, op2) \ +__builtin_rvv_vfwadd_wf_f64m1((vfloat64m1_t)(op0), (float)(op1), (size_t)(op2)) +#define vfwadd_wf_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwadd_wf_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (float)(op3), (size_t)(op4)) +#define vfwadd_wf_f64m2(op0, op1, op2) \ +__builtin_rvv_vfwadd_wf_f64m2((vfloat64m2_t)(op0), (float)(op1), (size_t)(op2)) +#define vfwadd_wf_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwadd_wf_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (float)(op3), (size_t)(op4)) +#define vfwadd_wf_f64m4(op0, op1, op2) \ +__builtin_rvv_vfwadd_wf_f64m4((vfloat64m4_t)(op0), (float)(op1), (size_t)(op2)) +#define vfwadd_wf_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwadd_wf_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (float)(op3), (size_t)(op4)) +#define vfwadd_wf_f64m8(op0, op1, op2) \ +__builtin_rvv_vfwadd_wf_f64m8((vfloat64m8_t)(op0), (float)(op1), (size_t)(op2)) +#define vfwadd_wf_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwadd_wf_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (float)(op3), (size_t)(op4)) +#define vfwsub_wv_f64m1(op0, op1, op2) \ +__builtin_rvv_vfwsub_wv_f64m1((vfloat64m1_t)(op0), (vfloat32mf2_t)(op1), (size_t)(op2)) +#define vfwsub_wv_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwsub_wv_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vfwsub_wv_f64m2(op0, op1, op2) \ +__builtin_rvv_vfwsub_wv_f64m2((vfloat64m2_t)(op0), (vfloat32m1_t)(op1), (size_t)(op2)) +#define vfwsub_wv_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwsub_wv_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfwsub_wv_f64m4(op0, op1, op2) \ +__builtin_rvv_vfwsub_wv_f64m4((vfloat64m4_t)(op0), (vfloat32m2_t)(op1), (size_t)(op2)) +#define vfwsub_wv_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwsub_wv_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vfwsub_wv_f64m8(op0, op1, op2) \ +__builtin_rvv_vfwsub_wv_f64m8((vfloat64m8_t)(op0), (vfloat32m4_t)(op1), (size_t)(op2)) +#define vfwsub_wv_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwsub_wv_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vfwsub_wf_f64m1(op0, op1, op2) \ +__builtin_rvv_vfwsub_wf_f64m1((vfloat64m1_t)(op0), (float)(op1), (size_t)(op2)) +#define vfwsub_wf_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwsub_wf_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (float)(op3), (size_t)(op4)) +#define vfwsub_wf_f64m2(op0, op1, op2) \ +__builtin_rvv_vfwsub_wf_f64m2((vfloat64m2_t)(op0), (float)(op1), (size_t)(op2)) +#define vfwsub_wf_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwsub_wf_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (float)(op3), (size_t)(op4)) +#define vfwsub_wf_f64m4(op0, op1, op2) \ +__builtin_rvv_vfwsub_wf_f64m4((vfloat64m4_t)(op0), (float)(op1), (size_t)(op2)) +#define vfwsub_wf_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwsub_wf_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (float)(op3), (size_t)(op4)) +#define vfwsub_wf_f64m8(op0, op1, op2) \ +__builtin_rvv_vfwsub_wf_f64m8((vfloat64m8_t)(op0), (float)(op1), (size_t)(op2)) +#define vfwsub_wf_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwsub_wf_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat64m8_t)(op2), (float)(op3), (size_t)(op4)) +#define vfwmul_vv_f64m1(op0, op1, op2) \ +__builtin_rvv_vfwmul_vv_f64m1((vfloat32mf2_t)(op0), (vfloat32mf2_t)(op1), (size_t)(op2)) +#define vfwmul_vv_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmul_vv_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vfwmul_vv_f64m2(op0, op1, op2) \ +__builtin_rvv_vfwmul_vv_f64m2((vfloat32m1_t)(op0), (vfloat32m1_t)(op1), (size_t)(op2)) +#define vfwmul_vv_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmul_vv_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfwmul_vv_f64m4(op0, op1, op2) \ +__builtin_rvv_vfwmul_vv_f64m4((vfloat32m2_t)(op0), (vfloat32m2_t)(op1), (size_t)(op2)) +#define vfwmul_vv_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmul_vv_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vfwmul_vv_f64m8(op0, op1, op2) \ +__builtin_rvv_vfwmul_vv_f64m8((vfloat32m4_t)(op0), (vfloat32m4_t)(op1), (size_t)(op2)) +#define vfwmul_vv_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmul_vv_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat32m4_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vfwmul_vf_f64m1(op0, op1, op2) \ +__builtin_rvv_vfwmul_vf_f64m1((vfloat32mf2_t)(op0), (float)(op1), (size_t)(op2)) +#define vfwmul_vf_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmul_vf_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat32mf2_t)(op2), (float)(op3), (size_t)(op4)) +#define vfwmul_vf_f64m2(op0, op1, op2) \ +__builtin_rvv_vfwmul_vf_f64m2((vfloat32m1_t)(op0), (float)(op1), (size_t)(op2)) +#define vfwmul_vf_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmul_vf_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat32m1_t)(op2), (float)(op3), (size_t)(op4)) +#define vfwmul_vf_f64m4(op0, op1, op2) \ +__builtin_rvv_vfwmul_vf_f64m4((vfloat32m2_t)(op0), (float)(op1), (size_t)(op2)) +#define vfwmul_vf_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmul_vf_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat32m2_t)(op2), (float)(op3), (size_t)(op4)) +#define vfwmul_vf_f64m8(op0, op1, op2) \ +__builtin_rvv_vfwmul_vf_f64m8((vfloat32m4_t)(op0), (float)(op1), (size_t)(op2)) +#define vfwmul_vf_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmul_vf_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat32m4_t)(op2), (float)(op3), (size_t)(op4)) +#define vfwmacc_vv_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfwmacc_vv_f64m1((vfloat64m1_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vfwmacc_vv_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmacc_vv_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vfwmacc_vv_f64m2(op0, op1, op2, op3) \ +__builtin_rvv_vfwmacc_vv_f64m2((vfloat64m2_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfwmacc_vv_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmacc_vv_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfwmacc_vv_f64m4(op0, op1, op2, op3) \ +__builtin_rvv_vfwmacc_vv_f64m4((vfloat64m4_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vfwmacc_vv_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmacc_vv_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vfwmacc_vv_f64m8(op0, op1, op2, op3) \ +__builtin_rvv_vfwmacc_vv_f64m8((vfloat64m8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vfwmacc_vv_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmacc_vv_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat32m4_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vfwmacc_vf_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfwmacc_vf_f64m1((vfloat64m1_t)(op0), (float)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vfwmacc_vf_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmacc_vf_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (float)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vfwmacc_vf_f64m2(op0, op1, op2, op3) \ +__builtin_rvv_vfwmacc_vf_f64m2((vfloat64m2_t)(op0), (float)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfwmacc_vf_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmacc_vf_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (float)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfwmacc_vf_f64m4(op0, op1, op2, op3) \ +__builtin_rvv_vfwmacc_vf_f64m4((vfloat64m4_t)(op0), (float)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vfwmacc_vf_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmacc_vf_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (float)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vfwmacc_vf_f64m8(op0, op1, op2, op3) \ +__builtin_rvv_vfwmacc_vf_f64m8((vfloat64m8_t)(op0), (float)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vfwmacc_vf_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmacc_vf_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (float)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vfwnmacc_vv_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfwnmacc_vv_f64m1((vfloat64m1_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vfwnmacc_vv_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwnmacc_vv_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vfwnmacc_vv_f64m2(op0, op1, op2, op3) \ +__builtin_rvv_vfwnmacc_vv_f64m2((vfloat64m2_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfwnmacc_vv_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwnmacc_vv_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfwnmacc_vv_f64m4(op0, op1, op2, op3) \ +__builtin_rvv_vfwnmacc_vv_f64m4((vfloat64m4_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vfwnmacc_vv_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwnmacc_vv_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vfwnmacc_vv_f64m8(op0, op1, op2, op3) \ +__builtin_rvv_vfwnmacc_vv_f64m8((vfloat64m8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vfwnmacc_vv_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwnmacc_vv_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat32m4_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vfwnmacc_vf_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfwnmacc_vf_f64m1((vfloat64m1_t)(op0), (float)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vfwnmacc_vf_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwnmacc_vf_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (float)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vfwnmacc_vf_f64m2(op0, op1, op2, op3) \ +__builtin_rvv_vfwnmacc_vf_f64m2((vfloat64m2_t)(op0), (float)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfwnmacc_vf_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwnmacc_vf_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (float)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfwnmacc_vf_f64m4(op0, op1, op2, op3) \ +__builtin_rvv_vfwnmacc_vf_f64m4((vfloat64m4_t)(op0), (float)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vfwnmacc_vf_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwnmacc_vf_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (float)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vfwnmacc_vf_f64m8(op0, op1, op2, op3) \ +__builtin_rvv_vfwnmacc_vf_f64m8((vfloat64m8_t)(op0), (float)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vfwnmacc_vf_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwnmacc_vf_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (float)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vfwmsac_vv_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfwmsac_vv_f64m1((vfloat64m1_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vfwmsac_vv_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmsac_vv_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vfwmsac_vv_f64m2(op0, op1, op2, op3) \ +__builtin_rvv_vfwmsac_vv_f64m2((vfloat64m2_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfwmsac_vv_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmsac_vv_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfwmsac_vv_f64m4(op0, op1, op2, op3) \ +__builtin_rvv_vfwmsac_vv_f64m4((vfloat64m4_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vfwmsac_vv_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmsac_vv_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vfwmsac_vv_f64m8(op0, op1, op2, op3) \ +__builtin_rvv_vfwmsac_vv_f64m8((vfloat64m8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vfwmsac_vv_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmsac_vv_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat32m4_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vfwmsac_vf_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfwmsac_vf_f64m1((vfloat64m1_t)(op0), (float)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vfwmsac_vf_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmsac_vf_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (float)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vfwmsac_vf_f64m2(op0, op1, op2, op3) \ +__builtin_rvv_vfwmsac_vf_f64m2((vfloat64m2_t)(op0), (float)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfwmsac_vf_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmsac_vf_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (float)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfwmsac_vf_f64m4(op0, op1, op2, op3) \ +__builtin_rvv_vfwmsac_vf_f64m4((vfloat64m4_t)(op0), (float)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vfwmsac_vf_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmsac_vf_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (float)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vfwmsac_vf_f64m8(op0, op1, op2, op3) \ +__builtin_rvv_vfwmsac_vf_f64m8((vfloat64m8_t)(op0), (float)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vfwmsac_vf_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmsac_vf_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (float)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vfwnmsac_vv_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfwnmsac_vv_f64m1((vfloat64m1_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vfwnmsac_vv_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwnmsac_vv_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vfwnmsac_vv_f64m2(op0, op1, op2, op3) \ +__builtin_rvv_vfwnmsac_vv_f64m2((vfloat64m2_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfwnmsac_vv_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwnmsac_vv_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfwnmsac_vv_f64m4(op0, op1, op2, op3) \ +__builtin_rvv_vfwnmsac_vv_f64m4((vfloat64m4_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vfwnmsac_vv_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwnmsac_vv_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vfwnmsac_vv_f64m8(op0, op1, op2, op3) \ +__builtin_rvv_vfwnmsac_vv_f64m8((vfloat64m8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vfwnmsac_vv_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwnmsac_vv_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat32m4_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vfwnmsac_vf_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfwnmsac_vf_f64m1((vfloat64m1_t)(op0), (float)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vfwnmsac_vf_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwnmsac_vf_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (float)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vfwnmsac_vf_f64m2(op0, op1, op2, op3) \ +__builtin_rvv_vfwnmsac_vf_f64m2((vfloat64m2_t)(op0), (float)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfwnmsac_vf_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwnmsac_vf_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (float)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfwnmsac_vf_f64m4(op0, op1, op2, op3) \ +__builtin_rvv_vfwnmsac_vf_f64m4((vfloat64m4_t)(op0), (float)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vfwnmsac_vf_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwnmsac_vf_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (float)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vfwnmsac_vf_f64m8(op0, op1, op2, op3) \ +__builtin_rvv_vfwnmsac_vf_f64m8((vfloat64m8_t)(op0), (float)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vfwnmsac_vf_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwnmsac_vf_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (float)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vfwredsum_vs_f32m1_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfwredsum_vs_f32m1_f64m1((vfloat64m1_t)(op0), (vfloat32m1_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfwredsum_vs_f32m1_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwredsum_vs_f32m1_f64m1_m((vbool32_t)(op0), (vfloat64m1_t)(op1), (vfloat32m1_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfwredsum_vs_f32m2_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfwredsum_vs_f32m2_f64m1((vfloat64m1_t)(op0), (vfloat32m2_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfwredsum_vs_f32m2_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwredsum_vs_f32m2_f64m1_m((vbool16_t)(op0), (vfloat64m1_t)(op1), (vfloat32m2_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfwredsum_vs_f32m4_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfwredsum_vs_f32m4_f64m1((vfloat64m1_t)(op0), (vfloat32m4_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfwredsum_vs_f32m4_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwredsum_vs_f32m4_f64m1_m((vbool8_t)(op0), (vfloat64m1_t)(op1), (vfloat32m4_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfwredsum_vs_f32m8_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfwredsum_vs_f32m8_f64m1((vfloat64m1_t)(op0), (vfloat32m8_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfwredsum_vs_f32m8_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwredsum_vs_f32m8_f64m1_m((vbool4_t)(op0), (vfloat64m1_t)(op1), (vfloat32m8_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfwredsum_vs_f32mf2_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfwredsum_vs_f32mf2_f64m1((vfloat64m1_t)(op0), (vfloat32mf2_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfwredsum_vs_f32mf2_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwredsum_vs_f32mf2_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat32mf2_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfwredosum_vs_f32m1_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfwredosum_vs_f32m1_f64m1((vfloat64m1_t)(op0), (vfloat32m1_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfwredosum_vs_f32m1_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwredosum_vs_f32m1_f64m1_m((vbool32_t)(op0), (vfloat64m1_t)(op1), (vfloat32m1_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfwredosum_vs_f32m2_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfwredosum_vs_f32m2_f64m1((vfloat64m1_t)(op0), (vfloat32m2_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfwredosum_vs_f32m2_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwredosum_vs_f32m2_f64m1_m((vbool16_t)(op0), (vfloat64m1_t)(op1), (vfloat32m2_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfwredosum_vs_f32m4_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfwredosum_vs_f32m4_f64m1((vfloat64m1_t)(op0), (vfloat32m4_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfwredosum_vs_f32m4_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwredosum_vs_f32m4_f64m1_m((vbool8_t)(op0), (vfloat64m1_t)(op1), (vfloat32m4_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfwredosum_vs_f32m8_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfwredosum_vs_f32m8_f64m1((vfloat64m1_t)(op0), (vfloat32m8_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfwredosum_vs_f32m8_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwredosum_vs_f32m8_f64m1_m((vbool4_t)(op0), (vfloat64m1_t)(op1), (vfloat32m8_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfwredosum_vs_f32mf2_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vfwredosum_vs_f32mf2_f64m1((vfloat64m1_t)(op0), (vfloat32mf2_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfwredosum_vs_f32mf2_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwredosum_vs_f32mf2_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat32mf2_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vfncvt_f_f_w_f32mf2(op0, op1) \ +__builtin_rvv_vfncvt_f_f_w_f32mf2((vfloat64m1_t)(op0), (size_t)(op1)) +#define vfncvt_f_f_w_f32mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_f_f_w_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfncvt_f_f_w_f32m1(op0, op1) \ +__builtin_rvv_vfncvt_f_f_w_f32m1((vfloat64m2_t)(op0), (size_t)(op1)) +#define vfncvt_f_f_w_f32m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_f_f_w_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vfncvt_f_f_w_f32m2(op0, op1) \ +__builtin_rvv_vfncvt_f_f_w_f32m2((vfloat64m4_t)(op0), (size_t)(op1)) +#define vfncvt_f_f_w_f32m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_f_f_w_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vfncvt_f_f_w_f32m4(op0, op1) \ +__builtin_rvv_vfncvt_f_f_w_f32m4((vfloat64m8_t)(op0), (size_t)(op1)) +#define vfncvt_f_f_w_f32m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_f_f_w_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vfncvt_rod_f_f_w_f32mf2(op0, op1) \ +__builtin_rvv_vfncvt_rod_f_f_w_f32mf2((vfloat64m1_t)(op0), (size_t)(op1)) +#define vfncvt_rod_f_f_w_f32mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_rod_f_f_w_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vfncvt_rod_f_f_w_f32m1(op0, op1) \ +__builtin_rvv_vfncvt_rod_f_f_w_f32m1((vfloat64m2_t)(op0), (size_t)(op1)) +#define vfncvt_rod_f_f_w_f32m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_rod_f_f_w_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vfncvt_rod_f_f_w_f32m2(op0, op1) \ +__builtin_rvv_vfncvt_rod_f_f_w_f32m2((vfloat64m4_t)(op0), (size_t)(op1)) +#define vfncvt_rod_f_f_w_f32m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_rod_f_f_w_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vfncvt_rod_f_f_w_f32m4(op0, op1) \ +__builtin_rvv_vfncvt_rod_f_f_w_f32m4((vfloat64m8_t)(op0), (size_t)(op1)) +#define vfncvt_rod_f_f_w_f32m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_rod_f_f_w_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vfwcvt_f_f_v_f64m1(op0, op1) \ +__builtin_rvv_vfwcvt_f_f_v_f64m1((vfloat32mf2_t)(op0), (size_t)(op1)) +#define vfwcvt_f_f_v_f64m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_f_f_v_f64m1_m((vbool64_t)(op0), (vfloat64m1_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vfwcvt_f_f_v_f64m2(op0, op1) \ +__builtin_rvv_vfwcvt_f_f_v_f64m2((vfloat32m1_t)(op0), (size_t)(op1)) +#define vfwcvt_f_f_v_f64m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_f_f_v_f64m2_m((vbool32_t)(op0), (vfloat64m2_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfwcvt_f_f_v_f64m4(op0, op1) \ +__builtin_rvv_vfwcvt_f_f_v_f64m4((vfloat32m2_t)(op0), (size_t)(op1)) +#define vfwcvt_f_f_v_f64m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_f_f_v_f64m4_m((vbool16_t)(op0), (vfloat64m4_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vfwcvt_f_f_v_f64m8(op0, op1) \ +__builtin_rvv_vfwcvt_f_f_v_f64m8((vfloat32m4_t)(op0), (size_t)(op1)) +#define vfwcvt_f_f_v_f64m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_f_f_v_f64m8_m((vbool8_t)(op0), (vfloat64m8_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#endif + +#if defined(__riscv_zfh) +#define vsse16_v_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vsse16_v_f16m1((_Float16 *)(op0), (ptrdiff_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vsse16_v_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse16_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (ptrdiff_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vsse16_v_f16m2(op0, op1, op2, op3) \ +__builtin_rvv_vsse16_v_f16m2((_Float16 *)(op0), (ptrdiff_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vsse16_v_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse16_v_f16m2_m((vbool8_t)(op0), (_Float16 *)(op1), (ptrdiff_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vsse16_v_f16m4(op0, op1, op2, op3) \ +__builtin_rvv_vsse16_v_f16m4((_Float16 *)(op0), (ptrdiff_t)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vsse16_v_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse16_v_f16m4_m((vbool4_t)(op0), (_Float16 *)(op1), (ptrdiff_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vsse16_v_f16m8(op0, op1, op2, op3) \ +__builtin_rvv_vsse16_v_f16m8((_Float16 *)(op0), (ptrdiff_t)(op1), (vfloat16m8_t)(op2), (size_t)(op3)) +#define vsse16_v_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse16_v_f16m8_m((vbool2_t)(op0), (_Float16 *)(op1), (ptrdiff_t)(op2), (vfloat16m8_t)(op3), (size_t)(op4)) +#define vsse16_v_f16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsse16_v_f16mf2((_Float16 *)(op0), (ptrdiff_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vsse16_v_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse16_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (ptrdiff_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vsse16_v_f16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsse16_v_f16mf4((_Float16 *)(op0), (ptrdiff_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vsse16_v_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsse16_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (ptrdiff_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vluxei8_v_f16m1(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_f16m1((const _Float16 *)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vluxei8_v_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (const _Float16 *)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vluxei8_v_f16m2(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_f16m2((const _Float16 *)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vluxei8_v_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (const _Float16 *)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vluxei8_v_f16m4(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_f16m4((const _Float16 *)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vluxei8_v_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (const _Float16 *)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vluxei8_v_f16m8(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_f16m8((const _Float16 *)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vluxei8_v_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (const _Float16 *)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vluxei8_v_f16mf2(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_f16mf2((const _Float16 *)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vluxei8_v_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (const _Float16 *)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vluxei8_v_f16mf4(op0, op1, op2) \ +__builtin_rvv_vluxei8_v_f16mf4((const _Float16 *)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vluxei8_v_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei8_v_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (const _Float16 *)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vluxei16_v_f16m1(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_f16m1((const _Float16 *)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vluxei16_v_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (const _Float16 *)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vluxei16_v_f16m2(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_f16m2((const _Float16 *)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vluxei16_v_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (const _Float16 *)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vluxei16_v_f16m4(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_f16m4((const _Float16 *)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vluxei16_v_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (const _Float16 *)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vluxei16_v_f16m8(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_f16m8((const _Float16 *)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vluxei16_v_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (const _Float16 *)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vluxei16_v_f16mf2(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_f16mf2((const _Float16 *)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vluxei16_v_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (const _Float16 *)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vluxei16_v_f16mf4(op0, op1, op2) \ +__builtin_rvv_vluxei16_v_f16mf4((const _Float16 *)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vluxei16_v_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei16_v_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (const _Float16 *)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vluxei32_v_f16m1(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_f16m1((const _Float16 *)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vluxei32_v_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (const _Float16 *)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vluxei32_v_f16m2(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_f16m2((const _Float16 *)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vluxei32_v_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (const _Float16 *)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vluxei32_v_f16m4(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_f16m4((const _Float16 *)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vluxei32_v_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (const _Float16 *)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vluxei32_v_f16mf2(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_f16mf2((const _Float16 *)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vluxei32_v_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (const _Float16 *)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vluxei32_v_f16mf4(op0, op1, op2) \ +__builtin_rvv_vluxei32_v_f16mf4((const _Float16 *)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vluxei32_v_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei32_v_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (const _Float16 *)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vluxei64_v_f16m1(op0, op1, op2) \ +__builtin_rvv_vluxei64_v_f16m1((const _Float16 *)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vluxei64_v_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei64_v_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (const _Float16 *)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vluxei64_v_f16m2(op0, op1, op2) \ +__builtin_rvv_vluxei64_v_f16m2((const _Float16 *)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vluxei64_v_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei64_v_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (const _Float16 *)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vluxei64_v_f16mf2(op0, op1, op2) \ +__builtin_rvv_vluxei64_v_f16mf2((const _Float16 *)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vluxei64_v_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei64_v_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (const _Float16 *)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vluxei64_v_f16mf4(op0, op1, op2) \ +__builtin_rvv_vluxei64_v_f16mf4((const _Float16 *)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vluxei64_v_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxei64_v_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (const _Float16 *)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vloxei8_v_f16m1(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_f16m1((const _Float16 *)(op0), (vuint8mf2_t)(op1), (size_t)(op2)) +#define vloxei8_v_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (const _Float16 *)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vloxei8_v_f16m2(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_f16m2((const _Float16 *)(op0), (vuint8m1_t)(op1), (size_t)(op2)) +#define vloxei8_v_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (const _Float16 *)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vloxei8_v_f16m4(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_f16m4((const _Float16 *)(op0), (vuint8m2_t)(op1), (size_t)(op2)) +#define vloxei8_v_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (const _Float16 *)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vloxei8_v_f16m8(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_f16m8((const _Float16 *)(op0), (vuint8m4_t)(op1), (size_t)(op2)) +#define vloxei8_v_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (const _Float16 *)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vloxei8_v_f16mf2(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_f16mf2((const _Float16 *)(op0), (vuint8mf4_t)(op1), (size_t)(op2)) +#define vloxei8_v_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (const _Float16 *)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vloxei8_v_f16mf4(op0, op1, op2) \ +__builtin_rvv_vloxei8_v_f16mf4((const _Float16 *)(op0), (vuint8mf8_t)(op1), (size_t)(op2)) +#define vloxei8_v_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei8_v_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (const _Float16 *)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vloxei16_v_f16m1(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_f16m1((const _Float16 *)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vloxei16_v_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (const _Float16 *)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vloxei16_v_f16m2(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_f16m2((const _Float16 *)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vloxei16_v_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (const _Float16 *)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vloxei16_v_f16m4(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_f16m4((const _Float16 *)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vloxei16_v_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (const _Float16 *)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vloxei16_v_f16m8(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_f16m8((const _Float16 *)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vloxei16_v_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (const _Float16 *)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vloxei16_v_f16mf2(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_f16mf2((const _Float16 *)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vloxei16_v_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (const _Float16 *)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vloxei16_v_f16mf4(op0, op1, op2) \ +__builtin_rvv_vloxei16_v_f16mf4((const _Float16 *)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vloxei16_v_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei16_v_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (const _Float16 *)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vloxei32_v_f16m1(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_f16m1((const _Float16 *)(op0), (vuint32m2_t)(op1), (size_t)(op2)) +#define vloxei32_v_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (const _Float16 *)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vloxei32_v_f16m2(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_f16m2((const _Float16 *)(op0), (vuint32m4_t)(op1), (size_t)(op2)) +#define vloxei32_v_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (const _Float16 *)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vloxei32_v_f16m4(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_f16m4((const _Float16 *)(op0), (vuint32m8_t)(op1), (size_t)(op2)) +#define vloxei32_v_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (const _Float16 *)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vloxei32_v_f16mf2(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_f16mf2((const _Float16 *)(op0), (vuint32m1_t)(op1), (size_t)(op2)) +#define vloxei32_v_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (const _Float16 *)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vloxei32_v_f16mf4(op0, op1, op2) \ +__builtin_rvv_vloxei32_v_f16mf4((const _Float16 *)(op0), (vuint32mf2_t)(op1), (size_t)(op2)) +#define vloxei32_v_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei32_v_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (const _Float16 *)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vloxei64_v_f16m1(op0, op1, op2) \ +__builtin_rvv_vloxei64_v_f16m1((const _Float16 *)(op0), (vuint64m4_t)(op1), (size_t)(op2)) +#define vloxei64_v_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei64_v_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (const _Float16 *)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vloxei64_v_f16m2(op0, op1, op2) \ +__builtin_rvv_vloxei64_v_f16m2((const _Float16 *)(op0), (vuint64m8_t)(op1), (size_t)(op2)) +#define vloxei64_v_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei64_v_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (const _Float16 *)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vloxei64_v_f16mf2(op0, op1, op2) \ +__builtin_rvv_vloxei64_v_f16mf2((const _Float16 *)(op0), (vuint64m2_t)(op1), (size_t)(op2)) +#define vloxei64_v_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei64_v_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (const _Float16 *)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vloxei64_v_f16mf4(op0, op1, op2) \ +__builtin_rvv_vloxei64_v_f16mf4((const _Float16 *)(op0), (vuint64m1_t)(op1), (size_t)(op2)) +#define vloxei64_v_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxei64_v_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (const _Float16 *)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vmv_v_v_f16m1(op0, op1) \ +__builtin_rvv_vmv_v_v_f16m1((vfloat16m1_t)(op0), (size_t)(op1)) +#define vmv_v_v_f16m2(op0, op1) \ +__builtin_rvv_vmv_v_v_f16m2((vfloat16m2_t)(op0), (size_t)(op1)) +#define vmv_v_v_f16m4(op0, op1) \ +__builtin_rvv_vmv_v_v_f16m4((vfloat16m4_t)(op0), (size_t)(op1)) +#define vmv_v_v_f16m8(op0, op1) \ +__builtin_rvv_vmv_v_v_f16m8((vfloat16m8_t)(op0), (size_t)(op1)) +#define vmv_v_v_f16mf2(op0, op1) \ +__builtin_rvv_vmv_v_v_f16mf2((vfloat16mf2_t)(op0), (size_t)(op1)) +#define vmv_v_v_f16mf4(op0, op1) \ +__builtin_rvv_vmv_v_v_f16mf4((vfloat16mf4_t)(op0), (size_t)(op1)) +#define vfadd_vv_f16m1(op0, op1, op2) \ +__builtin_rvv_vfadd_vv_f16m1((vfloat16m1_t)(op0), (vfloat16m1_t)(op1), (size_t)(op2)) +#define vfadd_vv_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfadd_vv_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfadd_vv_f16m2(op0, op1, op2) \ +__builtin_rvv_vfadd_vv_f16m2((vfloat16m2_t)(op0), (vfloat16m2_t)(op1), (size_t)(op2)) +#define vfadd_vv_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfadd_vv_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vfadd_vv_f16m4(op0, op1, op2) \ +__builtin_rvv_vfadd_vv_f16m4((vfloat16m4_t)(op0), (vfloat16m4_t)(op1), (size_t)(op2)) +#define vfadd_vv_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfadd_vv_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vfadd_vv_f16m8(op0, op1, op2) \ +__builtin_rvv_vfadd_vv_f16m8((vfloat16m8_t)(op0), (vfloat16m8_t)(op1), (size_t)(op2)) +#define vfadd_vv_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfadd_vv_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (vfloat16m8_t)(op3), (size_t)(op4)) +#define vfadd_vv_f16mf2(op0, op1, op2) \ +__builtin_rvv_vfadd_vv_f16mf2((vfloat16mf2_t)(op0), (vfloat16mf2_t)(op1), (size_t)(op2)) +#define vfadd_vv_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfadd_vv_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vfadd_vv_f16mf4(op0, op1, op2) \ +__builtin_rvv_vfadd_vv_f16mf4((vfloat16mf4_t)(op0), (vfloat16mf4_t)(op1), (size_t)(op2)) +#define vfadd_vv_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfadd_vv_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vfadd_vf_f16m1(op0, op1, op2) \ +__builtin_rvv_vfadd_vf_f16m1((vfloat16m1_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfadd_vf_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfadd_vf_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfadd_vf_f16m2(op0, op1, op2) \ +__builtin_rvv_vfadd_vf_f16m2((vfloat16m2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfadd_vf_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfadd_vf_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfadd_vf_f16m4(op0, op1, op2) \ +__builtin_rvv_vfadd_vf_f16m4((vfloat16m4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfadd_vf_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfadd_vf_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfadd_vf_f16m8(op0, op1, op2) \ +__builtin_rvv_vfadd_vf_f16m8((vfloat16m8_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfadd_vf_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfadd_vf_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfadd_vf_f16mf2(op0, op1, op2) \ +__builtin_rvv_vfadd_vf_f16mf2((vfloat16mf2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfadd_vf_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfadd_vf_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfadd_vf_f16mf4(op0, op1, op2) \ +__builtin_rvv_vfadd_vf_f16mf4((vfloat16mf4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfadd_vf_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfadd_vf_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfsub_vv_f16m1(op0, op1, op2) \ +__builtin_rvv_vfsub_vv_f16m1((vfloat16m1_t)(op0), (vfloat16m1_t)(op1), (size_t)(op2)) +#define vfsub_vv_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsub_vv_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfsub_vv_f16m2(op0, op1, op2) \ +__builtin_rvv_vfsub_vv_f16m2((vfloat16m2_t)(op0), (vfloat16m2_t)(op1), (size_t)(op2)) +#define vfsub_vv_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsub_vv_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vfsub_vv_f16m4(op0, op1, op2) \ +__builtin_rvv_vfsub_vv_f16m4((vfloat16m4_t)(op0), (vfloat16m4_t)(op1), (size_t)(op2)) +#define vfsub_vv_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsub_vv_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vfsub_vv_f16m8(op0, op1, op2) \ +__builtin_rvv_vfsub_vv_f16m8((vfloat16m8_t)(op0), (vfloat16m8_t)(op1), (size_t)(op2)) +#define vfsub_vv_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsub_vv_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (vfloat16m8_t)(op3), (size_t)(op4)) +#define vfsub_vv_f16mf2(op0, op1, op2) \ +__builtin_rvv_vfsub_vv_f16mf2((vfloat16mf2_t)(op0), (vfloat16mf2_t)(op1), (size_t)(op2)) +#define vfsub_vv_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsub_vv_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vfsub_vv_f16mf4(op0, op1, op2) \ +__builtin_rvv_vfsub_vv_f16mf4((vfloat16mf4_t)(op0), (vfloat16mf4_t)(op1), (size_t)(op2)) +#define vfsub_vv_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsub_vv_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vfsub_vf_f16m1(op0, op1, op2) \ +__builtin_rvv_vfsub_vf_f16m1((vfloat16m1_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfsub_vf_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsub_vf_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfsub_vf_f16m2(op0, op1, op2) \ +__builtin_rvv_vfsub_vf_f16m2((vfloat16m2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfsub_vf_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsub_vf_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfsub_vf_f16m4(op0, op1, op2) \ +__builtin_rvv_vfsub_vf_f16m4((vfloat16m4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfsub_vf_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsub_vf_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfsub_vf_f16m8(op0, op1, op2) \ +__builtin_rvv_vfsub_vf_f16m8((vfloat16m8_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfsub_vf_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsub_vf_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfsub_vf_f16mf2(op0, op1, op2) \ +__builtin_rvv_vfsub_vf_f16mf2((vfloat16mf2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfsub_vf_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsub_vf_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfsub_vf_f16mf4(op0, op1, op2) \ +__builtin_rvv_vfsub_vf_f16mf4((vfloat16mf4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfsub_vf_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsub_vf_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfrsub_vf_f16m1(op0, op1, op2) \ +__builtin_rvv_vfrsub_vf_f16m1((vfloat16m1_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfrsub_vf_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfrsub_vf_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfrsub_vf_f16m2(op0, op1, op2) \ +__builtin_rvv_vfrsub_vf_f16m2((vfloat16m2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfrsub_vf_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfrsub_vf_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfrsub_vf_f16m4(op0, op1, op2) \ +__builtin_rvv_vfrsub_vf_f16m4((vfloat16m4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfrsub_vf_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfrsub_vf_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfrsub_vf_f16m8(op0, op1, op2) \ +__builtin_rvv_vfrsub_vf_f16m8((vfloat16m8_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfrsub_vf_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfrsub_vf_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfrsub_vf_f16mf2(op0, op1, op2) \ +__builtin_rvv_vfrsub_vf_f16mf2((vfloat16mf2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfrsub_vf_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfrsub_vf_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfrsub_vf_f16mf4(op0, op1, op2) \ +__builtin_rvv_vfrsub_vf_f16mf4((vfloat16mf4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfrsub_vf_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfrsub_vf_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfmul_vv_f16m1(op0, op1, op2) \ +__builtin_rvv_vfmul_vv_f16m1((vfloat16m1_t)(op0), (vfloat16m1_t)(op1), (size_t)(op2)) +#define vfmul_vv_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmul_vv_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfmul_vv_f16m2(op0, op1, op2) \ +__builtin_rvv_vfmul_vv_f16m2((vfloat16m2_t)(op0), (vfloat16m2_t)(op1), (size_t)(op2)) +#define vfmul_vv_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmul_vv_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vfmul_vv_f16m4(op0, op1, op2) \ +__builtin_rvv_vfmul_vv_f16m4((vfloat16m4_t)(op0), (vfloat16m4_t)(op1), (size_t)(op2)) +#define vfmul_vv_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmul_vv_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vfmul_vv_f16m8(op0, op1, op2) \ +__builtin_rvv_vfmul_vv_f16m8((vfloat16m8_t)(op0), (vfloat16m8_t)(op1), (size_t)(op2)) +#define vfmul_vv_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmul_vv_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (vfloat16m8_t)(op3), (size_t)(op4)) +#define vfmul_vv_f16mf2(op0, op1, op2) \ +__builtin_rvv_vfmul_vv_f16mf2((vfloat16mf2_t)(op0), (vfloat16mf2_t)(op1), (size_t)(op2)) +#define vfmul_vv_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmul_vv_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vfmul_vv_f16mf4(op0, op1, op2) \ +__builtin_rvv_vfmul_vv_f16mf4((vfloat16mf4_t)(op0), (vfloat16mf4_t)(op1), (size_t)(op2)) +#define vfmul_vv_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmul_vv_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vfmul_vf_f16m1(op0, op1, op2) \ +__builtin_rvv_vfmul_vf_f16m1((vfloat16m1_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfmul_vf_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmul_vf_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfmul_vf_f16m2(op0, op1, op2) \ +__builtin_rvv_vfmul_vf_f16m2((vfloat16m2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfmul_vf_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmul_vf_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfmul_vf_f16m4(op0, op1, op2) \ +__builtin_rvv_vfmul_vf_f16m4((vfloat16m4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfmul_vf_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmul_vf_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfmul_vf_f16m8(op0, op1, op2) \ +__builtin_rvv_vfmul_vf_f16m8((vfloat16m8_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfmul_vf_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmul_vf_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfmul_vf_f16mf2(op0, op1, op2) \ +__builtin_rvv_vfmul_vf_f16mf2((vfloat16mf2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfmul_vf_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmul_vf_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfmul_vf_f16mf4(op0, op1, op2) \ +__builtin_rvv_vfmul_vf_f16mf4((vfloat16mf4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfmul_vf_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmul_vf_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfdiv_vv_f16m1(op0, op1, op2) \ +__builtin_rvv_vfdiv_vv_f16m1((vfloat16m1_t)(op0), (vfloat16m1_t)(op1), (size_t)(op2)) +#define vfdiv_vv_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfdiv_vv_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfdiv_vv_f16m2(op0, op1, op2) \ +__builtin_rvv_vfdiv_vv_f16m2((vfloat16m2_t)(op0), (vfloat16m2_t)(op1), (size_t)(op2)) +#define vfdiv_vv_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfdiv_vv_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vfdiv_vv_f16m4(op0, op1, op2) \ +__builtin_rvv_vfdiv_vv_f16m4((vfloat16m4_t)(op0), (vfloat16m4_t)(op1), (size_t)(op2)) +#define vfdiv_vv_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfdiv_vv_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vfdiv_vv_f16m8(op0, op1, op2) \ +__builtin_rvv_vfdiv_vv_f16m8((vfloat16m8_t)(op0), (vfloat16m8_t)(op1), (size_t)(op2)) +#define vfdiv_vv_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfdiv_vv_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (vfloat16m8_t)(op3), (size_t)(op4)) +#define vfdiv_vv_f16mf2(op0, op1, op2) \ +__builtin_rvv_vfdiv_vv_f16mf2((vfloat16mf2_t)(op0), (vfloat16mf2_t)(op1), (size_t)(op2)) +#define vfdiv_vv_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfdiv_vv_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vfdiv_vv_f16mf4(op0, op1, op2) \ +__builtin_rvv_vfdiv_vv_f16mf4((vfloat16mf4_t)(op0), (vfloat16mf4_t)(op1), (size_t)(op2)) +#define vfdiv_vv_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfdiv_vv_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vfdiv_vf_f16m1(op0, op1, op2) \ +__builtin_rvv_vfdiv_vf_f16m1((vfloat16m1_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfdiv_vf_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfdiv_vf_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfdiv_vf_f16m2(op0, op1, op2) \ +__builtin_rvv_vfdiv_vf_f16m2((vfloat16m2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfdiv_vf_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfdiv_vf_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfdiv_vf_f16m4(op0, op1, op2) \ +__builtin_rvv_vfdiv_vf_f16m4((vfloat16m4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfdiv_vf_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfdiv_vf_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfdiv_vf_f16m8(op0, op1, op2) \ +__builtin_rvv_vfdiv_vf_f16m8((vfloat16m8_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfdiv_vf_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfdiv_vf_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfdiv_vf_f16mf2(op0, op1, op2) \ +__builtin_rvv_vfdiv_vf_f16mf2((vfloat16mf2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfdiv_vf_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfdiv_vf_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfdiv_vf_f16mf4(op0, op1, op2) \ +__builtin_rvv_vfdiv_vf_f16mf4((vfloat16mf4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfdiv_vf_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfdiv_vf_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfrdiv_vf_f16m1(op0, op1, op2) \ +__builtin_rvv_vfrdiv_vf_f16m1((vfloat16m1_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfrdiv_vf_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfrdiv_vf_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfrdiv_vf_f16m2(op0, op1, op2) \ +__builtin_rvv_vfrdiv_vf_f16m2((vfloat16m2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfrdiv_vf_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfrdiv_vf_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfrdiv_vf_f16m4(op0, op1, op2) \ +__builtin_rvv_vfrdiv_vf_f16m4((vfloat16m4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfrdiv_vf_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfrdiv_vf_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfrdiv_vf_f16m8(op0, op1, op2) \ +__builtin_rvv_vfrdiv_vf_f16m8((vfloat16m8_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfrdiv_vf_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfrdiv_vf_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfrdiv_vf_f16mf2(op0, op1, op2) \ +__builtin_rvv_vfrdiv_vf_f16mf2((vfloat16mf2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfrdiv_vf_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfrdiv_vf_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfrdiv_vf_f16mf4(op0, op1, op2) \ +__builtin_rvv_vfrdiv_vf_f16mf4((vfloat16mf4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfrdiv_vf_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfrdiv_vf_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfmacc_vv_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vfmacc_vv_f16m1((vfloat16m1_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfmacc_vv_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmacc_vv_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfmacc_vv_f16m2(op0, op1, op2, op3) \ +__builtin_rvv_vfmacc_vv_f16m2((vfloat16m2_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vfmacc_vv_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmacc_vv_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vfmacc_vv_f16m4(op0, op1, op2, op3) \ +__builtin_rvv_vfmacc_vv_f16m4((vfloat16m4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vfmacc_vv_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmacc_vv_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vfmacc_vv_f16m8(op0, op1, op2, op3) \ +__builtin_rvv_vfmacc_vv_f16m8((vfloat16m8_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (size_t)(op3)) +#define vfmacc_vv_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmacc_vv_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (vfloat16m8_t)(op3), (size_t)(op4)) +#define vfmacc_vv_f16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vfmacc_vv_f16mf2((vfloat16mf2_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vfmacc_vv_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmacc_vv_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vfmacc_vv_f16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vfmacc_vv_f16mf4((vfloat16mf4_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vfmacc_vv_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmacc_vv_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vfmacc_vf_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vfmacc_vf_f16m1((vfloat16m1_t)(op0), (_Float16)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfmacc_vf_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmacc_vf_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (_Float16)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfmacc_vf_f16m2(op0, op1, op2, op3) \ +__builtin_rvv_vfmacc_vf_f16m2((vfloat16m2_t)(op0), (_Float16)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vfmacc_vf_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmacc_vf_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (_Float16)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vfmacc_vf_f16m4(op0, op1, op2, op3) \ +__builtin_rvv_vfmacc_vf_f16m4((vfloat16m4_t)(op0), (_Float16)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vfmacc_vf_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmacc_vf_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (_Float16)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vfmacc_vf_f16m8(op0, op1, op2, op3) \ +__builtin_rvv_vfmacc_vf_f16m8((vfloat16m8_t)(op0), (_Float16)(op1), (vfloat16m8_t)(op2), (size_t)(op3)) +#define vfmacc_vf_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmacc_vf_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (_Float16)(op2), (vfloat16m8_t)(op3), (size_t)(op4)) +#define vfmacc_vf_f16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vfmacc_vf_f16mf2((vfloat16mf2_t)(op0), (_Float16)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vfmacc_vf_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmacc_vf_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (_Float16)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vfmacc_vf_f16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vfmacc_vf_f16mf4((vfloat16mf4_t)(op0), (_Float16)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vfmacc_vf_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmacc_vf_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (_Float16)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vfnmacc_vv_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vfnmacc_vv_f16m1((vfloat16m1_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfnmacc_vv_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmacc_vv_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfnmacc_vv_f16m2(op0, op1, op2, op3) \ +__builtin_rvv_vfnmacc_vv_f16m2((vfloat16m2_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vfnmacc_vv_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmacc_vv_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vfnmacc_vv_f16m4(op0, op1, op2, op3) \ +__builtin_rvv_vfnmacc_vv_f16m4((vfloat16m4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vfnmacc_vv_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmacc_vv_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vfnmacc_vv_f16m8(op0, op1, op2, op3) \ +__builtin_rvv_vfnmacc_vv_f16m8((vfloat16m8_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (size_t)(op3)) +#define vfnmacc_vv_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmacc_vv_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (vfloat16m8_t)(op3), (size_t)(op4)) +#define vfnmacc_vv_f16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vfnmacc_vv_f16mf2((vfloat16mf2_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vfnmacc_vv_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmacc_vv_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vfnmacc_vv_f16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vfnmacc_vv_f16mf4((vfloat16mf4_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vfnmacc_vv_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmacc_vv_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vfnmacc_vf_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vfnmacc_vf_f16m1((vfloat16m1_t)(op0), (_Float16)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfnmacc_vf_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmacc_vf_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (_Float16)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfnmacc_vf_f16m2(op0, op1, op2, op3) \ +__builtin_rvv_vfnmacc_vf_f16m2((vfloat16m2_t)(op0), (_Float16)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vfnmacc_vf_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmacc_vf_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (_Float16)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vfnmacc_vf_f16m4(op0, op1, op2, op3) \ +__builtin_rvv_vfnmacc_vf_f16m4((vfloat16m4_t)(op0), (_Float16)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vfnmacc_vf_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmacc_vf_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (_Float16)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vfnmacc_vf_f16m8(op0, op1, op2, op3) \ +__builtin_rvv_vfnmacc_vf_f16m8((vfloat16m8_t)(op0), (_Float16)(op1), (vfloat16m8_t)(op2), (size_t)(op3)) +#define vfnmacc_vf_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmacc_vf_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (_Float16)(op2), (vfloat16m8_t)(op3), (size_t)(op4)) +#define vfnmacc_vf_f16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vfnmacc_vf_f16mf2((vfloat16mf2_t)(op0), (_Float16)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vfnmacc_vf_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmacc_vf_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (_Float16)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vfnmacc_vf_f16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vfnmacc_vf_f16mf4((vfloat16mf4_t)(op0), (_Float16)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vfnmacc_vf_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmacc_vf_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (_Float16)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vfmsac_vv_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vfmsac_vv_f16m1((vfloat16m1_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfmsac_vv_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsac_vv_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfmsac_vv_f16m2(op0, op1, op2, op3) \ +__builtin_rvv_vfmsac_vv_f16m2((vfloat16m2_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vfmsac_vv_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsac_vv_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vfmsac_vv_f16m4(op0, op1, op2, op3) \ +__builtin_rvv_vfmsac_vv_f16m4((vfloat16m4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vfmsac_vv_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsac_vv_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vfmsac_vv_f16m8(op0, op1, op2, op3) \ +__builtin_rvv_vfmsac_vv_f16m8((vfloat16m8_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (size_t)(op3)) +#define vfmsac_vv_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsac_vv_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (vfloat16m8_t)(op3), (size_t)(op4)) +#define vfmsac_vv_f16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vfmsac_vv_f16mf2((vfloat16mf2_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vfmsac_vv_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsac_vv_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vfmsac_vv_f16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vfmsac_vv_f16mf4((vfloat16mf4_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vfmsac_vv_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsac_vv_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vfmsac_vf_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vfmsac_vf_f16m1((vfloat16m1_t)(op0), (_Float16)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfmsac_vf_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsac_vf_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (_Float16)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfmsac_vf_f16m2(op0, op1, op2, op3) \ +__builtin_rvv_vfmsac_vf_f16m2((vfloat16m2_t)(op0), (_Float16)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vfmsac_vf_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsac_vf_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (_Float16)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vfmsac_vf_f16m4(op0, op1, op2, op3) \ +__builtin_rvv_vfmsac_vf_f16m4((vfloat16m4_t)(op0), (_Float16)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vfmsac_vf_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsac_vf_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (_Float16)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vfmsac_vf_f16m8(op0, op1, op2, op3) \ +__builtin_rvv_vfmsac_vf_f16m8((vfloat16m8_t)(op0), (_Float16)(op1), (vfloat16m8_t)(op2), (size_t)(op3)) +#define vfmsac_vf_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsac_vf_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (_Float16)(op2), (vfloat16m8_t)(op3), (size_t)(op4)) +#define vfmsac_vf_f16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vfmsac_vf_f16mf2((vfloat16mf2_t)(op0), (_Float16)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vfmsac_vf_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsac_vf_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (_Float16)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vfmsac_vf_f16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vfmsac_vf_f16mf4((vfloat16mf4_t)(op0), (_Float16)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vfmsac_vf_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsac_vf_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (_Float16)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vfnmsac_vv_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsac_vv_f16m1((vfloat16m1_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfnmsac_vv_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsac_vv_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfnmsac_vv_f16m2(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsac_vv_f16m2((vfloat16m2_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vfnmsac_vv_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsac_vv_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vfnmsac_vv_f16m4(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsac_vv_f16m4((vfloat16m4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vfnmsac_vv_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsac_vv_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vfnmsac_vv_f16m8(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsac_vv_f16m8((vfloat16m8_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (size_t)(op3)) +#define vfnmsac_vv_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsac_vv_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (vfloat16m8_t)(op3), (size_t)(op4)) +#define vfnmsac_vv_f16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsac_vv_f16mf2((vfloat16mf2_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vfnmsac_vv_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsac_vv_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vfnmsac_vv_f16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsac_vv_f16mf4((vfloat16mf4_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vfnmsac_vv_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsac_vv_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vfnmsac_vf_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsac_vf_f16m1((vfloat16m1_t)(op0), (_Float16)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfnmsac_vf_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsac_vf_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (_Float16)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfnmsac_vf_f16m2(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsac_vf_f16m2((vfloat16m2_t)(op0), (_Float16)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vfnmsac_vf_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsac_vf_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (_Float16)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vfnmsac_vf_f16m4(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsac_vf_f16m4((vfloat16m4_t)(op0), (_Float16)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vfnmsac_vf_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsac_vf_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (_Float16)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vfnmsac_vf_f16m8(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsac_vf_f16m8((vfloat16m8_t)(op0), (_Float16)(op1), (vfloat16m8_t)(op2), (size_t)(op3)) +#define vfnmsac_vf_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsac_vf_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (_Float16)(op2), (vfloat16m8_t)(op3), (size_t)(op4)) +#define vfnmsac_vf_f16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsac_vf_f16mf2((vfloat16mf2_t)(op0), (_Float16)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vfnmsac_vf_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsac_vf_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (_Float16)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vfnmsac_vf_f16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsac_vf_f16mf4((vfloat16mf4_t)(op0), (_Float16)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vfnmsac_vf_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsac_vf_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (_Float16)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vfmadd_vv_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vfmadd_vv_f16m1((vfloat16m1_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfmadd_vv_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmadd_vv_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfmadd_vv_f16m2(op0, op1, op2, op3) \ +__builtin_rvv_vfmadd_vv_f16m2((vfloat16m2_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vfmadd_vv_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmadd_vv_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vfmadd_vv_f16m4(op0, op1, op2, op3) \ +__builtin_rvv_vfmadd_vv_f16m4((vfloat16m4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vfmadd_vv_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmadd_vv_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vfmadd_vv_f16m8(op0, op1, op2, op3) \ +__builtin_rvv_vfmadd_vv_f16m8((vfloat16m8_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (size_t)(op3)) +#define vfmadd_vv_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmadd_vv_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (vfloat16m8_t)(op3), (size_t)(op4)) +#define vfmadd_vv_f16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vfmadd_vv_f16mf2((vfloat16mf2_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vfmadd_vv_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmadd_vv_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vfmadd_vv_f16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vfmadd_vv_f16mf4((vfloat16mf4_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vfmadd_vv_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmadd_vv_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vfmadd_vf_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vfmadd_vf_f16m1((vfloat16m1_t)(op0), (_Float16)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfmadd_vf_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmadd_vf_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (_Float16)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfmadd_vf_f16m2(op0, op1, op2, op3) \ +__builtin_rvv_vfmadd_vf_f16m2((vfloat16m2_t)(op0), (_Float16)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vfmadd_vf_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmadd_vf_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (_Float16)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vfmadd_vf_f16m4(op0, op1, op2, op3) \ +__builtin_rvv_vfmadd_vf_f16m4((vfloat16m4_t)(op0), (_Float16)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vfmadd_vf_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmadd_vf_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (_Float16)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vfmadd_vf_f16m8(op0, op1, op2, op3) \ +__builtin_rvv_vfmadd_vf_f16m8((vfloat16m8_t)(op0), (_Float16)(op1), (vfloat16m8_t)(op2), (size_t)(op3)) +#define vfmadd_vf_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmadd_vf_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (_Float16)(op2), (vfloat16m8_t)(op3), (size_t)(op4)) +#define vfmadd_vf_f16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vfmadd_vf_f16mf2((vfloat16mf2_t)(op0), (_Float16)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vfmadd_vf_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmadd_vf_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (_Float16)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vfmadd_vf_f16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vfmadd_vf_f16mf4((vfloat16mf4_t)(op0), (_Float16)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vfmadd_vf_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmadd_vf_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (_Float16)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vfnmadd_vv_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vfnmadd_vv_f16m1((vfloat16m1_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfnmadd_vv_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmadd_vv_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfnmadd_vv_f16m2(op0, op1, op2, op3) \ +__builtin_rvv_vfnmadd_vv_f16m2((vfloat16m2_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vfnmadd_vv_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmadd_vv_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vfnmadd_vv_f16m4(op0, op1, op2, op3) \ +__builtin_rvv_vfnmadd_vv_f16m4((vfloat16m4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vfnmadd_vv_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmadd_vv_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vfnmadd_vv_f16m8(op0, op1, op2, op3) \ +__builtin_rvv_vfnmadd_vv_f16m8((vfloat16m8_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (size_t)(op3)) +#define vfnmadd_vv_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmadd_vv_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (vfloat16m8_t)(op3), (size_t)(op4)) +#define vfnmadd_vv_f16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vfnmadd_vv_f16mf2((vfloat16mf2_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vfnmadd_vv_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmadd_vv_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vfnmadd_vv_f16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vfnmadd_vv_f16mf4((vfloat16mf4_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vfnmadd_vv_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmadd_vv_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vfnmadd_vf_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vfnmadd_vf_f16m1((vfloat16m1_t)(op0), (_Float16)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfnmadd_vf_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmadd_vf_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (_Float16)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfnmadd_vf_f16m2(op0, op1, op2, op3) \ +__builtin_rvv_vfnmadd_vf_f16m2((vfloat16m2_t)(op0), (_Float16)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vfnmadd_vf_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmadd_vf_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (_Float16)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vfnmadd_vf_f16m4(op0, op1, op2, op3) \ +__builtin_rvv_vfnmadd_vf_f16m4((vfloat16m4_t)(op0), (_Float16)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vfnmadd_vf_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmadd_vf_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (_Float16)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vfnmadd_vf_f16m8(op0, op1, op2, op3) \ +__builtin_rvv_vfnmadd_vf_f16m8((vfloat16m8_t)(op0), (_Float16)(op1), (vfloat16m8_t)(op2), (size_t)(op3)) +#define vfnmadd_vf_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmadd_vf_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (_Float16)(op2), (vfloat16m8_t)(op3), (size_t)(op4)) +#define vfnmadd_vf_f16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vfnmadd_vf_f16mf2((vfloat16mf2_t)(op0), (_Float16)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vfnmadd_vf_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmadd_vf_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (_Float16)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vfnmadd_vf_f16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vfnmadd_vf_f16mf4((vfloat16mf4_t)(op0), (_Float16)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vfnmadd_vf_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmadd_vf_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (_Float16)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vfmsub_vv_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vfmsub_vv_f16m1((vfloat16m1_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfmsub_vv_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsub_vv_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfmsub_vv_f16m2(op0, op1, op2, op3) \ +__builtin_rvv_vfmsub_vv_f16m2((vfloat16m2_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vfmsub_vv_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsub_vv_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vfmsub_vv_f16m4(op0, op1, op2, op3) \ +__builtin_rvv_vfmsub_vv_f16m4((vfloat16m4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vfmsub_vv_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsub_vv_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vfmsub_vv_f16m8(op0, op1, op2, op3) \ +__builtin_rvv_vfmsub_vv_f16m8((vfloat16m8_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (size_t)(op3)) +#define vfmsub_vv_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsub_vv_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (vfloat16m8_t)(op3), (size_t)(op4)) +#define vfmsub_vv_f16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vfmsub_vv_f16mf2((vfloat16mf2_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vfmsub_vv_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsub_vv_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vfmsub_vv_f16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vfmsub_vv_f16mf4((vfloat16mf4_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vfmsub_vv_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsub_vv_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vfmsub_vf_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vfmsub_vf_f16m1((vfloat16m1_t)(op0), (_Float16)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfmsub_vf_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsub_vf_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (_Float16)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfmsub_vf_f16m2(op0, op1, op2, op3) \ +__builtin_rvv_vfmsub_vf_f16m2((vfloat16m2_t)(op0), (_Float16)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vfmsub_vf_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsub_vf_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (_Float16)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vfmsub_vf_f16m4(op0, op1, op2, op3) \ +__builtin_rvv_vfmsub_vf_f16m4((vfloat16m4_t)(op0), (_Float16)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vfmsub_vf_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsub_vf_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (_Float16)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vfmsub_vf_f16m8(op0, op1, op2, op3) \ +__builtin_rvv_vfmsub_vf_f16m8((vfloat16m8_t)(op0), (_Float16)(op1), (vfloat16m8_t)(op2), (size_t)(op3)) +#define vfmsub_vf_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsub_vf_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (_Float16)(op2), (vfloat16m8_t)(op3), (size_t)(op4)) +#define vfmsub_vf_f16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vfmsub_vf_f16mf2((vfloat16mf2_t)(op0), (_Float16)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vfmsub_vf_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsub_vf_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (_Float16)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vfmsub_vf_f16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vfmsub_vf_f16mf4((vfloat16mf4_t)(op0), (_Float16)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vfmsub_vf_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmsub_vf_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (_Float16)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vfnmsub_vv_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsub_vv_f16m1((vfloat16m1_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfnmsub_vv_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsub_vv_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfnmsub_vv_f16m2(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsub_vv_f16m2((vfloat16m2_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vfnmsub_vv_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsub_vv_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vfnmsub_vv_f16m4(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsub_vv_f16m4((vfloat16m4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vfnmsub_vv_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsub_vv_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vfnmsub_vv_f16m8(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsub_vv_f16m8((vfloat16m8_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (size_t)(op3)) +#define vfnmsub_vv_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsub_vv_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (vfloat16m8_t)(op3), (size_t)(op4)) +#define vfnmsub_vv_f16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsub_vv_f16mf2((vfloat16mf2_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vfnmsub_vv_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsub_vv_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vfnmsub_vv_f16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsub_vv_f16mf4((vfloat16mf4_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vfnmsub_vv_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsub_vv_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vfnmsub_vf_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsub_vf_f16m1((vfloat16m1_t)(op0), (_Float16)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfnmsub_vf_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsub_vf_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (_Float16)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfnmsub_vf_f16m2(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsub_vf_f16m2((vfloat16m2_t)(op0), (_Float16)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vfnmsub_vf_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsub_vf_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (_Float16)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vfnmsub_vf_f16m4(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsub_vf_f16m4((vfloat16m4_t)(op0), (_Float16)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vfnmsub_vf_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsub_vf_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (_Float16)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vfnmsub_vf_f16m8(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsub_vf_f16m8((vfloat16m8_t)(op0), (_Float16)(op1), (vfloat16m8_t)(op2), (size_t)(op3)) +#define vfnmsub_vf_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsub_vf_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (_Float16)(op2), (vfloat16m8_t)(op3), (size_t)(op4)) +#define vfnmsub_vf_f16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsub_vf_f16mf2((vfloat16mf2_t)(op0), (_Float16)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vfnmsub_vf_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsub_vf_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (_Float16)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vfnmsub_vf_f16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vfnmsub_vf_f16mf4((vfloat16mf4_t)(op0), (_Float16)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vfnmsub_vf_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfnmsub_vf_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (_Float16)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vfmin_vv_f16m1(op0, op1, op2) \ +__builtin_rvv_vfmin_vv_f16m1((vfloat16m1_t)(op0), (vfloat16m1_t)(op1), (size_t)(op2)) +#define vfmin_vv_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmin_vv_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfmin_vv_f16m2(op0, op1, op2) \ +__builtin_rvv_vfmin_vv_f16m2((vfloat16m2_t)(op0), (vfloat16m2_t)(op1), (size_t)(op2)) +#define vfmin_vv_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmin_vv_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vfmin_vv_f16m4(op0, op1, op2) \ +__builtin_rvv_vfmin_vv_f16m4((vfloat16m4_t)(op0), (vfloat16m4_t)(op1), (size_t)(op2)) +#define vfmin_vv_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmin_vv_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vfmin_vv_f16m8(op0, op1, op2) \ +__builtin_rvv_vfmin_vv_f16m8((vfloat16m8_t)(op0), (vfloat16m8_t)(op1), (size_t)(op2)) +#define vfmin_vv_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmin_vv_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (vfloat16m8_t)(op3), (size_t)(op4)) +#define vfmin_vv_f16mf2(op0, op1, op2) \ +__builtin_rvv_vfmin_vv_f16mf2((vfloat16mf2_t)(op0), (vfloat16mf2_t)(op1), (size_t)(op2)) +#define vfmin_vv_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmin_vv_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vfmin_vv_f16mf4(op0, op1, op2) \ +__builtin_rvv_vfmin_vv_f16mf4((vfloat16mf4_t)(op0), (vfloat16mf4_t)(op1), (size_t)(op2)) +#define vfmin_vv_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmin_vv_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vfmin_vf_f16m1(op0, op1, op2) \ +__builtin_rvv_vfmin_vf_f16m1((vfloat16m1_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfmin_vf_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmin_vf_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfmin_vf_f16m2(op0, op1, op2) \ +__builtin_rvv_vfmin_vf_f16m2((vfloat16m2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfmin_vf_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmin_vf_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfmin_vf_f16m4(op0, op1, op2) \ +__builtin_rvv_vfmin_vf_f16m4((vfloat16m4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfmin_vf_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmin_vf_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfmin_vf_f16m8(op0, op1, op2) \ +__builtin_rvv_vfmin_vf_f16m8((vfloat16m8_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfmin_vf_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmin_vf_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfmin_vf_f16mf2(op0, op1, op2) \ +__builtin_rvv_vfmin_vf_f16mf2((vfloat16mf2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfmin_vf_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmin_vf_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfmin_vf_f16mf4(op0, op1, op2) \ +__builtin_rvv_vfmin_vf_f16mf4((vfloat16mf4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfmin_vf_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmin_vf_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfmax_vv_f16m1(op0, op1, op2) \ +__builtin_rvv_vfmax_vv_f16m1((vfloat16m1_t)(op0), (vfloat16m1_t)(op1), (size_t)(op2)) +#define vfmax_vv_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmax_vv_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfmax_vv_f16m2(op0, op1, op2) \ +__builtin_rvv_vfmax_vv_f16m2((vfloat16m2_t)(op0), (vfloat16m2_t)(op1), (size_t)(op2)) +#define vfmax_vv_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmax_vv_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vfmax_vv_f16m4(op0, op1, op2) \ +__builtin_rvv_vfmax_vv_f16m4((vfloat16m4_t)(op0), (vfloat16m4_t)(op1), (size_t)(op2)) +#define vfmax_vv_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmax_vv_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vfmax_vv_f16m8(op0, op1, op2) \ +__builtin_rvv_vfmax_vv_f16m8((vfloat16m8_t)(op0), (vfloat16m8_t)(op1), (size_t)(op2)) +#define vfmax_vv_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmax_vv_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (vfloat16m8_t)(op3), (size_t)(op4)) +#define vfmax_vv_f16mf2(op0, op1, op2) \ +__builtin_rvv_vfmax_vv_f16mf2((vfloat16mf2_t)(op0), (vfloat16mf2_t)(op1), (size_t)(op2)) +#define vfmax_vv_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmax_vv_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vfmax_vv_f16mf4(op0, op1, op2) \ +__builtin_rvv_vfmax_vv_f16mf4((vfloat16mf4_t)(op0), (vfloat16mf4_t)(op1), (size_t)(op2)) +#define vfmax_vv_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmax_vv_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vfmax_vf_f16m1(op0, op1, op2) \ +__builtin_rvv_vfmax_vf_f16m1((vfloat16m1_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfmax_vf_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmax_vf_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfmax_vf_f16m2(op0, op1, op2) \ +__builtin_rvv_vfmax_vf_f16m2((vfloat16m2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfmax_vf_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmax_vf_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfmax_vf_f16m4(op0, op1, op2) \ +__builtin_rvv_vfmax_vf_f16m4((vfloat16m4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfmax_vf_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmax_vf_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfmax_vf_f16m8(op0, op1, op2) \ +__builtin_rvv_vfmax_vf_f16m8((vfloat16m8_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfmax_vf_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmax_vf_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfmax_vf_f16mf2(op0, op1, op2) \ +__builtin_rvv_vfmax_vf_f16mf2((vfloat16mf2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfmax_vf_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmax_vf_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfmax_vf_f16mf4(op0, op1, op2) \ +__builtin_rvv_vfmax_vf_f16mf4((vfloat16mf4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfmax_vf_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfmax_vf_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfsgnj_vv_f16m1(op0, op1, op2) \ +__builtin_rvv_vfsgnj_vv_f16m1((vfloat16m1_t)(op0), (vfloat16m1_t)(op1), (size_t)(op2)) +#define vfsgnj_vv_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnj_vv_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfsgnj_vv_f16m2(op0, op1, op2) \ +__builtin_rvv_vfsgnj_vv_f16m2((vfloat16m2_t)(op0), (vfloat16m2_t)(op1), (size_t)(op2)) +#define vfsgnj_vv_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnj_vv_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vfsgnj_vv_f16m4(op0, op1, op2) \ +__builtin_rvv_vfsgnj_vv_f16m4((vfloat16m4_t)(op0), (vfloat16m4_t)(op1), (size_t)(op2)) +#define vfsgnj_vv_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnj_vv_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vfsgnj_vv_f16m8(op0, op1, op2) \ +__builtin_rvv_vfsgnj_vv_f16m8((vfloat16m8_t)(op0), (vfloat16m8_t)(op1), (size_t)(op2)) +#define vfsgnj_vv_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnj_vv_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (vfloat16m8_t)(op3), (size_t)(op4)) +#define vfsgnj_vv_f16mf2(op0, op1, op2) \ +__builtin_rvv_vfsgnj_vv_f16mf2((vfloat16mf2_t)(op0), (vfloat16mf2_t)(op1), (size_t)(op2)) +#define vfsgnj_vv_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnj_vv_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vfsgnj_vv_f16mf4(op0, op1, op2) \ +__builtin_rvv_vfsgnj_vv_f16mf4((vfloat16mf4_t)(op0), (vfloat16mf4_t)(op1), (size_t)(op2)) +#define vfsgnj_vv_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnj_vv_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vfsgnj_vf_f16m1(op0, op1, op2) \ +__builtin_rvv_vfsgnj_vf_f16m1((vfloat16m1_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfsgnj_vf_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnj_vf_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfsgnj_vf_f16m2(op0, op1, op2) \ +__builtin_rvv_vfsgnj_vf_f16m2((vfloat16m2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfsgnj_vf_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnj_vf_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfsgnj_vf_f16m4(op0, op1, op2) \ +__builtin_rvv_vfsgnj_vf_f16m4((vfloat16m4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfsgnj_vf_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnj_vf_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfsgnj_vf_f16m8(op0, op1, op2) \ +__builtin_rvv_vfsgnj_vf_f16m8((vfloat16m8_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfsgnj_vf_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnj_vf_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfsgnj_vf_f16mf2(op0, op1, op2) \ +__builtin_rvv_vfsgnj_vf_f16mf2((vfloat16mf2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfsgnj_vf_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnj_vf_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfsgnj_vf_f16mf4(op0, op1, op2) \ +__builtin_rvv_vfsgnj_vf_f16mf4((vfloat16mf4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfsgnj_vf_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnj_vf_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfsgnjn_vv_f16m1(op0, op1, op2) \ +__builtin_rvv_vfsgnjn_vv_f16m1((vfloat16m1_t)(op0), (vfloat16m1_t)(op1), (size_t)(op2)) +#define vfsgnjn_vv_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjn_vv_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfsgnjn_vv_f16m2(op0, op1, op2) \ +__builtin_rvv_vfsgnjn_vv_f16m2((vfloat16m2_t)(op0), (vfloat16m2_t)(op1), (size_t)(op2)) +#define vfsgnjn_vv_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjn_vv_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vfsgnjn_vv_f16m4(op0, op1, op2) \ +__builtin_rvv_vfsgnjn_vv_f16m4((vfloat16m4_t)(op0), (vfloat16m4_t)(op1), (size_t)(op2)) +#define vfsgnjn_vv_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjn_vv_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vfsgnjn_vv_f16m8(op0, op1, op2) \ +__builtin_rvv_vfsgnjn_vv_f16m8((vfloat16m8_t)(op0), (vfloat16m8_t)(op1), (size_t)(op2)) +#define vfsgnjn_vv_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjn_vv_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (vfloat16m8_t)(op3), (size_t)(op4)) +#define vfsgnjn_vv_f16mf2(op0, op1, op2) \ +__builtin_rvv_vfsgnjn_vv_f16mf2((vfloat16mf2_t)(op0), (vfloat16mf2_t)(op1), (size_t)(op2)) +#define vfsgnjn_vv_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjn_vv_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vfsgnjn_vv_f16mf4(op0, op1, op2) \ +__builtin_rvv_vfsgnjn_vv_f16mf4((vfloat16mf4_t)(op0), (vfloat16mf4_t)(op1), (size_t)(op2)) +#define vfsgnjn_vv_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjn_vv_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vfsgnjn_vf_f16m1(op0, op1, op2) \ +__builtin_rvv_vfsgnjn_vf_f16m1((vfloat16m1_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfsgnjn_vf_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjn_vf_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfsgnjn_vf_f16m2(op0, op1, op2) \ +__builtin_rvv_vfsgnjn_vf_f16m2((vfloat16m2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfsgnjn_vf_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjn_vf_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfsgnjn_vf_f16m4(op0, op1, op2) \ +__builtin_rvv_vfsgnjn_vf_f16m4((vfloat16m4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfsgnjn_vf_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjn_vf_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfsgnjn_vf_f16m8(op0, op1, op2) \ +__builtin_rvv_vfsgnjn_vf_f16m8((vfloat16m8_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfsgnjn_vf_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjn_vf_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfsgnjn_vf_f16mf2(op0, op1, op2) \ +__builtin_rvv_vfsgnjn_vf_f16mf2((vfloat16mf2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfsgnjn_vf_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjn_vf_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfsgnjn_vf_f16mf4(op0, op1, op2) \ +__builtin_rvv_vfsgnjn_vf_f16mf4((vfloat16mf4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfsgnjn_vf_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjn_vf_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfsgnjx_vv_f16m1(op0, op1, op2) \ +__builtin_rvv_vfsgnjx_vv_f16m1((vfloat16m1_t)(op0), (vfloat16m1_t)(op1), (size_t)(op2)) +#define vfsgnjx_vv_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjx_vv_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfsgnjx_vv_f16m2(op0, op1, op2) \ +__builtin_rvv_vfsgnjx_vv_f16m2((vfloat16m2_t)(op0), (vfloat16m2_t)(op1), (size_t)(op2)) +#define vfsgnjx_vv_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjx_vv_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vfsgnjx_vv_f16m4(op0, op1, op2) \ +__builtin_rvv_vfsgnjx_vv_f16m4((vfloat16m4_t)(op0), (vfloat16m4_t)(op1), (size_t)(op2)) +#define vfsgnjx_vv_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjx_vv_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vfsgnjx_vv_f16m8(op0, op1, op2) \ +__builtin_rvv_vfsgnjx_vv_f16m8((vfloat16m8_t)(op0), (vfloat16m8_t)(op1), (size_t)(op2)) +#define vfsgnjx_vv_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjx_vv_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (vfloat16m8_t)(op3), (size_t)(op4)) +#define vfsgnjx_vv_f16mf2(op0, op1, op2) \ +__builtin_rvv_vfsgnjx_vv_f16mf2((vfloat16mf2_t)(op0), (vfloat16mf2_t)(op1), (size_t)(op2)) +#define vfsgnjx_vv_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjx_vv_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vfsgnjx_vv_f16mf4(op0, op1, op2) \ +__builtin_rvv_vfsgnjx_vv_f16mf4((vfloat16mf4_t)(op0), (vfloat16mf4_t)(op1), (size_t)(op2)) +#define vfsgnjx_vv_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjx_vv_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vfsgnjx_vf_f16m1(op0, op1, op2) \ +__builtin_rvv_vfsgnjx_vf_f16m1((vfloat16m1_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfsgnjx_vf_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjx_vf_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfsgnjx_vf_f16m2(op0, op1, op2) \ +__builtin_rvv_vfsgnjx_vf_f16m2((vfloat16m2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfsgnjx_vf_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjx_vf_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfsgnjx_vf_f16m4(op0, op1, op2) \ +__builtin_rvv_vfsgnjx_vf_f16m4((vfloat16m4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfsgnjx_vf_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjx_vf_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfsgnjx_vf_f16m8(op0, op1, op2) \ +__builtin_rvv_vfsgnjx_vf_f16m8((vfloat16m8_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfsgnjx_vf_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjx_vf_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfsgnjx_vf_f16mf2(op0, op1, op2) \ +__builtin_rvv_vfsgnjx_vf_f16mf2((vfloat16mf2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfsgnjx_vf_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjx_vf_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfsgnjx_vf_f16mf4(op0, op1, op2) \ +__builtin_rvv_vfsgnjx_vf_f16mf4((vfloat16mf4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfsgnjx_vf_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfsgnjx_vf_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfabs_v_f16m1(op0, op1) \ +__builtin_rvv_vfabs_v_f16m1((vfloat16m1_t)(op0), (size_t)(op1)) +#define vfabs_v_f16m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfabs_v_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfabs_v_f16m2(op0, op1) \ +__builtin_rvv_vfabs_v_f16m2((vfloat16m2_t)(op0), (size_t)(op1)) +#define vfabs_v_f16m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfabs_v_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vfabs_v_f16m4(op0, op1) \ +__builtin_rvv_vfabs_v_f16m4((vfloat16m4_t)(op0), (size_t)(op1)) +#define vfabs_v_f16m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfabs_v_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vfabs_v_f16m8(op0, op1) \ +__builtin_rvv_vfabs_v_f16m8((vfloat16m8_t)(op0), (size_t)(op1)) +#define vfabs_v_f16m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfabs_v_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (size_t)(op3)) +#define vfabs_v_f16mf2(op0, op1) \ +__builtin_rvv_vfabs_v_f16mf2((vfloat16mf2_t)(op0), (size_t)(op1)) +#define vfabs_v_f16mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfabs_v_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vfabs_v_f16mf4(op0, op1) \ +__builtin_rvv_vfabs_v_f16mf4((vfloat16mf4_t)(op0), (size_t)(op1)) +#define vfabs_v_f16mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfabs_v_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vmfeq_vv_f16m1_b16(op0, op1, op2) \ +__builtin_rvv_vmfeq_vv_f16m1_b16((vfloat16m1_t)(op0), (vfloat16m1_t)(op1), (size_t)(op2)) +#define vmfeq_vv_f16m1_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfeq_vv_f16m1_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vmfeq_vv_f16m2_b8(op0, op1, op2) \ +__builtin_rvv_vmfeq_vv_f16m2_b8((vfloat16m2_t)(op0), (vfloat16m2_t)(op1), (size_t)(op2)) +#define vmfeq_vv_f16m2_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfeq_vv_f16m2_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vmfeq_vv_f16m4_b4(op0, op1, op2) \ +__builtin_rvv_vmfeq_vv_f16m4_b4((vfloat16m4_t)(op0), (vfloat16m4_t)(op1), (size_t)(op2)) +#define vmfeq_vv_f16m4_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfeq_vv_f16m4_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vfloat16m4_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vmfeq_vv_f16m8_b2(op0, op1, op2) \ +__builtin_rvv_vmfeq_vv_f16m8_b2((vfloat16m8_t)(op0), (vfloat16m8_t)(op1), (size_t)(op2)) +#define vmfeq_vv_f16m8_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfeq_vv_f16m8_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vfloat16m8_t)(op2), (vfloat16m8_t)(op3), (size_t)(op4)) +#define vmfeq_vv_f16mf2_b32(op0, op1, op2) \ +__builtin_rvv_vmfeq_vv_f16mf2_b32((vfloat16mf2_t)(op0), (vfloat16mf2_t)(op1), (size_t)(op2)) +#define vmfeq_vv_f16mf2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfeq_vv_f16mf2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vmfeq_vv_f16mf4_b64(op0, op1, op2) \ +__builtin_rvv_vmfeq_vv_f16mf4_b64((vfloat16mf4_t)(op0), (vfloat16mf4_t)(op1), (size_t)(op2)) +#define vmfeq_vv_f16mf4_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfeq_vv_f16mf4_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vmfeq_vf_f16m1_b16(op0, op1, op2) \ +__builtin_rvv_vmfeq_vf_f16m1_b16((vfloat16m1_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vmfeq_vf_f16m1_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfeq_vf_f16m1_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vfloat16m1_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vmfeq_vf_f16m2_b8(op0, op1, op2) \ +__builtin_rvv_vmfeq_vf_f16m2_b8((vfloat16m2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vmfeq_vf_f16m2_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfeq_vf_f16m2_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vfloat16m2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vmfeq_vf_f16m4_b4(op0, op1, op2) \ +__builtin_rvv_vmfeq_vf_f16m4_b4((vfloat16m4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vmfeq_vf_f16m4_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfeq_vf_f16m4_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vfloat16m4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vmfeq_vf_f16m8_b2(op0, op1, op2) \ +__builtin_rvv_vmfeq_vf_f16m8_b2((vfloat16m8_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vmfeq_vf_f16m8_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfeq_vf_f16m8_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vfloat16m8_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vmfeq_vf_f16mf2_b32(op0, op1, op2) \ +__builtin_rvv_vmfeq_vf_f16mf2_b32((vfloat16mf2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vmfeq_vf_f16mf2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfeq_vf_f16mf2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vfloat16mf2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vmfeq_vf_f16mf4_b64(op0, op1, op2) \ +__builtin_rvv_vmfeq_vf_f16mf4_b64((vfloat16mf4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vmfeq_vf_f16mf4_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfeq_vf_f16mf4_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vfloat16mf4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vmfne_vv_f16m1_b16(op0, op1, op2) \ +__builtin_rvv_vmfne_vv_f16m1_b16((vfloat16m1_t)(op0), (vfloat16m1_t)(op1), (size_t)(op2)) +#define vmfne_vv_f16m1_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfne_vv_f16m1_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vmfne_vv_f16m2_b8(op0, op1, op2) \ +__builtin_rvv_vmfne_vv_f16m2_b8((vfloat16m2_t)(op0), (vfloat16m2_t)(op1), (size_t)(op2)) +#define vmfne_vv_f16m2_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfne_vv_f16m2_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vmfne_vv_f16m4_b4(op0, op1, op2) \ +__builtin_rvv_vmfne_vv_f16m4_b4((vfloat16m4_t)(op0), (vfloat16m4_t)(op1), (size_t)(op2)) +#define vmfne_vv_f16m4_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfne_vv_f16m4_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vfloat16m4_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vmfne_vv_f16m8_b2(op0, op1, op2) \ +__builtin_rvv_vmfne_vv_f16m8_b2((vfloat16m8_t)(op0), (vfloat16m8_t)(op1), (size_t)(op2)) +#define vmfne_vv_f16m8_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfne_vv_f16m8_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vfloat16m8_t)(op2), (vfloat16m8_t)(op3), (size_t)(op4)) +#define vmfne_vv_f16mf2_b32(op0, op1, op2) \ +__builtin_rvv_vmfne_vv_f16mf2_b32((vfloat16mf2_t)(op0), (vfloat16mf2_t)(op1), (size_t)(op2)) +#define vmfne_vv_f16mf2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfne_vv_f16mf2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vmfne_vv_f16mf4_b64(op0, op1, op2) \ +__builtin_rvv_vmfne_vv_f16mf4_b64((vfloat16mf4_t)(op0), (vfloat16mf4_t)(op1), (size_t)(op2)) +#define vmfne_vv_f16mf4_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfne_vv_f16mf4_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vmfne_vf_f16m1_b16(op0, op1, op2) \ +__builtin_rvv_vmfne_vf_f16m1_b16((vfloat16m1_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vmfne_vf_f16m1_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfne_vf_f16m1_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vfloat16m1_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vmfne_vf_f16m2_b8(op0, op1, op2) \ +__builtin_rvv_vmfne_vf_f16m2_b8((vfloat16m2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vmfne_vf_f16m2_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfne_vf_f16m2_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vfloat16m2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vmfne_vf_f16m4_b4(op0, op1, op2) \ +__builtin_rvv_vmfne_vf_f16m4_b4((vfloat16m4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vmfne_vf_f16m4_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfne_vf_f16m4_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vfloat16m4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vmfne_vf_f16m8_b2(op0, op1, op2) \ +__builtin_rvv_vmfne_vf_f16m8_b2((vfloat16m8_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vmfne_vf_f16m8_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfne_vf_f16m8_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vfloat16m8_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vmfne_vf_f16mf2_b32(op0, op1, op2) \ +__builtin_rvv_vmfne_vf_f16mf2_b32((vfloat16mf2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vmfne_vf_f16mf2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfne_vf_f16mf2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vfloat16mf2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vmfne_vf_f16mf4_b64(op0, op1, op2) \ +__builtin_rvv_vmfne_vf_f16mf4_b64((vfloat16mf4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vmfne_vf_f16mf4_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfne_vf_f16mf4_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vfloat16mf4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vmflt_vv_f16m1_b16(op0, op1, op2) \ +__builtin_rvv_vmflt_vv_f16m1_b16((vfloat16m1_t)(op0), (vfloat16m1_t)(op1), (size_t)(op2)) +#define vmflt_vv_f16m1_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmflt_vv_f16m1_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vmflt_vv_f16m2_b8(op0, op1, op2) \ +__builtin_rvv_vmflt_vv_f16m2_b8((vfloat16m2_t)(op0), (vfloat16m2_t)(op1), (size_t)(op2)) +#define vmflt_vv_f16m2_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmflt_vv_f16m2_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vmflt_vv_f16m4_b4(op0, op1, op2) \ +__builtin_rvv_vmflt_vv_f16m4_b4((vfloat16m4_t)(op0), (vfloat16m4_t)(op1), (size_t)(op2)) +#define vmflt_vv_f16m4_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmflt_vv_f16m4_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vfloat16m4_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vmflt_vv_f16m8_b2(op0, op1, op2) \ +__builtin_rvv_vmflt_vv_f16m8_b2((vfloat16m8_t)(op0), (vfloat16m8_t)(op1), (size_t)(op2)) +#define vmflt_vv_f16m8_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmflt_vv_f16m8_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vfloat16m8_t)(op2), (vfloat16m8_t)(op3), (size_t)(op4)) +#define vmflt_vv_f16mf2_b32(op0, op1, op2) \ +__builtin_rvv_vmflt_vv_f16mf2_b32((vfloat16mf2_t)(op0), (vfloat16mf2_t)(op1), (size_t)(op2)) +#define vmflt_vv_f16mf2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmflt_vv_f16mf2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vmflt_vv_f16mf4_b64(op0, op1, op2) \ +__builtin_rvv_vmflt_vv_f16mf4_b64((vfloat16mf4_t)(op0), (vfloat16mf4_t)(op1), (size_t)(op2)) +#define vmflt_vv_f16mf4_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmflt_vv_f16mf4_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vmflt_vf_f16m1_b16(op0, op1, op2) \ +__builtin_rvv_vmflt_vf_f16m1_b16((vfloat16m1_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vmflt_vf_f16m1_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmflt_vf_f16m1_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vfloat16m1_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vmflt_vf_f16m2_b8(op0, op1, op2) \ +__builtin_rvv_vmflt_vf_f16m2_b8((vfloat16m2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vmflt_vf_f16m2_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmflt_vf_f16m2_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vfloat16m2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vmflt_vf_f16m4_b4(op0, op1, op2) \ +__builtin_rvv_vmflt_vf_f16m4_b4((vfloat16m4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vmflt_vf_f16m4_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmflt_vf_f16m4_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vfloat16m4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vmflt_vf_f16m8_b2(op0, op1, op2) \ +__builtin_rvv_vmflt_vf_f16m8_b2((vfloat16m8_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vmflt_vf_f16m8_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmflt_vf_f16m8_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vfloat16m8_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vmflt_vf_f16mf2_b32(op0, op1, op2) \ +__builtin_rvv_vmflt_vf_f16mf2_b32((vfloat16mf2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vmflt_vf_f16mf2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmflt_vf_f16mf2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vfloat16mf2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vmflt_vf_f16mf4_b64(op0, op1, op2) \ +__builtin_rvv_vmflt_vf_f16mf4_b64((vfloat16mf4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vmflt_vf_f16mf4_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmflt_vf_f16mf4_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vfloat16mf4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vmfle_vv_f16m1_b16(op0, op1, op2) \ +__builtin_rvv_vmfle_vv_f16m1_b16((vfloat16m1_t)(op0), (vfloat16m1_t)(op1), (size_t)(op2)) +#define vmfle_vv_f16m1_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfle_vv_f16m1_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vmfle_vv_f16m2_b8(op0, op1, op2) \ +__builtin_rvv_vmfle_vv_f16m2_b8((vfloat16m2_t)(op0), (vfloat16m2_t)(op1), (size_t)(op2)) +#define vmfle_vv_f16m2_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfle_vv_f16m2_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vmfle_vv_f16m4_b4(op0, op1, op2) \ +__builtin_rvv_vmfle_vv_f16m4_b4((vfloat16m4_t)(op0), (vfloat16m4_t)(op1), (size_t)(op2)) +#define vmfle_vv_f16m4_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfle_vv_f16m4_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vfloat16m4_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vmfle_vv_f16m8_b2(op0, op1, op2) \ +__builtin_rvv_vmfle_vv_f16m8_b2((vfloat16m8_t)(op0), (vfloat16m8_t)(op1), (size_t)(op2)) +#define vmfle_vv_f16m8_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfle_vv_f16m8_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vfloat16m8_t)(op2), (vfloat16m8_t)(op3), (size_t)(op4)) +#define vmfle_vv_f16mf2_b32(op0, op1, op2) \ +__builtin_rvv_vmfle_vv_f16mf2_b32((vfloat16mf2_t)(op0), (vfloat16mf2_t)(op1), (size_t)(op2)) +#define vmfle_vv_f16mf2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfle_vv_f16mf2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vmfle_vv_f16mf4_b64(op0, op1, op2) \ +__builtin_rvv_vmfle_vv_f16mf4_b64((vfloat16mf4_t)(op0), (vfloat16mf4_t)(op1), (size_t)(op2)) +#define vmfle_vv_f16mf4_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfle_vv_f16mf4_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vmfle_vf_f16m1_b16(op0, op1, op2) \ +__builtin_rvv_vmfle_vf_f16m1_b16((vfloat16m1_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vmfle_vf_f16m1_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfle_vf_f16m1_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vfloat16m1_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vmfle_vf_f16m2_b8(op0, op1, op2) \ +__builtin_rvv_vmfle_vf_f16m2_b8((vfloat16m2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vmfle_vf_f16m2_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfle_vf_f16m2_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vfloat16m2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vmfle_vf_f16m4_b4(op0, op1, op2) \ +__builtin_rvv_vmfle_vf_f16m4_b4((vfloat16m4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vmfle_vf_f16m4_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfle_vf_f16m4_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vfloat16m4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vmfle_vf_f16m8_b2(op0, op1, op2) \ +__builtin_rvv_vmfle_vf_f16m8_b2((vfloat16m8_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vmfle_vf_f16m8_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfle_vf_f16m8_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vfloat16m8_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vmfle_vf_f16mf2_b32(op0, op1, op2) \ +__builtin_rvv_vmfle_vf_f16mf2_b32((vfloat16mf2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vmfle_vf_f16mf2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfle_vf_f16mf2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vfloat16mf2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vmfle_vf_f16mf4_b64(op0, op1, op2) \ +__builtin_rvv_vmfle_vf_f16mf4_b64((vfloat16mf4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vmfle_vf_f16mf4_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfle_vf_f16mf4_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vfloat16mf4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vmfgt_vv_f16m1_b16(op0, op1, op2) \ +__builtin_rvv_vmfgt_vv_f16m1_b16((vfloat16m1_t)(op0), (vfloat16m1_t)(op1), (size_t)(op2)) +#define vmfgt_vv_f16m1_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfgt_vv_f16m1_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vmfgt_vv_f16m2_b8(op0, op1, op2) \ +__builtin_rvv_vmfgt_vv_f16m2_b8((vfloat16m2_t)(op0), (vfloat16m2_t)(op1), (size_t)(op2)) +#define vmfgt_vv_f16m2_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfgt_vv_f16m2_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vmfgt_vv_f16m4_b4(op0, op1, op2) \ +__builtin_rvv_vmfgt_vv_f16m4_b4((vfloat16m4_t)(op0), (vfloat16m4_t)(op1), (size_t)(op2)) +#define vmfgt_vv_f16m4_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfgt_vv_f16m4_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vfloat16m4_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vmfgt_vv_f16m8_b2(op0, op1, op2) \ +__builtin_rvv_vmfgt_vv_f16m8_b2((vfloat16m8_t)(op0), (vfloat16m8_t)(op1), (size_t)(op2)) +#define vmfgt_vv_f16m8_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfgt_vv_f16m8_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vfloat16m8_t)(op2), (vfloat16m8_t)(op3), (size_t)(op4)) +#define vmfgt_vv_f16mf2_b32(op0, op1, op2) \ +__builtin_rvv_vmfgt_vv_f16mf2_b32((vfloat16mf2_t)(op0), (vfloat16mf2_t)(op1), (size_t)(op2)) +#define vmfgt_vv_f16mf2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfgt_vv_f16mf2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vmfgt_vv_f16mf4_b64(op0, op1, op2) \ +__builtin_rvv_vmfgt_vv_f16mf4_b64((vfloat16mf4_t)(op0), (vfloat16mf4_t)(op1), (size_t)(op2)) +#define vmfgt_vv_f16mf4_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfgt_vv_f16mf4_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vmfgt_vf_f16m1_b16(op0, op1, op2) \ +__builtin_rvv_vmfgt_vf_f16m1_b16((vfloat16m1_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vmfgt_vf_f16m1_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfgt_vf_f16m1_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vfloat16m1_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vmfgt_vf_f16m2_b8(op0, op1, op2) \ +__builtin_rvv_vmfgt_vf_f16m2_b8((vfloat16m2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vmfgt_vf_f16m2_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfgt_vf_f16m2_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vfloat16m2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vmfgt_vf_f16m4_b4(op0, op1, op2) \ +__builtin_rvv_vmfgt_vf_f16m4_b4((vfloat16m4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vmfgt_vf_f16m4_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfgt_vf_f16m4_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vfloat16m4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vmfgt_vf_f16m8_b2(op0, op1, op2) \ +__builtin_rvv_vmfgt_vf_f16m8_b2((vfloat16m8_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vmfgt_vf_f16m8_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfgt_vf_f16m8_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vfloat16m8_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vmfgt_vf_f16mf2_b32(op0, op1, op2) \ +__builtin_rvv_vmfgt_vf_f16mf2_b32((vfloat16mf2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vmfgt_vf_f16mf2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfgt_vf_f16mf2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vfloat16mf2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vmfgt_vf_f16mf4_b64(op0, op1, op2) \ +__builtin_rvv_vmfgt_vf_f16mf4_b64((vfloat16mf4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vmfgt_vf_f16mf4_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfgt_vf_f16mf4_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vfloat16mf4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vmfge_vv_f16m1_b16(op0, op1, op2) \ +__builtin_rvv_vmfge_vv_f16m1_b16((vfloat16m1_t)(op0), (vfloat16m1_t)(op1), (size_t)(op2)) +#define vmfge_vv_f16m1_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfge_vv_f16m1_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vmfge_vv_f16m2_b8(op0, op1, op2) \ +__builtin_rvv_vmfge_vv_f16m2_b8((vfloat16m2_t)(op0), (vfloat16m2_t)(op1), (size_t)(op2)) +#define vmfge_vv_f16m2_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfge_vv_f16m2_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vmfge_vv_f16m4_b4(op0, op1, op2) \ +__builtin_rvv_vmfge_vv_f16m4_b4((vfloat16m4_t)(op0), (vfloat16m4_t)(op1), (size_t)(op2)) +#define vmfge_vv_f16m4_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfge_vv_f16m4_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vfloat16m4_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vmfge_vv_f16m8_b2(op0, op1, op2) \ +__builtin_rvv_vmfge_vv_f16m8_b2((vfloat16m8_t)(op0), (vfloat16m8_t)(op1), (size_t)(op2)) +#define vmfge_vv_f16m8_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfge_vv_f16m8_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vfloat16m8_t)(op2), (vfloat16m8_t)(op3), (size_t)(op4)) +#define vmfge_vv_f16mf2_b32(op0, op1, op2) \ +__builtin_rvv_vmfge_vv_f16mf2_b32((vfloat16mf2_t)(op0), (vfloat16mf2_t)(op1), (size_t)(op2)) +#define vmfge_vv_f16mf2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfge_vv_f16mf2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vmfge_vv_f16mf4_b64(op0, op1, op2) \ +__builtin_rvv_vmfge_vv_f16mf4_b64((vfloat16mf4_t)(op0), (vfloat16mf4_t)(op1), (size_t)(op2)) +#define vmfge_vv_f16mf4_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfge_vv_f16mf4_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vmfge_vf_f16m1_b16(op0, op1, op2) \ +__builtin_rvv_vmfge_vf_f16m1_b16((vfloat16m1_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vmfge_vf_f16m1_b16_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfge_vf_f16m1_b16_m((vbool16_t)(op0), (vbool16_t)(op1), (vfloat16m1_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vmfge_vf_f16m2_b8(op0, op1, op2) \ +__builtin_rvv_vmfge_vf_f16m2_b8((vfloat16m2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vmfge_vf_f16m2_b8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfge_vf_f16m2_b8_m((vbool8_t)(op0), (vbool8_t)(op1), (vfloat16m2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vmfge_vf_f16m4_b4(op0, op1, op2) \ +__builtin_rvv_vmfge_vf_f16m4_b4((vfloat16m4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vmfge_vf_f16m4_b4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfge_vf_f16m4_b4_m((vbool4_t)(op0), (vbool4_t)(op1), (vfloat16m4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vmfge_vf_f16m8_b2(op0, op1, op2) \ +__builtin_rvv_vmfge_vf_f16m8_b2((vfloat16m8_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vmfge_vf_f16m8_b2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfge_vf_f16m8_b2_m((vbool2_t)(op0), (vbool2_t)(op1), (vfloat16m8_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vmfge_vf_f16mf2_b32(op0, op1, op2) \ +__builtin_rvv_vmfge_vf_f16mf2_b32((vfloat16mf2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vmfge_vf_f16mf2_b32_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfge_vf_f16mf2_b32_m((vbool32_t)(op0), (vbool32_t)(op1), (vfloat16mf2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vmfge_vf_f16mf4_b64(op0, op1, op2) \ +__builtin_rvv_vmfge_vf_f16mf4_b64((vfloat16mf4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vmfge_vf_f16mf4_b64_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vmfge_vf_f16mf4_b64_m((vbool64_t)(op0), (vbool64_t)(op1), (vfloat16mf4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vmerge_vvm_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_f16m1((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vmerge_vvm_f16m2(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_f16m2((vbool8_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vmerge_vvm_f16m4(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_f16m4((vbool4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vmerge_vvm_f16m8(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_f16m8((vbool2_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (size_t)(op3)) +#define vmerge_vvm_f16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_f16mf2((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vmerge_vvm_f16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vmerge_vvm_f16mf4((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vfmerge_vfm_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vfmerge_vfm_f16m1((vbool16_t)(op0), (vfloat16m1_t)(op1), (_Float16)(op2), (size_t)(op3)) +#define vfmerge_vfm_f16m2(op0, op1, op2, op3) \ +__builtin_rvv_vfmerge_vfm_f16m2((vbool8_t)(op0), (vfloat16m2_t)(op1), (_Float16)(op2), (size_t)(op3)) +#define vfmerge_vfm_f16m4(op0, op1, op2, op3) \ +__builtin_rvv_vfmerge_vfm_f16m4((vbool4_t)(op0), (vfloat16m4_t)(op1), (_Float16)(op2), (size_t)(op3)) +#define vfmerge_vfm_f16m8(op0, op1, op2, op3) \ +__builtin_rvv_vfmerge_vfm_f16m8((vbool2_t)(op0), (vfloat16m8_t)(op1), (_Float16)(op2), (size_t)(op3)) +#define vfmerge_vfm_f16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vfmerge_vfm_f16mf2((vbool32_t)(op0), (vfloat16mf2_t)(op1), (_Float16)(op2), (size_t)(op3)) +#define vfmerge_vfm_f16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vfmerge_vfm_f16mf4((vbool64_t)(op0), (vfloat16mf4_t)(op1), (_Float16)(op2), (size_t)(op3)) +#define vfmv_v_f_f16m1(op0, op1) \ +__builtin_rvv_vfmv_v_f_f16m1((_Float16)(op0), (size_t)(op1)) +#define vfmv_v_f_f16m2(op0, op1) \ +__builtin_rvv_vfmv_v_f_f16m2((_Float16)(op0), (size_t)(op1)) +#define vfmv_v_f_f16m4(op0, op1) \ +__builtin_rvv_vfmv_v_f_f16m4((_Float16)(op0), (size_t)(op1)) +#define vfmv_v_f_f16m8(op0, op1) \ +__builtin_rvv_vfmv_v_f_f16m8((_Float16)(op0), (size_t)(op1)) +#define vfmv_v_f_f16mf2(op0, op1) \ +__builtin_rvv_vfmv_v_f_f16mf2((_Float16)(op0), (size_t)(op1)) +#define vfmv_v_f_f16mf4(op0, op1) \ +__builtin_rvv_vfmv_v_f_f16mf4((_Float16)(op0), (size_t)(op1)) +#define vfredmax_vs_f16m1_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredmax_vs_f16m1_f16m1((vfloat16m1_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfredmax_vs_f16m1_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredmax_vs_f16m1_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfredmax_vs_f16m2_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredmax_vs_f16m2_f16m1((vfloat16m1_t)(op0), (vfloat16m2_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfredmax_vs_f16m2_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredmax_vs_f16m2_f16m1_m((vbool8_t)(op0), (vfloat16m1_t)(op1), (vfloat16m2_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfredmax_vs_f16m4_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredmax_vs_f16m4_f16m1((vfloat16m1_t)(op0), (vfloat16m4_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfredmax_vs_f16m4_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredmax_vs_f16m4_f16m1_m((vbool4_t)(op0), (vfloat16m1_t)(op1), (vfloat16m4_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfredmax_vs_f16m8_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredmax_vs_f16m8_f16m1((vfloat16m1_t)(op0), (vfloat16m8_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfredmax_vs_f16m8_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredmax_vs_f16m8_f16m1_m((vbool2_t)(op0), (vfloat16m1_t)(op1), (vfloat16m8_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfredmax_vs_f16mf2_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredmax_vs_f16mf2_f16m1((vfloat16m1_t)(op0), (vfloat16mf2_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfredmax_vs_f16mf2_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredmax_vs_f16mf2_f16m1_m((vbool32_t)(op0), (vfloat16m1_t)(op1), (vfloat16mf2_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfredmax_vs_f16mf4_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredmax_vs_f16mf4_f16m1((vfloat16m1_t)(op0), (vfloat16mf4_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfredmax_vs_f16mf4_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredmax_vs_f16mf4_f16m1_m((vbool64_t)(op0), (vfloat16m1_t)(op1), (vfloat16mf4_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfredmin_vs_f16m1_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredmin_vs_f16m1_f16m1((vfloat16m1_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfredmin_vs_f16m1_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredmin_vs_f16m1_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfredmin_vs_f16m2_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredmin_vs_f16m2_f16m1((vfloat16m1_t)(op0), (vfloat16m2_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfredmin_vs_f16m2_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredmin_vs_f16m2_f16m1_m((vbool8_t)(op0), (vfloat16m1_t)(op1), (vfloat16m2_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfredmin_vs_f16m4_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredmin_vs_f16m4_f16m1((vfloat16m1_t)(op0), (vfloat16m4_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfredmin_vs_f16m4_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredmin_vs_f16m4_f16m1_m((vbool4_t)(op0), (vfloat16m1_t)(op1), (vfloat16m4_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfredmin_vs_f16m8_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredmin_vs_f16m8_f16m1((vfloat16m1_t)(op0), (vfloat16m8_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfredmin_vs_f16m8_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredmin_vs_f16m8_f16m1_m((vbool2_t)(op0), (vfloat16m1_t)(op1), (vfloat16m8_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfredmin_vs_f16mf2_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredmin_vs_f16mf2_f16m1((vfloat16m1_t)(op0), (vfloat16mf2_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfredmin_vs_f16mf2_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredmin_vs_f16mf2_f16m1_m((vbool32_t)(op0), (vfloat16m1_t)(op1), (vfloat16mf2_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfredmin_vs_f16mf4_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredmin_vs_f16mf4_f16m1((vfloat16m1_t)(op0), (vfloat16mf4_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfredmin_vs_f16mf4_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredmin_vs_f16mf4_f16m1_m((vbool64_t)(op0), (vfloat16m1_t)(op1), (vfloat16mf4_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfredsum_vs_f16m1_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredsum_vs_f16m1_f16m1((vfloat16m1_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfredsum_vs_f16m1_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredsum_vs_f16m1_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfredsum_vs_f16m2_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredsum_vs_f16m2_f16m1((vfloat16m1_t)(op0), (vfloat16m2_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfredsum_vs_f16m2_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredsum_vs_f16m2_f16m1_m((vbool8_t)(op0), (vfloat16m1_t)(op1), (vfloat16m2_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfredsum_vs_f16m4_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredsum_vs_f16m4_f16m1((vfloat16m1_t)(op0), (vfloat16m4_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfredsum_vs_f16m4_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredsum_vs_f16m4_f16m1_m((vbool4_t)(op0), (vfloat16m1_t)(op1), (vfloat16m4_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfredsum_vs_f16m8_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredsum_vs_f16m8_f16m1((vfloat16m1_t)(op0), (vfloat16m8_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfredsum_vs_f16m8_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredsum_vs_f16m8_f16m1_m((vbool2_t)(op0), (vfloat16m1_t)(op1), (vfloat16m8_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfredsum_vs_f16mf2_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredsum_vs_f16mf2_f16m1((vfloat16m1_t)(op0), (vfloat16mf2_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfredsum_vs_f16mf2_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredsum_vs_f16mf2_f16m1_m((vbool32_t)(op0), (vfloat16m1_t)(op1), (vfloat16mf2_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfredsum_vs_f16mf4_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredsum_vs_f16mf4_f16m1((vfloat16m1_t)(op0), (vfloat16mf4_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfredsum_vs_f16mf4_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredsum_vs_f16mf4_f16m1_m((vbool64_t)(op0), (vfloat16m1_t)(op1), (vfloat16mf4_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfredosum_vs_f16m1_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredosum_vs_f16m1_f16m1((vfloat16m1_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfredosum_vs_f16m1_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredosum_vs_f16m1_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfredosum_vs_f16m2_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredosum_vs_f16m2_f16m1((vfloat16m1_t)(op0), (vfloat16m2_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfredosum_vs_f16m2_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredosum_vs_f16m2_f16m1_m((vbool8_t)(op0), (vfloat16m1_t)(op1), (vfloat16m2_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfredosum_vs_f16m4_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredosum_vs_f16m4_f16m1((vfloat16m1_t)(op0), (vfloat16m4_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfredosum_vs_f16m4_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredosum_vs_f16m4_f16m1_m((vbool4_t)(op0), (vfloat16m1_t)(op1), (vfloat16m4_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfredosum_vs_f16m8_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredosum_vs_f16m8_f16m1((vfloat16m1_t)(op0), (vfloat16m8_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfredosum_vs_f16m8_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredosum_vs_f16m8_f16m1_m((vbool2_t)(op0), (vfloat16m1_t)(op1), (vfloat16m8_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfredosum_vs_f16mf2_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredosum_vs_f16mf2_f16m1((vfloat16m1_t)(op0), (vfloat16mf2_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfredosum_vs_f16mf2_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredosum_vs_f16mf2_f16m1_m((vbool32_t)(op0), (vfloat16m1_t)(op1), (vfloat16mf2_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfredosum_vs_f16mf4_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vfredosum_vs_f16mf4_f16m1((vfloat16m1_t)(op0), (vfloat16mf4_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfredosum_vs_f16mf4_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfredosum_vs_f16mf4_f16m1_m((vbool64_t)(op0), (vfloat16m1_t)(op1), (vfloat16mf4_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vsuxei8_v_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_f16m1((_Float16 *)(op0), (vuint8mf2_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vsuxei8_v_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint8mf2_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vsuxei8_v_f16m2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_f16m2((_Float16 *)(op0), (vuint8m1_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vsuxei8_v_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_f16m2_m((vbool8_t)(op0), (_Float16 *)(op1), (vuint8m1_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vsuxei8_v_f16m4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_f16m4((_Float16 *)(op0), (vuint8m2_t)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vsuxei8_v_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_f16m4_m((vbool4_t)(op0), (_Float16 *)(op1), (vuint8m2_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vsuxei8_v_f16m8(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_f16m8((_Float16 *)(op0), (vuint8m4_t)(op1), (vfloat16m8_t)(op2), (size_t)(op3)) +#define vsuxei8_v_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_f16m8_m((vbool2_t)(op0), (_Float16 *)(op1), (vuint8m4_t)(op2), (vfloat16m8_t)(op3), (size_t)(op4)) +#define vsuxei8_v_f16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_f16mf2((_Float16 *)(op0), (vuint8mf4_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vsuxei8_v_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint8mf4_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vsuxei8_v_f16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei8_v_f16mf4((_Float16 *)(op0), (vuint8mf8_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vsuxei8_v_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei8_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint8mf8_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vsuxei16_v_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_f16m1((_Float16 *)(op0), (vuint16m1_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vsuxei16_v_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vsuxei16_v_f16m2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_f16m2((_Float16 *)(op0), (vuint16m2_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vsuxei16_v_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_f16m2_m((vbool8_t)(op0), (_Float16 *)(op1), (vuint16m2_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vsuxei16_v_f16m4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_f16m4((_Float16 *)(op0), (vuint16m4_t)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vsuxei16_v_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_f16m4_m((vbool4_t)(op0), (_Float16 *)(op1), (vuint16m4_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vsuxei16_v_f16m8(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_f16m8((_Float16 *)(op0), (vuint16m8_t)(op1), (vfloat16m8_t)(op2), (size_t)(op3)) +#define vsuxei16_v_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_f16m8_m((vbool2_t)(op0), (_Float16 *)(op1), (vuint16m8_t)(op2), (vfloat16m8_t)(op3), (size_t)(op4)) +#define vsuxei16_v_f16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_f16mf2((_Float16 *)(op0), (vuint16mf2_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vsuxei16_v_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint16mf2_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vsuxei16_v_f16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei16_v_f16mf4((_Float16 *)(op0), (vuint16mf4_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vsuxei16_v_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei16_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint16mf4_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vfmv_f_s_f16m1_f16(op0) \ +__builtin_rvv_vfmv_f_s_f16m1_f16((vfloat16m1_t)(op0)) +#define vfmv_f_s_f16m2_f16(op0) \ +__builtin_rvv_vfmv_f_s_f16m2_f16((vfloat16m2_t)(op0)) +#define vfmv_f_s_f16m4_f16(op0) \ +__builtin_rvv_vfmv_f_s_f16m4_f16((vfloat16m4_t)(op0)) +#define vfmv_f_s_f16m8_f16(op0) \ +__builtin_rvv_vfmv_f_s_f16m8_f16((vfloat16m8_t)(op0)) +#define vfmv_f_s_f16mf2_f16(op0) \ +__builtin_rvv_vfmv_f_s_f16mf2_f16((vfloat16mf2_t)(op0)) +#define vfmv_f_s_f16mf4_f16(op0) \ +__builtin_rvv_vfmv_f_s_f16mf4_f16((vfloat16mf4_t)(op0)) +#define vfmv_s_f_f16m1(op0, op1, op2) \ +__builtin_rvv_vfmv_s_f_f16m1((vfloat16m1_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfmv_s_f_f16m2(op0, op1, op2) \ +__builtin_rvv_vfmv_s_f_f16m2((vfloat16m2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfmv_s_f_f16m4(op0, op1, op2) \ +__builtin_rvv_vfmv_s_f_f16m4((vfloat16m4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfmv_s_f_f16m8(op0, op1, op2) \ +__builtin_rvv_vfmv_s_f_f16m8((vfloat16m8_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfmv_s_f_f16mf2(op0, op1, op2) \ +__builtin_rvv_vfmv_s_f_f16mf2((vfloat16mf2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfmv_s_f_f16mf4(op0, op1, op2) \ +__builtin_rvv_vfmv_s_f_f16mf4((vfloat16mf4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vslideup_vx_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_f16m1((vfloat16m1_t)(op0), (vfloat16m1_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_f16m2(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_f16m2((vfloat16m2_t)(op0), (vfloat16m2_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_f16m4(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_f16m4((vfloat16m4_t)(op0), (vfloat16m4_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_f16m8(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_f16m8((vfloat16m8_t)(op0), (vfloat16m8_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_f16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_f16mf2((vfloat16mf2_t)(op0), (vfloat16mf2_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslideup_vx_f16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vslideup_vx_f16mf4((vfloat16mf4_t)(op0), (vfloat16mf4_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslideup_vx_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslideup_vx_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_f16m1((vfloat16m1_t)(op0), (vfloat16m1_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_f16m2(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_f16m2((vfloat16m2_t)(op0), (vfloat16m2_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_f16m4(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_f16m4((vfloat16m4_t)(op0), (vfloat16m4_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_f16m8(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_f16m8((vfloat16m8_t)(op0), (vfloat16m8_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_f16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_f16mf2((vfloat16mf2_t)(op0), (vfloat16mf2_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vslidedown_vx_f16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vslidedown_vx_f16mf4((vfloat16mf4_t)(op0), (vfloat16mf4_t)(op1), (size_t)(op2), (size_t)(op3)) +#define vslidedown_vx_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vslidedown_vx_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vfslide1up_vf_f16m1(op0, op1, op2) \ +__builtin_rvv_vfslide1up_vf_f16m1((vfloat16m1_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfslide1up_vf_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfslide1up_vf_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfslide1up_vf_f16m2(op0, op1, op2) \ +__builtin_rvv_vfslide1up_vf_f16m2((vfloat16m2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfslide1up_vf_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfslide1up_vf_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfslide1up_vf_f16m4(op0, op1, op2) \ +__builtin_rvv_vfslide1up_vf_f16m4((vfloat16m4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfslide1up_vf_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfslide1up_vf_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfslide1up_vf_f16m8(op0, op1, op2) \ +__builtin_rvv_vfslide1up_vf_f16m8((vfloat16m8_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfslide1up_vf_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfslide1up_vf_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfslide1up_vf_f16mf2(op0, op1, op2) \ +__builtin_rvv_vfslide1up_vf_f16mf2((vfloat16mf2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfslide1up_vf_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfslide1up_vf_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfslide1up_vf_f16mf4(op0, op1, op2) \ +__builtin_rvv_vfslide1up_vf_f16mf4((vfloat16mf4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfslide1up_vf_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfslide1up_vf_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vsuxei32_v_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_f16m1((_Float16 *)(op0), (vuint32m2_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vsuxei32_v_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint32m2_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vsuxei32_v_f16m2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_f16m2((_Float16 *)(op0), (vuint32m4_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vsuxei32_v_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_f16m2_m((vbool8_t)(op0), (_Float16 *)(op1), (vuint32m4_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vsuxei32_v_f16m4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_f16m4((_Float16 *)(op0), (vuint32m8_t)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vsuxei32_v_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_f16m4_m((vbool4_t)(op0), (_Float16 *)(op1), (vuint32m8_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vsuxei32_v_f16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_f16mf2((_Float16 *)(op0), (vuint32m1_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vsuxei32_v_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint32m1_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vsuxei32_v_f16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei32_v_f16mf4((_Float16 *)(op0), (vuint32mf2_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vsuxei32_v_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei32_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint32mf2_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vfslide1down_vf_f16m1(op0, op1, op2) \ +__builtin_rvv_vfslide1down_vf_f16m1((vfloat16m1_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfslide1down_vf_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfslide1down_vf_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfslide1down_vf_f16m2(op0, op1, op2) \ +__builtin_rvv_vfslide1down_vf_f16m2((vfloat16m2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfslide1down_vf_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfslide1down_vf_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfslide1down_vf_f16m4(op0, op1, op2) \ +__builtin_rvv_vfslide1down_vf_f16m4((vfloat16m4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfslide1down_vf_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfslide1down_vf_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfslide1down_vf_f16m8(op0, op1, op2) \ +__builtin_rvv_vfslide1down_vf_f16m8((vfloat16m8_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfslide1down_vf_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfslide1down_vf_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfslide1down_vf_f16mf2(op0, op1, op2) \ +__builtin_rvv_vfslide1down_vf_f16mf2((vfloat16mf2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfslide1down_vf_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfslide1down_vf_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfslide1down_vf_f16mf4(op0, op1, op2) \ +__builtin_rvv_vfslide1down_vf_f16mf4((vfloat16mf4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfslide1down_vf_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfslide1down_vf_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vrgather_vv_f16m1(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_f16m1((vfloat16m1_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vrgather_vv_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vrgather_vv_f16m2(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_f16m2((vfloat16m2_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vrgather_vv_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vrgather_vv_f16m4(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_f16m4((vfloat16m4_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vrgather_vv_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vrgather_vv_f16m8(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_f16m8((vfloat16m8_t)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vrgather_vv_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vrgather_vv_f16mf2(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_f16mf2((vfloat16mf2_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vrgather_vv_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vrgather_vv_f16mf4(op0, op1, op2) \ +__builtin_rvv_vrgather_vv_f16mf4((vfloat16mf4_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vrgather_vv_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vv_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vrgather_vx_f16m1(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_f16m1((vfloat16m1_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_f16m2(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_f16m2((vfloat16m2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_f16m4(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_f16m4((vfloat16m4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_f16m8(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_f16m8((vfloat16m8_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_f16mf2(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_f16mf2((vfloat16mf2_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgather_vx_f16mf4(op0, op1, op2) \ +__builtin_rvv_vrgather_vx_f16mf4((vfloat16mf4_t)(op0), (size_t)(op1), (size_t)(op2)) +#define vrgather_vx_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgather_vx_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_f16m1(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_f16m1((vfloat16m1_t)(op0), (vuint16m1_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_f16m2(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_f16m2((vfloat16m2_t)(op0), (vuint16m2_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_f16m4(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_f16m4((vfloat16m4_t)(op0), (vuint16m4_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_f16m8(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_f16m8((vfloat16m8_t)(op0), (vuint16m8_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_f16mf2(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_f16mf2((vfloat16mf2_t)(op0), (vuint16mf2_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vrgatherei16_vv_f16mf4(op0, op1, op2) \ +__builtin_rvv_vrgatherei16_vv_f16mf4((vfloat16mf4_t)(op0), (vuint16mf4_t)(op1), (size_t)(op2)) +#define vrgatherei16_vv_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vrgatherei16_vv_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vcompress_vm_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_f16m1((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vcompress_vm_f16m2(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_f16m2((vbool8_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vcompress_vm_f16m4(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_f16m4((vbool4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vcompress_vm_f16m8(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_f16m8((vbool2_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (size_t)(op3)) +#define vcompress_vm_f16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_f16mf2((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vcompress_vm_f16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vcompress_vm_f16mf4((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vsuxei64_v_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei64_v_f16m1((_Float16 *)(op0), (vuint64m4_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vsuxei64_v_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei64_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint64m4_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vsuxei64_v_f16m2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei64_v_f16m2((_Float16 *)(op0), (vuint64m8_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vsuxei64_v_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei64_v_f16m2_m((vbool8_t)(op0), (_Float16 *)(op1), (vuint64m8_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vsuxei64_v_f16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei64_v_f16mf2((_Float16 *)(op0), (vuint64m2_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vsuxei64_v_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei64_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint64m2_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vsuxei64_v_f16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsuxei64_v_f16mf4((_Float16 *)(op0), (vuint64m1_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vsuxei64_v_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxei64_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint64m1_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vsoxei8_v_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_f16m1((_Float16 *)(op0), (vuint8mf2_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vsoxei8_v_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint8mf2_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vsoxei8_v_f16m2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_f16m2((_Float16 *)(op0), (vuint8m1_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vsoxei8_v_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_f16m2_m((vbool8_t)(op0), (_Float16 *)(op1), (vuint8m1_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vsoxei8_v_f16m4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_f16m4((_Float16 *)(op0), (vuint8m2_t)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vsoxei8_v_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_f16m4_m((vbool4_t)(op0), (_Float16 *)(op1), (vuint8m2_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vsoxei8_v_f16m8(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_f16m8((_Float16 *)(op0), (vuint8m4_t)(op1), (vfloat16m8_t)(op2), (size_t)(op3)) +#define vsoxei8_v_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_f16m8_m((vbool2_t)(op0), (_Float16 *)(op1), (vuint8m4_t)(op2), (vfloat16m8_t)(op3), (size_t)(op4)) +#define vsoxei8_v_f16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_f16mf2((_Float16 *)(op0), (vuint8mf4_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vsoxei8_v_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint8mf4_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vsoxei8_v_f16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei8_v_f16mf4((_Float16 *)(op0), (vuint8mf8_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vsoxei8_v_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei8_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint8mf8_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vsoxei16_v_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_f16m1((_Float16 *)(op0), (vuint16m1_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vsoxei16_v_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vsoxei16_v_f16m2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_f16m2((_Float16 *)(op0), (vuint16m2_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vsoxei16_v_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_f16m2_m((vbool8_t)(op0), (_Float16 *)(op1), (vuint16m2_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vsoxei16_v_f16m4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_f16m4((_Float16 *)(op0), (vuint16m4_t)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vsoxei16_v_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_f16m4_m((vbool4_t)(op0), (_Float16 *)(op1), (vuint16m4_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vsoxei16_v_f16m8(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_f16m8((_Float16 *)(op0), (vuint16m8_t)(op1), (vfloat16m8_t)(op2), (size_t)(op3)) +#define vsoxei16_v_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_f16m8_m((vbool2_t)(op0), (_Float16 *)(op1), (vuint16m8_t)(op2), (vfloat16m8_t)(op3), (size_t)(op4)) +#define vsoxei16_v_f16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_f16mf2((_Float16 *)(op0), (vuint16mf2_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vsoxei16_v_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint16mf2_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vsoxei16_v_f16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei16_v_f16mf4((_Float16 *)(op0), (vuint16mf4_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vsoxei16_v_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei16_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint16mf4_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vsoxei32_v_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_f16m1((_Float16 *)(op0), (vuint32m2_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vsoxei32_v_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint32m2_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vsoxei32_v_f16m2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_f16m2((_Float16 *)(op0), (vuint32m4_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vsoxei32_v_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_f16m2_m((vbool8_t)(op0), (_Float16 *)(op1), (vuint32m4_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vsoxei32_v_f16m4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_f16m4((_Float16 *)(op0), (vuint32m8_t)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vsoxei32_v_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_f16m4_m((vbool4_t)(op0), (_Float16 *)(op1), (vuint32m8_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vsoxei32_v_f16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_f16mf2((_Float16 *)(op0), (vuint32m1_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vsoxei32_v_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint32m1_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vsoxei32_v_f16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei32_v_f16mf4((_Float16 *)(op0), (vuint32mf2_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vsoxei32_v_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei32_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint32mf2_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vsoxei64_v_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei64_v_f16m1((_Float16 *)(op0), (vuint64m4_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vsoxei64_v_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei64_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint64m4_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vsoxei64_v_f16m2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei64_v_f16m2((_Float16 *)(op0), (vuint64m8_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vsoxei64_v_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei64_v_f16m2_m((vbool8_t)(op0), (_Float16 *)(op1), (vuint64m8_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vsoxei64_v_f16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei64_v_f16mf2((_Float16 *)(op0), (vuint64m2_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vsoxei64_v_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei64_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint64m2_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vsoxei64_v_f16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsoxei64_v_f16mf4((_Float16 *)(op0), (vuint64m1_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vsoxei64_v_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxei64_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint64m1_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vle16ff_v_f16m1(op0, op1, op2) \ +__builtin_rvv_vle16ff_v_f16m1((const _Float16 *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle16ff_v_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle16ff_v_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (const _Float16 *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle16ff_v_f16m2(op0, op1, op2) \ +__builtin_rvv_vle16ff_v_f16m2((const _Float16 *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle16ff_v_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle16ff_v_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (const _Float16 *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle16ff_v_f16m4(op0, op1, op2) \ +__builtin_rvv_vle16ff_v_f16m4((const _Float16 *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle16ff_v_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle16ff_v_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (const _Float16 *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle16ff_v_f16m8(op0, op1, op2) \ +__builtin_rvv_vle16ff_v_f16m8((const _Float16 *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle16ff_v_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle16ff_v_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (const _Float16 *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle16ff_v_f16mf2(op0, op1, op2) \ +__builtin_rvv_vle16ff_v_f16mf2((const _Float16 *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle16ff_v_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle16ff_v_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (const _Float16 *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vle16ff_v_f16mf4(op0, op1, op2) \ +__builtin_rvv_vle16ff_v_f16mf4((const _Float16 *)(op0), (size_t *)(op1), (size_t)(op2)) +#define vle16ff_v_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vle16ff_v_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (const _Float16 *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vfneg_v_f16m1(op0, op1) \ +__builtin_rvv_vfneg_v_f16m1((vfloat16m1_t)(op0), (size_t)(op1)) +#define vfneg_v_f16m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfneg_v_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfneg_v_f16m2(op0, op1) \ +__builtin_rvv_vfneg_v_f16m2((vfloat16m2_t)(op0), (size_t)(op1)) +#define vfneg_v_f16m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfneg_v_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vfneg_v_f16m4(op0, op1) \ +__builtin_rvv_vfneg_v_f16m4((vfloat16m4_t)(op0), (size_t)(op1)) +#define vfneg_v_f16m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfneg_v_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vfneg_v_f16m8(op0, op1) \ +__builtin_rvv_vfneg_v_f16m8((vfloat16m8_t)(op0), (size_t)(op1)) +#define vfneg_v_f16m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfneg_v_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (size_t)(op3)) +#define vfneg_v_f16mf2(op0, op1) \ +__builtin_rvv_vfneg_v_f16mf2((vfloat16mf2_t)(op0), (size_t)(op1)) +#define vfneg_v_f16mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfneg_v_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vfneg_v_f16mf4(op0, op1) \ +__builtin_rvv_vfneg_v_f16mf4((vfloat16mf4_t)(op0), (size_t)(op1)) +#define vfneg_v_f16mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfneg_v_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vle16_v_f16m1(op0, op1) \ +__builtin_rvv_vle16_v_f16m1((const _Float16 *)(op0), (size_t)(op1)) +#define vle16_v_f16m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vle16_v_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (const _Float16 *)(op2), (size_t)(op3)) +#define vle16_v_f16m2(op0, op1) \ +__builtin_rvv_vle16_v_f16m2((const _Float16 *)(op0), (size_t)(op1)) +#define vle16_v_f16m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vle16_v_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (const _Float16 *)(op2), (size_t)(op3)) +#define vle16_v_f16m4(op0, op1) \ +__builtin_rvv_vle16_v_f16m4((const _Float16 *)(op0), (size_t)(op1)) +#define vle16_v_f16m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vle16_v_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (const _Float16 *)(op2), (size_t)(op3)) +#define vle16_v_f16m8(op0, op1) \ +__builtin_rvv_vle16_v_f16m8((const _Float16 *)(op0), (size_t)(op1)) +#define vle16_v_f16m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vle16_v_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (const _Float16 *)(op2), (size_t)(op3)) +#define vle16_v_f16mf2(op0, op1) \ +__builtin_rvv_vle16_v_f16mf2((const _Float16 *)(op0), (size_t)(op1)) +#define vle16_v_f16mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vle16_v_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (const _Float16 *)(op2), (size_t)(op3)) +#define vle16_v_f16mf4(op0, op1) \ +__builtin_rvv_vle16_v_f16mf4((const _Float16 *)(op0), (size_t)(op1)) +#define vle16_v_f16mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vle16_v_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (const _Float16 *)(op2), (size_t)(op3)) +#define vse16_v_f16m1(op0, op1, op2) \ +__builtin_rvv_vse16_v_f16m1((_Float16 *)(op0), (vfloat16m1_t)(op1), (size_t)(op2)) +#define vse16_v_f16m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vse16_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vse16_v_f16m2(op0, op1, op2) \ +__builtin_rvv_vse16_v_f16m2((_Float16 *)(op0), (vfloat16m2_t)(op1), (size_t)(op2)) +#define vse16_v_f16m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vse16_v_f16m2_m((vbool8_t)(op0), (_Float16 *)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vse16_v_f16m4(op0, op1, op2) \ +__builtin_rvv_vse16_v_f16m4((_Float16 *)(op0), (vfloat16m4_t)(op1), (size_t)(op2)) +#define vse16_v_f16m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vse16_v_f16m4_m((vbool4_t)(op0), (_Float16 *)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vse16_v_f16m8(op0, op1, op2) \ +__builtin_rvv_vse16_v_f16m8((_Float16 *)(op0), (vfloat16m8_t)(op1), (size_t)(op2)) +#define vse16_v_f16m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vse16_v_f16m8_m((vbool2_t)(op0), (_Float16 *)(op1), (vfloat16m8_t)(op2), (size_t)(op3)) +#define vse16_v_f16mf2(op0, op1, op2) \ +__builtin_rvv_vse16_v_f16mf2((_Float16 *)(op0), (vfloat16mf2_t)(op1), (size_t)(op2)) +#define vse16_v_f16mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vse16_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vse16_v_f16mf4(op0, op1, op2) \ +__builtin_rvv_vse16_v_f16mf4((_Float16 *)(op0), (vfloat16mf4_t)(op1), (size_t)(op2)) +#define vse16_v_f16mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vse16_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vlse16_v_f16m1(op0, op1, op2) \ +__builtin_rvv_vlse16_v_f16m1((const _Float16 *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse16_v_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse16_v_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (const _Float16 *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse16_v_f16m2(op0, op1, op2) \ +__builtin_rvv_vlse16_v_f16m2((const _Float16 *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse16_v_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse16_v_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (const _Float16 *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse16_v_f16m4(op0, op1, op2) \ +__builtin_rvv_vlse16_v_f16m4((const _Float16 *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse16_v_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse16_v_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (const _Float16 *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse16_v_f16m8(op0, op1, op2) \ +__builtin_rvv_vlse16_v_f16m8((const _Float16 *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse16_v_f16m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse16_v_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (const _Float16 *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse16_v_f16mf2(op0, op1, op2) \ +__builtin_rvv_vlse16_v_f16mf2((const _Float16 *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse16_v_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse16_v_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (const _Float16 *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlse16_v_f16mf4(op0, op1, op2) \ +__builtin_rvv_vlse16_v_f16mf4((const _Float16 *)(op0), (ptrdiff_t)(op1), (size_t)(op2)) +#define vlse16_v_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlse16_v_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (const _Float16 *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vfclass_v_u16m1(op0, op1) \ +__builtin_rvv_vfclass_v_u16m1((vfloat16m1_t)(op0), (size_t)(op1)) +#define vfclass_v_u16m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfclass_v_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfclass_v_u16m2(op0, op1) \ +__builtin_rvv_vfclass_v_u16m2((vfloat16m2_t)(op0), (size_t)(op1)) +#define vfclass_v_u16m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfclass_v_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vfclass_v_u16m4(op0, op1) \ +__builtin_rvv_vfclass_v_u16m4((vfloat16m4_t)(op0), (size_t)(op1)) +#define vfclass_v_u16m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfclass_v_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vfclass_v_u16m8(op0, op1) \ +__builtin_rvv_vfclass_v_u16m8((vfloat16m8_t)(op0), (size_t)(op1)) +#define vfclass_v_u16m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfclass_v_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vfloat16m8_t)(op2), (size_t)(op3)) +#define vfclass_v_u16mf2(op0, op1) \ +__builtin_rvv_vfclass_v_u16mf2((vfloat16mf2_t)(op0), (size_t)(op1)) +#define vfclass_v_u16mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfclass_v_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vfclass_v_u16mf4(op0, op1) \ +__builtin_rvv_vfclass_v_u16mf4((vfloat16mf4_t)(op0), (size_t)(op1)) +#define vfclass_v_u16mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfclass_v_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vfcvt_f_x_v_f16m1(op0, op1) \ +__builtin_rvv_vfcvt_f_x_v_f16m1((vint16m1_t)(op0), (size_t)(op1)) +#define vfcvt_f_x_v_f16m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_f_x_v_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vfcvt_f_x_v_f16m2(op0, op1) \ +__builtin_rvv_vfcvt_f_x_v_f16m2((vint16m2_t)(op0), (size_t)(op1)) +#define vfcvt_f_x_v_f16m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_f_x_v_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vint16m2_t)(op2), (size_t)(op3)) +#define vfcvt_f_x_v_f16m4(op0, op1) \ +__builtin_rvv_vfcvt_f_x_v_f16m4((vint16m4_t)(op0), (size_t)(op1)) +#define vfcvt_f_x_v_f16m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_f_x_v_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vint16m4_t)(op2), (size_t)(op3)) +#define vfcvt_f_x_v_f16m8(op0, op1) \ +__builtin_rvv_vfcvt_f_x_v_f16m8((vint16m8_t)(op0), (size_t)(op1)) +#define vfcvt_f_x_v_f16m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_f_x_v_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (vint16m8_t)(op2), (size_t)(op3)) +#define vfcvt_f_x_v_f16mf2(op0, op1) \ +__builtin_rvv_vfcvt_f_x_v_f16mf2((vint16mf2_t)(op0), (size_t)(op1)) +#define vfcvt_f_x_v_f16mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_f_x_v_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vint16mf2_t)(op2), (size_t)(op3)) +#define vfcvt_f_x_v_f16mf4(op0, op1) \ +__builtin_rvv_vfcvt_f_x_v_f16mf4((vint16mf4_t)(op0), (size_t)(op1)) +#define vfcvt_f_x_v_f16mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_f_x_v_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vint16mf4_t)(op2), (size_t)(op3)) +#define vfcvt_f_xu_v_f16m1(op0, op1) \ +__builtin_rvv_vfcvt_f_xu_v_f16m1((vuint16m1_t)(op0), (size_t)(op1)) +#define vfcvt_f_xu_v_f16m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_f_xu_v_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vfcvt_f_xu_v_f16m2(op0, op1) \ +__builtin_rvv_vfcvt_f_xu_v_f16m2((vuint16m2_t)(op0), (size_t)(op1)) +#define vfcvt_f_xu_v_f16m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_f_xu_v_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vuint16m2_t)(op2), (size_t)(op3)) +#define vfcvt_f_xu_v_f16m4(op0, op1) \ +__builtin_rvv_vfcvt_f_xu_v_f16m4((vuint16m4_t)(op0), (size_t)(op1)) +#define vfcvt_f_xu_v_f16m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_f_xu_v_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vuint16m4_t)(op2), (size_t)(op3)) +#define vfcvt_f_xu_v_f16m8(op0, op1) \ +__builtin_rvv_vfcvt_f_xu_v_f16m8((vuint16m8_t)(op0), (size_t)(op1)) +#define vfcvt_f_xu_v_f16m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_f_xu_v_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (vuint16m8_t)(op2), (size_t)(op3)) +#define vfcvt_f_xu_v_f16mf2(op0, op1) \ +__builtin_rvv_vfcvt_f_xu_v_f16mf2((vuint16mf2_t)(op0), (size_t)(op1)) +#define vfcvt_f_xu_v_f16mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_f_xu_v_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vuint16mf2_t)(op2), (size_t)(op3)) +#define vfcvt_f_xu_v_f16mf4(op0, op1) \ +__builtin_rvv_vfcvt_f_xu_v_f16mf4((vuint16mf4_t)(op0), (size_t)(op1)) +#define vfcvt_f_xu_v_f16mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_f_xu_v_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vuint16mf4_t)(op2), (size_t)(op3)) +#define vfcvt_rtz_x_f_v_i16m1(op0, op1) \ +__builtin_rvv_vfcvt_rtz_x_f_v_i16m1((vfloat16m1_t)(op0), (size_t)(op1)) +#define vfcvt_rtz_x_f_v_i16m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_rtz_x_f_v_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfcvt_rtz_x_f_v_i16m2(op0, op1) \ +__builtin_rvv_vfcvt_rtz_x_f_v_i16m2((vfloat16m2_t)(op0), (size_t)(op1)) +#define vfcvt_rtz_x_f_v_i16m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_rtz_x_f_v_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vfcvt_rtz_x_f_v_i16m4(op0, op1) \ +__builtin_rvv_vfcvt_rtz_x_f_v_i16m4((vfloat16m4_t)(op0), (size_t)(op1)) +#define vfcvt_rtz_x_f_v_i16m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_rtz_x_f_v_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vfcvt_rtz_x_f_v_i16m8(op0, op1) \ +__builtin_rvv_vfcvt_rtz_x_f_v_i16m8((vfloat16m8_t)(op0), (size_t)(op1)) +#define vfcvt_rtz_x_f_v_i16m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_rtz_x_f_v_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vfloat16m8_t)(op2), (size_t)(op3)) +#define vfcvt_rtz_x_f_v_i16mf2(op0, op1) \ +__builtin_rvv_vfcvt_rtz_x_f_v_i16mf2((vfloat16mf2_t)(op0), (size_t)(op1)) +#define vfcvt_rtz_x_f_v_i16mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_rtz_x_f_v_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vfcvt_rtz_x_f_v_i16mf4(op0, op1) \ +__builtin_rvv_vfcvt_rtz_x_f_v_i16mf4((vfloat16mf4_t)(op0), (size_t)(op1)) +#define vfcvt_rtz_x_f_v_i16mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_rtz_x_f_v_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vfcvt_rtz_xu_f_v_u16m1(op0, op1) \ +__builtin_rvv_vfcvt_rtz_xu_f_v_u16m1((vfloat16m1_t)(op0), (size_t)(op1)) +#define vfcvt_rtz_xu_f_v_u16m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_rtz_xu_f_v_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfcvt_rtz_xu_f_v_u16m2(op0, op1) \ +__builtin_rvv_vfcvt_rtz_xu_f_v_u16m2((vfloat16m2_t)(op0), (size_t)(op1)) +#define vfcvt_rtz_xu_f_v_u16m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_rtz_xu_f_v_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vfcvt_rtz_xu_f_v_u16m4(op0, op1) \ +__builtin_rvv_vfcvt_rtz_xu_f_v_u16m4((vfloat16m4_t)(op0), (size_t)(op1)) +#define vfcvt_rtz_xu_f_v_u16m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_rtz_xu_f_v_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vfcvt_rtz_xu_f_v_u16m8(op0, op1) \ +__builtin_rvv_vfcvt_rtz_xu_f_v_u16m8((vfloat16m8_t)(op0), (size_t)(op1)) +#define vfcvt_rtz_xu_f_v_u16m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_rtz_xu_f_v_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vfloat16m8_t)(op2), (size_t)(op3)) +#define vfcvt_rtz_xu_f_v_u16mf2(op0, op1) \ +__builtin_rvv_vfcvt_rtz_xu_f_v_u16mf2((vfloat16mf2_t)(op0), (size_t)(op1)) +#define vfcvt_rtz_xu_f_v_u16mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_rtz_xu_f_v_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vfcvt_rtz_xu_f_v_u16mf4(op0, op1) \ +__builtin_rvv_vfcvt_rtz_xu_f_v_u16mf4((vfloat16mf4_t)(op0), (size_t)(op1)) +#define vfcvt_rtz_xu_f_v_u16mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_rtz_xu_f_v_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vfcvt_x_f_v_i16m1(op0, op1) \ +__builtin_rvv_vfcvt_x_f_v_i16m1((vfloat16m1_t)(op0), (size_t)(op1)) +#define vfcvt_x_f_v_i16m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_x_f_v_i16m1_m((vbool16_t)(op0), (vint16m1_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfcvt_x_f_v_i16m2(op0, op1) \ +__builtin_rvv_vfcvt_x_f_v_i16m2((vfloat16m2_t)(op0), (size_t)(op1)) +#define vfcvt_x_f_v_i16m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_x_f_v_i16m2_m((vbool8_t)(op0), (vint16m2_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vfcvt_x_f_v_i16m4(op0, op1) \ +__builtin_rvv_vfcvt_x_f_v_i16m4((vfloat16m4_t)(op0), (size_t)(op1)) +#define vfcvt_x_f_v_i16m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_x_f_v_i16m4_m((vbool4_t)(op0), (vint16m4_t)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vfcvt_x_f_v_i16m8(op0, op1) \ +__builtin_rvv_vfcvt_x_f_v_i16m8((vfloat16m8_t)(op0), (size_t)(op1)) +#define vfcvt_x_f_v_i16m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_x_f_v_i16m8_m((vbool2_t)(op0), (vint16m8_t)(op1), (vfloat16m8_t)(op2), (size_t)(op3)) +#define vfcvt_x_f_v_i16mf2(op0, op1) \ +__builtin_rvv_vfcvt_x_f_v_i16mf2((vfloat16mf2_t)(op0), (size_t)(op1)) +#define vfcvt_x_f_v_i16mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_x_f_v_i16mf2_m((vbool32_t)(op0), (vint16mf2_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vfcvt_x_f_v_i16mf4(op0, op1) \ +__builtin_rvv_vfcvt_x_f_v_i16mf4((vfloat16mf4_t)(op0), (size_t)(op1)) +#define vfcvt_x_f_v_i16mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_x_f_v_i16mf4_m((vbool64_t)(op0), (vint16mf4_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vfcvt_xu_f_v_u16m1(op0, op1) \ +__builtin_rvv_vfcvt_xu_f_v_u16m1((vfloat16m1_t)(op0), (size_t)(op1)) +#define vfcvt_xu_f_v_u16m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_xu_f_v_u16m1_m((vbool16_t)(op0), (vuint16m1_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfcvt_xu_f_v_u16m2(op0, op1) \ +__builtin_rvv_vfcvt_xu_f_v_u16m2((vfloat16m2_t)(op0), (size_t)(op1)) +#define vfcvt_xu_f_v_u16m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_xu_f_v_u16m2_m((vbool8_t)(op0), (vuint16m2_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vfcvt_xu_f_v_u16m4(op0, op1) \ +__builtin_rvv_vfcvt_xu_f_v_u16m4((vfloat16m4_t)(op0), (size_t)(op1)) +#define vfcvt_xu_f_v_u16m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_xu_f_v_u16m4_m((vbool4_t)(op0), (vuint16m4_t)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vfcvt_xu_f_v_u16m8(op0, op1) \ +__builtin_rvv_vfcvt_xu_f_v_u16m8((vfloat16m8_t)(op0), (size_t)(op1)) +#define vfcvt_xu_f_v_u16m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_xu_f_v_u16m8_m((vbool2_t)(op0), (vuint16m8_t)(op1), (vfloat16m8_t)(op2), (size_t)(op3)) +#define vfcvt_xu_f_v_u16mf2(op0, op1) \ +__builtin_rvv_vfcvt_xu_f_v_u16mf2((vfloat16mf2_t)(op0), (size_t)(op1)) +#define vfcvt_xu_f_v_u16mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_xu_f_v_u16mf2_m((vbool32_t)(op0), (vuint16mf2_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vfcvt_xu_f_v_u16mf4(op0, op1) \ +__builtin_rvv_vfcvt_xu_f_v_u16mf4((vfloat16mf4_t)(op0), (size_t)(op1)) +#define vfcvt_xu_f_v_u16mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfcvt_xu_f_v_u16mf4_m((vbool64_t)(op0), (vuint16mf4_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vfncvt_f_x_w_f16mf4(op0, op1) \ +__builtin_rvv_vfncvt_f_x_w_f16mf4((vint32mf2_t)(op0), (size_t)(op1)) +#define vfncvt_f_x_w_f16mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_f_x_w_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vfncvt_f_x_w_f16mf2(op0, op1) \ +__builtin_rvv_vfncvt_f_x_w_f16mf2((vint32m1_t)(op0), (size_t)(op1)) +#define vfncvt_f_x_w_f16mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_f_x_w_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vfncvt_f_x_w_f16m1(op0, op1) \ +__builtin_rvv_vfncvt_f_x_w_f16m1((vint32m2_t)(op0), (size_t)(op1)) +#define vfncvt_f_x_w_f16m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_f_x_w_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vfncvt_f_x_w_f16m2(op0, op1) \ +__builtin_rvv_vfncvt_f_x_w_f16m2((vint32m4_t)(op0), (size_t)(op1)) +#define vfncvt_f_x_w_f16m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_f_x_w_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vfncvt_f_x_w_f16m4(op0, op1) \ +__builtin_rvv_vfncvt_f_x_w_f16m4((vint32m8_t)(op0), (size_t)(op1)) +#define vfncvt_f_x_w_f16m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_f_x_w_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vint32m8_t)(op2), (size_t)(op3)) +#define vfncvt_f_xu_w_f16mf4(op0, op1) \ +__builtin_rvv_vfncvt_f_xu_w_f16mf4((vuint32mf2_t)(op0), (size_t)(op1)) +#define vfncvt_f_xu_w_f16mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_f_xu_w_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vfncvt_f_xu_w_f16mf2(op0, op1) \ +__builtin_rvv_vfncvt_f_xu_w_f16mf2((vuint32m1_t)(op0), (size_t)(op1)) +#define vfncvt_f_xu_w_f16mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_f_xu_w_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vfncvt_f_xu_w_f16m1(op0, op1) \ +__builtin_rvv_vfncvt_f_xu_w_f16m1((vuint32m2_t)(op0), (size_t)(op1)) +#define vfncvt_f_xu_w_f16m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_f_xu_w_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vfncvt_f_xu_w_f16m2(op0, op1) \ +__builtin_rvv_vfncvt_f_xu_w_f16m2((vuint32m4_t)(op0), (size_t)(op1)) +#define vfncvt_f_xu_w_f16m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_f_xu_w_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vfncvt_f_xu_w_f16m4(op0, op1) \ +__builtin_rvv_vfncvt_f_xu_w_f16m4((vuint32m8_t)(op0), (size_t)(op1)) +#define vfncvt_f_xu_w_f16m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_f_xu_w_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vuint32m8_t)(op2), (size_t)(op3)) +#define vfncvt_rtz_x_f_w_i8mf8(op0, op1) \ +__builtin_rvv_vfncvt_rtz_x_f_w_i8mf8((vfloat16mf4_t)(op0), (size_t)(op1)) +#define vfncvt_rtz_x_f_w_i8mf8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_rtz_x_f_w_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vfncvt_rtz_x_f_w_i8mf4(op0, op1) \ +__builtin_rvv_vfncvt_rtz_x_f_w_i8mf4((vfloat16mf2_t)(op0), (size_t)(op1)) +#define vfncvt_rtz_x_f_w_i8mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_rtz_x_f_w_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vfncvt_rtz_x_f_w_i8mf2(op0, op1) \ +__builtin_rvv_vfncvt_rtz_x_f_w_i8mf2((vfloat16m1_t)(op0), (size_t)(op1)) +#define vfncvt_rtz_x_f_w_i8mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_rtz_x_f_w_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfncvt_rtz_x_f_w_i8m1(op0, op1) \ +__builtin_rvv_vfncvt_rtz_x_f_w_i8m1((vfloat16m2_t)(op0), (size_t)(op1)) +#define vfncvt_rtz_x_f_w_i8m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_rtz_x_f_w_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vfncvt_rtz_x_f_w_i8m2(op0, op1) \ +__builtin_rvv_vfncvt_rtz_x_f_w_i8m2((vfloat16m4_t)(op0), (size_t)(op1)) +#define vfncvt_rtz_x_f_w_i8m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_rtz_x_f_w_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vfncvt_rtz_x_f_w_i8m4(op0, op1) \ +__builtin_rvv_vfncvt_rtz_x_f_w_i8m4((vfloat16m8_t)(op0), (size_t)(op1)) +#define vfncvt_rtz_x_f_w_i8m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_rtz_x_f_w_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vfloat16m8_t)(op2), (size_t)(op3)) +#define vfncvt_rtz_xu_f_w_u8mf8(op0, op1) \ +__builtin_rvv_vfncvt_rtz_xu_f_w_u8mf8((vfloat16mf4_t)(op0), (size_t)(op1)) +#define vfncvt_rtz_xu_f_w_u8mf8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_rtz_xu_f_w_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vfncvt_rtz_xu_f_w_u8mf4(op0, op1) \ +__builtin_rvv_vfncvt_rtz_xu_f_w_u8mf4((vfloat16mf2_t)(op0), (size_t)(op1)) +#define vfncvt_rtz_xu_f_w_u8mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_rtz_xu_f_w_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vfncvt_rtz_xu_f_w_u8mf2(op0, op1) \ +__builtin_rvv_vfncvt_rtz_xu_f_w_u8mf2((vfloat16m1_t)(op0), (size_t)(op1)) +#define vfncvt_rtz_xu_f_w_u8mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_rtz_xu_f_w_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfncvt_rtz_xu_f_w_u8m1(op0, op1) \ +__builtin_rvv_vfncvt_rtz_xu_f_w_u8m1((vfloat16m2_t)(op0), (size_t)(op1)) +#define vfncvt_rtz_xu_f_w_u8m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_rtz_xu_f_w_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vfncvt_rtz_xu_f_w_u8m2(op0, op1) \ +__builtin_rvv_vfncvt_rtz_xu_f_w_u8m2((vfloat16m4_t)(op0), (size_t)(op1)) +#define vfncvt_rtz_xu_f_w_u8m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_rtz_xu_f_w_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vfncvt_rtz_xu_f_w_u8m4(op0, op1) \ +__builtin_rvv_vfncvt_rtz_xu_f_w_u8m4((vfloat16m8_t)(op0), (size_t)(op1)) +#define vfncvt_rtz_xu_f_w_u8m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_rtz_xu_f_w_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vfloat16m8_t)(op2), (size_t)(op3)) +#define vfncvt_x_f_w_i8mf8(op0, op1) \ +__builtin_rvv_vfncvt_x_f_w_i8mf8((vfloat16mf4_t)(op0), (size_t)(op1)) +#define vfncvt_x_f_w_i8mf8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_x_f_w_i8mf8_m((vbool64_t)(op0), (vint8mf8_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vfncvt_x_f_w_i8mf4(op0, op1) \ +__builtin_rvv_vfncvt_x_f_w_i8mf4((vfloat16mf2_t)(op0), (size_t)(op1)) +#define vfncvt_x_f_w_i8mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_x_f_w_i8mf4_m((vbool32_t)(op0), (vint8mf4_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vfncvt_x_f_w_i8mf2(op0, op1) \ +__builtin_rvv_vfncvt_x_f_w_i8mf2((vfloat16m1_t)(op0), (size_t)(op1)) +#define vfncvt_x_f_w_i8mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_x_f_w_i8mf2_m((vbool16_t)(op0), (vint8mf2_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfncvt_x_f_w_i8m1(op0, op1) \ +__builtin_rvv_vfncvt_x_f_w_i8m1((vfloat16m2_t)(op0), (size_t)(op1)) +#define vfncvt_x_f_w_i8m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_x_f_w_i8m1_m((vbool8_t)(op0), (vint8m1_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vfncvt_x_f_w_i8m2(op0, op1) \ +__builtin_rvv_vfncvt_x_f_w_i8m2((vfloat16m4_t)(op0), (size_t)(op1)) +#define vfncvt_x_f_w_i8m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_x_f_w_i8m2_m((vbool4_t)(op0), (vint8m2_t)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vfncvt_x_f_w_i8m4(op0, op1) \ +__builtin_rvv_vfncvt_x_f_w_i8m4((vfloat16m8_t)(op0), (size_t)(op1)) +#define vfncvt_x_f_w_i8m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_x_f_w_i8m4_m((vbool2_t)(op0), (vint8m4_t)(op1), (vfloat16m8_t)(op2), (size_t)(op3)) +#define vfncvt_xu_f_w_u8mf8(op0, op1) \ +__builtin_rvv_vfncvt_xu_f_w_u8mf8((vfloat16mf4_t)(op0), (size_t)(op1)) +#define vfncvt_xu_f_w_u8mf8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_xu_f_w_u8mf8_m((vbool64_t)(op0), (vuint8mf8_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vfncvt_xu_f_w_u8mf4(op0, op1) \ +__builtin_rvv_vfncvt_xu_f_w_u8mf4((vfloat16mf2_t)(op0), (size_t)(op1)) +#define vfncvt_xu_f_w_u8mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_xu_f_w_u8mf4_m((vbool32_t)(op0), (vuint8mf4_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vfncvt_xu_f_w_u8mf2(op0, op1) \ +__builtin_rvv_vfncvt_xu_f_w_u8mf2((vfloat16m1_t)(op0), (size_t)(op1)) +#define vfncvt_xu_f_w_u8mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_xu_f_w_u8mf2_m((vbool16_t)(op0), (vuint8mf2_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfncvt_xu_f_w_u8m1(op0, op1) \ +__builtin_rvv_vfncvt_xu_f_w_u8m1((vfloat16m2_t)(op0), (size_t)(op1)) +#define vfncvt_xu_f_w_u8m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_xu_f_w_u8m1_m((vbool8_t)(op0), (vuint8m1_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vfncvt_xu_f_w_u8m2(op0, op1) \ +__builtin_rvv_vfncvt_xu_f_w_u8m2((vfloat16m4_t)(op0), (size_t)(op1)) +#define vfncvt_xu_f_w_u8m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_xu_f_w_u8m2_m((vbool4_t)(op0), (vuint8m2_t)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vfncvt_xu_f_w_u8m4(op0, op1) \ +__builtin_rvv_vfncvt_xu_f_w_u8m4((vfloat16m8_t)(op0), (size_t)(op1)) +#define vfncvt_xu_f_w_u8m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_xu_f_w_u8m4_m((vbool2_t)(op0), (vuint8m4_t)(op1), (vfloat16m8_t)(op2), (size_t)(op3)) +#define vfrec7_v_f16m1(op0, op1) \ +__builtin_rvv_vfrec7_v_f16m1((vfloat16m1_t)(op0), (size_t)(op1)) +#define vfrec7_v_f16m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfrec7_v_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfrec7_v_f16m2(op0, op1) \ +__builtin_rvv_vfrec7_v_f16m2((vfloat16m2_t)(op0), (size_t)(op1)) +#define vfrec7_v_f16m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfrec7_v_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vfrec7_v_f16m4(op0, op1) \ +__builtin_rvv_vfrec7_v_f16m4((vfloat16m4_t)(op0), (size_t)(op1)) +#define vfrec7_v_f16m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfrec7_v_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vfrec7_v_f16m8(op0, op1) \ +__builtin_rvv_vfrec7_v_f16m8((vfloat16m8_t)(op0), (size_t)(op1)) +#define vfrec7_v_f16m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfrec7_v_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (size_t)(op3)) +#define vfrec7_v_f16mf2(op0, op1) \ +__builtin_rvv_vfrec7_v_f16mf2((vfloat16mf2_t)(op0), (size_t)(op1)) +#define vfrec7_v_f16mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfrec7_v_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vfrec7_v_f16mf4(op0, op1) \ +__builtin_rvv_vfrec7_v_f16mf4((vfloat16mf4_t)(op0), (size_t)(op1)) +#define vfrec7_v_f16mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfrec7_v_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vfrsqrt7_v_f16m1(op0, op1) \ +__builtin_rvv_vfrsqrt7_v_f16m1((vfloat16m1_t)(op0), (size_t)(op1)) +#define vfrsqrt7_v_f16m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfrsqrt7_v_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfrsqrt7_v_f16m2(op0, op1) \ +__builtin_rvv_vfrsqrt7_v_f16m2((vfloat16m2_t)(op0), (size_t)(op1)) +#define vfrsqrt7_v_f16m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfrsqrt7_v_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vfrsqrt7_v_f16m4(op0, op1) \ +__builtin_rvv_vfrsqrt7_v_f16m4((vfloat16m4_t)(op0), (size_t)(op1)) +#define vfrsqrt7_v_f16m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfrsqrt7_v_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vfrsqrt7_v_f16m8(op0, op1) \ +__builtin_rvv_vfrsqrt7_v_f16m8((vfloat16m8_t)(op0), (size_t)(op1)) +#define vfrsqrt7_v_f16m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfrsqrt7_v_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (size_t)(op3)) +#define vfrsqrt7_v_f16mf2(op0, op1) \ +__builtin_rvv_vfrsqrt7_v_f16mf2((vfloat16mf2_t)(op0), (size_t)(op1)) +#define vfrsqrt7_v_f16mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfrsqrt7_v_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vfrsqrt7_v_f16mf4(op0, op1) \ +__builtin_rvv_vfrsqrt7_v_f16mf4((vfloat16mf4_t)(op0), (size_t)(op1)) +#define vfrsqrt7_v_f16mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfrsqrt7_v_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vfsqrt_v_f16m1(op0, op1) \ +__builtin_rvv_vfsqrt_v_f16m1((vfloat16m1_t)(op0), (size_t)(op1)) +#define vfsqrt_v_f16m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfsqrt_v_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfsqrt_v_f16m2(op0, op1) \ +__builtin_rvv_vfsqrt_v_f16m2((vfloat16m2_t)(op0), (size_t)(op1)) +#define vfsqrt_v_f16m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfsqrt_v_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vfsqrt_v_f16m4(op0, op1) \ +__builtin_rvv_vfsqrt_v_f16m4((vfloat16m4_t)(op0), (size_t)(op1)) +#define vfsqrt_v_f16m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfsqrt_v_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vfsqrt_v_f16m8(op0, op1) \ +__builtin_rvv_vfsqrt_v_f16m8((vfloat16m8_t)(op0), (size_t)(op1)) +#define vfsqrt_v_f16m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfsqrt_v_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (vfloat16m8_t)(op2), (size_t)(op3)) +#define vfsqrt_v_f16mf2(op0, op1) \ +__builtin_rvv_vfsqrt_v_f16mf2((vfloat16mf2_t)(op0), (size_t)(op1)) +#define vfsqrt_v_f16mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfsqrt_v_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vfsqrt_v_f16mf4(op0, op1) \ +__builtin_rvv_vfsqrt_v_f16mf4((vfloat16mf4_t)(op0), (size_t)(op1)) +#define vfsqrt_v_f16mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfsqrt_v_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vfwcvt_f_x_v_f16mf4(op0, op1) \ +__builtin_rvv_vfwcvt_f_x_v_f16mf4((vint8mf8_t)(op0), (size_t)(op1)) +#define vfwcvt_f_x_v_f16mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_f_x_v_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vint8mf8_t)(op2), (size_t)(op3)) +#define vfwcvt_f_x_v_f16mf2(op0, op1) \ +__builtin_rvv_vfwcvt_f_x_v_f16mf2((vint8mf4_t)(op0), (size_t)(op1)) +#define vfwcvt_f_x_v_f16mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_f_x_v_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vint8mf4_t)(op2), (size_t)(op3)) +#define vfwcvt_f_x_v_f16m1(op0, op1) \ +__builtin_rvv_vfwcvt_f_x_v_f16m1((vint8mf2_t)(op0), (size_t)(op1)) +#define vfwcvt_f_x_v_f16m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_f_x_v_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vint8mf2_t)(op2), (size_t)(op3)) +#define vfwcvt_f_x_v_f16m2(op0, op1) \ +__builtin_rvv_vfwcvt_f_x_v_f16m2((vint8m1_t)(op0), (size_t)(op1)) +#define vfwcvt_f_x_v_f16m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_f_x_v_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vfwcvt_f_x_v_f16m4(op0, op1) \ +__builtin_rvv_vfwcvt_f_x_v_f16m4((vint8m2_t)(op0), (size_t)(op1)) +#define vfwcvt_f_x_v_f16m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_f_x_v_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vint8m2_t)(op2), (size_t)(op3)) +#define vfwcvt_f_x_v_f16m8(op0, op1) \ +__builtin_rvv_vfwcvt_f_x_v_f16m8((vint8m4_t)(op0), (size_t)(op1)) +#define vfwcvt_f_x_v_f16m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_f_x_v_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (vint8m4_t)(op2), (size_t)(op3)) +#define vfwcvt_f_xu_v_f16mf4(op0, op1) \ +__builtin_rvv_vfwcvt_f_xu_v_f16mf4((vuint8mf8_t)(op0), (size_t)(op1)) +#define vfwcvt_f_xu_v_f16mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_f_xu_v_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vuint8mf8_t)(op2), (size_t)(op3)) +#define vfwcvt_f_xu_v_f16mf2(op0, op1) \ +__builtin_rvv_vfwcvt_f_xu_v_f16mf2((vuint8mf4_t)(op0), (size_t)(op1)) +#define vfwcvt_f_xu_v_f16mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_f_xu_v_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vuint8mf4_t)(op2), (size_t)(op3)) +#define vfwcvt_f_xu_v_f16m1(op0, op1) \ +__builtin_rvv_vfwcvt_f_xu_v_f16m1((vuint8mf2_t)(op0), (size_t)(op1)) +#define vfwcvt_f_xu_v_f16m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_f_xu_v_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vuint8mf2_t)(op2), (size_t)(op3)) +#define vfwcvt_f_xu_v_f16m2(op0, op1) \ +__builtin_rvv_vfwcvt_f_xu_v_f16m2((vuint8m1_t)(op0), (size_t)(op1)) +#define vfwcvt_f_xu_v_f16m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_f_xu_v_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vfwcvt_f_xu_v_f16m4(op0, op1) \ +__builtin_rvv_vfwcvt_f_xu_v_f16m4((vuint8m2_t)(op0), (size_t)(op1)) +#define vfwcvt_f_xu_v_f16m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_f_xu_v_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vuint8m2_t)(op2), (size_t)(op3)) +#define vfwcvt_f_xu_v_f16m8(op0, op1) \ +__builtin_rvv_vfwcvt_f_xu_v_f16m8((vuint8m4_t)(op0), (size_t)(op1)) +#define vfwcvt_f_xu_v_f16m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_f_xu_v_f16m8_m((vbool2_t)(op0), (vfloat16m8_t)(op1), (vuint8m4_t)(op2), (size_t)(op3)) +#define vfwcvt_rtz_x_f_v_i32mf2(op0, op1) \ +__builtin_rvv_vfwcvt_rtz_x_f_v_i32mf2((vfloat16mf4_t)(op0), (size_t)(op1)) +#define vfwcvt_rtz_x_f_v_i32mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_rtz_x_f_v_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vfwcvt_rtz_x_f_v_i32m1(op0, op1) \ +__builtin_rvv_vfwcvt_rtz_x_f_v_i32m1((vfloat16mf2_t)(op0), (size_t)(op1)) +#define vfwcvt_rtz_x_f_v_i32m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_rtz_x_f_v_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vfwcvt_rtz_x_f_v_i32m2(op0, op1) \ +__builtin_rvv_vfwcvt_rtz_x_f_v_i32m2((vfloat16m1_t)(op0), (size_t)(op1)) +#define vfwcvt_rtz_x_f_v_i32m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_rtz_x_f_v_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfwcvt_rtz_x_f_v_i32m4(op0, op1) \ +__builtin_rvv_vfwcvt_rtz_x_f_v_i32m4((vfloat16m2_t)(op0), (size_t)(op1)) +#define vfwcvt_rtz_x_f_v_i32m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_rtz_x_f_v_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vfwcvt_rtz_x_f_v_i32m8(op0, op1) \ +__builtin_rvv_vfwcvt_rtz_x_f_v_i32m8((vfloat16m4_t)(op0), (size_t)(op1)) +#define vfwcvt_rtz_x_f_v_i32m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_rtz_x_f_v_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vfwcvt_rtz_xu_f_v_u32mf2(op0, op1) \ +__builtin_rvv_vfwcvt_rtz_xu_f_v_u32mf2((vfloat16mf4_t)(op0), (size_t)(op1)) +#define vfwcvt_rtz_xu_f_v_u32mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_rtz_xu_f_v_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vfwcvt_rtz_xu_f_v_u32m1(op0, op1) \ +__builtin_rvv_vfwcvt_rtz_xu_f_v_u32m1((vfloat16mf2_t)(op0), (size_t)(op1)) +#define vfwcvt_rtz_xu_f_v_u32m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_rtz_xu_f_v_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vfwcvt_rtz_xu_f_v_u32m2(op0, op1) \ +__builtin_rvv_vfwcvt_rtz_xu_f_v_u32m2((vfloat16m1_t)(op0), (size_t)(op1)) +#define vfwcvt_rtz_xu_f_v_u32m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_rtz_xu_f_v_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfwcvt_rtz_xu_f_v_u32m4(op0, op1) \ +__builtin_rvv_vfwcvt_rtz_xu_f_v_u32m4((vfloat16m2_t)(op0), (size_t)(op1)) +#define vfwcvt_rtz_xu_f_v_u32m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_rtz_xu_f_v_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vfwcvt_rtz_xu_f_v_u32m8(op0, op1) \ +__builtin_rvv_vfwcvt_rtz_xu_f_v_u32m8((vfloat16m4_t)(op0), (size_t)(op1)) +#define vfwcvt_rtz_xu_f_v_u32m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_rtz_xu_f_v_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vfwcvt_x_f_v_i32mf2(op0, op1) \ +__builtin_rvv_vfwcvt_x_f_v_i32mf2((vfloat16mf4_t)(op0), (size_t)(op1)) +#define vfwcvt_x_f_v_i32mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_x_f_v_i32mf2_m((vbool64_t)(op0), (vint32mf2_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vfwcvt_x_f_v_i32m1(op0, op1) \ +__builtin_rvv_vfwcvt_x_f_v_i32m1((vfloat16mf2_t)(op0), (size_t)(op1)) +#define vfwcvt_x_f_v_i32m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_x_f_v_i32m1_m((vbool32_t)(op0), (vint32m1_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vfwcvt_x_f_v_i32m2(op0, op1) \ +__builtin_rvv_vfwcvt_x_f_v_i32m2((vfloat16m1_t)(op0), (size_t)(op1)) +#define vfwcvt_x_f_v_i32m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_x_f_v_i32m2_m((vbool16_t)(op0), (vint32m2_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfwcvt_x_f_v_i32m4(op0, op1) \ +__builtin_rvv_vfwcvt_x_f_v_i32m4((vfloat16m2_t)(op0), (size_t)(op1)) +#define vfwcvt_x_f_v_i32m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_x_f_v_i32m4_m((vbool8_t)(op0), (vint32m4_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vfwcvt_x_f_v_i32m8(op0, op1) \ +__builtin_rvv_vfwcvt_x_f_v_i32m8((vfloat16m4_t)(op0), (size_t)(op1)) +#define vfwcvt_x_f_v_i32m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_x_f_v_i32m8_m((vbool4_t)(op0), (vint32m8_t)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vfwcvt_xu_f_v_u32mf2(op0, op1) \ +__builtin_rvv_vfwcvt_xu_f_v_u32mf2((vfloat16mf4_t)(op0), (size_t)(op1)) +#define vfwcvt_xu_f_v_u32mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_xu_f_v_u32mf2_m((vbool64_t)(op0), (vuint32mf2_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vfwcvt_xu_f_v_u32m1(op0, op1) \ +__builtin_rvv_vfwcvt_xu_f_v_u32m1((vfloat16mf2_t)(op0), (size_t)(op1)) +#define vfwcvt_xu_f_v_u32m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_xu_f_v_u32m1_m((vbool32_t)(op0), (vuint32m1_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vfwcvt_xu_f_v_u32m2(op0, op1) \ +__builtin_rvv_vfwcvt_xu_f_v_u32m2((vfloat16m1_t)(op0), (size_t)(op1)) +#define vfwcvt_xu_f_v_u32m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_xu_f_v_u32m2_m((vbool16_t)(op0), (vuint32m2_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfwcvt_xu_f_v_u32m4(op0, op1) \ +__builtin_rvv_vfwcvt_xu_f_v_u32m4((vfloat16m2_t)(op0), (size_t)(op1)) +#define vfwcvt_xu_f_v_u32m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_xu_f_v_u32m4_m((vbool8_t)(op0), (vuint32m4_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vfwcvt_xu_f_v_u32m8(op0, op1) \ +__builtin_rvv_vfwcvt_xu_f_v_u32m8((vfloat16m4_t)(op0), (size_t)(op1)) +#define vfwcvt_xu_f_v_u32m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_xu_f_v_u32m8_m((vbool4_t)(op0), (vuint32m8_t)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vlmul_ext_v_f16mf2_f16m1(op0) \ +__builtin_rvv_vlmul_ext_v_f16mf2_f16m1((vfloat16mf2_t)(op0)) +#define vlmul_ext_v_f16mf4_f16m1(op0) \ +__builtin_rvv_vlmul_ext_v_f16mf4_f16m1((vfloat16mf4_t)(op0)) +#define vlmul_ext_v_f16mf4_f16mf2(op0) \ +__builtin_rvv_vlmul_ext_v_f16mf4_f16mf2((vfloat16mf4_t)(op0)) +#define vlmul_ext_v_f16m1_f16m2(op0) \ +__builtin_rvv_vlmul_ext_v_f16m1_f16m2((vfloat16m1_t)(op0)) +#define vlmul_ext_v_f16mf2_f16m2(op0) \ +__builtin_rvv_vlmul_ext_v_f16mf2_f16m2((vfloat16mf2_t)(op0)) +#define vlmul_ext_v_f16mf4_f16m2(op0) \ +__builtin_rvv_vlmul_ext_v_f16mf4_f16m2((vfloat16mf4_t)(op0)) +#define vlmul_ext_v_f16m1_f16m4(op0) \ +__builtin_rvv_vlmul_ext_v_f16m1_f16m4((vfloat16m1_t)(op0)) +#define vlmul_ext_v_f16m2_f16m4(op0) \ +__builtin_rvv_vlmul_ext_v_f16m2_f16m4((vfloat16m2_t)(op0)) +#define vlmul_ext_v_f16mf2_f16m4(op0) \ +__builtin_rvv_vlmul_ext_v_f16mf2_f16m4((vfloat16mf2_t)(op0)) +#define vlmul_ext_v_f16mf4_f16m4(op0) \ +__builtin_rvv_vlmul_ext_v_f16mf4_f16m4((vfloat16mf4_t)(op0)) +#define vlmul_ext_v_f16m1_f16m8(op0) \ +__builtin_rvv_vlmul_ext_v_f16m1_f16m8((vfloat16m1_t)(op0)) +#define vlmul_ext_v_f16m2_f16m8(op0) \ +__builtin_rvv_vlmul_ext_v_f16m2_f16m8((vfloat16m2_t)(op0)) +#define vlmul_ext_v_f16m4_f16m8(op0) \ +__builtin_rvv_vlmul_ext_v_f16m4_f16m8((vfloat16m4_t)(op0)) +#define vlmul_ext_v_f16mf2_f16m8(op0) \ +__builtin_rvv_vlmul_ext_v_f16mf2_f16m8((vfloat16mf2_t)(op0)) +#define vlmul_ext_v_f16mf4_f16m8(op0) \ +__builtin_rvv_vlmul_ext_v_f16mf4_f16m8((vfloat16mf4_t)(op0)) +#define vlmul_trunc_v_f16m1_f16mf2(op0) \ +__builtin_rvv_vlmul_trunc_v_f16m1_f16mf2((vfloat16m1_t)(op0)) +#define vlmul_trunc_v_f16m2_f16mf2(op0) \ +__builtin_rvv_vlmul_trunc_v_f16m2_f16mf2((vfloat16m2_t)(op0)) +#define vlmul_trunc_v_f16m4_f16mf2(op0) \ +__builtin_rvv_vlmul_trunc_v_f16m4_f16mf2((vfloat16m4_t)(op0)) +#define vlmul_trunc_v_f16m8_f16mf2(op0) \ +__builtin_rvv_vlmul_trunc_v_f16m8_f16mf2((vfloat16m8_t)(op0)) +#define vlmul_trunc_v_f16m1_f16mf4(op0) \ +__builtin_rvv_vlmul_trunc_v_f16m1_f16mf4((vfloat16m1_t)(op0)) +#define vlmul_trunc_v_f16m2_f16mf4(op0) \ +__builtin_rvv_vlmul_trunc_v_f16m2_f16mf4((vfloat16m2_t)(op0)) +#define vlmul_trunc_v_f16m4_f16mf4(op0) \ +__builtin_rvv_vlmul_trunc_v_f16m4_f16mf4((vfloat16m4_t)(op0)) +#define vlmul_trunc_v_f16m8_f16mf4(op0) \ +__builtin_rvv_vlmul_trunc_v_f16m8_f16mf4((vfloat16m8_t)(op0)) +#define vlmul_trunc_v_f16mf2_f16mf4(op0) \ +__builtin_rvv_vlmul_trunc_v_f16mf2_f16mf4((vfloat16mf2_t)(op0)) +#define vlmul_trunc_v_f16m2_f16m1(op0) \ +__builtin_rvv_vlmul_trunc_v_f16m2_f16m1((vfloat16m2_t)(op0)) +#define vlmul_trunc_v_f16m4_f16m1(op0) \ +__builtin_rvv_vlmul_trunc_v_f16m4_f16m1((vfloat16m4_t)(op0)) +#define vlmul_trunc_v_f16m8_f16m1(op0) \ +__builtin_rvv_vlmul_trunc_v_f16m8_f16m1((vfloat16m8_t)(op0)) +#define vlmul_trunc_v_f16m4_f16m2(op0) \ +__builtin_rvv_vlmul_trunc_v_f16m4_f16m2((vfloat16m4_t)(op0)) +#define vlmul_trunc_v_f16m8_f16m2(op0) \ +__builtin_rvv_vlmul_trunc_v_f16m8_f16m2((vfloat16m8_t)(op0)) +#define vlmul_trunc_v_f16m8_f16m4(op0) \ +__builtin_rvv_vlmul_trunc_v_f16m8_f16m4((vfloat16m8_t)(op0)) +#define vreinterpret_v_i16m1_f16m1(op0) \ +__builtin_rvv_vreinterpret_v_i16m1_f16m1((vint16m1_t)(op0)) +#define vreinterpret_v_i16m2_f16m2(op0) \ +__builtin_rvv_vreinterpret_v_i16m2_f16m2((vint16m2_t)(op0)) +#define vreinterpret_v_i16m4_f16m4(op0) \ +__builtin_rvv_vreinterpret_v_i16m4_f16m4((vint16m4_t)(op0)) +#define vreinterpret_v_i16m8_f16m8(op0) \ +__builtin_rvv_vreinterpret_v_i16m8_f16m8((vint16m8_t)(op0)) +#define vreinterpret_v_i16mf2_f16mf2(op0) \ +__builtin_rvv_vreinterpret_v_i16mf2_f16mf2((vint16mf2_t)(op0)) +#define vreinterpret_v_i16mf4_f16mf4(op0) \ +__builtin_rvv_vreinterpret_v_i16mf4_f16mf4((vint16mf4_t)(op0)) +#define vreinterpret_v_u16m1_f16m1(op0) \ +__builtin_rvv_vreinterpret_v_u16m1_f16m1((vuint16m1_t)(op0)) +#define vreinterpret_v_u16m2_f16m2(op0) \ +__builtin_rvv_vreinterpret_v_u16m2_f16m2((vuint16m2_t)(op0)) +#define vreinterpret_v_u16m4_f16m4(op0) \ +__builtin_rvv_vreinterpret_v_u16m4_f16m4((vuint16m4_t)(op0)) +#define vreinterpret_v_u16m8_f16m8(op0) \ +__builtin_rvv_vreinterpret_v_u16m8_f16m8((vuint16m8_t)(op0)) +#define vreinterpret_v_u16mf2_f16mf2(op0) \ +__builtin_rvv_vreinterpret_v_u16mf2_f16mf2((vuint16mf2_t)(op0)) +#define vreinterpret_v_u16mf4_f16mf4(op0) \ +__builtin_rvv_vreinterpret_v_u16mf4_f16mf4((vuint16mf4_t)(op0)) +#define vreinterpret_v_f16m1_i16m1(op0) \ +__builtin_rvv_vreinterpret_v_f16m1_i16m1((vfloat16m1_t)(op0)) +#define vreinterpret_v_f16m2_i16m2(op0) \ +__builtin_rvv_vreinterpret_v_f16m2_i16m2((vfloat16m2_t)(op0)) +#define vreinterpret_v_f16m4_i16m4(op0) \ +__builtin_rvv_vreinterpret_v_f16m4_i16m4((vfloat16m4_t)(op0)) +#define vreinterpret_v_f16m8_i16m8(op0) \ +__builtin_rvv_vreinterpret_v_f16m8_i16m8((vfloat16m8_t)(op0)) +#define vreinterpret_v_f16mf2_i16mf2(op0) \ +__builtin_rvv_vreinterpret_v_f16mf2_i16mf2((vfloat16mf2_t)(op0)) +#define vreinterpret_v_f16mf4_i16mf4(op0) \ +__builtin_rvv_vreinterpret_v_f16mf4_i16mf4((vfloat16mf4_t)(op0)) +#define vreinterpret_v_f16m1_u16m1(op0) \ +__builtin_rvv_vreinterpret_v_f16m1_u16m1((vfloat16m1_t)(op0)) +#define vreinterpret_v_f16m2_u16m2(op0) \ +__builtin_rvv_vreinterpret_v_f16m2_u16m2((vfloat16m2_t)(op0)) +#define vreinterpret_v_f16m4_u16m4(op0) \ +__builtin_rvv_vreinterpret_v_f16m4_u16m4((vfloat16m4_t)(op0)) +#define vreinterpret_v_f16m8_u16m8(op0) \ +__builtin_rvv_vreinterpret_v_f16m8_u16m8((vfloat16m8_t)(op0)) +#define vreinterpret_v_f16mf2_u16mf2(op0) \ +__builtin_rvv_vreinterpret_v_f16mf2_u16mf2((vfloat16mf2_t)(op0)) +#define vreinterpret_v_f16mf4_u16mf4(op0) \ +__builtin_rvv_vreinterpret_v_f16mf4_u16mf4((vfloat16mf4_t)(op0)) +#define vundefined_f16m1() \ +__builtin_rvv_vundefined_f16m1() +#define vundefined_f16m2() \ +__builtin_rvv_vundefined_f16m2() +#define vundefined_f16m4() \ +__builtin_rvv_vundefined_f16m4() +#define vundefined_f16m8() \ +__builtin_rvv_vundefined_f16m8() +#define vundefined_f16mf2() \ +__builtin_rvv_vundefined_f16mf2() +#define vundefined_f16mf4() \ +__builtin_rvv_vundefined_f16mf4() +#endif + +#if defined(__riscv_f) && defined(__riscv_zfh) +#define vfwadd_vv_f32mf2(op0, op1, op2) \ +__builtin_rvv_vfwadd_vv_f32mf2((vfloat16mf4_t)(op0), (vfloat16mf4_t)(op1), (size_t)(op2)) +#define vfwadd_vv_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwadd_vv_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vfwadd_vv_f32m1(op0, op1, op2) \ +__builtin_rvv_vfwadd_vv_f32m1((vfloat16mf2_t)(op0), (vfloat16mf2_t)(op1), (size_t)(op2)) +#define vfwadd_vv_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwadd_vv_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vfwadd_vv_f32m2(op0, op1, op2) \ +__builtin_rvv_vfwadd_vv_f32m2((vfloat16m1_t)(op0), (vfloat16m1_t)(op1), (size_t)(op2)) +#define vfwadd_vv_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwadd_vv_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfwadd_vv_f32m4(op0, op1, op2) \ +__builtin_rvv_vfwadd_vv_f32m4((vfloat16m2_t)(op0), (vfloat16m2_t)(op1), (size_t)(op2)) +#define vfwadd_vv_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwadd_vv_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vfwadd_vv_f32m8(op0, op1, op2) \ +__builtin_rvv_vfwadd_vv_f32m8((vfloat16m4_t)(op0), (vfloat16m4_t)(op1), (size_t)(op2)) +#define vfwadd_vv_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwadd_vv_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat16m4_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vfwadd_vf_f32mf2(op0, op1, op2) \ +__builtin_rvv_vfwadd_vf_f32mf2((vfloat16mf4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfwadd_vf_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwadd_vf_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat16mf4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfwadd_vf_f32m1(op0, op1, op2) \ +__builtin_rvv_vfwadd_vf_f32m1((vfloat16mf2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfwadd_vf_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwadd_vf_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat16mf2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfwadd_vf_f32m2(op0, op1, op2) \ +__builtin_rvv_vfwadd_vf_f32m2((vfloat16m1_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfwadd_vf_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwadd_vf_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat16m1_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfwadd_vf_f32m4(op0, op1, op2) \ +__builtin_rvv_vfwadd_vf_f32m4((vfloat16m2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfwadd_vf_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwadd_vf_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat16m2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfwadd_vf_f32m8(op0, op1, op2) \ +__builtin_rvv_vfwadd_vf_f32m8((vfloat16m4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfwadd_vf_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwadd_vf_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat16m4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfwsub_vv_f32mf2(op0, op1, op2) \ +__builtin_rvv_vfwsub_vv_f32mf2((vfloat16mf4_t)(op0), (vfloat16mf4_t)(op1), (size_t)(op2)) +#define vfwsub_vv_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwsub_vv_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vfwsub_vv_f32m1(op0, op1, op2) \ +__builtin_rvv_vfwsub_vv_f32m1((vfloat16mf2_t)(op0), (vfloat16mf2_t)(op1), (size_t)(op2)) +#define vfwsub_vv_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwsub_vv_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vfwsub_vv_f32m2(op0, op1, op2) \ +__builtin_rvv_vfwsub_vv_f32m2((vfloat16m1_t)(op0), (vfloat16m1_t)(op1), (size_t)(op2)) +#define vfwsub_vv_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwsub_vv_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfwsub_vv_f32m4(op0, op1, op2) \ +__builtin_rvv_vfwsub_vv_f32m4((vfloat16m2_t)(op0), (vfloat16m2_t)(op1), (size_t)(op2)) +#define vfwsub_vv_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwsub_vv_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vfwsub_vv_f32m8(op0, op1, op2) \ +__builtin_rvv_vfwsub_vv_f32m8((vfloat16m4_t)(op0), (vfloat16m4_t)(op1), (size_t)(op2)) +#define vfwsub_vv_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwsub_vv_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat16m4_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vfwsub_vf_f32mf2(op0, op1, op2) \ +__builtin_rvv_vfwsub_vf_f32mf2((vfloat16mf4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfwsub_vf_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwsub_vf_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat16mf4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfwsub_vf_f32m1(op0, op1, op2) \ +__builtin_rvv_vfwsub_vf_f32m1((vfloat16mf2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfwsub_vf_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwsub_vf_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat16mf2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfwsub_vf_f32m2(op0, op1, op2) \ +__builtin_rvv_vfwsub_vf_f32m2((vfloat16m1_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfwsub_vf_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwsub_vf_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat16m1_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfwsub_vf_f32m4(op0, op1, op2) \ +__builtin_rvv_vfwsub_vf_f32m4((vfloat16m2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfwsub_vf_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwsub_vf_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat16m2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfwsub_vf_f32m8(op0, op1, op2) \ +__builtin_rvv_vfwsub_vf_f32m8((vfloat16m4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfwsub_vf_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwsub_vf_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat16m4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfwadd_wv_f32mf2(op0, op1, op2) \ +__builtin_rvv_vfwadd_wv_f32mf2((vfloat32mf2_t)(op0), (vfloat16mf4_t)(op1), (size_t)(op2)) +#define vfwadd_wv_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwadd_wv_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vfwadd_wv_f32m1(op0, op1, op2) \ +__builtin_rvv_vfwadd_wv_f32m1((vfloat32m1_t)(op0), (vfloat16mf2_t)(op1), (size_t)(op2)) +#define vfwadd_wv_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwadd_wv_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vfwadd_wv_f32m2(op0, op1, op2) \ +__builtin_rvv_vfwadd_wv_f32m2((vfloat32m2_t)(op0), (vfloat16m1_t)(op1), (size_t)(op2)) +#define vfwadd_wv_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwadd_wv_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfwadd_wv_f32m4(op0, op1, op2) \ +__builtin_rvv_vfwadd_wv_f32m4((vfloat32m4_t)(op0), (vfloat16m2_t)(op1), (size_t)(op2)) +#define vfwadd_wv_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwadd_wv_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vfwadd_wv_f32m8(op0, op1, op2) \ +__builtin_rvv_vfwadd_wv_f32m8((vfloat32m8_t)(op0), (vfloat16m4_t)(op1), (size_t)(op2)) +#define vfwadd_wv_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwadd_wv_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vfwadd_wf_f32mf2(op0, op1, op2) \ +__builtin_rvv_vfwadd_wf_f32mf2((vfloat32mf2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfwadd_wf_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwadd_wf_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfwadd_wf_f32m1(op0, op1, op2) \ +__builtin_rvv_vfwadd_wf_f32m1((vfloat32m1_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfwadd_wf_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwadd_wf_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfwadd_wf_f32m2(op0, op1, op2) \ +__builtin_rvv_vfwadd_wf_f32m2((vfloat32m2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfwadd_wf_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwadd_wf_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfwadd_wf_f32m4(op0, op1, op2) \ +__builtin_rvv_vfwadd_wf_f32m4((vfloat32m4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfwadd_wf_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwadd_wf_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfwadd_wf_f32m8(op0, op1, op2) \ +__builtin_rvv_vfwadd_wf_f32m8((vfloat32m8_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfwadd_wf_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwadd_wf_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfwsub_wv_f32mf2(op0, op1, op2) \ +__builtin_rvv_vfwsub_wv_f32mf2((vfloat32mf2_t)(op0), (vfloat16mf4_t)(op1), (size_t)(op2)) +#define vfwsub_wv_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwsub_wv_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vfwsub_wv_f32m1(op0, op1, op2) \ +__builtin_rvv_vfwsub_wv_f32m1((vfloat32m1_t)(op0), (vfloat16mf2_t)(op1), (size_t)(op2)) +#define vfwsub_wv_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwsub_wv_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vfwsub_wv_f32m2(op0, op1, op2) \ +__builtin_rvv_vfwsub_wv_f32m2((vfloat32m2_t)(op0), (vfloat16m1_t)(op1), (size_t)(op2)) +#define vfwsub_wv_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwsub_wv_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfwsub_wv_f32m4(op0, op1, op2) \ +__builtin_rvv_vfwsub_wv_f32m4((vfloat32m4_t)(op0), (vfloat16m2_t)(op1), (size_t)(op2)) +#define vfwsub_wv_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwsub_wv_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vfwsub_wv_f32m8(op0, op1, op2) \ +__builtin_rvv_vfwsub_wv_f32m8((vfloat32m8_t)(op0), (vfloat16m4_t)(op1), (size_t)(op2)) +#define vfwsub_wv_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwsub_wv_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vfwsub_wf_f32mf2(op0, op1, op2) \ +__builtin_rvv_vfwsub_wf_f32mf2((vfloat32mf2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfwsub_wf_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwsub_wf_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfwsub_wf_f32m1(op0, op1, op2) \ +__builtin_rvv_vfwsub_wf_f32m1((vfloat32m1_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfwsub_wf_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwsub_wf_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfwsub_wf_f32m2(op0, op1, op2) \ +__builtin_rvv_vfwsub_wf_f32m2((vfloat32m2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfwsub_wf_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwsub_wf_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfwsub_wf_f32m4(op0, op1, op2) \ +__builtin_rvv_vfwsub_wf_f32m4((vfloat32m4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfwsub_wf_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwsub_wf_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfwsub_wf_f32m8(op0, op1, op2) \ +__builtin_rvv_vfwsub_wf_f32m8((vfloat32m8_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfwsub_wf_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwsub_wf_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat32m8_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfwmul_vv_f32mf2(op0, op1, op2) \ +__builtin_rvv_vfwmul_vv_f32mf2((vfloat16mf4_t)(op0), (vfloat16mf4_t)(op1), (size_t)(op2)) +#define vfwmul_vv_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmul_vv_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vfwmul_vv_f32m1(op0, op1, op2) \ +__builtin_rvv_vfwmul_vv_f32m1((vfloat16mf2_t)(op0), (vfloat16mf2_t)(op1), (size_t)(op2)) +#define vfwmul_vv_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmul_vv_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vfwmul_vv_f32m2(op0, op1, op2) \ +__builtin_rvv_vfwmul_vv_f32m2((vfloat16m1_t)(op0), (vfloat16m1_t)(op1), (size_t)(op2)) +#define vfwmul_vv_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmul_vv_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfwmul_vv_f32m4(op0, op1, op2) \ +__builtin_rvv_vfwmul_vv_f32m4((vfloat16m2_t)(op0), (vfloat16m2_t)(op1), (size_t)(op2)) +#define vfwmul_vv_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmul_vv_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vfwmul_vv_f32m8(op0, op1, op2) \ +__builtin_rvv_vfwmul_vv_f32m8((vfloat16m4_t)(op0), (vfloat16m4_t)(op1), (size_t)(op2)) +#define vfwmul_vv_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmul_vv_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat16m4_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vfwmul_vf_f32mf2(op0, op1, op2) \ +__builtin_rvv_vfwmul_vf_f32mf2((vfloat16mf4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfwmul_vf_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmul_vf_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat16mf4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfwmul_vf_f32m1(op0, op1, op2) \ +__builtin_rvv_vfwmul_vf_f32m1((vfloat16mf2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfwmul_vf_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmul_vf_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat16mf2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfwmul_vf_f32m2(op0, op1, op2) \ +__builtin_rvv_vfwmul_vf_f32m2((vfloat16m1_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfwmul_vf_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmul_vf_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat16m1_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfwmul_vf_f32m4(op0, op1, op2) \ +__builtin_rvv_vfwmul_vf_f32m4((vfloat16m2_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfwmul_vf_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmul_vf_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat16m2_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfwmul_vf_f32m8(op0, op1, op2) \ +__builtin_rvv_vfwmul_vf_f32m8((vfloat16m4_t)(op0), (_Float16)(op1), (size_t)(op2)) +#define vfwmul_vf_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmul_vf_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat16m4_t)(op2), (_Float16)(op3), (size_t)(op4)) +#define vfwmacc_vv_f32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vfwmacc_vv_f32mf2((vfloat32mf2_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vfwmacc_vv_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmacc_vv_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vfwmacc_vv_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfwmacc_vv_f32m1((vfloat32m1_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vfwmacc_vv_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmacc_vv_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vfwmacc_vv_f32m2(op0, op1, op2, op3) \ +__builtin_rvv_vfwmacc_vv_f32m2((vfloat32m2_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfwmacc_vv_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmacc_vv_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfwmacc_vv_f32m4(op0, op1, op2, op3) \ +__builtin_rvv_vfwmacc_vv_f32m4((vfloat32m4_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vfwmacc_vv_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmacc_vv_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vfwmacc_vv_f32m8(op0, op1, op2, op3) \ +__builtin_rvv_vfwmacc_vv_f32m8((vfloat32m8_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vfwmacc_vv_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmacc_vv_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat16m4_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vfwmacc_vf_f32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vfwmacc_vf_f32mf2((vfloat32mf2_t)(op0), (_Float16)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vfwmacc_vf_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmacc_vf_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (_Float16)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vfwmacc_vf_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfwmacc_vf_f32m1((vfloat32m1_t)(op0), (_Float16)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vfwmacc_vf_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmacc_vf_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (_Float16)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vfwmacc_vf_f32m2(op0, op1, op2, op3) \ +__builtin_rvv_vfwmacc_vf_f32m2((vfloat32m2_t)(op0), (_Float16)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfwmacc_vf_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmacc_vf_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (_Float16)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfwmacc_vf_f32m4(op0, op1, op2, op3) \ +__builtin_rvv_vfwmacc_vf_f32m4((vfloat32m4_t)(op0), (_Float16)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vfwmacc_vf_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmacc_vf_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (_Float16)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vfwmacc_vf_f32m8(op0, op1, op2, op3) \ +__builtin_rvv_vfwmacc_vf_f32m8((vfloat32m8_t)(op0), (_Float16)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vfwmacc_vf_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmacc_vf_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (_Float16)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vfwnmacc_vv_f32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vfwnmacc_vv_f32mf2((vfloat32mf2_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vfwnmacc_vv_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwnmacc_vv_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vfwnmacc_vv_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfwnmacc_vv_f32m1((vfloat32m1_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vfwnmacc_vv_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwnmacc_vv_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vfwnmacc_vv_f32m2(op0, op1, op2, op3) \ +__builtin_rvv_vfwnmacc_vv_f32m2((vfloat32m2_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfwnmacc_vv_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwnmacc_vv_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfwnmacc_vv_f32m4(op0, op1, op2, op3) \ +__builtin_rvv_vfwnmacc_vv_f32m4((vfloat32m4_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vfwnmacc_vv_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwnmacc_vv_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vfwnmacc_vv_f32m8(op0, op1, op2, op3) \ +__builtin_rvv_vfwnmacc_vv_f32m8((vfloat32m8_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vfwnmacc_vv_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwnmacc_vv_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat16m4_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vfwnmacc_vf_f32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vfwnmacc_vf_f32mf2((vfloat32mf2_t)(op0), (_Float16)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vfwnmacc_vf_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwnmacc_vf_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (_Float16)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vfwnmacc_vf_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfwnmacc_vf_f32m1((vfloat32m1_t)(op0), (_Float16)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vfwnmacc_vf_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwnmacc_vf_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (_Float16)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vfwnmacc_vf_f32m2(op0, op1, op2, op3) \ +__builtin_rvv_vfwnmacc_vf_f32m2((vfloat32m2_t)(op0), (_Float16)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfwnmacc_vf_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwnmacc_vf_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (_Float16)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfwnmacc_vf_f32m4(op0, op1, op2, op3) \ +__builtin_rvv_vfwnmacc_vf_f32m4((vfloat32m4_t)(op0), (_Float16)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vfwnmacc_vf_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwnmacc_vf_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (_Float16)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vfwnmacc_vf_f32m8(op0, op1, op2, op3) \ +__builtin_rvv_vfwnmacc_vf_f32m8((vfloat32m8_t)(op0), (_Float16)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vfwnmacc_vf_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwnmacc_vf_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (_Float16)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vfwmsac_vv_f32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vfwmsac_vv_f32mf2((vfloat32mf2_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vfwmsac_vv_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmsac_vv_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vfwmsac_vv_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfwmsac_vv_f32m1((vfloat32m1_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vfwmsac_vv_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmsac_vv_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vfwmsac_vv_f32m2(op0, op1, op2, op3) \ +__builtin_rvv_vfwmsac_vv_f32m2((vfloat32m2_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfwmsac_vv_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmsac_vv_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfwmsac_vv_f32m4(op0, op1, op2, op3) \ +__builtin_rvv_vfwmsac_vv_f32m4((vfloat32m4_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vfwmsac_vv_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmsac_vv_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vfwmsac_vv_f32m8(op0, op1, op2, op3) \ +__builtin_rvv_vfwmsac_vv_f32m8((vfloat32m8_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vfwmsac_vv_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmsac_vv_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat16m4_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vfwmsac_vf_f32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vfwmsac_vf_f32mf2((vfloat32mf2_t)(op0), (_Float16)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vfwmsac_vf_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmsac_vf_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (_Float16)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vfwmsac_vf_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfwmsac_vf_f32m1((vfloat32m1_t)(op0), (_Float16)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vfwmsac_vf_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmsac_vf_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (_Float16)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vfwmsac_vf_f32m2(op0, op1, op2, op3) \ +__builtin_rvv_vfwmsac_vf_f32m2((vfloat32m2_t)(op0), (_Float16)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfwmsac_vf_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmsac_vf_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (_Float16)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfwmsac_vf_f32m4(op0, op1, op2, op3) \ +__builtin_rvv_vfwmsac_vf_f32m4((vfloat32m4_t)(op0), (_Float16)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vfwmsac_vf_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmsac_vf_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (_Float16)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vfwmsac_vf_f32m8(op0, op1, op2, op3) \ +__builtin_rvv_vfwmsac_vf_f32m8((vfloat32m8_t)(op0), (_Float16)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vfwmsac_vf_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwmsac_vf_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (_Float16)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vfwnmsac_vv_f32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vfwnmsac_vv_f32mf2((vfloat32mf2_t)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vfwnmsac_vv_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwnmsac_vv_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vfwnmsac_vv_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfwnmsac_vv_f32m1((vfloat32m1_t)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vfwnmsac_vv_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwnmsac_vv_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vfwnmsac_vv_f32m2(op0, op1, op2, op3) \ +__builtin_rvv_vfwnmsac_vv_f32m2((vfloat32m2_t)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfwnmsac_vv_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwnmsac_vv_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfwnmsac_vv_f32m4(op0, op1, op2, op3) \ +__builtin_rvv_vfwnmsac_vv_f32m4((vfloat32m4_t)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vfwnmsac_vv_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwnmsac_vv_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vfwnmsac_vv_f32m8(op0, op1, op2, op3) \ +__builtin_rvv_vfwnmsac_vv_f32m8((vfloat32m8_t)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vfwnmsac_vv_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwnmsac_vv_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat16m4_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vfwnmsac_vf_f32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vfwnmsac_vf_f32mf2((vfloat32mf2_t)(op0), (_Float16)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vfwnmsac_vf_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwnmsac_vf_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (_Float16)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vfwnmsac_vf_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfwnmsac_vf_f32m1((vfloat32m1_t)(op0), (_Float16)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vfwnmsac_vf_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwnmsac_vf_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (_Float16)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vfwnmsac_vf_f32m2(op0, op1, op2, op3) \ +__builtin_rvv_vfwnmsac_vf_f32m2((vfloat32m2_t)(op0), (_Float16)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfwnmsac_vf_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwnmsac_vf_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (_Float16)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vfwnmsac_vf_f32m4(op0, op1, op2, op3) \ +__builtin_rvv_vfwnmsac_vf_f32m4((vfloat32m4_t)(op0), (_Float16)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vfwnmsac_vf_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwnmsac_vf_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (_Float16)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vfwnmsac_vf_f32m8(op0, op1, op2, op3) \ +__builtin_rvv_vfwnmsac_vf_f32m8((vfloat32m8_t)(op0), (_Float16)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vfwnmsac_vf_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwnmsac_vf_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (_Float16)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vfwredsum_vs_f16m1_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfwredsum_vs_f16m1_f32m1((vfloat32m1_t)(op0), (vfloat16m1_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfwredsum_vs_f16m1_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwredsum_vs_f16m1_f32m1_m((vbool16_t)(op0), (vfloat32m1_t)(op1), (vfloat16m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfwredsum_vs_f16m2_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfwredsum_vs_f16m2_f32m1((vfloat32m1_t)(op0), (vfloat16m2_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfwredsum_vs_f16m2_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwredsum_vs_f16m2_f32m1_m((vbool8_t)(op0), (vfloat32m1_t)(op1), (vfloat16m2_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfwredsum_vs_f16m4_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfwredsum_vs_f16m4_f32m1((vfloat32m1_t)(op0), (vfloat16m4_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfwredsum_vs_f16m4_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwredsum_vs_f16m4_f32m1_m((vbool4_t)(op0), (vfloat32m1_t)(op1), (vfloat16m4_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfwredsum_vs_f16m8_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfwredsum_vs_f16m8_f32m1((vfloat32m1_t)(op0), (vfloat16m8_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfwredsum_vs_f16m8_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwredsum_vs_f16m8_f32m1_m((vbool2_t)(op0), (vfloat32m1_t)(op1), (vfloat16m8_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfwredsum_vs_f16mf2_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfwredsum_vs_f16mf2_f32m1((vfloat32m1_t)(op0), (vfloat16mf2_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfwredsum_vs_f16mf2_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwredsum_vs_f16mf2_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat16mf2_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfwredsum_vs_f16mf4_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfwredsum_vs_f16mf4_f32m1((vfloat32m1_t)(op0), (vfloat16mf4_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfwredsum_vs_f16mf4_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwredsum_vs_f16mf4_f32m1_m((vbool64_t)(op0), (vfloat32m1_t)(op1), (vfloat16mf4_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfwredosum_vs_f16m1_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfwredosum_vs_f16m1_f32m1((vfloat32m1_t)(op0), (vfloat16m1_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfwredosum_vs_f16m1_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwredosum_vs_f16m1_f32m1_m((vbool16_t)(op0), (vfloat32m1_t)(op1), (vfloat16m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfwredosum_vs_f16m2_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfwredosum_vs_f16m2_f32m1((vfloat32m1_t)(op0), (vfloat16m2_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfwredosum_vs_f16m2_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwredosum_vs_f16m2_f32m1_m((vbool8_t)(op0), (vfloat32m1_t)(op1), (vfloat16m2_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfwredosum_vs_f16m4_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfwredosum_vs_f16m4_f32m1((vfloat32m1_t)(op0), (vfloat16m4_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfwredosum_vs_f16m4_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwredosum_vs_f16m4_f32m1_m((vbool4_t)(op0), (vfloat32m1_t)(op1), (vfloat16m4_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfwredosum_vs_f16m8_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfwredosum_vs_f16m8_f32m1((vfloat32m1_t)(op0), (vfloat16m8_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfwredosum_vs_f16m8_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwredosum_vs_f16m8_f32m1_m((vbool2_t)(op0), (vfloat32m1_t)(op1), (vfloat16m8_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfwredosum_vs_f16mf2_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfwredosum_vs_f16mf2_f32m1((vfloat32m1_t)(op0), (vfloat16mf2_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfwredosum_vs_f16mf2_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwredosum_vs_f16mf2_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat16mf2_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfwredosum_vs_f16mf4_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vfwredosum_vs_f16mf4_f32m1((vfloat32m1_t)(op0), (vfloat16mf4_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfwredosum_vs_f16mf4_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vfwredosum_vs_f16mf4_f32m1_m((vbool64_t)(op0), (vfloat32m1_t)(op1), (vfloat16mf4_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vfncvt_f_f_w_f16mf4(op0, op1) \ +__builtin_rvv_vfncvt_f_f_w_f16mf4((vfloat32mf2_t)(op0), (size_t)(op1)) +#define vfncvt_f_f_w_f16mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_f_f_w_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vfncvt_f_f_w_f16mf2(op0, op1) \ +__builtin_rvv_vfncvt_f_f_w_f16mf2((vfloat32m1_t)(op0), (size_t)(op1)) +#define vfncvt_f_f_w_f16mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_f_f_w_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfncvt_f_f_w_f16m1(op0, op1) \ +__builtin_rvv_vfncvt_f_f_w_f16m1((vfloat32m2_t)(op0), (size_t)(op1)) +#define vfncvt_f_f_w_f16m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_f_f_w_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vfncvt_f_f_w_f16m2(op0, op1) \ +__builtin_rvv_vfncvt_f_f_w_f16m2((vfloat32m4_t)(op0), (size_t)(op1)) +#define vfncvt_f_f_w_f16m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_f_f_w_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vfncvt_f_f_w_f16m4(op0, op1) \ +__builtin_rvv_vfncvt_f_f_w_f16m4((vfloat32m8_t)(op0), (size_t)(op1)) +#define vfncvt_f_f_w_f16m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_f_f_w_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vfloat32m8_t)(op2), (size_t)(op3)) +#define vfncvt_rod_f_f_w_f16mf4(op0, op1) \ +__builtin_rvv_vfncvt_rod_f_f_w_f16mf4((vfloat32mf2_t)(op0), (size_t)(op1)) +#define vfncvt_rod_f_f_w_f16mf4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_rod_f_f_w_f16mf4_m((vbool64_t)(op0), (vfloat16mf4_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vfncvt_rod_f_f_w_f16mf2(op0, op1) \ +__builtin_rvv_vfncvt_rod_f_f_w_f16mf2((vfloat32m1_t)(op0), (size_t)(op1)) +#define vfncvt_rod_f_f_w_f16mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_rod_f_f_w_f16mf2_m((vbool32_t)(op0), (vfloat16mf2_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vfncvt_rod_f_f_w_f16m1(op0, op1) \ +__builtin_rvv_vfncvt_rod_f_f_w_f16m1((vfloat32m2_t)(op0), (size_t)(op1)) +#define vfncvt_rod_f_f_w_f16m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_rod_f_f_w_f16m1_m((vbool16_t)(op0), (vfloat16m1_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vfncvt_rod_f_f_w_f16m2(op0, op1) \ +__builtin_rvv_vfncvt_rod_f_f_w_f16m2((vfloat32m4_t)(op0), (size_t)(op1)) +#define vfncvt_rod_f_f_w_f16m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_rod_f_f_w_f16m2_m((vbool8_t)(op0), (vfloat16m2_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vfncvt_rod_f_f_w_f16m4(op0, op1) \ +__builtin_rvv_vfncvt_rod_f_f_w_f16m4((vfloat32m8_t)(op0), (size_t)(op1)) +#define vfncvt_rod_f_f_w_f16m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfncvt_rod_f_f_w_f16m4_m((vbool4_t)(op0), (vfloat16m4_t)(op1), (vfloat32m8_t)(op2), (size_t)(op3)) +#define vfwcvt_f_f_v_f32mf2(op0, op1) \ +__builtin_rvv_vfwcvt_f_f_v_f32mf2((vfloat16mf4_t)(op0), (size_t)(op1)) +#define vfwcvt_f_f_v_f32mf2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_f_f_v_f32mf2_m((vbool64_t)(op0), (vfloat32mf2_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vfwcvt_f_f_v_f32m1(op0, op1) \ +__builtin_rvv_vfwcvt_f_f_v_f32m1((vfloat16mf2_t)(op0), (size_t)(op1)) +#define vfwcvt_f_f_v_f32m1_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_f_f_v_f32m1_m((vbool32_t)(op0), (vfloat32m1_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vfwcvt_f_f_v_f32m2(op0, op1) \ +__builtin_rvv_vfwcvt_f_f_v_f32m2((vfloat16m1_t)(op0), (size_t)(op1)) +#define vfwcvt_f_f_v_f32m2_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_f_f_v_f32m2_m((vbool16_t)(op0), (vfloat32m2_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vfwcvt_f_f_v_f32m4(op0, op1) \ +__builtin_rvv_vfwcvt_f_f_v_f32m4((vfloat16m2_t)(op0), (size_t)(op1)) +#define vfwcvt_f_f_v_f32m4_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_f_f_v_f32m4_m((vbool8_t)(op0), (vfloat32m4_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vfwcvt_f_f_v_f32m8(op0, op1) \ +__builtin_rvv_vfwcvt_f_f_v_f32m8((vfloat16m4_t)(op0), (size_t)(op1)) +#define vfwcvt_f_f_v_f32m8_m(op0, op1, op2, op3) \ +__builtin_rvv_vfwcvt_f_f_v_f32m8_m((vbool4_t)(op0), (vfloat32m8_t)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#endif + +#if defined(__riscv_zvamo) +#define vamoswapei16_v_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei16_v_i32m1((int32_t *)(op0), (vuint16mf2_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vamoswapei16_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei16_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint16mf2_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vamoswapei16_v_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei16_v_i32m2((int32_t *)(op0), (vuint16m1_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vamoswapei16_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei16_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint16m1_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vamoswapei16_v_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei16_v_i32m4((int32_t *)(op0), (vuint16m2_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vamoswapei16_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei16_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vuint16m2_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vamoswapei16_v_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei16_v_i32m8((int32_t *)(op0), (vuint16m4_t)(op1), (vint32m8_t)(op2), (size_t)(op3)) +#define vamoswapei16_v_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei16_v_i32m8_m((vbool4_t)(op0), (int32_t *)(op1), (vuint16m4_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vamoswapei16_v_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei16_v_i32mf2((int32_t *)(op0), (vuint16mf4_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vamoswapei16_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei16_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint16mf4_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vamoswapei16_v_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei16_v_u32m1((uint32_t *)(op0), (vuint16mf2_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vamoswapei16_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei16_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint16mf2_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vamoswapei16_v_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei16_v_u32m2((uint32_t *)(op0), (vuint16m1_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vamoswapei16_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei16_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint16m1_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vamoswapei16_v_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei16_v_u32m4((uint32_t *)(op0), (vuint16m2_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vamoswapei16_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei16_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint16m2_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vamoswapei16_v_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei16_v_u32m8((uint32_t *)(op0), (vuint16m4_t)(op1), (vuint32m8_t)(op2), (size_t)(op3)) +#define vamoswapei16_v_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei16_v_u32m8_m((vbool4_t)(op0), (uint32_t *)(op1), (vuint16m4_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vamoswapei16_v_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei16_v_u32mf2((uint32_t *)(op0), (vuint16mf4_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vamoswapei16_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei16_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint16mf4_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vamoswapei32_v_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei32_v_i32m1((int32_t *)(op0), (vuint32m1_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vamoswapei32_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei32_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vamoswapei32_v_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei32_v_i32m2((int32_t *)(op0), (vuint32m2_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vamoswapei32_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei32_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vamoswapei32_v_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei32_v_i32m4((int32_t *)(op0), (vuint32m4_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vamoswapei32_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei32_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vuint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vamoswapei32_v_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei32_v_i32m8((int32_t *)(op0), (vuint32m8_t)(op1), (vint32m8_t)(op2), (size_t)(op3)) +#define vamoswapei32_v_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei32_v_i32m8_m((vbool4_t)(op0), (int32_t *)(op1), (vuint32m8_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vamoswapei32_v_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei32_v_i32mf2((int32_t *)(op0), (vuint32mf2_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vamoswapei32_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei32_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vamoswapei32_v_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei32_v_u32m1((uint32_t *)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vamoswapei32_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei32_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vamoswapei32_v_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei32_v_u32m2((uint32_t *)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vamoswapei32_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei32_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vamoswapei32_v_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei32_v_u32m4((uint32_t *)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vamoswapei32_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei32_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vamoswapei32_v_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei32_v_u32m8((uint32_t *)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (size_t)(op3)) +#define vamoswapei32_v_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei32_v_u32m8_m((vbool4_t)(op0), (uint32_t *)(op1), (vuint32m8_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vamoswapei32_v_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei32_v_u32mf2((uint32_t *)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vamoswapei32_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei32_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vamoswapei64_v_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei64_v_i32m1((int32_t *)(op0), (vuint64m2_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vamoswapei64_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei64_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint64m2_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vamoswapei64_v_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei64_v_i32m2((int32_t *)(op0), (vuint64m4_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vamoswapei64_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei64_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint64m4_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vamoswapei64_v_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei64_v_i32m4((int32_t *)(op0), (vuint64m8_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vamoswapei64_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei64_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vuint64m8_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vamoswapei64_v_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei64_v_i32mf2((int32_t *)(op0), (vuint64m1_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vamoswapei64_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei64_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint64m1_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vamoswapei64_v_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei64_v_u32m1((uint32_t *)(op0), (vuint64m2_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vamoswapei64_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei64_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint64m2_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vamoswapei64_v_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei64_v_u32m2((uint32_t *)(op0), (vuint64m4_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vamoswapei64_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei64_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint64m4_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vamoswapei64_v_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei64_v_u32m4((uint32_t *)(op0), (vuint64m8_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vamoswapei64_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei64_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint64m8_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vamoswapei64_v_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei64_v_u32mf2((uint32_t *)(op0), (vuint64m1_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vamoswapei64_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei64_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint64m1_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vamoswapei8_v_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei8_v_i64m1((int64_t *)(op0), (vuint8mf8_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vamoswapei8_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei8_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint8mf8_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vamoswapei8_v_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei8_v_i64m2((int64_t *)(op0), (vuint8mf4_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vamoswapei8_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei8_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint8mf4_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vamoswapei8_v_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei8_v_i64m4((int64_t *)(op0), (vuint8mf2_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vamoswapei8_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei8_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vuint8mf2_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vamoswapei8_v_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei8_v_i64m8((int64_t *)(op0), (vuint8m1_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vamoswapei8_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei8_v_i64m8_m((vbool8_t)(op0), (int64_t *)(op1), (vuint8m1_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vamoswapei8_v_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei8_v_u64m1((uint64_t *)(op0), (vuint8mf8_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vamoswapei8_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei8_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint8mf8_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vamoswapei8_v_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei8_v_u64m2((uint64_t *)(op0), (vuint8mf4_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vamoswapei8_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei8_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint8mf4_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vamoswapei8_v_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei8_v_u64m4((uint64_t *)(op0), (vuint8mf2_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vamoswapei8_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei8_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint8mf2_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vamoswapei8_v_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei8_v_u64m8((uint64_t *)(op0), (vuint8m1_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vamoswapei8_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei8_v_u64m8_m((vbool8_t)(op0), (uint64_t *)(op1), (vuint8m1_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vamoswapei16_v_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei16_v_i64m1((int64_t *)(op0), (vuint16mf4_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vamoswapei16_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei16_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint16mf4_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vamoswapei16_v_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei16_v_i64m2((int64_t *)(op0), (vuint16mf2_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vamoswapei16_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei16_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint16mf2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vamoswapei16_v_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei16_v_i64m4((int64_t *)(op0), (vuint16m1_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vamoswapei16_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei16_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vuint16m1_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vamoswapei16_v_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei16_v_i64m8((int64_t *)(op0), (vuint16m2_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vamoswapei16_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei16_v_i64m8_m((vbool8_t)(op0), (int64_t *)(op1), (vuint16m2_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vamoswapei16_v_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei16_v_u64m1((uint64_t *)(op0), (vuint16mf4_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vamoswapei16_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei16_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint16mf4_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vamoswapei16_v_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei16_v_u64m2((uint64_t *)(op0), (vuint16mf2_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vamoswapei16_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei16_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint16mf2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vamoswapei16_v_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei16_v_u64m4((uint64_t *)(op0), (vuint16m1_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vamoswapei16_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei16_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint16m1_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vamoswapei16_v_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei16_v_u64m8((uint64_t *)(op0), (vuint16m2_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vamoswapei16_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei16_v_u64m8_m((vbool8_t)(op0), (uint64_t *)(op1), (vuint16m2_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vamoswapei32_v_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei32_v_i64m1((int64_t *)(op0), (vuint32mf2_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vamoswapei32_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei32_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint32mf2_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vamoswapei32_v_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei32_v_i64m2((int64_t *)(op0), (vuint32m1_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vamoswapei32_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei32_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint32m1_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vamoswapei32_v_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei32_v_i64m4((int64_t *)(op0), (vuint32m2_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vamoswapei32_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei32_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vuint32m2_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vamoswapei32_v_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei32_v_i64m8((int64_t *)(op0), (vuint32m4_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vamoswapei32_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei32_v_i64m8_m((vbool8_t)(op0), (int64_t *)(op1), (vuint32m4_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vamoswapei32_v_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei32_v_u64m1((uint64_t *)(op0), (vuint32mf2_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vamoswapei32_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei32_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint32mf2_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vamoswapei32_v_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei32_v_u64m2((uint64_t *)(op0), (vuint32m1_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vamoswapei32_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei32_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint32m1_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vamoswapei32_v_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei32_v_u64m4((uint64_t *)(op0), (vuint32m2_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vamoswapei32_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei32_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint32m2_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vamoswapei32_v_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei32_v_u64m8((uint64_t *)(op0), (vuint32m4_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vamoswapei32_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei32_v_u64m8_m((vbool8_t)(op0), (uint64_t *)(op1), (vuint32m4_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vamoswapei64_v_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei64_v_i64m1((int64_t *)(op0), (vuint64m1_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vamoswapei64_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei64_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vamoswapei64_v_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei64_v_i64m2((int64_t *)(op0), (vuint64m2_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vamoswapei64_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei64_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vamoswapei64_v_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei64_v_i64m4((int64_t *)(op0), (vuint64m4_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vamoswapei64_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei64_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vuint64m4_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vamoswapei64_v_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei64_v_i64m8((int64_t *)(op0), (vuint64m8_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vamoswapei64_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei64_v_i64m8_m((vbool8_t)(op0), (int64_t *)(op1), (vuint64m8_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vamoswapei64_v_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei64_v_u64m1((uint64_t *)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vamoswapei64_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei64_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vamoswapei64_v_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei64_v_u64m2((uint64_t *)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vamoswapei64_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei64_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vamoswapei64_v_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei64_v_u64m4((uint64_t *)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vamoswapei64_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei64_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vamoswapei64_v_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei64_v_u64m8((uint64_t *)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vamoswapei64_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei64_v_u64m8_m((vbool8_t)(op0), (uint64_t *)(op1), (vuint64m8_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vamoaddei8_v_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei8_v_i32m1((int32_t *)(op0), (vuint8mf4_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vamoaddei8_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei8_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint8mf4_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vamoaddei8_v_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei8_v_i32m2((int32_t *)(op0), (vuint8mf2_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vamoaddei8_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei8_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint8mf2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vamoaddei8_v_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei8_v_i32m4((int32_t *)(op0), (vuint8m1_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vamoaddei8_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei8_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vuint8m1_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vamoaddei8_v_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei8_v_i32m8((int32_t *)(op0), (vuint8m2_t)(op1), (vint32m8_t)(op2), (size_t)(op3)) +#define vamoaddei8_v_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei8_v_i32m8_m((vbool4_t)(op0), (int32_t *)(op1), (vuint8m2_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vamoaddei8_v_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei8_v_i32mf2((int32_t *)(op0), (vuint8mf8_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vamoaddei8_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei8_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint8mf8_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vamoaddei8_v_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei8_v_u32m1((uint32_t *)(op0), (vuint8mf4_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vamoaddei8_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei8_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint8mf4_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vamoaddei8_v_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei8_v_u32m2((uint32_t *)(op0), (vuint8mf2_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vamoaddei8_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei8_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint8mf2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vamoaddei8_v_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei8_v_u32m4((uint32_t *)(op0), (vuint8m1_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vamoaddei8_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei8_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint8m1_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vamoaddei8_v_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei8_v_u32m8((uint32_t *)(op0), (vuint8m2_t)(op1), (vuint32m8_t)(op2), (size_t)(op3)) +#define vamoaddei8_v_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei8_v_u32m8_m((vbool4_t)(op0), (uint32_t *)(op1), (vuint8m2_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vamoaddei8_v_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei8_v_u32mf2((uint32_t *)(op0), (vuint8mf8_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vamoaddei8_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei8_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint8mf8_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vamoaddei16_v_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei16_v_i32m1((int32_t *)(op0), (vuint16mf2_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vamoaddei16_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei16_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint16mf2_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vamoaddei16_v_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei16_v_i32m2((int32_t *)(op0), (vuint16m1_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vamoaddei16_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei16_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint16m1_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vamoaddei16_v_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei16_v_i32m4((int32_t *)(op0), (vuint16m2_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vamoaddei16_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei16_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vuint16m2_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vamoaddei16_v_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei16_v_i32m8((int32_t *)(op0), (vuint16m4_t)(op1), (vint32m8_t)(op2), (size_t)(op3)) +#define vamoaddei16_v_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei16_v_i32m8_m((vbool4_t)(op0), (int32_t *)(op1), (vuint16m4_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vamoaddei16_v_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei16_v_i32mf2((int32_t *)(op0), (vuint16mf4_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vamoaddei16_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei16_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint16mf4_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vamoaddei16_v_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei16_v_u32m1((uint32_t *)(op0), (vuint16mf2_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vamoaddei16_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei16_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint16mf2_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vamoaddei16_v_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei16_v_u32m2((uint32_t *)(op0), (vuint16m1_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vamoaddei16_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei16_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint16m1_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vamoaddei16_v_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei16_v_u32m4((uint32_t *)(op0), (vuint16m2_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vamoaddei16_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei16_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint16m2_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vamoaddei16_v_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei16_v_u32m8((uint32_t *)(op0), (vuint16m4_t)(op1), (vuint32m8_t)(op2), (size_t)(op3)) +#define vamoaddei16_v_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei16_v_u32m8_m((vbool4_t)(op0), (uint32_t *)(op1), (vuint16m4_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vamoaddei16_v_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei16_v_u32mf2((uint32_t *)(op0), (vuint16mf4_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vamoaddei16_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei16_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint16mf4_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vamoaddei32_v_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei32_v_i32m1((int32_t *)(op0), (vuint32m1_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vamoaddei32_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei32_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vamoaddei32_v_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei32_v_i32m2((int32_t *)(op0), (vuint32m2_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vamoaddei32_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei32_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vamoaddei32_v_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei32_v_i32m4((int32_t *)(op0), (vuint32m4_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vamoaddei32_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei32_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vuint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vamoaddei32_v_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei32_v_i32m8((int32_t *)(op0), (vuint32m8_t)(op1), (vint32m8_t)(op2), (size_t)(op3)) +#define vamoaddei32_v_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei32_v_i32m8_m((vbool4_t)(op0), (int32_t *)(op1), (vuint32m8_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vamoaddei32_v_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei32_v_i32mf2((int32_t *)(op0), (vuint32mf2_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vamoaddei32_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei32_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vamoaddei32_v_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei32_v_u32m1((uint32_t *)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vamoaddei32_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei32_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vamoaddei32_v_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei32_v_u32m2((uint32_t *)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vamoaddei32_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei32_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vamoaddei32_v_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei32_v_u32m4((uint32_t *)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vamoaddei32_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei32_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vamoaddei32_v_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei32_v_u32m8((uint32_t *)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (size_t)(op3)) +#define vamoaddei32_v_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei32_v_u32m8_m((vbool4_t)(op0), (uint32_t *)(op1), (vuint32m8_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vamoaddei32_v_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei32_v_u32mf2((uint32_t *)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vamoaddei32_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei32_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vamoaddei64_v_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei64_v_i32m1((int32_t *)(op0), (vuint64m2_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vamoaddei64_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei64_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint64m2_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vamoaddei64_v_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei64_v_i32m2((int32_t *)(op0), (vuint64m4_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vamoaddei64_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei64_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint64m4_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vamoaddei64_v_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei64_v_i32m4((int32_t *)(op0), (vuint64m8_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vamoaddei64_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei64_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vuint64m8_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vamoaddei64_v_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei64_v_i32mf2((int32_t *)(op0), (vuint64m1_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vamoaddei64_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei64_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint64m1_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vamoaddei64_v_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei64_v_u32m1((uint32_t *)(op0), (vuint64m2_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vamoaddei64_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei64_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint64m2_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vamoaddei64_v_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei64_v_u32m2((uint32_t *)(op0), (vuint64m4_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vamoaddei64_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei64_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint64m4_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vamoaddei64_v_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei64_v_u32m4((uint32_t *)(op0), (vuint64m8_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vamoaddei64_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei64_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint64m8_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vamoaddei64_v_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei64_v_u32mf2((uint32_t *)(op0), (vuint64m1_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vamoaddei64_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei64_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint64m1_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vamoaddei8_v_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei8_v_i64m1((int64_t *)(op0), (vuint8mf8_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vamoaddei8_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei8_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint8mf8_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vamoaddei8_v_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei8_v_i64m2((int64_t *)(op0), (vuint8mf4_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vamoaddei8_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei8_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint8mf4_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vamoaddei8_v_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei8_v_i64m4((int64_t *)(op0), (vuint8mf2_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vamoaddei8_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei8_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vuint8mf2_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vamoaddei8_v_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei8_v_i64m8((int64_t *)(op0), (vuint8m1_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vamoaddei8_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei8_v_i64m8_m((vbool8_t)(op0), (int64_t *)(op1), (vuint8m1_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vamoaddei8_v_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei8_v_u64m1((uint64_t *)(op0), (vuint8mf8_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vamoaddei8_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei8_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint8mf8_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vamoaddei8_v_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei8_v_u64m2((uint64_t *)(op0), (vuint8mf4_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vamoaddei8_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei8_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint8mf4_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vamoaddei8_v_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei8_v_u64m4((uint64_t *)(op0), (vuint8mf2_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vamoaddei8_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei8_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint8mf2_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vamoaddei8_v_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei8_v_u64m8((uint64_t *)(op0), (vuint8m1_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vamoaddei8_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei8_v_u64m8_m((vbool8_t)(op0), (uint64_t *)(op1), (vuint8m1_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vamoaddei16_v_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei16_v_i64m1((int64_t *)(op0), (vuint16mf4_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vamoaddei16_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei16_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint16mf4_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vamoaddei16_v_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei16_v_i64m2((int64_t *)(op0), (vuint16mf2_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vamoaddei16_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei16_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint16mf2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vamoaddei16_v_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei16_v_i64m4((int64_t *)(op0), (vuint16m1_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vamoaddei16_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei16_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vuint16m1_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vamoaddei16_v_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei16_v_i64m8((int64_t *)(op0), (vuint16m2_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vamoaddei16_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei16_v_i64m8_m((vbool8_t)(op0), (int64_t *)(op1), (vuint16m2_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vamoaddei16_v_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei16_v_u64m1((uint64_t *)(op0), (vuint16mf4_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vamoaddei16_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei16_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint16mf4_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vamoaddei16_v_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei16_v_u64m2((uint64_t *)(op0), (vuint16mf2_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vamoaddei16_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei16_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint16mf2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vamoaddei16_v_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei16_v_u64m4((uint64_t *)(op0), (vuint16m1_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vamoaddei16_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei16_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint16m1_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vamoaddei16_v_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei16_v_u64m8((uint64_t *)(op0), (vuint16m2_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vamoaddei16_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei16_v_u64m8_m((vbool8_t)(op0), (uint64_t *)(op1), (vuint16m2_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vamoaddei32_v_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei32_v_i64m1((int64_t *)(op0), (vuint32mf2_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vamoaddei32_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei32_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint32mf2_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vamoaddei32_v_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei32_v_i64m2((int64_t *)(op0), (vuint32m1_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vamoaddei32_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei32_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint32m1_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vamoaddei32_v_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei32_v_i64m4((int64_t *)(op0), (vuint32m2_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vamoaddei32_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei32_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vuint32m2_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vamoaddei32_v_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei32_v_i64m8((int64_t *)(op0), (vuint32m4_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vamoaddei32_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei32_v_i64m8_m((vbool8_t)(op0), (int64_t *)(op1), (vuint32m4_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vamoaddei32_v_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei32_v_u64m1((uint64_t *)(op0), (vuint32mf2_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vamoaddei32_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei32_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint32mf2_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vamoaddei32_v_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei32_v_u64m2((uint64_t *)(op0), (vuint32m1_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vamoaddei32_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei32_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint32m1_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vamoaddei32_v_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei32_v_u64m4((uint64_t *)(op0), (vuint32m2_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vamoaddei32_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei32_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint32m2_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vamoaddei32_v_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei32_v_u64m8((uint64_t *)(op0), (vuint32m4_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vamoaddei32_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei32_v_u64m8_m((vbool8_t)(op0), (uint64_t *)(op1), (vuint32m4_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vamoaddei64_v_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei64_v_i64m1((int64_t *)(op0), (vuint64m1_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vamoaddei64_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei64_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vamoaddei64_v_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei64_v_i64m2((int64_t *)(op0), (vuint64m2_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vamoaddei64_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei64_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vamoaddei64_v_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei64_v_i64m4((int64_t *)(op0), (vuint64m4_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vamoaddei64_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei64_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vuint64m4_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vamoaddei64_v_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei64_v_i64m8((int64_t *)(op0), (vuint64m8_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vamoaddei64_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei64_v_i64m8_m((vbool8_t)(op0), (int64_t *)(op1), (vuint64m8_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vamoaddei64_v_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei64_v_u64m1((uint64_t *)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vamoaddei64_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei64_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vamoaddei64_v_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei64_v_u64m2((uint64_t *)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vamoaddei64_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei64_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vamoaddei64_v_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei64_v_u64m4((uint64_t *)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vamoaddei64_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei64_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vamoaddei64_v_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoaddei64_v_u64m8((uint64_t *)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vamoaddei64_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoaddei64_v_u64m8_m((vbool8_t)(op0), (uint64_t *)(op1), (vuint64m8_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vamoxorei8_v_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei8_v_i32m1((int32_t *)(op0), (vuint8mf4_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vamoxorei8_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei8_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint8mf4_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vamoxorei8_v_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei8_v_i32m2((int32_t *)(op0), (vuint8mf2_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vamoxorei8_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei8_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint8mf2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vamoxorei8_v_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei8_v_i32m4((int32_t *)(op0), (vuint8m1_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vamoxorei8_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei8_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vuint8m1_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vamoxorei8_v_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei8_v_i32m8((int32_t *)(op0), (vuint8m2_t)(op1), (vint32m8_t)(op2), (size_t)(op3)) +#define vamoxorei8_v_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei8_v_i32m8_m((vbool4_t)(op0), (int32_t *)(op1), (vuint8m2_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vamoxorei8_v_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei8_v_i32mf2((int32_t *)(op0), (vuint8mf8_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vamoxorei8_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei8_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint8mf8_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vamoxorei8_v_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei8_v_u32m1((uint32_t *)(op0), (vuint8mf4_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vamoxorei8_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei8_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint8mf4_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vamoxorei8_v_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei8_v_u32m2((uint32_t *)(op0), (vuint8mf2_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vamoxorei8_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei8_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint8mf2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vamoxorei8_v_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei8_v_u32m4((uint32_t *)(op0), (vuint8m1_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vamoxorei8_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei8_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint8m1_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vamoxorei8_v_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei8_v_u32m8((uint32_t *)(op0), (vuint8m2_t)(op1), (vuint32m8_t)(op2), (size_t)(op3)) +#define vamoxorei8_v_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei8_v_u32m8_m((vbool4_t)(op0), (uint32_t *)(op1), (vuint8m2_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vamoxorei8_v_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei8_v_u32mf2((uint32_t *)(op0), (vuint8mf8_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vamoxorei8_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei8_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint8mf8_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vamoxorei16_v_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei16_v_i32m1((int32_t *)(op0), (vuint16mf2_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vamoxorei16_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei16_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint16mf2_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vamoxorei16_v_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei16_v_i32m2((int32_t *)(op0), (vuint16m1_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vamoxorei16_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei16_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint16m1_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vamoxorei16_v_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei16_v_i32m4((int32_t *)(op0), (vuint16m2_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vamoxorei16_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei16_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vuint16m2_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vamoxorei16_v_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei16_v_i32m8((int32_t *)(op0), (vuint16m4_t)(op1), (vint32m8_t)(op2), (size_t)(op3)) +#define vamoxorei16_v_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei16_v_i32m8_m((vbool4_t)(op0), (int32_t *)(op1), (vuint16m4_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vamoxorei16_v_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei16_v_i32mf2((int32_t *)(op0), (vuint16mf4_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vamoxorei16_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei16_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint16mf4_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vamoxorei16_v_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei16_v_u32m1((uint32_t *)(op0), (vuint16mf2_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vamoxorei16_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei16_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint16mf2_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vamoxorei16_v_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei16_v_u32m2((uint32_t *)(op0), (vuint16m1_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vamoxorei16_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei16_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint16m1_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vamoxorei16_v_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei16_v_u32m4((uint32_t *)(op0), (vuint16m2_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vamoxorei16_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei16_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint16m2_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vamoxorei16_v_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei16_v_u32m8((uint32_t *)(op0), (vuint16m4_t)(op1), (vuint32m8_t)(op2), (size_t)(op3)) +#define vamoxorei16_v_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei16_v_u32m8_m((vbool4_t)(op0), (uint32_t *)(op1), (vuint16m4_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vamoxorei16_v_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei16_v_u32mf2((uint32_t *)(op0), (vuint16mf4_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vamoxorei16_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei16_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint16mf4_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vamoxorei32_v_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei32_v_i32m1((int32_t *)(op0), (vuint32m1_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vamoxorei32_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei32_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vamoxorei32_v_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei32_v_i32m2((int32_t *)(op0), (vuint32m2_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vamoxorei32_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei32_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vamoxorei32_v_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei32_v_i32m4((int32_t *)(op0), (vuint32m4_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vamoxorei32_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei32_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vuint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vamoxorei32_v_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei32_v_i32m8((int32_t *)(op0), (vuint32m8_t)(op1), (vint32m8_t)(op2), (size_t)(op3)) +#define vamoxorei32_v_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei32_v_i32m8_m((vbool4_t)(op0), (int32_t *)(op1), (vuint32m8_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vamoxorei32_v_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei32_v_i32mf2((int32_t *)(op0), (vuint32mf2_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vamoxorei32_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei32_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vamoxorei32_v_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei32_v_u32m1((uint32_t *)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vamoxorei32_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei32_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vamoxorei32_v_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei32_v_u32m2((uint32_t *)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vamoxorei32_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei32_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vamoxorei32_v_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei32_v_u32m4((uint32_t *)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vamoxorei32_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei32_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vamoxorei32_v_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei32_v_u32m8((uint32_t *)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (size_t)(op3)) +#define vamoxorei32_v_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei32_v_u32m8_m((vbool4_t)(op0), (uint32_t *)(op1), (vuint32m8_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vamoxorei32_v_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei32_v_u32mf2((uint32_t *)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vamoxorei32_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei32_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vamoxorei64_v_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei64_v_i32m1((int32_t *)(op0), (vuint64m2_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vamoxorei64_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei64_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint64m2_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vamoxorei64_v_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei64_v_i32m2((int32_t *)(op0), (vuint64m4_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vamoxorei64_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei64_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint64m4_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vamoxorei64_v_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei64_v_i32m4((int32_t *)(op0), (vuint64m8_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vamoxorei64_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei64_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vuint64m8_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vamoxorei64_v_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei64_v_i32mf2((int32_t *)(op0), (vuint64m1_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vamoxorei64_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei64_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint64m1_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vamoxorei64_v_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei64_v_u32m1((uint32_t *)(op0), (vuint64m2_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vamoxorei64_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei64_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint64m2_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vamoxorei64_v_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei64_v_u32m2((uint32_t *)(op0), (vuint64m4_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vamoxorei64_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei64_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint64m4_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vamoxorei64_v_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei64_v_u32m4((uint32_t *)(op0), (vuint64m8_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vamoxorei64_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei64_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint64m8_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vamoxorei64_v_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei64_v_u32mf2((uint32_t *)(op0), (vuint64m1_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vamoxorei64_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei64_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint64m1_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vamoxorei8_v_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei8_v_i64m1((int64_t *)(op0), (vuint8mf8_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vamoxorei8_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei8_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint8mf8_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vamoxorei8_v_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei8_v_i64m2((int64_t *)(op0), (vuint8mf4_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vamoxorei8_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei8_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint8mf4_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vamoxorei8_v_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei8_v_i64m4((int64_t *)(op0), (vuint8mf2_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vamoxorei8_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei8_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vuint8mf2_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vamoxorei8_v_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei8_v_i64m8((int64_t *)(op0), (vuint8m1_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vamoxorei8_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei8_v_i64m8_m((vbool8_t)(op0), (int64_t *)(op1), (vuint8m1_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vamoxorei8_v_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei8_v_u64m1((uint64_t *)(op0), (vuint8mf8_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vamoxorei8_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei8_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint8mf8_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vamoxorei8_v_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei8_v_u64m2((uint64_t *)(op0), (vuint8mf4_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vamoxorei8_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei8_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint8mf4_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vamoxorei8_v_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei8_v_u64m4((uint64_t *)(op0), (vuint8mf2_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vamoxorei8_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei8_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint8mf2_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vamoxorei8_v_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei8_v_u64m8((uint64_t *)(op0), (vuint8m1_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vamoxorei8_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei8_v_u64m8_m((vbool8_t)(op0), (uint64_t *)(op1), (vuint8m1_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vamoxorei16_v_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei16_v_i64m1((int64_t *)(op0), (vuint16mf4_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vamoxorei16_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei16_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint16mf4_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vamoxorei16_v_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei16_v_i64m2((int64_t *)(op0), (vuint16mf2_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vamoxorei16_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei16_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint16mf2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vamoxorei16_v_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei16_v_i64m4((int64_t *)(op0), (vuint16m1_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vamoxorei16_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei16_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vuint16m1_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vamoxorei16_v_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei16_v_i64m8((int64_t *)(op0), (vuint16m2_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vamoxorei16_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei16_v_i64m8_m((vbool8_t)(op0), (int64_t *)(op1), (vuint16m2_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vamoxorei16_v_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei16_v_u64m1((uint64_t *)(op0), (vuint16mf4_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vamoxorei16_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei16_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint16mf4_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vamoxorei16_v_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei16_v_u64m2((uint64_t *)(op0), (vuint16mf2_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vamoxorei16_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei16_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint16mf2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vamoxorei16_v_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei16_v_u64m4((uint64_t *)(op0), (vuint16m1_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vamoxorei16_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei16_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint16m1_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vamoxorei16_v_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei16_v_u64m8((uint64_t *)(op0), (vuint16m2_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vamoxorei16_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei16_v_u64m8_m((vbool8_t)(op0), (uint64_t *)(op1), (vuint16m2_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vamoxorei32_v_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei32_v_i64m1((int64_t *)(op0), (vuint32mf2_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vamoxorei32_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei32_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint32mf2_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vamoxorei32_v_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei32_v_i64m2((int64_t *)(op0), (vuint32m1_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vamoxorei32_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei32_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint32m1_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vamoxorei32_v_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei32_v_i64m4((int64_t *)(op0), (vuint32m2_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vamoxorei32_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei32_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vuint32m2_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vamoxorei32_v_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei32_v_i64m8((int64_t *)(op0), (vuint32m4_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vamoxorei32_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei32_v_i64m8_m((vbool8_t)(op0), (int64_t *)(op1), (vuint32m4_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vamoxorei32_v_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei32_v_u64m1((uint64_t *)(op0), (vuint32mf2_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vamoxorei32_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei32_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint32mf2_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vamoxorei32_v_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei32_v_u64m2((uint64_t *)(op0), (vuint32m1_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vamoxorei32_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei32_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint32m1_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vamoxorei32_v_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei32_v_u64m4((uint64_t *)(op0), (vuint32m2_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vamoxorei32_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei32_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint32m2_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vamoxorei32_v_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei32_v_u64m8((uint64_t *)(op0), (vuint32m4_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vamoxorei32_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei32_v_u64m8_m((vbool8_t)(op0), (uint64_t *)(op1), (vuint32m4_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vamoxorei64_v_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei64_v_i64m1((int64_t *)(op0), (vuint64m1_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vamoxorei64_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei64_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vamoxorei64_v_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei64_v_i64m2((int64_t *)(op0), (vuint64m2_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vamoxorei64_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei64_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vamoxorei64_v_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei64_v_i64m4((int64_t *)(op0), (vuint64m4_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vamoxorei64_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei64_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vuint64m4_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vamoxorei64_v_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei64_v_i64m8((int64_t *)(op0), (vuint64m8_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vamoxorei64_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei64_v_i64m8_m((vbool8_t)(op0), (int64_t *)(op1), (vuint64m8_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vamoxorei64_v_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei64_v_u64m1((uint64_t *)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vamoxorei64_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei64_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vamoxorei64_v_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei64_v_u64m2((uint64_t *)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vamoxorei64_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei64_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vamoxorei64_v_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei64_v_u64m4((uint64_t *)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vamoxorei64_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei64_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vamoxorei64_v_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoxorei64_v_u64m8((uint64_t *)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vamoxorei64_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoxorei64_v_u64m8_m((vbool8_t)(op0), (uint64_t *)(op1), (vuint64m8_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vamoandei8_v_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei8_v_i32m1((int32_t *)(op0), (vuint8mf4_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vamoandei8_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei8_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint8mf4_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vamoandei8_v_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei8_v_i32m2((int32_t *)(op0), (vuint8mf2_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vamoandei8_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei8_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint8mf2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vamoandei8_v_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei8_v_i32m4((int32_t *)(op0), (vuint8m1_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vamoandei8_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei8_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vuint8m1_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vamoandei8_v_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei8_v_i32m8((int32_t *)(op0), (vuint8m2_t)(op1), (vint32m8_t)(op2), (size_t)(op3)) +#define vamoandei8_v_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei8_v_i32m8_m((vbool4_t)(op0), (int32_t *)(op1), (vuint8m2_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vamoandei8_v_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei8_v_i32mf2((int32_t *)(op0), (vuint8mf8_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vamoandei8_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei8_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint8mf8_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vamoandei8_v_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei8_v_u32m1((uint32_t *)(op0), (vuint8mf4_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vamoandei8_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei8_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint8mf4_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vamoandei8_v_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei8_v_u32m2((uint32_t *)(op0), (vuint8mf2_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vamoandei8_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei8_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint8mf2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vamoandei8_v_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei8_v_u32m4((uint32_t *)(op0), (vuint8m1_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vamoandei8_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei8_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint8m1_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vamoandei8_v_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei8_v_u32m8((uint32_t *)(op0), (vuint8m2_t)(op1), (vuint32m8_t)(op2), (size_t)(op3)) +#define vamoandei8_v_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei8_v_u32m8_m((vbool4_t)(op0), (uint32_t *)(op1), (vuint8m2_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vamoandei8_v_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei8_v_u32mf2((uint32_t *)(op0), (vuint8mf8_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vamoandei8_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei8_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint8mf8_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vamoandei16_v_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei16_v_i32m1((int32_t *)(op0), (vuint16mf2_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vamoandei16_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei16_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint16mf2_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vamoandei16_v_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei16_v_i32m2((int32_t *)(op0), (vuint16m1_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vamoandei16_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei16_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint16m1_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vamoandei16_v_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei16_v_i32m4((int32_t *)(op0), (vuint16m2_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vamoandei16_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei16_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vuint16m2_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vamoandei16_v_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei16_v_i32m8((int32_t *)(op0), (vuint16m4_t)(op1), (vint32m8_t)(op2), (size_t)(op3)) +#define vamoandei16_v_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei16_v_i32m8_m((vbool4_t)(op0), (int32_t *)(op1), (vuint16m4_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vamoandei16_v_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei16_v_i32mf2((int32_t *)(op0), (vuint16mf4_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vamoandei16_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei16_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint16mf4_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vamoandei16_v_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei16_v_u32m1((uint32_t *)(op0), (vuint16mf2_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vamoandei16_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei16_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint16mf2_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vamoandei16_v_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei16_v_u32m2((uint32_t *)(op0), (vuint16m1_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vamoandei16_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei16_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint16m1_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vamoandei16_v_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei16_v_u32m4((uint32_t *)(op0), (vuint16m2_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vamoandei16_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei16_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint16m2_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vamoandei16_v_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei16_v_u32m8((uint32_t *)(op0), (vuint16m4_t)(op1), (vuint32m8_t)(op2), (size_t)(op3)) +#define vamoandei16_v_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei16_v_u32m8_m((vbool4_t)(op0), (uint32_t *)(op1), (vuint16m4_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vamoandei16_v_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei16_v_u32mf2((uint32_t *)(op0), (vuint16mf4_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vamoandei16_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei16_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint16mf4_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vamoandei32_v_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei32_v_i32m1((int32_t *)(op0), (vuint32m1_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vamoandei32_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei32_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vamoandei32_v_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei32_v_i32m2((int32_t *)(op0), (vuint32m2_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vamoandei32_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei32_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vamoandei32_v_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei32_v_i32m4((int32_t *)(op0), (vuint32m4_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vamoandei32_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei32_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vuint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vamoandei32_v_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei32_v_i32m8((int32_t *)(op0), (vuint32m8_t)(op1), (vint32m8_t)(op2), (size_t)(op3)) +#define vamoandei32_v_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei32_v_i32m8_m((vbool4_t)(op0), (int32_t *)(op1), (vuint32m8_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vamoandei32_v_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei32_v_i32mf2((int32_t *)(op0), (vuint32mf2_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vamoandei32_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei32_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vamoandei32_v_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei32_v_u32m1((uint32_t *)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vamoandei32_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei32_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vamoandei32_v_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei32_v_u32m2((uint32_t *)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vamoandei32_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei32_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vamoandei32_v_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei32_v_u32m4((uint32_t *)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vamoandei32_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei32_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vamoandei32_v_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei32_v_u32m8((uint32_t *)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (size_t)(op3)) +#define vamoandei32_v_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei32_v_u32m8_m((vbool4_t)(op0), (uint32_t *)(op1), (vuint32m8_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vamoandei32_v_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei32_v_u32mf2((uint32_t *)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vamoandei32_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei32_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vamoandei64_v_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei64_v_i32m1((int32_t *)(op0), (vuint64m2_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vamoandei64_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei64_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint64m2_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vamoandei64_v_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei64_v_i32m2((int32_t *)(op0), (vuint64m4_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vamoandei64_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei64_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint64m4_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vamoandei64_v_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei64_v_i32m4((int32_t *)(op0), (vuint64m8_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vamoandei64_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei64_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vuint64m8_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vamoandei64_v_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei64_v_i32mf2((int32_t *)(op0), (vuint64m1_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vamoandei64_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei64_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint64m1_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vamoandei64_v_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei64_v_u32m1((uint32_t *)(op0), (vuint64m2_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vamoandei64_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei64_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint64m2_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vamoandei64_v_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei64_v_u32m2((uint32_t *)(op0), (vuint64m4_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vamoandei64_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei64_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint64m4_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vamoandei64_v_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei64_v_u32m4((uint32_t *)(op0), (vuint64m8_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vamoandei64_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei64_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint64m8_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vamoandei64_v_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei64_v_u32mf2((uint32_t *)(op0), (vuint64m1_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vamoandei64_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei64_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint64m1_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vamoandei8_v_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei8_v_i64m1((int64_t *)(op0), (vuint8mf8_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vamoandei8_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei8_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint8mf8_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vamoandei8_v_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei8_v_i64m2((int64_t *)(op0), (vuint8mf4_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vamoandei8_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei8_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint8mf4_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vamoandei8_v_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei8_v_i64m4((int64_t *)(op0), (vuint8mf2_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vamoandei8_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei8_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vuint8mf2_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vamoandei8_v_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei8_v_i64m8((int64_t *)(op0), (vuint8m1_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vamoandei8_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei8_v_i64m8_m((vbool8_t)(op0), (int64_t *)(op1), (vuint8m1_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vamoandei8_v_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei8_v_u64m1((uint64_t *)(op0), (vuint8mf8_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vamoandei8_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei8_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint8mf8_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vamoandei8_v_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei8_v_u64m2((uint64_t *)(op0), (vuint8mf4_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vamoandei8_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei8_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint8mf4_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vamoandei8_v_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei8_v_u64m4((uint64_t *)(op0), (vuint8mf2_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vamoandei8_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei8_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint8mf2_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vamoandei8_v_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei8_v_u64m8((uint64_t *)(op0), (vuint8m1_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vamoandei8_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei8_v_u64m8_m((vbool8_t)(op0), (uint64_t *)(op1), (vuint8m1_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vamoandei16_v_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei16_v_i64m1((int64_t *)(op0), (vuint16mf4_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vamoandei16_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei16_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint16mf4_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vamoandei16_v_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei16_v_i64m2((int64_t *)(op0), (vuint16mf2_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vamoandei16_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei16_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint16mf2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vamoandei16_v_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei16_v_i64m4((int64_t *)(op0), (vuint16m1_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vamoandei16_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei16_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vuint16m1_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vamoandei16_v_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei16_v_i64m8((int64_t *)(op0), (vuint16m2_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vamoandei16_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei16_v_i64m8_m((vbool8_t)(op0), (int64_t *)(op1), (vuint16m2_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vamoandei16_v_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei16_v_u64m1((uint64_t *)(op0), (vuint16mf4_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vamoandei16_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei16_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint16mf4_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vamoandei16_v_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei16_v_u64m2((uint64_t *)(op0), (vuint16mf2_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vamoandei16_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei16_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint16mf2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vamoandei16_v_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei16_v_u64m4((uint64_t *)(op0), (vuint16m1_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vamoandei16_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei16_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint16m1_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vamoandei16_v_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei16_v_u64m8((uint64_t *)(op0), (vuint16m2_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vamoandei16_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei16_v_u64m8_m((vbool8_t)(op0), (uint64_t *)(op1), (vuint16m2_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vamoandei32_v_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei32_v_i64m1((int64_t *)(op0), (vuint32mf2_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vamoandei32_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei32_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint32mf2_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vamoandei32_v_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei32_v_i64m2((int64_t *)(op0), (vuint32m1_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vamoandei32_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei32_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint32m1_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vamoandei32_v_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei32_v_i64m4((int64_t *)(op0), (vuint32m2_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vamoandei32_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei32_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vuint32m2_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vamoandei32_v_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei32_v_i64m8((int64_t *)(op0), (vuint32m4_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vamoandei32_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei32_v_i64m8_m((vbool8_t)(op0), (int64_t *)(op1), (vuint32m4_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vamoandei32_v_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei32_v_u64m1((uint64_t *)(op0), (vuint32mf2_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vamoandei32_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei32_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint32mf2_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vamoandei32_v_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei32_v_u64m2((uint64_t *)(op0), (vuint32m1_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vamoandei32_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei32_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint32m1_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vamoandei32_v_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei32_v_u64m4((uint64_t *)(op0), (vuint32m2_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vamoandei32_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei32_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint32m2_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vamoandei32_v_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei32_v_u64m8((uint64_t *)(op0), (vuint32m4_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vamoandei32_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei32_v_u64m8_m((vbool8_t)(op0), (uint64_t *)(op1), (vuint32m4_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vamoandei64_v_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei64_v_i64m1((int64_t *)(op0), (vuint64m1_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vamoandei64_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei64_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vamoandei64_v_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei64_v_i64m2((int64_t *)(op0), (vuint64m2_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vamoandei64_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei64_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vamoandei64_v_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei64_v_i64m4((int64_t *)(op0), (vuint64m4_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vamoandei64_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei64_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vuint64m4_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vamoandei64_v_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei64_v_i64m8((int64_t *)(op0), (vuint64m8_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vamoandei64_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei64_v_i64m8_m((vbool8_t)(op0), (int64_t *)(op1), (vuint64m8_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vamoandei64_v_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei64_v_u64m1((uint64_t *)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vamoandei64_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei64_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vamoandei64_v_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei64_v_u64m2((uint64_t *)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vamoandei64_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei64_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vamoandei64_v_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei64_v_u64m4((uint64_t *)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vamoandei64_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei64_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vamoandei64_v_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoandei64_v_u64m8((uint64_t *)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vamoandei64_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoandei64_v_u64m8_m((vbool8_t)(op0), (uint64_t *)(op1), (vuint64m8_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vamoorei8_v_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei8_v_i32m1((int32_t *)(op0), (vuint8mf4_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vamoorei8_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei8_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint8mf4_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vamoorei8_v_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei8_v_i32m2((int32_t *)(op0), (vuint8mf2_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vamoorei8_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei8_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint8mf2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vamoorei8_v_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei8_v_i32m4((int32_t *)(op0), (vuint8m1_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vamoorei8_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei8_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vuint8m1_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vamoorei8_v_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei8_v_i32m8((int32_t *)(op0), (vuint8m2_t)(op1), (vint32m8_t)(op2), (size_t)(op3)) +#define vamoorei8_v_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei8_v_i32m8_m((vbool4_t)(op0), (int32_t *)(op1), (vuint8m2_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vamoorei8_v_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei8_v_i32mf2((int32_t *)(op0), (vuint8mf8_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vamoorei8_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei8_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint8mf8_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vamoorei8_v_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei8_v_u32m1((uint32_t *)(op0), (vuint8mf4_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vamoorei8_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei8_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint8mf4_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vamoorei8_v_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei8_v_u32m2((uint32_t *)(op0), (vuint8mf2_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vamoorei8_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei8_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint8mf2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vamoorei8_v_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei8_v_u32m4((uint32_t *)(op0), (vuint8m1_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vamoorei8_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei8_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint8m1_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vamoorei8_v_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei8_v_u32m8((uint32_t *)(op0), (vuint8m2_t)(op1), (vuint32m8_t)(op2), (size_t)(op3)) +#define vamoorei8_v_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei8_v_u32m8_m((vbool4_t)(op0), (uint32_t *)(op1), (vuint8m2_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vamoorei8_v_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei8_v_u32mf2((uint32_t *)(op0), (vuint8mf8_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vamoorei8_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei8_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint8mf8_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vamoorei16_v_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei16_v_i32m1((int32_t *)(op0), (vuint16mf2_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vamoorei16_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei16_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint16mf2_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vamoorei16_v_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei16_v_i32m2((int32_t *)(op0), (vuint16m1_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vamoorei16_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei16_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint16m1_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vamoorei16_v_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei16_v_i32m4((int32_t *)(op0), (vuint16m2_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vamoorei16_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei16_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vuint16m2_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vamoorei16_v_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei16_v_i32m8((int32_t *)(op0), (vuint16m4_t)(op1), (vint32m8_t)(op2), (size_t)(op3)) +#define vamoorei16_v_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei16_v_i32m8_m((vbool4_t)(op0), (int32_t *)(op1), (vuint16m4_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vamoorei16_v_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei16_v_i32mf2((int32_t *)(op0), (vuint16mf4_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vamoorei16_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei16_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint16mf4_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vamoorei16_v_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei16_v_u32m1((uint32_t *)(op0), (vuint16mf2_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vamoorei16_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei16_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint16mf2_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vamoorei16_v_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei16_v_u32m2((uint32_t *)(op0), (vuint16m1_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vamoorei16_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei16_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint16m1_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vamoorei16_v_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei16_v_u32m4((uint32_t *)(op0), (vuint16m2_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vamoorei16_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei16_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint16m2_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vamoorei16_v_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei16_v_u32m8((uint32_t *)(op0), (vuint16m4_t)(op1), (vuint32m8_t)(op2), (size_t)(op3)) +#define vamoorei16_v_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei16_v_u32m8_m((vbool4_t)(op0), (uint32_t *)(op1), (vuint16m4_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vamoorei16_v_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei16_v_u32mf2((uint32_t *)(op0), (vuint16mf4_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vamoorei16_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei16_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint16mf4_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vamoorei32_v_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei32_v_i32m1((int32_t *)(op0), (vuint32m1_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vamoorei32_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei32_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vamoorei32_v_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei32_v_i32m2((int32_t *)(op0), (vuint32m2_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vamoorei32_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei32_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vamoorei32_v_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei32_v_i32m4((int32_t *)(op0), (vuint32m4_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vamoorei32_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei32_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vuint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vamoorei32_v_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei32_v_i32m8((int32_t *)(op0), (vuint32m8_t)(op1), (vint32m8_t)(op2), (size_t)(op3)) +#define vamoorei32_v_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei32_v_i32m8_m((vbool4_t)(op0), (int32_t *)(op1), (vuint32m8_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vamoorei32_v_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei32_v_i32mf2((int32_t *)(op0), (vuint32mf2_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vamoorei32_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei32_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vamoorei32_v_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei32_v_u32m1((uint32_t *)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vamoorei32_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei32_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vamoorei32_v_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei32_v_u32m2((uint32_t *)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vamoorei32_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei32_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vamoorei32_v_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei32_v_u32m4((uint32_t *)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vamoorei32_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei32_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vamoorei32_v_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei32_v_u32m8((uint32_t *)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (size_t)(op3)) +#define vamoorei32_v_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei32_v_u32m8_m((vbool4_t)(op0), (uint32_t *)(op1), (vuint32m8_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vamoorei32_v_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei32_v_u32mf2((uint32_t *)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vamoorei32_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei32_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vamoorei64_v_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei64_v_i32m1((int32_t *)(op0), (vuint64m2_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vamoorei64_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei64_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint64m2_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vamoorei64_v_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei64_v_i32m2((int32_t *)(op0), (vuint64m4_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vamoorei64_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei64_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint64m4_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vamoorei64_v_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei64_v_i32m4((int32_t *)(op0), (vuint64m8_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vamoorei64_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei64_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vuint64m8_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vamoorei64_v_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei64_v_i32mf2((int32_t *)(op0), (vuint64m1_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vamoorei64_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei64_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint64m1_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vamoorei64_v_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei64_v_u32m1((uint32_t *)(op0), (vuint64m2_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vamoorei64_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei64_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint64m2_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vamoorei64_v_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei64_v_u32m2((uint32_t *)(op0), (vuint64m4_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vamoorei64_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei64_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint64m4_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vamoorei64_v_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei64_v_u32m4((uint32_t *)(op0), (vuint64m8_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vamoorei64_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei64_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint64m8_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vamoorei64_v_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei64_v_u32mf2((uint32_t *)(op0), (vuint64m1_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vamoorei64_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei64_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint64m1_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vamoorei8_v_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei8_v_i64m1((int64_t *)(op0), (vuint8mf8_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vamoorei8_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei8_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint8mf8_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vamoorei8_v_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei8_v_i64m2((int64_t *)(op0), (vuint8mf4_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vamoorei8_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei8_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint8mf4_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vamoorei8_v_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei8_v_i64m4((int64_t *)(op0), (vuint8mf2_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vamoorei8_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei8_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vuint8mf2_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vamoorei8_v_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei8_v_i64m8((int64_t *)(op0), (vuint8m1_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vamoorei8_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei8_v_i64m8_m((vbool8_t)(op0), (int64_t *)(op1), (vuint8m1_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vamoorei8_v_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei8_v_u64m1((uint64_t *)(op0), (vuint8mf8_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vamoorei8_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei8_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint8mf8_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vamoorei8_v_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei8_v_u64m2((uint64_t *)(op0), (vuint8mf4_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vamoorei8_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei8_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint8mf4_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vamoorei8_v_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei8_v_u64m4((uint64_t *)(op0), (vuint8mf2_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vamoorei8_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei8_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint8mf2_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vamoorei8_v_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei8_v_u64m8((uint64_t *)(op0), (vuint8m1_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vamoorei8_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei8_v_u64m8_m((vbool8_t)(op0), (uint64_t *)(op1), (vuint8m1_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vamoorei16_v_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei16_v_i64m1((int64_t *)(op0), (vuint16mf4_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vamoorei16_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei16_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint16mf4_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vamoorei16_v_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei16_v_i64m2((int64_t *)(op0), (vuint16mf2_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vamoorei16_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei16_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint16mf2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vamoorei16_v_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei16_v_i64m4((int64_t *)(op0), (vuint16m1_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vamoorei16_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei16_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vuint16m1_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vamoorei16_v_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei16_v_i64m8((int64_t *)(op0), (vuint16m2_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vamoorei16_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei16_v_i64m8_m((vbool8_t)(op0), (int64_t *)(op1), (vuint16m2_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vamoorei16_v_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei16_v_u64m1((uint64_t *)(op0), (vuint16mf4_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vamoorei16_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei16_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint16mf4_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vamoorei16_v_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei16_v_u64m2((uint64_t *)(op0), (vuint16mf2_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vamoorei16_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei16_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint16mf2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vamoorei16_v_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei16_v_u64m4((uint64_t *)(op0), (vuint16m1_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vamoorei16_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei16_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint16m1_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vamoorei16_v_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei16_v_u64m8((uint64_t *)(op0), (vuint16m2_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vamoorei16_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei16_v_u64m8_m((vbool8_t)(op0), (uint64_t *)(op1), (vuint16m2_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vamoorei32_v_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei32_v_i64m1((int64_t *)(op0), (vuint32mf2_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vamoorei32_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei32_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint32mf2_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vamoorei32_v_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei32_v_i64m2((int64_t *)(op0), (vuint32m1_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vamoorei32_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei32_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint32m1_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vamoorei32_v_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei32_v_i64m4((int64_t *)(op0), (vuint32m2_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vamoorei32_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei32_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vuint32m2_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vamoorei32_v_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei32_v_i64m8((int64_t *)(op0), (vuint32m4_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vamoorei32_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei32_v_i64m8_m((vbool8_t)(op0), (int64_t *)(op1), (vuint32m4_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vamoorei32_v_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei32_v_u64m1((uint64_t *)(op0), (vuint32mf2_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vamoorei32_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei32_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint32mf2_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vamoorei32_v_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei32_v_u64m2((uint64_t *)(op0), (vuint32m1_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vamoorei32_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei32_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint32m1_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vamoorei32_v_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei32_v_u64m4((uint64_t *)(op0), (vuint32m2_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vamoorei32_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei32_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint32m2_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vamoorei32_v_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei32_v_u64m8((uint64_t *)(op0), (vuint32m4_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vamoorei32_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei32_v_u64m8_m((vbool8_t)(op0), (uint64_t *)(op1), (vuint32m4_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vamoorei64_v_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei64_v_i64m1((int64_t *)(op0), (vuint64m1_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vamoorei64_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei64_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vamoorei64_v_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei64_v_i64m2((int64_t *)(op0), (vuint64m2_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vamoorei64_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei64_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vamoorei64_v_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei64_v_i64m4((int64_t *)(op0), (vuint64m4_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vamoorei64_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei64_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vuint64m4_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vamoorei64_v_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei64_v_i64m8((int64_t *)(op0), (vuint64m8_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vamoorei64_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei64_v_i64m8_m((vbool8_t)(op0), (int64_t *)(op1), (vuint64m8_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vamoorei64_v_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei64_v_u64m1((uint64_t *)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vamoorei64_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei64_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vamoorei64_v_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei64_v_u64m2((uint64_t *)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vamoorei64_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei64_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vamoorei64_v_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei64_v_u64m4((uint64_t *)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vamoorei64_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei64_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vamoorei64_v_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoorei64_v_u64m8((uint64_t *)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vamoorei64_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoorei64_v_u64m8_m((vbool8_t)(op0), (uint64_t *)(op1), (vuint64m8_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vamominei8_v_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamominei8_v_i32m1((int32_t *)(op0), (vuint8mf4_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vamominei8_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominei8_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint8mf4_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vamominei8_v_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamominei8_v_i32m2((int32_t *)(op0), (vuint8mf2_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vamominei8_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominei8_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint8mf2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vamominei8_v_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamominei8_v_i32m4((int32_t *)(op0), (vuint8m1_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vamominei8_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominei8_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vuint8m1_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vamominei8_v_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vamominei8_v_i32m8((int32_t *)(op0), (vuint8m2_t)(op1), (vint32m8_t)(op2), (size_t)(op3)) +#define vamominei8_v_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominei8_v_i32m8_m((vbool4_t)(op0), (int32_t *)(op1), (vuint8m2_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vamominei8_v_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamominei8_v_i32mf2((int32_t *)(op0), (vuint8mf8_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vamominei8_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominei8_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint8mf8_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vamominei16_v_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamominei16_v_i32m1((int32_t *)(op0), (vuint16mf2_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vamominei16_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominei16_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint16mf2_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vamominei16_v_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamominei16_v_i32m2((int32_t *)(op0), (vuint16m1_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vamominei16_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominei16_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint16m1_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vamominei16_v_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamominei16_v_i32m4((int32_t *)(op0), (vuint16m2_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vamominei16_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominei16_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vuint16m2_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vamominei16_v_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vamominei16_v_i32m8((int32_t *)(op0), (vuint16m4_t)(op1), (vint32m8_t)(op2), (size_t)(op3)) +#define vamominei16_v_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominei16_v_i32m8_m((vbool4_t)(op0), (int32_t *)(op1), (vuint16m4_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vamominei16_v_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamominei16_v_i32mf2((int32_t *)(op0), (vuint16mf4_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vamominei16_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominei16_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint16mf4_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vamominei32_v_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamominei32_v_i32m1((int32_t *)(op0), (vuint32m1_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vamominei32_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominei32_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vamominei32_v_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamominei32_v_i32m2((int32_t *)(op0), (vuint32m2_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vamominei32_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominei32_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vamominei32_v_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamominei32_v_i32m4((int32_t *)(op0), (vuint32m4_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vamominei32_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominei32_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vuint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vamominei32_v_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vamominei32_v_i32m8((int32_t *)(op0), (vuint32m8_t)(op1), (vint32m8_t)(op2), (size_t)(op3)) +#define vamominei32_v_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominei32_v_i32m8_m((vbool4_t)(op0), (int32_t *)(op1), (vuint32m8_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vamominei32_v_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamominei32_v_i32mf2((int32_t *)(op0), (vuint32mf2_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vamominei32_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominei32_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vamominei64_v_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamominei64_v_i32m1((int32_t *)(op0), (vuint64m2_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vamominei64_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominei64_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint64m2_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vamominei64_v_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamominei64_v_i32m2((int32_t *)(op0), (vuint64m4_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vamominei64_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominei64_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint64m4_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vamominei64_v_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamominei64_v_i32m4((int32_t *)(op0), (vuint64m8_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vamominei64_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominei64_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vuint64m8_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vamominei64_v_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamominei64_v_i32mf2((int32_t *)(op0), (vuint64m1_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vamominei64_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominei64_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint64m1_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vamominei8_v_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamominei8_v_i64m1((int64_t *)(op0), (vuint8mf8_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vamominei8_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominei8_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint8mf8_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vamominei8_v_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamominei8_v_i64m2((int64_t *)(op0), (vuint8mf4_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vamominei8_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominei8_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint8mf4_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vamominei8_v_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamominei8_v_i64m4((int64_t *)(op0), (vuint8mf2_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vamominei8_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominei8_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vuint8mf2_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vamominei8_v_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamominei8_v_i64m8((int64_t *)(op0), (vuint8m1_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vamominei8_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominei8_v_i64m8_m((vbool8_t)(op0), (int64_t *)(op1), (vuint8m1_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vamominei16_v_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamominei16_v_i64m1((int64_t *)(op0), (vuint16mf4_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vamominei16_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominei16_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint16mf4_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vamominei16_v_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamominei16_v_i64m2((int64_t *)(op0), (vuint16mf2_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vamominei16_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominei16_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint16mf2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vamominei16_v_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamominei16_v_i64m4((int64_t *)(op0), (vuint16m1_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vamominei16_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominei16_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vuint16m1_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vamominei16_v_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamominei16_v_i64m8((int64_t *)(op0), (vuint16m2_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vamominei16_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominei16_v_i64m8_m((vbool8_t)(op0), (int64_t *)(op1), (vuint16m2_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vamominei32_v_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamominei32_v_i64m1((int64_t *)(op0), (vuint32mf2_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vamominei32_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominei32_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint32mf2_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vamominei32_v_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamominei32_v_i64m2((int64_t *)(op0), (vuint32m1_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vamominei32_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominei32_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint32m1_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vamominei32_v_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamominei32_v_i64m4((int64_t *)(op0), (vuint32m2_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vamominei32_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominei32_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vuint32m2_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vamominei32_v_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamominei32_v_i64m8((int64_t *)(op0), (vuint32m4_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vamominei32_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominei32_v_i64m8_m((vbool8_t)(op0), (int64_t *)(op1), (vuint32m4_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vamominei64_v_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamominei64_v_i64m1((int64_t *)(op0), (vuint64m1_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vamominei64_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominei64_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vamominei64_v_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamominei64_v_i64m2((int64_t *)(op0), (vuint64m2_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vamominei64_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominei64_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vamominei64_v_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamominei64_v_i64m4((int64_t *)(op0), (vuint64m4_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vamominei64_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominei64_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vuint64m4_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vamominei64_v_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamominei64_v_i64m8((int64_t *)(op0), (vuint64m8_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vamominei64_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominei64_v_i64m8_m((vbool8_t)(op0), (int64_t *)(op1), (vuint64m8_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vamomaxei8_v_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxei8_v_i32m1((int32_t *)(op0), (vuint8mf4_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vamomaxei8_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxei8_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint8mf4_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vamomaxei8_v_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxei8_v_i32m2((int32_t *)(op0), (vuint8mf2_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vamomaxei8_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxei8_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint8mf2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vamomaxei8_v_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxei8_v_i32m4((int32_t *)(op0), (vuint8m1_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vamomaxei8_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxei8_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vuint8m1_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vamomaxei8_v_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxei8_v_i32m8((int32_t *)(op0), (vuint8m2_t)(op1), (vint32m8_t)(op2), (size_t)(op3)) +#define vamomaxei8_v_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxei8_v_i32m8_m((vbool4_t)(op0), (int32_t *)(op1), (vuint8m2_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vamomaxei8_v_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxei8_v_i32mf2((int32_t *)(op0), (vuint8mf8_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vamomaxei8_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxei8_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint8mf8_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vamomaxei16_v_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxei16_v_i32m1((int32_t *)(op0), (vuint16mf2_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vamomaxei16_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxei16_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint16mf2_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vamomaxei16_v_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxei16_v_i32m2((int32_t *)(op0), (vuint16m1_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vamomaxei16_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxei16_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint16m1_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vamomaxei16_v_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxei16_v_i32m4((int32_t *)(op0), (vuint16m2_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vamomaxei16_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxei16_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vuint16m2_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vamomaxei16_v_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxei16_v_i32m8((int32_t *)(op0), (vuint16m4_t)(op1), (vint32m8_t)(op2), (size_t)(op3)) +#define vamomaxei16_v_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxei16_v_i32m8_m((vbool4_t)(op0), (int32_t *)(op1), (vuint16m4_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vamomaxei16_v_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxei16_v_i32mf2((int32_t *)(op0), (vuint16mf4_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vamomaxei16_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxei16_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint16mf4_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vamomaxei32_v_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxei32_v_i32m1((int32_t *)(op0), (vuint32m1_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vamomaxei32_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxei32_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vamomaxei32_v_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxei32_v_i32m2((int32_t *)(op0), (vuint32m2_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vamomaxei32_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxei32_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vamomaxei32_v_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxei32_v_i32m4((int32_t *)(op0), (vuint32m4_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vamomaxei32_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxei32_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vuint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vamomaxei32_v_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxei32_v_i32m8((int32_t *)(op0), (vuint32m8_t)(op1), (vint32m8_t)(op2), (size_t)(op3)) +#define vamomaxei32_v_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxei32_v_i32m8_m((vbool4_t)(op0), (int32_t *)(op1), (vuint32m8_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vamomaxei32_v_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxei32_v_i32mf2((int32_t *)(op0), (vuint32mf2_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vamomaxei32_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxei32_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vamomaxei64_v_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxei64_v_i32m1((int32_t *)(op0), (vuint64m2_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vamomaxei64_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxei64_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint64m2_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vamomaxei64_v_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxei64_v_i32m2((int32_t *)(op0), (vuint64m4_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vamomaxei64_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxei64_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint64m4_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vamomaxei64_v_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxei64_v_i32m4((int32_t *)(op0), (vuint64m8_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vamomaxei64_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxei64_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vuint64m8_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vamomaxei64_v_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxei64_v_i32mf2((int32_t *)(op0), (vuint64m1_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vamomaxei64_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxei64_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint64m1_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vamomaxei8_v_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxei8_v_i64m1((int64_t *)(op0), (vuint8mf8_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vamomaxei8_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxei8_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint8mf8_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vamomaxei8_v_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxei8_v_i64m2((int64_t *)(op0), (vuint8mf4_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vamomaxei8_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxei8_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint8mf4_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vamomaxei8_v_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxei8_v_i64m4((int64_t *)(op0), (vuint8mf2_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vamomaxei8_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxei8_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vuint8mf2_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vamomaxei8_v_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxei8_v_i64m8((int64_t *)(op0), (vuint8m1_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vamomaxei8_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxei8_v_i64m8_m((vbool8_t)(op0), (int64_t *)(op1), (vuint8m1_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vamomaxei16_v_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxei16_v_i64m1((int64_t *)(op0), (vuint16mf4_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vamomaxei16_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxei16_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint16mf4_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vamomaxei16_v_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxei16_v_i64m2((int64_t *)(op0), (vuint16mf2_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vamomaxei16_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxei16_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint16mf2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vamomaxei16_v_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxei16_v_i64m4((int64_t *)(op0), (vuint16m1_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vamomaxei16_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxei16_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vuint16m1_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vamomaxei16_v_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxei16_v_i64m8((int64_t *)(op0), (vuint16m2_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vamomaxei16_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxei16_v_i64m8_m((vbool8_t)(op0), (int64_t *)(op1), (vuint16m2_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vamomaxei32_v_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxei32_v_i64m1((int64_t *)(op0), (vuint32mf2_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vamomaxei32_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxei32_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint32mf2_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vamomaxei32_v_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxei32_v_i64m2((int64_t *)(op0), (vuint32m1_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vamomaxei32_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxei32_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint32m1_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vamomaxei32_v_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxei32_v_i64m4((int64_t *)(op0), (vuint32m2_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vamomaxei32_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxei32_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vuint32m2_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vamomaxei32_v_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxei32_v_i64m8((int64_t *)(op0), (vuint32m4_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vamomaxei32_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxei32_v_i64m8_m((vbool8_t)(op0), (int64_t *)(op1), (vuint32m4_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vamomaxei64_v_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxei64_v_i64m1((int64_t *)(op0), (vuint64m1_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vamomaxei64_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxei64_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vamomaxei64_v_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxei64_v_i64m2((int64_t *)(op0), (vuint64m2_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vamomaxei64_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxei64_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vamomaxei64_v_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxei64_v_i64m4((int64_t *)(op0), (vuint64m4_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vamomaxei64_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxei64_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vuint64m4_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vamomaxei64_v_i64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxei64_v_i64m8((int64_t *)(op0), (vuint64m8_t)(op1), (vint64m8_t)(op2), (size_t)(op3)) +#define vamomaxei64_v_i64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxei64_v_i64m8_m((vbool8_t)(op0), (int64_t *)(op1), (vuint64m8_t)(op2), (vint64m8_t)(op3), (size_t)(op4)) +#define vamominuei8_v_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamominuei8_v_u32m1((uint32_t *)(op0), (vuint8mf4_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vamominuei8_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominuei8_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint8mf4_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vamominuei8_v_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamominuei8_v_u32m2((uint32_t *)(op0), (vuint8mf2_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vamominuei8_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominuei8_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint8mf2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vamominuei8_v_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamominuei8_v_u32m4((uint32_t *)(op0), (vuint8m1_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vamominuei8_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominuei8_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint8m1_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vamominuei8_v_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vamominuei8_v_u32m8((uint32_t *)(op0), (vuint8m2_t)(op1), (vuint32m8_t)(op2), (size_t)(op3)) +#define vamominuei8_v_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominuei8_v_u32m8_m((vbool4_t)(op0), (uint32_t *)(op1), (vuint8m2_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vamominuei8_v_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamominuei8_v_u32mf2((uint32_t *)(op0), (vuint8mf8_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vamominuei8_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominuei8_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint8mf8_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vamominuei16_v_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamominuei16_v_u32m1((uint32_t *)(op0), (vuint16mf2_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vamominuei16_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominuei16_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint16mf2_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vamominuei16_v_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamominuei16_v_u32m2((uint32_t *)(op0), (vuint16m1_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vamominuei16_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominuei16_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint16m1_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vamominuei16_v_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamominuei16_v_u32m4((uint32_t *)(op0), (vuint16m2_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vamominuei16_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominuei16_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint16m2_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vamominuei16_v_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vamominuei16_v_u32m8((uint32_t *)(op0), (vuint16m4_t)(op1), (vuint32m8_t)(op2), (size_t)(op3)) +#define vamominuei16_v_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominuei16_v_u32m8_m((vbool4_t)(op0), (uint32_t *)(op1), (vuint16m4_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vamominuei16_v_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamominuei16_v_u32mf2((uint32_t *)(op0), (vuint16mf4_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vamominuei16_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominuei16_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint16mf4_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vamominuei32_v_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamominuei32_v_u32m1((uint32_t *)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vamominuei32_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominuei32_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vamominuei32_v_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamominuei32_v_u32m2((uint32_t *)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vamominuei32_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominuei32_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vamominuei32_v_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamominuei32_v_u32m4((uint32_t *)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vamominuei32_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominuei32_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vamominuei32_v_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vamominuei32_v_u32m8((uint32_t *)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (size_t)(op3)) +#define vamominuei32_v_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominuei32_v_u32m8_m((vbool4_t)(op0), (uint32_t *)(op1), (vuint32m8_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vamominuei32_v_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamominuei32_v_u32mf2((uint32_t *)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vamominuei32_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominuei32_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vamominuei64_v_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamominuei64_v_u32m1((uint32_t *)(op0), (vuint64m2_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vamominuei64_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominuei64_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint64m2_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vamominuei64_v_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamominuei64_v_u32m2((uint32_t *)(op0), (vuint64m4_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vamominuei64_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominuei64_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint64m4_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vamominuei64_v_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamominuei64_v_u32m4((uint32_t *)(op0), (vuint64m8_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vamominuei64_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominuei64_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint64m8_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vamominuei64_v_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamominuei64_v_u32mf2((uint32_t *)(op0), (vuint64m1_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vamominuei64_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominuei64_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint64m1_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vamominuei8_v_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamominuei8_v_u64m1((uint64_t *)(op0), (vuint8mf8_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vamominuei8_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominuei8_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint8mf8_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vamominuei8_v_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamominuei8_v_u64m2((uint64_t *)(op0), (vuint8mf4_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vamominuei8_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominuei8_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint8mf4_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vamominuei8_v_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamominuei8_v_u64m4((uint64_t *)(op0), (vuint8mf2_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vamominuei8_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominuei8_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint8mf2_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vamominuei8_v_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamominuei8_v_u64m8((uint64_t *)(op0), (vuint8m1_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vamominuei8_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominuei8_v_u64m8_m((vbool8_t)(op0), (uint64_t *)(op1), (vuint8m1_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vamominuei16_v_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamominuei16_v_u64m1((uint64_t *)(op0), (vuint16mf4_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vamominuei16_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominuei16_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint16mf4_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vamominuei16_v_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamominuei16_v_u64m2((uint64_t *)(op0), (vuint16mf2_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vamominuei16_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominuei16_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint16mf2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vamominuei16_v_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamominuei16_v_u64m4((uint64_t *)(op0), (vuint16m1_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vamominuei16_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominuei16_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint16m1_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vamominuei16_v_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamominuei16_v_u64m8((uint64_t *)(op0), (vuint16m2_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vamominuei16_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominuei16_v_u64m8_m((vbool8_t)(op0), (uint64_t *)(op1), (vuint16m2_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vamominuei32_v_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamominuei32_v_u64m1((uint64_t *)(op0), (vuint32mf2_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vamominuei32_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominuei32_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint32mf2_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vamominuei32_v_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamominuei32_v_u64m2((uint64_t *)(op0), (vuint32m1_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vamominuei32_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominuei32_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint32m1_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vamominuei32_v_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamominuei32_v_u64m4((uint64_t *)(op0), (vuint32m2_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vamominuei32_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominuei32_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint32m2_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vamominuei32_v_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamominuei32_v_u64m8((uint64_t *)(op0), (vuint32m4_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vamominuei32_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominuei32_v_u64m8_m((vbool8_t)(op0), (uint64_t *)(op1), (vuint32m4_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vamominuei64_v_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamominuei64_v_u64m1((uint64_t *)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vamominuei64_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominuei64_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vamominuei64_v_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamominuei64_v_u64m2((uint64_t *)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vamominuei64_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominuei64_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vamominuei64_v_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamominuei64_v_u64m4((uint64_t *)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vamominuei64_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominuei64_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vamominuei64_v_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamominuei64_v_u64m8((uint64_t *)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vamominuei64_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamominuei64_v_u64m8_m((vbool8_t)(op0), (uint64_t *)(op1), (vuint64m8_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vamomaxuei8_v_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxuei8_v_u32m1((uint32_t *)(op0), (vuint8mf4_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vamomaxuei8_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxuei8_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint8mf4_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vamomaxuei8_v_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxuei8_v_u32m2((uint32_t *)(op0), (vuint8mf2_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vamomaxuei8_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxuei8_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint8mf2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vamomaxuei8_v_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxuei8_v_u32m4((uint32_t *)(op0), (vuint8m1_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vamomaxuei8_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxuei8_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint8m1_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vamomaxuei8_v_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxuei8_v_u32m8((uint32_t *)(op0), (vuint8m2_t)(op1), (vuint32m8_t)(op2), (size_t)(op3)) +#define vamomaxuei8_v_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxuei8_v_u32m8_m((vbool4_t)(op0), (uint32_t *)(op1), (vuint8m2_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vamomaxuei8_v_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxuei8_v_u32mf2((uint32_t *)(op0), (vuint8mf8_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vamomaxuei8_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxuei8_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint8mf8_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vamomaxuei16_v_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxuei16_v_u32m1((uint32_t *)(op0), (vuint16mf2_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vamomaxuei16_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxuei16_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint16mf2_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vamomaxuei16_v_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxuei16_v_u32m2((uint32_t *)(op0), (vuint16m1_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vamomaxuei16_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxuei16_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint16m1_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vamomaxuei16_v_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxuei16_v_u32m4((uint32_t *)(op0), (vuint16m2_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vamomaxuei16_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxuei16_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint16m2_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vamomaxuei16_v_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxuei16_v_u32m8((uint32_t *)(op0), (vuint16m4_t)(op1), (vuint32m8_t)(op2), (size_t)(op3)) +#define vamomaxuei16_v_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxuei16_v_u32m8_m((vbool4_t)(op0), (uint32_t *)(op1), (vuint16m4_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vamomaxuei16_v_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxuei16_v_u32mf2((uint32_t *)(op0), (vuint16mf4_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vamomaxuei16_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxuei16_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint16mf4_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vamomaxuei32_v_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxuei32_v_u32m1((uint32_t *)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vamomaxuei32_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxuei32_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vamomaxuei32_v_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxuei32_v_u32m2((uint32_t *)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vamomaxuei32_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxuei32_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vamomaxuei32_v_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxuei32_v_u32m4((uint32_t *)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vamomaxuei32_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxuei32_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vamomaxuei32_v_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxuei32_v_u32m8((uint32_t *)(op0), (vuint32m8_t)(op1), (vuint32m8_t)(op2), (size_t)(op3)) +#define vamomaxuei32_v_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxuei32_v_u32m8_m((vbool4_t)(op0), (uint32_t *)(op1), (vuint32m8_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vamomaxuei32_v_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxuei32_v_u32mf2((uint32_t *)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vamomaxuei32_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxuei32_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vamomaxuei64_v_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxuei64_v_u32m1((uint32_t *)(op0), (vuint64m2_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vamomaxuei64_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxuei64_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint64m2_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vamomaxuei64_v_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxuei64_v_u32m2((uint32_t *)(op0), (vuint64m4_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vamomaxuei64_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxuei64_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint64m4_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vamomaxuei64_v_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxuei64_v_u32m4((uint32_t *)(op0), (vuint64m8_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vamomaxuei64_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxuei64_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint64m8_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vamomaxuei64_v_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxuei64_v_u32mf2((uint32_t *)(op0), (vuint64m1_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vamomaxuei64_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxuei64_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint64m1_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vamomaxuei8_v_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxuei8_v_u64m1((uint64_t *)(op0), (vuint8mf8_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vamomaxuei8_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxuei8_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint8mf8_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vamomaxuei8_v_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxuei8_v_u64m2((uint64_t *)(op0), (vuint8mf4_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vamomaxuei8_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxuei8_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint8mf4_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vamomaxuei8_v_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxuei8_v_u64m4((uint64_t *)(op0), (vuint8mf2_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vamomaxuei8_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxuei8_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint8mf2_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vamomaxuei8_v_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxuei8_v_u64m8((uint64_t *)(op0), (vuint8m1_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vamomaxuei8_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxuei8_v_u64m8_m((vbool8_t)(op0), (uint64_t *)(op1), (vuint8m1_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vamomaxuei16_v_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxuei16_v_u64m1((uint64_t *)(op0), (vuint16mf4_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vamomaxuei16_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxuei16_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint16mf4_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vamomaxuei16_v_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxuei16_v_u64m2((uint64_t *)(op0), (vuint16mf2_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vamomaxuei16_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxuei16_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint16mf2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vamomaxuei16_v_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxuei16_v_u64m4((uint64_t *)(op0), (vuint16m1_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vamomaxuei16_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxuei16_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint16m1_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vamomaxuei16_v_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxuei16_v_u64m8((uint64_t *)(op0), (vuint16m2_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vamomaxuei16_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxuei16_v_u64m8_m((vbool8_t)(op0), (uint64_t *)(op1), (vuint16m2_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vamomaxuei32_v_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxuei32_v_u64m1((uint64_t *)(op0), (vuint32mf2_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vamomaxuei32_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxuei32_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint32mf2_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vamomaxuei32_v_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxuei32_v_u64m2((uint64_t *)(op0), (vuint32m1_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vamomaxuei32_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxuei32_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint32m1_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vamomaxuei32_v_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxuei32_v_u64m4((uint64_t *)(op0), (vuint32m2_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vamomaxuei32_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxuei32_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint32m2_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vamomaxuei32_v_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxuei32_v_u64m8((uint64_t *)(op0), (vuint32m4_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vamomaxuei32_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxuei32_v_u64m8_m((vbool8_t)(op0), (uint64_t *)(op1), (vuint32m4_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vamomaxuei64_v_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxuei64_v_u64m1((uint64_t *)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vamomaxuei64_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxuei64_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vamomaxuei64_v_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxuei64_v_u64m2((uint64_t *)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vamomaxuei64_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxuei64_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vamomaxuei64_v_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxuei64_v_u64m4((uint64_t *)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vamomaxuei64_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxuei64_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vamomaxuei64_v_u64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamomaxuei64_v_u64m8((uint64_t *)(op0), (vuint64m8_t)(op1), (vuint64m8_t)(op2), (size_t)(op3)) +#define vamomaxuei64_v_u64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamomaxuei64_v_u64m8_m((vbool8_t)(op0), (uint64_t *)(op1), (vuint64m8_t)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vamoswapei8_v_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei8_v_i32m1((int32_t *)(op0), (vuint8mf4_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vamoswapei8_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei8_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint8mf4_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vamoswapei8_v_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei8_v_i32m2((int32_t *)(op0), (vuint8mf2_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vamoswapei8_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei8_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint8mf2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vamoswapei8_v_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei8_v_i32m4((int32_t *)(op0), (vuint8m1_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vamoswapei8_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei8_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vuint8m1_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vamoswapei8_v_i32m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei8_v_i32m8((int32_t *)(op0), (vuint8m2_t)(op1), (vint32m8_t)(op2), (size_t)(op3)) +#define vamoswapei8_v_i32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei8_v_i32m8_m((vbool4_t)(op0), (int32_t *)(op1), (vuint8m2_t)(op2), (vint32m8_t)(op3), (size_t)(op4)) +#define vamoswapei8_v_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei8_v_i32mf2((int32_t *)(op0), (vuint8mf8_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vamoswapei8_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei8_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint8mf8_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vamoswapei8_v_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei8_v_u32m1((uint32_t *)(op0), (vuint8mf4_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vamoswapei8_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei8_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint8mf4_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vamoswapei8_v_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei8_v_u32m2((uint32_t *)(op0), (vuint8mf2_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vamoswapei8_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei8_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint8mf2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vamoswapei8_v_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei8_v_u32m4((uint32_t *)(op0), (vuint8m1_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vamoswapei8_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei8_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint8m1_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vamoswapei8_v_u32m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei8_v_u32m8((uint32_t *)(op0), (vuint8m2_t)(op1), (vuint32m8_t)(op2), (size_t)(op3)) +#define vamoswapei8_v_u32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei8_v_u32m8_m((vbool4_t)(op0), (uint32_t *)(op1), (vuint8m2_t)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vamoswapei8_v_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei8_v_u32mf2((uint32_t *)(op0), (vuint8mf8_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vamoswapei8_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei8_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint8mf8_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#endif + +#if defined(__riscv_f) && defined(__riscv_zvamo) +#define vamoswapei8_v_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei8_v_f32m1((float *)(op0), (vuint8mf4_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vamoswapei8_v_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei8_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint8mf4_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vamoswapei8_v_f32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei8_v_f32m2((float *)(op0), (vuint8mf2_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vamoswapei8_v_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei8_v_f32m2_m((vbool16_t)(op0), (float *)(op1), (vuint8mf2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vamoswapei8_v_f32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei8_v_f32m4((float *)(op0), (vuint8m1_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vamoswapei8_v_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei8_v_f32m4_m((vbool8_t)(op0), (float *)(op1), (vuint8m1_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vamoswapei8_v_f32m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei8_v_f32m8((float *)(op0), (vuint8m2_t)(op1), (vfloat32m8_t)(op2), (size_t)(op3)) +#define vamoswapei8_v_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei8_v_f32m8_m((vbool4_t)(op0), (float *)(op1), (vuint8m2_t)(op2), (vfloat32m8_t)(op3), (size_t)(op4)) +#define vamoswapei8_v_f32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei8_v_f32mf2((float *)(op0), (vuint8mf8_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vamoswapei8_v_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei8_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint8mf8_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vamoswapei16_v_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei16_v_f32m1((float *)(op0), (vuint16mf2_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vamoswapei16_v_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei16_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint16mf2_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vamoswapei16_v_f32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei16_v_f32m2((float *)(op0), (vuint16m1_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vamoswapei16_v_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei16_v_f32m2_m((vbool16_t)(op0), (float *)(op1), (vuint16m1_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vamoswapei16_v_f32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei16_v_f32m4((float *)(op0), (vuint16m2_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vamoswapei16_v_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei16_v_f32m4_m((vbool8_t)(op0), (float *)(op1), (vuint16m2_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vamoswapei16_v_f32m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei16_v_f32m8((float *)(op0), (vuint16m4_t)(op1), (vfloat32m8_t)(op2), (size_t)(op3)) +#define vamoswapei16_v_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei16_v_f32m8_m((vbool4_t)(op0), (float *)(op1), (vuint16m4_t)(op2), (vfloat32m8_t)(op3), (size_t)(op4)) +#define vamoswapei16_v_f32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei16_v_f32mf2((float *)(op0), (vuint16mf4_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vamoswapei16_v_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei16_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint16mf4_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vamoswapei32_v_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei32_v_f32m1((float *)(op0), (vuint32m1_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vamoswapei32_v_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei32_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vamoswapei32_v_f32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei32_v_f32m2((float *)(op0), (vuint32m2_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vamoswapei32_v_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei32_v_f32m2_m((vbool16_t)(op0), (float *)(op1), (vuint32m2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vamoswapei32_v_f32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei32_v_f32m4((float *)(op0), (vuint32m4_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vamoswapei32_v_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei32_v_f32m4_m((vbool8_t)(op0), (float *)(op1), (vuint32m4_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vamoswapei32_v_f32m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei32_v_f32m8((float *)(op0), (vuint32m8_t)(op1), (vfloat32m8_t)(op2), (size_t)(op3)) +#define vamoswapei32_v_f32m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei32_v_f32m8_m((vbool4_t)(op0), (float *)(op1), (vuint32m8_t)(op2), (vfloat32m8_t)(op3), (size_t)(op4)) +#define vamoswapei32_v_f32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei32_v_f32mf2((float *)(op0), (vuint32mf2_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vamoswapei32_v_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei32_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint32mf2_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vamoswapei64_v_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei64_v_f32m1((float *)(op0), (vuint64m2_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vamoswapei64_v_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei64_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint64m2_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vamoswapei64_v_f32m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei64_v_f32m2((float *)(op0), (vuint64m4_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vamoswapei64_v_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei64_v_f32m2_m((vbool16_t)(op0), (float *)(op1), (vuint64m4_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vamoswapei64_v_f32m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei64_v_f32m4((float *)(op0), (vuint64m8_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vamoswapei64_v_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei64_v_f32m4_m((vbool8_t)(op0), (float *)(op1), (vuint64m8_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vamoswapei64_v_f32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei64_v_f32mf2((float *)(op0), (vuint64m1_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vamoswapei64_v_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei64_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint64m1_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#endif + +#if defined(__riscv_d) && defined(__riscv_zvamo) +#define vamoswapei8_v_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei8_v_f64m1((double *)(op0), (vuint8mf8_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vamoswapei8_v_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei8_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint8mf8_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vamoswapei8_v_f64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei8_v_f64m2((double *)(op0), (vuint8mf4_t)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vamoswapei8_v_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei8_v_f64m2_m((vbool32_t)(op0), (double *)(op1), (vuint8mf4_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vamoswapei8_v_f64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei8_v_f64m4((double *)(op0), (vuint8mf2_t)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vamoswapei8_v_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei8_v_f64m4_m((vbool16_t)(op0), (double *)(op1), (vuint8mf2_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vamoswapei8_v_f64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei8_v_f64m8((double *)(op0), (vuint8m1_t)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vamoswapei8_v_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei8_v_f64m8_m((vbool8_t)(op0), (double *)(op1), (vuint8m1_t)(op2), (vfloat64m8_t)(op3), (size_t)(op4)) +#define vamoswapei16_v_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei16_v_f64m1((double *)(op0), (vuint16mf4_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vamoswapei16_v_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei16_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint16mf4_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vamoswapei16_v_f64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei16_v_f64m2((double *)(op0), (vuint16mf2_t)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vamoswapei16_v_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei16_v_f64m2_m((vbool32_t)(op0), (double *)(op1), (vuint16mf2_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vamoswapei16_v_f64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei16_v_f64m4((double *)(op0), (vuint16m1_t)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vamoswapei16_v_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei16_v_f64m4_m((vbool16_t)(op0), (double *)(op1), (vuint16m1_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vamoswapei16_v_f64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei16_v_f64m8((double *)(op0), (vuint16m2_t)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vamoswapei16_v_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei16_v_f64m8_m((vbool8_t)(op0), (double *)(op1), (vuint16m2_t)(op2), (vfloat64m8_t)(op3), (size_t)(op4)) +#define vamoswapei32_v_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei32_v_f64m1((double *)(op0), (vuint32mf2_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vamoswapei32_v_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei32_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint32mf2_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vamoswapei32_v_f64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei32_v_f64m2((double *)(op0), (vuint32m1_t)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vamoswapei32_v_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei32_v_f64m2_m((vbool32_t)(op0), (double *)(op1), (vuint32m1_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vamoswapei32_v_f64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei32_v_f64m4((double *)(op0), (vuint32m2_t)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vamoswapei32_v_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei32_v_f64m4_m((vbool16_t)(op0), (double *)(op1), (vuint32m2_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vamoswapei32_v_f64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei32_v_f64m8((double *)(op0), (vuint32m4_t)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vamoswapei32_v_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei32_v_f64m8_m((vbool8_t)(op0), (double *)(op1), (vuint32m4_t)(op2), (vfloat64m8_t)(op3), (size_t)(op4)) +#define vamoswapei64_v_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei64_v_f64m1((double *)(op0), (vuint64m1_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vamoswapei64_v_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei64_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint64m1_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vamoswapei64_v_f64m2(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei64_v_f64m2((double *)(op0), (vuint64m2_t)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vamoswapei64_v_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei64_v_f64m2_m((vbool32_t)(op0), (double *)(op1), (vuint64m2_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vamoswapei64_v_f64m4(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei64_v_f64m4((double *)(op0), (vuint64m4_t)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vamoswapei64_v_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei64_v_f64m4_m((vbool16_t)(op0), (double *)(op1), (vuint64m4_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vamoswapei64_v_f64m8(op0, op1, op2, op3) \ +__builtin_rvv_vamoswapei64_v_f64m8((double *)(op0), (vuint64m8_t)(op1), (vfloat64m8_t)(op2), (size_t)(op3)) +#define vamoswapei64_v_f64m8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vamoswapei64_v_f64m8_m((vbool8_t)(op0), (double *)(op1), (vuint64m8_t)(op2), (vfloat64m8_t)(op3), (size_t)(op4)) +#endif + +#if defined(__riscv_zvlsseg) +#define vloxseg2ei8_v_i64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (const int64_t *)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vbool64_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (const int64_t *)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vloxseg2ei8_v_i64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_i64m2((vint64m2_t *)(op0), (vint64m2_t *)(op1), (const int64_t *)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_i64m2_m((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vbool32_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (const int64_t *)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vloxseg2ei8_v_i64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_i64m4((vint64m4_t *)(op0), (vint64m4_t *)(op1), (const int64_t *)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_i64m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_i64m4_m((vint64m4_t *)(op0), (vint64m4_t *)(op1), (vbool16_t)(op2), (vint64m4_t)(op3), (vint64m4_t)(op4), (const int64_t *)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vloxseg2ei8_v_u64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (const uint64_t *)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vbool64_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (const uint64_t *)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vloxseg2ei8_v_u64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_u64m2((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (const uint64_t *)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_u64m2_m((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vbool32_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (const uint64_t *)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vloxseg2ei8_v_u64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_u64m4((vuint64m4_t *)(op0), (vuint64m4_t *)(op1), (const uint64_t *)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_u64m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_u64m4_m((vuint64m4_t *)(op0), (vuint64m4_t *)(op1), (vbool16_t)(op2), (vuint64m4_t)(op3), (vuint64m4_t)(op4), (const uint64_t *)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vloxseg3ei8_v_i64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei8_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (const int64_t *)(op3), (vuint8mf8_t)(op4), (size_t)(op5)) +#define vloxseg3ei8_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei8_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vbool64_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (const int64_t *)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vloxseg3ei8_v_i64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei8_v_i64m2((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vint64m2_t *)(op2), (const int64_t *)(op3), (vuint8mf4_t)(op4), (size_t)(op5)) +#define vloxseg3ei8_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei8_v_i64m2_m((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vint64m2_t *)(op2), (vbool32_t)(op3), (vint64m2_t)(op4), (vint64m2_t)(op5), (vint64m2_t)(op6), (const int64_t *)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vloxseg3ei8_v_u64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei8_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (const uint64_t *)(op3), (vuint8mf8_t)(op4), (size_t)(op5)) +#define vloxseg3ei8_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei8_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vbool64_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (const uint64_t *)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vloxseg3ei8_v_u64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei8_v_u64m2((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vuint64m2_t *)(op2), (const uint64_t *)(op3), (vuint8mf4_t)(op4), (size_t)(op5)) +#define vloxseg3ei8_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei8_v_u64m2_m((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vuint64m2_t *)(op2), (vbool32_t)(op3), (vuint64m2_t)(op4), (vuint64m2_t)(op5), (vuint64m2_t)(op6), (const uint64_t *)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vloxseg4ei8_v_i64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei8_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (const int64_t *)(op4), (vuint8mf8_t)(op5), (size_t)(op6)) +#define vloxseg4ei8_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei8_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vbool64_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (const int64_t *)(op9), (vuint8mf8_t)(op10), (size_t)(op11)) +#define vloxseg4ei8_v_i64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei8_v_i64m2((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vint64m2_t *)(op2), (vint64m2_t *)(op3), (const int64_t *)(op4), (vuint8mf4_t)(op5), (size_t)(op6)) +#define vloxseg4ei8_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei8_v_i64m2_m((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vint64m2_t *)(op2), (vint64m2_t *)(op3), (vbool32_t)(op4), (vint64m2_t)(op5), (vint64m2_t)(op6), (vint64m2_t)(op7), (vint64m2_t)(op8), (const int64_t *)(op9), (vuint8mf4_t)(op10), (size_t)(op11)) +#define vloxseg4ei8_v_u64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei8_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (const uint64_t *)(op4), (vuint8mf8_t)(op5), (size_t)(op6)) +#define vloxseg4ei8_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei8_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vbool64_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (const uint64_t *)(op9), (vuint8mf8_t)(op10), (size_t)(op11)) +#define vloxseg4ei8_v_u64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei8_v_u64m2((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vuint64m2_t *)(op2), (vuint64m2_t *)(op3), (const uint64_t *)(op4), (vuint8mf4_t)(op5), (size_t)(op6)) +#define vloxseg4ei8_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei8_v_u64m2_m((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vuint64m2_t *)(op2), (vuint64m2_t *)(op3), (vbool32_t)(op4), (vuint64m2_t)(op5), (vuint64m2_t)(op6), (vuint64m2_t)(op7), (vuint64m2_t)(op8), (const uint64_t *)(op9), (vuint8mf4_t)(op10), (size_t)(op11)) +#define vloxseg5ei8_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei8_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (const int64_t *)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vloxseg5ei8_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei8_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vbool64_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (const int64_t *)(op11), (vuint8mf8_t)(op12), (size_t)(op13)) +#define vloxseg5ei8_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei8_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (const uint64_t *)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vloxseg5ei8_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei8_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vbool64_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (const uint64_t *)(op11), (vuint8mf8_t)(op12), (size_t)(op13)) +#define vloxseg6ei8_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei8_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (const int64_t *)(op6), (vuint8mf8_t)(op7), (size_t)(op8)) +#define vloxseg6ei8_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei8_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vbool64_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (vint64m1_t)(op11), (vint64m1_t)(op12), (const int64_t *)(op13), (vuint8mf8_t)(op14), (size_t)(op15)) +#define vloxseg6ei8_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei8_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (const uint64_t *)(op6), (vuint8mf8_t)(op7), (size_t)(op8)) +#define vloxseg6ei8_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei8_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vbool64_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (vuint64m1_t)(op11), (vuint64m1_t)(op12), (const uint64_t *)(op13), (vuint8mf8_t)(op14), (size_t)(op15)) +#define vloxseg7ei8_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei8_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vint64m1_t *)(op6), (const int64_t *)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vloxseg7ei8_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei8_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vint64m1_t *)(op6), (vbool64_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (vint64m1_t)(op11), (vint64m1_t)(op12), (vint64m1_t)(op13), (vint64m1_t)(op14), (const int64_t *)(op15), (vuint8mf8_t)(op16), (size_t)(op17)) +#define vloxseg7ei8_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei8_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vuint64m1_t *)(op6), (const uint64_t *)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vloxseg7ei8_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei8_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vuint64m1_t *)(op6), (vbool64_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (vuint64m1_t)(op11), (vuint64m1_t)(op12), (vuint64m1_t)(op13), (vuint64m1_t)(op14), (const uint64_t *)(op15), (vuint8mf8_t)(op16), (size_t)(op17)) +#define vloxseg8ei8_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei8_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vint64m1_t *)(op6), (vint64m1_t *)(op7), (const int64_t *)(op8), (vuint8mf8_t)(op9), (size_t)(op10)) +#define vloxseg8ei8_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei8_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vint64m1_t *)(op6), (vint64m1_t *)(op7), (vbool64_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (vint64m1_t)(op11), (vint64m1_t)(op12), (vint64m1_t)(op13), (vint64m1_t)(op14), (vint64m1_t)(op15), (vint64m1_t)(op16), (const int64_t *)(op17), (vuint8mf8_t)(op18), (size_t)(op19)) +#define vloxseg8ei8_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei8_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vuint64m1_t *)(op6), (vuint64m1_t *)(op7), (const uint64_t *)(op8), (vuint8mf8_t)(op9), (size_t)(op10)) +#define vloxseg8ei8_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei8_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vuint64m1_t *)(op6), (vuint64m1_t *)(op7), (vbool64_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (vuint64m1_t)(op11), (vuint64m1_t)(op12), (vuint64m1_t)(op13), (vuint64m1_t)(op14), (vuint64m1_t)(op15), (vuint64m1_t)(op16), (const uint64_t *)(op17), (vuint8mf8_t)(op18), (size_t)(op19)) +#define vloxseg2ei16_v_i64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (const int64_t *)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vbool64_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (const int64_t *)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vloxseg2ei16_v_i64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_i64m2((vint64m2_t *)(op0), (vint64m2_t *)(op1), (const int64_t *)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_i64m2_m((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vbool32_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (const int64_t *)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vloxseg2ei16_v_i64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_i64m4((vint64m4_t *)(op0), (vint64m4_t *)(op1), (const int64_t *)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_i64m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_i64m4_m((vint64m4_t *)(op0), (vint64m4_t *)(op1), (vbool16_t)(op2), (vint64m4_t)(op3), (vint64m4_t)(op4), (const int64_t *)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vloxseg2ei16_v_u64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (const uint64_t *)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vbool64_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (const uint64_t *)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vloxseg2ei16_v_u64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_u64m2((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (const uint64_t *)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_u64m2_m((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vbool32_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (const uint64_t *)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vloxseg2ei16_v_u64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_u64m4((vuint64m4_t *)(op0), (vuint64m4_t *)(op1), (const uint64_t *)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_u64m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_u64m4_m((vuint64m4_t *)(op0), (vuint64m4_t *)(op1), (vbool16_t)(op2), (vuint64m4_t)(op3), (vuint64m4_t)(op4), (const uint64_t *)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vloxseg3ei16_v_i64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei16_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (const int64_t *)(op3), (vuint16mf4_t)(op4), (size_t)(op5)) +#define vloxseg3ei16_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei16_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vbool64_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (const int64_t *)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vloxseg3ei16_v_i64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei16_v_i64m2((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vint64m2_t *)(op2), (const int64_t *)(op3), (vuint16mf2_t)(op4), (size_t)(op5)) +#define vloxseg3ei16_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei16_v_i64m2_m((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vint64m2_t *)(op2), (vbool32_t)(op3), (vint64m2_t)(op4), (vint64m2_t)(op5), (vint64m2_t)(op6), (const int64_t *)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vloxseg3ei16_v_u64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei16_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (const uint64_t *)(op3), (vuint16mf4_t)(op4), (size_t)(op5)) +#define vloxseg3ei16_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei16_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vbool64_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (const uint64_t *)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vloxseg3ei16_v_u64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei16_v_u64m2((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vuint64m2_t *)(op2), (const uint64_t *)(op3), (vuint16mf2_t)(op4), (size_t)(op5)) +#define vloxseg3ei16_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei16_v_u64m2_m((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vuint64m2_t *)(op2), (vbool32_t)(op3), (vuint64m2_t)(op4), (vuint64m2_t)(op5), (vuint64m2_t)(op6), (const uint64_t *)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vloxseg4ei16_v_i64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei16_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (const int64_t *)(op4), (vuint16mf4_t)(op5), (size_t)(op6)) +#define vloxseg4ei16_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei16_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vbool64_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (const int64_t *)(op9), (vuint16mf4_t)(op10), (size_t)(op11)) +#define vloxseg4ei16_v_i64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei16_v_i64m2((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vint64m2_t *)(op2), (vint64m2_t *)(op3), (const int64_t *)(op4), (vuint16mf2_t)(op5), (size_t)(op6)) +#define vloxseg4ei16_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei16_v_i64m2_m((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vint64m2_t *)(op2), (vint64m2_t *)(op3), (vbool32_t)(op4), (vint64m2_t)(op5), (vint64m2_t)(op6), (vint64m2_t)(op7), (vint64m2_t)(op8), (const int64_t *)(op9), (vuint16mf2_t)(op10), (size_t)(op11)) +#define vloxseg4ei16_v_u64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei16_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (const uint64_t *)(op4), (vuint16mf4_t)(op5), (size_t)(op6)) +#define vloxseg4ei16_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei16_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vbool64_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (const uint64_t *)(op9), (vuint16mf4_t)(op10), (size_t)(op11)) +#define vloxseg4ei16_v_u64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei16_v_u64m2((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vuint64m2_t *)(op2), (vuint64m2_t *)(op3), (const uint64_t *)(op4), (vuint16mf2_t)(op5), (size_t)(op6)) +#define vloxseg4ei16_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei16_v_u64m2_m((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vuint64m2_t *)(op2), (vuint64m2_t *)(op3), (vbool32_t)(op4), (vuint64m2_t)(op5), (vuint64m2_t)(op6), (vuint64m2_t)(op7), (vuint64m2_t)(op8), (const uint64_t *)(op9), (vuint16mf2_t)(op10), (size_t)(op11)) +#define vloxseg5ei16_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei16_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (const int64_t *)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vloxseg5ei16_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei16_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vbool64_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (const int64_t *)(op11), (vuint16mf4_t)(op12), (size_t)(op13)) +#define vloxseg5ei16_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei16_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (const uint64_t *)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vloxseg5ei16_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei16_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vbool64_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (const uint64_t *)(op11), (vuint16mf4_t)(op12), (size_t)(op13)) +#define vloxseg6ei16_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei16_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (const int64_t *)(op6), (vuint16mf4_t)(op7), (size_t)(op8)) +#define vloxseg6ei16_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei16_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vbool64_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (vint64m1_t)(op11), (vint64m1_t)(op12), (const int64_t *)(op13), (vuint16mf4_t)(op14), (size_t)(op15)) +#define vloxseg6ei16_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei16_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (const uint64_t *)(op6), (vuint16mf4_t)(op7), (size_t)(op8)) +#define vloxseg6ei16_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei16_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vbool64_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (vuint64m1_t)(op11), (vuint64m1_t)(op12), (const uint64_t *)(op13), (vuint16mf4_t)(op14), (size_t)(op15)) +#define vloxseg7ei16_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei16_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vint64m1_t *)(op6), (const int64_t *)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vloxseg7ei16_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei16_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vint64m1_t *)(op6), (vbool64_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (vint64m1_t)(op11), (vint64m1_t)(op12), (vint64m1_t)(op13), (vint64m1_t)(op14), (const int64_t *)(op15), (vuint16mf4_t)(op16), (size_t)(op17)) +#define vloxseg7ei16_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei16_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vuint64m1_t *)(op6), (const uint64_t *)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vloxseg7ei16_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei16_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vuint64m1_t *)(op6), (vbool64_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (vuint64m1_t)(op11), (vuint64m1_t)(op12), (vuint64m1_t)(op13), (vuint64m1_t)(op14), (const uint64_t *)(op15), (vuint16mf4_t)(op16), (size_t)(op17)) +#define vloxseg8ei16_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei16_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vint64m1_t *)(op6), (vint64m1_t *)(op7), (const int64_t *)(op8), (vuint16mf4_t)(op9), (size_t)(op10)) +#define vloxseg8ei16_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei16_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vint64m1_t *)(op6), (vint64m1_t *)(op7), (vbool64_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (vint64m1_t)(op11), (vint64m1_t)(op12), (vint64m1_t)(op13), (vint64m1_t)(op14), (vint64m1_t)(op15), (vint64m1_t)(op16), (const int64_t *)(op17), (vuint16mf4_t)(op18), (size_t)(op19)) +#define vloxseg8ei16_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei16_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vuint64m1_t *)(op6), (vuint64m1_t *)(op7), (const uint64_t *)(op8), (vuint16mf4_t)(op9), (size_t)(op10)) +#define vloxseg8ei16_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei16_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vuint64m1_t *)(op6), (vuint64m1_t *)(op7), (vbool64_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (vuint64m1_t)(op11), (vuint64m1_t)(op12), (vuint64m1_t)(op13), (vuint64m1_t)(op14), (vuint64m1_t)(op15), (vuint64m1_t)(op16), (const uint64_t *)(op17), (vuint16mf4_t)(op18), (size_t)(op19)) +#define vloxseg2ei32_v_i64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (const int64_t *)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vbool64_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (const int64_t *)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vloxseg2ei32_v_i64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_i64m2((vint64m2_t *)(op0), (vint64m2_t *)(op1), (const int64_t *)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_i64m2_m((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vbool32_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (const int64_t *)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vloxseg2ei32_v_i64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_i64m4((vint64m4_t *)(op0), (vint64m4_t *)(op1), (const int64_t *)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_i64m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_i64m4_m((vint64m4_t *)(op0), (vint64m4_t *)(op1), (vbool16_t)(op2), (vint64m4_t)(op3), (vint64m4_t)(op4), (const int64_t *)(op5), (vuint32m2_t)(op6), (size_t)(op7)) +#define vloxseg2ei32_v_u64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (const uint64_t *)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vbool64_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (const uint64_t *)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vloxseg2ei32_v_u64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_u64m2((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (const uint64_t *)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_u64m2_m((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vbool32_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (const uint64_t *)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vloxseg2ei32_v_u64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_u64m4((vuint64m4_t *)(op0), (vuint64m4_t *)(op1), (const uint64_t *)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_u64m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_u64m4_m((vuint64m4_t *)(op0), (vuint64m4_t *)(op1), (vbool16_t)(op2), (vuint64m4_t)(op3), (vuint64m4_t)(op4), (const uint64_t *)(op5), (vuint32m2_t)(op6), (size_t)(op7)) +#define vloxseg3ei32_v_i64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei32_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (const int64_t *)(op3), (vuint32mf2_t)(op4), (size_t)(op5)) +#define vloxseg3ei32_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei32_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vbool64_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (const int64_t *)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vloxseg3ei32_v_i64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei32_v_i64m2((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vint64m2_t *)(op2), (const int64_t *)(op3), (vuint32m1_t)(op4), (size_t)(op5)) +#define vloxseg3ei32_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei32_v_i64m2_m((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vint64m2_t *)(op2), (vbool32_t)(op3), (vint64m2_t)(op4), (vint64m2_t)(op5), (vint64m2_t)(op6), (const int64_t *)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vloxseg3ei32_v_u64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei32_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (const uint64_t *)(op3), (vuint32mf2_t)(op4), (size_t)(op5)) +#define vloxseg3ei32_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei32_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vbool64_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (const uint64_t *)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vloxseg3ei32_v_u64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei32_v_u64m2((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vuint64m2_t *)(op2), (const uint64_t *)(op3), (vuint32m1_t)(op4), (size_t)(op5)) +#define vloxseg3ei32_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei32_v_u64m2_m((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vuint64m2_t *)(op2), (vbool32_t)(op3), (vuint64m2_t)(op4), (vuint64m2_t)(op5), (vuint64m2_t)(op6), (const uint64_t *)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vloxseg4ei32_v_i64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei32_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (const int64_t *)(op4), (vuint32mf2_t)(op5), (size_t)(op6)) +#define vloxseg4ei32_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei32_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vbool64_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (const int64_t *)(op9), (vuint32mf2_t)(op10), (size_t)(op11)) +#define vloxseg4ei32_v_i64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei32_v_i64m2((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vint64m2_t *)(op2), (vint64m2_t *)(op3), (const int64_t *)(op4), (vuint32m1_t)(op5), (size_t)(op6)) +#define vloxseg4ei32_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei32_v_i64m2_m((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vint64m2_t *)(op2), (vint64m2_t *)(op3), (vbool32_t)(op4), (vint64m2_t)(op5), (vint64m2_t)(op6), (vint64m2_t)(op7), (vint64m2_t)(op8), (const int64_t *)(op9), (vuint32m1_t)(op10), (size_t)(op11)) +#define vloxseg4ei32_v_u64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei32_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (const uint64_t *)(op4), (vuint32mf2_t)(op5), (size_t)(op6)) +#define vloxseg4ei32_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei32_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vbool64_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (const uint64_t *)(op9), (vuint32mf2_t)(op10), (size_t)(op11)) +#define vloxseg4ei32_v_u64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei32_v_u64m2((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vuint64m2_t *)(op2), (vuint64m2_t *)(op3), (const uint64_t *)(op4), (vuint32m1_t)(op5), (size_t)(op6)) +#define vloxseg4ei32_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei32_v_u64m2_m((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vuint64m2_t *)(op2), (vuint64m2_t *)(op3), (vbool32_t)(op4), (vuint64m2_t)(op5), (vuint64m2_t)(op6), (vuint64m2_t)(op7), (vuint64m2_t)(op8), (const uint64_t *)(op9), (vuint32m1_t)(op10), (size_t)(op11)) +#define vloxseg5ei32_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei32_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (const int64_t *)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vloxseg5ei32_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei32_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vbool64_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (const int64_t *)(op11), (vuint32mf2_t)(op12), (size_t)(op13)) +#define vloxseg5ei32_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei32_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (const uint64_t *)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vloxseg5ei32_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei32_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vbool64_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (const uint64_t *)(op11), (vuint32mf2_t)(op12), (size_t)(op13)) +#define vloxseg6ei32_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei32_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (const int64_t *)(op6), (vuint32mf2_t)(op7), (size_t)(op8)) +#define vloxseg6ei32_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei32_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vbool64_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (vint64m1_t)(op11), (vint64m1_t)(op12), (const int64_t *)(op13), (vuint32mf2_t)(op14), (size_t)(op15)) +#define vloxseg6ei32_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei32_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (const uint64_t *)(op6), (vuint32mf2_t)(op7), (size_t)(op8)) +#define vloxseg6ei32_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei32_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vbool64_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (vuint64m1_t)(op11), (vuint64m1_t)(op12), (const uint64_t *)(op13), (vuint32mf2_t)(op14), (size_t)(op15)) +#define vloxseg7ei32_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei32_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vint64m1_t *)(op6), (const int64_t *)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vloxseg7ei32_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei32_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vint64m1_t *)(op6), (vbool64_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (vint64m1_t)(op11), (vint64m1_t)(op12), (vint64m1_t)(op13), (vint64m1_t)(op14), (const int64_t *)(op15), (vuint32mf2_t)(op16), (size_t)(op17)) +#define vloxseg7ei32_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei32_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vuint64m1_t *)(op6), (const uint64_t *)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vloxseg7ei32_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei32_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vuint64m1_t *)(op6), (vbool64_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (vuint64m1_t)(op11), (vuint64m1_t)(op12), (vuint64m1_t)(op13), (vuint64m1_t)(op14), (const uint64_t *)(op15), (vuint32mf2_t)(op16), (size_t)(op17)) +#define vloxseg8ei32_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei32_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vint64m1_t *)(op6), (vint64m1_t *)(op7), (const int64_t *)(op8), (vuint32mf2_t)(op9), (size_t)(op10)) +#define vloxseg8ei32_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei32_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vint64m1_t *)(op6), (vint64m1_t *)(op7), (vbool64_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (vint64m1_t)(op11), (vint64m1_t)(op12), (vint64m1_t)(op13), (vint64m1_t)(op14), (vint64m1_t)(op15), (vint64m1_t)(op16), (const int64_t *)(op17), (vuint32mf2_t)(op18), (size_t)(op19)) +#define vloxseg8ei32_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei32_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vuint64m1_t *)(op6), (vuint64m1_t *)(op7), (const uint64_t *)(op8), (vuint32mf2_t)(op9), (size_t)(op10)) +#define vloxseg8ei32_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei32_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vuint64m1_t *)(op6), (vuint64m1_t *)(op7), (vbool64_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (vuint64m1_t)(op11), (vuint64m1_t)(op12), (vuint64m1_t)(op13), (vuint64m1_t)(op14), (vuint64m1_t)(op15), (vuint64m1_t)(op16), (const uint64_t *)(op17), (vuint32mf2_t)(op18), (size_t)(op19)) +#define vloxseg2ei64_v_i64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei64_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (const int64_t *)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vloxseg2ei64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei64_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vbool64_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (const int64_t *)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vloxseg2ei64_v_i64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei64_v_i64m2((vint64m2_t *)(op0), (vint64m2_t *)(op1), (const int64_t *)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vloxseg2ei64_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei64_v_i64m2_m((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vbool32_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (const int64_t *)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vloxseg2ei64_v_i64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei64_v_i64m4((vint64m4_t *)(op0), (vint64m4_t *)(op1), (const int64_t *)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vloxseg2ei64_v_i64m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei64_v_i64m4_m((vint64m4_t *)(op0), (vint64m4_t *)(op1), (vbool16_t)(op2), (vint64m4_t)(op3), (vint64m4_t)(op4), (const int64_t *)(op5), (vuint64m4_t)(op6), (size_t)(op7)) +#define vloxseg2ei64_v_u64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei64_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (const uint64_t *)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vloxseg2ei64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei64_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vbool64_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (const uint64_t *)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vloxseg2ei64_v_u64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei64_v_u64m2((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (const uint64_t *)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vloxseg2ei64_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei64_v_u64m2_m((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vbool32_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (const uint64_t *)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vloxseg2ei64_v_u64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei64_v_u64m4((vuint64m4_t *)(op0), (vuint64m4_t *)(op1), (const uint64_t *)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vloxseg2ei64_v_u64m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei64_v_u64m4_m((vuint64m4_t *)(op0), (vuint64m4_t *)(op1), (vbool16_t)(op2), (vuint64m4_t)(op3), (vuint64m4_t)(op4), (const uint64_t *)(op5), (vuint64m4_t)(op6), (size_t)(op7)) +#define vloxseg3ei64_v_i64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei64_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (const int64_t *)(op3), (vuint64m1_t)(op4), (size_t)(op5)) +#define vloxseg3ei64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei64_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vbool64_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (const int64_t *)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vloxseg3ei64_v_i64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei64_v_i64m2((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vint64m2_t *)(op2), (const int64_t *)(op3), (vuint64m2_t)(op4), (size_t)(op5)) +#define vloxseg3ei64_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei64_v_i64m2_m((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vint64m2_t *)(op2), (vbool32_t)(op3), (vint64m2_t)(op4), (vint64m2_t)(op5), (vint64m2_t)(op6), (const int64_t *)(op7), (vuint64m2_t)(op8), (size_t)(op9)) +#define vloxseg3ei64_v_u64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei64_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (const uint64_t *)(op3), (vuint64m1_t)(op4), (size_t)(op5)) +#define vloxseg3ei64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei64_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vbool64_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (const uint64_t *)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vloxseg3ei64_v_u64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei64_v_u64m2((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vuint64m2_t *)(op2), (const uint64_t *)(op3), (vuint64m2_t)(op4), (size_t)(op5)) +#define vloxseg3ei64_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei64_v_u64m2_m((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vuint64m2_t *)(op2), (vbool32_t)(op3), (vuint64m2_t)(op4), (vuint64m2_t)(op5), (vuint64m2_t)(op6), (const uint64_t *)(op7), (vuint64m2_t)(op8), (size_t)(op9)) +#define vloxseg4ei64_v_i64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei64_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (const int64_t *)(op4), (vuint64m1_t)(op5), (size_t)(op6)) +#define vloxseg4ei64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei64_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vbool64_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (const int64_t *)(op9), (vuint64m1_t)(op10), (size_t)(op11)) +#define vloxseg4ei64_v_i64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei64_v_i64m2((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vint64m2_t *)(op2), (vint64m2_t *)(op3), (const int64_t *)(op4), (vuint64m2_t)(op5), (size_t)(op6)) +#define vloxseg4ei64_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei64_v_i64m2_m((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vint64m2_t *)(op2), (vint64m2_t *)(op3), (vbool32_t)(op4), (vint64m2_t)(op5), (vint64m2_t)(op6), (vint64m2_t)(op7), (vint64m2_t)(op8), (const int64_t *)(op9), (vuint64m2_t)(op10), (size_t)(op11)) +#define vloxseg4ei64_v_u64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei64_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (const uint64_t *)(op4), (vuint64m1_t)(op5), (size_t)(op6)) +#define vloxseg4ei64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei64_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vbool64_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (const uint64_t *)(op9), (vuint64m1_t)(op10), (size_t)(op11)) +#define vloxseg4ei64_v_u64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei64_v_u64m2((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vuint64m2_t *)(op2), (vuint64m2_t *)(op3), (const uint64_t *)(op4), (vuint64m2_t)(op5), (size_t)(op6)) +#define vloxseg4ei64_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei64_v_u64m2_m((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vuint64m2_t *)(op2), (vuint64m2_t *)(op3), (vbool32_t)(op4), (vuint64m2_t)(op5), (vuint64m2_t)(op6), (vuint64m2_t)(op7), (vuint64m2_t)(op8), (const uint64_t *)(op9), (vuint64m2_t)(op10), (size_t)(op11)) +#define vloxseg5ei64_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei64_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (const int64_t *)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vloxseg5ei64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei64_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vbool64_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (const int64_t *)(op11), (vuint64m1_t)(op12), (size_t)(op13)) +#define vloxseg5ei64_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei64_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (const uint64_t *)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vloxseg5ei64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei64_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vbool64_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (const uint64_t *)(op11), (vuint64m1_t)(op12), (size_t)(op13)) +#define vloxseg6ei64_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei64_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (const int64_t *)(op6), (vuint64m1_t)(op7), (size_t)(op8)) +#define vloxseg6ei64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei64_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vbool64_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (vint64m1_t)(op11), (vint64m1_t)(op12), (const int64_t *)(op13), (vuint64m1_t)(op14), (size_t)(op15)) +#define vloxseg6ei64_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei64_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (const uint64_t *)(op6), (vuint64m1_t)(op7), (size_t)(op8)) +#define vloxseg6ei64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei64_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vbool64_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (vuint64m1_t)(op11), (vuint64m1_t)(op12), (const uint64_t *)(op13), (vuint64m1_t)(op14), (size_t)(op15)) +#define vloxseg7ei64_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei64_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vint64m1_t *)(op6), (const int64_t *)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vloxseg7ei64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei64_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vint64m1_t *)(op6), (vbool64_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (vint64m1_t)(op11), (vint64m1_t)(op12), (vint64m1_t)(op13), (vint64m1_t)(op14), (const int64_t *)(op15), (vuint64m1_t)(op16), (size_t)(op17)) +#define vloxseg7ei64_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei64_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vuint64m1_t *)(op6), (const uint64_t *)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vloxseg7ei64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei64_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vuint64m1_t *)(op6), (vbool64_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (vuint64m1_t)(op11), (vuint64m1_t)(op12), (vuint64m1_t)(op13), (vuint64m1_t)(op14), (const uint64_t *)(op15), (vuint64m1_t)(op16), (size_t)(op17)) +#define vloxseg8ei64_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei64_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vint64m1_t *)(op6), (vint64m1_t *)(op7), (const int64_t *)(op8), (vuint64m1_t)(op9), (size_t)(op10)) +#define vloxseg8ei64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei64_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vint64m1_t *)(op6), (vint64m1_t *)(op7), (vbool64_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (vint64m1_t)(op11), (vint64m1_t)(op12), (vint64m1_t)(op13), (vint64m1_t)(op14), (vint64m1_t)(op15), (vint64m1_t)(op16), (const int64_t *)(op17), (vuint64m1_t)(op18), (size_t)(op19)) +#define vloxseg8ei64_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei64_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vuint64m1_t *)(op6), (vuint64m1_t *)(op7), (const uint64_t *)(op8), (vuint64m1_t)(op9), (size_t)(op10)) +#define vloxseg8ei64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei64_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vuint64m1_t *)(op6), (vuint64m1_t *)(op7), (vbool64_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (vuint64m1_t)(op11), (vuint64m1_t)(op12), (vuint64m1_t)(op13), (vuint64m1_t)(op14), (vuint64m1_t)(op15), (vuint64m1_t)(op16), (const uint64_t *)(op17), (vuint64m1_t)(op18), (size_t)(op19)) +#define vsseg3e8_v_i8m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg3e8_v_i8m1((int8_t *)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vsseg3e8_v_i8m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg3e8_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (size_t)(op5)) +#define vsseg3e8_v_i8m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg3e8_v_i8m2((int8_t *)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vsseg3e8_v_i8m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg3e8_v_i8m2_m((vbool4_t)(op0), (int8_t *)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (size_t)(op5)) +#define vsseg3e8_v_i8mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg3e8_v_i8mf2((int8_t *)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vsseg3e8_v_i8mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg3e8_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (size_t)(op5)) +#define vsseg3e8_v_i8mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg3e8_v_i8mf4((int8_t *)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vsseg3e8_v_i8mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg3e8_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (size_t)(op5)) +#define vsseg3e8_v_i8mf8(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg3e8_v_i8mf8((int8_t *)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vsseg3e8_v_i8mf8_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg3e8_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (size_t)(op5)) +#define vsseg3e8_v_u8m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg3e8_v_u8m1((uint8_t *)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vsseg3e8_v_u8m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg3e8_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (size_t)(op5)) +#define vsseg3e8_v_u8m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg3e8_v_u8m2((uint8_t *)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vsseg3e8_v_u8m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg3e8_v_u8m2_m((vbool4_t)(op0), (uint8_t *)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (size_t)(op5)) +#define vsseg3e8_v_u8mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg3e8_v_u8mf2((uint8_t *)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vsseg3e8_v_u8mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg3e8_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (size_t)(op5)) +#define vsseg3e8_v_u8mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg3e8_v_u8mf4((uint8_t *)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vsseg3e8_v_u8mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg3e8_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (size_t)(op5)) +#define vsseg3e8_v_u8mf8(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg3e8_v_u8mf8((uint8_t *)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vsseg3e8_v_u8mf8_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg3e8_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (size_t)(op5)) +#define vsseg4e8_v_i8m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg4e8_v_i8m1((int8_t *)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (size_t)(op5)) +#define vsseg4e8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg4e8_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (size_t)(op6)) +#define vsseg4e8_v_i8m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg4e8_v_i8m2((int8_t *)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (size_t)(op5)) +#define vsseg4e8_v_i8m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg4e8_v_i8m2_m((vbool4_t)(op0), (int8_t *)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (vint8m2_t)(op5), (size_t)(op6)) +#define vsseg4e8_v_i8mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg4e8_v_i8mf2((int8_t *)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (size_t)(op5)) +#define vsseg4e8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg4e8_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (size_t)(op6)) +#define vsseg4e8_v_i8mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg4e8_v_i8mf4((int8_t *)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (size_t)(op5)) +#define vsseg4e8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg4e8_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (size_t)(op6)) +#define vsseg4e8_v_i8mf8(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg4e8_v_i8mf8((int8_t *)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (size_t)(op5)) +#define vsseg4e8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg4e8_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (size_t)(op6)) +#define vsseg4e8_v_u8m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg4e8_v_u8m1((uint8_t *)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (size_t)(op5)) +#define vsseg4e8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg4e8_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (size_t)(op6)) +#define vsseg4e8_v_u8m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg4e8_v_u8m2((uint8_t *)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (size_t)(op5)) +#define vsseg4e8_v_u8m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg4e8_v_u8m2_m((vbool4_t)(op0), (uint8_t *)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (vuint8m2_t)(op5), (size_t)(op6)) +#define vsseg4e8_v_u8mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg4e8_v_u8mf2((uint8_t *)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (size_t)(op5)) +#define vsseg4e8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg4e8_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (size_t)(op6)) +#define vsseg4e8_v_u8mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg4e8_v_u8mf4((uint8_t *)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (size_t)(op5)) +#define vsseg4e8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg4e8_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (size_t)(op6)) +#define vsseg4e8_v_u8mf8(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg4e8_v_u8mf8((uint8_t *)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (size_t)(op5)) +#define vsseg4e8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg4e8_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (size_t)(op6)) +#define vsseg5e8_v_i8m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg5e8_v_i8m1((int8_t *)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (size_t)(op6)) +#define vsseg5e8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg5e8_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (size_t)(op7)) +#define vsseg5e8_v_i8mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg5e8_v_i8mf2((int8_t *)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (size_t)(op6)) +#define vsseg5e8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg5e8_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (size_t)(op7)) +#define vsseg5e8_v_i8mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg5e8_v_i8mf4((int8_t *)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (size_t)(op6)) +#define vsseg5e8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg5e8_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (size_t)(op7)) +#define vsseg5e8_v_i8mf8(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg5e8_v_i8mf8((int8_t *)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (size_t)(op6)) +#define vsseg5e8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg5e8_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (size_t)(op7)) +#define vsseg5e8_v_u8m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg5e8_v_u8m1((uint8_t *)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (size_t)(op6)) +#define vsseg5e8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg5e8_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (size_t)(op7)) +#define vsseg5e8_v_u8mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg5e8_v_u8mf2((uint8_t *)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (size_t)(op6)) +#define vsseg5e8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg5e8_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vsseg5e8_v_u8mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg5e8_v_u8mf4((uint8_t *)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (size_t)(op6)) +#define vsseg5e8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg5e8_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vsseg5e8_v_u8mf8(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg5e8_v_u8mf8((uint8_t *)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (size_t)(op6)) +#define vsseg5e8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg5e8_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vsseg6e8_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg6e8_v_i8m1((int8_t *)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (size_t)(op7)) +#define vsseg6e8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg6e8_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (size_t)(op8)) +#define vsseg6e8_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg6e8_v_i8mf2((int8_t *)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (size_t)(op7)) +#define vsseg6e8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg6e8_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (size_t)(op8)) +#define vsseg6e8_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg6e8_v_i8mf4((int8_t *)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (size_t)(op7)) +#define vsseg6e8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg6e8_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (size_t)(op8)) +#define vsseg6e8_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg6e8_v_i8mf8((int8_t *)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (size_t)(op7)) +#define vsseg6e8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg6e8_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (size_t)(op8)) +#define vsseg6e8_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg6e8_v_u8m1((uint8_t *)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (size_t)(op7)) +#define vsseg6e8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg6e8_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (size_t)(op8)) +#define vsseg6e8_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg6e8_v_u8mf2((uint8_t *)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vsseg6e8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg6e8_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (size_t)(op8)) +#define vsseg6e8_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg6e8_v_u8mf4((uint8_t *)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vsseg6e8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg6e8_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (size_t)(op8)) +#define vsseg6e8_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg6e8_v_u8mf8((uint8_t *)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vsseg6e8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg6e8_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (size_t)(op8)) +#define vsseg7e8_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg7e8_v_i8m1((int8_t *)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (size_t)(op8)) +#define vsseg7e8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg7e8_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (size_t)(op9)) +#define vsseg7e8_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg7e8_v_i8mf2((int8_t *)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (size_t)(op8)) +#define vsseg7e8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg7e8_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (size_t)(op9)) +#define vsseg7e8_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg7e8_v_i8mf4((int8_t *)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (size_t)(op8)) +#define vsseg7e8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg7e8_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (size_t)(op9)) +#define vsseg7e8_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg7e8_v_i8mf8((int8_t *)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (size_t)(op8)) +#define vsseg7e8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg7e8_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (size_t)(op9)) +#define vsseg7e8_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg7e8_v_u8m1((uint8_t *)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (size_t)(op8)) +#define vsseg7e8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg7e8_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (size_t)(op9)) +#define vsseg7e8_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg7e8_v_u8mf2((uint8_t *)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (size_t)(op8)) +#define vsseg7e8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg7e8_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vsseg7e8_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg7e8_v_u8mf4((uint8_t *)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (size_t)(op8)) +#define vsseg7e8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg7e8_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vsseg7e8_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg7e8_v_u8mf8((uint8_t *)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (size_t)(op8)) +#define vsseg7e8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg7e8_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vsseg8e8_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg8e8_v_i8m1((int8_t *)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (size_t)(op9)) +#define vsseg8e8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsseg8e8_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (size_t)(op10)) +#define vsseg8e8_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg8e8_v_i8mf2((int8_t *)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (size_t)(op9)) +#define vsseg8e8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsseg8e8_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (size_t)(op10)) +#define vsseg8e8_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg8e8_v_i8mf4((int8_t *)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (size_t)(op9)) +#define vsseg8e8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsseg8e8_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (size_t)(op10)) +#define vsseg8e8_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg8e8_v_i8mf8((int8_t *)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (size_t)(op9)) +#define vsseg8e8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsseg8e8_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (size_t)(op10)) +#define vsseg8e8_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg8e8_v_u8m1((uint8_t *)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (size_t)(op9)) +#define vsseg8e8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsseg8e8_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (size_t)(op10)) +#define vsseg8e8_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg8e8_v_u8mf2((uint8_t *)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vsseg8e8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsseg8e8_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (size_t)(op10)) +#define vsseg8e8_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg8e8_v_u8mf4((uint8_t *)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vsseg8e8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsseg8e8_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (size_t)(op10)) +#define vsseg8e8_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg8e8_v_u8mf8((uint8_t *)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vsseg8e8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsseg8e8_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (size_t)(op10)) +#define vsseg2e16_v_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e16_v_i16m1((int16_t *)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (size_t)(op3)) +#define vsseg2e16_v_i16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e16_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vsseg2e16_v_i16m2(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e16_v_i16m2((int16_t *)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (size_t)(op3)) +#define vsseg2e16_v_i16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e16_v_i16m2_m((vbool8_t)(op0), (int16_t *)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vsseg2e16_v_i16m4(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e16_v_i16m4((int16_t *)(op0), (vint16m4_t)(op1), (vint16m4_t)(op2), (size_t)(op3)) +#define vsseg2e16_v_i16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e16_v_i16m4_m((vbool4_t)(op0), (int16_t *)(op1), (vint16m4_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vsseg2e16_v_i16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e16_v_i16mf2((int16_t *)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (size_t)(op3)) +#define vsseg2e16_v_i16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e16_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vsseg2e16_v_i16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e16_v_i16mf4((int16_t *)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (size_t)(op3)) +#define vsseg2e16_v_i16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e16_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vsseg2e16_v_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e16_v_u16m1((uint16_t *)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (size_t)(op3)) +#define vsseg2e16_v_u16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e16_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vsseg2e16_v_u16m2(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e16_v_u16m2((uint16_t *)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (size_t)(op3)) +#define vsseg2e16_v_u16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e16_v_u16m2_m((vbool8_t)(op0), (uint16_t *)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vsseg2e16_v_u16m4(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e16_v_u16m4((uint16_t *)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (size_t)(op3)) +#define vsseg2e16_v_u16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e16_v_u16m4_m((vbool4_t)(op0), (uint16_t *)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vsseg2e16_v_u16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e16_v_u16mf2((uint16_t *)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (size_t)(op3)) +#define vsseg2e16_v_u16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e16_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vsseg2e16_v_u16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e16_v_u16mf4((uint16_t *)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (size_t)(op3)) +#define vsseg2e16_v_u16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e16_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vsseg3e16_v_i16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg3e16_v_i16m1((int16_t *)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vsseg3e16_v_i16m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg3e16_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (size_t)(op5)) +#define vsseg3e16_v_i16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg3e16_v_i16m2((int16_t *)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vsseg3e16_v_i16m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg3e16_v_i16m2_m((vbool8_t)(op0), (int16_t *)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (size_t)(op5)) +#define vsseg3e16_v_i16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg3e16_v_i16mf2((int16_t *)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vsseg3e16_v_i16mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg3e16_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (size_t)(op5)) +#define vsseg3e16_v_i16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg3e16_v_i16mf4((int16_t *)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vsseg3e16_v_i16mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg3e16_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (size_t)(op5)) +#define vsseg3e16_v_u16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg3e16_v_u16m1((uint16_t *)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vsseg3e16_v_u16m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg3e16_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (size_t)(op5)) +#define vsseg3e16_v_u16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg3e16_v_u16m2((uint16_t *)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vsseg3e16_v_u16m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg3e16_v_u16m2_m((vbool8_t)(op0), (uint16_t *)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (size_t)(op5)) +#define vsseg3e16_v_u16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg3e16_v_u16mf2((uint16_t *)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vsseg3e16_v_u16mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg3e16_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (size_t)(op5)) +#define vsseg3e16_v_u16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg3e16_v_u16mf4((uint16_t *)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vsseg3e16_v_u16mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg3e16_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (size_t)(op5)) +#define vsseg4e16_v_i16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg4e16_v_i16m1((int16_t *)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (size_t)(op5)) +#define vsseg4e16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg4e16_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (size_t)(op6)) +#define vsseg4e16_v_i16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg4e16_v_i16m2((int16_t *)(op0), (vint16m2_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (size_t)(op5)) +#define vsseg4e16_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg4e16_v_i16m2_m((vbool8_t)(op0), (int16_t *)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (vint16m2_t)(op5), (size_t)(op6)) +#define vsseg4e16_v_i16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg4e16_v_i16mf2((int16_t *)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (size_t)(op5)) +#define vsseg4e16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg4e16_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (size_t)(op6)) +#define vsseg4e16_v_i16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg4e16_v_i16mf4((int16_t *)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (size_t)(op5)) +#define vsseg4e16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg4e16_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (size_t)(op6)) +#define vsseg4e16_v_u16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg4e16_v_u16m1((uint16_t *)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (size_t)(op5)) +#define vsseg4e16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg4e16_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (size_t)(op6)) +#define vsseg4e16_v_u16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg4e16_v_u16m2((uint16_t *)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (size_t)(op5)) +#define vsseg4e16_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg4e16_v_u16m2_m((vbool8_t)(op0), (uint16_t *)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (vuint16m2_t)(op5), (size_t)(op6)) +#define vsseg4e16_v_u16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg4e16_v_u16mf2((uint16_t *)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (size_t)(op5)) +#define vsseg4e16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg4e16_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (size_t)(op6)) +#define vsseg4e16_v_u16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg4e16_v_u16mf4((uint16_t *)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (size_t)(op5)) +#define vsseg4e16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg4e16_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (size_t)(op6)) +#define vsseg5e16_v_i16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg5e16_v_i16m1((int16_t *)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (size_t)(op6)) +#define vsseg5e16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg5e16_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (size_t)(op7)) +#define vsseg5e16_v_i16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg5e16_v_i16mf2((int16_t *)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (size_t)(op6)) +#define vsseg5e16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg5e16_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (size_t)(op7)) +#define vsseg5e16_v_i16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg5e16_v_i16mf4((int16_t *)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (size_t)(op6)) +#define vsseg5e16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg5e16_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (size_t)(op7)) +#define vsseg5e16_v_u16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg5e16_v_u16m1((uint16_t *)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (size_t)(op6)) +#define vsseg5e16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg5e16_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vsseg5e16_v_u16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg5e16_v_u16mf2((uint16_t *)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (size_t)(op6)) +#define vsseg5e16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg5e16_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vsseg5e16_v_u16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg5e16_v_u16mf4((uint16_t *)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (size_t)(op6)) +#define vsseg5e16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg5e16_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vsseg6e16_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg6e16_v_i16m1((int16_t *)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (size_t)(op7)) +#define vsseg6e16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg6e16_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (size_t)(op8)) +#define vsseg6e16_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg6e16_v_i16mf2((int16_t *)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (size_t)(op7)) +#define vsseg6e16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg6e16_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (size_t)(op8)) +#define vsseg6e16_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg6e16_v_i16mf4((int16_t *)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (size_t)(op7)) +#define vsseg6e16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg6e16_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (size_t)(op8)) +#define vsseg6e16_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg6e16_v_u16m1((uint16_t *)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vsseg6e16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg6e16_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (size_t)(op8)) +#define vsseg6e16_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg6e16_v_u16mf2((uint16_t *)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vsseg6e16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg6e16_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (size_t)(op8)) +#define vsseg6e16_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg6e16_v_u16mf4((uint16_t *)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vsseg6e16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg6e16_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (size_t)(op8)) +#define vsseg7e16_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg7e16_v_i16m1((int16_t *)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (size_t)(op8)) +#define vsseg7e16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg7e16_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (size_t)(op9)) +#define vsseg7e16_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg7e16_v_i16mf2((int16_t *)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (size_t)(op8)) +#define vsseg7e16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg7e16_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (size_t)(op9)) +#define vsseg7e16_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg7e16_v_i16mf4((int16_t *)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (size_t)(op8)) +#define vsseg7e16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg7e16_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (size_t)(op9)) +#define vsseg7e16_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg7e16_v_u16m1((uint16_t *)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (size_t)(op8)) +#define vsseg7e16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg7e16_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vsseg7e16_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg7e16_v_u16mf2((uint16_t *)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (size_t)(op8)) +#define vsseg7e16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg7e16_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vsseg7e16_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg7e16_v_u16mf4((uint16_t *)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (size_t)(op8)) +#define vsseg7e16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg7e16_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vsseg8e16_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg8e16_v_i16m1((int16_t *)(op0), (vint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (size_t)(op9)) +#define vsseg8e16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsseg8e16_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (size_t)(op10)) +#define vsseg8e16_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg8e16_v_i16mf2((int16_t *)(op0), (vint16mf2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (size_t)(op9)) +#define vsseg8e16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsseg8e16_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (size_t)(op10)) +#define vsseg8e16_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg8e16_v_i16mf4((int16_t *)(op0), (vint16mf4_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (size_t)(op9)) +#define vsseg8e16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsseg8e16_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (size_t)(op10)) +#define vsseg8e16_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg8e16_v_u16m1((uint16_t *)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vsseg8e16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsseg8e16_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (size_t)(op10)) +#define vsseg8e16_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg8e16_v_u16mf2((uint16_t *)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vsseg8e16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsseg8e16_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (size_t)(op10)) +#define vsseg8e16_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg8e16_v_u16mf4((uint16_t *)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vsseg8e16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsseg8e16_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (size_t)(op10)) +#define vsseg2e32_v_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e32_v_i32m1((int32_t *)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (size_t)(op3)) +#define vsseg2e32_v_i32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e32_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vsseg2e32_v_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e32_v_i32m2((int32_t *)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (size_t)(op3)) +#define vsseg2e32_v_i32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e32_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vsseg2e32_v_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e32_v_i32m4((int32_t *)(op0), (vint32m4_t)(op1), (vint32m4_t)(op2), (size_t)(op3)) +#define vsseg2e32_v_i32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e32_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vsseg2e32_v_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e32_v_i32mf2((int32_t *)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (size_t)(op3)) +#define vsseg2e32_v_i32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e32_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vsseg2e32_v_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e32_v_u32m1((uint32_t *)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (size_t)(op3)) +#define vsseg2e32_v_u32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e32_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vsseg2e32_v_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e32_v_u32m2((uint32_t *)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (size_t)(op3)) +#define vsseg2e32_v_u32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e32_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vsseg2e32_v_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e32_v_u32m4((uint32_t *)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (size_t)(op3)) +#define vsseg2e32_v_u32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e32_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vsseg2e32_v_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e32_v_u32mf2((uint32_t *)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (size_t)(op3)) +#define vsseg2e32_v_u32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e32_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vsseg3e32_v_i32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg3e32_v_i32m1((int32_t *)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vsseg3e32_v_i32m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg3e32_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (size_t)(op5)) +#define vsseg3e32_v_i32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg3e32_v_i32m2((int32_t *)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vsseg3e32_v_i32m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg3e32_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (size_t)(op5)) +#define vsseg3e32_v_i32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg3e32_v_i32mf2((int32_t *)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vsseg3e32_v_i32mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg3e32_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (size_t)(op5)) +#define vsseg3e32_v_u32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg3e32_v_u32m1((uint32_t *)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vsseg3e32_v_u32m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg3e32_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (size_t)(op5)) +#define vsseg3e32_v_u32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg3e32_v_u32m2((uint32_t *)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vsseg3e32_v_u32m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg3e32_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (size_t)(op5)) +#define vsseg3e32_v_u32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg3e32_v_u32mf2((uint32_t *)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vsseg3e32_v_u32mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg3e32_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (size_t)(op5)) +#define vsseg4e32_v_i32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg4e32_v_i32m1((int32_t *)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (size_t)(op5)) +#define vsseg4e32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg4e32_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (size_t)(op6)) +#define vsseg4e32_v_i32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg4e32_v_i32m2((int32_t *)(op0), (vint32m2_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (size_t)(op5)) +#define vsseg4e32_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg4e32_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (vint32m2_t)(op5), (size_t)(op6)) +#define vsseg4e32_v_i32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg4e32_v_i32mf2((int32_t *)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (size_t)(op5)) +#define vsseg4e32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg4e32_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (size_t)(op6)) +#define vsseg4e32_v_u32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg4e32_v_u32m1((uint32_t *)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (size_t)(op5)) +#define vsseg4e32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg4e32_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (size_t)(op6)) +#define vsseg4e32_v_u32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg4e32_v_u32m2((uint32_t *)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (size_t)(op5)) +#define vsseg4e32_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg4e32_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (vuint32m2_t)(op5), (size_t)(op6)) +#define vsseg4e32_v_u32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg4e32_v_u32mf2((uint32_t *)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (size_t)(op5)) +#define vsseg4e32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg4e32_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (size_t)(op6)) +#define vsseg5e32_v_i32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg5e32_v_i32m1((int32_t *)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (size_t)(op6)) +#define vsseg5e32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg5e32_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (size_t)(op7)) +#define vsseg5e32_v_i32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg5e32_v_i32mf2((int32_t *)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (size_t)(op6)) +#define vsseg5e32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg5e32_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (size_t)(op7)) +#define vsseg5e32_v_u32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg5e32_v_u32m1((uint32_t *)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (size_t)(op6)) +#define vsseg5e32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg5e32_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vsseg5e32_v_u32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg5e32_v_u32mf2((uint32_t *)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (size_t)(op6)) +#define vsseg5e32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg5e32_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vsseg6e32_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg6e32_v_i32m1((int32_t *)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (size_t)(op7)) +#define vsseg6e32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg6e32_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (size_t)(op8)) +#define vsseg6e32_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg6e32_v_i32mf2((int32_t *)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (size_t)(op7)) +#define vsseg6e32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg6e32_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (size_t)(op8)) +#define vsseg6e32_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg6e32_v_u32m1((uint32_t *)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vsseg6e32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg6e32_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (size_t)(op8)) +#define vsseg6e32_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg6e32_v_u32mf2((uint32_t *)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vsseg6e32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg6e32_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (size_t)(op8)) +#define vsseg7e32_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg7e32_v_i32m1((int32_t *)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (size_t)(op8)) +#define vsseg7e32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg7e32_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (size_t)(op9)) +#define vsseg7e32_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg7e32_v_i32mf2((int32_t *)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (size_t)(op8)) +#define vsseg7e32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg7e32_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (size_t)(op9)) +#define vsseg7e32_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg7e32_v_u32m1((uint32_t *)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (size_t)(op8)) +#define vsseg7e32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg7e32_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vsseg7e32_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg7e32_v_u32mf2((uint32_t *)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (size_t)(op8)) +#define vsseg7e32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg7e32_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vsseg8e32_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg8e32_v_i32m1((int32_t *)(op0), (vint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (size_t)(op9)) +#define vsseg8e32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsseg8e32_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (size_t)(op10)) +#define vsseg8e32_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg8e32_v_i32mf2((int32_t *)(op0), (vint32mf2_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (size_t)(op9)) +#define vsseg8e32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsseg8e32_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (size_t)(op10)) +#define vsseg8e32_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg8e32_v_u32m1((uint32_t *)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vsseg8e32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsseg8e32_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (size_t)(op10)) +#define vsseg8e32_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg8e32_v_u32mf2((uint32_t *)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vsseg8e32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsseg8e32_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (size_t)(op10)) +#define vsseg2e64_v_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e64_v_i64m1((int64_t *)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (size_t)(op3)) +#define vsseg2e64_v_i64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e64_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vsseg2e64_v_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e64_v_i64m2((int64_t *)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (size_t)(op3)) +#define vsseg2e64_v_i64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e64_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vsseg2e64_v_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e64_v_i64m4((int64_t *)(op0), (vint64m4_t)(op1), (vint64m4_t)(op2), (size_t)(op3)) +#define vsseg2e64_v_i64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e64_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vint64m4_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vsseg2e64_v_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e64_v_u64m1((uint64_t *)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (size_t)(op3)) +#define vsseg2e64_v_u64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e64_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vsseg2e64_v_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e64_v_u64m2((uint64_t *)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (size_t)(op3)) +#define vsseg2e64_v_u64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e64_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vsseg2e64_v_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e64_v_u64m4((uint64_t *)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (size_t)(op3)) +#define vsseg2e64_v_u64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e64_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vsseg3e64_v_i64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg3e64_v_i64m1((int64_t *)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vsseg3e64_v_i64m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg3e64_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (size_t)(op5)) +#define vsseg3e64_v_i64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg3e64_v_i64m2((int64_t *)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vsseg3e64_v_i64m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg3e64_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (size_t)(op5)) +#define vsseg3e64_v_u64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg3e64_v_u64m1((uint64_t *)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vsseg3e64_v_u64m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg3e64_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (size_t)(op5)) +#define vsseg3e64_v_u64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg3e64_v_u64m2((uint64_t *)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vsseg3e64_v_u64m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg3e64_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (size_t)(op5)) +#define vsseg4e64_v_i64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg4e64_v_i64m1((int64_t *)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (size_t)(op5)) +#define vsseg4e64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg4e64_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (size_t)(op6)) +#define vsseg4e64_v_i64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg4e64_v_i64m2((int64_t *)(op0), (vint64m2_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (size_t)(op5)) +#define vsseg4e64_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg4e64_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (vint64m2_t)(op5), (size_t)(op6)) +#define vsseg4e64_v_u64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg4e64_v_u64m1((uint64_t *)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (size_t)(op5)) +#define vsseg4e64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg4e64_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (size_t)(op6)) +#define vsseg4e64_v_u64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg4e64_v_u64m2((uint64_t *)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (size_t)(op5)) +#define vsseg4e64_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg4e64_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (vuint64m2_t)(op5), (size_t)(op6)) +#define vsseg5e64_v_i64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg5e64_v_i64m1((int64_t *)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (size_t)(op6)) +#define vsseg5e64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg5e64_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (size_t)(op7)) +#define vsseg5e64_v_u64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg5e64_v_u64m1((uint64_t *)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (size_t)(op6)) +#define vsseg5e64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg5e64_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vsseg6e64_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg6e64_v_i64m1((int64_t *)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (size_t)(op7)) +#define vsseg6e64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg6e64_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (size_t)(op8)) +#define vsseg6e64_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg6e64_v_u64m1((uint64_t *)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vsseg6e64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg6e64_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (size_t)(op8)) +#define vsseg7e64_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg7e64_v_i64m1((int64_t *)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (size_t)(op8)) +#define vsseg7e64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg7e64_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (size_t)(op9)) +#define vsseg7e64_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg7e64_v_u64m1((uint64_t *)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (size_t)(op8)) +#define vsseg7e64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg7e64_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vsseg8e64_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg8e64_v_i64m1((int64_t *)(op0), (vint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (size_t)(op9)) +#define vsseg8e64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsseg8e64_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (size_t)(op10)) +#define vsseg8e64_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg8e64_v_u64m1((uint64_t *)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vsseg8e64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsseg8e64_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (size_t)(op10)) +#define vssseg3e8_v_i8m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg3e8_v_i8m1((int8_t *)(op0), (ptrdiff_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (size_t)(op5)) +#define vssseg3e8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg3e8_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (ptrdiff_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (size_t)(op6)) +#define vssseg3e8_v_i8m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg3e8_v_i8m2((int8_t *)(op0), (ptrdiff_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (size_t)(op5)) +#define vssseg3e8_v_i8m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg3e8_v_i8m2_m((vbool4_t)(op0), (int8_t *)(op1), (ptrdiff_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (vint8m2_t)(op5), (size_t)(op6)) +#define vssseg3e8_v_i8mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg3e8_v_i8mf2((int8_t *)(op0), (ptrdiff_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (size_t)(op5)) +#define vssseg3e8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg3e8_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (ptrdiff_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (size_t)(op6)) +#define vssseg3e8_v_i8mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg3e8_v_i8mf4((int8_t *)(op0), (ptrdiff_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (size_t)(op5)) +#define vssseg3e8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg3e8_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (ptrdiff_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (size_t)(op6)) +#define vssseg3e8_v_i8mf8(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg3e8_v_i8mf8((int8_t *)(op0), (ptrdiff_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (size_t)(op5)) +#define vssseg3e8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg3e8_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (ptrdiff_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (size_t)(op6)) +#define vssseg3e8_v_u8m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg3e8_v_u8m1((uint8_t *)(op0), (ptrdiff_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (size_t)(op5)) +#define vssseg3e8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg3e8_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (ptrdiff_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (size_t)(op6)) +#define vssseg3e8_v_u8m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg3e8_v_u8m2((uint8_t *)(op0), (ptrdiff_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (size_t)(op5)) +#define vssseg3e8_v_u8m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg3e8_v_u8m2_m((vbool4_t)(op0), (uint8_t *)(op1), (ptrdiff_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (vuint8m2_t)(op5), (size_t)(op6)) +#define vssseg3e8_v_u8mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg3e8_v_u8mf2((uint8_t *)(op0), (ptrdiff_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (size_t)(op5)) +#define vssseg3e8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg3e8_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (ptrdiff_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (size_t)(op6)) +#define vssseg3e8_v_u8mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg3e8_v_u8mf4((uint8_t *)(op0), (ptrdiff_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (size_t)(op5)) +#define vssseg3e8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg3e8_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (ptrdiff_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (size_t)(op6)) +#define vssseg3e8_v_u8mf8(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg3e8_v_u8mf8((uint8_t *)(op0), (ptrdiff_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (size_t)(op5)) +#define vssseg3e8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg3e8_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (ptrdiff_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (size_t)(op6)) +#define vssseg4e8_v_i8m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg4e8_v_i8m1((int8_t *)(op0), (ptrdiff_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (size_t)(op6)) +#define vssseg4e8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg4e8_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (ptrdiff_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (size_t)(op7)) +#define vssseg4e8_v_i8m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg4e8_v_i8m2((int8_t *)(op0), (ptrdiff_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (vint8m2_t)(op5), (size_t)(op6)) +#define vssseg4e8_v_i8m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg4e8_v_i8m2_m((vbool4_t)(op0), (int8_t *)(op1), (ptrdiff_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (vint8m2_t)(op5), (vint8m2_t)(op6), (size_t)(op7)) +#define vssseg4e8_v_i8mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg4e8_v_i8mf2((int8_t *)(op0), (ptrdiff_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (size_t)(op6)) +#define vssseg4e8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg4e8_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (ptrdiff_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (size_t)(op7)) +#define vssseg4e8_v_i8mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg4e8_v_i8mf4((int8_t *)(op0), (ptrdiff_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (size_t)(op6)) +#define vssseg4e8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg4e8_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (ptrdiff_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (size_t)(op7)) +#define vssseg4e8_v_i8mf8(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg4e8_v_i8mf8((int8_t *)(op0), (ptrdiff_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (size_t)(op6)) +#define vssseg4e8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg4e8_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (ptrdiff_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (size_t)(op7)) +#define vssseg4e8_v_u8m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg4e8_v_u8m1((uint8_t *)(op0), (ptrdiff_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (size_t)(op6)) +#define vssseg4e8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg4e8_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (ptrdiff_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (size_t)(op7)) +#define vssseg4e8_v_u8m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg4e8_v_u8m2((uint8_t *)(op0), (ptrdiff_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (vuint8m2_t)(op5), (size_t)(op6)) +#define vssseg4e8_v_u8m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg4e8_v_u8m2_m((vbool4_t)(op0), (uint8_t *)(op1), (ptrdiff_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (vuint8m2_t)(op5), (vuint8m2_t)(op6), (size_t)(op7)) +#define vssseg4e8_v_u8mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg4e8_v_u8mf2((uint8_t *)(op0), (ptrdiff_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (size_t)(op6)) +#define vssseg4e8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg4e8_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (ptrdiff_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vssseg4e8_v_u8mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg4e8_v_u8mf4((uint8_t *)(op0), (ptrdiff_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (size_t)(op6)) +#define vssseg4e8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg4e8_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (ptrdiff_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vssseg4e8_v_u8mf8(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg4e8_v_u8mf8((uint8_t *)(op0), (ptrdiff_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (size_t)(op6)) +#define vssseg4e8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg4e8_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (ptrdiff_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vssseg5e8_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg5e8_v_i8m1((int8_t *)(op0), (ptrdiff_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (size_t)(op7)) +#define vssseg5e8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg5e8_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (ptrdiff_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (size_t)(op8)) +#define vssseg5e8_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg5e8_v_i8mf2((int8_t *)(op0), (ptrdiff_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (size_t)(op7)) +#define vssseg5e8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg5e8_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (ptrdiff_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (size_t)(op8)) +#define vssseg5e8_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg5e8_v_i8mf4((int8_t *)(op0), (ptrdiff_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (size_t)(op7)) +#define vssseg5e8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg5e8_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (ptrdiff_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (size_t)(op8)) +#define vssseg5e8_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg5e8_v_i8mf8((int8_t *)(op0), (ptrdiff_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (size_t)(op7)) +#define vssseg5e8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg5e8_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (ptrdiff_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (size_t)(op8)) +#define vssseg5e8_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg5e8_v_u8m1((uint8_t *)(op0), (ptrdiff_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (size_t)(op7)) +#define vssseg5e8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg5e8_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (ptrdiff_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (size_t)(op8)) +#define vssseg5e8_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg5e8_v_u8mf2((uint8_t *)(op0), (ptrdiff_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vssseg5e8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg5e8_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (ptrdiff_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (size_t)(op8)) +#define vssseg5e8_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg5e8_v_u8mf4((uint8_t *)(op0), (ptrdiff_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vssseg5e8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg5e8_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (ptrdiff_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (size_t)(op8)) +#define vssseg5e8_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg5e8_v_u8mf8((uint8_t *)(op0), (ptrdiff_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vssseg5e8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg5e8_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (ptrdiff_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (size_t)(op8)) +#define vssseg6e8_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg6e8_v_i8m1((int8_t *)(op0), (ptrdiff_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (size_t)(op8)) +#define vssseg6e8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg6e8_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (ptrdiff_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (size_t)(op9)) +#define vssseg6e8_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg6e8_v_i8mf2((int8_t *)(op0), (ptrdiff_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (size_t)(op8)) +#define vssseg6e8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg6e8_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (ptrdiff_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (size_t)(op9)) +#define vssseg6e8_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg6e8_v_i8mf4((int8_t *)(op0), (ptrdiff_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (size_t)(op8)) +#define vssseg6e8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg6e8_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (ptrdiff_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (size_t)(op9)) +#define vssseg6e8_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg6e8_v_i8mf8((int8_t *)(op0), (ptrdiff_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (size_t)(op8)) +#define vssseg6e8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg6e8_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (ptrdiff_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (size_t)(op9)) +#define vssseg6e8_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg6e8_v_u8m1((uint8_t *)(op0), (ptrdiff_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (size_t)(op8)) +#define vssseg6e8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg6e8_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (ptrdiff_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (size_t)(op9)) +#define vssseg6e8_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg6e8_v_u8mf2((uint8_t *)(op0), (ptrdiff_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (size_t)(op8)) +#define vssseg6e8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg6e8_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (ptrdiff_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vssseg6e8_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg6e8_v_u8mf4((uint8_t *)(op0), (ptrdiff_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (size_t)(op8)) +#define vssseg6e8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg6e8_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (ptrdiff_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vssseg6e8_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg6e8_v_u8mf8((uint8_t *)(op0), (ptrdiff_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (size_t)(op8)) +#define vssseg6e8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg6e8_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (ptrdiff_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vssseg7e8_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg7e8_v_i8m1((int8_t *)(op0), (ptrdiff_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (size_t)(op9)) +#define vssseg7e8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg7e8_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (ptrdiff_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (size_t)(op10)) +#define vssseg7e8_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg7e8_v_i8mf2((int8_t *)(op0), (ptrdiff_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (size_t)(op9)) +#define vssseg7e8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg7e8_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (ptrdiff_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (size_t)(op10)) +#define vssseg7e8_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg7e8_v_i8mf4((int8_t *)(op0), (ptrdiff_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (size_t)(op9)) +#define vssseg7e8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg7e8_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (ptrdiff_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (size_t)(op10)) +#define vssseg7e8_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg7e8_v_i8mf8((int8_t *)(op0), (ptrdiff_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (size_t)(op9)) +#define vssseg7e8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg7e8_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (ptrdiff_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (size_t)(op10)) +#define vssseg7e8_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg7e8_v_u8m1((uint8_t *)(op0), (ptrdiff_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (size_t)(op9)) +#define vssseg7e8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg7e8_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (ptrdiff_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (size_t)(op10)) +#define vssseg7e8_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg7e8_v_u8mf2((uint8_t *)(op0), (ptrdiff_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vssseg7e8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg7e8_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (ptrdiff_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (size_t)(op10)) +#define vssseg7e8_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg7e8_v_u8mf4((uint8_t *)(op0), (ptrdiff_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vssseg7e8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg7e8_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (ptrdiff_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (size_t)(op10)) +#define vssseg7e8_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg7e8_v_u8mf8((uint8_t *)(op0), (ptrdiff_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vssseg7e8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg7e8_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (ptrdiff_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (size_t)(op10)) +#define vssseg8e8_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg8e8_v_i8m1((int8_t *)(op0), (ptrdiff_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (size_t)(op10)) +#define vssseg8e8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vssseg8e8_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (ptrdiff_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (size_t)(op11)) +#define vssseg8e8_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg8e8_v_i8mf2((int8_t *)(op0), (ptrdiff_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (size_t)(op10)) +#define vssseg8e8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vssseg8e8_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (ptrdiff_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (size_t)(op11)) +#define vssseg8e8_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg8e8_v_i8mf4((int8_t *)(op0), (ptrdiff_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (size_t)(op10)) +#define vssseg8e8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vssseg8e8_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (ptrdiff_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (size_t)(op11)) +#define vssseg8e8_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg8e8_v_i8mf8((int8_t *)(op0), (ptrdiff_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (size_t)(op10)) +#define vssseg8e8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vssseg8e8_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (ptrdiff_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (size_t)(op11)) +#define vssseg8e8_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg8e8_v_u8m1((uint8_t *)(op0), (ptrdiff_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (size_t)(op10)) +#define vssseg8e8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vssseg8e8_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (ptrdiff_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (size_t)(op11)) +#define vssseg8e8_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg8e8_v_u8mf2((uint8_t *)(op0), (ptrdiff_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (size_t)(op10)) +#define vssseg8e8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vssseg8e8_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (ptrdiff_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (size_t)(op11)) +#define vssseg8e8_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg8e8_v_u8mf4((uint8_t *)(op0), (ptrdiff_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (size_t)(op10)) +#define vssseg8e8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vssseg8e8_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (ptrdiff_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (size_t)(op11)) +#define vssseg8e8_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg8e8_v_u8mf8((uint8_t *)(op0), (ptrdiff_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (size_t)(op10)) +#define vssseg8e8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vssseg8e8_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (ptrdiff_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (size_t)(op11)) +#define vssseg2e16_v_i16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e16_v_i16m1((int16_t *)(op0), (ptrdiff_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vssseg2e16_v_i16m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e16_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (ptrdiff_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (size_t)(op5)) +#define vssseg2e16_v_i16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e16_v_i16m2((int16_t *)(op0), (ptrdiff_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vssseg2e16_v_i16m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e16_v_i16m2_m((vbool8_t)(op0), (int16_t *)(op1), (ptrdiff_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (size_t)(op5)) +#define vssseg2e16_v_i16m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e16_v_i16m4((int16_t *)(op0), (ptrdiff_t)(op1), (vint16m4_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vssseg2e16_v_i16m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e16_v_i16m4_m((vbool4_t)(op0), (int16_t *)(op1), (ptrdiff_t)(op2), (vint16m4_t)(op3), (vint16m4_t)(op4), (size_t)(op5)) +#define vssseg2e16_v_i16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e16_v_i16mf2((int16_t *)(op0), (ptrdiff_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vssseg2e16_v_i16mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e16_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (ptrdiff_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (size_t)(op5)) +#define vssseg2e16_v_i16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e16_v_i16mf4((int16_t *)(op0), (ptrdiff_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vssseg2e16_v_i16mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e16_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (ptrdiff_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (size_t)(op5)) +#define vssseg2e16_v_u16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e16_v_u16m1((uint16_t *)(op0), (ptrdiff_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vssseg2e16_v_u16m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e16_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (ptrdiff_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (size_t)(op5)) +#define vssseg2e16_v_u16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e16_v_u16m2((uint16_t *)(op0), (ptrdiff_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vssseg2e16_v_u16m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e16_v_u16m2_m((vbool8_t)(op0), (uint16_t *)(op1), (ptrdiff_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (size_t)(op5)) +#define vssseg2e16_v_u16m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e16_v_u16m4((uint16_t *)(op0), (ptrdiff_t)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vssseg2e16_v_u16m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e16_v_u16m4_m((vbool4_t)(op0), (uint16_t *)(op1), (ptrdiff_t)(op2), (vuint16m4_t)(op3), (vuint16m4_t)(op4), (size_t)(op5)) +#define vssseg2e16_v_u16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e16_v_u16mf2((uint16_t *)(op0), (ptrdiff_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vssseg2e16_v_u16mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e16_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (ptrdiff_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (size_t)(op5)) +#define vssseg2e16_v_u16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e16_v_u16mf4((uint16_t *)(op0), (ptrdiff_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vssseg2e16_v_u16mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e16_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (ptrdiff_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (size_t)(op5)) +#define vssseg3e16_v_i16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg3e16_v_i16m1((int16_t *)(op0), (ptrdiff_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (size_t)(op5)) +#define vssseg3e16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg3e16_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (ptrdiff_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (size_t)(op6)) +#define vssseg3e16_v_i16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg3e16_v_i16m2((int16_t *)(op0), (ptrdiff_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (size_t)(op5)) +#define vssseg3e16_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg3e16_v_i16m2_m((vbool8_t)(op0), (int16_t *)(op1), (ptrdiff_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (vint16m2_t)(op5), (size_t)(op6)) +#define vssseg3e16_v_i16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg3e16_v_i16mf2((int16_t *)(op0), (ptrdiff_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (size_t)(op5)) +#define vssseg3e16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg3e16_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (ptrdiff_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (size_t)(op6)) +#define vssseg3e16_v_i16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg3e16_v_i16mf4((int16_t *)(op0), (ptrdiff_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (size_t)(op5)) +#define vssseg3e16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg3e16_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (ptrdiff_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (size_t)(op6)) +#define vssseg3e16_v_u16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg3e16_v_u16m1((uint16_t *)(op0), (ptrdiff_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (size_t)(op5)) +#define vssseg3e16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg3e16_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (ptrdiff_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (size_t)(op6)) +#define vssseg3e16_v_u16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg3e16_v_u16m2((uint16_t *)(op0), (ptrdiff_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (size_t)(op5)) +#define vssseg3e16_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg3e16_v_u16m2_m((vbool8_t)(op0), (uint16_t *)(op1), (ptrdiff_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (vuint16m2_t)(op5), (size_t)(op6)) +#define vssseg3e16_v_u16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg3e16_v_u16mf2((uint16_t *)(op0), (ptrdiff_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (size_t)(op5)) +#define vssseg3e16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg3e16_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (ptrdiff_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (size_t)(op6)) +#define vssseg3e16_v_u16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg3e16_v_u16mf4((uint16_t *)(op0), (ptrdiff_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (size_t)(op5)) +#define vssseg3e16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg3e16_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (ptrdiff_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (size_t)(op6)) +#define vssseg4e16_v_i16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg4e16_v_i16m1((int16_t *)(op0), (ptrdiff_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (size_t)(op6)) +#define vssseg4e16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg4e16_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (ptrdiff_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (size_t)(op7)) +#define vssseg4e16_v_i16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg4e16_v_i16m2((int16_t *)(op0), (ptrdiff_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (vint16m2_t)(op5), (size_t)(op6)) +#define vssseg4e16_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg4e16_v_i16m2_m((vbool8_t)(op0), (int16_t *)(op1), (ptrdiff_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (vint16m2_t)(op5), (vint16m2_t)(op6), (size_t)(op7)) +#define vssseg4e16_v_i16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg4e16_v_i16mf2((int16_t *)(op0), (ptrdiff_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (size_t)(op6)) +#define vssseg4e16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg4e16_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (ptrdiff_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (size_t)(op7)) +#define vssseg4e16_v_i16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg4e16_v_i16mf4((int16_t *)(op0), (ptrdiff_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (size_t)(op6)) +#define vssseg4e16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg4e16_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (ptrdiff_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (size_t)(op7)) +#define vssseg4e16_v_u16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg4e16_v_u16m1((uint16_t *)(op0), (ptrdiff_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (size_t)(op6)) +#define vssseg4e16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg4e16_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (ptrdiff_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vssseg4e16_v_u16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg4e16_v_u16m2((uint16_t *)(op0), (ptrdiff_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (vuint16m2_t)(op5), (size_t)(op6)) +#define vssseg4e16_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg4e16_v_u16m2_m((vbool8_t)(op0), (uint16_t *)(op1), (ptrdiff_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (vuint16m2_t)(op5), (vuint16m2_t)(op6), (size_t)(op7)) +#define vssseg4e16_v_u16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg4e16_v_u16mf2((uint16_t *)(op0), (ptrdiff_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (size_t)(op6)) +#define vssseg4e16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg4e16_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (ptrdiff_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vssseg4e16_v_u16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg4e16_v_u16mf4((uint16_t *)(op0), (ptrdiff_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (size_t)(op6)) +#define vssseg4e16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg4e16_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (ptrdiff_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vssseg5e16_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg5e16_v_i16m1((int16_t *)(op0), (ptrdiff_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (size_t)(op7)) +#define vssseg5e16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg5e16_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (ptrdiff_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (size_t)(op8)) +#define vssseg5e16_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg5e16_v_i16mf2((int16_t *)(op0), (ptrdiff_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (size_t)(op7)) +#define vssseg5e16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg5e16_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (ptrdiff_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (size_t)(op8)) +#define vssseg5e16_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg5e16_v_i16mf4((int16_t *)(op0), (ptrdiff_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (size_t)(op7)) +#define vssseg5e16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg5e16_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (ptrdiff_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (size_t)(op8)) +#define vssseg5e16_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg5e16_v_u16m1((uint16_t *)(op0), (ptrdiff_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vssseg5e16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg5e16_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (ptrdiff_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (size_t)(op8)) +#define vssseg5e16_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg5e16_v_u16mf2((uint16_t *)(op0), (ptrdiff_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vssseg5e16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg5e16_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (ptrdiff_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (size_t)(op8)) +#define vssseg5e16_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg5e16_v_u16mf4((uint16_t *)(op0), (ptrdiff_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vssseg5e16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg5e16_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (ptrdiff_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (size_t)(op8)) +#define vssseg6e16_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg6e16_v_i16m1((int16_t *)(op0), (ptrdiff_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (size_t)(op8)) +#define vssseg6e16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg6e16_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (ptrdiff_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (size_t)(op9)) +#define vssseg6e16_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg6e16_v_i16mf2((int16_t *)(op0), (ptrdiff_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (size_t)(op8)) +#define vssseg6e16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg6e16_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (ptrdiff_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (size_t)(op9)) +#define vssseg6e16_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg6e16_v_i16mf4((int16_t *)(op0), (ptrdiff_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (size_t)(op8)) +#define vssseg6e16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg6e16_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (ptrdiff_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (size_t)(op9)) +#define vssseg6e16_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg6e16_v_u16m1((uint16_t *)(op0), (ptrdiff_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (size_t)(op8)) +#define vssseg6e16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg6e16_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (ptrdiff_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vssseg6e16_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg6e16_v_u16mf2((uint16_t *)(op0), (ptrdiff_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (size_t)(op8)) +#define vssseg6e16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg6e16_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (ptrdiff_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vssseg6e16_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg6e16_v_u16mf4((uint16_t *)(op0), (ptrdiff_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (size_t)(op8)) +#define vssseg6e16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg6e16_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (ptrdiff_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vssseg7e16_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg7e16_v_i16m1((int16_t *)(op0), (ptrdiff_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (size_t)(op9)) +#define vssseg7e16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg7e16_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (ptrdiff_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (size_t)(op10)) +#define vssseg7e16_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg7e16_v_i16mf2((int16_t *)(op0), (ptrdiff_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (size_t)(op9)) +#define vssseg7e16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg7e16_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (ptrdiff_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (size_t)(op10)) +#define vssseg7e16_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg7e16_v_i16mf4((int16_t *)(op0), (ptrdiff_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (size_t)(op9)) +#define vssseg7e16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg7e16_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (ptrdiff_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (size_t)(op10)) +#define vssseg7e16_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg7e16_v_u16m1((uint16_t *)(op0), (ptrdiff_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vssseg7e16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg7e16_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (ptrdiff_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (size_t)(op10)) +#define vssseg7e16_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg7e16_v_u16mf2((uint16_t *)(op0), (ptrdiff_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vssseg7e16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg7e16_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (ptrdiff_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (size_t)(op10)) +#define vssseg7e16_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg7e16_v_u16mf4((uint16_t *)(op0), (ptrdiff_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vssseg7e16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg7e16_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (ptrdiff_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (size_t)(op10)) +#define vssseg8e16_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg8e16_v_i16m1((int16_t *)(op0), (ptrdiff_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (size_t)(op10)) +#define vssseg8e16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vssseg8e16_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (ptrdiff_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (size_t)(op11)) +#define vssseg8e16_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg8e16_v_i16mf2((int16_t *)(op0), (ptrdiff_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (size_t)(op10)) +#define vssseg8e16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vssseg8e16_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (ptrdiff_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (size_t)(op11)) +#define vssseg8e16_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg8e16_v_i16mf4((int16_t *)(op0), (ptrdiff_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (size_t)(op10)) +#define vssseg8e16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vssseg8e16_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (ptrdiff_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (size_t)(op11)) +#define vssseg8e16_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg8e16_v_u16m1((uint16_t *)(op0), (ptrdiff_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (size_t)(op10)) +#define vssseg8e16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vssseg8e16_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (ptrdiff_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (size_t)(op11)) +#define vssseg8e16_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg8e16_v_u16mf2((uint16_t *)(op0), (ptrdiff_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (size_t)(op10)) +#define vssseg8e16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vssseg8e16_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (ptrdiff_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (size_t)(op11)) +#define vssseg8e16_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg8e16_v_u16mf4((uint16_t *)(op0), (ptrdiff_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (size_t)(op10)) +#define vssseg8e16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vssseg8e16_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (ptrdiff_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (size_t)(op11)) +#define vssseg2e32_v_i32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e32_v_i32m1((int32_t *)(op0), (ptrdiff_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vssseg2e32_v_i32m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e32_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (ptrdiff_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (size_t)(op5)) +#define vssseg2e32_v_i32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e32_v_i32m2((int32_t *)(op0), (ptrdiff_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vssseg2e32_v_i32m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e32_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (ptrdiff_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (size_t)(op5)) +#define vssseg2e32_v_i32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e32_v_i32m4((int32_t *)(op0), (ptrdiff_t)(op1), (vint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vssseg2e32_v_i32m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e32_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (ptrdiff_t)(op2), (vint32m4_t)(op3), (vint32m4_t)(op4), (size_t)(op5)) +#define vssseg2e32_v_i32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e32_v_i32mf2((int32_t *)(op0), (ptrdiff_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vssseg2e32_v_i32mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e32_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (ptrdiff_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (size_t)(op5)) +#define vssseg2e32_v_u32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e32_v_u32m1((uint32_t *)(op0), (ptrdiff_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vssseg2e32_v_u32m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e32_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (ptrdiff_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (size_t)(op5)) +#define vssseg2e32_v_u32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e32_v_u32m2((uint32_t *)(op0), (ptrdiff_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vssseg2e32_v_u32m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e32_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (ptrdiff_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (size_t)(op5)) +#define vssseg2e32_v_u32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e32_v_u32m4((uint32_t *)(op0), (ptrdiff_t)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vssseg2e32_v_u32m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e32_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (ptrdiff_t)(op2), (vuint32m4_t)(op3), (vuint32m4_t)(op4), (size_t)(op5)) +#define vssseg2e32_v_u32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e32_v_u32mf2((uint32_t *)(op0), (ptrdiff_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vssseg2e32_v_u32mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e32_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (ptrdiff_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (size_t)(op5)) +#define vssseg3e32_v_i32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg3e32_v_i32m1((int32_t *)(op0), (ptrdiff_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (size_t)(op5)) +#define vssseg3e32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg3e32_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (ptrdiff_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (size_t)(op6)) +#define vssseg3e32_v_i32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg3e32_v_i32m2((int32_t *)(op0), (ptrdiff_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (size_t)(op5)) +#define vssseg3e32_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg3e32_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (ptrdiff_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (vint32m2_t)(op5), (size_t)(op6)) +#define vssseg3e32_v_i32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg3e32_v_i32mf2((int32_t *)(op0), (ptrdiff_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (size_t)(op5)) +#define vssseg3e32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg3e32_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (ptrdiff_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (size_t)(op6)) +#define vssseg3e32_v_u32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg3e32_v_u32m1((uint32_t *)(op0), (ptrdiff_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (size_t)(op5)) +#define vssseg3e32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg3e32_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (ptrdiff_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (size_t)(op6)) +#define vssseg3e32_v_u32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg3e32_v_u32m2((uint32_t *)(op0), (ptrdiff_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (size_t)(op5)) +#define vssseg3e32_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg3e32_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (ptrdiff_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (vuint32m2_t)(op5), (size_t)(op6)) +#define vssseg3e32_v_u32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg3e32_v_u32mf2((uint32_t *)(op0), (ptrdiff_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (size_t)(op5)) +#define vssseg3e32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg3e32_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (ptrdiff_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (size_t)(op6)) +#define vssseg4e32_v_i32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg4e32_v_i32m1((int32_t *)(op0), (ptrdiff_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (size_t)(op6)) +#define vssseg4e32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg4e32_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (ptrdiff_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (size_t)(op7)) +#define vssseg4e32_v_i32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg4e32_v_i32m2((int32_t *)(op0), (ptrdiff_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (vint32m2_t)(op5), (size_t)(op6)) +#define vssseg4e32_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg4e32_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (ptrdiff_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (vint32m2_t)(op5), (vint32m2_t)(op6), (size_t)(op7)) +#define vssseg4e32_v_i32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg4e32_v_i32mf2((int32_t *)(op0), (ptrdiff_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (size_t)(op6)) +#define vssseg4e32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg4e32_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (ptrdiff_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (size_t)(op7)) +#define vssseg4e32_v_u32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg4e32_v_u32m1((uint32_t *)(op0), (ptrdiff_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (size_t)(op6)) +#define vssseg4e32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg4e32_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (ptrdiff_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vssseg4e32_v_u32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg4e32_v_u32m2((uint32_t *)(op0), (ptrdiff_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (vuint32m2_t)(op5), (size_t)(op6)) +#define vssseg4e32_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg4e32_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (ptrdiff_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (vuint32m2_t)(op5), (vuint32m2_t)(op6), (size_t)(op7)) +#define vssseg4e32_v_u32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg4e32_v_u32mf2((uint32_t *)(op0), (ptrdiff_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (size_t)(op6)) +#define vssseg4e32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg4e32_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (ptrdiff_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vssseg5e32_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg5e32_v_i32m1((int32_t *)(op0), (ptrdiff_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (size_t)(op7)) +#define vssseg5e32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg5e32_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (ptrdiff_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (size_t)(op8)) +#define vssseg5e32_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg5e32_v_i32mf2((int32_t *)(op0), (ptrdiff_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (size_t)(op7)) +#define vssseg5e32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg5e32_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (ptrdiff_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (size_t)(op8)) +#define vssseg5e32_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg5e32_v_u32m1((uint32_t *)(op0), (ptrdiff_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vssseg5e32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg5e32_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (ptrdiff_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (size_t)(op8)) +#define vssseg5e32_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg5e32_v_u32mf2((uint32_t *)(op0), (ptrdiff_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vssseg5e32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg5e32_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (ptrdiff_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (size_t)(op8)) +#define vssseg6e32_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg6e32_v_i32m1((int32_t *)(op0), (ptrdiff_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (size_t)(op8)) +#define vssseg6e32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg6e32_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (ptrdiff_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (size_t)(op9)) +#define vssseg6e32_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg6e32_v_i32mf2((int32_t *)(op0), (ptrdiff_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (size_t)(op8)) +#define vssseg6e32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg6e32_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (ptrdiff_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (size_t)(op9)) +#define vssseg6e32_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg6e32_v_u32m1((uint32_t *)(op0), (ptrdiff_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (size_t)(op8)) +#define vssseg6e32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg6e32_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (ptrdiff_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vssseg6e32_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg6e32_v_u32mf2((uint32_t *)(op0), (ptrdiff_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (size_t)(op8)) +#define vssseg6e32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg6e32_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (ptrdiff_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vssseg7e32_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg7e32_v_i32m1((int32_t *)(op0), (ptrdiff_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (size_t)(op9)) +#define vssseg7e32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg7e32_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (ptrdiff_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (size_t)(op10)) +#define vssseg7e32_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg7e32_v_i32mf2((int32_t *)(op0), (ptrdiff_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (size_t)(op9)) +#define vssseg7e32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg7e32_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (ptrdiff_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (size_t)(op10)) +#define vssseg7e32_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg7e32_v_u32m1((uint32_t *)(op0), (ptrdiff_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vssseg7e32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg7e32_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (ptrdiff_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (size_t)(op10)) +#define vssseg7e32_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg7e32_v_u32mf2((uint32_t *)(op0), (ptrdiff_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vssseg7e32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg7e32_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (ptrdiff_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (size_t)(op10)) +#define vssseg8e32_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg8e32_v_i32m1((int32_t *)(op0), (ptrdiff_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (size_t)(op10)) +#define vssseg8e32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vssseg8e32_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (ptrdiff_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (size_t)(op11)) +#define vssseg8e32_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg8e32_v_i32mf2((int32_t *)(op0), (ptrdiff_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (size_t)(op10)) +#define vssseg8e32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vssseg8e32_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (ptrdiff_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (size_t)(op11)) +#define vssseg8e32_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg8e32_v_u32m1((uint32_t *)(op0), (ptrdiff_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (size_t)(op10)) +#define vssseg8e32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vssseg8e32_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (ptrdiff_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (size_t)(op11)) +#define vssseg8e32_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg8e32_v_u32mf2((uint32_t *)(op0), (ptrdiff_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (size_t)(op10)) +#define vssseg8e32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vssseg8e32_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (ptrdiff_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (size_t)(op11)) +#define vssseg2e64_v_i64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e64_v_i64m1((int64_t *)(op0), (ptrdiff_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vssseg2e64_v_i64m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e64_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (ptrdiff_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (size_t)(op5)) +#define vssseg2e64_v_i64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e64_v_i64m2((int64_t *)(op0), (ptrdiff_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vssseg2e64_v_i64m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e64_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (ptrdiff_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (size_t)(op5)) +#define vssseg2e64_v_i64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e64_v_i64m4((int64_t *)(op0), (ptrdiff_t)(op1), (vint64m4_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vssseg2e64_v_i64m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e64_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (ptrdiff_t)(op2), (vint64m4_t)(op3), (vint64m4_t)(op4), (size_t)(op5)) +#define vssseg2e64_v_u64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e64_v_u64m1((uint64_t *)(op0), (ptrdiff_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vssseg2e64_v_u64m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e64_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (ptrdiff_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (size_t)(op5)) +#define vssseg2e64_v_u64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e64_v_u64m2((uint64_t *)(op0), (ptrdiff_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vssseg2e64_v_u64m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e64_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (ptrdiff_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (size_t)(op5)) +#define vssseg2e64_v_u64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e64_v_u64m4((uint64_t *)(op0), (ptrdiff_t)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vssseg2e64_v_u64m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e64_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (ptrdiff_t)(op2), (vuint64m4_t)(op3), (vuint64m4_t)(op4), (size_t)(op5)) +#define vssseg3e64_v_i64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg3e64_v_i64m1((int64_t *)(op0), (ptrdiff_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (size_t)(op5)) +#define vssseg3e64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg3e64_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (ptrdiff_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (size_t)(op6)) +#define vssseg3e64_v_i64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg3e64_v_i64m2((int64_t *)(op0), (ptrdiff_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (size_t)(op5)) +#define vssseg3e64_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg3e64_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (ptrdiff_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (vint64m2_t)(op5), (size_t)(op6)) +#define vssseg3e64_v_u64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg3e64_v_u64m1((uint64_t *)(op0), (ptrdiff_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (size_t)(op5)) +#define vssseg3e64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg3e64_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (ptrdiff_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (size_t)(op6)) +#define vssseg3e64_v_u64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg3e64_v_u64m2((uint64_t *)(op0), (ptrdiff_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (size_t)(op5)) +#define vssseg3e64_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg3e64_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (ptrdiff_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (vuint64m2_t)(op5), (size_t)(op6)) +#define vssseg4e64_v_i64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg4e64_v_i64m1((int64_t *)(op0), (ptrdiff_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (size_t)(op6)) +#define vssseg4e64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg4e64_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (ptrdiff_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (size_t)(op7)) +#define vssseg4e64_v_i64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg4e64_v_i64m2((int64_t *)(op0), (ptrdiff_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (vint64m2_t)(op5), (size_t)(op6)) +#define vssseg4e64_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg4e64_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (ptrdiff_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (vint64m2_t)(op5), (vint64m2_t)(op6), (size_t)(op7)) +#define vssseg4e64_v_u64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg4e64_v_u64m1((uint64_t *)(op0), (ptrdiff_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (size_t)(op6)) +#define vssseg4e64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg4e64_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (ptrdiff_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vssseg4e64_v_u64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg4e64_v_u64m2((uint64_t *)(op0), (ptrdiff_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (vuint64m2_t)(op5), (size_t)(op6)) +#define vssseg4e64_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg4e64_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (ptrdiff_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (vuint64m2_t)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vssseg5e64_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg5e64_v_i64m1((int64_t *)(op0), (ptrdiff_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (size_t)(op7)) +#define vssseg5e64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg5e64_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (ptrdiff_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (size_t)(op8)) +#define vssseg5e64_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg5e64_v_u64m1((uint64_t *)(op0), (ptrdiff_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vssseg5e64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg5e64_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (ptrdiff_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (size_t)(op8)) +#define vssseg6e64_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg6e64_v_i64m1((int64_t *)(op0), (ptrdiff_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (size_t)(op8)) +#define vssseg6e64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg6e64_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (ptrdiff_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (size_t)(op9)) +#define vssseg6e64_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg6e64_v_u64m1((uint64_t *)(op0), (ptrdiff_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (size_t)(op8)) +#define vssseg6e64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg6e64_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (ptrdiff_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vssseg7e64_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg7e64_v_i64m1((int64_t *)(op0), (ptrdiff_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (size_t)(op9)) +#define vssseg7e64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg7e64_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (ptrdiff_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (size_t)(op10)) +#define vssseg7e64_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg7e64_v_u64m1((uint64_t *)(op0), (ptrdiff_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vssseg7e64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg7e64_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (ptrdiff_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (size_t)(op10)) +#define vssseg8e64_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg8e64_v_i64m1((int64_t *)(op0), (ptrdiff_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (size_t)(op10)) +#define vssseg8e64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vssseg8e64_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (ptrdiff_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (size_t)(op11)) +#define vssseg8e64_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg8e64_v_u64m1((uint64_t *)(op0), (ptrdiff_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (size_t)(op10)) +#define vssseg8e64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vssseg8e64_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (ptrdiff_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (size_t)(op11)) +#define vsuxseg3ei8_v_i8m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei8_v_i8m1((int8_t *)(op0), (vuint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (size_t)(op5)) +#define vsuxseg3ei8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei8_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (size_t)(op6)) +#define vsuxseg3ei8_v_i8m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei8_v_i8m2((int8_t *)(op0), (vuint8m2_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei8_v_i8m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei8_v_i8m2_m((vbool4_t)(op0), (int8_t *)(op1), (vuint8m2_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (vint8m2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei8_v_i8mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei8_v_i8mf2((int8_t *)(op0), (vuint8mf2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei8_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei8_v_i8mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei8_v_i8mf4((int8_t *)(op0), (vuint8mf4_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (size_t)(op5)) +#define vsuxseg3ei8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei8_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (size_t)(op6)) +#define vsuxseg3ei8_v_i8mf8(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei8_v_i8mf8((int8_t *)(op0), (vuint8mf8_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (size_t)(op5)) +#define vsuxseg3ei8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei8_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (size_t)(op6)) +#define vsuxseg3ei8_v_u8m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei8_v_u8m1((uint8_t *)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (size_t)(op5)) +#define vsuxseg3ei8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei8_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (size_t)(op6)) +#define vsuxseg3ei8_v_u8m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei8_v_u8m2((uint8_t *)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei8_v_u8m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei8_v_u8m2_m((vbool4_t)(op0), (uint8_t *)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (vuint8m2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei8_v_u8mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei8_v_u8mf2((uint8_t *)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei8_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei8_v_u8mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei8_v_u8mf4((uint8_t *)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (size_t)(op5)) +#define vsuxseg3ei8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei8_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (size_t)(op6)) +#define vsuxseg3ei8_v_u8mf8(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei8_v_u8mf8((uint8_t *)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (size_t)(op5)) +#define vsuxseg3ei8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei8_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (size_t)(op6)) +#define vsuxseg4ei8_v_i8m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei8_v_i8m1((int8_t *)(op0), (vuint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (size_t)(op6)) +#define vsuxseg4ei8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei8_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (size_t)(op7)) +#define vsuxseg4ei8_v_i8m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei8_v_i8m2((int8_t *)(op0), (vuint8m2_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (vint8m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei8_v_i8m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei8_v_i8m2_m((vbool4_t)(op0), (int8_t *)(op1), (vuint8m2_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (vint8m2_t)(op5), (vint8m2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei8_v_i8mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei8_v_i8mf2((int8_t *)(op0), (vuint8mf2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei8_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei8_v_i8mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei8_v_i8mf4((int8_t *)(op0), (vuint8mf4_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (size_t)(op6)) +#define vsuxseg4ei8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei8_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (size_t)(op7)) +#define vsuxseg4ei8_v_i8mf8(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei8_v_i8mf8((int8_t *)(op0), (vuint8mf8_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (size_t)(op6)) +#define vsuxseg4ei8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei8_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (size_t)(op7)) +#define vsuxseg4ei8_v_u8m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei8_v_u8m1((uint8_t *)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (size_t)(op6)) +#define vsuxseg4ei8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei8_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (size_t)(op7)) +#define vsuxseg4ei8_v_u8m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei8_v_u8m2((uint8_t *)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (vuint8m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei8_v_u8m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei8_v_u8m2_m((vbool4_t)(op0), (uint8_t *)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (vuint8m2_t)(op5), (vuint8m2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei8_v_u8mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei8_v_u8mf2((uint8_t *)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei8_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei8_v_u8mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei8_v_u8mf4((uint8_t *)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (size_t)(op6)) +#define vsuxseg4ei8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei8_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vsuxseg4ei8_v_u8mf8(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei8_v_u8mf8((uint8_t *)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (size_t)(op6)) +#define vsuxseg4ei8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei8_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vsuxseg5ei8_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei8_v_i8m1((int8_t *)(op0), (vuint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (size_t)(op7)) +#define vsuxseg5ei8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei8_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (size_t)(op8)) +#define vsuxseg5ei8_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei8_v_i8mf2((int8_t *)(op0), (vuint8mf2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei8_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (size_t)(op8)) +#define vsuxseg5ei8_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei8_v_i8mf4((int8_t *)(op0), (vuint8mf4_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (size_t)(op7)) +#define vsuxseg5ei8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei8_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (size_t)(op8)) +#define vsuxseg5ei8_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei8_v_i8mf8((int8_t *)(op0), (vuint8mf8_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (size_t)(op7)) +#define vsuxseg5ei8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei8_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (size_t)(op8)) +#define vsuxseg5ei8_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei8_v_u8m1((uint8_t *)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (size_t)(op7)) +#define vsuxseg5ei8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei8_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (size_t)(op8)) +#define vsuxseg5ei8_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei8_v_u8mf2((uint8_t *)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei8_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (size_t)(op8)) +#define vsuxseg5ei8_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei8_v_u8mf4((uint8_t *)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vsuxseg5ei8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei8_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (size_t)(op8)) +#define vsuxseg5ei8_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei8_v_u8mf8((uint8_t *)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vsuxseg5ei8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei8_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (size_t)(op8)) +#define vsuxseg6ei8_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei8_v_i8m1((int8_t *)(op0), (vuint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei8_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (size_t)(op9)) +#define vsuxseg6ei8_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei8_v_i8mf2((int8_t *)(op0), (vuint8mf2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (size_t)(op8)) +#define vsuxseg6ei8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei8_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (size_t)(op9)) +#define vsuxseg6ei8_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei8_v_i8mf4((int8_t *)(op0), (vuint8mf4_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (size_t)(op8)) +#define vsuxseg6ei8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei8_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (size_t)(op9)) +#define vsuxseg6ei8_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei8_v_i8mf8((int8_t *)(op0), (vuint8mf8_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (size_t)(op8)) +#define vsuxseg6ei8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei8_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (size_t)(op9)) +#define vsuxseg6ei8_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei8_v_u8m1((uint8_t *)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei8_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (size_t)(op9)) +#define vsuxseg6ei8_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei8_v_u8mf2((uint8_t *)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (size_t)(op8)) +#define vsuxseg6ei8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei8_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vsuxseg6ei8_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei8_v_u8mf4((uint8_t *)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (size_t)(op8)) +#define vsuxseg6ei8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei8_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vsuxseg6ei8_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei8_v_u8mf8((uint8_t *)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (size_t)(op8)) +#define vsuxseg6ei8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei8_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vsuxseg7ei8_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei8_v_i8m1((int8_t *)(op0), (vuint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei8_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (size_t)(op10)) +#define vsuxseg7ei8_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei8_v_i8mf2((int8_t *)(op0), (vuint8mf2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (size_t)(op9)) +#define vsuxseg7ei8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei8_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (size_t)(op10)) +#define vsuxseg7ei8_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei8_v_i8mf4((int8_t *)(op0), (vuint8mf4_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (size_t)(op9)) +#define vsuxseg7ei8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei8_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (size_t)(op10)) +#define vsuxseg7ei8_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei8_v_i8mf8((int8_t *)(op0), (vuint8mf8_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (size_t)(op9)) +#define vsuxseg7ei8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei8_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (size_t)(op10)) +#define vsuxseg7ei8_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei8_v_u8m1((uint8_t *)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei8_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (size_t)(op10)) +#define vsuxseg7ei8_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei8_v_u8mf2((uint8_t *)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vsuxseg7ei8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei8_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (size_t)(op10)) +#define vsuxseg7ei8_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei8_v_u8mf4((uint8_t *)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vsuxseg7ei8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei8_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (size_t)(op10)) +#define vsuxseg7ei8_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei8_v_u8mf8((uint8_t *)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vsuxseg7ei8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei8_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (size_t)(op10)) +#define vsuxseg8ei8_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei8_v_i8m1((int8_t *)(op0), (vuint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei8_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (size_t)(op11)) +#define vsuxseg8ei8_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei8_v_i8mf2((int8_t *)(op0), (vuint8mf2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (size_t)(op10)) +#define vsuxseg8ei8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei8_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (size_t)(op11)) +#define vsuxseg8ei8_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei8_v_i8mf4((int8_t *)(op0), (vuint8mf4_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (size_t)(op10)) +#define vsuxseg8ei8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei8_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (size_t)(op11)) +#define vsuxseg8ei8_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei8_v_i8mf8((int8_t *)(op0), (vuint8mf8_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (size_t)(op10)) +#define vsuxseg8ei8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei8_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (size_t)(op11)) +#define vsuxseg8ei8_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei8_v_u8m1((uint8_t *)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei8_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (size_t)(op11)) +#define vsuxseg8ei8_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei8_v_u8mf2((uint8_t *)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (size_t)(op10)) +#define vsuxseg8ei8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei8_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (size_t)(op11)) +#define vsuxseg8ei8_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei8_v_u8mf4((uint8_t *)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (size_t)(op10)) +#define vsuxseg8ei8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei8_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (size_t)(op11)) +#define vsuxseg8ei8_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei8_v_u8mf8((uint8_t *)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (size_t)(op10)) +#define vsuxseg8ei8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei8_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (size_t)(op11)) +#define vsuxseg2ei16_v_i8m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_i8m1((int8_t *)(op0), (vuint16m2_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_i8m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint16m2_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (size_t)(op5)) +#define vsuxseg2ei16_v_i8m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_i8m2((int8_t *)(op0), (vuint16m4_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_i8m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_i8m2_m((vbool4_t)(op0), (int8_t *)(op1), (vuint16m4_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei16_v_i8m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_i8m4((int8_t *)(op0), (vuint16m8_t)(op1), (vint8m4_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_i8m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_i8m4_m((vbool2_t)(op0), (int8_t *)(op1), (vuint16m8_t)(op2), (vint8m4_t)(op3), (vint8m4_t)(op4), (size_t)(op5)) +#define vsuxseg2ei16_v_i8mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_i8mf2((int8_t *)(op0), (vuint16m1_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_i8mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint16m1_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei16_v_i8mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_i8mf4((int8_t *)(op0), (vuint16mf2_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_i8mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint16mf2_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (size_t)(op5)) +#define vsuxseg2ei16_v_i8mf8(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_i8mf8((int8_t *)(op0), (vuint16mf4_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_i8mf8_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint16mf4_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (size_t)(op5)) +#define vsuxseg2ei16_v_u8m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_u8m1((uint8_t *)(op0), (vuint16m2_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_u8m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint16m2_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (size_t)(op5)) +#define vsuxseg2ei16_v_u8m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_u8m2((uint8_t *)(op0), (vuint16m4_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_u8m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_u8m2_m((vbool4_t)(op0), (uint8_t *)(op1), (vuint16m4_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei16_v_u8m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_u8m4((uint8_t *)(op0), (vuint16m8_t)(op1), (vuint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_u8m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_u8m4_m((vbool2_t)(op0), (uint8_t *)(op1), (vuint16m8_t)(op2), (vuint8m4_t)(op3), (vuint8m4_t)(op4), (size_t)(op5)) +#define vsuxseg2ei16_v_u8mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_u8mf2((uint8_t *)(op0), (vuint16m1_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_u8mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint16m1_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei16_v_u8mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_u8mf4((uint8_t *)(op0), (vuint16mf2_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_u8mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint16mf2_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (size_t)(op5)) +#define vsuxseg2ei16_v_u8mf8(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_u8mf8((uint8_t *)(op0), (vuint16mf4_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_u8mf8_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint16mf4_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (size_t)(op5)) +#define vsuxseg3ei16_v_i8m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei16_v_i8m1((int8_t *)(op0), (vuint16m2_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (size_t)(op5)) +#define vsuxseg3ei16_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei16_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint16m2_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (size_t)(op6)) +#define vsuxseg3ei16_v_i8m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei16_v_i8m2((int8_t *)(op0), (vuint16m4_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei16_v_i8m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei16_v_i8m2_m((vbool4_t)(op0), (int8_t *)(op1), (vuint16m4_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (vint8m2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei16_v_i8mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei16_v_i8mf2((int8_t *)(op0), (vuint16m1_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei16_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei16_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint16m1_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei16_v_i8mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei16_v_i8mf4((int8_t *)(op0), (vuint16mf2_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (size_t)(op5)) +#define vsuxseg3ei16_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei16_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint16mf2_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (size_t)(op6)) +#define vsuxseg3ei16_v_i8mf8(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei16_v_i8mf8((int8_t *)(op0), (vuint16mf4_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (size_t)(op5)) +#define vsuxseg3ei16_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei16_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint16mf4_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (size_t)(op6)) +#define vsuxseg3ei16_v_u8m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei16_v_u8m1((uint8_t *)(op0), (vuint16m2_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (size_t)(op5)) +#define vsuxseg3ei16_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei16_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint16m2_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (size_t)(op6)) +#define vsuxseg3ei16_v_u8m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei16_v_u8m2((uint8_t *)(op0), (vuint16m4_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei16_v_u8m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei16_v_u8m2_m((vbool4_t)(op0), (uint8_t *)(op1), (vuint16m4_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (vuint8m2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei16_v_u8mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei16_v_u8mf2((uint8_t *)(op0), (vuint16m1_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei16_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei16_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint16m1_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei16_v_u8mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei16_v_u8mf4((uint8_t *)(op0), (vuint16mf2_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (size_t)(op5)) +#define vsuxseg3ei16_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei16_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint16mf2_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (size_t)(op6)) +#define vsuxseg3ei16_v_u8mf8(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei16_v_u8mf8((uint8_t *)(op0), (vuint16mf4_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (size_t)(op5)) +#define vsuxseg3ei16_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei16_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint16mf4_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (size_t)(op6)) +#define vsuxseg4ei16_v_i8m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei16_v_i8m1((int8_t *)(op0), (vuint16m2_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (size_t)(op6)) +#define vsuxseg4ei16_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei16_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint16m2_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (size_t)(op7)) +#define vsuxseg4ei16_v_i8m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei16_v_i8m2((int8_t *)(op0), (vuint16m4_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (vint8m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei16_v_i8m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei16_v_i8m2_m((vbool4_t)(op0), (int8_t *)(op1), (vuint16m4_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (vint8m2_t)(op5), (vint8m2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei16_v_i8mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei16_v_i8mf2((int8_t *)(op0), (vuint16m1_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei16_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei16_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint16m1_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei16_v_i8mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei16_v_i8mf4((int8_t *)(op0), (vuint16mf2_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (size_t)(op6)) +#define vsuxseg4ei16_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei16_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint16mf2_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (size_t)(op7)) +#define vsuxseg4ei16_v_i8mf8(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei16_v_i8mf8((int8_t *)(op0), (vuint16mf4_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (size_t)(op6)) +#define vsuxseg4ei16_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei16_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint16mf4_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (size_t)(op7)) +#define vsuxseg4ei16_v_u8m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei16_v_u8m1((uint8_t *)(op0), (vuint16m2_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (size_t)(op6)) +#define vsuxseg4ei16_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei16_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint16m2_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (size_t)(op7)) +#define vsuxseg4ei16_v_u8m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei16_v_u8m2((uint8_t *)(op0), (vuint16m4_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (vuint8m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei16_v_u8m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei16_v_u8m2_m((vbool4_t)(op0), (uint8_t *)(op1), (vuint16m4_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (vuint8m2_t)(op5), (vuint8m2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei16_v_u8mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei16_v_u8mf2((uint8_t *)(op0), (vuint16m1_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei16_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei16_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint16m1_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei16_v_u8mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei16_v_u8mf4((uint8_t *)(op0), (vuint16mf2_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (size_t)(op6)) +#define vsuxseg4ei16_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei16_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint16mf2_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vsuxseg4ei16_v_u8mf8(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei16_v_u8mf8((uint8_t *)(op0), (vuint16mf4_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (size_t)(op6)) +#define vsuxseg4ei16_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei16_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint16mf4_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vsuxseg5ei16_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei16_v_i8m1((int8_t *)(op0), (vuint16m2_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (size_t)(op7)) +#define vsuxseg5ei16_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei16_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint16m2_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (size_t)(op8)) +#define vsuxseg5ei16_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei16_v_i8mf2((int8_t *)(op0), (vuint16m1_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei16_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei16_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint16m1_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (size_t)(op8)) +#define vsuxseg5ei16_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei16_v_i8mf4((int8_t *)(op0), (vuint16mf2_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (size_t)(op7)) +#define vsuxseg5ei16_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei16_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint16mf2_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (size_t)(op8)) +#define vsuxseg5ei16_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei16_v_i8mf8((int8_t *)(op0), (vuint16mf4_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (size_t)(op7)) +#define vsuxseg5ei16_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei16_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint16mf4_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (size_t)(op8)) +#define vsuxseg5ei16_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei16_v_u8m1((uint8_t *)(op0), (vuint16m2_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (size_t)(op7)) +#define vsuxseg5ei16_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei16_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint16m2_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (size_t)(op8)) +#define vsuxseg5ei16_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei16_v_u8mf2((uint8_t *)(op0), (vuint16m1_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei16_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei16_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint16m1_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (size_t)(op8)) +#define vsuxseg5ei16_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei16_v_u8mf4((uint8_t *)(op0), (vuint16mf2_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vsuxseg5ei16_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei16_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint16mf2_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (size_t)(op8)) +#define vsuxseg5ei16_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei16_v_u8mf8((uint8_t *)(op0), (vuint16mf4_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vsuxseg5ei16_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei16_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint16mf4_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (size_t)(op8)) +#define vsuxseg6ei16_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei16_v_i8m1((int8_t *)(op0), (vuint16m2_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei16_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei16_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint16m2_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (size_t)(op9)) +#define vsuxseg6ei16_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei16_v_i8mf2((int8_t *)(op0), (vuint16m1_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (size_t)(op8)) +#define vsuxseg6ei16_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei16_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint16m1_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (size_t)(op9)) +#define vsuxseg6ei16_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei16_v_i8mf4((int8_t *)(op0), (vuint16mf2_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (size_t)(op8)) +#define vsuxseg6ei16_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei16_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint16mf2_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (size_t)(op9)) +#define vsuxseg6ei16_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei16_v_i8mf8((int8_t *)(op0), (vuint16mf4_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (size_t)(op8)) +#define vsuxseg6ei16_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei16_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint16mf4_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (size_t)(op9)) +#define vsuxseg6ei16_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei16_v_u8m1((uint8_t *)(op0), (vuint16m2_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei16_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei16_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint16m2_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (size_t)(op9)) +#define vsuxseg6ei16_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei16_v_u8mf2((uint8_t *)(op0), (vuint16m1_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (size_t)(op8)) +#define vsuxseg6ei16_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei16_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint16m1_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vsuxseg6ei16_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei16_v_u8mf4((uint8_t *)(op0), (vuint16mf2_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (size_t)(op8)) +#define vsuxseg6ei16_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei16_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint16mf2_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vsuxseg6ei16_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei16_v_u8mf8((uint8_t *)(op0), (vuint16mf4_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (size_t)(op8)) +#define vsuxseg6ei16_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei16_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint16mf4_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vsuxseg7ei16_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei16_v_i8m1((int8_t *)(op0), (vuint16m2_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei16_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei16_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint16m2_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (size_t)(op10)) +#define vsuxseg7ei16_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei16_v_i8mf2((int8_t *)(op0), (vuint16m1_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (size_t)(op9)) +#define vsuxseg7ei16_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei16_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint16m1_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (size_t)(op10)) +#define vsuxseg7ei16_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei16_v_i8mf4((int8_t *)(op0), (vuint16mf2_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (size_t)(op9)) +#define vsuxseg7ei16_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei16_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint16mf2_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (size_t)(op10)) +#define vsuxseg7ei16_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei16_v_i8mf8((int8_t *)(op0), (vuint16mf4_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (size_t)(op9)) +#define vsuxseg7ei16_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei16_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint16mf4_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (size_t)(op10)) +#define vsuxseg7ei16_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei16_v_u8m1((uint8_t *)(op0), (vuint16m2_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei16_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei16_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint16m2_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (size_t)(op10)) +#define vsuxseg7ei16_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei16_v_u8mf2((uint8_t *)(op0), (vuint16m1_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vsuxseg7ei16_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei16_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint16m1_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (size_t)(op10)) +#define vsuxseg7ei16_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei16_v_u8mf4((uint8_t *)(op0), (vuint16mf2_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vsuxseg7ei16_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei16_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint16mf2_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (size_t)(op10)) +#define vsuxseg7ei16_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei16_v_u8mf8((uint8_t *)(op0), (vuint16mf4_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vsuxseg7ei16_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei16_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint16mf4_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (size_t)(op10)) +#define vsuxseg8ei16_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei16_v_i8m1((int8_t *)(op0), (vuint16m2_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei16_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei16_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint16m2_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (size_t)(op11)) +#define vsuxseg8ei16_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei16_v_i8mf2((int8_t *)(op0), (vuint16m1_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (size_t)(op10)) +#define vsuxseg8ei16_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei16_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint16m1_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (size_t)(op11)) +#define vsuxseg8ei16_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei16_v_i8mf4((int8_t *)(op0), (vuint16mf2_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (size_t)(op10)) +#define vsuxseg8ei16_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei16_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint16mf2_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (size_t)(op11)) +#define vsuxseg8ei16_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei16_v_i8mf8((int8_t *)(op0), (vuint16mf4_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (size_t)(op10)) +#define vsuxseg8ei16_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei16_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint16mf4_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (size_t)(op11)) +#define vsuxseg8ei16_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei16_v_u8m1((uint8_t *)(op0), (vuint16m2_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei16_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei16_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint16m2_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (size_t)(op11)) +#define vsuxseg8ei16_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei16_v_u8mf2((uint8_t *)(op0), (vuint16m1_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (size_t)(op10)) +#define vsuxseg8ei16_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei16_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint16m1_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (size_t)(op11)) +#define vsuxseg8ei16_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei16_v_u8mf4((uint8_t *)(op0), (vuint16mf2_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (size_t)(op10)) +#define vsuxseg8ei16_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei16_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint16mf2_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (size_t)(op11)) +#define vsuxseg8ei16_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei16_v_u8mf8((uint8_t *)(op0), (vuint16mf4_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (size_t)(op10)) +#define vsuxseg8ei16_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei16_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint16mf4_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (size_t)(op11)) +#define vsuxseg2ei32_v_i8m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_i8m1((int8_t *)(op0), (vuint32m4_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_i8m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint32m4_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (size_t)(op5)) +#define vsuxseg2ei32_v_i8m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_i8m2((int8_t *)(op0), (vuint32m8_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_i8m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_i8m2_m((vbool4_t)(op0), (int8_t *)(op1), (vuint32m8_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei32_v_i8mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_i8mf2((int8_t *)(op0), (vuint32m2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_i8mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint32m2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei32_v_i8mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_i8mf4((int8_t *)(op0), (vuint32m1_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_i8mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint32m1_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (size_t)(op5)) +#define vsuxseg2ei32_v_i8mf8(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_i8mf8((int8_t *)(op0), (vuint32mf2_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_i8mf8_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint32mf2_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (size_t)(op5)) +#define vsuxseg2ei32_v_u8m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_u8m1((uint8_t *)(op0), (vuint32m4_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_u8m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint32m4_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (size_t)(op5)) +#define vsuxseg2ei32_v_u8m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_u8m2((uint8_t *)(op0), (vuint32m8_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_u8m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_u8m2_m((vbool4_t)(op0), (uint8_t *)(op1), (vuint32m8_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei32_v_u8mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_u8mf2((uint8_t *)(op0), (vuint32m2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_u8mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint32m2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei32_v_u8mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_u8mf4((uint8_t *)(op0), (vuint32m1_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_u8mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint32m1_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (size_t)(op5)) +#define vsuxseg2ei32_v_u8mf8(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_u8mf8((uint8_t *)(op0), (vuint32mf2_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_u8mf8_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint32mf2_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (size_t)(op5)) +#define vsuxseg3ei32_v_i8m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei32_v_i8m1((int8_t *)(op0), (vuint32m4_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (size_t)(op5)) +#define vsuxseg3ei32_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei32_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint32m4_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (size_t)(op6)) +#define vsuxseg3ei32_v_i8m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei32_v_i8m2((int8_t *)(op0), (vuint32m8_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei32_v_i8m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei32_v_i8m2_m((vbool4_t)(op0), (int8_t *)(op1), (vuint32m8_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (vint8m2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei32_v_i8mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei32_v_i8mf2((int8_t *)(op0), (vuint32m2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei32_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei32_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint32m2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei32_v_i8mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei32_v_i8mf4((int8_t *)(op0), (vuint32m1_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (size_t)(op5)) +#define vsuxseg3ei32_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei32_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint32m1_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (size_t)(op6)) +#define vsuxseg3ei32_v_i8mf8(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei32_v_i8mf8((int8_t *)(op0), (vuint32mf2_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (size_t)(op5)) +#define vsuxseg3ei32_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei32_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint32mf2_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (size_t)(op6)) +#define vsuxseg3ei32_v_u8m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei32_v_u8m1((uint8_t *)(op0), (vuint32m4_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (size_t)(op5)) +#define vsuxseg3ei32_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei32_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint32m4_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (size_t)(op6)) +#define vsuxseg3ei32_v_u8m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei32_v_u8m2((uint8_t *)(op0), (vuint32m8_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei32_v_u8m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei32_v_u8m2_m((vbool4_t)(op0), (uint8_t *)(op1), (vuint32m8_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (vuint8m2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei32_v_u8mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei32_v_u8mf2((uint8_t *)(op0), (vuint32m2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei32_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei32_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint32m2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei32_v_u8mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei32_v_u8mf4((uint8_t *)(op0), (vuint32m1_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (size_t)(op5)) +#define vsuxseg3ei32_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei32_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint32m1_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (size_t)(op6)) +#define vsuxseg3ei32_v_u8mf8(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei32_v_u8mf8((uint8_t *)(op0), (vuint32mf2_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (size_t)(op5)) +#define vsuxseg3ei32_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei32_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint32mf2_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (size_t)(op6)) +#define vsuxseg4ei32_v_i8m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei32_v_i8m1((int8_t *)(op0), (vuint32m4_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (size_t)(op6)) +#define vsuxseg4ei32_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei32_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint32m4_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (size_t)(op7)) +#define vsuxseg4ei32_v_i8m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei32_v_i8m2((int8_t *)(op0), (vuint32m8_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (vint8m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei32_v_i8m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei32_v_i8m2_m((vbool4_t)(op0), (int8_t *)(op1), (vuint32m8_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (vint8m2_t)(op5), (vint8m2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei32_v_i8mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei32_v_i8mf2((int8_t *)(op0), (vuint32m2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei32_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei32_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint32m2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei32_v_i8mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei32_v_i8mf4((int8_t *)(op0), (vuint32m1_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (size_t)(op6)) +#define vsuxseg4ei32_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei32_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint32m1_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (size_t)(op7)) +#define vsuxseg4ei32_v_i8mf8(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei32_v_i8mf8((int8_t *)(op0), (vuint32mf2_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (size_t)(op6)) +#define vsuxseg4ei32_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei32_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint32mf2_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (size_t)(op7)) +#define vsuxseg4ei32_v_u8m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei32_v_u8m1((uint8_t *)(op0), (vuint32m4_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (size_t)(op6)) +#define vsuxseg4ei32_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei32_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint32m4_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (size_t)(op7)) +#define vsuxseg4ei32_v_u8m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei32_v_u8m2((uint8_t *)(op0), (vuint32m8_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (vuint8m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei32_v_u8m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei32_v_u8m2_m((vbool4_t)(op0), (uint8_t *)(op1), (vuint32m8_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (vuint8m2_t)(op5), (vuint8m2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei32_v_u8mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei32_v_u8mf2((uint8_t *)(op0), (vuint32m2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei32_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei32_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint32m2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei32_v_u8mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei32_v_u8mf4((uint8_t *)(op0), (vuint32m1_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (size_t)(op6)) +#define vsuxseg4ei32_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei32_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint32m1_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vsuxseg4ei32_v_u8mf8(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei32_v_u8mf8((uint8_t *)(op0), (vuint32mf2_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (size_t)(op6)) +#define vsuxseg4ei32_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei32_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint32mf2_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vsuxseg5ei32_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei32_v_i8m1((int8_t *)(op0), (vuint32m4_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (size_t)(op7)) +#define vsuxseg5ei32_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei32_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint32m4_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (size_t)(op8)) +#define vsuxseg5ei32_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei32_v_i8mf2((int8_t *)(op0), (vuint32m2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei32_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei32_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint32m2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (size_t)(op8)) +#define vsuxseg5ei32_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei32_v_i8mf4((int8_t *)(op0), (vuint32m1_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (size_t)(op7)) +#define vsuxseg5ei32_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei32_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint32m1_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (size_t)(op8)) +#define vsuxseg5ei32_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei32_v_i8mf8((int8_t *)(op0), (vuint32mf2_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (size_t)(op7)) +#define vsuxseg5ei32_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei32_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint32mf2_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (size_t)(op8)) +#define vsuxseg5ei32_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei32_v_u8m1((uint8_t *)(op0), (vuint32m4_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (size_t)(op7)) +#define vsuxseg5ei32_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei32_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint32m4_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (size_t)(op8)) +#define vsuxseg5ei32_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei32_v_u8mf2((uint8_t *)(op0), (vuint32m2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei32_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei32_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint32m2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (size_t)(op8)) +#define vsuxseg5ei32_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei32_v_u8mf4((uint8_t *)(op0), (vuint32m1_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vsuxseg5ei32_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei32_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint32m1_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (size_t)(op8)) +#define vsuxseg5ei32_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei32_v_u8mf8((uint8_t *)(op0), (vuint32mf2_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vsuxseg5ei32_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei32_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint32mf2_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (size_t)(op8)) +#define vsuxseg6ei32_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei32_v_i8m1((int8_t *)(op0), (vuint32m4_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei32_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei32_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint32m4_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (size_t)(op9)) +#define vsuxseg6ei32_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei32_v_i8mf2((int8_t *)(op0), (vuint32m2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (size_t)(op8)) +#define vsuxseg6ei32_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei32_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint32m2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (size_t)(op9)) +#define vsuxseg6ei32_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei32_v_i8mf4((int8_t *)(op0), (vuint32m1_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (size_t)(op8)) +#define vsuxseg6ei32_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei32_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint32m1_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (size_t)(op9)) +#define vsuxseg6ei32_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei32_v_i8mf8((int8_t *)(op0), (vuint32mf2_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (size_t)(op8)) +#define vsuxseg6ei32_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei32_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint32mf2_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (size_t)(op9)) +#define vsuxseg6ei32_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei32_v_u8m1((uint8_t *)(op0), (vuint32m4_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei32_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei32_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint32m4_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (size_t)(op9)) +#define vsuxseg6ei32_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei32_v_u8mf2((uint8_t *)(op0), (vuint32m2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (size_t)(op8)) +#define vsuxseg6ei32_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei32_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint32m2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vsuxseg6ei32_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei32_v_u8mf4((uint8_t *)(op0), (vuint32m1_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (size_t)(op8)) +#define vsuxseg6ei32_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei32_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint32m1_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vsuxseg6ei32_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei32_v_u8mf8((uint8_t *)(op0), (vuint32mf2_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (size_t)(op8)) +#define vsuxseg6ei32_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei32_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint32mf2_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vsuxseg7ei32_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei32_v_i8m1((int8_t *)(op0), (vuint32m4_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei32_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei32_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint32m4_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (size_t)(op10)) +#define vsuxseg7ei32_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei32_v_i8mf2((int8_t *)(op0), (vuint32m2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (size_t)(op9)) +#define vsuxseg7ei32_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei32_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint32m2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (size_t)(op10)) +#define vsuxseg7ei32_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei32_v_i8mf4((int8_t *)(op0), (vuint32m1_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (size_t)(op9)) +#define vsuxseg7ei32_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei32_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint32m1_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (size_t)(op10)) +#define vsuxseg7ei32_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei32_v_i8mf8((int8_t *)(op0), (vuint32mf2_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (size_t)(op9)) +#define vsuxseg7ei32_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei32_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint32mf2_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (size_t)(op10)) +#define vsuxseg7ei32_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei32_v_u8m1((uint8_t *)(op0), (vuint32m4_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei32_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei32_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint32m4_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (size_t)(op10)) +#define vsuxseg7ei32_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei32_v_u8mf2((uint8_t *)(op0), (vuint32m2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vsuxseg7ei32_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei32_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint32m2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (size_t)(op10)) +#define vsuxseg7ei32_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei32_v_u8mf4((uint8_t *)(op0), (vuint32m1_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vsuxseg7ei32_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei32_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint32m1_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (size_t)(op10)) +#define vsuxseg7ei32_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei32_v_u8mf8((uint8_t *)(op0), (vuint32mf2_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vsuxseg7ei32_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei32_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint32mf2_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (size_t)(op10)) +#define vsuxseg8ei32_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei32_v_i8m1((int8_t *)(op0), (vuint32m4_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei32_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei32_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint32m4_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (size_t)(op11)) +#define vsuxseg8ei32_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei32_v_i8mf2((int8_t *)(op0), (vuint32m2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (size_t)(op10)) +#define vsuxseg8ei32_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei32_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint32m2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (size_t)(op11)) +#define vsuxseg8ei32_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei32_v_i8mf4((int8_t *)(op0), (vuint32m1_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (size_t)(op10)) +#define vsuxseg8ei32_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei32_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint32m1_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (size_t)(op11)) +#define vsuxseg8ei32_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei32_v_i8mf8((int8_t *)(op0), (vuint32mf2_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (size_t)(op10)) +#define vsuxseg8ei32_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei32_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint32mf2_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (size_t)(op11)) +#define vsuxseg8ei32_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei32_v_u8m1((uint8_t *)(op0), (vuint32m4_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei32_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei32_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint32m4_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (size_t)(op11)) +#define vsuxseg8ei32_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei32_v_u8mf2((uint8_t *)(op0), (vuint32m2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (size_t)(op10)) +#define vsuxseg8ei32_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei32_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint32m2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (size_t)(op11)) +#define vsuxseg8ei32_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei32_v_u8mf4((uint8_t *)(op0), (vuint32m1_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (size_t)(op10)) +#define vsuxseg8ei32_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei32_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint32m1_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (size_t)(op11)) +#define vsuxseg8ei32_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei32_v_u8mf8((uint8_t *)(op0), (vuint32mf2_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (size_t)(op10)) +#define vsuxseg8ei32_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei32_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint32mf2_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (size_t)(op11)) +#define vsuxseg2ei64_v_i8m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei64_v_i8m1((int8_t *)(op0), (vuint64m8_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vsuxseg2ei64_v_i8m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei64_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint64m8_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (size_t)(op5)) +#define vsuxseg2ei64_v_i8mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei64_v_i8mf2((int8_t *)(op0), (vuint64m4_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei64_v_i8mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei64_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint64m4_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei64_v_i8mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei64_v_i8mf4((int8_t *)(op0), (vuint64m2_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei64_v_i8mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei64_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint64m2_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (size_t)(op5)) +#define vsuxseg2ei64_v_i8mf8(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei64_v_i8mf8((int8_t *)(op0), (vuint64m1_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vsuxseg2ei64_v_i8mf8_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei64_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint64m1_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (size_t)(op5)) +#define vsuxseg2ei64_v_u8m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei64_v_u8m1((uint8_t *)(op0), (vuint64m8_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vsuxseg2ei64_v_u8m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei64_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint64m8_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (size_t)(op5)) +#define vsuxseg2ei64_v_u8mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei64_v_u8mf2((uint8_t *)(op0), (vuint64m4_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei64_v_u8mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei64_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint64m4_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei64_v_u8mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei64_v_u8mf4((uint8_t *)(op0), (vuint64m2_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei64_v_u8mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei64_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint64m2_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (size_t)(op5)) +#define vsuxseg2ei64_v_u8mf8(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei64_v_u8mf8((uint8_t *)(op0), (vuint64m1_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vsuxseg2ei64_v_u8mf8_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei64_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint64m1_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (size_t)(op5)) +#define vsuxseg3ei64_v_i8m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei64_v_i8m1((int8_t *)(op0), (vuint64m8_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (size_t)(op5)) +#define vsuxseg3ei64_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei64_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint64m8_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (size_t)(op6)) +#define vsuxseg3ei64_v_i8mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei64_v_i8mf2((int8_t *)(op0), (vuint64m4_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei64_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei64_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint64m4_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei64_v_i8mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei64_v_i8mf4((int8_t *)(op0), (vuint64m2_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (size_t)(op5)) +#define vsuxseg3ei64_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei64_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint64m2_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (size_t)(op6)) +#define vsuxseg3ei64_v_i8mf8(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei64_v_i8mf8((int8_t *)(op0), (vuint64m1_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (size_t)(op5)) +#define vsuxseg3ei64_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei64_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint64m1_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (size_t)(op6)) +#define vsuxseg3ei64_v_u8m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei64_v_u8m1((uint8_t *)(op0), (vuint64m8_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (size_t)(op5)) +#define vsuxseg3ei64_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei64_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint64m8_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (size_t)(op6)) +#define vsuxseg3ei64_v_u8mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei64_v_u8mf2((uint8_t *)(op0), (vuint64m4_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei64_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei64_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint64m4_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei64_v_u8mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei64_v_u8mf4((uint8_t *)(op0), (vuint64m2_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (size_t)(op5)) +#define vsuxseg3ei64_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei64_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint64m2_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (size_t)(op6)) +#define vsuxseg3ei64_v_u8mf8(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei64_v_u8mf8((uint8_t *)(op0), (vuint64m1_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (size_t)(op5)) +#define vsuxseg3ei64_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei64_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint64m1_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (size_t)(op6)) +#define vsuxseg4ei64_v_i8m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei64_v_i8m1((int8_t *)(op0), (vuint64m8_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (size_t)(op6)) +#define vsuxseg4ei64_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei64_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint64m8_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (size_t)(op7)) +#define vsuxseg4ei64_v_i8mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei64_v_i8mf2((int8_t *)(op0), (vuint64m4_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei64_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei64_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint64m4_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei64_v_i8mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei64_v_i8mf4((int8_t *)(op0), (vuint64m2_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (size_t)(op6)) +#define vsuxseg4ei64_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei64_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint64m2_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (size_t)(op7)) +#define vsuxseg4ei64_v_i8mf8(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei64_v_i8mf8((int8_t *)(op0), (vuint64m1_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (size_t)(op6)) +#define vsuxseg4ei64_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei64_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint64m1_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (size_t)(op7)) +#define vsuxseg4ei64_v_u8m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei64_v_u8m1((uint8_t *)(op0), (vuint64m8_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (size_t)(op6)) +#define vsuxseg4ei64_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei64_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint64m8_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (size_t)(op7)) +#define vsuxseg4ei64_v_u8mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei64_v_u8mf2((uint8_t *)(op0), (vuint64m4_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei64_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei64_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint64m4_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei64_v_u8mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei64_v_u8mf4((uint8_t *)(op0), (vuint64m2_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (size_t)(op6)) +#define vsuxseg4ei64_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei64_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint64m2_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vsuxseg4ei64_v_u8mf8(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei64_v_u8mf8((uint8_t *)(op0), (vuint64m1_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (size_t)(op6)) +#define vsuxseg4ei64_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei64_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint64m1_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vsuxseg5ei64_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei64_v_i8m1((int8_t *)(op0), (vuint64m8_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (size_t)(op7)) +#define vsuxseg5ei64_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei64_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint64m8_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (size_t)(op8)) +#define vsuxseg5ei64_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei64_v_i8mf2((int8_t *)(op0), (vuint64m4_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei64_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei64_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint64m4_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (size_t)(op8)) +#define vsuxseg5ei64_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei64_v_i8mf4((int8_t *)(op0), (vuint64m2_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (size_t)(op7)) +#define vsuxseg5ei64_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei64_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint64m2_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (size_t)(op8)) +#define vsuxseg5ei64_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei64_v_i8mf8((int8_t *)(op0), (vuint64m1_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (size_t)(op7)) +#define vsuxseg5ei64_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei64_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint64m1_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (size_t)(op8)) +#define vsuxseg5ei64_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei64_v_u8m1((uint8_t *)(op0), (vuint64m8_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (size_t)(op7)) +#define vsuxseg5ei64_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei64_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint64m8_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (size_t)(op8)) +#define vsuxseg5ei64_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei64_v_u8mf2((uint8_t *)(op0), (vuint64m4_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei64_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei64_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint64m4_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (size_t)(op8)) +#define vsuxseg5ei64_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei64_v_u8mf4((uint8_t *)(op0), (vuint64m2_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vsuxseg5ei64_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei64_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint64m2_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (size_t)(op8)) +#define vsuxseg5ei64_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei64_v_u8mf8((uint8_t *)(op0), (vuint64m1_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vsuxseg5ei64_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei64_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint64m1_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (size_t)(op8)) +#define vsuxseg6ei64_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei64_v_i8m1((int8_t *)(op0), (vuint64m8_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei64_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei64_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint64m8_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (size_t)(op9)) +#define vsuxseg6ei64_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei64_v_i8mf2((int8_t *)(op0), (vuint64m4_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (size_t)(op8)) +#define vsuxseg6ei64_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei64_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint64m4_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (size_t)(op9)) +#define vsuxseg6ei64_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei64_v_i8mf4((int8_t *)(op0), (vuint64m2_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (size_t)(op8)) +#define vsuxseg6ei64_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei64_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint64m2_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (size_t)(op9)) +#define vsuxseg6ei64_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei64_v_i8mf8((int8_t *)(op0), (vuint64m1_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (size_t)(op8)) +#define vsuxseg6ei64_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei64_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint64m1_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (size_t)(op9)) +#define vsuxseg6ei64_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei64_v_u8m1((uint8_t *)(op0), (vuint64m8_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei64_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei64_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint64m8_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (size_t)(op9)) +#define vsuxseg6ei64_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei64_v_u8mf2((uint8_t *)(op0), (vuint64m4_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (size_t)(op8)) +#define vsuxseg6ei64_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei64_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint64m4_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vsuxseg6ei64_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei64_v_u8mf4((uint8_t *)(op0), (vuint64m2_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (size_t)(op8)) +#define vsuxseg6ei64_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei64_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint64m2_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vsuxseg6ei64_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei64_v_u8mf8((uint8_t *)(op0), (vuint64m1_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (size_t)(op8)) +#define vsuxseg6ei64_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei64_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint64m1_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vsuxseg7ei64_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei64_v_i8m1((int8_t *)(op0), (vuint64m8_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei64_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei64_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint64m8_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (size_t)(op10)) +#define vsuxseg7ei64_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei64_v_i8mf2((int8_t *)(op0), (vuint64m4_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (size_t)(op9)) +#define vsuxseg7ei64_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei64_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint64m4_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (size_t)(op10)) +#define vsuxseg7ei64_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei64_v_i8mf4((int8_t *)(op0), (vuint64m2_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (size_t)(op9)) +#define vsuxseg7ei64_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei64_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint64m2_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (size_t)(op10)) +#define vsuxseg7ei64_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei64_v_i8mf8((int8_t *)(op0), (vuint64m1_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (size_t)(op9)) +#define vsuxseg7ei64_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei64_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint64m1_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (size_t)(op10)) +#define vsuxseg7ei64_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei64_v_u8m1((uint8_t *)(op0), (vuint64m8_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei64_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei64_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint64m8_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (size_t)(op10)) +#define vsuxseg7ei64_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei64_v_u8mf2((uint8_t *)(op0), (vuint64m4_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vsuxseg7ei64_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei64_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint64m4_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (size_t)(op10)) +#define vsuxseg7ei64_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei64_v_u8mf4((uint8_t *)(op0), (vuint64m2_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vsuxseg7ei64_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei64_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint64m2_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (size_t)(op10)) +#define vsuxseg7ei64_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei64_v_u8mf8((uint8_t *)(op0), (vuint64m1_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vsuxseg7ei64_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei64_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint64m1_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (size_t)(op10)) +#define vsuxseg8ei64_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei64_v_i8m1((int8_t *)(op0), (vuint64m8_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei64_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei64_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint64m8_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (size_t)(op11)) +#define vsuxseg8ei64_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei64_v_i8mf2((int8_t *)(op0), (vuint64m4_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (size_t)(op10)) +#define vsuxseg8ei64_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei64_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint64m4_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (size_t)(op11)) +#define vsuxseg8ei64_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei64_v_i8mf4((int8_t *)(op0), (vuint64m2_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (size_t)(op10)) +#define vsuxseg8ei64_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei64_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint64m2_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (size_t)(op11)) +#define vsuxseg8ei64_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei64_v_i8mf8((int8_t *)(op0), (vuint64m1_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (size_t)(op10)) +#define vsuxseg8ei64_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei64_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint64m1_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (size_t)(op11)) +#define vsuxseg8ei64_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei64_v_u8m1((uint8_t *)(op0), (vuint64m8_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei64_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei64_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint64m8_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (size_t)(op11)) +#define vsuxseg8ei64_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei64_v_u8mf2((uint8_t *)(op0), (vuint64m4_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (size_t)(op10)) +#define vsuxseg8ei64_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei64_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint64m4_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (size_t)(op11)) +#define vsuxseg8ei64_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei64_v_u8mf4((uint8_t *)(op0), (vuint64m2_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (size_t)(op10)) +#define vsuxseg8ei64_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei64_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint64m2_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (size_t)(op11)) +#define vsuxseg8ei64_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei64_v_u8mf8((uint8_t *)(op0), (vuint64m1_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (size_t)(op10)) +#define vsuxseg8ei64_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei64_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint64m1_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (size_t)(op11)) +#define vsuxseg2ei8_v_i16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_i16m1((int16_t *)(op0), (vuint8mf2_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_i16m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint8mf2_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (size_t)(op5)) +#define vsuxseg2ei8_v_i16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_i16m2((int16_t *)(op0), (vuint8m1_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_i16m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_i16m2_m((vbool8_t)(op0), (int16_t *)(op1), (vuint8m1_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei8_v_i16m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_i16m4((int16_t *)(op0), (vuint8m2_t)(op1), (vint16m4_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_i16m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_i16m4_m((vbool4_t)(op0), (int16_t *)(op1), (vuint8m2_t)(op2), (vint16m4_t)(op3), (vint16m4_t)(op4), (size_t)(op5)) +#define vsuxseg2ei8_v_i16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_i16mf2((int16_t *)(op0), (vuint8mf4_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_i16mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint8mf4_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei8_v_i16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_i16mf4((int16_t *)(op0), (vuint8mf8_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_i16mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint8mf8_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (size_t)(op5)) +#define vsuxseg2ei8_v_u16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_u16m1((uint16_t *)(op0), (vuint8mf2_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_u16m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint8mf2_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (size_t)(op5)) +#define vsuxseg2ei8_v_u16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_u16m2((uint16_t *)(op0), (vuint8m1_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_u16m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_u16m2_m((vbool8_t)(op0), (uint16_t *)(op1), (vuint8m1_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei8_v_u16m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_u16m4((uint16_t *)(op0), (vuint8m2_t)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_u16m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_u16m4_m((vbool4_t)(op0), (uint16_t *)(op1), (vuint8m2_t)(op2), (vuint16m4_t)(op3), (vuint16m4_t)(op4), (size_t)(op5)) +#define vsuxseg2ei8_v_u16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_u16mf2((uint16_t *)(op0), (vuint8mf4_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_u16mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint8mf4_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei8_v_u16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_u16mf4((uint16_t *)(op0), (vuint8mf8_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_u16mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint8mf8_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (size_t)(op5)) +#define vsuxseg3ei8_v_i16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei8_v_i16m1((int16_t *)(op0), (vuint8mf2_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (size_t)(op5)) +#define vsuxseg3ei8_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei8_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint8mf2_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (size_t)(op6)) +#define vsuxseg3ei8_v_i16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei8_v_i16m2((int16_t *)(op0), (vuint8m1_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei8_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei8_v_i16m2_m((vbool8_t)(op0), (int16_t *)(op1), (vuint8m1_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (vint16m2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei8_v_i16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei8_v_i16mf2((int16_t *)(op0), (vuint8mf4_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei8_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei8_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint8mf4_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei8_v_i16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei8_v_i16mf4((int16_t *)(op0), (vuint8mf8_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (size_t)(op5)) +#define vsuxseg3ei8_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei8_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint8mf8_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (size_t)(op6)) +#define vsuxseg3ei8_v_u16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei8_v_u16m1((uint16_t *)(op0), (vuint8mf2_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (size_t)(op5)) +#define vsuxseg3ei8_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei8_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint8mf2_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (size_t)(op6)) +#define vsuxseg3ei8_v_u16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei8_v_u16m2((uint16_t *)(op0), (vuint8m1_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei8_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei8_v_u16m2_m((vbool8_t)(op0), (uint16_t *)(op1), (vuint8m1_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (vuint16m2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei8_v_u16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei8_v_u16mf2((uint16_t *)(op0), (vuint8mf4_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei8_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei8_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint8mf4_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei8_v_u16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei8_v_u16mf4((uint16_t *)(op0), (vuint8mf8_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (size_t)(op5)) +#define vsuxseg3ei8_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei8_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint8mf8_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (size_t)(op6)) +#define vsuxseg4ei8_v_i16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei8_v_i16m1((int16_t *)(op0), (vuint8mf2_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (size_t)(op6)) +#define vsuxseg4ei8_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei8_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint8mf2_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (size_t)(op7)) +#define vsuxseg4ei8_v_i16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei8_v_i16m2((int16_t *)(op0), (vuint8m1_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (vint16m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei8_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei8_v_i16m2_m((vbool8_t)(op0), (int16_t *)(op1), (vuint8m1_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (vint16m2_t)(op5), (vint16m2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei8_v_i16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei8_v_i16mf2((int16_t *)(op0), (vuint8mf4_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei8_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei8_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint8mf4_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei8_v_i16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei8_v_i16mf4((int16_t *)(op0), (vuint8mf8_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (size_t)(op6)) +#define vsuxseg4ei8_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei8_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint8mf8_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (size_t)(op7)) +#define vsuxseg4ei8_v_u16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei8_v_u16m1((uint16_t *)(op0), (vuint8mf2_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (size_t)(op6)) +#define vsuxseg4ei8_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei8_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint8mf2_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vsuxseg4ei8_v_u16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei8_v_u16m2((uint16_t *)(op0), (vuint8m1_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (vuint16m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei8_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei8_v_u16m2_m((vbool8_t)(op0), (uint16_t *)(op1), (vuint8m1_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (vuint16m2_t)(op5), (vuint16m2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei8_v_u16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei8_v_u16mf2((uint16_t *)(op0), (vuint8mf4_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei8_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei8_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint8mf4_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei8_v_u16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei8_v_u16mf4((uint16_t *)(op0), (vuint8mf8_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (size_t)(op6)) +#define vsuxseg4ei8_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei8_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint8mf8_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vsuxseg5ei8_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei8_v_i16m1((int16_t *)(op0), (vuint8mf2_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (size_t)(op7)) +#define vsuxseg5ei8_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei8_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint8mf2_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (size_t)(op8)) +#define vsuxseg5ei8_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei8_v_i16mf2((int16_t *)(op0), (vuint8mf4_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei8_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei8_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint8mf4_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (size_t)(op8)) +#define vsuxseg5ei8_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei8_v_i16mf4((int16_t *)(op0), (vuint8mf8_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (size_t)(op7)) +#define vsuxseg5ei8_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei8_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint8mf8_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (size_t)(op8)) +#define vsuxseg5ei8_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei8_v_u16m1((uint16_t *)(op0), (vuint8mf2_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vsuxseg5ei8_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei8_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint8mf2_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (size_t)(op8)) +#define vsuxseg5ei8_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei8_v_u16mf2((uint16_t *)(op0), (vuint8mf4_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei8_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei8_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint8mf4_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (size_t)(op8)) +#define vsuxseg5ei8_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei8_v_u16mf4((uint16_t *)(op0), (vuint8mf8_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vsuxseg5ei8_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei8_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint8mf8_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (size_t)(op8)) +#define vsuxseg6ei8_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei8_v_i16m1((int16_t *)(op0), (vuint8mf2_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei8_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei8_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint8mf2_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (size_t)(op9)) +#define vsuxseg6ei8_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei8_v_i16mf2((int16_t *)(op0), (vuint8mf4_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (size_t)(op8)) +#define vsuxseg6ei8_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei8_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint8mf4_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (size_t)(op9)) +#define vsuxseg6ei8_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei8_v_i16mf4((int16_t *)(op0), (vuint8mf8_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (size_t)(op8)) +#define vsuxseg6ei8_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei8_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint8mf8_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (size_t)(op9)) +#define vsuxseg6ei8_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei8_v_u16m1((uint16_t *)(op0), (vuint8mf2_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei8_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei8_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint8mf2_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vsuxseg6ei8_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei8_v_u16mf2((uint16_t *)(op0), (vuint8mf4_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (size_t)(op8)) +#define vsuxseg6ei8_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei8_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint8mf4_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vsuxseg6ei8_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei8_v_u16mf4((uint16_t *)(op0), (vuint8mf8_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (size_t)(op8)) +#define vsuxseg6ei8_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei8_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint8mf8_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vsuxseg7ei8_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei8_v_i16m1((int16_t *)(op0), (vuint8mf2_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei8_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei8_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint8mf2_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (size_t)(op10)) +#define vsuxseg7ei8_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei8_v_i16mf2((int16_t *)(op0), (vuint8mf4_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (size_t)(op9)) +#define vsuxseg7ei8_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei8_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint8mf4_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (size_t)(op10)) +#define vsuxseg7ei8_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei8_v_i16mf4((int16_t *)(op0), (vuint8mf8_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (size_t)(op9)) +#define vsuxseg7ei8_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei8_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint8mf8_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (size_t)(op10)) +#define vsuxseg7ei8_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei8_v_u16m1((uint16_t *)(op0), (vuint8mf2_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei8_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei8_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint8mf2_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (size_t)(op10)) +#define vsuxseg7ei8_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei8_v_u16mf2((uint16_t *)(op0), (vuint8mf4_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vsuxseg7ei8_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei8_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint8mf4_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (size_t)(op10)) +#define vsuxseg7ei8_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei8_v_u16mf4((uint16_t *)(op0), (vuint8mf8_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vsuxseg7ei8_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei8_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint8mf8_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (size_t)(op10)) +#define vsuxseg8ei8_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei8_v_i16m1((int16_t *)(op0), (vuint8mf2_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei8_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei8_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint8mf2_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (size_t)(op11)) +#define vsuxseg8ei8_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei8_v_i16mf2((int16_t *)(op0), (vuint8mf4_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (size_t)(op10)) +#define vsuxseg8ei8_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei8_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint8mf4_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (size_t)(op11)) +#define vsuxseg8ei8_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei8_v_i16mf4((int16_t *)(op0), (vuint8mf8_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (size_t)(op10)) +#define vsuxseg8ei8_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei8_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint8mf8_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (size_t)(op11)) +#define vsuxseg8ei8_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei8_v_u16m1((uint16_t *)(op0), (vuint8mf2_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei8_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei8_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint8mf2_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (size_t)(op11)) +#define vsuxseg8ei8_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei8_v_u16mf2((uint16_t *)(op0), (vuint8mf4_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (size_t)(op10)) +#define vsuxseg8ei8_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei8_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint8mf4_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (size_t)(op11)) +#define vsuxseg8ei8_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei8_v_u16mf4((uint16_t *)(op0), (vuint8mf8_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (size_t)(op10)) +#define vsuxseg8ei8_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei8_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint8mf8_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (size_t)(op11)) +#define vsuxseg2ei16_v_i16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_i16m1((int16_t *)(op0), (vuint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_i16m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (size_t)(op5)) +#define vsuxseg2ei16_v_i16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_i16m2((int16_t *)(op0), (vuint16m2_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_i16m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_i16m2_m((vbool8_t)(op0), (int16_t *)(op1), (vuint16m2_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei16_v_i16m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_i16m4((int16_t *)(op0), (vuint16m4_t)(op1), (vint16m4_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_i16m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_i16m4_m((vbool4_t)(op0), (int16_t *)(op1), (vuint16m4_t)(op2), (vint16m4_t)(op3), (vint16m4_t)(op4), (size_t)(op5)) +#define vsuxseg2ei16_v_i16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_i16mf2((int16_t *)(op0), (vuint16mf2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_i16mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei16_v_i16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_i16mf4((int16_t *)(op0), (vuint16mf4_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_i16mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (size_t)(op5)) +#define vsuxseg2ei16_v_u16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_u16m1((uint16_t *)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_u16m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (size_t)(op5)) +#define vsuxseg2ei16_v_u16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_u16m2((uint16_t *)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_u16m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_u16m2_m((vbool8_t)(op0), (uint16_t *)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei16_v_u16m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_u16m4((uint16_t *)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_u16m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_u16m4_m((vbool4_t)(op0), (uint16_t *)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (vuint16m4_t)(op4), (size_t)(op5)) +#define vsuxseg2ei16_v_u16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_u16mf2((uint16_t *)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_u16mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei16_v_u16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_u16mf4((uint16_t *)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_u16mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (size_t)(op5)) +#define vsuxseg3ei16_v_i16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei16_v_i16m1((int16_t *)(op0), (vuint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (size_t)(op5)) +#define vsuxseg3ei16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei16_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (size_t)(op6)) +#define vsuxseg3ei16_v_i16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei16_v_i16m2((int16_t *)(op0), (vuint16m2_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei16_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei16_v_i16m2_m((vbool8_t)(op0), (int16_t *)(op1), (vuint16m2_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (vint16m2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei16_v_i16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei16_v_i16mf2((int16_t *)(op0), (vuint16mf2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei16_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei16_v_i16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei16_v_i16mf4((int16_t *)(op0), (vuint16mf4_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (size_t)(op5)) +#define vsuxseg3ei16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei16_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (size_t)(op6)) +#define vsuxseg3ei16_v_u16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei16_v_u16m1((uint16_t *)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (size_t)(op5)) +#define vsuxseg3ei16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei16_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (size_t)(op6)) +#define vsuxseg3ei16_v_u16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei16_v_u16m2((uint16_t *)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei16_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei16_v_u16m2_m((vbool8_t)(op0), (uint16_t *)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (vuint16m2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei16_v_u16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei16_v_u16mf2((uint16_t *)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei16_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei16_v_u16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei16_v_u16mf4((uint16_t *)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (size_t)(op5)) +#define vsuxseg3ei16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei16_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (size_t)(op6)) +#define vsuxseg4ei16_v_i16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei16_v_i16m1((int16_t *)(op0), (vuint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (size_t)(op6)) +#define vsuxseg4ei16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei16_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (size_t)(op7)) +#define vsuxseg4ei16_v_i16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei16_v_i16m2((int16_t *)(op0), (vuint16m2_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (vint16m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei16_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei16_v_i16m2_m((vbool8_t)(op0), (int16_t *)(op1), (vuint16m2_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (vint16m2_t)(op5), (vint16m2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei16_v_i16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei16_v_i16mf2((int16_t *)(op0), (vuint16mf2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei16_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei16_v_i16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei16_v_i16mf4((int16_t *)(op0), (vuint16mf4_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (size_t)(op6)) +#define vsuxseg4ei16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei16_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (size_t)(op7)) +#define vsuxseg4ei16_v_u16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei16_v_u16m1((uint16_t *)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (size_t)(op6)) +#define vsuxseg4ei16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei16_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vsuxseg4ei16_v_u16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei16_v_u16m2((uint16_t *)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (vuint16m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei16_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei16_v_u16m2_m((vbool8_t)(op0), (uint16_t *)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (vuint16m2_t)(op5), (vuint16m2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei16_v_u16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei16_v_u16mf2((uint16_t *)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei16_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei16_v_u16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei16_v_u16mf4((uint16_t *)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (size_t)(op6)) +#define vsuxseg4ei16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei16_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vsuxseg5ei16_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei16_v_i16m1((int16_t *)(op0), (vuint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (size_t)(op7)) +#define vsuxseg5ei16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei16_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (size_t)(op8)) +#define vsuxseg5ei16_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei16_v_i16mf2((int16_t *)(op0), (vuint16mf2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei16_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (size_t)(op8)) +#define vsuxseg5ei16_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei16_v_i16mf4((int16_t *)(op0), (vuint16mf4_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (size_t)(op7)) +#define vsuxseg5ei16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei16_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (size_t)(op8)) +#define vsuxseg5ei16_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei16_v_u16m1((uint16_t *)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vsuxseg5ei16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei16_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (size_t)(op8)) +#define vsuxseg5ei16_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei16_v_u16mf2((uint16_t *)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei16_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (size_t)(op8)) +#define vsuxseg5ei16_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei16_v_u16mf4((uint16_t *)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vsuxseg5ei16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei16_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (size_t)(op8)) +#define vsuxseg6ei16_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei16_v_i16m1((int16_t *)(op0), (vuint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei16_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (size_t)(op9)) +#define vsuxseg6ei16_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei16_v_i16mf2((int16_t *)(op0), (vuint16mf2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (size_t)(op8)) +#define vsuxseg6ei16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei16_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (size_t)(op9)) +#define vsuxseg6ei16_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei16_v_i16mf4((int16_t *)(op0), (vuint16mf4_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (size_t)(op8)) +#define vsuxseg6ei16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei16_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (size_t)(op9)) +#define vsuxseg6ei16_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei16_v_u16m1((uint16_t *)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei16_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vsuxseg6ei16_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei16_v_u16mf2((uint16_t *)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (size_t)(op8)) +#define vsuxseg6ei16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei16_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vsuxseg6ei16_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei16_v_u16mf4((uint16_t *)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (size_t)(op8)) +#define vsuxseg6ei16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei16_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vsuxseg7ei16_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei16_v_i16m1((int16_t *)(op0), (vuint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei16_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (size_t)(op10)) +#define vsuxseg7ei16_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei16_v_i16mf2((int16_t *)(op0), (vuint16mf2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (size_t)(op9)) +#define vsuxseg7ei16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei16_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (size_t)(op10)) +#define vsuxseg7ei16_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei16_v_i16mf4((int16_t *)(op0), (vuint16mf4_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (size_t)(op9)) +#define vsuxseg7ei16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei16_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (size_t)(op10)) +#define vsuxseg7ei16_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei16_v_u16m1((uint16_t *)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei16_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (size_t)(op10)) +#define vsuxseg7ei16_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei16_v_u16mf2((uint16_t *)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vsuxseg7ei16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei16_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (size_t)(op10)) +#define vsuxseg7ei16_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei16_v_u16mf4((uint16_t *)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vsuxseg7ei16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei16_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (size_t)(op10)) +#define vsuxseg8ei16_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei16_v_i16m1((int16_t *)(op0), (vuint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei16_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (size_t)(op11)) +#define vsuxseg8ei16_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei16_v_i16mf2((int16_t *)(op0), (vuint16mf2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (size_t)(op10)) +#define vsuxseg8ei16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei16_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (size_t)(op11)) +#define vsuxseg8ei16_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei16_v_i16mf4((int16_t *)(op0), (vuint16mf4_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (size_t)(op10)) +#define vsuxseg8ei16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei16_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (size_t)(op11)) +#define vsuxseg8ei16_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei16_v_u16m1((uint16_t *)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei16_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (size_t)(op11)) +#define vsuxseg8ei16_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei16_v_u16mf2((uint16_t *)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (size_t)(op10)) +#define vsuxseg8ei16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei16_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (size_t)(op11)) +#define vsuxseg8ei16_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei16_v_u16mf4((uint16_t *)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (size_t)(op10)) +#define vsuxseg8ei16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei16_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (size_t)(op11)) +#define vsuxseg2ei32_v_i16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_i16m1((int16_t *)(op0), (vuint32m2_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_i16m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint32m2_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (size_t)(op5)) +#define vsuxseg2ei32_v_i16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_i16m2((int16_t *)(op0), (vuint32m4_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_i16m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_i16m2_m((vbool8_t)(op0), (int16_t *)(op1), (vuint32m4_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei32_v_i16m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_i16m4((int16_t *)(op0), (vuint32m8_t)(op1), (vint16m4_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_i16m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_i16m4_m((vbool4_t)(op0), (int16_t *)(op1), (vuint32m8_t)(op2), (vint16m4_t)(op3), (vint16m4_t)(op4), (size_t)(op5)) +#define vsuxseg2ei32_v_i16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_i16mf2((int16_t *)(op0), (vuint32m1_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_i16mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint32m1_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei32_v_i16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_i16mf4((int16_t *)(op0), (vuint32mf2_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_i16mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint32mf2_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (size_t)(op5)) +#define vsuxseg2ei32_v_u16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_u16m1((uint16_t *)(op0), (vuint32m2_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_u16m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint32m2_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (size_t)(op5)) +#define vsuxseg2ei32_v_u16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_u16m2((uint16_t *)(op0), (vuint32m4_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_u16m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_u16m2_m((vbool8_t)(op0), (uint16_t *)(op1), (vuint32m4_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei32_v_u16m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_u16m4((uint16_t *)(op0), (vuint32m8_t)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_u16m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_u16m4_m((vbool4_t)(op0), (uint16_t *)(op1), (vuint32m8_t)(op2), (vuint16m4_t)(op3), (vuint16m4_t)(op4), (size_t)(op5)) +#define vsuxseg2ei32_v_u16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_u16mf2((uint16_t *)(op0), (vuint32m1_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_u16mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint32m1_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei32_v_u16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_u16mf4((uint16_t *)(op0), (vuint32mf2_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_u16mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint32mf2_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (size_t)(op5)) +#define vsuxseg3ei32_v_i16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei32_v_i16m1((int16_t *)(op0), (vuint32m2_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (size_t)(op5)) +#define vsuxseg3ei32_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei32_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint32m2_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (size_t)(op6)) +#define vsuxseg3ei32_v_i16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei32_v_i16m2((int16_t *)(op0), (vuint32m4_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei32_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei32_v_i16m2_m((vbool8_t)(op0), (int16_t *)(op1), (vuint32m4_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (vint16m2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei32_v_i16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei32_v_i16mf2((int16_t *)(op0), (vuint32m1_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei32_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei32_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint32m1_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei32_v_i16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei32_v_i16mf4((int16_t *)(op0), (vuint32mf2_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (size_t)(op5)) +#define vsuxseg3ei32_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei32_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint32mf2_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (size_t)(op6)) +#define vsuxseg3ei32_v_u16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei32_v_u16m1((uint16_t *)(op0), (vuint32m2_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (size_t)(op5)) +#define vsuxseg3ei32_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei32_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint32m2_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (size_t)(op6)) +#define vsuxseg3ei32_v_u16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei32_v_u16m2((uint16_t *)(op0), (vuint32m4_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei32_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei32_v_u16m2_m((vbool8_t)(op0), (uint16_t *)(op1), (vuint32m4_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (vuint16m2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei32_v_u16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei32_v_u16mf2((uint16_t *)(op0), (vuint32m1_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei32_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei32_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint32m1_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei32_v_u16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei32_v_u16mf4((uint16_t *)(op0), (vuint32mf2_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (size_t)(op5)) +#define vsuxseg3ei32_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei32_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint32mf2_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (size_t)(op6)) +#define vsuxseg4ei32_v_i16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei32_v_i16m1((int16_t *)(op0), (vuint32m2_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (size_t)(op6)) +#define vsuxseg4ei32_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei32_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint32m2_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (size_t)(op7)) +#define vsuxseg4ei32_v_i16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei32_v_i16m2((int16_t *)(op0), (vuint32m4_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (vint16m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei32_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei32_v_i16m2_m((vbool8_t)(op0), (int16_t *)(op1), (vuint32m4_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (vint16m2_t)(op5), (vint16m2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei32_v_i16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei32_v_i16mf2((int16_t *)(op0), (vuint32m1_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei32_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei32_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint32m1_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei32_v_i16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei32_v_i16mf4((int16_t *)(op0), (vuint32mf2_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (size_t)(op6)) +#define vsuxseg4ei32_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei32_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint32mf2_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (size_t)(op7)) +#define vsuxseg4ei32_v_u16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei32_v_u16m1((uint16_t *)(op0), (vuint32m2_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (size_t)(op6)) +#define vsuxseg4ei32_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei32_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint32m2_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vsuxseg4ei32_v_u16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei32_v_u16m2((uint16_t *)(op0), (vuint32m4_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (vuint16m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei32_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei32_v_u16m2_m((vbool8_t)(op0), (uint16_t *)(op1), (vuint32m4_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (vuint16m2_t)(op5), (vuint16m2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei32_v_u16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei32_v_u16mf2((uint16_t *)(op0), (vuint32m1_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei32_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei32_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint32m1_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei32_v_u16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei32_v_u16mf4((uint16_t *)(op0), (vuint32mf2_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (size_t)(op6)) +#define vsuxseg4ei32_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei32_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint32mf2_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vsuxseg5ei32_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei32_v_i16m1((int16_t *)(op0), (vuint32m2_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (size_t)(op7)) +#define vsuxseg5ei32_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei32_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint32m2_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (size_t)(op8)) +#define vsuxseg5ei32_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei32_v_i16mf2((int16_t *)(op0), (vuint32m1_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei32_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei32_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint32m1_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (size_t)(op8)) +#define vsuxseg5ei32_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei32_v_i16mf4((int16_t *)(op0), (vuint32mf2_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (size_t)(op7)) +#define vsuxseg5ei32_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei32_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint32mf2_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (size_t)(op8)) +#define vsuxseg5ei32_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei32_v_u16m1((uint16_t *)(op0), (vuint32m2_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vsuxseg5ei32_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei32_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint32m2_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (size_t)(op8)) +#define vsuxseg5ei32_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei32_v_u16mf2((uint16_t *)(op0), (vuint32m1_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei32_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei32_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint32m1_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (size_t)(op8)) +#define vsuxseg5ei32_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei32_v_u16mf4((uint16_t *)(op0), (vuint32mf2_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vsuxseg5ei32_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei32_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint32mf2_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (size_t)(op8)) +#define vsuxseg6ei32_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei32_v_i16m1((int16_t *)(op0), (vuint32m2_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei32_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei32_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint32m2_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (size_t)(op9)) +#define vsuxseg6ei32_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei32_v_i16mf2((int16_t *)(op0), (vuint32m1_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (size_t)(op8)) +#define vsuxseg6ei32_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei32_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint32m1_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (size_t)(op9)) +#define vsuxseg6ei32_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei32_v_i16mf4((int16_t *)(op0), (vuint32mf2_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (size_t)(op8)) +#define vsuxseg6ei32_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei32_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint32mf2_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (size_t)(op9)) +#define vsuxseg6ei32_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei32_v_u16m1((uint16_t *)(op0), (vuint32m2_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei32_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei32_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint32m2_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vsuxseg6ei32_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei32_v_u16mf2((uint16_t *)(op0), (vuint32m1_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (size_t)(op8)) +#define vsuxseg6ei32_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei32_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint32m1_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vsuxseg6ei32_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei32_v_u16mf4((uint16_t *)(op0), (vuint32mf2_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (size_t)(op8)) +#define vsuxseg6ei32_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei32_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint32mf2_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vsuxseg7ei32_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei32_v_i16m1((int16_t *)(op0), (vuint32m2_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei32_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei32_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint32m2_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (size_t)(op10)) +#define vsuxseg7ei32_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei32_v_i16mf2((int16_t *)(op0), (vuint32m1_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (size_t)(op9)) +#define vsuxseg7ei32_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei32_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint32m1_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (size_t)(op10)) +#define vsuxseg7ei32_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei32_v_i16mf4((int16_t *)(op0), (vuint32mf2_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (size_t)(op9)) +#define vsuxseg7ei32_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei32_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint32mf2_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (size_t)(op10)) +#define vsuxseg7ei32_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei32_v_u16m1((uint16_t *)(op0), (vuint32m2_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei32_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei32_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint32m2_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (size_t)(op10)) +#define vsuxseg7ei32_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei32_v_u16mf2((uint16_t *)(op0), (vuint32m1_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vsuxseg7ei32_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei32_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint32m1_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (size_t)(op10)) +#define vsuxseg7ei32_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei32_v_u16mf4((uint16_t *)(op0), (vuint32mf2_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vsuxseg7ei32_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei32_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint32mf2_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (size_t)(op10)) +#define vsuxseg8ei32_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei32_v_i16m1((int16_t *)(op0), (vuint32m2_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei32_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei32_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint32m2_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (size_t)(op11)) +#define vsuxseg8ei32_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei32_v_i16mf2((int16_t *)(op0), (vuint32m1_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (size_t)(op10)) +#define vsuxseg8ei32_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei32_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint32m1_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (size_t)(op11)) +#define vsuxseg8ei32_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei32_v_i16mf4((int16_t *)(op0), (vuint32mf2_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (size_t)(op10)) +#define vsuxseg8ei32_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei32_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint32mf2_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (size_t)(op11)) +#define vsuxseg8ei32_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei32_v_u16m1((uint16_t *)(op0), (vuint32m2_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei32_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei32_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint32m2_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (size_t)(op11)) +#define vsuxseg8ei32_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei32_v_u16mf2((uint16_t *)(op0), (vuint32m1_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (size_t)(op10)) +#define vsuxseg8ei32_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei32_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint32m1_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (size_t)(op11)) +#define vsuxseg8ei32_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei32_v_u16mf4((uint16_t *)(op0), (vuint32mf2_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (size_t)(op10)) +#define vsuxseg8ei32_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei32_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint32mf2_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (size_t)(op11)) +#define vsuxseg2ei64_v_i16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei64_v_i16m1((int16_t *)(op0), (vuint64m4_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vsuxseg2ei64_v_i16m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei64_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint64m4_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (size_t)(op5)) +#define vsuxseg2ei64_v_i16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei64_v_i16m2((int16_t *)(op0), (vuint64m8_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei64_v_i16m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei64_v_i16m2_m((vbool8_t)(op0), (int16_t *)(op1), (vuint64m8_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei64_v_i16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei64_v_i16mf2((int16_t *)(op0), (vuint64m2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei64_v_i16mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei64_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint64m2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei64_v_i16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei64_v_i16mf4((int16_t *)(op0), (vuint64m1_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei64_v_i16mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei64_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint64m1_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (size_t)(op5)) +#define vsuxseg2ei64_v_u16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei64_v_u16m1((uint16_t *)(op0), (vuint64m4_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vsuxseg2ei64_v_u16m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei64_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint64m4_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (size_t)(op5)) +#define vsuxseg2ei64_v_u16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei64_v_u16m2((uint16_t *)(op0), (vuint64m8_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei64_v_u16m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei64_v_u16m2_m((vbool8_t)(op0), (uint16_t *)(op1), (vuint64m8_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei64_v_u16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei64_v_u16mf2((uint16_t *)(op0), (vuint64m2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei64_v_u16mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei64_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint64m2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei64_v_u16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei64_v_u16mf4((uint16_t *)(op0), (vuint64m1_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei64_v_u16mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei64_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint64m1_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (size_t)(op5)) +#define vsuxseg3ei64_v_i16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei64_v_i16m1((int16_t *)(op0), (vuint64m4_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (size_t)(op5)) +#define vsuxseg3ei64_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei64_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint64m4_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (size_t)(op6)) +#define vsuxseg3ei64_v_i16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei64_v_i16m2((int16_t *)(op0), (vuint64m8_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei64_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei64_v_i16m2_m((vbool8_t)(op0), (int16_t *)(op1), (vuint64m8_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (vint16m2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei64_v_i16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei64_v_i16mf2((int16_t *)(op0), (vuint64m2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei64_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei64_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint64m2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei64_v_i16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei64_v_i16mf4((int16_t *)(op0), (vuint64m1_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (size_t)(op5)) +#define vsuxseg3ei64_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei64_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint64m1_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (size_t)(op6)) +#define vsuxseg3ei64_v_u16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei64_v_u16m1((uint16_t *)(op0), (vuint64m4_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (size_t)(op5)) +#define vsuxseg3ei64_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei64_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint64m4_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (size_t)(op6)) +#define vsuxseg3ei64_v_u16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei64_v_u16m2((uint16_t *)(op0), (vuint64m8_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei64_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei64_v_u16m2_m((vbool8_t)(op0), (uint16_t *)(op1), (vuint64m8_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (vuint16m2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei64_v_u16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei64_v_u16mf2((uint16_t *)(op0), (vuint64m2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei64_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei64_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint64m2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei64_v_u16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei64_v_u16mf4((uint16_t *)(op0), (vuint64m1_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (size_t)(op5)) +#define vsuxseg3ei64_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei64_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint64m1_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (size_t)(op6)) +#define vsuxseg4ei64_v_i16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei64_v_i16m1((int16_t *)(op0), (vuint64m4_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (size_t)(op6)) +#define vsuxseg4ei64_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei64_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint64m4_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (size_t)(op7)) +#define vsuxseg4ei64_v_i16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei64_v_i16m2((int16_t *)(op0), (vuint64m8_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (vint16m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei64_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei64_v_i16m2_m((vbool8_t)(op0), (int16_t *)(op1), (vuint64m8_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (vint16m2_t)(op5), (vint16m2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei64_v_i16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei64_v_i16mf2((int16_t *)(op0), (vuint64m2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei64_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei64_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint64m2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei64_v_i16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei64_v_i16mf4((int16_t *)(op0), (vuint64m1_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (size_t)(op6)) +#define vsuxseg4ei64_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei64_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint64m1_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (size_t)(op7)) +#define vsuxseg4ei64_v_u16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei64_v_u16m1((uint16_t *)(op0), (vuint64m4_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (size_t)(op6)) +#define vsuxseg4ei64_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei64_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint64m4_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vsuxseg4ei64_v_u16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei64_v_u16m2((uint16_t *)(op0), (vuint64m8_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (vuint16m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei64_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei64_v_u16m2_m((vbool8_t)(op0), (uint16_t *)(op1), (vuint64m8_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (vuint16m2_t)(op5), (vuint16m2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei64_v_u16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei64_v_u16mf2((uint16_t *)(op0), (vuint64m2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei64_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei64_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint64m2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei64_v_u16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei64_v_u16mf4((uint16_t *)(op0), (vuint64m1_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (size_t)(op6)) +#define vsuxseg4ei64_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei64_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint64m1_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vsuxseg5ei64_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei64_v_i16m1((int16_t *)(op0), (vuint64m4_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (size_t)(op7)) +#define vsuxseg5ei64_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei64_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint64m4_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (size_t)(op8)) +#define vsuxseg5ei64_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei64_v_i16mf2((int16_t *)(op0), (vuint64m2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei64_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei64_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint64m2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (size_t)(op8)) +#define vsuxseg5ei64_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei64_v_i16mf4((int16_t *)(op0), (vuint64m1_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (size_t)(op7)) +#define vsuxseg5ei64_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei64_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint64m1_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (size_t)(op8)) +#define vsuxseg5ei64_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei64_v_u16m1((uint16_t *)(op0), (vuint64m4_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vsuxseg5ei64_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei64_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint64m4_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (size_t)(op8)) +#define vsuxseg5ei64_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei64_v_u16mf2((uint16_t *)(op0), (vuint64m2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei64_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei64_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint64m2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (size_t)(op8)) +#define vsuxseg5ei64_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei64_v_u16mf4((uint16_t *)(op0), (vuint64m1_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vsuxseg5ei64_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei64_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint64m1_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (size_t)(op8)) +#define vsuxseg6ei64_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei64_v_i16m1((int16_t *)(op0), (vuint64m4_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei64_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei64_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint64m4_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (size_t)(op9)) +#define vsuxseg6ei64_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei64_v_i16mf2((int16_t *)(op0), (vuint64m2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (size_t)(op8)) +#define vsuxseg6ei64_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei64_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint64m2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (size_t)(op9)) +#define vsuxseg6ei64_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei64_v_i16mf4((int16_t *)(op0), (vuint64m1_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (size_t)(op8)) +#define vsuxseg6ei64_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei64_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint64m1_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (size_t)(op9)) +#define vsuxseg6ei64_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei64_v_u16m1((uint16_t *)(op0), (vuint64m4_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei64_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei64_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint64m4_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vsuxseg6ei64_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei64_v_u16mf2((uint16_t *)(op0), (vuint64m2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (size_t)(op8)) +#define vsuxseg6ei64_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei64_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint64m2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vsuxseg6ei64_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei64_v_u16mf4((uint16_t *)(op0), (vuint64m1_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (size_t)(op8)) +#define vsuxseg6ei64_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei64_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint64m1_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vsuxseg7ei64_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei64_v_i16m1((int16_t *)(op0), (vuint64m4_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei64_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei64_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint64m4_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (size_t)(op10)) +#define vsuxseg7ei64_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei64_v_i16mf2((int16_t *)(op0), (vuint64m2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (size_t)(op9)) +#define vsuxseg7ei64_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei64_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint64m2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (size_t)(op10)) +#define vsuxseg7ei64_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei64_v_i16mf4((int16_t *)(op0), (vuint64m1_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (size_t)(op9)) +#define vsuxseg7ei64_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei64_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint64m1_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (size_t)(op10)) +#define vsuxseg7ei64_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei64_v_u16m1((uint16_t *)(op0), (vuint64m4_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei64_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei64_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint64m4_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (size_t)(op10)) +#define vsuxseg7ei64_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei64_v_u16mf2((uint16_t *)(op0), (vuint64m2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vsuxseg7ei64_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei64_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint64m2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (size_t)(op10)) +#define vsuxseg7ei64_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei64_v_u16mf4((uint16_t *)(op0), (vuint64m1_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vsuxseg7ei64_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei64_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint64m1_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (size_t)(op10)) +#define vsuxseg8ei64_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei64_v_i16m1((int16_t *)(op0), (vuint64m4_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei64_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei64_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint64m4_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (size_t)(op11)) +#define vsuxseg8ei64_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei64_v_i16mf2((int16_t *)(op0), (vuint64m2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (size_t)(op10)) +#define vsuxseg8ei64_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei64_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint64m2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (size_t)(op11)) +#define vsuxseg8ei64_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei64_v_i16mf4((int16_t *)(op0), (vuint64m1_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (size_t)(op10)) +#define vsuxseg8ei64_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei64_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint64m1_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (size_t)(op11)) +#define vsuxseg8ei64_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei64_v_u16m1((uint16_t *)(op0), (vuint64m4_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei64_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei64_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint64m4_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (size_t)(op11)) +#define vsuxseg8ei64_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei64_v_u16mf2((uint16_t *)(op0), (vuint64m2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (size_t)(op10)) +#define vsuxseg8ei64_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei64_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint64m2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (size_t)(op11)) +#define vsuxseg8ei64_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei64_v_u16mf4((uint16_t *)(op0), (vuint64m1_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (size_t)(op10)) +#define vsuxseg8ei64_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei64_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint64m1_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (size_t)(op11)) +#define vsuxseg2ei8_v_i32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_i32m1((int32_t *)(op0), (vuint8mf4_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_i32m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint8mf4_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (size_t)(op5)) +#define vsuxseg2ei8_v_i32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_i32m2((int32_t *)(op0), (vuint8mf2_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_i32m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint8mf2_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei8_v_i32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_i32m4((int32_t *)(op0), (vuint8m1_t)(op1), (vint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_i32m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vuint8m1_t)(op2), (vint32m4_t)(op3), (vint32m4_t)(op4), (size_t)(op5)) +#define vsuxseg2ei8_v_i32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_i32mf2((int32_t *)(op0), (vuint8mf8_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_i32mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint8mf8_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei8_v_u32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_u32m1((uint32_t *)(op0), (vuint8mf4_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_u32m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint8mf4_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (size_t)(op5)) +#define vsuxseg2ei8_v_u32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_u32m2((uint32_t *)(op0), (vuint8mf2_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_u32m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint8mf2_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei8_v_u32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_u32m4((uint32_t *)(op0), (vuint8m1_t)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_u32m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint8m1_t)(op2), (vuint32m4_t)(op3), (vuint32m4_t)(op4), (size_t)(op5)) +#define vsuxseg2ei8_v_u32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_u32mf2((uint32_t *)(op0), (vuint8mf8_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_u32mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint8mf8_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei8_v_i32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei8_v_i32m1((int32_t *)(op0), (vuint8mf4_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (size_t)(op5)) +#define vsuxseg3ei8_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei8_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint8mf4_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (size_t)(op6)) +#define vsuxseg3ei8_v_i32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei8_v_i32m2((int32_t *)(op0), (vuint8mf2_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei8_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei8_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint8mf2_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (vint32m2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei8_v_i32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei8_v_i32mf2((int32_t *)(op0), (vuint8mf8_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei8_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei8_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint8mf8_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei8_v_u32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei8_v_u32m1((uint32_t *)(op0), (vuint8mf4_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (size_t)(op5)) +#define vsuxseg3ei8_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei8_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint8mf4_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (size_t)(op6)) +#define vsuxseg3ei8_v_u32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei8_v_u32m2((uint32_t *)(op0), (vuint8mf2_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei8_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei8_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint8mf2_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (vuint32m2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei8_v_u32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei8_v_u32mf2((uint32_t *)(op0), (vuint8mf8_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei8_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei8_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint8mf8_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei8_v_i32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei8_v_i32m1((int32_t *)(op0), (vuint8mf4_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (size_t)(op6)) +#define vsuxseg4ei8_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei8_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint8mf4_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (size_t)(op7)) +#define vsuxseg4ei8_v_i32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei8_v_i32m2((int32_t *)(op0), (vuint8mf2_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (vint32m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei8_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei8_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint8mf2_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (vint32m2_t)(op5), (vint32m2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei8_v_i32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei8_v_i32mf2((int32_t *)(op0), (vuint8mf8_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei8_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei8_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint8mf8_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei8_v_u32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei8_v_u32m1((uint32_t *)(op0), (vuint8mf4_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (size_t)(op6)) +#define vsuxseg4ei8_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei8_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint8mf4_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vsuxseg4ei8_v_u32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei8_v_u32m2((uint32_t *)(op0), (vuint8mf2_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (vuint32m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei8_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei8_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint8mf2_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (vuint32m2_t)(op5), (vuint32m2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei8_v_u32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei8_v_u32mf2((uint32_t *)(op0), (vuint8mf8_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei8_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei8_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint8mf8_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei8_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei8_v_i32m1((int32_t *)(op0), (vuint8mf4_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (size_t)(op7)) +#define vsuxseg5ei8_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei8_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint8mf4_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (size_t)(op8)) +#define vsuxseg5ei8_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei8_v_i32mf2((int32_t *)(op0), (vuint8mf8_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei8_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei8_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint8mf8_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (size_t)(op8)) +#define vsuxseg5ei8_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei8_v_u32m1((uint32_t *)(op0), (vuint8mf4_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vsuxseg5ei8_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei8_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint8mf4_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (size_t)(op8)) +#define vsuxseg5ei8_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei8_v_u32mf2((uint32_t *)(op0), (vuint8mf8_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei8_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei8_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint8mf8_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (size_t)(op8)) +#define vsuxseg6ei8_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei8_v_i32m1((int32_t *)(op0), (vuint8mf4_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei8_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei8_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint8mf4_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (size_t)(op9)) +#define vsuxseg6ei8_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei8_v_i32mf2((int32_t *)(op0), (vuint8mf8_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (size_t)(op8)) +#define vsuxseg6ei8_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei8_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint8mf8_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (size_t)(op9)) +#define vsuxseg6ei8_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei8_v_u32m1((uint32_t *)(op0), (vuint8mf4_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei8_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei8_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint8mf4_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vsuxseg6ei8_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei8_v_u32mf2((uint32_t *)(op0), (vuint8mf8_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (size_t)(op8)) +#define vsuxseg6ei8_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei8_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint8mf8_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vsuxseg7ei8_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei8_v_i32m1((int32_t *)(op0), (vuint8mf4_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei8_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei8_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint8mf4_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (size_t)(op10)) +#define vsuxseg7ei8_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei8_v_i32mf2((int32_t *)(op0), (vuint8mf8_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (size_t)(op9)) +#define vsuxseg7ei8_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei8_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint8mf8_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (size_t)(op10)) +#define vsuxseg7ei8_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei8_v_u32m1((uint32_t *)(op0), (vuint8mf4_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei8_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei8_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint8mf4_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (size_t)(op10)) +#define vsuxseg7ei8_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei8_v_u32mf2((uint32_t *)(op0), (vuint8mf8_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vsuxseg7ei8_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei8_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint8mf8_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (size_t)(op10)) +#define vsuxseg8ei8_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei8_v_i32m1((int32_t *)(op0), (vuint8mf4_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei8_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei8_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint8mf4_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (size_t)(op11)) +#define vsuxseg8ei8_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei8_v_i32mf2((int32_t *)(op0), (vuint8mf8_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (size_t)(op10)) +#define vsuxseg8ei8_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei8_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint8mf8_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (size_t)(op11)) +#define vsuxseg8ei8_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei8_v_u32m1((uint32_t *)(op0), (vuint8mf4_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei8_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei8_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint8mf4_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (size_t)(op11)) +#define vsuxseg8ei8_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei8_v_u32mf2((uint32_t *)(op0), (vuint8mf8_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (size_t)(op10)) +#define vsuxseg8ei8_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei8_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint8mf8_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (size_t)(op11)) +#define vsuxseg2ei16_v_i32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_i32m1((int32_t *)(op0), (vuint16mf2_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_i32m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint16mf2_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (size_t)(op5)) +#define vsuxseg2ei16_v_i32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_i32m2((int32_t *)(op0), (vuint16m1_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_i32m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint16m1_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei16_v_i32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_i32m4((int32_t *)(op0), (vuint16m2_t)(op1), (vint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_i32m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vuint16m2_t)(op2), (vint32m4_t)(op3), (vint32m4_t)(op4), (size_t)(op5)) +#define vsuxseg2ei16_v_i32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_i32mf2((int32_t *)(op0), (vuint16mf4_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_i32mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint16mf4_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei16_v_u32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_u32m1((uint32_t *)(op0), (vuint16mf2_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_u32m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint16mf2_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (size_t)(op5)) +#define vsuxseg2ei16_v_u32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_u32m2((uint32_t *)(op0), (vuint16m1_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_u32m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint16m1_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei16_v_u32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_u32m4((uint32_t *)(op0), (vuint16m2_t)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_u32m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint16m2_t)(op2), (vuint32m4_t)(op3), (vuint32m4_t)(op4), (size_t)(op5)) +#define vsuxseg2ei16_v_u32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_u32mf2((uint32_t *)(op0), (vuint16mf4_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_u32mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint16mf4_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei16_v_i32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei16_v_i32m1((int32_t *)(op0), (vuint16mf2_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (size_t)(op5)) +#define vsuxseg3ei16_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei16_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint16mf2_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (size_t)(op6)) +#define vsuxseg3ei16_v_i32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei16_v_i32m2((int32_t *)(op0), (vuint16m1_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei16_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei16_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint16m1_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (vint32m2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei16_v_i32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei16_v_i32mf2((int32_t *)(op0), (vuint16mf4_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei16_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei16_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint16mf4_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei16_v_u32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei16_v_u32m1((uint32_t *)(op0), (vuint16mf2_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (size_t)(op5)) +#define vsuxseg3ei16_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei16_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint16mf2_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (size_t)(op6)) +#define vsuxseg3ei16_v_u32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei16_v_u32m2((uint32_t *)(op0), (vuint16m1_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei16_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei16_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint16m1_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (vuint32m2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei16_v_u32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei16_v_u32mf2((uint32_t *)(op0), (vuint16mf4_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei16_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei16_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint16mf4_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei16_v_i32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei16_v_i32m1((int32_t *)(op0), (vuint16mf2_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (size_t)(op6)) +#define vsuxseg4ei16_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei16_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint16mf2_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (size_t)(op7)) +#define vsuxseg4ei16_v_i32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei16_v_i32m2((int32_t *)(op0), (vuint16m1_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (vint32m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei16_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei16_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint16m1_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (vint32m2_t)(op5), (vint32m2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei16_v_i32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei16_v_i32mf2((int32_t *)(op0), (vuint16mf4_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei16_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei16_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint16mf4_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei16_v_u32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei16_v_u32m1((uint32_t *)(op0), (vuint16mf2_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (size_t)(op6)) +#define vsuxseg4ei16_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei16_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint16mf2_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vsuxseg4ei16_v_u32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei16_v_u32m2((uint32_t *)(op0), (vuint16m1_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (vuint32m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei16_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei16_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint16m1_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (vuint32m2_t)(op5), (vuint32m2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei16_v_u32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei16_v_u32mf2((uint32_t *)(op0), (vuint16mf4_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei16_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei16_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint16mf4_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei16_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei16_v_i32m1((int32_t *)(op0), (vuint16mf2_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (size_t)(op7)) +#define vsuxseg5ei16_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei16_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint16mf2_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (size_t)(op8)) +#define vsuxseg5ei16_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei16_v_i32mf2((int32_t *)(op0), (vuint16mf4_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei16_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei16_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint16mf4_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (size_t)(op8)) +#define vsuxseg5ei16_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei16_v_u32m1((uint32_t *)(op0), (vuint16mf2_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vsuxseg5ei16_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei16_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint16mf2_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (size_t)(op8)) +#define vsuxseg5ei16_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei16_v_u32mf2((uint32_t *)(op0), (vuint16mf4_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei16_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei16_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint16mf4_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (size_t)(op8)) +#define vsuxseg6ei16_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei16_v_i32m1((int32_t *)(op0), (vuint16mf2_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei16_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei16_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint16mf2_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (size_t)(op9)) +#define vsuxseg6ei16_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei16_v_i32mf2((int32_t *)(op0), (vuint16mf4_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (size_t)(op8)) +#define vsuxseg6ei16_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei16_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint16mf4_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (size_t)(op9)) +#define vsuxseg6ei16_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei16_v_u32m1((uint32_t *)(op0), (vuint16mf2_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei16_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei16_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint16mf2_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vsuxseg6ei16_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei16_v_u32mf2((uint32_t *)(op0), (vuint16mf4_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (size_t)(op8)) +#define vsuxseg6ei16_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei16_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint16mf4_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vsuxseg7ei16_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei16_v_i32m1((int32_t *)(op0), (vuint16mf2_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei16_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei16_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint16mf2_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (size_t)(op10)) +#define vsuxseg7ei16_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei16_v_i32mf2((int32_t *)(op0), (vuint16mf4_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (size_t)(op9)) +#define vsuxseg7ei16_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei16_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint16mf4_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (size_t)(op10)) +#define vsuxseg7ei16_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei16_v_u32m1((uint32_t *)(op0), (vuint16mf2_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei16_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei16_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint16mf2_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (size_t)(op10)) +#define vsuxseg7ei16_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei16_v_u32mf2((uint32_t *)(op0), (vuint16mf4_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vsuxseg7ei16_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei16_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint16mf4_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (size_t)(op10)) +#define vsuxseg8ei16_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei16_v_i32m1((int32_t *)(op0), (vuint16mf2_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei16_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei16_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint16mf2_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (size_t)(op11)) +#define vsuxseg8ei16_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei16_v_i32mf2((int32_t *)(op0), (vuint16mf4_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (size_t)(op10)) +#define vsuxseg8ei16_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei16_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint16mf4_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (size_t)(op11)) +#define vsuxseg8ei16_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei16_v_u32m1((uint32_t *)(op0), (vuint16mf2_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei16_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei16_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint16mf2_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (size_t)(op11)) +#define vsuxseg8ei16_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei16_v_u32mf2((uint32_t *)(op0), (vuint16mf4_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (size_t)(op10)) +#define vsuxseg8ei16_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei16_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint16mf4_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (size_t)(op11)) +#define vsuxseg2ei32_v_i32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_i32m1((int32_t *)(op0), (vuint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_i32m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (size_t)(op5)) +#define vsuxseg2ei32_v_i32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_i32m2((int32_t *)(op0), (vuint32m2_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_i32m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint32m2_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei32_v_i32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_i32m4((int32_t *)(op0), (vuint32m4_t)(op1), (vint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_i32m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vuint32m4_t)(op2), (vint32m4_t)(op3), (vint32m4_t)(op4), (size_t)(op5)) +#define vsuxseg2ei32_v_i32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_i32mf2((int32_t *)(op0), (vuint32mf2_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_i32mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei32_v_u32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_u32m1((uint32_t *)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_u32m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (size_t)(op5)) +#define vsuxseg2ei32_v_u32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_u32m2((uint32_t *)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_u32m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei32_v_u32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_u32m4((uint32_t *)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_u32m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (vuint32m4_t)(op4), (size_t)(op5)) +#define vsuxseg2ei32_v_u32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_u32mf2((uint32_t *)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_u32mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei32_v_i32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei32_v_i32m1((int32_t *)(op0), (vuint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (size_t)(op5)) +#define vsuxseg3ei32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei32_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (size_t)(op6)) +#define vsuxseg3ei32_v_i32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei32_v_i32m2((int32_t *)(op0), (vuint32m2_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei32_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei32_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint32m2_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (vint32m2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei32_v_i32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei32_v_i32mf2((int32_t *)(op0), (vuint32mf2_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei32_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei32_v_u32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei32_v_u32m1((uint32_t *)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (size_t)(op5)) +#define vsuxseg3ei32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei32_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (size_t)(op6)) +#define vsuxseg3ei32_v_u32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei32_v_u32m2((uint32_t *)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei32_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei32_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (vuint32m2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei32_v_u32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei32_v_u32mf2((uint32_t *)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei32_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei32_v_i32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei32_v_i32m1((int32_t *)(op0), (vuint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (size_t)(op6)) +#define vsuxseg4ei32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei32_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (size_t)(op7)) +#define vsuxseg4ei32_v_i32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei32_v_i32m2((int32_t *)(op0), (vuint32m2_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (vint32m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei32_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei32_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint32m2_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (vint32m2_t)(op5), (vint32m2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei32_v_i32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei32_v_i32mf2((int32_t *)(op0), (vuint32mf2_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei32_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei32_v_u32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei32_v_u32m1((uint32_t *)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (size_t)(op6)) +#define vsuxseg4ei32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei32_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vsuxseg4ei32_v_u32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei32_v_u32m2((uint32_t *)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (vuint32m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei32_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei32_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (vuint32m2_t)(op5), (vuint32m2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei32_v_u32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei32_v_u32mf2((uint32_t *)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei32_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei32_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei32_v_i32m1((int32_t *)(op0), (vuint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (size_t)(op7)) +#define vsuxseg5ei32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei32_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (size_t)(op8)) +#define vsuxseg5ei32_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei32_v_i32mf2((int32_t *)(op0), (vuint32mf2_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei32_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (size_t)(op8)) +#define vsuxseg5ei32_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei32_v_u32m1((uint32_t *)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vsuxseg5ei32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei32_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (size_t)(op8)) +#define vsuxseg5ei32_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei32_v_u32mf2((uint32_t *)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei32_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (size_t)(op8)) +#define vsuxseg6ei32_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei32_v_i32m1((int32_t *)(op0), (vuint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei32_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (size_t)(op9)) +#define vsuxseg6ei32_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei32_v_i32mf2((int32_t *)(op0), (vuint32mf2_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (size_t)(op8)) +#define vsuxseg6ei32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei32_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (size_t)(op9)) +#define vsuxseg6ei32_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei32_v_u32m1((uint32_t *)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei32_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vsuxseg6ei32_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei32_v_u32mf2((uint32_t *)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (size_t)(op8)) +#define vsuxseg6ei32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei32_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vsuxseg7ei32_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei32_v_i32m1((int32_t *)(op0), (vuint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei32_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (size_t)(op10)) +#define vsuxseg7ei32_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei32_v_i32mf2((int32_t *)(op0), (vuint32mf2_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (size_t)(op9)) +#define vsuxseg7ei32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei32_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (size_t)(op10)) +#define vsuxseg7ei32_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei32_v_u32m1((uint32_t *)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei32_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (size_t)(op10)) +#define vsuxseg7ei32_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei32_v_u32mf2((uint32_t *)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vsuxseg7ei32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei32_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (size_t)(op10)) +#define vsuxseg8ei32_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei32_v_i32m1((int32_t *)(op0), (vuint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei32_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (size_t)(op11)) +#define vsuxseg8ei32_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei32_v_i32mf2((int32_t *)(op0), (vuint32mf2_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (size_t)(op10)) +#define vsuxseg8ei32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei32_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (size_t)(op11)) +#define vsuxseg8ei32_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei32_v_u32m1((uint32_t *)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei32_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (size_t)(op11)) +#define vsuxseg8ei32_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei32_v_u32mf2((uint32_t *)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (size_t)(op10)) +#define vsuxseg8ei32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei32_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (size_t)(op11)) +#define vsuxseg2ei64_v_i32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei64_v_i32m1((int32_t *)(op0), (vuint64m2_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vsuxseg2ei64_v_i32m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei64_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint64m2_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (size_t)(op5)) +#define vsuxseg2ei64_v_i32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei64_v_i32m2((int32_t *)(op0), (vuint64m4_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei64_v_i32m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei64_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint64m4_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei64_v_i32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei64_v_i32m4((int32_t *)(op0), (vuint64m8_t)(op1), (vint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei64_v_i32m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei64_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vuint64m8_t)(op2), (vint32m4_t)(op3), (vint32m4_t)(op4), (size_t)(op5)) +#define vsuxseg2ei64_v_i32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei64_v_i32mf2((int32_t *)(op0), (vuint64m1_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei64_v_i32mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei64_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint64m1_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei64_v_u32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei64_v_u32m1((uint32_t *)(op0), (vuint64m2_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vsuxseg2ei64_v_u32m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei64_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint64m2_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (size_t)(op5)) +#define vsuxseg2ei64_v_u32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei64_v_u32m2((uint32_t *)(op0), (vuint64m4_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei64_v_u32m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei64_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint64m4_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei64_v_u32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei64_v_u32m4((uint32_t *)(op0), (vuint64m8_t)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei64_v_u32m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei64_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint64m8_t)(op2), (vuint32m4_t)(op3), (vuint32m4_t)(op4), (size_t)(op5)) +#define vsuxseg2ei64_v_u32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei64_v_u32mf2((uint32_t *)(op0), (vuint64m1_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei64_v_u32mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei64_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint64m1_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei64_v_i32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei64_v_i32m1((int32_t *)(op0), (vuint64m2_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (size_t)(op5)) +#define vsuxseg3ei64_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei64_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint64m2_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (size_t)(op6)) +#define vsuxseg3ei64_v_i32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei64_v_i32m2((int32_t *)(op0), (vuint64m4_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei64_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei64_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint64m4_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (vint32m2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei64_v_i32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei64_v_i32mf2((int32_t *)(op0), (vuint64m1_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei64_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei64_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint64m1_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei64_v_u32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei64_v_u32m1((uint32_t *)(op0), (vuint64m2_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (size_t)(op5)) +#define vsuxseg3ei64_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei64_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint64m2_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (size_t)(op6)) +#define vsuxseg3ei64_v_u32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei64_v_u32m2((uint32_t *)(op0), (vuint64m4_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei64_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei64_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint64m4_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (vuint32m2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei64_v_u32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei64_v_u32mf2((uint32_t *)(op0), (vuint64m1_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei64_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei64_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint64m1_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei64_v_i32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei64_v_i32m1((int32_t *)(op0), (vuint64m2_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (size_t)(op6)) +#define vsuxseg4ei64_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei64_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint64m2_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (size_t)(op7)) +#define vsuxseg4ei64_v_i32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei64_v_i32m2((int32_t *)(op0), (vuint64m4_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (vint32m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei64_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei64_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint64m4_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (vint32m2_t)(op5), (vint32m2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei64_v_i32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei64_v_i32mf2((int32_t *)(op0), (vuint64m1_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei64_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei64_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint64m1_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei64_v_u32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei64_v_u32m1((uint32_t *)(op0), (vuint64m2_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (size_t)(op6)) +#define vsuxseg4ei64_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei64_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint64m2_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vsuxseg4ei64_v_u32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei64_v_u32m2((uint32_t *)(op0), (vuint64m4_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (vuint32m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei64_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei64_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint64m4_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (vuint32m2_t)(op5), (vuint32m2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei64_v_u32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei64_v_u32mf2((uint32_t *)(op0), (vuint64m1_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei64_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei64_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint64m1_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei64_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei64_v_i32m1((int32_t *)(op0), (vuint64m2_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (size_t)(op7)) +#define vsuxseg5ei64_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei64_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint64m2_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (size_t)(op8)) +#define vsuxseg5ei64_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei64_v_i32mf2((int32_t *)(op0), (vuint64m1_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei64_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei64_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint64m1_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (size_t)(op8)) +#define vsuxseg5ei64_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei64_v_u32m1((uint32_t *)(op0), (vuint64m2_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vsuxseg5ei64_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei64_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint64m2_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (size_t)(op8)) +#define vsuxseg5ei64_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei64_v_u32mf2((uint32_t *)(op0), (vuint64m1_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei64_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei64_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint64m1_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (size_t)(op8)) +#define vsuxseg6ei64_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei64_v_i32m1((int32_t *)(op0), (vuint64m2_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei64_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei64_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint64m2_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (size_t)(op9)) +#define vsuxseg6ei64_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei64_v_i32mf2((int32_t *)(op0), (vuint64m1_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (size_t)(op8)) +#define vsuxseg6ei64_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei64_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint64m1_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (size_t)(op9)) +#define vsuxseg6ei64_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei64_v_u32m1((uint32_t *)(op0), (vuint64m2_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei64_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei64_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint64m2_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vsuxseg6ei64_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei64_v_u32mf2((uint32_t *)(op0), (vuint64m1_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (size_t)(op8)) +#define vsuxseg6ei64_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei64_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint64m1_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vsuxseg7ei64_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei64_v_i32m1((int32_t *)(op0), (vuint64m2_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei64_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei64_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint64m2_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (size_t)(op10)) +#define vsuxseg7ei64_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei64_v_i32mf2((int32_t *)(op0), (vuint64m1_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (size_t)(op9)) +#define vsuxseg7ei64_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei64_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint64m1_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (size_t)(op10)) +#define vsuxseg7ei64_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei64_v_u32m1((uint32_t *)(op0), (vuint64m2_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei64_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei64_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint64m2_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (size_t)(op10)) +#define vsuxseg7ei64_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei64_v_u32mf2((uint32_t *)(op0), (vuint64m1_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vsuxseg7ei64_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei64_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint64m1_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (size_t)(op10)) +#define vsuxseg8ei64_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei64_v_i32m1((int32_t *)(op0), (vuint64m2_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei64_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei64_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint64m2_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (size_t)(op11)) +#define vsuxseg8ei64_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei64_v_i32mf2((int32_t *)(op0), (vuint64m1_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (size_t)(op10)) +#define vsuxseg8ei64_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei64_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint64m1_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (size_t)(op11)) +#define vsuxseg8ei64_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei64_v_u32m1((uint32_t *)(op0), (vuint64m2_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei64_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei64_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint64m2_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (size_t)(op11)) +#define vsuxseg8ei64_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei64_v_u32mf2((uint32_t *)(op0), (vuint64m1_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (size_t)(op10)) +#define vsuxseg8ei64_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei64_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint64m1_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (size_t)(op11)) +#define vsuxseg2ei8_v_i64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_i64m1((int64_t *)(op0), (vuint8mf8_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_i64m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint8mf8_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (size_t)(op5)) +#define vsuxseg2ei8_v_i64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_i64m2((int64_t *)(op0), (vuint8mf4_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_i64m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint8mf4_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei8_v_i64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_i64m4((int64_t *)(op0), (vuint8mf2_t)(op1), (vint64m4_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_i64m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vuint8mf2_t)(op2), (vint64m4_t)(op3), (vint64m4_t)(op4), (size_t)(op5)) +#define vsuxseg2ei8_v_u64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_u64m1((uint64_t *)(op0), (vuint8mf8_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_u64m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint8mf8_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (size_t)(op5)) +#define vsuxseg2ei8_v_u64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_u64m2((uint64_t *)(op0), (vuint8mf4_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_u64m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint8mf4_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei8_v_u64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_u64m4((uint64_t *)(op0), (vuint8mf2_t)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_u64m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint8mf2_t)(op2), (vuint64m4_t)(op3), (vuint64m4_t)(op4), (size_t)(op5)) +#define vsuxseg3ei8_v_i64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei8_v_i64m1((int64_t *)(op0), (vuint8mf8_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (size_t)(op5)) +#define vsuxseg3ei8_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei8_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint8mf8_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (size_t)(op6)) +#define vsuxseg3ei8_v_i64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei8_v_i64m2((int64_t *)(op0), (vuint8mf4_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei8_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei8_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint8mf4_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (vint64m2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei8_v_u64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei8_v_u64m1((uint64_t *)(op0), (vuint8mf8_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (size_t)(op5)) +#define vsuxseg3ei8_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei8_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint8mf8_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (size_t)(op6)) +#define vsuxseg3ei8_v_u64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei8_v_u64m2((uint64_t *)(op0), (vuint8mf4_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei8_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei8_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint8mf4_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (vuint64m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei8_v_i64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei8_v_i64m1((int64_t *)(op0), (vuint8mf8_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (size_t)(op6)) +#define vsuxseg4ei8_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei8_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint8mf8_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (size_t)(op7)) +#define vsuxseg4ei8_v_i64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei8_v_i64m2((int64_t *)(op0), (vuint8mf4_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (vint64m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei8_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei8_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint8mf4_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (vint64m2_t)(op5), (vint64m2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei8_v_u64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei8_v_u64m1((uint64_t *)(op0), (vuint8mf8_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (size_t)(op6)) +#define vsuxseg4ei8_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei8_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint8mf8_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vsuxseg4ei8_v_u64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei8_v_u64m2((uint64_t *)(op0), (vuint8mf4_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (vuint64m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei8_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei8_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint8mf4_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (vuint64m2_t)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei8_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei8_v_i64m1((int64_t *)(op0), (vuint8mf8_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (size_t)(op7)) +#define vsuxseg5ei8_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei8_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint8mf8_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (size_t)(op8)) +#define vsuxseg5ei8_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei8_v_u64m1((uint64_t *)(op0), (vuint8mf8_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vsuxseg5ei8_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei8_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint8mf8_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei8_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei8_v_i64m1((int64_t *)(op0), (vuint8mf8_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei8_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei8_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint8mf8_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (size_t)(op9)) +#define vsuxseg6ei8_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei8_v_u64m1((uint64_t *)(op0), (vuint8mf8_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei8_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei8_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint8mf8_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei8_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei8_v_i64m1((int64_t *)(op0), (vuint8mf8_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei8_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei8_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint8mf8_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (size_t)(op10)) +#define vsuxseg7ei8_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei8_v_u64m1((uint64_t *)(op0), (vuint8mf8_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei8_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei8_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint8mf8_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei8_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei8_v_i64m1((int64_t *)(op0), (vuint8mf8_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei8_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei8_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint8mf8_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (size_t)(op11)) +#define vsuxseg8ei8_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei8_v_u64m1((uint64_t *)(op0), (vuint8mf8_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei8_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei8_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint8mf8_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (size_t)(op11)) +#define vsuxseg2ei16_v_i64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_i64m1((int64_t *)(op0), (vuint16mf4_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_i64m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint16mf4_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (size_t)(op5)) +#define vsuxseg2ei16_v_i64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_i64m2((int64_t *)(op0), (vuint16mf2_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_i64m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint16mf2_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei16_v_i64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_i64m4((int64_t *)(op0), (vuint16m1_t)(op1), (vint64m4_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_i64m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vuint16m1_t)(op2), (vint64m4_t)(op3), (vint64m4_t)(op4), (size_t)(op5)) +#define vsuxseg2ei16_v_u64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_u64m1((uint64_t *)(op0), (vuint16mf4_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_u64m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint16mf4_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (size_t)(op5)) +#define vsuxseg2ei16_v_u64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_u64m2((uint64_t *)(op0), (vuint16mf2_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_u64m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint16mf2_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei16_v_u64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_u64m4((uint64_t *)(op0), (vuint16m1_t)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_u64m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint16m1_t)(op2), (vuint64m4_t)(op3), (vuint64m4_t)(op4), (size_t)(op5)) +#define vsuxseg3ei16_v_i64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei16_v_i64m1((int64_t *)(op0), (vuint16mf4_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (size_t)(op5)) +#define vsuxseg3ei16_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei16_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint16mf4_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (size_t)(op6)) +#define vsuxseg3ei16_v_i64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei16_v_i64m2((int64_t *)(op0), (vuint16mf2_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei16_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei16_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint16mf2_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (vint64m2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei16_v_u64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei16_v_u64m1((uint64_t *)(op0), (vuint16mf4_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (size_t)(op5)) +#define vsuxseg3ei16_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei16_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint16mf4_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (size_t)(op6)) +#define vsuxseg3ei16_v_u64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei16_v_u64m2((uint64_t *)(op0), (vuint16mf2_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei16_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei16_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint16mf2_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (vuint64m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei16_v_i64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei16_v_i64m1((int64_t *)(op0), (vuint16mf4_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (size_t)(op6)) +#define vsuxseg4ei16_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei16_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint16mf4_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (size_t)(op7)) +#define vsuxseg4ei16_v_i64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei16_v_i64m2((int64_t *)(op0), (vuint16mf2_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (vint64m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei16_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei16_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint16mf2_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (vint64m2_t)(op5), (vint64m2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei16_v_u64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei16_v_u64m1((uint64_t *)(op0), (vuint16mf4_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (size_t)(op6)) +#define vsuxseg4ei16_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei16_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint16mf4_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vsuxseg4ei16_v_u64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei16_v_u64m2((uint64_t *)(op0), (vuint16mf2_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (vuint64m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei16_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei16_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint16mf2_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (vuint64m2_t)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei16_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei16_v_i64m1((int64_t *)(op0), (vuint16mf4_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (size_t)(op7)) +#define vsuxseg5ei16_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei16_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint16mf4_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (size_t)(op8)) +#define vsuxseg5ei16_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei16_v_u64m1((uint64_t *)(op0), (vuint16mf4_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vsuxseg5ei16_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei16_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint16mf4_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei16_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei16_v_i64m1((int64_t *)(op0), (vuint16mf4_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei16_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei16_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint16mf4_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (size_t)(op9)) +#define vsuxseg6ei16_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei16_v_u64m1((uint64_t *)(op0), (vuint16mf4_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei16_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei16_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint16mf4_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei16_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei16_v_i64m1((int64_t *)(op0), (vuint16mf4_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei16_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei16_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint16mf4_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (size_t)(op10)) +#define vsuxseg7ei16_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei16_v_u64m1((uint64_t *)(op0), (vuint16mf4_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei16_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei16_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint16mf4_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei16_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei16_v_i64m1((int64_t *)(op0), (vuint16mf4_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei16_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei16_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint16mf4_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (size_t)(op11)) +#define vsuxseg8ei16_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei16_v_u64m1((uint64_t *)(op0), (vuint16mf4_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei16_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei16_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint16mf4_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (size_t)(op11)) +#define vsuxseg2ei32_v_i64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_i64m1((int64_t *)(op0), (vuint32mf2_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_i64m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint32mf2_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (size_t)(op5)) +#define vsuxseg2ei32_v_i64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_i64m2((int64_t *)(op0), (vuint32m1_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_i64m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint32m1_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei32_v_i64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_i64m4((int64_t *)(op0), (vuint32m2_t)(op1), (vint64m4_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_i64m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vuint32m2_t)(op2), (vint64m4_t)(op3), (vint64m4_t)(op4), (size_t)(op5)) +#define vsuxseg2ei32_v_u64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_u64m1((uint64_t *)(op0), (vuint32mf2_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_u64m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint32mf2_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (size_t)(op5)) +#define vsuxseg2ei32_v_u64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_u64m2((uint64_t *)(op0), (vuint32m1_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_u64m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint32m1_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei32_v_u64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_u64m4((uint64_t *)(op0), (vuint32m2_t)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_u64m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint32m2_t)(op2), (vuint64m4_t)(op3), (vuint64m4_t)(op4), (size_t)(op5)) +#define vsuxseg3ei32_v_i64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei32_v_i64m1((int64_t *)(op0), (vuint32mf2_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (size_t)(op5)) +#define vsuxseg3ei32_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei32_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint32mf2_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (size_t)(op6)) +#define vsuxseg3ei32_v_i64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei32_v_i64m2((int64_t *)(op0), (vuint32m1_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei32_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei32_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint32m1_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (vint64m2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei32_v_u64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei32_v_u64m1((uint64_t *)(op0), (vuint32mf2_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (size_t)(op5)) +#define vsuxseg3ei32_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei32_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint32mf2_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (size_t)(op6)) +#define vsuxseg3ei32_v_u64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei32_v_u64m2((uint64_t *)(op0), (vuint32m1_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei32_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei32_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint32m1_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (vuint64m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei32_v_i64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei32_v_i64m1((int64_t *)(op0), (vuint32mf2_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (size_t)(op6)) +#define vsuxseg4ei32_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei32_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint32mf2_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (size_t)(op7)) +#define vsuxseg4ei32_v_i64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei32_v_i64m2((int64_t *)(op0), (vuint32m1_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (vint64m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei32_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei32_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint32m1_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (vint64m2_t)(op5), (vint64m2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei32_v_u64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei32_v_u64m1((uint64_t *)(op0), (vuint32mf2_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (size_t)(op6)) +#define vsuxseg4ei32_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei32_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint32mf2_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vsuxseg4ei32_v_u64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei32_v_u64m2((uint64_t *)(op0), (vuint32m1_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (vuint64m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei32_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei32_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint32m1_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (vuint64m2_t)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei32_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei32_v_i64m1((int64_t *)(op0), (vuint32mf2_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (size_t)(op7)) +#define vsuxseg5ei32_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei32_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint32mf2_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (size_t)(op8)) +#define vsuxseg5ei32_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei32_v_u64m1((uint64_t *)(op0), (vuint32mf2_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vsuxseg5ei32_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei32_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint32mf2_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei32_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei32_v_i64m1((int64_t *)(op0), (vuint32mf2_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei32_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei32_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint32mf2_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (size_t)(op9)) +#define vsuxseg6ei32_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei32_v_u64m1((uint64_t *)(op0), (vuint32mf2_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei32_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei32_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint32mf2_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei32_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei32_v_i64m1((int64_t *)(op0), (vuint32mf2_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei32_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei32_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint32mf2_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (size_t)(op10)) +#define vsuxseg7ei32_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei32_v_u64m1((uint64_t *)(op0), (vuint32mf2_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei32_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei32_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint32mf2_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei32_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei32_v_i64m1((int64_t *)(op0), (vuint32mf2_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei32_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei32_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint32mf2_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (size_t)(op11)) +#define vsuxseg8ei32_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei32_v_u64m1((uint64_t *)(op0), (vuint32mf2_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei32_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei32_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint32mf2_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (size_t)(op11)) +#define vsuxseg2ei64_v_i64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei64_v_i64m1((int64_t *)(op0), (vuint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vsuxseg2ei64_v_i64m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei64_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (size_t)(op5)) +#define vsuxseg2ei64_v_i64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei64_v_i64m2((int64_t *)(op0), (vuint64m2_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei64_v_i64m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei64_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint64m2_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei64_v_i64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei64_v_i64m4((int64_t *)(op0), (vuint64m4_t)(op1), (vint64m4_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei64_v_i64m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei64_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vuint64m4_t)(op2), (vint64m4_t)(op3), (vint64m4_t)(op4), (size_t)(op5)) +#define vsuxseg2ei64_v_u64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei64_v_u64m1((uint64_t *)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vsuxseg2ei64_v_u64m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei64_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (size_t)(op5)) +#define vsuxseg2ei64_v_u64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei64_v_u64m2((uint64_t *)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei64_v_u64m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei64_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei64_v_u64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei64_v_u64m4((uint64_t *)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei64_v_u64m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei64_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (vuint64m4_t)(op4), (size_t)(op5)) +#define vsuxseg3ei64_v_i64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei64_v_i64m1((int64_t *)(op0), (vuint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (size_t)(op5)) +#define vsuxseg3ei64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei64_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (size_t)(op6)) +#define vsuxseg3ei64_v_i64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei64_v_i64m2((int64_t *)(op0), (vuint64m2_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei64_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei64_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint64m2_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (vint64m2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei64_v_u64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei64_v_u64m1((uint64_t *)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (size_t)(op5)) +#define vsuxseg3ei64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei64_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (size_t)(op6)) +#define vsuxseg3ei64_v_u64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei64_v_u64m2((uint64_t *)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei64_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei64_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (vuint64m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei64_v_i64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei64_v_i64m1((int64_t *)(op0), (vuint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (size_t)(op6)) +#define vsuxseg4ei64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei64_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (size_t)(op7)) +#define vsuxseg4ei64_v_i64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei64_v_i64m2((int64_t *)(op0), (vuint64m2_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (vint64m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei64_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei64_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint64m2_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (vint64m2_t)(op5), (vint64m2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei64_v_u64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei64_v_u64m1((uint64_t *)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (size_t)(op6)) +#define vsuxseg4ei64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei64_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vsuxseg4ei64_v_u64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei64_v_u64m2((uint64_t *)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (vuint64m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei64_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei64_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (vuint64m2_t)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei64_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei64_v_i64m1((int64_t *)(op0), (vuint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (size_t)(op7)) +#define vsuxseg5ei64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei64_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (size_t)(op8)) +#define vsuxseg5ei64_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei64_v_u64m1((uint64_t *)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vsuxseg5ei64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei64_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei64_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei64_v_i64m1((int64_t *)(op0), (vuint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei64_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (size_t)(op9)) +#define vsuxseg6ei64_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei64_v_u64m1((uint64_t *)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei64_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei64_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei64_v_i64m1((int64_t *)(op0), (vuint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei64_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (size_t)(op10)) +#define vsuxseg7ei64_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei64_v_u64m1((uint64_t *)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei64_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei64_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei64_v_i64m1((int64_t *)(op0), (vuint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei64_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (size_t)(op11)) +#define vsuxseg8ei64_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei64_v_u64m1((uint64_t *)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei64_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (size_t)(op11)) +#define vsoxseg2ei8_v_i8m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_i8m1((int8_t *)(op0), (vuint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_i8m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (size_t)(op5)) +#define vsoxseg2ei8_v_i8m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_i8m2((int8_t *)(op0), (vuint8m2_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_i8m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_i8m2_m((vbool4_t)(op0), (int8_t *)(op1), (vuint8m2_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei8_v_i8m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_i8m4((int8_t *)(op0), (vuint8m4_t)(op1), (vint8m4_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_i8m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_i8m4_m((vbool2_t)(op0), (int8_t *)(op1), (vuint8m4_t)(op2), (vint8m4_t)(op3), (vint8m4_t)(op4), (size_t)(op5)) +#define vsoxseg2ei8_v_i8mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_i8mf2((int8_t *)(op0), (vuint8mf2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_i8mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei8_v_i8mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_i8mf4((int8_t *)(op0), (vuint8mf4_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_i8mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (size_t)(op5)) +#define vsoxseg2ei8_v_i8mf8(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_i8mf8((int8_t *)(op0), (vuint8mf8_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_i8mf8_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (size_t)(op5)) +#define vsoxseg2ei8_v_u8m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_u8m1((uint8_t *)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_u8m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (size_t)(op5)) +#define vsoxseg2ei8_v_u8m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_u8m2((uint8_t *)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_u8m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_u8m2_m((vbool4_t)(op0), (uint8_t *)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei8_v_u8m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_u8m4((uint8_t *)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_u8m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_u8m4_m((vbool2_t)(op0), (uint8_t *)(op1), (vuint8m4_t)(op2), (vuint8m4_t)(op3), (vuint8m4_t)(op4), (size_t)(op5)) +#define vsoxseg2ei8_v_u8mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_u8mf2((uint8_t *)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_u8mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei8_v_u8mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_u8mf4((uint8_t *)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_u8mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (size_t)(op5)) +#define vsoxseg2ei8_v_u8mf8(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_u8mf8((uint8_t *)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_u8mf8_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (size_t)(op5)) +#define vsoxseg3ei8_v_i8m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei8_v_i8m1((int8_t *)(op0), (vuint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (size_t)(op5)) +#define vsoxseg3ei8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei8_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (size_t)(op6)) +#define vsoxseg3ei8_v_i8m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei8_v_i8m2((int8_t *)(op0), (vuint8m2_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei8_v_i8m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei8_v_i8m2_m((vbool4_t)(op0), (int8_t *)(op1), (vuint8m2_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (vint8m2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei8_v_i8mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei8_v_i8mf2((int8_t *)(op0), (vuint8mf2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei8_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei8_v_i8mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei8_v_i8mf4((int8_t *)(op0), (vuint8mf4_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (size_t)(op5)) +#define vsoxseg3ei8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei8_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (size_t)(op6)) +#define vsoxseg3ei8_v_i8mf8(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei8_v_i8mf8((int8_t *)(op0), (vuint8mf8_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (size_t)(op5)) +#define vsoxseg3ei8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei8_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (size_t)(op6)) +#define vsoxseg3ei8_v_u8m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei8_v_u8m1((uint8_t *)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (size_t)(op5)) +#define vsoxseg3ei8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei8_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (size_t)(op6)) +#define vsoxseg3ei8_v_u8m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei8_v_u8m2((uint8_t *)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei8_v_u8m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei8_v_u8m2_m((vbool4_t)(op0), (uint8_t *)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (vuint8m2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei8_v_u8mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei8_v_u8mf2((uint8_t *)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei8_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei8_v_u8mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei8_v_u8mf4((uint8_t *)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (size_t)(op5)) +#define vsoxseg3ei8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei8_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (size_t)(op6)) +#define vsoxseg3ei8_v_u8mf8(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei8_v_u8mf8((uint8_t *)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (size_t)(op5)) +#define vsoxseg3ei8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei8_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (size_t)(op6)) +#define vsoxseg4ei8_v_i8m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei8_v_i8m1((int8_t *)(op0), (vuint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (size_t)(op6)) +#define vsoxseg4ei8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei8_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (size_t)(op7)) +#define vsoxseg4ei8_v_i8m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei8_v_i8m2((int8_t *)(op0), (vuint8m2_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (vint8m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei8_v_i8m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei8_v_i8m2_m((vbool4_t)(op0), (int8_t *)(op1), (vuint8m2_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (vint8m2_t)(op5), (vint8m2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei8_v_i8mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei8_v_i8mf2((int8_t *)(op0), (vuint8mf2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei8_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei8_v_i8mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei8_v_i8mf4((int8_t *)(op0), (vuint8mf4_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (size_t)(op6)) +#define vsoxseg4ei8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei8_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (size_t)(op7)) +#define vsoxseg4ei8_v_i8mf8(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei8_v_i8mf8((int8_t *)(op0), (vuint8mf8_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (size_t)(op6)) +#define vsoxseg4ei8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei8_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (size_t)(op7)) +#define vsoxseg4ei8_v_u8m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei8_v_u8m1((uint8_t *)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (size_t)(op6)) +#define vsoxseg4ei8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei8_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (size_t)(op7)) +#define vsoxseg4ei8_v_u8m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei8_v_u8m2((uint8_t *)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (vuint8m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei8_v_u8m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei8_v_u8m2_m((vbool4_t)(op0), (uint8_t *)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (vuint8m2_t)(op5), (vuint8m2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei8_v_u8mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei8_v_u8mf2((uint8_t *)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei8_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei8_v_u8mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei8_v_u8mf4((uint8_t *)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (size_t)(op6)) +#define vsoxseg4ei8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei8_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vsoxseg4ei8_v_u8mf8(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei8_v_u8mf8((uint8_t *)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (size_t)(op6)) +#define vsoxseg4ei8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei8_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vsoxseg5ei8_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei8_v_i8m1((int8_t *)(op0), (vuint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (size_t)(op7)) +#define vsoxseg5ei8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei8_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (size_t)(op8)) +#define vsoxseg5ei8_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei8_v_i8mf2((int8_t *)(op0), (vuint8mf2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei8_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (size_t)(op8)) +#define vsoxseg5ei8_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei8_v_i8mf4((int8_t *)(op0), (vuint8mf4_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (size_t)(op7)) +#define vsoxseg5ei8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei8_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (size_t)(op8)) +#define vsoxseg5ei8_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei8_v_i8mf8((int8_t *)(op0), (vuint8mf8_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (size_t)(op7)) +#define vsoxseg5ei8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei8_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (size_t)(op8)) +#define vsoxseg5ei8_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei8_v_u8m1((uint8_t *)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (size_t)(op7)) +#define vsoxseg5ei8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei8_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (size_t)(op8)) +#define vsoxseg5ei8_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei8_v_u8mf2((uint8_t *)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei8_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (size_t)(op8)) +#define vsoxseg5ei8_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei8_v_u8mf4((uint8_t *)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vsoxseg5ei8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei8_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (size_t)(op8)) +#define vsoxseg5ei8_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei8_v_u8mf8((uint8_t *)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vsoxseg5ei8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei8_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (size_t)(op8)) +#define vsoxseg6ei8_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei8_v_i8m1((int8_t *)(op0), (vuint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei8_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (size_t)(op9)) +#define vsoxseg6ei8_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei8_v_i8mf2((int8_t *)(op0), (vuint8mf2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (size_t)(op8)) +#define vsoxseg6ei8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei8_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (size_t)(op9)) +#define vsoxseg6ei8_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei8_v_i8mf4((int8_t *)(op0), (vuint8mf4_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (size_t)(op8)) +#define vsoxseg6ei8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei8_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (size_t)(op9)) +#define vsoxseg6ei8_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei8_v_i8mf8((int8_t *)(op0), (vuint8mf8_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (size_t)(op8)) +#define vsoxseg6ei8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei8_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (size_t)(op9)) +#define vsoxseg6ei8_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei8_v_u8m1((uint8_t *)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei8_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (size_t)(op9)) +#define vsoxseg6ei8_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei8_v_u8mf2((uint8_t *)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (size_t)(op8)) +#define vsoxseg6ei8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei8_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vsoxseg6ei8_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei8_v_u8mf4((uint8_t *)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (size_t)(op8)) +#define vsoxseg6ei8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei8_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vsoxseg6ei8_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei8_v_u8mf8((uint8_t *)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (size_t)(op8)) +#define vsoxseg6ei8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei8_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vsoxseg7ei8_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei8_v_i8m1((int8_t *)(op0), (vuint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei8_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (size_t)(op10)) +#define vsoxseg7ei8_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei8_v_i8mf2((int8_t *)(op0), (vuint8mf2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (size_t)(op9)) +#define vsoxseg7ei8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei8_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (size_t)(op10)) +#define vsoxseg7ei8_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei8_v_i8mf4((int8_t *)(op0), (vuint8mf4_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (size_t)(op9)) +#define vsoxseg7ei8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei8_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (size_t)(op10)) +#define vsoxseg7ei8_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei8_v_i8mf8((int8_t *)(op0), (vuint8mf8_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (size_t)(op9)) +#define vsoxseg7ei8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei8_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (size_t)(op10)) +#define vsoxseg7ei8_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei8_v_u8m1((uint8_t *)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei8_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (size_t)(op10)) +#define vsoxseg7ei8_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei8_v_u8mf2((uint8_t *)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vsoxseg7ei8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei8_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (size_t)(op10)) +#define vsoxseg7ei8_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei8_v_u8mf4((uint8_t *)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vsoxseg7ei8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei8_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (size_t)(op10)) +#define vsoxseg7ei8_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei8_v_u8mf8((uint8_t *)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vsoxseg7ei8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei8_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (size_t)(op10)) +#define vsoxseg8ei8_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei8_v_i8m1((int8_t *)(op0), (vuint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei8_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (size_t)(op11)) +#define vsoxseg8ei8_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei8_v_i8mf2((int8_t *)(op0), (vuint8mf2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (size_t)(op10)) +#define vsoxseg8ei8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei8_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (size_t)(op11)) +#define vsoxseg8ei8_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei8_v_i8mf4((int8_t *)(op0), (vuint8mf4_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (size_t)(op10)) +#define vsoxseg8ei8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei8_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (size_t)(op11)) +#define vsoxseg8ei8_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei8_v_i8mf8((int8_t *)(op0), (vuint8mf8_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (size_t)(op10)) +#define vsoxseg8ei8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei8_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (size_t)(op11)) +#define vsoxseg8ei8_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei8_v_u8m1((uint8_t *)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei8_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (size_t)(op11)) +#define vsoxseg8ei8_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei8_v_u8mf2((uint8_t *)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (size_t)(op10)) +#define vsoxseg8ei8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei8_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (size_t)(op11)) +#define vsoxseg8ei8_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei8_v_u8mf4((uint8_t *)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (size_t)(op10)) +#define vsoxseg8ei8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei8_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (size_t)(op11)) +#define vsoxseg8ei8_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei8_v_u8mf8((uint8_t *)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (size_t)(op10)) +#define vsoxseg8ei8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei8_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (size_t)(op11)) +#define vsoxseg2ei16_v_i8m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_i8m1((int8_t *)(op0), (vuint16m2_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_i8m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint16m2_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (size_t)(op5)) +#define vsoxseg2ei16_v_i8m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_i8m2((int8_t *)(op0), (vuint16m4_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_i8m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_i8m2_m((vbool4_t)(op0), (int8_t *)(op1), (vuint16m4_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei16_v_i8m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_i8m4((int8_t *)(op0), (vuint16m8_t)(op1), (vint8m4_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_i8m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_i8m4_m((vbool2_t)(op0), (int8_t *)(op1), (vuint16m8_t)(op2), (vint8m4_t)(op3), (vint8m4_t)(op4), (size_t)(op5)) +#define vsoxseg2ei16_v_i8mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_i8mf2((int8_t *)(op0), (vuint16m1_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_i8mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint16m1_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei16_v_i8mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_i8mf4((int8_t *)(op0), (vuint16mf2_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_i8mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint16mf2_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (size_t)(op5)) +#define vsoxseg2ei16_v_i8mf8(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_i8mf8((int8_t *)(op0), (vuint16mf4_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_i8mf8_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint16mf4_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (size_t)(op5)) +#define vsoxseg2ei16_v_u8m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_u8m1((uint8_t *)(op0), (vuint16m2_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_u8m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint16m2_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (size_t)(op5)) +#define vsoxseg2ei16_v_u8m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_u8m2((uint8_t *)(op0), (vuint16m4_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_u8m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_u8m2_m((vbool4_t)(op0), (uint8_t *)(op1), (vuint16m4_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei16_v_u8m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_u8m4((uint8_t *)(op0), (vuint16m8_t)(op1), (vuint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_u8m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_u8m4_m((vbool2_t)(op0), (uint8_t *)(op1), (vuint16m8_t)(op2), (vuint8m4_t)(op3), (vuint8m4_t)(op4), (size_t)(op5)) +#define vsoxseg2ei16_v_u8mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_u8mf2((uint8_t *)(op0), (vuint16m1_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_u8mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint16m1_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei16_v_u8mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_u8mf4((uint8_t *)(op0), (vuint16mf2_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_u8mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint16mf2_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (size_t)(op5)) +#define vsoxseg2ei16_v_u8mf8(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_u8mf8((uint8_t *)(op0), (vuint16mf4_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_u8mf8_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint16mf4_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (size_t)(op5)) +#define vsoxseg3ei16_v_i8m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei16_v_i8m1((int8_t *)(op0), (vuint16m2_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (size_t)(op5)) +#define vsoxseg3ei16_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei16_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint16m2_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (size_t)(op6)) +#define vsoxseg3ei16_v_i8m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei16_v_i8m2((int8_t *)(op0), (vuint16m4_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei16_v_i8m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei16_v_i8m2_m((vbool4_t)(op0), (int8_t *)(op1), (vuint16m4_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (vint8m2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei16_v_i8mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei16_v_i8mf2((int8_t *)(op0), (vuint16m1_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei16_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei16_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint16m1_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei16_v_i8mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei16_v_i8mf4((int8_t *)(op0), (vuint16mf2_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (size_t)(op5)) +#define vsoxseg3ei16_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei16_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint16mf2_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (size_t)(op6)) +#define vsoxseg3ei16_v_i8mf8(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei16_v_i8mf8((int8_t *)(op0), (vuint16mf4_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (size_t)(op5)) +#define vsoxseg3ei16_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei16_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint16mf4_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (size_t)(op6)) +#define vsoxseg3ei16_v_u8m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei16_v_u8m1((uint8_t *)(op0), (vuint16m2_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (size_t)(op5)) +#define vsoxseg3ei16_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei16_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint16m2_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (size_t)(op6)) +#define vsoxseg3ei16_v_u8m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei16_v_u8m2((uint8_t *)(op0), (vuint16m4_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei16_v_u8m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei16_v_u8m2_m((vbool4_t)(op0), (uint8_t *)(op1), (vuint16m4_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (vuint8m2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei16_v_u8mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei16_v_u8mf2((uint8_t *)(op0), (vuint16m1_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei16_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei16_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint16m1_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei16_v_u8mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei16_v_u8mf4((uint8_t *)(op0), (vuint16mf2_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (size_t)(op5)) +#define vsoxseg3ei16_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei16_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint16mf2_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (size_t)(op6)) +#define vsoxseg3ei16_v_u8mf8(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei16_v_u8mf8((uint8_t *)(op0), (vuint16mf4_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (size_t)(op5)) +#define vsoxseg3ei16_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei16_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint16mf4_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (size_t)(op6)) +#define vsoxseg4ei16_v_i8m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei16_v_i8m1((int8_t *)(op0), (vuint16m2_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (size_t)(op6)) +#define vsoxseg4ei16_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei16_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint16m2_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (size_t)(op7)) +#define vsoxseg4ei16_v_i8m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei16_v_i8m2((int8_t *)(op0), (vuint16m4_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (vint8m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei16_v_i8m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei16_v_i8m2_m((vbool4_t)(op0), (int8_t *)(op1), (vuint16m4_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (vint8m2_t)(op5), (vint8m2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei16_v_i8mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei16_v_i8mf2((int8_t *)(op0), (vuint16m1_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei16_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei16_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint16m1_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei16_v_i8mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei16_v_i8mf4((int8_t *)(op0), (vuint16mf2_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (size_t)(op6)) +#define vsoxseg4ei16_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei16_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint16mf2_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (size_t)(op7)) +#define vsoxseg4ei16_v_i8mf8(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei16_v_i8mf8((int8_t *)(op0), (vuint16mf4_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (size_t)(op6)) +#define vsoxseg4ei16_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei16_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint16mf4_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (size_t)(op7)) +#define vsoxseg4ei16_v_u8m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei16_v_u8m1((uint8_t *)(op0), (vuint16m2_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (size_t)(op6)) +#define vsoxseg4ei16_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei16_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint16m2_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (size_t)(op7)) +#define vsoxseg4ei16_v_u8m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei16_v_u8m2((uint8_t *)(op0), (vuint16m4_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (vuint8m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei16_v_u8m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei16_v_u8m2_m((vbool4_t)(op0), (uint8_t *)(op1), (vuint16m4_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (vuint8m2_t)(op5), (vuint8m2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei16_v_u8mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei16_v_u8mf2((uint8_t *)(op0), (vuint16m1_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei16_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei16_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint16m1_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei16_v_u8mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei16_v_u8mf4((uint8_t *)(op0), (vuint16mf2_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (size_t)(op6)) +#define vsoxseg4ei16_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei16_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint16mf2_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vsoxseg4ei16_v_u8mf8(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei16_v_u8mf8((uint8_t *)(op0), (vuint16mf4_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (size_t)(op6)) +#define vsoxseg4ei16_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei16_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint16mf4_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vsoxseg5ei16_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei16_v_i8m1((int8_t *)(op0), (vuint16m2_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (size_t)(op7)) +#define vsoxseg5ei16_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei16_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint16m2_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (size_t)(op8)) +#define vsoxseg5ei16_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei16_v_i8mf2((int8_t *)(op0), (vuint16m1_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei16_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei16_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint16m1_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (size_t)(op8)) +#define vsoxseg5ei16_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei16_v_i8mf4((int8_t *)(op0), (vuint16mf2_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (size_t)(op7)) +#define vsoxseg5ei16_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei16_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint16mf2_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (size_t)(op8)) +#define vsoxseg5ei16_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei16_v_i8mf8((int8_t *)(op0), (vuint16mf4_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (size_t)(op7)) +#define vsoxseg5ei16_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei16_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint16mf4_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (size_t)(op8)) +#define vsoxseg5ei16_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei16_v_u8m1((uint8_t *)(op0), (vuint16m2_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (size_t)(op7)) +#define vsoxseg5ei16_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei16_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint16m2_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (size_t)(op8)) +#define vsoxseg5ei16_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei16_v_u8mf2((uint8_t *)(op0), (vuint16m1_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei16_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei16_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint16m1_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (size_t)(op8)) +#define vsoxseg5ei16_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei16_v_u8mf4((uint8_t *)(op0), (vuint16mf2_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vsoxseg5ei16_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei16_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint16mf2_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (size_t)(op8)) +#define vsoxseg5ei16_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei16_v_u8mf8((uint8_t *)(op0), (vuint16mf4_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vsoxseg5ei16_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei16_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint16mf4_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (size_t)(op8)) +#define vsoxseg6ei16_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei16_v_i8m1((int8_t *)(op0), (vuint16m2_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei16_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei16_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint16m2_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (size_t)(op9)) +#define vsoxseg6ei16_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei16_v_i8mf2((int8_t *)(op0), (vuint16m1_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (size_t)(op8)) +#define vsoxseg6ei16_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei16_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint16m1_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (size_t)(op9)) +#define vsoxseg6ei16_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei16_v_i8mf4((int8_t *)(op0), (vuint16mf2_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (size_t)(op8)) +#define vsoxseg6ei16_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei16_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint16mf2_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (size_t)(op9)) +#define vsoxseg6ei16_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei16_v_i8mf8((int8_t *)(op0), (vuint16mf4_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (size_t)(op8)) +#define vsoxseg6ei16_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei16_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint16mf4_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (size_t)(op9)) +#define vsoxseg6ei16_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei16_v_u8m1((uint8_t *)(op0), (vuint16m2_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei16_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei16_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint16m2_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (size_t)(op9)) +#define vsoxseg6ei16_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei16_v_u8mf2((uint8_t *)(op0), (vuint16m1_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (size_t)(op8)) +#define vsoxseg6ei16_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei16_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint16m1_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vsoxseg6ei16_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei16_v_u8mf4((uint8_t *)(op0), (vuint16mf2_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (size_t)(op8)) +#define vsoxseg6ei16_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei16_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint16mf2_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vsoxseg6ei16_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei16_v_u8mf8((uint8_t *)(op0), (vuint16mf4_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (size_t)(op8)) +#define vsoxseg6ei16_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei16_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint16mf4_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vsoxseg7ei16_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei16_v_i8m1((int8_t *)(op0), (vuint16m2_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei16_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei16_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint16m2_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (size_t)(op10)) +#define vsoxseg7ei16_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei16_v_i8mf2((int8_t *)(op0), (vuint16m1_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (size_t)(op9)) +#define vsoxseg7ei16_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei16_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint16m1_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (size_t)(op10)) +#define vsoxseg7ei16_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei16_v_i8mf4((int8_t *)(op0), (vuint16mf2_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (size_t)(op9)) +#define vsoxseg7ei16_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei16_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint16mf2_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (size_t)(op10)) +#define vsoxseg7ei16_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei16_v_i8mf8((int8_t *)(op0), (vuint16mf4_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (size_t)(op9)) +#define vsoxseg7ei16_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei16_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint16mf4_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (size_t)(op10)) +#define vsoxseg7ei16_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei16_v_u8m1((uint8_t *)(op0), (vuint16m2_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei16_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei16_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint16m2_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (size_t)(op10)) +#define vsoxseg7ei16_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei16_v_u8mf2((uint8_t *)(op0), (vuint16m1_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vsoxseg7ei16_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei16_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint16m1_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (size_t)(op10)) +#define vsoxseg7ei16_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei16_v_u8mf4((uint8_t *)(op0), (vuint16mf2_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vsoxseg7ei16_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei16_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint16mf2_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (size_t)(op10)) +#define vsoxseg7ei16_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei16_v_u8mf8((uint8_t *)(op0), (vuint16mf4_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vsoxseg7ei16_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei16_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint16mf4_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (size_t)(op10)) +#define vsoxseg8ei16_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei16_v_i8m1((int8_t *)(op0), (vuint16m2_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei16_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei16_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint16m2_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (size_t)(op11)) +#define vsoxseg8ei16_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei16_v_i8mf2((int8_t *)(op0), (vuint16m1_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (size_t)(op10)) +#define vsoxseg8ei16_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei16_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint16m1_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (size_t)(op11)) +#define vsoxseg8ei16_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei16_v_i8mf4((int8_t *)(op0), (vuint16mf2_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (size_t)(op10)) +#define vsoxseg8ei16_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei16_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint16mf2_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (size_t)(op11)) +#define vsoxseg8ei16_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei16_v_i8mf8((int8_t *)(op0), (vuint16mf4_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (size_t)(op10)) +#define vsoxseg8ei16_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei16_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint16mf4_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (size_t)(op11)) +#define vsoxseg8ei16_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei16_v_u8m1((uint8_t *)(op0), (vuint16m2_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei16_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei16_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint16m2_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (size_t)(op11)) +#define vsoxseg8ei16_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei16_v_u8mf2((uint8_t *)(op0), (vuint16m1_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (size_t)(op10)) +#define vsoxseg8ei16_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei16_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint16m1_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (size_t)(op11)) +#define vsoxseg8ei16_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei16_v_u8mf4((uint8_t *)(op0), (vuint16mf2_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (size_t)(op10)) +#define vsoxseg8ei16_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei16_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint16mf2_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (size_t)(op11)) +#define vsoxseg8ei16_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei16_v_u8mf8((uint8_t *)(op0), (vuint16mf4_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (size_t)(op10)) +#define vsoxseg8ei16_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei16_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint16mf4_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (size_t)(op11)) +#define vsoxseg2ei32_v_i8m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_i8m1((int8_t *)(op0), (vuint32m4_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_i8m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint32m4_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (size_t)(op5)) +#define vsoxseg2ei32_v_i8m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_i8m2((int8_t *)(op0), (vuint32m8_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_i8m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_i8m2_m((vbool4_t)(op0), (int8_t *)(op1), (vuint32m8_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei32_v_i8mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_i8mf2((int8_t *)(op0), (vuint32m2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_i8mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint32m2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei32_v_i8mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_i8mf4((int8_t *)(op0), (vuint32m1_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_i8mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint32m1_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (size_t)(op5)) +#define vsoxseg2ei32_v_i8mf8(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_i8mf8((int8_t *)(op0), (vuint32mf2_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_i8mf8_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint32mf2_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (size_t)(op5)) +#define vsoxseg2ei32_v_u8m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_u8m1((uint8_t *)(op0), (vuint32m4_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_u8m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint32m4_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (size_t)(op5)) +#define vsoxseg2ei32_v_u8m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_u8m2((uint8_t *)(op0), (vuint32m8_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_u8m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_u8m2_m((vbool4_t)(op0), (uint8_t *)(op1), (vuint32m8_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei32_v_u8mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_u8mf2((uint8_t *)(op0), (vuint32m2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_u8mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint32m2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei32_v_u8mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_u8mf4((uint8_t *)(op0), (vuint32m1_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_u8mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint32m1_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (size_t)(op5)) +#define vsoxseg2ei32_v_u8mf8(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_u8mf8((uint8_t *)(op0), (vuint32mf2_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_u8mf8_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint32mf2_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (size_t)(op5)) +#define vsoxseg3ei32_v_i8m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei32_v_i8m1((int8_t *)(op0), (vuint32m4_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (size_t)(op5)) +#define vsoxseg3ei32_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei32_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint32m4_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (size_t)(op6)) +#define vsoxseg3ei32_v_i8m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei32_v_i8m2((int8_t *)(op0), (vuint32m8_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei32_v_i8m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei32_v_i8m2_m((vbool4_t)(op0), (int8_t *)(op1), (vuint32m8_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (vint8m2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei32_v_i8mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei32_v_i8mf2((int8_t *)(op0), (vuint32m2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei32_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei32_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint32m2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei32_v_i8mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei32_v_i8mf4((int8_t *)(op0), (vuint32m1_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (size_t)(op5)) +#define vsoxseg3ei32_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei32_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint32m1_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (size_t)(op6)) +#define vsoxseg3ei32_v_i8mf8(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei32_v_i8mf8((int8_t *)(op0), (vuint32mf2_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (size_t)(op5)) +#define vsoxseg3ei32_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei32_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint32mf2_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (size_t)(op6)) +#define vsoxseg3ei32_v_u8m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei32_v_u8m1((uint8_t *)(op0), (vuint32m4_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (size_t)(op5)) +#define vsoxseg3ei32_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei32_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint32m4_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (size_t)(op6)) +#define vsoxseg3ei32_v_u8m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei32_v_u8m2((uint8_t *)(op0), (vuint32m8_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei32_v_u8m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei32_v_u8m2_m((vbool4_t)(op0), (uint8_t *)(op1), (vuint32m8_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (vuint8m2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei32_v_u8mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei32_v_u8mf2((uint8_t *)(op0), (vuint32m2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei32_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei32_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint32m2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei32_v_u8mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei32_v_u8mf4((uint8_t *)(op0), (vuint32m1_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (size_t)(op5)) +#define vsoxseg3ei32_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei32_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint32m1_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (size_t)(op6)) +#define vsoxseg3ei32_v_u8mf8(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei32_v_u8mf8((uint8_t *)(op0), (vuint32mf2_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (size_t)(op5)) +#define vsoxseg3ei32_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei32_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint32mf2_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (size_t)(op6)) +#define vsoxseg4ei32_v_i8m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei32_v_i8m1((int8_t *)(op0), (vuint32m4_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (size_t)(op6)) +#define vsoxseg4ei32_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei32_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint32m4_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (size_t)(op7)) +#define vsoxseg4ei32_v_i8m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei32_v_i8m2((int8_t *)(op0), (vuint32m8_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (vint8m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei32_v_i8m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei32_v_i8m2_m((vbool4_t)(op0), (int8_t *)(op1), (vuint32m8_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (vint8m2_t)(op5), (vint8m2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei32_v_i8mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei32_v_i8mf2((int8_t *)(op0), (vuint32m2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei32_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei32_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint32m2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei32_v_i8mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei32_v_i8mf4((int8_t *)(op0), (vuint32m1_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (size_t)(op6)) +#define vsoxseg4ei32_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei32_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint32m1_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (size_t)(op7)) +#define vsoxseg4ei32_v_i8mf8(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei32_v_i8mf8((int8_t *)(op0), (vuint32mf2_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (size_t)(op6)) +#define vsoxseg4ei32_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei32_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint32mf2_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (size_t)(op7)) +#define vsoxseg4ei32_v_u8m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei32_v_u8m1((uint8_t *)(op0), (vuint32m4_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (size_t)(op6)) +#define vsoxseg4ei32_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei32_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint32m4_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (size_t)(op7)) +#define vsoxseg4ei32_v_u8m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei32_v_u8m2((uint8_t *)(op0), (vuint32m8_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (vuint8m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei32_v_u8m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei32_v_u8m2_m((vbool4_t)(op0), (uint8_t *)(op1), (vuint32m8_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (vuint8m2_t)(op5), (vuint8m2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei32_v_u8mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei32_v_u8mf2((uint8_t *)(op0), (vuint32m2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei32_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei32_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint32m2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei32_v_u8mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei32_v_u8mf4((uint8_t *)(op0), (vuint32m1_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (size_t)(op6)) +#define vsoxseg4ei32_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei32_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint32m1_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vsoxseg4ei32_v_u8mf8(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei32_v_u8mf8((uint8_t *)(op0), (vuint32mf2_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (size_t)(op6)) +#define vsoxseg4ei32_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei32_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint32mf2_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vsoxseg5ei32_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei32_v_i8m1((int8_t *)(op0), (vuint32m4_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (size_t)(op7)) +#define vsoxseg5ei32_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei32_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint32m4_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (size_t)(op8)) +#define vsoxseg5ei32_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei32_v_i8mf2((int8_t *)(op0), (vuint32m2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei32_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei32_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint32m2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (size_t)(op8)) +#define vsoxseg5ei32_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei32_v_i8mf4((int8_t *)(op0), (vuint32m1_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (size_t)(op7)) +#define vsoxseg5ei32_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei32_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint32m1_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (size_t)(op8)) +#define vsoxseg5ei32_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei32_v_i8mf8((int8_t *)(op0), (vuint32mf2_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (size_t)(op7)) +#define vsoxseg5ei32_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei32_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint32mf2_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (size_t)(op8)) +#define vsoxseg5ei32_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei32_v_u8m1((uint8_t *)(op0), (vuint32m4_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (size_t)(op7)) +#define vsoxseg5ei32_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei32_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint32m4_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (size_t)(op8)) +#define vsoxseg5ei32_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei32_v_u8mf2((uint8_t *)(op0), (vuint32m2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei32_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei32_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint32m2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (size_t)(op8)) +#define vsoxseg5ei32_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei32_v_u8mf4((uint8_t *)(op0), (vuint32m1_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vsoxseg5ei32_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei32_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint32m1_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (size_t)(op8)) +#define vsoxseg5ei32_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei32_v_u8mf8((uint8_t *)(op0), (vuint32mf2_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vsoxseg5ei32_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei32_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint32mf2_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (size_t)(op8)) +#define vsoxseg6ei32_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei32_v_i8m1((int8_t *)(op0), (vuint32m4_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei32_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei32_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint32m4_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (size_t)(op9)) +#define vsoxseg6ei32_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei32_v_i8mf2((int8_t *)(op0), (vuint32m2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (size_t)(op8)) +#define vsoxseg6ei32_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei32_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint32m2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (size_t)(op9)) +#define vsoxseg6ei32_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei32_v_i8mf4((int8_t *)(op0), (vuint32m1_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (size_t)(op8)) +#define vsoxseg6ei32_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei32_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint32m1_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (size_t)(op9)) +#define vsoxseg6ei32_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei32_v_i8mf8((int8_t *)(op0), (vuint32mf2_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (size_t)(op8)) +#define vsoxseg6ei32_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei32_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint32mf2_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (size_t)(op9)) +#define vsoxseg6ei32_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei32_v_u8m1((uint8_t *)(op0), (vuint32m4_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei32_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei32_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint32m4_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (size_t)(op9)) +#define vsoxseg6ei32_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei32_v_u8mf2((uint8_t *)(op0), (vuint32m2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (size_t)(op8)) +#define vsoxseg6ei32_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei32_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint32m2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vsoxseg6ei32_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei32_v_u8mf4((uint8_t *)(op0), (vuint32m1_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (size_t)(op8)) +#define vsoxseg6ei32_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei32_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint32m1_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vsoxseg6ei32_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei32_v_u8mf8((uint8_t *)(op0), (vuint32mf2_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (size_t)(op8)) +#define vsoxseg6ei32_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei32_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint32mf2_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vsoxseg7ei32_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei32_v_i8m1((int8_t *)(op0), (vuint32m4_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei32_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei32_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint32m4_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (size_t)(op10)) +#define vsoxseg7ei32_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei32_v_i8mf2((int8_t *)(op0), (vuint32m2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (size_t)(op9)) +#define vsoxseg7ei32_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei32_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint32m2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (size_t)(op10)) +#define vsoxseg7ei32_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei32_v_i8mf4((int8_t *)(op0), (vuint32m1_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (size_t)(op9)) +#define vsoxseg7ei32_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei32_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint32m1_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (size_t)(op10)) +#define vsoxseg7ei32_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei32_v_i8mf8((int8_t *)(op0), (vuint32mf2_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (size_t)(op9)) +#define vsoxseg7ei32_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei32_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint32mf2_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (size_t)(op10)) +#define vsoxseg7ei32_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei32_v_u8m1((uint8_t *)(op0), (vuint32m4_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei32_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei32_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint32m4_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (size_t)(op10)) +#define vsoxseg7ei32_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei32_v_u8mf2((uint8_t *)(op0), (vuint32m2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vsoxseg7ei32_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei32_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint32m2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (size_t)(op10)) +#define vsoxseg7ei32_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei32_v_u8mf4((uint8_t *)(op0), (vuint32m1_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vsoxseg7ei32_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei32_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint32m1_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (size_t)(op10)) +#define vsoxseg7ei32_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei32_v_u8mf8((uint8_t *)(op0), (vuint32mf2_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vsoxseg7ei32_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei32_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint32mf2_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (size_t)(op10)) +#define vsoxseg8ei32_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei32_v_i8m1((int8_t *)(op0), (vuint32m4_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei32_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei32_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint32m4_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (size_t)(op11)) +#define vsoxseg8ei32_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei32_v_i8mf2((int8_t *)(op0), (vuint32m2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (size_t)(op10)) +#define vsoxseg8ei32_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei32_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint32m2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (size_t)(op11)) +#define vsoxseg8ei32_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei32_v_i8mf4((int8_t *)(op0), (vuint32m1_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (size_t)(op10)) +#define vsoxseg8ei32_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei32_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint32m1_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (size_t)(op11)) +#define vsoxseg8ei32_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei32_v_i8mf8((int8_t *)(op0), (vuint32mf2_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (size_t)(op10)) +#define vsoxseg8ei32_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei32_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint32mf2_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (size_t)(op11)) +#define vsoxseg8ei32_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei32_v_u8m1((uint8_t *)(op0), (vuint32m4_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei32_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei32_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint32m4_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (size_t)(op11)) +#define vsoxseg8ei32_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei32_v_u8mf2((uint8_t *)(op0), (vuint32m2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (size_t)(op10)) +#define vsoxseg8ei32_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei32_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint32m2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (size_t)(op11)) +#define vsoxseg8ei32_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei32_v_u8mf4((uint8_t *)(op0), (vuint32m1_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (size_t)(op10)) +#define vsoxseg8ei32_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei32_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint32m1_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (size_t)(op11)) +#define vsoxseg8ei32_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei32_v_u8mf8((uint8_t *)(op0), (vuint32mf2_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (size_t)(op10)) +#define vsoxseg8ei32_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei32_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint32mf2_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (size_t)(op11)) +#define vsoxseg2ei64_v_i8m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei64_v_i8m1((int8_t *)(op0), (vuint64m8_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vsoxseg2ei64_v_i8m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei64_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint64m8_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (size_t)(op5)) +#define vsoxseg2ei64_v_i8mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei64_v_i8mf2((int8_t *)(op0), (vuint64m4_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei64_v_i8mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei64_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint64m4_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei64_v_i8mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei64_v_i8mf4((int8_t *)(op0), (vuint64m2_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei64_v_i8mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei64_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint64m2_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (size_t)(op5)) +#define vsoxseg2ei64_v_i8mf8(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei64_v_i8mf8((int8_t *)(op0), (vuint64m1_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vsoxseg2ei64_v_i8mf8_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei64_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint64m1_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (size_t)(op5)) +#define vsoxseg2ei64_v_u8m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei64_v_u8m1((uint8_t *)(op0), (vuint64m8_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vsoxseg2ei64_v_u8m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei64_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint64m8_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (size_t)(op5)) +#define vsoxseg2ei64_v_u8mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei64_v_u8mf2((uint8_t *)(op0), (vuint64m4_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei64_v_u8mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei64_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint64m4_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei64_v_u8mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei64_v_u8mf4((uint8_t *)(op0), (vuint64m2_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei64_v_u8mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei64_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint64m2_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (size_t)(op5)) +#define vsoxseg2ei64_v_u8mf8(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei64_v_u8mf8((uint8_t *)(op0), (vuint64m1_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vsoxseg2ei64_v_u8mf8_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei64_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint64m1_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (size_t)(op5)) +#define vsoxseg3ei64_v_i8m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei64_v_i8m1((int8_t *)(op0), (vuint64m8_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (size_t)(op5)) +#define vsoxseg3ei64_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei64_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint64m8_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (size_t)(op6)) +#define vsoxseg3ei64_v_i8mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei64_v_i8mf2((int8_t *)(op0), (vuint64m4_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei64_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei64_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint64m4_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei64_v_i8mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei64_v_i8mf4((int8_t *)(op0), (vuint64m2_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (size_t)(op5)) +#define vsoxseg3ei64_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei64_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint64m2_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (size_t)(op6)) +#define vsoxseg3ei64_v_i8mf8(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei64_v_i8mf8((int8_t *)(op0), (vuint64m1_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (size_t)(op5)) +#define vsoxseg3ei64_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei64_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint64m1_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (size_t)(op6)) +#define vsoxseg3ei64_v_u8m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei64_v_u8m1((uint8_t *)(op0), (vuint64m8_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (size_t)(op5)) +#define vsoxseg3ei64_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei64_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint64m8_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (size_t)(op6)) +#define vsoxseg3ei64_v_u8mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei64_v_u8mf2((uint8_t *)(op0), (vuint64m4_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei64_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei64_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint64m4_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei64_v_u8mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei64_v_u8mf4((uint8_t *)(op0), (vuint64m2_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (size_t)(op5)) +#define vsoxseg3ei64_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei64_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint64m2_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (size_t)(op6)) +#define vsoxseg3ei64_v_u8mf8(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei64_v_u8mf8((uint8_t *)(op0), (vuint64m1_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (size_t)(op5)) +#define vsoxseg3ei64_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei64_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint64m1_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (size_t)(op6)) +#define vsoxseg4ei64_v_i8m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei64_v_i8m1((int8_t *)(op0), (vuint64m8_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (size_t)(op6)) +#define vsoxseg4ei64_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei64_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint64m8_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (size_t)(op7)) +#define vsoxseg4ei64_v_i8mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei64_v_i8mf2((int8_t *)(op0), (vuint64m4_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei64_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei64_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint64m4_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei64_v_i8mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei64_v_i8mf4((int8_t *)(op0), (vuint64m2_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (size_t)(op6)) +#define vsoxseg4ei64_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei64_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint64m2_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (size_t)(op7)) +#define vsoxseg4ei64_v_i8mf8(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei64_v_i8mf8((int8_t *)(op0), (vuint64m1_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (size_t)(op6)) +#define vsoxseg4ei64_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei64_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint64m1_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (size_t)(op7)) +#define vsoxseg4ei64_v_u8m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei64_v_u8m1((uint8_t *)(op0), (vuint64m8_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (size_t)(op6)) +#define vsoxseg4ei64_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei64_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint64m8_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (size_t)(op7)) +#define vsoxseg4ei64_v_u8mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei64_v_u8mf2((uint8_t *)(op0), (vuint64m4_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei64_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei64_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint64m4_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei64_v_u8mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei64_v_u8mf4((uint8_t *)(op0), (vuint64m2_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (size_t)(op6)) +#define vsoxseg4ei64_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei64_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint64m2_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vsoxseg4ei64_v_u8mf8(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei64_v_u8mf8((uint8_t *)(op0), (vuint64m1_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (size_t)(op6)) +#define vsoxseg4ei64_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei64_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint64m1_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vsoxseg5ei64_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei64_v_i8m1((int8_t *)(op0), (vuint64m8_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (size_t)(op7)) +#define vsoxseg5ei64_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei64_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint64m8_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (size_t)(op8)) +#define vsoxseg5ei64_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei64_v_i8mf2((int8_t *)(op0), (vuint64m4_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei64_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei64_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint64m4_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (size_t)(op8)) +#define vsoxseg5ei64_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei64_v_i8mf4((int8_t *)(op0), (vuint64m2_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (size_t)(op7)) +#define vsoxseg5ei64_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei64_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint64m2_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (size_t)(op8)) +#define vsoxseg5ei64_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei64_v_i8mf8((int8_t *)(op0), (vuint64m1_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (size_t)(op7)) +#define vsoxseg5ei64_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei64_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint64m1_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (size_t)(op8)) +#define vsoxseg5ei64_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei64_v_u8m1((uint8_t *)(op0), (vuint64m8_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (size_t)(op7)) +#define vsoxseg5ei64_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei64_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint64m8_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (size_t)(op8)) +#define vsoxseg5ei64_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei64_v_u8mf2((uint8_t *)(op0), (vuint64m4_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei64_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei64_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint64m4_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (size_t)(op8)) +#define vsoxseg5ei64_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei64_v_u8mf4((uint8_t *)(op0), (vuint64m2_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vsoxseg5ei64_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei64_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint64m2_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (size_t)(op8)) +#define vsoxseg5ei64_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei64_v_u8mf8((uint8_t *)(op0), (vuint64m1_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vsoxseg5ei64_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei64_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint64m1_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (size_t)(op8)) +#define vsoxseg6ei64_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei64_v_i8m1((int8_t *)(op0), (vuint64m8_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei64_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei64_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint64m8_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (size_t)(op9)) +#define vsoxseg6ei64_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei64_v_i8mf2((int8_t *)(op0), (vuint64m4_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (size_t)(op8)) +#define vsoxseg6ei64_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei64_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint64m4_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (size_t)(op9)) +#define vsoxseg6ei64_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei64_v_i8mf4((int8_t *)(op0), (vuint64m2_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (size_t)(op8)) +#define vsoxseg6ei64_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei64_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint64m2_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (size_t)(op9)) +#define vsoxseg6ei64_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei64_v_i8mf8((int8_t *)(op0), (vuint64m1_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (size_t)(op8)) +#define vsoxseg6ei64_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei64_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint64m1_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (size_t)(op9)) +#define vsoxseg6ei64_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei64_v_u8m1((uint8_t *)(op0), (vuint64m8_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei64_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei64_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint64m8_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (size_t)(op9)) +#define vsoxseg6ei64_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei64_v_u8mf2((uint8_t *)(op0), (vuint64m4_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (size_t)(op8)) +#define vsoxseg6ei64_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei64_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint64m4_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vsoxseg6ei64_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei64_v_u8mf4((uint8_t *)(op0), (vuint64m2_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (size_t)(op8)) +#define vsoxseg6ei64_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei64_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint64m2_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vsoxseg6ei64_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei64_v_u8mf8((uint8_t *)(op0), (vuint64m1_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (size_t)(op8)) +#define vsoxseg6ei64_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei64_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint64m1_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vsoxseg7ei64_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei64_v_i8m1((int8_t *)(op0), (vuint64m8_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei64_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei64_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint64m8_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (size_t)(op10)) +#define vsoxseg7ei64_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei64_v_i8mf2((int8_t *)(op0), (vuint64m4_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (size_t)(op9)) +#define vsoxseg7ei64_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei64_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint64m4_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (size_t)(op10)) +#define vsoxseg7ei64_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei64_v_i8mf4((int8_t *)(op0), (vuint64m2_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (size_t)(op9)) +#define vsoxseg7ei64_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei64_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint64m2_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (size_t)(op10)) +#define vsoxseg7ei64_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei64_v_i8mf8((int8_t *)(op0), (vuint64m1_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (size_t)(op9)) +#define vsoxseg7ei64_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei64_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint64m1_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (size_t)(op10)) +#define vsoxseg7ei64_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei64_v_u8m1((uint8_t *)(op0), (vuint64m8_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei64_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei64_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint64m8_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (size_t)(op10)) +#define vsoxseg7ei64_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei64_v_u8mf2((uint8_t *)(op0), (vuint64m4_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vsoxseg7ei64_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei64_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint64m4_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (size_t)(op10)) +#define vsoxseg7ei64_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei64_v_u8mf4((uint8_t *)(op0), (vuint64m2_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vsoxseg7ei64_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei64_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint64m2_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (size_t)(op10)) +#define vsoxseg7ei64_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei64_v_u8mf8((uint8_t *)(op0), (vuint64m1_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vsoxseg7ei64_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei64_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint64m1_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (size_t)(op10)) +#define vsoxseg8ei64_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei64_v_i8m1((int8_t *)(op0), (vuint64m8_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei64_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei64_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint64m8_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (size_t)(op11)) +#define vsoxseg8ei64_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei64_v_i8mf2((int8_t *)(op0), (vuint64m4_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (size_t)(op10)) +#define vsoxseg8ei64_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei64_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint64m4_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (size_t)(op11)) +#define vsoxseg8ei64_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei64_v_i8mf4((int8_t *)(op0), (vuint64m2_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (size_t)(op10)) +#define vsoxseg8ei64_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei64_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint64m2_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (size_t)(op11)) +#define vsoxseg8ei64_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei64_v_i8mf8((int8_t *)(op0), (vuint64m1_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (size_t)(op10)) +#define vsoxseg8ei64_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei64_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint64m1_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (size_t)(op11)) +#define vsoxseg8ei64_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei64_v_u8m1((uint8_t *)(op0), (vuint64m8_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei64_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei64_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint64m8_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (size_t)(op11)) +#define vsoxseg8ei64_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei64_v_u8mf2((uint8_t *)(op0), (vuint64m4_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (size_t)(op10)) +#define vsoxseg8ei64_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei64_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint64m4_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (size_t)(op11)) +#define vsoxseg8ei64_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei64_v_u8mf4((uint8_t *)(op0), (vuint64m2_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (size_t)(op10)) +#define vsoxseg8ei64_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei64_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint64m2_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (size_t)(op11)) +#define vsoxseg8ei64_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei64_v_u8mf8((uint8_t *)(op0), (vuint64m1_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (size_t)(op10)) +#define vsoxseg8ei64_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei64_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint64m1_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (size_t)(op11)) +#define vsoxseg2ei8_v_i16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_i16m1((int16_t *)(op0), (vuint8mf2_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_i16m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint8mf2_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (size_t)(op5)) +#define vsoxseg2ei8_v_i16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_i16m2((int16_t *)(op0), (vuint8m1_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_i16m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_i16m2_m((vbool8_t)(op0), (int16_t *)(op1), (vuint8m1_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei8_v_i16m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_i16m4((int16_t *)(op0), (vuint8m2_t)(op1), (vint16m4_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_i16m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_i16m4_m((vbool4_t)(op0), (int16_t *)(op1), (vuint8m2_t)(op2), (vint16m4_t)(op3), (vint16m4_t)(op4), (size_t)(op5)) +#define vsoxseg2ei8_v_i16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_i16mf2((int16_t *)(op0), (vuint8mf4_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_i16mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint8mf4_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei8_v_i16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_i16mf4((int16_t *)(op0), (vuint8mf8_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_i16mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint8mf8_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (size_t)(op5)) +#define vsoxseg2ei8_v_u16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_u16m1((uint16_t *)(op0), (vuint8mf2_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_u16m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint8mf2_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (size_t)(op5)) +#define vsoxseg2ei8_v_u16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_u16m2((uint16_t *)(op0), (vuint8m1_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_u16m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_u16m2_m((vbool8_t)(op0), (uint16_t *)(op1), (vuint8m1_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei8_v_u16m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_u16m4((uint16_t *)(op0), (vuint8m2_t)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_u16m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_u16m4_m((vbool4_t)(op0), (uint16_t *)(op1), (vuint8m2_t)(op2), (vuint16m4_t)(op3), (vuint16m4_t)(op4), (size_t)(op5)) +#define vsoxseg2ei8_v_u16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_u16mf2((uint16_t *)(op0), (vuint8mf4_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_u16mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint8mf4_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei8_v_u16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_u16mf4((uint16_t *)(op0), (vuint8mf8_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_u16mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint8mf8_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (size_t)(op5)) +#define vsoxseg3ei8_v_i16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei8_v_i16m1((int16_t *)(op0), (vuint8mf2_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (size_t)(op5)) +#define vsoxseg3ei8_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei8_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint8mf2_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (size_t)(op6)) +#define vsoxseg3ei8_v_i16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei8_v_i16m2((int16_t *)(op0), (vuint8m1_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei8_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei8_v_i16m2_m((vbool8_t)(op0), (int16_t *)(op1), (vuint8m1_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (vint16m2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei8_v_i16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei8_v_i16mf2((int16_t *)(op0), (vuint8mf4_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei8_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei8_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint8mf4_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei8_v_i16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei8_v_i16mf4((int16_t *)(op0), (vuint8mf8_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (size_t)(op5)) +#define vsoxseg3ei8_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei8_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint8mf8_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (size_t)(op6)) +#define vsoxseg3ei8_v_u16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei8_v_u16m1((uint16_t *)(op0), (vuint8mf2_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (size_t)(op5)) +#define vsoxseg3ei8_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei8_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint8mf2_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (size_t)(op6)) +#define vsoxseg3ei8_v_u16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei8_v_u16m2((uint16_t *)(op0), (vuint8m1_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei8_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei8_v_u16m2_m((vbool8_t)(op0), (uint16_t *)(op1), (vuint8m1_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (vuint16m2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei8_v_u16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei8_v_u16mf2((uint16_t *)(op0), (vuint8mf4_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei8_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei8_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint8mf4_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei8_v_u16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei8_v_u16mf4((uint16_t *)(op0), (vuint8mf8_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (size_t)(op5)) +#define vsoxseg3ei8_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei8_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint8mf8_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (size_t)(op6)) +#define vsoxseg4ei8_v_i16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei8_v_i16m1((int16_t *)(op0), (vuint8mf2_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (size_t)(op6)) +#define vsoxseg4ei8_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei8_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint8mf2_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (size_t)(op7)) +#define vsoxseg4ei8_v_i16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei8_v_i16m2((int16_t *)(op0), (vuint8m1_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (vint16m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei8_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei8_v_i16m2_m((vbool8_t)(op0), (int16_t *)(op1), (vuint8m1_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (vint16m2_t)(op5), (vint16m2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei8_v_i16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei8_v_i16mf2((int16_t *)(op0), (vuint8mf4_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei8_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei8_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint8mf4_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei8_v_i16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei8_v_i16mf4((int16_t *)(op0), (vuint8mf8_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (size_t)(op6)) +#define vsoxseg4ei8_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei8_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint8mf8_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (size_t)(op7)) +#define vsoxseg4ei8_v_u16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei8_v_u16m1((uint16_t *)(op0), (vuint8mf2_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (size_t)(op6)) +#define vsoxseg4ei8_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei8_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint8mf2_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vsoxseg4ei8_v_u16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei8_v_u16m2((uint16_t *)(op0), (vuint8m1_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (vuint16m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei8_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei8_v_u16m2_m((vbool8_t)(op0), (uint16_t *)(op1), (vuint8m1_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (vuint16m2_t)(op5), (vuint16m2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei8_v_u16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei8_v_u16mf2((uint16_t *)(op0), (vuint8mf4_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei8_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei8_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint8mf4_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei8_v_u16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei8_v_u16mf4((uint16_t *)(op0), (vuint8mf8_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (size_t)(op6)) +#define vsoxseg4ei8_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei8_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint8mf8_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vsoxseg5ei8_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei8_v_i16m1((int16_t *)(op0), (vuint8mf2_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (size_t)(op7)) +#define vsoxseg5ei8_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei8_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint8mf2_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (size_t)(op8)) +#define vsoxseg5ei8_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei8_v_i16mf2((int16_t *)(op0), (vuint8mf4_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei8_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei8_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint8mf4_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (size_t)(op8)) +#define vsoxseg5ei8_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei8_v_i16mf4((int16_t *)(op0), (vuint8mf8_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (size_t)(op7)) +#define vsoxseg5ei8_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei8_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint8mf8_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (size_t)(op8)) +#define vsoxseg5ei8_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei8_v_u16m1((uint16_t *)(op0), (vuint8mf2_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vsoxseg5ei8_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei8_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint8mf2_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (size_t)(op8)) +#define vsoxseg5ei8_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei8_v_u16mf2((uint16_t *)(op0), (vuint8mf4_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei8_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei8_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint8mf4_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (size_t)(op8)) +#define vsoxseg5ei8_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei8_v_u16mf4((uint16_t *)(op0), (vuint8mf8_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vsoxseg5ei8_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei8_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint8mf8_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (size_t)(op8)) +#define vsoxseg6ei8_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei8_v_i16m1((int16_t *)(op0), (vuint8mf2_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei8_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei8_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint8mf2_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (size_t)(op9)) +#define vsoxseg6ei8_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei8_v_i16mf2((int16_t *)(op0), (vuint8mf4_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (size_t)(op8)) +#define vsoxseg6ei8_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei8_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint8mf4_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (size_t)(op9)) +#define vsoxseg6ei8_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei8_v_i16mf4((int16_t *)(op0), (vuint8mf8_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (size_t)(op8)) +#define vsoxseg6ei8_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei8_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint8mf8_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (size_t)(op9)) +#define vsoxseg6ei8_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei8_v_u16m1((uint16_t *)(op0), (vuint8mf2_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei8_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei8_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint8mf2_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vsoxseg6ei8_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei8_v_u16mf2((uint16_t *)(op0), (vuint8mf4_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (size_t)(op8)) +#define vsoxseg6ei8_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei8_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint8mf4_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vsoxseg6ei8_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei8_v_u16mf4((uint16_t *)(op0), (vuint8mf8_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (size_t)(op8)) +#define vsoxseg6ei8_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei8_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint8mf8_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vsoxseg7ei8_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei8_v_i16m1((int16_t *)(op0), (vuint8mf2_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei8_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei8_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint8mf2_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (size_t)(op10)) +#define vsoxseg7ei8_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei8_v_i16mf2((int16_t *)(op0), (vuint8mf4_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (size_t)(op9)) +#define vsoxseg7ei8_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei8_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint8mf4_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (size_t)(op10)) +#define vsoxseg7ei8_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei8_v_i16mf4((int16_t *)(op0), (vuint8mf8_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (size_t)(op9)) +#define vsoxseg7ei8_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei8_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint8mf8_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (size_t)(op10)) +#define vsoxseg7ei8_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei8_v_u16m1((uint16_t *)(op0), (vuint8mf2_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei8_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei8_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint8mf2_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (size_t)(op10)) +#define vsoxseg7ei8_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei8_v_u16mf2((uint16_t *)(op0), (vuint8mf4_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vsoxseg7ei8_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei8_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint8mf4_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (size_t)(op10)) +#define vsoxseg7ei8_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei8_v_u16mf4((uint16_t *)(op0), (vuint8mf8_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vsoxseg7ei8_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei8_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint8mf8_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (size_t)(op10)) +#define vsoxseg8ei8_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei8_v_i16m1((int16_t *)(op0), (vuint8mf2_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei8_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei8_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint8mf2_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (size_t)(op11)) +#define vsoxseg8ei8_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei8_v_i16mf2((int16_t *)(op0), (vuint8mf4_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (size_t)(op10)) +#define vsoxseg8ei8_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei8_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint8mf4_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (size_t)(op11)) +#define vsoxseg8ei8_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei8_v_i16mf4((int16_t *)(op0), (vuint8mf8_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (size_t)(op10)) +#define vsoxseg8ei8_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei8_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint8mf8_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (size_t)(op11)) +#define vsoxseg8ei8_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei8_v_u16m1((uint16_t *)(op0), (vuint8mf2_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei8_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei8_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint8mf2_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (size_t)(op11)) +#define vsoxseg8ei8_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei8_v_u16mf2((uint16_t *)(op0), (vuint8mf4_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (size_t)(op10)) +#define vsoxseg8ei8_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei8_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint8mf4_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (size_t)(op11)) +#define vsoxseg8ei8_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei8_v_u16mf4((uint16_t *)(op0), (vuint8mf8_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (size_t)(op10)) +#define vsoxseg8ei8_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei8_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint8mf8_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (size_t)(op11)) +#define vsoxseg2ei16_v_i16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_i16m1((int16_t *)(op0), (vuint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_i16m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (size_t)(op5)) +#define vsoxseg2ei16_v_i16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_i16m2((int16_t *)(op0), (vuint16m2_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_i16m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_i16m2_m((vbool8_t)(op0), (int16_t *)(op1), (vuint16m2_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei16_v_i16m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_i16m4((int16_t *)(op0), (vuint16m4_t)(op1), (vint16m4_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_i16m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_i16m4_m((vbool4_t)(op0), (int16_t *)(op1), (vuint16m4_t)(op2), (vint16m4_t)(op3), (vint16m4_t)(op4), (size_t)(op5)) +#define vsoxseg2ei16_v_i16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_i16mf2((int16_t *)(op0), (vuint16mf2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_i16mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei16_v_i16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_i16mf4((int16_t *)(op0), (vuint16mf4_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_i16mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (size_t)(op5)) +#define vsoxseg2ei16_v_u16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_u16m1((uint16_t *)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_u16m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (size_t)(op5)) +#define vsoxseg2ei16_v_u16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_u16m2((uint16_t *)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_u16m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_u16m2_m((vbool8_t)(op0), (uint16_t *)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei16_v_u16m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_u16m4((uint16_t *)(op0), (vuint16m4_t)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_u16m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_u16m4_m((vbool4_t)(op0), (uint16_t *)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (vuint16m4_t)(op4), (size_t)(op5)) +#define vsoxseg2ei16_v_u16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_u16mf2((uint16_t *)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_u16mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei16_v_u16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_u16mf4((uint16_t *)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_u16mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (size_t)(op5)) +#define vsoxseg3ei16_v_i16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei16_v_i16m1((int16_t *)(op0), (vuint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (size_t)(op5)) +#define vsoxseg3ei16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei16_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (size_t)(op6)) +#define vsoxseg3ei16_v_i16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei16_v_i16m2((int16_t *)(op0), (vuint16m2_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei16_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei16_v_i16m2_m((vbool8_t)(op0), (int16_t *)(op1), (vuint16m2_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (vint16m2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei16_v_i16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei16_v_i16mf2((int16_t *)(op0), (vuint16mf2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei16_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei16_v_i16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei16_v_i16mf4((int16_t *)(op0), (vuint16mf4_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (size_t)(op5)) +#define vsoxseg3ei16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei16_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (size_t)(op6)) +#define vsoxseg3ei16_v_u16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei16_v_u16m1((uint16_t *)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (size_t)(op5)) +#define vsoxseg3ei16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei16_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (size_t)(op6)) +#define vsoxseg3ei16_v_u16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei16_v_u16m2((uint16_t *)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei16_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei16_v_u16m2_m((vbool8_t)(op0), (uint16_t *)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (vuint16m2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei16_v_u16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei16_v_u16mf2((uint16_t *)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei16_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei16_v_u16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei16_v_u16mf4((uint16_t *)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (size_t)(op5)) +#define vsoxseg3ei16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei16_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (size_t)(op6)) +#define vsoxseg4ei16_v_i16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei16_v_i16m1((int16_t *)(op0), (vuint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (size_t)(op6)) +#define vsoxseg4ei16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei16_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (size_t)(op7)) +#define vsoxseg4ei16_v_i16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei16_v_i16m2((int16_t *)(op0), (vuint16m2_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (vint16m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei16_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei16_v_i16m2_m((vbool8_t)(op0), (int16_t *)(op1), (vuint16m2_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (vint16m2_t)(op5), (vint16m2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei16_v_i16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei16_v_i16mf2((int16_t *)(op0), (vuint16mf2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei16_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei16_v_i16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei16_v_i16mf4((int16_t *)(op0), (vuint16mf4_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (size_t)(op6)) +#define vsoxseg4ei16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei16_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (size_t)(op7)) +#define vsoxseg4ei16_v_u16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei16_v_u16m1((uint16_t *)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (size_t)(op6)) +#define vsoxseg4ei16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei16_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vsoxseg4ei16_v_u16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei16_v_u16m2((uint16_t *)(op0), (vuint16m2_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (vuint16m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei16_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei16_v_u16m2_m((vbool8_t)(op0), (uint16_t *)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (vuint16m2_t)(op5), (vuint16m2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei16_v_u16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei16_v_u16mf2((uint16_t *)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei16_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei16_v_u16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei16_v_u16mf4((uint16_t *)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (size_t)(op6)) +#define vsoxseg4ei16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei16_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vsoxseg5ei16_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei16_v_i16m1((int16_t *)(op0), (vuint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (size_t)(op7)) +#define vsoxseg5ei16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei16_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (size_t)(op8)) +#define vsoxseg5ei16_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei16_v_i16mf2((int16_t *)(op0), (vuint16mf2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei16_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (size_t)(op8)) +#define vsoxseg5ei16_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei16_v_i16mf4((int16_t *)(op0), (vuint16mf4_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (size_t)(op7)) +#define vsoxseg5ei16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei16_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (size_t)(op8)) +#define vsoxseg5ei16_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei16_v_u16m1((uint16_t *)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vsoxseg5ei16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei16_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (size_t)(op8)) +#define vsoxseg5ei16_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei16_v_u16mf2((uint16_t *)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei16_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (size_t)(op8)) +#define vsoxseg5ei16_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei16_v_u16mf4((uint16_t *)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vsoxseg5ei16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei16_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (size_t)(op8)) +#define vsoxseg6ei16_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei16_v_i16m1((int16_t *)(op0), (vuint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei16_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (size_t)(op9)) +#define vsoxseg6ei16_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei16_v_i16mf2((int16_t *)(op0), (vuint16mf2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (size_t)(op8)) +#define vsoxseg6ei16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei16_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (size_t)(op9)) +#define vsoxseg6ei16_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei16_v_i16mf4((int16_t *)(op0), (vuint16mf4_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (size_t)(op8)) +#define vsoxseg6ei16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei16_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (size_t)(op9)) +#define vsoxseg6ei16_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei16_v_u16m1((uint16_t *)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei16_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vsoxseg6ei16_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei16_v_u16mf2((uint16_t *)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (size_t)(op8)) +#define vsoxseg6ei16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei16_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vsoxseg6ei16_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei16_v_u16mf4((uint16_t *)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (size_t)(op8)) +#define vsoxseg6ei16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei16_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vsoxseg7ei16_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei16_v_i16m1((int16_t *)(op0), (vuint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei16_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (size_t)(op10)) +#define vsoxseg7ei16_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei16_v_i16mf2((int16_t *)(op0), (vuint16mf2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (size_t)(op9)) +#define vsoxseg7ei16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei16_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (size_t)(op10)) +#define vsoxseg7ei16_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei16_v_i16mf4((int16_t *)(op0), (vuint16mf4_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (size_t)(op9)) +#define vsoxseg7ei16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei16_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (size_t)(op10)) +#define vsoxseg7ei16_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei16_v_u16m1((uint16_t *)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei16_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (size_t)(op10)) +#define vsoxseg7ei16_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei16_v_u16mf2((uint16_t *)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vsoxseg7ei16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei16_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (size_t)(op10)) +#define vsoxseg7ei16_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei16_v_u16mf4((uint16_t *)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vsoxseg7ei16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei16_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (size_t)(op10)) +#define vsoxseg8ei16_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei16_v_i16m1((int16_t *)(op0), (vuint16m1_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei16_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (size_t)(op11)) +#define vsoxseg8ei16_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei16_v_i16mf2((int16_t *)(op0), (vuint16mf2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (size_t)(op10)) +#define vsoxseg8ei16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei16_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (size_t)(op11)) +#define vsoxseg8ei16_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei16_v_i16mf4((int16_t *)(op0), (vuint16mf4_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (size_t)(op10)) +#define vsoxseg8ei16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei16_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (size_t)(op11)) +#define vsoxseg8ei16_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei16_v_u16m1((uint16_t *)(op0), (vuint16m1_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei16_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (size_t)(op11)) +#define vsoxseg8ei16_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei16_v_u16mf2((uint16_t *)(op0), (vuint16mf2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (size_t)(op10)) +#define vsoxseg8ei16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei16_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (size_t)(op11)) +#define vsoxseg8ei16_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei16_v_u16mf4((uint16_t *)(op0), (vuint16mf4_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (size_t)(op10)) +#define vsoxseg8ei16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei16_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (size_t)(op11)) +#define vsoxseg2ei32_v_i16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_i16m1((int16_t *)(op0), (vuint32m2_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_i16m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint32m2_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (size_t)(op5)) +#define vsoxseg2ei32_v_i16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_i16m2((int16_t *)(op0), (vuint32m4_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_i16m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_i16m2_m((vbool8_t)(op0), (int16_t *)(op1), (vuint32m4_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei32_v_i16m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_i16m4((int16_t *)(op0), (vuint32m8_t)(op1), (vint16m4_t)(op2), (vint16m4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_i16m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_i16m4_m((vbool4_t)(op0), (int16_t *)(op1), (vuint32m8_t)(op2), (vint16m4_t)(op3), (vint16m4_t)(op4), (size_t)(op5)) +#define vsoxseg2ei32_v_i16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_i16mf2((int16_t *)(op0), (vuint32m1_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_i16mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint32m1_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei32_v_i16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_i16mf4((int16_t *)(op0), (vuint32mf2_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_i16mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint32mf2_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (size_t)(op5)) +#define vsoxseg2ei32_v_u16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_u16m1((uint16_t *)(op0), (vuint32m2_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_u16m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint32m2_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (size_t)(op5)) +#define vsoxseg2ei32_v_u16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_u16m2((uint16_t *)(op0), (vuint32m4_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_u16m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_u16m2_m((vbool8_t)(op0), (uint16_t *)(op1), (vuint32m4_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei32_v_u16m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_u16m4((uint16_t *)(op0), (vuint32m8_t)(op1), (vuint16m4_t)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_u16m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_u16m4_m((vbool4_t)(op0), (uint16_t *)(op1), (vuint32m8_t)(op2), (vuint16m4_t)(op3), (vuint16m4_t)(op4), (size_t)(op5)) +#define vsoxseg2ei32_v_u16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_u16mf2((uint16_t *)(op0), (vuint32m1_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_u16mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint32m1_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei32_v_u16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_u16mf4((uint16_t *)(op0), (vuint32mf2_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_u16mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint32mf2_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (size_t)(op5)) +#define vsoxseg3ei32_v_i16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei32_v_i16m1((int16_t *)(op0), (vuint32m2_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (size_t)(op5)) +#define vsoxseg3ei32_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei32_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint32m2_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (size_t)(op6)) +#define vsoxseg3ei32_v_i16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei32_v_i16m2((int16_t *)(op0), (vuint32m4_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei32_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei32_v_i16m2_m((vbool8_t)(op0), (int16_t *)(op1), (vuint32m4_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (vint16m2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei32_v_i16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei32_v_i16mf2((int16_t *)(op0), (vuint32m1_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei32_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei32_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint32m1_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei32_v_i16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei32_v_i16mf4((int16_t *)(op0), (vuint32mf2_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (size_t)(op5)) +#define vsoxseg3ei32_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei32_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint32mf2_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (size_t)(op6)) +#define vsoxseg3ei32_v_u16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei32_v_u16m1((uint16_t *)(op0), (vuint32m2_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (size_t)(op5)) +#define vsoxseg3ei32_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei32_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint32m2_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (size_t)(op6)) +#define vsoxseg3ei32_v_u16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei32_v_u16m2((uint16_t *)(op0), (vuint32m4_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei32_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei32_v_u16m2_m((vbool8_t)(op0), (uint16_t *)(op1), (vuint32m4_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (vuint16m2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei32_v_u16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei32_v_u16mf2((uint16_t *)(op0), (vuint32m1_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei32_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei32_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint32m1_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei32_v_u16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei32_v_u16mf4((uint16_t *)(op0), (vuint32mf2_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (size_t)(op5)) +#define vsoxseg3ei32_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei32_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint32mf2_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (size_t)(op6)) +#define vsoxseg4ei32_v_i16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei32_v_i16m1((int16_t *)(op0), (vuint32m2_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (size_t)(op6)) +#define vsoxseg4ei32_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei32_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint32m2_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (size_t)(op7)) +#define vsoxseg4ei32_v_i16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei32_v_i16m2((int16_t *)(op0), (vuint32m4_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (vint16m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei32_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei32_v_i16m2_m((vbool8_t)(op0), (int16_t *)(op1), (vuint32m4_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (vint16m2_t)(op5), (vint16m2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei32_v_i16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei32_v_i16mf2((int16_t *)(op0), (vuint32m1_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei32_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei32_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint32m1_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei32_v_i16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei32_v_i16mf4((int16_t *)(op0), (vuint32mf2_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (size_t)(op6)) +#define vsoxseg4ei32_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei32_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint32mf2_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (size_t)(op7)) +#define vsoxseg4ei32_v_u16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei32_v_u16m1((uint16_t *)(op0), (vuint32m2_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (size_t)(op6)) +#define vsoxseg4ei32_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei32_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint32m2_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vsoxseg4ei32_v_u16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei32_v_u16m2((uint16_t *)(op0), (vuint32m4_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (vuint16m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei32_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei32_v_u16m2_m((vbool8_t)(op0), (uint16_t *)(op1), (vuint32m4_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (vuint16m2_t)(op5), (vuint16m2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei32_v_u16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei32_v_u16mf2((uint16_t *)(op0), (vuint32m1_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei32_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei32_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint32m1_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei32_v_u16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei32_v_u16mf4((uint16_t *)(op0), (vuint32mf2_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (size_t)(op6)) +#define vsoxseg4ei32_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei32_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint32mf2_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vsoxseg5ei32_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei32_v_i16m1((int16_t *)(op0), (vuint32m2_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (size_t)(op7)) +#define vsoxseg5ei32_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei32_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint32m2_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (size_t)(op8)) +#define vsoxseg5ei32_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei32_v_i16mf2((int16_t *)(op0), (vuint32m1_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei32_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei32_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint32m1_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (size_t)(op8)) +#define vsoxseg5ei32_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei32_v_i16mf4((int16_t *)(op0), (vuint32mf2_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (size_t)(op7)) +#define vsoxseg5ei32_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei32_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint32mf2_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (size_t)(op8)) +#define vsoxseg5ei32_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei32_v_u16m1((uint16_t *)(op0), (vuint32m2_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vsoxseg5ei32_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei32_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint32m2_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (size_t)(op8)) +#define vsoxseg5ei32_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei32_v_u16mf2((uint16_t *)(op0), (vuint32m1_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei32_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei32_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint32m1_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (size_t)(op8)) +#define vsoxseg5ei32_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei32_v_u16mf4((uint16_t *)(op0), (vuint32mf2_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vsoxseg5ei32_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei32_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint32mf2_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (size_t)(op8)) +#define vsoxseg6ei32_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei32_v_i16m1((int16_t *)(op0), (vuint32m2_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei32_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei32_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint32m2_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (size_t)(op9)) +#define vsoxseg6ei32_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei32_v_i16mf2((int16_t *)(op0), (vuint32m1_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (size_t)(op8)) +#define vsoxseg6ei32_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei32_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint32m1_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (size_t)(op9)) +#define vsoxseg6ei32_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei32_v_i16mf4((int16_t *)(op0), (vuint32mf2_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (size_t)(op8)) +#define vsoxseg6ei32_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei32_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint32mf2_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (size_t)(op9)) +#define vsoxseg6ei32_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei32_v_u16m1((uint16_t *)(op0), (vuint32m2_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei32_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei32_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint32m2_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vsoxseg6ei32_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei32_v_u16mf2((uint16_t *)(op0), (vuint32m1_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (size_t)(op8)) +#define vsoxseg6ei32_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei32_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint32m1_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vsoxseg6ei32_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei32_v_u16mf4((uint16_t *)(op0), (vuint32mf2_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (size_t)(op8)) +#define vsoxseg6ei32_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei32_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint32mf2_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vsoxseg7ei32_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei32_v_i16m1((int16_t *)(op0), (vuint32m2_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei32_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei32_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint32m2_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (size_t)(op10)) +#define vsoxseg7ei32_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei32_v_i16mf2((int16_t *)(op0), (vuint32m1_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (size_t)(op9)) +#define vsoxseg7ei32_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei32_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint32m1_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (size_t)(op10)) +#define vsoxseg7ei32_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei32_v_i16mf4((int16_t *)(op0), (vuint32mf2_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (size_t)(op9)) +#define vsoxseg7ei32_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei32_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint32mf2_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (size_t)(op10)) +#define vsoxseg7ei32_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei32_v_u16m1((uint16_t *)(op0), (vuint32m2_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei32_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei32_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint32m2_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (size_t)(op10)) +#define vsoxseg7ei32_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei32_v_u16mf2((uint16_t *)(op0), (vuint32m1_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vsoxseg7ei32_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei32_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint32m1_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (size_t)(op10)) +#define vsoxseg7ei32_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei32_v_u16mf4((uint16_t *)(op0), (vuint32mf2_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vsoxseg7ei32_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei32_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint32mf2_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (size_t)(op10)) +#define vsoxseg8ei32_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei32_v_i16m1((int16_t *)(op0), (vuint32m2_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei32_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei32_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint32m2_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (size_t)(op11)) +#define vsoxseg8ei32_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei32_v_i16mf2((int16_t *)(op0), (vuint32m1_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (size_t)(op10)) +#define vsoxseg8ei32_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei32_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint32m1_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (size_t)(op11)) +#define vsoxseg8ei32_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei32_v_i16mf4((int16_t *)(op0), (vuint32mf2_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (size_t)(op10)) +#define vsoxseg8ei32_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei32_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint32mf2_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (size_t)(op11)) +#define vsoxseg8ei32_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei32_v_u16m1((uint16_t *)(op0), (vuint32m2_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei32_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei32_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint32m2_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (size_t)(op11)) +#define vsoxseg8ei32_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei32_v_u16mf2((uint16_t *)(op0), (vuint32m1_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (size_t)(op10)) +#define vsoxseg8ei32_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei32_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint32m1_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (size_t)(op11)) +#define vsoxseg8ei32_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei32_v_u16mf4((uint16_t *)(op0), (vuint32mf2_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (size_t)(op10)) +#define vsoxseg8ei32_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei32_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint32mf2_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (size_t)(op11)) +#define vsoxseg2ei64_v_i16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei64_v_i16m1((int16_t *)(op0), (vuint64m4_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (size_t)(op4)) +#define vsoxseg2ei64_v_i16m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei64_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint64m4_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (size_t)(op5)) +#define vsoxseg2ei64_v_i16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei64_v_i16m2((int16_t *)(op0), (vuint64m8_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei64_v_i16m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei64_v_i16m2_m((vbool8_t)(op0), (int16_t *)(op1), (vuint64m8_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei64_v_i16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei64_v_i16mf2((int16_t *)(op0), (vuint64m2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei64_v_i16mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei64_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint64m2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei64_v_i16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei64_v_i16mf4((int16_t *)(op0), (vuint64m1_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei64_v_i16mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei64_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint64m1_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (size_t)(op5)) +#define vsoxseg2ei64_v_u16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei64_v_u16m1((uint16_t *)(op0), (vuint64m4_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vsoxseg2ei64_v_u16m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei64_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint64m4_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (size_t)(op5)) +#define vsoxseg2ei64_v_u16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei64_v_u16m2((uint16_t *)(op0), (vuint64m8_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei64_v_u16m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei64_v_u16m2_m((vbool8_t)(op0), (uint16_t *)(op1), (vuint64m8_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei64_v_u16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei64_v_u16mf2((uint16_t *)(op0), (vuint64m2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei64_v_u16mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei64_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint64m2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei64_v_u16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei64_v_u16mf4((uint16_t *)(op0), (vuint64m1_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei64_v_u16mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei64_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint64m1_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (size_t)(op5)) +#define vsoxseg3ei64_v_i16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei64_v_i16m1((int16_t *)(op0), (vuint64m4_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (size_t)(op5)) +#define vsoxseg3ei64_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei64_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint64m4_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (size_t)(op6)) +#define vsoxseg3ei64_v_i16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei64_v_i16m2((int16_t *)(op0), (vuint64m8_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei64_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei64_v_i16m2_m((vbool8_t)(op0), (int16_t *)(op1), (vuint64m8_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (vint16m2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei64_v_i16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei64_v_i16mf2((int16_t *)(op0), (vuint64m2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei64_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei64_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint64m2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei64_v_i16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei64_v_i16mf4((int16_t *)(op0), (vuint64m1_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (size_t)(op5)) +#define vsoxseg3ei64_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei64_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint64m1_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (size_t)(op6)) +#define vsoxseg3ei64_v_u16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei64_v_u16m1((uint16_t *)(op0), (vuint64m4_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (size_t)(op5)) +#define vsoxseg3ei64_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei64_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint64m4_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (size_t)(op6)) +#define vsoxseg3ei64_v_u16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei64_v_u16m2((uint16_t *)(op0), (vuint64m8_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei64_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei64_v_u16m2_m((vbool8_t)(op0), (uint16_t *)(op1), (vuint64m8_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (vuint16m2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei64_v_u16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei64_v_u16mf2((uint16_t *)(op0), (vuint64m2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei64_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei64_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint64m2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei64_v_u16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei64_v_u16mf4((uint16_t *)(op0), (vuint64m1_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (size_t)(op5)) +#define vsoxseg3ei64_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei64_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint64m1_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (size_t)(op6)) +#define vsoxseg4ei64_v_i16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei64_v_i16m1((int16_t *)(op0), (vuint64m4_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (size_t)(op6)) +#define vsoxseg4ei64_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei64_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint64m4_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (size_t)(op7)) +#define vsoxseg4ei64_v_i16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei64_v_i16m2((int16_t *)(op0), (vuint64m8_t)(op1), (vint16m2_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (vint16m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei64_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei64_v_i16m2_m((vbool8_t)(op0), (int16_t *)(op1), (vuint64m8_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (vint16m2_t)(op5), (vint16m2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei64_v_i16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei64_v_i16mf2((int16_t *)(op0), (vuint64m2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei64_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei64_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint64m2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei64_v_i16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei64_v_i16mf4((int16_t *)(op0), (vuint64m1_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (size_t)(op6)) +#define vsoxseg4ei64_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei64_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint64m1_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (size_t)(op7)) +#define vsoxseg4ei64_v_u16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei64_v_u16m1((uint16_t *)(op0), (vuint64m4_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (size_t)(op6)) +#define vsoxseg4ei64_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei64_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint64m4_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vsoxseg4ei64_v_u16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei64_v_u16m2((uint16_t *)(op0), (vuint64m8_t)(op1), (vuint16m2_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (vuint16m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei64_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei64_v_u16m2_m((vbool8_t)(op0), (uint16_t *)(op1), (vuint64m8_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (vuint16m2_t)(op5), (vuint16m2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei64_v_u16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei64_v_u16mf2((uint16_t *)(op0), (vuint64m2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei64_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei64_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint64m2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei64_v_u16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei64_v_u16mf4((uint16_t *)(op0), (vuint64m1_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (size_t)(op6)) +#define vsoxseg4ei64_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei64_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint64m1_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vsoxseg5ei64_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei64_v_i16m1((int16_t *)(op0), (vuint64m4_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (size_t)(op7)) +#define vsoxseg5ei64_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei64_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint64m4_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (size_t)(op8)) +#define vsoxseg5ei64_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei64_v_i16mf2((int16_t *)(op0), (vuint64m2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei64_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei64_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint64m2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (size_t)(op8)) +#define vsoxseg5ei64_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei64_v_i16mf4((int16_t *)(op0), (vuint64m1_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (size_t)(op7)) +#define vsoxseg5ei64_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei64_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint64m1_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (size_t)(op8)) +#define vsoxseg5ei64_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei64_v_u16m1((uint16_t *)(op0), (vuint64m4_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vsoxseg5ei64_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei64_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint64m4_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (size_t)(op8)) +#define vsoxseg5ei64_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei64_v_u16mf2((uint16_t *)(op0), (vuint64m2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei64_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei64_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint64m2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (size_t)(op8)) +#define vsoxseg5ei64_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei64_v_u16mf4((uint16_t *)(op0), (vuint64m1_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vsoxseg5ei64_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei64_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint64m1_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (size_t)(op8)) +#define vsoxseg6ei64_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei64_v_i16m1((int16_t *)(op0), (vuint64m4_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei64_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei64_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint64m4_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (size_t)(op9)) +#define vsoxseg6ei64_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei64_v_i16mf2((int16_t *)(op0), (vuint64m2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (size_t)(op8)) +#define vsoxseg6ei64_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei64_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint64m2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (size_t)(op9)) +#define vsoxseg6ei64_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei64_v_i16mf4((int16_t *)(op0), (vuint64m1_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (size_t)(op8)) +#define vsoxseg6ei64_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei64_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint64m1_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (size_t)(op9)) +#define vsoxseg6ei64_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei64_v_u16m1((uint16_t *)(op0), (vuint64m4_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei64_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei64_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint64m4_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vsoxseg6ei64_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei64_v_u16mf2((uint16_t *)(op0), (vuint64m2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (size_t)(op8)) +#define vsoxseg6ei64_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei64_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint64m2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vsoxseg6ei64_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei64_v_u16mf4((uint16_t *)(op0), (vuint64m1_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (size_t)(op8)) +#define vsoxseg6ei64_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei64_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint64m1_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vsoxseg7ei64_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei64_v_i16m1((int16_t *)(op0), (vuint64m4_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei64_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei64_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint64m4_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (size_t)(op10)) +#define vsoxseg7ei64_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei64_v_i16mf2((int16_t *)(op0), (vuint64m2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (size_t)(op9)) +#define vsoxseg7ei64_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei64_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint64m2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (size_t)(op10)) +#define vsoxseg7ei64_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei64_v_i16mf4((int16_t *)(op0), (vuint64m1_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (size_t)(op9)) +#define vsoxseg7ei64_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei64_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint64m1_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (size_t)(op10)) +#define vsoxseg7ei64_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei64_v_u16m1((uint16_t *)(op0), (vuint64m4_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei64_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei64_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint64m4_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (size_t)(op10)) +#define vsoxseg7ei64_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei64_v_u16mf2((uint16_t *)(op0), (vuint64m2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vsoxseg7ei64_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei64_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint64m2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (size_t)(op10)) +#define vsoxseg7ei64_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei64_v_u16mf4((uint16_t *)(op0), (vuint64m1_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vsoxseg7ei64_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei64_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint64m1_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (size_t)(op10)) +#define vsoxseg8ei64_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei64_v_i16m1((int16_t *)(op0), (vuint64m4_t)(op1), (vint16m1_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei64_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei64_v_i16m1_m((vbool16_t)(op0), (int16_t *)(op1), (vuint64m4_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (size_t)(op11)) +#define vsoxseg8ei64_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei64_v_i16mf2((int16_t *)(op0), (vuint64m2_t)(op1), (vint16mf2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (size_t)(op10)) +#define vsoxseg8ei64_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei64_v_i16mf2_m((vbool32_t)(op0), (int16_t *)(op1), (vuint64m2_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (size_t)(op11)) +#define vsoxseg8ei64_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei64_v_i16mf4((int16_t *)(op0), (vuint64m1_t)(op1), (vint16mf4_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (size_t)(op10)) +#define vsoxseg8ei64_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei64_v_i16mf4_m((vbool64_t)(op0), (int16_t *)(op1), (vuint64m1_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (size_t)(op11)) +#define vsoxseg8ei64_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei64_v_u16m1((uint16_t *)(op0), (vuint64m4_t)(op1), (vuint16m1_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei64_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei64_v_u16m1_m((vbool16_t)(op0), (uint16_t *)(op1), (vuint64m4_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (size_t)(op11)) +#define vsoxseg8ei64_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei64_v_u16mf2((uint16_t *)(op0), (vuint64m2_t)(op1), (vuint16mf2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (size_t)(op10)) +#define vsoxseg8ei64_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei64_v_u16mf2_m((vbool32_t)(op0), (uint16_t *)(op1), (vuint64m2_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (size_t)(op11)) +#define vsoxseg8ei64_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei64_v_u16mf4((uint16_t *)(op0), (vuint64m1_t)(op1), (vuint16mf4_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (size_t)(op10)) +#define vsoxseg8ei64_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei64_v_u16mf4_m((vbool64_t)(op0), (uint16_t *)(op1), (vuint64m1_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (size_t)(op11)) +#define vsoxseg2ei8_v_i32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_i32m1((int32_t *)(op0), (vuint8mf4_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_i32m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint8mf4_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (size_t)(op5)) +#define vsoxseg2ei8_v_i32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_i32m2((int32_t *)(op0), (vuint8mf2_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_i32m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint8mf2_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei8_v_i32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_i32m4((int32_t *)(op0), (vuint8m1_t)(op1), (vint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_i32m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vuint8m1_t)(op2), (vint32m4_t)(op3), (vint32m4_t)(op4), (size_t)(op5)) +#define vsoxseg2ei8_v_i32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_i32mf2((int32_t *)(op0), (vuint8mf8_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_i32mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint8mf8_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei8_v_u32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_u32m1((uint32_t *)(op0), (vuint8mf4_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_u32m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint8mf4_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (size_t)(op5)) +#define vsoxseg2ei8_v_u32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_u32m2((uint32_t *)(op0), (vuint8mf2_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_u32m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint8mf2_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei8_v_u32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_u32m4((uint32_t *)(op0), (vuint8m1_t)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_u32m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint8m1_t)(op2), (vuint32m4_t)(op3), (vuint32m4_t)(op4), (size_t)(op5)) +#define vsoxseg2ei8_v_u32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_u32mf2((uint32_t *)(op0), (vuint8mf8_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_u32mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint8mf8_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei8_v_i32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei8_v_i32m1((int32_t *)(op0), (vuint8mf4_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (size_t)(op5)) +#define vsoxseg3ei8_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei8_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint8mf4_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (size_t)(op6)) +#define vsoxseg3ei8_v_i32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei8_v_i32m2((int32_t *)(op0), (vuint8mf2_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei8_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei8_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint8mf2_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (vint32m2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei8_v_i32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei8_v_i32mf2((int32_t *)(op0), (vuint8mf8_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei8_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei8_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint8mf8_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei8_v_u32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei8_v_u32m1((uint32_t *)(op0), (vuint8mf4_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (size_t)(op5)) +#define vsoxseg3ei8_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei8_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint8mf4_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (size_t)(op6)) +#define vsoxseg3ei8_v_u32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei8_v_u32m2((uint32_t *)(op0), (vuint8mf2_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei8_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei8_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint8mf2_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (vuint32m2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei8_v_u32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei8_v_u32mf2((uint32_t *)(op0), (vuint8mf8_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei8_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei8_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint8mf8_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei8_v_i32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei8_v_i32m1((int32_t *)(op0), (vuint8mf4_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (size_t)(op6)) +#define vsoxseg4ei8_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei8_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint8mf4_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (size_t)(op7)) +#define vsoxseg4ei8_v_i32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei8_v_i32m2((int32_t *)(op0), (vuint8mf2_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (vint32m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei8_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei8_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint8mf2_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (vint32m2_t)(op5), (vint32m2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei8_v_i32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei8_v_i32mf2((int32_t *)(op0), (vuint8mf8_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei8_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei8_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint8mf8_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei8_v_u32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei8_v_u32m1((uint32_t *)(op0), (vuint8mf4_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (size_t)(op6)) +#define vsoxseg4ei8_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei8_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint8mf4_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vsoxseg4ei8_v_u32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei8_v_u32m2((uint32_t *)(op0), (vuint8mf2_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (vuint32m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei8_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei8_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint8mf2_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (vuint32m2_t)(op5), (vuint32m2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei8_v_u32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei8_v_u32mf2((uint32_t *)(op0), (vuint8mf8_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei8_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei8_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint8mf8_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei8_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei8_v_i32m1((int32_t *)(op0), (vuint8mf4_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (size_t)(op7)) +#define vsoxseg5ei8_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei8_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint8mf4_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (size_t)(op8)) +#define vsoxseg5ei8_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei8_v_i32mf2((int32_t *)(op0), (vuint8mf8_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei8_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei8_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint8mf8_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (size_t)(op8)) +#define vsoxseg5ei8_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei8_v_u32m1((uint32_t *)(op0), (vuint8mf4_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vsoxseg5ei8_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei8_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint8mf4_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (size_t)(op8)) +#define vsoxseg5ei8_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei8_v_u32mf2((uint32_t *)(op0), (vuint8mf8_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei8_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei8_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint8mf8_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (size_t)(op8)) +#define vsoxseg6ei8_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei8_v_i32m1((int32_t *)(op0), (vuint8mf4_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei8_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei8_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint8mf4_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (size_t)(op9)) +#define vsoxseg6ei8_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei8_v_i32mf2((int32_t *)(op0), (vuint8mf8_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (size_t)(op8)) +#define vsoxseg6ei8_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei8_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint8mf8_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (size_t)(op9)) +#define vsoxseg6ei8_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei8_v_u32m1((uint32_t *)(op0), (vuint8mf4_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei8_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei8_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint8mf4_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vsoxseg6ei8_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei8_v_u32mf2((uint32_t *)(op0), (vuint8mf8_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (size_t)(op8)) +#define vsoxseg6ei8_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei8_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint8mf8_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vsoxseg7ei8_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei8_v_i32m1((int32_t *)(op0), (vuint8mf4_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei8_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei8_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint8mf4_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (size_t)(op10)) +#define vsoxseg7ei8_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei8_v_i32mf2((int32_t *)(op0), (vuint8mf8_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (size_t)(op9)) +#define vsoxseg7ei8_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei8_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint8mf8_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (size_t)(op10)) +#define vsoxseg7ei8_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei8_v_u32m1((uint32_t *)(op0), (vuint8mf4_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei8_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei8_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint8mf4_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (size_t)(op10)) +#define vsoxseg7ei8_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei8_v_u32mf2((uint32_t *)(op0), (vuint8mf8_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vsoxseg7ei8_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei8_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint8mf8_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (size_t)(op10)) +#define vsoxseg8ei8_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei8_v_i32m1((int32_t *)(op0), (vuint8mf4_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei8_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei8_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint8mf4_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (size_t)(op11)) +#define vsoxseg8ei8_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei8_v_i32mf2((int32_t *)(op0), (vuint8mf8_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (size_t)(op10)) +#define vsoxseg8ei8_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei8_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint8mf8_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (size_t)(op11)) +#define vsoxseg8ei8_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei8_v_u32m1((uint32_t *)(op0), (vuint8mf4_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei8_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei8_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint8mf4_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (size_t)(op11)) +#define vsoxseg8ei8_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei8_v_u32mf2((uint32_t *)(op0), (vuint8mf8_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (size_t)(op10)) +#define vsoxseg8ei8_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei8_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint8mf8_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (size_t)(op11)) +#define vsoxseg2ei16_v_i32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_i32m1((int32_t *)(op0), (vuint16mf2_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_i32m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint16mf2_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (size_t)(op5)) +#define vsoxseg2ei16_v_i32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_i32m2((int32_t *)(op0), (vuint16m1_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_i32m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint16m1_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei16_v_i32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_i32m4((int32_t *)(op0), (vuint16m2_t)(op1), (vint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_i32m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vuint16m2_t)(op2), (vint32m4_t)(op3), (vint32m4_t)(op4), (size_t)(op5)) +#define vsoxseg2ei16_v_i32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_i32mf2((int32_t *)(op0), (vuint16mf4_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_i32mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint16mf4_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei16_v_u32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_u32m1((uint32_t *)(op0), (vuint16mf2_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_u32m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint16mf2_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (size_t)(op5)) +#define vsoxseg2ei16_v_u32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_u32m2((uint32_t *)(op0), (vuint16m1_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_u32m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint16m1_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei16_v_u32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_u32m4((uint32_t *)(op0), (vuint16m2_t)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_u32m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint16m2_t)(op2), (vuint32m4_t)(op3), (vuint32m4_t)(op4), (size_t)(op5)) +#define vsoxseg2ei16_v_u32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_u32mf2((uint32_t *)(op0), (vuint16mf4_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_u32mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint16mf4_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei16_v_i32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei16_v_i32m1((int32_t *)(op0), (vuint16mf2_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (size_t)(op5)) +#define vsoxseg3ei16_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei16_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint16mf2_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (size_t)(op6)) +#define vsoxseg3ei16_v_i32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei16_v_i32m2((int32_t *)(op0), (vuint16m1_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei16_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei16_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint16m1_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (vint32m2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei16_v_i32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei16_v_i32mf2((int32_t *)(op0), (vuint16mf4_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei16_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei16_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint16mf4_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei16_v_u32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei16_v_u32m1((uint32_t *)(op0), (vuint16mf2_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (size_t)(op5)) +#define vsoxseg3ei16_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei16_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint16mf2_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (size_t)(op6)) +#define vsoxseg3ei16_v_u32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei16_v_u32m2((uint32_t *)(op0), (vuint16m1_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei16_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei16_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint16m1_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (vuint32m2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei16_v_u32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei16_v_u32mf2((uint32_t *)(op0), (vuint16mf4_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei16_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei16_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint16mf4_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei16_v_i32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei16_v_i32m1((int32_t *)(op0), (vuint16mf2_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (size_t)(op6)) +#define vsoxseg4ei16_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei16_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint16mf2_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (size_t)(op7)) +#define vsoxseg4ei16_v_i32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei16_v_i32m2((int32_t *)(op0), (vuint16m1_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (vint32m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei16_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei16_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint16m1_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (vint32m2_t)(op5), (vint32m2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei16_v_i32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei16_v_i32mf2((int32_t *)(op0), (vuint16mf4_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei16_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei16_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint16mf4_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei16_v_u32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei16_v_u32m1((uint32_t *)(op0), (vuint16mf2_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (size_t)(op6)) +#define vsoxseg4ei16_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei16_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint16mf2_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vsoxseg4ei16_v_u32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei16_v_u32m2((uint32_t *)(op0), (vuint16m1_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (vuint32m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei16_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei16_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint16m1_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (vuint32m2_t)(op5), (vuint32m2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei16_v_u32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei16_v_u32mf2((uint32_t *)(op0), (vuint16mf4_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei16_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei16_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint16mf4_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei16_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei16_v_i32m1((int32_t *)(op0), (vuint16mf2_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (size_t)(op7)) +#define vsoxseg5ei16_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei16_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint16mf2_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (size_t)(op8)) +#define vsoxseg5ei16_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei16_v_i32mf2((int32_t *)(op0), (vuint16mf4_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei16_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei16_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint16mf4_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (size_t)(op8)) +#define vsoxseg5ei16_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei16_v_u32m1((uint32_t *)(op0), (vuint16mf2_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vsoxseg5ei16_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei16_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint16mf2_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (size_t)(op8)) +#define vsoxseg5ei16_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei16_v_u32mf2((uint32_t *)(op0), (vuint16mf4_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei16_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei16_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint16mf4_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (size_t)(op8)) +#define vsoxseg6ei16_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei16_v_i32m1((int32_t *)(op0), (vuint16mf2_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei16_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei16_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint16mf2_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (size_t)(op9)) +#define vsoxseg6ei16_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei16_v_i32mf2((int32_t *)(op0), (vuint16mf4_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (size_t)(op8)) +#define vsoxseg6ei16_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei16_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint16mf4_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (size_t)(op9)) +#define vsoxseg6ei16_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei16_v_u32m1((uint32_t *)(op0), (vuint16mf2_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei16_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei16_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint16mf2_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vsoxseg6ei16_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei16_v_u32mf2((uint32_t *)(op0), (vuint16mf4_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (size_t)(op8)) +#define vsoxseg6ei16_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei16_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint16mf4_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vsoxseg7ei16_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei16_v_i32m1((int32_t *)(op0), (vuint16mf2_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei16_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei16_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint16mf2_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (size_t)(op10)) +#define vsoxseg7ei16_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei16_v_i32mf2((int32_t *)(op0), (vuint16mf4_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (size_t)(op9)) +#define vsoxseg7ei16_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei16_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint16mf4_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (size_t)(op10)) +#define vsoxseg7ei16_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei16_v_u32m1((uint32_t *)(op0), (vuint16mf2_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei16_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei16_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint16mf2_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (size_t)(op10)) +#define vsoxseg7ei16_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei16_v_u32mf2((uint32_t *)(op0), (vuint16mf4_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vsoxseg7ei16_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei16_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint16mf4_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (size_t)(op10)) +#define vsoxseg8ei16_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei16_v_i32m1((int32_t *)(op0), (vuint16mf2_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei16_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei16_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint16mf2_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (size_t)(op11)) +#define vsoxseg8ei16_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei16_v_i32mf2((int32_t *)(op0), (vuint16mf4_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (size_t)(op10)) +#define vsoxseg8ei16_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei16_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint16mf4_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (size_t)(op11)) +#define vsoxseg8ei16_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei16_v_u32m1((uint32_t *)(op0), (vuint16mf2_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei16_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei16_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint16mf2_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (size_t)(op11)) +#define vsoxseg8ei16_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei16_v_u32mf2((uint32_t *)(op0), (vuint16mf4_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (size_t)(op10)) +#define vsoxseg8ei16_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei16_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint16mf4_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (size_t)(op11)) +#define vsoxseg2ei32_v_i32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_i32m1((int32_t *)(op0), (vuint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_i32m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (size_t)(op5)) +#define vsoxseg2ei32_v_i32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_i32m2((int32_t *)(op0), (vuint32m2_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_i32m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint32m2_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei32_v_i32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_i32m4((int32_t *)(op0), (vuint32m4_t)(op1), (vint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_i32m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vuint32m4_t)(op2), (vint32m4_t)(op3), (vint32m4_t)(op4), (size_t)(op5)) +#define vsoxseg2ei32_v_i32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_i32mf2((int32_t *)(op0), (vuint32mf2_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_i32mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei32_v_u32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_u32m1((uint32_t *)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_u32m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (size_t)(op5)) +#define vsoxseg2ei32_v_u32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_u32m2((uint32_t *)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_u32m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei32_v_u32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_u32m4((uint32_t *)(op0), (vuint32m4_t)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_u32m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (vuint32m4_t)(op4), (size_t)(op5)) +#define vsoxseg2ei32_v_u32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_u32mf2((uint32_t *)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_u32mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei32_v_i32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei32_v_i32m1((int32_t *)(op0), (vuint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (size_t)(op5)) +#define vsoxseg3ei32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei32_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (size_t)(op6)) +#define vsoxseg3ei32_v_i32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei32_v_i32m2((int32_t *)(op0), (vuint32m2_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei32_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei32_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint32m2_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (vint32m2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei32_v_i32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei32_v_i32mf2((int32_t *)(op0), (vuint32mf2_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei32_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei32_v_u32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei32_v_u32m1((uint32_t *)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (size_t)(op5)) +#define vsoxseg3ei32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei32_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (size_t)(op6)) +#define vsoxseg3ei32_v_u32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei32_v_u32m2((uint32_t *)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei32_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei32_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (vuint32m2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei32_v_u32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei32_v_u32mf2((uint32_t *)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei32_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei32_v_i32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei32_v_i32m1((int32_t *)(op0), (vuint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (size_t)(op6)) +#define vsoxseg4ei32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei32_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (size_t)(op7)) +#define vsoxseg4ei32_v_i32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei32_v_i32m2((int32_t *)(op0), (vuint32m2_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (vint32m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei32_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei32_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint32m2_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (vint32m2_t)(op5), (vint32m2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei32_v_i32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei32_v_i32mf2((int32_t *)(op0), (vuint32mf2_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei32_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei32_v_u32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei32_v_u32m1((uint32_t *)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (size_t)(op6)) +#define vsoxseg4ei32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei32_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vsoxseg4ei32_v_u32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei32_v_u32m2((uint32_t *)(op0), (vuint32m2_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (vuint32m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei32_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei32_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (vuint32m2_t)(op5), (vuint32m2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei32_v_u32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei32_v_u32mf2((uint32_t *)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei32_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei32_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei32_v_i32m1((int32_t *)(op0), (vuint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (size_t)(op7)) +#define vsoxseg5ei32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei32_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (size_t)(op8)) +#define vsoxseg5ei32_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei32_v_i32mf2((int32_t *)(op0), (vuint32mf2_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei32_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (size_t)(op8)) +#define vsoxseg5ei32_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei32_v_u32m1((uint32_t *)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vsoxseg5ei32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei32_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (size_t)(op8)) +#define vsoxseg5ei32_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei32_v_u32mf2((uint32_t *)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei32_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (size_t)(op8)) +#define vsoxseg6ei32_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei32_v_i32m1((int32_t *)(op0), (vuint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei32_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (size_t)(op9)) +#define vsoxseg6ei32_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei32_v_i32mf2((int32_t *)(op0), (vuint32mf2_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (size_t)(op8)) +#define vsoxseg6ei32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei32_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (size_t)(op9)) +#define vsoxseg6ei32_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei32_v_u32m1((uint32_t *)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei32_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vsoxseg6ei32_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei32_v_u32mf2((uint32_t *)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (size_t)(op8)) +#define vsoxseg6ei32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei32_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vsoxseg7ei32_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei32_v_i32m1((int32_t *)(op0), (vuint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei32_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (size_t)(op10)) +#define vsoxseg7ei32_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei32_v_i32mf2((int32_t *)(op0), (vuint32mf2_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (size_t)(op9)) +#define vsoxseg7ei32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei32_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (size_t)(op10)) +#define vsoxseg7ei32_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei32_v_u32m1((uint32_t *)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei32_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (size_t)(op10)) +#define vsoxseg7ei32_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei32_v_u32mf2((uint32_t *)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vsoxseg7ei32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei32_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (size_t)(op10)) +#define vsoxseg8ei32_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei32_v_i32m1((int32_t *)(op0), (vuint32m1_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei32_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (size_t)(op11)) +#define vsoxseg8ei32_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei32_v_i32mf2((int32_t *)(op0), (vuint32mf2_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (size_t)(op10)) +#define vsoxseg8ei32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei32_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (size_t)(op11)) +#define vsoxseg8ei32_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei32_v_u32m1((uint32_t *)(op0), (vuint32m1_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei32_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (size_t)(op11)) +#define vsoxseg8ei32_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei32_v_u32mf2((uint32_t *)(op0), (vuint32mf2_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (size_t)(op10)) +#define vsoxseg8ei32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei32_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (size_t)(op11)) +#define vsoxseg2ei64_v_i32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei64_v_i32m1((int32_t *)(op0), (vuint64m2_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (size_t)(op4)) +#define vsoxseg2ei64_v_i32m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei64_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint64m2_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (size_t)(op5)) +#define vsoxseg2ei64_v_i32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei64_v_i32m2((int32_t *)(op0), (vuint64m4_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei64_v_i32m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei64_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint64m4_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei64_v_i32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei64_v_i32m4((int32_t *)(op0), (vuint64m8_t)(op1), (vint32m4_t)(op2), (vint32m4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei64_v_i32m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei64_v_i32m4_m((vbool8_t)(op0), (int32_t *)(op1), (vuint64m8_t)(op2), (vint32m4_t)(op3), (vint32m4_t)(op4), (size_t)(op5)) +#define vsoxseg2ei64_v_i32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei64_v_i32mf2((int32_t *)(op0), (vuint64m1_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei64_v_i32mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei64_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint64m1_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei64_v_u32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei64_v_u32m1((uint32_t *)(op0), (vuint64m2_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vsoxseg2ei64_v_u32m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei64_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint64m2_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (size_t)(op5)) +#define vsoxseg2ei64_v_u32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei64_v_u32m2((uint32_t *)(op0), (vuint64m4_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei64_v_u32m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei64_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint64m4_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei64_v_u32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei64_v_u32m4((uint32_t *)(op0), (vuint64m8_t)(op1), (vuint32m4_t)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei64_v_u32m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei64_v_u32m4_m((vbool8_t)(op0), (uint32_t *)(op1), (vuint64m8_t)(op2), (vuint32m4_t)(op3), (vuint32m4_t)(op4), (size_t)(op5)) +#define vsoxseg2ei64_v_u32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei64_v_u32mf2((uint32_t *)(op0), (vuint64m1_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei64_v_u32mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei64_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint64m1_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei64_v_i32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei64_v_i32m1((int32_t *)(op0), (vuint64m2_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (size_t)(op5)) +#define vsoxseg3ei64_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei64_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint64m2_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (size_t)(op6)) +#define vsoxseg3ei64_v_i32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei64_v_i32m2((int32_t *)(op0), (vuint64m4_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei64_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei64_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint64m4_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (vint32m2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei64_v_i32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei64_v_i32mf2((int32_t *)(op0), (vuint64m1_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei64_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei64_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint64m1_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei64_v_u32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei64_v_u32m1((uint32_t *)(op0), (vuint64m2_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (size_t)(op5)) +#define vsoxseg3ei64_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei64_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint64m2_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (size_t)(op6)) +#define vsoxseg3ei64_v_u32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei64_v_u32m2((uint32_t *)(op0), (vuint64m4_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei64_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei64_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint64m4_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (vuint32m2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei64_v_u32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei64_v_u32mf2((uint32_t *)(op0), (vuint64m1_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei64_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei64_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint64m1_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei64_v_i32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei64_v_i32m1((int32_t *)(op0), (vuint64m2_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (size_t)(op6)) +#define vsoxseg4ei64_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei64_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint64m2_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (size_t)(op7)) +#define vsoxseg4ei64_v_i32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei64_v_i32m2((int32_t *)(op0), (vuint64m4_t)(op1), (vint32m2_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (vint32m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei64_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei64_v_i32m2_m((vbool16_t)(op0), (int32_t *)(op1), (vuint64m4_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (vint32m2_t)(op5), (vint32m2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei64_v_i32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei64_v_i32mf2((int32_t *)(op0), (vuint64m1_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei64_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei64_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint64m1_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei64_v_u32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei64_v_u32m1((uint32_t *)(op0), (vuint64m2_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (size_t)(op6)) +#define vsoxseg4ei64_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei64_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint64m2_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vsoxseg4ei64_v_u32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei64_v_u32m2((uint32_t *)(op0), (vuint64m4_t)(op1), (vuint32m2_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (vuint32m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei64_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei64_v_u32m2_m((vbool16_t)(op0), (uint32_t *)(op1), (vuint64m4_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (vuint32m2_t)(op5), (vuint32m2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei64_v_u32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei64_v_u32mf2((uint32_t *)(op0), (vuint64m1_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei64_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei64_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint64m1_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei64_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei64_v_i32m1((int32_t *)(op0), (vuint64m2_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (size_t)(op7)) +#define vsoxseg5ei64_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei64_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint64m2_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (size_t)(op8)) +#define vsoxseg5ei64_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei64_v_i32mf2((int32_t *)(op0), (vuint64m1_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei64_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei64_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint64m1_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (size_t)(op8)) +#define vsoxseg5ei64_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei64_v_u32m1((uint32_t *)(op0), (vuint64m2_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vsoxseg5ei64_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei64_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint64m2_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (size_t)(op8)) +#define vsoxseg5ei64_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei64_v_u32mf2((uint32_t *)(op0), (vuint64m1_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei64_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei64_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint64m1_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (size_t)(op8)) +#define vsoxseg6ei64_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei64_v_i32m1((int32_t *)(op0), (vuint64m2_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei64_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei64_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint64m2_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (size_t)(op9)) +#define vsoxseg6ei64_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei64_v_i32mf2((int32_t *)(op0), (vuint64m1_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (size_t)(op8)) +#define vsoxseg6ei64_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei64_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint64m1_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (size_t)(op9)) +#define vsoxseg6ei64_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei64_v_u32m1((uint32_t *)(op0), (vuint64m2_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei64_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei64_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint64m2_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vsoxseg6ei64_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei64_v_u32mf2((uint32_t *)(op0), (vuint64m1_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (size_t)(op8)) +#define vsoxseg6ei64_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei64_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint64m1_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vsoxseg7ei64_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei64_v_i32m1((int32_t *)(op0), (vuint64m2_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei64_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei64_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint64m2_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (size_t)(op10)) +#define vsoxseg7ei64_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei64_v_i32mf2((int32_t *)(op0), (vuint64m1_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (size_t)(op9)) +#define vsoxseg7ei64_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei64_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint64m1_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (size_t)(op10)) +#define vsoxseg7ei64_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei64_v_u32m1((uint32_t *)(op0), (vuint64m2_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei64_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei64_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint64m2_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (size_t)(op10)) +#define vsoxseg7ei64_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei64_v_u32mf2((uint32_t *)(op0), (vuint64m1_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vsoxseg7ei64_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei64_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint64m1_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (size_t)(op10)) +#define vsoxseg8ei64_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei64_v_i32m1((int32_t *)(op0), (vuint64m2_t)(op1), (vint32m1_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei64_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei64_v_i32m1_m((vbool32_t)(op0), (int32_t *)(op1), (vuint64m2_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (size_t)(op11)) +#define vsoxseg8ei64_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei64_v_i32mf2((int32_t *)(op0), (vuint64m1_t)(op1), (vint32mf2_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (size_t)(op10)) +#define vsoxseg8ei64_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei64_v_i32mf2_m((vbool64_t)(op0), (int32_t *)(op1), (vuint64m1_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (size_t)(op11)) +#define vsoxseg8ei64_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei64_v_u32m1((uint32_t *)(op0), (vuint64m2_t)(op1), (vuint32m1_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei64_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei64_v_u32m1_m((vbool32_t)(op0), (uint32_t *)(op1), (vuint64m2_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (size_t)(op11)) +#define vsoxseg8ei64_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei64_v_u32mf2((uint32_t *)(op0), (vuint64m1_t)(op1), (vuint32mf2_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (size_t)(op10)) +#define vsoxseg8ei64_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei64_v_u32mf2_m((vbool64_t)(op0), (uint32_t *)(op1), (vuint64m1_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (size_t)(op11)) +#define vsoxseg2ei8_v_i64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_i64m1((int64_t *)(op0), (vuint8mf8_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_i64m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint8mf8_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (size_t)(op5)) +#define vsoxseg2ei8_v_i64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_i64m2((int64_t *)(op0), (vuint8mf4_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_i64m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint8mf4_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei8_v_i64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_i64m4((int64_t *)(op0), (vuint8mf2_t)(op1), (vint64m4_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_i64m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vuint8mf2_t)(op2), (vint64m4_t)(op3), (vint64m4_t)(op4), (size_t)(op5)) +#define vsoxseg2ei8_v_u64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_u64m1((uint64_t *)(op0), (vuint8mf8_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_u64m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint8mf8_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (size_t)(op5)) +#define vsoxseg2ei8_v_u64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_u64m2((uint64_t *)(op0), (vuint8mf4_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_u64m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint8mf4_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei8_v_u64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_u64m4((uint64_t *)(op0), (vuint8mf2_t)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_u64m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint8mf2_t)(op2), (vuint64m4_t)(op3), (vuint64m4_t)(op4), (size_t)(op5)) +#define vsoxseg3ei8_v_i64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei8_v_i64m1((int64_t *)(op0), (vuint8mf8_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (size_t)(op5)) +#define vsoxseg3ei8_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei8_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint8mf8_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (size_t)(op6)) +#define vsoxseg3ei8_v_i64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei8_v_i64m2((int64_t *)(op0), (vuint8mf4_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei8_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei8_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint8mf4_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (vint64m2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei8_v_u64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei8_v_u64m1((uint64_t *)(op0), (vuint8mf8_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (size_t)(op5)) +#define vsoxseg3ei8_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei8_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint8mf8_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (size_t)(op6)) +#define vsoxseg3ei8_v_u64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei8_v_u64m2((uint64_t *)(op0), (vuint8mf4_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei8_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei8_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint8mf4_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (vuint64m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei8_v_i64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei8_v_i64m1((int64_t *)(op0), (vuint8mf8_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (size_t)(op6)) +#define vsoxseg4ei8_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei8_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint8mf8_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (size_t)(op7)) +#define vsoxseg4ei8_v_i64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei8_v_i64m2((int64_t *)(op0), (vuint8mf4_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (vint64m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei8_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei8_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint8mf4_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (vint64m2_t)(op5), (vint64m2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei8_v_u64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei8_v_u64m1((uint64_t *)(op0), (vuint8mf8_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (size_t)(op6)) +#define vsoxseg4ei8_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei8_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint8mf8_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vsoxseg4ei8_v_u64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei8_v_u64m2((uint64_t *)(op0), (vuint8mf4_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (vuint64m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei8_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei8_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint8mf4_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (vuint64m2_t)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei8_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei8_v_i64m1((int64_t *)(op0), (vuint8mf8_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (size_t)(op7)) +#define vsoxseg5ei8_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei8_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint8mf8_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (size_t)(op8)) +#define vsoxseg5ei8_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei8_v_u64m1((uint64_t *)(op0), (vuint8mf8_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vsoxseg5ei8_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei8_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint8mf8_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei8_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei8_v_i64m1((int64_t *)(op0), (vuint8mf8_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei8_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei8_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint8mf8_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (size_t)(op9)) +#define vsoxseg6ei8_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei8_v_u64m1((uint64_t *)(op0), (vuint8mf8_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei8_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei8_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint8mf8_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei8_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei8_v_i64m1((int64_t *)(op0), (vuint8mf8_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei8_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei8_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint8mf8_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (size_t)(op10)) +#define vsoxseg7ei8_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei8_v_u64m1((uint64_t *)(op0), (vuint8mf8_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei8_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei8_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint8mf8_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei8_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei8_v_i64m1((int64_t *)(op0), (vuint8mf8_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei8_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei8_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint8mf8_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (size_t)(op11)) +#define vsoxseg8ei8_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei8_v_u64m1((uint64_t *)(op0), (vuint8mf8_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei8_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei8_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint8mf8_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (size_t)(op11)) +#define vsoxseg2ei16_v_i64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_i64m1((int64_t *)(op0), (vuint16mf4_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_i64m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint16mf4_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (size_t)(op5)) +#define vsoxseg2ei16_v_i64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_i64m2((int64_t *)(op0), (vuint16mf2_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_i64m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint16mf2_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei16_v_i64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_i64m4((int64_t *)(op0), (vuint16m1_t)(op1), (vint64m4_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_i64m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vuint16m1_t)(op2), (vint64m4_t)(op3), (vint64m4_t)(op4), (size_t)(op5)) +#define vsoxseg2ei16_v_u64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_u64m1((uint64_t *)(op0), (vuint16mf4_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_u64m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint16mf4_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (size_t)(op5)) +#define vsoxseg2ei16_v_u64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_u64m2((uint64_t *)(op0), (vuint16mf2_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_u64m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint16mf2_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei16_v_u64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_u64m4((uint64_t *)(op0), (vuint16m1_t)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_u64m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint16m1_t)(op2), (vuint64m4_t)(op3), (vuint64m4_t)(op4), (size_t)(op5)) +#define vsoxseg3ei16_v_i64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei16_v_i64m1((int64_t *)(op0), (vuint16mf4_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (size_t)(op5)) +#define vsoxseg3ei16_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei16_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint16mf4_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (size_t)(op6)) +#define vsoxseg3ei16_v_i64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei16_v_i64m2((int64_t *)(op0), (vuint16mf2_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei16_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei16_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint16mf2_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (vint64m2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei16_v_u64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei16_v_u64m1((uint64_t *)(op0), (vuint16mf4_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (size_t)(op5)) +#define vsoxseg3ei16_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei16_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint16mf4_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (size_t)(op6)) +#define vsoxseg3ei16_v_u64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei16_v_u64m2((uint64_t *)(op0), (vuint16mf2_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei16_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei16_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint16mf2_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (vuint64m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei16_v_i64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei16_v_i64m1((int64_t *)(op0), (vuint16mf4_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (size_t)(op6)) +#define vsoxseg4ei16_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei16_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint16mf4_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (size_t)(op7)) +#define vsoxseg4ei16_v_i64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei16_v_i64m2((int64_t *)(op0), (vuint16mf2_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (vint64m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei16_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei16_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint16mf2_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (vint64m2_t)(op5), (vint64m2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei16_v_u64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei16_v_u64m1((uint64_t *)(op0), (vuint16mf4_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (size_t)(op6)) +#define vsoxseg4ei16_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei16_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint16mf4_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vsoxseg4ei16_v_u64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei16_v_u64m2((uint64_t *)(op0), (vuint16mf2_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (vuint64m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei16_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei16_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint16mf2_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (vuint64m2_t)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei16_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei16_v_i64m1((int64_t *)(op0), (vuint16mf4_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (size_t)(op7)) +#define vsoxseg5ei16_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei16_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint16mf4_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (size_t)(op8)) +#define vsoxseg5ei16_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei16_v_u64m1((uint64_t *)(op0), (vuint16mf4_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vsoxseg5ei16_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei16_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint16mf4_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei16_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei16_v_i64m1((int64_t *)(op0), (vuint16mf4_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei16_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei16_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint16mf4_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (size_t)(op9)) +#define vsoxseg6ei16_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei16_v_u64m1((uint64_t *)(op0), (vuint16mf4_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei16_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei16_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint16mf4_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei16_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei16_v_i64m1((int64_t *)(op0), (vuint16mf4_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei16_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei16_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint16mf4_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (size_t)(op10)) +#define vsoxseg7ei16_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei16_v_u64m1((uint64_t *)(op0), (vuint16mf4_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei16_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei16_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint16mf4_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei16_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei16_v_i64m1((int64_t *)(op0), (vuint16mf4_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei16_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei16_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint16mf4_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (size_t)(op11)) +#define vsoxseg8ei16_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei16_v_u64m1((uint64_t *)(op0), (vuint16mf4_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei16_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei16_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint16mf4_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (size_t)(op11)) +#define vsoxseg2ei32_v_i64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_i64m1((int64_t *)(op0), (vuint32mf2_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_i64m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint32mf2_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (size_t)(op5)) +#define vsoxseg2ei32_v_i64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_i64m2((int64_t *)(op0), (vuint32m1_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_i64m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint32m1_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei32_v_i64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_i64m4((int64_t *)(op0), (vuint32m2_t)(op1), (vint64m4_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_i64m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vuint32m2_t)(op2), (vint64m4_t)(op3), (vint64m4_t)(op4), (size_t)(op5)) +#define vsoxseg2ei32_v_u64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_u64m1((uint64_t *)(op0), (vuint32mf2_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_u64m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint32mf2_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (size_t)(op5)) +#define vsoxseg2ei32_v_u64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_u64m2((uint64_t *)(op0), (vuint32m1_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_u64m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint32m1_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei32_v_u64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_u64m4((uint64_t *)(op0), (vuint32m2_t)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_u64m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint32m2_t)(op2), (vuint64m4_t)(op3), (vuint64m4_t)(op4), (size_t)(op5)) +#define vsoxseg3ei32_v_i64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei32_v_i64m1((int64_t *)(op0), (vuint32mf2_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (size_t)(op5)) +#define vsoxseg3ei32_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei32_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint32mf2_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (size_t)(op6)) +#define vsoxseg3ei32_v_i64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei32_v_i64m2((int64_t *)(op0), (vuint32m1_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei32_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei32_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint32m1_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (vint64m2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei32_v_u64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei32_v_u64m1((uint64_t *)(op0), (vuint32mf2_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (size_t)(op5)) +#define vsoxseg3ei32_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei32_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint32mf2_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (size_t)(op6)) +#define vsoxseg3ei32_v_u64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei32_v_u64m2((uint64_t *)(op0), (vuint32m1_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei32_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei32_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint32m1_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (vuint64m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei32_v_i64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei32_v_i64m1((int64_t *)(op0), (vuint32mf2_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (size_t)(op6)) +#define vsoxseg4ei32_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei32_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint32mf2_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (size_t)(op7)) +#define vsoxseg4ei32_v_i64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei32_v_i64m2((int64_t *)(op0), (vuint32m1_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (vint64m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei32_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei32_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint32m1_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (vint64m2_t)(op5), (vint64m2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei32_v_u64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei32_v_u64m1((uint64_t *)(op0), (vuint32mf2_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (size_t)(op6)) +#define vsoxseg4ei32_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei32_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint32mf2_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vsoxseg4ei32_v_u64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei32_v_u64m2((uint64_t *)(op0), (vuint32m1_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (vuint64m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei32_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei32_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint32m1_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (vuint64m2_t)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei32_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei32_v_i64m1((int64_t *)(op0), (vuint32mf2_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (size_t)(op7)) +#define vsoxseg5ei32_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei32_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint32mf2_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (size_t)(op8)) +#define vsoxseg5ei32_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei32_v_u64m1((uint64_t *)(op0), (vuint32mf2_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vsoxseg5ei32_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei32_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint32mf2_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei32_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei32_v_i64m1((int64_t *)(op0), (vuint32mf2_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei32_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei32_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint32mf2_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (size_t)(op9)) +#define vsoxseg6ei32_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei32_v_u64m1((uint64_t *)(op0), (vuint32mf2_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei32_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei32_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint32mf2_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei32_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei32_v_i64m1((int64_t *)(op0), (vuint32mf2_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei32_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei32_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint32mf2_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (size_t)(op10)) +#define vsoxseg7ei32_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei32_v_u64m1((uint64_t *)(op0), (vuint32mf2_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei32_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei32_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint32mf2_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei32_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei32_v_i64m1((int64_t *)(op0), (vuint32mf2_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei32_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei32_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint32mf2_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (size_t)(op11)) +#define vsoxseg8ei32_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei32_v_u64m1((uint64_t *)(op0), (vuint32mf2_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei32_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei32_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint32mf2_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (size_t)(op11)) +#define vsoxseg2ei64_v_i64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei64_v_i64m1((int64_t *)(op0), (vuint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (size_t)(op4)) +#define vsoxseg2ei64_v_i64m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei64_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (size_t)(op5)) +#define vsoxseg2ei64_v_i64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei64_v_i64m2((int64_t *)(op0), (vuint64m2_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei64_v_i64m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei64_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint64m2_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei64_v_i64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei64_v_i64m4((int64_t *)(op0), (vuint64m4_t)(op1), (vint64m4_t)(op2), (vint64m4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei64_v_i64m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei64_v_i64m4_m((vbool16_t)(op0), (int64_t *)(op1), (vuint64m4_t)(op2), (vint64m4_t)(op3), (vint64m4_t)(op4), (size_t)(op5)) +#define vsoxseg2ei64_v_u64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei64_v_u64m1((uint64_t *)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vsoxseg2ei64_v_u64m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei64_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (size_t)(op5)) +#define vsoxseg2ei64_v_u64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei64_v_u64m2((uint64_t *)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei64_v_u64m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei64_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei64_v_u64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei64_v_u64m4((uint64_t *)(op0), (vuint64m4_t)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei64_v_u64m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei64_v_u64m4_m((vbool16_t)(op0), (uint64_t *)(op1), (vuint64m4_t)(op2), (vuint64m4_t)(op3), (vuint64m4_t)(op4), (size_t)(op5)) +#define vsoxseg3ei64_v_i64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei64_v_i64m1((int64_t *)(op0), (vuint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (size_t)(op5)) +#define vsoxseg3ei64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei64_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (size_t)(op6)) +#define vsoxseg3ei64_v_i64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei64_v_i64m2((int64_t *)(op0), (vuint64m2_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei64_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei64_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint64m2_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (vint64m2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei64_v_u64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei64_v_u64m1((uint64_t *)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (size_t)(op5)) +#define vsoxseg3ei64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei64_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (size_t)(op6)) +#define vsoxseg3ei64_v_u64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei64_v_u64m2((uint64_t *)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei64_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei64_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (vuint64m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei64_v_i64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei64_v_i64m1((int64_t *)(op0), (vuint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (size_t)(op6)) +#define vsoxseg4ei64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei64_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (size_t)(op7)) +#define vsoxseg4ei64_v_i64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei64_v_i64m2((int64_t *)(op0), (vuint64m2_t)(op1), (vint64m2_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (vint64m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei64_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei64_v_i64m2_m((vbool32_t)(op0), (int64_t *)(op1), (vuint64m2_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (vint64m2_t)(op5), (vint64m2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei64_v_u64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei64_v_u64m1((uint64_t *)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (size_t)(op6)) +#define vsoxseg4ei64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei64_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vsoxseg4ei64_v_u64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei64_v_u64m2((uint64_t *)(op0), (vuint64m2_t)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (vuint64m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei64_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei64_v_u64m2_m((vbool32_t)(op0), (uint64_t *)(op1), (vuint64m2_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (vuint64m2_t)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei64_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei64_v_i64m1((int64_t *)(op0), (vuint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (size_t)(op7)) +#define vsoxseg5ei64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei64_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (size_t)(op8)) +#define vsoxseg5ei64_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei64_v_u64m1((uint64_t *)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vsoxseg5ei64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei64_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei64_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei64_v_i64m1((int64_t *)(op0), (vuint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei64_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (size_t)(op9)) +#define vsoxseg6ei64_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei64_v_u64m1((uint64_t *)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei64_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei64_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei64_v_i64m1((int64_t *)(op0), (vuint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei64_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (size_t)(op10)) +#define vsoxseg7ei64_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei64_v_u64m1((uint64_t *)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei64_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei64_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei64_v_i64m1((int64_t *)(op0), (vuint64m1_t)(op1), (vint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei64_v_i64m1_m((vbool64_t)(op0), (int64_t *)(op1), (vuint64m1_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (size_t)(op11)) +#define vsoxseg8ei64_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei64_v_u64m1((uint64_t *)(op0), (vuint64m1_t)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei64_v_u64m1_m((vbool64_t)(op0), (uint64_t *)(op1), (vuint64m1_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (size_t)(op11)) +#define vlseg2e8_v_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e8_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (const int8_t *)(op2), (size_t)(op3)) +#define vlseg2e8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e8_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vbool8_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (const int8_t *)(op5), (size_t)(op6)) +#define vlseg2e8_v_i8m2(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e8_v_i8m2((vint8m2_t *)(op0), (vint8m2_t *)(op1), (const int8_t *)(op2), (size_t)(op3)) +#define vlseg2e8_v_i8m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e8_v_i8m2_m((vint8m2_t *)(op0), (vint8m2_t *)(op1), (vbool4_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (const int8_t *)(op5), (size_t)(op6)) +#define vlseg2e8_v_i8m4(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e8_v_i8m4((vint8m4_t *)(op0), (vint8m4_t *)(op1), (const int8_t *)(op2), (size_t)(op3)) +#define vlseg2e8_v_i8m4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e8_v_i8m4_m((vint8m4_t *)(op0), (vint8m4_t *)(op1), (vbool2_t)(op2), (vint8m4_t)(op3), (vint8m4_t)(op4), (const int8_t *)(op5), (size_t)(op6)) +#define vlseg2e8_v_i8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e8_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (const int8_t *)(op2), (size_t)(op3)) +#define vlseg2e8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e8_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vbool16_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (const int8_t *)(op5), (size_t)(op6)) +#define vlseg2e8_v_i8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e8_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (const int8_t *)(op2), (size_t)(op3)) +#define vlseg2e8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e8_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vbool32_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (const int8_t *)(op5), (size_t)(op6)) +#define vlseg2e8_v_i8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e8_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (const int8_t *)(op2), (size_t)(op3)) +#define vlseg2e8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e8_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vbool64_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (const int8_t *)(op5), (size_t)(op6)) +#define vlseg2e8_v_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e8_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (const uint8_t *)(op2), (size_t)(op3)) +#define vlseg2e8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e8_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vbool8_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (const uint8_t *)(op5), (size_t)(op6)) +#define vlseg2e8_v_u8m2(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e8_v_u8m2((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (const uint8_t *)(op2), (size_t)(op3)) +#define vlseg2e8_v_u8m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e8_v_u8m2_m((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (vbool4_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (const uint8_t *)(op5), (size_t)(op6)) +#define vlseg2e8_v_u8m4(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e8_v_u8m4((vuint8m4_t *)(op0), (vuint8m4_t *)(op1), (const uint8_t *)(op2), (size_t)(op3)) +#define vlseg2e8_v_u8m4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e8_v_u8m4_m((vuint8m4_t *)(op0), (vuint8m4_t *)(op1), (vbool2_t)(op2), (vuint8m4_t)(op3), (vuint8m4_t)(op4), (const uint8_t *)(op5), (size_t)(op6)) +#define vlseg2e8_v_u8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e8_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (const uint8_t *)(op2), (size_t)(op3)) +#define vlseg2e8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e8_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vbool16_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (const uint8_t *)(op5), (size_t)(op6)) +#define vlseg2e8_v_u8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e8_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (const uint8_t *)(op2), (size_t)(op3)) +#define vlseg2e8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e8_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vbool32_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (const uint8_t *)(op5), (size_t)(op6)) +#define vlseg2e8_v_u8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e8_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (const uint8_t *)(op2), (size_t)(op3)) +#define vlseg2e8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e8_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vbool64_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (const uint8_t *)(op5), (size_t)(op6)) +#define vlseg3e8_v_i8m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg3e8_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (const int8_t *)(op3), (size_t)(op4)) +#define vlseg3e8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg3e8_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vbool8_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (const int8_t *)(op7), (size_t)(op8)) +#define vlseg3e8_v_i8m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg3e8_v_i8m2((vint8m2_t *)(op0), (vint8m2_t *)(op1), (vint8m2_t *)(op2), (const int8_t *)(op3), (size_t)(op4)) +#define vlseg3e8_v_i8m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg3e8_v_i8m2_m((vint8m2_t *)(op0), (vint8m2_t *)(op1), (vint8m2_t *)(op2), (vbool4_t)(op3), (vint8m2_t)(op4), (vint8m2_t)(op5), (vint8m2_t)(op6), (const int8_t *)(op7), (size_t)(op8)) +#define vlseg3e8_v_i8mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg3e8_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (const int8_t *)(op3), (size_t)(op4)) +#define vlseg3e8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg3e8_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vbool16_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (const int8_t *)(op7), (size_t)(op8)) +#define vlseg3e8_v_i8mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg3e8_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (const int8_t *)(op3), (size_t)(op4)) +#define vlseg3e8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg3e8_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vbool32_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (const int8_t *)(op7), (size_t)(op8)) +#define vlseg3e8_v_i8mf8(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg3e8_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (const int8_t *)(op3), (size_t)(op4)) +#define vlseg3e8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg3e8_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vbool64_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (const int8_t *)(op7), (size_t)(op8)) +#define vlseg3e8_v_u8m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg3e8_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (const uint8_t *)(op3), (size_t)(op4)) +#define vlseg3e8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg3e8_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vbool8_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (const uint8_t *)(op7), (size_t)(op8)) +#define vlseg3e8_v_u8m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg3e8_v_u8m2((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (vuint8m2_t *)(op2), (const uint8_t *)(op3), (size_t)(op4)) +#define vlseg3e8_v_u8m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg3e8_v_u8m2_m((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (vuint8m2_t *)(op2), (vbool4_t)(op3), (vuint8m2_t)(op4), (vuint8m2_t)(op5), (vuint8m2_t)(op6), (const uint8_t *)(op7), (size_t)(op8)) +#define vlseg3e8_v_u8mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg3e8_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (const uint8_t *)(op3), (size_t)(op4)) +#define vlseg3e8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg3e8_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vbool16_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (const uint8_t *)(op7), (size_t)(op8)) +#define vlseg3e8_v_u8mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg3e8_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (const uint8_t *)(op3), (size_t)(op4)) +#define vlseg3e8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg3e8_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vbool32_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (const uint8_t *)(op7), (size_t)(op8)) +#define vlseg3e8_v_u8mf8(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg3e8_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (const uint8_t *)(op3), (size_t)(op4)) +#define vlseg3e8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg3e8_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vbool64_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (const uint8_t *)(op7), (size_t)(op8)) +#define vlseg4e8_v_i8m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg4e8_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (const int8_t *)(op4), (size_t)(op5)) +#define vlseg4e8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg4e8_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vbool8_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (const int8_t *)(op9), (size_t)(op10)) +#define vlseg4e8_v_i8m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg4e8_v_i8m2((vint8m2_t *)(op0), (vint8m2_t *)(op1), (vint8m2_t *)(op2), (vint8m2_t *)(op3), (const int8_t *)(op4), (size_t)(op5)) +#define vlseg4e8_v_i8m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg4e8_v_i8m2_m((vint8m2_t *)(op0), (vint8m2_t *)(op1), (vint8m2_t *)(op2), (vint8m2_t *)(op3), (vbool4_t)(op4), (vint8m2_t)(op5), (vint8m2_t)(op6), (vint8m2_t)(op7), (vint8m2_t)(op8), (const int8_t *)(op9), (size_t)(op10)) +#define vlseg4e8_v_i8mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg4e8_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (const int8_t *)(op4), (size_t)(op5)) +#define vlseg4e8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg4e8_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vbool16_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (const int8_t *)(op9), (size_t)(op10)) +#define vlseg4e8_v_i8mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg4e8_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (const int8_t *)(op4), (size_t)(op5)) +#define vlseg4e8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg4e8_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vbool32_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (const int8_t *)(op9), (size_t)(op10)) +#define vlseg4e8_v_i8mf8(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg4e8_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (const int8_t *)(op4), (size_t)(op5)) +#define vlseg4e8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg4e8_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vbool64_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (const int8_t *)(op9), (size_t)(op10)) +#define vlseg4e8_v_u8m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg4e8_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (const uint8_t *)(op4), (size_t)(op5)) +#define vlseg4e8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg4e8_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vbool8_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (const uint8_t *)(op9), (size_t)(op10)) +#define vlseg4e8_v_u8m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg4e8_v_u8m2((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (vuint8m2_t *)(op2), (vuint8m2_t *)(op3), (const uint8_t *)(op4), (size_t)(op5)) +#define vlseg4e8_v_u8m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg4e8_v_u8m2_m((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (vuint8m2_t *)(op2), (vuint8m2_t *)(op3), (vbool4_t)(op4), (vuint8m2_t)(op5), (vuint8m2_t)(op6), (vuint8m2_t)(op7), (vuint8m2_t)(op8), (const uint8_t *)(op9), (size_t)(op10)) +#define vlseg4e8_v_u8mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg4e8_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (const uint8_t *)(op4), (size_t)(op5)) +#define vlseg4e8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg4e8_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vbool16_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (const uint8_t *)(op9), (size_t)(op10)) +#define vlseg4e8_v_u8mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg4e8_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (const uint8_t *)(op4), (size_t)(op5)) +#define vlseg4e8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg4e8_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vbool32_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (const uint8_t *)(op9), (size_t)(op10)) +#define vlseg4e8_v_u8mf8(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg4e8_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (const uint8_t *)(op4), (size_t)(op5)) +#define vlseg4e8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg4e8_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vbool64_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (const uint8_t *)(op9), (size_t)(op10)) +#define vlseg5e8_v_i8m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg5e8_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (const int8_t *)(op5), (size_t)(op6)) +#define vlseg5e8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12) \ +__builtin_rvv_vlseg5e8_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vbool8_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (const int8_t *)(op11), (size_t)(op12)) +#define vlseg5e8_v_i8mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg5e8_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (const int8_t *)(op5), (size_t)(op6)) +#define vlseg5e8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12) \ +__builtin_rvv_vlseg5e8_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vbool16_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (const int8_t *)(op11), (size_t)(op12)) +#define vlseg5e8_v_i8mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg5e8_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (const int8_t *)(op5), (size_t)(op6)) +#define vlseg5e8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12) \ +__builtin_rvv_vlseg5e8_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vbool32_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (const int8_t *)(op11), (size_t)(op12)) +#define vlseg5e8_v_i8mf8(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg5e8_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (const int8_t *)(op5), (size_t)(op6)) +#define vlseg5e8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12) \ +__builtin_rvv_vlseg5e8_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vbool64_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (const int8_t *)(op11), (size_t)(op12)) +#define vlseg5e8_v_u8m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg5e8_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (const uint8_t *)(op5), (size_t)(op6)) +#define vlseg5e8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12) \ +__builtin_rvv_vlseg5e8_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vbool8_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (const uint8_t *)(op11), (size_t)(op12)) +#define vlseg5e8_v_u8mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg5e8_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (const uint8_t *)(op5), (size_t)(op6)) +#define vlseg5e8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12) \ +__builtin_rvv_vlseg5e8_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vbool16_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (const uint8_t *)(op11), (size_t)(op12)) +#define vlseg5e8_v_u8mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg5e8_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (const uint8_t *)(op5), (size_t)(op6)) +#define vlseg5e8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12) \ +__builtin_rvv_vlseg5e8_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vbool32_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (const uint8_t *)(op11), (size_t)(op12)) +#define vlseg5e8_v_u8mf8(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg5e8_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (const uint8_t *)(op5), (size_t)(op6)) +#define vlseg5e8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12) \ +__builtin_rvv_vlseg5e8_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vbool64_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (const uint8_t *)(op11), (size_t)(op12)) +#define vlseg6e8_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg6e8_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (const int8_t *)(op6), (size_t)(op7)) +#define vlseg6e8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14) \ +__builtin_rvv_vlseg6e8_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vbool8_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (vint8m1_t)(op11), (vint8m1_t)(op12), (const int8_t *)(op13), (size_t)(op14)) +#define vlseg6e8_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg6e8_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (const int8_t *)(op6), (size_t)(op7)) +#define vlseg6e8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14) \ +__builtin_rvv_vlseg6e8_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vbool16_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (vint8mf2_t)(op11), (vint8mf2_t)(op12), (const int8_t *)(op13), (size_t)(op14)) +#define vlseg6e8_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg6e8_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (const int8_t *)(op6), (size_t)(op7)) +#define vlseg6e8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14) \ +__builtin_rvv_vlseg6e8_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vbool32_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (vint8mf4_t)(op11), (vint8mf4_t)(op12), (const int8_t *)(op13), (size_t)(op14)) +#define vlseg6e8_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg6e8_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (const int8_t *)(op6), (size_t)(op7)) +#define vlseg6e8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14) \ +__builtin_rvv_vlseg6e8_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vbool64_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (vint8mf8_t)(op11), (vint8mf8_t)(op12), (const int8_t *)(op13), (size_t)(op14)) +#define vlseg6e8_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg6e8_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (const uint8_t *)(op6), (size_t)(op7)) +#define vlseg6e8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14) \ +__builtin_rvv_vlseg6e8_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vbool8_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (vuint8m1_t)(op11), (vuint8m1_t)(op12), (const uint8_t *)(op13), (size_t)(op14)) +#define vlseg6e8_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg6e8_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (const uint8_t *)(op6), (size_t)(op7)) +#define vlseg6e8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14) \ +__builtin_rvv_vlseg6e8_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vbool16_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (vuint8mf2_t)(op11), (vuint8mf2_t)(op12), (const uint8_t *)(op13), (size_t)(op14)) +#define vlseg6e8_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg6e8_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (const uint8_t *)(op6), (size_t)(op7)) +#define vlseg6e8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14) \ +__builtin_rvv_vlseg6e8_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vbool32_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (vuint8mf4_t)(op11), (vuint8mf4_t)(op12), (const uint8_t *)(op13), (size_t)(op14)) +#define vlseg6e8_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg6e8_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (const uint8_t *)(op6), (size_t)(op7)) +#define vlseg6e8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14) \ +__builtin_rvv_vlseg6e8_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vbool64_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (vuint8mf8_t)(op11), (vuint8mf8_t)(op12), (const uint8_t *)(op13), (size_t)(op14)) +#define vlseg7e8_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg7e8_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vint8m1_t *)(op6), (const int8_t *)(op7), (size_t)(op8)) +#define vlseg7e8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16) \ +__builtin_rvv_vlseg7e8_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vint8m1_t *)(op6), (vbool8_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (vint8m1_t)(op11), (vint8m1_t)(op12), (vint8m1_t)(op13), (vint8m1_t)(op14), (const int8_t *)(op15), (size_t)(op16)) +#define vlseg7e8_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg7e8_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vint8mf2_t *)(op6), (const int8_t *)(op7), (size_t)(op8)) +#define vlseg7e8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16) \ +__builtin_rvv_vlseg7e8_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vint8mf2_t *)(op6), (vbool16_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (vint8mf2_t)(op11), (vint8mf2_t)(op12), (vint8mf2_t)(op13), (vint8mf2_t)(op14), (const int8_t *)(op15), (size_t)(op16)) +#define vlseg7e8_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg7e8_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vint8mf4_t *)(op6), (const int8_t *)(op7), (size_t)(op8)) +#define vlseg7e8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16) \ +__builtin_rvv_vlseg7e8_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vint8mf4_t *)(op6), (vbool32_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (vint8mf4_t)(op11), (vint8mf4_t)(op12), (vint8mf4_t)(op13), (vint8mf4_t)(op14), (const int8_t *)(op15), (size_t)(op16)) +#define vlseg7e8_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg7e8_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vint8mf8_t *)(op6), (const int8_t *)(op7), (size_t)(op8)) +#define vlseg7e8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16) \ +__builtin_rvv_vlseg7e8_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vint8mf8_t *)(op6), (vbool64_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (vint8mf8_t)(op11), (vint8mf8_t)(op12), (vint8mf8_t)(op13), (vint8mf8_t)(op14), (const int8_t *)(op15), (size_t)(op16)) +#define vlseg7e8_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg7e8_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vuint8m1_t *)(op6), (const uint8_t *)(op7), (size_t)(op8)) +#define vlseg7e8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16) \ +__builtin_rvv_vlseg7e8_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vuint8m1_t *)(op6), (vbool8_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (vuint8m1_t)(op11), (vuint8m1_t)(op12), (vuint8m1_t)(op13), (vuint8m1_t)(op14), (const uint8_t *)(op15), (size_t)(op16)) +#define vlseg7e8_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg7e8_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vuint8mf2_t *)(op6), (const uint8_t *)(op7), (size_t)(op8)) +#define vlseg7e8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16) \ +__builtin_rvv_vlseg7e8_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vuint8mf2_t *)(op6), (vbool16_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (vuint8mf2_t)(op11), (vuint8mf2_t)(op12), (vuint8mf2_t)(op13), (vuint8mf2_t)(op14), (const uint8_t *)(op15), (size_t)(op16)) +#define vlseg7e8_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg7e8_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vuint8mf4_t *)(op6), (const uint8_t *)(op7), (size_t)(op8)) +#define vlseg7e8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16) \ +__builtin_rvv_vlseg7e8_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vuint8mf4_t *)(op6), (vbool32_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (vuint8mf4_t)(op11), (vuint8mf4_t)(op12), (vuint8mf4_t)(op13), (vuint8mf4_t)(op14), (const uint8_t *)(op15), (size_t)(op16)) +#define vlseg7e8_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg7e8_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vuint8mf8_t *)(op6), (const uint8_t *)(op7), (size_t)(op8)) +#define vlseg7e8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16) \ +__builtin_rvv_vlseg7e8_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vuint8mf8_t *)(op6), (vbool64_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (vuint8mf8_t)(op11), (vuint8mf8_t)(op12), (vuint8mf8_t)(op13), (vuint8mf8_t)(op14), (const uint8_t *)(op15), (size_t)(op16)) +#define vlseg8e8_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg8e8_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vint8m1_t *)(op6), (vint8m1_t *)(op7), (const int8_t *)(op8), (size_t)(op9)) +#define vlseg8e8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18) \ +__builtin_rvv_vlseg8e8_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vint8m1_t *)(op6), (vint8m1_t *)(op7), (vbool8_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (vint8m1_t)(op11), (vint8m1_t)(op12), (vint8m1_t)(op13), (vint8m1_t)(op14), (vint8m1_t)(op15), (vint8m1_t)(op16), (const int8_t *)(op17), (size_t)(op18)) +#define vlseg8e8_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg8e8_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vint8mf2_t *)(op6), (vint8mf2_t *)(op7), (const int8_t *)(op8), (size_t)(op9)) +#define vlseg8e8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18) \ +__builtin_rvv_vlseg8e8_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vint8mf2_t *)(op6), (vint8mf2_t *)(op7), (vbool16_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (vint8mf2_t)(op11), (vint8mf2_t)(op12), (vint8mf2_t)(op13), (vint8mf2_t)(op14), (vint8mf2_t)(op15), (vint8mf2_t)(op16), (const int8_t *)(op17), (size_t)(op18)) +#define vlseg8e8_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg8e8_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vint8mf4_t *)(op6), (vint8mf4_t *)(op7), (const int8_t *)(op8), (size_t)(op9)) +#define vlseg8e8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18) \ +__builtin_rvv_vlseg8e8_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vint8mf4_t *)(op6), (vint8mf4_t *)(op7), (vbool32_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (vint8mf4_t)(op11), (vint8mf4_t)(op12), (vint8mf4_t)(op13), (vint8mf4_t)(op14), (vint8mf4_t)(op15), (vint8mf4_t)(op16), (const int8_t *)(op17), (size_t)(op18)) +#define vlseg8e8_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg8e8_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vint8mf8_t *)(op6), (vint8mf8_t *)(op7), (const int8_t *)(op8), (size_t)(op9)) +#define vlseg8e8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18) \ +__builtin_rvv_vlseg8e8_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vint8mf8_t *)(op6), (vint8mf8_t *)(op7), (vbool64_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (vint8mf8_t)(op11), (vint8mf8_t)(op12), (vint8mf8_t)(op13), (vint8mf8_t)(op14), (vint8mf8_t)(op15), (vint8mf8_t)(op16), (const int8_t *)(op17), (size_t)(op18)) +#define vlseg8e8_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg8e8_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vuint8m1_t *)(op6), (vuint8m1_t *)(op7), (const uint8_t *)(op8), (size_t)(op9)) +#define vlseg8e8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18) \ +__builtin_rvv_vlseg8e8_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vuint8m1_t *)(op6), (vuint8m1_t *)(op7), (vbool8_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (vuint8m1_t)(op11), (vuint8m1_t)(op12), (vuint8m1_t)(op13), (vuint8m1_t)(op14), (vuint8m1_t)(op15), (vuint8m1_t)(op16), (const uint8_t *)(op17), (size_t)(op18)) +#define vlseg8e8_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg8e8_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vuint8mf2_t *)(op6), (vuint8mf2_t *)(op7), (const uint8_t *)(op8), (size_t)(op9)) +#define vlseg8e8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18) \ +__builtin_rvv_vlseg8e8_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vuint8mf2_t *)(op6), (vuint8mf2_t *)(op7), (vbool16_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (vuint8mf2_t)(op11), (vuint8mf2_t)(op12), (vuint8mf2_t)(op13), (vuint8mf2_t)(op14), (vuint8mf2_t)(op15), (vuint8mf2_t)(op16), (const uint8_t *)(op17), (size_t)(op18)) +#define vlseg8e8_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg8e8_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vuint8mf4_t *)(op6), (vuint8mf4_t *)(op7), (const uint8_t *)(op8), (size_t)(op9)) +#define vlseg8e8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18) \ +__builtin_rvv_vlseg8e8_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vuint8mf4_t *)(op6), (vuint8mf4_t *)(op7), (vbool32_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (vuint8mf4_t)(op11), (vuint8mf4_t)(op12), (vuint8mf4_t)(op13), (vuint8mf4_t)(op14), (vuint8mf4_t)(op15), (vuint8mf4_t)(op16), (const uint8_t *)(op17), (size_t)(op18)) +#define vlseg8e8_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg8e8_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vuint8mf8_t *)(op6), (vuint8mf8_t *)(op7), (const uint8_t *)(op8), (size_t)(op9)) +#define vlseg8e8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18) \ +__builtin_rvv_vlseg8e8_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vuint8mf8_t *)(op6), (vuint8mf8_t *)(op7), (vbool64_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (vuint8mf8_t)(op11), (vuint8mf8_t)(op12), (vuint8mf8_t)(op13), (vuint8mf8_t)(op14), (vuint8mf8_t)(op15), (vuint8mf8_t)(op16), (const uint8_t *)(op17), (size_t)(op18)) +#define vlseg2e16_v_i16m1(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e16_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (const int16_t *)(op2), (size_t)(op3)) +#define vlseg2e16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e16_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vbool16_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (const int16_t *)(op5), (size_t)(op6)) +#define vlseg2e16_v_i16m2(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e16_v_i16m2((vint16m2_t *)(op0), (vint16m2_t *)(op1), (const int16_t *)(op2), (size_t)(op3)) +#define vlseg2e16_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e16_v_i16m2_m((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vbool8_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (const int16_t *)(op5), (size_t)(op6)) +#define vlseg2e16_v_i16m4(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e16_v_i16m4((vint16m4_t *)(op0), (vint16m4_t *)(op1), (const int16_t *)(op2), (size_t)(op3)) +#define vlseg2e16_v_i16m4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e16_v_i16m4_m((vint16m4_t *)(op0), (vint16m4_t *)(op1), (vbool4_t)(op2), (vint16m4_t)(op3), (vint16m4_t)(op4), (const int16_t *)(op5), (size_t)(op6)) +#define vlseg2e16_v_i16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e16_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (const int16_t *)(op2), (size_t)(op3)) +#define vlseg2e16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e16_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vbool32_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (const int16_t *)(op5), (size_t)(op6)) +#define vlseg2e16_v_i16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e16_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (const int16_t *)(op2), (size_t)(op3)) +#define vlseg2e16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e16_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vbool64_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (const int16_t *)(op5), (size_t)(op6)) +#define vlseg2e16_v_u16m1(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e16_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (const uint16_t *)(op2), (size_t)(op3)) +#define vlseg2e16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e16_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vbool16_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (const uint16_t *)(op5), (size_t)(op6)) +#define vlseg2e16_v_u16m2(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e16_v_u16m2((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (const uint16_t *)(op2), (size_t)(op3)) +#define vlseg2e16_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e16_v_u16m2_m((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vbool8_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (const uint16_t *)(op5), (size_t)(op6)) +#define vlseg2e16_v_u16m4(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e16_v_u16m4((vuint16m4_t *)(op0), (vuint16m4_t *)(op1), (const uint16_t *)(op2), (size_t)(op3)) +#define vlseg2e16_v_u16m4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e16_v_u16m4_m((vuint16m4_t *)(op0), (vuint16m4_t *)(op1), (vbool4_t)(op2), (vuint16m4_t)(op3), (vuint16m4_t)(op4), (const uint16_t *)(op5), (size_t)(op6)) +#define vlseg2e16_v_u16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e16_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (const uint16_t *)(op2), (size_t)(op3)) +#define vlseg2e16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e16_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vbool32_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (const uint16_t *)(op5), (size_t)(op6)) +#define vlseg2e16_v_u16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e16_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (const uint16_t *)(op2), (size_t)(op3)) +#define vlseg2e16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e16_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vbool64_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (const uint16_t *)(op5), (size_t)(op6)) +#define vlseg3e16_v_i16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg3e16_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (const int16_t *)(op3), (size_t)(op4)) +#define vlseg3e16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg3e16_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vbool16_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (const int16_t *)(op7), (size_t)(op8)) +#define vlseg3e16_v_i16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg3e16_v_i16m2((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vint16m2_t *)(op2), (const int16_t *)(op3), (size_t)(op4)) +#define vlseg3e16_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg3e16_v_i16m2_m((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vint16m2_t *)(op2), (vbool8_t)(op3), (vint16m2_t)(op4), (vint16m2_t)(op5), (vint16m2_t)(op6), (const int16_t *)(op7), (size_t)(op8)) +#define vlseg3e16_v_i16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg3e16_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (const int16_t *)(op3), (size_t)(op4)) +#define vlseg3e16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg3e16_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vbool32_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (const int16_t *)(op7), (size_t)(op8)) +#define vlseg3e16_v_i16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg3e16_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (const int16_t *)(op3), (size_t)(op4)) +#define vlseg3e16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg3e16_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vbool64_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (const int16_t *)(op7), (size_t)(op8)) +#define vlseg3e16_v_u16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg3e16_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (const uint16_t *)(op3), (size_t)(op4)) +#define vlseg3e16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg3e16_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vbool16_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (const uint16_t *)(op7), (size_t)(op8)) +#define vlseg3e16_v_u16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg3e16_v_u16m2((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vuint16m2_t *)(op2), (const uint16_t *)(op3), (size_t)(op4)) +#define vlseg3e16_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg3e16_v_u16m2_m((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vuint16m2_t *)(op2), (vbool8_t)(op3), (vuint16m2_t)(op4), (vuint16m2_t)(op5), (vuint16m2_t)(op6), (const uint16_t *)(op7), (size_t)(op8)) +#define vlseg3e16_v_u16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg3e16_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (const uint16_t *)(op3), (size_t)(op4)) +#define vlseg3e16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg3e16_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vbool32_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (const uint16_t *)(op7), (size_t)(op8)) +#define vlseg3e16_v_u16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg3e16_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (const uint16_t *)(op3), (size_t)(op4)) +#define vlseg3e16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg3e16_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vbool64_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (const uint16_t *)(op7), (size_t)(op8)) +#define vlseg4e16_v_i16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg4e16_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (const int16_t *)(op4), (size_t)(op5)) +#define vlseg4e16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg4e16_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vbool16_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (const int16_t *)(op9), (size_t)(op10)) +#define vlseg4e16_v_i16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg4e16_v_i16m2((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vint16m2_t *)(op2), (vint16m2_t *)(op3), (const int16_t *)(op4), (size_t)(op5)) +#define vlseg4e16_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg4e16_v_i16m2_m((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vint16m2_t *)(op2), (vint16m2_t *)(op3), (vbool8_t)(op4), (vint16m2_t)(op5), (vint16m2_t)(op6), (vint16m2_t)(op7), (vint16m2_t)(op8), (const int16_t *)(op9), (size_t)(op10)) +#define vlseg4e16_v_i16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg4e16_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (const int16_t *)(op4), (size_t)(op5)) +#define vlseg4e16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg4e16_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vbool32_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (const int16_t *)(op9), (size_t)(op10)) +#define vlseg4e16_v_i16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg4e16_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (const int16_t *)(op4), (size_t)(op5)) +#define vlseg4e16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg4e16_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vbool64_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (const int16_t *)(op9), (size_t)(op10)) +#define vlseg4e16_v_u16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg4e16_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (const uint16_t *)(op4), (size_t)(op5)) +#define vlseg4e16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg4e16_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vbool16_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (const uint16_t *)(op9), (size_t)(op10)) +#define vlseg4e16_v_u16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg4e16_v_u16m2((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vuint16m2_t *)(op2), (vuint16m2_t *)(op3), (const uint16_t *)(op4), (size_t)(op5)) +#define vlseg4e16_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg4e16_v_u16m2_m((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vuint16m2_t *)(op2), (vuint16m2_t *)(op3), (vbool8_t)(op4), (vuint16m2_t)(op5), (vuint16m2_t)(op6), (vuint16m2_t)(op7), (vuint16m2_t)(op8), (const uint16_t *)(op9), (size_t)(op10)) +#define vlseg4e16_v_u16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg4e16_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (const uint16_t *)(op4), (size_t)(op5)) +#define vlseg4e16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg4e16_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vbool32_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (const uint16_t *)(op9), (size_t)(op10)) +#define vlseg4e16_v_u16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg4e16_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (const uint16_t *)(op4), (size_t)(op5)) +#define vlseg4e16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg4e16_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vbool64_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (const uint16_t *)(op9), (size_t)(op10)) +#define vlseg5e16_v_i16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg5e16_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (const int16_t *)(op5), (size_t)(op6)) +#define vlseg5e16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12) \ +__builtin_rvv_vlseg5e16_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vbool16_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (const int16_t *)(op11), (size_t)(op12)) +#define vlseg5e16_v_i16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg5e16_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (const int16_t *)(op5), (size_t)(op6)) +#define vlseg5e16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12) \ +__builtin_rvv_vlseg5e16_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vbool32_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (const int16_t *)(op11), (size_t)(op12)) +#define vlseg5e16_v_i16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg5e16_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (const int16_t *)(op5), (size_t)(op6)) +#define vlseg5e16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12) \ +__builtin_rvv_vlseg5e16_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vbool64_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (const int16_t *)(op11), (size_t)(op12)) +#define vlseg5e16_v_u16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg5e16_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (const uint16_t *)(op5), (size_t)(op6)) +#define vlseg5e16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12) \ +__builtin_rvv_vlseg5e16_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vbool16_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (const uint16_t *)(op11), (size_t)(op12)) +#define vlseg5e16_v_u16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg5e16_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (const uint16_t *)(op5), (size_t)(op6)) +#define vlseg5e16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12) \ +__builtin_rvv_vlseg5e16_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vbool32_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (const uint16_t *)(op11), (size_t)(op12)) +#define vlseg5e16_v_u16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg5e16_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (const uint16_t *)(op5), (size_t)(op6)) +#define vlseg5e16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12) \ +__builtin_rvv_vlseg5e16_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vbool64_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (const uint16_t *)(op11), (size_t)(op12)) +#define vlseg6e16_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg6e16_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (const int16_t *)(op6), (size_t)(op7)) +#define vlseg6e16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14) \ +__builtin_rvv_vlseg6e16_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vbool16_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (vint16m1_t)(op11), (vint16m1_t)(op12), (const int16_t *)(op13), (size_t)(op14)) +#define vlseg6e16_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg6e16_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (const int16_t *)(op6), (size_t)(op7)) +#define vlseg6e16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14) \ +__builtin_rvv_vlseg6e16_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vbool32_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (vint16mf2_t)(op11), (vint16mf2_t)(op12), (const int16_t *)(op13), (size_t)(op14)) +#define vlseg6e16_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg6e16_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (const int16_t *)(op6), (size_t)(op7)) +#define vlseg6e16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14) \ +__builtin_rvv_vlseg6e16_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vbool64_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (vint16mf4_t)(op11), (vint16mf4_t)(op12), (const int16_t *)(op13), (size_t)(op14)) +#define vlseg6e16_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg6e16_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (const uint16_t *)(op6), (size_t)(op7)) +#define vlseg6e16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14) \ +__builtin_rvv_vlseg6e16_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vbool16_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (vuint16m1_t)(op11), (vuint16m1_t)(op12), (const uint16_t *)(op13), (size_t)(op14)) +#define vlseg6e16_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg6e16_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (const uint16_t *)(op6), (size_t)(op7)) +#define vlseg6e16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14) \ +__builtin_rvv_vlseg6e16_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vbool32_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (vuint16mf2_t)(op11), (vuint16mf2_t)(op12), (const uint16_t *)(op13), (size_t)(op14)) +#define vlseg6e16_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg6e16_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (const uint16_t *)(op6), (size_t)(op7)) +#define vlseg6e16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14) \ +__builtin_rvv_vlseg6e16_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vbool64_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (vuint16mf4_t)(op11), (vuint16mf4_t)(op12), (const uint16_t *)(op13), (size_t)(op14)) +#define vlseg7e16_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg7e16_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vint16m1_t *)(op6), (const int16_t *)(op7), (size_t)(op8)) +#define vlseg7e16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16) \ +__builtin_rvv_vlseg7e16_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vint16m1_t *)(op6), (vbool16_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (vint16m1_t)(op11), (vint16m1_t)(op12), (vint16m1_t)(op13), (vint16m1_t)(op14), (const int16_t *)(op15), (size_t)(op16)) +#define vlseg7e16_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg7e16_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vint16mf2_t *)(op6), (const int16_t *)(op7), (size_t)(op8)) +#define vlseg7e16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16) \ +__builtin_rvv_vlseg7e16_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vint16mf2_t *)(op6), (vbool32_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (vint16mf2_t)(op11), (vint16mf2_t)(op12), (vint16mf2_t)(op13), (vint16mf2_t)(op14), (const int16_t *)(op15), (size_t)(op16)) +#define vlseg7e16_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg7e16_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vint16mf4_t *)(op6), (const int16_t *)(op7), (size_t)(op8)) +#define vlseg7e16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16) \ +__builtin_rvv_vlseg7e16_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vint16mf4_t *)(op6), (vbool64_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (vint16mf4_t)(op11), (vint16mf4_t)(op12), (vint16mf4_t)(op13), (vint16mf4_t)(op14), (const int16_t *)(op15), (size_t)(op16)) +#define vlseg7e16_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg7e16_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vuint16m1_t *)(op6), (const uint16_t *)(op7), (size_t)(op8)) +#define vlseg7e16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16) \ +__builtin_rvv_vlseg7e16_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vuint16m1_t *)(op6), (vbool16_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (vuint16m1_t)(op11), (vuint16m1_t)(op12), (vuint16m1_t)(op13), (vuint16m1_t)(op14), (const uint16_t *)(op15), (size_t)(op16)) +#define vlseg7e16_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg7e16_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vuint16mf2_t *)(op6), (const uint16_t *)(op7), (size_t)(op8)) +#define vlseg7e16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16) \ +__builtin_rvv_vlseg7e16_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vuint16mf2_t *)(op6), (vbool32_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (vuint16mf2_t)(op11), (vuint16mf2_t)(op12), (vuint16mf2_t)(op13), (vuint16mf2_t)(op14), (const uint16_t *)(op15), (size_t)(op16)) +#define vlseg7e16_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg7e16_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vuint16mf4_t *)(op6), (const uint16_t *)(op7), (size_t)(op8)) +#define vlseg7e16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16) \ +__builtin_rvv_vlseg7e16_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vuint16mf4_t *)(op6), (vbool64_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (vuint16mf4_t)(op11), (vuint16mf4_t)(op12), (vuint16mf4_t)(op13), (vuint16mf4_t)(op14), (const uint16_t *)(op15), (size_t)(op16)) +#define vlseg8e16_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg8e16_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vint16m1_t *)(op6), (vint16m1_t *)(op7), (const int16_t *)(op8), (size_t)(op9)) +#define vlseg8e16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18) \ +__builtin_rvv_vlseg8e16_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vint16m1_t *)(op6), (vint16m1_t *)(op7), (vbool16_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (vint16m1_t)(op11), (vint16m1_t)(op12), (vint16m1_t)(op13), (vint16m1_t)(op14), (vint16m1_t)(op15), (vint16m1_t)(op16), (const int16_t *)(op17), (size_t)(op18)) +#define vlseg8e16_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg8e16_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vint16mf2_t *)(op6), (vint16mf2_t *)(op7), (const int16_t *)(op8), (size_t)(op9)) +#define vlseg8e16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18) \ +__builtin_rvv_vlseg8e16_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vint16mf2_t *)(op6), (vint16mf2_t *)(op7), (vbool32_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (vint16mf2_t)(op11), (vint16mf2_t)(op12), (vint16mf2_t)(op13), (vint16mf2_t)(op14), (vint16mf2_t)(op15), (vint16mf2_t)(op16), (const int16_t *)(op17), (size_t)(op18)) +#define vlseg8e16_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg8e16_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vint16mf4_t *)(op6), (vint16mf4_t *)(op7), (const int16_t *)(op8), (size_t)(op9)) +#define vlseg8e16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18) \ +__builtin_rvv_vlseg8e16_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vint16mf4_t *)(op6), (vint16mf4_t *)(op7), (vbool64_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (vint16mf4_t)(op11), (vint16mf4_t)(op12), (vint16mf4_t)(op13), (vint16mf4_t)(op14), (vint16mf4_t)(op15), (vint16mf4_t)(op16), (const int16_t *)(op17), (size_t)(op18)) +#define vlseg8e16_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg8e16_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vuint16m1_t *)(op6), (vuint16m1_t *)(op7), (const uint16_t *)(op8), (size_t)(op9)) +#define vlseg8e16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18) \ +__builtin_rvv_vlseg8e16_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vuint16m1_t *)(op6), (vuint16m1_t *)(op7), (vbool16_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (vuint16m1_t)(op11), (vuint16m1_t)(op12), (vuint16m1_t)(op13), (vuint16m1_t)(op14), (vuint16m1_t)(op15), (vuint16m1_t)(op16), (const uint16_t *)(op17), (size_t)(op18)) +#define vlseg8e16_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg8e16_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vuint16mf2_t *)(op6), (vuint16mf2_t *)(op7), (const uint16_t *)(op8), (size_t)(op9)) +#define vlseg8e16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18) \ +__builtin_rvv_vlseg8e16_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vuint16mf2_t *)(op6), (vuint16mf2_t *)(op7), (vbool32_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (vuint16mf2_t)(op11), (vuint16mf2_t)(op12), (vuint16mf2_t)(op13), (vuint16mf2_t)(op14), (vuint16mf2_t)(op15), (vuint16mf2_t)(op16), (const uint16_t *)(op17), (size_t)(op18)) +#define vlseg8e16_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg8e16_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vuint16mf4_t *)(op6), (vuint16mf4_t *)(op7), (const uint16_t *)(op8), (size_t)(op9)) +#define vlseg8e16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18) \ +__builtin_rvv_vlseg8e16_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vuint16mf4_t *)(op6), (vuint16mf4_t *)(op7), (vbool64_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (vuint16mf4_t)(op11), (vuint16mf4_t)(op12), (vuint16mf4_t)(op13), (vuint16mf4_t)(op14), (vuint16mf4_t)(op15), (vuint16mf4_t)(op16), (const uint16_t *)(op17), (size_t)(op18)) +#define vlseg2e32_v_i32m1(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e32_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (const int32_t *)(op2), (size_t)(op3)) +#define vlseg2e32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e32_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vbool32_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (const int32_t *)(op5), (size_t)(op6)) +#define vlseg2e32_v_i32m2(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e32_v_i32m2((vint32m2_t *)(op0), (vint32m2_t *)(op1), (const int32_t *)(op2), (size_t)(op3)) +#define vlseg2e32_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e32_v_i32m2_m((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vbool16_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (const int32_t *)(op5), (size_t)(op6)) +#define vlseg2e32_v_i32m4(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e32_v_i32m4((vint32m4_t *)(op0), (vint32m4_t *)(op1), (const int32_t *)(op2), (size_t)(op3)) +#define vlseg2e32_v_i32m4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e32_v_i32m4_m((vint32m4_t *)(op0), (vint32m4_t *)(op1), (vbool8_t)(op2), (vint32m4_t)(op3), (vint32m4_t)(op4), (const int32_t *)(op5), (size_t)(op6)) +#define vlseg2e32_v_i32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e32_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (const int32_t *)(op2), (size_t)(op3)) +#define vlseg2e32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e32_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vbool64_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (const int32_t *)(op5), (size_t)(op6)) +#define vlseg2e32_v_u32m1(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e32_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (const uint32_t *)(op2), (size_t)(op3)) +#define vlseg2e32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e32_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vbool32_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (const uint32_t *)(op5), (size_t)(op6)) +#define vlseg2e32_v_u32m2(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e32_v_u32m2((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (const uint32_t *)(op2), (size_t)(op3)) +#define vlseg2e32_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e32_v_u32m2_m((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vbool16_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (const uint32_t *)(op5), (size_t)(op6)) +#define vlseg2e32_v_u32m4(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e32_v_u32m4((vuint32m4_t *)(op0), (vuint32m4_t *)(op1), (const uint32_t *)(op2), (size_t)(op3)) +#define vlseg2e32_v_u32m4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e32_v_u32m4_m((vuint32m4_t *)(op0), (vuint32m4_t *)(op1), (vbool8_t)(op2), (vuint32m4_t)(op3), (vuint32m4_t)(op4), (const uint32_t *)(op5), (size_t)(op6)) +#define vlseg2e32_v_u32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e32_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (const uint32_t *)(op2), (size_t)(op3)) +#define vlseg2e32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e32_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vbool64_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (const uint32_t *)(op5), (size_t)(op6)) +#define vlseg3e32_v_i32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg3e32_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (const int32_t *)(op3), (size_t)(op4)) +#define vlseg3e32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg3e32_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vbool32_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (const int32_t *)(op7), (size_t)(op8)) +#define vlseg3e32_v_i32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg3e32_v_i32m2((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vint32m2_t *)(op2), (const int32_t *)(op3), (size_t)(op4)) +#define vlseg3e32_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg3e32_v_i32m2_m((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vint32m2_t *)(op2), (vbool16_t)(op3), (vint32m2_t)(op4), (vint32m2_t)(op5), (vint32m2_t)(op6), (const int32_t *)(op7), (size_t)(op8)) +#define vlseg3e32_v_i32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg3e32_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (const int32_t *)(op3), (size_t)(op4)) +#define vlseg3e32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg3e32_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vbool64_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (const int32_t *)(op7), (size_t)(op8)) +#define vlseg3e32_v_u32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg3e32_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (const uint32_t *)(op3), (size_t)(op4)) +#define vlseg3e32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg3e32_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vbool32_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (const uint32_t *)(op7), (size_t)(op8)) +#define vlseg3e32_v_u32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg3e32_v_u32m2((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vuint32m2_t *)(op2), (const uint32_t *)(op3), (size_t)(op4)) +#define vlseg3e32_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg3e32_v_u32m2_m((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vuint32m2_t *)(op2), (vbool16_t)(op3), (vuint32m2_t)(op4), (vuint32m2_t)(op5), (vuint32m2_t)(op6), (const uint32_t *)(op7), (size_t)(op8)) +#define vlseg3e32_v_u32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg3e32_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (const uint32_t *)(op3), (size_t)(op4)) +#define vlseg3e32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg3e32_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vbool64_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (const uint32_t *)(op7), (size_t)(op8)) +#define vlseg4e32_v_i32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg4e32_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (const int32_t *)(op4), (size_t)(op5)) +#define vlseg4e32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg4e32_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vbool32_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (const int32_t *)(op9), (size_t)(op10)) +#define vlseg4e32_v_i32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg4e32_v_i32m2((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vint32m2_t *)(op2), (vint32m2_t *)(op3), (const int32_t *)(op4), (size_t)(op5)) +#define vlseg4e32_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg4e32_v_i32m2_m((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vint32m2_t *)(op2), (vint32m2_t *)(op3), (vbool16_t)(op4), (vint32m2_t)(op5), (vint32m2_t)(op6), (vint32m2_t)(op7), (vint32m2_t)(op8), (const int32_t *)(op9), (size_t)(op10)) +#define vlseg4e32_v_i32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg4e32_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (const int32_t *)(op4), (size_t)(op5)) +#define vlseg4e32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg4e32_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vbool64_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (const int32_t *)(op9), (size_t)(op10)) +#define vlseg4e32_v_u32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg4e32_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (const uint32_t *)(op4), (size_t)(op5)) +#define vlseg4e32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg4e32_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vbool32_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (const uint32_t *)(op9), (size_t)(op10)) +#define vlseg4e32_v_u32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg4e32_v_u32m2((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vuint32m2_t *)(op2), (vuint32m2_t *)(op3), (const uint32_t *)(op4), (size_t)(op5)) +#define vlseg4e32_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg4e32_v_u32m2_m((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vuint32m2_t *)(op2), (vuint32m2_t *)(op3), (vbool16_t)(op4), (vuint32m2_t)(op5), (vuint32m2_t)(op6), (vuint32m2_t)(op7), (vuint32m2_t)(op8), (const uint32_t *)(op9), (size_t)(op10)) +#define vlseg4e32_v_u32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg4e32_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (const uint32_t *)(op4), (size_t)(op5)) +#define vlseg4e32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg4e32_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vbool64_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (const uint32_t *)(op9), (size_t)(op10)) +#define vlseg5e32_v_i32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg5e32_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (const int32_t *)(op5), (size_t)(op6)) +#define vlseg5e32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12) \ +__builtin_rvv_vlseg5e32_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vbool32_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (const int32_t *)(op11), (size_t)(op12)) +#define vlseg5e32_v_i32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg5e32_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (const int32_t *)(op5), (size_t)(op6)) +#define vlseg5e32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12) \ +__builtin_rvv_vlseg5e32_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vbool64_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (const int32_t *)(op11), (size_t)(op12)) +#define vlseg5e32_v_u32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg5e32_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (const uint32_t *)(op5), (size_t)(op6)) +#define vlseg5e32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12) \ +__builtin_rvv_vlseg5e32_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vbool32_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (const uint32_t *)(op11), (size_t)(op12)) +#define vlseg5e32_v_u32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg5e32_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (const uint32_t *)(op5), (size_t)(op6)) +#define vlseg5e32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12) \ +__builtin_rvv_vlseg5e32_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vbool64_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (const uint32_t *)(op11), (size_t)(op12)) +#define vlseg6e32_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg6e32_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (const int32_t *)(op6), (size_t)(op7)) +#define vlseg6e32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14) \ +__builtin_rvv_vlseg6e32_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vbool32_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (vint32m1_t)(op11), (vint32m1_t)(op12), (const int32_t *)(op13), (size_t)(op14)) +#define vlseg6e32_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg6e32_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (const int32_t *)(op6), (size_t)(op7)) +#define vlseg6e32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14) \ +__builtin_rvv_vlseg6e32_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vbool64_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (vint32mf2_t)(op11), (vint32mf2_t)(op12), (const int32_t *)(op13), (size_t)(op14)) +#define vlseg6e32_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg6e32_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (const uint32_t *)(op6), (size_t)(op7)) +#define vlseg6e32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14) \ +__builtin_rvv_vlseg6e32_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vbool32_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (vuint32m1_t)(op11), (vuint32m1_t)(op12), (const uint32_t *)(op13), (size_t)(op14)) +#define vlseg6e32_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg6e32_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (const uint32_t *)(op6), (size_t)(op7)) +#define vlseg6e32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14) \ +__builtin_rvv_vlseg6e32_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vbool64_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (vuint32mf2_t)(op11), (vuint32mf2_t)(op12), (const uint32_t *)(op13), (size_t)(op14)) +#define vlseg7e32_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg7e32_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vint32m1_t *)(op6), (const int32_t *)(op7), (size_t)(op8)) +#define vlseg7e32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16) \ +__builtin_rvv_vlseg7e32_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vint32m1_t *)(op6), (vbool32_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (vint32m1_t)(op11), (vint32m1_t)(op12), (vint32m1_t)(op13), (vint32m1_t)(op14), (const int32_t *)(op15), (size_t)(op16)) +#define vlseg7e32_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg7e32_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vint32mf2_t *)(op6), (const int32_t *)(op7), (size_t)(op8)) +#define vlseg7e32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16) \ +__builtin_rvv_vlseg7e32_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vint32mf2_t *)(op6), (vbool64_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (vint32mf2_t)(op11), (vint32mf2_t)(op12), (vint32mf2_t)(op13), (vint32mf2_t)(op14), (const int32_t *)(op15), (size_t)(op16)) +#define vlseg7e32_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg7e32_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vuint32m1_t *)(op6), (const uint32_t *)(op7), (size_t)(op8)) +#define vlseg7e32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16) \ +__builtin_rvv_vlseg7e32_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vuint32m1_t *)(op6), (vbool32_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (vuint32m1_t)(op11), (vuint32m1_t)(op12), (vuint32m1_t)(op13), (vuint32m1_t)(op14), (const uint32_t *)(op15), (size_t)(op16)) +#define vlseg7e32_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg7e32_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vuint32mf2_t *)(op6), (const uint32_t *)(op7), (size_t)(op8)) +#define vlseg7e32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16) \ +__builtin_rvv_vlseg7e32_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vuint32mf2_t *)(op6), (vbool64_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (vuint32mf2_t)(op11), (vuint32mf2_t)(op12), (vuint32mf2_t)(op13), (vuint32mf2_t)(op14), (const uint32_t *)(op15), (size_t)(op16)) +#define vlseg8e32_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg8e32_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vint32m1_t *)(op6), (vint32m1_t *)(op7), (const int32_t *)(op8), (size_t)(op9)) +#define vlseg8e32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18) \ +__builtin_rvv_vlseg8e32_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vint32m1_t *)(op6), (vint32m1_t *)(op7), (vbool32_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (vint32m1_t)(op11), (vint32m1_t)(op12), (vint32m1_t)(op13), (vint32m1_t)(op14), (vint32m1_t)(op15), (vint32m1_t)(op16), (const int32_t *)(op17), (size_t)(op18)) +#define vlseg8e32_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg8e32_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vint32mf2_t *)(op6), (vint32mf2_t *)(op7), (const int32_t *)(op8), (size_t)(op9)) +#define vlseg8e32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18) \ +__builtin_rvv_vlseg8e32_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vint32mf2_t *)(op6), (vint32mf2_t *)(op7), (vbool64_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (vint32mf2_t)(op11), (vint32mf2_t)(op12), (vint32mf2_t)(op13), (vint32mf2_t)(op14), (vint32mf2_t)(op15), (vint32mf2_t)(op16), (const int32_t *)(op17), (size_t)(op18)) +#define vlseg8e32_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg8e32_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vuint32m1_t *)(op6), (vuint32m1_t *)(op7), (const uint32_t *)(op8), (size_t)(op9)) +#define vlseg8e32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18) \ +__builtin_rvv_vlseg8e32_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vuint32m1_t *)(op6), (vuint32m1_t *)(op7), (vbool32_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (vuint32m1_t)(op11), (vuint32m1_t)(op12), (vuint32m1_t)(op13), (vuint32m1_t)(op14), (vuint32m1_t)(op15), (vuint32m1_t)(op16), (const uint32_t *)(op17), (size_t)(op18)) +#define vlseg8e32_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg8e32_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vuint32mf2_t *)(op6), (vuint32mf2_t *)(op7), (const uint32_t *)(op8), (size_t)(op9)) +#define vlseg8e32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18) \ +__builtin_rvv_vlseg8e32_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vuint32mf2_t *)(op6), (vuint32mf2_t *)(op7), (vbool64_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (vuint32mf2_t)(op11), (vuint32mf2_t)(op12), (vuint32mf2_t)(op13), (vuint32mf2_t)(op14), (vuint32mf2_t)(op15), (vuint32mf2_t)(op16), (const uint32_t *)(op17), (size_t)(op18)) +#define vlseg2e64_v_i64m1(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e64_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (const int64_t *)(op2), (size_t)(op3)) +#define vlseg2e64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e64_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vbool64_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (const int64_t *)(op5), (size_t)(op6)) +#define vlseg2e64_v_i64m2(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e64_v_i64m2((vint64m2_t *)(op0), (vint64m2_t *)(op1), (const int64_t *)(op2), (size_t)(op3)) +#define vlseg2e64_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e64_v_i64m2_m((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vbool32_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (const int64_t *)(op5), (size_t)(op6)) +#define vlseg2e64_v_i64m4(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e64_v_i64m4((vint64m4_t *)(op0), (vint64m4_t *)(op1), (const int64_t *)(op2), (size_t)(op3)) +#define vlseg2e64_v_i64m4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e64_v_i64m4_m((vint64m4_t *)(op0), (vint64m4_t *)(op1), (vbool16_t)(op2), (vint64m4_t)(op3), (vint64m4_t)(op4), (const int64_t *)(op5), (size_t)(op6)) +#define vlseg2e64_v_u64m1(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e64_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (const uint64_t *)(op2), (size_t)(op3)) +#define vlseg2e64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e64_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vbool64_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (const uint64_t *)(op5), (size_t)(op6)) +#define vlseg2e64_v_u64m2(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e64_v_u64m2((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (const uint64_t *)(op2), (size_t)(op3)) +#define vlseg2e64_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e64_v_u64m2_m((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vbool32_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (const uint64_t *)(op5), (size_t)(op6)) +#define vlseg2e64_v_u64m4(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e64_v_u64m4((vuint64m4_t *)(op0), (vuint64m4_t *)(op1), (const uint64_t *)(op2), (size_t)(op3)) +#define vlseg2e64_v_u64m4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e64_v_u64m4_m((vuint64m4_t *)(op0), (vuint64m4_t *)(op1), (vbool16_t)(op2), (vuint64m4_t)(op3), (vuint64m4_t)(op4), (const uint64_t *)(op5), (size_t)(op6)) +#define vlseg3e64_v_i64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg3e64_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (const int64_t *)(op3), (size_t)(op4)) +#define vlseg3e64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg3e64_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vbool64_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (const int64_t *)(op7), (size_t)(op8)) +#define vlseg3e64_v_i64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg3e64_v_i64m2((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vint64m2_t *)(op2), (const int64_t *)(op3), (size_t)(op4)) +#define vlseg3e64_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg3e64_v_i64m2_m((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vint64m2_t *)(op2), (vbool32_t)(op3), (vint64m2_t)(op4), (vint64m2_t)(op5), (vint64m2_t)(op6), (const int64_t *)(op7), (size_t)(op8)) +#define vlseg3e64_v_u64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg3e64_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (const uint64_t *)(op3), (size_t)(op4)) +#define vlseg3e64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg3e64_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vbool64_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (const uint64_t *)(op7), (size_t)(op8)) +#define vlseg3e64_v_u64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg3e64_v_u64m2((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vuint64m2_t *)(op2), (const uint64_t *)(op3), (size_t)(op4)) +#define vlseg3e64_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg3e64_v_u64m2_m((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vuint64m2_t *)(op2), (vbool32_t)(op3), (vuint64m2_t)(op4), (vuint64m2_t)(op5), (vuint64m2_t)(op6), (const uint64_t *)(op7), (size_t)(op8)) +#define vlseg4e64_v_i64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg4e64_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (const int64_t *)(op4), (size_t)(op5)) +#define vlseg4e64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg4e64_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vbool64_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (const int64_t *)(op9), (size_t)(op10)) +#define vlseg4e64_v_i64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg4e64_v_i64m2((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vint64m2_t *)(op2), (vint64m2_t *)(op3), (const int64_t *)(op4), (size_t)(op5)) +#define vlseg4e64_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg4e64_v_i64m2_m((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vint64m2_t *)(op2), (vint64m2_t *)(op3), (vbool32_t)(op4), (vint64m2_t)(op5), (vint64m2_t)(op6), (vint64m2_t)(op7), (vint64m2_t)(op8), (const int64_t *)(op9), (size_t)(op10)) +#define vlseg4e64_v_u64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg4e64_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (const uint64_t *)(op4), (size_t)(op5)) +#define vlseg4e64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg4e64_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vbool64_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (const uint64_t *)(op9), (size_t)(op10)) +#define vlseg4e64_v_u64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg4e64_v_u64m2((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vuint64m2_t *)(op2), (vuint64m2_t *)(op3), (const uint64_t *)(op4), (size_t)(op5)) +#define vlseg4e64_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg4e64_v_u64m2_m((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vuint64m2_t *)(op2), (vuint64m2_t *)(op3), (vbool32_t)(op4), (vuint64m2_t)(op5), (vuint64m2_t)(op6), (vuint64m2_t)(op7), (vuint64m2_t)(op8), (const uint64_t *)(op9), (size_t)(op10)) +#define vlseg5e64_v_i64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg5e64_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (const int64_t *)(op5), (size_t)(op6)) +#define vlseg5e64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12) \ +__builtin_rvv_vlseg5e64_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vbool64_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (const int64_t *)(op11), (size_t)(op12)) +#define vlseg5e64_v_u64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg5e64_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (const uint64_t *)(op5), (size_t)(op6)) +#define vlseg5e64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12) \ +__builtin_rvv_vlseg5e64_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vbool64_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (const uint64_t *)(op11), (size_t)(op12)) +#define vlseg6e64_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg6e64_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (const int64_t *)(op6), (size_t)(op7)) +#define vlseg6e64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14) \ +__builtin_rvv_vlseg6e64_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vbool64_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (vint64m1_t)(op11), (vint64m1_t)(op12), (const int64_t *)(op13), (size_t)(op14)) +#define vlseg6e64_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg6e64_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (const uint64_t *)(op6), (size_t)(op7)) +#define vlseg6e64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14) \ +__builtin_rvv_vlseg6e64_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vbool64_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (vuint64m1_t)(op11), (vuint64m1_t)(op12), (const uint64_t *)(op13), (size_t)(op14)) +#define vlseg7e64_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg7e64_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vint64m1_t *)(op6), (const int64_t *)(op7), (size_t)(op8)) +#define vlseg7e64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16) \ +__builtin_rvv_vlseg7e64_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vint64m1_t *)(op6), (vbool64_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (vint64m1_t)(op11), (vint64m1_t)(op12), (vint64m1_t)(op13), (vint64m1_t)(op14), (const int64_t *)(op15), (size_t)(op16)) +#define vlseg7e64_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg7e64_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vuint64m1_t *)(op6), (const uint64_t *)(op7), (size_t)(op8)) +#define vlseg7e64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16) \ +__builtin_rvv_vlseg7e64_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vuint64m1_t *)(op6), (vbool64_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (vuint64m1_t)(op11), (vuint64m1_t)(op12), (vuint64m1_t)(op13), (vuint64m1_t)(op14), (const uint64_t *)(op15), (size_t)(op16)) +#define vlseg8e64_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg8e64_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vint64m1_t *)(op6), (vint64m1_t *)(op7), (const int64_t *)(op8), (size_t)(op9)) +#define vlseg8e64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18) \ +__builtin_rvv_vlseg8e64_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vint64m1_t *)(op6), (vint64m1_t *)(op7), (vbool64_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (vint64m1_t)(op11), (vint64m1_t)(op12), (vint64m1_t)(op13), (vint64m1_t)(op14), (vint64m1_t)(op15), (vint64m1_t)(op16), (const int64_t *)(op17), (size_t)(op18)) +#define vlseg8e64_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg8e64_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vuint64m1_t *)(op6), (vuint64m1_t *)(op7), (const uint64_t *)(op8), (size_t)(op9)) +#define vlseg8e64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18) \ +__builtin_rvv_vlseg8e64_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vuint64m1_t *)(op6), (vuint64m1_t *)(op7), (vbool64_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (vuint64m1_t)(op11), (vuint64m1_t)(op12), (vuint64m1_t)(op13), (vuint64m1_t)(op14), (vuint64m1_t)(op15), (vuint64m1_t)(op16), (const uint64_t *)(op17), (size_t)(op18)) +#define vlseg3e8ff_v_i8m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg3e8ff_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (const int8_t *)(op3), (size_t *)(op4), (size_t)(op5)) +#define vlseg3e8ff_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg3e8ff_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vbool8_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (const int8_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg3e8ff_v_i8m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg3e8ff_v_i8m2((vint8m2_t *)(op0), (vint8m2_t *)(op1), (vint8m2_t *)(op2), (const int8_t *)(op3), (size_t *)(op4), (size_t)(op5)) +#define vlseg3e8ff_v_i8m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg3e8ff_v_i8m2_m((vint8m2_t *)(op0), (vint8m2_t *)(op1), (vint8m2_t *)(op2), (vbool4_t)(op3), (vint8m2_t)(op4), (vint8m2_t)(op5), (vint8m2_t)(op6), (const int8_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg3e8ff_v_i8mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg3e8ff_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (const int8_t *)(op3), (size_t *)(op4), (size_t)(op5)) +#define vlseg3e8ff_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg3e8ff_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vbool16_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (const int8_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg3e8ff_v_i8mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg3e8ff_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (const int8_t *)(op3), (size_t *)(op4), (size_t)(op5)) +#define vlseg3e8ff_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg3e8ff_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vbool32_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (const int8_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg3e8ff_v_i8mf8(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg3e8ff_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (const int8_t *)(op3), (size_t *)(op4), (size_t)(op5)) +#define vlseg3e8ff_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg3e8ff_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vbool64_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (const int8_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg3e8ff_v_u8m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg3e8ff_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (const uint8_t *)(op3), (size_t *)(op4), (size_t)(op5)) +#define vlseg3e8ff_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg3e8ff_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vbool8_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (const uint8_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg3e8ff_v_u8m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg3e8ff_v_u8m2((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (vuint8m2_t *)(op2), (const uint8_t *)(op3), (size_t *)(op4), (size_t)(op5)) +#define vlseg3e8ff_v_u8m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg3e8ff_v_u8m2_m((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (vuint8m2_t *)(op2), (vbool4_t)(op3), (vuint8m2_t)(op4), (vuint8m2_t)(op5), (vuint8m2_t)(op6), (const uint8_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg3e8ff_v_u8mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg3e8ff_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (const uint8_t *)(op3), (size_t *)(op4), (size_t)(op5)) +#define vlseg3e8ff_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg3e8ff_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vbool16_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (const uint8_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg3e8ff_v_u8mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg3e8ff_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (const uint8_t *)(op3), (size_t *)(op4), (size_t)(op5)) +#define vlseg3e8ff_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg3e8ff_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vbool32_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (const uint8_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg3e8ff_v_u8mf8(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg3e8ff_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (const uint8_t *)(op3), (size_t *)(op4), (size_t)(op5)) +#define vlseg3e8ff_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg3e8ff_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vbool64_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (const uint8_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg4e8ff_v_i8m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg4e8ff_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (const int8_t *)(op4), (size_t *)(op5), (size_t)(op6)) +#define vlseg4e8ff_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlseg4e8ff_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vbool8_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (const int8_t *)(op9), (size_t *)(op10), (size_t)(op11)) +#define vlseg4e8ff_v_i8m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg4e8ff_v_i8m2((vint8m2_t *)(op0), (vint8m2_t *)(op1), (vint8m2_t *)(op2), (vint8m2_t *)(op3), (const int8_t *)(op4), (size_t *)(op5), (size_t)(op6)) +#define vlseg4e8ff_v_i8m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlseg4e8ff_v_i8m2_m((vint8m2_t *)(op0), (vint8m2_t *)(op1), (vint8m2_t *)(op2), (vint8m2_t *)(op3), (vbool4_t)(op4), (vint8m2_t)(op5), (vint8m2_t)(op6), (vint8m2_t)(op7), (vint8m2_t)(op8), (const int8_t *)(op9), (size_t *)(op10), (size_t)(op11)) +#define vlseg4e8ff_v_i8mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg4e8ff_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (const int8_t *)(op4), (size_t *)(op5), (size_t)(op6)) +#define vlseg4e8ff_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlseg4e8ff_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vbool16_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (const int8_t *)(op9), (size_t *)(op10), (size_t)(op11)) +#define vlseg4e8ff_v_i8mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg4e8ff_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (const int8_t *)(op4), (size_t *)(op5), (size_t)(op6)) +#define vlseg4e8ff_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlseg4e8ff_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vbool32_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (const int8_t *)(op9), (size_t *)(op10), (size_t)(op11)) +#define vlseg4e8ff_v_i8mf8(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg4e8ff_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (const int8_t *)(op4), (size_t *)(op5), (size_t)(op6)) +#define vlseg4e8ff_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlseg4e8ff_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vbool64_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (const int8_t *)(op9), (size_t *)(op10), (size_t)(op11)) +#define vlseg4e8ff_v_u8m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg4e8ff_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (const uint8_t *)(op4), (size_t *)(op5), (size_t)(op6)) +#define vlseg4e8ff_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlseg4e8ff_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vbool8_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (const uint8_t *)(op9), (size_t *)(op10), (size_t)(op11)) +#define vlseg4e8ff_v_u8m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg4e8ff_v_u8m2((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (vuint8m2_t *)(op2), (vuint8m2_t *)(op3), (const uint8_t *)(op4), (size_t *)(op5), (size_t)(op6)) +#define vlseg4e8ff_v_u8m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlseg4e8ff_v_u8m2_m((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (vuint8m2_t *)(op2), (vuint8m2_t *)(op3), (vbool4_t)(op4), (vuint8m2_t)(op5), (vuint8m2_t)(op6), (vuint8m2_t)(op7), (vuint8m2_t)(op8), (const uint8_t *)(op9), (size_t *)(op10), (size_t)(op11)) +#define vlseg4e8ff_v_u8mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg4e8ff_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (const uint8_t *)(op4), (size_t *)(op5), (size_t)(op6)) +#define vlseg4e8ff_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlseg4e8ff_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vbool16_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (const uint8_t *)(op9), (size_t *)(op10), (size_t)(op11)) +#define vlseg4e8ff_v_u8mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg4e8ff_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (const uint8_t *)(op4), (size_t *)(op5), (size_t)(op6)) +#define vlseg4e8ff_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlseg4e8ff_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vbool32_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (const uint8_t *)(op9), (size_t *)(op10), (size_t)(op11)) +#define vlseg4e8ff_v_u8mf8(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg4e8ff_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (const uint8_t *)(op4), (size_t *)(op5), (size_t)(op6)) +#define vlseg4e8ff_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlseg4e8ff_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vbool64_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (const uint8_t *)(op9), (size_t *)(op10), (size_t)(op11)) +#define vlseg5e8ff_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg5e8ff_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (const int8_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg5e8ff_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlseg5e8ff_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vbool8_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (const int8_t *)(op11), (size_t *)(op12), (size_t)(op13)) +#define vlseg5e8ff_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg5e8ff_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (const int8_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg5e8ff_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlseg5e8ff_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vbool16_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (const int8_t *)(op11), (size_t *)(op12), (size_t)(op13)) +#define vlseg5e8ff_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg5e8ff_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (const int8_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg5e8ff_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlseg5e8ff_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vbool32_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (const int8_t *)(op11), (size_t *)(op12), (size_t)(op13)) +#define vlseg5e8ff_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg5e8ff_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (const int8_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg5e8ff_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlseg5e8ff_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vbool64_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (const int8_t *)(op11), (size_t *)(op12), (size_t)(op13)) +#define vlseg5e8ff_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg5e8ff_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (const uint8_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg5e8ff_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlseg5e8ff_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vbool8_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (const uint8_t *)(op11), (size_t *)(op12), (size_t)(op13)) +#define vlseg5e8ff_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg5e8ff_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (const uint8_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg5e8ff_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlseg5e8ff_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vbool16_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (const uint8_t *)(op11), (size_t *)(op12), (size_t)(op13)) +#define vlseg5e8ff_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg5e8ff_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (const uint8_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg5e8ff_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlseg5e8ff_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vbool32_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (const uint8_t *)(op11), (size_t *)(op12), (size_t)(op13)) +#define vlseg5e8ff_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg5e8ff_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (const uint8_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg5e8ff_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlseg5e8ff_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vbool64_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (const uint8_t *)(op11), (size_t *)(op12), (size_t)(op13)) +#define vlseg6e8ff_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg6e8ff_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (const int8_t *)(op6), (size_t *)(op7), (size_t)(op8)) +#define vlseg6e8ff_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlseg6e8ff_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vbool8_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (vint8m1_t)(op11), (vint8m1_t)(op12), (const int8_t *)(op13), (size_t *)(op14), (size_t)(op15)) +#define vlseg6e8ff_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg6e8ff_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (const int8_t *)(op6), (size_t *)(op7), (size_t)(op8)) +#define vlseg6e8ff_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlseg6e8ff_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vbool16_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (vint8mf2_t)(op11), (vint8mf2_t)(op12), (const int8_t *)(op13), (size_t *)(op14), (size_t)(op15)) +#define vlseg6e8ff_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg6e8ff_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (const int8_t *)(op6), (size_t *)(op7), (size_t)(op8)) +#define vlseg6e8ff_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlseg6e8ff_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vbool32_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (vint8mf4_t)(op11), (vint8mf4_t)(op12), (const int8_t *)(op13), (size_t *)(op14), (size_t)(op15)) +#define vlseg6e8ff_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg6e8ff_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (const int8_t *)(op6), (size_t *)(op7), (size_t)(op8)) +#define vlseg6e8ff_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlseg6e8ff_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vbool64_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (vint8mf8_t)(op11), (vint8mf8_t)(op12), (const int8_t *)(op13), (size_t *)(op14), (size_t)(op15)) +#define vlseg6e8ff_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg6e8ff_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (const uint8_t *)(op6), (size_t *)(op7), (size_t)(op8)) +#define vlseg6e8ff_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlseg6e8ff_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vbool8_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (vuint8m1_t)(op11), (vuint8m1_t)(op12), (const uint8_t *)(op13), (size_t *)(op14), (size_t)(op15)) +#define vlseg6e8ff_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg6e8ff_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (const uint8_t *)(op6), (size_t *)(op7), (size_t)(op8)) +#define vlseg6e8ff_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlseg6e8ff_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vbool16_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (vuint8mf2_t)(op11), (vuint8mf2_t)(op12), (const uint8_t *)(op13), (size_t *)(op14), (size_t)(op15)) +#define vlseg6e8ff_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg6e8ff_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (const uint8_t *)(op6), (size_t *)(op7), (size_t)(op8)) +#define vlseg6e8ff_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlseg6e8ff_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vbool32_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (vuint8mf4_t)(op11), (vuint8mf4_t)(op12), (const uint8_t *)(op13), (size_t *)(op14), (size_t)(op15)) +#define vlseg6e8ff_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg6e8ff_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (const uint8_t *)(op6), (size_t *)(op7), (size_t)(op8)) +#define vlseg6e8ff_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlseg6e8ff_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vbool64_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (vuint8mf8_t)(op11), (vuint8mf8_t)(op12), (const uint8_t *)(op13), (size_t *)(op14), (size_t)(op15)) +#define vlseg7e8ff_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg7e8ff_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vint8m1_t *)(op6), (const int8_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg7e8ff_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlseg7e8ff_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vint8m1_t *)(op6), (vbool8_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (vint8m1_t)(op11), (vint8m1_t)(op12), (vint8m1_t)(op13), (vint8m1_t)(op14), (const int8_t *)(op15), (size_t *)(op16), (size_t)(op17)) +#define vlseg7e8ff_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg7e8ff_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vint8mf2_t *)(op6), (const int8_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg7e8ff_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlseg7e8ff_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vint8mf2_t *)(op6), (vbool16_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (vint8mf2_t)(op11), (vint8mf2_t)(op12), (vint8mf2_t)(op13), (vint8mf2_t)(op14), (const int8_t *)(op15), (size_t *)(op16), (size_t)(op17)) +#define vlseg7e8ff_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg7e8ff_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vint8mf4_t *)(op6), (const int8_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg7e8ff_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlseg7e8ff_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vint8mf4_t *)(op6), (vbool32_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (vint8mf4_t)(op11), (vint8mf4_t)(op12), (vint8mf4_t)(op13), (vint8mf4_t)(op14), (const int8_t *)(op15), (size_t *)(op16), (size_t)(op17)) +#define vlseg7e8ff_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg7e8ff_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vint8mf8_t *)(op6), (const int8_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg7e8ff_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlseg7e8ff_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vint8mf8_t *)(op6), (vbool64_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (vint8mf8_t)(op11), (vint8mf8_t)(op12), (vint8mf8_t)(op13), (vint8mf8_t)(op14), (const int8_t *)(op15), (size_t *)(op16), (size_t)(op17)) +#define vlseg7e8ff_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg7e8ff_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vuint8m1_t *)(op6), (const uint8_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg7e8ff_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlseg7e8ff_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vuint8m1_t *)(op6), (vbool8_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (vuint8m1_t)(op11), (vuint8m1_t)(op12), (vuint8m1_t)(op13), (vuint8m1_t)(op14), (const uint8_t *)(op15), (size_t *)(op16), (size_t)(op17)) +#define vlseg7e8ff_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg7e8ff_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vuint8mf2_t *)(op6), (const uint8_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg7e8ff_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlseg7e8ff_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vuint8mf2_t *)(op6), (vbool16_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (vuint8mf2_t)(op11), (vuint8mf2_t)(op12), (vuint8mf2_t)(op13), (vuint8mf2_t)(op14), (const uint8_t *)(op15), (size_t *)(op16), (size_t)(op17)) +#define vlseg7e8ff_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg7e8ff_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vuint8mf4_t *)(op6), (const uint8_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg7e8ff_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlseg7e8ff_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vuint8mf4_t *)(op6), (vbool32_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (vuint8mf4_t)(op11), (vuint8mf4_t)(op12), (vuint8mf4_t)(op13), (vuint8mf4_t)(op14), (const uint8_t *)(op15), (size_t *)(op16), (size_t)(op17)) +#define vlseg7e8ff_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg7e8ff_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vuint8mf8_t *)(op6), (const uint8_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg7e8ff_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlseg7e8ff_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vuint8mf8_t *)(op6), (vbool64_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (vuint8mf8_t)(op11), (vuint8mf8_t)(op12), (vuint8mf8_t)(op13), (vuint8mf8_t)(op14), (const uint8_t *)(op15), (size_t *)(op16), (size_t)(op17)) +#define vlseg8e8ff_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg8e8ff_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vint8m1_t *)(op6), (vint8m1_t *)(op7), (const int8_t *)(op8), (size_t *)(op9), (size_t)(op10)) +#define vlseg8e8ff_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlseg8e8ff_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vint8m1_t *)(op6), (vint8m1_t *)(op7), (vbool8_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (vint8m1_t)(op11), (vint8m1_t)(op12), (vint8m1_t)(op13), (vint8m1_t)(op14), (vint8m1_t)(op15), (vint8m1_t)(op16), (const int8_t *)(op17), (size_t *)(op18), (size_t)(op19)) +#define vlseg8e8ff_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg8e8ff_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vint8mf2_t *)(op6), (vint8mf2_t *)(op7), (const int8_t *)(op8), (size_t *)(op9), (size_t)(op10)) +#define vlseg8e8ff_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlseg8e8ff_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vint8mf2_t *)(op6), (vint8mf2_t *)(op7), (vbool16_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (vint8mf2_t)(op11), (vint8mf2_t)(op12), (vint8mf2_t)(op13), (vint8mf2_t)(op14), (vint8mf2_t)(op15), (vint8mf2_t)(op16), (const int8_t *)(op17), (size_t *)(op18), (size_t)(op19)) +#define vlseg8e8ff_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg8e8ff_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vint8mf4_t *)(op6), (vint8mf4_t *)(op7), (const int8_t *)(op8), (size_t *)(op9), (size_t)(op10)) +#define vlseg8e8ff_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlseg8e8ff_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vint8mf4_t *)(op6), (vint8mf4_t *)(op7), (vbool32_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (vint8mf4_t)(op11), (vint8mf4_t)(op12), (vint8mf4_t)(op13), (vint8mf4_t)(op14), (vint8mf4_t)(op15), (vint8mf4_t)(op16), (const int8_t *)(op17), (size_t *)(op18), (size_t)(op19)) +#define vlseg8e8ff_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg8e8ff_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vint8mf8_t *)(op6), (vint8mf8_t *)(op7), (const int8_t *)(op8), (size_t *)(op9), (size_t)(op10)) +#define vlseg8e8ff_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlseg8e8ff_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vint8mf8_t *)(op6), (vint8mf8_t *)(op7), (vbool64_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (vint8mf8_t)(op11), (vint8mf8_t)(op12), (vint8mf8_t)(op13), (vint8mf8_t)(op14), (vint8mf8_t)(op15), (vint8mf8_t)(op16), (const int8_t *)(op17), (size_t *)(op18), (size_t)(op19)) +#define vlseg8e8ff_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg8e8ff_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vuint8m1_t *)(op6), (vuint8m1_t *)(op7), (const uint8_t *)(op8), (size_t *)(op9), (size_t)(op10)) +#define vlseg8e8ff_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlseg8e8ff_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vuint8m1_t *)(op6), (vuint8m1_t *)(op7), (vbool8_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (vuint8m1_t)(op11), (vuint8m1_t)(op12), (vuint8m1_t)(op13), (vuint8m1_t)(op14), (vuint8m1_t)(op15), (vuint8m1_t)(op16), (const uint8_t *)(op17), (size_t *)(op18), (size_t)(op19)) +#define vlseg8e8ff_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg8e8ff_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vuint8mf2_t *)(op6), (vuint8mf2_t *)(op7), (const uint8_t *)(op8), (size_t *)(op9), (size_t)(op10)) +#define vlseg8e8ff_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlseg8e8ff_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vuint8mf2_t *)(op6), (vuint8mf2_t *)(op7), (vbool16_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (vuint8mf2_t)(op11), (vuint8mf2_t)(op12), (vuint8mf2_t)(op13), (vuint8mf2_t)(op14), (vuint8mf2_t)(op15), (vuint8mf2_t)(op16), (const uint8_t *)(op17), (size_t *)(op18), (size_t)(op19)) +#define vlseg8e8ff_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg8e8ff_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vuint8mf4_t *)(op6), (vuint8mf4_t *)(op7), (const uint8_t *)(op8), (size_t *)(op9), (size_t)(op10)) +#define vlseg8e8ff_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlseg8e8ff_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vuint8mf4_t *)(op6), (vuint8mf4_t *)(op7), (vbool32_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (vuint8mf4_t)(op11), (vuint8mf4_t)(op12), (vuint8mf4_t)(op13), (vuint8mf4_t)(op14), (vuint8mf4_t)(op15), (vuint8mf4_t)(op16), (const uint8_t *)(op17), (size_t *)(op18), (size_t)(op19)) +#define vlseg8e8ff_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg8e8ff_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vuint8mf8_t *)(op6), (vuint8mf8_t *)(op7), (const uint8_t *)(op8), (size_t *)(op9), (size_t)(op10)) +#define vlseg8e8ff_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlseg8e8ff_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vuint8mf8_t *)(op6), (vuint8mf8_t *)(op7), (vbool64_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (vuint8mf8_t)(op11), (vuint8mf8_t)(op12), (vuint8mf8_t)(op13), (vuint8mf8_t)(op14), (vuint8mf8_t)(op15), (vuint8mf8_t)(op16), (const uint8_t *)(op17), (size_t *)(op18), (size_t)(op19)) +#define vlseg2e16ff_v_i16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e16ff_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (const int16_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e16ff_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e16ff_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vbool16_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (const int16_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg2e16ff_v_i16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e16ff_v_i16m2((vint16m2_t *)(op0), (vint16m2_t *)(op1), (const int16_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e16ff_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e16ff_v_i16m2_m((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vbool8_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (const int16_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg2e16ff_v_i16m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e16ff_v_i16m4((vint16m4_t *)(op0), (vint16m4_t *)(op1), (const int16_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e16ff_v_i16m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e16ff_v_i16m4_m((vint16m4_t *)(op0), (vint16m4_t *)(op1), (vbool4_t)(op2), (vint16m4_t)(op3), (vint16m4_t)(op4), (const int16_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg2e16ff_v_i16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e16ff_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (const int16_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e16ff_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e16ff_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vbool32_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (const int16_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg2e16ff_v_i16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e16ff_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (const int16_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e16ff_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e16ff_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vbool64_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (const int16_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg2e16ff_v_u16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e16ff_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (const uint16_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e16ff_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e16ff_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vbool16_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (const uint16_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg2e16ff_v_u16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e16ff_v_u16m2((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (const uint16_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e16ff_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e16ff_v_u16m2_m((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vbool8_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (const uint16_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg2e16ff_v_u16m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e16ff_v_u16m4((vuint16m4_t *)(op0), (vuint16m4_t *)(op1), (const uint16_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e16ff_v_u16m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e16ff_v_u16m4_m((vuint16m4_t *)(op0), (vuint16m4_t *)(op1), (vbool4_t)(op2), (vuint16m4_t)(op3), (vuint16m4_t)(op4), (const uint16_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg2e16ff_v_u16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e16ff_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (const uint16_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e16ff_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e16ff_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vbool32_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (const uint16_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg2e16ff_v_u16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e16ff_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (const uint16_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e16ff_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e16ff_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vbool64_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (const uint16_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg3e16ff_v_i16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg3e16ff_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (const int16_t *)(op3), (size_t *)(op4), (size_t)(op5)) +#define vlseg3e16ff_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg3e16ff_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vbool16_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (const int16_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg3e16ff_v_i16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg3e16ff_v_i16m2((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vint16m2_t *)(op2), (const int16_t *)(op3), (size_t *)(op4), (size_t)(op5)) +#define vlseg3e16ff_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg3e16ff_v_i16m2_m((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vint16m2_t *)(op2), (vbool8_t)(op3), (vint16m2_t)(op4), (vint16m2_t)(op5), (vint16m2_t)(op6), (const int16_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg3e16ff_v_i16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg3e16ff_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (const int16_t *)(op3), (size_t *)(op4), (size_t)(op5)) +#define vlseg3e16ff_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg3e16ff_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vbool32_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (const int16_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg3e16ff_v_i16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg3e16ff_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (const int16_t *)(op3), (size_t *)(op4), (size_t)(op5)) +#define vlseg3e16ff_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg3e16ff_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vbool64_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (const int16_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg3e16ff_v_u16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg3e16ff_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (const uint16_t *)(op3), (size_t *)(op4), (size_t)(op5)) +#define vlseg3e16ff_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg3e16ff_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vbool16_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (const uint16_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg3e16ff_v_u16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg3e16ff_v_u16m2((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vuint16m2_t *)(op2), (const uint16_t *)(op3), (size_t *)(op4), (size_t)(op5)) +#define vlseg3e16ff_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg3e16ff_v_u16m2_m((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vuint16m2_t *)(op2), (vbool8_t)(op3), (vuint16m2_t)(op4), (vuint16m2_t)(op5), (vuint16m2_t)(op6), (const uint16_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg3e16ff_v_u16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg3e16ff_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (const uint16_t *)(op3), (size_t *)(op4), (size_t)(op5)) +#define vlseg3e16ff_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg3e16ff_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vbool32_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (const uint16_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg3e16ff_v_u16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg3e16ff_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (const uint16_t *)(op3), (size_t *)(op4), (size_t)(op5)) +#define vlseg3e16ff_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg3e16ff_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vbool64_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (const uint16_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg4e16ff_v_i16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg4e16ff_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (const int16_t *)(op4), (size_t *)(op5), (size_t)(op6)) +#define vlseg4e16ff_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlseg4e16ff_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vbool16_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (const int16_t *)(op9), (size_t *)(op10), (size_t)(op11)) +#define vlseg4e16ff_v_i16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg4e16ff_v_i16m2((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vint16m2_t *)(op2), (vint16m2_t *)(op3), (const int16_t *)(op4), (size_t *)(op5), (size_t)(op6)) +#define vlseg4e16ff_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlseg4e16ff_v_i16m2_m((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vint16m2_t *)(op2), (vint16m2_t *)(op3), (vbool8_t)(op4), (vint16m2_t)(op5), (vint16m2_t)(op6), (vint16m2_t)(op7), (vint16m2_t)(op8), (const int16_t *)(op9), (size_t *)(op10), (size_t)(op11)) +#define vlseg4e16ff_v_i16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg4e16ff_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (const int16_t *)(op4), (size_t *)(op5), (size_t)(op6)) +#define vlseg4e16ff_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlseg4e16ff_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vbool32_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (const int16_t *)(op9), (size_t *)(op10), (size_t)(op11)) +#define vlseg4e16ff_v_i16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg4e16ff_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (const int16_t *)(op4), (size_t *)(op5), (size_t)(op6)) +#define vlseg4e16ff_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlseg4e16ff_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vbool64_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (const int16_t *)(op9), (size_t *)(op10), (size_t)(op11)) +#define vlseg4e16ff_v_u16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg4e16ff_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (const uint16_t *)(op4), (size_t *)(op5), (size_t)(op6)) +#define vlseg4e16ff_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlseg4e16ff_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vbool16_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (const uint16_t *)(op9), (size_t *)(op10), (size_t)(op11)) +#define vlseg4e16ff_v_u16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg4e16ff_v_u16m2((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vuint16m2_t *)(op2), (vuint16m2_t *)(op3), (const uint16_t *)(op4), (size_t *)(op5), (size_t)(op6)) +#define vlseg4e16ff_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlseg4e16ff_v_u16m2_m((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vuint16m2_t *)(op2), (vuint16m2_t *)(op3), (vbool8_t)(op4), (vuint16m2_t)(op5), (vuint16m2_t)(op6), (vuint16m2_t)(op7), (vuint16m2_t)(op8), (const uint16_t *)(op9), (size_t *)(op10), (size_t)(op11)) +#define vlseg4e16ff_v_u16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg4e16ff_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (const uint16_t *)(op4), (size_t *)(op5), (size_t)(op6)) +#define vlseg4e16ff_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlseg4e16ff_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vbool32_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (const uint16_t *)(op9), (size_t *)(op10), (size_t)(op11)) +#define vlseg4e16ff_v_u16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg4e16ff_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (const uint16_t *)(op4), (size_t *)(op5), (size_t)(op6)) +#define vlseg4e16ff_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlseg4e16ff_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vbool64_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (const uint16_t *)(op9), (size_t *)(op10), (size_t)(op11)) +#define vlseg5e16ff_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg5e16ff_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (const int16_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg5e16ff_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlseg5e16ff_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vbool16_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (const int16_t *)(op11), (size_t *)(op12), (size_t)(op13)) +#define vlseg5e16ff_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg5e16ff_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (const int16_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg5e16ff_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlseg5e16ff_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vbool32_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (const int16_t *)(op11), (size_t *)(op12), (size_t)(op13)) +#define vlseg5e16ff_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg5e16ff_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (const int16_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg5e16ff_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlseg5e16ff_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vbool64_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (const int16_t *)(op11), (size_t *)(op12), (size_t)(op13)) +#define vlseg5e16ff_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg5e16ff_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (const uint16_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg5e16ff_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlseg5e16ff_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vbool16_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (const uint16_t *)(op11), (size_t *)(op12), (size_t)(op13)) +#define vlseg5e16ff_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg5e16ff_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (const uint16_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg5e16ff_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlseg5e16ff_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vbool32_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (const uint16_t *)(op11), (size_t *)(op12), (size_t)(op13)) +#define vlseg5e16ff_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg5e16ff_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (const uint16_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg5e16ff_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlseg5e16ff_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vbool64_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (const uint16_t *)(op11), (size_t *)(op12), (size_t)(op13)) +#define vlseg6e16ff_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg6e16ff_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (const int16_t *)(op6), (size_t *)(op7), (size_t)(op8)) +#define vlseg6e16ff_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlseg6e16ff_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vbool16_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (vint16m1_t)(op11), (vint16m1_t)(op12), (const int16_t *)(op13), (size_t *)(op14), (size_t)(op15)) +#define vlseg6e16ff_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg6e16ff_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (const int16_t *)(op6), (size_t *)(op7), (size_t)(op8)) +#define vlseg6e16ff_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlseg6e16ff_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vbool32_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (vint16mf2_t)(op11), (vint16mf2_t)(op12), (const int16_t *)(op13), (size_t *)(op14), (size_t)(op15)) +#define vlseg6e16ff_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg6e16ff_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (const int16_t *)(op6), (size_t *)(op7), (size_t)(op8)) +#define vlseg6e16ff_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlseg6e16ff_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vbool64_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (vint16mf4_t)(op11), (vint16mf4_t)(op12), (const int16_t *)(op13), (size_t *)(op14), (size_t)(op15)) +#define vlseg6e16ff_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg6e16ff_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (const uint16_t *)(op6), (size_t *)(op7), (size_t)(op8)) +#define vlseg6e16ff_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlseg6e16ff_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vbool16_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (vuint16m1_t)(op11), (vuint16m1_t)(op12), (const uint16_t *)(op13), (size_t *)(op14), (size_t)(op15)) +#define vlseg6e16ff_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg6e16ff_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (const uint16_t *)(op6), (size_t *)(op7), (size_t)(op8)) +#define vlseg6e16ff_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlseg6e16ff_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vbool32_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (vuint16mf2_t)(op11), (vuint16mf2_t)(op12), (const uint16_t *)(op13), (size_t *)(op14), (size_t)(op15)) +#define vlseg6e16ff_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg6e16ff_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (const uint16_t *)(op6), (size_t *)(op7), (size_t)(op8)) +#define vlseg6e16ff_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlseg6e16ff_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vbool64_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (vuint16mf4_t)(op11), (vuint16mf4_t)(op12), (const uint16_t *)(op13), (size_t *)(op14), (size_t)(op15)) +#define vlseg7e16ff_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg7e16ff_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vint16m1_t *)(op6), (const int16_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg7e16ff_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlseg7e16ff_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vint16m1_t *)(op6), (vbool16_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (vint16m1_t)(op11), (vint16m1_t)(op12), (vint16m1_t)(op13), (vint16m1_t)(op14), (const int16_t *)(op15), (size_t *)(op16), (size_t)(op17)) +#define vlseg7e16ff_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg7e16ff_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vint16mf2_t *)(op6), (const int16_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg7e16ff_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlseg7e16ff_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vint16mf2_t *)(op6), (vbool32_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (vint16mf2_t)(op11), (vint16mf2_t)(op12), (vint16mf2_t)(op13), (vint16mf2_t)(op14), (const int16_t *)(op15), (size_t *)(op16), (size_t)(op17)) +#define vlseg7e16ff_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg7e16ff_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vint16mf4_t *)(op6), (const int16_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg7e16ff_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlseg7e16ff_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vint16mf4_t *)(op6), (vbool64_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (vint16mf4_t)(op11), (vint16mf4_t)(op12), (vint16mf4_t)(op13), (vint16mf4_t)(op14), (const int16_t *)(op15), (size_t *)(op16), (size_t)(op17)) +#define vlseg7e16ff_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg7e16ff_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vuint16m1_t *)(op6), (const uint16_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg7e16ff_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlseg7e16ff_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vuint16m1_t *)(op6), (vbool16_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (vuint16m1_t)(op11), (vuint16m1_t)(op12), (vuint16m1_t)(op13), (vuint16m1_t)(op14), (const uint16_t *)(op15), (size_t *)(op16), (size_t)(op17)) +#define vlseg7e16ff_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg7e16ff_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vuint16mf2_t *)(op6), (const uint16_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg7e16ff_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlseg7e16ff_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vuint16mf2_t *)(op6), (vbool32_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (vuint16mf2_t)(op11), (vuint16mf2_t)(op12), (vuint16mf2_t)(op13), (vuint16mf2_t)(op14), (const uint16_t *)(op15), (size_t *)(op16), (size_t)(op17)) +#define vlseg7e16ff_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg7e16ff_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vuint16mf4_t *)(op6), (const uint16_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg7e16ff_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlseg7e16ff_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vuint16mf4_t *)(op6), (vbool64_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (vuint16mf4_t)(op11), (vuint16mf4_t)(op12), (vuint16mf4_t)(op13), (vuint16mf4_t)(op14), (const uint16_t *)(op15), (size_t *)(op16), (size_t)(op17)) +#define vlseg8e16ff_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg8e16ff_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vint16m1_t *)(op6), (vint16m1_t *)(op7), (const int16_t *)(op8), (size_t *)(op9), (size_t)(op10)) +#define vlseg8e16ff_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlseg8e16ff_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vint16m1_t *)(op6), (vint16m1_t *)(op7), (vbool16_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (vint16m1_t)(op11), (vint16m1_t)(op12), (vint16m1_t)(op13), (vint16m1_t)(op14), (vint16m1_t)(op15), (vint16m1_t)(op16), (const int16_t *)(op17), (size_t *)(op18), (size_t)(op19)) +#define vlseg8e16ff_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg8e16ff_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vint16mf2_t *)(op6), (vint16mf2_t *)(op7), (const int16_t *)(op8), (size_t *)(op9), (size_t)(op10)) +#define vlseg8e16ff_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlseg8e16ff_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vint16mf2_t *)(op6), (vint16mf2_t *)(op7), (vbool32_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (vint16mf2_t)(op11), (vint16mf2_t)(op12), (vint16mf2_t)(op13), (vint16mf2_t)(op14), (vint16mf2_t)(op15), (vint16mf2_t)(op16), (const int16_t *)(op17), (size_t *)(op18), (size_t)(op19)) +#define vlseg8e16ff_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg8e16ff_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vint16mf4_t *)(op6), (vint16mf4_t *)(op7), (const int16_t *)(op8), (size_t *)(op9), (size_t)(op10)) +#define vlseg8e16ff_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlseg8e16ff_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vint16mf4_t *)(op6), (vint16mf4_t *)(op7), (vbool64_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (vint16mf4_t)(op11), (vint16mf4_t)(op12), (vint16mf4_t)(op13), (vint16mf4_t)(op14), (vint16mf4_t)(op15), (vint16mf4_t)(op16), (const int16_t *)(op17), (size_t *)(op18), (size_t)(op19)) +#define vlseg8e16ff_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg8e16ff_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vuint16m1_t *)(op6), (vuint16m1_t *)(op7), (const uint16_t *)(op8), (size_t *)(op9), (size_t)(op10)) +#define vlseg8e16ff_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlseg8e16ff_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vuint16m1_t *)(op6), (vuint16m1_t *)(op7), (vbool16_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (vuint16m1_t)(op11), (vuint16m1_t)(op12), (vuint16m1_t)(op13), (vuint16m1_t)(op14), (vuint16m1_t)(op15), (vuint16m1_t)(op16), (const uint16_t *)(op17), (size_t *)(op18), (size_t)(op19)) +#define vlseg8e16ff_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg8e16ff_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vuint16mf2_t *)(op6), (vuint16mf2_t *)(op7), (const uint16_t *)(op8), (size_t *)(op9), (size_t)(op10)) +#define vlseg8e16ff_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlseg8e16ff_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vuint16mf2_t *)(op6), (vuint16mf2_t *)(op7), (vbool32_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (vuint16mf2_t)(op11), (vuint16mf2_t)(op12), (vuint16mf2_t)(op13), (vuint16mf2_t)(op14), (vuint16mf2_t)(op15), (vuint16mf2_t)(op16), (const uint16_t *)(op17), (size_t *)(op18), (size_t)(op19)) +#define vlseg8e16ff_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg8e16ff_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vuint16mf4_t *)(op6), (vuint16mf4_t *)(op7), (const uint16_t *)(op8), (size_t *)(op9), (size_t)(op10)) +#define vlseg8e16ff_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlseg8e16ff_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vuint16mf4_t *)(op6), (vuint16mf4_t *)(op7), (vbool64_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (vuint16mf4_t)(op11), (vuint16mf4_t)(op12), (vuint16mf4_t)(op13), (vuint16mf4_t)(op14), (vuint16mf4_t)(op15), (vuint16mf4_t)(op16), (const uint16_t *)(op17), (size_t *)(op18), (size_t)(op19)) +#define vlseg2e32ff_v_i32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e32ff_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (const int32_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e32ff_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e32ff_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vbool32_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (const int32_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg2e32ff_v_i32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e32ff_v_i32m2((vint32m2_t *)(op0), (vint32m2_t *)(op1), (const int32_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e32ff_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e32ff_v_i32m2_m((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vbool16_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (const int32_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg2e32ff_v_i32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e32ff_v_i32m4((vint32m4_t *)(op0), (vint32m4_t *)(op1), (const int32_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e32ff_v_i32m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e32ff_v_i32m4_m((vint32m4_t *)(op0), (vint32m4_t *)(op1), (vbool8_t)(op2), (vint32m4_t)(op3), (vint32m4_t)(op4), (const int32_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg2e32ff_v_i32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e32ff_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (const int32_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e32ff_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e32ff_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vbool64_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (const int32_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg2e8ff_v_i8m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e8ff_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (const int8_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e8ff_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e8ff_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vbool8_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (const int8_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg2e8ff_v_i8m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e8ff_v_i8m2((vint8m2_t *)(op0), (vint8m2_t *)(op1), (const int8_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e8ff_v_i8m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e8ff_v_i8m2_m((vint8m2_t *)(op0), (vint8m2_t *)(op1), (vbool4_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (const int8_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg2e8ff_v_i8m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e8ff_v_i8m4((vint8m4_t *)(op0), (vint8m4_t *)(op1), (const int8_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e8ff_v_i8m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e8ff_v_i8m4_m((vint8m4_t *)(op0), (vint8m4_t *)(op1), (vbool2_t)(op2), (vint8m4_t)(op3), (vint8m4_t)(op4), (const int8_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg2e8ff_v_i8mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e8ff_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (const int8_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e8ff_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e8ff_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vbool16_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (const int8_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg2e8ff_v_i8mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e8ff_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (const int8_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e8ff_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e8ff_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vbool32_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (const int8_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg2e8ff_v_i8mf8(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e8ff_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (const int8_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e8ff_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e8ff_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vbool64_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (const int8_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg2e32ff_v_u32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e32ff_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (const uint32_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e32ff_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e32ff_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vbool32_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (const uint32_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg2e32ff_v_u32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e32ff_v_u32m2((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (const uint32_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e32ff_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e32ff_v_u32m2_m((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vbool16_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (const uint32_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg2e32ff_v_u32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e32ff_v_u32m4((vuint32m4_t *)(op0), (vuint32m4_t *)(op1), (const uint32_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e32ff_v_u32m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e32ff_v_u32m4_m((vuint32m4_t *)(op0), (vuint32m4_t *)(op1), (vbool8_t)(op2), (vuint32m4_t)(op3), (vuint32m4_t)(op4), (const uint32_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg2e32ff_v_u32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e32ff_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (const uint32_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e32ff_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e32ff_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vbool64_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (const uint32_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg3e32ff_v_i32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg3e32ff_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (const int32_t *)(op3), (size_t *)(op4), (size_t)(op5)) +#define vlseg3e32ff_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg3e32ff_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vbool32_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (const int32_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg3e32ff_v_i32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg3e32ff_v_i32m2((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vint32m2_t *)(op2), (const int32_t *)(op3), (size_t *)(op4), (size_t)(op5)) +#define vlseg3e32ff_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg3e32ff_v_i32m2_m((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vint32m2_t *)(op2), (vbool16_t)(op3), (vint32m2_t)(op4), (vint32m2_t)(op5), (vint32m2_t)(op6), (const int32_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg3e32ff_v_i32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg3e32ff_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (const int32_t *)(op3), (size_t *)(op4), (size_t)(op5)) +#define vlseg3e32ff_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg3e32ff_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vbool64_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (const int32_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg3e32ff_v_u32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg3e32ff_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (const uint32_t *)(op3), (size_t *)(op4), (size_t)(op5)) +#define vlseg3e32ff_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg3e32ff_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vbool32_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (const uint32_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg3e32ff_v_u32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg3e32ff_v_u32m2((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vuint32m2_t *)(op2), (const uint32_t *)(op3), (size_t *)(op4), (size_t)(op5)) +#define vlseg3e32ff_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg3e32ff_v_u32m2_m((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vuint32m2_t *)(op2), (vbool16_t)(op3), (vuint32m2_t)(op4), (vuint32m2_t)(op5), (vuint32m2_t)(op6), (const uint32_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg3e32ff_v_u32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg3e32ff_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (const uint32_t *)(op3), (size_t *)(op4), (size_t)(op5)) +#define vlseg3e32ff_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg3e32ff_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vbool64_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (const uint32_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg4e32ff_v_i32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg4e32ff_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (const int32_t *)(op4), (size_t *)(op5), (size_t)(op6)) +#define vlseg4e32ff_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlseg4e32ff_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vbool32_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (const int32_t *)(op9), (size_t *)(op10), (size_t)(op11)) +#define vlseg4e32ff_v_i32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg4e32ff_v_i32m2((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vint32m2_t *)(op2), (vint32m2_t *)(op3), (const int32_t *)(op4), (size_t *)(op5), (size_t)(op6)) +#define vlseg4e32ff_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlseg4e32ff_v_i32m2_m((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vint32m2_t *)(op2), (vint32m2_t *)(op3), (vbool16_t)(op4), (vint32m2_t)(op5), (vint32m2_t)(op6), (vint32m2_t)(op7), (vint32m2_t)(op8), (const int32_t *)(op9), (size_t *)(op10), (size_t)(op11)) +#define vlseg4e32ff_v_i32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg4e32ff_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (const int32_t *)(op4), (size_t *)(op5), (size_t)(op6)) +#define vlseg4e32ff_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlseg4e32ff_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vbool64_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (const int32_t *)(op9), (size_t *)(op10), (size_t)(op11)) +#define vlseg4e32ff_v_u32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg4e32ff_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (const uint32_t *)(op4), (size_t *)(op5), (size_t)(op6)) +#define vlseg4e32ff_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlseg4e32ff_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vbool32_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (const uint32_t *)(op9), (size_t *)(op10), (size_t)(op11)) +#define vlseg4e32ff_v_u32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg4e32ff_v_u32m2((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vuint32m2_t *)(op2), (vuint32m2_t *)(op3), (const uint32_t *)(op4), (size_t *)(op5), (size_t)(op6)) +#define vlseg4e32ff_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlseg4e32ff_v_u32m2_m((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vuint32m2_t *)(op2), (vuint32m2_t *)(op3), (vbool16_t)(op4), (vuint32m2_t)(op5), (vuint32m2_t)(op6), (vuint32m2_t)(op7), (vuint32m2_t)(op8), (const uint32_t *)(op9), (size_t *)(op10), (size_t)(op11)) +#define vlseg4e32ff_v_u32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg4e32ff_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (const uint32_t *)(op4), (size_t *)(op5), (size_t)(op6)) +#define vlseg4e32ff_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlseg4e32ff_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vbool64_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (const uint32_t *)(op9), (size_t *)(op10), (size_t)(op11)) +#define vlseg5e32ff_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg5e32ff_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (const int32_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg5e32ff_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlseg5e32ff_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vbool32_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (const int32_t *)(op11), (size_t *)(op12), (size_t)(op13)) +#define vlseg5e32ff_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg5e32ff_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (const int32_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg5e32ff_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlseg5e32ff_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vbool64_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (const int32_t *)(op11), (size_t *)(op12), (size_t)(op13)) +#define vlseg5e32ff_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg5e32ff_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (const uint32_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg5e32ff_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlseg5e32ff_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vbool32_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (const uint32_t *)(op11), (size_t *)(op12), (size_t)(op13)) +#define vlseg5e32ff_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg5e32ff_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (const uint32_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg5e32ff_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlseg5e32ff_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vbool64_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (const uint32_t *)(op11), (size_t *)(op12), (size_t)(op13)) +#define vlseg6e32ff_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg6e32ff_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (const int32_t *)(op6), (size_t *)(op7), (size_t)(op8)) +#define vlseg6e32ff_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlseg6e32ff_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vbool32_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (vint32m1_t)(op11), (vint32m1_t)(op12), (const int32_t *)(op13), (size_t *)(op14), (size_t)(op15)) +#define vlseg6e32ff_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg6e32ff_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (const int32_t *)(op6), (size_t *)(op7), (size_t)(op8)) +#define vlseg6e32ff_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlseg6e32ff_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vbool64_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (vint32mf2_t)(op11), (vint32mf2_t)(op12), (const int32_t *)(op13), (size_t *)(op14), (size_t)(op15)) +#define vlseg6e32ff_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg6e32ff_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (const uint32_t *)(op6), (size_t *)(op7), (size_t)(op8)) +#define vlseg6e32ff_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlseg6e32ff_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vbool32_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (vuint32m1_t)(op11), (vuint32m1_t)(op12), (const uint32_t *)(op13), (size_t *)(op14), (size_t)(op15)) +#define vlseg6e32ff_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg6e32ff_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (const uint32_t *)(op6), (size_t *)(op7), (size_t)(op8)) +#define vlseg6e32ff_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlseg6e32ff_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vbool64_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (vuint32mf2_t)(op11), (vuint32mf2_t)(op12), (const uint32_t *)(op13), (size_t *)(op14), (size_t)(op15)) +#define vlseg7e32ff_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg7e32ff_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vint32m1_t *)(op6), (const int32_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg7e32ff_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlseg7e32ff_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vint32m1_t *)(op6), (vbool32_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (vint32m1_t)(op11), (vint32m1_t)(op12), (vint32m1_t)(op13), (vint32m1_t)(op14), (const int32_t *)(op15), (size_t *)(op16), (size_t)(op17)) +#define vlseg7e32ff_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg7e32ff_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vint32mf2_t *)(op6), (const int32_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg7e32ff_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlseg7e32ff_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vint32mf2_t *)(op6), (vbool64_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (vint32mf2_t)(op11), (vint32mf2_t)(op12), (vint32mf2_t)(op13), (vint32mf2_t)(op14), (const int32_t *)(op15), (size_t *)(op16), (size_t)(op17)) +#define vlseg2e8ff_v_u8m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e8ff_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (const uint8_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e8ff_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e8ff_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vbool8_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (const uint8_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg2e8ff_v_u8m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e8ff_v_u8m2((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (const uint8_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e8ff_v_u8m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e8ff_v_u8m2_m((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (vbool4_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (const uint8_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg2e8ff_v_u8m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e8ff_v_u8m4((vuint8m4_t *)(op0), (vuint8m4_t *)(op1), (const uint8_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e8ff_v_u8m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e8ff_v_u8m4_m((vuint8m4_t *)(op0), (vuint8m4_t *)(op1), (vbool2_t)(op2), (vuint8m4_t)(op3), (vuint8m4_t)(op4), (const uint8_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg2e8ff_v_u8mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e8ff_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (const uint8_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e8ff_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e8ff_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vbool16_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (const uint8_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg2e8ff_v_u8mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e8ff_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (const uint8_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e8ff_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e8ff_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vbool32_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (const uint8_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg2e8ff_v_u8mf8(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e8ff_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (const uint8_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e8ff_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e8ff_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vbool64_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (const uint8_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg7e32ff_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg7e32ff_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vuint32m1_t *)(op6), (const uint32_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg7e32ff_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlseg7e32ff_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vuint32m1_t *)(op6), (vbool32_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (vuint32m1_t)(op11), (vuint32m1_t)(op12), (vuint32m1_t)(op13), (vuint32m1_t)(op14), (const uint32_t *)(op15), (size_t *)(op16), (size_t)(op17)) +#define vlseg7e32ff_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg7e32ff_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vuint32mf2_t *)(op6), (const uint32_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg7e32ff_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlseg7e32ff_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vuint32mf2_t *)(op6), (vbool64_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (vuint32mf2_t)(op11), (vuint32mf2_t)(op12), (vuint32mf2_t)(op13), (vuint32mf2_t)(op14), (const uint32_t *)(op15), (size_t *)(op16), (size_t)(op17)) +#define vlseg8e32ff_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg8e32ff_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vint32m1_t *)(op6), (vint32m1_t *)(op7), (const int32_t *)(op8), (size_t *)(op9), (size_t)(op10)) +#define vlseg8e32ff_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlseg8e32ff_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vint32m1_t *)(op6), (vint32m1_t *)(op7), (vbool32_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (vint32m1_t)(op11), (vint32m1_t)(op12), (vint32m1_t)(op13), (vint32m1_t)(op14), (vint32m1_t)(op15), (vint32m1_t)(op16), (const int32_t *)(op17), (size_t *)(op18), (size_t)(op19)) +#define vlseg8e32ff_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg8e32ff_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vint32mf2_t *)(op6), (vint32mf2_t *)(op7), (const int32_t *)(op8), (size_t *)(op9), (size_t)(op10)) +#define vlseg8e32ff_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlseg8e32ff_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vint32mf2_t *)(op6), (vint32mf2_t *)(op7), (vbool64_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (vint32mf2_t)(op11), (vint32mf2_t)(op12), (vint32mf2_t)(op13), (vint32mf2_t)(op14), (vint32mf2_t)(op15), (vint32mf2_t)(op16), (const int32_t *)(op17), (size_t *)(op18), (size_t)(op19)) +#define vlseg8e32ff_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg8e32ff_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vuint32m1_t *)(op6), (vuint32m1_t *)(op7), (const uint32_t *)(op8), (size_t *)(op9), (size_t)(op10)) +#define vlseg8e32ff_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlseg8e32ff_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vuint32m1_t *)(op6), (vuint32m1_t *)(op7), (vbool32_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (vuint32m1_t)(op11), (vuint32m1_t)(op12), (vuint32m1_t)(op13), (vuint32m1_t)(op14), (vuint32m1_t)(op15), (vuint32m1_t)(op16), (const uint32_t *)(op17), (size_t *)(op18), (size_t)(op19)) +#define vlseg8e32ff_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg8e32ff_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vuint32mf2_t *)(op6), (vuint32mf2_t *)(op7), (const uint32_t *)(op8), (size_t *)(op9), (size_t)(op10)) +#define vlseg8e32ff_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlseg8e32ff_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vuint32mf2_t *)(op6), (vuint32mf2_t *)(op7), (vbool64_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (vuint32mf2_t)(op11), (vuint32mf2_t)(op12), (vuint32mf2_t)(op13), (vuint32mf2_t)(op14), (vuint32mf2_t)(op15), (vuint32mf2_t)(op16), (const uint32_t *)(op17), (size_t *)(op18), (size_t)(op19)) +#define vlseg2e64ff_v_i64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e64ff_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (const int64_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e64ff_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e64ff_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vbool64_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (const int64_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg2e64ff_v_i64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e64ff_v_i64m2((vint64m2_t *)(op0), (vint64m2_t *)(op1), (const int64_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e64ff_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e64ff_v_i64m2_m((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vbool32_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (const int64_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg2e64ff_v_i64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e64ff_v_i64m4((vint64m4_t *)(op0), (vint64m4_t *)(op1), (const int64_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e64ff_v_i64m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e64ff_v_i64m4_m((vint64m4_t *)(op0), (vint64m4_t *)(op1), (vbool16_t)(op2), (vint64m4_t)(op3), (vint64m4_t)(op4), (const int64_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg2e64ff_v_u64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e64ff_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (const uint64_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e64ff_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e64ff_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vbool64_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (const uint64_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg2e64ff_v_u64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e64ff_v_u64m2((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (const uint64_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e64ff_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e64ff_v_u64m2_m((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vbool32_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (const uint64_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg2e64ff_v_u64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e64ff_v_u64m4((vuint64m4_t *)(op0), (vuint64m4_t *)(op1), (const uint64_t *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e64ff_v_u64m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e64ff_v_u64m4_m((vuint64m4_t *)(op0), (vuint64m4_t *)(op1), (vbool16_t)(op2), (vuint64m4_t)(op3), (vuint64m4_t)(op4), (const uint64_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg3e64ff_v_i64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg3e64ff_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (const int64_t *)(op3), (size_t *)(op4), (size_t)(op5)) +#define vlseg3e64ff_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg3e64ff_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vbool64_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (const int64_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg3e64ff_v_i64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg3e64ff_v_i64m2((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vint64m2_t *)(op2), (const int64_t *)(op3), (size_t *)(op4), (size_t)(op5)) +#define vlseg3e64ff_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg3e64ff_v_i64m2_m((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vint64m2_t *)(op2), (vbool32_t)(op3), (vint64m2_t)(op4), (vint64m2_t)(op5), (vint64m2_t)(op6), (const int64_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg3e64ff_v_u64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg3e64ff_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (const uint64_t *)(op3), (size_t *)(op4), (size_t)(op5)) +#define vlseg3e64ff_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg3e64ff_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vbool64_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (const uint64_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg3e64ff_v_u64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg3e64ff_v_u64m2((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vuint64m2_t *)(op2), (const uint64_t *)(op3), (size_t *)(op4), (size_t)(op5)) +#define vlseg3e64ff_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg3e64ff_v_u64m2_m((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vuint64m2_t *)(op2), (vbool32_t)(op3), (vuint64m2_t)(op4), (vuint64m2_t)(op5), (vuint64m2_t)(op6), (const uint64_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg4e64ff_v_i64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg4e64ff_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (const int64_t *)(op4), (size_t *)(op5), (size_t)(op6)) +#define vlseg4e64ff_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlseg4e64ff_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vbool64_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (const int64_t *)(op9), (size_t *)(op10), (size_t)(op11)) +#define vlseg4e64ff_v_i64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg4e64ff_v_i64m2((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vint64m2_t *)(op2), (vint64m2_t *)(op3), (const int64_t *)(op4), (size_t *)(op5), (size_t)(op6)) +#define vlseg4e64ff_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlseg4e64ff_v_i64m2_m((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vint64m2_t *)(op2), (vint64m2_t *)(op3), (vbool32_t)(op4), (vint64m2_t)(op5), (vint64m2_t)(op6), (vint64m2_t)(op7), (vint64m2_t)(op8), (const int64_t *)(op9), (size_t *)(op10), (size_t)(op11)) +#define vlseg4e64ff_v_u64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg4e64ff_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (const uint64_t *)(op4), (size_t *)(op5), (size_t)(op6)) +#define vlseg4e64ff_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlseg4e64ff_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vbool64_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (const uint64_t *)(op9), (size_t *)(op10), (size_t)(op11)) +#define vlseg4e64ff_v_u64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg4e64ff_v_u64m2((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vuint64m2_t *)(op2), (vuint64m2_t *)(op3), (const uint64_t *)(op4), (size_t *)(op5), (size_t)(op6)) +#define vlseg4e64ff_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlseg4e64ff_v_u64m2_m((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vuint64m2_t *)(op2), (vuint64m2_t *)(op3), (vbool32_t)(op4), (vuint64m2_t)(op5), (vuint64m2_t)(op6), (vuint64m2_t)(op7), (vuint64m2_t)(op8), (const uint64_t *)(op9), (size_t *)(op10), (size_t)(op11)) +#define vlseg5e64ff_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg5e64ff_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (const int64_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg5e64ff_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlseg5e64ff_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vbool64_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (const int64_t *)(op11), (size_t *)(op12), (size_t)(op13)) +#define vlsseg2e8_v_i8m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e8_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (const int8_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e8_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vbool8_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (const int8_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg2e8_v_i8m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e8_v_i8m2((vint8m2_t *)(op0), (vint8m2_t *)(op1), (const int8_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e8_v_i8m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e8_v_i8m2_m((vint8m2_t *)(op0), (vint8m2_t *)(op1), (vbool4_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (const int8_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg2e8_v_i8m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e8_v_i8m4((vint8m4_t *)(op0), (vint8m4_t *)(op1), (const int8_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e8_v_i8m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e8_v_i8m4_m((vint8m4_t *)(op0), (vint8m4_t *)(op1), (vbool2_t)(op2), (vint8m4_t)(op3), (vint8m4_t)(op4), (const int8_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg2e8_v_i8mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e8_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (const int8_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e8_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vbool16_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (const int8_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg2e8_v_i8mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e8_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (const int8_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e8_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vbool32_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (const int8_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg2e8_v_i8mf8(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e8_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (const int8_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e8_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vbool64_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (const int8_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlseg5e64ff_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg5e64ff_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (const uint64_t *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg5e64ff_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlseg5e64ff_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vbool64_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (const uint64_t *)(op11), (size_t *)(op12), (size_t)(op13)) +#define vlseg6e64ff_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg6e64ff_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (const int64_t *)(op6), (size_t *)(op7), (size_t)(op8)) +#define vlseg6e64ff_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlseg6e64ff_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vbool64_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (vint64m1_t)(op11), (vint64m1_t)(op12), (const int64_t *)(op13), (size_t *)(op14), (size_t)(op15)) +#define vlseg6e64ff_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg6e64ff_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (const uint64_t *)(op6), (size_t *)(op7), (size_t)(op8)) +#define vlseg6e64ff_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlseg6e64ff_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vbool64_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (vuint64m1_t)(op11), (vuint64m1_t)(op12), (const uint64_t *)(op13), (size_t *)(op14), (size_t)(op15)) +#define vlseg7e64ff_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg7e64ff_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vint64m1_t *)(op6), (const int64_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg7e64ff_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlseg7e64ff_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vint64m1_t *)(op6), (vbool64_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (vint64m1_t)(op11), (vint64m1_t)(op12), (vint64m1_t)(op13), (vint64m1_t)(op14), (const int64_t *)(op15), (size_t *)(op16), (size_t)(op17)) +#define vlseg7e64ff_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg7e64ff_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vuint64m1_t *)(op6), (const uint64_t *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg7e64ff_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlseg7e64ff_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vuint64m1_t *)(op6), (vbool64_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (vuint64m1_t)(op11), (vuint64m1_t)(op12), (vuint64m1_t)(op13), (vuint64m1_t)(op14), (const uint64_t *)(op15), (size_t *)(op16), (size_t)(op17)) +#define vlseg8e64ff_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg8e64ff_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vint64m1_t *)(op6), (vint64m1_t *)(op7), (const int64_t *)(op8), (size_t *)(op9), (size_t)(op10)) +#define vlseg8e64ff_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlseg8e64ff_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vint64m1_t *)(op6), (vint64m1_t *)(op7), (vbool64_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (vint64m1_t)(op11), (vint64m1_t)(op12), (vint64m1_t)(op13), (vint64m1_t)(op14), (vint64m1_t)(op15), (vint64m1_t)(op16), (const int64_t *)(op17), (size_t *)(op18), (size_t)(op19)) +#define vlseg8e64ff_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg8e64ff_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vuint64m1_t *)(op6), (vuint64m1_t *)(op7), (const uint64_t *)(op8), (size_t *)(op9), (size_t)(op10)) +#define vlseg8e64ff_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlseg8e64ff_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vuint64m1_t *)(op6), (vuint64m1_t *)(op7), (vbool64_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (vuint64m1_t)(op11), (vuint64m1_t)(op12), (vuint64m1_t)(op13), (vuint64m1_t)(op14), (vuint64m1_t)(op15), (vuint64m1_t)(op16), (const uint64_t *)(op17), (size_t *)(op18), (size_t)(op19)) +#define vlsseg2e8_v_u8m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e8_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (const uint8_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e8_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vbool8_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (const uint8_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg2e8_v_u8m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e8_v_u8m2((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (const uint8_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e8_v_u8m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e8_v_u8m2_m((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (vbool4_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (const uint8_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg2e8_v_u8m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e8_v_u8m4((vuint8m4_t *)(op0), (vuint8m4_t *)(op1), (const uint8_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e8_v_u8m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e8_v_u8m4_m((vuint8m4_t *)(op0), (vuint8m4_t *)(op1), (vbool2_t)(op2), (vuint8m4_t)(op3), (vuint8m4_t)(op4), (const uint8_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg2e8_v_u8mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e8_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (const uint8_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e8_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vbool16_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (const uint8_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg2e8_v_u8mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e8_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (const uint8_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e8_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vbool32_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (const uint8_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg2e8_v_u8mf8(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e8_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (const uint8_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e8_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vbool64_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (const uint8_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vluxseg2ei8_v_i8m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (const int8_t *)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vbool8_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (const int8_t *)(op5), (vuint8m1_t)(op6), (size_t)(op7)) +#define vluxseg2ei8_v_i8m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_i8m2((vint8m2_t *)(op0), (vint8m2_t *)(op1), (const int8_t *)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_i8m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_i8m2_m((vint8m2_t *)(op0), (vint8m2_t *)(op1), (vbool4_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (const int8_t *)(op5), (vuint8m2_t)(op6), (size_t)(op7)) +#define vluxseg2ei8_v_i8m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_i8m4((vint8m4_t *)(op0), (vint8m4_t *)(op1), (const int8_t *)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_i8m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_i8m4_m((vint8m4_t *)(op0), (vint8m4_t *)(op1), (vbool2_t)(op2), (vint8m4_t)(op3), (vint8m4_t)(op4), (const int8_t *)(op5), (vuint8m4_t)(op6), (size_t)(op7)) +#define vluxseg2ei8_v_i8mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (const int8_t *)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vbool16_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (const int8_t *)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vluxseg2ei8_v_i8mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (const int8_t *)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vbool32_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (const int8_t *)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vluxseg2ei8_v_i8mf8(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (const int8_t *)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vbool64_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (const int8_t *)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vlsseg3e8_v_i8m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlsseg3e8_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (const int8_t *)(op3), (ptrdiff_t)(op4), (size_t)(op5)) +#define vlsseg3e8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg3e8_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vbool8_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (const int8_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg3e8_v_i8m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlsseg3e8_v_i8m2((vint8m2_t *)(op0), (vint8m2_t *)(op1), (vint8m2_t *)(op2), (const int8_t *)(op3), (ptrdiff_t)(op4), (size_t)(op5)) +#define vlsseg3e8_v_i8m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg3e8_v_i8m2_m((vint8m2_t *)(op0), (vint8m2_t *)(op1), (vint8m2_t *)(op2), (vbool4_t)(op3), (vint8m2_t)(op4), (vint8m2_t)(op5), (vint8m2_t)(op6), (const int8_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg3e8_v_i8mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlsseg3e8_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (const int8_t *)(op3), (ptrdiff_t)(op4), (size_t)(op5)) +#define vlsseg3e8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg3e8_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vbool16_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (const int8_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg3e8_v_i8mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlsseg3e8_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (const int8_t *)(op3), (ptrdiff_t)(op4), (size_t)(op5)) +#define vlsseg3e8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg3e8_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vbool32_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (const int8_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg3e8_v_i8mf8(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlsseg3e8_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (const int8_t *)(op3), (ptrdiff_t)(op4), (size_t)(op5)) +#define vlsseg3e8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg3e8_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vbool64_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (const int8_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vluxseg2ei8_v_u8m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (const uint8_t *)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vbool8_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (const uint8_t *)(op5), (vuint8m1_t)(op6), (size_t)(op7)) +#define vluxseg2ei8_v_u8m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_u8m2((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (const uint8_t *)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_u8m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_u8m2_m((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (vbool4_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (const uint8_t *)(op5), (vuint8m2_t)(op6), (size_t)(op7)) +#define vluxseg2ei8_v_u8m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_u8m4((vuint8m4_t *)(op0), (vuint8m4_t *)(op1), (const uint8_t *)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_u8m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_u8m4_m((vuint8m4_t *)(op0), (vuint8m4_t *)(op1), (vbool2_t)(op2), (vuint8m4_t)(op3), (vuint8m4_t)(op4), (const uint8_t *)(op5), (vuint8m4_t)(op6), (size_t)(op7)) +#define vluxseg2ei8_v_u8mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (const uint8_t *)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vbool16_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (const uint8_t *)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vluxseg2ei8_v_u8mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (const uint8_t *)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vbool32_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (const uint8_t *)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vluxseg2ei8_v_u8mf8(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (const uint8_t *)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vbool64_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (const uint8_t *)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vlsseg3e8_v_u8m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlsseg3e8_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (const uint8_t *)(op3), (ptrdiff_t)(op4), (size_t)(op5)) +#define vlsseg3e8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg3e8_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vbool8_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (const uint8_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg3e8_v_u8m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlsseg3e8_v_u8m2((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (vuint8m2_t *)(op2), (const uint8_t *)(op3), (ptrdiff_t)(op4), (size_t)(op5)) +#define vlsseg3e8_v_u8m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg3e8_v_u8m2_m((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (vuint8m2_t *)(op2), (vbool4_t)(op3), (vuint8m2_t)(op4), (vuint8m2_t)(op5), (vuint8m2_t)(op6), (const uint8_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg3e8_v_u8mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlsseg3e8_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (const uint8_t *)(op3), (ptrdiff_t)(op4), (size_t)(op5)) +#define vlsseg3e8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg3e8_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vbool16_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (const uint8_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg3e8_v_u8mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlsseg3e8_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (const uint8_t *)(op3), (ptrdiff_t)(op4), (size_t)(op5)) +#define vlsseg3e8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg3e8_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vbool32_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (const uint8_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg3e8_v_u8mf8(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlsseg3e8_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (const uint8_t *)(op3), (ptrdiff_t)(op4), (size_t)(op5)) +#define vlsseg3e8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg3e8_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vbool64_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (const uint8_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg4e8_v_i8m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlsseg4e8_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (const int8_t *)(op4), (ptrdiff_t)(op5), (size_t)(op6)) +#define vlsseg4e8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlsseg4e8_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vbool8_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (const int8_t *)(op9), (ptrdiff_t)(op10), (size_t)(op11)) +#define vlsseg4e8_v_i8m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlsseg4e8_v_i8m2((vint8m2_t *)(op0), (vint8m2_t *)(op1), (vint8m2_t *)(op2), (vint8m2_t *)(op3), (const int8_t *)(op4), (ptrdiff_t)(op5), (size_t)(op6)) +#define vlsseg4e8_v_i8m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlsseg4e8_v_i8m2_m((vint8m2_t *)(op0), (vint8m2_t *)(op1), (vint8m2_t *)(op2), (vint8m2_t *)(op3), (vbool4_t)(op4), (vint8m2_t)(op5), (vint8m2_t)(op6), (vint8m2_t)(op7), (vint8m2_t)(op8), (const int8_t *)(op9), (ptrdiff_t)(op10), (size_t)(op11)) +#define vlsseg4e8_v_i8mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlsseg4e8_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (const int8_t *)(op4), (ptrdiff_t)(op5), (size_t)(op6)) +#define vlsseg4e8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlsseg4e8_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vbool16_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (const int8_t *)(op9), (ptrdiff_t)(op10), (size_t)(op11)) +#define vlsseg4e8_v_i8mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlsseg4e8_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (const int8_t *)(op4), (ptrdiff_t)(op5), (size_t)(op6)) +#define vlsseg4e8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlsseg4e8_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vbool32_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (const int8_t *)(op9), (ptrdiff_t)(op10), (size_t)(op11)) +#define vlsseg4e8_v_i8mf8(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlsseg4e8_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (const int8_t *)(op4), (ptrdiff_t)(op5), (size_t)(op6)) +#define vlsseg4e8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlsseg4e8_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vbool64_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (const int8_t *)(op9), (ptrdiff_t)(op10), (size_t)(op11)) +#define vlsseg4e8_v_u8m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlsseg4e8_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (const uint8_t *)(op4), (ptrdiff_t)(op5), (size_t)(op6)) +#define vlsseg4e8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlsseg4e8_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vbool8_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (const uint8_t *)(op9), (ptrdiff_t)(op10), (size_t)(op11)) +#define vlsseg4e8_v_u8m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlsseg4e8_v_u8m2((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (vuint8m2_t *)(op2), (vuint8m2_t *)(op3), (const uint8_t *)(op4), (ptrdiff_t)(op5), (size_t)(op6)) +#define vlsseg4e8_v_u8m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlsseg4e8_v_u8m2_m((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (vuint8m2_t *)(op2), (vuint8m2_t *)(op3), (vbool4_t)(op4), (vuint8m2_t)(op5), (vuint8m2_t)(op6), (vuint8m2_t)(op7), (vuint8m2_t)(op8), (const uint8_t *)(op9), (ptrdiff_t)(op10), (size_t)(op11)) +#define vlsseg4e8_v_u8mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlsseg4e8_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (const uint8_t *)(op4), (ptrdiff_t)(op5), (size_t)(op6)) +#define vlsseg4e8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlsseg4e8_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vbool16_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (const uint8_t *)(op9), (ptrdiff_t)(op10), (size_t)(op11)) +#define vlsseg4e8_v_u8mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlsseg4e8_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (const uint8_t *)(op4), (ptrdiff_t)(op5), (size_t)(op6)) +#define vlsseg4e8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlsseg4e8_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vbool32_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (const uint8_t *)(op9), (ptrdiff_t)(op10), (size_t)(op11)) +#define vlsseg4e8_v_u8mf8(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlsseg4e8_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (const uint8_t *)(op4), (ptrdiff_t)(op5), (size_t)(op6)) +#define vlsseg4e8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlsseg4e8_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vbool64_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (const uint8_t *)(op9), (ptrdiff_t)(op10), (size_t)(op11)) +#define vlsseg5e8_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg5e8_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (const int8_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg5e8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlsseg5e8_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vbool8_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (const int8_t *)(op11), (ptrdiff_t)(op12), (size_t)(op13)) +#define vlsseg5e8_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg5e8_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (const int8_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg5e8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlsseg5e8_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vbool16_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (const int8_t *)(op11), (ptrdiff_t)(op12), (size_t)(op13)) +#define vlsseg5e8_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg5e8_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (const int8_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg5e8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlsseg5e8_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vbool32_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (const int8_t *)(op11), (ptrdiff_t)(op12), (size_t)(op13)) +#define vlsseg5e8_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg5e8_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (const int8_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg5e8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlsseg5e8_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vbool64_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (const int8_t *)(op11), (ptrdiff_t)(op12), (size_t)(op13)) +#define vlsseg5e8_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg5e8_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (const uint8_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg5e8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlsseg5e8_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vbool8_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (const uint8_t *)(op11), (ptrdiff_t)(op12), (size_t)(op13)) +#define vlsseg5e8_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg5e8_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (const uint8_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg5e8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlsseg5e8_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vbool16_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (const uint8_t *)(op11), (ptrdiff_t)(op12), (size_t)(op13)) +#define vlsseg5e8_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg5e8_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (const uint8_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg5e8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlsseg5e8_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vbool32_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (const uint8_t *)(op11), (ptrdiff_t)(op12), (size_t)(op13)) +#define vlsseg5e8_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg5e8_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (const uint8_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg5e8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlsseg5e8_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vbool64_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (const uint8_t *)(op11), (ptrdiff_t)(op12), (size_t)(op13)) +#define vlsseg6e8_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlsseg6e8_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (const int8_t *)(op6), (ptrdiff_t)(op7), (size_t)(op8)) +#define vlsseg6e8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlsseg6e8_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vbool8_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (vint8m1_t)(op11), (vint8m1_t)(op12), (const int8_t *)(op13), (ptrdiff_t)(op14), (size_t)(op15)) +#define vlsseg6e8_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlsseg6e8_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (const int8_t *)(op6), (ptrdiff_t)(op7), (size_t)(op8)) +#define vlsseg6e8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlsseg6e8_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vbool16_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (vint8mf2_t)(op11), (vint8mf2_t)(op12), (const int8_t *)(op13), (ptrdiff_t)(op14), (size_t)(op15)) +#define vlsseg6e8_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlsseg6e8_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (const int8_t *)(op6), (ptrdiff_t)(op7), (size_t)(op8)) +#define vlsseg6e8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlsseg6e8_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vbool32_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (vint8mf4_t)(op11), (vint8mf4_t)(op12), (const int8_t *)(op13), (ptrdiff_t)(op14), (size_t)(op15)) +#define vlsseg6e8_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlsseg6e8_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (const int8_t *)(op6), (ptrdiff_t)(op7), (size_t)(op8)) +#define vlsseg6e8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlsseg6e8_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vbool64_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (vint8mf8_t)(op11), (vint8mf8_t)(op12), (const int8_t *)(op13), (ptrdiff_t)(op14), (size_t)(op15)) +#define vlsseg6e8_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlsseg6e8_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (const uint8_t *)(op6), (ptrdiff_t)(op7), (size_t)(op8)) +#define vlsseg6e8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlsseg6e8_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vbool8_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (vuint8m1_t)(op11), (vuint8m1_t)(op12), (const uint8_t *)(op13), (ptrdiff_t)(op14), (size_t)(op15)) +#define vlsseg6e8_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlsseg6e8_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (const uint8_t *)(op6), (ptrdiff_t)(op7), (size_t)(op8)) +#define vlsseg6e8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlsseg6e8_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vbool16_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (vuint8mf2_t)(op11), (vuint8mf2_t)(op12), (const uint8_t *)(op13), (ptrdiff_t)(op14), (size_t)(op15)) +#define vlsseg6e8_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlsseg6e8_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (const uint8_t *)(op6), (ptrdiff_t)(op7), (size_t)(op8)) +#define vlsseg6e8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlsseg6e8_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vbool32_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (vuint8mf4_t)(op11), (vuint8mf4_t)(op12), (const uint8_t *)(op13), (ptrdiff_t)(op14), (size_t)(op15)) +#define vlsseg6e8_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlsseg6e8_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (const uint8_t *)(op6), (ptrdiff_t)(op7), (size_t)(op8)) +#define vlsseg6e8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlsseg6e8_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vbool64_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (vuint8mf8_t)(op11), (vuint8mf8_t)(op12), (const uint8_t *)(op13), (ptrdiff_t)(op14), (size_t)(op15)) +#define vlsseg7e8_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg7e8_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vint8m1_t *)(op6), (const int8_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg7e8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlsseg7e8_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vint8m1_t *)(op6), (vbool8_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (vint8m1_t)(op11), (vint8m1_t)(op12), (vint8m1_t)(op13), (vint8m1_t)(op14), (const int8_t *)(op15), (ptrdiff_t)(op16), (size_t)(op17)) +#define vlsseg7e8_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg7e8_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vint8mf2_t *)(op6), (const int8_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg7e8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlsseg7e8_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vint8mf2_t *)(op6), (vbool16_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (vint8mf2_t)(op11), (vint8mf2_t)(op12), (vint8mf2_t)(op13), (vint8mf2_t)(op14), (const int8_t *)(op15), (ptrdiff_t)(op16), (size_t)(op17)) +#define vlsseg7e8_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg7e8_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vint8mf4_t *)(op6), (const int8_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg7e8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlsseg7e8_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vint8mf4_t *)(op6), (vbool32_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (vint8mf4_t)(op11), (vint8mf4_t)(op12), (vint8mf4_t)(op13), (vint8mf4_t)(op14), (const int8_t *)(op15), (ptrdiff_t)(op16), (size_t)(op17)) +#define vlsseg7e8_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg7e8_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vint8mf8_t *)(op6), (const int8_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg7e8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlsseg7e8_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vint8mf8_t *)(op6), (vbool64_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (vint8mf8_t)(op11), (vint8mf8_t)(op12), (vint8mf8_t)(op13), (vint8mf8_t)(op14), (const int8_t *)(op15), (ptrdiff_t)(op16), (size_t)(op17)) +#define vlsseg7e8_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg7e8_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vuint8m1_t *)(op6), (const uint8_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg7e8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlsseg7e8_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vuint8m1_t *)(op6), (vbool8_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (vuint8m1_t)(op11), (vuint8m1_t)(op12), (vuint8m1_t)(op13), (vuint8m1_t)(op14), (const uint8_t *)(op15), (ptrdiff_t)(op16), (size_t)(op17)) +#define vlsseg7e8_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg7e8_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vuint8mf2_t *)(op6), (const uint8_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg7e8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlsseg7e8_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vuint8mf2_t *)(op6), (vbool16_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (vuint8mf2_t)(op11), (vuint8mf2_t)(op12), (vuint8mf2_t)(op13), (vuint8mf2_t)(op14), (const uint8_t *)(op15), (ptrdiff_t)(op16), (size_t)(op17)) +#define vlsseg7e8_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg7e8_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vuint8mf4_t *)(op6), (const uint8_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg7e8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlsseg7e8_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vuint8mf4_t *)(op6), (vbool32_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (vuint8mf4_t)(op11), (vuint8mf4_t)(op12), (vuint8mf4_t)(op13), (vuint8mf4_t)(op14), (const uint8_t *)(op15), (ptrdiff_t)(op16), (size_t)(op17)) +#define vlsseg7e8_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg7e8_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vuint8mf8_t *)(op6), (const uint8_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg7e8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlsseg7e8_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vuint8mf8_t *)(op6), (vbool64_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (vuint8mf8_t)(op11), (vuint8mf8_t)(op12), (vuint8mf8_t)(op13), (vuint8mf8_t)(op14), (const uint8_t *)(op15), (ptrdiff_t)(op16), (size_t)(op17)) +#define vlsseg8e8_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlsseg8e8_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vint8m1_t *)(op6), (vint8m1_t *)(op7), (const int8_t *)(op8), (ptrdiff_t)(op9), (size_t)(op10)) +#define vlsseg8e8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlsseg8e8_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vint8m1_t *)(op6), (vint8m1_t *)(op7), (vbool8_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (vint8m1_t)(op11), (vint8m1_t)(op12), (vint8m1_t)(op13), (vint8m1_t)(op14), (vint8m1_t)(op15), (vint8m1_t)(op16), (const int8_t *)(op17), (ptrdiff_t)(op18), (size_t)(op19)) +#define vlsseg8e8_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlsseg8e8_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vint8mf2_t *)(op6), (vint8mf2_t *)(op7), (const int8_t *)(op8), (ptrdiff_t)(op9), (size_t)(op10)) +#define vlsseg8e8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlsseg8e8_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vint8mf2_t *)(op6), (vint8mf2_t *)(op7), (vbool16_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (vint8mf2_t)(op11), (vint8mf2_t)(op12), (vint8mf2_t)(op13), (vint8mf2_t)(op14), (vint8mf2_t)(op15), (vint8mf2_t)(op16), (const int8_t *)(op17), (ptrdiff_t)(op18), (size_t)(op19)) +#define vlsseg8e8_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlsseg8e8_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vint8mf4_t *)(op6), (vint8mf4_t *)(op7), (const int8_t *)(op8), (ptrdiff_t)(op9), (size_t)(op10)) +#define vlsseg8e8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlsseg8e8_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vint8mf4_t *)(op6), (vint8mf4_t *)(op7), (vbool32_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (vint8mf4_t)(op11), (vint8mf4_t)(op12), (vint8mf4_t)(op13), (vint8mf4_t)(op14), (vint8mf4_t)(op15), (vint8mf4_t)(op16), (const int8_t *)(op17), (ptrdiff_t)(op18), (size_t)(op19)) +#define vlsseg8e8_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlsseg8e8_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vint8mf8_t *)(op6), (vint8mf8_t *)(op7), (const int8_t *)(op8), (ptrdiff_t)(op9), (size_t)(op10)) +#define vlsseg8e8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlsseg8e8_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vint8mf8_t *)(op6), (vint8mf8_t *)(op7), (vbool64_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (vint8mf8_t)(op11), (vint8mf8_t)(op12), (vint8mf8_t)(op13), (vint8mf8_t)(op14), (vint8mf8_t)(op15), (vint8mf8_t)(op16), (const int8_t *)(op17), (ptrdiff_t)(op18), (size_t)(op19)) +#define vsseg2e8_v_i8m1(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e8_v_i8m1((int8_t *)(op0), (vint8m1_t)(op1), (vint8m1_t)(op2), (size_t)(op3)) +#define vsseg2e8_v_i8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e8_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vsseg2e8_v_i8m2(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e8_v_i8m2((int8_t *)(op0), (vint8m2_t)(op1), (vint8m2_t)(op2), (size_t)(op3)) +#define vsseg2e8_v_i8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e8_v_i8m2_m((vbool4_t)(op0), (int8_t *)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vsseg2e8_v_i8m4(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e8_v_i8m4((int8_t *)(op0), (vint8m4_t)(op1), (vint8m4_t)(op2), (size_t)(op3)) +#define vsseg2e8_v_i8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e8_v_i8m4_m((vbool2_t)(op0), (int8_t *)(op1), (vint8m4_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vsseg2e8_v_i8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e8_v_i8mf2((int8_t *)(op0), (vint8mf2_t)(op1), (vint8mf2_t)(op2), (size_t)(op3)) +#define vsseg2e8_v_i8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e8_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vsseg2e8_v_i8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e8_v_i8mf4((int8_t *)(op0), (vint8mf4_t)(op1), (vint8mf4_t)(op2), (size_t)(op3)) +#define vsseg2e8_v_i8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e8_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vsseg2e8_v_i8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e8_v_i8mf8((int8_t *)(op0), (vint8mf8_t)(op1), (vint8mf8_t)(op2), (size_t)(op3)) +#define vsseg2e8_v_i8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e8_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vlsseg8e8_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlsseg8e8_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vuint8m1_t *)(op6), (vuint8m1_t *)(op7), (const uint8_t *)(op8), (ptrdiff_t)(op9), (size_t)(op10)) +#define vlsseg8e8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlsseg8e8_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vuint8m1_t *)(op6), (vuint8m1_t *)(op7), (vbool8_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (vuint8m1_t)(op11), (vuint8m1_t)(op12), (vuint8m1_t)(op13), (vuint8m1_t)(op14), (vuint8m1_t)(op15), (vuint8m1_t)(op16), (const uint8_t *)(op17), (ptrdiff_t)(op18), (size_t)(op19)) +#define vlsseg8e8_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlsseg8e8_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vuint8mf2_t *)(op6), (vuint8mf2_t *)(op7), (const uint8_t *)(op8), (ptrdiff_t)(op9), (size_t)(op10)) +#define vlsseg8e8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlsseg8e8_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vuint8mf2_t *)(op6), (vuint8mf2_t *)(op7), (vbool16_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (vuint8mf2_t)(op11), (vuint8mf2_t)(op12), (vuint8mf2_t)(op13), (vuint8mf2_t)(op14), (vuint8mf2_t)(op15), (vuint8mf2_t)(op16), (const uint8_t *)(op17), (ptrdiff_t)(op18), (size_t)(op19)) +#define vlsseg8e8_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlsseg8e8_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vuint8mf4_t *)(op6), (vuint8mf4_t *)(op7), (const uint8_t *)(op8), (ptrdiff_t)(op9), (size_t)(op10)) +#define vlsseg8e8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlsseg8e8_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vuint8mf4_t *)(op6), (vuint8mf4_t *)(op7), (vbool32_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (vuint8mf4_t)(op11), (vuint8mf4_t)(op12), (vuint8mf4_t)(op13), (vuint8mf4_t)(op14), (vuint8mf4_t)(op15), (vuint8mf4_t)(op16), (const uint8_t *)(op17), (ptrdiff_t)(op18), (size_t)(op19)) +#define vlsseg8e8_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlsseg8e8_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vuint8mf8_t *)(op6), (vuint8mf8_t *)(op7), (const uint8_t *)(op8), (ptrdiff_t)(op9), (size_t)(op10)) +#define vlsseg8e8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlsseg8e8_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vuint8mf8_t *)(op6), (vuint8mf8_t *)(op7), (vbool64_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (vuint8mf8_t)(op11), (vuint8mf8_t)(op12), (vuint8mf8_t)(op13), (vuint8mf8_t)(op14), (vuint8mf8_t)(op15), (vuint8mf8_t)(op16), (const uint8_t *)(op17), (ptrdiff_t)(op18), (size_t)(op19)) +#define vlsseg2e16_v_i16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e16_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (const int16_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e16_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vbool16_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (const int16_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg2e16_v_i16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e16_v_i16m2((vint16m2_t *)(op0), (vint16m2_t *)(op1), (const int16_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e16_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e16_v_i16m2_m((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vbool8_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (const int16_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg2e16_v_i16m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e16_v_i16m4((vint16m4_t *)(op0), (vint16m4_t *)(op1), (const int16_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e16_v_i16m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e16_v_i16m4_m((vint16m4_t *)(op0), (vint16m4_t *)(op1), (vbool4_t)(op2), (vint16m4_t)(op3), (vint16m4_t)(op4), (const int16_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg2e16_v_i16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e16_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (const int16_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e16_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vbool32_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (const int16_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg2e16_v_i16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e16_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (const int16_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e16_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vbool64_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (const int16_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg2e16_v_u16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e16_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (const uint16_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e16_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vbool16_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (const uint16_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg2e16_v_u16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e16_v_u16m2((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (const uint16_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e16_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e16_v_u16m2_m((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vbool8_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (const uint16_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg2e16_v_u16m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e16_v_u16m4((vuint16m4_t *)(op0), (vuint16m4_t *)(op1), (const uint16_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e16_v_u16m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e16_v_u16m4_m((vuint16m4_t *)(op0), (vuint16m4_t *)(op1), (vbool4_t)(op2), (vuint16m4_t)(op3), (vuint16m4_t)(op4), (const uint16_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg2e16_v_u16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e16_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (const uint16_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e16_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vbool32_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (const uint16_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg2e16_v_u16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e16_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (const uint16_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e16_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vbool64_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (const uint16_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg3e16_v_i16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlsseg3e16_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (const int16_t *)(op3), (ptrdiff_t)(op4), (size_t)(op5)) +#define vlsseg3e16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg3e16_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vbool16_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (const int16_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg3e16_v_i16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlsseg3e16_v_i16m2((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vint16m2_t *)(op2), (const int16_t *)(op3), (ptrdiff_t)(op4), (size_t)(op5)) +#define vlsseg3e16_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg3e16_v_i16m2_m((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vint16m2_t *)(op2), (vbool8_t)(op3), (vint16m2_t)(op4), (vint16m2_t)(op5), (vint16m2_t)(op6), (const int16_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg3e16_v_i16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlsseg3e16_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (const int16_t *)(op3), (ptrdiff_t)(op4), (size_t)(op5)) +#define vlsseg3e16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg3e16_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vbool32_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (const int16_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg3e16_v_i16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlsseg3e16_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (const int16_t *)(op3), (ptrdiff_t)(op4), (size_t)(op5)) +#define vlsseg3e16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg3e16_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vbool64_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (const int16_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg3e16_v_u16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlsseg3e16_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (const uint16_t *)(op3), (ptrdiff_t)(op4), (size_t)(op5)) +#define vlsseg3e16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg3e16_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vbool16_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (const uint16_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg3e16_v_u16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlsseg3e16_v_u16m2((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vuint16m2_t *)(op2), (const uint16_t *)(op3), (ptrdiff_t)(op4), (size_t)(op5)) +#define vlsseg3e16_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg3e16_v_u16m2_m((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vuint16m2_t *)(op2), (vbool8_t)(op3), (vuint16m2_t)(op4), (vuint16m2_t)(op5), (vuint16m2_t)(op6), (const uint16_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg3e16_v_u16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlsseg3e16_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (const uint16_t *)(op3), (ptrdiff_t)(op4), (size_t)(op5)) +#define vlsseg3e16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg3e16_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vbool32_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (const uint16_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg3e16_v_u16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlsseg3e16_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (const uint16_t *)(op3), (ptrdiff_t)(op4), (size_t)(op5)) +#define vlsseg3e16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg3e16_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vbool64_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (const uint16_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg4e16_v_i16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlsseg4e16_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (const int16_t *)(op4), (ptrdiff_t)(op5), (size_t)(op6)) +#define vlsseg4e16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlsseg4e16_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vbool16_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (const int16_t *)(op9), (ptrdiff_t)(op10), (size_t)(op11)) +#define vlsseg4e16_v_i16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlsseg4e16_v_i16m2((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vint16m2_t *)(op2), (vint16m2_t *)(op3), (const int16_t *)(op4), (ptrdiff_t)(op5), (size_t)(op6)) +#define vlsseg4e16_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlsseg4e16_v_i16m2_m((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vint16m2_t *)(op2), (vint16m2_t *)(op3), (vbool8_t)(op4), (vint16m2_t)(op5), (vint16m2_t)(op6), (vint16m2_t)(op7), (vint16m2_t)(op8), (const int16_t *)(op9), (ptrdiff_t)(op10), (size_t)(op11)) +#define vlsseg4e16_v_i16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlsseg4e16_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (const int16_t *)(op4), (ptrdiff_t)(op5), (size_t)(op6)) +#define vlsseg4e16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlsseg4e16_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vbool32_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (const int16_t *)(op9), (ptrdiff_t)(op10), (size_t)(op11)) +#define vlsseg4e16_v_i16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlsseg4e16_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (const int16_t *)(op4), (ptrdiff_t)(op5), (size_t)(op6)) +#define vlsseg4e16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlsseg4e16_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vbool64_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (const int16_t *)(op9), (ptrdiff_t)(op10), (size_t)(op11)) +#define vlsseg4e16_v_u16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlsseg4e16_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (const uint16_t *)(op4), (ptrdiff_t)(op5), (size_t)(op6)) +#define vlsseg4e16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlsseg4e16_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vbool16_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (const uint16_t *)(op9), (ptrdiff_t)(op10), (size_t)(op11)) +#define vlsseg4e16_v_u16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlsseg4e16_v_u16m2((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vuint16m2_t *)(op2), (vuint16m2_t *)(op3), (const uint16_t *)(op4), (ptrdiff_t)(op5), (size_t)(op6)) +#define vlsseg4e16_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlsseg4e16_v_u16m2_m((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vuint16m2_t *)(op2), (vuint16m2_t *)(op3), (vbool8_t)(op4), (vuint16m2_t)(op5), (vuint16m2_t)(op6), (vuint16m2_t)(op7), (vuint16m2_t)(op8), (const uint16_t *)(op9), (ptrdiff_t)(op10), (size_t)(op11)) +#define vlsseg4e16_v_u16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlsseg4e16_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (const uint16_t *)(op4), (ptrdiff_t)(op5), (size_t)(op6)) +#define vlsseg4e16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlsseg4e16_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vbool32_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (const uint16_t *)(op9), (ptrdiff_t)(op10), (size_t)(op11)) +#define vlsseg4e16_v_u16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlsseg4e16_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (const uint16_t *)(op4), (ptrdiff_t)(op5), (size_t)(op6)) +#define vlsseg4e16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlsseg4e16_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vbool64_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (const uint16_t *)(op9), (ptrdiff_t)(op10), (size_t)(op11)) +#define vlsseg5e16_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg5e16_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (const int16_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg5e16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlsseg5e16_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vbool16_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (const int16_t *)(op11), (ptrdiff_t)(op12), (size_t)(op13)) +#define vlsseg5e16_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg5e16_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (const int16_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg5e16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlsseg5e16_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vbool32_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (const int16_t *)(op11), (ptrdiff_t)(op12), (size_t)(op13)) +#define vlsseg5e16_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg5e16_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (const int16_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg5e16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlsseg5e16_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vbool64_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (const int16_t *)(op11), (ptrdiff_t)(op12), (size_t)(op13)) +#define vlsseg5e16_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg5e16_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (const uint16_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg5e16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlsseg5e16_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vbool16_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (const uint16_t *)(op11), (ptrdiff_t)(op12), (size_t)(op13)) +#define vlsseg5e16_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg5e16_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (const uint16_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg5e16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlsseg5e16_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vbool32_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (const uint16_t *)(op11), (ptrdiff_t)(op12), (size_t)(op13)) +#define vlsseg5e16_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg5e16_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (const uint16_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg5e16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlsseg5e16_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vbool64_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (const uint16_t *)(op11), (ptrdiff_t)(op12), (size_t)(op13)) +#define vlsseg6e16_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlsseg6e16_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (const int16_t *)(op6), (ptrdiff_t)(op7), (size_t)(op8)) +#define vlsseg6e16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlsseg6e16_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vbool16_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (vint16m1_t)(op11), (vint16m1_t)(op12), (const int16_t *)(op13), (ptrdiff_t)(op14), (size_t)(op15)) +#define vlsseg6e16_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlsseg6e16_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (const int16_t *)(op6), (ptrdiff_t)(op7), (size_t)(op8)) +#define vlsseg6e16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlsseg6e16_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vbool32_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (vint16mf2_t)(op11), (vint16mf2_t)(op12), (const int16_t *)(op13), (ptrdiff_t)(op14), (size_t)(op15)) +#define vlsseg6e16_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlsseg6e16_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (const int16_t *)(op6), (ptrdiff_t)(op7), (size_t)(op8)) +#define vlsseg6e16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlsseg6e16_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vbool64_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (vint16mf4_t)(op11), (vint16mf4_t)(op12), (const int16_t *)(op13), (ptrdiff_t)(op14), (size_t)(op15)) +#define vsseg2e8_v_u8m1(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e8_v_u8m1((uint8_t *)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (size_t)(op3)) +#define vsseg2e8_v_u8m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e8_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vsseg2e8_v_u8m2(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e8_v_u8m2((uint8_t *)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (size_t)(op3)) +#define vsseg2e8_v_u8m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e8_v_u8m2_m((vbool4_t)(op0), (uint8_t *)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vsseg2e8_v_u8m4(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e8_v_u8m4((uint8_t *)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (size_t)(op3)) +#define vsseg2e8_v_u8m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e8_v_u8m4_m((vbool2_t)(op0), (uint8_t *)(op1), (vuint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vsseg2e8_v_u8mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e8_v_u8mf2((uint8_t *)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (size_t)(op3)) +#define vsseg2e8_v_u8mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e8_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vsseg2e8_v_u8mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e8_v_u8mf4((uint8_t *)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (size_t)(op3)) +#define vsseg2e8_v_u8mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e8_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vsseg2e8_v_u8mf8(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e8_v_u8mf8((uint8_t *)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (size_t)(op3)) +#define vsseg2e8_v_u8mf8_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e8_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vlsseg6e16_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlsseg6e16_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (const uint16_t *)(op6), (ptrdiff_t)(op7), (size_t)(op8)) +#define vlsseg6e16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlsseg6e16_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vbool16_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (vuint16m1_t)(op11), (vuint16m1_t)(op12), (const uint16_t *)(op13), (ptrdiff_t)(op14), (size_t)(op15)) +#define vlsseg6e16_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlsseg6e16_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (const uint16_t *)(op6), (ptrdiff_t)(op7), (size_t)(op8)) +#define vlsseg6e16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlsseg6e16_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vbool32_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (vuint16mf2_t)(op11), (vuint16mf2_t)(op12), (const uint16_t *)(op13), (ptrdiff_t)(op14), (size_t)(op15)) +#define vlsseg6e16_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlsseg6e16_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (const uint16_t *)(op6), (ptrdiff_t)(op7), (size_t)(op8)) +#define vlsseg6e16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlsseg6e16_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vbool64_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (vuint16mf4_t)(op11), (vuint16mf4_t)(op12), (const uint16_t *)(op13), (ptrdiff_t)(op14), (size_t)(op15)) +#define vlsseg7e16_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg7e16_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vint16m1_t *)(op6), (const int16_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg7e16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlsseg7e16_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vint16m1_t *)(op6), (vbool16_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (vint16m1_t)(op11), (vint16m1_t)(op12), (vint16m1_t)(op13), (vint16m1_t)(op14), (const int16_t *)(op15), (ptrdiff_t)(op16), (size_t)(op17)) +#define vlsseg7e16_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg7e16_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vint16mf2_t *)(op6), (const int16_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg7e16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlsseg7e16_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vint16mf2_t *)(op6), (vbool32_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (vint16mf2_t)(op11), (vint16mf2_t)(op12), (vint16mf2_t)(op13), (vint16mf2_t)(op14), (const int16_t *)(op15), (ptrdiff_t)(op16), (size_t)(op17)) +#define vlsseg7e16_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg7e16_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vint16mf4_t *)(op6), (const int16_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg7e16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlsseg7e16_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vint16mf4_t *)(op6), (vbool64_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (vint16mf4_t)(op11), (vint16mf4_t)(op12), (vint16mf4_t)(op13), (vint16mf4_t)(op14), (const int16_t *)(op15), (ptrdiff_t)(op16), (size_t)(op17)) +#define vlsseg7e16_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg7e16_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vuint16m1_t *)(op6), (const uint16_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg7e16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlsseg7e16_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vuint16m1_t *)(op6), (vbool16_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (vuint16m1_t)(op11), (vuint16m1_t)(op12), (vuint16m1_t)(op13), (vuint16m1_t)(op14), (const uint16_t *)(op15), (ptrdiff_t)(op16), (size_t)(op17)) +#define vlsseg7e16_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg7e16_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vuint16mf2_t *)(op6), (const uint16_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg7e16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlsseg7e16_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vuint16mf2_t *)(op6), (vbool32_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (vuint16mf2_t)(op11), (vuint16mf2_t)(op12), (vuint16mf2_t)(op13), (vuint16mf2_t)(op14), (const uint16_t *)(op15), (ptrdiff_t)(op16), (size_t)(op17)) +#define vlsseg7e16_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg7e16_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vuint16mf4_t *)(op6), (const uint16_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg7e16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlsseg7e16_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vuint16mf4_t *)(op6), (vbool64_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (vuint16mf4_t)(op11), (vuint16mf4_t)(op12), (vuint16mf4_t)(op13), (vuint16mf4_t)(op14), (const uint16_t *)(op15), (ptrdiff_t)(op16), (size_t)(op17)) +#define vlsseg8e16_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlsseg8e16_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vint16m1_t *)(op6), (vint16m1_t *)(op7), (const int16_t *)(op8), (ptrdiff_t)(op9), (size_t)(op10)) +#define vlsseg8e16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlsseg8e16_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vint16m1_t *)(op6), (vint16m1_t *)(op7), (vbool16_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (vint16m1_t)(op11), (vint16m1_t)(op12), (vint16m1_t)(op13), (vint16m1_t)(op14), (vint16m1_t)(op15), (vint16m1_t)(op16), (const int16_t *)(op17), (ptrdiff_t)(op18), (size_t)(op19)) +#define vlsseg8e16_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlsseg8e16_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vint16mf2_t *)(op6), (vint16mf2_t *)(op7), (const int16_t *)(op8), (ptrdiff_t)(op9), (size_t)(op10)) +#define vlsseg8e16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlsseg8e16_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vint16mf2_t *)(op6), (vint16mf2_t *)(op7), (vbool32_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (vint16mf2_t)(op11), (vint16mf2_t)(op12), (vint16mf2_t)(op13), (vint16mf2_t)(op14), (vint16mf2_t)(op15), (vint16mf2_t)(op16), (const int16_t *)(op17), (ptrdiff_t)(op18), (size_t)(op19)) +#define vlsseg8e16_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlsseg8e16_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vint16mf4_t *)(op6), (vint16mf4_t *)(op7), (const int16_t *)(op8), (ptrdiff_t)(op9), (size_t)(op10)) +#define vlsseg8e16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlsseg8e16_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vint16mf4_t *)(op6), (vint16mf4_t *)(op7), (vbool64_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (vint16mf4_t)(op11), (vint16mf4_t)(op12), (vint16mf4_t)(op13), (vint16mf4_t)(op14), (vint16mf4_t)(op15), (vint16mf4_t)(op16), (const int16_t *)(op17), (ptrdiff_t)(op18), (size_t)(op19)) +#define vlsseg8e16_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlsseg8e16_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vuint16m1_t *)(op6), (vuint16m1_t *)(op7), (const uint16_t *)(op8), (ptrdiff_t)(op9), (size_t)(op10)) +#define vlsseg8e16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlsseg8e16_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vuint16m1_t *)(op6), (vuint16m1_t *)(op7), (vbool16_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (vuint16m1_t)(op11), (vuint16m1_t)(op12), (vuint16m1_t)(op13), (vuint16m1_t)(op14), (vuint16m1_t)(op15), (vuint16m1_t)(op16), (const uint16_t *)(op17), (ptrdiff_t)(op18), (size_t)(op19)) +#define vlsseg8e16_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlsseg8e16_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vuint16mf2_t *)(op6), (vuint16mf2_t *)(op7), (const uint16_t *)(op8), (ptrdiff_t)(op9), (size_t)(op10)) +#define vlsseg8e16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlsseg8e16_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vuint16mf2_t *)(op6), (vuint16mf2_t *)(op7), (vbool32_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (vuint16mf2_t)(op11), (vuint16mf2_t)(op12), (vuint16mf2_t)(op13), (vuint16mf2_t)(op14), (vuint16mf2_t)(op15), (vuint16mf2_t)(op16), (const uint16_t *)(op17), (ptrdiff_t)(op18), (size_t)(op19)) +#define vlsseg8e16_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlsseg8e16_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vuint16mf4_t *)(op6), (vuint16mf4_t *)(op7), (const uint16_t *)(op8), (ptrdiff_t)(op9), (size_t)(op10)) +#define vlsseg8e16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlsseg8e16_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vuint16mf4_t *)(op6), (vuint16mf4_t *)(op7), (vbool64_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (vuint16mf4_t)(op11), (vuint16mf4_t)(op12), (vuint16mf4_t)(op13), (vuint16mf4_t)(op14), (vuint16mf4_t)(op15), (vuint16mf4_t)(op16), (const uint16_t *)(op17), (ptrdiff_t)(op18), (size_t)(op19)) +#define vlsseg2e32_v_i32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e32_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (const int32_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e32_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vbool32_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (const int32_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg2e32_v_i32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e32_v_i32m2((vint32m2_t *)(op0), (vint32m2_t *)(op1), (const int32_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e32_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e32_v_i32m2_m((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vbool16_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (const int32_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg2e32_v_i32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e32_v_i32m4((vint32m4_t *)(op0), (vint32m4_t *)(op1), (const int32_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e32_v_i32m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e32_v_i32m4_m((vint32m4_t *)(op0), (vint32m4_t *)(op1), (vbool8_t)(op2), (vint32m4_t)(op3), (vint32m4_t)(op4), (const int32_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg2e32_v_i32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e32_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (const int32_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e32_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vbool64_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (const int32_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg2e32_v_u32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e32_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (const uint32_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e32_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vbool32_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (const uint32_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg2e32_v_u32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e32_v_u32m2((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (const uint32_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e32_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e32_v_u32m2_m((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vbool16_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (const uint32_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg2e32_v_u32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e32_v_u32m4((vuint32m4_t *)(op0), (vuint32m4_t *)(op1), (const uint32_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e32_v_u32m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e32_v_u32m4_m((vuint32m4_t *)(op0), (vuint32m4_t *)(op1), (vbool8_t)(op2), (vuint32m4_t)(op3), (vuint32m4_t)(op4), (const uint32_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg2e32_v_u32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e32_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (const uint32_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e32_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vbool64_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (const uint32_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg3e32_v_i32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlsseg3e32_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (const int32_t *)(op3), (ptrdiff_t)(op4), (size_t)(op5)) +#define vlsseg3e32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg3e32_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vbool32_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (const int32_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg3e32_v_i32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlsseg3e32_v_i32m2((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vint32m2_t *)(op2), (const int32_t *)(op3), (ptrdiff_t)(op4), (size_t)(op5)) +#define vlsseg3e32_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg3e32_v_i32m2_m((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vint32m2_t *)(op2), (vbool16_t)(op3), (vint32m2_t)(op4), (vint32m2_t)(op5), (vint32m2_t)(op6), (const int32_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg3e32_v_i32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlsseg3e32_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (const int32_t *)(op3), (ptrdiff_t)(op4), (size_t)(op5)) +#define vlsseg3e32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg3e32_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vbool64_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (const int32_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg3e32_v_u32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlsseg3e32_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (const uint32_t *)(op3), (ptrdiff_t)(op4), (size_t)(op5)) +#define vlsseg3e32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg3e32_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vbool32_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (const uint32_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg3e32_v_u32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlsseg3e32_v_u32m2((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vuint32m2_t *)(op2), (const uint32_t *)(op3), (ptrdiff_t)(op4), (size_t)(op5)) +#define vlsseg3e32_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg3e32_v_u32m2_m((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vuint32m2_t *)(op2), (vbool16_t)(op3), (vuint32m2_t)(op4), (vuint32m2_t)(op5), (vuint32m2_t)(op6), (const uint32_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg3e32_v_u32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlsseg3e32_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (const uint32_t *)(op3), (ptrdiff_t)(op4), (size_t)(op5)) +#define vlsseg3e32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg3e32_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vbool64_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (const uint32_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg4e32_v_i32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlsseg4e32_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (const int32_t *)(op4), (ptrdiff_t)(op5), (size_t)(op6)) +#define vlsseg4e32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlsseg4e32_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vbool32_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (const int32_t *)(op9), (ptrdiff_t)(op10), (size_t)(op11)) +#define vlsseg4e32_v_i32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlsseg4e32_v_i32m2((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vint32m2_t *)(op2), (vint32m2_t *)(op3), (const int32_t *)(op4), (ptrdiff_t)(op5), (size_t)(op6)) +#define vlsseg4e32_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlsseg4e32_v_i32m2_m((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vint32m2_t *)(op2), (vint32m2_t *)(op3), (vbool16_t)(op4), (vint32m2_t)(op5), (vint32m2_t)(op6), (vint32m2_t)(op7), (vint32m2_t)(op8), (const int32_t *)(op9), (ptrdiff_t)(op10), (size_t)(op11)) +#define vlsseg4e32_v_i32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlsseg4e32_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (const int32_t *)(op4), (ptrdiff_t)(op5), (size_t)(op6)) +#define vlsseg4e32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlsseg4e32_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vbool64_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (const int32_t *)(op9), (ptrdiff_t)(op10), (size_t)(op11)) +#define vlsseg4e32_v_u32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlsseg4e32_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (const uint32_t *)(op4), (ptrdiff_t)(op5), (size_t)(op6)) +#define vlsseg4e32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlsseg4e32_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vbool32_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (const uint32_t *)(op9), (ptrdiff_t)(op10), (size_t)(op11)) +#define vlsseg4e32_v_u32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlsseg4e32_v_u32m2((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vuint32m2_t *)(op2), (vuint32m2_t *)(op3), (const uint32_t *)(op4), (ptrdiff_t)(op5), (size_t)(op6)) +#define vlsseg4e32_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlsseg4e32_v_u32m2_m((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vuint32m2_t *)(op2), (vuint32m2_t *)(op3), (vbool16_t)(op4), (vuint32m2_t)(op5), (vuint32m2_t)(op6), (vuint32m2_t)(op7), (vuint32m2_t)(op8), (const uint32_t *)(op9), (ptrdiff_t)(op10), (size_t)(op11)) +#define vlsseg4e32_v_u32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlsseg4e32_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (const uint32_t *)(op4), (ptrdiff_t)(op5), (size_t)(op6)) +#define vlsseg4e32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlsseg4e32_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vbool64_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (const uint32_t *)(op9), (ptrdiff_t)(op10), (size_t)(op11)) +#define vlsseg5e32_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg5e32_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (const int32_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg5e32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlsseg5e32_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vbool32_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (const int32_t *)(op11), (ptrdiff_t)(op12), (size_t)(op13)) +#define vlsseg5e32_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg5e32_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (const int32_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg5e32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlsseg5e32_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vbool64_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (const int32_t *)(op11), (ptrdiff_t)(op12), (size_t)(op13)) +#define vlsseg5e32_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg5e32_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (const uint32_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg5e32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlsseg5e32_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vbool32_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (const uint32_t *)(op11), (ptrdiff_t)(op12), (size_t)(op13)) +#define vlsseg5e32_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg5e32_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (const uint32_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg5e32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlsseg5e32_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vbool64_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (const uint32_t *)(op11), (ptrdiff_t)(op12), (size_t)(op13)) +#define vlsseg6e32_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlsseg6e32_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (const int32_t *)(op6), (ptrdiff_t)(op7), (size_t)(op8)) +#define vlsseg6e32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlsseg6e32_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vbool32_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (vint32m1_t)(op11), (vint32m1_t)(op12), (const int32_t *)(op13), (ptrdiff_t)(op14), (size_t)(op15)) +#define vlsseg6e32_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlsseg6e32_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (const int32_t *)(op6), (ptrdiff_t)(op7), (size_t)(op8)) +#define vlsseg6e32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlsseg6e32_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vbool64_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (vint32mf2_t)(op11), (vint32mf2_t)(op12), (const int32_t *)(op13), (ptrdiff_t)(op14), (size_t)(op15)) +#define vlsseg6e32_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlsseg6e32_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (const uint32_t *)(op6), (ptrdiff_t)(op7), (size_t)(op8)) +#define vlsseg6e32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlsseg6e32_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vbool32_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (vuint32m1_t)(op11), (vuint32m1_t)(op12), (const uint32_t *)(op13), (ptrdiff_t)(op14), (size_t)(op15)) +#define vlsseg6e32_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlsseg6e32_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (const uint32_t *)(op6), (ptrdiff_t)(op7), (size_t)(op8)) +#define vlsseg6e32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlsseg6e32_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vbool64_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (vuint32mf2_t)(op11), (vuint32mf2_t)(op12), (const uint32_t *)(op13), (ptrdiff_t)(op14), (size_t)(op15)) +#define vlsseg7e32_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg7e32_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vint32m1_t *)(op6), (const int32_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg7e32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlsseg7e32_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vint32m1_t *)(op6), (vbool32_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (vint32m1_t)(op11), (vint32m1_t)(op12), (vint32m1_t)(op13), (vint32m1_t)(op14), (const int32_t *)(op15), (ptrdiff_t)(op16), (size_t)(op17)) +#define vlsseg7e32_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg7e32_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vint32mf2_t *)(op6), (const int32_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg7e32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlsseg7e32_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vint32mf2_t *)(op6), (vbool64_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (vint32mf2_t)(op11), (vint32mf2_t)(op12), (vint32mf2_t)(op13), (vint32mf2_t)(op14), (const int32_t *)(op15), (ptrdiff_t)(op16), (size_t)(op17)) +#define vlsseg7e32_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg7e32_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vuint32m1_t *)(op6), (const uint32_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg7e32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlsseg7e32_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vuint32m1_t *)(op6), (vbool32_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (vuint32m1_t)(op11), (vuint32m1_t)(op12), (vuint32m1_t)(op13), (vuint32m1_t)(op14), (const uint32_t *)(op15), (ptrdiff_t)(op16), (size_t)(op17)) +#define vlsseg7e32_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg7e32_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vuint32mf2_t *)(op6), (const uint32_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg7e32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlsseg7e32_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vuint32mf2_t *)(op6), (vbool64_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (vuint32mf2_t)(op11), (vuint32mf2_t)(op12), (vuint32mf2_t)(op13), (vuint32mf2_t)(op14), (const uint32_t *)(op15), (ptrdiff_t)(op16), (size_t)(op17)) +#define vlsseg8e32_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlsseg8e32_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vint32m1_t *)(op6), (vint32m1_t *)(op7), (const int32_t *)(op8), (ptrdiff_t)(op9), (size_t)(op10)) +#define vlsseg8e32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlsseg8e32_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vint32m1_t *)(op6), (vint32m1_t *)(op7), (vbool32_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (vint32m1_t)(op11), (vint32m1_t)(op12), (vint32m1_t)(op13), (vint32m1_t)(op14), (vint32m1_t)(op15), (vint32m1_t)(op16), (const int32_t *)(op17), (ptrdiff_t)(op18), (size_t)(op19)) +#define vlsseg8e32_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlsseg8e32_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vint32mf2_t *)(op6), (vint32mf2_t *)(op7), (const int32_t *)(op8), (ptrdiff_t)(op9), (size_t)(op10)) +#define vlsseg8e32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlsseg8e32_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vint32mf2_t *)(op6), (vint32mf2_t *)(op7), (vbool64_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (vint32mf2_t)(op11), (vint32mf2_t)(op12), (vint32mf2_t)(op13), (vint32mf2_t)(op14), (vint32mf2_t)(op15), (vint32mf2_t)(op16), (const int32_t *)(op17), (ptrdiff_t)(op18), (size_t)(op19)) +#define vlsseg8e32_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlsseg8e32_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vuint32m1_t *)(op6), (vuint32m1_t *)(op7), (const uint32_t *)(op8), (ptrdiff_t)(op9), (size_t)(op10)) +#define vlsseg8e32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlsseg8e32_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vuint32m1_t *)(op6), (vuint32m1_t *)(op7), (vbool32_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (vuint32m1_t)(op11), (vuint32m1_t)(op12), (vuint32m1_t)(op13), (vuint32m1_t)(op14), (vuint32m1_t)(op15), (vuint32m1_t)(op16), (const uint32_t *)(op17), (ptrdiff_t)(op18), (size_t)(op19)) +#define vlsseg8e32_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlsseg8e32_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vuint32mf2_t *)(op6), (vuint32mf2_t *)(op7), (const uint32_t *)(op8), (ptrdiff_t)(op9), (size_t)(op10)) +#define vlsseg8e32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlsseg8e32_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vuint32mf2_t *)(op6), (vuint32mf2_t *)(op7), (vbool64_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (vuint32mf2_t)(op11), (vuint32mf2_t)(op12), (vuint32mf2_t)(op13), (vuint32mf2_t)(op14), (vuint32mf2_t)(op15), (vuint32mf2_t)(op16), (const uint32_t *)(op17), (ptrdiff_t)(op18), (size_t)(op19)) +#define vlsseg2e64_v_i64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e64_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (const int64_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e64_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vbool64_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (const int64_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg2e64_v_i64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e64_v_i64m2((vint64m2_t *)(op0), (vint64m2_t *)(op1), (const int64_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e64_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e64_v_i64m2_m((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vbool32_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (const int64_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg2e64_v_i64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e64_v_i64m4((vint64m4_t *)(op0), (vint64m4_t *)(op1), (const int64_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e64_v_i64m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e64_v_i64m4_m((vint64m4_t *)(op0), (vint64m4_t *)(op1), (vbool16_t)(op2), (vint64m4_t)(op3), (vint64m4_t)(op4), (const int64_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg2e64_v_u64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e64_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (const uint64_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e64_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vbool64_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (const uint64_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg2e64_v_u64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e64_v_u64m2((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (const uint64_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e64_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e64_v_u64m2_m((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vbool32_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (const uint64_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg2e64_v_u64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e64_v_u64m4((vuint64m4_t *)(op0), (vuint64m4_t *)(op1), (const uint64_t *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e64_v_u64m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e64_v_u64m4_m((vuint64m4_t *)(op0), (vuint64m4_t *)(op1), (vbool16_t)(op2), (vuint64m4_t)(op3), (vuint64m4_t)(op4), (const uint64_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg3e64_v_i64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlsseg3e64_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (const int64_t *)(op3), (ptrdiff_t)(op4), (size_t)(op5)) +#define vlsseg3e64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg3e64_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vbool64_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (const int64_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg3e64_v_i64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlsseg3e64_v_i64m2((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vint64m2_t *)(op2), (const int64_t *)(op3), (ptrdiff_t)(op4), (size_t)(op5)) +#define vlsseg3e64_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg3e64_v_i64m2_m((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vint64m2_t *)(op2), (vbool32_t)(op3), (vint64m2_t)(op4), (vint64m2_t)(op5), (vint64m2_t)(op6), (const int64_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg3e64_v_u64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlsseg3e64_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (const uint64_t *)(op3), (ptrdiff_t)(op4), (size_t)(op5)) +#define vlsseg3e64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg3e64_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vbool64_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (const uint64_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg3e64_v_u64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlsseg3e64_v_u64m2((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vuint64m2_t *)(op2), (const uint64_t *)(op3), (ptrdiff_t)(op4), (size_t)(op5)) +#define vlsseg3e64_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg3e64_v_u64m2_m((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vuint64m2_t *)(op2), (vbool32_t)(op3), (vuint64m2_t)(op4), (vuint64m2_t)(op5), (vuint64m2_t)(op6), (const uint64_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg4e64_v_i64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlsseg4e64_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (const int64_t *)(op4), (ptrdiff_t)(op5), (size_t)(op6)) +#define vlsseg4e64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlsseg4e64_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vbool64_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (const int64_t *)(op9), (ptrdiff_t)(op10), (size_t)(op11)) +#define vlsseg4e64_v_i64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlsseg4e64_v_i64m2((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vint64m2_t *)(op2), (vint64m2_t *)(op3), (const int64_t *)(op4), (ptrdiff_t)(op5), (size_t)(op6)) +#define vlsseg4e64_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlsseg4e64_v_i64m2_m((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vint64m2_t *)(op2), (vint64m2_t *)(op3), (vbool32_t)(op4), (vint64m2_t)(op5), (vint64m2_t)(op6), (vint64m2_t)(op7), (vint64m2_t)(op8), (const int64_t *)(op9), (ptrdiff_t)(op10), (size_t)(op11)) +#define vlsseg4e64_v_u64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlsseg4e64_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (const uint64_t *)(op4), (ptrdiff_t)(op5), (size_t)(op6)) +#define vlsseg4e64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlsseg4e64_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vbool64_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (const uint64_t *)(op9), (ptrdiff_t)(op10), (size_t)(op11)) +#define vlsseg4e64_v_u64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlsseg4e64_v_u64m2((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vuint64m2_t *)(op2), (vuint64m2_t *)(op3), (const uint64_t *)(op4), (ptrdiff_t)(op5), (size_t)(op6)) +#define vlsseg4e64_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlsseg4e64_v_u64m2_m((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vuint64m2_t *)(op2), (vuint64m2_t *)(op3), (vbool32_t)(op4), (vuint64m2_t)(op5), (vuint64m2_t)(op6), (vuint64m2_t)(op7), (vuint64m2_t)(op8), (const uint64_t *)(op9), (ptrdiff_t)(op10), (size_t)(op11)) +#define vlsseg5e64_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg5e64_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (const int64_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg5e64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlsseg5e64_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vbool64_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (const int64_t *)(op11), (ptrdiff_t)(op12), (size_t)(op13)) +#define vlsseg5e64_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg5e64_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (const uint64_t *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg5e64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlsseg5e64_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vbool64_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (const uint64_t *)(op11), (ptrdiff_t)(op12), (size_t)(op13)) +#define vlsseg6e64_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlsseg6e64_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (const int64_t *)(op6), (ptrdiff_t)(op7), (size_t)(op8)) +#define vlsseg6e64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlsseg6e64_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vbool64_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (vint64m1_t)(op11), (vint64m1_t)(op12), (const int64_t *)(op13), (ptrdiff_t)(op14), (size_t)(op15)) +#define vlsseg6e64_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlsseg6e64_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (const uint64_t *)(op6), (ptrdiff_t)(op7), (size_t)(op8)) +#define vlsseg6e64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlsseg6e64_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vbool64_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (vuint64m1_t)(op11), (vuint64m1_t)(op12), (const uint64_t *)(op13), (ptrdiff_t)(op14), (size_t)(op15)) +#define vlsseg7e64_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg7e64_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vint64m1_t *)(op6), (const int64_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg7e64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlsseg7e64_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vint64m1_t *)(op6), (vbool64_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (vint64m1_t)(op11), (vint64m1_t)(op12), (vint64m1_t)(op13), (vint64m1_t)(op14), (const int64_t *)(op15), (ptrdiff_t)(op16), (size_t)(op17)) +#define vlsseg7e64_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg7e64_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vuint64m1_t *)(op6), (const uint64_t *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg7e64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlsseg7e64_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vuint64m1_t *)(op6), (vbool64_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (vuint64m1_t)(op11), (vuint64m1_t)(op12), (vuint64m1_t)(op13), (vuint64m1_t)(op14), (const uint64_t *)(op15), (ptrdiff_t)(op16), (size_t)(op17)) +#define vlsseg8e64_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlsseg8e64_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vint64m1_t *)(op6), (vint64m1_t *)(op7), (const int64_t *)(op8), (ptrdiff_t)(op9), (size_t)(op10)) +#define vlsseg8e64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlsseg8e64_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vint64m1_t *)(op6), (vint64m1_t *)(op7), (vbool64_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (vint64m1_t)(op11), (vint64m1_t)(op12), (vint64m1_t)(op13), (vint64m1_t)(op14), (vint64m1_t)(op15), (vint64m1_t)(op16), (const int64_t *)(op17), (ptrdiff_t)(op18), (size_t)(op19)) +#define vlsseg8e64_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlsseg8e64_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vuint64m1_t *)(op6), (vuint64m1_t *)(op7), (const uint64_t *)(op8), (ptrdiff_t)(op9), (size_t)(op10)) +#define vlsseg8e64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlsseg8e64_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vuint64m1_t *)(op6), (vuint64m1_t *)(op7), (vbool64_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (vuint64m1_t)(op11), (vuint64m1_t)(op12), (vuint64m1_t)(op13), (vuint64m1_t)(op14), (vuint64m1_t)(op15), (vuint64m1_t)(op16), (const uint64_t *)(op17), (ptrdiff_t)(op18), (size_t)(op19)) +#define vluxseg3ei8_v_i8m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei8_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (const int8_t *)(op3), (vuint8m1_t)(op4), (size_t)(op5)) +#define vluxseg3ei8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei8_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vbool8_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (const int8_t *)(op7), (vuint8m1_t)(op8), (size_t)(op9)) +#define vluxseg3ei8_v_i8m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei8_v_i8m2((vint8m2_t *)(op0), (vint8m2_t *)(op1), (vint8m2_t *)(op2), (const int8_t *)(op3), (vuint8m2_t)(op4), (size_t)(op5)) +#define vluxseg3ei8_v_i8m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei8_v_i8m2_m((vint8m2_t *)(op0), (vint8m2_t *)(op1), (vint8m2_t *)(op2), (vbool4_t)(op3), (vint8m2_t)(op4), (vint8m2_t)(op5), (vint8m2_t)(op6), (const int8_t *)(op7), (vuint8m2_t)(op8), (size_t)(op9)) +#define vluxseg3ei8_v_i8mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei8_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (const int8_t *)(op3), (vuint8mf2_t)(op4), (size_t)(op5)) +#define vluxseg3ei8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei8_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vbool16_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (const int8_t *)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vluxseg3ei8_v_i8mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei8_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (const int8_t *)(op3), (vuint8mf4_t)(op4), (size_t)(op5)) +#define vluxseg3ei8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei8_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vbool32_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (const int8_t *)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vluxseg3ei8_v_i8mf8(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei8_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (const int8_t *)(op3), (vuint8mf8_t)(op4), (size_t)(op5)) +#define vluxseg3ei8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei8_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vbool64_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (const int8_t *)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vluxseg3ei8_v_u8m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei8_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (const uint8_t *)(op3), (vuint8m1_t)(op4), (size_t)(op5)) +#define vluxseg3ei8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei8_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vbool8_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (const uint8_t *)(op7), (vuint8m1_t)(op8), (size_t)(op9)) +#define vluxseg3ei8_v_u8m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei8_v_u8m2((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (vuint8m2_t *)(op2), (const uint8_t *)(op3), (vuint8m2_t)(op4), (size_t)(op5)) +#define vluxseg3ei8_v_u8m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei8_v_u8m2_m((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (vuint8m2_t *)(op2), (vbool4_t)(op3), (vuint8m2_t)(op4), (vuint8m2_t)(op5), (vuint8m2_t)(op6), (const uint8_t *)(op7), (vuint8m2_t)(op8), (size_t)(op9)) +#define vluxseg3ei8_v_u8mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei8_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (const uint8_t *)(op3), (vuint8mf2_t)(op4), (size_t)(op5)) +#define vluxseg3ei8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei8_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vbool16_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (const uint8_t *)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vluxseg3ei8_v_u8mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei8_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (const uint8_t *)(op3), (vuint8mf4_t)(op4), (size_t)(op5)) +#define vluxseg3ei8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei8_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vbool32_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (const uint8_t *)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vluxseg3ei8_v_u8mf8(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei8_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (const uint8_t *)(op3), (vuint8mf8_t)(op4), (size_t)(op5)) +#define vluxseg3ei8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei8_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vbool64_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (const uint8_t *)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vluxseg4ei8_v_i8m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei8_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (const int8_t *)(op4), (vuint8m1_t)(op5), (size_t)(op6)) +#define vluxseg4ei8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei8_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vbool8_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (const int8_t *)(op9), (vuint8m1_t)(op10), (size_t)(op11)) +#define vluxseg4ei8_v_i8m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei8_v_i8m2((vint8m2_t *)(op0), (vint8m2_t *)(op1), (vint8m2_t *)(op2), (vint8m2_t *)(op3), (const int8_t *)(op4), (vuint8m2_t)(op5), (size_t)(op6)) +#define vluxseg4ei8_v_i8m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei8_v_i8m2_m((vint8m2_t *)(op0), (vint8m2_t *)(op1), (vint8m2_t *)(op2), (vint8m2_t *)(op3), (vbool4_t)(op4), (vint8m2_t)(op5), (vint8m2_t)(op6), (vint8m2_t)(op7), (vint8m2_t)(op8), (const int8_t *)(op9), (vuint8m2_t)(op10), (size_t)(op11)) +#define vluxseg4ei8_v_i8mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei8_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (const int8_t *)(op4), (vuint8mf2_t)(op5), (size_t)(op6)) +#define vluxseg4ei8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei8_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vbool16_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (const int8_t *)(op9), (vuint8mf2_t)(op10), (size_t)(op11)) +#define vluxseg4ei8_v_i8mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei8_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (const int8_t *)(op4), (vuint8mf4_t)(op5), (size_t)(op6)) +#define vluxseg4ei8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei8_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vbool32_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (const int8_t *)(op9), (vuint8mf4_t)(op10), (size_t)(op11)) +#define vluxseg4ei8_v_i8mf8(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei8_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (const int8_t *)(op4), (vuint8mf8_t)(op5), (size_t)(op6)) +#define vluxseg4ei8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei8_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vbool64_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (const int8_t *)(op9), (vuint8mf8_t)(op10), (size_t)(op11)) +#define vluxseg4ei8_v_u8m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei8_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (const uint8_t *)(op4), (vuint8m1_t)(op5), (size_t)(op6)) +#define vluxseg4ei8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei8_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vbool8_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (const uint8_t *)(op9), (vuint8m1_t)(op10), (size_t)(op11)) +#define vluxseg4ei8_v_u8m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei8_v_u8m2((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (vuint8m2_t *)(op2), (vuint8m2_t *)(op3), (const uint8_t *)(op4), (vuint8m2_t)(op5), (size_t)(op6)) +#define vluxseg4ei8_v_u8m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei8_v_u8m2_m((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (vuint8m2_t *)(op2), (vuint8m2_t *)(op3), (vbool4_t)(op4), (vuint8m2_t)(op5), (vuint8m2_t)(op6), (vuint8m2_t)(op7), (vuint8m2_t)(op8), (const uint8_t *)(op9), (vuint8m2_t)(op10), (size_t)(op11)) +#define vluxseg4ei8_v_u8mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei8_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (const uint8_t *)(op4), (vuint8mf2_t)(op5), (size_t)(op6)) +#define vluxseg4ei8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei8_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vbool16_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (const uint8_t *)(op9), (vuint8mf2_t)(op10), (size_t)(op11)) +#define vluxseg4ei8_v_u8mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei8_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (const uint8_t *)(op4), (vuint8mf4_t)(op5), (size_t)(op6)) +#define vluxseg4ei8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei8_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vbool32_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (const uint8_t *)(op9), (vuint8mf4_t)(op10), (size_t)(op11)) +#define vluxseg4ei8_v_u8mf8(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei8_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (const uint8_t *)(op4), (vuint8mf8_t)(op5), (size_t)(op6)) +#define vluxseg4ei8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei8_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vbool64_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (const uint8_t *)(op9), (vuint8mf8_t)(op10), (size_t)(op11)) +#define vluxseg5ei8_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei8_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (const int8_t *)(op5), (vuint8m1_t)(op6), (size_t)(op7)) +#define vluxseg5ei8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei8_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vbool8_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (const int8_t *)(op11), (vuint8m1_t)(op12), (size_t)(op13)) +#define vluxseg5ei8_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei8_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (const int8_t *)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vluxseg5ei8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei8_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vbool16_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (const int8_t *)(op11), (vuint8mf2_t)(op12), (size_t)(op13)) +#define vluxseg5ei8_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei8_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (const int8_t *)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vluxseg5ei8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei8_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vbool32_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (const int8_t *)(op11), (vuint8mf4_t)(op12), (size_t)(op13)) +#define vluxseg5ei8_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei8_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (const int8_t *)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vluxseg5ei8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei8_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vbool64_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (const int8_t *)(op11), (vuint8mf8_t)(op12), (size_t)(op13)) +#define vluxseg5ei8_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei8_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (const uint8_t *)(op5), (vuint8m1_t)(op6), (size_t)(op7)) +#define vluxseg5ei8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei8_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vbool8_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (const uint8_t *)(op11), (vuint8m1_t)(op12), (size_t)(op13)) +#define vluxseg5ei8_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei8_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (const uint8_t *)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vluxseg5ei8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei8_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vbool16_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (const uint8_t *)(op11), (vuint8mf2_t)(op12), (size_t)(op13)) +#define vluxseg5ei8_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei8_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (const uint8_t *)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vluxseg5ei8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei8_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vbool32_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (const uint8_t *)(op11), (vuint8mf4_t)(op12), (size_t)(op13)) +#define vluxseg5ei8_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei8_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (const uint8_t *)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vluxseg5ei8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei8_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vbool64_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (const uint8_t *)(op11), (vuint8mf8_t)(op12), (size_t)(op13)) +#define vluxseg6ei8_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei8_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (const int8_t *)(op6), (vuint8m1_t)(op7), (size_t)(op8)) +#define vluxseg6ei8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei8_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vbool8_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (vint8m1_t)(op11), (vint8m1_t)(op12), (const int8_t *)(op13), (vuint8m1_t)(op14), (size_t)(op15)) +#define vluxseg6ei8_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei8_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (const int8_t *)(op6), (vuint8mf2_t)(op7), (size_t)(op8)) +#define vluxseg6ei8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei8_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vbool16_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (vint8mf2_t)(op11), (vint8mf2_t)(op12), (const int8_t *)(op13), (vuint8mf2_t)(op14), (size_t)(op15)) +#define vluxseg6ei8_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei8_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (const int8_t *)(op6), (vuint8mf4_t)(op7), (size_t)(op8)) +#define vluxseg6ei8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei8_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vbool32_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (vint8mf4_t)(op11), (vint8mf4_t)(op12), (const int8_t *)(op13), (vuint8mf4_t)(op14), (size_t)(op15)) +#define vluxseg6ei8_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei8_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (const int8_t *)(op6), (vuint8mf8_t)(op7), (size_t)(op8)) +#define vluxseg6ei8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei8_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vbool64_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (vint8mf8_t)(op11), (vint8mf8_t)(op12), (const int8_t *)(op13), (vuint8mf8_t)(op14), (size_t)(op15)) +#define vluxseg6ei8_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei8_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (const uint8_t *)(op6), (vuint8m1_t)(op7), (size_t)(op8)) +#define vluxseg6ei8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei8_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vbool8_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (vuint8m1_t)(op11), (vuint8m1_t)(op12), (const uint8_t *)(op13), (vuint8m1_t)(op14), (size_t)(op15)) +#define vluxseg6ei8_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei8_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (const uint8_t *)(op6), (vuint8mf2_t)(op7), (size_t)(op8)) +#define vluxseg6ei8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei8_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vbool16_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (vuint8mf2_t)(op11), (vuint8mf2_t)(op12), (const uint8_t *)(op13), (vuint8mf2_t)(op14), (size_t)(op15)) +#define vluxseg6ei8_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei8_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (const uint8_t *)(op6), (vuint8mf4_t)(op7), (size_t)(op8)) +#define vluxseg6ei8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei8_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vbool32_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (vuint8mf4_t)(op11), (vuint8mf4_t)(op12), (const uint8_t *)(op13), (vuint8mf4_t)(op14), (size_t)(op15)) +#define vluxseg6ei8_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei8_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (const uint8_t *)(op6), (vuint8mf8_t)(op7), (size_t)(op8)) +#define vluxseg6ei8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei8_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vbool64_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (vuint8mf8_t)(op11), (vuint8mf8_t)(op12), (const uint8_t *)(op13), (vuint8mf8_t)(op14), (size_t)(op15)) +#define vluxseg7ei8_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei8_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vint8m1_t *)(op6), (const int8_t *)(op7), (vuint8m1_t)(op8), (size_t)(op9)) +#define vluxseg7ei8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei8_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vint8m1_t *)(op6), (vbool8_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (vint8m1_t)(op11), (vint8m1_t)(op12), (vint8m1_t)(op13), (vint8m1_t)(op14), (const int8_t *)(op15), (vuint8m1_t)(op16), (size_t)(op17)) +#define vluxseg7ei8_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei8_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vint8mf2_t *)(op6), (const int8_t *)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vluxseg7ei8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei8_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vint8mf2_t *)(op6), (vbool16_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (vint8mf2_t)(op11), (vint8mf2_t)(op12), (vint8mf2_t)(op13), (vint8mf2_t)(op14), (const int8_t *)(op15), (vuint8mf2_t)(op16), (size_t)(op17)) +#define vluxseg7ei8_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei8_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vint8mf4_t *)(op6), (const int8_t *)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vluxseg7ei8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei8_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vint8mf4_t *)(op6), (vbool32_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (vint8mf4_t)(op11), (vint8mf4_t)(op12), (vint8mf4_t)(op13), (vint8mf4_t)(op14), (const int8_t *)(op15), (vuint8mf4_t)(op16), (size_t)(op17)) +#define vluxseg7ei8_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei8_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vint8mf8_t *)(op6), (const int8_t *)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vluxseg7ei8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei8_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vint8mf8_t *)(op6), (vbool64_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (vint8mf8_t)(op11), (vint8mf8_t)(op12), (vint8mf8_t)(op13), (vint8mf8_t)(op14), (const int8_t *)(op15), (vuint8mf8_t)(op16), (size_t)(op17)) +#define vluxseg7ei8_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei8_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vuint8m1_t *)(op6), (const uint8_t *)(op7), (vuint8m1_t)(op8), (size_t)(op9)) +#define vluxseg7ei8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei8_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vuint8m1_t *)(op6), (vbool8_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (vuint8m1_t)(op11), (vuint8m1_t)(op12), (vuint8m1_t)(op13), (vuint8m1_t)(op14), (const uint8_t *)(op15), (vuint8m1_t)(op16), (size_t)(op17)) +#define vluxseg7ei8_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei8_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vuint8mf2_t *)(op6), (const uint8_t *)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vluxseg7ei8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei8_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vuint8mf2_t *)(op6), (vbool16_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (vuint8mf2_t)(op11), (vuint8mf2_t)(op12), (vuint8mf2_t)(op13), (vuint8mf2_t)(op14), (const uint8_t *)(op15), (vuint8mf2_t)(op16), (size_t)(op17)) +#define vluxseg7ei8_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei8_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vuint8mf4_t *)(op6), (const uint8_t *)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vluxseg7ei8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei8_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vuint8mf4_t *)(op6), (vbool32_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (vuint8mf4_t)(op11), (vuint8mf4_t)(op12), (vuint8mf4_t)(op13), (vuint8mf4_t)(op14), (const uint8_t *)(op15), (vuint8mf4_t)(op16), (size_t)(op17)) +#define vluxseg7ei8_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei8_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vuint8mf8_t *)(op6), (const uint8_t *)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vluxseg7ei8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei8_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vuint8mf8_t *)(op6), (vbool64_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (vuint8mf8_t)(op11), (vuint8mf8_t)(op12), (vuint8mf8_t)(op13), (vuint8mf8_t)(op14), (const uint8_t *)(op15), (vuint8mf8_t)(op16), (size_t)(op17)) +#define vluxseg8ei8_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei8_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vint8m1_t *)(op6), (vint8m1_t *)(op7), (const int8_t *)(op8), (vuint8m1_t)(op9), (size_t)(op10)) +#define vluxseg8ei8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei8_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vint8m1_t *)(op6), (vint8m1_t *)(op7), (vbool8_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (vint8m1_t)(op11), (vint8m1_t)(op12), (vint8m1_t)(op13), (vint8m1_t)(op14), (vint8m1_t)(op15), (vint8m1_t)(op16), (const int8_t *)(op17), (vuint8m1_t)(op18), (size_t)(op19)) +#define vluxseg8ei8_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei8_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vint8mf2_t *)(op6), (vint8mf2_t *)(op7), (const int8_t *)(op8), (vuint8mf2_t)(op9), (size_t)(op10)) +#define vluxseg8ei8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei8_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vint8mf2_t *)(op6), (vint8mf2_t *)(op7), (vbool16_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (vint8mf2_t)(op11), (vint8mf2_t)(op12), (vint8mf2_t)(op13), (vint8mf2_t)(op14), (vint8mf2_t)(op15), (vint8mf2_t)(op16), (const int8_t *)(op17), (vuint8mf2_t)(op18), (size_t)(op19)) +#define vluxseg8ei8_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei8_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vint8mf4_t *)(op6), (vint8mf4_t *)(op7), (const int8_t *)(op8), (vuint8mf4_t)(op9), (size_t)(op10)) +#define vluxseg8ei8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei8_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vint8mf4_t *)(op6), (vint8mf4_t *)(op7), (vbool32_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (vint8mf4_t)(op11), (vint8mf4_t)(op12), (vint8mf4_t)(op13), (vint8mf4_t)(op14), (vint8mf4_t)(op15), (vint8mf4_t)(op16), (const int8_t *)(op17), (vuint8mf4_t)(op18), (size_t)(op19)) +#define vluxseg8ei8_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei8_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vint8mf8_t *)(op6), (vint8mf8_t *)(op7), (const int8_t *)(op8), (vuint8mf8_t)(op9), (size_t)(op10)) +#define vluxseg8ei8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei8_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vint8mf8_t *)(op6), (vint8mf8_t *)(op7), (vbool64_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (vint8mf8_t)(op11), (vint8mf8_t)(op12), (vint8mf8_t)(op13), (vint8mf8_t)(op14), (vint8mf8_t)(op15), (vint8mf8_t)(op16), (const int8_t *)(op17), (vuint8mf8_t)(op18), (size_t)(op19)) +#define vluxseg8ei8_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei8_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vuint8m1_t *)(op6), (vuint8m1_t *)(op7), (const uint8_t *)(op8), (vuint8m1_t)(op9), (size_t)(op10)) +#define vluxseg8ei8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei8_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vuint8m1_t *)(op6), (vuint8m1_t *)(op7), (vbool8_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (vuint8m1_t)(op11), (vuint8m1_t)(op12), (vuint8m1_t)(op13), (vuint8m1_t)(op14), (vuint8m1_t)(op15), (vuint8m1_t)(op16), (const uint8_t *)(op17), (vuint8m1_t)(op18), (size_t)(op19)) +#define vluxseg8ei8_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei8_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vuint8mf2_t *)(op6), (vuint8mf2_t *)(op7), (const uint8_t *)(op8), (vuint8mf2_t)(op9), (size_t)(op10)) +#define vluxseg8ei8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei8_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vuint8mf2_t *)(op6), (vuint8mf2_t *)(op7), (vbool16_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (vuint8mf2_t)(op11), (vuint8mf2_t)(op12), (vuint8mf2_t)(op13), (vuint8mf2_t)(op14), (vuint8mf2_t)(op15), (vuint8mf2_t)(op16), (const uint8_t *)(op17), (vuint8mf2_t)(op18), (size_t)(op19)) +#define vluxseg8ei8_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei8_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vuint8mf4_t *)(op6), (vuint8mf4_t *)(op7), (const uint8_t *)(op8), (vuint8mf4_t)(op9), (size_t)(op10)) +#define vluxseg8ei8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei8_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vuint8mf4_t *)(op6), (vuint8mf4_t *)(op7), (vbool32_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (vuint8mf4_t)(op11), (vuint8mf4_t)(op12), (vuint8mf4_t)(op13), (vuint8mf4_t)(op14), (vuint8mf4_t)(op15), (vuint8mf4_t)(op16), (const uint8_t *)(op17), (vuint8mf4_t)(op18), (size_t)(op19)) +#define vluxseg8ei8_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei8_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vuint8mf8_t *)(op6), (vuint8mf8_t *)(op7), (const uint8_t *)(op8), (vuint8mf8_t)(op9), (size_t)(op10)) +#define vluxseg8ei8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei8_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vuint8mf8_t *)(op6), (vuint8mf8_t *)(op7), (vbool64_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (vuint8mf8_t)(op11), (vuint8mf8_t)(op12), (vuint8mf8_t)(op13), (vuint8mf8_t)(op14), (vuint8mf8_t)(op15), (vuint8mf8_t)(op16), (const uint8_t *)(op17), (vuint8mf8_t)(op18), (size_t)(op19)) +#define vluxseg2ei16_v_i8m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (const int8_t *)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vbool8_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (const int8_t *)(op5), (vuint16m2_t)(op6), (size_t)(op7)) +#define vluxseg2ei16_v_i8m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_i8m2((vint8m2_t *)(op0), (vint8m2_t *)(op1), (const int8_t *)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_i8m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_i8m2_m((vint8m2_t *)(op0), (vint8m2_t *)(op1), (vbool4_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (const int8_t *)(op5), (vuint16m4_t)(op6), (size_t)(op7)) +#define vluxseg2ei16_v_i8m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_i8m4((vint8m4_t *)(op0), (vint8m4_t *)(op1), (const int8_t *)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_i8m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_i8m4_m((vint8m4_t *)(op0), (vint8m4_t *)(op1), (vbool2_t)(op2), (vint8m4_t)(op3), (vint8m4_t)(op4), (const int8_t *)(op5), (vuint16m8_t)(op6), (size_t)(op7)) +#define vluxseg2ei16_v_i8mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (const int8_t *)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vbool16_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (const int8_t *)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vluxseg2ei16_v_i8mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (const int8_t *)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vbool32_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (const int8_t *)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vluxseg2ei16_v_i8mf8(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (const int8_t *)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vbool64_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (const int8_t *)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vluxseg2ei16_v_u8m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (const uint8_t *)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vbool8_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (const uint8_t *)(op5), (vuint16m2_t)(op6), (size_t)(op7)) +#define vluxseg2ei16_v_u8m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_u8m2((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (const uint8_t *)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_u8m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_u8m2_m((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (vbool4_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (const uint8_t *)(op5), (vuint16m4_t)(op6), (size_t)(op7)) +#define vluxseg2ei16_v_u8m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_u8m4((vuint8m4_t *)(op0), (vuint8m4_t *)(op1), (const uint8_t *)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_u8m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_u8m4_m((vuint8m4_t *)(op0), (vuint8m4_t *)(op1), (vbool2_t)(op2), (vuint8m4_t)(op3), (vuint8m4_t)(op4), (const uint8_t *)(op5), (vuint16m8_t)(op6), (size_t)(op7)) +#define vluxseg2ei16_v_u8mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (const uint8_t *)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vbool16_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (const uint8_t *)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vluxseg2ei16_v_u8mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (const uint8_t *)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vbool32_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (const uint8_t *)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vluxseg2ei16_v_u8mf8(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (const uint8_t *)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vbool64_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (const uint8_t *)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vluxseg3ei16_v_i8m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei16_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (const int8_t *)(op3), (vuint16m2_t)(op4), (size_t)(op5)) +#define vluxseg3ei16_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei16_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vbool8_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (const int8_t *)(op7), (vuint16m2_t)(op8), (size_t)(op9)) +#define vluxseg3ei16_v_i8m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei16_v_i8m2((vint8m2_t *)(op0), (vint8m2_t *)(op1), (vint8m2_t *)(op2), (const int8_t *)(op3), (vuint16m4_t)(op4), (size_t)(op5)) +#define vluxseg3ei16_v_i8m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei16_v_i8m2_m((vint8m2_t *)(op0), (vint8m2_t *)(op1), (vint8m2_t *)(op2), (vbool4_t)(op3), (vint8m2_t)(op4), (vint8m2_t)(op5), (vint8m2_t)(op6), (const int8_t *)(op7), (vuint16m4_t)(op8), (size_t)(op9)) +#define vluxseg3ei16_v_i8mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei16_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (const int8_t *)(op3), (vuint16m1_t)(op4), (size_t)(op5)) +#define vluxseg3ei16_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei16_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vbool16_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (const int8_t *)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vluxseg3ei16_v_i8mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei16_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (const int8_t *)(op3), (vuint16mf2_t)(op4), (size_t)(op5)) +#define vluxseg3ei16_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei16_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vbool32_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (const int8_t *)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vluxseg3ei16_v_i8mf8(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei16_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (const int8_t *)(op3), (vuint16mf4_t)(op4), (size_t)(op5)) +#define vluxseg3ei16_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei16_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vbool64_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (const int8_t *)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vluxseg3ei16_v_u8m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei16_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (const uint8_t *)(op3), (vuint16m2_t)(op4), (size_t)(op5)) +#define vluxseg3ei16_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei16_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vbool8_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (const uint8_t *)(op7), (vuint16m2_t)(op8), (size_t)(op9)) +#define vluxseg3ei16_v_u8m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei16_v_u8m2((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (vuint8m2_t *)(op2), (const uint8_t *)(op3), (vuint16m4_t)(op4), (size_t)(op5)) +#define vluxseg3ei16_v_u8m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei16_v_u8m2_m((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (vuint8m2_t *)(op2), (vbool4_t)(op3), (vuint8m2_t)(op4), (vuint8m2_t)(op5), (vuint8m2_t)(op6), (const uint8_t *)(op7), (vuint16m4_t)(op8), (size_t)(op9)) +#define vluxseg3ei16_v_u8mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei16_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (const uint8_t *)(op3), (vuint16m1_t)(op4), (size_t)(op5)) +#define vluxseg3ei16_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei16_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vbool16_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (const uint8_t *)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vluxseg3ei16_v_u8mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei16_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (const uint8_t *)(op3), (vuint16mf2_t)(op4), (size_t)(op5)) +#define vluxseg3ei16_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei16_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vbool32_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (const uint8_t *)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vluxseg3ei16_v_u8mf8(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei16_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (const uint8_t *)(op3), (vuint16mf4_t)(op4), (size_t)(op5)) +#define vluxseg3ei16_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei16_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vbool64_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (const uint8_t *)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vluxseg4ei16_v_i8m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei16_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (const int8_t *)(op4), (vuint16m2_t)(op5), (size_t)(op6)) +#define vluxseg4ei16_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei16_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vbool8_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (const int8_t *)(op9), (vuint16m2_t)(op10), (size_t)(op11)) +#define vluxseg4ei16_v_i8m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei16_v_i8m2((vint8m2_t *)(op0), (vint8m2_t *)(op1), (vint8m2_t *)(op2), (vint8m2_t *)(op3), (const int8_t *)(op4), (vuint16m4_t)(op5), (size_t)(op6)) +#define vluxseg4ei16_v_i8m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei16_v_i8m2_m((vint8m2_t *)(op0), (vint8m2_t *)(op1), (vint8m2_t *)(op2), (vint8m2_t *)(op3), (vbool4_t)(op4), (vint8m2_t)(op5), (vint8m2_t)(op6), (vint8m2_t)(op7), (vint8m2_t)(op8), (const int8_t *)(op9), (vuint16m4_t)(op10), (size_t)(op11)) +#define vluxseg4ei16_v_i8mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei16_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (const int8_t *)(op4), (vuint16m1_t)(op5), (size_t)(op6)) +#define vluxseg4ei16_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei16_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vbool16_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (const int8_t *)(op9), (vuint16m1_t)(op10), (size_t)(op11)) +#define vluxseg4ei16_v_i8mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei16_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (const int8_t *)(op4), (vuint16mf2_t)(op5), (size_t)(op6)) +#define vluxseg4ei16_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei16_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vbool32_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (const int8_t *)(op9), (vuint16mf2_t)(op10), (size_t)(op11)) +#define vluxseg4ei16_v_i8mf8(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei16_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (const int8_t *)(op4), (vuint16mf4_t)(op5), (size_t)(op6)) +#define vluxseg4ei16_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei16_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vbool64_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (const int8_t *)(op9), (vuint16mf4_t)(op10), (size_t)(op11)) +#define vluxseg4ei16_v_u8m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei16_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (const uint8_t *)(op4), (vuint16m2_t)(op5), (size_t)(op6)) +#define vluxseg4ei16_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei16_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vbool8_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (const uint8_t *)(op9), (vuint16m2_t)(op10), (size_t)(op11)) +#define vluxseg4ei16_v_u8m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei16_v_u8m2((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (vuint8m2_t *)(op2), (vuint8m2_t *)(op3), (const uint8_t *)(op4), (vuint16m4_t)(op5), (size_t)(op6)) +#define vluxseg4ei16_v_u8m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei16_v_u8m2_m((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (vuint8m2_t *)(op2), (vuint8m2_t *)(op3), (vbool4_t)(op4), (vuint8m2_t)(op5), (vuint8m2_t)(op6), (vuint8m2_t)(op7), (vuint8m2_t)(op8), (const uint8_t *)(op9), (vuint16m4_t)(op10), (size_t)(op11)) +#define vluxseg4ei16_v_u8mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei16_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (const uint8_t *)(op4), (vuint16m1_t)(op5), (size_t)(op6)) +#define vluxseg4ei16_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei16_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vbool16_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (const uint8_t *)(op9), (vuint16m1_t)(op10), (size_t)(op11)) +#define vluxseg4ei16_v_u8mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei16_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (const uint8_t *)(op4), (vuint16mf2_t)(op5), (size_t)(op6)) +#define vluxseg4ei16_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei16_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vbool32_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (const uint8_t *)(op9), (vuint16mf2_t)(op10), (size_t)(op11)) +#define vluxseg4ei16_v_u8mf8(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei16_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (const uint8_t *)(op4), (vuint16mf4_t)(op5), (size_t)(op6)) +#define vluxseg4ei16_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei16_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vbool64_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (const uint8_t *)(op9), (vuint16mf4_t)(op10), (size_t)(op11)) +#define vluxseg5ei16_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei16_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (const int8_t *)(op5), (vuint16m2_t)(op6), (size_t)(op7)) +#define vluxseg5ei16_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei16_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vbool8_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (const int8_t *)(op11), (vuint16m2_t)(op12), (size_t)(op13)) +#define vluxseg5ei16_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei16_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (const int8_t *)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vluxseg5ei16_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei16_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vbool16_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (const int8_t *)(op11), (vuint16m1_t)(op12), (size_t)(op13)) +#define vluxseg5ei16_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei16_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (const int8_t *)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vluxseg5ei16_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei16_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vbool32_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (const int8_t *)(op11), (vuint16mf2_t)(op12), (size_t)(op13)) +#define vluxseg5ei16_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei16_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (const int8_t *)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vluxseg5ei16_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei16_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vbool64_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (const int8_t *)(op11), (vuint16mf4_t)(op12), (size_t)(op13)) +#define vluxseg5ei16_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei16_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (const uint8_t *)(op5), (vuint16m2_t)(op6), (size_t)(op7)) +#define vluxseg5ei16_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei16_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vbool8_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (const uint8_t *)(op11), (vuint16m2_t)(op12), (size_t)(op13)) +#define vluxseg5ei16_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei16_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (const uint8_t *)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vluxseg5ei16_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei16_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vbool16_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (const uint8_t *)(op11), (vuint16m1_t)(op12), (size_t)(op13)) +#define vluxseg5ei16_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei16_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (const uint8_t *)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vluxseg5ei16_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei16_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vbool32_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (const uint8_t *)(op11), (vuint16mf2_t)(op12), (size_t)(op13)) +#define vluxseg5ei16_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei16_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (const uint8_t *)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vluxseg5ei16_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei16_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vbool64_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (const uint8_t *)(op11), (vuint16mf4_t)(op12), (size_t)(op13)) +#define vluxseg6ei16_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei16_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (const int8_t *)(op6), (vuint16m2_t)(op7), (size_t)(op8)) +#define vluxseg6ei16_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei16_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vbool8_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (vint8m1_t)(op11), (vint8m1_t)(op12), (const int8_t *)(op13), (vuint16m2_t)(op14), (size_t)(op15)) +#define vluxseg6ei16_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei16_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (const int8_t *)(op6), (vuint16m1_t)(op7), (size_t)(op8)) +#define vluxseg6ei16_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei16_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vbool16_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (vint8mf2_t)(op11), (vint8mf2_t)(op12), (const int8_t *)(op13), (vuint16m1_t)(op14), (size_t)(op15)) +#define vluxseg6ei16_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei16_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (const int8_t *)(op6), (vuint16mf2_t)(op7), (size_t)(op8)) +#define vluxseg6ei16_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei16_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vbool32_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (vint8mf4_t)(op11), (vint8mf4_t)(op12), (const int8_t *)(op13), (vuint16mf2_t)(op14), (size_t)(op15)) +#define vluxseg6ei16_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei16_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (const int8_t *)(op6), (vuint16mf4_t)(op7), (size_t)(op8)) +#define vluxseg6ei16_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei16_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vbool64_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (vint8mf8_t)(op11), (vint8mf8_t)(op12), (const int8_t *)(op13), (vuint16mf4_t)(op14), (size_t)(op15)) +#define vluxseg6ei16_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei16_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (const uint8_t *)(op6), (vuint16m2_t)(op7), (size_t)(op8)) +#define vluxseg6ei16_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei16_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vbool8_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (vuint8m1_t)(op11), (vuint8m1_t)(op12), (const uint8_t *)(op13), (vuint16m2_t)(op14), (size_t)(op15)) +#define vluxseg6ei16_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei16_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (const uint8_t *)(op6), (vuint16m1_t)(op7), (size_t)(op8)) +#define vluxseg6ei16_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei16_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vbool16_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (vuint8mf2_t)(op11), (vuint8mf2_t)(op12), (const uint8_t *)(op13), (vuint16m1_t)(op14), (size_t)(op15)) +#define vluxseg6ei16_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei16_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (const uint8_t *)(op6), (vuint16mf2_t)(op7), (size_t)(op8)) +#define vluxseg6ei16_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei16_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vbool32_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (vuint8mf4_t)(op11), (vuint8mf4_t)(op12), (const uint8_t *)(op13), (vuint16mf2_t)(op14), (size_t)(op15)) +#define vluxseg6ei16_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei16_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (const uint8_t *)(op6), (vuint16mf4_t)(op7), (size_t)(op8)) +#define vluxseg6ei16_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei16_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vbool64_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (vuint8mf8_t)(op11), (vuint8mf8_t)(op12), (const uint8_t *)(op13), (vuint16mf4_t)(op14), (size_t)(op15)) +#define vluxseg7ei16_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei16_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vint8m1_t *)(op6), (const int8_t *)(op7), (vuint16m2_t)(op8), (size_t)(op9)) +#define vluxseg7ei16_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei16_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vint8m1_t *)(op6), (vbool8_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (vint8m1_t)(op11), (vint8m1_t)(op12), (vint8m1_t)(op13), (vint8m1_t)(op14), (const int8_t *)(op15), (vuint16m2_t)(op16), (size_t)(op17)) +#define vluxseg7ei16_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei16_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vint8mf2_t *)(op6), (const int8_t *)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vluxseg7ei16_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei16_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vint8mf2_t *)(op6), (vbool16_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (vint8mf2_t)(op11), (vint8mf2_t)(op12), (vint8mf2_t)(op13), (vint8mf2_t)(op14), (const int8_t *)(op15), (vuint16m1_t)(op16), (size_t)(op17)) +#define vluxseg7ei16_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei16_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vint8mf4_t *)(op6), (const int8_t *)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vluxseg7ei16_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei16_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vint8mf4_t *)(op6), (vbool32_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (vint8mf4_t)(op11), (vint8mf4_t)(op12), (vint8mf4_t)(op13), (vint8mf4_t)(op14), (const int8_t *)(op15), (vuint16mf2_t)(op16), (size_t)(op17)) +#define vluxseg7ei16_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei16_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vint8mf8_t *)(op6), (const int8_t *)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vluxseg7ei16_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei16_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vint8mf8_t *)(op6), (vbool64_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (vint8mf8_t)(op11), (vint8mf8_t)(op12), (vint8mf8_t)(op13), (vint8mf8_t)(op14), (const int8_t *)(op15), (vuint16mf4_t)(op16), (size_t)(op17)) +#define vluxseg7ei16_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei16_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vuint8m1_t *)(op6), (const uint8_t *)(op7), (vuint16m2_t)(op8), (size_t)(op9)) +#define vluxseg7ei16_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei16_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vuint8m1_t *)(op6), (vbool8_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (vuint8m1_t)(op11), (vuint8m1_t)(op12), (vuint8m1_t)(op13), (vuint8m1_t)(op14), (const uint8_t *)(op15), (vuint16m2_t)(op16), (size_t)(op17)) +#define vluxseg7ei16_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei16_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vuint8mf2_t *)(op6), (const uint8_t *)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vluxseg7ei16_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei16_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vuint8mf2_t *)(op6), (vbool16_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (vuint8mf2_t)(op11), (vuint8mf2_t)(op12), (vuint8mf2_t)(op13), (vuint8mf2_t)(op14), (const uint8_t *)(op15), (vuint16m1_t)(op16), (size_t)(op17)) +#define vluxseg7ei16_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei16_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vuint8mf4_t *)(op6), (const uint8_t *)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vluxseg7ei16_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei16_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vuint8mf4_t *)(op6), (vbool32_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (vuint8mf4_t)(op11), (vuint8mf4_t)(op12), (vuint8mf4_t)(op13), (vuint8mf4_t)(op14), (const uint8_t *)(op15), (vuint16mf2_t)(op16), (size_t)(op17)) +#define vluxseg7ei16_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei16_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vuint8mf8_t *)(op6), (const uint8_t *)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vluxseg7ei16_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei16_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vuint8mf8_t *)(op6), (vbool64_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (vuint8mf8_t)(op11), (vuint8mf8_t)(op12), (vuint8mf8_t)(op13), (vuint8mf8_t)(op14), (const uint8_t *)(op15), (vuint16mf4_t)(op16), (size_t)(op17)) +#define vluxseg8ei16_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei16_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vint8m1_t *)(op6), (vint8m1_t *)(op7), (const int8_t *)(op8), (vuint16m2_t)(op9), (size_t)(op10)) +#define vluxseg8ei16_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei16_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vint8m1_t *)(op6), (vint8m1_t *)(op7), (vbool8_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (vint8m1_t)(op11), (vint8m1_t)(op12), (vint8m1_t)(op13), (vint8m1_t)(op14), (vint8m1_t)(op15), (vint8m1_t)(op16), (const int8_t *)(op17), (vuint16m2_t)(op18), (size_t)(op19)) +#define vluxseg8ei16_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei16_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vint8mf2_t *)(op6), (vint8mf2_t *)(op7), (const int8_t *)(op8), (vuint16m1_t)(op9), (size_t)(op10)) +#define vluxseg8ei16_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei16_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vint8mf2_t *)(op6), (vint8mf2_t *)(op7), (vbool16_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (vint8mf2_t)(op11), (vint8mf2_t)(op12), (vint8mf2_t)(op13), (vint8mf2_t)(op14), (vint8mf2_t)(op15), (vint8mf2_t)(op16), (const int8_t *)(op17), (vuint16m1_t)(op18), (size_t)(op19)) +#define vluxseg8ei16_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei16_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vint8mf4_t *)(op6), (vint8mf4_t *)(op7), (const int8_t *)(op8), (vuint16mf2_t)(op9), (size_t)(op10)) +#define vluxseg8ei16_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei16_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vint8mf4_t *)(op6), (vint8mf4_t *)(op7), (vbool32_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (vint8mf4_t)(op11), (vint8mf4_t)(op12), (vint8mf4_t)(op13), (vint8mf4_t)(op14), (vint8mf4_t)(op15), (vint8mf4_t)(op16), (const int8_t *)(op17), (vuint16mf2_t)(op18), (size_t)(op19)) +#define vluxseg8ei16_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei16_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vint8mf8_t *)(op6), (vint8mf8_t *)(op7), (const int8_t *)(op8), (vuint16mf4_t)(op9), (size_t)(op10)) +#define vluxseg8ei16_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei16_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vint8mf8_t *)(op6), (vint8mf8_t *)(op7), (vbool64_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (vint8mf8_t)(op11), (vint8mf8_t)(op12), (vint8mf8_t)(op13), (vint8mf8_t)(op14), (vint8mf8_t)(op15), (vint8mf8_t)(op16), (const int8_t *)(op17), (vuint16mf4_t)(op18), (size_t)(op19)) +#define vluxseg8ei16_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei16_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vuint8m1_t *)(op6), (vuint8m1_t *)(op7), (const uint8_t *)(op8), (vuint16m2_t)(op9), (size_t)(op10)) +#define vluxseg8ei16_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei16_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vuint8m1_t *)(op6), (vuint8m1_t *)(op7), (vbool8_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (vuint8m1_t)(op11), (vuint8m1_t)(op12), (vuint8m1_t)(op13), (vuint8m1_t)(op14), (vuint8m1_t)(op15), (vuint8m1_t)(op16), (const uint8_t *)(op17), (vuint16m2_t)(op18), (size_t)(op19)) +#define vluxseg8ei16_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei16_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vuint8mf2_t *)(op6), (vuint8mf2_t *)(op7), (const uint8_t *)(op8), (vuint16m1_t)(op9), (size_t)(op10)) +#define vluxseg8ei16_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei16_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vuint8mf2_t *)(op6), (vuint8mf2_t *)(op7), (vbool16_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (vuint8mf2_t)(op11), (vuint8mf2_t)(op12), (vuint8mf2_t)(op13), (vuint8mf2_t)(op14), (vuint8mf2_t)(op15), (vuint8mf2_t)(op16), (const uint8_t *)(op17), (vuint16m1_t)(op18), (size_t)(op19)) +#define vluxseg8ei16_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei16_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vuint8mf4_t *)(op6), (vuint8mf4_t *)(op7), (const uint8_t *)(op8), (vuint16mf2_t)(op9), (size_t)(op10)) +#define vluxseg8ei16_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei16_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vuint8mf4_t *)(op6), (vuint8mf4_t *)(op7), (vbool32_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (vuint8mf4_t)(op11), (vuint8mf4_t)(op12), (vuint8mf4_t)(op13), (vuint8mf4_t)(op14), (vuint8mf4_t)(op15), (vuint8mf4_t)(op16), (const uint8_t *)(op17), (vuint16mf2_t)(op18), (size_t)(op19)) +#define vluxseg8ei16_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei16_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vuint8mf8_t *)(op6), (vuint8mf8_t *)(op7), (const uint8_t *)(op8), (vuint16mf4_t)(op9), (size_t)(op10)) +#define vluxseg8ei16_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei16_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vuint8mf8_t *)(op6), (vuint8mf8_t *)(op7), (vbool64_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (vuint8mf8_t)(op11), (vuint8mf8_t)(op12), (vuint8mf8_t)(op13), (vuint8mf8_t)(op14), (vuint8mf8_t)(op15), (vuint8mf8_t)(op16), (const uint8_t *)(op17), (vuint16mf4_t)(op18), (size_t)(op19)) +#define vluxseg2ei32_v_i8m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (const int8_t *)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vbool8_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (const int8_t *)(op5), (vuint32m4_t)(op6), (size_t)(op7)) +#define vluxseg2ei32_v_i8m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_i8m2((vint8m2_t *)(op0), (vint8m2_t *)(op1), (const int8_t *)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_i8m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_i8m2_m((vint8m2_t *)(op0), (vint8m2_t *)(op1), (vbool4_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (const int8_t *)(op5), (vuint32m8_t)(op6), (size_t)(op7)) +#define vluxseg2ei32_v_i8mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (const int8_t *)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vbool16_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (const int8_t *)(op5), (vuint32m2_t)(op6), (size_t)(op7)) +#define vluxseg2ei32_v_i8mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (const int8_t *)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vbool32_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (const int8_t *)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vluxseg2ei32_v_i8mf8(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (const int8_t *)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vbool64_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (const int8_t *)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vluxseg2ei32_v_u8m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (const uint8_t *)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vbool8_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (const uint8_t *)(op5), (vuint32m4_t)(op6), (size_t)(op7)) +#define vluxseg2ei32_v_u8m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_u8m2((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (const uint8_t *)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_u8m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_u8m2_m((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (vbool4_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (const uint8_t *)(op5), (vuint32m8_t)(op6), (size_t)(op7)) +#define vluxseg2ei32_v_u8mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (const uint8_t *)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vbool16_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (const uint8_t *)(op5), (vuint32m2_t)(op6), (size_t)(op7)) +#define vluxseg2ei32_v_u8mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (const uint8_t *)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vbool32_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (const uint8_t *)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vluxseg2ei32_v_u8mf8(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (const uint8_t *)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vbool64_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (const uint8_t *)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vluxseg3ei32_v_i8m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei32_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (const int8_t *)(op3), (vuint32m4_t)(op4), (size_t)(op5)) +#define vluxseg3ei32_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei32_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vbool8_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (const int8_t *)(op7), (vuint32m4_t)(op8), (size_t)(op9)) +#define vluxseg3ei32_v_i8m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei32_v_i8m2((vint8m2_t *)(op0), (vint8m2_t *)(op1), (vint8m2_t *)(op2), (const int8_t *)(op3), (vuint32m8_t)(op4), (size_t)(op5)) +#define vluxseg3ei32_v_i8m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei32_v_i8m2_m((vint8m2_t *)(op0), (vint8m2_t *)(op1), (vint8m2_t *)(op2), (vbool4_t)(op3), (vint8m2_t)(op4), (vint8m2_t)(op5), (vint8m2_t)(op6), (const int8_t *)(op7), (vuint32m8_t)(op8), (size_t)(op9)) +#define vluxseg3ei32_v_i8mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei32_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (const int8_t *)(op3), (vuint32m2_t)(op4), (size_t)(op5)) +#define vluxseg3ei32_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei32_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vbool16_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (const int8_t *)(op7), (vuint32m2_t)(op8), (size_t)(op9)) +#define vluxseg3ei32_v_i8mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei32_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (const int8_t *)(op3), (vuint32m1_t)(op4), (size_t)(op5)) +#define vluxseg3ei32_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei32_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vbool32_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (const int8_t *)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vluxseg3ei32_v_i8mf8(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei32_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (const int8_t *)(op3), (vuint32mf2_t)(op4), (size_t)(op5)) +#define vluxseg3ei32_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei32_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vbool64_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (const int8_t *)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vluxseg3ei32_v_u8m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei32_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (const uint8_t *)(op3), (vuint32m4_t)(op4), (size_t)(op5)) +#define vluxseg3ei32_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei32_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vbool8_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (const uint8_t *)(op7), (vuint32m4_t)(op8), (size_t)(op9)) +#define vluxseg3ei32_v_u8m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei32_v_u8m2((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (vuint8m2_t *)(op2), (const uint8_t *)(op3), (vuint32m8_t)(op4), (size_t)(op5)) +#define vluxseg3ei32_v_u8m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei32_v_u8m2_m((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (vuint8m2_t *)(op2), (vbool4_t)(op3), (vuint8m2_t)(op4), (vuint8m2_t)(op5), (vuint8m2_t)(op6), (const uint8_t *)(op7), (vuint32m8_t)(op8), (size_t)(op9)) +#define vluxseg3ei32_v_u8mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei32_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (const uint8_t *)(op3), (vuint32m2_t)(op4), (size_t)(op5)) +#define vluxseg3ei32_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei32_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vbool16_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (const uint8_t *)(op7), (vuint32m2_t)(op8), (size_t)(op9)) +#define vluxseg3ei32_v_u8mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei32_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (const uint8_t *)(op3), (vuint32m1_t)(op4), (size_t)(op5)) +#define vluxseg3ei32_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei32_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vbool32_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (const uint8_t *)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vluxseg3ei32_v_u8mf8(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei32_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (const uint8_t *)(op3), (vuint32mf2_t)(op4), (size_t)(op5)) +#define vluxseg3ei32_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei32_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vbool64_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (const uint8_t *)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vluxseg4ei32_v_i8m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei32_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (const int8_t *)(op4), (vuint32m4_t)(op5), (size_t)(op6)) +#define vluxseg4ei32_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei32_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vbool8_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (const int8_t *)(op9), (vuint32m4_t)(op10), (size_t)(op11)) +#define vluxseg4ei32_v_i8m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei32_v_i8m2((vint8m2_t *)(op0), (vint8m2_t *)(op1), (vint8m2_t *)(op2), (vint8m2_t *)(op3), (const int8_t *)(op4), (vuint32m8_t)(op5), (size_t)(op6)) +#define vluxseg4ei32_v_i8m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei32_v_i8m2_m((vint8m2_t *)(op0), (vint8m2_t *)(op1), (vint8m2_t *)(op2), (vint8m2_t *)(op3), (vbool4_t)(op4), (vint8m2_t)(op5), (vint8m2_t)(op6), (vint8m2_t)(op7), (vint8m2_t)(op8), (const int8_t *)(op9), (vuint32m8_t)(op10), (size_t)(op11)) +#define vluxseg4ei32_v_i8mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei32_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (const int8_t *)(op4), (vuint32m2_t)(op5), (size_t)(op6)) +#define vluxseg4ei32_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei32_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vbool16_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (const int8_t *)(op9), (vuint32m2_t)(op10), (size_t)(op11)) +#define vluxseg4ei32_v_i8mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei32_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (const int8_t *)(op4), (vuint32m1_t)(op5), (size_t)(op6)) +#define vluxseg4ei32_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei32_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vbool32_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (const int8_t *)(op9), (vuint32m1_t)(op10), (size_t)(op11)) +#define vluxseg4ei32_v_i8mf8(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei32_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (const int8_t *)(op4), (vuint32mf2_t)(op5), (size_t)(op6)) +#define vluxseg4ei32_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei32_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vbool64_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (const int8_t *)(op9), (vuint32mf2_t)(op10), (size_t)(op11)) +#define vluxseg4ei32_v_u8m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei32_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (const uint8_t *)(op4), (vuint32m4_t)(op5), (size_t)(op6)) +#define vluxseg4ei32_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei32_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vbool8_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (const uint8_t *)(op9), (vuint32m4_t)(op10), (size_t)(op11)) +#define vluxseg4ei32_v_u8m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei32_v_u8m2((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (vuint8m2_t *)(op2), (vuint8m2_t *)(op3), (const uint8_t *)(op4), (vuint32m8_t)(op5), (size_t)(op6)) +#define vluxseg4ei32_v_u8m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei32_v_u8m2_m((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (vuint8m2_t *)(op2), (vuint8m2_t *)(op3), (vbool4_t)(op4), (vuint8m2_t)(op5), (vuint8m2_t)(op6), (vuint8m2_t)(op7), (vuint8m2_t)(op8), (const uint8_t *)(op9), (vuint32m8_t)(op10), (size_t)(op11)) +#define vluxseg4ei32_v_u8mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei32_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (const uint8_t *)(op4), (vuint32m2_t)(op5), (size_t)(op6)) +#define vluxseg4ei32_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei32_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vbool16_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (const uint8_t *)(op9), (vuint32m2_t)(op10), (size_t)(op11)) +#define vluxseg4ei32_v_u8mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei32_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (const uint8_t *)(op4), (vuint32m1_t)(op5), (size_t)(op6)) +#define vluxseg4ei32_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei32_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vbool32_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (const uint8_t *)(op9), (vuint32m1_t)(op10), (size_t)(op11)) +#define vluxseg4ei32_v_u8mf8(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei32_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (const uint8_t *)(op4), (vuint32mf2_t)(op5), (size_t)(op6)) +#define vluxseg4ei32_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei32_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vbool64_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (const uint8_t *)(op9), (vuint32mf2_t)(op10), (size_t)(op11)) +#define vluxseg5ei32_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei32_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (const int8_t *)(op5), (vuint32m4_t)(op6), (size_t)(op7)) +#define vluxseg5ei32_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei32_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vbool8_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (const int8_t *)(op11), (vuint32m4_t)(op12), (size_t)(op13)) +#define vluxseg5ei32_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei32_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (const int8_t *)(op5), (vuint32m2_t)(op6), (size_t)(op7)) +#define vluxseg5ei32_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei32_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vbool16_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (const int8_t *)(op11), (vuint32m2_t)(op12), (size_t)(op13)) +#define vluxseg5ei32_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei32_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (const int8_t *)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vluxseg5ei32_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei32_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vbool32_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (const int8_t *)(op11), (vuint32m1_t)(op12), (size_t)(op13)) +#define vluxseg5ei32_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei32_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (const int8_t *)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vluxseg5ei32_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei32_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vbool64_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (const int8_t *)(op11), (vuint32mf2_t)(op12), (size_t)(op13)) +#define vluxseg5ei32_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei32_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (const uint8_t *)(op5), (vuint32m4_t)(op6), (size_t)(op7)) +#define vluxseg5ei32_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei32_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vbool8_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (const uint8_t *)(op11), (vuint32m4_t)(op12), (size_t)(op13)) +#define vluxseg5ei32_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei32_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (const uint8_t *)(op5), (vuint32m2_t)(op6), (size_t)(op7)) +#define vluxseg5ei32_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei32_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vbool16_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (const uint8_t *)(op11), (vuint32m2_t)(op12), (size_t)(op13)) +#define vluxseg5ei32_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei32_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (const uint8_t *)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vluxseg5ei32_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei32_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vbool32_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (const uint8_t *)(op11), (vuint32m1_t)(op12), (size_t)(op13)) +#define vluxseg5ei32_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei32_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (const uint8_t *)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vluxseg5ei32_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei32_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vbool64_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (const uint8_t *)(op11), (vuint32mf2_t)(op12), (size_t)(op13)) +#define vluxseg6ei32_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei32_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (const int8_t *)(op6), (vuint32m4_t)(op7), (size_t)(op8)) +#define vluxseg6ei32_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei32_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vbool8_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (vint8m1_t)(op11), (vint8m1_t)(op12), (const int8_t *)(op13), (vuint32m4_t)(op14), (size_t)(op15)) +#define vluxseg6ei32_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei32_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (const int8_t *)(op6), (vuint32m2_t)(op7), (size_t)(op8)) +#define vluxseg6ei32_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei32_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vbool16_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (vint8mf2_t)(op11), (vint8mf2_t)(op12), (const int8_t *)(op13), (vuint32m2_t)(op14), (size_t)(op15)) +#define vluxseg6ei32_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei32_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (const int8_t *)(op6), (vuint32m1_t)(op7), (size_t)(op8)) +#define vluxseg6ei32_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei32_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vbool32_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (vint8mf4_t)(op11), (vint8mf4_t)(op12), (const int8_t *)(op13), (vuint32m1_t)(op14), (size_t)(op15)) +#define vluxseg6ei32_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei32_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (const int8_t *)(op6), (vuint32mf2_t)(op7), (size_t)(op8)) +#define vluxseg6ei32_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei32_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vbool64_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (vint8mf8_t)(op11), (vint8mf8_t)(op12), (const int8_t *)(op13), (vuint32mf2_t)(op14), (size_t)(op15)) +#define vluxseg6ei32_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei32_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (const uint8_t *)(op6), (vuint32m4_t)(op7), (size_t)(op8)) +#define vluxseg6ei32_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei32_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vbool8_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (vuint8m1_t)(op11), (vuint8m1_t)(op12), (const uint8_t *)(op13), (vuint32m4_t)(op14), (size_t)(op15)) +#define vluxseg6ei32_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei32_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (const uint8_t *)(op6), (vuint32m2_t)(op7), (size_t)(op8)) +#define vluxseg6ei32_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei32_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vbool16_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (vuint8mf2_t)(op11), (vuint8mf2_t)(op12), (const uint8_t *)(op13), (vuint32m2_t)(op14), (size_t)(op15)) +#define vluxseg6ei32_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei32_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (const uint8_t *)(op6), (vuint32m1_t)(op7), (size_t)(op8)) +#define vluxseg6ei32_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei32_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vbool32_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (vuint8mf4_t)(op11), (vuint8mf4_t)(op12), (const uint8_t *)(op13), (vuint32m1_t)(op14), (size_t)(op15)) +#define vluxseg6ei32_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei32_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (const uint8_t *)(op6), (vuint32mf2_t)(op7), (size_t)(op8)) +#define vluxseg6ei32_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei32_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vbool64_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (vuint8mf8_t)(op11), (vuint8mf8_t)(op12), (const uint8_t *)(op13), (vuint32mf2_t)(op14), (size_t)(op15)) +#define vluxseg7ei32_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei32_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vint8m1_t *)(op6), (const int8_t *)(op7), (vuint32m4_t)(op8), (size_t)(op9)) +#define vluxseg7ei32_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei32_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vint8m1_t *)(op6), (vbool8_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (vint8m1_t)(op11), (vint8m1_t)(op12), (vint8m1_t)(op13), (vint8m1_t)(op14), (const int8_t *)(op15), (vuint32m4_t)(op16), (size_t)(op17)) +#define vluxseg7ei32_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei32_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vint8mf2_t *)(op6), (const int8_t *)(op7), (vuint32m2_t)(op8), (size_t)(op9)) +#define vluxseg7ei32_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei32_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vint8mf2_t *)(op6), (vbool16_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (vint8mf2_t)(op11), (vint8mf2_t)(op12), (vint8mf2_t)(op13), (vint8mf2_t)(op14), (const int8_t *)(op15), (vuint32m2_t)(op16), (size_t)(op17)) +#define vluxseg7ei32_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei32_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vint8mf4_t *)(op6), (const int8_t *)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vluxseg7ei32_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei32_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vint8mf4_t *)(op6), (vbool32_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (vint8mf4_t)(op11), (vint8mf4_t)(op12), (vint8mf4_t)(op13), (vint8mf4_t)(op14), (const int8_t *)(op15), (vuint32m1_t)(op16), (size_t)(op17)) +#define vluxseg7ei32_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei32_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vint8mf8_t *)(op6), (const int8_t *)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vluxseg7ei32_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei32_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vint8mf8_t *)(op6), (vbool64_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (vint8mf8_t)(op11), (vint8mf8_t)(op12), (vint8mf8_t)(op13), (vint8mf8_t)(op14), (const int8_t *)(op15), (vuint32mf2_t)(op16), (size_t)(op17)) +#define vluxseg7ei32_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei32_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vuint8m1_t *)(op6), (const uint8_t *)(op7), (vuint32m4_t)(op8), (size_t)(op9)) +#define vluxseg7ei32_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei32_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vuint8m1_t *)(op6), (vbool8_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (vuint8m1_t)(op11), (vuint8m1_t)(op12), (vuint8m1_t)(op13), (vuint8m1_t)(op14), (const uint8_t *)(op15), (vuint32m4_t)(op16), (size_t)(op17)) +#define vluxseg7ei32_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei32_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vuint8mf2_t *)(op6), (const uint8_t *)(op7), (vuint32m2_t)(op8), (size_t)(op9)) +#define vluxseg7ei32_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei32_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vuint8mf2_t *)(op6), (vbool16_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (vuint8mf2_t)(op11), (vuint8mf2_t)(op12), (vuint8mf2_t)(op13), (vuint8mf2_t)(op14), (const uint8_t *)(op15), (vuint32m2_t)(op16), (size_t)(op17)) +#define vluxseg7ei32_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei32_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vuint8mf4_t *)(op6), (const uint8_t *)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vluxseg7ei32_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei32_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vuint8mf4_t *)(op6), (vbool32_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (vuint8mf4_t)(op11), (vuint8mf4_t)(op12), (vuint8mf4_t)(op13), (vuint8mf4_t)(op14), (const uint8_t *)(op15), (vuint32m1_t)(op16), (size_t)(op17)) +#define vluxseg7ei32_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei32_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vuint8mf8_t *)(op6), (const uint8_t *)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vluxseg7ei32_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei32_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vuint8mf8_t *)(op6), (vbool64_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (vuint8mf8_t)(op11), (vuint8mf8_t)(op12), (vuint8mf8_t)(op13), (vuint8mf8_t)(op14), (const uint8_t *)(op15), (vuint32mf2_t)(op16), (size_t)(op17)) +#define vluxseg8ei32_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei32_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vint8m1_t *)(op6), (vint8m1_t *)(op7), (const int8_t *)(op8), (vuint32m4_t)(op9), (size_t)(op10)) +#define vluxseg8ei32_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei32_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vint8m1_t *)(op6), (vint8m1_t *)(op7), (vbool8_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (vint8m1_t)(op11), (vint8m1_t)(op12), (vint8m1_t)(op13), (vint8m1_t)(op14), (vint8m1_t)(op15), (vint8m1_t)(op16), (const int8_t *)(op17), (vuint32m4_t)(op18), (size_t)(op19)) +#define vluxseg8ei32_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei32_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vint8mf2_t *)(op6), (vint8mf2_t *)(op7), (const int8_t *)(op8), (vuint32m2_t)(op9), (size_t)(op10)) +#define vluxseg8ei32_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei32_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vint8mf2_t *)(op6), (vint8mf2_t *)(op7), (vbool16_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (vint8mf2_t)(op11), (vint8mf2_t)(op12), (vint8mf2_t)(op13), (vint8mf2_t)(op14), (vint8mf2_t)(op15), (vint8mf2_t)(op16), (const int8_t *)(op17), (vuint32m2_t)(op18), (size_t)(op19)) +#define vluxseg8ei32_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei32_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vint8mf4_t *)(op6), (vint8mf4_t *)(op7), (const int8_t *)(op8), (vuint32m1_t)(op9), (size_t)(op10)) +#define vluxseg8ei32_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei32_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vint8mf4_t *)(op6), (vint8mf4_t *)(op7), (vbool32_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (vint8mf4_t)(op11), (vint8mf4_t)(op12), (vint8mf4_t)(op13), (vint8mf4_t)(op14), (vint8mf4_t)(op15), (vint8mf4_t)(op16), (const int8_t *)(op17), (vuint32m1_t)(op18), (size_t)(op19)) +#define vluxseg8ei32_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei32_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vint8mf8_t *)(op6), (vint8mf8_t *)(op7), (const int8_t *)(op8), (vuint32mf2_t)(op9), (size_t)(op10)) +#define vluxseg8ei32_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei32_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vint8mf8_t *)(op6), (vint8mf8_t *)(op7), (vbool64_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (vint8mf8_t)(op11), (vint8mf8_t)(op12), (vint8mf8_t)(op13), (vint8mf8_t)(op14), (vint8mf8_t)(op15), (vint8mf8_t)(op16), (const int8_t *)(op17), (vuint32mf2_t)(op18), (size_t)(op19)) +#define vluxseg8ei32_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei32_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vuint8m1_t *)(op6), (vuint8m1_t *)(op7), (const uint8_t *)(op8), (vuint32m4_t)(op9), (size_t)(op10)) +#define vluxseg8ei32_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei32_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vuint8m1_t *)(op6), (vuint8m1_t *)(op7), (vbool8_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (vuint8m1_t)(op11), (vuint8m1_t)(op12), (vuint8m1_t)(op13), (vuint8m1_t)(op14), (vuint8m1_t)(op15), (vuint8m1_t)(op16), (const uint8_t *)(op17), (vuint32m4_t)(op18), (size_t)(op19)) +#define vluxseg8ei32_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei32_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vuint8mf2_t *)(op6), (vuint8mf2_t *)(op7), (const uint8_t *)(op8), (vuint32m2_t)(op9), (size_t)(op10)) +#define vluxseg8ei32_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei32_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vuint8mf2_t *)(op6), (vuint8mf2_t *)(op7), (vbool16_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (vuint8mf2_t)(op11), (vuint8mf2_t)(op12), (vuint8mf2_t)(op13), (vuint8mf2_t)(op14), (vuint8mf2_t)(op15), (vuint8mf2_t)(op16), (const uint8_t *)(op17), (vuint32m2_t)(op18), (size_t)(op19)) +#define vluxseg8ei32_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei32_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vuint8mf4_t *)(op6), (vuint8mf4_t *)(op7), (const uint8_t *)(op8), (vuint32m1_t)(op9), (size_t)(op10)) +#define vluxseg8ei32_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei32_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vuint8mf4_t *)(op6), (vuint8mf4_t *)(op7), (vbool32_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (vuint8mf4_t)(op11), (vuint8mf4_t)(op12), (vuint8mf4_t)(op13), (vuint8mf4_t)(op14), (vuint8mf4_t)(op15), (vuint8mf4_t)(op16), (const uint8_t *)(op17), (vuint32m1_t)(op18), (size_t)(op19)) +#define vluxseg8ei32_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei32_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vuint8mf8_t *)(op6), (vuint8mf8_t *)(op7), (const uint8_t *)(op8), (vuint32mf2_t)(op9), (size_t)(op10)) +#define vluxseg8ei32_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei32_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vuint8mf8_t *)(op6), (vuint8mf8_t *)(op7), (vbool64_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (vuint8mf8_t)(op11), (vuint8mf8_t)(op12), (vuint8mf8_t)(op13), (vuint8mf8_t)(op14), (vuint8mf8_t)(op15), (vuint8mf8_t)(op16), (const uint8_t *)(op17), (vuint32mf2_t)(op18), (size_t)(op19)) +#define vluxseg2ei64_v_i8m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei64_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (const int8_t *)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vluxseg2ei64_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei64_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vbool8_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (const int8_t *)(op5), (vuint64m8_t)(op6), (size_t)(op7)) +#define vluxseg2ei64_v_i8mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei64_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (const int8_t *)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vluxseg2ei64_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei64_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vbool16_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (const int8_t *)(op5), (vuint64m4_t)(op6), (size_t)(op7)) +#define vluxseg2ei64_v_i8mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei64_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (const int8_t *)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vluxseg2ei64_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei64_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vbool32_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (const int8_t *)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vluxseg2ei64_v_i8mf8(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei64_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (const int8_t *)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vluxseg2ei64_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei64_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vbool64_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (const int8_t *)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vluxseg2ei64_v_u8m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei64_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (const uint8_t *)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vluxseg2ei64_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei64_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vbool8_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (const uint8_t *)(op5), (vuint64m8_t)(op6), (size_t)(op7)) +#define vluxseg2ei64_v_u8mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei64_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (const uint8_t *)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vluxseg2ei64_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei64_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vbool16_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (const uint8_t *)(op5), (vuint64m4_t)(op6), (size_t)(op7)) +#define vluxseg2ei64_v_u8mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei64_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (const uint8_t *)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vluxseg2ei64_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei64_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vbool32_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (const uint8_t *)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vluxseg2ei64_v_u8mf8(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei64_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (const uint8_t *)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vluxseg2ei64_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei64_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vbool64_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (const uint8_t *)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vluxseg3ei64_v_i8m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei64_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (const int8_t *)(op3), (vuint64m8_t)(op4), (size_t)(op5)) +#define vluxseg3ei64_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei64_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vbool8_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (const int8_t *)(op7), (vuint64m8_t)(op8), (size_t)(op9)) +#define vluxseg3ei64_v_i8mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei64_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (const int8_t *)(op3), (vuint64m4_t)(op4), (size_t)(op5)) +#define vluxseg3ei64_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei64_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vbool16_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (const int8_t *)(op7), (vuint64m4_t)(op8), (size_t)(op9)) +#define vluxseg3ei64_v_i8mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei64_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (const int8_t *)(op3), (vuint64m2_t)(op4), (size_t)(op5)) +#define vluxseg3ei64_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei64_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vbool32_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (const int8_t *)(op7), (vuint64m2_t)(op8), (size_t)(op9)) +#define vluxseg3ei64_v_i8mf8(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei64_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (const int8_t *)(op3), (vuint64m1_t)(op4), (size_t)(op5)) +#define vluxseg3ei64_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei64_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vbool64_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (const int8_t *)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vluxseg3ei64_v_u8m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei64_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (const uint8_t *)(op3), (vuint64m8_t)(op4), (size_t)(op5)) +#define vluxseg3ei64_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei64_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vbool8_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (const uint8_t *)(op7), (vuint64m8_t)(op8), (size_t)(op9)) +#define vluxseg3ei64_v_u8mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei64_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (const uint8_t *)(op3), (vuint64m4_t)(op4), (size_t)(op5)) +#define vluxseg3ei64_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei64_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vbool16_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (const uint8_t *)(op7), (vuint64m4_t)(op8), (size_t)(op9)) +#define vluxseg3ei64_v_u8mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei64_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (const uint8_t *)(op3), (vuint64m2_t)(op4), (size_t)(op5)) +#define vluxseg3ei64_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei64_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vbool32_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (const uint8_t *)(op7), (vuint64m2_t)(op8), (size_t)(op9)) +#define vluxseg3ei64_v_u8mf8(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei64_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (const uint8_t *)(op3), (vuint64m1_t)(op4), (size_t)(op5)) +#define vluxseg3ei64_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei64_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vbool64_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (const uint8_t *)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vluxseg4ei64_v_i8m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei64_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (const int8_t *)(op4), (vuint64m8_t)(op5), (size_t)(op6)) +#define vluxseg4ei64_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei64_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vbool8_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (const int8_t *)(op9), (vuint64m8_t)(op10), (size_t)(op11)) +#define vluxseg4ei64_v_i8mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei64_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (const int8_t *)(op4), (vuint64m4_t)(op5), (size_t)(op6)) +#define vluxseg4ei64_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei64_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vbool16_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (const int8_t *)(op9), (vuint64m4_t)(op10), (size_t)(op11)) +#define vluxseg4ei64_v_i8mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei64_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (const int8_t *)(op4), (vuint64m2_t)(op5), (size_t)(op6)) +#define vluxseg4ei64_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei64_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vbool32_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (const int8_t *)(op9), (vuint64m2_t)(op10), (size_t)(op11)) +#define vluxseg4ei64_v_i8mf8(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei64_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (const int8_t *)(op4), (vuint64m1_t)(op5), (size_t)(op6)) +#define vluxseg4ei64_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei64_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vbool64_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (const int8_t *)(op9), (vuint64m1_t)(op10), (size_t)(op11)) +#define vluxseg4ei64_v_u8m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei64_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (const uint8_t *)(op4), (vuint64m8_t)(op5), (size_t)(op6)) +#define vluxseg4ei64_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei64_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vbool8_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (const uint8_t *)(op9), (vuint64m8_t)(op10), (size_t)(op11)) +#define vluxseg4ei64_v_u8mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei64_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (const uint8_t *)(op4), (vuint64m4_t)(op5), (size_t)(op6)) +#define vluxseg4ei64_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei64_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vbool16_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (const uint8_t *)(op9), (vuint64m4_t)(op10), (size_t)(op11)) +#define vluxseg4ei64_v_u8mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei64_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (const uint8_t *)(op4), (vuint64m2_t)(op5), (size_t)(op6)) +#define vluxseg4ei64_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei64_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vbool32_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (const uint8_t *)(op9), (vuint64m2_t)(op10), (size_t)(op11)) +#define vluxseg4ei64_v_u8mf8(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei64_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (const uint8_t *)(op4), (vuint64m1_t)(op5), (size_t)(op6)) +#define vluxseg4ei64_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei64_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vbool64_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (const uint8_t *)(op9), (vuint64m1_t)(op10), (size_t)(op11)) +#define vluxseg5ei64_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei64_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (const int8_t *)(op5), (vuint64m8_t)(op6), (size_t)(op7)) +#define vluxseg5ei64_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei64_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vbool8_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (const int8_t *)(op11), (vuint64m8_t)(op12), (size_t)(op13)) +#define vluxseg5ei64_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei64_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (const int8_t *)(op5), (vuint64m4_t)(op6), (size_t)(op7)) +#define vluxseg5ei64_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei64_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vbool16_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (const int8_t *)(op11), (vuint64m4_t)(op12), (size_t)(op13)) +#define vluxseg5ei64_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei64_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (const int8_t *)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vluxseg5ei64_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei64_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vbool32_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (const int8_t *)(op11), (vuint64m2_t)(op12), (size_t)(op13)) +#define vluxseg5ei64_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei64_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (const int8_t *)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vluxseg5ei64_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei64_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vbool64_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (const int8_t *)(op11), (vuint64m1_t)(op12), (size_t)(op13)) +#define vluxseg5ei64_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei64_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (const uint8_t *)(op5), (vuint64m8_t)(op6), (size_t)(op7)) +#define vluxseg5ei64_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei64_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vbool8_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (const uint8_t *)(op11), (vuint64m8_t)(op12), (size_t)(op13)) +#define vluxseg5ei64_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei64_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (const uint8_t *)(op5), (vuint64m4_t)(op6), (size_t)(op7)) +#define vluxseg5ei64_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei64_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vbool16_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (const uint8_t *)(op11), (vuint64m4_t)(op12), (size_t)(op13)) +#define vluxseg5ei64_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei64_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (const uint8_t *)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vluxseg5ei64_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei64_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vbool32_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (const uint8_t *)(op11), (vuint64m2_t)(op12), (size_t)(op13)) +#define vluxseg5ei64_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei64_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (const uint8_t *)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vluxseg5ei64_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei64_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vbool64_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (const uint8_t *)(op11), (vuint64m1_t)(op12), (size_t)(op13)) +#define vluxseg6ei64_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei64_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (const int8_t *)(op6), (vuint64m8_t)(op7), (size_t)(op8)) +#define vluxseg6ei64_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei64_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vbool8_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (vint8m1_t)(op11), (vint8m1_t)(op12), (const int8_t *)(op13), (vuint64m8_t)(op14), (size_t)(op15)) +#define vluxseg6ei64_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei64_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (const int8_t *)(op6), (vuint64m4_t)(op7), (size_t)(op8)) +#define vluxseg6ei64_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei64_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vbool16_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (vint8mf2_t)(op11), (vint8mf2_t)(op12), (const int8_t *)(op13), (vuint64m4_t)(op14), (size_t)(op15)) +#define vluxseg6ei64_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei64_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (const int8_t *)(op6), (vuint64m2_t)(op7), (size_t)(op8)) +#define vluxseg6ei64_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei64_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vbool32_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (vint8mf4_t)(op11), (vint8mf4_t)(op12), (const int8_t *)(op13), (vuint64m2_t)(op14), (size_t)(op15)) +#define vluxseg6ei64_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei64_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (const int8_t *)(op6), (vuint64m1_t)(op7), (size_t)(op8)) +#define vluxseg6ei64_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei64_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vbool64_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (vint8mf8_t)(op11), (vint8mf8_t)(op12), (const int8_t *)(op13), (vuint64m1_t)(op14), (size_t)(op15)) +#define vluxseg6ei64_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei64_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (const uint8_t *)(op6), (vuint64m8_t)(op7), (size_t)(op8)) +#define vluxseg6ei64_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei64_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vbool8_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (vuint8m1_t)(op11), (vuint8m1_t)(op12), (const uint8_t *)(op13), (vuint64m8_t)(op14), (size_t)(op15)) +#define vluxseg6ei64_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei64_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (const uint8_t *)(op6), (vuint64m4_t)(op7), (size_t)(op8)) +#define vluxseg6ei64_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei64_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vbool16_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (vuint8mf2_t)(op11), (vuint8mf2_t)(op12), (const uint8_t *)(op13), (vuint64m4_t)(op14), (size_t)(op15)) +#define vluxseg6ei64_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei64_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (const uint8_t *)(op6), (vuint64m2_t)(op7), (size_t)(op8)) +#define vluxseg6ei64_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei64_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vbool32_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (vuint8mf4_t)(op11), (vuint8mf4_t)(op12), (const uint8_t *)(op13), (vuint64m2_t)(op14), (size_t)(op15)) +#define vluxseg6ei64_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei64_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (const uint8_t *)(op6), (vuint64m1_t)(op7), (size_t)(op8)) +#define vluxseg6ei64_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei64_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vbool64_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (vuint8mf8_t)(op11), (vuint8mf8_t)(op12), (const uint8_t *)(op13), (vuint64m1_t)(op14), (size_t)(op15)) +#define vluxseg7ei64_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei64_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vint8m1_t *)(op6), (const int8_t *)(op7), (vuint64m8_t)(op8), (size_t)(op9)) +#define vluxseg7ei64_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei64_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vint8m1_t *)(op6), (vbool8_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (vint8m1_t)(op11), (vint8m1_t)(op12), (vint8m1_t)(op13), (vint8m1_t)(op14), (const int8_t *)(op15), (vuint64m8_t)(op16), (size_t)(op17)) +#define vluxseg7ei64_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei64_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vint8mf2_t *)(op6), (const int8_t *)(op7), (vuint64m4_t)(op8), (size_t)(op9)) +#define vluxseg7ei64_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei64_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vint8mf2_t *)(op6), (vbool16_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (vint8mf2_t)(op11), (vint8mf2_t)(op12), (vint8mf2_t)(op13), (vint8mf2_t)(op14), (const int8_t *)(op15), (vuint64m4_t)(op16), (size_t)(op17)) +#define vluxseg7ei64_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei64_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vint8mf4_t *)(op6), (const int8_t *)(op7), (vuint64m2_t)(op8), (size_t)(op9)) +#define vluxseg7ei64_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei64_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vint8mf4_t *)(op6), (vbool32_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (vint8mf4_t)(op11), (vint8mf4_t)(op12), (vint8mf4_t)(op13), (vint8mf4_t)(op14), (const int8_t *)(op15), (vuint64m2_t)(op16), (size_t)(op17)) +#define vluxseg7ei64_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei64_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vint8mf8_t *)(op6), (const int8_t *)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vluxseg7ei64_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei64_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vint8mf8_t *)(op6), (vbool64_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (vint8mf8_t)(op11), (vint8mf8_t)(op12), (vint8mf8_t)(op13), (vint8mf8_t)(op14), (const int8_t *)(op15), (vuint64m1_t)(op16), (size_t)(op17)) +#define vluxseg7ei64_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei64_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vuint8m1_t *)(op6), (const uint8_t *)(op7), (vuint64m8_t)(op8), (size_t)(op9)) +#define vluxseg7ei64_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei64_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vuint8m1_t *)(op6), (vbool8_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (vuint8m1_t)(op11), (vuint8m1_t)(op12), (vuint8m1_t)(op13), (vuint8m1_t)(op14), (const uint8_t *)(op15), (vuint64m8_t)(op16), (size_t)(op17)) +#define vluxseg7ei64_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei64_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vuint8mf2_t *)(op6), (const uint8_t *)(op7), (vuint64m4_t)(op8), (size_t)(op9)) +#define vluxseg7ei64_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei64_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vuint8mf2_t *)(op6), (vbool16_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (vuint8mf2_t)(op11), (vuint8mf2_t)(op12), (vuint8mf2_t)(op13), (vuint8mf2_t)(op14), (const uint8_t *)(op15), (vuint64m4_t)(op16), (size_t)(op17)) +#define vluxseg7ei64_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei64_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vuint8mf4_t *)(op6), (const uint8_t *)(op7), (vuint64m2_t)(op8), (size_t)(op9)) +#define vluxseg7ei64_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei64_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vuint8mf4_t *)(op6), (vbool32_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (vuint8mf4_t)(op11), (vuint8mf4_t)(op12), (vuint8mf4_t)(op13), (vuint8mf4_t)(op14), (const uint8_t *)(op15), (vuint64m2_t)(op16), (size_t)(op17)) +#define vluxseg7ei64_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei64_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vuint8mf8_t *)(op6), (const uint8_t *)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vluxseg7ei64_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei64_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vuint8mf8_t *)(op6), (vbool64_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (vuint8mf8_t)(op11), (vuint8mf8_t)(op12), (vuint8mf8_t)(op13), (vuint8mf8_t)(op14), (const uint8_t *)(op15), (vuint64m1_t)(op16), (size_t)(op17)) +#define vluxseg8ei64_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei64_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vint8m1_t *)(op6), (vint8m1_t *)(op7), (const int8_t *)(op8), (vuint64m8_t)(op9), (size_t)(op10)) +#define vluxseg8ei64_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei64_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vint8m1_t *)(op6), (vint8m1_t *)(op7), (vbool8_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (vint8m1_t)(op11), (vint8m1_t)(op12), (vint8m1_t)(op13), (vint8m1_t)(op14), (vint8m1_t)(op15), (vint8m1_t)(op16), (const int8_t *)(op17), (vuint64m8_t)(op18), (size_t)(op19)) +#define vluxseg8ei64_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei64_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vint8mf2_t *)(op6), (vint8mf2_t *)(op7), (const int8_t *)(op8), (vuint64m4_t)(op9), (size_t)(op10)) +#define vluxseg8ei64_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei64_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vint8mf2_t *)(op6), (vint8mf2_t *)(op7), (vbool16_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (vint8mf2_t)(op11), (vint8mf2_t)(op12), (vint8mf2_t)(op13), (vint8mf2_t)(op14), (vint8mf2_t)(op15), (vint8mf2_t)(op16), (const int8_t *)(op17), (vuint64m4_t)(op18), (size_t)(op19)) +#define vluxseg8ei64_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei64_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vint8mf4_t *)(op6), (vint8mf4_t *)(op7), (const int8_t *)(op8), (vuint64m2_t)(op9), (size_t)(op10)) +#define vluxseg8ei64_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei64_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vint8mf4_t *)(op6), (vint8mf4_t *)(op7), (vbool32_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (vint8mf4_t)(op11), (vint8mf4_t)(op12), (vint8mf4_t)(op13), (vint8mf4_t)(op14), (vint8mf4_t)(op15), (vint8mf4_t)(op16), (const int8_t *)(op17), (vuint64m2_t)(op18), (size_t)(op19)) +#define vluxseg8ei64_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei64_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vint8mf8_t *)(op6), (vint8mf8_t *)(op7), (const int8_t *)(op8), (vuint64m1_t)(op9), (size_t)(op10)) +#define vluxseg8ei64_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei64_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vint8mf8_t *)(op6), (vint8mf8_t *)(op7), (vbool64_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (vint8mf8_t)(op11), (vint8mf8_t)(op12), (vint8mf8_t)(op13), (vint8mf8_t)(op14), (vint8mf8_t)(op15), (vint8mf8_t)(op16), (const int8_t *)(op17), (vuint64m1_t)(op18), (size_t)(op19)) +#define vluxseg8ei64_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei64_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vuint8m1_t *)(op6), (vuint8m1_t *)(op7), (const uint8_t *)(op8), (vuint64m8_t)(op9), (size_t)(op10)) +#define vluxseg8ei64_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei64_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vuint8m1_t *)(op6), (vuint8m1_t *)(op7), (vbool8_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (vuint8m1_t)(op11), (vuint8m1_t)(op12), (vuint8m1_t)(op13), (vuint8m1_t)(op14), (vuint8m1_t)(op15), (vuint8m1_t)(op16), (const uint8_t *)(op17), (vuint64m8_t)(op18), (size_t)(op19)) +#define vluxseg8ei64_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei64_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vuint8mf2_t *)(op6), (vuint8mf2_t *)(op7), (const uint8_t *)(op8), (vuint64m4_t)(op9), (size_t)(op10)) +#define vluxseg8ei64_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei64_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vuint8mf2_t *)(op6), (vuint8mf2_t *)(op7), (vbool16_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (vuint8mf2_t)(op11), (vuint8mf2_t)(op12), (vuint8mf2_t)(op13), (vuint8mf2_t)(op14), (vuint8mf2_t)(op15), (vuint8mf2_t)(op16), (const uint8_t *)(op17), (vuint64m4_t)(op18), (size_t)(op19)) +#define vluxseg8ei64_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei64_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vuint8mf4_t *)(op6), (vuint8mf4_t *)(op7), (const uint8_t *)(op8), (vuint64m2_t)(op9), (size_t)(op10)) +#define vluxseg8ei64_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei64_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vuint8mf4_t *)(op6), (vuint8mf4_t *)(op7), (vbool32_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (vuint8mf4_t)(op11), (vuint8mf4_t)(op12), (vuint8mf4_t)(op13), (vuint8mf4_t)(op14), (vuint8mf4_t)(op15), (vuint8mf4_t)(op16), (const uint8_t *)(op17), (vuint64m2_t)(op18), (size_t)(op19)) +#define vluxseg8ei64_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei64_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vuint8mf8_t *)(op6), (vuint8mf8_t *)(op7), (const uint8_t *)(op8), (vuint64m1_t)(op9), (size_t)(op10)) +#define vluxseg8ei64_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei64_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vuint8mf8_t *)(op6), (vuint8mf8_t *)(op7), (vbool64_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (vuint8mf8_t)(op11), (vuint8mf8_t)(op12), (vuint8mf8_t)(op13), (vuint8mf8_t)(op14), (vuint8mf8_t)(op15), (vuint8mf8_t)(op16), (const uint8_t *)(op17), (vuint64m1_t)(op18), (size_t)(op19)) +#define vluxseg2ei8_v_i16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (const int16_t *)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vbool16_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (const int16_t *)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vluxseg2ei8_v_i16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_i16m2((vint16m2_t *)(op0), (vint16m2_t *)(op1), (const int16_t *)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_i16m2_m((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vbool8_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (const int16_t *)(op5), (vuint8m1_t)(op6), (size_t)(op7)) +#define vluxseg2ei8_v_i16m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_i16m4((vint16m4_t *)(op0), (vint16m4_t *)(op1), (const int16_t *)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_i16m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_i16m4_m((vint16m4_t *)(op0), (vint16m4_t *)(op1), (vbool4_t)(op2), (vint16m4_t)(op3), (vint16m4_t)(op4), (const int16_t *)(op5), (vuint8m2_t)(op6), (size_t)(op7)) +#define vluxseg2ei8_v_i16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (const int16_t *)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vbool32_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (const int16_t *)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vluxseg2ei8_v_i16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (const int16_t *)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vbool64_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (const int16_t *)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vluxseg2ei8_v_u16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (const uint16_t *)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vbool16_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (const uint16_t *)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vluxseg2ei8_v_u16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_u16m2((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (const uint16_t *)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_u16m2_m((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vbool8_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (const uint16_t *)(op5), (vuint8m1_t)(op6), (size_t)(op7)) +#define vluxseg2ei8_v_u16m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_u16m4((vuint16m4_t *)(op0), (vuint16m4_t *)(op1), (const uint16_t *)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_u16m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_u16m4_m((vuint16m4_t *)(op0), (vuint16m4_t *)(op1), (vbool4_t)(op2), (vuint16m4_t)(op3), (vuint16m4_t)(op4), (const uint16_t *)(op5), (vuint8m2_t)(op6), (size_t)(op7)) +#define vluxseg2ei8_v_u16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (const uint16_t *)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vbool32_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (const uint16_t *)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vluxseg2ei8_v_u16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (const uint16_t *)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vbool64_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (const uint16_t *)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vluxseg3ei8_v_i16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei8_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (const int16_t *)(op3), (vuint8mf2_t)(op4), (size_t)(op5)) +#define vluxseg3ei8_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei8_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vbool16_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (const int16_t *)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vluxseg3ei8_v_i16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei8_v_i16m2((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vint16m2_t *)(op2), (const int16_t *)(op3), (vuint8m1_t)(op4), (size_t)(op5)) +#define vluxseg3ei8_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei8_v_i16m2_m((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vint16m2_t *)(op2), (vbool8_t)(op3), (vint16m2_t)(op4), (vint16m2_t)(op5), (vint16m2_t)(op6), (const int16_t *)(op7), (vuint8m1_t)(op8), (size_t)(op9)) +#define vluxseg3ei8_v_i16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei8_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (const int16_t *)(op3), (vuint8mf4_t)(op4), (size_t)(op5)) +#define vluxseg3ei8_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei8_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vbool32_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (const int16_t *)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vluxseg3ei8_v_i16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei8_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (const int16_t *)(op3), (vuint8mf8_t)(op4), (size_t)(op5)) +#define vluxseg3ei8_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei8_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vbool64_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (const int16_t *)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vluxseg3ei8_v_u16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei8_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (const uint16_t *)(op3), (vuint8mf2_t)(op4), (size_t)(op5)) +#define vluxseg3ei8_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei8_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vbool16_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (const uint16_t *)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vluxseg3ei8_v_u16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei8_v_u16m2((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vuint16m2_t *)(op2), (const uint16_t *)(op3), (vuint8m1_t)(op4), (size_t)(op5)) +#define vluxseg3ei8_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei8_v_u16m2_m((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vuint16m2_t *)(op2), (vbool8_t)(op3), (vuint16m2_t)(op4), (vuint16m2_t)(op5), (vuint16m2_t)(op6), (const uint16_t *)(op7), (vuint8m1_t)(op8), (size_t)(op9)) +#define vluxseg3ei8_v_u16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei8_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (const uint16_t *)(op3), (vuint8mf4_t)(op4), (size_t)(op5)) +#define vluxseg3ei8_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei8_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vbool32_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (const uint16_t *)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vluxseg3ei8_v_u16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei8_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (const uint16_t *)(op3), (vuint8mf8_t)(op4), (size_t)(op5)) +#define vluxseg3ei8_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei8_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vbool64_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (const uint16_t *)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vluxseg4ei8_v_i16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei8_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (const int16_t *)(op4), (vuint8mf2_t)(op5), (size_t)(op6)) +#define vluxseg4ei8_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei8_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vbool16_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (const int16_t *)(op9), (vuint8mf2_t)(op10), (size_t)(op11)) +#define vluxseg4ei8_v_i16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei8_v_i16m2((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vint16m2_t *)(op2), (vint16m2_t *)(op3), (const int16_t *)(op4), (vuint8m1_t)(op5), (size_t)(op6)) +#define vluxseg4ei8_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei8_v_i16m2_m((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vint16m2_t *)(op2), (vint16m2_t *)(op3), (vbool8_t)(op4), (vint16m2_t)(op5), (vint16m2_t)(op6), (vint16m2_t)(op7), (vint16m2_t)(op8), (const int16_t *)(op9), (vuint8m1_t)(op10), (size_t)(op11)) +#define vluxseg4ei8_v_i16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei8_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (const int16_t *)(op4), (vuint8mf4_t)(op5), (size_t)(op6)) +#define vluxseg4ei8_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei8_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vbool32_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (const int16_t *)(op9), (vuint8mf4_t)(op10), (size_t)(op11)) +#define vluxseg4ei8_v_i16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei8_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (const int16_t *)(op4), (vuint8mf8_t)(op5), (size_t)(op6)) +#define vluxseg4ei8_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei8_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vbool64_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (const int16_t *)(op9), (vuint8mf8_t)(op10), (size_t)(op11)) +#define vluxseg4ei8_v_u16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei8_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (const uint16_t *)(op4), (vuint8mf2_t)(op5), (size_t)(op6)) +#define vluxseg4ei8_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei8_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vbool16_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (const uint16_t *)(op9), (vuint8mf2_t)(op10), (size_t)(op11)) +#define vluxseg4ei8_v_u16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei8_v_u16m2((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vuint16m2_t *)(op2), (vuint16m2_t *)(op3), (const uint16_t *)(op4), (vuint8m1_t)(op5), (size_t)(op6)) +#define vluxseg4ei8_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei8_v_u16m2_m((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vuint16m2_t *)(op2), (vuint16m2_t *)(op3), (vbool8_t)(op4), (vuint16m2_t)(op5), (vuint16m2_t)(op6), (vuint16m2_t)(op7), (vuint16m2_t)(op8), (const uint16_t *)(op9), (vuint8m1_t)(op10), (size_t)(op11)) +#define vluxseg4ei8_v_u16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei8_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (const uint16_t *)(op4), (vuint8mf4_t)(op5), (size_t)(op6)) +#define vluxseg4ei8_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei8_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vbool32_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (const uint16_t *)(op9), (vuint8mf4_t)(op10), (size_t)(op11)) +#define vluxseg4ei8_v_u16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei8_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (const uint16_t *)(op4), (vuint8mf8_t)(op5), (size_t)(op6)) +#define vluxseg4ei8_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei8_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vbool64_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (const uint16_t *)(op9), (vuint8mf8_t)(op10), (size_t)(op11)) +#define vluxseg5ei8_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei8_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (const int16_t *)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vluxseg5ei8_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei8_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vbool16_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (const int16_t *)(op11), (vuint8mf2_t)(op12), (size_t)(op13)) +#define vluxseg5ei8_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei8_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (const int16_t *)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vluxseg5ei8_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei8_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vbool32_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (const int16_t *)(op11), (vuint8mf4_t)(op12), (size_t)(op13)) +#define vluxseg5ei8_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei8_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (const int16_t *)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vluxseg5ei8_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei8_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vbool64_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (const int16_t *)(op11), (vuint8mf8_t)(op12), (size_t)(op13)) +#define vluxseg5ei8_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei8_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (const uint16_t *)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vluxseg5ei8_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei8_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vbool16_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (const uint16_t *)(op11), (vuint8mf2_t)(op12), (size_t)(op13)) +#define vluxseg5ei8_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei8_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (const uint16_t *)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vluxseg5ei8_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei8_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vbool32_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (const uint16_t *)(op11), (vuint8mf4_t)(op12), (size_t)(op13)) +#define vluxseg5ei8_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei8_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (const uint16_t *)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vluxseg5ei8_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei8_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vbool64_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (const uint16_t *)(op11), (vuint8mf8_t)(op12), (size_t)(op13)) +#define vluxseg6ei8_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei8_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (const int16_t *)(op6), (vuint8mf2_t)(op7), (size_t)(op8)) +#define vluxseg6ei8_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei8_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vbool16_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (vint16m1_t)(op11), (vint16m1_t)(op12), (const int16_t *)(op13), (vuint8mf2_t)(op14), (size_t)(op15)) +#define vluxseg6ei8_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei8_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (const int16_t *)(op6), (vuint8mf4_t)(op7), (size_t)(op8)) +#define vluxseg6ei8_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei8_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vbool32_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (vint16mf2_t)(op11), (vint16mf2_t)(op12), (const int16_t *)(op13), (vuint8mf4_t)(op14), (size_t)(op15)) +#define vluxseg6ei8_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei8_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (const int16_t *)(op6), (vuint8mf8_t)(op7), (size_t)(op8)) +#define vluxseg6ei8_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei8_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vbool64_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (vint16mf4_t)(op11), (vint16mf4_t)(op12), (const int16_t *)(op13), (vuint8mf8_t)(op14), (size_t)(op15)) +#define vluxseg6ei8_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei8_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (const uint16_t *)(op6), (vuint8mf2_t)(op7), (size_t)(op8)) +#define vluxseg6ei8_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei8_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vbool16_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (vuint16m1_t)(op11), (vuint16m1_t)(op12), (const uint16_t *)(op13), (vuint8mf2_t)(op14), (size_t)(op15)) +#define vluxseg6ei8_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei8_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (const uint16_t *)(op6), (vuint8mf4_t)(op7), (size_t)(op8)) +#define vluxseg6ei8_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei8_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vbool32_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (vuint16mf2_t)(op11), (vuint16mf2_t)(op12), (const uint16_t *)(op13), (vuint8mf4_t)(op14), (size_t)(op15)) +#define vluxseg6ei8_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei8_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (const uint16_t *)(op6), (vuint8mf8_t)(op7), (size_t)(op8)) +#define vluxseg6ei8_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei8_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vbool64_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (vuint16mf4_t)(op11), (vuint16mf4_t)(op12), (const uint16_t *)(op13), (vuint8mf8_t)(op14), (size_t)(op15)) +#define vluxseg7ei8_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei8_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vint16m1_t *)(op6), (const int16_t *)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vluxseg7ei8_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei8_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vint16m1_t *)(op6), (vbool16_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (vint16m1_t)(op11), (vint16m1_t)(op12), (vint16m1_t)(op13), (vint16m1_t)(op14), (const int16_t *)(op15), (vuint8mf2_t)(op16), (size_t)(op17)) +#define vluxseg7ei8_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei8_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vint16mf2_t *)(op6), (const int16_t *)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vluxseg7ei8_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei8_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vint16mf2_t *)(op6), (vbool32_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (vint16mf2_t)(op11), (vint16mf2_t)(op12), (vint16mf2_t)(op13), (vint16mf2_t)(op14), (const int16_t *)(op15), (vuint8mf4_t)(op16), (size_t)(op17)) +#define vluxseg7ei8_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei8_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vint16mf4_t *)(op6), (const int16_t *)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vluxseg7ei8_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei8_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vint16mf4_t *)(op6), (vbool64_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (vint16mf4_t)(op11), (vint16mf4_t)(op12), (vint16mf4_t)(op13), (vint16mf4_t)(op14), (const int16_t *)(op15), (vuint8mf8_t)(op16), (size_t)(op17)) +#define vluxseg7ei8_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei8_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vuint16m1_t *)(op6), (const uint16_t *)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vluxseg7ei8_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei8_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vuint16m1_t *)(op6), (vbool16_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (vuint16m1_t)(op11), (vuint16m1_t)(op12), (vuint16m1_t)(op13), (vuint16m1_t)(op14), (const uint16_t *)(op15), (vuint8mf2_t)(op16), (size_t)(op17)) +#define vluxseg7ei8_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei8_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vuint16mf2_t *)(op6), (const uint16_t *)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vluxseg7ei8_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei8_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vuint16mf2_t *)(op6), (vbool32_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (vuint16mf2_t)(op11), (vuint16mf2_t)(op12), (vuint16mf2_t)(op13), (vuint16mf2_t)(op14), (const uint16_t *)(op15), (vuint8mf4_t)(op16), (size_t)(op17)) +#define vluxseg7ei8_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei8_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vuint16mf4_t *)(op6), (const uint16_t *)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vluxseg7ei8_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei8_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vuint16mf4_t *)(op6), (vbool64_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (vuint16mf4_t)(op11), (vuint16mf4_t)(op12), (vuint16mf4_t)(op13), (vuint16mf4_t)(op14), (const uint16_t *)(op15), (vuint8mf8_t)(op16), (size_t)(op17)) +#define vluxseg8ei8_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei8_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vint16m1_t *)(op6), (vint16m1_t *)(op7), (const int16_t *)(op8), (vuint8mf2_t)(op9), (size_t)(op10)) +#define vluxseg8ei8_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei8_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vint16m1_t *)(op6), (vint16m1_t *)(op7), (vbool16_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (vint16m1_t)(op11), (vint16m1_t)(op12), (vint16m1_t)(op13), (vint16m1_t)(op14), (vint16m1_t)(op15), (vint16m1_t)(op16), (const int16_t *)(op17), (vuint8mf2_t)(op18), (size_t)(op19)) +#define vluxseg8ei8_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei8_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vint16mf2_t *)(op6), (vint16mf2_t *)(op7), (const int16_t *)(op8), (vuint8mf4_t)(op9), (size_t)(op10)) +#define vluxseg8ei8_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei8_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vint16mf2_t *)(op6), (vint16mf2_t *)(op7), (vbool32_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (vint16mf2_t)(op11), (vint16mf2_t)(op12), (vint16mf2_t)(op13), (vint16mf2_t)(op14), (vint16mf2_t)(op15), (vint16mf2_t)(op16), (const int16_t *)(op17), (vuint8mf4_t)(op18), (size_t)(op19)) +#define vluxseg8ei8_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei8_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vint16mf4_t *)(op6), (vint16mf4_t *)(op7), (const int16_t *)(op8), (vuint8mf8_t)(op9), (size_t)(op10)) +#define vluxseg8ei8_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei8_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vint16mf4_t *)(op6), (vint16mf4_t *)(op7), (vbool64_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (vint16mf4_t)(op11), (vint16mf4_t)(op12), (vint16mf4_t)(op13), (vint16mf4_t)(op14), (vint16mf4_t)(op15), (vint16mf4_t)(op16), (const int16_t *)(op17), (vuint8mf8_t)(op18), (size_t)(op19)) +#define vluxseg8ei8_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei8_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vuint16m1_t *)(op6), (vuint16m1_t *)(op7), (const uint16_t *)(op8), (vuint8mf2_t)(op9), (size_t)(op10)) +#define vluxseg8ei8_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei8_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vuint16m1_t *)(op6), (vuint16m1_t *)(op7), (vbool16_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (vuint16m1_t)(op11), (vuint16m1_t)(op12), (vuint16m1_t)(op13), (vuint16m1_t)(op14), (vuint16m1_t)(op15), (vuint16m1_t)(op16), (const uint16_t *)(op17), (vuint8mf2_t)(op18), (size_t)(op19)) +#define vluxseg8ei8_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei8_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vuint16mf2_t *)(op6), (vuint16mf2_t *)(op7), (const uint16_t *)(op8), (vuint8mf4_t)(op9), (size_t)(op10)) +#define vluxseg8ei8_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei8_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vuint16mf2_t *)(op6), (vuint16mf2_t *)(op7), (vbool32_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (vuint16mf2_t)(op11), (vuint16mf2_t)(op12), (vuint16mf2_t)(op13), (vuint16mf2_t)(op14), (vuint16mf2_t)(op15), (vuint16mf2_t)(op16), (const uint16_t *)(op17), (vuint8mf4_t)(op18), (size_t)(op19)) +#define vluxseg8ei8_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei8_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vuint16mf4_t *)(op6), (vuint16mf4_t *)(op7), (const uint16_t *)(op8), (vuint8mf8_t)(op9), (size_t)(op10)) +#define vluxseg8ei8_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei8_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vuint16mf4_t *)(op6), (vuint16mf4_t *)(op7), (vbool64_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (vuint16mf4_t)(op11), (vuint16mf4_t)(op12), (vuint16mf4_t)(op13), (vuint16mf4_t)(op14), (vuint16mf4_t)(op15), (vuint16mf4_t)(op16), (const uint16_t *)(op17), (vuint8mf8_t)(op18), (size_t)(op19)) +#define vluxseg2ei16_v_i16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (const int16_t *)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vbool16_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (const int16_t *)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vluxseg2ei16_v_i16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_i16m2((vint16m2_t *)(op0), (vint16m2_t *)(op1), (const int16_t *)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_i16m2_m((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vbool8_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (const int16_t *)(op5), (vuint16m2_t)(op6), (size_t)(op7)) +#define vluxseg2ei16_v_i16m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_i16m4((vint16m4_t *)(op0), (vint16m4_t *)(op1), (const int16_t *)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_i16m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_i16m4_m((vint16m4_t *)(op0), (vint16m4_t *)(op1), (vbool4_t)(op2), (vint16m4_t)(op3), (vint16m4_t)(op4), (const int16_t *)(op5), (vuint16m4_t)(op6), (size_t)(op7)) +#define vluxseg2ei16_v_i16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (const int16_t *)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vbool32_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (const int16_t *)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vluxseg2ei16_v_i16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (const int16_t *)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vbool64_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (const int16_t *)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vluxseg2ei16_v_u16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (const uint16_t *)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vbool16_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (const uint16_t *)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vluxseg2ei16_v_u16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_u16m2((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (const uint16_t *)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_u16m2_m((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vbool8_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (const uint16_t *)(op5), (vuint16m2_t)(op6), (size_t)(op7)) +#define vluxseg2ei16_v_u16m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_u16m4((vuint16m4_t *)(op0), (vuint16m4_t *)(op1), (const uint16_t *)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_u16m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_u16m4_m((vuint16m4_t *)(op0), (vuint16m4_t *)(op1), (vbool4_t)(op2), (vuint16m4_t)(op3), (vuint16m4_t)(op4), (const uint16_t *)(op5), (vuint16m4_t)(op6), (size_t)(op7)) +#define vluxseg2ei16_v_u16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (const uint16_t *)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vbool32_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (const uint16_t *)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vluxseg2ei16_v_u16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (const uint16_t *)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vbool64_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (const uint16_t *)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vluxseg3ei16_v_i16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei16_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (const int16_t *)(op3), (vuint16m1_t)(op4), (size_t)(op5)) +#define vluxseg3ei16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei16_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vbool16_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (const int16_t *)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vluxseg3ei16_v_i16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei16_v_i16m2((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vint16m2_t *)(op2), (const int16_t *)(op3), (vuint16m2_t)(op4), (size_t)(op5)) +#define vluxseg3ei16_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei16_v_i16m2_m((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vint16m2_t *)(op2), (vbool8_t)(op3), (vint16m2_t)(op4), (vint16m2_t)(op5), (vint16m2_t)(op6), (const int16_t *)(op7), (vuint16m2_t)(op8), (size_t)(op9)) +#define vluxseg3ei16_v_i16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei16_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (const int16_t *)(op3), (vuint16mf2_t)(op4), (size_t)(op5)) +#define vluxseg3ei16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei16_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vbool32_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (const int16_t *)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vluxseg3ei16_v_i16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei16_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (const int16_t *)(op3), (vuint16mf4_t)(op4), (size_t)(op5)) +#define vluxseg3ei16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei16_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vbool64_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (const int16_t *)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vluxseg3ei16_v_u16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei16_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (const uint16_t *)(op3), (vuint16m1_t)(op4), (size_t)(op5)) +#define vluxseg3ei16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei16_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vbool16_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (const uint16_t *)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vluxseg3ei16_v_u16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei16_v_u16m2((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vuint16m2_t *)(op2), (const uint16_t *)(op3), (vuint16m2_t)(op4), (size_t)(op5)) +#define vluxseg3ei16_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei16_v_u16m2_m((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vuint16m2_t *)(op2), (vbool8_t)(op3), (vuint16m2_t)(op4), (vuint16m2_t)(op5), (vuint16m2_t)(op6), (const uint16_t *)(op7), (vuint16m2_t)(op8), (size_t)(op9)) +#define vluxseg3ei16_v_u16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei16_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (const uint16_t *)(op3), (vuint16mf2_t)(op4), (size_t)(op5)) +#define vluxseg3ei16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei16_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vbool32_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (const uint16_t *)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vluxseg3ei16_v_u16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei16_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (const uint16_t *)(op3), (vuint16mf4_t)(op4), (size_t)(op5)) +#define vluxseg3ei16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei16_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vbool64_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (const uint16_t *)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vluxseg4ei16_v_i16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei16_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (const int16_t *)(op4), (vuint16m1_t)(op5), (size_t)(op6)) +#define vluxseg4ei16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei16_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vbool16_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (const int16_t *)(op9), (vuint16m1_t)(op10), (size_t)(op11)) +#define vluxseg4ei16_v_i16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei16_v_i16m2((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vint16m2_t *)(op2), (vint16m2_t *)(op3), (const int16_t *)(op4), (vuint16m2_t)(op5), (size_t)(op6)) +#define vluxseg4ei16_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei16_v_i16m2_m((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vint16m2_t *)(op2), (vint16m2_t *)(op3), (vbool8_t)(op4), (vint16m2_t)(op5), (vint16m2_t)(op6), (vint16m2_t)(op7), (vint16m2_t)(op8), (const int16_t *)(op9), (vuint16m2_t)(op10), (size_t)(op11)) +#define vluxseg4ei16_v_i16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei16_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (const int16_t *)(op4), (vuint16mf2_t)(op5), (size_t)(op6)) +#define vluxseg4ei16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei16_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vbool32_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (const int16_t *)(op9), (vuint16mf2_t)(op10), (size_t)(op11)) +#define vluxseg4ei16_v_i16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei16_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (const int16_t *)(op4), (vuint16mf4_t)(op5), (size_t)(op6)) +#define vluxseg4ei16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei16_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vbool64_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (const int16_t *)(op9), (vuint16mf4_t)(op10), (size_t)(op11)) +#define vluxseg4ei16_v_u16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei16_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (const uint16_t *)(op4), (vuint16m1_t)(op5), (size_t)(op6)) +#define vluxseg4ei16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei16_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vbool16_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (const uint16_t *)(op9), (vuint16m1_t)(op10), (size_t)(op11)) +#define vluxseg4ei16_v_u16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei16_v_u16m2((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vuint16m2_t *)(op2), (vuint16m2_t *)(op3), (const uint16_t *)(op4), (vuint16m2_t)(op5), (size_t)(op6)) +#define vluxseg4ei16_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei16_v_u16m2_m((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vuint16m2_t *)(op2), (vuint16m2_t *)(op3), (vbool8_t)(op4), (vuint16m2_t)(op5), (vuint16m2_t)(op6), (vuint16m2_t)(op7), (vuint16m2_t)(op8), (const uint16_t *)(op9), (vuint16m2_t)(op10), (size_t)(op11)) +#define vluxseg4ei16_v_u16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei16_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (const uint16_t *)(op4), (vuint16mf2_t)(op5), (size_t)(op6)) +#define vluxseg4ei16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei16_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vbool32_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (const uint16_t *)(op9), (vuint16mf2_t)(op10), (size_t)(op11)) +#define vluxseg4ei16_v_u16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei16_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (const uint16_t *)(op4), (vuint16mf4_t)(op5), (size_t)(op6)) +#define vluxseg4ei16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei16_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vbool64_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (const uint16_t *)(op9), (vuint16mf4_t)(op10), (size_t)(op11)) +#define vluxseg5ei16_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei16_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (const int16_t *)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vluxseg5ei16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei16_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vbool16_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (const int16_t *)(op11), (vuint16m1_t)(op12), (size_t)(op13)) +#define vluxseg5ei16_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei16_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (const int16_t *)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vluxseg5ei16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei16_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vbool32_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (const int16_t *)(op11), (vuint16mf2_t)(op12), (size_t)(op13)) +#define vluxseg5ei16_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei16_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (const int16_t *)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vluxseg5ei16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei16_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vbool64_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (const int16_t *)(op11), (vuint16mf4_t)(op12), (size_t)(op13)) +#define vluxseg5ei16_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei16_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (const uint16_t *)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vluxseg5ei16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei16_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vbool16_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (const uint16_t *)(op11), (vuint16m1_t)(op12), (size_t)(op13)) +#define vluxseg5ei16_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei16_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (const uint16_t *)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vluxseg5ei16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei16_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vbool32_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (const uint16_t *)(op11), (vuint16mf2_t)(op12), (size_t)(op13)) +#define vluxseg5ei16_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei16_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (const uint16_t *)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vluxseg5ei16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei16_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vbool64_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (const uint16_t *)(op11), (vuint16mf4_t)(op12), (size_t)(op13)) +#define vluxseg6ei16_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei16_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (const int16_t *)(op6), (vuint16m1_t)(op7), (size_t)(op8)) +#define vluxseg6ei16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei16_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vbool16_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (vint16m1_t)(op11), (vint16m1_t)(op12), (const int16_t *)(op13), (vuint16m1_t)(op14), (size_t)(op15)) +#define vluxseg6ei16_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei16_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (const int16_t *)(op6), (vuint16mf2_t)(op7), (size_t)(op8)) +#define vluxseg6ei16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei16_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vbool32_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (vint16mf2_t)(op11), (vint16mf2_t)(op12), (const int16_t *)(op13), (vuint16mf2_t)(op14), (size_t)(op15)) +#define vluxseg6ei16_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei16_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (const int16_t *)(op6), (vuint16mf4_t)(op7), (size_t)(op8)) +#define vluxseg6ei16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei16_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vbool64_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (vint16mf4_t)(op11), (vint16mf4_t)(op12), (const int16_t *)(op13), (vuint16mf4_t)(op14), (size_t)(op15)) +#define vluxseg6ei16_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei16_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (const uint16_t *)(op6), (vuint16m1_t)(op7), (size_t)(op8)) +#define vluxseg6ei16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei16_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vbool16_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (vuint16m1_t)(op11), (vuint16m1_t)(op12), (const uint16_t *)(op13), (vuint16m1_t)(op14), (size_t)(op15)) +#define vluxseg6ei16_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei16_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (const uint16_t *)(op6), (vuint16mf2_t)(op7), (size_t)(op8)) +#define vluxseg6ei16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei16_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vbool32_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (vuint16mf2_t)(op11), (vuint16mf2_t)(op12), (const uint16_t *)(op13), (vuint16mf2_t)(op14), (size_t)(op15)) +#define vluxseg6ei16_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei16_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (const uint16_t *)(op6), (vuint16mf4_t)(op7), (size_t)(op8)) +#define vluxseg6ei16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei16_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vbool64_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (vuint16mf4_t)(op11), (vuint16mf4_t)(op12), (const uint16_t *)(op13), (vuint16mf4_t)(op14), (size_t)(op15)) +#define vluxseg7ei16_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei16_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vint16m1_t *)(op6), (const int16_t *)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vluxseg7ei16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei16_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vint16m1_t *)(op6), (vbool16_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (vint16m1_t)(op11), (vint16m1_t)(op12), (vint16m1_t)(op13), (vint16m1_t)(op14), (const int16_t *)(op15), (vuint16m1_t)(op16), (size_t)(op17)) +#define vluxseg7ei16_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei16_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vint16mf2_t *)(op6), (const int16_t *)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vluxseg7ei16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei16_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vint16mf2_t *)(op6), (vbool32_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (vint16mf2_t)(op11), (vint16mf2_t)(op12), (vint16mf2_t)(op13), (vint16mf2_t)(op14), (const int16_t *)(op15), (vuint16mf2_t)(op16), (size_t)(op17)) +#define vluxseg7ei16_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei16_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vint16mf4_t *)(op6), (const int16_t *)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vluxseg7ei16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei16_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vint16mf4_t *)(op6), (vbool64_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (vint16mf4_t)(op11), (vint16mf4_t)(op12), (vint16mf4_t)(op13), (vint16mf4_t)(op14), (const int16_t *)(op15), (vuint16mf4_t)(op16), (size_t)(op17)) +#define vluxseg7ei16_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei16_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vuint16m1_t *)(op6), (const uint16_t *)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vluxseg7ei16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei16_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vuint16m1_t *)(op6), (vbool16_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (vuint16m1_t)(op11), (vuint16m1_t)(op12), (vuint16m1_t)(op13), (vuint16m1_t)(op14), (const uint16_t *)(op15), (vuint16m1_t)(op16), (size_t)(op17)) +#define vluxseg7ei16_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei16_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vuint16mf2_t *)(op6), (const uint16_t *)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vluxseg7ei16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei16_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vuint16mf2_t *)(op6), (vbool32_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (vuint16mf2_t)(op11), (vuint16mf2_t)(op12), (vuint16mf2_t)(op13), (vuint16mf2_t)(op14), (const uint16_t *)(op15), (vuint16mf2_t)(op16), (size_t)(op17)) +#define vluxseg7ei16_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei16_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vuint16mf4_t *)(op6), (const uint16_t *)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vluxseg7ei16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei16_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vuint16mf4_t *)(op6), (vbool64_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (vuint16mf4_t)(op11), (vuint16mf4_t)(op12), (vuint16mf4_t)(op13), (vuint16mf4_t)(op14), (const uint16_t *)(op15), (vuint16mf4_t)(op16), (size_t)(op17)) +#define vluxseg8ei16_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei16_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vint16m1_t *)(op6), (vint16m1_t *)(op7), (const int16_t *)(op8), (vuint16m1_t)(op9), (size_t)(op10)) +#define vluxseg8ei16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei16_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vint16m1_t *)(op6), (vint16m1_t *)(op7), (vbool16_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (vint16m1_t)(op11), (vint16m1_t)(op12), (vint16m1_t)(op13), (vint16m1_t)(op14), (vint16m1_t)(op15), (vint16m1_t)(op16), (const int16_t *)(op17), (vuint16m1_t)(op18), (size_t)(op19)) +#define vluxseg8ei16_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei16_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vint16mf2_t *)(op6), (vint16mf2_t *)(op7), (const int16_t *)(op8), (vuint16mf2_t)(op9), (size_t)(op10)) +#define vluxseg8ei16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei16_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vint16mf2_t *)(op6), (vint16mf2_t *)(op7), (vbool32_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (vint16mf2_t)(op11), (vint16mf2_t)(op12), (vint16mf2_t)(op13), (vint16mf2_t)(op14), (vint16mf2_t)(op15), (vint16mf2_t)(op16), (const int16_t *)(op17), (vuint16mf2_t)(op18), (size_t)(op19)) +#define vluxseg8ei16_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei16_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vint16mf4_t *)(op6), (vint16mf4_t *)(op7), (const int16_t *)(op8), (vuint16mf4_t)(op9), (size_t)(op10)) +#define vluxseg8ei16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei16_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vint16mf4_t *)(op6), (vint16mf4_t *)(op7), (vbool64_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (vint16mf4_t)(op11), (vint16mf4_t)(op12), (vint16mf4_t)(op13), (vint16mf4_t)(op14), (vint16mf4_t)(op15), (vint16mf4_t)(op16), (const int16_t *)(op17), (vuint16mf4_t)(op18), (size_t)(op19)) +#define vluxseg8ei16_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei16_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vuint16m1_t *)(op6), (vuint16m1_t *)(op7), (const uint16_t *)(op8), (vuint16m1_t)(op9), (size_t)(op10)) +#define vluxseg8ei16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei16_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vuint16m1_t *)(op6), (vuint16m1_t *)(op7), (vbool16_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (vuint16m1_t)(op11), (vuint16m1_t)(op12), (vuint16m1_t)(op13), (vuint16m1_t)(op14), (vuint16m1_t)(op15), (vuint16m1_t)(op16), (const uint16_t *)(op17), (vuint16m1_t)(op18), (size_t)(op19)) +#define vluxseg8ei16_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei16_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vuint16mf2_t *)(op6), (vuint16mf2_t *)(op7), (const uint16_t *)(op8), (vuint16mf2_t)(op9), (size_t)(op10)) +#define vluxseg8ei16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei16_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vuint16mf2_t *)(op6), (vuint16mf2_t *)(op7), (vbool32_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (vuint16mf2_t)(op11), (vuint16mf2_t)(op12), (vuint16mf2_t)(op13), (vuint16mf2_t)(op14), (vuint16mf2_t)(op15), (vuint16mf2_t)(op16), (const uint16_t *)(op17), (vuint16mf2_t)(op18), (size_t)(op19)) +#define vluxseg8ei16_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei16_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vuint16mf4_t *)(op6), (vuint16mf4_t *)(op7), (const uint16_t *)(op8), (vuint16mf4_t)(op9), (size_t)(op10)) +#define vluxseg8ei16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei16_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vuint16mf4_t *)(op6), (vuint16mf4_t *)(op7), (vbool64_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (vuint16mf4_t)(op11), (vuint16mf4_t)(op12), (vuint16mf4_t)(op13), (vuint16mf4_t)(op14), (vuint16mf4_t)(op15), (vuint16mf4_t)(op16), (const uint16_t *)(op17), (vuint16mf4_t)(op18), (size_t)(op19)) +#define vluxseg2ei32_v_i16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (const int16_t *)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vbool16_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (const int16_t *)(op5), (vuint32m2_t)(op6), (size_t)(op7)) +#define vluxseg2ei32_v_i16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_i16m2((vint16m2_t *)(op0), (vint16m2_t *)(op1), (const int16_t *)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_i16m2_m((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vbool8_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (const int16_t *)(op5), (vuint32m4_t)(op6), (size_t)(op7)) +#define vluxseg2ei32_v_i16m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_i16m4((vint16m4_t *)(op0), (vint16m4_t *)(op1), (const int16_t *)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_i16m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_i16m4_m((vint16m4_t *)(op0), (vint16m4_t *)(op1), (vbool4_t)(op2), (vint16m4_t)(op3), (vint16m4_t)(op4), (const int16_t *)(op5), (vuint32m8_t)(op6), (size_t)(op7)) +#define vluxseg2ei32_v_i16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (const int16_t *)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vbool32_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (const int16_t *)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vluxseg2ei32_v_i16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (const int16_t *)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vbool64_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (const int16_t *)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vluxseg2ei32_v_u16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (const uint16_t *)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vbool16_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (const uint16_t *)(op5), (vuint32m2_t)(op6), (size_t)(op7)) +#define vluxseg2ei32_v_u16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_u16m2((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (const uint16_t *)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_u16m2_m((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vbool8_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (const uint16_t *)(op5), (vuint32m4_t)(op6), (size_t)(op7)) +#define vluxseg2ei32_v_u16m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_u16m4((vuint16m4_t *)(op0), (vuint16m4_t *)(op1), (const uint16_t *)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_u16m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_u16m4_m((vuint16m4_t *)(op0), (vuint16m4_t *)(op1), (vbool4_t)(op2), (vuint16m4_t)(op3), (vuint16m4_t)(op4), (const uint16_t *)(op5), (vuint32m8_t)(op6), (size_t)(op7)) +#define vluxseg2ei32_v_u16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (const uint16_t *)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vbool32_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (const uint16_t *)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vluxseg2ei32_v_u16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (const uint16_t *)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vbool64_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (const uint16_t *)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vluxseg3ei32_v_i16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei32_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (const int16_t *)(op3), (vuint32m2_t)(op4), (size_t)(op5)) +#define vluxseg3ei32_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei32_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vbool16_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (const int16_t *)(op7), (vuint32m2_t)(op8), (size_t)(op9)) +#define vluxseg3ei32_v_i16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei32_v_i16m2((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vint16m2_t *)(op2), (const int16_t *)(op3), (vuint32m4_t)(op4), (size_t)(op5)) +#define vluxseg3ei32_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei32_v_i16m2_m((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vint16m2_t *)(op2), (vbool8_t)(op3), (vint16m2_t)(op4), (vint16m2_t)(op5), (vint16m2_t)(op6), (const int16_t *)(op7), (vuint32m4_t)(op8), (size_t)(op9)) +#define vluxseg3ei32_v_i16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei32_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (const int16_t *)(op3), (vuint32m1_t)(op4), (size_t)(op5)) +#define vluxseg3ei32_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei32_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vbool32_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (const int16_t *)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vluxseg3ei32_v_i16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei32_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (const int16_t *)(op3), (vuint32mf2_t)(op4), (size_t)(op5)) +#define vluxseg3ei32_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei32_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vbool64_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (const int16_t *)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vluxseg3ei32_v_u16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei32_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (const uint16_t *)(op3), (vuint32m2_t)(op4), (size_t)(op5)) +#define vluxseg3ei32_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei32_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vbool16_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (const uint16_t *)(op7), (vuint32m2_t)(op8), (size_t)(op9)) +#define vluxseg3ei32_v_u16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei32_v_u16m2((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vuint16m2_t *)(op2), (const uint16_t *)(op3), (vuint32m4_t)(op4), (size_t)(op5)) +#define vluxseg3ei32_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei32_v_u16m2_m((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vuint16m2_t *)(op2), (vbool8_t)(op3), (vuint16m2_t)(op4), (vuint16m2_t)(op5), (vuint16m2_t)(op6), (const uint16_t *)(op7), (vuint32m4_t)(op8), (size_t)(op9)) +#define vluxseg3ei32_v_u16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei32_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (const uint16_t *)(op3), (vuint32m1_t)(op4), (size_t)(op5)) +#define vluxseg3ei32_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei32_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vbool32_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (const uint16_t *)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vluxseg3ei32_v_u16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei32_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (const uint16_t *)(op3), (vuint32mf2_t)(op4), (size_t)(op5)) +#define vluxseg3ei32_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei32_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vbool64_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (const uint16_t *)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vluxseg4ei32_v_i16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei32_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (const int16_t *)(op4), (vuint32m2_t)(op5), (size_t)(op6)) +#define vluxseg4ei32_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei32_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vbool16_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (const int16_t *)(op9), (vuint32m2_t)(op10), (size_t)(op11)) +#define vluxseg4ei32_v_i16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei32_v_i16m2((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vint16m2_t *)(op2), (vint16m2_t *)(op3), (const int16_t *)(op4), (vuint32m4_t)(op5), (size_t)(op6)) +#define vluxseg4ei32_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei32_v_i16m2_m((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vint16m2_t *)(op2), (vint16m2_t *)(op3), (vbool8_t)(op4), (vint16m2_t)(op5), (vint16m2_t)(op6), (vint16m2_t)(op7), (vint16m2_t)(op8), (const int16_t *)(op9), (vuint32m4_t)(op10), (size_t)(op11)) +#define vluxseg4ei32_v_i16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei32_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (const int16_t *)(op4), (vuint32m1_t)(op5), (size_t)(op6)) +#define vluxseg4ei32_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei32_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vbool32_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (const int16_t *)(op9), (vuint32m1_t)(op10), (size_t)(op11)) +#define vluxseg4ei32_v_i16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei32_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (const int16_t *)(op4), (vuint32mf2_t)(op5), (size_t)(op6)) +#define vluxseg4ei32_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei32_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vbool64_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (const int16_t *)(op9), (vuint32mf2_t)(op10), (size_t)(op11)) +#define vluxseg4ei32_v_u16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei32_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (const uint16_t *)(op4), (vuint32m2_t)(op5), (size_t)(op6)) +#define vluxseg4ei32_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei32_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vbool16_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (const uint16_t *)(op9), (vuint32m2_t)(op10), (size_t)(op11)) +#define vluxseg4ei32_v_u16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei32_v_u16m2((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vuint16m2_t *)(op2), (vuint16m2_t *)(op3), (const uint16_t *)(op4), (vuint32m4_t)(op5), (size_t)(op6)) +#define vluxseg4ei32_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei32_v_u16m2_m((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vuint16m2_t *)(op2), (vuint16m2_t *)(op3), (vbool8_t)(op4), (vuint16m2_t)(op5), (vuint16m2_t)(op6), (vuint16m2_t)(op7), (vuint16m2_t)(op8), (const uint16_t *)(op9), (vuint32m4_t)(op10), (size_t)(op11)) +#define vluxseg4ei32_v_u16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei32_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (const uint16_t *)(op4), (vuint32m1_t)(op5), (size_t)(op6)) +#define vluxseg4ei32_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei32_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vbool32_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (const uint16_t *)(op9), (vuint32m1_t)(op10), (size_t)(op11)) +#define vluxseg4ei32_v_u16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei32_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (const uint16_t *)(op4), (vuint32mf2_t)(op5), (size_t)(op6)) +#define vluxseg4ei32_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei32_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vbool64_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (const uint16_t *)(op9), (vuint32mf2_t)(op10), (size_t)(op11)) +#define vluxseg5ei32_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei32_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (const int16_t *)(op5), (vuint32m2_t)(op6), (size_t)(op7)) +#define vluxseg5ei32_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei32_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vbool16_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (const int16_t *)(op11), (vuint32m2_t)(op12), (size_t)(op13)) +#define vluxseg5ei32_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei32_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (const int16_t *)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vluxseg5ei32_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei32_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vbool32_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (const int16_t *)(op11), (vuint32m1_t)(op12), (size_t)(op13)) +#define vluxseg5ei32_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei32_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (const int16_t *)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vluxseg5ei32_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei32_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vbool64_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (const int16_t *)(op11), (vuint32mf2_t)(op12), (size_t)(op13)) +#define vluxseg5ei32_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei32_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (const uint16_t *)(op5), (vuint32m2_t)(op6), (size_t)(op7)) +#define vluxseg5ei32_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei32_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vbool16_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (const uint16_t *)(op11), (vuint32m2_t)(op12), (size_t)(op13)) +#define vluxseg5ei32_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei32_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (const uint16_t *)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vluxseg5ei32_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei32_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vbool32_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (const uint16_t *)(op11), (vuint32m1_t)(op12), (size_t)(op13)) +#define vluxseg5ei32_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei32_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (const uint16_t *)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vluxseg5ei32_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei32_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vbool64_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (const uint16_t *)(op11), (vuint32mf2_t)(op12), (size_t)(op13)) +#define vluxseg6ei32_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei32_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (const int16_t *)(op6), (vuint32m2_t)(op7), (size_t)(op8)) +#define vluxseg6ei32_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei32_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vbool16_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (vint16m1_t)(op11), (vint16m1_t)(op12), (const int16_t *)(op13), (vuint32m2_t)(op14), (size_t)(op15)) +#define vluxseg6ei32_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei32_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (const int16_t *)(op6), (vuint32m1_t)(op7), (size_t)(op8)) +#define vluxseg6ei32_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei32_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vbool32_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (vint16mf2_t)(op11), (vint16mf2_t)(op12), (const int16_t *)(op13), (vuint32m1_t)(op14), (size_t)(op15)) +#define vluxseg6ei32_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei32_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (const int16_t *)(op6), (vuint32mf2_t)(op7), (size_t)(op8)) +#define vluxseg6ei32_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei32_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vbool64_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (vint16mf4_t)(op11), (vint16mf4_t)(op12), (const int16_t *)(op13), (vuint32mf2_t)(op14), (size_t)(op15)) +#define vluxseg6ei32_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei32_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (const uint16_t *)(op6), (vuint32m2_t)(op7), (size_t)(op8)) +#define vluxseg6ei32_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei32_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vbool16_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (vuint16m1_t)(op11), (vuint16m1_t)(op12), (const uint16_t *)(op13), (vuint32m2_t)(op14), (size_t)(op15)) +#define vluxseg6ei32_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei32_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (const uint16_t *)(op6), (vuint32m1_t)(op7), (size_t)(op8)) +#define vluxseg6ei32_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei32_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vbool32_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (vuint16mf2_t)(op11), (vuint16mf2_t)(op12), (const uint16_t *)(op13), (vuint32m1_t)(op14), (size_t)(op15)) +#define vluxseg6ei32_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei32_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (const uint16_t *)(op6), (vuint32mf2_t)(op7), (size_t)(op8)) +#define vluxseg6ei32_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei32_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vbool64_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (vuint16mf4_t)(op11), (vuint16mf4_t)(op12), (const uint16_t *)(op13), (vuint32mf2_t)(op14), (size_t)(op15)) +#define vluxseg7ei32_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei32_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vint16m1_t *)(op6), (const int16_t *)(op7), (vuint32m2_t)(op8), (size_t)(op9)) +#define vluxseg7ei32_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei32_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vint16m1_t *)(op6), (vbool16_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (vint16m1_t)(op11), (vint16m1_t)(op12), (vint16m1_t)(op13), (vint16m1_t)(op14), (const int16_t *)(op15), (vuint32m2_t)(op16), (size_t)(op17)) +#define vluxseg7ei32_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei32_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vint16mf2_t *)(op6), (const int16_t *)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vluxseg7ei32_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei32_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vint16mf2_t *)(op6), (vbool32_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (vint16mf2_t)(op11), (vint16mf2_t)(op12), (vint16mf2_t)(op13), (vint16mf2_t)(op14), (const int16_t *)(op15), (vuint32m1_t)(op16), (size_t)(op17)) +#define vluxseg7ei32_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei32_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vint16mf4_t *)(op6), (const int16_t *)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vluxseg7ei32_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei32_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vint16mf4_t *)(op6), (vbool64_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (vint16mf4_t)(op11), (vint16mf4_t)(op12), (vint16mf4_t)(op13), (vint16mf4_t)(op14), (const int16_t *)(op15), (vuint32mf2_t)(op16), (size_t)(op17)) +#define vluxseg7ei32_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei32_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vuint16m1_t *)(op6), (const uint16_t *)(op7), (vuint32m2_t)(op8), (size_t)(op9)) +#define vluxseg7ei32_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei32_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vuint16m1_t *)(op6), (vbool16_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (vuint16m1_t)(op11), (vuint16m1_t)(op12), (vuint16m1_t)(op13), (vuint16m1_t)(op14), (const uint16_t *)(op15), (vuint32m2_t)(op16), (size_t)(op17)) +#define vluxseg7ei32_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei32_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vuint16mf2_t *)(op6), (const uint16_t *)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vluxseg7ei32_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei32_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vuint16mf2_t *)(op6), (vbool32_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (vuint16mf2_t)(op11), (vuint16mf2_t)(op12), (vuint16mf2_t)(op13), (vuint16mf2_t)(op14), (const uint16_t *)(op15), (vuint32m1_t)(op16), (size_t)(op17)) +#define vluxseg7ei32_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei32_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vuint16mf4_t *)(op6), (const uint16_t *)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vluxseg7ei32_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei32_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vuint16mf4_t *)(op6), (vbool64_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (vuint16mf4_t)(op11), (vuint16mf4_t)(op12), (vuint16mf4_t)(op13), (vuint16mf4_t)(op14), (const uint16_t *)(op15), (vuint32mf2_t)(op16), (size_t)(op17)) +#define vluxseg8ei32_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei32_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vint16m1_t *)(op6), (vint16m1_t *)(op7), (const int16_t *)(op8), (vuint32m2_t)(op9), (size_t)(op10)) +#define vluxseg8ei32_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei32_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vint16m1_t *)(op6), (vint16m1_t *)(op7), (vbool16_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (vint16m1_t)(op11), (vint16m1_t)(op12), (vint16m1_t)(op13), (vint16m1_t)(op14), (vint16m1_t)(op15), (vint16m1_t)(op16), (const int16_t *)(op17), (vuint32m2_t)(op18), (size_t)(op19)) +#define vluxseg8ei32_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei32_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vint16mf2_t *)(op6), (vint16mf2_t *)(op7), (const int16_t *)(op8), (vuint32m1_t)(op9), (size_t)(op10)) +#define vluxseg8ei32_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei32_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vint16mf2_t *)(op6), (vint16mf2_t *)(op7), (vbool32_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (vint16mf2_t)(op11), (vint16mf2_t)(op12), (vint16mf2_t)(op13), (vint16mf2_t)(op14), (vint16mf2_t)(op15), (vint16mf2_t)(op16), (const int16_t *)(op17), (vuint32m1_t)(op18), (size_t)(op19)) +#define vluxseg8ei32_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei32_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vint16mf4_t *)(op6), (vint16mf4_t *)(op7), (const int16_t *)(op8), (vuint32mf2_t)(op9), (size_t)(op10)) +#define vluxseg8ei32_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei32_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vint16mf4_t *)(op6), (vint16mf4_t *)(op7), (vbool64_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (vint16mf4_t)(op11), (vint16mf4_t)(op12), (vint16mf4_t)(op13), (vint16mf4_t)(op14), (vint16mf4_t)(op15), (vint16mf4_t)(op16), (const int16_t *)(op17), (vuint32mf2_t)(op18), (size_t)(op19)) +#define vssseg2e8_v_i8m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e8_v_i8m1((int8_t *)(op0), (ptrdiff_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vssseg2e8_v_i8m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e8_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (ptrdiff_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (size_t)(op5)) +#define vssseg2e8_v_i8m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e8_v_i8m2((int8_t *)(op0), (ptrdiff_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vssseg2e8_v_i8m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e8_v_i8m2_m((vbool4_t)(op0), (int8_t *)(op1), (ptrdiff_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (size_t)(op5)) +#define vssseg2e8_v_i8m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e8_v_i8m4((int8_t *)(op0), (ptrdiff_t)(op1), (vint8m4_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vssseg2e8_v_i8m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e8_v_i8m4_m((vbool2_t)(op0), (int8_t *)(op1), (ptrdiff_t)(op2), (vint8m4_t)(op3), (vint8m4_t)(op4), (size_t)(op5)) +#define vssseg2e8_v_i8mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e8_v_i8mf2((int8_t *)(op0), (ptrdiff_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vssseg2e8_v_i8mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e8_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (ptrdiff_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (size_t)(op5)) +#define vssseg2e8_v_i8mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e8_v_i8mf4((int8_t *)(op0), (ptrdiff_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vssseg2e8_v_i8mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e8_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (ptrdiff_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (size_t)(op5)) +#define vssseg2e8_v_i8mf8(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e8_v_i8mf8((int8_t *)(op0), (ptrdiff_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vssseg2e8_v_i8mf8_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e8_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (ptrdiff_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (size_t)(op5)) +#define vluxseg8ei32_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei32_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vuint16m1_t *)(op6), (vuint16m1_t *)(op7), (const uint16_t *)(op8), (vuint32m2_t)(op9), (size_t)(op10)) +#define vluxseg8ei32_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei32_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vuint16m1_t *)(op6), (vuint16m1_t *)(op7), (vbool16_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (vuint16m1_t)(op11), (vuint16m1_t)(op12), (vuint16m1_t)(op13), (vuint16m1_t)(op14), (vuint16m1_t)(op15), (vuint16m1_t)(op16), (const uint16_t *)(op17), (vuint32m2_t)(op18), (size_t)(op19)) +#define vluxseg8ei32_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei32_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vuint16mf2_t *)(op6), (vuint16mf2_t *)(op7), (const uint16_t *)(op8), (vuint32m1_t)(op9), (size_t)(op10)) +#define vluxseg8ei32_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei32_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vuint16mf2_t *)(op6), (vuint16mf2_t *)(op7), (vbool32_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (vuint16mf2_t)(op11), (vuint16mf2_t)(op12), (vuint16mf2_t)(op13), (vuint16mf2_t)(op14), (vuint16mf2_t)(op15), (vuint16mf2_t)(op16), (const uint16_t *)(op17), (vuint32m1_t)(op18), (size_t)(op19)) +#define vluxseg8ei32_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei32_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vuint16mf4_t *)(op6), (vuint16mf4_t *)(op7), (const uint16_t *)(op8), (vuint32mf2_t)(op9), (size_t)(op10)) +#define vluxseg8ei32_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei32_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vuint16mf4_t *)(op6), (vuint16mf4_t *)(op7), (vbool64_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (vuint16mf4_t)(op11), (vuint16mf4_t)(op12), (vuint16mf4_t)(op13), (vuint16mf4_t)(op14), (vuint16mf4_t)(op15), (vuint16mf4_t)(op16), (const uint16_t *)(op17), (vuint32mf2_t)(op18), (size_t)(op19)) +#define vluxseg2ei64_v_i16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei64_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (const int16_t *)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vluxseg2ei64_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei64_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vbool16_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (const int16_t *)(op5), (vuint64m4_t)(op6), (size_t)(op7)) +#define vluxseg2ei64_v_i16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei64_v_i16m2((vint16m2_t *)(op0), (vint16m2_t *)(op1), (const int16_t *)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vluxseg2ei64_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei64_v_i16m2_m((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vbool8_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (const int16_t *)(op5), (vuint64m8_t)(op6), (size_t)(op7)) +#define vluxseg2ei64_v_i16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei64_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (const int16_t *)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vluxseg2ei64_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei64_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vbool32_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (const int16_t *)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vluxseg2ei64_v_i16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei64_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (const int16_t *)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vluxseg2ei64_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei64_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vbool64_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (const int16_t *)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vluxseg2ei64_v_u16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei64_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (const uint16_t *)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vluxseg2ei64_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei64_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vbool16_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (const uint16_t *)(op5), (vuint64m4_t)(op6), (size_t)(op7)) +#define vluxseg2ei64_v_u16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei64_v_u16m2((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (const uint16_t *)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vluxseg2ei64_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei64_v_u16m2_m((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vbool8_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (const uint16_t *)(op5), (vuint64m8_t)(op6), (size_t)(op7)) +#define vluxseg2ei64_v_u16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei64_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (const uint16_t *)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vluxseg2ei64_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei64_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vbool32_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (const uint16_t *)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vluxseg2ei64_v_u16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei64_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (const uint16_t *)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vluxseg2ei64_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei64_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vbool64_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (const uint16_t *)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vluxseg3ei64_v_i16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei64_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (const int16_t *)(op3), (vuint64m4_t)(op4), (size_t)(op5)) +#define vluxseg3ei64_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei64_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vbool16_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (const int16_t *)(op7), (vuint64m4_t)(op8), (size_t)(op9)) +#define vluxseg3ei64_v_i16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei64_v_i16m2((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vint16m2_t *)(op2), (const int16_t *)(op3), (vuint64m8_t)(op4), (size_t)(op5)) +#define vluxseg3ei64_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei64_v_i16m2_m((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vint16m2_t *)(op2), (vbool8_t)(op3), (vint16m2_t)(op4), (vint16m2_t)(op5), (vint16m2_t)(op6), (const int16_t *)(op7), (vuint64m8_t)(op8), (size_t)(op9)) +#define vluxseg3ei64_v_i16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei64_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (const int16_t *)(op3), (vuint64m2_t)(op4), (size_t)(op5)) +#define vluxseg3ei64_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei64_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vbool32_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (const int16_t *)(op7), (vuint64m2_t)(op8), (size_t)(op9)) +#define vluxseg3ei64_v_i16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei64_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (const int16_t *)(op3), (vuint64m1_t)(op4), (size_t)(op5)) +#define vluxseg3ei64_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei64_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vbool64_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (const int16_t *)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vluxseg3ei64_v_u16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei64_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (const uint16_t *)(op3), (vuint64m4_t)(op4), (size_t)(op5)) +#define vluxseg3ei64_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei64_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vbool16_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (const uint16_t *)(op7), (vuint64m4_t)(op8), (size_t)(op9)) +#define vluxseg3ei64_v_u16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei64_v_u16m2((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vuint16m2_t *)(op2), (const uint16_t *)(op3), (vuint64m8_t)(op4), (size_t)(op5)) +#define vluxseg3ei64_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei64_v_u16m2_m((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vuint16m2_t *)(op2), (vbool8_t)(op3), (vuint16m2_t)(op4), (vuint16m2_t)(op5), (vuint16m2_t)(op6), (const uint16_t *)(op7), (vuint64m8_t)(op8), (size_t)(op9)) +#define vluxseg3ei64_v_u16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei64_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (const uint16_t *)(op3), (vuint64m2_t)(op4), (size_t)(op5)) +#define vluxseg3ei64_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei64_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vbool32_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (const uint16_t *)(op7), (vuint64m2_t)(op8), (size_t)(op9)) +#define vluxseg3ei64_v_u16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei64_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (const uint16_t *)(op3), (vuint64m1_t)(op4), (size_t)(op5)) +#define vluxseg3ei64_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei64_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vbool64_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (const uint16_t *)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vluxseg4ei64_v_i16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei64_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (const int16_t *)(op4), (vuint64m4_t)(op5), (size_t)(op6)) +#define vluxseg4ei64_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei64_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vbool16_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (const int16_t *)(op9), (vuint64m4_t)(op10), (size_t)(op11)) +#define vluxseg4ei64_v_i16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei64_v_i16m2((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vint16m2_t *)(op2), (vint16m2_t *)(op3), (const int16_t *)(op4), (vuint64m8_t)(op5), (size_t)(op6)) +#define vluxseg4ei64_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei64_v_i16m2_m((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vint16m2_t *)(op2), (vint16m2_t *)(op3), (vbool8_t)(op4), (vint16m2_t)(op5), (vint16m2_t)(op6), (vint16m2_t)(op7), (vint16m2_t)(op8), (const int16_t *)(op9), (vuint64m8_t)(op10), (size_t)(op11)) +#define vluxseg4ei64_v_i16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei64_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (const int16_t *)(op4), (vuint64m2_t)(op5), (size_t)(op6)) +#define vluxseg4ei64_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei64_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vbool32_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (const int16_t *)(op9), (vuint64m2_t)(op10), (size_t)(op11)) +#define vluxseg4ei64_v_i16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei64_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (const int16_t *)(op4), (vuint64m1_t)(op5), (size_t)(op6)) +#define vluxseg4ei64_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei64_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vbool64_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (const int16_t *)(op9), (vuint64m1_t)(op10), (size_t)(op11)) +#define vluxseg4ei64_v_u16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei64_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (const uint16_t *)(op4), (vuint64m4_t)(op5), (size_t)(op6)) +#define vluxseg4ei64_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei64_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vbool16_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (const uint16_t *)(op9), (vuint64m4_t)(op10), (size_t)(op11)) +#define vluxseg4ei64_v_u16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei64_v_u16m2((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vuint16m2_t *)(op2), (vuint16m2_t *)(op3), (const uint16_t *)(op4), (vuint64m8_t)(op5), (size_t)(op6)) +#define vluxseg4ei64_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei64_v_u16m2_m((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vuint16m2_t *)(op2), (vuint16m2_t *)(op3), (vbool8_t)(op4), (vuint16m2_t)(op5), (vuint16m2_t)(op6), (vuint16m2_t)(op7), (vuint16m2_t)(op8), (const uint16_t *)(op9), (vuint64m8_t)(op10), (size_t)(op11)) +#define vluxseg4ei64_v_u16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei64_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (const uint16_t *)(op4), (vuint64m2_t)(op5), (size_t)(op6)) +#define vluxseg4ei64_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei64_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vbool32_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (const uint16_t *)(op9), (vuint64m2_t)(op10), (size_t)(op11)) +#define vluxseg4ei64_v_u16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei64_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (const uint16_t *)(op4), (vuint64m1_t)(op5), (size_t)(op6)) +#define vluxseg4ei64_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei64_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vbool64_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (const uint16_t *)(op9), (vuint64m1_t)(op10), (size_t)(op11)) +#define vluxseg5ei64_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei64_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (const int16_t *)(op5), (vuint64m4_t)(op6), (size_t)(op7)) +#define vluxseg5ei64_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei64_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vbool16_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (const int16_t *)(op11), (vuint64m4_t)(op12), (size_t)(op13)) +#define vluxseg5ei64_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei64_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (const int16_t *)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vluxseg5ei64_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei64_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vbool32_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (const int16_t *)(op11), (vuint64m2_t)(op12), (size_t)(op13)) +#define vluxseg5ei64_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei64_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (const int16_t *)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vluxseg5ei64_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei64_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vbool64_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (const int16_t *)(op11), (vuint64m1_t)(op12), (size_t)(op13)) +#define vluxseg5ei64_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei64_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (const uint16_t *)(op5), (vuint64m4_t)(op6), (size_t)(op7)) +#define vluxseg5ei64_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei64_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vbool16_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (const uint16_t *)(op11), (vuint64m4_t)(op12), (size_t)(op13)) +#define vluxseg5ei64_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei64_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (const uint16_t *)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vluxseg5ei64_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei64_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vbool32_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (const uint16_t *)(op11), (vuint64m2_t)(op12), (size_t)(op13)) +#define vluxseg5ei64_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei64_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (const uint16_t *)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vluxseg5ei64_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei64_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vbool64_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (const uint16_t *)(op11), (vuint64m1_t)(op12), (size_t)(op13)) +#define vluxseg6ei64_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei64_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (const int16_t *)(op6), (vuint64m4_t)(op7), (size_t)(op8)) +#define vluxseg6ei64_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei64_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vbool16_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (vint16m1_t)(op11), (vint16m1_t)(op12), (const int16_t *)(op13), (vuint64m4_t)(op14), (size_t)(op15)) +#define vluxseg6ei64_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei64_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (const int16_t *)(op6), (vuint64m2_t)(op7), (size_t)(op8)) +#define vluxseg6ei64_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei64_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vbool32_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (vint16mf2_t)(op11), (vint16mf2_t)(op12), (const int16_t *)(op13), (vuint64m2_t)(op14), (size_t)(op15)) +#define vluxseg6ei64_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei64_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (const int16_t *)(op6), (vuint64m1_t)(op7), (size_t)(op8)) +#define vluxseg6ei64_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei64_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vbool64_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (vint16mf4_t)(op11), (vint16mf4_t)(op12), (const int16_t *)(op13), (vuint64m1_t)(op14), (size_t)(op15)) +#define vssseg2e8_v_u8m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e8_v_u8m1((uint8_t *)(op0), (ptrdiff_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vssseg2e8_v_u8m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e8_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (ptrdiff_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (size_t)(op5)) +#define vssseg2e8_v_u8m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e8_v_u8m2((uint8_t *)(op0), (ptrdiff_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vssseg2e8_v_u8m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e8_v_u8m2_m((vbool4_t)(op0), (uint8_t *)(op1), (ptrdiff_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (size_t)(op5)) +#define vssseg2e8_v_u8m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e8_v_u8m4((uint8_t *)(op0), (ptrdiff_t)(op1), (vuint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vssseg2e8_v_u8m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e8_v_u8m4_m((vbool2_t)(op0), (uint8_t *)(op1), (ptrdiff_t)(op2), (vuint8m4_t)(op3), (vuint8m4_t)(op4), (size_t)(op5)) +#define vssseg2e8_v_u8mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e8_v_u8mf2((uint8_t *)(op0), (ptrdiff_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vssseg2e8_v_u8mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e8_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (ptrdiff_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (size_t)(op5)) +#define vssseg2e8_v_u8mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e8_v_u8mf4((uint8_t *)(op0), (ptrdiff_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vssseg2e8_v_u8mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e8_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (ptrdiff_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (size_t)(op5)) +#define vssseg2e8_v_u8mf8(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e8_v_u8mf8((uint8_t *)(op0), (ptrdiff_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vssseg2e8_v_u8mf8_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e8_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (ptrdiff_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (size_t)(op5)) +#define vluxseg6ei64_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei64_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (const uint16_t *)(op6), (vuint64m4_t)(op7), (size_t)(op8)) +#define vluxseg6ei64_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei64_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vbool16_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (vuint16m1_t)(op11), (vuint16m1_t)(op12), (const uint16_t *)(op13), (vuint64m4_t)(op14), (size_t)(op15)) +#define vluxseg6ei64_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei64_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (const uint16_t *)(op6), (vuint64m2_t)(op7), (size_t)(op8)) +#define vluxseg6ei64_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei64_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vbool32_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (vuint16mf2_t)(op11), (vuint16mf2_t)(op12), (const uint16_t *)(op13), (vuint64m2_t)(op14), (size_t)(op15)) +#define vluxseg6ei64_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei64_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (const uint16_t *)(op6), (vuint64m1_t)(op7), (size_t)(op8)) +#define vluxseg6ei64_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei64_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vbool64_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (vuint16mf4_t)(op11), (vuint16mf4_t)(op12), (const uint16_t *)(op13), (vuint64m1_t)(op14), (size_t)(op15)) +#define vluxseg7ei64_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei64_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vint16m1_t *)(op6), (const int16_t *)(op7), (vuint64m4_t)(op8), (size_t)(op9)) +#define vluxseg7ei64_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei64_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vint16m1_t *)(op6), (vbool16_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (vint16m1_t)(op11), (vint16m1_t)(op12), (vint16m1_t)(op13), (vint16m1_t)(op14), (const int16_t *)(op15), (vuint64m4_t)(op16), (size_t)(op17)) +#define vluxseg7ei64_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei64_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vint16mf2_t *)(op6), (const int16_t *)(op7), (vuint64m2_t)(op8), (size_t)(op9)) +#define vluxseg7ei64_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei64_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vint16mf2_t *)(op6), (vbool32_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (vint16mf2_t)(op11), (vint16mf2_t)(op12), (vint16mf2_t)(op13), (vint16mf2_t)(op14), (const int16_t *)(op15), (vuint64m2_t)(op16), (size_t)(op17)) +#define vluxseg7ei64_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei64_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vint16mf4_t *)(op6), (const int16_t *)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vluxseg7ei64_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei64_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vint16mf4_t *)(op6), (vbool64_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (vint16mf4_t)(op11), (vint16mf4_t)(op12), (vint16mf4_t)(op13), (vint16mf4_t)(op14), (const int16_t *)(op15), (vuint64m1_t)(op16), (size_t)(op17)) +#define vluxseg7ei64_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei64_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vuint16m1_t *)(op6), (const uint16_t *)(op7), (vuint64m4_t)(op8), (size_t)(op9)) +#define vluxseg7ei64_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei64_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vuint16m1_t *)(op6), (vbool16_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (vuint16m1_t)(op11), (vuint16m1_t)(op12), (vuint16m1_t)(op13), (vuint16m1_t)(op14), (const uint16_t *)(op15), (vuint64m4_t)(op16), (size_t)(op17)) +#define vluxseg7ei64_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei64_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vuint16mf2_t *)(op6), (const uint16_t *)(op7), (vuint64m2_t)(op8), (size_t)(op9)) +#define vluxseg7ei64_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei64_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vuint16mf2_t *)(op6), (vbool32_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (vuint16mf2_t)(op11), (vuint16mf2_t)(op12), (vuint16mf2_t)(op13), (vuint16mf2_t)(op14), (const uint16_t *)(op15), (vuint64m2_t)(op16), (size_t)(op17)) +#define vluxseg7ei64_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei64_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vuint16mf4_t *)(op6), (const uint16_t *)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vluxseg7ei64_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei64_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vuint16mf4_t *)(op6), (vbool64_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (vuint16mf4_t)(op11), (vuint16mf4_t)(op12), (vuint16mf4_t)(op13), (vuint16mf4_t)(op14), (const uint16_t *)(op15), (vuint64m1_t)(op16), (size_t)(op17)) +#define vluxseg8ei64_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei64_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vint16m1_t *)(op6), (vint16m1_t *)(op7), (const int16_t *)(op8), (vuint64m4_t)(op9), (size_t)(op10)) +#define vluxseg8ei64_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei64_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vint16m1_t *)(op6), (vint16m1_t *)(op7), (vbool16_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (vint16m1_t)(op11), (vint16m1_t)(op12), (vint16m1_t)(op13), (vint16m1_t)(op14), (vint16m1_t)(op15), (vint16m1_t)(op16), (const int16_t *)(op17), (vuint64m4_t)(op18), (size_t)(op19)) +#define vluxseg8ei64_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei64_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vint16mf2_t *)(op6), (vint16mf2_t *)(op7), (const int16_t *)(op8), (vuint64m2_t)(op9), (size_t)(op10)) +#define vluxseg8ei64_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei64_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vint16mf2_t *)(op6), (vint16mf2_t *)(op7), (vbool32_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (vint16mf2_t)(op11), (vint16mf2_t)(op12), (vint16mf2_t)(op13), (vint16mf2_t)(op14), (vint16mf2_t)(op15), (vint16mf2_t)(op16), (const int16_t *)(op17), (vuint64m2_t)(op18), (size_t)(op19)) +#define vluxseg8ei64_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei64_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vint16mf4_t *)(op6), (vint16mf4_t *)(op7), (const int16_t *)(op8), (vuint64m1_t)(op9), (size_t)(op10)) +#define vluxseg8ei64_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei64_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vint16mf4_t *)(op6), (vint16mf4_t *)(op7), (vbool64_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (vint16mf4_t)(op11), (vint16mf4_t)(op12), (vint16mf4_t)(op13), (vint16mf4_t)(op14), (vint16mf4_t)(op15), (vint16mf4_t)(op16), (const int16_t *)(op17), (vuint64m1_t)(op18), (size_t)(op19)) +#define vluxseg8ei64_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei64_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vuint16m1_t *)(op6), (vuint16m1_t *)(op7), (const uint16_t *)(op8), (vuint64m4_t)(op9), (size_t)(op10)) +#define vluxseg8ei64_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei64_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vuint16m1_t *)(op6), (vuint16m1_t *)(op7), (vbool16_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (vuint16m1_t)(op11), (vuint16m1_t)(op12), (vuint16m1_t)(op13), (vuint16m1_t)(op14), (vuint16m1_t)(op15), (vuint16m1_t)(op16), (const uint16_t *)(op17), (vuint64m4_t)(op18), (size_t)(op19)) +#define vluxseg8ei64_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei64_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vuint16mf2_t *)(op6), (vuint16mf2_t *)(op7), (const uint16_t *)(op8), (vuint64m2_t)(op9), (size_t)(op10)) +#define vluxseg8ei64_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei64_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vuint16mf2_t *)(op6), (vuint16mf2_t *)(op7), (vbool32_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (vuint16mf2_t)(op11), (vuint16mf2_t)(op12), (vuint16mf2_t)(op13), (vuint16mf2_t)(op14), (vuint16mf2_t)(op15), (vuint16mf2_t)(op16), (const uint16_t *)(op17), (vuint64m2_t)(op18), (size_t)(op19)) +#define vluxseg8ei64_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei64_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vuint16mf4_t *)(op6), (vuint16mf4_t *)(op7), (const uint16_t *)(op8), (vuint64m1_t)(op9), (size_t)(op10)) +#define vluxseg8ei64_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei64_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vuint16mf4_t *)(op6), (vuint16mf4_t *)(op7), (vbool64_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (vuint16mf4_t)(op11), (vuint16mf4_t)(op12), (vuint16mf4_t)(op13), (vuint16mf4_t)(op14), (vuint16mf4_t)(op15), (vuint16mf4_t)(op16), (const uint16_t *)(op17), (vuint64m1_t)(op18), (size_t)(op19)) +#define vluxseg2ei8_v_i32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (const int32_t *)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vbool32_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (const int32_t *)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vluxseg2ei8_v_i32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_i32m2((vint32m2_t *)(op0), (vint32m2_t *)(op1), (const int32_t *)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_i32m2_m((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vbool16_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (const int32_t *)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vluxseg2ei8_v_i32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_i32m4((vint32m4_t *)(op0), (vint32m4_t *)(op1), (const int32_t *)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_i32m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_i32m4_m((vint32m4_t *)(op0), (vint32m4_t *)(op1), (vbool8_t)(op2), (vint32m4_t)(op3), (vint32m4_t)(op4), (const int32_t *)(op5), (vuint8m1_t)(op6), (size_t)(op7)) +#define vluxseg2ei8_v_i32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (const int32_t *)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vbool64_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (const int32_t *)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vluxseg2ei8_v_u32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (const uint32_t *)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vbool32_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (const uint32_t *)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vluxseg2ei8_v_u32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_u32m2((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (const uint32_t *)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_u32m2_m((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vbool16_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (const uint32_t *)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vluxseg2ei8_v_u32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_u32m4((vuint32m4_t *)(op0), (vuint32m4_t *)(op1), (const uint32_t *)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_u32m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_u32m4_m((vuint32m4_t *)(op0), (vuint32m4_t *)(op1), (vbool8_t)(op2), (vuint32m4_t)(op3), (vuint32m4_t)(op4), (const uint32_t *)(op5), (vuint8m1_t)(op6), (size_t)(op7)) +#define vluxseg2ei8_v_u32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (const uint32_t *)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vbool64_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (const uint32_t *)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vluxseg3ei8_v_i32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei8_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (const int32_t *)(op3), (vuint8mf4_t)(op4), (size_t)(op5)) +#define vluxseg3ei8_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei8_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vbool32_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (const int32_t *)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vluxseg3ei8_v_i32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei8_v_i32m2((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vint32m2_t *)(op2), (const int32_t *)(op3), (vuint8mf2_t)(op4), (size_t)(op5)) +#define vluxseg3ei8_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei8_v_i32m2_m((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vint32m2_t *)(op2), (vbool16_t)(op3), (vint32m2_t)(op4), (vint32m2_t)(op5), (vint32m2_t)(op6), (const int32_t *)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vluxseg3ei8_v_i32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei8_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (const int32_t *)(op3), (vuint8mf8_t)(op4), (size_t)(op5)) +#define vluxseg3ei8_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei8_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vbool64_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (const int32_t *)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vluxseg3ei8_v_u32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei8_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (const uint32_t *)(op3), (vuint8mf4_t)(op4), (size_t)(op5)) +#define vluxseg3ei8_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei8_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vbool32_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (const uint32_t *)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vluxseg3ei8_v_u32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei8_v_u32m2((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vuint32m2_t *)(op2), (const uint32_t *)(op3), (vuint8mf2_t)(op4), (size_t)(op5)) +#define vluxseg3ei8_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei8_v_u32m2_m((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vuint32m2_t *)(op2), (vbool16_t)(op3), (vuint32m2_t)(op4), (vuint32m2_t)(op5), (vuint32m2_t)(op6), (const uint32_t *)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vluxseg3ei8_v_u32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei8_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (const uint32_t *)(op3), (vuint8mf8_t)(op4), (size_t)(op5)) +#define vluxseg3ei8_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei8_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vbool64_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (const uint32_t *)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vluxseg4ei8_v_i32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei8_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (const int32_t *)(op4), (vuint8mf4_t)(op5), (size_t)(op6)) +#define vluxseg4ei8_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei8_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vbool32_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (const int32_t *)(op9), (vuint8mf4_t)(op10), (size_t)(op11)) +#define vluxseg4ei8_v_i32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei8_v_i32m2((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vint32m2_t *)(op2), (vint32m2_t *)(op3), (const int32_t *)(op4), (vuint8mf2_t)(op5), (size_t)(op6)) +#define vluxseg4ei8_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei8_v_i32m2_m((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vint32m2_t *)(op2), (vint32m2_t *)(op3), (vbool16_t)(op4), (vint32m2_t)(op5), (vint32m2_t)(op6), (vint32m2_t)(op7), (vint32m2_t)(op8), (const int32_t *)(op9), (vuint8mf2_t)(op10), (size_t)(op11)) +#define vluxseg4ei8_v_i32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei8_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (const int32_t *)(op4), (vuint8mf8_t)(op5), (size_t)(op6)) +#define vluxseg4ei8_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei8_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vbool64_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (const int32_t *)(op9), (vuint8mf8_t)(op10), (size_t)(op11)) +#define vsuxseg2ei8_v_i8m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_i8m1((int8_t *)(op0), (vuint8m1_t)(op1), (vint8m1_t)(op2), (vint8m1_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_i8m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_i8m1_m((vbool8_t)(op0), (int8_t *)(op1), (vuint8m1_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (size_t)(op5)) +#define vsuxseg2ei8_v_i8m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_i8m2((int8_t *)(op0), (vuint8m2_t)(op1), (vint8m2_t)(op2), (vint8m2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_i8m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_i8m2_m((vbool4_t)(op0), (int8_t *)(op1), (vuint8m2_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei8_v_i8m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_i8m4((int8_t *)(op0), (vuint8m4_t)(op1), (vint8m4_t)(op2), (vint8m4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_i8m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_i8m4_m((vbool2_t)(op0), (int8_t *)(op1), (vuint8m4_t)(op2), (vint8m4_t)(op3), (vint8m4_t)(op4), (size_t)(op5)) +#define vsuxseg2ei8_v_i8mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_i8mf2((int8_t *)(op0), (vuint8mf2_t)(op1), (vint8mf2_t)(op2), (vint8mf2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_i8mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_i8mf2_m((vbool16_t)(op0), (int8_t *)(op1), (vuint8mf2_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei8_v_i8mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_i8mf4((int8_t *)(op0), (vuint8mf4_t)(op1), (vint8mf4_t)(op2), (vint8mf4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_i8mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_i8mf4_m((vbool32_t)(op0), (int8_t *)(op1), (vuint8mf4_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (size_t)(op5)) +#define vsuxseg2ei8_v_i8mf8(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_i8mf8((int8_t *)(op0), (vuint8mf8_t)(op1), (vint8mf8_t)(op2), (vint8mf8_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_i8mf8_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_i8mf8_m((vbool64_t)(op0), (int8_t *)(op1), (vuint8mf8_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (size_t)(op5)) +#define vluxseg4ei8_v_u32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei8_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (const uint32_t *)(op4), (vuint8mf4_t)(op5), (size_t)(op6)) +#define vluxseg4ei8_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei8_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vbool32_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (const uint32_t *)(op9), (vuint8mf4_t)(op10), (size_t)(op11)) +#define vluxseg4ei8_v_u32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei8_v_u32m2((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vuint32m2_t *)(op2), (vuint32m2_t *)(op3), (const uint32_t *)(op4), (vuint8mf2_t)(op5), (size_t)(op6)) +#define vluxseg4ei8_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei8_v_u32m2_m((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vuint32m2_t *)(op2), (vuint32m2_t *)(op3), (vbool16_t)(op4), (vuint32m2_t)(op5), (vuint32m2_t)(op6), (vuint32m2_t)(op7), (vuint32m2_t)(op8), (const uint32_t *)(op9), (vuint8mf2_t)(op10), (size_t)(op11)) +#define vluxseg4ei8_v_u32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei8_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (const uint32_t *)(op4), (vuint8mf8_t)(op5), (size_t)(op6)) +#define vluxseg4ei8_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei8_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vbool64_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (const uint32_t *)(op9), (vuint8mf8_t)(op10), (size_t)(op11)) +#define vluxseg5ei8_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei8_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (const int32_t *)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vluxseg5ei8_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei8_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vbool32_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (const int32_t *)(op11), (vuint8mf4_t)(op12), (size_t)(op13)) +#define vluxseg5ei8_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei8_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (const int32_t *)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vluxseg5ei8_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei8_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vbool64_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (const int32_t *)(op11), (vuint8mf8_t)(op12), (size_t)(op13)) +#define vluxseg5ei8_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei8_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (const uint32_t *)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vluxseg5ei8_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei8_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vbool32_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (const uint32_t *)(op11), (vuint8mf4_t)(op12), (size_t)(op13)) +#define vluxseg5ei8_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei8_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (const uint32_t *)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vluxseg5ei8_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei8_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vbool64_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (const uint32_t *)(op11), (vuint8mf8_t)(op12), (size_t)(op13)) +#define vluxseg6ei8_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei8_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (const int32_t *)(op6), (vuint8mf4_t)(op7), (size_t)(op8)) +#define vluxseg6ei8_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei8_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vbool32_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (vint32m1_t)(op11), (vint32m1_t)(op12), (const int32_t *)(op13), (vuint8mf4_t)(op14), (size_t)(op15)) +#define vluxseg6ei8_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei8_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (const int32_t *)(op6), (vuint8mf8_t)(op7), (size_t)(op8)) +#define vluxseg6ei8_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei8_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vbool64_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (vint32mf2_t)(op11), (vint32mf2_t)(op12), (const int32_t *)(op13), (vuint8mf8_t)(op14), (size_t)(op15)) +#define vluxseg6ei8_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei8_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (const uint32_t *)(op6), (vuint8mf4_t)(op7), (size_t)(op8)) +#define vluxseg6ei8_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei8_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vbool32_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (vuint32m1_t)(op11), (vuint32m1_t)(op12), (const uint32_t *)(op13), (vuint8mf4_t)(op14), (size_t)(op15)) +#define vluxseg6ei8_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei8_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (const uint32_t *)(op6), (vuint8mf8_t)(op7), (size_t)(op8)) +#define vluxseg6ei8_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei8_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vbool64_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (vuint32mf2_t)(op11), (vuint32mf2_t)(op12), (const uint32_t *)(op13), (vuint8mf8_t)(op14), (size_t)(op15)) +#define vluxseg7ei8_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei8_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vint32m1_t *)(op6), (const int32_t *)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vluxseg7ei8_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei8_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vint32m1_t *)(op6), (vbool32_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (vint32m1_t)(op11), (vint32m1_t)(op12), (vint32m1_t)(op13), (vint32m1_t)(op14), (const int32_t *)(op15), (vuint8mf4_t)(op16), (size_t)(op17)) +#define vluxseg7ei8_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei8_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vint32mf2_t *)(op6), (const int32_t *)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vluxseg7ei8_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei8_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vint32mf2_t *)(op6), (vbool64_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (vint32mf2_t)(op11), (vint32mf2_t)(op12), (vint32mf2_t)(op13), (vint32mf2_t)(op14), (const int32_t *)(op15), (vuint8mf8_t)(op16), (size_t)(op17)) +#define vluxseg7ei8_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei8_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vuint32m1_t *)(op6), (const uint32_t *)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vluxseg7ei8_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei8_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vuint32m1_t *)(op6), (vbool32_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (vuint32m1_t)(op11), (vuint32m1_t)(op12), (vuint32m1_t)(op13), (vuint32m1_t)(op14), (const uint32_t *)(op15), (vuint8mf4_t)(op16), (size_t)(op17)) +#define vluxseg7ei8_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei8_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vuint32mf2_t *)(op6), (const uint32_t *)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vluxseg7ei8_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei8_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vuint32mf2_t *)(op6), (vbool64_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (vuint32mf2_t)(op11), (vuint32mf2_t)(op12), (vuint32mf2_t)(op13), (vuint32mf2_t)(op14), (const uint32_t *)(op15), (vuint8mf8_t)(op16), (size_t)(op17)) +#define vluxseg8ei8_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei8_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vint32m1_t *)(op6), (vint32m1_t *)(op7), (const int32_t *)(op8), (vuint8mf4_t)(op9), (size_t)(op10)) +#define vluxseg8ei8_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei8_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vint32m1_t *)(op6), (vint32m1_t *)(op7), (vbool32_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (vint32m1_t)(op11), (vint32m1_t)(op12), (vint32m1_t)(op13), (vint32m1_t)(op14), (vint32m1_t)(op15), (vint32m1_t)(op16), (const int32_t *)(op17), (vuint8mf4_t)(op18), (size_t)(op19)) +#define vluxseg8ei8_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei8_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vint32mf2_t *)(op6), (vint32mf2_t *)(op7), (const int32_t *)(op8), (vuint8mf8_t)(op9), (size_t)(op10)) +#define vluxseg8ei8_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei8_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vint32mf2_t *)(op6), (vint32mf2_t *)(op7), (vbool64_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (vint32mf2_t)(op11), (vint32mf2_t)(op12), (vint32mf2_t)(op13), (vint32mf2_t)(op14), (vint32mf2_t)(op15), (vint32mf2_t)(op16), (const int32_t *)(op17), (vuint8mf8_t)(op18), (size_t)(op19)) +#define vluxseg8ei8_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei8_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vuint32m1_t *)(op6), (vuint32m1_t *)(op7), (const uint32_t *)(op8), (vuint8mf4_t)(op9), (size_t)(op10)) +#define vluxseg8ei8_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei8_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vuint32m1_t *)(op6), (vuint32m1_t *)(op7), (vbool32_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (vuint32m1_t)(op11), (vuint32m1_t)(op12), (vuint32m1_t)(op13), (vuint32m1_t)(op14), (vuint32m1_t)(op15), (vuint32m1_t)(op16), (const uint32_t *)(op17), (vuint8mf4_t)(op18), (size_t)(op19)) +#define vluxseg8ei8_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei8_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vuint32mf2_t *)(op6), (vuint32mf2_t *)(op7), (const uint32_t *)(op8), (vuint8mf8_t)(op9), (size_t)(op10)) +#define vluxseg8ei8_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei8_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vuint32mf2_t *)(op6), (vuint32mf2_t *)(op7), (vbool64_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (vuint32mf2_t)(op11), (vuint32mf2_t)(op12), (vuint32mf2_t)(op13), (vuint32mf2_t)(op14), (vuint32mf2_t)(op15), (vuint32mf2_t)(op16), (const uint32_t *)(op17), (vuint8mf8_t)(op18), (size_t)(op19)) +#define vluxseg2ei16_v_i32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (const int32_t *)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vbool32_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (const int32_t *)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vluxseg2ei16_v_i32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_i32m2((vint32m2_t *)(op0), (vint32m2_t *)(op1), (const int32_t *)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_i32m2_m((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vbool16_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (const int32_t *)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vluxseg2ei16_v_i32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_i32m4((vint32m4_t *)(op0), (vint32m4_t *)(op1), (const int32_t *)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_i32m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_i32m4_m((vint32m4_t *)(op0), (vint32m4_t *)(op1), (vbool8_t)(op2), (vint32m4_t)(op3), (vint32m4_t)(op4), (const int32_t *)(op5), (vuint16m2_t)(op6), (size_t)(op7)) +#define vluxseg2ei16_v_i32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (const int32_t *)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vbool64_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (const int32_t *)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vsuxseg2ei8_v_u8m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_u8m1((uint8_t *)(op0), (vuint8m1_t)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_u8m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_u8m1_m((vbool8_t)(op0), (uint8_t *)(op1), (vuint8m1_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (size_t)(op5)) +#define vsuxseg2ei8_v_u8m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_u8m2((uint8_t *)(op0), (vuint8m2_t)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_u8m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_u8m2_m((vbool4_t)(op0), (uint8_t *)(op1), (vuint8m2_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei8_v_u8m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_u8m4((uint8_t *)(op0), (vuint8m4_t)(op1), (vuint8m4_t)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_u8m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_u8m4_m((vbool2_t)(op0), (uint8_t *)(op1), (vuint8m4_t)(op2), (vuint8m4_t)(op3), (vuint8m4_t)(op4), (size_t)(op5)) +#define vsuxseg2ei8_v_u8mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_u8mf2((uint8_t *)(op0), (vuint8mf2_t)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_u8mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_u8mf2_m((vbool16_t)(op0), (uint8_t *)(op1), (vuint8mf2_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei8_v_u8mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_u8mf4((uint8_t *)(op0), (vuint8mf4_t)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_u8mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_u8mf4_m((vbool32_t)(op0), (uint8_t *)(op1), (vuint8mf4_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (size_t)(op5)) +#define vsuxseg2ei8_v_u8mf8(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_u8mf8((uint8_t *)(op0), (vuint8mf8_t)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_u8mf8_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_u8mf8_m((vbool64_t)(op0), (uint8_t *)(op1), (vuint8mf8_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (size_t)(op5)) +#define vluxseg2ei16_v_u32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (const uint32_t *)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vbool32_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (const uint32_t *)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vluxseg2ei16_v_u32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_u32m2((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (const uint32_t *)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_u32m2_m((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vbool16_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (const uint32_t *)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vluxseg2ei16_v_u32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_u32m4((vuint32m4_t *)(op0), (vuint32m4_t *)(op1), (const uint32_t *)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_u32m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_u32m4_m((vuint32m4_t *)(op0), (vuint32m4_t *)(op1), (vbool8_t)(op2), (vuint32m4_t)(op3), (vuint32m4_t)(op4), (const uint32_t *)(op5), (vuint16m2_t)(op6), (size_t)(op7)) +#define vluxseg2ei16_v_u32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (const uint32_t *)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vbool64_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (const uint32_t *)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vluxseg3ei16_v_i32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei16_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (const int32_t *)(op3), (vuint16mf2_t)(op4), (size_t)(op5)) +#define vluxseg3ei16_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei16_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vbool32_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (const int32_t *)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vluxseg3ei16_v_i32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei16_v_i32m2((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vint32m2_t *)(op2), (const int32_t *)(op3), (vuint16m1_t)(op4), (size_t)(op5)) +#define vluxseg3ei16_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei16_v_i32m2_m((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vint32m2_t *)(op2), (vbool16_t)(op3), (vint32m2_t)(op4), (vint32m2_t)(op5), (vint32m2_t)(op6), (const int32_t *)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vluxseg3ei16_v_i32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei16_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (const int32_t *)(op3), (vuint16mf4_t)(op4), (size_t)(op5)) +#define vluxseg3ei16_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei16_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vbool64_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (const int32_t *)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vluxseg3ei16_v_u32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei16_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (const uint32_t *)(op3), (vuint16mf2_t)(op4), (size_t)(op5)) +#define vluxseg3ei16_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei16_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vbool32_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (const uint32_t *)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vluxseg3ei16_v_u32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei16_v_u32m2((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vuint32m2_t *)(op2), (const uint32_t *)(op3), (vuint16m1_t)(op4), (size_t)(op5)) +#define vluxseg3ei16_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei16_v_u32m2_m((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vuint32m2_t *)(op2), (vbool16_t)(op3), (vuint32m2_t)(op4), (vuint32m2_t)(op5), (vuint32m2_t)(op6), (const uint32_t *)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vluxseg3ei16_v_u32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei16_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (const uint32_t *)(op3), (vuint16mf4_t)(op4), (size_t)(op5)) +#define vluxseg3ei16_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei16_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vbool64_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (const uint32_t *)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vluxseg4ei16_v_i32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei16_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (const int32_t *)(op4), (vuint16mf2_t)(op5), (size_t)(op6)) +#define vluxseg4ei16_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei16_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vbool32_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (const int32_t *)(op9), (vuint16mf2_t)(op10), (size_t)(op11)) +#define vluxseg4ei16_v_i32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei16_v_i32m2((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vint32m2_t *)(op2), (vint32m2_t *)(op3), (const int32_t *)(op4), (vuint16m1_t)(op5), (size_t)(op6)) +#define vluxseg4ei16_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei16_v_i32m2_m((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vint32m2_t *)(op2), (vint32m2_t *)(op3), (vbool16_t)(op4), (vint32m2_t)(op5), (vint32m2_t)(op6), (vint32m2_t)(op7), (vint32m2_t)(op8), (const int32_t *)(op9), (vuint16m1_t)(op10), (size_t)(op11)) +#define vluxseg4ei16_v_i32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei16_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (const int32_t *)(op4), (vuint16mf4_t)(op5), (size_t)(op6)) +#define vluxseg4ei16_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei16_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vbool64_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (const int32_t *)(op9), (vuint16mf4_t)(op10), (size_t)(op11)) +#define vluxseg4ei16_v_u32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei16_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (const uint32_t *)(op4), (vuint16mf2_t)(op5), (size_t)(op6)) +#define vluxseg4ei16_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei16_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vbool32_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (const uint32_t *)(op9), (vuint16mf2_t)(op10), (size_t)(op11)) +#define vluxseg4ei16_v_u32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei16_v_u32m2((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vuint32m2_t *)(op2), (vuint32m2_t *)(op3), (const uint32_t *)(op4), (vuint16m1_t)(op5), (size_t)(op6)) +#define vluxseg4ei16_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei16_v_u32m2_m((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vuint32m2_t *)(op2), (vuint32m2_t *)(op3), (vbool16_t)(op4), (vuint32m2_t)(op5), (vuint32m2_t)(op6), (vuint32m2_t)(op7), (vuint32m2_t)(op8), (const uint32_t *)(op9), (vuint16m1_t)(op10), (size_t)(op11)) +#define vluxseg4ei16_v_u32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei16_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (const uint32_t *)(op4), (vuint16mf4_t)(op5), (size_t)(op6)) +#define vluxseg4ei16_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei16_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vbool64_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (const uint32_t *)(op9), (vuint16mf4_t)(op10), (size_t)(op11)) +#define vluxseg5ei16_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei16_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (const int32_t *)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vluxseg5ei16_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei16_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vbool32_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (const int32_t *)(op11), (vuint16mf2_t)(op12), (size_t)(op13)) +#define vluxseg5ei16_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei16_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (const int32_t *)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vluxseg5ei16_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei16_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vbool64_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (const int32_t *)(op11), (vuint16mf4_t)(op12), (size_t)(op13)) +#define vluxseg5ei16_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei16_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (const uint32_t *)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vluxseg5ei16_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei16_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vbool32_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (const uint32_t *)(op11), (vuint16mf2_t)(op12), (size_t)(op13)) +#define vluxseg5ei16_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei16_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (const uint32_t *)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vluxseg5ei16_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei16_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vbool64_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (const uint32_t *)(op11), (vuint16mf4_t)(op12), (size_t)(op13)) +#define vluxseg6ei16_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei16_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (const int32_t *)(op6), (vuint16mf2_t)(op7), (size_t)(op8)) +#define vluxseg6ei16_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei16_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vbool32_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (vint32m1_t)(op11), (vint32m1_t)(op12), (const int32_t *)(op13), (vuint16mf2_t)(op14), (size_t)(op15)) +#define vluxseg6ei16_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei16_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (const int32_t *)(op6), (vuint16mf4_t)(op7), (size_t)(op8)) +#define vluxseg6ei16_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei16_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vbool64_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (vint32mf2_t)(op11), (vint32mf2_t)(op12), (const int32_t *)(op13), (vuint16mf4_t)(op14), (size_t)(op15)) +#define vluxseg6ei16_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei16_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (const uint32_t *)(op6), (vuint16mf2_t)(op7), (size_t)(op8)) +#define vluxseg6ei16_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei16_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vbool32_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (vuint32m1_t)(op11), (vuint32m1_t)(op12), (const uint32_t *)(op13), (vuint16mf2_t)(op14), (size_t)(op15)) +#define vluxseg6ei16_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei16_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (const uint32_t *)(op6), (vuint16mf4_t)(op7), (size_t)(op8)) +#define vluxseg6ei16_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei16_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vbool64_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (vuint32mf2_t)(op11), (vuint32mf2_t)(op12), (const uint32_t *)(op13), (vuint16mf4_t)(op14), (size_t)(op15)) +#define vluxseg7ei16_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei16_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vint32m1_t *)(op6), (const int32_t *)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vluxseg7ei16_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei16_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vint32m1_t *)(op6), (vbool32_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (vint32m1_t)(op11), (vint32m1_t)(op12), (vint32m1_t)(op13), (vint32m1_t)(op14), (const int32_t *)(op15), (vuint16mf2_t)(op16), (size_t)(op17)) +#define vluxseg7ei16_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei16_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vint32mf2_t *)(op6), (const int32_t *)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vluxseg7ei16_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei16_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vint32mf2_t *)(op6), (vbool64_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (vint32mf2_t)(op11), (vint32mf2_t)(op12), (vint32mf2_t)(op13), (vint32mf2_t)(op14), (const int32_t *)(op15), (vuint16mf4_t)(op16), (size_t)(op17)) +#define vluxseg7ei16_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei16_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vuint32m1_t *)(op6), (const uint32_t *)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vluxseg7ei16_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei16_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vuint32m1_t *)(op6), (vbool32_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (vuint32m1_t)(op11), (vuint32m1_t)(op12), (vuint32m1_t)(op13), (vuint32m1_t)(op14), (const uint32_t *)(op15), (vuint16mf2_t)(op16), (size_t)(op17)) +#define vluxseg7ei16_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei16_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vuint32mf2_t *)(op6), (const uint32_t *)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vluxseg7ei16_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei16_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vuint32mf2_t *)(op6), (vbool64_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (vuint32mf2_t)(op11), (vuint32mf2_t)(op12), (vuint32mf2_t)(op13), (vuint32mf2_t)(op14), (const uint32_t *)(op15), (vuint16mf4_t)(op16), (size_t)(op17)) +#define vluxseg8ei16_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei16_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vint32m1_t *)(op6), (vint32m1_t *)(op7), (const int32_t *)(op8), (vuint16mf2_t)(op9), (size_t)(op10)) +#define vluxseg8ei16_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei16_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vint32m1_t *)(op6), (vint32m1_t *)(op7), (vbool32_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (vint32m1_t)(op11), (vint32m1_t)(op12), (vint32m1_t)(op13), (vint32m1_t)(op14), (vint32m1_t)(op15), (vint32m1_t)(op16), (const int32_t *)(op17), (vuint16mf2_t)(op18), (size_t)(op19)) +#define vluxseg8ei16_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei16_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vint32mf2_t *)(op6), (vint32mf2_t *)(op7), (const int32_t *)(op8), (vuint16mf4_t)(op9), (size_t)(op10)) +#define vluxseg8ei16_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei16_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vint32mf2_t *)(op6), (vint32mf2_t *)(op7), (vbool64_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (vint32mf2_t)(op11), (vint32mf2_t)(op12), (vint32mf2_t)(op13), (vint32mf2_t)(op14), (vint32mf2_t)(op15), (vint32mf2_t)(op16), (const int32_t *)(op17), (vuint16mf4_t)(op18), (size_t)(op19)) +#define vluxseg8ei16_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei16_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vuint32m1_t *)(op6), (vuint32m1_t *)(op7), (const uint32_t *)(op8), (vuint16mf2_t)(op9), (size_t)(op10)) +#define vluxseg8ei16_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei16_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vuint32m1_t *)(op6), (vuint32m1_t *)(op7), (vbool32_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (vuint32m1_t)(op11), (vuint32m1_t)(op12), (vuint32m1_t)(op13), (vuint32m1_t)(op14), (vuint32m1_t)(op15), (vuint32m1_t)(op16), (const uint32_t *)(op17), (vuint16mf2_t)(op18), (size_t)(op19)) +#define vluxseg8ei16_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei16_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vuint32mf2_t *)(op6), (vuint32mf2_t *)(op7), (const uint32_t *)(op8), (vuint16mf4_t)(op9), (size_t)(op10)) +#define vluxseg8ei16_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei16_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vuint32mf2_t *)(op6), (vuint32mf2_t *)(op7), (vbool64_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (vuint32mf2_t)(op11), (vuint32mf2_t)(op12), (vuint32mf2_t)(op13), (vuint32mf2_t)(op14), (vuint32mf2_t)(op15), (vuint32mf2_t)(op16), (const uint32_t *)(op17), (vuint16mf4_t)(op18), (size_t)(op19)) +#define vluxseg2ei32_v_i32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (const int32_t *)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vbool32_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (const int32_t *)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vluxseg2ei32_v_i32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_i32m2((vint32m2_t *)(op0), (vint32m2_t *)(op1), (const int32_t *)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_i32m2_m((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vbool16_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (const int32_t *)(op5), (vuint32m2_t)(op6), (size_t)(op7)) +#define vluxseg2ei32_v_i32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_i32m4((vint32m4_t *)(op0), (vint32m4_t *)(op1), (const int32_t *)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_i32m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_i32m4_m((vint32m4_t *)(op0), (vint32m4_t *)(op1), (vbool8_t)(op2), (vint32m4_t)(op3), (vint32m4_t)(op4), (const int32_t *)(op5), (vuint32m4_t)(op6), (size_t)(op7)) +#define vluxseg2ei32_v_i32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (const int32_t *)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vbool64_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (const int32_t *)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vluxseg2ei32_v_u32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (const uint32_t *)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vbool32_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (const uint32_t *)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vluxseg2ei32_v_u32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_u32m2((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (const uint32_t *)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_u32m2_m((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vbool16_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (const uint32_t *)(op5), (vuint32m2_t)(op6), (size_t)(op7)) +#define vluxseg2ei32_v_u32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_u32m4((vuint32m4_t *)(op0), (vuint32m4_t *)(op1), (const uint32_t *)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_u32m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_u32m4_m((vuint32m4_t *)(op0), (vuint32m4_t *)(op1), (vbool8_t)(op2), (vuint32m4_t)(op3), (vuint32m4_t)(op4), (const uint32_t *)(op5), (vuint32m4_t)(op6), (size_t)(op7)) +#define vluxseg2ei32_v_u32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (const uint32_t *)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vbool64_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (const uint32_t *)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vluxseg3ei32_v_i32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei32_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (const int32_t *)(op3), (vuint32m1_t)(op4), (size_t)(op5)) +#define vluxseg3ei32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei32_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vbool32_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (const int32_t *)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vluxseg3ei32_v_i32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei32_v_i32m2((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vint32m2_t *)(op2), (const int32_t *)(op3), (vuint32m2_t)(op4), (size_t)(op5)) +#define vluxseg3ei32_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei32_v_i32m2_m((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vint32m2_t *)(op2), (vbool16_t)(op3), (vint32m2_t)(op4), (vint32m2_t)(op5), (vint32m2_t)(op6), (const int32_t *)(op7), (vuint32m2_t)(op8), (size_t)(op9)) +#define vluxseg3ei32_v_i32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei32_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (const int32_t *)(op3), (vuint32mf2_t)(op4), (size_t)(op5)) +#define vluxseg3ei32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei32_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vbool64_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (const int32_t *)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vluxseg3ei32_v_u32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei32_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (const uint32_t *)(op3), (vuint32m1_t)(op4), (size_t)(op5)) +#define vluxseg3ei32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei32_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vbool32_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (const uint32_t *)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vluxseg3ei32_v_u32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei32_v_u32m2((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vuint32m2_t *)(op2), (const uint32_t *)(op3), (vuint32m2_t)(op4), (size_t)(op5)) +#define vluxseg3ei32_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei32_v_u32m2_m((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vuint32m2_t *)(op2), (vbool16_t)(op3), (vuint32m2_t)(op4), (vuint32m2_t)(op5), (vuint32m2_t)(op6), (const uint32_t *)(op7), (vuint32m2_t)(op8), (size_t)(op9)) +#define vluxseg3ei32_v_u32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei32_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (const uint32_t *)(op3), (vuint32mf2_t)(op4), (size_t)(op5)) +#define vluxseg3ei32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei32_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vbool64_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (const uint32_t *)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vluxseg4ei32_v_i32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei32_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (const int32_t *)(op4), (vuint32m1_t)(op5), (size_t)(op6)) +#define vluxseg4ei32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei32_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vbool32_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (const int32_t *)(op9), (vuint32m1_t)(op10), (size_t)(op11)) +#define vluxseg4ei32_v_i32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei32_v_i32m2((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vint32m2_t *)(op2), (vint32m2_t *)(op3), (const int32_t *)(op4), (vuint32m2_t)(op5), (size_t)(op6)) +#define vluxseg4ei32_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei32_v_i32m2_m((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vint32m2_t *)(op2), (vint32m2_t *)(op3), (vbool16_t)(op4), (vint32m2_t)(op5), (vint32m2_t)(op6), (vint32m2_t)(op7), (vint32m2_t)(op8), (const int32_t *)(op9), (vuint32m2_t)(op10), (size_t)(op11)) +#define vluxseg4ei32_v_i32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei32_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (const int32_t *)(op4), (vuint32mf2_t)(op5), (size_t)(op6)) +#define vluxseg4ei32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei32_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vbool64_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (const int32_t *)(op9), (vuint32mf2_t)(op10), (size_t)(op11)) +#define vluxseg4ei32_v_u32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei32_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (const uint32_t *)(op4), (vuint32m1_t)(op5), (size_t)(op6)) +#define vluxseg4ei32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei32_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vbool32_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (const uint32_t *)(op9), (vuint32m1_t)(op10), (size_t)(op11)) +#define vluxseg4ei32_v_u32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei32_v_u32m2((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vuint32m2_t *)(op2), (vuint32m2_t *)(op3), (const uint32_t *)(op4), (vuint32m2_t)(op5), (size_t)(op6)) +#define vluxseg4ei32_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei32_v_u32m2_m((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vuint32m2_t *)(op2), (vuint32m2_t *)(op3), (vbool16_t)(op4), (vuint32m2_t)(op5), (vuint32m2_t)(op6), (vuint32m2_t)(op7), (vuint32m2_t)(op8), (const uint32_t *)(op9), (vuint32m2_t)(op10), (size_t)(op11)) +#define vluxseg4ei32_v_u32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei32_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (const uint32_t *)(op4), (vuint32mf2_t)(op5), (size_t)(op6)) +#define vluxseg4ei32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei32_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vbool64_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (const uint32_t *)(op9), (vuint32mf2_t)(op10), (size_t)(op11)) +#define vluxseg5ei32_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei32_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (const int32_t *)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vluxseg5ei32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei32_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vbool32_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (const int32_t *)(op11), (vuint32m1_t)(op12), (size_t)(op13)) +#define vluxseg5ei32_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei32_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (const int32_t *)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vluxseg5ei32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei32_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vbool64_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (const int32_t *)(op11), (vuint32mf2_t)(op12), (size_t)(op13)) +#define vluxseg5ei32_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei32_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (const uint32_t *)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vluxseg5ei32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei32_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vbool32_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (const uint32_t *)(op11), (vuint32m1_t)(op12), (size_t)(op13)) +#define vluxseg5ei32_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei32_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (const uint32_t *)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vluxseg5ei32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei32_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vbool64_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (const uint32_t *)(op11), (vuint32mf2_t)(op12), (size_t)(op13)) +#define vluxseg6ei32_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei32_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (const int32_t *)(op6), (vuint32m1_t)(op7), (size_t)(op8)) +#define vluxseg6ei32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei32_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vbool32_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (vint32m1_t)(op11), (vint32m1_t)(op12), (const int32_t *)(op13), (vuint32m1_t)(op14), (size_t)(op15)) +#define vluxseg6ei32_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei32_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (const int32_t *)(op6), (vuint32mf2_t)(op7), (size_t)(op8)) +#define vluxseg6ei32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei32_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vbool64_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (vint32mf2_t)(op11), (vint32mf2_t)(op12), (const int32_t *)(op13), (vuint32mf2_t)(op14), (size_t)(op15)) +#define vluxseg6ei32_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei32_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (const uint32_t *)(op6), (vuint32m1_t)(op7), (size_t)(op8)) +#define vluxseg6ei32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei32_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vbool32_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (vuint32m1_t)(op11), (vuint32m1_t)(op12), (const uint32_t *)(op13), (vuint32m1_t)(op14), (size_t)(op15)) +#define vluxseg6ei32_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei32_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (const uint32_t *)(op6), (vuint32mf2_t)(op7), (size_t)(op8)) +#define vluxseg6ei32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei32_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vbool64_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (vuint32mf2_t)(op11), (vuint32mf2_t)(op12), (const uint32_t *)(op13), (vuint32mf2_t)(op14), (size_t)(op15)) +#define vluxseg7ei32_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei32_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vint32m1_t *)(op6), (const int32_t *)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vluxseg7ei32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei32_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vint32m1_t *)(op6), (vbool32_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (vint32m1_t)(op11), (vint32m1_t)(op12), (vint32m1_t)(op13), (vint32m1_t)(op14), (const int32_t *)(op15), (vuint32m1_t)(op16), (size_t)(op17)) +#define vluxseg7ei32_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei32_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vint32mf2_t *)(op6), (const int32_t *)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vluxseg7ei32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei32_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vint32mf2_t *)(op6), (vbool64_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (vint32mf2_t)(op11), (vint32mf2_t)(op12), (vint32mf2_t)(op13), (vint32mf2_t)(op14), (const int32_t *)(op15), (vuint32mf2_t)(op16), (size_t)(op17)) +#define vluxseg7ei32_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei32_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vuint32m1_t *)(op6), (const uint32_t *)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vluxseg7ei32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei32_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vuint32m1_t *)(op6), (vbool32_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (vuint32m1_t)(op11), (vuint32m1_t)(op12), (vuint32m1_t)(op13), (vuint32m1_t)(op14), (const uint32_t *)(op15), (vuint32m1_t)(op16), (size_t)(op17)) +#define vluxseg7ei32_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei32_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vuint32mf2_t *)(op6), (const uint32_t *)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vluxseg7ei32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei32_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vuint32mf2_t *)(op6), (vbool64_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (vuint32mf2_t)(op11), (vuint32mf2_t)(op12), (vuint32mf2_t)(op13), (vuint32mf2_t)(op14), (const uint32_t *)(op15), (vuint32mf2_t)(op16), (size_t)(op17)) +#define vluxseg8ei32_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei32_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vint32m1_t *)(op6), (vint32m1_t *)(op7), (const int32_t *)(op8), (vuint32m1_t)(op9), (size_t)(op10)) +#define vluxseg8ei32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei32_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vint32m1_t *)(op6), (vint32m1_t *)(op7), (vbool32_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (vint32m1_t)(op11), (vint32m1_t)(op12), (vint32m1_t)(op13), (vint32m1_t)(op14), (vint32m1_t)(op15), (vint32m1_t)(op16), (const int32_t *)(op17), (vuint32m1_t)(op18), (size_t)(op19)) +#define vluxseg8ei32_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei32_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vint32mf2_t *)(op6), (vint32mf2_t *)(op7), (const int32_t *)(op8), (vuint32mf2_t)(op9), (size_t)(op10)) +#define vluxseg8ei32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei32_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vint32mf2_t *)(op6), (vint32mf2_t *)(op7), (vbool64_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (vint32mf2_t)(op11), (vint32mf2_t)(op12), (vint32mf2_t)(op13), (vint32mf2_t)(op14), (vint32mf2_t)(op15), (vint32mf2_t)(op16), (const int32_t *)(op17), (vuint32mf2_t)(op18), (size_t)(op19)) +#define vluxseg8ei32_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei32_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vuint32m1_t *)(op6), (vuint32m1_t *)(op7), (const uint32_t *)(op8), (vuint32m1_t)(op9), (size_t)(op10)) +#define vluxseg8ei32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei32_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vuint32m1_t *)(op6), (vuint32m1_t *)(op7), (vbool32_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (vuint32m1_t)(op11), (vuint32m1_t)(op12), (vuint32m1_t)(op13), (vuint32m1_t)(op14), (vuint32m1_t)(op15), (vuint32m1_t)(op16), (const uint32_t *)(op17), (vuint32m1_t)(op18), (size_t)(op19)) +#define vluxseg8ei32_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei32_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vuint32mf2_t *)(op6), (vuint32mf2_t *)(op7), (const uint32_t *)(op8), (vuint32mf2_t)(op9), (size_t)(op10)) +#define vluxseg8ei32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei32_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vuint32mf2_t *)(op6), (vuint32mf2_t *)(op7), (vbool64_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (vuint32mf2_t)(op11), (vuint32mf2_t)(op12), (vuint32mf2_t)(op13), (vuint32mf2_t)(op14), (vuint32mf2_t)(op15), (vuint32mf2_t)(op16), (const uint32_t *)(op17), (vuint32mf2_t)(op18), (size_t)(op19)) +#define vluxseg2ei64_v_i32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei64_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (const int32_t *)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vluxseg2ei64_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei64_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vbool32_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (const int32_t *)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vluxseg2ei64_v_i32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei64_v_i32m2((vint32m2_t *)(op0), (vint32m2_t *)(op1), (const int32_t *)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vluxseg2ei64_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei64_v_i32m2_m((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vbool16_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (const int32_t *)(op5), (vuint64m4_t)(op6), (size_t)(op7)) +#define vluxseg2ei64_v_i32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei64_v_i32m4((vint32m4_t *)(op0), (vint32m4_t *)(op1), (const int32_t *)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vluxseg2ei64_v_i32m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei64_v_i32m4_m((vint32m4_t *)(op0), (vint32m4_t *)(op1), (vbool8_t)(op2), (vint32m4_t)(op3), (vint32m4_t)(op4), (const int32_t *)(op5), (vuint64m8_t)(op6), (size_t)(op7)) +#define vluxseg2ei64_v_i32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei64_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (const int32_t *)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vluxseg2ei64_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei64_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vbool64_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (const int32_t *)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vluxseg2ei64_v_u32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei64_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (const uint32_t *)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vluxseg2ei64_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei64_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vbool32_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (const uint32_t *)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vluxseg2ei64_v_u32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei64_v_u32m2((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (const uint32_t *)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vluxseg2ei64_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei64_v_u32m2_m((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vbool16_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (const uint32_t *)(op5), (vuint64m4_t)(op6), (size_t)(op7)) +#define vluxseg2ei64_v_u32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei64_v_u32m4((vuint32m4_t *)(op0), (vuint32m4_t *)(op1), (const uint32_t *)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vluxseg2ei64_v_u32m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei64_v_u32m4_m((vuint32m4_t *)(op0), (vuint32m4_t *)(op1), (vbool8_t)(op2), (vuint32m4_t)(op3), (vuint32m4_t)(op4), (const uint32_t *)(op5), (vuint64m8_t)(op6), (size_t)(op7)) +#define vluxseg2ei64_v_u32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei64_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (const uint32_t *)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vluxseg2ei64_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei64_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vbool64_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (const uint32_t *)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vluxseg3ei64_v_i32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei64_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (const int32_t *)(op3), (vuint64m2_t)(op4), (size_t)(op5)) +#define vluxseg3ei64_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei64_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vbool32_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (const int32_t *)(op7), (vuint64m2_t)(op8), (size_t)(op9)) +#define vluxseg3ei64_v_i32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei64_v_i32m2((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vint32m2_t *)(op2), (const int32_t *)(op3), (vuint64m4_t)(op4), (size_t)(op5)) +#define vluxseg3ei64_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei64_v_i32m2_m((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vint32m2_t *)(op2), (vbool16_t)(op3), (vint32m2_t)(op4), (vint32m2_t)(op5), (vint32m2_t)(op6), (const int32_t *)(op7), (vuint64m4_t)(op8), (size_t)(op9)) +#define vluxseg3ei64_v_i32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei64_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (const int32_t *)(op3), (vuint64m1_t)(op4), (size_t)(op5)) +#define vluxseg3ei64_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei64_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vbool64_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (const int32_t *)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vluxseg3ei64_v_u32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei64_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (const uint32_t *)(op3), (vuint64m2_t)(op4), (size_t)(op5)) +#define vluxseg3ei64_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei64_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vbool32_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (const uint32_t *)(op7), (vuint64m2_t)(op8), (size_t)(op9)) +#define vluxseg3ei64_v_u32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei64_v_u32m2((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vuint32m2_t *)(op2), (const uint32_t *)(op3), (vuint64m4_t)(op4), (size_t)(op5)) +#define vluxseg3ei64_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei64_v_u32m2_m((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vuint32m2_t *)(op2), (vbool16_t)(op3), (vuint32m2_t)(op4), (vuint32m2_t)(op5), (vuint32m2_t)(op6), (const uint32_t *)(op7), (vuint64m4_t)(op8), (size_t)(op9)) +#define vluxseg3ei64_v_u32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei64_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (const uint32_t *)(op3), (vuint64m1_t)(op4), (size_t)(op5)) +#define vluxseg3ei64_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei64_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vbool64_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (const uint32_t *)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vluxseg4ei64_v_i32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei64_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (const int32_t *)(op4), (vuint64m2_t)(op5), (size_t)(op6)) +#define vluxseg4ei64_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei64_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vbool32_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (const int32_t *)(op9), (vuint64m2_t)(op10), (size_t)(op11)) +#define vluxseg4ei64_v_i32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei64_v_i32m2((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vint32m2_t *)(op2), (vint32m2_t *)(op3), (const int32_t *)(op4), (vuint64m4_t)(op5), (size_t)(op6)) +#define vluxseg4ei64_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei64_v_i32m2_m((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vint32m2_t *)(op2), (vint32m2_t *)(op3), (vbool16_t)(op4), (vint32m2_t)(op5), (vint32m2_t)(op6), (vint32m2_t)(op7), (vint32m2_t)(op8), (const int32_t *)(op9), (vuint64m4_t)(op10), (size_t)(op11)) +#define vluxseg4ei64_v_i32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei64_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (const int32_t *)(op4), (vuint64m1_t)(op5), (size_t)(op6)) +#define vluxseg4ei64_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei64_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vbool64_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (const int32_t *)(op9), (vuint64m1_t)(op10), (size_t)(op11)) +#define vluxseg4ei64_v_u32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei64_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (const uint32_t *)(op4), (vuint64m2_t)(op5), (size_t)(op6)) +#define vluxseg4ei64_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei64_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vbool32_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (const uint32_t *)(op9), (vuint64m2_t)(op10), (size_t)(op11)) +#define vluxseg4ei64_v_u32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei64_v_u32m2((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vuint32m2_t *)(op2), (vuint32m2_t *)(op3), (const uint32_t *)(op4), (vuint64m4_t)(op5), (size_t)(op6)) +#define vluxseg4ei64_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei64_v_u32m2_m((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vuint32m2_t *)(op2), (vuint32m2_t *)(op3), (vbool16_t)(op4), (vuint32m2_t)(op5), (vuint32m2_t)(op6), (vuint32m2_t)(op7), (vuint32m2_t)(op8), (const uint32_t *)(op9), (vuint64m4_t)(op10), (size_t)(op11)) +#define vluxseg4ei64_v_u32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei64_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (const uint32_t *)(op4), (vuint64m1_t)(op5), (size_t)(op6)) +#define vluxseg4ei64_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei64_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vbool64_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (const uint32_t *)(op9), (vuint64m1_t)(op10), (size_t)(op11)) +#define vluxseg5ei64_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei64_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (const int32_t *)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vluxseg5ei64_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei64_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vbool32_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (const int32_t *)(op11), (vuint64m2_t)(op12), (size_t)(op13)) +#define vluxseg5ei64_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei64_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (const int32_t *)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vluxseg5ei64_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei64_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vbool64_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (const int32_t *)(op11), (vuint64m1_t)(op12), (size_t)(op13)) +#define vluxseg5ei64_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei64_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (const uint32_t *)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vluxseg5ei64_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei64_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vbool32_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (const uint32_t *)(op11), (vuint64m2_t)(op12), (size_t)(op13)) +#define vluxseg5ei64_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei64_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (const uint32_t *)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vluxseg5ei64_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei64_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vbool64_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (const uint32_t *)(op11), (vuint64m1_t)(op12), (size_t)(op13)) +#define vluxseg6ei64_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei64_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (const int32_t *)(op6), (vuint64m2_t)(op7), (size_t)(op8)) +#define vluxseg6ei64_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei64_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vbool32_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (vint32m1_t)(op11), (vint32m1_t)(op12), (const int32_t *)(op13), (vuint64m2_t)(op14), (size_t)(op15)) +#define vluxseg6ei64_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei64_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (const int32_t *)(op6), (vuint64m1_t)(op7), (size_t)(op8)) +#define vluxseg6ei64_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei64_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vbool64_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (vint32mf2_t)(op11), (vint32mf2_t)(op12), (const int32_t *)(op13), (vuint64m1_t)(op14), (size_t)(op15)) +#define vluxseg6ei64_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei64_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (const uint32_t *)(op6), (vuint64m2_t)(op7), (size_t)(op8)) +#define vluxseg6ei64_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei64_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vbool32_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (vuint32m1_t)(op11), (vuint32m1_t)(op12), (const uint32_t *)(op13), (vuint64m2_t)(op14), (size_t)(op15)) +#define vluxseg6ei64_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei64_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (const uint32_t *)(op6), (vuint64m1_t)(op7), (size_t)(op8)) +#define vluxseg6ei64_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei64_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vbool64_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (vuint32mf2_t)(op11), (vuint32mf2_t)(op12), (const uint32_t *)(op13), (vuint64m1_t)(op14), (size_t)(op15)) +#define vluxseg7ei64_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei64_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vint32m1_t *)(op6), (const int32_t *)(op7), (vuint64m2_t)(op8), (size_t)(op9)) +#define vluxseg7ei64_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei64_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vint32m1_t *)(op6), (vbool32_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (vint32m1_t)(op11), (vint32m1_t)(op12), (vint32m1_t)(op13), (vint32m1_t)(op14), (const int32_t *)(op15), (vuint64m2_t)(op16), (size_t)(op17)) +#define vluxseg7ei64_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei64_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vint32mf2_t *)(op6), (const int32_t *)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vluxseg7ei64_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei64_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vint32mf2_t *)(op6), (vbool64_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (vint32mf2_t)(op11), (vint32mf2_t)(op12), (vint32mf2_t)(op13), (vint32mf2_t)(op14), (const int32_t *)(op15), (vuint64m1_t)(op16), (size_t)(op17)) +#define vluxseg7ei64_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei64_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vuint32m1_t *)(op6), (const uint32_t *)(op7), (vuint64m2_t)(op8), (size_t)(op9)) +#define vluxseg7ei64_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei64_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vuint32m1_t *)(op6), (vbool32_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (vuint32m1_t)(op11), (vuint32m1_t)(op12), (vuint32m1_t)(op13), (vuint32m1_t)(op14), (const uint32_t *)(op15), (vuint64m2_t)(op16), (size_t)(op17)) +#define vluxseg7ei64_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei64_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vuint32mf2_t *)(op6), (const uint32_t *)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vluxseg7ei64_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei64_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vuint32mf2_t *)(op6), (vbool64_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (vuint32mf2_t)(op11), (vuint32mf2_t)(op12), (vuint32mf2_t)(op13), (vuint32mf2_t)(op14), (const uint32_t *)(op15), (vuint64m1_t)(op16), (size_t)(op17)) +#define vluxseg8ei64_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei64_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vint32m1_t *)(op6), (vint32m1_t *)(op7), (const int32_t *)(op8), (vuint64m2_t)(op9), (size_t)(op10)) +#define vluxseg8ei64_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei64_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vint32m1_t *)(op6), (vint32m1_t *)(op7), (vbool32_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (vint32m1_t)(op11), (vint32m1_t)(op12), (vint32m1_t)(op13), (vint32m1_t)(op14), (vint32m1_t)(op15), (vint32m1_t)(op16), (const int32_t *)(op17), (vuint64m2_t)(op18), (size_t)(op19)) +#define vluxseg8ei64_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei64_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vint32mf2_t *)(op6), (vint32mf2_t *)(op7), (const int32_t *)(op8), (vuint64m1_t)(op9), (size_t)(op10)) +#define vluxseg8ei64_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei64_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vint32mf2_t *)(op6), (vint32mf2_t *)(op7), (vbool64_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (vint32mf2_t)(op11), (vint32mf2_t)(op12), (vint32mf2_t)(op13), (vint32mf2_t)(op14), (vint32mf2_t)(op15), (vint32mf2_t)(op16), (const int32_t *)(op17), (vuint64m1_t)(op18), (size_t)(op19)) +#define vluxseg8ei64_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei64_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vuint32m1_t *)(op6), (vuint32m1_t *)(op7), (const uint32_t *)(op8), (vuint64m2_t)(op9), (size_t)(op10)) +#define vluxseg8ei64_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei64_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vuint32m1_t *)(op6), (vuint32m1_t *)(op7), (vbool32_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (vuint32m1_t)(op11), (vuint32m1_t)(op12), (vuint32m1_t)(op13), (vuint32m1_t)(op14), (vuint32m1_t)(op15), (vuint32m1_t)(op16), (const uint32_t *)(op17), (vuint64m2_t)(op18), (size_t)(op19)) +#define vluxseg8ei64_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei64_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vuint32mf2_t *)(op6), (vuint32mf2_t *)(op7), (const uint32_t *)(op8), (vuint64m1_t)(op9), (size_t)(op10)) +#define vluxseg8ei64_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei64_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vuint32mf2_t *)(op6), (vuint32mf2_t *)(op7), (vbool64_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (vuint32mf2_t)(op11), (vuint32mf2_t)(op12), (vuint32mf2_t)(op13), (vuint32mf2_t)(op14), (vuint32mf2_t)(op15), (vuint32mf2_t)(op16), (const uint32_t *)(op17), (vuint64m1_t)(op18), (size_t)(op19)) +#define vluxseg2ei8_v_i64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (const int64_t *)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vbool64_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (const int64_t *)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vluxseg2ei8_v_i64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_i64m2((vint64m2_t *)(op0), (vint64m2_t *)(op1), (const int64_t *)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_i64m2_m((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vbool32_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (const int64_t *)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vluxseg2ei8_v_i64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_i64m4((vint64m4_t *)(op0), (vint64m4_t *)(op1), (const int64_t *)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_i64m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_i64m4_m((vint64m4_t *)(op0), (vint64m4_t *)(op1), (vbool16_t)(op2), (vint64m4_t)(op3), (vint64m4_t)(op4), (const int64_t *)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vluxseg2ei8_v_u64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (const uint64_t *)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vbool64_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (const uint64_t *)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vluxseg2ei8_v_u64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_u64m2((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (const uint64_t *)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_u64m2_m((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vbool32_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (const uint64_t *)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vluxseg2ei8_v_u64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_u64m4((vuint64m4_t *)(op0), (vuint64m4_t *)(op1), (const uint64_t *)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_u64m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_u64m4_m((vuint64m4_t *)(op0), (vuint64m4_t *)(op1), (vbool16_t)(op2), (vuint64m4_t)(op3), (vuint64m4_t)(op4), (const uint64_t *)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vluxseg3ei8_v_i64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei8_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (const int64_t *)(op3), (vuint8mf8_t)(op4), (size_t)(op5)) +#define vluxseg3ei8_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei8_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vbool64_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (const int64_t *)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vluxseg3ei8_v_i64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei8_v_i64m2((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vint64m2_t *)(op2), (const int64_t *)(op3), (vuint8mf4_t)(op4), (size_t)(op5)) +#define vluxseg3ei8_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei8_v_i64m2_m((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vint64m2_t *)(op2), (vbool32_t)(op3), (vint64m2_t)(op4), (vint64m2_t)(op5), (vint64m2_t)(op6), (const int64_t *)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vluxseg3ei8_v_u64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei8_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (const uint64_t *)(op3), (vuint8mf8_t)(op4), (size_t)(op5)) +#define vluxseg3ei8_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei8_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vbool64_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (const uint64_t *)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vluxseg3ei8_v_u64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei8_v_u64m2((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vuint64m2_t *)(op2), (const uint64_t *)(op3), (vuint8mf4_t)(op4), (size_t)(op5)) +#define vluxseg3ei8_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei8_v_u64m2_m((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vuint64m2_t *)(op2), (vbool32_t)(op3), (vuint64m2_t)(op4), (vuint64m2_t)(op5), (vuint64m2_t)(op6), (const uint64_t *)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vluxseg4ei8_v_i64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei8_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (const int64_t *)(op4), (vuint8mf8_t)(op5), (size_t)(op6)) +#define vluxseg4ei8_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei8_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vbool64_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (const int64_t *)(op9), (vuint8mf8_t)(op10), (size_t)(op11)) +#define vluxseg4ei8_v_i64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei8_v_i64m2((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vint64m2_t *)(op2), (vint64m2_t *)(op3), (const int64_t *)(op4), (vuint8mf4_t)(op5), (size_t)(op6)) +#define vluxseg4ei8_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei8_v_i64m2_m((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vint64m2_t *)(op2), (vint64m2_t *)(op3), (vbool32_t)(op4), (vint64m2_t)(op5), (vint64m2_t)(op6), (vint64m2_t)(op7), (vint64m2_t)(op8), (const int64_t *)(op9), (vuint8mf4_t)(op10), (size_t)(op11)) +#define vluxseg4ei8_v_u64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei8_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (const uint64_t *)(op4), (vuint8mf8_t)(op5), (size_t)(op6)) +#define vluxseg4ei8_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei8_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vbool64_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (const uint64_t *)(op9), (vuint8mf8_t)(op10), (size_t)(op11)) +#define vluxseg4ei8_v_u64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei8_v_u64m2((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vuint64m2_t *)(op2), (vuint64m2_t *)(op3), (const uint64_t *)(op4), (vuint8mf4_t)(op5), (size_t)(op6)) +#define vluxseg4ei8_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei8_v_u64m2_m((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vuint64m2_t *)(op2), (vuint64m2_t *)(op3), (vbool32_t)(op4), (vuint64m2_t)(op5), (vuint64m2_t)(op6), (vuint64m2_t)(op7), (vuint64m2_t)(op8), (const uint64_t *)(op9), (vuint8mf4_t)(op10), (size_t)(op11)) +#define vluxseg5ei8_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei8_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (const int64_t *)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vluxseg5ei8_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei8_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vbool64_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (const int64_t *)(op11), (vuint8mf8_t)(op12), (size_t)(op13)) +#define vluxseg5ei8_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei8_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (const uint64_t *)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vluxseg5ei8_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei8_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vbool64_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (const uint64_t *)(op11), (vuint8mf8_t)(op12), (size_t)(op13)) +#define vluxseg6ei8_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei8_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (const int64_t *)(op6), (vuint8mf8_t)(op7), (size_t)(op8)) +#define vluxseg6ei8_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei8_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vbool64_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (vint64m1_t)(op11), (vint64m1_t)(op12), (const int64_t *)(op13), (vuint8mf8_t)(op14), (size_t)(op15)) +#define vluxseg6ei8_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei8_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (const uint64_t *)(op6), (vuint8mf8_t)(op7), (size_t)(op8)) +#define vluxseg6ei8_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei8_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vbool64_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (vuint64m1_t)(op11), (vuint64m1_t)(op12), (const uint64_t *)(op13), (vuint8mf8_t)(op14), (size_t)(op15)) +#define vluxseg7ei8_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei8_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vint64m1_t *)(op6), (const int64_t *)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vluxseg7ei8_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei8_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vint64m1_t *)(op6), (vbool64_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (vint64m1_t)(op11), (vint64m1_t)(op12), (vint64m1_t)(op13), (vint64m1_t)(op14), (const int64_t *)(op15), (vuint8mf8_t)(op16), (size_t)(op17)) +#define vluxseg7ei8_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei8_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vuint64m1_t *)(op6), (const uint64_t *)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vluxseg7ei8_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei8_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vuint64m1_t *)(op6), (vbool64_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (vuint64m1_t)(op11), (vuint64m1_t)(op12), (vuint64m1_t)(op13), (vuint64m1_t)(op14), (const uint64_t *)(op15), (vuint8mf8_t)(op16), (size_t)(op17)) +#define vluxseg8ei8_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei8_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vint64m1_t *)(op6), (vint64m1_t *)(op7), (const int64_t *)(op8), (vuint8mf8_t)(op9), (size_t)(op10)) +#define vluxseg8ei8_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei8_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vint64m1_t *)(op6), (vint64m1_t *)(op7), (vbool64_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (vint64m1_t)(op11), (vint64m1_t)(op12), (vint64m1_t)(op13), (vint64m1_t)(op14), (vint64m1_t)(op15), (vint64m1_t)(op16), (const int64_t *)(op17), (vuint8mf8_t)(op18), (size_t)(op19)) +#define vluxseg8ei8_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei8_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vuint64m1_t *)(op6), (vuint64m1_t *)(op7), (const uint64_t *)(op8), (vuint8mf8_t)(op9), (size_t)(op10)) +#define vluxseg8ei8_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei8_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vuint64m1_t *)(op6), (vuint64m1_t *)(op7), (vbool64_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (vuint64m1_t)(op11), (vuint64m1_t)(op12), (vuint64m1_t)(op13), (vuint64m1_t)(op14), (vuint64m1_t)(op15), (vuint64m1_t)(op16), (const uint64_t *)(op17), (vuint8mf8_t)(op18), (size_t)(op19)) +#define vluxseg2ei16_v_i64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (const int64_t *)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vbool64_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (const int64_t *)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vluxseg2ei16_v_i64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_i64m2((vint64m2_t *)(op0), (vint64m2_t *)(op1), (const int64_t *)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_i64m2_m((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vbool32_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (const int64_t *)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vluxseg2ei16_v_i64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_i64m4((vint64m4_t *)(op0), (vint64m4_t *)(op1), (const int64_t *)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_i64m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_i64m4_m((vint64m4_t *)(op0), (vint64m4_t *)(op1), (vbool16_t)(op2), (vint64m4_t)(op3), (vint64m4_t)(op4), (const int64_t *)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vluxseg2ei16_v_u64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (const uint64_t *)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vbool64_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (const uint64_t *)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vluxseg2ei16_v_u64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_u64m2((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (const uint64_t *)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_u64m2_m((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vbool32_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (const uint64_t *)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vluxseg2ei16_v_u64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_u64m4((vuint64m4_t *)(op0), (vuint64m4_t *)(op1), (const uint64_t *)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_u64m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_u64m4_m((vuint64m4_t *)(op0), (vuint64m4_t *)(op1), (vbool16_t)(op2), (vuint64m4_t)(op3), (vuint64m4_t)(op4), (const uint64_t *)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vluxseg3ei16_v_i64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei16_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (const int64_t *)(op3), (vuint16mf4_t)(op4), (size_t)(op5)) +#define vluxseg3ei16_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei16_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vbool64_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (const int64_t *)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vluxseg3ei16_v_i64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei16_v_i64m2((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vint64m2_t *)(op2), (const int64_t *)(op3), (vuint16mf2_t)(op4), (size_t)(op5)) +#define vluxseg3ei16_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei16_v_i64m2_m((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vint64m2_t *)(op2), (vbool32_t)(op3), (vint64m2_t)(op4), (vint64m2_t)(op5), (vint64m2_t)(op6), (const int64_t *)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vluxseg3ei16_v_u64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei16_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (const uint64_t *)(op3), (vuint16mf4_t)(op4), (size_t)(op5)) +#define vluxseg3ei16_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei16_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vbool64_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (const uint64_t *)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vluxseg3ei16_v_u64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei16_v_u64m2((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vuint64m2_t *)(op2), (const uint64_t *)(op3), (vuint16mf2_t)(op4), (size_t)(op5)) +#define vluxseg3ei16_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei16_v_u64m2_m((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vuint64m2_t *)(op2), (vbool32_t)(op3), (vuint64m2_t)(op4), (vuint64m2_t)(op5), (vuint64m2_t)(op6), (const uint64_t *)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vluxseg4ei16_v_i64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei16_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (const int64_t *)(op4), (vuint16mf4_t)(op5), (size_t)(op6)) +#define vluxseg4ei16_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei16_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vbool64_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (const int64_t *)(op9), (vuint16mf4_t)(op10), (size_t)(op11)) +#define vluxseg4ei16_v_i64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei16_v_i64m2((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vint64m2_t *)(op2), (vint64m2_t *)(op3), (const int64_t *)(op4), (vuint16mf2_t)(op5), (size_t)(op6)) +#define vluxseg4ei16_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei16_v_i64m2_m((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vint64m2_t *)(op2), (vint64m2_t *)(op3), (vbool32_t)(op4), (vint64m2_t)(op5), (vint64m2_t)(op6), (vint64m2_t)(op7), (vint64m2_t)(op8), (const int64_t *)(op9), (vuint16mf2_t)(op10), (size_t)(op11)) +#define vluxseg4ei16_v_u64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei16_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (const uint64_t *)(op4), (vuint16mf4_t)(op5), (size_t)(op6)) +#define vluxseg4ei16_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei16_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vbool64_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (const uint64_t *)(op9), (vuint16mf4_t)(op10), (size_t)(op11)) +#define vluxseg4ei16_v_u64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei16_v_u64m2((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vuint64m2_t *)(op2), (vuint64m2_t *)(op3), (const uint64_t *)(op4), (vuint16mf2_t)(op5), (size_t)(op6)) +#define vluxseg4ei16_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei16_v_u64m2_m((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vuint64m2_t *)(op2), (vuint64m2_t *)(op3), (vbool32_t)(op4), (vuint64m2_t)(op5), (vuint64m2_t)(op6), (vuint64m2_t)(op7), (vuint64m2_t)(op8), (const uint64_t *)(op9), (vuint16mf2_t)(op10), (size_t)(op11)) +#define vluxseg5ei16_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei16_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (const int64_t *)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vluxseg5ei16_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei16_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vbool64_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (const int64_t *)(op11), (vuint16mf4_t)(op12), (size_t)(op13)) +#define vluxseg5ei16_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei16_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (const uint64_t *)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vluxseg5ei16_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei16_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vbool64_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (const uint64_t *)(op11), (vuint16mf4_t)(op12), (size_t)(op13)) +#define vluxseg6ei16_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei16_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (const int64_t *)(op6), (vuint16mf4_t)(op7), (size_t)(op8)) +#define vluxseg6ei16_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei16_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vbool64_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (vint64m1_t)(op11), (vint64m1_t)(op12), (const int64_t *)(op13), (vuint16mf4_t)(op14), (size_t)(op15)) +#define vluxseg6ei16_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei16_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (const uint64_t *)(op6), (vuint16mf4_t)(op7), (size_t)(op8)) +#define vluxseg6ei16_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei16_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vbool64_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (vuint64m1_t)(op11), (vuint64m1_t)(op12), (const uint64_t *)(op13), (vuint16mf4_t)(op14), (size_t)(op15)) +#define vluxseg7ei16_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei16_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vint64m1_t *)(op6), (const int64_t *)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vluxseg7ei16_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei16_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vint64m1_t *)(op6), (vbool64_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (vint64m1_t)(op11), (vint64m1_t)(op12), (vint64m1_t)(op13), (vint64m1_t)(op14), (const int64_t *)(op15), (vuint16mf4_t)(op16), (size_t)(op17)) +#define vluxseg7ei16_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei16_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vuint64m1_t *)(op6), (const uint64_t *)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vluxseg7ei16_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei16_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vuint64m1_t *)(op6), (vbool64_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (vuint64m1_t)(op11), (vuint64m1_t)(op12), (vuint64m1_t)(op13), (vuint64m1_t)(op14), (const uint64_t *)(op15), (vuint16mf4_t)(op16), (size_t)(op17)) +#define vluxseg8ei16_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei16_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vint64m1_t *)(op6), (vint64m1_t *)(op7), (const int64_t *)(op8), (vuint16mf4_t)(op9), (size_t)(op10)) +#define vluxseg8ei16_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei16_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vint64m1_t *)(op6), (vint64m1_t *)(op7), (vbool64_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (vint64m1_t)(op11), (vint64m1_t)(op12), (vint64m1_t)(op13), (vint64m1_t)(op14), (vint64m1_t)(op15), (vint64m1_t)(op16), (const int64_t *)(op17), (vuint16mf4_t)(op18), (size_t)(op19)) +#define vluxseg8ei16_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei16_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vuint64m1_t *)(op6), (vuint64m1_t *)(op7), (const uint64_t *)(op8), (vuint16mf4_t)(op9), (size_t)(op10)) +#define vluxseg8ei16_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei16_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vuint64m1_t *)(op6), (vuint64m1_t *)(op7), (vbool64_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (vuint64m1_t)(op11), (vuint64m1_t)(op12), (vuint64m1_t)(op13), (vuint64m1_t)(op14), (vuint64m1_t)(op15), (vuint64m1_t)(op16), (const uint64_t *)(op17), (vuint16mf4_t)(op18), (size_t)(op19)) +#define vluxseg2ei32_v_i64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (const int64_t *)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vbool64_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (const int64_t *)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vluxseg2ei32_v_i64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_i64m2((vint64m2_t *)(op0), (vint64m2_t *)(op1), (const int64_t *)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_i64m2_m((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vbool32_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (const int64_t *)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vluxseg2ei32_v_i64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_i64m4((vint64m4_t *)(op0), (vint64m4_t *)(op1), (const int64_t *)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_i64m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_i64m4_m((vint64m4_t *)(op0), (vint64m4_t *)(op1), (vbool16_t)(op2), (vint64m4_t)(op3), (vint64m4_t)(op4), (const int64_t *)(op5), (vuint32m2_t)(op6), (size_t)(op7)) +#define vluxseg2ei32_v_u64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (const uint64_t *)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vbool64_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (const uint64_t *)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vluxseg2ei32_v_u64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_u64m2((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (const uint64_t *)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_u64m2_m((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vbool32_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (const uint64_t *)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vluxseg2ei32_v_u64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_u64m4((vuint64m4_t *)(op0), (vuint64m4_t *)(op1), (const uint64_t *)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_u64m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_u64m4_m((vuint64m4_t *)(op0), (vuint64m4_t *)(op1), (vbool16_t)(op2), (vuint64m4_t)(op3), (vuint64m4_t)(op4), (const uint64_t *)(op5), (vuint32m2_t)(op6), (size_t)(op7)) +#define vluxseg3ei32_v_i64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei32_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (const int64_t *)(op3), (vuint32mf2_t)(op4), (size_t)(op5)) +#define vluxseg3ei32_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei32_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vbool64_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (const int64_t *)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vluxseg3ei32_v_i64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei32_v_i64m2((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vint64m2_t *)(op2), (const int64_t *)(op3), (vuint32m1_t)(op4), (size_t)(op5)) +#define vluxseg3ei32_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei32_v_i64m2_m((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vint64m2_t *)(op2), (vbool32_t)(op3), (vint64m2_t)(op4), (vint64m2_t)(op5), (vint64m2_t)(op6), (const int64_t *)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vluxseg3ei32_v_u64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei32_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (const uint64_t *)(op3), (vuint32mf2_t)(op4), (size_t)(op5)) +#define vluxseg3ei32_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei32_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vbool64_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (const uint64_t *)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vluxseg3ei32_v_u64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei32_v_u64m2((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vuint64m2_t *)(op2), (const uint64_t *)(op3), (vuint32m1_t)(op4), (size_t)(op5)) +#define vluxseg3ei32_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei32_v_u64m2_m((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vuint64m2_t *)(op2), (vbool32_t)(op3), (vuint64m2_t)(op4), (vuint64m2_t)(op5), (vuint64m2_t)(op6), (const uint64_t *)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vluxseg4ei32_v_i64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei32_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (const int64_t *)(op4), (vuint32mf2_t)(op5), (size_t)(op6)) +#define vluxseg4ei32_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei32_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vbool64_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (const int64_t *)(op9), (vuint32mf2_t)(op10), (size_t)(op11)) +#define vluxseg4ei32_v_i64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei32_v_i64m2((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vint64m2_t *)(op2), (vint64m2_t *)(op3), (const int64_t *)(op4), (vuint32m1_t)(op5), (size_t)(op6)) +#define vluxseg4ei32_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei32_v_i64m2_m((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vint64m2_t *)(op2), (vint64m2_t *)(op3), (vbool32_t)(op4), (vint64m2_t)(op5), (vint64m2_t)(op6), (vint64m2_t)(op7), (vint64m2_t)(op8), (const int64_t *)(op9), (vuint32m1_t)(op10), (size_t)(op11)) +#define vluxseg4ei32_v_u64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei32_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (const uint64_t *)(op4), (vuint32mf2_t)(op5), (size_t)(op6)) +#define vluxseg4ei32_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei32_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vbool64_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (const uint64_t *)(op9), (vuint32mf2_t)(op10), (size_t)(op11)) +#define vluxseg4ei32_v_u64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei32_v_u64m2((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vuint64m2_t *)(op2), (vuint64m2_t *)(op3), (const uint64_t *)(op4), (vuint32m1_t)(op5), (size_t)(op6)) +#define vluxseg4ei32_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei32_v_u64m2_m((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vuint64m2_t *)(op2), (vuint64m2_t *)(op3), (vbool32_t)(op4), (vuint64m2_t)(op5), (vuint64m2_t)(op6), (vuint64m2_t)(op7), (vuint64m2_t)(op8), (const uint64_t *)(op9), (vuint32m1_t)(op10), (size_t)(op11)) +#define vluxseg5ei32_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei32_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (const int64_t *)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vluxseg5ei32_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei32_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vbool64_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (const int64_t *)(op11), (vuint32mf2_t)(op12), (size_t)(op13)) +#define vluxseg5ei32_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei32_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (const uint64_t *)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vluxseg5ei32_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei32_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vbool64_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (const uint64_t *)(op11), (vuint32mf2_t)(op12), (size_t)(op13)) +#define vluxseg6ei32_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei32_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (const int64_t *)(op6), (vuint32mf2_t)(op7), (size_t)(op8)) +#define vluxseg6ei32_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei32_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vbool64_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (vint64m1_t)(op11), (vint64m1_t)(op12), (const int64_t *)(op13), (vuint32mf2_t)(op14), (size_t)(op15)) +#define vluxseg6ei32_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei32_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (const uint64_t *)(op6), (vuint32mf2_t)(op7), (size_t)(op8)) +#define vluxseg6ei32_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei32_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vbool64_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (vuint64m1_t)(op11), (vuint64m1_t)(op12), (const uint64_t *)(op13), (vuint32mf2_t)(op14), (size_t)(op15)) +#define vluxseg7ei32_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei32_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vint64m1_t *)(op6), (const int64_t *)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vluxseg7ei32_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei32_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vint64m1_t *)(op6), (vbool64_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (vint64m1_t)(op11), (vint64m1_t)(op12), (vint64m1_t)(op13), (vint64m1_t)(op14), (const int64_t *)(op15), (vuint32mf2_t)(op16), (size_t)(op17)) +#define vluxseg7ei32_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei32_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vuint64m1_t *)(op6), (const uint64_t *)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vluxseg7ei32_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei32_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vuint64m1_t *)(op6), (vbool64_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (vuint64m1_t)(op11), (vuint64m1_t)(op12), (vuint64m1_t)(op13), (vuint64m1_t)(op14), (const uint64_t *)(op15), (vuint32mf2_t)(op16), (size_t)(op17)) +#define vluxseg8ei32_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei32_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vint64m1_t *)(op6), (vint64m1_t *)(op7), (const int64_t *)(op8), (vuint32mf2_t)(op9), (size_t)(op10)) +#define vluxseg8ei32_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei32_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vint64m1_t *)(op6), (vint64m1_t *)(op7), (vbool64_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (vint64m1_t)(op11), (vint64m1_t)(op12), (vint64m1_t)(op13), (vint64m1_t)(op14), (vint64m1_t)(op15), (vint64m1_t)(op16), (const int64_t *)(op17), (vuint32mf2_t)(op18), (size_t)(op19)) +#define vluxseg8ei32_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei32_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vuint64m1_t *)(op6), (vuint64m1_t *)(op7), (const uint64_t *)(op8), (vuint32mf2_t)(op9), (size_t)(op10)) +#define vluxseg8ei32_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei32_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vuint64m1_t *)(op6), (vuint64m1_t *)(op7), (vbool64_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (vuint64m1_t)(op11), (vuint64m1_t)(op12), (vuint64m1_t)(op13), (vuint64m1_t)(op14), (vuint64m1_t)(op15), (vuint64m1_t)(op16), (const uint64_t *)(op17), (vuint32mf2_t)(op18), (size_t)(op19)) +#define vluxseg2ei64_v_i64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei64_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (const int64_t *)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vluxseg2ei64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei64_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vbool64_t)(op2), (vint64m1_t)(op3), (vint64m1_t)(op4), (const int64_t *)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vluxseg2ei64_v_i64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei64_v_i64m2((vint64m2_t *)(op0), (vint64m2_t *)(op1), (const int64_t *)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vluxseg2ei64_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei64_v_i64m2_m((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vbool32_t)(op2), (vint64m2_t)(op3), (vint64m2_t)(op4), (const int64_t *)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vluxseg2ei64_v_i64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei64_v_i64m4((vint64m4_t *)(op0), (vint64m4_t *)(op1), (const int64_t *)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vluxseg2ei64_v_i64m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei64_v_i64m4_m((vint64m4_t *)(op0), (vint64m4_t *)(op1), (vbool16_t)(op2), (vint64m4_t)(op3), (vint64m4_t)(op4), (const int64_t *)(op5), (vuint64m4_t)(op6), (size_t)(op7)) +#define vluxseg2ei64_v_u64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei64_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (const uint64_t *)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vluxseg2ei64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei64_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vbool64_t)(op2), (vuint64m1_t)(op3), (vuint64m1_t)(op4), (const uint64_t *)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vluxseg2ei64_v_u64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei64_v_u64m2((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (const uint64_t *)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vluxseg2ei64_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei64_v_u64m2_m((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vbool32_t)(op2), (vuint64m2_t)(op3), (vuint64m2_t)(op4), (const uint64_t *)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vluxseg2ei64_v_u64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei64_v_u64m4((vuint64m4_t *)(op0), (vuint64m4_t *)(op1), (const uint64_t *)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vluxseg2ei64_v_u64m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei64_v_u64m4_m((vuint64m4_t *)(op0), (vuint64m4_t *)(op1), (vbool16_t)(op2), (vuint64m4_t)(op3), (vuint64m4_t)(op4), (const uint64_t *)(op5), (vuint64m4_t)(op6), (size_t)(op7)) +#define vluxseg3ei64_v_i64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei64_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (const int64_t *)(op3), (vuint64m1_t)(op4), (size_t)(op5)) +#define vluxseg3ei64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei64_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vbool64_t)(op3), (vint64m1_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (const int64_t *)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vluxseg3ei64_v_i64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei64_v_i64m2((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vint64m2_t *)(op2), (const int64_t *)(op3), (vuint64m2_t)(op4), (size_t)(op5)) +#define vluxseg3ei64_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei64_v_i64m2_m((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vint64m2_t *)(op2), (vbool32_t)(op3), (vint64m2_t)(op4), (vint64m2_t)(op5), (vint64m2_t)(op6), (const int64_t *)(op7), (vuint64m2_t)(op8), (size_t)(op9)) +#define vluxseg3ei64_v_u64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei64_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (const uint64_t *)(op3), (vuint64m1_t)(op4), (size_t)(op5)) +#define vluxseg3ei64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei64_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vbool64_t)(op3), (vuint64m1_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (const uint64_t *)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vluxseg3ei64_v_u64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei64_v_u64m2((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vuint64m2_t *)(op2), (const uint64_t *)(op3), (vuint64m2_t)(op4), (size_t)(op5)) +#define vluxseg3ei64_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei64_v_u64m2_m((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vuint64m2_t *)(op2), (vbool32_t)(op3), (vuint64m2_t)(op4), (vuint64m2_t)(op5), (vuint64m2_t)(op6), (const uint64_t *)(op7), (vuint64m2_t)(op8), (size_t)(op9)) +#define vluxseg4ei64_v_i64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei64_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (const int64_t *)(op4), (vuint64m1_t)(op5), (size_t)(op6)) +#define vluxseg4ei64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei64_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vbool64_t)(op4), (vint64m1_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (const int64_t *)(op9), (vuint64m1_t)(op10), (size_t)(op11)) +#define vluxseg4ei64_v_i64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei64_v_i64m2((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vint64m2_t *)(op2), (vint64m2_t *)(op3), (const int64_t *)(op4), (vuint64m2_t)(op5), (size_t)(op6)) +#define vluxseg4ei64_v_i64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei64_v_i64m2_m((vint64m2_t *)(op0), (vint64m2_t *)(op1), (vint64m2_t *)(op2), (vint64m2_t *)(op3), (vbool32_t)(op4), (vint64m2_t)(op5), (vint64m2_t)(op6), (vint64m2_t)(op7), (vint64m2_t)(op8), (const int64_t *)(op9), (vuint64m2_t)(op10), (size_t)(op11)) +#define vluxseg4ei64_v_u64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei64_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (const uint64_t *)(op4), (vuint64m1_t)(op5), (size_t)(op6)) +#define vluxseg4ei64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei64_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vbool64_t)(op4), (vuint64m1_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (const uint64_t *)(op9), (vuint64m1_t)(op10), (size_t)(op11)) +#define vluxseg4ei64_v_u64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei64_v_u64m2((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vuint64m2_t *)(op2), (vuint64m2_t *)(op3), (const uint64_t *)(op4), (vuint64m2_t)(op5), (size_t)(op6)) +#define vluxseg4ei64_v_u64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei64_v_u64m2_m((vuint64m2_t *)(op0), (vuint64m2_t *)(op1), (vuint64m2_t *)(op2), (vuint64m2_t *)(op3), (vbool32_t)(op4), (vuint64m2_t)(op5), (vuint64m2_t)(op6), (vuint64m2_t)(op7), (vuint64m2_t)(op8), (const uint64_t *)(op9), (vuint64m2_t)(op10), (size_t)(op11)) +#define vluxseg5ei64_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei64_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (const int64_t *)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vluxseg5ei64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei64_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vbool64_t)(op5), (vint64m1_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (const int64_t *)(op11), (vuint64m1_t)(op12), (size_t)(op13)) +#define vluxseg5ei64_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei64_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (const uint64_t *)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vluxseg5ei64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei64_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vbool64_t)(op5), (vuint64m1_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (const uint64_t *)(op11), (vuint64m1_t)(op12), (size_t)(op13)) +#define vluxseg6ei64_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei64_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (const int64_t *)(op6), (vuint64m1_t)(op7), (size_t)(op8)) +#define vluxseg6ei64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei64_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vbool64_t)(op6), (vint64m1_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (vint64m1_t)(op11), (vint64m1_t)(op12), (const int64_t *)(op13), (vuint64m1_t)(op14), (size_t)(op15)) +#define vluxseg6ei64_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei64_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (const uint64_t *)(op6), (vuint64m1_t)(op7), (size_t)(op8)) +#define vluxseg6ei64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei64_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vbool64_t)(op6), (vuint64m1_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (vuint64m1_t)(op11), (vuint64m1_t)(op12), (const uint64_t *)(op13), (vuint64m1_t)(op14), (size_t)(op15)) +#define vluxseg7ei64_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei64_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vint64m1_t *)(op6), (const int64_t *)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vluxseg7ei64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei64_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vint64m1_t *)(op6), (vbool64_t)(op7), (vint64m1_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (vint64m1_t)(op11), (vint64m1_t)(op12), (vint64m1_t)(op13), (vint64m1_t)(op14), (const int64_t *)(op15), (vuint64m1_t)(op16), (size_t)(op17)) +#define vluxseg7ei64_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei64_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vuint64m1_t *)(op6), (const uint64_t *)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vluxseg7ei64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei64_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vuint64m1_t *)(op6), (vbool64_t)(op7), (vuint64m1_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (vuint64m1_t)(op11), (vuint64m1_t)(op12), (vuint64m1_t)(op13), (vuint64m1_t)(op14), (const uint64_t *)(op15), (vuint64m1_t)(op16), (size_t)(op17)) +#define vluxseg8ei64_v_i64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei64_v_i64m1((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vint64m1_t *)(op6), (vint64m1_t *)(op7), (const int64_t *)(op8), (vuint64m1_t)(op9), (size_t)(op10)) +#define vluxseg8ei64_v_i64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei64_v_i64m1_m((vint64m1_t *)(op0), (vint64m1_t *)(op1), (vint64m1_t *)(op2), (vint64m1_t *)(op3), (vint64m1_t *)(op4), (vint64m1_t *)(op5), (vint64m1_t *)(op6), (vint64m1_t *)(op7), (vbool64_t)(op8), (vint64m1_t)(op9), (vint64m1_t)(op10), (vint64m1_t)(op11), (vint64m1_t)(op12), (vint64m1_t)(op13), (vint64m1_t)(op14), (vint64m1_t)(op15), (vint64m1_t)(op16), (const int64_t *)(op17), (vuint64m1_t)(op18), (size_t)(op19)) +#define vluxseg8ei64_v_u64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei64_v_u64m1((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vuint64m1_t *)(op6), (vuint64m1_t *)(op7), (const uint64_t *)(op8), (vuint64m1_t)(op9), (size_t)(op10)) +#define vluxseg8ei64_v_u64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei64_v_u64m1_m((vuint64m1_t *)(op0), (vuint64m1_t *)(op1), (vuint64m1_t *)(op2), (vuint64m1_t *)(op3), (vuint64m1_t *)(op4), (vuint64m1_t *)(op5), (vuint64m1_t *)(op6), (vuint64m1_t *)(op7), (vbool64_t)(op8), (vuint64m1_t)(op9), (vuint64m1_t)(op10), (vuint64m1_t)(op11), (vuint64m1_t)(op12), (vuint64m1_t)(op13), (vuint64m1_t)(op14), (vuint64m1_t)(op15), (vuint64m1_t)(op16), (const uint64_t *)(op17), (vuint64m1_t)(op18), (size_t)(op19)) +#define vloxseg2ei8_v_i8m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (const int8_t *)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vbool8_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (const int8_t *)(op5), (vuint8m1_t)(op6), (size_t)(op7)) +#define vloxseg2ei8_v_i8m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_i8m2((vint8m2_t *)(op0), (vint8m2_t *)(op1), (const int8_t *)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_i8m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_i8m2_m((vint8m2_t *)(op0), (vint8m2_t *)(op1), (vbool4_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (const int8_t *)(op5), (vuint8m2_t)(op6), (size_t)(op7)) +#define vloxseg2ei8_v_i8m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_i8m4((vint8m4_t *)(op0), (vint8m4_t *)(op1), (const int8_t *)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_i8m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_i8m4_m((vint8m4_t *)(op0), (vint8m4_t *)(op1), (vbool2_t)(op2), (vint8m4_t)(op3), (vint8m4_t)(op4), (const int8_t *)(op5), (vuint8m4_t)(op6), (size_t)(op7)) +#define vloxseg2ei8_v_i8mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (const int8_t *)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vbool16_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (const int8_t *)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vloxseg2ei8_v_i8mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (const int8_t *)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vbool32_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (const int8_t *)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vloxseg2ei8_v_i8mf8(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (const int8_t *)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vbool64_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (const int8_t *)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vloxseg2ei8_v_u8m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (const uint8_t *)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vbool8_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (const uint8_t *)(op5), (vuint8m1_t)(op6), (size_t)(op7)) +#define vloxseg2ei8_v_u8m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_u8m2((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (const uint8_t *)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_u8m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_u8m2_m((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (vbool4_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (const uint8_t *)(op5), (vuint8m2_t)(op6), (size_t)(op7)) +#define vloxseg2ei8_v_u8m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_u8m4((vuint8m4_t *)(op0), (vuint8m4_t *)(op1), (const uint8_t *)(op2), (vuint8m4_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_u8m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_u8m4_m((vuint8m4_t *)(op0), (vuint8m4_t *)(op1), (vbool2_t)(op2), (vuint8m4_t)(op3), (vuint8m4_t)(op4), (const uint8_t *)(op5), (vuint8m4_t)(op6), (size_t)(op7)) +#define vloxseg2ei8_v_u8mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (const uint8_t *)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vbool16_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (const uint8_t *)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vloxseg2ei8_v_u8mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (const uint8_t *)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vbool32_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (const uint8_t *)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vloxseg2ei8_v_u8mf8(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (const uint8_t *)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vbool64_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (const uint8_t *)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vloxseg3ei8_v_i8m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei8_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (const int8_t *)(op3), (vuint8m1_t)(op4), (size_t)(op5)) +#define vloxseg3ei8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei8_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vbool8_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (const int8_t *)(op7), (vuint8m1_t)(op8), (size_t)(op9)) +#define vloxseg3ei8_v_i8m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei8_v_i8m2((vint8m2_t *)(op0), (vint8m2_t *)(op1), (vint8m2_t *)(op2), (const int8_t *)(op3), (vuint8m2_t)(op4), (size_t)(op5)) +#define vloxseg3ei8_v_i8m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei8_v_i8m2_m((vint8m2_t *)(op0), (vint8m2_t *)(op1), (vint8m2_t *)(op2), (vbool4_t)(op3), (vint8m2_t)(op4), (vint8m2_t)(op5), (vint8m2_t)(op6), (const int8_t *)(op7), (vuint8m2_t)(op8), (size_t)(op9)) +#define vloxseg3ei8_v_i8mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei8_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (const int8_t *)(op3), (vuint8mf2_t)(op4), (size_t)(op5)) +#define vloxseg3ei8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei8_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vbool16_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (const int8_t *)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vloxseg3ei8_v_i8mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei8_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (const int8_t *)(op3), (vuint8mf4_t)(op4), (size_t)(op5)) +#define vloxseg3ei8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei8_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vbool32_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (const int8_t *)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vloxseg3ei8_v_i8mf8(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei8_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (const int8_t *)(op3), (vuint8mf8_t)(op4), (size_t)(op5)) +#define vloxseg3ei8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei8_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vbool64_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (const int8_t *)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vloxseg3ei8_v_u8m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei8_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (const uint8_t *)(op3), (vuint8m1_t)(op4), (size_t)(op5)) +#define vloxseg3ei8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei8_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vbool8_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (const uint8_t *)(op7), (vuint8m1_t)(op8), (size_t)(op9)) +#define vloxseg3ei8_v_u8m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei8_v_u8m2((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (vuint8m2_t *)(op2), (const uint8_t *)(op3), (vuint8m2_t)(op4), (size_t)(op5)) +#define vloxseg3ei8_v_u8m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei8_v_u8m2_m((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (vuint8m2_t *)(op2), (vbool4_t)(op3), (vuint8m2_t)(op4), (vuint8m2_t)(op5), (vuint8m2_t)(op6), (const uint8_t *)(op7), (vuint8m2_t)(op8), (size_t)(op9)) +#define vloxseg3ei8_v_u8mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei8_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (const uint8_t *)(op3), (vuint8mf2_t)(op4), (size_t)(op5)) +#define vloxseg3ei8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei8_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vbool16_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (const uint8_t *)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vloxseg3ei8_v_u8mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei8_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (const uint8_t *)(op3), (vuint8mf4_t)(op4), (size_t)(op5)) +#define vloxseg3ei8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei8_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vbool32_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (const uint8_t *)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vloxseg3ei8_v_u8mf8(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei8_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (const uint8_t *)(op3), (vuint8mf8_t)(op4), (size_t)(op5)) +#define vloxseg3ei8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei8_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vbool64_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (const uint8_t *)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vloxseg4ei8_v_i8m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei8_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (const int8_t *)(op4), (vuint8m1_t)(op5), (size_t)(op6)) +#define vloxseg4ei8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei8_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vbool8_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (const int8_t *)(op9), (vuint8m1_t)(op10), (size_t)(op11)) +#define vloxseg4ei8_v_i8m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei8_v_i8m2((vint8m2_t *)(op0), (vint8m2_t *)(op1), (vint8m2_t *)(op2), (vint8m2_t *)(op3), (const int8_t *)(op4), (vuint8m2_t)(op5), (size_t)(op6)) +#define vloxseg4ei8_v_i8m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei8_v_i8m2_m((vint8m2_t *)(op0), (vint8m2_t *)(op1), (vint8m2_t *)(op2), (vint8m2_t *)(op3), (vbool4_t)(op4), (vint8m2_t)(op5), (vint8m2_t)(op6), (vint8m2_t)(op7), (vint8m2_t)(op8), (const int8_t *)(op9), (vuint8m2_t)(op10), (size_t)(op11)) +#define vloxseg4ei8_v_i8mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei8_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (const int8_t *)(op4), (vuint8mf2_t)(op5), (size_t)(op6)) +#define vloxseg4ei8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei8_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vbool16_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (const int8_t *)(op9), (vuint8mf2_t)(op10), (size_t)(op11)) +#define vloxseg4ei8_v_i8mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei8_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (const int8_t *)(op4), (vuint8mf4_t)(op5), (size_t)(op6)) +#define vloxseg4ei8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei8_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vbool32_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (const int8_t *)(op9), (vuint8mf4_t)(op10), (size_t)(op11)) +#define vloxseg4ei8_v_i8mf8(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei8_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (const int8_t *)(op4), (vuint8mf8_t)(op5), (size_t)(op6)) +#define vloxseg4ei8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei8_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vbool64_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (const int8_t *)(op9), (vuint8mf8_t)(op10), (size_t)(op11)) +#define vloxseg4ei8_v_u8m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei8_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (const uint8_t *)(op4), (vuint8m1_t)(op5), (size_t)(op6)) +#define vloxseg4ei8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei8_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vbool8_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (const uint8_t *)(op9), (vuint8m1_t)(op10), (size_t)(op11)) +#define vloxseg4ei8_v_u8m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei8_v_u8m2((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (vuint8m2_t *)(op2), (vuint8m2_t *)(op3), (const uint8_t *)(op4), (vuint8m2_t)(op5), (size_t)(op6)) +#define vloxseg4ei8_v_u8m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei8_v_u8m2_m((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (vuint8m2_t *)(op2), (vuint8m2_t *)(op3), (vbool4_t)(op4), (vuint8m2_t)(op5), (vuint8m2_t)(op6), (vuint8m2_t)(op7), (vuint8m2_t)(op8), (const uint8_t *)(op9), (vuint8m2_t)(op10), (size_t)(op11)) +#define vloxseg4ei8_v_u8mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei8_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (const uint8_t *)(op4), (vuint8mf2_t)(op5), (size_t)(op6)) +#define vloxseg4ei8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei8_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vbool16_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (const uint8_t *)(op9), (vuint8mf2_t)(op10), (size_t)(op11)) +#define vloxseg4ei8_v_u8mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei8_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (const uint8_t *)(op4), (vuint8mf4_t)(op5), (size_t)(op6)) +#define vloxseg4ei8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei8_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vbool32_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (const uint8_t *)(op9), (vuint8mf4_t)(op10), (size_t)(op11)) +#define vloxseg4ei8_v_u8mf8(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei8_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (const uint8_t *)(op4), (vuint8mf8_t)(op5), (size_t)(op6)) +#define vloxseg4ei8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei8_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vbool64_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (const uint8_t *)(op9), (vuint8mf8_t)(op10), (size_t)(op11)) +#define vloxseg5ei8_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei8_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (const int8_t *)(op5), (vuint8m1_t)(op6), (size_t)(op7)) +#define vloxseg5ei8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei8_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vbool8_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (const int8_t *)(op11), (vuint8m1_t)(op12), (size_t)(op13)) +#define vloxseg5ei8_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei8_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (const int8_t *)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vloxseg5ei8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei8_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vbool16_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (const int8_t *)(op11), (vuint8mf2_t)(op12), (size_t)(op13)) +#define vloxseg5ei8_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei8_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (const int8_t *)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vloxseg5ei8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei8_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vbool32_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (const int8_t *)(op11), (vuint8mf4_t)(op12), (size_t)(op13)) +#define vloxseg5ei8_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei8_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (const int8_t *)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vloxseg5ei8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei8_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vbool64_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (const int8_t *)(op11), (vuint8mf8_t)(op12), (size_t)(op13)) +#define vloxseg5ei8_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei8_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (const uint8_t *)(op5), (vuint8m1_t)(op6), (size_t)(op7)) +#define vloxseg5ei8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei8_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vbool8_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (const uint8_t *)(op11), (vuint8m1_t)(op12), (size_t)(op13)) +#define vloxseg5ei8_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei8_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (const uint8_t *)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vloxseg5ei8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei8_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vbool16_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (const uint8_t *)(op11), (vuint8mf2_t)(op12), (size_t)(op13)) +#define vloxseg5ei8_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei8_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (const uint8_t *)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vloxseg5ei8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei8_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vbool32_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (const uint8_t *)(op11), (vuint8mf4_t)(op12), (size_t)(op13)) +#define vloxseg5ei8_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei8_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (const uint8_t *)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vloxseg5ei8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei8_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vbool64_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (const uint8_t *)(op11), (vuint8mf8_t)(op12), (size_t)(op13)) +#define vloxseg6ei8_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei8_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (const int8_t *)(op6), (vuint8m1_t)(op7), (size_t)(op8)) +#define vloxseg6ei8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei8_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vbool8_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (vint8m1_t)(op11), (vint8m1_t)(op12), (const int8_t *)(op13), (vuint8m1_t)(op14), (size_t)(op15)) +#define vloxseg6ei8_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei8_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (const int8_t *)(op6), (vuint8mf2_t)(op7), (size_t)(op8)) +#define vloxseg6ei8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei8_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vbool16_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (vint8mf2_t)(op11), (vint8mf2_t)(op12), (const int8_t *)(op13), (vuint8mf2_t)(op14), (size_t)(op15)) +#define vloxseg6ei8_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei8_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (const int8_t *)(op6), (vuint8mf4_t)(op7), (size_t)(op8)) +#define vloxseg6ei8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei8_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vbool32_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (vint8mf4_t)(op11), (vint8mf4_t)(op12), (const int8_t *)(op13), (vuint8mf4_t)(op14), (size_t)(op15)) +#define vloxseg6ei8_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei8_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (const int8_t *)(op6), (vuint8mf8_t)(op7), (size_t)(op8)) +#define vloxseg6ei8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei8_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vbool64_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (vint8mf8_t)(op11), (vint8mf8_t)(op12), (const int8_t *)(op13), (vuint8mf8_t)(op14), (size_t)(op15)) +#define vloxseg6ei8_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei8_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (const uint8_t *)(op6), (vuint8m1_t)(op7), (size_t)(op8)) +#define vloxseg6ei8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei8_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vbool8_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (vuint8m1_t)(op11), (vuint8m1_t)(op12), (const uint8_t *)(op13), (vuint8m1_t)(op14), (size_t)(op15)) +#define vloxseg6ei8_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei8_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (const uint8_t *)(op6), (vuint8mf2_t)(op7), (size_t)(op8)) +#define vloxseg6ei8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei8_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vbool16_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (vuint8mf2_t)(op11), (vuint8mf2_t)(op12), (const uint8_t *)(op13), (vuint8mf2_t)(op14), (size_t)(op15)) +#define vloxseg6ei8_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei8_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (const uint8_t *)(op6), (vuint8mf4_t)(op7), (size_t)(op8)) +#define vloxseg6ei8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei8_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vbool32_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (vuint8mf4_t)(op11), (vuint8mf4_t)(op12), (const uint8_t *)(op13), (vuint8mf4_t)(op14), (size_t)(op15)) +#define vloxseg6ei8_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei8_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (const uint8_t *)(op6), (vuint8mf8_t)(op7), (size_t)(op8)) +#define vloxseg6ei8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei8_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vbool64_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (vuint8mf8_t)(op11), (vuint8mf8_t)(op12), (const uint8_t *)(op13), (vuint8mf8_t)(op14), (size_t)(op15)) +#define vloxseg7ei8_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei8_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vint8m1_t *)(op6), (const int8_t *)(op7), (vuint8m1_t)(op8), (size_t)(op9)) +#define vloxseg7ei8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei8_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vint8m1_t *)(op6), (vbool8_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (vint8m1_t)(op11), (vint8m1_t)(op12), (vint8m1_t)(op13), (vint8m1_t)(op14), (const int8_t *)(op15), (vuint8m1_t)(op16), (size_t)(op17)) +#define vloxseg7ei8_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei8_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vint8mf2_t *)(op6), (const int8_t *)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vloxseg7ei8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei8_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vint8mf2_t *)(op6), (vbool16_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (vint8mf2_t)(op11), (vint8mf2_t)(op12), (vint8mf2_t)(op13), (vint8mf2_t)(op14), (const int8_t *)(op15), (vuint8mf2_t)(op16), (size_t)(op17)) +#define vloxseg7ei8_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei8_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vint8mf4_t *)(op6), (const int8_t *)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vloxseg7ei8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei8_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vint8mf4_t *)(op6), (vbool32_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (vint8mf4_t)(op11), (vint8mf4_t)(op12), (vint8mf4_t)(op13), (vint8mf4_t)(op14), (const int8_t *)(op15), (vuint8mf4_t)(op16), (size_t)(op17)) +#define vloxseg7ei8_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei8_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vint8mf8_t *)(op6), (const int8_t *)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vloxseg7ei8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei8_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vint8mf8_t *)(op6), (vbool64_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (vint8mf8_t)(op11), (vint8mf8_t)(op12), (vint8mf8_t)(op13), (vint8mf8_t)(op14), (const int8_t *)(op15), (vuint8mf8_t)(op16), (size_t)(op17)) +#define vloxseg7ei8_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei8_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vuint8m1_t *)(op6), (const uint8_t *)(op7), (vuint8m1_t)(op8), (size_t)(op9)) +#define vloxseg7ei8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei8_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vuint8m1_t *)(op6), (vbool8_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (vuint8m1_t)(op11), (vuint8m1_t)(op12), (vuint8m1_t)(op13), (vuint8m1_t)(op14), (const uint8_t *)(op15), (vuint8m1_t)(op16), (size_t)(op17)) +#define vloxseg7ei8_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei8_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vuint8mf2_t *)(op6), (const uint8_t *)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vloxseg7ei8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei8_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vuint8mf2_t *)(op6), (vbool16_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (vuint8mf2_t)(op11), (vuint8mf2_t)(op12), (vuint8mf2_t)(op13), (vuint8mf2_t)(op14), (const uint8_t *)(op15), (vuint8mf2_t)(op16), (size_t)(op17)) +#define vloxseg7ei8_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei8_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vuint8mf4_t *)(op6), (const uint8_t *)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vloxseg7ei8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei8_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vuint8mf4_t *)(op6), (vbool32_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (vuint8mf4_t)(op11), (vuint8mf4_t)(op12), (vuint8mf4_t)(op13), (vuint8mf4_t)(op14), (const uint8_t *)(op15), (vuint8mf4_t)(op16), (size_t)(op17)) +#define vloxseg7ei8_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei8_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vuint8mf8_t *)(op6), (const uint8_t *)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vloxseg7ei8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei8_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vuint8mf8_t *)(op6), (vbool64_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (vuint8mf8_t)(op11), (vuint8mf8_t)(op12), (vuint8mf8_t)(op13), (vuint8mf8_t)(op14), (const uint8_t *)(op15), (vuint8mf8_t)(op16), (size_t)(op17)) +#define vloxseg8ei8_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei8_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vint8m1_t *)(op6), (vint8m1_t *)(op7), (const int8_t *)(op8), (vuint8m1_t)(op9), (size_t)(op10)) +#define vloxseg8ei8_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei8_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vint8m1_t *)(op6), (vint8m1_t *)(op7), (vbool8_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (vint8m1_t)(op11), (vint8m1_t)(op12), (vint8m1_t)(op13), (vint8m1_t)(op14), (vint8m1_t)(op15), (vint8m1_t)(op16), (const int8_t *)(op17), (vuint8m1_t)(op18), (size_t)(op19)) +#define vloxseg8ei8_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei8_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vint8mf2_t *)(op6), (vint8mf2_t *)(op7), (const int8_t *)(op8), (vuint8mf2_t)(op9), (size_t)(op10)) +#define vloxseg8ei8_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei8_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vint8mf2_t *)(op6), (vint8mf2_t *)(op7), (vbool16_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (vint8mf2_t)(op11), (vint8mf2_t)(op12), (vint8mf2_t)(op13), (vint8mf2_t)(op14), (vint8mf2_t)(op15), (vint8mf2_t)(op16), (const int8_t *)(op17), (vuint8mf2_t)(op18), (size_t)(op19)) +#define vloxseg8ei8_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei8_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vint8mf4_t *)(op6), (vint8mf4_t *)(op7), (const int8_t *)(op8), (vuint8mf4_t)(op9), (size_t)(op10)) +#define vloxseg8ei8_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei8_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vint8mf4_t *)(op6), (vint8mf4_t *)(op7), (vbool32_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (vint8mf4_t)(op11), (vint8mf4_t)(op12), (vint8mf4_t)(op13), (vint8mf4_t)(op14), (vint8mf4_t)(op15), (vint8mf4_t)(op16), (const int8_t *)(op17), (vuint8mf4_t)(op18), (size_t)(op19)) +#define vloxseg8ei8_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei8_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vint8mf8_t *)(op6), (vint8mf8_t *)(op7), (const int8_t *)(op8), (vuint8mf8_t)(op9), (size_t)(op10)) +#define vloxseg8ei8_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei8_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vint8mf8_t *)(op6), (vint8mf8_t *)(op7), (vbool64_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (vint8mf8_t)(op11), (vint8mf8_t)(op12), (vint8mf8_t)(op13), (vint8mf8_t)(op14), (vint8mf8_t)(op15), (vint8mf8_t)(op16), (const int8_t *)(op17), (vuint8mf8_t)(op18), (size_t)(op19)) +#define vloxseg8ei8_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei8_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vuint8m1_t *)(op6), (vuint8m1_t *)(op7), (const uint8_t *)(op8), (vuint8m1_t)(op9), (size_t)(op10)) +#define vloxseg8ei8_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei8_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vuint8m1_t *)(op6), (vuint8m1_t *)(op7), (vbool8_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (vuint8m1_t)(op11), (vuint8m1_t)(op12), (vuint8m1_t)(op13), (vuint8m1_t)(op14), (vuint8m1_t)(op15), (vuint8m1_t)(op16), (const uint8_t *)(op17), (vuint8m1_t)(op18), (size_t)(op19)) +#define vloxseg8ei8_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei8_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vuint8mf2_t *)(op6), (vuint8mf2_t *)(op7), (const uint8_t *)(op8), (vuint8mf2_t)(op9), (size_t)(op10)) +#define vloxseg8ei8_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei8_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vuint8mf2_t *)(op6), (vuint8mf2_t *)(op7), (vbool16_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (vuint8mf2_t)(op11), (vuint8mf2_t)(op12), (vuint8mf2_t)(op13), (vuint8mf2_t)(op14), (vuint8mf2_t)(op15), (vuint8mf2_t)(op16), (const uint8_t *)(op17), (vuint8mf2_t)(op18), (size_t)(op19)) +#define vloxseg8ei8_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei8_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vuint8mf4_t *)(op6), (vuint8mf4_t *)(op7), (const uint8_t *)(op8), (vuint8mf4_t)(op9), (size_t)(op10)) +#define vloxseg8ei8_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei8_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vuint8mf4_t *)(op6), (vuint8mf4_t *)(op7), (vbool32_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (vuint8mf4_t)(op11), (vuint8mf4_t)(op12), (vuint8mf4_t)(op13), (vuint8mf4_t)(op14), (vuint8mf4_t)(op15), (vuint8mf4_t)(op16), (const uint8_t *)(op17), (vuint8mf4_t)(op18), (size_t)(op19)) +#define vloxseg8ei8_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei8_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vuint8mf8_t *)(op6), (vuint8mf8_t *)(op7), (const uint8_t *)(op8), (vuint8mf8_t)(op9), (size_t)(op10)) +#define vloxseg8ei8_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei8_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vuint8mf8_t *)(op6), (vuint8mf8_t *)(op7), (vbool64_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (vuint8mf8_t)(op11), (vuint8mf8_t)(op12), (vuint8mf8_t)(op13), (vuint8mf8_t)(op14), (vuint8mf8_t)(op15), (vuint8mf8_t)(op16), (const uint8_t *)(op17), (vuint8mf8_t)(op18), (size_t)(op19)) +#define vloxseg2ei16_v_i8m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (const int8_t *)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vbool8_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (const int8_t *)(op5), (vuint16m2_t)(op6), (size_t)(op7)) +#define vloxseg2ei16_v_i8m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_i8m2((vint8m2_t *)(op0), (vint8m2_t *)(op1), (const int8_t *)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_i8m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_i8m2_m((vint8m2_t *)(op0), (vint8m2_t *)(op1), (vbool4_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (const int8_t *)(op5), (vuint16m4_t)(op6), (size_t)(op7)) +#define vloxseg2ei16_v_i8m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_i8m4((vint8m4_t *)(op0), (vint8m4_t *)(op1), (const int8_t *)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_i8m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_i8m4_m((vint8m4_t *)(op0), (vint8m4_t *)(op1), (vbool2_t)(op2), (vint8m4_t)(op3), (vint8m4_t)(op4), (const int8_t *)(op5), (vuint16m8_t)(op6), (size_t)(op7)) +#define vloxseg2ei16_v_i8mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (const int8_t *)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vbool16_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (const int8_t *)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vloxseg2ei16_v_i8mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (const int8_t *)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vbool32_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (const int8_t *)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vloxseg2ei16_v_i8mf8(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (const int8_t *)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vbool64_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (const int8_t *)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vloxseg2ei16_v_u8m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (const uint8_t *)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vbool8_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (const uint8_t *)(op5), (vuint16m2_t)(op6), (size_t)(op7)) +#define vloxseg2ei16_v_u8m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_u8m2((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (const uint8_t *)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_u8m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_u8m2_m((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (vbool4_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (const uint8_t *)(op5), (vuint16m4_t)(op6), (size_t)(op7)) +#define vloxseg2ei16_v_u8m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_u8m4((vuint8m4_t *)(op0), (vuint8m4_t *)(op1), (const uint8_t *)(op2), (vuint16m8_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_u8m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_u8m4_m((vuint8m4_t *)(op0), (vuint8m4_t *)(op1), (vbool2_t)(op2), (vuint8m4_t)(op3), (vuint8m4_t)(op4), (const uint8_t *)(op5), (vuint16m8_t)(op6), (size_t)(op7)) +#define vloxseg2ei16_v_u8mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (const uint8_t *)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vbool16_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (const uint8_t *)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vloxseg2ei16_v_u8mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (const uint8_t *)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vbool32_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (const uint8_t *)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vloxseg2ei16_v_u8mf8(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (const uint8_t *)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vbool64_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (const uint8_t *)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vloxseg3ei16_v_i8m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei16_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (const int8_t *)(op3), (vuint16m2_t)(op4), (size_t)(op5)) +#define vloxseg3ei16_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei16_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vbool8_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (const int8_t *)(op7), (vuint16m2_t)(op8), (size_t)(op9)) +#define vloxseg3ei16_v_i8m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei16_v_i8m2((vint8m2_t *)(op0), (vint8m2_t *)(op1), (vint8m2_t *)(op2), (const int8_t *)(op3), (vuint16m4_t)(op4), (size_t)(op5)) +#define vloxseg3ei16_v_i8m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei16_v_i8m2_m((vint8m2_t *)(op0), (vint8m2_t *)(op1), (vint8m2_t *)(op2), (vbool4_t)(op3), (vint8m2_t)(op4), (vint8m2_t)(op5), (vint8m2_t)(op6), (const int8_t *)(op7), (vuint16m4_t)(op8), (size_t)(op9)) +#define vloxseg3ei16_v_i8mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei16_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (const int8_t *)(op3), (vuint16m1_t)(op4), (size_t)(op5)) +#define vloxseg3ei16_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei16_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vbool16_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (const int8_t *)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vloxseg3ei16_v_i8mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei16_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (const int8_t *)(op3), (vuint16mf2_t)(op4), (size_t)(op5)) +#define vloxseg3ei16_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei16_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vbool32_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (const int8_t *)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vloxseg3ei16_v_i8mf8(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei16_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (const int8_t *)(op3), (vuint16mf4_t)(op4), (size_t)(op5)) +#define vloxseg3ei16_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei16_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vbool64_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (const int8_t *)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vloxseg3ei16_v_u8m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei16_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (const uint8_t *)(op3), (vuint16m2_t)(op4), (size_t)(op5)) +#define vloxseg3ei16_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei16_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vbool8_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (const uint8_t *)(op7), (vuint16m2_t)(op8), (size_t)(op9)) +#define vloxseg3ei16_v_u8m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei16_v_u8m2((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (vuint8m2_t *)(op2), (const uint8_t *)(op3), (vuint16m4_t)(op4), (size_t)(op5)) +#define vloxseg3ei16_v_u8m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei16_v_u8m2_m((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (vuint8m2_t *)(op2), (vbool4_t)(op3), (vuint8m2_t)(op4), (vuint8m2_t)(op5), (vuint8m2_t)(op6), (const uint8_t *)(op7), (vuint16m4_t)(op8), (size_t)(op9)) +#define vloxseg3ei16_v_u8mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei16_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (const uint8_t *)(op3), (vuint16m1_t)(op4), (size_t)(op5)) +#define vloxseg3ei16_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei16_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vbool16_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (const uint8_t *)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vloxseg3ei16_v_u8mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei16_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (const uint8_t *)(op3), (vuint16mf2_t)(op4), (size_t)(op5)) +#define vloxseg3ei16_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei16_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vbool32_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (const uint8_t *)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vloxseg3ei16_v_u8mf8(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei16_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (const uint8_t *)(op3), (vuint16mf4_t)(op4), (size_t)(op5)) +#define vloxseg3ei16_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei16_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vbool64_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (const uint8_t *)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vloxseg4ei16_v_i8m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei16_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (const int8_t *)(op4), (vuint16m2_t)(op5), (size_t)(op6)) +#define vloxseg4ei16_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei16_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vbool8_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (const int8_t *)(op9), (vuint16m2_t)(op10), (size_t)(op11)) +#define vloxseg4ei16_v_i8m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei16_v_i8m2((vint8m2_t *)(op0), (vint8m2_t *)(op1), (vint8m2_t *)(op2), (vint8m2_t *)(op3), (const int8_t *)(op4), (vuint16m4_t)(op5), (size_t)(op6)) +#define vloxseg4ei16_v_i8m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei16_v_i8m2_m((vint8m2_t *)(op0), (vint8m2_t *)(op1), (vint8m2_t *)(op2), (vint8m2_t *)(op3), (vbool4_t)(op4), (vint8m2_t)(op5), (vint8m2_t)(op6), (vint8m2_t)(op7), (vint8m2_t)(op8), (const int8_t *)(op9), (vuint16m4_t)(op10), (size_t)(op11)) +#define vloxseg4ei16_v_i8mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei16_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (const int8_t *)(op4), (vuint16m1_t)(op5), (size_t)(op6)) +#define vloxseg4ei16_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei16_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vbool16_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (const int8_t *)(op9), (vuint16m1_t)(op10), (size_t)(op11)) +#define vloxseg4ei16_v_i8mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei16_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (const int8_t *)(op4), (vuint16mf2_t)(op5), (size_t)(op6)) +#define vloxseg4ei16_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei16_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vbool32_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (const int8_t *)(op9), (vuint16mf2_t)(op10), (size_t)(op11)) +#define vloxseg4ei16_v_i8mf8(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei16_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (const int8_t *)(op4), (vuint16mf4_t)(op5), (size_t)(op6)) +#define vloxseg4ei16_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei16_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vbool64_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (const int8_t *)(op9), (vuint16mf4_t)(op10), (size_t)(op11)) +#define vloxseg4ei16_v_u8m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei16_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (const uint8_t *)(op4), (vuint16m2_t)(op5), (size_t)(op6)) +#define vloxseg4ei16_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei16_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vbool8_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (const uint8_t *)(op9), (vuint16m2_t)(op10), (size_t)(op11)) +#define vloxseg4ei16_v_u8m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei16_v_u8m2((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (vuint8m2_t *)(op2), (vuint8m2_t *)(op3), (const uint8_t *)(op4), (vuint16m4_t)(op5), (size_t)(op6)) +#define vloxseg4ei16_v_u8m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei16_v_u8m2_m((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (vuint8m2_t *)(op2), (vuint8m2_t *)(op3), (vbool4_t)(op4), (vuint8m2_t)(op5), (vuint8m2_t)(op6), (vuint8m2_t)(op7), (vuint8m2_t)(op8), (const uint8_t *)(op9), (vuint16m4_t)(op10), (size_t)(op11)) +#define vloxseg4ei16_v_u8mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei16_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (const uint8_t *)(op4), (vuint16m1_t)(op5), (size_t)(op6)) +#define vloxseg4ei16_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei16_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vbool16_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (const uint8_t *)(op9), (vuint16m1_t)(op10), (size_t)(op11)) +#define vloxseg4ei16_v_u8mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei16_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (const uint8_t *)(op4), (vuint16mf2_t)(op5), (size_t)(op6)) +#define vloxseg4ei16_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei16_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vbool32_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (const uint8_t *)(op9), (vuint16mf2_t)(op10), (size_t)(op11)) +#define vloxseg4ei16_v_u8mf8(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei16_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (const uint8_t *)(op4), (vuint16mf4_t)(op5), (size_t)(op6)) +#define vloxseg4ei16_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei16_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vbool64_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (const uint8_t *)(op9), (vuint16mf4_t)(op10), (size_t)(op11)) +#define vloxseg5ei16_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei16_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (const int8_t *)(op5), (vuint16m2_t)(op6), (size_t)(op7)) +#define vloxseg5ei16_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei16_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vbool8_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (const int8_t *)(op11), (vuint16m2_t)(op12), (size_t)(op13)) +#define vloxseg5ei16_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei16_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (const int8_t *)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vloxseg5ei16_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei16_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vbool16_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (const int8_t *)(op11), (vuint16m1_t)(op12), (size_t)(op13)) +#define vloxseg5ei16_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei16_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (const int8_t *)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vloxseg5ei16_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei16_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vbool32_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (const int8_t *)(op11), (vuint16mf2_t)(op12), (size_t)(op13)) +#define vloxseg5ei16_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei16_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (const int8_t *)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vloxseg5ei16_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei16_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vbool64_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (const int8_t *)(op11), (vuint16mf4_t)(op12), (size_t)(op13)) +#define vloxseg5ei16_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei16_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (const uint8_t *)(op5), (vuint16m2_t)(op6), (size_t)(op7)) +#define vloxseg5ei16_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei16_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vbool8_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (const uint8_t *)(op11), (vuint16m2_t)(op12), (size_t)(op13)) +#define vloxseg5ei16_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei16_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (const uint8_t *)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vloxseg5ei16_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei16_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vbool16_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (const uint8_t *)(op11), (vuint16m1_t)(op12), (size_t)(op13)) +#define vloxseg5ei16_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei16_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (const uint8_t *)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vloxseg5ei16_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei16_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vbool32_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (const uint8_t *)(op11), (vuint16mf2_t)(op12), (size_t)(op13)) +#define vloxseg5ei16_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei16_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (const uint8_t *)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vloxseg5ei16_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei16_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vbool64_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (const uint8_t *)(op11), (vuint16mf4_t)(op12), (size_t)(op13)) +#define vloxseg6ei16_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei16_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (const int8_t *)(op6), (vuint16m2_t)(op7), (size_t)(op8)) +#define vloxseg6ei16_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei16_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vbool8_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (vint8m1_t)(op11), (vint8m1_t)(op12), (const int8_t *)(op13), (vuint16m2_t)(op14), (size_t)(op15)) +#define vloxseg6ei16_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei16_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (const int8_t *)(op6), (vuint16m1_t)(op7), (size_t)(op8)) +#define vloxseg6ei16_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei16_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vbool16_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (vint8mf2_t)(op11), (vint8mf2_t)(op12), (const int8_t *)(op13), (vuint16m1_t)(op14), (size_t)(op15)) +#define vloxseg6ei16_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei16_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (const int8_t *)(op6), (vuint16mf2_t)(op7), (size_t)(op8)) +#define vloxseg6ei16_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei16_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vbool32_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (vint8mf4_t)(op11), (vint8mf4_t)(op12), (const int8_t *)(op13), (vuint16mf2_t)(op14), (size_t)(op15)) +#define vloxseg6ei16_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei16_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (const int8_t *)(op6), (vuint16mf4_t)(op7), (size_t)(op8)) +#define vloxseg6ei16_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei16_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vbool64_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (vint8mf8_t)(op11), (vint8mf8_t)(op12), (const int8_t *)(op13), (vuint16mf4_t)(op14), (size_t)(op15)) +#define vloxseg6ei16_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei16_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (const uint8_t *)(op6), (vuint16m2_t)(op7), (size_t)(op8)) +#define vloxseg6ei16_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei16_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vbool8_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (vuint8m1_t)(op11), (vuint8m1_t)(op12), (const uint8_t *)(op13), (vuint16m2_t)(op14), (size_t)(op15)) +#define vloxseg6ei16_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei16_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (const uint8_t *)(op6), (vuint16m1_t)(op7), (size_t)(op8)) +#define vloxseg6ei16_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei16_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vbool16_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (vuint8mf2_t)(op11), (vuint8mf2_t)(op12), (const uint8_t *)(op13), (vuint16m1_t)(op14), (size_t)(op15)) +#define vloxseg6ei16_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei16_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (const uint8_t *)(op6), (vuint16mf2_t)(op7), (size_t)(op8)) +#define vloxseg6ei16_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei16_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vbool32_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (vuint8mf4_t)(op11), (vuint8mf4_t)(op12), (const uint8_t *)(op13), (vuint16mf2_t)(op14), (size_t)(op15)) +#define vloxseg6ei16_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei16_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (const uint8_t *)(op6), (vuint16mf4_t)(op7), (size_t)(op8)) +#define vloxseg6ei16_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei16_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vbool64_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (vuint8mf8_t)(op11), (vuint8mf8_t)(op12), (const uint8_t *)(op13), (vuint16mf4_t)(op14), (size_t)(op15)) +#define vloxseg7ei16_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei16_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vint8m1_t *)(op6), (const int8_t *)(op7), (vuint16m2_t)(op8), (size_t)(op9)) +#define vloxseg7ei16_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei16_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vint8m1_t *)(op6), (vbool8_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (vint8m1_t)(op11), (vint8m1_t)(op12), (vint8m1_t)(op13), (vint8m1_t)(op14), (const int8_t *)(op15), (vuint16m2_t)(op16), (size_t)(op17)) +#define vloxseg7ei16_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei16_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vint8mf2_t *)(op6), (const int8_t *)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vloxseg7ei16_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei16_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vint8mf2_t *)(op6), (vbool16_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (vint8mf2_t)(op11), (vint8mf2_t)(op12), (vint8mf2_t)(op13), (vint8mf2_t)(op14), (const int8_t *)(op15), (vuint16m1_t)(op16), (size_t)(op17)) +#define vloxseg7ei16_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei16_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vint8mf4_t *)(op6), (const int8_t *)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vloxseg7ei16_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei16_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vint8mf4_t *)(op6), (vbool32_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (vint8mf4_t)(op11), (vint8mf4_t)(op12), (vint8mf4_t)(op13), (vint8mf4_t)(op14), (const int8_t *)(op15), (vuint16mf2_t)(op16), (size_t)(op17)) +#define vloxseg7ei16_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei16_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vint8mf8_t *)(op6), (const int8_t *)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vloxseg7ei16_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei16_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vint8mf8_t *)(op6), (vbool64_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (vint8mf8_t)(op11), (vint8mf8_t)(op12), (vint8mf8_t)(op13), (vint8mf8_t)(op14), (const int8_t *)(op15), (vuint16mf4_t)(op16), (size_t)(op17)) +#define vloxseg7ei16_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei16_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vuint8m1_t *)(op6), (const uint8_t *)(op7), (vuint16m2_t)(op8), (size_t)(op9)) +#define vloxseg7ei16_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei16_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vuint8m1_t *)(op6), (vbool8_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (vuint8m1_t)(op11), (vuint8m1_t)(op12), (vuint8m1_t)(op13), (vuint8m1_t)(op14), (const uint8_t *)(op15), (vuint16m2_t)(op16), (size_t)(op17)) +#define vloxseg7ei16_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei16_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vuint8mf2_t *)(op6), (const uint8_t *)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vloxseg7ei16_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei16_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vuint8mf2_t *)(op6), (vbool16_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (vuint8mf2_t)(op11), (vuint8mf2_t)(op12), (vuint8mf2_t)(op13), (vuint8mf2_t)(op14), (const uint8_t *)(op15), (vuint16m1_t)(op16), (size_t)(op17)) +#define vloxseg7ei16_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei16_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vuint8mf4_t *)(op6), (const uint8_t *)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vloxseg7ei16_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei16_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vuint8mf4_t *)(op6), (vbool32_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (vuint8mf4_t)(op11), (vuint8mf4_t)(op12), (vuint8mf4_t)(op13), (vuint8mf4_t)(op14), (const uint8_t *)(op15), (vuint16mf2_t)(op16), (size_t)(op17)) +#define vloxseg7ei16_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei16_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vuint8mf8_t *)(op6), (const uint8_t *)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vloxseg7ei16_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei16_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vuint8mf8_t *)(op6), (vbool64_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (vuint8mf8_t)(op11), (vuint8mf8_t)(op12), (vuint8mf8_t)(op13), (vuint8mf8_t)(op14), (const uint8_t *)(op15), (vuint16mf4_t)(op16), (size_t)(op17)) +#define vloxseg8ei16_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei16_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vint8m1_t *)(op6), (vint8m1_t *)(op7), (const int8_t *)(op8), (vuint16m2_t)(op9), (size_t)(op10)) +#define vloxseg8ei16_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei16_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vint8m1_t *)(op6), (vint8m1_t *)(op7), (vbool8_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (vint8m1_t)(op11), (vint8m1_t)(op12), (vint8m1_t)(op13), (vint8m1_t)(op14), (vint8m1_t)(op15), (vint8m1_t)(op16), (const int8_t *)(op17), (vuint16m2_t)(op18), (size_t)(op19)) +#define vloxseg8ei16_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei16_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vint8mf2_t *)(op6), (vint8mf2_t *)(op7), (const int8_t *)(op8), (vuint16m1_t)(op9), (size_t)(op10)) +#define vloxseg8ei16_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei16_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vint8mf2_t *)(op6), (vint8mf2_t *)(op7), (vbool16_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (vint8mf2_t)(op11), (vint8mf2_t)(op12), (vint8mf2_t)(op13), (vint8mf2_t)(op14), (vint8mf2_t)(op15), (vint8mf2_t)(op16), (const int8_t *)(op17), (vuint16m1_t)(op18), (size_t)(op19)) +#define vloxseg8ei16_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei16_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vint8mf4_t *)(op6), (vint8mf4_t *)(op7), (const int8_t *)(op8), (vuint16mf2_t)(op9), (size_t)(op10)) +#define vloxseg8ei16_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei16_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vint8mf4_t *)(op6), (vint8mf4_t *)(op7), (vbool32_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (vint8mf4_t)(op11), (vint8mf4_t)(op12), (vint8mf4_t)(op13), (vint8mf4_t)(op14), (vint8mf4_t)(op15), (vint8mf4_t)(op16), (const int8_t *)(op17), (vuint16mf2_t)(op18), (size_t)(op19)) +#define vloxseg8ei16_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei16_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vint8mf8_t *)(op6), (vint8mf8_t *)(op7), (const int8_t *)(op8), (vuint16mf4_t)(op9), (size_t)(op10)) +#define vloxseg8ei16_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei16_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vint8mf8_t *)(op6), (vint8mf8_t *)(op7), (vbool64_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (vint8mf8_t)(op11), (vint8mf8_t)(op12), (vint8mf8_t)(op13), (vint8mf8_t)(op14), (vint8mf8_t)(op15), (vint8mf8_t)(op16), (const int8_t *)(op17), (vuint16mf4_t)(op18), (size_t)(op19)) +#define vloxseg8ei16_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei16_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vuint8m1_t *)(op6), (vuint8m1_t *)(op7), (const uint8_t *)(op8), (vuint16m2_t)(op9), (size_t)(op10)) +#define vloxseg8ei16_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei16_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vuint8m1_t *)(op6), (vuint8m1_t *)(op7), (vbool8_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (vuint8m1_t)(op11), (vuint8m1_t)(op12), (vuint8m1_t)(op13), (vuint8m1_t)(op14), (vuint8m1_t)(op15), (vuint8m1_t)(op16), (const uint8_t *)(op17), (vuint16m2_t)(op18), (size_t)(op19)) +#define vloxseg8ei16_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei16_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vuint8mf2_t *)(op6), (vuint8mf2_t *)(op7), (const uint8_t *)(op8), (vuint16m1_t)(op9), (size_t)(op10)) +#define vloxseg8ei16_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei16_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vuint8mf2_t *)(op6), (vuint8mf2_t *)(op7), (vbool16_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (vuint8mf2_t)(op11), (vuint8mf2_t)(op12), (vuint8mf2_t)(op13), (vuint8mf2_t)(op14), (vuint8mf2_t)(op15), (vuint8mf2_t)(op16), (const uint8_t *)(op17), (vuint16m1_t)(op18), (size_t)(op19)) +#define vloxseg8ei16_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei16_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vuint8mf4_t *)(op6), (vuint8mf4_t *)(op7), (const uint8_t *)(op8), (vuint16mf2_t)(op9), (size_t)(op10)) +#define vloxseg8ei16_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei16_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vuint8mf4_t *)(op6), (vuint8mf4_t *)(op7), (vbool32_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (vuint8mf4_t)(op11), (vuint8mf4_t)(op12), (vuint8mf4_t)(op13), (vuint8mf4_t)(op14), (vuint8mf4_t)(op15), (vuint8mf4_t)(op16), (const uint8_t *)(op17), (vuint16mf2_t)(op18), (size_t)(op19)) +#define vloxseg8ei16_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei16_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vuint8mf8_t *)(op6), (vuint8mf8_t *)(op7), (const uint8_t *)(op8), (vuint16mf4_t)(op9), (size_t)(op10)) +#define vloxseg8ei16_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei16_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vuint8mf8_t *)(op6), (vuint8mf8_t *)(op7), (vbool64_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (vuint8mf8_t)(op11), (vuint8mf8_t)(op12), (vuint8mf8_t)(op13), (vuint8mf8_t)(op14), (vuint8mf8_t)(op15), (vuint8mf8_t)(op16), (const uint8_t *)(op17), (vuint16mf4_t)(op18), (size_t)(op19)) +#define vloxseg2ei32_v_i8m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (const int8_t *)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vbool8_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (const int8_t *)(op5), (vuint32m4_t)(op6), (size_t)(op7)) +#define vloxseg2ei32_v_i8m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_i8m2((vint8m2_t *)(op0), (vint8m2_t *)(op1), (const int8_t *)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_i8m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_i8m2_m((vint8m2_t *)(op0), (vint8m2_t *)(op1), (vbool4_t)(op2), (vint8m2_t)(op3), (vint8m2_t)(op4), (const int8_t *)(op5), (vuint32m8_t)(op6), (size_t)(op7)) +#define vloxseg2ei32_v_i8mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (const int8_t *)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vbool16_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (const int8_t *)(op5), (vuint32m2_t)(op6), (size_t)(op7)) +#define vloxseg2ei32_v_i8mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (const int8_t *)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vbool32_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (const int8_t *)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vloxseg2ei32_v_i8mf8(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (const int8_t *)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vbool64_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (const int8_t *)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vloxseg2ei32_v_u8m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (const uint8_t *)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vbool8_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (const uint8_t *)(op5), (vuint32m4_t)(op6), (size_t)(op7)) +#define vloxseg2ei32_v_u8m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_u8m2((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (const uint8_t *)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_u8m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_u8m2_m((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (vbool4_t)(op2), (vuint8m2_t)(op3), (vuint8m2_t)(op4), (const uint8_t *)(op5), (vuint32m8_t)(op6), (size_t)(op7)) +#define vloxseg2ei32_v_u8mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (const uint8_t *)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vbool16_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (const uint8_t *)(op5), (vuint32m2_t)(op6), (size_t)(op7)) +#define vloxseg2ei32_v_u8mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (const uint8_t *)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vbool32_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (const uint8_t *)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vloxseg2ei32_v_u8mf8(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (const uint8_t *)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vbool64_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (const uint8_t *)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vloxseg3ei32_v_i8m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei32_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (const int8_t *)(op3), (vuint32m4_t)(op4), (size_t)(op5)) +#define vloxseg3ei32_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei32_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vbool8_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (const int8_t *)(op7), (vuint32m4_t)(op8), (size_t)(op9)) +#define vloxseg3ei32_v_i8m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei32_v_i8m2((vint8m2_t *)(op0), (vint8m2_t *)(op1), (vint8m2_t *)(op2), (const int8_t *)(op3), (vuint32m8_t)(op4), (size_t)(op5)) +#define vloxseg3ei32_v_i8m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei32_v_i8m2_m((vint8m2_t *)(op0), (vint8m2_t *)(op1), (vint8m2_t *)(op2), (vbool4_t)(op3), (vint8m2_t)(op4), (vint8m2_t)(op5), (vint8m2_t)(op6), (const int8_t *)(op7), (vuint32m8_t)(op8), (size_t)(op9)) +#define vloxseg3ei32_v_i8mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei32_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (const int8_t *)(op3), (vuint32m2_t)(op4), (size_t)(op5)) +#define vloxseg3ei32_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei32_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vbool16_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (const int8_t *)(op7), (vuint32m2_t)(op8), (size_t)(op9)) +#define vloxseg3ei32_v_i8mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei32_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (const int8_t *)(op3), (vuint32m1_t)(op4), (size_t)(op5)) +#define vloxseg3ei32_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei32_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vbool32_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (const int8_t *)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vloxseg3ei32_v_i8mf8(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei32_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (const int8_t *)(op3), (vuint32mf2_t)(op4), (size_t)(op5)) +#define vloxseg3ei32_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei32_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vbool64_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (const int8_t *)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vloxseg3ei32_v_u8m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei32_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (const uint8_t *)(op3), (vuint32m4_t)(op4), (size_t)(op5)) +#define vloxseg3ei32_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei32_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vbool8_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (const uint8_t *)(op7), (vuint32m4_t)(op8), (size_t)(op9)) +#define vloxseg3ei32_v_u8m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei32_v_u8m2((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (vuint8m2_t *)(op2), (const uint8_t *)(op3), (vuint32m8_t)(op4), (size_t)(op5)) +#define vloxseg3ei32_v_u8m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei32_v_u8m2_m((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (vuint8m2_t *)(op2), (vbool4_t)(op3), (vuint8m2_t)(op4), (vuint8m2_t)(op5), (vuint8m2_t)(op6), (const uint8_t *)(op7), (vuint32m8_t)(op8), (size_t)(op9)) +#define vloxseg3ei32_v_u8mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei32_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (const uint8_t *)(op3), (vuint32m2_t)(op4), (size_t)(op5)) +#define vloxseg3ei32_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei32_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vbool16_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (const uint8_t *)(op7), (vuint32m2_t)(op8), (size_t)(op9)) +#define vloxseg3ei32_v_u8mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei32_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (const uint8_t *)(op3), (vuint32m1_t)(op4), (size_t)(op5)) +#define vloxseg3ei32_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei32_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vbool32_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (const uint8_t *)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vloxseg3ei32_v_u8mf8(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei32_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (const uint8_t *)(op3), (vuint32mf2_t)(op4), (size_t)(op5)) +#define vloxseg3ei32_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei32_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vbool64_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (const uint8_t *)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vloxseg4ei32_v_i8m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei32_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (const int8_t *)(op4), (vuint32m4_t)(op5), (size_t)(op6)) +#define vloxseg4ei32_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei32_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vbool8_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (const int8_t *)(op9), (vuint32m4_t)(op10), (size_t)(op11)) +#define vloxseg4ei32_v_i8m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei32_v_i8m2((vint8m2_t *)(op0), (vint8m2_t *)(op1), (vint8m2_t *)(op2), (vint8m2_t *)(op3), (const int8_t *)(op4), (vuint32m8_t)(op5), (size_t)(op6)) +#define vloxseg4ei32_v_i8m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei32_v_i8m2_m((vint8m2_t *)(op0), (vint8m2_t *)(op1), (vint8m2_t *)(op2), (vint8m2_t *)(op3), (vbool4_t)(op4), (vint8m2_t)(op5), (vint8m2_t)(op6), (vint8m2_t)(op7), (vint8m2_t)(op8), (const int8_t *)(op9), (vuint32m8_t)(op10), (size_t)(op11)) +#define vloxseg4ei32_v_i8mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei32_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (const int8_t *)(op4), (vuint32m2_t)(op5), (size_t)(op6)) +#define vloxseg4ei32_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei32_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vbool16_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (const int8_t *)(op9), (vuint32m2_t)(op10), (size_t)(op11)) +#define vloxseg4ei32_v_i8mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei32_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (const int8_t *)(op4), (vuint32m1_t)(op5), (size_t)(op6)) +#define vloxseg4ei32_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei32_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vbool32_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (const int8_t *)(op9), (vuint32m1_t)(op10), (size_t)(op11)) +#define vloxseg4ei32_v_i8mf8(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei32_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (const int8_t *)(op4), (vuint32mf2_t)(op5), (size_t)(op6)) +#define vloxseg4ei32_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei32_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vbool64_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (const int8_t *)(op9), (vuint32mf2_t)(op10), (size_t)(op11)) +#define vloxseg4ei32_v_u8m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei32_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (const uint8_t *)(op4), (vuint32m4_t)(op5), (size_t)(op6)) +#define vloxseg4ei32_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei32_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vbool8_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (const uint8_t *)(op9), (vuint32m4_t)(op10), (size_t)(op11)) +#define vloxseg4ei32_v_u8m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei32_v_u8m2((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (vuint8m2_t *)(op2), (vuint8m2_t *)(op3), (const uint8_t *)(op4), (vuint32m8_t)(op5), (size_t)(op6)) +#define vloxseg4ei32_v_u8m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei32_v_u8m2_m((vuint8m2_t *)(op0), (vuint8m2_t *)(op1), (vuint8m2_t *)(op2), (vuint8m2_t *)(op3), (vbool4_t)(op4), (vuint8m2_t)(op5), (vuint8m2_t)(op6), (vuint8m2_t)(op7), (vuint8m2_t)(op8), (const uint8_t *)(op9), (vuint32m8_t)(op10), (size_t)(op11)) +#define vloxseg4ei32_v_u8mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei32_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (const uint8_t *)(op4), (vuint32m2_t)(op5), (size_t)(op6)) +#define vloxseg4ei32_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei32_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vbool16_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (const uint8_t *)(op9), (vuint32m2_t)(op10), (size_t)(op11)) +#define vloxseg4ei32_v_u8mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei32_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (const uint8_t *)(op4), (vuint32m1_t)(op5), (size_t)(op6)) +#define vloxseg4ei32_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei32_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vbool32_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (const uint8_t *)(op9), (vuint32m1_t)(op10), (size_t)(op11)) +#define vloxseg4ei32_v_u8mf8(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei32_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (const uint8_t *)(op4), (vuint32mf2_t)(op5), (size_t)(op6)) +#define vloxseg4ei32_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei32_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vbool64_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (const uint8_t *)(op9), (vuint32mf2_t)(op10), (size_t)(op11)) +#define vloxseg5ei32_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei32_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (const int8_t *)(op5), (vuint32m4_t)(op6), (size_t)(op7)) +#define vloxseg5ei32_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei32_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vbool8_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (const int8_t *)(op11), (vuint32m4_t)(op12), (size_t)(op13)) +#define vloxseg5ei32_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei32_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (const int8_t *)(op5), (vuint32m2_t)(op6), (size_t)(op7)) +#define vloxseg5ei32_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei32_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vbool16_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (const int8_t *)(op11), (vuint32m2_t)(op12), (size_t)(op13)) +#define vloxseg5ei32_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei32_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (const int8_t *)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vloxseg5ei32_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei32_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vbool32_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (const int8_t *)(op11), (vuint32m1_t)(op12), (size_t)(op13)) +#define vloxseg5ei32_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei32_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (const int8_t *)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vloxseg5ei32_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei32_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vbool64_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (const int8_t *)(op11), (vuint32mf2_t)(op12), (size_t)(op13)) +#define vloxseg5ei32_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei32_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (const uint8_t *)(op5), (vuint32m4_t)(op6), (size_t)(op7)) +#define vloxseg5ei32_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei32_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vbool8_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (const uint8_t *)(op11), (vuint32m4_t)(op12), (size_t)(op13)) +#define vloxseg5ei32_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei32_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (const uint8_t *)(op5), (vuint32m2_t)(op6), (size_t)(op7)) +#define vloxseg5ei32_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei32_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vbool16_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (const uint8_t *)(op11), (vuint32m2_t)(op12), (size_t)(op13)) +#define vloxseg5ei32_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei32_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (const uint8_t *)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vloxseg5ei32_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei32_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vbool32_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (const uint8_t *)(op11), (vuint32m1_t)(op12), (size_t)(op13)) +#define vloxseg5ei32_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei32_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (const uint8_t *)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vloxseg5ei32_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei32_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vbool64_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (const uint8_t *)(op11), (vuint32mf2_t)(op12), (size_t)(op13)) +#define vloxseg6ei32_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei32_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (const int8_t *)(op6), (vuint32m4_t)(op7), (size_t)(op8)) +#define vloxseg6ei32_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei32_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vbool8_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (vint8m1_t)(op11), (vint8m1_t)(op12), (const int8_t *)(op13), (vuint32m4_t)(op14), (size_t)(op15)) +#define vloxseg6ei32_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei32_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (const int8_t *)(op6), (vuint32m2_t)(op7), (size_t)(op8)) +#define vloxseg6ei32_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei32_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vbool16_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (vint8mf2_t)(op11), (vint8mf2_t)(op12), (const int8_t *)(op13), (vuint32m2_t)(op14), (size_t)(op15)) +#define vloxseg6ei32_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei32_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (const int8_t *)(op6), (vuint32m1_t)(op7), (size_t)(op8)) +#define vloxseg6ei32_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei32_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vbool32_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (vint8mf4_t)(op11), (vint8mf4_t)(op12), (const int8_t *)(op13), (vuint32m1_t)(op14), (size_t)(op15)) +#define vloxseg6ei32_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei32_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (const int8_t *)(op6), (vuint32mf2_t)(op7), (size_t)(op8)) +#define vloxseg6ei32_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei32_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vbool64_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (vint8mf8_t)(op11), (vint8mf8_t)(op12), (const int8_t *)(op13), (vuint32mf2_t)(op14), (size_t)(op15)) +#define vloxseg6ei32_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei32_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (const uint8_t *)(op6), (vuint32m4_t)(op7), (size_t)(op8)) +#define vloxseg6ei32_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei32_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vbool8_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (vuint8m1_t)(op11), (vuint8m1_t)(op12), (const uint8_t *)(op13), (vuint32m4_t)(op14), (size_t)(op15)) +#define vloxseg6ei32_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei32_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (const uint8_t *)(op6), (vuint32m2_t)(op7), (size_t)(op8)) +#define vloxseg6ei32_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei32_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vbool16_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (vuint8mf2_t)(op11), (vuint8mf2_t)(op12), (const uint8_t *)(op13), (vuint32m2_t)(op14), (size_t)(op15)) +#define vloxseg6ei32_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei32_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (const uint8_t *)(op6), (vuint32m1_t)(op7), (size_t)(op8)) +#define vloxseg6ei32_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei32_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vbool32_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (vuint8mf4_t)(op11), (vuint8mf4_t)(op12), (const uint8_t *)(op13), (vuint32m1_t)(op14), (size_t)(op15)) +#define vloxseg6ei32_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei32_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (const uint8_t *)(op6), (vuint32mf2_t)(op7), (size_t)(op8)) +#define vloxseg6ei32_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei32_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vbool64_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (vuint8mf8_t)(op11), (vuint8mf8_t)(op12), (const uint8_t *)(op13), (vuint32mf2_t)(op14), (size_t)(op15)) +#define vloxseg7ei32_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei32_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vint8m1_t *)(op6), (const int8_t *)(op7), (vuint32m4_t)(op8), (size_t)(op9)) +#define vloxseg7ei32_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei32_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vint8m1_t *)(op6), (vbool8_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (vint8m1_t)(op11), (vint8m1_t)(op12), (vint8m1_t)(op13), (vint8m1_t)(op14), (const int8_t *)(op15), (vuint32m4_t)(op16), (size_t)(op17)) +#define vloxseg7ei32_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei32_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vint8mf2_t *)(op6), (const int8_t *)(op7), (vuint32m2_t)(op8), (size_t)(op9)) +#define vloxseg7ei32_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei32_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vint8mf2_t *)(op6), (vbool16_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (vint8mf2_t)(op11), (vint8mf2_t)(op12), (vint8mf2_t)(op13), (vint8mf2_t)(op14), (const int8_t *)(op15), (vuint32m2_t)(op16), (size_t)(op17)) +#define vloxseg7ei32_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei32_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vint8mf4_t *)(op6), (const int8_t *)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vloxseg7ei32_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei32_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vint8mf4_t *)(op6), (vbool32_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (vint8mf4_t)(op11), (vint8mf4_t)(op12), (vint8mf4_t)(op13), (vint8mf4_t)(op14), (const int8_t *)(op15), (vuint32m1_t)(op16), (size_t)(op17)) +#define vloxseg7ei32_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei32_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vint8mf8_t *)(op6), (const int8_t *)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vloxseg7ei32_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei32_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vint8mf8_t *)(op6), (vbool64_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (vint8mf8_t)(op11), (vint8mf8_t)(op12), (vint8mf8_t)(op13), (vint8mf8_t)(op14), (const int8_t *)(op15), (vuint32mf2_t)(op16), (size_t)(op17)) +#define vloxseg7ei32_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei32_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vuint8m1_t *)(op6), (const uint8_t *)(op7), (vuint32m4_t)(op8), (size_t)(op9)) +#define vloxseg7ei32_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei32_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vuint8m1_t *)(op6), (vbool8_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (vuint8m1_t)(op11), (vuint8m1_t)(op12), (vuint8m1_t)(op13), (vuint8m1_t)(op14), (const uint8_t *)(op15), (vuint32m4_t)(op16), (size_t)(op17)) +#define vloxseg7ei32_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei32_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vuint8mf2_t *)(op6), (const uint8_t *)(op7), (vuint32m2_t)(op8), (size_t)(op9)) +#define vloxseg7ei32_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei32_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vuint8mf2_t *)(op6), (vbool16_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (vuint8mf2_t)(op11), (vuint8mf2_t)(op12), (vuint8mf2_t)(op13), (vuint8mf2_t)(op14), (const uint8_t *)(op15), (vuint32m2_t)(op16), (size_t)(op17)) +#define vloxseg7ei32_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei32_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vuint8mf4_t *)(op6), (const uint8_t *)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vloxseg7ei32_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei32_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vuint8mf4_t *)(op6), (vbool32_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (vuint8mf4_t)(op11), (vuint8mf4_t)(op12), (vuint8mf4_t)(op13), (vuint8mf4_t)(op14), (const uint8_t *)(op15), (vuint32m1_t)(op16), (size_t)(op17)) +#define vloxseg7ei32_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei32_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vuint8mf8_t *)(op6), (const uint8_t *)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vloxseg7ei32_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei32_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vuint8mf8_t *)(op6), (vbool64_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (vuint8mf8_t)(op11), (vuint8mf8_t)(op12), (vuint8mf8_t)(op13), (vuint8mf8_t)(op14), (const uint8_t *)(op15), (vuint32mf2_t)(op16), (size_t)(op17)) +#define vloxseg8ei32_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei32_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vint8m1_t *)(op6), (vint8m1_t *)(op7), (const int8_t *)(op8), (vuint32m4_t)(op9), (size_t)(op10)) +#define vloxseg8ei32_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei32_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vint8m1_t *)(op6), (vint8m1_t *)(op7), (vbool8_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (vint8m1_t)(op11), (vint8m1_t)(op12), (vint8m1_t)(op13), (vint8m1_t)(op14), (vint8m1_t)(op15), (vint8m1_t)(op16), (const int8_t *)(op17), (vuint32m4_t)(op18), (size_t)(op19)) +#define vloxseg8ei32_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei32_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vint8mf2_t *)(op6), (vint8mf2_t *)(op7), (const int8_t *)(op8), (vuint32m2_t)(op9), (size_t)(op10)) +#define vloxseg8ei32_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei32_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vint8mf2_t *)(op6), (vint8mf2_t *)(op7), (vbool16_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (vint8mf2_t)(op11), (vint8mf2_t)(op12), (vint8mf2_t)(op13), (vint8mf2_t)(op14), (vint8mf2_t)(op15), (vint8mf2_t)(op16), (const int8_t *)(op17), (vuint32m2_t)(op18), (size_t)(op19)) +#define vloxseg8ei32_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei32_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vint8mf4_t *)(op6), (vint8mf4_t *)(op7), (const int8_t *)(op8), (vuint32m1_t)(op9), (size_t)(op10)) +#define vloxseg8ei32_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei32_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vint8mf4_t *)(op6), (vint8mf4_t *)(op7), (vbool32_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (vint8mf4_t)(op11), (vint8mf4_t)(op12), (vint8mf4_t)(op13), (vint8mf4_t)(op14), (vint8mf4_t)(op15), (vint8mf4_t)(op16), (const int8_t *)(op17), (vuint32m1_t)(op18), (size_t)(op19)) +#define vloxseg8ei32_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei32_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vint8mf8_t *)(op6), (vint8mf8_t *)(op7), (const int8_t *)(op8), (vuint32mf2_t)(op9), (size_t)(op10)) +#define vloxseg8ei32_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei32_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vint8mf8_t *)(op6), (vint8mf8_t *)(op7), (vbool64_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (vint8mf8_t)(op11), (vint8mf8_t)(op12), (vint8mf8_t)(op13), (vint8mf8_t)(op14), (vint8mf8_t)(op15), (vint8mf8_t)(op16), (const int8_t *)(op17), (vuint32mf2_t)(op18), (size_t)(op19)) +#define vloxseg8ei32_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei32_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vuint8m1_t *)(op6), (vuint8m1_t *)(op7), (const uint8_t *)(op8), (vuint32m4_t)(op9), (size_t)(op10)) +#define vloxseg8ei32_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei32_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vuint8m1_t *)(op6), (vuint8m1_t *)(op7), (vbool8_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (vuint8m1_t)(op11), (vuint8m1_t)(op12), (vuint8m1_t)(op13), (vuint8m1_t)(op14), (vuint8m1_t)(op15), (vuint8m1_t)(op16), (const uint8_t *)(op17), (vuint32m4_t)(op18), (size_t)(op19)) +#define vloxseg8ei32_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei32_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vuint8mf2_t *)(op6), (vuint8mf2_t *)(op7), (const uint8_t *)(op8), (vuint32m2_t)(op9), (size_t)(op10)) +#define vloxseg8ei32_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei32_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vuint8mf2_t *)(op6), (vuint8mf2_t *)(op7), (vbool16_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (vuint8mf2_t)(op11), (vuint8mf2_t)(op12), (vuint8mf2_t)(op13), (vuint8mf2_t)(op14), (vuint8mf2_t)(op15), (vuint8mf2_t)(op16), (const uint8_t *)(op17), (vuint32m2_t)(op18), (size_t)(op19)) +#define vloxseg8ei32_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei32_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vuint8mf4_t *)(op6), (vuint8mf4_t *)(op7), (const uint8_t *)(op8), (vuint32m1_t)(op9), (size_t)(op10)) +#define vloxseg8ei32_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei32_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vuint8mf4_t *)(op6), (vuint8mf4_t *)(op7), (vbool32_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (vuint8mf4_t)(op11), (vuint8mf4_t)(op12), (vuint8mf4_t)(op13), (vuint8mf4_t)(op14), (vuint8mf4_t)(op15), (vuint8mf4_t)(op16), (const uint8_t *)(op17), (vuint32m1_t)(op18), (size_t)(op19)) +#define vloxseg8ei32_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei32_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vuint8mf8_t *)(op6), (vuint8mf8_t *)(op7), (const uint8_t *)(op8), (vuint32mf2_t)(op9), (size_t)(op10)) +#define vloxseg8ei32_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei32_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vuint8mf8_t *)(op6), (vuint8mf8_t *)(op7), (vbool64_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (vuint8mf8_t)(op11), (vuint8mf8_t)(op12), (vuint8mf8_t)(op13), (vuint8mf8_t)(op14), (vuint8mf8_t)(op15), (vuint8mf8_t)(op16), (const uint8_t *)(op17), (vuint32mf2_t)(op18), (size_t)(op19)) +#define vloxseg2ei64_v_i8m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei64_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (const int8_t *)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vloxseg2ei64_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei64_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vbool8_t)(op2), (vint8m1_t)(op3), (vint8m1_t)(op4), (const int8_t *)(op5), (vuint64m8_t)(op6), (size_t)(op7)) +#define vloxseg2ei64_v_i8mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei64_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (const int8_t *)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vloxseg2ei64_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei64_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vbool16_t)(op2), (vint8mf2_t)(op3), (vint8mf2_t)(op4), (const int8_t *)(op5), (vuint64m4_t)(op6), (size_t)(op7)) +#define vloxseg2ei64_v_i8mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei64_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (const int8_t *)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vloxseg2ei64_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei64_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vbool32_t)(op2), (vint8mf4_t)(op3), (vint8mf4_t)(op4), (const int8_t *)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vloxseg2ei64_v_i8mf8(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei64_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (const int8_t *)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vloxseg2ei64_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei64_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vbool64_t)(op2), (vint8mf8_t)(op3), (vint8mf8_t)(op4), (const int8_t *)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vloxseg2ei64_v_u8m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei64_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (const uint8_t *)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vloxseg2ei64_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei64_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vbool8_t)(op2), (vuint8m1_t)(op3), (vuint8m1_t)(op4), (const uint8_t *)(op5), (vuint64m8_t)(op6), (size_t)(op7)) +#define vloxseg2ei64_v_u8mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei64_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (const uint8_t *)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vloxseg2ei64_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei64_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vbool16_t)(op2), (vuint8mf2_t)(op3), (vuint8mf2_t)(op4), (const uint8_t *)(op5), (vuint64m4_t)(op6), (size_t)(op7)) +#define vloxseg2ei64_v_u8mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei64_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (const uint8_t *)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vloxseg2ei64_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei64_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vbool32_t)(op2), (vuint8mf4_t)(op3), (vuint8mf4_t)(op4), (const uint8_t *)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vloxseg2ei64_v_u8mf8(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei64_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (const uint8_t *)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vloxseg2ei64_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei64_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vbool64_t)(op2), (vuint8mf8_t)(op3), (vuint8mf8_t)(op4), (const uint8_t *)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vloxseg3ei64_v_i8m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei64_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (const int8_t *)(op3), (vuint64m8_t)(op4), (size_t)(op5)) +#define vloxseg3ei64_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei64_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vbool8_t)(op3), (vint8m1_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (const int8_t *)(op7), (vuint64m8_t)(op8), (size_t)(op9)) +#define vloxseg3ei64_v_i8mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei64_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (const int8_t *)(op3), (vuint64m4_t)(op4), (size_t)(op5)) +#define vloxseg3ei64_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei64_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vbool16_t)(op3), (vint8mf2_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (const int8_t *)(op7), (vuint64m4_t)(op8), (size_t)(op9)) +#define vloxseg3ei64_v_i8mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei64_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (const int8_t *)(op3), (vuint64m2_t)(op4), (size_t)(op5)) +#define vloxseg3ei64_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei64_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vbool32_t)(op3), (vint8mf4_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (const int8_t *)(op7), (vuint64m2_t)(op8), (size_t)(op9)) +#define vloxseg3ei64_v_i8mf8(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei64_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (const int8_t *)(op3), (vuint64m1_t)(op4), (size_t)(op5)) +#define vloxseg3ei64_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei64_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vbool64_t)(op3), (vint8mf8_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (const int8_t *)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vloxseg3ei64_v_u8m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei64_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (const uint8_t *)(op3), (vuint64m8_t)(op4), (size_t)(op5)) +#define vloxseg3ei64_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei64_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vbool8_t)(op3), (vuint8m1_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (const uint8_t *)(op7), (vuint64m8_t)(op8), (size_t)(op9)) +#define vloxseg3ei64_v_u8mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei64_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (const uint8_t *)(op3), (vuint64m4_t)(op4), (size_t)(op5)) +#define vloxseg3ei64_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei64_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vbool16_t)(op3), (vuint8mf2_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (const uint8_t *)(op7), (vuint64m4_t)(op8), (size_t)(op9)) +#define vloxseg3ei64_v_u8mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei64_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (const uint8_t *)(op3), (vuint64m2_t)(op4), (size_t)(op5)) +#define vloxseg3ei64_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei64_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vbool32_t)(op3), (vuint8mf4_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (const uint8_t *)(op7), (vuint64m2_t)(op8), (size_t)(op9)) +#define vloxseg3ei64_v_u8mf8(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei64_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (const uint8_t *)(op3), (vuint64m1_t)(op4), (size_t)(op5)) +#define vloxseg3ei64_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei64_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vbool64_t)(op3), (vuint8mf8_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (const uint8_t *)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vloxseg4ei64_v_i8m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei64_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (const int8_t *)(op4), (vuint64m8_t)(op5), (size_t)(op6)) +#define vloxseg4ei64_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei64_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vbool8_t)(op4), (vint8m1_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (const int8_t *)(op9), (vuint64m8_t)(op10), (size_t)(op11)) +#define vloxseg4ei64_v_i8mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei64_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (const int8_t *)(op4), (vuint64m4_t)(op5), (size_t)(op6)) +#define vloxseg4ei64_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei64_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vbool16_t)(op4), (vint8mf2_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (const int8_t *)(op9), (vuint64m4_t)(op10), (size_t)(op11)) +#define vloxseg4ei64_v_i8mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei64_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (const int8_t *)(op4), (vuint64m2_t)(op5), (size_t)(op6)) +#define vloxseg4ei64_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei64_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vbool32_t)(op4), (vint8mf4_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (const int8_t *)(op9), (vuint64m2_t)(op10), (size_t)(op11)) +#define vloxseg4ei64_v_i8mf8(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei64_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (const int8_t *)(op4), (vuint64m1_t)(op5), (size_t)(op6)) +#define vloxseg4ei64_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei64_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vbool64_t)(op4), (vint8mf8_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (const int8_t *)(op9), (vuint64m1_t)(op10), (size_t)(op11)) +#define vloxseg4ei64_v_u8m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei64_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (const uint8_t *)(op4), (vuint64m8_t)(op5), (size_t)(op6)) +#define vloxseg4ei64_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei64_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vbool8_t)(op4), (vuint8m1_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (const uint8_t *)(op9), (vuint64m8_t)(op10), (size_t)(op11)) +#define vloxseg4ei64_v_u8mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei64_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (const uint8_t *)(op4), (vuint64m4_t)(op5), (size_t)(op6)) +#define vloxseg4ei64_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei64_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vbool16_t)(op4), (vuint8mf2_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (const uint8_t *)(op9), (vuint64m4_t)(op10), (size_t)(op11)) +#define vloxseg4ei64_v_u8mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei64_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (const uint8_t *)(op4), (vuint64m2_t)(op5), (size_t)(op6)) +#define vloxseg4ei64_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei64_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vbool32_t)(op4), (vuint8mf4_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (const uint8_t *)(op9), (vuint64m2_t)(op10), (size_t)(op11)) +#define vloxseg4ei64_v_u8mf8(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei64_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (const uint8_t *)(op4), (vuint64m1_t)(op5), (size_t)(op6)) +#define vloxseg4ei64_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei64_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vbool64_t)(op4), (vuint8mf8_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (const uint8_t *)(op9), (vuint64m1_t)(op10), (size_t)(op11)) +#define vloxseg5ei64_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei64_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (const int8_t *)(op5), (vuint64m8_t)(op6), (size_t)(op7)) +#define vloxseg5ei64_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei64_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vbool8_t)(op5), (vint8m1_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (const int8_t *)(op11), (vuint64m8_t)(op12), (size_t)(op13)) +#define vloxseg5ei64_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei64_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (const int8_t *)(op5), (vuint64m4_t)(op6), (size_t)(op7)) +#define vloxseg5ei64_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei64_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vbool16_t)(op5), (vint8mf2_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (const int8_t *)(op11), (vuint64m4_t)(op12), (size_t)(op13)) +#define vloxseg5ei64_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei64_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (const int8_t *)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vloxseg5ei64_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei64_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vbool32_t)(op5), (vint8mf4_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (const int8_t *)(op11), (vuint64m2_t)(op12), (size_t)(op13)) +#define vloxseg5ei64_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei64_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (const int8_t *)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vloxseg5ei64_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei64_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vbool64_t)(op5), (vint8mf8_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (const int8_t *)(op11), (vuint64m1_t)(op12), (size_t)(op13)) +#define vloxseg5ei64_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei64_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (const uint8_t *)(op5), (vuint64m8_t)(op6), (size_t)(op7)) +#define vloxseg5ei64_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei64_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vbool8_t)(op5), (vuint8m1_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (const uint8_t *)(op11), (vuint64m8_t)(op12), (size_t)(op13)) +#define vloxseg5ei64_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei64_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (const uint8_t *)(op5), (vuint64m4_t)(op6), (size_t)(op7)) +#define vloxseg5ei64_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei64_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vbool16_t)(op5), (vuint8mf2_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (const uint8_t *)(op11), (vuint64m4_t)(op12), (size_t)(op13)) +#define vloxseg5ei64_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei64_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (const uint8_t *)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vloxseg5ei64_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei64_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vbool32_t)(op5), (vuint8mf4_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (const uint8_t *)(op11), (vuint64m2_t)(op12), (size_t)(op13)) +#define vloxseg5ei64_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei64_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (const uint8_t *)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vloxseg5ei64_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei64_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vbool64_t)(op5), (vuint8mf8_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (const uint8_t *)(op11), (vuint64m1_t)(op12), (size_t)(op13)) +#define vloxseg6ei64_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei64_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (const int8_t *)(op6), (vuint64m8_t)(op7), (size_t)(op8)) +#define vloxseg6ei64_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei64_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vbool8_t)(op6), (vint8m1_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (vint8m1_t)(op11), (vint8m1_t)(op12), (const int8_t *)(op13), (vuint64m8_t)(op14), (size_t)(op15)) +#define vloxseg6ei64_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei64_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (const int8_t *)(op6), (vuint64m4_t)(op7), (size_t)(op8)) +#define vloxseg6ei64_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei64_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vbool16_t)(op6), (vint8mf2_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (vint8mf2_t)(op11), (vint8mf2_t)(op12), (const int8_t *)(op13), (vuint64m4_t)(op14), (size_t)(op15)) +#define vloxseg6ei64_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei64_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (const int8_t *)(op6), (vuint64m2_t)(op7), (size_t)(op8)) +#define vloxseg6ei64_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei64_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vbool32_t)(op6), (vint8mf4_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (vint8mf4_t)(op11), (vint8mf4_t)(op12), (const int8_t *)(op13), (vuint64m2_t)(op14), (size_t)(op15)) +#define vloxseg6ei64_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei64_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (const int8_t *)(op6), (vuint64m1_t)(op7), (size_t)(op8)) +#define vloxseg6ei64_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei64_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vbool64_t)(op6), (vint8mf8_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (vint8mf8_t)(op11), (vint8mf8_t)(op12), (const int8_t *)(op13), (vuint64m1_t)(op14), (size_t)(op15)) +#define vloxseg6ei64_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei64_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (const uint8_t *)(op6), (vuint64m8_t)(op7), (size_t)(op8)) +#define vloxseg6ei64_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei64_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vbool8_t)(op6), (vuint8m1_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (vuint8m1_t)(op11), (vuint8m1_t)(op12), (const uint8_t *)(op13), (vuint64m8_t)(op14), (size_t)(op15)) +#define vloxseg6ei64_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei64_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (const uint8_t *)(op6), (vuint64m4_t)(op7), (size_t)(op8)) +#define vloxseg6ei64_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei64_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vbool16_t)(op6), (vuint8mf2_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (vuint8mf2_t)(op11), (vuint8mf2_t)(op12), (const uint8_t *)(op13), (vuint64m4_t)(op14), (size_t)(op15)) +#define vloxseg6ei64_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei64_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (const uint8_t *)(op6), (vuint64m2_t)(op7), (size_t)(op8)) +#define vloxseg6ei64_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei64_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vbool32_t)(op6), (vuint8mf4_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (vuint8mf4_t)(op11), (vuint8mf4_t)(op12), (const uint8_t *)(op13), (vuint64m2_t)(op14), (size_t)(op15)) +#define vloxseg6ei64_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei64_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (const uint8_t *)(op6), (vuint64m1_t)(op7), (size_t)(op8)) +#define vloxseg6ei64_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei64_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vbool64_t)(op6), (vuint8mf8_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (vuint8mf8_t)(op11), (vuint8mf8_t)(op12), (const uint8_t *)(op13), (vuint64m1_t)(op14), (size_t)(op15)) +#define vloxseg7ei64_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei64_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vint8m1_t *)(op6), (const int8_t *)(op7), (vuint64m8_t)(op8), (size_t)(op9)) +#define vloxseg7ei64_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei64_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vint8m1_t *)(op6), (vbool8_t)(op7), (vint8m1_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (vint8m1_t)(op11), (vint8m1_t)(op12), (vint8m1_t)(op13), (vint8m1_t)(op14), (const int8_t *)(op15), (vuint64m8_t)(op16), (size_t)(op17)) +#define vloxseg7ei64_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei64_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vint8mf2_t *)(op6), (const int8_t *)(op7), (vuint64m4_t)(op8), (size_t)(op9)) +#define vloxseg7ei64_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei64_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vint8mf2_t *)(op6), (vbool16_t)(op7), (vint8mf2_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (vint8mf2_t)(op11), (vint8mf2_t)(op12), (vint8mf2_t)(op13), (vint8mf2_t)(op14), (const int8_t *)(op15), (vuint64m4_t)(op16), (size_t)(op17)) +#define vloxseg7ei64_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei64_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vint8mf4_t *)(op6), (const int8_t *)(op7), (vuint64m2_t)(op8), (size_t)(op9)) +#define vloxseg7ei64_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei64_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vint8mf4_t *)(op6), (vbool32_t)(op7), (vint8mf4_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (vint8mf4_t)(op11), (vint8mf4_t)(op12), (vint8mf4_t)(op13), (vint8mf4_t)(op14), (const int8_t *)(op15), (vuint64m2_t)(op16), (size_t)(op17)) +#define vloxseg7ei64_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei64_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vint8mf8_t *)(op6), (const int8_t *)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vloxseg7ei64_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei64_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vint8mf8_t *)(op6), (vbool64_t)(op7), (vint8mf8_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (vint8mf8_t)(op11), (vint8mf8_t)(op12), (vint8mf8_t)(op13), (vint8mf8_t)(op14), (const int8_t *)(op15), (vuint64m1_t)(op16), (size_t)(op17)) +#define vloxseg7ei64_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei64_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vuint8m1_t *)(op6), (const uint8_t *)(op7), (vuint64m8_t)(op8), (size_t)(op9)) +#define vloxseg7ei64_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei64_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vuint8m1_t *)(op6), (vbool8_t)(op7), (vuint8m1_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (vuint8m1_t)(op11), (vuint8m1_t)(op12), (vuint8m1_t)(op13), (vuint8m1_t)(op14), (const uint8_t *)(op15), (vuint64m8_t)(op16), (size_t)(op17)) +#define vloxseg7ei64_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei64_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vuint8mf2_t *)(op6), (const uint8_t *)(op7), (vuint64m4_t)(op8), (size_t)(op9)) +#define vloxseg7ei64_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei64_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vuint8mf2_t *)(op6), (vbool16_t)(op7), (vuint8mf2_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (vuint8mf2_t)(op11), (vuint8mf2_t)(op12), (vuint8mf2_t)(op13), (vuint8mf2_t)(op14), (const uint8_t *)(op15), (vuint64m4_t)(op16), (size_t)(op17)) +#define vloxseg7ei64_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei64_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vuint8mf4_t *)(op6), (const uint8_t *)(op7), (vuint64m2_t)(op8), (size_t)(op9)) +#define vloxseg7ei64_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei64_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vuint8mf4_t *)(op6), (vbool32_t)(op7), (vuint8mf4_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (vuint8mf4_t)(op11), (vuint8mf4_t)(op12), (vuint8mf4_t)(op13), (vuint8mf4_t)(op14), (const uint8_t *)(op15), (vuint64m2_t)(op16), (size_t)(op17)) +#define vloxseg7ei64_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei64_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vuint8mf8_t *)(op6), (const uint8_t *)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vloxseg7ei64_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei64_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vuint8mf8_t *)(op6), (vbool64_t)(op7), (vuint8mf8_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (vuint8mf8_t)(op11), (vuint8mf8_t)(op12), (vuint8mf8_t)(op13), (vuint8mf8_t)(op14), (const uint8_t *)(op15), (vuint64m1_t)(op16), (size_t)(op17)) +#define vloxseg8ei64_v_i8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei64_v_i8m1((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vint8m1_t *)(op6), (vint8m1_t *)(op7), (const int8_t *)(op8), (vuint64m8_t)(op9), (size_t)(op10)) +#define vloxseg8ei64_v_i8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei64_v_i8m1_m((vint8m1_t *)(op0), (vint8m1_t *)(op1), (vint8m1_t *)(op2), (vint8m1_t *)(op3), (vint8m1_t *)(op4), (vint8m1_t *)(op5), (vint8m1_t *)(op6), (vint8m1_t *)(op7), (vbool8_t)(op8), (vint8m1_t)(op9), (vint8m1_t)(op10), (vint8m1_t)(op11), (vint8m1_t)(op12), (vint8m1_t)(op13), (vint8m1_t)(op14), (vint8m1_t)(op15), (vint8m1_t)(op16), (const int8_t *)(op17), (vuint64m8_t)(op18), (size_t)(op19)) +#define vloxseg8ei64_v_i8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei64_v_i8mf2((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vint8mf2_t *)(op6), (vint8mf2_t *)(op7), (const int8_t *)(op8), (vuint64m4_t)(op9), (size_t)(op10)) +#define vloxseg8ei64_v_i8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei64_v_i8mf2_m((vint8mf2_t *)(op0), (vint8mf2_t *)(op1), (vint8mf2_t *)(op2), (vint8mf2_t *)(op3), (vint8mf2_t *)(op4), (vint8mf2_t *)(op5), (vint8mf2_t *)(op6), (vint8mf2_t *)(op7), (vbool16_t)(op8), (vint8mf2_t)(op9), (vint8mf2_t)(op10), (vint8mf2_t)(op11), (vint8mf2_t)(op12), (vint8mf2_t)(op13), (vint8mf2_t)(op14), (vint8mf2_t)(op15), (vint8mf2_t)(op16), (const int8_t *)(op17), (vuint64m4_t)(op18), (size_t)(op19)) +#define vloxseg8ei64_v_i8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei64_v_i8mf4((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vint8mf4_t *)(op6), (vint8mf4_t *)(op7), (const int8_t *)(op8), (vuint64m2_t)(op9), (size_t)(op10)) +#define vloxseg8ei64_v_i8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei64_v_i8mf4_m((vint8mf4_t *)(op0), (vint8mf4_t *)(op1), (vint8mf4_t *)(op2), (vint8mf4_t *)(op3), (vint8mf4_t *)(op4), (vint8mf4_t *)(op5), (vint8mf4_t *)(op6), (vint8mf4_t *)(op7), (vbool32_t)(op8), (vint8mf4_t)(op9), (vint8mf4_t)(op10), (vint8mf4_t)(op11), (vint8mf4_t)(op12), (vint8mf4_t)(op13), (vint8mf4_t)(op14), (vint8mf4_t)(op15), (vint8mf4_t)(op16), (const int8_t *)(op17), (vuint64m2_t)(op18), (size_t)(op19)) +#define vloxseg8ei64_v_i8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei64_v_i8mf8((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vint8mf8_t *)(op6), (vint8mf8_t *)(op7), (const int8_t *)(op8), (vuint64m1_t)(op9), (size_t)(op10)) +#define vloxseg8ei64_v_i8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei64_v_i8mf8_m((vint8mf8_t *)(op0), (vint8mf8_t *)(op1), (vint8mf8_t *)(op2), (vint8mf8_t *)(op3), (vint8mf8_t *)(op4), (vint8mf8_t *)(op5), (vint8mf8_t *)(op6), (vint8mf8_t *)(op7), (vbool64_t)(op8), (vint8mf8_t)(op9), (vint8mf8_t)(op10), (vint8mf8_t)(op11), (vint8mf8_t)(op12), (vint8mf8_t)(op13), (vint8mf8_t)(op14), (vint8mf8_t)(op15), (vint8mf8_t)(op16), (const int8_t *)(op17), (vuint64m1_t)(op18), (size_t)(op19)) +#define vloxseg8ei64_v_u8m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei64_v_u8m1((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vuint8m1_t *)(op6), (vuint8m1_t *)(op7), (const uint8_t *)(op8), (vuint64m8_t)(op9), (size_t)(op10)) +#define vloxseg8ei64_v_u8m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei64_v_u8m1_m((vuint8m1_t *)(op0), (vuint8m1_t *)(op1), (vuint8m1_t *)(op2), (vuint8m1_t *)(op3), (vuint8m1_t *)(op4), (vuint8m1_t *)(op5), (vuint8m1_t *)(op6), (vuint8m1_t *)(op7), (vbool8_t)(op8), (vuint8m1_t)(op9), (vuint8m1_t)(op10), (vuint8m1_t)(op11), (vuint8m1_t)(op12), (vuint8m1_t)(op13), (vuint8m1_t)(op14), (vuint8m1_t)(op15), (vuint8m1_t)(op16), (const uint8_t *)(op17), (vuint64m8_t)(op18), (size_t)(op19)) +#define vloxseg8ei64_v_u8mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei64_v_u8mf2((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vuint8mf2_t *)(op6), (vuint8mf2_t *)(op7), (const uint8_t *)(op8), (vuint64m4_t)(op9), (size_t)(op10)) +#define vloxseg8ei64_v_u8mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei64_v_u8mf2_m((vuint8mf2_t *)(op0), (vuint8mf2_t *)(op1), (vuint8mf2_t *)(op2), (vuint8mf2_t *)(op3), (vuint8mf2_t *)(op4), (vuint8mf2_t *)(op5), (vuint8mf2_t *)(op6), (vuint8mf2_t *)(op7), (vbool16_t)(op8), (vuint8mf2_t)(op9), (vuint8mf2_t)(op10), (vuint8mf2_t)(op11), (vuint8mf2_t)(op12), (vuint8mf2_t)(op13), (vuint8mf2_t)(op14), (vuint8mf2_t)(op15), (vuint8mf2_t)(op16), (const uint8_t *)(op17), (vuint64m4_t)(op18), (size_t)(op19)) +#define vloxseg8ei64_v_u8mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei64_v_u8mf4((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vuint8mf4_t *)(op6), (vuint8mf4_t *)(op7), (const uint8_t *)(op8), (vuint64m2_t)(op9), (size_t)(op10)) +#define vloxseg8ei64_v_u8mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei64_v_u8mf4_m((vuint8mf4_t *)(op0), (vuint8mf4_t *)(op1), (vuint8mf4_t *)(op2), (vuint8mf4_t *)(op3), (vuint8mf4_t *)(op4), (vuint8mf4_t *)(op5), (vuint8mf4_t *)(op6), (vuint8mf4_t *)(op7), (vbool32_t)(op8), (vuint8mf4_t)(op9), (vuint8mf4_t)(op10), (vuint8mf4_t)(op11), (vuint8mf4_t)(op12), (vuint8mf4_t)(op13), (vuint8mf4_t)(op14), (vuint8mf4_t)(op15), (vuint8mf4_t)(op16), (const uint8_t *)(op17), (vuint64m2_t)(op18), (size_t)(op19)) +#define vloxseg8ei64_v_u8mf8(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei64_v_u8mf8((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vuint8mf8_t *)(op6), (vuint8mf8_t *)(op7), (const uint8_t *)(op8), (vuint64m1_t)(op9), (size_t)(op10)) +#define vloxseg8ei64_v_u8mf8_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei64_v_u8mf8_m((vuint8mf8_t *)(op0), (vuint8mf8_t *)(op1), (vuint8mf8_t *)(op2), (vuint8mf8_t *)(op3), (vuint8mf8_t *)(op4), (vuint8mf8_t *)(op5), (vuint8mf8_t *)(op6), (vuint8mf8_t *)(op7), (vbool64_t)(op8), (vuint8mf8_t)(op9), (vuint8mf8_t)(op10), (vuint8mf8_t)(op11), (vuint8mf8_t)(op12), (vuint8mf8_t)(op13), (vuint8mf8_t)(op14), (vuint8mf8_t)(op15), (vuint8mf8_t)(op16), (const uint8_t *)(op17), (vuint64m1_t)(op18), (size_t)(op19)) +#define vloxseg2ei8_v_i16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (const int16_t *)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vbool16_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (const int16_t *)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vloxseg2ei8_v_i16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_i16m2((vint16m2_t *)(op0), (vint16m2_t *)(op1), (const int16_t *)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_i16m2_m((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vbool8_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (const int16_t *)(op5), (vuint8m1_t)(op6), (size_t)(op7)) +#define vloxseg2ei8_v_i16m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_i16m4((vint16m4_t *)(op0), (vint16m4_t *)(op1), (const int16_t *)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_i16m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_i16m4_m((vint16m4_t *)(op0), (vint16m4_t *)(op1), (vbool4_t)(op2), (vint16m4_t)(op3), (vint16m4_t)(op4), (const int16_t *)(op5), (vuint8m2_t)(op6), (size_t)(op7)) +#define vloxseg2ei8_v_i16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (const int16_t *)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vbool32_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (const int16_t *)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vloxseg2ei8_v_i16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (const int16_t *)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vbool64_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (const int16_t *)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vloxseg2ei8_v_u16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (const uint16_t *)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vbool16_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (const uint16_t *)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vloxseg2ei8_v_u16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_u16m2((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (const uint16_t *)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_u16m2_m((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vbool8_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (const uint16_t *)(op5), (vuint8m1_t)(op6), (size_t)(op7)) +#define vloxseg2ei8_v_u16m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_u16m4((vuint16m4_t *)(op0), (vuint16m4_t *)(op1), (const uint16_t *)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_u16m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_u16m4_m((vuint16m4_t *)(op0), (vuint16m4_t *)(op1), (vbool4_t)(op2), (vuint16m4_t)(op3), (vuint16m4_t)(op4), (const uint16_t *)(op5), (vuint8m2_t)(op6), (size_t)(op7)) +#define vloxseg2ei8_v_u16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (const uint16_t *)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vbool32_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (const uint16_t *)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vloxseg2ei8_v_u16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (const uint16_t *)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vbool64_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (const uint16_t *)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vloxseg3ei8_v_i16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei8_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (const int16_t *)(op3), (vuint8mf2_t)(op4), (size_t)(op5)) +#define vloxseg3ei8_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei8_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vbool16_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (const int16_t *)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vloxseg3ei8_v_i16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei8_v_i16m2((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vint16m2_t *)(op2), (const int16_t *)(op3), (vuint8m1_t)(op4), (size_t)(op5)) +#define vloxseg3ei8_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei8_v_i16m2_m((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vint16m2_t *)(op2), (vbool8_t)(op3), (vint16m2_t)(op4), (vint16m2_t)(op5), (vint16m2_t)(op6), (const int16_t *)(op7), (vuint8m1_t)(op8), (size_t)(op9)) +#define vloxseg3ei8_v_i16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei8_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (const int16_t *)(op3), (vuint8mf4_t)(op4), (size_t)(op5)) +#define vloxseg3ei8_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei8_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vbool32_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (const int16_t *)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vloxseg3ei8_v_i16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei8_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (const int16_t *)(op3), (vuint8mf8_t)(op4), (size_t)(op5)) +#define vloxseg3ei8_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei8_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vbool64_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (const int16_t *)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vloxseg3ei8_v_u16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei8_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (const uint16_t *)(op3), (vuint8mf2_t)(op4), (size_t)(op5)) +#define vloxseg3ei8_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei8_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vbool16_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (const uint16_t *)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vloxseg3ei8_v_u16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei8_v_u16m2((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vuint16m2_t *)(op2), (const uint16_t *)(op3), (vuint8m1_t)(op4), (size_t)(op5)) +#define vloxseg3ei8_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei8_v_u16m2_m((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vuint16m2_t *)(op2), (vbool8_t)(op3), (vuint16m2_t)(op4), (vuint16m2_t)(op5), (vuint16m2_t)(op6), (const uint16_t *)(op7), (vuint8m1_t)(op8), (size_t)(op9)) +#define vloxseg3ei8_v_u16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei8_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (const uint16_t *)(op3), (vuint8mf4_t)(op4), (size_t)(op5)) +#define vloxseg3ei8_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei8_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vbool32_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (const uint16_t *)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vloxseg3ei8_v_u16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei8_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (const uint16_t *)(op3), (vuint8mf8_t)(op4), (size_t)(op5)) +#define vloxseg3ei8_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei8_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vbool64_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (const uint16_t *)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vloxseg4ei8_v_i16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei8_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (const int16_t *)(op4), (vuint8mf2_t)(op5), (size_t)(op6)) +#define vloxseg4ei8_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei8_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vbool16_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (const int16_t *)(op9), (vuint8mf2_t)(op10), (size_t)(op11)) +#define vloxseg4ei8_v_i16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei8_v_i16m2((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vint16m2_t *)(op2), (vint16m2_t *)(op3), (const int16_t *)(op4), (vuint8m1_t)(op5), (size_t)(op6)) +#define vloxseg4ei8_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei8_v_i16m2_m((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vint16m2_t *)(op2), (vint16m2_t *)(op3), (vbool8_t)(op4), (vint16m2_t)(op5), (vint16m2_t)(op6), (vint16m2_t)(op7), (vint16m2_t)(op8), (const int16_t *)(op9), (vuint8m1_t)(op10), (size_t)(op11)) +#define vloxseg4ei8_v_i16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei8_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (const int16_t *)(op4), (vuint8mf4_t)(op5), (size_t)(op6)) +#define vloxseg4ei8_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei8_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vbool32_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (const int16_t *)(op9), (vuint8mf4_t)(op10), (size_t)(op11)) +#define vloxseg4ei8_v_i16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei8_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (const int16_t *)(op4), (vuint8mf8_t)(op5), (size_t)(op6)) +#define vloxseg4ei8_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei8_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vbool64_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (const int16_t *)(op9), (vuint8mf8_t)(op10), (size_t)(op11)) +#define vloxseg4ei8_v_u16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei8_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (const uint16_t *)(op4), (vuint8mf2_t)(op5), (size_t)(op6)) +#define vloxseg4ei8_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei8_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vbool16_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (const uint16_t *)(op9), (vuint8mf2_t)(op10), (size_t)(op11)) +#define vloxseg4ei8_v_u16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei8_v_u16m2((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vuint16m2_t *)(op2), (vuint16m2_t *)(op3), (const uint16_t *)(op4), (vuint8m1_t)(op5), (size_t)(op6)) +#define vloxseg4ei8_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei8_v_u16m2_m((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vuint16m2_t *)(op2), (vuint16m2_t *)(op3), (vbool8_t)(op4), (vuint16m2_t)(op5), (vuint16m2_t)(op6), (vuint16m2_t)(op7), (vuint16m2_t)(op8), (const uint16_t *)(op9), (vuint8m1_t)(op10), (size_t)(op11)) +#define vloxseg4ei8_v_u16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei8_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (const uint16_t *)(op4), (vuint8mf4_t)(op5), (size_t)(op6)) +#define vloxseg4ei8_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei8_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vbool32_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (const uint16_t *)(op9), (vuint8mf4_t)(op10), (size_t)(op11)) +#define vloxseg4ei8_v_u16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei8_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (const uint16_t *)(op4), (vuint8mf8_t)(op5), (size_t)(op6)) +#define vloxseg4ei8_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei8_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vbool64_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (const uint16_t *)(op9), (vuint8mf8_t)(op10), (size_t)(op11)) +#define vloxseg5ei8_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei8_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (const int16_t *)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vloxseg5ei8_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei8_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vbool16_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (const int16_t *)(op11), (vuint8mf2_t)(op12), (size_t)(op13)) +#define vloxseg5ei8_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei8_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (const int16_t *)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vloxseg5ei8_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei8_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vbool32_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (const int16_t *)(op11), (vuint8mf4_t)(op12), (size_t)(op13)) +#define vloxseg5ei8_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei8_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (const int16_t *)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vloxseg5ei8_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei8_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vbool64_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (const int16_t *)(op11), (vuint8mf8_t)(op12), (size_t)(op13)) +#define vloxseg5ei8_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei8_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (const uint16_t *)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vloxseg5ei8_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei8_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vbool16_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (const uint16_t *)(op11), (vuint8mf2_t)(op12), (size_t)(op13)) +#define vloxseg5ei8_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei8_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (const uint16_t *)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vloxseg5ei8_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei8_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vbool32_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (const uint16_t *)(op11), (vuint8mf4_t)(op12), (size_t)(op13)) +#define vloxseg5ei8_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei8_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (const uint16_t *)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vloxseg5ei8_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei8_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vbool64_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (const uint16_t *)(op11), (vuint8mf8_t)(op12), (size_t)(op13)) +#define vloxseg6ei8_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei8_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (const int16_t *)(op6), (vuint8mf2_t)(op7), (size_t)(op8)) +#define vloxseg6ei8_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei8_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vbool16_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (vint16m1_t)(op11), (vint16m1_t)(op12), (const int16_t *)(op13), (vuint8mf2_t)(op14), (size_t)(op15)) +#define vloxseg6ei8_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei8_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (const int16_t *)(op6), (vuint8mf4_t)(op7), (size_t)(op8)) +#define vloxseg6ei8_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei8_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vbool32_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (vint16mf2_t)(op11), (vint16mf2_t)(op12), (const int16_t *)(op13), (vuint8mf4_t)(op14), (size_t)(op15)) +#define vloxseg6ei8_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei8_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (const int16_t *)(op6), (vuint8mf8_t)(op7), (size_t)(op8)) +#define vloxseg6ei8_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei8_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vbool64_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (vint16mf4_t)(op11), (vint16mf4_t)(op12), (const int16_t *)(op13), (vuint8mf8_t)(op14), (size_t)(op15)) +#define vloxseg6ei8_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei8_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (const uint16_t *)(op6), (vuint8mf2_t)(op7), (size_t)(op8)) +#define vloxseg6ei8_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei8_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vbool16_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (vuint16m1_t)(op11), (vuint16m1_t)(op12), (const uint16_t *)(op13), (vuint8mf2_t)(op14), (size_t)(op15)) +#define vloxseg6ei8_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei8_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (const uint16_t *)(op6), (vuint8mf4_t)(op7), (size_t)(op8)) +#define vloxseg6ei8_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei8_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vbool32_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (vuint16mf2_t)(op11), (vuint16mf2_t)(op12), (const uint16_t *)(op13), (vuint8mf4_t)(op14), (size_t)(op15)) +#define vloxseg6ei8_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei8_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (const uint16_t *)(op6), (vuint8mf8_t)(op7), (size_t)(op8)) +#define vloxseg6ei8_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei8_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vbool64_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (vuint16mf4_t)(op11), (vuint16mf4_t)(op12), (const uint16_t *)(op13), (vuint8mf8_t)(op14), (size_t)(op15)) +#define vloxseg7ei8_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei8_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vint16m1_t *)(op6), (const int16_t *)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vloxseg7ei8_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei8_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vint16m1_t *)(op6), (vbool16_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (vint16m1_t)(op11), (vint16m1_t)(op12), (vint16m1_t)(op13), (vint16m1_t)(op14), (const int16_t *)(op15), (vuint8mf2_t)(op16), (size_t)(op17)) +#define vloxseg7ei8_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei8_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vint16mf2_t *)(op6), (const int16_t *)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vloxseg7ei8_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei8_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vint16mf2_t *)(op6), (vbool32_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (vint16mf2_t)(op11), (vint16mf2_t)(op12), (vint16mf2_t)(op13), (vint16mf2_t)(op14), (const int16_t *)(op15), (vuint8mf4_t)(op16), (size_t)(op17)) +#define vloxseg7ei8_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei8_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vint16mf4_t *)(op6), (const int16_t *)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vloxseg7ei8_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei8_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vint16mf4_t *)(op6), (vbool64_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (vint16mf4_t)(op11), (vint16mf4_t)(op12), (vint16mf4_t)(op13), (vint16mf4_t)(op14), (const int16_t *)(op15), (vuint8mf8_t)(op16), (size_t)(op17)) +#define vloxseg7ei8_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei8_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vuint16m1_t *)(op6), (const uint16_t *)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vloxseg7ei8_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei8_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vuint16m1_t *)(op6), (vbool16_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (vuint16m1_t)(op11), (vuint16m1_t)(op12), (vuint16m1_t)(op13), (vuint16m1_t)(op14), (const uint16_t *)(op15), (vuint8mf2_t)(op16), (size_t)(op17)) +#define vloxseg7ei8_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei8_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vuint16mf2_t *)(op6), (const uint16_t *)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vloxseg7ei8_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei8_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vuint16mf2_t *)(op6), (vbool32_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (vuint16mf2_t)(op11), (vuint16mf2_t)(op12), (vuint16mf2_t)(op13), (vuint16mf2_t)(op14), (const uint16_t *)(op15), (vuint8mf4_t)(op16), (size_t)(op17)) +#define vloxseg7ei8_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei8_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vuint16mf4_t *)(op6), (const uint16_t *)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vloxseg7ei8_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei8_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vuint16mf4_t *)(op6), (vbool64_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (vuint16mf4_t)(op11), (vuint16mf4_t)(op12), (vuint16mf4_t)(op13), (vuint16mf4_t)(op14), (const uint16_t *)(op15), (vuint8mf8_t)(op16), (size_t)(op17)) +#define vloxseg8ei8_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei8_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vint16m1_t *)(op6), (vint16m1_t *)(op7), (const int16_t *)(op8), (vuint8mf2_t)(op9), (size_t)(op10)) +#define vloxseg8ei8_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei8_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vint16m1_t *)(op6), (vint16m1_t *)(op7), (vbool16_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (vint16m1_t)(op11), (vint16m1_t)(op12), (vint16m1_t)(op13), (vint16m1_t)(op14), (vint16m1_t)(op15), (vint16m1_t)(op16), (const int16_t *)(op17), (vuint8mf2_t)(op18), (size_t)(op19)) +#define vloxseg8ei8_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei8_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vint16mf2_t *)(op6), (vint16mf2_t *)(op7), (const int16_t *)(op8), (vuint8mf4_t)(op9), (size_t)(op10)) +#define vloxseg8ei8_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei8_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vint16mf2_t *)(op6), (vint16mf2_t *)(op7), (vbool32_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (vint16mf2_t)(op11), (vint16mf2_t)(op12), (vint16mf2_t)(op13), (vint16mf2_t)(op14), (vint16mf2_t)(op15), (vint16mf2_t)(op16), (const int16_t *)(op17), (vuint8mf4_t)(op18), (size_t)(op19)) +#define vloxseg8ei8_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei8_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vint16mf4_t *)(op6), (vint16mf4_t *)(op7), (const int16_t *)(op8), (vuint8mf8_t)(op9), (size_t)(op10)) +#define vloxseg8ei8_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei8_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vint16mf4_t *)(op6), (vint16mf4_t *)(op7), (vbool64_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (vint16mf4_t)(op11), (vint16mf4_t)(op12), (vint16mf4_t)(op13), (vint16mf4_t)(op14), (vint16mf4_t)(op15), (vint16mf4_t)(op16), (const int16_t *)(op17), (vuint8mf8_t)(op18), (size_t)(op19)) +#define vloxseg8ei8_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei8_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vuint16m1_t *)(op6), (vuint16m1_t *)(op7), (const uint16_t *)(op8), (vuint8mf2_t)(op9), (size_t)(op10)) +#define vloxseg8ei8_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei8_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vuint16m1_t *)(op6), (vuint16m1_t *)(op7), (vbool16_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (vuint16m1_t)(op11), (vuint16m1_t)(op12), (vuint16m1_t)(op13), (vuint16m1_t)(op14), (vuint16m1_t)(op15), (vuint16m1_t)(op16), (const uint16_t *)(op17), (vuint8mf2_t)(op18), (size_t)(op19)) +#define vloxseg8ei8_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei8_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vuint16mf2_t *)(op6), (vuint16mf2_t *)(op7), (const uint16_t *)(op8), (vuint8mf4_t)(op9), (size_t)(op10)) +#define vloxseg8ei8_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei8_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vuint16mf2_t *)(op6), (vuint16mf2_t *)(op7), (vbool32_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (vuint16mf2_t)(op11), (vuint16mf2_t)(op12), (vuint16mf2_t)(op13), (vuint16mf2_t)(op14), (vuint16mf2_t)(op15), (vuint16mf2_t)(op16), (const uint16_t *)(op17), (vuint8mf4_t)(op18), (size_t)(op19)) +#define vloxseg8ei8_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei8_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vuint16mf4_t *)(op6), (vuint16mf4_t *)(op7), (const uint16_t *)(op8), (vuint8mf8_t)(op9), (size_t)(op10)) +#define vloxseg8ei8_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei8_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vuint16mf4_t *)(op6), (vuint16mf4_t *)(op7), (vbool64_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (vuint16mf4_t)(op11), (vuint16mf4_t)(op12), (vuint16mf4_t)(op13), (vuint16mf4_t)(op14), (vuint16mf4_t)(op15), (vuint16mf4_t)(op16), (const uint16_t *)(op17), (vuint8mf8_t)(op18), (size_t)(op19)) +#define vloxseg2ei16_v_i16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (const int16_t *)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vbool16_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (const int16_t *)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vloxseg2ei16_v_i16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_i16m2((vint16m2_t *)(op0), (vint16m2_t *)(op1), (const int16_t *)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_i16m2_m((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vbool8_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (const int16_t *)(op5), (vuint16m2_t)(op6), (size_t)(op7)) +#define vloxseg2ei16_v_i16m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_i16m4((vint16m4_t *)(op0), (vint16m4_t *)(op1), (const int16_t *)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_i16m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_i16m4_m((vint16m4_t *)(op0), (vint16m4_t *)(op1), (vbool4_t)(op2), (vint16m4_t)(op3), (vint16m4_t)(op4), (const int16_t *)(op5), (vuint16m4_t)(op6), (size_t)(op7)) +#define vloxseg2ei16_v_i16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (const int16_t *)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vbool32_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (const int16_t *)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vloxseg2ei16_v_i16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (const int16_t *)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vbool64_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (const int16_t *)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vloxseg2ei16_v_u16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (const uint16_t *)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vbool16_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (const uint16_t *)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vloxseg2ei16_v_u16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_u16m2((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (const uint16_t *)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_u16m2_m((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vbool8_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (const uint16_t *)(op5), (vuint16m2_t)(op6), (size_t)(op7)) +#define vloxseg2ei16_v_u16m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_u16m4((vuint16m4_t *)(op0), (vuint16m4_t *)(op1), (const uint16_t *)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_u16m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_u16m4_m((vuint16m4_t *)(op0), (vuint16m4_t *)(op1), (vbool4_t)(op2), (vuint16m4_t)(op3), (vuint16m4_t)(op4), (const uint16_t *)(op5), (vuint16m4_t)(op6), (size_t)(op7)) +#define vloxseg2ei16_v_u16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (const uint16_t *)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vbool32_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (const uint16_t *)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vloxseg2ei16_v_u16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (const uint16_t *)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vbool64_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (const uint16_t *)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vloxseg3ei16_v_i16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei16_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (const int16_t *)(op3), (vuint16m1_t)(op4), (size_t)(op5)) +#define vloxseg3ei16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei16_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vbool16_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (const int16_t *)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vloxseg3ei16_v_i16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei16_v_i16m2((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vint16m2_t *)(op2), (const int16_t *)(op3), (vuint16m2_t)(op4), (size_t)(op5)) +#define vloxseg3ei16_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei16_v_i16m2_m((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vint16m2_t *)(op2), (vbool8_t)(op3), (vint16m2_t)(op4), (vint16m2_t)(op5), (vint16m2_t)(op6), (const int16_t *)(op7), (vuint16m2_t)(op8), (size_t)(op9)) +#define vloxseg3ei16_v_i16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei16_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (const int16_t *)(op3), (vuint16mf2_t)(op4), (size_t)(op5)) +#define vloxseg3ei16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei16_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vbool32_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (const int16_t *)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vloxseg3ei16_v_i16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei16_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (const int16_t *)(op3), (vuint16mf4_t)(op4), (size_t)(op5)) +#define vloxseg3ei16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei16_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vbool64_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (const int16_t *)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vloxseg3ei16_v_u16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei16_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (const uint16_t *)(op3), (vuint16m1_t)(op4), (size_t)(op5)) +#define vloxseg3ei16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei16_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vbool16_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (const uint16_t *)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vloxseg3ei16_v_u16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei16_v_u16m2((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vuint16m2_t *)(op2), (const uint16_t *)(op3), (vuint16m2_t)(op4), (size_t)(op5)) +#define vloxseg3ei16_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei16_v_u16m2_m((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vuint16m2_t *)(op2), (vbool8_t)(op3), (vuint16m2_t)(op4), (vuint16m2_t)(op5), (vuint16m2_t)(op6), (const uint16_t *)(op7), (vuint16m2_t)(op8), (size_t)(op9)) +#define vloxseg3ei16_v_u16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei16_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (const uint16_t *)(op3), (vuint16mf2_t)(op4), (size_t)(op5)) +#define vloxseg3ei16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei16_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vbool32_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (const uint16_t *)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vloxseg3ei16_v_u16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei16_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (const uint16_t *)(op3), (vuint16mf4_t)(op4), (size_t)(op5)) +#define vloxseg3ei16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei16_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vbool64_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (const uint16_t *)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vloxseg4ei16_v_i16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei16_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (const int16_t *)(op4), (vuint16m1_t)(op5), (size_t)(op6)) +#define vloxseg4ei16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei16_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vbool16_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (const int16_t *)(op9), (vuint16m1_t)(op10), (size_t)(op11)) +#define vloxseg4ei16_v_i16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei16_v_i16m2((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vint16m2_t *)(op2), (vint16m2_t *)(op3), (const int16_t *)(op4), (vuint16m2_t)(op5), (size_t)(op6)) +#define vloxseg4ei16_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei16_v_i16m2_m((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vint16m2_t *)(op2), (vint16m2_t *)(op3), (vbool8_t)(op4), (vint16m2_t)(op5), (vint16m2_t)(op6), (vint16m2_t)(op7), (vint16m2_t)(op8), (const int16_t *)(op9), (vuint16m2_t)(op10), (size_t)(op11)) +#define vloxseg4ei16_v_i16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei16_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (const int16_t *)(op4), (vuint16mf2_t)(op5), (size_t)(op6)) +#define vloxseg4ei16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei16_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vbool32_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (const int16_t *)(op9), (vuint16mf2_t)(op10), (size_t)(op11)) +#define vloxseg4ei16_v_i16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei16_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (const int16_t *)(op4), (vuint16mf4_t)(op5), (size_t)(op6)) +#define vloxseg4ei16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei16_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vbool64_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (const int16_t *)(op9), (vuint16mf4_t)(op10), (size_t)(op11)) +#define vloxseg4ei16_v_u16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei16_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (const uint16_t *)(op4), (vuint16m1_t)(op5), (size_t)(op6)) +#define vloxseg4ei16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei16_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vbool16_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (const uint16_t *)(op9), (vuint16m1_t)(op10), (size_t)(op11)) +#define vloxseg4ei16_v_u16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei16_v_u16m2((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vuint16m2_t *)(op2), (vuint16m2_t *)(op3), (const uint16_t *)(op4), (vuint16m2_t)(op5), (size_t)(op6)) +#define vloxseg4ei16_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei16_v_u16m2_m((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vuint16m2_t *)(op2), (vuint16m2_t *)(op3), (vbool8_t)(op4), (vuint16m2_t)(op5), (vuint16m2_t)(op6), (vuint16m2_t)(op7), (vuint16m2_t)(op8), (const uint16_t *)(op9), (vuint16m2_t)(op10), (size_t)(op11)) +#define vloxseg4ei16_v_u16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei16_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (const uint16_t *)(op4), (vuint16mf2_t)(op5), (size_t)(op6)) +#define vloxseg4ei16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei16_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vbool32_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (const uint16_t *)(op9), (vuint16mf2_t)(op10), (size_t)(op11)) +#define vloxseg4ei16_v_u16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei16_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (const uint16_t *)(op4), (vuint16mf4_t)(op5), (size_t)(op6)) +#define vloxseg4ei16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei16_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vbool64_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (const uint16_t *)(op9), (vuint16mf4_t)(op10), (size_t)(op11)) +#define vloxseg5ei16_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei16_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (const int16_t *)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vloxseg5ei16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei16_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vbool16_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (const int16_t *)(op11), (vuint16m1_t)(op12), (size_t)(op13)) +#define vloxseg5ei16_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei16_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (const int16_t *)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vloxseg5ei16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei16_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vbool32_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (const int16_t *)(op11), (vuint16mf2_t)(op12), (size_t)(op13)) +#define vloxseg5ei16_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei16_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (const int16_t *)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vloxseg5ei16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei16_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vbool64_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (const int16_t *)(op11), (vuint16mf4_t)(op12), (size_t)(op13)) +#define vloxseg5ei16_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei16_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (const uint16_t *)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vloxseg5ei16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei16_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vbool16_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (const uint16_t *)(op11), (vuint16m1_t)(op12), (size_t)(op13)) +#define vloxseg5ei16_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei16_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (const uint16_t *)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vloxseg5ei16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei16_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vbool32_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (const uint16_t *)(op11), (vuint16mf2_t)(op12), (size_t)(op13)) +#define vloxseg5ei16_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei16_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (const uint16_t *)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vloxseg5ei16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei16_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vbool64_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (const uint16_t *)(op11), (vuint16mf4_t)(op12), (size_t)(op13)) +#define vloxseg6ei16_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei16_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (const int16_t *)(op6), (vuint16m1_t)(op7), (size_t)(op8)) +#define vloxseg6ei16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei16_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vbool16_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (vint16m1_t)(op11), (vint16m1_t)(op12), (const int16_t *)(op13), (vuint16m1_t)(op14), (size_t)(op15)) +#define vloxseg6ei16_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei16_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (const int16_t *)(op6), (vuint16mf2_t)(op7), (size_t)(op8)) +#define vloxseg6ei16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei16_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vbool32_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (vint16mf2_t)(op11), (vint16mf2_t)(op12), (const int16_t *)(op13), (vuint16mf2_t)(op14), (size_t)(op15)) +#define vloxseg6ei16_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei16_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (const int16_t *)(op6), (vuint16mf4_t)(op7), (size_t)(op8)) +#define vloxseg6ei16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei16_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vbool64_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (vint16mf4_t)(op11), (vint16mf4_t)(op12), (const int16_t *)(op13), (vuint16mf4_t)(op14), (size_t)(op15)) +#define vloxseg6ei16_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei16_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (const uint16_t *)(op6), (vuint16m1_t)(op7), (size_t)(op8)) +#define vloxseg6ei16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei16_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vbool16_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (vuint16m1_t)(op11), (vuint16m1_t)(op12), (const uint16_t *)(op13), (vuint16m1_t)(op14), (size_t)(op15)) +#define vloxseg6ei16_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei16_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (const uint16_t *)(op6), (vuint16mf2_t)(op7), (size_t)(op8)) +#define vloxseg6ei16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei16_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vbool32_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (vuint16mf2_t)(op11), (vuint16mf2_t)(op12), (const uint16_t *)(op13), (vuint16mf2_t)(op14), (size_t)(op15)) +#define vloxseg6ei16_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei16_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (const uint16_t *)(op6), (vuint16mf4_t)(op7), (size_t)(op8)) +#define vloxseg6ei16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei16_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vbool64_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (vuint16mf4_t)(op11), (vuint16mf4_t)(op12), (const uint16_t *)(op13), (vuint16mf4_t)(op14), (size_t)(op15)) +#define vloxseg7ei16_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei16_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vint16m1_t *)(op6), (const int16_t *)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vloxseg7ei16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei16_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vint16m1_t *)(op6), (vbool16_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (vint16m1_t)(op11), (vint16m1_t)(op12), (vint16m1_t)(op13), (vint16m1_t)(op14), (const int16_t *)(op15), (vuint16m1_t)(op16), (size_t)(op17)) +#define vloxseg7ei16_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei16_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vint16mf2_t *)(op6), (const int16_t *)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vloxseg7ei16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei16_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vint16mf2_t *)(op6), (vbool32_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (vint16mf2_t)(op11), (vint16mf2_t)(op12), (vint16mf2_t)(op13), (vint16mf2_t)(op14), (const int16_t *)(op15), (vuint16mf2_t)(op16), (size_t)(op17)) +#define vloxseg7ei16_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei16_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vint16mf4_t *)(op6), (const int16_t *)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vloxseg7ei16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei16_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vint16mf4_t *)(op6), (vbool64_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (vint16mf4_t)(op11), (vint16mf4_t)(op12), (vint16mf4_t)(op13), (vint16mf4_t)(op14), (const int16_t *)(op15), (vuint16mf4_t)(op16), (size_t)(op17)) +#define vloxseg7ei16_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei16_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vuint16m1_t *)(op6), (const uint16_t *)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vloxseg7ei16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei16_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vuint16m1_t *)(op6), (vbool16_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (vuint16m1_t)(op11), (vuint16m1_t)(op12), (vuint16m1_t)(op13), (vuint16m1_t)(op14), (const uint16_t *)(op15), (vuint16m1_t)(op16), (size_t)(op17)) +#define vloxseg7ei16_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei16_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vuint16mf2_t *)(op6), (const uint16_t *)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vloxseg7ei16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei16_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vuint16mf2_t *)(op6), (vbool32_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (vuint16mf2_t)(op11), (vuint16mf2_t)(op12), (vuint16mf2_t)(op13), (vuint16mf2_t)(op14), (const uint16_t *)(op15), (vuint16mf2_t)(op16), (size_t)(op17)) +#define vloxseg7ei16_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei16_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vuint16mf4_t *)(op6), (const uint16_t *)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vloxseg7ei16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei16_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vuint16mf4_t *)(op6), (vbool64_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (vuint16mf4_t)(op11), (vuint16mf4_t)(op12), (vuint16mf4_t)(op13), (vuint16mf4_t)(op14), (const uint16_t *)(op15), (vuint16mf4_t)(op16), (size_t)(op17)) +#define vloxseg8ei16_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei16_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vint16m1_t *)(op6), (vint16m1_t *)(op7), (const int16_t *)(op8), (vuint16m1_t)(op9), (size_t)(op10)) +#define vloxseg8ei16_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei16_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vint16m1_t *)(op6), (vint16m1_t *)(op7), (vbool16_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (vint16m1_t)(op11), (vint16m1_t)(op12), (vint16m1_t)(op13), (vint16m1_t)(op14), (vint16m1_t)(op15), (vint16m1_t)(op16), (const int16_t *)(op17), (vuint16m1_t)(op18), (size_t)(op19)) +#define vloxseg8ei16_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei16_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vint16mf2_t *)(op6), (vint16mf2_t *)(op7), (const int16_t *)(op8), (vuint16mf2_t)(op9), (size_t)(op10)) +#define vloxseg8ei16_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei16_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vint16mf2_t *)(op6), (vint16mf2_t *)(op7), (vbool32_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (vint16mf2_t)(op11), (vint16mf2_t)(op12), (vint16mf2_t)(op13), (vint16mf2_t)(op14), (vint16mf2_t)(op15), (vint16mf2_t)(op16), (const int16_t *)(op17), (vuint16mf2_t)(op18), (size_t)(op19)) +#define vloxseg8ei16_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei16_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vint16mf4_t *)(op6), (vint16mf4_t *)(op7), (const int16_t *)(op8), (vuint16mf4_t)(op9), (size_t)(op10)) +#define vloxseg8ei16_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei16_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vint16mf4_t *)(op6), (vint16mf4_t *)(op7), (vbool64_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (vint16mf4_t)(op11), (vint16mf4_t)(op12), (vint16mf4_t)(op13), (vint16mf4_t)(op14), (vint16mf4_t)(op15), (vint16mf4_t)(op16), (const int16_t *)(op17), (vuint16mf4_t)(op18), (size_t)(op19)) +#define vloxseg8ei16_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei16_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vuint16m1_t *)(op6), (vuint16m1_t *)(op7), (const uint16_t *)(op8), (vuint16m1_t)(op9), (size_t)(op10)) +#define vloxseg8ei16_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei16_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vuint16m1_t *)(op6), (vuint16m1_t *)(op7), (vbool16_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (vuint16m1_t)(op11), (vuint16m1_t)(op12), (vuint16m1_t)(op13), (vuint16m1_t)(op14), (vuint16m1_t)(op15), (vuint16m1_t)(op16), (const uint16_t *)(op17), (vuint16m1_t)(op18), (size_t)(op19)) +#define vloxseg8ei16_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei16_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vuint16mf2_t *)(op6), (vuint16mf2_t *)(op7), (const uint16_t *)(op8), (vuint16mf2_t)(op9), (size_t)(op10)) +#define vloxseg8ei16_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei16_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vuint16mf2_t *)(op6), (vuint16mf2_t *)(op7), (vbool32_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (vuint16mf2_t)(op11), (vuint16mf2_t)(op12), (vuint16mf2_t)(op13), (vuint16mf2_t)(op14), (vuint16mf2_t)(op15), (vuint16mf2_t)(op16), (const uint16_t *)(op17), (vuint16mf2_t)(op18), (size_t)(op19)) +#define vloxseg8ei16_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei16_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vuint16mf4_t *)(op6), (vuint16mf4_t *)(op7), (const uint16_t *)(op8), (vuint16mf4_t)(op9), (size_t)(op10)) +#define vloxseg8ei16_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei16_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vuint16mf4_t *)(op6), (vuint16mf4_t *)(op7), (vbool64_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (vuint16mf4_t)(op11), (vuint16mf4_t)(op12), (vuint16mf4_t)(op13), (vuint16mf4_t)(op14), (vuint16mf4_t)(op15), (vuint16mf4_t)(op16), (const uint16_t *)(op17), (vuint16mf4_t)(op18), (size_t)(op19)) +#define vloxseg2ei32_v_i16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (const int16_t *)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vbool16_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (const int16_t *)(op5), (vuint32m2_t)(op6), (size_t)(op7)) +#define vloxseg2ei32_v_i16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_i16m2((vint16m2_t *)(op0), (vint16m2_t *)(op1), (const int16_t *)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_i16m2_m((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vbool8_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (const int16_t *)(op5), (vuint32m4_t)(op6), (size_t)(op7)) +#define vloxseg2ei32_v_i16m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_i16m4((vint16m4_t *)(op0), (vint16m4_t *)(op1), (const int16_t *)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_i16m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_i16m4_m((vint16m4_t *)(op0), (vint16m4_t *)(op1), (vbool4_t)(op2), (vint16m4_t)(op3), (vint16m4_t)(op4), (const int16_t *)(op5), (vuint32m8_t)(op6), (size_t)(op7)) +#define vloxseg2ei32_v_i16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (const int16_t *)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vbool32_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (const int16_t *)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vloxseg2ei32_v_i16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (const int16_t *)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vbool64_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (const int16_t *)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vloxseg2ei32_v_u16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (const uint16_t *)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vbool16_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (const uint16_t *)(op5), (vuint32m2_t)(op6), (size_t)(op7)) +#define vloxseg2ei32_v_u16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_u16m2((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (const uint16_t *)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_u16m2_m((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vbool8_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (const uint16_t *)(op5), (vuint32m4_t)(op6), (size_t)(op7)) +#define vloxseg2ei32_v_u16m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_u16m4((vuint16m4_t *)(op0), (vuint16m4_t *)(op1), (const uint16_t *)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_u16m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_u16m4_m((vuint16m4_t *)(op0), (vuint16m4_t *)(op1), (vbool4_t)(op2), (vuint16m4_t)(op3), (vuint16m4_t)(op4), (const uint16_t *)(op5), (vuint32m8_t)(op6), (size_t)(op7)) +#define vloxseg2ei32_v_u16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (const uint16_t *)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vbool32_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (const uint16_t *)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vloxseg2ei32_v_u16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (const uint16_t *)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vbool64_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (const uint16_t *)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vloxseg3ei32_v_i16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei32_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (const int16_t *)(op3), (vuint32m2_t)(op4), (size_t)(op5)) +#define vloxseg3ei32_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei32_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vbool16_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (const int16_t *)(op7), (vuint32m2_t)(op8), (size_t)(op9)) +#define vloxseg3ei32_v_i16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei32_v_i16m2((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vint16m2_t *)(op2), (const int16_t *)(op3), (vuint32m4_t)(op4), (size_t)(op5)) +#define vloxseg3ei32_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei32_v_i16m2_m((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vint16m2_t *)(op2), (vbool8_t)(op3), (vint16m2_t)(op4), (vint16m2_t)(op5), (vint16m2_t)(op6), (const int16_t *)(op7), (vuint32m4_t)(op8), (size_t)(op9)) +#define vloxseg3ei32_v_i16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei32_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (const int16_t *)(op3), (vuint32m1_t)(op4), (size_t)(op5)) +#define vloxseg3ei32_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei32_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vbool32_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (const int16_t *)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vloxseg3ei32_v_i16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei32_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (const int16_t *)(op3), (vuint32mf2_t)(op4), (size_t)(op5)) +#define vloxseg3ei32_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei32_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vbool64_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (const int16_t *)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vloxseg3ei32_v_u16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei32_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (const uint16_t *)(op3), (vuint32m2_t)(op4), (size_t)(op5)) +#define vloxseg3ei32_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei32_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vbool16_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (const uint16_t *)(op7), (vuint32m2_t)(op8), (size_t)(op9)) +#define vloxseg3ei32_v_u16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei32_v_u16m2((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vuint16m2_t *)(op2), (const uint16_t *)(op3), (vuint32m4_t)(op4), (size_t)(op5)) +#define vloxseg3ei32_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei32_v_u16m2_m((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vuint16m2_t *)(op2), (vbool8_t)(op3), (vuint16m2_t)(op4), (vuint16m2_t)(op5), (vuint16m2_t)(op6), (const uint16_t *)(op7), (vuint32m4_t)(op8), (size_t)(op9)) +#define vloxseg3ei32_v_u16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei32_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (const uint16_t *)(op3), (vuint32m1_t)(op4), (size_t)(op5)) +#define vloxseg3ei32_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei32_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vbool32_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (const uint16_t *)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vloxseg3ei32_v_u16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei32_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (const uint16_t *)(op3), (vuint32mf2_t)(op4), (size_t)(op5)) +#define vloxseg3ei32_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei32_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vbool64_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (const uint16_t *)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vloxseg4ei32_v_i16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei32_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (const int16_t *)(op4), (vuint32m2_t)(op5), (size_t)(op6)) +#define vloxseg4ei32_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei32_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vbool16_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (const int16_t *)(op9), (vuint32m2_t)(op10), (size_t)(op11)) +#define vloxseg4ei32_v_i16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei32_v_i16m2((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vint16m2_t *)(op2), (vint16m2_t *)(op3), (const int16_t *)(op4), (vuint32m4_t)(op5), (size_t)(op6)) +#define vloxseg4ei32_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei32_v_i16m2_m((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vint16m2_t *)(op2), (vint16m2_t *)(op3), (vbool8_t)(op4), (vint16m2_t)(op5), (vint16m2_t)(op6), (vint16m2_t)(op7), (vint16m2_t)(op8), (const int16_t *)(op9), (vuint32m4_t)(op10), (size_t)(op11)) +#define vloxseg4ei32_v_i16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei32_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (const int16_t *)(op4), (vuint32m1_t)(op5), (size_t)(op6)) +#define vloxseg4ei32_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei32_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vbool32_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (const int16_t *)(op9), (vuint32m1_t)(op10), (size_t)(op11)) +#define vloxseg4ei32_v_i16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei32_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (const int16_t *)(op4), (vuint32mf2_t)(op5), (size_t)(op6)) +#define vloxseg4ei32_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei32_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vbool64_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (const int16_t *)(op9), (vuint32mf2_t)(op10), (size_t)(op11)) +#define vloxseg4ei32_v_u16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei32_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (const uint16_t *)(op4), (vuint32m2_t)(op5), (size_t)(op6)) +#define vloxseg4ei32_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei32_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vbool16_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (const uint16_t *)(op9), (vuint32m2_t)(op10), (size_t)(op11)) +#define vloxseg4ei32_v_u16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei32_v_u16m2((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vuint16m2_t *)(op2), (vuint16m2_t *)(op3), (const uint16_t *)(op4), (vuint32m4_t)(op5), (size_t)(op6)) +#define vloxseg4ei32_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei32_v_u16m2_m((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vuint16m2_t *)(op2), (vuint16m2_t *)(op3), (vbool8_t)(op4), (vuint16m2_t)(op5), (vuint16m2_t)(op6), (vuint16m2_t)(op7), (vuint16m2_t)(op8), (const uint16_t *)(op9), (vuint32m4_t)(op10), (size_t)(op11)) +#define vloxseg4ei32_v_u16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei32_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (const uint16_t *)(op4), (vuint32m1_t)(op5), (size_t)(op6)) +#define vloxseg4ei32_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei32_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vbool32_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (const uint16_t *)(op9), (vuint32m1_t)(op10), (size_t)(op11)) +#define vloxseg4ei32_v_u16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei32_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (const uint16_t *)(op4), (vuint32mf2_t)(op5), (size_t)(op6)) +#define vloxseg4ei32_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei32_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vbool64_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (const uint16_t *)(op9), (vuint32mf2_t)(op10), (size_t)(op11)) +#define vloxseg5ei32_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei32_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (const int16_t *)(op5), (vuint32m2_t)(op6), (size_t)(op7)) +#define vloxseg5ei32_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei32_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vbool16_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (const int16_t *)(op11), (vuint32m2_t)(op12), (size_t)(op13)) +#define vloxseg5ei32_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei32_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (const int16_t *)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vloxseg5ei32_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei32_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vbool32_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (const int16_t *)(op11), (vuint32m1_t)(op12), (size_t)(op13)) +#define vloxseg5ei32_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei32_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (const int16_t *)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vloxseg5ei32_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei32_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vbool64_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (const int16_t *)(op11), (vuint32mf2_t)(op12), (size_t)(op13)) +#define vloxseg5ei32_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei32_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (const uint16_t *)(op5), (vuint32m2_t)(op6), (size_t)(op7)) +#define vloxseg5ei32_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei32_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vbool16_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (const uint16_t *)(op11), (vuint32m2_t)(op12), (size_t)(op13)) +#define vloxseg5ei32_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei32_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (const uint16_t *)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vloxseg5ei32_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei32_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vbool32_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (const uint16_t *)(op11), (vuint32m1_t)(op12), (size_t)(op13)) +#define vloxseg5ei32_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei32_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (const uint16_t *)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vloxseg5ei32_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei32_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vbool64_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (const uint16_t *)(op11), (vuint32mf2_t)(op12), (size_t)(op13)) +#define vloxseg6ei32_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei32_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (const int16_t *)(op6), (vuint32m2_t)(op7), (size_t)(op8)) +#define vloxseg6ei32_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei32_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vbool16_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (vint16m1_t)(op11), (vint16m1_t)(op12), (const int16_t *)(op13), (vuint32m2_t)(op14), (size_t)(op15)) +#define vloxseg6ei32_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei32_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (const int16_t *)(op6), (vuint32m1_t)(op7), (size_t)(op8)) +#define vloxseg6ei32_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei32_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vbool32_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (vint16mf2_t)(op11), (vint16mf2_t)(op12), (const int16_t *)(op13), (vuint32m1_t)(op14), (size_t)(op15)) +#define vloxseg6ei32_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei32_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (const int16_t *)(op6), (vuint32mf2_t)(op7), (size_t)(op8)) +#define vloxseg6ei32_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei32_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vbool64_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (vint16mf4_t)(op11), (vint16mf4_t)(op12), (const int16_t *)(op13), (vuint32mf2_t)(op14), (size_t)(op15)) +#define vloxseg6ei32_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei32_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (const uint16_t *)(op6), (vuint32m2_t)(op7), (size_t)(op8)) +#define vloxseg6ei32_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei32_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vbool16_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (vuint16m1_t)(op11), (vuint16m1_t)(op12), (const uint16_t *)(op13), (vuint32m2_t)(op14), (size_t)(op15)) +#define vloxseg6ei32_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei32_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (const uint16_t *)(op6), (vuint32m1_t)(op7), (size_t)(op8)) +#define vloxseg6ei32_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei32_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vbool32_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (vuint16mf2_t)(op11), (vuint16mf2_t)(op12), (const uint16_t *)(op13), (vuint32m1_t)(op14), (size_t)(op15)) +#define vloxseg6ei32_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei32_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (const uint16_t *)(op6), (vuint32mf2_t)(op7), (size_t)(op8)) +#define vloxseg6ei32_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei32_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vbool64_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (vuint16mf4_t)(op11), (vuint16mf4_t)(op12), (const uint16_t *)(op13), (vuint32mf2_t)(op14), (size_t)(op15)) +#define vloxseg7ei32_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei32_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vint16m1_t *)(op6), (const int16_t *)(op7), (vuint32m2_t)(op8), (size_t)(op9)) +#define vloxseg7ei32_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei32_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vint16m1_t *)(op6), (vbool16_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (vint16m1_t)(op11), (vint16m1_t)(op12), (vint16m1_t)(op13), (vint16m1_t)(op14), (const int16_t *)(op15), (vuint32m2_t)(op16), (size_t)(op17)) +#define vloxseg7ei32_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei32_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vint16mf2_t *)(op6), (const int16_t *)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vloxseg7ei32_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei32_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vint16mf2_t *)(op6), (vbool32_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (vint16mf2_t)(op11), (vint16mf2_t)(op12), (vint16mf2_t)(op13), (vint16mf2_t)(op14), (const int16_t *)(op15), (vuint32m1_t)(op16), (size_t)(op17)) +#define vloxseg7ei32_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei32_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vint16mf4_t *)(op6), (const int16_t *)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vloxseg7ei32_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei32_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vint16mf4_t *)(op6), (vbool64_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (vint16mf4_t)(op11), (vint16mf4_t)(op12), (vint16mf4_t)(op13), (vint16mf4_t)(op14), (const int16_t *)(op15), (vuint32mf2_t)(op16), (size_t)(op17)) +#define vloxseg7ei32_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei32_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vuint16m1_t *)(op6), (const uint16_t *)(op7), (vuint32m2_t)(op8), (size_t)(op9)) +#define vloxseg7ei32_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei32_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vuint16m1_t *)(op6), (vbool16_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (vuint16m1_t)(op11), (vuint16m1_t)(op12), (vuint16m1_t)(op13), (vuint16m1_t)(op14), (const uint16_t *)(op15), (vuint32m2_t)(op16), (size_t)(op17)) +#define vloxseg7ei32_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei32_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vuint16mf2_t *)(op6), (const uint16_t *)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vloxseg7ei32_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei32_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vuint16mf2_t *)(op6), (vbool32_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (vuint16mf2_t)(op11), (vuint16mf2_t)(op12), (vuint16mf2_t)(op13), (vuint16mf2_t)(op14), (const uint16_t *)(op15), (vuint32m1_t)(op16), (size_t)(op17)) +#define vloxseg7ei32_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei32_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vuint16mf4_t *)(op6), (const uint16_t *)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vloxseg7ei32_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei32_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vuint16mf4_t *)(op6), (vbool64_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (vuint16mf4_t)(op11), (vuint16mf4_t)(op12), (vuint16mf4_t)(op13), (vuint16mf4_t)(op14), (const uint16_t *)(op15), (vuint32mf2_t)(op16), (size_t)(op17)) +#define vloxseg8ei32_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei32_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vint16m1_t *)(op6), (vint16m1_t *)(op7), (const int16_t *)(op8), (vuint32m2_t)(op9), (size_t)(op10)) +#define vloxseg8ei32_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei32_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vint16m1_t *)(op6), (vint16m1_t *)(op7), (vbool16_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (vint16m1_t)(op11), (vint16m1_t)(op12), (vint16m1_t)(op13), (vint16m1_t)(op14), (vint16m1_t)(op15), (vint16m1_t)(op16), (const int16_t *)(op17), (vuint32m2_t)(op18), (size_t)(op19)) +#define vloxseg8ei32_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei32_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vint16mf2_t *)(op6), (vint16mf2_t *)(op7), (const int16_t *)(op8), (vuint32m1_t)(op9), (size_t)(op10)) +#define vloxseg8ei32_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei32_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vint16mf2_t *)(op6), (vint16mf2_t *)(op7), (vbool32_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (vint16mf2_t)(op11), (vint16mf2_t)(op12), (vint16mf2_t)(op13), (vint16mf2_t)(op14), (vint16mf2_t)(op15), (vint16mf2_t)(op16), (const int16_t *)(op17), (vuint32m1_t)(op18), (size_t)(op19)) +#define vloxseg8ei32_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei32_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vint16mf4_t *)(op6), (vint16mf4_t *)(op7), (const int16_t *)(op8), (vuint32mf2_t)(op9), (size_t)(op10)) +#define vloxseg8ei32_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei32_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vint16mf4_t *)(op6), (vint16mf4_t *)(op7), (vbool64_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (vint16mf4_t)(op11), (vint16mf4_t)(op12), (vint16mf4_t)(op13), (vint16mf4_t)(op14), (vint16mf4_t)(op15), (vint16mf4_t)(op16), (const int16_t *)(op17), (vuint32mf2_t)(op18), (size_t)(op19)) +#define vloxseg8ei32_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei32_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vuint16m1_t *)(op6), (vuint16m1_t *)(op7), (const uint16_t *)(op8), (vuint32m2_t)(op9), (size_t)(op10)) +#define vloxseg8ei32_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei32_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vuint16m1_t *)(op6), (vuint16m1_t *)(op7), (vbool16_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (vuint16m1_t)(op11), (vuint16m1_t)(op12), (vuint16m1_t)(op13), (vuint16m1_t)(op14), (vuint16m1_t)(op15), (vuint16m1_t)(op16), (const uint16_t *)(op17), (vuint32m2_t)(op18), (size_t)(op19)) +#define vloxseg8ei32_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei32_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vuint16mf2_t *)(op6), (vuint16mf2_t *)(op7), (const uint16_t *)(op8), (vuint32m1_t)(op9), (size_t)(op10)) +#define vloxseg8ei32_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei32_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vuint16mf2_t *)(op6), (vuint16mf2_t *)(op7), (vbool32_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (vuint16mf2_t)(op11), (vuint16mf2_t)(op12), (vuint16mf2_t)(op13), (vuint16mf2_t)(op14), (vuint16mf2_t)(op15), (vuint16mf2_t)(op16), (const uint16_t *)(op17), (vuint32m1_t)(op18), (size_t)(op19)) +#define vloxseg8ei32_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei32_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vuint16mf4_t *)(op6), (vuint16mf4_t *)(op7), (const uint16_t *)(op8), (vuint32mf2_t)(op9), (size_t)(op10)) +#define vloxseg8ei32_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei32_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vuint16mf4_t *)(op6), (vuint16mf4_t *)(op7), (vbool64_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (vuint16mf4_t)(op11), (vuint16mf4_t)(op12), (vuint16mf4_t)(op13), (vuint16mf4_t)(op14), (vuint16mf4_t)(op15), (vuint16mf4_t)(op16), (const uint16_t *)(op17), (vuint32mf2_t)(op18), (size_t)(op19)) +#define vloxseg2ei64_v_i16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei64_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (const int16_t *)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vloxseg2ei64_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei64_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vbool16_t)(op2), (vint16m1_t)(op3), (vint16m1_t)(op4), (const int16_t *)(op5), (vuint64m4_t)(op6), (size_t)(op7)) +#define vloxseg2ei64_v_i16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei64_v_i16m2((vint16m2_t *)(op0), (vint16m2_t *)(op1), (const int16_t *)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vloxseg2ei64_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei64_v_i16m2_m((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vbool8_t)(op2), (vint16m2_t)(op3), (vint16m2_t)(op4), (const int16_t *)(op5), (vuint64m8_t)(op6), (size_t)(op7)) +#define vloxseg2ei64_v_i16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei64_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (const int16_t *)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vloxseg2ei64_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei64_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vbool32_t)(op2), (vint16mf2_t)(op3), (vint16mf2_t)(op4), (const int16_t *)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vloxseg2ei64_v_i16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei64_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (const int16_t *)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vloxseg2ei64_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei64_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vbool64_t)(op2), (vint16mf4_t)(op3), (vint16mf4_t)(op4), (const int16_t *)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vloxseg2ei64_v_u16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei64_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (const uint16_t *)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vloxseg2ei64_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei64_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vbool16_t)(op2), (vuint16m1_t)(op3), (vuint16m1_t)(op4), (const uint16_t *)(op5), (vuint64m4_t)(op6), (size_t)(op7)) +#define vloxseg2ei64_v_u16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei64_v_u16m2((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (const uint16_t *)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vloxseg2ei64_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei64_v_u16m2_m((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vbool8_t)(op2), (vuint16m2_t)(op3), (vuint16m2_t)(op4), (const uint16_t *)(op5), (vuint64m8_t)(op6), (size_t)(op7)) +#define vloxseg2ei64_v_u16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei64_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (const uint16_t *)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vloxseg2ei64_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei64_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vbool32_t)(op2), (vuint16mf2_t)(op3), (vuint16mf2_t)(op4), (const uint16_t *)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vloxseg2ei64_v_u16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei64_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (const uint16_t *)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vloxseg2ei64_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei64_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vbool64_t)(op2), (vuint16mf4_t)(op3), (vuint16mf4_t)(op4), (const uint16_t *)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vloxseg3ei64_v_i16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei64_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (const int16_t *)(op3), (vuint64m4_t)(op4), (size_t)(op5)) +#define vloxseg3ei64_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei64_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vbool16_t)(op3), (vint16m1_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (const int16_t *)(op7), (vuint64m4_t)(op8), (size_t)(op9)) +#define vloxseg3ei64_v_i16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei64_v_i16m2((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vint16m2_t *)(op2), (const int16_t *)(op3), (vuint64m8_t)(op4), (size_t)(op5)) +#define vloxseg3ei64_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei64_v_i16m2_m((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vint16m2_t *)(op2), (vbool8_t)(op3), (vint16m2_t)(op4), (vint16m2_t)(op5), (vint16m2_t)(op6), (const int16_t *)(op7), (vuint64m8_t)(op8), (size_t)(op9)) +#define vloxseg3ei64_v_i16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei64_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (const int16_t *)(op3), (vuint64m2_t)(op4), (size_t)(op5)) +#define vloxseg3ei64_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei64_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vbool32_t)(op3), (vint16mf2_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (const int16_t *)(op7), (vuint64m2_t)(op8), (size_t)(op9)) +#define vloxseg3ei64_v_i16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei64_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (const int16_t *)(op3), (vuint64m1_t)(op4), (size_t)(op5)) +#define vloxseg3ei64_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei64_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vbool64_t)(op3), (vint16mf4_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (const int16_t *)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vloxseg3ei64_v_u16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei64_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (const uint16_t *)(op3), (vuint64m4_t)(op4), (size_t)(op5)) +#define vloxseg3ei64_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei64_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vbool16_t)(op3), (vuint16m1_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (const uint16_t *)(op7), (vuint64m4_t)(op8), (size_t)(op9)) +#define vloxseg3ei64_v_u16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei64_v_u16m2((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vuint16m2_t *)(op2), (const uint16_t *)(op3), (vuint64m8_t)(op4), (size_t)(op5)) +#define vloxseg3ei64_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei64_v_u16m2_m((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vuint16m2_t *)(op2), (vbool8_t)(op3), (vuint16m2_t)(op4), (vuint16m2_t)(op5), (vuint16m2_t)(op6), (const uint16_t *)(op7), (vuint64m8_t)(op8), (size_t)(op9)) +#define vloxseg3ei64_v_u16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei64_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (const uint16_t *)(op3), (vuint64m2_t)(op4), (size_t)(op5)) +#define vloxseg3ei64_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei64_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vbool32_t)(op3), (vuint16mf2_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (const uint16_t *)(op7), (vuint64m2_t)(op8), (size_t)(op9)) +#define vloxseg3ei64_v_u16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei64_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (const uint16_t *)(op3), (vuint64m1_t)(op4), (size_t)(op5)) +#define vloxseg3ei64_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei64_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vbool64_t)(op3), (vuint16mf4_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (const uint16_t *)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vloxseg4ei64_v_i16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei64_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (const int16_t *)(op4), (vuint64m4_t)(op5), (size_t)(op6)) +#define vloxseg4ei64_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei64_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vbool16_t)(op4), (vint16m1_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (const int16_t *)(op9), (vuint64m4_t)(op10), (size_t)(op11)) +#define vloxseg4ei64_v_i16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei64_v_i16m2((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vint16m2_t *)(op2), (vint16m2_t *)(op3), (const int16_t *)(op4), (vuint64m8_t)(op5), (size_t)(op6)) +#define vloxseg4ei64_v_i16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei64_v_i16m2_m((vint16m2_t *)(op0), (vint16m2_t *)(op1), (vint16m2_t *)(op2), (vint16m2_t *)(op3), (vbool8_t)(op4), (vint16m2_t)(op5), (vint16m2_t)(op6), (vint16m2_t)(op7), (vint16m2_t)(op8), (const int16_t *)(op9), (vuint64m8_t)(op10), (size_t)(op11)) +#define vloxseg4ei64_v_i16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei64_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (const int16_t *)(op4), (vuint64m2_t)(op5), (size_t)(op6)) +#define vloxseg4ei64_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei64_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vbool32_t)(op4), (vint16mf2_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (const int16_t *)(op9), (vuint64m2_t)(op10), (size_t)(op11)) +#define vloxseg4ei64_v_i16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei64_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (const int16_t *)(op4), (vuint64m1_t)(op5), (size_t)(op6)) +#define vloxseg4ei64_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei64_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vbool64_t)(op4), (vint16mf4_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (const int16_t *)(op9), (vuint64m1_t)(op10), (size_t)(op11)) +#define vloxseg4ei64_v_u16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei64_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (const uint16_t *)(op4), (vuint64m4_t)(op5), (size_t)(op6)) +#define vloxseg4ei64_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei64_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vbool16_t)(op4), (vuint16m1_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (const uint16_t *)(op9), (vuint64m4_t)(op10), (size_t)(op11)) +#define vloxseg4ei64_v_u16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei64_v_u16m2((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vuint16m2_t *)(op2), (vuint16m2_t *)(op3), (const uint16_t *)(op4), (vuint64m8_t)(op5), (size_t)(op6)) +#define vloxseg4ei64_v_u16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei64_v_u16m2_m((vuint16m2_t *)(op0), (vuint16m2_t *)(op1), (vuint16m2_t *)(op2), (vuint16m2_t *)(op3), (vbool8_t)(op4), (vuint16m2_t)(op5), (vuint16m2_t)(op6), (vuint16m2_t)(op7), (vuint16m2_t)(op8), (const uint16_t *)(op9), (vuint64m8_t)(op10), (size_t)(op11)) +#define vloxseg4ei64_v_u16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei64_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (const uint16_t *)(op4), (vuint64m2_t)(op5), (size_t)(op6)) +#define vloxseg4ei64_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei64_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vbool32_t)(op4), (vuint16mf2_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (const uint16_t *)(op9), (vuint64m2_t)(op10), (size_t)(op11)) +#define vloxseg4ei64_v_u16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei64_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (const uint16_t *)(op4), (vuint64m1_t)(op5), (size_t)(op6)) +#define vloxseg4ei64_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei64_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vbool64_t)(op4), (vuint16mf4_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (const uint16_t *)(op9), (vuint64m1_t)(op10), (size_t)(op11)) +#define vloxseg5ei64_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei64_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (const int16_t *)(op5), (vuint64m4_t)(op6), (size_t)(op7)) +#define vloxseg5ei64_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei64_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vbool16_t)(op5), (vint16m1_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (const int16_t *)(op11), (vuint64m4_t)(op12), (size_t)(op13)) +#define vloxseg5ei64_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei64_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (const int16_t *)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vloxseg5ei64_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei64_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vbool32_t)(op5), (vint16mf2_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (const int16_t *)(op11), (vuint64m2_t)(op12), (size_t)(op13)) +#define vloxseg5ei64_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei64_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (const int16_t *)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vloxseg5ei64_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei64_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vbool64_t)(op5), (vint16mf4_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (const int16_t *)(op11), (vuint64m1_t)(op12), (size_t)(op13)) +#define vloxseg5ei64_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei64_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (const uint16_t *)(op5), (vuint64m4_t)(op6), (size_t)(op7)) +#define vloxseg5ei64_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei64_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vbool16_t)(op5), (vuint16m1_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (const uint16_t *)(op11), (vuint64m4_t)(op12), (size_t)(op13)) +#define vloxseg5ei64_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei64_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (const uint16_t *)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vloxseg5ei64_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei64_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vbool32_t)(op5), (vuint16mf2_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (const uint16_t *)(op11), (vuint64m2_t)(op12), (size_t)(op13)) +#define vloxseg5ei64_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei64_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (const uint16_t *)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vloxseg5ei64_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei64_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vbool64_t)(op5), (vuint16mf4_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (const uint16_t *)(op11), (vuint64m1_t)(op12), (size_t)(op13)) +#define vloxseg6ei64_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei64_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (const int16_t *)(op6), (vuint64m4_t)(op7), (size_t)(op8)) +#define vloxseg6ei64_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei64_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vbool16_t)(op6), (vint16m1_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (vint16m1_t)(op11), (vint16m1_t)(op12), (const int16_t *)(op13), (vuint64m4_t)(op14), (size_t)(op15)) +#define vloxseg6ei64_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei64_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (const int16_t *)(op6), (vuint64m2_t)(op7), (size_t)(op8)) +#define vloxseg6ei64_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei64_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vbool32_t)(op6), (vint16mf2_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (vint16mf2_t)(op11), (vint16mf2_t)(op12), (const int16_t *)(op13), (vuint64m2_t)(op14), (size_t)(op15)) +#define vloxseg6ei64_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei64_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (const int16_t *)(op6), (vuint64m1_t)(op7), (size_t)(op8)) +#define vloxseg6ei64_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei64_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vbool64_t)(op6), (vint16mf4_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (vint16mf4_t)(op11), (vint16mf4_t)(op12), (const int16_t *)(op13), (vuint64m1_t)(op14), (size_t)(op15)) +#define vloxseg6ei64_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei64_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (const uint16_t *)(op6), (vuint64m4_t)(op7), (size_t)(op8)) +#define vloxseg6ei64_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei64_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vbool16_t)(op6), (vuint16m1_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (vuint16m1_t)(op11), (vuint16m1_t)(op12), (const uint16_t *)(op13), (vuint64m4_t)(op14), (size_t)(op15)) +#define vloxseg6ei64_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei64_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (const uint16_t *)(op6), (vuint64m2_t)(op7), (size_t)(op8)) +#define vloxseg6ei64_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei64_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vbool32_t)(op6), (vuint16mf2_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (vuint16mf2_t)(op11), (vuint16mf2_t)(op12), (const uint16_t *)(op13), (vuint64m2_t)(op14), (size_t)(op15)) +#define vloxseg6ei64_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei64_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (const uint16_t *)(op6), (vuint64m1_t)(op7), (size_t)(op8)) +#define vloxseg6ei64_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei64_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vbool64_t)(op6), (vuint16mf4_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (vuint16mf4_t)(op11), (vuint16mf4_t)(op12), (const uint16_t *)(op13), (vuint64m1_t)(op14), (size_t)(op15)) +#define vloxseg7ei64_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei64_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vint16m1_t *)(op6), (const int16_t *)(op7), (vuint64m4_t)(op8), (size_t)(op9)) +#define vloxseg7ei64_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei64_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vint16m1_t *)(op6), (vbool16_t)(op7), (vint16m1_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (vint16m1_t)(op11), (vint16m1_t)(op12), (vint16m1_t)(op13), (vint16m1_t)(op14), (const int16_t *)(op15), (vuint64m4_t)(op16), (size_t)(op17)) +#define vloxseg7ei64_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei64_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vint16mf2_t *)(op6), (const int16_t *)(op7), (vuint64m2_t)(op8), (size_t)(op9)) +#define vloxseg7ei64_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei64_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vint16mf2_t *)(op6), (vbool32_t)(op7), (vint16mf2_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (vint16mf2_t)(op11), (vint16mf2_t)(op12), (vint16mf2_t)(op13), (vint16mf2_t)(op14), (const int16_t *)(op15), (vuint64m2_t)(op16), (size_t)(op17)) +#define vloxseg7ei64_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei64_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vint16mf4_t *)(op6), (const int16_t *)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vloxseg7ei64_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei64_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vint16mf4_t *)(op6), (vbool64_t)(op7), (vint16mf4_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (vint16mf4_t)(op11), (vint16mf4_t)(op12), (vint16mf4_t)(op13), (vint16mf4_t)(op14), (const int16_t *)(op15), (vuint64m1_t)(op16), (size_t)(op17)) +#define vloxseg7ei64_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei64_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vuint16m1_t *)(op6), (const uint16_t *)(op7), (vuint64m4_t)(op8), (size_t)(op9)) +#define vloxseg7ei64_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei64_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vuint16m1_t *)(op6), (vbool16_t)(op7), (vuint16m1_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (vuint16m1_t)(op11), (vuint16m1_t)(op12), (vuint16m1_t)(op13), (vuint16m1_t)(op14), (const uint16_t *)(op15), (vuint64m4_t)(op16), (size_t)(op17)) +#define vloxseg7ei64_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei64_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vuint16mf2_t *)(op6), (const uint16_t *)(op7), (vuint64m2_t)(op8), (size_t)(op9)) +#define vloxseg7ei64_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei64_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vuint16mf2_t *)(op6), (vbool32_t)(op7), (vuint16mf2_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (vuint16mf2_t)(op11), (vuint16mf2_t)(op12), (vuint16mf2_t)(op13), (vuint16mf2_t)(op14), (const uint16_t *)(op15), (vuint64m2_t)(op16), (size_t)(op17)) +#define vloxseg7ei64_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei64_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vuint16mf4_t *)(op6), (const uint16_t *)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vloxseg7ei64_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei64_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vuint16mf4_t *)(op6), (vbool64_t)(op7), (vuint16mf4_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (vuint16mf4_t)(op11), (vuint16mf4_t)(op12), (vuint16mf4_t)(op13), (vuint16mf4_t)(op14), (const uint16_t *)(op15), (vuint64m1_t)(op16), (size_t)(op17)) +#define vloxseg8ei64_v_i16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei64_v_i16m1((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vint16m1_t *)(op6), (vint16m1_t *)(op7), (const int16_t *)(op8), (vuint64m4_t)(op9), (size_t)(op10)) +#define vloxseg8ei64_v_i16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei64_v_i16m1_m((vint16m1_t *)(op0), (vint16m1_t *)(op1), (vint16m1_t *)(op2), (vint16m1_t *)(op3), (vint16m1_t *)(op4), (vint16m1_t *)(op5), (vint16m1_t *)(op6), (vint16m1_t *)(op7), (vbool16_t)(op8), (vint16m1_t)(op9), (vint16m1_t)(op10), (vint16m1_t)(op11), (vint16m1_t)(op12), (vint16m1_t)(op13), (vint16m1_t)(op14), (vint16m1_t)(op15), (vint16m1_t)(op16), (const int16_t *)(op17), (vuint64m4_t)(op18), (size_t)(op19)) +#define vloxseg8ei64_v_i16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei64_v_i16mf2((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vint16mf2_t *)(op6), (vint16mf2_t *)(op7), (const int16_t *)(op8), (vuint64m2_t)(op9), (size_t)(op10)) +#define vloxseg8ei64_v_i16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei64_v_i16mf2_m((vint16mf2_t *)(op0), (vint16mf2_t *)(op1), (vint16mf2_t *)(op2), (vint16mf2_t *)(op3), (vint16mf2_t *)(op4), (vint16mf2_t *)(op5), (vint16mf2_t *)(op6), (vint16mf2_t *)(op7), (vbool32_t)(op8), (vint16mf2_t)(op9), (vint16mf2_t)(op10), (vint16mf2_t)(op11), (vint16mf2_t)(op12), (vint16mf2_t)(op13), (vint16mf2_t)(op14), (vint16mf2_t)(op15), (vint16mf2_t)(op16), (const int16_t *)(op17), (vuint64m2_t)(op18), (size_t)(op19)) +#define vloxseg8ei64_v_i16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei64_v_i16mf4((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vint16mf4_t *)(op6), (vint16mf4_t *)(op7), (const int16_t *)(op8), (vuint64m1_t)(op9), (size_t)(op10)) +#define vloxseg8ei64_v_i16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei64_v_i16mf4_m((vint16mf4_t *)(op0), (vint16mf4_t *)(op1), (vint16mf4_t *)(op2), (vint16mf4_t *)(op3), (vint16mf4_t *)(op4), (vint16mf4_t *)(op5), (vint16mf4_t *)(op6), (vint16mf4_t *)(op7), (vbool64_t)(op8), (vint16mf4_t)(op9), (vint16mf4_t)(op10), (vint16mf4_t)(op11), (vint16mf4_t)(op12), (vint16mf4_t)(op13), (vint16mf4_t)(op14), (vint16mf4_t)(op15), (vint16mf4_t)(op16), (const int16_t *)(op17), (vuint64m1_t)(op18), (size_t)(op19)) +#define vloxseg8ei64_v_u16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei64_v_u16m1((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vuint16m1_t *)(op6), (vuint16m1_t *)(op7), (const uint16_t *)(op8), (vuint64m4_t)(op9), (size_t)(op10)) +#define vloxseg8ei64_v_u16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei64_v_u16m1_m((vuint16m1_t *)(op0), (vuint16m1_t *)(op1), (vuint16m1_t *)(op2), (vuint16m1_t *)(op3), (vuint16m1_t *)(op4), (vuint16m1_t *)(op5), (vuint16m1_t *)(op6), (vuint16m1_t *)(op7), (vbool16_t)(op8), (vuint16m1_t)(op9), (vuint16m1_t)(op10), (vuint16m1_t)(op11), (vuint16m1_t)(op12), (vuint16m1_t)(op13), (vuint16m1_t)(op14), (vuint16m1_t)(op15), (vuint16m1_t)(op16), (const uint16_t *)(op17), (vuint64m4_t)(op18), (size_t)(op19)) +#define vloxseg8ei64_v_u16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei64_v_u16mf2((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vuint16mf2_t *)(op6), (vuint16mf2_t *)(op7), (const uint16_t *)(op8), (vuint64m2_t)(op9), (size_t)(op10)) +#define vloxseg8ei64_v_u16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei64_v_u16mf2_m((vuint16mf2_t *)(op0), (vuint16mf2_t *)(op1), (vuint16mf2_t *)(op2), (vuint16mf2_t *)(op3), (vuint16mf2_t *)(op4), (vuint16mf2_t *)(op5), (vuint16mf2_t *)(op6), (vuint16mf2_t *)(op7), (vbool32_t)(op8), (vuint16mf2_t)(op9), (vuint16mf2_t)(op10), (vuint16mf2_t)(op11), (vuint16mf2_t)(op12), (vuint16mf2_t)(op13), (vuint16mf2_t)(op14), (vuint16mf2_t)(op15), (vuint16mf2_t)(op16), (const uint16_t *)(op17), (vuint64m2_t)(op18), (size_t)(op19)) +#define vloxseg8ei64_v_u16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei64_v_u16mf4((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vuint16mf4_t *)(op6), (vuint16mf4_t *)(op7), (const uint16_t *)(op8), (vuint64m1_t)(op9), (size_t)(op10)) +#define vloxseg8ei64_v_u16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei64_v_u16mf4_m((vuint16mf4_t *)(op0), (vuint16mf4_t *)(op1), (vuint16mf4_t *)(op2), (vuint16mf4_t *)(op3), (vuint16mf4_t *)(op4), (vuint16mf4_t *)(op5), (vuint16mf4_t *)(op6), (vuint16mf4_t *)(op7), (vbool64_t)(op8), (vuint16mf4_t)(op9), (vuint16mf4_t)(op10), (vuint16mf4_t)(op11), (vuint16mf4_t)(op12), (vuint16mf4_t)(op13), (vuint16mf4_t)(op14), (vuint16mf4_t)(op15), (vuint16mf4_t)(op16), (const uint16_t *)(op17), (vuint64m1_t)(op18), (size_t)(op19)) +#define vloxseg2ei8_v_i32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (const int32_t *)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vbool32_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (const int32_t *)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vloxseg2ei8_v_i32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_i32m2((vint32m2_t *)(op0), (vint32m2_t *)(op1), (const int32_t *)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_i32m2_m((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vbool16_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (const int32_t *)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vloxseg2ei8_v_i32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_i32m4((vint32m4_t *)(op0), (vint32m4_t *)(op1), (const int32_t *)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_i32m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_i32m4_m((vint32m4_t *)(op0), (vint32m4_t *)(op1), (vbool8_t)(op2), (vint32m4_t)(op3), (vint32m4_t)(op4), (const int32_t *)(op5), (vuint8m1_t)(op6), (size_t)(op7)) +#define vloxseg2ei8_v_i32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (const int32_t *)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vbool64_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (const int32_t *)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vloxseg2ei8_v_u32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (const uint32_t *)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vbool32_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (const uint32_t *)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vloxseg2ei8_v_u32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_u32m2((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (const uint32_t *)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_u32m2_m((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vbool16_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (const uint32_t *)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vloxseg2ei8_v_u32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_u32m4((vuint32m4_t *)(op0), (vuint32m4_t *)(op1), (const uint32_t *)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_u32m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_u32m4_m((vuint32m4_t *)(op0), (vuint32m4_t *)(op1), (vbool8_t)(op2), (vuint32m4_t)(op3), (vuint32m4_t)(op4), (const uint32_t *)(op5), (vuint8m1_t)(op6), (size_t)(op7)) +#define vloxseg2ei8_v_u32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (const uint32_t *)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vbool64_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (const uint32_t *)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vloxseg3ei8_v_i32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei8_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (const int32_t *)(op3), (vuint8mf4_t)(op4), (size_t)(op5)) +#define vloxseg3ei8_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei8_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vbool32_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (const int32_t *)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vloxseg3ei8_v_i32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei8_v_i32m2((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vint32m2_t *)(op2), (const int32_t *)(op3), (vuint8mf2_t)(op4), (size_t)(op5)) +#define vloxseg3ei8_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei8_v_i32m2_m((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vint32m2_t *)(op2), (vbool16_t)(op3), (vint32m2_t)(op4), (vint32m2_t)(op5), (vint32m2_t)(op6), (const int32_t *)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vloxseg3ei8_v_i32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei8_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (const int32_t *)(op3), (vuint8mf8_t)(op4), (size_t)(op5)) +#define vloxseg3ei8_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei8_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vbool64_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (const int32_t *)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vloxseg3ei8_v_u32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei8_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (const uint32_t *)(op3), (vuint8mf4_t)(op4), (size_t)(op5)) +#define vloxseg3ei8_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei8_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vbool32_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (const uint32_t *)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vloxseg3ei8_v_u32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei8_v_u32m2((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vuint32m2_t *)(op2), (const uint32_t *)(op3), (vuint8mf2_t)(op4), (size_t)(op5)) +#define vloxseg3ei8_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei8_v_u32m2_m((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vuint32m2_t *)(op2), (vbool16_t)(op3), (vuint32m2_t)(op4), (vuint32m2_t)(op5), (vuint32m2_t)(op6), (const uint32_t *)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vloxseg3ei8_v_u32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei8_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (const uint32_t *)(op3), (vuint8mf8_t)(op4), (size_t)(op5)) +#define vloxseg3ei8_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei8_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vbool64_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (const uint32_t *)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vloxseg4ei8_v_i32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei8_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (const int32_t *)(op4), (vuint8mf4_t)(op5), (size_t)(op6)) +#define vloxseg4ei8_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei8_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vbool32_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (const int32_t *)(op9), (vuint8mf4_t)(op10), (size_t)(op11)) +#define vloxseg4ei8_v_i32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei8_v_i32m2((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vint32m2_t *)(op2), (vint32m2_t *)(op3), (const int32_t *)(op4), (vuint8mf2_t)(op5), (size_t)(op6)) +#define vloxseg4ei8_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei8_v_i32m2_m((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vint32m2_t *)(op2), (vint32m2_t *)(op3), (vbool16_t)(op4), (vint32m2_t)(op5), (vint32m2_t)(op6), (vint32m2_t)(op7), (vint32m2_t)(op8), (const int32_t *)(op9), (vuint8mf2_t)(op10), (size_t)(op11)) +#define vloxseg4ei8_v_i32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei8_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (const int32_t *)(op4), (vuint8mf8_t)(op5), (size_t)(op6)) +#define vloxseg4ei8_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei8_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vbool64_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (const int32_t *)(op9), (vuint8mf8_t)(op10), (size_t)(op11)) +#define vloxseg4ei8_v_u32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei8_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (const uint32_t *)(op4), (vuint8mf4_t)(op5), (size_t)(op6)) +#define vloxseg4ei8_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei8_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vbool32_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (const uint32_t *)(op9), (vuint8mf4_t)(op10), (size_t)(op11)) +#define vloxseg4ei8_v_u32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei8_v_u32m2((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vuint32m2_t *)(op2), (vuint32m2_t *)(op3), (const uint32_t *)(op4), (vuint8mf2_t)(op5), (size_t)(op6)) +#define vloxseg4ei8_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei8_v_u32m2_m((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vuint32m2_t *)(op2), (vuint32m2_t *)(op3), (vbool16_t)(op4), (vuint32m2_t)(op5), (vuint32m2_t)(op6), (vuint32m2_t)(op7), (vuint32m2_t)(op8), (const uint32_t *)(op9), (vuint8mf2_t)(op10), (size_t)(op11)) +#define vloxseg4ei8_v_u32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei8_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (const uint32_t *)(op4), (vuint8mf8_t)(op5), (size_t)(op6)) +#define vloxseg4ei8_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei8_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vbool64_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (const uint32_t *)(op9), (vuint8mf8_t)(op10), (size_t)(op11)) +#define vloxseg5ei8_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei8_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (const int32_t *)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vloxseg5ei8_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei8_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vbool32_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (const int32_t *)(op11), (vuint8mf4_t)(op12), (size_t)(op13)) +#define vloxseg5ei8_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei8_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (const int32_t *)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vloxseg5ei8_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei8_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vbool64_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (const int32_t *)(op11), (vuint8mf8_t)(op12), (size_t)(op13)) +#define vloxseg5ei8_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei8_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (const uint32_t *)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vloxseg5ei8_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei8_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vbool32_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (const uint32_t *)(op11), (vuint8mf4_t)(op12), (size_t)(op13)) +#define vloxseg5ei8_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei8_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (const uint32_t *)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vloxseg5ei8_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei8_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vbool64_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (const uint32_t *)(op11), (vuint8mf8_t)(op12), (size_t)(op13)) +#define vloxseg6ei8_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei8_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (const int32_t *)(op6), (vuint8mf4_t)(op7), (size_t)(op8)) +#define vloxseg6ei8_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei8_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vbool32_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (vint32m1_t)(op11), (vint32m1_t)(op12), (const int32_t *)(op13), (vuint8mf4_t)(op14), (size_t)(op15)) +#define vloxseg6ei8_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei8_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (const int32_t *)(op6), (vuint8mf8_t)(op7), (size_t)(op8)) +#define vloxseg6ei8_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei8_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vbool64_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (vint32mf2_t)(op11), (vint32mf2_t)(op12), (const int32_t *)(op13), (vuint8mf8_t)(op14), (size_t)(op15)) +#define vloxseg6ei8_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei8_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (const uint32_t *)(op6), (vuint8mf4_t)(op7), (size_t)(op8)) +#define vloxseg6ei8_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei8_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vbool32_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (vuint32m1_t)(op11), (vuint32m1_t)(op12), (const uint32_t *)(op13), (vuint8mf4_t)(op14), (size_t)(op15)) +#define vloxseg6ei8_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei8_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (const uint32_t *)(op6), (vuint8mf8_t)(op7), (size_t)(op8)) +#define vloxseg6ei8_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei8_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vbool64_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (vuint32mf2_t)(op11), (vuint32mf2_t)(op12), (const uint32_t *)(op13), (vuint8mf8_t)(op14), (size_t)(op15)) +#define vloxseg7ei8_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei8_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vint32m1_t *)(op6), (const int32_t *)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vloxseg7ei8_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei8_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vint32m1_t *)(op6), (vbool32_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (vint32m1_t)(op11), (vint32m1_t)(op12), (vint32m1_t)(op13), (vint32m1_t)(op14), (const int32_t *)(op15), (vuint8mf4_t)(op16), (size_t)(op17)) +#define vloxseg7ei8_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei8_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vint32mf2_t *)(op6), (const int32_t *)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vloxseg7ei8_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei8_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vint32mf2_t *)(op6), (vbool64_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (vint32mf2_t)(op11), (vint32mf2_t)(op12), (vint32mf2_t)(op13), (vint32mf2_t)(op14), (const int32_t *)(op15), (vuint8mf8_t)(op16), (size_t)(op17)) +#define vloxseg7ei8_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei8_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vuint32m1_t *)(op6), (const uint32_t *)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vloxseg7ei8_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei8_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vuint32m1_t *)(op6), (vbool32_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (vuint32m1_t)(op11), (vuint32m1_t)(op12), (vuint32m1_t)(op13), (vuint32m1_t)(op14), (const uint32_t *)(op15), (vuint8mf4_t)(op16), (size_t)(op17)) +#define vloxseg7ei8_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei8_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vuint32mf2_t *)(op6), (const uint32_t *)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vloxseg7ei8_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei8_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vuint32mf2_t *)(op6), (vbool64_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (vuint32mf2_t)(op11), (vuint32mf2_t)(op12), (vuint32mf2_t)(op13), (vuint32mf2_t)(op14), (const uint32_t *)(op15), (vuint8mf8_t)(op16), (size_t)(op17)) +#define vloxseg8ei8_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei8_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vint32m1_t *)(op6), (vint32m1_t *)(op7), (const int32_t *)(op8), (vuint8mf4_t)(op9), (size_t)(op10)) +#define vloxseg8ei8_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei8_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vint32m1_t *)(op6), (vint32m1_t *)(op7), (vbool32_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (vint32m1_t)(op11), (vint32m1_t)(op12), (vint32m1_t)(op13), (vint32m1_t)(op14), (vint32m1_t)(op15), (vint32m1_t)(op16), (const int32_t *)(op17), (vuint8mf4_t)(op18), (size_t)(op19)) +#define vloxseg8ei8_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei8_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vint32mf2_t *)(op6), (vint32mf2_t *)(op7), (const int32_t *)(op8), (vuint8mf8_t)(op9), (size_t)(op10)) +#define vloxseg8ei8_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei8_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vint32mf2_t *)(op6), (vint32mf2_t *)(op7), (vbool64_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (vint32mf2_t)(op11), (vint32mf2_t)(op12), (vint32mf2_t)(op13), (vint32mf2_t)(op14), (vint32mf2_t)(op15), (vint32mf2_t)(op16), (const int32_t *)(op17), (vuint8mf8_t)(op18), (size_t)(op19)) +#define vloxseg8ei8_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei8_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vuint32m1_t *)(op6), (vuint32m1_t *)(op7), (const uint32_t *)(op8), (vuint8mf4_t)(op9), (size_t)(op10)) +#define vloxseg8ei8_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei8_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vuint32m1_t *)(op6), (vuint32m1_t *)(op7), (vbool32_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (vuint32m1_t)(op11), (vuint32m1_t)(op12), (vuint32m1_t)(op13), (vuint32m1_t)(op14), (vuint32m1_t)(op15), (vuint32m1_t)(op16), (const uint32_t *)(op17), (vuint8mf4_t)(op18), (size_t)(op19)) +#define vloxseg8ei8_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei8_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vuint32mf2_t *)(op6), (vuint32mf2_t *)(op7), (const uint32_t *)(op8), (vuint8mf8_t)(op9), (size_t)(op10)) +#define vloxseg8ei8_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei8_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vuint32mf2_t *)(op6), (vuint32mf2_t *)(op7), (vbool64_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (vuint32mf2_t)(op11), (vuint32mf2_t)(op12), (vuint32mf2_t)(op13), (vuint32mf2_t)(op14), (vuint32mf2_t)(op15), (vuint32mf2_t)(op16), (const uint32_t *)(op17), (vuint8mf8_t)(op18), (size_t)(op19)) +#define vloxseg2ei16_v_i32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (const int32_t *)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vbool32_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (const int32_t *)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vloxseg2ei16_v_i32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_i32m2((vint32m2_t *)(op0), (vint32m2_t *)(op1), (const int32_t *)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_i32m2_m((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vbool16_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (const int32_t *)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vloxseg2ei16_v_i32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_i32m4((vint32m4_t *)(op0), (vint32m4_t *)(op1), (const int32_t *)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_i32m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_i32m4_m((vint32m4_t *)(op0), (vint32m4_t *)(op1), (vbool8_t)(op2), (vint32m4_t)(op3), (vint32m4_t)(op4), (const int32_t *)(op5), (vuint16m2_t)(op6), (size_t)(op7)) +#define vloxseg2ei16_v_i32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (const int32_t *)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vbool64_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (const int32_t *)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vloxseg2ei16_v_u32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (const uint32_t *)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vbool32_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (const uint32_t *)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vloxseg2ei16_v_u32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_u32m2((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (const uint32_t *)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_u32m2_m((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vbool16_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (const uint32_t *)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vloxseg2ei16_v_u32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_u32m4((vuint32m4_t *)(op0), (vuint32m4_t *)(op1), (const uint32_t *)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_u32m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_u32m4_m((vuint32m4_t *)(op0), (vuint32m4_t *)(op1), (vbool8_t)(op2), (vuint32m4_t)(op3), (vuint32m4_t)(op4), (const uint32_t *)(op5), (vuint16m2_t)(op6), (size_t)(op7)) +#define vloxseg2ei16_v_u32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (const uint32_t *)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vbool64_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (const uint32_t *)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vloxseg3ei16_v_i32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei16_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (const int32_t *)(op3), (vuint16mf2_t)(op4), (size_t)(op5)) +#define vloxseg3ei16_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei16_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vbool32_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (const int32_t *)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vloxseg3ei16_v_i32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei16_v_i32m2((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vint32m2_t *)(op2), (const int32_t *)(op3), (vuint16m1_t)(op4), (size_t)(op5)) +#define vloxseg3ei16_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei16_v_i32m2_m((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vint32m2_t *)(op2), (vbool16_t)(op3), (vint32m2_t)(op4), (vint32m2_t)(op5), (vint32m2_t)(op6), (const int32_t *)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vloxseg3ei16_v_i32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei16_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (const int32_t *)(op3), (vuint16mf4_t)(op4), (size_t)(op5)) +#define vloxseg3ei16_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei16_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vbool64_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (const int32_t *)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vloxseg3ei16_v_u32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei16_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (const uint32_t *)(op3), (vuint16mf2_t)(op4), (size_t)(op5)) +#define vloxseg3ei16_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei16_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vbool32_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (const uint32_t *)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vloxseg3ei16_v_u32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei16_v_u32m2((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vuint32m2_t *)(op2), (const uint32_t *)(op3), (vuint16m1_t)(op4), (size_t)(op5)) +#define vloxseg3ei16_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei16_v_u32m2_m((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vuint32m2_t *)(op2), (vbool16_t)(op3), (vuint32m2_t)(op4), (vuint32m2_t)(op5), (vuint32m2_t)(op6), (const uint32_t *)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vloxseg3ei16_v_u32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei16_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (const uint32_t *)(op3), (vuint16mf4_t)(op4), (size_t)(op5)) +#define vloxseg3ei16_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei16_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vbool64_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (const uint32_t *)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vloxseg4ei16_v_i32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei16_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (const int32_t *)(op4), (vuint16mf2_t)(op5), (size_t)(op6)) +#define vloxseg4ei16_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei16_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vbool32_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (const int32_t *)(op9), (vuint16mf2_t)(op10), (size_t)(op11)) +#define vloxseg4ei16_v_i32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei16_v_i32m2((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vint32m2_t *)(op2), (vint32m2_t *)(op3), (const int32_t *)(op4), (vuint16m1_t)(op5), (size_t)(op6)) +#define vloxseg4ei16_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei16_v_i32m2_m((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vint32m2_t *)(op2), (vint32m2_t *)(op3), (vbool16_t)(op4), (vint32m2_t)(op5), (vint32m2_t)(op6), (vint32m2_t)(op7), (vint32m2_t)(op8), (const int32_t *)(op9), (vuint16m1_t)(op10), (size_t)(op11)) +#define vloxseg4ei16_v_i32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei16_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (const int32_t *)(op4), (vuint16mf4_t)(op5), (size_t)(op6)) +#define vloxseg4ei16_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei16_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vbool64_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (const int32_t *)(op9), (vuint16mf4_t)(op10), (size_t)(op11)) +#define vloxseg4ei16_v_u32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei16_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (const uint32_t *)(op4), (vuint16mf2_t)(op5), (size_t)(op6)) +#define vloxseg4ei16_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei16_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vbool32_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (const uint32_t *)(op9), (vuint16mf2_t)(op10), (size_t)(op11)) +#define vloxseg4ei16_v_u32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei16_v_u32m2((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vuint32m2_t *)(op2), (vuint32m2_t *)(op3), (const uint32_t *)(op4), (vuint16m1_t)(op5), (size_t)(op6)) +#define vloxseg4ei16_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei16_v_u32m2_m((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vuint32m2_t *)(op2), (vuint32m2_t *)(op3), (vbool16_t)(op4), (vuint32m2_t)(op5), (vuint32m2_t)(op6), (vuint32m2_t)(op7), (vuint32m2_t)(op8), (const uint32_t *)(op9), (vuint16m1_t)(op10), (size_t)(op11)) +#define vloxseg4ei16_v_u32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei16_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (const uint32_t *)(op4), (vuint16mf4_t)(op5), (size_t)(op6)) +#define vloxseg4ei16_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei16_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vbool64_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (const uint32_t *)(op9), (vuint16mf4_t)(op10), (size_t)(op11)) +#define vloxseg5ei16_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei16_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (const int32_t *)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vloxseg5ei16_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei16_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vbool32_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (const int32_t *)(op11), (vuint16mf2_t)(op12), (size_t)(op13)) +#define vloxseg5ei16_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei16_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (const int32_t *)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vloxseg5ei16_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei16_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vbool64_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (const int32_t *)(op11), (vuint16mf4_t)(op12), (size_t)(op13)) +#define vloxseg5ei16_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei16_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (const uint32_t *)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vloxseg5ei16_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei16_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vbool32_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (const uint32_t *)(op11), (vuint16mf2_t)(op12), (size_t)(op13)) +#define vloxseg5ei16_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei16_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (const uint32_t *)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vloxseg5ei16_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei16_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vbool64_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (const uint32_t *)(op11), (vuint16mf4_t)(op12), (size_t)(op13)) +#define vloxseg6ei16_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei16_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (const int32_t *)(op6), (vuint16mf2_t)(op7), (size_t)(op8)) +#define vloxseg6ei16_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei16_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vbool32_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (vint32m1_t)(op11), (vint32m1_t)(op12), (const int32_t *)(op13), (vuint16mf2_t)(op14), (size_t)(op15)) +#define vloxseg6ei16_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei16_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (const int32_t *)(op6), (vuint16mf4_t)(op7), (size_t)(op8)) +#define vloxseg6ei16_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei16_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vbool64_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (vint32mf2_t)(op11), (vint32mf2_t)(op12), (const int32_t *)(op13), (vuint16mf4_t)(op14), (size_t)(op15)) +#define vloxseg6ei16_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei16_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (const uint32_t *)(op6), (vuint16mf2_t)(op7), (size_t)(op8)) +#define vloxseg6ei16_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei16_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vbool32_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (vuint32m1_t)(op11), (vuint32m1_t)(op12), (const uint32_t *)(op13), (vuint16mf2_t)(op14), (size_t)(op15)) +#define vloxseg6ei16_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei16_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (const uint32_t *)(op6), (vuint16mf4_t)(op7), (size_t)(op8)) +#define vloxseg6ei16_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei16_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vbool64_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (vuint32mf2_t)(op11), (vuint32mf2_t)(op12), (const uint32_t *)(op13), (vuint16mf4_t)(op14), (size_t)(op15)) +#define vloxseg7ei16_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei16_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vint32m1_t *)(op6), (const int32_t *)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vloxseg7ei16_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei16_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vint32m1_t *)(op6), (vbool32_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (vint32m1_t)(op11), (vint32m1_t)(op12), (vint32m1_t)(op13), (vint32m1_t)(op14), (const int32_t *)(op15), (vuint16mf2_t)(op16), (size_t)(op17)) +#define vloxseg7ei16_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei16_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vint32mf2_t *)(op6), (const int32_t *)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vloxseg7ei16_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei16_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vint32mf2_t *)(op6), (vbool64_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (vint32mf2_t)(op11), (vint32mf2_t)(op12), (vint32mf2_t)(op13), (vint32mf2_t)(op14), (const int32_t *)(op15), (vuint16mf4_t)(op16), (size_t)(op17)) +#define vloxseg7ei16_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei16_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vuint32m1_t *)(op6), (const uint32_t *)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vloxseg7ei16_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei16_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vuint32m1_t *)(op6), (vbool32_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (vuint32m1_t)(op11), (vuint32m1_t)(op12), (vuint32m1_t)(op13), (vuint32m1_t)(op14), (const uint32_t *)(op15), (vuint16mf2_t)(op16), (size_t)(op17)) +#define vloxseg7ei16_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei16_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vuint32mf2_t *)(op6), (const uint32_t *)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vloxseg7ei16_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei16_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vuint32mf2_t *)(op6), (vbool64_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (vuint32mf2_t)(op11), (vuint32mf2_t)(op12), (vuint32mf2_t)(op13), (vuint32mf2_t)(op14), (const uint32_t *)(op15), (vuint16mf4_t)(op16), (size_t)(op17)) +#define vloxseg8ei16_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei16_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vint32m1_t *)(op6), (vint32m1_t *)(op7), (const int32_t *)(op8), (vuint16mf2_t)(op9), (size_t)(op10)) +#define vloxseg8ei16_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei16_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vint32m1_t *)(op6), (vint32m1_t *)(op7), (vbool32_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (vint32m1_t)(op11), (vint32m1_t)(op12), (vint32m1_t)(op13), (vint32m1_t)(op14), (vint32m1_t)(op15), (vint32m1_t)(op16), (const int32_t *)(op17), (vuint16mf2_t)(op18), (size_t)(op19)) +#define vloxseg8ei16_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei16_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vint32mf2_t *)(op6), (vint32mf2_t *)(op7), (const int32_t *)(op8), (vuint16mf4_t)(op9), (size_t)(op10)) +#define vloxseg8ei16_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei16_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vint32mf2_t *)(op6), (vint32mf2_t *)(op7), (vbool64_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (vint32mf2_t)(op11), (vint32mf2_t)(op12), (vint32mf2_t)(op13), (vint32mf2_t)(op14), (vint32mf2_t)(op15), (vint32mf2_t)(op16), (const int32_t *)(op17), (vuint16mf4_t)(op18), (size_t)(op19)) +#define vloxseg8ei16_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei16_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vuint32m1_t *)(op6), (vuint32m1_t *)(op7), (const uint32_t *)(op8), (vuint16mf2_t)(op9), (size_t)(op10)) +#define vloxseg8ei16_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei16_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vuint32m1_t *)(op6), (vuint32m1_t *)(op7), (vbool32_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (vuint32m1_t)(op11), (vuint32m1_t)(op12), (vuint32m1_t)(op13), (vuint32m1_t)(op14), (vuint32m1_t)(op15), (vuint32m1_t)(op16), (const uint32_t *)(op17), (vuint16mf2_t)(op18), (size_t)(op19)) +#define vloxseg8ei16_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei16_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vuint32mf2_t *)(op6), (vuint32mf2_t *)(op7), (const uint32_t *)(op8), (vuint16mf4_t)(op9), (size_t)(op10)) +#define vloxseg8ei16_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei16_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vuint32mf2_t *)(op6), (vuint32mf2_t *)(op7), (vbool64_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (vuint32mf2_t)(op11), (vuint32mf2_t)(op12), (vuint32mf2_t)(op13), (vuint32mf2_t)(op14), (vuint32mf2_t)(op15), (vuint32mf2_t)(op16), (const uint32_t *)(op17), (vuint16mf4_t)(op18), (size_t)(op19)) +#define vloxseg2ei32_v_i32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (const int32_t *)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vbool32_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (const int32_t *)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vloxseg2ei32_v_i32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_i32m2((vint32m2_t *)(op0), (vint32m2_t *)(op1), (const int32_t *)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_i32m2_m((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vbool16_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (const int32_t *)(op5), (vuint32m2_t)(op6), (size_t)(op7)) +#define vloxseg2ei32_v_i32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_i32m4((vint32m4_t *)(op0), (vint32m4_t *)(op1), (const int32_t *)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_i32m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_i32m4_m((vint32m4_t *)(op0), (vint32m4_t *)(op1), (vbool8_t)(op2), (vint32m4_t)(op3), (vint32m4_t)(op4), (const int32_t *)(op5), (vuint32m4_t)(op6), (size_t)(op7)) +#define vloxseg2ei32_v_i32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (const int32_t *)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vbool64_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (const int32_t *)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vloxseg2ei32_v_u32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (const uint32_t *)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vbool32_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (const uint32_t *)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vloxseg2ei32_v_u32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_u32m2((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (const uint32_t *)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_u32m2_m((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vbool16_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (const uint32_t *)(op5), (vuint32m2_t)(op6), (size_t)(op7)) +#define vloxseg2ei32_v_u32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_u32m4((vuint32m4_t *)(op0), (vuint32m4_t *)(op1), (const uint32_t *)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_u32m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_u32m4_m((vuint32m4_t *)(op0), (vuint32m4_t *)(op1), (vbool8_t)(op2), (vuint32m4_t)(op3), (vuint32m4_t)(op4), (const uint32_t *)(op5), (vuint32m4_t)(op6), (size_t)(op7)) +#define vloxseg2ei32_v_u32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (const uint32_t *)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vbool64_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (const uint32_t *)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vloxseg3ei32_v_i32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei32_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (const int32_t *)(op3), (vuint32m1_t)(op4), (size_t)(op5)) +#define vloxseg3ei32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei32_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vbool32_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (const int32_t *)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vloxseg3ei32_v_i32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei32_v_i32m2((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vint32m2_t *)(op2), (const int32_t *)(op3), (vuint32m2_t)(op4), (size_t)(op5)) +#define vloxseg3ei32_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei32_v_i32m2_m((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vint32m2_t *)(op2), (vbool16_t)(op3), (vint32m2_t)(op4), (vint32m2_t)(op5), (vint32m2_t)(op6), (const int32_t *)(op7), (vuint32m2_t)(op8), (size_t)(op9)) +#define vloxseg3ei32_v_i32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei32_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (const int32_t *)(op3), (vuint32mf2_t)(op4), (size_t)(op5)) +#define vloxseg3ei32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei32_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vbool64_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (const int32_t *)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vloxseg3ei32_v_u32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei32_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (const uint32_t *)(op3), (vuint32m1_t)(op4), (size_t)(op5)) +#define vloxseg3ei32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei32_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vbool32_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (const uint32_t *)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vloxseg3ei32_v_u32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei32_v_u32m2((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vuint32m2_t *)(op2), (const uint32_t *)(op3), (vuint32m2_t)(op4), (size_t)(op5)) +#define vloxseg3ei32_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei32_v_u32m2_m((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vuint32m2_t *)(op2), (vbool16_t)(op3), (vuint32m2_t)(op4), (vuint32m2_t)(op5), (vuint32m2_t)(op6), (const uint32_t *)(op7), (vuint32m2_t)(op8), (size_t)(op9)) +#define vloxseg3ei32_v_u32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei32_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (const uint32_t *)(op3), (vuint32mf2_t)(op4), (size_t)(op5)) +#define vloxseg3ei32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei32_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vbool64_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (const uint32_t *)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vloxseg4ei32_v_i32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei32_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (const int32_t *)(op4), (vuint32m1_t)(op5), (size_t)(op6)) +#define vloxseg4ei32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei32_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vbool32_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (const int32_t *)(op9), (vuint32m1_t)(op10), (size_t)(op11)) +#define vloxseg4ei32_v_i32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei32_v_i32m2((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vint32m2_t *)(op2), (vint32m2_t *)(op3), (const int32_t *)(op4), (vuint32m2_t)(op5), (size_t)(op6)) +#define vloxseg4ei32_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei32_v_i32m2_m((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vint32m2_t *)(op2), (vint32m2_t *)(op3), (vbool16_t)(op4), (vint32m2_t)(op5), (vint32m2_t)(op6), (vint32m2_t)(op7), (vint32m2_t)(op8), (const int32_t *)(op9), (vuint32m2_t)(op10), (size_t)(op11)) +#define vloxseg4ei32_v_i32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei32_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (const int32_t *)(op4), (vuint32mf2_t)(op5), (size_t)(op6)) +#define vloxseg4ei32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei32_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vbool64_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (const int32_t *)(op9), (vuint32mf2_t)(op10), (size_t)(op11)) +#define vloxseg4ei32_v_u32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei32_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (const uint32_t *)(op4), (vuint32m1_t)(op5), (size_t)(op6)) +#define vloxseg4ei32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei32_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vbool32_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (const uint32_t *)(op9), (vuint32m1_t)(op10), (size_t)(op11)) +#define vloxseg4ei32_v_u32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei32_v_u32m2((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vuint32m2_t *)(op2), (vuint32m2_t *)(op3), (const uint32_t *)(op4), (vuint32m2_t)(op5), (size_t)(op6)) +#define vloxseg4ei32_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei32_v_u32m2_m((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vuint32m2_t *)(op2), (vuint32m2_t *)(op3), (vbool16_t)(op4), (vuint32m2_t)(op5), (vuint32m2_t)(op6), (vuint32m2_t)(op7), (vuint32m2_t)(op8), (const uint32_t *)(op9), (vuint32m2_t)(op10), (size_t)(op11)) +#define vloxseg4ei32_v_u32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei32_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (const uint32_t *)(op4), (vuint32mf2_t)(op5), (size_t)(op6)) +#define vloxseg4ei32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei32_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vbool64_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (const uint32_t *)(op9), (vuint32mf2_t)(op10), (size_t)(op11)) +#define vloxseg5ei32_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei32_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (const int32_t *)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vloxseg5ei32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei32_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vbool32_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (const int32_t *)(op11), (vuint32m1_t)(op12), (size_t)(op13)) +#define vloxseg5ei32_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei32_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (const int32_t *)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vloxseg5ei32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei32_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vbool64_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (const int32_t *)(op11), (vuint32mf2_t)(op12), (size_t)(op13)) +#define vloxseg5ei32_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei32_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (const uint32_t *)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vloxseg5ei32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei32_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vbool32_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (const uint32_t *)(op11), (vuint32m1_t)(op12), (size_t)(op13)) +#define vloxseg5ei32_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei32_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (const uint32_t *)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vloxseg5ei32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei32_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vbool64_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (const uint32_t *)(op11), (vuint32mf2_t)(op12), (size_t)(op13)) +#define vloxseg6ei32_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei32_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (const int32_t *)(op6), (vuint32m1_t)(op7), (size_t)(op8)) +#define vloxseg6ei32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei32_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vbool32_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (vint32m1_t)(op11), (vint32m1_t)(op12), (const int32_t *)(op13), (vuint32m1_t)(op14), (size_t)(op15)) +#define vloxseg6ei32_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei32_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (const int32_t *)(op6), (vuint32mf2_t)(op7), (size_t)(op8)) +#define vloxseg6ei32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei32_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vbool64_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (vint32mf2_t)(op11), (vint32mf2_t)(op12), (const int32_t *)(op13), (vuint32mf2_t)(op14), (size_t)(op15)) +#define vloxseg6ei32_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei32_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (const uint32_t *)(op6), (vuint32m1_t)(op7), (size_t)(op8)) +#define vloxseg6ei32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei32_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vbool32_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (vuint32m1_t)(op11), (vuint32m1_t)(op12), (const uint32_t *)(op13), (vuint32m1_t)(op14), (size_t)(op15)) +#define vloxseg6ei32_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei32_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (const uint32_t *)(op6), (vuint32mf2_t)(op7), (size_t)(op8)) +#define vloxseg6ei32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei32_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vbool64_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (vuint32mf2_t)(op11), (vuint32mf2_t)(op12), (const uint32_t *)(op13), (vuint32mf2_t)(op14), (size_t)(op15)) +#define vloxseg7ei32_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei32_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vint32m1_t *)(op6), (const int32_t *)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vloxseg7ei32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei32_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vint32m1_t *)(op6), (vbool32_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (vint32m1_t)(op11), (vint32m1_t)(op12), (vint32m1_t)(op13), (vint32m1_t)(op14), (const int32_t *)(op15), (vuint32m1_t)(op16), (size_t)(op17)) +#define vloxseg7ei32_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei32_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vint32mf2_t *)(op6), (const int32_t *)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vloxseg7ei32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei32_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vint32mf2_t *)(op6), (vbool64_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (vint32mf2_t)(op11), (vint32mf2_t)(op12), (vint32mf2_t)(op13), (vint32mf2_t)(op14), (const int32_t *)(op15), (vuint32mf2_t)(op16), (size_t)(op17)) +#define vloxseg7ei32_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei32_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vuint32m1_t *)(op6), (const uint32_t *)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vloxseg7ei32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei32_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vuint32m1_t *)(op6), (vbool32_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (vuint32m1_t)(op11), (vuint32m1_t)(op12), (vuint32m1_t)(op13), (vuint32m1_t)(op14), (const uint32_t *)(op15), (vuint32m1_t)(op16), (size_t)(op17)) +#define vloxseg7ei32_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei32_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vuint32mf2_t *)(op6), (const uint32_t *)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vloxseg7ei32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei32_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vuint32mf2_t *)(op6), (vbool64_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (vuint32mf2_t)(op11), (vuint32mf2_t)(op12), (vuint32mf2_t)(op13), (vuint32mf2_t)(op14), (const uint32_t *)(op15), (vuint32mf2_t)(op16), (size_t)(op17)) +#define vloxseg8ei32_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei32_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vint32m1_t *)(op6), (vint32m1_t *)(op7), (const int32_t *)(op8), (vuint32m1_t)(op9), (size_t)(op10)) +#define vloxseg8ei32_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei32_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vint32m1_t *)(op6), (vint32m1_t *)(op7), (vbool32_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (vint32m1_t)(op11), (vint32m1_t)(op12), (vint32m1_t)(op13), (vint32m1_t)(op14), (vint32m1_t)(op15), (vint32m1_t)(op16), (const int32_t *)(op17), (vuint32m1_t)(op18), (size_t)(op19)) +#define vloxseg8ei32_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei32_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vint32mf2_t *)(op6), (vint32mf2_t *)(op7), (const int32_t *)(op8), (vuint32mf2_t)(op9), (size_t)(op10)) +#define vloxseg8ei32_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei32_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vint32mf2_t *)(op6), (vint32mf2_t *)(op7), (vbool64_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (vint32mf2_t)(op11), (vint32mf2_t)(op12), (vint32mf2_t)(op13), (vint32mf2_t)(op14), (vint32mf2_t)(op15), (vint32mf2_t)(op16), (const int32_t *)(op17), (vuint32mf2_t)(op18), (size_t)(op19)) +#define vloxseg8ei32_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei32_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vuint32m1_t *)(op6), (vuint32m1_t *)(op7), (const uint32_t *)(op8), (vuint32m1_t)(op9), (size_t)(op10)) +#define vloxseg8ei32_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei32_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vuint32m1_t *)(op6), (vuint32m1_t *)(op7), (vbool32_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (vuint32m1_t)(op11), (vuint32m1_t)(op12), (vuint32m1_t)(op13), (vuint32m1_t)(op14), (vuint32m1_t)(op15), (vuint32m1_t)(op16), (const uint32_t *)(op17), (vuint32m1_t)(op18), (size_t)(op19)) +#define vloxseg8ei32_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei32_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vuint32mf2_t *)(op6), (vuint32mf2_t *)(op7), (const uint32_t *)(op8), (vuint32mf2_t)(op9), (size_t)(op10)) +#define vloxseg8ei32_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei32_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vuint32mf2_t *)(op6), (vuint32mf2_t *)(op7), (vbool64_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (vuint32mf2_t)(op11), (vuint32mf2_t)(op12), (vuint32mf2_t)(op13), (vuint32mf2_t)(op14), (vuint32mf2_t)(op15), (vuint32mf2_t)(op16), (const uint32_t *)(op17), (vuint32mf2_t)(op18), (size_t)(op19)) +#define vloxseg2ei64_v_i32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei64_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (const int32_t *)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vloxseg2ei64_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei64_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vbool32_t)(op2), (vint32m1_t)(op3), (vint32m1_t)(op4), (const int32_t *)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vloxseg2ei64_v_i32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei64_v_i32m2((vint32m2_t *)(op0), (vint32m2_t *)(op1), (const int32_t *)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vloxseg2ei64_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei64_v_i32m2_m((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vbool16_t)(op2), (vint32m2_t)(op3), (vint32m2_t)(op4), (const int32_t *)(op5), (vuint64m4_t)(op6), (size_t)(op7)) +#define vloxseg2ei64_v_i32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei64_v_i32m4((vint32m4_t *)(op0), (vint32m4_t *)(op1), (const int32_t *)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vloxseg2ei64_v_i32m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei64_v_i32m4_m((vint32m4_t *)(op0), (vint32m4_t *)(op1), (vbool8_t)(op2), (vint32m4_t)(op3), (vint32m4_t)(op4), (const int32_t *)(op5), (vuint64m8_t)(op6), (size_t)(op7)) +#define vloxseg2ei64_v_i32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei64_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (const int32_t *)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vloxseg2ei64_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei64_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vbool64_t)(op2), (vint32mf2_t)(op3), (vint32mf2_t)(op4), (const int32_t *)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vloxseg2ei64_v_u32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei64_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (const uint32_t *)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vloxseg2ei64_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei64_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vbool32_t)(op2), (vuint32m1_t)(op3), (vuint32m1_t)(op4), (const uint32_t *)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vloxseg2ei64_v_u32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei64_v_u32m2((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (const uint32_t *)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vloxseg2ei64_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei64_v_u32m2_m((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vbool16_t)(op2), (vuint32m2_t)(op3), (vuint32m2_t)(op4), (const uint32_t *)(op5), (vuint64m4_t)(op6), (size_t)(op7)) +#define vloxseg2ei64_v_u32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei64_v_u32m4((vuint32m4_t *)(op0), (vuint32m4_t *)(op1), (const uint32_t *)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vloxseg2ei64_v_u32m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei64_v_u32m4_m((vuint32m4_t *)(op0), (vuint32m4_t *)(op1), (vbool8_t)(op2), (vuint32m4_t)(op3), (vuint32m4_t)(op4), (const uint32_t *)(op5), (vuint64m8_t)(op6), (size_t)(op7)) +#define vloxseg2ei64_v_u32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei64_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (const uint32_t *)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vloxseg2ei64_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei64_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vbool64_t)(op2), (vuint32mf2_t)(op3), (vuint32mf2_t)(op4), (const uint32_t *)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vloxseg3ei64_v_i32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei64_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (const int32_t *)(op3), (vuint64m2_t)(op4), (size_t)(op5)) +#define vloxseg3ei64_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei64_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vbool32_t)(op3), (vint32m1_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (const int32_t *)(op7), (vuint64m2_t)(op8), (size_t)(op9)) +#define vloxseg3ei64_v_i32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei64_v_i32m2((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vint32m2_t *)(op2), (const int32_t *)(op3), (vuint64m4_t)(op4), (size_t)(op5)) +#define vloxseg3ei64_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei64_v_i32m2_m((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vint32m2_t *)(op2), (vbool16_t)(op3), (vint32m2_t)(op4), (vint32m2_t)(op5), (vint32m2_t)(op6), (const int32_t *)(op7), (vuint64m4_t)(op8), (size_t)(op9)) +#define vloxseg3ei64_v_i32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei64_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (const int32_t *)(op3), (vuint64m1_t)(op4), (size_t)(op5)) +#define vloxseg3ei64_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei64_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vbool64_t)(op3), (vint32mf2_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (const int32_t *)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vloxseg3ei64_v_u32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei64_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (const uint32_t *)(op3), (vuint64m2_t)(op4), (size_t)(op5)) +#define vloxseg3ei64_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei64_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vbool32_t)(op3), (vuint32m1_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (const uint32_t *)(op7), (vuint64m2_t)(op8), (size_t)(op9)) +#define vloxseg3ei64_v_u32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei64_v_u32m2((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vuint32m2_t *)(op2), (const uint32_t *)(op3), (vuint64m4_t)(op4), (size_t)(op5)) +#define vloxseg3ei64_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei64_v_u32m2_m((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vuint32m2_t *)(op2), (vbool16_t)(op3), (vuint32m2_t)(op4), (vuint32m2_t)(op5), (vuint32m2_t)(op6), (const uint32_t *)(op7), (vuint64m4_t)(op8), (size_t)(op9)) +#define vloxseg3ei64_v_u32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei64_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (const uint32_t *)(op3), (vuint64m1_t)(op4), (size_t)(op5)) +#define vloxseg3ei64_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei64_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vbool64_t)(op3), (vuint32mf2_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (const uint32_t *)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vloxseg4ei64_v_i32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei64_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (const int32_t *)(op4), (vuint64m2_t)(op5), (size_t)(op6)) +#define vloxseg4ei64_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei64_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vbool32_t)(op4), (vint32m1_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (const int32_t *)(op9), (vuint64m2_t)(op10), (size_t)(op11)) +#define vloxseg4ei64_v_i32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei64_v_i32m2((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vint32m2_t *)(op2), (vint32m2_t *)(op3), (const int32_t *)(op4), (vuint64m4_t)(op5), (size_t)(op6)) +#define vloxseg4ei64_v_i32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei64_v_i32m2_m((vint32m2_t *)(op0), (vint32m2_t *)(op1), (vint32m2_t *)(op2), (vint32m2_t *)(op3), (vbool16_t)(op4), (vint32m2_t)(op5), (vint32m2_t)(op6), (vint32m2_t)(op7), (vint32m2_t)(op8), (const int32_t *)(op9), (vuint64m4_t)(op10), (size_t)(op11)) +#define vloxseg4ei64_v_i32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei64_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (const int32_t *)(op4), (vuint64m1_t)(op5), (size_t)(op6)) +#define vloxseg4ei64_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei64_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vbool64_t)(op4), (vint32mf2_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (const int32_t *)(op9), (vuint64m1_t)(op10), (size_t)(op11)) +#define vloxseg4ei64_v_u32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei64_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (const uint32_t *)(op4), (vuint64m2_t)(op5), (size_t)(op6)) +#define vloxseg4ei64_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei64_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vbool32_t)(op4), (vuint32m1_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (const uint32_t *)(op9), (vuint64m2_t)(op10), (size_t)(op11)) +#define vloxseg4ei64_v_u32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei64_v_u32m2((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vuint32m2_t *)(op2), (vuint32m2_t *)(op3), (const uint32_t *)(op4), (vuint64m4_t)(op5), (size_t)(op6)) +#define vloxseg4ei64_v_u32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei64_v_u32m2_m((vuint32m2_t *)(op0), (vuint32m2_t *)(op1), (vuint32m2_t *)(op2), (vuint32m2_t *)(op3), (vbool16_t)(op4), (vuint32m2_t)(op5), (vuint32m2_t)(op6), (vuint32m2_t)(op7), (vuint32m2_t)(op8), (const uint32_t *)(op9), (vuint64m4_t)(op10), (size_t)(op11)) +#define vloxseg4ei64_v_u32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei64_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (const uint32_t *)(op4), (vuint64m1_t)(op5), (size_t)(op6)) +#define vloxseg4ei64_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei64_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vbool64_t)(op4), (vuint32mf2_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (const uint32_t *)(op9), (vuint64m1_t)(op10), (size_t)(op11)) +#define vloxseg5ei64_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei64_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (const int32_t *)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vloxseg5ei64_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei64_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vbool32_t)(op5), (vint32m1_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (const int32_t *)(op11), (vuint64m2_t)(op12), (size_t)(op13)) +#define vloxseg5ei64_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei64_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (const int32_t *)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vloxseg5ei64_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei64_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vbool64_t)(op5), (vint32mf2_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (const int32_t *)(op11), (vuint64m1_t)(op12), (size_t)(op13)) +#define vloxseg5ei64_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei64_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (const uint32_t *)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vloxseg5ei64_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei64_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vbool32_t)(op5), (vuint32m1_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (const uint32_t *)(op11), (vuint64m2_t)(op12), (size_t)(op13)) +#define vloxseg5ei64_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei64_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (const uint32_t *)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vloxseg5ei64_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei64_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vbool64_t)(op5), (vuint32mf2_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (const uint32_t *)(op11), (vuint64m1_t)(op12), (size_t)(op13)) +#define vloxseg6ei64_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei64_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (const int32_t *)(op6), (vuint64m2_t)(op7), (size_t)(op8)) +#define vloxseg6ei64_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei64_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vbool32_t)(op6), (vint32m1_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (vint32m1_t)(op11), (vint32m1_t)(op12), (const int32_t *)(op13), (vuint64m2_t)(op14), (size_t)(op15)) +#define vloxseg6ei64_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei64_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (const int32_t *)(op6), (vuint64m1_t)(op7), (size_t)(op8)) +#define vloxseg6ei64_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei64_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vbool64_t)(op6), (vint32mf2_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (vint32mf2_t)(op11), (vint32mf2_t)(op12), (const int32_t *)(op13), (vuint64m1_t)(op14), (size_t)(op15)) +#define vloxseg6ei64_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei64_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (const uint32_t *)(op6), (vuint64m2_t)(op7), (size_t)(op8)) +#define vloxseg6ei64_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei64_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vbool32_t)(op6), (vuint32m1_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (vuint32m1_t)(op11), (vuint32m1_t)(op12), (const uint32_t *)(op13), (vuint64m2_t)(op14), (size_t)(op15)) +#define vloxseg6ei64_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei64_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (const uint32_t *)(op6), (vuint64m1_t)(op7), (size_t)(op8)) +#define vloxseg6ei64_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei64_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vbool64_t)(op6), (vuint32mf2_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (vuint32mf2_t)(op11), (vuint32mf2_t)(op12), (const uint32_t *)(op13), (vuint64m1_t)(op14), (size_t)(op15)) +#define vloxseg7ei64_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei64_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vint32m1_t *)(op6), (const int32_t *)(op7), (vuint64m2_t)(op8), (size_t)(op9)) +#define vloxseg7ei64_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei64_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vint32m1_t *)(op6), (vbool32_t)(op7), (vint32m1_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (vint32m1_t)(op11), (vint32m1_t)(op12), (vint32m1_t)(op13), (vint32m1_t)(op14), (const int32_t *)(op15), (vuint64m2_t)(op16), (size_t)(op17)) +#define vloxseg7ei64_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei64_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vint32mf2_t *)(op6), (const int32_t *)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vloxseg7ei64_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei64_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vint32mf2_t *)(op6), (vbool64_t)(op7), (vint32mf2_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (vint32mf2_t)(op11), (vint32mf2_t)(op12), (vint32mf2_t)(op13), (vint32mf2_t)(op14), (const int32_t *)(op15), (vuint64m1_t)(op16), (size_t)(op17)) +#define vloxseg7ei64_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei64_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vuint32m1_t *)(op6), (const uint32_t *)(op7), (vuint64m2_t)(op8), (size_t)(op9)) +#define vloxseg7ei64_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei64_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vuint32m1_t *)(op6), (vbool32_t)(op7), (vuint32m1_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (vuint32m1_t)(op11), (vuint32m1_t)(op12), (vuint32m1_t)(op13), (vuint32m1_t)(op14), (const uint32_t *)(op15), (vuint64m2_t)(op16), (size_t)(op17)) +#define vloxseg7ei64_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei64_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vuint32mf2_t *)(op6), (const uint32_t *)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vloxseg7ei64_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei64_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vuint32mf2_t *)(op6), (vbool64_t)(op7), (vuint32mf2_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (vuint32mf2_t)(op11), (vuint32mf2_t)(op12), (vuint32mf2_t)(op13), (vuint32mf2_t)(op14), (const uint32_t *)(op15), (vuint64m1_t)(op16), (size_t)(op17)) +#define vloxseg8ei64_v_i32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei64_v_i32m1((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vint32m1_t *)(op6), (vint32m1_t *)(op7), (const int32_t *)(op8), (vuint64m2_t)(op9), (size_t)(op10)) +#define vloxseg8ei64_v_i32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei64_v_i32m1_m((vint32m1_t *)(op0), (vint32m1_t *)(op1), (vint32m1_t *)(op2), (vint32m1_t *)(op3), (vint32m1_t *)(op4), (vint32m1_t *)(op5), (vint32m1_t *)(op6), (vint32m1_t *)(op7), (vbool32_t)(op8), (vint32m1_t)(op9), (vint32m1_t)(op10), (vint32m1_t)(op11), (vint32m1_t)(op12), (vint32m1_t)(op13), (vint32m1_t)(op14), (vint32m1_t)(op15), (vint32m1_t)(op16), (const int32_t *)(op17), (vuint64m2_t)(op18), (size_t)(op19)) +#define vloxseg8ei64_v_i32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei64_v_i32mf2((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vint32mf2_t *)(op6), (vint32mf2_t *)(op7), (const int32_t *)(op8), (vuint64m1_t)(op9), (size_t)(op10)) +#define vloxseg8ei64_v_i32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei64_v_i32mf2_m((vint32mf2_t *)(op0), (vint32mf2_t *)(op1), (vint32mf2_t *)(op2), (vint32mf2_t *)(op3), (vint32mf2_t *)(op4), (vint32mf2_t *)(op5), (vint32mf2_t *)(op6), (vint32mf2_t *)(op7), (vbool64_t)(op8), (vint32mf2_t)(op9), (vint32mf2_t)(op10), (vint32mf2_t)(op11), (vint32mf2_t)(op12), (vint32mf2_t)(op13), (vint32mf2_t)(op14), (vint32mf2_t)(op15), (vint32mf2_t)(op16), (const int32_t *)(op17), (vuint64m1_t)(op18), (size_t)(op19)) +#define vloxseg8ei64_v_u32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei64_v_u32m1((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vuint32m1_t *)(op6), (vuint32m1_t *)(op7), (const uint32_t *)(op8), (vuint64m2_t)(op9), (size_t)(op10)) +#define vloxseg8ei64_v_u32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei64_v_u32m1_m((vuint32m1_t *)(op0), (vuint32m1_t *)(op1), (vuint32m1_t *)(op2), (vuint32m1_t *)(op3), (vuint32m1_t *)(op4), (vuint32m1_t *)(op5), (vuint32m1_t *)(op6), (vuint32m1_t *)(op7), (vbool32_t)(op8), (vuint32m1_t)(op9), (vuint32m1_t)(op10), (vuint32m1_t)(op11), (vuint32m1_t)(op12), (vuint32m1_t)(op13), (vuint32m1_t)(op14), (vuint32m1_t)(op15), (vuint32m1_t)(op16), (const uint32_t *)(op17), (vuint64m2_t)(op18), (size_t)(op19)) +#define vloxseg8ei64_v_u32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei64_v_u32mf2((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vuint32mf2_t *)(op6), (vuint32mf2_t *)(op7), (const uint32_t *)(op8), (vuint64m1_t)(op9), (size_t)(op10)) +#define vloxseg8ei64_v_u32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei64_v_u32mf2_m((vuint32mf2_t *)(op0), (vuint32mf2_t *)(op1), (vuint32mf2_t *)(op2), (vuint32mf2_t *)(op3), (vuint32mf2_t *)(op4), (vuint32mf2_t *)(op5), (vuint32mf2_t *)(op6), (vuint32mf2_t *)(op7), (vbool64_t)(op8), (vuint32mf2_t)(op9), (vuint32mf2_t)(op10), (vuint32mf2_t)(op11), (vuint32mf2_t)(op12), (vuint32mf2_t)(op13), (vuint32mf2_t)(op14), (vuint32mf2_t)(op15), (vuint32mf2_t)(op16), (const uint32_t *)(op17), (vuint64m1_t)(op18), (size_t)(op19)) +#endif + +#if defined(__riscv_f) && defined(__riscv_zvlsseg) +#define vloxseg2ei8_v_f32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (const float *)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vbool32_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (const float *)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vloxseg2ei8_v_f32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_f32m2((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (const float *)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_f32m2_m((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vbool16_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (const float *)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vloxseg2ei8_v_f32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_f32m4((vfloat32m4_t *)(op0), (vfloat32m4_t *)(op1), (const float *)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_f32m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_f32m4_m((vfloat32m4_t *)(op0), (vfloat32m4_t *)(op1), (vbool8_t)(op2), (vfloat32m4_t)(op3), (vfloat32m4_t)(op4), (const float *)(op5), (vuint8m1_t)(op6), (size_t)(op7)) +#define vloxseg2ei8_v_f32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (const float *)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vbool64_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (const float *)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vloxseg3ei8_v_f32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei8_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (const float *)(op3), (vuint8mf4_t)(op4), (size_t)(op5)) +#define vloxseg3ei8_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei8_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vbool32_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (const float *)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vloxseg3ei8_v_f32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei8_v_f32m2((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vfloat32m2_t *)(op2), (const float *)(op3), (vuint8mf2_t)(op4), (size_t)(op5)) +#define vloxseg3ei8_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei8_v_f32m2_m((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vfloat32m2_t *)(op2), (vbool16_t)(op3), (vfloat32m2_t)(op4), (vfloat32m2_t)(op5), (vfloat32m2_t)(op6), (const float *)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vloxseg3ei8_v_f32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei8_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (const float *)(op3), (vuint8mf8_t)(op4), (size_t)(op5)) +#define vloxseg3ei8_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei8_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vbool64_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (const float *)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vloxseg4ei8_v_f32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei8_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (const float *)(op4), (vuint8mf4_t)(op5), (size_t)(op6)) +#define vloxseg4ei8_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei8_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vbool32_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (const float *)(op9), (vuint8mf4_t)(op10), (size_t)(op11)) +#define vloxseg4ei8_v_f32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei8_v_f32m2((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vfloat32m2_t *)(op2), (vfloat32m2_t *)(op3), (const float *)(op4), (vuint8mf2_t)(op5), (size_t)(op6)) +#define vloxseg4ei8_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei8_v_f32m2_m((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vfloat32m2_t *)(op2), (vfloat32m2_t *)(op3), (vbool16_t)(op4), (vfloat32m2_t)(op5), (vfloat32m2_t)(op6), (vfloat32m2_t)(op7), (vfloat32m2_t)(op8), (const float *)(op9), (vuint8mf2_t)(op10), (size_t)(op11)) +#define vloxseg4ei8_v_f32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei8_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (const float *)(op4), (vuint8mf8_t)(op5), (size_t)(op6)) +#define vloxseg4ei8_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei8_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vbool64_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (const float *)(op9), (vuint8mf8_t)(op10), (size_t)(op11)) +#define vloxseg5ei8_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei8_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (const float *)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vloxseg5ei8_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei8_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vbool32_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (const float *)(op11), (vuint8mf4_t)(op12), (size_t)(op13)) +#define vloxseg5ei8_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei8_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (const float *)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vloxseg5ei8_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei8_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vbool64_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (const float *)(op11), (vuint8mf8_t)(op12), (size_t)(op13)) +#define vloxseg6ei8_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei8_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (const float *)(op6), (vuint8mf4_t)(op7), (size_t)(op8)) +#define vloxseg6ei8_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei8_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vbool32_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (vfloat32m1_t)(op11), (vfloat32m1_t)(op12), (const float *)(op13), (vuint8mf4_t)(op14), (size_t)(op15)) +#define vloxseg6ei8_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei8_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (const float *)(op6), (vuint8mf8_t)(op7), (size_t)(op8)) +#define vloxseg6ei8_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei8_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vbool64_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (vfloat32mf2_t)(op11), (vfloat32mf2_t)(op12), (const float *)(op13), (vuint8mf8_t)(op14), (size_t)(op15)) +#define vloxseg7ei8_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei8_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vfloat32m1_t *)(op6), (const float *)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vloxseg7ei8_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei8_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vfloat32m1_t *)(op6), (vbool32_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (vfloat32m1_t)(op11), (vfloat32m1_t)(op12), (vfloat32m1_t)(op13), (vfloat32m1_t)(op14), (const float *)(op15), (vuint8mf4_t)(op16), (size_t)(op17)) +#define vloxseg7ei8_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei8_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vfloat32mf2_t *)(op6), (const float *)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vloxseg7ei8_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei8_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vfloat32mf2_t *)(op6), (vbool64_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (vfloat32mf2_t)(op11), (vfloat32mf2_t)(op12), (vfloat32mf2_t)(op13), (vfloat32mf2_t)(op14), (const float *)(op15), (vuint8mf8_t)(op16), (size_t)(op17)) +#define vloxseg8ei8_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei8_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vfloat32m1_t *)(op6), (vfloat32m1_t *)(op7), (const float *)(op8), (vuint8mf4_t)(op9), (size_t)(op10)) +#define vloxseg8ei8_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei8_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vfloat32m1_t *)(op6), (vfloat32m1_t *)(op7), (vbool32_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (vfloat32m1_t)(op11), (vfloat32m1_t)(op12), (vfloat32m1_t)(op13), (vfloat32m1_t)(op14), (vfloat32m1_t)(op15), (vfloat32m1_t)(op16), (const float *)(op17), (vuint8mf4_t)(op18), (size_t)(op19)) +#define vloxseg8ei8_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei8_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vfloat32mf2_t *)(op6), (vfloat32mf2_t *)(op7), (const float *)(op8), (vuint8mf8_t)(op9), (size_t)(op10)) +#define vloxseg8ei8_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei8_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vfloat32mf2_t *)(op6), (vfloat32mf2_t *)(op7), (vbool64_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (vfloat32mf2_t)(op11), (vfloat32mf2_t)(op12), (vfloat32mf2_t)(op13), (vfloat32mf2_t)(op14), (vfloat32mf2_t)(op15), (vfloat32mf2_t)(op16), (const float *)(op17), (vuint8mf8_t)(op18), (size_t)(op19)) +#define vloxseg2ei16_v_f32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (const float *)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vbool32_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (const float *)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vloxseg2ei16_v_f32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_f32m2((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (const float *)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_f32m2_m((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vbool16_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (const float *)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vloxseg2ei16_v_f32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_f32m4((vfloat32m4_t *)(op0), (vfloat32m4_t *)(op1), (const float *)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_f32m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_f32m4_m((vfloat32m4_t *)(op0), (vfloat32m4_t *)(op1), (vbool8_t)(op2), (vfloat32m4_t)(op3), (vfloat32m4_t)(op4), (const float *)(op5), (vuint16m2_t)(op6), (size_t)(op7)) +#define vloxseg2ei16_v_f32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (const float *)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vbool64_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (const float *)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vloxseg3ei16_v_f32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei16_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (const float *)(op3), (vuint16mf2_t)(op4), (size_t)(op5)) +#define vloxseg3ei16_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei16_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vbool32_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (const float *)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vloxseg3ei16_v_f32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei16_v_f32m2((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vfloat32m2_t *)(op2), (const float *)(op3), (vuint16m1_t)(op4), (size_t)(op5)) +#define vloxseg3ei16_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei16_v_f32m2_m((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vfloat32m2_t *)(op2), (vbool16_t)(op3), (vfloat32m2_t)(op4), (vfloat32m2_t)(op5), (vfloat32m2_t)(op6), (const float *)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vloxseg3ei16_v_f32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei16_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (const float *)(op3), (vuint16mf4_t)(op4), (size_t)(op5)) +#define vloxseg3ei16_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei16_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vbool64_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (const float *)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vloxseg4ei16_v_f32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei16_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (const float *)(op4), (vuint16mf2_t)(op5), (size_t)(op6)) +#define vloxseg4ei16_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei16_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vbool32_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (const float *)(op9), (vuint16mf2_t)(op10), (size_t)(op11)) +#define vloxseg4ei16_v_f32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei16_v_f32m2((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vfloat32m2_t *)(op2), (vfloat32m2_t *)(op3), (const float *)(op4), (vuint16m1_t)(op5), (size_t)(op6)) +#define vloxseg4ei16_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei16_v_f32m2_m((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vfloat32m2_t *)(op2), (vfloat32m2_t *)(op3), (vbool16_t)(op4), (vfloat32m2_t)(op5), (vfloat32m2_t)(op6), (vfloat32m2_t)(op7), (vfloat32m2_t)(op8), (const float *)(op9), (vuint16m1_t)(op10), (size_t)(op11)) +#define vloxseg4ei16_v_f32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei16_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (const float *)(op4), (vuint16mf4_t)(op5), (size_t)(op6)) +#define vloxseg4ei16_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei16_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vbool64_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (const float *)(op9), (vuint16mf4_t)(op10), (size_t)(op11)) +#define vloxseg5ei16_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei16_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (const float *)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vloxseg5ei16_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei16_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vbool32_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (const float *)(op11), (vuint16mf2_t)(op12), (size_t)(op13)) +#define vloxseg5ei16_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei16_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (const float *)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vloxseg5ei16_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei16_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vbool64_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (const float *)(op11), (vuint16mf4_t)(op12), (size_t)(op13)) +#define vloxseg6ei16_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei16_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (const float *)(op6), (vuint16mf2_t)(op7), (size_t)(op8)) +#define vloxseg6ei16_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei16_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vbool32_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (vfloat32m1_t)(op11), (vfloat32m1_t)(op12), (const float *)(op13), (vuint16mf2_t)(op14), (size_t)(op15)) +#define vloxseg6ei16_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei16_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (const float *)(op6), (vuint16mf4_t)(op7), (size_t)(op8)) +#define vloxseg6ei16_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei16_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vbool64_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (vfloat32mf2_t)(op11), (vfloat32mf2_t)(op12), (const float *)(op13), (vuint16mf4_t)(op14), (size_t)(op15)) +#define vloxseg7ei16_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei16_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vfloat32m1_t *)(op6), (const float *)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vloxseg7ei16_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei16_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vfloat32m1_t *)(op6), (vbool32_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (vfloat32m1_t)(op11), (vfloat32m1_t)(op12), (vfloat32m1_t)(op13), (vfloat32m1_t)(op14), (const float *)(op15), (vuint16mf2_t)(op16), (size_t)(op17)) +#define vloxseg7ei16_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei16_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vfloat32mf2_t *)(op6), (const float *)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vloxseg7ei16_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei16_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vfloat32mf2_t *)(op6), (vbool64_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (vfloat32mf2_t)(op11), (vfloat32mf2_t)(op12), (vfloat32mf2_t)(op13), (vfloat32mf2_t)(op14), (const float *)(op15), (vuint16mf4_t)(op16), (size_t)(op17)) +#define vloxseg8ei16_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei16_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vfloat32m1_t *)(op6), (vfloat32m1_t *)(op7), (const float *)(op8), (vuint16mf2_t)(op9), (size_t)(op10)) +#define vloxseg8ei16_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei16_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vfloat32m1_t *)(op6), (vfloat32m1_t *)(op7), (vbool32_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (vfloat32m1_t)(op11), (vfloat32m1_t)(op12), (vfloat32m1_t)(op13), (vfloat32m1_t)(op14), (vfloat32m1_t)(op15), (vfloat32m1_t)(op16), (const float *)(op17), (vuint16mf2_t)(op18), (size_t)(op19)) +#define vloxseg8ei16_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei16_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vfloat32mf2_t *)(op6), (vfloat32mf2_t *)(op7), (const float *)(op8), (vuint16mf4_t)(op9), (size_t)(op10)) +#define vloxseg8ei16_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei16_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vfloat32mf2_t *)(op6), (vfloat32mf2_t *)(op7), (vbool64_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (vfloat32mf2_t)(op11), (vfloat32mf2_t)(op12), (vfloat32mf2_t)(op13), (vfloat32mf2_t)(op14), (vfloat32mf2_t)(op15), (vfloat32mf2_t)(op16), (const float *)(op17), (vuint16mf4_t)(op18), (size_t)(op19)) +#define vloxseg2ei32_v_f32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (const float *)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vbool32_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (const float *)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vloxseg2ei32_v_f32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_f32m2((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (const float *)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_f32m2_m((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vbool16_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (const float *)(op5), (vuint32m2_t)(op6), (size_t)(op7)) +#define vloxseg2ei32_v_f32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_f32m4((vfloat32m4_t *)(op0), (vfloat32m4_t *)(op1), (const float *)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_f32m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_f32m4_m((vfloat32m4_t *)(op0), (vfloat32m4_t *)(op1), (vbool8_t)(op2), (vfloat32m4_t)(op3), (vfloat32m4_t)(op4), (const float *)(op5), (vuint32m4_t)(op6), (size_t)(op7)) +#define vloxseg2ei32_v_f32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (const float *)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vbool64_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (const float *)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vloxseg3ei32_v_f32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei32_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (const float *)(op3), (vuint32m1_t)(op4), (size_t)(op5)) +#define vloxseg3ei32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei32_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vbool32_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (const float *)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vloxseg3ei32_v_f32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei32_v_f32m2((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vfloat32m2_t *)(op2), (const float *)(op3), (vuint32m2_t)(op4), (size_t)(op5)) +#define vloxseg3ei32_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei32_v_f32m2_m((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vfloat32m2_t *)(op2), (vbool16_t)(op3), (vfloat32m2_t)(op4), (vfloat32m2_t)(op5), (vfloat32m2_t)(op6), (const float *)(op7), (vuint32m2_t)(op8), (size_t)(op9)) +#define vloxseg3ei32_v_f32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei32_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (const float *)(op3), (vuint32mf2_t)(op4), (size_t)(op5)) +#define vloxseg3ei32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei32_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vbool64_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (const float *)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vloxseg4ei32_v_f32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei32_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (const float *)(op4), (vuint32m1_t)(op5), (size_t)(op6)) +#define vloxseg4ei32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei32_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vbool32_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (const float *)(op9), (vuint32m1_t)(op10), (size_t)(op11)) +#define vloxseg4ei32_v_f32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei32_v_f32m2((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vfloat32m2_t *)(op2), (vfloat32m2_t *)(op3), (const float *)(op4), (vuint32m2_t)(op5), (size_t)(op6)) +#define vloxseg4ei32_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei32_v_f32m2_m((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vfloat32m2_t *)(op2), (vfloat32m2_t *)(op3), (vbool16_t)(op4), (vfloat32m2_t)(op5), (vfloat32m2_t)(op6), (vfloat32m2_t)(op7), (vfloat32m2_t)(op8), (const float *)(op9), (vuint32m2_t)(op10), (size_t)(op11)) +#define vloxseg4ei32_v_f32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei32_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (const float *)(op4), (vuint32mf2_t)(op5), (size_t)(op6)) +#define vloxseg4ei32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei32_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vbool64_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (const float *)(op9), (vuint32mf2_t)(op10), (size_t)(op11)) +#define vloxseg5ei32_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei32_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (const float *)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vloxseg5ei32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei32_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vbool32_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (const float *)(op11), (vuint32m1_t)(op12), (size_t)(op13)) +#define vloxseg5ei32_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei32_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (const float *)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vloxseg5ei32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei32_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vbool64_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (const float *)(op11), (vuint32mf2_t)(op12), (size_t)(op13)) +#define vloxseg6ei32_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei32_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (const float *)(op6), (vuint32m1_t)(op7), (size_t)(op8)) +#define vloxseg6ei32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei32_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vbool32_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (vfloat32m1_t)(op11), (vfloat32m1_t)(op12), (const float *)(op13), (vuint32m1_t)(op14), (size_t)(op15)) +#define vloxseg6ei32_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei32_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (const float *)(op6), (vuint32mf2_t)(op7), (size_t)(op8)) +#define vloxseg6ei32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei32_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vbool64_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (vfloat32mf2_t)(op11), (vfloat32mf2_t)(op12), (const float *)(op13), (vuint32mf2_t)(op14), (size_t)(op15)) +#define vloxseg7ei32_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei32_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vfloat32m1_t *)(op6), (const float *)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vloxseg7ei32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei32_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vfloat32m1_t *)(op6), (vbool32_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (vfloat32m1_t)(op11), (vfloat32m1_t)(op12), (vfloat32m1_t)(op13), (vfloat32m1_t)(op14), (const float *)(op15), (vuint32m1_t)(op16), (size_t)(op17)) +#define vloxseg7ei32_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei32_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vfloat32mf2_t *)(op6), (const float *)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vloxseg7ei32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei32_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vfloat32mf2_t *)(op6), (vbool64_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (vfloat32mf2_t)(op11), (vfloat32mf2_t)(op12), (vfloat32mf2_t)(op13), (vfloat32mf2_t)(op14), (const float *)(op15), (vuint32mf2_t)(op16), (size_t)(op17)) +#define vloxseg8ei32_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei32_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vfloat32m1_t *)(op6), (vfloat32m1_t *)(op7), (const float *)(op8), (vuint32m1_t)(op9), (size_t)(op10)) +#define vloxseg8ei32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei32_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vfloat32m1_t *)(op6), (vfloat32m1_t *)(op7), (vbool32_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (vfloat32m1_t)(op11), (vfloat32m1_t)(op12), (vfloat32m1_t)(op13), (vfloat32m1_t)(op14), (vfloat32m1_t)(op15), (vfloat32m1_t)(op16), (const float *)(op17), (vuint32m1_t)(op18), (size_t)(op19)) +#define vloxseg8ei32_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei32_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vfloat32mf2_t *)(op6), (vfloat32mf2_t *)(op7), (const float *)(op8), (vuint32mf2_t)(op9), (size_t)(op10)) +#define vloxseg8ei32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei32_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vfloat32mf2_t *)(op6), (vfloat32mf2_t *)(op7), (vbool64_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (vfloat32mf2_t)(op11), (vfloat32mf2_t)(op12), (vfloat32mf2_t)(op13), (vfloat32mf2_t)(op14), (vfloat32mf2_t)(op15), (vfloat32mf2_t)(op16), (const float *)(op17), (vuint32mf2_t)(op18), (size_t)(op19)) +#define vloxseg2ei64_v_f32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei64_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (const float *)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vloxseg2ei64_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei64_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vbool32_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (const float *)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vloxseg2ei64_v_f32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei64_v_f32m2((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (const float *)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vloxseg2ei64_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei64_v_f32m2_m((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vbool16_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (const float *)(op5), (vuint64m4_t)(op6), (size_t)(op7)) +#define vloxseg2ei64_v_f32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei64_v_f32m4((vfloat32m4_t *)(op0), (vfloat32m4_t *)(op1), (const float *)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vloxseg2ei64_v_f32m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei64_v_f32m4_m((vfloat32m4_t *)(op0), (vfloat32m4_t *)(op1), (vbool8_t)(op2), (vfloat32m4_t)(op3), (vfloat32m4_t)(op4), (const float *)(op5), (vuint64m8_t)(op6), (size_t)(op7)) +#define vloxseg2ei64_v_f32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei64_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (const float *)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vloxseg2ei64_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei64_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vbool64_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (const float *)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vloxseg3ei64_v_f32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei64_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (const float *)(op3), (vuint64m2_t)(op4), (size_t)(op5)) +#define vloxseg3ei64_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei64_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vbool32_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (const float *)(op7), (vuint64m2_t)(op8), (size_t)(op9)) +#define vloxseg3ei64_v_f32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei64_v_f32m2((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vfloat32m2_t *)(op2), (const float *)(op3), (vuint64m4_t)(op4), (size_t)(op5)) +#define vloxseg3ei64_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei64_v_f32m2_m((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vfloat32m2_t *)(op2), (vbool16_t)(op3), (vfloat32m2_t)(op4), (vfloat32m2_t)(op5), (vfloat32m2_t)(op6), (const float *)(op7), (vuint64m4_t)(op8), (size_t)(op9)) +#define vloxseg3ei64_v_f32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei64_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (const float *)(op3), (vuint64m1_t)(op4), (size_t)(op5)) +#define vloxseg3ei64_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei64_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vbool64_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (const float *)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vloxseg4ei64_v_f32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei64_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (const float *)(op4), (vuint64m2_t)(op5), (size_t)(op6)) +#define vloxseg4ei64_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei64_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vbool32_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (const float *)(op9), (vuint64m2_t)(op10), (size_t)(op11)) +#define vloxseg4ei64_v_f32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei64_v_f32m2((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vfloat32m2_t *)(op2), (vfloat32m2_t *)(op3), (const float *)(op4), (vuint64m4_t)(op5), (size_t)(op6)) +#define vloxseg4ei64_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei64_v_f32m2_m((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vfloat32m2_t *)(op2), (vfloat32m2_t *)(op3), (vbool16_t)(op4), (vfloat32m2_t)(op5), (vfloat32m2_t)(op6), (vfloat32m2_t)(op7), (vfloat32m2_t)(op8), (const float *)(op9), (vuint64m4_t)(op10), (size_t)(op11)) +#define vloxseg4ei64_v_f32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei64_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (const float *)(op4), (vuint64m1_t)(op5), (size_t)(op6)) +#define vloxseg4ei64_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei64_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vbool64_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (const float *)(op9), (vuint64m1_t)(op10), (size_t)(op11)) +#define vloxseg5ei64_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei64_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (const float *)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vloxseg5ei64_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei64_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vbool32_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (const float *)(op11), (vuint64m2_t)(op12), (size_t)(op13)) +#define vloxseg5ei64_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei64_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (const float *)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vloxseg5ei64_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei64_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vbool64_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (const float *)(op11), (vuint64m1_t)(op12), (size_t)(op13)) +#define vloxseg6ei64_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei64_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (const float *)(op6), (vuint64m2_t)(op7), (size_t)(op8)) +#define vloxseg6ei64_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei64_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vbool32_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (vfloat32m1_t)(op11), (vfloat32m1_t)(op12), (const float *)(op13), (vuint64m2_t)(op14), (size_t)(op15)) +#define vloxseg6ei64_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei64_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (const float *)(op6), (vuint64m1_t)(op7), (size_t)(op8)) +#define vloxseg6ei64_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei64_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vbool64_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (vfloat32mf2_t)(op11), (vfloat32mf2_t)(op12), (const float *)(op13), (vuint64m1_t)(op14), (size_t)(op15)) +#define vloxseg7ei64_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei64_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vfloat32m1_t *)(op6), (const float *)(op7), (vuint64m2_t)(op8), (size_t)(op9)) +#define vloxseg7ei64_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei64_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vfloat32m1_t *)(op6), (vbool32_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (vfloat32m1_t)(op11), (vfloat32m1_t)(op12), (vfloat32m1_t)(op13), (vfloat32m1_t)(op14), (const float *)(op15), (vuint64m2_t)(op16), (size_t)(op17)) +#define vloxseg7ei64_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei64_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vfloat32mf2_t *)(op6), (const float *)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vloxseg7ei64_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei64_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vfloat32mf2_t *)(op6), (vbool64_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (vfloat32mf2_t)(op11), (vfloat32mf2_t)(op12), (vfloat32mf2_t)(op13), (vfloat32mf2_t)(op14), (const float *)(op15), (vuint64m1_t)(op16), (size_t)(op17)) +#define vloxseg8ei64_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei64_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vfloat32m1_t *)(op6), (vfloat32m1_t *)(op7), (const float *)(op8), (vuint64m2_t)(op9), (size_t)(op10)) +#define vloxseg8ei64_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei64_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vfloat32m1_t *)(op6), (vfloat32m1_t *)(op7), (vbool32_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (vfloat32m1_t)(op11), (vfloat32m1_t)(op12), (vfloat32m1_t)(op13), (vfloat32m1_t)(op14), (vfloat32m1_t)(op15), (vfloat32m1_t)(op16), (const float *)(op17), (vuint64m2_t)(op18), (size_t)(op19)) +#define vloxseg8ei64_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei64_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vfloat32mf2_t *)(op6), (vfloat32mf2_t *)(op7), (const float *)(op8), (vuint64m1_t)(op9), (size_t)(op10)) +#define vloxseg8ei64_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei64_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vfloat32mf2_t *)(op6), (vfloat32mf2_t *)(op7), (vbool64_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (vfloat32mf2_t)(op11), (vfloat32mf2_t)(op12), (vfloat32mf2_t)(op13), (vfloat32mf2_t)(op14), (vfloat32mf2_t)(op15), (vfloat32mf2_t)(op16), (const float *)(op17), (vuint64m1_t)(op18), (size_t)(op19)) +#define vsseg2e32_v_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e32_v_f32m1((float *)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (size_t)(op3)) +#define vsseg2e32_v_f32m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e32_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vsseg2e32_v_f32m2(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e32_v_f32m2((float *)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (size_t)(op3)) +#define vsseg2e32_v_f32m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e32_v_f32m2_m((vbool16_t)(op0), (float *)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vsseg2e32_v_f32m4(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e32_v_f32m4((float *)(op0), (vfloat32m4_t)(op1), (vfloat32m4_t)(op2), (size_t)(op3)) +#define vsseg2e32_v_f32m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e32_v_f32m4_m((vbool8_t)(op0), (float *)(op1), (vfloat32m4_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vsseg2e32_v_f32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e32_v_f32mf2((float *)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (size_t)(op3)) +#define vsseg2e32_v_f32mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e32_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vsseg3e32_v_f32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg3e32_v_f32m1((float *)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vsseg3e32_v_f32m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg3e32_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (size_t)(op5)) +#define vsseg3e32_v_f32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg3e32_v_f32m2((float *)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vsseg3e32_v_f32m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg3e32_v_f32m2_m((vbool16_t)(op0), (float *)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (size_t)(op5)) +#define vsseg3e32_v_f32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg3e32_v_f32mf2((float *)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vsseg3e32_v_f32mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg3e32_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (size_t)(op5)) +#define vsseg4e32_v_f32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg4e32_v_f32m1((float *)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (size_t)(op5)) +#define vsseg4e32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg4e32_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (size_t)(op6)) +#define vsseg4e32_v_f32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg4e32_v_f32m2((float *)(op0), (vfloat32m2_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (size_t)(op5)) +#define vsseg4e32_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg4e32_v_f32m2_m((vbool16_t)(op0), (float *)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (vfloat32m2_t)(op5), (size_t)(op6)) +#define vsseg4e32_v_f32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg4e32_v_f32mf2((float *)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (size_t)(op5)) +#define vsseg4e32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg4e32_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (size_t)(op6)) +#define vsseg5e32_v_f32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg5e32_v_f32m1((float *)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (size_t)(op6)) +#define vsseg5e32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg5e32_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (size_t)(op7)) +#define vsseg5e32_v_f32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg5e32_v_f32mf2((float *)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (size_t)(op6)) +#define vsseg5e32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg5e32_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (size_t)(op7)) +#define vsseg6e32_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg6e32_v_f32m1((float *)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (size_t)(op7)) +#define vsseg6e32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg6e32_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (size_t)(op8)) +#define vsseg6e32_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg6e32_v_f32mf2((float *)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (size_t)(op7)) +#define vsseg6e32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg6e32_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (size_t)(op8)) +#define vsseg7e32_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg7e32_v_f32m1((float *)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (size_t)(op8)) +#define vsseg7e32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg7e32_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (size_t)(op9)) +#define vsseg7e32_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg7e32_v_f32mf2((float *)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (size_t)(op8)) +#define vsseg7e32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg7e32_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (size_t)(op9)) +#define vsseg8e32_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg8e32_v_f32m1((float *)(op0), (vfloat32m1_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (size_t)(op9)) +#define vsseg8e32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsseg8e32_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (size_t)(op10)) +#define vsseg8e32_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg8e32_v_f32mf2((float *)(op0), (vfloat32mf2_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (size_t)(op9)) +#define vsseg8e32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsseg8e32_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (size_t)(op10)) +#define vssseg2e32_v_f32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e32_v_f32m1((float *)(op0), (ptrdiff_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vssseg2e32_v_f32m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e32_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (ptrdiff_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (size_t)(op5)) +#define vssseg2e32_v_f32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e32_v_f32m2((float *)(op0), (ptrdiff_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vssseg2e32_v_f32m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e32_v_f32m2_m((vbool16_t)(op0), (float *)(op1), (ptrdiff_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (size_t)(op5)) +#define vssseg2e32_v_f32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e32_v_f32m4((float *)(op0), (ptrdiff_t)(op1), (vfloat32m4_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vssseg2e32_v_f32m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e32_v_f32m4_m((vbool8_t)(op0), (float *)(op1), (ptrdiff_t)(op2), (vfloat32m4_t)(op3), (vfloat32m4_t)(op4), (size_t)(op5)) +#define vssseg2e32_v_f32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e32_v_f32mf2((float *)(op0), (ptrdiff_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vssseg2e32_v_f32mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e32_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (ptrdiff_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (size_t)(op5)) +#define vssseg3e32_v_f32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg3e32_v_f32m1((float *)(op0), (ptrdiff_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (size_t)(op5)) +#define vssseg3e32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg3e32_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (ptrdiff_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (size_t)(op6)) +#define vssseg3e32_v_f32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg3e32_v_f32m2((float *)(op0), (ptrdiff_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (size_t)(op5)) +#define vssseg3e32_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg3e32_v_f32m2_m((vbool16_t)(op0), (float *)(op1), (ptrdiff_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (vfloat32m2_t)(op5), (size_t)(op6)) +#define vssseg3e32_v_f32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg3e32_v_f32mf2((float *)(op0), (ptrdiff_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (size_t)(op5)) +#define vssseg3e32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg3e32_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (ptrdiff_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (size_t)(op6)) +#define vssseg4e32_v_f32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg4e32_v_f32m1((float *)(op0), (ptrdiff_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (size_t)(op6)) +#define vssseg4e32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg4e32_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (ptrdiff_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (size_t)(op7)) +#define vssseg4e32_v_f32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg4e32_v_f32m2((float *)(op0), (ptrdiff_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (vfloat32m2_t)(op5), (size_t)(op6)) +#define vssseg4e32_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg4e32_v_f32m2_m((vbool16_t)(op0), (float *)(op1), (ptrdiff_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (vfloat32m2_t)(op5), (vfloat32m2_t)(op6), (size_t)(op7)) +#define vssseg4e32_v_f32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg4e32_v_f32mf2((float *)(op0), (ptrdiff_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (size_t)(op6)) +#define vssseg4e32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg4e32_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (ptrdiff_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (size_t)(op7)) +#define vssseg5e32_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg5e32_v_f32m1((float *)(op0), (ptrdiff_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (size_t)(op7)) +#define vssseg5e32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg5e32_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (ptrdiff_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (size_t)(op8)) +#define vssseg5e32_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg5e32_v_f32mf2((float *)(op0), (ptrdiff_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (size_t)(op7)) +#define vssseg5e32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg5e32_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (ptrdiff_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (size_t)(op8)) +#define vssseg6e32_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg6e32_v_f32m1((float *)(op0), (ptrdiff_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (size_t)(op8)) +#define vssseg6e32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg6e32_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (ptrdiff_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (size_t)(op9)) +#define vssseg6e32_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg6e32_v_f32mf2((float *)(op0), (ptrdiff_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (size_t)(op8)) +#define vssseg6e32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg6e32_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (ptrdiff_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (size_t)(op9)) +#define vssseg7e32_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg7e32_v_f32m1((float *)(op0), (ptrdiff_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (size_t)(op9)) +#define vssseg7e32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg7e32_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (ptrdiff_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (size_t)(op10)) +#define vssseg7e32_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg7e32_v_f32mf2((float *)(op0), (ptrdiff_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (size_t)(op9)) +#define vssseg7e32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg7e32_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (ptrdiff_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (size_t)(op10)) +#define vssseg8e32_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg8e32_v_f32m1((float *)(op0), (ptrdiff_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (size_t)(op10)) +#define vssseg8e32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vssseg8e32_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (ptrdiff_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (size_t)(op11)) +#define vssseg8e32_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg8e32_v_f32mf2((float *)(op0), (ptrdiff_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (size_t)(op10)) +#define vssseg8e32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vssseg8e32_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (ptrdiff_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (size_t)(op11)) +#define vsuxseg2ei8_v_f32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_f32m1((float *)(op0), (vuint8mf4_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_f32m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint8mf4_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (size_t)(op5)) +#define vsuxseg2ei8_v_f32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_f32m2((float *)(op0), (vuint8mf2_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_f32m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_f32m2_m((vbool16_t)(op0), (float *)(op1), (vuint8mf2_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei8_v_f32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_f32m4((float *)(op0), (vuint8m1_t)(op1), (vfloat32m4_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_f32m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_f32m4_m((vbool8_t)(op0), (float *)(op1), (vuint8m1_t)(op2), (vfloat32m4_t)(op3), (vfloat32m4_t)(op4), (size_t)(op5)) +#define vsuxseg2ei8_v_f32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_f32mf2((float *)(op0), (vuint8mf8_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_f32mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint8mf8_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei8_v_f32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei8_v_f32m1((float *)(op0), (vuint8mf4_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (size_t)(op5)) +#define vsuxseg3ei8_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei8_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint8mf4_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (size_t)(op6)) +#define vsuxseg3ei8_v_f32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei8_v_f32m2((float *)(op0), (vuint8mf2_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei8_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei8_v_f32m2_m((vbool16_t)(op0), (float *)(op1), (vuint8mf2_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (vfloat32m2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei8_v_f32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei8_v_f32mf2((float *)(op0), (vuint8mf8_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei8_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei8_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint8mf8_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei8_v_f32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei8_v_f32m1((float *)(op0), (vuint8mf4_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (size_t)(op6)) +#define vsuxseg4ei8_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei8_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint8mf4_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (size_t)(op7)) +#define vsuxseg4ei8_v_f32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei8_v_f32m2((float *)(op0), (vuint8mf2_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (vfloat32m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei8_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei8_v_f32m2_m((vbool16_t)(op0), (float *)(op1), (vuint8mf2_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (vfloat32m2_t)(op5), (vfloat32m2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei8_v_f32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei8_v_f32mf2((float *)(op0), (vuint8mf8_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei8_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei8_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint8mf8_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei8_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei8_v_f32m1((float *)(op0), (vuint8mf4_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (size_t)(op7)) +#define vsuxseg5ei8_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei8_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint8mf4_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (size_t)(op8)) +#define vsuxseg5ei8_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei8_v_f32mf2((float *)(op0), (vuint8mf8_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei8_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei8_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint8mf8_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (size_t)(op8)) +#define vsuxseg6ei8_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei8_v_f32m1((float *)(op0), (vuint8mf4_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei8_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei8_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint8mf4_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (size_t)(op9)) +#define vsuxseg6ei8_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei8_v_f32mf2((float *)(op0), (vuint8mf8_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (size_t)(op8)) +#define vsuxseg6ei8_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei8_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint8mf8_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (size_t)(op9)) +#define vsuxseg7ei8_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei8_v_f32m1((float *)(op0), (vuint8mf4_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei8_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei8_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint8mf4_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (size_t)(op10)) +#define vsuxseg7ei8_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei8_v_f32mf2((float *)(op0), (vuint8mf8_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (size_t)(op9)) +#define vsuxseg7ei8_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei8_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint8mf8_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (size_t)(op10)) +#define vsuxseg8ei8_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei8_v_f32m1((float *)(op0), (vuint8mf4_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei8_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei8_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint8mf4_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (size_t)(op11)) +#define vsuxseg8ei8_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei8_v_f32mf2((float *)(op0), (vuint8mf8_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (size_t)(op10)) +#define vsuxseg8ei8_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei8_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint8mf8_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (size_t)(op11)) +#define vsuxseg2ei16_v_f32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_f32m1((float *)(op0), (vuint16mf2_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_f32m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint16mf2_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (size_t)(op5)) +#define vsuxseg2ei16_v_f32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_f32m2((float *)(op0), (vuint16m1_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_f32m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_f32m2_m((vbool16_t)(op0), (float *)(op1), (vuint16m1_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei16_v_f32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_f32m4((float *)(op0), (vuint16m2_t)(op1), (vfloat32m4_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_f32m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_f32m4_m((vbool8_t)(op0), (float *)(op1), (vuint16m2_t)(op2), (vfloat32m4_t)(op3), (vfloat32m4_t)(op4), (size_t)(op5)) +#define vsuxseg2ei16_v_f32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_f32mf2((float *)(op0), (vuint16mf4_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_f32mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint16mf4_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei16_v_f32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei16_v_f32m1((float *)(op0), (vuint16mf2_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (size_t)(op5)) +#define vsuxseg3ei16_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei16_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint16mf2_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (size_t)(op6)) +#define vsuxseg3ei16_v_f32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei16_v_f32m2((float *)(op0), (vuint16m1_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei16_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei16_v_f32m2_m((vbool16_t)(op0), (float *)(op1), (vuint16m1_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (vfloat32m2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei16_v_f32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei16_v_f32mf2((float *)(op0), (vuint16mf4_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei16_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei16_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint16mf4_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei16_v_f32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei16_v_f32m1((float *)(op0), (vuint16mf2_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (size_t)(op6)) +#define vsuxseg4ei16_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei16_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint16mf2_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (size_t)(op7)) +#define vsuxseg4ei16_v_f32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei16_v_f32m2((float *)(op0), (vuint16m1_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (vfloat32m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei16_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei16_v_f32m2_m((vbool16_t)(op0), (float *)(op1), (vuint16m1_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (vfloat32m2_t)(op5), (vfloat32m2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei16_v_f32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei16_v_f32mf2((float *)(op0), (vuint16mf4_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei16_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei16_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint16mf4_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei16_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei16_v_f32m1((float *)(op0), (vuint16mf2_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (size_t)(op7)) +#define vsuxseg5ei16_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei16_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint16mf2_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (size_t)(op8)) +#define vsuxseg5ei16_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei16_v_f32mf2((float *)(op0), (vuint16mf4_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei16_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei16_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint16mf4_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (size_t)(op8)) +#define vsuxseg6ei16_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei16_v_f32m1((float *)(op0), (vuint16mf2_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei16_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei16_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint16mf2_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (size_t)(op9)) +#define vsuxseg6ei16_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei16_v_f32mf2((float *)(op0), (vuint16mf4_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (size_t)(op8)) +#define vsuxseg6ei16_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei16_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint16mf4_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (size_t)(op9)) +#define vsuxseg7ei16_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei16_v_f32m1((float *)(op0), (vuint16mf2_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei16_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei16_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint16mf2_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (size_t)(op10)) +#define vsuxseg7ei16_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei16_v_f32mf2((float *)(op0), (vuint16mf4_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (size_t)(op9)) +#define vsuxseg7ei16_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei16_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint16mf4_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (size_t)(op10)) +#define vsuxseg8ei16_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei16_v_f32m1((float *)(op0), (vuint16mf2_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei16_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei16_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint16mf2_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (size_t)(op11)) +#define vsuxseg8ei16_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei16_v_f32mf2((float *)(op0), (vuint16mf4_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (size_t)(op10)) +#define vsuxseg8ei16_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei16_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint16mf4_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (size_t)(op11)) +#define vsuxseg2ei32_v_f32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_f32m1((float *)(op0), (vuint32m1_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_f32m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (size_t)(op5)) +#define vsuxseg2ei32_v_f32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_f32m2((float *)(op0), (vuint32m2_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_f32m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_f32m2_m((vbool16_t)(op0), (float *)(op1), (vuint32m2_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei32_v_f32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_f32m4((float *)(op0), (vuint32m4_t)(op1), (vfloat32m4_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_f32m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_f32m4_m((vbool8_t)(op0), (float *)(op1), (vuint32m4_t)(op2), (vfloat32m4_t)(op3), (vfloat32m4_t)(op4), (size_t)(op5)) +#define vsuxseg2ei32_v_f32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_f32mf2((float *)(op0), (vuint32mf2_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_f32mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei32_v_f32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei32_v_f32m1((float *)(op0), (vuint32m1_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (size_t)(op5)) +#define vsuxseg3ei32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei32_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (size_t)(op6)) +#define vsuxseg3ei32_v_f32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei32_v_f32m2((float *)(op0), (vuint32m2_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei32_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei32_v_f32m2_m((vbool16_t)(op0), (float *)(op1), (vuint32m2_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (vfloat32m2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei32_v_f32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei32_v_f32mf2((float *)(op0), (vuint32mf2_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei32_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei32_v_f32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei32_v_f32m1((float *)(op0), (vuint32m1_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (size_t)(op6)) +#define vsuxseg4ei32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei32_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (size_t)(op7)) +#define vsuxseg4ei32_v_f32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei32_v_f32m2((float *)(op0), (vuint32m2_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (vfloat32m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei32_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei32_v_f32m2_m((vbool16_t)(op0), (float *)(op1), (vuint32m2_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (vfloat32m2_t)(op5), (vfloat32m2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei32_v_f32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei32_v_f32mf2((float *)(op0), (vuint32mf2_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei32_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei32_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei32_v_f32m1((float *)(op0), (vuint32m1_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (size_t)(op7)) +#define vsuxseg5ei32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei32_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (size_t)(op8)) +#define vsuxseg5ei32_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei32_v_f32mf2((float *)(op0), (vuint32mf2_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei32_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (size_t)(op8)) +#define vsuxseg6ei32_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei32_v_f32m1((float *)(op0), (vuint32m1_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei32_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (size_t)(op9)) +#define vsuxseg6ei32_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei32_v_f32mf2((float *)(op0), (vuint32mf2_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (size_t)(op8)) +#define vsuxseg6ei32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei32_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (size_t)(op9)) +#define vsuxseg7ei32_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei32_v_f32m1((float *)(op0), (vuint32m1_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei32_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (size_t)(op10)) +#define vsuxseg7ei32_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei32_v_f32mf2((float *)(op0), (vuint32mf2_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (size_t)(op9)) +#define vsuxseg7ei32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei32_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (size_t)(op10)) +#define vsuxseg8ei32_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei32_v_f32m1((float *)(op0), (vuint32m1_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei32_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (size_t)(op11)) +#define vsuxseg8ei32_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei32_v_f32mf2((float *)(op0), (vuint32mf2_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (size_t)(op10)) +#define vsuxseg8ei32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei32_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (size_t)(op11)) +#define vsuxseg2ei64_v_f32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei64_v_f32m1((float *)(op0), (vuint64m2_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vsuxseg2ei64_v_f32m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei64_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint64m2_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (size_t)(op5)) +#define vsuxseg2ei64_v_f32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei64_v_f32m2((float *)(op0), (vuint64m4_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei64_v_f32m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei64_v_f32m2_m((vbool16_t)(op0), (float *)(op1), (vuint64m4_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei64_v_f32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei64_v_f32m4((float *)(op0), (vuint64m8_t)(op1), (vfloat32m4_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei64_v_f32m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei64_v_f32m4_m((vbool8_t)(op0), (float *)(op1), (vuint64m8_t)(op2), (vfloat32m4_t)(op3), (vfloat32m4_t)(op4), (size_t)(op5)) +#define vsuxseg2ei64_v_f32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei64_v_f32mf2((float *)(op0), (vuint64m1_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei64_v_f32mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei64_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint64m1_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei64_v_f32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei64_v_f32m1((float *)(op0), (vuint64m2_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (size_t)(op5)) +#define vsuxseg3ei64_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei64_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint64m2_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (size_t)(op6)) +#define vsuxseg3ei64_v_f32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei64_v_f32m2((float *)(op0), (vuint64m4_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei64_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei64_v_f32m2_m((vbool16_t)(op0), (float *)(op1), (vuint64m4_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (vfloat32m2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei64_v_f32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei64_v_f32mf2((float *)(op0), (vuint64m1_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei64_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei64_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint64m1_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei64_v_f32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei64_v_f32m1((float *)(op0), (vuint64m2_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (size_t)(op6)) +#define vsuxseg4ei64_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei64_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint64m2_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (size_t)(op7)) +#define vsuxseg4ei64_v_f32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei64_v_f32m2((float *)(op0), (vuint64m4_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (vfloat32m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei64_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei64_v_f32m2_m((vbool16_t)(op0), (float *)(op1), (vuint64m4_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (vfloat32m2_t)(op5), (vfloat32m2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei64_v_f32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei64_v_f32mf2((float *)(op0), (vuint64m1_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei64_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei64_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint64m1_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei64_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei64_v_f32m1((float *)(op0), (vuint64m2_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (size_t)(op7)) +#define vsuxseg5ei64_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei64_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint64m2_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (size_t)(op8)) +#define vsuxseg5ei64_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei64_v_f32mf2((float *)(op0), (vuint64m1_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei64_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei64_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint64m1_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (size_t)(op8)) +#define vsuxseg6ei64_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei64_v_f32m1((float *)(op0), (vuint64m2_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei64_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei64_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint64m2_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (size_t)(op9)) +#define vsuxseg6ei64_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei64_v_f32mf2((float *)(op0), (vuint64m1_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (size_t)(op8)) +#define vsuxseg6ei64_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei64_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint64m1_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (size_t)(op9)) +#define vsuxseg7ei64_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei64_v_f32m1((float *)(op0), (vuint64m2_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei64_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei64_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint64m2_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (size_t)(op10)) +#define vsuxseg7ei64_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei64_v_f32mf2((float *)(op0), (vuint64m1_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (size_t)(op9)) +#define vsuxseg7ei64_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei64_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint64m1_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (size_t)(op10)) +#define vsuxseg8ei64_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei64_v_f32m1((float *)(op0), (vuint64m2_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei64_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei64_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint64m2_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (size_t)(op11)) +#define vsuxseg8ei64_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei64_v_f32mf2((float *)(op0), (vuint64m1_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (size_t)(op10)) +#define vsuxseg8ei64_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei64_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint64m1_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (size_t)(op11)) +#define vsoxseg2ei8_v_f32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_f32m1((float *)(op0), (vuint8mf4_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_f32m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint8mf4_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (size_t)(op5)) +#define vsoxseg2ei8_v_f32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_f32m2((float *)(op0), (vuint8mf2_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_f32m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_f32m2_m((vbool16_t)(op0), (float *)(op1), (vuint8mf2_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei8_v_f32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_f32m4((float *)(op0), (vuint8m1_t)(op1), (vfloat32m4_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_f32m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_f32m4_m((vbool8_t)(op0), (float *)(op1), (vuint8m1_t)(op2), (vfloat32m4_t)(op3), (vfloat32m4_t)(op4), (size_t)(op5)) +#define vsoxseg2ei8_v_f32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_f32mf2((float *)(op0), (vuint8mf8_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_f32mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint8mf8_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei8_v_f32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei8_v_f32m1((float *)(op0), (vuint8mf4_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (size_t)(op5)) +#define vsoxseg3ei8_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei8_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint8mf4_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (size_t)(op6)) +#define vsoxseg3ei8_v_f32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei8_v_f32m2((float *)(op0), (vuint8mf2_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei8_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei8_v_f32m2_m((vbool16_t)(op0), (float *)(op1), (vuint8mf2_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (vfloat32m2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei8_v_f32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei8_v_f32mf2((float *)(op0), (vuint8mf8_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei8_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei8_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint8mf8_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei8_v_f32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei8_v_f32m1((float *)(op0), (vuint8mf4_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (size_t)(op6)) +#define vsoxseg4ei8_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei8_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint8mf4_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (size_t)(op7)) +#define vsoxseg4ei8_v_f32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei8_v_f32m2((float *)(op0), (vuint8mf2_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (vfloat32m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei8_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei8_v_f32m2_m((vbool16_t)(op0), (float *)(op1), (vuint8mf2_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (vfloat32m2_t)(op5), (vfloat32m2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei8_v_f32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei8_v_f32mf2((float *)(op0), (vuint8mf8_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei8_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei8_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint8mf8_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei8_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei8_v_f32m1((float *)(op0), (vuint8mf4_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (size_t)(op7)) +#define vsoxseg5ei8_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei8_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint8mf4_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (size_t)(op8)) +#define vsoxseg5ei8_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei8_v_f32mf2((float *)(op0), (vuint8mf8_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei8_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei8_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint8mf8_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (size_t)(op8)) +#define vsoxseg6ei8_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei8_v_f32m1((float *)(op0), (vuint8mf4_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei8_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei8_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint8mf4_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (size_t)(op9)) +#define vsoxseg6ei8_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei8_v_f32mf2((float *)(op0), (vuint8mf8_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (size_t)(op8)) +#define vsoxseg6ei8_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei8_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint8mf8_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (size_t)(op9)) +#define vsoxseg7ei8_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei8_v_f32m1((float *)(op0), (vuint8mf4_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei8_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei8_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint8mf4_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (size_t)(op10)) +#define vsoxseg7ei8_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei8_v_f32mf2((float *)(op0), (vuint8mf8_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (size_t)(op9)) +#define vsoxseg7ei8_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei8_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint8mf8_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (size_t)(op10)) +#define vsoxseg8ei8_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei8_v_f32m1((float *)(op0), (vuint8mf4_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei8_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei8_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint8mf4_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (size_t)(op11)) +#define vsoxseg8ei8_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei8_v_f32mf2((float *)(op0), (vuint8mf8_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (size_t)(op10)) +#define vsoxseg8ei8_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei8_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint8mf8_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (size_t)(op11)) +#define vsoxseg2ei16_v_f32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_f32m1((float *)(op0), (vuint16mf2_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_f32m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint16mf2_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (size_t)(op5)) +#define vsoxseg2ei16_v_f32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_f32m2((float *)(op0), (vuint16m1_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_f32m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_f32m2_m((vbool16_t)(op0), (float *)(op1), (vuint16m1_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei16_v_f32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_f32m4((float *)(op0), (vuint16m2_t)(op1), (vfloat32m4_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_f32m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_f32m4_m((vbool8_t)(op0), (float *)(op1), (vuint16m2_t)(op2), (vfloat32m4_t)(op3), (vfloat32m4_t)(op4), (size_t)(op5)) +#define vsoxseg2ei16_v_f32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_f32mf2((float *)(op0), (vuint16mf4_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_f32mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint16mf4_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei16_v_f32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei16_v_f32m1((float *)(op0), (vuint16mf2_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (size_t)(op5)) +#define vsoxseg3ei16_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei16_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint16mf2_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (size_t)(op6)) +#define vsoxseg3ei16_v_f32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei16_v_f32m2((float *)(op0), (vuint16m1_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei16_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei16_v_f32m2_m((vbool16_t)(op0), (float *)(op1), (vuint16m1_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (vfloat32m2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei16_v_f32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei16_v_f32mf2((float *)(op0), (vuint16mf4_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei16_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei16_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint16mf4_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei16_v_f32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei16_v_f32m1((float *)(op0), (vuint16mf2_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (size_t)(op6)) +#define vsoxseg4ei16_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei16_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint16mf2_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (size_t)(op7)) +#define vsoxseg4ei16_v_f32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei16_v_f32m2((float *)(op0), (vuint16m1_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (vfloat32m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei16_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei16_v_f32m2_m((vbool16_t)(op0), (float *)(op1), (vuint16m1_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (vfloat32m2_t)(op5), (vfloat32m2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei16_v_f32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei16_v_f32mf2((float *)(op0), (vuint16mf4_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei16_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei16_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint16mf4_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei16_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei16_v_f32m1((float *)(op0), (vuint16mf2_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (size_t)(op7)) +#define vsoxseg5ei16_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei16_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint16mf2_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (size_t)(op8)) +#define vsoxseg5ei16_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei16_v_f32mf2((float *)(op0), (vuint16mf4_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei16_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei16_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint16mf4_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (size_t)(op8)) +#define vsoxseg6ei16_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei16_v_f32m1((float *)(op0), (vuint16mf2_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei16_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei16_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint16mf2_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (size_t)(op9)) +#define vsoxseg6ei16_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei16_v_f32mf2((float *)(op0), (vuint16mf4_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (size_t)(op8)) +#define vsoxseg6ei16_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei16_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint16mf4_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (size_t)(op9)) +#define vsoxseg7ei16_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei16_v_f32m1((float *)(op0), (vuint16mf2_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei16_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei16_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint16mf2_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (size_t)(op10)) +#define vsoxseg7ei16_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei16_v_f32mf2((float *)(op0), (vuint16mf4_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (size_t)(op9)) +#define vsoxseg7ei16_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei16_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint16mf4_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (size_t)(op10)) +#define vsoxseg8ei16_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei16_v_f32m1((float *)(op0), (vuint16mf2_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei16_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei16_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint16mf2_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (size_t)(op11)) +#define vsoxseg8ei16_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei16_v_f32mf2((float *)(op0), (vuint16mf4_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (size_t)(op10)) +#define vsoxseg8ei16_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei16_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint16mf4_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (size_t)(op11)) +#define vsoxseg2ei32_v_f32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_f32m1((float *)(op0), (vuint32m1_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_f32m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (size_t)(op5)) +#define vsoxseg2ei32_v_f32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_f32m2((float *)(op0), (vuint32m2_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_f32m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_f32m2_m((vbool16_t)(op0), (float *)(op1), (vuint32m2_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei32_v_f32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_f32m4((float *)(op0), (vuint32m4_t)(op1), (vfloat32m4_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_f32m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_f32m4_m((vbool8_t)(op0), (float *)(op1), (vuint32m4_t)(op2), (vfloat32m4_t)(op3), (vfloat32m4_t)(op4), (size_t)(op5)) +#define vsoxseg2ei32_v_f32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_f32mf2((float *)(op0), (vuint32mf2_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_f32mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei32_v_f32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei32_v_f32m1((float *)(op0), (vuint32m1_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (size_t)(op5)) +#define vsoxseg3ei32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei32_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (size_t)(op6)) +#define vsoxseg3ei32_v_f32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei32_v_f32m2((float *)(op0), (vuint32m2_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei32_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei32_v_f32m2_m((vbool16_t)(op0), (float *)(op1), (vuint32m2_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (vfloat32m2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei32_v_f32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei32_v_f32mf2((float *)(op0), (vuint32mf2_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei32_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei32_v_f32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei32_v_f32m1((float *)(op0), (vuint32m1_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (size_t)(op6)) +#define vsoxseg4ei32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei32_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (size_t)(op7)) +#define vsoxseg4ei32_v_f32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei32_v_f32m2((float *)(op0), (vuint32m2_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (vfloat32m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei32_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei32_v_f32m2_m((vbool16_t)(op0), (float *)(op1), (vuint32m2_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (vfloat32m2_t)(op5), (vfloat32m2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei32_v_f32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei32_v_f32mf2((float *)(op0), (vuint32mf2_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei32_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei32_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei32_v_f32m1((float *)(op0), (vuint32m1_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (size_t)(op7)) +#define vsoxseg5ei32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei32_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (size_t)(op8)) +#define vsoxseg5ei32_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei32_v_f32mf2((float *)(op0), (vuint32mf2_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei32_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (size_t)(op8)) +#define vsoxseg6ei32_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei32_v_f32m1((float *)(op0), (vuint32m1_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei32_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (size_t)(op9)) +#define vsoxseg6ei32_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei32_v_f32mf2((float *)(op0), (vuint32mf2_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (size_t)(op8)) +#define vsoxseg6ei32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei32_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (size_t)(op9)) +#define vsoxseg7ei32_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei32_v_f32m1((float *)(op0), (vuint32m1_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei32_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (size_t)(op10)) +#define vsoxseg7ei32_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei32_v_f32mf2((float *)(op0), (vuint32mf2_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (size_t)(op9)) +#define vsoxseg7ei32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei32_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (size_t)(op10)) +#define vsoxseg8ei32_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei32_v_f32m1((float *)(op0), (vuint32m1_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei32_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (size_t)(op11)) +#define vsoxseg8ei32_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei32_v_f32mf2((float *)(op0), (vuint32mf2_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (size_t)(op10)) +#define vsoxseg8ei32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei32_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (size_t)(op11)) +#define vsoxseg2ei64_v_f32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei64_v_f32m1((float *)(op0), (vuint64m2_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (size_t)(op4)) +#define vsoxseg2ei64_v_f32m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei64_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint64m2_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (size_t)(op5)) +#define vsoxseg2ei64_v_f32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei64_v_f32m2((float *)(op0), (vuint64m4_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei64_v_f32m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei64_v_f32m2_m((vbool16_t)(op0), (float *)(op1), (vuint64m4_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei64_v_f32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei64_v_f32m4((float *)(op0), (vuint64m8_t)(op1), (vfloat32m4_t)(op2), (vfloat32m4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei64_v_f32m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei64_v_f32m4_m((vbool8_t)(op0), (float *)(op1), (vuint64m8_t)(op2), (vfloat32m4_t)(op3), (vfloat32m4_t)(op4), (size_t)(op5)) +#define vsoxseg2ei64_v_f32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei64_v_f32mf2((float *)(op0), (vuint64m1_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei64_v_f32mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei64_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint64m1_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei64_v_f32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei64_v_f32m1((float *)(op0), (vuint64m2_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (size_t)(op5)) +#define vsoxseg3ei64_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei64_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint64m2_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (size_t)(op6)) +#define vsoxseg3ei64_v_f32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei64_v_f32m2((float *)(op0), (vuint64m4_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei64_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei64_v_f32m2_m((vbool16_t)(op0), (float *)(op1), (vuint64m4_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (vfloat32m2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei64_v_f32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei64_v_f32mf2((float *)(op0), (vuint64m1_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei64_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei64_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint64m1_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei64_v_f32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei64_v_f32m1((float *)(op0), (vuint64m2_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (size_t)(op6)) +#define vsoxseg4ei64_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei64_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint64m2_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (size_t)(op7)) +#define vsoxseg4ei64_v_f32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei64_v_f32m2((float *)(op0), (vuint64m4_t)(op1), (vfloat32m2_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (vfloat32m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei64_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei64_v_f32m2_m((vbool16_t)(op0), (float *)(op1), (vuint64m4_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (vfloat32m2_t)(op5), (vfloat32m2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei64_v_f32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei64_v_f32mf2((float *)(op0), (vuint64m1_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei64_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei64_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint64m1_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei64_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei64_v_f32m1((float *)(op0), (vuint64m2_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (size_t)(op7)) +#define vsoxseg5ei64_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei64_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint64m2_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (size_t)(op8)) +#define vsoxseg5ei64_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei64_v_f32mf2((float *)(op0), (vuint64m1_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei64_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei64_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint64m1_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (size_t)(op8)) +#define vsoxseg6ei64_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei64_v_f32m1((float *)(op0), (vuint64m2_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei64_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei64_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint64m2_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (size_t)(op9)) +#define vsoxseg6ei64_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei64_v_f32mf2((float *)(op0), (vuint64m1_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (size_t)(op8)) +#define vsoxseg6ei64_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei64_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint64m1_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (size_t)(op9)) +#define vsoxseg7ei64_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei64_v_f32m1((float *)(op0), (vuint64m2_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei64_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei64_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint64m2_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (size_t)(op10)) +#define vsoxseg7ei64_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei64_v_f32mf2((float *)(op0), (vuint64m1_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (size_t)(op9)) +#define vsoxseg7ei64_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei64_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint64m1_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (size_t)(op10)) +#define vsoxseg8ei64_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei64_v_f32m1((float *)(op0), (vuint64m2_t)(op1), (vfloat32m1_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei64_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei64_v_f32m1_m((vbool32_t)(op0), (float *)(op1), (vuint64m2_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (size_t)(op11)) +#define vsoxseg8ei64_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei64_v_f32mf2((float *)(op0), (vuint64m1_t)(op1), (vfloat32mf2_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (size_t)(op10)) +#define vsoxseg8ei64_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei64_v_f32mf2_m((vbool64_t)(op0), (float *)(op1), (vuint64m1_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (size_t)(op11)) +#define vlseg2e32_v_f32m1(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e32_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (const float *)(op2), (size_t)(op3)) +#define vlseg2e32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e32_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vbool32_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (const float *)(op5), (size_t)(op6)) +#define vlseg2e32_v_f32m2(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e32_v_f32m2((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (const float *)(op2), (size_t)(op3)) +#define vlseg2e32_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e32_v_f32m2_m((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vbool16_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (const float *)(op5), (size_t)(op6)) +#define vlseg2e32_v_f32m4(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e32_v_f32m4((vfloat32m4_t *)(op0), (vfloat32m4_t *)(op1), (const float *)(op2), (size_t)(op3)) +#define vlseg2e32_v_f32m4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e32_v_f32m4_m((vfloat32m4_t *)(op0), (vfloat32m4_t *)(op1), (vbool8_t)(op2), (vfloat32m4_t)(op3), (vfloat32m4_t)(op4), (const float *)(op5), (size_t)(op6)) +#define vlseg2e32_v_f32mf2(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e32_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (const float *)(op2), (size_t)(op3)) +#define vlseg2e32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e32_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vbool64_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (const float *)(op5), (size_t)(op6)) +#define vlseg3e32_v_f32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg3e32_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (const float *)(op3), (size_t)(op4)) +#define vlseg3e32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg3e32_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vbool32_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (const float *)(op7), (size_t)(op8)) +#define vlseg3e32_v_f32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg3e32_v_f32m2((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vfloat32m2_t *)(op2), (const float *)(op3), (size_t)(op4)) +#define vlseg3e32_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg3e32_v_f32m2_m((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vfloat32m2_t *)(op2), (vbool16_t)(op3), (vfloat32m2_t)(op4), (vfloat32m2_t)(op5), (vfloat32m2_t)(op6), (const float *)(op7), (size_t)(op8)) +#define vlseg3e32_v_f32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg3e32_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (const float *)(op3), (size_t)(op4)) +#define vlseg3e32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg3e32_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vbool64_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (const float *)(op7), (size_t)(op8)) +#define vlseg4e32_v_f32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg4e32_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (const float *)(op4), (size_t)(op5)) +#define vlseg4e32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg4e32_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vbool32_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (const float *)(op9), (size_t)(op10)) +#define vlseg4e32_v_f32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg4e32_v_f32m2((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vfloat32m2_t *)(op2), (vfloat32m2_t *)(op3), (const float *)(op4), (size_t)(op5)) +#define vlseg4e32_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg4e32_v_f32m2_m((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vfloat32m2_t *)(op2), (vfloat32m2_t *)(op3), (vbool16_t)(op4), (vfloat32m2_t)(op5), (vfloat32m2_t)(op6), (vfloat32m2_t)(op7), (vfloat32m2_t)(op8), (const float *)(op9), (size_t)(op10)) +#define vlseg4e32_v_f32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg4e32_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (const float *)(op4), (size_t)(op5)) +#define vlseg4e32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg4e32_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vbool64_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (const float *)(op9), (size_t)(op10)) +#define vlseg5e32_v_f32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg5e32_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (const float *)(op5), (size_t)(op6)) +#define vlseg5e32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12) \ +__builtin_rvv_vlseg5e32_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vbool32_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (const float *)(op11), (size_t)(op12)) +#define vlseg5e32_v_f32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg5e32_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (const float *)(op5), (size_t)(op6)) +#define vlseg5e32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12) \ +__builtin_rvv_vlseg5e32_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vbool64_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (const float *)(op11), (size_t)(op12)) +#define vlseg6e32_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg6e32_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (const float *)(op6), (size_t)(op7)) +#define vlseg6e32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14) \ +__builtin_rvv_vlseg6e32_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vbool32_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (vfloat32m1_t)(op11), (vfloat32m1_t)(op12), (const float *)(op13), (size_t)(op14)) +#define vlseg6e32_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg6e32_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (const float *)(op6), (size_t)(op7)) +#define vlseg6e32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14) \ +__builtin_rvv_vlseg6e32_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vbool64_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (vfloat32mf2_t)(op11), (vfloat32mf2_t)(op12), (const float *)(op13), (size_t)(op14)) +#define vlseg7e32_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg7e32_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vfloat32m1_t *)(op6), (const float *)(op7), (size_t)(op8)) +#define vlseg7e32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16) \ +__builtin_rvv_vlseg7e32_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vfloat32m1_t *)(op6), (vbool32_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (vfloat32m1_t)(op11), (vfloat32m1_t)(op12), (vfloat32m1_t)(op13), (vfloat32m1_t)(op14), (const float *)(op15), (size_t)(op16)) +#define vlseg7e32_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg7e32_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vfloat32mf2_t *)(op6), (const float *)(op7), (size_t)(op8)) +#define vlseg7e32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16) \ +__builtin_rvv_vlseg7e32_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vfloat32mf2_t *)(op6), (vbool64_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (vfloat32mf2_t)(op11), (vfloat32mf2_t)(op12), (vfloat32mf2_t)(op13), (vfloat32mf2_t)(op14), (const float *)(op15), (size_t)(op16)) +#define vlseg8e32_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg8e32_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vfloat32m1_t *)(op6), (vfloat32m1_t *)(op7), (const float *)(op8), (size_t)(op9)) +#define vlseg8e32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18) \ +__builtin_rvv_vlseg8e32_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vfloat32m1_t *)(op6), (vfloat32m1_t *)(op7), (vbool32_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (vfloat32m1_t)(op11), (vfloat32m1_t)(op12), (vfloat32m1_t)(op13), (vfloat32m1_t)(op14), (vfloat32m1_t)(op15), (vfloat32m1_t)(op16), (const float *)(op17), (size_t)(op18)) +#define vlseg8e32_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg8e32_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vfloat32mf2_t *)(op6), (vfloat32mf2_t *)(op7), (const float *)(op8), (size_t)(op9)) +#define vlseg8e32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18) \ +__builtin_rvv_vlseg8e32_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vfloat32mf2_t *)(op6), (vfloat32mf2_t *)(op7), (vbool64_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (vfloat32mf2_t)(op11), (vfloat32mf2_t)(op12), (vfloat32mf2_t)(op13), (vfloat32mf2_t)(op14), (vfloat32mf2_t)(op15), (vfloat32mf2_t)(op16), (const float *)(op17), (size_t)(op18)) +#define vlseg2e32ff_v_f32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e32ff_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (const float *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e32ff_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e32ff_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vbool32_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (const float *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg2e32ff_v_f32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e32ff_v_f32m2((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (const float *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e32ff_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e32ff_v_f32m2_m((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vbool16_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (const float *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg2e32ff_v_f32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e32ff_v_f32m4((vfloat32m4_t *)(op0), (vfloat32m4_t *)(op1), (const float *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e32ff_v_f32m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e32ff_v_f32m4_m((vfloat32m4_t *)(op0), (vfloat32m4_t *)(op1), (vbool8_t)(op2), (vfloat32m4_t)(op3), (vfloat32m4_t)(op4), (const float *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg2e32ff_v_f32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e32ff_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (const float *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e32ff_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e32ff_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vbool64_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (const float *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg3e32ff_v_f32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg3e32ff_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (const float *)(op3), (size_t *)(op4), (size_t)(op5)) +#define vlseg3e32ff_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg3e32ff_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vbool32_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (const float *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg3e32ff_v_f32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg3e32ff_v_f32m2((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vfloat32m2_t *)(op2), (const float *)(op3), (size_t *)(op4), (size_t)(op5)) +#define vlseg3e32ff_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg3e32ff_v_f32m2_m((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vfloat32m2_t *)(op2), (vbool16_t)(op3), (vfloat32m2_t)(op4), (vfloat32m2_t)(op5), (vfloat32m2_t)(op6), (const float *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg3e32ff_v_f32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg3e32ff_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (const float *)(op3), (size_t *)(op4), (size_t)(op5)) +#define vlseg3e32ff_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg3e32ff_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vbool64_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (const float *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg4e32ff_v_f32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg4e32ff_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (const float *)(op4), (size_t *)(op5), (size_t)(op6)) +#define vlseg4e32ff_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlseg4e32ff_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vbool32_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (const float *)(op9), (size_t *)(op10), (size_t)(op11)) +#define vlseg4e32ff_v_f32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg4e32ff_v_f32m2((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vfloat32m2_t *)(op2), (vfloat32m2_t *)(op3), (const float *)(op4), (size_t *)(op5), (size_t)(op6)) +#define vlseg4e32ff_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlseg4e32ff_v_f32m2_m((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vfloat32m2_t *)(op2), (vfloat32m2_t *)(op3), (vbool16_t)(op4), (vfloat32m2_t)(op5), (vfloat32m2_t)(op6), (vfloat32m2_t)(op7), (vfloat32m2_t)(op8), (const float *)(op9), (size_t *)(op10), (size_t)(op11)) +#define vlseg4e32ff_v_f32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg4e32ff_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (const float *)(op4), (size_t *)(op5), (size_t)(op6)) +#define vlseg4e32ff_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlseg4e32ff_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vbool64_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (const float *)(op9), (size_t *)(op10), (size_t)(op11)) +#define vlseg5e32ff_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg5e32ff_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (const float *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg5e32ff_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlseg5e32ff_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vbool32_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (const float *)(op11), (size_t *)(op12), (size_t)(op13)) +#define vlseg5e32ff_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg5e32ff_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (const float *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg5e32ff_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlseg5e32ff_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vbool64_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (const float *)(op11), (size_t *)(op12), (size_t)(op13)) +#define vlseg6e32ff_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg6e32ff_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (const float *)(op6), (size_t *)(op7), (size_t)(op8)) +#define vlseg6e32ff_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlseg6e32ff_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vbool32_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (vfloat32m1_t)(op11), (vfloat32m1_t)(op12), (const float *)(op13), (size_t *)(op14), (size_t)(op15)) +#define vlseg6e32ff_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg6e32ff_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (const float *)(op6), (size_t *)(op7), (size_t)(op8)) +#define vlseg6e32ff_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlseg6e32ff_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vbool64_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (vfloat32mf2_t)(op11), (vfloat32mf2_t)(op12), (const float *)(op13), (size_t *)(op14), (size_t)(op15)) +#define vlseg7e32ff_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg7e32ff_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vfloat32m1_t *)(op6), (const float *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg7e32ff_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlseg7e32ff_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vfloat32m1_t *)(op6), (vbool32_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (vfloat32m1_t)(op11), (vfloat32m1_t)(op12), (vfloat32m1_t)(op13), (vfloat32m1_t)(op14), (const float *)(op15), (size_t *)(op16), (size_t)(op17)) +#define vlseg7e32ff_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg7e32ff_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vfloat32mf2_t *)(op6), (const float *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg7e32ff_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlseg7e32ff_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vfloat32mf2_t *)(op6), (vbool64_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (vfloat32mf2_t)(op11), (vfloat32mf2_t)(op12), (vfloat32mf2_t)(op13), (vfloat32mf2_t)(op14), (const float *)(op15), (size_t *)(op16), (size_t)(op17)) +#define vlseg8e32ff_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg8e32ff_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vfloat32m1_t *)(op6), (vfloat32m1_t *)(op7), (const float *)(op8), (size_t *)(op9), (size_t)(op10)) +#define vlseg8e32ff_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlseg8e32ff_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vfloat32m1_t *)(op6), (vfloat32m1_t *)(op7), (vbool32_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (vfloat32m1_t)(op11), (vfloat32m1_t)(op12), (vfloat32m1_t)(op13), (vfloat32m1_t)(op14), (vfloat32m1_t)(op15), (vfloat32m1_t)(op16), (const float *)(op17), (size_t *)(op18), (size_t)(op19)) +#define vlseg8e32ff_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg8e32ff_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vfloat32mf2_t *)(op6), (vfloat32mf2_t *)(op7), (const float *)(op8), (size_t *)(op9), (size_t)(op10)) +#define vlseg8e32ff_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlseg8e32ff_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vfloat32mf2_t *)(op6), (vfloat32mf2_t *)(op7), (vbool64_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (vfloat32mf2_t)(op11), (vfloat32mf2_t)(op12), (vfloat32mf2_t)(op13), (vfloat32mf2_t)(op14), (vfloat32mf2_t)(op15), (vfloat32mf2_t)(op16), (const float *)(op17), (size_t *)(op18), (size_t)(op19)) +#define vlsseg2e32_v_f32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e32_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (const float *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e32_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vbool32_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (const float *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg2e32_v_f32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e32_v_f32m2((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (const float *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e32_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e32_v_f32m2_m((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vbool16_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (const float *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg2e32_v_f32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e32_v_f32m4((vfloat32m4_t *)(op0), (vfloat32m4_t *)(op1), (const float *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e32_v_f32m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e32_v_f32m4_m((vfloat32m4_t *)(op0), (vfloat32m4_t *)(op1), (vbool8_t)(op2), (vfloat32m4_t)(op3), (vfloat32m4_t)(op4), (const float *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg2e32_v_f32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e32_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (const float *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e32_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vbool64_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (const float *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg3e32_v_f32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlsseg3e32_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (const float *)(op3), (ptrdiff_t)(op4), (size_t)(op5)) +#define vlsseg3e32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg3e32_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vbool32_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (const float *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg3e32_v_f32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlsseg3e32_v_f32m2((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vfloat32m2_t *)(op2), (const float *)(op3), (ptrdiff_t)(op4), (size_t)(op5)) +#define vlsseg3e32_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg3e32_v_f32m2_m((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vfloat32m2_t *)(op2), (vbool16_t)(op3), (vfloat32m2_t)(op4), (vfloat32m2_t)(op5), (vfloat32m2_t)(op6), (const float *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg3e32_v_f32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlsseg3e32_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (const float *)(op3), (ptrdiff_t)(op4), (size_t)(op5)) +#define vlsseg3e32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg3e32_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vbool64_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (const float *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg4e32_v_f32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlsseg4e32_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (const float *)(op4), (ptrdiff_t)(op5), (size_t)(op6)) +#define vlsseg4e32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlsseg4e32_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vbool32_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (const float *)(op9), (ptrdiff_t)(op10), (size_t)(op11)) +#define vlsseg4e32_v_f32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlsseg4e32_v_f32m2((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vfloat32m2_t *)(op2), (vfloat32m2_t *)(op3), (const float *)(op4), (ptrdiff_t)(op5), (size_t)(op6)) +#define vlsseg4e32_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlsseg4e32_v_f32m2_m((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vfloat32m2_t *)(op2), (vfloat32m2_t *)(op3), (vbool16_t)(op4), (vfloat32m2_t)(op5), (vfloat32m2_t)(op6), (vfloat32m2_t)(op7), (vfloat32m2_t)(op8), (const float *)(op9), (ptrdiff_t)(op10), (size_t)(op11)) +#define vlsseg4e32_v_f32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlsseg4e32_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (const float *)(op4), (ptrdiff_t)(op5), (size_t)(op6)) +#define vlsseg4e32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlsseg4e32_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vbool64_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (const float *)(op9), (ptrdiff_t)(op10), (size_t)(op11)) +#define vlsseg5e32_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg5e32_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (const float *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg5e32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlsseg5e32_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vbool32_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (const float *)(op11), (ptrdiff_t)(op12), (size_t)(op13)) +#define vlsseg5e32_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg5e32_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (const float *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg5e32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlsseg5e32_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vbool64_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (const float *)(op11), (ptrdiff_t)(op12), (size_t)(op13)) +#define vlsseg6e32_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlsseg6e32_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (const float *)(op6), (ptrdiff_t)(op7), (size_t)(op8)) +#define vlsseg6e32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlsseg6e32_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vbool32_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (vfloat32m1_t)(op11), (vfloat32m1_t)(op12), (const float *)(op13), (ptrdiff_t)(op14), (size_t)(op15)) +#define vlsseg6e32_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlsseg6e32_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (const float *)(op6), (ptrdiff_t)(op7), (size_t)(op8)) +#define vlsseg6e32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlsseg6e32_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vbool64_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (vfloat32mf2_t)(op11), (vfloat32mf2_t)(op12), (const float *)(op13), (ptrdiff_t)(op14), (size_t)(op15)) +#define vlsseg7e32_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg7e32_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vfloat32m1_t *)(op6), (const float *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg7e32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlsseg7e32_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vfloat32m1_t *)(op6), (vbool32_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (vfloat32m1_t)(op11), (vfloat32m1_t)(op12), (vfloat32m1_t)(op13), (vfloat32m1_t)(op14), (const float *)(op15), (ptrdiff_t)(op16), (size_t)(op17)) +#define vlsseg7e32_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg7e32_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vfloat32mf2_t *)(op6), (const float *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg7e32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlsseg7e32_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vfloat32mf2_t *)(op6), (vbool64_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (vfloat32mf2_t)(op11), (vfloat32mf2_t)(op12), (vfloat32mf2_t)(op13), (vfloat32mf2_t)(op14), (const float *)(op15), (ptrdiff_t)(op16), (size_t)(op17)) +#define vlsseg8e32_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlsseg8e32_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vfloat32m1_t *)(op6), (vfloat32m1_t *)(op7), (const float *)(op8), (ptrdiff_t)(op9), (size_t)(op10)) +#define vlsseg8e32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlsseg8e32_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vfloat32m1_t *)(op6), (vfloat32m1_t *)(op7), (vbool32_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (vfloat32m1_t)(op11), (vfloat32m1_t)(op12), (vfloat32m1_t)(op13), (vfloat32m1_t)(op14), (vfloat32m1_t)(op15), (vfloat32m1_t)(op16), (const float *)(op17), (ptrdiff_t)(op18), (size_t)(op19)) +#define vlsseg8e32_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlsseg8e32_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vfloat32mf2_t *)(op6), (vfloat32mf2_t *)(op7), (const float *)(op8), (ptrdiff_t)(op9), (size_t)(op10)) +#define vlsseg8e32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlsseg8e32_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vfloat32mf2_t *)(op6), (vfloat32mf2_t *)(op7), (vbool64_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (vfloat32mf2_t)(op11), (vfloat32mf2_t)(op12), (vfloat32mf2_t)(op13), (vfloat32mf2_t)(op14), (vfloat32mf2_t)(op15), (vfloat32mf2_t)(op16), (const float *)(op17), (ptrdiff_t)(op18), (size_t)(op19)) +#define vluxseg2ei8_v_f32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (const float *)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vbool32_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (const float *)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vluxseg2ei8_v_f32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_f32m2((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (const float *)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_f32m2_m((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vbool16_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (const float *)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vluxseg2ei8_v_f32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_f32m4((vfloat32m4_t *)(op0), (vfloat32m4_t *)(op1), (const float *)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_f32m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_f32m4_m((vfloat32m4_t *)(op0), (vfloat32m4_t *)(op1), (vbool8_t)(op2), (vfloat32m4_t)(op3), (vfloat32m4_t)(op4), (const float *)(op5), (vuint8m1_t)(op6), (size_t)(op7)) +#define vluxseg2ei8_v_f32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (const float *)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vbool64_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (const float *)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vluxseg3ei8_v_f32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei8_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (const float *)(op3), (vuint8mf4_t)(op4), (size_t)(op5)) +#define vluxseg3ei8_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei8_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vbool32_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (const float *)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vluxseg3ei8_v_f32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei8_v_f32m2((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vfloat32m2_t *)(op2), (const float *)(op3), (vuint8mf2_t)(op4), (size_t)(op5)) +#define vluxseg3ei8_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei8_v_f32m2_m((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vfloat32m2_t *)(op2), (vbool16_t)(op3), (vfloat32m2_t)(op4), (vfloat32m2_t)(op5), (vfloat32m2_t)(op6), (const float *)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vluxseg3ei8_v_f32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei8_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (const float *)(op3), (vuint8mf8_t)(op4), (size_t)(op5)) +#define vluxseg3ei8_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei8_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vbool64_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (const float *)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vluxseg4ei8_v_f32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei8_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (const float *)(op4), (vuint8mf4_t)(op5), (size_t)(op6)) +#define vluxseg4ei8_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei8_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vbool32_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (const float *)(op9), (vuint8mf4_t)(op10), (size_t)(op11)) +#define vluxseg4ei8_v_f32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei8_v_f32m2((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vfloat32m2_t *)(op2), (vfloat32m2_t *)(op3), (const float *)(op4), (vuint8mf2_t)(op5), (size_t)(op6)) +#define vluxseg4ei8_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei8_v_f32m2_m((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vfloat32m2_t *)(op2), (vfloat32m2_t *)(op3), (vbool16_t)(op4), (vfloat32m2_t)(op5), (vfloat32m2_t)(op6), (vfloat32m2_t)(op7), (vfloat32m2_t)(op8), (const float *)(op9), (vuint8mf2_t)(op10), (size_t)(op11)) +#define vluxseg4ei8_v_f32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei8_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (const float *)(op4), (vuint8mf8_t)(op5), (size_t)(op6)) +#define vluxseg4ei8_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei8_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vbool64_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (const float *)(op9), (vuint8mf8_t)(op10), (size_t)(op11)) +#define vluxseg5ei8_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei8_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (const float *)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vluxseg5ei8_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei8_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vbool32_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (const float *)(op11), (vuint8mf4_t)(op12), (size_t)(op13)) +#define vluxseg5ei8_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei8_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (const float *)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vluxseg5ei8_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei8_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vbool64_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (const float *)(op11), (vuint8mf8_t)(op12), (size_t)(op13)) +#define vluxseg6ei8_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei8_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (const float *)(op6), (vuint8mf4_t)(op7), (size_t)(op8)) +#define vluxseg6ei8_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei8_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vbool32_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (vfloat32m1_t)(op11), (vfloat32m1_t)(op12), (const float *)(op13), (vuint8mf4_t)(op14), (size_t)(op15)) +#define vluxseg6ei8_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei8_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (const float *)(op6), (vuint8mf8_t)(op7), (size_t)(op8)) +#define vluxseg6ei8_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei8_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vbool64_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (vfloat32mf2_t)(op11), (vfloat32mf2_t)(op12), (const float *)(op13), (vuint8mf8_t)(op14), (size_t)(op15)) +#define vluxseg7ei8_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei8_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vfloat32m1_t *)(op6), (const float *)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vluxseg7ei8_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei8_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vfloat32m1_t *)(op6), (vbool32_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (vfloat32m1_t)(op11), (vfloat32m1_t)(op12), (vfloat32m1_t)(op13), (vfloat32m1_t)(op14), (const float *)(op15), (vuint8mf4_t)(op16), (size_t)(op17)) +#define vluxseg7ei8_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei8_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vfloat32mf2_t *)(op6), (const float *)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vluxseg7ei8_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei8_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vfloat32mf2_t *)(op6), (vbool64_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (vfloat32mf2_t)(op11), (vfloat32mf2_t)(op12), (vfloat32mf2_t)(op13), (vfloat32mf2_t)(op14), (const float *)(op15), (vuint8mf8_t)(op16), (size_t)(op17)) +#define vluxseg8ei8_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei8_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vfloat32m1_t *)(op6), (vfloat32m1_t *)(op7), (const float *)(op8), (vuint8mf4_t)(op9), (size_t)(op10)) +#define vluxseg8ei8_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei8_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vfloat32m1_t *)(op6), (vfloat32m1_t *)(op7), (vbool32_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (vfloat32m1_t)(op11), (vfloat32m1_t)(op12), (vfloat32m1_t)(op13), (vfloat32m1_t)(op14), (vfloat32m1_t)(op15), (vfloat32m1_t)(op16), (const float *)(op17), (vuint8mf4_t)(op18), (size_t)(op19)) +#define vluxseg8ei8_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei8_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vfloat32mf2_t *)(op6), (vfloat32mf2_t *)(op7), (const float *)(op8), (vuint8mf8_t)(op9), (size_t)(op10)) +#define vluxseg8ei8_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei8_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vfloat32mf2_t *)(op6), (vfloat32mf2_t *)(op7), (vbool64_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (vfloat32mf2_t)(op11), (vfloat32mf2_t)(op12), (vfloat32mf2_t)(op13), (vfloat32mf2_t)(op14), (vfloat32mf2_t)(op15), (vfloat32mf2_t)(op16), (const float *)(op17), (vuint8mf8_t)(op18), (size_t)(op19)) +#define vluxseg2ei16_v_f32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (const float *)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vbool32_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (const float *)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vluxseg2ei16_v_f32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_f32m2((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (const float *)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_f32m2_m((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vbool16_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (const float *)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vluxseg2ei16_v_f32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_f32m4((vfloat32m4_t *)(op0), (vfloat32m4_t *)(op1), (const float *)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_f32m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_f32m4_m((vfloat32m4_t *)(op0), (vfloat32m4_t *)(op1), (vbool8_t)(op2), (vfloat32m4_t)(op3), (vfloat32m4_t)(op4), (const float *)(op5), (vuint16m2_t)(op6), (size_t)(op7)) +#define vluxseg2ei16_v_f32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (const float *)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vbool64_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (const float *)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vluxseg3ei16_v_f32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei16_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (const float *)(op3), (vuint16mf2_t)(op4), (size_t)(op5)) +#define vluxseg3ei16_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei16_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vbool32_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (const float *)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vluxseg3ei16_v_f32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei16_v_f32m2((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vfloat32m2_t *)(op2), (const float *)(op3), (vuint16m1_t)(op4), (size_t)(op5)) +#define vluxseg3ei16_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei16_v_f32m2_m((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vfloat32m2_t *)(op2), (vbool16_t)(op3), (vfloat32m2_t)(op4), (vfloat32m2_t)(op5), (vfloat32m2_t)(op6), (const float *)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vluxseg3ei16_v_f32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei16_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (const float *)(op3), (vuint16mf4_t)(op4), (size_t)(op5)) +#define vluxseg3ei16_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei16_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vbool64_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (const float *)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vluxseg4ei16_v_f32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei16_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (const float *)(op4), (vuint16mf2_t)(op5), (size_t)(op6)) +#define vluxseg4ei16_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei16_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vbool32_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (const float *)(op9), (vuint16mf2_t)(op10), (size_t)(op11)) +#define vluxseg4ei16_v_f32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei16_v_f32m2((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vfloat32m2_t *)(op2), (vfloat32m2_t *)(op3), (const float *)(op4), (vuint16m1_t)(op5), (size_t)(op6)) +#define vluxseg4ei16_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei16_v_f32m2_m((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vfloat32m2_t *)(op2), (vfloat32m2_t *)(op3), (vbool16_t)(op4), (vfloat32m2_t)(op5), (vfloat32m2_t)(op6), (vfloat32m2_t)(op7), (vfloat32m2_t)(op8), (const float *)(op9), (vuint16m1_t)(op10), (size_t)(op11)) +#define vluxseg4ei16_v_f32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei16_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (const float *)(op4), (vuint16mf4_t)(op5), (size_t)(op6)) +#define vluxseg4ei16_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei16_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vbool64_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (const float *)(op9), (vuint16mf4_t)(op10), (size_t)(op11)) +#define vluxseg5ei16_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei16_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (const float *)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vluxseg5ei16_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei16_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vbool32_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (const float *)(op11), (vuint16mf2_t)(op12), (size_t)(op13)) +#define vluxseg5ei16_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei16_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (const float *)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vluxseg5ei16_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei16_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vbool64_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (const float *)(op11), (vuint16mf4_t)(op12), (size_t)(op13)) +#define vluxseg6ei16_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei16_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (const float *)(op6), (vuint16mf2_t)(op7), (size_t)(op8)) +#define vluxseg6ei16_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei16_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vbool32_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (vfloat32m1_t)(op11), (vfloat32m1_t)(op12), (const float *)(op13), (vuint16mf2_t)(op14), (size_t)(op15)) +#define vluxseg6ei16_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei16_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (const float *)(op6), (vuint16mf4_t)(op7), (size_t)(op8)) +#define vluxseg6ei16_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei16_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vbool64_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (vfloat32mf2_t)(op11), (vfloat32mf2_t)(op12), (const float *)(op13), (vuint16mf4_t)(op14), (size_t)(op15)) +#define vluxseg7ei16_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei16_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vfloat32m1_t *)(op6), (const float *)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vluxseg7ei16_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei16_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vfloat32m1_t *)(op6), (vbool32_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (vfloat32m1_t)(op11), (vfloat32m1_t)(op12), (vfloat32m1_t)(op13), (vfloat32m1_t)(op14), (const float *)(op15), (vuint16mf2_t)(op16), (size_t)(op17)) +#define vluxseg7ei16_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei16_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vfloat32mf2_t *)(op6), (const float *)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vluxseg7ei16_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei16_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vfloat32mf2_t *)(op6), (vbool64_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (vfloat32mf2_t)(op11), (vfloat32mf2_t)(op12), (vfloat32mf2_t)(op13), (vfloat32mf2_t)(op14), (const float *)(op15), (vuint16mf4_t)(op16), (size_t)(op17)) +#define vluxseg8ei16_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei16_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vfloat32m1_t *)(op6), (vfloat32m1_t *)(op7), (const float *)(op8), (vuint16mf2_t)(op9), (size_t)(op10)) +#define vluxseg8ei16_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei16_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vfloat32m1_t *)(op6), (vfloat32m1_t *)(op7), (vbool32_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (vfloat32m1_t)(op11), (vfloat32m1_t)(op12), (vfloat32m1_t)(op13), (vfloat32m1_t)(op14), (vfloat32m1_t)(op15), (vfloat32m1_t)(op16), (const float *)(op17), (vuint16mf2_t)(op18), (size_t)(op19)) +#define vluxseg8ei16_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei16_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vfloat32mf2_t *)(op6), (vfloat32mf2_t *)(op7), (const float *)(op8), (vuint16mf4_t)(op9), (size_t)(op10)) +#define vluxseg8ei16_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei16_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vfloat32mf2_t *)(op6), (vfloat32mf2_t *)(op7), (vbool64_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (vfloat32mf2_t)(op11), (vfloat32mf2_t)(op12), (vfloat32mf2_t)(op13), (vfloat32mf2_t)(op14), (vfloat32mf2_t)(op15), (vfloat32mf2_t)(op16), (const float *)(op17), (vuint16mf4_t)(op18), (size_t)(op19)) +#define vluxseg2ei32_v_f32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (const float *)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vbool32_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (const float *)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vluxseg2ei32_v_f32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_f32m2((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (const float *)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_f32m2_m((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vbool16_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (const float *)(op5), (vuint32m2_t)(op6), (size_t)(op7)) +#define vluxseg2ei32_v_f32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_f32m4((vfloat32m4_t *)(op0), (vfloat32m4_t *)(op1), (const float *)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_f32m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_f32m4_m((vfloat32m4_t *)(op0), (vfloat32m4_t *)(op1), (vbool8_t)(op2), (vfloat32m4_t)(op3), (vfloat32m4_t)(op4), (const float *)(op5), (vuint32m4_t)(op6), (size_t)(op7)) +#define vluxseg2ei32_v_f32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (const float *)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vbool64_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (const float *)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vluxseg3ei32_v_f32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei32_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (const float *)(op3), (vuint32m1_t)(op4), (size_t)(op5)) +#define vluxseg3ei32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei32_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vbool32_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (const float *)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vluxseg3ei32_v_f32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei32_v_f32m2((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vfloat32m2_t *)(op2), (const float *)(op3), (vuint32m2_t)(op4), (size_t)(op5)) +#define vluxseg3ei32_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei32_v_f32m2_m((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vfloat32m2_t *)(op2), (vbool16_t)(op3), (vfloat32m2_t)(op4), (vfloat32m2_t)(op5), (vfloat32m2_t)(op6), (const float *)(op7), (vuint32m2_t)(op8), (size_t)(op9)) +#define vluxseg3ei32_v_f32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei32_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (const float *)(op3), (vuint32mf2_t)(op4), (size_t)(op5)) +#define vluxseg3ei32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei32_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vbool64_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (const float *)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vluxseg4ei32_v_f32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei32_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (const float *)(op4), (vuint32m1_t)(op5), (size_t)(op6)) +#define vluxseg4ei32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei32_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vbool32_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (const float *)(op9), (vuint32m1_t)(op10), (size_t)(op11)) +#define vluxseg4ei32_v_f32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei32_v_f32m2((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vfloat32m2_t *)(op2), (vfloat32m2_t *)(op3), (const float *)(op4), (vuint32m2_t)(op5), (size_t)(op6)) +#define vluxseg4ei32_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei32_v_f32m2_m((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vfloat32m2_t *)(op2), (vfloat32m2_t *)(op3), (vbool16_t)(op4), (vfloat32m2_t)(op5), (vfloat32m2_t)(op6), (vfloat32m2_t)(op7), (vfloat32m2_t)(op8), (const float *)(op9), (vuint32m2_t)(op10), (size_t)(op11)) +#define vluxseg4ei32_v_f32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei32_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (const float *)(op4), (vuint32mf2_t)(op5), (size_t)(op6)) +#define vluxseg4ei32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei32_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vbool64_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (const float *)(op9), (vuint32mf2_t)(op10), (size_t)(op11)) +#define vluxseg5ei32_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei32_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (const float *)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vluxseg5ei32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei32_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vbool32_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (const float *)(op11), (vuint32m1_t)(op12), (size_t)(op13)) +#define vluxseg5ei32_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei32_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (const float *)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vluxseg5ei32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei32_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vbool64_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (const float *)(op11), (vuint32mf2_t)(op12), (size_t)(op13)) +#define vluxseg6ei32_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei32_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (const float *)(op6), (vuint32m1_t)(op7), (size_t)(op8)) +#define vluxseg6ei32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei32_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vbool32_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (vfloat32m1_t)(op11), (vfloat32m1_t)(op12), (const float *)(op13), (vuint32m1_t)(op14), (size_t)(op15)) +#define vluxseg6ei32_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei32_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (const float *)(op6), (vuint32mf2_t)(op7), (size_t)(op8)) +#define vluxseg6ei32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei32_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vbool64_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (vfloat32mf2_t)(op11), (vfloat32mf2_t)(op12), (const float *)(op13), (vuint32mf2_t)(op14), (size_t)(op15)) +#define vluxseg7ei32_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei32_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vfloat32m1_t *)(op6), (const float *)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vluxseg7ei32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei32_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vfloat32m1_t *)(op6), (vbool32_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (vfloat32m1_t)(op11), (vfloat32m1_t)(op12), (vfloat32m1_t)(op13), (vfloat32m1_t)(op14), (const float *)(op15), (vuint32m1_t)(op16), (size_t)(op17)) +#define vluxseg7ei32_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei32_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vfloat32mf2_t *)(op6), (const float *)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vluxseg7ei32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei32_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vfloat32mf2_t *)(op6), (vbool64_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (vfloat32mf2_t)(op11), (vfloat32mf2_t)(op12), (vfloat32mf2_t)(op13), (vfloat32mf2_t)(op14), (const float *)(op15), (vuint32mf2_t)(op16), (size_t)(op17)) +#define vluxseg8ei32_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei32_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vfloat32m1_t *)(op6), (vfloat32m1_t *)(op7), (const float *)(op8), (vuint32m1_t)(op9), (size_t)(op10)) +#define vluxseg8ei32_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei32_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vfloat32m1_t *)(op6), (vfloat32m1_t *)(op7), (vbool32_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (vfloat32m1_t)(op11), (vfloat32m1_t)(op12), (vfloat32m1_t)(op13), (vfloat32m1_t)(op14), (vfloat32m1_t)(op15), (vfloat32m1_t)(op16), (const float *)(op17), (vuint32m1_t)(op18), (size_t)(op19)) +#define vluxseg8ei32_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei32_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vfloat32mf2_t *)(op6), (vfloat32mf2_t *)(op7), (const float *)(op8), (vuint32mf2_t)(op9), (size_t)(op10)) +#define vluxseg8ei32_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei32_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vfloat32mf2_t *)(op6), (vfloat32mf2_t *)(op7), (vbool64_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (vfloat32mf2_t)(op11), (vfloat32mf2_t)(op12), (vfloat32mf2_t)(op13), (vfloat32mf2_t)(op14), (vfloat32mf2_t)(op15), (vfloat32mf2_t)(op16), (const float *)(op17), (vuint32mf2_t)(op18), (size_t)(op19)) +#define vluxseg2ei64_v_f32m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei64_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (const float *)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vluxseg2ei64_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei64_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vbool32_t)(op2), (vfloat32m1_t)(op3), (vfloat32m1_t)(op4), (const float *)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vluxseg2ei64_v_f32m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei64_v_f32m2((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (const float *)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vluxseg2ei64_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei64_v_f32m2_m((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vbool16_t)(op2), (vfloat32m2_t)(op3), (vfloat32m2_t)(op4), (const float *)(op5), (vuint64m4_t)(op6), (size_t)(op7)) +#define vluxseg2ei64_v_f32m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei64_v_f32m4((vfloat32m4_t *)(op0), (vfloat32m4_t *)(op1), (const float *)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vluxseg2ei64_v_f32m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei64_v_f32m4_m((vfloat32m4_t *)(op0), (vfloat32m4_t *)(op1), (vbool8_t)(op2), (vfloat32m4_t)(op3), (vfloat32m4_t)(op4), (const float *)(op5), (vuint64m8_t)(op6), (size_t)(op7)) +#define vluxseg2ei64_v_f32mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei64_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (const float *)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vluxseg2ei64_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei64_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vbool64_t)(op2), (vfloat32mf2_t)(op3), (vfloat32mf2_t)(op4), (const float *)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vluxseg3ei64_v_f32m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei64_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (const float *)(op3), (vuint64m2_t)(op4), (size_t)(op5)) +#define vluxseg3ei64_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei64_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vbool32_t)(op3), (vfloat32m1_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (const float *)(op7), (vuint64m2_t)(op8), (size_t)(op9)) +#define vluxseg3ei64_v_f32m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei64_v_f32m2((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vfloat32m2_t *)(op2), (const float *)(op3), (vuint64m4_t)(op4), (size_t)(op5)) +#define vluxseg3ei64_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei64_v_f32m2_m((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vfloat32m2_t *)(op2), (vbool16_t)(op3), (vfloat32m2_t)(op4), (vfloat32m2_t)(op5), (vfloat32m2_t)(op6), (const float *)(op7), (vuint64m4_t)(op8), (size_t)(op9)) +#define vluxseg3ei64_v_f32mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei64_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (const float *)(op3), (vuint64m1_t)(op4), (size_t)(op5)) +#define vluxseg3ei64_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei64_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vbool64_t)(op3), (vfloat32mf2_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (const float *)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vluxseg4ei64_v_f32m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei64_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (const float *)(op4), (vuint64m2_t)(op5), (size_t)(op6)) +#define vluxseg4ei64_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei64_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vbool32_t)(op4), (vfloat32m1_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (const float *)(op9), (vuint64m2_t)(op10), (size_t)(op11)) +#define vluxseg4ei64_v_f32m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei64_v_f32m2((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vfloat32m2_t *)(op2), (vfloat32m2_t *)(op3), (const float *)(op4), (vuint64m4_t)(op5), (size_t)(op6)) +#define vluxseg4ei64_v_f32m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei64_v_f32m2_m((vfloat32m2_t *)(op0), (vfloat32m2_t *)(op1), (vfloat32m2_t *)(op2), (vfloat32m2_t *)(op3), (vbool16_t)(op4), (vfloat32m2_t)(op5), (vfloat32m2_t)(op6), (vfloat32m2_t)(op7), (vfloat32m2_t)(op8), (const float *)(op9), (vuint64m4_t)(op10), (size_t)(op11)) +#define vluxseg4ei64_v_f32mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei64_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (const float *)(op4), (vuint64m1_t)(op5), (size_t)(op6)) +#define vluxseg4ei64_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei64_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vbool64_t)(op4), (vfloat32mf2_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (const float *)(op9), (vuint64m1_t)(op10), (size_t)(op11)) +#define vluxseg5ei64_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei64_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (const float *)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vluxseg5ei64_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei64_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vbool32_t)(op5), (vfloat32m1_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (const float *)(op11), (vuint64m2_t)(op12), (size_t)(op13)) +#define vluxseg5ei64_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei64_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (const float *)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vluxseg5ei64_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei64_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vbool64_t)(op5), (vfloat32mf2_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (const float *)(op11), (vuint64m1_t)(op12), (size_t)(op13)) +#define vluxseg6ei64_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei64_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (const float *)(op6), (vuint64m2_t)(op7), (size_t)(op8)) +#define vluxseg6ei64_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei64_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vbool32_t)(op6), (vfloat32m1_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (vfloat32m1_t)(op11), (vfloat32m1_t)(op12), (const float *)(op13), (vuint64m2_t)(op14), (size_t)(op15)) +#define vluxseg6ei64_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei64_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (const float *)(op6), (vuint64m1_t)(op7), (size_t)(op8)) +#define vluxseg6ei64_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei64_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vbool64_t)(op6), (vfloat32mf2_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (vfloat32mf2_t)(op11), (vfloat32mf2_t)(op12), (const float *)(op13), (vuint64m1_t)(op14), (size_t)(op15)) +#define vluxseg7ei64_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei64_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vfloat32m1_t *)(op6), (const float *)(op7), (vuint64m2_t)(op8), (size_t)(op9)) +#define vluxseg7ei64_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei64_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vfloat32m1_t *)(op6), (vbool32_t)(op7), (vfloat32m1_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (vfloat32m1_t)(op11), (vfloat32m1_t)(op12), (vfloat32m1_t)(op13), (vfloat32m1_t)(op14), (const float *)(op15), (vuint64m2_t)(op16), (size_t)(op17)) +#define vluxseg7ei64_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei64_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vfloat32mf2_t *)(op6), (const float *)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vluxseg7ei64_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei64_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vfloat32mf2_t *)(op6), (vbool64_t)(op7), (vfloat32mf2_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (vfloat32mf2_t)(op11), (vfloat32mf2_t)(op12), (vfloat32mf2_t)(op13), (vfloat32mf2_t)(op14), (const float *)(op15), (vuint64m1_t)(op16), (size_t)(op17)) +#define vluxseg8ei64_v_f32m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei64_v_f32m1((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vfloat32m1_t *)(op6), (vfloat32m1_t *)(op7), (const float *)(op8), (vuint64m2_t)(op9), (size_t)(op10)) +#define vluxseg8ei64_v_f32m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei64_v_f32m1_m((vfloat32m1_t *)(op0), (vfloat32m1_t *)(op1), (vfloat32m1_t *)(op2), (vfloat32m1_t *)(op3), (vfloat32m1_t *)(op4), (vfloat32m1_t *)(op5), (vfloat32m1_t *)(op6), (vfloat32m1_t *)(op7), (vbool32_t)(op8), (vfloat32m1_t)(op9), (vfloat32m1_t)(op10), (vfloat32m1_t)(op11), (vfloat32m1_t)(op12), (vfloat32m1_t)(op13), (vfloat32m1_t)(op14), (vfloat32m1_t)(op15), (vfloat32m1_t)(op16), (const float *)(op17), (vuint64m2_t)(op18), (size_t)(op19)) +#define vluxseg8ei64_v_f32mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei64_v_f32mf2((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vfloat32mf2_t *)(op6), (vfloat32mf2_t *)(op7), (const float *)(op8), (vuint64m1_t)(op9), (size_t)(op10)) +#define vluxseg8ei64_v_f32mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei64_v_f32mf2_m((vfloat32mf2_t *)(op0), (vfloat32mf2_t *)(op1), (vfloat32mf2_t *)(op2), (vfloat32mf2_t *)(op3), (vfloat32mf2_t *)(op4), (vfloat32mf2_t *)(op5), (vfloat32mf2_t *)(op6), (vfloat32mf2_t *)(op7), (vbool64_t)(op8), (vfloat32mf2_t)(op9), (vfloat32mf2_t)(op10), (vfloat32mf2_t)(op11), (vfloat32mf2_t)(op12), (vfloat32mf2_t)(op13), (vfloat32mf2_t)(op14), (vfloat32mf2_t)(op15), (vfloat32mf2_t)(op16), (const float *)(op17), (vuint64m1_t)(op18), (size_t)(op19)) +#endif + +#if defined(__riscv_d) && defined(__riscv_zvlsseg) +#define vloxseg2ei8_v_f64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (const double *)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vbool64_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (const double *)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vloxseg2ei8_v_f64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_f64m2((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (const double *)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_f64m2_m((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vbool32_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (const double *)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vloxseg2ei8_v_f64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_f64m4((vfloat64m4_t *)(op0), (vfloat64m4_t *)(op1), (const double *)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_f64m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_f64m4_m((vfloat64m4_t *)(op0), (vfloat64m4_t *)(op1), (vbool16_t)(op2), (vfloat64m4_t)(op3), (vfloat64m4_t)(op4), (const double *)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vloxseg3ei8_v_f64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei8_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (const double *)(op3), (vuint8mf8_t)(op4), (size_t)(op5)) +#define vloxseg3ei8_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei8_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vbool64_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (const double *)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vloxseg3ei8_v_f64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei8_v_f64m2((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vfloat64m2_t *)(op2), (const double *)(op3), (vuint8mf4_t)(op4), (size_t)(op5)) +#define vloxseg3ei8_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei8_v_f64m2_m((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vfloat64m2_t *)(op2), (vbool32_t)(op3), (vfloat64m2_t)(op4), (vfloat64m2_t)(op5), (vfloat64m2_t)(op6), (const double *)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vloxseg4ei8_v_f64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei8_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (const double *)(op4), (vuint8mf8_t)(op5), (size_t)(op6)) +#define vloxseg4ei8_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei8_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vbool64_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (const double *)(op9), (vuint8mf8_t)(op10), (size_t)(op11)) +#define vloxseg4ei8_v_f64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei8_v_f64m2((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vfloat64m2_t *)(op2), (vfloat64m2_t *)(op3), (const double *)(op4), (vuint8mf4_t)(op5), (size_t)(op6)) +#define vloxseg4ei8_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei8_v_f64m2_m((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vfloat64m2_t *)(op2), (vfloat64m2_t *)(op3), (vbool32_t)(op4), (vfloat64m2_t)(op5), (vfloat64m2_t)(op6), (vfloat64m2_t)(op7), (vfloat64m2_t)(op8), (const double *)(op9), (vuint8mf4_t)(op10), (size_t)(op11)) +#define vloxseg5ei8_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei8_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (const double *)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vloxseg5ei8_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei8_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vbool64_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (const double *)(op11), (vuint8mf8_t)(op12), (size_t)(op13)) +#define vloxseg6ei8_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei8_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (const double *)(op6), (vuint8mf8_t)(op7), (size_t)(op8)) +#define vloxseg6ei8_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei8_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vbool64_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (vfloat64m1_t)(op11), (vfloat64m1_t)(op12), (const double *)(op13), (vuint8mf8_t)(op14), (size_t)(op15)) +#define vloxseg7ei8_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei8_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vfloat64m1_t *)(op6), (const double *)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vloxseg7ei8_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei8_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vfloat64m1_t *)(op6), (vbool64_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (vfloat64m1_t)(op11), (vfloat64m1_t)(op12), (vfloat64m1_t)(op13), (vfloat64m1_t)(op14), (const double *)(op15), (vuint8mf8_t)(op16), (size_t)(op17)) +#define vloxseg8ei8_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei8_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vfloat64m1_t *)(op6), (vfloat64m1_t *)(op7), (const double *)(op8), (vuint8mf8_t)(op9), (size_t)(op10)) +#define vloxseg8ei8_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei8_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vfloat64m1_t *)(op6), (vfloat64m1_t *)(op7), (vbool64_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (vfloat64m1_t)(op11), (vfloat64m1_t)(op12), (vfloat64m1_t)(op13), (vfloat64m1_t)(op14), (vfloat64m1_t)(op15), (vfloat64m1_t)(op16), (const double *)(op17), (vuint8mf8_t)(op18), (size_t)(op19)) +#define vloxseg2ei16_v_f64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (const double *)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vbool64_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (const double *)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vloxseg2ei16_v_f64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_f64m2((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (const double *)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_f64m2_m((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vbool32_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (const double *)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vloxseg2ei16_v_f64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_f64m4((vfloat64m4_t *)(op0), (vfloat64m4_t *)(op1), (const double *)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_f64m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_f64m4_m((vfloat64m4_t *)(op0), (vfloat64m4_t *)(op1), (vbool16_t)(op2), (vfloat64m4_t)(op3), (vfloat64m4_t)(op4), (const double *)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vloxseg3ei16_v_f64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei16_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (const double *)(op3), (vuint16mf4_t)(op4), (size_t)(op5)) +#define vloxseg3ei16_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei16_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vbool64_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (const double *)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vloxseg3ei16_v_f64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei16_v_f64m2((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vfloat64m2_t *)(op2), (const double *)(op3), (vuint16mf2_t)(op4), (size_t)(op5)) +#define vloxseg3ei16_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei16_v_f64m2_m((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vfloat64m2_t *)(op2), (vbool32_t)(op3), (vfloat64m2_t)(op4), (vfloat64m2_t)(op5), (vfloat64m2_t)(op6), (const double *)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vloxseg4ei16_v_f64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei16_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (const double *)(op4), (vuint16mf4_t)(op5), (size_t)(op6)) +#define vloxseg4ei16_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei16_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vbool64_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (const double *)(op9), (vuint16mf4_t)(op10), (size_t)(op11)) +#define vloxseg4ei16_v_f64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei16_v_f64m2((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vfloat64m2_t *)(op2), (vfloat64m2_t *)(op3), (const double *)(op4), (vuint16mf2_t)(op5), (size_t)(op6)) +#define vloxseg4ei16_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei16_v_f64m2_m((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vfloat64m2_t *)(op2), (vfloat64m2_t *)(op3), (vbool32_t)(op4), (vfloat64m2_t)(op5), (vfloat64m2_t)(op6), (vfloat64m2_t)(op7), (vfloat64m2_t)(op8), (const double *)(op9), (vuint16mf2_t)(op10), (size_t)(op11)) +#define vloxseg5ei16_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei16_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (const double *)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vloxseg5ei16_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei16_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vbool64_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (const double *)(op11), (vuint16mf4_t)(op12), (size_t)(op13)) +#define vloxseg6ei16_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei16_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (const double *)(op6), (vuint16mf4_t)(op7), (size_t)(op8)) +#define vloxseg6ei16_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei16_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vbool64_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (vfloat64m1_t)(op11), (vfloat64m1_t)(op12), (const double *)(op13), (vuint16mf4_t)(op14), (size_t)(op15)) +#define vloxseg7ei16_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei16_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vfloat64m1_t *)(op6), (const double *)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vloxseg7ei16_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei16_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vfloat64m1_t *)(op6), (vbool64_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (vfloat64m1_t)(op11), (vfloat64m1_t)(op12), (vfloat64m1_t)(op13), (vfloat64m1_t)(op14), (const double *)(op15), (vuint16mf4_t)(op16), (size_t)(op17)) +#define vloxseg8ei16_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei16_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vfloat64m1_t *)(op6), (vfloat64m1_t *)(op7), (const double *)(op8), (vuint16mf4_t)(op9), (size_t)(op10)) +#define vloxseg8ei16_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei16_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vfloat64m1_t *)(op6), (vfloat64m1_t *)(op7), (vbool64_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (vfloat64m1_t)(op11), (vfloat64m1_t)(op12), (vfloat64m1_t)(op13), (vfloat64m1_t)(op14), (vfloat64m1_t)(op15), (vfloat64m1_t)(op16), (const double *)(op17), (vuint16mf4_t)(op18), (size_t)(op19)) +#define vloxseg2ei32_v_f64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (const double *)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vbool64_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (const double *)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vloxseg2ei32_v_f64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_f64m2((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (const double *)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_f64m2_m((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vbool32_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (const double *)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vloxseg2ei32_v_f64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_f64m4((vfloat64m4_t *)(op0), (vfloat64m4_t *)(op1), (const double *)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_f64m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_f64m4_m((vfloat64m4_t *)(op0), (vfloat64m4_t *)(op1), (vbool16_t)(op2), (vfloat64m4_t)(op3), (vfloat64m4_t)(op4), (const double *)(op5), (vuint32m2_t)(op6), (size_t)(op7)) +#define vloxseg3ei32_v_f64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei32_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (const double *)(op3), (vuint32mf2_t)(op4), (size_t)(op5)) +#define vloxseg3ei32_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei32_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vbool64_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (const double *)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vloxseg3ei32_v_f64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei32_v_f64m2((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vfloat64m2_t *)(op2), (const double *)(op3), (vuint32m1_t)(op4), (size_t)(op5)) +#define vloxseg3ei32_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei32_v_f64m2_m((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vfloat64m2_t *)(op2), (vbool32_t)(op3), (vfloat64m2_t)(op4), (vfloat64m2_t)(op5), (vfloat64m2_t)(op6), (const double *)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vloxseg4ei32_v_f64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei32_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (const double *)(op4), (vuint32mf2_t)(op5), (size_t)(op6)) +#define vloxseg4ei32_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei32_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vbool64_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (const double *)(op9), (vuint32mf2_t)(op10), (size_t)(op11)) +#define vloxseg4ei32_v_f64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei32_v_f64m2((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vfloat64m2_t *)(op2), (vfloat64m2_t *)(op3), (const double *)(op4), (vuint32m1_t)(op5), (size_t)(op6)) +#define vloxseg4ei32_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei32_v_f64m2_m((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vfloat64m2_t *)(op2), (vfloat64m2_t *)(op3), (vbool32_t)(op4), (vfloat64m2_t)(op5), (vfloat64m2_t)(op6), (vfloat64m2_t)(op7), (vfloat64m2_t)(op8), (const double *)(op9), (vuint32m1_t)(op10), (size_t)(op11)) +#define vloxseg5ei32_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei32_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (const double *)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vloxseg5ei32_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei32_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vbool64_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (const double *)(op11), (vuint32mf2_t)(op12), (size_t)(op13)) +#define vloxseg6ei32_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei32_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (const double *)(op6), (vuint32mf2_t)(op7), (size_t)(op8)) +#define vloxseg6ei32_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei32_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vbool64_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (vfloat64m1_t)(op11), (vfloat64m1_t)(op12), (const double *)(op13), (vuint32mf2_t)(op14), (size_t)(op15)) +#define vloxseg7ei32_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei32_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vfloat64m1_t *)(op6), (const double *)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vloxseg7ei32_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei32_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vfloat64m1_t *)(op6), (vbool64_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (vfloat64m1_t)(op11), (vfloat64m1_t)(op12), (vfloat64m1_t)(op13), (vfloat64m1_t)(op14), (const double *)(op15), (vuint32mf2_t)(op16), (size_t)(op17)) +#define vloxseg8ei32_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei32_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vfloat64m1_t *)(op6), (vfloat64m1_t *)(op7), (const double *)(op8), (vuint32mf2_t)(op9), (size_t)(op10)) +#define vloxseg8ei32_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei32_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vfloat64m1_t *)(op6), (vfloat64m1_t *)(op7), (vbool64_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (vfloat64m1_t)(op11), (vfloat64m1_t)(op12), (vfloat64m1_t)(op13), (vfloat64m1_t)(op14), (vfloat64m1_t)(op15), (vfloat64m1_t)(op16), (const double *)(op17), (vuint32mf2_t)(op18), (size_t)(op19)) +#define vloxseg2ei64_v_f64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei64_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (const double *)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vloxseg2ei64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei64_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vbool64_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (const double *)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vloxseg2ei64_v_f64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei64_v_f64m2((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (const double *)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vloxseg2ei64_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei64_v_f64m2_m((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vbool32_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (const double *)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vloxseg2ei64_v_f64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei64_v_f64m4((vfloat64m4_t *)(op0), (vfloat64m4_t *)(op1), (const double *)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vloxseg2ei64_v_f64m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei64_v_f64m4_m((vfloat64m4_t *)(op0), (vfloat64m4_t *)(op1), (vbool16_t)(op2), (vfloat64m4_t)(op3), (vfloat64m4_t)(op4), (const double *)(op5), (vuint64m4_t)(op6), (size_t)(op7)) +#define vloxseg3ei64_v_f64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei64_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (const double *)(op3), (vuint64m1_t)(op4), (size_t)(op5)) +#define vloxseg3ei64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei64_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vbool64_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (const double *)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vloxseg3ei64_v_f64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei64_v_f64m2((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vfloat64m2_t *)(op2), (const double *)(op3), (vuint64m2_t)(op4), (size_t)(op5)) +#define vloxseg3ei64_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei64_v_f64m2_m((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vfloat64m2_t *)(op2), (vbool32_t)(op3), (vfloat64m2_t)(op4), (vfloat64m2_t)(op5), (vfloat64m2_t)(op6), (const double *)(op7), (vuint64m2_t)(op8), (size_t)(op9)) +#define vloxseg4ei64_v_f64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei64_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (const double *)(op4), (vuint64m1_t)(op5), (size_t)(op6)) +#define vloxseg4ei64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei64_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vbool64_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (const double *)(op9), (vuint64m1_t)(op10), (size_t)(op11)) +#define vloxseg4ei64_v_f64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei64_v_f64m2((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vfloat64m2_t *)(op2), (vfloat64m2_t *)(op3), (const double *)(op4), (vuint64m2_t)(op5), (size_t)(op6)) +#define vloxseg4ei64_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei64_v_f64m2_m((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vfloat64m2_t *)(op2), (vfloat64m2_t *)(op3), (vbool32_t)(op4), (vfloat64m2_t)(op5), (vfloat64m2_t)(op6), (vfloat64m2_t)(op7), (vfloat64m2_t)(op8), (const double *)(op9), (vuint64m2_t)(op10), (size_t)(op11)) +#define vloxseg5ei64_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei64_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (const double *)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vloxseg5ei64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei64_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vbool64_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (const double *)(op11), (vuint64m1_t)(op12), (size_t)(op13)) +#define vloxseg6ei64_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei64_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (const double *)(op6), (vuint64m1_t)(op7), (size_t)(op8)) +#define vloxseg6ei64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei64_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vbool64_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (vfloat64m1_t)(op11), (vfloat64m1_t)(op12), (const double *)(op13), (vuint64m1_t)(op14), (size_t)(op15)) +#define vloxseg7ei64_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei64_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vfloat64m1_t *)(op6), (const double *)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vloxseg7ei64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei64_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vfloat64m1_t *)(op6), (vbool64_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (vfloat64m1_t)(op11), (vfloat64m1_t)(op12), (vfloat64m1_t)(op13), (vfloat64m1_t)(op14), (const double *)(op15), (vuint64m1_t)(op16), (size_t)(op17)) +#define vloxseg8ei64_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei64_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vfloat64m1_t *)(op6), (vfloat64m1_t *)(op7), (const double *)(op8), (vuint64m1_t)(op9), (size_t)(op10)) +#define vloxseg8ei64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei64_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vfloat64m1_t *)(op6), (vfloat64m1_t *)(op7), (vbool64_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (vfloat64m1_t)(op11), (vfloat64m1_t)(op12), (vfloat64m1_t)(op13), (vfloat64m1_t)(op14), (vfloat64m1_t)(op15), (vfloat64m1_t)(op16), (const double *)(op17), (vuint64m1_t)(op18), (size_t)(op19)) +#define vsseg2e64_v_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e64_v_f64m1((double *)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (size_t)(op3)) +#define vsseg2e64_v_f64m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e64_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vsseg2e64_v_f64m2(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e64_v_f64m2((double *)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (size_t)(op3)) +#define vsseg2e64_v_f64m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e64_v_f64m2_m((vbool32_t)(op0), (double *)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vsseg2e64_v_f64m4(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e64_v_f64m4((double *)(op0), (vfloat64m4_t)(op1), (vfloat64m4_t)(op2), (size_t)(op3)) +#define vsseg2e64_v_f64m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e64_v_f64m4_m((vbool16_t)(op0), (double *)(op1), (vfloat64m4_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vsseg3e64_v_f64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg3e64_v_f64m1((double *)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vsseg3e64_v_f64m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg3e64_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (size_t)(op5)) +#define vsseg3e64_v_f64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg3e64_v_f64m2((double *)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vsseg3e64_v_f64m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg3e64_v_f64m2_m((vbool32_t)(op0), (double *)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (size_t)(op5)) +#define vsseg4e64_v_f64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg4e64_v_f64m1((double *)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (size_t)(op5)) +#define vsseg4e64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg4e64_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (size_t)(op6)) +#define vsseg4e64_v_f64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg4e64_v_f64m2((double *)(op0), (vfloat64m2_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (size_t)(op5)) +#define vsseg4e64_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg4e64_v_f64m2_m((vbool32_t)(op0), (double *)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (vfloat64m2_t)(op5), (size_t)(op6)) +#define vsseg5e64_v_f64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg5e64_v_f64m1((double *)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (size_t)(op6)) +#define vsseg5e64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg5e64_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (size_t)(op7)) +#define vsseg6e64_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg6e64_v_f64m1((double *)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (size_t)(op7)) +#define vsseg6e64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg6e64_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (size_t)(op8)) +#define vsseg7e64_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg7e64_v_f64m1((double *)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (size_t)(op8)) +#define vsseg7e64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg7e64_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (size_t)(op9)) +#define vsseg8e64_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg8e64_v_f64m1((double *)(op0), (vfloat64m1_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (size_t)(op9)) +#define vsseg8e64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsseg8e64_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (size_t)(op10)) +#define vssseg2e64_v_f64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e64_v_f64m1((double *)(op0), (ptrdiff_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vssseg2e64_v_f64m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e64_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (ptrdiff_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (size_t)(op5)) +#define vssseg2e64_v_f64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e64_v_f64m2((double *)(op0), (ptrdiff_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vssseg2e64_v_f64m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e64_v_f64m2_m((vbool32_t)(op0), (double *)(op1), (ptrdiff_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (size_t)(op5)) +#define vssseg2e64_v_f64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e64_v_f64m4((double *)(op0), (ptrdiff_t)(op1), (vfloat64m4_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vssseg2e64_v_f64m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e64_v_f64m4_m((vbool16_t)(op0), (double *)(op1), (ptrdiff_t)(op2), (vfloat64m4_t)(op3), (vfloat64m4_t)(op4), (size_t)(op5)) +#define vssseg3e64_v_f64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg3e64_v_f64m1((double *)(op0), (ptrdiff_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (size_t)(op5)) +#define vssseg3e64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg3e64_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (ptrdiff_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (size_t)(op6)) +#define vssseg3e64_v_f64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg3e64_v_f64m2((double *)(op0), (ptrdiff_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (size_t)(op5)) +#define vssseg3e64_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg3e64_v_f64m2_m((vbool32_t)(op0), (double *)(op1), (ptrdiff_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (vfloat64m2_t)(op5), (size_t)(op6)) +#define vssseg4e64_v_f64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg4e64_v_f64m1((double *)(op0), (ptrdiff_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (size_t)(op6)) +#define vssseg4e64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg4e64_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (ptrdiff_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (size_t)(op7)) +#define vssseg4e64_v_f64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg4e64_v_f64m2((double *)(op0), (ptrdiff_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (vfloat64m2_t)(op5), (size_t)(op6)) +#define vssseg4e64_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg4e64_v_f64m2_m((vbool32_t)(op0), (double *)(op1), (ptrdiff_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (vfloat64m2_t)(op5), (vfloat64m2_t)(op6), (size_t)(op7)) +#define vssseg5e64_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg5e64_v_f64m1((double *)(op0), (ptrdiff_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (size_t)(op7)) +#define vssseg5e64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg5e64_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (ptrdiff_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (size_t)(op8)) +#define vssseg6e64_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg6e64_v_f64m1((double *)(op0), (ptrdiff_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (size_t)(op8)) +#define vssseg6e64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg6e64_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (ptrdiff_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (size_t)(op9)) +#define vssseg7e64_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg7e64_v_f64m1((double *)(op0), (ptrdiff_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (size_t)(op9)) +#define vssseg7e64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg7e64_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (ptrdiff_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (size_t)(op10)) +#define vssseg8e64_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg8e64_v_f64m1((double *)(op0), (ptrdiff_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (size_t)(op10)) +#define vssseg8e64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vssseg8e64_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (ptrdiff_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (size_t)(op11)) +#define vsuxseg2ei8_v_f64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_f64m1((double *)(op0), (vuint8mf8_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_f64m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint8mf8_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (size_t)(op5)) +#define vsuxseg2ei8_v_f64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_f64m2((double *)(op0), (vuint8mf4_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_f64m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_f64m2_m((vbool32_t)(op0), (double *)(op1), (vuint8mf4_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei8_v_f64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_f64m4((double *)(op0), (vuint8mf2_t)(op1), (vfloat64m4_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_f64m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_f64m4_m((vbool16_t)(op0), (double *)(op1), (vuint8mf2_t)(op2), (vfloat64m4_t)(op3), (vfloat64m4_t)(op4), (size_t)(op5)) +#define vsuxseg3ei8_v_f64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei8_v_f64m1((double *)(op0), (vuint8mf8_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (size_t)(op5)) +#define vsuxseg3ei8_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei8_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint8mf8_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (size_t)(op6)) +#define vsuxseg3ei8_v_f64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei8_v_f64m2((double *)(op0), (vuint8mf4_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei8_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei8_v_f64m2_m((vbool32_t)(op0), (double *)(op1), (vuint8mf4_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (vfloat64m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei8_v_f64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei8_v_f64m1((double *)(op0), (vuint8mf8_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (size_t)(op6)) +#define vsuxseg4ei8_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei8_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint8mf8_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (size_t)(op7)) +#define vsuxseg4ei8_v_f64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei8_v_f64m2((double *)(op0), (vuint8mf4_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (vfloat64m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei8_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei8_v_f64m2_m((vbool32_t)(op0), (double *)(op1), (vuint8mf4_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (vfloat64m2_t)(op5), (vfloat64m2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei8_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei8_v_f64m1((double *)(op0), (vuint8mf8_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (size_t)(op7)) +#define vsuxseg5ei8_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei8_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint8mf8_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei8_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei8_v_f64m1((double *)(op0), (vuint8mf8_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei8_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei8_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint8mf8_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei8_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei8_v_f64m1((double *)(op0), (vuint8mf8_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei8_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei8_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint8mf8_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei8_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei8_v_f64m1((double *)(op0), (vuint8mf8_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei8_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei8_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint8mf8_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (size_t)(op11)) +#define vsuxseg2ei16_v_f64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_f64m1((double *)(op0), (vuint16mf4_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_f64m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint16mf4_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (size_t)(op5)) +#define vsuxseg2ei16_v_f64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_f64m2((double *)(op0), (vuint16mf2_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_f64m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_f64m2_m((vbool32_t)(op0), (double *)(op1), (vuint16mf2_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei16_v_f64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_f64m4((double *)(op0), (vuint16m1_t)(op1), (vfloat64m4_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_f64m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_f64m4_m((vbool16_t)(op0), (double *)(op1), (vuint16m1_t)(op2), (vfloat64m4_t)(op3), (vfloat64m4_t)(op4), (size_t)(op5)) +#define vsuxseg3ei16_v_f64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei16_v_f64m1((double *)(op0), (vuint16mf4_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (size_t)(op5)) +#define vsuxseg3ei16_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei16_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint16mf4_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (size_t)(op6)) +#define vsuxseg3ei16_v_f64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei16_v_f64m2((double *)(op0), (vuint16mf2_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei16_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei16_v_f64m2_m((vbool32_t)(op0), (double *)(op1), (vuint16mf2_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (vfloat64m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei16_v_f64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei16_v_f64m1((double *)(op0), (vuint16mf4_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (size_t)(op6)) +#define vsuxseg4ei16_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei16_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint16mf4_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (size_t)(op7)) +#define vsuxseg4ei16_v_f64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei16_v_f64m2((double *)(op0), (vuint16mf2_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (vfloat64m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei16_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei16_v_f64m2_m((vbool32_t)(op0), (double *)(op1), (vuint16mf2_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (vfloat64m2_t)(op5), (vfloat64m2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei16_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei16_v_f64m1((double *)(op0), (vuint16mf4_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (size_t)(op7)) +#define vsuxseg5ei16_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei16_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint16mf4_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei16_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei16_v_f64m1((double *)(op0), (vuint16mf4_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei16_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei16_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint16mf4_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei16_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei16_v_f64m1((double *)(op0), (vuint16mf4_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei16_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei16_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint16mf4_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei16_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei16_v_f64m1((double *)(op0), (vuint16mf4_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei16_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei16_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint16mf4_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (size_t)(op11)) +#define vsuxseg2ei32_v_f64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_f64m1((double *)(op0), (vuint32mf2_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_f64m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint32mf2_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (size_t)(op5)) +#define vsuxseg2ei32_v_f64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_f64m2((double *)(op0), (vuint32m1_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_f64m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_f64m2_m((vbool32_t)(op0), (double *)(op1), (vuint32m1_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei32_v_f64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_f64m4((double *)(op0), (vuint32m2_t)(op1), (vfloat64m4_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_f64m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_f64m4_m((vbool16_t)(op0), (double *)(op1), (vuint32m2_t)(op2), (vfloat64m4_t)(op3), (vfloat64m4_t)(op4), (size_t)(op5)) +#define vsuxseg3ei32_v_f64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei32_v_f64m1((double *)(op0), (vuint32mf2_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (size_t)(op5)) +#define vsuxseg3ei32_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei32_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint32mf2_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (size_t)(op6)) +#define vsuxseg3ei32_v_f64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei32_v_f64m2((double *)(op0), (vuint32m1_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei32_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei32_v_f64m2_m((vbool32_t)(op0), (double *)(op1), (vuint32m1_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (vfloat64m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei32_v_f64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei32_v_f64m1((double *)(op0), (vuint32mf2_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (size_t)(op6)) +#define vsuxseg4ei32_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei32_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint32mf2_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (size_t)(op7)) +#define vsuxseg4ei32_v_f64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei32_v_f64m2((double *)(op0), (vuint32m1_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (vfloat64m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei32_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei32_v_f64m2_m((vbool32_t)(op0), (double *)(op1), (vuint32m1_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (vfloat64m2_t)(op5), (vfloat64m2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei32_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei32_v_f64m1((double *)(op0), (vuint32mf2_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (size_t)(op7)) +#define vsuxseg5ei32_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei32_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint32mf2_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei32_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei32_v_f64m1((double *)(op0), (vuint32mf2_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei32_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei32_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint32mf2_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei32_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei32_v_f64m1((double *)(op0), (vuint32mf2_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei32_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei32_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint32mf2_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei32_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei32_v_f64m1((double *)(op0), (vuint32mf2_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei32_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei32_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint32mf2_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (size_t)(op11)) +#define vsuxseg2ei64_v_f64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei64_v_f64m1((double *)(op0), (vuint64m1_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vsuxseg2ei64_v_f64m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei64_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (size_t)(op5)) +#define vsuxseg2ei64_v_f64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei64_v_f64m2((double *)(op0), (vuint64m2_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei64_v_f64m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei64_v_f64m2_m((vbool32_t)(op0), (double *)(op1), (vuint64m2_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei64_v_f64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei64_v_f64m4((double *)(op0), (vuint64m4_t)(op1), (vfloat64m4_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei64_v_f64m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei64_v_f64m4_m((vbool16_t)(op0), (double *)(op1), (vuint64m4_t)(op2), (vfloat64m4_t)(op3), (vfloat64m4_t)(op4), (size_t)(op5)) +#define vsuxseg3ei64_v_f64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei64_v_f64m1((double *)(op0), (vuint64m1_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (size_t)(op5)) +#define vsuxseg3ei64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei64_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (size_t)(op6)) +#define vsuxseg3ei64_v_f64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei64_v_f64m2((double *)(op0), (vuint64m2_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei64_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei64_v_f64m2_m((vbool32_t)(op0), (double *)(op1), (vuint64m2_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (vfloat64m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei64_v_f64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei64_v_f64m1((double *)(op0), (vuint64m1_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (size_t)(op6)) +#define vsuxseg4ei64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei64_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (size_t)(op7)) +#define vsuxseg4ei64_v_f64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei64_v_f64m2((double *)(op0), (vuint64m2_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (vfloat64m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei64_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei64_v_f64m2_m((vbool32_t)(op0), (double *)(op1), (vuint64m2_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (vfloat64m2_t)(op5), (vfloat64m2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei64_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei64_v_f64m1((double *)(op0), (vuint64m1_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (size_t)(op7)) +#define vsuxseg5ei64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei64_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei64_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei64_v_f64m1((double *)(op0), (vuint64m1_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei64_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei64_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei64_v_f64m1((double *)(op0), (vuint64m1_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei64_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei64_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei64_v_f64m1((double *)(op0), (vuint64m1_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei64_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (size_t)(op11)) +#define vsoxseg2ei8_v_f64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_f64m1((double *)(op0), (vuint8mf8_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_f64m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint8mf8_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (size_t)(op5)) +#define vsoxseg2ei8_v_f64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_f64m2((double *)(op0), (vuint8mf4_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_f64m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_f64m2_m((vbool32_t)(op0), (double *)(op1), (vuint8mf4_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei8_v_f64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_f64m4((double *)(op0), (vuint8mf2_t)(op1), (vfloat64m4_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_f64m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_f64m4_m((vbool16_t)(op0), (double *)(op1), (vuint8mf2_t)(op2), (vfloat64m4_t)(op3), (vfloat64m4_t)(op4), (size_t)(op5)) +#define vsoxseg3ei8_v_f64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei8_v_f64m1((double *)(op0), (vuint8mf8_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (size_t)(op5)) +#define vsoxseg3ei8_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei8_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint8mf8_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (size_t)(op6)) +#define vsoxseg3ei8_v_f64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei8_v_f64m2((double *)(op0), (vuint8mf4_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei8_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei8_v_f64m2_m((vbool32_t)(op0), (double *)(op1), (vuint8mf4_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (vfloat64m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei8_v_f64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei8_v_f64m1((double *)(op0), (vuint8mf8_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (size_t)(op6)) +#define vsoxseg4ei8_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei8_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint8mf8_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (size_t)(op7)) +#define vsoxseg4ei8_v_f64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei8_v_f64m2((double *)(op0), (vuint8mf4_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (vfloat64m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei8_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei8_v_f64m2_m((vbool32_t)(op0), (double *)(op1), (vuint8mf4_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (vfloat64m2_t)(op5), (vfloat64m2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei8_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei8_v_f64m1((double *)(op0), (vuint8mf8_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (size_t)(op7)) +#define vsoxseg5ei8_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei8_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint8mf8_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei8_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei8_v_f64m1((double *)(op0), (vuint8mf8_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei8_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei8_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint8mf8_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei8_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei8_v_f64m1((double *)(op0), (vuint8mf8_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei8_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei8_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint8mf8_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei8_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei8_v_f64m1((double *)(op0), (vuint8mf8_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei8_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei8_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint8mf8_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (size_t)(op11)) +#define vsoxseg2ei16_v_f64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_f64m1((double *)(op0), (vuint16mf4_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_f64m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint16mf4_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (size_t)(op5)) +#define vsoxseg2ei16_v_f64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_f64m2((double *)(op0), (vuint16mf2_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_f64m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_f64m2_m((vbool32_t)(op0), (double *)(op1), (vuint16mf2_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei16_v_f64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_f64m4((double *)(op0), (vuint16m1_t)(op1), (vfloat64m4_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_f64m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_f64m4_m((vbool16_t)(op0), (double *)(op1), (vuint16m1_t)(op2), (vfloat64m4_t)(op3), (vfloat64m4_t)(op4), (size_t)(op5)) +#define vsoxseg3ei16_v_f64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei16_v_f64m1((double *)(op0), (vuint16mf4_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (size_t)(op5)) +#define vsoxseg3ei16_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei16_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint16mf4_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (size_t)(op6)) +#define vsoxseg3ei16_v_f64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei16_v_f64m2((double *)(op0), (vuint16mf2_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei16_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei16_v_f64m2_m((vbool32_t)(op0), (double *)(op1), (vuint16mf2_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (vfloat64m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei16_v_f64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei16_v_f64m1((double *)(op0), (vuint16mf4_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (size_t)(op6)) +#define vsoxseg4ei16_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei16_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint16mf4_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (size_t)(op7)) +#define vsoxseg4ei16_v_f64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei16_v_f64m2((double *)(op0), (vuint16mf2_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (vfloat64m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei16_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei16_v_f64m2_m((vbool32_t)(op0), (double *)(op1), (vuint16mf2_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (vfloat64m2_t)(op5), (vfloat64m2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei16_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei16_v_f64m1((double *)(op0), (vuint16mf4_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (size_t)(op7)) +#define vsoxseg5ei16_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei16_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint16mf4_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei16_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei16_v_f64m1((double *)(op0), (vuint16mf4_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei16_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei16_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint16mf4_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei16_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei16_v_f64m1((double *)(op0), (vuint16mf4_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei16_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei16_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint16mf4_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei16_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei16_v_f64m1((double *)(op0), (vuint16mf4_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei16_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei16_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint16mf4_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (size_t)(op11)) +#define vsoxseg2ei32_v_f64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_f64m1((double *)(op0), (vuint32mf2_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_f64m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint32mf2_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (size_t)(op5)) +#define vsoxseg2ei32_v_f64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_f64m2((double *)(op0), (vuint32m1_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_f64m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_f64m2_m((vbool32_t)(op0), (double *)(op1), (vuint32m1_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei32_v_f64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_f64m4((double *)(op0), (vuint32m2_t)(op1), (vfloat64m4_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_f64m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_f64m4_m((vbool16_t)(op0), (double *)(op1), (vuint32m2_t)(op2), (vfloat64m4_t)(op3), (vfloat64m4_t)(op4), (size_t)(op5)) +#define vsoxseg3ei32_v_f64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei32_v_f64m1((double *)(op0), (vuint32mf2_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (size_t)(op5)) +#define vsoxseg3ei32_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei32_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint32mf2_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (size_t)(op6)) +#define vsoxseg3ei32_v_f64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei32_v_f64m2((double *)(op0), (vuint32m1_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei32_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei32_v_f64m2_m((vbool32_t)(op0), (double *)(op1), (vuint32m1_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (vfloat64m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei32_v_f64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei32_v_f64m1((double *)(op0), (vuint32mf2_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (size_t)(op6)) +#define vsoxseg4ei32_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei32_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint32mf2_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (size_t)(op7)) +#define vsoxseg4ei32_v_f64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei32_v_f64m2((double *)(op0), (vuint32m1_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (vfloat64m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei32_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei32_v_f64m2_m((vbool32_t)(op0), (double *)(op1), (vuint32m1_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (vfloat64m2_t)(op5), (vfloat64m2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei32_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei32_v_f64m1((double *)(op0), (vuint32mf2_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (size_t)(op7)) +#define vsoxseg5ei32_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei32_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint32mf2_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei32_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei32_v_f64m1((double *)(op0), (vuint32mf2_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei32_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei32_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint32mf2_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei32_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei32_v_f64m1((double *)(op0), (vuint32mf2_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei32_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei32_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint32mf2_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei32_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei32_v_f64m1((double *)(op0), (vuint32mf2_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei32_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei32_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint32mf2_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (size_t)(op11)) +#define vsoxseg2ei64_v_f64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei64_v_f64m1((double *)(op0), (vuint64m1_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (size_t)(op4)) +#define vsoxseg2ei64_v_f64m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei64_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (size_t)(op5)) +#define vsoxseg2ei64_v_f64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei64_v_f64m2((double *)(op0), (vuint64m2_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei64_v_f64m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei64_v_f64m2_m((vbool32_t)(op0), (double *)(op1), (vuint64m2_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei64_v_f64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei64_v_f64m4((double *)(op0), (vuint64m4_t)(op1), (vfloat64m4_t)(op2), (vfloat64m4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei64_v_f64m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei64_v_f64m4_m((vbool16_t)(op0), (double *)(op1), (vuint64m4_t)(op2), (vfloat64m4_t)(op3), (vfloat64m4_t)(op4), (size_t)(op5)) +#define vsoxseg3ei64_v_f64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei64_v_f64m1((double *)(op0), (vuint64m1_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (size_t)(op5)) +#define vsoxseg3ei64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei64_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (size_t)(op6)) +#define vsoxseg3ei64_v_f64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei64_v_f64m2((double *)(op0), (vuint64m2_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei64_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei64_v_f64m2_m((vbool32_t)(op0), (double *)(op1), (vuint64m2_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (vfloat64m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei64_v_f64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei64_v_f64m1((double *)(op0), (vuint64m1_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (size_t)(op6)) +#define vsoxseg4ei64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei64_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (size_t)(op7)) +#define vsoxseg4ei64_v_f64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei64_v_f64m2((double *)(op0), (vuint64m2_t)(op1), (vfloat64m2_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (vfloat64m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei64_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei64_v_f64m2_m((vbool32_t)(op0), (double *)(op1), (vuint64m2_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (vfloat64m2_t)(op5), (vfloat64m2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei64_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei64_v_f64m1((double *)(op0), (vuint64m1_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (size_t)(op7)) +#define vsoxseg5ei64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei64_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei64_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei64_v_f64m1((double *)(op0), (vuint64m1_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei64_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei64_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei64_v_f64m1((double *)(op0), (vuint64m1_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei64_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei64_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei64_v_f64m1((double *)(op0), (vuint64m1_t)(op1), (vfloat64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei64_v_f64m1_m((vbool64_t)(op0), (double *)(op1), (vuint64m1_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (size_t)(op11)) +#define vlseg2e64_v_f64m1(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e64_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (const double *)(op2), (size_t)(op3)) +#define vlseg2e64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e64_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vbool64_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (const double *)(op5), (size_t)(op6)) +#define vlseg2e64_v_f64m2(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e64_v_f64m2((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (const double *)(op2), (size_t)(op3)) +#define vlseg2e64_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e64_v_f64m2_m((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vbool32_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (const double *)(op5), (size_t)(op6)) +#define vlseg2e64_v_f64m4(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e64_v_f64m4((vfloat64m4_t *)(op0), (vfloat64m4_t *)(op1), (const double *)(op2), (size_t)(op3)) +#define vlseg2e64_v_f64m4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e64_v_f64m4_m((vfloat64m4_t *)(op0), (vfloat64m4_t *)(op1), (vbool16_t)(op2), (vfloat64m4_t)(op3), (vfloat64m4_t)(op4), (const double *)(op5), (size_t)(op6)) +#define vlseg3e64_v_f64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg3e64_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (const double *)(op3), (size_t)(op4)) +#define vlseg3e64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg3e64_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vbool64_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (const double *)(op7), (size_t)(op8)) +#define vlseg3e64_v_f64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg3e64_v_f64m2((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vfloat64m2_t *)(op2), (const double *)(op3), (size_t)(op4)) +#define vlseg3e64_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg3e64_v_f64m2_m((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vfloat64m2_t *)(op2), (vbool32_t)(op3), (vfloat64m2_t)(op4), (vfloat64m2_t)(op5), (vfloat64m2_t)(op6), (const double *)(op7), (size_t)(op8)) +#define vlseg4e64_v_f64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg4e64_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (const double *)(op4), (size_t)(op5)) +#define vlseg4e64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg4e64_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vbool64_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (const double *)(op9), (size_t)(op10)) +#define vlseg4e64_v_f64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg4e64_v_f64m2((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vfloat64m2_t *)(op2), (vfloat64m2_t *)(op3), (const double *)(op4), (size_t)(op5)) +#define vlseg4e64_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg4e64_v_f64m2_m((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vfloat64m2_t *)(op2), (vfloat64m2_t *)(op3), (vbool32_t)(op4), (vfloat64m2_t)(op5), (vfloat64m2_t)(op6), (vfloat64m2_t)(op7), (vfloat64m2_t)(op8), (const double *)(op9), (size_t)(op10)) +#define vlseg5e64_v_f64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg5e64_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (const double *)(op5), (size_t)(op6)) +#define vlseg5e64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12) \ +__builtin_rvv_vlseg5e64_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vbool64_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (const double *)(op11), (size_t)(op12)) +#define vlseg6e64_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg6e64_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (const double *)(op6), (size_t)(op7)) +#define vlseg6e64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14) \ +__builtin_rvv_vlseg6e64_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vbool64_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (vfloat64m1_t)(op11), (vfloat64m1_t)(op12), (const double *)(op13), (size_t)(op14)) +#define vlseg7e64_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg7e64_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vfloat64m1_t *)(op6), (const double *)(op7), (size_t)(op8)) +#define vlseg7e64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16) \ +__builtin_rvv_vlseg7e64_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vfloat64m1_t *)(op6), (vbool64_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (vfloat64m1_t)(op11), (vfloat64m1_t)(op12), (vfloat64m1_t)(op13), (vfloat64m1_t)(op14), (const double *)(op15), (size_t)(op16)) +#define vlseg8e64_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg8e64_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vfloat64m1_t *)(op6), (vfloat64m1_t *)(op7), (const double *)(op8), (size_t)(op9)) +#define vlseg8e64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18) \ +__builtin_rvv_vlseg8e64_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vfloat64m1_t *)(op6), (vfloat64m1_t *)(op7), (vbool64_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (vfloat64m1_t)(op11), (vfloat64m1_t)(op12), (vfloat64m1_t)(op13), (vfloat64m1_t)(op14), (vfloat64m1_t)(op15), (vfloat64m1_t)(op16), (const double *)(op17), (size_t)(op18)) +#define vlseg2e64ff_v_f64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e64ff_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (const double *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e64ff_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e64ff_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vbool64_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (const double *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg2e64ff_v_f64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e64ff_v_f64m2((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (const double *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e64ff_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e64ff_v_f64m2_m((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vbool32_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (const double *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg2e64ff_v_f64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e64ff_v_f64m4((vfloat64m4_t *)(op0), (vfloat64m4_t *)(op1), (const double *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e64ff_v_f64m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e64ff_v_f64m4_m((vfloat64m4_t *)(op0), (vfloat64m4_t *)(op1), (vbool16_t)(op2), (vfloat64m4_t)(op3), (vfloat64m4_t)(op4), (const double *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg3e64ff_v_f64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg3e64ff_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (const double *)(op3), (size_t *)(op4), (size_t)(op5)) +#define vlseg3e64ff_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg3e64ff_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vbool64_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (const double *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg3e64ff_v_f64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg3e64ff_v_f64m2((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vfloat64m2_t *)(op2), (const double *)(op3), (size_t *)(op4), (size_t)(op5)) +#define vlseg3e64ff_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg3e64ff_v_f64m2_m((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vfloat64m2_t *)(op2), (vbool32_t)(op3), (vfloat64m2_t)(op4), (vfloat64m2_t)(op5), (vfloat64m2_t)(op6), (const double *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg4e64ff_v_f64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg4e64ff_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (const double *)(op4), (size_t *)(op5), (size_t)(op6)) +#define vlseg4e64ff_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlseg4e64ff_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vbool64_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (const double *)(op9), (size_t *)(op10), (size_t)(op11)) +#define vlseg4e64ff_v_f64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg4e64ff_v_f64m2((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vfloat64m2_t *)(op2), (vfloat64m2_t *)(op3), (const double *)(op4), (size_t *)(op5), (size_t)(op6)) +#define vlseg4e64ff_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlseg4e64ff_v_f64m2_m((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vfloat64m2_t *)(op2), (vfloat64m2_t *)(op3), (vbool32_t)(op4), (vfloat64m2_t)(op5), (vfloat64m2_t)(op6), (vfloat64m2_t)(op7), (vfloat64m2_t)(op8), (const double *)(op9), (size_t *)(op10), (size_t)(op11)) +#define vlseg5e64ff_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg5e64ff_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (const double *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg5e64ff_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlseg5e64ff_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vbool64_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (const double *)(op11), (size_t *)(op12), (size_t)(op13)) +#define vlseg6e64ff_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg6e64ff_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (const double *)(op6), (size_t *)(op7), (size_t)(op8)) +#define vlseg6e64ff_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlseg6e64ff_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vbool64_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (vfloat64m1_t)(op11), (vfloat64m1_t)(op12), (const double *)(op13), (size_t *)(op14), (size_t)(op15)) +#define vlseg7e64ff_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg7e64ff_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vfloat64m1_t *)(op6), (const double *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg7e64ff_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlseg7e64ff_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vfloat64m1_t *)(op6), (vbool64_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (vfloat64m1_t)(op11), (vfloat64m1_t)(op12), (vfloat64m1_t)(op13), (vfloat64m1_t)(op14), (const double *)(op15), (size_t *)(op16), (size_t)(op17)) +#define vlseg8e64ff_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg8e64ff_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vfloat64m1_t *)(op6), (vfloat64m1_t *)(op7), (const double *)(op8), (size_t *)(op9), (size_t)(op10)) +#define vlseg8e64ff_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlseg8e64ff_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vfloat64m1_t *)(op6), (vfloat64m1_t *)(op7), (vbool64_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (vfloat64m1_t)(op11), (vfloat64m1_t)(op12), (vfloat64m1_t)(op13), (vfloat64m1_t)(op14), (vfloat64m1_t)(op15), (vfloat64m1_t)(op16), (const double *)(op17), (size_t *)(op18), (size_t)(op19)) +#define vlsseg2e64_v_f64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e64_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (const double *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e64_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vbool64_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (const double *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg2e64_v_f64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e64_v_f64m2((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (const double *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e64_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e64_v_f64m2_m((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vbool32_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (const double *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg2e64_v_f64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e64_v_f64m4((vfloat64m4_t *)(op0), (vfloat64m4_t *)(op1), (const double *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e64_v_f64m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e64_v_f64m4_m((vfloat64m4_t *)(op0), (vfloat64m4_t *)(op1), (vbool16_t)(op2), (vfloat64m4_t)(op3), (vfloat64m4_t)(op4), (const double *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg3e64_v_f64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlsseg3e64_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (const double *)(op3), (ptrdiff_t)(op4), (size_t)(op5)) +#define vlsseg3e64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg3e64_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vbool64_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (const double *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg3e64_v_f64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlsseg3e64_v_f64m2((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vfloat64m2_t *)(op2), (const double *)(op3), (ptrdiff_t)(op4), (size_t)(op5)) +#define vlsseg3e64_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg3e64_v_f64m2_m((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vfloat64m2_t *)(op2), (vbool32_t)(op3), (vfloat64m2_t)(op4), (vfloat64m2_t)(op5), (vfloat64m2_t)(op6), (const double *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg4e64_v_f64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlsseg4e64_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (const double *)(op4), (ptrdiff_t)(op5), (size_t)(op6)) +#define vlsseg4e64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlsseg4e64_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vbool64_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (const double *)(op9), (ptrdiff_t)(op10), (size_t)(op11)) +#define vlsseg4e64_v_f64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlsseg4e64_v_f64m2((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vfloat64m2_t *)(op2), (vfloat64m2_t *)(op3), (const double *)(op4), (ptrdiff_t)(op5), (size_t)(op6)) +#define vlsseg4e64_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlsseg4e64_v_f64m2_m((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vfloat64m2_t *)(op2), (vfloat64m2_t *)(op3), (vbool32_t)(op4), (vfloat64m2_t)(op5), (vfloat64m2_t)(op6), (vfloat64m2_t)(op7), (vfloat64m2_t)(op8), (const double *)(op9), (ptrdiff_t)(op10), (size_t)(op11)) +#define vlsseg5e64_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg5e64_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (const double *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg5e64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlsseg5e64_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vbool64_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (const double *)(op11), (ptrdiff_t)(op12), (size_t)(op13)) +#define vlsseg6e64_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlsseg6e64_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (const double *)(op6), (ptrdiff_t)(op7), (size_t)(op8)) +#define vlsseg6e64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlsseg6e64_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vbool64_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (vfloat64m1_t)(op11), (vfloat64m1_t)(op12), (const double *)(op13), (ptrdiff_t)(op14), (size_t)(op15)) +#define vlsseg7e64_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg7e64_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vfloat64m1_t *)(op6), (const double *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg7e64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlsseg7e64_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vfloat64m1_t *)(op6), (vbool64_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (vfloat64m1_t)(op11), (vfloat64m1_t)(op12), (vfloat64m1_t)(op13), (vfloat64m1_t)(op14), (const double *)(op15), (ptrdiff_t)(op16), (size_t)(op17)) +#define vlsseg8e64_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlsseg8e64_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vfloat64m1_t *)(op6), (vfloat64m1_t *)(op7), (const double *)(op8), (ptrdiff_t)(op9), (size_t)(op10)) +#define vlsseg8e64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlsseg8e64_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vfloat64m1_t *)(op6), (vfloat64m1_t *)(op7), (vbool64_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (vfloat64m1_t)(op11), (vfloat64m1_t)(op12), (vfloat64m1_t)(op13), (vfloat64m1_t)(op14), (vfloat64m1_t)(op15), (vfloat64m1_t)(op16), (const double *)(op17), (ptrdiff_t)(op18), (size_t)(op19)) +#define vluxseg2ei8_v_f64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (const double *)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vbool64_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (const double *)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vluxseg2ei8_v_f64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_f64m2((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (const double *)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_f64m2_m((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vbool32_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (const double *)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vluxseg2ei8_v_f64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_f64m4((vfloat64m4_t *)(op0), (vfloat64m4_t *)(op1), (const double *)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_f64m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_f64m4_m((vfloat64m4_t *)(op0), (vfloat64m4_t *)(op1), (vbool16_t)(op2), (vfloat64m4_t)(op3), (vfloat64m4_t)(op4), (const double *)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vluxseg3ei8_v_f64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei8_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (const double *)(op3), (vuint8mf8_t)(op4), (size_t)(op5)) +#define vluxseg3ei8_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei8_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vbool64_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (const double *)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vluxseg3ei8_v_f64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei8_v_f64m2((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vfloat64m2_t *)(op2), (const double *)(op3), (vuint8mf4_t)(op4), (size_t)(op5)) +#define vluxseg3ei8_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei8_v_f64m2_m((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vfloat64m2_t *)(op2), (vbool32_t)(op3), (vfloat64m2_t)(op4), (vfloat64m2_t)(op5), (vfloat64m2_t)(op6), (const double *)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vluxseg4ei8_v_f64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei8_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (const double *)(op4), (vuint8mf8_t)(op5), (size_t)(op6)) +#define vluxseg4ei8_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei8_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vbool64_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (const double *)(op9), (vuint8mf8_t)(op10), (size_t)(op11)) +#define vluxseg4ei8_v_f64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei8_v_f64m2((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vfloat64m2_t *)(op2), (vfloat64m2_t *)(op3), (const double *)(op4), (vuint8mf4_t)(op5), (size_t)(op6)) +#define vluxseg4ei8_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei8_v_f64m2_m((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vfloat64m2_t *)(op2), (vfloat64m2_t *)(op3), (vbool32_t)(op4), (vfloat64m2_t)(op5), (vfloat64m2_t)(op6), (vfloat64m2_t)(op7), (vfloat64m2_t)(op8), (const double *)(op9), (vuint8mf4_t)(op10), (size_t)(op11)) +#define vluxseg5ei8_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei8_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (const double *)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vluxseg5ei8_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei8_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vbool64_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (const double *)(op11), (vuint8mf8_t)(op12), (size_t)(op13)) +#define vluxseg6ei8_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei8_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (const double *)(op6), (vuint8mf8_t)(op7), (size_t)(op8)) +#define vluxseg6ei8_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei8_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vbool64_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (vfloat64m1_t)(op11), (vfloat64m1_t)(op12), (const double *)(op13), (vuint8mf8_t)(op14), (size_t)(op15)) +#define vluxseg7ei8_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei8_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vfloat64m1_t *)(op6), (const double *)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vluxseg7ei8_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei8_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vfloat64m1_t *)(op6), (vbool64_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (vfloat64m1_t)(op11), (vfloat64m1_t)(op12), (vfloat64m1_t)(op13), (vfloat64m1_t)(op14), (const double *)(op15), (vuint8mf8_t)(op16), (size_t)(op17)) +#define vluxseg8ei8_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei8_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vfloat64m1_t *)(op6), (vfloat64m1_t *)(op7), (const double *)(op8), (vuint8mf8_t)(op9), (size_t)(op10)) +#define vluxseg8ei8_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei8_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vfloat64m1_t *)(op6), (vfloat64m1_t *)(op7), (vbool64_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (vfloat64m1_t)(op11), (vfloat64m1_t)(op12), (vfloat64m1_t)(op13), (vfloat64m1_t)(op14), (vfloat64m1_t)(op15), (vfloat64m1_t)(op16), (const double *)(op17), (vuint8mf8_t)(op18), (size_t)(op19)) +#define vluxseg2ei16_v_f64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (const double *)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vbool64_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (const double *)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vluxseg2ei16_v_f64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_f64m2((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (const double *)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_f64m2_m((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vbool32_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (const double *)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vluxseg2ei16_v_f64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_f64m4((vfloat64m4_t *)(op0), (vfloat64m4_t *)(op1), (const double *)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_f64m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_f64m4_m((vfloat64m4_t *)(op0), (vfloat64m4_t *)(op1), (vbool16_t)(op2), (vfloat64m4_t)(op3), (vfloat64m4_t)(op4), (const double *)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vluxseg3ei16_v_f64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei16_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (const double *)(op3), (vuint16mf4_t)(op4), (size_t)(op5)) +#define vluxseg3ei16_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei16_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vbool64_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (const double *)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vluxseg3ei16_v_f64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei16_v_f64m2((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vfloat64m2_t *)(op2), (const double *)(op3), (vuint16mf2_t)(op4), (size_t)(op5)) +#define vluxseg3ei16_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei16_v_f64m2_m((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vfloat64m2_t *)(op2), (vbool32_t)(op3), (vfloat64m2_t)(op4), (vfloat64m2_t)(op5), (vfloat64m2_t)(op6), (const double *)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vluxseg4ei16_v_f64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei16_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (const double *)(op4), (vuint16mf4_t)(op5), (size_t)(op6)) +#define vluxseg4ei16_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei16_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vbool64_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (const double *)(op9), (vuint16mf4_t)(op10), (size_t)(op11)) +#define vluxseg4ei16_v_f64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei16_v_f64m2((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vfloat64m2_t *)(op2), (vfloat64m2_t *)(op3), (const double *)(op4), (vuint16mf2_t)(op5), (size_t)(op6)) +#define vluxseg4ei16_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei16_v_f64m2_m((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vfloat64m2_t *)(op2), (vfloat64m2_t *)(op3), (vbool32_t)(op4), (vfloat64m2_t)(op5), (vfloat64m2_t)(op6), (vfloat64m2_t)(op7), (vfloat64m2_t)(op8), (const double *)(op9), (vuint16mf2_t)(op10), (size_t)(op11)) +#define vluxseg5ei16_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei16_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (const double *)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vluxseg5ei16_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei16_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vbool64_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (const double *)(op11), (vuint16mf4_t)(op12), (size_t)(op13)) +#define vluxseg6ei16_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei16_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (const double *)(op6), (vuint16mf4_t)(op7), (size_t)(op8)) +#define vluxseg6ei16_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei16_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vbool64_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (vfloat64m1_t)(op11), (vfloat64m1_t)(op12), (const double *)(op13), (vuint16mf4_t)(op14), (size_t)(op15)) +#define vluxseg7ei16_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei16_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vfloat64m1_t *)(op6), (const double *)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vluxseg7ei16_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei16_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vfloat64m1_t *)(op6), (vbool64_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (vfloat64m1_t)(op11), (vfloat64m1_t)(op12), (vfloat64m1_t)(op13), (vfloat64m1_t)(op14), (const double *)(op15), (vuint16mf4_t)(op16), (size_t)(op17)) +#define vluxseg8ei16_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei16_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vfloat64m1_t *)(op6), (vfloat64m1_t *)(op7), (const double *)(op8), (vuint16mf4_t)(op9), (size_t)(op10)) +#define vluxseg8ei16_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei16_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vfloat64m1_t *)(op6), (vfloat64m1_t *)(op7), (vbool64_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (vfloat64m1_t)(op11), (vfloat64m1_t)(op12), (vfloat64m1_t)(op13), (vfloat64m1_t)(op14), (vfloat64m1_t)(op15), (vfloat64m1_t)(op16), (const double *)(op17), (vuint16mf4_t)(op18), (size_t)(op19)) +#define vluxseg2ei32_v_f64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (const double *)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vbool64_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (const double *)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vluxseg2ei32_v_f64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_f64m2((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (const double *)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_f64m2_m((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vbool32_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (const double *)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vluxseg2ei32_v_f64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_f64m4((vfloat64m4_t *)(op0), (vfloat64m4_t *)(op1), (const double *)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_f64m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_f64m4_m((vfloat64m4_t *)(op0), (vfloat64m4_t *)(op1), (vbool16_t)(op2), (vfloat64m4_t)(op3), (vfloat64m4_t)(op4), (const double *)(op5), (vuint32m2_t)(op6), (size_t)(op7)) +#define vluxseg3ei32_v_f64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei32_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (const double *)(op3), (vuint32mf2_t)(op4), (size_t)(op5)) +#define vluxseg3ei32_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei32_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vbool64_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (const double *)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vluxseg3ei32_v_f64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei32_v_f64m2((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vfloat64m2_t *)(op2), (const double *)(op3), (vuint32m1_t)(op4), (size_t)(op5)) +#define vluxseg3ei32_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei32_v_f64m2_m((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vfloat64m2_t *)(op2), (vbool32_t)(op3), (vfloat64m2_t)(op4), (vfloat64m2_t)(op5), (vfloat64m2_t)(op6), (const double *)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vluxseg4ei32_v_f64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei32_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (const double *)(op4), (vuint32mf2_t)(op5), (size_t)(op6)) +#define vluxseg4ei32_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei32_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vbool64_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (const double *)(op9), (vuint32mf2_t)(op10), (size_t)(op11)) +#define vluxseg4ei32_v_f64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei32_v_f64m2((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vfloat64m2_t *)(op2), (vfloat64m2_t *)(op3), (const double *)(op4), (vuint32m1_t)(op5), (size_t)(op6)) +#define vluxseg4ei32_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei32_v_f64m2_m((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vfloat64m2_t *)(op2), (vfloat64m2_t *)(op3), (vbool32_t)(op4), (vfloat64m2_t)(op5), (vfloat64m2_t)(op6), (vfloat64m2_t)(op7), (vfloat64m2_t)(op8), (const double *)(op9), (vuint32m1_t)(op10), (size_t)(op11)) +#define vluxseg5ei32_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei32_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (const double *)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vluxseg5ei32_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei32_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vbool64_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (const double *)(op11), (vuint32mf2_t)(op12), (size_t)(op13)) +#define vluxseg6ei32_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei32_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (const double *)(op6), (vuint32mf2_t)(op7), (size_t)(op8)) +#define vluxseg6ei32_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei32_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vbool64_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (vfloat64m1_t)(op11), (vfloat64m1_t)(op12), (const double *)(op13), (vuint32mf2_t)(op14), (size_t)(op15)) +#define vluxseg7ei32_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei32_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vfloat64m1_t *)(op6), (const double *)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vluxseg7ei32_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei32_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vfloat64m1_t *)(op6), (vbool64_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (vfloat64m1_t)(op11), (vfloat64m1_t)(op12), (vfloat64m1_t)(op13), (vfloat64m1_t)(op14), (const double *)(op15), (vuint32mf2_t)(op16), (size_t)(op17)) +#define vluxseg8ei32_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei32_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vfloat64m1_t *)(op6), (vfloat64m1_t *)(op7), (const double *)(op8), (vuint32mf2_t)(op9), (size_t)(op10)) +#define vluxseg8ei32_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei32_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vfloat64m1_t *)(op6), (vfloat64m1_t *)(op7), (vbool64_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (vfloat64m1_t)(op11), (vfloat64m1_t)(op12), (vfloat64m1_t)(op13), (vfloat64m1_t)(op14), (vfloat64m1_t)(op15), (vfloat64m1_t)(op16), (const double *)(op17), (vuint32mf2_t)(op18), (size_t)(op19)) +#define vluxseg2ei64_v_f64m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei64_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (const double *)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vluxseg2ei64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei64_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vbool64_t)(op2), (vfloat64m1_t)(op3), (vfloat64m1_t)(op4), (const double *)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vluxseg2ei64_v_f64m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei64_v_f64m2((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (const double *)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vluxseg2ei64_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei64_v_f64m2_m((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vbool32_t)(op2), (vfloat64m2_t)(op3), (vfloat64m2_t)(op4), (const double *)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vluxseg2ei64_v_f64m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei64_v_f64m4((vfloat64m4_t *)(op0), (vfloat64m4_t *)(op1), (const double *)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vluxseg2ei64_v_f64m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei64_v_f64m4_m((vfloat64m4_t *)(op0), (vfloat64m4_t *)(op1), (vbool16_t)(op2), (vfloat64m4_t)(op3), (vfloat64m4_t)(op4), (const double *)(op5), (vuint64m4_t)(op6), (size_t)(op7)) +#define vluxseg3ei64_v_f64m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei64_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (const double *)(op3), (vuint64m1_t)(op4), (size_t)(op5)) +#define vluxseg3ei64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei64_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vbool64_t)(op3), (vfloat64m1_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (const double *)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vluxseg3ei64_v_f64m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei64_v_f64m2((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vfloat64m2_t *)(op2), (const double *)(op3), (vuint64m2_t)(op4), (size_t)(op5)) +#define vluxseg3ei64_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei64_v_f64m2_m((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vfloat64m2_t *)(op2), (vbool32_t)(op3), (vfloat64m2_t)(op4), (vfloat64m2_t)(op5), (vfloat64m2_t)(op6), (const double *)(op7), (vuint64m2_t)(op8), (size_t)(op9)) +#define vluxseg4ei64_v_f64m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei64_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (const double *)(op4), (vuint64m1_t)(op5), (size_t)(op6)) +#define vluxseg4ei64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei64_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vbool64_t)(op4), (vfloat64m1_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (const double *)(op9), (vuint64m1_t)(op10), (size_t)(op11)) +#define vluxseg4ei64_v_f64m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei64_v_f64m2((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vfloat64m2_t *)(op2), (vfloat64m2_t *)(op3), (const double *)(op4), (vuint64m2_t)(op5), (size_t)(op6)) +#define vluxseg4ei64_v_f64m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei64_v_f64m2_m((vfloat64m2_t *)(op0), (vfloat64m2_t *)(op1), (vfloat64m2_t *)(op2), (vfloat64m2_t *)(op3), (vbool32_t)(op4), (vfloat64m2_t)(op5), (vfloat64m2_t)(op6), (vfloat64m2_t)(op7), (vfloat64m2_t)(op8), (const double *)(op9), (vuint64m2_t)(op10), (size_t)(op11)) +#define vluxseg5ei64_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei64_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (const double *)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vluxseg5ei64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei64_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vbool64_t)(op5), (vfloat64m1_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (const double *)(op11), (vuint64m1_t)(op12), (size_t)(op13)) +#define vluxseg6ei64_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei64_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (const double *)(op6), (vuint64m1_t)(op7), (size_t)(op8)) +#define vluxseg6ei64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei64_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vbool64_t)(op6), (vfloat64m1_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (vfloat64m1_t)(op11), (vfloat64m1_t)(op12), (const double *)(op13), (vuint64m1_t)(op14), (size_t)(op15)) +#define vluxseg7ei64_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei64_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vfloat64m1_t *)(op6), (const double *)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vluxseg7ei64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei64_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vfloat64m1_t *)(op6), (vbool64_t)(op7), (vfloat64m1_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (vfloat64m1_t)(op11), (vfloat64m1_t)(op12), (vfloat64m1_t)(op13), (vfloat64m1_t)(op14), (const double *)(op15), (vuint64m1_t)(op16), (size_t)(op17)) +#define vluxseg8ei64_v_f64m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei64_v_f64m1((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vfloat64m1_t *)(op6), (vfloat64m1_t *)(op7), (const double *)(op8), (vuint64m1_t)(op9), (size_t)(op10)) +#define vluxseg8ei64_v_f64m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei64_v_f64m1_m((vfloat64m1_t *)(op0), (vfloat64m1_t *)(op1), (vfloat64m1_t *)(op2), (vfloat64m1_t *)(op3), (vfloat64m1_t *)(op4), (vfloat64m1_t *)(op5), (vfloat64m1_t *)(op6), (vfloat64m1_t *)(op7), (vbool64_t)(op8), (vfloat64m1_t)(op9), (vfloat64m1_t)(op10), (vfloat64m1_t)(op11), (vfloat64m1_t)(op12), (vfloat64m1_t)(op13), (vfloat64m1_t)(op14), (vfloat64m1_t)(op15), (vfloat64m1_t)(op16), (const double *)(op17), (vuint64m1_t)(op18), (size_t)(op19)) +#endif + +#if defined(__riscv_zfh) && defined(__riscv_zvlsseg) +#define vloxseg2ei8_v_f16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (const _Float16 *)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vbool16_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (const _Float16 *)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vloxseg2ei8_v_f16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_f16m2((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (const _Float16 *)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_f16m2_m((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vbool8_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (const _Float16 *)(op5), (vuint8m1_t)(op6), (size_t)(op7)) +#define vloxseg2ei8_v_f16m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_f16m4((vfloat16m4_t *)(op0), (vfloat16m4_t *)(op1), (const _Float16 *)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_f16m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_f16m4_m((vfloat16m4_t *)(op0), (vfloat16m4_t *)(op1), (vbool4_t)(op2), (vfloat16m4_t)(op3), (vfloat16m4_t)(op4), (const _Float16 *)(op5), (vuint8m2_t)(op6), (size_t)(op7)) +#define vloxseg2ei8_v_f16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (const _Float16 *)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vbool32_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (const _Float16 *)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vloxseg2ei8_v_f16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei8_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (const _Float16 *)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vloxseg2ei8_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei8_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vbool64_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (const _Float16 *)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vloxseg3ei8_v_f16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei8_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (const _Float16 *)(op3), (vuint8mf2_t)(op4), (size_t)(op5)) +#define vloxseg3ei8_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei8_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vbool16_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (const _Float16 *)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vloxseg3ei8_v_f16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei8_v_f16m2((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vfloat16m2_t *)(op2), (const _Float16 *)(op3), (vuint8m1_t)(op4), (size_t)(op5)) +#define vloxseg3ei8_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei8_v_f16m2_m((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vfloat16m2_t *)(op2), (vbool8_t)(op3), (vfloat16m2_t)(op4), (vfloat16m2_t)(op5), (vfloat16m2_t)(op6), (const _Float16 *)(op7), (vuint8m1_t)(op8), (size_t)(op9)) +#define vloxseg3ei8_v_f16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei8_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (const _Float16 *)(op3), (vuint8mf4_t)(op4), (size_t)(op5)) +#define vloxseg3ei8_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei8_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vbool32_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (const _Float16 *)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vloxseg3ei8_v_f16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei8_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (const _Float16 *)(op3), (vuint8mf8_t)(op4), (size_t)(op5)) +#define vloxseg3ei8_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei8_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vbool64_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (const _Float16 *)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vloxseg4ei8_v_f16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei8_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (const _Float16 *)(op4), (vuint8mf2_t)(op5), (size_t)(op6)) +#define vloxseg4ei8_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei8_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vbool16_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (const _Float16 *)(op9), (vuint8mf2_t)(op10), (size_t)(op11)) +#define vloxseg4ei8_v_f16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei8_v_f16m2((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vfloat16m2_t *)(op2), (vfloat16m2_t *)(op3), (const _Float16 *)(op4), (vuint8m1_t)(op5), (size_t)(op6)) +#define vloxseg4ei8_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei8_v_f16m2_m((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vfloat16m2_t *)(op2), (vfloat16m2_t *)(op3), (vbool8_t)(op4), (vfloat16m2_t)(op5), (vfloat16m2_t)(op6), (vfloat16m2_t)(op7), (vfloat16m2_t)(op8), (const _Float16 *)(op9), (vuint8m1_t)(op10), (size_t)(op11)) +#define vloxseg4ei8_v_f16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei8_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (const _Float16 *)(op4), (vuint8mf4_t)(op5), (size_t)(op6)) +#define vloxseg4ei8_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei8_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vbool32_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (const _Float16 *)(op9), (vuint8mf4_t)(op10), (size_t)(op11)) +#define vloxseg4ei8_v_f16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei8_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (const _Float16 *)(op4), (vuint8mf8_t)(op5), (size_t)(op6)) +#define vloxseg4ei8_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei8_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vbool64_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (const _Float16 *)(op9), (vuint8mf8_t)(op10), (size_t)(op11)) +#define vloxseg5ei8_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei8_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (const _Float16 *)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vloxseg5ei8_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei8_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vbool16_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (const _Float16 *)(op11), (vuint8mf2_t)(op12), (size_t)(op13)) +#define vloxseg5ei8_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei8_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (const _Float16 *)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vloxseg5ei8_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei8_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vbool32_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (const _Float16 *)(op11), (vuint8mf4_t)(op12), (size_t)(op13)) +#define vloxseg5ei8_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei8_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (const _Float16 *)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vloxseg5ei8_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei8_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vbool64_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (const _Float16 *)(op11), (vuint8mf8_t)(op12), (size_t)(op13)) +#define vloxseg6ei8_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei8_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (const _Float16 *)(op6), (vuint8mf2_t)(op7), (size_t)(op8)) +#define vloxseg6ei8_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei8_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vbool16_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (vfloat16m1_t)(op11), (vfloat16m1_t)(op12), (const _Float16 *)(op13), (vuint8mf2_t)(op14), (size_t)(op15)) +#define vloxseg6ei8_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei8_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (const _Float16 *)(op6), (vuint8mf4_t)(op7), (size_t)(op8)) +#define vloxseg6ei8_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei8_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vbool32_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (vfloat16mf2_t)(op11), (vfloat16mf2_t)(op12), (const _Float16 *)(op13), (vuint8mf4_t)(op14), (size_t)(op15)) +#define vloxseg6ei8_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei8_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (const _Float16 *)(op6), (vuint8mf8_t)(op7), (size_t)(op8)) +#define vloxseg6ei8_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei8_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vbool64_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (vfloat16mf4_t)(op11), (vfloat16mf4_t)(op12), (const _Float16 *)(op13), (vuint8mf8_t)(op14), (size_t)(op15)) +#define vloxseg7ei8_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei8_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vfloat16m1_t *)(op6), (const _Float16 *)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vloxseg7ei8_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei8_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vfloat16m1_t *)(op6), (vbool16_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (vfloat16m1_t)(op11), (vfloat16m1_t)(op12), (vfloat16m1_t)(op13), (vfloat16m1_t)(op14), (const _Float16 *)(op15), (vuint8mf2_t)(op16), (size_t)(op17)) +#define vloxseg7ei8_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei8_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vfloat16mf2_t *)(op6), (const _Float16 *)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vloxseg7ei8_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei8_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vfloat16mf2_t *)(op6), (vbool32_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (vfloat16mf2_t)(op11), (vfloat16mf2_t)(op12), (vfloat16mf2_t)(op13), (vfloat16mf2_t)(op14), (const _Float16 *)(op15), (vuint8mf4_t)(op16), (size_t)(op17)) +#define vloxseg7ei8_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei8_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vfloat16mf4_t *)(op6), (const _Float16 *)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vloxseg7ei8_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei8_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vfloat16mf4_t *)(op6), (vbool64_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (vfloat16mf4_t)(op11), (vfloat16mf4_t)(op12), (vfloat16mf4_t)(op13), (vfloat16mf4_t)(op14), (const _Float16 *)(op15), (vuint8mf8_t)(op16), (size_t)(op17)) +#define vloxseg8ei8_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei8_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vfloat16m1_t *)(op6), (vfloat16m1_t *)(op7), (const _Float16 *)(op8), (vuint8mf2_t)(op9), (size_t)(op10)) +#define vloxseg8ei8_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei8_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vfloat16m1_t *)(op6), (vfloat16m1_t *)(op7), (vbool16_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (vfloat16m1_t)(op11), (vfloat16m1_t)(op12), (vfloat16m1_t)(op13), (vfloat16m1_t)(op14), (vfloat16m1_t)(op15), (vfloat16m1_t)(op16), (const _Float16 *)(op17), (vuint8mf2_t)(op18), (size_t)(op19)) +#define vloxseg8ei8_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei8_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vfloat16mf2_t *)(op6), (vfloat16mf2_t *)(op7), (const _Float16 *)(op8), (vuint8mf4_t)(op9), (size_t)(op10)) +#define vloxseg8ei8_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei8_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vfloat16mf2_t *)(op6), (vfloat16mf2_t *)(op7), (vbool32_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (vfloat16mf2_t)(op11), (vfloat16mf2_t)(op12), (vfloat16mf2_t)(op13), (vfloat16mf2_t)(op14), (vfloat16mf2_t)(op15), (vfloat16mf2_t)(op16), (const _Float16 *)(op17), (vuint8mf4_t)(op18), (size_t)(op19)) +#define vloxseg8ei8_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei8_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vfloat16mf4_t *)(op6), (vfloat16mf4_t *)(op7), (const _Float16 *)(op8), (vuint8mf8_t)(op9), (size_t)(op10)) +#define vloxseg8ei8_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei8_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vfloat16mf4_t *)(op6), (vfloat16mf4_t *)(op7), (vbool64_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (vfloat16mf4_t)(op11), (vfloat16mf4_t)(op12), (vfloat16mf4_t)(op13), (vfloat16mf4_t)(op14), (vfloat16mf4_t)(op15), (vfloat16mf4_t)(op16), (const _Float16 *)(op17), (vuint8mf8_t)(op18), (size_t)(op19)) +#define vloxseg2ei16_v_f16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (const _Float16 *)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vbool16_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (const _Float16 *)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vloxseg2ei16_v_f16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_f16m2((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (const _Float16 *)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_f16m2_m((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vbool8_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (const _Float16 *)(op5), (vuint16m2_t)(op6), (size_t)(op7)) +#define vloxseg2ei16_v_f16m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_f16m4((vfloat16m4_t *)(op0), (vfloat16m4_t *)(op1), (const _Float16 *)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_f16m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_f16m4_m((vfloat16m4_t *)(op0), (vfloat16m4_t *)(op1), (vbool4_t)(op2), (vfloat16m4_t)(op3), (vfloat16m4_t)(op4), (const _Float16 *)(op5), (vuint16m4_t)(op6), (size_t)(op7)) +#define vloxseg2ei16_v_f16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (const _Float16 *)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vbool32_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (const _Float16 *)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vloxseg2ei16_v_f16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei16_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (const _Float16 *)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vloxseg2ei16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei16_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vbool64_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (const _Float16 *)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vloxseg3ei16_v_f16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei16_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (const _Float16 *)(op3), (vuint16m1_t)(op4), (size_t)(op5)) +#define vloxseg3ei16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei16_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vbool16_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (const _Float16 *)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vloxseg3ei16_v_f16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei16_v_f16m2((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vfloat16m2_t *)(op2), (const _Float16 *)(op3), (vuint16m2_t)(op4), (size_t)(op5)) +#define vloxseg3ei16_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei16_v_f16m2_m((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vfloat16m2_t *)(op2), (vbool8_t)(op3), (vfloat16m2_t)(op4), (vfloat16m2_t)(op5), (vfloat16m2_t)(op6), (const _Float16 *)(op7), (vuint16m2_t)(op8), (size_t)(op9)) +#define vloxseg3ei16_v_f16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei16_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (const _Float16 *)(op3), (vuint16mf2_t)(op4), (size_t)(op5)) +#define vloxseg3ei16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei16_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vbool32_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (const _Float16 *)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vloxseg3ei16_v_f16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei16_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (const _Float16 *)(op3), (vuint16mf4_t)(op4), (size_t)(op5)) +#define vloxseg3ei16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei16_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vbool64_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (const _Float16 *)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vloxseg4ei16_v_f16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei16_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (const _Float16 *)(op4), (vuint16m1_t)(op5), (size_t)(op6)) +#define vloxseg4ei16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei16_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vbool16_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (const _Float16 *)(op9), (vuint16m1_t)(op10), (size_t)(op11)) +#define vloxseg4ei16_v_f16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei16_v_f16m2((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vfloat16m2_t *)(op2), (vfloat16m2_t *)(op3), (const _Float16 *)(op4), (vuint16m2_t)(op5), (size_t)(op6)) +#define vloxseg4ei16_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei16_v_f16m2_m((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vfloat16m2_t *)(op2), (vfloat16m2_t *)(op3), (vbool8_t)(op4), (vfloat16m2_t)(op5), (vfloat16m2_t)(op6), (vfloat16m2_t)(op7), (vfloat16m2_t)(op8), (const _Float16 *)(op9), (vuint16m2_t)(op10), (size_t)(op11)) +#define vloxseg4ei16_v_f16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei16_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (const _Float16 *)(op4), (vuint16mf2_t)(op5), (size_t)(op6)) +#define vloxseg4ei16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei16_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vbool32_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (const _Float16 *)(op9), (vuint16mf2_t)(op10), (size_t)(op11)) +#define vloxseg4ei16_v_f16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei16_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (const _Float16 *)(op4), (vuint16mf4_t)(op5), (size_t)(op6)) +#define vloxseg4ei16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei16_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vbool64_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (const _Float16 *)(op9), (vuint16mf4_t)(op10), (size_t)(op11)) +#define vloxseg5ei16_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei16_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (const _Float16 *)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vloxseg5ei16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei16_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vbool16_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (const _Float16 *)(op11), (vuint16m1_t)(op12), (size_t)(op13)) +#define vloxseg5ei16_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei16_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (const _Float16 *)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vloxseg5ei16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei16_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vbool32_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (const _Float16 *)(op11), (vuint16mf2_t)(op12), (size_t)(op13)) +#define vloxseg5ei16_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei16_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (const _Float16 *)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vloxseg5ei16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei16_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vbool64_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (const _Float16 *)(op11), (vuint16mf4_t)(op12), (size_t)(op13)) +#define vloxseg6ei16_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei16_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (const _Float16 *)(op6), (vuint16m1_t)(op7), (size_t)(op8)) +#define vloxseg6ei16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei16_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vbool16_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (vfloat16m1_t)(op11), (vfloat16m1_t)(op12), (const _Float16 *)(op13), (vuint16m1_t)(op14), (size_t)(op15)) +#define vloxseg6ei16_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei16_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (const _Float16 *)(op6), (vuint16mf2_t)(op7), (size_t)(op8)) +#define vloxseg6ei16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei16_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vbool32_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (vfloat16mf2_t)(op11), (vfloat16mf2_t)(op12), (const _Float16 *)(op13), (vuint16mf2_t)(op14), (size_t)(op15)) +#define vloxseg6ei16_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei16_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (const _Float16 *)(op6), (vuint16mf4_t)(op7), (size_t)(op8)) +#define vloxseg6ei16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei16_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vbool64_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (vfloat16mf4_t)(op11), (vfloat16mf4_t)(op12), (const _Float16 *)(op13), (vuint16mf4_t)(op14), (size_t)(op15)) +#define vloxseg7ei16_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei16_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vfloat16m1_t *)(op6), (const _Float16 *)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vloxseg7ei16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei16_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vfloat16m1_t *)(op6), (vbool16_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (vfloat16m1_t)(op11), (vfloat16m1_t)(op12), (vfloat16m1_t)(op13), (vfloat16m1_t)(op14), (const _Float16 *)(op15), (vuint16m1_t)(op16), (size_t)(op17)) +#define vloxseg7ei16_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei16_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vfloat16mf2_t *)(op6), (const _Float16 *)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vloxseg7ei16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei16_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vfloat16mf2_t *)(op6), (vbool32_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (vfloat16mf2_t)(op11), (vfloat16mf2_t)(op12), (vfloat16mf2_t)(op13), (vfloat16mf2_t)(op14), (const _Float16 *)(op15), (vuint16mf2_t)(op16), (size_t)(op17)) +#define vloxseg7ei16_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei16_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vfloat16mf4_t *)(op6), (const _Float16 *)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vloxseg7ei16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei16_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vfloat16mf4_t *)(op6), (vbool64_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (vfloat16mf4_t)(op11), (vfloat16mf4_t)(op12), (vfloat16mf4_t)(op13), (vfloat16mf4_t)(op14), (const _Float16 *)(op15), (vuint16mf4_t)(op16), (size_t)(op17)) +#define vloxseg8ei16_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei16_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vfloat16m1_t *)(op6), (vfloat16m1_t *)(op7), (const _Float16 *)(op8), (vuint16m1_t)(op9), (size_t)(op10)) +#define vloxseg8ei16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei16_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vfloat16m1_t *)(op6), (vfloat16m1_t *)(op7), (vbool16_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (vfloat16m1_t)(op11), (vfloat16m1_t)(op12), (vfloat16m1_t)(op13), (vfloat16m1_t)(op14), (vfloat16m1_t)(op15), (vfloat16m1_t)(op16), (const _Float16 *)(op17), (vuint16m1_t)(op18), (size_t)(op19)) +#define vloxseg8ei16_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei16_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vfloat16mf2_t *)(op6), (vfloat16mf2_t *)(op7), (const _Float16 *)(op8), (vuint16mf2_t)(op9), (size_t)(op10)) +#define vloxseg8ei16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei16_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vfloat16mf2_t *)(op6), (vfloat16mf2_t *)(op7), (vbool32_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (vfloat16mf2_t)(op11), (vfloat16mf2_t)(op12), (vfloat16mf2_t)(op13), (vfloat16mf2_t)(op14), (vfloat16mf2_t)(op15), (vfloat16mf2_t)(op16), (const _Float16 *)(op17), (vuint16mf2_t)(op18), (size_t)(op19)) +#define vloxseg8ei16_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei16_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vfloat16mf4_t *)(op6), (vfloat16mf4_t *)(op7), (const _Float16 *)(op8), (vuint16mf4_t)(op9), (size_t)(op10)) +#define vloxseg8ei16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei16_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vfloat16mf4_t *)(op6), (vfloat16mf4_t *)(op7), (vbool64_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (vfloat16mf4_t)(op11), (vfloat16mf4_t)(op12), (vfloat16mf4_t)(op13), (vfloat16mf4_t)(op14), (vfloat16mf4_t)(op15), (vfloat16mf4_t)(op16), (const _Float16 *)(op17), (vuint16mf4_t)(op18), (size_t)(op19)) +#define vloxseg2ei32_v_f16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (const _Float16 *)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vbool16_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (const _Float16 *)(op5), (vuint32m2_t)(op6), (size_t)(op7)) +#define vloxseg2ei32_v_f16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_f16m2((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (const _Float16 *)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_f16m2_m((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vbool8_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (const _Float16 *)(op5), (vuint32m4_t)(op6), (size_t)(op7)) +#define vloxseg2ei32_v_f16m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_f16m4((vfloat16m4_t *)(op0), (vfloat16m4_t *)(op1), (const _Float16 *)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_f16m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_f16m4_m((vfloat16m4_t *)(op0), (vfloat16m4_t *)(op1), (vbool4_t)(op2), (vfloat16m4_t)(op3), (vfloat16m4_t)(op4), (const _Float16 *)(op5), (vuint32m8_t)(op6), (size_t)(op7)) +#define vloxseg2ei32_v_f16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (const _Float16 *)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vbool32_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (const _Float16 *)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vloxseg2ei32_v_f16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei32_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (const _Float16 *)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vloxseg2ei32_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei32_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vbool64_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (const _Float16 *)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vloxseg3ei32_v_f16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei32_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (const _Float16 *)(op3), (vuint32m2_t)(op4), (size_t)(op5)) +#define vloxseg3ei32_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei32_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vbool16_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (const _Float16 *)(op7), (vuint32m2_t)(op8), (size_t)(op9)) +#define vloxseg3ei32_v_f16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei32_v_f16m2((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vfloat16m2_t *)(op2), (const _Float16 *)(op3), (vuint32m4_t)(op4), (size_t)(op5)) +#define vloxseg3ei32_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei32_v_f16m2_m((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vfloat16m2_t *)(op2), (vbool8_t)(op3), (vfloat16m2_t)(op4), (vfloat16m2_t)(op5), (vfloat16m2_t)(op6), (const _Float16 *)(op7), (vuint32m4_t)(op8), (size_t)(op9)) +#define vloxseg3ei32_v_f16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei32_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (const _Float16 *)(op3), (vuint32m1_t)(op4), (size_t)(op5)) +#define vloxseg3ei32_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei32_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vbool32_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (const _Float16 *)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vloxseg3ei32_v_f16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei32_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (const _Float16 *)(op3), (vuint32mf2_t)(op4), (size_t)(op5)) +#define vloxseg3ei32_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei32_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vbool64_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (const _Float16 *)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vloxseg4ei32_v_f16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei32_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (const _Float16 *)(op4), (vuint32m2_t)(op5), (size_t)(op6)) +#define vloxseg4ei32_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei32_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vbool16_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (const _Float16 *)(op9), (vuint32m2_t)(op10), (size_t)(op11)) +#define vloxseg4ei32_v_f16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei32_v_f16m2((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vfloat16m2_t *)(op2), (vfloat16m2_t *)(op3), (const _Float16 *)(op4), (vuint32m4_t)(op5), (size_t)(op6)) +#define vloxseg4ei32_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei32_v_f16m2_m((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vfloat16m2_t *)(op2), (vfloat16m2_t *)(op3), (vbool8_t)(op4), (vfloat16m2_t)(op5), (vfloat16m2_t)(op6), (vfloat16m2_t)(op7), (vfloat16m2_t)(op8), (const _Float16 *)(op9), (vuint32m4_t)(op10), (size_t)(op11)) +#define vloxseg4ei32_v_f16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei32_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (const _Float16 *)(op4), (vuint32m1_t)(op5), (size_t)(op6)) +#define vloxseg4ei32_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei32_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vbool32_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (const _Float16 *)(op9), (vuint32m1_t)(op10), (size_t)(op11)) +#define vloxseg4ei32_v_f16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei32_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (const _Float16 *)(op4), (vuint32mf2_t)(op5), (size_t)(op6)) +#define vloxseg4ei32_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei32_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vbool64_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (const _Float16 *)(op9), (vuint32mf2_t)(op10), (size_t)(op11)) +#define vloxseg5ei32_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei32_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (const _Float16 *)(op5), (vuint32m2_t)(op6), (size_t)(op7)) +#define vloxseg5ei32_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei32_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vbool16_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (const _Float16 *)(op11), (vuint32m2_t)(op12), (size_t)(op13)) +#define vloxseg5ei32_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei32_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (const _Float16 *)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vloxseg5ei32_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei32_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vbool32_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (const _Float16 *)(op11), (vuint32m1_t)(op12), (size_t)(op13)) +#define vloxseg5ei32_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei32_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (const _Float16 *)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vloxseg5ei32_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei32_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vbool64_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (const _Float16 *)(op11), (vuint32mf2_t)(op12), (size_t)(op13)) +#define vloxseg6ei32_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei32_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (const _Float16 *)(op6), (vuint32m2_t)(op7), (size_t)(op8)) +#define vloxseg6ei32_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei32_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vbool16_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (vfloat16m1_t)(op11), (vfloat16m1_t)(op12), (const _Float16 *)(op13), (vuint32m2_t)(op14), (size_t)(op15)) +#define vloxseg6ei32_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei32_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (const _Float16 *)(op6), (vuint32m1_t)(op7), (size_t)(op8)) +#define vloxseg6ei32_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei32_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vbool32_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (vfloat16mf2_t)(op11), (vfloat16mf2_t)(op12), (const _Float16 *)(op13), (vuint32m1_t)(op14), (size_t)(op15)) +#define vloxseg6ei32_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei32_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (const _Float16 *)(op6), (vuint32mf2_t)(op7), (size_t)(op8)) +#define vloxseg6ei32_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei32_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vbool64_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (vfloat16mf4_t)(op11), (vfloat16mf4_t)(op12), (const _Float16 *)(op13), (vuint32mf2_t)(op14), (size_t)(op15)) +#define vloxseg7ei32_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei32_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vfloat16m1_t *)(op6), (const _Float16 *)(op7), (vuint32m2_t)(op8), (size_t)(op9)) +#define vloxseg7ei32_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei32_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vfloat16m1_t *)(op6), (vbool16_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (vfloat16m1_t)(op11), (vfloat16m1_t)(op12), (vfloat16m1_t)(op13), (vfloat16m1_t)(op14), (const _Float16 *)(op15), (vuint32m2_t)(op16), (size_t)(op17)) +#define vloxseg7ei32_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei32_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vfloat16mf2_t *)(op6), (const _Float16 *)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vloxseg7ei32_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei32_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vfloat16mf2_t *)(op6), (vbool32_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (vfloat16mf2_t)(op11), (vfloat16mf2_t)(op12), (vfloat16mf2_t)(op13), (vfloat16mf2_t)(op14), (const _Float16 *)(op15), (vuint32m1_t)(op16), (size_t)(op17)) +#define vloxseg7ei32_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei32_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vfloat16mf4_t *)(op6), (const _Float16 *)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vloxseg7ei32_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei32_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vfloat16mf4_t *)(op6), (vbool64_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (vfloat16mf4_t)(op11), (vfloat16mf4_t)(op12), (vfloat16mf4_t)(op13), (vfloat16mf4_t)(op14), (const _Float16 *)(op15), (vuint32mf2_t)(op16), (size_t)(op17)) +#define vloxseg8ei32_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei32_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vfloat16m1_t *)(op6), (vfloat16m1_t *)(op7), (const _Float16 *)(op8), (vuint32m2_t)(op9), (size_t)(op10)) +#define vloxseg8ei32_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei32_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vfloat16m1_t *)(op6), (vfloat16m1_t *)(op7), (vbool16_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (vfloat16m1_t)(op11), (vfloat16m1_t)(op12), (vfloat16m1_t)(op13), (vfloat16m1_t)(op14), (vfloat16m1_t)(op15), (vfloat16m1_t)(op16), (const _Float16 *)(op17), (vuint32m2_t)(op18), (size_t)(op19)) +#define vloxseg8ei32_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei32_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vfloat16mf2_t *)(op6), (vfloat16mf2_t *)(op7), (const _Float16 *)(op8), (vuint32m1_t)(op9), (size_t)(op10)) +#define vloxseg8ei32_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei32_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vfloat16mf2_t *)(op6), (vfloat16mf2_t *)(op7), (vbool32_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (vfloat16mf2_t)(op11), (vfloat16mf2_t)(op12), (vfloat16mf2_t)(op13), (vfloat16mf2_t)(op14), (vfloat16mf2_t)(op15), (vfloat16mf2_t)(op16), (const _Float16 *)(op17), (vuint32m1_t)(op18), (size_t)(op19)) +#define vloxseg8ei32_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei32_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vfloat16mf4_t *)(op6), (vfloat16mf4_t *)(op7), (const _Float16 *)(op8), (vuint32mf2_t)(op9), (size_t)(op10)) +#define vloxseg8ei32_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei32_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vfloat16mf4_t *)(op6), (vfloat16mf4_t *)(op7), (vbool64_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (vfloat16mf4_t)(op11), (vfloat16mf4_t)(op12), (vfloat16mf4_t)(op13), (vfloat16mf4_t)(op14), (vfloat16mf4_t)(op15), (vfloat16mf4_t)(op16), (const _Float16 *)(op17), (vuint32mf2_t)(op18), (size_t)(op19)) +#define vloxseg2ei64_v_f16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei64_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (const _Float16 *)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vloxseg2ei64_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei64_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vbool16_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (const _Float16 *)(op5), (vuint64m4_t)(op6), (size_t)(op7)) +#define vloxseg2ei64_v_f16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei64_v_f16m2((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (const _Float16 *)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vloxseg2ei64_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei64_v_f16m2_m((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vbool8_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (const _Float16 *)(op5), (vuint64m8_t)(op6), (size_t)(op7)) +#define vloxseg2ei64_v_f16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei64_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (const _Float16 *)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vloxseg2ei64_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei64_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vbool32_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (const _Float16 *)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vloxseg2ei64_v_f16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vloxseg2ei64_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (const _Float16 *)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vloxseg2ei64_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg2ei64_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vbool64_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (const _Float16 *)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vloxseg3ei64_v_f16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei64_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (const _Float16 *)(op3), (vuint64m4_t)(op4), (size_t)(op5)) +#define vloxseg3ei64_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei64_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vbool16_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (const _Float16 *)(op7), (vuint64m4_t)(op8), (size_t)(op9)) +#define vloxseg3ei64_v_f16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei64_v_f16m2((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vfloat16m2_t *)(op2), (const _Float16 *)(op3), (vuint64m8_t)(op4), (size_t)(op5)) +#define vloxseg3ei64_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei64_v_f16m2_m((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vfloat16m2_t *)(op2), (vbool8_t)(op3), (vfloat16m2_t)(op4), (vfloat16m2_t)(op5), (vfloat16m2_t)(op6), (const _Float16 *)(op7), (vuint64m8_t)(op8), (size_t)(op9)) +#define vloxseg3ei64_v_f16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei64_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (const _Float16 *)(op3), (vuint64m2_t)(op4), (size_t)(op5)) +#define vloxseg3ei64_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei64_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vbool32_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (const _Float16 *)(op7), (vuint64m2_t)(op8), (size_t)(op9)) +#define vloxseg3ei64_v_f16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vloxseg3ei64_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (const _Float16 *)(op3), (vuint64m1_t)(op4), (size_t)(op5)) +#define vloxseg3ei64_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg3ei64_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vbool64_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (const _Float16 *)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vloxseg4ei64_v_f16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei64_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (const _Float16 *)(op4), (vuint64m4_t)(op5), (size_t)(op6)) +#define vloxseg4ei64_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei64_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vbool16_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (const _Float16 *)(op9), (vuint64m4_t)(op10), (size_t)(op11)) +#define vloxseg4ei64_v_f16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei64_v_f16m2((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vfloat16m2_t *)(op2), (vfloat16m2_t *)(op3), (const _Float16 *)(op4), (vuint64m8_t)(op5), (size_t)(op6)) +#define vloxseg4ei64_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei64_v_f16m2_m((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vfloat16m2_t *)(op2), (vfloat16m2_t *)(op3), (vbool8_t)(op4), (vfloat16m2_t)(op5), (vfloat16m2_t)(op6), (vfloat16m2_t)(op7), (vfloat16m2_t)(op8), (const _Float16 *)(op9), (vuint64m8_t)(op10), (size_t)(op11)) +#define vloxseg4ei64_v_f16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei64_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (const _Float16 *)(op4), (vuint64m2_t)(op5), (size_t)(op6)) +#define vloxseg4ei64_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei64_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vbool32_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (const _Float16 *)(op9), (vuint64m2_t)(op10), (size_t)(op11)) +#define vloxseg4ei64_v_f16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vloxseg4ei64_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (const _Float16 *)(op4), (vuint64m1_t)(op5), (size_t)(op6)) +#define vloxseg4ei64_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vloxseg4ei64_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vbool64_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (const _Float16 *)(op9), (vuint64m1_t)(op10), (size_t)(op11)) +#define vloxseg5ei64_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei64_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (const _Float16 *)(op5), (vuint64m4_t)(op6), (size_t)(op7)) +#define vloxseg5ei64_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei64_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vbool16_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (const _Float16 *)(op11), (vuint64m4_t)(op12), (size_t)(op13)) +#define vloxseg5ei64_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei64_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (const _Float16 *)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vloxseg5ei64_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei64_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vbool32_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (const _Float16 *)(op11), (vuint64m2_t)(op12), (size_t)(op13)) +#define vloxseg5ei64_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vloxseg5ei64_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (const _Float16 *)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vloxseg5ei64_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vloxseg5ei64_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vbool64_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (const _Float16 *)(op11), (vuint64m1_t)(op12), (size_t)(op13)) +#define vloxseg6ei64_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei64_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (const _Float16 *)(op6), (vuint64m4_t)(op7), (size_t)(op8)) +#define vloxseg6ei64_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei64_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vbool16_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (vfloat16m1_t)(op11), (vfloat16m1_t)(op12), (const _Float16 *)(op13), (vuint64m4_t)(op14), (size_t)(op15)) +#define vloxseg6ei64_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei64_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (const _Float16 *)(op6), (vuint64m2_t)(op7), (size_t)(op8)) +#define vloxseg6ei64_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei64_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vbool32_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (vfloat16mf2_t)(op11), (vfloat16mf2_t)(op12), (const _Float16 *)(op13), (vuint64m2_t)(op14), (size_t)(op15)) +#define vloxseg6ei64_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vloxseg6ei64_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (const _Float16 *)(op6), (vuint64m1_t)(op7), (size_t)(op8)) +#define vloxseg6ei64_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vloxseg6ei64_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vbool64_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (vfloat16mf4_t)(op11), (vfloat16mf4_t)(op12), (const _Float16 *)(op13), (vuint64m1_t)(op14), (size_t)(op15)) +#define vloxseg7ei64_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei64_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vfloat16m1_t *)(op6), (const _Float16 *)(op7), (vuint64m4_t)(op8), (size_t)(op9)) +#define vloxseg7ei64_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei64_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vfloat16m1_t *)(op6), (vbool16_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (vfloat16m1_t)(op11), (vfloat16m1_t)(op12), (vfloat16m1_t)(op13), (vfloat16m1_t)(op14), (const _Float16 *)(op15), (vuint64m4_t)(op16), (size_t)(op17)) +#define vloxseg7ei64_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei64_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vfloat16mf2_t *)(op6), (const _Float16 *)(op7), (vuint64m2_t)(op8), (size_t)(op9)) +#define vloxseg7ei64_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei64_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vfloat16mf2_t *)(op6), (vbool32_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (vfloat16mf2_t)(op11), (vfloat16mf2_t)(op12), (vfloat16mf2_t)(op13), (vfloat16mf2_t)(op14), (const _Float16 *)(op15), (vuint64m2_t)(op16), (size_t)(op17)) +#define vloxseg7ei64_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vloxseg7ei64_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vfloat16mf4_t *)(op6), (const _Float16 *)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vloxseg7ei64_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vloxseg7ei64_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vfloat16mf4_t *)(op6), (vbool64_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (vfloat16mf4_t)(op11), (vfloat16mf4_t)(op12), (vfloat16mf4_t)(op13), (vfloat16mf4_t)(op14), (const _Float16 *)(op15), (vuint64m1_t)(op16), (size_t)(op17)) +#define vloxseg8ei64_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei64_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vfloat16m1_t *)(op6), (vfloat16m1_t *)(op7), (const _Float16 *)(op8), (vuint64m4_t)(op9), (size_t)(op10)) +#define vloxseg8ei64_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei64_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vfloat16m1_t *)(op6), (vfloat16m1_t *)(op7), (vbool16_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (vfloat16m1_t)(op11), (vfloat16m1_t)(op12), (vfloat16m1_t)(op13), (vfloat16m1_t)(op14), (vfloat16m1_t)(op15), (vfloat16m1_t)(op16), (const _Float16 *)(op17), (vuint64m4_t)(op18), (size_t)(op19)) +#define vloxseg8ei64_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei64_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vfloat16mf2_t *)(op6), (vfloat16mf2_t *)(op7), (const _Float16 *)(op8), (vuint64m2_t)(op9), (size_t)(op10)) +#define vloxseg8ei64_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei64_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vfloat16mf2_t *)(op6), (vfloat16mf2_t *)(op7), (vbool32_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (vfloat16mf2_t)(op11), (vfloat16mf2_t)(op12), (vfloat16mf2_t)(op13), (vfloat16mf2_t)(op14), (vfloat16mf2_t)(op15), (vfloat16mf2_t)(op16), (const _Float16 *)(op17), (vuint64m2_t)(op18), (size_t)(op19)) +#define vloxseg8ei64_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vloxseg8ei64_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vfloat16mf4_t *)(op6), (vfloat16mf4_t *)(op7), (const _Float16 *)(op8), (vuint64m1_t)(op9), (size_t)(op10)) +#define vloxseg8ei64_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vloxseg8ei64_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vfloat16mf4_t *)(op6), (vfloat16mf4_t *)(op7), (vbool64_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (vfloat16mf4_t)(op11), (vfloat16mf4_t)(op12), (vfloat16mf4_t)(op13), (vfloat16mf4_t)(op14), (vfloat16mf4_t)(op15), (vfloat16mf4_t)(op16), (const _Float16 *)(op17), (vuint64m1_t)(op18), (size_t)(op19)) +#define vsseg2e16_v_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e16_v_f16m1((_Float16 *)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (size_t)(op3)) +#define vsseg2e16_v_f16m1_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e16_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vsseg2e16_v_f16m2(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e16_v_f16m2((_Float16 *)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (size_t)(op3)) +#define vsseg2e16_v_f16m2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e16_v_f16m2_m((vbool8_t)(op0), (_Float16 *)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vsseg2e16_v_f16m4(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e16_v_f16m4((_Float16 *)(op0), (vfloat16m4_t)(op1), (vfloat16m4_t)(op2), (size_t)(op3)) +#define vsseg2e16_v_f16m4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e16_v_f16m4_m((vbool4_t)(op0), (_Float16 *)(op1), (vfloat16m4_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vsseg2e16_v_f16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e16_v_f16mf2((_Float16 *)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (size_t)(op3)) +#define vsseg2e16_v_f16mf2_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e16_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vsseg2e16_v_f16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vsseg2e16_v_f16mf4((_Float16 *)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (size_t)(op3)) +#define vsseg2e16_v_f16mf4_m(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg2e16_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vsseg3e16_v_f16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg3e16_v_f16m1((_Float16 *)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vsseg3e16_v_f16m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg3e16_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (size_t)(op5)) +#define vsseg3e16_v_f16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg3e16_v_f16m2((_Float16 *)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vsseg3e16_v_f16m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg3e16_v_f16m2_m((vbool8_t)(op0), (_Float16 *)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (size_t)(op5)) +#define vsseg3e16_v_f16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg3e16_v_f16mf2((_Float16 *)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vsseg3e16_v_f16mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg3e16_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (size_t)(op5)) +#define vsseg3e16_v_f16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsseg3e16_v_f16mf4((_Float16 *)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vsseg3e16_v_f16mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg3e16_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (size_t)(op5)) +#define vsseg4e16_v_f16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg4e16_v_f16m1((_Float16 *)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (size_t)(op5)) +#define vsseg4e16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg4e16_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (size_t)(op6)) +#define vsseg4e16_v_f16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg4e16_v_f16m2((_Float16 *)(op0), (vfloat16m2_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (size_t)(op5)) +#define vsseg4e16_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg4e16_v_f16m2_m((vbool8_t)(op0), (_Float16 *)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (vfloat16m2_t)(op5), (size_t)(op6)) +#define vsseg4e16_v_f16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg4e16_v_f16mf2((_Float16 *)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (size_t)(op5)) +#define vsseg4e16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg4e16_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (size_t)(op6)) +#define vsseg4e16_v_f16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsseg4e16_v_f16mf4((_Float16 *)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (size_t)(op5)) +#define vsseg4e16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg4e16_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (size_t)(op6)) +#define vsseg5e16_v_f16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg5e16_v_f16m1((_Float16 *)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (size_t)(op6)) +#define vsseg5e16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg5e16_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (size_t)(op7)) +#define vsseg5e16_v_f16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg5e16_v_f16mf2((_Float16 *)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (size_t)(op6)) +#define vsseg5e16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg5e16_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (size_t)(op7)) +#define vsseg5e16_v_f16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsseg5e16_v_f16mf4((_Float16 *)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (size_t)(op6)) +#define vsseg5e16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg5e16_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (size_t)(op7)) +#define vsseg6e16_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg6e16_v_f16m1((_Float16 *)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (size_t)(op7)) +#define vsseg6e16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg6e16_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (size_t)(op8)) +#define vsseg6e16_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg6e16_v_f16mf2((_Float16 *)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (size_t)(op7)) +#define vsseg6e16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg6e16_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (size_t)(op8)) +#define vsseg6e16_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsseg6e16_v_f16mf4((_Float16 *)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (size_t)(op7)) +#define vsseg6e16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg6e16_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (size_t)(op8)) +#define vsseg7e16_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg7e16_v_f16m1((_Float16 *)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (size_t)(op8)) +#define vsseg7e16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg7e16_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (size_t)(op9)) +#define vsseg7e16_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg7e16_v_f16mf2((_Float16 *)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (size_t)(op8)) +#define vsseg7e16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg7e16_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (size_t)(op9)) +#define vsseg7e16_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsseg7e16_v_f16mf4((_Float16 *)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (size_t)(op8)) +#define vsseg7e16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg7e16_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (size_t)(op9)) +#define vsseg8e16_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg8e16_v_f16m1((_Float16 *)(op0), (vfloat16m1_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (size_t)(op9)) +#define vsseg8e16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsseg8e16_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (size_t)(op10)) +#define vsseg8e16_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg8e16_v_f16mf2((_Float16 *)(op0), (vfloat16mf2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (size_t)(op9)) +#define vsseg8e16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsseg8e16_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (size_t)(op10)) +#define vsseg8e16_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsseg8e16_v_f16mf4((_Float16 *)(op0), (vfloat16mf4_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (size_t)(op9)) +#define vsseg8e16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsseg8e16_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (size_t)(op10)) +#define vssseg2e16_v_f16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e16_v_f16m1((_Float16 *)(op0), (ptrdiff_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vssseg2e16_v_f16m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e16_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (ptrdiff_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (size_t)(op5)) +#define vssseg2e16_v_f16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e16_v_f16m2((_Float16 *)(op0), (ptrdiff_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vssseg2e16_v_f16m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e16_v_f16m2_m((vbool8_t)(op0), (_Float16 *)(op1), (ptrdiff_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (size_t)(op5)) +#define vssseg2e16_v_f16m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e16_v_f16m4((_Float16 *)(op0), (ptrdiff_t)(op1), (vfloat16m4_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vssseg2e16_v_f16m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e16_v_f16m4_m((vbool4_t)(op0), (_Float16 *)(op1), (ptrdiff_t)(op2), (vfloat16m4_t)(op3), (vfloat16m4_t)(op4), (size_t)(op5)) +#define vssseg2e16_v_f16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e16_v_f16mf2((_Float16 *)(op0), (ptrdiff_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vssseg2e16_v_f16mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e16_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (ptrdiff_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (size_t)(op5)) +#define vssseg2e16_v_f16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vssseg2e16_v_f16mf4((_Float16 *)(op0), (ptrdiff_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vssseg2e16_v_f16mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg2e16_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (ptrdiff_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (size_t)(op5)) +#define vssseg3e16_v_f16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg3e16_v_f16m1((_Float16 *)(op0), (ptrdiff_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (size_t)(op5)) +#define vssseg3e16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg3e16_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (ptrdiff_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (size_t)(op6)) +#define vssseg3e16_v_f16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg3e16_v_f16m2((_Float16 *)(op0), (ptrdiff_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (size_t)(op5)) +#define vssseg3e16_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg3e16_v_f16m2_m((vbool8_t)(op0), (_Float16 *)(op1), (ptrdiff_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (vfloat16m2_t)(op5), (size_t)(op6)) +#define vssseg3e16_v_f16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg3e16_v_f16mf2((_Float16 *)(op0), (ptrdiff_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (size_t)(op5)) +#define vssseg3e16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg3e16_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (ptrdiff_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (size_t)(op6)) +#define vssseg3e16_v_f16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vssseg3e16_v_f16mf4((_Float16 *)(op0), (ptrdiff_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (size_t)(op5)) +#define vssseg3e16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg3e16_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (ptrdiff_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (size_t)(op6)) +#define vssseg4e16_v_f16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg4e16_v_f16m1((_Float16 *)(op0), (ptrdiff_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (size_t)(op6)) +#define vssseg4e16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg4e16_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (ptrdiff_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (size_t)(op7)) +#define vssseg4e16_v_f16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg4e16_v_f16m2((_Float16 *)(op0), (ptrdiff_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (vfloat16m2_t)(op5), (size_t)(op6)) +#define vssseg4e16_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg4e16_v_f16m2_m((vbool8_t)(op0), (_Float16 *)(op1), (ptrdiff_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (vfloat16m2_t)(op5), (vfloat16m2_t)(op6), (size_t)(op7)) +#define vssseg4e16_v_f16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg4e16_v_f16mf2((_Float16 *)(op0), (ptrdiff_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (size_t)(op6)) +#define vssseg4e16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg4e16_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (ptrdiff_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (size_t)(op7)) +#define vssseg4e16_v_f16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vssseg4e16_v_f16mf4((_Float16 *)(op0), (ptrdiff_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (size_t)(op6)) +#define vssseg4e16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg4e16_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (ptrdiff_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (size_t)(op7)) +#define vssseg5e16_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg5e16_v_f16m1((_Float16 *)(op0), (ptrdiff_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (size_t)(op7)) +#define vssseg5e16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg5e16_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (ptrdiff_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (size_t)(op8)) +#define vssseg5e16_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg5e16_v_f16mf2((_Float16 *)(op0), (ptrdiff_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (size_t)(op7)) +#define vssseg5e16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg5e16_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (ptrdiff_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (size_t)(op8)) +#define vssseg5e16_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vssseg5e16_v_f16mf4((_Float16 *)(op0), (ptrdiff_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (size_t)(op7)) +#define vssseg5e16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg5e16_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (ptrdiff_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (size_t)(op8)) +#define vssseg6e16_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg6e16_v_f16m1((_Float16 *)(op0), (ptrdiff_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (size_t)(op8)) +#define vssseg6e16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg6e16_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (ptrdiff_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (size_t)(op9)) +#define vssseg6e16_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg6e16_v_f16mf2((_Float16 *)(op0), (ptrdiff_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (size_t)(op8)) +#define vssseg6e16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg6e16_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (ptrdiff_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (size_t)(op9)) +#define vssseg6e16_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vssseg6e16_v_f16mf4((_Float16 *)(op0), (ptrdiff_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (size_t)(op8)) +#define vssseg6e16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg6e16_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (ptrdiff_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (size_t)(op9)) +#define vssseg7e16_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg7e16_v_f16m1((_Float16 *)(op0), (ptrdiff_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (size_t)(op9)) +#define vssseg7e16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg7e16_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (ptrdiff_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (size_t)(op10)) +#define vssseg7e16_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg7e16_v_f16mf2((_Float16 *)(op0), (ptrdiff_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (size_t)(op9)) +#define vssseg7e16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg7e16_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (ptrdiff_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (size_t)(op10)) +#define vssseg7e16_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vssseg7e16_v_f16mf4((_Float16 *)(op0), (ptrdiff_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (size_t)(op9)) +#define vssseg7e16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg7e16_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (ptrdiff_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (size_t)(op10)) +#define vssseg8e16_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg8e16_v_f16m1((_Float16 *)(op0), (ptrdiff_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (size_t)(op10)) +#define vssseg8e16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vssseg8e16_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (ptrdiff_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (size_t)(op11)) +#define vssseg8e16_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg8e16_v_f16mf2((_Float16 *)(op0), (ptrdiff_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (size_t)(op10)) +#define vssseg8e16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vssseg8e16_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (ptrdiff_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (size_t)(op11)) +#define vssseg8e16_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vssseg8e16_v_f16mf4((_Float16 *)(op0), (ptrdiff_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (size_t)(op10)) +#define vssseg8e16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vssseg8e16_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (ptrdiff_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (size_t)(op11)) +#define vsuxseg2ei8_v_f16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_f16m1((_Float16 *)(op0), (vuint8mf2_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_f16m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint8mf2_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (size_t)(op5)) +#define vsuxseg2ei8_v_f16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_f16m2((_Float16 *)(op0), (vuint8m1_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_f16m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_f16m2_m((vbool8_t)(op0), (_Float16 *)(op1), (vuint8m1_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei8_v_f16m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_f16m4((_Float16 *)(op0), (vuint8m2_t)(op1), (vfloat16m4_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_f16m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_f16m4_m((vbool4_t)(op0), (_Float16 *)(op1), (vuint8m2_t)(op2), (vfloat16m4_t)(op3), (vfloat16m4_t)(op4), (size_t)(op5)) +#define vsuxseg2ei8_v_f16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_f16mf2((_Float16 *)(op0), (vuint8mf4_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_f16mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint8mf4_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei8_v_f16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei8_v_f16mf4((_Float16 *)(op0), (vuint8mf8_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei8_v_f16mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei8_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint8mf8_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (size_t)(op5)) +#define vsuxseg3ei8_v_f16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei8_v_f16m1((_Float16 *)(op0), (vuint8mf2_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (size_t)(op5)) +#define vsuxseg3ei8_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei8_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint8mf2_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (size_t)(op6)) +#define vsuxseg3ei8_v_f16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei8_v_f16m2((_Float16 *)(op0), (vuint8m1_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei8_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei8_v_f16m2_m((vbool8_t)(op0), (_Float16 *)(op1), (vuint8m1_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (vfloat16m2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei8_v_f16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei8_v_f16mf2((_Float16 *)(op0), (vuint8mf4_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei8_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei8_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint8mf4_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei8_v_f16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei8_v_f16mf4((_Float16 *)(op0), (vuint8mf8_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (size_t)(op5)) +#define vsuxseg3ei8_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei8_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint8mf8_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (size_t)(op6)) +#define vsuxseg4ei8_v_f16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei8_v_f16m1((_Float16 *)(op0), (vuint8mf2_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (size_t)(op6)) +#define vsuxseg4ei8_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei8_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint8mf2_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (size_t)(op7)) +#define vsuxseg4ei8_v_f16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei8_v_f16m2((_Float16 *)(op0), (vuint8m1_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (vfloat16m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei8_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei8_v_f16m2_m((vbool8_t)(op0), (_Float16 *)(op1), (vuint8m1_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (vfloat16m2_t)(op5), (vfloat16m2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei8_v_f16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei8_v_f16mf2((_Float16 *)(op0), (vuint8mf4_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei8_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei8_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint8mf4_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei8_v_f16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei8_v_f16mf4((_Float16 *)(op0), (vuint8mf8_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (size_t)(op6)) +#define vsuxseg4ei8_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei8_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint8mf8_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (size_t)(op7)) +#define vsuxseg5ei8_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei8_v_f16m1((_Float16 *)(op0), (vuint8mf2_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (size_t)(op7)) +#define vsuxseg5ei8_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei8_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint8mf2_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (size_t)(op8)) +#define vsuxseg5ei8_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei8_v_f16mf2((_Float16 *)(op0), (vuint8mf4_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei8_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei8_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint8mf4_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (size_t)(op8)) +#define vsuxseg5ei8_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei8_v_f16mf4((_Float16 *)(op0), (vuint8mf8_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (size_t)(op7)) +#define vsuxseg5ei8_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei8_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint8mf8_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (size_t)(op8)) +#define vsuxseg6ei8_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei8_v_f16m1((_Float16 *)(op0), (vuint8mf2_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei8_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei8_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint8mf2_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (size_t)(op9)) +#define vsuxseg6ei8_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei8_v_f16mf2((_Float16 *)(op0), (vuint8mf4_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (size_t)(op8)) +#define vsuxseg6ei8_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei8_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint8mf4_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (size_t)(op9)) +#define vsuxseg6ei8_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei8_v_f16mf4((_Float16 *)(op0), (vuint8mf8_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (size_t)(op8)) +#define vsuxseg6ei8_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei8_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint8mf8_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (size_t)(op9)) +#define vsuxseg7ei8_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei8_v_f16m1((_Float16 *)(op0), (vuint8mf2_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei8_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei8_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint8mf2_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (size_t)(op10)) +#define vsuxseg7ei8_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei8_v_f16mf2((_Float16 *)(op0), (vuint8mf4_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (size_t)(op9)) +#define vsuxseg7ei8_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei8_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint8mf4_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (size_t)(op10)) +#define vsuxseg7ei8_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei8_v_f16mf4((_Float16 *)(op0), (vuint8mf8_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (size_t)(op9)) +#define vsuxseg7ei8_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei8_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint8mf8_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (size_t)(op10)) +#define vsuxseg8ei8_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei8_v_f16m1((_Float16 *)(op0), (vuint8mf2_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei8_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei8_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint8mf2_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (size_t)(op11)) +#define vsuxseg8ei8_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei8_v_f16mf2((_Float16 *)(op0), (vuint8mf4_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (size_t)(op10)) +#define vsuxseg8ei8_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei8_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint8mf4_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (size_t)(op11)) +#define vsuxseg8ei8_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei8_v_f16mf4((_Float16 *)(op0), (vuint8mf8_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (size_t)(op10)) +#define vsuxseg8ei8_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei8_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint8mf8_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (size_t)(op11)) +#define vsuxseg2ei16_v_f16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_f16m1((_Float16 *)(op0), (vuint16m1_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_f16m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (size_t)(op5)) +#define vsuxseg2ei16_v_f16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_f16m2((_Float16 *)(op0), (vuint16m2_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_f16m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_f16m2_m((vbool8_t)(op0), (_Float16 *)(op1), (vuint16m2_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei16_v_f16m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_f16m4((_Float16 *)(op0), (vuint16m4_t)(op1), (vfloat16m4_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_f16m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_f16m4_m((vbool4_t)(op0), (_Float16 *)(op1), (vuint16m4_t)(op2), (vfloat16m4_t)(op3), (vfloat16m4_t)(op4), (size_t)(op5)) +#define vsuxseg2ei16_v_f16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_f16mf2((_Float16 *)(op0), (vuint16mf2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_f16mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei16_v_f16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei16_v_f16mf4((_Float16 *)(op0), (vuint16mf4_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei16_v_f16mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei16_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (size_t)(op5)) +#define vsuxseg3ei16_v_f16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei16_v_f16m1((_Float16 *)(op0), (vuint16m1_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (size_t)(op5)) +#define vsuxseg3ei16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei16_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (size_t)(op6)) +#define vsuxseg3ei16_v_f16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei16_v_f16m2((_Float16 *)(op0), (vuint16m2_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei16_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei16_v_f16m2_m((vbool8_t)(op0), (_Float16 *)(op1), (vuint16m2_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (vfloat16m2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei16_v_f16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei16_v_f16mf2((_Float16 *)(op0), (vuint16mf2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei16_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei16_v_f16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei16_v_f16mf4((_Float16 *)(op0), (vuint16mf4_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (size_t)(op5)) +#define vsuxseg3ei16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei16_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (size_t)(op6)) +#define vsuxseg4ei16_v_f16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei16_v_f16m1((_Float16 *)(op0), (vuint16m1_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (size_t)(op6)) +#define vsuxseg4ei16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei16_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (size_t)(op7)) +#define vsuxseg4ei16_v_f16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei16_v_f16m2((_Float16 *)(op0), (vuint16m2_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (vfloat16m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei16_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei16_v_f16m2_m((vbool8_t)(op0), (_Float16 *)(op1), (vuint16m2_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (vfloat16m2_t)(op5), (vfloat16m2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei16_v_f16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei16_v_f16mf2((_Float16 *)(op0), (vuint16mf2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei16_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei16_v_f16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei16_v_f16mf4((_Float16 *)(op0), (vuint16mf4_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (size_t)(op6)) +#define vsuxseg4ei16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei16_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (size_t)(op7)) +#define vsuxseg5ei16_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei16_v_f16m1((_Float16 *)(op0), (vuint16m1_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (size_t)(op7)) +#define vsuxseg5ei16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei16_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (size_t)(op8)) +#define vsuxseg5ei16_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei16_v_f16mf2((_Float16 *)(op0), (vuint16mf2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei16_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (size_t)(op8)) +#define vsuxseg5ei16_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei16_v_f16mf4((_Float16 *)(op0), (vuint16mf4_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (size_t)(op7)) +#define vsuxseg5ei16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei16_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (size_t)(op8)) +#define vsuxseg6ei16_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei16_v_f16m1((_Float16 *)(op0), (vuint16m1_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei16_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (size_t)(op9)) +#define vsuxseg6ei16_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei16_v_f16mf2((_Float16 *)(op0), (vuint16mf2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (size_t)(op8)) +#define vsuxseg6ei16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei16_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (size_t)(op9)) +#define vsuxseg6ei16_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei16_v_f16mf4((_Float16 *)(op0), (vuint16mf4_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (size_t)(op8)) +#define vsuxseg6ei16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei16_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (size_t)(op9)) +#define vsuxseg7ei16_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei16_v_f16m1((_Float16 *)(op0), (vuint16m1_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei16_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (size_t)(op10)) +#define vsuxseg7ei16_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei16_v_f16mf2((_Float16 *)(op0), (vuint16mf2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (size_t)(op9)) +#define vsuxseg7ei16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei16_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (size_t)(op10)) +#define vsuxseg7ei16_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei16_v_f16mf4((_Float16 *)(op0), (vuint16mf4_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (size_t)(op9)) +#define vsuxseg7ei16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei16_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (size_t)(op10)) +#define vsuxseg8ei16_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei16_v_f16m1((_Float16 *)(op0), (vuint16m1_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei16_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (size_t)(op11)) +#define vsuxseg8ei16_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei16_v_f16mf2((_Float16 *)(op0), (vuint16mf2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (size_t)(op10)) +#define vsuxseg8ei16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei16_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (size_t)(op11)) +#define vsuxseg8ei16_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei16_v_f16mf4((_Float16 *)(op0), (vuint16mf4_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (size_t)(op10)) +#define vsuxseg8ei16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei16_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (size_t)(op11)) +#define vsuxseg2ei32_v_f16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_f16m1((_Float16 *)(op0), (vuint32m2_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_f16m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint32m2_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (size_t)(op5)) +#define vsuxseg2ei32_v_f16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_f16m2((_Float16 *)(op0), (vuint32m4_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_f16m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_f16m2_m((vbool8_t)(op0), (_Float16 *)(op1), (vuint32m4_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei32_v_f16m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_f16m4((_Float16 *)(op0), (vuint32m8_t)(op1), (vfloat16m4_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_f16m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_f16m4_m((vbool4_t)(op0), (_Float16 *)(op1), (vuint32m8_t)(op2), (vfloat16m4_t)(op3), (vfloat16m4_t)(op4), (size_t)(op5)) +#define vsuxseg2ei32_v_f16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_f16mf2((_Float16 *)(op0), (vuint32m1_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_f16mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint32m1_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei32_v_f16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei32_v_f16mf4((_Float16 *)(op0), (vuint32mf2_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei32_v_f16mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei32_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint32mf2_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (size_t)(op5)) +#define vsuxseg3ei32_v_f16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei32_v_f16m1((_Float16 *)(op0), (vuint32m2_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (size_t)(op5)) +#define vsuxseg3ei32_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei32_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint32m2_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (size_t)(op6)) +#define vsuxseg3ei32_v_f16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei32_v_f16m2((_Float16 *)(op0), (vuint32m4_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei32_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei32_v_f16m2_m((vbool8_t)(op0), (_Float16 *)(op1), (vuint32m4_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (vfloat16m2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei32_v_f16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei32_v_f16mf2((_Float16 *)(op0), (vuint32m1_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei32_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei32_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint32m1_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei32_v_f16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei32_v_f16mf4((_Float16 *)(op0), (vuint32mf2_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (size_t)(op5)) +#define vsuxseg3ei32_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei32_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint32mf2_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (size_t)(op6)) +#define vsuxseg4ei32_v_f16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei32_v_f16m1((_Float16 *)(op0), (vuint32m2_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (size_t)(op6)) +#define vsuxseg4ei32_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei32_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint32m2_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (size_t)(op7)) +#define vsuxseg4ei32_v_f16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei32_v_f16m2((_Float16 *)(op0), (vuint32m4_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (vfloat16m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei32_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei32_v_f16m2_m((vbool8_t)(op0), (_Float16 *)(op1), (vuint32m4_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (vfloat16m2_t)(op5), (vfloat16m2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei32_v_f16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei32_v_f16mf2((_Float16 *)(op0), (vuint32m1_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei32_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei32_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint32m1_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei32_v_f16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei32_v_f16mf4((_Float16 *)(op0), (vuint32mf2_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (size_t)(op6)) +#define vsuxseg4ei32_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei32_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint32mf2_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (size_t)(op7)) +#define vsuxseg5ei32_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei32_v_f16m1((_Float16 *)(op0), (vuint32m2_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (size_t)(op7)) +#define vsuxseg5ei32_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei32_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint32m2_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (size_t)(op8)) +#define vsuxseg5ei32_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei32_v_f16mf2((_Float16 *)(op0), (vuint32m1_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei32_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei32_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint32m1_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (size_t)(op8)) +#define vsuxseg5ei32_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei32_v_f16mf4((_Float16 *)(op0), (vuint32mf2_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (size_t)(op7)) +#define vsuxseg5ei32_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei32_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint32mf2_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (size_t)(op8)) +#define vsuxseg6ei32_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei32_v_f16m1((_Float16 *)(op0), (vuint32m2_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei32_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei32_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint32m2_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (size_t)(op9)) +#define vsuxseg6ei32_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei32_v_f16mf2((_Float16 *)(op0), (vuint32m1_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (size_t)(op8)) +#define vsuxseg6ei32_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei32_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint32m1_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (size_t)(op9)) +#define vsuxseg6ei32_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei32_v_f16mf4((_Float16 *)(op0), (vuint32mf2_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (size_t)(op8)) +#define vsuxseg6ei32_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei32_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint32mf2_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (size_t)(op9)) +#define vsuxseg7ei32_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei32_v_f16m1((_Float16 *)(op0), (vuint32m2_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei32_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei32_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint32m2_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (size_t)(op10)) +#define vsuxseg7ei32_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei32_v_f16mf2((_Float16 *)(op0), (vuint32m1_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (size_t)(op9)) +#define vsuxseg7ei32_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei32_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint32m1_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (size_t)(op10)) +#define vsuxseg7ei32_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei32_v_f16mf4((_Float16 *)(op0), (vuint32mf2_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (size_t)(op9)) +#define vsuxseg7ei32_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei32_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint32mf2_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (size_t)(op10)) +#define vsuxseg8ei32_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei32_v_f16m1((_Float16 *)(op0), (vuint32m2_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei32_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei32_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint32m2_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (size_t)(op11)) +#define vsuxseg8ei32_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei32_v_f16mf2((_Float16 *)(op0), (vuint32m1_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (size_t)(op10)) +#define vsuxseg8ei32_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei32_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint32m1_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (size_t)(op11)) +#define vsuxseg8ei32_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei32_v_f16mf4((_Float16 *)(op0), (vuint32mf2_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (size_t)(op10)) +#define vsuxseg8ei32_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei32_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint32mf2_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (size_t)(op11)) +#define vsuxseg2ei64_v_f16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei64_v_f16m1((_Float16 *)(op0), (vuint64m4_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vsuxseg2ei64_v_f16m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei64_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint64m4_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (size_t)(op5)) +#define vsuxseg2ei64_v_f16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei64_v_f16m2((_Float16 *)(op0), (vuint64m8_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei64_v_f16m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei64_v_f16m2_m((vbool8_t)(op0), (_Float16 *)(op1), (vuint64m8_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei64_v_f16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei64_v_f16mf2((_Float16 *)(op0), (vuint64m2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vsuxseg2ei64_v_f16mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei64_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint64m2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (size_t)(op5)) +#define vsuxseg2ei64_v_f16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsuxseg2ei64_v_f16mf4((_Float16 *)(op0), (vuint64m1_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vsuxseg2ei64_v_f16mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg2ei64_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint64m1_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (size_t)(op5)) +#define vsuxseg3ei64_v_f16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei64_v_f16m1((_Float16 *)(op0), (vuint64m4_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (size_t)(op5)) +#define vsuxseg3ei64_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei64_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint64m4_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (size_t)(op6)) +#define vsuxseg3ei64_v_f16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei64_v_f16m2((_Float16 *)(op0), (vuint64m8_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei64_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei64_v_f16m2_m((vbool8_t)(op0), (_Float16 *)(op1), (vuint64m8_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (vfloat16m2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei64_v_f16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei64_v_f16mf2((_Float16 *)(op0), (vuint64m2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (size_t)(op5)) +#define vsuxseg3ei64_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei64_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint64m2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (size_t)(op6)) +#define vsuxseg3ei64_v_f16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsuxseg3ei64_v_f16mf4((_Float16 *)(op0), (vuint64m1_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (size_t)(op5)) +#define vsuxseg3ei64_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg3ei64_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint64m1_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (size_t)(op6)) +#define vsuxseg4ei64_v_f16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei64_v_f16m1((_Float16 *)(op0), (vuint64m4_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (size_t)(op6)) +#define vsuxseg4ei64_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei64_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint64m4_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (size_t)(op7)) +#define vsuxseg4ei64_v_f16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei64_v_f16m2((_Float16 *)(op0), (vuint64m8_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (vfloat16m2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei64_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei64_v_f16m2_m((vbool8_t)(op0), (_Float16 *)(op1), (vuint64m8_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (vfloat16m2_t)(op5), (vfloat16m2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei64_v_f16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei64_v_f16mf2((_Float16 *)(op0), (vuint64m2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (size_t)(op6)) +#define vsuxseg4ei64_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei64_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint64m2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (size_t)(op7)) +#define vsuxseg4ei64_v_f16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsuxseg4ei64_v_f16mf4((_Float16 *)(op0), (vuint64m1_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (size_t)(op6)) +#define vsuxseg4ei64_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg4ei64_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint64m1_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (size_t)(op7)) +#define vsuxseg5ei64_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei64_v_f16m1((_Float16 *)(op0), (vuint64m4_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (size_t)(op7)) +#define vsuxseg5ei64_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei64_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint64m4_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (size_t)(op8)) +#define vsuxseg5ei64_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei64_v_f16mf2((_Float16 *)(op0), (vuint64m2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (size_t)(op7)) +#define vsuxseg5ei64_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei64_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint64m2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (size_t)(op8)) +#define vsuxseg5ei64_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsuxseg5ei64_v_f16mf4((_Float16 *)(op0), (vuint64m1_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (size_t)(op7)) +#define vsuxseg5ei64_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg5ei64_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint64m1_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (size_t)(op8)) +#define vsuxseg6ei64_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei64_v_f16m1((_Float16 *)(op0), (vuint64m4_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (size_t)(op8)) +#define vsuxseg6ei64_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei64_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint64m4_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (size_t)(op9)) +#define vsuxseg6ei64_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei64_v_f16mf2((_Float16 *)(op0), (vuint64m2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (size_t)(op8)) +#define vsuxseg6ei64_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei64_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint64m2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (size_t)(op9)) +#define vsuxseg6ei64_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsuxseg6ei64_v_f16mf4((_Float16 *)(op0), (vuint64m1_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (size_t)(op8)) +#define vsuxseg6ei64_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg6ei64_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint64m1_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (size_t)(op9)) +#define vsuxseg7ei64_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei64_v_f16m1((_Float16 *)(op0), (vuint64m4_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (size_t)(op9)) +#define vsuxseg7ei64_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei64_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint64m4_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (size_t)(op10)) +#define vsuxseg7ei64_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei64_v_f16mf2((_Float16 *)(op0), (vuint64m2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (size_t)(op9)) +#define vsuxseg7ei64_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei64_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint64m2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (size_t)(op10)) +#define vsuxseg7ei64_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsuxseg7ei64_v_f16mf4((_Float16 *)(op0), (vuint64m1_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (size_t)(op9)) +#define vsuxseg7ei64_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg7ei64_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint64m1_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (size_t)(op10)) +#define vsuxseg8ei64_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei64_v_f16m1((_Float16 *)(op0), (vuint64m4_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (size_t)(op10)) +#define vsuxseg8ei64_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei64_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint64m4_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (size_t)(op11)) +#define vsuxseg8ei64_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei64_v_f16mf2((_Float16 *)(op0), (vuint64m2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (size_t)(op10)) +#define vsuxseg8ei64_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei64_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint64m2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (size_t)(op11)) +#define vsuxseg8ei64_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsuxseg8ei64_v_f16mf4((_Float16 *)(op0), (vuint64m1_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (size_t)(op10)) +#define vsuxseg8ei64_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsuxseg8ei64_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint64m1_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (size_t)(op11)) +#define vsoxseg2ei8_v_f16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_f16m1((_Float16 *)(op0), (vuint8mf2_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_f16m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint8mf2_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (size_t)(op5)) +#define vsoxseg2ei8_v_f16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_f16m2((_Float16 *)(op0), (vuint8m1_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_f16m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_f16m2_m((vbool8_t)(op0), (_Float16 *)(op1), (vuint8m1_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei8_v_f16m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_f16m4((_Float16 *)(op0), (vuint8m2_t)(op1), (vfloat16m4_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_f16m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_f16m4_m((vbool4_t)(op0), (_Float16 *)(op1), (vuint8m2_t)(op2), (vfloat16m4_t)(op3), (vfloat16m4_t)(op4), (size_t)(op5)) +#define vsoxseg2ei8_v_f16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_f16mf2((_Float16 *)(op0), (vuint8mf4_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_f16mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint8mf4_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei8_v_f16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei8_v_f16mf4((_Float16 *)(op0), (vuint8mf8_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei8_v_f16mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei8_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint8mf8_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (size_t)(op5)) +#define vsoxseg3ei8_v_f16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei8_v_f16m1((_Float16 *)(op0), (vuint8mf2_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (size_t)(op5)) +#define vsoxseg3ei8_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei8_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint8mf2_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (size_t)(op6)) +#define vsoxseg3ei8_v_f16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei8_v_f16m2((_Float16 *)(op0), (vuint8m1_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei8_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei8_v_f16m2_m((vbool8_t)(op0), (_Float16 *)(op1), (vuint8m1_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (vfloat16m2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei8_v_f16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei8_v_f16mf2((_Float16 *)(op0), (vuint8mf4_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei8_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei8_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint8mf4_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei8_v_f16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei8_v_f16mf4((_Float16 *)(op0), (vuint8mf8_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (size_t)(op5)) +#define vsoxseg3ei8_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei8_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint8mf8_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (size_t)(op6)) +#define vsoxseg4ei8_v_f16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei8_v_f16m1((_Float16 *)(op0), (vuint8mf2_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (size_t)(op6)) +#define vsoxseg4ei8_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei8_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint8mf2_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (size_t)(op7)) +#define vsoxseg4ei8_v_f16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei8_v_f16m2((_Float16 *)(op0), (vuint8m1_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (vfloat16m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei8_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei8_v_f16m2_m((vbool8_t)(op0), (_Float16 *)(op1), (vuint8m1_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (vfloat16m2_t)(op5), (vfloat16m2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei8_v_f16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei8_v_f16mf2((_Float16 *)(op0), (vuint8mf4_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei8_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei8_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint8mf4_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei8_v_f16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei8_v_f16mf4((_Float16 *)(op0), (vuint8mf8_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (size_t)(op6)) +#define vsoxseg4ei8_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei8_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint8mf8_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (size_t)(op7)) +#define vsoxseg5ei8_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei8_v_f16m1((_Float16 *)(op0), (vuint8mf2_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (size_t)(op7)) +#define vsoxseg5ei8_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei8_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint8mf2_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (size_t)(op8)) +#define vsoxseg5ei8_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei8_v_f16mf2((_Float16 *)(op0), (vuint8mf4_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei8_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei8_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint8mf4_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (size_t)(op8)) +#define vsoxseg5ei8_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei8_v_f16mf4((_Float16 *)(op0), (vuint8mf8_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (size_t)(op7)) +#define vsoxseg5ei8_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei8_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint8mf8_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (size_t)(op8)) +#define vsoxseg6ei8_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei8_v_f16m1((_Float16 *)(op0), (vuint8mf2_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei8_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei8_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint8mf2_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (size_t)(op9)) +#define vsoxseg6ei8_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei8_v_f16mf2((_Float16 *)(op0), (vuint8mf4_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (size_t)(op8)) +#define vsoxseg6ei8_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei8_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint8mf4_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (size_t)(op9)) +#define vsoxseg6ei8_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei8_v_f16mf4((_Float16 *)(op0), (vuint8mf8_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (size_t)(op8)) +#define vsoxseg6ei8_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei8_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint8mf8_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (size_t)(op9)) +#define vsoxseg7ei8_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei8_v_f16m1((_Float16 *)(op0), (vuint8mf2_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei8_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei8_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint8mf2_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (size_t)(op10)) +#define vsoxseg7ei8_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei8_v_f16mf2((_Float16 *)(op0), (vuint8mf4_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (size_t)(op9)) +#define vsoxseg7ei8_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei8_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint8mf4_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (size_t)(op10)) +#define vsoxseg7ei8_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei8_v_f16mf4((_Float16 *)(op0), (vuint8mf8_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (size_t)(op9)) +#define vsoxseg7ei8_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei8_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint8mf8_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (size_t)(op10)) +#define vsoxseg8ei8_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei8_v_f16m1((_Float16 *)(op0), (vuint8mf2_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei8_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei8_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint8mf2_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (size_t)(op11)) +#define vsoxseg8ei8_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei8_v_f16mf2((_Float16 *)(op0), (vuint8mf4_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (size_t)(op10)) +#define vsoxseg8ei8_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei8_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint8mf4_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (size_t)(op11)) +#define vsoxseg8ei8_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei8_v_f16mf4((_Float16 *)(op0), (vuint8mf8_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (size_t)(op10)) +#define vsoxseg8ei8_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei8_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint8mf8_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (size_t)(op11)) +#define vsoxseg2ei16_v_f16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_f16m1((_Float16 *)(op0), (vuint16m1_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_f16m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (size_t)(op5)) +#define vsoxseg2ei16_v_f16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_f16m2((_Float16 *)(op0), (vuint16m2_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_f16m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_f16m2_m((vbool8_t)(op0), (_Float16 *)(op1), (vuint16m2_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei16_v_f16m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_f16m4((_Float16 *)(op0), (vuint16m4_t)(op1), (vfloat16m4_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_f16m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_f16m4_m((vbool4_t)(op0), (_Float16 *)(op1), (vuint16m4_t)(op2), (vfloat16m4_t)(op3), (vfloat16m4_t)(op4), (size_t)(op5)) +#define vsoxseg2ei16_v_f16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_f16mf2((_Float16 *)(op0), (vuint16mf2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_f16mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei16_v_f16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei16_v_f16mf4((_Float16 *)(op0), (vuint16mf4_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei16_v_f16mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei16_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (size_t)(op5)) +#define vsoxseg3ei16_v_f16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei16_v_f16m1((_Float16 *)(op0), (vuint16m1_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (size_t)(op5)) +#define vsoxseg3ei16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei16_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (size_t)(op6)) +#define vsoxseg3ei16_v_f16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei16_v_f16m2((_Float16 *)(op0), (vuint16m2_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei16_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei16_v_f16m2_m((vbool8_t)(op0), (_Float16 *)(op1), (vuint16m2_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (vfloat16m2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei16_v_f16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei16_v_f16mf2((_Float16 *)(op0), (vuint16mf2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei16_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei16_v_f16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei16_v_f16mf4((_Float16 *)(op0), (vuint16mf4_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (size_t)(op5)) +#define vsoxseg3ei16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei16_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (size_t)(op6)) +#define vsoxseg4ei16_v_f16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei16_v_f16m1((_Float16 *)(op0), (vuint16m1_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (size_t)(op6)) +#define vsoxseg4ei16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei16_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (size_t)(op7)) +#define vsoxseg4ei16_v_f16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei16_v_f16m2((_Float16 *)(op0), (vuint16m2_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (vfloat16m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei16_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei16_v_f16m2_m((vbool8_t)(op0), (_Float16 *)(op1), (vuint16m2_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (vfloat16m2_t)(op5), (vfloat16m2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei16_v_f16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei16_v_f16mf2((_Float16 *)(op0), (vuint16mf2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei16_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei16_v_f16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei16_v_f16mf4((_Float16 *)(op0), (vuint16mf4_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (size_t)(op6)) +#define vsoxseg4ei16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei16_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (size_t)(op7)) +#define vsoxseg5ei16_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei16_v_f16m1((_Float16 *)(op0), (vuint16m1_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (size_t)(op7)) +#define vsoxseg5ei16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei16_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (size_t)(op8)) +#define vsoxseg5ei16_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei16_v_f16mf2((_Float16 *)(op0), (vuint16mf2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei16_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (size_t)(op8)) +#define vsoxseg5ei16_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei16_v_f16mf4((_Float16 *)(op0), (vuint16mf4_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (size_t)(op7)) +#define vsoxseg5ei16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei16_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (size_t)(op8)) +#define vsoxseg6ei16_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei16_v_f16m1((_Float16 *)(op0), (vuint16m1_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei16_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (size_t)(op9)) +#define vsoxseg6ei16_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei16_v_f16mf2((_Float16 *)(op0), (vuint16mf2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (size_t)(op8)) +#define vsoxseg6ei16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei16_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (size_t)(op9)) +#define vsoxseg6ei16_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei16_v_f16mf4((_Float16 *)(op0), (vuint16mf4_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (size_t)(op8)) +#define vsoxseg6ei16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei16_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (size_t)(op9)) +#define vsoxseg7ei16_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei16_v_f16m1((_Float16 *)(op0), (vuint16m1_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei16_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (size_t)(op10)) +#define vsoxseg7ei16_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei16_v_f16mf2((_Float16 *)(op0), (vuint16mf2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (size_t)(op9)) +#define vsoxseg7ei16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei16_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (size_t)(op10)) +#define vsoxseg7ei16_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei16_v_f16mf4((_Float16 *)(op0), (vuint16mf4_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (size_t)(op9)) +#define vsoxseg7ei16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei16_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (size_t)(op10)) +#define vsoxseg8ei16_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei16_v_f16m1((_Float16 *)(op0), (vuint16m1_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei16_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (size_t)(op11)) +#define vsoxseg8ei16_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei16_v_f16mf2((_Float16 *)(op0), (vuint16mf2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (size_t)(op10)) +#define vsoxseg8ei16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei16_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (size_t)(op11)) +#define vsoxseg8ei16_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei16_v_f16mf4((_Float16 *)(op0), (vuint16mf4_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (size_t)(op10)) +#define vsoxseg8ei16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei16_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (size_t)(op11)) +#define vsoxseg2ei32_v_f16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_f16m1((_Float16 *)(op0), (vuint32m2_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_f16m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint32m2_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (size_t)(op5)) +#define vsoxseg2ei32_v_f16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_f16m2((_Float16 *)(op0), (vuint32m4_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_f16m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_f16m2_m((vbool8_t)(op0), (_Float16 *)(op1), (vuint32m4_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei32_v_f16m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_f16m4((_Float16 *)(op0), (vuint32m8_t)(op1), (vfloat16m4_t)(op2), (vfloat16m4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_f16m4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_f16m4_m((vbool4_t)(op0), (_Float16 *)(op1), (vuint32m8_t)(op2), (vfloat16m4_t)(op3), (vfloat16m4_t)(op4), (size_t)(op5)) +#define vsoxseg2ei32_v_f16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_f16mf2((_Float16 *)(op0), (vuint32m1_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_f16mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint32m1_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei32_v_f16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei32_v_f16mf4((_Float16 *)(op0), (vuint32mf2_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei32_v_f16mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei32_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint32mf2_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (size_t)(op5)) +#define vsoxseg3ei32_v_f16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei32_v_f16m1((_Float16 *)(op0), (vuint32m2_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (size_t)(op5)) +#define vsoxseg3ei32_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei32_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint32m2_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (size_t)(op6)) +#define vsoxseg3ei32_v_f16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei32_v_f16m2((_Float16 *)(op0), (vuint32m4_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei32_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei32_v_f16m2_m((vbool8_t)(op0), (_Float16 *)(op1), (vuint32m4_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (vfloat16m2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei32_v_f16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei32_v_f16mf2((_Float16 *)(op0), (vuint32m1_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei32_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei32_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint32m1_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei32_v_f16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei32_v_f16mf4((_Float16 *)(op0), (vuint32mf2_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (size_t)(op5)) +#define vsoxseg3ei32_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei32_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint32mf2_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (size_t)(op6)) +#define vsoxseg4ei32_v_f16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei32_v_f16m1((_Float16 *)(op0), (vuint32m2_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (size_t)(op6)) +#define vsoxseg4ei32_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei32_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint32m2_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (size_t)(op7)) +#define vsoxseg4ei32_v_f16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei32_v_f16m2((_Float16 *)(op0), (vuint32m4_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (vfloat16m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei32_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei32_v_f16m2_m((vbool8_t)(op0), (_Float16 *)(op1), (vuint32m4_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (vfloat16m2_t)(op5), (vfloat16m2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei32_v_f16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei32_v_f16mf2((_Float16 *)(op0), (vuint32m1_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei32_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei32_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint32m1_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei32_v_f16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei32_v_f16mf4((_Float16 *)(op0), (vuint32mf2_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (size_t)(op6)) +#define vsoxseg4ei32_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei32_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint32mf2_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (size_t)(op7)) +#define vsoxseg5ei32_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei32_v_f16m1((_Float16 *)(op0), (vuint32m2_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (size_t)(op7)) +#define vsoxseg5ei32_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei32_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint32m2_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (size_t)(op8)) +#define vsoxseg5ei32_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei32_v_f16mf2((_Float16 *)(op0), (vuint32m1_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei32_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei32_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint32m1_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (size_t)(op8)) +#define vsoxseg5ei32_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei32_v_f16mf4((_Float16 *)(op0), (vuint32mf2_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (size_t)(op7)) +#define vsoxseg5ei32_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei32_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint32mf2_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (size_t)(op8)) +#define vsoxseg6ei32_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei32_v_f16m1((_Float16 *)(op0), (vuint32m2_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei32_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei32_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint32m2_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (size_t)(op9)) +#define vsoxseg6ei32_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei32_v_f16mf2((_Float16 *)(op0), (vuint32m1_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (size_t)(op8)) +#define vsoxseg6ei32_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei32_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint32m1_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (size_t)(op9)) +#define vsoxseg6ei32_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei32_v_f16mf4((_Float16 *)(op0), (vuint32mf2_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (size_t)(op8)) +#define vsoxseg6ei32_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei32_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint32mf2_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (size_t)(op9)) +#define vsoxseg7ei32_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei32_v_f16m1((_Float16 *)(op0), (vuint32m2_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei32_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei32_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint32m2_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (size_t)(op10)) +#define vsoxseg7ei32_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei32_v_f16mf2((_Float16 *)(op0), (vuint32m1_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (size_t)(op9)) +#define vsoxseg7ei32_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei32_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint32m1_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (size_t)(op10)) +#define vsoxseg7ei32_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei32_v_f16mf4((_Float16 *)(op0), (vuint32mf2_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (size_t)(op9)) +#define vsoxseg7ei32_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei32_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint32mf2_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (size_t)(op10)) +#define vsoxseg8ei32_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei32_v_f16m1((_Float16 *)(op0), (vuint32m2_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei32_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei32_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint32m2_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (size_t)(op11)) +#define vsoxseg8ei32_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei32_v_f16mf2((_Float16 *)(op0), (vuint32m1_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (size_t)(op10)) +#define vsoxseg8ei32_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei32_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint32m1_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (size_t)(op11)) +#define vsoxseg8ei32_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei32_v_f16mf4((_Float16 *)(op0), (vuint32mf2_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (size_t)(op10)) +#define vsoxseg8ei32_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei32_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint32mf2_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (size_t)(op11)) +#define vsoxseg2ei64_v_f16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei64_v_f16m1((_Float16 *)(op0), (vuint64m4_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (size_t)(op4)) +#define vsoxseg2ei64_v_f16m1_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei64_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint64m4_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (size_t)(op5)) +#define vsoxseg2ei64_v_f16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei64_v_f16m2((_Float16 *)(op0), (vuint64m8_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei64_v_f16m2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei64_v_f16m2_m((vbool8_t)(op0), (_Float16 *)(op1), (vuint64m8_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei64_v_f16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei64_v_f16mf2((_Float16 *)(op0), (vuint64m2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (size_t)(op4)) +#define vsoxseg2ei64_v_f16mf2_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei64_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint64m2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (size_t)(op5)) +#define vsoxseg2ei64_v_f16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vsoxseg2ei64_v_f16mf4((_Float16 *)(op0), (vuint64m1_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (size_t)(op4)) +#define vsoxseg2ei64_v_f16mf4_m(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg2ei64_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint64m1_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (size_t)(op5)) +#define vsoxseg3ei64_v_f16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei64_v_f16m1((_Float16 *)(op0), (vuint64m4_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (size_t)(op5)) +#define vsoxseg3ei64_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei64_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint64m4_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (size_t)(op6)) +#define vsoxseg3ei64_v_f16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei64_v_f16m2((_Float16 *)(op0), (vuint64m8_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei64_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei64_v_f16m2_m((vbool8_t)(op0), (_Float16 *)(op1), (vuint64m8_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (vfloat16m2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei64_v_f16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei64_v_f16mf2((_Float16 *)(op0), (vuint64m2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (size_t)(op5)) +#define vsoxseg3ei64_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei64_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint64m2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (size_t)(op6)) +#define vsoxseg3ei64_v_f16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vsoxseg3ei64_v_f16mf4((_Float16 *)(op0), (vuint64m1_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (size_t)(op5)) +#define vsoxseg3ei64_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg3ei64_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint64m1_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (size_t)(op6)) +#define vsoxseg4ei64_v_f16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei64_v_f16m1((_Float16 *)(op0), (vuint64m4_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (size_t)(op6)) +#define vsoxseg4ei64_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei64_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint64m4_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (size_t)(op7)) +#define vsoxseg4ei64_v_f16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei64_v_f16m2((_Float16 *)(op0), (vuint64m8_t)(op1), (vfloat16m2_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (vfloat16m2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei64_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei64_v_f16m2_m((vbool8_t)(op0), (_Float16 *)(op1), (vuint64m8_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (vfloat16m2_t)(op5), (vfloat16m2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei64_v_f16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei64_v_f16mf2((_Float16 *)(op0), (vuint64m2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (size_t)(op6)) +#define vsoxseg4ei64_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei64_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint64m2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (size_t)(op7)) +#define vsoxseg4ei64_v_f16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vsoxseg4ei64_v_f16mf4((_Float16 *)(op0), (vuint64m1_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (size_t)(op6)) +#define vsoxseg4ei64_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg4ei64_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint64m1_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (size_t)(op7)) +#define vsoxseg5ei64_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei64_v_f16m1((_Float16 *)(op0), (vuint64m4_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (size_t)(op7)) +#define vsoxseg5ei64_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei64_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint64m4_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (size_t)(op8)) +#define vsoxseg5ei64_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei64_v_f16mf2((_Float16 *)(op0), (vuint64m2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (size_t)(op7)) +#define vsoxseg5ei64_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei64_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint64m2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (size_t)(op8)) +#define vsoxseg5ei64_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vsoxseg5ei64_v_f16mf4((_Float16 *)(op0), (vuint64m1_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (size_t)(op7)) +#define vsoxseg5ei64_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg5ei64_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint64m1_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (size_t)(op8)) +#define vsoxseg6ei64_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei64_v_f16m1((_Float16 *)(op0), (vuint64m4_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (size_t)(op8)) +#define vsoxseg6ei64_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei64_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint64m4_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (size_t)(op9)) +#define vsoxseg6ei64_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei64_v_f16mf2((_Float16 *)(op0), (vuint64m2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (size_t)(op8)) +#define vsoxseg6ei64_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei64_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint64m2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (size_t)(op9)) +#define vsoxseg6ei64_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vsoxseg6ei64_v_f16mf4((_Float16 *)(op0), (vuint64m1_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (size_t)(op8)) +#define vsoxseg6ei64_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg6ei64_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint64m1_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (size_t)(op9)) +#define vsoxseg7ei64_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei64_v_f16m1((_Float16 *)(op0), (vuint64m4_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (size_t)(op9)) +#define vsoxseg7ei64_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei64_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint64m4_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (size_t)(op10)) +#define vsoxseg7ei64_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei64_v_f16mf2((_Float16 *)(op0), (vuint64m2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (size_t)(op9)) +#define vsoxseg7ei64_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei64_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint64m2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (size_t)(op10)) +#define vsoxseg7ei64_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vsoxseg7ei64_v_f16mf4((_Float16 *)(op0), (vuint64m1_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (size_t)(op9)) +#define vsoxseg7ei64_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg7ei64_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint64m1_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (size_t)(op10)) +#define vsoxseg8ei64_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei64_v_f16m1((_Float16 *)(op0), (vuint64m4_t)(op1), (vfloat16m1_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (size_t)(op10)) +#define vsoxseg8ei64_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei64_v_f16m1_m((vbool16_t)(op0), (_Float16 *)(op1), (vuint64m4_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (size_t)(op11)) +#define vsoxseg8ei64_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei64_v_f16mf2((_Float16 *)(op0), (vuint64m2_t)(op1), (vfloat16mf2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (size_t)(op10)) +#define vsoxseg8ei64_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei64_v_f16mf2_m((vbool32_t)(op0), (_Float16 *)(op1), (vuint64m2_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (size_t)(op11)) +#define vsoxseg8ei64_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vsoxseg8ei64_v_f16mf4((_Float16 *)(op0), (vuint64m1_t)(op1), (vfloat16mf4_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (size_t)(op10)) +#define vsoxseg8ei64_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vsoxseg8ei64_v_f16mf4_m((vbool64_t)(op0), (_Float16 *)(op1), (vuint64m1_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (size_t)(op11)) +#define vlseg2e16_v_f16m1(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e16_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (const _Float16 *)(op2), (size_t)(op3)) +#define vlseg2e16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e16_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vbool16_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (const _Float16 *)(op5), (size_t)(op6)) +#define vlseg2e16_v_f16m2(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e16_v_f16m2((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (const _Float16 *)(op2), (size_t)(op3)) +#define vlseg2e16_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e16_v_f16m2_m((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vbool8_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (const _Float16 *)(op5), (size_t)(op6)) +#define vlseg2e16_v_f16m4(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e16_v_f16m4((vfloat16m4_t *)(op0), (vfloat16m4_t *)(op1), (const _Float16 *)(op2), (size_t)(op3)) +#define vlseg2e16_v_f16m4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e16_v_f16m4_m((vfloat16m4_t *)(op0), (vfloat16m4_t *)(op1), (vbool4_t)(op2), (vfloat16m4_t)(op3), (vfloat16m4_t)(op4), (const _Float16 *)(op5), (size_t)(op6)) +#define vlseg2e16_v_f16mf2(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e16_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (const _Float16 *)(op2), (size_t)(op3)) +#define vlseg2e16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e16_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vbool32_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (const _Float16 *)(op5), (size_t)(op6)) +#define vlseg2e16_v_f16mf4(op0, op1, op2, op3) \ +__builtin_rvv_vlseg2e16_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (const _Float16 *)(op2), (size_t)(op3)) +#define vlseg2e16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg2e16_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vbool64_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (const _Float16 *)(op5), (size_t)(op6)) +#define vlseg3e16_v_f16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg3e16_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (const _Float16 *)(op3), (size_t)(op4)) +#define vlseg3e16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg3e16_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vbool16_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (const _Float16 *)(op7), (size_t)(op8)) +#define vlseg3e16_v_f16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg3e16_v_f16m2((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vfloat16m2_t *)(op2), (const _Float16 *)(op3), (size_t)(op4)) +#define vlseg3e16_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg3e16_v_f16m2_m((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vfloat16m2_t *)(op2), (vbool8_t)(op3), (vfloat16m2_t)(op4), (vfloat16m2_t)(op5), (vfloat16m2_t)(op6), (const _Float16 *)(op7), (size_t)(op8)) +#define vlseg3e16_v_f16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg3e16_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (const _Float16 *)(op3), (size_t)(op4)) +#define vlseg3e16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg3e16_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vbool32_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (const _Float16 *)(op7), (size_t)(op8)) +#define vlseg3e16_v_f16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg3e16_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (const _Float16 *)(op3), (size_t)(op4)) +#define vlseg3e16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg3e16_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vbool64_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (const _Float16 *)(op7), (size_t)(op8)) +#define vlseg4e16_v_f16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg4e16_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (const _Float16 *)(op4), (size_t)(op5)) +#define vlseg4e16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg4e16_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vbool16_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (const _Float16 *)(op9), (size_t)(op10)) +#define vlseg4e16_v_f16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg4e16_v_f16m2((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vfloat16m2_t *)(op2), (vfloat16m2_t *)(op3), (const _Float16 *)(op4), (size_t)(op5)) +#define vlseg4e16_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg4e16_v_f16m2_m((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vfloat16m2_t *)(op2), (vfloat16m2_t *)(op3), (vbool8_t)(op4), (vfloat16m2_t)(op5), (vfloat16m2_t)(op6), (vfloat16m2_t)(op7), (vfloat16m2_t)(op8), (const _Float16 *)(op9), (size_t)(op10)) +#define vlseg4e16_v_f16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg4e16_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (const _Float16 *)(op4), (size_t)(op5)) +#define vlseg4e16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg4e16_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vbool32_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (const _Float16 *)(op9), (size_t)(op10)) +#define vlseg4e16_v_f16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg4e16_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (const _Float16 *)(op4), (size_t)(op5)) +#define vlseg4e16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg4e16_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vbool64_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (const _Float16 *)(op9), (size_t)(op10)) +#define vlseg5e16_v_f16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg5e16_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (const _Float16 *)(op5), (size_t)(op6)) +#define vlseg5e16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12) \ +__builtin_rvv_vlseg5e16_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vbool16_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (const _Float16 *)(op11), (size_t)(op12)) +#define vlseg5e16_v_f16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg5e16_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (const _Float16 *)(op5), (size_t)(op6)) +#define vlseg5e16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12) \ +__builtin_rvv_vlseg5e16_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vbool32_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (const _Float16 *)(op11), (size_t)(op12)) +#define vlseg5e16_v_f16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg5e16_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (const _Float16 *)(op5), (size_t)(op6)) +#define vlseg5e16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12) \ +__builtin_rvv_vlseg5e16_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vbool64_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (const _Float16 *)(op11), (size_t)(op12)) +#define vlseg6e16_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg6e16_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (const _Float16 *)(op6), (size_t)(op7)) +#define vlseg6e16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14) \ +__builtin_rvv_vlseg6e16_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vbool16_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (vfloat16m1_t)(op11), (vfloat16m1_t)(op12), (const _Float16 *)(op13), (size_t)(op14)) +#define vlseg6e16_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg6e16_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (const _Float16 *)(op6), (size_t)(op7)) +#define vlseg6e16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14) \ +__builtin_rvv_vlseg6e16_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vbool32_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (vfloat16mf2_t)(op11), (vfloat16mf2_t)(op12), (const _Float16 *)(op13), (size_t)(op14)) +#define vlseg6e16_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg6e16_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (const _Float16 *)(op6), (size_t)(op7)) +#define vlseg6e16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14) \ +__builtin_rvv_vlseg6e16_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vbool64_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (vfloat16mf4_t)(op11), (vfloat16mf4_t)(op12), (const _Float16 *)(op13), (size_t)(op14)) +#define vlseg7e16_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg7e16_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vfloat16m1_t *)(op6), (const _Float16 *)(op7), (size_t)(op8)) +#define vlseg7e16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16) \ +__builtin_rvv_vlseg7e16_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vfloat16m1_t *)(op6), (vbool16_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (vfloat16m1_t)(op11), (vfloat16m1_t)(op12), (vfloat16m1_t)(op13), (vfloat16m1_t)(op14), (const _Float16 *)(op15), (size_t)(op16)) +#define vlseg7e16_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg7e16_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vfloat16mf2_t *)(op6), (const _Float16 *)(op7), (size_t)(op8)) +#define vlseg7e16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16) \ +__builtin_rvv_vlseg7e16_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vfloat16mf2_t *)(op6), (vbool32_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (vfloat16mf2_t)(op11), (vfloat16mf2_t)(op12), (vfloat16mf2_t)(op13), (vfloat16mf2_t)(op14), (const _Float16 *)(op15), (size_t)(op16)) +#define vlseg7e16_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg7e16_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vfloat16mf4_t *)(op6), (const _Float16 *)(op7), (size_t)(op8)) +#define vlseg7e16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16) \ +__builtin_rvv_vlseg7e16_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vfloat16mf4_t *)(op6), (vbool64_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (vfloat16mf4_t)(op11), (vfloat16mf4_t)(op12), (vfloat16mf4_t)(op13), (vfloat16mf4_t)(op14), (const _Float16 *)(op15), (size_t)(op16)) +#define vlseg8e16_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg8e16_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vfloat16m1_t *)(op6), (vfloat16m1_t *)(op7), (const _Float16 *)(op8), (size_t)(op9)) +#define vlseg8e16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18) \ +__builtin_rvv_vlseg8e16_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vfloat16m1_t *)(op6), (vfloat16m1_t *)(op7), (vbool16_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (vfloat16m1_t)(op11), (vfloat16m1_t)(op12), (vfloat16m1_t)(op13), (vfloat16m1_t)(op14), (vfloat16m1_t)(op15), (vfloat16m1_t)(op16), (const _Float16 *)(op17), (size_t)(op18)) +#define vlseg8e16_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg8e16_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vfloat16mf2_t *)(op6), (vfloat16mf2_t *)(op7), (const _Float16 *)(op8), (size_t)(op9)) +#define vlseg8e16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18) \ +__builtin_rvv_vlseg8e16_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vfloat16mf2_t *)(op6), (vfloat16mf2_t *)(op7), (vbool32_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (vfloat16mf2_t)(op11), (vfloat16mf2_t)(op12), (vfloat16mf2_t)(op13), (vfloat16mf2_t)(op14), (vfloat16mf2_t)(op15), (vfloat16mf2_t)(op16), (const _Float16 *)(op17), (size_t)(op18)) +#define vlseg8e16_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg8e16_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vfloat16mf4_t *)(op6), (vfloat16mf4_t *)(op7), (const _Float16 *)(op8), (size_t)(op9)) +#define vlseg8e16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18) \ +__builtin_rvv_vlseg8e16_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vfloat16mf4_t *)(op6), (vfloat16mf4_t *)(op7), (vbool64_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (vfloat16mf4_t)(op11), (vfloat16mf4_t)(op12), (vfloat16mf4_t)(op13), (vfloat16mf4_t)(op14), (vfloat16mf4_t)(op15), (vfloat16mf4_t)(op16), (const _Float16 *)(op17), (size_t)(op18)) +#define vlseg2e16ff_v_f16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e16ff_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (const _Float16 *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e16ff_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e16ff_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vbool16_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (const _Float16 *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg2e16ff_v_f16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e16ff_v_f16m2((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (const _Float16 *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e16ff_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e16ff_v_f16m2_m((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vbool8_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (const _Float16 *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg2e16ff_v_f16m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e16ff_v_f16m4((vfloat16m4_t *)(op0), (vfloat16m4_t *)(op1), (const _Float16 *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e16ff_v_f16m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e16ff_v_f16m4_m((vfloat16m4_t *)(op0), (vfloat16m4_t *)(op1), (vbool4_t)(op2), (vfloat16m4_t)(op3), (vfloat16m4_t)(op4), (const _Float16 *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg2e16ff_v_f16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e16ff_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (const _Float16 *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e16ff_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e16ff_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vbool32_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (const _Float16 *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg2e16ff_v_f16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlseg2e16ff_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (const _Float16 *)(op2), (size_t *)(op3), (size_t)(op4)) +#define vlseg2e16ff_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg2e16ff_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vbool64_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (const _Float16 *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg3e16ff_v_f16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg3e16ff_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (const _Float16 *)(op3), (size_t *)(op4), (size_t)(op5)) +#define vlseg3e16ff_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg3e16ff_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vbool16_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (const _Float16 *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg3e16ff_v_f16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg3e16ff_v_f16m2((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vfloat16m2_t *)(op2), (const _Float16 *)(op3), (size_t *)(op4), (size_t)(op5)) +#define vlseg3e16ff_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg3e16ff_v_f16m2_m((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vfloat16m2_t *)(op2), (vbool8_t)(op3), (vfloat16m2_t)(op4), (vfloat16m2_t)(op5), (vfloat16m2_t)(op6), (const _Float16 *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg3e16ff_v_f16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg3e16ff_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (const _Float16 *)(op3), (size_t *)(op4), (size_t)(op5)) +#define vlseg3e16ff_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg3e16ff_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vbool32_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (const _Float16 *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg3e16ff_v_f16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlseg3e16ff_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (const _Float16 *)(op3), (size_t *)(op4), (size_t)(op5)) +#define vlseg3e16ff_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg3e16ff_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vbool64_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (const _Float16 *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg4e16ff_v_f16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg4e16ff_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (const _Float16 *)(op4), (size_t *)(op5), (size_t)(op6)) +#define vlseg4e16ff_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlseg4e16ff_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vbool16_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (const _Float16 *)(op9), (size_t *)(op10), (size_t)(op11)) +#define vlseg4e16ff_v_f16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg4e16ff_v_f16m2((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vfloat16m2_t *)(op2), (vfloat16m2_t *)(op3), (const _Float16 *)(op4), (size_t *)(op5), (size_t)(op6)) +#define vlseg4e16ff_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlseg4e16ff_v_f16m2_m((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vfloat16m2_t *)(op2), (vfloat16m2_t *)(op3), (vbool8_t)(op4), (vfloat16m2_t)(op5), (vfloat16m2_t)(op6), (vfloat16m2_t)(op7), (vfloat16m2_t)(op8), (const _Float16 *)(op9), (size_t *)(op10), (size_t)(op11)) +#define vlseg4e16ff_v_f16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg4e16ff_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (const _Float16 *)(op4), (size_t *)(op5), (size_t)(op6)) +#define vlseg4e16ff_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlseg4e16ff_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vbool32_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (const _Float16 *)(op9), (size_t *)(op10), (size_t)(op11)) +#define vlseg4e16ff_v_f16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlseg4e16ff_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (const _Float16 *)(op4), (size_t *)(op5), (size_t)(op6)) +#define vlseg4e16ff_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlseg4e16ff_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vbool64_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (const _Float16 *)(op9), (size_t *)(op10), (size_t)(op11)) +#define vlseg5e16ff_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg5e16ff_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (const _Float16 *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg5e16ff_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlseg5e16ff_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vbool16_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (const _Float16 *)(op11), (size_t *)(op12), (size_t)(op13)) +#define vlseg5e16ff_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg5e16ff_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (const _Float16 *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg5e16ff_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlseg5e16ff_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vbool32_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (const _Float16 *)(op11), (size_t *)(op12), (size_t)(op13)) +#define vlseg5e16ff_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlseg5e16ff_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (const _Float16 *)(op5), (size_t *)(op6), (size_t)(op7)) +#define vlseg5e16ff_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlseg5e16ff_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vbool64_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (const _Float16 *)(op11), (size_t *)(op12), (size_t)(op13)) +#define vlseg6e16ff_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg6e16ff_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (const _Float16 *)(op6), (size_t *)(op7), (size_t)(op8)) +#define vlseg6e16ff_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlseg6e16ff_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vbool16_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (vfloat16m1_t)(op11), (vfloat16m1_t)(op12), (const _Float16 *)(op13), (size_t *)(op14), (size_t)(op15)) +#define vlseg6e16ff_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg6e16ff_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (const _Float16 *)(op6), (size_t *)(op7), (size_t)(op8)) +#define vlseg6e16ff_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlseg6e16ff_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vbool32_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (vfloat16mf2_t)(op11), (vfloat16mf2_t)(op12), (const _Float16 *)(op13), (size_t *)(op14), (size_t)(op15)) +#define vlseg6e16ff_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlseg6e16ff_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (const _Float16 *)(op6), (size_t *)(op7), (size_t)(op8)) +#define vlseg6e16ff_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlseg6e16ff_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vbool64_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (vfloat16mf4_t)(op11), (vfloat16mf4_t)(op12), (const _Float16 *)(op13), (size_t *)(op14), (size_t)(op15)) +#define vlseg7e16ff_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg7e16ff_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vfloat16m1_t *)(op6), (const _Float16 *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg7e16ff_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlseg7e16ff_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vfloat16m1_t *)(op6), (vbool16_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (vfloat16m1_t)(op11), (vfloat16m1_t)(op12), (vfloat16m1_t)(op13), (vfloat16m1_t)(op14), (const _Float16 *)(op15), (size_t *)(op16), (size_t)(op17)) +#define vlseg7e16ff_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg7e16ff_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vfloat16mf2_t *)(op6), (const _Float16 *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg7e16ff_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlseg7e16ff_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vfloat16mf2_t *)(op6), (vbool32_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (vfloat16mf2_t)(op11), (vfloat16mf2_t)(op12), (vfloat16mf2_t)(op13), (vfloat16mf2_t)(op14), (const _Float16 *)(op15), (size_t *)(op16), (size_t)(op17)) +#define vlseg7e16ff_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlseg7e16ff_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vfloat16mf4_t *)(op6), (const _Float16 *)(op7), (size_t *)(op8), (size_t)(op9)) +#define vlseg7e16ff_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlseg7e16ff_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vfloat16mf4_t *)(op6), (vbool64_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (vfloat16mf4_t)(op11), (vfloat16mf4_t)(op12), (vfloat16mf4_t)(op13), (vfloat16mf4_t)(op14), (const _Float16 *)(op15), (size_t *)(op16), (size_t)(op17)) +#define vlseg8e16ff_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg8e16ff_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vfloat16m1_t *)(op6), (vfloat16m1_t *)(op7), (const _Float16 *)(op8), (size_t *)(op9), (size_t)(op10)) +#define vlseg8e16ff_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlseg8e16ff_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vfloat16m1_t *)(op6), (vfloat16m1_t *)(op7), (vbool16_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (vfloat16m1_t)(op11), (vfloat16m1_t)(op12), (vfloat16m1_t)(op13), (vfloat16m1_t)(op14), (vfloat16m1_t)(op15), (vfloat16m1_t)(op16), (const _Float16 *)(op17), (size_t *)(op18), (size_t)(op19)) +#define vlseg8e16ff_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg8e16ff_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vfloat16mf2_t *)(op6), (vfloat16mf2_t *)(op7), (const _Float16 *)(op8), (size_t *)(op9), (size_t)(op10)) +#define vlseg8e16ff_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlseg8e16ff_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vfloat16mf2_t *)(op6), (vfloat16mf2_t *)(op7), (vbool32_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (vfloat16mf2_t)(op11), (vfloat16mf2_t)(op12), (vfloat16mf2_t)(op13), (vfloat16mf2_t)(op14), (vfloat16mf2_t)(op15), (vfloat16mf2_t)(op16), (const _Float16 *)(op17), (size_t *)(op18), (size_t)(op19)) +#define vlseg8e16ff_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlseg8e16ff_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vfloat16mf4_t *)(op6), (vfloat16mf4_t *)(op7), (const _Float16 *)(op8), (size_t *)(op9), (size_t)(op10)) +#define vlseg8e16ff_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlseg8e16ff_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vfloat16mf4_t *)(op6), (vfloat16mf4_t *)(op7), (vbool64_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (vfloat16mf4_t)(op11), (vfloat16mf4_t)(op12), (vfloat16mf4_t)(op13), (vfloat16mf4_t)(op14), (vfloat16mf4_t)(op15), (vfloat16mf4_t)(op16), (const _Float16 *)(op17), (size_t *)(op18), (size_t)(op19)) +#define vlsseg2e16_v_f16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e16_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (const _Float16 *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e16_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vbool16_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (const _Float16 *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg2e16_v_f16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e16_v_f16m2((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (const _Float16 *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e16_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e16_v_f16m2_m((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vbool8_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (const _Float16 *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg2e16_v_f16m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e16_v_f16m4((vfloat16m4_t *)(op0), (vfloat16m4_t *)(op1), (const _Float16 *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e16_v_f16m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e16_v_f16m4_m((vfloat16m4_t *)(op0), (vfloat16m4_t *)(op1), (vbool4_t)(op2), (vfloat16m4_t)(op3), (vfloat16m4_t)(op4), (const _Float16 *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg2e16_v_f16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e16_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (const _Float16 *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e16_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vbool32_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (const _Float16 *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg2e16_v_f16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vlsseg2e16_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (const _Float16 *)(op2), (ptrdiff_t)(op3), (size_t)(op4)) +#define vlsseg2e16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg2e16_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vbool64_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (const _Float16 *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg3e16_v_f16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlsseg3e16_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (const _Float16 *)(op3), (ptrdiff_t)(op4), (size_t)(op5)) +#define vlsseg3e16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg3e16_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vbool16_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (const _Float16 *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg3e16_v_f16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlsseg3e16_v_f16m2((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vfloat16m2_t *)(op2), (const _Float16 *)(op3), (ptrdiff_t)(op4), (size_t)(op5)) +#define vlsseg3e16_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg3e16_v_f16m2_m((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vfloat16m2_t *)(op2), (vbool8_t)(op3), (vfloat16m2_t)(op4), (vfloat16m2_t)(op5), (vfloat16m2_t)(op6), (const _Float16 *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg3e16_v_f16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlsseg3e16_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (const _Float16 *)(op3), (ptrdiff_t)(op4), (size_t)(op5)) +#define vlsseg3e16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg3e16_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vbool32_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (const _Float16 *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg3e16_v_f16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vlsseg3e16_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (const _Float16 *)(op3), (ptrdiff_t)(op4), (size_t)(op5)) +#define vlsseg3e16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg3e16_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vbool64_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (const _Float16 *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg4e16_v_f16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlsseg4e16_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (const _Float16 *)(op4), (ptrdiff_t)(op5), (size_t)(op6)) +#define vlsseg4e16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlsseg4e16_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vbool16_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (const _Float16 *)(op9), (ptrdiff_t)(op10), (size_t)(op11)) +#define vlsseg4e16_v_f16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlsseg4e16_v_f16m2((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vfloat16m2_t *)(op2), (vfloat16m2_t *)(op3), (const _Float16 *)(op4), (ptrdiff_t)(op5), (size_t)(op6)) +#define vlsseg4e16_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlsseg4e16_v_f16m2_m((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vfloat16m2_t *)(op2), (vfloat16m2_t *)(op3), (vbool8_t)(op4), (vfloat16m2_t)(op5), (vfloat16m2_t)(op6), (vfloat16m2_t)(op7), (vfloat16m2_t)(op8), (const _Float16 *)(op9), (ptrdiff_t)(op10), (size_t)(op11)) +#define vlsseg4e16_v_f16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlsseg4e16_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (const _Float16 *)(op4), (ptrdiff_t)(op5), (size_t)(op6)) +#define vlsseg4e16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlsseg4e16_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vbool32_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (const _Float16 *)(op9), (ptrdiff_t)(op10), (size_t)(op11)) +#define vlsseg4e16_v_f16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vlsseg4e16_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (const _Float16 *)(op4), (ptrdiff_t)(op5), (size_t)(op6)) +#define vlsseg4e16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vlsseg4e16_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vbool64_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (const _Float16 *)(op9), (ptrdiff_t)(op10), (size_t)(op11)) +#define vlsseg5e16_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg5e16_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (const _Float16 *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg5e16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlsseg5e16_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vbool16_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (const _Float16 *)(op11), (ptrdiff_t)(op12), (size_t)(op13)) +#define vlsseg5e16_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg5e16_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (const _Float16 *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg5e16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlsseg5e16_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vbool32_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (const _Float16 *)(op11), (ptrdiff_t)(op12), (size_t)(op13)) +#define vlsseg5e16_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vlsseg5e16_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (const _Float16 *)(op5), (ptrdiff_t)(op6), (size_t)(op7)) +#define vlsseg5e16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vlsseg5e16_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vbool64_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (const _Float16 *)(op11), (ptrdiff_t)(op12), (size_t)(op13)) +#define vlsseg6e16_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlsseg6e16_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (const _Float16 *)(op6), (ptrdiff_t)(op7), (size_t)(op8)) +#define vlsseg6e16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlsseg6e16_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vbool16_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (vfloat16m1_t)(op11), (vfloat16m1_t)(op12), (const _Float16 *)(op13), (ptrdiff_t)(op14), (size_t)(op15)) +#define vlsseg6e16_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlsseg6e16_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (const _Float16 *)(op6), (ptrdiff_t)(op7), (size_t)(op8)) +#define vlsseg6e16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlsseg6e16_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vbool32_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (vfloat16mf2_t)(op11), (vfloat16mf2_t)(op12), (const _Float16 *)(op13), (ptrdiff_t)(op14), (size_t)(op15)) +#define vlsseg6e16_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vlsseg6e16_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (const _Float16 *)(op6), (ptrdiff_t)(op7), (size_t)(op8)) +#define vlsseg6e16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vlsseg6e16_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vbool64_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (vfloat16mf4_t)(op11), (vfloat16mf4_t)(op12), (const _Float16 *)(op13), (ptrdiff_t)(op14), (size_t)(op15)) +#define vlsseg7e16_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg7e16_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vfloat16m1_t *)(op6), (const _Float16 *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg7e16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlsseg7e16_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vfloat16m1_t *)(op6), (vbool16_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (vfloat16m1_t)(op11), (vfloat16m1_t)(op12), (vfloat16m1_t)(op13), (vfloat16m1_t)(op14), (const _Float16 *)(op15), (ptrdiff_t)(op16), (size_t)(op17)) +#define vlsseg7e16_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg7e16_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vfloat16mf2_t *)(op6), (const _Float16 *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg7e16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlsseg7e16_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vfloat16mf2_t *)(op6), (vbool32_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (vfloat16mf2_t)(op11), (vfloat16mf2_t)(op12), (vfloat16mf2_t)(op13), (vfloat16mf2_t)(op14), (const _Float16 *)(op15), (ptrdiff_t)(op16), (size_t)(op17)) +#define vlsseg7e16_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vlsseg7e16_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vfloat16mf4_t *)(op6), (const _Float16 *)(op7), (ptrdiff_t)(op8), (size_t)(op9)) +#define vlsseg7e16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vlsseg7e16_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vfloat16mf4_t *)(op6), (vbool64_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (vfloat16mf4_t)(op11), (vfloat16mf4_t)(op12), (vfloat16mf4_t)(op13), (vfloat16mf4_t)(op14), (const _Float16 *)(op15), (ptrdiff_t)(op16), (size_t)(op17)) +#define vlsseg8e16_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlsseg8e16_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vfloat16m1_t *)(op6), (vfloat16m1_t *)(op7), (const _Float16 *)(op8), (ptrdiff_t)(op9), (size_t)(op10)) +#define vlsseg8e16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlsseg8e16_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vfloat16m1_t *)(op6), (vfloat16m1_t *)(op7), (vbool16_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (vfloat16m1_t)(op11), (vfloat16m1_t)(op12), (vfloat16m1_t)(op13), (vfloat16m1_t)(op14), (vfloat16m1_t)(op15), (vfloat16m1_t)(op16), (const _Float16 *)(op17), (ptrdiff_t)(op18), (size_t)(op19)) +#define vlsseg8e16_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlsseg8e16_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vfloat16mf2_t *)(op6), (vfloat16mf2_t *)(op7), (const _Float16 *)(op8), (ptrdiff_t)(op9), (size_t)(op10)) +#define vlsseg8e16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlsseg8e16_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vfloat16mf2_t *)(op6), (vfloat16mf2_t *)(op7), (vbool32_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (vfloat16mf2_t)(op11), (vfloat16mf2_t)(op12), (vfloat16mf2_t)(op13), (vfloat16mf2_t)(op14), (vfloat16mf2_t)(op15), (vfloat16mf2_t)(op16), (const _Float16 *)(op17), (ptrdiff_t)(op18), (size_t)(op19)) +#define vlsseg8e16_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vlsseg8e16_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vfloat16mf4_t *)(op6), (vfloat16mf4_t *)(op7), (const _Float16 *)(op8), (ptrdiff_t)(op9), (size_t)(op10)) +#define vlsseg8e16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vlsseg8e16_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vfloat16mf4_t *)(op6), (vfloat16mf4_t *)(op7), (vbool64_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (vfloat16mf4_t)(op11), (vfloat16mf4_t)(op12), (vfloat16mf4_t)(op13), (vfloat16mf4_t)(op14), (vfloat16mf4_t)(op15), (vfloat16mf4_t)(op16), (const _Float16 *)(op17), (ptrdiff_t)(op18), (size_t)(op19)) +#define vluxseg2ei8_v_f16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (const _Float16 *)(op2), (vuint8mf2_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vbool16_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (const _Float16 *)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vluxseg2ei8_v_f16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_f16m2((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (const _Float16 *)(op2), (vuint8m1_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_f16m2_m((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vbool8_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (const _Float16 *)(op5), (vuint8m1_t)(op6), (size_t)(op7)) +#define vluxseg2ei8_v_f16m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_f16m4((vfloat16m4_t *)(op0), (vfloat16m4_t *)(op1), (const _Float16 *)(op2), (vuint8m2_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_f16m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_f16m4_m((vfloat16m4_t *)(op0), (vfloat16m4_t *)(op1), (vbool4_t)(op2), (vfloat16m4_t)(op3), (vfloat16m4_t)(op4), (const _Float16 *)(op5), (vuint8m2_t)(op6), (size_t)(op7)) +#define vluxseg2ei8_v_f16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (const _Float16 *)(op2), (vuint8mf4_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vbool32_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (const _Float16 *)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vluxseg2ei8_v_f16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei8_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (const _Float16 *)(op2), (vuint8mf8_t)(op3), (size_t)(op4)) +#define vluxseg2ei8_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei8_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vbool64_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (const _Float16 *)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vluxseg3ei8_v_f16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei8_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (const _Float16 *)(op3), (vuint8mf2_t)(op4), (size_t)(op5)) +#define vluxseg3ei8_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei8_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vbool16_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (const _Float16 *)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vluxseg3ei8_v_f16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei8_v_f16m2((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vfloat16m2_t *)(op2), (const _Float16 *)(op3), (vuint8m1_t)(op4), (size_t)(op5)) +#define vluxseg3ei8_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei8_v_f16m2_m((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vfloat16m2_t *)(op2), (vbool8_t)(op3), (vfloat16m2_t)(op4), (vfloat16m2_t)(op5), (vfloat16m2_t)(op6), (const _Float16 *)(op7), (vuint8m1_t)(op8), (size_t)(op9)) +#define vluxseg3ei8_v_f16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei8_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (const _Float16 *)(op3), (vuint8mf4_t)(op4), (size_t)(op5)) +#define vluxseg3ei8_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei8_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vbool32_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (const _Float16 *)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vluxseg3ei8_v_f16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei8_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (const _Float16 *)(op3), (vuint8mf8_t)(op4), (size_t)(op5)) +#define vluxseg3ei8_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei8_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vbool64_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (const _Float16 *)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vluxseg4ei8_v_f16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei8_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (const _Float16 *)(op4), (vuint8mf2_t)(op5), (size_t)(op6)) +#define vluxseg4ei8_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei8_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vbool16_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (const _Float16 *)(op9), (vuint8mf2_t)(op10), (size_t)(op11)) +#define vluxseg4ei8_v_f16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei8_v_f16m2((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vfloat16m2_t *)(op2), (vfloat16m2_t *)(op3), (const _Float16 *)(op4), (vuint8m1_t)(op5), (size_t)(op6)) +#define vluxseg4ei8_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei8_v_f16m2_m((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vfloat16m2_t *)(op2), (vfloat16m2_t *)(op3), (vbool8_t)(op4), (vfloat16m2_t)(op5), (vfloat16m2_t)(op6), (vfloat16m2_t)(op7), (vfloat16m2_t)(op8), (const _Float16 *)(op9), (vuint8m1_t)(op10), (size_t)(op11)) +#define vluxseg4ei8_v_f16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei8_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (const _Float16 *)(op4), (vuint8mf4_t)(op5), (size_t)(op6)) +#define vluxseg4ei8_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei8_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vbool32_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (const _Float16 *)(op9), (vuint8mf4_t)(op10), (size_t)(op11)) +#define vluxseg4ei8_v_f16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei8_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (const _Float16 *)(op4), (vuint8mf8_t)(op5), (size_t)(op6)) +#define vluxseg4ei8_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei8_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vbool64_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (const _Float16 *)(op9), (vuint8mf8_t)(op10), (size_t)(op11)) +#define vluxseg5ei8_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei8_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (const _Float16 *)(op5), (vuint8mf2_t)(op6), (size_t)(op7)) +#define vluxseg5ei8_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei8_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vbool16_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (const _Float16 *)(op11), (vuint8mf2_t)(op12), (size_t)(op13)) +#define vluxseg5ei8_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei8_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (const _Float16 *)(op5), (vuint8mf4_t)(op6), (size_t)(op7)) +#define vluxseg5ei8_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei8_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vbool32_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (const _Float16 *)(op11), (vuint8mf4_t)(op12), (size_t)(op13)) +#define vluxseg5ei8_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei8_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (const _Float16 *)(op5), (vuint8mf8_t)(op6), (size_t)(op7)) +#define vluxseg5ei8_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei8_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vbool64_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (const _Float16 *)(op11), (vuint8mf8_t)(op12), (size_t)(op13)) +#define vluxseg6ei8_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei8_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (const _Float16 *)(op6), (vuint8mf2_t)(op7), (size_t)(op8)) +#define vluxseg6ei8_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei8_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vbool16_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (vfloat16m1_t)(op11), (vfloat16m1_t)(op12), (const _Float16 *)(op13), (vuint8mf2_t)(op14), (size_t)(op15)) +#define vluxseg6ei8_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei8_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (const _Float16 *)(op6), (vuint8mf4_t)(op7), (size_t)(op8)) +#define vluxseg6ei8_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei8_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vbool32_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (vfloat16mf2_t)(op11), (vfloat16mf2_t)(op12), (const _Float16 *)(op13), (vuint8mf4_t)(op14), (size_t)(op15)) +#define vluxseg6ei8_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei8_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (const _Float16 *)(op6), (vuint8mf8_t)(op7), (size_t)(op8)) +#define vluxseg6ei8_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei8_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vbool64_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (vfloat16mf4_t)(op11), (vfloat16mf4_t)(op12), (const _Float16 *)(op13), (vuint8mf8_t)(op14), (size_t)(op15)) +#define vluxseg7ei8_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei8_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vfloat16m1_t *)(op6), (const _Float16 *)(op7), (vuint8mf2_t)(op8), (size_t)(op9)) +#define vluxseg7ei8_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei8_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vfloat16m1_t *)(op6), (vbool16_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (vfloat16m1_t)(op11), (vfloat16m1_t)(op12), (vfloat16m1_t)(op13), (vfloat16m1_t)(op14), (const _Float16 *)(op15), (vuint8mf2_t)(op16), (size_t)(op17)) +#define vluxseg7ei8_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei8_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vfloat16mf2_t *)(op6), (const _Float16 *)(op7), (vuint8mf4_t)(op8), (size_t)(op9)) +#define vluxseg7ei8_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei8_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vfloat16mf2_t *)(op6), (vbool32_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (vfloat16mf2_t)(op11), (vfloat16mf2_t)(op12), (vfloat16mf2_t)(op13), (vfloat16mf2_t)(op14), (const _Float16 *)(op15), (vuint8mf4_t)(op16), (size_t)(op17)) +#define vluxseg7ei8_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei8_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vfloat16mf4_t *)(op6), (const _Float16 *)(op7), (vuint8mf8_t)(op8), (size_t)(op9)) +#define vluxseg7ei8_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei8_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vfloat16mf4_t *)(op6), (vbool64_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (vfloat16mf4_t)(op11), (vfloat16mf4_t)(op12), (vfloat16mf4_t)(op13), (vfloat16mf4_t)(op14), (const _Float16 *)(op15), (vuint8mf8_t)(op16), (size_t)(op17)) +#define vluxseg8ei8_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei8_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vfloat16m1_t *)(op6), (vfloat16m1_t *)(op7), (const _Float16 *)(op8), (vuint8mf2_t)(op9), (size_t)(op10)) +#define vluxseg8ei8_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei8_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vfloat16m1_t *)(op6), (vfloat16m1_t *)(op7), (vbool16_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (vfloat16m1_t)(op11), (vfloat16m1_t)(op12), (vfloat16m1_t)(op13), (vfloat16m1_t)(op14), (vfloat16m1_t)(op15), (vfloat16m1_t)(op16), (const _Float16 *)(op17), (vuint8mf2_t)(op18), (size_t)(op19)) +#define vluxseg8ei8_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei8_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vfloat16mf2_t *)(op6), (vfloat16mf2_t *)(op7), (const _Float16 *)(op8), (vuint8mf4_t)(op9), (size_t)(op10)) +#define vluxseg8ei8_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei8_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vfloat16mf2_t *)(op6), (vfloat16mf2_t *)(op7), (vbool32_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (vfloat16mf2_t)(op11), (vfloat16mf2_t)(op12), (vfloat16mf2_t)(op13), (vfloat16mf2_t)(op14), (vfloat16mf2_t)(op15), (vfloat16mf2_t)(op16), (const _Float16 *)(op17), (vuint8mf4_t)(op18), (size_t)(op19)) +#define vluxseg8ei8_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei8_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vfloat16mf4_t *)(op6), (vfloat16mf4_t *)(op7), (const _Float16 *)(op8), (vuint8mf8_t)(op9), (size_t)(op10)) +#define vluxseg8ei8_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei8_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vfloat16mf4_t *)(op6), (vfloat16mf4_t *)(op7), (vbool64_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (vfloat16mf4_t)(op11), (vfloat16mf4_t)(op12), (vfloat16mf4_t)(op13), (vfloat16mf4_t)(op14), (vfloat16mf4_t)(op15), (vfloat16mf4_t)(op16), (const _Float16 *)(op17), (vuint8mf8_t)(op18), (size_t)(op19)) +#define vluxseg2ei16_v_f16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (const _Float16 *)(op2), (vuint16m1_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vbool16_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (const _Float16 *)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vluxseg2ei16_v_f16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_f16m2((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (const _Float16 *)(op2), (vuint16m2_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_f16m2_m((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vbool8_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (const _Float16 *)(op5), (vuint16m2_t)(op6), (size_t)(op7)) +#define vluxseg2ei16_v_f16m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_f16m4((vfloat16m4_t *)(op0), (vfloat16m4_t *)(op1), (const _Float16 *)(op2), (vuint16m4_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_f16m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_f16m4_m((vfloat16m4_t *)(op0), (vfloat16m4_t *)(op1), (vbool4_t)(op2), (vfloat16m4_t)(op3), (vfloat16m4_t)(op4), (const _Float16 *)(op5), (vuint16m4_t)(op6), (size_t)(op7)) +#define vluxseg2ei16_v_f16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (const _Float16 *)(op2), (vuint16mf2_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vbool32_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (const _Float16 *)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vluxseg2ei16_v_f16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei16_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (const _Float16 *)(op2), (vuint16mf4_t)(op3), (size_t)(op4)) +#define vluxseg2ei16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei16_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vbool64_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (const _Float16 *)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vluxseg3ei16_v_f16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei16_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (const _Float16 *)(op3), (vuint16m1_t)(op4), (size_t)(op5)) +#define vluxseg3ei16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei16_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vbool16_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (const _Float16 *)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vluxseg3ei16_v_f16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei16_v_f16m2((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vfloat16m2_t *)(op2), (const _Float16 *)(op3), (vuint16m2_t)(op4), (size_t)(op5)) +#define vluxseg3ei16_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei16_v_f16m2_m((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vfloat16m2_t *)(op2), (vbool8_t)(op3), (vfloat16m2_t)(op4), (vfloat16m2_t)(op5), (vfloat16m2_t)(op6), (const _Float16 *)(op7), (vuint16m2_t)(op8), (size_t)(op9)) +#define vluxseg3ei16_v_f16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei16_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (const _Float16 *)(op3), (vuint16mf2_t)(op4), (size_t)(op5)) +#define vluxseg3ei16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei16_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vbool32_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (const _Float16 *)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vluxseg3ei16_v_f16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei16_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (const _Float16 *)(op3), (vuint16mf4_t)(op4), (size_t)(op5)) +#define vluxseg3ei16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei16_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vbool64_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (const _Float16 *)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vluxseg4ei16_v_f16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei16_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (const _Float16 *)(op4), (vuint16m1_t)(op5), (size_t)(op6)) +#define vluxseg4ei16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei16_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vbool16_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (const _Float16 *)(op9), (vuint16m1_t)(op10), (size_t)(op11)) +#define vluxseg4ei16_v_f16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei16_v_f16m2((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vfloat16m2_t *)(op2), (vfloat16m2_t *)(op3), (const _Float16 *)(op4), (vuint16m2_t)(op5), (size_t)(op6)) +#define vluxseg4ei16_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei16_v_f16m2_m((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vfloat16m2_t *)(op2), (vfloat16m2_t *)(op3), (vbool8_t)(op4), (vfloat16m2_t)(op5), (vfloat16m2_t)(op6), (vfloat16m2_t)(op7), (vfloat16m2_t)(op8), (const _Float16 *)(op9), (vuint16m2_t)(op10), (size_t)(op11)) +#define vluxseg4ei16_v_f16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei16_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (const _Float16 *)(op4), (vuint16mf2_t)(op5), (size_t)(op6)) +#define vluxseg4ei16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei16_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vbool32_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (const _Float16 *)(op9), (vuint16mf2_t)(op10), (size_t)(op11)) +#define vluxseg4ei16_v_f16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei16_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (const _Float16 *)(op4), (vuint16mf4_t)(op5), (size_t)(op6)) +#define vluxseg4ei16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei16_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vbool64_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (const _Float16 *)(op9), (vuint16mf4_t)(op10), (size_t)(op11)) +#define vluxseg5ei16_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei16_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (const _Float16 *)(op5), (vuint16m1_t)(op6), (size_t)(op7)) +#define vluxseg5ei16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei16_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vbool16_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (const _Float16 *)(op11), (vuint16m1_t)(op12), (size_t)(op13)) +#define vluxseg5ei16_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei16_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (const _Float16 *)(op5), (vuint16mf2_t)(op6), (size_t)(op7)) +#define vluxseg5ei16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei16_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vbool32_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (const _Float16 *)(op11), (vuint16mf2_t)(op12), (size_t)(op13)) +#define vluxseg5ei16_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei16_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (const _Float16 *)(op5), (vuint16mf4_t)(op6), (size_t)(op7)) +#define vluxseg5ei16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei16_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vbool64_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (const _Float16 *)(op11), (vuint16mf4_t)(op12), (size_t)(op13)) +#define vluxseg6ei16_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei16_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (const _Float16 *)(op6), (vuint16m1_t)(op7), (size_t)(op8)) +#define vluxseg6ei16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei16_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vbool16_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (vfloat16m1_t)(op11), (vfloat16m1_t)(op12), (const _Float16 *)(op13), (vuint16m1_t)(op14), (size_t)(op15)) +#define vluxseg6ei16_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei16_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (const _Float16 *)(op6), (vuint16mf2_t)(op7), (size_t)(op8)) +#define vluxseg6ei16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei16_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vbool32_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (vfloat16mf2_t)(op11), (vfloat16mf2_t)(op12), (const _Float16 *)(op13), (vuint16mf2_t)(op14), (size_t)(op15)) +#define vluxseg6ei16_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei16_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (const _Float16 *)(op6), (vuint16mf4_t)(op7), (size_t)(op8)) +#define vluxseg6ei16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei16_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vbool64_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (vfloat16mf4_t)(op11), (vfloat16mf4_t)(op12), (const _Float16 *)(op13), (vuint16mf4_t)(op14), (size_t)(op15)) +#define vluxseg7ei16_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei16_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vfloat16m1_t *)(op6), (const _Float16 *)(op7), (vuint16m1_t)(op8), (size_t)(op9)) +#define vluxseg7ei16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei16_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vfloat16m1_t *)(op6), (vbool16_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (vfloat16m1_t)(op11), (vfloat16m1_t)(op12), (vfloat16m1_t)(op13), (vfloat16m1_t)(op14), (const _Float16 *)(op15), (vuint16m1_t)(op16), (size_t)(op17)) +#define vluxseg7ei16_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei16_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vfloat16mf2_t *)(op6), (const _Float16 *)(op7), (vuint16mf2_t)(op8), (size_t)(op9)) +#define vluxseg7ei16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei16_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vfloat16mf2_t *)(op6), (vbool32_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (vfloat16mf2_t)(op11), (vfloat16mf2_t)(op12), (vfloat16mf2_t)(op13), (vfloat16mf2_t)(op14), (const _Float16 *)(op15), (vuint16mf2_t)(op16), (size_t)(op17)) +#define vluxseg7ei16_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei16_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vfloat16mf4_t *)(op6), (const _Float16 *)(op7), (vuint16mf4_t)(op8), (size_t)(op9)) +#define vluxseg7ei16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei16_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vfloat16mf4_t *)(op6), (vbool64_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (vfloat16mf4_t)(op11), (vfloat16mf4_t)(op12), (vfloat16mf4_t)(op13), (vfloat16mf4_t)(op14), (const _Float16 *)(op15), (vuint16mf4_t)(op16), (size_t)(op17)) +#define vluxseg8ei16_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei16_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vfloat16m1_t *)(op6), (vfloat16m1_t *)(op7), (const _Float16 *)(op8), (vuint16m1_t)(op9), (size_t)(op10)) +#define vluxseg8ei16_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei16_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vfloat16m1_t *)(op6), (vfloat16m1_t *)(op7), (vbool16_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (vfloat16m1_t)(op11), (vfloat16m1_t)(op12), (vfloat16m1_t)(op13), (vfloat16m1_t)(op14), (vfloat16m1_t)(op15), (vfloat16m1_t)(op16), (const _Float16 *)(op17), (vuint16m1_t)(op18), (size_t)(op19)) +#define vluxseg8ei16_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei16_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vfloat16mf2_t *)(op6), (vfloat16mf2_t *)(op7), (const _Float16 *)(op8), (vuint16mf2_t)(op9), (size_t)(op10)) +#define vluxseg8ei16_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei16_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vfloat16mf2_t *)(op6), (vfloat16mf2_t *)(op7), (vbool32_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (vfloat16mf2_t)(op11), (vfloat16mf2_t)(op12), (vfloat16mf2_t)(op13), (vfloat16mf2_t)(op14), (vfloat16mf2_t)(op15), (vfloat16mf2_t)(op16), (const _Float16 *)(op17), (vuint16mf2_t)(op18), (size_t)(op19)) +#define vluxseg8ei16_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei16_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vfloat16mf4_t *)(op6), (vfloat16mf4_t *)(op7), (const _Float16 *)(op8), (vuint16mf4_t)(op9), (size_t)(op10)) +#define vluxseg8ei16_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei16_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vfloat16mf4_t *)(op6), (vfloat16mf4_t *)(op7), (vbool64_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (vfloat16mf4_t)(op11), (vfloat16mf4_t)(op12), (vfloat16mf4_t)(op13), (vfloat16mf4_t)(op14), (vfloat16mf4_t)(op15), (vfloat16mf4_t)(op16), (const _Float16 *)(op17), (vuint16mf4_t)(op18), (size_t)(op19)) +#define vluxseg2ei32_v_f16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (const _Float16 *)(op2), (vuint32m2_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vbool16_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (const _Float16 *)(op5), (vuint32m2_t)(op6), (size_t)(op7)) +#define vluxseg2ei32_v_f16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_f16m2((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (const _Float16 *)(op2), (vuint32m4_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_f16m2_m((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vbool8_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (const _Float16 *)(op5), (vuint32m4_t)(op6), (size_t)(op7)) +#define vluxseg2ei32_v_f16m4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_f16m4((vfloat16m4_t *)(op0), (vfloat16m4_t *)(op1), (const _Float16 *)(op2), (vuint32m8_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_f16m4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_f16m4_m((vfloat16m4_t *)(op0), (vfloat16m4_t *)(op1), (vbool4_t)(op2), (vfloat16m4_t)(op3), (vfloat16m4_t)(op4), (const _Float16 *)(op5), (vuint32m8_t)(op6), (size_t)(op7)) +#define vluxseg2ei32_v_f16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (const _Float16 *)(op2), (vuint32m1_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vbool32_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (const _Float16 *)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vluxseg2ei32_v_f16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei32_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (const _Float16 *)(op2), (vuint32mf2_t)(op3), (size_t)(op4)) +#define vluxseg2ei32_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei32_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vbool64_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (const _Float16 *)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vluxseg3ei32_v_f16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei32_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (const _Float16 *)(op3), (vuint32m2_t)(op4), (size_t)(op5)) +#define vluxseg3ei32_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei32_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vbool16_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (const _Float16 *)(op7), (vuint32m2_t)(op8), (size_t)(op9)) +#define vluxseg3ei32_v_f16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei32_v_f16m2((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vfloat16m2_t *)(op2), (const _Float16 *)(op3), (vuint32m4_t)(op4), (size_t)(op5)) +#define vluxseg3ei32_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei32_v_f16m2_m((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vfloat16m2_t *)(op2), (vbool8_t)(op3), (vfloat16m2_t)(op4), (vfloat16m2_t)(op5), (vfloat16m2_t)(op6), (const _Float16 *)(op7), (vuint32m4_t)(op8), (size_t)(op9)) +#define vluxseg3ei32_v_f16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei32_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (const _Float16 *)(op3), (vuint32m1_t)(op4), (size_t)(op5)) +#define vluxseg3ei32_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei32_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vbool32_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (const _Float16 *)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vluxseg3ei32_v_f16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei32_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (const _Float16 *)(op3), (vuint32mf2_t)(op4), (size_t)(op5)) +#define vluxseg3ei32_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei32_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vbool64_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (const _Float16 *)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vluxseg4ei32_v_f16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei32_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (const _Float16 *)(op4), (vuint32m2_t)(op5), (size_t)(op6)) +#define vluxseg4ei32_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei32_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vbool16_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (const _Float16 *)(op9), (vuint32m2_t)(op10), (size_t)(op11)) +#define vluxseg4ei32_v_f16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei32_v_f16m2((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vfloat16m2_t *)(op2), (vfloat16m2_t *)(op3), (const _Float16 *)(op4), (vuint32m4_t)(op5), (size_t)(op6)) +#define vluxseg4ei32_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei32_v_f16m2_m((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vfloat16m2_t *)(op2), (vfloat16m2_t *)(op3), (vbool8_t)(op4), (vfloat16m2_t)(op5), (vfloat16m2_t)(op6), (vfloat16m2_t)(op7), (vfloat16m2_t)(op8), (const _Float16 *)(op9), (vuint32m4_t)(op10), (size_t)(op11)) +#define vluxseg4ei32_v_f16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei32_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (const _Float16 *)(op4), (vuint32m1_t)(op5), (size_t)(op6)) +#define vluxseg4ei32_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei32_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vbool32_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (const _Float16 *)(op9), (vuint32m1_t)(op10), (size_t)(op11)) +#define vluxseg4ei32_v_f16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei32_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (const _Float16 *)(op4), (vuint32mf2_t)(op5), (size_t)(op6)) +#define vluxseg4ei32_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei32_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vbool64_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (const _Float16 *)(op9), (vuint32mf2_t)(op10), (size_t)(op11)) +#define vluxseg5ei32_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei32_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (const _Float16 *)(op5), (vuint32m2_t)(op6), (size_t)(op7)) +#define vluxseg5ei32_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei32_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vbool16_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (const _Float16 *)(op11), (vuint32m2_t)(op12), (size_t)(op13)) +#define vluxseg5ei32_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei32_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (const _Float16 *)(op5), (vuint32m1_t)(op6), (size_t)(op7)) +#define vluxseg5ei32_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei32_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vbool32_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (const _Float16 *)(op11), (vuint32m1_t)(op12), (size_t)(op13)) +#define vluxseg5ei32_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei32_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (const _Float16 *)(op5), (vuint32mf2_t)(op6), (size_t)(op7)) +#define vluxseg5ei32_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei32_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vbool64_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (const _Float16 *)(op11), (vuint32mf2_t)(op12), (size_t)(op13)) +#define vluxseg6ei32_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei32_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (const _Float16 *)(op6), (vuint32m2_t)(op7), (size_t)(op8)) +#define vluxseg6ei32_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei32_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vbool16_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (vfloat16m1_t)(op11), (vfloat16m1_t)(op12), (const _Float16 *)(op13), (vuint32m2_t)(op14), (size_t)(op15)) +#define vluxseg6ei32_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei32_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (const _Float16 *)(op6), (vuint32m1_t)(op7), (size_t)(op8)) +#define vluxseg6ei32_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei32_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vbool32_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (vfloat16mf2_t)(op11), (vfloat16mf2_t)(op12), (const _Float16 *)(op13), (vuint32m1_t)(op14), (size_t)(op15)) +#define vluxseg6ei32_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei32_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (const _Float16 *)(op6), (vuint32mf2_t)(op7), (size_t)(op8)) +#define vluxseg6ei32_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei32_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vbool64_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (vfloat16mf4_t)(op11), (vfloat16mf4_t)(op12), (const _Float16 *)(op13), (vuint32mf2_t)(op14), (size_t)(op15)) +#define vluxseg7ei32_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei32_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vfloat16m1_t *)(op6), (const _Float16 *)(op7), (vuint32m2_t)(op8), (size_t)(op9)) +#define vluxseg7ei32_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei32_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vfloat16m1_t *)(op6), (vbool16_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (vfloat16m1_t)(op11), (vfloat16m1_t)(op12), (vfloat16m1_t)(op13), (vfloat16m1_t)(op14), (const _Float16 *)(op15), (vuint32m2_t)(op16), (size_t)(op17)) +#define vluxseg7ei32_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei32_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vfloat16mf2_t *)(op6), (const _Float16 *)(op7), (vuint32m1_t)(op8), (size_t)(op9)) +#define vluxseg7ei32_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei32_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vfloat16mf2_t *)(op6), (vbool32_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (vfloat16mf2_t)(op11), (vfloat16mf2_t)(op12), (vfloat16mf2_t)(op13), (vfloat16mf2_t)(op14), (const _Float16 *)(op15), (vuint32m1_t)(op16), (size_t)(op17)) +#define vluxseg7ei32_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei32_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vfloat16mf4_t *)(op6), (const _Float16 *)(op7), (vuint32mf2_t)(op8), (size_t)(op9)) +#define vluxseg7ei32_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei32_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vfloat16mf4_t *)(op6), (vbool64_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (vfloat16mf4_t)(op11), (vfloat16mf4_t)(op12), (vfloat16mf4_t)(op13), (vfloat16mf4_t)(op14), (const _Float16 *)(op15), (vuint32mf2_t)(op16), (size_t)(op17)) +#define vluxseg8ei32_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei32_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vfloat16m1_t *)(op6), (vfloat16m1_t *)(op7), (const _Float16 *)(op8), (vuint32m2_t)(op9), (size_t)(op10)) +#define vluxseg8ei32_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei32_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vfloat16m1_t *)(op6), (vfloat16m1_t *)(op7), (vbool16_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (vfloat16m1_t)(op11), (vfloat16m1_t)(op12), (vfloat16m1_t)(op13), (vfloat16m1_t)(op14), (vfloat16m1_t)(op15), (vfloat16m1_t)(op16), (const _Float16 *)(op17), (vuint32m2_t)(op18), (size_t)(op19)) +#define vluxseg8ei32_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei32_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vfloat16mf2_t *)(op6), (vfloat16mf2_t *)(op7), (const _Float16 *)(op8), (vuint32m1_t)(op9), (size_t)(op10)) +#define vluxseg8ei32_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei32_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vfloat16mf2_t *)(op6), (vfloat16mf2_t *)(op7), (vbool32_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (vfloat16mf2_t)(op11), (vfloat16mf2_t)(op12), (vfloat16mf2_t)(op13), (vfloat16mf2_t)(op14), (vfloat16mf2_t)(op15), (vfloat16mf2_t)(op16), (const _Float16 *)(op17), (vuint32m1_t)(op18), (size_t)(op19)) +#define vluxseg8ei32_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei32_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vfloat16mf4_t *)(op6), (vfloat16mf4_t *)(op7), (const _Float16 *)(op8), (vuint32mf2_t)(op9), (size_t)(op10)) +#define vluxseg8ei32_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei32_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vfloat16mf4_t *)(op6), (vfloat16mf4_t *)(op7), (vbool64_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (vfloat16mf4_t)(op11), (vfloat16mf4_t)(op12), (vfloat16mf4_t)(op13), (vfloat16mf4_t)(op14), (vfloat16mf4_t)(op15), (vfloat16mf4_t)(op16), (const _Float16 *)(op17), (vuint32mf2_t)(op18), (size_t)(op19)) +#define vluxseg2ei64_v_f16m1(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei64_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (const _Float16 *)(op2), (vuint64m4_t)(op3), (size_t)(op4)) +#define vluxseg2ei64_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei64_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vbool16_t)(op2), (vfloat16m1_t)(op3), (vfloat16m1_t)(op4), (const _Float16 *)(op5), (vuint64m4_t)(op6), (size_t)(op7)) +#define vluxseg2ei64_v_f16m2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei64_v_f16m2((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (const _Float16 *)(op2), (vuint64m8_t)(op3), (size_t)(op4)) +#define vluxseg2ei64_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei64_v_f16m2_m((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vbool8_t)(op2), (vfloat16m2_t)(op3), (vfloat16m2_t)(op4), (const _Float16 *)(op5), (vuint64m8_t)(op6), (size_t)(op7)) +#define vluxseg2ei64_v_f16mf2(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei64_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (const _Float16 *)(op2), (vuint64m2_t)(op3), (size_t)(op4)) +#define vluxseg2ei64_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei64_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vbool32_t)(op2), (vfloat16mf2_t)(op3), (vfloat16mf2_t)(op4), (const _Float16 *)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vluxseg2ei64_v_f16mf4(op0, op1, op2, op3, op4) \ +__builtin_rvv_vluxseg2ei64_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (const _Float16 *)(op2), (vuint64m1_t)(op3), (size_t)(op4)) +#define vluxseg2ei64_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg2ei64_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vbool64_t)(op2), (vfloat16mf4_t)(op3), (vfloat16mf4_t)(op4), (const _Float16 *)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vluxseg3ei64_v_f16m1(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei64_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (const _Float16 *)(op3), (vuint64m4_t)(op4), (size_t)(op5)) +#define vluxseg3ei64_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei64_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vbool16_t)(op3), (vfloat16m1_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (const _Float16 *)(op7), (vuint64m4_t)(op8), (size_t)(op9)) +#define vluxseg3ei64_v_f16m2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei64_v_f16m2((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vfloat16m2_t *)(op2), (const _Float16 *)(op3), (vuint64m8_t)(op4), (size_t)(op5)) +#define vluxseg3ei64_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei64_v_f16m2_m((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vfloat16m2_t *)(op2), (vbool8_t)(op3), (vfloat16m2_t)(op4), (vfloat16m2_t)(op5), (vfloat16m2_t)(op6), (const _Float16 *)(op7), (vuint64m8_t)(op8), (size_t)(op9)) +#define vluxseg3ei64_v_f16mf2(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei64_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (const _Float16 *)(op3), (vuint64m2_t)(op4), (size_t)(op5)) +#define vluxseg3ei64_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei64_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vbool32_t)(op3), (vfloat16mf2_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (const _Float16 *)(op7), (vuint64m2_t)(op8), (size_t)(op9)) +#define vluxseg3ei64_v_f16mf4(op0, op1, op2, op3, op4, op5) \ +__builtin_rvv_vluxseg3ei64_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (const _Float16 *)(op3), (vuint64m1_t)(op4), (size_t)(op5)) +#define vluxseg3ei64_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg3ei64_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vbool64_t)(op3), (vfloat16mf4_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (const _Float16 *)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vluxseg4ei64_v_f16m1(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei64_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (const _Float16 *)(op4), (vuint64m4_t)(op5), (size_t)(op6)) +#define vluxseg4ei64_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei64_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vbool16_t)(op4), (vfloat16m1_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (const _Float16 *)(op9), (vuint64m4_t)(op10), (size_t)(op11)) +#define vluxseg4ei64_v_f16m2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei64_v_f16m2((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vfloat16m2_t *)(op2), (vfloat16m2_t *)(op3), (const _Float16 *)(op4), (vuint64m8_t)(op5), (size_t)(op6)) +#define vluxseg4ei64_v_f16m2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei64_v_f16m2_m((vfloat16m2_t *)(op0), (vfloat16m2_t *)(op1), (vfloat16m2_t *)(op2), (vfloat16m2_t *)(op3), (vbool8_t)(op4), (vfloat16m2_t)(op5), (vfloat16m2_t)(op6), (vfloat16m2_t)(op7), (vfloat16m2_t)(op8), (const _Float16 *)(op9), (vuint64m8_t)(op10), (size_t)(op11)) +#define vluxseg4ei64_v_f16mf2(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei64_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (const _Float16 *)(op4), (vuint64m2_t)(op5), (size_t)(op6)) +#define vluxseg4ei64_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei64_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vbool32_t)(op4), (vfloat16mf2_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (const _Float16 *)(op9), (vuint64m2_t)(op10), (size_t)(op11)) +#define vluxseg4ei64_v_f16mf4(op0, op1, op2, op3, op4, op5, op6) \ +__builtin_rvv_vluxseg4ei64_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (const _Float16 *)(op4), (vuint64m1_t)(op5), (size_t)(op6)) +#define vluxseg4ei64_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11) \ +__builtin_rvv_vluxseg4ei64_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vbool64_t)(op4), (vfloat16mf4_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (const _Float16 *)(op9), (vuint64m1_t)(op10), (size_t)(op11)) +#define vluxseg5ei64_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei64_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (const _Float16 *)(op5), (vuint64m4_t)(op6), (size_t)(op7)) +#define vluxseg5ei64_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei64_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vbool16_t)(op5), (vfloat16m1_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (const _Float16 *)(op11), (vuint64m4_t)(op12), (size_t)(op13)) +#define vluxseg5ei64_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei64_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (const _Float16 *)(op5), (vuint64m2_t)(op6), (size_t)(op7)) +#define vluxseg5ei64_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei64_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vbool32_t)(op5), (vfloat16mf2_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (const _Float16 *)(op11), (vuint64m2_t)(op12), (size_t)(op13)) +#define vluxseg5ei64_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7) \ +__builtin_rvv_vluxseg5ei64_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (const _Float16 *)(op5), (vuint64m1_t)(op6), (size_t)(op7)) +#define vluxseg5ei64_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13) \ +__builtin_rvv_vluxseg5ei64_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vbool64_t)(op5), (vfloat16mf4_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (const _Float16 *)(op11), (vuint64m1_t)(op12), (size_t)(op13)) +#define vluxseg6ei64_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei64_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (const _Float16 *)(op6), (vuint64m4_t)(op7), (size_t)(op8)) +#define vluxseg6ei64_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei64_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vbool16_t)(op6), (vfloat16m1_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (vfloat16m1_t)(op11), (vfloat16m1_t)(op12), (const _Float16 *)(op13), (vuint64m4_t)(op14), (size_t)(op15)) +#define vluxseg6ei64_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei64_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (const _Float16 *)(op6), (vuint64m2_t)(op7), (size_t)(op8)) +#define vluxseg6ei64_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei64_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vbool32_t)(op6), (vfloat16mf2_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (vfloat16mf2_t)(op11), (vfloat16mf2_t)(op12), (const _Float16 *)(op13), (vuint64m2_t)(op14), (size_t)(op15)) +#define vluxseg6ei64_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8) \ +__builtin_rvv_vluxseg6ei64_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (const _Float16 *)(op6), (vuint64m1_t)(op7), (size_t)(op8)) +#define vluxseg6ei64_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15) \ +__builtin_rvv_vluxseg6ei64_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vbool64_t)(op6), (vfloat16mf4_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (vfloat16mf4_t)(op11), (vfloat16mf4_t)(op12), (const _Float16 *)(op13), (vuint64m1_t)(op14), (size_t)(op15)) +#define vluxseg7ei64_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei64_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vfloat16m1_t *)(op6), (const _Float16 *)(op7), (vuint64m4_t)(op8), (size_t)(op9)) +#define vluxseg7ei64_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei64_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vfloat16m1_t *)(op6), (vbool16_t)(op7), (vfloat16m1_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (vfloat16m1_t)(op11), (vfloat16m1_t)(op12), (vfloat16m1_t)(op13), (vfloat16m1_t)(op14), (const _Float16 *)(op15), (vuint64m4_t)(op16), (size_t)(op17)) +#define vluxseg7ei64_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei64_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vfloat16mf2_t *)(op6), (const _Float16 *)(op7), (vuint64m2_t)(op8), (size_t)(op9)) +#define vluxseg7ei64_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei64_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vfloat16mf2_t *)(op6), (vbool32_t)(op7), (vfloat16mf2_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (vfloat16mf2_t)(op11), (vfloat16mf2_t)(op12), (vfloat16mf2_t)(op13), (vfloat16mf2_t)(op14), (const _Float16 *)(op15), (vuint64m2_t)(op16), (size_t)(op17)) +#define vluxseg7ei64_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9) \ +__builtin_rvv_vluxseg7ei64_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vfloat16mf4_t *)(op6), (const _Float16 *)(op7), (vuint64m1_t)(op8), (size_t)(op9)) +#define vluxseg7ei64_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17) \ +__builtin_rvv_vluxseg7ei64_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vfloat16mf4_t *)(op6), (vbool64_t)(op7), (vfloat16mf4_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (vfloat16mf4_t)(op11), (vfloat16mf4_t)(op12), (vfloat16mf4_t)(op13), (vfloat16mf4_t)(op14), (const _Float16 *)(op15), (vuint64m1_t)(op16), (size_t)(op17)) +#define vluxseg8ei64_v_f16m1(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei64_v_f16m1((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vfloat16m1_t *)(op6), (vfloat16m1_t *)(op7), (const _Float16 *)(op8), (vuint64m4_t)(op9), (size_t)(op10)) +#define vluxseg8ei64_v_f16m1_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei64_v_f16m1_m((vfloat16m1_t *)(op0), (vfloat16m1_t *)(op1), (vfloat16m1_t *)(op2), (vfloat16m1_t *)(op3), (vfloat16m1_t *)(op4), (vfloat16m1_t *)(op5), (vfloat16m1_t *)(op6), (vfloat16m1_t *)(op7), (vbool16_t)(op8), (vfloat16m1_t)(op9), (vfloat16m1_t)(op10), (vfloat16m1_t)(op11), (vfloat16m1_t)(op12), (vfloat16m1_t)(op13), (vfloat16m1_t)(op14), (vfloat16m1_t)(op15), (vfloat16m1_t)(op16), (const _Float16 *)(op17), (vuint64m4_t)(op18), (size_t)(op19)) +#define vluxseg8ei64_v_f16mf2(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei64_v_f16mf2((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vfloat16mf2_t *)(op6), (vfloat16mf2_t *)(op7), (const _Float16 *)(op8), (vuint64m2_t)(op9), (size_t)(op10)) +#define vluxseg8ei64_v_f16mf2_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei64_v_f16mf2_m((vfloat16mf2_t *)(op0), (vfloat16mf2_t *)(op1), (vfloat16mf2_t *)(op2), (vfloat16mf2_t *)(op3), (vfloat16mf2_t *)(op4), (vfloat16mf2_t *)(op5), (vfloat16mf2_t *)(op6), (vfloat16mf2_t *)(op7), (vbool32_t)(op8), (vfloat16mf2_t)(op9), (vfloat16mf2_t)(op10), (vfloat16mf2_t)(op11), (vfloat16mf2_t)(op12), (vfloat16mf2_t)(op13), (vfloat16mf2_t)(op14), (vfloat16mf2_t)(op15), (vfloat16mf2_t)(op16), (const _Float16 *)(op17), (vuint64m2_t)(op18), (size_t)(op19)) +#define vluxseg8ei64_v_f16mf4(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10) \ +__builtin_rvv_vluxseg8ei64_v_f16mf4((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vfloat16mf4_t *)(op6), (vfloat16mf4_t *)(op7), (const _Float16 *)(op8), (vuint64m1_t)(op9), (size_t)(op10)) +#define vluxseg8ei64_v_f16mf4_m(op0, op1, op2, op3, op4, op5, op6, op7, op8, op9, op10, op11, op12, op13, op14, op15, op16, op17, op18, op19) \ +__builtin_rvv_vluxseg8ei64_v_f16mf4_m((vfloat16mf4_t *)(op0), (vfloat16mf4_t *)(op1), (vfloat16mf4_t *)(op2), (vfloat16mf4_t *)(op3), (vfloat16mf4_t *)(op4), (vfloat16mf4_t *)(op5), (vfloat16mf4_t *)(op6), (vfloat16mf4_t *)(op7), (vbool64_t)(op8), (vfloat16mf4_t)(op9), (vfloat16mf4_t)(op10), (vfloat16mf4_t)(op11), (vfloat16mf4_t)(op12), (vfloat16mf4_t)(op13), (vfloat16mf4_t)(op14), (vfloat16mf4_t)(op15), (vfloat16mf4_t)(op16), (const _Float16 *)(op17), (vuint64m1_t)(op18), (size_t)(op19)) +#endif + +#define __riscv_v_intrinsic_overloading 1 +#define __rvv_overloaded static inline __attribute__((__always_inline__, __nodebug__, __overloadable__)) +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_i8m1))) +vint8m1_t vadd(vint8m1_t op0, vint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_i8m1_m))) +vint8m1_t vadd(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_i8m2))) +vint8m2_t vadd(vint8m2_t op0, vint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_i8m2_m))) +vint8m2_t vadd(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_i8m4))) +vint8m4_t vadd(vint8m4_t op0, vint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_i8m4_m))) +vint8m4_t vadd(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_i8m8))) +vint8m8_t vadd(vint8m8_t op0, vint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_i8m8_m))) +vint8m8_t vadd(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, vint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_i8mf2))) +vint8mf2_t vadd(vint8mf2_t op0, vint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_i8mf2_m))) +vint8mf2_t vadd(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_i8mf4))) +vint8mf4_t vadd(vint8mf4_t op0, vint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_i8mf4_m))) +vint8mf4_t vadd(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_i8mf8))) +vint8mf8_t vadd(vint8mf8_t op0, vint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_i8mf8_m))) +vint8mf8_t vadd(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_i16m1))) +vint16m1_t vadd(vint16m1_t op0, vint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_i16m1_m))) +vint16m1_t vadd(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_i16m2))) +vint16m2_t vadd(vint16m2_t op0, vint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_i16m2_m))) +vint16m2_t vadd(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_i16m4))) +vint16m4_t vadd(vint16m4_t op0, vint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_i16m4_m))) +vint16m4_t vadd(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_i16m8))) +vint16m8_t vadd(vint16m8_t op0, vint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_i16m8_m))) +vint16m8_t vadd(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, vint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_i16mf2))) +vint16mf2_t vadd(vint16mf2_t op0, vint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_i16mf2_m))) +vint16mf2_t vadd(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_i16mf4))) +vint16mf4_t vadd(vint16mf4_t op0, vint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_i16mf4_m))) +vint16mf4_t vadd(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_i32m1))) +vint32m1_t vadd(vint32m1_t op0, vint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_i32m1_m))) +vint32m1_t vadd(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_i32m2))) +vint32m2_t vadd(vint32m2_t op0, vint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_i32m2_m))) +vint32m2_t vadd(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_i32m4))) +vint32m4_t vadd(vint32m4_t op0, vint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_i32m4_m))) +vint32m4_t vadd(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_i32m8))) +vint32m8_t vadd(vint32m8_t op0, vint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_i32m8_m))) +vint32m8_t vadd(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_i32mf2))) +vint32mf2_t vadd(vint32mf2_t op0, vint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_i32mf2_m))) +vint32mf2_t vadd(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_i64m1))) +vint64m1_t vadd(vint64m1_t op0, vint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_i64m1_m))) +vint64m1_t vadd(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_i64m2))) +vint64m2_t vadd(vint64m2_t op0, vint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_i64m2_m))) +vint64m2_t vadd(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_i64m4))) +vint64m4_t vadd(vint64m4_t op0, vint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_i64m4_m))) +vint64m4_t vadd(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_i64m8))) +vint64m8_t vadd(vint64m8_t op0, vint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_i64m8_m))) +vint64m8_t vadd(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vv_u16mf4))) +vuint16mf4_t vwaddu_vv(vuint8mf8_t op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vv_u16mf4_m))) +vuint16mf4_t vwaddu_vv(vbool64_t op0, vuint16mf4_t op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vv_u16mf2))) +vuint16mf2_t vwaddu_vv(vuint8mf4_t op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vv_u16mf2_m))) +vuint16mf2_t vwaddu_vv(vbool32_t op0, vuint16mf2_t op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vv_u16m1))) +vuint16m1_t vwaddu_vv(vuint8mf2_t op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vv_u16m1_m))) +vuint16m1_t vwaddu_vv(vbool16_t op0, vuint16m1_t op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vv_u16m2))) +vuint16m2_t vwaddu_vv(vuint8m1_t op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vv_u16m2_m))) +vuint16m2_t vwaddu_vv(vbool8_t op0, vuint16m2_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vv_u16m4))) +vuint16m4_t vwaddu_vv(vuint8m2_t op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vv_u16m4_m))) +vuint16m4_t vwaddu_vv(vbool4_t op0, vuint16m4_t op1, vuint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vv_u16m8))) +vuint16m8_t vwaddu_vv(vuint8m4_t op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vv_u16m8_m))) +vuint16m8_t vwaddu_vv(vbool2_t op0, vuint16m8_t op1, vuint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vv_u32mf2))) +vuint32mf2_t vwaddu_vv(vuint16mf4_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vv_u32mf2_m))) +vuint32mf2_t vwaddu_vv(vbool64_t op0, vuint32mf2_t op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vv_u32m1))) +vuint32m1_t vwaddu_vv(vuint16mf2_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vv_u32m1_m))) +vuint32m1_t vwaddu_vv(vbool32_t op0, vuint32m1_t op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vv_u32m2))) +vuint32m2_t vwaddu_vv(vuint16m1_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vv_u32m2_m))) +vuint32m2_t vwaddu_vv(vbool16_t op0, vuint32m2_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vv_u32m4))) +vuint32m4_t vwaddu_vv(vuint16m2_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vv_u32m4_m))) +vuint32m4_t vwaddu_vv(vbool8_t op0, vuint32m4_t op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vv_u32m8))) +vuint32m8_t vwaddu_vv(vuint16m4_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vv_u32m8_m))) +vuint32m8_t vwaddu_vv(vbool4_t op0, vuint32m8_t op1, vuint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vv_u64m1))) +vuint64m1_t vwaddu_vv(vuint32mf2_t op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vv_u64m1_m))) +vuint64m1_t vwaddu_vv(vbool64_t op0, vuint64m1_t op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vv_u64m2))) +vuint64m2_t vwaddu_vv(vuint32m1_t op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vv_u64m2_m))) +vuint64m2_t vwaddu_vv(vbool32_t op0, vuint64m2_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vv_u64m4))) +vuint64m4_t vwaddu_vv(vuint32m2_t op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vv_u64m4_m))) +vuint64m4_t vwaddu_vv(vbool16_t op0, vuint64m4_t op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vv_u64m8))) +vuint64m8_t vwaddu_vv(vuint32m4_t op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vv_u64m8_m))) +vuint64m8_t vwaddu_vv(vbool8_t op0, vuint64m8_t op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_u8m1))) +vuint8m1_t vluxei8(const uint8_t * op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_u8m1_m))) +vuint8m1_t vluxei8(vbool8_t op0, vuint8m1_t op1, const uint8_t * op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_u8m2))) +vuint8m2_t vluxei8(const uint8_t * op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_u8m2_m))) +vuint8m2_t vluxei8(vbool4_t op0, vuint8m2_t op1, const uint8_t * op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_u8m4))) +vuint8m4_t vluxei8(const uint8_t * op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_u8m4_m))) +vuint8m4_t vluxei8(vbool2_t op0, vuint8m4_t op1, const uint8_t * op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_u8m8))) +vuint8m8_t vluxei8(const uint8_t * op0, vuint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_u8m8_m))) +vuint8m8_t vluxei8(vbool1_t op0, vuint8m8_t op1, const uint8_t * op2, vuint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_u8mf2))) +vuint8mf2_t vluxei8(const uint8_t * op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_u8mf2_m))) +vuint8mf2_t vluxei8(vbool16_t op0, vuint8mf2_t op1, const uint8_t * op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_u8mf4))) +vuint8mf4_t vluxei8(const uint8_t * op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_u8mf4_m))) +vuint8mf4_t vluxei8(vbool32_t op0, vuint8mf4_t op1, const uint8_t * op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_u8mf8))) +vuint8mf8_t vluxei8(const uint8_t * op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_u8mf8_m))) +vuint8mf8_t vluxei8(vbool64_t op0, vuint8mf8_t op1, const uint8_t * op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse64_v_u64m1_m))) +vuint64m1_t vlse64(vbool64_t op0, vuint64m1_t op1, const uint64_t * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse64_v_u64m2_m))) +vuint64m2_t vlse64(vbool32_t op0, vuint64m2_t op1, const uint64_t * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse64_v_u64m4_m))) +vuint64m4_t vlse64(vbool16_t op0, vuint64m4_t op1, const uint64_t * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse64_v_u64m8_m))) +vuint64m8_t vlse64(vbool8_t op0, vuint64m8_t op1, const uint64_t * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse16_v_i16m1))) +void vsse16(int16_t * op0, ptrdiff_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse16_v_i16m1_m))) +void vsse16(vbool16_t op0, int16_t * op1, ptrdiff_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse16_v_i16m2))) +void vsse16(int16_t * op0, ptrdiff_t op1, vint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse16_v_i16m2_m))) +void vsse16(vbool8_t op0, int16_t * op1, ptrdiff_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse16_v_i16m4))) +void vsse16(int16_t * op0, ptrdiff_t op1, vint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse16_v_i16m4_m))) +void vsse16(vbool4_t op0, int16_t * op1, ptrdiff_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse16_v_i16m8))) +void vsse16(int16_t * op0, ptrdiff_t op1, vint16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse16_v_i16m8_m))) +void vsse16(vbool2_t op0, int16_t * op1, ptrdiff_t op2, vint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse16_v_i16mf2))) +void vsse16(int16_t * op0, ptrdiff_t op1, vint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse16_v_i16mf2_m))) +void vsse16(vbool32_t op0, int16_t * op1, ptrdiff_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse16_v_i16mf4))) +void vsse16(int16_t * op0, ptrdiff_t op1, vint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse16_v_i16mf4_m))) +void vsse16(vbool64_t op0, int16_t * op1, ptrdiff_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse16_v_u16m1))) +void vsse16(uint16_t * op0, ptrdiff_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse16_v_u16m1_m))) +void vsse16(vbool16_t op0, uint16_t * op1, ptrdiff_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse16_v_u16m2))) +void vsse16(uint16_t * op0, ptrdiff_t op1, vuint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse16_v_u16m2_m))) +void vsse16(vbool8_t op0, uint16_t * op1, ptrdiff_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse16_v_u16m4))) +void vsse16(uint16_t * op0, ptrdiff_t op1, vuint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse16_v_u16m4_m))) +void vsse16(vbool4_t op0, uint16_t * op1, ptrdiff_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse16_v_u16m8))) +void vsse16(uint16_t * op0, ptrdiff_t op1, vuint16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse16_v_u16m8_m))) +void vsse16(vbool2_t op0, uint16_t * op1, ptrdiff_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse16_v_u16mf2))) +void vsse16(uint16_t * op0, ptrdiff_t op1, vuint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse16_v_u16mf2_m))) +void vsse16(vbool32_t op0, uint16_t * op1, ptrdiff_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse16_v_u16mf4))) +void vsse16(uint16_t * op0, ptrdiff_t op1, vuint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse16_v_u16mf4_m))) +void vsse16(vbool64_t op0, uint16_t * op1, ptrdiff_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse32_v_i32m1))) +void vsse32(int32_t * op0, ptrdiff_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse32_v_i32m1_m))) +void vsse32(vbool32_t op0, int32_t * op1, ptrdiff_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse32_v_i32m2))) +void vsse32(int32_t * op0, ptrdiff_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse32_v_i32m2_m))) +void vsse32(vbool16_t op0, int32_t * op1, ptrdiff_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse32_v_i32m4))) +void vsse32(int32_t * op0, ptrdiff_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse32_v_i32m4_m))) +void vsse32(vbool8_t op0, int32_t * op1, ptrdiff_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse32_v_i32m8))) +void vsse32(int32_t * op0, ptrdiff_t op1, vint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse32_v_i32m8_m))) +void vsse32(vbool4_t op0, int32_t * op1, ptrdiff_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse32_v_i32mf2))) +void vsse32(int32_t * op0, ptrdiff_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse32_v_i32mf2_m))) +void vsse32(vbool64_t op0, int32_t * op1, ptrdiff_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse32_v_u32m1))) +void vsse32(uint32_t * op0, ptrdiff_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse32_v_u32m1_m))) +void vsse32(vbool32_t op0, uint32_t * op1, ptrdiff_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse32_v_u32m2))) +void vsse32(uint32_t * op0, ptrdiff_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse32_v_u32m2_m))) +void vsse32(vbool16_t op0, uint32_t * op1, ptrdiff_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse32_v_u32m4))) +void vsse32(uint32_t * op0, ptrdiff_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse32_v_u32m4_m))) +void vsse32(vbool8_t op0, uint32_t * op1, ptrdiff_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse32_v_u32m8))) +void vsse32(uint32_t * op0, ptrdiff_t op1, vuint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse32_v_u32m8_m))) +void vsse32(vbool4_t op0, uint32_t * op1, ptrdiff_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse32_v_u32mf2))) +void vsse32(uint32_t * op0, ptrdiff_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse32_v_u32mf2_m))) +void vsse32(vbool64_t op0, uint32_t * op1, ptrdiff_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse64_v_i64m1))) +void vsse64(int64_t * op0, ptrdiff_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse64_v_i64m1_m))) +void vsse64(vbool64_t op0, int64_t * op1, ptrdiff_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse64_v_i64m2))) +void vsse64(int64_t * op0, ptrdiff_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse64_v_i64m2_m))) +void vsse64(vbool32_t op0, int64_t * op1, ptrdiff_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse64_v_i64m4))) +void vsse64(int64_t * op0, ptrdiff_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse64_v_i64m4_m))) +void vsse64(vbool16_t op0, int64_t * op1, ptrdiff_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse64_v_i64m8))) +void vsse64(int64_t * op0, ptrdiff_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse64_v_i64m8_m))) +void vsse64(vbool8_t op0, int64_t * op1, ptrdiff_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse64_v_u64m1))) +void vsse64(uint64_t * op0, ptrdiff_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse64_v_u64m1_m))) +void vsse64(vbool64_t op0, uint64_t * op1, ptrdiff_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse64_v_u64m2))) +void vsse64(uint64_t * op0, ptrdiff_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse64_v_u64m2_m))) +void vsse64(vbool32_t op0, uint64_t * op1, ptrdiff_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse64_v_u64m4))) +void vsse64(uint64_t * op0, ptrdiff_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse64_v_u64m4_m))) +void vsse64(vbool16_t op0, uint64_t * op1, ptrdiff_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse64_v_u64m8))) +void vsse64(uint64_t * op0, ptrdiff_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse64_v_u64m8_m))) +void vsse64(vbool8_t op0, uint64_t * op1, ptrdiff_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_i8m1))) +vint8m1_t vluxei16(const int8_t * op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_i8m1_m))) +vint8m1_t vluxei16(vbool8_t op0, vint8m1_t op1, const int8_t * op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_i8m2))) +vint8m2_t vluxei16(const int8_t * op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_i8m2_m))) +vint8m2_t vluxei16(vbool4_t op0, vint8m2_t op1, const int8_t * op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_i8m4))) +vint8m4_t vluxei16(const int8_t * op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_i8m4_m))) +vint8m4_t vluxei16(vbool2_t op0, vint8m4_t op1, const int8_t * op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_i8mf2))) +vint8mf2_t vluxei16(const int8_t * op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_i8mf2_m))) +vint8mf2_t vluxei16(vbool16_t op0, vint8mf2_t op1, const int8_t * op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_i8mf4))) +vint8mf4_t vluxei16(const int8_t * op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_i8mf4_m))) +vint8mf4_t vluxei16(vbool32_t op0, vint8mf4_t op1, const int8_t * op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_i8mf8))) +vint8mf8_t vluxei16(const int8_t * op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_i8mf8_m))) +vint8mf8_t vluxei16(vbool64_t op0, vint8mf8_t op1, const int8_t * op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_u8m1))) +vuint8m1_t vluxei16(const uint8_t * op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_u8m1_m))) +vuint8m1_t vluxei16(vbool8_t op0, vuint8m1_t op1, const uint8_t * op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_u8m2))) +vuint8m2_t vluxei16(const uint8_t * op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_u8m2_m))) +vuint8m2_t vluxei16(vbool4_t op0, vuint8m2_t op1, const uint8_t * op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_u8m4))) +vuint8m4_t vluxei16(const uint8_t * op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_u8m4_m))) +vuint8m4_t vluxei16(vbool2_t op0, vuint8m4_t op1, const uint8_t * op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_u8mf2))) +vuint8mf2_t vluxei16(const uint8_t * op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_u8mf2_m))) +vuint8mf2_t vluxei16(vbool16_t op0, vuint8mf2_t op1, const uint8_t * op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_u8mf4))) +vuint8mf4_t vluxei16(const uint8_t * op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_u8mf4_m))) +vuint8mf4_t vluxei16(vbool32_t op0, vuint8mf4_t op1, const uint8_t * op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_u8mf8))) +vuint8mf8_t vluxei16(const uint8_t * op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_u8mf8_m))) +vuint8mf8_t vluxei16(vbool64_t op0, vuint8mf8_t op1, const uint8_t * op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_i8m1))) +vint8m1_t vluxei32(const int8_t * op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_i8m1_m))) +vint8m1_t vluxei32(vbool8_t op0, vint8m1_t op1, const int8_t * op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_i8m2))) +vint8m2_t vluxei32(const int8_t * op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_i8m2_m))) +vint8m2_t vluxei32(vbool4_t op0, vint8m2_t op1, const int8_t * op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_i8mf2))) +vint8mf2_t vluxei32(const int8_t * op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_i8mf2_m))) +vint8mf2_t vluxei32(vbool16_t op0, vint8mf2_t op1, const int8_t * op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_i8mf4))) +vint8mf4_t vluxei32(const int8_t * op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_i8mf4_m))) +vint8mf4_t vluxei32(vbool32_t op0, vint8mf4_t op1, const int8_t * op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_i8mf8))) +vint8mf8_t vluxei32(const int8_t * op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_i8mf8_m))) +vint8mf8_t vluxei32(vbool64_t op0, vint8mf8_t op1, const int8_t * op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_u8m1))) +vuint8m1_t vluxei32(const uint8_t * op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_u8m1_m))) +vuint8m1_t vluxei32(vbool8_t op0, vuint8m1_t op1, const uint8_t * op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_u8m2))) +vuint8m2_t vluxei32(const uint8_t * op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_u8m2_m))) +vuint8m2_t vluxei32(vbool4_t op0, vuint8m2_t op1, const uint8_t * op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_u8mf2))) +vuint8mf2_t vluxei32(const uint8_t * op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_u8mf2_m))) +vuint8mf2_t vluxei32(vbool16_t op0, vuint8mf2_t op1, const uint8_t * op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_u8mf4))) +vuint8mf4_t vluxei32(const uint8_t * op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_u8mf4_m))) +vuint8mf4_t vluxei32(vbool32_t op0, vuint8mf4_t op1, const uint8_t * op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_u8mf8))) +vuint8mf8_t vluxei32(const uint8_t * op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_u8mf8_m))) +vuint8mf8_t vluxei32(vbool64_t op0, vuint8mf8_t op1, const uint8_t * op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_i8m1))) +vint8m1_t vluxei64(const int8_t * op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_i8m1_m))) +vint8m1_t vluxei64(vbool8_t op0, vint8m1_t op1, const int8_t * op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_i8mf2))) +vint8mf2_t vluxei64(const int8_t * op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_i8mf2_m))) +vint8mf2_t vluxei64(vbool16_t op0, vint8mf2_t op1, const int8_t * op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_i8mf4))) +vint8mf4_t vluxei64(const int8_t * op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_i8mf4_m))) +vint8mf4_t vluxei64(vbool32_t op0, vint8mf4_t op1, const int8_t * op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_i8mf8))) +vint8mf8_t vluxei64(const int8_t * op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_i8mf8_m))) +vint8mf8_t vluxei64(vbool64_t op0, vint8mf8_t op1, const int8_t * op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_u8m1))) +vuint8m1_t vluxei64(const uint8_t * op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_u8m1_m))) +vuint8m1_t vluxei64(vbool8_t op0, vuint8m1_t op1, const uint8_t * op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_u8mf2))) +vuint8mf2_t vluxei64(const uint8_t * op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_u8mf2_m))) +vuint8mf2_t vluxei64(vbool16_t op0, vuint8mf2_t op1, const uint8_t * op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_u8mf4))) +vuint8mf4_t vluxei64(const uint8_t * op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_u8mf4_m))) +vuint8mf4_t vluxei64(vbool32_t op0, vuint8mf4_t op1, const uint8_t * op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_u8mf8))) +vuint8mf8_t vluxei64(const uint8_t * op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_u8mf8_m))) +vuint8mf8_t vluxei64(vbool64_t op0, vuint8mf8_t op1, const uint8_t * op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_i16m1))) +vint16m1_t vluxei8(const int16_t * op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_i16m1_m))) +vint16m1_t vluxei8(vbool16_t op0, vint16m1_t op1, const int16_t * op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_i16m2))) +vint16m2_t vluxei8(const int16_t * op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_i16m2_m))) +vint16m2_t vluxei8(vbool8_t op0, vint16m2_t op1, const int16_t * op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_i16m4))) +vint16m4_t vluxei8(const int16_t * op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_i16m4_m))) +vint16m4_t vluxei8(vbool4_t op0, vint16m4_t op1, const int16_t * op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_i16m8))) +vint16m8_t vluxei8(const int16_t * op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_i16m8_m))) +vint16m8_t vluxei8(vbool2_t op0, vint16m8_t op1, const int16_t * op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_i16mf2))) +vint16mf2_t vluxei8(const int16_t * op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_i16mf2_m))) +vint16mf2_t vluxei8(vbool32_t op0, vint16mf2_t op1, const int16_t * op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_i16mf4))) +vint16mf4_t vluxei8(const int16_t * op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_i16mf4_m))) +vint16mf4_t vluxei8(vbool64_t op0, vint16mf4_t op1, const int16_t * op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_u16m1))) +vuint16m1_t vluxei8(const uint16_t * op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_u16m1_m))) +vuint16m1_t vluxei8(vbool16_t op0, vuint16m1_t op1, const uint16_t * op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_u16m2))) +vuint16m2_t vluxei8(const uint16_t * op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_u16m2_m))) +vuint16m2_t vluxei8(vbool8_t op0, vuint16m2_t op1, const uint16_t * op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_u16m4))) +vuint16m4_t vluxei8(const uint16_t * op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_u16m4_m))) +vuint16m4_t vluxei8(vbool4_t op0, vuint16m4_t op1, const uint16_t * op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_u16m8))) +vuint16m8_t vluxei8(const uint16_t * op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_u16m8_m))) +vuint16m8_t vluxei8(vbool2_t op0, vuint16m8_t op1, const uint16_t * op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_u16mf2))) +vuint16mf2_t vluxei8(const uint16_t * op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_u16mf2_m))) +vuint16mf2_t vluxei8(vbool32_t op0, vuint16mf2_t op1, const uint16_t * op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_u16mf4))) +vuint16mf4_t vluxei8(const uint16_t * op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_u16mf4_m))) +vuint16mf4_t vluxei8(vbool64_t op0, vuint16mf4_t op1, const uint16_t * op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_i16m1))) +vint16m1_t vluxei16(const int16_t * op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_i16m1_m))) +vint16m1_t vluxei16(vbool16_t op0, vint16m1_t op1, const int16_t * op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_i16m2))) +vint16m2_t vluxei16(const int16_t * op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_i16m2_m))) +vint16m2_t vluxei16(vbool8_t op0, vint16m2_t op1, const int16_t * op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_i16m4))) +vint16m4_t vluxei16(const int16_t * op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_i16m4_m))) +vint16m4_t vluxei16(vbool4_t op0, vint16m4_t op1, const int16_t * op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_i16m8))) +vint16m8_t vluxei16(const int16_t * op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_i16m8_m))) +vint16m8_t vluxei16(vbool2_t op0, vint16m8_t op1, const int16_t * op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_i16mf2))) +vint16mf2_t vluxei16(const int16_t * op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_i16mf2_m))) +vint16mf2_t vluxei16(vbool32_t op0, vint16mf2_t op1, const int16_t * op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_i16mf4))) +vint16mf4_t vluxei16(const int16_t * op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_i16mf4_m))) +vint16mf4_t vluxei16(vbool64_t op0, vint16mf4_t op1, const int16_t * op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_u16m1))) +vuint16m1_t vluxei16(const uint16_t * op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_u16m1_m))) +vuint16m1_t vluxei16(vbool16_t op0, vuint16m1_t op1, const uint16_t * op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_u16m2))) +vuint16m2_t vluxei16(const uint16_t * op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_u16m2_m))) +vuint16m2_t vluxei16(vbool8_t op0, vuint16m2_t op1, const uint16_t * op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_u16m4))) +vuint16m4_t vluxei16(const uint16_t * op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_u16m4_m))) +vuint16m4_t vluxei16(vbool4_t op0, vuint16m4_t op1, const uint16_t * op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_u16m8))) +vuint16m8_t vluxei16(const uint16_t * op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_u16m8_m))) +vuint16m8_t vluxei16(vbool2_t op0, vuint16m8_t op1, const uint16_t * op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_u16mf2))) +vuint16mf2_t vluxei16(const uint16_t * op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_u16mf2_m))) +vuint16mf2_t vluxei16(vbool32_t op0, vuint16mf2_t op1, const uint16_t * op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_u16mf4))) +vuint16mf4_t vluxei16(const uint16_t * op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_u16mf4_m))) +vuint16mf4_t vluxei16(vbool64_t op0, vuint16mf4_t op1, const uint16_t * op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_i16m1))) +vint16m1_t vluxei32(const int16_t * op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_i16m1_m))) +vint16m1_t vluxei32(vbool16_t op0, vint16m1_t op1, const int16_t * op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_i16m2))) +vint16m2_t vluxei32(const int16_t * op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_i16m2_m))) +vint16m2_t vluxei32(vbool8_t op0, vint16m2_t op1, const int16_t * op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_i16m4))) +vint16m4_t vluxei32(const int16_t * op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_i16m4_m))) +vint16m4_t vluxei32(vbool4_t op0, vint16m4_t op1, const int16_t * op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_i16mf2))) +vint16mf2_t vluxei32(const int16_t * op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_i16mf2_m))) +vint16mf2_t vluxei32(vbool32_t op0, vint16mf2_t op1, const int16_t * op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_i16mf4))) +vint16mf4_t vluxei32(const int16_t * op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_i16mf4_m))) +vint16mf4_t vluxei32(vbool64_t op0, vint16mf4_t op1, const int16_t * op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_u16m1))) +vuint16m1_t vluxei32(const uint16_t * op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_u16m1_m))) +vuint16m1_t vluxei32(vbool16_t op0, vuint16m1_t op1, const uint16_t * op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_u16m2))) +vuint16m2_t vluxei32(const uint16_t * op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_u16m2_m))) +vuint16m2_t vluxei32(vbool8_t op0, vuint16m2_t op1, const uint16_t * op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_u16m4))) +vuint16m4_t vluxei32(const uint16_t * op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_u16m4_m))) +vuint16m4_t vluxei32(vbool4_t op0, vuint16m4_t op1, const uint16_t * op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_u16mf2))) +vuint16mf2_t vluxei32(const uint16_t * op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_u16mf2_m))) +vuint16mf2_t vluxei32(vbool32_t op0, vuint16mf2_t op1, const uint16_t * op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_u16mf4))) +vuint16mf4_t vluxei32(const uint16_t * op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_u16mf4_m))) +vuint16mf4_t vluxei32(vbool64_t op0, vuint16mf4_t op1, const uint16_t * op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_i16m1))) +vint16m1_t vluxei64(const int16_t * op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_i16m1_m))) +vint16m1_t vluxei64(vbool16_t op0, vint16m1_t op1, const int16_t * op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_i16m2))) +vint16m2_t vluxei64(const int16_t * op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_i16m2_m))) +vint16m2_t vluxei64(vbool8_t op0, vint16m2_t op1, const int16_t * op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_i16mf2))) +vint16mf2_t vluxei64(const int16_t * op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_i16mf2_m))) +vint16mf2_t vluxei64(vbool32_t op0, vint16mf2_t op1, const int16_t * op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_i16mf4))) +vint16mf4_t vluxei64(const int16_t * op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_i16mf4_m))) +vint16mf4_t vluxei64(vbool64_t op0, vint16mf4_t op1, const int16_t * op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_u16m1))) +vuint16m1_t vluxei64(const uint16_t * op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_u16m1_m))) +vuint16m1_t vluxei64(vbool16_t op0, vuint16m1_t op1, const uint16_t * op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_u16m2))) +vuint16m2_t vluxei64(const uint16_t * op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_u16m2_m))) +vuint16m2_t vluxei64(vbool8_t op0, vuint16m2_t op1, const uint16_t * op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_u16mf2))) +vuint16mf2_t vluxei64(const uint16_t * op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_u16mf2_m))) +vuint16mf2_t vluxei64(vbool32_t op0, vuint16mf2_t op1, const uint16_t * op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_u16mf4))) +vuint16mf4_t vluxei64(const uint16_t * op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_u16mf4_m))) +vuint16mf4_t vluxei64(vbool64_t op0, vuint16mf4_t op1, const uint16_t * op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_i32m1))) +vint32m1_t vluxei8(const int32_t * op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_i32m1_m))) +vint32m1_t vluxei8(vbool32_t op0, vint32m1_t op1, const int32_t * op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_i32m2))) +vint32m2_t vluxei8(const int32_t * op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_i32m2_m))) +vint32m2_t vluxei8(vbool16_t op0, vint32m2_t op1, const int32_t * op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_i32m4))) +vint32m4_t vluxei8(const int32_t * op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_i32m4_m))) +vint32m4_t vluxei8(vbool8_t op0, vint32m4_t op1, const int32_t * op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_i32m8))) +vint32m8_t vluxei8(const int32_t * op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_i32m8_m))) +vint32m8_t vluxei8(vbool4_t op0, vint32m8_t op1, const int32_t * op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_i32mf2))) +vint32mf2_t vluxei8(const int32_t * op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_i32mf2_m))) +vint32mf2_t vluxei8(vbool64_t op0, vint32mf2_t op1, const int32_t * op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_u32m1))) +vuint32m1_t vluxei8(const uint32_t * op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_u32m1_m))) +vuint32m1_t vluxei8(vbool32_t op0, vuint32m1_t op1, const uint32_t * op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_u32m2))) +vuint32m2_t vluxei8(const uint32_t * op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_u32m2_m))) +vuint32m2_t vluxei8(vbool16_t op0, vuint32m2_t op1, const uint32_t * op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_u32m4))) +vuint32m4_t vluxei8(const uint32_t * op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_u32m4_m))) +vuint32m4_t vluxei8(vbool8_t op0, vuint32m4_t op1, const uint32_t * op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_u32m8))) +vuint32m8_t vluxei8(const uint32_t * op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_u32m8_m))) +vuint32m8_t vluxei8(vbool4_t op0, vuint32m8_t op1, const uint32_t * op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_u32mf2))) +vuint32mf2_t vluxei8(const uint32_t * op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_u32mf2_m))) +vuint32mf2_t vluxei8(vbool64_t op0, vuint32mf2_t op1, const uint32_t * op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_i32m1))) +vint32m1_t vluxei16(const int32_t * op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_i32m1_m))) +vint32m1_t vluxei16(vbool32_t op0, vint32m1_t op1, const int32_t * op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_i32m2))) +vint32m2_t vluxei16(const int32_t * op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_i32m2_m))) +vint32m2_t vluxei16(vbool16_t op0, vint32m2_t op1, const int32_t * op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_i32m4))) +vint32m4_t vluxei16(const int32_t * op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_i32m4_m))) +vint32m4_t vluxei16(vbool8_t op0, vint32m4_t op1, const int32_t * op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_i32m8))) +vint32m8_t vluxei16(const int32_t * op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_i32m8_m))) +vint32m8_t vluxei16(vbool4_t op0, vint32m8_t op1, const int32_t * op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_i32mf2))) +vint32mf2_t vluxei16(const int32_t * op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_i32mf2_m))) +vint32mf2_t vluxei16(vbool64_t op0, vint32mf2_t op1, const int32_t * op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_u32m1))) +vuint32m1_t vluxei16(const uint32_t * op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_u32m1_m))) +vuint32m1_t vluxei16(vbool32_t op0, vuint32m1_t op1, const uint32_t * op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_u32m2))) +vuint32m2_t vluxei16(const uint32_t * op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_u32m2_m))) +vuint32m2_t vluxei16(vbool16_t op0, vuint32m2_t op1, const uint32_t * op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_u32m4))) +vuint32m4_t vluxei16(const uint32_t * op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_u32m4_m))) +vuint32m4_t vluxei16(vbool8_t op0, vuint32m4_t op1, const uint32_t * op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_u32m8))) +vuint32m8_t vluxei16(const uint32_t * op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_u32m8_m))) +vuint32m8_t vluxei16(vbool4_t op0, vuint32m8_t op1, const uint32_t * op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_u32mf2))) +vuint32mf2_t vluxei16(const uint32_t * op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_u32mf2_m))) +vuint32mf2_t vluxei16(vbool64_t op0, vuint32mf2_t op1, const uint32_t * op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_i32m1))) +vint32m1_t vluxei32(const int32_t * op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_i32m1_m))) +vint32m1_t vluxei32(vbool32_t op0, vint32m1_t op1, const int32_t * op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_i32m2))) +vint32m2_t vluxei32(const int32_t * op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_i32m2_m))) +vint32m2_t vluxei32(vbool16_t op0, vint32m2_t op1, const int32_t * op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_i32m4))) +vint32m4_t vluxei32(const int32_t * op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_i32m4_m))) +vint32m4_t vluxei32(vbool8_t op0, vint32m4_t op1, const int32_t * op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_i32m8))) +vint32m8_t vluxei32(const int32_t * op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_i32m8_m))) +vint32m8_t vluxei32(vbool4_t op0, vint32m8_t op1, const int32_t * op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_i32mf2))) +vint32mf2_t vluxei32(const int32_t * op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_i32mf2_m))) +vint32mf2_t vluxei32(vbool64_t op0, vint32mf2_t op1, const int32_t * op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_u32m1))) +vuint32m1_t vluxei32(const uint32_t * op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_u32m1_m))) +vuint32m1_t vluxei32(vbool32_t op0, vuint32m1_t op1, const uint32_t * op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_u32m2))) +vuint32m2_t vluxei32(const uint32_t * op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_u32m2_m))) +vuint32m2_t vluxei32(vbool16_t op0, vuint32m2_t op1, const uint32_t * op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_u32m4))) +vuint32m4_t vluxei32(const uint32_t * op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_u32m4_m))) +vuint32m4_t vluxei32(vbool8_t op0, vuint32m4_t op1, const uint32_t * op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_u32m8))) +vuint32m8_t vluxei32(const uint32_t * op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_u32m8_m))) +vuint32m8_t vluxei32(vbool4_t op0, vuint32m8_t op1, const uint32_t * op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_u32mf2))) +vuint32mf2_t vluxei32(const uint32_t * op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_u32mf2_m))) +vuint32mf2_t vluxei32(vbool64_t op0, vuint32mf2_t op1, const uint32_t * op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_i32m1))) +vint32m1_t vluxei64(const int32_t * op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_i32m1_m))) +vint32m1_t vluxei64(vbool32_t op0, vint32m1_t op1, const int32_t * op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_i32m2))) +vint32m2_t vluxei64(const int32_t * op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_i32m2_m))) +vint32m2_t vluxei64(vbool16_t op0, vint32m2_t op1, const int32_t * op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_i32m4))) +vint32m4_t vluxei64(const int32_t * op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_i32m4_m))) +vint32m4_t vluxei64(vbool8_t op0, vint32m4_t op1, const int32_t * op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_i32mf2))) +vint32mf2_t vluxei64(const int32_t * op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_i32mf2_m))) +vint32mf2_t vluxei64(vbool64_t op0, vint32mf2_t op1, const int32_t * op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_u32m1))) +vuint32m1_t vluxei64(const uint32_t * op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_u32m1_m))) +vuint32m1_t vluxei64(vbool32_t op0, vuint32m1_t op1, const uint32_t * op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_u32m2))) +vuint32m2_t vluxei64(const uint32_t * op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_u32m2_m))) +vuint32m2_t vluxei64(vbool16_t op0, vuint32m2_t op1, const uint32_t * op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_u32m4))) +vuint32m4_t vluxei64(const uint32_t * op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_u32m4_m))) +vuint32m4_t vluxei64(vbool8_t op0, vuint32m4_t op1, const uint32_t * op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_u32mf2))) +vuint32mf2_t vluxei64(const uint32_t * op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_u32mf2_m))) +vuint32mf2_t vluxei64(vbool64_t op0, vuint32mf2_t op1, const uint32_t * op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_i64m1))) +vint64m1_t vluxei8(const int64_t * op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_i64m1_m))) +vint64m1_t vluxei8(vbool64_t op0, vint64m1_t op1, const int64_t * op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_i64m2))) +vint64m2_t vluxei8(const int64_t * op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_i64m2_m))) +vint64m2_t vluxei8(vbool32_t op0, vint64m2_t op1, const int64_t * op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_i64m4))) +vint64m4_t vluxei8(const int64_t * op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_i64m4_m))) +vint64m4_t vluxei8(vbool16_t op0, vint64m4_t op1, const int64_t * op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_i64m8))) +vint64m8_t vluxei8(const int64_t * op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_i64m8_m))) +vint64m8_t vluxei8(vbool8_t op0, vint64m8_t op1, const int64_t * op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_u64m1))) +vuint64m1_t vluxei8(const uint64_t * op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_u64m1_m))) +vuint64m1_t vluxei8(vbool64_t op0, vuint64m1_t op1, const uint64_t * op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_u64m2))) +vuint64m2_t vluxei8(const uint64_t * op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_u64m2_m))) +vuint64m2_t vluxei8(vbool32_t op0, vuint64m2_t op1, const uint64_t * op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_u64m4))) +vuint64m4_t vluxei8(const uint64_t * op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_u64m4_m))) +vuint64m4_t vluxei8(vbool16_t op0, vuint64m4_t op1, const uint64_t * op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_u64m8))) +vuint64m8_t vluxei8(const uint64_t * op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_u64m8_m))) +vuint64m8_t vluxei8(vbool8_t op0, vuint64m8_t op1, const uint64_t * op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_i64m1))) +vint64m1_t vluxei16(const int64_t * op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_i64m1_m))) +vint64m1_t vluxei16(vbool64_t op0, vint64m1_t op1, const int64_t * op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_i64m2))) +vint64m2_t vluxei16(const int64_t * op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_i64m2_m))) +vint64m2_t vluxei16(vbool32_t op0, vint64m2_t op1, const int64_t * op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_i64m4))) +vint64m4_t vluxei16(const int64_t * op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_i64m4_m))) +vint64m4_t vluxei16(vbool16_t op0, vint64m4_t op1, const int64_t * op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_i64m8))) +vint64m8_t vluxei16(const int64_t * op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_i64m8_m))) +vint64m8_t vluxei16(vbool8_t op0, vint64m8_t op1, const int64_t * op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_u64m1))) +vuint64m1_t vluxei16(const uint64_t * op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_u64m1_m))) +vuint64m1_t vluxei16(vbool64_t op0, vuint64m1_t op1, const uint64_t * op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_u64m2))) +vuint64m2_t vluxei16(const uint64_t * op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_u64m2_m))) +vuint64m2_t vluxei16(vbool32_t op0, vuint64m2_t op1, const uint64_t * op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_u64m4))) +vuint64m4_t vluxei16(const uint64_t * op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_u64m4_m))) +vuint64m4_t vluxei16(vbool16_t op0, vuint64m4_t op1, const uint64_t * op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_u64m8))) +vuint64m8_t vluxei16(const uint64_t * op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_u64m8_m))) +vuint64m8_t vluxei16(vbool8_t op0, vuint64m8_t op1, const uint64_t * op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_i64m1))) +vint64m1_t vluxei32(const int64_t * op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_i64m1_m))) +vint64m1_t vluxei32(vbool64_t op0, vint64m1_t op1, const int64_t * op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_i64m2))) +vint64m2_t vluxei32(const int64_t * op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_i64m2_m))) +vint64m2_t vluxei32(vbool32_t op0, vint64m2_t op1, const int64_t * op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_i64m4))) +vint64m4_t vluxei32(const int64_t * op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_i64m4_m))) +vint64m4_t vluxei32(vbool16_t op0, vint64m4_t op1, const int64_t * op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_i64m8))) +vint64m8_t vluxei32(const int64_t * op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_i64m8_m))) +vint64m8_t vluxei32(vbool8_t op0, vint64m8_t op1, const int64_t * op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_u64m1))) +vuint64m1_t vluxei32(const uint64_t * op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_u64m1_m))) +vuint64m1_t vluxei32(vbool64_t op0, vuint64m1_t op1, const uint64_t * op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_u64m2))) +vuint64m2_t vluxei32(const uint64_t * op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_u64m2_m))) +vuint64m2_t vluxei32(vbool32_t op0, vuint64m2_t op1, const uint64_t * op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_u64m4))) +vuint64m4_t vluxei32(const uint64_t * op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_u64m4_m))) +vuint64m4_t vluxei32(vbool16_t op0, vuint64m4_t op1, const uint64_t * op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_u64m8))) +vuint64m8_t vluxei32(const uint64_t * op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_u64m8_m))) +vuint64m8_t vluxei32(vbool8_t op0, vuint64m8_t op1, const uint64_t * op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_i64m1))) +vint64m1_t vluxei64(const int64_t * op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_i64m1_m))) +vint64m1_t vluxei64(vbool64_t op0, vint64m1_t op1, const int64_t * op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_i64m2))) +vint64m2_t vluxei64(const int64_t * op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_i64m2_m))) +vint64m2_t vluxei64(vbool32_t op0, vint64m2_t op1, const int64_t * op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_i64m4))) +vint64m4_t vluxei64(const int64_t * op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_i64m4_m))) +vint64m4_t vluxei64(vbool16_t op0, vint64m4_t op1, const int64_t * op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_i64m8))) +vint64m8_t vluxei64(const int64_t * op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_i64m8_m))) +vint64m8_t vluxei64(vbool8_t op0, vint64m8_t op1, const int64_t * op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_u64m1))) +vuint64m1_t vluxei64(const uint64_t * op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_u64m1_m))) +vuint64m1_t vluxei64(vbool64_t op0, vuint64m1_t op1, const uint64_t * op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_u64m2))) +vuint64m2_t vluxei64(const uint64_t * op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_u64m2_m))) +vuint64m2_t vluxei64(vbool32_t op0, vuint64m2_t op1, const uint64_t * op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_u64m4))) +vuint64m4_t vluxei64(const uint64_t * op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_u64m4_m))) +vuint64m4_t vluxei64(vbool16_t op0, vuint64m4_t op1, const uint64_t * op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_u64m8))) +vuint64m8_t vluxei64(const uint64_t * op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_u64m8_m))) +vuint64m8_t vluxei64(vbool8_t op0, vuint64m8_t op1, const uint64_t * op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_i8m1))) +vint8m1_t vloxei8(const int8_t * op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_i8m1_m))) +vint8m1_t vloxei8(vbool8_t op0, vint8m1_t op1, const int8_t * op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_i8m2))) +vint8m2_t vloxei8(const int8_t * op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_i8m2_m))) +vint8m2_t vloxei8(vbool4_t op0, vint8m2_t op1, const int8_t * op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_i8m4))) +vint8m4_t vloxei8(const int8_t * op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_i8m4_m))) +vint8m4_t vloxei8(vbool2_t op0, vint8m4_t op1, const int8_t * op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_i8m8))) +vint8m8_t vloxei8(const int8_t * op0, vuint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_i8m8_m))) +vint8m8_t vloxei8(vbool1_t op0, vint8m8_t op1, const int8_t * op2, vuint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_i8mf2))) +vint8mf2_t vloxei8(const int8_t * op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_i8mf2_m))) +vint8mf2_t vloxei8(vbool16_t op0, vint8mf2_t op1, const int8_t * op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_i8mf4))) +vint8mf4_t vloxei8(const int8_t * op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_i8mf4_m))) +vint8mf4_t vloxei8(vbool32_t op0, vint8mf4_t op1, const int8_t * op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_i8mf8))) +vint8mf8_t vloxei8(const int8_t * op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_i8mf8_m))) +vint8mf8_t vloxei8(vbool64_t op0, vint8mf8_t op1, const int8_t * op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_u8m1))) +vuint8m1_t vloxei8(const uint8_t * op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_u8m1_m))) +vuint8m1_t vloxei8(vbool8_t op0, vuint8m1_t op1, const uint8_t * op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_u8m2))) +vuint8m2_t vloxei8(const uint8_t * op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_u8m2_m))) +vuint8m2_t vloxei8(vbool4_t op0, vuint8m2_t op1, const uint8_t * op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_u8m4))) +vuint8m4_t vloxei8(const uint8_t * op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_u8m4_m))) +vuint8m4_t vloxei8(vbool2_t op0, vuint8m4_t op1, const uint8_t * op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_u8m8))) +vuint8m8_t vloxei8(const uint8_t * op0, vuint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_u8m8_m))) +vuint8m8_t vloxei8(vbool1_t op0, vuint8m8_t op1, const uint8_t * op2, vuint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_u8mf2))) +vuint8mf2_t vloxei8(const uint8_t * op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_u8mf2_m))) +vuint8mf2_t vloxei8(vbool16_t op0, vuint8mf2_t op1, const uint8_t * op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_u8mf4))) +vuint8mf4_t vloxei8(const uint8_t * op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_u8mf4_m))) +vuint8mf4_t vloxei8(vbool32_t op0, vuint8mf4_t op1, const uint8_t * op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_u8mf8))) +vuint8mf8_t vloxei8(const uint8_t * op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_u8mf8_m))) +vuint8mf8_t vloxei8(vbool64_t op0, vuint8mf8_t op1, const uint8_t * op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_i8m1))) +vint8m1_t vloxei16(const int8_t * op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_i8m1_m))) +vint8m1_t vloxei16(vbool8_t op0, vint8m1_t op1, const int8_t * op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_i8m2))) +vint8m2_t vloxei16(const int8_t * op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_i8m2_m))) +vint8m2_t vloxei16(vbool4_t op0, vint8m2_t op1, const int8_t * op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_i8m4))) +vint8m4_t vloxei16(const int8_t * op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_i8m4_m))) +vint8m4_t vloxei16(vbool2_t op0, vint8m4_t op1, const int8_t * op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_i8mf2))) +vint8mf2_t vloxei16(const int8_t * op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_i8mf2_m))) +vint8mf2_t vloxei16(vbool16_t op0, vint8mf2_t op1, const int8_t * op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_i8mf4))) +vint8mf4_t vloxei16(const int8_t * op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_i8mf4_m))) +vint8mf4_t vloxei16(vbool32_t op0, vint8mf4_t op1, const int8_t * op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_i8mf8))) +vint8mf8_t vloxei16(const int8_t * op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_i8mf8_m))) +vint8mf8_t vloxei16(vbool64_t op0, vint8mf8_t op1, const int8_t * op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_u8m1))) +vuint8m1_t vloxei16(const uint8_t * op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_u8m1_m))) +vuint8m1_t vloxei16(vbool8_t op0, vuint8m1_t op1, const uint8_t * op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_u8m2))) +vuint8m2_t vloxei16(const uint8_t * op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_u8m2_m))) +vuint8m2_t vloxei16(vbool4_t op0, vuint8m2_t op1, const uint8_t * op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_u8m4))) +vuint8m4_t vloxei16(const uint8_t * op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_u8m4_m))) +vuint8m4_t vloxei16(vbool2_t op0, vuint8m4_t op1, const uint8_t * op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_u8mf2))) +vuint8mf2_t vloxei16(const uint8_t * op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_u8mf2_m))) +vuint8mf2_t vloxei16(vbool16_t op0, vuint8mf2_t op1, const uint8_t * op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_u8mf4))) +vuint8mf4_t vloxei16(const uint8_t * op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_u8mf4_m))) +vuint8mf4_t vloxei16(vbool32_t op0, vuint8mf4_t op1, const uint8_t * op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_u8mf8))) +vuint8mf8_t vloxei16(const uint8_t * op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_u8mf8_m))) +vuint8mf8_t vloxei16(vbool64_t op0, vuint8mf8_t op1, const uint8_t * op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_i8m1))) +vint8m1_t vloxei32(const int8_t * op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_i8m1_m))) +vint8m1_t vloxei32(vbool8_t op0, vint8m1_t op1, const int8_t * op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_i8m2))) +vint8m2_t vloxei32(const int8_t * op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_i8m2_m))) +vint8m2_t vloxei32(vbool4_t op0, vint8m2_t op1, const int8_t * op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_i8mf2))) +vint8mf2_t vloxei32(const int8_t * op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_i8mf2_m))) +vint8mf2_t vloxei32(vbool16_t op0, vint8mf2_t op1, const int8_t * op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_i8mf4))) +vint8mf4_t vloxei32(const int8_t * op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_i8mf4_m))) +vint8mf4_t vloxei32(vbool32_t op0, vint8mf4_t op1, const int8_t * op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_i8mf8))) +vint8mf8_t vloxei32(const int8_t * op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_i8mf8_m))) +vint8mf8_t vloxei32(vbool64_t op0, vint8mf8_t op1, const int8_t * op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_u8m1))) +vuint8m1_t vloxei32(const uint8_t * op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_u8m1_m))) +vuint8m1_t vloxei32(vbool8_t op0, vuint8m1_t op1, const uint8_t * op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_u8m2))) +vuint8m2_t vloxei32(const uint8_t * op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_u8m2_m))) +vuint8m2_t vloxei32(vbool4_t op0, vuint8m2_t op1, const uint8_t * op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_u8mf2))) +vuint8mf2_t vloxei32(const uint8_t * op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_u8mf2_m))) +vuint8mf2_t vloxei32(vbool16_t op0, vuint8mf2_t op1, const uint8_t * op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_u8mf4))) +vuint8mf4_t vloxei32(const uint8_t * op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_u8mf4_m))) +vuint8mf4_t vloxei32(vbool32_t op0, vuint8mf4_t op1, const uint8_t * op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_u8mf8))) +vuint8mf8_t vloxei32(const uint8_t * op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_u8mf8_m))) +vuint8mf8_t vloxei32(vbool64_t op0, vuint8mf8_t op1, const uint8_t * op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_i8m1))) +vint8m1_t vloxei64(const int8_t * op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_i8m1_m))) +vint8m1_t vloxei64(vbool8_t op0, vint8m1_t op1, const int8_t * op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_i8mf2))) +vint8mf2_t vloxei64(const int8_t * op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_i8mf2_m))) +vint8mf2_t vloxei64(vbool16_t op0, vint8mf2_t op1, const int8_t * op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_i8mf4))) +vint8mf4_t vloxei64(const int8_t * op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_i8mf4_m))) +vint8mf4_t vloxei64(vbool32_t op0, vint8mf4_t op1, const int8_t * op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_i8mf8))) +vint8mf8_t vloxei64(const int8_t * op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_i8mf8_m))) +vint8mf8_t vloxei64(vbool64_t op0, vint8mf8_t op1, const int8_t * op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_u8m1))) +vuint8m1_t vloxei64(const uint8_t * op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_u8m1_m))) +vuint8m1_t vloxei64(vbool8_t op0, vuint8m1_t op1, const uint8_t * op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_u8mf2))) +vuint8mf2_t vloxei64(const uint8_t * op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_u8mf2_m))) +vuint8mf2_t vloxei64(vbool16_t op0, vuint8mf2_t op1, const uint8_t * op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_u8mf4))) +vuint8mf4_t vloxei64(const uint8_t * op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_u8mf4_m))) +vuint8mf4_t vloxei64(vbool32_t op0, vuint8mf4_t op1, const uint8_t * op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_u8mf8))) +vuint8mf8_t vloxei64(const uint8_t * op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_u8mf8_m))) +vuint8mf8_t vloxei64(vbool64_t op0, vuint8mf8_t op1, const uint8_t * op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_i16m1))) +vint16m1_t vloxei8(const int16_t * op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_i16m1_m))) +vint16m1_t vloxei8(vbool16_t op0, vint16m1_t op1, const int16_t * op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_i16m2))) +vint16m2_t vloxei8(const int16_t * op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_i16m2_m))) +vint16m2_t vloxei8(vbool8_t op0, vint16m2_t op1, const int16_t * op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_i16m4))) +vint16m4_t vloxei8(const int16_t * op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_i16m4_m))) +vint16m4_t vloxei8(vbool4_t op0, vint16m4_t op1, const int16_t * op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_i16m8))) +vint16m8_t vloxei8(const int16_t * op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_i16m8_m))) +vint16m8_t vloxei8(vbool2_t op0, vint16m8_t op1, const int16_t * op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_i16mf2))) +vint16mf2_t vloxei8(const int16_t * op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_i16mf2_m))) +vint16mf2_t vloxei8(vbool32_t op0, vint16mf2_t op1, const int16_t * op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_i16mf4))) +vint16mf4_t vloxei8(const int16_t * op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_i16mf4_m))) +vint16mf4_t vloxei8(vbool64_t op0, vint16mf4_t op1, const int16_t * op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_u16m1))) +vuint16m1_t vloxei8(const uint16_t * op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_u16m1_m))) +vuint16m1_t vloxei8(vbool16_t op0, vuint16m1_t op1, const uint16_t * op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_u16m2))) +vuint16m2_t vloxei8(const uint16_t * op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_u16m2_m))) +vuint16m2_t vloxei8(vbool8_t op0, vuint16m2_t op1, const uint16_t * op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_u16m4))) +vuint16m4_t vloxei8(const uint16_t * op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_u16m4_m))) +vuint16m4_t vloxei8(vbool4_t op0, vuint16m4_t op1, const uint16_t * op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_u16m8))) +vuint16m8_t vloxei8(const uint16_t * op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_u16m8_m))) +vuint16m8_t vloxei8(vbool2_t op0, vuint16m8_t op1, const uint16_t * op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_u16mf2))) +vuint16mf2_t vloxei8(const uint16_t * op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_u16mf2_m))) +vuint16mf2_t vloxei8(vbool32_t op0, vuint16mf2_t op1, const uint16_t * op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_u16mf4))) +vuint16mf4_t vloxei8(const uint16_t * op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_u16mf4_m))) +vuint16mf4_t vloxei8(vbool64_t op0, vuint16mf4_t op1, const uint16_t * op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_i16m1))) +vint16m1_t vloxei16(const int16_t * op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_i16m1_m))) +vint16m1_t vloxei16(vbool16_t op0, vint16m1_t op1, const int16_t * op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_i16m2))) +vint16m2_t vloxei16(const int16_t * op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_i16m2_m))) +vint16m2_t vloxei16(vbool8_t op0, vint16m2_t op1, const int16_t * op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_i16m4))) +vint16m4_t vloxei16(const int16_t * op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_i16m4_m))) +vint16m4_t vloxei16(vbool4_t op0, vint16m4_t op1, const int16_t * op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_i16m8))) +vint16m8_t vloxei16(const int16_t * op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_i16m8_m))) +vint16m8_t vloxei16(vbool2_t op0, vint16m8_t op1, const int16_t * op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_i16mf2))) +vint16mf2_t vloxei16(const int16_t * op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_i16mf2_m))) +vint16mf2_t vloxei16(vbool32_t op0, vint16mf2_t op1, const int16_t * op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_i16mf4))) +vint16mf4_t vloxei16(const int16_t * op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_i16mf4_m))) +vint16mf4_t vloxei16(vbool64_t op0, vint16mf4_t op1, const int16_t * op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_u16m1))) +vuint16m1_t vloxei16(const uint16_t * op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_u16m1_m))) +vuint16m1_t vloxei16(vbool16_t op0, vuint16m1_t op1, const uint16_t * op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_u16m2))) +vuint16m2_t vloxei16(const uint16_t * op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_u16m2_m))) +vuint16m2_t vloxei16(vbool8_t op0, vuint16m2_t op1, const uint16_t * op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_u16m4))) +vuint16m4_t vloxei16(const uint16_t * op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_u16m4_m))) +vuint16m4_t vloxei16(vbool4_t op0, vuint16m4_t op1, const uint16_t * op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_u16m8))) +vuint16m8_t vloxei16(const uint16_t * op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_u16m8_m))) +vuint16m8_t vloxei16(vbool2_t op0, vuint16m8_t op1, const uint16_t * op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_u16mf2))) +vuint16mf2_t vloxei16(const uint16_t * op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_u16mf2_m))) +vuint16mf2_t vloxei16(vbool32_t op0, vuint16mf2_t op1, const uint16_t * op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_u16mf4))) +vuint16mf4_t vloxei16(const uint16_t * op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_u16mf4_m))) +vuint16mf4_t vloxei16(vbool64_t op0, vuint16mf4_t op1, const uint16_t * op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_i16m1))) +vint16m1_t vloxei32(const int16_t * op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_i16m1_m))) +vint16m1_t vloxei32(vbool16_t op0, vint16m1_t op1, const int16_t * op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_i16m2))) +vint16m2_t vloxei32(const int16_t * op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_i16m2_m))) +vint16m2_t vloxei32(vbool8_t op0, vint16m2_t op1, const int16_t * op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_i16m4))) +vint16m4_t vloxei32(const int16_t * op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_i16m4_m))) +vint16m4_t vloxei32(vbool4_t op0, vint16m4_t op1, const int16_t * op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_i16mf2))) +vint16mf2_t vloxei32(const int16_t * op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_i16mf2_m))) +vint16mf2_t vloxei32(vbool32_t op0, vint16mf2_t op1, const int16_t * op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_i16mf4))) +vint16mf4_t vloxei32(const int16_t * op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_i16mf4_m))) +vint16mf4_t vloxei32(vbool64_t op0, vint16mf4_t op1, const int16_t * op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_u16m1))) +vuint16m1_t vloxei32(const uint16_t * op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_u16m1_m))) +vuint16m1_t vloxei32(vbool16_t op0, vuint16m1_t op1, const uint16_t * op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_u16m2))) +vuint16m2_t vloxei32(const uint16_t * op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_u16m2_m))) +vuint16m2_t vloxei32(vbool8_t op0, vuint16m2_t op1, const uint16_t * op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_u16m4))) +vuint16m4_t vloxei32(const uint16_t * op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_u16m4_m))) +vuint16m4_t vloxei32(vbool4_t op0, vuint16m4_t op1, const uint16_t * op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_u16mf2))) +vuint16mf2_t vloxei32(const uint16_t * op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_u16mf2_m))) +vuint16mf2_t vloxei32(vbool32_t op0, vuint16mf2_t op1, const uint16_t * op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_u16mf4))) +vuint16mf4_t vloxei32(const uint16_t * op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_u16mf4_m))) +vuint16mf4_t vloxei32(vbool64_t op0, vuint16mf4_t op1, const uint16_t * op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_i16m1))) +vint16m1_t vloxei64(const int16_t * op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_i16m1_m))) +vint16m1_t vloxei64(vbool16_t op0, vint16m1_t op1, const int16_t * op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_i16m2))) +vint16m2_t vloxei64(const int16_t * op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_i16m2_m))) +vint16m2_t vloxei64(vbool8_t op0, vint16m2_t op1, const int16_t * op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_i16mf2))) +vint16mf2_t vloxei64(const int16_t * op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_i16mf2_m))) +vint16mf2_t vloxei64(vbool32_t op0, vint16mf2_t op1, const int16_t * op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_i16mf4))) +vint16mf4_t vloxei64(const int16_t * op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_i16mf4_m))) +vint16mf4_t vloxei64(vbool64_t op0, vint16mf4_t op1, const int16_t * op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_u16m1))) +vuint16m1_t vloxei64(const uint16_t * op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_u16m1_m))) +vuint16m1_t vloxei64(vbool16_t op0, vuint16m1_t op1, const uint16_t * op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_u16m2))) +vuint16m2_t vloxei64(const uint16_t * op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_u16m2_m))) +vuint16m2_t vloxei64(vbool8_t op0, vuint16m2_t op1, const uint16_t * op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_u16mf2))) +vuint16mf2_t vloxei64(const uint16_t * op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_u16mf2_m))) +vuint16mf2_t vloxei64(vbool32_t op0, vuint16mf2_t op1, const uint16_t * op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_u16mf4))) +vuint16mf4_t vloxei64(const uint16_t * op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_u16mf4_m))) +vuint16mf4_t vloxei64(vbool64_t op0, vuint16mf4_t op1, const uint16_t * op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_i32m1))) +vint32m1_t vloxei8(const int32_t * op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_i32m1_m))) +vint32m1_t vloxei8(vbool32_t op0, vint32m1_t op1, const int32_t * op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_i32m2))) +vint32m2_t vloxei8(const int32_t * op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_i32m2_m))) +vint32m2_t vloxei8(vbool16_t op0, vint32m2_t op1, const int32_t * op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_i32m4))) +vint32m4_t vloxei8(const int32_t * op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_i32m4_m))) +vint32m4_t vloxei8(vbool8_t op0, vint32m4_t op1, const int32_t * op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_i32m8))) +vint32m8_t vloxei8(const int32_t * op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_i32m8_m))) +vint32m8_t vloxei8(vbool4_t op0, vint32m8_t op1, const int32_t * op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_i32mf2))) +vint32mf2_t vloxei8(const int32_t * op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_i32mf2_m))) +vint32mf2_t vloxei8(vbool64_t op0, vint32mf2_t op1, const int32_t * op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_u32m1))) +vuint32m1_t vloxei8(const uint32_t * op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_u32m1_m))) +vuint32m1_t vloxei8(vbool32_t op0, vuint32m1_t op1, const uint32_t * op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_u32m2))) +vuint32m2_t vloxei8(const uint32_t * op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_u32m2_m))) +vuint32m2_t vloxei8(vbool16_t op0, vuint32m2_t op1, const uint32_t * op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_u32m4))) +vuint32m4_t vloxei8(const uint32_t * op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_u32m4_m))) +vuint32m4_t vloxei8(vbool8_t op0, vuint32m4_t op1, const uint32_t * op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_u32m8))) +vuint32m8_t vloxei8(const uint32_t * op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_u32m8_m))) +vuint32m8_t vloxei8(vbool4_t op0, vuint32m8_t op1, const uint32_t * op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_u32mf2))) +vuint32mf2_t vloxei8(const uint32_t * op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_u32mf2_m))) +vuint32mf2_t vloxei8(vbool64_t op0, vuint32mf2_t op1, const uint32_t * op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_i32m1))) +vint32m1_t vloxei16(const int32_t * op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_i32m1_m))) +vint32m1_t vloxei16(vbool32_t op0, vint32m1_t op1, const int32_t * op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_i32m2))) +vint32m2_t vloxei16(const int32_t * op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_i32m2_m))) +vint32m2_t vloxei16(vbool16_t op0, vint32m2_t op1, const int32_t * op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_i32m4))) +vint32m4_t vloxei16(const int32_t * op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_i32m4_m))) +vint32m4_t vloxei16(vbool8_t op0, vint32m4_t op1, const int32_t * op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_i32m8))) +vint32m8_t vloxei16(const int32_t * op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_i32m8_m))) +vint32m8_t vloxei16(vbool4_t op0, vint32m8_t op1, const int32_t * op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_i32mf2))) +vint32mf2_t vloxei16(const int32_t * op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_i32mf2_m))) +vint32mf2_t vloxei16(vbool64_t op0, vint32mf2_t op1, const int32_t * op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_u32m1))) +vuint32m1_t vloxei16(const uint32_t * op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_u32m1_m))) +vuint32m1_t vloxei16(vbool32_t op0, vuint32m1_t op1, const uint32_t * op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_u32m2))) +vuint32m2_t vloxei16(const uint32_t * op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_u32m2_m))) +vuint32m2_t vloxei16(vbool16_t op0, vuint32m2_t op1, const uint32_t * op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_u32m4))) +vuint32m4_t vloxei16(const uint32_t * op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_u32m4_m))) +vuint32m4_t vloxei16(vbool8_t op0, vuint32m4_t op1, const uint32_t * op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_u32m8))) +vuint32m8_t vloxei16(const uint32_t * op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_u32m8_m))) +vuint32m8_t vloxei16(vbool4_t op0, vuint32m8_t op1, const uint32_t * op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_u32mf2))) +vuint32mf2_t vloxei16(const uint32_t * op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_u32mf2_m))) +vuint32mf2_t vloxei16(vbool64_t op0, vuint32mf2_t op1, const uint32_t * op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_i32m1))) +vint32m1_t vloxei32(const int32_t * op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_i32m1_m))) +vint32m1_t vloxei32(vbool32_t op0, vint32m1_t op1, const int32_t * op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_i32m2))) +vint32m2_t vloxei32(const int32_t * op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_i32m2_m))) +vint32m2_t vloxei32(vbool16_t op0, vint32m2_t op1, const int32_t * op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_i32m4))) +vint32m4_t vloxei32(const int32_t * op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_i32m4_m))) +vint32m4_t vloxei32(vbool8_t op0, vint32m4_t op1, const int32_t * op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_i32m8))) +vint32m8_t vloxei32(const int32_t * op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_i32m8_m))) +vint32m8_t vloxei32(vbool4_t op0, vint32m8_t op1, const int32_t * op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_i32mf2))) +vint32mf2_t vloxei32(const int32_t * op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_i32mf2_m))) +vint32mf2_t vloxei32(vbool64_t op0, vint32mf2_t op1, const int32_t * op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_u32m1))) +vuint32m1_t vloxei32(const uint32_t * op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_u32m1_m))) +vuint32m1_t vloxei32(vbool32_t op0, vuint32m1_t op1, const uint32_t * op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_u32m2))) +vuint32m2_t vloxei32(const uint32_t * op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_u32m2_m))) +vuint32m2_t vloxei32(vbool16_t op0, vuint32m2_t op1, const uint32_t * op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_u32m4))) +vuint32m4_t vloxei32(const uint32_t * op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_u32m4_m))) +vuint32m4_t vloxei32(vbool8_t op0, vuint32m4_t op1, const uint32_t * op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_u32m8))) +vuint32m8_t vloxei32(const uint32_t * op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_u32m8_m))) +vuint32m8_t vloxei32(vbool4_t op0, vuint32m8_t op1, const uint32_t * op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_u32mf2))) +vuint32mf2_t vloxei32(const uint32_t * op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_u32mf2_m))) +vuint32mf2_t vloxei32(vbool64_t op0, vuint32mf2_t op1, const uint32_t * op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_i32m1))) +vint32m1_t vloxei64(const int32_t * op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_i32m1_m))) +vint32m1_t vloxei64(vbool32_t op0, vint32m1_t op1, const int32_t * op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_i32m2))) +vint32m2_t vloxei64(const int32_t * op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_i32m2_m))) +vint32m2_t vloxei64(vbool16_t op0, vint32m2_t op1, const int32_t * op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_i32m4))) +vint32m4_t vloxei64(const int32_t * op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_i32m4_m))) +vint32m4_t vloxei64(vbool8_t op0, vint32m4_t op1, const int32_t * op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_i32mf2))) +vint32mf2_t vloxei64(const int32_t * op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_i32mf2_m))) +vint32mf2_t vloxei64(vbool64_t op0, vint32mf2_t op1, const int32_t * op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_u32m1))) +vuint32m1_t vloxei64(const uint32_t * op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_u32m1_m))) +vuint32m1_t vloxei64(vbool32_t op0, vuint32m1_t op1, const uint32_t * op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_u32m2))) +vuint32m2_t vloxei64(const uint32_t * op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_u32m2_m))) +vuint32m2_t vloxei64(vbool16_t op0, vuint32m2_t op1, const uint32_t * op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_u32m4))) +vuint32m4_t vloxei64(const uint32_t * op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_u32m4_m))) +vuint32m4_t vloxei64(vbool8_t op0, vuint32m4_t op1, const uint32_t * op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_u32mf2))) +vuint32mf2_t vloxei64(const uint32_t * op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_u32mf2_m))) +vuint32mf2_t vloxei64(vbool64_t op0, vuint32mf2_t op1, const uint32_t * op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_i64m1))) +vint64m1_t vloxei8(const int64_t * op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_i64m1_m))) +vint64m1_t vloxei8(vbool64_t op0, vint64m1_t op1, const int64_t * op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_i64m2))) +vint64m2_t vloxei8(const int64_t * op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_i64m2_m))) +vint64m2_t vloxei8(vbool32_t op0, vint64m2_t op1, const int64_t * op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_i64m4))) +vint64m4_t vloxei8(const int64_t * op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_i64m4_m))) +vint64m4_t vloxei8(vbool16_t op0, vint64m4_t op1, const int64_t * op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_i64m8))) +vint64m8_t vloxei8(const int64_t * op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_i64m8_m))) +vint64m8_t vloxei8(vbool8_t op0, vint64m8_t op1, const int64_t * op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse8_v_i8m1))) +void vse8(int8_t * op0, vint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse8_v_i8m1_m))) +void vse8(vbool8_t op0, int8_t * op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse8_v_i8m2))) +void vse8(int8_t * op0, vint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse8_v_i8m2_m))) +void vse8(vbool4_t op0, int8_t * op1, vint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse8_v_i8m4))) +void vse8(int8_t * op0, vint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse8_v_i8m4_m))) +void vse8(vbool2_t op0, int8_t * op1, vint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse8_v_i8m8))) +void vse8(int8_t * op0, vint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse8_v_i8m8_m))) +void vse8(vbool1_t op0, int8_t * op1, vint8m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse8_v_i8mf2))) +void vse8(int8_t * op0, vint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse8_v_i8mf2_m))) +void vse8(vbool16_t op0, int8_t * op1, vint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse8_v_i8mf4))) +void vse8(int8_t * op0, vint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse8_v_i8mf4_m))) +void vse8(vbool32_t op0, int8_t * op1, vint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse8_v_i8mf8))) +void vse8(int8_t * op0, vint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse8_v_i8mf8_m))) +void vse8(vbool64_t op0, int8_t * op1, vint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_u64m1))) +vuint64m1_t vloxei8(const uint64_t * op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_u64m1_m))) +vuint64m1_t vloxei8(vbool64_t op0, vuint64m1_t op1, const uint64_t * op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_u64m2))) +vuint64m2_t vloxei8(const uint64_t * op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_u64m2_m))) +vuint64m2_t vloxei8(vbool32_t op0, vuint64m2_t op1, const uint64_t * op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_u64m4))) +vuint64m4_t vloxei8(const uint64_t * op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_u64m4_m))) +vuint64m4_t vloxei8(vbool16_t op0, vuint64m4_t op1, const uint64_t * op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_u64m8))) +vuint64m8_t vloxei8(const uint64_t * op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_u64m8_m))) +vuint64m8_t vloxei8(vbool8_t op0, vuint64m8_t op1, const uint64_t * op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_i64m1))) +vint64m1_t vloxei16(const int64_t * op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_i64m1_m))) +vint64m1_t vloxei16(vbool64_t op0, vint64m1_t op1, const int64_t * op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_i64m2))) +vint64m2_t vloxei16(const int64_t * op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_i64m2_m))) +vint64m2_t vloxei16(vbool32_t op0, vint64m2_t op1, const int64_t * op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_i64m4))) +vint64m4_t vloxei16(const int64_t * op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_i64m4_m))) +vint64m4_t vloxei16(vbool16_t op0, vint64m4_t op1, const int64_t * op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_i64m8))) +vint64m8_t vloxei16(const int64_t * op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_i64m8_m))) +vint64m8_t vloxei16(vbool8_t op0, vint64m8_t op1, const int64_t * op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_u64m1))) +vuint64m1_t vloxei16(const uint64_t * op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_u64m1_m))) +vuint64m1_t vloxei16(vbool64_t op0, vuint64m1_t op1, const uint64_t * op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_u64m2))) +vuint64m2_t vloxei16(const uint64_t * op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_u64m2_m))) +vuint64m2_t vloxei16(vbool32_t op0, vuint64m2_t op1, const uint64_t * op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_u64m4))) +vuint64m4_t vloxei16(const uint64_t * op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_u64m4_m))) +vuint64m4_t vloxei16(vbool16_t op0, vuint64m4_t op1, const uint64_t * op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_u64m8))) +vuint64m8_t vloxei16(const uint64_t * op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_u64m8_m))) +vuint64m8_t vloxei16(vbool8_t op0, vuint64m8_t op1, const uint64_t * op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_i64m1))) +vint64m1_t vloxei32(const int64_t * op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_i64m1_m))) +vint64m1_t vloxei32(vbool64_t op0, vint64m1_t op1, const int64_t * op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_i64m2))) +vint64m2_t vloxei32(const int64_t * op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_i64m2_m))) +vint64m2_t vloxei32(vbool32_t op0, vint64m2_t op1, const int64_t * op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_i64m4))) +vint64m4_t vloxei32(const int64_t * op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_i64m4_m))) +vint64m4_t vloxei32(vbool16_t op0, vint64m4_t op1, const int64_t * op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_i64m8))) +vint64m8_t vloxei32(const int64_t * op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_i64m8_m))) +vint64m8_t vloxei32(vbool8_t op0, vint64m8_t op1, const int64_t * op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_u64m1))) +vuint64m1_t vloxei32(const uint64_t * op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_u64m1_m))) +vuint64m1_t vloxei32(vbool64_t op0, vuint64m1_t op1, const uint64_t * op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_u64m2))) +vuint64m2_t vloxei32(const uint64_t * op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_u64m2_m))) +vuint64m2_t vloxei32(vbool32_t op0, vuint64m2_t op1, const uint64_t * op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_u64m4))) +vuint64m4_t vloxei32(const uint64_t * op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_u64m4_m))) +vuint64m4_t vloxei32(vbool16_t op0, vuint64m4_t op1, const uint64_t * op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_u64m8))) +vuint64m8_t vloxei32(const uint64_t * op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_u64m8_m))) +vuint64m8_t vloxei32(vbool8_t op0, vuint64m8_t op1, const uint64_t * op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_i64m1))) +vint64m1_t vloxei64(const int64_t * op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_i64m1_m))) +vint64m1_t vloxei64(vbool64_t op0, vint64m1_t op1, const int64_t * op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_i64m2))) +vint64m2_t vloxei64(const int64_t * op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_i64m2_m))) +vint64m2_t vloxei64(vbool32_t op0, vint64m2_t op1, const int64_t * op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_i64m4))) +vint64m4_t vloxei64(const int64_t * op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_i64m4_m))) +vint64m4_t vloxei64(vbool16_t op0, vint64m4_t op1, const int64_t * op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_i64m8))) +vint64m8_t vloxei64(const int64_t * op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_i64m8_m))) +vint64m8_t vloxei64(vbool8_t op0, vint64m8_t op1, const int64_t * op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_u64m1))) +vuint64m1_t vloxei64(const uint64_t * op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_u64m1_m))) +vuint64m1_t vloxei64(vbool64_t op0, vuint64m1_t op1, const uint64_t * op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_u64m2))) +vuint64m2_t vloxei64(const uint64_t * op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_u64m2_m))) +vuint64m2_t vloxei64(vbool32_t op0, vuint64m2_t op1, const uint64_t * op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_u64m4))) +vuint64m4_t vloxei64(const uint64_t * op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_u64m4_m))) +vuint64m4_t vloxei64(vbool16_t op0, vuint64m4_t op1, const uint64_t * op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_u64m8))) +vuint64m8_t vloxei64(const uint64_t * op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_u64m8_m))) +vuint64m8_t vloxei64(vbool8_t op0, vuint64m8_t op1, const uint64_t * op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse8_v_u8m1))) +void vse8(uint8_t * op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse8_v_u8m1_m))) +void vse8(vbool8_t op0, uint8_t * op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse8_v_u8m2))) +void vse8(uint8_t * op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse8_v_u8m2_m))) +void vse8(vbool4_t op0, uint8_t * op1, vuint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse8_v_u8m4))) +void vse8(uint8_t * op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse8_v_u8m4_m))) +void vse8(vbool2_t op0, uint8_t * op1, vuint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse8_v_u8m8))) +void vse8(uint8_t * op0, vuint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse8_v_u8m8_m))) +void vse8(vbool1_t op0, uint8_t * op1, vuint8m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse8_v_u8mf2))) +void vse8(uint8_t * op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse8_v_u8mf2_m))) +void vse8(vbool16_t op0, uint8_t * op1, vuint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse8_v_u8mf4))) +void vse8(uint8_t * op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse8_v_u8mf4_m))) +void vse8(vbool32_t op0, uint8_t * op1, vuint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse8_v_u8mf8))) +void vse8(uint8_t * op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse8_v_u8mf8_m))) +void vse8(vbool64_t op0, uint8_t * op1, vuint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wv_u16mf4))) +vuint16mf4_t vwaddu_wv(vuint16mf4_t op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wv_u16mf4_m))) +vuint16mf4_t vwaddu_wv(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wv_u16mf2))) +vuint16mf2_t vwaddu_wv(vuint16mf2_t op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wv_u16mf2_m))) +vuint16mf2_t vwaddu_wv(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wv_u16m1))) +vuint16m1_t vwaddu_wv(vuint16m1_t op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wv_u16m1_m))) +vuint16m1_t vwaddu_wv(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wv_u16m2))) +vuint16m2_t vwaddu_wv(vuint16m2_t op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wv_u16m2_m))) +vuint16m2_t vwaddu_wv(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wv_u16m4))) +vuint16m4_t vwaddu_wv(vuint16m4_t op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wv_u16m4_m))) +vuint16m4_t vwaddu_wv(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wv_u16m8))) +vuint16m8_t vwaddu_wv(vuint16m8_t op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wv_u16m8_m))) +vuint16m8_t vwaddu_wv(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wv_u32mf2))) +vuint32mf2_t vwaddu_wv(vuint32mf2_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wv_u32mf2_m))) +vuint32mf2_t vwaddu_wv(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wv_u32m1))) +vuint32m1_t vwaddu_wv(vuint32m1_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wv_u32m1_m))) +vuint32m1_t vwaddu_wv(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wv_u32m2))) +vuint32m2_t vwaddu_wv(vuint32m2_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wv_u32m2_m))) +vuint32m2_t vwaddu_wv(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wv_u32m4))) +vuint32m4_t vwaddu_wv(vuint32m4_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wv_u32m4_m))) +vuint32m4_t vwaddu_wv(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wv_u32m8))) +vuint32m8_t vwaddu_wv(vuint32m8_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wv_u32m8_m))) +vuint32m8_t vwaddu_wv(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wv_u64m1))) +vuint64m1_t vwaddu_wv(vuint64m1_t op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wv_u64m1_m))) +vuint64m1_t vwaddu_wv(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wv_u64m2))) +vuint64m2_t vwaddu_wv(vuint64m2_t op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wv_u64m2_m))) +vuint64m2_t vwaddu_wv(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wv_u64m4))) +vuint64m4_t vwaddu_wv(vuint64m4_t op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wv_u64m4_m))) +vuint64m4_t vwaddu_wv(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wv_u64m8))) +vuint64m8_t vwaddu_wv(vuint64m8_t op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wv_u64m8_m))) +vuint64m8_t vwaddu_wv(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse8_v_i8m1))) +void vsse8(int8_t * op0, ptrdiff_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse8_v_i8m1_m))) +void vsse8(vbool8_t op0, int8_t * op1, ptrdiff_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse8_v_i8m2))) +void vsse8(int8_t * op0, ptrdiff_t op1, vint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse8_v_i8m2_m))) +void vsse8(vbool4_t op0, int8_t * op1, ptrdiff_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse8_v_i8m4))) +void vsse8(int8_t * op0, ptrdiff_t op1, vint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse8_v_i8m4_m))) +void vsse8(vbool2_t op0, int8_t * op1, ptrdiff_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse8_v_i8m8))) +void vsse8(int8_t * op0, ptrdiff_t op1, vint8m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse8_v_i8m8_m))) +void vsse8(vbool1_t op0, int8_t * op1, ptrdiff_t op2, vint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse8_v_i8mf2))) +void vsse8(int8_t * op0, ptrdiff_t op1, vint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse8_v_i8mf2_m))) +void vsse8(vbool16_t op0, int8_t * op1, ptrdiff_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse8_v_i8mf4))) +void vsse8(int8_t * op0, ptrdiff_t op1, vint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse8_v_i8mf4_m))) +void vsse8(vbool32_t op0, int8_t * op1, ptrdiff_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse8_v_i8mf8))) +void vsse8(int8_t * op0, ptrdiff_t op1, vint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse8_v_i8mf8_m))) +void vsse8(vbool64_t op0, int8_t * op1, ptrdiff_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_i8m1))) +void vsuxei16(int8_t * op0, vuint16m2_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_i8m1_m))) +void vsuxei16(vbool8_t op0, int8_t * op1, vuint16m2_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_i8m2))) +void vsuxei16(int8_t * op0, vuint16m4_t op1, vint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_i8m2_m))) +void vsuxei16(vbool4_t op0, int8_t * op1, vuint16m4_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_i8m4))) +void vsuxei16(int8_t * op0, vuint16m8_t op1, vint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_i8m4_m))) +void vsuxei16(vbool2_t op0, int8_t * op1, vuint16m8_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_i8mf2))) +void vsuxei16(int8_t * op0, vuint16m1_t op1, vint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_i8mf2_m))) +void vsuxei16(vbool16_t op0, int8_t * op1, vuint16m1_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_i8mf4))) +void vsuxei16(int8_t * op0, vuint16mf2_t op1, vint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_i8mf4_m))) +void vsuxei16(vbool32_t op0, int8_t * op1, vuint16mf2_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_i8mf8))) +void vsuxei16(int8_t * op0, vuint16mf4_t op1, vint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_i8mf8_m))) +void vsuxei16(vbool64_t op0, int8_t * op1, vuint16mf4_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_u8m1))) +void vsuxei16(uint8_t * op0, vuint16m2_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_u8m1_m))) +void vsuxei16(vbool8_t op0, uint8_t * op1, vuint16m2_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_u8m2))) +void vsuxei16(uint8_t * op0, vuint16m4_t op1, vuint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_u8m2_m))) +void vsuxei16(vbool4_t op0, uint8_t * op1, vuint16m4_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_u8m4))) +void vsuxei16(uint8_t * op0, vuint16m8_t op1, vuint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_u8m4_m))) +void vsuxei16(vbool2_t op0, uint8_t * op1, vuint16m8_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_u8mf2))) +void vsuxei16(uint8_t * op0, vuint16m1_t op1, vuint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_u8mf2_m))) +void vsuxei16(vbool16_t op0, uint8_t * op1, vuint16m1_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_u8mf4))) +void vsuxei16(uint8_t * op0, vuint16mf2_t op1, vuint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_u8mf4_m))) +void vsuxei16(vbool32_t op0, uint8_t * op1, vuint16mf2_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_u8mf8))) +void vsuxei16(uint8_t * op0, vuint16mf4_t op1, vuint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_u8mf8_m))) +void vsuxei16(vbool64_t op0, uint8_t * op1, vuint16mf4_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_i8m1))) +void vsuxei32(int8_t * op0, vuint32m4_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_i8m1_m))) +void vsuxei32(vbool8_t op0, int8_t * op1, vuint32m4_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_i8m2))) +void vsuxei32(int8_t * op0, vuint32m8_t op1, vint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_i8m2_m))) +void vsuxei32(vbool4_t op0, int8_t * op1, vuint32m8_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_i8mf2))) +void vsuxei32(int8_t * op0, vuint32m2_t op1, vint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_i8mf2_m))) +void vsuxei32(vbool16_t op0, int8_t * op1, vuint32m2_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_i8mf4))) +void vsuxei32(int8_t * op0, vuint32m1_t op1, vint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_i8mf4_m))) +void vsuxei32(vbool32_t op0, int8_t * op1, vuint32m1_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_i8mf8))) +void vsuxei32(int8_t * op0, vuint32mf2_t op1, vint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_i8mf8_m))) +void vsuxei32(vbool64_t op0, int8_t * op1, vuint32mf2_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_i8m1))) +vint8m1_t vadd(vint8m1_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_i8m1_m))) +vint8m1_t vadd(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_i8m2))) +vint8m2_t vadd(vint8m2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_i8m2_m))) +vint8m2_t vadd(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_i8m4))) +vint8m4_t vadd(vint8m4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_i8m4_m))) +vint8m4_t vadd(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_i8m8))) +vint8m8_t vadd(vint8m8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_i8m8_m))) +vint8m8_t vadd(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_i8mf2))) +vint8mf2_t vadd(vint8mf2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_i8mf2_m))) +vint8mf2_t vadd(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_i8mf4))) +vint8mf4_t vadd(vint8mf4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_i8mf4_m))) +vint8mf4_t vadd(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_i8mf8))) +vint8mf8_t vadd(vint8mf8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_i8mf8_m))) +vint8mf8_t vadd(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_i16m1))) +vint16m1_t vadd(vint16m1_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_i16m1_m))) +vint16m1_t vadd(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_i16m2))) +vint16m2_t vadd(vint16m2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_i16m2_m))) +vint16m2_t vadd(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_i16m4))) +vint16m4_t vadd(vint16m4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_i16m4_m))) +vint16m4_t vadd(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_i16m8))) +vint16m8_t vadd(vint16m8_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_i16m8_m))) +vint16m8_t vadd(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_i16mf2))) +vint16mf2_t vadd(vint16mf2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_i16mf2_m))) +vint16mf2_t vadd(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_i16mf4))) +vint16mf4_t vadd(vint16mf4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_i16mf4_m))) +vint16mf4_t vadd(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_i32m1))) +vint32m1_t vadd(vint32m1_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_i32m1_m))) +vint32m1_t vadd(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_i32m2))) +vint32m2_t vadd(vint32m2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_i32m2_m))) +vint32m2_t vadd(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_i32m4))) +vint32m4_t vadd(vint32m4_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_i32m4_m))) +vint32m4_t vadd(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_i32m8))) +vint32m8_t vadd(vint32m8_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_i32m8_m))) +vint32m8_t vadd(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_i32mf2))) +vint32mf2_t vadd(vint32mf2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_i32mf2_m))) +vint32mf2_t vadd(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_i64m1))) +vint64m1_t vadd(vint64m1_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_i64m1_m))) +vint64m1_t vadd(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_i64m2))) +vint64m2_t vadd(vint64m2_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_i64m2_m))) +vint64m2_t vadd(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_i64m4))) +vint64m4_t vadd(vint64m4_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_i64m4_m))) +vint64m4_t vadd(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_i64m8))) +vint64m8_t vadd(vint64m8_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_i64m8_m))) +vint64m8_t vadd(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_u8m1))) +vuint8m1_t vadd(vuint8m1_t op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_u8m1_m))) +vuint8m1_t vadd(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_u8m2))) +vuint8m2_t vadd(vuint8m2_t op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_u8m2_m))) +vuint8m2_t vadd(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_u8m4))) +vuint8m4_t vadd(vuint8m4_t op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_u8m4_m))) +vuint8m4_t vadd(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_u8m8))) +vuint8m8_t vadd(vuint8m8_t op0, vuint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_u8m8_m))) +vuint8m8_t vadd(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, vuint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_u8mf2))) +vuint8mf2_t vadd(vuint8mf2_t op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_u8mf2_m))) +vuint8mf2_t vadd(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_u8mf4))) +vuint8mf4_t vadd(vuint8mf4_t op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_u8mf4_m))) +vuint8mf4_t vadd(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_u8mf8))) +vuint8mf8_t vadd(vuint8mf8_t op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_u8mf8_m))) +vuint8mf8_t vadd(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_u16m1))) +vuint16m1_t vadd(vuint16m1_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_u16m1_m))) +vuint16m1_t vadd(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_u16m2))) +vuint16m2_t vadd(vuint16m2_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_u16m2_m))) +vuint16m2_t vadd(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_u16m4))) +vuint16m4_t vadd(vuint16m4_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_u16m4_m))) +vuint16m4_t vadd(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_u16m8))) +vuint16m8_t vadd(vuint16m8_t op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_u16m8_m))) +vuint16m8_t vadd(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_u16mf2))) +vuint16mf2_t vadd(vuint16mf2_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_u16mf2_m))) +vuint16mf2_t vadd(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_u16mf4))) +vuint16mf4_t vadd(vuint16mf4_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_u16mf4_m))) +vuint16mf4_t vadd(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_u32m1))) +vuint32m1_t vadd(vuint32m1_t op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_u32m1_m))) +vuint32m1_t vadd(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_u32m2))) +vuint32m2_t vadd(vuint32m2_t op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_u32m2_m))) +vuint32m2_t vadd(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_u32m4))) +vuint32m4_t vadd(vuint32m4_t op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_u32m4_m))) +vuint32m4_t vadd(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_u32m8))) +vuint32m8_t vadd(vuint32m8_t op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_u32m8_m))) +vuint32m8_t vadd(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_u32mf2))) +vuint32mf2_t vadd(vuint32mf2_t op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_u32mf2_m))) +vuint32mf2_t vadd(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_u64m1))) +vuint64m1_t vadd(vuint64m1_t op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_u64m1_m))) +vuint64m1_t vadd(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_u64m2))) +vuint64m2_t vadd(vuint64m2_t op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_u64m2_m))) +vuint64m2_t vadd(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_u64m4))) +vuint64m4_t vadd(vuint64m4_t op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_u64m4_m))) +vuint64m4_t vadd(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_u64m8))) +vuint64m8_t vadd(vuint64m8_t op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vv_u64m8_m))) +vuint64m8_t vadd(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_u8m1))) +vuint8m1_t vadd(vuint8m1_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_u8m1_m))) +vuint8m1_t vadd(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_u8m2))) +vuint8m2_t vadd(vuint8m2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_u8m2_m))) +vuint8m2_t vadd(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_u8m4))) +vuint8m4_t vadd(vuint8m4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_u8m4_m))) +vuint8m4_t vadd(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_u8m8))) +vuint8m8_t vadd(vuint8m8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_u8m8_m))) +vuint8m8_t vadd(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_u8mf2))) +vuint8mf2_t vadd(vuint8mf2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_u8mf2_m))) +vuint8mf2_t vadd(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_u8mf4))) +vuint8mf4_t vadd(vuint8mf4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_u8mf4_m))) +vuint8mf4_t vadd(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_u8mf8))) +vuint8mf8_t vadd(vuint8mf8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_u8mf8_m))) +vuint8mf8_t vadd(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_u16m1))) +vuint16m1_t vadd(vuint16m1_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_u16m1_m))) +vuint16m1_t vadd(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_u16m2))) +vuint16m2_t vadd(vuint16m2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_u16m2_m))) +vuint16m2_t vadd(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_u16m4))) +vuint16m4_t vadd(vuint16m4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_u16m4_m))) +vuint16m4_t vadd(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_u16m8))) +vuint16m8_t vadd(vuint16m8_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_u16m8_m))) +vuint16m8_t vadd(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_u16mf2))) +vuint16mf2_t vadd(vuint16mf2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_u16mf2_m))) +vuint16mf2_t vadd(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_u16mf4))) +vuint16mf4_t vadd(vuint16mf4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_u16mf4_m))) +vuint16mf4_t vadd(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_u32m1))) +vuint32m1_t vadd(vuint32m1_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_u32m1_m))) +vuint32m1_t vadd(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_u32m2))) +vuint32m2_t vadd(vuint32m2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_u32m2_m))) +vuint32m2_t vadd(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_u32m4))) +vuint32m4_t vadd(vuint32m4_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_u32m4_m))) +vuint32m4_t vadd(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_u32m8))) +vuint32m8_t vadd(vuint32m8_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_u32m8_m))) +vuint32m8_t vadd(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_u32mf2))) +vuint32mf2_t vadd(vuint32mf2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_u32mf2_m))) +vuint32mf2_t vadd(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_u64m1))) +vuint64m1_t vadd(vuint64m1_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_u64m1_m))) +vuint64m1_t vadd(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_u64m2))) +vuint64m2_t vadd(vuint64m2_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_u64m2_m))) +vuint64m2_t vadd(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_u64m4))) +vuint64m4_t vadd(vuint64m4_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_u64m4_m))) +vuint64m4_t vadd(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_u64m8))) +vuint64m8_t vadd(vuint64m8_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadd_vx_u64m8_m))) +vuint64m8_t vadd(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_i8m1))) +vint8m1_t vsub(vint8m1_t op0, vint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_i8m1_m))) +vint8m1_t vsub(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_i8m2))) +vint8m2_t vsub(vint8m2_t op0, vint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_i8m2_m))) +vint8m2_t vsub(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_i8m4))) +vint8m4_t vsub(vint8m4_t op0, vint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_i8m4_m))) +vint8m4_t vsub(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_i8m8))) +vint8m8_t vsub(vint8m8_t op0, vint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_i8m8_m))) +vint8m8_t vsub(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, vint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_i8mf2))) +vint8mf2_t vsub(vint8mf2_t op0, vint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_i8mf2_m))) +vint8mf2_t vsub(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_i8mf4))) +vint8mf4_t vsub(vint8mf4_t op0, vint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_i8mf4_m))) +vint8mf4_t vsub(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_i8mf8))) +vint8mf8_t vsub(vint8mf8_t op0, vint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_i8mf8_m))) +vint8mf8_t vsub(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_i16m1))) +vint16m1_t vsub(vint16m1_t op0, vint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_i16m1_m))) +vint16m1_t vsub(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_i16m2))) +vint16m2_t vsub(vint16m2_t op0, vint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_i16m2_m))) +vint16m2_t vsub(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_i16m4))) +vint16m4_t vsub(vint16m4_t op0, vint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_i16m4_m))) +vint16m4_t vsub(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_i16m8))) +vint16m8_t vsub(vint16m8_t op0, vint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_i16m8_m))) +vint16m8_t vsub(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, vint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_i16mf2))) +vint16mf2_t vsub(vint16mf2_t op0, vint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_i16mf2_m))) +vint16mf2_t vsub(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_i16mf4))) +vint16mf4_t vsub(vint16mf4_t op0, vint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_i16mf4_m))) +vint16mf4_t vsub(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_i32m1))) +vint32m1_t vsub(vint32m1_t op0, vint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_i32m1_m))) +vint32m1_t vsub(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_i32m2))) +vint32m2_t vsub(vint32m2_t op0, vint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_i32m2_m))) +vint32m2_t vsub(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_i32m4))) +vint32m4_t vsub(vint32m4_t op0, vint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_i32m4_m))) +vint32m4_t vsub(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_i32m8))) +vint32m8_t vsub(vint32m8_t op0, vint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_i32m8_m))) +vint32m8_t vsub(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_i32mf2))) +vint32mf2_t vsub(vint32mf2_t op0, vint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_i32mf2_m))) +vint32mf2_t vsub(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_i64m1))) +vint64m1_t vsub(vint64m1_t op0, vint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_i64m1_m))) +vint64m1_t vsub(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_i64m2))) +vint64m2_t vsub(vint64m2_t op0, vint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_i64m2_m))) +vint64m2_t vsub(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_i64m4))) +vint64m4_t vsub(vint64m4_t op0, vint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_i64m4_m))) +vint64m4_t vsub(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_i64m8))) +vint64m8_t vsub(vint64m8_t op0, vint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_i64m8_m))) +vint64m8_t vsub(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_u8m1))) +void vsuxei32(uint8_t * op0, vuint32m4_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_u8m1_m))) +void vsuxei32(vbool8_t op0, uint8_t * op1, vuint32m4_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_u8m2))) +void vsuxei32(uint8_t * op0, vuint32m8_t op1, vuint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_u8m2_m))) +void vsuxei32(vbool4_t op0, uint8_t * op1, vuint32m8_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_u8mf2))) +void vsuxei32(uint8_t * op0, vuint32m2_t op1, vuint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_u8mf2_m))) +void vsuxei32(vbool16_t op0, uint8_t * op1, vuint32m2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_u8mf4))) +void vsuxei32(uint8_t * op0, vuint32m1_t op1, vuint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_u8mf4_m))) +void vsuxei32(vbool32_t op0, uint8_t * op1, vuint32m1_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_u8mf8))) +void vsuxei32(uint8_t * op0, vuint32mf2_t op1, vuint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_u8mf8_m))) +void vsuxei32(vbool64_t op0, uint8_t * op1, vuint32mf2_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_i8m1))) +vint8m1_t vsub(vint8m1_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_i8m1_m))) +vint8m1_t vsub(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_i8m2))) +vint8m2_t vsub(vint8m2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_i8m2_m))) +vint8m2_t vsub(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_i8m4))) +vint8m4_t vsub(vint8m4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_i8m4_m))) +vint8m4_t vsub(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_i8m8))) +vint8m8_t vsub(vint8m8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_i8m8_m))) +vint8m8_t vsub(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_i8mf2))) +vint8mf2_t vsub(vint8mf2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_i8mf2_m))) +vint8mf2_t vsub(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_i8mf4))) +vint8mf4_t vsub(vint8mf4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_i8mf4_m))) +vint8mf4_t vsub(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_i8mf8))) +vint8mf8_t vsub(vint8mf8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_i8mf8_m))) +vint8mf8_t vsub(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_i16m1))) +vint16m1_t vsub(vint16m1_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_i16m1_m))) +vint16m1_t vsub(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_i16m2))) +vint16m2_t vsub(vint16m2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_i16m2_m))) +vint16m2_t vsub(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_i16m4))) +vint16m4_t vsub(vint16m4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_i16m4_m))) +vint16m4_t vsub(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_i16m8))) +vint16m8_t vsub(vint16m8_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_i16m8_m))) +vint16m8_t vsub(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_i16mf2))) +vint16mf2_t vsub(vint16mf2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_i16mf2_m))) +vint16mf2_t vsub(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_i16mf4))) +vint16mf4_t vsub(vint16mf4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_i16mf4_m))) +vint16mf4_t vsub(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_i32m1))) +vint32m1_t vsub(vint32m1_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_i32m1_m))) +vint32m1_t vsub(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_i32m2))) +vint32m2_t vsub(vint32m2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_i32m2_m))) +vint32m2_t vsub(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_i32m4))) +vint32m4_t vsub(vint32m4_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_i32m4_m))) +vint32m4_t vsub(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_i32m8))) +vint32m8_t vsub(vint32m8_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_i32m8_m))) +vint32m8_t vsub(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_i32mf2))) +vint32mf2_t vsub(vint32mf2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_i32mf2_m))) +vint32mf2_t vsub(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_i64m1))) +vint64m1_t vsub(vint64m1_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_i64m1_m))) +vint64m1_t vsub(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_i64m2))) +vint64m2_t vsub(vint64m2_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_i64m2_m))) +vint64m2_t vsub(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_i64m4))) +vint64m4_t vsub(vint64m4_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_i64m4_m))) +vint64m4_t vsub(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_i64m8))) +vint64m8_t vsub(vint64m8_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_i64m8_m))) +vint64m8_t vsub(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_u8m1))) +vuint8m1_t vsub(vuint8m1_t op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_u8m1_m))) +vuint8m1_t vsub(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_u8m2))) +vuint8m2_t vsub(vuint8m2_t op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_u8m2_m))) +vuint8m2_t vsub(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_u8m4))) +vuint8m4_t vsub(vuint8m4_t op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_u8m4_m))) +vuint8m4_t vsub(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_u8m8))) +vuint8m8_t vsub(vuint8m8_t op0, vuint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_u8m8_m))) +vuint8m8_t vsub(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, vuint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_u8mf2))) +vuint8mf2_t vsub(vuint8mf2_t op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_u8mf2_m))) +vuint8mf2_t vsub(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_u8mf4))) +vuint8mf4_t vsub(vuint8mf4_t op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_u8mf4_m))) +vuint8mf4_t vsub(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_u8mf8))) +vuint8mf8_t vsub(vuint8mf8_t op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_u8mf8_m))) +vuint8mf8_t vsub(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_u16m1))) +vuint16m1_t vsub(vuint16m1_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_u16m1_m))) +vuint16m1_t vsub(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_u16m2))) +vuint16m2_t vsub(vuint16m2_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_u16m2_m))) +vuint16m2_t vsub(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_u16m4))) +vuint16m4_t vsub(vuint16m4_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_u16m4_m))) +vuint16m4_t vsub(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_u16m8))) +vuint16m8_t vsub(vuint16m8_t op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_u16m8_m))) +vuint16m8_t vsub(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_u16mf2))) +vuint16mf2_t vsub(vuint16mf2_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_u16mf2_m))) +vuint16mf2_t vsub(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_u16mf4))) +vuint16mf4_t vsub(vuint16mf4_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_u16mf4_m))) +vuint16mf4_t vsub(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_u32m1))) +vuint32m1_t vsub(vuint32m1_t op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_u32m1_m))) +vuint32m1_t vsub(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_u32m2))) +vuint32m2_t vsub(vuint32m2_t op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_u32m2_m))) +vuint32m2_t vsub(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_u32m4))) +vuint32m4_t vsub(vuint32m4_t op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_u32m4_m))) +vuint32m4_t vsub(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_u32m8))) +vuint32m8_t vsub(vuint32m8_t op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_u32m8_m))) +vuint32m8_t vsub(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_u32mf2))) +vuint32mf2_t vsub(vuint32mf2_t op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_u32mf2_m))) +vuint32mf2_t vsub(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_u64m1))) +vuint64m1_t vsub(vuint64m1_t op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_u64m1_m))) +vuint64m1_t vsub(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_u64m2))) +vuint64m2_t vsub(vuint64m2_t op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_u64m2_m))) +vuint64m2_t vsub(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_u64m4))) +vuint64m4_t vsub(vuint64m4_t op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_u64m4_m))) +vuint64m4_t vsub(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_u64m8))) +vuint64m8_t vsub(vuint64m8_t op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vv_u64m8_m))) +vuint64m8_t vsub(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_u8m1))) +vuint8m1_t vsub(vuint8m1_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_u8m1_m))) +vuint8m1_t vsub(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_u8m2))) +vuint8m2_t vsub(vuint8m2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_u8m2_m))) +vuint8m2_t vsub(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_u8m4))) +vuint8m4_t vsub(vuint8m4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_u8m4_m))) +vuint8m4_t vsub(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_u8m8))) +vuint8m8_t vsub(vuint8m8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_u8m8_m))) +vuint8m8_t vsub(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_u8mf2))) +vuint8mf2_t vsub(vuint8mf2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_u8mf2_m))) +vuint8mf2_t vsub(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_u8mf4))) +vuint8mf4_t vsub(vuint8mf4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_u8mf4_m))) +vuint8mf4_t vsub(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_u8mf8))) +vuint8mf8_t vsub(vuint8mf8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_u8mf8_m))) +vuint8mf8_t vsub(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_u16m1))) +vuint16m1_t vsub(vuint16m1_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_u16m1_m))) +vuint16m1_t vsub(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_u16m2))) +vuint16m2_t vsub(vuint16m2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_u16m2_m))) +vuint16m2_t vsub(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_u16m4))) +vuint16m4_t vsub(vuint16m4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_u16m4_m))) +vuint16m4_t vsub(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_u16m8))) +vuint16m8_t vsub(vuint16m8_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_u16m8_m))) +vuint16m8_t vsub(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_u16mf2))) +vuint16mf2_t vsub(vuint16mf2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_u16mf2_m))) +vuint16mf2_t vsub(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_u16mf4))) +vuint16mf4_t vsub(vuint16mf4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_u16mf4_m))) +vuint16mf4_t vsub(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_u32m1))) +vuint32m1_t vsub(vuint32m1_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_u32m1_m))) +vuint32m1_t vsub(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_u32m2))) +vuint32m2_t vsub(vuint32m2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_u32m2_m))) +vuint32m2_t vsub(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_u32m4))) +vuint32m4_t vsub(vuint32m4_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_u32m4_m))) +vuint32m4_t vsub(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_u32m8))) +vuint32m8_t vsub(vuint32m8_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_u32m8_m))) +vuint32m8_t vsub(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_u32mf2))) +vuint32mf2_t vsub(vuint32mf2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_u32mf2_m))) +vuint32mf2_t vsub(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_u64m1))) +vuint64m1_t vsub(vuint64m1_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_u64m1_m))) +vuint64m1_t vsub(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_u64m2))) +vuint64m2_t vsub(vuint64m2_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_u64m2_m))) +vuint64m2_t vsub(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_u64m4))) +vuint64m4_t vsub(vuint64m4_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_u64m4_m))) +vuint64m4_t vsub(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_u64m8))) +vuint64m8_t vsub(vuint64m8_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsub_vx_u64m8_m))) +vuint64m8_t vsub(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_i8m1))) +vint8m1_t vrsub(vint8m1_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_i8m1_m))) +vint8m1_t vrsub(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_i8m2))) +vint8m2_t vrsub(vint8m2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_i8m2_m))) +vint8m2_t vrsub(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_i8m4))) +vint8m4_t vrsub(vint8m4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_i8m4_m))) +vint8m4_t vrsub(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_i8m8))) +vint8m8_t vrsub(vint8m8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_i8m8_m))) +vint8m8_t vrsub(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_i8mf2))) +vint8mf2_t vrsub(vint8mf2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_i8mf2_m))) +vint8mf2_t vrsub(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_i8mf4))) +vint8mf4_t vrsub(vint8mf4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_i8mf4_m))) +vint8mf4_t vrsub(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_i8mf8))) +vint8mf8_t vrsub(vint8mf8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_i8mf8_m))) +vint8mf8_t vrsub(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_i16m1))) +vint16m1_t vrsub(vint16m1_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_i16m1_m))) +vint16m1_t vrsub(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_i16m2))) +vint16m2_t vrsub(vint16m2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_i16m2_m))) +vint16m2_t vrsub(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_i16m4))) +vint16m4_t vrsub(vint16m4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_i16m4_m))) +vint16m4_t vrsub(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_i16m8))) +vint16m8_t vrsub(vint16m8_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_i16m8_m))) +vint16m8_t vrsub(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_i16mf2))) +vint16mf2_t vrsub(vint16mf2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_i16mf2_m))) +vint16mf2_t vrsub(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_i16mf4))) +vint16mf4_t vrsub(vint16mf4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_i16mf4_m))) +vint16mf4_t vrsub(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_i32m1))) +vint32m1_t vrsub(vint32m1_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_i32m1_m))) +vint32m1_t vrsub(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_i32m2))) +vint32m2_t vrsub(vint32m2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_i32m2_m))) +vint32m2_t vrsub(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_i32m4))) +vint32m4_t vrsub(vint32m4_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_i32m4_m))) +vint32m4_t vrsub(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_i32m8))) +vint32m8_t vrsub(vint32m8_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_i32m8_m))) +vint32m8_t vrsub(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_i32mf2))) +vint32mf2_t vrsub(vint32mf2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_i32mf2_m))) +vint32mf2_t vrsub(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_i64m1))) +vint64m1_t vrsub(vint64m1_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_i64m1_m))) +vint64m1_t vrsub(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_i64m2))) +vint64m2_t vrsub(vint64m2_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_i64m2_m))) +vint64m2_t vrsub(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_i64m4))) +vint64m4_t vrsub(vint64m4_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_i64m4_m))) +vint64m4_t vrsub(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_i64m8))) +vint64m8_t vrsub(vint64m8_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_i64m8_m))) +vint64m8_t vrsub(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_u8m1))) +vuint8m1_t vrsub(vuint8m1_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_u8m1_m))) +vuint8m1_t vrsub(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_u8m2))) +vuint8m2_t vrsub(vuint8m2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_u8m2_m))) +vuint8m2_t vrsub(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_u8m4))) +vuint8m4_t vrsub(vuint8m4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_u8m4_m))) +vuint8m4_t vrsub(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_u8m8))) +vuint8m8_t vrsub(vuint8m8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_u8m8_m))) +vuint8m8_t vrsub(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_u8mf2))) +vuint8mf2_t vrsub(vuint8mf2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_u8mf2_m))) +vuint8mf2_t vrsub(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_u8mf4))) +vuint8mf4_t vrsub(vuint8mf4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_u8mf4_m))) +vuint8mf4_t vrsub(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_u8mf8))) +vuint8mf8_t vrsub(vuint8mf8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_u8mf8_m))) +vuint8mf8_t vrsub(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_u16m1))) +vuint16m1_t vrsub(vuint16m1_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_u16m1_m))) +vuint16m1_t vrsub(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_u16m2))) +vuint16m2_t vrsub(vuint16m2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_u16m2_m))) +vuint16m2_t vrsub(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_u16m4))) +vuint16m4_t vrsub(vuint16m4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_u16m4_m))) +vuint16m4_t vrsub(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_u16m8))) +vuint16m8_t vrsub(vuint16m8_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_u16m8_m))) +vuint16m8_t vrsub(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_u16mf2))) +vuint16mf2_t vrsub(vuint16mf2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_u16mf2_m))) +vuint16mf2_t vrsub(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_u16mf4))) +vuint16mf4_t vrsub(vuint16mf4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_u16mf4_m))) +vuint16mf4_t vrsub(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_u32m1))) +vuint32m1_t vrsub(vuint32m1_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_u32m1_m))) +vuint32m1_t vrsub(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_u32m2))) +vuint32m2_t vrsub(vuint32m2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_u32m2_m))) +vuint32m2_t vrsub(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_u32m4))) +vuint32m4_t vrsub(vuint32m4_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_u32m4_m))) +vuint32m4_t vrsub(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_u32m8))) +vuint32m8_t vrsub(vuint32m8_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_u32m8_m))) +vuint32m8_t vrsub(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_u32mf2))) +vuint32mf2_t vrsub(vuint32mf2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_u32mf2_m))) +vuint32mf2_t vrsub(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_u64m1))) +vuint64m1_t vrsub(vuint64m1_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_u64m1_m))) +vuint64m1_t vrsub(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_u64m2))) +vuint64m2_t vrsub(vuint64m2_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_u64m2_m))) +vuint64m2_t vrsub(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_u64m4))) +vuint64m4_t vrsub(vuint64m4_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_u64m4_m))) +vuint64m4_t vrsub(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_u64m8))) +vuint64m8_t vrsub(vuint64m8_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrsub_vx_u64m8_m))) +vuint64m8_t vrsub(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vx_u16mf4))) +vuint16mf4_t vwaddu_vx(vuint8mf8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vx_u16mf4_m))) +vuint16mf4_t vwaddu_vx(vbool64_t op0, vuint16mf4_t op1, vuint8mf8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vx_u16mf2))) +vuint16mf2_t vwaddu_vx(vuint8mf4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vx_u16mf2_m))) +vuint16mf2_t vwaddu_vx(vbool32_t op0, vuint16mf2_t op1, vuint8mf4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vx_u16m1))) +vuint16m1_t vwaddu_vx(vuint8mf2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vx_u16m1_m))) +vuint16m1_t vwaddu_vx(vbool16_t op0, vuint16m1_t op1, vuint8mf2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vx_u16m2))) +vuint16m2_t vwaddu_vx(vuint8m1_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vx_u16m2_m))) +vuint16m2_t vwaddu_vx(vbool8_t op0, vuint16m2_t op1, vuint8m1_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vx_u16m4))) +vuint16m4_t vwaddu_vx(vuint8m2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vx_u16m4_m))) +vuint16m4_t vwaddu_vx(vbool4_t op0, vuint16m4_t op1, vuint8m2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vx_u16m8))) +vuint16m8_t vwaddu_vx(vuint8m4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vx_u16m8_m))) +vuint16m8_t vwaddu_vx(vbool2_t op0, vuint16m8_t op1, vuint8m4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vx_u32mf2))) +vuint32mf2_t vwaddu_vx(vuint16mf4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vx_u32mf2_m))) +vuint32mf2_t vwaddu_vx(vbool64_t op0, vuint32mf2_t op1, vuint16mf4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vx_u32m1))) +vuint32m1_t vwaddu_vx(vuint16mf2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vx_u32m1_m))) +vuint32m1_t vwaddu_vx(vbool32_t op0, vuint32m1_t op1, vuint16mf2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vx_u32m2))) +vuint32m2_t vwaddu_vx(vuint16m1_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vx_u32m2_m))) +vuint32m2_t vwaddu_vx(vbool16_t op0, vuint32m2_t op1, vuint16m1_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vx_u32m4))) +vuint32m4_t vwaddu_vx(vuint16m2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vx_u32m4_m))) +vuint32m4_t vwaddu_vx(vbool8_t op0, vuint32m4_t op1, vuint16m2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vx_u32m8))) +vuint32m8_t vwaddu_vx(vuint16m4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vx_u32m8_m))) +vuint32m8_t vwaddu_vx(vbool4_t op0, vuint32m8_t op1, vuint16m4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vx_u64m1))) +vuint64m1_t vwaddu_vx(vuint32mf2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vx_u64m1_m))) +vuint64m1_t vwaddu_vx(vbool64_t op0, vuint64m1_t op1, vuint32mf2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vx_u64m2))) +vuint64m2_t vwaddu_vx(vuint32m1_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vx_u64m2_m))) +vuint64m2_t vwaddu_vx(vbool32_t op0, vuint64m2_t op1, vuint32m1_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vx_u64m4))) +vuint64m4_t vwaddu_vx(vuint32m2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vx_u64m4_m))) +vuint64m4_t vwaddu_vx(vbool16_t op0, vuint64m4_t op1, vuint32m2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vx_u64m8))) +vuint64m8_t vwaddu_vx(vuint32m4_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_vx_u64m8_m))) +vuint64m8_t vwaddu_vx(vbool8_t op0, vuint64m8_t op1, vuint32m4_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vv_u16mf4))) +vuint16mf4_t vwsubu_vv(vuint8mf8_t op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vv_u16mf4_m))) +vuint16mf4_t vwsubu_vv(vbool64_t op0, vuint16mf4_t op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vv_u16mf2))) +vuint16mf2_t vwsubu_vv(vuint8mf4_t op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vv_u16mf2_m))) +vuint16mf2_t vwsubu_vv(vbool32_t op0, vuint16mf2_t op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vv_u16m1))) +vuint16m1_t vwsubu_vv(vuint8mf2_t op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vv_u16m1_m))) +vuint16m1_t vwsubu_vv(vbool16_t op0, vuint16m1_t op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vv_u16m2))) +vuint16m2_t vwsubu_vv(vuint8m1_t op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vv_u16m2_m))) +vuint16m2_t vwsubu_vv(vbool8_t op0, vuint16m2_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vv_u16m4))) +vuint16m4_t vwsubu_vv(vuint8m2_t op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vv_u16m4_m))) +vuint16m4_t vwsubu_vv(vbool4_t op0, vuint16m4_t op1, vuint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vv_u16m8))) +vuint16m8_t vwsubu_vv(vuint8m4_t op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vv_u16m8_m))) +vuint16m8_t vwsubu_vv(vbool2_t op0, vuint16m8_t op1, vuint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vv_u32mf2))) +vuint32mf2_t vwsubu_vv(vuint16mf4_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vv_u32mf2_m))) +vuint32mf2_t vwsubu_vv(vbool64_t op0, vuint32mf2_t op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vv_u32m1))) +vuint32m1_t vwsubu_vv(vuint16mf2_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vv_u32m1_m))) +vuint32m1_t vwsubu_vv(vbool32_t op0, vuint32m1_t op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vv_u32m2))) +vuint32m2_t vwsubu_vv(vuint16m1_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vv_u32m2_m))) +vuint32m2_t vwsubu_vv(vbool16_t op0, vuint32m2_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vv_u32m4))) +vuint32m4_t vwsubu_vv(vuint16m2_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vv_u32m4_m))) +vuint32m4_t vwsubu_vv(vbool8_t op0, vuint32m4_t op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vv_u32m8))) +vuint32m8_t vwsubu_vv(vuint16m4_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vv_u32m8_m))) +vuint32m8_t vwsubu_vv(vbool4_t op0, vuint32m8_t op1, vuint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vv_u64m1))) +vuint64m1_t vwsubu_vv(vuint32mf2_t op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vv_u64m1_m))) +vuint64m1_t vwsubu_vv(vbool64_t op0, vuint64m1_t op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vv_u64m2))) +vuint64m2_t vwsubu_vv(vuint32m1_t op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vv_u64m2_m))) +vuint64m2_t vwsubu_vv(vbool32_t op0, vuint64m2_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vv_u64m4))) +vuint64m4_t vwsubu_vv(vuint32m2_t op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vv_u64m4_m))) +vuint64m4_t vwsubu_vv(vbool16_t op0, vuint64m4_t op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vv_u64m8))) +vuint64m8_t vwsubu_vv(vuint32m4_t op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vv_u64m8_m))) +vuint64m8_t vwsubu_vv(vbool8_t op0, vuint64m8_t op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vx_u16mf4))) +vuint16mf4_t vwsubu_vx(vuint8mf8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vx_u16mf4_m))) +vuint16mf4_t vwsubu_vx(vbool64_t op0, vuint16mf4_t op1, vuint8mf8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vx_u16mf2))) +vuint16mf2_t vwsubu_vx(vuint8mf4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vx_u16mf2_m))) +vuint16mf2_t vwsubu_vx(vbool32_t op0, vuint16mf2_t op1, vuint8mf4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vx_u16m1))) +vuint16m1_t vwsubu_vx(vuint8mf2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vx_u16m1_m))) +vuint16m1_t vwsubu_vx(vbool16_t op0, vuint16m1_t op1, vuint8mf2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vx_u16m2))) +vuint16m2_t vwsubu_vx(vuint8m1_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vx_u16m2_m))) +vuint16m2_t vwsubu_vx(vbool8_t op0, vuint16m2_t op1, vuint8m1_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vx_u16m4))) +vuint16m4_t vwsubu_vx(vuint8m2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vx_u16m4_m))) +vuint16m4_t vwsubu_vx(vbool4_t op0, vuint16m4_t op1, vuint8m2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vx_u16m8))) +vuint16m8_t vwsubu_vx(vuint8m4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vx_u16m8_m))) +vuint16m8_t vwsubu_vx(vbool2_t op0, vuint16m8_t op1, vuint8m4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vx_u32mf2))) +vuint32mf2_t vwsubu_vx(vuint16mf4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vx_u32mf2_m))) +vuint32mf2_t vwsubu_vx(vbool64_t op0, vuint32mf2_t op1, vuint16mf4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vx_u32m1))) +vuint32m1_t vwsubu_vx(vuint16mf2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vx_u32m1_m))) +vuint32m1_t vwsubu_vx(vbool32_t op0, vuint32m1_t op1, vuint16mf2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vx_u32m2))) +vuint32m2_t vwsubu_vx(vuint16m1_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vx_u32m2_m))) +vuint32m2_t vwsubu_vx(vbool16_t op0, vuint32m2_t op1, vuint16m1_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vx_u32m4))) +vuint32m4_t vwsubu_vx(vuint16m2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vx_u32m4_m))) +vuint32m4_t vwsubu_vx(vbool8_t op0, vuint32m4_t op1, vuint16m2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vx_u32m8))) +vuint32m8_t vwsubu_vx(vuint16m4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vx_u32m8_m))) +vuint32m8_t vwsubu_vx(vbool4_t op0, vuint32m8_t op1, vuint16m4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vx_u64m1))) +vuint64m1_t vwsubu_vx(vuint32mf2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vx_u64m1_m))) +vuint64m1_t vwsubu_vx(vbool64_t op0, vuint64m1_t op1, vuint32mf2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vx_u64m2))) +vuint64m2_t vwsubu_vx(vuint32m1_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vx_u64m2_m))) +vuint64m2_t vwsubu_vx(vbool32_t op0, vuint64m2_t op1, vuint32m1_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vx_u64m4))) +vuint64m4_t vwsubu_vx(vuint32m2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vx_u64m4_m))) +vuint64m4_t vwsubu_vx(vbool16_t op0, vuint64m4_t op1, vuint32m2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vx_u64m8))) +vuint64m8_t vwsubu_vx(vuint32m4_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_vx_u64m8_m))) +vuint64m8_t vwsubu_vx(vbool8_t op0, vuint64m8_t op1, vuint32m4_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vv_i16mf4))) +vint16mf4_t vwadd_vv(vint8mf8_t op0, vint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vv_i16mf4_m))) +vint16mf4_t vwadd_vv(vbool64_t op0, vint16mf4_t op1, vint8mf8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vv_i16mf2))) +vint16mf2_t vwadd_vv(vint8mf4_t op0, vint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vv_i16mf2_m))) +vint16mf2_t vwadd_vv(vbool32_t op0, vint16mf2_t op1, vint8mf4_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vv_i16m1))) +vint16m1_t vwadd_vv(vint8mf2_t op0, vint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vv_i16m1_m))) +vint16m1_t vwadd_vv(vbool16_t op0, vint16m1_t op1, vint8mf2_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vv_i16m2))) +vint16m2_t vwadd_vv(vint8m1_t op0, vint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vv_i16m2_m))) +vint16m2_t vwadd_vv(vbool8_t op0, vint16m2_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vv_i16m4))) +vint16m4_t vwadd_vv(vint8m2_t op0, vint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vv_i16m4_m))) +vint16m4_t vwadd_vv(vbool4_t op0, vint16m4_t op1, vint8m2_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vv_i16m8))) +vint16m8_t vwadd_vv(vint8m4_t op0, vint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vv_i16m8_m))) +vint16m8_t vwadd_vv(vbool2_t op0, vint16m8_t op1, vint8m4_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vv_i32mf2))) +vint32mf2_t vwadd_vv(vint16mf4_t op0, vint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vv_i32mf2_m))) +vint32mf2_t vwadd_vv(vbool64_t op0, vint32mf2_t op1, vint16mf4_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vv_i32m1))) +vint32m1_t vwadd_vv(vint16mf2_t op0, vint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vv_i32m1_m))) +vint32m1_t vwadd_vv(vbool32_t op0, vint32m1_t op1, vint16mf2_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vv_i32m2))) +vint32m2_t vwadd_vv(vint16m1_t op0, vint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vv_i32m2_m))) +vint32m2_t vwadd_vv(vbool16_t op0, vint32m2_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vv_i32m4))) +vint32m4_t vwadd_vv(vint16m2_t op0, vint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vv_i32m4_m))) +vint32m4_t vwadd_vv(vbool8_t op0, vint32m4_t op1, vint16m2_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vv_i32m8))) +vint32m8_t vwadd_vv(vint16m4_t op0, vint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vv_i32m8_m))) +vint32m8_t vwadd_vv(vbool4_t op0, vint32m8_t op1, vint16m4_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vv_i64m1))) +vint64m1_t vwadd_vv(vint32mf2_t op0, vint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vv_i64m1_m))) +vint64m1_t vwadd_vv(vbool64_t op0, vint64m1_t op1, vint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vv_i64m2))) +vint64m2_t vwadd_vv(vint32m1_t op0, vint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vv_i64m2_m))) +vint64m2_t vwadd_vv(vbool32_t op0, vint64m2_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vv_i64m4))) +vint64m4_t vwadd_vv(vint32m2_t op0, vint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vv_i64m4_m))) +vint64m4_t vwadd_vv(vbool16_t op0, vint64m4_t op1, vint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vv_i64m8))) +vint64m8_t vwadd_vv(vint32m4_t op0, vint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vv_i64m8_m))) +vint64m8_t vwadd_vv(vbool8_t op0, vint64m8_t op1, vint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vx_i16mf4))) +vint16mf4_t vwadd_vx(vint8mf8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vx_i16mf4_m))) +vint16mf4_t vwadd_vx(vbool64_t op0, vint16mf4_t op1, vint8mf8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vx_i16mf2))) +vint16mf2_t vwadd_vx(vint8mf4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vx_i16mf2_m))) +vint16mf2_t vwadd_vx(vbool32_t op0, vint16mf2_t op1, vint8mf4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vx_i16m1))) +vint16m1_t vwadd_vx(vint8mf2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vx_i16m1_m))) +vint16m1_t vwadd_vx(vbool16_t op0, vint16m1_t op1, vint8mf2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vx_i16m2))) +vint16m2_t vwadd_vx(vint8m1_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vx_i16m2_m))) +vint16m2_t vwadd_vx(vbool8_t op0, vint16m2_t op1, vint8m1_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vx_i16m4))) +vint16m4_t vwadd_vx(vint8m2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vx_i16m4_m))) +vint16m4_t vwadd_vx(vbool4_t op0, vint16m4_t op1, vint8m2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vx_i16m8))) +vint16m8_t vwadd_vx(vint8m4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vx_i16m8_m))) +vint16m8_t vwadd_vx(vbool2_t op0, vint16m8_t op1, vint8m4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vx_i32mf2))) +vint32mf2_t vwadd_vx(vint16mf4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vx_i32mf2_m))) +vint32mf2_t vwadd_vx(vbool64_t op0, vint32mf2_t op1, vint16mf4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vx_i32m1))) +vint32m1_t vwadd_vx(vint16mf2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vx_i32m1_m))) +vint32m1_t vwadd_vx(vbool32_t op0, vint32m1_t op1, vint16mf2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vx_i32m2))) +vint32m2_t vwadd_vx(vint16m1_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vx_i32m2_m))) +vint32m2_t vwadd_vx(vbool16_t op0, vint32m2_t op1, vint16m1_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vx_i32m4))) +vint32m4_t vwadd_vx(vint16m2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vx_i32m4_m))) +vint32m4_t vwadd_vx(vbool8_t op0, vint32m4_t op1, vint16m2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vx_i32m8))) +vint32m8_t vwadd_vx(vint16m4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vx_i32m8_m))) +vint32m8_t vwadd_vx(vbool4_t op0, vint32m8_t op1, vint16m4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vx_i64m1))) +vint64m1_t vwadd_vx(vint32mf2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vx_i64m1_m))) +vint64m1_t vwadd_vx(vbool64_t op0, vint64m1_t op1, vint32mf2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vx_i64m2))) +vint64m2_t vwadd_vx(vint32m1_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vx_i64m2_m))) +vint64m2_t vwadd_vx(vbool32_t op0, vint64m2_t op1, vint32m1_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vx_i64m4))) +vint64m4_t vwadd_vx(vint32m2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vx_i64m4_m))) +vint64m4_t vwadd_vx(vbool16_t op0, vint64m4_t op1, vint32m2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vx_i64m8))) +vint64m8_t vwadd_vx(vint32m4_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_vx_i64m8_m))) +vint64m8_t vwadd_vx(vbool8_t op0, vint64m8_t op1, vint32m4_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_i8m1))) +void vsuxei64(int8_t * op0, vuint64m8_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_i8m1_m))) +void vsuxei64(vbool8_t op0, int8_t * op1, vuint64m8_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_i8mf2))) +void vsuxei64(int8_t * op0, vuint64m4_t op1, vint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_i8mf2_m))) +void vsuxei64(vbool16_t op0, int8_t * op1, vuint64m4_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_i8mf4))) +void vsuxei64(int8_t * op0, vuint64m2_t op1, vint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_i8mf4_m))) +void vsuxei64(vbool32_t op0, int8_t * op1, vuint64m2_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_i8mf8))) +void vsuxei64(int8_t * op0, vuint64m1_t op1, vint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_i8mf8_m))) +void vsuxei64(vbool64_t op0, int8_t * op1, vuint64m1_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vv_i16mf4))) +vint16mf4_t vwsub_vv(vint8mf8_t op0, vint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vv_i16mf4_m))) +vint16mf4_t vwsub_vv(vbool64_t op0, vint16mf4_t op1, vint8mf8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vv_i16mf2))) +vint16mf2_t vwsub_vv(vint8mf4_t op0, vint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vv_i16mf2_m))) +vint16mf2_t vwsub_vv(vbool32_t op0, vint16mf2_t op1, vint8mf4_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vv_i16m1))) +vint16m1_t vwsub_vv(vint8mf2_t op0, vint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vv_i16m1_m))) +vint16m1_t vwsub_vv(vbool16_t op0, vint16m1_t op1, vint8mf2_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vv_i16m2))) +vint16m2_t vwsub_vv(vint8m1_t op0, vint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vv_i16m2_m))) +vint16m2_t vwsub_vv(vbool8_t op0, vint16m2_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vv_i16m4))) +vint16m4_t vwsub_vv(vint8m2_t op0, vint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vv_i16m4_m))) +vint16m4_t vwsub_vv(vbool4_t op0, vint16m4_t op1, vint8m2_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vv_i16m8))) +vint16m8_t vwsub_vv(vint8m4_t op0, vint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vv_i16m8_m))) +vint16m8_t vwsub_vv(vbool2_t op0, vint16m8_t op1, vint8m4_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vv_i32mf2))) +vint32mf2_t vwsub_vv(vint16mf4_t op0, vint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vv_i32mf2_m))) +vint32mf2_t vwsub_vv(vbool64_t op0, vint32mf2_t op1, vint16mf4_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vv_i32m1))) +vint32m1_t vwsub_vv(vint16mf2_t op0, vint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vv_i32m1_m))) +vint32m1_t vwsub_vv(vbool32_t op0, vint32m1_t op1, vint16mf2_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vv_i32m2))) +vint32m2_t vwsub_vv(vint16m1_t op0, vint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vv_i32m2_m))) +vint32m2_t vwsub_vv(vbool16_t op0, vint32m2_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vv_i32m4))) +vint32m4_t vwsub_vv(vint16m2_t op0, vint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vv_i32m4_m))) +vint32m4_t vwsub_vv(vbool8_t op0, vint32m4_t op1, vint16m2_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vv_i32m8))) +vint32m8_t vwsub_vv(vint16m4_t op0, vint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vv_i32m8_m))) +vint32m8_t vwsub_vv(vbool4_t op0, vint32m8_t op1, vint16m4_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vv_i64m1))) +vint64m1_t vwsub_vv(vint32mf2_t op0, vint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vv_i64m1_m))) +vint64m1_t vwsub_vv(vbool64_t op0, vint64m1_t op1, vint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vv_i64m2))) +vint64m2_t vwsub_vv(vint32m1_t op0, vint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vv_i64m2_m))) +vint64m2_t vwsub_vv(vbool32_t op0, vint64m2_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vv_i64m4))) +vint64m4_t vwsub_vv(vint32m2_t op0, vint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vv_i64m4_m))) +vint64m4_t vwsub_vv(vbool16_t op0, vint64m4_t op1, vint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vv_i64m8))) +vint64m8_t vwsub_vv(vint32m4_t op0, vint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vv_i64m8_m))) +vint64m8_t vwsub_vv(vbool8_t op0, vint64m8_t op1, vint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vx_i16mf4))) +vint16mf4_t vwsub_vx(vint8mf8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vx_i16mf4_m))) +vint16mf4_t vwsub_vx(vbool64_t op0, vint16mf4_t op1, vint8mf8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vx_i16mf2))) +vint16mf2_t vwsub_vx(vint8mf4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vx_i16mf2_m))) +vint16mf2_t vwsub_vx(vbool32_t op0, vint16mf2_t op1, vint8mf4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vx_i16m1))) +vint16m1_t vwsub_vx(vint8mf2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vx_i16m1_m))) +vint16m1_t vwsub_vx(vbool16_t op0, vint16m1_t op1, vint8mf2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vx_i16m2))) +vint16m2_t vwsub_vx(vint8m1_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vx_i16m2_m))) +vint16m2_t vwsub_vx(vbool8_t op0, vint16m2_t op1, vint8m1_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vx_i16m4))) +vint16m4_t vwsub_vx(vint8m2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vx_i16m4_m))) +vint16m4_t vwsub_vx(vbool4_t op0, vint16m4_t op1, vint8m2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vx_i16m8))) +vint16m8_t vwsub_vx(vint8m4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vx_i16m8_m))) +vint16m8_t vwsub_vx(vbool2_t op0, vint16m8_t op1, vint8m4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vx_i32mf2))) +vint32mf2_t vwsub_vx(vint16mf4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vx_i32mf2_m))) +vint32mf2_t vwsub_vx(vbool64_t op0, vint32mf2_t op1, vint16mf4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vx_i32m1))) +vint32m1_t vwsub_vx(vint16mf2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vx_i32m1_m))) +vint32m1_t vwsub_vx(vbool32_t op0, vint32m1_t op1, vint16mf2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vx_i32m2))) +vint32m2_t vwsub_vx(vint16m1_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vx_i32m2_m))) +vint32m2_t vwsub_vx(vbool16_t op0, vint32m2_t op1, vint16m1_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vx_i32m4))) +vint32m4_t vwsub_vx(vint16m2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vx_i32m4_m))) +vint32m4_t vwsub_vx(vbool8_t op0, vint32m4_t op1, vint16m2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vx_i32m8))) +vint32m8_t vwsub_vx(vint16m4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vx_i32m8_m))) +vint32m8_t vwsub_vx(vbool4_t op0, vint32m8_t op1, vint16m4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vx_i64m1))) +vint64m1_t vwsub_vx(vint32mf2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vx_i64m1_m))) +vint64m1_t vwsub_vx(vbool64_t op0, vint64m1_t op1, vint32mf2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vx_i64m2))) +vint64m2_t vwsub_vx(vint32m1_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vx_i64m2_m))) +vint64m2_t vwsub_vx(vbool32_t op0, vint64m2_t op1, vint32m1_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vx_i64m4))) +vint64m4_t vwsub_vx(vint32m2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vx_i64m4_m))) +vint64m4_t vwsub_vx(vbool16_t op0, vint64m4_t op1, vint32m2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vx_i64m8))) +vint64m8_t vwsub_vx(vint32m4_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_vx_i64m8_m))) +vint64m8_t vwsub_vx(vbool8_t op0, vint64m8_t op1, vint32m4_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wx_u16mf4))) +vuint16mf4_t vwaddu_wx(vuint16mf4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wx_u16mf4_m))) +vuint16mf4_t vwaddu_wx(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wx_u16mf2))) +vuint16mf2_t vwaddu_wx(vuint16mf2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wx_u16mf2_m))) +vuint16mf2_t vwaddu_wx(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wx_u16m1))) +vuint16m1_t vwaddu_wx(vuint16m1_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wx_u16m1_m))) +vuint16m1_t vwaddu_wx(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wx_u16m2))) +vuint16m2_t vwaddu_wx(vuint16m2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wx_u16m2_m))) +vuint16m2_t vwaddu_wx(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wx_u16m4))) +vuint16m4_t vwaddu_wx(vuint16m4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wx_u16m4_m))) +vuint16m4_t vwaddu_wx(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wx_u16m8))) +vuint16m8_t vwaddu_wx(vuint16m8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wx_u16m8_m))) +vuint16m8_t vwaddu_wx(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wx_u32mf2))) +vuint32mf2_t vwaddu_wx(vuint32mf2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wx_u32mf2_m))) +vuint32mf2_t vwaddu_wx(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wx_u32m1))) +vuint32m1_t vwaddu_wx(vuint32m1_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wx_u32m1_m))) +vuint32m1_t vwaddu_wx(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wx_u32m2))) +vuint32m2_t vwaddu_wx(vuint32m2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wx_u32m2_m))) +vuint32m2_t vwaddu_wx(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wx_u32m4))) +vuint32m4_t vwaddu_wx(vuint32m4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wx_u32m4_m))) +vuint32m4_t vwaddu_wx(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wx_u32m8))) +vuint32m8_t vwaddu_wx(vuint32m8_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wx_u32m8_m))) +vuint32m8_t vwaddu_wx(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wx_u64m1))) +vuint64m1_t vwaddu_wx(vuint64m1_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wx_u64m1_m))) +vuint64m1_t vwaddu_wx(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wx_u64m2))) +vuint64m2_t vwaddu_wx(vuint64m2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wx_u64m2_m))) +vuint64m2_t vwaddu_wx(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wx_u64m4))) +vuint64m4_t vwaddu_wx(vuint64m4_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wx_u64m4_m))) +vuint64m4_t vwaddu_wx(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wx_u64m8))) +vuint64m8_t vwaddu_wx(vuint64m8_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwaddu_wx_u64m8_m))) +vuint64m8_t vwaddu_wx(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wv_u16mf4))) +vuint16mf4_t vwsubu_wv(vuint16mf4_t op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wv_u16mf4_m))) +vuint16mf4_t vwsubu_wv(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wv_u16mf2))) +vuint16mf2_t vwsubu_wv(vuint16mf2_t op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wv_u16mf2_m))) +vuint16mf2_t vwsubu_wv(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wv_u16m1))) +vuint16m1_t vwsubu_wv(vuint16m1_t op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wv_u16m1_m))) +vuint16m1_t vwsubu_wv(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wv_u16m2))) +vuint16m2_t vwsubu_wv(vuint16m2_t op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wv_u16m2_m))) +vuint16m2_t vwsubu_wv(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wv_u16m4))) +vuint16m4_t vwsubu_wv(vuint16m4_t op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wv_u16m4_m))) +vuint16m4_t vwsubu_wv(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wv_u16m8))) +vuint16m8_t vwsubu_wv(vuint16m8_t op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wv_u16m8_m))) +vuint16m8_t vwsubu_wv(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wv_u32mf2))) +vuint32mf2_t vwsubu_wv(vuint32mf2_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wv_u32mf2_m))) +vuint32mf2_t vwsubu_wv(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wv_u32m1))) +vuint32m1_t vwsubu_wv(vuint32m1_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wv_u32m1_m))) +vuint32m1_t vwsubu_wv(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wv_u32m2))) +vuint32m2_t vwsubu_wv(vuint32m2_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wv_u32m2_m))) +vuint32m2_t vwsubu_wv(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wv_u32m4))) +vuint32m4_t vwsubu_wv(vuint32m4_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wv_u32m4_m))) +vuint32m4_t vwsubu_wv(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wv_u32m8))) +vuint32m8_t vwsubu_wv(vuint32m8_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wv_u32m8_m))) +vuint32m8_t vwsubu_wv(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wv_u64m1))) +vuint64m1_t vwsubu_wv(vuint64m1_t op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wv_u64m1_m))) +vuint64m1_t vwsubu_wv(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wv_u64m2))) +vuint64m2_t vwsubu_wv(vuint64m2_t op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wv_u64m2_m))) +vuint64m2_t vwsubu_wv(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wv_u64m4))) +vuint64m4_t vwsubu_wv(vuint64m4_t op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wv_u64m4_m))) +vuint64m4_t vwsubu_wv(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wv_u64m8))) +vuint64m8_t vwsubu_wv(vuint64m8_t op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wv_u64m8_m))) +vuint64m8_t vwsubu_wv(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wx_u16mf4))) +vuint16mf4_t vwsubu_wx(vuint16mf4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wx_u16mf4_m))) +vuint16mf4_t vwsubu_wx(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wx_u16mf2))) +vuint16mf2_t vwsubu_wx(vuint16mf2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wx_u16mf2_m))) +vuint16mf2_t vwsubu_wx(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wx_u16m1))) +vuint16m1_t vwsubu_wx(vuint16m1_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wx_u16m1_m))) +vuint16m1_t vwsubu_wx(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wx_u16m2))) +vuint16m2_t vwsubu_wx(vuint16m2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wx_u16m2_m))) +vuint16m2_t vwsubu_wx(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wx_u16m4))) +vuint16m4_t vwsubu_wx(vuint16m4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wx_u16m4_m))) +vuint16m4_t vwsubu_wx(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wx_u16m8))) +vuint16m8_t vwsubu_wx(vuint16m8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wx_u16m8_m))) +vuint16m8_t vwsubu_wx(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wx_u32mf2))) +vuint32mf2_t vwsubu_wx(vuint32mf2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wx_u32mf2_m))) +vuint32mf2_t vwsubu_wx(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wx_u32m1))) +vuint32m1_t vwsubu_wx(vuint32m1_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wx_u32m1_m))) +vuint32m1_t vwsubu_wx(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wx_u32m2))) +vuint32m2_t vwsubu_wx(vuint32m2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wx_u32m2_m))) +vuint32m2_t vwsubu_wx(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wx_u32m4))) +vuint32m4_t vwsubu_wx(vuint32m4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wx_u32m4_m))) +vuint32m4_t vwsubu_wx(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wx_u32m8))) +vuint32m8_t vwsubu_wx(vuint32m8_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wx_u32m8_m))) +vuint32m8_t vwsubu_wx(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wx_u64m1))) +vuint64m1_t vwsubu_wx(vuint64m1_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wx_u64m1_m))) +vuint64m1_t vwsubu_wx(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wx_u64m2))) +vuint64m2_t vwsubu_wx(vuint64m2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wx_u64m2_m))) +vuint64m2_t vwsubu_wx(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wx_u64m4))) +vuint64m4_t vwsubu_wx(vuint64m4_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wx_u64m4_m))) +vuint64m4_t vwsubu_wx(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wx_u64m8))) +vuint64m8_t vwsubu_wx(vuint64m8_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsubu_wx_u64m8_m))) +vuint64m8_t vwsubu_wx(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wv_i16mf4))) +vint16mf4_t vwadd_wv(vint16mf4_t op0, vint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wv_i16mf4_m))) +vint16mf4_t vwadd_wv(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wv_i16mf2))) +vint16mf2_t vwadd_wv(vint16mf2_t op0, vint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wv_i16mf2_m))) +vint16mf2_t vwadd_wv(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wv_i16m1))) +vint16m1_t vwadd_wv(vint16m1_t op0, vint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wv_i16m1_m))) +vint16m1_t vwadd_wv(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wv_i16m2))) +vint16m2_t vwadd_wv(vint16m2_t op0, vint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wv_i16m2_m))) +vint16m2_t vwadd_wv(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wv_i16m4))) +vint16m4_t vwadd_wv(vint16m4_t op0, vint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wv_i16m4_m))) +vint16m4_t vwadd_wv(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wv_i16m8))) +vint16m8_t vwadd_wv(vint16m8_t op0, vint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wv_i16m8_m))) +vint16m8_t vwadd_wv(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wv_i32mf2))) +vint32mf2_t vwadd_wv(vint32mf2_t op0, vint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wv_i32mf2_m))) +vint32mf2_t vwadd_wv(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wv_i32m1))) +vint32m1_t vwadd_wv(vint32m1_t op0, vint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wv_i32m1_m))) +vint32m1_t vwadd_wv(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wv_i32m2))) +vint32m2_t vwadd_wv(vint32m2_t op0, vint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wv_i32m2_m))) +vint32m2_t vwadd_wv(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wv_i32m4))) +vint32m4_t vwadd_wv(vint32m4_t op0, vint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wv_i32m4_m))) +vint32m4_t vwadd_wv(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wv_i32m8))) +vint32m8_t vwadd_wv(vint32m8_t op0, vint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wv_i32m8_m))) +vint32m8_t vwadd_wv(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wv_i64m1))) +vint64m1_t vwadd_wv(vint64m1_t op0, vint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wv_i64m1_m))) +vint64m1_t vwadd_wv(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wv_i64m2))) +vint64m2_t vwadd_wv(vint64m2_t op0, vint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wv_i64m2_m))) +vint64m2_t vwadd_wv(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wv_i64m4))) +vint64m4_t vwadd_wv(vint64m4_t op0, vint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wv_i64m4_m))) +vint64m4_t vwadd_wv(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wv_i64m8))) +vint64m8_t vwadd_wv(vint64m8_t op0, vint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wv_i64m8_m))) +vint64m8_t vwadd_wv(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wx_i16mf4))) +vint16mf4_t vwadd_wx(vint16mf4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wx_i16mf4_m))) +vint16mf4_t vwadd_wx(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wx_i16mf2))) +vint16mf2_t vwadd_wx(vint16mf2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wx_i16mf2_m))) +vint16mf2_t vwadd_wx(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wx_i16m1))) +vint16m1_t vwadd_wx(vint16m1_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wx_i16m1_m))) +vint16m1_t vwadd_wx(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wx_i16m2))) +vint16m2_t vwadd_wx(vint16m2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wx_i16m2_m))) +vint16m2_t vwadd_wx(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wx_i16m4))) +vint16m4_t vwadd_wx(vint16m4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wx_i16m4_m))) +vint16m4_t vwadd_wx(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wx_i16m8))) +vint16m8_t vwadd_wx(vint16m8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wx_i16m8_m))) +vint16m8_t vwadd_wx(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wx_i32mf2))) +vint32mf2_t vwadd_wx(vint32mf2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wx_i32mf2_m))) +vint32mf2_t vwadd_wx(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wx_i32m1))) +vint32m1_t vwadd_wx(vint32m1_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wx_i32m1_m))) +vint32m1_t vwadd_wx(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wx_i32m2))) +vint32m2_t vwadd_wx(vint32m2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wx_i32m2_m))) +vint32m2_t vwadd_wx(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wx_i32m4))) +vint32m4_t vwadd_wx(vint32m4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wx_i32m4_m))) +vint32m4_t vwadd_wx(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wx_i32m8))) +vint32m8_t vwadd_wx(vint32m8_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wx_i32m8_m))) +vint32m8_t vwadd_wx(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wx_i64m1))) +vint64m1_t vwadd_wx(vint64m1_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wx_i64m1_m))) +vint64m1_t vwadd_wx(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wx_i64m2))) +vint64m2_t vwadd_wx(vint64m2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wx_i64m2_m))) +vint64m2_t vwadd_wx(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wx_i64m4))) +vint64m4_t vwadd_wx(vint64m4_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wx_i64m4_m))) +vint64m4_t vwadd_wx(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wx_i64m8))) +vint64m8_t vwadd_wx(vint64m8_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwadd_wx_i64m8_m))) +vint64m8_t vwadd_wx(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wv_i16mf4))) +vint16mf4_t vwsub_wv(vint16mf4_t op0, vint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wv_i16mf4_m))) +vint16mf4_t vwsub_wv(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wv_i16mf2))) +vint16mf2_t vwsub_wv(vint16mf2_t op0, vint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wv_i16mf2_m))) +vint16mf2_t vwsub_wv(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wv_i16m1))) +vint16m1_t vwsub_wv(vint16m1_t op0, vint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wv_i16m1_m))) +vint16m1_t vwsub_wv(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wv_i16m2))) +vint16m2_t vwsub_wv(vint16m2_t op0, vint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wv_i16m2_m))) +vint16m2_t vwsub_wv(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wv_i16m4))) +vint16m4_t vwsub_wv(vint16m4_t op0, vint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wv_i16m4_m))) +vint16m4_t vwsub_wv(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wv_i16m8))) +vint16m8_t vwsub_wv(vint16m8_t op0, vint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wv_i16m8_m))) +vint16m8_t vwsub_wv(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wv_i32mf2))) +vint32mf2_t vwsub_wv(vint32mf2_t op0, vint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wv_i32mf2_m))) +vint32mf2_t vwsub_wv(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wv_i32m1))) +vint32m1_t vwsub_wv(vint32m1_t op0, vint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wv_i32m1_m))) +vint32m1_t vwsub_wv(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wv_i32m2))) +vint32m2_t vwsub_wv(vint32m2_t op0, vint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wv_i32m2_m))) +vint32m2_t vwsub_wv(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wv_i32m4))) +vint32m4_t vwsub_wv(vint32m4_t op0, vint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wv_i32m4_m))) +vint32m4_t vwsub_wv(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wv_i32m8))) +vint32m8_t vwsub_wv(vint32m8_t op0, vint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wv_i32m8_m))) +vint32m8_t vwsub_wv(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wv_i64m1))) +vint64m1_t vwsub_wv(vint64m1_t op0, vint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wv_i64m1_m))) +vint64m1_t vwsub_wv(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wv_i64m2))) +vint64m2_t vwsub_wv(vint64m2_t op0, vint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wv_i64m2_m))) +vint64m2_t vwsub_wv(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wv_i64m4))) +vint64m4_t vwsub_wv(vint64m4_t op0, vint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wv_i64m4_m))) +vint64m4_t vwsub_wv(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wv_i64m8))) +vint64m8_t vwsub_wv(vint64m8_t op0, vint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wv_i64m8_m))) +vint64m8_t vwsub_wv(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wx_i16mf4))) +vint16mf4_t vwsub_wx(vint16mf4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wx_i16mf4_m))) +vint16mf4_t vwsub_wx(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wx_i16mf2))) +vint16mf2_t vwsub_wx(vint16mf2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wx_i16mf2_m))) +vint16mf2_t vwsub_wx(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wx_i16m1))) +vint16m1_t vwsub_wx(vint16m1_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wx_i16m1_m))) +vint16m1_t vwsub_wx(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wx_i16m2))) +vint16m2_t vwsub_wx(vint16m2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wx_i16m2_m))) +vint16m2_t vwsub_wx(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wx_i16m4))) +vint16m4_t vwsub_wx(vint16m4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wx_i16m4_m))) +vint16m4_t vwsub_wx(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wx_i16m8))) +vint16m8_t vwsub_wx(vint16m8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wx_i16m8_m))) +vint16m8_t vwsub_wx(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wx_i32mf2))) +vint32mf2_t vwsub_wx(vint32mf2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wx_i32mf2_m))) +vint32mf2_t vwsub_wx(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wx_i32m1))) +vint32m1_t vwsub_wx(vint32m1_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wx_i32m1_m))) +vint32m1_t vwsub_wx(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wx_i32m2))) +vint32m2_t vwsub_wx(vint32m2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wx_i32m2_m))) +vint32m2_t vwsub_wx(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wx_i32m4))) +vint32m4_t vwsub_wx(vint32m4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wx_i32m4_m))) +vint32m4_t vwsub_wx(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wx_i32m8))) +vint32m8_t vwsub_wx(vint32m8_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wx_i32m8_m))) +vint32m8_t vwsub_wx(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wx_i64m1))) +vint64m1_t vwsub_wx(vint64m1_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wx_i64m1_m))) +vint64m1_t vwsub_wx(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wx_i64m2))) +vint64m2_t vwsub_wx(vint64m2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wx_i64m2_m))) +vint64m2_t vwsub_wx(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wx_i64m4))) +vint64m4_t vwsub_wx(vint64m4_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wx_i64m4_m))) +vint64m4_t vwsub_wx(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wx_i64m8))) +vint64m8_t vwsub_wx(vint64m8_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwsub_wx_i64m8_m))) +vint64m8_t vwsub_wx(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvt_x_x_v_i16m2))) +vint16m2_t vwcvt_x(vint8m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvt_x_x_v_i16m2_m))) +vint16m2_t vwcvt_x(vbool8_t op0, vint16m2_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvt_x_x_v_i16m4))) +vint16m4_t vwcvt_x(vint8m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvt_x_x_v_i16m4_m))) +vint16m4_t vwcvt_x(vbool4_t op0, vint16m4_t op1, vint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvt_x_x_v_i16m8))) +vint16m8_t vwcvt_x(vint8m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvt_x_x_v_i16m8_m))) +vint16m8_t vwcvt_x(vbool2_t op0, vint16m8_t op1, vint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvt_x_x_v_i16m1))) +vint16m1_t vwcvt_x(vint8mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvt_x_x_v_i16m1_m))) +vint16m1_t vwcvt_x(vbool16_t op0, vint16m1_t op1, vint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvt_x_x_v_i16mf2))) +vint16mf2_t vwcvt_x(vint8mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvt_x_x_v_i16mf2_m))) +vint16mf2_t vwcvt_x(vbool32_t op0, vint16mf2_t op1, vint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvt_x_x_v_i16mf4))) +vint16mf4_t vwcvt_x(vint8mf8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvt_x_x_v_i16mf4_m))) +vint16mf4_t vwcvt_x(vbool64_t op0, vint16mf4_t op1, vint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvt_x_x_v_i32m2))) +vint32m2_t vwcvt_x(vint16m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvt_x_x_v_i32m2_m))) +vint32m2_t vwcvt_x(vbool16_t op0, vint32m2_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvt_x_x_v_i32m4))) +vint32m4_t vwcvt_x(vint16m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvt_x_x_v_i32m4_m))) +vint32m4_t vwcvt_x(vbool8_t op0, vint32m4_t op1, vint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvt_x_x_v_i32m8))) +vint32m8_t vwcvt_x(vint16m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvt_x_x_v_i32m8_m))) +vint32m8_t vwcvt_x(vbool4_t op0, vint32m8_t op1, vint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvt_x_x_v_i32m1))) +vint32m1_t vwcvt_x(vint16mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvt_x_x_v_i32m1_m))) +vint32m1_t vwcvt_x(vbool32_t op0, vint32m1_t op1, vint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvt_x_x_v_i32mf2))) +vint32mf2_t vwcvt_x(vint16mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvt_x_x_v_i32mf2_m))) +vint32mf2_t vwcvt_x(vbool64_t op0, vint32mf2_t op1, vint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvt_x_x_v_i64m2))) +vint64m2_t vwcvt_x(vint32m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvt_x_x_v_i64m2_m))) +vint64m2_t vwcvt_x(vbool32_t op0, vint64m2_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvt_x_x_v_i64m4))) +vint64m4_t vwcvt_x(vint32m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvt_x_x_v_i64m4_m))) +vint64m4_t vwcvt_x(vbool16_t op0, vint64m4_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvt_x_x_v_i64m8))) +vint64m8_t vwcvt_x(vint32m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvt_x_x_v_i64m8_m))) +vint64m8_t vwcvt_x(vbool8_t op0, vint64m8_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvt_x_x_v_i64m1))) +vint64m1_t vwcvt_x(vint32mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvt_x_x_v_i64m1_m))) +vint64m1_t vwcvt_x(vbool64_t op0, vint64m1_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_u8m1))) +void vsuxei64(uint8_t * op0, vuint64m8_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_u8m1_m))) +void vsuxei64(vbool8_t op0, uint8_t * op1, vuint64m8_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_u8mf2))) +void vsuxei64(uint8_t * op0, vuint64m4_t op1, vuint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_u8mf2_m))) +void vsuxei64(vbool16_t op0, uint8_t * op1, vuint64m4_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_u8mf4))) +void vsuxei64(uint8_t * op0, vuint64m2_t op1, vuint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_u8mf4_m))) +void vsuxei64(vbool32_t op0, uint8_t * op1, vuint64m2_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_u8mf8))) +void vsuxei64(uint8_t * op0, vuint64m1_t op1, vuint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_u8mf8_m))) +void vsuxei64(vbool64_t op0, uint8_t * op1, vuint64m1_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vvm_i8m1))) +vint8m1_t vadc(vint8m1_t op0, vint8m1_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vvm_i8m2))) +vint8m2_t vadc(vint8m2_t op0, vint8m2_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vvm_i8m4))) +vint8m4_t vadc(vint8m4_t op0, vint8m4_t op1, vbool2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vvm_i8m8))) +vint8m8_t vadc(vint8m8_t op0, vint8m8_t op1, vbool1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vvm_i8mf2))) +vint8mf2_t vadc(vint8mf2_t op0, vint8mf2_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vvm_i8mf4))) +vint8mf4_t vadc(vint8mf4_t op0, vint8mf4_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vvm_i8mf8))) +vint8mf8_t vadc(vint8mf8_t op0, vint8mf8_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vvm_i16m1))) +vint16m1_t vadc(vint16m1_t op0, vint16m1_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vvm_i16m2))) +vint16m2_t vadc(vint16m2_t op0, vint16m2_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vvm_i16m4))) +vint16m4_t vadc(vint16m4_t op0, vint16m4_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vvm_i16m8))) +vint16m8_t vadc(vint16m8_t op0, vint16m8_t op1, vbool2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vvm_i16mf2))) +vint16mf2_t vadc(vint16mf2_t op0, vint16mf2_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vvm_i16mf4))) +vint16mf4_t vadc(vint16mf4_t op0, vint16mf4_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vvm_i32m1))) +vint32m1_t vadc(vint32m1_t op0, vint32m1_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vvm_i32m2))) +vint32m2_t vadc(vint32m2_t op0, vint32m2_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vvm_i32m4))) +vint32m4_t vadc(vint32m4_t op0, vint32m4_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vvm_i32m8))) +vint32m8_t vadc(vint32m8_t op0, vint32m8_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vvm_i32mf2))) +vint32mf2_t vadc(vint32mf2_t op0, vint32mf2_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vvm_i64m1))) +vint64m1_t vadc(vint64m1_t op0, vint64m1_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vvm_i64m2))) +vint64m2_t vadc(vint64m2_t op0, vint64m2_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vvm_i64m4))) +vint64m4_t vadc(vint64m4_t op0, vint64m4_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vvm_i64m8))) +vint64m8_t vadc(vint64m8_t op0, vint64m8_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vxm_i8m1))) +vint8m1_t vadc(vint8m1_t op0, int8_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vxm_i8m2))) +vint8m2_t vadc(vint8m2_t op0, int8_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vxm_i8m4))) +vint8m4_t vadc(vint8m4_t op0, int8_t op1, vbool2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vxm_i8m8))) +vint8m8_t vadc(vint8m8_t op0, int8_t op1, vbool1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vxm_i8mf2))) +vint8mf2_t vadc(vint8mf2_t op0, int8_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vxm_i8mf4))) +vint8mf4_t vadc(vint8mf4_t op0, int8_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vxm_i8mf8))) +vint8mf8_t vadc(vint8mf8_t op0, int8_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vxm_i16m1))) +vint16m1_t vadc(vint16m1_t op0, int16_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vxm_i16m2))) +vint16m2_t vadc(vint16m2_t op0, int16_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vxm_i16m4))) +vint16m4_t vadc(vint16m4_t op0, int16_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vxm_i16m8))) +vint16m8_t vadc(vint16m8_t op0, int16_t op1, vbool2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vxm_i16mf2))) +vint16mf2_t vadc(vint16mf2_t op0, int16_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vxm_i16mf4))) +vint16mf4_t vadc(vint16mf4_t op0, int16_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vxm_i32m1))) +vint32m1_t vadc(vint32m1_t op0, int32_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vxm_i32m2))) +vint32m2_t vadc(vint32m2_t op0, int32_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vxm_i32m4))) +vint32m4_t vadc(vint32m4_t op0, int32_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vxm_i32m8))) +vint32m8_t vadc(vint32m8_t op0, int32_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vxm_i32mf2))) +vint32mf2_t vadc(vint32mf2_t op0, int32_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vxm_i64m1))) +vint64m1_t vadc(vint64m1_t op0, int64_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vxm_i64m2))) +vint64m2_t vadc(vint64m2_t op0, int64_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vxm_i64m4))) +vint64m4_t vadc(vint64m4_t op0, int64_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vxm_i64m8))) +vint64m8_t vadc(vint64m8_t op0, int64_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vvm_u8m1))) +vuint8m1_t vadc(vuint8m1_t op0, vuint8m1_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vvm_u8m2))) +vuint8m2_t vadc(vuint8m2_t op0, vuint8m2_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vvm_u8m4))) +vuint8m4_t vadc(vuint8m4_t op0, vuint8m4_t op1, vbool2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vvm_u8m8))) +vuint8m8_t vadc(vuint8m8_t op0, vuint8m8_t op1, vbool1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vvm_u8mf2))) +vuint8mf2_t vadc(vuint8mf2_t op0, vuint8mf2_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vvm_u8mf4))) +vuint8mf4_t vadc(vuint8mf4_t op0, vuint8mf4_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vvm_u8mf8))) +vuint8mf8_t vadc(vuint8mf8_t op0, vuint8mf8_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vvm_u16m1))) +vuint16m1_t vadc(vuint16m1_t op0, vuint16m1_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vvm_u16m2))) +vuint16m2_t vadc(vuint16m2_t op0, vuint16m2_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vvm_u16m4))) +vuint16m4_t vadc(vuint16m4_t op0, vuint16m4_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vvm_u16m8))) +vuint16m8_t vadc(vuint16m8_t op0, vuint16m8_t op1, vbool2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vvm_u16mf2))) +vuint16mf2_t vadc(vuint16mf2_t op0, vuint16mf2_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vvm_u16mf4))) +vuint16mf4_t vadc(vuint16mf4_t op0, vuint16mf4_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vvm_u32m1))) +vuint32m1_t vadc(vuint32m1_t op0, vuint32m1_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vvm_u32m2))) +vuint32m2_t vadc(vuint32m2_t op0, vuint32m2_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vvm_u32m4))) +vuint32m4_t vadc(vuint32m4_t op0, vuint32m4_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vvm_u32m8))) +vuint32m8_t vadc(vuint32m8_t op0, vuint32m8_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vvm_u32mf2))) +vuint32mf2_t vadc(vuint32mf2_t op0, vuint32mf2_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vvm_u64m1))) +vuint64m1_t vadc(vuint64m1_t op0, vuint64m1_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vvm_u64m2))) +vuint64m2_t vadc(vuint64m2_t op0, vuint64m2_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vvm_u64m4))) +vuint64m4_t vadc(vuint64m4_t op0, vuint64m4_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vvm_u64m8))) +vuint64m8_t vadc(vuint64m8_t op0, vuint64m8_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vxm_u8m1))) +vuint8m1_t vadc(vuint8m1_t op0, uint8_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vxm_u8m2))) +vuint8m2_t vadc(vuint8m2_t op0, uint8_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vxm_u8m4))) +vuint8m4_t vadc(vuint8m4_t op0, uint8_t op1, vbool2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vxm_u8m8))) +vuint8m8_t vadc(vuint8m8_t op0, uint8_t op1, vbool1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vxm_u8mf2))) +vuint8mf2_t vadc(vuint8mf2_t op0, uint8_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vxm_u8mf4))) +vuint8mf4_t vadc(vuint8mf4_t op0, uint8_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vxm_u8mf8))) +vuint8mf8_t vadc(vuint8mf8_t op0, uint8_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vxm_u16m1))) +vuint16m1_t vadc(vuint16m1_t op0, uint16_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vxm_u16m2))) +vuint16m2_t vadc(vuint16m2_t op0, uint16_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vxm_u16m4))) +vuint16m4_t vadc(vuint16m4_t op0, uint16_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vxm_u16m8))) +vuint16m8_t vadc(vuint16m8_t op0, uint16_t op1, vbool2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vxm_u16mf2))) +vuint16mf2_t vadc(vuint16mf2_t op0, uint16_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vxm_u16mf4))) +vuint16mf4_t vadc(vuint16mf4_t op0, uint16_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vxm_u32m1))) +vuint32m1_t vadc(vuint32m1_t op0, uint32_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vxm_u32m2))) +vuint32m2_t vadc(vuint32m2_t op0, uint32_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vxm_u32m4))) +vuint32m4_t vadc(vuint32m4_t op0, uint32_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vxm_u32m8))) +vuint32m8_t vadc(vuint32m8_t op0, uint32_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vxm_u32mf2))) +vuint32mf2_t vadc(vuint32mf2_t op0, uint32_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vxm_u64m1))) +vuint64m1_t vadc(vuint64m1_t op0, uint64_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vxm_u64m2))) +vuint64m2_t vadc(vuint64m2_t op0, uint64_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vxm_u64m4))) +vuint64m4_t vadc(vuint64m4_t op0, uint64_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vadc_vxm_u64m8))) +vuint64m8_t vadc(vuint64m8_t op0, uint64_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vvm_i8m1_b8))) +vbool8_t vmadc(vint8m1_t op0, vint8m1_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vvm_i8m2_b4))) +vbool4_t vmadc(vint8m2_t op0, vint8m2_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vvm_i8m4_b2))) +vbool2_t vmadc(vint8m4_t op0, vint8m4_t op1, vbool2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vvm_i8m8_b1))) +vbool1_t vmadc(vint8m8_t op0, vint8m8_t op1, vbool1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vvm_i8mf2_b16))) +vbool16_t vmadc(vint8mf2_t op0, vint8mf2_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vvm_i8mf4_b32))) +vbool32_t vmadc(vint8mf4_t op0, vint8mf4_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vvm_i8mf8_b64))) +vbool64_t vmadc(vint8mf8_t op0, vint8mf8_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vvm_i16m1_b16))) +vbool16_t vmadc(vint16m1_t op0, vint16m1_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vvm_i16m2_b8))) +vbool8_t vmadc(vint16m2_t op0, vint16m2_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vvm_i16m4_b4))) +vbool4_t vmadc(vint16m4_t op0, vint16m4_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vvm_i16m8_b2))) +vbool2_t vmadc(vint16m8_t op0, vint16m8_t op1, vbool2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vvm_i16mf2_b32))) +vbool32_t vmadc(vint16mf2_t op0, vint16mf2_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vvm_i16mf4_b64))) +vbool64_t vmadc(vint16mf4_t op0, vint16mf4_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vvm_i32m1_b32))) +vbool32_t vmadc(vint32m1_t op0, vint32m1_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vvm_i32m2_b16))) +vbool16_t vmadc(vint32m2_t op0, vint32m2_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vvm_i32m4_b8))) +vbool8_t vmadc(vint32m4_t op0, vint32m4_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vvm_i32m8_b4))) +vbool4_t vmadc(vint32m8_t op0, vint32m8_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vvm_i32mf2_b64))) +vbool64_t vmadc(vint32mf2_t op0, vint32mf2_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vvm_i64m1_b64))) +vbool64_t vmadc(vint64m1_t op0, vint64m1_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vvm_i64m2_b32))) +vbool32_t vmadc(vint64m2_t op0, vint64m2_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vvm_i64m4_b16))) +vbool16_t vmadc(vint64m4_t op0, vint64m4_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vvm_i64m8_b8))) +vbool8_t vmadc(vint64m8_t op0, vint64m8_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vxm_i8m1_b8))) +vbool8_t vmadc(vint8m1_t op0, int8_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vxm_i8m2_b4))) +vbool4_t vmadc(vint8m2_t op0, int8_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vxm_i8m4_b2))) +vbool2_t vmadc(vint8m4_t op0, int8_t op1, vbool2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vxm_i8m8_b1))) +vbool1_t vmadc(vint8m8_t op0, int8_t op1, vbool1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vxm_i8mf2_b16))) +vbool16_t vmadc(vint8mf2_t op0, int8_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vxm_i8mf4_b32))) +vbool32_t vmadc(vint8mf4_t op0, int8_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vxm_i8mf8_b64))) +vbool64_t vmadc(vint8mf8_t op0, int8_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vxm_i16m1_b16))) +vbool16_t vmadc(vint16m1_t op0, int16_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vxm_i16m2_b8))) +vbool8_t vmadc(vint16m2_t op0, int16_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vxm_i16m4_b4))) +vbool4_t vmadc(vint16m4_t op0, int16_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vxm_i16m8_b2))) +vbool2_t vmadc(vint16m8_t op0, int16_t op1, vbool2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vxm_i16mf2_b32))) +vbool32_t vmadc(vint16mf2_t op0, int16_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vxm_i16mf4_b64))) +vbool64_t vmadc(vint16mf4_t op0, int16_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vxm_i32m1_b32))) +vbool32_t vmadc(vint32m1_t op0, int32_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vxm_i32m2_b16))) +vbool16_t vmadc(vint32m2_t op0, int32_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vxm_i32m4_b8))) +vbool8_t vmadc(vint32m4_t op0, int32_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vxm_i32m8_b4))) +vbool4_t vmadc(vint32m8_t op0, int32_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vxm_i32mf2_b64))) +vbool64_t vmadc(vint32mf2_t op0, int32_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vxm_i64m1_b64))) +vbool64_t vmadc(vint64m1_t op0, int64_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vxm_i64m2_b32))) +vbool32_t vmadc(vint64m2_t op0, int64_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vxm_i64m4_b16))) +vbool16_t vmadc(vint64m4_t op0, int64_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vxm_i64m8_b8))) +vbool8_t vmadc(vint64m8_t op0, int64_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vvm_u8m1_b8))) +vbool8_t vmadc(vuint8m1_t op0, vuint8m1_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vvm_u8m2_b4))) +vbool4_t vmadc(vuint8m2_t op0, vuint8m2_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vvm_u8m4_b2))) +vbool2_t vmadc(vuint8m4_t op0, vuint8m4_t op1, vbool2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vvm_u8m8_b1))) +vbool1_t vmadc(vuint8m8_t op0, vuint8m8_t op1, vbool1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vvm_u8mf2_b16))) +vbool16_t vmadc(vuint8mf2_t op0, vuint8mf2_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vvm_u8mf4_b32))) +vbool32_t vmadc(vuint8mf4_t op0, vuint8mf4_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vvm_u8mf8_b64))) +vbool64_t vmadc(vuint8mf8_t op0, vuint8mf8_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vvm_u16m1_b16))) +vbool16_t vmadc(vuint16m1_t op0, vuint16m1_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vvm_u16m2_b8))) +vbool8_t vmadc(vuint16m2_t op0, vuint16m2_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vvm_u16m4_b4))) +vbool4_t vmadc(vuint16m4_t op0, vuint16m4_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vvm_u16m8_b2))) +vbool2_t vmadc(vuint16m8_t op0, vuint16m8_t op1, vbool2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vvm_u16mf2_b32))) +vbool32_t vmadc(vuint16mf2_t op0, vuint16mf2_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vvm_u16mf4_b64))) +vbool64_t vmadc(vuint16mf4_t op0, vuint16mf4_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vvm_u32m1_b32))) +vbool32_t vmadc(vuint32m1_t op0, vuint32m1_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vvm_u32m2_b16))) +vbool16_t vmadc(vuint32m2_t op0, vuint32m2_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vvm_u32m4_b8))) +vbool8_t vmadc(vuint32m4_t op0, vuint32m4_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vvm_u32m8_b4))) +vbool4_t vmadc(vuint32m8_t op0, vuint32m8_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vvm_u32mf2_b64))) +vbool64_t vmadc(vuint32mf2_t op0, vuint32mf2_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vvm_u64m1_b64))) +vbool64_t vmadc(vuint64m1_t op0, vuint64m1_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vvm_u64m2_b32))) +vbool32_t vmadc(vuint64m2_t op0, vuint64m2_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vvm_u64m4_b16))) +vbool16_t vmadc(vuint64m4_t op0, vuint64m4_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vvm_u64m8_b8))) +vbool8_t vmadc(vuint64m8_t op0, vuint64m8_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vxm_u8m1_b8))) +vbool8_t vmadc(vuint8m1_t op0, uint8_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vxm_u8m2_b4))) +vbool4_t vmadc(vuint8m2_t op0, uint8_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vxm_u8m4_b2))) +vbool2_t vmadc(vuint8m4_t op0, uint8_t op1, vbool2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vxm_u8m8_b1))) +vbool1_t vmadc(vuint8m8_t op0, uint8_t op1, vbool1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vxm_u8mf2_b16))) +vbool16_t vmadc(vuint8mf2_t op0, uint8_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vxm_u8mf4_b32))) +vbool32_t vmadc(vuint8mf4_t op0, uint8_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vxm_u8mf8_b64))) +vbool64_t vmadc(vuint8mf8_t op0, uint8_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vxm_u16m1_b16))) +vbool16_t vmadc(vuint16m1_t op0, uint16_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vxm_u16m2_b8))) +vbool8_t vmadc(vuint16m2_t op0, uint16_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vxm_u16m4_b4))) +vbool4_t vmadc(vuint16m4_t op0, uint16_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vxm_u16m8_b2))) +vbool2_t vmadc(vuint16m8_t op0, uint16_t op1, vbool2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vxm_u16mf2_b32))) +vbool32_t vmadc(vuint16mf2_t op0, uint16_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vxm_u16mf4_b64))) +vbool64_t vmadc(vuint16mf4_t op0, uint16_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vxm_u32m1_b32))) +vbool32_t vmadc(vuint32m1_t op0, uint32_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vxm_u32m2_b16))) +vbool16_t vmadc(vuint32m2_t op0, uint32_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vxm_u32m4_b8))) +vbool8_t vmadc(vuint32m4_t op0, uint32_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vxm_u32m8_b4))) +vbool4_t vmadc(vuint32m8_t op0, uint32_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vxm_u32mf2_b64))) +vbool64_t vmadc(vuint32mf2_t op0, uint32_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vxm_u64m1_b64))) +vbool64_t vmadc(vuint64m1_t op0, uint64_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vxm_u64m2_b32))) +vbool32_t vmadc(vuint64m2_t op0, uint64_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vxm_u64m4_b16))) +vbool16_t vmadc(vuint64m4_t op0, uint64_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vxm_u64m8_b8))) +vbool8_t vmadc(vuint64m8_t op0, uint64_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vv_i8m1_b8))) +vbool8_t vmadc(vint8m1_t op0, vint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vv_i8m2_b4))) +vbool4_t vmadc(vint8m2_t op0, vint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vv_i8m4_b2))) +vbool2_t vmadc(vint8m4_t op0, vint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vv_i8m8_b1))) +vbool1_t vmadc(vint8m8_t op0, vint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vv_i8mf2_b16))) +vbool16_t vmadc(vint8mf2_t op0, vint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vv_i8mf4_b32))) +vbool32_t vmadc(vint8mf4_t op0, vint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vv_i8mf8_b64))) +vbool64_t vmadc(vint8mf8_t op0, vint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vv_i16m1_b16))) +vbool16_t vmadc(vint16m1_t op0, vint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vv_i16m2_b8))) +vbool8_t vmadc(vint16m2_t op0, vint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vv_i16m4_b4))) +vbool4_t vmadc(vint16m4_t op0, vint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vv_i16m8_b2))) +vbool2_t vmadc(vint16m8_t op0, vint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vv_i16mf2_b32))) +vbool32_t vmadc(vint16mf2_t op0, vint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vv_i16mf4_b64))) +vbool64_t vmadc(vint16mf4_t op0, vint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vv_i32m1_b32))) +vbool32_t vmadc(vint32m1_t op0, vint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vv_i32m2_b16))) +vbool16_t vmadc(vint32m2_t op0, vint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vv_i32m4_b8))) +vbool8_t vmadc(vint32m4_t op0, vint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vv_i32m8_b4))) +vbool4_t vmadc(vint32m8_t op0, vint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vv_i32mf2_b64))) +vbool64_t vmadc(vint32mf2_t op0, vint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vv_i64m1_b64))) +vbool64_t vmadc(vint64m1_t op0, vint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vv_i64m2_b32))) +vbool32_t vmadc(vint64m2_t op0, vint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vv_i64m4_b16))) +vbool16_t vmadc(vint64m4_t op0, vint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vv_i64m8_b8))) +vbool8_t vmadc(vint64m8_t op0, vint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vx_i8m1_b8))) +vbool8_t vmadc(vint8m1_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vx_i8m2_b4))) +vbool4_t vmadc(vint8m2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vx_i8m4_b2))) +vbool2_t vmadc(vint8m4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vx_i8m8_b1))) +vbool1_t vmadc(vint8m8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vx_i8mf2_b16))) +vbool16_t vmadc(vint8mf2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vx_i8mf4_b32))) +vbool32_t vmadc(vint8mf4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vx_i8mf8_b64))) +vbool64_t vmadc(vint8mf8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vx_i16m1_b16))) +vbool16_t vmadc(vint16m1_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vx_i16m2_b8))) +vbool8_t vmadc(vint16m2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vx_i16m4_b4))) +vbool4_t vmadc(vint16m4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vx_i16m8_b2))) +vbool2_t vmadc(vint16m8_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vx_i16mf2_b32))) +vbool32_t vmadc(vint16mf2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vx_i16mf4_b64))) +vbool64_t vmadc(vint16mf4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vx_i32m1_b32))) +vbool32_t vmadc(vint32m1_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vx_i32m2_b16))) +vbool16_t vmadc(vint32m2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vx_i32m4_b8))) +vbool8_t vmadc(vint32m4_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vx_i32m8_b4))) +vbool4_t vmadc(vint32m8_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vx_i32mf2_b64))) +vbool64_t vmadc(vint32mf2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vx_i64m1_b64))) +vbool64_t vmadc(vint64m1_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vx_i64m2_b32))) +vbool32_t vmadc(vint64m2_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vx_i64m4_b16))) +vbool16_t vmadc(vint64m4_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vx_i64m8_b8))) +vbool8_t vmadc(vint64m8_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_i16m1))) +void vsuxei8(int16_t * op0, vuint8mf2_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_i16m1_m))) +void vsuxei8(vbool16_t op0, int16_t * op1, vuint8mf2_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_i16m2))) +void vsuxei8(int16_t * op0, vuint8m1_t op1, vint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_i16m2_m))) +void vsuxei8(vbool8_t op0, int16_t * op1, vuint8m1_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_i16m4))) +void vsuxei8(int16_t * op0, vuint8m2_t op1, vint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_i16m4_m))) +void vsuxei8(vbool4_t op0, int16_t * op1, vuint8m2_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_i16m8))) +void vsuxei8(int16_t * op0, vuint8m4_t op1, vint16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_i16m8_m))) +void vsuxei8(vbool2_t op0, int16_t * op1, vuint8m4_t op2, vint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_i16mf2))) +void vsuxei8(int16_t * op0, vuint8mf4_t op1, vint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_i16mf2_m))) +void vsuxei8(vbool32_t op0, int16_t * op1, vuint8mf4_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_i16mf4))) +void vsuxei8(int16_t * op0, vuint8mf8_t op1, vint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_i16mf4_m))) +void vsuxei8(vbool64_t op0, int16_t * op1, vuint8mf8_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vv_u8m1_b8))) +vbool8_t vmadc(vuint8m1_t op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vv_u8m2_b4))) +vbool4_t vmadc(vuint8m2_t op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vv_u8m4_b2))) +vbool2_t vmadc(vuint8m4_t op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vv_u8m8_b1))) +vbool1_t vmadc(vuint8m8_t op0, vuint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vv_u8mf2_b16))) +vbool16_t vmadc(vuint8mf2_t op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vv_u8mf4_b32))) +vbool32_t vmadc(vuint8mf4_t op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vv_u8mf8_b64))) +vbool64_t vmadc(vuint8mf8_t op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vv_u16m1_b16))) +vbool16_t vmadc(vuint16m1_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vv_u16m2_b8))) +vbool8_t vmadc(vuint16m2_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vv_u16m4_b4))) +vbool4_t vmadc(vuint16m4_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vv_u16m8_b2))) +vbool2_t vmadc(vuint16m8_t op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vv_u16mf2_b32))) +vbool32_t vmadc(vuint16mf2_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vv_u16mf4_b64))) +vbool64_t vmadc(vuint16mf4_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vv_u32m1_b32))) +vbool32_t vmadc(vuint32m1_t op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vv_u32m2_b16))) +vbool16_t vmadc(vuint32m2_t op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vv_u32m4_b8))) +vbool8_t vmadc(vuint32m4_t op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vv_u32m8_b4))) +vbool4_t vmadc(vuint32m8_t op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vv_u32mf2_b64))) +vbool64_t vmadc(vuint32mf2_t op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vv_u64m1_b64))) +vbool64_t vmadc(vuint64m1_t op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vv_u64m2_b32))) +vbool32_t vmadc(vuint64m2_t op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vv_u64m4_b16))) +vbool16_t vmadc(vuint64m4_t op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vv_u64m8_b8))) +vbool8_t vmadc(vuint64m8_t op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vx_u8m1_b8))) +vbool8_t vmadc(vuint8m1_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vx_u8m2_b4))) +vbool4_t vmadc(vuint8m2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vx_u8m4_b2))) +vbool2_t vmadc(vuint8m4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vx_u8m8_b1))) +vbool1_t vmadc(vuint8m8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vx_u8mf2_b16))) +vbool16_t vmadc(vuint8mf2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vx_u8mf4_b32))) +vbool32_t vmadc(vuint8mf4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vx_u8mf8_b64))) +vbool64_t vmadc(vuint8mf8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vx_u16m1_b16))) +vbool16_t vmadc(vuint16m1_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vx_u16m2_b8))) +vbool8_t vmadc(vuint16m2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vx_u16m4_b4))) +vbool4_t vmadc(vuint16m4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vx_u16m8_b2))) +vbool2_t vmadc(vuint16m8_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vx_u16mf2_b32))) +vbool32_t vmadc(vuint16mf2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vx_u16mf4_b64))) +vbool64_t vmadc(vuint16mf4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vx_u32m1_b32))) +vbool32_t vmadc(vuint32m1_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vx_u32m2_b16))) +vbool16_t vmadc(vuint32m2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vx_u32m4_b8))) +vbool8_t vmadc(vuint32m4_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vx_u32m8_b4))) +vbool4_t vmadc(vuint32m8_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vx_u32mf2_b64))) +vbool64_t vmadc(vuint32mf2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vx_u64m1_b64))) +vbool64_t vmadc(vuint64m1_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vx_u64m2_b32))) +vbool32_t vmadc(vuint64m2_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vx_u64m4_b16))) +vbool16_t vmadc(vuint64m4_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadc_vx_u64m8_b8))) +vbool8_t vmadc(vuint64m8_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vvm_i8m1))) +vint8m1_t vsbc(vint8m1_t op0, vint8m1_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vvm_i8m2))) +vint8m2_t vsbc(vint8m2_t op0, vint8m2_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vvm_i8m4))) +vint8m4_t vsbc(vint8m4_t op0, vint8m4_t op1, vbool2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vvm_i8m8))) +vint8m8_t vsbc(vint8m8_t op0, vint8m8_t op1, vbool1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vvm_i8mf2))) +vint8mf2_t vsbc(vint8mf2_t op0, vint8mf2_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vvm_i8mf4))) +vint8mf4_t vsbc(vint8mf4_t op0, vint8mf4_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vvm_i8mf8))) +vint8mf8_t vsbc(vint8mf8_t op0, vint8mf8_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vvm_i16m1))) +vint16m1_t vsbc(vint16m1_t op0, vint16m1_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vvm_i16m2))) +vint16m2_t vsbc(vint16m2_t op0, vint16m2_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vvm_i16m4))) +vint16m4_t vsbc(vint16m4_t op0, vint16m4_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vvm_i16m8))) +vint16m8_t vsbc(vint16m8_t op0, vint16m8_t op1, vbool2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vvm_i16mf2))) +vint16mf2_t vsbc(vint16mf2_t op0, vint16mf2_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vvm_i16mf4))) +vint16mf4_t vsbc(vint16mf4_t op0, vint16mf4_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vvm_i32m1))) +vint32m1_t vsbc(vint32m1_t op0, vint32m1_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vvm_i32m2))) +vint32m2_t vsbc(vint32m2_t op0, vint32m2_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vvm_i32m4))) +vint32m4_t vsbc(vint32m4_t op0, vint32m4_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vvm_i32m8))) +vint32m8_t vsbc(vint32m8_t op0, vint32m8_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vvm_i32mf2))) +vint32mf2_t vsbc(vint32mf2_t op0, vint32mf2_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vvm_i64m1))) +vint64m1_t vsbc(vint64m1_t op0, vint64m1_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vvm_i64m2))) +vint64m2_t vsbc(vint64m2_t op0, vint64m2_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vvm_i64m4))) +vint64m4_t vsbc(vint64m4_t op0, vint64m4_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vvm_i64m8))) +vint64m8_t vsbc(vint64m8_t op0, vint64m8_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vxm_i8m1))) +vint8m1_t vsbc(vint8m1_t op0, int8_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vxm_i8m2))) +vint8m2_t vsbc(vint8m2_t op0, int8_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vxm_i8m4))) +vint8m4_t vsbc(vint8m4_t op0, int8_t op1, vbool2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vxm_i8m8))) +vint8m8_t vsbc(vint8m8_t op0, int8_t op1, vbool1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vxm_i8mf2))) +vint8mf2_t vsbc(vint8mf2_t op0, int8_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vxm_i8mf4))) +vint8mf4_t vsbc(vint8mf4_t op0, int8_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vxm_i8mf8))) +vint8mf8_t vsbc(vint8mf8_t op0, int8_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vxm_i16m1))) +vint16m1_t vsbc(vint16m1_t op0, int16_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vxm_i16m2))) +vint16m2_t vsbc(vint16m2_t op0, int16_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vxm_i16m4))) +vint16m4_t vsbc(vint16m4_t op0, int16_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vxm_i16m8))) +vint16m8_t vsbc(vint16m8_t op0, int16_t op1, vbool2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vxm_i16mf2))) +vint16mf2_t vsbc(vint16mf2_t op0, int16_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vxm_i16mf4))) +vint16mf4_t vsbc(vint16mf4_t op0, int16_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vxm_i32m1))) +vint32m1_t vsbc(vint32m1_t op0, int32_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vxm_i32m2))) +vint32m2_t vsbc(vint32m2_t op0, int32_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vxm_i32m4))) +vint32m4_t vsbc(vint32m4_t op0, int32_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vxm_i32m8))) +vint32m8_t vsbc(vint32m8_t op0, int32_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vxm_i32mf2))) +vint32mf2_t vsbc(vint32mf2_t op0, int32_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vxm_i64m1))) +vint64m1_t vsbc(vint64m1_t op0, int64_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vxm_i64m2))) +vint64m2_t vsbc(vint64m2_t op0, int64_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vxm_i64m4))) +vint64m4_t vsbc(vint64m4_t op0, int64_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vxm_i64m8))) +vint64m8_t vsbc(vint64m8_t op0, int64_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vvm_u8m1))) +vuint8m1_t vsbc(vuint8m1_t op0, vuint8m1_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vvm_u8m2))) +vuint8m2_t vsbc(vuint8m2_t op0, vuint8m2_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vvm_u8m4))) +vuint8m4_t vsbc(vuint8m4_t op0, vuint8m4_t op1, vbool2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vvm_u8m8))) +vuint8m8_t vsbc(vuint8m8_t op0, vuint8m8_t op1, vbool1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vvm_u8mf2))) +vuint8mf2_t vsbc(vuint8mf2_t op0, vuint8mf2_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vvm_u8mf4))) +vuint8mf4_t vsbc(vuint8mf4_t op0, vuint8mf4_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vvm_u8mf8))) +vuint8mf8_t vsbc(vuint8mf8_t op0, vuint8mf8_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vvm_u16m1))) +vuint16m1_t vsbc(vuint16m1_t op0, vuint16m1_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vvm_u16m2))) +vuint16m2_t vsbc(vuint16m2_t op0, vuint16m2_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vvm_u16m4))) +vuint16m4_t vsbc(vuint16m4_t op0, vuint16m4_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vvm_u16m8))) +vuint16m8_t vsbc(vuint16m8_t op0, vuint16m8_t op1, vbool2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vvm_u16mf2))) +vuint16mf2_t vsbc(vuint16mf2_t op0, vuint16mf2_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vvm_u16mf4))) +vuint16mf4_t vsbc(vuint16mf4_t op0, vuint16mf4_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vvm_u32m1))) +vuint32m1_t vsbc(vuint32m1_t op0, vuint32m1_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vvm_u32m2))) +vuint32m2_t vsbc(vuint32m2_t op0, vuint32m2_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vvm_u32m4))) +vuint32m4_t vsbc(vuint32m4_t op0, vuint32m4_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vvm_u32m8))) +vuint32m8_t vsbc(vuint32m8_t op0, vuint32m8_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vvm_u32mf2))) +vuint32mf2_t vsbc(vuint32mf2_t op0, vuint32mf2_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vvm_u64m1))) +vuint64m1_t vsbc(vuint64m1_t op0, vuint64m1_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vvm_u64m2))) +vuint64m2_t vsbc(vuint64m2_t op0, vuint64m2_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vvm_u64m4))) +vuint64m4_t vsbc(vuint64m4_t op0, vuint64m4_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vvm_u64m8))) +vuint64m8_t vsbc(vuint64m8_t op0, vuint64m8_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vxm_u8m1))) +vuint8m1_t vsbc(vuint8m1_t op0, uint8_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vxm_u8m2))) +vuint8m2_t vsbc(vuint8m2_t op0, uint8_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vxm_u8m4))) +vuint8m4_t vsbc(vuint8m4_t op0, uint8_t op1, vbool2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vxm_u8m8))) +vuint8m8_t vsbc(vuint8m8_t op0, uint8_t op1, vbool1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vxm_u8mf2))) +vuint8mf2_t vsbc(vuint8mf2_t op0, uint8_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vxm_u8mf4))) +vuint8mf4_t vsbc(vuint8mf4_t op0, uint8_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vxm_u8mf8))) +vuint8mf8_t vsbc(vuint8mf8_t op0, uint8_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vxm_u16m1))) +vuint16m1_t vsbc(vuint16m1_t op0, uint16_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vxm_u16m2))) +vuint16m2_t vsbc(vuint16m2_t op0, uint16_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vxm_u16m4))) +vuint16m4_t vsbc(vuint16m4_t op0, uint16_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vxm_u16m8))) +vuint16m8_t vsbc(vuint16m8_t op0, uint16_t op1, vbool2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vxm_u16mf2))) +vuint16mf2_t vsbc(vuint16mf2_t op0, uint16_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vxm_u16mf4))) +vuint16mf4_t vsbc(vuint16mf4_t op0, uint16_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vxm_u32m1))) +vuint32m1_t vsbc(vuint32m1_t op0, uint32_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vxm_u32m2))) +vuint32m2_t vsbc(vuint32m2_t op0, uint32_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vxm_u32m4))) +vuint32m4_t vsbc(vuint32m4_t op0, uint32_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vxm_u32m8))) +vuint32m8_t vsbc(vuint32m8_t op0, uint32_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vxm_u32mf2))) +vuint32mf2_t vsbc(vuint32mf2_t op0, uint32_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vxm_u64m1))) +vuint64m1_t vsbc(vuint64m1_t op0, uint64_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vxm_u64m2))) +vuint64m2_t vsbc(vuint64m2_t op0, uint64_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vxm_u64m4))) +vuint64m4_t vsbc(vuint64m4_t op0, uint64_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsbc_vxm_u64m8))) +vuint64m8_t vsbc(vuint64m8_t op0, uint64_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vvm_i8m1_b8))) +vbool8_t vmsbc(vint8m1_t op0, vint8m1_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vvm_i8m2_b4))) +vbool4_t vmsbc(vint8m2_t op0, vint8m2_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vvm_i8m4_b2))) +vbool2_t vmsbc(vint8m4_t op0, vint8m4_t op1, vbool2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vvm_i8m8_b1))) +vbool1_t vmsbc(vint8m8_t op0, vint8m8_t op1, vbool1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vvm_i8mf2_b16))) +vbool16_t vmsbc(vint8mf2_t op0, vint8mf2_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vvm_i8mf4_b32))) +vbool32_t vmsbc(vint8mf4_t op0, vint8mf4_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vvm_i8mf8_b64))) +vbool64_t vmsbc(vint8mf8_t op0, vint8mf8_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vvm_i16m1_b16))) +vbool16_t vmsbc(vint16m1_t op0, vint16m1_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vvm_i16m2_b8))) +vbool8_t vmsbc(vint16m2_t op0, vint16m2_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vvm_i16m4_b4))) +vbool4_t vmsbc(vint16m4_t op0, vint16m4_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vvm_i16m8_b2))) +vbool2_t vmsbc(vint16m8_t op0, vint16m8_t op1, vbool2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vvm_i16mf2_b32))) +vbool32_t vmsbc(vint16mf2_t op0, vint16mf2_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vvm_i16mf4_b64))) +vbool64_t vmsbc(vint16mf4_t op0, vint16mf4_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vvm_i32m1_b32))) +vbool32_t vmsbc(vint32m1_t op0, vint32m1_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vvm_i32m2_b16))) +vbool16_t vmsbc(vint32m2_t op0, vint32m2_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vvm_i32m4_b8))) +vbool8_t vmsbc(vint32m4_t op0, vint32m4_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vvm_i32m8_b4))) +vbool4_t vmsbc(vint32m8_t op0, vint32m8_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vvm_i32mf2_b64))) +vbool64_t vmsbc(vint32mf2_t op0, vint32mf2_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vvm_i64m1_b64))) +vbool64_t vmsbc(vint64m1_t op0, vint64m1_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vvm_i64m2_b32))) +vbool32_t vmsbc(vint64m2_t op0, vint64m2_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vvm_i64m4_b16))) +vbool16_t vmsbc(vint64m4_t op0, vint64m4_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vvm_i64m8_b8))) +vbool8_t vmsbc(vint64m8_t op0, vint64m8_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vxm_i8m1_b8))) +vbool8_t vmsbc(vint8m1_t op0, int8_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vxm_i8m2_b4))) +vbool4_t vmsbc(vint8m2_t op0, int8_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vxm_i8m4_b2))) +vbool2_t vmsbc(vint8m4_t op0, int8_t op1, vbool2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vxm_i8m8_b1))) +vbool1_t vmsbc(vint8m8_t op0, int8_t op1, vbool1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vxm_i8mf2_b16))) +vbool16_t vmsbc(vint8mf2_t op0, int8_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vxm_i8mf4_b32))) +vbool32_t vmsbc(vint8mf4_t op0, int8_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vxm_i8mf8_b64))) +vbool64_t vmsbc(vint8mf8_t op0, int8_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vxm_i16m1_b16))) +vbool16_t vmsbc(vint16m1_t op0, int16_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vxm_i16m2_b8))) +vbool8_t vmsbc(vint16m2_t op0, int16_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vxm_i16m4_b4))) +vbool4_t vmsbc(vint16m4_t op0, int16_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vxm_i16m8_b2))) +vbool2_t vmsbc(vint16m8_t op0, int16_t op1, vbool2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vxm_i16mf2_b32))) +vbool32_t vmsbc(vint16mf2_t op0, int16_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vxm_i16mf4_b64))) +vbool64_t vmsbc(vint16mf4_t op0, int16_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vxm_i32m1_b32))) +vbool32_t vmsbc(vint32m1_t op0, int32_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vxm_i32m2_b16))) +vbool16_t vmsbc(vint32m2_t op0, int32_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vxm_i32m4_b8))) +vbool8_t vmsbc(vint32m4_t op0, int32_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vxm_i32m8_b4))) +vbool4_t vmsbc(vint32m8_t op0, int32_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vxm_i32mf2_b64))) +vbool64_t vmsbc(vint32mf2_t op0, int32_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vxm_i64m1_b64))) +vbool64_t vmsbc(vint64m1_t op0, int64_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vxm_i64m2_b32))) +vbool32_t vmsbc(vint64m2_t op0, int64_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vxm_i64m4_b16))) +vbool16_t vmsbc(vint64m4_t op0, int64_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vxm_i64m8_b8))) +vbool8_t vmsbc(vint64m8_t op0, int64_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vvm_u8m1_b8))) +vbool8_t vmsbc(vuint8m1_t op0, vuint8m1_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vvm_u8m2_b4))) +vbool4_t vmsbc(vuint8m2_t op0, vuint8m2_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vvm_u8m4_b2))) +vbool2_t vmsbc(vuint8m4_t op0, vuint8m4_t op1, vbool2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vvm_u8m8_b1))) +vbool1_t vmsbc(vuint8m8_t op0, vuint8m8_t op1, vbool1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vvm_u8mf2_b16))) +vbool16_t vmsbc(vuint8mf2_t op0, vuint8mf2_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vvm_u8mf4_b32))) +vbool32_t vmsbc(vuint8mf4_t op0, vuint8mf4_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vvm_u8mf8_b64))) +vbool64_t vmsbc(vuint8mf8_t op0, vuint8mf8_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vvm_u16m1_b16))) +vbool16_t vmsbc(vuint16m1_t op0, vuint16m1_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vvm_u16m2_b8))) +vbool8_t vmsbc(vuint16m2_t op0, vuint16m2_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vvm_u16m4_b4))) +vbool4_t vmsbc(vuint16m4_t op0, vuint16m4_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vvm_u16m8_b2))) +vbool2_t vmsbc(vuint16m8_t op0, vuint16m8_t op1, vbool2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vvm_u16mf2_b32))) +vbool32_t vmsbc(vuint16mf2_t op0, vuint16mf2_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vvm_u16mf4_b64))) +vbool64_t vmsbc(vuint16mf4_t op0, vuint16mf4_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vvm_u32m1_b32))) +vbool32_t vmsbc(vuint32m1_t op0, vuint32m1_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vvm_u32m2_b16))) +vbool16_t vmsbc(vuint32m2_t op0, vuint32m2_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vvm_u32m4_b8))) +vbool8_t vmsbc(vuint32m4_t op0, vuint32m4_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vvm_u32m8_b4))) +vbool4_t vmsbc(vuint32m8_t op0, vuint32m8_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vvm_u32mf2_b64))) +vbool64_t vmsbc(vuint32mf2_t op0, vuint32mf2_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vvm_u64m1_b64))) +vbool64_t vmsbc(vuint64m1_t op0, vuint64m1_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vvm_u64m2_b32))) +vbool32_t vmsbc(vuint64m2_t op0, vuint64m2_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vvm_u64m4_b16))) +vbool16_t vmsbc(vuint64m4_t op0, vuint64m4_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vvm_u64m8_b8))) +vbool8_t vmsbc(vuint64m8_t op0, vuint64m8_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vxm_u8m1_b8))) +vbool8_t vmsbc(vuint8m1_t op0, uint8_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vxm_u8m2_b4))) +vbool4_t vmsbc(vuint8m2_t op0, uint8_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vxm_u8m4_b2))) +vbool2_t vmsbc(vuint8m4_t op0, uint8_t op1, vbool2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vxm_u8m8_b1))) +vbool1_t vmsbc(vuint8m8_t op0, uint8_t op1, vbool1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vxm_u8mf2_b16))) +vbool16_t vmsbc(vuint8mf2_t op0, uint8_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vxm_u8mf4_b32))) +vbool32_t vmsbc(vuint8mf4_t op0, uint8_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vxm_u8mf8_b64))) +vbool64_t vmsbc(vuint8mf8_t op0, uint8_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vxm_u16m1_b16))) +vbool16_t vmsbc(vuint16m1_t op0, uint16_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vxm_u16m2_b8))) +vbool8_t vmsbc(vuint16m2_t op0, uint16_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vxm_u16m4_b4))) +vbool4_t vmsbc(vuint16m4_t op0, uint16_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vxm_u16m8_b2))) +vbool2_t vmsbc(vuint16m8_t op0, uint16_t op1, vbool2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vxm_u16mf2_b32))) +vbool32_t vmsbc(vuint16mf2_t op0, uint16_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vxm_u16mf4_b64))) +vbool64_t vmsbc(vuint16mf4_t op0, uint16_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vxm_u32m1_b32))) +vbool32_t vmsbc(vuint32m1_t op0, uint32_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vxm_u32m2_b16))) +vbool16_t vmsbc(vuint32m2_t op0, uint32_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vxm_u32m4_b8))) +vbool8_t vmsbc(vuint32m4_t op0, uint32_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vxm_u32m8_b4))) +vbool4_t vmsbc(vuint32m8_t op0, uint32_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vxm_u32mf2_b64))) +vbool64_t vmsbc(vuint32mf2_t op0, uint32_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vxm_u64m1_b64))) +vbool64_t vmsbc(vuint64m1_t op0, uint64_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vxm_u64m2_b32))) +vbool32_t vmsbc(vuint64m2_t op0, uint64_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vxm_u64m4_b16))) +vbool16_t vmsbc(vuint64m4_t op0, uint64_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vxm_u64m8_b8))) +vbool8_t vmsbc(vuint64m8_t op0, uint64_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_u16m1))) +void vsuxei8(uint16_t * op0, vuint8mf2_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_u16m1_m))) +void vsuxei8(vbool16_t op0, uint16_t * op1, vuint8mf2_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_u16m2))) +void vsuxei8(uint16_t * op0, vuint8m1_t op1, vuint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_u16m2_m))) +void vsuxei8(vbool8_t op0, uint16_t * op1, vuint8m1_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_u16m4))) +void vsuxei8(uint16_t * op0, vuint8m2_t op1, vuint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_u16m4_m))) +void vsuxei8(vbool4_t op0, uint16_t * op1, vuint8m2_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_u16m8))) +void vsuxei8(uint16_t * op0, vuint8m4_t op1, vuint16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_u16m8_m))) +void vsuxei8(vbool2_t op0, uint16_t * op1, vuint8m4_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_u16mf2))) +void vsuxei8(uint16_t * op0, vuint8mf4_t op1, vuint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_u16mf2_m))) +void vsuxei8(vbool32_t op0, uint16_t * op1, vuint8mf4_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_u16mf4))) +void vsuxei8(uint16_t * op0, vuint8mf8_t op1, vuint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_u16mf4_m))) +void vsuxei8(vbool64_t op0, uint16_t * op1, vuint8mf8_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vv_i8m1_b8))) +vbool8_t vmsbc(vint8m1_t op0, vint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vv_i8m2_b4))) +vbool4_t vmsbc(vint8m2_t op0, vint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vv_i8m4_b2))) +vbool2_t vmsbc(vint8m4_t op0, vint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vv_i8m8_b1))) +vbool1_t vmsbc(vint8m8_t op0, vint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vv_i8mf2_b16))) +vbool16_t vmsbc(vint8mf2_t op0, vint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vv_i8mf4_b32))) +vbool32_t vmsbc(vint8mf4_t op0, vint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vv_i8mf8_b64))) +vbool64_t vmsbc(vint8mf8_t op0, vint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vv_i16m1_b16))) +vbool16_t vmsbc(vint16m1_t op0, vint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vv_i16m2_b8))) +vbool8_t vmsbc(vint16m2_t op0, vint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vv_i16m4_b4))) +vbool4_t vmsbc(vint16m4_t op0, vint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vv_i16m8_b2))) +vbool2_t vmsbc(vint16m8_t op0, vint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vv_i16mf2_b32))) +vbool32_t vmsbc(vint16mf2_t op0, vint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vv_i16mf4_b64))) +vbool64_t vmsbc(vint16mf4_t op0, vint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vv_i32m1_b32))) +vbool32_t vmsbc(vint32m1_t op0, vint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vv_i32m2_b16))) +vbool16_t vmsbc(vint32m2_t op0, vint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vv_i32m4_b8))) +vbool8_t vmsbc(vint32m4_t op0, vint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vv_i32m8_b4))) +vbool4_t vmsbc(vint32m8_t op0, vint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vv_i32mf2_b64))) +vbool64_t vmsbc(vint32mf2_t op0, vint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vv_i64m1_b64))) +vbool64_t vmsbc(vint64m1_t op0, vint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vv_i64m2_b32))) +vbool32_t vmsbc(vint64m2_t op0, vint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vv_i64m4_b16))) +vbool16_t vmsbc(vint64m4_t op0, vint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vv_i64m8_b8))) +vbool8_t vmsbc(vint64m8_t op0, vint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vx_i8m1_b8))) +vbool8_t vmsbc(vint8m1_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vx_i8m2_b4))) +vbool4_t vmsbc(vint8m2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vx_i8m4_b2))) +vbool2_t vmsbc(vint8m4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vx_i8m8_b1))) +vbool1_t vmsbc(vint8m8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vx_i8mf2_b16))) +vbool16_t vmsbc(vint8mf2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vx_i8mf4_b32))) +vbool32_t vmsbc(vint8mf4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vx_i8mf8_b64))) +vbool64_t vmsbc(vint8mf8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vx_i16m1_b16))) +vbool16_t vmsbc(vint16m1_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vx_i16m2_b8))) +vbool8_t vmsbc(vint16m2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vx_i16m4_b4))) +vbool4_t vmsbc(vint16m4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vx_i16m8_b2))) +vbool2_t vmsbc(vint16m8_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vx_i16mf2_b32))) +vbool32_t vmsbc(vint16mf2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vx_i16mf4_b64))) +vbool64_t vmsbc(vint16mf4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vx_i32m1_b32))) +vbool32_t vmsbc(vint32m1_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vx_i32m2_b16))) +vbool16_t vmsbc(vint32m2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vx_i32m4_b8))) +vbool8_t vmsbc(vint32m4_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vx_i32m8_b4))) +vbool4_t vmsbc(vint32m8_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vx_i32mf2_b64))) +vbool64_t vmsbc(vint32mf2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vx_i64m1_b64))) +vbool64_t vmsbc(vint64m1_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vx_i64m2_b32))) +vbool32_t vmsbc(vint64m2_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vx_i64m4_b16))) +vbool16_t vmsbc(vint64m4_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vx_i64m8_b8))) +vbool8_t vmsbc(vint64m8_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vv_u8m1_b8))) +vbool8_t vmsbc(vuint8m1_t op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vv_u8m2_b4))) +vbool4_t vmsbc(vuint8m2_t op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vv_u8m4_b2))) +vbool2_t vmsbc(vuint8m4_t op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vv_u8m8_b1))) +vbool1_t vmsbc(vuint8m8_t op0, vuint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vv_u8mf2_b16))) +vbool16_t vmsbc(vuint8mf2_t op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vv_u8mf4_b32))) +vbool32_t vmsbc(vuint8mf4_t op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vv_u8mf8_b64))) +vbool64_t vmsbc(vuint8mf8_t op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vv_u16m1_b16))) +vbool16_t vmsbc(vuint16m1_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vv_u16m2_b8))) +vbool8_t vmsbc(vuint16m2_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vv_u16m4_b4))) +vbool4_t vmsbc(vuint16m4_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vv_u16m8_b2))) +vbool2_t vmsbc(vuint16m8_t op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vv_u16mf2_b32))) +vbool32_t vmsbc(vuint16mf2_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vv_u16mf4_b64))) +vbool64_t vmsbc(vuint16mf4_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vv_u32m1_b32))) +vbool32_t vmsbc(vuint32m1_t op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vv_u32m2_b16))) +vbool16_t vmsbc(vuint32m2_t op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vv_u32m4_b8))) +vbool8_t vmsbc(vuint32m4_t op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vv_u32m8_b4))) +vbool4_t vmsbc(vuint32m8_t op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vv_u32mf2_b64))) +vbool64_t vmsbc(vuint32mf2_t op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vv_u64m1_b64))) +vbool64_t vmsbc(vuint64m1_t op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vv_u64m2_b32))) +vbool32_t vmsbc(vuint64m2_t op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vv_u64m4_b16))) +vbool16_t vmsbc(vuint64m4_t op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vv_u64m8_b8))) +vbool8_t vmsbc(vuint64m8_t op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vx_u8m1_b8))) +vbool8_t vmsbc(vuint8m1_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vx_u8m2_b4))) +vbool4_t vmsbc(vuint8m2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vx_u8m4_b2))) +vbool2_t vmsbc(vuint8m4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vx_u8m8_b1))) +vbool1_t vmsbc(vuint8m8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vx_u8mf2_b16))) +vbool16_t vmsbc(vuint8mf2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vx_u8mf4_b32))) +vbool32_t vmsbc(vuint8mf4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vx_u8mf8_b64))) +vbool64_t vmsbc(vuint8mf8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vx_u16m1_b16))) +vbool16_t vmsbc(vuint16m1_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vx_u16m2_b8))) +vbool8_t vmsbc(vuint16m2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vx_u16m4_b4))) +vbool4_t vmsbc(vuint16m4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vx_u16m8_b2))) +vbool2_t vmsbc(vuint16m8_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vx_u16mf2_b32))) +vbool32_t vmsbc(vuint16mf2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vx_u16mf4_b64))) +vbool64_t vmsbc(vuint16mf4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vx_u32m1_b32))) +vbool32_t vmsbc(vuint32m1_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vx_u32m2_b16))) +vbool16_t vmsbc(vuint32m2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vx_u32m4_b8))) +vbool8_t vmsbc(vuint32m4_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vx_u32m8_b4))) +vbool4_t vmsbc(vuint32m8_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vx_u32mf2_b64))) +vbool64_t vmsbc(vuint32mf2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vx_u64m1_b64))) +vbool64_t vmsbc(vuint64m1_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vx_u64m2_b32))) +vbool32_t vmsbc(vuint64m2_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vx_u64m4_b16))) +vbool16_t vmsbc(vuint64m4_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbc_vx_u64m8_b8))) +vbool8_t vmsbc(vuint64m8_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_i8m1))) +vint8m1_t vand(vint8m1_t op0, vint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_i8m1_m))) +vint8m1_t vand(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_i8m2))) +vint8m2_t vand(vint8m2_t op0, vint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_i8m2_m))) +vint8m2_t vand(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_i8m4))) +vint8m4_t vand(vint8m4_t op0, vint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_i8m4_m))) +vint8m4_t vand(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_i8m8))) +vint8m8_t vand(vint8m8_t op0, vint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_i8m8_m))) +vint8m8_t vand(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, vint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_i8mf2))) +vint8mf2_t vand(vint8mf2_t op0, vint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_i8mf2_m))) +vint8mf2_t vand(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_i8mf4))) +vint8mf4_t vand(vint8mf4_t op0, vint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_i8mf4_m))) +vint8mf4_t vand(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_i8mf8))) +vint8mf8_t vand(vint8mf8_t op0, vint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_i8mf8_m))) +vint8mf8_t vand(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_i16m1))) +vint16m1_t vand(vint16m1_t op0, vint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_i16m1_m))) +vint16m1_t vand(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_i16m2))) +vint16m2_t vand(vint16m2_t op0, vint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_i16m2_m))) +vint16m2_t vand(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_i16m4))) +vint16m4_t vand(vint16m4_t op0, vint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_i16m4_m))) +vint16m4_t vand(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_i16m8))) +vint16m8_t vand(vint16m8_t op0, vint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_i16m8_m))) +vint16m8_t vand(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, vint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_i16mf2))) +vint16mf2_t vand(vint16mf2_t op0, vint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_i16mf2_m))) +vint16mf2_t vand(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_i16mf4))) +vint16mf4_t vand(vint16mf4_t op0, vint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_i16mf4_m))) +vint16mf4_t vand(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_i32m1))) +vint32m1_t vand(vint32m1_t op0, vint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_i32m1_m))) +vint32m1_t vand(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_i32m2))) +vint32m2_t vand(vint32m2_t op0, vint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_i32m2_m))) +vint32m2_t vand(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_i32m4))) +vint32m4_t vand(vint32m4_t op0, vint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_i32m4_m))) +vint32m4_t vand(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_i32m8))) +vint32m8_t vand(vint32m8_t op0, vint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_i32m8_m))) +vint32m8_t vand(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_i32mf2))) +vint32mf2_t vand(vint32mf2_t op0, vint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_i32mf2_m))) +vint32mf2_t vand(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_i64m1))) +vint64m1_t vand(vint64m1_t op0, vint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_i64m1_m))) +vint64m1_t vand(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_i64m2))) +vint64m2_t vand(vint64m2_t op0, vint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_i64m2_m))) +vint64m2_t vand(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_i64m4))) +vint64m4_t vand(vint64m4_t op0, vint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_i64m4_m))) +vint64m4_t vand(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_i64m8))) +vint64m8_t vand(vint64m8_t op0, vint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_i64m8_m))) +vint64m8_t vand(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_i8m1))) +vint8m1_t vand(vint8m1_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_i8m1_m))) +vint8m1_t vand(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_i8m2))) +vint8m2_t vand(vint8m2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_i8m2_m))) +vint8m2_t vand(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_i8m4))) +vint8m4_t vand(vint8m4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_i8m4_m))) +vint8m4_t vand(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_i8m8))) +vint8m8_t vand(vint8m8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_i8m8_m))) +vint8m8_t vand(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_i8mf2))) +vint8mf2_t vand(vint8mf2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_i8mf2_m))) +vint8mf2_t vand(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_i8mf4))) +vint8mf4_t vand(vint8mf4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_i8mf4_m))) +vint8mf4_t vand(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_i8mf8))) +vint8mf8_t vand(vint8mf8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_i8mf8_m))) +vint8mf8_t vand(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_i16m1))) +vint16m1_t vand(vint16m1_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_i16m1_m))) +vint16m1_t vand(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_i16m2))) +vint16m2_t vand(vint16m2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_i16m2_m))) +vint16m2_t vand(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_i16m4))) +vint16m4_t vand(vint16m4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_i16m4_m))) +vint16m4_t vand(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_i16m8))) +vint16m8_t vand(vint16m8_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_i16m8_m))) +vint16m8_t vand(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_i16mf2))) +vint16mf2_t vand(vint16mf2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_i16mf2_m))) +vint16mf2_t vand(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_i16mf4))) +vint16mf4_t vand(vint16mf4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_i16mf4_m))) +vint16mf4_t vand(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_i32m1))) +vint32m1_t vand(vint32m1_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_i32m1_m))) +vint32m1_t vand(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_i32m2))) +vint32m2_t vand(vint32m2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_i32m2_m))) +vint32m2_t vand(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_i32m4))) +vint32m4_t vand(vint32m4_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_i32m4_m))) +vint32m4_t vand(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_i32m8))) +vint32m8_t vand(vint32m8_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_i32m8_m))) +vint32m8_t vand(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_i32mf2))) +vint32mf2_t vand(vint32mf2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_i32mf2_m))) +vint32mf2_t vand(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_i64m1))) +vint64m1_t vand(vint64m1_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_i64m1_m))) +vint64m1_t vand(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_i64m2))) +vint64m2_t vand(vint64m2_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_i64m2_m))) +vint64m2_t vand(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_i64m4))) +vint64m4_t vand(vint64m4_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_i64m4_m))) +vint64m4_t vand(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_i64m8))) +vint64m8_t vand(vint64m8_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_i64m8_m))) +vint64m8_t vand(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_u8m1))) +vuint8m1_t vand(vuint8m1_t op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_u8m1_m))) +vuint8m1_t vand(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_u8m2))) +vuint8m2_t vand(vuint8m2_t op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_u8m2_m))) +vuint8m2_t vand(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_u8m4))) +vuint8m4_t vand(vuint8m4_t op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_u8m4_m))) +vuint8m4_t vand(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_u8m8))) +vuint8m8_t vand(vuint8m8_t op0, vuint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_u8m8_m))) +vuint8m8_t vand(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, vuint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_u8mf2))) +vuint8mf2_t vand(vuint8mf2_t op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_u8mf2_m))) +vuint8mf2_t vand(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_u8mf4))) +vuint8mf4_t vand(vuint8mf4_t op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_u8mf4_m))) +vuint8mf4_t vand(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_u8mf8))) +vuint8mf8_t vand(vuint8mf8_t op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_u8mf8_m))) +vuint8mf8_t vand(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_u16m1))) +vuint16m1_t vand(vuint16m1_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_u16m1_m))) +vuint16m1_t vand(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_u16m2))) +vuint16m2_t vand(vuint16m2_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_u16m2_m))) +vuint16m2_t vand(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_u16m4))) +vuint16m4_t vand(vuint16m4_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_u16m4_m))) +vuint16m4_t vand(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_u16m8))) +vuint16m8_t vand(vuint16m8_t op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_u16m8_m))) +vuint16m8_t vand(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_u16mf2))) +vuint16mf2_t vand(vuint16mf2_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_u16mf2_m))) +vuint16mf2_t vand(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_u16mf4))) +vuint16mf4_t vand(vuint16mf4_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_u16mf4_m))) +vuint16mf4_t vand(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_u32m1))) +vuint32m1_t vand(vuint32m1_t op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_u32m1_m))) +vuint32m1_t vand(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_u32m2))) +vuint32m2_t vand(vuint32m2_t op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_u32m2_m))) +vuint32m2_t vand(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_u32m4))) +vuint32m4_t vand(vuint32m4_t op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_u32m4_m))) +vuint32m4_t vand(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_u32m8))) +vuint32m8_t vand(vuint32m8_t op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_u32m8_m))) +vuint32m8_t vand(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_u32mf2))) +vuint32mf2_t vand(vuint32mf2_t op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_u32mf2_m))) +vuint32mf2_t vand(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_u64m1))) +vuint64m1_t vand(vuint64m1_t op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_u64m1_m))) +vuint64m1_t vand(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_u64m2))) +vuint64m2_t vand(vuint64m2_t op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_u64m2_m))) +vuint64m2_t vand(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_u64m4))) +vuint64m4_t vand(vuint64m4_t op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_u64m4_m))) +vuint64m4_t vand(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_u64m8))) +vuint64m8_t vand(vuint64m8_t op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vv_u64m8_m))) +vuint64m8_t vand(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_u8m1))) +vuint8m1_t vand(vuint8m1_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_u8m1_m))) +vuint8m1_t vand(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_u8m2))) +vuint8m2_t vand(vuint8m2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_u8m2_m))) +vuint8m2_t vand(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_u8m4))) +vuint8m4_t vand(vuint8m4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_u8m4_m))) +vuint8m4_t vand(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_u8m8))) +vuint8m8_t vand(vuint8m8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_u8m8_m))) +vuint8m8_t vand(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_u8mf2))) +vuint8mf2_t vand(vuint8mf2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_u8mf2_m))) +vuint8mf2_t vand(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_u8mf4))) +vuint8mf4_t vand(vuint8mf4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_u8mf4_m))) +vuint8mf4_t vand(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_u8mf8))) +vuint8mf8_t vand(vuint8mf8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_u8mf8_m))) +vuint8mf8_t vand(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_u16m1))) +vuint16m1_t vand(vuint16m1_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_u16m1_m))) +vuint16m1_t vand(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_u16m2))) +vuint16m2_t vand(vuint16m2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_u16m2_m))) +vuint16m2_t vand(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_u16m4))) +vuint16m4_t vand(vuint16m4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_u16m4_m))) +vuint16m4_t vand(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_u16m8))) +vuint16m8_t vand(vuint16m8_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_u16m8_m))) +vuint16m8_t vand(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_u16mf2))) +vuint16mf2_t vand(vuint16mf2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_u16mf2_m))) +vuint16mf2_t vand(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_u16mf4))) +vuint16mf4_t vand(vuint16mf4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_u16mf4_m))) +vuint16mf4_t vand(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_u32m1))) +vuint32m1_t vand(vuint32m1_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_u32m1_m))) +vuint32m1_t vand(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_u32m2))) +vuint32m2_t vand(vuint32m2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_u32m2_m))) +vuint32m2_t vand(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_u32m4))) +vuint32m4_t vand(vuint32m4_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_u32m4_m))) +vuint32m4_t vand(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_u32m8))) +vuint32m8_t vand(vuint32m8_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_u32m8_m))) +vuint32m8_t vand(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_u32mf2))) +vuint32mf2_t vand(vuint32mf2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_u32mf2_m))) +vuint32mf2_t vand(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_u64m1))) +vuint64m1_t vand(vuint64m1_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_u64m1_m))) +vuint64m1_t vand(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_u64m2))) +vuint64m2_t vand(vuint64m2_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_u64m2_m))) +vuint64m2_t vand(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_u64m4))) +vuint64m4_t vand(vuint64m4_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_u64m4_m))) +vuint64m4_t vand(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_u64m8))) +vuint64m8_t vand(vuint64m8_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vand_vx_u64m8_m))) +vuint64m8_t vand(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_i8m1))) +vint8m1_t vxor(vint8m1_t op0, vint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_i8m1_m))) +vint8m1_t vxor(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_i8m2))) +vint8m2_t vxor(vint8m2_t op0, vint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_i8m2_m))) +vint8m2_t vxor(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_i8m4))) +vint8m4_t vxor(vint8m4_t op0, vint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_i8m4_m))) +vint8m4_t vxor(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_i8m8))) +vint8m8_t vxor(vint8m8_t op0, vint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_i8m8_m))) +vint8m8_t vxor(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, vint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_i8mf2))) +vint8mf2_t vxor(vint8mf2_t op0, vint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_i8mf2_m))) +vint8mf2_t vxor(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_i8mf4))) +vint8mf4_t vxor(vint8mf4_t op0, vint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_i8mf4_m))) +vint8mf4_t vxor(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_i8mf8))) +vint8mf8_t vxor(vint8mf8_t op0, vint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_i8mf8_m))) +vint8mf8_t vxor(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_i16m1))) +vint16m1_t vxor(vint16m1_t op0, vint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_i16m1_m))) +vint16m1_t vxor(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_i16m2))) +vint16m2_t vxor(vint16m2_t op0, vint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_i16m2_m))) +vint16m2_t vxor(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_i16m4))) +vint16m4_t vxor(vint16m4_t op0, vint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_i16m4_m))) +vint16m4_t vxor(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_i16m8))) +vint16m8_t vxor(vint16m8_t op0, vint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_i16m8_m))) +vint16m8_t vxor(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, vint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_i16mf2))) +vint16mf2_t vxor(vint16mf2_t op0, vint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_i16mf2_m))) +vint16mf2_t vxor(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_i16mf4))) +vint16mf4_t vxor(vint16mf4_t op0, vint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_i16mf4_m))) +vint16mf4_t vxor(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_i32m1))) +vint32m1_t vxor(vint32m1_t op0, vint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_i32m1_m))) +vint32m1_t vxor(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_i32m2))) +vint32m2_t vxor(vint32m2_t op0, vint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_i32m2_m))) +vint32m2_t vxor(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_i32m4))) +vint32m4_t vxor(vint32m4_t op0, vint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_i32m4_m))) +vint32m4_t vxor(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_i32m8))) +vint32m8_t vxor(vint32m8_t op0, vint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_i32m8_m))) +vint32m8_t vxor(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_i32mf2))) +vint32mf2_t vxor(vint32mf2_t op0, vint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_i32mf2_m))) +vint32mf2_t vxor(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_i64m1))) +vint64m1_t vxor(vint64m1_t op0, vint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_i64m1_m))) +vint64m1_t vxor(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_i64m2))) +vint64m2_t vxor(vint64m2_t op0, vint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_i64m2_m))) +vint64m2_t vxor(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_i64m4))) +vint64m4_t vxor(vint64m4_t op0, vint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_i64m4_m))) +vint64m4_t vxor(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_i64m8))) +vint64m8_t vxor(vint64m8_t op0, vint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_i64m8_m))) +vint64m8_t vxor(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_i8m1))) +vint8m1_t vxor(vint8m1_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_i8m1_m))) +vint8m1_t vxor(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_i8m2))) +vint8m2_t vxor(vint8m2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_i8m2_m))) +vint8m2_t vxor(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_i8m4))) +vint8m4_t vxor(vint8m4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_i8m4_m))) +vint8m4_t vxor(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_i8m8))) +vint8m8_t vxor(vint8m8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_i8m8_m))) +vint8m8_t vxor(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_i8mf2))) +vint8mf2_t vxor(vint8mf2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_i8mf2_m))) +vint8mf2_t vxor(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_i8mf4))) +vint8mf4_t vxor(vint8mf4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_i8mf4_m))) +vint8mf4_t vxor(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_i8mf8))) +vint8mf8_t vxor(vint8mf8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_i8mf8_m))) +vint8mf8_t vxor(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_i16m1))) +vint16m1_t vxor(vint16m1_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_i16m1_m))) +vint16m1_t vxor(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_i16m2))) +vint16m2_t vxor(vint16m2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_i16m2_m))) +vint16m2_t vxor(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_i16m4))) +vint16m4_t vxor(vint16m4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_i16m4_m))) +vint16m4_t vxor(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_i16m8))) +vint16m8_t vxor(vint16m8_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_i16m8_m))) +vint16m8_t vxor(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_i16mf2))) +vint16mf2_t vxor(vint16mf2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_i16mf2_m))) +vint16mf2_t vxor(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_i16mf4))) +vint16mf4_t vxor(vint16mf4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_i16mf4_m))) +vint16mf4_t vxor(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_i32m1))) +vint32m1_t vxor(vint32m1_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_i32m1_m))) +vint32m1_t vxor(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_i32m2))) +vint32m2_t vxor(vint32m2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_i32m2_m))) +vint32m2_t vxor(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_i32m4))) +vint32m4_t vxor(vint32m4_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_i32m4_m))) +vint32m4_t vxor(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_i32m8))) +vint32m8_t vxor(vint32m8_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_i32m8_m))) +vint32m8_t vxor(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_i32mf2))) +vint32mf2_t vxor(vint32mf2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_i32mf2_m))) +vint32mf2_t vxor(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_i64m1))) +vint64m1_t vxor(vint64m1_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_i64m1_m))) +vint64m1_t vxor(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_i64m2))) +vint64m2_t vxor(vint64m2_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_i64m2_m))) +vint64m2_t vxor(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_i64m4))) +vint64m4_t vxor(vint64m4_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_i64m4_m))) +vint64m4_t vxor(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_i64m8))) +vint64m8_t vxor(vint64m8_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_i64m8_m))) +vint64m8_t vxor(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_i16m1))) +void vsuxei16(int16_t * op0, vuint16m1_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_i16m1_m))) +void vsuxei16(vbool16_t op0, int16_t * op1, vuint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_i16m2))) +void vsuxei16(int16_t * op0, vuint16m2_t op1, vint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_i16m2_m))) +void vsuxei16(vbool8_t op0, int16_t * op1, vuint16m2_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_i16m4))) +void vsuxei16(int16_t * op0, vuint16m4_t op1, vint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_i16m4_m))) +void vsuxei16(vbool4_t op0, int16_t * op1, vuint16m4_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_i16m8))) +void vsuxei16(int16_t * op0, vuint16m8_t op1, vint16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_i16m8_m))) +void vsuxei16(vbool2_t op0, int16_t * op1, vuint16m8_t op2, vint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_i16mf2))) +void vsuxei16(int16_t * op0, vuint16mf2_t op1, vint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_i16mf2_m))) +void vsuxei16(vbool32_t op0, int16_t * op1, vuint16mf2_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_i16mf4))) +void vsuxei16(int16_t * op0, vuint16mf4_t op1, vint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_i16mf4_m))) +void vsuxei16(vbool64_t op0, int16_t * op1, vuint16mf4_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_u8m1))) +vuint8m1_t vxor(vuint8m1_t op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_u8m1_m))) +vuint8m1_t vxor(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_u8m2))) +vuint8m2_t vxor(vuint8m2_t op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_u8m2_m))) +vuint8m2_t vxor(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_u8m4))) +vuint8m4_t vxor(vuint8m4_t op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_u8m4_m))) +vuint8m4_t vxor(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_u8m8))) +vuint8m8_t vxor(vuint8m8_t op0, vuint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_u8m8_m))) +vuint8m8_t vxor(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, vuint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_u8mf2))) +vuint8mf2_t vxor(vuint8mf2_t op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_u8mf2_m))) +vuint8mf2_t vxor(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_u8mf4))) +vuint8mf4_t vxor(vuint8mf4_t op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_u8mf4_m))) +vuint8mf4_t vxor(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_u8mf8))) +vuint8mf8_t vxor(vuint8mf8_t op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_u8mf8_m))) +vuint8mf8_t vxor(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_u16m1))) +vuint16m1_t vxor(vuint16m1_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_u16m1_m))) +vuint16m1_t vxor(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_u16m2))) +vuint16m2_t vxor(vuint16m2_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_u16m2_m))) +vuint16m2_t vxor(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_u16m4))) +vuint16m4_t vxor(vuint16m4_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_u16m4_m))) +vuint16m4_t vxor(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_u16m8))) +vuint16m8_t vxor(vuint16m8_t op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_u16m8_m))) +vuint16m8_t vxor(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_u16mf2))) +vuint16mf2_t vxor(vuint16mf2_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_u16mf2_m))) +vuint16mf2_t vxor(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_u16mf4))) +vuint16mf4_t vxor(vuint16mf4_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_u16mf4_m))) +vuint16mf4_t vxor(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_u32m1))) +vuint32m1_t vxor(vuint32m1_t op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_u32m1_m))) +vuint32m1_t vxor(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_u32m2))) +vuint32m2_t vxor(vuint32m2_t op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_u32m2_m))) +vuint32m2_t vxor(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_u32m4))) +vuint32m4_t vxor(vuint32m4_t op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_u32m4_m))) +vuint32m4_t vxor(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_u32m8))) +vuint32m8_t vxor(vuint32m8_t op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_u32m8_m))) +vuint32m8_t vxor(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_u32mf2))) +vuint32mf2_t vxor(vuint32mf2_t op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_u32mf2_m))) +vuint32mf2_t vxor(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_u64m1))) +vuint64m1_t vxor(vuint64m1_t op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_u64m1_m))) +vuint64m1_t vxor(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_u64m2))) +vuint64m2_t vxor(vuint64m2_t op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_u64m2_m))) +vuint64m2_t vxor(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_u64m4))) +vuint64m4_t vxor(vuint64m4_t op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_u64m4_m))) +vuint64m4_t vxor(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_u64m8))) +vuint64m8_t vxor(vuint64m8_t op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vv_u64m8_m))) +vuint64m8_t vxor(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_u8m1))) +vuint8m1_t vxor(vuint8m1_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_u8m1_m))) +vuint8m1_t vxor(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_u8m2))) +vuint8m2_t vxor(vuint8m2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_u8m2_m))) +vuint8m2_t vxor(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_u8m4))) +vuint8m4_t vxor(vuint8m4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_u8m4_m))) +vuint8m4_t vxor(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_u8m8))) +vuint8m8_t vxor(vuint8m8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_u8m8_m))) +vuint8m8_t vxor(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_u8mf2))) +vuint8mf2_t vxor(vuint8mf2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_u8mf2_m))) +vuint8mf2_t vxor(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_u8mf4))) +vuint8mf4_t vxor(vuint8mf4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_u8mf4_m))) +vuint8mf4_t vxor(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_u8mf8))) +vuint8mf8_t vxor(vuint8mf8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_u8mf8_m))) +vuint8mf8_t vxor(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_u16m1))) +vuint16m1_t vxor(vuint16m1_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_u16m1_m))) +vuint16m1_t vxor(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_u16m2))) +vuint16m2_t vxor(vuint16m2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_u16m2_m))) +vuint16m2_t vxor(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_u16m4))) +vuint16m4_t vxor(vuint16m4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_u16m4_m))) +vuint16m4_t vxor(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_u16m8))) +vuint16m8_t vxor(vuint16m8_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_u16m8_m))) +vuint16m8_t vxor(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_u16mf2))) +vuint16mf2_t vxor(vuint16mf2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_u16mf2_m))) +vuint16mf2_t vxor(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_u16mf4))) +vuint16mf4_t vxor(vuint16mf4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_u16mf4_m))) +vuint16mf4_t vxor(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_u32m1))) +vuint32m1_t vxor(vuint32m1_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_u32m1_m))) +vuint32m1_t vxor(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_u32m2))) +vuint32m2_t vxor(vuint32m2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_u32m2_m))) +vuint32m2_t vxor(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_u32m4))) +vuint32m4_t vxor(vuint32m4_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_u32m4_m))) +vuint32m4_t vxor(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_u32m8))) +vuint32m8_t vxor(vuint32m8_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_u32m8_m))) +vuint32m8_t vxor(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_u32mf2))) +vuint32mf2_t vxor(vuint32mf2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_u32mf2_m))) +vuint32mf2_t vxor(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_u64m1))) +vuint64m1_t vxor(vuint64m1_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_u64m1_m))) +vuint64m1_t vxor(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_u64m2))) +vuint64m2_t vxor(vuint64m2_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_u64m2_m))) +vuint64m2_t vxor(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_u64m4))) +vuint64m4_t vxor(vuint64m4_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_u64m4_m))) +vuint64m4_t vxor(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_u64m8))) +vuint64m8_t vxor(vuint64m8_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vxor_vx_u64m8_m))) +vuint64m8_t vxor(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_i8m1))) +vint8m1_t vor(vint8m1_t op0, vint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_i8m1_m))) +vint8m1_t vor(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_i8m2))) +vint8m2_t vor(vint8m2_t op0, vint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_i8m2_m))) +vint8m2_t vor(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_i8m4))) +vint8m4_t vor(vint8m4_t op0, vint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_i8m4_m))) +vint8m4_t vor(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_i8m8))) +vint8m8_t vor(vint8m8_t op0, vint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_i8m8_m))) +vint8m8_t vor(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, vint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_i8mf2))) +vint8mf2_t vor(vint8mf2_t op0, vint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_i8mf2_m))) +vint8mf2_t vor(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_i8mf4))) +vint8mf4_t vor(vint8mf4_t op0, vint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_i8mf4_m))) +vint8mf4_t vor(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_i8mf8))) +vint8mf8_t vor(vint8mf8_t op0, vint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_i8mf8_m))) +vint8mf8_t vor(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_i16m1))) +vint16m1_t vor(vint16m1_t op0, vint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_i16m1_m))) +vint16m1_t vor(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_i16m2))) +vint16m2_t vor(vint16m2_t op0, vint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_i16m2_m))) +vint16m2_t vor(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_i16m4))) +vint16m4_t vor(vint16m4_t op0, vint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_i16m4_m))) +vint16m4_t vor(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_i16m8))) +vint16m8_t vor(vint16m8_t op0, vint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_i16m8_m))) +vint16m8_t vor(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, vint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_i16mf2))) +vint16mf2_t vor(vint16mf2_t op0, vint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_i16mf2_m))) +vint16mf2_t vor(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_i16mf4))) +vint16mf4_t vor(vint16mf4_t op0, vint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_i16mf4_m))) +vint16mf4_t vor(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_i32m1))) +vint32m1_t vor(vint32m1_t op0, vint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_i32m1_m))) +vint32m1_t vor(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_i32m2))) +vint32m2_t vor(vint32m2_t op0, vint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_i32m2_m))) +vint32m2_t vor(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_i32m4))) +vint32m4_t vor(vint32m4_t op0, vint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_i32m4_m))) +vint32m4_t vor(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_i32m8))) +vint32m8_t vor(vint32m8_t op0, vint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_i32m8_m))) +vint32m8_t vor(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_i32mf2))) +vint32mf2_t vor(vint32mf2_t op0, vint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_i32mf2_m))) +vint32mf2_t vor(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_i64m1))) +vint64m1_t vor(vint64m1_t op0, vint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_i64m1_m))) +vint64m1_t vor(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_i64m2))) +vint64m2_t vor(vint64m2_t op0, vint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_i64m2_m))) +vint64m2_t vor(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_i64m4))) +vint64m4_t vor(vint64m4_t op0, vint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_i64m4_m))) +vint64m4_t vor(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_i64m8))) +vint64m8_t vor(vint64m8_t op0, vint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_i64m8_m))) +vint64m8_t vor(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_i8m1))) +vint8m1_t vor(vint8m1_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_i8m1_m))) +vint8m1_t vor(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_i8m2))) +vint8m2_t vor(vint8m2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_i8m2_m))) +vint8m2_t vor(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_i8m4))) +vint8m4_t vor(vint8m4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_i8m4_m))) +vint8m4_t vor(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_i8m8))) +vint8m8_t vor(vint8m8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_i8m8_m))) +vint8m8_t vor(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_i8mf2))) +vint8mf2_t vor(vint8mf2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_i8mf2_m))) +vint8mf2_t vor(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_i8mf4))) +vint8mf4_t vor(vint8mf4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_i8mf4_m))) +vint8mf4_t vor(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_i8mf8))) +vint8mf8_t vor(vint8mf8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_i8mf8_m))) +vint8mf8_t vor(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_i16m1))) +vint16m1_t vor(vint16m1_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_i16m1_m))) +vint16m1_t vor(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_i16m2))) +vint16m2_t vor(vint16m2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_i16m2_m))) +vint16m2_t vor(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_i16m4))) +vint16m4_t vor(vint16m4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_i16m4_m))) +vint16m4_t vor(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_i16m8))) +vint16m8_t vor(vint16m8_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_i16m8_m))) +vint16m8_t vor(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_i16mf2))) +vint16mf2_t vor(vint16mf2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_i16mf2_m))) +vint16mf2_t vor(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_i16mf4))) +vint16mf4_t vor(vint16mf4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_i16mf4_m))) +vint16mf4_t vor(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_i32m1))) +vint32m1_t vor(vint32m1_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_i32m1_m))) +vint32m1_t vor(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_i32m2))) +vint32m2_t vor(vint32m2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_i32m2_m))) +vint32m2_t vor(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_i32m4))) +vint32m4_t vor(vint32m4_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_i32m4_m))) +vint32m4_t vor(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_i32m8))) +vint32m8_t vor(vint32m8_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_i32m8_m))) +vint32m8_t vor(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_i32mf2))) +vint32mf2_t vor(vint32mf2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_i32mf2_m))) +vint32mf2_t vor(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_i64m1))) +vint64m1_t vor(vint64m1_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_i64m1_m))) +vint64m1_t vor(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_i64m2))) +vint64m2_t vor(vint64m2_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_i64m2_m))) +vint64m2_t vor(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_i64m4))) +vint64m4_t vor(vint64m4_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_i64m4_m))) +vint64m4_t vor(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_i64m8))) +vint64m8_t vor(vint64m8_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_i64m8_m))) +vint64m8_t vor(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_u8m1))) +vuint8m1_t vor(vuint8m1_t op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_u8m1_m))) +vuint8m1_t vor(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_u8m2))) +vuint8m2_t vor(vuint8m2_t op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_u8m2_m))) +vuint8m2_t vor(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_u8m4))) +vuint8m4_t vor(vuint8m4_t op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_u8m4_m))) +vuint8m4_t vor(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_u8m8))) +vuint8m8_t vor(vuint8m8_t op0, vuint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_u8m8_m))) +vuint8m8_t vor(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, vuint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_u8mf2))) +vuint8mf2_t vor(vuint8mf2_t op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_u8mf2_m))) +vuint8mf2_t vor(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_u8mf4))) +vuint8mf4_t vor(vuint8mf4_t op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_u8mf4_m))) +vuint8mf4_t vor(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_u8mf8))) +vuint8mf8_t vor(vuint8mf8_t op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_u8mf8_m))) +vuint8mf8_t vor(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_u16m1))) +vuint16m1_t vor(vuint16m1_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_u16m1_m))) +vuint16m1_t vor(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_u16m2))) +vuint16m2_t vor(vuint16m2_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_u16m2_m))) +vuint16m2_t vor(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_u16m4))) +vuint16m4_t vor(vuint16m4_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_u16m4_m))) +vuint16m4_t vor(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_u16m8))) +vuint16m8_t vor(vuint16m8_t op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_u16m8_m))) +vuint16m8_t vor(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_u16mf2))) +vuint16mf2_t vor(vuint16mf2_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_u16mf2_m))) +vuint16mf2_t vor(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_u16mf4))) +vuint16mf4_t vor(vuint16mf4_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_u16mf4_m))) +vuint16mf4_t vor(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_u32m1))) +vuint32m1_t vor(vuint32m1_t op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_u32m1_m))) +vuint32m1_t vor(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_u32m2))) +vuint32m2_t vor(vuint32m2_t op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_u32m2_m))) +vuint32m2_t vor(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_u32m4))) +vuint32m4_t vor(vuint32m4_t op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_u32m4_m))) +vuint32m4_t vor(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_u32m8))) +vuint32m8_t vor(vuint32m8_t op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_u32m8_m))) +vuint32m8_t vor(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_u32mf2))) +vuint32mf2_t vor(vuint32mf2_t op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_u32mf2_m))) +vuint32mf2_t vor(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_u64m1))) +vuint64m1_t vor(vuint64m1_t op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_u64m1_m))) +vuint64m1_t vor(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_u64m2))) +vuint64m2_t vor(vuint64m2_t op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_u64m2_m))) +vuint64m2_t vor(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_u64m4))) +vuint64m4_t vor(vuint64m4_t op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_u64m4_m))) +vuint64m4_t vor(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_u64m8))) +vuint64m8_t vor(vuint64m8_t op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vv_u64m8_m))) +vuint64m8_t vor(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_u8m1))) +vuint8m1_t vor(vuint8m1_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_u8m1_m))) +vuint8m1_t vor(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_u8m2))) +vuint8m2_t vor(vuint8m2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_u8m2_m))) +vuint8m2_t vor(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_u8m4))) +vuint8m4_t vor(vuint8m4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_u8m4_m))) +vuint8m4_t vor(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_u8m8))) +vuint8m8_t vor(vuint8m8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_u8m8_m))) +vuint8m8_t vor(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_u8mf2))) +vuint8mf2_t vor(vuint8mf2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_u8mf2_m))) +vuint8mf2_t vor(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_u8mf4))) +vuint8mf4_t vor(vuint8mf4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_u8mf4_m))) +vuint8mf4_t vor(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_u8mf8))) +vuint8mf8_t vor(vuint8mf8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_u8mf8_m))) +vuint8mf8_t vor(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_u16m1))) +vuint16m1_t vor(vuint16m1_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_u16m1_m))) +vuint16m1_t vor(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_u16m2))) +vuint16m2_t vor(vuint16m2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_u16m2_m))) +vuint16m2_t vor(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_u16m4))) +vuint16m4_t vor(vuint16m4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_u16m4_m))) +vuint16m4_t vor(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_u16m8))) +vuint16m8_t vor(vuint16m8_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_u16m8_m))) +vuint16m8_t vor(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_u16mf2))) +vuint16mf2_t vor(vuint16mf2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_u16mf2_m))) +vuint16mf2_t vor(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_u16mf4))) +vuint16mf4_t vor(vuint16mf4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_u16mf4_m))) +vuint16mf4_t vor(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_u32m1))) +vuint32m1_t vor(vuint32m1_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_u32m1_m))) +vuint32m1_t vor(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_u32m2))) +vuint32m2_t vor(vuint32m2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_u32m2_m))) +vuint32m2_t vor(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_u32m4))) +vuint32m4_t vor(vuint32m4_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_u32m4_m))) +vuint32m4_t vor(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_u32m8))) +vuint32m8_t vor(vuint32m8_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_u32m8_m))) +vuint32m8_t vor(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_u32mf2))) +vuint32mf2_t vor(vuint32mf2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_u32mf2_m))) +vuint32mf2_t vor(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_u64m1))) +vuint64m1_t vor(vuint64m1_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_u64m1_m))) +vuint64m1_t vor(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_u64m2))) +vuint64m2_t vor(vuint64m2_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_u64m2_m))) +vuint64m2_t vor(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_u64m4))) +vuint64m4_t vor(vuint64m4_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_u64m4_m))) +vuint64m4_t vor(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_u64m8))) +vuint64m8_t vor(vuint64m8_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vor_vx_u64m8_m))) +vuint64m8_t vor(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_i8m1))) +vint8m1_t vsll(vint8m1_t op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_i8m1_m))) +vint8m1_t vsll(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_i8m2))) +vint8m2_t vsll(vint8m2_t op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_i8m2_m))) +vint8m2_t vsll(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_i8m4))) +vint8m4_t vsll(vint8m4_t op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_i8m4_m))) +vint8m4_t vsll(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_i8m8))) +vint8m8_t vsll(vint8m8_t op0, vuint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_i8m8_m))) +vint8m8_t vsll(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, vuint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_i8mf2))) +vint8mf2_t vsll(vint8mf2_t op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_i8mf2_m))) +vint8mf2_t vsll(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_i8mf4))) +vint8mf4_t vsll(vint8mf4_t op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_i8mf4_m))) +vint8mf4_t vsll(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_i8mf8))) +vint8mf8_t vsll(vint8mf8_t op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_i8mf8_m))) +vint8mf8_t vsll(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_i16m1))) +vint16m1_t vsll(vint16m1_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_i16m1_m))) +vint16m1_t vsll(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_i16m2))) +vint16m2_t vsll(vint16m2_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_i16m2_m))) +vint16m2_t vsll(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_i16m4))) +vint16m4_t vsll(vint16m4_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_i16m4_m))) +vint16m4_t vsll(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_i16m8))) +vint16m8_t vsll(vint16m8_t op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_i16m8_m))) +vint16m8_t vsll(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_i16mf2))) +vint16mf2_t vsll(vint16mf2_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_i16mf2_m))) +vint16mf2_t vsll(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_i16mf4))) +vint16mf4_t vsll(vint16mf4_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_i16mf4_m))) +vint16mf4_t vsll(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_i32m1))) +vint32m1_t vsll(vint32m1_t op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_i32m1_m))) +vint32m1_t vsll(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_i32m2))) +vint32m2_t vsll(vint32m2_t op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_i32m2_m))) +vint32m2_t vsll(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_i32m4))) +vint32m4_t vsll(vint32m4_t op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_i32m4_m))) +vint32m4_t vsll(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_i32m8))) +vint32m8_t vsll(vint32m8_t op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_i32m8_m))) +vint32m8_t vsll(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_i32mf2))) +vint32mf2_t vsll(vint32mf2_t op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_i32mf2_m))) +vint32mf2_t vsll(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_i64m1))) +vint64m1_t vsll(vint64m1_t op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_i64m1_m))) +vint64m1_t vsll(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_i64m2))) +vint64m2_t vsll(vint64m2_t op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_i64m2_m))) +vint64m2_t vsll(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_i64m4))) +vint64m4_t vsll(vint64m4_t op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_i64m4_m))) +vint64m4_t vsll(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_i64m8))) +vint64m8_t vsll(vint64m8_t op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_i64m8_m))) +vint64m8_t vsll(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_i8m1))) +vint8m1_t vsll(vint8m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_i8m1_m))) +vint8m1_t vsll(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_i8m2))) +vint8m2_t vsll(vint8m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_i8m2_m))) +vint8m2_t vsll(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_i8m4))) +vint8m4_t vsll(vint8m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_i8m4_m))) +vint8m4_t vsll(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_i8m8))) +vint8m8_t vsll(vint8m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_i8m8_m))) +vint8m8_t vsll(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_i8mf2))) +vint8mf2_t vsll(vint8mf2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_i8mf2_m))) +vint8mf2_t vsll(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_i8mf4))) +vint8mf4_t vsll(vint8mf4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_i8mf4_m))) +vint8mf4_t vsll(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_i8mf8))) +vint8mf8_t vsll(vint8mf8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_i8mf8_m))) +vint8mf8_t vsll(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_i16m1))) +vint16m1_t vsll(vint16m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_i16m1_m))) +vint16m1_t vsll(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_i16m2))) +vint16m2_t vsll(vint16m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_i16m2_m))) +vint16m2_t vsll(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_i16m4))) +vint16m4_t vsll(vint16m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_i16m4_m))) +vint16m4_t vsll(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_i16m8))) +vint16m8_t vsll(vint16m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_i16m8_m))) +vint16m8_t vsll(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_i16mf2))) +vint16mf2_t vsll(vint16mf2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_i16mf2_m))) +vint16mf2_t vsll(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_i16mf4))) +vint16mf4_t vsll(vint16mf4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_i16mf4_m))) +vint16mf4_t vsll(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_i32m1))) +vint32m1_t vsll(vint32m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_i32m1_m))) +vint32m1_t vsll(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_i32m2))) +vint32m2_t vsll(vint32m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_i32m2_m))) +vint32m2_t vsll(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_i32m4))) +vint32m4_t vsll(vint32m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_i32m4_m))) +vint32m4_t vsll(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_i32m8))) +vint32m8_t vsll(vint32m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_i32m8_m))) +vint32m8_t vsll(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_i32mf2))) +vint32mf2_t vsll(vint32mf2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_i32mf2_m))) +vint32mf2_t vsll(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_i64m1))) +vint64m1_t vsll(vint64m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_i64m1_m))) +vint64m1_t vsll(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_i64m2))) +vint64m2_t vsll(vint64m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_i64m2_m))) +vint64m2_t vsll(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_i64m4))) +vint64m4_t vsll(vint64m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_i64m4_m))) +vint64m4_t vsll(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_i64m8))) +vint64m8_t vsll(vint64m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_i64m8_m))) +vint64m8_t vsll(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_u8m1))) +vuint8m1_t vsll(vuint8m1_t op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_u8m1_m))) +vuint8m1_t vsll(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_u8m2))) +vuint8m2_t vsll(vuint8m2_t op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_u8m2_m))) +vuint8m2_t vsll(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_u8m4))) +vuint8m4_t vsll(vuint8m4_t op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_u8m4_m))) +vuint8m4_t vsll(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_u8m8))) +vuint8m8_t vsll(vuint8m8_t op0, vuint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_u8m8_m))) +vuint8m8_t vsll(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, vuint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_u8mf2))) +vuint8mf2_t vsll(vuint8mf2_t op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_u8mf2_m))) +vuint8mf2_t vsll(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_u8mf4))) +vuint8mf4_t vsll(vuint8mf4_t op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_u8mf4_m))) +vuint8mf4_t vsll(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_u8mf8))) +vuint8mf8_t vsll(vuint8mf8_t op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_u8mf8_m))) +vuint8mf8_t vsll(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_u16m1))) +vuint16m1_t vsll(vuint16m1_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_u16m1_m))) +vuint16m1_t vsll(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_u16m2))) +vuint16m2_t vsll(vuint16m2_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_u16m2_m))) +vuint16m2_t vsll(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_u16m4))) +vuint16m4_t vsll(vuint16m4_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_u16m4_m))) +vuint16m4_t vsll(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_u16m8))) +vuint16m8_t vsll(vuint16m8_t op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_u16m8_m))) +vuint16m8_t vsll(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_u16mf2))) +vuint16mf2_t vsll(vuint16mf2_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_u16mf2_m))) +vuint16mf2_t vsll(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_u16mf4))) +vuint16mf4_t vsll(vuint16mf4_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_u16mf4_m))) +vuint16mf4_t vsll(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_u32m1))) +vuint32m1_t vsll(vuint32m1_t op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_u32m1_m))) +vuint32m1_t vsll(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_u32m2))) +vuint32m2_t vsll(vuint32m2_t op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_u32m2_m))) +vuint32m2_t vsll(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_u32m4))) +vuint32m4_t vsll(vuint32m4_t op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_u32m4_m))) +vuint32m4_t vsll(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_u32m8))) +vuint32m8_t vsll(vuint32m8_t op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_u32m8_m))) +vuint32m8_t vsll(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_u32mf2))) +vuint32mf2_t vsll(vuint32mf2_t op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_u32mf2_m))) +vuint32mf2_t vsll(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_u64m1))) +vuint64m1_t vsll(vuint64m1_t op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_u64m1_m))) +vuint64m1_t vsll(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_u64m2))) +vuint64m2_t vsll(vuint64m2_t op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_u64m2_m))) +vuint64m2_t vsll(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_u64m4))) +vuint64m4_t vsll(vuint64m4_t op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_u64m4_m))) +vuint64m4_t vsll(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_u64m8))) +vuint64m8_t vsll(vuint64m8_t op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vv_u64m8_m))) +vuint64m8_t vsll(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_u8m1))) +vuint8m1_t vsll(vuint8m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_u8m1_m))) +vuint8m1_t vsll(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_u8m2))) +vuint8m2_t vsll(vuint8m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_u8m2_m))) +vuint8m2_t vsll(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_u8m4))) +vuint8m4_t vsll(vuint8m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_u8m4_m))) +vuint8m4_t vsll(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_u8m8))) +vuint8m8_t vsll(vuint8m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_u8m8_m))) +vuint8m8_t vsll(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_u8mf2))) +vuint8mf2_t vsll(vuint8mf2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_u8mf2_m))) +vuint8mf2_t vsll(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_u8mf4))) +vuint8mf4_t vsll(vuint8mf4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_u8mf4_m))) +vuint8mf4_t vsll(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_u8mf8))) +vuint8mf8_t vsll(vuint8mf8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_u8mf8_m))) +vuint8mf8_t vsll(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_u16m1))) +vuint16m1_t vsll(vuint16m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_u16m1_m))) +vuint16m1_t vsll(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_u16m2))) +vuint16m2_t vsll(vuint16m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_u16m2_m))) +vuint16m2_t vsll(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_u16m4))) +vuint16m4_t vsll(vuint16m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_u16m4_m))) +vuint16m4_t vsll(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_u16m8))) +vuint16m8_t vsll(vuint16m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_u16m8_m))) +vuint16m8_t vsll(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_u16mf2))) +vuint16mf2_t vsll(vuint16mf2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_u16mf2_m))) +vuint16mf2_t vsll(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_u16mf4))) +vuint16mf4_t vsll(vuint16mf4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_u16mf4_m))) +vuint16mf4_t vsll(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_u32m1))) +vuint32m1_t vsll(vuint32m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_u32m1_m))) +vuint32m1_t vsll(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_u32m2))) +vuint32m2_t vsll(vuint32m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_u32m2_m))) +vuint32m2_t vsll(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_u32m4))) +vuint32m4_t vsll(vuint32m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_u32m4_m))) +vuint32m4_t vsll(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_u32m8))) +vuint32m8_t vsll(vuint32m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_u32m8_m))) +vuint32m8_t vsll(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_u32mf2))) +vuint32mf2_t vsll(vuint32mf2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_u32mf2_m))) +vuint32mf2_t vsll(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_u64m1))) +vuint64m1_t vsll(vuint64m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_u64m1_m))) +vuint64m1_t vsll(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_u64m2))) +vuint64m2_t vsll(vuint64m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_u64m2_m))) +vuint64m2_t vsll(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_u64m4))) +vuint64m4_t vsll(vuint64m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_u64m4_m))) +vuint64m4_t vsll(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_u64m8))) +vuint64m8_t vsll(vuint64m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsll_vx_u64m8_m))) +vuint64m8_t vsll(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_u16m1))) +void vsuxei16(uint16_t * op0, vuint16m1_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_u16m1_m))) +void vsuxei16(vbool16_t op0, uint16_t * op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_u16m2))) +void vsuxei16(uint16_t * op0, vuint16m2_t op1, vuint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_u16m2_m))) +void vsuxei16(vbool8_t op0, uint16_t * op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_u16m4))) +void vsuxei16(uint16_t * op0, vuint16m4_t op1, vuint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_u16m4_m))) +void vsuxei16(vbool4_t op0, uint16_t * op1, vuint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_u16m8))) +void vsuxei16(uint16_t * op0, vuint16m8_t op1, vuint16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_u16m8_m))) +void vsuxei16(vbool2_t op0, uint16_t * op1, vuint16m8_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_u16mf2))) +void vsuxei16(uint16_t * op0, vuint16mf2_t op1, vuint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_u16mf2_m))) +void vsuxei16(vbool32_t op0, uint16_t * op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_u16mf4))) +void vsuxei16(uint16_t * op0, vuint16mf4_t op1, vuint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_u16mf4_m))) +void vsuxei16(vbool64_t op0, uint16_t * op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vv_u8m1))) +vuint8m1_t vsrl(vuint8m1_t op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vv_u8m1_m))) +vuint8m1_t vsrl(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vv_u8m2))) +vuint8m2_t vsrl(vuint8m2_t op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vv_u8m2_m))) +vuint8m2_t vsrl(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vv_u8m4))) +vuint8m4_t vsrl(vuint8m4_t op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vv_u8m4_m))) +vuint8m4_t vsrl(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vv_u8m8))) +vuint8m8_t vsrl(vuint8m8_t op0, vuint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vv_u8m8_m))) +vuint8m8_t vsrl(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, vuint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vv_u8mf2))) +vuint8mf2_t vsrl(vuint8mf2_t op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vv_u8mf2_m))) +vuint8mf2_t vsrl(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vv_u8mf4))) +vuint8mf4_t vsrl(vuint8mf4_t op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vv_u8mf4_m))) +vuint8mf4_t vsrl(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vv_u8mf8))) +vuint8mf8_t vsrl(vuint8mf8_t op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vv_u8mf8_m))) +vuint8mf8_t vsrl(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vv_u16m1))) +vuint16m1_t vsrl(vuint16m1_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vv_u16m1_m))) +vuint16m1_t vsrl(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vv_u16m2))) +vuint16m2_t vsrl(vuint16m2_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vv_u16m2_m))) +vuint16m2_t vsrl(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vv_u16m4))) +vuint16m4_t vsrl(vuint16m4_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vv_u16m4_m))) +vuint16m4_t vsrl(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vv_u16m8))) +vuint16m8_t vsrl(vuint16m8_t op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vv_u16m8_m))) +vuint16m8_t vsrl(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vv_u16mf2))) +vuint16mf2_t vsrl(vuint16mf2_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vv_u16mf2_m))) +vuint16mf2_t vsrl(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vv_u16mf4))) +vuint16mf4_t vsrl(vuint16mf4_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vv_u16mf4_m))) +vuint16mf4_t vsrl(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vv_u32m1))) +vuint32m1_t vsrl(vuint32m1_t op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vv_u32m1_m))) +vuint32m1_t vsrl(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vv_u32m2))) +vuint32m2_t vsrl(vuint32m2_t op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vv_u32m2_m))) +vuint32m2_t vsrl(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vv_u32m4))) +vuint32m4_t vsrl(vuint32m4_t op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vv_u32m4_m))) +vuint32m4_t vsrl(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vv_u32m8))) +vuint32m8_t vsrl(vuint32m8_t op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vv_u32m8_m))) +vuint32m8_t vsrl(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vv_u32mf2))) +vuint32mf2_t vsrl(vuint32mf2_t op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vv_u32mf2_m))) +vuint32mf2_t vsrl(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vv_u64m1))) +vuint64m1_t vsrl(vuint64m1_t op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vv_u64m1_m))) +vuint64m1_t vsrl(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vv_u64m2))) +vuint64m2_t vsrl(vuint64m2_t op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vv_u64m2_m))) +vuint64m2_t vsrl(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vv_u64m4))) +vuint64m4_t vsrl(vuint64m4_t op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vv_u64m4_m))) +vuint64m4_t vsrl(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vv_u64m8))) +vuint64m8_t vsrl(vuint64m8_t op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vv_u64m8_m))) +vuint64m8_t vsrl(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vx_u8m1))) +vuint8m1_t vsrl(vuint8m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vx_u8m1_m))) +vuint8m1_t vsrl(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vx_u8m2))) +vuint8m2_t vsrl(vuint8m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vx_u8m2_m))) +vuint8m2_t vsrl(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vx_u8m4))) +vuint8m4_t vsrl(vuint8m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vx_u8m4_m))) +vuint8m4_t vsrl(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vx_u8m8))) +vuint8m8_t vsrl(vuint8m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vx_u8m8_m))) +vuint8m8_t vsrl(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vx_u8mf2))) +vuint8mf2_t vsrl(vuint8mf2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vx_u8mf2_m))) +vuint8mf2_t vsrl(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vx_u8mf4))) +vuint8mf4_t vsrl(vuint8mf4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vx_u8mf4_m))) +vuint8mf4_t vsrl(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vx_u8mf8))) +vuint8mf8_t vsrl(vuint8mf8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vx_u8mf8_m))) +vuint8mf8_t vsrl(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vx_u16m1))) +vuint16m1_t vsrl(vuint16m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vx_u16m1_m))) +vuint16m1_t vsrl(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vx_u16m2))) +vuint16m2_t vsrl(vuint16m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vx_u16m2_m))) +vuint16m2_t vsrl(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vx_u16m4))) +vuint16m4_t vsrl(vuint16m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vx_u16m4_m))) +vuint16m4_t vsrl(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vx_u16m8))) +vuint16m8_t vsrl(vuint16m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vx_u16m8_m))) +vuint16m8_t vsrl(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vx_u16mf2))) +vuint16mf2_t vsrl(vuint16mf2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vx_u16mf2_m))) +vuint16mf2_t vsrl(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vx_u16mf4))) +vuint16mf4_t vsrl(vuint16mf4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vx_u16mf4_m))) +vuint16mf4_t vsrl(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vx_u32m1))) +vuint32m1_t vsrl(vuint32m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vx_u32m1_m))) +vuint32m1_t vsrl(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vx_u32m2))) +vuint32m2_t vsrl(vuint32m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vx_u32m2_m))) +vuint32m2_t vsrl(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vx_u32m4))) +vuint32m4_t vsrl(vuint32m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vx_u32m4_m))) +vuint32m4_t vsrl(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vx_u32m8))) +vuint32m8_t vsrl(vuint32m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vx_u32m8_m))) +vuint32m8_t vsrl(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vx_u32mf2))) +vuint32mf2_t vsrl(vuint32mf2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vx_u32mf2_m))) +vuint32mf2_t vsrl(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vx_u64m1))) +vuint64m1_t vsrl(vuint64m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vx_u64m1_m))) +vuint64m1_t vsrl(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vx_u64m2))) +vuint64m2_t vsrl(vuint64m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vx_u64m2_m))) +vuint64m2_t vsrl(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vx_u64m4))) +vuint64m4_t vsrl(vuint64m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vx_u64m4_m))) +vuint64m4_t vsrl(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vx_u64m8))) +vuint64m8_t vsrl(vuint64m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsrl_vx_u64m8_m))) +vuint64m8_t vsrl(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vv_i8m1))) +vint8m1_t vsra(vint8m1_t op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vv_i8m1_m))) +vint8m1_t vsra(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vv_i8m2))) +vint8m2_t vsra(vint8m2_t op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vv_i8m2_m))) +vint8m2_t vsra(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vv_i8m4))) +vint8m4_t vsra(vint8m4_t op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vv_i8m4_m))) +vint8m4_t vsra(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vv_i8m8))) +vint8m8_t vsra(vint8m8_t op0, vuint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vv_i8m8_m))) +vint8m8_t vsra(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, vuint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vv_i8mf2))) +vint8mf2_t vsra(vint8mf2_t op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vv_i8mf2_m))) +vint8mf2_t vsra(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vv_i8mf4))) +vint8mf4_t vsra(vint8mf4_t op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vv_i8mf4_m))) +vint8mf4_t vsra(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vv_i8mf8))) +vint8mf8_t vsra(vint8mf8_t op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vv_i8mf8_m))) +vint8mf8_t vsra(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vv_i16m1))) +vint16m1_t vsra(vint16m1_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vv_i16m1_m))) +vint16m1_t vsra(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vv_i16m2))) +vint16m2_t vsra(vint16m2_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vv_i16m2_m))) +vint16m2_t vsra(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vv_i16m4))) +vint16m4_t vsra(vint16m4_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vv_i16m4_m))) +vint16m4_t vsra(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vv_i16m8))) +vint16m8_t vsra(vint16m8_t op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vv_i16m8_m))) +vint16m8_t vsra(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vv_i16mf2))) +vint16mf2_t vsra(vint16mf2_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vv_i16mf2_m))) +vint16mf2_t vsra(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vv_i16mf4))) +vint16mf4_t vsra(vint16mf4_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vv_i16mf4_m))) +vint16mf4_t vsra(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vv_i32m1))) +vint32m1_t vsra(vint32m1_t op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vv_i32m1_m))) +vint32m1_t vsra(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vv_i32m2))) +vint32m2_t vsra(vint32m2_t op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vv_i32m2_m))) +vint32m2_t vsra(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vv_i32m4))) +vint32m4_t vsra(vint32m4_t op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vv_i32m4_m))) +vint32m4_t vsra(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vv_i32m8))) +vint32m8_t vsra(vint32m8_t op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vv_i32m8_m))) +vint32m8_t vsra(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vv_i32mf2))) +vint32mf2_t vsra(vint32mf2_t op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vv_i32mf2_m))) +vint32mf2_t vsra(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vv_i64m1))) +vint64m1_t vsra(vint64m1_t op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vv_i64m1_m))) +vint64m1_t vsra(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vv_i64m2))) +vint64m2_t vsra(vint64m2_t op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vv_i64m2_m))) +vint64m2_t vsra(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vv_i64m4))) +vint64m4_t vsra(vint64m4_t op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vv_i64m4_m))) +vint64m4_t vsra(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vv_i64m8))) +vint64m8_t vsra(vint64m8_t op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vv_i64m8_m))) +vint64m8_t vsra(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vx_i8m1))) +vint8m1_t vsra(vint8m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vx_i8m1_m))) +vint8m1_t vsra(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vx_i8m2))) +vint8m2_t vsra(vint8m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vx_i8m2_m))) +vint8m2_t vsra(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vx_i8m4))) +vint8m4_t vsra(vint8m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vx_i8m4_m))) +vint8m4_t vsra(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vx_i8m8))) +vint8m8_t vsra(vint8m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vx_i8m8_m))) +vint8m8_t vsra(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vx_i8mf2))) +vint8mf2_t vsra(vint8mf2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vx_i8mf2_m))) +vint8mf2_t vsra(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vx_i8mf4))) +vint8mf4_t vsra(vint8mf4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vx_i8mf4_m))) +vint8mf4_t vsra(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vx_i8mf8))) +vint8mf8_t vsra(vint8mf8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vx_i8mf8_m))) +vint8mf8_t vsra(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vx_i16m1))) +vint16m1_t vsra(vint16m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vx_i16m1_m))) +vint16m1_t vsra(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vx_i16m2))) +vint16m2_t vsra(vint16m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vx_i16m2_m))) +vint16m2_t vsra(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vx_i16m4))) +vint16m4_t vsra(vint16m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vx_i16m4_m))) +vint16m4_t vsra(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vx_i16m8))) +vint16m8_t vsra(vint16m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vx_i16m8_m))) +vint16m8_t vsra(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vx_i16mf2))) +vint16mf2_t vsra(vint16mf2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vx_i16mf2_m))) +vint16mf2_t vsra(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vx_i16mf4))) +vint16mf4_t vsra(vint16mf4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vx_i16mf4_m))) +vint16mf4_t vsra(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vx_i32m1))) +vint32m1_t vsra(vint32m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vx_i32m1_m))) +vint32m1_t vsra(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vx_i32m2))) +vint32m2_t vsra(vint32m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vx_i32m2_m))) +vint32m2_t vsra(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vx_i32m4))) +vint32m4_t vsra(vint32m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vx_i32m4_m))) +vint32m4_t vsra(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vx_i32m8))) +vint32m8_t vsra(vint32m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vx_i32m8_m))) +vint32m8_t vsra(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vx_i32mf2))) +vint32mf2_t vsra(vint32mf2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vx_i32mf2_m))) +vint32mf2_t vsra(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vx_i64m1))) +vint64m1_t vsra(vint64m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vx_i64m1_m))) +vint64m1_t vsra(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vx_i64m2))) +vint64m2_t vsra(vint64m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vx_i64m2_m))) +vint64m2_t vsra(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vx_i64m4))) +vint64m4_t vsra(vint64m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vx_i64m4_m))) +vint64m4_t vsra(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vx_i64m8))) +vint64m8_t vsra(vint64m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsra_vx_i64m8_m))) +vint64m8_t vsra(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wv_u8m1))) +vuint8m1_t vnsrl(vuint16m2_t op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wv_u8m1_m))) +vuint8m1_t vnsrl(vbool8_t op0, vuint8m1_t op1, vuint16m2_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wv_u8m2))) +vuint8m2_t vnsrl(vuint16m4_t op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wv_u8m2_m))) +vuint8m2_t vnsrl(vbool4_t op0, vuint8m2_t op1, vuint16m4_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wv_u8m4))) +vuint8m4_t vnsrl(vuint16m8_t op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wv_u8m4_m))) +vuint8m4_t vnsrl(vbool2_t op0, vuint8m4_t op1, vuint16m8_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wv_u8mf2))) +vuint8mf2_t vnsrl(vuint16m1_t op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wv_u8mf2_m))) +vuint8mf2_t vnsrl(vbool16_t op0, vuint8mf2_t op1, vuint16m1_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wv_u8mf4))) +vuint8mf4_t vnsrl(vuint16mf2_t op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wv_u8mf4_m))) +vuint8mf4_t vnsrl(vbool32_t op0, vuint8mf4_t op1, vuint16mf2_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wv_u8mf8))) +vuint8mf8_t vnsrl(vuint16mf4_t op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wv_u8mf8_m))) +vuint8mf8_t vnsrl(vbool64_t op0, vuint8mf8_t op1, vuint16mf4_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wv_u16m1))) +vuint16m1_t vnsrl(vuint32m2_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wv_u16m1_m))) +vuint16m1_t vnsrl(vbool16_t op0, vuint16m1_t op1, vuint32m2_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wv_u16m2))) +vuint16m2_t vnsrl(vuint32m4_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wv_u16m2_m))) +vuint16m2_t vnsrl(vbool8_t op0, vuint16m2_t op1, vuint32m4_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wv_u16m4))) +vuint16m4_t vnsrl(vuint32m8_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wv_u16m4_m))) +vuint16m4_t vnsrl(vbool4_t op0, vuint16m4_t op1, vuint32m8_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wv_u16mf2))) +vuint16mf2_t vnsrl(vuint32m1_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wv_u16mf2_m))) +vuint16mf2_t vnsrl(vbool32_t op0, vuint16mf2_t op1, vuint32m1_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wv_u16mf4))) +vuint16mf4_t vnsrl(vuint32mf2_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wv_u16mf4_m))) +vuint16mf4_t vnsrl(vbool64_t op0, vuint16mf4_t op1, vuint32mf2_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wv_u32m1))) +vuint32m1_t vnsrl(vuint64m2_t op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wv_u32m1_m))) +vuint32m1_t vnsrl(vbool32_t op0, vuint32m1_t op1, vuint64m2_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wv_u32m2))) +vuint32m2_t vnsrl(vuint64m4_t op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wv_u32m2_m))) +vuint32m2_t vnsrl(vbool16_t op0, vuint32m2_t op1, vuint64m4_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wv_u32m4))) +vuint32m4_t vnsrl(vuint64m8_t op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wv_u32m4_m))) +vuint32m4_t vnsrl(vbool8_t op0, vuint32m4_t op1, vuint64m8_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wv_u32mf2))) +vuint32mf2_t vnsrl(vuint64m1_t op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wv_u32mf2_m))) +vuint32mf2_t vnsrl(vbool64_t op0, vuint32mf2_t op1, vuint64m1_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wx_u8m1))) +vuint8m1_t vnsrl(vuint16m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wx_u8m1_m))) +vuint8m1_t vnsrl(vbool8_t op0, vuint8m1_t op1, vuint16m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wx_u8m2))) +vuint8m2_t vnsrl(vuint16m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wx_u8m2_m))) +vuint8m2_t vnsrl(vbool4_t op0, vuint8m2_t op1, vuint16m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wx_u8m4))) +vuint8m4_t vnsrl(vuint16m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wx_u8m4_m))) +vuint8m4_t vnsrl(vbool2_t op0, vuint8m4_t op1, vuint16m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wx_u8mf2))) +vuint8mf2_t vnsrl(vuint16m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wx_u8mf2_m))) +vuint8mf2_t vnsrl(vbool16_t op0, vuint8mf2_t op1, vuint16m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wx_u8mf4))) +vuint8mf4_t vnsrl(vuint16mf2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wx_u8mf4_m))) +vuint8mf4_t vnsrl(vbool32_t op0, vuint8mf4_t op1, vuint16mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wx_u8mf8))) +vuint8mf8_t vnsrl(vuint16mf4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wx_u8mf8_m))) +vuint8mf8_t vnsrl(vbool64_t op0, vuint8mf8_t op1, vuint16mf4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wx_u16m1))) +vuint16m1_t vnsrl(vuint32m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wx_u16m1_m))) +vuint16m1_t vnsrl(vbool16_t op0, vuint16m1_t op1, vuint32m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wx_u16m2))) +vuint16m2_t vnsrl(vuint32m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wx_u16m2_m))) +vuint16m2_t vnsrl(vbool8_t op0, vuint16m2_t op1, vuint32m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wx_u16m4))) +vuint16m4_t vnsrl(vuint32m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wx_u16m4_m))) +vuint16m4_t vnsrl(vbool4_t op0, vuint16m4_t op1, vuint32m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wx_u16mf2))) +vuint16mf2_t vnsrl(vuint32m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wx_u16mf2_m))) +vuint16mf2_t vnsrl(vbool32_t op0, vuint16mf2_t op1, vuint32m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wx_u16mf4))) +vuint16mf4_t vnsrl(vuint32mf2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wx_u16mf4_m))) +vuint16mf4_t vnsrl(vbool64_t op0, vuint16mf4_t op1, vuint32mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wx_u32m1))) +vuint32m1_t vnsrl(vuint64m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wx_u32m1_m))) +vuint32m1_t vnsrl(vbool32_t op0, vuint32m1_t op1, vuint64m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wx_u32m2))) +vuint32m2_t vnsrl(vuint64m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wx_u32m2_m))) +vuint32m2_t vnsrl(vbool16_t op0, vuint32m2_t op1, vuint64m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wx_u32m4))) +vuint32m4_t vnsrl(vuint64m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wx_u32m4_m))) +vuint32m4_t vnsrl(vbool8_t op0, vuint32m4_t op1, vuint64m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wx_u32mf2))) +vuint32mf2_t vnsrl(vuint64m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsrl_wx_u32mf2_m))) +vuint32mf2_t vnsrl(vbool64_t op0, vuint32mf2_t op1, vuint64m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wv_i8m1))) +vint8m1_t vnsra(vint16m2_t op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wv_i8m1_m))) +vint8m1_t vnsra(vbool8_t op0, vint8m1_t op1, vint16m2_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wv_i8m2))) +vint8m2_t vnsra(vint16m4_t op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wv_i8m2_m))) +vint8m2_t vnsra(vbool4_t op0, vint8m2_t op1, vint16m4_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wv_i8m4))) +vint8m4_t vnsra(vint16m8_t op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wv_i8m4_m))) +vint8m4_t vnsra(vbool2_t op0, vint8m4_t op1, vint16m8_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wv_i8mf2))) +vint8mf2_t vnsra(vint16m1_t op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wv_i8mf2_m))) +vint8mf2_t vnsra(vbool16_t op0, vint8mf2_t op1, vint16m1_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wv_i8mf4))) +vint8mf4_t vnsra(vint16mf2_t op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wv_i8mf4_m))) +vint8mf4_t vnsra(vbool32_t op0, vint8mf4_t op1, vint16mf2_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wv_i8mf8))) +vint8mf8_t vnsra(vint16mf4_t op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wv_i8mf8_m))) +vint8mf8_t vnsra(vbool64_t op0, vint8mf8_t op1, vint16mf4_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wv_i16m1))) +vint16m1_t vnsra(vint32m2_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wv_i16m1_m))) +vint16m1_t vnsra(vbool16_t op0, vint16m1_t op1, vint32m2_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wv_i16m2))) +vint16m2_t vnsra(vint32m4_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wv_i16m2_m))) +vint16m2_t vnsra(vbool8_t op0, vint16m2_t op1, vint32m4_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wv_i16m4))) +vint16m4_t vnsra(vint32m8_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wv_i16m4_m))) +vint16m4_t vnsra(vbool4_t op0, vint16m4_t op1, vint32m8_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wv_i16mf2))) +vint16mf2_t vnsra(vint32m1_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wv_i16mf2_m))) +vint16mf2_t vnsra(vbool32_t op0, vint16mf2_t op1, vint32m1_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wv_i16mf4))) +vint16mf4_t vnsra(vint32mf2_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wv_i16mf4_m))) +vint16mf4_t vnsra(vbool64_t op0, vint16mf4_t op1, vint32mf2_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wv_i32m1))) +vint32m1_t vnsra(vint64m2_t op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wv_i32m1_m))) +vint32m1_t vnsra(vbool32_t op0, vint32m1_t op1, vint64m2_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wv_i32m2))) +vint32m2_t vnsra(vint64m4_t op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wv_i32m2_m))) +vint32m2_t vnsra(vbool16_t op0, vint32m2_t op1, vint64m4_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wv_i32m4))) +vint32m4_t vnsra(vint64m8_t op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wv_i32m4_m))) +vint32m4_t vnsra(vbool8_t op0, vint32m4_t op1, vint64m8_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wv_i32mf2))) +vint32mf2_t vnsra(vint64m1_t op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wv_i32mf2_m))) +vint32mf2_t vnsra(vbool64_t op0, vint32mf2_t op1, vint64m1_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wx_i8m1))) +vint8m1_t vnsra(vint16m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wx_i8m1_m))) +vint8m1_t vnsra(vbool8_t op0, vint8m1_t op1, vint16m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wx_i8m2))) +vint8m2_t vnsra(vint16m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wx_i8m2_m))) +vint8m2_t vnsra(vbool4_t op0, vint8m2_t op1, vint16m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wx_i8m4))) +vint8m4_t vnsra(vint16m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wx_i8m4_m))) +vint8m4_t vnsra(vbool2_t op0, vint8m4_t op1, vint16m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wx_i8mf2))) +vint8mf2_t vnsra(vint16m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wx_i8mf2_m))) +vint8mf2_t vnsra(vbool16_t op0, vint8mf2_t op1, vint16m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wx_i8mf4))) +vint8mf4_t vnsra(vint16mf2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wx_i8mf4_m))) +vint8mf4_t vnsra(vbool32_t op0, vint8mf4_t op1, vint16mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wx_i8mf8))) +vint8mf8_t vnsra(vint16mf4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wx_i8mf8_m))) +vint8mf8_t vnsra(vbool64_t op0, vint8mf8_t op1, vint16mf4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wx_i16m1))) +vint16m1_t vnsra(vint32m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wx_i16m1_m))) +vint16m1_t vnsra(vbool16_t op0, vint16m1_t op1, vint32m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wx_i16m2))) +vint16m2_t vnsra(vint32m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wx_i16m2_m))) +vint16m2_t vnsra(vbool8_t op0, vint16m2_t op1, vint32m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wx_i16m4))) +vint16m4_t vnsra(vint32m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wx_i16m4_m))) +vint16m4_t vnsra(vbool4_t op0, vint16m4_t op1, vint32m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wx_i16mf2))) +vint16mf2_t vnsra(vint32m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wx_i16mf2_m))) +vint16mf2_t vnsra(vbool32_t op0, vint16mf2_t op1, vint32m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wx_i16mf4))) +vint16mf4_t vnsra(vint32mf2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wx_i16mf4_m))) +vint16mf4_t vnsra(vbool64_t op0, vint16mf4_t op1, vint32mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wx_i32m1))) +vint32m1_t vnsra(vint64m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wx_i32m1_m))) +vint32m1_t vnsra(vbool32_t op0, vint32m1_t op1, vint64m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wx_i32m2))) +vint32m2_t vnsra(vint64m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wx_i32m2_m))) +vint32m2_t vnsra(vbool16_t op0, vint32m2_t op1, vint64m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wx_i32m4))) +vint32m4_t vnsra(vint64m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wx_i32m4_m))) +vint32m4_t vnsra(vbool8_t op0, vint32m4_t op1, vint64m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wx_i32mf2))) +vint32mf2_t vnsra(vint64m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnsra_wx_i32mf2_m))) +vint32mf2_t vnsra(vbool64_t op0, vint32mf2_t op1, vint64m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_u8m1))) +vuint8m1_t vncvt_x(vuint16m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_u8m1_m))) +vuint8m1_t vncvt_x(vbool8_t op0, vuint8m1_t op1, vuint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_u8m2))) +vuint8m2_t vncvt_x(vuint16m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_u8m2_m))) +vuint8m2_t vncvt_x(vbool4_t op0, vuint8m2_t op1, vuint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_u8m4))) +vuint8m4_t vncvt_x(vuint16m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_u8m4_m))) +vuint8m4_t vncvt_x(vbool2_t op0, vuint8m4_t op1, vuint16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_u8mf2))) +vuint8mf2_t vncvt_x(vuint16m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_u8mf2_m))) +vuint8mf2_t vncvt_x(vbool16_t op0, vuint8mf2_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_u8mf4))) +vuint8mf4_t vncvt_x(vuint16mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_u8mf4_m))) +vuint8mf4_t vncvt_x(vbool32_t op0, vuint8mf4_t op1, vuint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_u8mf8))) +vuint8mf8_t vncvt_x(vuint16mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_u8mf8_m))) +vuint8mf8_t vncvt_x(vbool64_t op0, vuint8mf8_t op1, vuint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_u16m1))) +vuint16m1_t vncvt_x(vuint32m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_u16m1_m))) +vuint16m1_t vncvt_x(vbool16_t op0, vuint16m1_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_u16m2))) +vuint16m2_t vncvt_x(vuint32m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_u16m2_m))) +vuint16m2_t vncvt_x(vbool8_t op0, vuint16m2_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_u16m4))) +vuint16m4_t vncvt_x(vuint32m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_u16m4_m))) +vuint16m4_t vncvt_x(vbool4_t op0, vuint16m4_t op1, vuint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_u16mf2))) +vuint16mf2_t vncvt_x(vuint32m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_u16mf2_m))) +vuint16mf2_t vncvt_x(vbool32_t op0, vuint16mf2_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_u16mf4))) +vuint16mf4_t vncvt_x(vuint32mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_u16mf4_m))) +vuint16mf4_t vncvt_x(vbool64_t op0, vuint16mf4_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_u32m1))) +vuint32m1_t vncvt_x(vuint64m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_u32m1_m))) +vuint32m1_t vncvt_x(vbool32_t op0, vuint32m1_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_u32m2))) +vuint32m2_t vncvt_x(vuint64m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_u32m2_m))) +vuint32m2_t vncvt_x(vbool16_t op0, vuint32m2_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_u32m4))) +vuint32m4_t vncvt_x(vuint64m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_u32m4_m))) +vuint32m4_t vncvt_x(vbool8_t op0, vuint32m4_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_u32mf2))) +vuint32mf2_t vncvt_x(vuint64m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_u32mf2_m))) +vuint32mf2_t vncvt_x(vbool64_t op0, vuint32mf2_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_i8m1_b8))) +vbool8_t vmseq(vint8m1_t op0, vint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_i8m1_b8_m))) +vbool8_t vmseq(vbool8_t op0, vbool8_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_i8m2_b4))) +vbool4_t vmseq(vint8m2_t op0, vint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_i8m2_b4_m))) +vbool4_t vmseq(vbool4_t op0, vbool4_t op1, vint8m2_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_i8m4_b2))) +vbool2_t vmseq(vint8m4_t op0, vint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_i8m4_b2_m))) +vbool2_t vmseq(vbool2_t op0, vbool2_t op1, vint8m4_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_i8m8_b1))) +vbool1_t vmseq(vint8m8_t op0, vint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_i8m8_b1_m))) +vbool1_t vmseq(vbool1_t op0, vbool1_t op1, vint8m8_t op2, vint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_i8mf2_b16))) +vbool16_t vmseq(vint8mf2_t op0, vint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_i8mf2_b16_m))) +vbool16_t vmseq(vbool16_t op0, vbool16_t op1, vint8mf2_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_i8mf4_b32))) +vbool32_t vmseq(vint8mf4_t op0, vint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_i8mf4_b32_m))) +vbool32_t vmseq(vbool32_t op0, vbool32_t op1, vint8mf4_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_i8mf8_b64))) +vbool64_t vmseq(vint8mf8_t op0, vint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_i8mf8_b64_m))) +vbool64_t vmseq(vbool64_t op0, vbool64_t op1, vint8mf8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_i16m1_b16))) +vbool16_t vmseq(vint16m1_t op0, vint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_i16m1_b16_m))) +vbool16_t vmseq(vbool16_t op0, vbool16_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_i16m2_b8))) +vbool8_t vmseq(vint16m2_t op0, vint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_i16m2_b8_m))) +vbool8_t vmseq(vbool8_t op0, vbool8_t op1, vint16m2_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_i16m4_b4))) +vbool4_t vmseq(vint16m4_t op0, vint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_i16m4_b4_m))) +vbool4_t vmseq(vbool4_t op0, vbool4_t op1, vint16m4_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_i16m8_b2))) +vbool2_t vmseq(vint16m8_t op0, vint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_i16m8_b2_m))) +vbool2_t vmseq(vbool2_t op0, vbool2_t op1, vint16m8_t op2, vint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_i16mf2_b32))) +vbool32_t vmseq(vint16mf2_t op0, vint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_i16mf2_b32_m))) +vbool32_t vmseq(vbool32_t op0, vbool32_t op1, vint16mf2_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_i16mf4_b64))) +vbool64_t vmseq(vint16mf4_t op0, vint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_i16mf4_b64_m))) +vbool64_t vmseq(vbool64_t op0, vbool64_t op1, vint16mf4_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_i32m1_b32))) +vbool32_t vmseq(vint32m1_t op0, vint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_i32m1_b32_m))) +vbool32_t vmseq(vbool32_t op0, vbool32_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_i32m2_b16))) +vbool16_t vmseq(vint32m2_t op0, vint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_i32m2_b16_m))) +vbool16_t vmseq(vbool16_t op0, vbool16_t op1, vint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_i32m4_b8))) +vbool8_t vmseq(vint32m4_t op0, vint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_i32m4_b8_m))) +vbool8_t vmseq(vbool8_t op0, vbool8_t op1, vint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_i32m8_b4))) +vbool4_t vmseq(vint32m8_t op0, vint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_i32m8_b4_m))) +vbool4_t vmseq(vbool4_t op0, vbool4_t op1, vint32m8_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_i32mf2_b64))) +vbool64_t vmseq(vint32mf2_t op0, vint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_i32mf2_b64_m))) +vbool64_t vmseq(vbool64_t op0, vbool64_t op1, vint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_i64m1_b64))) +vbool64_t vmseq(vint64m1_t op0, vint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_i64m1_b64_m))) +vbool64_t vmseq(vbool64_t op0, vbool64_t op1, vint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_i64m2_b32))) +vbool32_t vmseq(vint64m2_t op0, vint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_i64m2_b32_m))) +vbool32_t vmseq(vbool32_t op0, vbool32_t op1, vint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_i64m4_b16))) +vbool16_t vmseq(vint64m4_t op0, vint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_i64m4_b16_m))) +vbool16_t vmseq(vbool16_t op0, vbool16_t op1, vint64m4_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_i64m8_b8))) +vbool8_t vmseq(vint64m8_t op0, vint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_i64m8_b8_m))) +vbool8_t vmseq(vbool8_t op0, vbool8_t op1, vint64m8_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse8_v_u8m1))) +void vsse8(uint8_t * op0, ptrdiff_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse8_v_u8m1_m))) +void vsse8(vbool8_t op0, uint8_t * op1, ptrdiff_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse8_v_u8m2))) +void vsse8(uint8_t * op0, ptrdiff_t op1, vuint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse8_v_u8m2_m))) +void vsse8(vbool4_t op0, uint8_t * op1, ptrdiff_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse8_v_u8m4))) +void vsse8(uint8_t * op0, ptrdiff_t op1, vuint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse8_v_u8m4_m))) +void vsse8(vbool2_t op0, uint8_t * op1, ptrdiff_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse8_v_u8m8))) +void vsse8(uint8_t * op0, ptrdiff_t op1, vuint8m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse8_v_u8m8_m))) +void vsse8(vbool1_t op0, uint8_t * op1, ptrdiff_t op2, vuint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse8_v_u8mf2))) +void vsse8(uint8_t * op0, ptrdiff_t op1, vuint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse8_v_u8mf2_m))) +void vsse8(vbool16_t op0, uint8_t * op1, ptrdiff_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse8_v_u8mf4))) +void vsse8(uint8_t * op0, ptrdiff_t op1, vuint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse8_v_u8mf4_m))) +void vsse8(vbool32_t op0, uint8_t * op1, ptrdiff_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse8_v_u8mf8))) +void vsse8(uint8_t * op0, ptrdiff_t op1, vuint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse8_v_u8mf8_m))) +void vsse8(vbool64_t op0, uint8_t * op1, ptrdiff_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_i16m1))) +void vsuxei32(int16_t * op0, vuint32m2_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_i16m1_m))) +void vsuxei32(vbool16_t op0, int16_t * op1, vuint32m2_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_i16m2))) +void vsuxei32(int16_t * op0, vuint32m4_t op1, vint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_i16m2_m))) +void vsuxei32(vbool8_t op0, int16_t * op1, vuint32m4_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_i16m4))) +void vsuxei32(int16_t * op0, vuint32m8_t op1, vint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_i16m4_m))) +void vsuxei32(vbool4_t op0, int16_t * op1, vuint32m8_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_i16mf2))) +void vsuxei32(int16_t * op0, vuint32m1_t op1, vint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_i16mf2_m))) +void vsuxei32(vbool32_t op0, int16_t * op1, vuint32m1_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_i16mf4))) +void vsuxei32(int16_t * op0, vuint32mf2_t op1, vint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_i16mf4_m))) +void vsuxei32(vbool64_t op0, int16_t * op1, vuint32mf2_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_i8m1_b8))) +vbool8_t vmseq(vint8m1_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_i8m1_b8_m))) +vbool8_t vmseq(vbool8_t op0, vbool8_t op1, vint8m1_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_i8m2_b4))) +vbool4_t vmseq(vint8m2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_i8m2_b4_m))) +vbool4_t vmseq(vbool4_t op0, vbool4_t op1, vint8m2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_i8m4_b2))) +vbool2_t vmseq(vint8m4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_i8m4_b2_m))) +vbool2_t vmseq(vbool2_t op0, vbool2_t op1, vint8m4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_i8m8_b1))) +vbool1_t vmseq(vint8m8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_i8m8_b1_m))) +vbool1_t vmseq(vbool1_t op0, vbool1_t op1, vint8m8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_i8mf2_b16))) +vbool16_t vmseq(vint8mf2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_i8mf2_b16_m))) +vbool16_t vmseq(vbool16_t op0, vbool16_t op1, vint8mf2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_i8mf4_b32))) +vbool32_t vmseq(vint8mf4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_i8mf4_b32_m))) +vbool32_t vmseq(vbool32_t op0, vbool32_t op1, vint8mf4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_i8mf8_b64))) +vbool64_t vmseq(vint8mf8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_i8mf8_b64_m))) +vbool64_t vmseq(vbool64_t op0, vbool64_t op1, vint8mf8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_i16m1_b16))) +vbool16_t vmseq(vint16m1_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_i16m1_b16_m))) +vbool16_t vmseq(vbool16_t op0, vbool16_t op1, vint16m1_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_i16m2_b8))) +vbool8_t vmseq(vint16m2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_i16m2_b8_m))) +vbool8_t vmseq(vbool8_t op0, vbool8_t op1, vint16m2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_i16m4_b4))) +vbool4_t vmseq(vint16m4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_i16m4_b4_m))) +vbool4_t vmseq(vbool4_t op0, vbool4_t op1, vint16m4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_i16m8_b2))) +vbool2_t vmseq(vint16m8_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_i16m8_b2_m))) +vbool2_t vmseq(vbool2_t op0, vbool2_t op1, vint16m8_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_i16mf2_b32))) +vbool32_t vmseq(vint16mf2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_i16mf2_b32_m))) +vbool32_t vmseq(vbool32_t op0, vbool32_t op1, vint16mf2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_i16mf4_b64))) +vbool64_t vmseq(vint16mf4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_i16mf4_b64_m))) +vbool64_t vmseq(vbool64_t op0, vbool64_t op1, vint16mf4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_i32m1_b32))) +vbool32_t vmseq(vint32m1_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_i32m1_b32_m))) +vbool32_t vmseq(vbool32_t op0, vbool32_t op1, vint32m1_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_i32m2_b16))) +vbool16_t vmseq(vint32m2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_i32m2_b16_m))) +vbool16_t vmseq(vbool16_t op0, vbool16_t op1, vint32m2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_i32m4_b8))) +vbool8_t vmseq(vint32m4_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_i32m4_b8_m))) +vbool8_t vmseq(vbool8_t op0, vbool8_t op1, vint32m4_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_i32m8_b4))) +vbool4_t vmseq(vint32m8_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_i32m8_b4_m))) +vbool4_t vmseq(vbool4_t op0, vbool4_t op1, vint32m8_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_i32mf2_b64))) +vbool64_t vmseq(vint32mf2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_i32mf2_b64_m))) +vbool64_t vmseq(vbool64_t op0, vbool64_t op1, vint32mf2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_i64m1_b64))) +vbool64_t vmseq(vint64m1_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_i64m1_b64_m))) +vbool64_t vmseq(vbool64_t op0, vbool64_t op1, vint64m1_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_i64m2_b32))) +vbool32_t vmseq(vint64m2_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_i64m2_b32_m))) +vbool32_t vmseq(vbool32_t op0, vbool32_t op1, vint64m2_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_i64m4_b16))) +vbool16_t vmseq(vint64m4_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_i64m4_b16_m))) +vbool16_t vmseq(vbool16_t op0, vbool16_t op1, vint64m4_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_i64m8_b8))) +vbool8_t vmseq(vint64m8_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_i64m8_b8_m))) +vbool8_t vmseq(vbool8_t op0, vbool8_t op1, vint64m8_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_u8m1_b8))) +vbool8_t vmseq(vuint8m1_t op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_u8m1_b8_m))) +vbool8_t vmseq(vbool8_t op0, vbool8_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_u8m2_b4))) +vbool4_t vmseq(vuint8m2_t op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_u8m2_b4_m))) +vbool4_t vmseq(vbool4_t op0, vbool4_t op1, vuint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_u8m4_b2))) +vbool2_t vmseq(vuint8m4_t op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_u8m4_b2_m))) +vbool2_t vmseq(vbool2_t op0, vbool2_t op1, vuint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_u8m8_b1))) +vbool1_t vmseq(vuint8m8_t op0, vuint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_u8m8_b1_m))) +vbool1_t vmseq(vbool1_t op0, vbool1_t op1, vuint8m8_t op2, vuint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_u8mf2_b16))) +vbool16_t vmseq(vuint8mf2_t op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_u8mf2_b16_m))) +vbool16_t vmseq(vbool16_t op0, vbool16_t op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_u8mf4_b32))) +vbool32_t vmseq(vuint8mf4_t op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_u8mf4_b32_m))) +vbool32_t vmseq(vbool32_t op0, vbool32_t op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_u8mf8_b64))) +vbool64_t vmseq(vuint8mf8_t op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_u8mf8_b64_m))) +vbool64_t vmseq(vbool64_t op0, vbool64_t op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_u16m1_b16))) +vbool16_t vmseq(vuint16m1_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_u16m1_b16_m))) +vbool16_t vmseq(vbool16_t op0, vbool16_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_u16m2_b8))) +vbool8_t vmseq(vuint16m2_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_u16m2_b8_m))) +vbool8_t vmseq(vbool8_t op0, vbool8_t op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_u16m4_b4))) +vbool4_t vmseq(vuint16m4_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_u16m4_b4_m))) +vbool4_t vmseq(vbool4_t op0, vbool4_t op1, vuint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_u16m8_b2))) +vbool2_t vmseq(vuint16m8_t op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_u16m8_b2_m))) +vbool2_t vmseq(vbool2_t op0, vbool2_t op1, vuint16m8_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_u16mf2_b32))) +vbool32_t vmseq(vuint16mf2_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_u16mf2_b32_m))) +vbool32_t vmseq(vbool32_t op0, vbool32_t op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_u16mf4_b64))) +vbool64_t vmseq(vuint16mf4_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_u16mf4_b64_m))) +vbool64_t vmseq(vbool64_t op0, vbool64_t op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_u32m1_b32))) +vbool32_t vmseq(vuint32m1_t op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_u32m1_b32_m))) +vbool32_t vmseq(vbool32_t op0, vbool32_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_u32m2_b16))) +vbool16_t vmseq(vuint32m2_t op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_u32m2_b16_m))) +vbool16_t vmseq(vbool16_t op0, vbool16_t op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_u32m4_b8))) +vbool8_t vmseq(vuint32m4_t op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_u32m4_b8_m))) +vbool8_t vmseq(vbool8_t op0, vbool8_t op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_u32m8_b4))) +vbool4_t vmseq(vuint32m8_t op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_u32m8_b4_m))) +vbool4_t vmseq(vbool4_t op0, vbool4_t op1, vuint32m8_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_u32mf2_b64))) +vbool64_t vmseq(vuint32mf2_t op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_u32mf2_b64_m))) +vbool64_t vmseq(vbool64_t op0, vbool64_t op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_u64m1_b64))) +vbool64_t vmseq(vuint64m1_t op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_u64m1_b64_m))) +vbool64_t vmseq(vbool64_t op0, vbool64_t op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_u64m2_b32))) +vbool32_t vmseq(vuint64m2_t op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_u64m2_b32_m))) +vbool32_t vmseq(vbool32_t op0, vbool32_t op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_u64m4_b16))) +vbool16_t vmseq(vuint64m4_t op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_u64m4_b16_m))) +vbool16_t vmseq(vbool16_t op0, vbool16_t op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_u64m8_b8))) +vbool8_t vmseq(vuint64m8_t op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vv_u64m8_b8_m))) +vbool8_t vmseq(vbool8_t op0, vbool8_t op1, vuint64m8_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_u8m1_b8))) +vbool8_t vmseq(vuint8m1_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_u8m1_b8_m))) +vbool8_t vmseq(vbool8_t op0, vbool8_t op1, vuint8m1_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_u8m2_b4))) +vbool4_t vmseq(vuint8m2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_u8m2_b4_m))) +vbool4_t vmseq(vbool4_t op0, vbool4_t op1, vuint8m2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_u8m4_b2))) +vbool2_t vmseq(vuint8m4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_u8m4_b2_m))) +vbool2_t vmseq(vbool2_t op0, vbool2_t op1, vuint8m4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_u8m8_b1))) +vbool1_t vmseq(vuint8m8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_u8m8_b1_m))) +vbool1_t vmseq(vbool1_t op0, vbool1_t op1, vuint8m8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_u8mf2_b16))) +vbool16_t vmseq(vuint8mf2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_u8mf2_b16_m))) +vbool16_t vmseq(vbool16_t op0, vbool16_t op1, vuint8mf2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_u8mf4_b32))) +vbool32_t vmseq(vuint8mf4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_u8mf4_b32_m))) +vbool32_t vmseq(vbool32_t op0, vbool32_t op1, vuint8mf4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_u8mf8_b64))) +vbool64_t vmseq(vuint8mf8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_u8mf8_b64_m))) +vbool64_t vmseq(vbool64_t op0, vbool64_t op1, vuint8mf8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_u16m1_b16))) +vbool16_t vmseq(vuint16m1_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_u16m1_b16_m))) +vbool16_t vmseq(vbool16_t op0, vbool16_t op1, vuint16m1_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_u16m2_b8))) +vbool8_t vmseq(vuint16m2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_u16m2_b8_m))) +vbool8_t vmseq(vbool8_t op0, vbool8_t op1, vuint16m2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_u16m4_b4))) +vbool4_t vmseq(vuint16m4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_u16m4_b4_m))) +vbool4_t vmseq(vbool4_t op0, vbool4_t op1, vuint16m4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_u16m8_b2))) +vbool2_t vmseq(vuint16m8_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_u16m8_b2_m))) +vbool2_t vmseq(vbool2_t op0, vbool2_t op1, vuint16m8_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_u16mf2_b32))) +vbool32_t vmseq(vuint16mf2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_u16mf2_b32_m))) +vbool32_t vmseq(vbool32_t op0, vbool32_t op1, vuint16mf2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_u16mf4_b64))) +vbool64_t vmseq(vuint16mf4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_u16mf4_b64_m))) +vbool64_t vmseq(vbool64_t op0, vbool64_t op1, vuint16mf4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_u32m1_b32))) +vbool32_t vmseq(vuint32m1_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_u32m1_b32_m))) +vbool32_t vmseq(vbool32_t op0, vbool32_t op1, vuint32m1_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_u32m2_b16))) +vbool16_t vmseq(vuint32m2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_u32m2_b16_m))) +vbool16_t vmseq(vbool16_t op0, vbool16_t op1, vuint32m2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_u32m4_b8))) +vbool8_t vmseq(vuint32m4_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_u32m4_b8_m))) +vbool8_t vmseq(vbool8_t op0, vbool8_t op1, vuint32m4_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_u32m8_b4))) +vbool4_t vmseq(vuint32m8_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_u32m8_b4_m))) +vbool4_t vmseq(vbool4_t op0, vbool4_t op1, vuint32m8_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_u32mf2_b64))) +vbool64_t vmseq(vuint32mf2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_u32mf2_b64_m))) +vbool64_t vmseq(vbool64_t op0, vbool64_t op1, vuint32mf2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_u64m1_b64))) +vbool64_t vmseq(vuint64m1_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_u64m1_b64_m))) +vbool64_t vmseq(vbool64_t op0, vbool64_t op1, vuint64m1_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_u64m2_b32))) +vbool32_t vmseq(vuint64m2_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_u64m2_b32_m))) +vbool32_t vmseq(vbool32_t op0, vbool32_t op1, vuint64m2_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_u64m4_b16))) +vbool16_t vmseq(vuint64m4_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_u64m4_b16_m))) +vbool16_t vmseq(vbool16_t op0, vbool16_t op1, vuint64m4_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_u64m8_b8))) +vbool8_t vmseq(vuint64m8_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmseq_vx_u64m8_b8_m))) +vbool8_t vmseq(vbool8_t op0, vbool8_t op1, vuint64m8_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_i8m1_b8))) +vbool8_t vmsne(vint8m1_t op0, vint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_i8m1_b8_m))) +vbool8_t vmsne(vbool8_t op0, vbool8_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_i8m2_b4))) +vbool4_t vmsne(vint8m2_t op0, vint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_i8m2_b4_m))) +vbool4_t vmsne(vbool4_t op0, vbool4_t op1, vint8m2_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_i8m4_b2))) +vbool2_t vmsne(vint8m4_t op0, vint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_i8m4_b2_m))) +vbool2_t vmsne(vbool2_t op0, vbool2_t op1, vint8m4_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_i8m8_b1))) +vbool1_t vmsne(vint8m8_t op0, vint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_i8m8_b1_m))) +vbool1_t vmsne(vbool1_t op0, vbool1_t op1, vint8m8_t op2, vint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_i8mf2_b16))) +vbool16_t vmsne(vint8mf2_t op0, vint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_i8mf2_b16_m))) +vbool16_t vmsne(vbool16_t op0, vbool16_t op1, vint8mf2_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_i8mf4_b32))) +vbool32_t vmsne(vint8mf4_t op0, vint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_i8mf4_b32_m))) +vbool32_t vmsne(vbool32_t op0, vbool32_t op1, vint8mf4_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_i8mf8_b64))) +vbool64_t vmsne(vint8mf8_t op0, vint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_i8mf8_b64_m))) +vbool64_t vmsne(vbool64_t op0, vbool64_t op1, vint8mf8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_i16m1_b16))) +vbool16_t vmsne(vint16m1_t op0, vint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_i16m1_b16_m))) +vbool16_t vmsne(vbool16_t op0, vbool16_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_i16m2_b8))) +vbool8_t vmsne(vint16m2_t op0, vint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_i16m2_b8_m))) +vbool8_t vmsne(vbool8_t op0, vbool8_t op1, vint16m2_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_i16m4_b4))) +vbool4_t vmsne(vint16m4_t op0, vint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_i16m4_b4_m))) +vbool4_t vmsne(vbool4_t op0, vbool4_t op1, vint16m4_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_i16m8_b2))) +vbool2_t vmsne(vint16m8_t op0, vint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_i16m8_b2_m))) +vbool2_t vmsne(vbool2_t op0, vbool2_t op1, vint16m8_t op2, vint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_i16mf2_b32))) +vbool32_t vmsne(vint16mf2_t op0, vint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_i16mf2_b32_m))) +vbool32_t vmsne(vbool32_t op0, vbool32_t op1, vint16mf2_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_i16mf4_b64))) +vbool64_t vmsne(vint16mf4_t op0, vint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_i16mf4_b64_m))) +vbool64_t vmsne(vbool64_t op0, vbool64_t op1, vint16mf4_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_i32m1_b32))) +vbool32_t vmsne(vint32m1_t op0, vint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_i32m1_b32_m))) +vbool32_t vmsne(vbool32_t op0, vbool32_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_i32m2_b16))) +vbool16_t vmsne(vint32m2_t op0, vint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_i32m2_b16_m))) +vbool16_t vmsne(vbool16_t op0, vbool16_t op1, vint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_i32m4_b8))) +vbool8_t vmsne(vint32m4_t op0, vint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_i32m4_b8_m))) +vbool8_t vmsne(vbool8_t op0, vbool8_t op1, vint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_i32m8_b4))) +vbool4_t vmsne(vint32m8_t op0, vint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_i32m8_b4_m))) +vbool4_t vmsne(vbool4_t op0, vbool4_t op1, vint32m8_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_i32mf2_b64))) +vbool64_t vmsne(vint32mf2_t op0, vint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_i32mf2_b64_m))) +vbool64_t vmsne(vbool64_t op0, vbool64_t op1, vint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_i64m1_b64))) +vbool64_t vmsne(vint64m1_t op0, vint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_i64m1_b64_m))) +vbool64_t vmsne(vbool64_t op0, vbool64_t op1, vint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_i64m2_b32))) +vbool32_t vmsne(vint64m2_t op0, vint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_i64m2_b32_m))) +vbool32_t vmsne(vbool32_t op0, vbool32_t op1, vint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_i64m4_b16))) +vbool16_t vmsne(vint64m4_t op0, vint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_i64m4_b16_m))) +vbool16_t vmsne(vbool16_t op0, vbool16_t op1, vint64m4_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_i64m8_b8))) +vbool8_t vmsne(vint64m8_t op0, vint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_i64m8_b8_m))) +vbool8_t vmsne(vbool8_t op0, vbool8_t op1, vint64m8_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_i8m1_b8))) +vbool8_t vmsne(vint8m1_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_i8m1_b8_m))) +vbool8_t vmsne(vbool8_t op0, vbool8_t op1, vint8m1_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_i8m2_b4))) +vbool4_t vmsne(vint8m2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_i8m2_b4_m))) +vbool4_t vmsne(vbool4_t op0, vbool4_t op1, vint8m2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_i8m4_b2))) +vbool2_t vmsne(vint8m4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_i8m4_b2_m))) +vbool2_t vmsne(vbool2_t op0, vbool2_t op1, vint8m4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_i8m8_b1))) +vbool1_t vmsne(vint8m8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_i8m8_b1_m))) +vbool1_t vmsne(vbool1_t op0, vbool1_t op1, vint8m8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_i8mf2_b16))) +vbool16_t vmsne(vint8mf2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_i8mf2_b16_m))) +vbool16_t vmsne(vbool16_t op0, vbool16_t op1, vint8mf2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_i8mf4_b32))) +vbool32_t vmsne(vint8mf4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_i8mf4_b32_m))) +vbool32_t vmsne(vbool32_t op0, vbool32_t op1, vint8mf4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_i8mf8_b64))) +vbool64_t vmsne(vint8mf8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_i8mf8_b64_m))) +vbool64_t vmsne(vbool64_t op0, vbool64_t op1, vint8mf8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_i16m1_b16))) +vbool16_t vmsne(vint16m1_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_i16m1_b16_m))) +vbool16_t vmsne(vbool16_t op0, vbool16_t op1, vint16m1_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_i16m2_b8))) +vbool8_t vmsne(vint16m2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_i16m2_b8_m))) +vbool8_t vmsne(vbool8_t op0, vbool8_t op1, vint16m2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_i16m4_b4))) +vbool4_t vmsne(vint16m4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_i16m4_b4_m))) +vbool4_t vmsne(vbool4_t op0, vbool4_t op1, vint16m4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_i16m8_b2))) +vbool2_t vmsne(vint16m8_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_i16m8_b2_m))) +vbool2_t vmsne(vbool2_t op0, vbool2_t op1, vint16m8_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_i16mf2_b32))) +vbool32_t vmsne(vint16mf2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_i16mf2_b32_m))) +vbool32_t vmsne(vbool32_t op0, vbool32_t op1, vint16mf2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_i16mf4_b64))) +vbool64_t vmsne(vint16mf4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_i16mf4_b64_m))) +vbool64_t vmsne(vbool64_t op0, vbool64_t op1, vint16mf4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_i32m1_b32))) +vbool32_t vmsne(vint32m1_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_i32m1_b32_m))) +vbool32_t vmsne(vbool32_t op0, vbool32_t op1, vint32m1_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_i32m2_b16))) +vbool16_t vmsne(vint32m2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_i32m2_b16_m))) +vbool16_t vmsne(vbool16_t op0, vbool16_t op1, vint32m2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_i32m4_b8))) +vbool8_t vmsne(vint32m4_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_i32m4_b8_m))) +vbool8_t vmsne(vbool8_t op0, vbool8_t op1, vint32m4_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_i32m8_b4))) +vbool4_t vmsne(vint32m8_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_i32m8_b4_m))) +vbool4_t vmsne(vbool4_t op0, vbool4_t op1, vint32m8_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_i32mf2_b64))) +vbool64_t vmsne(vint32mf2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_i32mf2_b64_m))) +vbool64_t vmsne(vbool64_t op0, vbool64_t op1, vint32mf2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_i64m1_b64))) +vbool64_t vmsne(vint64m1_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_i64m1_b64_m))) +vbool64_t vmsne(vbool64_t op0, vbool64_t op1, vint64m1_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_i64m2_b32))) +vbool32_t vmsne(vint64m2_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_i64m2_b32_m))) +vbool32_t vmsne(vbool32_t op0, vbool32_t op1, vint64m2_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_i64m4_b16))) +vbool16_t vmsne(vint64m4_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_i64m4_b16_m))) +vbool16_t vmsne(vbool16_t op0, vbool16_t op1, vint64m4_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_i64m8_b8))) +vbool8_t vmsne(vint64m8_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_i64m8_b8_m))) +vbool8_t vmsne(vbool8_t op0, vbool8_t op1, vint64m8_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_u8m1_b8))) +vbool8_t vmsne(vuint8m1_t op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_u8m1_b8_m))) +vbool8_t vmsne(vbool8_t op0, vbool8_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_u8m2_b4))) +vbool4_t vmsne(vuint8m2_t op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_u8m2_b4_m))) +vbool4_t vmsne(vbool4_t op0, vbool4_t op1, vuint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_u8m4_b2))) +vbool2_t vmsne(vuint8m4_t op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_u8m4_b2_m))) +vbool2_t vmsne(vbool2_t op0, vbool2_t op1, vuint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_u8m8_b1))) +vbool1_t vmsne(vuint8m8_t op0, vuint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_u8m8_b1_m))) +vbool1_t vmsne(vbool1_t op0, vbool1_t op1, vuint8m8_t op2, vuint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_u8mf2_b16))) +vbool16_t vmsne(vuint8mf2_t op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_u8mf2_b16_m))) +vbool16_t vmsne(vbool16_t op0, vbool16_t op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_u8mf4_b32))) +vbool32_t vmsne(vuint8mf4_t op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_u8mf4_b32_m))) +vbool32_t vmsne(vbool32_t op0, vbool32_t op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_u8mf8_b64))) +vbool64_t vmsne(vuint8mf8_t op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_u8mf8_b64_m))) +vbool64_t vmsne(vbool64_t op0, vbool64_t op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_u16m1_b16))) +vbool16_t vmsne(vuint16m1_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_u16m1_b16_m))) +vbool16_t vmsne(vbool16_t op0, vbool16_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_u16m2_b8))) +vbool8_t vmsne(vuint16m2_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_u16m2_b8_m))) +vbool8_t vmsne(vbool8_t op0, vbool8_t op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_u16m4_b4))) +vbool4_t vmsne(vuint16m4_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_u16m4_b4_m))) +vbool4_t vmsne(vbool4_t op0, vbool4_t op1, vuint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_u16m8_b2))) +vbool2_t vmsne(vuint16m8_t op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_u16m8_b2_m))) +vbool2_t vmsne(vbool2_t op0, vbool2_t op1, vuint16m8_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_u16mf2_b32))) +vbool32_t vmsne(vuint16mf2_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_u16mf2_b32_m))) +vbool32_t vmsne(vbool32_t op0, vbool32_t op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_u16mf4_b64))) +vbool64_t vmsne(vuint16mf4_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_u16mf4_b64_m))) +vbool64_t vmsne(vbool64_t op0, vbool64_t op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_u32m1_b32))) +vbool32_t vmsne(vuint32m1_t op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_u32m1_b32_m))) +vbool32_t vmsne(vbool32_t op0, vbool32_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_u32m2_b16))) +vbool16_t vmsne(vuint32m2_t op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_u32m2_b16_m))) +vbool16_t vmsne(vbool16_t op0, vbool16_t op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_u32m4_b8))) +vbool8_t vmsne(vuint32m4_t op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_u32m4_b8_m))) +vbool8_t vmsne(vbool8_t op0, vbool8_t op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_u32m8_b4))) +vbool4_t vmsne(vuint32m8_t op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_u32m8_b4_m))) +vbool4_t vmsne(vbool4_t op0, vbool4_t op1, vuint32m8_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_u32mf2_b64))) +vbool64_t vmsne(vuint32mf2_t op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_u32mf2_b64_m))) +vbool64_t vmsne(vbool64_t op0, vbool64_t op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_u64m1_b64))) +vbool64_t vmsne(vuint64m1_t op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_u64m1_b64_m))) +vbool64_t vmsne(vbool64_t op0, vbool64_t op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_u64m2_b32))) +vbool32_t vmsne(vuint64m2_t op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_u64m2_b32_m))) +vbool32_t vmsne(vbool32_t op0, vbool32_t op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_u64m4_b16))) +vbool16_t vmsne(vuint64m4_t op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_u64m4_b16_m))) +vbool16_t vmsne(vbool16_t op0, vbool16_t op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_u64m8_b8))) +vbool8_t vmsne(vuint64m8_t op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vv_u64m8_b8_m))) +vbool8_t vmsne(vbool8_t op0, vbool8_t op1, vuint64m8_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_u8m1_b8))) +vbool8_t vmsne(vuint8m1_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_u8m1_b8_m))) +vbool8_t vmsne(vbool8_t op0, vbool8_t op1, vuint8m1_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_u8m2_b4))) +vbool4_t vmsne(vuint8m2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_u8m2_b4_m))) +vbool4_t vmsne(vbool4_t op0, vbool4_t op1, vuint8m2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_u8m4_b2))) +vbool2_t vmsne(vuint8m4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_u8m4_b2_m))) +vbool2_t vmsne(vbool2_t op0, vbool2_t op1, vuint8m4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_u8m8_b1))) +vbool1_t vmsne(vuint8m8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_u8m8_b1_m))) +vbool1_t vmsne(vbool1_t op0, vbool1_t op1, vuint8m8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_u8mf2_b16))) +vbool16_t vmsne(vuint8mf2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_u8mf2_b16_m))) +vbool16_t vmsne(vbool16_t op0, vbool16_t op1, vuint8mf2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_u8mf4_b32))) +vbool32_t vmsne(vuint8mf4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_u8mf4_b32_m))) +vbool32_t vmsne(vbool32_t op0, vbool32_t op1, vuint8mf4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_u8mf8_b64))) +vbool64_t vmsne(vuint8mf8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_u8mf8_b64_m))) +vbool64_t vmsne(vbool64_t op0, vbool64_t op1, vuint8mf8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_u16m1_b16))) +vbool16_t vmsne(vuint16m1_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_u16m1_b16_m))) +vbool16_t vmsne(vbool16_t op0, vbool16_t op1, vuint16m1_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_u16m2_b8))) +vbool8_t vmsne(vuint16m2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_u16m2_b8_m))) +vbool8_t vmsne(vbool8_t op0, vbool8_t op1, vuint16m2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_u16m4_b4))) +vbool4_t vmsne(vuint16m4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_u16m4_b4_m))) +vbool4_t vmsne(vbool4_t op0, vbool4_t op1, vuint16m4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_u16m8_b2))) +vbool2_t vmsne(vuint16m8_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_u16m8_b2_m))) +vbool2_t vmsne(vbool2_t op0, vbool2_t op1, vuint16m8_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_u16mf2_b32))) +vbool32_t vmsne(vuint16mf2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_u16mf2_b32_m))) +vbool32_t vmsne(vbool32_t op0, vbool32_t op1, vuint16mf2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_u16mf4_b64))) +vbool64_t vmsne(vuint16mf4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_u16mf4_b64_m))) +vbool64_t vmsne(vbool64_t op0, vbool64_t op1, vuint16mf4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_u32m1_b32))) +vbool32_t vmsne(vuint32m1_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_u32m1_b32_m))) +vbool32_t vmsne(vbool32_t op0, vbool32_t op1, vuint32m1_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_u32m2_b16))) +vbool16_t vmsne(vuint32m2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_u32m2_b16_m))) +vbool16_t vmsne(vbool16_t op0, vbool16_t op1, vuint32m2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_u32m4_b8))) +vbool8_t vmsne(vuint32m4_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_u32m4_b8_m))) +vbool8_t vmsne(vbool8_t op0, vbool8_t op1, vuint32m4_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_u32m8_b4))) +vbool4_t vmsne(vuint32m8_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_u32m8_b4_m))) +vbool4_t vmsne(vbool4_t op0, vbool4_t op1, vuint32m8_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_u32mf2_b64))) +vbool64_t vmsne(vuint32mf2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_u32mf2_b64_m))) +vbool64_t vmsne(vbool64_t op0, vbool64_t op1, vuint32mf2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_u64m1_b64))) +vbool64_t vmsne(vuint64m1_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_u64m1_b64_m))) +vbool64_t vmsne(vbool64_t op0, vbool64_t op1, vuint64m1_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_u64m2_b32))) +vbool32_t vmsne(vuint64m2_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_u64m2_b32_m))) +vbool32_t vmsne(vbool32_t op0, vbool32_t op1, vuint64m2_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_u64m4_b16))) +vbool16_t vmsne(vuint64m4_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_u64m4_b16_m))) +vbool16_t vmsne(vbool16_t op0, vbool16_t op1, vuint64m4_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_u64m8_b8))) +vbool8_t vmsne(vuint64m8_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsne_vx_u64m8_b8_m))) +vbool8_t vmsne(vbool8_t op0, vbool8_t op1, vuint64m8_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vv_u8m1_b8))) +vbool8_t vmsltu(vuint8m1_t op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vv_u8m1_b8_m))) +vbool8_t vmsltu(vbool8_t op0, vbool8_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vv_u8m2_b4))) +vbool4_t vmsltu(vuint8m2_t op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vv_u8m2_b4_m))) +vbool4_t vmsltu(vbool4_t op0, vbool4_t op1, vuint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vv_u8m4_b2))) +vbool2_t vmsltu(vuint8m4_t op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vv_u8m4_b2_m))) +vbool2_t vmsltu(vbool2_t op0, vbool2_t op1, vuint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vv_u8m8_b1))) +vbool1_t vmsltu(vuint8m8_t op0, vuint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vv_u8m8_b1_m))) +vbool1_t vmsltu(vbool1_t op0, vbool1_t op1, vuint8m8_t op2, vuint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vv_u8mf2_b16))) +vbool16_t vmsltu(vuint8mf2_t op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vv_u8mf2_b16_m))) +vbool16_t vmsltu(vbool16_t op0, vbool16_t op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vv_u8mf4_b32))) +vbool32_t vmsltu(vuint8mf4_t op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vv_u8mf4_b32_m))) +vbool32_t vmsltu(vbool32_t op0, vbool32_t op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vv_u8mf8_b64))) +vbool64_t vmsltu(vuint8mf8_t op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vv_u8mf8_b64_m))) +vbool64_t vmsltu(vbool64_t op0, vbool64_t op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vv_u16m1_b16))) +vbool16_t vmsltu(vuint16m1_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vv_u16m1_b16_m))) +vbool16_t vmsltu(vbool16_t op0, vbool16_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vv_u16m2_b8))) +vbool8_t vmsltu(vuint16m2_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vv_u16m2_b8_m))) +vbool8_t vmsltu(vbool8_t op0, vbool8_t op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vv_u16m4_b4))) +vbool4_t vmsltu(vuint16m4_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vv_u16m4_b4_m))) +vbool4_t vmsltu(vbool4_t op0, vbool4_t op1, vuint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vv_u16m8_b2))) +vbool2_t vmsltu(vuint16m8_t op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vv_u16m8_b2_m))) +vbool2_t vmsltu(vbool2_t op0, vbool2_t op1, vuint16m8_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vv_u16mf2_b32))) +vbool32_t vmsltu(vuint16mf2_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vv_u16mf2_b32_m))) +vbool32_t vmsltu(vbool32_t op0, vbool32_t op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vv_u16mf4_b64))) +vbool64_t vmsltu(vuint16mf4_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vv_u16mf4_b64_m))) +vbool64_t vmsltu(vbool64_t op0, vbool64_t op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vv_u32m1_b32))) +vbool32_t vmsltu(vuint32m1_t op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vv_u32m1_b32_m))) +vbool32_t vmsltu(vbool32_t op0, vbool32_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vv_u32m2_b16))) +vbool16_t vmsltu(vuint32m2_t op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vv_u32m2_b16_m))) +vbool16_t vmsltu(vbool16_t op0, vbool16_t op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vv_u32m4_b8))) +vbool8_t vmsltu(vuint32m4_t op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vv_u32m4_b8_m))) +vbool8_t vmsltu(vbool8_t op0, vbool8_t op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vv_u32m8_b4))) +vbool4_t vmsltu(vuint32m8_t op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vv_u32m8_b4_m))) +vbool4_t vmsltu(vbool4_t op0, vbool4_t op1, vuint32m8_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vv_u32mf2_b64))) +vbool64_t vmsltu(vuint32mf2_t op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vv_u32mf2_b64_m))) +vbool64_t vmsltu(vbool64_t op0, vbool64_t op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vv_u64m1_b64))) +vbool64_t vmsltu(vuint64m1_t op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vv_u64m1_b64_m))) +vbool64_t vmsltu(vbool64_t op0, vbool64_t op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vv_u64m2_b32))) +vbool32_t vmsltu(vuint64m2_t op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vv_u64m2_b32_m))) +vbool32_t vmsltu(vbool32_t op0, vbool32_t op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vv_u64m4_b16))) +vbool16_t vmsltu(vuint64m4_t op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vv_u64m4_b16_m))) +vbool16_t vmsltu(vbool16_t op0, vbool16_t op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vv_u64m8_b8))) +vbool8_t vmsltu(vuint64m8_t op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vv_u64m8_b8_m))) +vbool8_t vmsltu(vbool8_t op0, vbool8_t op1, vuint64m8_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vx_u8m1_b8))) +vbool8_t vmsltu(vuint8m1_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vx_u8m1_b8_m))) +vbool8_t vmsltu(vbool8_t op0, vbool8_t op1, vuint8m1_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vx_u8m2_b4))) +vbool4_t vmsltu(vuint8m2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vx_u8m2_b4_m))) +vbool4_t vmsltu(vbool4_t op0, vbool4_t op1, vuint8m2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vx_u8m4_b2))) +vbool2_t vmsltu(vuint8m4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vx_u8m4_b2_m))) +vbool2_t vmsltu(vbool2_t op0, vbool2_t op1, vuint8m4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vx_u8m8_b1))) +vbool1_t vmsltu(vuint8m8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vx_u8m8_b1_m))) +vbool1_t vmsltu(vbool1_t op0, vbool1_t op1, vuint8m8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vx_u8mf2_b16))) +vbool16_t vmsltu(vuint8mf2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vx_u8mf2_b16_m))) +vbool16_t vmsltu(vbool16_t op0, vbool16_t op1, vuint8mf2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vx_u8mf4_b32))) +vbool32_t vmsltu(vuint8mf4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vx_u8mf4_b32_m))) +vbool32_t vmsltu(vbool32_t op0, vbool32_t op1, vuint8mf4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vx_u8mf8_b64))) +vbool64_t vmsltu(vuint8mf8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vx_u8mf8_b64_m))) +vbool64_t vmsltu(vbool64_t op0, vbool64_t op1, vuint8mf8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vx_u16m1_b16))) +vbool16_t vmsltu(vuint16m1_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vx_u16m1_b16_m))) +vbool16_t vmsltu(vbool16_t op0, vbool16_t op1, vuint16m1_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vx_u16m2_b8))) +vbool8_t vmsltu(vuint16m2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vx_u16m2_b8_m))) +vbool8_t vmsltu(vbool8_t op0, vbool8_t op1, vuint16m2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vx_u16m4_b4))) +vbool4_t vmsltu(vuint16m4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vx_u16m4_b4_m))) +vbool4_t vmsltu(vbool4_t op0, vbool4_t op1, vuint16m4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vx_u16m8_b2))) +vbool2_t vmsltu(vuint16m8_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vx_u16m8_b2_m))) +vbool2_t vmsltu(vbool2_t op0, vbool2_t op1, vuint16m8_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vx_u16mf2_b32))) +vbool32_t vmsltu(vuint16mf2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vx_u16mf2_b32_m))) +vbool32_t vmsltu(vbool32_t op0, vbool32_t op1, vuint16mf2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vx_u16mf4_b64))) +vbool64_t vmsltu(vuint16mf4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vx_u16mf4_b64_m))) +vbool64_t vmsltu(vbool64_t op0, vbool64_t op1, vuint16mf4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vx_u32m1_b32))) +vbool32_t vmsltu(vuint32m1_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vx_u32m1_b32_m))) +vbool32_t vmsltu(vbool32_t op0, vbool32_t op1, vuint32m1_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vx_u32m2_b16))) +vbool16_t vmsltu(vuint32m2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vx_u32m2_b16_m))) +vbool16_t vmsltu(vbool16_t op0, vbool16_t op1, vuint32m2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vx_u32m4_b8))) +vbool8_t vmsltu(vuint32m4_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vx_u32m4_b8_m))) +vbool8_t vmsltu(vbool8_t op0, vbool8_t op1, vuint32m4_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vx_u32m8_b4))) +vbool4_t vmsltu(vuint32m8_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vx_u32m8_b4_m))) +vbool4_t vmsltu(vbool4_t op0, vbool4_t op1, vuint32m8_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vx_u32mf2_b64))) +vbool64_t vmsltu(vuint32mf2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vx_u32mf2_b64_m))) +vbool64_t vmsltu(vbool64_t op0, vbool64_t op1, vuint32mf2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vx_u64m1_b64))) +vbool64_t vmsltu(vuint64m1_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vx_u64m1_b64_m))) +vbool64_t vmsltu(vbool64_t op0, vbool64_t op1, vuint64m1_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vx_u64m2_b32))) +vbool32_t vmsltu(vuint64m2_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vx_u64m2_b32_m))) +vbool32_t vmsltu(vbool32_t op0, vbool32_t op1, vuint64m2_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vx_u64m4_b16))) +vbool16_t vmsltu(vuint64m4_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vx_u64m4_b16_m))) +vbool16_t vmsltu(vbool16_t op0, vbool16_t op1, vuint64m4_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vx_u64m8_b8))) +vbool8_t vmsltu(vuint64m8_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsltu_vx_u64m8_b8_m))) +vbool8_t vmsltu(vbool8_t op0, vbool8_t op1, vuint64m8_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vv_i8m1_b8))) +vbool8_t vmslt(vint8m1_t op0, vint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vv_i8m1_b8_m))) +vbool8_t vmslt(vbool8_t op0, vbool8_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vv_i8m2_b4))) +vbool4_t vmslt(vint8m2_t op0, vint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vv_i8m2_b4_m))) +vbool4_t vmslt(vbool4_t op0, vbool4_t op1, vint8m2_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vv_i8m4_b2))) +vbool2_t vmslt(vint8m4_t op0, vint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vv_i8m4_b2_m))) +vbool2_t vmslt(vbool2_t op0, vbool2_t op1, vint8m4_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vv_i8m8_b1))) +vbool1_t vmslt(vint8m8_t op0, vint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vv_i8m8_b1_m))) +vbool1_t vmslt(vbool1_t op0, vbool1_t op1, vint8m8_t op2, vint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vv_i8mf2_b16))) +vbool16_t vmslt(vint8mf2_t op0, vint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vv_i8mf2_b16_m))) +vbool16_t vmslt(vbool16_t op0, vbool16_t op1, vint8mf2_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vv_i8mf4_b32))) +vbool32_t vmslt(vint8mf4_t op0, vint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vv_i8mf4_b32_m))) +vbool32_t vmslt(vbool32_t op0, vbool32_t op1, vint8mf4_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vv_i8mf8_b64))) +vbool64_t vmslt(vint8mf8_t op0, vint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vv_i8mf8_b64_m))) +vbool64_t vmslt(vbool64_t op0, vbool64_t op1, vint8mf8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vv_i16m1_b16))) +vbool16_t vmslt(vint16m1_t op0, vint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vv_i16m1_b16_m))) +vbool16_t vmslt(vbool16_t op0, vbool16_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vv_i16m2_b8))) +vbool8_t vmslt(vint16m2_t op0, vint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vv_i16m2_b8_m))) +vbool8_t vmslt(vbool8_t op0, vbool8_t op1, vint16m2_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vv_i16m4_b4))) +vbool4_t vmslt(vint16m4_t op0, vint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vv_i16m4_b4_m))) +vbool4_t vmslt(vbool4_t op0, vbool4_t op1, vint16m4_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vv_i16m8_b2))) +vbool2_t vmslt(vint16m8_t op0, vint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vv_i16m8_b2_m))) +vbool2_t vmslt(vbool2_t op0, vbool2_t op1, vint16m8_t op2, vint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vv_i16mf2_b32))) +vbool32_t vmslt(vint16mf2_t op0, vint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vv_i16mf2_b32_m))) +vbool32_t vmslt(vbool32_t op0, vbool32_t op1, vint16mf2_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vv_i16mf4_b64))) +vbool64_t vmslt(vint16mf4_t op0, vint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vv_i16mf4_b64_m))) +vbool64_t vmslt(vbool64_t op0, vbool64_t op1, vint16mf4_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vv_i32m1_b32))) +vbool32_t vmslt(vint32m1_t op0, vint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vv_i32m1_b32_m))) +vbool32_t vmslt(vbool32_t op0, vbool32_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vv_i32m2_b16))) +vbool16_t vmslt(vint32m2_t op0, vint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vv_i32m2_b16_m))) +vbool16_t vmslt(vbool16_t op0, vbool16_t op1, vint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vv_i32m4_b8))) +vbool8_t vmslt(vint32m4_t op0, vint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vv_i32m4_b8_m))) +vbool8_t vmslt(vbool8_t op0, vbool8_t op1, vint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vv_i32m8_b4))) +vbool4_t vmslt(vint32m8_t op0, vint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vv_i32m8_b4_m))) +vbool4_t vmslt(vbool4_t op0, vbool4_t op1, vint32m8_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vv_i32mf2_b64))) +vbool64_t vmslt(vint32mf2_t op0, vint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vv_i32mf2_b64_m))) +vbool64_t vmslt(vbool64_t op0, vbool64_t op1, vint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vv_i64m1_b64))) +vbool64_t vmslt(vint64m1_t op0, vint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vv_i64m1_b64_m))) +vbool64_t vmslt(vbool64_t op0, vbool64_t op1, vint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vv_i64m2_b32))) +vbool32_t vmslt(vint64m2_t op0, vint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vv_i64m2_b32_m))) +vbool32_t vmslt(vbool32_t op0, vbool32_t op1, vint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vv_i64m4_b16))) +vbool16_t vmslt(vint64m4_t op0, vint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vv_i64m4_b16_m))) +vbool16_t vmslt(vbool16_t op0, vbool16_t op1, vint64m4_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vv_i64m8_b8))) +vbool8_t vmslt(vint64m8_t op0, vint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vv_i64m8_b8_m))) +vbool8_t vmslt(vbool8_t op0, vbool8_t op1, vint64m8_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_u16m1))) +void vsuxei32(uint16_t * op0, vuint32m2_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_u16m1_m))) +void vsuxei32(vbool16_t op0, uint16_t * op1, vuint32m2_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_u16m2))) +void vsuxei32(uint16_t * op0, vuint32m4_t op1, vuint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_u16m2_m))) +void vsuxei32(vbool8_t op0, uint16_t * op1, vuint32m4_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_u16m4))) +void vsuxei32(uint16_t * op0, vuint32m8_t op1, vuint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_u16m4_m))) +void vsuxei32(vbool4_t op0, uint16_t * op1, vuint32m8_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_u16mf2))) +void vsuxei32(uint16_t * op0, vuint32m1_t op1, vuint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_u16mf2_m))) +void vsuxei32(vbool32_t op0, uint16_t * op1, vuint32m1_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_u16mf4))) +void vsuxei32(uint16_t * op0, vuint32mf2_t op1, vuint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_u16mf4_m))) +void vsuxei32(vbool64_t op0, uint16_t * op1, vuint32mf2_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vx_i8m1_b8))) +vbool8_t vmslt(vint8m1_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vx_i8m1_b8_m))) +vbool8_t vmslt(vbool8_t op0, vbool8_t op1, vint8m1_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vx_i8m2_b4))) +vbool4_t vmslt(vint8m2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vx_i8m2_b4_m))) +vbool4_t vmslt(vbool4_t op0, vbool4_t op1, vint8m2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vx_i8m4_b2))) +vbool2_t vmslt(vint8m4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vx_i8m4_b2_m))) +vbool2_t vmslt(vbool2_t op0, vbool2_t op1, vint8m4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vx_i8m8_b1))) +vbool1_t vmslt(vint8m8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vx_i8m8_b1_m))) +vbool1_t vmslt(vbool1_t op0, vbool1_t op1, vint8m8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vx_i8mf2_b16))) +vbool16_t vmslt(vint8mf2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vx_i8mf2_b16_m))) +vbool16_t vmslt(vbool16_t op0, vbool16_t op1, vint8mf2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vx_i8mf4_b32))) +vbool32_t vmslt(vint8mf4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vx_i8mf4_b32_m))) +vbool32_t vmslt(vbool32_t op0, vbool32_t op1, vint8mf4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vx_i8mf8_b64))) +vbool64_t vmslt(vint8mf8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vx_i8mf8_b64_m))) +vbool64_t vmslt(vbool64_t op0, vbool64_t op1, vint8mf8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vx_i16m1_b16))) +vbool16_t vmslt(vint16m1_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vx_i16m1_b16_m))) +vbool16_t vmslt(vbool16_t op0, vbool16_t op1, vint16m1_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vx_i16m2_b8))) +vbool8_t vmslt(vint16m2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vx_i16m2_b8_m))) +vbool8_t vmslt(vbool8_t op0, vbool8_t op1, vint16m2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vx_i16m4_b4))) +vbool4_t vmslt(vint16m4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vx_i16m4_b4_m))) +vbool4_t vmslt(vbool4_t op0, vbool4_t op1, vint16m4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vx_i16m8_b2))) +vbool2_t vmslt(vint16m8_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vx_i16m8_b2_m))) +vbool2_t vmslt(vbool2_t op0, vbool2_t op1, vint16m8_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vx_i16mf2_b32))) +vbool32_t vmslt(vint16mf2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vx_i16mf2_b32_m))) +vbool32_t vmslt(vbool32_t op0, vbool32_t op1, vint16mf2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vx_i16mf4_b64))) +vbool64_t vmslt(vint16mf4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vx_i16mf4_b64_m))) +vbool64_t vmslt(vbool64_t op0, vbool64_t op1, vint16mf4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vx_i32m1_b32))) +vbool32_t vmslt(vint32m1_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vx_i32m1_b32_m))) +vbool32_t vmslt(vbool32_t op0, vbool32_t op1, vint32m1_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vx_i32m2_b16))) +vbool16_t vmslt(vint32m2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vx_i32m2_b16_m))) +vbool16_t vmslt(vbool16_t op0, vbool16_t op1, vint32m2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vx_i32m4_b8))) +vbool8_t vmslt(vint32m4_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vx_i32m4_b8_m))) +vbool8_t vmslt(vbool8_t op0, vbool8_t op1, vint32m4_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vx_i32m8_b4))) +vbool4_t vmslt(vint32m8_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vx_i32m8_b4_m))) +vbool4_t vmslt(vbool4_t op0, vbool4_t op1, vint32m8_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vx_i32mf2_b64))) +vbool64_t vmslt(vint32mf2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vx_i32mf2_b64_m))) +vbool64_t vmslt(vbool64_t op0, vbool64_t op1, vint32mf2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vx_i64m1_b64))) +vbool64_t vmslt(vint64m1_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vx_i64m1_b64_m))) +vbool64_t vmslt(vbool64_t op0, vbool64_t op1, vint64m1_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vx_i64m2_b32))) +vbool32_t vmslt(vint64m2_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vx_i64m2_b32_m))) +vbool32_t vmslt(vbool32_t op0, vbool32_t op1, vint64m2_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vx_i64m4_b16))) +vbool16_t vmslt(vint64m4_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vx_i64m4_b16_m))) +vbool16_t vmslt(vbool16_t op0, vbool16_t op1, vint64m4_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vx_i64m8_b8))) +vbool8_t vmslt(vint64m8_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmslt_vx_i64m8_b8_m))) +vbool8_t vmslt(vbool8_t op0, vbool8_t op1, vint64m8_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vv_u8m1_b8))) +vbool8_t vmsleu(vuint8m1_t op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vv_u8m1_b8_m))) +vbool8_t vmsleu(vbool8_t op0, vbool8_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vv_u8m2_b4))) +vbool4_t vmsleu(vuint8m2_t op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vv_u8m2_b4_m))) +vbool4_t vmsleu(vbool4_t op0, vbool4_t op1, vuint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vv_u8m4_b2))) +vbool2_t vmsleu(vuint8m4_t op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vv_u8m4_b2_m))) +vbool2_t vmsleu(vbool2_t op0, vbool2_t op1, vuint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vv_u8m8_b1))) +vbool1_t vmsleu(vuint8m8_t op0, vuint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vv_u8m8_b1_m))) +vbool1_t vmsleu(vbool1_t op0, vbool1_t op1, vuint8m8_t op2, vuint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vv_u8mf2_b16))) +vbool16_t vmsleu(vuint8mf2_t op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vv_u8mf2_b16_m))) +vbool16_t vmsleu(vbool16_t op0, vbool16_t op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vv_u8mf4_b32))) +vbool32_t vmsleu(vuint8mf4_t op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vv_u8mf4_b32_m))) +vbool32_t vmsleu(vbool32_t op0, vbool32_t op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vv_u8mf8_b64))) +vbool64_t vmsleu(vuint8mf8_t op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vv_u8mf8_b64_m))) +vbool64_t vmsleu(vbool64_t op0, vbool64_t op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vv_u16m1_b16))) +vbool16_t vmsleu(vuint16m1_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vv_u16m1_b16_m))) +vbool16_t vmsleu(vbool16_t op0, vbool16_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vv_u16m2_b8))) +vbool8_t vmsleu(vuint16m2_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vv_u16m2_b8_m))) +vbool8_t vmsleu(vbool8_t op0, vbool8_t op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vv_u16m4_b4))) +vbool4_t vmsleu(vuint16m4_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vv_u16m4_b4_m))) +vbool4_t vmsleu(vbool4_t op0, vbool4_t op1, vuint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vv_u16m8_b2))) +vbool2_t vmsleu(vuint16m8_t op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vv_u16m8_b2_m))) +vbool2_t vmsleu(vbool2_t op0, vbool2_t op1, vuint16m8_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vv_u16mf2_b32))) +vbool32_t vmsleu(vuint16mf2_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vv_u16mf2_b32_m))) +vbool32_t vmsleu(vbool32_t op0, vbool32_t op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vv_u16mf4_b64))) +vbool64_t vmsleu(vuint16mf4_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vv_u16mf4_b64_m))) +vbool64_t vmsleu(vbool64_t op0, vbool64_t op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vv_u32m1_b32))) +vbool32_t vmsleu(vuint32m1_t op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vv_u32m1_b32_m))) +vbool32_t vmsleu(vbool32_t op0, vbool32_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vv_u32m2_b16))) +vbool16_t vmsleu(vuint32m2_t op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vv_u32m2_b16_m))) +vbool16_t vmsleu(vbool16_t op0, vbool16_t op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vv_u32m4_b8))) +vbool8_t vmsleu(vuint32m4_t op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vv_u32m4_b8_m))) +vbool8_t vmsleu(vbool8_t op0, vbool8_t op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vv_u32m8_b4))) +vbool4_t vmsleu(vuint32m8_t op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vv_u32m8_b4_m))) +vbool4_t vmsleu(vbool4_t op0, vbool4_t op1, vuint32m8_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vv_u32mf2_b64))) +vbool64_t vmsleu(vuint32mf2_t op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vv_u32mf2_b64_m))) +vbool64_t vmsleu(vbool64_t op0, vbool64_t op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vv_u64m1_b64))) +vbool64_t vmsleu(vuint64m1_t op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vv_u64m1_b64_m))) +vbool64_t vmsleu(vbool64_t op0, vbool64_t op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vv_u64m2_b32))) +vbool32_t vmsleu(vuint64m2_t op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vv_u64m2_b32_m))) +vbool32_t vmsleu(vbool32_t op0, vbool32_t op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vv_u64m4_b16))) +vbool16_t vmsleu(vuint64m4_t op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vv_u64m4_b16_m))) +vbool16_t vmsleu(vbool16_t op0, vbool16_t op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vv_u64m8_b8))) +vbool8_t vmsleu(vuint64m8_t op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vv_u64m8_b8_m))) +vbool8_t vmsleu(vbool8_t op0, vbool8_t op1, vuint64m8_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vx_u8m1_b8))) +vbool8_t vmsleu(vuint8m1_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vx_u8m1_b8_m))) +vbool8_t vmsleu(vbool8_t op0, vbool8_t op1, vuint8m1_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vx_u8m2_b4))) +vbool4_t vmsleu(vuint8m2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vx_u8m2_b4_m))) +vbool4_t vmsleu(vbool4_t op0, vbool4_t op1, vuint8m2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vx_u8m4_b2))) +vbool2_t vmsleu(vuint8m4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vx_u8m4_b2_m))) +vbool2_t vmsleu(vbool2_t op0, vbool2_t op1, vuint8m4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vx_u8m8_b1))) +vbool1_t vmsleu(vuint8m8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vx_u8m8_b1_m))) +vbool1_t vmsleu(vbool1_t op0, vbool1_t op1, vuint8m8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vx_u8mf2_b16))) +vbool16_t vmsleu(vuint8mf2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vx_u8mf2_b16_m))) +vbool16_t vmsleu(vbool16_t op0, vbool16_t op1, vuint8mf2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vx_u8mf4_b32))) +vbool32_t vmsleu(vuint8mf4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vx_u8mf4_b32_m))) +vbool32_t vmsleu(vbool32_t op0, vbool32_t op1, vuint8mf4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vx_u8mf8_b64))) +vbool64_t vmsleu(vuint8mf8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vx_u8mf8_b64_m))) +vbool64_t vmsleu(vbool64_t op0, vbool64_t op1, vuint8mf8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vx_u16m1_b16))) +vbool16_t vmsleu(vuint16m1_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vx_u16m1_b16_m))) +vbool16_t vmsleu(vbool16_t op0, vbool16_t op1, vuint16m1_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vx_u16m2_b8))) +vbool8_t vmsleu(vuint16m2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vx_u16m2_b8_m))) +vbool8_t vmsleu(vbool8_t op0, vbool8_t op1, vuint16m2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vx_u16m4_b4))) +vbool4_t vmsleu(vuint16m4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vx_u16m4_b4_m))) +vbool4_t vmsleu(vbool4_t op0, vbool4_t op1, vuint16m4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vx_u16m8_b2))) +vbool2_t vmsleu(vuint16m8_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vx_u16m8_b2_m))) +vbool2_t vmsleu(vbool2_t op0, vbool2_t op1, vuint16m8_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vx_u16mf2_b32))) +vbool32_t vmsleu(vuint16mf2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vx_u16mf2_b32_m))) +vbool32_t vmsleu(vbool32_t op0, vbool32_t op1, vuint16mf2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vx_u16mf4_b64))) +vbool64_t vmsleu(vuint16mf4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vx_u16mf4_b64_m))) +vbool64_t vmsleu(vbool64_t op0, vbool64_t op1, vuint16mf4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vx_u32m1_b32))) +vbool32_t vmsleu(vuint32m1_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vx_u32m1_b32_m))) +vbool32_t vmsleu(vbool32_t op0, vbool32_t op1, vuint32m1_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vx_u32m2_b16))) +vbool16_t vmsleu(vuint32m2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vx_u32m2_b16_m))) +vbool16_t vmsleu(vbool16_t op0, vbool16_t op1, vuint32m2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vx_u32m4_b8))) +vbool8_t vmsleu(vuint32m4_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vx_u32m4_b8_m))) +vbool8_t vmsleu(vbool8_t op0, vbool8_t op1, vuint32m4_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vx_u32m8_b4))) +vbool4_t vmsleu(vuint32m8_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vx_u32m8_b4_m))) +vbool4_t vmsleu(vbool4_t op0, vbool4_t op1, vuint32m8_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vx_u32mf2_b64))) +vbool64_t vmsleu(vuint32mf2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vx_u32mf2_b64_m))) +vbool64_t vmsleu(vbool64_t op0, vbool64_t op1, vuint32mf2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vx_u64m1_b64))) +vbool64_t vmsleu(vuint64m1_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vx_u64m1_b64_m))) +vbool64_t vmsleu(vbool64_t op0, vbool64_t op1, vuint64m1_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vx_u64m2_b32))) +vbool32_t vmsleu(vuint64m2_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vx_u64m2_b32_m))) +vbool32_t vmsleu(vbool32_t op0, vbool32_t op1, vuint64m2_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vx_u64m4_b16))) +vbool16_t vmsleu(vuint64m4_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vx_u64m4_b16_m))) +vbool16_t vmsleu(vbool16_t op0, vbool16_t op1, vuint64m4_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vx_u64m8_b8))) +vbool8_t vmsleu(vuint64m8_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsleu_vx_u64m8_b8_m))) +vbool8_t vmsleu(vbool8_t op0, vbool8_t op1, vuint64m8_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vv_i8m1_b8))) +vbool8_t vmsle(vint8m1_t op0, vint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vv_i8m1_b8_m))) +vbool8_t vmsle(vbool8_t op0, vbool8_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vv_i8m2_b4))) +vbool4_t vmsle(vint8m2_t op0, vint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vv_i8m2_b4_m))) +vbool4_t vmsle(vbool4_t op0, vbool4_t op1, vint8m2_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vv_i8m4_b2))) +vbool2_t vmsle(vint8m4_t op0, vint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vv_i8m4_b2_m))) +vbool2_t vmsle(vbool2_t op0, vbool2_t op1, vint8m4_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vv_i8m8_b1))) +vbool1_t vmsle(vint8m8_t op0, vint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vv_i8m8_b1_m))) +vbool1_t vmsle(vbool1_t op0, vbool1_t op1, vint8m8_t op2, vint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vv_i8mf2_b16))) +vbool16_t vmsle(vint8mf2_t op0, vint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vv_i8mf2_b16_m))) +vbool16_t vmsle(vbool16_t op0, vbool16_t op1, vint8mf2_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vv_i8mf4_b32))) +vbool32_t vmsle(vint8mf4_t op0, vint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vv_i8mf4_b32_m))) +vbool32_t vmsle(vbool32_t op0, vbool32_t op1, vint8mf4_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vv_i8mf8_b64))) +vbool64_t vmsle(vint8mf8_t op0, vint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vv_i8mf8_b64_m))) +vbool64_t vmsle(vbool64_t op0, vbool64_t op1, vint8mf8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vv_i16m1_b16))) +vbool16_t vmsle(vint16m1_t op0, vint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vv_i16m1_b16_m))) +vbool16_t vmsle(vbool16_t op0, vbool16_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vv_i16m2_b8))) +vbool8_t vmsle(vint16m2_t op0, vint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vv_i16m2_b8_m))) +vbool8_t vmsle(vbool8_t op0, vbool8_t op1, vint16m2_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vv_i16m4_b4))) +vbool4_t vmsle(vint16m4_t op0, vint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vv_i16m4_b4_m))) +vbool4_t vmsle(vbool4_t op0, vbool4_t op1, vint16m4_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vv_i16m8_b2))) +vbool2_t vmsle(vint16m8_t op0, vint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vv_i16m8_b2_m))) +vbool2_t vmsle(vbool2_t op0, vbool2_t op1, vint16m8_t op2, vint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vv_i16mf2_b32))) +vbool32_t vmsle(vint16mf2_t op0, vint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vv_i16mf2_b32_m))) +vbool32_t vmsle(vbool32_t op0, vbool32_t op1, vint16mf2_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vv_i16mf4_b64))) +vbool64_t vmsle(vint16mf4_t op0, vint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vv_i16mf4_b64_m))) +vbool64_t vmsle(vbool64_t op0, vbool64_t op1, vint16mf4_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vv_i32m1_b32))) +vbool32_t vmsle(vint32m1_t op0, vint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vv_i32m1_b32_m))) +vbool32_t vmsle(vbool32_t op0, vbool32_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vv_i32m2_b16))) +vbool16_t vmsle(vint32m2_t op0, vint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vv_i32m2_b16_m))) +vbool16_t vmsle(vbool16_t op0, vbool16_t op1, vint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vv_i32m4_b8))) +vbool8_t vmsle(vint32m4_t op0, vint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vv_i32m4_b8_m))) +vbool8_t vmsle(vbool8_t op0, vbool8_t op1, vint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vv_i32m8_b4))) +vbool4_t vmsle(vint32m8_t op0, vint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vv_i32m8_b4_m))) +vbool4_t vmsle(vbool4_t op0, vbool4_t op1, vint32m8_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vv_i32mf2_b64))) +vbool64_t vmsle(vint32mf2_t op0, vint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vv_i32mf2_b64_m))) +vbool64_t vmsle(vbool64_t op0, vbool64_t op1, vint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vv_i64m1_b64))) +vbool64_t vmsle(vint64m1_t op0, vint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vv_i64m1_b64_m))) +vbool64_t vmsle(vbool64_t op0, vbool64_t op1, vint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vv_i64m2_b32))) +vbool32_t vmsle(vint64m2_t op0, vint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vv_i64m2_b32_m))) +vbool32_t vmsle(vbool32_t op0, vbool32_t op1, vint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vv_i64m4_b16))) +vbool16_t vmsle(vint64m4_t op0, vint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vv_i64m4_b16_m))) +vbool16_t vmsle(vbool16_t op0, vbool16_t op1, vint64m4_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vv_i64m8_b8))) +vbool8_t vmsle(vint64m8_t op0, vint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vv_i64m8_b8_m))) +vbool8_t vmsle(vbool8_t op0, vbool8_t op1, vint64m8_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vx_i8m1_b8))) +vbool8_t vmsle(vint8m1_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vx_i8m1_b8_m))) +vbool8_t vmsle(vbool8_t op0, vbool8_t op1, vint8m1_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vx_i8m2_b4))) +vbool4_t vmsle(vint8m2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vx_i8m2_b4_m))) +vbool4_t vmsle(vbool4_t op0, vbool4_t op1, vint8m2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vx_i8m4_b2))) +vbool2_t vmsle(vint8m4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vx_i8m4_b2_m))) +vbool2_t vmsle(vbool2_t op0, vbool2_t op1, vint8m4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vx_i8m8_b1))) +vbool1_t vmsle(vint8m8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vx_i8m8_b1_m))) +vbool1_t vmsle(vbool1_t op0, vbool1_t op1, vint8m8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vx_i8mf2_b16))) +vbool16_t vmsle(vint8mf2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vx_i8mf2_b16_m))) +vbool16_t vmsle(vbool16_t op0, vbool16_t op1, vint8mf2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vx_i8mf4_b32))) +vbool32_t vmsle(vint8mf4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vx_i8mf4_b32_m))) +vbool32_t vmsle(vbool32_t op0, vbool32_t op1, vint8mf4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vx_i8mf8_b64))) +vbool64_t vmsle(vint8mf8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vx_i8mf8_b64_m))) +vbool64_t vmsle(vbool64_t op0, vbool64_t op1, vint8mf8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vx_i16m1_b16))) +vbool16_t vmsle(vint16m1_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vx_i16m1_b16_m))) +vbool16_t vmsle(vbool16_t op0, vbool16_t op1, vint16m1_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vx_i16m2_b8))) +vbool8_t vmsle(vint16m2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vx_i16m2_b8_m))) +vbool8_t vmsle(vbool8_t op0, vbool8_t op1, vint16m2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vx_i16m4_b4))) +vbool4_t vmsle(vint16m4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vx_i16m4_b4_m))) +vbool4_t vmsle(vbool4_t op0, vbool4_t op1, vint16m4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vx_i16m8_b2))) +vbool2_t vmsle(vint16m8_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vx_i16m8_b2_m))) +vbool2_t vmsle(vbool2_t op0, vbool2_t op1, vint16m8_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vx_i16mf2_b32))) +vbool32_t vmsle(vint16mf2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vx_i16mf2_b32_m))) +vbool32_t vmsle(vbool32_t op0, vbool32_t op1, vint16mf2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vx_i16mf4_b64))) +vbool64_t vmsle(vint16mf4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vx_i16mf4_b64_m))) +vbool64_t vmsle(vbool64_t op0, vbool64_t op1, vint16mf4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vx_i32m1_b32))) +vbool32_t vmsle(vint32m1_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vx_i32m1_b32_m))) +vbool32_t vmsle(vbool32_t op0, vbool32_t op1, vint32m1_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vx_i32m2_b16))) +vbool16_t vmsle(vint32m2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vx_i32m2_b16_m))) +vbool16_t vmsle(vbool16_t op0, vbool16_t op1, vint32m2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vx_i32m4_b8))) +vbool8_t vmsle(vint32m4_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vx_i32m4_b8_m))) +vbool8_t vmsle(vbool8_t op0, vbool8_t op1, vint32m4_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vx_i32m8_b4))) +vbool4_t vmsle(vint32m8_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vx_i32m8_b4_m))) +vbool4_t vmsle(vbool4_t op0, vbool4_t op1, vint32m8_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vx_i32mf2_b64))) +vbool64_t vmsle(vint32mf2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vx_i32mf2_b64_m))) +vbool64_t vmsle(vbool64_t op0, vbool64_t op1, vint32mf2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vx_i64m1_b64))) +vbool64_t vmsle(vint64m1_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vx_i64m1_b64_m))) +vbool64_t vmsle(vbool64_t op0, vbool64_t op1, vint64m1_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vx_i64m2_b32))) +vbool32_t vmsle(vint64m2_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vx_i64m2_b32_m))) +vbool32_t vmsle(vbool32_t op0, vbool32_t op1, vint64m2_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vx_i64m4_b16))) +vbool16_t vmsle(vint64m4_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vx_i64m4_b16_m))) +vbool16_t vmsle(vbool16_t op0, vbool16_t op1, vint64m4_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vx_i64m8_b8))) +vbool8_t vmsle(vint64m8_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsle_vx_i64m8_b8_m))) +vbool8_t vmsle(vbool8_t op0, vbool8_t op1, vint64m8_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vv_u8m1_b8))) +vbool8_t vmsgtu(vuint8m1_t op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vv_u8m1_b8_m))) +vbool8_t vmsgtu(vbool8_t op0, vbool8_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vv_u8m2_b4))) +vbool4_t vmsgtu(vuint8m2_t op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vv_u8m2_b4_m))) +vbool4_t vmsgtu(vbool4_t op0, vbool4_t op1, vuint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vv_u8m4_b2))) +vbool2_t vmsgtu(vuint8m4_t op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vv_u8m4_b2_m))) +vbool2_t vmsgtu(vbool2_t op0, vbool2_t op1, vuint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vv_u8m8_b1))) +vbool1_t vmsgtu(vuint8m8_t op0, vuint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vv_u8m8_b1_m))) +vbool1_t vmsgtu(vbool1_t op0, vbool1_t op1, vuint8m8_t op2, vuint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vv_u8mf2_b16))) +vbool16_t vmsgtu(vuint8mf2_t op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vv_u8mf2_b16_m))) +vbool16_t vmsgtu(vbool16_t op0, vbool16_t op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vv_u8mf4_b32))) +vbool32_t vmsgtu(vuint8mf4_t op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vv_u8mf4_b32_m))) +vbool32_t vmsgtu(vbool32_t op0, vbool32_t op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vv_u8mf8_b64))) +vbool64_t vmsgtu(vuint8mf8_t op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vv_u8mf8_b64_m))) +vbool64_t vmsgtu(vbool64_t op0, vbool64_t op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vv_u16m1_b16))) +vbool16_t vmsgtu(vuint16m1_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vv_u16m1_b16_m))) +vbool16_t vmsgtu(vbool16_t op0, vbool16_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vv_u16m2_b8))) +vbool8_t vmsgtu(vuint16m2_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vv_u16m2_b8_m))) +vbool8_t vmsgtu(vbool8_t op0, vbool8_t op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vv_u16m4_b4))) +vbool4_t vmsgtu(vuint16m4_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vv_u16m4_b4_m))) +vbool4_t vmsgtu(vbool4_t op0, vbool4_t op1, vuint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vv_u16m8_b2))) +vbool2_t vmsgtu(vuint16m8_t op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vv_u16m8_b2_m))) +vbool2_t vmsgtu(vbool2_t op0, vbool2_t op1, vuint16m8_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vv_u16mf2_b32))) +vbool32_t vmsgtu(vuint16mf2_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vv_u16mf2_b32_m))) +vbool32_t vmsgtu(vbool32_t op0, vbool32_t op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vv_u16mf4_b64))) +vbool64_t vmsgtu(vuint16mf4_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vv_u16mf4_b64_m))) +vbool64_t vmsgtu(vbool64_t op0, vbool64_t op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vv_u32m1_b32))) +vbool32_t vmsgtu(vuint32m1_t op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vv_u32m1_b32_m))) +vbool32_t vmsgtu(vbool32_t op0, vbool32_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vv_u32m2_b16))) +vbool16_t vmsgtu(vuint32m2_t op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vv_u32m2_b16_m))) +vbool16_t vmsgtu(vbool16_t op0, vbool16_t op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vv_u32m4_b8))) +vbool8_t vmsgtu(vuint32m4_t op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vv_u32m4_b8_m))) +vbool8_t vmsgtu(vbool8_t op0, vbool8_t op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vv_u32m8_b4))) +vbool4_t vmsgtu(vuint32m8_t op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vv_u32m8_b4_m))) +vbool4_t vmsgtu(vbool4_t op0, vbool4_t op1, vuint32m8_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vv_u32mf2_b64))) +vbool64_t vmsgtu(vuint32mf2_t op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vv_u32mf2_b64_m))) +vbool64_t vmsgtu(vbool64_t op0, vbool64_t op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vv_u64m1_b64))) +vbool64_t vmsgtu(vuint64m1_t op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vv_u64m1_b64_m))) +vbool64_t vmsgtu(vbool64_t op0, vbool64_t op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vv_u64m2_b32))) +vbool32_t vmsgtu(vuint64m2_t op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vv_u64m2_b32_m))) +vbool32_t vmsgtu(vbool32_t op0, vbool32_t op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vv_u64m4_b16))) +vbool16_t vmsgtu(vuint64m4_t op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vv_u64m4_b16_m))) +vbool16_t vmsgtu(vbool16_t op0, vbool16_t op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vv_u64m8_b8))) +vbool8_t vmsgtu(vuint64m8_t op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vv_u64m8_b8_m))) +vbool8_t vmsgtu(vbool8_t op0, vbool8_t op1, vuint64m8_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vx_u8m1_b8))) +vbool8_t vmsgtu(vuint8m1_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vx_u8m1_b8_m))) +vbool8_t vmsgtu(vbool8_t op0, vbool8_t op1, vuint8m1_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vx_u8m2_b4))) +vbool4_t vmsgtu(vuint8m2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vx_u8m2_b4_m))) +vbool4_t vmsgtu(vbool4_t op0, vbool4_t op1, vuint8m2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vx_u8m4_b2))) +vbool2_t vmsgtu(vuint8m4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vx_u8m4_b2_m))) +vbool2_t vmsgtu(vbool2_t op0, vbool2_t op1, vuint8m4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vx_u8m8_b1))) +vbool1_t vmsgtu(vuint8m8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vx_u8m8_b1_m))) +vbool1_t vmsgtu(vbool1_t op0, vbool1_t op1, vuint8m8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vx_u8mf2_b16))) +vbool16_t vmsgtu(vuint8mf2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vx_u8mf2_b16_m))) +vbool16_t vmsgtu(vbool16_t op0, vbool16_t op1, vuint8mf2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vx_u8mf4_b32))) +vbool32_t vmsgtu(vuint8mf4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vx_u8mf4_b32_m))) +vbool32_t vmsgtu(vbool32_t op0, vbool32_t op1, vuint8mf4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vx_u8mf8_b64))) +vbool64_t vmsgtu(vuint8mf8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vx_u8mf8_b64_m))) +vbool64_t vmsgtu(vbool64_t op0, vbool64_t op1, vuint8mf8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vx_u16m1_b16))) +vbool16_t vmsgtu(vuint16m1_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vx_u16m1_b16_m))) +vbool16_t vmsgtu(vbool16_t op0, vbool16_t op1, vuint16m1_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vx_u16m2_b8))) +vbool8_t vmsgtu(vuint16m2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vx_u16m2_b8_m))) +vbool8_t vmsgtu(vbool8_t op0, vbool8_t op1, vuint16m2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vx_u16m4_b4))) +vbool4_t vmsgtu(vuint16m4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vx_u16m4_b4_m))) +vbool4_t vmsgtu(vbool4_t op0, vbool4_t op1, vuint16m4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vx_u16m8_b2))) +vbool2_t vmsgtu(vuint16m8_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vx_u16m8_b2_m))) +vbool2_t vmsgtu(vbool2_t op0, vbool2_t op1, vuint16m8_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vx_u16mf2_b32))) +vbool32_t vmsgtu(vuint16mf2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vx_u16mf2_b32_m))) +vbool32_t vmsgtu(vbool32_t op0, vbool32_t op1, vuint16mf2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vx_u16mf4_b64))) +vbool64_t vmsgtu(vuint16mf4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vx_u16mf4_b64_m))) +vbool64_t vmsgtu(vbool64_t op0, vbool64_t op1, vuint16mf4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vx_u32m1_b32))) +vbool32_t vmsgtu(vuint32m1_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vx_u32m1_b32_m))) +vbool32_t vmsgtu(vbool32_t op0, vbool32_t op1, vuint32m1_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vx_u32m2_b16))) +vbool16_t vmsgtu(vuint32m2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vx_u32m2_b16_m))) +vbool16_t vmsgtu(vbool16_t op0, vbool16_t op1, vuint32m2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vx_u32m4_b8))) +vbool8_t vmsgtu(vuint32m4_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vx_u32m4_b8_m))) +vbool8_t vmsgtu(vbool8_t op0, vbool8_t op1, vuint32m4_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vx_u32m8_b4))) +vbool4_t vmsgtu(vuint32m8_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vx_u32m8_b4_m))) +vbool4_t vmsgtu(vbool4_t op0, vbool4_t op1, vuint32m8_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vx_u32mf2_b64))) +vbool64_t vmsgtu(vuint32mf2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vx_u32mf2_b64_m))) +vbool64_t vmsgtu(vbool64_t op0, vbool64_t op1, vuint32mf2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vx_u64m1_b64))) +vbool64_t vmsgtu(vuint64m1_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vx_u64m1_b64_m))) +vbool64_t vmsgtu(vbool64_t op0, vbool64_t op1, vuint64m1_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vx_u64m2_b32))) +vbool32_t vmsgtu(vuint64m2_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vx_u64m2_b32_m))) +vbool32_t vmsgtu(vbool32_t op0, vbool32_t op1, vuint64m2_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vx_u64m4_b16))) +vbool16_t vmsgtu(vuint64m4_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vx_u64m4_b16_m))) +vbool16_t vmsgtu(vbool16_t op0, vbool16_t op1, vuint64m4_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vx_u64m8_b8))) +vbool8_t vmsgtu(vuint64m8_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgtu_vx_u64m8_b8_m))) +vbool8_t vmsgtu(vbool8_t op0, vbool8_t op1, vuint64m8_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vv_i8m1_b8))) +vbool8_t vmsgt(vint8m1_t op0, vint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vv_i8m1_b8_m))) +vbool8_t vmsgt(vbool8_t op0, vbool8_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vv_i8m2_b4))) +vbool4_t vmsgt(vint8m2_t op0, vint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vv_i8m2_b4_m))) +vbool4_t vmsgt(vbool4_t op0, vbool4_t op1, vint8m2_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vv_i8m4_b2))) +vbool2_t vmsgt(vint8m4_t op0, vint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vv_i8m4_b2_m))) +vbool2_t vmsgt(vbool2_t op0, vbool2_t op1, vint8m4_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vv_i8m8_b1))) +vbool1_t vmsgt(vint8m8_t op0, vint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vv_i8m8_b1_m))) +vbool1_t vmsgt(vbool1_t op0, vbool1_t op1, vint8m8_t op2, vint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vv_i8mf2_b16))) +vbool16_t vmsgt(vint8mf2_t op0, vint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vv_i8mf2_b16_m))) +vbool16_t vmsgt(vbool16_t op0, vbool16_t op1, vint8mf2_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vv_i8mf4_b32))) +vbool32_t vmsgt(vint8mf4_t op0, vint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vv_i8mf4_b32_m))) +vbool32_t vmsgt(vbool32_t op0, vbool32_t op1, vint8mf4_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vv_i8mf8_b64))) +vbool64_t vmsgt(vint8mf8_t op0, vint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vv_i8mf8_b64_m))) +vbool64_t vmsgt(vbool64_t op0, vbool64_t op1, vint8mf8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vv_i16m1_b16))) +vbool16_t vmsgt(vint16m1_t op0, vint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vv_i16m1_b16_m))) +vbool16_t vmsgt(vbool16_t op0, vbool16_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vv_i16m2_b8))) +vbool8_t vmsgt(vint16m2_t op0, vint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vv_i16m2_b8_m))) +vbool8_t vmsgt(vbool8_t op0, vbool8_t op1, vint16m2_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vv_i16m4_b4))) +vbool4_t vmsgt(vint16m4_t op0, vint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vv_i16m4_b4_m))) +vbool4_t vmsgt(vbool4_t op0, vbool4_t op1, vint16m4_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vv_i16m8_b2))) +vbool2_t vmsgt(vint16m8_t op0, vint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vv_i16m8_b2_m))) +vbool2_t vmsgt(vbool2_t op0, vbool2_t op1, vint16m8_t op2, vint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vv_i16mf2_b32))) +vbool32_t vmsgt(vint16mf2_t op0, vint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vv_i16mf2_b32_m))) +vbool32_t vmsgt(vbool32_t op0, vbool32_t op1, vint16mf2_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vv_i16mf4_b64))) +vbool64_t vmsgt(vint16mf4_t op0, vint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vv_i16mf4_b64_m))) +vbool64_t vmsgt(vbool64_t op0, vbool64_t op1, vint16mf4_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vv_i32m1_b32))) +vbool32_t vmsgt(vint32m1_t op0, vint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vv_i32m1_b32_m))) +vbool32_t vmsgt(vbool32_t op0, vbool32_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vv_i32m2_b16))) +vbool16_t vmsgt(vint32m2_t op0, vint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vv_i32m2_b16_m))) +vbool16_t vmsgt(vbool16_t op0, vbool16_t op1, vint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vv_i32m4_b8))) +vbool8_t vmsgt(vint32m4_t op0, vint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vv_i32m4_b8_m))) +vbool8_t vmsgt(vbool8_t op0, vbool8_t op1, vint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vv_i32m8_b4))) +vbool4_t vmsgt(vint32m8_t op0, vint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vv_i32m8_b4_m))) +vbool4_t vmsgt(vbool4_t op0, vbool4_t op1, vint32m8_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vv_i32mf2_b64))) +vbool64_t vmsgt(vint32mf2_t op0, vint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vv_i32mf2_b64_m))) +vbool64_t vmsgt(vbool64_t op0, vbool64_t op1, vint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vv_i64m1_b64))) +vbool64_t vmsgt(vint64m1_t op0, vint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vv_i64m1_b64_m))) +vbool64_t vmsgt(vbool64_t op0, vbool64_t op1, vint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vv_i64m2_b32))) +vbool32_t vmsgt(vint64m2_t op0, vint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vv_i64m2_b32_m))) +vbool32_t vmsgt(vbool32_t op0, vbool32_t op1, vint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vv_i64m4_b16))) +vbool16_t vmsgt(vint64m4_t op0, vint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vv_i64m4_b16_m))) +vbool16_t vmsgt(vbool16_t op0, vbool16_t op1, vint64m4_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vv_i64m8_b8))) +vbool8_t vmsgt(vint64m8_t op0, vint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vv_i64m8_b8_m))) +vbool8_t vmsgt(vbool8_t op0, vbool8_t op1, vint64m8_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vx_i8m1_b8))) +vbool8_t vmsgt(vint8m1_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vx_i8m1_b8_m))) +vbool8_t vmsgt(vbool8_t op0, vbool8_t op1, vint8m1_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vx_i8m2_b4))) +vbool4_t vmsgt(vint8m2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vx_i8m2_b4_m))) +vbool4_t vmsgt(vbool4_t op0, vbool4_t op1, vint8m2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vx_i8m4_b2))) +vbool2_t vmsgt(vint8m4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vx_i8m4_b2_m))) +vbool2_t vmsgt(vbool2_t op0, vbool2_t op1, vint8m4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vx_i8m8_b1))) +vbool1_t vmsgt(vint8m8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vx_i8m8_b1_m))) +vbool1_t vmsgt(vbool1_t op0, vbool1_t op1, vint8m8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vx_i8mf2_b16))) +vbool16_t vmsgt(vint8mf2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vx_i8mf2_b16_m))) +vbool16_t vmsgt(vbool16_t op0, vbool16_t op1, vint8mf2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vx_i8mf4_b32))) +vbool32_t vmsgt(vint8mf4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vx_i8mf4_b32_m))) +vbool32_t vmsgt(vbool32_t op0, vbool32_t op1, vint8mf4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vx_i8mf8_b64))) +vbool64_t vmsgt(vint8mf8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vx_i8mf8_b64_m))) +vbool64_t vmsgt(vbool64_t op0, vbool64_t op1, vint8mf8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vx_i16m1_b16))) +vbool16_t vmsgt(vint16m1_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vx_i16m1_b16_m))) +vbool16_t vmsgt(vbool16_t op0, vbool16_t op1, vint16m1_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vx_i16m2_b8))) +vbool8_t vmsgt(vint16m2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vx_i16m2_b8_m))) +vbool8_t vmsgt(vbool8_t op0, vbool8_t op1, vint16m2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vx_i16m4_b4))) +vbool4_t vmsgt(vint16m4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vx_i16m4_b4_m))) +vbool4_t vmsgt(vbool4_t op0, vbool4_t op1, vint16m4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vx_i16m8_b2))) +vbool2_t vmsgt(vint16m8_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vx_i16m8_b2_m))) +vbool2_t vmsgt(vbool2_t op0, vbool2_t op1, vint16m8_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vx_i16mf2_b32))) +vbool32_t vmsgt(vint16mf2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vx_i16mf2_b32_m))) +vbool32_t vmsgt(vbool32_t op0, vbool32_t op1, vint16mf2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vx_i16mf4_b64))) +vbool64_t vmsgt(vint16mf4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vx_i16mf4_b64_m))) +vbool64_t vmsgt(vbool64_t op0, vbool64_t op1, vint16mf4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vx_i32m1_b32))) +vbool32_t vmsgt(vint32m1_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vx_i32m1_b32_m))) +vbool32_t vmsgt(vbool32_t op0, vbool32_t op1, vint32m1_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vx_i32m2_b16))) +vbool16_t vmsgt(vint32m2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vx_i32m2_b16_m))) +vbool16_t vmsgt(vbool16_t op0, vbool16_t op1, vint32m2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vx_i32m4_b8))) +vbool8_t vmsgt(vint32m4_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vx_i32m4_b8_m))) +vbool8_t vmsgt(vbool8_t op0, vbool8_t op1, vint32m4_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vx_i32m8_b4))) +vbool4_t vmsgt(vint32m8_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vx_i32m8_b4_m))) +vbool4_t vmsgt(vbool4_t op0, vbool4_t op1, vint32m8_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vx_i32mf2_b64))) +vbool64_t vmsgt(vint32mf2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vx_i32mf2_b64_m))) +vbool64_t vmsgt(vbool64_t op0, vbool64_t op1, vint32mf2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vx_i64m1_b64))) +vbool64_t vmsgt(vint64m1_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vx_i64m1_b64_m))) +vbool64_t vmsgt(vbool64_t op0, vbool64_t op1, vint64m1_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vx_i64m2_b32))) +vbool32_t vmsgt(vint64m2_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vx_i64m2_b32_m))) +vbool32_t vmsgt(vbool32_t op0, vbool32_t op1, vint64m2_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vx_i64m4_b16))) +vbool16_t vmsgt(vint64m4_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vx_i64m4_b16_m))) +vbool16_t vmsgt(vbool16_t op0, vbool16_t op1, vint64m4_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vx_i64m8_b8))) +vbool8_t vmsgt(vint64m8_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgt_vx_i64m8_b8_m))) +vbool8_t vmsgt(vbool8_t op0, vbool8_t op1, vint64m8_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vv_u8m1_b8))) +vbool8_t vmsgeu(vuint8m1_t op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vv_u8m1_b8_m))) +vbool8_t vmsgeu(vbool8_t op0, vbool8_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vv_u8m2_b4))) +vbool4_t vmsgeu(vuint8m2_t op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vv_u8m2_b4_m))) +vbool4_t vmsgeu(vbool4_t op0, vbool4_t op1, vuint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vv_u8m4_b2))) +vbool2_t vmsgeu(vuint8m4_t op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vv_u8m4_b2_m))) +vbool2_t vmsgeu(vbool2_t op0, vbool2_t op1, vuint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vv_u8m8_b1))) +vbool1_t vmsgeu(vuint8m8_t op0, vuint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vv_u8m8_b1_m))) +vbool1_t vmsgeu(vbool1_t op0, vbool1_t op1, vuint8m8_t op2, vuint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vv_u8mf2_b16))) +vbool16_t vmsgeu(vuint8mf2_t op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vv_u8mf2_b16_m))) +vbool16_t vmsgeu(vbool16_t op0, vbool16_t op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vv_u8mf4_b32))) +vbool32_t vmsgeu(vuint8mf4_t op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vv_u8mf4_b32_m))) +vbool32_t vmsgeu(vbool32_t op0, vbool32_t op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vv_u8mf8_b64))) +vbool64_t vmsgeu(vuint8mf8_t op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vv_u8mf8_b64_m))) +vbool64_t vmsgeu(vbool64_t op0, vbool64_t op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vv_u16m1_b16))) +vbool16_t vmsgeu(vuint16m1_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vv_u16m1_b16_m))) +vbool16_t vmsgeu(vbool16_t op0, vbool16_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vv_u16m2_b8))) +vbool8_t vmsgeu(vuint16m2_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vv_u16m2_b8_m))) +vbool8_t vmsgeu(vbool8_t op0, vbool8_t op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vv_u16m4_b4))) +vbool4_t vmsgeu(vuint16m4_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vv_u16m4_b4_m))) +vbool4_t vmsgeu(vbool4_t op0, vbool4_t op1, vuint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vv_u16m8_b2))) +vbool2_t vmsgeu(vuint16m8_t op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vv_u16m8_b2_m))) +vbool2_t vmsgeu(vbool2_t op0, vbool2_t op1, vuint16m8_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vv_u16mf2_b32))) +vbool32_t vmsgeu(vuint16mf2_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vv_u16mf2_b32_m))) +vbool32_t vmsgeu(vbool32_t op0, vbool32_t op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vv_u16mf4_b64))) +vbool64_t vmsgeu(vuint16mf4_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vv_u16mf4_b64_m))) +vbool64_t vmsgeu(vbool64_t op0, vbool64_t op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vv_u32m1_b32))) +vbool32_t vmsgeu(vuint32m1_t op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vv_u32m1_b32_m))) +vbool32_t vmsgeu(vbool32_t op0, vbool32_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vv_u32m2_b16))) +vbool16_t vmsgeu(vuint32m2_t op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vv_u32m2_b16_m))) +vbool16_t vmsgeu(vbool16_t op0, vbool16_t op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vv_u32m4_b8))) +vbool8_t vmsgeu(vuint32m4_t op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vv_u32m4_b8_m))) +vbool8_t vmsgeu(vbool8_t op0, vbool8_t op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vv_u32m8_b4))) +vbool4_t vmsgeu(vuint32m8_t op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vv_u32m8_b4_m))) +vbool4_t vmsgeu(vbool4_t op0, vbool4_t op1, vuint32m8_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vv_u32mf2_b64))) +vbool64_t vmsgeu(vuint32mf2_t op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vv_u32mf2_b64_m))) +vbool64_t vmsgeu(vbool64_t op0, vbool64_t op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vv_u64m1_b64))) +vbool64_t vmsgeu(vuint64m1_t op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vv_u64m1_b64_m))) +vbool64_t vmsgeu(vbool64_t op0, vbool64_t op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vv_u64m2_b32))) +vbool32_t vmsgeu(vuint64m2_t op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vv_u64m2_b32_m))) +vbool32_t vmsgeu(vbool32_t op0, vbool32_t op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vv_u64m4_b16))) +vbool16_t vmsgeu(vuint64m4_t op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vv_u64m4_b16_m))) +vbool16_t vmsgeu(vbool16_t op0, vbool16_t op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vv_u64m8_b8))) +vbool8_t vmsgeu(vuint64m8_t op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vv_u64m8_b8_m))) +vbool8_t vmsgeu(vbool8_t op0, vbool8_t op1, vuint64m8_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_i16m1))) +void vsuxei64(int16_t * op0, vuint64m4_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_i16m1_m))) +void vsuxei64(vbool16_t op0, int16_t * op1, vuint64m4_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_i16m2))) +void vsuxei64(int16_t * op0, vuint64m8_t op1, vint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_i16m2_m))) +void vsuxei64(vbool8_t op0, int16_t * op1, vuint64m8_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_i16mf2))) +void vsuxei64(int16_t * op0, vuint64m2_t op1, vint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_i16mf2_m))) +void vsuxei64(vbool32_t op0, int16_t * op1, vuint64m2_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_i16mf4))) +void vsuxei64(int16_t * op0, vuint64m1_t op1, vint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_i16mf4_m))) +void vsuxei64(vbool64_t op0, int16_t * op1, vuint64m1_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vx_u8m1_b8))) +vbool8_t vmsgeu(vuint8m1_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vx_u8m1_b8_m))) +vbool8_t vmsgeu(vbool8_t op0, vbool8_t op1, vuint8m1_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vx_u8m2_b4))) +vbool4_t vmsgeu(vuint8m2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vx_u8m2_b4_m))) +vbool4_t vmsgeu(vbool4_t op0, vbool4_t op1, vuint8m2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vx_u8m4_b2))) +vbool2_t vmsgeu(vuint8m4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vx_u8m4_b2_m))) +vbool2_t vmsgeu(vbool2_t op0, vbool2_t op1, vuint8m4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vx_u8m8_b1))) +vbool1_t vmsgeu(vuint8m8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vx_u8m8_b1_m))) +vbool1_t vmsgeu(vbool1_t op0, vbool1_t op1, vuint8m8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vx_u8mf2_b16))) +vbool16_t vmsgeu(vuint8mf2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vx_u8mf2_b16_m))) +vbool16_t vmsgeu(vbool16_t op0, vbool16_t op1, vuint8mf2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vx_u8mf4_b32))) +vbool32_t vmsgeu(vuint8mf4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vx_u8mf4_b32_m))) +vbool32_t vmsgeu(vbool32_t op0, vbool32_t op1, vuint8mf4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vx_u8mf8_b64))) +vbool64_t vmsgeu(vuint8mf8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vx_u8mf8_b64_m))) +vbool64_t vmsgeu(vbool64_t op0, vbool64_t op1, vuint8mf8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vx_u16m1_b16))) +vbool16_t vmsgeu(vuint16m1_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vx_u16m1_b16_m))) +vbool16_t vmsgeu(vbool16_t op0, vbool16_t op1, vuint16m1_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vx_u16m2_b8))) +vbool8_t vmsgeu(vuint16m2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vx_u16m2_b8_m))) +vbool8_t vmsgeu(vbool8_t op0, vbool8_t op1, vuint16m2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vx_u16m4_b4))) +vbool4_t vmsgeu(vuint16m4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vx_u16m4_b4_m))) +vbool4_t vmsgeu(vbool4_t op0, vbool4_t op1, vuint16m4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vx_u16m8_b2))) +vbool2_t vmsgeu(vuint16m8_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vx_u16m8_b2_m))) +vbool2_t vmsgeu(vbool2_t op0, vbool2_t op1, vuint16m8_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vx_u16mf2_b32))) +vbool32_t vmsgeu(vuint16mf2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vx_u16mf2_b32_m))) +vbool32_t vmsgeu(vbool32_t op0, vbool32_t op1, vuint16mf2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vx_u16mf4_b64))) +vbool64_t vmsgeu(vuint16mf4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vx_u16mf4_b64_m))) +vbool64_t vmsgeu(vbool64_t op0, vbool64_t op1, vuint16mf4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vx_u32m1_b32))) +vbool32_t vmsgeu(vuint32m1_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vx_u32m1_b32_m))) +vbool32_t vmsgeu(vbool32_t op0, vbool32_t op1, vuint32m1_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vx_u32m2_b16))) +vbool16_t vmsgeu(vuint32m2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vx_u32m2_b16_m))) +vbool16_t vmsgeu(vbool16_t op0, vbool16_t op1, vuint32m2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vx_u32m4_b8))) +vbool8_t vmsgeu(vuint32m4_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vx_u32m4_b8_m))) +vbool8_t vmsgeu(vbool8_t op0, vbool8_t op1, vuint32m4_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vx_u32m8_b4))) +vbool4_t vmsgeu(vuint32m8_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vx_u32m8_b4_m))) +vbool4_t vmsgeu(vbool4_t op0, vbool4_t op1, vuint32m8_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vx_u32mf2_b64))) +vbool64_t vmsgeu(vuint32mf2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vx_u32mf2_b64_m))) +vbool64_t vmsgeu(vbool64_t op0, vbool64_t op1, vuint32mf2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vx_u64m1_b64))) +vbool64_t vmsgeu(vuint64m1_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vx_u64m1_b64_m))) +vbool64_t vmsgeu(vbool64_t op0, vbool64_t op1, vuint64m1_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vx_u64m2_b32))) +vbool32_t vmsgeu(vuint64m2_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vx_u64m2_b32_m))) +vbool32_t vmsgeu(vbool32_t op0, vbool32_t op1, vuint64m2_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vx_u64m4_b16))) +vbool16_t vmsgeu(vuint64m4_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vx_u64m4_b16_m))) +vbool16_t vmsgeu(vbool16_t op0, vbool16_t op1, vuint64m4_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vx_u64m8_b8))) +vbool8_t vmsgeu(vuint64m8_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsgeu_vx_u64m8_b8_m))) +vbool8_t vmsgeu(vbool8_t op0, vbool8_t op1, vuint64m8_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vv_i8m1_b8))) +vbool8_t vmsge(vint8m1_t op0, vint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vv_i8m1_b8_m))) +vbool8_t vmsge(vbool8_t op0, vbool8_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vv_i8m2_b4))) +vbool4_t vmsge(vint8m2_t op0, vint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vv_i8m2_b4_m))) +vbool4_t vmsge(vbool4_t op0, vbool4_t op1, vint8m2_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vv_i8m4_b2))) +vbool2_t vmsge(vint8m4_t op0, vint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vv_i8m4_b2_m))) +vbool2_t vmsge(vbool2_t op0, vbool2_t op1, vint8m4_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vv_i8m8_b1))) +vbool1_t vmsge(vint8m8_t op0, vint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vv_i8m8_b1_m))) +vbool1_t vmsge(vbool1_t op0, vbool1_t op1, vint8m8_t op2, vint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vv_i8mf2_b16))) +vbool16_t vmsge(vint8mf2_t op0, vint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vv_i8mf2_b16_m))) +vbool16_t vmsge(vbool16_t op0, vbool16_t op1, vint8mf2_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vv_i8mf4_b32))) +vbool32_t vmsge(vint8mf4_t op0, vint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vv_i8mf4_b32_m))) +vbool32_t vmsge(vbool32_t op0, vbool32_t op1, vint8mf4_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vv_i8mf8_b64))) +vbool64_t vmsge(vint8mf8_t op0, vint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vv_i8mf8_b64_m))) +vbool64_t vmsge(vbool64_t op0, vbool64_t op1, vint8mf8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vv_i16m1_b16))) +vbool16_t vmsge(vint16m1_t op0, vint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vv_i16m1_b16_m))) +vbool16_t vmsge(vbool16_t op0, vbool16_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vv_i16m2_b8))) +vbool8_t vmsge(vint16m2_t op0, vint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vv_i16m2_b8_m))) +vbool8_t vmsge(vbool8_t op0, vbool8_t op1, vint16m2_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vv_i16m4_b4))) +vbool4_t vmsge(vint16m4_t op0, vint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vv_i16m4_b4_m))) +vbool4_t vmsge(vbool4_t op0, vbool4_t op1, vint16m4_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vv_i16m8_b2))) +vbool2_t vmsge(vint16m8_t op0, vint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vv_i16m8_b2_m))) +vbool2_t vmsge(vbool2_t op0, vbool2_t op1, vint16m8_t op2, vint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vv_i16mf2_b32))) +vbool32_t vmsge(vint16mf2_t op0, vint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vv_i16mf2_b32_m))) +vbool32_t vmsge(vbool32_t op0, vbool32_t op1, vint16mf2_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vv_i16mf4_b64))) +vbool64_t vmsge(vint16mf4_t op0, vint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vv_i16mf4_b64_m))) +vbool64_t vmsge(vbool64_t op0, vbool64_t op1, vint16mf4_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vv_i32m1_b32))) +vbool32_t vmsge(vint32m1_t op0, vint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vv_i32m1_b32_m))) +vbool32_t vmsge(vbool32_t op0, vbool32_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vv_i32m2_b16))) +vbool16_t vmsge(vint32m2_t op0, vint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vv_i32m2_b16_m))) +vbool16_t vmsge(vbool16_t op0, vbool16_t op1, vint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vv_i32m4_b8))) +vbool8_t vmsge(vint32m4_t op0, vint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vv_i32m4_b8_m))) +vbool8_t vmsge(vbool8_t op0, vbool8_t op1, vint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vv_i32m8_b4))) +vbool4_t vmsge(vint32m8_t op0, vint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vv_i32m8_b4_m))) +vbool4_t vmsge(vbool4_t op0, vbool4_t op1, vint32m8_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vv_i32mf2_b64))) +vbool64_t vmsge(vint32mf2_t op0, vint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vv_i32mf2_b64_m))) +vbool64_t vmsge(vbool64_t op0, vbool64_t op1, vint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vv_i64m1_b64))) +vbool64_t vmsge(vint64m1_t op0, vint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vv_i64m1_b64_m))) +vbool64_t vmsge(vbool64_t op0, vbool64_t op1, vint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vv_i64m2_b32))) +vbool32_t vmsge(vint64m2_t op0, vint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vv_i64m2_b32_m))) +vbool32_t vmsge(vbool32_t op0, vbool32_t op1, vint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vv_i64m4_b16))) +vbool16_t vmsge(vint64m4_t op0, vint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vv_i64m4_b16_m))) +vbool16_t vmsge(vbool16_t op0, vbool16_t op1, vint64m4_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vv_i64m8_b8))) +vbool8_t vmsge(vint64m8_t op0, vint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vv_i64m8_b8_m))) +vbool8_t vmsge(vbool8_t op0, vbool8_t op1, vint64m8_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vx_i8m1_b8))) +vbool8_t vmsge(vint8m1_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vx_i8m1_b8_m))) +vbool8_t vmsge(vbool8_t op0, vbool8_t op1, vint8m1_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vx_i8m2_b4))) +vbool4_t vmsge(vint8m2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vx_i8m2_b4_m))) +vbool4_t vmsge(vbool4_t op0, vbool4_t op1, vint8m2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vx_i8m4_b2))) +vbool2_t vmsge(vint8m4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vx_i8m4_b2_m))) +vbool2_t vmsge(vbool2_t op0, vbool2_t op1, vint8m4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vx_i8m8_b1))) +vbool1_t vmsge(vint8m8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vx_i8m8_b1_m))) +vbool1_t vmsge(vbool1_t op0, vbool1_t op1, vint8m8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vx_i8mf2_b16))) +vbool16_t vmsge(vint8mf2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vx_i8mf2_b16_m))) +vbool16_t vmsge(vbool16_t op0, vbool16_t op1, vint8mf2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vx_i8mf4_b32))) +vbool32_t vmsge(vint8mf4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vx_i8mf4_b32_m))) +vbool32_t vmsge(vbool32_t op0, vbool32_t op1, vint8mf4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vx_i8mf8_b64))) +vbool64_t vmsge(vint8mf8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vx_i8mf8_b64_m))) +vbool64_t vmsge(vbool64_t op0, vbool64_t op1, vint8mf8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vx_i16m1_b16))) +vbool16_t vmsge(vint16m1_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vx_i16m1_b16_m))) +vbool16_t vmsge(vbool16_t op0, vbool16_t op1, vint16m1_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vx_i16m2_b8))) +vbool8_t vmsge(vint16m2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vx_i16m2_b8_m))) +vbool8_t vmsge(vbool8_t op0, vbool8_t op1, vint16m2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vx_i16m4_b4))) +vbool4_t vmsge(vint16m4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vx_i16m4_b4_m))) +vbool4_t vmsge(vbool4_t op0, vbool4_t op1, vint16m4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vx_i16m8_b2))) +vbool2_t vmsge(vint16m8_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vx_i16m8_b2_m))) +vbool2_t vmsge(vbool2_t op0, vbool2_t op1, vint16m8_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vx_i16mf2_b32))) +vbool32_t vmsge(vint16mf2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vx_i16mf2_b32_m))) +vbool32_t vmsge(vbool32_t op0, vbool32_t op1, vint16mf2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vx_i16mf4_b64))) +vbool64_t vmsge(vint16mf4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vx_i16mf4_b64_m))) +vbool64_t vmsge(vbool64_t op0, vbool64_t op1, vint16mf4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vx_i32m1_b32))) +vbool32_t vmsge(vint32m1_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vx_i32m1_b32_m))) +vbool32_t vmsge(vbool32_t op0, vbool32_t op1, vint32m1_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vx_i32m2_b16))) +vbool16_t vmsge(vint32m2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vx_i32m2_b16_m))) +vbool16_t vmsge(vbool16_t op0, vbool16_t op1, vint32m2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vx_i32m4_b8))) +vbool8_t vmsge(vint32m4_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vx_i32m4_b8_m))) +vbool8_t vmsge(vbool8_t op0, vbool8_t op1, vint32m4_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vx_i32m8_b4))) +vbool4_t vmsge(vint32m8_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vx_i32m8_b4_m))) +vbool4_t vmsge(vbool4_t op0, vbool4_t op1, vint32m8_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vx_i32mf2_b64))) +vbool64_t vmsge(vint32mf2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vx_i32mf2_b64_m))) +vbool64_t vmsge(vbool64_t op0, vbool64_t op1, vint32mf2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vx_i64m1_b64))) +vbool64_t vmsge(vint64m1_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vx_i64m1_b64_m))) +vbool64_t vmsge(vbool64_t op0, vbool64_t op1, vint64m1_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vx_i64m2_b32))) +vbool32_t vmsge(vint64m2_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vx_i64m2_b32_m))) +vbool32_t vmsge(vbool32_t op0, vbool32_t op1, vint64m2_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vx_i64m4_b16))) +vbool16_t vmsge(vint64m4_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vx_i64m4_b16_m))) +vbool16_t vmsge(vbool16_t op0, vbool16_t op1, vint64m4_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vx_i64m8_b8))) +vbool8_t vmsge(vint64m8_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsge_vx_i64m8_b8_m))) +vbool8_t vmsge(vbool8_t op0, vbool8_t op1, vint64m8_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vv_u8m1))) +vuint8m1_t vminu(vuint8m1_t op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vv_u8m1_m))) +vuint8m1_t vminu(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vv_u8m2))) +vuint8m2_t vminu(vuint8m2_t op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vv_u8m2_m))) +vuint8m2_t vminu(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vv_u8m4))) +vuint8m4_t vminu(vuint8m4_t op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vv_u8m4_m))) +vuint8m4_t vminu(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vv_u8m8))) +vuint8m8_t vminu(vuint8m8_t op0, vuint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vv_u8m8_m))) +vuint8m8_t vminu(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, vuint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vv_u8mf2))) +vuint8mf2_t vminu(vuint8mf2_t op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vv_u8mf2_m))) +vuint8mf2_t vminu(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vv_u8mf4))) +vuint8mf4_t vminu(vuint8mf4_t op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vv_u8mf4_m))) +vuint8mf4_t vminu(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vv_u8mf8))) +vuint8mf8_t vminu(vuint8mf8_t op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vv_u8mf8_m))) +vuint8mf8_t vminu(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vv_u16m1))) +vuint16m1_t vminu(vuint16m1_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vv_u16m1_m))) +vuint16m1_t vminu(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vv_u16m2))) +vuint16m2_t vminu(vuint16m2_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vv_u16m2_m))) +vuint16m2_t vminu(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vv_u16m4))) +vuint16m4_t vminu(vuint16m4_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vv_u16m4_m))) +vuint16m4_t vminu(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vv_u16m8))) +vuint16m8_t vminu(vuint16m8_t op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vv_u16m8_m))) +vuint16m8_t vminu(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vv_u16mf2))) +vuint16mf2_t vminu(vuint16mf2_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vv_u16mf2_m))) +vuint16mf2_t vminu(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vv_u16mf4))) +vuint16mf4_t vminu(vuint16mf4_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vv_u16mf4_m))) +vuint16mf4_t vminu(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vv_u32m1))) +vuint32m1_t vminu(vuint32m1_t op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vv_u32m1_m))) +vuint32m1_t vminu(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vv_u32m2))) +vuint32m2_t vminu(vuint32m2_t op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vv_u32m2_m))) +vuint32m2_t vminu(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vv_u32m4))) +vuint32m4_t vminu(vuint32m4_t op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vv_u32m4_m))) +vuint32m4_t vminu(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vv_u32m8))) +vuint32m8_t vminu(vuint32m8_t op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vv_u32m8_m))) +vuint32m8_t vminu(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vv_u32mf2))) +vuint32mf2_t vminu(vuint32mf2_t op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vv_u32mf2_m))) +vuint32mf2_t vminu(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vv_u64m1))) +vuint64m1_t vminu(vuint64m1_t op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vv_u64m1_m))) +vuint64m1_t vminu(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vv_u64m2))) +vuint64m2_t vminu(vuint64m2_t op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vv_u64m2_m))) +vuint64m2_t vminu(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vv_u64m4))) +vuint64m4_t vminu(vuint64m4_t op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vv_u64m4_m))) +vuint64m4_t vminu(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vv_u64m8))) +vuint64m8_t vminu(vuint64m8_t op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vv_u64m8_m))) +vuint64m8_t vminu(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vx_u8m1))) +vuint8m1_t vminu(vuint8m1_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vx_u8m1_m))) +vuint8m1_t vminu(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vx_u8m2))) +vuint8m2_t vminu(vuint8m2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vx_u8m2_m))) +vuint8m2_t vminu(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vx_u8m4))) +vuint8m4_t vminu(vuint8m4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vx_u8m4_m))) +vuint8m4_t vminu(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vx_u8m8))) +vuint8m8_t vminu(vuint8m8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vx_u8m8_m))) +vuint8m8_t vminu(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vx_u8mf2))) +vuint8mf2_t vminu(vuint8mf2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vx_u8mf2_m))) +vuint8mf2_t vminu(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vx_u8mf4))) +vuint8mf4_t vminu(vuint8mf4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vx_u8mf4_m))) +vuint8mf4_t vminu(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vx_u8mf8))) +vuint8mf8_t vminu(vuint8mf8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vx_u8mf8_m))) +vuint8mf8_t vminu(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vx_u16m1))) +vuint16m1_t vminu(vuint16m1_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vx_u16m1_m))) +vuint16m1_t vminu(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vx_u16m2))) +vuint16m2_t vminu(vuint16m2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vx_u16m2_m))) +vuint16m2_t vminu(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vx_u16m4))) +vuint16m4_t vminu(vuint16m4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vx_u16m4_m))) +vuint16m4_t vminu(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vx_u16m8))) +vuint16m8_t vminu(vuint16m8_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vx_u16m8_m))) +vuint16m8_t vminu(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vx_u16mf2))) +vuint16mf2_t vminu(vuint16mf2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vx_u16mf2_m))) +vuint16mf2_t vminu(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vx_u16mf4))) +vuint16mf4_t vminu(vuint16mf4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vx_u16mf4_m))) +vuint16mf4_t vminu(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vx_u32m1))) +vuint32m1_t vminu(vuint32m1_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vx_u32m1_m))) +vuint32m1_t vminu(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vx_u32m2))) +vuint32m2_t vminu(vuint32m2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vx_u32m2_m))) +vuint32m2_t vminu(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vx_u32m4))) +vuint32m4_t vminu(vuint32m4_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vx_u32m4_m))) +vuint32m4_t vminu(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vx_u32m8))) +vuint32m8_t vminu(vuint32m8_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vx_u32m8_m))) +vuint32m8_t vminu(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vx_u32mf2))) +vuint32mf2_t vminu(vuint32mf2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vx_u32mf2_m))) +vuint32mf2_t vminu(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vx_u64m1))) +vuint64m1_t vminu(vuint64m1_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vx_u64m1_m))) +vuint64m1_t vminu(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vx_u64m2))) +vuint64m2_t vminu(vuint64m2_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vx_u64m2_m))) +vuint64m2_t vminu(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vx_u64m4))) +vuint64m4_t vminu(vuint64m4_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vx_u64m4_m))) +vuint64m4_t vminu(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vx_u64m8))) +vuint64m8_t vminu(vuint64m8_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vminu_vx_u64m8_m))) +vuint64m8_t vminu(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vv_i8m1))) +vint8m1_t vmin(vint8m1_t op0, vint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vv_i8m1_m))) +vint8m1_t vmin(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vv_i8m2))) +vint8m2_t vmin(vint8m2_t op0, vint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vv_i8m2_m))) +vint8m2_t vmin(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vv_i8m4))) +vint8m4_t vmin(vint8m4_t op0, vint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vv_i8m4_m))) +vint8m4_t vmin(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vv_i8m8))) +vint8m8_t vmin(vint8m8_t op0, vint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vv_i8m8_m))) +vint8m8_t vmin(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, vint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vv_i8mf2))) +vint8mf2_t vmin(vint8mf2_t op0, vint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vv_i8mf2_m))) +vint8mf2_t vmin(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vv_i8mf4))) +vint8mf4_t vmin(vint8mf4_t op0, vint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vv_i8mf4_m))) +vint8mf4_t vmin(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vv_i8mf8))) +vint8mf8_t vmin(vint8mf8_t op0, vint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vv_i8mf8_m))) +vint8mf8_t vmin(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vv_i16m1))) +vint16m1_t vmin(vint16m1_t op0, vint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vv_i16m1_m))) +vint16m1_t vmin(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vv_i16m2))) +vint16m2_t vmin(vint16m2_t op0, vint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vv_i16m2_m))) +vint16m2_t vmin(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vv_i16m4))) +vint16m4_t vmin(vint16m4_t op0, vint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vv_i16m4_m))) +vint16m4_t vmin(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vv_i16m8))) +vint16m8_t vmin(vint16m8_t op0, vint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vv_i16m8_m))) +vint16m8_t vmin(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, vint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vv_i16mf2))) +vint16mf2_t vmin(vint16mf2_t op0, vint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vv_i16mf2_m))) +vint16mf2_t vmin(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vv_i16mf4))) +vint16mf4_t vmin(vint16mf4_t op0, vint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vv_i16mf4_m))) +vint16mf4_t vmin(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vv_i32m1))) +vint32m1_t vmin(vint32m1_t op0, vint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vv_i32m1_m))) +vint32m1_t vmin(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vv_i32m2))) +vint32m2_t vmin(vint32m2_t op0, vint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vv_i32m2_m))) +vint32m2_t vmin(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vv_i32m4))) +vint32m4_t vmin(vint32m4_t op0, vint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vv_i32m4_m))) +vint32m4_t vmin(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vv_i32m8))) +vint32m8_t vmin(vint32m8_t op0, vint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vv_i32m8_m))) +vint32m8_t vmin(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vv_i32mf2))) +vint32mf2_t vmin(vint32mf2_t op0, vint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vv_i32mf2_m))) +vint32mf2_t vmin(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vv_i64m1))) +vint64m1_t vmin(vint64m1_t op0, vint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vv_i64m1_m))) +vint64m1_t vmin(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vv_i64m2))) +vint64m2_t vmin(vint64m2_t op0, vint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vv_i64m2_m))) +vint64m2_t vmin(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vv_i64m4))) +vint64m4_t vmin(vint64m4_t op0, vint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vv_i64m4_m))) +vint64m4_t vmin(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vv_i64m8))) +vint64m8_t vmin(vint64m8_t op0, vint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vv_i64m8_m))) +vint64m8_t vmin(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vx_i8m1))) +vint8m1_t vmin(vint8m1_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vx_i8m1_m))) +vint8m1_t vmin(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vx_i8m2))) +vint8m2_t vmin(vint8m2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vx_i8m2_m))) +vint8m2_t vmin(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vx_i8m4))) +vint8m4_t vmin(vint8m4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vx_i8m4_m))) +vint8m4_t vmin(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vx_i8m8))) +vint8m8_t vmin(vint8m8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vx_i8m8_m))) +vint8m8_t vmin(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vx_i8mf2))) +vint8mf2_t vmin(vint8mf2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vx_i8mf2_m))) +vint8mf2_t vmin(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vx_i8mf4))) +vint8mf4_t vmin(vint8mf4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vx_i8mf4_m))) +vint8mf4_t vmin(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vx_i8mf8))) +vint8mf8_t vmin(vint8mf8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vx_i8mf8_m))) +vint8mf8_t vmin(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vx_i16m1))) +vint16m1_t vmin(vint16m1_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vx_i16m1_m))) +vint16m1_t vmin(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vx_i16m2))) +vint16m2_t vmin(vint16m2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vx_i16m2_m))) +vint16m2_t vmin(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vx_i16m4))) +vint16m4_t vmin(vint16m4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vx_i16m4_m))) +vint16m4_t vmin(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vx_i16m8))) +vint16m8_t vmin(vint16m8_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vx_i16m8_m))) +vint16m8_t vmin(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vx_i16mf2))) +vint16mf2_t vmin(vint16mf2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vx_i16mf2_m))) +vint16mf2_t vmin(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vx_i16mf4))) +vint16mf4_t vmin(vint16mf4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vx_i16mf4_m))) +vint16mf4_t vmin(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vx_i32m1))) +vint32m1_t vmin(vint32m1_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vx_i32m1_m))) +vint32m1_t vmin(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vx_i32m2))) +vint32m2_t vmin(vint32m2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vx_i32m2_m))) +vint32m2_t vmin(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vx_i32m4))) +vint32m4_t vmin(vint32m4_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vx_i32m4_m))) +vint32m4_t vmin(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vx_i32m8))) +vint32m8_t vmin(vint32m8_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vx_i32m8_m))) +vint32m8_t vmin(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vx_i32mf2))) +vint32mf2_t vmin(vint32mf2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vx_i32mf2_m))) +vint32mf2_t vmin(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vx_i64m1))) +vint64m1_t vmin(vint64m1_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vx_i64m1_m))) +vint64m1_t vmin(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vx_i64m2))) +vint64m2_t vmin(vint64m2_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vx_i64m2_m))) +vint64m2_t vmin(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vx_i64m4))) +vint64m4_t vmin(vint64m4_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vx_i64m4_m))) +vint64m4_t vmin(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vx_i64m8))) +vint64m8_t vmin(vint64m8_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmin_vx_i64m8_m))) +vint64m8_t vmin(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vv_u8m1))) +vuint8m1_t vmaxu(vuint8m1_t op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vv_u8m1_m))) +vuint8m1_t vmaxu(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vv_u8m2))) +vuint8m2_t vmaxu(vuint8m2_t op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vv_u8m2_m))) +vuint8m2_t vmaxu(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vv_u8m4))) +vuint8m4_t vmaxu(vuint8m4_t op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vv_u8m4_m))) +vuint8m4_t vmaxu(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vv_u8m8))) +vuint8m8_t vmaxu(vuint8m8_t op0, vuint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vv_u8m8_m))) +vuint8m8_t vmaxu(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, vuint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vv_u8mf2))) +vuint8mf2_t vmaxu(vuint8mf2_t op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vv_u8mf2_m))) +vuint8mf2_t vmaxu(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vv_u8mf4))) +vuint8mf4_t vmaxu(vuint8mf4_t op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vv_u8mf4_m))) +vuint8mf4_t vmaxu(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vv_u8mf8))) +vuint8mf8_t vmaxu(vuint8mf8_t op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vv_u8mf8_m))) +vuint8mf8_t vmaxu(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vv_u16m1))) +vuint16m1_t vmaxu(vuint16m1_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vv_u16m1_m))) +vuint16m1_t vmaxu(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vv_u16m2))) +vuint16m2_t vmaxu(vuint16m2_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vv_u16m2_m))) +vuint16m2_t vmaxu(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vv_u16m4))) +vuint16m4_t vmaxu(vuint16m4_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vv_u16m4_m))) +vuint16m4_t vmaxu(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vv_u16m8))) +vuint16m8_t vmaxu(vuint16m8_t op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vv_u16m8_m))) +vuint16m8_t vmaxu(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vv_u16mf2))) +vuint16mf2_t vmaxu(vuint16mf2_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vv_u16mf2_m))) +vuint16mf2_t vmaxu(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vv_u16mf4))) +vuint16mf4_t vmaxu(vuint16mf4_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vv_u16mf4_m))) +vuint16mf4_t vmaxu(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vv_u32m1))) +vuint32m1_t vmaxu(vuint32m1_t op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vv_u32m1_m))) +vuint32m1_t vmaxu(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vv_u32m2))) +vuint32m2_t vmaxu(vuint32m2_t op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vv_u32m2_m))) +vuint32m2_t vmaxu(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vv_u32m4))) +vuint32m4_t vmaxu(vuint32m4_t op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vv_u32m4_m))) +vuint32m4_t vmaxu(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vv_u32m8))) +vuint32m8_t vmaxu(vuint32m8_t op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vv_u32m8_m))) +vuint32m8_t vmaxu(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vv_u32mf2))) +vuint32mf2_t vmaxu(vuint32mf2_t op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vv_u32mf2_m))) +vuint32mf2_t vmaxu(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vv_u64m1))) +vuint64m1_t vmaxu(vuint64m1_t op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vv_u64m1_m))) +vuint64m1_t vmaxu(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vv_u64m2))) +vuint64m2_t vmaxu(vuint64m2_t op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vv_u64m2_m))) +vuint64m2_t vmaxu(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vv_u64m4))) +vuint64m4_t vmaxu(vuint64m4_t op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vv_u64m4_m))) +vuint64m4_t vmaxu(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vv_u64m8))) +vuint64m8_t vmaxu(vuint64m8_t op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vv_u64m8_m))) +vuint64m8_t vmaxu(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vx_u8m1))) +vuint8m1_t vmaxu(vuint8m1_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vx_u8m1_m))) +vuint8m1_t vmaxu(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vx_u8m2))) +vuint8m2_t vmaxu(vuint8m2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vx_u8m2_m))) +vuint8m2_t vmaxu(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vx_u8m4))) +vuint8m4_t vmaxu(vuint8m4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vx_u8m4_m))) +vuint8m4_t vmaxu(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vx_u8m8))) +vuint8m8_t vmaxu(vuint8m8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vx_u8m8_m))) +vuint8m8_t vmaxu(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vx_u8mf2))) +vuint8mf2_t vmaxu(vuint8mf2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vx_u8mf2_m))) +vuint8mf2_t vmaxu(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vx_u8mf4))) +vuint8mf4_t vmaxu(vuint8mf4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vx_u8mf4_m))) +vuint8mf4_t vmaxu(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vx_u8mf8))) +vuint8mf8_t vmaxu(vuint8mf8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vx_u8mf8_m))) +vuint8mf8_t vmaxu(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vx_u16m1))) +vuint16m1_t vmaxu(vuint16m1_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vx_u16m1_m))) +vuint16m1_t vmaxu(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vx_u16m2))) +vuint16m2_t vmaxu(vuint16m2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vx_u16m2_m))) +vuint16m2_t vmaxu(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vx_u16m4))) +vuint16m4_t vmaxu(vuint16m4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vx_u16m4_m))) +vuint16m4_t vmaxu(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vx_u16m8))) +vuint16m8_t vmaxu(vuint16m8_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vx_u16m8_m))) +vuint16m8_t vmaxu(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vx_u16mf2))) +vuint16mf2_t vmaxu(vuint16mf2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vx_u16mf2_m))) +vuint16mf2_t vmaxu(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vx_u16mf4))) +vuint16mf4_t vmaxu(vuint16mf4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vx_u16mf4_m))) +vuint16mf4_t vmaxu(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vx_u32m1))) +vuint32m1_t vmaxu(vuint32m1_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vx_u32m1_m))) +vuint32m1_t vmaxu(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vx_u32m2))) +vuint32m2_t vmaxu(vuint32m2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vx_u32m2_m))) +vuint32m2_t vmaxu(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vx_u32m4))) +vuint32m4_t vmaxu(vuint32m4_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vx_u32m4_m))) +vuint32m4_t vmaxu(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vx_u32m8))) +vuint32m8_t vmaxu(vuint32m8_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vx_u32m8_m))) +vuint32m8_t vmaxu(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vx_u32mf2))) +vuint32mf2_t vmaxu(vuint32mf2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vx_u32mf2_m))) +vuint32mf2_t vmaxu(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vx_u64m1))) +vuint64m1_t vmaxu(vuint64m1_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vx_u64m1_m))) +vuint64m1_t vmaxu(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vx_u64m2))) +vuint64m2_t vmaxu(vuint64m2_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vx_u64m2_m))) +vuint64m2_t vmaxu(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vx_u64m4))) +vuint64m4_t vmaxu(vuint64m4_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vx_u64m4_m))) +vuint64m4_t vmaxu(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vx_u64m8))) +vuint64m8_t vmaxu(vuint64m8_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmaxu_vx_u64m8_m))) +vuint64m8_t vmaxu(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vv_i8m1))) +vint8m1_t vmax(vint8m1_t op0, vint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vv_i8m1_m))) +vint8m1_t vmax(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vv_i8m2))) +vint8m2_t vmax(vint8m2_t op0, vint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vv_i8m2_m))) +vint8m2_t vmax(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vv_i8m4))) +vint8m4_t vmax(vint8m4_t op0, vint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vv_i8m4_m))) +vint8m4_t vmax(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vv_i8m8))) +vint8m8_t vmax(vint8m8_t op0, vint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vv_i8m8_m))) +vint8m8_t vmax(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, vint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vv_i8mf2))) +vint8mf2_t vmax(vint8mf2_t op0, vint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vv_i8mf2_m))) +vint8mf2_t vmax(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vv_i8mf4))) +vint8mf4_t vmax(vint8mf4_t op0, vint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vv_i8mf4_m))) +vint8mf4_t vmax(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vv_i8mf8))) +vint8mf8_t vmax(vint8mf8_t op0, vint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vv_i8mf8_m))) +vint8mf8_t vmax(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vv_i16m1))) +vint16m1_t vmax(vint16m1_t op0, vint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vv_i16m1_m))) +vint16m1_t vmax(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vv_i16m2))) +vint16m2_t vmax(vint16m2_t op0, vint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vv_i16m2_m))) +vint16m2_t vmax(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vv_i16m4))) +vint16m4_t vmax(vint16m4_t op0, vint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vv_i16m4_m))) +vint16m4_t vmax(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vv_i16m8))) +vint16m8_t vmax(vint16m8_t op0, vint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vv_i16m8_m))) +vint16m8_t vmax(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, vint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vv_i16mf2))) +vint16mf2_t vmax(vint16mf2_t op0, vint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vv_i16mf2_m))) +vint16mf2_t vmax(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vv_i16mf4))) +vint16mf4_t vmax(vint16mf4_t op0, vint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vv_i16mf4_m))) +vint16mf4_t vmax(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vv_i32m1))) +vint32m1_t vmax(vint32m1_t op0, vint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vv_i32m1_m))) +vint32m1_t vmax(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vv_i32m2))) +vint32m2_t vmax(vint32m2_t op0, vint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vv_i32m2_m))) +vint32m2_t vmax(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vv_i32m4))) +vint32m4_t vmax(vint32m4_t op0, vint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vv_i32m4_m))) +vint32m4_t vmax(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vv_i32m8))) +vint32m8_t vmax(vint32m8_t op0, vint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vv_i32m8_m))) +vint32m8_t vmax(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vv_i32mf2))) +vint32mf2_t vmax(vint32mf2_t op0, vint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vv_i32mf2_m))) +vint32mf2_t vmax(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vv_i64m1))) +vint64m1_t vmax(vint64m1_t op0, vint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vv_i64m1_m))) +vint64m1_t vmax(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vv_i64m2))) +vint64m2_t vmax(vint64m2_t op0, vint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vv_i64m2_m))) +vint64m2_t vmax(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vv_i64m4))) +vint64m4_t vmax(vint64m4_t op0, vint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vv_i64m4_m))) +vint64m4_t vmax(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vv_i64m8))) +vint64m8_t vmax(vint64m8_t op0, vint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vv_i64m8_m))) +vint64m8_t vmax(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_u16m1))) +void vsuxei64(uint16_t * op0, vuint64m4_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_u16m1_m))) +void vsuxei64(vbool16_t op0, uint16_t * op1, vuint64m4_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_u16m2))) +void vsuxei64(uint16_t * op0, vuint64m8_t op1, vuint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_u16m2_m))) +void vsuxei64(vbool8_t op0, uint16_t * op1, vuint64m8_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_u16mf2))) +void vsuxei64(uint16_t * op0, vuint64m2_t op1, vuint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_u16mf2_m))) +void vsuxei64(vbool32_t op0, uint16_t * op1, vuint64m2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_u16mf4))) +void vsuxei64(uint16_t * op0, vuint64m1_t op1, vuint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_u16mf4_m))) +void vsuxei64(vbool64_t op0, uint16_t * op1, vuint64m1_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vx_i8m1))) +vint8m1_t vmax(vint8m1_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vx_i8m1_m))) +vint8m1_t vmax(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vx_i8m2))) +vint8m2_t vmax(vint8m2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vx_i8m2_m))) +vint8m2_t vmax(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vx_i8m4))) +vint8m4_t vmax(vint8m4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vx_i8m4_m))) +vint8m4_t vmax(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vx_i8m8))) +vint8m8_t vmax(vint8m8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vx_i8m8_m))) +vint8m8_t vmax(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vx_i8mf2))) +vint8mf2_t vmax(vint8mf2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vx_i8mf2_m))) +vint8mf2_t vmax(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vx_i8mf4))) +vint8mf4_t vmax(vint8mf4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vx_i8mf4_m))) +vint8mf4_t vmax(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vx_i8mf8))) +vint8mf8_t vmax(vint8mf8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vx_i8mf8_m))) +vint8mf8_t vmax(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vx_i16m1))) +vint16m1_t vmax(vint16m1_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vx_i16m1_m))) +vint16m1_t vmax(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vx_i16m2))) +vint16m2_t vmax(vint16m2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vx_i16m2_m))) +vint16m2_t vmax(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vx_i16m4))) +vint16m4_t vmax(vint16m4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vx_i16m4_m))) +vint16m4_t vmax(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vx_i16m8))) +vint16m8_t vmax(vint16m8_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vx_i16m8_m))) +vint16m8_t vmax(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vx_i16mf2))) +vint16mf2_t vmax(vint16mf2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vx_i16mf2_m))) +vint16mf2_t vmax(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vx_i16mf4))) +vint16mf4_t vmax(vint16mf4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vx_i16mf4_m))) +vint16mf4_t vmax(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vx_i32m1))) +vint32m1_t vmax(vint32m1_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vx_i32m1_m))) +vint32m1_t vmax(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vx_i32m2))) +vint32m2_t vmax(vint32m2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vx_i32m2_m))) +vint32m2_t vmax(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vx_i32m4))) +vint32m4_t vmax(vint32m4_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vx_i32m4_m))) +vint32m4_t vmax(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vx_i32m8))) +vint32m8_t vmax(vint32m8_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vx_i32m8_m))) +vint32m8_t vmax(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vx_i32mf2))) +vint32mf2_t vmax(vint32mf2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vx_i32mf2_m))) +vint32mf2_t vmax(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vx_i64m1))) +vint64m1_t vmax(vint64m1_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vx_i64m1_m))) +vint64m1_t vmax(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vx_i64m2))) +vint64m2_t vmax(vint64m2_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vx_i64m2_m))) +vint64m2_t vmax(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vx_i64m4))) +vint64m4_t vmax(vint64m4_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vx_i64m4_m))) +vint64m4_t vmax(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vx_i64m8))) +vint64m8_t vmax(vint64m8_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmax_vx_i64m8_m))) +vint64m8_t vmax(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_i8m1))) +vint8m1_t vmul(vint8m1_t op0, vint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_i8m1_m))) +vint8m1_t vmul(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_i8m2))) +vint8m2_t vmul(vint8m2_t op0, vint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_i8m2_m))) +vint8m2_t vmul(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_i8m4))) +vint8m4_t vmul(vint8m4_t op0, vint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_i8m4_m))) +vint8m4_t vmul(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_i8m8))) +vint8m8_t vmul(vint8m8_t op0, vint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_i8m8_m))) +vint8m8_t vmul(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, vint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_i8mf2))) +vint8mf2_t vmul(vint8mf2_t op0, vint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_i8mf2_m))) +vint8mf2_t vmul(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_i8mf4))) +vint8mf4_t vmul(vint8mf4_t op0, vint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_i8mf4_m))) +vint8mf4_t vmul(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_i8mf8))) +vint8mf8_t vmul(vint8mf8_t op0, vint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_i8mf8_m))) +vint8mf8_t vmul(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_i16m1))) +vint16m1_t vmul(vint16m1_t op0, vint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_i16m1_m))) +vint16m1_t vmul(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_i16m2))) +vint16m2_t vmul(vint16m2_t op0, vint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_i16m2_m))) +vint16m2_t vmul(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_i16m4))) +vint16m4_t vmul(vint16m4_t op0, vint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_i16m4_m))) +vint16m4_t vmul(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_i16m8))) +vint16m8_t vmul(vint16m8_t op0, vint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_i16m8_m))) +vint16m8_t vmul(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, vint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_i16mf2))) +vint16mf2_t vmul(vint16mf2_t op0, vint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_i16mf2_m))) +vint16mf2_t vmul(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_i16mf4))) +vint16mf4_t vmul(vint16mf4_t op0, vint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_i16mf4_m))) +vint16mf4_t vmul(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_i32m1))) +vint32m1_t vmul(vint32m1_t op0, vint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_i32m1_m))) +vint32m1_t vmul(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_i32m2))) +vint32m2_t vmul(vint32m2_t op0, vint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_i32m2_m))) +vint32m2_t vmul(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_i32m4))) +vint32m4_t vmul(vint32m4_t op0, vint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_i32m4_m))) +vint32m4_t vmul(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_i32m8))) +vint32m8_t vmul(vint32m8_t op0, vint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_i32m8_m))) +vint32m8_t vmul(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_i32mf2))) +vint32mf2_t vmul(vint32mf2_t op0, vint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_i32mf2_m))) +vint32mf2_t vmul(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_i64m1))) +vint64m1_t vmul(vint64m1_t op0, vint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_i64m1_m))) +vint64m1_t vmul(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_i64m2))) +vint64m2_t vmul(vint64m2_t op0, vint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_i64m2_m))) +vint64m2_t vmul(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_i64m4))) +vint64m4_t vmul(vint64m4_t op0, vint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_i64m4_m))) +vint64m4_t vmul(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_i64m8))) +vint64m8_t vmul(vint64m8_t op0, vint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_i64m8_m))) +vint64m8_t vmul(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_i8m1))) +vint8m1_t vmul(vint8m1_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_i8m1_m))) +vint8m1_t vmul(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_i8m2))) +vint8m2_t vmul(vint8m2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_i8m2_m))) +vint8m2_t vmul(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_i8m4))) +vint8m4_t vmul(vint8m4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_i8m4_m))) +vint8m4_t vmul(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_i8m8))) +vint8m8_t vmul(vint8m8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_i8m8_m))) +vint8m8_t vmul(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_i8mf2))) +vint8mf2_t vmul(vint8mf2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_i8mf2_m))) +vint8mf2_t vmul(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_i8mf4))) +vint8mf4_t vmul(vint8mf4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_i8mf4_m))) +vint8mf4_t vmul(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_i8mf8))) +vint8mf8_t vmul(vint8mf8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_i8mf8_m))) +vint8mf8_t vmul(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_i16m1))) +vint16m1_t vmul(vint16m1_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_i16m1_m))) +vint16m1_t vmul(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_i16m2))) +vint16m2_t vmul(vint16m2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_i16m2_m))) +vint16m2_t vmul(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_i16m4))) +vint16m4_t vmul(vint16m4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_i16m4_m))) +vint16m4_t vmul(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_i16m8))) +vint16m8_t vmul(vint16m8_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_i16m8_m))) +vint16m8_t vmul(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_i16mf2))) +vint16mf2_t vmul(vint16mf2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_i16mf2_m))) +vint16mf2_t vmul(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_i16mf4))) +vint16mf4_t vmul(vint16mf4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_i16mf4_m))) +vint16mf4_t vmul(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_i32m1))) +vint32m1_t vmul(vint32m1_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_i32m1_m))) +vint32m1_t vmul(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_i32m2))) +vint32m2_t vmul(vint32m2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_i32m2_m))) +vint32m2_t vmul(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_i32m4))) +vint32m4_t vmul(vint32m4_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_i32m4_m))) +vint32m4_t vmul(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_i32m8))) +vint32m8_t vmul(vint32m8_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_i32m8_m))) +vint32m8_t vmul(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_i32mf2))) +vint32mf2_t vmul(vint32mf2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_i32mf2_m))) +vint32mf2_t vmul(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_i64m1))) +vint64m1_t vmul(vint64m1_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_i64m1_m))) +vint64m1_t vmul(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_i64m2))) +vint64m2_t vmul(vint64m2_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_i64m2_m))) +vint64m2_t vmul(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_i64m4))) +vint64m4_t vmul(vint64m4_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_i64m4_m))) +vint64m4_t vmul(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_i64m8))) +vint64m8_t vmul(vint64m8_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_i64m8_m))) +vint64m8_t vmul(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_u8m1))) +vuint8m1_t vmul(vuint8m1_t op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_u8m1_m))) +vuint8m1_t vmul(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_u8m2))) +vuint8m2_t vmul(vuint8m2_t op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_u8m2_m))) +vuint8m2_t vmul(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_u8m4))) +vuint8m4_t vmul(vuint8m4_t op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_u8m4_m))) +vuint8m4_t vmul(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_u8m8))) +vuint8m8_t vmul(vuint8m8_t op0, vuint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_u8m8_m))) +vuint8m8_t vmul(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, vuint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_u8mf2))) +vuint8mf2_t vmul(vuint8mf2_t op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_u8mf2_m))) +vuint8mf2_t vmul(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_u8mf4))) +vuint8mf4_t vmul(vuint8mf4_t op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_u8mf4_m))) +vuint8mf4_t vmul(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_u8mf8))) +vuint8mf8_t vmul(vuint8mf8_t op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_u8mf8_m))) +vuint8mf8_t vmul(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_u16m1))) +vuint16m1_t vmul(vuint16m1_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_u16m1_m))) +vuint16m1_t vmul(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_u16m2))) +vuint16m2_t vmul(vuint16m2_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_u16m2_m))) +vuint16m2_t vmul(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_u16m4))) +vuint16m4_t vmul(vuint16m4_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_u16m4_m))) +vuint16m4_t vmul(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_u16m8))) +vuint16m8_t vmul(vuint16m8_t op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_u16m8_m))) +vuint16m8_t vmul(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_u16mf2))) +vuint16mf2_t vmul(vuint16mf2_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_u16mf2_m))) +vuint16mf2_t vmul(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_u16mf4))) +vuint16mf4_t vmul(vuint16mf4_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_u16mf4_m))) +vuint16mf4_t vmul(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_u32m1))) +vuint32m1_t vmul(vuint32m1_t op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_u32m1_m))) +vuint32m1_t vmul(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_u32m2))) +vuint32m2_t vmul(vuint32m2_t op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_u32m2_m))) +vuint32m2_t vmul(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_u32m4))) +vuint32m4_t vmul(vuint32m4_t op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_u32m4_m))) +vuint32m4_t vmul(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_u32m8))) +vuint32m8_t vmul(vuint32m8_t op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_u32m8_m))) +vuint32m8_t vmul(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_u32mf2))) +vuint32mf2_t vmul(vuint32mf2_t op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_u32mf2_m))) +vuint32mf2_t vmul(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_u64m1))) +vuint64m1_t vmul(vuint64m1_t op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_u64m1_m))) +vuint64m1_t vmul(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_u64m2))) +vuint64m2_t vmul(vuint64m2_t op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_u64m2_m))) +vuint64m2_t vmul(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_u64m4))) +vuint64m4_t vmul(vuint64m4_t op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_u64m4_m))) +vuint64m4_t vmul(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_u64m8))) +vuint64m8_t vmul(vuint64m8_t op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vv_u64m8_m))) +vuint64m8_t vmul(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_u8m1))) +vuint8m1_t vmul(vuint8m1_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_u8m1_m))) +vuint8m1_t vmul(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_u8m2))) +vuint8m2_t vmul(vuint8m2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_u8m2_m))) +vuint8m2_t vmul(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_u8m4))) +vuint8m4_t vmul(vuint8m4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_u8m4_m))) +vuint8m4_t vmul(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_u8m8))) +vuint8m8_t vmul(vuint8m8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_u8m8_m))) +vuint8m8_t vmul(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_u8mf2))) +vuint8mf2_t vmul(vuint8mf2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_u8mf2_m))) +vuint8mf2_t vmul(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_u8mf4))) +vuint8mf4_t vmul(vuint8mf4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_u8mf4_m))) +vuint8mf4_t vmul(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_u8mf8))) +vuint8mf8_t vmul(vuint8mf8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_u8mf8_m))) +vuint8mf8_t vmul(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_u16m1))) +vuint16m1_t vmul(vuint16m1_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_u16m1_m))) +vuint16m1_t vmul(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_u16m2))) +vuint16m2_t vmul(vuint16m2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_u16m2_m))) +vuint16m2_t vmul(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_u16m4))) +vuint16m4_t vmul(vuint16m4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_u16m4_m))) +vuint16m4_t vmul(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_u16m8))) +vuint16m8_t vmul(vuint16m8_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_u16m8_m))) +vuint16m8_t vmul(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_u16mf2))) +vuint16mf2_t vmul(vuint16mf2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_u16mf2_m))) +vuint16mf2_t vmul(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_u16mf4))) +vuint16mf4_t vmul(vuint16mf4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_u16mf4_m))) +vuint16mf4_t vmul(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_u32m1))) +vuint32m1_t vmul(vuint32m1_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_u32m1_m))) +vuint32m1_t vmul(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_u32m2))) +vuint32m2_t vmul(vuint32m2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_u32m2_m))) +vuint32m2_t vmul(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_u32m4))) +vuint32m4_t vmul(vuint32m4_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_u32m4_m))) +vuint32m4_t vmul(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_u32m8))) +vuint32m8_t vmul(vuint32m8_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_u32m8_m))) +vuint32m8_t vmul(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_u32mf2))) +vuint32mf2_t vmul(vuint32mf2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_u32mf2_m))) +vuint32mf2_t vmul(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_u64m1))) +vuint64m1_t vmul(vuint64m1_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_u64m1_m))) +vuint64m1_t vmul(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_u64m2))) +vuint64m2_t vmul(vuint64m2_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_u64m2_m))) +vuint64m2_t vmul(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_u64m4))) +vuint64m4_t vmul(vuint64m4_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_u64m4_m))) +vuint64m4_t vmul(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_u64m8))) +vuint64m8_t vmul(vuint64m8_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmul_vx_u64m8_m))) +vuint64m8_t vmul(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vv_i8m1))) +vint8m1_t vmulh(vint8m1_t op0, vint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vv_i8m1_m))) +vint8m1_t vmulh(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vv_i8m2))) +vint8m2_t vmulh(vint8m2_t op0, vint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vv_i8m2_m))) +vint8m2_t vmulh(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vv_i8m4))) +vint8m4_t vmulh(vint8m4_t op0, vint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vv_i8m4_m))) +vint8m4_t vmulh(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vv_i8m8))) +vint8m8_t vmulh(vint8m8_t op0, vint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vv_i8m8_m))) +vint8m8_t vmulh(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, vint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vv_i8mf2))) +vint8mf2_t vmulh(vint8mf2_t op0, vint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vv_i8mf2_m))) +vint8mf2_t vmulh(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vv_i8mf4))) +vint8mf4_t vmulh(vint8mf4_t op0, vint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vv_i8mf4_m))) +vint8mf4_t vmulh(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vv_i8mf8))) +vint8mf8_t vmulh(vint8mf8_t op0, vint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vv_i8mf8_m))) +vint8mf8_t vmulh(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vv_i16m1))) +vint16m1_t vmulh(vint16m1_t op0, vint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vv_i16m1_m))) +vint16m1_t vmulh(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vv_i16m2))) +vint16m2_t vmulh(vint16m2_t op0, vint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vv_i16m2_m))) +vint16m2_t vmulh(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vv_i16m4))) +vint16m4_t vmulh(vint16m4_t op0, vint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vv_i16m4_m))) +vint16m4_t vmulh(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vv_i16m8))) +vint16m8_t vmulh(vint16m8_t op0, vint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vv_i16m8_m))) +vint16m8_t vmulh(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, vint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vv_i16mf2))) +vint16mf2_t vmulh(vint16mf2_t op0, vint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vv_i16mf2_m))) +vint16mf2_t vmulh(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vv_i16mf4))) +vint16mf4_t vmulh(vint16mf4_t op0, vint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vv_i16mf4_m))) +vint16mf4_t vmulh(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vv_i32m1))) +vint32m1_t vmulh(vint32m1_t op0, vint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vv_i32m1_m))) +vint32m1_t vmulh(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vv_i32m2))) +vint32m2_t vmulh(vint32m2_t op0, vint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vv_i32m2_m))) +vint32m2_t vmulh(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vv_i32m4))) +vint32m4_t vmulh(vint32m4_t op0, vint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vv_i32m4_m))) +vint32m4_t vmulh(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vv_i32m8))) +vint32m8_t vmulh(vint32m8_t op0, vint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vv_i32m8_m))) +vint32m8_t vmulh(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vv_i32mf2))) +vint32mf2_t vmulh(vint32mf2_t op0, vint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vv_i32mf2_m))) +vint32mf2_t vmulh(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vv_i64m1))) +vint64m1_t vmulh(vint64m1_t op0, vint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vv_i64m1_m))) +vint64m1_t vmulh(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vv_i64m2))) +vint64m2_t vmulh(vint64m2_t op0, vint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vv_i64m2_m))) +vint64m2_t vmulh(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vv_i64m4))) +vint64m4_t vmulh(vint64m4_t op0, vint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vv_i64m4_m))) +vint64m4_t vmulh(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vv_i64m8))) +vint64m8_t vmulh(vint64m8_t op0, vint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vv_i64m8_m))) +vint64m8_t vmulh(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vx_i8m1))) +vint8m1_t vmulh(vint8m1_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vx_i8m1_m))) +vint8m1_t vmulh(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vx_i8m2))) +vint8m2_t vmulh(vint8m2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vx_i8m2_m))) +vint8m2_t vmulh(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vx_i8m4))) +vint8m4_t vmulh(vint8m4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vx_i8m4_m))) +vint8m4_t vmulh(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vx_i8m8))) +vint8m8_t vmulh(vint8m8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vx_i8m8_m))) +vint8m8_t vmulh(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vx_i8mf2))) +vint8mf2_t vmulh(vint8mf2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vx_i8mf2_m))) +vint8mf2_t vmulh(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vx_i8mf4))) +vint8mf4_t vmulh(vint8mf4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vx_i8mf4_m))) +vint8mf4_t vmulh(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vx_i8mf8))) +vint8mf8_t vmulh(vint8mf8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vx_i8mf8_m))) +vint8mf8_t vmulh(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vx_i16m1))) +vint16m1_t vmulh(vint16m1_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vx_i16m1_m))) +vint16m1_t vmulh(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vx_i16m2))) +vint16m2_t vmulh(vint16m2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vx_i16m2_m))) +vint16m2_t vmulh(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vx_i16m4))) +vint16m4_t vmulh(vint16m4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vx_i16m4_m))) +vint16m4_t vmulh(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vx_i16m8))) +vint16m8_t vmulh(vint16m8_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vx_i16m8_m))) +vint16m8_t vmulh(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vx_i16mf2))) +vint16mf2_t vmulh(vint16mf2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vx_i16mf2_m))) +vint16mf2_t vmulh(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vx_i16mf4))) +vint16mf4_t vmulh(vint16mf4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vx_i16mf4_m))) +vint16mf4_t vmulh(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vx_i32m1))) +vint32m1_t vmulh(vint32m1_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vx_i32m1_m))) +vint32m1_t vmulh(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vx_i32m2))) +vint32m2_t vmulh(vint32m2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vx_i32m2_m))) +vint32m2_t vmulh(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vx_i32m4))) +vint32m4_t vmulh(vint32m4_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vx_i32m4_m))) +vint32m4_t vmulh(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vx_i32m8))) +vint32m8_t vmulh(vint32m8_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vx_i32m8_m))) +vint32m8_t vmulh(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vx_i32mf2))) +vint32mf2_t vmulh(vint32mf2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vx_i32mf2_m))) +vint32mf2_t vmulh(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vx_i64m1))) +vint64m1_t vmulh(vint64m1_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vx_i64m1_m))) +vint64m1_t vmulh(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vx_i64m2))) +vint64m2_t vmulh(vint64m2_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vx_i64m2_m))) +vint64m2_t vmulh(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vx_i64m4))) +vint64m4_t vmulh(vint64m4_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vx_i64m4_m))) +vint64m4_t vmulh(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vx_i64m8))) +vint64m8_t vmulh(vint64m8_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulh_vx_i64m8_m))) +vint64m8_t vmulh(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vv_u8m1))) +vuint8m1_t vmulhu(vuint8m1_t op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vv_u8m1_m))) +vuint8m1_t vmulhu(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vv_u8m2))) +vuint8m2_t vmulhu(vuint8m2_t op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vv_u8m2_m))) +vuint8m2_t vmulhu(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vv_u8m4))) +vuint8m4_t vmulhu(vuint8m4_t op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vv_u8m4_m))) +vuint8m4_t vmulhu(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vv_u8m8))) +vuint8m8_t vmulhu(vuint8m8_t op0, vuint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vv_u8m8_m))) +vuint8m8_t vmulhu(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, vuint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vv_u8mf2))) +vuint8mf2_t vmulhu(vuint8mf2_t op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vv_u8mf2_m))) +vuint8mf2_t vmulhu(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vv_u8mf4))) +vuint8mf4_t vmulhu(vuint8mf4_t op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vv_u8mf4_m))) +vuint8mf4_t vmulhu(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vv_u8mf8))) +vuint8mf8_t vmulhu(vuint8mf8_t op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vv_u8mf8_m))) +vuint8mf8_t vmulhu(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vv_u16m1))) +vuint16m1_t vmulhu(vuint16m1_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vv_u16m1_m))) +vuint16m1_t vmulhu(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vv_u16m2))) +vuint16m2_t vmulhu(vuint16m2_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vv_u16m2_m))) +vuint16m2_t vmulhu(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vv_u16m4))) +vuint16m4_t vmulhu(vuint16m4_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vv_u16m4_m))) +vuint16m4_t vmulhu(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vv_u16m8))) +vuint16m8_t vmulhu(vuint16m8_t op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vv_u16m8_m))) +vuint16m8_t vmulhu(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vv_u16mf2))) +vuint16mf2_t vmulhu(vuint16mf2_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vv_u16mf2_m))) +vuint16mf2_t vmulhu(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vv_u16mf4))) +vuint16mf4_t vmulhu(vuint16mf4_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vv_u16mf4_m))) +vuint16mf4_t vmulhu(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vv_u32m1))) +vuint32m1_t vmulhu(vuint32m1_t op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vv_u32m1_m))) +vuint32m1_t vmulhu(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vv_u32m2))) +vuint32m2_t vmulhu(vuint32m2_t op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vv_u32m2_m))) +vuint32m2_t vmulhu(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vv_u32m4))) +vuint32m4_t vmulhu(vuint32m4_t op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vv_u32m4_m))) +vuint32m4_t vmulhu(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vv_u32m8))) +vuint32m8_t vmulhu(vuint32m8_t op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vv_u32m8_m))) +vuint32m8_t vmulhu(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vv_u32mf2))) +vuint32mf2_t vmulhu(vuint32mf2_t op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vv_u32mf2_m))) +vuint32mf2_t vmulhu(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vv_u64m1))) +vuint64m1_t vmulhu(vuint64m1_t op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vv_u64m1_m))) +vuint64m1_t vmulhu(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vv_u64m2))) +vuint64m2_t vmulhu(vuint64m2_t op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vv_u64m2_m))) +vuint64m2_t vmulhu(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vv_u64m4))) +vuint64m4_t vmulhu(vuint64m4_t op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vv_u64m4_m))) +vuint64m4_t vmulhu(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vv_u64m8))) +vuint64m8_t vmulhu(vuint64m8_t op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vv_u64m8_m))) +vuint64m8_t vmulhu(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vx_u8m1))) +vuint8m1_t vmulhu(vuint8m1_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vx_u8m1_m))) +vuint8m1_t vmulhu(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vx_u8m2))) +vuint8m2_t vmulhu(vuint8m2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vx_u8m2_m))) +vuint8m2_t vmulhu(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vx_u8m4))) +vuint8m4_t vmulhu(vuint8m4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vx_u8m4_m))) +vuint8m4_t vmulhu(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vx_u8m8))) +vuint8m8_t vmulhu(vuint8m8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vx_u8m8_m))) +vuint8m8_t vmulhu(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vx_u8mf2))) +vuint8mf2_t vmulhu(vuint8mf2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vx_u8mf2_m))) +vuint8mf2_t vmulhu(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vx_u8mf4))) +vuint8mf4_t vmulhu(vuint8mf4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vx_u8mf4_m))) +vuint8mf4_t vmulhu(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vx_u8mf8))) +vuint8mf8_t vmulhu(vuint8mf8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vx_u8mf8_m))) +vuint8mf8_t vmulhu(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vx_u16m1))) +vuint16m1_t vmulhu(vuint16m1_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vx_u16m1_m))) +vuint16m1_t vmulhu(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vx_u16m2))) +vuint16m2_t vmulhu(vuint16m2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vx_u16m2_m))) +vuint16m2_t vmulhu(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vx_u16m4))) +vuint16m4_t vmulhu(vuint16m4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vx_u16m4_m))) +vuint16m4_t vmulhu(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vx_u16m8))) +vuint16m8_t vmulhu(vuint16m8_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vx_u16m8_m))) +vuint16m8_t vmulhu(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vx_u16mf2))) +vuint16mf2_t vmulhu(vuint16mf2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vx_u16mf2_m))) +vuint16mf2_t vmulhu(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vx_u16mf4))) +vuint16mf4_t vmulhu(vuint16mf4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vx_u16mf4_m))) +vuint16mf4_t vmulhu(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vx_u32m1))) +vuint32m1_t vmulhu(vuint32m1_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vx_u32m1_m))) +vuint32m1_t vmulhu(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vx_u32m2))) +vuint32m2_t vmulhu(vuint32m2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vx_u32m2_m))) +vuint32m2_t vmulhu(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vx_u32m4))) +vuint32m4_t vmulhu(vuint32m4_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vx_u32m4_m))) +vuint32m4_t vmulhu(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vx_u32m8))) +vuint32m8_t vmulhu(vuint32m8_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vx_u32m8_m))) +vuint32m8_t vmulhu(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vx_u32mf2))) +vuint32mf2_t vmulhu(vuint32mf2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vx_u32mf2_m))) +vuint32mf2_t vmulhu(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vx_u64m1))) +vuint64m1_t vmulhu(vuint64m1_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vx_u64m1_m))) +vuint64m1_t vmulhu(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vx_u64m2))) +vuint64m2_t vmulhu(vuint64m2_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vx_u64m2_m))) +vuint64m2_t vmulhu(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vx_u64m4))) +vuint64m4_t vmulhu(vuint64m4_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vx_u64m4_m))) +vuint64m4_t vmulhu(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vx_u64m8))) +vuint64m8_t vmulhu(vuint64m8_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhu_vx_u64m8_m))) +vuint64m8_t vmulhu(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vv_i8m1))) +vint8m1_t vmulhsu(vint8m1_t op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vv_i8m1_m))) +vint8m1_t vmulhsu(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vv_i8m2))) +vint8m2_t vmulhsu(vint8m2_t op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vv_i8m2_m))) +vint8m2_t vmulhsu(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vv_i8m4))) +vint8m4_t vmulhsu(vint8m4_t op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vv_i8m4_m))) +vint8m4_t vmulhsu(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vv_i8m8))) +vint8m8_t vmulhsu(vint8m8_t op0, vuint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vv_i8m8_m))) +vint8m8_t vmulhsu(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, vuint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vv_i8mf2))) +vint8mf2_t vmulhsu(vint8mf2_t op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vv_i8mf2_m))) +vint8mf2_t vmulhsu(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vv_i8mf4))) +vint8mf4_t vmulhsu(vint8mf4_t op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vv_i8mf4_m))) +vint8mf4_t vmulhsu(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vv_i8mf8))) +vint8mf8_t vmulhsu(vint8mf8_t op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vv_i8mf8_m))) +vint8mf8_t vmulhsu(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vv_i16m1))) +vint16m1_t vmulhsu(vint16m1_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vv_i16m1_m))) +vint16m1_t vmulhsu(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vv_i16m2))) +vint16m2_t vmulhsu(vint16m2_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vv_i16m2_m))) +vint16m2_t vmulhsu(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vv_i16m4))) +vint16m4_t vmulhsu(vint16m4_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vv_i16m4_m))) +vint16m4_t vmulhsu(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vv_i16m8))) +vint16m8_t vmulhsu(vint16m8_t op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vv_i16m8_m))) +vint16m8_t vmulhsu(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vv_i16mf2))) +vint16mf2_t vmulhsu(vint16mf2_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vv_i16mf2_m))) +vint16mf2_t vmulhsu(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vv_i16mf4))) +vint16mf4_t vmulhsu(vint16mf4_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vv_i16mf4_m))) +vint16mf4_t vmulhsu(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vv_i32m1))) +vint32m1_t vmulhsu(vint32m1_t op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vv_i32m1_m))) +vint32m1_t vmulhsu(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vv_i32m2))) +vint32m2_t vmulhsu(vint32m2_t op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vv_i32m2_m))) +vint32m2_t vmulhsu(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vv_i32m4))) +vint32m4_t vmulhsu(vint32m4_t op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vv_i32m4_m))) +vint32m4_t vmulhsu(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vv_i32m8))) +vint32m8_t vmulhsu(vint32m8_t op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vv_i32m8_m))) +vint32m8_t vmulhsu(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vv_i32mf2))) +vint32mf2_t vmulhsu(vint32mf2_t op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vv_i32mf2_m))) +vint32mf2_t vmulhsu(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vv_i64m1))) +vint64m1_t vmulhsu(vint64m1_t op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vv_i64m1_m))) +vint64m1_t vmulhsu(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vv_i64m2))) +vint64m2_t vmulhsu(vint64m2_t op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vv_i64m2_m))) +vint64m2_t vmulhsu(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vv_i64m4))) +vint64m4_t vmulhsu(vint64m4_t op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vv_i64m4_m))) +vint64m4_t vmulhsu(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vv_i64m8))) +vint64m8_t vmulhsu(vint64m8_t op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vv_i64m8_m))) +vint64m8_t vmulhsu(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_i32m1))) +void vsuxei8(int32_t * op0, vuint8mf4_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_i32m1_m))) +void vsuxei8(vbool32_t op0, int32_t * op1, vuint8mf4_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_i32m2))) +void vsuxei8(int32_t * op0, vuint8mf2_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_i32m2_m))) +void vsuxei8(vbool16_t op0, int32_t * op1, vuint8mf2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_i32m4))) +void vsuxei8(int32_t * op0, vuint8m1_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_i32m4_m))) +void vsuxei8(vbool8_t op0, int32_t * op1, vuint8m1_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_i32m8))) +void vsuxei8(int32_t * op0, vuint8m2_t op1, vint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_i32m8_m))) +void vsuxei8(vbool4_t op0, int32_t * op1, vuint8m2_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_i32mf2))) +void vsuxei8(int32_t * op0, vuint8mf8_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_i32mf2_m))) +void vsuxei8(vbool64_t op0, int32_t * op1, vuint8mf8_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vx_i8m1))) +vint8m1_t vmulhsu(vint8m1_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vx_i8m1_m))) +vint8m1_t vmulhsu(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vx_i8m2))) +vint8m2_t vmulhsu(vint8m2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vx_i8m2_m))) +vint8m2_t vmulhsu(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vx_i8m4))) +vint8m4_t vmulhsu(vint8m4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vx_i8m4_m))) +vint8m4_t vmulhsu(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vx_i8m8))) +vint8m8_t vmulhsu(vint8m8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vx_i8m8_m))) +vint8m8_t vmulhsu(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vx_i8mf2))) +vint8mf2_t vmulhsu(vint8mf2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vx_i8mf2_m))) +vint8mf2_t vmulhsu(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vx_i8mf4))) +vint8mf4_t vmulhsu(vint8mf4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vx_i8mf4_m))) +vint8mf4_t vmulhsu(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vx_i8mf8))) +vint8mf8_t vmulhsu(vint8mf8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vx_i8mf8_m))) +vint8mf8_t vmulhsu(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vx_i16m1))) +vint16m1_t vmulhsu(vint16m1_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vx_i16m1_m))) +vint16m1_t vmulhsu(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vx_i16m2))) +vint16m2_t vmulhsu(vint16m2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vx_i16m2_m))) +vint16m2_t vmulhsu(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vx_i16m4))) +vint16m4_t vmulhsu(vint16m4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vx_i16m4_m))) +vint16m4_t vmulhsu(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vx_i16m8))) +vint16m8_t vmulhsu(vint16m8_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vx_i16m8_m))) +vint16m8_t vmulhsu(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vx_i16mf2))) +vint16mf2_t vmulhsu(vint16mf2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vx_i16mf2_m))) +vint16mf2_t vmulhsu(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vx_i16mf4))) +vint16mf4_t vmulhsu(vint16mf4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vx_i16mf4_m))) +vint16mf4_t vmulhsu(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vx_i32m1))) +vint32m1_t vmulhsu(vint32m1_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vx_i32m1_m))) +vint32m1_t vmulhsu(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vx_i32m2))) +vint32m2_t vmulhsu(vint32m2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vx_i32m2_m))) +vint32m2_t vmulhsu(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vx_i32m4))) +vint32m4_t vmulhsu(vint32m4_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vx_i32m4_m))) +vint32m4_t vmulhsu(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vx_i32m8))) +vint32m8_t vmulhsu(vint32m8_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vx_i32m8_m))) +vint32m8_t vmulhsu(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vx_i32mf2))) +vint32mf2_t vmulhsu(vint32mf2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vx_i32mf2_m))) +vint32mf2_t vmulhsu(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vx_i64m1))) +vint64m1_t vmulhsu(vint64m1_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vx_i64m1_m))) +vint64m1_t vmulhsu(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vx_i64m2))) +vint64m2_t vmulhsu(vint64m2_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vx_i64m2_m))) +vint64m2_t vmulhsu(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vx_i64m4))) +vint64m4_t vmulhsu(vint64m4_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vx_i64m4_m))) +vint64m4_t vmulhsu(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vx_i64m8))) +vint64m8_t vmulhsu(vint64m8_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmulhsu_vx_i64m8_m))) +vint64m8_t vmulhsu(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vv_u8m1))) +vuint8m1_t vdivu(vuint8m1_t op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vv_u8m1_m))) +vuint8m1_t vdivu(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vv_u8m2))) +vuint8m2_t vdivu(vuint8m2_t op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vv_u8m2_m))) +vuint8m2_t vdivu(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vv_u8m4))) +vuint8m4_t vdivu(vuint8m4_t op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vv_u8m4_m))) +vuint8m4_t vdivu(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vv_u8m8))) +vuint8m8_t vdivu(vuint8m8_t op0, vuint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vv_u8m8_m))) +vuint8m8_t vdivu(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, vuint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vv_u8mf2))) +vuint8mf2_t vdivu(vuint8mf2_t op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vv_u8mf2_m))) +vuint8mf2_t vdivu(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vv_u8mf4))) +vuint8mf4_t vdivu(vuint8mf4_t op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vv_u8mf4_m))) +vuint8mf4_t vdivu(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vv_u8mf8))) +vuint8mf8_t vdivu(vuint8mf8_t op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vv_u8mf8_m))) +vuint8mf8_t vdivu(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vv_u16m1))) +vuint16m1_t vdivu(vuint16m1_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vv_u16m1_m))) +vuint16m1_t vdivu(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vv_u16m2))) +vuint16m2_t vdivu(vuint16m2_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vv_u16m2_m))) +vuint16m2_t vdivu(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vv_u16m4))) +vuint16m4_t vdivu(vuint16m4_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vv_u16m4_m))) +vuint16m4_t vdivu(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vv_u16m8))) +vuint16m8_t vdivu(vuint16m8_t op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vv_u16m8_m))) +vuint16m8_t vdivu(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vv_u16mf2))) +vuint16mf2_t vdivu(vuint16mf2_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vv_u16mf2_m))) +vuint16mf2_t vdivu(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vv_u16mf4))) +vuint16mf4_t vdivu(vuint16mf4_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vv_u16mf4_m))) +vuint16mf4_t vdivu(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vv_u32m1))) +vuint32m1_t vdivu(vuint32m1_t op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vv_u32m1_m))) +vuint32m1_t vdivu(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vv_u32m2))) +vuint32m2_t vdivu(vuint32m2_t op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vv_u32m2_m))) +vuint32m2_t vdivu(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vv_u32m4))) +vuint32m4_t vdivu(vuint32m4_t op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vv_u32m4_m))) +vuint32m4_t vdivu(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vv_u32m8))) +vuint32m8_t vdivu(vuint32m8_t op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vv_u32m8_m))) +vuint32m8_t vdivu(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vv_u32mf2))) +vuint32mf2_t vdivu(vuint32mf2_t op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vv_u32mf2_m))) +vuint32mf2_t vdivu(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vv_u64m1))) +vuint64m1_t vdivu(vuint64m1_t op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vv_u64m1_m))) +vuint64m1_t vdivu(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vv_u64m2))) +vuint64m2_t vdivu(vuint64m2_t op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vv_u64m2_m))) +vuint64m2_t vdivu(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vv_u64m4))) +vuint64m4_t vdivu(vuint64m4_t op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vv_u64m4_m))) +vuint64m4_t vdivu(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vv_u64m8))) +vuint64m8_t vdivu(vuint64m8_t op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vv_u64m8_m))) +vuint64m8_t vdivu(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vx_u8m1))) +vuint8m1_t vdivu(vuint8m1_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vx_u8m1_m))) +vuint8m1_t vdivu(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vx_u8m2))) +vuint8m2_t vdivu(vuint8m2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vx_u8m2_m))) +vuint8m2_t vdivu(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vx_u8m4))) +vuint8m4_t vdivu(vuint8m4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vx_u8m4_m))) +vuint8m4_t vdivu(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vx_u8m8))) +vuint8m8_t vdivu(vuint8m8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vx_u8m8_m))) +vuint8m8_t vdivu(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vx_u8mf2))) +vuint8mf2_t vdivu(vuint8mf2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vx_u8mf2_m))) +vuint8mf2_t vdivu(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vx_u8mf4))) +vuint8mf4_t vdivu(vuint8mf4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vx_u8mf4_m))) +vuint8mf4_t vdivu(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vx_u8mf8))) +vuint8mf8_t vdivu(vuint8mf8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vx_u8mf8_m))) +vuint8mf8_t vdivu(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vx_u16m1))) +vuint16m1_t vdivu(vuint16m1_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vx_u16m1_m))) +vuint16m1_t vdivu(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vx_u16m2))) +vuint16m2_t vdivu(vuint16m2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vx_u16m2_m))) +vuint16m2_t vdivu(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vx_u16m4))) +vuint16m4_t vdivu(vuint16m4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vx_u16m4_m))) +vuint16m4_t vdivu(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vx_u16m8))) +vuint16m8_t vdivu(vuint16m8_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vx_u16m8_m))) +vuint16m8_t vdivu(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vx_u16mf2))) +vuint16mf2_t vdivu(vuint16mf2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vx_u16mf2_m))) +vuint16mf2_t vdivu(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vx_u16mf4))) +vuint16mf4_t vdivu(vuint16mf4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vx_u16mf4_m))) +vuint16mf4_t vdivu(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vx_u32m1))) +vuint32m1_t vdivu(vuint32m1_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vx_u32m1_m))) +vuint32m1_t vdivu(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vx_u32m2))) +vuint32m2_t vdivu(vuint32m2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vx_u32m2_m))) +vuint32m2_t vdivu(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vx_u32m4))) +vuint32m4_t vdivu(vuint32m4_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vx_u32m4_m))) +vuint32m4_t vdivu(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vx_u32m8))) +vuint32m8_t vdivu(vuint32m8_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vx_u32m8_m))) +vuint32m8_t vdivu(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vx_u32mf2))) +vuint32mf2_t vdivu(vuint32mf2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vx_u32mf2_m))) +vuint32mf2_t vdivu(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vx_u64m1))) +vuint64m1_t vdivu(vuint64m1_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vx_u64m1_m))) +vuint64m1_t vdivu(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vx_u64m2))) +vuint64m2_t vdivu(vuint64m2_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vx_u64m2_m))) +vuint64m2_t vdivu(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vx_u64m4))) +vuint64m4_t vdivu(vuint64m4_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vx_u64m4_m))) +vuint64m4_t vdivu(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vx_u64m8))) +vuint64m8_t vdivu(vuint64m8_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdivu_vx_u64m8_m))) +vuint64m8_t vdivu(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vv_i8m1))) +vint8m1_t vdiv(vint8m1_t op0, vint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vv_i8m1_m))) +vint8m1_t vdiv(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vv_i8m2))) +vint8m2_t vdiv(vint8m2_t op0, vint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vv_i8m2_m))) +vint8m2_t vdiv(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vv_i8m4))) +vint8m4_t vdiv(vint8m4_t op0, vint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vv_i8m4_m))) +vint8m4_t vdiv(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vv_i8m8))) +vint8m8_t vdiv(vint8m8_t op0, vint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vv_i8m8_m))) +vint8m8_t vdiv(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, vint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vv_i8mf2))) +vint8mf2_t vdiv(vint8mf2_t op0, vint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vv_i8mf2_m))) +vint8mf2_t vdiv(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vv_i8mf4))) +vint8mf4_t vdiv(vint8mf4_t op0, vint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vv_i8mf4_m))) +vint8mf4_t vdiv(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vv_i8mf8))) +vint8mf8_t vdiv(vint8mf8_t op0, vint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vv_i8mf8_m))) +vint8mf8_t vdiv(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vv_i16m1))) +vint16m1_t vdiv(vint16m1_t op0, vint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vv_i16m1_m))) +vint16m1_t vdiv(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vv_i16m2))) +vint16m2_t vdiv(vint16m2_t op0, vint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vv_i16m2_m))) +vint16m2_t vdiv(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vv_i16m4))) +vint16m4_t vdiv(vint16m4_t op0, vint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vv_i16m4_m))) +vint16m4_t vdiv(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vv_i16m8))) +vint16m8_t vdiv(vint16m8_t op0, vint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vv_i16m8_m))) +vint16m8_t vdiv(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, vint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vv_i16mf2))) +vint16mf2_t vdiv(vint16mf2_t op0, vint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vv_i16mf2_m))) +vint16mf2_t vdiv(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vv_i16mf4))) +vint16mf4_t vdiv(vint16mf4_t op0, vint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vv_i16mf4_m))) +vint16mf4_t vdiv(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vv_i32m1))) +vint32m1_t vdiv(vint32m1_t op0, vint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vv_i32m1_m))) +vint32m1_t vdiv(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vv_i32m2))) +vint32m2_t vdiv(vint32m2_t op0, vint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vv_i32m2_m))) +vint32m2_t vdiv(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vv_i32m4))) +vint32m4_t vdiv(vint32m4_t op0, vint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vv_i32m4_m))) +vint32m4_t vdiv(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vv_i32m8))) +vint32m8_t vdiv(vint32m8_t op0, vint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vv_i32m8_m))) +vint32m8_t vdiv(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vv_i32mf2))) +vint32mf2_t vdiv(vint32mf2_t op0, vint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vv_i32mf2_m))) +vint32mf2_t vdiv(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vv_i64m1))) +vint64m1_t vdiv(vint64m1_t op0, vint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vv_i64m1_m))) +vint64m1_t vdiv(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vv_i64m2))) +vint64m2_t vdiv(vint64m2_t op0, vint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vv_i64m2_m))) +vint64m2_t vdiv(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vv_i64m4))) +vint64m4_t vdiv(vint64m4_t op0, vint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vv_i64m4_m))) +vint64m4_t vdiv(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vv_i64m8))) +vint64m8_t vdiv(vint64m8_t op0, vint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vv_i64m8_m))) +vint64m8_t vdiv(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vx_i8m1))) +vint8m1_t vdiv(vint8m1_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vx_i8m1_m))) +vint8m1_t vdiv(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vx_i8m2))) +vint8m2_t vdiv(vint8m2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vx_i8m2_m))) +vint8m2_t vdiv(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vx_i8m4))) +vint8m4_t vdiv(vint8m4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vx_i8m4_m))) +vint8m4_t vdiv(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vx_i8m8))) +vint8m8_t vdiv(vint8m8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vx_i8m8_m))) +vint8m8_t vdiv(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vx_i8mf2))) +vint8mf2_t vdiv(vint8mf2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vx_i8mf2_m))) +vint8mf2_t vdiv(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vx_i8mf4))) +vint8mf4_t vdiv(vint8mf4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vx_i8mf4_m))) +vint8mf4_t vdiv(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vx_i8mf8))) +vint8mf8_t vdiv(vint8mf8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vx_i8mf8_m))) +vint8mf8_t vdiv(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vx_i16m1))) +vint16m1_t vdiv(vint16m1_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vx_i16m1_m))) +vint16m1_t vdiv(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vx_i16m2))) +vint16m2_t vdiv(vint16m2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vx_i16m2_m))) +vint16m2_t vdiv(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vx_i16m4))) +vint16m4_t vdiv(vint16m4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vx_i16m4_m))) +vint16m4_t vdiv(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vx_i16m8))) +vint16m8_t vdiv(vint16m8_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vx_i16m8_m))) +vint16m8_t vdiv(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vx_i16mf2))) +vint16mf2_t vdiv(vint16mf2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vx_i16mf2_m))) +vint16mf2_t vdiv(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vx_i16mf4))) +vint16mf4_t vdiv(vint16mf4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vx_i16mf4_m))) +vint16mf4_t vdiv(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vx_i32m1))) +vint32m1_t vdiv(vint32m1_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vx_i32m1_m))) +vint32m1_t vdiv(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vx_i32m2))) +vint32m2_t vdiv(vint32m2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vx_i32m2_m))) +vint32m2_t vdiv(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vx_i32m4))) +vint32m4_t vdiv(vint32m4_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vx_i32m4_m))) +vint32m4_t vdiv(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vx_i32m8))) +vint32m8_t vdiv(vint32m8_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vx_i32m8_m))) +vint32m8_t vdiv(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vx_i32mf2))) +vint32mf2_t vdiv(vint32mf2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vx_i32mf2_m))) +vint32mf2_t vdiv(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vx_i64m1))) +vint64m1_t vdiv(vint64m1_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vx_i64m1_m))) +vint64m1_t vdiv(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vx_i64m2))) +vint64m2_t vdiv(vint64m2_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vx_i64m2_m))) +vint64m2_t vdiv(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vx_i64m4))) +vint64m4_t vdiv(vint64m4_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vx_i64m4_m))) +vint64m4_t vdiv(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vx_i64m8))) +vint64m8_t vdiv(vint64m8_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vdiv_vx_i64m8_m))) +vint64m8_t vdiv(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vv_u8m1))) +vuint8m1_t vremu(vuint8m1_t op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vv_u8m1_m))) +vuint8m1_t vremu(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vv_u8m2))) +vuint8m2_t vremu(vuint8m2_t op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vv_u8m2_m))) +vuint8m2_t vremu(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vv_u8m4))) +vuint8m4_t vremu(vuint8m4_t op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vv_u8m4_m))) +vuint8m4_t vremu(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vv_u8m8))) +vuint8m8_t vremu(vuint8m8_t op0, vuint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vv_u8m8_m))) +vuint8m8_t vremu(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, vuint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vv_u8mf2))) +vuint8mf2_t vremu(vuint8mf2_t op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vv_u8mf2_m))) +vuint8mf2_t vremu(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vv_u8mf4))) +vuint8mf4_t vremu(vuint8mf4_t op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vv_u8mf4_m))) +vuint8mf4_t vremu(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vv_u8mf8))) +vuint8mf8_t vremu(vuint8mf8_t op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vv_u8mf8_m))) +vuint8mf8_t vremu(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vv_u16m1))) +vuint16m1_t vremu(vuint16m1_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vv_u16m1_m))) +vuint16m1_t vremu(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vv_u16m2))) +vuint16m2_t vremu(vuint16m2_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vv_u16m2_m))) +vuint16m2_t vremu(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vv_u16m4))) +vuint16m4_t vremu(vuint16m4_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vv_u16m4_m))) +vuint16m4_t vremu(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vv_u16m8))) +vuint16m8_t vremu(vuint16m8_t op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vv_u16m8_m))) +vuint16m8_t vremu(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vv_u16mf2))) +vuint16mf2_t vremu(vuint16mf2_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vv_u16mf2_m))) +vuint16mf2_t vremu(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vv_u16mf4))) +vuint16mf4_t vremu(vuint16mf4_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vv_u16mf4_m))) +vuint16mf4_t vremu(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vv_u32m1))) +vuint32m1_t vremu(vuint32m1_t op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vv_u32m1_m))) +vuint32m1_t vremu(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vv_u32m2))) +vuint32m2_t vremu(vuint32m2_t op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vv_u32m2_m))) +vuint32m2_t vremu(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vv_u32m4))) +vuint32m4_t vremu(vuint32m4_t op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vv_u32m4_m))) +vuint32m4_t vremu(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vv_u32m8))) +vuint32m8_t vremu(vuint32m8_t op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vv_u32m8_m))) +vuint32m8_t vremu(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vv_u32mf2))) +vuint32mf2_t vremu(vuint32mf2_t op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vv_u32mf2_m))) +vuint32mf2_t vremu(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vv_u64m1))) +vuint64m1_t vremu(vuint64m1_t op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vv_u64m1_m))) +vuint64m1_t vremu(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vv_u64m2))) +vuint64m2_t vremu(vuint64m2_t op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vv_u64m2_m))) +vuint64m2_t vremu(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vv_u64m4))) +vuint64m4_t vremu(vuint64m4_t op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vv_u64m4_m))) +vuint64m4_t vremu(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vv_u64m8))) +vuint64m8_t vremu(vuint64m8_t op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vv_u64m8_m))) +vuint64m8_t vremu(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vx_u8m1))) +vuint8m1_t vremu(vuint8m1_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vx_u8m1_m))) +vuint8m1_t vremu(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vx_u8m2))) +vuint8m2_t vremu(vuint8m2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vx_u8m2_m))) +vuint8m2_t vremu(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vx_u8m4))) +vuint8m4_t vremu(vuint8m4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vx_u8m4_m))) +vuint8m4_t vremu(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vx_u8m8))) +vuint8m8_t vremu(vuint8m8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vx_u8m8_m))) +vuint8m8_t vremu(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vx_u8mf2))) +vuint8mf2_t vremu(vuint8mf2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vx_u8mf2_m))) +vuint8mf2_t vremu(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vx_u8mf4))) +vuint8mf4_t vremu(vuint8mf4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vx_u8mf4_m))) +vuint8mf4_t vremu(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vx_u8mf8))) +vuint8mf8_t vremu(vuint8mf8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vx_u8mf8_m))) +vuint8mf8_t vremu(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vx_u16m1))) +vuint16m1_t vremu(vuint16m1_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vx_u16m1_m))) +vuint16m1_t vremu(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vx_u16m2))) +vuint16m2_t vremu(vuint16m2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vx_u16m2_m))) +vuint16m2_t vremu(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vx_u16m4))) +vuint16m4_t vremu(vuint16m4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vx_u16m4_m))) +vuint16m4_t vremu(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vx_u16m8))) +vuint16m8_t vremu(vuint16m8_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vx_u16m8_m))) +vuint16m8_t vremu(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vx_u16mf2))) +vuint16mf2_t vremu(vuint16mf2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vx_u16mf2_m))) +vuint16mf2_t vremu(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vx_u16mf4))) +vuint16mf4_t vremu(vuint16mf4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vx_u16mf4_m))) +vuint16mf4_t vremu(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vx_u32m1))) +vuint32m1_t vremu(vuint32m1_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vx_u32m1_m))) +vuint32m1_t vremu(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vx_u32m2))) +vuint32m2_t vremu(vuint32m2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vx_u32m2_m))) +vuint32m2_t vremu(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vx_u32m4))) +vuint32m4_t vremu(vuint32m4_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vx_u32m4_m))) +vuint32m4_t vremu(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vx_u32m8))) +vuint32m8_t vremu(vuint32m8_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vx_u32m8_m))) +vuint32m8_t vremu(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vx_u32mf2))) +vuint32mf2_t vremu(vuint32mf2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vx_u32mf2_m))) +vuint32mf2_t vremu(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vx_u64m1))) +vuint64m1_t vremu(vuint64m1_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vx_u64m1_m))) +vuint64m1_t vremu(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vx_u64m2))) +vuint64m2_t vremu(vuint64m2_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vx_u64m2_m))) +vuint64m2_t vremu(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vx_u64m4))) +vuint64m4_t vremu(vuint64m4_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vx_u64m4_m))) +vuint64m4_t vremu(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vx_u64m8))) +vuint64m8_t vremu(vuint64m8_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vremu_vx_u64m8_m))) +vuint64m8_t vremu(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vv_i8m1))) +vint8m1_t vrem(vint8m1_t op0, vint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vv_i8m1_m))) +vint8m1_t vrem(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vv_i8m2))) +vint8m2_t vrem(vint8m2_t op0, vint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vv_i8m2_m))) +vint8m2_t vrem(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vv_i8m4))) +vint8m4_t vrem(vint8m4_t op0, vint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vv_i8m4_m))) +vint8m4_t vrem(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vv_i8m8))) +vint8m8_t vrem(vint8m8_t op0, vint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vv_i8m8_m))) +vint8m8_t vrem(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, vint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vv_i8mf2))) +vint8mf2_t vrem(vint8mf2_t op0, vint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vv_i8mf2_m))) +vint8mf2_t vrem(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vv_i8mf4))) +vint8mf4_t vrem(vint8mf4_t op0, vint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vv_i8mf4_m))) +vint8mf4_t vrem(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vv_i8mf8))) +vint8mf8_t vrem(vint8mf8_t op0, vint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vv_i8mf8_m))) +vint8mf8_t vrem(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vv_i16m1))) +vint16m1_t vrem(vint16m1_t op0, vint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vv_i16m1_m))) +vint16m1_t vrem(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vv_i16m2))) +vint16m2_t vrem(vint16m2_t op0, vint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vv_i16m2_m))) +vint16m2_t vrem(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vv_i16m4))) +vint16m4_t vrem(vint16m4_t op0, vint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vv_i16m4_m))) +vint16m4_t vrem(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vv_i16m8))) +vint16m8_t vrem(vint16m8_t op0, vint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vv_i16m8_m))) +vint16m8_t vrem(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, vint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vv_i16mf2))) +vint16mf2_t vrem(vint16mf2_t op0, vint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vv_i16mf2_m))) +vint16mf2_t vrem(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vv_i16mf4))) +vint16mf4_t vrem(vint16mf4_t op0, vint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vv_i16mf4_m))) +vint16mf4_t vrem(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vv_i32m1))) +vint32m1_t vrem(vint32m1_t op0, vint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vv_i32m1_m))) +vint32m1_t vrem(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vv_i32m2))) +vint32m2_t vrem(vint32m2_t op0, vint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vv_i32m2_m))) +vint32m2_t vrem(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vv_i32m4))) +vint32m4_t vrem(vint32m4_t op0, vint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vv_i32m4_m))) +vint32m4_t vrem(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vv_i32m8))) +vint32m8_t vrem(vint32m8_t op0, vint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vv_i32m8_m))) +vint32m8_t vrem(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vv_i32mf2))) +vint32mf2_t vrem(vint32mf2_t op0, vint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vv_i32mf2_m))) +vint32mf2_t vrem(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vv_i64m1))) +vint64m1_t vrem(vint64m1_t op0, vint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vv_i64m1_m))) +vint64m1_t vrem(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vv_i64m2))) +vint64m2_t vrem(vint64m2_t op0, vint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vv_i64m2_m))) +vint64m2_t vrem(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vv_i64m4))) +vint64m4_t vrem(vint64m4_t op0, vint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vv_i64m4_m))) +vint64m4_t vrem(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vv_i64m8))) +vint64m8_t vrem(vint64m8_t op0, vint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vv_i64m8_m))) +vint64m8_t vrem(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vx_i8m1))) +vint8m1_t vrem(vint8m1_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vx_i8m1_m))) +vint8m1_t vrem(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vx_i8m2))) +vint8m2_t vrem(vint8m2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vx_i8m2_m))) +vint8m2_t vrem(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vx_i8m4))) +vint8m4_t vrem(vint8m4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vx_i8m4_m))) +vint8m4_t vrem(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vx_i8m8))) +vint8m8_t vrem(vint8m8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vx_i8m8_m))) +vint8m8_t vrem(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vx_i8mf2))) +vint8mf2_t vrem(vint8mf2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vx_i8mf2_m))) +vint8mf2_t vrem(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vx_i8mf4))) +vint8mf4_t vrem(vint8mf4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vx_i8mf4_m))) +vint8mf4_t vrem(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vx_i8mf8))) +vint8mf8_t vrem(vint8mf8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vx_i8mf8_m))) +vint8mf8_t vrem(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vx_i16m1))) +vint16m1_t vrem(vint16m1_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vx_i16m1_m))) +vint16m1_t vrem(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vx_i16m2))) +vint16m2_t vrem(vint16m2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vx_i16m2_m))) +vint16m2_t vrem(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vx_i16m4))) +vint16m4_t vrem(vint16m4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vx_i16m4_m))) +vint16m4_t vrem(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vx_i16m8))) +vint16m8_t vrem(vint16m8_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vx_i16m8_m))) +vint16m8_t vrem(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vx_i16mf2))) +vint16mf2_t vrem(vint16mf2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vx_i16mf2_m))) +vint16mf2_t vrem(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vx_i16mf4))) +vint16mf4_t vrem(vint16mf4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vx_i16mf4_m))) +vint16mf4_t vrem(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vx_i32m1))) +vint32m1_t vrem(vint32m1_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vx_i32m1_m))) +vint32m1_t vrem(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vx_i32m2))) +vint32m2_t vrem(vint32m2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vx_i32m2_m))) +vint32m2_t vrem(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vx_i32m4))) +vint32m4_t vrem(vint32m4_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vx_i32m4_m))) +vint32m4_t vrem(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vx_i32m8))) +vint32m8_t vrem(vint32m8_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vx_i32m8_m))) +vint32m8_t vrem(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vx_i32mf2))) +vint32mf2_t vrem(vint32mf2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vx_i32mf2_m))) +vint32mf2_t vrem(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vx_i64m1))) +vint64m1_t vrem(vint64m1_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vx_i64m1_m))) +vint64m1_t vrem(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vx_i64m2))) +vint64m2_t vrem(vint64m2_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vx_i64m2_m))) +vint64m2_t vrem(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vx_i64m4))) +vint64m4_t vrem(vint64m4_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vx_i64m4_m))) +vint64m4_t vrem(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vx_i64m8))) +vint64m8_t vrem(vint64m8_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrem_vx_i64m8_m))) +vint64m8_t vrem(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vv_i16mf4))) +vint16mf4_t vwmul(vint8mf8_t op0, vint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vv_i16mf4_m))) +vint16mf4_t vwmul(vbool64_t op0, vint16mf4_t op1, vint8mf8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vv_i16mf2))) +vint16mf2_t vwmul(vint8mf4_t op0, vint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vv_i16mf2_m))) +vint16mf2_t vwmul(vbool32_t op0, vint16mf2_t op1, vint8mf4_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vv_i16m1))) +vint16m1_t vwmul(vint8mf2_t op0, vint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vv_i16m1_m))) +vint16m1_t vwmul(vbool16_t op0, vint16m1_t op1, vint8mf2_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vv_i16m2))) +vint16m2_t vwmul(vint8m1_t op0, vint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vv_i16m2_m))) +vint16m2_t vwmul(vbool8_t op0, vint16m2_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vv_i16m4))) +vint16m4_t vwmul(vint8m2_t op0, vint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vv_i16m4_m))) +vint16m4_t vwmul(vbool4_t op0, vint16m4_t op1, vint8m2_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vv_i16m8))) +vint16m8_t vwmul(vint8m4_t op0, vint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vv_i16m8_m))) +vint16m8_t vwmul(vbool2_t op0, vint16m8_t op1, vint8m4_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vv_i32mf2))) +vint32mf2_t vwmul(vint16mf4_t op0, vint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vv_i32mf2_m))) +vint32mf2_t vwmul(vbool64_t op0, vint32mf2_t op1, vint16mf4_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vv_i32m1))) +vint32m1_t vwmul(vint16mf2_t op0, vint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vv_i32m1_m))) +vint32m1_t vwmul(vbool32_t op0, vint32m1_t op1, vint16mf2_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vv_i32m2))) +vint32m2_t vwmul(vint16m1_t op0, vint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vv_i32m2_m))) +vint32m2_t vwmul(vbool16_t op0, vint32m2_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vv_i32m4))) +vint32m4_t vwmul(vint16m2_t op0, vint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vv_i32m4_m))) +vint32m4_t vwmul(vbool8_t op0, vint32m4_t op1, vint16m2_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vv_i32m8))) +vint32m8_t vwmul(vint16m4_t op0, vint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vv_i32m8_m))) +vint32m8_t vwmul(vbool4_t op0, vint32m8_t op1, vint16m4_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vv_i64m1))) +vint64m1_t vwmul(vint32mf2_t op0, vint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vv_i64m1_m))) +vint64m1_t vwmul(vbool64_t op0, vint64m1_t op1, vint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vv_i64m2))) +vint64m2_t vwmul(vint32m1_t op0, vint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vv_i64m2_m))) +vint64m2_t vwmul(vbool32_t op0, vint64m2_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vv_i64m4))) +vint64m4_t vwmul(vint32m2_t op0, vint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vv_i64m4_m))) +vint64m4_t vwmul(vbool16_t op0, vint64m4_t op1, vint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vv_i64m8))) +vint64m8_t vwmul(vint32m4_t op0, vint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vv_i64m8_m))) +vint64m8_t vwmul(vbool8_t op0, vint64m8_t op1, vint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_u32m1))) +void vsuxei8(uint32_t * op0, vuint8mf4_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_u32m1_m))) +void vsuxei8(vbool32_t op0, uint32_t * op1, vuint8mf4_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_u32m2))) +void vsuxei8(uint32_t * op0, vuint8mf2_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_u32m2_m))) +void vsuxei8(vbool16_t op0, uint32_t * op1, vuint8mf2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_u32m4))) +void vsuxei8(uint32_t * op0, vuint8m1_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_u32m4_m))) +void vsuxei8(vbool8_t op0, uint32_t * op1, vuint8m1_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_u32m8))) +void vsuxei8(uint32_t * op0, vuint8m2_t op1, vuint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_u32m8_m))) +void vsuxei8(vbool4_t op0, uint32_t * op1, vuint8m2_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_u32mf2))) +void vsuxei8(uint32_t * op0, vuint8mf8_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_u32mf2_m))) +void vsuxei8(vbool64_t op0, uint32_t * op1, vuint8mf8_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vx_i16mf4))) +vint16mf4_t vwmul(vint8mf8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vx_i16mf4_m))) +vint16mf4_t vwmul(vbool64_t op0, vint16mf4_t op1, vint8mf8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vx_i16mf2))) +vint16mf2_t vwmul(vint8mf4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vx_i16mf2_m))) +vint16mf2_t vwmul(vbool32_t op0, vint16mf2_t op1, vint8mf4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vx_i16m1))) +vint16m1_t vwmul(vint8mf2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vx_i16m1_m))) +vint16m1_t vwmul(vbool16_t op0, vint16m1_t op1, vint8mf2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vx_i16m2))) +vint16m2_t vwmul(vint8m1_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vx_i16m2_m))) +vint16m2_t vwmul(vbool8_t op0, vint16m2_t op1, vint8m1_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vx_i16m4))) +vint16m4_t vwmul(vint8m2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vx_i16m4_m))) +vint16m4_t vwmul(vbool4_t op0, vint16m4_t op1, vint8m2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vx_i16m8))) +vint16m8_t vwmul(vint8m4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vx_i16m8_m))) +vint16m8_t vwmul(vbool2_t op0, vint16m8_t op1, vint8m4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vx_i32mf2))) +vint32mf2_t vwmul(vint16mf4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vx_i32mf2_m))) +vint32mf2_t vwmul(vbool64_t op0, vint32mf2_t op1, vint16mf4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vx_i32m1))) +vint32m1_t vwmul(vint16mf2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vx_i32m1_m))) +vint32m1_t vwmul(vbool32_t op0, vint32m1_t op1, vint16mf2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vx_i32m2))) +vint32m2_t vwmul(vint16m1_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vx_i32m2_m))) +vint32m2_t vwmul(vbool16_t op0, vint32m2_t op1, vint16m1_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vx_i32m4))) +vint32m4_t vwmul(vint16m2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vx_i32m4_m))) +vint32m4_t vwmul(vbool8_t op0, vint32m4_t op1, vint16m2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vx_i32m8))) +vint32m8_t vwmul(vint16m4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vx_i32m8_m))) +vint32m8_t vwmul(vbool4_t op0, vint32m8_t op1, vint16m4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vx_i64m1))) +vint64m1_t vwmul(vint32mf2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vx_i64m1_m))) +vint64m1_t vwmul(vbool64_t op0, vint64m1_t op1, vint32mf2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vx_i64m2))) +vint64m2_t vwmul(vint32m1_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vx_i64m2_m))) +vint64m2_t vwmul(vbool32_t op0, vint64m2_t op1, vint32m1_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vx_i64m4))) +vint64m4_t vwmul(vint32m2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vx_i64m4_m))) +vint64m4_t vwmul(vbool16_t op0, vint64m4_t op1, vint32m2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vx_i64m8))) +vint64m8_t vwmul(vint32m4_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmul_vx_i64m8_m))) +vint64m8_t vwmul(vbool8_t op0, vint64m8_t op1, vint32m4_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vv_u16mf4))) +vuint16mf4_t vwmulu(vuint8mf8_t op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vv_u16mf4_m))) +vuint16mf4_t vwmulu(vbool64_t op0, vuint16mf4_t op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vv_u16mf2))) +vuint16mf2_t vwmulu(vuint8mf4_t op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vv_u16mf2_m))) +vuint16mf2_t vwmulu(vbool32_t op0, vuint16mf2_t op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vv_u16m1))) +vuint16m1_t vwmulu(vuint8mf2_t op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vv_u16m1_m))) +vuint16m1_t vwmulu(vbool16_t op0, vuint16m1_t op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vv_u16m2))) +vuint16m2_t vwmulu(vuint8m1_t op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vv_u16m2_m))) +vuint16m2_t vwmulu(vbool8_t op0, vuint16m2_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vv_u16m4))) +vuint16m4_t vwmulu(vuint8m2_t op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vv_u16m4_m))) +vuint16m4_t vwmulu(vbool4_t op0, vuint16m4_t op1, vuint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vv_u16m8))) +vuint16m8_t vwmulu(vuint8m4_t op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vv_u16m8_m))) +vuint16m8_t vwmulu(vbool2_t op0, vuint16m8_t op1, vuint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vv_u32mf2))) +vuint32mf2_t vwmulu(vuint16mf4_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vv_u32mf2_m))) +vuint32mf2_t vwmulu(vbool64_t op0, vuint32mf2_t op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vv_u32m1))) +vuint32m1_t vwmulu(vuint16mf2_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vv_u32m1_m))) +vuint32m1_t vwmulu(vbool32_t op0, vuint32m1_t op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vv_u32m2))) +vuint32m2_t vwmulu(vuint16m1_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vv_u32m2_m))) +vuint32m2_t vwmulu(vbool16_t op0, vuint32m2_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vv_u32m4))) +vuint32m4_t vwmulu(vuint16m2_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vv_u32m4_m))) +vuint32m4_t vwmulu(vbool8_t op0, vuint32m4_t op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vv_u32m8))) +vuint32m8_t vwmulu(vuint16m4_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vv_u32m8_m))) +vuint32m8_t vwmulu(vbool4_t op0, vuint32m8_t op1, vuint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vv_u64m1))) +vuint64m1_t vwmulu(vuint32mf2_t op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vv_u64m1_m))) +vuint64m1_t vwmulu(vbool64_t op0, vuint64m1_t op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vv_u64m2))) +vuint64m2_t vwmulu(vuint32m1_t op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vv_u64m2_m))) +vuint64m2_t vwmulu(vbool32_t op0, vuint64m2_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vv_u64m4))) +vuint64m4_t vwmulu(vuint32m2_t op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vv_u64m4_m))) +vuint64m4_t vwmulu(vbool16_t op0, vuint64m4_t op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vv_u64m8))) +vuint64m8_t vwmulu(vuint32m4_t op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vv_u64m8_m))) +vuint64m8_t vwmulu(vbool8_t op0, vuint64m8_t op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vx_u16mf4))) +vuint16mf4_t vwmulu(vuint8mf8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vx_u16mf4_m))) +vuint16mf4_t vwmulu(vbool64_t op0, vuint16mf4_t op1, vuint8mf8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vx_u16mf2))) +vuint16mf2_t vwmulu(vuint8mf4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vx_u16mf2_m))) +vuint16mf2_t vwmulu(vbool32_t op0, vuint16mf2_t op1, vuint8mf4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vx_u16m1))) +vuint16m1_t vwmulu(vuint8mf2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vx_u16m1_m))) +vuint16m1_t vwmulu(vbool16_t op0, vuint16m1_t op1, vuint8mf2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vx_u16m2))) +vuint16m2_t vwmulu(vuint8m1_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vx_u16m2_m))) +vuint16m2_t vwmulu(vbool8_t op0, vuint16m2_t op1, vuint8m1_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vx_u16m4))) +vuint16m4_t vwmulu(vuint8m2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vx_u16m4_m))) +vuint16m4_t vwmulu(vbool4_t op0, vuint16m4_t op1, vuint8m2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vx_u16m8))) +vuint16m8_t vwmulu(vuint8m4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vx_u16m8_m))) +vuint16m8_t vwmulu(vbool2_t op0, vuint16m8_t op1, vuint8m4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vx_u32mf2))) +vuint32mf2_t vwmulu(vuint16mf4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vx_u32mf2_m))) +vuint32mf2_t vwmulu(vbool64_t op0, vuint32mf2_t op1, vuint16mf4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vx_u32m1))) +vuint32m1_t vwmulu(vuint16mf2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vx_u32m1_m))) +vuint32m1_t vwmulu(vbool32_t op0, vuint32m1_t op1, vuint16mf2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vx_u32m2))) +vuint32m2_t vwmulu(vuint16m1_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vx_u32m2_m))) +vuint32m2_t vwmulu(vbool16_t op0, vuint32m2_t op1, vuint16m1_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vx_u32m4))) +vuint32m4_t vwmulu(vuint16m2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vx_u32m4_m))) +vuint32m4_t vwmulu(vbool8_t op0, vuint32m4_t op1, vuint16m2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vx_u32m8))) +vuint32m8_t vwmulu(vuint16m4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vx_u32m8_m))) +vuint32m8_t vwmulu(vbool4_t op0, vuint32m8_t op1, vuint16m4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vx_u64m1))) +vuint64m1_t vwmulu(vuint32mf2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vx_u64m1_m))) +vuint64m1_t vwmulu(vbool64_t op0, vuint64m1_t op1, vuint32mf2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vx_u64m2))) +vuint64m2_t vwmulu(vuint32m1_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vx_u64m2_m))) +vuint64m2_t vwmulu(vbool32_t op0, vuint64m2_t op1, vuint32m1_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vx_u64m4))) +vuint64m4_t vwmulu(vuint32m2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vx_u64m4_m))) +vuint64m4_t vwmulu(vbool16_t op0, vuint64m4_t op1, vuint32m2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vx_u64m8))) +vuint64m8_t vwmulu(vuint32m4_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulu_vx_u64m8_m))) +vuint64m8_t vwmulu(vbool8_t op0, vuint64m8_t op1, vuint32m4_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vv_i16mf4))) +vint16mf4_t vwmulsu(vint8mf8_t op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vv_i16mf4_m))) +vint16mf4_t vwmulsu(vbool64_t op0, vint16mf4_t op1, vint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vv_i16mf2))) +vint16mf2_t vwmulsu(vint8mf4_t op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vv_i16mf2_m))) +vint16mf2_t vwmulsu(vbool32_t op0, vint16mf2_t op1, vint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vv_i16m1))) +vint16m1_t vwmulsu(vint8mf2_t op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vv_i16m1_m))) +vint16m1_t vwmulsu(vbool16_t op0, vint16m1_t op1, vint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vv_i16m2))) +vint16m2_t vwmulsu(vint8m1_t op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vv_i16m2_m))) +vint16m2_t vwmulsu(vbool8_t op0, vint16m2_t op1, vint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vv_i16m4))) +vint16m4_t vwmulsu(vint8m2_t op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vv_i16m4_m))) +vint16m4_t vwmulsu(vbool4_t op0, vint16m4_t op1, vint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vv_i16m8))) +vint16m8_t vwmulsu(vint8m4_t op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vv_i16m8_m))) +vint16m8_t vwmulsu(vbool2_t op0, vint16m8_t op1, vint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vv_i32mf2))) +vint32mf2_t vwmulsu(vint16mf4_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vv_i32mf2_m))) +vint32mf2_t vwmulsu(vbool64_t op0, vint32mf2_t op1, vint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vv_i32m1))) +vint32m1_t vwmulsu(vint16mf2_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vv_i32m1_m))) +vint32m1_t vwmulsu(vbool32_t op0, vint32m1_t op1, vint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vv_i32m2))) +vint32m2_t vwmulsu(vint16m1_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vv_i32m2_m))) +vint32m2_t vwmulsu(vbool16_t op0, vint32m2_t op1, vint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vv_i32m4))) +vint32m4_t vwmulsu(vint16m2_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vv_i32m4_m))) +vint32m4_t vwmulsu(vbool8_t op0, vint32m4_t op1, vint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vv_i32m8))) +vint32m8_t vwmulsu(vint16m4_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vv_i32m8_m))) +vint32m8_t vwmulsu(vbool4_t op0, vint32m8_t op1, vint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vv_i64m1))) +vint64m1_t vwmulsu(vint32mf2_t op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vv_i64m1_m))) +vint64m1_t vwmulsu(vbool64_t op0, vint64m1_t op1, vint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vv_i64m2))) +vint64m2_t vwmulsu(vint32m1_t op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vv_i64m2_m))) +vint64m2_t vwmulsu(vbool32_t op0, vint64m2_t op1, vint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vv_i64m4))) +vint64m4_t vwmulsu(vint32m2_t op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vv_i64m4_m))) +vint64m4_t vwmulsu(vbool16_t op0, vint64m4_t op1, vint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vv_i64m8))) +vint64m8_t vwmulsu(vint32m4_t op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vv_i64m8_m))) +vint64m8_t vwmulsu(vbool8_t op0, vint64m8_t op1, vint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vx_i16mf4))) +vint16mf4_t vwmulsu(vint8mf8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vx_i16mf4_m))) +vint16mf4_t vwmulsu(vbool64_t op0, vint16mf4_t op1, vint8mf8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vx_i16mf2))) +vint16mf2_t vwmulsu(vint8mf4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vx_i16mf2_m))) +vint16mf2_t vwmulsu(vbool32_t op0, vint16mf2_t op1, vint8mf4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vx_i16m1))) +vint16m1_t vwmulsu(vint8mf2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vx_i16m1_m))) +vint16m1_t vwmulsu(vbool16_t op0, vint16m1_t op1, vint8mf2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vx_i16m2))) +vint16m2_t vwmulsu(vint8m1_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vx_i16m2_m))) +vint16m2_t vwmulsu(vbool8_t op0, vint16m2_t op1, vint8m1_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vx_i16m4))) +vint16m4_t vwmulsu(vint8m2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vx_i16m4_m))) +vint16m4_t vwmulsu(vbool4_t op0, vint16m4_t op1, vint8m2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vx_i16m8))) +vint16m8_t vwmulsu(vint8m4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vx_i16m8_m))) +vint16m8_t vwmulsu(vbool2_t op0, vint16m8_t op1, vint8m4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vx_i32mf2))) +vint32mf2_t vwmulsu(vint16mf4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vx_i32mf2_m))) +vint32mf2_t vwmulsu(vbool64_t op0, vint32mf2_t op1, vint16mf4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vx_i32m1))) +vint32m1_t vwmulsu(vint16mf2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vx_i32m1_m))) +vint32m1_t vwmulsu(vbool32_t op0, vint32m1_t op1, vint16mf2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vx_i32m2))) +vint32m2_t vwmulsu(vint16m1_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vx_i32m2_m))) +vint32m2_t vwmulsu(vbool16_t op0, vint32m2_t op1, vint16m1_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vx_i32m4))) +vint32m4_t vwmulsu(vint16m2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vx_i32m4_m))) +vint32m4_t vwmulsu(vbool8_t op0, vint32m4_t op1, vint16m2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vx_i32m8))) +vint32m8_t vwmulsu(vint16m4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vx_i32m8_m))) +vint32m8_t vwmulsu(vbool4_t op0, vint32m8_t op1, vint16m4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vx_i64m1))) +vint64m1_t vwmulsu(vint32mf2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vx_i64m1_m))) +vint64m1_t vwmulsu(vbool64_t op0, vint64m1_t op1, vint32mf2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vx_i64m2))) +vint64m2_t vwmulsu(vint32m1_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vx_i64m2_m))) +vint64m2_t vwmulsu(vbool32_t op0, vint64m2_t op1, vint32m1_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vx_i64m4))) +vint64m4_t vwmulsu(vint32m2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vx_i64m4_m))) +vint64m4_t vwmulsu(vbool16_t op0, vint64m4_t op1, vint32m2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vx_i64m8))) +vint64m8_t vwmulsu(vint32m4_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmulsu_vx_i64m8_m))) +vint64m8_t vwmulsu(vbool8_t op0, vint64m8_t op1, vint32m4_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_i8m1))) +vint8m1_t vmacc(vint8m1_t op0, vint8m1_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_i8m1_m))) +vint8m1_t vmacc(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_i8m2))) +vint8m2_t vmacc(vint8m2_t op0, vint8m2_t op1, vint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_i8m2_m))) +vint8m2_t vmacc(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_i8m4))) +vint8m4_t vmacc(vint8m4_t op0, vint8m4_t op1, vint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_i8m4_m))) +vint8m4_t vmacc(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_i8m8))) +vint8m8_t vmacc(vint8m8_t op0, vint8m8_t op1, vint8m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_i8m8_m))) +vint8m8_t vmacc(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, vint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_i8mf2))) +vint8mf2_t vmacc(vint8mf2_t op0, vint8mf2_t op1, vint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_i8mf2_m))) +vint8mf2_t vmacc(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_i8mf4))) +vint8mf4_t vmacc(vint8mf4_t op0, vint8mf4_t op1, vint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_i8mf4_m))) +vint8mf4_t vmacc(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_i8mf8))) +vint8mf8_t vmacc(vint8mf8_t op0, vint8mf8_t op1, vint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_i8mf8_m))) +vint8mf8_t vmacc(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_i16m1))) +vint16m1_t vmacc(vint16m1_t op0, vint16m1_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_i16m1_m))) +vint16m1_t vmacc(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_i16m2))) +vint16m2_t vmacc(vint16m2_t op0, vint16m2_t op1, vint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_i16m2_m))) +vint16m2_t vmacc(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_i16m4))) +vint16m4_t vmacc(vint16m4_t op0, vint16m4_t op1, vint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_i16m4_m))) +vint16m4_t vmacc(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_i16m8))) +vint16m8_t vmacc(vint16m8_t op0, vint16m8_t op1, vint16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_i16m8_m))) +vint16m8_t vmacc(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, vint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_i16mf2))) +vint16mf2_t vmacc(vint16mf2_t op0, vint16mf2_t op1, vint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_i16mf2_m))) +vint16mf2_t vmacc(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_i16mf4))) +vint16mf4_t vmacc(vint16mf4_t op0, vint16mf4_t op1, vint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_i16mf4_m))) +vint16mf4_t vmacc(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_i32m1))) +vint32m1_t vmacc(vint32m1_t op0, vint32m1_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_i32m1_m))) +vint32m1_t vmacc(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_i32m2))) +vint32m2_t vmacc(vint32m2_t op0, vint32m2_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_i32m2_m))) +vint32m2_t vmacc(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_i32m4))) +vint32m4_t vmacc(vint32m4_t op0, vint32m4_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_i32m4_m))) +vint32m4_t vmacc(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_i32m8))) +vint32m8_t vmacc(vint32m8_t op0, vint32m8_t op1, vint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_i32m8_m))) +vint32m8_t vmacc(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_i32mf2))) +vint32mf2_t vmacc(vint32mf2_t op0, vint32mf2_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_i32mf2_m))) +vint32mf2_t vmacc(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_i64m1))) +vint64m1_t vmacc(vint64m1_t op0, vint64m1_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_i64m1_m))) +vint64m1_t vmacc(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_i64m2))) +vint64m2_t vmacc(vint64m2_t op0, vint64m2_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_i64m2_m))) +vint64m2_t vmacc(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_i64m4))) +vint64m4_t vmacc(vint64m4_t op0, vint64m4_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_i64m4_m))) +vint64m4_t vmacc(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_i64m8))) +vint64m8_t vmacc(vint64m8_t op0, vint64m8_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_i64m8_m))) +vint64m8_t vmacc(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_i8m1))) +vint8m1_t vmacc(vint8m1_t op0, int8_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_i8m1_m))) +vint8m1_t vmacc(vbool8_t op0, vint8m1_t op1, int8_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_i8m2))) +vint8m2_t vmacc(vint8m2_t op0, int8_t op1, vint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_i8m2_m))) +vint8m2_t vmacc(vbool4_t op0, vint8m2_t op1, int8_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_i8m4))) +vint8m4_t vmacc(vint8m4_t op0, int8_t op1, vint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_i8m4_m))) +vint8m4_t vmacc(vbool2_t op0, vint8m4_t op1, int8_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_i8m8))) +vint8m8_t vmacc(vint8m8_t op0, int8_t op1, vint8m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_i8m8_m))) +vint8m8_t vmacc(vbool1_t op0, vint8m8_t op1, int8_t op2, vint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_i8mf2))) +vint8mf2_t vmacc(vint8mf2_t op0, int8_t op1, vint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_i8mf2_m))) +vint8mf2_t vmacc(vbool16_t op0, vint8mf2_t op1, int8_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_i8mf4))) +vint8mf4_t vmacc(vint8mf4_t op0, int8_t op1, vint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_i8mf4_m))) +vint8mf4_t vmacc(vbool32_t op0, vint8mf4_t op1, int8_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_i8mf8))) +vint8mf8_t vmacc(vint8mf8_t op0, int8_t op1, vint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_i8mf8_m))) +vint8mf8_t vmacc(vbool64_t op0, vint8mf8_t op1, int8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_i16m1))) +vint16m1_t vmacc(vint16m1_t op0, int16_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_i16m1_m))) +vint16m1_t vmacc(vbool16_t op0, vint16m1_t op1, int16_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_i16m2))) +vint16m2_t vmacc(vint16m2_t op0, int16_t op1, vint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_i16m2_m))) +vint16m2_t vmacc(vbool8_t op0, vint16m2_t op1, int16_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_i16m4))) +vint16m4_t vmacc(vint16m4_t op0, int16_t op1, vint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_i16m4_m))) +vint16m4_t vmacc(vbool4_t op0, vint16m4_t op1, int16_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_i16m8))) +vint16m8_t vmacc(vint16m8_t op0, int16_t op1, vint16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_i16m8_m))) +vint16m8_t vmacc(vbool2_t op0, vint16m8_t op1, int16_t op2, vint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_i16mf2))) +vint16mf2_t vmacc(vint16mf2_t op0, int16_t op1, vint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_i16mf2_m))) +vint16mf2_t vmacc(vbool32_t op0, vint16mf2_t op1, int16_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_i16mf4))) +vint16mf4_t vmacc(vint16mf4_t op0, int16_t op1, vint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_i16mf4_m))) +vint16mf4_t vmacc(vbool64_t op0, vint16mf4_t op1, int16_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_i32m1))) +vint32m1_t vmacc(vint32m1_t op0, int32_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_i32m1_m))) +vint32m1_t vmacc(vbool32_t op0, vint32m1_t op1, int32_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_i32m2))) +vint32m2_t vmacc(vint32m2_t op0, int32_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_i32m2_m))) +vint32m2_t vmacc(vbool16_t op0, vint32m2_t op1, int32_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_i32m4))) +vint32m4_t vmacc(vint32m4_t op0, int32_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_i32m4_m))) +vint32m4_t vmacc(vbool8_t op0, vint32m4_t op1, int32_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_i32m8))) +vint32m8_t vmacc(vint32m8_t op0, int32_t op1, vint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_i32m8_m))) +vint32m8_t vmacc(vbool4_t op0, vint32m8_t op1, int32_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_i32mf2))) +vint32mf2_t vmacc(vint32mf2_t op0, int32_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_i32mf2_m))) +vint32mf2_t vmacc(vbool64_t op0, vint32mf2_t op1, int32_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_i64m1))) +vint64m1_t vmacc(vint64m1_t op0, int64_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_i64m1_m))) +vint64m1_t vmacc(vbool64_t op0, vint64m1_t op1, int64_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_i64m2))) +vint64m2_t vmacc(vint64m2_t op0, int64_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_i64m2_m))) +vint64m2_t vmacc(vbool32_t op0, vint64m2_t op1, int64_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_i64m4))) +vint64m4_t vmacc(vint64m4_t op0, int64_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_i64m4_m))) +vint64m4_t vmacc(vbool16_t op0, vint64m4_t op1, int64_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_i64m8))) +vint64m8_t vmacc(vint64m8_t op0, int64_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_i64m8_m))) +vint64m8_t vmacc(vbool8_t op0, vint64m8_t op1, int64_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_u8m1))) +vuint8m1_t vmacc(vuint8m1_t op0, vuint8m1_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_u8m1_m))) +vuint8m1_t vmacc(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_u8m2))) +vuint8m2_t vmacc(vuint8m2_t op0, vuint8m2_t op1, vuint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_u8m2_m))) +vuint8m2_t vmacc(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_u8m4))) +vuint8m4_t vmacc(vuint8m4_t op0, vuint8m4_t op1, vuint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_u8m4_m))) +vuint8m4_t vmacc(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_u8m8))) +vuint8m8_t vmacc(vuint8m8_t op0, vuint8m8_t op1, vuint8m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_u8m8_m))) +vuint8m8_t vmacc(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, vuint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_u8mf2))) +vuint8mf2_t vmacc(vuint8mf2_t op0, vuint8mf2_t op1, vuint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_u8mf2_m))) +vuint8mf2_t vmacc(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_u8mf4))) +vuint8mf4_t vmacc(vuint8mf4_t op0, vuint8mf4_t op1, vuint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_u8mf4_m))) +vuint8mf4_t vmacc(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_u8mf8))) +vuint8mf8_t vmacc(vuint8mf8_t op0, vuint8mf8_t op1, vuint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_u8mf8_m))) +vuint8mf8_t vmacc(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_u16m1))) +vuint16m1_t vmacc(vuint16m1_t op0, vuint16m1_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_u16m1_m))) +vuint16m1_t vmacc(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_u16m2))) +vuint16m2_t vmacc(vuint16m2_t op0, vuint16m2_t op1, vuint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_u16m2_m))) +vuint16m2_t vmacc(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_u16m4))) +vuint16m4_t vmacc(vuint16m4_t op0, vuint16m4_t op1, vuint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_u16m4_m))) +vuint16m4_t vmacc(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_u16m8))) +vuint16m8_t vmacc(vuint16m8_t op0, vuint16m8_t op1, vuint16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_u16m8_m))) +vuint16m8_t vmacc(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_u16mf2))) +vuint16mf2_t vmacc(vuint16mf2_t op0, vuint16mf2_t op1, vuint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_u16mf2_m))) +vuint16mf2_t vmacc(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_u16mf4))) +vuint16mf4_t vmacc(vuint16mf4_t op0, vuint16mf4_t op1, vuint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_u16mf4_m))) +vuint16mf4_t vmacc(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_u32m1))) +vuint32m1_t vmacc(vuint32m1_t op0, vuint32m1_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_u32m1_m))) +vuint32m1_t vmacc(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_u32m2))) +vuint32m2_t vmacc(vuint32m2_t op0, vuint32m2_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_u32m2_m))) +vuint32m2_t vmacc(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_u32m4))) +vuint32m4_t vmacc(vuint32m4_t op0, vuint32m4_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_u32m4_m))) +vuint32m4_t vmacc(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_u32m8))) +vuint32m8_t vmacc(vuint32m8_t op0, vuint32m8_t op1, vuint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_u32m8_m))) +vuint32m8_t vmacc(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_u32mf2))) +vuint32mf2_t vmacc(vuint32mf2_t op0, vuint32mf2_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_u32mf2_m))) +vuint32mf2_t vmacc(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_u64m1))) +vuint64m1_t vmacc(vuint64m1_t op0, vuint64m1_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_u64m1_m))) +vuint64m1_t vmacc(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_u64m2))) +vuint64m2_t vmacc(vuint64m2_t op0, vuint64m2_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_u64m2_m))) +vuint64m2_t vmacc(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_u64m4))) +vuint64m4_t vmacc(vuint64m4_t op0, vuint64m4_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_u64m4_m))) +vuint64m4_t vmacc(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_u64m8))) +vuint64m8_t vmacc(vuint64m8_t op0, vuint64m8_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vv_u64m8_m))) +vuint64m8_t vmacc(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_u8m1))) +vuint8m1_t vmacc(vuint8m1_t op0, uint8_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_u8m1_m))) +vuint8m1_t vmacc(vbool8_t op0, vuint8m1_t op1, uint8_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_u8m2))) +vuint8m2_t vmacc(vuint8m2_t op0, uint8_t op1, vuint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_u8m2_m))) +vuint8m2_t vmacc(vbool4_t op0, vuint8m2_t op1, uint8_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_u8m4))) +vuint8m4_t vmacc(vuint8m4_t op0, uint8_t op1, vuint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_u8m4_m))) +vuint8m4_t vmacc(vbool2_t op0, vuint8m4_t op1, uint8_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_u8m8))) +vuint8m8_t vmacc(vuint8m8_t op0, uint8_t op1, vuint8m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_u8m8_m))) +vuint8m8_t vmacc(vbool1_t op0, vuint8m8_t op1, uint8_t op2, vuint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_u8mf2))) +vuint8mf2_t vmacc(vuint8mf2_t op0, uint8_t op1, vuint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_u8mf2_m))) +vuint8mf2_t vmacc(vbool16_t op0, vuint8mf2_t op1, uint8_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_u8mf4))) +vuint8mf4_t vmacc(vuint8mf4_t op0, uint8_t op1, vuint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_u8mf4_m))) +vuint8mf4_t vmacc(vbool32_t op0, vuint8mf4_t op1, uint8_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_u8mf8))) +vuint8mf8_t vmacc(vuint8mf8_t op0, uint8_t op1, vuint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_u8mf8_m))) +vuint8mf8_t vmacc(vbool64_t op0, vuint8mf8_t op1, uint8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_u16m1))) +vuint16m1_t vmacc(vuint16m1_t op0, uint16_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_u16m1_m))) +vuint16m1_t vmacc(vbool16_t op0, vuint16m1_t op1, uint16_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_u16m2))) +vuint16m2_t vmacc(vuint16m2_t op0, uint16_t op1, vuint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_u16m2_m))) +vuint16m2_t vmacc(vbool8_t op0, vuint16m2_t op1, uint16_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_u16m4))) +vuint16m4_t vmacc(vuint16m4_t op0, uint16_t op1, vuint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_u16m4_m))) +vuint16m4_t vmacc(vbool4_t op0, vuint16m4_t op1, uint16_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_u16m8))) +vuint16m8_t vmacc(vuint16m8_t op0, uint16_t op1, vuint16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_u16m8_m))) +vuint16m8_t vmacc(vbool2_t op0, vuint16m8_t op1, uint16_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_u16mf2))) +vuint16mf2_t vmacc(vuint16mf2_t op0, uint16_t op1, vuint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_u16mf2_m))) +vuint16mf2_t vmacc(vbool32_t op0, vuint16mf2_t op1, uint16_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_u16mf4))) +vuint16mf4_t vmacc(vuint16mf4_t op0, uint16_t op1, vuint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_u16mf4_m))) +vuint16mf4_t vmacc(vbool64_t op0, vuint16mf4_t op1, uint16_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_u32m1))) +vuint32m1_t vmacc(vuint32m1_t op0, uint32_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_u32m1_m))) +vuint32m1_t vmacc(vbool32_t op0, vuint32m1_t op1, uint32_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_u32m2))) +vuint32m2_t vmacc(vuint32m2_t op0, uint32_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_u32m2_m))) +vuint32m2_t vmacc(vbool16_t op0, vuint32m2_t op1, uint32_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_u32m4))) +vuint32m4_t vmacc(vuint32m4_t op0, uint32_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_u32m4_m))) +vuint32m4_t vmacc(vbool8_t op0, vuint32m4_t op1, uint32_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_u32m8))) +vuint32m8_t vmacc(vuint32m8_t op0, uint32_t op1, vuint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_u32m8_m))) +vuint32m8_t vmacc(vbool4_t op0, vuint32m8_t op1, uint32_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_u32mf2))) +vuint32mf2_t vmacc(vuint32mf2_t op0, uint32_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_u32mf2_m))) +vuint32mf2_t vmacc(vbool64_t op0, vuint32mf2_t op1, uint32_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_u64m1))) +vuint64m1_t vmacc(vuint64m1_t op0, uint64_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_u64m1_m))) +vuint64m1_t vmacc(vbool64_t op0, vuint64m1_t op1, uint64_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_u64m2))) +vuint64m2_t vmacc(vuint64m2_t op0, uint64_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_u64m2_m))) +vuint64m2_t vmacc(vbool32_t op0, vuint64m2_t op1, uint64_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_u64m4))) +vuint64m4_t vmacc(vuint64m4_t op0, uint64_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_u64m4_m))) +vuint64m4_t vmacc(vbool16_t op0, vuint64m4_t op1, uint64_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_u64m8))) +vuint64m8_t vmacc(vuint64m8_t op0, uint64_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmacc_vx_u64m8_m))) +vuint64m8_t vmacc(vbool8_t op0, vuint64m8_t op1, uint64_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_i8m1))) +vint8m1_t vnmsac(vint8m1_t op0, vint8m1_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_i8m1_m))) +vint8m1_t vnmsac(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_i8m2))) +vint8m2_t vnmsac(vint8m2_t op0, vint8m2_t op1, vint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_i8m2_m))) +vint8m2_t vnmsac(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_i8m4))) +vint8m4_t vnmsac(vint8m4_t op0, vint8m4_t op1, vint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_i8m4_m))) +vint8m4_t vnmsac(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_i8m8))) +vint8m8_t vnmsac(vint8m8_t op0, vint8m8_t op1, vint8m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_i8m8_m))) +vint8m8_t vnmsac(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, vint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_i8mf2))) +vint8mf2_t vnmsac(vint8mf2_t op0, vint8mf2_t op1, vint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_i8mf2_m))) +vint8mf2_t vnmsac(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_i8mf4))) +vint8mf4_t vnmsac(vint8mf4_t op0, vint8mf4_t op1, vint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_i8mf4_m))) +vint8mf4_t vnmsac(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_i8mf8))) +vint8mf8_t vnmsac(vint8mf8_t op0, vint8mf8_t op1, vint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_i8mf8_m))) +vint8mf8_t vnmsac(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_i16m1))) +vint16m1_t vnmsac(vint16m1_t op0, vint16m1_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_i16m1_m))) +vint16m1_t vnmsac(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_i16m2))) +vint16m2_t vnmsac(vint16m2_t op0, vint16m2_t op1, vint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_i16m2_m))) +vint16m2_t vnmsac(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_i16m4))) +vint16m4_t vnmsac(vint16m4_t op0, vint16m4_t op1, vint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_i16m4_m))) +vint16m4_t vnmsac(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_i16m8))) +vint16m8_t vnmsac(vint16m8_t op0, vint16m8_t op1, vint16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_i16m8_m))) +vint16m8_t vnmsac(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, vint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_i16mf2))) +vint16mf2_t vnmsac(vint16mf2_t op0, vint16mf2_t op1, vint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_i16mf2_m))) +vint16mf2_t vnmsac(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_i16mf4))) +vint16mf4_t vnmsac(vint16mf4_t op0, vint16mf4_t op1, vint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_i16mf4_m))) +vint16mf4_t vnmsac(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_i32m1))) +vint32m1_t vnmsac(vint32m1_t op0, vint32m1_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_i32m1_m))) +vint32m1_t vnmsac(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_i32m2))) +vint32m2_t vnmsac(vint32m2_t op0, vint32m2_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_i32m2_m))) +vint32m2_t vnmsac(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_i32m4))) +vint32m4_t vnmsac(vint32m4_t op0, vint32m4_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_i32m4_m))) +vint32m4_t vnmsac(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_i32m8))) +vint32m8_t vnmsac(vint32m8_t op0, vint32m8_t op1, vint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_i32m8_m))) +vint32m8_t vnmsac(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_i32mf2))) +vint32mf2_t vnmsac(vint32mf2_t op0, vint32mf2_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_i32mf2_m))) +vint32mf2_t vnmsac(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_i64m1))) +vint64m1_t vnmsac(vint64m1_t op0, vint64m1_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_i64m1_m))) +vint64m1_t vnmsac(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_i64m2))) +vint64m2_t vnmsac(vint64m2_t op0, vint64m2_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_i64m2_m))) +vint64m2_t vnmsac(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_i64m4))) +vint64m4_t vnmsac(vint64m4_t op0, vint64m4_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_i64m4_m))) +vint64m4_t vnmsac(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_i64m8))) +vint64m8_t vnmsac(vint64m8_t op0, vint64m8_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_i64m8_m))) +vint64m8_t vnmsac(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_i32m1))) +void vsuxei16(int32_t * op0, vuint16mf2_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_i32m1_m))) +void vsuxei16(vbool32_t op0, int32_t * op1, vuint16mf2_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_i32m2))) +void vsuxei16(int32_t * op0, vuint16m1_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_i32m2_m))) +void vsuxei16(vbool16_t op0, int32_t * op1, vuint16m1_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_i32m4))) +void vsuxei16(int32_t * op0, vuint16m2_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_i32m4_m))) +void vsuxei16(vbool8_t op0, int32_t * op1, vuint16m2_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_i32m8))) +void vsuxei16(int32_t * op0, vuint16m4_t op1, vint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_i32m8_m))) +void vsuxei16(vbool4_t op0, int32_t * op1, vuint16m4_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_i32mf2))) +void vsuxei16(int32_t * op0, vuint16mf4_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_i32mf2_m))) +void vsuxei16(vbool64_t op0, int32_t * op1, vuint16mf4_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_i8m1))) +vint8m1_t vnmsac(vint8m1_t op0, int8_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_i8m1_m))) +vint8m1_t vnmsac(vbool8_t op0, vint8m1_t op1, int8_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_i8m2))) +vint8m2_t vnmsac(vint8m2_t op0, int8_t op1, vint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_i8m2_m))) +vint8m2_t vnmsac(vbool4_t op0, vint8m2_t op1, int8_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_i8m4))) +vint8m4_t vnmsac(vint8m4_t op0, int8_t op1, vint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_i8m4_m))) +vint8m4_t vnmsac(vbool2_t op0, vint8m4_t op1, int8_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_i8m8))) +vint8m8_t vnmsac(vint8m8_t op0, int8_t op1, vint8m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_i8m8_m))) +vint8m8_t vnmsac(vbool1_t op0, vint8m8_t op1, int8_t op2, vint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_i8mf2))) +vint8mf2_t vnmsac(vint8mf2_t op0, int8_t op1, vint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_i8mf2_m))) +vint8mf2_t vnmsac(vbool16_t op0, vint8mf2_t op1, int8_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_i8mf4))) +vint8mf4_t vnmsac(vint8mf4_t op0, int8_t op1, vint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_i8mf4_m))) +vint8mf4_t vnmsac(vbool32_t op0, vint8mf4_t op1, int8_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_i8mf8))) +vint8mf8_t vnmsac(vint8mf8_t op0, int8_t op1, vint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_i8mf8_m))) +vint8mf8_t vnmsac(vbool64_t op0, vint8mf8_t op1, int8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_i16m1))) +vint16m1_t vnmsac(vint16m1_t op0, int16_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_i16m1_m))) +vint16m1_t vnmsac(vbool16_t op0, vint16m1_t op1, int16_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_i16m2))) +vint16m2_t vnmsac(vint16m2_t op0, int16_t op1, vint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_i16m2_m))) +vint16m2_t vnmsac(vbool8_t op0, vint16m2_t op1, int16_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_i16m4))) +vint16m4_t vnmsac(vint16m4_t op0, int16_t op1, vint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_i16m4_m))) +vint16m4_t vnmsac(vbool4_t op0, vint16m4_t op1, int16_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_i16m8))) +vint16m8_t vnmsac(vint16m8_t op0, int16_t op1, vint16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_i16m8_m))) +vint16m8_t vnmsac(vbool2_t op0, vint16m8_t op1, int16_t op2, vint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_i16mf2))) +vint16mf2_t vnmsac(vint16mf2_t op0, int16_t op1, vint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_i16mf2_m))) +vint16mf2_t vnmsac(vbool32_t op0, vint16mf2_t op1, int16_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_i16mf4))) +vint16mf4_t vnmsac(vint16mf4_t op0, int16_t op1, vint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_i16mf4_m))) +vint16mf4_t vnmsac(vbool64_t op0, vint16mf4_t op1, int16_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_i32m1))) +vint32m1_t vnmsac(vint32m1_t op0, int32_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_i32m1_m))) +vint32m1_t vnmsac(vbool32_t op0, vint32m1_t op1, int32_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_i32m2))) +vint32m2_t vnmsac(vint32m2_t op0, int32_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_i32m2_m))) +vint32m2_t vnmsac(vbool16_t op0, vint32m2_t op1, int32_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_i32m4))) +vint32m4_t vnmsac(vint32m4_t op0, int32_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_i32m4_m))) +vint32m4_t vnmsac(vbool8_t op0, vint32m4_t op1, int32_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_i32m8))) +vint32m8_t vnmsac(vint32m8_t op0, int32_t op1, vint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_i32m8_m))) +vint32m8_t vnmsac(vbool4_t op0, vint32m8_t op1, int32_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_i32mf2))) +vint32mf2_t vnmsac(vint32mf2_t op0, int32_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_i32mf2_m))) +vint32mf2_t vnmsac(vbool64_t op0, vint32mf2_t op1, int32_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_i64m1))) +vint64m1_t vnmsac(vint64m1_t op0, int64_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_i64m1_m))) +vint64m1_t vnmsac(vbool64_t op0, vint64m1_t op1, int64_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_i64m2))) +vint64m2_t vnmsac(vint64m2_t op0, int64_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_i64m2_m))) +vint64m2_t vnmsac(vbool32_t op0, vint64m2_t op1, int64_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_i64m4))) +vint64m4_t vnmsac(vint64m4_t op0, int64_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_i64m4_m))) +vint64m4_t vnmsac(vbool16_t op0, vint64m4_t op1, int64_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_i64m8))) +vint64m8_t vnmsac(vint64m8_t op0, int64_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_i64m8_m))) +vint64m8_t vnmsac(vbool8_t op0, vint64m8_t op1, int64_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_u8m1))) +vuint8m1_t vnmsac(vuint8m1_t op0, vuint8m1_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_u8m1_m))) +vuint8m1_t vnmsac(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_u8m2))) +vuint8m2_t vnmsac(vuint8m2_t op0, vuint8m2_t op1, vuint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_u8m2_m))) +vuint8m2_t vnmsac(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_u8m4))) +vuint8m4_t vnmsac(vuint8m4_t op0, vuint8m4_t op1, vuint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_u8m4_m))) +vuint8m4_t vnmsac(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_u8m8))) +vuint8m8_t vnmsac(vuint8m8_t op0, vuint8m8_t op1, vuint8m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_u8m8_m))) +vuint8m8_t vnmsac(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, vuint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_u8mf2))) +vuint8mf2_t vnmsac(vuint8mf2_t op0, vuint8mf2_t op1, vuint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_u8mf2_m))) +vuint8mf2_t vnmsac(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_u8mf4))) +vuint8mf4_t vnmsac(vuint8mf4_t op0, vuint8mf4_t op1, vuint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_u8mf4_m))) +vuint8mf4_t vnmsac(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_u8mf8))) +vuint8mf8_t vnmsac(vuint8mf8_t op0, vuint8mf8_t op1, vuint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_u8mf8_m))) +vuint8mf8_t vnmsac(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_u16m1))) +vuint16m1_t vnmsac(vuint16m1_t op0, vuint16m1_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_u16m1_m))) +vuint16m1_t vnmsac(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_u16m2))) +vuint16m2_t vnmsac(vuint16m2_t op0, vuint16m2_t op1, vuint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_u16m2_m))) +vuint16m2_t vnmsac(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_u16m4))) +vuint16m4_t vnmsac(vuint16m4_t op0, vuint16m4_t op1, vuint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_u16m4_m))) +vuint16m4_t vnmsac(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_u16m8))) +vuint16m8_t vnmsac(vuint16m8_t op0, vuint16m8_t op1, vuint16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_u16m8_m))) +vuint16m8_t vnmsac(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_u16mf2))) +vuint16mf2_t vnmsac(vuint16mf2_t op0, vuint16mf2_t op1, vuint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_u16mf2_m))) +vuint16mf2_t vnmsac(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_u16mf4))) +vuint16mf4_t vnmsac(vuint16mf4_t op0, vuint16mf4_t op1, vuint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_u16mf4_m))) +vuint16mf4_t vnmsac(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_u32m1))) +vuint32m1_t vnmsac(vuint32m1_t op0, vuint32m1_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_u32m1_m))) +vuint32m1_t vnmsac(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_u32m2))) +vuint32m2_t vnmsac(vuint32m2_t op0, vuint32m2_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_u32m2_m))) +vuint32m2_t vnmsac(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_u32m4))) +vuint32m4_t vnmsac(vuint32m4_t op0, vuint32m4_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_u32m4_m))) +vuint32m4_t vnmsac(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_u32m8))) +vuint32m8_t vnmsac(vuint32m8_t op0, vuint32m8_t op1, vuint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_u32m8_m))) +vuint32m8_t vnmsac(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_u32mf2))) +vuint32mf2_t vnmsac(vuint32mf2_t op0, vuint32mf2_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_u32mf2_m))) +vuint32mf2_t vnmsac(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_u64m1))) +vuint64m1_t vnmsac(vuint64m1_t op0, vuint64m1_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_u64m1_m))) +vuint64m1_t vnmsac(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_u64m2))) +vuint64m2_t vnmsac(vuint64m2_t op0, vuint64m2_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_u64m2_m))) +vuint64m2_t vnmsac(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_u64m4))) +vuint64m4_t vnmsac(vuint64m4_t op0, vuint64m4_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_u64m4_m))) +vuint64m4_t vnmsac(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_u64m8))) +vuint64m8_t vnmsac(vuint64m8_t op0, vuint64m8_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vv_u64m8_m))) +vuint64m8_t vnmsac(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_u8m1))) +vuint8m1_t vnmsac(vuint8m1_t op0, uint8_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_u8m1_m))) +vuint8m1_t vnmsac(vbool8_t op0, vuint8m1_t op1, uint8_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_u8m2))) +vuint8m2_t vnmsac(vuint8m2_t op0, uint8_t op1, vuint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_u8m2_m))) +vuint8m2_t vnmsac(vbool4_t op0, vuint8m2_t op1, uint8_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_u8m4))) +vuint8m4_t vnmsac(vuint8m4_t op0, uint8_t op1, vuint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_u8m4_m))) +vuint8m4_t vnmsac(vbool2_t op0, vuint8m4_t op1, uint8_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_u8m8))) +vuint8m8_t vnmsac(vuint8m8_t op0, uint8_t op1, vuint8m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_u8m8_m))) +vuint8m8_t vnmsac(vbool1_t op0, vuint8m8_t op1, uint8_t op2, vuint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_u8mf2))) +vuint8mf2_t vnmsac(vuint8mf2_t op0, uint8_t op1, vuint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_u8mf2_m))) +vuint8mf2_t vnmsac(vbool16_t op0, vuint8mf2_t op1, uint8_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_u8mf4))) +vuint8mf4_t vnmsac(vuint8mf4_t op0, uint8_t op1, vuint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_u8mf4_m))) +vuint8mf4_t vnmsac(vbool32_t op0, vuint8mf4_t op1, uint8_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_u8mf8))) +vuint8mf8_t vnmsac(vuint8mf8_t op0, uint8_t op1, vuint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_u8mf8_m))) +vuint8mf8_t vnmsac(vbool64_t op0, vuint8mf8_t op1, uint8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_u16m1))) +vuint16m1_t vnmsac(vuint16m1_t op0, uint16_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_u16m1_m))) +vuint16m1_t vnmsac(vbool16_t op0, vuint16m1_t op1, uint16_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_u16m2))) +vuint16m2_t vnmsac(vuint16m2_t op0, uint16_t op1, vuint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_u16m2_m))) +vuint16m2_t vnmsac(vbool8_t op0, vuint16m2_t op1, uint16_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_u16m4))) +vuint16m4_t vnmsac(vuint16m4_t op0, uint16_t op1, vuint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_u16m4_m))) +vuint16m4_t vnmsac(vbool4_t op0, vuint16m4_t op1, uint16_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_u16m8))) +vuint16m8_t vnmsac(vuint16m8_t op0, uint16_t op1, vuint16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_u16m8_m))) +vuint16m8_t vnmsac(vbool2_t op0, vuint16m8_t op1, uint16_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_u16mf2))) +vuint16mf2_t vnmsac(vuint16mf2_t op0, uint16_t op1, vuint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_u16mf2_m))) +vuint16mf2_t vnmsac(vbool32_t op0, vuint16mf2_t op1, uint16_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_u16mf4))) +vuint16mf4_t vnmsac(vuint16mf4_t op0, uint16_t op1, vuint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_u16mf4_m))) +vuint16mf4_t vnmsac(vbool64_t op0, vuint16mf4_t op1, uint16_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_u32m1))) +vuint32m1_t vnmsac(vuint32m1_t op0, uint32_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_u32m1_m))) +vuint32m1_t vnmsac(vbool32_t op0, vuint32m1_t op1, uint32_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_u32m2))) +vuint32m2_t vnmsac(vuint32m2_t op0, uint32_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_u32m2_m))) +vuint32m2_t vnmsac(vbool16_t op0, vuint32m2_t op1, uint32_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_u32m4))) +vuint32m4_t vnmsac(vuint32m4_t op0, uint32_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_u32m4_m))) +vuint32m4_t vnmsac(vbool8_t op0, vuint32m4_t op1, uint32_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_u32m8))) +vuint32m8_t vnmsac(vuint32m8_t op0, uint32_t op1, vuint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_u32m8_m))) +vuint32m8_t vnmsac(vbool4_t op0, vuint32m8_t op1, uint32_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_u32mf2))) +vuint32mf2_t vnmsac(vuint32mf2_t op0, uint32_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_u32mf2_m))) +vuint32mf2_t vnmsac(vbool64_t op0, vuint32mf2_t op1, uint32_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_u64m1))) +vuint64m1_t vnmsac(vuint64m1_t op0, uint64_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_u64m1_m))) +vuint64m1_t vnmsac(vbool64_t op0, vuint64m1_t op1, uint64_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_u64m2))) +vuint64m2_t vnmsac(vuint64m2_t op0, uint64_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_u64m2_m))) +vuint64m2_t vnmsac(vbool32_t op0, vuint64m2_t op1, uint64_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_u64m4))) +vuint64m4_t vnmsac(vuint64m4_t op0, uint64_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_u64m4_m))) +vuint64m4_t vnmsac(vbool16_t op0, vuint64m4_t op1, uint64_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_u64m8))) +vuint64m8_t vnmsac(vuint64m8_t op0, uint64_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsac_vx_u64m8_m))) +vuint64m8_t vnmsac(vbool8_t op0, vuint64m8_t op1, uint64_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_i8m1))) +vint8m1_t vmadd(vint8m1_t op0, vint8m1_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_i8m1_m))) +vint8m1_t vmadd(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_i8m2))) +vint8m2_t vmadd(vint8m2_t op0, vint8m2_t op1, vint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_i8m2_m))) +vint8m2_t vmadd(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_i8m4))) +vint8m4_t vmadd(vint8m4_t op0, vint8m4_t op1, vint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_i8m4_m))) +vint8m4_t vmadd(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_i8m8))) +vint8m8_t vmadd(vint8m8_t op0, vint8m8_t op1, vint8m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_i8m8_m))) +vint8m8_t vmadd(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, vint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_i8mf2))) +vint8mf2_t vmadd(vint8mf2_t op0, vint8mf2_t op1, vint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_i8mf2_m))) +vint8mf2_t vmadd(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_i8mf4))) +vint8mf4_t vmadd(vint8mf4_t op0, vint8mf4_t op1, vint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_i8mf4_m))) +vint8mf4_t vmadd(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_i8mf8))) +vint8mf8_t vmadd(vint8mf8_t op0, vint8mf8_t op1, vint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_i8mf8_m))) +vint8mf8_t vmadd(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_i16m1))) +vint16m1_t vmadd(vint16m1_t op0, vint16m1_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_i16m1_m))) +vint16m1_t vmadd(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_i16m2))) +vint16m2_t vmadd(vint16m2_t op0, vint16m2_t op1, vint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_i16m2_m))) +vint16m2_t vmadd(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_i16m4))) +vint16m4_t vmadd(vint16m4_t op0, vint16m4_t op1, vint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_i16m4_m))) +vint16m4_t vmadd(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_i16m8))) +vint16m8_t vmadd(vint16m8_t op0, vint16m8_t op1, vint16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_i16m8_m))) +vint16m8_t vmadd(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, vint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_i16mf2))) +vint16mf2_t vmadd(vint16mf2_t op0, vint16mf2_t op1, vint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_i16mf2_m))) +vint16mf2_t vmadd(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_i16mf4))) +vint16mf4_t vmadd(vint16mf4_t op0, vint16mf4_t op1, vint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_i16mf4_m))) +vint16mf4_t vmadd(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_i32m1))) +vint32m1_t vmadd(vint32m1_t op0, vint32m1_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_i32m1_m))) +vint32m1_t vmadd(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_i32m2))) +vint32m2_t vmadd(vint32m2_t op0, vint32m2_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_i32m2_m))) +vint32m2_t vmadd(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_i32m4))) +vint32m4_t vmadd(vint32m4_t op0, vint32m4_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_i32m4_m))) +vint32m4_t vmadd(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_i32m8))) +vint32m8_t vmadd(vint32m8_t op0, vint32m8_t op1, vint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_i32m8_m))) +vint32m8_t vmadd(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_i32mf2))) +vint32mf2_t vmadd(vint32mf2_t op0, vint32mf2_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_i32mf2_m))) +vint32mf2_t vmadd(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_i64m1))) +vint64m1_t vmadd(vint64m1_t op0, vint64m1_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_i64m1_m))) +vint64m1_t vmadd(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_i64m2))) +vint64m2_t vmadd(vint64m2_t op0, vint64m2_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_i64m2_m))) +vint64m2_t vmadd(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_i64m4))) +vint64m4_t vmadd(vint64m4_t op0, vint64m4_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_i64m4_m))) +vint64m4_t vmadd(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_i64m8))) +vint64m8_t vmadd(vint64m8_t op0, vint64m8_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_i64m8_m))) +vint64m8_t vmadd(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_i8m1))) +vint8m1_t vmadd(vint8m1_t op0, int8_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_i8m1_m))) +vint8m1_t vmadd(vbool8_t op0, vint8m1_t op1, int8_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_i8m2))) +vint8m2_t vmadd(vint8m2_t op0, int8_t op1, vint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_i8m2_m))) +vint8m2_t vmadd(vbool4_t op0, vint8m2_t op1, int8_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_i8m4))) +vint8m4_t vmadd(vint8m4_t op0, int8_t op1, vint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_i8m4_m))) +vint8m4_t vmadd(vbool2_t op0, vint8m4_t op1, int8_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_i8m8))) +vint8m8_t vmadd(vint8m8_t op0, int8_t op1, vint8m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_i8m8_m))) +vint8m8_t vmadd(vbool1_t op0, vint8m8_t op1, int8_t op2, vint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_i8mf2))) +vint8mf2_t vmadd(vint8mf2_t op0, int8_t op1, vint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_i8mf2_m))) +vint8mf2_t vmadd(vbool16_t op0, vint8mf2_t op1, int8_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_i8mf4))) +vint8mf4_t vmadd(vint8mf4_t op0, int8_t op1, vint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_i8mf4_m))) +vint8mf4_t vmadd(vbool32_t op0, vint8mf4_t op1, int8_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_i8mf8))) +vint8mf8_t vmadd(vint8mf8_t op0, int8_t op1, vint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_i8mf8_m))) +vint8mf8_t vmadd(vbool64_t op0, vint8mf8_t op1, int8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_i16m1))) +vint16m1_t vmadd(vint16m1_t op0, int16_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_i16m1_m))) +vint16m1_t vmadd(vbool16_t op0, vint16m1_t op1, int16_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_i16m2))) +vint16m2_t vmadd(vint16m2_t op0, int16_t op1, vint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_i16m2_m))) +vint16m2_t vmadd(vbool8_t op0, vint16m2_t op1, int16_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_i16m4))) +vint16m4_t vmadd(vint16m4_t op0, int16_t op1, vint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_i16m4_m))) +vint16m4_t vmadd(vbool4_t op0, vint16m4_t op1, int16_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_i16m8))) +vint16m8_t vmadd(vint16m8_t op0, int16_t op1, vint16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_i16m8_m))) +vint16m8_t vmadd(vbool2_t op0, vint16m8_t op1, int16_t op2, vint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_i16mf2))) +vint16mf2_t vmadd(vint16mf2_t op0, int16_t op1, vint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_i16mf2_m))) +vint16mf2_t vmadd(vbool32_t op0, vint16mf2_t op1, int16_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_i16mf4))) +vint16mf4_t vmadd(vint16mf4_t op0, int16_t op1, vint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_i16mf4_m))) +vint16mf4_t vmadd(vbool64_t op0, vint16mf4_t op1, int16_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_i32m1))) +vint32m1_t vmadd(vint32m1_t op0, int32_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_i32m1_m))) +vint32m1_t vmadd(vbool32_t op0, vint32m1_t op1, int32_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_i32m2))) +vint32m2_t vmadd(vint32m2_t op0, int32_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_i32m2_m))) +vint32m2_t vmadd(vbool16_t op0, vint32m2_t op1, int32_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_i32m4))) +vint32m4_t vmadd(vint32m4_t op0, int32_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_i32m4_m))) +vint32m4_t vmadd(vbool8_t op0, vint32m4_t op1, int32_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_i32m8))) +vint32m8_t vmadd(vint32m8_t op0, int32_t op1, vint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_i32m8_m))) +vint32m8_t vmadd(vbool4_t op0, vint32m8_t op1, int32_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_i32mf2))) +vint32mf2_t vmadd(vint32mf2_t op0, int32_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_i32mf2_m))) +vint32mf2_t vmadd(vbool64_t op0, vint32mf2_t op1, int32_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_i64m1))) +vint64m1_t vmadd(vint64m1_t op0, int64_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_i64m1_m))) +vint64m1_t vmadd(vbool64_t op0, vint64m1_t op1, int64_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_i64m2))) +vint64m2_t vmadd(vint64m2_t op0, int64_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_i64m2_m))) +vint64m2_t vmadd(vbool32_t op0, vint64m2_t op1, int64_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_i64m4))) +vint64m4_t vmadd(vint64m4_t op0, int64_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_i64m4_m))) +vint64m4_t vmadd(vbool16_t op0, vint64m4_t op1, int64_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_i64m8))) +vint64m8_t vmadd(vint64m8_t op0, int64_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_i64m8_m))) +vint64m8_t vmadd(vbool8_t op0, vint64m8_t op1, int64_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_u8m1))) +vuint8m1_t vmadd(vuint8m1_t op0, vuint8m1_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_u8m1_m))) +vuint8m1_t vmadd(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_u8m2))) +vuint8m2_t vmadd(vuint8m2_t op0, vuint8m2_t op1, vuint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_u8m2_m))) +vuint8m2_t vmadd(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_u8m4))) +vuint8m4_t vmadd(vuint8m4_t op0, vuint8m4_t op1, vuint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_u8m4_m))) +vuint8m4_t vmadd(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_u8m8))) +vuint8m8_t vmadd(vuint8m8_t op0, vuint8m8_t op1, vuint8m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_u8m8_m))) +vuint8m8_t vmadd(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, vuint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_u8mf2))) +vuint8mf2_t vmadd(vuint8mf2_t op0, vuint8mf2_t op1, vuint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_u8mf2_m))) +vuint8mf2_t vmadd(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_u8mf4))) +vuint8mf4_t vmadd(vuint8mf4_t op0, vuint8mf4_t op1, vuint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_u8mf4_m))) +vuint8mf4_t vmadd(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_u8mf8))) +vuint8mf8_t vmadd(vuint8mf8_t op0, vuint8mf8_t op1, vuint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_u8mf8_m))) +vuint8mf8_t vmadd(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_u16m1))) +vuint16m1_t vmadd(vuint16m1_t op0, vuint16m1_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_u16m1_m))) +vuint16m1_t vmadd(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_u16m2))) +vuint16m2_t vmadd(vuint16m2_t op0, vuint16m2_t op1, vuint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_u16m2_m))) +vuint16m2_t vmadd(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_u16m4))) +vuint16m4_t vmadd(vuint16m4_t op0, vuint16m4_t op1, vuint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_u16m4_m))) +vuint16m4_t vmadd(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_u16m8))) +vuint16m8_t vmadd(vuint16m8_t op0, vuint16m8_t op1, vuint16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_u16m8_m))) +vuint16m8_t vmadd(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_u16mf2))) +vuint16mf2_t vmadd(vuint16mf2_t op0, vuint16mf2_t op1, vuint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_u16mf2_m))) +vuint16mf2_t vmadd(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_u16mf4))) +vuint16mf4_t vmadd(vuint16mf4_t op0, vuint16mf4_t op1, vuint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_u16mf4_m))) +vuint16mf4_t vmadd(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_u32m1))) +vuint32m1_t vmadd(vuint32m1_t op0, vuint32m1_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_u32m1_m))) +vuint32m1_t vmadd(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_u32m2))) +vuint32m2_t vmadd(vuint32m2_t op0, vuint32m2_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_u32m2_m))) +vuint32m2_t vmadd(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_u32m4))) +vuint32m4_t vmadd(vuint32m4_t op0, vuint32m4_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_u32m4_m))) +vuint32m4_t vmadd(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_u32m8))) +vuint32m8_t vmadd(vuint32m8_t op0, vuint32m8_t op1, vuint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_u32m8_m))) +vuint32m8_t vmadd(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_u32mf2))) +vuint32mf2_t vmadd(vuint32mf2_t op0, vuint32mf2_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_u32mf2_m))) +vuint32mf2_t vmadd(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_u64m1))) +vuint64m1_t vmadd(vuint64m1_t op0, vuint64m1_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_u64m1_m))) +vuint64m1_t vmadd(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_u64m2))) +vuint64m2_t vmadd(vuint64m2_t op0, vuint64m2_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_u64m2_m))) +vuint64m2_t vmadd(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_u64m4))) +vuint64m4_t vmadd(vuint64m4_t op0, vuint64m4_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_u64m4_m))) +vuint64m4_t vmadd(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_u64m8))) +vuint64m8_t vmadd(vuint64m8_t op0, vuint64m8_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vv_u64m8_m))) +vuint64m8_t vmadd(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_u8m1))) +vuint8m1_t vmadd(vuint8m1_t op0, uint8_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_u8m1_m))) +vuint8m1_t vmadd(vbool8_t op0, vuint8m1_t op1, uint8_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_u8m2))) +vuint8m2_t vmadd(vuint8m2_t op0, uint8_t op1, vuint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_u8m2_m))) +vuint8m2_t vmadd(vbool4_t op0, vuint8m2_t op1, uint8_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_u8m4))) +vuint8m4_t vmadd(vuint8m4_t op0, uint8_t op1, vuint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_u8m4_m))) +vuint8m4_t vmadd(vbool2_t op0, vuint8m4_t op1, uint8_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_u8m8))) +vuint8m8_t vmadd(vuint8m8_t op0, uint8_t op1, vuint8m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_u8m8_m))) +vuint8m8_t vmadd(vbool1_t op0, vuint8m8_t op1, uint8_t op2, vuint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_u8mf2))) +vuint8mf2_t vmadd(vuint8mf2_t op0, uint8_t op1, vuint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_u8mf2_m))) +vuint8mf2_t vmadd(vbool16_t op0, vuint8mf2_t op1, uint8_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_u8mf4))) +vuint8mf4_t vmadd(vuint8mf4_t op0, uint8_t op1, vuint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_u8mf4_m))) +vuint8mf4_t vmadd(vbool32_t op0, vuint8mf4_t op1, uint8_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_u8mf8))) +vuint8mf8_t vmadd(vuint8mf8_t op0, uint8_t op1, vuint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_u8mf8_m))) +vuint8mf8_t vmadd(vbool64_t op0, vuint8mf8_t op1, uint8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_u16m1))) +vuint16m1_t vmadd(vuint16m1_t op0, uint16_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_u16m1_m))) +vuint16m1_t vmadd(vbool16_t op0, vuint16m1_t op1, uint16_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_u16m2))) +vuint16m2_t vmadd(vuint16m2_t op0, uint16_t op1, vuint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_u16m2_m))) +vuint16m2_t vmadd(vbool8_t op0, vuint16m2_t op1, uint16_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_u16m4))) +vuint16m4_t vmadd(vuint16m4_t op0, uint16_t op1, vuint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_u16m4_m))) +vuint16m4_t vmadd(vbool4_t op0, vuint16m4_t op1, uint16_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_u16m8))) +vuint16m8_t vmadd(vuint16m8_t op0, uint16_t op1, vuint16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_u16m8_m))) +vuint16m8_t vmadd(vbool2_t op0, vuint16m8_t op1, uint16_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_u16mf2))) +vuint16mf2_t vmadd(vuint16mf2_t op0, uint16_t op1, vuint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_u16mf2_m))) +vuint16mf2_t vmadd(vbool32_t op0, vuint16mf2_t op1, uint16_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_u16mf4))) +vuint16mf4_t vmadd(vuint16mf4_t op0, uint16_t op1, vuint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_u16mf4_m))) +vuint16mf4_t vmadd(vbool64_t op0, vuint16mf4_t op1, uint16_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_u32m1))) +vuint32m1_t vmadd(vuint32m1_t op0, uint32_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_u32m1_m))) +vuint32m1_t vmadd(vbool32_t op0, vuint32m1_t op1, uint32_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_u32m2))) +vuint32m2_t vmadd(vuint32m2_t op0, uint32_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_u32m2_m))) +vuint32m2_t vmadd(vbool16_t op0, vuint32m2_t op1, uint32_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_u32m4))) +vuint32m4_t vmadd(vuint32m4_t op0, uint32_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_u32m4_m))) +vuint32m4_t vmadd(vbool8_t op0, vuint32m4_t op1, uint32_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_u32m8))) +vuint32m8_t vmadd(vuint32m8_t op0, uint32_t op1, vuint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_u32m8_m))) +vuint32m8_t vmadd(vbool4_t op0, vuint32m8_t op1, uint32_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_u32mf2))) +vuint32mf2_t vmadd(vuint32mf2_t op0, uint32_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_u32mf2_m))) +vuint32mf2_t vmadd(vbool64_t op0, vuint32mf2_t op1, uint32_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_u64m1))) +vuint64m1_t vmadd(vuint64m1_t op0, uint64_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_u64m1_m))) +vuint64m1_t vmadd(vbool64_t op0, vuint64m1_t op1, uint64_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_u64m2))) +vuint64m2_t vmadd(vuint64m2_t op0, uint64_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_u64m2_m))) +vuint64m2_t vmadd(vbool32_t op0, vuint64m2_t op1, uint64_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_u64m4))) +vuint64m4_t vmadd(vuint64m4_t op0, uint64_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_u64m4_m))) +vuint64m4_t vmadd(vbool16_t op0, vuint64m4_t op1, uint64_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_u64m8))) +vuint64m8_t vmadd(vuint64m8_t op0, uint64_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmadd_vx_u64m8_m))) +vuint64m8_t vmadd(vbool8_t op0, vuint64m8_t op1, uint64_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_i8m1))) +vint8m1_t vnmsub(vint8m1_t op0, vint8m1_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_i8m1_m))) +vint8m1_t vnmsub(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_i8m2))) +vint8m2_t vnmsub(vint8m2_t op0, vint8m2_t op1, vint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_i8m2_m))) +vint8m2_t vnmsub(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_i8m4))) +vint8m4_t vnmsub(vint8m4_t op0, vint8m4_t op1, vint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_i8m4_m))) +vint8m4_t vnmsub(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_i8m8))) +vint8m8_t vnmsub(vint8m8_t op0, vint8m8_t op1, vint8m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_i8m8_m))) +vint8m8_t vnmsub(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, vint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_i8mf2))) +vint8mf2_t vnmsub(vint8mf2_t op0, vint8mf2_t op1, vint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_i8mf2_m))) +vint8mf2_t vnmsub(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_i8mf4))) +vint8mf4_t vnmsub(vint8mf4_t op0, vint8mf4_t op1, vint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_i8mf4_m))) +vint8mf4_t vnmsub(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_i8mf8))) +vint8mf8_t vnmsub(vint8mf8_t op0, vint8mf8_t op1, vint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_i8mf8_m))) +vint8mf8_t vnmsub(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_i16m1))) +vint16m1_t vnmsub(vint16m1_t op0, vint16m1_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_i16m1_m))) +vint16m1_t vnmsub(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_i16m2))) +vint16m2_t vnmsub(vint16m2_t op0, vint16m2_t op1, vint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_i16m2_m))) +vint16m2_t vnmsub(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_i16m4))) +vint16m4_t vnmsub(vint16m4_t op0, vint16m4_t op1, vint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_i16m4_m))) +vint16m4_t vnmsub(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_i16m8))) +vint16m8_t vnmsub(vint16m8_t op0, vint16m8_t op1, vint16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_i16m8_m))) +vint16m8_t vnmsub(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, vint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_i16mf2))) +vint16mf2_t vnmsub(vint16mf2_t op0, vint16mf2_t op1, vint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_i16mf2_m))) +vint16mf2_t vnmsub(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_i16mf4))) +vint16mf4_t vnmsub(vint16mf4_t op0, vint16mf4_t op1, vint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_i16mf4_m))) +vint16mf4_t vnmsub(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_i32m1))) +vint32m1_t vnmsub(vint32m1_t op0, vint32m1_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_i32m1_m))) +vint32m1_t vnmsub(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_i32m2))) +vint32m2_t vnmsub(vint32m2_t op0, vint32m2_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_i32m2_m))) +vint32m2_t vnmsub(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_i32m4))) +vint32m4_t vnmsub(vint32m4_t op0, vint32m4_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_i32m4_m))) +vint32m4_t vnmsub(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_i32m8))) +vint32m8_t vnmsub(vint32m8_t op0, vint32m8_t op1, vint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_i32m8_m))) +vint32m8_t vnmsub(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_i32mf2))) +vint32mf2_t vnmsub(vint32mf2_t op0, vint32mf2_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_i32mf2_m))) +vint32mf2_t vnmsub(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_i64m1))) +vint64m1_t vnmsub(vint64m1_t op0, vint64m1_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_i64m1_m))) +vint64m1_t vnmsub(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_i64m2))) +vint64m2_t vnmsub(vint64m2_t op0, vint64m2_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_i64m2_m))) +vint64m2_t vnmsub(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_i64m4))) +vint64m4_t vnmsub(vint64m4_t op0, vint64m4_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_i64m4_m))) +vint64m4_t vnmsub(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_i64m8))) +vint64m8_t vnmsub(vint64m8_t op0, vint64m8_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_i64m8_m))) +vint64m8_t vnmsub(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_i8m1))) +vint8m1_t vnmsub(vint8m1_t op0, int8_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_i8m1_m))) +vint8m1_t vnmsub(vbool8_t op0, vint8m1_t op1, int8_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_i8m2))) +vint8m2_t vnmsub(vint8m2_t op0, int8_t op1, vint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_i8m2_m))) +vint8m2_t vnmsub(vbool4_t op0, vint8m2_t op1, int8_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_i8m4))) +vint8m4_t vnmsub(vint8m4_t op0, int8_t op1, vint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_i8m4_m))) +vint8m4_t vnmsub(vbool2_t op0, vint8m4_t op1, int8_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_i8m8))) +vint8m8_t vnmsub(vint8m8_t op0, int8_t op1, vint8m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_i8m8_m))) +vint8m8_t vnmsub(vbool1_t op0, vint8m8_t op1, int8_t op2, vint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_i8mf2))) +vint8mf2_t vnmsub(vint8mf2_t op0, int8_t op1, vint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_i8mf2_m))) +vint8mf2_t vnmsub(vbool16_t op0, vint8mf2_t op1, int8_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_i8mf4))) +vint8mf4_t vnmsub(vint8mf4_t op0, int8_t op1, vint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_i8mf4_m))) +vint8mf4_t vnmsub(vbool32_t op0, vint8mf4_t op1, int8_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_i8mf8))) +vint8mf8_t vnmsub(vint8mf8_t op0, int8_t op1, vint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_i8mf8_m))) +vint8mf8_t vnmsub(vbool64_t op0, vint8mf8_t op1, int8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_i16m1))) +vint16m1_t vnmsub(vint16m1_t op0, int16_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_i16m1_m))) +vint16m1_t vnmsub(vbool16_t op0, vint16m1_t op1, int16_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_i16m2))) +vint16m2_t vnmsub(vint16m2_t op0, int16_t op1, vint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_i16m2_m))) +vint16m2_t vnmsub(vbool8_t op0, vint16m2_t op1, int16_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_i16m4))) +vint16m4_t vnmsub(vint16m4_t op0, int16_t op1, vint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_i16m4_m))) +vint16m4_t vnmsub(vbool4_t op0, vint16m4_t op1, int16_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_i16m8))) +vint16m8_t vnmsub(vint16m8_t op0, int16_t op1, vint16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_i16m8_m))) +vint16m8_t vnmsub(vbool2_t op0, vint16m8_t op1, int16_t op2, vint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_i16mf2))) +vint16mf2_t vnmsub(vint16mf2_t op0, int16_t op1, vint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_i16mf2_m))) +vint16mf2_t vnmsub(vbool32_t op0, vint16mf2_t op1, int16_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_i16mf4))) +vint16mf4_t vnmsub(vint16mf4_t op0, int16_t op1, vint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_i16mf4_m))) +vint16mf4_t vnmsub(vbool64_t op0, vint16mf4_t op1, int16_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_i32m1))) +vint32m1_t vnmsub(vint32m1_t op0, int32_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_i32m1_m))) +vint32m1_t vnmsub(vbool32_t op0, vint32m1_t op1, int32_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_i32m2))) +vint32m2_t vnmsub(vint32m2_t op0, int32_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_i32m2_m))) +vint32m2_t vnmsub(vbool16_t op0, vint32m2_t op1, int32_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_i32m4))) +vint32m4_t vnmsub(vint32m4_t op0, int32_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_i32m4_m))) +vint32m4_t vnmsub(vbool8_t op0, vint32m4_t op1, int32_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_i32m8))) +vint32m8_t vnmsub(vint32m8_t op0, int32_t op1, vint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_i32m8_m))) +vint32m8_t vnmsub(vbool4_t op0, vint32m8_t op1, int32_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_i32mf2))) +vint32mf2_t vnmsub(vint32mf2_t op0, int32_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_i32mf2_m))) +vint32mf2_t vnmsub(vbool64_t op0, vint32mf2_t op1, int32_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_i64m1))) +vint64m1_t vnmsub(vint64m1_t op0, int64_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_i64m1_m))) +vint64m1_t vnmsub(vbool64_t op0, vint64m1_t op1, int64_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_i64m2))) +vint64m2_t vnmsub(vint64m2_t op0, int64_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_i64m2_m))) +vint64m2_t vnmsub(vbool32_t op0, vint64m2_t op1, int64_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_i64m4))) +vint64m4_t vnmsub(vint64m4_t op0, int64_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_i64m4_m))) +vint64m4_t vnmsub(vbool16_t op0, vint64m4_t op1, int64_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_i64m8))) +vint64m8_t vnmsub(vint64m8_t op0, int64_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_i64m8_m))) +vint64m8_t vnmsub(vbool8_t op0, vint64m8_t op1, int64_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_u8m1))) +vuint8m1_t vnmsub(vuint8m1_t op0, vuint8m1_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_u8m1_m))) +vuint8m1_t vnmsub(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_u8m2))) +vuint8m2_t vnmsub(vuint8m2_t op0, vuint8m2_t op1, vuint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_u8m2_m))) +vuint8m2_t vnmsub(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_u8m4))) +vuint8m4_t vnmsub(vuint8m4_t op0, vuint8m4_t op1, vuint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_u8m4_m))) +vuint8m4_t vnmsub(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_u8m8))) +vuint8m8_t vnmsub(vuint8m8_t op0, vuint8m8_t op1, vuint8m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_u8m8_m))) +vuint8m8_t vnmsub(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, vuint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_u8mf2))) +vuint8mf2_t vnmsub(vuint8mf2_t op0, vuint8mf2_t op1, vuint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_u8mf2_m))) +vuint8mf2_t vnmsub(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_u8mf4))) +vuint8mf4_t vnmsub(vuint8mf4_t op0, vuint8mf4_t op1, vuint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_u8mf4_m))) +vuint8mf4_t vnmsub(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_u8mf8))) +vuint8mf8_t vnmsub(vuint8mf8_t op0, vuint8mf8_t op1, vuint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_u8mf8_m))) +vuint8mf8_t vnmsub(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_u16m1))) +vuint16m1_t vnmsub(vuint16m1_t op0, vuint16m1_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_u16m1_m))) +vuint16m1_t vnmsub(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_u16m2))) +vuint16m2_t vnmsub(vuint16m2_t op0, vuint16m2_t op1, vuint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_u16m2_m))) +vuint16m2_t vnmsub(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_u16m4))) +vuint16m4_t vnmsub(vuint16m4_t op0, vuint16m4_t op1, vuint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_u16m4_m))) +vuint16m4_t vnmsub(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_u16m8))) +vuint16m8_t vnmsub(vuint16m8_t op0, vuint16m8_t op1, vuint16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_u16m8_m))) +vuint16m8_t vnmsub(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_u16mf2))) +vuint16mf2_t vnmsub(vuint16mf2_t op0, vuint16mf2_t op1, vuint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_u16mf2_m))) +vuint16mf2_t vnmsub(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_u16mf4))) +vuint16mf4_t vnmsub(vuint16mf4_t op0, vuint16mf4_t op1, vuint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_u16mf4_m))) +vuint16mf4_t vnmsub(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_u32m1))) +vuint32m1_t vnmsub(vuint32m1_t op0, vuint32m1_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_u32m1_m))) +vuint32m1_t vnmsub(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_u32m2))) +vuint32m2_t vnmsub(vuint32m2_t op0, vuint32m2_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_u32m2_m))) +vuint32m2_t vnmsub(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_u32m4))) +vuint32m4_t vnmsub(vuint32m4_t op0, vuint32m4_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_u32m4_m))) +vuint32m4_t vnmsub(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_u32m8))) +vuint32m8_t vnmsub(vuint32m8_t op0, vuint32m8_t op1, vuint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_u32m8_m))) +vuint32m8_t vnmsub(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_u32mf2))) +vuint32mf2_t vnmsub(vuint32mf2_t op0, vuint32mf2_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_u32mf2_m))) +vuint32mf2_t vnmsub(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_u64m1))) +vuint64m1_t vnmsub(vuint64m1_t op0, vuint64m1_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_u64m1_m))) +vuint64m1_t vnmsub(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_u64m2))) +vuint64m2_t vnmsub(vuint64m2_t op0, vuint64m2_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_u64m2_m))) +vuint64m2_t vnmsub(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_u64m4))) +vuint64m4_t vnmsub(vuint64m4_t op0, vuint64m4_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_u64m4_m))) +vuint64m4_t vnmsub(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_u64m8))) +vuint64m8_t vnmsub(vuint64m8_t op0, vuint64m8_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vv_u64m8_m))) +vuint64m8_t vnmsub(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_u32m1))) +void vsuxei16(uint32_t * op0, vuint16mf2_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_u32m1_m))) +void vsuxei16(vbool32_t op0, uint32_t * op1, vuint16mf2_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_u32m2))) +void vsuxei16(uint32_t * op0, vuint16m1_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_u32m2_m))) +void vsuxei16(vbool16_t op0, uint32_t * op1, vuint16m1_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_u32m4))) +void vsuxei16(uint32_t * op0, vuint16m2_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_u32m4_m))) +void vsuxei16(vbool8_t op0, uint32_t * op1, vuint16m2_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_u32m8))) +void vsuxei16(uint32_t * op0, vuint16m4_t op1, vuint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_u32m8_m))) +void vsuxei16(vbool4_t op0, uint32_t * op1, vuint16m4_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_u32mf2))) +void vsuxei16(uint32_t * op0, vuint16mf4_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_u32mf2_m))) +void vsuxei16(vbool64_t op0, uint32_t * op1, vuint16mf4_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_u8m1))) +vuint8m1_t vnmsub(vuint8m1_t op0, uint8_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_u8m1_m))) +vuint8m1_t vnmsub(vbool8_t op0, vuint8m1_t op1, uint8_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_u8m2))) +vuint8m2_t vnmsub(vuint8m2_t op0, uint8_t op1, vuint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_u8m2_m))) +vuint8m2_t vnmsub(vbool4_t op0, vuint8m2_t op1, uint8_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_u8m4))) +vuint8m4_t vnmsub(vuint8m4_t op0, uint8_t op1, vuint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_u8m4_m))) +vuint8m4_t vnmsub(vbool2_t op0, vuint8m4_t op1, uint8_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_u8m8))) +vuint8m8_t vnmsub(vuint8m8_t op0, uint8_t op1, vuint8m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_u8m8_m))) +vuint8m8_t vnmsub(vbool1_t op0, vuint8m8_t op1, uint8_t op2, vuint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_u8mf2))) +vuint8mf2_t vnmsub(vuint8mf2_t op0, uint8_t op1, vuint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_u8mf2_m))) +vuint8mf2_t vnmsub(vbool16_t op0, vuint8mf2_t op1, uint8_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_u8mf4))) +vuint8mf4_t vnmsub(vuint8mf4_t op0, uint8_t op1, vuint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_u8mf4_m))) +vuint8mf4_t vnmsub(vbool32_t op0, vuint8mf4_t op1, uint8_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_u8mf8))) +vuint8mf8_t vnmsub(vuint8mf8_t op0, uint8_t op1, vuint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_u8mf8_m))) +vuint8mf8_t vnmsub(vbool64_t op0, vuint8mf8_t op1, uint8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_u16m1))) +vuint16m1_t vnmsub(vuint16m1_t op0, uint16_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_u16m1_m))) +vuint16m1_t vnmsub(vbool16_t op0, vuint16m1_t op1, uint16_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_u16m2))) +vuint16m2_t vnmsub(vuint16m2_t op0, uint16_t op1, vuint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_u16m2_m))) +vuint16m2_t vnmsub(vbool8_t op0, vuint16m2_t op1, uint16_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_u16m4))) +vuint16m4_t vnmsub(vuint16m4_t op0, uint16_t op1, vuint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_u16m4_m))) +vuint16m4_t vnmsub(vbool4_t op0, vuint16m4_t op1, uint16_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_u16m8))) +vuint16m8_t vnmsub(vuint16m8_t op0, uint16_t op1, vuint16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_u16m8_m))) +vuint16m8_t vnmsub(vbool2_t op0, vuint16m8_t op1, uint16_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_u16mf2))) +vuint16mf2_t vnmsub(vuint16mf2_t op0, uint16_t op1, vuint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_u16mf2_m))) +vuint16mf2_t vnmsub(vbool32_t op0, vuint16mf2_t op1, uint16_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_u16mf4))) +vuint16mf4_t vnmsub(vuint16mf4_t op0, uint16_t op1, vuint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_u16mf4_m))) +vuint16mf4_t vnmsub(vbool64_t op0, vuint16mf4_t op1, uint16_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_u32m1))) +vuint32m1_t vnmsub(vuint32m1_t op0, uint32_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_u32m1_m))) +vuint32m1_t vnmsub(vbool32_t op0, vuint32m1_t op1, uint32_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_u32m2))) +vuint32m2_t vnmsub(vuint32m2_t op0, uint32_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_u32m2_m))) +vuint32m2_t vnmsub(vbool16_t op0, vuint32m2_t op1, uint32_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_u32m4))) +vuint32m4_t vnmsub(vuint32m4_t op0, uint32_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_u32m4_m))) +vuint32m4_t vnmsub(vbool8_t op0, vuint32m4_t op1, uint32_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_u32m8))) +vuint32m8_t vnmsub(vuint32m8_t op0, uint32_t op1, vuint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_u32m8_m))) +vuint32m8_t vnmsub(vbool4_t op0, vuint32m8_t op1, uint32_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_u32mf2))) +vuint32mf2_t vnmsub(vuint32mf2_t op0, uint32_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_u32mf2_m))) +vuint32mf2_t vnmsub(vbool64_t op0, vuint32mf2_t op1, uint32_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_u64m1))) +vuint64m1_t vnmsub(vuint64m1_t op0, uint64_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_u64m1_m))) +vuint64m1_t vnmsub(vbool64_t op0, vuint64m1_t op1, uint64_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_u64m2))) +vuint64m2_t vnmsub(vuint64m2_t op0, uint64_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_u64m2_m))) +vuint64m2_t vnmsub(vbool32_t op0, vuint64m2_t op1, uint64_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_u64m4))) +vuint64m4_t vnmsub(vuint64m4_t op0, uint64_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_u64m4_m))) +vuint64m4_t vnmsub(vbool16_t op0, vuint64m4_t op1, uint64_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_u64m8))) +vuint64m8_t vnmsub(vuint64m8_t op0, uint64_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnmsub_vx_u64m8_m))) +vuint64m8_t vnmsub(vbool8_t op0, vuint64m8_t op1, uint64_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vv_u16mf4))) +vuint16mf4_t vwmaccu(vuint16mf4_t op0, vuint8mf8_t op1, vuint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vv_u16mf4_m))) +vuint16mf4_t vwmaccu(vbool64_t op0, vuint16mf4_t op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vv_u16mf2))) +vuint16mf2_t vwmaccu(vuint16mf2_t op0, vuint8mf4_t op1, vuint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vv_u16mf2_m))) +vuint16mf2_t vwmaccu(vbool32_t op0, vuint16mf2_t op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vv_u16m1))) +vuint16m1_t vwmaccu(vuint16m1_t op0, vuint8mf2_t op1, vuint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vv_u16m1_m))) +vuint16m1_t vwmaccu(vbool16_t op0, vuint16m1_t op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vv_u16m2))) +vuint16m2_t vwmaccu(vuint16m2_t op0, vuint8m1_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vv_u16m2_m))) +vuint16m2_t vwmaccu(vbool8_t op0, vuint16m2_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vv_u16m4))) +vuint16m4_t vwmaccu(vuint16m4_t op0, vuint8m2_t op1, vuint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vv_u16m4_m))) +vuint16m4_t vwmaccu(vbool4_t op0, vuint16m4_t op1, vuint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vv_u16m8))) +vuint16m8_t vwmaccu(vuint16m8_t op0, vuint8m4_t op1, vuint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vv_u16m8_m))) +vuint16m8_t vwmaccu(vbool2_t op0, vuint16m8_t op1, vuint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vv_u32mf2))) +vuint32mf2_t vwmaccu(vuint32mf2_t op0, vuint16mf4_t op1, vuint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vv_u32mf2_m))) +vuint32mf2_t vwmaccu(vbool64_t op0, vuint32mf2_t op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vv_u32m1))) +vuint32m1_t vwmaccu(vuint32m1_t op0, vuint16mf2_t op1, vuint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vv_u32m1_m))) +vuint32m1_t vwmaccu(vbool32_t op0, vuint32m1_t op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vv_u32m2))) +vuint32m2_t vwmaccu(vuint32m2_t op0, vuint16m1_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vv_u32m2_m))) +vuint32m2_t vwmaccu(vbool16_t op0, vuint32m2_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vv_u32m4))) +vuint32m4_t vwmaccu(vuint32m4_t op0, vuint16m2_t op1, vuint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vv_u32m4_m))) +vuint32m4_t vwmaccu(vbool8_t op0, vuint32m4_t op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vv_u32m8))) +vuint32m8_t vwmaccu(vuint32m8_t op0, vuint16m4_t op1, vuint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vv_u32m8_m))) +vuint32m8_t vwmaccu(vbool4_t op0, vuint32m8_t op1, vuint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vv_u64m1))) +vuint64m1_t vwmaccu(vuint64m1_t op0, vuint32mf2_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vv_u64m1_m))) +vuint64m1_t vwmaccu(vbool64_t op0, vuint64m1_t op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vv_u64m2))) +vuint64m2_t vwmaccu(vuint64m2_t op0, vuint32m1_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vv_u64m2_m))) +vuint64m2_t vwmaccu(vbool32_t op0, vuint64m2_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vv_u64m4))) +vuint64m4_t vwmaccu(vuint64m4_t op0, vuint32m2_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vv_u64m4_m))) +vuint64m4_t vwmaccu(vbool16_t op0, vuint64m4_t op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vv_u64m8))) +vuint64m8_t vwmaccu(vuint64m8_t op0, vuint32m4_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vv_u64m8_m))) +vuint64m8_t vwmaccu(vbool8_t op0, vuint64m8_t op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vx_u16mf4))) +vuint16mf4_t vwmaccu(vuint16mf4_t op0, uint8_t op1, vuint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vx_u16mf4_m))) +vuint16mf4_t vwmaccu(vbool64_t op0, vuint16mf4_t op1, uint8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vx_u16mf2))) +vuint16mf2_t vwmaccu(vuint16mf2_t op0, uint8_t op1, vuint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vx_u16mf2_m))) +vuint16mf2_t vwmaccu(vbool32_t op0, vuint16mf2_t op1, uint8_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vx_u16m1))) +vuint16m1_t vwmaccu(vuint16m1_t op0, uint8_t op1, vuint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vx_u16m1_m))) +vuint16m1_t vwmaccu(vbool16_t op0, vuint16m1_t op1, uint8_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vx_u16m2))) +vuint16m2_t vwmaccu(vuint16m2_t op0, uint8_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vx_u16m2_m))) +vuint16m2_t vwmaccu(vbool8_t op0, vuint16m2_t op1, uint8_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vx_u16m4))) +vuint16m4_t vwmaccu(vuint16m4_t op0, uint8_t op1, vuint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vx_u16m4_m))) +vuint16m4_t vwmaccu(vbool4_t op0, vuint16m4_t op1, uint8_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vx_u16m8))) +vuint16m8_t vwmaccu(vuint16m8_t op0, uint8_t op1, vuint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vx_u16m8_m))) +vuint16m8_t vwmaccu(vbool2_t op0, vuint16m8_t op1, uint8_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vx_u32mf2))) +vuint32mf2_t vwmaccu(vuint32mf2_t op0, uint16_t op1, vuint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vx_u32mf2_m))) +vuint32mf2_t vwmaccu(vbool64_t op0, vuint32mf2_t op1, uint16_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vx_u32m1))) +vuint32m1_t vwmaccu(vuint32m1_t op0, uint16_t op1, vuint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vx_u32m1_m))) +vuint32m1_t vwmaccu(vbool32_t op0, vuint32m1_t op1, uint16_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vx_u32m2))) +vuint32m2_t vwmaccu(vuint32m2_t op0, uint16_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vx_u32m2_m))) +vuint32m2_t vwmaccu(vbool16_t op0, vuint32m2_t op1, uint16_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vx_u32m4))) +vuint32m4_t vwmaccu(vuint32m4_t op0, uint16_t op1, vuint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vx_u32m4_m))) +vuint32m4_t vwmaccu(vbool8_t op0, vuint32m4_t op1, uint16_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vx_u32m8))) +vuint32m8_t vwmaccu(vuint32m8_t op0, uint16_t op1, vuint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vx_u32m8_m))) +vuint32m8_t vwmaccu(vbool4_t op0, vuint32m8_t op1, uint16_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vx_u64m1))) +vuint64m1_t vwmaccu(vuint64m1_t op0, uint32_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vx_u64m1_m))) +vuint64m1_t vwmaccu(vbool64_t op0, vuint64m1_t op1, uint32_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vx_u64m2))) +vuint64m2_t vwmaccu(vuint64m2_t op0, uint32_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vx_u64m2_m))) +vuint64m2_t vwmaccu(vbool32_t op0, vuint64m2_t op1, uint32_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vx_u64m4))) +vuint64m4_t vwmaccu(vuint64m4_t op0, uint32_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vx_u64m4_m))) +vuint64m4_t vwmaccu(vbool16_t op0, vuint64m4_t op1, uint32_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vx_u64m8))) +vuint64m8_t vwmaccu(vuint64m8_t op0, uint32_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccu_vx_u64m8_m))) +vuint64m8_t vwmaccu(vbool8_t op0, vuint64m8_t op1, uint32_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vv_i16mf4))) +vint16mf4_t vwmacc(vint16mf4_t op0, vint8mf8_t op1, vint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vv_i16mf4_m))) +vint16mf4_t vwmacc(vbool64_t op0, vint16mf4_t op1, vint8mf8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vv_i16mf2))) +vint16mf2_t vwmacc(vint16mf2_t op0, vint8mf4_t op1, vint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vv_i16mf2_m))) +vint16mf2_t vwmacc(vbool32_t op0, vint16mf2_t op1, vint8mf4_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vv_i16m1))) +vint16m1_t vwmacc(vint16m1_t op0, vint8mf2_t op1, vint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vv_i16m1_m))) +vint16m1_t vwmacc(vbool16_t op0, vint16m1_t op1, vint8mf2_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vv_i16m2))) +vint16m2_t vwmacc(vint16m2_t op0, vint8m1_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vv_i16m2_m))) +vint16m2_t vwmacc(vbool8_t op0, vint16m2_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vv_i16m4))) +vint16m4_t vwmacc(vint16m4_t op0, vint8m2_t op1, vint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vv_i16m4_m))) +vint16m4_t vwmacc(vbool4_t op0, vint16m4_t op1, vint8m2_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vv_i16m8))) +vint16m8_t vwmacc(vint16m8_t op0, vint8m4_t op1, vint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vv_i16m8_m))) +vint16m8_t vwmacc(vbool2_t op0, vint16m8_t op1, vint8m4_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vv_i32mf2))) +vint32mf2_t vwmacc(vint32mf2_t op0, vint16mf4_t op1, vint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vv_i32mf2_m))) +vint32mf2_t vwmacc(vbool64_t op0, vint32mf2_t op1, vint16mf4_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vv_i32m1))) +vint32m1_t vwmacc(vint32m1_t op0, vint16mf2_t op1, vint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vv_i32m1_m))) +vint32m1_t vwmacc(vbool32_t op0, vint32m1_t op1, vint16mf2_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vv_i32m2))) +vint32m2_t vwmacc(vint32m2_t op0, vint16m1_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vv_i32m2_m))) +vint32m2_t vwmacc(vbool16_t op0, vint32m2_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vv_i32m4))) +vint32m4_t vwmacc(vint32m4_t op0, vint16m2_t op1, vint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vv_i32m4_m))) +vint32m4_t vwmacc(vbool8_t op0, vint32m4_t op1, vint16m2_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vv_i32m8))) +vint32m8_t vwmacc(vint32m8_t op0, vint16m4_t op1, vint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vv_i32m8_m))) +vint32m8_t vwmacc(vbool4_t op0, vint32m8_t op1, vint16m4_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vv_i64m1))) +vint64m1_t vwmacc(vint64m1_t op0, vint32mf2_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vv_i64m1_m))) +vint64m1_t vwmacc(vbool64_t op0, vint64m1_t op1, vint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vv_i64m2))) +vint64m2_t vwmacc(vint64m2_t op0, vint32m1_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vv_i64m2_m))) +vint64m2_t vwmacc(vbool32_t op0, vint64m2_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vv_i64m4))) +vint64m4_t vwmacc(vint64m4_t op0, vint32m2_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vv_i64m4_m))) +vint64m4_t vwmacc(vbool16_t op0, vint64m4_t op1, vint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vv_i64m8))) +vint64m8_t vwmacc(vint64m8_t op0, vint32m4_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vv_i64m8_m))) +vint64m8_t vwmacc(vbool8_t op0, vint64m8_t op1, vint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vx_i16mf4))) +vint16mf4_t vwmacc(vint16mf4_t op0, int8_t op1, vint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vx_i16mf4_m))) +vint16mf4_t vwmacc(vbool64_t op0, vint16mf4_t op1, int8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vx_i16mf2))) +vint16mf2_t vwmacc(vint16mf2_t op0, int8_t op1, vint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vx_i16mf2_m))) +vint16mf2_t vwmacc(vbool32_t op0, vint16mf2_t op1, int8_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vx_i16m1))) +vint16m1_t vwmacc(vint16m1_t op0, int8_t op1, vint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vx_i16m1_m))) +vint16m1_t vwmacc(vbool16_t op0, vint16m1_t op1, int8_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vx_i16m2))) +vint16m2_t vwmacc(vint16m2_t op0, int8_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vx_i16m2_m))) +vint16m2_t vwmacc(vbool8_t op0, vint16m2_t op1, int8_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vx_i16m4))) +vint16m4_t vwmacc(vint16m4_t op0, int8_t op1, vint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vx_i16m4_m))) +vint16m4_t vwmacc(vbool4_t op0, vint16m4_t op1, int8_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vx_i16m8))) +vint16m8_t vwmacc(vint16m8_t op0, int8_t op1, vint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vx_i16m8_m))) +vint16m8_t vwmacc(vbool2_t op0, vint16m8_t op1, int8_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vx_i32mf2))) +vint32mf2_t vwmacc(vint32mf2_t op0, int16_t op1, vint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vx_i32mf2_m))) +vint32mf2_t vwmacc(vbool64_t op0, vint32mf2_t op1, int16_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vx_i32m1))) +vint32m1_t vwmacc(vint32m1_t op0, int16_t op1, vint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vx_i32m1_m))) +vint32m1_t vwmacc(vbool32_t op0, vint32m1_t op1, int16_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vx_i32m2))) +vint32m2_t vwmacc(vint32m2_t op0, int16_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vx_i32m2_m))) +vint32m2_t vwmacc(vbool16_t op0, vint32m2_t op1, int16_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vx_i32m4))) +vint32m4_t vwmacc(vint32m4_t op0, int16_t op1, vint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vx_i32m4_m))) +vint32m4_t vwmacc(vbool8_t op0, vint32m4_t op1, int16_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vx_i32m8))) +vint32m8_t vwmacc(vint32m8_t op0, int16_t op1, vint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vx_i32m8_m))) +vint32m8_t vwmacc(vbool4_t op0, vint32m8_t op1, int16_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vx_i64m1))) +vint64m1_t vwmacc(vint64m1_t op0, int32_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vx_i64m1_m))) +vint64m1_t vwmacc(vbool64_t op0, vint64m1_t op1, int32_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vx_i64m2))) +vint64m2_t vwmacc(vint64m2_t op0, int32_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vx_i64m2_m))) +vint64m2_t vwmacc(vbool32_t op0, vint64m2_t op1, int32_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vx_i64m4))) +vint64m4_t vwmacc(vint64m4_t op0, int32_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vx_i64m4_m))) +vint64m4_t vwmacc(vbool16_t op0, vint64m4_t op1, int32_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vx_i64m8))) +vint64m8_t vwmacc(vint64m8_t op0, int32_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmacc_vx_i64m8_m))) +vint64m8_t vwmacc(vbool8_t op0, vint64m8_t op1, int32_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vv_i16mf4))) +vint16mf4_t vwmaccsu(vint16mf4_t op0, vint8mf8_t op1, vuint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vv_i16mf4_m))) +vint16mf4_t vwmaccsu(vbool64_t op0, vint16mf4_t op1, vint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vv_i16mf2))) +vint16mf2_t vwmaccsu(vint16mf2_t op0, vint8mf4_t op1, vuint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vv_i16mf2_m))) +vint16mf2_t vwmaccsu(vbool32_t op0, vint16mf2_t op1, vint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vv_i16m1))) +vint16m1_t vwmaccsu(vint16m1_t op0, vint8mf2_t op1, vuint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vv_i16m1_m))) +vint16m1_t vwmaccsu(vbool16_t op0, vint16m1_t op1, vint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vv_i16m2))) +vint16m2_t vwmaccsu(vint16m2_t op0, vint8m1_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vv_i16m2_m))) +vint16m2_t vwmaccsu(vbool8_t op0, vint16m2_t op1, vint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vv_i16m4))) +vint16m4_t vwmaccsu(vint16m4_t op0, vint8m2_t op1, vuint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vv_i16m4_m))) +vint16m4_t vwmaccsu(vbool4_t op0, vint16m4_t op1, vint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vv_i16m8))) +vint16m8_t vwmaccsu(vint16m8_t op0, vint8m4_t op1, vuint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vv_i16m8_m))) +vint16m8_t vwmaccsu(vbool2_t op0, vint16m8_t op1, vint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vv_i32mf2))) +vint32mf2_t vwmaccsu(vint32mf2_t op0, vint16mf4_t op1, vuint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vv_i32mf2_m))) +vint32mf2_t vwmaccsu(vbool64_t op0, vint32mf2_t op1, vint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vv_i32m1))) +vint32m1_t vwmaccsu(vint32m1_t op0, vint16mf2_t op1, vuint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vv_i32m1_m))) +vint32m1_t vwmaccsu(vbool32_t op0, vint32m1_t op1, vint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vv_i32m2))) +vint32m2_t vwmaccsu(vint32m2_t op0, vint16m1_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vv_i32m2_m))) +vint32m2_t vwmaccsu(vbool16_t op0, vint32m2_t op1, vint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vv_i32m4))) +vint32m4_t vwmaccsu(vint32m4_t op0, vint16m2_t op1, vuint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vv_i32m4_m))) +vint32m4_t vwmaccsu(vbool8_t op0, vint32m4_t op1, vint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vv_i32m8))) +vint32m8_t vwmaccsu(vint32m8_t op0, vint16m4_t op1, vuint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vv_i32m8_m))) +vint32m8_t vwmaccsu(vbool4_t op0, vint32m8_t op1, vint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vv_i64m1))) +vint64m1_t vwmaccsu(vint64m1_t op0, vint32mf2_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vv_i64m1_m))) +vint64m1_t vwmaccsu(vbool64_t op0, vint64m1_t op1, vint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vv_i64m2))) +vint64m2_t vwmaccsu(vint64m2_t op0, vint32m1_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vv_i64m2_m))) +vint64m2_t vwmaccsu(vbool32_t op0, vint64m2_t op1, vint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vv_i64m4))) +vint64m4_t vwmaccsu(vint64m4_t op0, vint32m2_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vv_i64m4_m))) +vint64m4_t vwmaccsu(vbool16_t op0, vint64m4_t op1, vint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vv_i64m8))) +vint64m8_t vwmaccsu(vint64m8_t op0, vint32m4_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vv_i64m8_m))) +vint64m8_t vwmaccsu(vbool8_t op0, vint64m8_t op1, vint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vx_i16mf4))) +vint16mf4_t vwmaccsu(vint16mf4_t op0, int8_t op1, vuint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vx_i16mf4_m))) +vint16mf4_t vwmaccsu(vbool64_t op0, vint16mf4_t op1, int8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vx_i16mf2))) +vint16mf2_t vwmaccsu(vint16mf2_t op0, int8_t op1, vuint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vx_i16mf2_m))) +vint16mf2_t vwmaccsu(vbool32_t op0, vint16mf2_t op1, int8_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vx_i16m1))) +vint16m1_t vwmaccsu(vint16m1_t op0, int8_t op1, vuint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vx_i16m1_m))) +vint16m1_t vwmaccsu(vbool16_t op0, vint16m1_t op1, int8_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vx_i16m2))) +vint16m2_t vwmaccsu(vint16m2_t op0, int8_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vx_i16m2_m))) +vint16m2_t vwmaccsu(vbool8_t op0, vint16m2_t op1, int8_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vx_i16m4))) +vint16m4_t vwmaccsu(vint16m4_t op0, int8_t op1, vuint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vx_i16m4_m))) +vint16m4_t vwmaccsu(vbool4_t op0, vint16m4_t op1, int8_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vx_i16m8))) +vint16m8_t vwmaccsu(vint16m8_t op0, int8_t op1, vuint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vx_i16m8_m))) +vint16m8_t vwmaccsu(vbool2_t op0, vint16m8_t op1, int8_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vx_i32mf2))) +vint32mf2_t vwmaccsu(vint32mf2_t op0, int16_t op1, vuint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vx_i32mf2_m))) +vint32mf2_t vwmaccsu(vbool64_t op0, vint32mf2_t op1, int16_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vx_i32m1))) +vint32m1_t vwmaccsu(vint32m1_t op0, int16_t op1, vuint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vx_i32m1_m))) +vint32m1_t vwmaccsu(vbool32_t op0, vint32m1_t op1, int16_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vx_i32m2))) +vint32m2_t vwmaccsu(vint32m2_t op0, int16_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vx_i32m2_m))) +vint32m2_t vwmaccsu(vbool16_t op0, vint32m2_t op1, int16_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vx_i32m4))) +vint32m4_t vwmaccsu(vint32m4_t op0, int16_t op1, vuint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vx_i32m4_m))) +vint32m4_t vwmaccsu(vbool8_t op0, vint32m4_t op1, int16_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vx_i32m8))) +vint32m8_t vwmaccsu(vint32m8_t op0, int16_t op1, vuint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vx_i32m8_m))) +vint32m8_t vwmaccsu(vbool4_t op0, vint32m8_t op1, int16_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vx_i64m1))) +vint64m1_t vwmaccsu(vint64m1_t op0, int32_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vx_i64m1_m))) +vint64m1_t vwmaccsu(vbool64_t op0, vint64m1_t op1, int32_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vx_i64m2))) +vint64m2_t vwmaccsu(vint64m2_t op0, int32_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vx_i64m2_m))) +vint64m2_t vwmaccsu(vbool32_t op0, vint64m2_t op1, int32_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vx_i64m4))) +vint64m4_t vwmaccsu(vint64m4_t op0, int32_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vx_i64m4_m))) +vint64m4_t vwmaccsu(vbool16_t op0, vint64m4_t op1, int32_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vx_i64m8))) +vint64m8_t vwmaccsu(vint64m8_t op0, int32_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccsu_vx_i64m8_m))) +vint64m8_t vwmaccsu(vbool8_t op0, vint64m8_t op1, int32_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccus_vx_i16mf4))) +vint16mf4_t vwmaccus(vint16mf4_t op0, uint8_t op1, vint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccus_vx_i16mf4_m))) +vint16mf4_t vwmaccus(vbool64_t op0, vint16mf4_t op1, uint8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccus_vx_i16mf2))) +vint16mf2_t vwmaccus(vint16mf2_t op0, uint8_t op1, vint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccus_vx_i16mf2_m))) +vint16mf2_t vwmaccus(vbool32_t op0, vint16mf2_t op1, uint8_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccus_vx_i16m1))) +vint16m1_t vwmaccus(vint16m1_t op0, uint8_t op1, vint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccus_vx_i16m1_m))) +vint16m1_t vwmaccus(vbool16_t op0, vint16m1_t op1, uint8_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccus_vx_i16m2))) +vint16m2_t vwmaccus(vint16m2_t op0, uint8_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccus_vx_i16m2_m))) +vint16m2_t vwmaccus(vbool8_t op0, vint16m2_t op1, uint8_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccus_vx_i16m4))) +vint16m4_t vwmaccus(vint16m4_t op0, uint8_t op1, vint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccus_vx_i16m4_m))) +vint16m4_t vwmaccus(vbool4_t op0, vint16m4_t op1, uint8_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccus_vx_i16m8))) +vint16m8_t vwmaccus(vint16m8_t op0, uint8_t op1, vint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccus_vx_i16m8_m))) +vint16m8_t vwmaccus(vbool2_t op0, vint16m8_t op1, uint8_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccus_vx_i32mf2))) +vint32mf2_t vwmaccus(vint32mf2_t op0, uint16_t op1, vint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccus_vx_i32mf2_m))) +vint32mf2_t vwmaccus(vbool64_t op0, vint32mf2_t op1, uint16_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccus_vx_i32m1))) +vint32m1_t vwmaccus(vint32m1_t op0, uint16_t op1, vint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccus_vx_i32m1_m))) +vint32m1_t vwmaccus(vbool32_t op0, vint32m1_t op1, uint16_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccus_vx_i32m2))) +vint32m2_t vwmaccus(vint32m2_t op0, uint16_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccus_vx_i32m2_m))) +vint32m2_t vwmaccus(vbool16_t op0, vint32m2_t op1, uint16_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccus_vx_i32m4))) +vint32m4_t vwmaccus(vint32m4_t op0, uint16_t op1, vint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccus_vx_i32m4_m))) +vint32m4_t vwmaccus(vbool8_t op0, vint32m4_t op1, uint16_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccus_vx_i32m8))) +vint32m8_t vwmaccus(vint32m8_t op0, uint16_t op1, vint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccus_vx_i32m8_m))) +vint32m8_t vwmaccus(vbool4_t op0, vint32m8_t op1, uint16_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccus_vx_i64m1))) +vint64m1_t vwmaccus(vint64m1_t op0, uint32_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccus_vx_i64m1_m))) +vint64m1_t vwmaccus(vbool64_t op0, vint64m1_t op1, uint32_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccus_vx_i64m2))) +vint64m2_t vwmaccus(vint64m2_t op0, uint32_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccus_vx_i64m2_m))) +vint64m2_t vwmaccus(vbool32_t op0, vint64m2_t op1, uint32_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccus_vx_i64m4))) +vint64m4_t vwmaccus(vint64m4_t op0, uint32_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccus_vx_i64m4_m))) +vint64m4_t vwmaccus(vbool16_t op0, vint64m4_t op1, uint32_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccus_vx_i64m8))) +vint64m8_t vwmaccus(vint64m8_t op0, uint32_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwmaccus_vx_i64m8_m))) +vint64m8_t vwmaccus(vbool8_t op0, vint64m8_t op1, uint32_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_i8m1))) +vint8m1_t vmerge(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_i8m2))) +vint8m2_t vmerge(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_i8m4))) +vint8m4_t vmerge(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_i8m8))) +vint8m8_t vmerge(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_i8mf2))) +vint8mf2_t vmerge(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_i8mf4))) +vint8mf4_t vmerge(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_i8mf8))) +vint8mf8_t vmerge(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_i16m1))) +vint16m1_t vmerge(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_i16m2))) +vint16m2_t vmerge(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_i16m4))) +vint16m4_t vmerge(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_i16m8))) +vint16m8_t vmerge(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_i16mf2))) +vint16mf2_t vmerge(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_i16mf4))) +vint16mf4_t vmerge(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_i32m1))) +vint32m1_t vmerge(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_i32m2))) +vint32m2_t vmerge(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_i32m4))) +vint32m4_t vmerge(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_i32m8))) +vint32m8_t vmerge(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_i32mf2))) +vint32mf2_t vmerge(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_i64m1))) +vint64m1_t vmerge(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_i64m2))) +vint64m2_t vmerge(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_i64m4))) +vint64m4_t vmerge(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_i64m8))) +vint64m8_t vmerge(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vxm_i8m1))) +vint8m1_t vmerge(vbool8_t op0, vint8m1_t op1, int8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vxm_i8m2))) +vint8m2_t vmerge(vbool4_t op0, vint8m2_t op1, int8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vxm_i8m4))) +vint8m4_t vmerge(vbool2_t op0, vint8m4_t op1, int8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vxm_i8m8))) +vint8m8_t vmerge(vbool1_t op0, vint8m8_t op1, int8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vxm_i8mf2))) +vint8mf2_t vmerge(vbool16_t op0, vint8mf2_t op1, int8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vxm_i8mf4))) +vint8mf4_t vmerge(vbool32_t op0, vint8mf4_t op1, int8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vxm_i8mf8))) +vint8mf8_t vmerge(vbool64_t op0, vint8mf8_t op1, int8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vxm_i16m1))) +vint16m1_t vmerge(vbool16_t op0, vint16m1_t op1, int16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vxm_i16m2))) +vint16m2_t vmerge(vbool8_t op0, vint16m2_t op1, int16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vxm_i16m4))) +vint16m4_t vmerge(vbool4_t op0, vint16m4_t op1, int16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vxm_i16m8))) +vint16m8_t vmerge(vbool2_t op0, vint16m8_t op1, int16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vxm_i16mf2))) +vint16mf2_t vmerge(vbool32_t op0, vint16mf2_t op1, int16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vxm_i16mf4))) +vint16mf4_t vmerge(vbool64_t op0, vint16mf4_t op1, int16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vxm_i32m1))) +vint32m1_t vmerge(vbool32_t op0, vint32m1_t op1, int32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vxm_i32m2))) +vint32m2_t vmerge(vbool16_t op0, vint32m2_t op1, int32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vxm_i32m4))) +vint32m4_t vmerge(vbool8_t op0, vint32m4_t op1, int32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vxm_i32m8))) +vint32m8_t vmerge(vbool4_t op0, vint32m8_t op1, int32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vxm_i32mf2))) +vint32mf2_t vmerge(vbool64_t op0, vint32mf2_t op1, int32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vxm_i64m1))) +vint64m1_t vmerge(vbool64_t op0, vint64m1_t op1, int64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vxm_i64m2))) +vint64m2_t vmerge(vbool32_t op0, vint64m2_t op1, int64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vxm_i64m4))) +vint64m4_t vmerge(vbool16_t op0, vint64m4_t op1, int64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vxm_i64m8))) +vint64m8_t vmerge(vbool8_t op0, vint64m8_t op1, int64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_i32m1))) +void vsuxei32(int32_t * op0, vuint32m1_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_i32m1_m))) +void vsuxei32(vbool32_t op0, int32_t * op1, vuint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_i32m2))) +void vsuxei32(int32_t * op0, vuint32m2_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_i32m2_m))) +void vsuxei32(vbool16_t op0, int32_t * op1, vuint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_i32m4))) +void vsuxei32(int32_t * op0, vuint32m4_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_i32m4_m))) +void vsuxei32(vbool8_t op0, int32_t * op1, vuint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_i32m8))) +void vsuxei32(int32_t * op0, vuint32m8_t op1, vint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_i32m8_m))) +void vsuxei32(vbool4_t op0, int32_t * op1, vuint32m8_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_i32mf2))) +void vsuxei32(int32_t * op0, vuint32mf2_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_i32mf2_m))) +void vsuxei32(vbool64_t op0, int32_t * op1, vuint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_u8m1))) +vuint8m1_t vmerge(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_u8m2))) +vuint8m2_t vmerge(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_u8m4))) +vuint8m4_t vmerge(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_u8m8))) +vuint8m8_t vmerge(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_u8mf2))) +vuint8mf2_t vmerge(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_u8mf4))) +vuint8mf4_t vmerge(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_u8mf8))) +vuint8mf8_t vmerge(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_u16m1))) +vuint16m1_t vmerge(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_u16m2))) +vuint16m2_t vmerge(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_u16m4))) +vuint16m4_t vmerge(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_u16m8))) +vuint16m8_t vmerge(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_u16mf2))) +vuint16mf2_t vmerge(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_u16mf4))) +vuint16mf4_t vmerge(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_u32m1))) +vuint32m1_t vmerge(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_u32m2))) +vuint32m2_t vmerge(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_u32m4))) +vuint32m4_t vmerge(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_u32m8))) +vuint32m8_t vmerge(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_u32mf2))) +vuint32mf2_t vmerge(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_u64m1))) +vuint64m1_t vmerge(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_u64m2))) +vuint64m2_t vmerge(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_u64m4))) +vuint64m4_t vmerge(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_u64m8))) +vuint64m8_t vmerge(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vxm_u8m1))) +vuint8m1_t vmerge(vbool8_t op0, vuint8m1_t op1, uint8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vxm_u8m2))) +vuint8m2_t vmerge(vbool4_t op0, vuint8m2_t op1, uint8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vxm_u8m4))) +vuint8m4_t vmerge(vbool2_t op0, vuint8m4_t op1, uint8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vxm_u8m8))) +vuint8m8_t vmerge(vbool1_t op0, vuint8m8_t op1, uint8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vxm_u8mf2))) +vuint8mf2_t vmerge(vbool16_t op0, vuint8mf2_t op1, uint8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vxm_u8mf4))) +vuint8mf4_t vmerge(vbool32_t op0, vuint8mf4_t op1, uint8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vxm_u8mf8))) +vuint8mf8_t vmerge(vbool64_t op0, vuint8mf8_t op1, uint8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vxm_u16m1))) +vuint16m1_t vmerge(vbool16_t op0, vuint16m1_t op1, uint16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vxm_u16m2))) +vuint16m2_t vmerge(vbool8_t op0, vuint16m2_t op1, uint16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vxm_u16m4))) +vuint16m4_t vmerge(vbool4_t op0, vuint16m4_t op1, uint16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vxm_u16m8))) +vuint16m8_t vmerge(vbool2_t op0, vuint16m8_t op1, uint16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vxm_u16mf2))) +vuint16mf2_t vmerge(vbool32_t op0, vuint16mf2_t op1, uint16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vxm_u16mf4))) +vuint16mf4_t vmerge(vbool64_t op0, vuint16mf4_t op1, uint16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vxm_u32m1))) +vuint32m1_t vmerge(vbool32_t op0, vuint32m1_t op1, uint32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vxm_u32m2))) +vuint32m2_t vmerge(vbool16_t op0, vuint32m2_t op1, uint32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vxm_u32m4))) +vuint32m4_t vmerge(vbool8_t op0, vuint32m4_t op1, uint32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vxm_u32m8))) +vuint32m8_t vmerge(vbool4_t op0, vuint32m8_t op1, uint32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vxm_u32mf2))) +vuint32mf2_t vmerge(vbool64_t op0, vuint32mf2_t op1, uint32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vxm_u64m1))) +vuint64m1_t vmerge(vbool64_t op0, vuint64m1_t op1, uint64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vxm_u64m2))) +vuint64m2_t vmerge(vbool32_t op0, vuint64m2_t op1, uint64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vxm_u64m4))) +vuint64m4_t vmerge(vbool16_t op0, vuint64m4_t op1, uint64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vxm_u64m8))) +vuint64m8_t vmerge(vbool8_t op0, vuint64m8_t op1, uint64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_u8m1))) +vuint8m1_t vmv_v(vuint8m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_u8m2))) +vuint8m2_t vmv_v(vuint8m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_u8m4))) +vuint8m4_t vmv_v(vuint8m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_u8m8))) +vuint8m8_t vmv_v(vuint8m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_u8mf2))) +vuint8mf2_t vmv_v(vuint8mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_u8mf4))) +vuint8mf4_t vmv_v(vuint8mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_u8mf8))) +vuint8mf8_t vmv_v(vuint8mf8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_u16m1))) +vuint16m1_t vmv_v(vuint16m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_u16m2))) +vuint16m2_t vmv_v(vuint16m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_u16m4))) +vuint16m4_t vmv_v(vuint16m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_u16m8))) +vuint16m8_t vmv_v(vuint16m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_u16mf2))) +vuint16mf2_t vmv_v(vuint16mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_u16mf4))) +vuint16mf4_t vmv_v(vuint16mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_u32m1))) +vuint32m1_t vmv_v(vuint32m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_u32m2))) +vuint32m2_t vmv_v(vuint32m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_u32m4))) +vuint32m4_t vmv_v(vuint32m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_u32m8))) +vuint32m8_t vmv_v(vuint32m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_u32mf2))) +vuint32mf2_t vmv_v(vuint32mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_u64m1))) +vuint64m1_t vmv_v(vuint64m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_u64m2))) +vuint64m2_t vmv_v(vuint64m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_u64m4))) +vuint64m4_t vmv_v(vuint64m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_u64m8))) +vuint64m8_t vmv_v(vuint64m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_i8m1))) +vint8m1_t vmv_v(vint8m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_i8m2))) +vint8m2_t vmv_v(vint8m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_i8m4))) +vint8m4_t vmv_v(vint8m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_i8m8))) +vint8m8_t vmv_v(vint8m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_i8mf2))) +vint8mf2_t vmv_v(vint8mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_i8mf4))) +vint8mf4_t vmv_v(vint8mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_i8mf8))) +vint8mf8_t vmv_v(vint8mf8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_i16m1))) +vint16m1_t vmv_v(vint16m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_i16m2))) +vint16m2_t vmv_v(vint16m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_i16m4))) +vint16m4_t vmv_v(vint16m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_i16m8))) +vint16m8_t vmv_v(vint16m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_i16mf2))) +vint16mf2_t vmv_v(vint16mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_i16mf4))) +vint16mf4_t vmv_v(vint16mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_i32m1))) +vint32m1_t vmv_v(vint32m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_i32m2))) +vint32m2_t vmv_v(vint32m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_i32m4))) +vint32m4_t vmv_v(vint32m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_i32m8))) +vint32m8_t vmv_v(vint32m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_i32mf2))) +vint32mf2_t vmv_v(vint32mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_i64m1))) +vint64m1_t vmv_v(vint64m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_i64m2))) +vint64m2_t vmv_v(vint64m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_i64m4))) +vint64m4_t vmv_v(vint64m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_i64m8))) +vint64m8_t vmv_v(vint64m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vv_u8m1))) +vuint8m1_t vsaddu(vuint8m1_t op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vv_u8m1_m))) +vuint8m1_t vsaddu(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vv_u8m2))) +vuint8m2_t vsaddu(vuint8m2_t op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vv_u8m2_m))) +vuint8m2_t vsaddu(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vv_u8m4))) +vuint8m4_t vsaddu(vuint8m4_t op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vv_u8m4_m))) +vuint8m4_t vsaddu(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vv_u8m8))) +vuint8m8_t vsaddu(vuint8m8_t op0, vuint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vv_u8m8_m))) +vuint8m8_t vsaddu(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, vuint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vv_u8mf2))) +vuint8mf2_t vsaddu(vuint8mf2_t op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vv_u8mf2_m))) +vuint8mf2_t vsaddu(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vv_u8mf4))) +vuint8mf4_t vsaddu(vuint8mf4_t op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vv_u8mf4_m))) +vuint8mf4_t vsaddu(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vv_u8mf8))) +vuint8mf8_t vsaddu(vuint8mf8_t op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vv_u8mf8_m))) +vuint8mf8_t vsaddu(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vv_u16m1))) +vuint16m1_t vsaddu(vuint16m1_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vv_u16m1_m))) +vuint16m1_t vsaddu(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vv_u16m2))) +vuint16m2_t vsaddu(vuint16m2_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vv_u16m2_m))) +vuint16m2_t vsaddu(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vv_u16m4))) +vuint16m4_t vsaddu(vuint16m4_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vv_u16m4_m))) +vuint16m4_t vsaddu(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vv_u16m8))) +vuint16m8_t vsaddu(vuint16m8_t op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vv_u16m8_m))) +vuint16m8_t vsaddu(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vv_u16mf2))) +vuint16mf2_t vsaddu(vuint16mf2_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vv_u16mf2_m))) +vuint16mf2_t vsaddu(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vv_u16mf4))) +vuint16mf4_t vsaddu(vuint16mf4_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vv_u16mf4_m))) +vuint16mf4_t vsaddu(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vv_u32m1))) +vuint32m1_t vsaddu(vuint32m1_t op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vv_u32m1_m))) +vuint32m1_t vsaddu(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vv_u32m2))) +vuint32m2_t vsaddu(vuint32m2_t op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vv_u32m2_m))) +vuint32m2_t vsaddu(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vv_u32m4))) +vuint32m4_t vsaddu(vuint32m4_t op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vv_u32m4_m))) +vuint32m4_t vsaddu(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vv_u32m8))) +vuint32m8_t vsaddu(vuint32m8_t op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vv_u32m8_m))) +vuint32m8_t vsaddu(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vv_u32mf2))) +vuint32mf2_t vsaddu(vuint32mf2_t op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vv_u32mf2_m))) +vuint32mf2_t vsaddu(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vv_u64m1))) +vuint64m1_t vsaddu(vuint64m1_t op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vv_u64m1_m))) +vuint64m1_t vsaddu(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vv_u64m2))) +vuint64m2_t vsaddu(vuint64m2_t op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vv_u64m2_m))) +vuint64m2_t vsaddu(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vv_u64m4))) +vuint64m4_t vsaddu(vuint64m4_t op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vv_u64m4_m))) +vuint64m4_t vsaddu(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vv_u64m8))) +vuint64m8_t vsaddu(vuint64m8_t op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vv_u64m8_m))) +vuint64m8_t vsaddu(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vx_u8m1))) +vuint8m1_t vsaddu(vuint8m1_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vx_u8m1_m))) +vuint8m1_t vsaddu(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vx_u8m2))) +vuint8m2_t vsaddu(vuint8m2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vx_u8m2_m))) +vuint8m2_t vsaddu(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vx_u8m4))) +vuint8m4_t vsaddu(vuint8m4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vx_u8m4_m))) +vuint8m4_t vsaddu(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vx_u8m8))) +vuint8m8_t vsaddu(vuint8m8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vx_u8m8_m))) +vuint8m8_t vsaddu(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vx_u8mf2))) +vuint8mf2_t vsaddu(vuint8mf2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vx_u8mf2_m))) +vuint8mf2_t vsaddu(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vx_u8mf4))) +vuint8mf4_t vsaddu(vuint8mf4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vx_u8mf4_m))) +vuint8mf4_t vsaddu(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vx_u8mf8))) +vuint8mf8_t vsaddu(vuint8mf8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vx_u8mf8_m))) +vuint8mf8_t vsaddu(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vx_u16m1))) +vuint16m1_t vsaddu(vuint16m1_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vx_u16m1_m))) +vuint16m1_t vsaddu(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vx_u16m2))) +vuint16m2_t vsaddu(vuint16m2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vx_u16m2_m))) +vuint16m2_t vsaddu(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vx_u16m4))) +vuint16m4_t vsaddu(vuint16m4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vx_u16m4_m))) +vuint16m4_t vsaddu(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vx_u16m8))) +vuint16m8_t vsaddu(vuint16m8_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vx_u16m8_m))) +vuint16m8_t vsaddu(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vx_u16mf2))) +vuint16mf2_t vsaddu(vuint16mf2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vx_u16mf2_m))) +vuint16mf2_t vsaddu(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vx_u16mf4))) +vuint16mf4_t vsaddu(vuint16mf4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vx_u16mf4_m))) +vuint16mf4_t vsaddu(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vx_u32m1))) +vuint32m1_t vsaddu(vuint32m1_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vx_u32m1_m))) +vuint32m1_t vsaddu(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vx_u32m2))) +vuint32m2_t vsaddu(vuint32m2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vx_u32m2_m))) +vuint32m2_t vsaddu(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vx_u32m4))) +vuint32m4_t vsaddu(vuint32m4_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vx_u32m4_m))) +vuint32m4_t vsaddu(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vx_u32m8))) +vuint32m8_t vsaddu(vuint32m8_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vx_u32m8_m))) +vuint32m8_t vsaddu(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vx_u32mf2))) +vuint32mf2_t vsaddu(vuint32mf2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vx_u32mf2_m))) +vuint32mf2_t vsaddu(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vx_u64m1))) +vuint64m1_t vsaddu(vuint64m1_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vx_u64m1_m))) +vuint64m1_t vsaddu(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vx_u64m2))) +vuint64m2_t vsaddu(vuint64m2_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vx_u64m2_m))) +vuint64m2_t vsaddu(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vx_u64m4))) +vuint64m4_t vsaddu(vuint64m4_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vx_u64m4_m))) +vuint64m4_t vsaddu(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vx_u64m8))) +vuint64m8_t vsaddu(vuint64m8_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsaddu_vx_u64m8_m))) +vuint64m8_t vsaddu(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vv_i8m1))) +vint8m1_t vsadd(vint8m1_t op0, vint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vv_i8m1_m))) +vint8m1_t vsadd(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vv_i8m2))) +vint8m2_t vsadd(vint8m2_t op0, vint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vv_i8m2_m))) +vint8m2_t vsadd(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vv_i8m4))) +vint8m4_t vsadd(vint8m4_t op0, vint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vv_i8m4_m))) +vint8m4_t vsadd(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vv_i8m8))) +vint8m8_t vsadd(vint8m8_t op0, vint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vv_i8m8_m))) +vint8m8_t vsadd(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, vint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vv_i8mf2))) +vint8mf2_t vsadd(vint8mf2_t op0, vint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vv_i8mf2_m))) +vint8mf2_t vsadd(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vv_i8mf4))) +vint8mf4_t vsadd(vint8mf4_t op0, vint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vv_i8mf4_m))) +vint8mf4_t vsadd(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vv_i8mf8))) +vint8mf8_t vsadd(vint8mf8_t op0, vint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vv_i8mf8_m))) +vint8mf8_t vsadd(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vv_i16m1))) +vint16m1_t vsadd(vint16m1_t op0, vint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vv_i16m1_m))) +vint16m1_t vsadd(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vv_i16m2))) +vint16m2_t vsadd(vint16m2_t op0, vint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vv_i16m2_m))) +vint16m2_t vsadd(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vv_i16m4))) +vint16m4_t vsadd(vint16m4_t op0, vint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vv_i16m4_m))) +vint16m4_t vsadd(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vv_i16m8))) +vint16m8_t vsadd(vint16m8_t op0, vint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vv_i16m8_m))) +vint16m8_t vsadd(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, vint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vv_i16mf2))) +vint16mf2_t vsadd(vint16mf2_t op0, vint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vv_i16mf2_m))) +vint16mf2_t vsadd(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vv_i16mf4))) +vint16mf4_t vsadd(vint16mf4_t op0, vint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vv_i16mf4_m))) +vint16mf4_t vsadd(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vv_i32m1))) +vint32m1_t vsadd(vint32m1_t op0, vint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vv_i32m1_m))) +vint32m1_t vsadd(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vv_i32m2))) +vint32m2_t vsadd(vint32m2_t op0, vint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vv_i32m2_m))) +vint32m2_t vsadd(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vv_i32m4))) +vint32m4_t vsadd(vint32m4_t op0, vint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vv_i32m4_m))) +vint32m4_t vsadd(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vv_i32m8))) +vint32m8_t vsadd(vint32m8_t op0, vint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vv_i32m8_m))) +vint32m8_t vsadd(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vv_i32mf2))) +vint32mf2_t vsadd(vint32mf2_t op0, vint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vv_i32mf2_m))) +vint32mf2_t vsadd(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vv_i64m1))) +vint64m1_t vsadd(vint64m1_t op0, vint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vv_i64m1_m))) +vint64m1_t vsadd(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vv_i64m2))) +vint64m2_t vsadd(vint64m2_t op0, vint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vv_i64m2_m))) +vint64m2_t vsadd(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vv_i64m4))) +vint64m4_t vsadd(vint64m4_t op0, vint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vv_i64m4_m))) +vint64m4_t vsadd(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vv_i64m8))) +vint64m8_t vsadd(vint64m8_t op0, vint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vv_i64m8_m))) +vint64m8_t vsadd(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vx_i8m1))) +vint8m1_t vsadd(vint8m1_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vx_i8m1_m))) +vint8m1_t vsadd(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vx_i8m2))) +vint8m2_t vsadd(vint8m2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vx_i8m2_m))) +vint8m2_t vsadd(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vx_i8m4))) +vint8m4_t vsadd(vint8m4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vx_i8m4_m))) +vint8m4_t vsadd(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vx_i8m8))) +vint8m8_t vsadd(vint8m8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vx_i8m8_m))) +vint8m8_t vsadd(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vx_i8mf2))) +vint8mf2_t vsadd(vint8mf2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vx_i8mf2_m))) +vint8mf2_t vsadd(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vx_i8mf4))) +vint8mf4_t vsadd(vint8mf4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vx_i8mf4_m))) +vint8mf4_t vsadd(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vx_i8mf8))) +vint8mf8_t vsadd(vint8mf8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vx_i8mf8_m))) +vint8mf8_t vsadd(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vx_i16m1))) +vint16m1_t vsadd(vint16m1_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vx_i16m1_m))) +vint16m1_t vsadd(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vx_i16m2))) +vint16m2_t vsadd(vint16m2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vx_i16m2_m))) +vint16m2_t vsadd(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vx_i16m4))) +vint16m4_t vsadd(vint16m4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vx_i16m4_m))) +vint16m4_t vsadd(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vx_i16m8))) +vint16m8_t vsadd(vint16m8_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vx_i16m8_m))) +vint16m8_t vsadd(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vx_i16mf2))) +vint16mf2_t vsadd(vint16mf2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vx_i16mf2_m))) +vint16mf2_t vsadd(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vx_i16mf4))) +vint16mf4_t vsadd(vint16mf4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vx_i16mf4_m))) +vint16mf4_t vsadd(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vx_i32m1))) +vint32m1_t vsadd(vint32m1_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vx_i32m1_m))) +vint32m1_t vsadd(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vx_i32m2))) +vint32m2_t vsadd(vint32m2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vx_i32m2_m))) +vint32m2_t vsadd(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vx_i32m4))) +vint32m4_t vsadd(vint32m4_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vx_i32m4_m))) +vint32m4_t vsadd(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vx_i32m8))) +vint32m8_t vsadd(vint32m8_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vx_i32m8_m))) +vint32m8_t vsadd(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vx_i32mf2))) +vint32mf2_t vsadd(vint32mf2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vx_i32mf2_m))) +vint32mf2_t vsadd(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vx_i64m1))) +vint64m1_t vsadd(vint64m1_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vx_i64m1_m))) +vint64m1_t vsadd(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vx_i64m2))) +vint64m2_t vsadd(vint64m2_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vx_i64m2_m))) +vint64m2_t vsadd(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vx_i64m4))) +vint64m4_t vsadd(vint64m4_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vx_i64m4_m))) +vint64m4_t vsadd(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vx_i64m8))) +vint64m8_t vsadd(vint64m8_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsadd_vx_i64m8_m))) +vint64m8_t vsadd(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_u32m1))) +void vsuxei32(uint32_t * op0, vuint32m1_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_u32m1_m))) +void vsuxei32(vbool32_t op0, uint32_t * op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_u32m2))) +void vsuxei32(uint32_t * op0, vuint32m2_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_u32m2_m))) +void vsuxei32(vbool16_t op0, uint32_t * op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_u32m4))) +void vsuxei32(uint32_t * op0, vuint32m4_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_u32m4_m))) +void vsuxei32(vbool8_t op0, uint32_t * op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_u32m8))) +void vsuxei32(uint32_t * op0, vuint32m8_t op1, vuint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_u32m8_m))) +void vsuxei32(vbool4_t op0, uint32_t * op1, vuint32m8_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_u32mf2))) +void vsuxei32(uint32_t * op0, vuint32mf2_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_u32mf2_m))) +void vsuxei32(vbool64_t op0, uint32_t * op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vv_u8m1))) +vuint8m1_t vssubu(vuint8m1_t op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vv_u8m1_m))) +vuint8m1_t vssubu(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vv_u8m2))) +vuint8m2_t vssubu(vuint8m2_t op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vv_u8m2_m))) +vuint8m2_t vssubu(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vv_u8m4))) +vuint8m4_t vssubu(vuint8m4_t op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vv_u8m4_m))) +vuint8m4_t vssubu(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vv_u8m8))) +vuint8m8_t vssubu(vuint8m8_t op0, vuint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vv_u8m8_m))) +vuint8m8_t vssubu(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, vuint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vv_u8mf2))) +vuint8mf2_t vssubu(vuint8mf2_t op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vv_u8mf2_m))) +vuint8mf2_t vssubu(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vv_u8mf4))) +vuint8mf4_t vssubu(vuint8mf4_t op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vv_u8mf4_m))) +vuint8mf4_t vssubu(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vv_u8mf8))) +vuint8mf8_t vssubu(vuint8mf8_t op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vv_u8mf8_m))) +vuint8mf8_t vssubu(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vv_u16m1))) +vuint16m1_t vssubu(vuint16m1_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vv_u16m1_m))) +vuint16m1_t vssubu(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vv_u16m2))) +vuint16m2_t vssubu(vuint16m2_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vv_u16m2_m))) +vuint16m2_t vssubu(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vv_u16m4))) +vuint16m4_t vssubu(vuint16m4_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vv_u16m4_m))) +vuint16m4_t vssubu(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vv_u16m8))) +vuint16m8_t vssubu(vuint16m8_t op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vv_u16m8_m))) +vuint16m8_t vssubu(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vv_u16mf2))) +vuint16mf2_t vssubu(vuint16mf2_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vv_u16mf2_m))) +vuint16mf2_t vssubu(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vv_u16mf4))) +vuint16mf4_t vssubu(vuint16mf4_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vv_u16mf4_m))) +vuint16mf4_t vssubu(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vv_u32m1))) +vuint32m1_t vssubu(vuint32m1_t op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vv_u32m1_m))) +vuint32m1_t vssubu(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vv_u32m2))) +vuint32m2_t vssubu(vuint32m2_t op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vv_u32m2_m))) +vuint32m2_t vssubu(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vv_u32m4))) +vuint32m4_t vssubu(vuint32m4_t op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vv_u32m4_m))) +vuint32m4_t vssubu(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vv_u32m8))) +vuint32m8_t vssubu(vuint32m8_t op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vv_u32m8_m))) +vuint32m8_t vssubu(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vv_u32mf2))) +vuint32mf2_t vssubu(vuint32mf2_t op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vv_u32mf2_m))) +vuint32mf2_t vssubu(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vv_u64m1))) +vuint64m1_t vssubu(vuint64m1_t op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vv_u64m1_m))) +vuint64m1_t vssubu(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vv_u64m2))) +vuint64m2_t vssubu(vuint64m2_t op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vv_u64m2_m))) +vuint64m2_t vssubu(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vv_u64m4))) +vuint64m4_t vssubu(vuint64m4_t op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vv_u64m4_m))) +vuint64m4_t vssubu(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vv_u64m8))) +vuint64m8_t vssubu(vuint64m8_t op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vv_u64m8_m))) +vuint64m8_t vssubu(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vx_u8m1))) +vuint8m1_t vssubu(vuint8m1_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vx_u8m1_m))) +vuint8m1_t vssubu(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vx_u8m2))) +vuint8m2_t vssubu(vuint8m2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vx_u8m2_m))) +vuint8m2_t vssubu(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vx_u8m4))) +vuint8m4_t vssubu(vuint8m4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vx_u8m4_m))) +vuint8m4_t vssubu(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vx_u8m8))) +vuint8m8_t vssubu(vuint8m8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vx_u8m8_m))) +vuint8m8_t vssubu(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vx_u8mf2))) +vuint8mf2_t vssubu(vuint8mf2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vx_u8mf2_m))) +vuint8mf2_t vssubu(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vx_u8mf4))) +vuint8mf4_t vssubu(vuint8mf4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vx_u8mf4_m))) +vuint8mf4_t vssubu(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vx_u8mf8))) +vuint8mf8_t vssubu(vuint8mf8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vx_u8mf8_m))) +vuint8mf8_t vssubu(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vx_u16m1))) +vuint16m1_t vssubu(vuint16m1_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vx_u16m1_m))) +vuint16m1_t vssubu(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vx_u16m2))) +vuint16m2_t vssubu(vuint16m2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vx_u16m2_m))) +vuint16m2_t vssubu(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vx_u16m4))) +vuint16m4_t vssubu(vuint16m4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vx_u16m4_m))) +vuint16m4_t vssubu(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vx_u16m8))) +vuint16m8_t vssubu(vuint16m8_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vx_u16m8_m))) +vuint16m8_t vssubu(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vx_u16mf2))) +vuint16mf2_t vssubu(vuint16mf2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vx_u16mf2_m))) +vuint16mf2_t vssubu(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vx_u16mf4))) +vuint16mf4_t vssubu(vuint16mf4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vx_u16mf4_m))) +vuint16mf4_t vssubu(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vx_u32m1))) +vuint32m1_t vssubu(vuint32m1_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vx_u32m1_m))) +vuint32m1_t vssubu(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vx_u32m2))) +vuint32m2_t vssubu(vuint32m2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vx_u32m2_m))) +vuint32m2_t vssubu(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vx_u32m4))) +vuint32m4_t vssubu(vuint32m4_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vx_u32m4_m))) +vuint32m4_t vssubu(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vx_u32m8))) +vuint32m8_t vssubu(vuint32m8_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vx_u32m8_m))) +vuint32m8_t vssubu(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vx_u32mf2))) +vuint32mf2_t vssubu(vuint32mf2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vx_u32mf2_m))) +vuint32mf2_t vssubu(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vx_u64m1))) +vuint64m1_t vssubu(vuint64m1_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vx_u64m1_m))) +vuint64m1_t vssubu(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vx_u64m2))) +vuint64m2_t vssubu(vuint64m2_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vx_u64m2_m))) +vuint64m2_t vssubu(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vx_u64m4))) +vuint64m4_t vssubu(vuint64m4_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vx_u64m4_m))) +vuint64m4_t vssubu(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vx_u64m8))) +vuint64m8_t vssubu(vuint64m8_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssubu_vx_u64m8_m))) +vuint64m8_t vssubu(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vv_i8m1))) +vint8m1_t vssub(vint8m1_t op0, vint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vv_i8m1_m))) +vint8m1_t vssub(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vv_i8m2))) +vint8m2_t vssub(vint8m2_t op0, vint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vv_i8m2_m))) +vint8m2_t vssub(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vv_i8m4))) +vint8m4_t vssub(vint8m4_t op0, vint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vv_i8m4_m))) +vint8m4_t vssub(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vv_i8m8))) +vint8m8_t vssub(vint8m8_t op0, vint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vv_i8m8_m))) +vint8m8_t vssub(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, vint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vv_i8mf2))) +vint8mf2_t vssub(vint8mf2_t op0, vint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vv_i8mf2_m))) +vint8mf2_t vssub(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vv_i8mf4))) +vint8mf4_t vssub(vint8mf4_t op0, vint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vv_i8mf4_m))) +vint8mf4_t vssub(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vv_i8mf8))) +vint8mf8_t vssub(vint8mf8_t op0, vint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vv_i8mf8_m))) +vint8mf8_t vssub(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vv_i16m1))) +vint16m1_t vssub(vint16m1_t op0, vint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vv_i16m1_m))) +vint16m1_t vssub(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vv_i16m2))) +vint16m2_t vssub(vint16m2_t op0, vint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vv_i16m2_m))) +vint16m2_t vssub(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vv_i16m4))) +vint16m4_t vssub(vint16m4_t op0, vint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vv_i16m4_m))) +vint16m4_t vssub(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vv_i16m8))) +vint16m8_t vssub(vint16m8_t op0, vint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vv_i16m8_m))) +vint16m8_t vssub(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, vint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vv_i16mf2))) +vint16mf2_t vssub(vint16mf2_t op0, vint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vv_i16mf2_m))) +vint16mf2_t vssub(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vv_i16mf4))) +vint16mf4_t vssub(vint16mf4_t op0, vint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vv_i16mf4_m))) +vint16mf4_t vssub(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vv_i32m1))) +vint32m1_t vssub(vint32m1_t op0, vint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vv_i32m1_m))) +vint32m1_t vssub(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vv_i32m2))) +vint32m2_t vssub(vint32m2_t op0, vint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vv_i32m2_m))) +vint32m2_t vssub(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vv_i32m4))) +vint32m4_t vssub(vint32m4_t op0, vint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vv_i32m4_m))) +vint32m4_t vssub(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vv_i32m8))) +vint32m8_t vssub(vint32m8_t op0, vint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vv_i32m8_m))) +vint32m8_t vssub(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vv_i32mf2))) +vint32mf2_t vssub(vint32mf2_t op0, vint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vv_i32mf2_m))) +vint32mf2_t vssub(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vv_i64m1))) +vint64m1_t vssub(vint64m1_t op0, vint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vv_i64m1_m))) +vint64m1_t vssub(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vv_i64m2))) +vint64m2_t vssub(vint64m2_t op0, vint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vv_i64m2_m))) +vint64m2_t vssub(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vv_i64m4))) +vint64m4_t vssub(vint64m4_t op0, vint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vv_i64m4_m))) +vint64m4_t vssub(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vv_i64m8))) +vint64m8_t vssub(vint64m8_t op0, vint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vv_i64m8_m))) +vint64m8_t vssub(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vx_i8m1))) +vint8m1_t vssub(vint8m1_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vx_i8m1_m))) +vint8m1_t vssub(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vx_i8m2))) +vint8m2_t vssub(vint8m2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vx_i8m2_m))) +vint8m2_t vssub(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vx_i8m4))) +vint8m4_t vssub(vint8m4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vx_i8m4_m))) +vint8m4_t vssub(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vx_i8m8))) +vint8m8_t vssub(vint8m8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vx_i8m8_m))) +vint8m8_t vssub(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vx_i8mf2))) +vint8mf2_t vssub(vint8mf2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vx_i8mf2_m))) +vint8mf2_t vssub(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vx_i8mf4))) +vint8mf4_t vssub(vint8mf4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vx_i8mf4_m))) +vint8mf4_t vssub(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vx_i8mf8))) +vint8mf8_t vssub(vint8mf8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vx_i8mf8_m))) +vint8mf8_t vssub(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vx_i16m1))) +vint16m1_t vssub(vint16m1_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vx_i16m1_m))) +vint16m1_t vssub(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vx_i16m2))) +vint16m2_t vssub(vint16m2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vx_i16m2_m))) +vint16m2_t vssub(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vx_i16m4))) +vint16m4_t vssub(vint16m4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vx_i16m4_m))) +vint16m4_t vssub(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vx_i16m8))) +vint16m8_t vssub(vint16m8_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vx_i16m8_m))) +vint16m8_t vssub(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vx_i16mf2))) +vint16mf2_t vssub(vint16mf2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vx_i16mf2_m))) +vint16mf2_t vssub(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vx_i16mf4))) +vint16mf4_t vssub(vint16mf4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vx_i16mf4_m))) +vint16mf4_t vssub(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vx_i32m1))) +vint32m1_t vssub(vint32m1_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vx_i32m1_m))) +vint32m1_t vssub(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vx_i32m2))) +vint32m2_t vssub(vint32m2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vx_i32m2_m))) +vint32m2_t vssub(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vx_i32m4))) +vint32m4_t vssub(vint32m4_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vx_i32m4_m))) +vint32m4_t vssub(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vx_i32m8))) +vint32m8_t vssub(vint32m8_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vx_i32m8_m))) +vint32m8_t vssub(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vx_i32mf2))) +vint32mf2_t vssub(vint32mf2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vx_i32mf2_m))) +vint32mf2_t vssub(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vx_i64m1))) +vint64m1_t vssub(vint64m1_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vx_i64m1_m))) +vint64m1_t vssub(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vx_i64m2))) +vint64m2_t vssub(vint64m2_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vx_i64m2_m))) +vint64m2_t vssub(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vx_i64m4))) +vint64m4_t vssub(vint64m4_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vx_i64m4_m))) +vint64m4_t vssub(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vx_i64m8))) +vint64m8_t vssub(vint64m8_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssub_vx_i64m8_m))) +vint64m8_t vssub(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vv_u8m1))) +vuint8m1_t vaaddu(vuint8m1_t op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vv_u8m1_m))) +vuint8m1_t vaaddu(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vv_u8m2))) +vuint8m2_t vaaddu(vuint8m2_t op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vv_u8m2_m))) +vuint8m2_t vaaddu(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vv_u8m4))) +vuint8m4_t vaaddu(vuint8m4_t op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vv_u8m4_m))) +vuint8m4_t vaaddu(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vv_u8m8))) +vuint8m8_t vaaddu(vuint8m8_t op0, vuint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vv_u8m8_m))) +vuint8m8_t vaaddu(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, vuint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vv_u8mf2))) +vuint8mf2_t vaaddu(vuint8mf2_t op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vv_u8mf2_m))) +vuint8mf2_t vaaddu(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vv_u8mf4))) +vuint8mf4_t vaaddu(vuint8mf4_t op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vv_u8mf4_m))) +vuint8mf4_t vaaddu(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vv_u8mf8))) +vuint8mf8_t vaaddu(vuint8mf8_t op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vv_u8mf8_m))) +vuint8mf8_t vaaddu(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vv_u16m1))) +vuint16m1_t vaaddu(vuint16m1_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vv_u16m1_m))) +vuint16m1_t vaaddu(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vv_u16m2))) +vuint16m2_t vaaddu(vuint16m2_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vv_u16m2_m))) +vuint16m2_t vaaddu(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vv_u16m4))) +vuint16m4_t vaaddu(vuint16m4_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vv_u16m4_m))) +vuint16m4_t vaaddu(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vv_u16m8))) +vuint16m8_t vaaddu(vuint16m8_t op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vv_u16m8_m))) +vuint16m8_t vaaddu(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vv_u16mf2))) +vuint16mf2_t vaaddu(vuint16mf2_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vv_u16mf2_m))) +vuint16mf2_t vaaddu(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vv_u16mf4))) +vuint16mf4_t vaaddu(vuint16mf4_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vv_u16mf4_m))) +vuint16mf4_t vaaddu(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vv_u32m1))) +vuint32m1_t vaaddu(vuint32m1_t op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vv_u32m1_m))) +vuint32m1_t vaaddu(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vv_u32m2))) +vuint32m2_t vaaddu(vuint32m2_t op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vv_u32m2_m))) +vuint32m2_t vaaddu(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vv_u32m4))) +vuint32m4_t vaaddu(vuint32m4_t op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vv_u32m4_m))) +vuint32m4_t vaaddu(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vv_u32m8))) +vuint32m8_t vaaddu(vuint32m8_t op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vv_u32m8_m))) +vuint32m8_t vaaddu(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vv_u32mf2))) +vuint32mf2_t vaaddu(vuint32mf2_t op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vv_u32mf2_m))) +vuint32mf2_t vaaddu(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vv_u64m1))) +vuint64m1_t vaaddu(vuint64m1_t op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vv_u64m1_m))) +vuint64m1_t vaaddu(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vv_u64m2))) +vuint64m2_t vaaddu(vuint64m2_t op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vv_u64m2_m))) +vuint64m2_t vaaddu(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vv_u64m4))) +vuint64m4_t vaaddu(vuint64m4_t op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vv_u64m4_m))) +vuint64m4_t vaaddu(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vv_u64m8))) +vuint64m8_t vaaddu(vuint64m8_t op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vv_u64m8_m))) +vuint64m8_t vaaddu(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vx_u8m1))) +vuint8m1_t vaaddu(vuint8m1_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vx_u8m1_m))) +vuint8m1_t vaaddu(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vx_u8m2))) +vuint8m2_t vaaddu(vuint8m2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vx_u8m2_m))) +vuint8m2_t vaaddu(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vx_u8m4))) +vuint8m4_t vaaddu(vuint8m4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vx_u8m4_m))) +vuint8m4_t vaaddu(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vx_u8m8))) +vuint8m8_t vaaddu(vuint8m8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vx_u8m8_m))) +vuint8m8_t vaaddu(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vx_u8mf2))) +vuint8mf2_t vaaddu(vuint8mf2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vx_u8mf2_m))) +vuint8mf2_t vaaddu(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vx_u8mf4))) +vuint8mf4_t vaaddu(vuint8mf4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vx_u8mf4_m))) +vuint8mf4_t vaaddu(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vx_u8mf8))) +vuint8mf8_t vaaddu(vuint8mf8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vx_u8mf8_m))) +vuint8mf8_t vaaddu(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vx_u16m1))) +vuint16m1_t vaaddu(vuint16m1_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vx_u16m1_m))) +vuint16m1_t vaaddu(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vx_u16m2))) +vuint16m2_t vaaddu(vuint16m2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vx_u16m2_m))) +vuint16m2_t vaaddu(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vx_u16m4))) +vuint16m4_t vaaddu(vuint16m4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vx_u16m4_m))) +vuint16m4_t vaaddu(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vx_u16m8))) +vuint16m8_t vaaddu(vuint16m8_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vx_u16m8_m))) +vuint16m8_t vaaddu(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vx_u16mf2))) +vuint16mf2_t vaaddu(vuint16mf2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vx_u16mf2_m))) +vuint16mf2_t vaaddu(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vx_u16mf4))) +vuint16mf4_t vaaddu(vuint16mf4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vx_u16mf4_m))) +vuint16mf4_t vaaddu(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vx_u32m1))) +vuint32m1_t vaaddu(vuint32m1_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vx_u32m1_m))) +vuint32m1_t vaaddu(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vx_u32m2))) +vuint32m2_t vaaddu(vuint32m2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vx_u32m2_m))) +vuint32m2_t vaaddu(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vx_u32m4))) +vuint32m4_t vaaddu(vuint32m4_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vx_u32m4_m))) +vuint32m4_t vaaddu(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vx_u32m8))) +vuint32m8_t vaaddu(vuint32m8_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vx_u32m8_m))) +vuint32m8_t vaaddu(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vx_u32mf2))) +vuint32mf2_t vaaddu(vuint32mf2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vx_u32mf2_m))) +vuint32mf2_t vaaddu(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vx_u64m1))) +vuint64m1_t vaaddu(vuint64m1_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vx_u64m1_m))) +vuint64m1_t vaaddu(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vx_u64m2))) +vuint64m2_t vaaddu(vuint64m2_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vx_u64m2_m))) +vuint64m2_t vaaddu(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vx_u64m4))) +vuint64m4_t vaaddu(vuint64m4_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vx_u64m4_m))) +vuint64m4_t vaaddu(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vx_u64m8))) +vuint64m8_t vaaddu(vuint64m8_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaaddu_vx_u64m8_m))) +vuint64m8_t vaaddu(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vv_i8m1))) +vint8m1_t vaadd(vint8m1_t op0, vint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vv_i8m1_m))) +vint8m1_t vaadd(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vv_i8m2))) +vint8m2_t vaadd(vint8m2_t op0, vint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vv_i8m2_m))) +vint8m2_t vaadd(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vv_i8m4))) +vint8m4_t vaadd(vint8m4_t op0, vint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vv_i8m4_m))) +vint8m4_t vaadd(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vv_i8m8))) +vint8m8_t vaadd(vint8m8_t op0, vint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vv_i8m8_m))) +vint8m8_t vaadd(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, vint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vv_i8mf2))) +vint8mf2_t vaadd(vint8mf2_t op0, vint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vv_i8mf2_m))) +vint8mf2_t vaadd(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vv_i8mf4))) +vint8mf4_t vaadd(vint8mf4_t op0, vint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vv_i8mf4_m))) +vint8mf4_t vaadd(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vv_i8mf8))) +vint8mf8_t vaadd(vint8mf8_t op0, vint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vv_i8mf8_m))) +vint8mf8_t vaadd(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vv_i16m1))) +vint16m1_t vaadd(vint16m1_t op0, vint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vv_i16m1_m))) +vint16m1_t vaadd(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vv_i16m2))) +vint16m2_t vaadd(vint16m2_t op0, vint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vv_i16m2_m))) +vint16m2_t vaadd(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vv_i16m4))) +vint16m4_t vaadd(vint16m4_t op0, vint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vv_i16m4_m))) +vint16m4_t vaadd(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vv_i16m8))) +vint16m8_t vaadd(vint16m8_t op0, vint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vv_i16m8_m))) +vint16m8_t vaadd(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, vint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vv_i16mf2))) +vint16mf2_t vaadd(vint16mf2_t op0, vint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vv_i16mf2_m))) +vint16mf2_t vaadd(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vv_i16mf4))) +vint16mf4_t vaadd(vint16mf4_t op0, vint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vv_i16mf4_m))) +vint16mf4_t vaadd(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vv_i32m1))) +vint32m1_t vaadd(vint32m1_t op0, vint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vv_i32m1_m))) +vint32m1_t vaadd(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vv_i32m2))) +vint32m2_t vaadd(vint32m2_t op0, vint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vv_i32m2_m))) +vint32m2_t vaadd(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vv_i32m4))) +vint32m4_t vaadd(vint32m4_t op0, vint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vv_i32m4_m))) +vint32m4_t vaadd(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vv_i32m8))) +vint32m8_t vaadd(vint32m8_t op0, vint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vv_i32m8_m))) +vint32m8_t vaadd(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vv_i32mf2))) +vint32mf2_t vaadd(vint32mf2_t op0, vint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vv_i32mf2_m))) +vint32mf2_t vaadd(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vv_i64m1))) +vint64m1_t vaadd(vint64m1_t op0, vint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vv_i64m1_m))) +vint64m1_t vaadd(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vv_i64m2))) +vint64m2_t vaadd(vint64m2_t op0, vint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vv_i64m2_m))) +vint64m2_t vaadd(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vv_i64m4))) +vint64m4_t vaadd(vint64m4_t op0, vint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vv_i64m4_m))) +vint64m4_t vaadd(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vv_i64m8))) +vint64m8_t vaadd(vint64m8_t op0, vint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vv_i64m8_m))) +vint64m8_t vaadd(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vx_i8m1))) +vint8m1_t vaadd(vint8m1_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vx_i8m1_m))) +vint8m1_t vaadd(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vx_i8m2))) +vint8m2_t vaadd(vint8m2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vx_i8m2_m))) +vint8m2_t vaadd(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vx_i8m4))) +vint8m4_t vaadd(vint8m4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vx_i8m4_m))) +vint8m4_t vaadd(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vx_i8m8))) +vint8m8_t vaadd(vint8m8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vx_i8m8_m))) +vint8m8_t vaadd(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vx_i8mf2))) +vint8mf2_t vaadd(vint8mf2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vx_i8mf2_m))) +vint8mf2_t vaadd(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vx_i8mf4))) +vint8mf4_t vaadd(vint8mf4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vx_i8mf4_m))) +vint8mf4_t vaadd(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vx_i8mf8))) +vint8mf8_t vaadd(vint8mf8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vx_i8mf8_m))) +vint8mf8_t vaadd(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vx_i16m1))) +vint16m1_t vaadd(vint16m1_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vx_i16m1_m))) +vint16m1_t vaadd(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vx_i16m2))) +vint16m2_t vaadd(vint16m2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vx_i16m2_m))) +vint16m2_t vaadd(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vx_i16m4))) +vint16m4_t vaadd(vint16m4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vx_i16m4_m))) +vint16m4_t vaadd(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vx_i16m8))) +vint16m8_t vaadd(vint16m8_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vx_i16m8_m))) +vint16m8_t vaadd(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vx_i16mf2))) +vint16mf2_t vaadd(vint16mf2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vx_i16mf2_m))) +vint16mf2_t vaadd(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vx_i16mf4))) +vint16mf4_t vaadd(vint16mf4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vx_i16mf4_m))) +vint16mf4_t vaadd(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vx_i32m1))) +vint32m1_t vaadd(vint32m1_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vx_i32m1_m))) +vint32m1_t vaadd(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vx_i32m2))) +vint32m2_t vaadd(vint32m2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vx_i32m2_m))) +vint32m2_t vaadd(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vx_i32m4))) +vint32m4_t vaadd(vint32m4_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vx_i32m4_m))) +vint32m4_t vaadd(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vx_i32m8))) +vint32m8_t vaadd(vint32m8_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vx_i32m8_m))) +vint32m8_t vaadd(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vx_i32mf2))) +vint32mf2_t vaadd(vint32mf2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vx_i32mf2_m))) +vint32mf2_t vaadd(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vx_i64m1))) +vint64m1_t vaadd(vint64m1_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vx_i64m1_m))) +vint64m1_t vaadd(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vx_i64m2))) +vint64m2_t vaadd(vint64m2_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vx_i64m2_m))) +vint64m2_t vaadd(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vx_i64m4))) +vint64m4_t vaadd(vint64m4_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vx_i64m4_m))) +vint64m4_t vaadd(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vx_i64m8))) +vint64m8_t vaadd(vint64m8_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vaadd_vx_i64m8_m))) +vint64m8_t vaadd(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vv_u8m1))) +vuint8m1_t vasubu(vuint8m1_t op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vv_u8m1_m))) +vuint8m1_t vasubu(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vv_u8m2))) +vuint8m2_t vasubu(vuint8m2_t op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vv_u8m2_m))) +vuint8m2_t vasubu(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vv_u8m4))) +vuint8m4_t vasubu(vuint8m4_t op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vv_u8m4_m))) +vuint8m4_t vasubu(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vv_u8m8))) +vuint8m8_t vasubu(vuint8m8_t op0, vuint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vv_u8m8_m))) +vuint8m8_t vasubu(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, vuint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vv_u8mf2))) +vuint8mf2_t vasubu(vuint8mf2_t op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vv_u8mf2_m))) +vuint8mf2_t vasubu(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vv_u8mf4))) +vuint8mf4_t vasubu(vuint8mf4_t op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vv_u8mf4_m))) +vuint8mf4_t vasubu(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vv_u8mf8))) +vuint8mf8_t vasubu(vuint8mf8_t op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vv_u8mf8_m))) +vuint8mf8_t vasubu(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vv_u16m1))) +vuint16m1_t vasubu(vuint16m1_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vv_u16m1_m))) +vuint16m1_t vasubu(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vv_u16m2))) +vuint16m2_t vasubu(vuint16m2_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vv_u16m2_m))) +vuint16m2_t vasubu(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vv_u16m4))) +vuint16m4_t vasubu(vuint16m4_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vv_u16m4_m))) +vuint16m4_t vasubu(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vv_u16m8))) +vuint16m8_t vasubu(vuint16m8_t op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vv_u16m8_m))) +vuint16m8_t vasubu(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vv_u16mf2))) +vuint16mf2_t vasubu(vuint16mf2_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vv_u16mf2_m))) +vuint16mf2_t vasubu(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vv_u16mf4))) +vuint16mf4_t vasubu(vuint16mf4_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vv_u16mf4_m))) +vuint16mf4_t vasubu(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vv_u32m1))) +vuint32m1_t vasubu(vuint32m1_t op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vv_u32m1_m))) +vuint32m1_t vasubu(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vv_u32m2))) +vuint32m2_t vasubu(vuint32m2_t op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vv_u32m2_m))) +vuint32m2_t vasubu(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vv_u32m4))) +vuint32m4_t vasubu(vuint32m4_t op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vv_u32m4_m))) +vuint32m4_t vasubu(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vv_u32m8))) +vuint32m8_t vasubu(vuint32m8_t op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vv_u32m8_m))) +vuint32m8_t vasubu(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vv_u32mf2))) +vuint32mf2_t vasubu(vuint32mf2_t op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vv_u32mf2_m))) +vuint32mf2_t vasubu(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vv_u64m1))) +vuint64m1_t vasubu(vuint64m1_t op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vv_u64m1_m))) +vuint64m1_t vasubu(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vv_u64m2))) +vuint64m2_t vasubu(vuint64m2_t op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vv_u64m2_m))) +vuint64m2_t vasubu(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vv_u64m4))) +vuint64m4_t vasubu(vuint64m4_t op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vv_u64m4_m))) +vuint64m4_t vasubu(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vv_u64m8))) +vuint64m8_t vasubu(vuint64m8_t op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vv_u64m8_m))) +vuint64m8_t vasubu(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vx_u8m1))) +vuint8m1_t vasubu(vuint8m1_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vx_u8m1_m))) +vuint8m1_t vasubu(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vx_u8m2))) +vuint8m2_t vasubu(vuint8m2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vx_u8m2_m))) +vuint8m2_t vasubu(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vx_u8m4))) +vuint8m4_t vasubu(vuint8m4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vx_u8m4_m))) +vuint8m4_t vasubu(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vx_u8m8))) +vuint8m8_t vasubu(vuint8m8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vx_u8m8_m))) +vuint8m8_t vasubu(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vx_u8mf2))) +vuint8mf2_t vasubu(vuint8mf2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vx_u8mf2_m))) +vuint8mf2_t vasubu(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vx_u8mf4))) +vuint8mf4_t vasubu(vuint8mf4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vx_u8mf4_m))) +vuint8mf4_t vasubu(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vx_u8mf8))) +vuint8mf8_t vasubu(vuint8mf8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vx_u8mf8_m))) +vuint8mf8_t vasubu(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, uint8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vx_u16m1))) +vuint16m1_t vasubu(vuint16m1_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vx_u16m1_m))) +vuint16m1_t vasubu(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vx_u16m2))) +vuint16m2_t vasubu(vuint16m2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vx_u16m2_m))) +vuint16m2_t vasubu(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vx_u16m4))) +vuint16m4_t vasubu(vuint16m4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vx_u16m4_m))) +vuint16m4_t vasubu(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vx_u16m8))) +vuint16m8_t vasubu(vuint16m8_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vx_u16m8_m))) +vuint16m8_t vasubu(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vx_u16mf2))) +vuint16mf2_t vasubu(vuint16mf2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vx_u16mf2_m))) +vuint16mf2_t vasubu(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vx_u16mf4))) +vuint16mf4_t vasubu(vuint16mf4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vx_u16mf4_m))) +vuint16mf4_t vasubu(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, uint16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vx_u32m1))) +vuint32m1_t vasubu(vuint32m1_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vx_u32m1_m))) +vuint32m1_t vasubu(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vx_u32m2))) +vuint32m2_t vasubu(vuint32m2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vx_u32m2_m))) +vuint32m2_t vasubu(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vx_u32m4))) +vuint32m4_t vasubu(vuint32m4_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vx_u32m4_m))) +vuint32m4_t vasubu(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vx_u32m8))) +vuint32m8_t vasubu(vuint32m8_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vx_u32m8_m))) +vuint32m8_t vasubu(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vx_u32mf2))) +vuint32mf2_t vasubu(vuint32mf2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vx_u32mf2_m))) +vuint32mf2_t vasubu(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, uint32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vx_u64m1))) +vuint64m1_t vasubu(vuint64m1_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vx_u64m1_m))) +vuint64m1_t vasubu(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vx_u64m2))) +vuint64m2_t vasubu(vuint64m2_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vx_u64m2_m))) +vuint64m2_t vasubu(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vx_u64m4))) +vuint64m4_t vasubu(vuint64m4_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vx_u64m4_m))) +vuint64m4_t vasubu(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vx_u64m8))) +vuint64m8_t vasubu(vuint64m8_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasubu_vx_u64m8_m))) +vuint64m8_t vasubu(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, uint64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_i8m1))) +void vsuxei8(int8_t * op0, vuint8m1_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_i8m1_m))) +void vsuxei8(vbool8_t op0, int8_t * op1, vuint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_i8m2))) +void vsuxei8(int8_t * op0, vuint8m2_t op1, vint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_i8m2_m))) +void vsuxei8(vbool4_t op0, int8_t * op1, vuint8m2_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_i8m4))) +void vsuxei8(int8_t * op0, vuint8m4_t op1, vint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_i8m4_m))) +void vsuxei8(vbool2_t op0, int8_t * op1, vuint8m4_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_i8m8))) +void vsuxei8(int8_t * op0, vuint8m8_t op1, vint8m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_i8m8_m))) +void vsuxei8(vbool1_t op0, int8_t * op1, vuint8m8_t op2, vint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_i8mf2))) +void vsuxei8(int8_t * op0, vuint8mf2_t op1, vint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_i8mf2_m))) +void vsuxei8(vbool16_t op0, int8_t * op1, vuint8mf2_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_i8mf4))) +void vsuxei8(int8_t * op0, vuint8mf4_t op1, vint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_i8mf4_m))) +void vsuxei8(vbool32_t op0, int8_t * op1, vuint8mf4_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_i8mf8))) +void vsuxei8(int8_t * op0, vuint8mf8_t op1, vint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_i8mf8_m))) +void vsuxei8(vbool64_t op0, int8_t * op1, vuint8mf8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_i32m1))) +void vsuxei64(int32_t * op0, vuint64m2_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_i32m1_m))) +void vsuxei64(vbool32_t op0, int32_t * op1, vuint64m2_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_i32m2))) +void vsuxei64(int32_t * op0, vuint64m4_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_i32m2_m))) +void vsuxei64(vbool16_t op0, int32_t * op1, vuint64m4_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_i32m4))) +void vsuxei64(int32_t * op0, vuint64m8_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_i32m4_m))) +void vsuxei64(vbool8_t op0, int32_t * op1, vuint64m8_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_i32mf2))) +void vsuxei64(int32_t * op0, vuint64m1_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_i32mf2_m))) +void vsuxei64(vbool64_t op0, int32_t * op1, vuint64m1_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vv_i8m1))) +vint8m1_t vasub(vint8m1_t op0, vint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vv_i8m1_m))) +vint8m1_t vasub(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vv_i8m2))) +vint8m2_t vasub(vint8m2_t op0, vint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vv_i8m2_m))) +vint8m2_t vasub(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vv_i8m4))) +vint8m4_t vasub(vint8m4_t op0, vint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vv_i8m4_m))) +vint8m4_t vasub(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vv_i8m8))) +vint8m8_t vasub(vint8m8_t op0, vint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vv_i8m8_m))) +vint8m8_t vasub(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, vint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vv_i8mf2))) +vint8mf2_t vasub(vint8mf2_t op0, vint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vv_i8mf2_m))) +vint8mf2_t vasub(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vv_i8mf4))) +vint8mf4_t vasub(vint8mf4_t op0, vint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vv_i8mf4_m))) +vint8mf4_t vasub(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vv_i8mf8))) +vint8mf8_t vasub(vint8mf8_t op0, vint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vv_i8mf8_m))) +vint8mf8_t vasub(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vv_i16m1))) +vint16m1_t vasub(vint16m1_t op0, vint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vv_i16m1_m))) +vint16m1_t vasub(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vv_i16m2))) +vint16m2_t vasub(vint16m2_t op0, vint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vv_i16m2_m))) +vint16m2_t vasub(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vv_i16m4))) +vint16m4_t vasub(vint16m4_t op0, vint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vv_i16m4_m))) +vint16m4_t vasub(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vv_i16m8))) +vint16m8_t vasub(vint16m8_t op0, vint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vv_i16m8_m))) +vint16m8_t vasub(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, vint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vv_i16mf2))) +vint16mf2_t vasub(vint16mf2_t op0, vint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vv_i16mf2_m))) +vint16mf2_t vasub(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vv_i16mf4))) +vint16mf4_t vasub(vint16mf4_t op0, vint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vv_i16mf4_m))) +vint16mf4_t vasub(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vv_i32m1))) +vint32m1_t vasub(vint32m1_t op0, vint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vv_i32m1_m))) +vint32m1_t vasub(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vv_i32m2))) +vint32m2_t vasub(vint32m2_t op0, vint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vv_i32m2_m))) +vint32m2_t vasub(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vv_i32m4))) +vint32m4_t vasub(vint32m4_t op0, vint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vv_i32m4_m))) +vint32m4_t vasub(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vv_i32m8))) +vint32m8_t vasub(vint32m8_t op0, vint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vv_i32m8_m))) +vint32m8_t vasub(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vv_i32mf2))) +vint32mf2_t vasub(vint32mf2_t op0, vint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vv_i32mf2_m))) +vint32mf2_t vasub(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vv_i64m1))) +vint64m1_t vasub(vint64m1_t op0, vint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vv_i64m1_m))) +vint64m1_t vasub(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vv_i64m2))) +vint64m2_t vasub(vint64m2_t op0, vint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vv_i64m2_m))) +vint64m2_t vasub(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vv_i64m4))) +vint64m4_t vasub(vint64m4_t op0, vint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vv_i64m4_m))) +vint64m4_t vasub(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vv_i64m8))) +vint64m8_t vasub(vint64m8_t op0, vint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vv_i64m8_m))) +vint64m8_t vasub(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vx_i8m1))) +vint8m1_t vasub(vint8m1_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vx_i8m1_m))) +vint8m1_t vasub(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vx_i8m2))) +vint8m2_t vasub(vint8m2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vx_i8m2_m))) +vint8m2_t vasub(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vx_i8m4))) +vint8m4_t vasub(vint8m4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vx_i8m4_m))) +vint8m4_t vasub(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vx_i8m8))) +vint8m8_t vasub(vint8m8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vx_i8m8_m))) +vint8m8_t vasub(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vx_i8mf2))) +vint8mf2_t vasub(vint8mf2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vx_i8mf2_m))) +vint8mf2_t vasub(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vx_i8mf4))) +vint8mf4_t vasub(vint8mf4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vx_i8mf4_m))) +vint8mf4_t vasub(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vx_i8mf8))) +vint8mf8_t vasub(vint8mf8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vx_i8mf8_m))) +vint8mf8_t vasub(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vx_i16m1))) +vint16m1_t vasub(vint16m1_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vx_i16m1_m))) +vint16m1_t vasub(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vx_i16m2))) +vint16m2_t vasub(vint16m2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vx_i16m2_m))) +vint16m2_t vasub(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vx_i16m4))) +vint16m4_t vasub(vint16m4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vx_i16m4_m))) +vint16m4_t vasub(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vx_i16m8))) +vint16m8_t vasub(vint16m8_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vx_i16m8_m))) +vint16m8_t vasub(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vx_i16mf2))) +vint16mf2_t vasub(vint16mf2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vx_i16mf2_m))) +vint16mf2_t vasub(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vx_i16mf4))) +vint16mf4_t vasub(vint16mf4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vx_i16mf4_m))) +vint16mf4_t vasub(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vx_i32m1))) +vint32m1_t vasub(vint32m1_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vx_i32m1_m))) +vint32m1_t vasub(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vx_i32m2))) +vint32m2_t vasub(vint32m2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vx_i32m2_m))) +vint32m2_t vasub(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vx_i32m4))) +vint32m4_t vasub(vint32m4_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vx_i32m4_m))) +vint32m4_t vasub(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vx_i32m8))) +vint32m8_t vasub(vint32m8_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vx_i32m8_m))) +vint32m8_t vasub(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vx_i32mf2))) +vint32mf2_t vasub(vint32mf2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vx_i32mf2_m))) +vint32mf2_t vasub(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vx_i64m1))) +vint64m1_t vasub(vint64m1_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vx_i64m1_m))) +vint64m1_t vasub(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vx_i64m2))) +vint64m2_t vasub(vint64m2_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vx_i64m2_m))) +vint64m2_t vasub(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vx_i64m4))) +vint64m4_t vasub(vint64m4_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vx_i64m4_m))) +vint64m4_t vasub(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vx_i64m8))) +vint64m8_t vasub(vint64m8_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vasub_vx_i64m8_m))) +vint64m8_t vasub(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vv_i8m1))) +vint8m1_t vsmul(vint8m1_t op0, vint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vv_i8m1_m))) +vint8m1_t vsmul(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vv_i8m2))) +vint8m2_t vsmul(vint8m2_t op0, vint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vv_i8m2_m))) +vint8m2_t vsmul(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vv_i8m4))) +vint8m4_t vsmul(vint8m4_t op0, vint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vv_i8m4_m))) +vint8m4_t vsmul(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vv_i8m8))) +vint8m8_t vsmul(vint8m8_t op0, vint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vv_i8m8_m))) +vint8m8_t vsmul(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, vint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vv_i8mf2))) +vint8mf2_t vsmul(vint8mf2_t op0, vint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vv_i8mf2_m))) +vint8mf2_t vsmul(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vv_i8mf4))) +vint8mf4_t vsmul(vint8mf4_t op0, vint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vv_i8mf4_m))) +vint8mf4_t vsmul(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vv_i8mf8))) +vint8mf8_t vsmul(vint8mf8_t op0, vint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vv_i8mf8_m))) +vint8mf8_t vsmul(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vv_i16m1))) +vint16m1_t vsmul(vint16m1_t op0, vint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vv_i16m1_m))) +vint16m1_t vsmul(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vv_i16m2))) +vint16m2_t vsmul(vint16m2_t op0, vint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vv_i16m2_m))) +vint16m2_t vsmul(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vv_i16m4))) +vint16m4_t vsmul(vint16m4_t op0, vint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vv_i16m4_m))) +vint16m4_t vsmul(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vv_i16m8))) +vint16m8_t vsmul(vint16m8_t op0, vint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vv_i16m8_m))) +vint16m8_t vsmul(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, vint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vv_i16mf2))) +vint16mf2_t vsmul(vint16mf2_t op0, vint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vv_i16mf2_m))) +vint16mf2_t vsmul(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vv_i16mf4))) +vint16mf4_t vsmul(vint16mf4_t op0, vint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vv_i16mf4_m))) +vint16mf4_t vsmul(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vv_i32m1))) +vint32m1_t vsmul(vint32m1_t op0, vint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vv_i32m1_m))) +vint32m1_t vsmul(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vv_i32m2))) +vint32m2_t vsmul(vint32m2_t op0, vint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vv_i32m2_m))) +vint32m2_t vsmul(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vv_i32m4))) +vint32m4_t vsmul(vint32m4_t op0, vint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vv_i32m4_m))) +vint32m4_t vsmul(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vv_i32m8))) +vint32m8_t vsmul(vint32m8_t op0, vint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vv_i32m8_m))) +vint32m8_t vsmul(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vv_i32mf2))) +vint32mf2_t vsmul(vint32mf2_t op0, vint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vv_i32mf2_m))) +vint32mf2_t vsmul(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vv_i64m1))) +vint64m1_t vsmul(vint64m1_t op0, vint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vv_i64m1_m))) +vint64m1_t vsmul(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vv_i64m2))) +vint64m2_t vsmul(vint64m2_t op0, vint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vv_i64m2_m))) +vint64m2_t vsmul(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vv_i64m4))) +vint64m4_t vsmul(vint64m4_t op0, vint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vv_i64m4_m))) +vint64m4_t vsmul(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vv_i64m8))) +vint64m8_t vsmul(vint64m8_t op0, vint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vv_i64m8_m))) +vint64m8_t vsmul(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vx_i8m1))) +vint8m1_t vsmul(vint8m1_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vx_i8m1_m))) +vint8m1_t vsmul(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vx_i8m2))) +vint8m2_t vsmul(vint8m2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vx_i8m2_m))) +vint8m2_t vsmul(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vx_i8m4))) +vint8m4_t vsmul(vint8m4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vx_i8m4_m))) +vint8m4_t vsmul(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vx_i8m8))) +vint8m8_t vsmul(vint8m8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vx_i8m8_m))) +vint8m8_t vsmul(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vx_i8mf2))) +vint8mf2_t vsmul(vint8mf2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vx_i8mf2_m))) +vint8mf2_t vsmul(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vx_i8mf4))) +vint8mf4_t vsmul(vint8mf4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vx_i8mf4_m))) +vint8mf4_t vsmul(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vx_i8mf8))) +vint8mf8_t vsmul(vint8mf8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vx_i8mf8_m))) +vint8mf8_t vsmul(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vx_i16m1))) +vint16m1_t vsmul(vint16m1_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vx_i16m1_m))) +vint16m1_t vsmul(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vx_i16m2))) +vint16m2_t vsmul(vint16m2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vx_i16m2_m))) +vint16m2_t vsmul(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vx_i16m4))) +vint16m4_t vsmul(vint16m4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vx_i16m4_m))) +vint16m4_t vsmul(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vx_i16m8))) +vint16m8_t vsmul(vint16m8_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vx_i16m8_m))) +vint16m8_t vsmul(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vx_i16mf2))) +vint16mf2_t vsmul(vint16mf2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vx_i16mf2_m))) +vint16mf2_t vsmul(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vx_i16mf4))) +vint16mf4_t vsmul(vint16mf4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vx_i16mf4_m))) +vint16mf4_t vsmul(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vx_i32m1))) +vint32m1_t vsmul(vint32m1_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vx_i32m1_m))) +vint32m1_t vsmul(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vx_i32m2))) +vint32m2_t vsmul(vint32m2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vx_i32m2_m))) +vint32m2_t vsmul(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vx_i32m4))) +vint32m4_t vsmul(vint32m4_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vx_i32m4_m))) +vint32m4_t vsmul(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vx_i32m8))) +vint32m8_t vsmul(vint32m8_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vx_i32m8_m))) +vint32m8_t vsmul(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vx_i32mf2))) +vint32mf2_t vsmul(vint32mf2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vx_i32mf2_m))) +vint32mf2_t vsmul(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vx_i64m1))) +vint64m1_t vsmul(vint64m1_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vx_i64m1_m))) +vint64m1_t vsmul(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vx_i64m2))) +vint64m2_t vsmul(vint64m2_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vx_i64m2_m))) +vint64m2_t vsmul(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vx_i64m4))) +vint64m4_t vsmul(vint64m4_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vx_i64m4_m))) +vint64m4_t vsmul(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vx_i64m8))) +vint64m8_t vsmul(vint64m8_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsmul_vx_i64m8_m))) +vint64m8_t vsmul(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vv_u8m1))) +vuint8m1_t vssrl(vuint8m1_t op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vv_u8m1_m))) +vuint8m1_t vssrl(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vv_u8m2))) +vuint8m2_t vssrl(vuint8m2_t op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vv_u8m2_m))) +vuint8m2_t vssrl(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vv_u8m4))) +vuint8m4_t vssrl(vuint8m4_t op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vv_u8m4_m))) +vuint8m4_t vssrl(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vv_u8m8))) +vuint8m8_t vssrl(vuint8m8_t op0, vuint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vv_u8m8_m))) +vuint8m8_t vssrl(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, vuint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vv_u8mf2))) +vuint8mf2_t vssrl(vuint8mf2_t op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vv_u8mf2_m))) +vuint8mf2_t vssrl(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vv_u8mf4))) +vuint8mf4_t vssrl(vuint8mf4_t op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vv_u8mf4_m))) +vuint8mf4_t vssrl(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vv_u8mf8))) +vuint8mf8_t vssrl(vuint8mf8_t op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vv_u8mf8_m))) +vuint8mf8_t vssrl(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vv_u16m1))) +vuint16m1_t vssrl(vuint16m1_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vv_u16m1_m))) +vuint16m1_t vssrl(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vv_u16m2))) +vuint16m2_t vssrl(vuint16m2_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vv_u16m2_m))) +vuint16m2_t vssrl(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vv_u16m4))) +vuint16m4_t vssrl(vuint16m4_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vv_u16m4_m))) +vuint16m4_t vssrl(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vv_u16m8))) +vuint16m8_t vssrl(vuint16m8_t op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vv_u16m8_m))) +vuint16m8_t vssrl(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vv_u16mf2))) +vuint16mf2_t vssrl(vuint16mf2_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vv_u16mf2_m))) +vuint16mf2_t vssrl(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vv_u16mf4))) +vuint16mf4_t vssrl(vuint16mf4_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vv_u16mf4_m))) +vuint16mf4_t vssrl(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vv_u32m1))) +vuint32m1_t vssrl(vuint32m1_t op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vv_u32m1_m))) +vuint32m1_t vssrl(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vv_u32m2))) +vuint32m2_t vssrl(vuint32m2_t op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vv_u32m2_m))) +vuint32m2_t vssrl(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vv_u32m4))) +vuint32m4_t vssrl(vuint32m4_t op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vv_u32m4_m))) +vuint32m4_t vssrl(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vv_u32m8))) +vuint32m8_t vssrl(vuint32m8_t op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vv_u32m8_m))) +vuint32m8_t vssrl(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vv_u32mf2))) +vuint32mf2_t vssrl(vuint32mf2_t op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vv_u32mf2_m))) +vuint32mf2_t vssrl(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vv_u64m1))) +vuint64m1_t vssrl(vuint64m1_t op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vv_u64m1_m))) +vuint64m1_t vssrl(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vv_u64m2))) +vuint64m2_t vssrl(vuint64m2_t op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vv_u64m2_m))) +vuint64m2_t vssrl(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vv_u64m4))) +vuint64m4_t vssrl(vuint64m4_t op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vv_u64m4_m))) +vuint64m4_t vssrl(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vv_u64m8))) +vuint64m8_t vssrl(vuint64m8_t op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vv_u64m8_m))) +vuint64m8_t vssrl(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vx_u8m1))) +vuint8m1_t vssrl(vuint8m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vx_u8m1_m))) +vuint8m1_t vssrl(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vx_u8m2))) +vuint8m2_t vssrl(vuint8m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vx_u8m2_m))) +vuint8m2_t vssrl(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vx_u8m4))) +vuint8m4_t vssrl(vuint8m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vx_u8m4_m))) +vuint8m4_t vssrl(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vx_u8m8))) +vuint8m8_t vssrl(vuint8m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vx_u8m8_m))) +vuint8m8_t vssrl(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vx_u8mf2))) +vuint8mf2_t vssrl(vuint8mf2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vx_u8mf2_m))) +vuint8mf2_t vssrl(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vx_u8mf4))) +vuint8mf4_t vssrl(vuint8mf4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vx_u8mf4_m))) +vuint8mf4_t vssrl(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vx_u8mf8))) +vuint8mf8_t vssrl(vuint8mf8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vx_u8mf8_m))) +vuint8mf8_t vssrl(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vx_u16m1))) +vuint16m1_t vssrl(vuint16m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vx_u16m1_m))) +vuint16m1_t vssrl(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vx_u16m2))) +vuint16m2_t vssrl(vuint16m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vx_u16m2_m))) +vuint16m2_t vssrl(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vx_u16m4))) +vuint16m4_t vssrl(vuint16m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vx_u16m4_m))) +vuint16m4_t vssrl(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vx_u16m8))) +vuint16m8_t vssrl(vuint16m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vx_u16m8_m))) +vuint16m8_t vssrl(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vx_u16mf2))) +vuint16mf2_t vssrl(vuint16mf2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vx_u16mf2_m))) +vuint16mf2_t vssrl(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vx_u16mf4))) +vuint16mf4_t vssrl(vuint16mf4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vx_u16mf4_m))) +vuint16mf4_t vssrl(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vx_u32m1))) +vuint32m1_t vssrl(vuint32m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vx_u32m1_m))) +vuint32m1_t vssrl(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vx_u32m2))) +vuint32m2_t vssrl(vuint32m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vx_u32m2_m))) +vuint32m2_t vssrl(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vx_u32m4))) +vuint32m4_t vssrl(vuint32m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vx_u32m4_m))) +vuint32m4_t vssrl(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vx_u32m8))) +vuint32m8_t vssrl(vuint32m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vx_u32m8_m))) +vuint32m8_t vssrl(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vx_u32mf2))) +vuint32mf2_t vssrl(vuint32mf2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vx_u32mf2_m))) +vuint32mf2_t vssrl(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vx_u64m1))) +vuint64m1_t vssrl(vuint64m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vx_u64m1_m))) +vuint64m1_t vssrl(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vx_u64m2))) +vuint64m2_t vssrl(vuint64m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vx_u64m2_m))) +vuint64m2_t vssrl(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vx_u64m4))) +vuint64m4_t vssrl(vuint64m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vx_u64m4_m))) +vuint64m4_t vssrl(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vx_u64m8))) +vuint64m8_t vssrl(vuint64m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssrl_vx_u64m8_m))) +vuint64m8_t vssrl(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vv_i8m1))) +vint8m1_t vssra(vint8m1_t op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vv_i8m1_m))) +vint8m1_t vssra(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vv_i8m2))) +vint8m2_t vssra(vint8m2_t op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vv_i8m2_m))) +vint8m2_t vssra(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vv_i8m4))) +vint8m4_t vssra(vint8m4_t op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vv_i8m4_m))) +vint8m4_t vssra(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vv_i8m8))) +vint8m8_t vssra(vint8m8_t op0, vuint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vv_i8m8_m))) +vint8m8_t vssra(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, vuint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vv_i8mf2))) +vint8mf2_t vssra(vint8mf2_t op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vv_i8mf2_m))) +vint8mf2_t vssra(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vv_i8mf4))) +vint8mf4_t vssra(vint8mf4_t op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vv_i8mf4_m))) +vint8mf4_t vssra(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vv_i8mf8))) +vint8mf8_t vssra(vint8mf8_t op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vv_i8mf8_m))) +vint8mf8_t vssra(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vv_i16m1))) +vint16m1_t vssra(vint16m1_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vv_i16m1_m))) +vint16m1_t vssra(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vv_i16m2))) +vint16m2_t vssra(vint16m2_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vv_i16m2_m))) +vint16m2_t vssra(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vv_i16m4))) +vint16m4_t vssra(vint16m4_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vv_i16m4_m))) +vint16m4_t vssra(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vv_i16m8))) +vint16m8_t vssra(vint16m8_t op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vv_i16m8_m))) +vint16m8_t vssra(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vv_i16mf2))) +vint16mf2_t vssra(vint16mf2_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vv_i16mf2_m))) +vint16mf2_t vssra(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vv_i16mf4))) +vint16mf4_t vssra(vint16mf4_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vv_i16mf4_m))) +vint16mf4_t vssra(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vv_i32m1))) +vint32m1_t vssra(vint32m1_t op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vv_i32m1_m))) +vint32m1_t vssra(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vv_i32m2))) +vint32m2_t vssra(vint32m2_t op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vv_i32m2_m))) +vint32m2_t vssra(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vv_i32m4))) +vint32m4_t vssra(vint32m4_t op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vv_i32m4_m))) +vint32m4_t vssra(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vv_i32m8))) +vint32m8_t vssra(vint32m8_t op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vv_i32m8_m))) +vint32m8_t vssra(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vv_i32mf2))) +vint32mf2_t vssra(vint32mf2_t op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vv_i32mf2_m))) +vint32mf2_t vssra(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vv_i64m1))) +vint64m1_t vssra(vint64m1_t op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vv_i64m1_m))) +vint64m1_t vssra(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vv_i64m2))) +vint64m2_t vssra(vint64m2_t op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vv_i64m2_m))) +vint64m2_t vssra(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vv_i64m4))) +vint64m4_t vssra(vint64m4_t op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vv_i64m4_m))) +vint64m4_t vssra(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vv_i64m8))) +vint64m8_t vssra(vint64m8_t op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vv_i64m8_m))) +vint64m8_t vssra(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vx_i8m1))) +vint8m1_t vssra(vint8m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vx_i8m1_m))) +vint8m1_t vssra(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vx_i8m2))) +vint8m2_t vssra(vint8m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vx_i8m2_m))) +vint8m2_t vssra(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vx_i8m4))) +vint8m4_t vssra(vint8m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vx_i8m4_m))) +vint8m4_t vssra(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vx_i8m8))) +vint8m8_t vssra(vint8m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vx_i8m8_m))) +vint8m8_t vssra(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vx_i8mf2))) +vint8mf2_t vssra(vint8mf2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vx_i8mf2_m))) +vint8mf2_t vssra(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vx_i8mf4))) +vint8mf4_t vssra(vint8mf4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vx_i8mf4_m))) +vint8mf4_t vssra(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vx_i8mf8))) +vint8mf8_t vssra(vint8mf8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vx_i8mf8_m))) +vint8mf8_t vssra(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vx_i16m1))) +vint16m1_t vssra(vint16m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vx_i16m1_m))) +vint16m1_t vssra(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vx_i16m2))) +vint16m2_t vssra(vint16m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vx_i16m2_m))) +vint16m2_t vssra(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vx_i16m4))) +vint16m4_t vssra(vint16m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vx_i16m4_m))) +vint16m4_t vssra(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vx_i16m8))) +vint16m8_t vssra(vint16m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vx_i16m8_m))) +vint16m8_t vssra(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vx_i16mf2))) +vint16mf2_t vssra(vint16mf2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vx_i16mf2_m))) +vint16mf2_t vssra(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vx_i16mf4))) +vint16mf4_t vssra(vint16mf4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vx_i16mf4_m))) +vint16mf4_t vssra(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vx_i32m1))) +vint32m1_t vssra(vint32m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vx_i32m1_m))) +vint32m1_t vssra(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vx_i32m2))) +vint32m2_t vssra(vint32m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vx_i32m2_m))) +vint32m2_t vssra(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vx_i32m4))) +vint32m4_t vssra(vint32m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vx_i32m4_m))) +vint32m4_t vssra(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vx_i32m8))) +vint32m8_t vssra(vint32m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vx_i32m8_m))) +vint32m8_t vssra(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vx_i32mf2))) +vint32mf2_t vssra(vint32mf2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vx_i32mf2_m))) +vint32mf2_t vssra(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vx_i64m1))) +vint64m1_t vssra(vint64m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vx_i64m1_m))) +vint64m1_t vssra(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vx_i64m2))) +vint64m2_t vssra(vint64m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vx_i64m2_m))) +vint64m2_t vssra(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vx_i64m4))) +vint64m4_t vssra(vint64m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vx_i64m4_m))) +vint64m4_t vssra(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vx_i64m8))) +vint64m8_t vssra(vint64m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssra_vx_i64m8_m))) +vint64m8_t vssra(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wv_u8m1))) +vuint8m1_t vnclipu(vuint16m2_t op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wv_u8m1_m))) +vuint8m1_t vnclipu(vbool8_t op0, vuint8m1_t op1, vuint16m2_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wv_u8m2))) +vuint8m2_t vnclipu(vuint16m4_t op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wv_u8m2_m))) +vuint8m2_t vnclipu(vbool4_t op0, vuint8m2_t op1, vuint16m4_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wv_u8m4))) +vuint8m4_t vnclipu(vuint16m8_t op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wv_u8m4_m))) +vuint8m4_t vnclipu(vbool2_t op0, vuint8m4_t op1, vuint16m8_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wv_u8mf2))) +vuint8mf2_t vnclipu(vuint16m1_t op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wv_u8mf2_m))) +vuint8mf2_t vnclipu(vbool16_t op0, vuint8mf2_t op1, vuint16m1_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wv_u8mf4))) +vuint8mf4_t vnclipu(vuint16mf2_t op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wv_u8mf4_m))) +vuint8mf4_t vnclipu(vbool32_t op0, vuint8mf4_t op1, vuint16mf2_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wv_u8mf8))) +vuint8mf8_t vnclipu(vuint16mf4_t op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wv_u8mf8_m))) +vuint8mf8_t vnclipu(vbool64_t op0, vuint8mf8_t op1, vuint16mf4_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wv_u16m1))) +vuint16m1_t vnclipu(vuint32m2_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wv_u16m1_m))) +vuint16m1_t vnclipu(vbool16_t op0, vuint16m1_t op1, vuint32m2_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wv_u16m2))) +vuint16m2_t vnclipu(vuint32m4_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wv_u16m2_m))) +vuint16m2_t vnclipu(vbool8_t op0, vuint16m2_t op1, vuint32m4_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wv_u16m4))) +vuint16m4_t vnclipu(vuint32m8_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wv_u16m4_m))) +vuint16m4_t vnclipu(vbool4_t op0, vuint16m4_t op1, vuint32m8_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wv_u16mf2))) +vuint16mf2_t vnclipu(vuint32m1_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wv_u16mf2_m))) +vuint16mf2_t vnclipu(vbool32_t op0, vuint16mf2_t op1, vuint32m1_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wv_u16mf4))) +vuint16mf4_t vnclipu(vuint32mf2_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wv_u16mf4_m))) +vuint16mf4_t vnclipu(vbool64_t op0, vuint16mf4_t op1, vuint32mf2_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wv_u32m1))) +vuint32m1_t vnclipu(vuint64m2_t op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wv_u32m1_m))) +vuint32m1_t vnclipu(vbool32_t op0, vuint32m1_t op1, vuint64m2_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wv_u32m2))) +vuint32m2_t vnclipu(vuint64m4_t op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wv_u32m2_m))) +vuint32m2_t vnclipu(vbool16_t op0, vuint32m2_t op1, vuint64m4_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wv_u32m4))) +vuint32m4_t vnclipu(vuint64m8_t op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wv_u32m4_m))) +vuint32m4_t vnclipu(vbool8_t op0, vuint32m4_t op1, vuint64m8_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wv_u32mf2))) +vuint32mf2_t vnclipu(vuint64m1_t op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wv_u32mf2_m))) +vuint32mf2_t vnclipu(vbool64_t op0, vuint32mf2_t op1, vuint64m1_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wx_u8m1))) +vuint8m1_t vnclipu(vuint16m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wx_u8m1_m))) +vuint8m1_t vnclipu(vbool8_t op0, vuint8m1_t op1, vuint16m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wx_u8m2))) +vuint8m2_t vnclipu(vuint16m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wx_u8m2_m))) +vuint8m2_t vnclipu(vbool4_t op0, vuint8m2_t op1, vuint16m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wx_u8m4))) +vuint8m4_t vnclipu(vuint16m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wx_u8m4_m))) +vuint8m4_t vnclipu(vbool2_t op0, vuint8m4_t op1, vuint16m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wx_u8mf2))) +vuint8mf2_t vnclipu(vuint16m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wx_u8mf2_m))) +vuint8mf2_t vnclipu(vbool16_t op0, vuint8mf2_t op1, vuint16m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wx_u8mf4))) +vuint8mf4_t vnclipu(vuint16mf2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wx_u8mf4_m))) +vuint8mf4_t vnclipu(vbool32_t op0, vuint8mf4_t op1, vuint16mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wx_u8mf8))) +vuint8mf8_t vnclipu(vuint16mf4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wx_u8mf8_m))) +vuint8mf8_t vnclipu(vbool64_t op0, vuint8mf8_t op1, vuint16mf4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wx_u16m1))) +vuint16m1_t vnclipu(vuint32m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wx_u16m1_m))) +vuint16m1_t vnclipu(vbool16_t op0, vuint16m1_t op1, vuint32m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wx_u16m2))) +vuint16m2_t vnclipu(vuint32m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wx_u16m2_m))) +vuint16m2_t vnclipu(vbool8_t op0, vuint16m2_t op1, vuint32m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wx_u16m4))) +vuint16m4_t vnclipu(vuint32m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wx_u16m4_m))) +vuint16m4_t vnclipu(vbool4_t op0, vuint16m4_t op1, vuint32m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wx_u16mf2))) +vuint16mf2_t vnclipu(vuint32m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wx_u16mf2_m))) +vuint16mf2_t vnclipu(vbool32_t op0, vuint16mf2_t op1, vuint32m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wx_u16mf4))) +vuint16mf4_t vnclipu(vuint32mf2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wx_u16mf4_m))) +vuint16mf4_t vnclipu(vbool64_t op0, vuint16mf4_t op1, vuint32mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wx_u32m1))) +vuint32m1_t vnclipu(vuint64m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wx_u32m1_m))) +vuint32m1_t vnclipu(vbool32_t op0, vuint32m1_t op1, vuint64m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wx_u32m2))) +vuint32m2_t vnclipu(vuint64m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wx_u32m2_m))) +vuint32m2_t vnclipu(vbool16_t op0, vuint32m2_t op1, vuint64m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wx_u32m4))) +vuint32m4_t vnclipu(vuint64m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wx_u32m4_m))) +vuint32m4_t vnclipu(vbool8_t op0, vuint32m4_t op1, vuint64m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wx_u32mf2))) +vuint32mf2_t vnclipu(vuint64m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclipu_wx_u32mf2_m))) +vuint32mf2_t vnclipu(vbool64_t op0, vuint32mf2_t op1, vuint64m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_u32m1))) +void vsuxei64(uint32_t * op0, vuint64m2_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_u32m1_m))) +void vsuxei64(vbool32_t op0, uint32_t * op1, vuint64m2_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_u32m2))) +void vsuxei64(uint32_t * op0, vuint64m4_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_u32m2_m))) +void vsuxei64(vbool16_t op0, uint32_t * op1, vuint64m4_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_u32m4))) +void vsuxei64(uint32_t * op0, vuint64m8_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_u32m4_m))) +void vsuxei64(vbool8_t op0, uint32_t * op1, vuint64m8_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_u32mf2))) +void vsuxei64(uint32_t * op0, vuint64m1_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_u32mf2_m))) +void vsuxei64(vbool64_t op0, uint32_t * op1, vuint64m1_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wv_i8m1))) +vint8m1_t vnclip(vint16m2_t op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wv_i8m1_m))) +vint8m1_t vnclip(vbool8_t op0, vint8m1_t op1, vint16m2_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wv_i8m2))) +vint8m2_t vnclip(vint16m4_t op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wv_i8m2_m))) +vint8m2_t vnclip(vbool4_t op0, vint8m2_t op1, vint16m4_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wv_i8m4))) +vint8m4_t vnclip(vint16m8_t op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wv_i8m4_m))) +vint8m4_t vnclip(vbool2_t op0, vint8m4_t op1, vint16m8_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wv_i8mf2))) +vint8mf2_t vnclip(vint16m1_t op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wv_i8mf2_m))) +vint8mf2_t vnclip(vbool16_t op0, vint8mf2_t op1, vint16m1_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wv_i8mf4))) +vint8mf4_t vnclip(vint16mf2_t op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wv_i8mf4_m))) +vint8mf4_t vnclip(vbool32_t op0, vint8mf4_t op1, vint16mf2_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wv_i8mf8))) +vint8mf8_t vnclip(vint16mf4_t op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wv_i8mf8_m))) +vint8mf8_t vnclip(vbool64_t op0, vint8mf8_t op1, vint16mf4_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wv_i16m1))) +vint16m1_t vnclip(vint32m2_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wv_i16m1_m))) +vint16m1_t vnclip(vbool16_t op0, vint16m1_t op1, vint32m2_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wv_i16m2))) +vint16m2_t vnclip(vint32m4_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wv_i16m2_m))) +vint16m2_t vnclip(vbool8_t op0, vint16m2_t op1, vint32m4_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wv_i16m4))) +vint16m4_t vnclip(vint32m8_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wv_i16m4_m))) +vint16m4_t vnclip(vbool4_t op0, vint16m4_t op1, vint32m8_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wv_i16mf2))) +vint16mf2_t vnclip(vint32m1_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wv_i16mf2_m))) +vint16mf2_t vnclip(vbool32_t op0, vint16mf2_t op1, vint32m1_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wv_i16mf4))) +vint16mf4_t vnclip(vint32mf2_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wv_i16mf4_m))) +vint16mf4_t vnclip(vbool64_t op0, vint16mf4_t op1, vint32mf2_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wv_i32m1))) +vint32m1_t vnclip(vint64m2_t op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wv_i32m1_m))) +vint32m1_t vnclip(vbool32_t op0, vint32m1_t op1, vint64m2_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wv_i32m2))) +vint32m2_t vnclip(vint64m4_t op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wv_i32m2_m))) +vint32m2_t vnclip(vbool16_t op0, vint32m2_t op1, vint64m4_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wv_i32m4))) +vint32m4_t vnclip(vint64m8_t op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wv_i32m4_m))) +vint32m4_t vnclip(vbool8_t op0, vint32m4_t op1, vint64m8_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wv_i32mf2))) +vint32mf2_t vnclip(vint64m1_t op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wv_i32mf2_m))) +vint32mf2_t vnclip(vbool64_t op0, vint32mf2_t op1, vint64m1_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wx_i8m1))) +vint8m1_t vnclip(vint16m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wx_i8m1_m))) +vint8m1_t vnclip(vbool8_t op0, vint8m1_t op1, vint16m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wx_i8m2))) +vint8m2_t vnclip(vint16m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wx_i8m2_m))) +vint8m2_t vnclip(vbool4_t op0, vint8m2_t op1, vint16m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wx_i8m4))) +vint8m4_t vnclip(vint16m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wx_i8m4_m))) +vint8m4_t vnclip(vbool2_t op0, vint8m4_t op1, vint16m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wx_i8mf2))) +vint8mf2_t vnclip(vint16m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wx_i8mf2_m))) +vint8mf2_t vnclip(vbool16_t op0, vint8mf2_t op1, vint16m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wx_i8mf4))) +vint8mf4_t vnclip(vint16mf2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wx_i8mf4_m))) +vint8mf4_t vnclip(vbool32_t op0, vint8mf4_t op1, vint16mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wx_i8mf8))) +vint8mf8_t vnclip(vint16mf4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wx_i8mf8_m))) +vint8mf8_t vnclip(vbool64_t op0, vint8mf8_t op1, vint16mf4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wx_i16m1))) +vint16m1_t vnclip(vint32m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wx_i16m1_m))) +vint16m1_t vnclip(vbool16_t op0, vint16m1_t op1, vint32m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wx_i16m2))) +vint16m2_t vnclip(vint32m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wx_i16m2_m))) +vint16m2_t vnclip(vbool8_t op0, vint16m2_t op1, vint32m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wx_i16m4))) +vint16m4_t vnclip(vint32m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wx_i16m4_m))) +vint16m4_t vnclip(vbool4_t op0, vint16m4_t op1, vint32m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wx_i16mf2))) +vint16mf2_t vnclip(vint32m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wx_i16mf2_m))) +vint16mf2_t vnclip(vbool32_t op0, vint16mf2_t op1, vint32m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wx_i16mf4))) +vint16mf4_t vnclip(vint32mf2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wx_i16mf4_m))) +vint16mf4_t vnclip(vbool64_t op0, vint16mf4_t op1, vint32mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wx_i32m1))) +vint32m1_t vnclip(vint64m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wx_i32m1_m))) +vint32m1_t vnclip(vbool32_t op0, vint32m1_t op1, vint64m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wx_i32m2))) +vint32m2_t vnclip(vint64m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wx_i32m2_m))) +vint32m2_t vnclip(vbool16_t op0, vint32m2_t op1, vint64m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wx_i32m4))) +vint32m4_t vnclip(vint64m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wx_i32m4_m))) +vint32m4_t vnclip(vbool8_t op0, vint32m4_t op1, vint64m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wx_i32mf2))) +vint32mf2_t vnclip(vint64m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnclip_wx_i32mf2_m))) +vint32mf2_t vnclip(vbool64_t op0, vint32mf2_t op1, vint64m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_i64m1))) +void vsuxei8(int64_t * op0, vuint8mf8_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_i64m1_m))) +void vsuxei8(vbool64_t op0, int64_t * op1, vuint8mf8_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_i64m2))) +void vsuxei8(int64_t * op0, vuint8mf4_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_i64m2_m))) +void vsuxei8(vbool32_t op0, int64_t * op1, vuint8mf4_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_i64m4))) +void vsuxei8(int64_t * op0, vuint8mf2_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_i64m4_m))) +void vsuxei8(vbool16_t op0, int64_t * op1, vuint8mf2_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_i64m8))) +void vsuxei8(int64_t * op0, vuint8m1_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_i64m8_m))) +void vsuxei8(vbool8_t op0, int64_t * op1, vuint8m1_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_u64m1))) +void vsuxei8(uint64_t * op0, vuint8mf8_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_u64m1_m))) +void vsuxei8(vbool64_t op0, uint64_t * op1, vuint8mf8_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_u64m2))) +void vsuxei8(uint64_t * op0, vuint8mf4_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_u64m2_m))) +void vsuxei8(vbool32_t op0, uint64_t * op1, vuint8mf4_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_u64m4))) +void vsuxei8(uint64_t * op0, vuint8mf2_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_u64m4_m))) +void vsuxei8(vbool16_t op0, uint64_t * op1, vuint8mf2_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_u64m8))) +void vsuxei8(uint64_t * op0, vuint8m1_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_u64m8_m))) +void vsuxei8(vbool8_t op0, uint64_t * op1, vuint8m1_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_i64m1))) +void vsuxei16(int64_t * op0, vuint16mf4_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_i64m1_m))) +void vsuxei16(vbool64_t op0, int64_t * op1, vuint16mf4_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_i64m2))) +void vsuxei16(int64_t * op0, vuint16mf2_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_i64m2_m))) +void vsuxei16(vbool32_t op0, int64_t * op1, vuint16mf2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_i64m4))) +void vsuxei16(int64_t * op0, vuint16m1_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_i64m4_m))) +void vsuxei16(vbool16_t op0, int64_t * op1, vuint16m1_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_i64m8))) +void vsuxei16(int64_t * op0, vuint16m2_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_i64m8_m))) +void vsuxei16(vbool8_t op0, int64_t * op1, vuint16m2_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_u64m1))) +void vsuxei16(uint64_t * op0, vuint16mf4_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_u64m1_m))) +void vsuxei16(vbool64_t op0, uint64_t * op1, vuint16mf4_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_u64m2))) +void vsuxei16(uint64_t * op0, vuint16mf2_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_u64m2_m))) +void vsuxei16(vbool32_t op0, uint64_t * op1, vuint16mf2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_u64m4))) +void vsuxei16(uint64_t * op0, vuint16m1_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_u64m4_m))) +void vsuxei16(vbool16_t op0, uint64_t * op1, vuint16m1_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_u64m8))) +void vsuxei16(uint64_t * op0, vuint16m2_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_u64m8_m))) +void vsuxei16(vbool8_t op0, uint64_t * op1, vuint16m2_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_i64m1))) +void vsuxei32(int64_t * op0, vuint32mf2_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_i64m1_m))) +void vsuxei32(vbool64_t op0, int64_t * op1, vuint32mf2_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_i64m2))) +void vsuxei32(int64_t * op0, vuint32m1_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_i64m2_m))) +void vsuxei32(vbool32_t op0, int64_t * op1, vuint32m1_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_i64m4))) +void vsuxei32(int64_t * op0, vuint32m2_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_i64m4_m))) +void vsuxei32(vbool16_t op0, int64_t * op1, vuint32m2_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_i64m8))) +void vsuxei32(int64_t * op0, vuint32m4_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_i64m8_m))) +void vsuxei32(vbool8_t op0, int64_t * op1, vuint32m4_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_u64m1))) +void vsuxei32(uint64_t * op0, vuint32mf2_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_u64m1_m))) +void vsuxei32(vbool64_t op0, uint64_t * op1, vuint32mf2_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_u64m2))) +void vsuxei32(uint64_t * op0, vuint32m1_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_u64m2_m))) +void vsuxei32(vbool32_t op0, uint64_t * op1, vuint32m1_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_u64m4))) +void vsuxei32(uint64_t * op0, vuint32m2_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_u64m4_m))) +void vsuxei32(vbool16_t op0, uint64_t * op1, vuint32m2_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_u64m8))) +void vsuxei32(uint64_t * op0, vuint32m4_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_u64m8_m))) +void vsuxei32(vbool8_t op0, uint64_t * op1, vuint32m4_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_i64m1))) +void vsuxei64(int64_t * op0, vuint64m1_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_i64m1_m))) +void vsuxei64(vbool64_t op0, int64_t * op1, vuint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_i64m2))) +void vsuxei64(int64_t * op0, vuint64m2_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_i64m2_m))) +void vsuxei64(vbool32_t op0, int64_t * op1, vuint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_i64m4))) +void vsuxei64(int64_t * op0, vuint64m4_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_i64m4_m))) +void vsuxei64(vbool16_t op0, int64_t * op1, vuint64m4_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_i64m8))) +void vsuxei64(int64_t * op0, vuint64m8_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_i64m8_m))) +void vsuxei64(vbool8_t op0, int64_t * op1, vuint64m8_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_i8m1_i8m1))) +vint8m1_t vredsum(vint8m1_t op0, vint8m1_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_i8m1_i8m1_m))) +vint8m1_t vredsum(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_i8m2_i8m1))) +vint8m1_t vredsum(vint8m1_t op0, vint8m2_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_i8m2_i8m1_m))) +vint8m1_t vredsum(vbool4_t op0, vint8m1_t op1, vint8m2_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_i8m4_i8m1))) +vint8m1_t vredsum(vint8m1_t op0, vint8m4_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_i8m4_i8m1_m))) +vint8m1_t vredsum(vbool2_t op0, vint8m1_t op1, vint8m4_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_i8m8_i8m1))) +vint8m1_t vredsum(vint8m1_t op0, vint8m8_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_i8m8_i8m1_m))) +vint8m1_t vredsum(vbool1_t op0, vint8m1_t op1, vint8m8_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_i8mf2_i8m1))) +vint8m1_t vredsum(vint8m1_t op0, vint8mf2_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_i8mf2_i8m1_m))) +vint8m1_t vredsum(vbool16_t op0, vint8m1_t op1, vint8mf2_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_i8mf4_i8m1))) +vint8m1_t vredsum(vint8m1_t op0, vint8mf4_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_i8mf4_i8m1_m))) +vint8m1_t vredsum(vbool32_t op0, vint8m1_t op1, vint8mf4_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_i8mf8_i8m1))) +vint8m1_t vredsum(vint8m1_t op0, vint8mf8_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_i8mf8_i8m1_m))) +vint8m1_t vredsum(vbool64_t op0, vint8m1_t op1, vint8mf8_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_i16m1_i16m1))) +vint16m1_t vredsum(vint16m1_t op0, vint16m1_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_i16m1_i16m1_m))) +vint16m1_t vredsum(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_i16m2_i16m1))) +vint16m1_t vredsum(vint16m1_t op0, vint16m2_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_i16m2_i16m1_m))) +vint16m1_t vredsum(vbool8_t op0, vint16m1_t op1, vint16m2_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_i16m4_i16m1))) +vint16m1_t vredsum(vint16m1_t op0, vint16m4_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_i16m4_i16m1_m))) +vint16m1_t vredsum(vbool4_t op0, vint16m1_t op1, vint16m4_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_i16m8_i16m1))) +vint16m1_t vredsum(vint16m1_t op0, vint16m8_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_i16m8_i16m1_m))) +vint16m1_t vredsum(vbool2_t op0, vint16m1_t op1, vint16m8_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_i16mf2_i16m1))) +vint16m1_t vredsum(vint16m1_t op0, vint16mf2_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_i16mf2_i16m1_m))) +vint16m1_t vredsum(vbool32_t op0, vint16m1_t op1, vint16mf2_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_i16mf4_i16m1))) +vint16m1_t vredsum(vint16m1_t op0, vint16mf4_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_i16mf4_i16m1_m))) +vint16m1_t vredsum(vbool64_t op0, vint16m1_t op1, vint16mf4_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_i32m1_i32m1))) +vint32m1_t vredsum(vint32m1_t op0, vint32m1_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_i32m1_i32m1_m))) +vint32m1_t vredsum(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_i32m2_i32m1))) +vint32m1_t vredsum(vint32m1_t op0, vint32m2_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_i32m2_i32m1_m))) +vint32m1_t vredsum(vbool16_t op0, vint32m1_t op1, vint32m2_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_i32m4_i32m1))) +vint32m1_t vredsum(vint32m1_t op0, vint32m4_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_i32m4_i32m1_m))) +vint32m1_t vredsum(vbool8_t op0, vint32m1_t op1, vint32m4_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_i32m8_i32m1))) +vint32m1_t vredsum(vint32m1_t op0, vint32m8_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_i32m8_i32m1_m))) +vint32m1_t vredsum(vbool4_t op0, vint32m1_t op1, vint32m8_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_i32mf2_i32m1))) +vint32m1_t vredsum(vint32m1_t op0, vint32mf2_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_i32mf2_i32m1_m))) +vint32m1_t vredsum(vbool64_t op0, vint32m1_t op1, vint32mf2_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_i64m1_i64m1))) +vint64m1_t vredsum(vint64m1_t op0, vint64m1_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_i64m1_i64m1_m))) +vint64m1_t vredsum(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_i64m2_i64m1))) +vint64m1_t vredsum(vint64m1_t op0, vint64m2_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_i64m2_i64m1_m))) +vint64m1_t vredsum(vbool32_t op0, vint64m1_t op1, vint64m2_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_i64m4_i64m1))) +vint64m1_t vredsum(vint64m1_t op0, vint64m4_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_i64m4_i64m1_m))) +vint64m1_t vredsum(vbool16_t op0, vint64m1_t op1, vint64m4_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_i64m8_i64m1))) +vint64m1_t vredsum(vint64m1_t op0, vint64m8_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_i64m8_i64m1_m))) +vint64m1_t vredsum(vbool8_t op0, vint64m1_t op1, vint64m8_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_u8m1_u8m1))) +vuint8m1_t vredsum(vuint8m1_t op0, vuint8m1_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_u8m1_u8m1_m))) +vuint8m1_t vredsum(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_u8m2_u8m1))) +vuint8m1_t vredsum(vuint8m1_t op0, vuint8m2_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_u8m2_u8m1_m))) +vuint8m1_t vredsum(vbool4_t op0, vuint8m1_t op1, vuint8m2_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_u8m4_u8m1))) +vuint8m1_t vredsum(vuint8m1_t op0, vuint8m4_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_u8m4_u8m1_m))) +vuint8m1_t vredsum(vbool2_t op0, vuint8m1_t op1, vuint8m4_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_u8m8_u8m1))) +vuint8m1_t vredsum(vuint8m1_t op0, vuint8m8_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_u8m8_u8m1_m))) +vuint8m1_t vredsum(vbool1_t op0, vuint8m1_t op1, vuint8m8_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_u8mf2_u8m1))) +vuint8m1_t vredsum(vuint8m1_t op0, vuint8mf2_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_u8mf2_u8m1_m))) +vuint8m1_t vredsum(vbool16_t op0, vuint8m1_t op1, vuint8mf2_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_u8mf4_u8m1))) +vuint8m1_t vredsum(vuint8m1_t op0, vuint8mf4_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_u8mf4_u8m1_m))) +vuint8m1_t vredsum(vbool32_t op0, vuint8m1_t op1, vuint8mf4_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_u8mf8_u8m1))) +vuint8m1_t vredsum(vuint8m1_t op0, vuint8mf8_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_u8mf8_u8m1_m))) +vuint8m1_t vredsum(vbool64_t op0, vuint8m1_t op1, vuint8mf8_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_u16m1_u16m1))) +vuint16m1_t vredsum(vuint16m1_t op0, vuint16m1_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_u16m1_u16m1_m))) +vuint16m1_t vredsum(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_u16m2_u16m1))) +vuint16m1_t vredsum(vuint16m1_t op0, vuint16m2_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_u16m2_u16m1_m))) +vuint16m1_t vredsum(vbool8_t op0, vuint16m1_t op1, vuint16m2_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_u16m4_u16m1))) +vuint16m1_t vredsum(vuint16m1_t op0, vuint16m4_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_u16m4_u16m1_m))) +vuint16m1_t vredsum(vbool4_t op0, vuint16m1_t op1, vuint16m4_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_u16m8_u16m1))) +vuint16m1_t vredsum(vuint16m1_t op0, vuint16m8_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_u16m8_u16m1_m))) +vuint16m1_t vredsum(vbool2_t op0, vuint16m1_t op1, vuint16m8_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_u16mf2_u16m1))) +vuint16m1_t vredsum(vuint16m1_t op0, vuint16mf2_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_u16mf2_u16m1_m))) +vuint16m1_t vredsum(vbool32_t op0, vuint16m1_t op1, vuint16mf2_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_u16mf4_u16m1))) +vuint16m1_t vredsum(vuint16m1_t op0, vuint16mf4_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_u16mf4_u16m1_m))) +vuint16m1_t vredsum(vbool64_t op0, vuint16m1_t op1, vuint16mf4_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_u32m1_u32m1))) +vuint32m1_t vredsum(vuint32m1_t op0, vuint32m1_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_u32m1_u32m1_m))) +vuint32m1_t vredsum(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_u32m2_u32m1))) +vuint32m1_t vredsum(vuint32m1_t op0, vuint32m2_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_u32m2_u32m1_m))) +vuint32m1_t vredsum(vbool16_t op0, vuint32m1_t op1, vuint32m2_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_u32m4_u32m1))) +vuint32m1_t vredsum(vuint32m1_t op0, vuint32m4_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_u32m4_u32m1_m))) +vuint32m1_t vredsum(vbool8_t op0, vuint32m1_t op1, vuint32m4_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_u32m8_u32m1))) +vuint32m1_t vredsum(vuint32m1_t op0, vuint32m8_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_u32m8_u32m1_m))) +vuint32m1_t vredsum(vbool4_t op0, vuint32m1_t op1, vuint32m8_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_u32mf2_u32m1))) +vuint32m1_t vredsum(vuint32m1_t op0, vuint32mf2_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_u32mf2_u32m1_m))) +vuint32m1_t vredsum(vbool64_t op0, vuint32m1_t op1, vuint32mf2_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_u64m1_u64m1))) +vuint64m1_t vredsum(vuint64m1_t op0, vuint64m1_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_u64m1_u64m1_m))) +vuint64m1_t vredsum(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_u64m2_u64m1))) +vuint64m1_t vredsum(vuint64m1_t op0, vuint64m2_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_u64m2_u64m1_m))) +vuint64m1_t vredsum(vbool32_t op0, vuint64m1_t op1, vuint64m2_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_u64m4_u64m1))) +vuint64m1_t vredsum(vuint64m1_t op0, vuint64m4_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_u64m4_u64m1_m))) +vuint64m1_t vredsum(vbool16_t op0, vuint64m1_t op1, vuint64m4_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_u64m8_u64m1))) +vuint64m1_t vredsum(vuint64m1_t op0, vuint64m8_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredsum_vs_u64m8_u64m1_m))) +vuint64m1_t vredsum(vbool8_t op0, vuint64m1_t op1, vuint64m8_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmaxu_vs_u8m1_u8m1))) +vuint8m1_t vredmaxu(vuint8m1_t op0, vuint8m1_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmaxu_vs_u8m1_u8m1_m))) +vuint8m1_t vredmaxu(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmaxu_vs_u8m2_u8m1))) +vuint8m1_t vredmaxu(vuint8m1_t op0, vuint8m2_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmaxu_vs_u8m2_u8m1_m))) +vuint8m1_t vredmaxu(vbool4_t op0, vuint8m1_t op1, vuint8m2_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmaxu_vs_u8m4_u8m1))) +vuint8m1_t vredmaxu(vuint8m1_t op0, vuint8m4_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmaxu_vs_u8m4_u8m1_m))) +vuint8m1_t vredmaxu(vbool2_t op0, vuint8m1_t op1, vuint8m4_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmaxu_vs_u8m8_u8m1))) +vuint8m1_t vredmaxu(vuint8m1_t op0, vuint8m8_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmaxu_vs_u8m8_u8m1_m))) +vuint8m1_t vredmaxu(vbool1_t op0, vuint8m1_t op1, vuint8m8_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmaxu_vs_u8mf2_u8m1))) +vuint8m1_t vredmaxu(vuint8m1_t op0, vuint8mf2_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmaxu_vs_u8mf2_u8m1_m))) +vuint8m1_t vredmaxu(vbool16_t op0, vuint8m1_t op1, vuint8mf2_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmaxu_vs_u8mf4_u8m1))) +vuint8m1_t vredmaxu(vuint8m1_t op0, vuint8mf4_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmaxu_vs_u8mf4_u8m1_m))) +vuint8m1_t vredmaxu(vbool32_t op0, vuint8m1_t op1, vuint8mf4_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmaxu_vs_u8mf8_u8m1))) +vuint8m1_t vredmaxu(vuint8m1_t op0, vuint8mf8_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmaxu_vs_u8mf8_u8m1_m))) +vuint8m1_t vredmaxu(vbool64_t op0, vuint8m1_t op1, vuint8mf8_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmaxu_vs_u16m1_u16m1))) +vuint16m1_t vredmaxu(vuint16m1_t op0, vuint16m1_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmaxu_vs_u16m1_u16m1_m))) +vuint16m1_t vredmaxu(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmaxu_vs_u16m2_u16m1))) +vuint16m1_t vredmaxu(vuint16m1_t op0, vuint16m2_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmaxu_vs_u16m2_u16m1_m))) +vuint16m1_t vredmaxu(vbool8_t op0, vuint16m1_t op1, vuint16m2_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmaxu_vs_u16m4_u16m1))) +vuint16m1_t vredmaxu(vuint16m1_t op0, vuint16m4_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmaxu_vs_u16m4_u16m1_m))) +vuint16m1_t vredmaxu(vbool4_t op0, vuint16m1_t op1, vuint16m4_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmaxu_vs_u16m8_u16m1))) +vuint16m1_t vredmaxu(vuint16m1_t op0, vuint16m8_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmaxu_vs_u16m8_u16m1_m))) +vuint16m1_t vredmaxu(vbool2_t op0, vuint16m1_t op1, vuint16m8_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmaxu_vs_u16mf2_u16m1))) +vuint16m1_t vredmaxu(vuint16m1_t op0, vuint16mf2_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmaxu_vs_u16mf2_u16m1_m))) +vuint16m1_t vredmaxu(vbool32_t op0, vuint16m1_t op1, vuint16mf2_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmaxu_vs_u16mf4_u16m1))) +vuint16m1_t vredmaxu(vuint16m1_t op0, vuint16mf4_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmaxu_vs_u16mf4_u16m1_m))) +vuint16m1_t vredmaxu(vbool64_t op0, vuint16m1_t op1, vuint16mf4_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmaxu_vs_u32m1_u32m1))) +vuint32m1_t vredmaxu(vuint32m1_t op0, vuint32m1_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmaxu_vs_u32m1_u32m1_m))) +vuint32m1_t vredmaxu(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmaxu_vs_u32m2_u32m1))) +vuint32m1_t vredmaxu(vuint32m1_t op0, vuint32m2_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmaxu_vs_u32m2_u32m1_m))) +vuint32m1_t vredmaxu(vbool16_t op0, vuint32m1_t op1, vuint32m2_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmaxu_vs_u32m4_u32m1))) +vuint32m1_t vredmaxu(vuint32m1_t op0, vuint32m4_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmaxu_vs_u32m4_u32m1_m))) +vuint32m1_t vredmaxu(vbool8_t op0, vuint32m1_t op1, vuint32m4_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmaxu_vs_u32m8_u32m1))) +vuint32m1_t vredmaxu(vuint32m1_t op0, vuint32m8_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmaxu_vs_u32m8_u32m1_m))) +vuint32m1_t vredmaxu(vbool4_t op0, vuint32m1_t op1, vuint32m8_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmaxu_vs_u32mf2_u32m1))) +vuint32m1_t vredmaxu(vuint32m1_t op0, vuint32mf2_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmaxu_vs_u32mf2_u32m1_m))) +vuint32m1_t vredmaxu(vbool64_t op0, vuint32m1_t op1, vuint32mf2_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmaxu_vs_u64m1_u64m1))) +vuint64m1_t vredmaxu(vuint64m1_t op0, vuint64m1_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmaxu_vs_u64m1_u64m1_m))) +vuint64m1_t vredmaxu(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmaxu_vs_u64m2_u64m1))) +vuint64m1_t vredmaxu(vuint64m1_t op0, vuint64m2_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmaxu_vs_u64m2_u64m1_m))) +vuint64m1_t vredmaxu(vbool32_t op0, vuint64m1_t op1, vuint64m2_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmaxu_vs_u64m4_u64m1))) +vuint64m1_t vredmaxu(vuint64m1_t op0, vuint64m4_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmaxu_vs_u64m4_u64m1_m))) +vuint64m1_t vredmaxu(vbool16_t op0, vuint64m1_t op1, vuint64m4_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmaxu_vs_u64m8_u64m1))) +vuint64m1_t vredmaxu(vuint64m1_t op0, vuint64m8_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmaxu_vs_u64m8_u64m1_m))) +vuint64m1_t vredmaxu(vbool8_t op0, vuint64m1_t op1, vuint64m8_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmax_vs_i8m1_i8m1))) +vint8m1_t vredmax(vint8m1_t op0, vint8m1_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmax_vs_i8m1_i8m1_m))) +vint8m1_t vredmax(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmax_vs_i8m2_i8m1))) +vint8m1_t vredmax(vint8m1_t op0, vint8m2_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmax_vs_i8m2_i8m1_m))) +vint8m1_t vredmax(vbool4_t op0, vint8m1_t op1, vint8m2_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmax_vs_i8m4_i8m1))) +vint8m1_t vredmax(vint8m1_t op0, vint8m4_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmax_vs_i8m4_i8m1_m))) +vint8m1_t vredmax(vbool2_t op0, vint8m1_t op1, vint8m4_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmax_vs_i8m8_i8m1))) +vint8m1_t vredmax(vint8m1_t op0, vint8m8_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmax_vs_i8m8_i8m1_m))) +vint8m1_t vredmax(vbool1_t op0, vint8m1_t op1, vint8m8_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmax_vs_i8mf2_i8m1))) +vint8m1_t vredmax(vint8m1_t op0, vint8mf2_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmax_vs_i8mf2_i8m1_m))) +vint8m1_t vredmax(vbool16_t op0, vint8m1_t op1, vint8mf2_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmax_vs_i8mf4_i8m1))) +vint8m1_t vredmax(vint8m1_t op0, vint8mf4_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmax_vs_i8mf4_i8m1_m))) +vint8m1_t vredmax(vbool32_t op0, vint8m1_t op1, vint8mf4_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmax_vs_i8mf8_i8m1))) +vint8m1_t vredmax(vint8m1_t op0, vint8mf8_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmax_vs_i8mf8_i8m1_m))) +vint8m1_t vredmax(vbool64_t op0, vint8m1_t op1, vint8mf8_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmax_vs_i16m1_i16m1))) +vint16m1_t vredmax(vint16m1_t op0, vint16m1_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmax_vs_i16m1_i16m1_m))) +vint16m1_t vredmax(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmax_vs_i16m2_i16m1))) +vint16m1_t vredmax(vint16m1_t op0, vint16m2_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmax_vs_i16m2_i16m1_m))) +vint16m1_t vredmax(vbool8_t op0, vint16m1_t op1, vint16m2_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmax_vs_i16m4_i16m1))) +vint16m1_t vredmax(vint16m1_t op0, vint16m4_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmax_vs_i16m4_i16m1_m))) +vint16m1_t vredmax(vbool4_t op0, vint16m1_t op1, vint16m4_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmax_vs_i16m8_i16m1))) +vint16m1_t vredmax(vint16m1_t op0, vint16m8_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmax_vs_i16m8_i16m1_m))) +vint16m1_t vredmax(vbool2_t op0, vint16m1_t op1, vint16m8_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmax_vs_i16mf2_i16m1))) +vint16m1_t vredmax(vint16m1_t op0, vint16mf2_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmax_vs_i16mf2_i16m1_m))) +vint16m1_t vredmax(vbool32_t op0, vint16m1_t op1, vint16mf2_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmax_vs_i16mf4_i16m1))) +vint16m1_t vredmax(vint16m1_t op0, vint16mf4_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmax_vs_i16mf4_i16m1_m))) +vint16m1_t vredmax(vbool64_t op0, vint16m1_t op1, vint16mf4_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmax_vs_i32m1_i32m1))) +vint32m1_t vredmax(vint32m1_t op0, vint32m1_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmax_vs_i32m1_i32m1_m))) +vint32m1_t vredmax(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmax_vs_i32m2_i32m1))) +vint32m1_t vredmax(vint32m1_t op0, vint32m2_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmax_vs_i32m2_i32m1_m))) +vint32m1_t vredmax(vbool16_t op0, vint32m1_t op1, vint32m2_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmax_vs_i32m4_i32m1))) +vint32m1_t vredmax(vint32m1_t op0, vint32m4_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmax_vs_i32m4_i32m1_m))) +vint32m1_t vredmax(vbool8_t op0, vint32m1_t op1, vint32m4_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmax_vs_i32m8_i32m1))) +vint32m1_t vredmax(vint32m1_t op0, vint32m8_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmax_vs_i32m8_i32m1_m))) +vint32m1_t vredmax(vbool4_t op0, vint32m1_t op1, vint32m8_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmax_vs_i32mf2_i32m1))) +vint32m1_t vredmax(vint32m1_t op0, vint32mf2_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmax_vs_i32mf2_i32m1_m))) +vint32m1_t vredmax(vbool64_t op0, vint32m1_t op1, vint32mf2_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmax_vs_i64m1_i64m1))) +vint64m1_t vredmax(vint64m1_t op0, vint64m1_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmax_vs_i64m1_i64m1_m))) +vint64m1_t vredmax(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmax_vs_i64m2_i64m1))) +vint64m1_t vredmax(vint64m1_t op0, vint64m2_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmax_vs_i64m2_i64m1_m))) +vint64m1_t vredmax(vbool32_t op0, vint64m1_t op1, vint64m2_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmax_vs_i64m4_i64m1))) +vint64m1_t vredmax(vint64m1_t op0, vint64m4_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmax_vs_i64m4_i64m1_m))) +vint64m1_t vredmax(vbool16_t op0, vint64m1_t op1, vint64m4_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmax_vs_i64m8_i64m1))) +vint64m1_t vredmax(vint64m1_t op0, vint64m8_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmax_vs_i64m8_i64m1_m))) +vint64m1_t vredmax(vbool8_t op0, vint64m1_t op1, vint64m8_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredminu_vs_u8m1_u8m1))) +vuint8m1_t vredminu(vuint8m1_t op0, vuint8m1_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredminu_vs_u8m1_u8m1_m))) +vuint8m1_t vredminu(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredminu_vs_u8m2_u8m1))) +vuint8m1_t vredminu(vuint8m1_t op0, vuint8m2_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredminu_vs_u8m2_u8m1_m))) +vuint8m1_t vredminu(vbool4_t op0, vuint8m1_t op1, vuint8m2_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredminu_vs_u8m4_u8m1))) +vuint8m1_t vredminu(vuint8m1_t op0, vuint8m4_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredminu_vs_u8m4_u8m1_m))) +vuint8m1_t vredminu(vbool2_t op0, vuint8m1_t op1, vuint8m4_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredminu_vs_u8m8_u8m1))) +vuint8m1_t vredminu(vuint8m1_t op0, vuint8m8_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredminu_vs_u8m8_u8m1_m))) +vuint8m1_t vredminu(vbool1_t op0, vuint8m1_t op1, vuint8m8_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredminu_vs_u8mf2_u8m1))) +vuint8m1_t vredminu(vuint8m1_t op0, vuint8mf2_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredminu_vs_u8mf2_u8m1_m))) +vuint8m1_t vredminu(vbool16_t op0, vuint8m1_t op1, vuint8mf2_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredminu_vs_u8mf4_u8m1))) +vuint8m1_t vredminu(vuint8m1_t op0, vuint8mf4_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredminu_vs_u8mf4_u8m1_m))) +vuint8m1_t vredminu(vbool32_t op0, vuint8m1_t op1, vuint8mf4_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredminu_vs_u8mf8_u8m1))) +vuint8m1_t vredminu(vuint8m1_t op0, vuint8mf8_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredminu_vs_u8mf8_u8m1_m))) +vuint8m1_t vredminu(vbool64_t op0, vuint8m1_t op1, vuint8mf8_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredminu_vs_u16m1_u16m1))) +vuint16m1_t vredminu(vuint16m1_t op0, vuint16m1_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredminu_vs_u16m1_u16m1_m))) +vuint16m1_t vredminu(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredminu_vs_u16m2_u16m1))) +vuint16m1_t vredminu(vuint16m1_t op0, vuint16m2_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredminu_vs_u16m2_u16m1_m))) +vuint16m1_t vredminu(vbool8_t op0, vuint16m1_t op1, vuint16m2_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredminu_vs_u16m4_u16m1))) +vuint16m1_t vredminu(vuint16m1_t op0, vuint16m4_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredminu_vs_u16m4_u16m1_m))) +vuint16m1_t vredminu(vbool4_t op0, vuint16m1_t op1, vuint16m4_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredminu_vs_u16m8_u16m1))) +vuint16m1_t vredminu(vuint16m1_t op0, vuint16m8_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredminu_vs_u16m8_u16m1_m))) +vuint16m1_t vredminu(vbool2_t op0, vuint16m1_t op1, vuint16m8_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredminu_vs_u16mf2_u16m1))) +vuint16m1_t vredminu(vuint16m1_t op0, vuint16mf2_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredminu_vs_u16mf2_u16m1_m))) +vuint16m1_t vredminu(vbool32_t op0, vuint16m1_t op1, vuint16mf2_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredminu_vs_u16mf4_u16m1))) +vuint16m1_t vredminu(vuint16m1_t op0, vuint16mf4_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredminu_vs_u16mf4_u16m1_m))) +vuint16m1_t vredminu(vbool64_t op0, vuint16m1_t op1, vuint16mf4_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredminu_vs_u32m1_u32m1))) +vuint32m1_t vredminu(vuint32m1_t op0, vuint32m1_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredminu_vs_u32m1_u32m1_m))) +vuint32m1_t vredminu(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredminu_vs_u32m2_u32m1))) +vuint32m1_t vredminu(vuint32m1_t op0, vuint32m2_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredminu_vs_u32m2_u32m1_m))) +vuint32m1_t vredminu(vbool16_t op0, vuint32m1_t op1, vuint32m2_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredminu_vs_u32m4_u32m1))) +vuint32m1_t vredminu(vuint32m1_t op0, vuint32m4_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredminu_vs_u32m4_u32m1_m))) +vuint32m1_t vredminu(vbool8_t op0, vuint32m1_t op1, vuint32m4_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredminu_vs_u32m8_u32m1))) +vuint32m1_t vredminu(vuint32m1_t op0, vuint32m8_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredminu_vs_u32m8_u32m1_m))) +vuint32m1_t vredminu(vbool4_t op0, vuint32m1_t op1, vuint32m8_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredminu_vs_u32mf2_u32m1))) +vuint32m1_t vredminu(vuint32m1_t op0, vuint32mf2_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredminu_vs_u32mf2_u32m1_m))) +vuint32m1_t vredminu(vbool64_t op0, vuint32m1_t op1, vuint32mf2_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredminu_vs_u64m1_u64m1))) +vuint64m1_t vredminu(vuint64m1_t op0, vuint64m1_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredminu_vs_u64m1_u64m1_m))) +vuint64m1_t vredminu(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredminu_vs_u64m2_u64m1))) +vuint64m1_t vredminu(vuint64m1_t op0, vuint64m2_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredminu_vs_u64m2_u64m1_m))) +vuint64m1_t vredminu(vbool32_t op0, vuint64m1_t op1, vuint64m2_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredminu_vs_u64m4_u64m1))) +vuint64m1_t vredminu(vuint64m1_t op0, vuint64m4_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredminu_vs_u64m4_u64m1_m))) +vuint64m1_t vredminu(vbool16_t op0, vuint64m1_t op1, vuint64m4_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredminu_vs_u64m8_u64m1))) +vuint64m1_t vredminu(vuint64m1_t op0, vuint64m8_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredminu_vs_u64m8_u64m1_m))) +vuint64m1_t vredminu(vbool8_t op0, vuint64m1_t op1, vuint64m8_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmin_vs_i8m1_i8m1))) +vint8m1_t vredmin(vint8m1_t op0, vint8m1_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmin_vs_i8m1_i8m1_m))) +vint8m1_t vredmin(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmin_vs_i8m2_i8m1))) +vint8m1_t vredmin(vint8m1_t op0, vint8m2_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmin_vs_i8m2_i8m1_m))) +vint8m1_t vredmin(vbool4_t op0, vint8m1_t op1, vint8m2_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmin_vs_i8m4_i8m1))) +vint8m1_t vredmin(vint8m1_t op0, vint8m4_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmin_vs_i8m4_i8m1_m))) +vint8m1_t vredmin(vbool2_t op0, vint8m1_t op1, vint8m4_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmin_vs_i8m8_i8m1))) +vint8m1_t vredmin(vint8m1_t op0, vint8m8_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmin_vs_i8m8_i8m1_m))) +vint8m1_t vredmin(vbool1_t op0, vint8m1_t op1, vint8m8_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmin_vs_i8mf2_i8m1))) +vint8m1_t vredmin(vint8m1_t op0, vint8mf2_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmin_vs_i8mf2_i8m1_m))) +vint8m1_t vredmin(vbool16_t op0, vint8m1_t op1, vint8mf2_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmin_vs_i8mf4_i8m1))) +vint8m1_t vredmin(vint8m1_t op0, vint8mf4_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmin_vs_i8mf4_i8m1_m))) +vint8m1_t vredmin(vbool32_t op0, vint8m1_t op1, vint8mf4_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmin_vs_i8mf8_i8m1))) +vint8m1_t vredmin(vint8m1_t op0, vint8mf8_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmin_vs_i8mf8_i8m1_m))) +vint8m1_t vredmin(vbool64_t op0, vint8m1_t op1, vint8mf8_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmin_vs_i16m1_i16m1))) +vint16m1_t vredmin(vint16m1_t op0, vint16m1_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmin_vs_i16m1_i16m1_m))) +vint16m1_t vredmin(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmin_vs_i16m2_i16m1))) +vint16m1_t vredmin(vint16m1_t op0, vint16m2_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmin_vs_i16m2_i16m1_m))) +vint16m1_t vredmin(vbool8_t op0, vint16m1_t op1, vint16m2_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmin_vs_i16m4_i16m1))) +vint16m1_t vredmin(vint16m1_t op0, vint16m4_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmin_vs_i16m4_i16m1_m))) +vint16m1_t vredmin(vbool4_t op0, vint16m1_t op1, vint16m4_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmin_vs_i16m8_i16m1))) +vint16m1_t vredmin(vint16m1_t op0, vint16m8_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmin_vs_i16m8_i16m1_m))) +vint16m1_t vredmin(vbool2_t op0, vint16m1_t op1, vint16m8_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmin_vs_i16mf2_i16m1))) +vint16m1_t vredmin(vint16m1_t op0, vint16mf2_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmin_vs_i16mf2_i16m1_m))) +vint16m1_t vredmin(vbool32_t op0, vint16m1_t op1, vint16mf2_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmin_vs_i16mf4_i16m1))) +vint16m1_t vredmin(vint16m1_t op0, vint16mf4_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmin_vs_i16mf4_i16m1_m))) +vint16m1_t vredmin(vbool64_t op0, vint16m1_t op1, vint16mf4_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmin_vs_i32m1_i32m1))) +vint32m1_t vredmin(vint32m1_t op0, vint32m1_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmin_vs_i32m1_i32m1_m))) +vint32m1_t vredmin(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmin_vs_i32m2_i32m1))) +vint32m1_t vredmin(vint32m1_t op0, vint32m2_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmin_vs_i32m2_i32m1_m))) +vint32m1_t vredmin(vbool16_t op0, vint32m1_t op1, vint32m2_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmin_vs_i32m4_i32m1))) +vint32m1_t vredmin(vint32m1_t op0, vint32m4_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmin_vs_i32m4_i32m1_m))) +vint32m1_t vredmin(vbool8_t op0, vint32m1_t op1, vint32m4_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmin_vs_i32m8_i32m1))) +vint32m1_t vredmin(vint32m1_t op0, vint32m8_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmin_vs_i32m8_i32m1_m))) +vint32m1_t vredmin(vbool4_t op0, vint32m1_t op1, vint32m8_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmin_vs_i32mf2_i32m1))) +vint32m1_t vredmin(vint32m1_t op0, vint32mf2_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmin_vs_i32mf2_i32m1_m))) +vint32m1_t vredmin(vbool64_t op0, vint32m1_t op1, vint32mf2_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmin_vs_i64m1_i64m1))) +vint64m1_t vredmin(vint64m1_t op0, vint64m1_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmin_vs_i64m1_i64m1_m))) +vint64m1_t vredmin(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmin_vs_i64m2_i64m1))) +vint64m1_t vredmin(vint64m1_t op0, vint64m2_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmin_vs_i64m2_i64m1_m))) +vint64m1_t vredmin(vbool32_t op0, vint64m1_t op1, vint64m2_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmin_vs_i64m4_i64m1))) +vint64m1_t vredmin(vint64m1_t op0, vint64m4_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmin_vs_i64m4_i64m1_m))) +vint64m1_t vredmin(vbool16_t op0, vint64m1_t op1, vint64m4_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmin_vs_i64m8_i64m1))) +vint64m1_t vredmin(vint64m1_t op0, vint64m8_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredmin_vs_i64m8_i64m1_m))) +vint64m1_t vredmin(vbool8_t op0, vint64m1_t op1, vint64m8_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_i8m1_i8m1))) +vint8m1_t vredand(vint8m1_t op0, vint8m1_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_i8m1_i8m1_m))) +vint8m1_t vredand(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_i8m2_i8m1))) +vint8m1_t vredand(vint8m1_t op0, vint8m2_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_i8m2_i8m1_m))) +vint8m1_t vredand(vbool4_t op0, vint8m1_t op1, vint8m2_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_i8m4_i8m1))) +vint8m1_t vredand(vint8m1_t op0, vint8m4_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_i8m4_i8m1_m))) +vint8m1_t vredand(vbool2_t op0, vint8m1_t op1, vint8m4_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_i8m8_i8m1))) +vint8m1_t vredand(vint8m1_t op0, vint8m8_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_i8m8_i8m1_m))) +vint8m1_t vredand(vbool1_t op0, vint8m1_t op1, vint8m8_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_i8mf2_i8m1))) +vint8m1_t vredand(vint8m1_t op0, vint8mf2_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_i8mf2_i8m1_m))) +vint8m1_t vredand(vbool16_t op0, vint8m1_t op1, vint8mf2_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_i8mf4_i8m1))) +vint8m1_t vredand(vint8m1_t op0, vint8mf4_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_i8mf4_i8m1_m))) +vint8m1_t vredand(vbool32_t op0, vint8m1_t op1, vint8mf4_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_i8mf8_i8m1))) +vint8m1_t vredand(vint8m1_t op0, vint8mf8_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_i8mf8_i8m1_m))) +vint8m1_t vredand(vbool64_t op0, vint8m1_t op1, vint8mf8_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_i16m1_i16m1))) +vint16m1_t vredand(vint16m1_t op0, vint16m1_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_i16m1_i16m1_m))) +vint16m1_t vredand(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_i16m2_i16m1))) +vint16m1_t vredand(vint16m1_t op0, vint16m2_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_i16m2_i16m1_m))) +vint16m1_t vredand(vbool8_t op0, vint16m1_t op1, vint16m2_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_i16m4_i16m1))) +vint16m1_t vredand(vint16m1_t op0, vint16m4_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_i16m4_i16m1_m))) +vint16m1_t vredand(vbool4_t op0, vint16m1_t op1, vint16m4_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_i16m8_i16m1))) +vint16m1_t vredand(vint16m1_t op0, vint16m8_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_i16m8_i16m1_m))) +vint16m1_t vredand(vbool2_t op0, vint16m1_t op1, vint16m8_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_i16mf2_i16m1))) +vint16m1_t vredand(vint16m1_t op0, vint16mf2_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_i16mf2_i16m1_m))) +vint16m1_t vredand(vbool32_t op0, vint16m1_t op1, vint16mf2_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_i16mf4_i16m1))) +vint16m1_t vredand(vint16m1_t op0, vint16mf4_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_i16mf4_i16m1_m))) +vint16m1_t vredand(vbool64_t op0, vint16m1_t op1, vint16mf4_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_i32m1_i32m1))) +vint32m1_t vredand(vint32m1_t op0, vint32m1_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_i32m1_i32m1_m))) +vint32m1_t vredand(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_i32m2_i32m1))) +vint32m1_t vredand(vint32m1_t op0, vint32m2_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_i32m2_i32m1_m))) +vint32m1_t vredand(vbool16_t op0, vint32m1_t op1, vint32m2_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_i32m4_i32m1))) +vint32m1_t vredand(vint32m1_t op0, vint32m4_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_i32m4_i32m1_m))) +vint32m1_t vredand(vbool8_t op0, vint32m1_t op1, vint32m4_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_i32m8_i32m1))) +vint32m1_t vredand(vint32m1_t op0, vint32m8_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_i32m8_i32m1_m))) +vint32m1_t vredand(vbool4_t op0, vint32m1_t op1, vint32m8_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_i32mf2_i32m1))) +vint32m1_t vredand(vint32m1_t op0, vint32mf2_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_i32mf2_i32m1_m))) +vint32m1_t vredand(vbool64_t op0, vint32m1_t op1, vint32mf2_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_i64m1_i64m1))) +vint64m1_t vredand(vint64m1_t op0, vint64m1_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_i64m1_i64m1_m))) +vint64m1_t vredand(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_i64m2_i64m1))) +vint64m1_t vredand(vint64m1_t op0, vint64m2_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_i64m2_i64m1_m))) +vint64m1_t vredand(vbool32_t op0, vint64m1_t op1, vint64m2_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_i64m4_i64m1))) +vint64m1_t vredand(vint64m1_t op0, vint64m4_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_i64m4_i64m1_m))) +vint64m1_t vredand(vbool16_t op0, vint64m1_t op1, vint64m4_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_i64m8_i64m1))) +vint64m1_t vredand(vint64m1_t op0, vint64m8_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_i64m8_i64m1_m))) +vint64m1_t vredand(vbool8_t op0, vint64m1_t op1, vint64m8_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_u8m1_u8m1))) +vuint8m1_t vredand(vuint8m1_t op0, vuint8m1_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_u8m1_u8m1_m))) +vuint8m1_t vredand(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_u8m2_u8m1))) +vuint8m1_t vredand(vuint8m1_t op0, vuint8m2_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_u8m2_u8m1_m))) +vuint8m1_t vredand(vbool4_t op0, vuint8m1_t op1, vuint8m2_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_u8m4_u8m1))) +vuint8m1_t vredand(vuint8m1_t op0, vuint8m4_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_u8m4_u8m1_m))) +vuint8m1_t vredand(vbool2_t op0, vuint8m1_t op1, vuint8m4_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_u8m8_u8m1))) +vuint8m1_t vredand(vuint8m1_t op0, vuint8m8_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_u8m8_u8m1_m))) +vuint8m1_t vredand(vbool1_t op0, vuint8m1_t op1, vuint8m8_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_u8mf2_u8m1))) +vuint8m1_t vredand(vuint8m1_t op0, vuint8mf2_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_u8mf2_u8m1_m))) +vuint8m1_t vredand(vbool16_t op0, vuint8m1_t op1, vuint8mf2_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_u8mf4_u8m1))) +vuint8m1_t vredand(vuint8m1_t op0, vuint8mf4_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_u8mf4_u8m1_m))) +vuint8m1_t vredand(vbool32_t op0, vuint8m1_t op1, vuint8mf4_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_u8mf8_u8m1))) +vuint8m1_t vredand(vuint8m1_t op0, vuint8mf8_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_u8mf8_u8m1_m))) +vuint8m1_t vredand(vbool64_t op0, vuint8m1_t op1, vuint8mf8_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_u16m1_u16m1))) +vuint16m1_t vredand(vuint16m1_t op0, vuint16m1_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_u16m1_u16m1_m))) +vuint16m1_t vredand(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_u16m2_u16m1))) +vuint16m1_t vredand(vuint16m1_t op0, vuint16m2_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_u16m2_u16m1_m))) +vuint16m1_t vredand(vbool8_t op0, vuint16m1_t op1, vuint16m2_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_u16m4_u16m1))) +vuint16m1_t vredand(vuint16m1_t op0, vuint16m4_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_u16m4_u16m1_m))) +vuint16m1_t vredand(vbool4_t op0, vuint16m1_t op1, vuint16m4_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_u16m8_u16m1))) +vuint16m1_t vredand(vuint16m1_t op0, vuint16m8_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_u16m8_u16m1_m))) +vuint16m1_t vredand(vbool2_t op0, vuint16m1_t op1, vuint16m8_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_u16mf2_u16m1))) +vuint16m1_t vredand(vuint16m1_t op0, vuint16mf2_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_u16mf2_u16m1_m))) +vuint16m1_t vredand(vbool32_t op0, vuint16m1_t op1, vuint16mf2_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_u16mf4_u16m1))) +vuint16m1_t vredand(vuint16m1_t op0, vuint16mf4_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_u16mf4_u16m1_m))) +vuint16m1_t vredand(vbool64_t op0, vuint16m1_t op1, vuint16mf4_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_u32m1_u32m1))) +vuint32m1_t vredand(vuint32m1_t op0, vuint32m1_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_u32m1_u32m1_m))) +vuint32m1_t vredand(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_u32m2_u32m1))) +vuint32m1_t vredand(vuint32m1_t op0, vuint32m2_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_u32m2_u32m1_m))) +vuint32m1_t vredand(vbool16_t op0, vuint32m1_t op1, vuint32m2_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_u32m4_u32m1))) +vuint32m1_t vredand(vuint32m1_t op0, vuint32m4_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_u32m4_u32m1_m))) +vuint32m1_t vredand(vbool8_t op0, vuint32m1_t op1, vuint32m4_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_u32m8_u32m1))) +vuint32m1_t vredand(vuint32m1_t op0, vuint32m8_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_u32m8_u32m1_m))) +vuint32m1_t vredand(vbool4_t op0, vuint32m1_t op1, vuint32m8_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_u32mf2_u32m1))) +vuint32m1_t vredand(vuint32m1_t op0, vuint32mf2_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_u32mf2_u32m1_m))) +vuint32m1_t vredand(vbool64_t op0, vuint32m1_t op1, vuint32mf2_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_u64m1_u64m1))) +vuint64m1_t vredand(vuint64m1_t op0, vuint64m1_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_u64m1_u64m1_m))) +vuint64m1_t vredand(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_u64m2_u64m1))) +vuint64m1_t vredand(vuint64m1_t op0, vuint64m2_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_u64m2_u64m1_m))) +vuint64m1_t vredand(vbool32_t op0, vuint64m1_t op1, vuint64m2_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_u64m4_u64m1))) +vuint64m1_t vredand(vuint64m1_t op0, vuint64m4_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_u64m4_u64m1_m))) +vuint64m1_t vredand(vbool16_t op0, vuint64m1_t op1, vuint64m4_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_u64m8_u64m1))) +vuint64m1_t vredand(vuint64m1_t op0, vuint64m8_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredand_vs_u64m8_u64m1_m))) +vuint64m1_t vredand(vbool8_t op0, vuint64m1_t op1, vuint64m8_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_u64m1))) +void vsuxei64(uint64_t * op0, vuint64m1_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_u64m1_m))) +void vsuxei64(vbool64_t op0, uint64_t * op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_u64m2))) +void vsuxei64(uint64_t * op0, vuint64m2_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_u64m2_m))) +void vsuxei64(vbool32_t op0, uint64_t * op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_u64m4))) +void vsuxei64(uint64_t * op0, vuint64m4_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_u64m4_m))) +void vsuxei64(vbool16_t op0, uint64_t * op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_u64m8))) +void vsuxei64(uint64_t * op0, vuint64m8_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_u64m8_m))) +void vsuxei64(vbool8_t op0, uint64_t * op1, vuint64m8_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_i8m1_i8m1))) +vint8m1_t vredor(vint8m1_t op0, vint8m1_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_i8m1_i8m1_m))) +vint8m1_t vredor(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_i8m2_i8m1))) +vint8m1_t vredor(vint8m1_t op0, vint8m2_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_i8m2_i8m1_m))) +vint8m1_t vredor(vbool4_t op0, vint8m1_t op1, vint8m2_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_i8m4_i8m1))) +vint8m1_t vredor(vint8m1_t op0, vint8m4_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_i8m4_i8m1_m))) +vint8m1_t vredor(vbool2_t op0, vint8m1_t op1, vint8m4_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_i8m8_i8m1))) +vint8m1_t vredor(vint8m1_t op0, vint8m8_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_i8m8_i8m1_m))) +vint8m1_t vredor(vbool1_t op0, vint8m1_t op1, vint8m8_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_i8mf2_i8m1))) +vint8m1_t vredor(vint8m1_t op0, vint8mf2_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_i8mf2_i8m1_m))) +vint8m1_t vredor(vbool16_t op0, vint8m1_t op1, vint8mf2_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_i8mf4_i8m1))) +vint8m1_t vredor(vint8m1_t op0, vint8mf4_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_i8mf4_i8m1_m))) +vint8m1_t vredor(vbool32_t op0, vint8m1_t op1, vint8mf4_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_i8mf8_i8m1))) +vint8m1_t vredor(vint8m1_t op0, vint8mf8_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_i8mf8_i8m1_m))) +vint8m1_t vredor(vbool64_t op0, vint8m1_t op1, vint8mf8_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_i16m1_i16m1))) +vint16m1_t vredor(vint16m1_t op0, vint16m1_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_i16m1_i16m1_m))) +vint16m1_t vredor(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_i16m2_i16m1))) +vint16m1_t vredor(vint16m1_t op0, vint16m2_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_i16m2_i16m1_m))) +vint16m1_t vredor(vbool8_t op0, vint16m1_t op1, vint16m2_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_i16m4_i16m1))) +vint16m1_t vredor(vint16m1_t op0, vint16m4_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_i16m4_i16m1_m))) +vint16m1_t vredor(vbool4_t op0, vint16m1_t op1, vint16m4_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_i16m8_i16m1))) +vint16m1_t vredor(vint16m1_t op0, vint16m8_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_i16m8_i16m1_m))) +vint16m1_t vredor(vbool2_t op0, vint16m1_t op1, vint16m8_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_i16mf2_i16m1))) +vint16m1_t vredor(vint16m1_t op0, vint16mf2_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_i16mf2_i16m1_m))) +vint16m1_t vredor(vbool32_t op0, vint16m1_t op1, vint16mf2_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_i16mf4_i16m1))) +vint16m1_t vredor(vint16m1_t op0, vint16mf4_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_i16mf4_i16m1_m))) +vint16m1_t vredor(vbool64_t op0, vint16m1_t op1, vint16mf4_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_i32m1_i32m1))) +vint32m1_t vredor(vint32m1_t op0, vint32m1_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_i32m1_i32m1_m))) +vint32m1_t vredor(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_i32m2_i32m1))) +vint32m1_t vredor(vint32m1_t op0, vint32m2_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_i32m2_i32m1_m))) +vint32m1_t vredor(vbool16_t op0, vint32m1_t op1, vint32m2_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_i32m4_i32m1))) +vint32m1_t vredor(vint32m1_t op0, vint32m4_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_i32m4_i32m1_m))) +vint32m1_t vredor(vbool8_t op0, vint32m1_t op1, vint32m4_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_i32m8_i32m1))) +vint32m1_t vredor(vint32m1_t op0, vint32m8_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_i32m8_i32m1_m))) +vint32m1_t vredor(vbool4_t op0, vint32m1_t op1, vint32m8_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_i32mf2_i32m1))) +vint32m1_t vredor(vint32m1_t op0, vint32mf2_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_i32mf2_i32m1_m))) +vint32m1_t vredor(vbool64_t op0, vint32m1_t op1, vint32mf2_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_i64m1_i64m1))) +vint64m1_t vredor(vint64m1_t op0, vint64m1_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_i64m1_i64m1_m))) +vint64m1_t vredor(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_i64m2_i64m1))) +vint64m1_t vredor(vint64m1_t op0, vint64m2_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_i64m2_i64m1_m))) +vint64m1_t vredor(vbool32_t op0, vint64m1_t op1, vint64m2_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_i64m4_i64m1))) +vint64m1_t vredor(vint64m1_t op0, vint64m4_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_i64m4_i64m1_m))) +vint64m1_t vredor(vbool16_t op0, vint64m1_t op1, vint64m4_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_i64m8_i64m1))) +vint64m1_t vredor(vint64m1_t op0, vint64m8_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_i64m8_i64m1_m))) +vint64m1_t vredor(vbool8_t op0, vint64m1_t op1, vint64m8_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_u8m1_u8m1))) +vuint8m1_t vredor(vuint8m1_t op0, vuint8m1_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_u8m1_u8m1_m))) +vuint8m1_t vredor(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_u8m2_u8m1))) +vuint8m1_t vredor(vuint8m1_t op0, vuint8m2_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_u8m2_u8m1_m))) +vuint8m1_t vredor(vbool4_t op0, vuint8m1_t op1, vuint8m2_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_u8m4_u8m1))) +vuint8m1_t vredor(vuint8m1_t op0, vuint8m4_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_u8m4_u8m1_m))) +vuint8m1_t vredor(vbool2_t op0, vuint8m1_t op1, vuint8m4_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_u8m8_u8m1))) +vuint8m1_t vredor(vuint8m1_t op0, vuint8m8_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_u8m8_u8m1_m))) +vuint8m1_t vredor(vbool1_t op0, vuint8m1_t op1, vuint8m8_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_u8mf2_u8m1))) +vuint8m1_t vredor(vuint8m1_t op0, vuint8mf2_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_u8mf2_u8m1_m))) +vuint8m1_t vredor(vbool16_t op0, vuint8m1_t op1, vuint8mf2_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_u8mf4_u8m1))) +vuint8m1_t vredor(vuint8m1_t op0, vuint8mf4_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_u8mf4_u8m1_m))) +vuint8m1_t vredor(vbool32_t op0, vuint8m1_t op1, vuint8mf4_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_u8mf8_u8m1))) +vuint8m1_t vredor(vuint8m1_t op0, vuint8mf8_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_u8mf8_u8m1_m))) +vuint8m1_t vredor(vbool64_t op0, vuint8m1_t op1, vuint8mf8_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_u16m1_u16m1))) +vuint16m1_t vredor(vuint16m1_t op0, vuint16m1_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_u16m1_u16m1_m))) +vuint16m1_t vredor(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_u16m2_u16m1))) +vuint16m1_t vredor(vuint16m1_t op0, vuint16m2_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_u16m2_u16m1_m))) +vuint16m1_t vredor(vbool8_t op0, vuint16m1_t op1, vuint16m2_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_u16m4_u16m1))) +vuint16m1_t vredor(vuint16m1_t op0, vuint16m4_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_u16m4_u16m1_m))) +vuint16m1_t vredor(vbool4_t op0, vuint16m1_t op1, vuint16m4_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_u16m8_u16m1))) +vuint16m1_t vredor(vuint16m1_t op0, vuint16m8_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_u16m8_u16m1_m))) +vuint16m1_t vredor(vbool2_t op0, vuint16m1_t op1, vuint16m8_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_u16mf2_u16m1))) +vuint16m1_t vredor(vuint16m1_t op0, vuint16mf2_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_u16mf2_u16m1_m))) +vuint16m1_t vredor(vbool32_t op0, vuint16m1_t op1, vuint16mf2_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_u16mf4_u16m1))) +vuint16m1_t vredor(vuint16m1_t op0, vuint16mf4_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_u16mf4_u16m1_m))) +vuint16m1_t vredor(vbool64_t op0, vuint16m1_t op1, vuint16mf4_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_u32m1_u32m1))) +vuint32m1_t vredor(vuint32m1_t op0, vuint32m1_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_u32m1_u32m1_m))) +vuint32m1_t vredor(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_u32m2_u32m1))) +vuint32m1_t vredor(vuint32m1_t op0, vuint32m2_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_u32m2_u32m1_m))) +vuint32m1_t vredor(vbool16_t op0, vuint32m1_t op1, vuint32m2_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_u32m4_u32m1))) +vuint32m1_t vredor(vuint32m1_t op0, vuint32m4_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_u32m4_u32m1_m))) +vuint32m1_t vredor(vbool8_t op0, vuint32m1_t op1, vuint32m4_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_u32m8_u32m1))) +vuint32m1_t vredor(vuint32m1_t op0, vuint32m8_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_u32m8_u32m1_m))) +vuint32m1_t vredor(vbool4_t op0, vuint32m1_t op1, vuint32m8_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_u32mf2_u32m1))) +vuint32m1_t vredor(vuint32m1_t op0, vuint32mf2_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_u32mf2_u32m1_m))) +vuint32m1_t vredor(vbool64_t op0, vuint32m1_t op1, vuint32mf2_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_u64m1_u64m1))) +vuint64m1_t vredor(vuint64m1_t op0, vuint64m1_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_u64m1_u64m1_m))) +vuint64m1_t vredor(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_u64m2_u64m1))) +vuint64m1_t vredor(vuint64m1_t op0, vuint64m2_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_u64m2_u64m1_m))) +vuint64m1_t vredor(vbool32_t op0, vuint64m1_t op1, vuint64m2_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_u64m4_u64m1))) +vuint64m1_t vredor(vuint64m1_t op0, vuint64m4_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_u64m4_u64m1_m))) +vuint64m1_t vredor(vbool16_t op0, vuint64m1_t op1, vuint64m4_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_u64m8_u64m1))) +vuint64m1_t vredor(vuint64m1_t op0, vuint64m8_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredor_vs_u64m8_u64m1_m))) +vuint64m1_t vredor(vbool8_t op0, vuint64m1_t op1, vuint64m8_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_i8m1_i8m1))) +vint8m1_t vredxor(vint8m1_t op0, vint8m1_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_i8m1_i8m1_m))) +vint8m1_t vredxor(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_i8m2_i8m1))) +vint8m1_t vredxor(vint8m1_t op0, vint8m2_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_i8m2_i8m1_m))) +vint8m1_t vredxor(vbool4_t op0, vint8m1_t op1, vint8m2_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_i8m4_i8m1))) +vint8m1_t vredxor(vint8m1_t op0, vint8m4_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_i8m4_i8m1_m))) +vint8m1_t vredxor(vbool2_t op0, vint8m1_t op1, vint8m4_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_i8m8_i8m1))) +vint8m1_t vredxor(vint8m1_t op0, vint8m8_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_i8m8_i8m1_m))) +vint8m1_t vredxor(vbool1_t op0, vint8m1_t op1, vint8m8_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_i8mf2_i8m1))) +vint8m1_t vredxor(vint8m1_t op0, vint8mf2_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_i8mf2_i8m1_m))) +vint8m1_t vredxor(vbool16_t op0, vint8m1_t op1, vint8mf2_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_i8mf4_i8m1))) +vint8m1_t vredxor(vint8m1_t op0, vint8mf4_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_i8mf4_i8m1_m))) +vint8m1_t vredxor(vbool32_t op0, vint8m1_t op1, vint8mf4_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_i8mf8_i8m1))) +vint8m1_t vredxor(vint8m1_t op0, vint8mf8_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_i8mf8_i8m1_m))) +vint8m1_t vredxor(vbool64_t op0, vint8m1_t op1, vint8mf8_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_i16m1_i16m1))) +vint16m1_t vredxor(vint16m1_t op0, vint16m1_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_i16m1_i16m1_m))) +vint16m1_t vredxor(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_i16m2_i16m1))) +vint16m1_t vredxor(vint16m1_t op0, vint16m2_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_i16m2_i16m1_m))) +vint16m1_t vredxor(vbool8_t op0, vint16m1_t op1, vint16m2_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_i16m4_i16m1))) +vint16m1_t vredxor(vint16m1_t op0, vint16m4_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_i16m4_i16m1_m))) +vint16m1_t vredxor(vbool4_t op0, vint16m1_t op1, vint16m4_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_i16m8_i16m1))) +vint16m1_t vredxor(vint16m1_t op0, vint16m8_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_i16m8_i16m1_m))) +vint16m1_t vredxor(vbool2_t op0, vint16m1_t op1, vint16m8_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_i16mf2_i16m1))) +vint16m1_t vredxor(vint16m1_t op0, vint16mf2_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_i16mf2_i16m1_m))) +vint16m1_t vredxor(vbool32_t op0, vint16m1_t op1, vint16mf2_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_i16mf4_i16m1))) +vint16m1_t vredxor(vint16m1_t op0, vint16mf4_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_i16mf4_i16m1_m))) +vint16m1_t vredxor(vbool64_t op0, vint16m1_t op1, vint16mf4_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_i32m1_i32m1))) +vint32m1_t vredxor(vint32m1_t op0, vint32m1_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_i32m1_i32m1_m))) +vint32m1_t vredxor(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_i32m2_i32m1))) +vint32m1_t vredxor(vint32m1_t op0, vint32m2_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_i32m2_i32m1_m))) +vint32m1_t vredxor(vbool16_t op0, vint32m1_t op1, vint32m2_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_i32m4_i32m1))) +vint32m1_t vredxor(vint32m1_t op0, vint32m4_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_i32m4_i32m1_m))) +vint32m1_t vredxor(vbool8_t op0, vint32m1_t op1, vint32m4_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_i32m8_i32m1))) +vint32m1_t vredxor(vint32m1_t op0, vint32m8_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_i32m8_i32m1_m))) +vint32m1_t vredxor(vbool4_t op0, vint32m1_t op1, vint32m8_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_i32mf2_i32m1))) +vint32m1_t vredxor(vint32m1_t op0, vint32mf2_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_i32mf2_i32m1_m))) +vint32m1_t vredxor(vbool64_t op0, vint32m1_t op1, vint32mf2_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_i64m1_i64m1))) +vint64m1_t vredxor(vint64m1_t op0, vint64m1_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_i64m1_i64m1_m))) +vint64m1_t vredxor(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_i64m2_i64m1))) +vint64m1_t vredxor(vint64m1_t op0, vint64m2_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_i64m2_i64m1_m))) +vint64m1_t vredxor(vbool32_t op0, vint64m1_t op1, vint64m2_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_i64m4_i64m1))) +vint64m1_t vredxor(vint64m1_t op0, vint64m4_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_i64m4_i64m1_m))) +vint64m1_t vredxor(vbool16_t op0, vint64m1_t op1, vint64m4_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_i64m8_i64m1))) +vint64m1_t vredxor(vint64m1_t op0, vint64m8_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_i64m8_i64m1_m))) +vint64m1_t vredxor(vbool8_t op0, vint64m1_t op1, vint64m8_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_u8m1_u8m1))) +vuint8m1_t vredxor(vuint8m1_t op0, vuint8m1_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_u8m1_u8m1_m))) +vuint8m1_t vredxor(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_u8m2_u8m1))) +vuint8m1_t vredxor(vuint8m1_t op0, vuint8m2_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_u8m2_u8m1_m))) +vuint8m1_t vredxor(vbool4_t op0, vuint8m1_t op1, vuint8m2_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_u8m4_u8m1))) +vuint8m1_t vredxor(vuint8m1_t op0, vuint8m4_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_u8m4_u8m1_m))) +vuint8m1_t vredxor(vbool2_t op0, vuint8m1_t op1, vuint8m4_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_u8m8_u8m1))) +vuint8m1_t vredxor(vuint8m1_t op0, vuint8m8_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_u8m8_u8m1_m))) +vuint8m1_t vredxor(vbool1_t op0, vuint8m1_t op1, vuint8m8_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_u8mf2_u8m1))) +vuint8m1_t vredxor(vuint8m1_t op0, vuint8mf2_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_u8mf2_u8m1_m))) +vuint8m1_t vredxor(vbool16_t op0, vuint8m1_t op1, vuint8mf2_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_u8mf4_u8m1))) +vuint8m1_t vredxor(vuint8m1_t op0, vuint8mf4_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_u8mf4_u8m1_m))) +vuint8m1_t vredxor(vbool32_t op0, vuint8m1_t op1, vuint8mf4_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_u8mf8_u8m1))) +vuint8m1_t vredxor(vuint8m1_t op0, vuint8mf8_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_u8mf8_u8m1_m))) +vuint8m1_t vredxor(vbool64_t op0, vuint8m1_t op1, vuint8mf8_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_u16m1_u16m1))) +vuint16m1_t vredxor(vuint16m1_t op0, vuint16m1_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_u16m1_u16m1_m))) +vuint16m1_t vredxor(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_u16m2_u16m1))) +vuint16m1_t vredxor(vuint16m1_t op0, vuint16m2_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_u16m2_u16m1_m))) +vuint16m1_t vredxor(vbool8_t op0, vuint16m1_t op1, vuint16m2_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_u16m4_u16m1))) +vuint16m1_t vredxor(vuint16m1_t op0, vuint16m4_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_u16m4_u16m1_m))) +vuint16m1_t vredxor(vbool4_t op0, vuint16m1_t op1, vuint16m4_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_u16m8_u16m1))) +vuint16m1_t vredxor(vuint16m1_t op0, vuint16m8_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_u16m8_u16m1_m))) +vuint16m1_t vredxor(vbool2_t op0, vuint16m1_t op1, vuint16m8_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_u16mf2_u16m1))) +vuint16m1_t vredxor(vuint16m1_t op0, vuint16mf2_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_u16mf2_u16m1_m))) +vuint16m1_t vredxor(vbool32_t op0, vuint16m1_t op1, vuint16mf2_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_u16mf4_u16m1))) +vuint16m1_t vredxor(vuint16m1_t op0, vuint16mf4_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_u16mf4_u16m1_m))) +vuint16m1_t vredxor(vbool64_t op0, vuint16m1_t op1, vuint16mf4_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_u32m1_u32m1))) +vuint32m1_t vredxor(vuint32m1_t op0, vuint32m1_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_u32m1_u32m1_m))) +vuint32m1_t vredxor(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_u32m2_u32m1))) +vuint32m1_t vredxor(vuint32m1_t op0, vuint32m2_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_u32m2_u32m1_m))) +vuint32m1_t vredxor(vbool16_t op0, vuint32m1_t op1, vuint32m2_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_u32m4_u32m1))) +vuint32m1_t vredxor(vuint32m1_t op0, vuint32m4_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_u32m4_u32m1_m))) +vuint32m1_t vredxor(vbool8_t op0, vuint32m1_t op1, vuint32m4_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_u32m8_u32m1))) +vuint32m1_t vredxor(vuint32m1_t op0, vuint32m8_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_u32m8_u32m1_m))) +vuint32m1_t vredxor(vbool4_t op0, vuint32m1_t op1, vuint32m8_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_u32mf2_u32m1))) +vuint32m1_t vredxor(vuint32m1_t op0, vuint32mf2_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_u32mf2_u32m1_m))) +vuint32m1_t vredxor(vbool64_t op0, vuint32m1_t op1, vuint32mf2_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_u64m1_u64m1))) +vuint64m1_t vredxor(vuint64m1_t op0, vuint64m1_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_u64m1_u64m1_m))) +vuint64m1_t vredxor(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_u64m2_u64m1))) +vuint64m1_t vredxor(vuint64m1_t op0, vuint64m2_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_u64m2_u64m1_m))) +vuint64m1_t vredxor(vbool32_t op0, vuint64m1_t op1, vuint64m2_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_u64m4_u64m1))) +vuint64m1_t vredxor(vuint64m1_t op0, vuint64m4_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_u64m4_u64m1_m))) +vuint64m1_t vredxor(vbool16_t op0, vuint64m1_t op1, vuint64m4_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_u64m8_u64m1))) +vuint64m1_t vredxor(vuint64m1_t op0, vuint64m8_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vredxor_vs_u64m8_u64m1_m))) +vuint64m1_t vredxor(vbool8_t op0, vuint64m1_t op1, vuint64m8_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsum_vs_i8m1_i16m1))) +vint16m1_t vwredsum(vint16m1_t op0, vint8m1_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsum_vs_i8m1_i16m1_m))) +vint16m1_t vwredsum(vbool8_t op0, vint16m1_t op1, vint8m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsum_vs_i8m2_i16m1))) +vint16m1_t vwredsum(vint16m1_t op0, vint8m2_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsum_vs_i8m2_i16m1_m))) +vint16m1_t vwredsum(vbool4_t op0, vint16m1_t op1, vint8m2_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsum_vs_i8m4_i16m1))) +vint16m1_t vwredsum(vint16m1_t op0, vint8m4_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsum_vs_i8m4_i16m1_m))) +vint16m1_t vwredsum(vbool2_t op0, vint16m1_t op1, vint8m4_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsum_vs_i8m8_i16m1))) +vint16m1_t vwredsum(vint16m1_t op0, vint8m8_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsum_vs_i8m8_i16m1_m))) +vint16m1_t vwredsum(vbool1_t op0, vint16m1_t op1, vint8m8_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsum_vs_i8mf2_i16m1))) +vint16m1_t vwredsum(vint16m1_t op0, vint8mf2_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsum_vs_i8mf2_i16m1_m))) +vint16m1_t vwredsum(vbool16_t op0, vint16m1_t op1, vint8mf2_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsum_vs_i8mf4_i16m1))) +vint16m1_t vwredsum(vint16m1_t op0, vint8mf4_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsum_vs_i8mf4_i16m1_m))) +vint16m1_t vwredsum(vbool32_t op0, vint16m1_t op1, vint8mf4_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsum_vs_i8mf8_i16m1))) +vint16m1_t vwredsum(vint16m1_t op0, vint8mf8_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsum_vs_i8mf8_i16m1_m))) +vint16m1_t vwredsum(vbool64_t op0, vint16m1_t op1, vint8mf8_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsum_vs_i16m1_i32m1))) +vint32m1_t vwredsum(vint32m1_t op0, vint16m1_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsum_vs_i16m1_i32m1_m))) +vint32m1_t vwredsum(vbool16_t op0, vint32m1_t op1, vint16m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsum_vs_i16m2_i32m1))) +vint32m1_t vwredsum(vint32m1_t op0, vint16m2_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsum_vs_i16m2_i32m1_m))) +vint32m1_t vwredsum(vbool8_t op0, vint32m1_t op1, vint16m2_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsum_vs_i16m4_i32m1))) +vint32m1_t vwredsum(vint32m1_t op0, vint16m4_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsum_vs_i16m4_i32m1_m))) +vint32m1_t vwredsum(vbool4_t op0, vint32m1_t op1, vint16m4_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsum_vs_i16m8_i32m1))) +vint32m1_t vwredsum(vint32m1_t op0, vint16m8_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsum_vs_i16m8_i32m1_m))) +vint32m1_t vwredsum(vbool2_t op0, vint32m1_t op1, vint16m8_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsum_vs_i16mf2_i32m1))) +vint32m1_t vwredsum(vint32m1_t op0, vint16mf2_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsum_vs_i16mf2_i32m1_m))) +vint32m1_t vwredsum(vbool32_t op0, vint32m1_t op1, vint16mf2_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsum_vs_i16mf4_i32m1))) +vint32m1_t vwredsum(vint32m1_t op0, vint16mf4_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsum_vs_i16mf4_i32m1_m))) +vint32m1_t vwredsum(vbool64_t op0, vint32m1_t op1, vint16mf4_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsum_vs_i32m1_i64m1))) +vint64m1_t vwredsum(vint64m1_t op0, vint32m1_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsum_vs_i32m1_i64m1_m))) +vint64m1_t vwredsum(vbool32_t op0, vint64m1_t op1, vint32m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsum_vs_i32m2_i64m1))) +vint64m1_t vwredsum(vint64m1_t op0, vint32m2_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsum_vs_i32m2_i64m1_m))) +vint64m1_t vwredsum(vbool16_t op0, vint64m1_t op1, vint32m2_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsum_vs_i32m4_i64m1))) +vint64m1_t vwredsum(vint64m1_t op0, vint32m4_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsum_vs_i32m4_i64m1_m))) +vint64m1_t vwredsum(vbool8_t op0, vint64m1_t op1, vint32m4_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsum_vs_i32m8_i64m1))) +vint64m1_t vwredsum(vint64m1_t op0, vint32m8_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsum_vs_i32m8_i64m1_m))) +vint64m1_t vwredsum(vbool4_t op0, vint64m1_t op1, vint32m8_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsum_vs_i32mf2_i64m1))) +vint64m1_t vwredsum(vint64m1_t op0, vint32mf2_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsum_vs_i32mf2_i64m1_m))) +vint64m1_t vwredsum(vbool64_t op0, vint64m1_t op1, vint32mf2_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsumu_vs_u8m1_u16m1))) +vuint16m1_t vwredsumu(vuint16m1_t op0, vuint8m1_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsumu_vs_u8m1_u16m1_m))) +vuint16m1_t vwredsumu(vbool8_t op0, vuint16m1_t op1, vuint8m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsumu_vs_u8m2_u16m1))) +vuint16m1_t vwredsumu(vuint16m1_t op0, vuint8m2_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsumu_vs_u8m2_u16m1_m))) +vuint16m1_t vwredsumu(vbool4_t op0, vuint16m1_t op1, vuint8m2_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsumu_vs_u8m4_u16m1))) +vuint16m1_t vwredsumu(vuint16m1_t op0, vuint8m4_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsumu_vs_u8m4_u16m1_m))) +vuint16m1_t vwredsumu(vbool2_t op0, vuint16m1_t op1, vuint8m4_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsumu_vs_u8m8_u16m1))) +vuint16m1_t vwredsumu(vuint16m1_t op0, vuint8m8_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsumu_vs_u8m8_u16m1_m))) +vuint16m1_t vwredsumu(vbool1_t op0, vuint16m1_t op1, vuint8m8_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsumu_vs_u8mf2_u16m1))) +vuint16m1_t vwredsumu(vuint16m1_t op0, vuint8mf2_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsumu_vs_u8mf2_u16m1_m))) +vuint16m1_t vwredsumu(vbool16_t op0, vuint16m1_t op1, vuint8mf2_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsumu_vs_u8mf4_u16m1))) +vuint16m1_t vwredsumu(vuint16m1_t op0, vuint8mf4_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsumu_vs_u8mf4_u16m1_m))) +vuint16m1_t vwredsumu(vbool32_t op0, vuint16m1_t op1, vuint8mf4_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsumu_vs_u8mf8_u16m1))) +vuint16m1_t vwredsumu(vuint16m1_t op0, vuint8mf8_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsumu_vs_u8mf8_u16m1_m))) +vuint16m1_t vwredsumu(vbool64_t op0, vuint16m1_t op1, vuint8mf8_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsumu_vs_u16m1_u32m1))) +vuint32m1_t vwredsumu(vuint32m1_t op0, vuint16m1_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsumu_vs_u16m1_u32m1_m))) +vuint32m1_t vwredsumu(vbool16_t op0, vuint32m1_t op1, vuint16m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsumu_vs_u16m2_u32m1))) +vuint32m1_t vwredsumu(vuint32m1_t op0, vuint16m2_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsumu_vs_u16m2_u32m1_m))) +vuint32m1_t vwredsumu(vbool8_t op0, vuint32m1_t op1, vuint16m2_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsumu_vs_u16m4_u32m1))) +vuint32m1_t vwredsumu(vuint32m1_t op0, vuint16m4_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsumu_vs_u16m4_u32m1_m))) +vuint32m1_t vwredsumu(vbool4_t op0, vuint32m1_t op1, vuint16m4_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsumu_vs_u16m8_u32m1))) +vuint32m1_t vwredsumu(vuint32m1_t op0, vuint16m8_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsumu_vs_u16m8_u32m1_m))) +vuint32m1_t vwredsumu(vbool2_t op0, vuint32m1_t op1, vuint16m8_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsumu_vs_u16mf2_u32m1))) +vuint32m1_t vwredsumu(vuint32m1_t op0, vuint16mf2_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsumu_vs_u16mf2_u32m1_m))) +vuint32m1_t vwredsumu(vbool32_t op0, vuint32m1_t op1, vuint16mf2_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsumu_vs_u16mf4_u32m1))) +vuint32m1_t vwredsumu(vuint32m1_t op0, vuint16mf4_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsumu_vs_u16mf4_u32m1_m))) +vuint32m1_t vwredsumu(vbool64_t op0, vuint32m1_t op1, vuint16mf4_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsumu_vs_u32m1_u64m1))) +vuint64m1_t vwredsumu(vuint64m1_t op0, vuint32m1_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsumu_vs_u32m1_u64m1_m))) +vuint64m1_t vwredsumu(vbool32_t op0, vuint64m1_t op1, vuint32m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsumu_vs_u32m2_u64m1))) +vuint64m1_t vwredsumu(vuint64m1_t op0, vuint32m2_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsumu_vs_u32m2_u64m1_m))) +vuint64m1_t vwredsumu(vbool16_t op0, vuint64m1_t op1, vuint32m2_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsumu_vs_u32m4_u64m1))) +vuint64m1_t vwredsumu(vuint64m1_t op0, vuint32m4_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsumu_vs_u32m4_u64m1_m))) +vuint64m1_t vwredsumu(vbool8_t op0, vuint64m1_t op1, vuint32m4_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsumu_vs_u32m8_u64m1))) +vuint64m1_t vwredsumu(vuint64m1_t op0, vuint32m8_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsumu_vs_u32m8_u64m1_m))) +vuint64m1_t vwredsumu(vbool4_t op0, vuint64m1_t op1, vuint32m8_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsumu_vs_u32mf2_u64m1))) +vuint64m1_t vwredsumu(vuint64m1_t op0, vuint32mf2_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwredsumu_vs_u32mf2_u64m1_m))) +vuint64m1_t vwredsumu(vbool64_t op0, vuint64m1_t op1, vuint32mf2_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_u8m1))) +void vsuxei8(uint8_t * op0, vuint8m1_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_u8m1_m))) +void vsuxei8(vbool8_t op0, uint8_t * op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_u8m2))) +void vsuxei8(uint8_t * op0, vuint8m2_t op1, vuint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_u8m2_m))) +void vsuxei8(vbool4_t op0, uint8_t * op1, vuint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_u8m4))) +void vsuxei8(uint8_t * op0, vuint8m4_t op1, vuint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_u8m4_m))) +void vsuxei8(vbool2_t op0, uint8_t * op1, vuint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_u8m8))) +void vsuxei8(uint8_t * op0, vuint8m8_t op1, vuint8m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_u8m8_m))) +void vsuxei8(vbool1_t op0, uint8_t * op1, vuint8m8_t op2, vuint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_u8mf2))) +void vsuxei8(uint8_t * op0, vuint8mf2_t op1, vuint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_u8mf2_m))) +void vsuxei8(vbool16_t op0, uint8_t * op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_u8mf4))) +void vsuxei8(uint8_t * op0, vuint8mf4_t op1, vuint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_u8mf4_m))) +void vsuxei8(vbool32_t op0, uint8_t * op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_u8mf8))) +void vsuxei8(uint8_t * op0, vuint8mf8_t op1, vuint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_u8mf8_m))) +void vsuxei8(vbool64_t op0, uint8_t * op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmnot_m_b8))) +vbool8_t vmnot(vbool8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmnot_m_b4))) +vbool4_t vmnot(vbool4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmnot_m_b2))) +vbool2_t vmnot(vbool2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmnot_m_b1))) +vbool1_t vmnot(vbool1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmnot_m_b16))) +vbool16_t vmnot(vbool16_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmnot_m_b32))) +vbool32_t vmnot(vbool32_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmnot_m_b64))) +vbool64_t vmnot(vbool64_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_viota_m_u8m1_m))) +vuint8m1_t viota(vbool8_t op0, vuint8m1_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_viota_m_u8m2_m))) +vuint8m2_t viota(vbool4_t op0, vuint8m2_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_viota_m_u8m4_m))) +vuint8m4_t viota(vbool2_t op0, vuint8m4_t op1, vbool2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_viota_m_u8m8_m))) +vuint8m8_t viota(vbool1_t op0, vuint8m8_t op1, vbool1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_viota_m_u8mf2_m))) +vuint8mf2_t viota(vbool16_t op0, vuint8mf2_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_viota_m_u8mf4_m))) +vuint8mf4_t viota(vbool32_t op0, vuint8mf4_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_viota_m_u8mf8_m))) +vuint8mf8_t viota(vbool64_t op0, vuint8mf8_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_viota_m_u16m1_m))) +vuint16m1_t viota(vbool16_t op0, vuint16m1_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_viota_m_u16m2_m))) +vuint16m2_t viota(vbool8_t op0, vuint16m2_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_viota_m_u16m4_m))) +vuint16m4_t viota(vbool4_t op0, vuint16m4_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_viota_m_u16m8_m))) +vuint16m8_t viota(vbool2_t op0, vuint16m8_t op1, vbool2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_viota_m_u16mf2_m))) +vuint16mf2_t viota(vbool32_t op0, vuint16mf2_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_viota_m_u16mf4_m))) +vuint16mf4_t viota(vbool64_t op0, vuint16mf4_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_viota_m_u32m1_m))) +vuint32m1_t viota(vbool32_t op0, vuint32m1_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_viota_m_u32m2_m))) +vuint32m2_t viota(vbool16_t op0, vuint32m2_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_viota_m_u32m4_m))) +vuint32m4_t viota(vbool8_t op0, vuint32m4_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_viota_m_u32m8_m))) +vuint32m8_t viota(vbool4_t op0, vuint32m8_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_viota_m_u32mf2_m))) +vuint32mf2_t viota(vbool64_t op0, vuint32mf2_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_viota_m_u64m1_m))) +vuint64m1_t viota(vbool64_t op0, vuint64m1_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_viota_m_u64m2_m))) +vuint64m2_t viota(vbool32_t op0, vuint64m2_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_viota_m_u64m4_m))) +vuint64m4_t viota(vbool16_t op0, vuint64m4_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_viota_m_u64m8_m))) +vuint64m8_t viota(vbool8_t op0, vuint64m8_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vid_v_i8m1_m))) +vint8m1_t vid(vbool8_t op0, vint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vid_v_i8m2_m))) +vint8m2_t vid(vbool4_t op0, vint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vid_v_i8m4_m))) +vint8m4_t vid(vbool2_t op0, vint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vid_v_i8m8_m))) +vint8m8_t vid(vbool1_t op0, vint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vid_v_i8mf2_m))) +vint8mf2_t vid(vbool16_t op0, vint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vid_v_i8mf4_m))) +vint8mf4_t vid(vbool32_t op0, vint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vid_v_i8mf8_m))) +vint8mf8_t vid(vbool64_t op0, vint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vid_v_i16m1_m))) +vint16m1_t vid(vbool16_t op0, vint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vid_v_i16m2_m))) +vint16m2_t vid(vbool8_t op0, vint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vid_v_i16m4_m))) +vint16m4_t vid(vbool4_t op0, vint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vid_v_i16m8_m))) +vint16m8_t vid(vbool2_t op0, vint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vid_v_i16mf2_m))) +vint16mf2_t vid(vbool32_t op0, vint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vid_v_i16mf4_m))) +vint16mf4_t vid(vbool64_t op0, vint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vid_v_i32m1_m))) +vint32m1_t vid(vbool32_t op0, vint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vid_v_i32m2_m))) +vint32m2_t vid(vbool16_t op0, vint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vid_v_i32m4_m))) +vint32m4_t vid(vbool8_t op0, vint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vid_v_i32m8_m))) +vint32m8_t vid(vbool4_t op0, vint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vid_v_i32mf2_m))) +vint32mf2_t vid(vbool64_t op0, vint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vid_v_i64m1_m))) +vint64m1_t vid(vbool64_t op0, vint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vid_v_i64m2_m))) +vint64m2_t vid(vbool32_t op0, vint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vid_v_i64m4_m))) +vint64m4_t vid(vbool16_t op0, vint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vid_v_i64m8_m))) +vint64m8_t vid(vbool8_t op0, vint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vid_v_u8m1_m))) +vuint8m1_t vid(vbool8_t op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vid_v_u8m2_m))) +vuint8m2_t vid(vbool4_t op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vid_v_u8m4_m))) +vuint8m4_t vid(vbool2_t op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vid_v_u8m8_m))) +vuint8m8_t vid(vbool1_t op0, vuint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vid_v_u8mf2_m))) +vuint8mf2_t vid(vbool16_t op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vid_v_u8mf4_m))) +vuint8mf4_t vid(vbool32_t op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vid_v_u8mf8_m))) +vuint8mf8_t vid(vbool64_t op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vid_v_u16m1_m))) +vuint16m1_t vid(vbool16_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vid_v_u16m2_m))) +vuint16m2_t vid(vbool8_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vid_v_u16m4_m))) +vuint16m4_t vid(vbool4_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vid_v_u16m8_m))) +vuint16m8_t vid(vbool2_t op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vid_v_u16mf2_m))) +vuint16mf2_t vid(vbool32_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vid_v_u16mf4_m))) +vuint16mf4_t vid(vbool64_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vid_v_u32m1_m))) +vuint32m1_t vid(vbool32_t op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vid_v_u32m2_m))) +vuint32m2_t vid(vbool16_t op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vid_v_u32m4_m))) +vuint32m4_t vid(vbool8_t op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vid_v_u32m8_m))) +vuint32m8_t vid(vbool4_t op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vid_v_u32mf2_m))) +vuint32mf2_t vid(vbool64_t op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vid_v_u64m1_m))) +vuint64m1_t vid(vbool64_t op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vid_v_u64m2_m))) +vuint64m2_t vid(vbool32_t op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vid_v_u64m4_m))) +vuint64m4_t vid(vbool16_t op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vid_v_u64m8_m))) +vuint64m8_t vid(vbool8_t op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_x_s_i8m1_i8))) +int8_t vmv_x(vint8m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_x_s_i8m2_i8))) +int8_t vmv_x(vint8m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_x_s_i8m4_i8))) +int8_t vmv_x(vint8m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_x_s_i8m8_i8))) +int8_t vmv_x(vint8m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_x_s_i8mf2_i8))) +int8_t vmv_x(vint8mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_x_s_i8mf4_i8))) +int8_t vmv_x(vint8mf4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_x_s_i8mf8_i8))) +int8_t vmv_x(vint8mf8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_x_s_i16m1_i16))) +int16_t vmv_x(vint16m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_x_s_i16m2_i16))) +int16_t vmv_x(vint16m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_x_s_i16m4_i16))) +int16_t vmv_x(vint16m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_x_s_i16m8_i16))) +int16_t vmv_x(vint16m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_x_s_i16mf2_i16))) +int16_t vmv_x(vint16mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_x_s_i16mf4_i16))) +int16_t vmv_x(vint16mf4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_x_s_i32m1_i32))) +int32_t vmv_x(vint32m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_x_s_i32m2_i32))) +int32_t vmv_x(vint32m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_x_s_i32m4_i32))) +int32_t vmv_x(vint32m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_x_s_i32m8_i32))) +int32_t vmv_x(vint32m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_x_s_i32mf2_i32))) +int32_t vmv_x(vint32mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_x_s_i64m1_i64))) +int64_t vmv_x(vint64m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_x_s_i64m2_i64))) +int64_t vmv_x(vint64m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_x_s_i64m4_i64))) +int64_t vmv_x(vint64m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_x_s_i64m8_i64))) +int64_t vmv_x(vint64m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_x_s_u8m1_u8))) +uint8_t vmv_x(vuint8m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_x_s_u8m2_u8))) +uint8_t vmv_x(vuint8m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_x_s_u8m4_u8))) +uint8_t vmv_x(vuint8m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_x_s_u8m8_u8))) +uint8_t vmv_x(vuint8m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_x_s_u8mf2_u8))) +uint8_t vmv_x(vuint8mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_x_s_u8mf4_u8))) +uint8_t vmv_x(vuint8mf4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_x_s_u8mf8_u8))) +uint8_t vmv_x(vuint8mf8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_x_s_u16m1_u16))) +uint16_t vmv_x(vuint16m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_x_s_u16m2_u16))) +uint16_t vmv_x(vuint16m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_x_s_u16m4_u16))) +uint16_t vmv_x(vuint16m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_x_s_u16m8_u16))) +uint16_t vmv_x(vuint16m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_x_s_u16mf2_u16))) +uint16_t vmv_x(vuint16mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_x_s_u16mf4_u16))) +uint16_t vmv_x(vuint16mf4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_x_s_u32m1_u32))) +uint32_t vmv_x(vuint32m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_x_s_u32m2_u32))) +uint32_t vmv_x(vuint32m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_x_s_u32m4_u32))) +uint32_t vmv_x(vuint32m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_x_s_u32m8_u32))) +uint32_t vmv_x(vuint32m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_x_s_u32mf2_u32))) +uint32_t vmv_x(vuint32mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_x_s_u64m1_u64))) +uint64_t vmv_x(vuint64m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_x_s_u64m2_u64))) +uint64_t vmv_x(vuint64m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_x_s_u64m4_u64))) +uint64_t vmv_x(vuint64m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_x_s_u64m8_u64))) +uint64_t vmv_x(vuint64m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_s_x_i8m1))) +vint8m1_t vmv_s(vint8m1_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_s_x_i8m2))) +vint8m2_t vmv_s(vint8m2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_s_x_i8m4))) +vint8m4_t vmv_s(vint8m4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_s_x_i8m8))) +vint8m8_t vmv_s(vint8m8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_s_x_i8mf2))) +vint8mf2_t vmv_s(vint8mf2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_s_x_i8mf4))) +vint8mf4_t vmv_s(vint8mf4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_s_x_i8mf8))) +vint8mf8_t vmv_s(vint8mf8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_s_x_i16m1))) +vint16m1_t vmv_s(vint16m1_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_s_x_i16m2))) +vint16m2_t vmv_s(vint16m2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_s_x_i16m4))) +vint16m4_t vmv_s(vint16m4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_s_x_i16m8))) +vint16m8_t vmv_s(vint16m8_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_s_x_i16mf2))) +vint16mf2_t vmv_s(vint16mf2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_s_x_i16mf4))) +vint16mf4_t vmv_s(vint16mf4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_s_x_i32m1))) +vint32m1_t vmv_s(vint32m1_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_s_x_i32m2))) +vint32m2_t vmv_s(vint32m2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_s_x_i32m4))) +vint32m4_t vmv_s(vint32m4_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_s_x_i32m8))) +vint32m8_t vmv_s(vint32m8_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_s_x_i32mf2))) +vint32mf2_t vmv_s(vint32mf2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_s_x_i64m1))) +vint64m1_t vmv_s(vint64m1_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_s_x_i64m2))) +vint64m2_t vmv_s(vint64m2_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_s_x_i64m4))) +vint64m4_t vmv_s(vint64m4_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_s_x_i64m8))) +vint64m8_t vmv_s(vint64m8_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_s_x_u8m1))) +vuint8m1_t vmv_s(vuint8m1_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_s_x_u8m2))) +vuint8m2_t vmv_s(vuint8m2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_s_x_u8m4))) +vuint8m4_t vmv_s(vuint8m4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_s_x_u8m8))) +vuint8m8_t vmv_s(vuint8m8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_s_x_u8mf2))) +vuint8mf2_t vmv_s(vuint8mf2_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_s_x_u8mf4))) +vuint8mf4_t vmv_s(vuint8mf4_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_s_x_u8mf8))) +vuint8mf8_t vmv_s(vuint8mf8_t op0, uint8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_s_x_u16m1))) +vuint16m1_t vmv_s(vuint16m1_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_s_x_u16m2))) +vuint16m2_t vmv_s(vuint16m2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_s_x_u16m4))) +vuint16m4_t vmv_s(vuint16m4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_s_x_u16m8))) +vuint16m8_t vmv_s(vuint16m8_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_s_x_u16mf2))) +vuint16mf2_t vmv_s(vuint16mf2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_s_x_u16mf4))) +vuint16mf4_t vmv_s(vuint16mf4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_s_x_u32m1))) +vuint32m1_t vmv_s(vuint32m1_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_s_x_u32m2))) +vuint32m2_t vmv_s(vuint32m2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_s_x_u32m4))) +vuint32m4_t vmv_s(vuint32m4_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_s_x_u32m8))) +vuint32m8_t vmv_s(vuint32m8_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_s_x_u32mf2))) +vuint32mf2_t vmv_s(vuint32mf2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_s_x_u64m1))) +vuint64m1_t vmv_s(vuint64m1_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_s_x_u64m2))) +vuint64m2_t vmv_s(vuint64m2_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_s_x_u64m4))) +vuint64m4_t vmv_s(vuint64m4_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_s_x_u64m8))) +vuint64m8_t vmv_s(vuint64m8_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmv_s_x_u16m1))) +vuint16m1_t vfmv_s(vuint16m1_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmv_s_x_u16m2))) +vuint16m2_t vfmv_s(vuint16m2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmv_s_x_u16m4))) +vuint16m4_t vfmv_s(vuint16m4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmv_s_x_u16m8))) +vuint16m8_t vfmv_s(vuint16m8_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmv_s_x_u16mf2))) +vuint16mf2_t vfmv_s(vuint16mf2_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmv_s_x_u16mf4))) +vuint16mf4_t vfmv_s(vuint16mf4_t op0, uint16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmv_s_x_u32m1))) +vuint32m1_t vfmv_s(vuint32m1_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmv_s_x_u32m2))) +vuint32m2_t vfmv_s(vuint32m2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmv_s_x_u32m4))) +vuint32m4_t vfmv_s(vuint32m4_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmv_s_x_u32m8))) +vuint32m8_t vfmv_s(vuint32m8_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmv_s_x_u32mf2))) +vuint32mf2_t vfmv_s(vuint32mf2_t op0, uint32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmv_s_x_u64m1))) +vuint64m1_t vfmv_s(vuint64m1_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmv_s_x_u64m2))) +vuint64m2_t vfmv_s(vuint64m2_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmv_s_x_u64m4))) +vuint64m4_t vfmv_s(vuint64m4_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmv_s_x_u64m8))) +vuint64m8_t vfmv_s(vuint64m8_t op0, uint64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_i8m1))) +vint8m1_t vslideup(vint8m1_t op0, vint8m1_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_i8m1_m))) +vint8m1_t vslideup(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_i8m2))) +vint8m2_t vslideup(vint8m2_t op0, vint8m2_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_i8m2_m))) +vint8m2_t vslideup(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_i8m4))) +vint8m4_t vslideup(vint8m4_t op0, vint8m4_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_i8m4_m))) +vint8m4_t vslideup(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_i8m8))) +vint8m8_t vslideup(vint8m8_t op0, vint8m8_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_i8m8_m))) +vint8m8_t vslideup(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_i8mf2))) +vint8mf2_t vslideup(vint8mf2_t op0, vint8mf2_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_i8mf2_m))) +vint8mf2_t vslideup(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_i8mf4))) +vint8mf4_t vslideup(vint8mf4_t op0, vint8mf4_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_i8mf4_m))) +vint8mf4_t vslideup(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_i8mf8))) +vint8mf8_t vslideup(vint8mf8_t op0, vint8mf8_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_i8mf8_m))) +vint8mf8_t vslideup(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_i16m1))) +vint16m1_t vslideup(vint16m1_t op0, vint16m1_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_i16m1_m))) +vint16m1_t vslideup(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_i16m2))) +vint16m2_t vslideup(vint16m2_t op0, vint16m2_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_i16m2_m))) +vint16m2_t vslideup(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_i16m4))) +vint16m4_t vslideup(vint16m4_t op0, vint16m4_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_i16m4_m))) +vint16m4_t vslideup(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_i16m8))) +vint16m8_t vslideup(vint16m8_t op0, vint16m8_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_i16m8_m))) +vint16m8_t vslideup(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_i16mf2))) +vint16mf2_t vslideup(vint16mf2_t op0, vint16mf2_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_i16mf2_m))) +vint16mf2_t vslideup(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_i16mf4))) +vint16mf4_t vslideup(vint16mf4_t op0, vint16mf4_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_i16mf4_m))) +vint16mf4_t vslideup(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_i32m1))) +vint32m1_t vslideup(vint32m1_t op0, vint32m1_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_i32m1_m))) +vint32m1_t vslideup(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_i32m2))) +vint32m2_t vslideup(vint32m2_t op0, vint32m2_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_i32m2_m))) +vint32m2_t vslideup(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_i32m4))) +vint32m4_t vslideup(vint32m4_t op0, vint32m4_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_i32m4_m))) +vint32m4_t vslideup(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_i32m8))) +vint32m8_t vslideup(vint32m8_t op0, vint32m8_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_i32m8_m))) +vint32m8_t vslideup(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_i32mf2))) +vint32mf2_t vslideup(vint32mf2_t op0, vint32mf2_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_i32mf2_m))) +vint32mf2_t vslideup(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_i64m1))) +vint64m1_t vslideup(vint64m1_t op0, vint64m1_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_i64m1_m))) +vint64m1_t vslideup(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_i64m2))) +vint64m2_t vslideup(vint64m2_t op0, vint64m2_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_i64m2_m))) +vint64m2_t vslideup(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_i64m4))) +vint64m4_t vslideup(vint64m4_t op0, vint64m4_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_i64m4_m))) +vint64m4_t vslideup(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_i64m8))) +vint64m8_t vslideup(vint64m8_t op0, vint64m8_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_i64m8_m))) +vint64m8_t vslideup(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_u8m1))) +vuint8m1_t vslideup(vuint8m1_t op0, vuint8m1_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_u8m1_m))) +vuint8m1_t vslideup(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_u8m2))) +vuint8m2_t vslideup(vuint8m2_t op0, vuint8m2_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_u8m2_m))) +vuint8m2_t vslideup(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_u8m4))) +vuint8m4_t vslideup(vuint8m4_t op0, vuint8m4_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_u8m4_m))) +vuint8m4_t vslideup(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_u8m8))) +vuint8m8_t vslideup(vuint8m8_t op0, vuint8m8_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_u8m8_m))) +vuint8m8_t vslideup(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_u8mf2))) +vuint8mf2_t vslideup(vuint8mf2_t op0, vuint8mf2_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_u8mf2_m))) +vuint8mf2_t vslideup(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_u8mf4))) +vuint8mf4_t vslideup(vuint8mf4_t op0, vuint8mf4_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_u8mf4_m))) +vuint8mf4_t vslideup(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_u8mf8))) +vuint8mf8_t vslideup(vuint8mf8_t op0, vuint8mf8_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_u8mf8_m))) +vuint8mf8_t vslideup(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_u16m1))) +vuint16m1_t vslideup(vuint16m1_t op0, vuint16m1_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_u16m1_m))) +vuint16m1_t vslideup(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_u16m2))) +vuint16m2_t vslideup(vuint16m2_t op0, vuint16m2_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_u16m2_m))) +vuint16m2_t vslideup(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_u16m4))) +vuint16m4_t vslideup(vuint16m4_t op0, vuint16m4_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_u16m4_m))) +vuint16m4_t vslideup(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_u16m8))) +vuint16m8_t vslideup(vuint16m8_t op0, vuint16m8_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_u16m8_m))) +vuint16m8_t vslideup(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_u16mf2))) +vuint16mf2_t vslideup(vuint16mf2_t op0, vuint16mf2_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_u16mf2_m))) +vuint16mf2_t vslideup(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_u16mf4))) +vuint16mf4_t vslideup(vuint16mf4_t op0, vuint16mf4_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_u16mf4_m))) +vuint16mf4_t vslideup(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_u32m1))) +vuint32m1_t vslideup(vuint32m1_t op0, vuint32m1_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_u32m1_m))) +vuint32m1_t vslideup(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_u32m2))) +vuint32m2_t vslideup(vuint32m2_t op0, vuint32m2_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_u32m2_m))) +vuint32m2_t vslideup(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_u32m4))) +vuint32m4_t vslideup(vuint32m4_t op0, vuint32m4_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_u32m4_m))) +vuint32m4_t vslideup(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_u32m8))) +vuint32m8_t vslideup(vuint32m8_t op0, vuint32m8_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_u32m8_m))) +vuint32m8_t vslideup(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_u32mf2))) +vuint32mf2_t vslideup(vuint32mf2_t op0, vuint32mf2_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_u32mf2_m))) +vuint32mf2_t vslideup(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_u64m1))) +vuint64m1_t vslideup(vuint64m1_t op0, vuint64m1_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_u64m1_m))) +vuint64m1_t vslideup(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_u64m2))) +vuint64m2_t vslideup(vuint64m2_t op0, vuint64m2_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_u64m2_m))) +vuint64m2_t vslideup(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_u64m4))) +vuint64m4_t vslideup(vuint64m4_t op0, vuint64m4_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_u64m4_m))) +vuint64m4_t vslideup(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_u64m8))) +vuint64m8_t vslideup(vuint64m8_t op0, vuint64m8_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_u64m8_m))) +vuint64m8_t vslideup(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_i8m1))) +vint8m1_t vslidedown(vint8m1_t op0, vint8m1_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_i8m1_m))) +vint8m1_t vslidedown(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_i8m2))) +vint8m2_t vslidedown(vint8m2_t op0, vint8m2_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_i8m2_m))) +vint8m2_t vslidedown(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_i8m4))) +vint8m4_t vslidedown(vint8m4_t op0, vint8m4_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_i8m4_m))) +vint8m4_t vslidedown(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_i8m8))) +vint8m8_t vslidedown(vint8m8_t op0, vint8m8_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_i8m8_m))) +vint8m8_t vslidedown(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_i8mf2))) +vint8mf2_t vslidedown(vint8mf2_t op0, vint8mf2_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_i8mf2_m))) +vint8mf2_t vslidedown(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_i8mf4))) +vint8mf4_t vslidedown(vint8mf4_t op0, vint8mf4_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_i8mf4_m))) +vint8mf4_t vslidedown(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_i8mf8))) +vint8mf8_t vslidedown(vint8mf8_t op0, vint8mf8_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_i8mf8_m))) +vint8mf8_t vslidedown(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_i16m1))) +vint16m1_t vslidedown(vint16m1_t op0, vint16m1_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_i16m1_m))) +vint16m1_t vslidedown(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_i16m2))) +vint16m2_t vslidedown(vint16m2_t op0, vint16m2_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_i16m2_m))) +vint16m2_t vslidedown(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_i16m4))) +vint16m4_t vslidedown(vint16m4_t op0, vint16m4_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_i16m4_m))) +vint16m4_t vslidedown(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_i16m8))) +vint16m8_t vslidedown(vint16m8_t op0, vint16m8_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_i16m8_m))) +vint16m8_t vslidedown(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_i16mf2))) +vint16mf2_t vslidedown(vint16mf2_t op0, vint16mf2_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_i16mf2_m))) +vint16mf2_t vslidedown(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_i16mf4))) +vint16mf4_t vslidedown(vint16mf4_t op0, vint16mf4_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_i16mf4_m))) +vint16mf4_t vslidedown(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_i32m1))) +vint32m1_t vslidedown(vint32m1_t op0, vint32m1_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_i32m1_m))) +vint32m1_t vslidedown(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_i32m2))) +vint32m2_t vslidedown(vint32m2_t op0, vint32m2_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_i32m2_m))) +vint32m2_t vslidedown(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_i32m4))) +vint32m4_t vslidedown(vint32m4_t op0, vint32m4_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_i32m4_m))) +vint32m4_t vslidedown(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_i32m8))) +vint32m8_t vslidedown(vint32m8_t op0, vint32m8_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_i32m8_m))) +vint32m8_t vslidedown(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_i32mf2))) +vint32mf2_t vslidedown(vint32mf2_t op0, vint32mf2_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_i32mf2_m))) +vint32mf2_t vslidedown(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_i64m1))) +vint64m1_t vslidedown(vint64m1_t op0, vint64m1_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_i64m1_m))) +vint64m1_t vslidedown(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_i64m2))) +vint64m2_t vslidedown(vint64m2_t op0, vint64m2_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_i64m2_m))) +vint64m2_t vslidedown(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_i64m4))) +vint64m4_t vslidedown(vint64m4_t op0, vint64m4_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_i64m4_m))) +vint64m4_t vslidedown(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_i64m8))) +vint64m8_t vslidedown(vint64m8_t op0, vint64m8_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_i64m8_m))) +vint64m8_t vslidedown(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_u8m1))) +vuint8m1_t vslidedown(vuint8m1_t op0, vuint8m1_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_u8m1_m))) +vuint8m1_t vslidedown(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_u8m2))) +vuint8m2_t vslidedown(vuint8m2_t op0, vuint8m2_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_u8m2_m))) +vuint8m2_t vslidedown(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_u8m4))) +vuint8m4_t vslidedown(vuint8m4_t op0, vuint8m4_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_u8m4_m))) +vuint8m4_t vslidedown(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_u8m8))) +vuint8m8_t vslidedown(vuint8m8_t op0, vuint8m8_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_u8m8_m))) +vuint8m8_t vslidedown(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_u8mf2))) +vuint8mf2_t vslidedown(vuint8mf2_t op0, vuint8mf2_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_u8mf2_m))) +vuint8mf2_t vslidedown(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_u8mf4))) +vuint8mf4_t vslidedown(vuint8mf4_t op0, vuint8mf4_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_u8mf4_m))) +vuint8mf4_t vslidedown(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_u8mf8))) +vuint8mf8_t vslidedown(vuint8mf8_t op0, vuint8mf8_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_u8mf8_m))) +vuint8mf8_t vslidedown(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_u16m1))) +vuint16m1_t vslidedown(vuint16m1_t op0, vuint16m1_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_u16m1_m))) +vuint16m1_t vslidedown(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_u16m2))) +vuint16m2_t vslidedown(vuint16m2_t op0, vuint16m2_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_u16m2_m))) +vuint16m2_t vslidedown(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_u16m4))) +vuint16m4_t vslidedown(vuint16m4_t op0, vuint16m4_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_u16m4_m))) +vuint16m4_t vslidedown(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_u16m8))) +vuint16m8_t vslidedown(vuint16m8_t op0, vuint16m8_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_u16m8_m))) +vuint16m8_t vslidedown(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_u16mf2))) +vuint16mf2_t vslidedown(vuint16mf2_t op0, vuint16mf2_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_u16mf2_m))) +vuint16mf2_t vslidedown(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_u16mf4))) +vuint16mf4_t vslidedown(vuint16mf4_t op0, vuint16mf4_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_u16mf4_m))) +vuint16mf4_t vslidedown(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_u32m1))) +vuint32m1_t vslidedown(vuint32m1_t op0, vuint32m1_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_u32m1_m))) +vuint32m1_t vslidedown(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_u32m2))) +vuint32m2_t vslidedown(vuint32m2_t op0, vuint32m2_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_u32m2_m))) +vuint32m2_t vslidedown(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_u32m4))) +vuint32m4_t vslidedown(vuint32m4_t op0, vuint32m4_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_u32m4_m))) +vuint32m4_t vslidedown(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_u32m8))) +vuint32m8_t vslidedown(vuint32m8_t op0, vuint32m8_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_u32m8_m))) +vuint32m8_t vslidedown(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_u32mf2))) +vuint32mf2_t vslidedown(vuint32mf2_t op0, vuint32mf2_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_u32mf2_m))) +vuint32mf2_t vslidedown(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_u64m1))) +vuint64m1_t vslidedown(vuint64m1_t op0, vuint64m1_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_u64m1_m))) +vuint64m1_t vslidedown(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_u64m2))) +vuint64m2_t vslidedown(vuint64m2_t op0, vuint64m2_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_u64m2_m))) +vuint64m2_t vslidedown(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_u64m4))) +vuint64m4_t vslidedown(vuint64m4_t op0, vuint64m4_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_u64m4_m))) +vuint64m4_t vslidedown(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_u64m8))) +vuint64m8_t vslidedown(vuint64m8_t op0, vuint64m8_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_u64m8_m))) +vuint64m8_t vslidedown(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_i8m1))) +vint8m1_t vslide1up(vint8m1_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_i8m1_m))) +vint8m1_t vslide1up(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_i8m2))) +vint8m2_t vslide1up(vint8m2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_i8m2_m))) +vint8m2_t vslide1up(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_i8m4))) +vint8m4_t vslide1up(vint8m4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_i8m4_m))) +vint8m4_t vslide1up(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_i8m8))) +vint8m8_t vslide1up(vint8m8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_i8m8_m))) +vint8m8_t vslide1up(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_i8mf2))) +vint8mf2_t vslide1up(vint8mf2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_i8mf2_m))) +vint8mf2_t vslide1up(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_i8mf4))) +vint8mf4_t vslide1up(vint8mf4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_i8mf4_m))) +vint8mf4_t vslide1up(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_i8mf8))) +vint8mf8_t vslide1up(vint8mf8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_i8mf8_m))) +vint8mf8_t vslide1up(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_i16m1))) +vint16m1_t vslide1up(vint16m1_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_i16m1_m))) +vint16m1_t vslide1up(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_i16m2))) +vint16m2_t vslide1up(vint16m2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_i16m2_m))) +vint16m2_t vslide1up(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_i16m4))) +vint16m4_t vslide1up(vint16m4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_i16m4_m))) +vint16m4_t vslide1up(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_i16m8))) +vint16m8_t vslide1up(vint16m8_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_i16m8_m))) +vint16m8_t vslide1up(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_i16mf2))) +vint16mf2_t vslide1up(vint16mf2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_i16mf2_m))) +vint16mf2_t vslide1up(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_i16mf4))) +vint16mf4_t vslide1up(vint16mf4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_i16mf4_m))) +vint16mf4_t vslide1up(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_i32m1))) +vint32m1_t vslide1up(vint32m1_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_i32m1_m))) +vint32m1_t vslide1up(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_i32m2))) +vint32m2_t vslide1up(vint32m2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_i32m2_m))) +vint32m2_t vslide1up(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_i32m4))) +vint32m4_t vslide1up(vint32m4_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_i32m4_m))) +vint32m4_t vslide1up(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_i32m8))) +vint32m8_t vslide1up(vint32m8_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_i32m8_m))) +vint32m8_t vslide1up(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_i32mf2))) +vint32mf2_t vslide1up(vint32mf2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_i32mf2_m))) +vint32mf2_t vslide1up(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_i64m1))) +vint64m1_t vslide1up(vint64m1_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_i64m1_m))) +vint64m1_t vslide1up(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_i64m2))) +vint64m2_t vslide1up(vint64m2_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_i64m2_m))) +vint64m2_t vslide1up(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_i64m4))) +vint64m4_t vslide1up(vint64m4_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_i64m4_m))) +vint64m4_t vslide1up(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_i64m8))) +vint64m8_t vslide1up(vint64m8_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_i64m8_m))) +vint64m8_t vslide1up(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_u8m1))) +vuint8m1_t vslide1up(vuint8m1_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_u8m1_m))) +vuint8m1_t vslide1up(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_u8m2))) +vuint8m2_t vslide1up(vuint8m2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_u8m2_m))) +vuint8m2_t vslide1up(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_u8m4))) +vuint8m4_t vslide1up(vuint8m4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_u8m4_m))) +vuint8m4_t vslide1up(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_u8m8))) +vuint8m8_t vslide1up(vuint8m8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_u8m8_m))) +vuint8m8_t vslide1up(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_u8mf2))) +vuint8mf2_t vslide1up(vuint8mf2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_u8mf2_m))) +vuint8mf2_t vslide1up(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_u8mf4))) +vuint8mf4_t vslide1up(vuint8mf4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_u8mf4_m))) +vuint8mf4_t vslide1up(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_u8mf8))) +vuint8mf8_t vslide1up(vuint8mf8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_u8mf8_m))) +vuint8mf8_t vslide1up(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_u16m1))) +vuint16m1_t vslide1up(vuint16m1_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_u16m1_m))) +vuint16m1_t vslide1up(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_u16m2))) +vuint16m2_t vslide1up(vuint16m2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_u16m2_m))) +vuint16m2_t vslide1up(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_u16m4))) +vuint16m4_t vslide1up(vuint16m4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_u16m4_m))) +vuint16m4_t vslide1up(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_u16m8))) +vuint16m8_t vslide1up(vuint16m8_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_u16m8_m))) +vuint16m8_t vslide1up(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_u16mf2))) +vuint16mf2_t vslide1up(vuint16mf2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_u16mf2_m))) +vuint16mf2_t vslide1up(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_u16mf4))) +vuint16mf4_t vslide1up(vuint16mf4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_u16mf4_m))) +vuint16mf4_t vslide1up(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_u32m1))) +vuint32m1_t vslide1up(vuint32m1_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_u32m1_m))) +vuint32m1_t vslide1up(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_u32m2))) +vuint32m2_t vslide1up(vuint32m2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_u32m2_m))) +vuint32m2_t vslide1up(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_u32m4))) +vuint32m4_t vslide1up(vuint32m4_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_u32m4_m))) +vuint32m4_t vslide1up(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_u32m8))) +vuint32m8_t vslide1up(vuint32m8_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_u32m8_m))) +vuint32m8_t vslide1up(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_u32mf2))) +vuint32mf2_t vslide1up(vuint32mf2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_u32mf2_m))) +vuint32mf2_t vslide1up(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_u64m1))) +vuint64m1_t vslide1up(vuint64m1_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_u64m1_m))) +vuint64m1_t vslide1up(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_u64m2))) +vuint64m2_t vslide1up(vuint64m2_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_u64m2_m))) +vuint64m2_t vslide1up(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_u64m4))) +vuint64m4_t vslide1up(vuint64m4_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_u64m4_m))) +vuint64m4_t vslide1up(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_u64m8))) +vuint64m8_t vslide1up(vuint64m8_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1up_vx_u64m8_m))) +vuint64m8_t vslide1up(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_i8m1))) +vint8m1_t vslide1down(vint8m1_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_i8m1_m))) +vint8m1_t vslide1down(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_i8m2))) +vint8m2_t vslide1down(vint8m2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_i8m2_m))) +vint8m2_t vslide1down(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_i8m4))) +vint8m4_t vslide1down(vint8m4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_i8m4_m))) +vint8m4_t vslide1down(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_i8m8))) +vint8m8_t vslide1down(vint8m8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_i8m8_m))) +vint8m8_t vslide1down(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_i8mf2))) +vint8mf2_t vslide1down(vint8mf2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_i8mf2_m))) +vint8mf2_t vslide1down(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_i8mf4))) +vint8mf4_t vslide1down(vint8mf4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_i8mf4_m))) +vint8mf4_t vslide1down(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_i8mf8))) +vint8mf8_t vslide1down(vint8mf8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_i8mf8_m))) +vint8mf8_t vslide1down(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_i16m1))) +vint16m1_t vslide1down(vint16m1_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_i16m1_m))) +vint16m1_t vslide1down(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_i16m2))) +vint16m2_t vslide1down(vint16m2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_i16m2_m))) +vint16m2_t vslide1down(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_i16m4))) +vint16m4_t vslide1down(vint16m4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_i16m4_m))) +vint16m4_t vslide1down(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_i16m8))) +vint16m8_t vslide1down(vint16m8_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_i16m8_m))) +vint16m8_t vslide1down(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_i16mf2))) +vint16mf2_t vslide1down(vint16mf2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_i16mf2_m))) +vint16mf2_t vslide1down(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_i16mf4))) +vint16mf4_t vslide1down(vint16mf4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_i16mf4_m))) +vint16mf4_t vslide1down(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_i32m1))) +vint32m1_t vslide1down(vint32m1_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_i32m1_m))) +vint32m1_t vslide1down(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_i32m2))) +vint32m2_t vslide1down(vint32m2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_i32m2_m))) +vint32m2_t vslide1down(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_i32m4))) +vint32m4_t vslide1down(vint32m4_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_i32m4_m))) +vint32m4_t vslide1down(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_i32m8))) +vint32m8_t vslide1down(vint32m8_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_i32m8_m))) +vint32m8_t vslide1down(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_i32mf2))) +vint32mf2_t vslide1down(vint32mf2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_i32mf2_m))) +vint32mf2_t vslide1down(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_i64m1))) +vint64m1_t vslide1down(vint64m1_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_i64m1_m))) +vint64m1_t vslide1down(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_i64m2))) +vint64m2_t vslide1down(vint64m2_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_i64m2_m))) +vint64m2_t vslide1down(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_i64m4))) +vint64m4_t vslide1down(vint64m4_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_i64m4_m))) +vint64m4_t vslide1down(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_i64m8))) +vint64m8_t vslide1down(vint64m8_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_i64m8_m))) +vint64m8_t vslide1down(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_u8m1))) +vuint8m1_t vslide1down(vuint8m1_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_u8m1_m))) +vuint8m1_t vslide1down(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_u8m2))) +vuint8m2_t vslide1down(vuint8m2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_u8m2_m))) +vuint8m2_t vslide1down(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_u8m4))) +vuint8m4_t vslide1down(vuint8m4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_u8m4_m))) +vuint8m4_t vslide1down(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_u8m8))) +vuint8m8_t vslide1down(vuint8m8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_u8m8_m))) +vuint8m8_t vslide1down(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_u8mf2))) +vuint8mf2_t vslide1down(vuint8mf2_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_u8mf2_m))) +vuint8mf2_t vslide1down(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_u8mf4))) +vuint8mf4_t vslide1down(vuint8mf4_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_u8mf4_m))) +vuint8mf4_t vslide1down(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_u8mf8))) +vuint8mf8_t vslide1down(vuint8mf8_t op0, int8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_u8mf8_m))) +vuint8mf8_t vslide1down(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, int8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_u16m1))) +vuint16m1_t vslide1down(vuint16m1_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_u16m1_m))) +vuint16m1_t vslide1down(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_u16m2))) +vuint16m2_t vslide1down(vuint16m2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_u16m2_m))) +vuint16m2_t vslide1down(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_u16m4))) +vuint16m4_t vslide1down(vuint16m4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_u16m4_m))) +vuint16m4_t vslide1down(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_u16m8))) +vuint16m8_t vslide1down(vuint16m8_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_u16m8_m))) +vuint16m8_t vslide1down(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_u16mf2))) +vuint16mf2_t vslide1down(vuint16mf2_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_u16mf2_m))) +vuint16mf2_t vslide1down(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_u16mf4))) +vuint16mf4_t vslide1down(vuint16mf4_t op0, int16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_u16mf4_m))) +vuint16mf4_t vslide1down(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, int16_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_u32m1))) +vuint32m1_t vslide1down(vuint32m1_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_u32m1_m))) +vuint32m1_t vslide1down(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_u32m2))) +vuint32m2_t vslide1down(vuint32m2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_u32m2_m))) +vuint32m2_t vslide1down(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_u32m4))) +vuint32m4_t vslide1down(vuint32m4_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_u32m4_m))) +vuint32m4_t vslide1down(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_u32m8))) +vuint32m8_t vslide1down(vuint32m8_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_u32m8_m))) +vuint32m8_t vslide1down(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_u32mf2))) +vuint32mf2_t vslide1down(vuint32mf2_t op0, int32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_u32mf2_m))) +vuint32mf2_t vslide1down(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, int32_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_u64m1))) +vuint64m1_t vslide1down(vuint64m1_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_u64m1_m))) +vuint64m1_t vslide1down(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_u64m2))) +vuint64m2_t vslide1down(vuint64m2_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_u64m2_m))) +vuint64m2_t vslide1down(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_u64m4))) +vuint64m4_t vslide1down(vuint64m4_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_u64m4_m))) +vuint64m4_t vslide1down(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_u64m8))) +vuint64m8_t vslide1down(vuint64m8_t op0, int64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslide1down_vx_u64m8_m))) +vuint64m8_t vslide1down(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, int64_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_i8m1))) +vint8m1_t vrgather(vint8m1_t op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_i8m1_m))) +vint8m1_t vrgather(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_i8m2))) +vint8m2_t vrgather(vint8m2_t op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_i8m2_m))) +vint8m2_t vrgather(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_i8m4))) +vint8m4_t vrgather(vint8m4_t op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_i8m4_m))) +vint8m4_t vrgather(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_i8m8))) +vint8m8_t vrgather(vint8m8_t op0, vuint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_i8m8_m))) +vint8m8_t vrgather(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, vuint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_i8mf2))) +vint8mf2_t vrgather(vint8mf2_t op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_i8mf2_m))) +vint8mf2_t vrgather(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_i8mf4))) +vint8mf4_t vrgather(vint8mf4_t op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_i8mf4_m))) +vint8mf4_t vrgather(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_i8mf8))) +vint8mf8_t vrgather(vint8mf8_t op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_i8mf8_m))) +vint8mf8_t vrgather(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_i16m1))) +vint16m1_t vrgather(vint16m1_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_i16m1_m))) +vint16m1_t vrgather(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_i16m2))) +vint16m2_t vrgather(vint16m2_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_i16m2_m))) +vint16m2_t vrgather(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_i16m4))) +vint16m4_t vrgather(vint16m4_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_i16m4_m))) +vint16m4_t vrgather(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_i16m8))) +vint16m8_t vrgather(vint16m8_t op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_i16m8_m))) +vint16m8_t vrgather(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_i16mf2))) +vint16mf2_t vrgather(vint16mf2_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_i16mf2_m))) +vint16mf2_t vrgather(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_i16mf4))) +vint16mf4_t vrgather(vint16mf4_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_i16mf4_m))) +vint16mf4_t vrgather(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_i32m1))) +vint32m1_t vrgather(vint32m1_t op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_i32m1_m))) +vint32m1_t vrgather(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_i32m2))) +vint32m2_t vrgather(vint32m2_t op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_i32m2_m))) +vint32m2_t vrgather(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_i32m4))) +vint32m4_t vrgather(vint32m4_t op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_i32m4_m))) +vint32m4_t vrgather(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_i32m8))) +vint32m8_t vrgather(vint32m8_t op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_i32m8_m))) +vint32m8_t vrgather(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_i32mf2))) +vint32mf2_t vrgather(vint32mf2_t op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_i32mf2_m))) +vint32mf2_t vrgather(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_i64m1))) +vint64m1_t vrgather(vint64m1_t op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_i64m1_m))) +vint64m1_t vrgather(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_i64m2))) +vint64m2_t vrgather(vint64m2_t op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_i64m2_m))) +vint64m2_t vrgather(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_i64m4))) +vint64m4_t vrgather(vint64m4_t op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_i64m4_m))) +vint64m4_t vrgather(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_i64m8))) +vint64m8_t vrgather(vint64m8_t op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_i64m8_m))) +vint64m8_t vrgather(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_i8m1))) +vint8m1_t vrgather(vint8m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_i8m1_m))) +vint8m1_t vrgather(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_i8m2))) +vint8m2_t vrgather(vint8m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_i8m2_m))) +vint8m2_t vrgather(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_i8m4))) +vint8m4_t vrgather(vint8m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_i8m4_m))) +vint8m4_t vrgather(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_i8m8))) +vint8m8_t vrgather(vint8m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_i8m8_m))) +vint8m8_t vrgather(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_i8mf2))) +vint8mf2_t vrgather(vint8mf2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_i8mf2_m))) +vint8mf2_t vrgather(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_i8mf4))) +vint8mf4_t vrgather(vint8mf4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_i8mf4_m))) +vint8mf4_t vrgather(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_i8mf8))) +vint8mf8_t vrgather(vint8mf8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_i8mf8_m))) +vint8mf8_t vrgather(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_i16m1))) +vint16m1_t vrgather(vint16m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_i16m1_m))) +vint16m1_t vrgather(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_i16m2))) +vint16m2_t vrgather(vint16m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_i16m2_m))) +vint16m2_t vrgather(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_i16m4))) +vint16m4_t vrgather(vint16m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_i16m4_m))) +vint16m4_t vrgather(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_i16m8))) +vint16m8_t vrgather(vint16m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_i16m8_m))) +vint16m8_t vrgather(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_i16mf2))) +vint16mf2_t vrgather(vint16mf2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_i16mf2_m))) +vint16mf2_t vrgather(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_i16mf4))) +vint16mf4_t vrgather(vint16mf4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_i16mf4_m))) +vint16mf4_t vrgather(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_i32m1))) +vint32m1_t vrgather(vint32m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_i32m1_m))) +vint32m1_t vrgather(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_i32m2))) +vint32m2_t vrgather(vint32m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_i32m2_m))) +vint32m2_t vrgather(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_i32m4))) +vint32m4_t vrgather(vint32m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_i32m4_m))) +vint32m4_t vrgather(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_i32m8))) +vint32m8_t vrgather(vint32m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_i32m8_m))) +vint32m8_t vrgather(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_i32mf2))) +vint32mf2_t vrgather(vint32mf2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_i32mf2_m))) +vint32mf2_t vrgather(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_i64m1))) +vint64m1_t vrgather(vint64m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_i64m1_m))) +vint64m1_t vrgather(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_i64m2))) +vint64m2_t vrgather(vint64m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_i64m2_m))) +vint64m2_t vrgather(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_i64m4))) +vint64m4_t vrgather(vint64m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_i64m4_m))) +vint64m4_t vrgather(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_i64m8))) +vint64m8_t vrgather(vint64m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_i64m8_m))) +vint64m8_t vrgather(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_i8m1))) +vint8m1_t vrgatherei16(vint8m1_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_i8m1_m))) +vint8m1_t vrgatherei16(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_i8m2))) +vint8m2_t vrgatherei16(vint8m2_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_i8m2_m))) +vint8m2_t vrgatherei16(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_i8m4))) +vint8m4_t vrgatherei16(vint8m4_t op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_i8m4_m))) +vint8m4_t vrgatherei16(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_i8mf2))) +vint8mf2_t vrgatherei16(vint8mf2_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_i8mf2_m))) +vint8mf2_t vrgatherei16(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_i8mf4))) +vint8mf4_t vrgatherei16(vint8mf4_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_i8mf4_m))) +vint8mf4_t vrgatherei16(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_i8mf8))) +vint8mf8_t vrgatherei16(vint8mf8_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_i8mf8_m))) +vint8mf8_t vrgatherei16(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_i16m1))) +vint16m1_t vrgatherei16(vint16m1_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_i16m1_m))) +vint16m1_t vrgatherei16(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_i16m2))) +vint16m2_t vrgatherei16(vint16m2_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_i16m2_m))) +vint16m2_t vrgatherei16(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_i16m4))) +vint16m4_t vrgatherei16(vint16m4_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_i16m4_m))) +vint16m4_t vrgatherei16(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_i16m8))) +vint16m8_t vrgatherei16(vint16m8_t op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_i16m8_m))) +vint16m8_t vrgatherei16(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_i16mf2))) +vint16mf2_t vrgatherei16(vint16mf2_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_i16mf2_m))) +vint16mf2_t vrgatherei16(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_i16mf4))) +vint16mf4_t vrgatherei16(vint16mf4_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_i16mf4_m))) +vint16mf4_t vrgatherei16(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_i32m1))) +vint32m1_t vrgatherei16(vint32m1_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_i32m1_m))) +vint32m1_t vrgatherei16(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_i32m2))) +vint32m2_t vrgatherei16(vint32m2_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_i32m2_m))) +vint32m2_t vrgatherei16(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_i32m4))) +vint32m4_t vrgatherei16(vint32m4_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_i32m4_m))) +vint32m4_t vrgatherei16(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_i32m8))) +vint32m8_t vrgatherei16(vint32m8_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_i32m8_m))) +vint32m8_t vrgatherei16(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_i32mf2))) +vint32mf2_t vrgatherei16(vint32mf2_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_i32mf2_m))) +vint32mf2_t vrgatherei16(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_i64m1))) +vint64m1_t vrgatherei16(vint64m1_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_i64m1_m))) +vint64m1_t vrgatherei16(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_i64m2))) +vint64m2_t vrgatherei16(vint64m2_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_i64m2_m))) +vint64m2_t vrgatherei16(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_i64m4))) +vint64m4_t vrgatherei16(vint64m4_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_i64m4_m))) +vint64m4_t vrgatherei16(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_i64m8))) +vint64m8_t vrgatherei16(vint64m8_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_i64m8_m))) +vint64m8_t vrgatherei16(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_u8m1))) +vuint8m1_t vrgather(vuint8m1_t op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_u8m1_m))) +vuint8m1_t vrgather(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_u8m2))) +vuint8m2_t vrgather(vuint8m2_t op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_u8m2_m))) +vuint8m2_t vrgather(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_u8m4))) +vuint8m4_t vrgather(vuint8m4_t op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_u8m4_m))) +vuint8m4_t vrgather(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_u8m8))) +vuint8m8_t vrgather(vuint8m8_t op0, vuint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_u8m8_m))) +vuint8m8_t vrgather(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, vuint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_u8mf2))) +vuint8mf2_t vrgather(vuint8mf2_t op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_u8mf2_m))) +vuint8mf2_t vrgather(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_u8mf4))) +vuint8mf4_t vrgather(vuint8mf4_t op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_u8mf4_m))) +vuint8mf4_t vrgather(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_u8mf8))) +vuint8mf8_t vrgather(vuint8mf8_t op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_u8mf8_m))) +vuint8mf8_t vrgather(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_u16m1))) +vuint16m1_t vrgather(vuint16m1_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_u16m1_m))) +vuint16m1_t vrgather(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_u16m2))) +vuint16m2_t vrgather(vuint16m2_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_u16m2_m))) +vuint16m2_t vrgather(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_u16m4))) +vuint16m4_t vrgather(vuint16m4_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_u16m4_m))) +vuint16m4_t vrgather(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_u16m8))) +vuint16m8_t vrgather(vuint16m8_t op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_u16m8_m))) +vuint16m8_t vrgather(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_u16mf2))) +vuint16mf2_t vrgather(vuint16mf2_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_u16mf2_m))) +vuint16mf2_t vrgather(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_u16mf4))) +vuint16mf4_t vrgather(vuint16mf4_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_u16mf4_m))) +vuint16mf4_t vrgather(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_u32m1))) +vuint32m1_t vrgather(vuint32m1_t op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_u32m1_m))) +vuint32m1_t vrgather(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_u32m2))) +vuint32m2_t vrgather(vuint32m2_t op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_u32m2_m))) +vuint32m2_t vrgather(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_u32m4))) +vuint32m4_t vrgather(vuint32m4_t op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_u32m4_m))) +vuint32m4_t vrgather(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_u32m8))) +vuint32m8_t vrgather(vuint32m8_t op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_u32m8_m))) +vuint32m8_t vrgather(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_u32mf2))) +vuint32mf2_t vrgather(vuint32mf2_t op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_u32mf2_m))) +vuint32mf2_t vrgather(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_u64m1))) +vuint64m1_t vrgather(vuint64m1_t op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_u64m1_m))) +vuint64m1_t vrgather(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_u64m2))) +vuint64m2_t vrgather(vuint64m2_t op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_u64m2_m))) +vuint64m2_t vrgather(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_u64m4))) +vuint64m4_t vrgather(vuint64m4_t op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_u64m4_m))) +vuint64m4_t vrgather(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_u64m8))) +vuint64m8_t vrgather(vuint64m8_t op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_u64m8_m))) +vuint64m8_t vrgather(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_u8m1))) +vuint8m1_t vrgather(vuint8m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_u8m1_m))) +vuint8m1_t vrgather(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_u8m2))) +vuint8m2_t vrgather(vuint8m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_u8m2_m))) +vuint8m2_t vrgather(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_u8m4))) +vuint8m4_t vrgather(vuint8m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_u8m4_m))) +vuint8m4_t vrgather(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_u8m8))) +vuint8m8_t vrgather(vuint8m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_u8m8_m))) +vuint8m8_t vrgather(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_u8mf2))) +vuint8mf2_t vrgather(vuint8mf2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_u8mf2_m))) +vuint8mf2_t vrgather(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_u8mf4))) +vuint8mf4_t vrgather(vuint8mf4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_u8mf4_m))) +vuint8mf4_t vrgather(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_u8mf8))) +vuint8mf8_t vrgather(vuint8mf8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_u8mf8_m))) +vuint8mf8_t vrgather(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_u16m1))) +vuint16m1_t vrgather(vuint16m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_u16m1_m))) +vuint16m1_t vrgather(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_u16m2))) +vuint16m2_t vrgather(vuint16m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_u16m2_m))) +vuint16m2_t vrgather(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_u16m4))) +vuint16m4_t vrgather(vuint16m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_u16m4_m))) +vuint16m4_t vrgather(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_u16m8))) +vuint16m8_t vrgather(vuint16m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_u16m8_m))) +vuint16m8_t vrgather(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_u16mf2))) +vuint16mf2_t vrgather(vuint16mf2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_u16mf2_m))) +vuint16mf2_t vrgather(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_u16mf4))) +vuint16mf4_t vrgather(vuint16mf4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_u16mf4_m))) +vuint16mf4_t vrgather(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_u32m1))) +vuint32m1_t vrgather(vuint32m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_u32m1_m))) +vuint32m1_t vrgather(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_u32m2))) +vuint32m2_t vrgather(vuint32m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_u32m2_m))) +vuint32m2_t vrgather(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_u32m4))) +vuint32m4_t vrgather(vuint32m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_u32m4_m))) +vuint32m4_t vrgather(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_u32m8))) +vuint32m8_t vrgather(vuint32m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_u32m8_m))) +vuint32m8_t vrgather(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_u32mf2))) +vuint32mf2_t vrgather(vuint32mf2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_u32mf2_m))) +vuint32mf2_t vrgather(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_u64m1))) +vuint64m1_t vrgather(vuint64m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_u64m1_m))) +vuint64m1_t vrgather(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_u64m2))) +vuint64m2_t vrgather(vuint64m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_u64m2_m))) +vuint64m2_t vrgather(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_u64m4))) +vuint64m4_t vrgather(vuint64m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_u64m4_m))) +vuint64m4_t vrgather(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_u64m8))) +vuint64m8_t vrgather(vuint64m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_u64m8_m))) +vuint64m8_t vrgather(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_u8m1))) +vuint8m1_t vrgatherei16(vuint8m1_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_u8m1_m))) +vuint8m1_t vrgatherei16(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_u8m2))) +vuint8m2_t vrgatherei16(vuint8m2_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_u8m2_m))) +vuint8m2_t vrgatherei16(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_u8m4))) +vuint8m4_t vrgatherei16(vuint8m4_t op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_u8m4_m))) +vuint8m4_t vrgatherei16(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_u8mf2))) +vuint8mf2_t vrgatherei16(vuint8mf2_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_u8mf2_m))) +vuint8mf2_t vrgatherei16(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_u8mf4))) +vuint8mf4_t vrgatherei16(vuint8mf4_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_u8mf4_m))) +vuint8mf4_t vrgatherei16(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_u8mf8))) +vuint8mf8_t vrgatherei16(vuint8mf8_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_u8mf8_m))) +vuint8mf8_t vrgatherei16(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_u16m1))) +vuint16m1_t vrgatherei16(vuint16m1_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_u16m1_m))) +vuint16m1_t vrgatherei16(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_u16m2))) +vuint16m2_t vrgatherei16(vuint16m2_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_u16m2_m))) +vuint16m2_t vrgatherei16(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_u16m4))) +vuint16m4_t vrgatherei16(vuint16m4_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_u16m4_m))) +vuint16m4_t vrgatherei16(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_u16m8))) +vuint16m8_t vrgatherei16(vuint16m8_t op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_u16m8_m))) +vuint16m8_t vrgatherei16(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_u16mf2))) +vuint16mf2_t vrgatherei16(vuint16mf2_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_u16mf2_m))) +vuint16mf2_t vrgatherei16(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_u16mf4))) +vuint16mf4_t vrgatherei16(vuint16mf4_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_u16mf4_m))) +vuint16mf4_t vrgatherei16(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_u32m1))) +vuint32m1_t vrgatherei16(vuint32m1_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_u32m1_m))) +vuint32m1_t vrgatherei16(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_u32m2))) +vuint32m2_t vrgatherei16(vuint32m2_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_u32m2_m))) +vuint32m2_t vrgatherei16(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_u32m4))) +vuint32m4_t vrgatherei16(vuint32m4_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_u32m4_m))) +vuint32m4_t vrgatherei16(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_u32m8))) +vuint32m8_t vrgatherei16(vuint32m8_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_u32m8_m))) +vuint32m8_t vrgatherei16(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_u32mf2))) +vuint32mf2_t vrgatherei16(vuint32mf2_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_u32mf2_m))) +vuint32mf2_t vrgatherei16(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_u64m1))) +vuint64m1_t vrgatherei16(vuint64m1_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_u64m1_m))) +vuint64m1_t vrgatherei16(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_u64m2))) +vuint64m2_t vrgatherei16(vuint64m2_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_u64m2_m))) +vuint64m2_t vrgatherei16(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_u64m4))) +vuint64m4_t vrgatherei16(vuint64m4_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_u64m4_m))) +vuint64m4_t vrgatherei16(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_u64m8))) +vuint64m8_t vrgatherei16(vuint64m8_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_u64m8_m))) +vuint64m8_t vrgatherei16(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_i8m1))) +vint8m1_t vcompress(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_i8m2))) +vint8m2_t vcompress(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_i8m4))) +vint8m4_t vcompress(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_i8m8))) +vint8m8_t vcompress(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_i8mf2))) +vint8mf2_t vcompress(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_i8mf4))) +vint8mf4_t vcompress(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_i8mf8))) +vint8mf8_t vcompress(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_i16m1))) +vint16m1_t vcompress(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_i16m2))) +vint16m2_t vcompress(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_i16m4))) +vint16m4_t vcompress(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_i16m8))) +vint16m8_t vcompress(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_i16mf2))) +vint16mf2_t vcompress(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_i16mf4))) +vint16mf4_t vcompress(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_i32m1))) +vint32m1_t vcompress(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_i32m2))) +vint32m2_t vcompress(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_i32m4))) +vint32m4_t vcompress(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_i32m8))) +vint32m8_t vcompress(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_i32mf2))) +vint32mf2_t vcompress(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_i64m1))) +vint64m1_t vcompress(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_i64m2))) +vint64m2_t vcompress(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_i64m4))) +vint64m4_t vcompress(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_i64m8))) +vint64m8_t vcompress(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_u8m1))) +vuint8m1_t vcompress(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_u8m2))) +vuint8m2_t vcompress(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_u8m4))) +vuint8m4_t vcompress(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_u8m8))) +vuint8m8_t vcompress(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_u8mf2))) +vuint8mf2_t vcompress(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_u8mf4))) +vuint8mf4_t vcompress(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_u8mf8))) +vuint8mf8_t vcompress(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_u16m1))) +vuint16m1_t vcompress(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_u16m2))) +vuint16m2_t vcompress(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_u16m4))) +vuint16m4_t vcompress(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_u16m8))) +vuint16m8_t vcompress(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_u16mf2))) +vuint16mf2_t vcompress(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_u16mf4))) +vuint16mf4_t vcompress(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_u32m1))) +vuint32m1_t vcompress(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_u32m2))) +vuint32m2_t vcompress(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_u32m4))) +vuint32m4_t vcompress(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_u32m8))) +vuint32m8_t vcompress(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_u32mf2))) +vuint32mf2_t vcompress(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_u64m1))) +vuint64m1_t vcompress(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_u64m2))) +vuint64m2_t vcompress(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_u64m4))) +vuint64m4_t vcompress(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_u64m8))) +vuint64m8_t vcompress(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_i8m2_i8m1))) +vint8m1_t vget_i8m1(vint8m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_i8m4_i8m1))) +vint8m1_t vget_i8m1(vint8m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_i8m8_i8m1))) +vint8m1_t vget_i8m1(vint8m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_i16m2_i16m1))) +vint16m1_t vget_i16m1(vint16m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_i16m4_i16m1))) +vint16m1_t vget_i16m1(vint16m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_i16m8_i16m1))) +vint16m1_t vget_i16m1(vint16m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_i32m2_i32m1))) +vint32m1_t vget_i32m1(vint32m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_i32m4_i32m1))) +vint32m1_t vget_i32m1(vint32m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_i32m8_i32m1))) +vint32m1_t vget_i32m1(vint32m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_i64m2_i64m1))) +vint64m1_t vget_i64m1(vint64m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_i64m4_i64m1))) +vint64m1_t vget_i64m1(vint64m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_i64m8_i64m1))) +vint64m1_t vget_i64m1(vint64m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_u8m2_u8m1))) +vuint8m1_t vget_u8m1(vuint8m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_u8m4_u8m1))) +vuint8m1_t vget_u8m1(vuint8m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_u8m8_u8m1))) +vuint8m1_t vget_u8m1(vuint8m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_u16m2_u16m1))) +vuint16m1_t vget_u16m1(vuint16m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_u16m4_u16m1))) +vuint16m1_t vget_u16m1(vuint16m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_u16m8_u16m1))) +vuint16m1_t vget_u16m1(vuint16m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_u32m2_u32m1))) +vuint32m1_t vget_u32m1(vuint32m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_u32m4_u32m1))) +vuint32m1_t vget_u32m1(vuint32m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_u32m8_u32m1))) +vuint32m1_t vget_u32m1(vuint32m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_u64m2_u64m1))) +vuint64m1_t vget_u64m1(vuint64m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_u64m4_u64m1))) +vuint64m1_t vget_u64m1(vuint64m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_u64m8_u64m1))) +vuint64m1_t vget_u64m1(vuint64m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_i8m4_i8m2))) +vint8m2_t vget_i8m2(vint8m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_i8m8_i8m2))) +vint8m2_t vget_i8m2(vint8m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_i16m4_i16m2))) +vint16m2_t vget_i16m2(vint16m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_i16m8_i16m2))) +vint16m2_t vget_i16m2(vint16m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_i32m4_i32m2))) +vint32m2_t vget_i32m2(vint32m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_i32m8_i32m2))) +vint32m2_t vget_i32m2(vint32m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_i64m4_i64m2))) +vint64m2_t vget_i64m2(vint64m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_i64m8_i64m2))) +vint64m2_t vget_i64m2(vint64m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_u8m4_u8m2))) +vuint8m2_t vget_u8m2(vuint8m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_u8m8_u8m2))) +vuint8m2_t vget_u8m2(vuint8m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_u16m4_u16m2))) +vuint16m2_t vget_u16m2(vuint16m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_u16m8_u16m2))) +vuint16m2_t vget_u16m2(vuint16m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_u32m4_u32m2))) +vuint32m2_t vget_u32m2(vuint32m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_u32m8_u32m2))) +vuint32m2_t vget_u32m2(vuint32m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_u64m4_u64m2))) +vuint64m2_t vget_u64m2(vuint64m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_u64m8_u64m2))) +vuint64m2_t vget_u64m2(vuint64m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_i8m8_i8m4))) +vint8m4_t vget_i8m4(vint8m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_i16m8_i16m4))) +vint16m4_t vget_i16m4(vint16m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_i32m8_i32m4))) +vint32m4_t vget_i32m4(vint32m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_i64m8_i64m4))) +vint64m4_t vget_i64m4(vint64m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_u8m8_u8m4))) +vuint8m4_t vget_u8m4(vuint8m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_u16m8_u16m4))) +vuint16m4_t vget_u16m4(vuint16m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_u32m8_u32m4))) +vuint32m4_t vget_u32m4(vuint32m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_u64m8_u64m4))) +vuint64m4_t vget_u64m4(vuint64m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_i8m1_i8m2))) +vint8m2_t vset(vint8m2_t op0, size_t op1, vint8m1_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_i16m1_i16m2))) +vint16m2_t vset(vint16m2_t op0, size_t op1, vint16m1_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_i32m1_i32m2))) +vint32m2_t vset(vint32m2_t op0, size_t op1, vint32m1_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_i64m1_i64m2))) +vint64m2_t vset(vint64m2_t op0, size_t op1, vint64m1_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_u8m1_u8m2))) +vuint8m2_t vset(vuint8m2_t op0, size_t op1, vuint8m1_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_u16m1_u16m2))) +vuint16m2_t vset(vuint16m2_t op0, size_t op1, vuint16m1_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_u32m1_u32m2))) +vuint32m2_t vset(vuint32m2_t op0, size_t op1, vuint32m1_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_u64m1_u64m2))) +vuint64m2_t vset(vuint64m2_t op0, size_t op1, vuint64m1_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_i8m1_i8m4))) +vint8m4_t vset(vint8m4_t op0, size_t op1, vint8m1_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_i8m2_i8m4))) +vint8m4_t vset(vint8m4_t op0, size_t op1, vint8m2_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_i16m1_i16m4))) +vint16m4_t vset(vint16m4_t op0, size_t op1, vint16m1_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_i16m2_i16m4))) +vint16m4_t vset(vint16m4_t op0, size_t op1, vint16m2_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_i32m1_i32m4))) +vint32m4_t vset(vint32m4_t op0, size_t op1, vint32m1_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_i32m2_i32m4))) +vint32m4_t vset(vint32m4_t op0, size_t op1, vint32m2_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_i64m1_i64m4))) +vint64m4_t vset(vint64m4_t op0, size_t op1, vint64m1_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_i64m2_i64m4))) +vint64m4_t vset(vint64m4_t op0, size_t op1, vint64m2_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_u8m1_u8m4))) +vuint8m4_t vset(vuint8m4_t op0, size_t op1, vuint8m1_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_u8m2_u8m4))) +vuint8m4_t vset(vuint8m4_t op0, size_t op1, vuint8m2_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_u16m1_u16m4))) +vuint16m4_t vset(vuint16m4_t op0, size_t op1, vuint16m1_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_u16m2_u16m4))) +vuint16m4_t vset(vuint16m4_t op0, size_t op1, vuint16m2_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_u32m1_u32m4))) +vuint32m4_t vset(vuint32m4_t op0, size_t op1, vuint32m1_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_u32m2_u32m4))) +vuint32m4_t vset(vuint32m4_t op0, size_t op1, vuint32m2_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_u64m1_u64m4))) +vuint64m4_t vset(vuint64m4_t op0, size_t op1, vuint64m1_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_u64m2_u64m4))) +vuint64m4_t vset(vuint64m4_t op0, size_t op1, vuint64m2_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_i8m1_i8m8))) +vint8m8_t vset(vint8m8_t op0, size_t op1, vint8m1_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_i8m2_i8m8))) +vint8m8_t vset(vint8m8_t op0, size_t op1, vint8m2_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_i8m4_i8m8))) +vint8m8_t vset(vint8m8_t op0, size_t op1, vint8m4_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_i16m1_i16m8))) +vint16m8_t vset(vint16m8_t op0, size_t op1, vint16m1_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_i16m2_i16m8))) +vint16m8_t vset(vint16m8_t op0, size_t op1, vint16m2_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_i16m4_i16m8))) +vint16m8_t vset(vint16m8_t op0, size_t op1, vint16m4_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_i32m1_i32m8))) +vint32m8_t vset(vint32m8_t op0, size_t op1, vint32m1_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_i32m2_i32m8))) +vint32m8_t vset(vint32m8_t op0, size_t op1, vint32m2_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_i32m4_i32m8))) +vint32m8_t vset(vint32m8_t op0, size_t op1, vint32m4_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_i64m1_i64m8))) +vint64m8_t vset(vint64m8_t op0, size_t op1, vint64m1_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_i64m2_i64m8))) +vint64m8_t vset(vint64m8_t op0, size_t op1, vint64m2_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_i64m4_i64m8))) +vint64m8_t vset(vint64m8_t op0, size_t op1, vint64m4_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_u8m1_u8m8))) +vuint8m8_t vset(vuint8m8_t op0, size_t op1, vuint8m1_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_u8m2_u8m8))) +vuint8m8_t vset(vuint8m8_t op0, size_t op1, vuint8m2_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_u8m4_u8m8))) +vuint8m8_t vset(vuint8m8_t op0, size_t op1, vuint8m4_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_u16m1_u16m8))) +vuint16m8_t vset(vuint16m8_t op0, size_t op1, vuint16m1_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_u16m2_u16m8))) +vuint16m8_t vset(vuint16m8_t op0, size_t op1, vuint16m2_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_u16m4_u16m8))) +vuint16m8_t vset(vuint16m8_t op0, size_t op1, vuint16m4_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_u32m1_u32m8))) +vuint32m8_t vset(vuint32m8_t op0, size_t op1, vuint32m1_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_u32m2_u32m8))) +vuint32m8_t vset(vuint32m8_t op0, size_t op1, vuint32m2_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_u32m4_u32m8))) +vuint32m8_t vset(vuint32m8_t op0, size_t op1, vuint32m4_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_u64m1_u64m8))) +vuint64m8_t vset(vuint64m8_t op0, size_t op1, vuint64m1_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_u64m2_u64m8))) +vuint64m8_t vset(vuint64m8_t op0, size_t op1, vuint64m2_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_u64m4_u64m8))) +vuint64m8_t vset(vuint64m8_t op0, size_t op1, vuint64m4_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_i8m1))) +void vsoxei8(int8_t * op0, vuint8m1_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_i8m1_m))) +void vsoxei8(vbool8_t op0, int8_t * op1, vuint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_i8m2))) +void vsoxei8(int8_t * op0, vuint8m2_t op1, vint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_i8m2_m))) +void vsoxei8(vbool4_t op0, int8_t * op1, vuint8m2_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_i8m4))) +void vsoxei8(int8_t * op0, vuint8m4_t op1, vint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_i8m4_m))) +void vsoxei8(vbool2_t op0, int8_t * op1, vuint8m4_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_i8m8))) +void vsoxei8(int8_t * op0, vuint8m8_t op1, vint8m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_i8m8_m))) +void vsoxei8(vbool1_t op0, int8_t * op1, vuint8m8_t op2, vint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_i8mf2))) +void vsoxei8(int8_t * op0, vuint8mf2_t op1, vint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_i8mf2_m))) +void vsoxei8(vbool16_t op0, int8_t * op1, vuint8mf2_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_i8mf4))) +void vsoxei8(int8_t * op0, vuint8mf4_t op1, vint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_i8mf4_m))) +void vsoxei8(vbool32_t op0, int8_t * op1, vuint8mf4_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_i8mf8))) +void vsoxei8(int8_t * op0, vuint8mf8_t op1, vint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_i8mf8_m))) +void vsoxei8(vbool64_t op0, int8_t * op1, vuint8mf8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_u8m1))) +void vsoxei8(uint8_t * op0, vuint8m1_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_u8m1_m))) +void vsoxei8(vbool8_t op0, uint8_t * op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_u8m2))) +void vsoxei8(uint8_t * op0, vuint8m2_t op1, vuint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_u8m2_m))) +void vsoxei8(vbool4_t op0, uint8_t * op1, vuint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_u8m4))) +void vsoxei8(uint8_t * op0, vuint8m4_t op1, vuint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_u8m4_m))) +void vsoxei8(vbool2_t op0, uint8_t * op1, vuint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_u8m8))) +void vsoxei8(uint8_t * op0, vuint8m8_t op1, vuint8m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_u8m8_m))) +void vsoxei8(vbool1_t op0, uint8_t * op1, vuint8m8_t op2, vuint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_u8mf2))) +void vsoxei8(uint8_t * op0, vuint8mf2_t op1, vuint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_u8mf2_m))) +void vsoxei8(vbool16_t op0, uint8_t * op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_u8mf4))) +void vsoxei8(uint8_t * op0, vuint8mf4_t op1, vuint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_u8mf4_m))) +void vsoxei8(vbool32_t op0, uint8_t * op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_u8mf8))) +void vsoxei8(uint8_t * op0, vuint8mf8_t op1, vuint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_u8mf8_m))) +void vsoxei8(vbool64_t op0, uint8_t * op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_i8m1))) +void vsoxei16(int8_t * op0, vuint16m2_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_i8m1_m))) +void vsoxei16(vbool8_t op0, int8_t * op1, vuint16m2_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_i8m2))) +void vsoxei16(int8_t * op0, vuint16m4_t op1, vint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_i8m2_m))) +void vsoxei16(vbool4_t op0, int8_t * op1, vuint16m4_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_i8m4))) +void vsoxei16(int8_t * op0, vuint16m8_t op1, vint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_i8m4_m))) +void vsoxei16(vbool2_t op0, int8_t * op1, vuint16m8_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_i8mf2))) +void vsoxei16(int8_t * op0, vuint16m1_t op1, vint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_i8mf2_m))) +void vsoxei16(vbool16_t op0, int8_t * op1, vuint16m1_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_i8mf4))) +void vsoxei16(int8_t * op0, vuint16mf2_t op1, vint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_i8mf4_m))) +void vsoxei16(vbool32_t op0, int8_t * op1, vuint16mf2_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_i8mf8))) +void vsoxei16(int8_t * op0, vuint16mf4_t op1, vint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_i8mf8_m))) +void vsoxei16(vbool64_t op0, int8_t * op1, vuint16mf4_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_u8m1))) +void vsoxei16(uint8_t * op0, vuint16m2_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_u8m1_m))) +void vsoxei16(vbool8_t op0, uint8_t * op1, vuint16m2_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_u8m2))) +void vsoxei16(uint8_t * op0, vuint16m4_t op1, vuint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_u8m2_m))) +void vsoxei16(vbool4_t op0, uint8_t * op1, vuint16m4_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_u8m4))) +void vsoxei16(uint8_t * op0, vuint16m8_t op1, vuint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_u8m4_m))) +void vsoxei16(vbool2_t op0, uint8_t * op1, vuint16m8_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_u8mf2))) +void vsoxei16(uint8_t * op0, vuint16m1_t op1, vuint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_u8mf2_m))) +void vsoxei16(vbool16_t op0, uint8_t * op1, vuint16m1_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_u8mf4))) +void vsoxei16(uint8_t * op0, vuint16mf2_t op1, vuint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_u8mf4_m))) +void vsoxei16(vbool32_t op0, uint8_t * op1, vuint16mf2_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_u8mf8))) +void vsoxei16(uint8_t * op0, vuint16mf4_t op1, vuint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_u8mf8_m))) +void vsoxei16(vbool64_t op0, uint8_t * op1, vuint16mf4_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_i8m1))) +void vsoxei32(int8_t * op0, vuint32m4_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_i8m1_m))) +void vsoxei32(vbool8_t op0, int8_t * op1, vuint32m4_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_i8m2))) +void vsoxei32(int8_t * op0, vuint32m8_t op1, vint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_i8m2_m))) +void vsoxei32(vbool4_t op0, int8_t * op1, vuint32m8_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_i8mf2))) +void vsoxei32(int8_t * op0, vuint32m2_t op1, vint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_i8mf2_m))) +void vsoxei32(vbool16_t op0, int8_t * op1, vuint32m2_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_i8mf4))) +void vsoxei32(int8_t * op0, vuint32m1_t op1, vint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_i8mf4_m))) +void vsoxei32(vbool32_t op0, int8_t * op1, vuint32m1_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_i8mf8))) +void vsoxei32(int8_t * op0, vuint32mf2_t op1, vint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_i8mf8_m))) +void vsoxei32(vbool64_t op0, int8_t * op1, vuint32mf2_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_u8m1))) +void vsoxei32(uint8_t * op0, vuint32m4_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_u8m1_m))) +void vsoxei32(vbool8_t op0, uint8_t * op1, vuint32m4_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_u8m2))) +void vsoxei32(uint8_t * op0, vuint32m8_t op1, vuint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_u8m2_m))) +void vsoxei32(vbool4_t op0, uint8_t * op1, vuint32m8_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_u8mf2))) +void vsoxei32(uint8_t * op0, vuint32m2_t op1, vuint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_u8mf2_m))) +void vsoxei32(vbool16_t op0, uint8_t * op1, vuint32m2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_u8mf4))) +void vsoxei32(uint8_t * op0, vuint32m1_t op1, vuint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_u8mf4_m))) +void vsoxei32(vbool32_t op0, uint8_t * op1, vuint32m1_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_u8mf8))) +void vsoxei32(uint8_t * op0, vuint32mf2_t op1, vuint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_u8mf8_m))) +void vsoxei32(vbool64_t op0, uint8_t * op1, vuint32mf2_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_i8m1))) +void vsoxei64(int8_t * op0, vuint64m8_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_i8m1_m))) +void vsoxei64(vbool8_t op0, int8_t * op1, vuint64m8_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_i8mf2))) +void vsoxei64(int8_t * op0, vuint64m4_t op1, vint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_i8mf2_m))) +void vsoxei64(vbool16_t op0, int8_t * op1, vuint64m4_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_i8mf4))) +void vsoxei64(int8_t * op0, vuint64m2_t op1, vint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_i8mf4_m))) +void vsoxei64(vbool32_t op0, int8_t * op1, vuint64m2_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_i8mf8))) +void vsoxei64(int8_t * op0, vuint64m1_t op1, vint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_i8mf8_m))) +void vsoxei64(vbool64_t op0, int8_t * op1, vuint64m1_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_u8m1))) +void vsoxei64(uint8_t * op0, vuint64m8_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_u8m1_m))) +void vsoxei64(vbool8_t op0, uint8_t * op1, vuint64m8_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_u8mf2))) +void vsoxei64(uint8_t * op0, vuint64m4_t op1, vuint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_u8mf2_m))) +void vsoxei64(vbool16_t op0, uint8_t * op1, vuint64m4_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_u8mf4))) +void vsoxei64(uint8_t * op0, vuint64m2_t op1, vuint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_u8mf4_m))) +void vsoxei64(vbool32_t op0, uint8_t * op1, vuint64m2_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_u8mf8))) +void vsoxei64(uint8_t * op0, vuint64m1_t op1, vuint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_u8mf8_m))) +void vsoxei64(vbool64_t op0, uint8_t * op1, vuint64m1_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_i16m1))) +void vsoxei8(int16_t * op0, vuint8mf2_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_i16m1_m))) +void vsoxei8(vbool16_t op0, int16_t * op1, vuint8mf2_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_i16m2))) +void vsoxei8(int16_t * op0, vuint8m1_t op1, vint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_i16m2_m))) +void vsoxei8(vbool8_t op0, int16_t * op1, vuint8m1_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_i16m4))) +void vsoxei8(int16_t * op0, vuint8m2_t op1, vint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_i16m4_m))) +void vsoxei8(vbool4_t op0, int16_t * op1, vuint8m2_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_i16m8))) +void vsoxei8(int16_t * op0, vuint8m4_t op1, vint16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_i16m8_m))) +void vsoxei8(vbool2_t op0, int16_t * op1, vuint8m4_t op2, vint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_i16mf2))) +void vsoxei8(int16_t * op0, vuint8mf4_t op1, vint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_i16mf2_m))) +void vsoxei8(vbool32_t op0, int16_t * op1, vuint8mf4_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_i16mf4))) +void vsoxei8(int16_t * op0, vuint8mf8_t op1, vint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_i16mf4_m))) +void vsoxei8(vbool64_t op0, int16_t * op1, vuint8mf8_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_u16m1))) +void vsoxei8(uint16_t * op0, vuint8mf2_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_u16m1_m))) +void vsoxei8(vbool16_t op0, uint16_t * op1, vuint8mf2_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_u16m2))) +void vsoxei8(uint16_t * op0, vuint8m1_t op1, vuint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_u16m2_m))) +void vsoxei8(vbool8_t op0, uint16_t * op1, vuint8m1_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_u16m4))) +void vsoxei8(uint16_t * op0, vuint8m2_t op1, vuint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_u16m4_m))) +void vsoxei8(vbool4_t op0, uint16_t * op1, vuint8m2_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_u16m8))) +void vsoxei8(uint16_t * op0, vuint8m4_t op1, vuint16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_u16m8_m))) +void vsoxei8(vbool2_t op0, uint16_t * op1, vuint8m4_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_u16mf2))) +void vsoxei8(uint16_t * op0, vuint8mf4_t op1, vuint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_u16mf2_m))) +void vsoxei8(vbool32_t op0, uint16_t * op1, vuint8mf4_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_u16mf4))) +void vsoxei8(uint16_t * op0, vuint8mf8_t op1, vuint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_u16mf4_m))) +void vsoxei8(vbool64_t op0, uint16_t * op1, vuint8mf8_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_i16m1))) +void vsoxei16(int16_t * op0, vuint16m1_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_i16m1_m))) +void vsoxei16(vbool16_t op0, int16_t * op1, vuint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_i16m2))) +void vsoxei16(int16_t * op0, vuint16m2_t op1, vint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_i16m2_m))) +void vsoxei16(vbool8_t op0, int16_t * op1, vuint16m2_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_i16m4))) +void vsoxei16(int16_t * op0, vuint16m4_t op1, vint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_i16m4_m))) +void vsoxei16(vbool4_t op0, int16_t * op1, vuint16m4_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_i16m8))) +void vsoxei16(int16_t * op0, vuint16m8_t op1, vint16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_i16m8_m))) +void vsoxei16(vbool2_t op0, int16_t * op1, vuint16m8_t op2, vint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_i16mf2))) +void vsoxei16(int16_t * op0, vuint16mf2_t op1, vint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_i16mf2_m))) +void vsoxei16(vbool32_t op0, int16_t * op1, vuint16mf2_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_i16mf4))) +void vsoxei16(int16_t * op0, vuint16mf4_t op1, vint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_i16mf4_m))) +void vsoxei16(vbool64_t op0, int16_t * op1, vuint16mf4_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_u16m1))) +void vsoxei16(uint16_t * op0, vuint16m1_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_u16m1_m))) +void vsoxei16(vbool16_t op0, uint16_t * op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_u16m2))) +void vsoxei16(uint16_t * op0, vuint16m2_t op1, vuint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_u16m2_m))) +void vsoxei16(vbool8_t op0, uint16_t * op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_u16m4))) +void vsoxei16(uint16_t * op0, vuint16m4_t op1, vuint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_u16m4_m))) +void vsoxei16(vbool4_t op0, uint16_t * op1, vuint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_u16m8))) +void vsoxei16(uint16_t * op0, vuint16m8_t op1, vuint16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_u16m8_m))) +void vsoxei16(vbool2_t op0, uint16_t * op1, vuint16m8_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_u16mf2))) +void vsoxei16(uint16_t * op0, vuint16mf2_t op1, vuint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_u16mf2_m))) +void vsoxei16(vbool32_t op0, uint16_t * op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_u16mf4))) +void vsoxei16(uint16_t * op0, vuint16mf4_t op1, vuint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_u16mf4_m))) +void vsoxei16(vbool64_t op0, uint16_t * op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_i16m1))) +void vsoxei32(int16_t * op0, vuint32m2_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_i16m1_m))) +void vsoxei32(vbool16_t op0, int16_t * op1, vuint32m2_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_i16m2))) +void vsoxei32(int16_t * op0, vuint32m4_t op1, vint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_i16m2_m))) +void vsoxei32(vbool8_t op0, int16_t * op1, vuint32m4_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_i16m4))) +void vsoxei32(int16_t * op0, vuint32m8_t op1, vint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_i16m4_m))) +void vsoxei32(vbool4_t op0, int16_t * op1, vuint32m8_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_i16mf2))) +void vsoxei32(int16_t * op0, vuint32m1_t op1, vint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_i16mf2_m))) +void vsoxei32(vbool32_t op0, int16_t * op1, vuint32m1_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_i16mf4))) +void vsoxei32(int16_t * op0, vuint32mf2_t op1, vint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_i16mf4_m))) +void vsoxei32(vbool64_t op0, int16_t * op1, vuint32mf2_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_u16m1))) +void vsoxei32(uint16_t * op0, vuint32m2_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_u16m1_m))) +void vsoxei32(vbool16_t op0, uint16_t * op1, vuint32m2_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_u16m2))) +void vsoxei32(uint16_t * op0, vuint32m4_t op1, vuint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_u16m2_m))) +void vsoxei32(vbool8_t op0, uint16_t * op1, vuint32m4_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_u16m4))) +void vsoxei32(uint16_t * op0, vuint32m8_t op1, vuint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_u16m4_m))) +void vsoxei32(vbool4_t op0, uint16_t * op1, vuint32m8_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_u16mf2))) +void vsoxei32(uint16_t * op0, vuint32m1_t op1, vuint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_u16mf2_m))) +void vsoxei32(vbool32_t op0, uint16_t * op1, vuint32m1_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_u16mf4))) +void vsoxei32(uint16_t * op0, vuint32mf2_t op1, vuint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_u16mf4_m))) +void vsoxei32(vbool64_t op0, uint16_t * op1, vuint32mf2_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_i16m1))) +void vsoxei64(int16_t * op0, vuint64m4_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_i16m1_m))) +void vsoxei64(vbool16_t op0, int16_t * op1, vuint64m4_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_i16m2))) +void vsoxei64(int16_t * op0, vuint64m8_t op1, vint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_i16m2_m))) +void vsoxei64(vbool8_t op0, int16_t * op1, vuint64m8_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_i16mf2))) +void vsoxei64(int16_t * op0, vuint64m2_t op1, vint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_i16mf2_m))) +void vsoxei64(vbool32_t op0, int16_t * op1, vuint64m2_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_i16mf4))) +void vsoxei64(int16_t * op0, vuint64m1_t op1, vint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_i16mf4_m))) +void vsoxei64(vbool64_t op0, int16_t * op1, vuint64m1_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_u16m1))) +void vsoxei64(uint16_t * op0, vuint64m4_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_u16m1_m))) +void vsoxei64(vbool16_t op0, uint16_t * op1, vuint64m4_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_u16m2))) +void vsoxei64(uint16_t * op0, vuint64m8_t op1, vuint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_u16m2_m))) +void vsoxei64(vbool8_t op0, uint16_t * op1, vuint64m8_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_u16mf2))) +void vsoxei64(uint16_t * op0, vuint64m2_t op1, vuint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_u16mf2_m))) +void vsoxei64(vbool32_t op0, uint16_t * op1, vuint64m2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_u16mf4))) +void vsoxei64(uint16_t * op0, vuint64m1_t op1, vuint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_u16mf4_m))) +void vsoxei64(vbool64_t op0, uint16_t * op1, vuint64m1_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_i32m1))) +void vsoxei8(int32_t * op0, vuint8mf4_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_i32m1_m))) +void vsoxei8(vbool32_t op0, int32_t * op1, vuint8mf4_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_i32m2))) +void vsoxei8(int32_t * op0, vuint8mf2_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_i32m2_m))) +void vsoxei8(vbool16_t op0, int32_t * op1, vuint8mf2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_i32m4))) +void vsoxei8(int32_t * op0, vuint8m1_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_i32m4_m))) +void vsoxei8(vbool8_t op0, int32_t * op1, vuint8m1_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_i32m8))) +void vsoxei8(int32_t * op0, vuint8m2_t op1, vint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_i32m8_m))) +void vsoxei8(vbool4_t op0, int32_t * op1, vuint8m2_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_i32mf2))) +void vsoxei8(int32_t * op0, vuint8mf8_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_i32mf2_m))) +void vsoxei8(vbool64_t op0, int32_t * op1, vuint8mf8_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_u32m1))) +void vsoxei8(uint32_t * op0, vuint8mf4_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_u32m1_m))) +void vsoxei8(vbool32_t op0, uint32_t * op1, vuint8mf4_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_u32m2))) +void vsoxei8(uint32_t * op0, vuint8mf2_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_u32m2_m))) +void vsoxei8(vbool16_t op0, uint32_t * op1, vuint8mf2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_u32m4))) +void vsoxei8(uint32_t * op0, vuint8m1_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_u32m4_m))) +void vsoxei8(vbool8_t op0, uint32_t * op1, vuint8m1_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_u32m8))) +void vsoxei8(uint32_t * op0, vuint8m2_t op1, vuint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_u32m8_m))) +void vsoxei8(vbool4_t op0, uint32_t * op1, vuint8m2_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_u32mf2))) +void vsoxei8(uint32_t * op0, vuint8mf8_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_u32mf2_m))) +void vsoxei8(vbool64_t op0, uint32_t * op1, vuint8mf8_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_i32m1))) +void vsoxei16(int32_t * op0, vuint16mf2_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_i32m1_m))) +void vsoxei16(vbool32_t op0, int32_t * op1, vuint16mf2_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_i32m2))) +void vsoxei16(int32_t * op0, vuint16m1_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_i32m2_m))) +void vsoxei16(vbool16_t op0, int32_t * op1, vuint16m1_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_i32m4))) +void vsoxei16(int32_t * op0, vuint16m2_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_i32m4_m))) +void vsoxei16(vbool8_t op0, int32_t * op1, vuint16m2_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_i32m8))) +void vsoxei16(int32_t * op0, vuint16m4_t op1, vint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_i32m8_m))) +void vsoxei16(vbool4_t op0, int32_t * op1, vuint16m4_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_i32mf2))) +void vsoxei16(int32_t * op0, vuint16mf4_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_i32mf2_m))) +void vsoxei16(vbool64_t op0, int32_t * op1, vuint16mf4_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_u32m1))) +void vsoxei16(uint32_t * op0, vuint16mf2_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_u32m1_m))) +void vsoxei16(vbool32_t op0, uint32_t * op1, vuint16mf2_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_u32m2))) +void vsoxei16(uint32_t * op0, vuint16m1_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_u32m2_m))) +void vsoxei16(vbool16_t op0, uint32_t * op1, vuint16m1_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_u32m4))) +void vsoxei16(uint32_t * op0, vuint16m2_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_u32m4_m))) +void vsoxei16(vbool8_t op0, uint32_t * op1, vuint16m2_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_u32m8))) +void vsoxei16(uint32_t * op0, vuint16m4_t op1, vuint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_u32m8_m))) +void vsoxei16(vbool4_t op0, uint32_t * op1, vuint16m4_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_u32mf2))) +void vsoxei16(uint32_t * op0, vuint16mf4_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_u32mf2_m))) +void vsoxei16(vbool64_t op0, uint32_t * op1, vuint16mf4_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_i32m1))) +void vsoxei32(int32_t * op0, vuint32m1_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_i32m1_m))) +void vsoxei32(vbool32_t op0, int32_t * op1, vuint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_i32m2))) +void vsoxei32(int32_t * op0, vuint32m2_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_i32m2_m))) +void vsoxei32(vbool16_t op0, int32_t * op1, vuint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_i32m4))) +void vsoxei32(int32_t * op0, vuint32m4_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_i32m4_m))) +void vsoxei32(vbool8_t op0, int32_t * op1, vuint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_i32m8))) +void vsoxei32(int32_t * op0, vuint32m8_t op1, vint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_i32m8_m))) +void vsoxei32(vbool4_t op0, int32_t * op1, vuint32m8_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_i32mf2))) +void vsoxei32(int32_t * op0, vuint32mf2_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_i32mf2_m))) +void vsoxei32(vbool64_t op0, int32_t * op1, vuint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_u32m1))) +void vsoxei32(uint32_t * op0, vuint32m1_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_u32m1_m))) +void vsoxei32(vbool32_t op0, uint32_t * op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_u32m2))) +void vsoxei32(uint32_t * op0, vuint32m2_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_u32m2_m))) +void vsoxei32(vbool16_t op0, uint32_t * op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_u32m4))) +void vsoxei32(uint32_t * op0, vuint32m4_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_u32m4_m))) +void vsoxei32(vbool8_t op0, uint32_t * op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_u32m8))) +void vsoxei32(uint32_t * op0, vuint32m8_t op1, vuint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_u32m8_m))) +void vsoxei32(vbool4_t op0, uint32_t * op1, vuint32m8_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_u32mf2))) +void vsoxei32(uint32_t * op0, vuint32mf2_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_u32mf2_m))) +void vsoxei32(vbool64_t op0, uint32_t * op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_i32m1))) +void vsoxei64(int32_t * op0, vuint64m2_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_i32m1_m))) +void vsoxei64(vbool32_t op0, int32_t * op1, vuint64m2_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_i32m2))) +void vsoxei64(int32_t * op0, vuint64m4_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_i32m2_m))) +void vsoxei64(vbool16_t op0, int32_t * op1, vuint64m4_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_i32m4))) +void vsoxei64(int32_t * op0, vuint64m8_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_i32m4_m))) +void vsoxei64(vbool8_t op0, int32_t * op1, vuint64m8_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_i32mf2))) +void vsoxei64(int32_t * op0, vuint64m1_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_i32mf2_m))) +void vsoxei64(vbool64_t op0, int32_t * op1, vuint64m1_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_u32m1))) +void vsoxei64(uint32_t * op0, vuint64m2_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_u32m1_m))) +void vsoxei64(vbool32_t op0, uint32_t * op1, vuint64m2_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_u32m2))) +void vsoxei64(uint32_t * op0, vuint64m4_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_u32m2_m))) +void vsoxei64(vbool16_t op0, uint32_t * op1, vuint64m4_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_u32m4))) +void vsoxei64(uint32_t * op0, vuint64m8_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_u32m4_m))) +void vsoxei64(vbool8_t op0, uint32_t * op1, vuint64m8_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_u32mf2))) +void vsoxei64(uint32_t * op0, vuint64m1_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_u32mf2_m))) +void vsoxei64(vbool64_t op0, uint32_t * op1, vuint64m1_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_i64m1))) +void vsoxei8(int64_t * op0, vuint8mf8_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_i64m1_m))) +void vsoxei8(vbool64_t op0, int64_t * op1, vuint8mf8_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_i64m2))) +void vsoxei8(int64_t * op0, vuint8mf4_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_i64m2_m))) +void vsoxei8(vbool32_t op0, int64_t * op1, vuint8mf4_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_i64m4))) +void vsoxei8(int64_t * op0, vuint8mf2_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_i64m4_m))) +void vsoxei8(vbool16_t op0, int64_t * op1, vuint8mf2_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_i64m8))) +void vsoxei8(int64_t * op0, vuint8m1_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_i64m8_m))) +void vsoxei8(vbool8_t op0, int64_t * op1, vuint8m1_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_u64m1))) +void vsoxei8(uint64_t * op0, vuint8mf8_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_u64m1_m))) +void vsoxei8(vbool64_t op0, uint64_t * op1, vuint8mf8_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_u64m2))) +void vsoxei8(uint64_t * op0, vuint8mf4_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_u64m2_m))) +void vsoxei8(vbool32_t op0, uint64_t * op1, vuint8mf4_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_u64m4))) +void vsoxei8(uint64_t * op0, vuint8mf2_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_u64m4_m))) +void vsoxei8(vbool16_t op0, uint64_t * op1, vuint8mf2_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_u64m8))) +void vsoxei8(uint64_t * op0, vuint8m1_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_u64m8_m))) +void vsoxei8(vbool8_t op0, uint64_t * op1, vuint8m1_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_i64m1))) +void vsoxei16(int64_t * op0, vuint16mf4_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_i64m1_m))) +void vsoxei16(vbool64_t op0, int64_t * op1, vuint16mf4_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_i64m2))) +void vsoxei16(int64_t * op0, vuint16mf2_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_i64m2_m))) +void vsoxei16(vbool32_t op0, int64_t * op1, vuint16mf2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_i64m4))) +void vsoxei16(int64_t * op0, vuint16m1_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_i64m4_m))) +void vsoxei16(vbool16_t op0, int64_t * op1, vuint16m1_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_i64m8))) +void vsoxei16(int64_t * op0, vuint16m2_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_i64m8_m))) +void vsoxei16(vbool8_t op0, int64_t * op1, vuint16m2_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_u64m1))) +void vsoxei16(uint64_t * op0, vuint16mf4_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_u64m1_m))) +void vsoxei16(vbool64_t op0, uint64_t * op1, vuint16mf4_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_u64m2))) +void vsoxei16(uint64_t * op0, vuint16mf2_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_u64m2_m))) +void vsoxei16(vbool32_t op0, uint64_t * op1, vuint16mf2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_u64m4))) +void vsoxei16(uint64_t * op0, vuint16m1_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_u64m4_m))) +void vsoxei16(vbool16_t op0, uint64_t * op1, vuint16m1_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_u64m8))) +void vsoxei16(uint64_t * op0, vuint16m2_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_u64m8_m))) +void vsoxei16(vbool8_t op0, uint64_t * op1, vuint16m2_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_i64m1))) +void vsoxei32(int64_t * op0, vuint32mf2_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_i64m1_m))) +void vsoxei32(vbool64_t op0, int64_t * op1, vuint32mf2_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_i64m2))) +void vsoxei32(int64_t * op0, vuint32m1_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_i64m2_m))) +void vsoxei32(vbool32_t op0, int64_t * op1, vuint32m1_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_i64m4))) +void vsoxei32(int64_t * op0, vuint32m2_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_i64m4_m))) +void vsoxei32(vbool16_t op0, int64_t * op1, vuint32m2_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_i64m8))) +void vsoxei32(int64_t * op0, vuint32m4_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_i64m8_m))) +void vsoxei32(vbool8_t op0, int64_t * op1, vuint32m4_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_u64m1))) +void vsoxei32(uint64_t * op0, vuint32mf2_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_u64m1_m))) +void vsoxei32(vbool64_t op0, uint64_t * op1, vuint32mf2_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_u64m2))) +void vsoxei32(uint64_t * op0, vuint32m1_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_u64m2_m))) +void vsoxei32(vbool32_t op0, uint64_t * op1, vuint32m1_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_u64m4))) +void vsoxei32(uint64_t * op0, vuint32m2_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_u64m4_m))) +void vsoxei32(vbool16_t op0, uint64_t * op1, vuint32m2_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_u64m8))) +void vsoxei32(uint64_t * op0, vuint32m4_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_u64m8_m))) +void vsoxei32(vbool8_t op0, uint64_t * op1, vuint32m4_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_i64m1))) +void vsoxei64(int64_t * op0, vuint64m1_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_i64m1_m))) +void vsoxei64(vbool64_t op0, int64_t * op1, vuint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_i64m2))) +void vsoxei64(int64_t * op0, vuint64m2_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_i64m2_m))) +void vsoxei64(vbool32_t op0, int64_t * op1, vuint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_i64m4))) +void vsoxei64(int64_t * op0, vuint64m4_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_i64m4_m))) +void vsoxei64(vbool16_t op0, int64_t * op1, vuint64m4_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_i64m8))) +void vsoxei64(int64_t * op0, vuint64m8_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_i64m8_m))) +void vsoxei64(vbool8_t op0, int64_t * op1, vuint64m8_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_u64m1))) +void vsoxei64(uint64_t * op0, vuint64m1_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_u64m1_m))) +void vsoxei64(vbool64_t op0, uint64_t * op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_u64m2))) +void vsoxei64(uint64_t * op0, vuint64m2_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_u64m2_m))) +void vsoxei64(vbool32_t op0, uint64_t * op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_u64m4))) +void vsoxei64(uint64_t * op0, vuint64m4_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_u64m4_m))) +void vsoxei64(vbool16_t op0, uint64_t * op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_u64m8))) +void vsoxei64(uint64_t * op0, vuint64m8_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_u64m8_m))) +void vsoxei64(vbool8_t op0, uint64_t * op1, vuint64m8_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle16ff_v_i16m1_m))) +vint16m1_t vle16ff(vbool16_t op0, vint16m1_t op1, const int16_t * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle16ff_v_i16m2_m))) +vint16m2_t vle16ff(vbool8_t op0, vint16m2_t op1, const int16_t * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle16ff_v_i16m4_m))) +vint16m4_t vle16ff(vbool4_t op0, vint16m4_t op1, const int16_t * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle16ff_v_i16m8_m))) +vint16m8_t vle16ff(vbool2_t op0, vint16m8_t op1, const int16_t * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle16ff_v_i16mf2_m))) +vint16mf2_t vle16ff(vbool32_t op0, vint16mf2_t op1, const int16_t * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle16ff_v_i16mf4_m))) +vint16mf4_t vle16ff(vbool64_t op0, vint16mf4_t op1, const int16_t * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle16ff_v_u16m1_m))) +vuint16m1_t vle16ff(vbool16_t op0, vuint16m1_t op1, const uint16_t * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle16ff_v_u16m2_m))) +vuint16m2_t vle16ff(vbool8_t op0, vuint16m2_t op1, const uint16_t * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle16ff_v_u16m4_m))) +vuint16m4_t vle16ff(vbool4_t op0, vuint16m4_t op1, const uint16_t * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle16ff_v_u16m8_m))) +vuint16m8_t vle16ff(vbool2_t op0, vuint16m8_t op1, const uint16_t * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle16ff_v_u16mf2_m))) +vuint16mf2_t vle16ff(vbool32_t op0, vuint16mf2_t op1, const uint16_t * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle16ff_v_u16mf4_m))) +vuint16mf4_t vle16ff(vbool64_t op0, vuint16mf4_t op1, const uint16_t * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle32ff_v_i32m1_m))) +vint32m1_t vle32ff(vbool32_t op0, vint32m1_t op1, const int32_t * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle32ff_v_i32m2_m))) +vint32m2_t vle32ff(vbool16_t op0, vint32m2_t op1, const int32_t * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle32ff_v_i32m4_m))) +vint32m4_t vle32ff(vbool8_t op0, vint32m4_t op1, const int32_t * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle32ff_v_i32m8_m))) +vint32m8_t vle32ff(vbool4_t op0, vint32m8_t op1, const int32_t * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle32ff_v_i32mf2_m))) +vint32mf2_t vle32ff(vbool64_t op0, vint32mf2_t op1, const int32_t * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle32ff_v_u32m1_m))) +vuint32m1_t vle32ff(vbool32_t op0, vuint32m1_t op1, const uint32_t * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle32ff_v_u32m2_m))) +vuint32m2_t vle32ff(vbool16_t op0, vuint32m2_t op1, const uint32_t * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle32ff_v_u32m4_m))) +vuint32m4_t vle32ff(vbool8_t op0, vuint32m4_t op1, const uint32_t * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle32ff_v_u32m8_m))) +vuint32m8_t vle32ff(vbool4_t op0, vuint32m8_t op1, const uint32_t * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle32ff_v_u32mf2_m))) +vuint32mf2_t vle32ff(vbool64_t op0, vuint32mf2_t op1, const uint32_t * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle64ff_v_i64m1_m))) +vint64m1_t vle64ff(vbool64_t op0, vint64m1_t op1, const int64_t * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle64ff_v_i64m2_m))) +vint64m2_t vle64ff(vbool32_t op0, vint64m2_t op1, const int64_t * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle64ff_v_i64m4_m))) +vint64m4_t vle64ff(vbool16_t op0, vint64m4_t op1, const int64_t * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle64ff_v_i64m8_m))) +vint64m8_t vle64ff(vbool8_t op0, vint64m8_t op1, const int64_t * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle64ff_v_u64m1_m))) +vuint64m1_t vle64ff(vbool64_t op0, vuint64m1_t op1, const uint64_t * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle64ff_v_u64m2_m))) +vuint64m2_t vle64ff(vbool32_t op0, vuint64m2_t op1, const uint64_t * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle64ff_v_u64m4_m))) +vuint64m4_t vle64ff(vbool16_t op0, vuint64m4_t op1, const uint64_t * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle64ff_v_u64m8_m))) +vuint64m8_t vle64ff(vbool8_t op0, vuint64m8_t op1, const uint64_t * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle8_v_i8m1_m))) +vint8m1_t vle8(vbool8_t op0, vint8m1_t op1, const int8_t * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle8_v_i8m2_m))) +vint8m2_t vle8(vbool4_t op0, vint8m2_t op1, const int8_t * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle8_v_i8m4_m))) +vint8m4_t vle8(vbool2_t op0, vint8m4_t op1, const int8_t * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle8_v_i8m8_m))) +vint8m8_t vle8(vbool1_t op0, vint8m8_t op1, const int8_t * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle8_v_i8mf2_m))) +vint8mf2_t vle8(vbool16_t op0, vint8mf2_t op1, const int8_t * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle8_v_i8mf4_m))) +vint8mf4_t vle8(vbool32_t op0, vint8mf4_t op1, const int8_t * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle8_v_i8mf8_m))) +vint8mf8_t vle8(vbool64_t op0, vint8mf8_t op1, const int8_t * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle8_v_u8m1_m))) +vuint8m1_t vle8(vbool8_t op0, vuint8m1_t op1, const uint8_t * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle8_v_u8m2_m))) +vuint8m2_t vle8(vbool4_t op0, vuint8m2_t op1, const uint8_t * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle8_v_u8m4_m))) +vuint8m4_t vle8(vbool2_t op0, vuint8m4_t op1, const uint8_t * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle8_v_u8m8_m))) +vuint8m8_t vle8(vbool1_t op0, vuint8m8_t op1, const uint8_t * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle8_v_u8mf2_m))) +vuint8mf2_t vle8(vbool16_t op0, vuint8mf2_t op1, const uint8_t * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle8_v_u8mf4_m))) +vuint8mf4_t vle8(vbool32_t op0, vuint8mf4_t op1, const uint8_t * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle8_v_u8mf8_m))) +vuint8mf8_t vle8(vbool64_t op0, vuint8mf8_t op1, const uint8_t * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle8ff_v_i8m1_m))) +vint8m1_t vle8ff(vbool8_t op0, vint8m1_t op1, const int8_t * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle8ff_v_i8m2_m))) +vint8m2_t vle8ff(vbool4_t op0, vint8m2_t op1, const int8_t * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle8ff_v_i8m4_m))) +vint8m4_t vle8ff(vbool2_t op0, vint8m4_t op1, const int8_t * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle8ff_v_i8m8_m))) +vint8m8_t vle8ff(vbool1_t op0, vint8m8_t op1, const int8_t * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle8ff_v_i8mf2_m))) +vint8mf2_t vle8ff(vbool16_t op0, vint8mf2_t op1, const int8_t * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle8ff_v_i8mf4_m))) +vint8mf4_t vle8ff(vbool32_t op0, vint8mf4_t op1, const int8_t * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle8ff_v_i8mf8_m))) +vint8mf8_t vle8ff(vbool64_t op0, vint8mf8_t op1, const int8_t * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle8ff_v_u8m1_m))) +vuint8m1_t vle8ff(vbool8_t op0, vuint8m1_t op1, const uint8_t * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle8ff_v_u8m2_m))) +vuint8m2_t vle8ff(vbool4_t op0, vuint8m2_t op1, const uint8_t * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle8ff_v_u8m4_m))) +vuint8m4_t vle8ff(vbool2_t op0, vuint8m4_t op1, const uint8_t * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle8ff_v_u8m8_m))) +vuint8m8_t vle8ff(vbool1_t op0, vuint8m8_t op1, const uint8_t * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle8ff_v_u8mf2_m))) +vuint8mf2_t vle8ff(vbool16_t op0, vuint8mf2_t op1, const uint8_t * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle8ff_v_u8mf4_m))) +vuint8mf4_t vle8ff(vbool32_t op0, vuint8mf4_t op1, const uint8_t * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle8ff_v_u8mf8_m))) +vuint8mf8_t vle8ff(vbool64_t op0, vuint8mf8_t op1, const uint8_t * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vneg_v_i8m1))) +vint8m1_t vneg(vint8m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vneg_v_i8m1_m))) +vint8m1_t vneg(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vneg_v_i8m2))) +vint8m2_t vneg(vint8m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vneg_v_i8m2_m))) +vint8m2_t vneg(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vneg_v_i8m4))) +vint8m4_t vneg(vint8m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vneg_v_i8m4_m))) +vint8m4_t vneg(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vneg_v_i8m8))) +vint8m8_t vneg(vint8m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vneg_v_i8m8_m))) +vint8m8_t vneg(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vneg_v_i8mf2))) +vint8mf2_t vneg(vint8mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vneg_v_i8mf2_m))) +vint8mf2_t vneg(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vneg_v_i8mf4))) +vint8mf4_t vneg(vint8mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vneg_v_i8mf4_m))) +vint8mf4_t vneg(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vneg_v_i8mf8))) +vint8mf8_t vneg(vint8mf8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vneg_v_i8mf8_m))) +vint8mf8_t vneg(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vneg_v_i16m1))) +vint16m1_t vneg(vint16m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vneg_v_i16m1_m))) +vint16m1_t vneg(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vneg_v_i16m2))) +vint16m2_t vneg(vint16m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vneg_v_i16m2_m))) +vint16m2_t vneg(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vneg_v_i16m4))) +vint16m4_t vneg(vint16m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vneg_v_i16m4_m))) +vint16m4_t vneg(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vneg_v_i16m8))) +vint16m8_t vneg(vint16m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vneg_v_i16m8_m))) +vint16m8_t vneg(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vneg_v_i16mf2))) +vint16mf2_t vneg(vint16mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vneg_v_i16mf2_m))) +vint16mf2_t vneg(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vneg_v_i16mf4))) +vint16mf4_t vneg(vint16mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vneg_v_i16mf4_m))) +vint16mf4_t vneg(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vneg_v_i32m1))) +vint32m1_t vneg(vint32m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vneg_v_i32m1_m))) +vint32m1_t vneg(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vneg_v_i32m2))) +vint32m2_t vneg(vint32m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vneg_v_i32m2_m))) +vint32m2_t vneg(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vneg_v_i32m4))) +vint32m4_t vneg(vint32m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vneg_v_i32m4_m))) +vint32m4_t vneg(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vneg_v_i32m8))) +vint32m8_t vneg(vint32m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vneg_v_i32m8_m))) +vint32m8_t vneg(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vneg_v_i32mf2))) +vint32mf2_t vneg(vint32mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vneg_v_i32mf2_m))) +vint32mf2_t vneg(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vneg_v_i64m1))) +vint64m1_t vneg(vint64m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vneg_v_i64m1_m))) +vint64m1_t vneg(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vneg_v_i64m2))) +vint64m2_t vneg(vint64m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vneg_v_i64m2_m))) +vint64m2_t vneg(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vneg_v_i64m4))) +vint64m4_t vneg(vint64m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vneg_v_i64m4_m))) +vint64m4_t vneg(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vneg_v_i64m8))) +vint64m8_t vneg(vint64m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vneg_v_i64m8_m))) +vint64m8_t vneg(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_i8m1))) +vint8m1_t vnot(vint8m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_i8m1_m))) +vint8m1_t vnot(vbool8_t op0, vint8m1_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_i8m2))) +vint8m2_t vnot(vint8m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_i8m2_m))) +vint8m2_t vnot(vbool4_t op0, vint8m2_t op1, vint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_i8m4))) +vint8m4_t vnot(vint8m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_i8m4_m))) +vint8m4_t vnot(vbool2_t op0, vint8m4_t op1, vint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_i8m8))) +vint8m8_t vnot(vint8m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_i8m8_m))) +vint8m8_t vnot(vbool1_t op0, vint8m8_t op1, vint8m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_i8mf2))) +vint8mf2_t vnot(vint8mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_i8mf2_m))) +vint8mf2_t vnot(vbool16_t op0, vint8mf2_t op1, vint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_i8mf4))) +vint8mf4_t vnot(vint8mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_i8mf4_m))) +vint8mf4_t vnot(vbool32_t op0, vint8mf4_t op1, vint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_i8mf8))) +vint8mf8_t vnot(vint8mf8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_i8mf8_m))) +vint8mf8_t vnot(vbool64_t op0, vint8mf8_t op1, vint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_i16m1))) +vint16m1_t vnot(vint16m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_i16m1_m))) +vint16m1_t vnot(vbool16_t op0, vint16m1_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_i16m2))) +vint16m2_t vnot(vint16m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_i16m2_m))) +vint16m2_t vnot(vbool8_t op0, vint16m2_t op1, vint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_i16m4))) +vint16m4_t vnot(vint16m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_i16m4_m))) +vint16m4_t vnot(vbool4_t op0, vint16m4_t op1, vint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_i16m8))) +vint16m8_t vnot(vint16m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_i16m8_m))) +vint16m8_t vnot(vbool2_t op0, vint16m8_t op1, vint16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_i16mf2))) +vint16mf2_t vnot(vint16mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_i16mf2_m))) +vint16mf2_t vnot(vbool32_t op0, vint16mf2_t op1, vint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_i16mf4))) +vint16mf4_t vnot(vint16mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_i16mf4_m))) +vint16mf4_t vnot(vbool64_t op0, vint16mf4_t op1, vint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_i32m1))) +vint32m1_t vnot(vint32m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_i32m1_m))) +vint32m1_t vnot(vbool32_t op0, vint32m1_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_i32m2))) +vint32m2_t vnot(vint32m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_i32m2_m))) +vint32m2_t vnot(vbool16_t op0, vint32m2_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_i32m4))) +vint32m4_t vnot(vint32m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_i32m4_m))) +vint32m4_t vnot(vbool8_t op0, vint32m4_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_i32m8))) +vint32m8_t vnot(vint32m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_i32m8_m))) +vint32m8_t vnot(vbool4_t op0, vint32m8_t op1, vint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_i32mf2))) +vint32mf2_t vnot(vint32mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_i32mf2_m))) +vint32mf2_t vnot(vbool64_t op0, vint32mf2_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_i64m1))) +vint64m1_t vnot(vint64m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_i64m1_m))) +vint64m1_t vnot(vbool64_t op0, vint64m1_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_i64m2))) +vint64m2_t vnot(vint64m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_i64m2_m))) +vint64m2_t vnot(vbool32_t op0, vint64m2_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_i64m4))) +vint64m4_t vnot(vint64m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_i64m4_m))) +vint64m4_t vnot(vbool16_t op0, vint64m4_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_i64m8))) +vint64m8_t vnot(vint64m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_i64m8_m))) +vint64m8_t vnot(vbool8_t op0, vint64m8_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse8_v_i8m1_m))) +vint8m1_t vlse8(vbool8_t op0, vint8m1_t op1, const int8_t * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse8_v_i8m2_m))) +vint8m2_t vlse8(vbool4_t op0, vint8m2_t op1, const int8_t * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse8_v_i8m4_m))) +vint8m4_t vlse8(vbool2_t op0, vint8m4_t op1, const int8_t * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse8_v_i8m8_m))) +vint8m8_t vlse8(vbool1_t op0, vint8m8_t op1, const int8_t * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse8_v_i8mf2_m))) +vint8mf2_t vlse8(vbool16_t op0, vint8mf2_t op1, const int8_t * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse8_v_i8mf4_m))) +vint8mf4_t vlse8(vbool32_t op0, vint8mf4_t op1, const int8_t * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse8_v_i8mf8_m))) +vint8mf8_t vlse8(vbool64_t op0, vint8mf8_t op1, const int8_t * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_u8m1))) +vuint8m1_t vnot(vuint8m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_u8m1_m))) +vuint8m1_t vnot(vbool8_t op0, vuint8m1_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_u8m2))) +vuint8m2_t vnot(vuint8m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_u8m2_m))) +vuint8m2_t vnot(vbool4_t op0, vuint8m2_t op1, vuint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_u8m4))) +vuint8m4_t vnot(vuint8m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_u8m4_m))) +vuint8m4_t vnot(vbool2_t op0, vuint8m4_t op1, vuint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_u8m8))) +vuint8m8_t vnot(vuint8m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_u8m8_m))) +vuint8m8_t vnot(vbool1_t op0, vuint8m8_t op1, vuint8m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_u8mf2))) +vuint8mf2_t vnot(vuint8mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_u8mf2_m))) +vuint8mf2_t vnot(vbool16_t op0, vuint8mf2_t op1, vuint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_u8mf4))) +vuint8mf4_t vnot(vuint8mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_u8mf4_m))) +vuint8mf4_t vnot(vbool32_t op0, vuint8mf4_t op1, vuint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_u8mf8))) +vuint8mf8_t vnot(vuint8mf8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_u8mf8_m))) +vuint8mf8_t vnot(vbool64_t op0, vuint8mf8_t op1, vuint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_u16m1))) +vuint16m1_t vnot(vuint16m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_u16m1_m))) +vuint16m1_t vnot(vbool16_t op0, vuint16m1_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_u16m2))) +vuint16m2_t vnot(vuint16m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_u16m2_m))) +vuint16m2_t vnot(vbool8_t op0, vuint16m2_t op1, vuint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_u16m4))) +vuint16m4_t vnot(vuint16m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_u16m4_m))) +vuint16m4_t vnot(vbool4_t op0, vuint16m4_t op1, vuint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_u16m8))) +vuint16m8_t vnot(vuint16m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_u16m8_m))) +vuint16m8_t vnot(vbool2_t op0, vuint16m8_t op1, vuint16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_u16mf2))) +vuint16mf2_t vnot(vuint16mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_u16mf2_m))) +vuint16mf2_t vnot(vbool32_t op0, vuint16mf2_t op1, vuint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_u16mf4))) +vuint16mf4_t vnot(vuint16mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_u16mf4_m))) +vuint16mf4_t vnot(vbool64_t op0, vuint16mf4_t op1, vuint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_u32m1))) +vuint32m1_t vnot(vuint32m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_u32m1_m))) +vuint32m1_t vnot(vbool32_t op0, vuint32m1_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_u32m2))) +vuint32m2_t vnot(vuint32m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_u32m2_m))) +vuint32m2_t vnot(vbool16_t op0, vuint32m2_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_u32m4))) +vuint32m4_t vnot(vuint32m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_u32m4_m))) +vuint32m4_t vnot(vbool8_t op0, vuint32m4_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_u32m8))) +vuint32m8_t vnot(vuint32m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_u32m8_m))) +vuint32m8_t vnot(vbool4_t op0, vuint32m8_t op1, vuint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_u32mf2))) +vuint32mf2_t vnot(vuint32mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_u32mf2_m))) +vuint32mf2_t vnot(vbool64_t op0, vuint32mf2_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_u64m1))) +vuint64m1_t vnot(vuint64m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_u64m1_m))) +vuint64m1_t vnot(vbool64_t op0, vuint64m1_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_u64m2))) +vuint64m2_t vnot(vuint64m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_u64m2_m))) +vuint64m2_t vnot(vbool32_t op0, vuint64m2_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_u64m4))) +vuint64m4_t vnot(vuint64m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_u64m4_m))) +vuint64m4_t vnot(vbool16_t op0, vuint64m4_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_u64m8))) +vuint64m8_t vnot(vuint64m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vnot_v_u64m8_m))) +vuint64m8_t vnot(vbool8_t op0, vuint64m8_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmmv_m_b8))) +vbool8_t vmmv(vbool8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmmv_m_b4))) +vbool4_t vmmv(vbool4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmmv_m_b2))) +vbool2_t vmmv(vbool2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmmv_m_b1))) +vbool1_t vmmv(vbool1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmmv_m_b16))) +vbool16_t vmmv(vbool16_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmmv_m_b32))) +vbool32_t vmmv(vbool32_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmmv_m_b64))) +vbool64_t vmmv(vbool64_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvtu_x_x_v_u16m2))) +vuint16m2_t vwcvtu_x(vuint8m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvtu_x_x_v_u16m2_m))) +vuint16m2_t vwcvtu_x(vbool8_t op0, vuint16m2_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvtu_x_x_v_u16m4))) +vuint16m4_t vwcvtu_x(vuint8m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvtu_x_x_v_u16m4_m))) +vuint16m4_t vwcvtu_x(vbool4_t op0, vuint16m4_t op1, vuint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvtu_x_x_v_u16m8))) +vuint16m8_t vwcvtu_x(vuint8m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvtu_x_x_v_u16m8_m))) +vuint16m8_t vwcvtu_x(vbool2_t op0, vuint16m8_t op1, vuint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvtu_x_x_v_u16m1))) +vuint16m1_t vwcvtu_x(vuint8mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvtu_x_x_v_u16m1_m))) +vuint16m1_t vwcvtu_x(vbool16_t op0, vuint16m1_t op1, vuint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvtu_x_x_v_u16mf2))) +vuint16mf2_t vwcvtu_x(vuint8mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvtu_x_x_v_u16mf2_m))) +vuint16mf2_t vwcvtu_x(vbool32_t op0, vuint16mf2_t op1, vuint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvtu_x_x_v_u16mf4))) +vuint16mf4_t vwcvtu_x(vuint8mf8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvtu_x_x_v_u16mf4_m))) +vuint16mf4_t vwcvtu_x(vbool64_t op0, vuint16mf4_t op1, vuint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvtu_x_x_v_u32m2))) +vuint32m2_t vwcvtu_x(vuint16m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvtu_x_x_v_u32m2_m))) +vuint32m2_t vwcvtu_x(vbool16_t op0, vuint32m2_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvtu_x_x_v_u32m4))) +vuint32m4_t vwcvtu_x(vuint16m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvtu_x_x_v_u32m4_m))) +vuint32m4_t vwcvtu_x(vbool8_t op0, vuint32m4_t op1, vuint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvtu_x_x_v_u32m8))) +vuint32m8_t vwcvtu_x(vuint16m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvtu_x_x_v_u32m8_m))) +vuint32m8_t vwcvtu_x(vbool4_t op0, vuint32m8_t op1, vuint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvtu_x_x_v_u32m1))) +vuint32m1_t vwcvtu_x(vuint16mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvtu_x_x_v_u32m1_m))) +vuint32m1_t vwcvtu_x(vbool32_t op0, vuint32m1_t op1, vuint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvtu_x_x_v_u32mf2))) +vuint32mf2_t vwcvtu_x(vuint16mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvtu_x_x_v_u32mf2_m))) +vuint32mf2_t vwcvtu_x(vbool64_t op0, vuint32mf2_t op1, vuint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvtu_x_x_v_u64m2))) +vuint64m2_t vwcvtu_x(vuint32m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvtu_x_x_v_u64m2_m))) +vuint64m2_t vwcvtu_x(vbool32_t op0, vuint64m2_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvtu_x_x_v_u64m4))) +vuint64m4_t vwcvtu_x(vuint32m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvtu_x_x_v_u64m4_m))) +vuint64m4_t vwcvtu_x(vbool16_t op0, vuint64m4_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvtu_x_x_v_u64m8))) +vuint64m8_t vwcvtu_x(vuint32m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvtu_x_x_v_u64m8_m))) +vuint64m8_t vwcvtu_x(vbool8_t op0, vuint64m8_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvtu_x_x_v_u64m1))) +vuint64m1_t vwcvtu_x(vuint32mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vwcvtu_x_x_v_u64m1_m))) +vuint64m1_t vwcvtu_x(vbool64_t op0, vuint64m1_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_i8m1))) +vint8m1_t vncvt_x(vint16m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_i8m1_m))) +vint8m1_t vncvt_x(vbool8_t op0, vint8m1_t op1, vint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_i8m2))) +vint8m2_t vncvt_x(vint16m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_i8m2_m))) +vint8m2_t vncvt_x(vbool4_t op0, vint8m2_t op1, vint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_i8m4))) +vint8m4_t vncvt_x(vint16m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_i8m4_m))) +vint8m4_t vncvt_x(vbool2_t op0, vint8m4_t op1, vint16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_i8mf2))) +vint8mf2_t vncvt_x(vint16m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_i8mf2_m))) +vint8mf2_t vncvt_x(vbool16_t op0, vint8mf2_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_i8mf4))) +vint8mf4_t vncvt_x(vint16mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_i8mf4_m))) +vint8mf4_t vncvt_x(vbool32_t op0, vint8mf4_t op1, vint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_i8mf8))) +vint8mf8_t vncvt_x(vint16mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_i8mf8_m))) +vint8mf8_t vncvt_x(vbool64_t op0, vint8mf8_t op1, vint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_i16m1))) +vint16m1_t vncvt_x(vint32m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_i16m1_m))) +vint16m1_t vncvt_x(vbool16_t op0, vint16m1_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_i16m2))) +vint16m2_t vncvt_x(vint32m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_i16m2_m))) +vint16m2_t vncvt_x(vbool8_t op0, vint16m2_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_i16m4))) +vint16m4_t vncvt_x(vint32m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_i16m4_m))) +vint16m4_t vncvt_x(vbool4_t op0, vint16m4_t op1, vint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_i16mf2))) +vint16mf2_t vncvt_x(vint32m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_i16mf2_m))) +vint16mf2_t vncvt_x(vbool32_t op0, vint16mf2_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_i16mf4))) +vint16mf4_t vncvt_x(vint32mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_i16mf4_m))) +vint16mf4_t vncvt_x(vbool64_t op0, vint16mf4_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_i32m1))) +vint32m1_t vncvt_x(vint64m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_i32m1_m))) +vint32m1_t vncvt_x(vbool32_t op0, vint32m1_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_i32m2))) +vint32m2_t vncvt_x(vint64m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_i32m2_m))) +vint32m2_t vncvt_x(vbool16_t op0, vint32m2_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_i32m4))) +vint32m4_t vncvt_x(vint64m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_i32m4_m))) +vint32m4_t vncvt_x(vbool8_t op0, vint32m4_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_i32mf2))) +vint32mf2_t vncvt_x(vint64m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vncvt_x_x_w_i32mf2_m))) +vint32mf2_t vncvt_x(vbool64_t op0, vint32mf2_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle16_v_i16m1_m))) +vint16m1_t vle16(vbool16_t op0, vint16m1_t op1, const int16_t * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle16_v_i16m2_m))) +vint16m2_t vle16(vbool8_t op0, vint16m2_t op1, const int16_t * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle16_v_i16m4_m))) +vint16m4_t vle16(vbool4_t op0, vint16m4_t op1, const int16_t * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle16_v_i16m8_m))) +vint16m8_t vle16(vbool2_t op0, vint16m8_t op1, const int16_t * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle16_v_i16mf2_m))) +vint16mf2_t vle16(vbool32_t op0, vint16mf2_t op1, const int16_t * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle16_v_i16mf4_m))) +vint16mf4_t vle16(vbool64_t op0, vint16mf4_t op1, const int16_t * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle16_v_u16m1_m))) +vuint16m1_t vle16(vbool16_t op0, vuint16m1_t op1, const uint16_t * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle16_v_u16m2_m))) +vuint16m2_t vle16(vbool8_t op0, vuint16m2_t op1, const uint16_t * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle16_v_u16m4_m))) +vuint16m4_t vle16(vbool4_t op0, vuint16m4_t op1, const uint16_t * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle16_v_u16m8_m))) +vuint16m8_t vle16(vbool2_t op0, vuint16m8_t op1, const uint16_t * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle16_v_u16mf2_m))) +vuint16mf2_t vle16(vbool32_t op0, vuint16mf2_t op1, const uint16_t * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle16_v_u16mf4_m))) +vuint16mf4_t vle16(vbool64_t op0, vuint16mf4_t op1, const uint16_t * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle32_v_i32m1_m))) +vint32m1_t vle32(vbool32_t op0, vint32m1_t op1, const int32_t * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle32_v_i32m2_m))) +vint32m2_t vle32(vbool16_t op0, vint32m2_t op1, const int32_t * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle32_v_i32m4_m))) +vint32m4_t vle32(vbool8_t op0, vint32m4_t op1, const int32_t * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle32_v_i32m8_m))) +vint32m8_t vle32(vbool4_t op0, vint32m8_t op1, const int32_t * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle32_v_i32mf2_m))) +vint32mf2_t vle32(vbool64_t op0, vint32mf2_t op1, const int32_t * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle32_v_u32m1_m))) +vuint32m1_t vle32(vbool32_t op0, vuint32m1_t op1, const uint32_t * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle32_v_u32m2_m))) +vuint32m2_t vle32(vbool16_t op0, vuint32m2_t op1, const uint32_t * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle32_v_u32m4_m))) +vuint32m4_t vle32(vbool8_t op0, vuint32m4_t op1, const uint32_t * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle32_v_u32m8_m))) +vuint32m8_t vle32(vbool4_t op0, vuint32m8_t op1, const uint32_t * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle32_v_u32mf2_m))) +vuint32mf2_t vle32(vbool64_t op0, vuint32mf2_t op1, const uint32_t * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse8_v_u8m1_m))) +vuint8m1_t vlse8(vbool8_t op0, vuint8m1_t op1, const uint8_t * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse8_v_u8m2_m))) +vuint8m2_t vlse8(vbool4_t op0, vuint8m2_t op1, const uint8_t * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse8_v_u8m4_m))) +vuint8m4_t vlse8(vbool2_t op0, vuint8m4_t op1, const uint8_t * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse8_v_u8m8_m))) +vuint8m8_t vlse8(vbool1_t op0, vuint8m8_t op1, const uint8_t * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse8_v_u8mf2_m))) +vuint8mf2_t vlse8(vbool16_t op0, vuint8mf2_t op1, const uint8_t * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse8_v_u8mf4_m))) +vuint8mf4_t vlse8(vbool32_t op0, vuint8mf4_t op1, const uint8_t * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse8_v_u8mf8_m))) +vuint8mf8_t vlse8(vbool64_t op0, vuint8mf8_t op1, const uint8_t * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle64_v_i64m1_m))) +vint64m1_t vle64(vbool64_t op0, vint64m1_t op1, const int64_t * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle64_v_i64m2_m))) +vint64m2_t vle64(vbool32_t op0, vint64m2_t op1, const int64_t * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle64_v_i64m4_m))) +vint64m4_t vle64(vbool16_t op0, vint64m4_t op1, const int64_t * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle64_v_i64m8_m))) +vint64m8_t vle64(vbool8_t op0, vint64m8_t op1, const int64_t * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle64_v_u64m1_m))) +vuint64m1_t vle64(vbool64_t op0, vuint64m1_t op1, const uint64_t * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle64_v_u64m2_m))) +vuint64m2_t vle64(vbool32_t op0, vuint64m2_t op1, const uint64_t * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle64_v_u64m4_m))) +vuint64m4_t vle64(vbool16_t op0, vuint64m4_t op1, const uint64_t * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle64_v_u64m8_m))) +vuint64m8_t vle64(vbool8_t op0, vuint64m8_t op1, const uint64_t * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse16_v_i16m1))) +void vse16(int16_t * op0, vint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse16_v_i16m1_m))) +void vse16(vbool16_t op0, int16_t * op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse16_v_i16m2))) +void vse16(int16_t * op0, vint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse16_v_i16m2_m))) +void vse16(vbool8_t op0, int16_t * op1, vint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse16_v_i16m4))) +void vse16(int16_t * op0, vint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse16_v_i16m4_m))) +void vse16(vbool4_t op0, int16_t * op1, vint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse16_v_i16m8))) +void vse16(int16_t * op0, vint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse16_v_i16m8_m))) +void vse16(vbool2_t op0, int16_t * op1, vint16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse16_v_i16mf2))) +void vse16(int16_t * op0, vint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse16_v_i16mf2_m))) +void vse16(vbool32_t op0, int16_t * op1, vint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse16_v_i16mf4))) +void vse16(int16_t * op0, vint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse16_v_i16mf4_m))) +void vse16(vbool64_t op0, int16_t * op1, vint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse16_v_u16m1))) +void vse16(uint16_t * op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse16_v_u16m1_m))) +void vse16(vbool16_t op0, uint16_t * op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse16_v_u16m2))) +void vse16(uint16_t * op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse16_v_u16m2_m))) +void vse16(vbool8_t op0, uint16_t * op1, vuint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse16_v_u16m4))) +void vse16(uint16_t * op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse16_v_u16m4_m))) +void vse16(vbool4_t op0, uint16_t * op1, vuint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse16_v_u16m8))) +void vse16(uint16_t * op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse16_v_u16m8_m))) +void vse16(vbool2_t op0, uint16_t * op1, vuint16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse16_v_u16mf2))) +void vse16(uint16_t * op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse16_v_u16mf2_m))) +void vse16(vbool32_t op0, uint16_t * op1, vuint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse16_v_u16mf4))) +void vse16(uint16_t * op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse16_v_u16mf4_m))) +void vse16(vbool64_t op0, uint16_t * op1, vuint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse32_v_i32m1))) +void vse32(int32_t * op0, vint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse32_v_i32m1_m))) +void vse32(vbool32_t op0, int32_t * op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse32_v_i32m2))) +void vse32(int32_t * op0, vint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse32_v_i32m2_m))) +void vse32(vbool16_t op0, int32_t * op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse32_v_i32m4))) +void vse32(int32_t * op0, vint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse32_v_i32m4_m))) +void vse32(vbool8_t op0, int32_t * op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse32_v_i32m8))) +void vse32(int32_t * op0, vint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse32_v_i32m8_m))) +void vse32(vbool4_t op0, int32_t * op1, vint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse32_v_i32mf2))) +void vse32(int32_t * op0, vint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse32_v_i32mf2_m))) +void vse32(vbool64_t op0, int32_t * op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse32_v_u32m1))) +void vse32(uint32_t * op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse32_v_u32m1_m))) +void vse32(vbool32_t op0, uint32_t * op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse32_v_u32m2))) +void vse32(uint32_t * op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse32_v_u32m2_m))) +void vse32(vbool16_t op0, uint32_t * op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse32_v_u32m4))) +void vse32(uint32_t * op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse32_v_u32m4_m))) +void vse32(vbool8_t op0, uint32_t * op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse32_v_u32m8))) +void vse32(uint32_t * op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse32_v_u32m8_m))) +void vse32(vbool4_t op0, uint32_t * op1, vuint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse32_v_u32mf2))) +void vse32(uint32_t * op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse32_v_u32mf2_m))) +void vse32(vbool64_t op0, uint32_t * op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_i8m1))) +vint8m1_t vluxei8(const int8_t * op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_i8m1_m))) +vint8m1_t vluxei8(vbool8_t op0, vint8m1_t op1, const int8_t * op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_i8m2))) +vint8m2_t vluxei8(const int8_t * op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_i8m2_m))) +vint8m2_t vluxei8(vbool4_t op0, vint8m2_t op1, const int8_t * op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_i8m4))) +vint8m4_t vluxei8(const int8_t * op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_i8m4_m))) +vint8m4_t vluxei8(vbool2_t op0, vint8m4_t op1, const int8_t * op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_i8m8))) +vint8m8_t vluxei8(const int8_t * op0, vuint8m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_i8m8_m))) +vint8m8_t vluxei8(vbool1_t op0, vint8m8_t op1, const int8_t * op2, vuint8m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_i8mf2))) +vint8mf2_t vluxei8(const int8_t * op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_i8mf2_m))) +vint8mf2_t vluxei8(vbool16_t op0, vint8mf2_t op1, const int8_t * op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_i8mf4))) +vint8mf4_t vluxei8(const int8_t * op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_i8mf4_m))) +vint8mf4_t vluxei8(vbool32_t op0, vint8mf4_t op1, const int8_t * op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_i8mf8))) +vint8mf8_t vluxei8(const int8_t * op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_i8mf8_m))) +vint8mf8_t vluxei8(vbool64_t op0, vint8mf8_t op1, const int8_t * op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse64_v_i64m1))) +void vse64(int64_t * op0, vint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse64_v_i64m1_m))) +void vse64(vbool64_t op0, int64_t * op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse64_v_i64m2))) +void vse64(int64_t * op0, vint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse64_v_i64m2_m))) +void vse64(vbool32_t op0, int64_t * op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse64_v_i64m4))) +void vse64(int64_t * op0, vint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse64_v_i64m4_m))) +void vse64(vbool16_t op0, int64_t * op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse64_v_i64m8))) +void vse64(int64_t * op0, vint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse64_v_i64m8_m))) +void vse64(vbool8_t op0, int64_t * op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse64_v_u64m1))) +void vse64(uint64_t * op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse64_v_u64m1_m))) +void vse64(vbool64_t op0, uint64_t * op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse64_v_u64m2))) +void vse64(uint64_t * op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse64_v_u64m2_m))) +void vse64(vbool32_t op0, uint64_t * op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse64_v_u64m4))) +void vse64(uint64_t * op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse64_v_u64m4_m))) +void vse64(vbool16_t op0, uint64_t * op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse64_v_u64m8))) +void vse64(uint64_t * op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse64_v_u64m8_m))) +void vse64(vbool8_t op0, uint64_t * op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse16_v_i16m1_m))) +vint16m1_t vlse16(vbool16_t op0, vint16m1_t op1, const int16_t * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse16_v_i16m2_m))) +vint16m2_t vlse16(vbool8_t op0, vint16m2_t op1, const int16_t * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse16_v_i16m4_m))) +vint16m4_t vlse16(vbool4_t op0, vint16m4_t op1, const int16_t * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse16_v_i16m8_m))) +vint16m8_t vlse16(vbool2_t op0, vint16m8_t op1, const int16_t * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse16_v_i16mf2_m))) +vint16mf2_t vlse16(vbool32_t op0, vint16mf2_t op1, const int16_t * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse16_v_i16mf4_m))) +vint16mf4_t vlse16(vbool64_t op0, vint16mf4_t op1, const int16_t * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse16_v_u16m1_m))) +vuint16m1_t vlse16(vbool16_t op0, vuint16m1_t op1, const uint16_t * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse16_v_u16m2_m))) +vuint16m2_t vlse16(vbool8_t op0, vuint16m2_t op1, const uint16_t * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse16_v_u16m4_m))) +vuint16m4_t vlse16(vbool4_t op0, vuint16m4_t op1, const uint16_t * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse16_v_u16m8_m))) +vuint16m8_t vlse16(vbool2_t op0, vuint16m8_t op1, const uint16_t * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse16_v_u16mf2_m))) +vuint16mf2_t vlse16(vbool32_t op0, vuint16mf2_t op1, const uint16_t * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse16_v_u16mf4_m))) +vuint16mf4_t vlse16(vbool64_t op0, vuint16mf4_t op1, const uint16_t * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse32_v_i32m1_m))) +vint32m1_t vlse32(vbool32_t op0, vint32m1_t op1, const int32_t * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse32_v_i32m2_m))) +vint32m2_t vlse32(vbool16_t op0, vint32m2_t op1, const int32_t * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse32_v_i32m4_m))) +vint32m4_t vlse32(vbool8_t op0, vint32m4_t op1, const int32_t * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse32_v_i32m8_m))) +vint32m8_t vlse32(vbool4_t op0, vint32m8_t op1, const int32_t * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse32_v_i32mf2_m))) +vint32mf2_t vlse32(vbool64_t op0, vint32mf2_t op1, const int32_t * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse32_v_u32m1_m))) +vuint32m1_t vlse32(vbool32_t op0, vuint32m1_t op1, const uint32_t * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse32_v_u32m2_m))) +vuint32m2_t vlse32(vbool16_t op0, vuint32m2_t op1, const uint32_t * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse32_v_u32m4_m))) +vuint32m4_t vlse32(vbool8_t op0, vuint32m4_t op1, const uint32_t * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse32_v_u32m8_m))) +vuint32m8_t vlse32(vbool4_t op0, vuint32m8_t op1, const uint32_t * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse32_v_u32mf2_m))) +vuint32mf2_t vlse32(vbool64_t op0, vuint32mf2_t op1, const uint32_t * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse64_v_i64m1_m))) +vint64m1_t vlse64(vbool64_t op0, vint64m1_t op1, const int64_t * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse64_v_i64m2_m))) +vint64m2_t vlse64(vbool32_t op0, vint64m2_t op1, const int64_t * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse64_v_i64m4_m))) +vint64m4_t vlse64(vbool16_t op0, vint64m4_t op1, const int64_t * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse64_v_i64m8_m))) +vint64m8_t vlse64(vbool8_t op0, vint64m8_t op1, const int64_t * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfirst_m_b8))) +long vfirst(vbool8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfirst_m_b8_m))) +long vfirst(vbool8_t op0, vbool8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfirst_m_b4))) +long vfirst(vbool4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfirst_m_b4_m))) +long vfirst(vbool4_t op0, vbool4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfirst_m_b2))) +long vfirst(vbool2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfirst_m_b2_m))) +long vfirst(vbool2_t op0, vbool2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfirst_m_b1))) +long vfirst(vbool1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfirst_m_b1_m))) +long vfirst(vbool1_t op0, vbool1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfirst_m_b16))) +long vfirst(vbool16_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfirst_m_b16_m))) +long vfirst(vbool16_t op0, vbool16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfirst_m_b32))) +long vfirst(vbool32_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfirst_m_b32_m))) +long vfirst(vbool32_t op0, vbool32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfirst_m_b64))) +long vfirst(vbool64_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfirst_m_b64_m))) +long vfirst(vbool64_t op0, vbool64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i8mf2_i8m1))) +vint8m1_t vlmul_ext_i8m1(vint8mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i8mf4_i8m1))) +vint8m1_t vlmul_ext_i8m1(vint8mf4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i8mf8_i8m1))) +vint8m1_t vlmul_ext_i8m1(vint8mf8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i16mf2_i16m1))) +vint16m1_t vlmul_ext_i16m1(vint16mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i16mf4_i16m1))) +vint16m1_t vlmul_ext_i16m1(vint16mf4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i32mf2_i32m1))) +vint32m1_t vlmul_ext_i32m1(vint32mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i8mf4_i8mf2))) +vint8mf2_t vlmul_ext_i8mf2(vint8mf4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i8mf8_i8mf2))) +vint8mf2_t vlmul_ext_i8mf2(vint8mf8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i16mf4_i16mf2))) +vint16mf2_t vlmul_ext_i16mf2(vint16mf4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i8mf8_i8mf4))) +vint8mf4_t vlmul_ext_i8mf4(vint8mf8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i8m1_i8m2))) +vint8m2_t vlmul_ext_i8m2(vint8m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i8mf2_i8m2))) +vint8m2_t vlmul_ext_i8m2(vint8mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i8mf4_i8m2))) +vint8m2_t vlmul_ext_i8m2(vint8mf4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i8mf8_i8m2))) +vint8m2_t vlmul_ext_i8m2(vint8mf8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i16m1_i16m2))) +vint16m2_t vlmul_ext_i16m2(vint16m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i16mf2_i16m2))) +vint16m2_t vlmul_ext_i16m2(vint16mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i16mf4_i16m2))) +vint16m2_t vlmul_ext_i16m2(vint16mf4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i32m1_i32m2))) +vint32m2_t vlmul_ext_i32m2(vint32m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i32mf2_i32m2))) +vint32m2_t vlmul_ext_i32m2(vint32mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i64m1_i64m2))) +vint64m2_t vlmul_ext_i64m2(vint64m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i8m1_i8m4))) +vint8m4_t vlmul_ext_i8m4(vint8m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i8m2_i8m4))) +vint8m4_t vlmul_ext_i8m4(vint8m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i8mf2_i8m4))) +vint8m4_t vlmul_ext_i8m4(vint8mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i8mf4_i8m4))) +vint8m4_t vlmul_ext_i8m4(vint8mf4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i8mf8_i8m4))) +vint8m4_t vlmul_ext_i8m4(vint8mf8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i16m1_i16m4))) +vint16m4_t vlmul_ext_i16m4(vint16m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i16m2_i16m4))) +vint16m4_t vlmul_ext_i16m4(vint16m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i16mf2_i16m4))) +vint16m4_t vlmul_ext_i16m4(vint16mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i16mf4_i16m4))) +vint16m4_t vlmul_ext_i16m4(vint16mf4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i32m1_i32m4))) +vint32m4_t vlmul_ext_i32m4(vint32m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i32m2_i32m4))) +vint32m4_t vlmul_ext_i32m4(vint32m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i32mf2_i32m4))) +vint32m4_t vlmul_ext_i32m4(vint32mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i64m1_i64m4))) +vint64m4_t vlmul_ext_i64m4(vint64m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i64m2_i64m4))) +vint64m4_t vlmul_ext_i64m4(vint64m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i8m1_i8m8))) +vint8m8_t vlmul_ext_i8m8(vint8m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i8m2_i8m8))) +vint8m8_t vlmul_ext_i8m8(vint8m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i8m4_i8m8))) +vint8m8_t vlmul_ext_i8m8(vint8m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i8mf2_i8m8))) +vint8m8_t vlmul_ext_i8m8(vint8mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i8mf4_i8m8))) +vint8m8_t vlmul_ext_i8m8(vint8mf4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i8mf8_i8m8))) +vint8m8_t vlmul_ext_i8m8(vint8mf8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i16m1_i16m8))) +vint16m8_t vlmul_ext_i16m8(vint16m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i16m2_i16m8))) +vint16m8_t vlmul_ext_i16m8(vint16m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i16m4_i16m8))) +vint16m8_t vlmul_ext_i16m8(vint16m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i16mf2_i16m8))) +vint16m8_t vlmul_ext_i16m8(vint16mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i16mf4_i16m8))) +vint16m8_t vlmul_ext_i16m8(vint16mf4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i32m1_i32m8))) +vint32m8_t vlmul_ext_i32m8(vint32m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i32m2_i32m8))) +vint32m8_t vlmul_ext_i32m8(vint32m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i32m4_i32m8))) +vint32m8_t vlmul_ext_i32m8(vint32m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i32mf2_i32m8))) +vint32m8_t vlmul_ext_i32m8(vint32mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i64m1_i64m8))) +vint64m8_t vlmul_ext_i64m8(vint64m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i64m2_i64m8))) +vint64m8_t vlmul_ext_i64m8(vint64m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_i64m4_i64m8))) +vint64m8_t vlmul_ext_i64m8(vint64m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u8mf2_u8m1))) +vuint8m1_t vlmul_ext_u8m1(vuint8mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u8mf4_u8m1))) +vuint8m1_t vlmul_ext_u8m1(vuint8mf4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u8mf8_u8m1))) +vuint8m1_t vlmul_ext_u8m1(vuint8mf8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u16mf2_u16m1))) +vuint16m1_t vlmul_ext_u16m1(vuint16mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u16mf4_u16m1))) +vuint16m1_t vlmul_ext_u16m1(vuint16mf4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u32mf2_u32m1))) +vuint32m1_t vlmul_ext_u32m1(vuint32mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u8mf4_u8mf2))) +vuint8mf2_t vlmul_ext_u8mf2(vuint8mf4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u8mf8_u8mf2))) +vuint8mf2_t vlmul_ext_u8mf2(vuint8mf8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u16mf4_u16mf2))) +vuint16mf2_t vlmul_ext_u16mf2(vuint16mf4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u8mf8_u8mf4))) +vuint8mf4_t vlmul_ext_u8mf4(vuint8mf8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u8m1_u8m2))) +vuint8m2_t vlmul_ext_u8m2(vuint8m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u8mf2_u8m2))) +vuint8m2_t vlmul_ext_u8m2(vuint8mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u8mf4_u8m2))) +vuint8m2_t vlmul_ext_u8m2(vuint8mf4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u8mf8_u8m2))) +vuint8m2_t vlmul_ext_u8m2(vuint8mf8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u16m1_u16m2))) +vuint16m2_t vlmul_ext_u16m2(vuint16m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u16mf2_u16m2))) +vuint16m2_t vlmul_ext_u16m2(vuint16mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u16mf4_u16m2))) +vuint16m2_t vlmul_ext_u16m2(vuint16mf4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u32m1_u32m2))) +vuint32m2_t vlmul_ext_u32m2(vuint32m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u32mf2_u32m2))) +vuint32m2_t vlmul_ext_u32m2(vuint32mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u64m1_u64m2))) +vuint64m2_t vlmul_ext_u64m2(vuint64m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u8m1_u8m4))) +vuint8m4_t vlmul_ext_u8m4(vuint8m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u8m2_u8m4))) +vuint8m4_t vlmul_ext_u8m4(vuint8m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u8mf2_u8m4))) +vuint8m4_t vlmul_ext_u8m4(vuint8mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u8mf4_u8m4))) +vuint8m4_t vlmul_ext_u8m4(vuint8mf4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u8mf8_u8m4))) +vuint8m4_t vlmul_ext_u8m4(vuint8mf8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u16m1_u16m4))) +vuint16m4_t vlmul_ext_u16m4(vuint16m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u16m2_u16m4))) +vuint16m4_t vlmul_ext_u16m4(vuint16m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u16mf2_u16m4))) +vuint16m4_t vlmul_ext_u16m4(vuint16mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u16mf4_u16m4))) +vuint16m4_t vlmul_ext_u16m4(vuint16mf4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u32m1_u32m4))) +vuint32m4_t vlmul_ext_u32m4(vuint32m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u32m2_u32m4))) +vuint32m4_t vlmul_ext_u32m4(vuint32m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u32mf2_u32m4))) +vuint32m4_t vlmul_ext_u32m4(vuint32mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u64m1_u64m4))) +vuint64m4_t vlmul_ext_u64m4(vuint64m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u64m2_u64m4))) +vuint64m4_t vlmul_ext_u64m4(vuint64m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u8m1_u8m8))) +vuint8m8_t vlmul_ext_u8m8(vuint8m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u8m2_u8m8))) +vuint8m8_t vlmul_ext_u8m8(vuint8m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u8m4_u8m8))) +vuint8m8_t vlmul_ext_u8m8(vuint8m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u8mf2_u8m8))) +vuint8m8_t vlmul_ext_u8m8(vuint8mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u8mf4_u8m8))) +vuint8m8_t vlmul_ext_u8m8(vuint8mf4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u8mf8_u8m8))) +vuint8m8_t vlmul_ext_u8m8(vuint8mf8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u16m1_u16m8))) +vuint16m8_t vlmul_ext_u16m8(vuint16m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u16m2_u16m8))) +vuint16m8_t vlmul_ext_u16m8(vuint16m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u16m4_u16m8))) +vuint16m8_t vlmul_ext_u16m8(vuint16m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u16mf2_u16m8))) +vuint16m8_t vlmul_ext_u16m8(vuint16mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u16mf4_u16m8))) +vuint16m8_t vlmul_ext_u16m8(vuint16mf4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u32m1_u32m8))) +vuint32m8_t vlmul_ext_u32m8(vuint32m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u32m2_u32m8))) +vuint32m8_t vlmul_ext_u32m8(vuint32m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u32m4_u32m8))) +vuint32m8_t vlmul_ext_u32m8(vuint32m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u32mf2_u32m8))) +vuint32m8_t vlmul_ext_u32m8(vuint32mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u64m1_u64m8))) +vuint64m8_t vlmul_ext_u64m8(vuint64m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u64m2_u64m8))) +vuint64m8_t vlmul_ext_u64m8(vuint64m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_u64m4_u64m8))) +vuint64m8_t vlmul_ext_u64m8(vuint64m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i8m1_i8mf2))) +vint8mf2_t vlmul_trunc_i8mf2(vint8m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i8m2_i8mf2))) +vint8mf2_t vlmul_trunc_i8mf2(vint8m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i8m4_i8mf2))) +vint8mf2_t vlmul_trunc_i8mf2(vint8m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i8m8_i8mf2))) +vint8mf2_t vlmul_trunc_i8mf2(vint8m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i16m1_i16mf2))) +vint16mf2_t vlmul_trunc_i16mf2(vint16m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i16m2_i16mf2))) +vint16mf2_t vlmul_trunc_i16mf2(vint16m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i16m4_i16mf2))) +vint16mf2_t vlmul_trunc_i16mf2(vint16m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i16m8_i16mf2))) +vint16mf2_t vlmul_trunc_i16mf2(vint16m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i32m1_i32mf2))) +vint32mf2_t vlmul_trunc_i32mf2(vint32m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i32m2_i32mf2))) +vint32mf2_t vlmul_trunc_i32mf2(vint32m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i32m4_i32mf2))) +vint32mf2_t vlmul_trunc_i32mf2(vint32m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i32m8_i32mf2))) +vint32mf2_t vlmul_trunc_i32mf2(vint32m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i8m1_i8mf4))) +vint8mf4_t vlmul_trunc_i8mf4(vint8m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i8m2_i8mf4))) +vint8mf4_t vlmul_trunc_i8mf4(vint8m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i8m4_i8mf4))) +vint8mf4_t vlmul_trunc_i8mf4(vint8m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i8m8_i8mf4))) +vint8mf4_t vlmul_trunc_i8mf4(vint8m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i8mf2_i8mf4))) +vint8mf4_t vlmul_trunc_i8mf4(vint8mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i16m1_i16mf4))) +vint16mf4_t vlmul_trunc_i16mf4(vint16m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i16m2_i16mf4))) +vint16mf4_t vlmul_trunc_i16mf4(vint16m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i16m4_i16mf4))) +vint16mf4_t vlmul_trunc_i16mf4(vint16m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i16m8_i16mf4))) +vint16mf4_t vlmul_trunc_i16mf4(vint16m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i16mf2_i16mf4))) +vint16mf4_t vlmul_trunc_i16mf4(vint16mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i8m1_i8mf8))) +vint8mf8_t vlmul_trunc_i8mf8(vint8m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i8m2_i8mf8))) +vint8mf8_t vlmul_trunc_i8mf8(vint8m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i8m4_i8mf8))) +vint8mf8_t vlmul_trunc_i8mf8(vint8m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i8m8_i8mf8))) +vint8mf8_t vlmul_trunc_i8mf8(vint8m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i8mf2_i8mf8))) +vint8mf8_t vlmul_trunc_i8mf8(vint8mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i8mf4_i8mf8))) +vint8mf8_t vlmul_trunc_i8mf8(vint8mf4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i8m2_i8m1))) +vint8m1_t vlmul_trunc_i8m1(vint8m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i8m4_i8m1))) +vint8m1_t vlmul_trunc_i8m1(vint8m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i8m8_i8m1))) +vint8m1_t vlmul_trunc_i8m1(vint8m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i16m2_i16m1))) +vint16m1_t vlmul_trunc_i16m1(vint16m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i16m4_i16m1))) +vint16m1_t vlmul_trunc_i16m1(vint16m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i16m8_i16m1))) +vint16m1_t vlmul_trunc_i16m1(vint16m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i32m2_i32m1))) +vint32m1_t vlmul_trunc_i32m1(vint32m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i32m4_i32m1))) +vint32m1_t vlmul_trunc_i32m1(vint32m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i32m8_i32m1))) +vint32m1_t vlmul_trunc_i32m1(vint32m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i64m2_i64m1))) +vint64m1_t vlmul_trunc_i64m1(vint64m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i64m4_i64m1))) +vint64m1_t vlmul_trunc_i64m1(vint64m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i64m8_i64m1))) +vint64m1_t vlmul_trunc_i64m1(vint64m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i8m4_i8m2))) +vint8m2_t vlmul_trunc_i8m2(vint8m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i8m8_i8m2))) +vint8m2_t vlmul_trunc_i8m2(vint8m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i16m4_i16m2))) +vint16m2_t vlmul_trunc_i16m2(vint16m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i16m8_i16m2))) +vint16m2_t vlmul_trunc_i16m2(vint16m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i32m4_i32m2))) +vint32m2_t vlmul_trunc_i32m2(vint32m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i32m8_i32m2))) +vint32m2_t vlmul_trunc_i32m2(vint32m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i64m4_i64m2))) +vint64m2_t vlmul_trunc_i64m2(vint64m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i64m8_i64m2))) +vint64m2_t vlmul_trunc_i64m2(vint64m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i8m8_i8m4))) +vint8m4_t vlmul_trunc_i8m4(vint8m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i16m8_i16m4))) +vint16m4_t vlmul_trunc_i16m4(vint16m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i32m8_i32m4))) +vint32m4_t vlmul_trunc_i32m4(vint32m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_i64m8_i64m4))) +vint64m4_t vlmul_trunc_i64m4(vint64m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u8m1_u8mf2))) +vuint8mf2_t vlmul_trunc_u8mf2(vuint8m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u8m2_u8mf2))) +vuint8mf2_t vlmul_trunc_u8mf2(vuint8m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u8m4_u8mf2))) +vuint8mf2_t vlmul_trunc_u8mf2(vuint8m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u8m8_u8mf2))) +vuint8mf2_t vlmul_trunc_u8mf2(vuint8m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u16m1_u16mf2))) +vuint16mf2_t vlmul_trunc_u16mf2(vuint16m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u16m2_u16mf2))) +vuint16mf2_t vlmul_trunc_u16mf2(vuint16m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u16m4_u16mf2))) +vuint16mf2_t vlmul_trunc_u16mf2(vuint16m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u16m8_u16mf2))) +vuint16mf2_t vlmul_trunc_u16mf2(vuint16m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u32m1_u32mf2))) +vuint32mf2_t vlmul_trunc_u32mf2(vuint32m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u32m2_u32mf2))) +vuint32mf2_t vlmul_trunc_u32mf2(vuint32m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u32m4_u32mf2))) +vuint32mf2_t vlmul_trunc_u32mf2(vuint32m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u32m8_u32mf2))) +vuint32mf2_t vlmul_trunc_u32mf2(vuint32m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u8m1_u8mf4))) +vuint8mf4_t vlmul_trunc_u8mf4(vuint8m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u8m2_u8mf4))) +vuint8mf4_t vlmul_trunc_u8mf4(vuint8m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u8m4_u8mf4))) +vuint8mf4_t vlmul_trunc_u8mf4(vuint8m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u8m8_u8mf4))) +vuint8mf4_t vlmul_trunc_u8mf4(vuint8m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u8mf2_u8mf4))) +vuint8mf4_t vlmul_trunc_u8mf4(vuint8mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u16m1_u16mf4))) +vuint16mf4_t vlmul_trunc_u16mf4(vuint16m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u16m2_u16mf4))) +vuint16mf4_t vlmul_trunc_u16mf4(vuint16m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u16m4_u16mf4))) +vuint16mf4_t vlmul_trunc_u16mf4(vuint16m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u16m8_u16mf4))) +vuint16mf4_t vlmul_trunc_u16mf4(vuint16m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u16mf2_u16mf4))) +vuint16mf4_t vlmul_trunc_u16mf4(vuint16mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u8m1_u8mf8))) +vuint8mf8_t vlmul_trunc_u8mf8(vuint8m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u8m2_u8mf8))) +vuint8mf8_t vlmul_trunc_u8mf8(vuint8m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u8m4_u8mf8))) +vuint8mf8_t vlmul_trunc_u8mf8(vuint8m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u8m8_u8mf8))) +vuint8mf8_t vlmul_trunc_u8mf8(vuint8m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u8mf2_u8mf8))) +vuint8mf8_t vlmul_trunc_u8mf8(vuint8mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u8mf4_u8mf8))) +vuint8mf8_t vlmul_trunc_u8mf8(vuint8mf4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u8m2_u8m1))) +vuint8m1_t vlmul_trunc_u8m1(vuint8m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u8m4_u8m1))) +vuint8m1_t vlmul_trunc_u8m1(vuint8m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u8m8_u8m1))) +vuint8m1_t vlmul_trunc_u8m1(vuint8m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u16m2_u16m1))) +vuint16m1_t vlmul_trunc_u16m1(vuint16m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u16m4_u16m1))) +vuint16m1_t vlmul_trunc_u16m1(vuint16m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u16m8_u16m1))) +vuint16m1_t vlmul_trunc_u16m1(vuint16m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u32m2_u32m1))) +vuint32m1_t vlmul_trunc_u32m1(vuint32m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u32m4_u32m1))) +vuint32m1_t vlmul_trunc_u32m1(vuint32m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u32m8_u32m1))) +vuint32m1_t vlmul_trunc_u32m1(vuint32m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u64m2_u64m1))) +vuint64m1_t vlmul_trunc_u64m1(vuint64m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u64m4_u64m1))) +vuint64m1_t vlmul_trunc_u64m1(vuint64m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u64m8_u64m1))) +vuint64m1_t vlmul_trunc_u64m1(vuint64m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u8m4_u8m2))) +vuint8m2_t vlmul_trunc_u8m2(vuint8m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u8m8_u8m2))) +vuint8m2_t vlmul_trunc_u8m2(vuint8m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u16m4_u16m2))) +vuint16m2_t vlmul_trunc_u16m2(vuint16m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u16m8_u16m2))) +vuint16m2_t vlmul_trunc_u16m2(vuint16m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u32m4_u32m2))) +vuint32m2_t vlmul_trunc_u32m2(vuint32m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u32m8_u32m2))) +vuint32m2_t vlmul_trunc_u32m2(vuint32m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u64m4_u64m2))) +vuint64m2_t vlmul_trunc_u64m2(vuint64m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u64m8_u64m2))) +vuint64m2_t vlmul_trunc_u64m2(vuint64m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u8m8_u8m4))) +vuint8m4_t vlmul_trunc_u8m4(vuint8m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u16m8_u16m4))) +vuint16m4_t vlmul_trunc_u16m4(vuint16m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u32m8_u32m4))) +vuint32m4_t vlmul_trunc_u32m4(vuint32m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_u64m8_u64m4))) +vuint64m4_t vlmul_trunc_u64m4(vuint64m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmand_mm_b8))) +vbool8_t vmand(vbool8_t op0, vbool8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmand_mm_b4))) +vbool4_t vmand(vbool4_t op0, vbool4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmand_mm_b2))) +vbool2_t vmand(vbool2_t op0, vbool2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmand_mm_b1))) +vbool1_t vmand(vbool1_t op0, vbool1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmand_mm_b16))) +vbool16_t vmand(vbool16_t op0, vbool16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmand_mm_b32))) +vbool32_t vmand(vbool32_t op0, vbool32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmand_mm_b64))) +vbool64_t vmand(vbool64_t op0, vbool64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmandnot_mm_b8))) +vbool8_t vmandnot(vbool8_t op0, vbool8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmandnot_mm_b4))) +vbool4_t vmandnot(vbool4_t op0, vbool4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmandnot_mm_b2))) +vbool2_t vmandnot(vbool2_t op0, vbool2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmandnot_mm_b1))) +vbool1_t vmandnot(vbool1_t op0, vbool1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmandnot_mm_b16))) +vbool16_t vmandnot(vbool16_t op0, vbool16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmandnot_mm_b32))) +vbool32_t vmandnot(vbool32_t op0, vbool32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmandnot_mm_b64))) +vbool64_t vmandnot(vbool64_t op0, vbool64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmnand_mm_b8))) +vbool8_t vmnand(vbool8_t op0, vbool8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmnand_mm_b4))) +vbool4_t vmnand(vbool4_t op0, vbool4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmnand_mm_b2))) +vbool2_t vmnand(vbool2_t op0, vbool2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmnand_mm_b1))) +vbool1_t vmnand(vbool1_t op0, vbool1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmnand_mm_b16))) +vbool16_t vmnand(vbool16_t op0, vbool16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmnand_mm_b32))) +vbool32_t vmnand(vbool32_t op0, vbool32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmnand_mm_b64))) +vbool64_t vmnand(vbool64_t op0, vbool64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmnor_mm_b8))) +vbool8_t vmnor(vbool8_t op0, vbool8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmnor_mm_b4))) +vbool4_t vmnor(vbool4_t op0, vbool4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmnor_mm_b2))) +vbool2_t vmnor(vbool2_t op0, vbool2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmnor_mm_b1))) +vbool1_t vmnor(vbool1_t op0, vbool1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmnor_mm_b16))) +vbool16_t vmnor(vbool16_t op0, vbool16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmnor_mm_b32))) +vbool32_t vmnor(vbool32_t op0, vbool32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmnor_mm_b64))) +vbool64_t vmnor(vbool64_t op0, vbool64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmor_mm_b8))) +vbool8_t vmor(vbool8_t op0, vbool8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmor_mm_b4))) +vbool4_t vmor(vbool4_t op0, vbool4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmor_mm_b2))) +vbool2_t vmor(vbool2_t op0, vbool2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmor_mm_b1))) +vbool1_t vmor(vbool1_t op0, vbool1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmor_mm_b16))) +vbool16_t vmor(vbool16_t op0, vbool16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmor_mm_b32))) +vbool32_t vmor(vbool32_t op0, vbool32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmor_mm_b64))) +vbool64_t vmor(vbool64_t op0, vbool64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmornot_mm_b8))) +vbool8_t vmornot(vbool8_t op0, vbool8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmornot_mm_b4))) +vbool4_t vmornot(vbool4_t op0, vbool4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmornot_mm_b2))) +vbool2_t vmornot(vbool2_t op0, vbool2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmornot_mm_b1))) +vbool1_t vmornot(vbool1_t op0, vbool1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmornot_mm_b16))) +vbool16_t vmornot(vbool16_t op0, vbool16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmornot_mm_b32))) +vbool32_t vmornot(vbool32_t op0, vbool32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmornot_mm_b64))) +vbool64_t vmornot(vbool64_t op0, vbool64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbf_m_b8))) +vbool8_t vmsbf(vbool8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbf_m_b8_m))) +vbool8_t vmsbf(vbool8_t op0, vbool8_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbf_m_b4))) +vbool4_t vmsbf(vbool4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbf_m_b4_m))) +vbool4_t vmsbf(vbool4_t op0, vbool4_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbf_m_b2))) +vbool2_t vmsbf(vbool2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbf_m_b2_m))) +vbool2_t vmsbf(vbool2_t op0, vbool2_t op1, vbool2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbf_m_b1))) +vbool1_t vmsbf(vbool1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbf_m_b1_m))) +vbool1_t vmsbf(vbool1_t op0, vbool1_t op1, vbool1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbf_m_b16))) +vbool16_t vmsbf(vbool16_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbf_m_b16_m))) +vbool16_t vmsbf(vbool16_t op0, vbool16_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbf_m_b32))) +vbool32_t vmsbf(vbool32_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbf_m_b32_m))) +vbool32_t vmsbf(vbool32_t op0, vbool32_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbf_m_b64))) +vbool64_t vmsbf(vbool64_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsbf_m_b64_m))) +vbool64_t vmsbf(vbool64_t op0, vbool64_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsif_m_b8))) +vbool8_t vmsif(vbool8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsif_m_b8_m))) +vbool8_t vmsif(vbool8_t op0, vbool8_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsif_m_b4))) +vbool4_t vmsif(vbool4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsif_m_b4_m))) +vbool4_t vmsif(vbool4_t op0, vbool4_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsif_m_b2))) +vbool2_t vmsif(vbool2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsif_m_b2_m))) +vbool2_t vmsif(vbool2_t op0, vbool2_t op1, vbool2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsif_m_b1))) +vbool1_t vmsif(vbool1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsif_m_b1_m))) +vbool1_t vmsif(vbool1_t op0, vbool1_t op1, vbool1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsif_m_b16))) +vbool16_t vmsif(vbool16_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsif_m_b16_m))) +vbool16_t vmsif(vbool16_t op0, vbool16_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsif_m_b32))) +vbool32_t vmsif(vbool32_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsif_m_b32_m))) +vbool32_t vmsif(vbool32_t op0, vbool32_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsif_m_b64))) +vbool64_t vmsif(vbool64_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsif_m_b64_m))) +vbool64_t vmsif(vbool64_t op0, vbool64_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsof_m_b8))) +vbool8_t vmsof(vbool8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsof_m_b8_m))) +vbool8_t vmsof(vbool8_t op0, vbool8_t op1, vbool8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsof_m_b4))) +vbool4_t vmsof(vbool4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsof_m_b4_m))) +vbool4_t vmsof(vbool4_t op0, vbool4_t op1, vbool4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsof_m_b2))) +vbool2_t vmsof(vbool2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsof_m_b2_m))) +vbool2_t vmsof(vbool2_t op0, vbool2_t op1, vbool2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsof_m_b1))) +vbool1_t vmsof(vbool1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsof_m_b1_m))) +vbool1_t vmsof(vbool1_t op0, vbool1_t op1, vbool1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsof_m_b16))) +vbool16_t vmsof(vbool16_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsof_m_b16_m))) +vbool16_t vmsof(vbool16_t op0, vbool16_t op1, vbool16_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsof_m_b32))) +vbool32_t vmsof(vbool32_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsof_m_b32_m))) +vbool32_t vmsof(vbool32_t op0, vbool32_t op1, vbool32_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsof_m_b64))) +vbool64_t vmsof(vbool64_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmsof_m_b64_m))) +vbool64_t vmsof(vbool64_t op0, vbool64_t op1, vbool64_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmxnor_mm_b8))) +vbool8_t vmxnor(vbool8_t op0, vbool8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmxnor_mm_b4))) +vbool4_t vmxnor(vbool4_t op0, vbool4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmxnor_mm_b2))) +vbool2_t vmxnor(vbool2_t op0, vbool2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmxnor_mm_b1))) +vbool1_t vmxnor(vbool1_t op0, vbool1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmxnor_mm_b16))) +vbool16_t vmxnor(vbool16_t op0, vbool16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmxnor_mm_b32))) +vbool32_t vmxnor(vbool32_t op0, vbool32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmxnor_mm_b64))) +vbool64_t vmxnor(vbool64_t op0, vbool64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmxor_mm_b8))) +vbool8_t vmxor(vbool8_t op0, vbool8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmxor_mm_b4))) +vbool4_t vmxor(vbool4_t op0, vbool4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmxor_mm_b2))) +vbool2_t vmxor(vbool2_t op0, vbool2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmxor_mm_b1))) +vbool1_t vmxor(vbool1_t op0, vbool1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmxor_mm_b16))) +vbool16_t vmxor(vbool16_t op0, vbool16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmxor_mm_b32))) +vbool32_t vmxor(vbool32_t op0, vbool32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmxor_mm_b64))) +vbool64_t vmxor(vbool64_t op0, vbool64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vpopc_m_b8))) +unsigned long vpopc(vbool8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vpopc_m_b8_m))) +unsigned long vpopc(vbool8_t op0, vbool8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vpopc_m_b4))) +unsigned long vpopc(vbool4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vpopc_m_b4_m))) +unsigned long vpopc(vbool4_t op0, vbool4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vpopc_m_b2))) +unsigned long vpopc(vbool2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vpopc_m_b2_m))) +unsigned long vpopc(vbool2_t op0, vbool2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vpopc_m_b1))) +unsigned long vpopc(vbool1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vpopc_m_b1_m))) +unsigned long vpopc(vbool1_t op0, vbool1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vpopc_m_b16))) +unsigned long vpopc(vbool16_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vpopc_m_b16_m))) +unsigned long vpopc(vbool16_t op0, vbool16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vpopc_m_b32))) +unsigned long vpopc(vbool32_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vpopc_m_b32_m))) +unsigned long vpopc(vbool32_t op0, vbool32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vpopc_m_b64))) +unsigned long vpopc(vbool64_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vpopc_m_b64_m))) +unsigned long vpopc(vbool64_t op0, vbool64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i8m1_i16m1))) +vint16m1_t vreinterpret_i16m1(vint8m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i8m2_i16m2))) +vint16m2_t vreinterpret_i16m2(vint8m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i8m4_i16m4))) +vint16m4_t vreinterpret_i16m4(vint8m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i8m8_i16m8))) +vint16m8_t vreinterpret_i16m8(vint8m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i8mf2_i16mf2))) +vint16mf2_t vreinterpret_i16mf2(vint8mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i8mf4_i16mf4))) +vint16mf4_t vreinterpret_i16mf4(vint8mf4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i32m1_i16m1))) +vint16m1_t vreinterpret_i16m1(vint32m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i32m2_i16m2))) +vint16m2_t vreinterpret_i16m2(vint32m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i32m4_i16m4))) +vint16m4_t vreinterpret_i16m4(vint32m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i32m8_i16m8))) +vint16m8_t vreinterpret_i16m8(vint32m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i32mf2_i16mf2))) +vint16mf2_t vreinterpret_i16mf2(vint32mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i64m1_i16m1))) +vint16m1_t vreinterpret_i16m1(vint64m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i64m2_i16m2))) +vint16m2_t vreinterpret_i16m2(vint64m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i64m4_i16m4))) +vint16m4_t vreinterpret_i16m4(vint64m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i64m8_i16m8))) +vint16m8_t vreinterpret_i16m8(vint64m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i8m1_i32m1))) +vint32m1_t vreinterpret_i32m1(vint8m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i8m2_i32m2))) +vint32m2_t vreinterpret_i32m2(vint8m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i8m4_i32m4))) +vint32m4_t vreinterpret_i32m4(vint8m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i8m8_i32m8))) +vint32m8_t vreinterpret_i32m8(vint8m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i8mf2_i32mf2))) +vint32mf2_t vreinterpret_i32mf2(vint8mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i16m1_i32m1))) +vint32m1_t vreinterpret_i32m1(vint16m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i16m2_i32m2))) +vint32m2_t vreinterpret_i32m2(vint16m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i16m4_i32m4))) +vint32m4_t vreinterpret_i32m4(vint16m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i16m8_i32m8))) +vint32m8_t vreinterpret_i32m8(vint16m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i16mf2_i32mf2))) +vint32mf2_t vreinterpret_i32mf2(vint16mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i64m1_i32m1))) +vint32m1_t vreinterpret_i32m1(vint64m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i64m2_i32m2))) +vint32m2_t vreinterpret_i32m2(vint64m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i64m4_i32m4))) +vint32m4_t vreinterpret_i32m4(vint64m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i64m8_i32m8))) +vint32m8_t vreinterpret_i32m8(vint64m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i8m1_i64m1))) +vint64m1_t vreinterpret_i64m1(vint8m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i8m2_i64m2))) +vint64m2_t vreinterpret_i64m2(vint8m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i8m4_i64m4))) +vint64m4_t vreinterpret_i64m4(vint8m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i8m8_i64m8))) +vint64m8_t vreinterpret_i64m8(vint8m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i16m1_i64m1))) +vint64m1_t vreinterpret_i64m1(vint16m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i16m2_i64m2))) +vint64m2_t vreinterpret_i64m2(vint16m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i16m4_i64m4))) +vint64m4_t vreinterpret_i64m4(vint16m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i16m8_i64m8))) +vint64m8_t vreinterpret_i64m8(vint16m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i32m1_i64m1))) +vint64m1_t vreinterpret_i64m1(vint32m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i32m2_i64m2))) +vint64m2_t vreinterpret_i64m2(vint32m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i32m4_i64m4))) +vint64m4_t vreinterpret_i64m4(vint32m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i32m8_i64m8))) +vint64m8_t vreinterpret_i64m8(vint32m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i16m1_i8m1))) +vint8m1_t vreinterpret_i8m1(vint16m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i16m2_i8m2))) +vint8m2_t vreinterpret_i8m2(vint16m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i16m4_i8m4))) +vint8m4_t vreinterpret_i8m4(vint16m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i16m8_i8m8))) +vint8m8_t vreinterpret_i8m8(vint16m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i16mf2_i8mf2))) +vint8mf2_t vreinterpret_i8mf2(vint16mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i16mf4_i8mf4))) +vint8mf4_t vreinterpret_i8mf4(vint16mf4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i32m1_i8m1))) +vint8m1_t vreinterpret_i8m1(vint32m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i32m2_i8m2))) +vint8m2_t vreinterpret_i8m2(vint32m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i32m4_i8m4))) +vint8m4_t vreinterpret_i8m4(vint32m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i32m8_i8m8))) +vint8m8_t vreinterpret_i8m8(vint32m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i32mf2_i8mf2))) +vint8mf2_t vreinterpret_i8mf2(vint32mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i64m1_i8m1))) +vint8m1_t vreinterpret_i8m1(vint64m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i64m2_i8m2))) +vint8m2_t vreinterpret_i8m2(vint64m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i64m4_i8m4))) +vint8m4_t vreinterpret_i8m4(vint64m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i64m8_i8m8))) +vint8m8_t vreinterpret_i8m8(vint64m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u8m1_i8m1))) +vint8m1_t vreinterpret_i8m1(vuint8m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u8m2_i8m2))) +vint8m2_t vreinterpret_i8m2(vuint8m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u8m4_i8m4))) +vint8m4_t vreinterpret_i8m4(vuint8m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u8m8_i8m8))) +vint8m8_t vreinterpret_i8m8(vuint8m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u8mf2_i8mf2))) +vint8mf2_t vreinterpret_i8mf2(vuint8mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u8mf4_i8mf4))) +vint8mf4_t vreinterpret_i8mf4(vuint8mf4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u8mf8_i8mf8))) +vint8mf8_t vreinterpret_i8mf8(vuint8mf8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u16m1_i16m1))) +vint16m1_t vreinterpret_i16m1(vuint16m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u16m2_i16m2))) +vint16m2_t vreinterpret_i16m2(vuint16m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u16m4_i16m4))) +vint16m4_t vreinterpret_i16m4(vuint16m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u16m8_i16m8))) +vint16m8_t vreinterpret_i16m8(vuint16m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u16mf2_i16mf2))) +vint16mf2_t vreinterpret_i16mf2(vuint16mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u16mf4_i16mf4))) +vint16mf4_t vreinterpret_i16mf4(vuint16mf4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u32m1_i32m1))) +vint32m1_t vreinterpret_i32m1(vuint32m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u32m2_i32m2))) +vint32m2_t vreinterpret_i32m2(vuint32m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u32m4_i32m4))) +vint32m4_t vreinterpret_i32m4(vuint32m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u32m8_i32m8))) +vint32m8_t vreinterpret_i32m8(vuint32m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u32mf2_i32mf2))) +vint32mf2_t vreinterpret_i32mf2(vuint32mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u64m1_i64m1))) +vint64m1_t vreinterpret_i64m1(vuint64m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u64m2_i64m2))) +vint64m2_t vreinterpret_i64m2(vuint64m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u64m4_i64m4))) +vint64m4_t vreinterpret_i64m4(vuint64m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u64m8_i64m8))) +vint64m8_t vreinterpret_i64m8(vuint64m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u8m1_u16m1))) +vuint16m1_t vreinterpret_u16m1(vuint8m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u8m2_u16m2))) +vuint16m2_t vreinterpret_u16m2(vuint8m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u8m4_u16m4))) +vuint16m4_t vreinterpret_u16m4(vuint8m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u8m8_u16m8))) +vuint16m8_t vreinterpret_u16m8(vuint8m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u8mf2_u16mf2))) +vuint16mf2_t vreinterpret_u16mf2(vuint8mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u8mf4_u16mf4))) +vuint16mf4_t vreinterpret_u16mf4(vuint8mf4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u32m1_u16m1))) +vuint16m1_t vreinterpret_u16m1(vuint32m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u32m2_u16m2))) +vuint16m2_t vreinterpret_u16m2(vuint32m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u32m4_u16m4))) +vuint16m4_t vreinterpret_u16m4(vuint32m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u32m8_u16m8))) +vuint16m8_t vreinterpret_u16m8(vuint32m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u32mf2_u16mf2))) +vuint16mf2_t vreinterpret_u16mf2(vuint32mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u64m1_u16m1))) +vuint16m1_t vreinterpret_u16m1(vuint64m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u64m2_u16m2))) +vuint16m2_t vreinterpret_u16m2(vuint64m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u64m4_u16m4))) +vuint16m4_t vreinterpret_u16m4(vuint64m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u64m8_u16m8))) +vuint16m8_t vreinterpret_u16m8(vuint64m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u8m1_u32m1))) +vuint32m1_t vreinterpret_u32m1(vuint8m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u8m2_u32m2))) +vuint32m2_t vreinterpret_u32m2(vuint8m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u8m4_u32m4))) +vuint32m4_t vreinterpret_u32m4(vuint8m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u8m8_u32m8))) +vuint32m8_t vreinterpret_u32m8(vuint8m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u8mf2_u32mf2))) +vuint32mf2_t vreinterpret_u32mf2(vuint8mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u16m1_u32m1))) +vuint32m1_t vreinterpret_u32m1(vuint16m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u16m2_u32m2))) +vuint32m2_t vreinterpret_u32m2(vuint16m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u16m4_u32m4))) +vuint32m4_t vreinterpret_u32m4(vuint16m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u16m8_u32m8))) +vuint32m8_t vreinterpret_u32m8(vuint16m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u16mf2_u32mf2))) +vuint32mf2_t vreinterpret_u32mf2(vuint16mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u64m1_u32m1))) +vuint32m1_t vreinterpret_u32m1(vuint64m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u64m2_u32m2))) +vuint32m2_t vreinterpret_u32m2(vuint64m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u64m4_u32m4))) +vuint32m4_t vreinterpret_u32m4(vuint64m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u64m8_u32m8))) +vuint32m8_t vreinterpret_u32m8(vuint64m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u8m1_u64m1))) +vuint64m1_t vreinterpret_u64m1(vuint8m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u8m2_u64m2))) +vuint64m2_t vreinterpret_u64m2(vuint8m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u8m4_u64m4))) +vuint64m4_t vreinterpret_u64m4(vuint8m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u8m8_u64m8))) +vuint64m8_t vreinterpret_u64m8(vuint8m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u16m1_u64m1))) +vuint64m1_t vreinterpret_u64m1(vuint16m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u16m2_u64m2))) +vuint64m2_t vreinterpret_u64m2(vuint16m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u16m4_u64m4))) +vuint64m4_t vreinterpret_u64m4(vuint16m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u16m8_u64m8))) +vuint64m8_t vreinterpret_u64m8(vuint16m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u32m1_u64m1))) +vuint64m1_t vreinterpret_u64m1(vuint32m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u32m2_u64m2))) +vuint64m2_t vreinterpret_u64m2(vuint32m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u32m4_u64m4))) +vuint64m4_t vreinterpret_u64m4(vuint32m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u32m8_u64m8))) +vuint64m8_t vreinterpret_u64m8(vuint32m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u16m1_u8m1))) +vuint8m1_t vreinterpret_u8m1(vuint16m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u16m2_u8m2))) +vuint8m2_t vreinterpret_u8m2(vuint16m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u16m4_u8m4))) +vuint8m4_t vreinterpret_u8m4(vuint16m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u16m8_u8m8))) +vuint8m8_t vreinterpret_u8m8(vuint16m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u16mf2_u8mf2))) +vuint8mf2_t vreinterpret_u8mf2(vuint16mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u16mf4_u8mf4))) +vuint8mf4_t vreinterpret_u8mf4(vuint16mf4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u32m1_u8m1))) +vuint8m1_t vreinterpret_u8m1(vuint32m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u32m2_u8m2))) +vuint8m2_t vreinterpret_u8m2(vuint32m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u32m4_u8m4))) +vuint8m4_t vreinterpret_u8m4(vuint32m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u32m8_u8m8))) +vuint8m8_t vreinterpret_u8m8(vuint32m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u32mf2_u8mf2))) +vuint8mf2_t vreinterpret_u8mf2(vuint32mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u64m1_u8m1))) +vuint8m1_t vreinterpret_u8m1(vuint64m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u64m2_u8m2))) +vuint8m2_t vreinterpret_u8m2(vuint64m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u64m4_u8m4))) +vuint8m4_t vreinterpret_u8m4(vuint64m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u64m8_u8m8))) +vuint8m8_t vreinterpret_u8m8(vuint64m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i8m1_u8m1))) +vuint8m1_t vreinterpret_u8m1(vint8m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i8m2_u8m2))) +vuint8m2_t vreinterpret_u8m2(vint8m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i8m4_u8m4))) +vuint8m4_t vreinterpret_u8m4(vint8m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i8m8_u8m8))) +vuint8m8_t vreinterpret_u8m8(vint8m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i8mf2_u8mf2))) +vuint8mf2_t vreinterpret_u8mf2(vint8mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i8mf4_u8mf4))) +vuint8mf4_t vreinterpret_u8mf4(vint8mf4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i8mf8_u8mf8))) +vuint8mf8_t vreinterpret_u8mf8(vint8mf8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i16m1_u16m1))) +vuint16m1_t vreinterpret_u16m1(vint16m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i16m2_u16m2))) +vuint16m2_t vreinterpret_u16m2(vint16m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i16m4_u16m4))) +vuint16m4_t vreinterpret_u16m4(vint16m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i16m8_u16m8))) +vuint16m8_t vreinterpret_u16m8(vint16m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i16mf2_u16mf2))) +vuint16mf2_t vreinterpret_u16mf2(vint16mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i16mf4_u16mf4))) +vuint16mf4_t vreinterpret_u16mf4(vint16mf4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i32m1_u32m1))) +vuint32m1_t vreinterpret_u32m1(vint32m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i32m2_u32m2))) +vuint32m2_t vreinterpret_u32m2(vint32m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i32m4_u32m4))) +vuint32m4_t vreinterpret_u32m4(vint32m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i32m8_u32m8))) +vuint32m8_t vreinterpret_u32m8(vint32m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i32mf2_u32mf2))) +vuint32mf2_t vreinterpret_u32mf2(vint32mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i64m1_u64m1))) +vuint64m1_t vreinterpret_u64m1(vint64m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i64m2_u64m2))) +vuint64m2_t vreinterpret_u64m2(vint64m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i64m4_u64m4))) +vuint64m4_t vreinterpret_u64m4(vint64m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i64m8_u64m8))) +vuint64m8_t vreinterpret_u64m8(vint64m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse1_v_b8))) +void vse1(uint8_t * op0, vbool8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse1_v_b4))) +void vse1(uint8_t * op0, vbool4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse1_v_b2))) +void vse1(uint8_t * op0, vbool2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse1_v_b1))) +void vse1(uint8_t * op0, vbool1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse1_v_b16))) +void vse1(uint8_t * op0, vbool16_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse1_v_b32))) +void vse1(uint8_t * op0, vbool32_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse1_v_b64))) +void vse1(uint8_t * op0, vbool64_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf2_i16mf4))) +vint16mf4_t vsext_vf2(vint8mf8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf2_i16mf4_m))) +vint16mf4_t vsext_vf2(vbool64_t op0, vint16mf4_t op1, vint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf2_i16mf2))) +vint16mf2_t vsext_vf2(vint8mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf2_i16mf2_m))) +vint16mf2_t vsext_vf2(vbool32_t op0, vint16mf2_t op1, vint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf2_i16m1))) +vint16m1_t vsext_vf2(vint8mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf2_i16m1_m))) +vint16m1_t vsext_vf2(vbool16_t op0, vint16m1_t op1, vint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf2_i16m2))) +vint16m2_t vsext_vf2(vint8m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf2_i16m2_m))) +vint16m2_t vsext_vf2(vbool8_t op0, vint16m2_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf2_i16m4))) +vint16m4_t vsext_vf2(vint8m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf2_i16m4_m))) +vint16m4_t vsext_vf2(vbool4_t op0, vint16m4_t op1, vint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf2_i16m8))) +vint16m8_t vsext_vf2(vint8m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf2_i16m8_m))) +vint16m8_t vsext_vf2(vbool2_t op0, vint16m8_t op1, vint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf2_i32mf2))) +vint32mf2_t vsext_vf2(vint16mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf2_i32mf2_m))) +vint32mf2_t vsext_vf2(vbool64_t op0, vint32mf2_t op1, vint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf2_i32m1))) +vint32m1_t vsext_vf2(vint16mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf2_i32m1_m))) +vint32m1_t vsext_vf2(vbool32_t op0, vint32m1_t op1, vint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf2_i32m2))) +vint32m2_t vsext_vf2(vint16m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf2_i32m2_m))) +vint32m2_t vsext_vf2(vbool16_t op0, vint32m2_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf2_i32m4))) +vint32m4_t vsext_vf2(vint16m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf2_i32m4_m))) +vint32m4_t vsext_vf2(vbool8_t op0, vint32m4_t op1, vint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf2_i32m8))) +vint32m8_t vsext_vf2(vint16m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf2_i32m8_m))) +vint32m8_t vsext_vf2(vbool4_t op0, vint32m8_t op1, vint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf2_i64m1))) +vint64m1_t vsext_vf2(vint32mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf2_i64m1_m))) +vint64m1_t vsext_vf2(vbool64_t op0, vint64m1_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf2_i64m2))) +vint64m2_t vsext_vf2(vint32m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf2_i64m2_m))) +vint64m2_t vsext_vf2(vbool32_t op0, vint64m2_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf2_i64m4))) +vint64m4_t vsext_vf2(vint32m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf2_i64m4_m))) +vint64m4_t vsext_vf2(vbool16_t op0, vint64m4_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf2_i64m8))) +vint64m8_t vsext_vf2(vint32m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf2_i64m8_m))) +vint64m8_t vsext_vf2(vbool8_t op0, vint64m8_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf4_i32mf2))) +vint32mf2_t vsext_vf4(vint8mf8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf4_i32mf2_m))) +vint32mf2_t vsext_vf4(vbool64_t op0, vint32mf2_t op1, vint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf4_i32m1))) +vint32m1_t vsext_vf4(vint8mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf4_i32m1_m))) +vint32m1_t vsext_vf4(vbool32_t op0, vint32m1_t op1, vint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf4_i32m2))) +vint32m2_t vsext_vf4(vint8mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf4_i32m2_m))) +vint32m2_t vsext_vf4(vbool16_t op0, vint32m2_t op1, vint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf4_i32m4))) +vint32m4_t vsext_vf4(vint8m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf4_i32m4_m))) +vint32m4_t vsext_vf4(vbool8_t op0, vint32m4_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf4_i32m8))) +vint32m8_t vsext_vf4(vint8m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf4_i32m8_m))) +vint32m8_t vsext_vf4(vbool4_t op0, vint32m8_t op1, vint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf4_i64m1))) +vint64m1_t vsext_vf4(vint16mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf4_i64m1_m))) +vint64m1_t vsext_vf4(vbool64_t op0, vint64m1_t op1, vint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf4_i64m2))) +vint64m2_t vsext_vf4(vint16mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf4_i64m2_m))) +vint64m2_t vsext_vf4(vbool32_t op0, vint64m2_t op1, vint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf4_i64m4))) +vint64m4_t vsext_vf4(vint16m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf4_i64m4_m))) +vint64m4_t vsext_vf4(vbool16_t op0, vint64m4_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf4_i64m8))) +vint64m8_t vsext_vf4(vint16m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf4_i64m8_m))) +vint64m8_t vsext_vf4(vbool8_t op0, vint64m8_t op1, vint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf8_i64m1))) +vint64m1_t vsext_vf8(vint8mf8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf8_i64m1_m))) +vint64m1_t vsext_vf8(vbool64_t op0, vint64m1_t op1, vint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf8_i64m2))) +vint64m2_t vsext_vf8(vint8mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf8_i64m2_m))) +vint64m2_t vsext_vf8(vbool32_t op0, vint64m2_t op1, vint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf8_i64m4))) +vint64m4_t vsext_vf8(vint8mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf8_i64m4_m))) +vint64m4_t vsext_vf8(vbool16_t op0, vint64m4_t op1, vint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf8_i64m8))) +vint64m8_t vsext_vf8(vint8m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsext_vf8_i64m8_m))) +vint64m8_t vsext_vf8(vbool8_t op0, vint64m8_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf2_u16mf4))) +vuint16mf4_t vzext_vf2(vuint8mf8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf2_u16mf4_m))) +vuint16mf4_t vzext_vf2(vbool64_t op0, vuint16mf4_t op1, vuint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf2_u16mf2))) +vuint16mf2_t vzext_vf2(vuint8mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf2_u16mf2_m))) +vuint16mf2_t vzext_vf2(vbool32_t op0, vuint16mf2_t op1, vuint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf2_u16m1))) +vuint16m1_t vzext_vf2(vuint8mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf2_u16m1_m))) +vuint16m1_t vzext_vf2(vbool16_t op0, vuint16m1_t op1, vuint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf2_u16m2))) +vuint16m2_t vzext_vf2(vuint8m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf2_u16m2_m))) +vuint16m2_t vzext_vf2(vbool8_t op0, vuint16m2_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf2_u16m4))) +vuint16m4_t vzext_vf2(vuint8m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf2_u16m4_m))) +vuint16m4_t vzext_vf2(vbool4_t op0, vuint16m4_t op1, vuint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf2_u16m8))) +vuint16m8_t vzext_vf2(vuint8m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf2_u16m8_m))) +vuint16m8_t vzext_vf2(vbool2_t op0, vuint16m8_t op1, vuint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf2_u32mf2))) +vuint32mf2_t vzext_vf2(vuint16mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf2_u32mf2_m))) +vuint32mf2_t vzext_vf2(vbool64_t op0, vuint32mf2_t op1, vuint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf2_u32m1))) +vuint32m1_t vzext_vf2(vuint16mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf2_u32m1_m))) +vuint32m1_t vzext_vf2(vbool32_t op0, vuint32m1_t op1, vuint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf2_u32m2))) +vuint32m2_t vzext_vf2(vuint16m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf2_u32m2_m))) +vuint32m2_t vzext_vf2(vbool16_t op0, vuint32m2_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf2_u32m4))) +vuint32m4_t vzext_vf2(vuint16m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf2_u32m4_m))) +vuint32m4_t vzext_vf2(vbool8_t op0, vuint32m4_t op1, vuint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf2_u32m8))) +vuint32m8_t vzext_vf2(vuint16m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf2_u32m8_m))) +vuint32m8_t vzext_vf2(vbool4_t op0, vuint32m8_t op1, vuint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf2_u64m1))) +vuint64m1_t vzext_vf2(vuint32mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf2_u64m1_m))) +vuint64m1_t vzext_vf2(vbool64_t op0, vuint64m1_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf2_u64m2))) +vuint64m2_t vzext_vf2(vuint32m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf2_u64m2_m))) +vuint64m2_t vzext_vf2(vbool32_t op0, vuint64m2_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf2_u64m4))) +vuint64m4_t vzext_vf2(vuint32m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf2_u64m4_m))) +vuint64m4_t vzext_vf2(vbool16_t op0, vuint64m4_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf2_u64m8))) +vuint64m8_t vzext_vf2(vuint32m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf2_u64m8_m))) +vuint64m8_t vzext_vf2(vbool8_t op0, vuint64m8_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf4_u32mf2))) +vuint32mf2_t vzext_vf4(vuint8mf8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf4_u32mf2_m))) +vuint32mf2_t vzext_vf4(vbool64_t op0, vuint32mf2_t op1, vuint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf4_u32m1))) +vuint32m1_t vzext_vf4(vuint8mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf4_u32m1_m))) +vuint32m1_t vzext_vf4(vbool32_t op0, vuint32m1_t op1, vuint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf4_u32m2))) +vuint32m2_t vzext_vf4(vuint8mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf4_u32m2_m))) +vuint32m2_t vzext_vf4(vbool16_t op0, vuint32m2_t op1, vuint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf4_u32m4))) +vuint32m4_t vzext_vf4(vuint8m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf4_u32m4_m))) +vuint32m4_t vzext_vf4(vbool8_t op0, vuint32m4_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf4_u32m8))) +vuint32m8_t vzext_vf4(vuint8m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf4_u32m8_m))) +vuint32m8_t vzext_vf4(vbool4_t op0, vuint32m8_t op1, vuint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf4_u64m1))) +vuint64m1_t vzext_vf4(vuint16mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf4_u64m1_m))) +vuint64m1_t vzext_vf4(vbool64_t op0, vuint64m1_t op1, vuint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf4_u64m2))) +vuint64m2_t vzext_vf4(vuint16mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf4_u64m2_m))) +vuint64m2_t vzext_vf4(vbool32_t op0, vuint64m2_t op1, vuint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf4_u64m4))) +vuint64m4_t vzext_vf4(vuint16m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf4_u64m4_m))) +vuint64m4_t vzext_vf4(vbool16_t op0, vuint64m4_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf4_u64m8))) +vuint64m8_t vzext_vf4(vuint16m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf4_u64m8_m))) +vuint64m8_t vzext_vf4(vbool8_t op0, vuint64m8_t op1, vuint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf8_u64m1))) +vuint64m1_t vzext_vf8(vuint8mf8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf8_u64m1_m))) +vuint64m1_t vzext_vf8(vbool64_t op0, vuint64m1_t op1, vuint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf8_u64m2))) +vuint64m2_t vzext_vf8(vuint8mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf8_u64m2_m))) +vuint64m2_t vzext_vf8(vbool32_t op0, vuint64m2_t op1, vuint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf8_u64m4))) +vuint64m4_t vzext_vf8(vuint8mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf8_u64m4_m))) +vuint64m4_t vzext_vf8(vbool16_t op0, vuint64m4_t op1, vuint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf8_u64m8))) +vuint64m8_t vzext_vf8(vuint8m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vzext_vf8_u64m8_m))) +vuint64m8_t vzext_vf8(vbool8_t op0, vuint64m8_t op1, vuint8m1_t op2, size_t op3); + +#if defined(__riscv_f) +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse32_v_f32m1))) +void vsse32(float * op0, ptrdiff_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse32_v_f32m1_m))) +void vsse32(vbool32_t op0, float * op1, ptrdiff_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse32_v_f32m2))) +void vsse32(float * op0, ptrdiff_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse32_v_f32m2_m))) +void vsse32(vbool16_t op0, float * op1, ptrdiff_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse32_v_f32m4))) +void vsse32(float * op0, ptrdiff_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse32_v_f32m4_m))) +void vsse32(vbool8_t op0, float * op1, ptrdiff_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse32_v_f32m8))) +void vsse32(float * op0, ptrdiff_t op1, vfloat32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse32_v_f32m8_m))) +void vsse32(vbool4_t op0, float * op1, ptrdiff_t op2, vfloat32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse32_v_f32mf2))) +void vsse32(float * op0, ptrdiff_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse32_v_f32mf2_m))) +void vsse32(vbool64_t op0, float * op1, ptrdiff_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_f32m1))) +vfloat32m1_t vluxei8(const float * op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_f32m1_m))) +vfloat32m1_t vluxei8(vbool32_t op0, vfloat32m1_t op1, const float * op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_f32m2))) +vfloat32m2_t vluxei8(const float * op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_f32m2_m))) +vfloat32m2_t vluxei8(vbool16_t op0, vfloat32m2_t op1, const float * op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_f32m4))) +vfloat32m4_t vluxei8(const float * op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_f32m4_m))) +vfloat32m4_t vluxei8(vbool8_t op0, vfloat32m4_t op1, const float * op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_f32m8))) +vfloat32m8_t vluxei8(const float * op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_f32m8_m))) +vfloat32m8_t vluxei8(vbool4_t op0, vfloat32m8_t op1, const float * op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_f32mf2))) +vfloat32mf2_t vluxei8(const float * op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_f32mf2_m))) +vfloat32mf2_t vluxei8(vbool64_t op0, vfloat32mf2_t op1, const float * op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_f32m1))) +vfloat32m1_t vluxei16(const float * op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_f32m1_m))) +vfloat32m1_t vluxei16(vbool32_t op0, vfloat32m1_t op1, const float * op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_f32m2))) +vfloat32m2_t vluxei16(const float * op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_f32m2_m))) +vfloat32m2_t vluxei16(vbool16_t op0, vfloat32m2_t op1, const float * op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_f32m4))) +vfloat32m4_t vluxei16(const float * op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_f32m4_m))) +vfloat32m4_t vluxei16(vbool8_t op0, vfloat32m4_t op1, const float * op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_f32m8))) +vfloat32m8_t vluxei16(const float * op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_f32m8_m))) +vfloat32m8_t vluxei16(vbool4_t op0, vfloat32m8_t op1, const float * op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_f32mf2))) +vfloat32mf2_t vluxei16(const float * op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_f32mf2_m))) +vfloat32mf2_t vluxei16(vbool64_t op0, vfloat32mf2_t op1, const float * op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_f32m1))) +vfloat32m1_t vluxei32(const float * op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_f32m1_m))) +vfloat32m1_t vluxei32(vbool32_t op0, vfloat32m1_t op1, const float * op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_f32m2))) +vfloat32m2_t vluxei32(const float * op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_f32m2_m))) +vfloat32m2_t vluxei32(vbool16_t op0, vfloat32m2_t op1, const float * op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_f32m4))) +vfloat32m4_t vluxei32(const float * op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_f32m4_m))) +vfloat32m4_t vluxei32(vbool8_t op0, vfloat32m4_t op1, const float * op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_f32m8))) +vfloat32m8_t vluxei32(const float * op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_f32m8_m))) +vfloat32m8_t vluxei32(vbool4_t op0, vfloat32m8_t op1, const float * op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_f32mf2))) +vfloat32mf2_t vluxei32(const float * op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_f32mf2_m))) +vfloat32mf2_t vluxei32(vbool64_t op0, vfloat32mf2_t op1, const float * op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_f32m1))) +vfloat32m1_t vluxei64(const float * op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_f32m1_m))) +vfloat32m1_t vluxei64(vbool32_t op0, vfloat32m1_t op1, const float * op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_f32m2))) +vfloat32m2_t vluxei64(const float * op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_f32m2_m))) +vfloat32m2_t vluxei64(vbool16_t op0, vfloat32m2_t op1, const float * op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_f32m4))) +vfloat32m4_t vluxei64(const float * op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_f32m4_m))) +vfloat32m4_t vluxei64(vbool8_t op0, vfloat32m4_t op1, const float * op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_f32mf2))) +vfloat32mf2_t vluxei64(const float * op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_f32mf2_m))) +vfloat32mf2_t vluxei64(vbool64_t op0, vfloat32mf2_t op1, const float * op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_f32m1))) +vfloat32m1_t vloxei8(const float * op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_f32m1_m))) +vfloat32m1_t vloxei8(vbool32_t op0, vfloat32m1_t op1, const float * op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_f32m2))) +vfloat32m2_t vloxei8(const float * op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_f32m2_m))) +vfloat32m2_t vloxei8(vbool16_t op0, vfloat32m2_t op1, const float * op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_f32m4))) +vfloat32m4_t vloxei8(const float * op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_f32m4_m))) +vfloat32m4_t vloxei8(vbool8_t op0, vfloat32m4_t op1, const float * op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_f32m8))) +vfloat32m8_t vloxei8(const float * op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_f32m8_m))) +vfloat32m8_t vloxei8(vbool4_t op0, vfloat32m8_t op1, const float * op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_f32mf2))) +vfloat32mf2_t vloxei8(const float * op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_f32mf2_m))) +vfloat32mf2_t vloxei8(vbool64_t op0, vfloat32mf2_t op1, const float * op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_f32m1))) +vfloat32m1_t vloxei16(const float * op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_f32m1_m))) +vfloat32m1_t vloxei16(vbool32_t op0, vfloat32m1_t op1, const float * op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_f32m2))) +vfloat32m2_t vloxei16(const float * op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_f32m2_m))) +vfloat32m2_t vloxei16(vbool16_t op0, vfloat32m2_t op1, const float * op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_f32m4))) +vfloat32m4_t vloxei16(const float * op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_f32m4_m))) +vfloat32m4_t vloxei16(vbool8_t op0, vfloat32m4_t op1, const float * op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_f32m8))) +vfloat32m8_t vloxei16(const float * op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_f32m8_m))) +vfloat32m8_t vloxei16(vbool4_t op0, vfloat32m8_t op1, const float * op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_f32mf2))) +vfloat32mf2_t vloxei16(const float * op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_f32mf2_m))) +vfloat32mf2_t vloxei16(vbool64_t op0, vfloat32mf2_t op1, const float * op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_f32m1))) +vfloat32m1_t vloxei32(const float * op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_f32m1_m))) +vfloat32m1_t vloxei32(vbool32_t op0, vfloat32m1_t op1, const float * op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_f32m2))) +vfloat32m2_t vloxei32(const float * op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_f32m2_m))) +vfloat32m2_t vloxei32(vbool16_t op0, vfloat32m2_t op1, const float * op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_f32m4))) +vfloat32m4_t vloxei32(const float * op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_f32m4_m))) +vfloat32m4_t vloxei32(vbool8_t op0, vfloat32m4_t op1, const float * op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_f32m8))) +vfloat32m8_t vloxei32(const float * op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_f32m8_m))) +vfloat32m8_t vloxei32(vbool4_t op0, vfloat32m8_t op1, const float * op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_f32mf2))) +vfloat32mf2_t vloxei32(const float * op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_f32mf2_m))) +vfloat32mf2_t vloxei32(vbool64_t op0, vfloat32mf2_t op1, const float * op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_f32m1))) +vfloat32m1_t vloxei64(const float * op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_f32m1_m))) +vfloat32m1_t vloxei64(vbool32_t op0, vfloat32m1_t op1, const float * op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_f32m2))) +vfloat32m2_t vloxei64(const float * op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_f32m2_m))) +vfloat32m2_t vloxei64(vbool16_t op0, vfloat32m2_t op1, const float * op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_f32m4))) +vfloat32m4_t vloxei64(const float * op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_f32m4_m))) +vfloat32m4_t vloxei64(vbool8_t op0, vfloat32m4_t op1, const float * op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_f32mf2))) +vfloat32mf2_t vloxei64(const float * op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_f32mf2_m))) +vfloat32mf2_t vloxei64(vbool64_t op0, vfloat32mf2_t op1, const float * op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_f32m1))) +vfloat32m1_t vmv_v(vfloat32m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_f32m2))) +vfloat32m2_t vmv_v(vfloat32m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_f32m4))) +vfloat32m4_t vmv_v(vfloat32m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_f32m8))) +vfloat32m8_t vmv_v(vfloat32m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_f32mf2))) +vfloat32mf2_t vmv_v(vfloat32mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vv_f32m1))) +vfloat32m1_t vfadd(vfloat32m1_t op0, vfloat32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vv_f32m1_m))) +vfloat32m1_t vfadd(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vv_f32m2))) +vfloat32m2_t vfadd(vfloat32m2_t op0, vfloat32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vv_f32m2_m))) +vfloat32m2_t vfadd(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vv_f32m4))) +vfloat32m4_t vfadd(vfloat32m4_t op0, vfloat32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vv_f32m4_m))) +vfloat32m4_t vfadd(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vv_f32m8))) +vfloat32m8_t vfadd(vfloat32m8_t op0, vfloat32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vv_f32m8_m))) +vfloat32m8_t vfadd(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, vfloat32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vv_f32mf2))) +vfloat32mf2_t vfadd(vfloat32mf2_t op0, vfloat32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vv_f32mf2_m))) +vfloat32mf2_t vfadd(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vf_f32m1))) +vfloat32m1_t vfadd(vfloat32m1_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vf_f32m1_m))) +vfloat32m1_t vfadd(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vf_f32m2))) +vfloat32m2_t vfadd(vfloat32m2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vf_f32m2_m))) +vfloat32m2_t vfadd(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vf_f32m4))) +vfloat32m4_t vfadd(vfloat32m4_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vf_f32m4_m))) +vfloat32m4_t vfadd(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vf_f32m8))) +vfloat32m8_t vfadd(vfloat32m8_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vf_f32m8_m))) +vfloat32m8_t vfadd(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vf_f32mf2))) +vfloat32mf2_t vfadd(vfloat32mf2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vf_f32mf2_m))) +vfloat32mf2_t vfadd(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vv_f32m1))) +vfloat32m1_t vfsub(vfloat32m1_t op0, vfloat32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vv_f32m1_m))) +vfloat32m1_t vfsub(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vv_f32m2))) +vfloat32m2_t vfsub(vfloat32m2_t op0, vfloat32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vv_f32m2_m))) +vfloat32m2_t vfsub(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vv_f32m4))) +vfloat32m4_t vfsub(vfloat32m4_t op0, vfloat32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vv_f32m4_m))) +vfloat32m4_t vfsub(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vv_f32m8))) +vfloat32m8_t vfsub(vfloat32m8_t op0, vfloat32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vv_f32m8_m))) +vfloat32m8_t vfsub(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, vfloat32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vv_f32mf2))) +vfloat32mf2_t vfsub(vfloat32mf2_t op0, vfloat32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vv_f32mf2_m))) +vfloat32mf2_t vfsub(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vf_f32m1))) +vfloat32m1_t vfsub(vfloat32m1_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vf_f32m1_m))) +vfloat32m1_t vfsub(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vf_f32m2))) +vfloat32m2_t vfsub(vfloat32m2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vf_f32m2_m))) +vfloat32m2_t vfsub(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vf_f32m4))) +vfloat32m4_t vfsub(vfloat32m4_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vf_f32m4_m))) +vfloat32m4_t vfsub(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vf_f32m8))) +vfloat32m8_t vfsub(vfloat32m8_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vf_f32m8_m))) +vfloat32m8_t vfsub(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vf_f32mf2))) +vfloat32mf2_t vfsub(vfloat32mf2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vf_f32mf2_m))) +vfloat32mf2_t vfsub(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsub_vf_f32m1))) +vfloat32m1_t vfrsub(vfloat32m1_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsub_vf_f32m1_m))) +vfloat32m1_t vfrsub(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsub_vf_f32m2))) +vfloat32m2_t vfrsub(vfloat32m2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsub_vf_f32m2_m))) +vfloat32m2_t vfrsub(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsub_vf_f32m4))) +vfloat32m4_t vfrsub(vfloat32m4_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsub_vf_f32m4_m))) +vfloat32m4_t vfrsub(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsub_vf_f32m8))) +vfloat32m8_t vfrsub(vfloat32m8_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsub_vf_f32m8_m))) +vfloat32m8_t vfrsub(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsub_vf_f32mf2))) +vfloat32mf2_t vfrsub(vfloat32mf2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsub_vf_f32mf2_m))) +vfloat32mf2_t vfrsub(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vv_f32m1))) +vfloat32m1_t vfmul(vfloat32m1_t op0, vfloat32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vv_f32m1_m))) +vfloat32m1_t vfmul(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vv_f32m2))) +vfloat32m2_t vfmul(vfloat32m2_t op0, vfloat32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vv_f32m2_m))) +vfloat32m2_t vfmul(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vv_f32m4))) +vfloat32m4_t vfmul(vfloat32m4_t op0, vfloat32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vv_f32m4_m))) +vfloat32m4_t vfmul(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vv_f32m8))) +vfloat32m8_t vfmul(vfloat32m8_t op0, vfloat32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vv_f32m8_m))) +vfloat32m8_t vfmul(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, vfloat32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vv_f32mf2))) +vfloat32mf2_t vfmul(vfloat32mf2_t op0, vfloat32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vv_f32mf2_m))) +vfloat32mf2_t vfmul(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vf_f32m1))) +vfloat32m1_t vfmul(vfloat32m1_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vf_f32m1_m))) +vfloat32m1_t vfmul(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vf_f32m2))) +vfloat32m2_t vfmul(vfloat32m2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vf_f32m2_m))) +vfloat32m2_t vfmul(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vf_f32m4))) +vfloat32m4_t vfmul(vfloat32m4_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vf_f32m4_m))) +vfloat32m4_t vfmul(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vf_f32m8))) +vfloat32m8_t vfmul(vfloat32m8_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vf_f32m8_m))) +vfloat32m8_t vfmul(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vf_f32mf2))) +vfloat32mf2_t vfmul(vfloat32mf2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vf_f32mf2_m))) +vfloat32mf2_t vfmul(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vv_f32m1))) +vfloat32m1_t vfdiv(vfloat32m1_t op0, vfloat32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vv_f32m1_m))) +vfloat32m1_t vfdiv(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vv_f32m2))) +vfloat32m2_t vfdiv(vfloat32m2_t op0, vfloat32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vv_f32m2_m))) +vfloat32m2_t vfdiv(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vv_f32m4))) +vfloat32m4_t vfdiv(vfloat32m4_t op0, vfloat32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vv_f32m4_m))) +vfloat32m4_t vfdiv(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vv_f32m8))) +vfloat32m8_t vfdiv(vfloat32m8_t op0, vfloat32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vv_f32m8_m))) +vfloat32m8_t vfdiv(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, vfloat32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vv_f32mf2))) +vfloat32mf2_t vfdiv(vfloat32mf2_t op0, vfloat32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vv_f32mf2_m))) +vfloat32mf2_t vfdiv(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vf_f32m1))) +vfloat32m1_t vfdiv(vfloat32m1_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vf_f32m1_m))) +vfloat32m1_t vfdiv(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vf_f32m2))) +vfloat32m2_t vfdiv(vfloat32m2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vf_f32m2_m))) +vfloat32m2_t vfdiv(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vf_f32m4))) +vfloat32m4_t vfdiv(vfloat32m4_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vf_f32m4_m))) +vfloat32m4_t vfdiv(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vf_f32m8))) +vfloat32m8_t vfdiv(vfloat32m8_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vf_f32m8_m))) +vfloat32m8_t vfdiv(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vf_f32mf2))) +vfloat32mf2_t vfdiv(vfloat32mf2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vf_f32mf2_m))) +vfloat32mf2_t vfdiv(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrdiv_vf_f32m1))) +vfloat32m1_t vfrdiv(vfloat32m1_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrdiv_vf_f32m1_m))) +vfloat32m1_t vfrdiv(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrdiv_vf_f32m2))) +vfloat32m2_t vfrdiv(vfloat32m2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrdiv_vf_f32m2_m))) +vfloat32m2_t vfrdiv(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrdiv_vf_f32m4))) +vfloat32m4_t vfrdiv(vfloat32m4_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrdiv_vf_f32m4_m))) +vfloat32m4_t vfrdiv(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrdiv_vf_f32m8))) +vfloat32m8_t vfrdiv(vfloat32m8_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrdiv_vf_f32m8_m))) +vfloat32m8_t vfrdiv(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrdiv_vf_f32mf2))) +vfloat32mf2_t vfrdiv(vfloat32mf2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrdiv_vf_f32mf2_m))) +vfloat32mf2_t vfrdiv(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vv_f32m1))) +vfloat32m1_t vfmacc(vfloat32m1_t op0, vfloat32m1_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vv_f32m1_m))) +vfloat32m1_t vfmacc(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vv_f32m2))) +vfloat32m2_t vfmacc(vfloat32m2_t op0, vfloat32m2_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vv_f32m2_m))) +vfloat32m2_t vfmacc(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vv_f32m4))) +vfloat32m4_t vfmacc(vfloat32m4_t op0, vfloat32m4_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vv_f32m4_m))) +vfloat32m4_t vfmacc(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vv_f32m8))) +vfloat32m8_t vfmacc(vfloat32m8_t op0, vfloat32m8_t op1, vfloat32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vv_f32m8_m))) +vfloat32m8_t vfmacc(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, vfloat32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vv_f32mf2))) +vfloat32mf2_t vfmacc(vfloat32mf2_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vv_f32mf2_m))) +vfloat32mf2_t vfmacc(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vf_f32m1))) +vfloat32m1_t vfmacc(vfloat32m1_t op0, float op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vf_f32m1_m))) +vfloat32m1_t vfmacc(vbool32_t op0, vfloat32m1_t op1, float op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vf_f32m2))) +vfloat32m2_t vfmacc(vfloat32m2_t op0, float op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vf_f32m2_m))) +vfloat32m2_t vfmacc(vbool16_t op0, vfloat32m2_t op1, float op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vf_f32m4))) +vfloat32m4_t vfmacc(vfloat32m4_t op0, float op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vf_f32m4_m))) +vfloat32m4_t vfmacc(vbool8_t op0, vfloat32m4_t op1, float op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vf_f32m8))) +vfloat32m8_t vfmacc(vfloat32m8_t op0, float op1, vfloat32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vf_f32m8_m))) +vfloat32m8_t vfmacc(vbool4_t op0, vfloat32m8_t op1, float op2, vfloat32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vf_f32mf2))) +vfloat32mf2_t vfmacc(vfloat32mf2_t op0, float op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vf_f32mf2_m))) +vfloat32mf2_t vfmacc(vbool64_t op0, vfloat32mf2_t op1, float op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vv_f32m1))) +vfloat32m1_t vfnmacc(vfloat32m1_t op0, vfloat32m1_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vv_f32m1_m))) +vfloat32m1_t vfnmacc(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vv_f32m2))) +vfloat32m2_t vfnmacc(vfloat32m2_t op0, vfloat32m2_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vv_f32m2_m))) +vfloat32m2_t vfnmacc(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vv_f32m4))) +vfloat32m4_t vfnmacc(vfloat32m4_t op0, vfloat32m4_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vv_f32m4_m))) +vfloat32m4_t vfnmacc(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vv_f32m8))) +vfloat32m8_t vfnmacc(vfloat32m8_t op0, vfloat32m8_t op1, vfloat32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vv_f32m8_m))) +vfloat32m8_t vfnmacc(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, vfloat32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vv_f32mf2))) +vfloat32mf2_t vfnmacc(vfloat32mf2_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vv_f32mf2_m))) +vfloat32mf2_t vfnmacc(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vf_f32m1))) +vfloat32m1_t vfnmacc(vfloat32m1_t op0, float op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vf_f32m1_m))) +vfloat32m1_t vfnmacc(vbool32_t op0, vfloat32m1_t op1, float op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vf_f32m2))) +vfloat32m2_t vfnmacc(vfloat32m2_t op0, float op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vf_f32m2_m))) +vfloat32m2_t vfnmacc(vbool16_t op0, vfloat32m2_t op1, float op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vf_f32m4))) +vfloat32m4_t vfnmacc(vfloat32m4_t op0, float op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vf_f32m4_m))) +vfloat32m4_t vfnmacc(vbool8_t op0, vfloat32m4_t op1, float op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vf_f32m8))) +vfloat32m8_t vfnmacc(vfloat32m8_t op0, float op1, vfloat32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vf_f32m8_m))) +vfloat32m8_t vfnmacc(vbool4_t op0, vfloat32m8_t op1, float op2, vfloat32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vf_f32mf2))) +vfloat32mf2_t vfnmacc(vfloat32mf2_t op0, float op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vf_f32mf2_m))) +vfloat32mf2_t vfnmacc(vbool64_t op0, vfloat32mf2_t op1, float op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vv_f32m1))) +vfloat32m1_t vfmsac(vfloat32m1_t op0, vfloat32m1_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vv_f32m1_m))) +vfloat32m1_t vfmsac(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vv_f32m2))) +vfloat32m2_t vfmsac(vfloat32m2_t op0, vfloat32m2_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vv_f32m2_m))) +vfloat32m2_t vfmsac(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vv_f32m4))) +vfloat32m4_t vfmsac(vfloat32m4_t op0, vfloat32m4_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vv_f32m4_m))) +vfloat32m4_t vfmsac(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vv_f32m8))) +vfloat32m8_t vfmsac(vfloat32m8_t op0, vfloat32m8_t op1, vfloat32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vv_f32m8_m))) +vfloat32m8_t vfmsac(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, vfloat32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vv_f32mf2))) +vfloat32mf2_t vfmsac(vfloat32mf2_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vv_f32mf2_m))) +vfloat32mf2_t vfmsac(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vf_f32m1))) +vfloat32m1_t vfmsac(vfloat32m1_t op0, float op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vf_f32m1_m))) +vfloat32m1_t vfmsac(vbool32_t op0, vfloat32m1_t op1, float op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vf_f32m2))) +vfloat32m2_t vfmsac(vfloat32m2_t op0, float op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vf_f32m2_m))) +vfloat32m2_t vfmsac(vbool16_t op0, vfloat32m2_t op1, float op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vf_f32m4))) +vfloat32m4_t vfmsac(vfloat32m4_t op0, float op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vf_f32m4_m))) +vfloat32m4_t vfmsac(vbool8_t op0, vfloat32m4_t op1, float op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vf_f32m8))) +vfloat32m8_t vfmsac(vfloat32m8_t op0, float op1, vfloat32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vf_f32m8_m))) +vfloat32m8_t vfmsac(vbool4_t op0, vfloat32m8_t op1, float op2, vfloat32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vf_f32mf2))) +vfloat32mf2_t vfmsac(vfloat32mf2_t op0, float op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vf_f32mf2_m))) +vfloat32mf2_t vfmsac(vbool64_t op0, vfloat32mf2_t op1, float op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vv_f32m1))) +vfloat32m1_t vfnmsac(vfloat32m1_t op0, vfloat32m1_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vv_f32m1_m))) +vfloat32m1_t vfnmsac(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vv_f32m2))) +vfloat32m2_t vfnmsac(vfloat32m2_t op0, vfloat32m2_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vv_f32m2_m))) +vfloat32m2_t vfnmsac(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vv_f32m4))) +vfloat32m4_t vfnmsac(vfloat32m4_t op0, vfloat32m4_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vv_f32m4_m))) +vfloat32m4_t vfnmsac(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vv_f32m8))) +vfloat32m8_t vfnmsac(vfloat32m8_t op0, vfloat32m8_t op1, vfloat32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vv_f32m8_m))) +vfloat32m8_t vfnmsac(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, vfloat32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vv_f32mf2))) +vfloat32mf2_t vfnmsac(vfloat32mf2_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vv_f32mf2_m))) +vfloat32mf2_t vfnmsac(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vf_f32m1))) +vfloat32m1_t vfnmsac(vfloat32m1_t op0, float op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vf_f32m1_m))) +vfloat32m1_t vfnmsac(vbool32_t op0, vfloat32m1_t op1, float op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vf_f32m2))) +vfloat32m2_t vfnmsac(vfloat32m2_t op0, float op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vf_f32m2_m))) +vfloat32m2_t vfnmsac(vbool16_t op0, vfloat32m2_t op1, float op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vf_f32m4))) +vfloat32m4_t vfnmsac(vfloat32m4_t op0, float op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vf_f32m4_m))) +vfloat32m4_t vfnmsac(vbool8_t op0, vfloat32m4_t op1, float op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vf_f32m8))) +vfloat32m8_t vfnmsac(vfloat32m8_t op0, float op1, vfloat32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vf_f32m8_m))) +vfloat32m8_t vfnmsac(vbool4_t op0, vfloat32m8_t op1, float op2, vfloat32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vf_f32mf2))) +vfloat32mf2_t vfnmsac(vfloat32mf2_t op0, float op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vf_f32mf2_m))) +vfloat32mf2_t vfnmsac(vbool64_t op0, vfloat32mf2_t op1, float op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vv_f32m1))) +vfloat32m1_t vfmadd(vfloat32m1_t op0, vfloat32m1_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vv_f32m1_m))) +vfloat32m1_t vfmadd(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vv_f32m2))) +vfloat32m2_t vfmadd(vfloat32m2_t op0, vfloat32m2_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vv_f32m2_m))) +vfloat32m2_t vfmadd(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vv_f32m4))) +vfloat32m4_t vfmadd(vfloat32m4_t op0, vfloat32m4_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vv_f32m4_m))) +vfloat32m4_t vfmadd(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vv_f32m8))) +vfloat32m8_t vfmadd(vfloat32m8_t op0, vfloat32m8_t op1, vfloat32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vv_f32m8_m))) +vfloat32m8_t vfmadd(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, vfloat32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vv_f32mf2))) +vfloat32mf2_t vfmadd(vfloat32mf2_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vv_f32mf2_m))) +vfloat32mf2_t vfmadd(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vf_f32m1))) +vfloat32m1_t vfmadd(vfloat32m1_t op0, float op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vf_f32m1_m))) +vfloat32m1_t vfmadd(vbool32_t op0, vfloat32m1_t op1, float op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vf_f32m2))) +vfloat32m2_t vfmadd(vfloat32m2_t op0, float op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vf_f32m2_m))) +vfloat32m2_t vfmadd(vbool16_t op0, vfloat32m2_t op1, float op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vf_f32m4))) +vfloat32m4_t vfmadd(vfloat32m4_t op0, float op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vf_f32m4_m))) +vfloat32m4_t vfmadd(vbool8_t op0, vfloat32m4_t op1, float op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vf_f32m8))) +vfloat32m8_t vfmadd(vfloat32m8_t op0, float op1, vfloat32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vf_f32m8_m))) +vfloat32m8_t vfmadd(vbool4_t op0, vfloat32m8_t op1, float op2, vfloat32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vf_f32mf2))) +vfloat32mf2_t vfmadd(vfloat32mf2_t op0, float op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vf_f32mf2_m))) +vfloat32mf2_t vfmadd(vbool64_t op0, vfloat32mf2_t op1, float op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vv_f32m1))) +vfloat32m1_t vfnmadd(vfloat32m1_t op0, vfloat32m1_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vv_f32m1_m))) +vfloat32m1_t vfnmadd(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vv_f32m2))) +vfloat32m2_t vfnmadd(vfloat32m2_t op0, vfloat32m2_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vv_f32m2_m))) +vfloat32m2_t vfnmadd(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vv_f32m4))) +vfloat32m4_t vfnmadd(vfloat32m4_t op0, vfloat32m4_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vv_f32m4_m))) +vfloat32m4_t vfnmadd(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vv_f32m8))) +vfloat32m8_t vfnmadd(vfloat32m8_t op0, vfloat32m8_t op1, vfloat32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vv_f32m8_m))) +vfloat32m8_t vfnmadd(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, vfloat32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vv_f32mf2))) +vfloat32mf2_t vfnmadd(vfloat32mf2_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vv_f32mf2_m))) +vfloat32mf2_t vfnmadd(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vf_f32m1))) +vfloat32m1_t vfnmadd(vfloat32m1_t op0, float op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vf_f32m1_m))) +vfloat32m1_t vfnmadd(vbool32_t op0, vfloat32m1_t op1, float op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vf_f32m2))) +vfloat32m2_t vfnmadd(vfloat32m2_t op0, float op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vf_f32m2_m))) +vfloat32m2_t vfnmadd(vbool16_t op0, vfloat32m2_t op1, float op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vf_f32m4))) +vfloat32m4_t vfnmadd(vfloat32m4_t op0, float op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vf_f32m4_m))) +vfloat32m4_t vfnmadd(vbool8_t op0, vfloat32m4_t op1, float op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vf_f32m8))) +vfloat32m8_t vfnmadd(vfloat32m8_t op0, float op1, vfloat32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vf_f32m8_m))) +vfloat32m8_t vfnmadd(vbool4_t op0, vfloat32m8_t op1, float op2, vfloat32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vf_f32mf2))) +vfloat32mf2_t vfnmadd(vfloat32mf2_t op0, float op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vf_f32mf2_m))) +vfloat32mf2_t vfnmadd(vbool64_t op0, vfloat32mf2_t op1, float op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vv_f32m1))) +vfloat32m1_t vfmsub(vfloat32m1_t op0, vfloat32m1_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vv_f32m1_m))) +vfloat32m1_t vfmsub(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vv_f32m2))) +vfloat32m2_t vfmsub(vfloat32m2_t op0, vfloat32m2_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vv_f32m2_m))) +vfloat32m2_t vfmsub(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vv_f32m4))) +vfloat32m4_t vfmsub(vfloat32m4_t op0, vfloat32m4_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vv_f32m4_m))) +vfloat32m4_t vfmsub(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vv_f32m8))) +vfloat32m8_t vfmsub(vfloat32m8_t op0, vfloat32m8_t op1, vfloat32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vv_f32m8_m))) +vfloat32m8_t vfmsub(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, vfloat32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vv_f32mf2))) +vfloat32mf2_t vfmsub(vfloat32mf2_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vv_f32mf2_m))) +vfloat32mf2_t vfmsub(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vf_f32m1))) +vfloat32m1_t vfmsub(vfloat32m1_t op0, float op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vf_f32m1_m))) +vfloat32m1_t vfmsub(vbool32_t op0, vfloat32m1_t op1, float op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vf_f32m2))) +vfloat32m2_t vfmsub(vfloat32m2_t op0, float op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vf_f32m2_m))) +vfloat32m2_t vfmsub(vbool16_t op0, vfloat32m2_t op1, float op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vf_f32m4))) +vfloat32m4_t vfmsub(vfloat32m4_t op0, float op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vf_f32m4_m))) +vfloat32m4_t vfmsub(vbool8_t op0, vfloat32m4_t op1, float op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vf_f32m8))) +vfloat32m8_t vfmsub(vfloat32m8_t op0, float op1, vfloat32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vf_f32m8_m))) +vfloat32m8_t vfmsub(vbool4_t op0, vfloat32m8_t op1, float op2, vfloat32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vf_f32mf2))) +vfloat32mf2_t vfmsub(vfloat32mf2_t op0, float op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vf_f32mf2_m))) +vfloat32mf2_t vfmsub(vbool64_t op0, vfloat32mf2_t op1, float op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vv_f32m1))) +vfloat32m1_t vfnmsub(vfloat32m1_t op0, vfloat32m1_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vv_f32m1_m))) +vfloat32m1_t vfnmsub(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vv_f32m2))) +vfloat32m2_t vfnmsub(vfloat32m2_t op0, vfloat32m2_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vv_f32m2_m))) +vfloat32m2_t vfnmsub(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vv_f32m4))) +vfloat32m4_t vfnmsub(vfloat32m4_t op0, vfloat32m4_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vv_f32m4_m))) +vfloat32m4_t vfnmsub(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vv_f32m8))) +vfloat32m8_t vfnmsub(vfloat32m8_t op0, vfloat32m8_t op1, vfloat32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vv_f32m8_m))) +vfloat32m8_t vfnmsub(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, vfloat32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vv_f32mf2))) +vfloat32mf2_t vfnmsub(vfloat32mf2_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vv_f32mf2_m))) +vfloat32mf2_t vfnmsub(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vf_f32m1))) +vfloat32m1_t vfnmsub(vfloat32m1_t op0, float op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vf_f32m1_m))) +vfloat32m1_t vfnmsub(vbool32_t op0, vfloat32m1_t op1, float op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vf_f32m2))) +vfloat32m2_t vfnmsub(vfloat32m2_t op0, float op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vf_f32m2_m))) +vfloat32m2_t vfnmsub(vbool16_t op0, vfloat32m2_t op1, float op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vf_f32m4))) +vfloat32m4_t vfnmsub(vfloat32m4_t op0, float op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vf_f32m4_m))) +vfloat32m4_t vfnmsub(vbool8_t op0, vfloat32m4_t op1, float op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vf_f32m8))) +vfloat32m8_t vfnmsub(vfloat32m8_t op0, float op1, vfloat32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vf_f32m8_m))) +vfloat32m8_t vfnmsub(vbool4_t op0, vfloat32m8_t op1, float op2, vfloat32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vf_f32mf2))) +vfloat32mf2_t vfnmsub(vfloat32mf2_t op0, float op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vf_f32mf2_m))) +vfloat32mf2_t vfnmsub(vbool64_t op0, vfloat32mf2_t op1, float op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vv_f32m1))) +vfloat32m1_t vfmin(vfloat32m1_t op0, vfloat32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vv_f32m1_m))) +vfloat32m1_t vfmin(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vv_f32m2))) +vfloat32m2_t vfmin(vfloat32m2_t op0, vfloat32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vv_f32m2_m))) +vfloat32m2_t vfmin(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vv_f32m4))) +vfloat32m4_t vfmin(vfloat32m4_t op0, vfloat32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vv_f32m4_m))) +vfloat32m4_t vfmin(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vv_f32m8))) +vfloat32m8_t vfmin(vfloat32m8_t op0, vfloat32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vv_f32m8_m))) +vfloat32m8_t vfmin(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, vfloat32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vv_f32mf2))) +vfloat32mf2_t vfmin(vfloat32mf2_t op0, vfloat32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vv_f32mf2_m))) +vfloat32mf2_t vfmin(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vf_f32m1))) +vfloat32m1_t vfmin(vfloat32m1_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vf_f32m1_m))) +vfloat32m1_t vfmin(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vf_f32m2))) +vfloat32m2_t vfmin(vfloat32m2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vf_f32m2_m))) +vfloat32m2_t vfmin(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vf_f32m4))) +vfloat32m4_t vfmin(vfloat32m4_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vf_f32m4_m))) +vfloat32m4_t vfmin(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vf_f32m8))) +vfloat32m8_t vfmin(vfloat32m8_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vf_f32m8_m))) +vfloat32m8_t vfmin(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vf_f32mf2))) +vfloat32mf2_t vfmin(vfloat32mf2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vf_f32mf2_m))) +vfloat32mf2_t vfmin(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vv_f32m1))) +vfloat32m1_t vfmax(vfloat32m1_t op0, vfloat32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vv_f32m1_m))) +vfloat32m1_t vfmax(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vv_f32m2))) +vfloat32m2_t vfmax(vfloat32m2_t op0, vfloat32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vv_f32m2_m))) +vfloat32m2_t vfmax(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vv_f32m4))) +vfloat32m4_t vfmax(vfloat32m4_t op0, vfloat32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vv_f32m4_m))) +vfloat32m4_t vfmax(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vv_f32m8))) +vfloat32m8_t vfmax(vfloat32m8_t op0, vfloat32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vv_f32m8_m))) +vfloat32m8_t vfmax(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, vfloat32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vv_f32mf2))) +vfloat32mf2_t vfmax(vfloat32mf2_t op0, vfloat32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vv_f32mf2_m))) +vfloat32mf2_t vfmax(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vf_f32m1))) +vfloat32m1_t vfmax(vfloat32m1_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vf_f32m1_m))) +vfloat32m1_t vfmax(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vf_f32m2))) +vfloat32m2_t vfmax(vfloat32m2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vf_f32m2_m))) +vfloat32m2_t vfmax(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vf_f32m4))) +vfloat32m4_t vfmax(vfloat32m4_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vf_f32m4_m))) +vfloat32m4_t vfmax(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vf_f32m8))) +vfloat32m8_t vfmax(vfloat32m8_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vf_f32m8_m))) +vfloat32m8_t vfmax(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vf_f32mf2))) +vfloat32mf2_t vfmax(vfloat32mf2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vf_f32mf2_m))) +vfloat32mf2_t vfmax(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vv_f32m1))) +vfloat32m1_t vfsgnj(vfloat32m1_t op0, vfloat32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vv_f32m1_m))) +vfloat32m1_t vfsgnj(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vv_f32m2))) +vfloat32m2_t vfsgnj(vfloat32m2_t op0, vfloat32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vv_f32m2_m))) +vfloat32m2_t vfsgnj(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vv_f32m4))) +vfloat32m4_t vfsgnj(vfloat32m4_t op0, vfloat32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vv_f32m4_m))) +vfloat32m4_t vfsgnj(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vv_f32m8))) +vfloat32m8_t vfsgnj(vfloat32m8_t op0, vfloat32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vv_f32m8_m))) +vfloat32m8_t vfsgnj(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, vfloat32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vv_f32mf2))) +vfloat32mf2_t vfsgnj(vfloat32mf2_t op0, vfloat32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vv_f32mf2_m))) +vfloat32mf2_t vfsgnj(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vf_f32m1))) +vfloat32m1_t vfsgnj(vfloat32m1_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vf_f32m1_m))) +vfloat32m1_t vfsgnj(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vf_f32m2))) +vfloat32m2_t vfsgnj(vfloat32m2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vf_f32m2_m))) +vfloat32m2_t vfsgnj(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vf_f32m4))) +vfloat32m4_t vfsgnj(vfloat32m4_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vf_f32m4_m))) +vfloat32m4_t vfsgnj(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vf_f32m8))) +vfloat32m8_t vfsgnj(vfloat32m8_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vf_f32m8_m))) +vfloat32m8_t vfsgnj(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vf_f32mf2))) +vfloat32mf2_t vfsgnj(vfloat32mf2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vf_f32mf2_m))) +vfloat32mf2_t vfsgnj(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vv_f32m1))) +vfloat32m1_t vfsgnjn(vfloat32m1_t op0, vfloat32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vv_f32m1_m))) +vfloat32m1_t vfsgnjn(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vv_f32m2))) +vfloat32m2_t vfsgnjn(vfloat32m2_t op0, vfloat32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vv_f32m2_m))) +vfloat32m2_t vfsgnjn(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vv_f32m4))) +vfloat32m4_t vfsgnjn(vfloat32m4_t op0, vfloat32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vv_f32m4_m))) +vfloat32m4_t vfsgnjn(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vv_f32m8))) +vfloat32m8_t vfsgnjn(vfloat32m8_t op0, vfloat32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vv_f32m8_m))) +vfloat32m8_t vfsgnjn(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, vfloat32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vv_f32mf2))) +vfloat32mf2_t vfsgnjn(vfloat32mf2_t op0, vfloat32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vv_f32mf2_m))) +vfloat32mf2_t vfsgnjn(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vf_f32m1))) +vfloat32m1_t vfsgnjn(vfloat32m1_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vf_f32m1_m))) +vfloat32m1_t vfsgnjn(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vf_f32m2))) +vfloat32m2_t vfsgnjn(vfloat32m2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vf_f32m2_m))) +vfloat32m2_t vfsgnjn(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vf_f32m4))) +vfloat32m4_t vfsgnjn(vfloat32m4_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vf_f32m4_m))) +vfloat32m4_t vfsgnjn(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vf_f32m8))) +vfloat32m8_t vfsgnjn(vfloat32m8_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vf_f32m8_m))) +vfloat32m8_t vfsgnjn(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vf_f32mf2))) +vfloat32mf2_t vfsgnjn(vfloat32mf2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vf_f32mf2_m))) +vfloat32mf2_t vfsgnjn(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vv_f32m1))) +vfloat32m1_t vfsgnjx(vfloat32m1_t op0, vfloat32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vv_f32m1_m))) +vfloat32m1_t vfsgnjx(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vv_f32m2))) +vfloat32m2_t vfsgnjx(vfloat32m2_t op0, vfloat32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vv_f32m2_m))) +vfloat32m2_t vfsgnjx(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vv_f32m4))) +vfloat32m4_t vfsgnjx(vfloat32m4_t op0, vfloat32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vv_f32m4_m))) +vfloat32m4_t vfsgnjx(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vv_f32m8))) +vfloat32m8_t vfsgnjx(vfloat32m8_t op0, vfloat32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vv_f32m8_m))) +vfloat32m8_t vfsgnjx(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, vfloat32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vv_f32mf2))) +vfloat32mf2_t vfsgnjx(vfloat32mf2_t op0, vfloat32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vv_f32mf2_m))) +vfloat32mf2_t vfsgnjx(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vf_f32m1))) +vfloat32m1_t vfsgnjx(vfloat32m1_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vf_f32m1_m))) +vfloat32m1_t vfsgnjx(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vf_f32m2))) +vfloat32m2_t vfsgnjx(vfloat32m2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vf_f32m2_m))) +vfloat32m2_t vfsgnjx(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vf_f32m4))) +vfloat32m4_t vfsgnjx(vfloat32m4_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vf_f32m4_m))) +vfloat32m4_t vfsgnjx(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vf_f32m8))) +vfloat32m8_t vfsgnjx(vfloat32m8_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vf_f32m8_m))) +vfloat32m8_t vfsgnjx(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vf_f32mf2))) +vfloat32mf2_t vfsgnjx(vfloat32mf2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vf_f32mf2_m))) +vfloat32mf2_t vfsgnjx(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfabs_v_f32m1))) +vfloat32m1_t vfabs(vfloat32m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfabs_v_f32m1_m))) +vfloat32m1_t vfabs(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfabs_v_f32m2))) +vfloat32m2_t vfabs(vfloat32m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfabs_v_f32m2_m))) +vfloat32m2_t vfabs(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfabs_v_f32m4))) +vfloat32m4_t vfabs(vfloat32m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfabs_v_f32m4_m))) +vfloat32m4_t vfabs(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfabs_v_f32m8))) +vfloat32m8_t vfabs(vfloat32m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfabs_v_f32m8_m))) +vfloat32m8_t vfabs(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfabs_v_f32mf2))) +vfloat32mf2_t vfabs(vfloat32mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfabs_v_f32mf2_m))) +vfloat32mf2_t vfabs(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vv_f32m1_b32))) +vbool32_t vmfeq(vfloat32m1_t op0, vfloat32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vv_f32m1_b32_m))) +vbool32_t vmfeq(vbool32_t op0, vbool32_t op1, vfloat32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vv_f32m2_b16))) +vbool16_t vmfeq(vfloat32m2_t op0, vfloat32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vv_f32m2_b16_m))) +vbool16_t vmfeq(vbool16_t op0, vbool16_t op1, vfloat32m2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vv_f32m4_b8))) +vbool8_t vmfeq(vfloat32m4_t op0, vfloat32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vv_f32m4_b8_m))) +vbool8_t vmfeq(vbool8_t op0, vbool8_t op1, vfloat32m4_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vv_f32m8_b4))) +vbool4_t vmfeq(vfloat32m8_t op0, vfloat32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vv_f32m8_b4_m))) +vbool4_t vmfeq(vbool4_t op0, vbool4_t op1, vfloat32m8_t op2, vfloat32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vv_f32mf2_b64))) +vbool64_t vmfeq(vfloat32mf2_t op0, vfloat32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vv_f32mf2_b64_m))) +vbool64_t vmfeq(vbool64_t op0, vbool64_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vf_f32m1_b32))) +vbool32_t vmfeq(vfloat32m1_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vf_f32m1_b32_m))) +vbool32_t vmfeq(vbool32_t op0, vbool32_t op1, vfloat32m1_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vf_f32m2_b16))) +vbool16_t vmfeq(vfloat32m2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vf_f32m2_b16_m))) +vbool16_t vmfeq(vbool16_t op0, vbool16_t op1, vfloat32m2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vf_f32m4_b8))) +vbool8_t vmfeq(vfloat32m4_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vf_f32m4_b8_m))) +vbool8_t vmfeq(vbool8_t op0, vbool8_t op1, vfloat32m4_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vf_f32m8_b4))) +vbool4_t vmfeq(vfloat32m8_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vf_f32m8_b4_m))) +vbool4_t vmfeq(vbool4_t op0, vbool4_t op1, vfloat32m8_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vf_f32mf2_b64))) +vbool64_t vmfeq(vfloat32mf2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vf_f32mf2_b64_m))) +vbool64_t vmfeq(vbool64_t op0, vbool64_t op1, vfloat32mf2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vv_f32m1_b32))) +vbool32_t vmfne(vfloat32m1_t op0, vfloat32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vv_f32m1_b32_m))) +vbool32_t vmfne(vbool32_t op0, vbool32_t op1, vfloat32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vv_f32m2_b16))) +vbool16_t vmfne(vfloat32m2_t op0, vfloat32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vv_f32m2_b16_m))) +vbool16_t vmfne(vbool16_t op0, vbool16_t op1, vfloat32m2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vv_f32m4_b8))) +vbool8_t vmfne(vfloat32m4_t op0, vfloat32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vv_f32m4_b8_m))) +vbool8_t vmfne(vbool8_t op0, vbool8_t op1, vfloat32m4_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vv_f32m8_b4))) +vbool4_t vmfne(vfloat32m8_t op0, vfloat32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vv_f32m8_b4_m))) +vbool4_t vmfne(vbool4_t op0, vbool4_t op1, vfloat32m8_t op2, vfloat32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vv_f32mf2_b64))) +vbool64_t vmfne(vfloat32mf2_t op0, vfloat32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vv_f32mf2_b64_m))) +vbool64_t vmfne(vbool64_t op0, vbool64_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vf_f32m1_b32))) +vbool32_t vmfne(vfloat32m1_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vf_f32m1_b32_m))) +vbool32_t vmfne(vbool32_t op0, vbool32_t op1, vfloat32m1_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vf_f32m2_b16))) +vbool16_t vmfne(vfloat32m2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vf_f32m2_b16_m))) +vbool16_t vmfne(vbool16_t op0, vbool16_t op1, vfloat32m2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vf_f32m4_b8))) +vbool8_t vmfne(vfloat32m4_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vf_f32m4_b8_m))) +vbool8_t vmfne(vbool8_t op0, vbool8_t op1, vfloat32m4_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vf_f32m8_b4))) +vbool4_t vmfne(vfloat32m8_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vf_f32m8_b4_m))) +vbool4_t vmfne(vbool4_t op0, vbool4_t op1, vfloat32m8_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vf_f32mf2_b64))) +vbool64_t vmfne(vfloat32mf2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vf_f32mf2_b64_m))) +vbool64_t vmfne(vbool64_t op0, vbool64_t op1, vfloat32mf2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vv_f32m1_b32))) +vbool32_t vmflt(vfloat32m1_t op0, vfloat32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vv_f32m1_b32_m))) +vbool32_t vmflt(vbool32_t op0, vbool32_t op1, vfloat32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vv_f32m2_b16))) +vbool16_t vmflt(vfloat32m2_t op0, vfloat32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vv_f32m2_b16_m))) +vbool16_t vmflt(vbool16_t op0, vbool16_t op1, vfloat32m2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vv_f32m4_b8))) +vbool8_t vmflt(vfloat32m4_t op0, vfloat32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vv_f32m4_b8_m))) +vbool8_t vmflt(vbool8_t op0, vbool8_t op1, vfloat32m4_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vv_f32m8_b4))) +vbool4_t vmflt(vfloat32m8_t op0, vfloat32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vv_f32m8_b4_m))) +vbool4_t vmflt(vbool4_t op0, vbool4_t op1, vfloat32m8_t op2, vfloat32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vv_f32mf2_b64))) +vbool64_t vmflt(vfloat32mf2_t op0, vfloat32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vv_f32mf2_b64_m))) +vbool64_t vmflt(vbool64_t op0, vbool64_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vf_f32m1_b32))) +vbool32_t vmflt(vfloat32m1_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vf_f32m1_b32_m))) +vbool32_t vmflt(vbool32_t op0, vbool32_t op1, vfloat32m1_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vf_f32m2_b16))) +vbool16_t vmflt(vfloat32m2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vf_f32m2_b16_m))) +vbool16_t vmflt(vbool16_t op0, vbool16_t op1, vfloat32m2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vf_f32m4_b8))) +vbool8_t vmflt(vfloat32m4_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vf_f32m4_b8_m))) +vbool8_t vmflt(vbool8_t op0, vbool8_t op1, vfloat32m4_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vf_f32m8_b4))) +vbool4_t vmflt(vfloat32m8_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vf_f32m8_b4_m))) +vbool4_t vmflt(vbool4_t op0, vbool4_t op1, vfloat32m8_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vf_f32mf2_b64))) +vbool64_t vmflt(vfloat32mf2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vf_f32mf2_b64_m))) +vbool64_t vmflt(vbool64_t op0, vbool64_t op1, vfloat32mf2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vv_f32m1_b32))) +vbool32_t vmfle(vfloat32m1_t op0, vfloat32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vv_f32m1_b32_m))) +vbool32_t vmfle(vbool32_t op0, vbool32_t op1, vfloat32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vv_f32m2_b16))) +vbool16_t vmfle(vfloat32m2_t op0, vfloat32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vv_f32m2_b16_m))) +vbool16_t vmfle(vbool16_t op0, vbool16_t op1, vfloat32m2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vv_f32m4_b8))) +vbool8_t vmfle(vfloat32m4_t op0, vfloat32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vv_f32m4_b8_m))) +vbool8_t vmfle(vbool8_t op0, vbool8_t op1, vfloat32m4_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vv_f32m8_b4))) +vbool4_t vmfle(vfloat32m8_t op0, vfloat32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vv_f32m8_b4_m))) +vbool4_t vmfle(vbool4_t op0, vbool4_t op1, vfloat32m8_t op2, vfloat32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vv_f32mf2_b64))) +vbool64_t vmfle(vfloat32mf2_t op0, vfloat32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vv_f32mf2_b64_m))) +vbool64_t vmfle(vbool64_t op0, vbool64_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vf_f32m1_b32))) +vbool32_t vmfle(vfloat32m1_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vf_f32m1_b32_m))) +vbool32_t vmfle(vbool32_t op0, vbool32_t op1, vfloat32m1_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vf_f32m2_b16))) +vbool16_t vmfle(vfloat32m2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vf_f32m2_b16_m))) +vbool16_t vmfle(vbool16_t op0, vbool16_t op1, vfloat32m2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vf_f32m4_b8))) +vbool8_t vmfle(vfloat32m4_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vf_f32m4_b8_m))) +vbool8_t vmfle(vbool8_t op0, vbool8_t op1, vfloat32m4_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vf_f32m8_b4))) +vbool4_t vmfle(vfloat32m8_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vf_f32m8_b4_m))) +vbool4_t vmfle(vbool4_t op0, vbool4_t op1, vfloat32m8_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vf_f32mf2_b64))) +vbool64_t vmfle(vfloat32mf2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vf_f32mf2_b64_m))) +vbool64_t vmfle(vbool64_t op0, vbool64_t op1, vfloat32mf2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vv_f32m1_b32))) +vbool32_t vmfgt(vfloat32m1_t op0, vfloat32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vv_f32m1_b32_m))) +vbool32_t vmfgt(vbool32_t op0, vbool32_t op1, vfloat32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vv_f32m2_b16))) +vbool16_t vmfgt(vfloat32m2_t op0, vfloat32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vv_f32m2_b16_m))) +vbool16_t vmfgt(vbool16_t op0, vbool16_t op1, vfloat32m2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vv_f32m4_b8))) +vbool8_t vmfgt(vfloat32m4_t op0, vfloat32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vv_f32m4_b8_m))) +vbool8_t vmfgt(vbool8_t op0, vbool8_t op1, vfloat32m4_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vv_f32m8_b4))) +vbool4_t vmfgt(vfloat32m8_t op0, vfloat32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vv_f32m8_b4_m))) +vbool4_t vmfgt(vbool4_t op0, vbool4_t op1, vfloat32m8_t op2, vfloat32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vv_f32mf2_b64))) +vbool64_t vmfgt(vfloat32mf2_t op0, vfloat32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vv_f32mf2_b64_m))) +vbool64_t vmfgt(vbool64_t op0, vbool64_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vf_f32m1_b32))) +vbool32_t vmfgt(vfloat32m1_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vf_f32m1_b32_m))) +vbool32_t vmfgt(vbool32_t op0, vbool32_t op1, vfloat32m1_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vf_f32m2_b16))) +vbool16_t vmfgt(vfloat32m2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vf_f32m2_b16_m))) +vbool16_t vmfgt(vbool16_t op0, vbool16_t op1, vfloat32m2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vf_f32m4_b8))) +vbool8_t vmfgt(vfloat32m4_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vf_f32m4_b8_m))) +vbool8_t vmfgt(vbool8_t op0, vbool8_t op1, vfloat32m4_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vf_f32m8_b4))) +vbool4_t vmfgt(vfloat32m8_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vf_f32m8_b4_m))) +vbool4_t vmfgt(vbool4_t op0, vbool4_t op1, vfloat32m8_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vf_f32mf2_b64))) +vbool64_t vmfgt(vfloat32mf2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vf_f32mf2_b64_m))) +vbool64_t vmfgt(vbool64_t op0, vbool64_t op1, vfloat32mf2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vv_f32m1_b32))) +vbool32_t vmfge(vfloat32m1_t op0, vfloat32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vv_f32m1_b32_m))) +vbool32_t vmfge(vbool32_t op0, vbool32_t op1, vfloat32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vv_f32m2_b16))) +vbool16_t vmfge(vfloat32m2_t op0, vfloat32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vv_f32m2_b16_m))) +vbool16_t vmfge(vbool16_t op0, vbool16_t op1, vfloat32m2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vv_f32m4_b8))) +vbool8_t vmfge(vfloat32m4_t op0, vfloat32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vv_f32m4_b8_m))) +vbool8_t vmfge(vbool8_t op0, vbool8_t op1, vfloat32m4_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vv_f32m8_b4))) +vbool4_t vmfge(vfloat32m8_t op0, vfloat32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vv_f32m8_b4_m))) +vbool4_t vmfge(vbool4_t op0, vbool4_t op1, vfloat32m8_t op2, vfloat32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vv_f32mf2_b64))) +vbool64_t vmfge(vfloat32mf2_t op0, vfloat32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vv_f32mf2_b64_m))) +vbool64_t vmfge(vbool64_t op0, vbool64_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vf_f32m1_b32))) +vbool32_t vmfge(vfloat32m1_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vf_f32m1_b32_m))) +vbool32_t vmfge(vbool32_t op0, vbool32_t op1, vfloat32m1_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vf_f32m2_b16))) +vbool16_t vmfge(vfloat32m2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vf_f32m2_b16_m))) +vbool16_t vmfge(vbool16_t op0, vbool16_t op1, vfloat32m2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vf_f32m4_b8))) +vbool8_t vmfge(vfloat32m4_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vf_f32m4_b8_m))) +vbool8_t vmfge(vbool8_t op0, vbool8_t op1, vfloat32m4_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vf_f32m8_b4))) +vbool4_t vmfge(vfloat32m8_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vf_f32m8_b4_m))) +vbool4_t vmfge(vbool4_t op0, vbool4_t op1, vfloat32m8_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vf_f32mf2_b64))) +vbool64_t vmfge(vfloat32mf2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vf_f32mf2_b64_m))) +vbool64_t vmfge(vbool64_t op0, vbool64_t op1, vfloat32mf2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_f32m1))) +vfloat32m1_t vmerge(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_f32m2))) +vfloat32m2_t vmerge(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_f32m4))) +vfloat32m4_t vmerge(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_f32m8))) +vfloat32m8_t vmerge(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_f32mf2))) +vfloat32mf2_t vmerge(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmerge_vfm_f32m1))) +vfloat32m1_t vfmerge(vbool32_t op0, vfloat32m1_t op1, float op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmerge_vfm_f32m2))) +vfloat32m2_t vfmerge(vbool16_t op0, vfloat32m2_t op1, float op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmerge_vfm_f32m4))) +vfloat32m4_t vfmerge(vbool8_t op0, vfloat32m4_t op1, float op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmerge_vfm_f32m8))) +vfloat32m8_t vfmerge(vbool4_t op0, vfloat32m8_t op1, float op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmerge_vfm_f32mf2))) +vfloat32mf2_t vfmerge(vbool64_t op0, vfloat32mf2_t op1, float op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmax_vs_f32m1_f32m1))) +vfloat32m1_t vfredmax(vfloat32m1_t op0, vfloat32m1_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmax_vs_f32m1_f32m1_m))) +vfloat32m1_t vfredmax(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmax_vs_f32m2_f32m1))) +vfloat32m1_t vfredmax(vfloat32m1_t op0, vfloat32m2_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmax_vs_f32m2_f32m1_m))) +vfloat32m1_t vfredmax(vbool16_t op0, vfloat32m1_t op1, vfloat32m2_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmax_vs_f32m4_f32m1))) +vfloat32m1_t vfredmax(vfloat32m1_t op0, vfloat32m4_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmax_vs_f32m4_f32m1_m))) +vfloat32m1_t vfredmax(vbool8_t op0, vfloat32m1_t op1, vfloat32m4_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmax_vs_f32m8_f32m1))) +vfloat32m1_t vfredmax(vfloat32m1_t op0, vfloat32m8_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmax_vs_f32m8_f32m1_m))) +vfloat32m1_t vfredmax(vbool4_t op0, vfloat32m1_t op1, vfloat32m8_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmax_vs_f32mf2_f32m1))) +vfloat32m1_t vfredmax(vfloat32m1_t op0, vfloat32mf2_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmax_vs_f32mf2_f32m1_m))) +vfloat32m1_t vfredmax(vbool64_t op0, vfloat32m1_t op1, vfloat32mf2_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmin_vs_f32m1_f32m1))) +vfloat32m1_t vfredmin(vfloat32m1_t op0, vfloat32m1_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmin_vs_f32m1_f32m1_m))) +vfloat32m1_t vfredmin(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmin_vs_f32m2_f32m1))) +vfloat32m1_t vfredmin(vfloat32m1_t op0, vfloat32m2_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmin_vs_f32m2_f32m1_m))) +vfloat32m1_t vfredmin(vbool16_t op0, vfloat32m1_t op1, vfloat32m2_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmin_vs_f32m4_f32m1))) +vfloat32m1_t vfredmin(vfloat32m1_t op0, vfloat32m4_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmin_vs_f32m4_f32m1_m))) +vfloat32m1_t vfredmin(vbool8_t op0, vfloat32m1_t op1, vfloat32m4_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmin_vs_f32m8_f32m1))) +vfloat32m1_t vfredmin(vfloat32m1_t op0, vfloat32m8_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmin_vs_f32m8_f32m1_m))) +vfloat32m1_t vfredmin(vbool4_t op0, vfloat32m1_t op1, vfloat32m8_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmin_vs_f32mf2_f32m1))) +vfloat32m1_t vfredmin(vfloat32m1_t op0, vfloat32mf2_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmin_vs_f32mf2_f32m1_m))) +vfloat32m1_t vfredmin(vbool64_t op0, vfloat32m1_t op1, vfloat32mf2_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredsum_vs_f32m1_f32m1))) +vfloat32m1_t vfredsum(vfloat32m1_t op0, vfloat32m1_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredsum_vs_f32m1_f32m1_m))) +vfloat32m1_t vfredsum(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredsum_vs_f32m2_f32m1))) +vfloat32m1_t vfredsum(vfloat32m1_t op0, vfloat32m2_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredsum_vs_f32m2_f32m1_m))) +vfloat32m1_t vfredsum(vbool16_t op0, vfloat32m1_t op1, vfloat32m2_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredsum_vs_f32m4_f32m1))) +vfloat32m1_t vfredsum(vfloat32m1_t op0, vfloat32m4_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredsum_vs_f32m4_f32m1_m))) +vfloat32m1_t vfredsum(vbool8_t op0, vfloat32m1_t op1, vfloat32m4_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredsum_vs_f32m8_f32m1))) +vfloat32m1_t vfredsum(vfloat32m1_t op0, vfloat32m8_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredsum_vs_f32m8_f32m1_m))) +vfloat32m1_t vfredsum(vbool4_t op0, vfloat32m1_t op1, vfloat32m8_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredsum_vs_f32mf2_f32m1))) +vfloat32m1_t vfredsum(vfloat32m1_t op0, vfloat32mf2_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredsum_vs_f32mf2_f32m1_m))) +vfloat32m1_t vfredsum(vbool64_t op0, vfloat32m1_t op1, vfloat32mf2_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredosum_vs_f32m1_f32m1))) +vfloat32m1_t vfredosum(vfloat32m1_t op0, vfloat32m1_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredosum_vs_f32m1_f32m1_m))) +vfloat32m1_t vfredosum(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredosum_vs_f32m2_f32m1))) +vfloat32m1_t vfredosum(vfloat32m1_t op0, vfloat32m2_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredosum_vs_f32m2_f32m1_m))) +vfloat32m1_t vfredosum(vbool16_t op0, vfloat32m1_t op1, vfloat32m2_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredosum_vs_f32m4_f32m1))) +vfloat32m1_t vfredosum(vfloat32m1_t op0, vfloat32m4_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredosum_vs_f32m4_f32m1_m))) +vfloat32m1_t vfredosum(vbool8_t op0, vfloat32m1_t op1, vfloat32m4_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredosum_vs_f32m8_f32m1))) +vfloat32m1_t vfredosum(vfloat32m1_t op0, vfloat32m8_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredosum_vs_f32m8_f32m1_m))) +vfloat32m1_t vfredosum(vbool4_t op0, vfloat32m1_t op1, vfloat32m8_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredosum_vs_f32mf2_f32m1))) +vfloat32m1_t vfredosum(vfloat32m1_t op0, vfloat32mf2_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredosum_vs_f32mf2_f32m1_m))) +vfloat32m1_t vfredosum(vbool64_t op0, vfloat32m1_t op1, vfloat32mf2_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmv_f_s_f32m1_f32))) +float vfmv_f(vfloat32m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmv_f_s_f32m2_f32))) +float vfmv_f(vfloat32m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmv_f_s_f32m4_f32))) +float vfmv_f(vfloat32m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmv_f_s_f32m8_f32))) +float vfmv_f(vfloat32m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmv_f_s_f32mf2_f32))) +float vfmv_f(vfloat32mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmv_s_f_f32m1))) +vfloat32m1_t vfmv_s(vfloat32m1_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmv_s_f_f32m2))) +vfloat32m2_t vfmv_s(vfloat32m2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmv_s_f_f32m4))) +vfloat32m4_t vfmv_s(vfloat32m4_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmv_s_f_f32m8))) +vfloat32m8_t vfmv_s(vfloat32m8_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmv_s_f_f32mf2))) +vfloat32mf2_t vfmv_s(vfloat32mf2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_f32m1))) +vfloat32m1_t vslideup(vfloat32m1_t op0, vfloat32m1_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_f32m1_m))) +vfloat32m1_t vslideup(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_f32m2))) +vfloat32m2_t vslideup(vfloat32m2_t op0, vfloat32m2_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_f32m2_m))) +vfloat32m2_t vslideup(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_f32m4))) +vfloat32m4_t vslideup(vfloat32m4_t op0, vfloat32m4_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_f32m4_m))) +vfloat32m4_t vslideup(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_f32m8))) +vfloat32m8_t vslideup(vfloat32m8_t op0, vfloat32m8_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_f32m8_m))) +vfloat32m8_t vslideup(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_f32mf2))) +vfloat32mf2_t vslideup(vfloat32mf2_t op0, vfloat32mf2_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_f32mf2_m))) +vfloat32mf2_t vslideup(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_f32m1))) +vfloat32m1_t vslidedown(vfloat32m1_t op0, vfloat32m1_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_f32m1_m))) +vfloat32m1_t vslidedown(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_f32m2))) +vfloat32m2_t vslidedown(vfloat32m2_t op0, vfloat32m2_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_f32m2_m))) +vfloat32m2_t vslidedown(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_f32m4))) +vfloat32m4_t vslidedown(vfloat32m4_t op0, vfloat32m4_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_f32m4_m))) +vfloat32m4_t vslidedown(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_f32m8))) +vfloat32m8_t vslidedown(vfloat32m8_t op0, vfloat32m8_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_f32m8_m))) +vfloat32m8_t vslidedown(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_f32mf2))) +vfloat32mf2_t vslidedown(vfloat32mf2_t op0, vfloat32mf2_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_f32mf2_m))) +vfloat32mf2_t vslidedown(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1up_vf_f32m1))) +vfloat32m1_t vfslide1up(vfloat32m1_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1up_vf_f32m1_m))) +vfloat32m1_t vfslide1up(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1up_vf_f32m2))) +vfloat32m2_t vfslide1up(vfloat32m2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1up_vf_f32m2_m))) +vfloat32m2_t vfslide1up(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1up_vf_f32m4))) +vfloat32m4_t vfslide1up(vfloat32m4_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1up_vf_f32m4_m))) +vfloat32m4_t vfslide1up(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1up_vf_f32m8))) +vfloat32m8_t vfslide1up(vfloat32m8_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1up_vf_f32m8_m))) +vfloat32m8_t vfslide1up(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1up_vf_f32mf2))) +vfloat32mf2_t vfslide1up(vfloat32mf2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1up_vf_f32mf2_m))) +vfloat32mf2_t vfslide1up(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1down_vf_f32m1))) +vfloat32m1_t vfslide1down(vfloat32m1_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1down_vf_f32m1_m))) +vfloat32m1_t vfslide1down(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1down_vf_f32m2))) +vfloat32m2_t vfslide1down(vfloat32m2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1down_vf_f32m2_m))) +vfloat32m2_t vfslide1down(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1down_vf_f32m4))) +vfloat32m4_t vfslide1down(vfloat32m4_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1down_vf_f32m4_m))) +vfloat32m4_t vfslide1down(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1down_vf_f32m8))) +vfloat32m8_t vfslide1down(vfloat32m8_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1down_vf_f32m8_m))) +vfloat32m8_t vfslide1down(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1down_vf_f32mf2))) +vfloat32mf2_t vfslide1down(vfloat32mf2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1down_vf_f32mf2_m))) +vfloat32mf2_t vfslide1down(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_f32m1))) +vfloat32m1_t vrgather(vfloat32m1_t op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_f32m1_m))) +vfloat32m1_t vrgather(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_f32m2))) +vfloat32m2_t vrgather(vfloat32m2_t op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_f32m2_m))) +vfloat32m2_t vrgather(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_f32m4))) +vfloat32m4_t vrgather(vfloat32m4_t op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_f32m4_m))) +vfloat32m4_t vrgather(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_f32m8))) +vfloat32m8_t vrgather(vfloat32m8_t op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_f32m8_m))) +vfloat32m8_t vrgather(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_f32mf2))) +vfloat32mf2_t vrgather(vfloat32mf2_t op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_f32mf2_m))) +vfloat32mf2_t vrgather(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_f32m1))) +vfloat32m1_t vrgather(vfloat32m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_f32m1_m))) +vfloat32m1_t vrgather(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_f32m2))) +vfloat32m2_t vrgather(vfloat32m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_f32m2_m))) +vfloat32m2_t vrgather(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_f32m4))) +vfloat32m4_t vrgather(vfloat32m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_f32m4_m))) +vfloat32m4_t vrgather(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_f32m8))) +vfloat32m8_t vrgather(vfloat32m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_f32m8_m))) +vfloat32m8_t vrgather(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_f32mf2))) +vfloat32mf2_t vrgather(vfloat32mf2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_f32mf2_m))) +vfloat32mf2_t vrgather(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_f32m1))) +vfloat32m1_t vrgatherei16(vfloat32m1_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_f32m1_m))) +vfloat32m1_t vrgatherei16(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_f32m2))) +vfloat32m2_t vrgatherei16(vfloat32m2_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_f32m2_m))) +vfloat32m2_t vrgatherei16(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_f32m4))) +vfloat32m4_t vrgatherei16(vfloat32m4_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_f32m4_m))) +vfloat32m4_t vrgatherei16(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_f32m8))) +vfloat32m8_t vrgatherei16(vfloat32m8_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_f32m8_m))) +vfloat32m8_t vrgatherei16(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_f32mf2))) +vfloat32mf2_t vrgatherei16(vfloat32mf2_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_f32mf2_m))) +vfloat32mf2_t vrgatherei16(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_f32m1))) +vfloat32m1_t vcompress(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_f32m2))) +vfloat32m2_t vcompress(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_f32m4))) +vfloat32m4_t vcompress(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_f32m8))) +vfloat32m8_t vcompress(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_f32mf2))) +vfloat32mf2_t vcompress(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_f32m2_f32m1))) +vfloat32m1_t vget_f32m1(vfloat32m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_f32m4_f32m1))) +vfloat32m1_t vget_f32m1(vfloat32m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_f32m8_f32m1))) +vfloat32m1_t vget_f32m1(vfloat32m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_f32m4_f32m2))) +vfloat32m2_t vget_f32m2(vfloat32m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_f32m8_f32m2))) +vfloat32m2_t vget_f32m2(vfloat32m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_f32m8_f32m4))) +vfloat32m4_t vget_f32m4(vfloat32m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_f32m1_f32m2))) +vfloat32m2_t vset(vfloat32m2_t op0, size_t op1, vfloat32m1_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_f32m1_f32m4))) +vfloat32m4_t vset(vfloat32m4_t op0, size_t op1, vfloat32m1_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_f32m2_f32m4))) +vfloat32m4_t vset(vfloat32m4_t op0, size_t op1, vfloat32m2_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_f32m1))) +void vsuxei8(float * op0, vuint8mf4_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_f32m1_m))) +void vsuxei8(vbool32_t op0, float * op1, vuint8mf4_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_f32m2))) +void vsuxei8(float * op0, vuint8mf2_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_f32m2_m))) +void vsuxei8(vbool16_t op0, float * op1, vuint8mf2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_f32m4))) +void vsuxei8(float * op0, vuint8m1_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_f32m4_m))) +void vsuxei8(vbool8_t op0, float * op1, vuint8m1_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_f32m8))) +void vsuxei8(float * op0, vuint8m2_t op1, vfloat32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_f32m8_m))) +void vsuxei8(vbool4_t op0, float * op1, vuint8m2_t op2, vfloat32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_f32mf2))) +void vsuxei8(float * op0, vuint8mf8_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_f32mf2_m))) +void vsuxei8(vbool64_t op0, float * op1, vuint8mf8_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_f32m1_f32m8))) +vfloat32m8_t vset(vfloat32m8_t op0, size_t op1, vfloat32m1_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_f32m2_f32m8))) +vfloat32m8_t vset(vfloat32m8_t op0, size_t op1, vfloat32m2_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_f32m4_f32m8))) +vfloat32m8_t vset(vfloat32m8_t op0, size_t op1, vfloat32m4_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_f32m1))) +void vsuxei16(float * op0, vuint16mf2_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_f32m1_m))) +void vsuxei16(vbool32_t op0, float * op1, vuint16mf2_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_f32m2))) +void vsuxei16(float * op0, vuint16m1_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_f32m2_m))) +void vsuxei16(vbool16_t op0, float * op1, vuint16m1_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_f32m4))) +void vsuxei16(float * op0, vuint16m2_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_f32m4_m))) +void vsuxei16(vbool8_t op0, float * op1, vuint16m2_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_f32m8))) +void vsuxei16(float * op0, vuint16m4_t op1, vfloat32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_f32m8_m))) +void vsuxei16(vbool4_t op0, float * op1, vuint16m4_t op2, vfloat32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_f32mf2))) +void vsuxei16(float * op0, vuint16mf4_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_f32mf2_m))) +void vsuxei16(vbool64_t op0, float * op1, vuint16mf4_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_f32m1))) +void vsuxei32(float * op0, vuint32m1_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_f32m1_m))) +void vsuxei32(vbool32_t op0, float * op1, vuint32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_f32m2))) +void vsuxei32(float * op0, vuint32m2_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_f32m2_m))) +void vsuxei32(vbool16_t op0, float * op1, vuint32m2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_f32m4))) +void vsuxei32(float * op0, vuint32m4_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_f32m4_m))) +void vsuxei32(vbool8_t op0, float * op1, vuint32m4_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_f32m8))) +void vsuxei32(float * op0, vuint32m8_t op1, vfloat32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_f32m8_m))) +void vsuxei32(vbool4_t op0, float * op1, vuint32m8_t op2, vfloat32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_f32mf2))) +void vsuxei32(float * op0, vuint32mf2_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_f32mf2_m))) +void vsuxei32(vbool64_t op0, float * op1, vuint32mf2_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_f32m1))) +void vsuxei64(float * op0, vuint64m2_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_f32m1_m))) +void vsuxei64(vbool32_t op0, float * op1, vuint64m2_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_f32m2))) +void vsuxei64(float * op0, vuint64m4_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_f32m2_m))) +void vsuxei64(vbool16_t op0, float * op1, vuint64m4_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_f32m4))) +void vsuxei64(float * op0, vuint64m8_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_f32m4_m))) +void vsuxei64(vbool8_t op0, float * op1, vuint64m8_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_f32mf2))) +void vsuxei64(float * op0, vuint64m1_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_f32mf2_m))) +void vsuxei64(vbool64_t op0, float * op1, vuint64m1_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_f32m1))) +void vsoxei8(float * op0, vuint8mf4_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_f32m1_m))) +void vsoxei8(vbool32_t op0, float * op1, vuint8mf4_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_f32m2))) +void vsoxei8(float * op0, vuint8mf2_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_f32m2_m))) +void vsoxei8(vbool16_t op0, float * op1, vuint8mf2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_f32m4))) +void vsoxei8(float * op0, vuint8m1_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_f32m4_m))) +void vsoxei8(vbool8_t op0, float * op1, vuint8m1_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_f32m8))) +void vsoxei8(float * op0, vuint8m2_t op1, vfloat32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_f32m8_m))) +void vsoxei8(vbool4_t op0, float * op1, vuint8m2_t op2, vfloat32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_f32mf2))) +void vsoxei8(float * op0, vuint8mf8_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_f32mf2_m))) +void vsoxei8(vbool64_t op0, float * op1, vuint8mf8_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_f32m1))) +void vsoxei16(float * op0, vuint16mf2_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_f32m1_m))) +void vsoxei16(vbool32_t op0, float * op1, vuint16mf2_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_f32m2))) +void vsoxei16(float * op0, vuint16m1_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_f32m2_m))) +void vsoxei16(vbool16_t op0, float * op1, vuint16m1_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_f32m4))) +void vsoxei16(float * op0, vuint16m2_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_f32m4_m))) +void vsoxei16(vbool8_t op0, float * op1, vuint16m2_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_f32m8))) +void vsoxei16(float * op0, vuint16m4_t op1, vfloat32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_f32m8_m))) +void vsoxei16(vbool4_t op0, float * op1, vuint16m4_t op2, vfloat32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_f32mf2))) +void vsoxei16(float * op0, vuint16mf4_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_f32mf2_m))) +void vsoxei16(vbool64_t op0, float * op1, vuint16mf4_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_f32m1))) +void vsoxei32(float * op0, vuint32m1_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_f32m1_m))) +void vsoxei32(vbool32_t op0, float * op1, vuint32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_f32m2))) +void vsoxei32(float * op0, vuint32m2_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_f32m2_m))) +void vsoxei32(vbool16_t op0, float * op1, vuint32m2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_f32m4))) +void vsoxei32(float * op0, vuint32m4_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_f32m4_m))) +void vsoxei32(vbool8_t op0, float * op1, vuint32m4_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_f32m8))) +void vsoxei32(float * op0, vuint32m8_t op1, vfloat32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_f32m8_m))) +void vsoxei32(vbool4_t op0, float * op1, vuint32m8_t op2, vfloat32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_f32mf2))) +void vsoxei32(float * op0, vuint32mf2_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_f32mf2_m))) +void vsoxei32(vbool64_t op0, float * op1, vuint32mf2_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_f32m1))) +void vsoxei64(float * op0, vuint64m2_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_f32m1_m))) +void vsoxei64(vbool32_t op0, float * op1, vuint64m2_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_f32m2))) +void vsoxei64(float * op0, vuint64m4_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_f32m2_m))) +void vsoxei64(vbool16_t op0, float * op1, vuint64m4_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_f32m4))) +void vsoxei64(float * op0, vuint64m8_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_f32m4_m))) +void vsoxei64(vbool8_t op0, float * op1, vuint64m8_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_f32mf2))) +void vsoxei64(float * op0, vuint64m1_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_f32mf2_m))) +void vsoxei64(vbool64_t op0, float * op1, vuint64m1_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle32ff_v_f32m1_m))) +vfloat32m1_t vle32ff(vbool32_t op0, vfloat32m1_t op1, const float * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle32ff_v_f32m2_m))) +vfloat32m2_t vle32ff(vbool16_t op0, vfloat32m2_t op1, const float * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle32ff_v_f32m4_m))) +vfloat32m4_t vle32ff(vbool8_t op0, vfloat32m4_t op1, const float * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle32ff_v_f32m8_m))) +vfloat32m8_t vle32ff(vbool4_t op0, vfloat32m8_t op1, const float * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle32ff_v_f32mf2_m))) +vfloat32mf2_t vle32ff(vbool64_t op0, vfloat32mf2_t op1, const float * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfneg_v_f32m1))) +vfloat32m1_t vfneg(vfloat32m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfneg_v_f32m1_m))) +vfloat32m1_t vfneg(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfneg_v_f32m2))) +vfloat32m2_t vfneg(vfloat32m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfneg_v_f32m2_m))) +vfloat32m2_t vfneg(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfneg_v_f32m4))) +vfloat32m4_t vfneg(vfloat32m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfneg_v_f32m4_m))) +vfloat32m4_t vfneg(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfneg_v_f32m8))) +vfloat32m8_t vfneg(vfloat32m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfneg_v_f32m8_m))) +vfloat32m8_t vfneg(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfneg_v_f32mf2))) +vfloat32mf2_t vfneg(vfloat32mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfneg_v_f32mf2_m))) +vfloat32mf2_t vfneg(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle32_v_f32m1_m))) +vfloat32m1_t vle32(vbool32_t op0, vfloat32m1_t op1, const float * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle32_v_f32m2_m))) +vfloat32m2_t vle32(vbool16_t op0, vfloat32m2_t op1, const float * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle32_v_f32m4_m))) +vfloat32m4_t vle32(vbool8_t op0, vfloat32m4_t op1, const float * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle32_v_f32m8_m))) +vfloat32m8_t vle32(vbool4_t op0, vfloat32m8_t op1, const float * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle32_v_f32mf2_m))) +vfloat32mf2_t vle32(vbool64_t op0, vfloat32mf2_t op1, const float * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse32_v_f32m1))) +void vse32(float * op0, vfloat32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse32_v_f32m1_m))) +void vse32(vbool32_t op0, float * op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse32_v_f32m2))) +void vse32(float * op0, vfloat32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse32_v_f32m2_m))) +void vse32(vbool16_t op0, float * op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse32_v_f32m4))) +void vse32(float * op0, vfloat32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse32_v_f32m4_m))) +void vse32(vbool8_t op0, float * op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse32_v_f32m8))) +void vse32(float * op0, vfloat32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse32_v_f32m8_m))) +void vse32(vbool4_t op0, float * op1, vfloat32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse32_v_f32mf2))) +void vse32(float * op0, vfloat32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse32_v_f32mf2_m))) +void vse32(vbool64_t op0, float * op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse32_v_f32m1_m))) +vfloat32m1_t vlse32(vbool32_t op0, vfloat32m1_t op1, const float * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse32_v_f32m2_m))) +vfloat32m2_t vlse32(vbool16_t op0, vfloat32m2_t op1, const float * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse32_v_f32m4_m))) +vfloat32m4_t vlse32(vbool8_t op0, vfloat32m4_t op1, const float * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse32_v_f32m8_m))) +vfloat32m8_t vlse32(vbool4_t op0, vfloat32m8_t op1, const float * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse32_v_f32mf2_m))) +vfloat32mf2_t vlse32(vbool64_t op0, vfloat32mf2_t op1, const float * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfclass_v_u32m1))) +vuint32m1_t vfclass(vfloat32m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfclass_v_u32m1_m))) +vuint32m1_t vfclass(vbool32_t op0, vuint32m1_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfclass_v_u32m2))) +vuint32m2_t vfclass(vfloat32m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfclass_v_u32m2_m))) +vuint32m2_t vfclass(vbool16_t op0, vuint32m2_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfclass_v_u32m4))) +vuint32m4_t vfclass(vfloat32m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfclass_v_u32m4_m))) +vuint32m4_t vfclass(vbool8_t op0, vuint32m4_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfclass_v_u32m8))) +vuint32m8_t vfclass(vfloat32m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfclass_v_u32m8_m))) +vuint32m8_t vfclass(vbool4_t op0, vuint32m8_t op1, vfloat32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfclass_v_u32mf2))) +vuint32mf2_t vfclass(vfloat32mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfclass_v_u32mf2_m))) +vuint32mf2_t vfclass(vbool64_t op0, vuint32mf2_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_x_v_f32m1))) +vfloat32m1_t vfcvt_f(vint32m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_x_v_f32m1_m))) +vfloat32m1_t vfcvt_f(vbool32_t op0, vfloat32m1_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_x_v_f32m2))) +vfloat32m2_t vfcvt_f(vint32m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_x_v_f32m2_m))) +vfloat32m2_t vfcvt_f(vbool16_t op0, vfloat32m2_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_x_v_f32m4))) +vfloat32m4_t vfcvt_f(vint32m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_x_v_f32m4_m))) +vfloat32m4_t vfcvt_f(vbool8_t op0, vfloat32m4_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_x_v_f32m8))) +vfloat32m8_t vfcvt_f(vint32m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_x_v_f32m8_m))) +vfloat32m8_t vfcvt_f(vbool4_t op0, vfloat32m8_t op1, vint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_x_v_f32mf2))) +vfloat32mf2_t vfcvt_f(vint32mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_x_v_f32mf2_m))) +vfloat32mf2_t vfcvt_f(vbool64_t op0, vfloat32mf2_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_xu_v_f32m1))) +vfloat32m1_t vfcvt_f(vuint32m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_xu_v_f32m1_m))) +vfloat32m1_t vfcvt_f(vbool32_t op0, vfloat32m1_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_xu_v_f32m2))) +vfloat32m2_t vfcvt_f(vuint32m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_xu_v_f32m2_m))) +vfloat32m2_t vfcvt_f(vbool16_t op0, vfloat32m2_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_xu_v_f32m4))) +vfloat32m4_t vfcvt_f(vuint32m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_xu_v_f32m4_m))) +vfloat32m4_t vfcvt_f(vbool8_t op0, vfloat32m4_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_xu_v_f32m8))) +vfloat32m8_t vfcvt_f(vuint32m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_xu_v_f32m8_m))) +vfloat32m8_t vfcvt_f(vbool4_t op0, vfloat32m8_t op1, vuint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_xu_v_f32mf2))) +vfloat32mf2_t vfcvt_f(vuint32mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_xu_v_f32mf2_m))) +vfloat32mf2_t vfcvt_f(vbool64_t op0, vfloat32mf2_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_x_f_v_i32m1))) +vint32m1_t vfcvt_rtz_x(vfloat32m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_x_f_v_i32m1_m))) +vint32m1_t vfcvt_rtz_x(vbool32_t op0, vint32m1_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_x_f_v_i32m2))) +vint32m2_t vfcvt_rtz_x(vfloat32m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_x_f_v_i32m2_m))) +vint32m2_t vfcvt_rtz_x(vbool16_t op0, vint32m2_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_x_f_v_i32m4))) +vint32m4_t vfcvt_rtz_x(vfloat32m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_x_f_v_i32m4_m))) +vint32m4_t vfcvt_rtz_x(vbool8_t op0, vint32m4_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_x_f_v_i32m8))) +vint32m8_t vfcvt_rtz_x(vfloat32m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_x_f_v_i32m8_m))) +vint32m8_t vfcvt_rtz_x(vbool4_t op0, vint32m8_t op1, vfloat32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_x_f_v_i32mf2))) +vint32mf2_t vfcvt_rtz_x(vfloat32mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_x_f_v_i32mf2_m))) +vint32mf2_t vfcvt_rtz_x(vbool64_t op0, vint32mf2_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_xu_f_v_u32m1))) +vuint32m1_t vfcvt_rtz_xu(vfloat32m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_xu_f_v_u32m1_m))) +vuint32m1_t vfcvt_rtz_xu(vbool32_t op0, vuint32m1_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_xu_f_v_u32m2))) +vuint32m2_t vfcvt_rtz_xu(vfloat32m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_xu_f_v_u32m2_m))) +vuint32m2_t vfcvt_rtz_xu(vbool16_t op0, vuint32m2_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_xu_f_v_u32m4))) +vuint32m4_t vfcvt_rtz_xu(vfloat32m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_xu_f_v_u32m4_m))) +vuint32m4_t vfcvt_rtz_xu(vbool8_t op0, vuint32m4_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_xu_f_v_u32m8))) +vuint32m8_t vfcvt_rtz_xu(vfloat32m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_xu_f_v_u32m8_m))) +vuint32m8_t vfcvt_rtz_xu(vbool4_t op0, vuint32m8_t op1, vfloat32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_xu_f_v_u32mf2))) +vuint32mf2_t vfcvt_rtz_xu(vfloat32mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_xu_f_v_u32mf2_m))) +vuint32mf2_t vfcvt_rtz_xu(vbool64_t op0, vuint32mf2_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_x_f_v_i32m1))) +vint32m1_t vfcvt_x(vfloat32m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_x_f_v_i32m1_m))) +vint32m1_t vfcvt_x(vbool32_t op0, vint32m1_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_x_f_v_i32m2))) +vint32m2_t vfcvt_x(vfloat32m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_x_f_v_i32m2_m))) +vint32m2_t vfcvt_x(vbool16_t op0, vint32m2_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_x_f_v_i32m4))) +vint32m4_t vfcvt_x(vfloat32m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_x_f_v_i32m4_m))) +vint32m4_t vfcvt_x(vbool8_t op0, vint32m4_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_x_f_v_i32m8))) +vint32m8_t vfcvt_x(vfloat32m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_x_f_v_i32m8_m))) +vint32m8_t vfcvt_x(vbool4_t op0, vint32m8_t op1, vfloat32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_x_f_v_i32mf2))) +vint32mf2_t vfcvt_x(vfloat32mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_x_f_v_i32mf2_m))) +vint32mf2_t vfcvt_x(vbool64_t op0, vint32mf2_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_xu_f_v_u32m1))) +vuint32m1_t vfcvt_xu(vfloat32m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_xu_f_v_u32m1_m))) +vuint32m1_t vfcvt_xu(vbool32_t op0, vuint32m1_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_xu_f_v_u32m2))) +vuint32m2_t vfcvt_xu(vfloat32m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_xu_f_v_u32m2_m))) +vuint32m2_t vfcvt_xu(vbool16_t op0, vuint32m2_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_xu_f_v_u32m4))) +vuint32m4_t vfcvt_xu(vfloat32m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_xu_f_v_u32m4_m))) +vuint32m4_t vfcvt_xu(vbool8_t op0, vuint32m4_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_xu_f_v_u32m8))) +vuint32m8_t vfcvt_xu(vfloat32m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_xu_f_v_u32m8_m))) +vuint32m8_t vfcvt_xu(vbool4_t op0, vuint32m8_t op1, vfloat32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_xu_f_v_u32mf2))) +vuint32mf2_t vfcvt_xu(vfloat32mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_xu_f_v_u32mf2_m))) +vuint32mf2_t vfcvt_xu(vbool64_t op0, vuint32mf2_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_x_w_f32mf2))) +vfloat32mf2_t vfncvt_f(vint64m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_x_w_f32mf2_m))) +vfloat32mf2_t vfncvt_f(vbool64_t op0, vfloat32mf2_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_x_w_f32m1))) +vfloat32m1_t vfncvt_f(vint64m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_x_w_f32m1_m))) +vfloat32m1_t vfncvt_f(vbool32_t op0, vfloat32m1_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_x_w_f32m2))) +vfloat32m2_t vfncvt_f(vint64m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_x_w_f32m2_m))) +vfloat32m2_t vfncvt_f(vbool16_t op0, vfloat32m2_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_x_w_f32m4))) +vfloat32m4_t vfncvt_f(vint64m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_x_w_f32m4_m))) +vfloat32m4_t vfncvt_f(vbool8_t op0, vfloat32m4_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_xu_w_f32mf2))) +vfloat32mf2_t vfncvt_f(vuint64m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_xu_w_f32mf2_m))) +vfloat32mf2_t vfncvt_f(vbool64_t op0, vfloat32mf2_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_xu_w_f32m1))) +vfloat32m1_t vfncvt_f(vuint64m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_xu_w_f32m1_m))) +vfloat32m1_t vfncvt_f(vbool32_t op0, vfloat32m1_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_xu_w_f32m2))) +vfloat32m2_t vfncvt_f(vuint64m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_xu_w_f32m2_m))) +vfloat32m2_t vfncvt_f(vbool16_t op0, vfloat32m2_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_xu_w_f32m4))) +vfloat32m4_t vfncvt_f(vuint64m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_xu_w_f32m4_m))) +vfloat32m4_t vfncvt_f(vbool8_t op0, vfloat32m4_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_x_f_w_i16mf4))) +vint16mf4_t vfncvt_rtz_x(vfloat32mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_x_f_w_i16mf4_m))) +vint16mf4_t vfncvt_rtz_x(vbool64_t op0, vint16mf4_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_x_f_w_i16mf2))) +vint16mf2_t vfncvt_rtz_x(vfloat32m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_x_f_w_i16mf2_m))) +vint16mf2_t vfncvt_rtz_x(vbool32_t op0, vint16mf2_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_x_f_w_i16m1))) +vint16m1_t vfncvt_rtz_x(vfloat32m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_x_f_w_i16m1_m))) +vint16m1_t vfncvt_rtz_x(vbool16_t op0, vint16m1_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_x_f_w_i16m2))) +vint16m2_t vfncvt_rtz_x(vfloat32m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_x_f_w_i16m2_m))) +vint16m2_t vfncvt_rtz_x(vbool8_t op0, vint16m2_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_x_f_w_i16m4))) +vint16m4_t vfncvt_rtz_x(vfloat32m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_x_f_w_i16m4_m))) +vint16m4_t vfncvt_rtz_x(vbool4_t op0, vint16m4_t op1, vfloat32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_xu_f_w_u16mf4))) +vuint16mf4_t vfncvt_rtz_xu(vfloat32mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_xu_f_w_u16mf4_m))) +vuint16mf4_t vfncvt_rtz_xu(vbool64_t op0, vuint16mf4_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_xu_f_w_u16mf2))) +vuint16mf2_t vfncvt_rtz_xu(vfloat32m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_xu_f_w_u16mf2_m))) +vuint16mf2_t vfncvt_rtz_xu(vbool32_t op0, vuint16mf2_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_xu_f_w_u16m1))) +vuint16m1_t vfncvt_rtz_xu(vfloat32m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_xu_f_w_u16m1_m))) +vuint16m1_t vfncvt_rtz_xu(vbool16_t op0, vuint16m1_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_xu_f_w_u16m2))) +vuint16m2_t vfncvt_rtz_xu(vfloat32m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_xu_f_w_u16m2_m))) +vuint16m2_t vfncvt_rtz_xu(vbool8_t op0, vuint16m2_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_xu_f_w_u16m4))) +vuint16m4_t vfncvt_rtz_xu(vfloat32m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_xu_f_w_u16m4_m))) +vuint16m4_t vfncvt_rtz_xu(vbool4_t op0, vuint16m4_t op1, vfloat32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_x_f_w_i16mf4))) +vint16mf4_t vfncvt_x(vfloat32mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_x_f_w_i16mf4_m))) +vint16mf4_t vfncvt_x(vbool64_t op0, vint16mf4_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_x_f_w_i16mf2))) +vint16mf2_t vfncvt_x(vfloat32m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_x_f_w_i16mf2_m))) +vint16mf2_t vfncvt_x(vbool32_t op0, vint16mf2_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_x_f_w_i16m1))) +vint16m1_t vfncvt_x(vfloat32m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_x_f_w_i16m1_m))) +vint16m1_t vfncvt_x(vbool16_t op0, vint16m1_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_x_f_w_i16m2))) +vint16m2_t vfncvt_x(vfloat32m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_x_f_w_i16m2_m))) +vint16m2_t vfncvt_x(vbool8_t op0, vint16m2_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_x_f_w_i16m4))) +vint16m4_t vfncvt_x(vfloat32m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_x_f_w_i16m4_m))) +vint16m4_t vfncvt_x(vbool4_t op0, vint16m4_t op1, vfloat32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_xu_f_w_u16mf4))) +vuint16mf4_t vfncvt_xu(vfloat32mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_xu_f_w_u16mf4_m))) +vuint16mf4_t vfncvt_xu(vbool64_t op0, vuint16mf4_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_xu_f_w_u16mf2))) +vuint16mf2_t vfncvt_xu(vfloat32m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_xu_f_w_u16mf2_m))) +vuint16mf2_t vfncvt_xu(vbool32_t op0, vuint16mf2_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_xu_f_w_u16m1))) +vuint16m1_t vfncvt_xu(vfloat32m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_xu_f_w_u16m1_m))) +vuint16m1_t vfncvt_xu(vbool16_t op0, vuint16m1_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_xu_f_w_u16m2))) +vuint16m2_t vfncvt_xu(vfloat32m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_xu_f_w_u16m2_m))) +vuint16m2_t vfncvt_xu(vbool8_t op0, vuint16m2_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_xu_f_w_u16m4))) +vuint16m4_t vfncvt_xu(vfloat32m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_xu_f_w_u16m4_m))) +vuint16m4_t vfncvt_xu(vbool4_t op0, vuint16m4_t op1, vfloat32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrec7_v_f32m1))) +vfloat32m1_t vfrec7(vfloat32m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrec7_v_f32m1_m))) +vfloat32m1_t vfrec7(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrec7_v_f32m2))) +vfloat32m2_t vfrec7(vfloat32m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrec7_v_f32m2_m))) +vfloat32m2_t vfrec7(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrec7_v_f32m4))) +vfloat32m4_t vfrec7(vfloat32m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrec7_v_f32m4_m))) +vfloat32m4_t vfrec7(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrec7_v_f32m8))) +vfloat32m8_t vfrec7(vfloat32m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrec7_v_f32m8_m))) +vfloat32m8_t vfrec7(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrec7_v_f32mf2))) +vfloat32mf2_t vfrec7(vfloat32mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrec7_v_f32mf2_m))) +vfloat32mf2_t vfrec7(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsqrt7_v_f32m1))) +vfloat32m1_t vfrsqrt7(vfloat32m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsqrt7_v_f32m1_m))) +vfloat32m1_t vfrsqrt7(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsqrt7_v_f32m2))) +vfloat32m2_t vfrsqrt7(vfloat32m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsqrt7_v_f32m2_m))) +vfloat32m2_t vfrsqrt7(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsqrt7_v_f32m4))) +vfloat32m4_t vfrsqrt7(vfloat32m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsqrt7_v_f32m4_m))) +vfloat32m4_t vfrsqrt7(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsqrt7_v_f32m8))) +vfloat32m8_t vfrsqrt7(vfloat32m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsqrt7_v_f32m8_m))) +vfloat32m8_t vfrsqrt7(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsqrt7_v_f32mf2))) +vfloat32mf2_t vfrsqrt7(vfloat32mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsqrt7_v_f32mf2_m))) +vfloat32mf2_t vfrsqrt7(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsqrt_v_f32m1))) +vfloat32m1_t vfsqrt(vfloat32m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsqrt_v_f32m1_m))) +vfloat32m1_t vfsqrt(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsqrt_v_f32m2))) +vfloat32m2_t vfsqrt(vfloat32m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsqrt_v_f32m2_m))) +vfloat32m2_t vfsqrt(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsqrt_v_f32m4))) +vfloat32m4_t vfsqrt(vfloat32m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsqrt_v_f32m4_m))) +vfloat32m4_t vfsqrt(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsqrt_v_f32m8))) +vfloat32m8_t vfsqrt(vfloat32m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsqrt_v_f32m8_m))) +vfloat32m8_t vfsqrt(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsqrt_v_f32mf2))) +vfloat32mf2_t vfsqrt(vfloat32mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsqrt_v_f32mf2_m))) +vfloat32mf2_t vfsqrt(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_x_v_f32mf2))) +vfloat32mf2_t vfwcvt_f(vint16mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_x_v_f32mf2_m))) +vfloat32mf2_t vfwcvt_f(vbool64_t op0, vfloat32mf2_t op1, vint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_x_v_f32m1))) +vfloat32m1_t vfwcvt_f(vint16mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_x_v_f32m1_m))) +vfloat32m1_t vfwcvt_f(vbool32_t op0, vfloat32m1_t op1, vint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_x_v_f32m2))) +vfloat32m2_t vfwcvt_f(vint16m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_x_v_f32m2_m))) +vfloat32m2_t vfwcvt_f(vbool16_t op0, vfloat32m2_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_x_v_f32m4))) +vfloat32m4_t vfwcvt_f(vint16m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_x_v_f32m4_m))) +vfloat32m4_t vfwcvt_f(vbool8_t op0, vfloat32m4_t op1, vint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_x_v_f32m8))) +vfloat32m8_t vfwcvt_f(vint16m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_x_v_f32m8_m))) +vfloat32m8_t vfwcvt_f(vbool4_t op0, vfloat32m8_t op1, vint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_xu_v_f32mf2))) +vfloat32mf2_t vfwcvt_f(vuint16mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_xu_v_f32mf2_m))) +vfloat32mf2_t vfwcvt_f(vbool64_t op0, vfloat32mf2_t op1, vuint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_xu_v_f32m1))) +vfloat32m1_t vfwcvt_f(vuint16mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_xu_v_f32m1_m))) +vfloat32m1_t vfwcvt_f(vbool32_t op0, vfloat32m1_t op1, vuint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_xu_v_f32m2))) +vfloat32m2_t vfwcvt_f(vuint16m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_xu_v_f32m2_m))) +vfloat32m2_t vfwcvt_f(vbool16_t op0, vfloat32m2_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_xu_v_f32m4))) +vfloat32m4_t vfwcvt_f(vuint16m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_xu_v_f32m4_m))) +vfloat32m4_t vfwcvt_f(vbool8_t op0, vfloat32m4_t op1, vuint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_xu_v_f32m8))) +vfloat32m8_t vfwcvt_f(vuint16m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_xu_v_f32m8_m))) +vfloat32m8_t vfwcvt_f(vbool4_t op0, vfloat32m8_t op1, vuint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_rtz_x_f_v_i64m1))) +vint64m1_t vfwcvt_rtz_x(vfloat32mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_rtz_x_f_v_i64m1_m))) +vint64m1_t vfwcvt_rtz_x(vbool64_t op0, vint64m1_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_rtz_x_f_v_i64m2))) +vint64m2_t vfwcvt_rtz_x(vfloat32m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_rtz_x_f_v_i64m2_m))) +vint64m2_t vfwcvt_rtz_x(vbool32_t op0, vint64m2_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_rtz_x_f_v_i64m4))) +vint64m4_t vfwcvt_rtz_x(vfloat32m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_rtz_x_f_v_i64m4_m))) +vint64m4_t vfwcvt_rtz_x(vbool16_t op0, vint64m4_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_rtz_x_f_v_i64m8))) +vint64m8_t vfwcvt_rtz_x(vfloat32m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_rtz_x_f_v_i64m8_m))) +vint64m8_t vfwcvt_rtz_x(vbool8_t op0, vint64m8_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_rtz_xu_f_v_u64m1))) +vuint64m1_t vfwcvt_rtz_xu(vfloat32mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_rtz_xu_f_v_u64m1_m))) +vuint64m1_t vfwcvt_rtz_xu(vbool64_t op0, vuint64m1_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_rtz_xu_f_v_u64m2))) +vuint64m2_t vfwcvt_rtz_xu(vfloat32m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_rtz_xu_f_v_u64m2_m))) +vuint64m2_t vfwcvt_rtz_xu(vbool32_t op0, vuint64m2_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_rtz_xu_f_v_u64m4))) +vuint64m4_t vfwcvt_rtz_xu(vfloat32m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_rtz_xu_f_v_u64m4_m))) +vuint64m4_t vfwcvt_rtz_xu(vbool16_t op0, vuint64m4_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_rtz_xu_f_v_u64m8))) +vuint64m8_t vfwcvt_rtz_xu(vfloat32m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_rtz_xu_f_v_u64m8_m))) +vuint64m8_t vfwcvt_rtz_xu(vbool8_t op0, vuint64m8_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_x_f_v_i64m1))) +vint64m1_t vfwcvt_x(vfloat32mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_x_f_v_i64m1_m))) +vint64m1_t vfwcvt_x(vbool64_t op0, vint64m1_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_x_f_v_i64m2))) +vint64m2_t vfwcvt_x(vfloat32m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_x_f_v_i64m2_m))) +vint64m2_t vfwcvt_x(vbool32_t op0, vint64m2_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_x_f_v_i64m4))) +vint64m4_t vfwcvt_x(vfloat32m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_x_f_v_i64m4_m))) +vint64m4_t vfwcvt_x(vbool16_t op0, vint64m4_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_x_f_v_i64m8))) +vint64m8_t vfwcvt_x(vfloat32m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_x_f_v_i64m8_m))) +vint64m8_t vfwcvt_x(vbool8_t op0, vint64m8_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_xu_f_v_u64m1))) +vuint64m1_t vfwcvt_xu(vfloat32mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_xu_f_v_u64m1_m))) +vuint64m1_t vfwcvt_xu(vbool64_t op0, vuint64m1_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_xu_f_v_u64m2))) +vuint64m2_t vfwcvt_xu(vfloat32m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_xu_f_v_u64m2_m))) +vuint64m2_t vfwcvt_xu(vbool32_t op0, vuint64m2_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_xu_f_v_u64m4))) +vuint64m4_t vfwcvt_xu(vfloat32m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_xu_f_v_u64m4_m))) +vuint64m4_t vfwcvt_xu(vbool16_t op0, vuint64m4_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_xu_f_v_u64m8))) +vuint64m8_t vfwcvt_xu(vfloat32m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_xu_f_v_u64m8_m))) +vuint64m8_t vfwcvt_xu(vbool8_t op0, vuint64m8_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_f32mf2_f32m1))) +vfloat32m1_t vlmul_ext_f32m1(vfloat32mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_f32m1_f32m2))) +vfloat32m2_t vlmul_ext_f32m2(vfloat32m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_f32mf2_f32m2))) +vfloat32m2_t vlmul_ext_f32m2(vfloat32mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_f32m1_f32m4))) +vfloat32m4_t vlmul_ext_f32m4(vfloat32m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_f32m2_f32m4))) +vfloat32m4_t vlmul_ext_f32m4(vfloat32m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_f32mf2_f32m4))) +vfloat32m4_t vlmul_ext_f32m4(vfloat32mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_f32m1_f32m8))) +vfloat32m8_t vlmul_ext_f32m8(vfloat32m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_f32m2_f32m8))) +vfloat32m8_t vlmul_ext_f32m8(vfloat32m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_f32m4_f32m8))) +vfloat32m8_t vlmul_ext_f32m8(vfloat32m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_f32mf2_f32m8))) +vfloat32m8_t vlmul_ext_f32m8(vfloat32mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_f32m1_f32mf2))) +vfloat32mf2_t vlmul_trunc_f32mf2(vfloat32m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_f32m2_f32mf2))) +vfloat32mf2_t vlmul_trunc_f32mf2(vfloat32m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_f32m4_f32mf2))) +vfloat32mf2_t vlmul_trunc_f32mf2(vfloat32m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_f32m8_f32mf2))) +vfloat32mf2_t vlmul_trunc_f32mf2(vfloat32m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_f32m2_f32m1))) +vfloat32m1_t vlmul_trunc_f32m1(vfloat32m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_f32m4_f32m1))) +vfloat32m1_t vlmul_trunc_f32m1(vfloat32m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_f32m8_f32m1))) +vfloat32m1_t vlmul_trunc_f32m1(vfloat32m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_f32m4_f32m2))) +vfloat32m2_t vlmul_trunc_f32m2(vfloat32m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_f32m8_f32m2))) +vfloat32m2_t vlmul_trunc_f32m2(vfloat32m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_f32m8_f32m4))) +vfloat32m4_t vlmul_trunc_f32m4(vfloat32m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i32m1_f32m1))) +vfloat32m1_t vreinterpret_f32m1(vint32m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i32m2_f32m2))) +vfloat32m2_t vreinterpret_f32m2(vint32m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i32m4_f32m4))) +vfloat32m4_t vreinterpret_f32m4(vint32m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i32m8_f32m8))) +vfloat32m8_t vreinterpret_f32m8(vint32m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i32mf2_f32mf2))) +vfloat32mf2_t vreinterpret_f32mf2(vint32mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u32m1_f32m1))) +vfloat32m1_t vreinterpret_f32m1(vuint32m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u32m2_f32m2))) +vfloat32m2_t vreinterpret_f32m2(vuint32m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u32m4_f32m4))) +vfloat32m4_t vreinterpret_f32m4(vuint32m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u32m8_f32m8))) +vfloat32m8_t vreinterpret_f32m8(vuint32m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u32mf2_f32mf2))) +vfloat32mf2_t vreinterpret_f32mf2(vuint32mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_f32m1_i32m1))) +vint32m1_t vreinterpret_i32m1(vfloat32m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_f32m2_i32m2))) +vint32m2_t vreinterpret_i32m2(vfloat32m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_f32m4_i32m4))) +vint32m4_t vreinterpret_i32m4(vfloat32m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_f32m8_i32m8))) +vint32m8_t vreinterpret_i32m8(vfloat32m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_f32mf2_i32mf2))) +vint32mf2_t vreinterpret_i32mf2(vfloat32mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_f32m1_u32m1))) +vuint32m1_t vreinterpret_u32m1(vfloat32m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_f32m2_u32m2))) +vuint32m2_t vreinterpret_u32m2(vfloat32m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_f32m4_u32m4))) +vuint32m4_t vreinterpret_u32m4(vfloat32m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_f32m8_u32m8))) +vuint32m8_t vreinterpret_u32m8(vfloat32m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_f32mf2_u32mf2))) +vuint32mf2_t vreinterpret_u32mf2(vfloat32mf2_t op0); + +#endif + +#if defined(__riscv_d) +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse64_v_f64m1_m))) +vfloat64m1_t vlse64(vbool64_t op0, vfloat64m1_t op1, const double * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse64_v_f64m2_m))) +vfloat64m2_t vlse64(vbool32_t op0, vfloat64m2_t op1, const double * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse64_v_f64m4_m))) +vfloat64m4_t vlse64(vbool16_t op0, vfloat64m4_t op1, const double * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse64_v_f64m8_m))) +vfloat64m8_t vlse64(vbool8_t op0, vfloat64m8_t op1, const double * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse64_v_f64m1))) +void vsse64(double * op0, ptrdiff_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse64_v_f64m1_m))) +void vsse64(vbool64_t op0, double * op1, ptrdiff_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse64_v_f64m2))) +void vsse64(double * op0, ptrdiff_t op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse64_v_f64m2_m))) +void vsse64(vbool32_t op0, double * op1, ptrdiff_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse64_v_f64m4))) +void vsse64(double * op0, ptrdiff_t op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse64_v_f64m4_m))) +void vsse64(vbool16_t op0, double * op1, ptrdiff_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse64_v_f64m8))) +void vsse64(double * op0, ptrdiff_t op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse64_v_f64m8_m))) +void vsse64(vbool8_t op0, double * op1, ptrdiff_t op2, vfloat64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_f64m1))) +vfloat64m1_t vluxei8(const double * op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_f64m1_m))) +vfloat64m1_t vluxei8(vbool64_t op0, vfloat64m1_t op1, const double * op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_f64m2))) +vfloat64m2_t vluxei8(const double * op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_f64m2_m))) +vfloat64m2_t vluxei8(vbool32_t op0, vfloat64m2_t op1, const double * op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_f64m4))) +vfloat64m4_t vluxei8(const double * op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_f64m4_m))) +vfloat64m4_t vluxei8(vbool16_t op0, vfloat64m4_t op1, const double * op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_f64m8))) +vfloat64m8_t vluxei8(const double * op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_f64m8_m))) +vfloat64m8_t vluxei8(vbool8_t op0, vfloat64m8_t op1, const double * op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_f64m1))) +vfloat64m1_t vluxei16(const double * op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_f64m1_m))) +vfloat64m1_t vluxei16(vbool64_t op0, vfloat64m1_t op1, const double * op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_f64m2))) +vfloat64m2_t vluxei16(const double * op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_f64m2_m))) +vfloat64m2_t vluxei16(vbool32_t op0, vfloat64m2_t op1, const double * op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_f64m4))) +vfloat64m4_t vluxei16(const double * op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_f64m4_m))) +vfloat64m4_t vluxei16(vbool16_t op0, vfloat64m4_t op1, const double * op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_f64m8))) +vfloat64m8_t vluxei16(const double * op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_f64m8_m))) +vfloat64m8_t vluxei16(vbool8_t op0, vfloat64m8_t op1, const double * op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_f64m1))) +vfloat64m1_t vluxei32(const double * op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_f64m1_m))) +vfloat64m1_t vluxei32(vbool64_t op0, vfloat64m1_t op1, const double * op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_f64m2))) +vfloat64m2_t vluxei32(const double * op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_f64m2_m))) +vfloat64m2_t vluxei32(vbool32_t op0, vfloat64m2_t op1, const double * op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_f64m4))) +vfloat64m4_t vluxei32(const double * op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_f64m4_m))) +vfloat64m4_t vluxei32(vbool16_t op0, vfloat64m4_t op1, const double * op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_f64m8))) +vfloat64m8_t vluxei32(const double * op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_f64m8_m))) +vfloat64m8_t vluxei32(vbool8_t op0, vfloat64m8_t op1, const double * op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_f64m1))) +vfloat64m1_t vluxei64(const double * op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_f64m1_m))) +vfloat64m1_t vluxei64(vbool64_t op0, vfloat64m1_t op1, const double * op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_f64m2))) +vfloat64m2_t vluxei64(const double * op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_f64m2_m))) +vfloat64m2_t vluxei64(vbool32_t op0, vfloat64m2_t op1, const double * op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_f64m4))) +vfloat64m4_t vluxei64(const double * op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_f64m4_m))) +vfloat64m4_t vluxei64(vbool16_t op0, vfloat64m4_t op1, const double * op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_f64m8))) +vfloat64m8_t vluxei64(const double * op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_f64m8_m))) +vfloat64m8_t vluxei64(vbool8_t op0, vfloat64m8_t op1, const double * op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_f64m1))) +vfloat64m1_t vloxei8(const double * op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_f64m1_m))) +vfloat64m1_t vloxei8(vbool64_t op0, vfloat64m1_t op1, const double * op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_f64m2))) +vfloat64m2_t vloxei8(const double * op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_f64m2_m))) +vfloat64m2_t vloxei8(vbool32_t op0, vfloat64m2_t op1, const double * op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_f64m4))) +vfloat64m4_t vloxei8(const double * op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_f64m4_m))) +vfloat64m4_t vloxei8(vbool16_t op0, vfloat64m4_t op1, const double * op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_f64m8))) +vfloat64m8_t vloxei8(const double * op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_f64m8_m))) +vfloat64m8_t vloxei8(vbool8_t op0, vfloat64m8_t op1, const double * op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_f64m1))) +vfloat64m1_t vloxei16(const double * op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_f64m1_m))) +vfloat64m1_t vloxei16(vbool64_t op0, vfloat64m1_t op1, const double * op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_f64m2))) +vfloat64m2_t vloxei16(const double * op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_f64m2_m))) +vfloat64m2_t vloxei16(vbool32_t op0, vfloat64m2_t op1, const double * op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_f64m4))) +vfloat64m4_t vloxei16(const double * op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_f64m4_m))) +vfloat64m4_t vloxei16(vbool16_t op0, vfloat64m4_t op1, const double * op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_f64m8))) +vfloat64m8_t vloxei16(const double * op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_f64m8_m))) +vfloat64m8_t vloxei16(vbool8_t op0, vfloat64m8_t op1, const double * op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_f64m1))) +vfloat64m1_t vloxei32(const double * op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_f64m1_m))) +vfloat64m1_t vloxei32(vbool64_t op0, vfloat64m1_t op1, const double * op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_f64m2))) +vfloat64m2_t vloxei32(const double * op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_f64m2_m))) +vfloat64m2_t vloxei32(vbool32_t op0, vfloat64m2_t op1, const double * op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_f64m4))) +vfloat64m4_t vloxei32(const double * op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_f64m4_m))) +vfloat64m4_t vloxei32(vbool16_t op0, vfloat64m4_t op1, const double * op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_f64m8))) +vfloat64m8_t vloxei32(const double * op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_f64m8_m))) +vfloat64m8_t vloxei32(vbool8_t op0, vfloat64m8_t op1, const double * op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_f64m1))) +vfloat64m1_t vloxei64(const double * op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_f64m1_m))) +vfloat64m1_t vloxei64(vbool64_t op0, vfloat64m1_t op1, const double * op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_f64m2))) +vfloat64m2_t vloxei64(const double * op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_f64m2_m))) +vfloat64m2_t vloxei64(vbool32_t op0, vfloat64m2_t op1, const double * op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_f64m4))) +vfloat64m4_t vloxei64(const double * op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_f64m4_m))) +vfloat64m4_t vloxei64(vbool16_t op0, vfloat64m4_t op1, const double * op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_f64m8))) +vfloat64m8_t vloxei64(const double * op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_f64m8_m))) +vfloat64m8_t vloxei64(vbool8_t op0, vfloat64m8_t op1, const double * op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_f64m1))) +vfloat64m1_t vmv_v(vfloat64m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_f64m2))) +vfloat64m2_t vmv_v(vfloat64m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_f64m4))) +vfloat64m4_t vmv_v(vfloat64m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_f64m8))) +vfloat64m8_t vmv_v(vfloat64m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vv_f64m1))) +vfloat64m1_t vfadd(vfloat64m1_t op0, vfloat64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vv_f64m1_m))) +vfloat64m1_t vfadd(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vv_f64m2))) +vfloat64m2_t vfadd(vfloat64m2_t op0, vfloat64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vv_f64m2_m))) +vfloat64m2_t vfadd(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vv_f64m4))) +vfloat64m4_t vfadd(vfloat64m4_t op0, vfloat64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vv_f64m4_m))) +vfloat64m4_t vfadd(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vv_f64m8))) +vfloat64m8_t vfadd(vfloat64m8_t op0, vfloat64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vv_f64m8_m))) +vfloat64m8_t vfadd(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, vfloat64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vf_f64m1))) +vfloat64m1_t vfadd(vfloat64m1_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vf_f64m1_m))) +vfloat64m1_t vfadd(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vf_f64m2))) +vfloat64m2_t vfadd(vfloat64m2_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vf_f64m2_m))) +vfloat64m2_t vfadd(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vf_f64m4))) +vfloat64m4_t vfadd(vfloat64m4_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vf_f64m4_m))) +vfloat64m4_t vfadd(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vf_f64m8))) +vfloat64m8_t vfadd(vfloat64m8_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vf_f64m8_m))) +vfloat64m8_t vfadd(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vv_f64m1))) +vfloat64m1_t vfsub(vfloat64m1_t op0, vfloat64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vv_f64m1_m))) +vfloat64m1_t vfsub(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vv_f64m2))) +vfloat64m2_t vfsub(vfloat64m2_t op0, vfloat64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vv_f64m2_m))) +vfloat64m2_t vfsub(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vv_f64m4))) +vfloat64m4_t vfsub(vfloat64m4_t op0, vfloat64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vv_f64m4_m))) +vfloat64m4_t vfsub(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vv_f64m8))) +vfloat64m8_t vfsub(vfloat64m8_t op0, vfloat64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vv_f64m8_m))) +vfloat64m8_t vfsub(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, vfloat64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vf_f64m1))) +vfloat64m1_t vfsub(vfloat64m1_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vf_f64m1_m))) +vfloat64m1_t vfsub(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vf_f64m2))) +vfloat64m2_t vfsub(vfloat64m2_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vf_f64m2_m))) +vfloat64m2_t vfsub(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vf_f64m4))) +vfloat64m4_t vfsub(vfloat64m4_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vf_f64m4_m))) +vfloat64m4_t vfsub(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vf_f64m8))) +vfloat64m8_t vfsub(vfloat64m8_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vf_f64m8_m))) +vfloat64m8_t vfsub(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsub_vf_f64m1))) +vfloat64m1_t vfrsub(vfloat64m1_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsub_vf_f64m1_m))) +vfloat64m1_t vfrsub(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsub_vf_f64m2))) +vfloat64m2_t vfrsub(vfloat64m2_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsub_vf_f64m2_m))) +vfloat64m2_t vfrsub(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsub_vf_f64m4))) +vfloat64m4_t vfrsub(vfloat64m4_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsub_vf_f64m4_m))) +vfloat64m4_t vfrsub(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsub_vf_f64m8))) +vfloat64m8_t vfrsub(vfloat64m8_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsub_vf_f64m8_m))) +vfloat64m8_t vfrsub(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vv_f64m1))) +vfloat64m1_t vfmul(vfloat64m1_t op0, vfloat64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vv_f64m1_m))) +vfloat64m1_t vfmul(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vv_f64m2))) +vfloat64m2_t vfmul(vfloat64m2_t op0, vfloat64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vv_f64m2_m))) +vfloat64m2_t vfmul(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vv_f64m4))) +vfloat64m4_t vfmul(vfloat64m4_t op0, vfloat64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vv_f64m4_m))) +vfloat64m4_t vfmul(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vv_f64m8))) +vfloat64m8_t vfmul(vfloat64m8_t op0, vfloat64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vv_f64m8_m))) +vfloat64m8_t vfmul(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, vfloat64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vf_f64m1))) +vfloat64m1_t vfmul(vfloat64m1_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vf_f64m1_m))) +vfloat64m1_t vfmul(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vf_f64m2))) +vfloat64m2_t vfmul(vfloat64m2_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vf_f64m2_m))) +vfloat64m2_t vfmul(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vf_f64m4))) +vfloat64m4_t vfmul(vfloat64m4_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vf_f64m4_m))) +vfloat64m4_t vfmul(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vf_f64m8))) +vfloat64m8_t vfmul(vfloat64m8_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vf_f64m8_m))) +vfloat64m8_t vfmul(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vv_f64m1))) +vfloat64m1_t vfdiv(vfloat64m1_t op0, vfloat64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vv_f64m1_m))) +vfloat64m1_t vfdiv(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vv_f64m2))) +vfloat64m2_t vfdiv(vfloat64m2_t op0, vfloat64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vv_f64m2_m))) +vfloat64m2_t vfdiv(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vv_f64m4))) +vfloat64m4_t vfdiv(vfloat64m4_t op0, vfloat64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vv_f64m4_m))) +vfloat64m4_t vfdiv(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vv_f64m8))) +vfloat64m8_t vfdiv(vfloat64m8_t op0, vfloat64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vv_f64m8_m))) +vfloat64m8_t vfdiv(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, vfloat64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vf_f64m1))) +vfloat64m1_t vfdiv(vfloat64m1_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vf_f64m1_m))) +vfloat64m1_t vfdiv(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vf_f64m2))) +vfloat64m2_t vfdiv(vfloat64m2_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vf_f64m2_m))) +vfloat64m2_t vfdiv(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vf_f64m4))) +vfloat64m4_t vfdiv(vfloat64m4_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vf_f64m4_m))) +vfloat64m4_t vfdiv(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vf_f64m8))) +vfloat64m8_t vfdiv(vfloat64m8_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vf_f64m8_m))) +vfloat64m8_t vfdiv(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrdiv_vf_f64m1))) +vfloat64m1_t vfrdiv(vfloat64m1_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrdiv_vf_f64m1_m))) +vfloat64m1_t vfrdiv(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrdiv_vf_f64m2))) +vfloat64m2_t vfrdiv(vfloat64m2_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrdiv_vf_f64m2_m))) +vfloat64m2_t vfrdiv(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrdiv_vf_f64m4))) +vfloat64m4_t vfrdiv(vfloat64m4_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrdiv_vf_f64m4_m))) +vfloat64m4_t vfrdiv(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrdiv_vf_f64m8))) +vfloat64m8_t vfrdiv(vfloat64m8_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrdiv_vf_f64m8_m))) +vfloat64m8_t vfrdiv(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vv_f64m1))) +vfloat64m1_t vfmacc(vfloat64m1_t op0, vfloat64m1_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vv_f64m1_m))) +vfloat64m1_t vfmacc(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vv_f64m2))) +vfloat64m2_t vfmacc(vfloat64m2_t op0, vfloat64m2_t op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vv_f64m2_m))) +vfloat64m2_t vfmacc(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vv_f64m4))) +vfloat64m4_t vfmacc(vfloat64m4_t op0, vfloat64m4_t op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vv_f64m4_m))) +vfloat64m4_t vfmacc(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vv_f64m8))) +vfloat64m8_t vfmacc(vfloat64m8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vv_f64m8_m))) +vfloat64m8_t vfmacc(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, vfloat64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vf_f64m1))) +vfloat64m1_t vfmacc(vfloat64m1_t op0, double op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vf_f64m1_m))) +vfloat64m1_t vfmacc(vbool64_t op0, vfloat64m1_t op1, double op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vf_f64m2))) +vfloat64m2_t vfmacc(vfloat64m2_t op0, double op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vf_f64m2_m))) +vfloat64m2_t vfmacc(vbool32_t op0, vfloat64m2_t op1, double op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vf_f64m4))) +vfloat64m4_t vfmacc(vfloat64m4_t op0, double op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vf_f64m4_m))) +vfloat64m4_t vfmacc(vbool16_t op0, vfloat64m4_t op1, double op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vf_f64m8))) +vfloat64m8_t vfmacc(vfloat64m8_t op0, double op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vf_f64m8_m))) +vfloat64m8_t vfmacc(vbool8_t op0, vfloat64m8_t op1, double op2, vfloat64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vv_f64m1))) +vfloat64m1_t vfnmacc(vfloat64m1_t op0, vfloat64m1_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vv_f64m1_m))) +vfloat64m1_t vfnmacc(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vv_f64m2))) +vfloat64m2_t vfnmacc(vfloat64m2_t op0, vfloat64m2_t op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vv_f64m2_m))) +vfloat64m2_t vfnmacc(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vv_f64m4))) +vfloat64m4_t vfnmacc(vfloat64m4_t op0, vfloat64m4_t op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vv_f64m4_m))) +vfloat64m4_t vfnmacc(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vv_f64m8))) +vfloat64m8_t vfnmacc(vfloat64m8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vv_f64m8_m))) +vfloat64m8_t vfnmacc(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, vfloat64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vf_f64m1))) +vfloat64m1_t vfnmacc(vfloat64m1_t op0, double op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vf_f64m1_m))) +vfloat64m1_t vfnmacc(vbool64_t op0, vfloat64m1_t op1, double op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vf_f64m2))) +vfloat64m2_t vfnmacc(vfloat64m2_t op0, double op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vf_f64m2_m))) +vfloat64m2_t vfnmacc(vbool32_t op0, vfloat64m2_t op1, double op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vf_f64m4))) +vfloat64m4_t vfnmacc(vfloat64m4_t op0, double op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vf_f64m4_m))) +vfloat64m4_t vfnmacc(vbool16_t op0, vfloat64m4_t op1, double op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vf_f64m8))) +vfloat64m8_t vfnmacc(vfloat64m8_t op0, double op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vf_f64m8_m))) +vfloat64m8_t vfnmacc(vbool8_t op0, vfloat64m8_t op1, double op2, vfloat64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vv_f64m1))) +vfloat64m1_t vfmsac(vfloat64m1_t op0, vfloat64m1_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vv_f64m1_m))) +vfloat64m1_t vfmsac(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vv_f64m2))) +vfloat64m2_t vfmsac(vfloat64m2_t op0, vfloat64m2_t op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vv_f64m2_m))) +vfloat64m2_t vfmsac(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vv_f64m4))) +vfloat64m4_t vfmsac(vfloat64m4_t op0, vfloat64m4_t op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vv_f64m4_m))) +vfloat64m4_t vfmsac(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vv_f64m8))) +vfloat64m8_t vfmsac(vfloat64m8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vv_f64m8_m))) +vfloat64m8_t vfmsac(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, vfloat64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vf_f64m1))) +vfloat64m1_t vfmsac(vfloat64m1_t op0, double op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vf_f64m1_m))) +vfloat64m1_t vfmsac(vbool64_t op0, vfloat64m1_t op1, double op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vf_f64m2))) +vfloat64m2_t vfmsac(vfloat64m2_t op0, double op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vf_f64m2_m))) +vfloat64m2_t vfmsac(vbool32_t op0, vfloat64m2_t op1, double op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vf_f64m4))) +vfloat64m4_t vfmsac(vfloat64m4_t op0, double op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vf_f64m4_m))) +vfloat64m4_t vfmsac(vbool16_t op0, vfloat64m4_t op1, double op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vf_f64m8))) +vfloat64m8_t vfmsac(vfloat64m8_t op0, double op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vf_f64m8_m))) +vfloat64m8_t vfmsac(vbool8_t op0, vfloat64m8_t op1, double op2, vfloat64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vv_f64m1))) +vfloat64m1_t vfnmsac(vfloat64m1_t op0, vfloat64m1_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vv_f64m1_m))) +vfloat64m1_t vfnmsac(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vv_f64m2))) +vfloat64m2_t vfnmsac(vfloat64m2_t op0, vfloat64m2_t op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vv_f64m2_m))) +vfloat64m2_t vfnmsac(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vv_f64m4))) +vfloat64m4_t vfnmsac(vfloat64m4_t op0, vfloat64m4_t op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vv_f64m4_m))) +vfloat64m4_t vfnmsac(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vv_f64m8))) +vfloat64m8_t vfnmsac(vfloat64m8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vv_f64m8_m))) +vfloat64m8_t vfnmsac(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, vfloat64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vf_f64m1))) +vfloat64m1_t vfnmsac(vfloat64m1_t op0, double op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vf_f64m1_m))) +vfloat64m1_t vfnmsac(vbool64_t op0, vfloat64m1_t op1, double op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vf_f64m2))) +vfloat64m2_t vfnmsac(vfloat64m2_t op0, double op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vf_f64m2_m))) +vfloat64m2_t vfnmsac(vbool32_t op0, vfloat64m2_t op1, double op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vf_f64m4))) +vfloat64m4_t vfnmsac(vfloat64m4_t op0, double op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vf_f64m4_m))) +vfloat64m4_t vfnmsac(vbool16_t op0, vfloat64m4_t op1, double op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vf_f64m8))) +vfloat64m8_t vfnmsac(vfloat64m8_t op0, double op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vf_f64m8_m))) +vfloat64m8_t vfnmsac(vbool8_t op0, vfloat64m8_t op1, double op2, vfloat64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vv_f64m1))) +vfloat64m1_t vfmadd(vfloat64m1_t op0, vfloat64m1_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vv_f64m1_m))) +vfloat64m1_t vfmadd(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vv_f64m2))) +vfloat64m2_t vfmadd(vfloat64m2_t op0, vfloat64m2_t op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vv_f64m2_m))) +vfloat64m2_t vfmadd(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vv_f64m4))) +vfloat64m4_t vfmadd(vfloat64m4_t op0, vfloat64m4_t op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vv_f64m4_m))) +vfloat64m4_t vfmadd(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vv_f64m8))) +vfloat64m8_t vfmadd(vfloat64m8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vv_f64m8_m))) +vfloat64m8_t vfmadd(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, vfloat64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vf_f64m1))) +vfloat64m1_t vfmadd(vfloat64m1_t op0, double op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vf_f64m1_m))) +vfloat64m1_t vfmadd(vbool64_t op0, vfloat64m1_t op1, double op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vf_f64m2))) +vfloat64m2_t vfmadd(vfloat64m2_t op0, double op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vf_f64m2_m))) +vfloat64m2_t vfmadd(vbool32_t op0, vfloat64m2_t op1, double op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vf_f64m4))) +vfloat64m4_t vfmadd(vfloat64m4_t op0, double op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vf_f64m4_m))) +vfloat64m4_t vfmadd(vbool16_t op0, vfloat64m4_t op1, double op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vf_f64m8))) +vfloat64m8_t vfmadd(vfloat64m8_t op0, double op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vf_f64m8_m))) +vfloat64m8_t vfmadd(vbool8_t op0, vfloat64m8_t op1, double op2, vfloat64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vv_f64m1))) +vfloat64m1_t vfnmadd(vfloat64m1_t op0, vfloat64m1_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vv_f64m1_m))) +vfloat64m1_t vfnmadd(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vv_f64m2))) +vfloat64m2_t vfnmadd(vfloat64m2_t op0, vfloat64m2_t op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vv_f64m2_m))) +vfloat64m2_t vfnmadd(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vv_f64m4))) +vfloat64m4_t vfnmadd(vfloat64m4_t op0, vfloat64m4_t op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vv_f64m4_m))) +vfloat64m4_t vfnmadd(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vv_f64m8))) +vfloat64m8_t vfnmadd(vfloat64m8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vv_f64m8_m))) +vfloat64m8_t vfnmadd(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, vfloat64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vf_f64m1))) +vfloat64m1_t vfnmadd(vfloat64m1_t op0, double op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vf_f64m1_m))) +vfloat64m1_t vfnmadd(vbool64_t op0, vfloat64m1_t op1, double op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vf_f64m2))) +vfloat64m2_t vfnmadd(vfloat64m2_t op0, double op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vf_f64m2_m))) +vfloat64m2_t vfnmadd(vbool32_t op0, vfloat64m2_t op1, double op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vf_f64m4))) +vfloat64m4_t vfnmadd(vfloat64m4_t op0, double op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vf_f64m4_m))) +vfloat64m4_t vfnmadd(vbool16_t op0, vfloat64m4_t op1, double op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vf_f64m8))) +vfloat64m8_t vfnmadd(vfloat64m8_t op0, double op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vf_f64m8_m))) +vfloat64m8_t vfnmadd(vbool8_t op0, vfloat64m8_t op1, double op2, vfloat64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vv_f64m1))) +vfloat64m1_t vfmsub(vfloat64m1_t op0, vfloat64m1_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vv_f64m1_m))) +vfloat64m1_t vfmsub(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vv_f64m2))) +vfloat64m2_t vfmsub(vfloat64m2_t op0, vfloat64m2_t op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vv_f64m2_m))) +vfloat64m2_t vfmsub(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vv_f64m4))) +vfloat64m4_t vfmsub(vfloat64m4_t op0, vfloat64m4_t op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vv_f64m4_m))) +vfloat64m4_t vfmsub(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vv_f64m8))) +vfloat64m8_t vfmsub(vfloat64m8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vv_f64m8_m))) +vfloat64m8_t vfmsub(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, vfloat64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vf_f64m1))) +vfloat64m1_t vfmsub(vfloat64m1_t op0, double op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vf_f64m1_m))) +vfloat64m1_t vfmsub(vbool64_t op0, vfloat64m1_t op1, double op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vf_f64m2))) +vfloat64m2_t vfmsub(vfloat64m2_t op0, double op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vf_f64m2_m))) +vfloat64m2_t vfmsub(vbool32_t op0, vfloat64m2_t op1, double op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vf_f64m4))) +vfloat64m4_t vfmsub(vfloat64m4_t op0, double op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vf_f64m4_m))) +vfloat64m4_t vfmsub(vbool16_t op0, vfloat64m4_t op1, double op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vf_f64m8))) +vfloat64m8_t vfmsub(vfloat64m8_t op0, double op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vf_f64m8_m))) +vfloat64m8_t vfmsub(vbool8_t op0, vfloat64m8_t op1, double op2, vfloat64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vv_f64m1))) +vfloat64m1_t vfnmsub(vfloat64m1_t op0, vfloat64m1_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vv_f64m1_m))) +vfloat64m1_t vfnmsub(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vv_f64m2))) +vfloat64m2_t vfnmsub(vfloat64m2_t op0, vfloat64m2_t op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vv_f64m2_m))) +vfloat64m2_t vfnmsub(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vv_f64m4))) +vfloat64m4_t vfnmsub(vfloat64m4_t op0, vfloat64m4_t op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vv_f64m4_m))) +vfloat64m4_t vfnmsub(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vv_f64m8))) +vfloat64m8_t vfnmsub(vfloat64m8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vv_f64m8_m))) +vfloat64m8_t vfnmsub(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, vfloat64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vf_f64m1))) +vfloat64m1_t vfnmsub(vfloat64m1_t op0, double op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vf_f64m1_m))) +vfloat64m1_t vfnmsub(vbool64_t op0, vfloat64m1_t op1, double op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vf_f64m2))) +vfloat64m2_t vfnmsub(vfloat64m2_t op0, double op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vf_f64m2_m))) +vfloat64m2_t vfnmsub(vbool32_t op0, vfloat64m2_t op1, double op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vf_f64m4))) +vfloat64m4_t vfnmsub(vfloat64m4_t op0, double op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vf_f64m4_m))) +vfloat64m4_t vfnmsub(vbool16_t op0, vfloat64m4_t op1, double op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vf_f64m8))) +vfloat64m8_t vfnmsub(vfloat64m8_t op0, double op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vf_f64m8_m))) +vfloat64m8_t vfnmsub(vbool8_t op0, vfloat64m8_t op1, double op2, vfloat64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vv_f64m1))) +vfloat64m1_t vfmin(vfloat64m1_t op0, vfloat64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vv_f64m1_m))) +vfloat64m1_t vfmin(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vv_f64m2))) +vfloat64m2_t vfmin(vfloat64m2_t op0, vfloat64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vv_f64m2_m))) +vfloat64m2_t vfmin(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vv_f64m4))) +vfloat64m4_t vfmin(vfloat64m4_t op0, vfloat64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vv_f64m4_m))) +vfloat64m4_t vfmin(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vv_f64m8))) +vfloat64m8_t vfmin(vfloat64m8_t op0, vfloat64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vv_f64m8_m))) +vfloat64m8_t vfmin(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, vfloat64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vf_f64m1))) +vfloat64m1_t vfmin(vfloat64m1_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vf_f64m1_m))) +vfloat64m1_t vfmin(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vf_f64m2))) +vfloat64m2_t vfmin(vfloat64m2_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vf_f64m2_m))) +vfloat64m2_t vfmin(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vf_f64m4))) +vfloat64m4_t vfmin(vfloat64m4_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vf_f64m4_m))) +vfloat64m4_t vfmin(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vf_f64m8))) +vfloat64m8_t vfmin(vfloat64m8_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vf_f64m8_m))) +vfloat64m8_t vfmin(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vv_f64m1))) +vfloat64m1_t vfmax(vfloat64m1_t op0, vfloat64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vv_f64m1_m))) +vfloat64m1_t vfmax(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vv_f64m2))) +vfloat64m2_t vfmax(vfloat64m2_t op0, vfloat64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vv_f64m2_m))) +vfloat64m2_t vfmax(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vv_f64m4))) +vfloat64m4_t vfmax(vfloat64m4_t op0, vfloat64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vv_f64m4_m))) +vfloat64m4_t vfmax(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vv_f64m8))) +vfloat64m8_t vfmax(vfloat64m8_t op0, vfloat64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vv_f64m8_m))) +vfloat64m8_t vfmax(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, vfloat64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vf_f64m1))) +vfloat64m1_t vfmax(vfloat64m1_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vf_f64m1_m))) +vfloat64m1_t vfmax(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vf_f64m2))) +vfloat64m2_t vfmax(vfloat64m2_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vf_f64m2_m))) +vfloat64m2_t vfmax(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vf_f64m4))) +vfloat64m4_t vfmax(vfloat64m4_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vf_f64m4_m))) +vfloat64m4_t vfmax(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vf_f64m8))) +vfloat64m8_t vfmax(vfloat64m8_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vf_f64m8_m))) +vfloat64m8_t vfmax(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vv_f64m1))) +vfloat64m1_t vfsgnj(vfloat64m1_t op0, vfloat64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vv_f64m1_m))) +vfloat64m1_t vfsgnj(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vv_f64m2))) +vfloat64m2_t vfsgnj(vfloat64m2_t op0, vfloat64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vv_f64m2_m))) +vfloat64m2_t vfsgnj(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vv_f64m4))) +vfloat64m4_t vfsgnj(vfloat64m4_t op0, vfloat64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vv_f64m4_m))) +vfloat64m4_t vfsgnj(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vv_f64m8))) +vfloat64m8_t vfsgnj(vfloat64m8_t op0, vfloat64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vv_f64m8_m))) +vfloat64m8_t vfsgnj(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, vfloat64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vf_f64m1))) +vfloat64m1_t vfsgnj(vfloat64m1_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vf_f64m1_m))) +vfloat64m1_t vfsgnj(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vf_f64m2))) +vfloat64m2_t vfsgnj(vfloat64m2_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vf_f64m2_m))) +vfloat64m2_t vfsgnj(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vf_f64m4))) +vfloat64m4_t vfsgnj(vfloat64m4_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vf_f64m4_m))) +vfloat64m4_t vfsgnj(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vf_f64m8))) +vfloat64m8_t vfsgnj(vfloat64m8_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vf_f64m8_m))) +vfloat64m8_t vfsgnj(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vv_f64m1))) +vfloat64m1_t vfsgnjn(vfloat64m1_t op0, vfloat64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vv_f64m1_m))) +vfloat64m1_t vfsgnjn(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vv_f64m2))) +vfloat64m2_t vfsgnjn(vfloat64m2_t op0, vfloat64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vv_f64m2_m))) +vfloat64m2_t vfsgnjn(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vv_f64m4))) +vfloat64m4_t vfsgnjn(vfloat64m4_t op0, vfloat64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vv_f64m4_m))) +vfloat64m4_t vfsgnjn(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vv_f64m8))) +vfloat64m8_t vfsgnjn(vfloat64m8_t op0, vfloat64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vv_f64m8_m))) +vfloat64m8_t vfsgnjn(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, vfloat64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vf_f64m1))) +vfloat64m1_t vfsgnjn(vfloat64m1_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vf_f64m1_m))) +vfloat64m1_t vfsgnjn(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vf_f64m2))) +vfloat64m2_t vfsgnjn(vfloat64m2_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vf_f64m2_m))) +vfloat64m2_t vfsgnjn(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vf_f64m4))) +vfloat64m4_t vfsgnjn(vfloat64m4_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vf_f64m4_m))) +vfloat64m4_t vfsgnjn(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vf_f64m8))) +vfloat64m8_t vfsgnjn(vfloat64m8_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vf_f64m8_m))) +vfloat64m8_t vfsgnjn(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vv_f64m1))) +vfloat64m1_t vfsgnjx(vfloat64m1_t op0, vfloat64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vv_f64m1_m))) +vfloat64m1_t vfsgnjx(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vv_f64m2))) +vfloat64m2_t vfsgnjx(vfloat64m2_t op0, vfloat64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vv_f64m2_m))) +vfloat64m2_t vfsgnjx(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vv_f64m4))) +vfloat64m4_t vfsgnjx(vfloat64m4_t op0, vfloat64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vv_f64m4_m))) +vfloat64m4_t vfsgnjx(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vv_f64m8))) +vfloat64m8_t vfsgnjx(vfloat64m8_t op0, vfloat64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vv_f64m8_m))) +vfloat64m8_t vfsgnjx(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, vfloat64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vf_f64m1))) +vfloat64m1_t vfsgnjx(vfloat64m1_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vf_f64m1_m))) +vfloat64m1_t vfsgnjx(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vf_f64m2))) +vfloat64m2_t vfsgnjx(vfloat64m2_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vf_f64m2_m))) +vfloat64m2_t vfsgnjx(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vf_f64m4))) +vfloat64m4_t vfsgnjx(vfloat64m4_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vf_f64m4_m))) +vfloat64m4_t vfsgnjx(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vf_f64m8))) +vfloat64m8_t vfsgnjx(vfloat64m8_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vf_f64m8_m))) +vfloat64m8_t vfsgnjx(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfabs_v_f64m1))) +vfloat64m1_t vfabs(vfloat64m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfabs_v_f64m1_m))) +vfloat64m1_t vfabs(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfabs_v_f64m2))) +vfloat64m2_t vfabs(vfloat64m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfabs_v_f64m2_m))) +vfloat64m2_t vfabs(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfabs_v_f64m4))) +vfloat64m4_t vfabs(vfloat64m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfabs_v_f64m4_m))) +vfloat64m4_t vfabs(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfabs_v_f64m8))) +vfloat64m8_t vfabs(vfloat64m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfabs_v_f64m8_m))) +vfloat64m8_t vfabs(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vv_f64m1_b64))) +vbool64_t vmfeq(vfloat64m1_t op0, vfloat64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vv_f64m1_b64_m))) +vbool64_t vmfeq(vbool64_t op0, vbool64_t op1, vfloat64m1_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vv_f64m2_b32))) +vbool32_t vmfeq(vfloat64m2_t op0, vfloat64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vv_f64m2_b32_m))) +vbool32_t vmfeq(vbool32_t op0, vbool32_t op1, vfloat64m2_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vv_f64m4_b16))) +vbool16_t vmfeq(vfloat64m4_t op0, vfloat64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vv_f64m4_b16_m))) +vbool16_t vmfeq(vbool16_t op0, vbool16_t op1, vfloat64m4_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vv_f64m8_b8))) +vbool8_t vmfeq(vfloat64m8_t op0, vfloat64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vv_f64m8_b8_m))) +vbool8_t vmfeq(vbool8_t op0, vbool8_t op1, vfloat64m8_t op2, vfloat64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vf_f64m1_b64))) +vbool64_t vmfeq(vfloat64m1_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vf_f64m1_b64_m))) +vbool64_t vmfeq(vbool64_t op0, vbool64_t op1, vfloat64m1_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vf_f64m2_b32))) +vbool32_t vmfeq(vfloat64m2_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vf_f64m2_b32_m))) +vbool32_t vmfeq(vbool32_t op0, vbool32_t op1, vfloat64m2_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vf_f64m4_b16))) +vbool16_t vmfeq(vfloat64m4_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vf_f64m4_b16_m))) +vbool16_t vmfeq(vbool16_t op0, vbool16_t op1, vfloat64m4_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vf_f64m8_b8))) +vbool8_t vmfeq(vfloat64m8_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vf_f64m8_b8_m))) +vbool8_t vmfeq(vbool8_t op0, vbool8_t op1, vfloat64m8_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vv_f64m1_b64))) +vbool64_t vmfne(vfloat64m1_t op0, vfloat64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vv_f64m1_b64_m))) +vbool64_t vmfne(vbool64_t op0, vbool64_t op1, vfloat64m1_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vv_f64m2_b32))) +vbool32_t vmfne(vfloat64m2_t op0, vfloat64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vv_f64m2_b32_m))) +vbool32_t vmfne(vbool32_t op0, vbool32_t op1, vfloat64m2_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vv_f64m4_b16))) +vbool16_t vmfne(vfloat64m4_t op0, vfloat64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vv_f64m4_b16_m))) +vbool16_t vmfne(vbool16_t op0, vbool16_t op1, vfloat64m4_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vv_f64m8_b8))) +vbool8_t vmfne(vfloat64m8_t op0, vfloat64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vv_f64m8_b8_m))) +vbool8_t vmfne(vbool8_t op0, vbool8_t op1, vfloat64m8_t op2, vfloat64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vf_f64m1_b64))) +vbool64_t vmfne(vfloat64m1_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vf_f64m1_b64_m))) +vbool64_t vmfne(vbool64_t op0, vbool64_t op1, vfloat64m1_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vf_f64m2_b32))) +vbool32_t vmfne(vfloat64m2_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vf_f64m2_b32_m))) +vbool32_t vmfne(vbool32_t op0, vbool32_t op1, vfloat64m2_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vf_f64m4_b16))) +vbool16_t vmfne(vfloat64m4_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vf_f64m4_b16_m))) +vbool16_t vmfne(vbool16_t op0, vbool16_t op1, vfloat64m4_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vf_f64m8_b8))) +vbool8_t vmfne(vfloat64m8_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vf_f64m8_b8_m))) +vbool8_t vmfne(vbool8_t op0, vbool8_t op1, vfloat64m8_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vv_f64m1_b64))) +vbool64_t vmflt(vfloat64m1_t op0, vfloat64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vv_f64m1_b64_m))) +vbool64_t vmflt(vbool64_t op0, vbool64_t op1, vfloat64m1_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vv_f64m2_b32))) +vbool32_t vmflt(vfloat64m2_t op0, vfloat64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vv_f64m2_b32_m))) +vbool32_t vmflt(vbool32_t op0, vbool32_t op1, vfloat64m2_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vv_f64m4_b16))) +vbool16_t vmflt(vfloat64m4_t op0, vfloat64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vv_f64m4_b16_m))) +vbool16_t vmflt(vbool16_t op0, vbool16_t op1, vfloat64m4_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vv_f64m8_b8))) +vbool8_t vmflt(vfloat64m8_t op0, vfloat64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vv_f64m8_b8_m))) +vbool8_t vmflt(vbool8_t op0, vbool8_t op1, vfloat64m8_t op2, vfloat64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vf_f64m1_b64))) +vbool64_t vmflt(vfloat64m1_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vf_f64m1_b64_m))) +vbool64_t vmflt(vbool64_t op0, vbool64_t op1, vfloat64m1_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vf_f64m2_b32))) +vbool32_t vmflt(vfloat64m2_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vf_f64m2_b32_m))) +vbool32_t vmflt(vbool32_t op0, vbool32_t op1, vfloat64m2_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vf_f64m4_b16))) +vbool16_t vmflt(vfloat64m4_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vf_f64m4_b16_m))) +vbool16_t vmflt(vbool16_t op0, vbool16_t op1, vfloat64m4_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vf_f64m8_b8))) +vbool8_t vmflt(vfloat64m8_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vf_f64m8_b8_m))) +vbool8_t vmflt(vbool8_t op0, vbool8_t op1, vfloat64m8_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vv_f64m1_b64))) +vbool64_t vmfle(vfloat64m1_t op0, vfloat64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vv_f64m1_b64_m))) +vbool64_t vmfle(vbool64_t op0, vbool64_t op1, vfloat64m1_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vv_f64m2_b32))) +vbool32_t vmfle(vfloat64m2_t op0, vfloat64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vv_f64m2_b32_m))) +vbool32_t vmfle(vbool32_t op0, vbool32_t op1, vfloat64m2_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vv_f64m4_b16))) +vbool16_t vmfle(vfloat64m4_t op0, vfloat64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vv_f64m4_b16_m))) +vbool16_t vmfle(vbool16_t op0, vbool16_t op1, vfloat64m4_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vv_f64m8_b8))) +vbool8_t vmfle(vfloat64m8_t op0, vfloat64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vv_f64m8_b8_m))) +vbool8_t vmfle(vbool8_t op0, vbool8_t op1, vfloat64m8_t op2, vfloat64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vf_f64m1_b64))) +vbool64_t vmfle(vfloat64m1_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vf_f64m1_b64_m))) +vbool64_t vmfle(vbool64_t op0, vbool64_t op1, vfloat64m1_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vf_f64m2_b32))) +vbool32_t vmfle(vfloat64m2_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vf_f64m2_b32_m))) +vbool32_t vmfle(vbool32_t op0, vbool32_t op1, vfloat64m2_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vf_f64m4_b16))) +vbool16_t vmfle(vfloat64m4_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vf_f64m4_b16_m))) +vbool16_t vmfle(vbool16_t op0, vbool16_t op1, vfloat64m4_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vf_f64m8_b8))) +vbool8_t vmfle(vfloat64m8_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vf_f64m8_b8_m))) +vbool8_t vmfle(vbool8_t op0, vbool8_t op1, vfloat64m8_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vv_f64m1_b64))) +vbool64_t vmfgt(vfloat64m1_t op0, vfloat64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vv_f64m1_b64_m))) +vbool64_t vmfgt(vbool64_t op0, vbool64_t op1, vfloat64m1_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vv_f64m2_b32))) +vbool32_t vmfgt(vfloat64m2_t op0, vfloat64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vv_f64m2_b32_m))) +vbool32_t vmfgt(vbool32_t op0, vbool32_t op1, vfloat64m2_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vv_f64m4_b16))) +vbool16_t vmfgt(vfloat64m4_t op0, vfloat64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vv_f64m4_b16_m))) +vbool16_t vmfgt(vbool16_t op0, vbool16_t op1, vfloat64m4_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vv_f64m8_b8))) +vbool8_t vmfgt(vfloat64m8_t op0, vfloat64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vv_f64m8_b8_m))) +vbool8_t vmfgt(vbool8_t op0, vbool8_t op1, vfloat64m8_t op2, vfloat64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vf_f64m1_b64))) +vbool64_t vmfgt(vfloat64m1_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vf_f64m1_b64_m))) +vbool64_t vmfgt(vbool64_t op0, vbool64_t op1, vfloat64m1_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vf_f64m2_b32))) +vbool32_t vmfgt(vfloat64m2_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vf_f64m2_b32_m))) +vbool32_t vmfgt(vbool32_t op0, vbool32_t op1, vfloat64m2_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vf_f64m4_b16))) +vbool16_t vmfgt(vfloat64m4_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vf_f64m4_b16_m))) +vbool16_t vmfgt(vbool16_t op0, vbool16_t op1, vfloat64m4_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vf_f64m8_b8))) +vbool8_t vmfgt(vfloat64m8_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vf_f64m8_b8_m))) +vbool8_t vmfgt(vbool8_t op0, vbool8_t op1, vfloat64m8_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vv_f64m1_b64))) +vbool64_t vmfge(vfloat64m1_t op0, vfloat64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vv_f64m1_b64_m))) +vbool64_t vmfge(vbool64_t op0, vbool64_t op1, vfloat64m1_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vv_f64m2_b32))) +vbool32_t vmfge(vfloat64m2_t op0, vfloat64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vv_f64m2_b32_m))) +vbool32_t vmfge(vbool32_t op0, vbool32_t op1, vfloat64m2_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vv_f64m4_b16))) +vbool16_t vmfge(vfloat64m4_t op0, vfloat64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vv_f64m4_b16_m))) +vbool16_t vmfge(vbool16_t op0, vbool16_t op1, vfloat64m4_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vv_f64m8_b8))) +vbool8_t vmfge(vfloat64m8_t op0, vfloat64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vv_f64m8_b8_m))) +vbool8_t vmfge(vbool8_t op0, vbool8_t op1, vfloat64m8_t op2, vfloat64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vf_f64m1_b64))) +vbool64_t vmfge(vfloat64m1_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vf_f64m1_b64_m))) +vbool64_t vmfge(vbool64_t op0, vbool64_t op1, vfloat64m1_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vf_f64m2_b32))) +vbool32_t vmfge(vfloat64m2_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vf_f64m2_b32_m))) +vbool32_t vmfge(vbool32_t op0, vbool32_t op1, vfloat64m2_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vf_f64m4_b16))) +vbool16_t vmfge(vfloat64m4_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vf_f64m4_b16_m))) +vbool16_t vmfge(vbool16_t op0, vbool16_t op1, vfloat64m4_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vf_f64m8_b8))) +vbool8_t vmfge(vfloat64m8_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vf_f64m8_b8_m))) +vbool8_t vmfge(vbool8_t op0, vbool8_t op1, vfloat64m8_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_f64m1))) +vfloat64m1_t vmerge(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_f64m2))) +vfloat64m2_t vmerge(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_f64m4))) +vfloat64m4_t vmerge(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_f64m8))) +vfloat64m8_t vmerge(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmerge_vfm_f64m1))) +vfloat64m1_t vfmerge(vbool64_t op0, vfloat64m1_t op1, double op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmerge_vfm_f64m2))) +vfloat64m2_t vfmerge(vbool32_t op0, vfloat64m2_t op1, double op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmerge_vfm_f64m4))) +vfloat64m4_t vfmerge(vbool16_t op0, vfloat64m4_t op1, double op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmerge_vfm_f64m8))) +vfloat64m8_t vfmerge(vbool8_t op0, vfloat64m8_t op1, double op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmax_vs_f64m1_f64m1))) +vfloat64m1_t vfredmax(vfloat64m1_t op0, vfloat64m1_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmax_vs_f64m1_f64m1_m))) +vfloat64m1_t vfredmax(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmax_vs_f64m2_f64m1))) +vfloat64m1_t vfredmax(vfloat64m1_t op0, vfloat64m2_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmax_vs_f64m2_f64m1_m))) +vfloat64m1_t vfredmax(vbool32_t op0, vfloat64m1_t op1, vfloat64m2_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmax_vs_f64m4_f64m1))) +vfloat64m1_t vfredmax(vfloat64m1_t op0, vfloat64m4_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmax_vs_f64m4_f64m1_m))) +vfloat64m1_t vfredmax(vbool16_t op0, vfloat64m1_t op1, vfloat64m4_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmax_vs_f64m8_f64m1))) +vfloat64m1_t vfredmax(vfloat64m1_t op0, vfloat64m8_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmax_vs_f64m8_f64m1_m))) +vfloat64m1_t vfredmax(vbool8_t op0, vfloat64m1_t op1, vfloat64m8_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmin_vs_f64m1_f64m1))) +vfloat64m1_t vfredmin(vfloat64m1_t op0, vfloat64m1_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmin_vs_f64m1_f64m1_m))) +vfloat64m1_t vfredmin(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmin_vs_f64m2_f64m1))) +vfloat64m1_t vfredmin(vfloat64m1_t op0, vfloat64m2_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmin_vs_f64m2_f64m1_m))) +vfloat64m1_t vfredmin(vbool32_t op0, vfloat64m1_t op1, vfloat64m2_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmin_vs_f64m4_f64m1))) +vfloat64m1_t vfredmin(vfloat64m1_t op0, vfloat64m4_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmin_vs_f64m4_f64m1_m))) +vfloat64m1_t vfredmin(vbool16_t op0, vfloat64m1_t op1, vfloat64m4_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmin_vs_f64m8_f64m1))) +vfloat64m1_t vfredmin(vfloat64m1_t op0, vfloat64m8_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmin_vs_f64m8_f64m1_m))) +vfloat64m1_t vfredmin(vbool8_t op0, vfloat64m1_t op1, vfloat64m8_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredsum_vs_f64m1_f64m1))) +vfloat64m1_t vfredsum(vfloat64m1_t op0, vfloat64m1_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredsum_vs_f64m1_f64m1_m))) +vfloat64m1_t vfredsum(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredsum_vs_f64m2_f64m1))) +vfloat64m1_t vfredsum(vfloat64m1_t op0, vfloat64m2_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredsum_vs_f64m2_f64m1_m))) +vfloat64m1_t vfredsum(vbool32_t op0, vfloat64m1_t op1, vfloat64m2_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredsum_vs_f64m4_f64m1))) +vfloat64m1_t vfredsum(vfloat64m1_t op0, vfloat64m4_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredsum_vs_f64m4_f64m1_m))) +vfloat64m1_t vfredsum(vbool16_t op0, vfloat64m1_t op1, vfloat64m4_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredsum_vs_f64m8_f64m1))) +vfloat64m1_t vfredsum(vfloat64m1_t op0, vfloat64m8_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredsum_vs_f64m8_f64m1_m))) +vfloat64m1_t vfredsum(vbool8_t op0, vfloat64m1_t op1, vfloat64m8_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredosum_vs_f64m1_f64m1))) +vfloat64m1_t vfredosum(vfloat64m1_t op0, vfloat64m1_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredosum_vs_f64m1_f64m1_m))) +vfloat64m1_t vfredosum(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredosum_vs_f64m2_f64m1))) +vfloat64m1_t vfredosum(vfloat64m1_t op0, vfloat64m2_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredosum_vs_f64m2_f64m1_m))) +vfloat64m1_t vfredosum(vbool32_t op0, vfloat64m1_t op1, vfloat64m2_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredosum_vs_f64m4_f64m1))) +vfloat64m1_t vfredosum(vfloat64m1_t op0, vfloat64m4_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredosum_vs_f64m4_f64m1_m))) +vfloat64m1_t vfredosum(vbool16_t op0, vfloat64m1_t op1, vfloat64m4_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredosum_vs_f64m8_f64m1))) +vfloat64m1_t vfredosum(vfloat64m1_t op0, vfloat64m8_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredosum_vs_f64m8_f64m1_m))) +vfloat64m1_t vfredosum(vbool8_t op0, vfloat64m1_t op1, vfloat64m8_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmv_f_s_f64m1_f64))) +double vfmv_f(vfloat64m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmv_f_s_f64m2_f64))) +double vfmv_f(vfloat64m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmv_f_s_f64m4_f64))) +double vfmv_f(vfloat64m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmv_f_s_f64m8_f64))) +double vfmv_f(vfloat64m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmv_s_f_f64m1))) +vfloat64m1_t vfmv_s(vfloat64m1_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmv_s_f_f64m2))) +vfloat64m2_t vfmv_s(vfloat64m2_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmv_s_f_f64m4))) +vfloat64m4_t vfmv_s(vfloat64m4_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmv_s_f_f64m8))) +vfloat64m8_t vfmv_s(vfloat64m8_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_f64m1))) +vfloat64m1_t vslideup(vfloat64m1_t op0, vfloat64m1_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_f64m1_m))) +vfloat64m1_t vslideup(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_f64m2))) +vfloat64m2_t vslideup(vfloat64m2_t op0, vfloat64m2_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_f64m2_m))) +vfloat64m2_t vslideup(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_f64m4))) +vfloat64m4_t vslideup(vfloat64m4_t op0, vfloat64m4_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_f64m4_m))) +vfloat64m4_t vslideup(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_f64m8))) +vfloat64m8_t vslideup(vfloat64m8_t op0, vfloat64m8_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_f64m8_m))) +vfloat64m8_t vslideup(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_f64m1))) +vfloat64m1_t vslidedown(vfloat64m1_t op0, vfloat64m1_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_f64m1_m))) +vfloat64m1_t vslidedown(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_f64m2))) +vfloat64m2_t vslidedown(vfloat64m2_t op0, vfloat64m2_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_f64m2_m))) +vfloat64m2_t vslidedown(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_f64m4))) +vfloat64m4_t vslidedown(vfloat64m4_t op0, vfloat64m4_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_f64m4_m))) +vfloat64m4_t vslidedown(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_f64m8))) +vfloat64m8_t vslidedown(vfloat64m8_t op0, vfloat64m8_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_f64m8_m))) +vfloat64m8_t vslidedown(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1up_vf_f64m1))) +vfloat64m1_t vfslide1up(vfloat64m1_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1up_vf_f64m1_m))) +vfloat64m1_t vfslide1up(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1up_vf_f64m2))) +vfloat64m2_t vfslide1up(vfloat64m2_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1up_vf_f64m2_m))) +vfloat64m2_t vfslide1up(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1up_vf_f64m4))) +vfloat64m4_t vfslide1up(vfloat64m4_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1up_vf_f64m4_m))) +vfloat64m4_t vfslide1up(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1up_vf_f64m8))) +vfloat64m8_t vfslide1up(vfloat64m8_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1up_vf_f64m8_m))) +vfloat64m8_t vfslide1up(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1down_vf_f64m1))) +vfloat64m1_t vfslide1down(vfloat64m1_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1down_vf_f64m1_m))) +vfloat64m1_t vfslide1down(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1down_vf_f64m2))) +vfloat64m2_t vfslide1down(vfloat64m2_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1down_vf_f64m2_m))) +vfloat64m2_t vfslide1down(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1down_vf_f64m4))) +vfloat64m4_t vfslide1down(vfloat64m4_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1down_vf_f64m4_m))) +vfloat64m4_t vfslide1down(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1down_vf_f64m8))) +vfloat64m8_t vfslide1down(vfloat64m8_t op0, double op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1down_vf_f64m8_m))) +vfloat64m8_t vfslide1down(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, double op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_f64m1))) +vfloat64m1_t vrgather(vfloat64m1_t op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_f64m1_m))) +vfloat64m1_t vrgather(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_f64m2))) +vfloat64m2_t vrgather(vfloat64m2_t op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_f64m2_m))) +vfloat64m2_t vrgather(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_f64m4))) +vfloat64m4_t vrgather(vfloat64m4_t op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_f64m4_m))) +vfloat64m4_t vrgather(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_f64m8))) +vfloat64m8_t vrgather(vfloat64m8_t op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_f64m8_m))) +vfloat64m8_t vrgather(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_f64m1))) +vfloat64m1_t vrgather(vfloat64m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_f64m1_m))) +vfloat64m1_t vrgather(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_f64m2))) +vfloat64m2_t vrgather(vfloat64m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_f64m2_m))) +vfloat64m2_t vrgather(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_f64m4))) +vfloat64m4_t vrgather(vfloat64m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_f64m4_m))) +vfloat64m4_t vrgather(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_f64m8))) +vfloat64m8_t vrgather(vfloat64m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_f64m8_m))) +vfloat64m8_t vrgather(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_f64m1))) +vfloat64m1_t vrgatherei16(vfloat64m1_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_f64m1_m))) +vfloat64m1_t vrgatherei16(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_f64m2))) +vfloat64m2_t vrgatherei16(vfloat64m2_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_f64m2_m))) +vfloat64m2_t vrgatherei16(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_f64m4))) +vfloat64m4_t vrgatherei16(vfloat64m4_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_f64m4_m))) +vfloat64m4_t vrgatherei16(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_f64m8))) +vfloat64m8_t vrgatherei16(vfloat64m8_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_f64m8_m))) +vfloat64m8_t vrgatherei16(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_f64m1))) +vfloat64m1_t vcompress(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_f64m2))) +vfloat64m2_t vcompress(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_f64m4))) +vfloat64m4_t vcompress(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_f64m8))) +vfloat64m8_t vcompress(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_f64m2_f64m1))) +vfloat64m1_t vget_f64m1(vfloat64m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_f64m4_f64m1))) +vfloat64m1_t vget_f64m1(vfloat64m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_f64m8_f64m1))) +vfloat64m1_t vget_f64m1(vfloat64m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_f64m4_f64m2))) +vfloat64m2_t vget_f64m2(vfloat64m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_f64m8_f64m2))) +vfloat64m2_t vget_f64m2(vfloat64m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vget_v_f64m8_f64m4))) +vfloat64m4_t vget_f64m4(vfloat64m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_f64m1_f64m2))) +vfloat64m2_t vset(vfloat64m2_t op0, size_t op1, vfloat64m1_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_f64m1_f64m4))) +vfloat64m4_t vset(vfloat64m4_t op0, size_t op1, vfloat64m1_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_f64m2_f64m4))) +vfloat64m4_t vset(vfloat64m4_t op0, size_t op1, vfloat64m2_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_f64m1_f64m8))) +vfloat64m8_t vset(vfloat64m8_t op0, size_t op1, vfloat64m1_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_f64m2_f64m8))) +vfloat64m8_t vset(vfloat64m8_t op0, size_t op1, vfloat64m2_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vset_v_f64m4_f64m8))) +vfloat64m8_t vset(vfloat64m8_t op0, size_t op1, vfloat64m4_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_f64m1))) +void vsuxei8(double * op0, vuint8mf8_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_f64m1_m))) +void vsuxei8(vbool64_t op0, double * op1, vuint8mf8_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_f64m2))) +void vsuxei8(double * op0, vuint8mf4_t op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_f64m2_m))) +void vsuxei8(vbool32_t op0, double * op1, vuint8mf4_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_f64m4))) +void vsuxei8(double * op0, vuint8mf2_t op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_f64m4_m))) +void vsuxei8(vbool16_t op0, double * op1, vuint8mf2_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_f64m8))) +void vsuxei8(double * op0, vuint8m1_t op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_f64m8_m))) +void vsuxei8(vbool8_t op0, double * op1, vuint8m1_t op2, vfloat64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_f64m1))) +void vsuxei16(double * op0, vuint16mf4_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_f64m1_m))) +void vsuxei16(vbool64_t op0, double * op1, vuint16mf4_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_f64m2))) +void vsuxei16(double * op0, vuint16mf2_t op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_f64m2_m))) +void vsuxei16(vbool32_t op0, double * op1, vuint16mf2_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_f64m4))) +void vsuxei16(double * op0, vuint16m1_t op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_f64m4_m))) +void vsuxei16(vbool16_t op0, double * op1, vuint16m1_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_f64m8))) +void vsuxei16(double * op0, vuint16m2_t op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_f64m8_m))) +void vsuxei16(vbool8_t op0, double * op1, vuint16m2_t op2, vfloat64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_f64m1))) +void vsuxei32(double * op0, vuint32mf2_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_f64m1_m))) +void vsuxei32(vbool64_t op0, double * op1, vuint32mf2_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_f64m2))) +void vsuxei32(double * op0, vuint32m1_t op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_f64m2_m))) +void vsuxei32(vbool32_t op0, double * op1, vuint32m1_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_f64m4))) +void vsuxei32(double * op0, vuint32m2_t op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_f64m4_m))) +void vsuxei32(vbool16_t op0, double * op1, vuint32m2_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_f64m8))) +void vsuxei32(double * op0, vuint32m4_t op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_f64m8_m))) +void vsuxei32(vbool8_t op0, double * op1, vuint32m4_t op2, vfloat64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_f64m1))) +void vsuxei64(double * op0, vuint64m1_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_f64m1_m))) +void vsuxei64(vbool64_t op0, double * op1, vuint64m1_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_f64m2))) +void vsuxei64(double * op0, vuint64m2_t op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_f64m2_m))) +void vsuxei64(vbool32_t op0, double * op1, vuint64m2_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_f64m4))) +void vsuxei64(double * op0, vuint64m4_t op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_f64m4_m))) +void vsuxei64(vbool16_t op0, double * op1, vuint64m4_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_f64m8))) +void vsuxei64(double * op0, vuint64m8_t op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_f64m8_m))) +void vsuxei64(vbool8_t op0, double * op1, vuint64m8_t op2, vfloat64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_f64m1))) +void vsoxei8(double * op0, vuint8mf8_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_f64m1_m))) +void vsoxei8(vbool64_t op0, double * op1, vuint8mf8_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_f64m2))) +void vsoxei8(double * op0, vuint8mf4_t op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_f64m2_m))) +void vsoxei8(vbool32_t op0, double * op1, vuint8mf4_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_f64m4))) +void vsoxei8(double * op0, vuint8mf2_t op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_f64m4_m))) +void vsoxei8(vbool16_t op0, double * op1, vuint8mf2_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_f64m8))) +void vsoxei8(double * op0, vuint8m1_t op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_f64m8_m))) +void vsoxei8(vbool8_t op0, double * op1, vuint8m1_t op2, vfloat64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_f64m1))) +void vsoxei16(double * op0, vuint16mf4_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_f64m1_m))) +void vsoxei16(vbool64_t op0, double * op1, vuint16mf4_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_f64m2))) +void vsoxei16(double * op0, vuint16mf2_t op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_f64m2_m))) +void vsoxei16(vbool32_t op0, double * op1, vuint16mf2_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_f64m4))) +void vsoxei16(double * op0, vuint16m1_t op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_f64m4_m))) +void vsoxei16(vbool16_t op0, double * op1, vuint16m1_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_f64m8))) +void vsoxei16(double * op0, vuint16m2_t op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_f64m8_m))) +void vsoxei16(vbool8_t op0, double * op1, vuint16m2_t op2, vfloat64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_f64m1))) +void vsoxei32(double * op0, vuint32mf2_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_f64m1_m))) +void vsoxei32(vbool64_t op0, double * op1, vuint32mf2_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_f64m2))) +void vsoxei32(double * op0, vuint32m1_t op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_f64m2_m))) +void vsoxei32(vbool32_t op0, double * op1, vuint32m1_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_f64m4))) +void vsoxei32(double * op0, vuint32m2_t op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_f64m4_m))) +void vsoxei32(vbool16_t op0, double * op1, vuint32m2_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_f64m8))) +void vsoxei32(double * op0, vuint32m4_t op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_f64m8_m))) +void vsoxei32(vbool8_t op0, double * op1, vuint32m4_t op2, vfloat64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_f64m1))) +void vsoxei64(double * op0, vuint64m1_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_f64m1_m))) +void vsoxei64(vbool64_t op0, double * op1, vuint64m1_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_f64m2))) +void vsoxei64(double * op0, vuint64m2_t op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_f64m2_m))) +void vsoxei64(vbool32_t op0, double * op1, vuint64m2_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_f64m4))) +void vsoxei64(double * op0, vuint64m4_t op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_f64m4_m))) +void vsoxei64(vbool16_t op0, double * op1, vuint64m4_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_f64m8))) +void vsoxei64(double * op0, vuint64m8_t op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_f64m8_m))) +void vsoxei64(vbool8_t op0, double * op1, vuint64m8_t op2, vfloat64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle64ff_v_f64m1_m))) +vfloat64m1_t vle64ff(vbool64_t op0, vfloat64m1_t op1, const double * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle64ff_v_f64m2_m))) +vfloat64m2_t vle64ff(vbool32_t op0, vfloat64m2_t op1, const double * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle64ff_v_f64m4_m))) +vfloat64m4_t vle64ff(vbool16_t op0, vfloat64m4_t op1, const double * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle64ff_v_f64m8_m))) +vfloat64m8_t vle64ff(vbool8_t op0, vfloat64m8_t op1, const double * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfneg_v_f64m1))) +vfloat64m1_t vfneg(vfloat64m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfneg_v_f64m1_m))) +vfloat64m1_t vfneg(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfneg_v_f64m2))) +vfloat64m2_t vfneg(vfloat64m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfneg_v_f64m2_m))) +vfloat64m2_t vfneg(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfneg_v_f64m4))) +vfloat64m4_t vfneg(vfloat64m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfneg_v_f64m4_m))) +vfloat64m4_t vfneg(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfneg_v_f64m8))) +vfloat64m8_t vfneg(vfloat64m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfneg_v_f64m8_m))) +vfloat64m8_t vfneg(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle64_v_f64m1_m))) +vfloat64m1_t vle64(vbool64_t op0, vfloat64m1_t op1, const double * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle64_v_f64m2_m))) +vfloat64m2_t vle64(vbool32_t op0, vfloat64m2_t op1, const double * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle64_v_f64m4_m))) +vfloat64m4_t vle64(vbool16_t op0, vfloat64m4_t op1, const double * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle64_v_f64m8_m))) +vfloat64m8_t vle64(vbool8_t op0, vfloat64m8_t op1, const double * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse64_v_f64m1))) +void vse64(double * op0, vfloat64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse64_v_f64m1_m))) +void vse64(vbool64_t op0, double * op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse64_v_f64m2))) +void vse64(double * op0, vfloat64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse64_v_f64m2_m))) +void vse64(vbool32_t op0, double * op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse64_v_f64m4))) +void vse64(double * op0, vfloat64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse64_v_f64m4_m))) +void vse64(vbool16_t op0, double * op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse64_v_f64m8))) +void vse64(double * op0, vfloat64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse64_v_f64m8_m))) +void vse64(vbool8_t op0, double * op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfclass_v_u64m1))) +vuint64m1_t vfclass(vfloat64m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfclass_v_u64m1_m))) +vuint64m1_t vfclass(vbool64_t op0, vuint64m1_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfclass_v_u64m2))) +vuint64m2_t vfclass(vfloat64m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfclass_v_u64m2_m))) +vuint64m2_t vfclass(vbool32_t op0, vuint64m2_t op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfclass_v_u64m4))) +vuint64m4_t vfclass(vfloat64m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfclass_v_u64m4_m))) +vuint64m4_t vfclass(vbool16_t op0, vuint64m4_t op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfclass_v_u64m8))) +vuint64m8_t vfclass(vfloat64m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfclass_v_u64m8_m))) +vuint64m8_t vfclass(vbool8_t op0, vuint64m8_t op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_x_v_f64m1))) +vfloat64m1_t vfcvt_f(vint64m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_x_v_f64m1_m))) +vfloat64m1_t vfcvt_f(vbool64_t op0, vfloat64m1_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_x_v_f64m2))) +vfloat64m2_t vfcvt_f(vint64m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_x_v_f64m2_m))) +vfloat64m2_t vfcvt_f(vbool32_t op0, vfloat64m2_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_x_v_f64m4))) +vfloat64m4_t vfcvt_f(vint64m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_x_v_f64m4_m))) +vfloat64m4_t vfcvt_f(vbool16_t op0, vfloat64m4_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_x_v_f64m8))) +vfloat64m8_t vfcvt_f(vint64m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_x_v_f64m8_m))) +vfloat64m8_t vfcvt_f(vbool8_t op0, vfloat64m8_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_xu_v_f64m1))) +vfloat64m1_t vfcvt_f(vuint64m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_xu_v_f64m1_m))) +vfloat64m1_t vfcvt_f(vbool64_t op0, vfloat64m1_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_xu_v_f64m2))) +vfloat64m2_t vfcvt_f(vuint64m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_xu_v_f64m2_m))) +vfloat64m2_t vfcvt_f(vbool32_t op0, vfloat64m2_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_xu_v_f64m4))) +vfloat64m4_t vfcvt_f(vuint64m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_xu_v_f64m4_m))) +vfloat64m4_t vfcvt_f(vbool16_t op0, vfloat64m4_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_xu_v_f64m8))) +vfloat64m8_t vfcvt_f(vuint64m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_xu_v_f64m8_m))) +vfloat64m8_t vfcvt_f(vbool8_t op0, vfloat64m8_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_x_f_v_i64m1))) +vint64m1_t vfcvt_rtz_x(vfloat64m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_x_f_v_i64m1_m))) +vint64m1_t vfcvt_rtz_x(vbool64_t op0, vint64m1_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_x_f_v_i64m2))) +vint64m2_t vfcvt_rtz_x(vfloat64m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_x_f_v_i64m2_m))) +vint64m2_t vfcvt_rtz_x(vbool32_t op0, vint64m2_t op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_x_f_v_i64m4))) +vint64m4_t vfcvt_rtz_x(vfloat64m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_x_f_v_i64m4_m))) +vint64m4_t vfcvt_rtz_x(vbool16_t op0, vint64m4_t op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_x_f_v_i64m8))) +vint64m8_t vfcvt_rtz_x(vfloat64m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_x_f_v_i64m8_m))) +vint64m8_t vfcvt_rtz_x(vbool8_t op0, vint64m8_t op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_xu_f_v_u64m1))) +vuint64m1_t vfcvt_rtz_xu(vfloat64m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_xu_f_v_u64m1_m))) +vuint64m1_t vfcvt_rtz_xu(vbool64_t op0, vuint64m1_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_xu_f_v_u64m2))) +vuint64m2_t vfcvt_rtz_xu(vfloat64m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_xu_f_v_u64m2_m))) +vuint64m2_t vfcvt_rtz_xu(vbool32_t op0, vuint64m2_t op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_xu_f_v_u64m4))) +vuint64m4_t vfcvt_rtz_xu(vfloat64m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_xu_f_v_u64m4_m))) +vuint64m4_t vfcvt_rtz_xu(vbool16_t op0, vuint64m4_t op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_xu_f_v_u64m8))) +vuint64m8_t vfcvt_rtz_xu(vfloat64m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_xu_f_v_u64m8_m))) +vuint64m8_t vfcvt_rtz_xu(vbool8_t op0, vuint64m8_t op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_x_f_v_i64m1))) +vint64m1_t vfcvt_x(vfloat64m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_x_f_v_i64m1_m))) +vint64m1_t vfcvt_x(vbool64_t op0, vint64m1_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_x_f_v_i64m2))) +vint64m2_t vfcvt_x(vfloat64m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_x_f_v_i64m2_m))) +vint64m2_t vfcvt_x(vbool32_t op0, vint64m2_t op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_x_f_v_i64m4))) +vint64m4_t vfcvt_x(vfloat64m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_x_f_v_i64m4_m))) +vint64m4_t vfcvt_x(vbool16_t op0, vint64m4_t op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_x_f_v_i64m8))) +vint64m8_t vfcvt_x(vfloat64m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_x_f_v_i64m8_m))) +vint64m8_t vfcvt_x(vbool8_t op0, vint64m8_t op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_xu_f_v_u64m1))) +vuint64m1_t vfcvt_xu(vfloat64m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_xu_f_v_u64m1_m))) +vuint64m1_t vfcvt_xu(vbool64_t op0, vuint64m1_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_xu_f_v_u64m2))) +vuint64m2_t vfcvt_xu(vfloat64m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_xu_f_v_u64m2_m))) +vuint64m2_t vfcvt_xu(vbool32_t op0, vuint64m2_t op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_xu_f_v_u64m4))) +vuint64m4_t vfcvt_xu(vfloat64m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_xu_f_v_u64m4_m))) +vuint64m4_t vfcvt_xu(vbool16_t op0, vuint64m4_t op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_xu_f_v_u64m8))) +vuint64m8_t vfcvt_xu(vfloat64m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_xu_f_v_u64m8_m))) +vuint64m8_t vfcvt_xu(vbool8_t op0, vuint64m8_t op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_x_f_w_i32mf2))) +vint32mf2_t vfncvt_rtz_x(vfloat64m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_x_f_w_i32mf2_m))) +vint32mf2_t vfncvt_rtz_x(vbool64_t op0, vint32mf2_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_x_f_w_i32m1))) +vint32m1_t vfncvt_rtz_x(vfloat64m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_x_f_w_i32m1_m))) +vint32m1_t vfncvt_rtz_x(vbool32_t op0, vint32m1_t op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_x_f_w_i32m2))) +vint32m2_t vfncvt_rtz_x(vfloat64m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_x_f_w_i32m2_m))) +vint32m2_t vfncvt_rtz_x(vbool16_t op0, vint32m2_t op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_x_f_w_i32m4))) +vint32m4_t vfncvt_rtz_x(vfloat64m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_x_f_w_i32m4_m))) +vint32m4_t vfncvt_rtz_x(vbool8_t op0, vint32m4_t op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_xu_f_w_u32mf2))) +vuint32mf2_t vfncvt_rtz_xu(vfloat64m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_xu_f_w_u32mf2_m))) +vuint32mf2_t vfncvt_rtz_xu(vbool64_t op0, vuint32mf2_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_xu_f_w_u32m1))) +vuint32m1_t vfncvt_rtz_xu(vfloat64m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_xu_f_w_u32m1_m))) +vuint32m1_t vfncvt_rtz_xu(vbool32_t op0, vuint32m1_t op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_xu_f_w_u32m2))) +vuint32m2_t vfncvt_rtz_xu(vfloat64m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_xu_f_w_u32m2_m))) +vuint32m2_t vfncvt_rtz_xu(vbool16_t op0, vuint32m2_t op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_xu_f_w_u32m4))) +vuint32m4_t vfncvt_rtz_xu(vfloat64m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_xu_f_w_u32m4_m))) +vuint32m4_t vfncvt_rtz_xu(vbool8_t op0, vuint32m4_t op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_x_f_w_i32mf2))) +vint32mf2_t vfncvt_x(vfloat64m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_x_f_w_i32mf2_m))) +vint32mf2_t vfncvt_x(vbool64_t op0, vint32mf2_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_x_f_w_i32m1))) +vint32m1_t vfncvt_x(vfloat64m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_x_f_w_i32m1_m))) +vint32m1_t vfncvt_x(vbool32_t op0, vint32m1_t op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_x_f_w_i32m2))) +vint32m2_t vfncvt_x(vfloat64m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_x_f_w_i32m2_m))) +vint32m2_t vfncvt_x(vbool16_t op0, vint32m2_t op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_x_f_w_i32m4))) +vint32m4_t vfncvt_x(vfloat64m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_x_f_w_i32m4_m))) +vint32m4_t vfncvt_x(vbool8_t op0, vint32m4_t op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_xu_f_w_u32mf2))) +vuint32mf2_t vfncvt_xu(vfloat64m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_xu_f_w_u32mf2_m))) +vuint32mf2_t vfncvt_xu(vbool64_t op0, vuint32mf2_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_xu_f_w_u32m1))) +vuint32m1_t vfncvt_xu(vfloat64m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_xu_f_w_u32m1_m))) +vuint32m1_t vfncvt_xu(vbool32_t op0, vuint32m1_t op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_xu_f_w_u32m2))) +vuint32m2_t vfncvt_xu(vfloat64m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_xu_f_w_u32m2_m))) +vuint32m2_t vfncvt_xu(vbool16_t op0, vuint32m2_t op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_xu_f_w_u32m4))) +vuint32m4_t vfncvt_xu(vfloat64m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_xu_f_w_u32m4_m))) +vuint32m4_t vfncvt_xu(vbool8_t op0, vuint32m4_t op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrec7_v_f64m1))) +vfloat64m1_t vfrec7(vfloat64m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrec7_v_f64m1_m))) +vfloat64m1_t vfrec7(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrec7_v_f64m2))) +vfloat64m2_t vfrec7(vfloat64m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrec7_v_f64m2_m))) +vfloat64m2_t vfrec7(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrec7_v_f64m4))) +vfloat64m4_t vfrec7(vfloat64m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrec7_v_f64m4_m))) +vfloat64m4_t vfrec7(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrec7_v_f64m8))) +vfloat64m8_t vfrec7(vfloat64m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrec7_v_f64m8_m))) +vfloat64m8_t vfrec7(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsqrt7_v_f64m1))) +vfloat64m1_t vfrsqrt7(vfloat64m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsqrt7_v_f64m1_m))) +vfloat64m1_t vfrsqrt7(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsqrt7_v_f64m2))) +vfloat64m2_t vfrsqrt7(vfloat64m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsqrt7_v_f64m2_m))) +vfloat64m2_t vfrsqrt7(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsqrt7_v_f64m4))) +vfloat64m4_t vfrsqrt7(vfloat64m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsqrt7_v_f64m4_m))) +vfloat64m4_t vfrsqrt7(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsqrt7_v_f64m8))) +vfloat64m8_t vfrsqrt7(vfloat64m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsqrt7_v_f64m8_m))) +vfloat64m8_t vfrsqrt7(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsqrt_v_f64m1))) +vfloat64m1_t vfsqrt(vfloat64m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsqrt_v_f64m1_m))) +vfloat64m1_t vfsqrt(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsqrt_v_f64m2))) +vfloat64m2_t vfsqrt(vfloat64m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsqrt_v_f64m2_m))) +vfloat64m2_t vfsqrt(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsqrt_v_f64m4))) +vfloat64m4_t vfsqrt(vfloat64m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsqrt_v_f64m4_m))) +vfloat64m4_t vfsqrt(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsqrt_v_f64m8))) +vfloat64m8_t vfsqrt(vfloat64m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsqrt_v_f64m8_m))) +vfloat64m8_t vfsqrt(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_x_v_f64m1))) +vfloat64m1_t vfwcvt_f(vint32mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_x_v_f64m1_m))) +vfloat64m1_t vfwcvt_f(vbool64_t op0, vfloat64m1_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_x_v_f64m2))) +vfloat64m2_t vfwcvt_f(vint32m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_x_v_f64m2_m))) +vfloat64m2_t vfwcvt_f(vbool32_t op0, vfloat64m2_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_x_v_f64m4))) +vfloat64m4_t vfwcvt_f(vint32m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_x_v_f64m4_m))) +vfloat64m4_t vfwcvt_f(vbool16_t op0, vfloat64m4_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_x_v_f64m8))) +vfloat64m8_t vfwcvt_f(vint32m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_x_v_f64m8_m))) +vfloat64m8_t vfwcvt_f(vbool8_t op0, vfloat64m8_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_xu_v_f64m1))) +vfloat64m1_t vfwcvt_f(vuint32mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_xu_v_f64m1_m))) +vfloat64m1_t vfwcvt_f(vbool64_t op0, vfloat64m1_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_xu_v_f64m2))) +vfloat64m2_t vfwcvt_f(vuint32m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_xu_v_f64m2_m))) +vfloat64m2_t vfwcvt_f(vbool32_t op0, vfloat64m2_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_xu_v_f64m4))) +vfloat64m4_t vfwcvt_f(vuint32m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_xu_v_f64m4_m))) +vfloat64m4_t vfwcvt_f(vbool16_t op0, vfloat64m4_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_xu_v_f64m8))) +vfloat64m8_t vfwcvt_f(vuint32m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_xu_v_f64m8_m))) +vfloat64m8_t vfwcvt_f(vbool8_t op0, vfloat64m8_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_f64m1_f64m2))) +vfloat64m2_t vlmul_ext_f64m2(vfloat64m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_f64m1_f64m4))) +vfloat64m4_t vlmul_ext_f64m4(vfloat64m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_f64m2_f64m4))) +vfloat64m4_t vlmul_ext_f64m4(vfloat64m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_f64m1_f64m8))) +vfloat64m8_t vlmul_ext_f64m8(vfloat64m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_f64m2_f64m8))) +vfloat64m8_t vlmul_ext_f64m8(vfloat64m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_f64m4_f64m8))) +vfloat64m8_t vlmul_ext_f64m8(vfloat64m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_f64m2_f64m1))) +vfloat64m1_t vlmul_trunc_f64m1(vfloat64m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_f64m4_f64m1))) +vfloat64m1_t vlmul_trunc_f64m1(vfloat64m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_f64m8_f64m1))) +vfloat64m1_t vlmul_trunc_f64m1(vfloat64m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_f64m4_f64m2))) +vfloat64m2_t vlmul_trunc_f64m2(vfloat64m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_f64m8_f64m2))) +vfloat64m2_t vlmul_trunc_f64m2(vfloat64m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_f64m8_f64m4))) +vfloat64m4_t vlmul_trunc_f64m4(vfloat64m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i64m1_f64m1))) +vfloat64m1_t vreinterpret_f64m1(vint64m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i64m2_f64m2))) +vfloat64m2_t vreinterpret_f64m2(vint64m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i64m4_f64m4))) +vfloat64m4_t vreinterpret_f64m4(vint64m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i64m8_f64m8))) +vfloat64m8_t vreinterpret_f64m8(vint64m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u64m1_f64m1))) +vfloat64m1_t vreinterpret_f64m1(vuint64m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u64m2_f64m2))) +vfloat64m2_t vreinterpret_f64m2(vuint64m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u64m4_f64m4))) +vfloat64m4_t vreinterpret_f64m4(vuint64m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u64m8_f64m8))) +vfloat64m8_t vreinterpret_f64m8(vuint64m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_f64m1_i64m1))) +vint64m1_t vreinterpret_i64m1(vfloat64m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_f64m2_i64m2))) +vint64m2_t vreinterpret_i64m2(vfloat64m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_f64m4_i64m4))) +vint64m4_t vreinterpret_i64m4(vfloat64m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_f64m8_i64m8))) +vint64m8_t vreinterpret_i64m8(vfloat64m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_f64m1_u64m1))) +vuint64m1_t vreinterpret_u64m1(vfloat64m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_f64m2_u64m2))) +vuint64m2_t vreinterpret_u64m2(vfloat64m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_f64m4_u64m4))) +vuint64m4_t vreinterpret_u64m4(vfloat64m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_f64m8_u64m8))) +vuint64m8_t vreinterpret_u64m8(vfloat64m8_t op0); + +#endif + +#if defined(__riscv_f) && defined(__riscv_d) +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_vv_f64m1))) +vfloat64m1_t vfwadd_vv(vfloat32mf2_t op0, vfloat32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_vv_f64m1_m))) +vfloat64m1_t vfwadd_vv(vbool64_t op0, vfloat64m1_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_vv_f64m2))) +vfloat64m2_t vfwadd_vv(vfloat32m1_t op0, vfloat32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_vv_f64m2_m))) +vfloat64m2_t vfwadd_vv(vbool32_t op0, vfloat64m2_t op1, vfloat32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_vv_f64m4))) +vfloat64m4_t vfwadd_vv(vfloat32m2_t op0, vfloat32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_vv_f64m4_m))) +vfloat64m4_t vfwadd_vv(vbool16_t op0, vfloat64m4_t op1, vfloat32m2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_vv_f64m8))) +vfloat64m8_t vfwadd_vv(vfloat32m4_t op0, vfloat32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_vv_f64m8_m))) +vfloat64m8_t vfwadd_vv(vbool8_t op0, vfloat64m8_t op1, vfloat32m4_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_vf_f64m1))) +vfloat64m1_t vfwadd_vf(vfloat32mf2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_vf_f64m1_m))) +vfloat64m1_t vfwadd_vf(vbool64_t op0, vfloat64m1_t op1, vfloat32mf2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_vf_f64m2))) +vfloat64m2_t vfwadd_vf(vfloat32m1_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_vf_f64m2_m))) +vfloat64m2_t vfwadd_vf(vbool32_t op0, vfloat64m2_t op1, vfloat32m1_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_vf_f64m4))) +vfloat64m4_t vfwadd_vf(vfloat32m2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_vf_f64m4_m))) +vfloat64m4_t vfwadd_vf(vbool16_t op0, vfloat64m4_t op1, vfloat32m2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_vf_f64m8))) +vfloat64m8_t vfwadd_vf(vfloat32m4_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_vf_f64m8_m))) +vfloat64m8_t vfwadd_vf(vbool8_t op0, vfloat64m8_t op1, vfloat32m4_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_vv_f64m1))) +vfloat64m1_t vfwsub_vv(vfloat32mf2_t op0, vfloat32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_vv_f64m1_m))) +vfloat64m1_t vfwsub_vv(vbool64_t op0, vfloat64m1_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_vv_f64m2))) +vfloat64m2_t vfwsub_vv(vfloat32m1_t op0, vfloat32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_vv_f64m2_m))) +vfloat64m2_t vfwsub_vv(vbool32_t op0, vfloat64m2_t op1, vfloat32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_vv_f64m4))) +vfloat64m4_t vfwsub_vv(vfloat32m2_t op0, vfloat32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_vv_f64m4_m))) +vfloat64m4_t vfwsub_vv(vbool16_t op0, vfloat64m4_t op1, vfloat32m2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_vv_f64m8))) +vfloat64m8_t vfwsub_vv(vfloat32m4_t op0, vfloat32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_vv_f64m8_m))) +vfloat64m8_t vfwsub_vv(vbool8_t op0, vfloat64m8_t op1, vfloat32m4_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_vf_f64m1))) +vfloat64m1_t vfwsub_vf(vfloat32mf2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_vf_f64m1_m))) +vfloat64m1_t vfwsub_vf(vbool64_t op0, vfloat64m1_t op1, vfloat32mf2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_vf_f64m2))) +vfloat64m2_t vfwsub_vf(vfloat32m1_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_vf_f64m2_m))) +vfloat64m2_t vfwsub_vf(vbool32_t op0, vfloat64m2_t op1, vfloat32m1_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_vf_f64m4))) +vfloat64m4_t vfwsub_vf(vfloat32m2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_vf_f64m4_m))) +vfloat64m4_t vfwsub_vf(vbool16_t op0, vfloat64m4_t op1, vfloat32m2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_vf_f64m8))) +vfloat64m8_t vfwsub_vf(vfloat32m4_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_vf_f64m8_m))) +vfloat64m8_t vfwsub_vf(vbool8_t op0, vfloat64m8_t op1, vfloat32m4_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_wv_f64m1))) +vfloat64m1_t vfwadd_wv(vfloat64m1_t op0, vfloat32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_wv_f64m1_m))) +vfloat64m1_t vfwadd_wv(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_wv_f64m2))) +vfloat64m2_t vfwadd_wv(vfloat64m2_t op0, vfloat32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_wv_f64m2_m))) +vfloat64m2_t vfwadd_wv(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_wv_f64m4))) +vfloat64m4_t vfwadd_wv(vfloat64m4_t op0, vfloat32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_wv_f64m4_m))) +vfloat64m4_t vfwadd_wv(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_wv_f64m8))) +vfloat64m8_t vfwadd_wv(vfloat64m8_t op0, vfloat32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_wv_f64m8_m))) +vfloat64m8_t vfwadd_wv(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_wf_f64m1))) +vfloat64m1_t vfwadd_wf(vfloat64m1_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_wf_f64m1_m))) +vfloat64m1_t vfwadd_wf(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_wf_f64m2))) +vfloat64m2_t vfwadd_wf(vfloat64m2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_wf_f64m2_m))) +vfloat64m2_t vfwadd_wf(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_wf_f64m4))) +vfloat64m4_t vfwadd_wf(vfloat64m4_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_wf_f64m4_m))) +vfloat64m4_t vfwadd_wf(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_wf_f64m8))) +vfloat64m8_t vfwadd_wf(vfloat64m8_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_wf_f64m8_m))) +vfloat64m8_t vfwadd_wf(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_wv_f64m1))) +vfloat64m1_t vfwsub_wv(vfloat64m1_t op0, vfloat32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_wv_f64m1_m))) +vfloat64m1_t vfwsub_wv(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_wv_f64m2))) +vfloat64m2_t vfwsub_wv(vfloat64m2_t op0, vfloat32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_wv_f64m2_m))) +vfloat64m2_t vfwsub_wv(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_wv_f64m4))) +vfloat64m4_t vfwsub_wv(vfloat64m4_t op0, vfloat32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_wv_f64m4_m))) +vfloat64m4_t vfwsub_wv(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_wv_f64m8))) +vfloat64m8_t vfwsub_wv(vfloat64m8_t op0, vfloat32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_wv_f64m8_m))) +vfloat64m8_t vfwsub_wv(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_wf_f64m1))) +vfloat64m1_t vfwsub_wf(vfloat64m1_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_wf_f64m1_m))) +vfloat64m1_t vfwsub_wf(vbool64_t op0, vfloat64m1_t op1, vfloat64m1_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_wf_f64m2))) +vfloat64m2_t vfwsub_wf(vfloat64m2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_wf_f64m2_m))) +vfloat64m2_t vfwsub_wf(vbool32_t op0, vfloat64m2_t op1, vfloat64m2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_wf_f64m4))) +vfloat64m4_t vfwsub_wf(vfloat64m4_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_wf_f64m4_m))) +vfloat64m4_t vfwsub_wf(vbool16_t op0, vfloat64m4_t op1, vfloat64m4_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_wf_f64m8))) +vfloat64m8_t vfwsub_wf(vfloat64m8_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_wf_f64m8_m))) +vfloat64m8_t vfwsub_wf(vbool8_t op0, vfloat64m8_t op1, vfloat64m8_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmul_vv_f64m1))) +vfloat64m1_t vfwmul(vfloat32mf2_t op0, vfloat32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmul_vv_f64m1_m))) +vfloat64m1_t vfwmul(vbool64_t op0, vfloat64m1_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmul_vv_f64m2))) +vfloat64m2_t vfwmul(vfloat32m1_t op0, vfloat32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmul_vv_f64m2_m))) +vfloat64m2_t vfwmul(vbool32_t op0, vfloat64m2_t op1, vfloat32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmul_vv_f64m4))) +vfloat64m4_t vfwmul(vfloat32m2_t op0, vfloat32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmul_vv_f64m4_m))) +vfloat64m4_t vfwmul(vbool16_t op0, vfloat64m4_t op1, vfloat32m2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmul_vv_f64m8))) +vfloat64m8_t vfwmul(vfloat32m4_t op0, vfloat32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmul_vv_f64m8_m))) +vfloat64m8_t vfwmul(vbool8_t op0, vfloat64m8_t op1, vfloat32m4_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmul_vf_f64m1))) +vfloat64m1_t vfwmul(vfloat32mf2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmul_vf_f64m1_m))) +vfloat64m1_t vfwmul(vbool64_t op0, vfloat64m1_t op1, vfloat32mf2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmul_vf_f64m2))) +vfloat64m2_t vfwmul(vfloat32m1_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmul_vf_f64m2_m))) +vfloat64m2_t vfwmul(vbool32_t op0, vfloat64m2_t op1, vfloat32m1_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmul_vf_f64m4))) +vfloat64m4_t vfwmul(vfloat32m2_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmul_vf_f64m4_m))) +vfloat64m4_t vfwmul(vbool16_t op0, vfloat64m4_t op1, vfloat32m2_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmul_vf_f64m8))) +vfloat64m8_t vfwmul(vfloat32m4_t op0, float op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmul_vf_f64m8_m))) +vfloat64m8_t vfwmul(vbool8_t op0, vfloat64m8_t op1, vfloat32m4_t op2, float op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmacc_vv_f64m1))) +vfloat64m1_t vfwmacc(vfloat64m1_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmacc_vv_f64m1_m))) +vfloat64m1_t vfwmacc(vbool64_t op0, vfloat64m1_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmacc_vv_f64m2))) +vfloat64m2_t vfwmacc(vfloat64m2_t op0, vfloat32m1_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmacc_vv_f64m2_m))) +vfloat64m2_t vfwmacc(vbool32_t op0, vfloat64m2_t op1, vfloat32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmacc_vv_f64m4))) +vfloat64m4_t vfwmacc(vfloat64m4_t op0, vfloat32m2_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmacc_vv_f64m4_m))) +vfloat64m4_t vfwmacc(vbool16_t op0, vfloat64m4_t op1, vfloat32m2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmacc_vv_f64m8))) +vfloat64m8_t vfwmacc(vfloat64m8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmacc_vv_f64m8_m))) +vfloat64m8_t vfwmacc(vbool8_t op0, vfloat64m8_t op1, vfloat32m4_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmacc_vf_f64m1))) +vfloat64m1_t vfwmacc(vfloat64m1_t op0, float op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmacc_vf_f64m1_m))) +vfloat64m1_t vfwmacc(vbool64_t op0, vfloat64m1_t op1, float op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmacc_vf_f64m2))) +vfloat64m2_t vfwmacc(vfloat64m2_t op0, float op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmacc_vf_f64m2_m))) +vfloat64m2_t vfwmacc(vbool32_t op0, vfloat64m2_t op1, float op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmacc_vf_f64m4))) +vfloat64m4_t vfwmacc(vfloat64m4_t op0, float op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmacc_vf_f64m4_m))) +vfloat64m4_t vfwmacc(vbool16_t op0, vfloat64m4_t op1, float op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmacc_vf_f64m8))) +vfloat64m8_t vfwmacc(vfloat64m8_t op0, float op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmacc_vf_f64m8_m))) +vfloat64m8_t vfwmacc(vbool8_t op0, vfloat64m8_t op1, float op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmacc_vv_f64m1))) +vfloat64m1_t vfwnmacc(vfloat64m1_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmacc_vv_f64m1_m))) +vfloat64m1_t vfwnmacc(vbool64_t op0, vfloat64m1_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmacc_vv_f64m2))) +vfloat64m2_t vfwnmacc(vfloat64m2_t op0, vfloat32m1_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmacc_vv_f64m2_m))) +vfloat64m2_t vfwnmacc(vbool32_t op0, vfloat64m2_t op1, vfloat32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmacc_vv_f64m4))) +vfloat64m4_t vfwnmacc(vfloat64m4_t op0, vfloat32m2_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmacc_vv_f64m4_m))) +vfloat64m4_t vfwnmacc(vbool16_t op0, vfloat64m4_t op1, vfloat32m2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmacc_vv_f64m8))) +vfloat64m8_t vfwnmacc(vfloat64m8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmacc_vv_f64m8_m))) +vfloat64m8_t vfwnmacc(vbool8_t op0, vfloat64m8_t op1, vfloat32m4_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmacc_vf_f64m1))) +vfloat64m1_t vfwnmacc(vfloat64m1_t op0, float op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmacc_vf_f64m1_m))) +vfloat64m1_t vfwnmacc(vbool64_t op0, vfloat64m1_t op1, float op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmacc_vf_f64m2))) +vfloat64m2_t vfwnmacc(vfloat64m2_t op0, float op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmacc_vf_f64m2_m))) +vfloat64m2_t vfwnmacc(vbool32_t op0, vfloat64m2_t op1, float op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmacc_vf_f64m4))) +vfloat64m4_t vfwnmacc(vfloat64m4_t op0, float op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmacc_vf_f64m4_m))) +vfloat64m4_t vfwnmacc(vbool16_t op0, vfloat64m4_t op1, float op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmacc_vf_f64m8))) +vfloat64m8_t vfwnmacc(vfloat64m8_t op0, float op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmacc_vf_f64m8_m))) +vfloat64m8_t vfwnmacc(vbool8_t op0, vfloat64m8_t op1, float op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmsac_vv_f64m1))) +vfloat64m1_t vfwmsac(vfloat64m1_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmsac_vv_f64m1_m))) +vfloat64m1_t vfwmsac(vbool64_t op0, vfloat64m1_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmsac_vv_f64m2))) +vfloat64m2_t vfwmsac(vfloat64m2_t op0, vfloat32m1_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmsac_vv_f64m2_m))) +vfloat64m2_t vfwmsac(vbool32_t op0, vfloat64m2_t op1, vfloat32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmsac_vv_f64m4))) +vfloat64m4_t vfwmsac(vfloat64m4_t op0, vfloat32m2_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmsac_vv_f64m4_m))) +vfloat64m4_t vfwmsac(vbool16_t op0, vfloat64m4_t op1, vfloat32m2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmsac_vv_f64m8))) +vfloat64m8_t vfwmsac(vfloat64m8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmsac_vv_f64m8_m))) +vfloat64m8_t vfwmsac(vbool8_t op0, vfloat64m8_t op1, vfloat32m4_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmsac_vf_f64m1))) +vfloat64m1_t vfwmsac(vfloat64m1_t op0, float op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmsac_vf_f64m1_m))) +vfloat64m1_t vfwmsac(vbool64_t op0, vfloat64m1_t op1, float op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmsac_vf_f64m2))) +vfloat64m2_t vfwmsac(vfloat64m2_t op0, float op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmsac_vf_f64m2_m))) +vfloat64m2_t vfwmsac(vbool32_t op0, vfloat64m2_t op1, float op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmsac_vf_f64m4))) +vfloat64m4_t vfwmsac(vfloat64m4_t op0, float op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmsac_vf_f64m4_m))) +vfloat64m4_t vfwmsac(vbool16_t op0, vfloat64m4_t op1, float op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmsac_vf_f64m8))) +vfloat64m8_t vfwmsac(vfloat64m8_t op0, float op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmsac_vf_f64m8_m))) +vfloat64m8_t vfwmsac(vbool8_t op0, vfloat64m8_t op1, float op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmsac_vv_f64m1))) +vfloat64m1_t vfwnmsac(vfloat64m1_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmsac_vv_f64m1_m))) +vfloat64m1_t vfwnmsac(vbool64_t op0, vfloat64m1_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmsac_vv_f64m2))) +vfloat64m2_t vfwnmsac(vfloat64m2_t op0, vfloat32m1_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmsac_vv_f64m2_m))) +vfloat64m2_t vfwnmsac(vbool32_t op0, vfloat64m2_t op1, vfloat32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmsac_vv_f64m4))) +vfloat64m4_t vfwnmsac(vfloat64m4_t op0, vfloat32m2_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmsac_vv_f64m4_m))) +vfloat64m4_t vfwnmsac(vbool16_t op0, vfloat64m4_t op1, vfloat32m2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmsac_vv_f64m8))) +vfloat64m8_t vfwnmsac(vfloat64m8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmsac_vv_f64m8_m))) +vfloat64m8_t vfwnmsac(vbool8_t op0, vfloat64m8_t op1, vfloat32m4_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmsac_vf_f64m1))) +vfloat64m1_t vfwnmsac(vfloat64m1_t op0, float op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmsac_vf_f64m1_m))) +vfloat64m1_t vfwnmsac(vbool64_t op0, vfloat64m1_t op1, float op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmsac_vf_f64m2))) +vfloat64m2_t vfwnmsac(vfloat64m2_t op0, float op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmsac_vf_f64m2_m))) +vfloat64m2_t vfwnmsac(vbool32_t op0, vfloat64m2_t op1, float op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmsac_vf_f64m4))) +vfloat64m4_t vfwnmsac(vfloat64m4_t op0, float op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmsac_vf_f64m4_m))) +vfloat64m4_t vfwnmsac(vbool16_t op0, vfloat64m4_t op1, float op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmsac_vf_f64m8))) +vfloat64m8_t vfwnmsac(vfloat64m8_t op0, float op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmsac_vf_f64m8_m))) +vfloat64m8_t vfwnmsac(vbool8_t op0, vfloat64m8_t op1, float op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwredsum_vs_f32m1_f64m1))) +vfloat64m1_t vfwredsum(vfloat64m1_t op0, vfloat32m1_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwredsum_vs_f32m1_f64m1_m))) +vfloat64m1_t vfwredsum(vbool32_t op0, vfloat64m1_t op1, vfloat32m1_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwredsum_vs_f32m2_f64m1))) +vfloat64m1_t vfwredsum(vfloat64m1_t op0, vfloat32m2_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwredsum_vs_f32m2_f64m1_m))) +vfloat64m1_t vfwredsum(vbool16_t op0, vfloat64m1_t op1, vfloat32m2_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwredsum_vs_f32m4_f64m1))) +vfloat64m1_t vfwredsum(vfloat64m1_t op0, vfloat32m4_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwredsum_vs_f32m4_f64m1_m))) +vfloat64m1_t vfwredsum(vbool8_t op0, vfloat64m1_t op1, vfloat32m4_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwredsum_vs_f32m8_f64m1))) +vfloat64m1_t vfwredsum(vfloat64m1_t op0, vfloat32m8_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwredsum_vs_f32m8_f64m1_m))) +vfloat64m1_t vfwredsum(vbool4_t op0, vfloat64m1_t op1, vfloat32m8_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwredsum_vs_f32mf2_f64m1))) +vfloat64m1_t vfwredsum(vfloat64m1_t op0, vfloat32mf2_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwredsum_vs_f32mf2_f64m1_m))) +vfloat64m1_t vfwredsum(vbool64_t op0, vfloat64m1_t op1, vfloat32mf2_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwredosum_vs_f32m1_f64m1))) +vfloat64m1_t vfwredosum(vfloat64m1_t op0, vfloat32m1_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwredosum_vs_f32m1_f64m1_m))) +vfloat64m1_t vfwredosum(vbool32_t op0, vfloat64m1_t op1, vfloat32m1_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwredosum_vs_f32m2_f64m1))) +vfloat64m1_t vfwredosum(vfloat64m1_t op0, vfloat32m2_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwredosum_vs_f32m2_f64m1_m))) +vfloat64m1_t vfwredosum(vbool16_t op0, vfloat64m1_t op1, vfloat32m2_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwredosum_vs_f32m4_f64m1))) +vfloat64m1_t vfwredosum(vfloat64m1_t op0, vfloat32m4_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwredosum_vs_f32m4_f64m1_m))) +vfloat64m1_t vfwredosum(vbool8_t op0, vfloat64m1_t op1, vfloat32m4_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwredosum_vs_f32m8_f64m1))) +vfloat64m1_t vfwredosum(vfloat64m1_t op0, vfloat32m8_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwredosum_vs_f32m8_f64m1_m))) +vfloat64m1_t vfwredosum(vbool4_t op0, vfloat64m1_t op1, vfloat32m8_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwredosum_vs_f32mf2_f64m1))) +vfloat64m1_t vfwredosum(vfloat64m1_t op0, vfloat32mf2_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwredosum_vs_f32mf2_f64m1_m))) +vfloat64m1_t vfwredosum(vbool64_t op0, vfloat64m1_t op1, vfloat32mf2_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_f_w_f32mf2))) +vfloat32mf2_t vfncvt_f(vfloat64m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_f_w_f32mf2_m))) +vfloat32mf2_t vfncvt_f(vbool64_t op0, vfloat32mf2_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_f_w_f32m1))) +vfloat32m1_t vfncvt_f(vfloat64m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_f_w_f32m1_m))) +vfloat32m1_t vfncvt_f(vbool32_t op0, vfloat32m1_t op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_f_w_f32m2))) +vfloat32m2_t vfncvt_f(vfloat64m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_f_w_f32m2_m))) +vfloat32m2_t vfncvt_f(vbool16_t op0, vfloat32m2_t op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_f_w_f32m4))) +vfloat32m4_t vfncvt_f(vfloat64m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_f_w_f32m4_m))) +vfloat32m4_t vfncvt_f(vbool8_t op0, vfloat32m4_t op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rod_f_f_w_f32mf2))) +vfloat32mf2_t vfncvt_rod_f(vfloat64m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rod_f_f_w_f32mf2_m))) +vfloat32mf2_t vfncvt_rod_f(vbool64_t op0, vfloat32mf2_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rod_f_f_w_f32m1))) +vfloat32m1_t vfncvt_rod_f(vfloat64m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rod_f_f_w_f32m1_m))) +vfloat32m1_t vfncvt_rod_f(vbool32_t op0, vfloat32m1_t op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rod_f_f_w_f32m2))) +vfloat32m2_t vfncvt_rod_f(vfloat64m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rod_f_f_w_f32m2_m))) +vfloat32m2_t vfncvt_rod_f(vbool16_t op0, vfloat32m2_t op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rod_f_f_w_f32m4))) +vfloat32m4_t vfncvt_rod_f(vfloat64m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rod_f_f_w_f32m4_m))) +vfloat32m4_t vfncvt_rod_f(vbool8_t op0, vfloat32m4_t op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_f_v_f64m1))) +vfloat64m1_t vfwcvt_f(vfloat32mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_f_v_f64m1_m))) +vfloat64m1_t vfwcvt_f(vbool64_t op0, vfloat64m1_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_f_v_f64m2))) +vfloat64m2_t vfwcvt_f(vfloat32m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_f_v_f64m2_m))) +vfloat64m2_t vfwcvt_f(vbool32_t op0, vfloat64m2_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_f_v_f64m4))) +vfloat64m4_t vfwcvt_f(vfloat32m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_f_v_f64m4_m))) +vfloat64m4_t vfwcvt_f(vbool16_t op0, vfloat64m4_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_f_v_f64m8))) +vfloat64m8_t vfwcvt_f(vfloat32m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_f_v_f64m8_m))) +vfloat64m8_t vfwcvt_f(vbool8_t op0, vfloat64m8_t op1, vfloat32m4_t op2, size_t op3); + +#endif + +#if defined(__riscv_zfh) +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse16_v_f16m1))) +void vsse16(_Float16 * op0, ptrdiff_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse16_v_f16m1_m))) +void vsse16(vbool16_t op0, _Float16 * op1, ptrdiff_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse16_v_f16m2))) +void vsse16(_Float16 * op0, ptrdiff_t op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse16_v_f16m2_m))) +void vsse16(vbool8_t op0, _Float16 * op1, ptrdiff_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse16_v_f16m4))) +void vsse16(_Float16 * op0, ptrdiff_t op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse16_v_f16m4_m))) +void vsse16(vbool4_t op0, _Float16 * op1, ptrdiff_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse16_v_f16m8))) +void vsse16(_Float16 * op0, ptrdiff_t op1, vfloat16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse16_v_f16m8_m))) +void vsse16(vbool2_t op0, _Float16 * op1, ptrdiff_t op2, vfloat16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse16_v_f16mf2))) +void vsse16(_Float16 * op0, ptrdiff_t op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse16_v_f16mf2_m))) +void vsse16(vbool32_t op0, _Float16 * op1, ptrdiff_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse16_v_f16mf4))) +void vsse16(_Float16 * op0, ptrdiff_t op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsse16_v_f16mf4_m))) +void vsse16(vbool64_t op0, _Float16 * op1, ptrdiff_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_f16m1))) +vfloat16m1_t vluxei8(const _Float16 * op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_f16m1_m))) +vfloat16m1_t vluxei8(vbool16_t op0, vfloat16m1_t op1, const _Float16 * op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_f16m2))) +vfloat16m2_t vluxei8(const _Float16 * op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_f16m2_m))) +vfloat16m2_t vluxei8(vbool8_t op0, vfloat16m2_t op1, const _Float16 * op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_f16m4))) +vfloat16m4_t vluxei8(const _Float16 * op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_f16m4_m))) +vfloat16m4_t vluxei8(vbool4_t op0, vfloat16m4_t op1, const _Float16 * op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_f16m8))) +vfloat16m8_t vluxei8(const _Float16 * op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_f16m8_m))) +vfloat16m8_t vluxei8(vbool2_t op0, vfloat16m8_t op1, const _Float16 * op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_f16mf2))) +vfloat16mf2_t vluxei8(const _Float16 * op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_f16mf2_m))) +vfloat16mf2_t vluxei8(vbool32_t op0, vfloat16mf2_t op1, const _Float16 * op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_f16mf4))) +vfloat16mf4_t vluxei8(const _Float16 * op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei8_v_f16mf4_m))) +vfloat16mf4_t vluxei8(vbool64_t op0, vfloat16mf4_t op1, const _Float16 * op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_f16m1))) +vfloat16m1_t vluxei16(const _Float16 * op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_f16m1_m))) +vfloat16m1_t vluxei16(vbool16_t op0, vfloat16m1_t op1, const _Float16 * op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_f16m2))) +vfloat16m2_t vluxei16(const _Float16 * op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_f16m2_m))) +vfloat16m2_t vluxei16(vbool8_t op0, vfloat16m2_t op1, const _Float16 * op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_f16m4))) +vfloat16m4_t vluxei16(const _Float16 * op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_f16m4_m))) +vfloat16m4_t vluxei16(vbool4_t op0, vfloat16m4_t op1, const _Float16 * op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_f16m8))) +vfloat16m8_t vluxei16(const _Float16 * op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_f16m8_m))) +vfloat16m8_t vluxei16(vbool2_t op0, vfloat16m8_t op1, const _Float16 * op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_f16mf2))) +vfloat16mf2_t vluxei16(const _Float16 * op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_f16mf2_m))) +vfloat16mf2_t vluxei16(vbool32_t op0, vfloat16mf2_t op1, const _Float16 * op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_f16mf4))) +vfloat16mf4_t vluxei16(const _Float16 * op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei16_v_f16mf4_m))) +vfloat16mf4_t vluxei16(vbool64_t op0, vfloat16mf4_t op1, const _Float16 * op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_f16m1))) +vfloat16m1_t vluxei32(const _Float16 * op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_f16m1_m))) +vfloat16m1_t vluxei32(vbool16_t op0, vfloat16m1_t op1, const _Float16 * op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_f16m2))) +vfloat16m2_t vluxei32(const _Float16 * op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_f16m2_m))) +vfloat16m2_t vluxei32(vbool8_t op0, vfloat16m2_t op1, const _Float16 * op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_f16m4))) +vfloat16m4_t vluxei32(const _Float16 * op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_f16m4_m))) +vfloat16m4_t vluxei32(vbool4_t op0, vfloat16m4_t op1, const _Float16 * op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_f16mf2))) +vfloat16mf2_t vluxei32(const _Float16 * op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_f16mf2_m))) +vfloat16mf2_t vluxei32(vbool32_t op0, vfloat16mf2_t op1, const _Float16 * op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_f16mf4))) +vfloat16mf4_t vluxei32(const _Float16 * op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei32_v_f16mf4_m))) +vfloat16mf4_t vluxei32(vbool64_t op0, vfloat16mf4_t op1, const _Float16 * op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_f16m1))) +vfloat16m1_t vluxei64(const _Float16 * op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_f16m1_m))) +vfloat16m1_t vluxei64(vbool16_t op0, vfloat16m1_t op1, const _Float16 * op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_f16m2))) +vfloat16m2_t vluxei64(const _Float16 * op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_f16m2_m))) +vfloat16m2_t vluxei64(vbool8_t op0, vfloat16m2_t op1, const _Float16 * op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_f16mf2))) +vfloat16mf2_t vluxei64(const _Float16 * op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_f16mf2_m))) +vfloat16mf2_t vluxei64(vbool32_t op0, vfloat16mf2_t op1, const _Float16 * op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_f16mf4))) +vfloat16mf4_t vluxei64(const _Float16 * op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxei64_v_f16mf4_m))) +vfloat16mf4_t vluxei64(vbool64_t op0, vfloat16mf4_t op1, const _Float16 * op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_f16m1))) +vfloat16m1_t vloxei8(const _Float16 * op0, vuint8mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_f16m1_m))) +vfloat16m1_t vloxei8(vbool16_t op0, vfloat16m1_t op1, const _Float16 * op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_f16m2))) +vfloat16m2_t vloxei8(const _Float16 * op0, vuint8m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_f16m2_m))) +vfloat16m2_t vloxei8(vbool8_t op0, vfloat16m2_t op1, const _Float16 * op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_f16m4))) +vfloat16m4_t vloxei8(const _Float16 * op0, vuint8m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_f16m4_m))) +vfloat16m4_t vloxei8(vbool4_t op0, vfloat16m4_t op1, const _Float16 * op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_f16m8))) +vfloat16m8_t vloxei8(const _Float16 * op0, vuint8m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_f16m8_m))) +vfloat16m8_t vloxei8(vbool2_t op0, vfloat16m8_t op1, const _Float16 * op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_f16mf2))) +vfloat16mf2_t vloxei8(const _Float16 * op0, vuint8mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_f16mf2_m))) +vfloat16mf2_t vloxei8(vbool32_t op0, vfloat16mf2_t op1, const _Float16 * op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_f16mf4))) +vfloat16mf4_t vloxei8(const _Float16 * op0, vuint8mf8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei8_v_f16mf4_m))) +vfloat16mf4_t vloxei8(vbool64_t op0, vfloat16mf4_t op1, const _Float16 * op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_f16m1))) +vfloat16m1_t vloxei16(const _Float16 * op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_f16m1_m))) +vfloat16m1_t vloxei16(vbool16_t op0, vfloat16m1_t op1, const _Float16 * op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_f16m2))) +vfloat16m2_t vloxei16(const _Float16 * op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_f16m2_m))) +vfloat16m2_t vloxei16(vbool8_t op0, vfloat16m2_t op1, const _Float16 * op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_f16m4))) +vfloat16m4_t vloxei16(const _Float16 * op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_f16m4_m))) +vfloat16m4_t vloxei16(vbool4_t op0, vfloat16m4_t op1, const _Float16 * op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_f16m8))) +vfloat16m8_t vloxei16(const _Float16 * op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_f16m8_m))) +vfloat16m8_t vloxei16(vbool2_t op0, vfloat16m8_t op1, const _Float16 * op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_f16mf2))) +vfloat16mf2_t vloxei16(const _Float16 * op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_f16mf2_m))) +vfloat16mf2_t vloxei16(vbool32_t op0, vfloat16mf2_t op1, const _Float16 * op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_f16mf4))) +vfloat16mf4_t vloxei16(const _Float16 * op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei16_v_f16mf4_m))) +vfloat16mf4_t vloxei16(vbool64_t op0, vfloat16mf4_t op1, const _Float16 * op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_f16m1))) +vfloat16m1_t vloxei32(const _Float16 * op0, vuint32m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_f16m1_m))) +vfloat16m1_t vloxei32(vbool16_t op0, vfloat16m1_t op1, const _Float16 * op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_f16m2))) +vfloat16m2_t vloxei32(const _Float16 * op0, vuint32m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_f16m2_m))) +vfloat16m2_t vloxei32(vbool8_t op0, vfloat16m2_t op1, const _Float16 * op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_f16m4))) +vfloat16m4_t vloxei32(const _Float16 * op0, vuint32m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_f16m4_m))) +vfloat16m4_t vloxei32(vbool4_t op0, vfloat16m4_t op1, const _Float16 * op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_f16mf2))) +vfloat16mf2_t vloxei32(const _Float16 * op0, vuint32m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_f16mf2_m))) +vfloat16mf2_t vloxei32(vbool32_t op0, vfloat16mf2_t op1, const _Float16 * op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_f16mf4))) +vfloat16mf4_t vloxei32(const _Float16 * op0, vuint32mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei32_v_f16mf4_m))) +vfloat16mf4_t vloxei32(vbool64_t op0, vfloat16mf4_t op1, const _Float16 * op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_f16m1))) +vfloat16m1_t vloxei64(const _Float16 * op0, vuint64m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_f16m1_m))) +vfloat16m1_t vloxei64(vbool16_t op0, vfloat16m1_t op1, const _Float16 * op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_f16m2))) +vfloat16m2_t vloxei64(const _Float16 * op0, vuint64m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_f16m2_m))) +vfloat16m2_t vloxei64(vbool8_t op0, vfloat16m2_t op1, const _Float16 * op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_f16mf2))) +vfloat16mf2_t vloxei64(const _Float16 * op0, vuint64m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_f16mf2_m))) +vfloat16mf2_t vloxei64(vbool32_t op0, vfloat16mf2_t op1, const _Float16 * op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_f16mf4))) +vfloat16mf4_t vloxei64(const _Float16 * op0, vuint64m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxei64_v_f16mf4_m))) +vfloat16mf4_t vloxei64(vbool64_t op0, vfloat16mf4_t op1, const _Float16 * op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_f16m1))) +vfloat16m1_t vmv_v(vfloat16m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_f16m2))) +vfloat16m2_t vmv_v(vfloat16m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_f16m4))) +vfloat16m4_t vmv_v(vfloat16m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_f16m8))) +vfloat16m8_t vmv_v(vfloat16m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_f16mf2))) +vfloat16mf2_t vmv_v(vfloat16mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmv_v_v_f16mf4))) +vfloat16mf4_t vmv_v(vfloat16mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vv_f16m1))) +vfloat16m1_t vfadd(vfloat16m1_t op0, vfloat16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vv_f16m1_m))) +vfloat16m1_t vfadd(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vv_f16m2))) +vfloat16m2_t vfadd(vfloat16m2_t op0, vfloat16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vv_f16m2_m))) +vfloat16m2_t vfadd(vbool8_t op0, vfloat16m2_t op1, vfloat16m2_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vv_f16m4))) +vfloat16m4_t vfadd(vfloat16m4_t op0, vfloat16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vv_f16m4_m))) +vfloat16m4_t vfadd(vbool4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vv_f16m8))) +vfloat16m8_t vfadd(vfloat16m8_t op0, vfloat16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vv_f16m8_m))) +vfloat16m8_t vfadd(vbool2_t op0, vfloat16m8_t op1, vfloat16m8_t op2, vfloat16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vv_f16mf2))) +vfloat16mf2_t vfadd(vfloat16mf2_t op0, vfloat16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vv_f16mf2_m))) +vfloat16mf2_t vfadd(vbool32_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vv_f16mf4))) +vfloat16mf4_t vfadd(vfloat16mf4_t op0, vfloat16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vv_f16mf4_m))) +vfloat16mf4_t vfadd(vbool64_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vf_f16m1))) +vfloat16m1_t vfadd(vfloat16m1_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vf_f16m1_m))) +vfloat16m1_t vfadd(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vf_f16m2))) +vfloat16m2_t vfadd(vfloat16m2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vf_f16m2_m))) +vfloat16m2_t vfadd(vbool8_t op0, vfloat16m2_t op1, vfloat16m2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vf_f16m4))) +vfloat16m4_t vfadd(vfloat16m4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vf_f16m4_m))) +vfloat16m4_t vfadd(vbool4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vf_f16m8))) +vfloat16m8_t vfadd(vfloat16m8_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vf_f16m8_m))) +vfloat16m8_t vfadd(vbool2_t op0, vfloat16m8_t op1, vfloat16m8_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vf_f16mf2))) +vfloat16mf2_t vfadd(vfloat16mf2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vf_f16mf2_m))) +vfloat16mf2_t vfadd(vbool32_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vf_f16mf4))) +vfloat16mf4_t vfadd(vfloat16mf4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfadd_vf_f16mf4_m))) +vfloat16mf4_t vfadd(vbool64_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vv_f16m1))) +vfloat16m1_t vfsub(vfloat16m1_t op0, vfloat16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vv_f16m1_m))) +vfloat16m1_t vfsub(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vv_f16m2))) +vfloat16m2_t vfsub(vfloat16m2_t op0, vfloat16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vv_f16m2_m))) +vfloat16m2_t vfsub(vbool8_t op0, vfloat16m2_t op1, vfloat16m2_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vv_f16m4))) +vfloat16m4_t vfsub(vfloat16m4_t op0, vfloat16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vv_f16m4_m))) +vfloat16m4_t vfsub(vbool4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vv_f16m8))) +vfloat16m8_t vfsub(vfloat16m8_t op0, vfloat16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vv_f16m8_m))) +vfloat16m8_t vfsub(vbool2_t op0, vfloat16m8_t op1, vfloat16m8_t op2, vfloat16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vv_f16mf2))) +vfloat16mf2_t vfsub(vfloat16mf2_t op0, vfloat16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vv_f16mf2_m))) +vfloat16mf2_t vfsub(vbool32_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vv_f16mf4))) +vfloat16mf4_t vfsub(vfloat16mf4_t op0, vfloat16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vv_f16mf4_m))) +vfloat16mf4_t vfsub(vbool64_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vf_f16m1))) +vfloat16m1_t vfsub(vfloat16m1_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vf_f16m1_m))) +vfloat16m1_t vfsub(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vf_f16m2))) +vfloat16m2_t vfsub(vfloat16m2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vf_f16m2_m))) +vfloat16m2_t vfsub(vbool8_t op0, vfloat16m2_t op1, vfloat16m2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vf_f16m4))) +vfloat16m4_t vfsub(vfloat16m4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vf_f16m4_m))) +vfloat16m4_t vfsub(vbool4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vf_f16m8))) +vfloat16m8_t vfsub(vfloat16m8_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vf_f16m8_m))) +vfloat16m8_t vfsub(vbool2_t op0, vfloat16m8_t op1, vfloat16m8_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vf_f16mf2))) +vfloat16mf2_t vfsub(vfloat16mf2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vf_f16mf2_m))) +vfloat16mf2_t vfsub(vbool32_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vf_f16mf4))) +vfloat16mf4_t vfsub(vfloat16mf4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsub_vf_f16mf4_m))) +vfloat16mf4_t vfsub(vbool64_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsub_vf_f16m1))) +vfloat16m1_t vfrsub(vfloat16m1_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsub_vf_f16m1_m))) +vfloat16m1_t vfrsub(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsub_vf_f16m2))) +vfloat16m2_t vfrsub(vfloat16m2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsub_vf_f16m2_m))) +vfloat16m2_t vfrsub(vbool8_t op0, vfloat16m2_t op1, vfloat16m2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsub_vf_f16m4))) +vfloat16m4_t vfrsub(vfloat16m4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsub_vf_f16m4_m))) +vfloat16m4_t vfrsub(vbool4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsub_vf_f16m8))) +vfloat16m8_t vfrsub(vfloat16m8_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsub_vf_f16m8_m))) +vfloat16m8_t vfrsub(vbool2_t op0, vfloat16m8_t op1, vfloat16m8_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsub_vf_f16mf2))) +vfloat16mf2_t vfrsub(vfloat16mf2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsub_vf_f16mf2_m))) +vfloat16mf2_t vfrsub(vbool32_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsub_vf_f16mf4))) +vfloat16mf4_t vfrsub(vfloat16mf4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsub_vf_f16mf4_m))) +vfloat16mf4_t vfrsub(vbool64_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vv_f16m1))) +vfloat16m1_t vfmul(vfloat16m1_t op0, vfloat16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vv_f16m1_m))) +vfloat16m1_t vfmul(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vv_f16m2))) +vfloat16m2_t vfmul(vfloat16m2_t op0, vfloat16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vv_f16m2_m))) +vfloat16m2_t vfmul(vbool8_t op0, vfloat16m2_t op1, vfloat16m2_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vv_f16m4))) +vfloat16m4_t vfmul(vfloat16m4_t op0, vfloat16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vv_f16m4_m))) +vfloat16m4_t vfmul(vbool4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vv_f16m8))) +vfloat16m8_t vfmul(vfloat16m8_t op0, vfloat16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vv_f16m8_m))) +vfloat16m8_t vfmul(vbool2_t op0, vfloat16m8_t op1, vfloat16m8_t op2, vfloat16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vv_f16mf2))) +vfloat16mf2_t vfmul(vfloat16mf2_t op0, vfloat16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vv_f16mf2_m))) +vfloat16mf2_t vfmul(vbool32_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vv_f16mf4))) +vfloat16mf4_t vfmul(vfloat16mf4_t op0, vfloat16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vv_f16mf4_m))) +vfloat16mf4_t vfmul(vbool64_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vf_f16m1))) +vfloat16m1_t vfmul(vfloat16m1_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vf_f16m1_m))) +vfloat16m1_t vfmul(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vf_f16m2))) +vfloat16m2_t vfmul(vfloat16m2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vf_f16m2_m))) +vfloat16m2_t vfmul(vbool8_t op0, vfloat16m2_t op1, vfloat16m2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vf_f16m4))) +vfloat16m4_t vfmul(vfloat16m4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vf_f16m4_m))) +vfloat16m4_t vfmul(vbool4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vf_f16m8))) +vfloat16m8_t vfmul(vfloat16m8_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vf_f16m8_m))) +vfloat16m8_t vfmul(vbool2_t op0, vfloat16m8_t op1, vfloat16m8_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vf_f16mf2))) +vfloat16mf2_t vfmul(vfloat16mf2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vf_f16mf2_m))) +vfloat16mf2_t vfmul(vbool32_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vf_f16mf4))) +vfloat16mf4_t vfmul(vfloat16mf4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmul_vf_f16mf4_m))) +vfloat16mf4_t vfmul(vbool64_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vv_f16m1))) +vfloat16m1_t vfdiv(vfloat16m1_t op0, vfloat16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vv_f16m1_m))) +vfloat16m1_t vfdiv(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vv_f16m2))) +vfloat16m2_t vfdiv(vfloat16m2_t op0, vfloat16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vv_f16m2_m))) +vfloat16m2_t vfdiv(vbool8_t op0, vfloat16m2_t op1, vfloat16m2_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vv_f16m4))) +vfloat16m4_t vfdiv(vfloat16m4_t op0, vfloat16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vv_f16m4_m))) +vfloat16m4_t vfdiv(vbool4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vv_f16m8))) +vfloat16m8_t vfdiv(vfloat16m8_t op0, vfloat16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vv_f16m8_m))) +vfloat16m8_t vfdiv(vbool2_t op0, vfloat16m8_t op1, vfloat16m8_t op2, vfloat16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vv_f16mf2))) +vfloat16mf2_t vfdiv(vfloat16mf2_t op0, vfloat16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vv_f16mf2_m))) +vfloat16mf2_t vfdiv(vbool32_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vv_f16mf4))) +vfloat16mf4_t vfdiv(vfloat16mf4_t op0, vfloat16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vv_f16mf4_m))) +vfloat16mf4_t vfdiv(vbool64_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vf_f16m1))) +vfloat16m1_t vfdiv(vfloat16m1_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vf_f16m1_m))) +vfloat16m1_t vfdiv(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vf_f16m2))) +vfloat16m2_t vfdiv(vfloat16m2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vf_f16m2_m))) +vfloat16m2_t vfdiv(vbool8_t op0, vfloat16m2_t op1, vfloat16m2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vf_f16m4))) +vfloat16m4_t vfdiv(vfloat16m4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vf_f16m4_m))) +vfloat16m4_t vfdiv(vbool4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vf_f16m8))) +vfloat16m8_t vfdiv(vfloat16m8_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vf_f16m8_m))) +vfloat16m8_t vfdiv(vbool2_t op0, vfloat16m8_t op1, vfloat16m8_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vf_f16mf2))) +vfloat16mf2_t vfdiv(vfloat16mf2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vf_f16mf2_m))) +vfloat16mf2_t vfdiv(vbool32_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vf_f16mf4))) +vfloat16mf4_t vfdiv(vfloat16mf4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfdiv_vf_f16mf4_m))) +vfloat16mf4_t vfdiv(vbool64_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrdiv_vf_f16m1))) +vfloat16m1_t vfrdiv(vfloat16m1_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrdiv_vf_f16m1_m))) +vfloat16m1_t vfrdiv(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrdiv_vf_f16m2))) +vfloat16m2_t vfrdiv(vfloat16m2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrdiv_vf_f16m2_m))) +vfloat16m2_t vfrdiv(vbool8_t op0, vfloat16m2_t op1, vfloat16m2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrdiv_vf_f16m4))) +vfloat16m4_t vfrdiv(vfloat16m4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrdiv_vf_f16m4_m))) +vfloat16m4_t vfrdiv(vbool4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrdiv_vf_f16m8))) +vfloat16m8_t vfrdiv(vfloat16m8_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrdiv_vf_f16m8_m))) +vfloat16m8_t vfrdiv(vbool2_t op0, vfloat16m8_t op1, vfloat16m8_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrdiv_vf_f16mf2))) +vfloat16mf2_t vfrdiv(vfloat16mf2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrdiv_vf_f16mf2_m))) +vfloat16mf2_t vfrdiv(vbool32_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrdiv_vf_f16mf4))) +vfloat16mf4_t vfrdiv(vfloat16mf4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrdiv_vf_f16mf4_m))) +vfloat16mf4_t vfrdiv(vbool64_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vv_f16m1))) +vfloat16m1_t vfmacc(vfloat16m1_t op0, vfloat16m1_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vv_f16m1_m))) +vfloat16m1_t vfmacc(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vv_f16m2))) +vfloat16m2_t vfmacc(vfloat16m2_t op0, vfloat16m2_t op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vv_f16m2_m))) +vfloat16m2_t vfmacc(vbool8_t op0, vfloat16m2_t op1, vfloat16m2_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vv_f16m4))) +vfloat16m4_t vfmacc(vfloat16m4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vv_f16m4_m))) +vfloat16m4_t vfmacc(vbool4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vv_f16m8))) +vfloat16m8_t vfmacc(vfloat16m8_t op0, vfloat16m8_t op1, vfloat16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vv_f16m8_m))) +vfloat16m8_t vfmacc(vbool2_t op0, vfloat16m8_t op1, vfloat16m8_t op2, vfloat16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vv_f16mf2))) +vfloat16mf2_t vfmacc(vfloat16mf2_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vv_f16mf2_m))) +vfloat16mf2_t vfmacc(vbool32_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vv_f16mf4))) +vfloat16mf4_t vfmacc(vfloat16mf4_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vv_f16mf4_m))) +vfloat16mf4_t vfmacc(vbool64_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vf_f16m1))) +vfloat16m1_t vfmacc(vfloat16m1_t op0, _Float16 op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vf_f16m1_m))) +vfloat16m1_t vfmacc(vbool16_t op0, vfloat16m1_t op1, _Float16 op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vf_f16m2))) +vfloat16m2_t vfmacc(vfloat16m2_t op0, _Float16 op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vf_f16m2_m))) +vfloat16m2_t vfmacc(vbool8_t op0, vfloat16m2_t op1, _Float16 op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vf_f16m4))) +vfloat16m4_t vfmacc(vfloat16m4_t op0, _Float16 op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vf_f16m4_m))) +vfloat16m4_t vfmacc(vbool4_t op0, vfloat16m4_t op1, _Float16 op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vf_f16m8))) +vfloat16m8_t vfmacc(vfloat16m8_t op0, _Float16 op1, vfloat16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vf_f16m8_m))) +vfloat16m8_t vfmacc(vbool2_t op0, vfloat16m8_t op1, _Float16 op2, vfloat16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vf_f16mf2))) +vfloat16mf2_t vfmacc(vfloat16mf2_t op0, _Float16 op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vf_f16mf2_m))) +vfloat16mf2_t vfmacc(vbool32_t op0, vfloat16mf2_t op1, _Float16 op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vf_f16mf4))) +vfloat16mf4_t vfmacc(vfloat16mf4_t op0, _Float16 op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmacc_vf_f16mf4_m))) +vfloat16mf4_t vfmacc(vbool64_t op0, vfloat16mf4_t op1, _Float16 op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vv_f16m1))) +vfloat16m1_t vfnmacc(vfloat16m1_t op0, vfloat16m1_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vv_f16m1_m))) +vfloat16m1_t vfnmacc(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vv_f16m2))) +vfloat16m2_t vfnmacc(vfloat16m2_t op0, vfloat16m2_t op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vv_f16m2_m))) +vfloat16m2_t vfnmacc(vbool8_t op0, vfloat16m2_t op1, vfloat16m2_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vv_f16m4))) +vfloat16m4_t vfnmacc(vfloat16m4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vv_f16m4_m))) +vfloat16m4_t vfnmacc(vbool4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vv_f16m8))) +vfloat16m8_t vfnmacc(vfloat16m8_t op0, vfloat16m8_t op1, vfloat16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vv_f16m8_m))) +vfloat16m8_t vfnmacc(vbool2_t op0, vfloat16m8_t op1, vfloat16m8_t op2, vfloat16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vv_f16mf2))) +vfloat16mf2_t vfnmacc(vfloat16mf2_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vv_f16mf2_m))) +vfloat16mf2_t vfnmacc(vbool32_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vv_f16mf4))) +vfloat16mf4_t vfnmacc(vfloat16mf4_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vv_f16mf4_m))) +vfloat16mf4_t vfnmacc(vbool64_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vf_f16m1))) +vfloat16m1_t vfnmacc(vfloat16m1_t op0, _Float16 op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vf_f16m1_m))) +vfloat16m1_t vfnmacc(vbool16_t op0, vfloat16m1_t op1, _Float16 op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vf_f16m2))) +vfloat16m2_t vfnmacc(vfloat16m2_t op0, _Float16 op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vf_f16m2_m))) +vfloat16m2_t vfnmacc(vbool8_t op0, vfloat16m2_t op1, _Float16 op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vf_f16m4))) +vfloat16m4_t vfnmacc(vfloat16m4_t op0, _Float16 op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vf_f16m4_m))) +vfloat16m4_t vfnmacc(vbool4_t op0, vfloat16m4_t op1, _Float16 op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vf_f16m8))) +vfloat16m8_t vfnmacc(vfloat16m8_t op0, _Float16 op1, vfloat16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vf_f16m8_m))) +vfloat16m8_t vfnmacc(vbool2_t op0, vfloat16m8_t op1, _Float16 op2, vfloat16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vf_f16mf2))) +vfloat16mf2_t vfnmacc(vfloat16mf2_t op0, _Float16 op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vf_f16mf2_m))) +vfloat16mf2_t vfnmacc(vbool32_t op0, vfloat16mf2_t op1, _Float16 op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vf_f16mf4))) +vfloat16mf4_t vfnmacc(vfloat16mf4_t op0, _Float16 op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmacc_vf_f16mf4_m))) +vfloat16mf4_t vfnmacc(vbool64_t op0, vfloat16mf4_t op1, _Float16 op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vv_f16m1))) +vfloat16m1_t vfmsac(vfloat16m1_t op0, vfloat16m1_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vv_f16m1_m))) +vfloat16m1_t vfmsac(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vv_f16m2))) +vfloat16m2_t vfmsac(vfloat16m2_t op0, vfloat16m2_t op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vv_f16m2_m))) +vfloat16m2_t vfmsac(vbool8_t op0, vfloat16m2_t op1, vfloat16m2_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vv_f16m4))) +vfloat16m4_t vfmsac(vfloat16m4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vv_f16m4_m))) +vfloat16m4_t vfmsac(vbool4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vv_f16m8))) +vfloat16m8_t vfmsac(vfloat16m8_t op0, vfloat16m8_t op1, vfloat16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vv_f16m8_m))) +vfloat16m8_t vfmsac(vbool2_t op0, vfloat16m8_t op1, vfloat16m8_t op2, vfloat16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vv_f16mf2))) +vfloat16mf2_t vfmsac(vfloat16mf2_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vv_f16mf2_m))) +vfloat16mf2_t vfmsac(vbool32_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vv_f16mf4))) +vfloat16mf4_t vfmsac(vfloat16mf4_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vv_f16mf4_m))) +vfloat16mf4_t vfmsac(vbool64_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vf_f16m1))) +vfloat16m1_t vfmsac(vfloat16m1_t op0, _Float16 op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vf_f16m1_m))) +vfloat16m1_t vfmsac(vbool16_t op0, vfloat16m1_t op1, _Float16 op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vf_f16m2))) +vfloat16m2_t vfmsac(vfloat16m2_t op0, _Float16 op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vf_f16m2_m))) +vfloat16m2_t vfmsac(vbool8_t op0, vfloat16m2_t op1, _Float16 op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vf_f16m4))) +vfloat16m4_t vfmsac(vfloat16m4_t op0, _Float16 op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vf_f16m4_m))) +vfloat16m4_t vfmsac(vbool4_t op0, vfloat16m4_t op1, _Float16 op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vf_f16m8))) +vfloat16m8_t vfmsac(vfloat16m8_t op0, _Float16 op1, vfloat16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vf_f16m8_m))) +vfloat16m8_t vfmsac(vbool2_t op0, vfloat16m8_t op1, _Float16 op2, vfloat16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vf_f16mf2))) +vfloat16mf2_t vfmsac(vfloat16mf2_t op0, _Float16 op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vf_f16mf2_m))) +vfloat16mf2_t vfmsac(vbool32_t op0, vfloat16mf2_t op1, _Float16 op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vf_f16mf4))) +vfloat16mf4_t vfmsac(vfloat16mf4_t op0, _Float16 op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsac_vf_f16mf4_m))) +vfloat16mf4_t vfmsac(vbool64_t op0, vfloat16mf4_t op1, _Float16 op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vv_f16m1))) +vfloat16m1_t vfnmsac(vfloat16m1_t op0, vfloat16m1_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vv_f16m1_m))) +vfloat16m1_t vfnmsac(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vv_f16m2))) +vfloat16m2_t vfnmsac(vfloat16m2_t op0, vfloat16m2_t op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vv_f16m2_m))) +vfloat16m2_t vfnmsac(vbool8_t op0, vfloat16m2_t op1, vfloat16m2_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vv_f16m4))) +vfloat16m4_t vfnmsac(vfloat16m4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vv_f16m4_m))) +vfloat16m4_t vfnmsac(vbool4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vv_f16m8))) +vfloat16m8_t vfnmsac(vfloat16m8_t op0, vfloat16m8_t op1, vfloat16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vv_f16m8_m))) +vfloat16m8_t vfnmsac(vbool2_t op0, vfloat16m8_t op1, vfloat16m8_t op2, vfloat16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vv_f16mf2))) +vfloat16mf2_t vfnmsac(vfloat16mf2_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vv_f16mf2_m))) +vfloat16mf2_t vfnmsac(vbool32_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vv_f16mf4))) +vfloat16mf4_t vfnmsac(vfloat16mf4_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vv_f16mf4_m))) +vfloat16mf4_t vfnmsac(vbool64_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vf_f16m1))) +vfloat16m1_t vfnmsac(vfloat16m1_t op0, _Float16 op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vf_f16m1_m))) +vfloat16m1_t vfnmsac(vbool16_t op0, vfloat16m1_t op1, _Float16 op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vf_f16m2))) +vfloat16m2_t vfnmsac(vfloat16m2_t op0, _Float16 op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vf_f16m2_m))) +vfloat16m2_t vfnmsac(vbool8_t op0, vfloat16m2_t op1, _Float16 op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vf_f16m4))) +vfloat16m4_t vfnmsac(vfloat16m4_t op0, _Float16 op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vf_f16m4_m))) +vfloat16m4_t vfnmsac(vbool4_t op0, vfloat16m4_t op1, _Float16 op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vf_f16m8))) +vfloat16m8_t vfnmsac(vfloat16m8_t op0, _Float16 op1, vfloat16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vf_f16m8_m))) +vfloat16m8_t vfnmsac(vbool2_t op0, vfloat16m8_t op1, _Float16 op2, vfloat16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vf_f16mf2))) +vfloat16mf2_t vfnmsac(vfloat16mf2_t op0, _Float16 op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vf_f16mf2_m))) +vfloat16mf2_t vfnmsac(vbool32_t op0, vfloat16mf2_t op1, _Float16 op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vf_f16mf4))) +vfloat16mf4_t vfnmsac(vfloat16mf4_t op0, _Float16 op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsac_vf_f16mf4_m))) +vfloat16mf4_t vfnmsac(vbool64_t op0, vfloat16mf4_t op1, _Float16 op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vv_f16m1))) +vfloat16m1_t vfmadd(vfloat16m1_t op0, vfloat16m1_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vv_f16m1_m))) +vfloat16m1_t vfmadd(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vv_f16m2))) +vfloat16m2_t vfmadd(vfloat16m2_t op0, vfloat16m2_t op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vv_f16m2_m))) +vfloat16m2_t vfmadd(vbool8_t op0, vfloat16m2_t op1, vfloat16m2_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vv_f16m4))) +vfloat16m4_t vfmadd(vfloat16m4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vv_f16m4_m))) +vfloat16m4_t vfmadd(vbool4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vv_f16m8))) +vfloat16m8_t vfmadd(vfloat16m8_t op0, vfloat16m8_t op1, vfloat16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vv_f16m8_m))) +vfloat16m8_t vfmadd(vbool2_t op0, vfloat16m8_t op1, vfloat16m8_t op2, vfloat16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vv_f16mf2))) +vfloat16mf2_t vfmadd(vfloat16mf2_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vv_f16mf2_m))) +vfloat16mf2_t vfmadd(vbool32_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vv_f16mf4))) +vfloat16mf4_t vfmadd(vfloat16mf4_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vv_f16mf4_m))) +vfloat16mf4_t vfmadd(vbool64_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vf_f16m1))) +vfloat16m1_t vfmadd(vfloat16m1_t op0, _Float16 op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vf_f16m1_m))) +vfloat16m1_t vfmadd(vbool16_t op0, vfloat16m1_t op1, _Float16 op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vf_f16m2))) +vfloat16m2_t vfmadd(vfloat16m2_t op0, _Float16 op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vf_f16m2_m))) +vfloat16m2_t vfmadd(vbool8_t op0, vfloat16m2_t op1, _Float16 op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vf_f16m4))) +vfloat16m4_t vfmadd(vfloat16m4_t op0, _Float16 op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vf_f16m4_m))) +vfloat16m4_t vfmadd(vbool4_t op0, vfloat16m4_t op1, _Float16 op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vf_f16m8))) +vfloat16m8_t vfmadd(vfloat16m8_t op0, _Float16 op1, vfloat16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vf_f16m8_m))) +vfloat16m8_t vfmadd(vbool2_t op0, vfloat16m8_t op1, _Float16 op2, vfloat16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vf_f16mf2))) +vfloat16mf2_t vfmadd(vfloat16mf2_t op0, _Float16 op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vf_f16mf2_m))) +vfloat16mf2_t vfmadd(vbool32_t op0, vfloat16mf2_t op1, _Float16 op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vf_f16mf4))) +vfloat16mf4_t vfmadd(vfloat16mf4_t op0, _Float16 op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmadd_vf_f16mf4_m))) +vfloat16mf4_t vfmadd(vbool64_t op0, vfloat16mf4_t op1, _Float16 op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vv_f16m1))) +vfloat16m1_t vfnmadd(vfloat16m1_t op0, vfloat16m1_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vv_f16m1_m))) +vfloat16m1_t vfnmadd(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vv_f16m2))) +vfloat16m2_t vfnmadd(vfloat16m2_t op0, vfloat16m2_t op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vv_f16m2_m))) +vfloat16m2_t vfnmadd(vbool8_t op0, vfloat16m2_t op1, vfloat16m2_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vv_f16m4))) +vfloat16m4_t vfnmadd(vfloat16m4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vv_f16m4_m))) +vfloat16m4_t vfnmadd(vbool4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vv_f16m8))) +vfloat16m8_t vfnmadd(vfloat16m8_t op0, vfloat16m8_t op1, vfloat16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vv_f16m8_m))) +vfloat16m8_t vfnmadd(vbool2_t op0, vfloat16m8_t op1, vfloat16m8_t op2, vfloat16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vv_f16mf2))) +vfloat16mf2_t vfnmadd(vfloat16mf2_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vv_f16mf2_m))) +vfloat16mf2_t vfnmadd(vbool32_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vv_f16mf4))) +vfloat16mf4_t vfnmadd(vfloat16mf4_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vv_f16mf4_m))) +vfloat16mf4_t vfnmadd(vbool64_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vf_f16m1))) +vfloat16m1_t vfnmadd(vfloat16m1_t op0, _Float16 op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vf_f16m1_m))) +vfloat16m1_t vfnmadd(vbool16_t op0, vfloat16m1_t op1, _Float16 op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vf_f16m2))) +vfloat16m2_t vfnmadd(vfloat16m2_t op0, _Float16 op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vf_f16m2_m))) +vfloat16m2_t vfnmadd(vbool8_t op0, vfloat16m2_t op1, _Float16 op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vf_f16m4))) +vfloat16m4_t vfnmadd(vfloat16m4_t op0, _Float16 op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vf_f16m4_m))) +vfloat16m4_t vfnmadd(vbool4_t op0, vfloat16m4_t op1, _Float16 op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vf_f16m8))) +vfloat16m8_t vfnmadd(vfloat16m8_t op0, _Float16 op1, vfloat16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vf_f16m8_m))) +vfloat16m8_t vfnmadd(vbool2_t op0, vfloat16m8_t op1, _Float16 op2, vfloat16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vf_f16mf2))) +vfloat16mf2_t vfnmadd(vfloat16mf2_t op0, _Float16 op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vf_f16mf2_m))) +vfloat16mf2_t vfnmadd(vbool32_t op0, vfloat16mf2_t op1, _Float16 op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vf_f16mf4))) +vfloat16mf4_t vfnmadd(vfloat16mf4_t op0, _Float16 op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmadd_vf_f16mf4_m))) +vfloat16mf4_t vfnmadd(vbool64_t op0, vfloat16mf4_t op1, _Float16 op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vv_f16m1))) +vfloat16m1_t vfmsub(vfloat16m1_t op0, vfloat16m1_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vv_f16m1_m))) +vfloat16m1_t vfmsub(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vv_f16m2))) +vfloat16m2_t vfmsub(vfloat16m2_t op0, vfloat16m2_t op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vv_f16m2_m))) +vfloat16m2_t vfmsub(vbool8_t op0, vfloat16m2_t op1, vfloat16m2_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vv_f16m4))) +vfloat16m4_t vfmsub(vfloat16m4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vv_f16m4_m))) +vfloat16m4_t vfmsub(vbool4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vv_f16m8))) +vfloat16m8_t vfmsub(vfloat16m8_t op0, vfloat16m8_t op1, vfloat16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vv_f16m8_m))) +vfloat16m8_t vfmsub(vbool2_t op0, vfloat16m8_t op1, vfloat16m8_t op2, vfloat16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vv_f16mf2))) +vfloat16mf2_t vfmsub(vfloat16mf2_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vv_f16mf2_m))) +vfloat16mf2_t vfmsub(vbool32_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vv_f16mf4))) +vfloat16mf4_t vfmsub(vfloat16mf4_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vv_f16mf4_m))) +vfloat16mf4_t vfmsub(vbool64_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vf_f16m1))) +vfloat16m1_t vfmsub(vfloat16m1_t op0, _Float16 op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vf_f16m1_m))) +vfloat16m1_t vfmsub(vbool16_t op0, vfloat16m1_t op1, _Float16 op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vf_f16m2))) +vfloat16m2_t vfmsub(vfloat16m2_t op0, _Float16 op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vf_f16m2_m))) +vfloat16m2_t vfmsub(vbool8_t op0, vfloat16m2_t op1, _Float16 op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vf_f16m4))) +vfloat16m4_t vfmsub(vfloat16m4_t op0, _Float16 op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vf_f16m4_m))) +vfloat16m4_t vfmsub(vbool4_t op0, vfloat16m4_t op1, _Float16 op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vf_f16m8))) +vfloat16m8_t vfmsub(vfloat16m8_t op0, _Float16 op1, vfloat16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vf_f16m8_m))) +vfloat16m8_t vfmsub(vbool2_t op0, vfloat16m8_t op1, _Float16 op2, vfloat16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vf_f16mf2))) +vfloat16mf2_t vfmsub(vfloat16mf2_t op0, _Float16 op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vf_f16mf2_m))) +vfloat16mf2_t vfmsub(vbool32_t op0, vfloat16mf2_t op1, _Float16 op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vf_f16mf4))) +vfloat16mf4_t vfmsub(vfloat16mf4_t op0, _Float16 op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmsub_vf_f16mf4_m))) +vfloat16mf4_t vfmsub(vbool64_t op0, vfloat16mf4_t op1, _Float16 op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vv_f16m1))) +vfloat16m1_t vfnmsub(vfloat16m1_t op0, vfloat16m1_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vv_f16m1_m))) +vfloat16m1_t vfnmsub(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vv_f16m2))) +vfloat16m2_t vfnmsub(vfloat16m2_t op0, vfloat16m2_t op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vv_f16m2_m))) +vfloat16m2_t vfnmsub(vbool8_t op0, vfloat16m2_t op1, vfloat16m2_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vv_f16m4))) +vfloat16m4_t vfnmsub(vfloat16m4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vv_f16m4_m))) +vfloat16m4_t vfnmsub(vbool4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vv_f16m8))) +vfloat16m8_t vfnmsub(vfloat16m8_t op0, vfloat16m8_t op1, vfloat16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vv_f16m8_m))) +vfloat16m8_t vfnmsub(vbool2_t op0, vfloat16m8_t op1, vfloat16m8_t op2, vfloat16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vv_f16mf2))) +vfloat16mf2_t vfnmsub(vfloat16mf2_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vv_f16mf2_m))) +vfloat16mf2_t vfnmsub(vbool32_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vv_f16mf4))) +vfloat16mf4_t vfnmsub(vfloat16mf4_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vv_f16mf4_m))) +vfloat16mf4_t vfnmsub(vbool64_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vf_f16m1))) +vfloat16m1_t vfnmsub(vfloat16m1_t op0, _Float16 op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vf_f16m1_m))) +vfloat16m1_t vfnmsub(vbool16_t op0, vfloat16m1_t op1, _Float16 op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vf_f16m2))) +vfloat16m2_t vfnmsub(vfloat16m2_t op0, _Float16 op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vf_f16m2_m))) +vfloat16m2_t vfnmsub(vbool8_t op0, vfloat16m2_t op1, _Float16 op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vf_f16m4))) +vfloat16m4_t vfnmsub(vfloat16m4_t op0, _Float16 op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vf_f16m4_m))) +vfloat16m4_t vfnmsub(vbool4_t op0, vfloat16m4_t op1, _Float16 op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vf_f16m8))) +vfloat16m8_t vfnmsub(vfloat16m8_t op0, _Float16 op1, vfloat16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vf_f16m8_m))) +vfloat16m8_t vfnmsub(vbool2_t op0, vfloat16m8_t op1, _Float16 op2, vfloat16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vf_f16mf2))) +vfloat16mf2_t vfnmsub(vfloat16mf2_t op0, _Float16 op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vf_f16mf2_m))) +vfloat16mf2_t vfnmsub(vbool32_t op0, vfloat16mf2_t op1, _Float16 op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vf_f16mf4))) +vfloat16mf4_t vfnmsub(vfloat16mf4_t op0, _Float16 op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfnmsub_vf_f16mf4_m))) +vfloat16mf4_t vfnmsub(vbool64_t op0, vfloat16mf4_t op1, _Float16 op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vv_f16m1))) +vfloat16m1_t vfmin(vfloat16m1_t op0, vfloat16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vv_f16m1_m))) +vfloat16m1_t vfmin(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vv_f16m2))) +vfloat16m2_t vfmin(vfloat16m2_t op0, vfloat16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vv_f16m2_m))) +vfloat16m2_t vfmin(vbool8_t op0, vfloat16m2_t op1, vfloat16m2_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vv_f16m4))) +vfloat16m4_t vfmin(vfloat16m4_t op0, vfloat16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vv_f16m4_m))) +vfloat16m4_t vfmin(vbool4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vv_f16m8))) +vfloat16m8_t vfmin(vfloat16m8_t op0, vfloat16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vv_f16m8_m))) +vfloat16m8_t vfmin(vbool2_t op0, vfloat16m8_t op1, vfloat16m8_t op2, vfloat16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vv_f16mf2))) +vfloat16mf2_t vfmin(vfloat16mf2_t op0, vfloat16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vv_f16mf2_m))) +vfloat16mf2_t vfmin(vbool32_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vv_f16mf4))) +vfloat16mf4_t vfmin(vfloat16mf4_t op0, vfloat16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vv_f16mf4_m))) +vfloat16mf4_t vfmin(vbool64_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vf_f16m1))) +vfloat16m1_t vfmin(vfloat16m1_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vf_f16m1_m))) +vfloat16m1_t vfmin(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vf_f16m2))) +vfloat16m2_t vfmin(vfloat16m2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vf_f16m2_m))) +vfloat16m2_t vfmin(vbool8_t op0, vfloat16m2_t op1, vfloat16m2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vf_f16m4))) +vfloat16m4_t vfmin(vfloat16m4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vf_f16m4_m))) +vfloat16m4_t vfmin(vbool4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vf_f16m8))) +vfloat16m8_t vfmin(vfloat16m8_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vf_f16m8_m))) +vfloat16m8_t vfmin(vbool2_t op0, vfloat16m8_t op1, vfloat16m8_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vf_f16mf2))) +vfloat16mf2_t vfmin(vfloat16mf2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vf_f16mf2_m))) +vfloat16mf2_t vfmin(vbool32_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vf_f16mf4))) +vfloat16mf4_t vfmin(vfloat16mf4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmin_vf_f16mf4_m))) +vfloat16mf4_t vfmin(vbool64_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vv_f16m1))) +vfloat16m1_t vfmax(vfloat16m1_t op0, vfloat16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vv_f16m1_m))) +vfloat16m1_t vfmax(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vv_f16m2))) +vfloat16m2_t vfmax(vfloat16m2_t op0, vfloat16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vv_f16m2_m))) +vfloat16m2_t vfmax(vbool8_t op0, vfloat16m2_t op1, vfloat16m2_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vv_f16m4))) +vfloat16m4_t vfmax(vfloat16m4_t op0, vfloat16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vv_f16m4_m))) +vfloat16m4_t vfmax(vbool4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vv_f16m8))) +vfloat16m8_t vfmax(vfloat16m8_t op0, vfloat16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vv_f16m8_m))) +vfloat16m8_t vfmax(vbool2_t op0, vfloat16m8_t op1, vfloat16m8_t op2, vfloat16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vv_f16mf2))) +vfloat16mf2_t vfmax(vfloat16mf2_t op0, vfloat16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vv_f16mf2_m))) +vfloat16mf2_t vfmax(vbool32_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vv_f16mf4))) +vfloat16mf4_t vfmax(vfloat16mf4_t op0, vfloat16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vv_f16mf4_m))) +vfloat16mf4_t vfmax(vbool64_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vf_f16m1))) +vfloat16m1_t vfmax(vfloat16m1_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vf_f16m1_m))) +vfloat16m1_t vfmax(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vf_f16m2))) +vfloat16m2_t vfmax(vfloat16m2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vf_f16m2_m))) +vfloat16m2_t vfmax(vbool8_t op0, vfloat16m2_t op1, vfloat16m2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vf_f16m4))) +vfloat16m4_t vfmax(vfloat16m4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vf_f16m4_m))) +vfloat16m4_t vfmax(vbool4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vf_f16m8))) +vfloat16m8_t vfmax(vfloat16m8_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vf_f16m8_m))) +vfloat16m8_t vfmax(vbool2_t op0, vfloat16m8_t op1, vfloat16m8_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vf_f16mf2))) +vfloat16mf2_t vfmax(vfloat16mf2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vf_f16mf2_m))) +vfloat16mf2_t vfmax(vbool32_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vf_f16mf4))) +vfloat16mf4_t vfmax(vfloat16mf4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmax_vf_f16mf4_m))) +vfloat16mf4_t vfmax(vbool64_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vv_f16m1))) +vfloat16m1_t vfsgnj(vfloat16m1_t op0, vfloat16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vv_f16m1_m))) +vfloat16m1_t vfsgnj(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vv_f16m2))) +vfloat16m2_t vfsgnj(vfloat16m2_t op0, vfloat16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vv_f16m2_m))) +vfloat16m2_t vfsgnj(vbool8_t op0, vfloat16m2_t op1, vfloat16m2_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vv_f16m4))) +vfloat16m4_t vfsgnj(vfloat16m4_t op0, vfloat16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vv_f16m4_m))) +vfloat16m4_t vfsgnj(vbool4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vv_f16m8))) +vfloat16m8_t vfsgnj(vfloat16m8_t op0, vfloat16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vv_f16m8_m))) +vfloat16m8_t vfsgnj(vbool2_t op0, vfloat16m8_t op1, vfloat16m8_t op2, vfloat16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vv_f16mf2))) +vfloat16mf2_t vfsgnj(vfloat16mf2_t op0, vfloat16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vv_f16mf2_m))) +vfloat16mf2_t vfsgnj(vbool32_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vv_f16mf4))) +vfloat16mf4_t vfsgnj(vfloat16mf4_t op0, vfloat16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vv_f16mf4_m))) +vfloat16mf4_t vfsgnj(vbool64_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vf_f16m1))) +vfloat16m1_t vfsgnj(vfloat16m1_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vf_f16m1_m))) +vfloat16m1_t vfsgnj(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vf_f16m2))) +vfloat16m2_t vfsgnj(vfloat16m2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vf_f16m2_m))) +vfloat16m2_t vfsgnj(vbool8_t op0, vfloat16m2_t op1, vfloat16m2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vf_f16m4))) +vfloat16m4_t vfsgnj(vfloat16m4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vf_f16m4_m))) +vfloat16m4_t vfsgnj(vbool4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vf_f16m8))) +vfloat16m8_t vfsgnj(vfloat16m8_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vf_f16m8_m))) +vfloat16m8_t vfsgnj(vbool2_t op0, vfloat16m8_t op1, vfloat16m8_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vf_f16mf2))) +vfloat16mf2_t vfsgnj(vfloat16mf2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vf_f16mf2_m))) +vfloat16mf2_t vfsgnj(vbool32_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vf_f16mf4))) +vfloat16mf4_t vfsgnj(vfloat16mf4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnj_vf_f16mf4_m))) +vfloat16mf4_t vfsgnj(vbool64_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vv_f16m1))) +vfloat16m1_t vfsgnjn(vfloat16m1_t op0, vfloat16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vv_f16m1_m))) +vfloat16m1_t vfsgnjn(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vv_f16m2))) +vfloat16m2_t vfsgnjn(vfloat16m2_t op0, vfloat16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vv_f16m2_m))) +vfloat16m2_t vfsgnjn(vbool8_t op0, vfloat16m2_t op1, vfloat16m2_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vv_f16m4))) +vfloat16m4_t vfsgnjn(vfloat16m4_t op0, vfloat16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vv_f16m4_m))) +vfloat16m4_t vfsgnjn(vbool4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vv_f16m8))) +vfloat16m8_t vfsgnjn(vfloat16m8_t op0, vfloat16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vv_f16m8_m))) +vfloat16m8_t vfsgnjn(vbool2_t op0, vfloat16m8_t op1, vfloat16m8_t op2, vfloat16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vv_f16mf2))) +vfloat16mf2_t vfsgnjn(vfloat16mf2_t op0, vfloat16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vv_f16mf2_m))) +vfloat16mf2_t vfsgnjn(vbool32_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vv_f16mf4))) +vfloat16mf4_t vfsgnjn(vfloat16mf4_t op0, vfloat16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vv_f16mf4_m))) +vfloat16mf4_t vfsgnjn(vbool64_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vf_f16m1))) +vfloat16m1_t vfsgnjn(vfloat16m1_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vf_f16m1_m))) +vfloat16m1_t vfsgnjn(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vf_f16m2))) +vfloat16m2_t vfsgnjn(vfloat16m2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vf_f16m2_m))) +vfloat16m2_t vfsgnjn(vbool8_t op0, vfloat16m2_t op1, vfloat16m2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vf_f16m4))) +vfloat16m4_t vfsgnjn(vfloat16m4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vf_f16m4_m))) +vfloat16m4_t vfsgnjn(vbool4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vf_f16m8))) +vfloat16m8_t vfsgnjn(vfloat16m8_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vf_f16m8_m))) +vfloat16m8_t vfsgnjn(vbool2_t op0, vfloat16m8_t op1, vfloat16m8_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vf_f16mf2))) +vfloat16mf2_t vfsgnjn(vfloat16mf2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vf_f16mf2_m))) +vfloat16mf2_t vfsgnjn(vbool32_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vf_f16mf4))) +vfloat16mf4_t vfsgnjn(vfloat16mf4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjn_vf_f16mf4_m))) +vfloat16mf4_t vfsgnjn(vbool64_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vv_f16m1))) +vfloat16m1_t vfsgnjx(vfloat16m1_t op0, vfloat16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vv_f16m1_m))) +vfloat16m1_t vfsgnjx(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vv_f16m2))) +vfloat16m2_t vfsgnjx(vfloat16m2_t op0, vfloat16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vv_f16m2_m))) +vfloat16m2_t vfsgnjx(vbool8_t op0, vfloat16m2_t op1, vfloat16m2_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vv_f16m4))) +vfloat16m4_t vfsgnjx(vfloat16m4_t op0, vfloat16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vv_f16m4_m))) +vfloat16m4_t vfsgnjx(vbool4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vv_f16m8))) +vfloat16m8_t vfsgnjx(vfloat16m8_t op0, vfloat16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vv_f16m8_m))) +vfloat16m8_t vfsgnjx(vbool2_t op0, vfloat16m8_t op1, vfloat16m8_t op2, vfloat16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vv_f16mf2))) +vfloat16mf2_t vfsgnjx(vfloat16mf2_t op0, vfloat16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vv_f16mf2_m))) +vfloat16mf2_t vfsgnjx(vbool32_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vv_f16mf4))) +vfloat16mf4_t vfsgnjx(vfloat16mf4_t op0, vfloat16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vv_f16mf4_m))) +vfloat16mf4_t vfsgnjx(vbool64_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vf_f16m1))) +vfloat16m1_t vfsgnjx(vfloat16m1_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vf_f16m1_m))) +vfloat16m1_t vfsgnjx(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vf_f16m2))) +vfloat16m2_t vfsgnjx(vfloat16m2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vf_f16m2_m))) +vfloat16m2_t vfsgnjx(vbool8_t op0, vfloat16m2_t op1, vfloat16m2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vf_f16m4))) +vfloat16m4_t vfsgnjx(vfloat16m4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vf_f16m4_m))) +vfloat16m4_t vfsgnjx(vbool4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vf_f16m8))) +vfloat16m8_t vfsgnjx(vfloat16m8_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vf_f16m8_m))) +vfloat16m8_t vfsgnjx(vbool2_t op0, vfloat16m8_t op1, vfloat16m8_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vf_f16mf2))) +vfloat16mf2_t vfsgnjx(vfloat16mf2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vf_f16mf2_m))) +vfloat16mf2_t vfsgnjx(vbool32_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vf_f16mf4))) +vfloat16mf4_t vfsgnjx(vfloat16mf4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsgnjx_vf_f16mf4_m))) +vfloat16mf4_t vfsgnjx(vbool64_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfabs_v_f16m1))) +vfloat16m1_t vfabs(vfloat16m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfabs_v_f16m1_m))) +vfloat16m1_t vfabs(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfabs_v_f16m2))) +vfloat16m2_t vfabs(vfloat16m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfabs_v_f16m2_m))) +vfloat16m2_t vfabs(vbool8_t op0, vfloat16m2_t op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfabs_v_f16m4))) +vfloat16m4_t vfabs(vfloat16m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfabs_v_f16m4_m))) +vfloat16m4_t vfabs(vbool4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfabs_v_f16m8))) +vfloat16m8_t vfabs(vfloat16m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfabs_v_f16m8_m))) +vfloat16m8_t vfabs(vbool2_t op0, vfloat16m8_t op1, vfloat16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfabs_v_f16mf2))) +vfloat16mf2_t vfabs(vfloat16mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfabs_v_f16mf2_m))) +vfloat16mf2_t vfabs(vbool32_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfabs_v_f16mf4))) +vfloat16mf4_t vfabs(vfloat16mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfabs_v_f16mf4_m))) +vfloat16mf4_t vfabs(vbool64_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vv_f16m1_b16))) +vbool16_t vmfeq(vfloat16m1_t op0, vfloat16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vv_f16m1_b16_m))) +vbool16_t vmfeq(vbool16_t op0, vbool16_t op1, vfloat16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vv_f16m2_b8))) +vbool8_t vmfeq(vfloat16m2_t op0, vfloat16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vv_f16m2_b8_m))) +vbool8_t vmfeq(vbool8_t op0, vbool8_t op1, vfloat16m2_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vv_f16m4_b4))) +vbool4_t vmfeq(vfloat16m4_t op0, vfloat16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vv_f16m4_b4_m))) +vbool4_t vmfeq(vbool4_t op0, vbool4_t op1, vfloat16m4_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vv_f16m8_b2))) +vbool2_t vmfeq(vfloat16m8_t op0, vfloat16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vv_f16m8_b2_m))) +vbool2_t vmfeq(vbool2_t op0, vbool2_t op1, vfloat16m8_t op2, vfloat16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vv_f16mf2_b32))) +vbool32_t vmfeq(vfloat16mf2_t op0, vfloat16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vv_f16mf2_b32_m))) +vbool32_t vmfeq(vbool32_t op0, vbool32_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vv_f16mf4_b64))) +vbool64_t vmfeq(vfloat16mf4_t op0, vfloat16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vv_f16mf4_b64_m))) +vbool64_t vmfeq(vbool64_t op0, vbool64_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vf_f16m1_b16))) +vbool16_t vmfeq(vfloat16m1_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vf_f16m1_b16_m))) +vbool16_t vmfeq(vbool16_t op0, vbool16_t op1, vfloat16m1_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vf_f16m2_b8))) +vbool8_t vmfeq(vfloat16m2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vf_f16m2_b8_m))) +vbool8_t vmfeq(vbool8_t op0, vbool8_t op1, vfloat16m2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vf_f16m4_b4))) +vbool4_t vmfeq(vfloat16m4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vf_f16m4_b4_m))) +vbool4_t vmfeq(vbool4_t op0, vbool4_t op1, vfloat16m4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vf_f16m8_b2))) +vbool2_t vmfeq(vfloat16m8_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vf_f16m8_b2_m))) +vbool2_t vmfeq(vbool2_t op0, vbool2_t op1, vfloat16m8_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vf_f16mf2_b32))) +vbool32_t vmfeq(vfloat16mf2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vf_f16mf2_b32_m))) +vbool32_t vmfeq(vbool32_t op0, vbool32_t op1, vfloat16mf2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vf_f16mf4_b64))) +vbool64_t vmfeq(vfloat16mf4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfeq_vf_f16mf4_b64_m))) +vbool64_t vmfeq(vbool64_t op0, vbool64_t op1, vfloat16mf4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vv_f16m1_b16))) +vbool16_t vmfne(vfloat16m1_t op0, vfloat16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vv_f16m1_b16_m))) +vbool16_t vmfne(vbool16_t op0, vbool16_t op1, vfloat16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vv_f16m2_b8))) +vbool8_t vmfne(vfloat16m2_t op0, vfloat16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vv_f16m2_b8_m))) +vbool8_t vmfne(vbool8_t op0, vbool8_t op1, vfloat16m2_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vv_f16m4_b4))) +vbool4_t vmfne(vfloat16m4_t op0, vfloat16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vv_f16m4_b4_m))) +vbool4_t vmfne(vbool4_t op0, vbool4_t op1, vfloat16m4_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vv_f16m8_b2))) +vbool2_t vmfne(vfloat16m8_t op0, vfloat16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vv_f16m8_b2_m))) +vbool2_t vmfne(vbool2_t op0, vbool2_t op1, vfloat16m8_t op2, vfloat16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vv_f16mf2_b32))) +vbool32_t vmfne(vfloat16mf2_t op0, vfloat16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vv_f16mf2_b32_m))) +vbool32_t vmfne(vbool32_t op0, vbool32_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vv_f16mf4_b64))) +vbool64_t vmfne(vfloat16mf4_t op0, vfloat16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vv_f16mf4_b64_m))) +vbool64_t vmfne(vbool64_t op0, vbool64_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vf_f16m1_b16))) +vbool16_t vmfne(vfloat16m1_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vf_f16m1_b16_m))) +vbool16_t vmfne(vbool16_t op0, vbool16_t op1, vfloat16m1_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vf_f16m2_b8))) +vbool8_t vmfne(vfloat16m2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vf_f16m2_b8_m))) +vbool8_t vmfne(vbool8_t op0, vbool8_t op1, vfloat16m2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vf_f16m4_b4))) +vbool4_t vmfne(vfloat16m4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vf_f16m4_b4_m))) +vbool4_t vmfne(vbool4_t op0, vbool4_t op1, vfloat16m4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vf_f16m8_b2))) +vbool2_t vmfne(vfloat16m8_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vf_f16m8_b2_m))) +vbool2_t vmfne(vbool2_t op0, vbool2_t op1, vfloat16m8_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vf_f16mf2_b32))) +vbool32_t vmfne(vfloat16mf2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vf_f16mf2_b32_m))) +vbool32_t vmfne(vbool32_t op0, vbool32_t op1, vfloat16mf2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vf_f16mf4_b64))) +vbool64_t vmfne(vfloat16mf4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfne_vf_f16mf4_b64_m))) +vbool64_t vmfne(vbool64_t op0, vbool64_t op1, vfloat16mf4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vv_f16m1_b16))) +vbool16_t vmflt(vfloat16m1_t op0, vfloat16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vv_f16m1_b16_m))) +vbool16_t vmflt(vbool16_t op0, vbool16_t op1, vfloat16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vv_f16m2_b8))) +vbool8_t vmflt(vfloat16m2_t op0, vfloat16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vv_f16m2_b8_m))) +vbool8_t vmflt(vbool8_t op0, vbool8_t op1, vfloat16m2_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vv_f16m4_b4))) +vbool4_t vmflt(vfloat16m4_t op0, vfloat16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vv_f16m4_b4_m))) +vbool4_t vmflt(vbool4_t op0, vbool4_t op1, vfloat16m4_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vv_f16m8_b2))) +vbool2_t vmflt(vfloat16m8_t op0, vfloat16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vv_f16m8_b2_m))) +vbool2_t vmflt(vbool2_t op0, vbool2_t op1, vfloat16m8_t op2, vfloat16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vv_f16mf2_b32))) +vbool32_t vmflt(vfloat16mf2_t op0, vfloat16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vv_f16mf2_b32_m))) +vbool32_t vmflt(vbool32_t op0, vbool32_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vv_f16mf4_b64))) +vbool64_t vmflt(vfloat16mf4_t op0, vfloat16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vv_f16mf4_b64_m))) +vbool64_t vmflt(vbool64_t op0, vbool64_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vf_f16m1_b16))) +vbool16_t vmflt(vfloat16m1_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vf_f16m1_b16_m))) +vbool16_t vmflt(vbool16_t op0, vbool16_t op1, vfloat16m1_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vf_f16m2_b8))) +vbool8_t vmflt(vfloat16m2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vf_f16m2_b8_m))) +vbool8_t vmflt(vbool8_t op0, vbool8_t op1, vfloat16m2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vf_f16m4_b4))) +vbool4_t vmflt(vfloat16m4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vf_f16m4_b4_m))) +vbool4_t vmflt(vbool4_t op0, vbool4_t op1, vfloat16m4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vf_f16m8_b2))) +vbool2_t vmflt(vfloat16m8_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vf_f16m8_b2_m))) +vbool2_t vmflt(vbool2_t op0, vbool2_t op1, vfloat16m8_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vf_f16mf2_b32))) +vbool32_t vmflt(vfloat16mf2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vf_f16mf2_b32_m))) +vbool32_t vmflt(vbool32_t op0, vbool32_t op1, vfloat16mf2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vf_f16mf4_b64))) +vbool64_t vmflt(vfloat16mf4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmflt_vf_f16mf4_b64_m))) +vbool64_t vmflt(vbool64_t op0, vbool64_t op1, vfloat16mf4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vv_f16m1_b16))) +vbool16_t vmfle(vfloat16m1_t op0, vfloat16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vv_f16m1_b16_m))) +vbool16_t vmfle(vbool16_t op0, vbool16_t op1, vfloat16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vv_f16m2_b8))) +vbool8_t vmfle(vfloat16m2_t op0, vfloat16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vv_f16m2_b8_m))) +vbool8_t vmfle(vbool8_t op0, vbool8_t op1, vfloat16m2_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vv_f16m4_b4))) +vbool4_t vmfle(vfloat16m4_t op0, vfloat16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vv_f16m4_b4_m))) +vbool4_t vmfle(vbool4_t op0, vbool4_t op1, vfloat16m4_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vv_f16m8_b2))) +vbool2_t vmfle(vfloat16m8_t op0, vfloat16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vv_f16m8_b2_m))) +vbool2_t vmfle(vbool2_t op0, vbool2_t op1, vfloat16m8_t op2, vfloat16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vv_f16mf2_b32))) +vbool32_t vmfle(vfloat16mf2_t op0, vfloat16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vv_f16mf2_b32_m))) +vbool32_t vmfle(vbool32_t op0, vbool32_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vv_f16mf4_b64))) +vbool64_t vmfle(vfloat16mf4_t op0, vfloat16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vv_f16mf4_b64_m))) +vbool64_t vmfle(vbool64_t op0, vbool64_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vf_f16m1_b16))) +vbool16_t vmfle(vfloat16m1_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vf_f16m1_b16_m))) +vbool16_t vmfle(vbool16_t op0, vbool16_t op1, vfloat16m1_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vf_f16m2_b8))) +vbool8_t vmfle(vfloat16m2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vf_f16m2_b8_m))) +vbool8_t vmfle(vbool8_t op0, vbool8_t op1, vfloat16m2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vf_f16m4_b4))) +vbool4_t vmfle(vfloat16m4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vf_f16m4_b4_m))) +vbool4_t vmfle(vbool4_t op0, vbool4_t op1, vfloat16m4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vf_f16m8_b2))) +vbool2_t vmfle(vfloat16m8_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vf_f16m8_b2_m))) +vbool2_t vmfle(vbool2_t op0, vbool2_t op1, vfloat16m8_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vf_f16mf2_b32))) +vbool32_t vmfle(vfloat16mf2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vf_f16mf2_b32_m))) +vbool32_t vmfle(vbool32_t op0, vbool32_t op1, vfloat16mf2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vf_f16mf4_b64))) +vbool64_t vmfle(vfloat16mf4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfle_vf_f16mf4_b64_m))) +vbool64_t vmfle(vbool64_t op0, vbool64_t op1, vfloat16mf4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vv_f16m1_b16))) +vbool16_t vmfgt(vfloat16m1_t op0, vfloat16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vv_f16m1_b16_m))) +vbool16_t vmfgt(vbool16_t op0, vbool16_t op1, vfloat16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vv_f16m2_b8))) +vbool8_t vmfgt(vfloat16m2_t op0, vfloat16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vv_f16m2_b8_m))) +vbool8_t vmfgt(vbool8_t op0, vbool8_t op1, vfloat16m2_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vv_f16m4_b4))) +vbool4_t vmfgt(vfloat16m4_t op0, vfloat16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vv_f16m4_b4_m))) +vbool4_t vmfgt(vbool4_t op0, vbool4_t op1, vfloat16m4_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vv_f16m8_b2))) +vbool2_t vmfgt(vfloat16m8_t op0, vfloat16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vv_f16m8_b2_m))) +vbool2_t vmfgt(vbool2_t op0, vbool2_t op1, vfloat16m8_t op2, vfloat16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vv_f16mf2_b32))) +vbool32_t vmfgt(vfloat16mf2_t op0, vfloat16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vv_f16mf2_b32_m))) +vbool32_t vmfgt(vbool32_t op0, vbool32_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vv_f16mf4_b64))) +vbool64_t vmfgt(vfloat16mf4_t op0, vfloat16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vv_f16mf4_b64_m))) +vbool64_t vmfgt(vbool64_t op0, vbool64_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vf_f16m1_b16))) +vbool16_t vmfgt(vfloat16m1_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vf_f16m1_b16_m))) +vbool16_t vmfgt(vbool16_t op0, vbool16_t op1, vfloat16m1_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vf_f16m2_b8))) +vbool8_t vmfgt(vfloat16m2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vf_f16m2_b8_m))) +vbool8_t vmfgt(vbool8_t op0, vbool8_t op1, vfloat16m2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vf_f16m4_b4))) +vbool4_t vmfgt(vfloat16m4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vf_f16m4_b4_m))) +vbool4_t vmfgt(vbool4_t op0, vbool4_t op1, vfloat16m4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vf_f16m8_b2))) +vbool2_t vmfgt(vfloat16m8_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vf_f16m8_b2_m))) +vbool2_t vmfgt(vbool2_t op0, vbool2_t op1, vfloat16m8_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vf_f16mf2_b32))) +vbool32_t vmfgt(vfloat16mf2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vf_f16mf2_b32_m))) +vbool32_t vmfgt(vbool32_t op0, vbool32_t op1, vfloat16mf2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vf_f16mf4_b64))) +vbool64_t vmfgt(vfloat16mf4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfgt_vf_f16mf4_b64_m))) +vbool64_t vmfgt(vbool64_t op0, vbool64_t op1, vfloat16mf4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vv_f16m1_b16))) +vbool16_t vmfge(vfloat16m1_t op0, vfloat16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vv_f16m1_b16_m))) +vbool16_t vmfge(vbool16_t op0, vbool16_t op1, vfloat16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vv_f16m2_b8))) +vbool8_t vmfge(vfloat16m2_t op0, vfloat16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vv_f16m2_b8_m))) +vbool8_t vmfge(vbool8_t op0, vbool8_t op1, vfloat16m2_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vv_f16m4_b4))) +vbool4_t vmfge(vfloat16m4_t op0, vfloat16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vv_f16m4_b4_m))) +vbool4_t vmfge(vbool4_t op0, vbool4_t op1, vfloat16m4_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vv_f16m8_b2))) +vbool2_t vmfge(vfloat16m8_t op0, vfloat16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vv_f16m8_b2_m))) +vbool2_t vmfge(vbool2_t op0, vbool2_t op1, vfloat16m8_t op2, vfloat16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vv_f16mf2_b32))) +vbool32_t vmfge(vfloat16mf2_t op0, vfloat16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vv_f16mf2_b32_m))) +vbool32_t vmfge(vbool32_t op0, vbool32_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vv_f16mf4_b64))) +vbool64_t vmfge(vfloat16mf4_t op0, vfloat16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vv_f16mf4_b64_m))) +vbool64_t vmfge(vbool64_t op0, vbool64_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vf_f16m1_b16))) +vbool16_t vmfge(vfloat16m1_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vf_f16m1_b16_m))) +vbool16_t vmfge(vbool16_t op0, vbool16_t op1, vfloat16m1_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vf_f16m2_b8))) +vbool8_t vmfge(vfloat16m2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vf_f16m2_b8_m))) +vbool8_t vmfge(vbool8_t op0, vbool8_t op1, vfloat16m2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vf_f16m4_b4))) +vbool4_t vmfge(vfloat16m4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vf_f16m4_b4_m))) +vbool4_t vmfge(vbool4_t op0, vbool4_t op1, vfloat16m4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vf_f16m8_b2))) +vbool2_t vmfge(vfloat16m8_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vf_f16m8_b2_m))) +vbool2_t vmfge(vbool2_t op0, vbool2_t op1, vfloat16m8_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vf_f16mf2_b32))) +vbool32_t vmfge(vfloat16mf2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vf_f16mf2_b32_m))) +vbool32_t vmfge(vbool32_t op0, vbool32_t op1, vfloat16mf2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vf_f16mf4_b64))) +vbool64_t vmfge(vfloat16mf4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmfge_vf_f16mf4_b64_m))) +vbool64_t vmfge(vbool64_t op0, vbool64_t op1, vfloat16mf4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_f16m1))) +vfloat16m1_t vmerge(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_f16m2))) +vfloat16m2_t vmerge(vbool8_t op0, vfloat16m2_t op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_f16m4))) +vfloat16m4_t vmerge(vbool4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_f16m8))) +vfloat16m8_t vmerge(vbool2_t op0, vfloat16m8_t op1, vfloat16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_f16mf2))) +vfloat16mf2_t vmerge(vbool32_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vmerge_vvm_f16mf4))) +vfloat16mf4_t vmerge(vbool64_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmerge_vfm_f16m1))) +vfloat16m1_t vfmerge(vbool16_t op0, vfloat16m1_t op1, _Float16 op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmerge_vfm_f16m2))) +vfloat16m2_t vfmerge(vbool8_t op0, vfloat16m2_t op1, _Float16 op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmerge_vfm_f16m4))) +vfloat16m4_t vfmerge(vbool4_t op0, vfloat16m4_t op1, _Float16 op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmerge_vfm_f16m8))) +vfloat16m8_t vfmerge(vbool2_t op0, vfloat16m8_t op1, _Float16 op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmerge_vfm_f16mf2))) +vfloat16mf2_t vfmerge(vbool32_t op0, vfloat16mf2_t op1, _Float16 op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmerge_vfm_f16mf4))) +vfloat16mf4_t vfmerge(vbool64_t op0, vfloat16mf4_t op1, _Float16 op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmax_vs_f16m1_f16m1))) +vfloat16m1_t vfredmax(vfloat16m1_t op0, vfloat16m1_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmax_vs_f16m1_f16m1_m))) +vfloat16m1_t vfredmax(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmax_vs_f16m2_f16m1))) +vfloat16m1_t vfredmax(vfloat16m1_t op0, vfloat16m2_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmax_vs_f16m2_f16m1_m))) +vfloat16m1_t vfredmax(vbool8_t op0, vfloat16m1_t op1, vfloat16m2_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmax_vs_f16m4_f16m1))) +vfloat16m1_t vfredmax(vfloat16m1_t op0, vfloat16m4_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmax_vs_f16m4_f16m1_m))) +vfloat16m1_t vfredmax(vbool4_t op0, vfloat16m1_t op1, vfloat16m4_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmax_vs_f16m8_f16m1))) +vfloat16m1_t vfredmax(vfloat16m1_t op0, vfloat16m8_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmax_vs_f16m8_f16m1_m))) +vfloat16m1_t vfredmax(vbool2_t op0, vfloat16m1_t op1, vfloat16m8_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmax_vs_f16mf2_f16m1))) +vfloat16m1_t vfredmax(vfloat16m1_t op0, vfloat16mf2_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmax_vs_f16mf2_f16m1_m))) +vfloat16m1_t vfredmax(vbool32_t op0, vfloat16m1_t op1, vfloat16mf2_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmax_vs_f16mf4_f16m1))) +vfloat16m1_t vfredmax(vfloat16m1_t op0, vfloat16mf4_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmax_vs_f16mf4_f16m1_m))) +vfloat16m1_t vfredmax(vbool64_t op0, vfloat16m1_t op1, vfloat16mf4_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmin_vs_f16m1_f16m1))) +vfloat16m1_t vfredmin(vfloat16m1_t op0, vfloat16m1_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmin_vs_f16m1_f16m1_m))) +vfloat16m1_t vfredmin(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmin_vs_f16m2_f16m1))) +vfloat16m1_t vfredmin(vfloat16m1_t op0, vfloat16m2_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmin_vs_f16m2_f16m1_m))) +vfloat16m1_t vfredmin(vbool8_t op0, vfloat16m1_t op1, vfloat16m2_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmin_vs_f16m4_f16m1))) +vfloat16m1_t vfredmin(vfloat16m1_t op0, vfloat16m4_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmin_vs_f16m4_f16m1_m))) +vfloat16m1_t vfredmin(vbool4_t op0, vfloat16m1_t op1, vfloat16m4_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmin_vs_f16m8_f16m1))) +vfloat16m1_t vfredmin(vfloat16m1_t op0, vfloat16m8_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmin_vs_f16m8_f16m1_m))) +vfloat16m1_t vfredmin(vbool2_t op0, vfloat16m1_t op1, vfloat16m8_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmin_vs_f16mf2_f16m1))) +vfloat16m1_t vfredmin(vfloat16m1_t op0, vfloat16mf2_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmin_vs_f16mf2_f16m1_m))) +vfloat16m1_t vfredmin(vbool32_t op0, vfloat16m1_t op1, vfloat16mf2_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmin_vs_f16mf4_f16m1))) +vfloat16m1_t vfredmin(vfloat16m1_t op0, vfloat16mf4_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredmin_vs_f16mf4_f16m1_m))) +vfloat16m1_t vfredmin(vbool64_t op0, vfloat16m1_t op1, vfloat16mf4_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredsum_vs_f16m1_f16m1))) +vfloat16m1_t vfredsum(vfloat16m1_t op0, vfloat16m1_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredsum_vs_f16m1_f16m1_m))) +vfloat16m1_t vfredsum(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredsum_vs_f16m2_f16m1))) +vfloat16m1_t vfredsum(vfloat16m1_t op0, vfloat16m2_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredsum_vs_f16m2_f16m1_m))) +vfloat16m1_t vfredsum(vbool8_t op0, vfloat16m1_t op1, vfloat16m2_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredsum_vs_f16m4_f16m1))) +vfloat16m1_t vfredsum(vfloat16m1_t op0, vfloat16m4_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredsum_vs_f16m4_f16m1_m))) +vfloat16m1_t vfredsum(vbool4_t op0, vfloat16m1_t op1, vfloat16m4_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredsum_vs_f16m8_f16m1))) +vfloat16m1_t vfredsum(vfloat16m1_t op0, vfloat16m8_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredsum_vs_f16m8_f16m1_m))) +vfloat16m1_t vfredsum(vbool2_t op0, vfloat16m1_t op1, vfloat16m8_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredsum_vs_f16mf2_f16m1))) +vfloat16m1_t vfredsum(vfloat16m1_t op0, vfloat16mf2_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredsum_vs_f16mf2_f16m1_m))) +vfloat16m1_t vfredsum(vbool32_t op0, vfloat16m1_t op1, vfloat16mf2_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredsum_vs_f16mf4_f16m1))) +vfloat16m1_t vfredsum(vfloat16m1_t op0, vfloat16mf4_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredsum_vs_f16mf4_f16m1_m))) +vfloat16m1_t vfredsum(vbool64_t op0, vfloat16m1_t op1, vfloat16mf4_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredosum_vs_f16m1_f16m1))) +vfloat16m1_t vfredosum(vfloat16m1_t op0, vfloat16m1_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredosum_vs_f16m1_f16m1_m))) +vfloat16m1_t vfredosum(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredosum_vs_f16m2_f16m1))) +vfloat16m1_t vfredosum(vfloat16m1_t op0, vfloat16m2_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredosum_vs_f16m2_f16m1_m))) +vfloat16m1_t vfredosum(vbool8_t op0, vfloat16m1_t op1, vfloat16m2_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredosum_vs_f16m4_f16m1))) +vfloat16m1_t vfredosum(vfloat16m1_t op0, vfloat16m4_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredosum_vs_f16m4_f16m1_m))) +vfloat16m1_t vfredosum(vbool4_t op0, vfloat16m1_t op1, vfloat16m4_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredosum_vs_f16m8_f16m1))) +vfloat16m1_t vfredosum(vfloat16m1_t op0, vfloat16m8_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredosum_vs_f16m8_f16m1_m))) +vfloat16m1_t vfredosum(vbool2_t op0, vfloat16m1_t op1, vfloat16m8_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredosum_vs_f16mf2_f16m1))) +vfloat16m1_t vfredosum(vfloat16m1_t op0, vfloat16mf2_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredosum_vs_f16mf2_f16m1_m))) +vfloat16m1_t vfredosum(vbool32_t op0, vfloat16m1_t op1, vfloat16mf2_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredosum_vs_f16mf4_f16m1))) +vfloat16m1_t vfredosum(vfloat16m1_t op0, vfloat16mf4_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfredosum_vs_f16mf4_f16m1_m))) +vfloat16m1_t vfredosum(vbool64_t op0, vfloat16m1_t op1, vfloat16mf4_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_f16m1))) +void vsuxei8(_Float16 * op0, vuint8mf2_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_f16m1_m))) +void vsuxei8(vbool16_t op0, _Float16 * op1, vuint8mf2_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_f16m2))) +void vsuxei8(_Float16 * op0, vuint8m1_t op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_f16m2_m))) +void vsuxei8(vbool8_t op0, _Float16 * op1, vuint8m1_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_f16m4))) +void vsuxei8(_Float16 * op0, vuint8m2_t op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_f16m4_m))) +void vsuxei8(vbool4_t op0, _Float16 * op1, vuint8m2_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_f16m8))) +void vsuxei8(_Float16 * op0, vuint8m4_t op1, vfloat16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_f16m8_m))) +void vsuxei8(vbool2_t op0, _Float16 * op1, vuint8m4_t op2, vfloat16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_f16mf2))) +void vsuxei8(_Float16 * op0, vuint8mf4_t op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_f16mf2_m))) +void vsuxei8(vbool32_t op0, _Float16 * op1, vuint8mf4_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_f16mf4))) +void vsuxei8(_Float16 * op0, vuint8mf8_t op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei8_v_f16mf4_m))) +void vsuxei8(vbool64_t op0, _Float16 * op1, vuint8mf8_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_f16m1))) +void vsuxei16(_Float16 * op0, vuint16m1_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_f16m1_m))) +void vsuxei16(vbool16_t op0, _Float16 * op1, vuint16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_f16m2))) +void vsuxei16(_Float16 * op0, vuint16m2_t op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_f16m2_m))) +void vsuxei16(vbool8_t op0, _Float16 * op1, vuint16m2_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_f16m4))) +void vsuxei16(_Float16 * op0, vuint16m4_t op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_f16m4_m))) +void vsuxei16(vbool4_t op0, _Float16 * op1, vuint16m4_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_f16m8))) +void vsuxei16(_Float16 * op0, vuint16m8_t op1, vfloat16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_f16m8_m))) +void vsuxei16(vbool2_t op0, _Float16 * op1, vuint16m8_t op2, vfloat16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_f16mf2))) +void vsuxei16(_Float16 * op0, vuint16mf2_t op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_f16mf2_m))) +void vsuxei16(vbool32_t op0, _Float16 * op1, vuint16mf2_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_f16mf4))) +void vsuxei16(_Float16 * op0, vuint16mf4_t op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei16_v_f16mf4_m))) +void vsuxei16(vbool64_t op0, _Float16 * op1, vuint16mf4_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmv_f_s_f16m1_f16))) +_Float16 vfmv_f(vfloat16m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmv_f_s_f16m2_f16))) +_Float16 vfmv_f(vfloat16m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmv_f_s_f16m4_f16))) +_Float16 vfmv_f(vfloat16m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmv_f_s_f16m8_f16))) +_Float16 vfmv_f(vfloat16m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmv_f_s_f16mf2_f16))) +_Float16 vfmv_f(vfloat16mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmv_f_s_f16mf4_f16))) +_Float16 vfmv_f(vfloat16mf4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmv_s_f_f16m1))) +vfloat16m1_t vfmv_s(vfloat16m1_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmv_s_f_f16m2))) +vfloat16m2_t vfmv_s(vfloat16m2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmv_s_f_f16m4))) +vfloat16m4_t vfmv_s(vfloat16m4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmv_s_f_f16m8))) +vfloat16m8_t vfmv_s(vfloat16m8_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmv_s_f_f16mf2))) +vfloat16mf2_t vfmv_s(vfloat16mf2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfmv_s_f_f16mf4))) +vfloat16mf4_t vfmv_s(vfloat16mf4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_f16m1))) +vfloat16m1_t vslideup(vfloat16m1_t op0, vfloat16m1_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_f16m1_m))) +vfloat16m1_t vslideup(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_f16m2))) +vfloat16m2_t vslideup(vfloat16m2_t op0, vfloat16m2_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_f16m2_m))) +vfloat16m2_t vslideup(vbool8_t op0, vfloat16m2_t op1, vfloat16m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_f16m4))) +vfloat16m4_t vslideup(vfloat16m4_t op0, vfloat16m4_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_f16m4_m))) +vfloat16m4_t vslideup(vbool4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_f16m8))) +vfloat16m8_t vslideup(vfloat16m8_t op0, vfloat16m8_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_f16m8_m))) +vfloat16m8_t vslideup(vbool2_t op0, vfloat16m8_t op1, vfloat16m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_f16mf2))) +vfloat16mf2_t vslideup(vfloat16mf2_t op0, vfloat16mf2_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_f16mf2_m))) +vfloat16mf2_t vslideup(vbool32_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_f16mf4))) +vfloat16mf4_t vslideup(vfloat16mf4_t op0, vfloat16mf4_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslideup_vx_f16mf4_m))) +vfloat16mf4_t vslideup(vbool64_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_f16m1))) +vfloat16m1_t vslidedown(vfloat16m1_t op0, vfloat16m1_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_f16m1_m))) +vfloat16m1_t vslidedown(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_f16m2))) +vfloat16m2_t vslidedown(vfloat16m2_t op0, vfloat16m2_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_f16m2_m))) +vfloat16m2_t vslidedown(vbool8_t op0, vfloat16m2_t op1, vfloat16m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_f16m4))) +vfloat16m4_t vslidedown(vfloat16m4_t op0, vfloat16m4_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_f16m4_m))) +vfloat16m4_t vslidedown(vbool4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_f16m8))) +vfloat16m8_t vslidedown(vfloat16m8_t op0, vfloat16m8_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_f16m8_m))) +vfloat16m8_t vslidedown(vbool2_t op0, vfloat16m8_t op1, vfloat16m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_f16mf2))) +vfloat16mf2_t vslidedown(vfloat16mf2_t op0, vfloat16mf2_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_f16mf2_m))) +vfloat16mf2_t vslidedown(vbool32_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_f16mf4))) +vfloat16mf4_t vslidedown(vfloat16mf4_t op0, vfloat16mf4_t op1, size_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vslidedown_vx_f16mf4_m))) +vfloat16mf4_t vslidedown(vbool64_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1up_vf_f16m1))) +vfloat16m1_t vfslide1up(vfloat16m1_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1up_vf_f16m1_m))) +vfloat16m1_t vfslide1up(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1up_vf_f16m2))) +vfloat16m2_t vfslide1up(vfloat16m2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1up_vf_f16m2_m))) +vfloat16m2_t vfslide1up(vbool8_t op0, vfloat16m2_t op1, vfloat16m2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1up_vf_f16m4))) +vfloat16m4_t vfslide1up(vfloat16m4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1up_vf_f16m4_m))) +vfloat16m4_t vfslide1up(vbool4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1up_vf_f16m8))) +vfloat16m8_t vfslide1up(vfloat16m8_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1up_vf_f16m8_m))) +vfloat16m8_t vfslide1up(vbool2_t op0, vfloat16m8_t op1, vfloat16m8_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1up_vf_f16mf2))) +vfloat16mf2_t vfslide1up(vfloat16mf2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1up_vf_f16mf2_m))) +vfloat16mf2_t vfslide1up(vbool32_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1up_vf_f16mf4))) +vfloat16mf4_t vfslide1up(vfloat16mf4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1up_vf_f16mf4_m))) +vfloat16mf4_t vfslide1up(vbool64_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_f16m1))) +void vsuxei32(_Float16 * op0, vuint32m2_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_f16m1_m))) +void vsuxei32(vbool16_t op0, _Float16 * op1, vuint32m2_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_f16m2))) +void vsuxei32(_Float16 * op0, vuint32m4_t op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_f16m2_m))) +void vsuxei32(vbool8_t op0, _Float16 * op1, vuint32m4_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_f16m4))) +void vsuxei32(_Float16 * op0, vuint32m8_t op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_f16m4_m))) +void vsuxei32(vbool4_t op0, _Float16 * op1, vuint32m8_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_f16mf2))) +void vsuxei32(_Float16 * op0, vuint32m1_t op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_f16mf2_m))) +void vsuxei32(vbool32_t op0, _Float16 * op1, vuint32m1_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_f16mf4))) +void vsuxei32(_Float16 * op0, vuint32mf2_t op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei32_v_f16mf4_m))) +void vsuxei32(vbool64_t op0, _Float16 * op1, vuint32mf2_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1down_vf_f16m1))) +vfloat16m1_t vfslide1down(vfloat16m1_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1down_vf_f16m1_m))) +vfloat16m1_t vfslide1down(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1down_vf_f16m2))) +vfloat16m2_t vfslide1down(vfloat16m2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1down_vf_f16m2_m))) +vfloat16m2_t vfslide1down(vbool8_t op0, vfloat16m2_t op1, vfloat16m2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1down_vf_f16m4))) +vfloat16m4_t vfslide1down(vfloat16m4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1down_vf_f16m4_m))) +vfloat16m4_t vfslide1down(vbool4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1down_vf_f16m8))) +vfloat16m8_t vfslide1down(vfloat16m8_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1down_vf_f16m8_m))) +vfloat16m8_t vfslide1down(vbool2_t op0, vfloat16m8_t op1, vfloat16m8_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1down_vf_f16mf2))) +vfloat16mf2_t vfslide1down(vfloat16mf2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1down_vf_f16mf2_m))) +vfloat16mf2_t vfslide1down(vbool32_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1down_vf_f16mf4))) +vfloat16mf4_t vfslide1down(vfloat16mf4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfslide1down_vf_f16mf4_m))) +vfloat16mf4_t vfslide1down(vbool64_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_f16m1))) +vfloat16m1_t vrgather(vfloat16m1_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_f16m1_m))) +vfloat16m1_t vrgather(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_f16m2))) +vfloat16m2_t vrgather(vfloat16m2_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_f16m2_m))) +vfloat16m2_t vrgather(vbool8_t op0, vfloat16m2_t op1, vfloat16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_f16m4))) +vfloat16m4_t vrgather(vfloat16m4_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_f16m4_m))) +vfloat16m4_t vrgather(vbool4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_f16m8))) +vfloat16m8_t vrgather(vfloat16m8_t op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_f16m8_m))) +vfloat16m8_t vrgather(vbool2_t op0, vfloat16m8_t op1, vfloat16m8_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_f16mf2))) +vfloat16mf2_t vrgather(vfloat16mf2_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_f16mf2_m))) +vfloat16mf2_t vrgather(vbool32_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_f16mf4))) +vfloat16mf4_t vrgather(vfloat16mf4_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vv_f16mf4_m))) +vfloat16mf4_t vrgather(vbool64_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_f16m1))) +vfloat16m1_t vrgather(vfloat16m1_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_f16m1_m))) +vfloat16m1_t vrgather(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_f16m2))) +vfloat16m2_t vrgather(vfloat16m2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_f16m2_m))) +vfloat16m2_t vrgather(vbool8_t op0, vfloat16m2_t op1, vfloat16m2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_f16m4))) +vfloat16m4_t vrgather(vfloat16m4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_f16m4_m))) +vfloat16m4_t vrgather(vbool4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_f16m8))) +vfloat16m8_t vrgather(vfloat16m8_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_f16m8_m))) +vfloat16m8_t vrgather(vbool2_t op0, vfloat16m8_t op1, vfloat16m8_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_f16mf2))) +vfloat16mf2_t vrgather(vfloat16mf2_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_f16mf2_m))) +vfloat16mf2_t vrgather(vbool32_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_f16mf4))) +vfloat16mf4_t vrgather(vfloat16mf4_t op0, size_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgather_vx_f16mf4_m))) +vfloat16mf4_t vrgather(vbool64_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, size_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_f16m1))) +vfloat16m1_t vrgatherei16(vfloat16m1_t op0, vuint16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_f16m1_m))) +vfloat16m1_t vrgatherei16(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_f16m2))) +vfloat16m2_t vrgatherei16(vfloat16m2_t op0, vuint16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_f16m2_m))) +vfloat16m2_t vrgatherei16(vbool8_t op0, vfloat16m2_t op1, vfloat16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_f16m4))) +vfloat16m4_t vrgatherei16(vfloat16m4_t op0, vuint16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_f16m4_m))) +vfloat16m4_t vrgatherei16(vbool4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_f16m8))) +vfloat16m8_t vrgatherei16(vfloat16m8_t op0, vuint16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_f16m8_m))) +vfloat16m8_t vrgatherei16(vbool2_t op0, vfloat16m8_t op1, vfloat16m8_t op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_f16mf2))) +vfloat16mf2_t vrgatherei16(vfloat16mf2_t op0, vuint16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_f16mf2_m))) +vfloat16mf2_t vrgatherei16(vbool32_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_f16mf4))) +vfloat16mf4_t vrgatherei16(vfloat16mf4_t op0, vuint16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vrgatherei16_vv_f16mf4_m))) +vfloat16mf4_t vrgatherei16(vbool64_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_f16m1))) +vfloat16m1_t vcompress(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_f16m2))) +vfloat16m2_t vcompress(vbool8_t op0, vfloat16m2_t op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_f16m4))) +vfloat16m4_t vcompress(vbool4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_f16m8))) +vfloat16m8_t vcompress(vbool2_t op0, vfloat16m8_t op1, vfloat16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_f16mf2))) +vfloat16mf2_t vcompress(vbool32_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vcompress_vm_f16mf4))) +vfloat16mf4_t vcompress(vbool64_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_f16m1))) +void vsuxei64(_Float16 * op0, vuint64m4_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_f16m1_m))) +void vsuxei64(vbool16_t op0, _Float16 * op1, vuint64m4_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_f16m2))) +void vsuxei64(_Float16 * op0, vuint64m8_t op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_f16m2_m))) +void vsuxei64(vbool8_t op0, _Float16 * op1, vuint64m8_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_f16mf2))) +void vsuxei64(_Float16 * op0, vuint64m2_t op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_f16mf2_m))) +void vsuxei64(vbool32_t op0, _Float16 * op1, vuint64m2_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_f16mf4))) +void vsuxei64(_Float16 * op0, vuint64m1_t op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxei64_v_f16mf4_m))) +void vsuxei64(vbool64_t op0, _Float16 * op1, vuint64m1_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_f16m1))) +void vsoxei8(_Float16 * op0, vuint8mf2_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_f16m1_m))) +void vsoxei8(vbool16_t op0, _Float16 * op1, vuint8mf2_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_f16m2))) +void vsoxei8(_Float16 * op0, vuint8m1_t op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_f16m2_m))) +void vsoxei8(vbool8_t op0, _Float16 * op1, vuint8m1_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_f16m4))) +void vsoxei8(_Float16 * op0, vuint8m2_t op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_f16m4_m))) +void vsoxei8(vbool4_t op0, _Float16 * op1, vuint8m2_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_f16m8))) +void vsoxei8(_Float16 * op0, vuint8m4_t op1, vfloat16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_f16m8_m))) +void vsoxei8(vbool2_t op0, _Float16 * op1, vuint8m4_t op2, vfloat16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_f16mf2))) +void vsoxei8(_Float16 * op0, vuint8mf4_t op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_f16mf2_m))) +void vsoxei8(vbool32_t op0, _Float16 * op1, vuint8mf4_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_f16mf4))) +void vsoxei8(_Float16 * op0, vuint8mf8_t op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei8_v_f16mf4_m))) +void vsoxei8(vbool64_t op0, _Float16 * op1, vuint8mf8_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_f16m1))) +void vsoxei16(_Float16 * op0, vuint16m1_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_f16m1_m))) +void vsoxei16(vbool16_t op0, _Float16 * op1, vuint16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_f16m2))) +void vsoxei16(_Float16 * op0, vuint16m2_t op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_f16m2_m))) +void vsoxei16(vbool8_t op0, _Float16 * op1, vuint16m2_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_f16m4))) +void vsoxei16(_Float16 * op0, vuint16m4_t op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_f16m4_m))) +void vsoxei16(vbool4_t op0, _Float16 * op1, vuint16m4_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_f16m8))) +void vsoxei16(_Float16 * op0, vuint16m8_t op1, vfloat16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_f16m8_m))) +void vsoxei16(vbool2_t op0, _Float16 * op1, vuint16m8_t op2, vfloat16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_f16mf2))) +void vsoxei16(_Float16 * op0, vuint16mf2_t op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_f16mf2_m))) +void vsoxei16(vbool32_t op0, _Float16 * op1, vuint16mf2_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_f16mf4))) +void vsoxei16(_Float16 * op0, vuint16mf4_t op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei16_v_f16mf4_m))) +void vsoxei16(vbool64_t op0, _Float16 * op1, vuint16mf4_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_f16m1))) +void vsoxei32(_Float16 * op0, vuint32m2_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_f16m1_m))) +void vsoxei32(vbool16_t op0, _Float16 * op1, vuint32m2_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_f16m2))) +void vsoxei32(_Float16 * op0, vuint32m4_t op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_f16m2_m))) +void vsoxei32(vbool8_t op0, _Float16 * op1, vuint32m4_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_f16m4))) +void vsoxei32(_Float16 * op0, vuint32m8_t op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_f16m4_m))) +void vsoxei32(vbool4_t op0, _Float16 * op1, vuint32m8_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_f16mf2))) +void vsoxei32(_Float16 * op0, vuint32m1_t op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_f16mf2_m))) +void vsoxei32(vbool32_t op0, _Float16 * op1, vuint32m1_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_f16mf4))) +void vsoxei32(_Float16 * op0, vuint32mf2_t op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei32_v_f16mf4_m))) +void vsoxei32(vbool64_t op0, _Float16 * op1, vuint32mf2_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_f16m1))) +void vsoxei64(_Float16 * op0, vuint64m4_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_f16m1_m))) +void vsoxei64(vbool16_t op0, _Float16 * op1, vuint64m4_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_f16m2))) +void vsoxei64(_Float16 * op0, vuint64m8_t op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_f16m2_m))) +void vsoxei64(vbool8_t op0, _Float16 * op1, vuint64m8_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_f16mf2))) +void vsoxei64(_Float16 * op0, vuint64m2_t op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_f16mf2_m))) +void vsoxei64(vbool32_t op0, _Float16 * op1, vuint64m2_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_f16mf4))) +void vsoxei64(_Float16 * op0, vuint64m1_t op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxei64_v_f16mf4_m))) +void vsoxei64(vbool64_t op0, _Float16 * op1, vuint64m1_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle16ff_v_f16m1_m))) +vfloat16m1_t vle16ff(vbool16_t op0, vfloat16m1_t op1, const _Float16 * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle16ff_v_f16m2_m))) +vfloat16m2_t vle16ff(vbool8_t op0, vfloat16m2_t op1, const _Float16 * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle16ff_v_f16m4_m))) +vfloat16m4_t vle16ff(vbool4_t op0, vfloat16m4_t op1, const _Float16 * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle16ff_v_f16m8_m))) +vfloat16m8_t vle16ff(vbool2_t op0, vfloat16m8_t op1, const _Float16 * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle16ff_v_f16mf2_m))) +vfloat16mf2_t vle16ff(vbool32_t op0, vfloat16mf2_t op1, const _Float16 * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle16ff_v_f16mf4_m))) +vfloat16mf4_t vle16ff(vbool64_t op0, vfloat16mf4_t op1, const _Float16 * op2, size_t * op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfneg_v_f16m1))) +vfloat16m1_t vfneg(vfloat16m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfneg_v_f16m1_m))) +vfloat16m1_t vfneg(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfneg_v_f16m2))) +vfloat16m2_t vfneg(vfloat16m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfneg_v_f16m2_m))) +vfloat16m2_t vfneg(vbool8_t op0, vfloat16m2_t op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfneg_v_f16m4))) +vfloat16m4_t vfneg(vfloat16m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfneg_v_f16m4_m))) +vfloat16m4_t vfneg(vbool4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfneg_v_f16m8))) +vfloat16m8_t vfneg(vfloat16m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfneg_v_f16m8_m))) +vfloat16m8_t vfneg(vbool2_t op0, vfloat16m8_t op1, vfloat16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfneg_v_f16mf2))) +vfloat16mf2_t vfneg(vfloat16mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfneg_v_f16mf2_m))) +vfloat16mf2_t vfneg(vbool32_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfneg_v_f16mf4))) +vfloat16mf4_t vfneg(vfloat16mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfneg_v_f16mf4_m))) +vfloat16mf4_t vfneg(vbool64_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle16_v_f16m1_m))) +vfloat16m1_t vle16(vbool16_t op0, vfloat16m1_t op1, const _Float16 * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle16_v_f16m2_m))) +vfloat16m2_t vle16(vbool8_t op0, vfloat16m2_t op1, const _Float16 * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle16_v_f16m4_m))) +vfloat16m4_t vle16(vbool4_t op0, vfloat16m4_t op1, const _Float16 * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle16_v_f16m8_m))) +vfloat16m8_t vle16(vbool2_t op0, vfloat16m8_t op1, const _Float16 * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle16_v_f16mf2_m))) +vfloat16mf2_t vle16(vbool32_t op0, vfloat16mf2_t op1, const _Float16 * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vle16_v_f16mf4_m))) +vfloat16mf4_t vle16(vbool64_t op0, vfloat16mf4_t op1, const _Float16 * op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse16_v_f16m1))) +void vse16(_Float16 * op0, vfloat16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse16_v_f16m1_m))) +void vse16(vbool16_t op0, _Float16 * op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse16_v_f16m2))) +void vse16(_Float16 * op0, vfloat16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse16_v_f16m2_m))) +void vse16(vbool8_t op0, _Float16 * op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse16_v_f16m4))) +void vse16(_Float16 * op0, vfloat16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse16_v_f16m4_m))) +void vse16(vbool4_t op0, _Float16 * op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse16_v_f16m8))) +void vse16(_Float16 * op0, vfloat16m8_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse16_v_f16m8_m))) +void vse16(vbool2_t op0, _Float16 * op1, vfloat16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse16_v_f16mf2))) +void vse16(_Float16 * op0, vfloat16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse16_v_f16mf2_m))) +void vse16(vbool32_t op0, _Float16 * op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse16_v_f16mf4))) +void vse16(_Float16 * op0, vfloat16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vse16_v_f16mf4_m))) +void vse16(vbool64_t op0, _Float16 * op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse16_v_f16m1_m))) +vfloat16m1_t vlse16(vbool16_t op0, vfloat16m1_t op1, const _Float16 * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse16_v_f16m2_m))) +vfloat16m2_t vlse16(vbool8_t op0, vfloat16m2_t op1, const _Float16 * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse16_v_f16m4_m))) +vfloat16m4_t vlse16(vbool4_t op0, vfloat16m4_t op1, const _Float16 * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse16_v_f16m8_m))) +vfloat16m8_t vlse16(vbool2_t op0, vfloat16m8_t op1, const _Float16 * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse16_v_f16mf2_m))) +vfloat16mf2_t vlse16(vbool32_t op0, vfloat16mf2_t op1, const _Float16 * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlse16_v_f16mf4_m))) +vfloat16mf4_t vlse16(vbool64_t op0, vfloat16mf4_t op1, const _Float16 * op2, ptrdiff_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfclass_v_u16m1))) +vuint16m1_t vfclass(vfloat16m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfclass_v_u16m1_m))) +vuint16m1_t vfclass(vbool16_t op0, vuint16m1_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfclass_v_u16m2))) +vuint16m2_t vfclass(vfloat16m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfclass_v_u16m2_m))) +vuint16m2_t vfclass(vbool8_t op0, vuint16m2_t op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfclass_v_u16m4))) +vuint16m4_t vfclass(vfloat16m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfclass_v_u16m4_m))) +vuint16m4_t vfclass(vbool4_t op0, vuint16m4_t op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfclass_v_u16m8))) +vuint16m8_t vfclass(vfloat16m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfclass_v_u16m8_m))) +vuint16m8_t vfclass(vbool2_t op0, vuint16m8_t op1, vfloat16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfclass_v_u16mf2))) +vuint16mf2_t vfclass(vfloat16mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfclass_v_u16mf2_m))) +vuint16mf2_t vfclass(vbool32_t op0, vuint16mf2_t op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfclass_v_u16mf4))) +vuint16mf4_t vfclass(vfloat16mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfclass_v_u16mf4_m))) +vuint16mf4_t vfclass(vbool64_t op0, vuint16mf4_t op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_x_v_f16m1))) +vfloat16m1_t vfcvt_f(vint16m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_x_v_f16m1_m))) +vfloat16m1_t vfcvt_f(vbool16_t op0, vfloat16m1_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_x_v_f16m2))) +vfloat16m2_t vfcvt_f(vint16m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_x_v_f16m2_m))) +vfloat16m2_t vfcvt_f(vbool8_t op0, vfloat16m2_t op1, vint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_x_v_f16m4))) +vfloat16m4_t vfcvt_f(vint16m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_x_v_f16m4_m))) +vfloat16m4_t vfcvt_f(vbool4_t op0, vfloat16m4_t op1, vint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_x_v_f16m8))) +vfloat16m8_t vfcvt_f(vint16m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_x_v_f16m8_m))) +vfloat16m8_t vfcvt_f(vbool2_t op0, vfloat16m8_t op1, vint16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_x_v_f16mf2))) +vfloat16mf2_t vfcvt_f(vint16mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_x_v_f16mf2_m))) +vfloat16mf2_t vfcvt_f(vbool32_t op0, vfloat16mf2_t op1, vint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_x_v_f16mf4))) +vfloat16mf4_t vfcvt_f(vint16mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_x_v_f16mf4_m))) +vfloat16mf4_t vfcvt_f(vbool64_t op0, vfloat16mf4_t op1, vint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_xu_v_f16m1))) +vfloat16m1_t vfcvt_f(vuint16m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_xu_v_f16m1_m))) +vfloat16m1_t vfcvt_f(vbool16_t op0, vfloat16m1_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_xu_v_f16m2))) +vfloat16m2_t vfcvt_f(vuint16m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_xu_v_f16m2_m))) +vfloat16m2_t vfcvt_f(vbool8_t op0, vfloat16m2_t op1, vuint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_xu_v_f16m4))) +vfloat16m4_t vfcvt_f(vuint16m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_xu_v_f16m4_m))) +vfloat16m4_t vfcvt_f(vbool4_t op0, vfloat16m4_t op1, vuint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_xu_v_f16m8))) +vfloat16m8_t vfcvt_f(vuint16m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_xu_v_f16m8_m))) +vfloat16m8_t vfcvt_f(vbool2_t op0, vfloat16m8_t op1, vuint16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_xu_v_f16mf2))) +vfloat16mf2_t vfcvt_f(vuint16mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_xu_v_f16mf2_m))) +vfloat16mf2_t vfcvt_f(vbool32_t op0, vfloat16mf2_t op1, vuint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_xu_v_f16mf4))) +vfloat16mf4_t vfcvt_f(vuint16mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_f_xu_v_f16mf4_m))) +vfloat16mf4_t vfcvt_f(vbool64_t op0, vfloat16mf4_t op1, vuint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_x_f_v_i16m1))) +vint16m1_t vfcvt_rtz_x(vfloat16m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_x_f_v_i16m1_m))) +vint16m1_t vfcvt_rtz_x(vbool16_t op0, vint16m1_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_x_f_v_i16m2))) +vint16m2_t vfcvt_rtz_x(vfloat16m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_x_f_v_i16m2_m))) +vint16m2_t vfcvt_rtz_x(vbool8_t op0, vint16m2_t op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_x_f_v_i16m4))) +vint16m4_t vfcvt_rtz_x(vfloat16m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_x_f_v_i16m4_m))) +vint16m4_t vfcvt_rtz_x(vbool4_t op0, vint16m4_t op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_x_f_v_i16m8))) +vint16m8_t vfcvt_rtz_x(vfloat16m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_x_f_v_i16m8_m))) +vint16m8_t vfcvt_rtz_x(vbool2_t op0, vint16m8_t op1, vfloat16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_x_f_v_i16mf2))) +vint16mf2_t vfcvt_rtz_x(vfloat16mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_x_f_v_i16mf2_m))) +vint16mf2_t vfcvt_rtz_x(vbool32_t op0, vint16mf2_t op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_x_f_v_i16mf4))) +vint16mf4_t vfcvt_rtz_x(vfloat16mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_x_f_v_i16mf4_m))) +vint16mf4_t vfcvt_rtz_x(vbool64_t op0, vint16mf4_t op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_xu_f_v_u16m1))) +vuint16m1_t vfcvt_rtz_xu(vfloat16m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_xu_f_v_u16m1_m))) +vuint16m1_t vfcvt_rtz_xu(vbool16_t op0, vuint16m1_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_xu_f_v_u16m2))) +vuint16m2_t vfcvt_rtz_xu(vfloat16m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_xu_f_v_u16m2_m))) +vuint16m2_t vfcvt_rtz_xu(vbool8_t op0, vuint16m2_t op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_xu_f_v_u16m4))) +vuint16m4_t vfcvt_rtz_xu(vfloat16m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_xu_f_v_u16m4_m))) +vuint16m4_t vfcvt_rtz_xu(vbool4_t op0, vuint16m4_t op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_xu_f_v_u16m8))) +vuint16m8_t vfcvt_rtz_xu(vfloat16m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_xu_f_v_u16m8_m))) +vuint16m8_t vfcvt_rtz_xu(vbool2_t op0, vuint16m8_t op1, vfloat16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_xu_f_v_u16mf2))) +vuint16mf2_t vfcvt_rtz_xu(vfloat16mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_xu_f_v_u16mf2_m))) +vuint16mf2_t vfcvt_rtz_xu(vbool32_t op0, vuint16mf2_t op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_xu_f_v_u16mf4))) +vuint16mf4_t vfcvt_rtz_xu(vfloat16mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_rtz_xu_f_v_u16mf4_m))) +vuint16mf4_t vfcvt_rtz_xu(vbool64_t op0, vuint16mf4_t op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_x_f_v_i16m1))) +vint16m1_t vfcvt_x(vfloat16m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_x_f_v_i16m1_m))) +vint16m1_t vfcvt_x(vbool16_t op0, vint16m1_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_x_f_v_i16m2))) +vint16m2_t vfcvt_x(vfloat16m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_x_f_v_i16m2_m))) +vint16m2_t vfcvt_x(vbool8_t op0, vint16m2_t op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_x_f_v_i16m4))) +vint16m4_t vfcvt_x(vfloat16m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_x_f_v_i16m4_m))) +vint16m4_t vfcvt_x(vbool4_t op0, vint16m4_t op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_x_f_v_i16m8))) +vint16m8_t vfcvt_x(vfloat16m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_x_f_v_i16m8_m))) +vint16m8_t vfcvt_x(vbool2_t op0, vint16m8_t op1, vfloat16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_x_f_v_i16mf2))) +vint16mf2_t vfcvt_x(vfloat16mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_x_f_v_i16mf2_m))) +vint16mf2_t vfcvt_x(vbool32_t op0, vint16mf2_t op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_x_f_v_i16mf4))) +vint16mf4_t vfcvt_x(vfloat16mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_x_f_v_i16mf4_m))) +vint16mf4_t vfcvt_x(vbool64_t op0, vint16mf4_t op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_xu_f_v_u16m1))) +vuint16m1_t vfcvt_xu(vfloat16m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_xu_f_v_u16m1_m))) +vuint16m1_t vfcvt_xu(vbool16_t op0, vuint16m1_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_xu_f_v_u16m2))) +vuint16m2_t vfcvt_xu(vfloat16m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_xu_f_v_u16m2_m))) +vuint16m2_t vfcvt_xu(vbool8_t op0, vuint16m2_t op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_xu_f_v_u16m4))) +vuint16m4_t vfcvt_xu(vfloat16m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_xu_f_v_u16m4_m))) +vuint16m4_t vfcvt_xu(vbool4_t op0, vuint16m4_t op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_xu_f_v_u16m8))) +vuint16m8_t vfcvt_xu(vfloat16m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_xu_f_v_u16m8_m))) +vuint16m8_t vfcvt_xu(vbool2_t op0, vuint16m8_t op1, vfloat16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_xu_f_v_u16mf2))) +vuint16mf2_t vfcvt_xu(vfloat16mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_xu_f_v_u16mf2_m))) +vuint16mf2_t vfcvt_xu(vbool32_t op0, vuint16mf2_t op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_xu_f_v_u16mf4))) +vuint16mf4_t vfcvt_xu(vfloat16mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfcvt_xu_f_v_u16mf4_m))) +vuint16mf4_t vfcvt_xu(vbool64_t op0, vuint16mf4_t op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_x_w_f16mf4))) +vfloat16mf4_t vfncvt_f(vint32mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_x_w_f16mf4_m))) +vfloat16mf4_t vfncvt_f(vbool64_t op0, vfloat16mf4_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_x_w_f16mf2))) +vfloat16mf2_t vfncvt_f(vint32m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_x_w_f16mf2_m))) +vfloat16mf2_t vfncvt_f(vbool32_t op0, vfloat16mf2_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_x_w_f16m1))) +vfloat16m1_t vfncvt_f(vint32m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_x_w_f16m1_m))) +vfloat16m1_t vfncvt_f(vbool16_t op0, vfloat16m1_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_x_w_f16m2))) +vfloat16m2_t vfncvt_f(vint32m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_x_w_f16m2_m))) +vfloat16m2_t vfncvt_f(vbool8_t op0, vfloat16m2_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_x_w_f16m4))) +vfloat16m4_t vfncvt_f(vint32m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_x_w_f16m4_m))) +vfloat16m4_t vfncvt_f(vbool4_t op0, vfloat16m4_t op1, vint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_xu_w_f16mf4))) +vfloat16mf4_t vfncvt_f(vuint32mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_xu_w_f16mf4_m))) +vfloat16mf4_t vfncvt_f(vbool64_t op0, vfloat16mf4_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_xu_w_f16mf2))) +vfloat16mf2_t vfncvt_f(vuint32m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_xu_w_f16mf2_m))) +vfloat16mf2_t vfncvt_f(vbool32_t op0, vfloat16mf2_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_xu_w_f16m1))) +vfloat16m1_t vfncvt_f(vuint32m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_xu_w_f16m1_m))) +vfloat16m1_t vfncvt_f(vbool16_t op0, vfloat16m1_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_xu_w_f16m2))) +vfloat16m2_t vfncvt_f(vuint32m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_xu_w_f16m2_m))) +vfloat16m2_t vfncvt_f(vbool8_t op0, vfloat16m2_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_xu_w_f16m4))) +vfloat16m4_t vfncvt_f(vuint32m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_xu_w_f16m4_m))) +vfloat16m4_t vfncvt_f(vbool4_t op0, vfloat16m4_t op1, vuint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_x_f_w_i8mf8))) +vint8mf8_t vfncvt_rtz_x(vfloat16mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_x_f_w_i8mf8_m))) +vint8mf8_t vfncvt_rtz_x(vbool64_t op0, vint8mf8_t op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_x_f_w_i8mf4))) +vint8mf4_t vfncvt_rtz_x(vfloat16mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_x_f_w_i8mf4_m))) +vint8mf4_t vfncvt_rtz_x(vbool32_t op0, vint8mf4_t op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_x_f_w_i8mf2))) +vint8mf2_t vfncvt_rtz_x(vfloat16m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_x_f_w_i8mf2_m))) +vint8mf2_t vfncvt_rtz_x(vbool16_t op0, vint8mf2_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_x_f_w_i8m1))) +vint8m1_t vfncvt_rtz_x(vfloat16m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_x_f_w_i8m1_m))) +vint8m1_t vfncvt_rtz_x(vbool8_t op0, vint8m1_t op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_x_f_w_i8m2))) +vint8m2_t vfncvt_rtz_x(vfloat16m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_x_f_w_i8m2_m))) +vint8m2_t vfncvt_rtz_x(vbool4_t op0, vint8m2_t op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_x_f_w_i8m4))) +vint8m4_t vfncvt_rtz_x(vfloat16m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_x_f_w_i8m4_m))) +vint8m4_t vfncvt_rtz_x(vbool2_t op0, vint8m4_t op1, vfloat16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_xu_f_w_u8mf8))) +vuint8mf8_t vfncvt_rtz_xu(vfloat16mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_xu_f_w_u8mf8_m))) +vuint8mf8_t vfncvt_rtz_xu(vbool64_t op0, vuint8mf8_t op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_xu_f_w_u8mf4))) +vuint8mf4_t vfncvt_rtz_xu(vfloat16mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_xu_f_w_u8mf4_m))) +vuint8mf4_t vfncvt_rtz_xu(vbool32_t op0, vuint8mf4_t op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_xu_f_w_u8mf2))) +vuint8mf2_t vfncvt_rtz_xu(vfloat16m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_xu_f_w_u8mf2_m))) +vuint8mf2_t vfncvt_rtz_xu(vbool16_t op0, vuint8mf2_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_xu_f_w_u8m1))) +vuint8m1_t vfncvt_rtz_xu(vfloat16m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_xu_f_w_u8m1_m))) +vuint8m1_t vfncvt_rtz_xu(vbool8_t op0, vuint8m1_t op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_xu_f_w_u8m2))) +vuint8m2_t vfncvt_rtz_xu(vfloat16m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_xu_f_w_u8m2_m))) +vuint8m2_t vfncvt_rtz_xu(vbool4_t op0, vuint8m2_t op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_xu_f_w_u8m4))) +vuint8m4_t vfncvt_rtz_xu(vfloat16m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rtz_xu_f_w_u8m4_m))) +vuint8m4_t vfncvt_rtz_xu(vbool2_t op0, vuint8m4_t op1, vfloat16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_x_f_w_i8mf8))) +vint8mf8_t vfncvt_x(vfloat16mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_x_f_w_i8mf8_m))) +vint8mf8_t vfncvt_x(vbool64_t op0, vint8mf8_t op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_x_f_w_i8mf4))) +vint8mf4_t vfncvt_x(vfloat16mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_x_f_w_i8mf4_m))) +vint8mf4_t vfncvt_x(vbool32_t op0, vint8mf4_t op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_x_f_w_i8mf2))) +vint8mf2_t vfncvt_x(vfloat16m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_x_f_w_i8mf2_m))) +vint8mf2_t vfncvt_x(vbool16_t op0, vint8mf2_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_x_f_w_i8m1))) +vint8m1_t vfncvt_x(vfloat16m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_x_f_w_i8m1_m))) +vint8m1_t vfncvt_x(vbool8_t op0, vint8m1_t op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_x_f_w_i8m2))) +vint8m2_t vfncvt_x(vfloat16m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_x_f_w_i8m2_m))) +vint8m2_t vfncvt_x(vbool4_t op0, vint8m2_t op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_x_f_w_i8m4))) +vint8m4_t vfncvt_x(vfloat16m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_x_f_w_i8m4_m))) +vint8m4_t vfncvt_x(vbool2_t op0, vint8m4_t op1, vfloat16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_xu_f_w_u8mf8))) +vuint8mf8_t vfncvt_xu(vfloat16mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_xu_f_w_u8mf8_m))) +vuint8mf8_t vfncvt_xu(vbool64_t op0, vuint8mf8_t op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_xu_f_w_u8mf4))) +vuint8mf4_t vfncvt_xu(vfloat16mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_xu_f_w_u8mf4_m))) +vuint8mf4_t vfncvt_xu(vbool32_t op0, vuint8mf4_t op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_xu_f_w_u8mf2))) +vuint8mf2_t vfncvt_xu(vfloat16m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_xu_f_w_u8mf2_m))) +vuint8mf2_t vfncvt_xu(vbool16_t op0, vuint8mf2_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_xu_f_w_u8m1))) +vuint8m1_t vfncvt_xu(vfloat16m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_xu_f_w_u8m1_m))) +vuint8m1_t vfncvt_xu(vbool8_t op0, vuint8m1_t op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_xu_f_w_u8m2))) +vuint8m2_t vfncvt_xu(vfloat16m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_xu_f_w_u8m2_m))) +vuint8m2_t vfncvt_xu(vbool4_t op0, vuint8m2_t op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_xu_f_w_u8m4))) +vuint8m4_t vfncvt_xu(vfloat16m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_xu_f_w_u8m4_m))) +vuint8m4_t vfncvt_xu(vbool2_t op0, vuint8m4_t op1, vfloat16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrec7_v_f16m1))) +vfloat16m1_t vfrec7(vfloat16m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrec7_v_f16m1_m))) +vfloat16m1_t vfrec7(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrec7_v_f16m2))) +vfloat16m2_t vfrec7(vfloat16m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrec7_v_f16m2_m))) +vfloat16m2_t vfrec7(vbool8_t op0, vfloat16m2_t op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrec7_v_f16m4))) +vfloat16m4_t vfrec7(vfloat16m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrec7_v_f16m4_m))) +vfloat16m4_t vfrec7(vbool4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrec7_v_f16m8))) +vfloat16m8_t vfrec7(vfloat16m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrec7_v_f16m8_m))) +vfloat16m8_t vfrec7(vbool2_t op0, vfloat16m8_t op1, vfloat16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrec7_v_f16mf2))) +vfloat16mf2_t vfrec7(vfloat16mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrec7_v_f16mf2_m))) +vfloat16mf2_t vfrec7(vbool32_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrec7_v_f16mf4))) +vfloat16mf4_t vfrec7(vfloat16mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrec7_v_f16mf4_m))) +vfloat16mf4_t vfrec7(vbool64_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsqrt7_v_f16m1))) +vfloat16m1_t vfrsqrt7(vfloat16m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsqrt7_v_f16m1_m))) +vfloat16m1_t vfrsqrt7(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsqrt7_v_f16m2))) +vfloat16m2_t vfrsqrt7(vfloat16m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsqrt7_v_f16m2_m))) +vfloat16m2_t vfrsqrt7(vbool8_t op0, vfloat16m2_t op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsqrt7_v_f16m4))) +vfloat16m4_t vfrsqrt7(vfloat16m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsqrt7_v_f16m4_m))) +vfloat16m4_t vfrsqrt7(vbool4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsqrt7_v_f16m8))) +vfloat16m8_t vfrsqrt7(vfloat16m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsqrt7_v_f16m8_m))) +vfloat16m8_t vfrsqrt7(vbool2_t op0, vfloat16m8_t op1, vfloat16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsqrt7_v_f16mf2))) +vfloat16mf2_t vfrsqrt7(vfloat16mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsqrt7_v_f16mf2_m))) +vfloat16mf2_t vfrsqrt7(vbool32_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsqrt7_v_f16mf4))) +vfloat16mf4_t vfrsqrt7(vfloat16mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfrsqrt7_v_f16mf4_m))) +vfloat16mf4_t vfrsqrt7(vbool64_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsqrt_v_f16m1))) +vfloat16m1_t vfsqrt(vfloat16m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsqrt_v_f16m1_m))) +vfloat16m1_t vfsqrt(vbool16_t op0, vfloat16m1_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsqrt_v_f16m2))) +vfloat16m2_t vfsqrt(vfloat16m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsqrt_v_f16m2_m))) +vfloat16m2_t vfsqrt(vbool8_t op0, vfloat16m2_t op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsqrt_v_f16m4))) +vfloat16m4_t vfsqrt(vfloat16m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsqrt_v_f16m4_m))) +vfloat16m4_t vfsqrt(vbool4_t op0, vfloat16m4_t op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsqrt_v_f16m8))) +vfloat16m8_t vfsqrt(vfloat16m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsqrt_v_f16m8_m))) +vfloat16m8_t vfsqrt(vbool2_t op0, vfloat16m8_t op1, vfloat16m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsqrt_v_f16mf2))) +vfloat16mf2_t vfsqrt(vfloat16mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsqrt_v_f16mf2_m))) +vfloat16mf2_t vfsqrt(vbool32_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsqrt_v_f16mf4))) +vfloat16mf4_t vfsqrt(vfloat16mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfsqrt_v_f16mf4_m))) +vfloat16mf4_t vfsqrt(vbool64_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_x_v_f16mf4))) +vfloat16mf4_t vfwcvt_f(vint8mf8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_x_v_f16mf4_m))) +vfloat16mf4_t vfwcvt_f(vbool64_t op0, vfloat16mf4_t op1, vint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_x_v_f16mf2))) +vfloat16mf2_t vfwcvt_f(vint8mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_x_v_f16mf2_m))) +vfloat16mf2_t vfwcvt_f(vbool32_t op0, vfloat16mf2_t op1, vint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_x_v_f16m1))) +vfloat16m1_t vfwcvt_f(vint8mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_x_v_f16m1_m))) +vfloat16m1_t vfwcvt_f(vbool16_t op0, vfloat16m1_t op1, vint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_x_v_f16m2))) +vfloat16m2_t vfwcvt_f(vint8m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_x_v_f16m2_m))) +vfloat16m2_t vfwcvt_f(vbool8_t op0, vfloat16m2_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_x_v_f16m4))) +vfloat16m4_t vfwcvt_f(vint8m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_x_v_f16m4_m))) +vfloat16m4_t vfwcvt_f(vbool4_t op0, vfloat16m4_t op1, vint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_x_v_f16m8))) +vfloat16m8_t vfwcvt_f(vint8m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_x_v_f16m8_m))) +vfloat16m8_t vfwcvt_f(vbool2_t op0, vfloat16m8_t op1, vint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_xu_v_f16mf4))) +vfloat16mf4_t vfwcvt_f(vuint8mf8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_xu_v_f16mf4_m))) +vfloat16mf4_t vfwcvt_f(vbool64_t op0, vfloat16mf4_t op1, vuint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_xu_v_f16mf2))) +vfloat16mf2_t vfwcvt_f(vuint8mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_xu_v_f16mf2_m))) +vfloat16mf2_t vfwcvt_f(vbool32_t op0, vfloat16mf2_t op1, vuint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_xu_v_f16m1))) +vfloat16m1_t vfwcvt_f(vuint8mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_xu_v_f16m1_m))) +vfloat16m1_t vfwcvt_f(vbool16_t op0, vfloat16m1_t op1, vuint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_xu_v_f16m2))) +vfloat16m2_t vfwcvt_f(vuint8m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_xu_v_f16m2_m))) +vfloat16m2_t vfwcvt_f(vbool8_t op0, vfloat16m2_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_xu_v_f16m4))) +vfloat16m4_t vfwcvt_f(vuint8m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_xu_v_f16m4_m))) +vfloat16m4_t vfwcvt_f(vbool4_t op0, vfloat16m4_t op1, vuint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_xu_v_f16m8))) +vfloat16m8_t vfwcvt_f(vuint8m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_xu_v_f16m8_m))) +vfloat16m8_t vfwcvt_f(vbool2_t op0, vfloat16m8_t op1, vuint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_rtz_x_f_v_i32mf2))) +vint32mf2_t vfwcvt_rtz_x(vfloat16mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_rtz_x_f_v_i32mf2_m))) +vint32mf2_t vfwcvt_rtz_x(vbool64_t op0, vint32mf2_t op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_rtz_x_f_v_i32m1))) +vint32m1_t vfwcvt_rtz_x(vfloat16mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_rtz_x_f_v_i32m1_m))) +vint32m1_t vfwcvt_rtz_x(vbool32_t op0, vint32m1_t op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_rtz_x_f_v_i32m2))) +vint32m2_t vfwcvt_rtz_x(vfloat16m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_rtz_x_f_v_i32m2_m))) +vint32m2_t vfwcvt_rtz_x(vbool16_t op0, vint32m2_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_rtz_x_f_v_i32m4))) +vint32m4_t vfwcvt_rtz_x(vfloat16m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_rtz_x_f_v_i32m4_m))) +vint32m4_t vfwcvt_rtz_x(vbool8_t op0, vint32m4_t op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_rtz_x_f_v_i32m8))) +vint32m8_t vfwcvt_rtz_x(vfloat16m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_rtz_x_f_v_i32m8_m))) +vint32m8_t vfwcvt_rtz_x(vbool4_t op0, vint32m8_t op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_rtz_xu_f_v_u32mf2))) +vuint32mf2_t vfwcvt_rtz_xu(vfloat16mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_rtz_xu_f_v_u32mf2_m))) +vuint32mf2_t vfwcvt_rtz_xu(vbool64_t op0, vuint32mf2_t op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_rtz_xu_f_v_u32m1))) +vuint32m1_t vfwcvt_rtz_xu(vfloat16mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_rtz_xu_f_v_u32m1_m))) +vuint32m1_t vfwcvt_rtz_xu(vbool32_t op0, vuint32m1_t op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_rtz_xu_f_v_u32m2))) +vuint32m2_t vfwcvt_rtz_xu(vfloat16m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_rtz_xu_f_v_u32m2_m))) +vuint32m2_t vfwcvt_rtz_xu(vbool16_t op0, vuint32m2_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_rtz_xu_f_v_u32m4))) +vuint32m4_t vfwcvt_rtz_xu(vfloat16m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_rtz_xu_f_v_u32m4_m))) +vuint32m4_t vfwcvt_rtz_xu(vbool8_t op0, vuint32m4_t op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_rtz_xu_f_v_u32m8))) +vuint32m8_t vfwcvt_rtz_xu(vfloat16m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_rtz_xu_f_v_u32m8_m))) +vuint32m8_t vfwcvt_rtz_xu(vbool4_t op0, vuint32m8_t op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_x_f_v_i32mf2))) +vint32mf2_t vfwcvt_x(vfloat16mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_x_f_v_i32mf2_m))) +vint32mf2_t vfwcvt_x(vbool64_t op0, vint32mf2_t op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_x_f_v_i32m1))) +vint32m1_t vfwcvt_x(vfloat16mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_x_f_v_i32m1_m))) +vint32m1_t vfwcvt_x(vbool32_t op0, vint32m1_t op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_x_f_v_i32m2))) +vint32m2_t vfwcvt_x(vfloat16m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_x_f_v_i32m2_m))) +vint32m2_t vfwcvt_x(vbool16_t op0, vint32m2_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_x_f_v_i32m4))) +vint32m4_t vfwcvt_x(vfloat16m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_x_f_v_i32m4_m))) +vint32m4_t vfwcvt_x(vbool8_t op0, vint32m4_t op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_x_f_v_i32m8))) +vint32m8_t vfwcvt_x(vfloat16m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_x_f_v_i32m8_m))) +vint32m8_t vfwcvt_x(vbool4_t op0, vint32m8_t op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_xu_f_v_u32mf2))) +vuint32mf2_t vfwcvt_xu(vfloat16mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_xu_f_v_u32mf2_m))) +vuint32mf2_t vfwcvt_xu(vbool64_t op0, vuint32mf2_t op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_xu_f_v_u32m1))) +vuint32m1_t vfwcvt_xu(vfloat16mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_xu_f_v_u32m1_m))) +vuint32m1_t vfwcvt_xu(vbool32_t op0, vuint32m1_t op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_xu_f_v_u32m2))) +vuint32m2_t vfwcvt_xu(vfloat16m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_xu_f_v_u32m2_m))) +vuint32m2_t vfwcvt_xu(vbool16_t op0, vuint32m2_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_xu_f_v_u32m4))) +vuint32m4_t vfwcvt_xu(vfloat16m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_xu_f_v_u32m4_m))) +vuint32m4_t vfwcvt_xu(vbool8_t op0, vuint32m4_t op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_xu_f_v_u32m8))) +vuint32m8_t vfwcvt_xu(vfloat16m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_xu_f_v_u32m8_m))) +vuint32m8_t vfwcvt_xu(vbool4_t op0, vuint32m8_t op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_f16mf2_f16m1))) +vfloat16m1_t vlmul_ext_f16m1(vfloat16mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_f16mf4_f16m1))) +vfloat16m1_t vlmul_ext_f16m1(vfloat16mf4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_f16mf4_f16mf2))) +vfloat16mf2_t vlmul_ext_f16mf2(vfloat16mf4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_f16m1_f16m2))) +vfloat16m2_t vlmul_ext_f16m2(vfloat16m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_f16mf2_f16m2))) +vfloat16m2_t vlmul_ext_f16m2(vfloat16mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_f16mf4_f16m2))) +vfloat16m2_t vlmul_ext_f16m2(vfloat16mf4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_f16m1_f16m4))) +vfloat16m4_t vlmul_ext_f16m4(vfloat16m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_f16m2_f16m4))) +vfloat16m4_t vlmul_ext_f16m4(vfloat16m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_f16mf2_f16m4))) +vfloat16m4_t vlmul_ext_f16m4(vfloat16mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_f16mf4_f16m4))) +vfloat16m4_t vlmul_ext_f16m4(vfloat16mf4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_f16m1_f16m8))) +vfloat16m8_t vlmul_ext_f16m8(vfloat16m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_f16m2_f16m8))) +vfloat16m8_t vlmul_ext_f16m8(vfloat16m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_f16m4_f16m8))) +vfloat16m8_t vlmul_ext_f16m8(vfloat16m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_f16mf2_f16m8))) +vfloat16m8_t vlmul_ext_f16m8(vfloat16mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_ext_v_f16mf4_f16m8))) +vfloat16m8_t vlmul_ext_f16m8(vfloat16mf4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_f16m1_f16mf2))) +vfloat16mf2_t vlmul_trunc_f16mf2(vfloat16m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_f16m2_f16mf2))) +vfloat16mf2_t vlmul_trunc_f16mf2(vfloat16m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_f16m4_f16mf2))) +vfloat16mf2_t vlmul_trunc_f16mf2(vfloat16m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_f16m8_f16mf2))) +vfloat16mf2_t vlmul_trunc_f16mf2(vfloat16m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_f16m1_f16mf4))) +vfloat16mf4_t vlmul_trunc_f16mf4(vfloat16m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_f16m2_f16mf4))) +vfloat16mf4_t vlmul_trunc_f16mf4(vfloat16m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_f16m4_f16mf4))) +vfloat16mf4_t vlmul_trunc_f16mf4(vfloat16m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_f16m8_f16mf4))) +vfloat16mf4_t vlmul_trunc_f16mf4(vfloat16m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_f16mf2_f16mf4))) +vfloat16mf4_t vlmul_trunc_f16mf4(vfloat16mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_f16m2_f16m1))) +vfloat16m1_t vlmul_trunc_f16m1(vfloat16m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_f16m4_f16m1))) +vfloat16m1_t vlmul_trunc_f16m1(vfloat16m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_f16m8_f16m1))) +vfloat16m1_t vlmul_trunc_f16m1(vfloat16m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_f16m4_f16m2))) +vfloat16m2_t vlmul_trunc_f16m2(vfloat16m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_f16m8_f16m2))) +vfloat16m2_t vlmul_trunc_f16m2(vfloat16m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlmul_trunc_v_f16m8_f16m4))) +vfloat16m4_t vlmul_trunc_f16m4(vfloat16m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i16m1_f16m1))) +vfloat16m1_t vreinterpret_f16m1(vint16m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i16m2_f16m2))) +vfloat16m2_t vreinterpret_f16m2(vint16m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i16m4_f16m4))) +vfloat16m4_t vreinterpret_f16m4(vint16m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i16m8_f16m8))) +vfloat16m8_t vreinterpret_f16m8(vint16m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i16mf2_f16mf2))) +vfloat16mf2_t vreinterpret_f16mf2(vint16mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_i16mf4_f16mf4))) +vfloat16mf4_t vreinterpret_f16mf4(vint16mf4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u16m1_f16m1))) +vfloat16m1_t vreinterpret_f16m1(vuint16m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u16m2_f16m2))) +vfloat16m2_t vreinterpret_f16m2(vuint16m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u16m4_f16m4))) +vfloat16m4_t vreinterpret_f16m4(vuint16m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u16m8_f16m8))) +vfloat16m8_t vreinterpret_f16m8(vuint16m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u16mf2_f16mf2))) +vfloat16mf2_t vreinterpret_f16mf2(vuint16mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_u16mf4_f16mf4))) +vfloat16mf4_t vreinterpret_f16mf4(vuint16mf4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_f16m1_i16m1))) +vint16m1_t vreinterpret_i16m1(vfloat16m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_f16m2_i16m2))) +vint16m2_t vreinterpret_i16m2(vfloat16m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_f16m4_i16m4))) +vint16m4_t vreinterpret_i16m4(vfloat16m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_f16m8_i16m8))) +vint16m8_t vreinterpret_i16m8(vfloat16m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_f16mf2_i16mf2))) +vint16mf2_t vreinterpret_i16mf2(vfloat16mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_f16mf4_i16mf4))) +vint16mf4_t vreinterpret_i16mf4(vfloat16mf4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_f16m1_u16m1))) +vuint16m1_t vreinterpret_u16m1(vfloat16m1_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_f16m2_u16m2))) +vuint16m2_t vreinterpret_u16m2(vfloat16m2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_f16m4_u16m4))) +vuint16m4_t vreinterpret_u16m4(vfloat16m4_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_f16m8_u16m8))) +vuint16m8_t vreinterpret_u16m8(vfloat16m8_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_f16mf2_u16mf2))) +vuint16mf2_t vreinterpret_u16mf2(vfloat16mf2_t op0); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vreinterpret_v_f16mf4_u16mf4))) +vuint16mf4_t vreinterpret_u16mf4(vfloat16mf4_t op0); + +#endif + +#if defined(__riscv_f) && defined(__riscv_zfh) +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_vv_f32mf2))) +vfloat32mf2_t vfwadd_vv(vfloat16mf4_t op0, vfloat16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_vv_f32mf2_m))) +vfloat32mf2_t vfwadd_vv(vbool64_t op0, vfloat32mf2_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_vv_f32m1))) +vfloat32m1_t vfwadd_vv(vfloat16mf2_t op0, vfloat16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_vv_f32m1_m))) +vfloat32m1_t vfwadd_vv(vbool32_t op0, vfloat32m1_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_vv_f32m2))) +vfloat32m2_t vfwadd_vv(vfloat16m1_t op0, vfloat16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_vv_f32m2_m))) +vfloat32m2_t vfwadd_vv(vbool16_t op0, vfloat32m2_t op1, vfloat16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_vv_f32m4))) +vfloat32m4_t vfwadd_vv(vfloat16m2_t op0, vfloat16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_vv_f32m4_m))) +vfloat32m4_t vfwadd_vv(vbool8_t op0, vfloat32m4_t op1, vfloat16m2_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_vv_f32m8))) +vfloat32m8_t vfwadd_vv(vfloat16m4_t op0, vfloat16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_vv_f32m8_m))) +vfloat32m8_t vfwadd_vv(vbool4_t op0, vfloat32m8_t op1, vfloat16m4_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_vf_f32mf2))) +vfloat32mf2_t vfwadd_vf(vfloat16mf4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_vf_f32mf2_m))) +vfloat32mf2_t vfwadd_vf(vbool64_t op0, vfloat32mf2_t op1, vfloat16mf4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_vf_f32m1))) +vfloat32m1_t vfwadd_vf(vfloat16mf2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_vf_f32m1_m))) +vfloat32m1_t vfwadd_vf(vbool32_t op0, vfloat32m1_t op1, vfloat16mf2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_vf_f32m2))) +vfloat32m2_t vfwadd_vf(vfloat16m1_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_vf_f32m2_m))) +vfloat32m2_t vfwadd_vf(vbool16_t op0, vfloat32m2_t op1, vfloat16m1_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_vf_f32m4))) +vfloat32m4_t vfwadd_vf(vfloat16m2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_vf_f32m4_m))) +vfloat32m4_t vfwadd_vf(vbool8_t op0, vfloat32m4_t op1, vfloat16m2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_vf_f32m8))) +vfloat32m8_t vfwadd_vf(vfloat16m4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_vf_f32m8_m))) +vfloat32m8_t vfwadd_vf(vbool4_t op0, vfloat32m8_t op1, vfloat16m4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_vv_f32mf2))) +vfloat32mf2_t vfwsub_vv(vfloat16mf4_t op0, vfloat16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_vv_f32mf2_m))) +vfloat32mf2_t vfwsub_vv(vbool64_t op0, vfloat32mf2_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_vv_f32m1))) +vfloat32m1_t vfwsub_vv(vfloat16mf2_t op0, vfloat16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_vv_f32m1_m))) +vfloat32m1_t vfwsub_vv(vbool32_t op0, vfloat32m1_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_vv_f32m2))) +vfloat32m2_t vfwsub_vv(vfloat16m1_t op0, vfloat16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_vv_f32m2_m))) +vfloat32m2_t vfwsub_vv(vbool16_t op0, vfloat32m2_t op1, vfloat16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_vv_f32m4))) +vfloat32m4_t vfwsub_vv(vfloat16m2_t op0, vfloat16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_vv_f32m4_m))) +vfloat32m4_t vfwsub_vv(vbool8_t op0, vfloat32m4_t op1, vfloat16m2_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_vv_f32m8))) +vfloat32m8_t vfwsub_vv(vfloat16m4_t op0, vfloat16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_vv_f32m8_m))) +vfloat32m8_t vfwsub_vv(vbool4_t op0, vfloat32m8_t op1, vfloat16m4_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_vf_f32mf2))) +vfloat32mf2_t vfwsub_vf(vfloat16mf4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_vf_f32mf2_m))) +vfloat32mf2_t vfwsub_vf(vbool64_t op0, vfloat32mf2_t op1, vfloat16mf4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_vf_f32m1))) +vfloat32m1_t vfwsub_vf(vfloat16mf2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_vf_f32m1_m))) +vfloat32m1_t vfwsub_vf(vbool32_t op0, vfloat32m1_t op1, vfloat16mf2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_vf_f32m2))) +vfloat32m2_t vfwsub_vf(vfloat16m1_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_vf_f32m2_m))) +vfloat32m2_t vfwsub_vf(vbool16_t op0, vfloat32m2_t op1, vfloat16m1_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_vf_f32m4))) +vfloat32m4_t vfwsub_vf(vfloat16m2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_vf_f32m4_m))) +vfloat32m4_t vfwsub_vf(vbool8_t op0, vfloat32m4_t op1, vfloat16m2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_vf_f32m8))) +vfloat32m8_t vfwsub_vf(vfloat16m4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_vf_f32m8_m))) +vfloat32m8_t vfwsub_vf(vbool4_t op0, vfloat32m8_t op1, vfloat16m4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_wv_f32mf2))) +vfloat32mf2_t vfwadd_wv(vfloat32mf2_t op0, vfloat16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_wv_f32mf2_m))) +vfloat32mf2_t vfwadd_wv(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_wv_f32m1))) +vfloat32m1_t vfwadd_wv(vfloat32m1_t op0, vfloat16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_wv_f32m1_m))) +vfloat32m1_t vfwadd_wv(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_wv_f32m2))) +vfloat32m2_t vfwadd_wv(vfloat32m2_t op0, vfloat16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_wv_f32m2_m))) +vfloat32m2_t vfwadd_wv(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_wv_f32m4))) +vfloat32m4_t vfwadd_wv(vfloat32m4_t op0, vfloat16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_wv_f32m4_m))) +vfloat32m4_t vfwadd_wv(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_wv_f32m8))) +vfloat32m8_t vfwadd_wv(vfloat32m8_t op0, vfloat16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_wv_f32m8_m))) +vfloat32m8_t vfwadd_wv(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_wf_f32mf2))) +vfloat32mf2_t vfwadd_wf(vfloat32mf2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_wf_f32mf2_m))) +vfloat32mf2_t vfwadd_wf(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_wf_f32m1))) +vfloat32m1_t vfwadd_wf(vfloat32m1_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_wf_f32m1_m))) +vfloat32m1_t vfwadd_wf(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_wf_f32m2))) +vfloat32m2_t vfwadd_wf(vfloat32m2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_wf_f32m2_m))) +vfloat32m2_t vfwadd_wf(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_wf_f32m4))) +vfloat32m4_t vfwadd_wf(vfloat32m4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_wf_f32m4_m))) +vfloat32m4_t vfwadd_wf(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_wf_f32m8))) +vfloat32m8_t vfwadd_wf(vfloat32m8_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwadd_wf_f32m8_m))) +vfloat32m8_t vfwadd_wf(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_wv_f32mf2))) +vfloat32mf2_t vfwsub_wv(vfloat32mf2_t op0, vfloat16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_wv_f32mf2_m))) +vfloat32mf2_t vfwsub_wv(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_wv_f32m1))) +vfloat32m1_t vfwsub_wv(vfloat32m1_t op0, vfloat16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_wv_f32m1_m))) +vfloat32m1_t vfwsub_wv(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_wv_f32m2))) +vfloat32m2_t vfwsub_wv(vfloat32m2_t op0, vfloat16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_wv_f32m2_m))) +vfloat32m2_t vfwsub_wv(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_wv_f32m4))) +vfloat32m4_t vfwsub_wv(vfloat32m4_t op0, vfloat16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_wv_f32m4_m))) +vfloat32m4_t vfwsub_wv(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_wv_f32m8))) +vfloat32m8_t vfwsub_wv(vfloat32m8_t op0, vfloat16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_wv_f32m8_m))) +vfloat32m8_t vfwsub_wv(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_wf_f32mf2))) +vfloat32mf2_t vfwsub_wf(vfloat32mf2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_wf_f32mf2_m))) +vfloat32mf2_t vfwsub_wf(vbool64_t op0, vfloat32mf2_t op1, vfloat32mf2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_wf_f32m1))) +vfloat32m1_t vfwsub_wf(vfloat32m1_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_wf_f32m1_m))) +vfloat32m1_t vfwsub_wf(vbool32_t op0, vfloat32m1_t op1, vfloat32m1_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_wf_f32m2))) +vfloat32m2_t vfwsub_wf(vfloat32m2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_wf_f32m2_m))) +vfloat32m2_t vfwsub_wf(vbool16_t op0, vfloat32m2_t op1, vfloat32m2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_wf_f32m4))) +vfloat32m4_t vfwsub_wf(vfloat32m4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_wf_f32m4_m))) +vfloat32m4_t vfwsub_wf(vbool8_t op0, vfloat32m4_t op1, vfloat32m4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_wf_f32m8))) +vfloat32m8_t vfwsub_wf(vfloat32m8_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwsub_wf_f32m8_m))) +vfloat32m8_t vfwsub_wf(vbool4_t op0, vfloat32m8_t op1, vfloat32m8_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmul_vv_f32mf2))) +vfloat32mf2_t vfwmul(vfloat16mf4_t op0, vfloat16mf4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmul_vv_f32mf2_m))) +vfloat32mf2_t vfwmul(vbool64_t op0, vfloat32mf2_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmul_vv_f32m1))) +vfloat32m1_t vfwmul(vfloat16mf2_t op0, vfloat16mf2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmul_vv_f32m1_m))) +vfloat32m1_t vfwmul(vbool32_t op0, vfloat32m1_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmul_vv_f32m2))) +vfloat32m2_t vfwmul(vfloat16m1_t op0, vfloat16m1_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmul_vv_f32m2_m))) +vfloat32m2_t vfwmul(vbool16_t op0, vfloat32m2_t op1, vfloat16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmul_vv_f32m4))) +vfloat32m4_t vfwmul(vfloat16m2_t op0, vfloat16m2_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmul_vv_f32m4_m))) +vfloat32m4_t vfwmul(vbool8_t op0, vfloat32m4_t op1, vfloat16m2_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmul_vv_f32m8))) +vfloat32m8_t vfwmul(vfloat16m4_t op0, vfloat16m4_t op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmul_vv_f32m8_m))) +vfloat32m8_t vfwmul(vbool4_t op0, vfloat32m8_t op1, vfloat16m4_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmul_vf_f32mf2))) +vfloat32mf2_t vfwmul(vfloat16mf4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmul_vf_f32mf2_m))) +vfloat32mf2_t vfwmul(vbool64_t op0, vfloat32mf2_t op1, vfloat16mf4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmul_vf_f32m1))) +vfloat32m1_t vfwmul(vfloat16mf2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmul_vf_f32m1_m))) +vfloat32m1_t vfwmul(vbool32_t op0, vfloat32m1_t op1, vfloat16mf2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmul_vf_f32m2))) +vfloat32m2_t vfwmul(vfloat16m1_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmul_vf_f32m2_m))) +vfloat32m2_t vfwmul(vbool16_t op0, vfloat32m2_t op1, vfloat16m1_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmul_vf_f32m4))) +vfloat32m4_t vfwmul(vfloat16m2_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmul_vf_f32m4_m))) +vfloat32m4_t vfwmul(vbool8_t op0, vfloat32m4_t op1, vfloat16m2_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmul_vf_f32m8))) +vfloat32m8_t vfwmul(vfloat16m4_t op0, _Float16 op1, size_t op2); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmul_vf_f32m8_m))) +vfloat32m8_t vfwmul(vbool4_t op0, vfloat32m8_t op1, vfloat16m4_t op2, _Float16 op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmacc_vv_f32mf2))) +vfloat32mf2_t vfwmacc(vfloat32mf2_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmacc_vv_f32mf2_m))) +vfloat32mf2_t vfwmacc(vbool64_t op0, vfloat32mf2_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmacc_vv_f32m1))) +vfloat32m1_t vfwmacc(vfloat32m1_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmacc_vv_f32m1_m))) +vfloat32m1_t vfwmacc(vbool32_t op0, vfloat32m1_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmacc_vv_f32m2))) +vfloat32m2_t vfwmacc(vfloat32m2_t op0, vfloat16m1_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmacc_vv_f32m2_m))) +vfloat32m2_t vfwmacc(vbool16_t op0, vfloat32m2_t op1, vfloat16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmacc_vv_f32m4))) +vfloat32m4_t vfwmacc(vfloat32m4_t op0, vfloat16m2_t op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmacc_vv_f32m4_m))) +vfloat32m4_t vfwmacc(vbool8_t op0, vfloat32m4_t op1, vfloat16m2_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmacc_vv_f32m8))) +vfloat32m8_t vfwmacc(vfloat32m8_t op0, vfloat16m4_t op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmacc_vv_f32m8_m))) +vfloat32m8_t vfwmacc(vbool4_t op0, vfloat32m8_t op1, vfloat16m4_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmacc_vf_f32mf2))) +vfloat32mf2_t vfwmacc(vfloat32mf2_t op0, _Float16 op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmacc_vf_f32mf2_m))) +vfloat32mf2_t vfwmacc(vbool64_t op0, vfloat32mf2_t op1, _Float16 op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmacc_vf_f32m1))) +vfloat32m1_t vfwmacc(vfloat32m1_t op0, _Float16 op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmacc_vf_f32m1_m))) +vfloat32m1_t vfwmacc(vbool32_t op0, vfloat32m1_t op1, _Float16 op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmacc_vf_f32m2))) +vfloat32m2_t vfwmacc(vfloat32m2_t op0, _Float16 op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmacc_vf_f32m2_m))) +vfloat32m2_t vfwmacc(vbool16_t op0, vfloat32m2_t op1, _Float16 op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmacc_vf_f32m4))) +vfloat32m4_t vfwmacc(vfloat32m4_t op0, _Float16 op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmacc_vf_f32m4_m))) +vfloat32m4_t vfwmacc(vbool8_t op0, vfloat32m4_t op1, _Float16 op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmacc_vf_f32m8))) +vfloat32m8_t vfwmacc(vfloat32m8_t op0, _Float16 op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmacc_vf_f32m8_m))) +vfloat32m8_t vfwmacc(vbool4_t op0, vfloat32m8_t op1, _Float16 op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmacc_vv_f32mf2))) +vfloat32mf2_t vfwnmacc(vfloat32mf2_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmacc_vv_f32mf2_m))) +vfloat32mf2_t vfwnmacc(vbool64_t op0, vfloat32mf2_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmacc_vv_f32m1))) +vfloat32m1_t vfwnmacc(vfloat32m1_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmacc_vv_f32m1_m))) +vfloat32m1_t vfwnmacc(vbool32_t op0, vfloat32m1_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmacc_vv_f32m2))) +vfloat32m2_t vfwnmacc(vfloat32m2_t op0, vfloat16m1_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmacc_vv_f32m2_m))) +vfloat32m2_t vfwnmacc(vbool16_t op0, vfloat32m2_t op1, vfloat16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmacc_vv_f32m4))) +vfloat32m4_t vfwnmacc(vfloat32m4_t op0, vfloat16m2_t op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmacc_vv_f32m4_m))) +vfloat32m4_t vfwnmacc(vbool8_t op0, vfloat32m4_t op1, vfloat16m2_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmacc_vv_f32m8))) +vfloat32m8_t vfwnmacc(vfloat32m8_t op0, vfloat16m4_t op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmacc_vv_f32m8_m))) +vfloat32m8_t vfwnmacc(vbool4_t op0, vfloat32m8_t op1, vfloat16m4_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmacc_vf_f32mf2))) +vfloat32mf2_t vfwnmacc(vfloat32mf2_t op0, _Float16 op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmacc_vf_f32mf2_m))) +vfloat32mf2_t vfwnmacc(vbool64_t op0, vfloat32mf2_t op1, _Float16 op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmacc_vf_f32m1))) +vfloat32m1_t vfwnmacc(vfloat32m1_t op0, _Float16 op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmacc_vf_f32m1_m))) +vfloat32m1_t vfwnmacc(vbool32_t op0, vfloat32m1_t op1, _Float16 op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmacc_vf_f32m2))) +vfloat32m2_t vfwnmacc(vfloat32m2_t op0, _Float16 op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmacc_vf_f32m2_m))) +vfloat32m2_t vfwnmacc(vbool16_t op0, vfloat32m2_t op1, _Float16 op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmacc_vf_f32m4))) +vfloat32m4_t vfwnmacc(vfloat32m4_t op0, _Float16 op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmacc_vf_f32m4_m))) +vfloat32m4_t vfwnmacc(vbool8_t op0, vfloat32m4_t op1, _Float16 op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmacc_vf_f32m8))) +vfloat32m8_t vfwnmacc(vfloat32m8_t op0, _Float16 op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmacc_vf_f32m8_m))) +vfloat32m8_t vfwnmacc(vbool4_t op0, vfloat32m8_t op1, _Float16 op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmsac_vv_f32mf2))) +vfloat32mf2_t vfwmsac(vfloat32mf2_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmsac_vv_f32mf2_m))) +vfloat32mf2_t vfwmsac(vbool64_t op0, vfloat32mf2_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmsac_vv_f32m1))) +vfloat32m1_t vfwmsac(vfloat32m1_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmsac_vv_f32m1_m))) +vfloat32m1_t vfwmsac(vbool32_t op0, vfloat32m1_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmsac_vv_f32m2))) +vfloat32m2_t vfwmsac(vfloat32m2_t op0, vfloat16m1_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmsac_vv_f32m2_m))) +vfloat32m2_t vfwmsac(vbool16_t op0, vfloat32m2_t op1, vfloat16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmsac_vv_f32m4))) +vfloat32m4_t vfwmsac(vfloat32m4_t op0, vfloat16m2_t op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmsac_vv_f32m4_m))) +vfloat32m4_t vfwmsac(vbool8_t op0, vfloat32m4_t op1, vfloat16m2_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmsac_vv_f32m8))) +vfloat32m8_t vfwmsac(vfloat32m8_t op0, vfloat16m4_t op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmsac_vv_f32m8_m))) +vfloat32m8_t vfwmsac(vbool4_t op0, vfloat32m8_t op1, vfloat16m4_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmsac_vf_f32mf2))) +vfloat32mf2_t vfwmsac(vfloat32mf2_t op0, _Float16 op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmsac_vf_f32mf2_m))) +vfloat32mf2_t vfwmsac(vbool64_t op0, vfloat32mf2_t op1, _Float16 op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmsac_vf_f32m1))) +vfloat32m1_t vfwmsac(vfloat32m1_t op0, _Float16 op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmsac_vf_f32m1_m))) +vfloat32m1_t vfwmsac(vbool32_t op0, vfloat32m1_t op1, _Float16 op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmsac_vf_f32m2))) +vfloat32m2_t vfwmsac(vfloat32m2_t op0, _Float16 op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmsac_vf_f32m2_m))) +vfloat32m2_t vfwmsac(vbool16_t op0, vfloat32m2_t op1, _Float16 op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmsac_vf_f32m4))) +vfloat32m4_t vfwmsac(vfloat32m4_t op0, _Float16 op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmsac_vf_f32m4_m))) +vfloat32m4_t vfwmsac(vbool8_t op0, vfloat32m4_t op1, _Float16 op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmsac_vf_f32m8))) +vfloat32m8_t vfwmsac(vfloat32m8_t op0, _Float16 op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwmsac_vf_f32m8_m))) +vfloat32m8_t vfwmsac(vbool4_t op0, vfloat32m8_t op1, _Float16 op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmsac_vv_f32mf2))) +vfloat32mf2_t vfwnmsac(vfloat32mf2_t op0, vfloat16mf4_t op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmsac_vv_f32mf2_m))) +vfloat32mf2_t vfwnmsac(vbool64_t op0, vfloat32mf2_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmsac_vv_f32m1))) +vfloat32m1_t vfwnmsac(vfloat32m1_t op0, vfloat16mf2_t op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmsac_vv_f32m1_m))) +vfloat32m1_t vfwnmsac(vbool32_t op0, vfloat32m1_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmsac_vv_f32m2))) +vfloat32m2_t vfwnmsac(vfloat32m2_t op0, vfloat16m1_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmsac_vv_f32m2_m))) +vfloat32m2_t vfwnmsac(vbool16_t op0, vfloat32m2_t op1, vfloat16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmsac_vv_f32m4))) +vfloat32m4_t vfwnmsac(vfloat32m4_t op0, vfloat16m2_t op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmsac_vv_f32m4_m))) +vfloat32m4_t vfwnmsac(vbool8_t op0, vfloat32m4_t op1, vfloat16m2_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmsac_vv_f32m8))) +vfloat32m8_t vfwnmsac(vfloat32m8_t op0, vfloat16m4_t op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmsac_vv_f32m8_m))) +vfloat32m8_t vfwnmsac(vbool4_t op0, vfloat32m8_t op1, vfloat16m4_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmsac_vf_f32mf2))) +vfloat32mf2_t vfwnmsac(vfloat32mf2_t op0, _Float16 op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmsac_vf_f32mf2_m))) +vfloat32mf2_t vfwnmsac(vbool64_t op0, vfloat32mf2_t op1, _Float16 op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmsac_vf_f32m1))) +vfloat32m1_t vfwnmsac(vfloat32m1_t op0, _Float16 op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmsac_vf_f32m1_m))) +vfloat32m1_t vfwnmsac(vbool32_t op0, vfloat32m1_t op1, _Float16 op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmsac_vf_f32m2))) +vfloat32m2_t vfwnmsac(vfloat32m2_t op0, _Float16 op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmsac_vf_f32m2_m))) +vfloat32m2_t vfwnmsac(vbool16_t op0, vfloat32m2_t op1, _Float16 op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmsac_vf_f32m4))) +vfloat32m4_t vfwnmsac(vfloat32m4_t op0, _Float16 op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmsac_vf_f32m4_m))) +vfloat32m4_t vfwnmsac(vbool8_t op0, vfloat32m4_t op1, _Float16 op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmsac_vf_f32m8))) +vfloat32m8_t vfwnmsac(vfloat32m8_t op0, _Float16 op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwnmsac_vf_f32m8_m))) +vfloat32m8_t vfwnmsac(vbool4_t op0, vfloat32m8_t op1, _Float16 op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwredsum_vs_f16m1_f32m1))) +vfloat32m1_t vfwredsum(vfloat32m1_t op0, vfloat16m1_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwredsum_vs_f16m1_f32m1_m))) +vfloat32m1_t vfwredsum(vbool16_t op0, vfloat32m1_t op1, vfloat16m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwredsum_vs_f16m2_f32m1))) +vfloat32m1_t vfwredsum(vfloat32m1_t op0, vfloat16m2_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwredsum_vs_f16m2_f32m1_m))) +vfloat32m1_t vfwredsum(vbool8_t op0, vfloat32m1_t op1, vfloat16m2_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwredsum_vs_f16m4_f32m1))) +vfloat32m1_t vfwredsum(vfloat32m1_t op0, vfloat16m4_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwredsum_vs_f16m4_f32m1_m))) +vfloat32m1_t vfwredsum(vbool4_t op0, vfloat32m1_t op1, vfloat16m4_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwredsum_vs_f16m8_f32m1))) +vfloat32m1_t vfwredsum(vfloat32m1_t op0, vfloat16m8_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwredsum_vs_f16m8_f32m1_m))) +vfloat32m1_t vfwredsum(vbool2_t op0, vfloat32m1_t op1, vfloat16m8_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwredsum_vs_f16mf2_f32m1))) +vfloat32m1_t vfwredsum(vfloat32m1_t op0, vfloat16mf2_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwredsum_vs_f16mf2_f32m1_m))) +vfloat32m1_t vfwredsum(vbool32_t op0, vfloat32m1_t op1, vfloat16mf2_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwredsum_vs_f16mf4_f32m1))) +vfloat32m1_t vfwredsum(vfloat32m1_t op0, vfloat16mf4_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwredsum_vs_f16mf4_f32m1_m))) +vfloat32m1_t vfwredsum(vbool64_t op0, vfloat32m1_t op1, vfloat16mf4_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwredosum_vs_f16m1_f32m1))) +vfloat32m1_t vfwredosum(vfloat32m1_t op0, vfloat16m1_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwredosum_vs_f16m1_f32m1_m))) +vfloat32m1_t vfwredosum(vbool16_t op0, vfloat32m1_t op1, vfloat16m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwredosum_vs_f16m2_f32m1))) +vfloat32m1_t vfwredosum(vfloat32m1_t op0, vfloat16m2_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwredosum_vs_f16m2_f32m1_m))) +vfloat32m1_t vfwredosum(vbool8_t op0, vfloat32m1_t op1, vfloat16m2_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwredosum_vs_f16m4_f32m1))) +vfloat32m1_t vfwredosum(vfloat32m1_t op0, vfloat16m4_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwredosum_vs_f16m4_f32m1_m))) +vfloat32m1_t vfwredosum(vbool4_t op0, vfloat32m1_t op1, vfloat16m4_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwredosum_vs_f16m8_f32m1))) +vfloat32m1_t vfwredosum(vfloat32m1_t op0, vfloat16m8_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwredosum_vs_f16m8_f32m1_m))) +vfloat32m1_t vfwredosum(vbool2_t op0, vfloat32m1_t op1, vfloat16m8_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwredosum_vs_f16mf2_f32m1))) +vfloat32m1_t vfwredosum(vfloat32m1_t op0, vfloat16mf2_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwredosum_vs_f16mf2_f32m1_m))) +vfloat32m1_t vfwredosum(vbool32_t op0, vfloat32m1_t op1, vfloat16mf2_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwredosum_vs_f16mf4_f32m1))) +vfloat32m1_t vfwredosum(vfloat32m1_t op0, vfloat16mf4_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwredosum_vs_f16mf4_f32m1_m))) +vfloat32m1_t vfwredosum(vbool64_t op0, vfloat32m1_t op1, vfloat16mf4_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_f_w_f16mf4))) +vfloat16mf4_t vfncvt_f(vfloat32mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_f_w_f16mf4_m))) +vfloat16mf4_t vfncvt_f(vbool64_t op0, vfloat16mf4_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_f_w_f16mf2))) +vfloat16mf2_t vfncvt_f(vfloat32m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_f_w_f16mf2_m))) +vfloat16mf2_t vfncvt_f(vbool32_t op0, vfloat16mf2_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_f_w_f16m1))) +vfloat16m1_t vfncvt_f(vfloat32m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_f_w_f16m1_m))) +vfloat16m1_t vfncvt_f(vbool16_t op0, vfloat16m1_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_f_w_f16m2))) +vfloat16m2_t vfncvt_f(vfloat32m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_f_w_f16m2_m))) +vfloat16m2_t vfncvt_f(vbool8_t op0, vfloat16m2_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_f_w_f16m4))) +vfloat16m4_t vfncvt_f(vfloat32m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_f_f_w_f16m4_m))) +vfloat16m4_t vfncvt_f(vbool4_t op0, vfloat16m4_t op1, vfloat32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rod_f_f_w_f16mf4))) +vfloat16mf4_t vfncvt_rod_f(vfloat32mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rod_f_f_w_f16mf4_m))) +vfloat16mf4_t vfncvt_rod_f(vbool64_t op0, vfloat16mf4_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rod_f_f_w_f16mf2))) +vfloat16mf2_t vfncvt_rod_f(vfloat32m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rod_f_f_w_f16mf2_m))) +vfloat16mf2_t vfncvt_rod_f(vbool32_t op0, vfloat16mf2_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rod_f_f_w_f16m1))) +vfloat16m1_t vfncvt_rod_f(vfloat32m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rod_f_f_w_f16m1_m))) +vfloat16m1_t vfncvt_rod_f(vbool16_t op0, vfloat16m1_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rod_f_f_w_f16m2))) +vfloat16m2_t vfncvt_rod_f(vfloat32m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rod_f_f_w_f16m2_m))) +vfloat16m2_t vfncvt_rod_f(vbool8_t op0, vfloat16m2_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rod_f_f_w_f16m4))) +vfloat16m4_t vfncvt_rod_f(vfloat32m8_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfncvt_rod_f_f_w_f16m4_m))) +vfloat16m4_t vfncvt_rod_f(vbool4_t op0, vfloat16m4_t op1, vfloat32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_f_v_f32mf2))) +vfloat32mf2_t vfwcvt_f(vfloat16mf4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_f_v_f32mf2_m))) +vfloat32mf2_t vfwcvt_f(vbool64_t op0, vfloat32mf2_t op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_f_v_f32m1))) +vfloat32m1_t vfwcvt_f(vfloat16mf2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_f_v_f32m1_m))) +vfloat32m1_t vfwcvt_f(vbool32_t op0, vfloat32m1_t op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_f_v_f32m2))) +vfloat32m2_t vfwcvt_f(vfloat16m1_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_f_v_f32m2_m))) +vfloat32m2_t vfwcvt_f(vbool16_t op0, vfloat32m2_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_f_v_f32m4))) +vfloat32m4_t vfwcvt_f(vfloat16m2_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_f_v_f32m4_m))) +vfloat32m4_t vfwcvt_f(vbool8_t op0, vfloat32m4_t op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_f_v_f32m8))) +vfloat32m8_t vfwcvt_f(vfloat16m4_t op0, size_t op1); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vfwcvt_f_f_v_f32m8_m))) +vfloat32m8_t vfwcvt_f(vbool4_t op0, vfloat32m8_t op1, vfloat16m4_t op2, size_t op3); + +#endif + +#if defined(__riscv_zvamo) +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_i32m1))) +vint32m1_t vamoswapei16(int32_t * op0, vuint16mf2_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_i32m1_m))) +vint32m1_t vamoswapei16(vbool32_t op0, int32_t * op1, vuint16mf2_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_i32m2))) +vint32m2_t vamoswapei16(int32_t * op0, vuint16m1_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_i32m2_m))) +vint32m2_t vamoswapei16(vbool16_t op0, int32_t * op1, vuint16m1_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_i32m4))) +vint32m4_t vamoswapei16(int32_t * op0, vuint16m2_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_i32m4_m))) +vint32m4_t vamoswapei16(vbool8_t op0, int32_t * op1, vuint16m2_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_i32m8))) +vint32m8_t vamoswapei16(int32_t * op0, vuint16m4_t op1, vint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_i32m8_m))) +vint32m8_t vamoswapei16(vbool4_t op0, int32_t * op1, vuint16m4_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_i32mf2))) +vint32mf2_t vamoswapei16(int32_t * op0, vuint16mf4_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_i32mf2_m))) +vint32mf2_t vamoswapei16(vbool64_t op0, int32_t * op1, vuint16mf4_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_u32m1))) +vuint32m1_t vamoswapei16(uint32_t * op0, vuint16mf2_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_u32m1_m))) +vuint32m1_t vamoswapei16(vbool32_t op0, uint32_t * op1, vuint16mf2_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_u32m2))) +vuint32m2_t vamoswapei16(uint32_t * op0, vuint16m1_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_u32m2_m))) +vuint32m2_t vamoswapei16(vbool16_t op0, uint32_t * op1, vuint16m1_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_u32m4))) +vuint32m4_t vamoswapei16(uint32_t * op0, vuint16m2_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_u32m4_m))) +vuint32m4_t vamoswapei16(vbool8_t op0, uint32_t * op1, vuint16m2_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_u32m8))) +vuint32m8_t vamoswapei16(uint32_t * op0, vuint16m4_t op1, vuint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_u32m8_m))) +vuint32m8_t vamoswapei16(vbool4_t op0, uint32_t * op1, vuint16m4_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_u32mf2))) +vuint32mf2_t vamoswapei16(uint32_t * op0, vuint16mf4_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_u32mf2_m))) +vuint32mf2_t vamoswapei16(vbool64_t op0, uint32_t * op1, vuint16mf4_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_i32m1))) +vint32m1_t vamoswapei32(int32_t * op0, vuint32m1_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_i32m1_m))) +vint32m1_t vamoswapei32(vbool32_t op0, int32_t * op1, vuint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_i32m2))) +vint32m2_t vamoswapei32(int32_t * op0, vuint32m2_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_i32m2_m))) +vint32m2_t vamoswapei32(vbool16_t op0, int32_t * op1, vuint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_i32m4))) +vint32m4_t vamoswapei32(int32_t * op0, vuint32m4_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_i32m4_m))) +vint32m4_t vamoswapei32(vbool8_t op0, int32_t * op1, vuint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_i32m8))) +vint32m8_t vamoswapei32(int32_t * op0, vuint32m8_t op1, vint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_i32m8_m))) +vint32m8_t vamoswapei32(vbool4_t op0, int32_t * op1, vuint32m8_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_i32mf2))) +vint32mf2_t vamoswapei32(int32_t * op0, vuint32mf2_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_i32mf2_m))) +vint32mf2_t vamoswapei32(vbool64_t op0, int32_t * op1, vuint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_u32m1))) +vuint32m1_t vamoswapei32(uint32_t * op0, vuint32m1_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_u32m1_m))) +vuint32m1_t vamoswapei32(vbool32_t op0, uint32_t * op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_u32m2))) +vuint32m2_t vamoswapei32(uint32_t * op0, vuint32m2_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_u32m2_m))) +vuint32m2_t vamoswapei32(vbool16_t op0, uint32_t * op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_u32m4))) +vuint32m4_t vamoswapei32(uint32_t * op0, vuint32m4_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_u32m4_m))) +vuint32m4_t vamoswapei32(vbool8_t op0, uint32_t * op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_u32m8))) +vuint32m8_t vamoswapei32(uint32_t * op0, vuint32m8_t op1, vuint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_u32m8_m))) +vuint32m8_t vamoswapei32(vbool4_t op0, uint32_t * op1, vuint32m8_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_u32mf2))) +vuint32mf2_t vamoswapei32(uint32_t * op0, vuint32mf2_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_u32mf2_m))) +vuint32mf2_t vamoswapei32(vbool64_t op0, uint32_t * op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_i32m1))) +vint32m1_t vamoswapei64(int32_t * op0, vuint64m2_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_i32m1_m))) +vint32m1_t vamoswapei64(vbool32_t op0, int32_t * op1, vuint64m2_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_i32m2))) +vint32m2_t vamoswapei64(int32_t * op0, vuint64m4_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_i32m2_m))) +vint32m2_t vamoswapei64(vbool16_t op0, int32_t * op1, vuint64m4_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_i32m4))) +vint32m4_t vamoswapei64(int32_t * op0, vuint64m8_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_i32m4_m))) +vint32m4_t vamoswapei64(vbool8_t op0, int32_t * op1, vuint64m8_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_i32mf2))) +vint32mf2_t vamoswapei64(int32_t * op0, vuint64m1_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_i32mf2_m))) +vint32mf2_t vamoswapei64(vbool64_t op0, int32_t * op1, vuint64m1_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_u32m1))) +vuint32m1_t vamoswapei64(uint32_t * op0, vuint64m2_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_u32m1_m))) +vuint32m1_t vamoswapei64(vbool32_t op0, uint32_t * op1, vuint64m2_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_u32m2))) +vuint32m2_t vamoswapei64(uint32_t * op0, vuint64m4_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_u32m2_m))) +vuint32m2_t vamoswapei64(vbool16_t op0, uint32_t * op1, vuint64m4_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_u32m4))) +vuint32m4_t vamoswapei64(uint32_t * op0, vuint64m8_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_u32m4_m))) +vuint32m4_t vamoswapei64(vbool8_t op0, uint32_t * op1, vuint64m8_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_u32mf2))) +vuint32mf2_t vamoswapei64(uint32_t * op0, vuint64m1_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_u32mf2_m))) +vuint32mf2_t vamoswapei64(vbool64_t op0, uint32_t * op1, vuint64m1_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_i64m1))) +vint64m1_t vamoswapei8(int64_t * op0, vuint8mf8_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_i64m1_m))) +vint64m1_t vamoswapei8(vbool64_t op0, int64_t * op1, vuint8mf8_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_i64m2))) +vint64m2_t vamoswapei8(int64_t * op0, vuint8mf4_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_i64m2_m))) +vint64m2_t vamoswapei8(vbool32_t op0, int64_t * op1, vuint8mf4_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_i64m4))) +vint64m4_t vamoswapei8(int64_t * op0, vuint8mf2_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_i64m4_m))) +vint64m4_t vamoswapei8(vbool16_t op0, int64_t * op1, vuint8mf2_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_i64m8))) +vint64m8_t vamoswapei8(int64_t * op0, vuint8m1_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_i64m8_m))) +vint64m8_t vamoswapei8(vbool8_t op0, int64_t * op1, vuint8m1_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_u64m1))) +vuint64m1_t vamoswapei8(uint64_t * op0, vuint8mf8_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_u64m1_m))) +vuint64m1_t vamoswapei8(vbool64_t op0, uint64_t * op1, vuint8mf8_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_u64m2))) +vuint64m2_t vamoswapei8(uint64_t * op0, vuint8mf4_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_u64m2_m))) +vuint64m2_t vamoswapei8(vbool32_t op0, uint64_t * op1, vuint8mf4_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_u64m4))) +vuint64m4_t vamoswapei8(uint64_t * op0, vuint8mf2_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_u64m4_m))) +vuint64m4_t vamoswapei8(vbool16_t op0, uint64_t * op1, vuint8mf2_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_u64m8))) +vuint64m8_t vamoswapei8(uint64_t * op0, vuint8m1_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_u64m8_m))) +vuint64m8_t vamoswapei8(vbool8_t op0, uint64_t * op1, vuint8m1_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_i64m1))) +vint64m1_t vamoswapei16(int64_t * op0, vuint16mf4_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_i64m1_m))) +vint64m1_t vamoswapei16(vbool64_t op0, int64_t * op1, vuint16mf4_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_i64m2))) +vint64m2_t vamoswapei16(int64_t * op0, vuint16mf2_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_i64m2_m))) +vint64m2_t vamoswapei16(vbool32_t op0, int64_t * op1, vuint16mf2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_i64m4))) +vint64m4_t vamoswapei16(int64_t * op0, vuint16m1_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_i64m4_m))) +vint64m4_t vamoswapei16(vbool16_t op0, int64_t * op1, vuint16m1_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_i64m8))) +vint64m8_t vamoswapei16(int64_t * op0, vuint16m2_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_i64m8_m))) +vint64m8_t vamoswapei16(vbool8_t op0, int64_t * op1, vuint16m2_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_u64m1))) +vuint64m1_t vamoswapei16(uint64_t * op0, vuint16mf4_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_u64m1_m))) +vuint64m1_t vamoswapei16(vbool64_t op0, uint64_t * op1, vuint16mf4_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_u64m2))) +vuint64m2_t vamoswapei16(uint64_t * op0, vuint16mf2_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_u64m2_m))) +vuint64m2_t vamoswapei16(vbool32_t op0, uint64_t * op1, vuint16mf2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_u64m4))) +vuint64m4_t vamoswapei16(uint64_t * op0, vuint16m1_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_u64m4_m))) +vuint64m4_t vamoswapei16(vbool16_t op0, uint64_t * op1, vuint16m1_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_u64m8))) +vuint64m8_t vamoswapei16(uint64_t * op0, vuint16m2_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_u64m8_m))) +vuint64m8_t vamoswapei16(vbool8_t op0, uint64_t * op1, vuint16m2_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_i64m1))) +vint64m1_t vamoswapei32(int64_t * op0, vuint32mf2_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_i64m1_m))) +vint64m1_t vamoswapei32(vbool64_t op0, int64_t * op1, vuint32mf2_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_i64m2))) +vint64m2_t vamoswapei32(int64_t * op0, vuint32m1_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_i64m2_m))) +vint64m2_t vamoswapei32(vbool32_t op0, int64_t * op1, vuint32m1_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_i64m4))) +vint64m4_t vamoswapei32(int64_t * op0, vuint32m2_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_i64m4_m))) +vint64m4_t vamoswapei32(vbool16_t op0, int64_t * op1, vuint32m2_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_i64m8))) +vint64m8_t vamoswapei32(int64_t * op0, vuint32m4_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_i64m8_m))) +vint64m8_t vamoswapei32(vbool8_t op0, int64_t * op1, vuint32m4_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_u64m1))) +vuint64m1_t vamoswapei32(uint64_t * op0, vuint32mf2_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_u64m1_m))) +vuint64m1_t vamoswapei32(vbool64_t op0, uint64_t * op1, vuint32mf2_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_u64m2))) +vuint64m2_t vamoswapei32(uint64_t * op0, vuint32m1_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_u64m2_m))) +vuint64m2_t vamoswapei32(vbool32_t op0, uint64_t * op1, vuint32m1_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_u64m4))) +vuint64m4_t vamoswapei32(uint64_t * op0, vuint32m2_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_u64m4_m))) +vuint64m4_t vamoswapei32(vbool16_t op0, uint64_t * op1, vuint32m2_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_u64m8))) +vuint64m8_t vamoswapei32(uint64_t * op0, vuint32m4_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_u64m8_m))) +vuint64m8_t vamoswapei32(vbool8_t op0, uint64_t * op1, vuint32m4_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_i64m1))) +vint64m1_t vamoswapei64(int64_t * op0, vuint64m1_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_i64m1_m))) +vint64m1_t vamoswapei64(vbool64_t op0, int64_t * op1, vuint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_i64m2))) +vint64m2_t vamoswapei64(int64_t * op0, vuint64m2_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_i64m2_m))) +vint64m2_t vamoswapei64(vbool32_t op0, int64_t * op1, vuint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_i64m4))) +vint64m4_t vamoswapei64(int64_t * op0, vuint64m4_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_i64m4_m))) +vint64m4_t vamoswapei64(vbool16_t op0, int64_t * op1, vuint64m4_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_i64m8))) +vint64m8_t vamoswapei64(int64_t * op0, vuint64m8_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_i64m8_m))) +vint64m8_t vamoswapei64(vbool8_t op0, int64_t * op1, vuint64m8_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_u64m1))) +vuint64m1_t vamoswapei64(uint64_t * op0, vuint64m1_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_u64m1_m))) +vuint64m1_t vamoswapei64(vbool64_t op0, uint64_t * op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_u64m2))) +vuint64m2_t vamoswapei64(uint64_t * op0, vuint64m2_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_u64m2_m))) +vuint64m2_t vamoswapei64(vbool32_t op0, uint64_t * op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_u64m4))) +vuint64m4_t vamoswapei64(uint64_t * op0, vuint64m4_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_u64m4_m))) +vuint64m4_t vamoswapei64(vbool16_t op0, uint64_t * op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_u64m8))) +vuint64m8_t vamoswapei64(uint64_t * op0, vuint64m8_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_u64m8_m))) +vuint64m8_t vamoswapei64(vbool8_t op0, uint64_t * op1, vuint64m8_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei8_v_i32m1))) +vint32m1_t vamoaddei8(int32_t * op0, vuint8mf4_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei8_v_i32m1_m))) +vint32m1_t vamoaddei8(vbool32_t op0, int32_t * op1, vuint8mf4_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei8_v_i32m2))) +vint32m2_t vamoaddei8(int32_t * op0, vuint8mf2_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei8_v_i32m2_m))) +vint32m2_t vamoaddei8(vbool16_t op0, int32_t * op1, vuint8mf2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei8_v_i32m4))) +vint32m4_t vamoaddei8(int32_t * op0, vuint8m1_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei8_v_i32m4_m))) +vint32m4_t vamoaddei8(vbool8_t op0, int32_t * op1, vuint8m1_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei8_v_i32m8))) +vint32m8_t vamoaddei8(int32_t * op0, vuint8m2_t op1, vint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei8_v_i32m8_m))) +vint32m8_t vamoaddei8(vbool4_t op0, int32_t * op1, vuint8m2_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei8_v_i32mf2))) +vint32mf2_t vamoaddei8(int32_t * op0, vuint8mf8_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei8_v_i32mf2_m))) +vint32mf2_t vamoaddei8(vbool64_t op0, int32_t * op1, vuint8mf8_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei8_v_u32m1))) +vuint32m1_t vamoaddei8(uint32_t * op0, vuint8mf4_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei8_v_u32m1_m))) +vuint32m1_t vamoaddei8(vbool32_t op0, uint32_t * op1, vuint8mf4_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei8_v_u32m2))) +vuint32m2_t vamoaddei8(uint32_t * op0, vuint8mf2_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei8_v_u32m2_m))) +vuint32m2_t vamoaddei8(vbool16_t op0, uint32_t * op1, vuint8mf2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei8_v_u32m4))) +vuint32m4_t vamoaddei8(uint32_t * op0, vuint8m1_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei8_v_u32m4_m))) +vuint32m4_t vamoaddei8(vbool8_t op0, uint32_t * op1, vuint8m1_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei8_v_u32m8))) +vuint32m8_t vamoaddei8(uint32_t * op0, vuint8m2_t op1, vuint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei8_v_u32m8_m))) +vuint32m8_t vamoaddei8(vbool4_t op0, uint32_t * op1, vuint8m2_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei8_v_u32mf2))) +vuint32mf2_t vamoaddei8(uint32_t * op0, vuint8mf8_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei8_v_u32mf2_m))) +vuint32mf2_t vamoaddei8(vbool64_t op0, uint32_t * op1, vuint8mf8_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei16_v_i32m1))) +vint32m1_t vamoaddei16(int32_t * op0, vuint16mf2_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei16_v_i32m1_m))) +vint32m1_t vamoaddei16(vbool32_t op0, int32_t * op1, vuint16mf2_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei16_v_i32m2))) +vint32m2_t vamoaddei16(int32_t * op0, vuint16m1_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei16_v_i32m2_m))) +vint32m2_t vamoaddei16(vbool16_t op0, int32_t * op1, vuint16m1_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei16_v_i32m4))) +vint32m4_t vamoaddei16(int32_t * op0, vuint16m2_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei16_v_i32m4_m))) +vint32m4_t vamoaddei16(vbool8_t op0, int32_t * op1, vuint16m2_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei16_v_i32m8))) +vint32m8_t vamoaddei16(int32_t * op0, vuint16m4_t op1, vint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei16_v_i32m8_m))) +vint32m8_t vamoaddei16(vbool4_t op0, int32_t * op1, vuint16m4_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei16_v_i32mf2))) +vint32mf2_t vamoaddei16(int32_t * op0, vuint16mf4_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei16_v_i32mf2_m))) +vint32mf2_t vamoaddei16(vbool64_t op0, int32_t * op1, vuint16mf4_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei16_v_u32m1))) +vuint32m1_t vamoaddei16(uint32_t * op0, vuint16mf2_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei16_v_u32m1_m))) +vuint32m1_t vamoaddei16(vbool32_t op0, uint32_t * op1, vuint16mf2_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei16_v_u32m2))) +vuint32m2_t vamoaddei16(uint32_t * op0, vuint16m1_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei16_v_u32m2_m))) +vuint32m2_t vamoaddei16(vbool16_t op0, uint32_t * op1, vuint16m1_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei16_v_u32m4))) +vuint32m4_t vamoaddei16(uint32_t * op0, vuint16m2_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei16_v_u32m4_m))) +vuint32m4_t vamoaddei16(vbool8_t op0, uint32_t * op1, vuint16m2_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei16_v_u32m8))) +vuint32m8_t vamoaddei16(uint32_t * op0, vuint16m4_t op1, vuint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei16_v_u32m8_m))) +vuint32m8_t vamoaddei16(vbool4_t op0, uint32_t * op1, vuint16m4_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei16_v_u32mf2))) +vuint32mf2_t vamoaddei16(uint32_t * op0, vuint16mf4_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei16_v_u32mf2_m))) +vuint32mf2_t vamoaddei16(vbool64_t op0, uint32_t * op1, vuint16mf4_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei32_v_i32m1))) +vint32m1_t vamoaddei32(int32_t * op0, vuint32m1_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei32_v_i32m1_m))) +vint32m1_t vamoaddei32(vbool32_t op0, int32_t * op1, vuint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei32_v_i32m2))) +vint32m2_t vamoaddei32(int32_t * op0, vuint32m2_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei32_v_i32m2_m))) +vint32m2_t vamoaddei32(vbool16_t op0, int32_t * op1, vuint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei32_v_i32m4))) +vint32m4_t vamoaddei32(int32_t * op0, vuint32m4_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei32_v_i32m4_m))) +vint32m4_t vamoaddei32(vbool8_t op0, int32_t * op1, vuint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei32_v_i32m8))) +vint32m8_t vamoaddei32(int32_t * op0, vuint32m8_t op1, vint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei32_v_i32m8_m))) +vint32m8_t vamoaddei32(vbool4_t op0, int32_t * op1, vuint32m8_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei32_v_i32mf2))) +vint32mf2_t vamoaddei32(int32_t * op0, vuint32mf2_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei32_v_i32mf2_m))) +vint32mf2_t vamoaddei32(vbool64_t op0, int32_t * op1, vuint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei32_v_u32m1))) +vuint32m1_t vamoaddei32(uint32_t * op0, vuint32m1_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei32_v_u32m1_m))) +vuint32m1_t vamoaddei32(vbool32_t op0, uint32_t * op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei32_v_u32m2))) +vuint32m2_t vamoaddei32(uint32_t * op0, vuint32m2_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei32_v_u32m2_m))) +vuint32m2_t vamoaddei32(vbool16_t op0, uint32_t * op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei32_v_u32m4))) +vuint32m4_t vamoaddei32(uint32_t * op0, vuint32m4_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei32_v_u32m4_m))) +vuint32m4_t vamoaddei32(vbool8_t op0, uint32_t * op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei32_v_u32m8))) +vuint32m8_t vamoaddei32(uint32_t * op0, vuint32m8_t op1, vuint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei32_v_u32m8_m))) +vuint32m8_t vamoaddei32(vbool4_t op0, uint32_t * op1, vuint32m8_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei32_v_u32mf2))) +vuint32mf2_t vamoaddei32(uint32_t * op0, vuint32mf2_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei32_v_u32mf2_m))) +vuint32mf2_t vamoaddei32(vbool64_t op0, uint32_t * op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei64_v_i32m1))) +vint32m1_t vamoaddei64(int32_t * op0, vuint64m2_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei64_v_i32m1_m))) +vint32m1_t vamoaddei64(vbool32_t op0, int32_t * op1, vuint64m2_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei64_v_i32m2))) +vint32m2_t vamoaddei64(int32_t * op0, vuint64m4_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei64_v_i32m2_m))) +vint32m2_t vamoaddei64(vbool16_t op0, int32_t * op1, vuint64m4_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei64_v_i32m4))) +vint32m4_t vamoaddei64(int32_t * op0, vuint64m8_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei64_v_i32m4_m))) +vint32m4_t vamoaddei64(vbool8_t op0, int32_t * op1, vuint64m8_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei64_v_i32mf2))) +vint32mf2_t vamoaddei64(int32_t * op0, vuint64m1_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei64_v_i32mf2_m))) +vint32mf2_t vamoaddei64(vbool64_t op0, int32_t * op1, vuint64m1_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei64_v_u32m1))) +vuint32m1_t vamoaddei64(uint32_t * op0, vuint64m2_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei64_v_u32m1_m))) +vuint32m1_t vamoaddei64(vbool32_t op0, uint32_t * op1, vuint64m2_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei64_v_u32m2))) +vuint32m2_t vamoaddei64(uint32_t * op0, vuint64m4_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei64_v_u32m2_m))) +vuint32m2_t vamoaddei64(vbool16_t op0, uint32_t * op1, vuint64m4_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei64_v_u32m4))) +vuint32m4_t vamoaddei64(uint32_t * op0, vuint64m8_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei64_v_u32m4_m))) +vuint32m4_t vamoaddei64(vbool8_t op0, uint32_t * op1, vuint64m8_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei64_v_u32mf2))) +vuint32mf2_t vamoaddei64(uint32_t * op0, vuint64m1_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei64_v_u32mf2_m))) +vuint32mf2_t vamoaddei64(vbool64_t op0, uint32_t * op1, vuint64m1_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei8_v_i64m1))) +vint64m1_t vamoaddei8(int64_t * op0, vuint8mf8_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei8_v_i64m1_m))) +vint64m1_t vamoaddei8(vbool64_t op0, int64_t * op1, vuint8mf8_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei8_v_i64m2))) +vint64m2_t vamoaddei8(int64_t * op0, vuint8mf4_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei8_v_i64m2_m))) +vint64m2_t vamoaddei8(vbool32_t op0, int64_t * op1, vuint8mf4_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei8_v_i64m4))) +vint64m4_t vamoaddei8(int64_t * op0, vuint8mf2_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei8_v_i64m4_m))) +vint64m4_t vamoaddei8(vbool16_t op0, int64_t * op1, vuint8mf2_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei8_v_i64m8))) +vint64m8_t vamoaddei8(int64_t * op0, vuint8m1_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei8_v_i64m8_m))) +vint64m8_t vamoaddei8(vbool8_t op0, int64_t * op1, vuint8m1_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei8_v_u64m1))) +vuint64m1_t vamoaddei8(uint64_t * op0, vuint8mf8_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei8_v_u64m1_m))) +vuint64m1_t vamoaddei8(vbool64_t op0, uint64_t * op1, vuint8mf8_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei8_v_u64m2))) +vuint64m2_t vamoaddei8(uint64_t * op0, vuint8mf4_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei8_v_u64m2_m))) +vuint64m2_t vamoaddei8(vbool32_t op0, uint64_t * op1, vuint8mf4_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei8_v_u64m4))) +vuint64m4_t vamoaddei8(uint64_t * op0, vuint8mf2_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei8_v_u64m4_m))) +vuint64m4_t vamoaddei8(vbool16_t op0, uint64_t * op1, vuint8mf2_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei8_v_u64m8))) +vuint64m8_t vamoaddei8(uint64_t * op0, vuint8m1_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei8_v_u64m8_m))) +vuint64m8_t vamoaddei8(vbool8_t op0, uint64_t * op1, vuint8m1_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei16_v_i64m1))) +vint64m1_t vamoaddei16(int64_t * op0, vuint16mf4_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei16_v_i64m1_m))) +vint64m1_t vamoaddei16(vbool64_t op0, int64_t * op1, vuint16mf4_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei16_v_i64m2))) +vint64m2_t vamoaddei16(int64_t * op0, vuint16mf2_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei16_v_i64m2_m))) +vint64m2_t vamoaddei16(vbool32_t op0, int64_t * op1, vuint16mf2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei16_v_i64m4))) +vint64m4_t vamoaddei16(int64_t * op0, vuint16m1_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei16_v_i64m4_m))) +vint64m4_t vamoaddei16(vbool16_t op0, int64_t * op1, vuint16m1_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei16_v_i64m8))) +vint64m8_t vamoaddei16(int64_t * op0, vuint16m2_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei16_v_i64m8_m))) +vint64m8_t vamoaddei16(vbool8_t op0, int64_t * op1, vuint16m2_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei16_v_u64m1))) +vuint64m1_t vamoaddei16(uint64_t * op0, vuint16mf4_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei16_v_u64m1_m))) +vuint64m1_t vamoaddei16(vbool64_t op0, uint64_t * op1, vuint16mf4_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei16_v_u64m2))) +vuint64m2_t vamoaddei16(uint64_t * op0, vuint16mf2_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei16_v_u64m2_m))) +vuint64m2_t vamoaddei16(vbool32_t op0, uint64_t * op1, vuint16mf2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei16_v_u64m4))) +vuint64m4_t vamoaddei16(uint64_t * op0, vuint16m1_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei16_v_u64m4_m))) +vuint64m4_t vamoaddei16(vbool16_t op0, uint64_t * op1, vuint16m1_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei16_v_u64m8))) +vuint64m8_t vamoaddei16(uint64_t * op0, vuint16m2_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei16_v_u64m8_m))) +vuint64m8_t vamoaddei16(vbool8_t op0, uint64_t * op1, vuint16m2_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei32_v_i64m1))) +vint64m1_t vamoaddei32(int64_t * op0, vuint32mf2_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei32_v_i64m1_m))) +vint64m1_t vamoaddei32(vbool64_t op0, int64_t * op1, vuint32mf2_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei32_v_i64m2))) +vint64m2_t vamoaddei32(int64_t * op0, vuint32m1_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei32_v_i64m2_m))) +vint64m2_t vamoaddei32(vbool32_t op0, int64_t * op1, vuint32m1_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei32_v_i64m4))) +vint64m4_t vamoaddei32(int64_t * op0, vuint32m2_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei32_v_i64m4_m))) +vint64m4_t vamoaddei32(vbool16_t op0, int64_t * op1, vuint32m2_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei32_v_i64m8))) +vint64m8_t vamoaddei32(int64_t * op0, vuint32m4_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei32_v_i64m8_m))) +vint64m8_t vamoaddei32(vbool8_t op0, int64_t * op1, vuint32m4_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei32_v_u64m1))) +vuint64m1_t vamoaddei32(uint64_t * op0, vuint32mf2_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei32_v_u64m1_m))) +vuint64m1_t vamoaddei32(vbool64_t op0, uint64_t * op1, vuint32mf2_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei32_v_u64m2))) +vuint64m2_t vamoaddei32(uint64_t * op0, vuint32m1_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei32_v_u64m2_m))) +vuint64m2_t vamoaddei32(vbool32_t op0, uint64_t * op1, vuint32m1_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei32_v_u64m4))) +vuint64m4_t vamoaddei32(uint64_t * op0, vuint32m2_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei32_v_u64m4_m))) +vuint64m4_t vamoaddei32(vbool16_t op0, uint64_t * op1, vuint32m2_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei32_v_u64m8))) +vuint64m8_t vamoaddei32(uint64_t * op0, vuint32m4_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei32_v_u64m8_m))) +vuint64m8_t vamoaddei32(vbool8_t op0, uint64_t * op1, vuint32m4_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei64_v_i64m1))) +vint64m1_t vamoaddei64(int64_t * op0, vuint64m1_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei64_v_i64m1_m))) +vint64m1_t vamoaddei64(vbool64_t op0, int64_t * op1, vuint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei64_v_i64m2))) +vint64m2_t vamoaddei64(int64_t * op0, vuint64m2_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei64_v_i64m2_m))) +vint64m2_t vamoaddei64(vbool32_t op0, int64_t * op1, vuint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei64_v_i64m4))) +vint64m4_t vamoaddei64(int64_t * op0, vuint64m4_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei64_v_i64m4_m))) +vint64m4_t vamoaddei64(vbool16_t op0, int64_t * op1, vuint64m4_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei64_v_i64m8))) +vint64m8_t vamoaddei64(int64_t * op0, vuint64m8_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei64_v_i64m8_m))) +vint64m8_t vamoaddei64(vbool8_t op0, int64_t * op1, vuint64m8_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei64_v_u64m1))) +vuint64m1_t vamoaddei64(uint64_t * op0, vuint64m1_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei64_v_u64m1_m))) +vuint64m1_t vamoaddei64(vbool64_t op0, uint64_t * op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei64_v_u64m2))) +vuint64m2_t vamoaddei64(uint64_t * op0, vuint64m2_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei64_v_u64m2_m))) +vuint64m2_t vamoaddei64(vbool32_t op0, uint64_t * op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei64_v_u64m4))) +vuint64m4_t vamoaddei64(uint64_t * op0, vuint64m4_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei64_v_u64m4_m))) +vuint64m4_t vamoaddei64(vbool16_t op0, uint64_t * op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei64_v_u64m8))) +vuint64m8_t vamoaddei64(uint64_t * op0, vuint64m8_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoaddei64_v_u64m8_m))) +vuint64m8_t vamoaddei64(vbool8_t op0, uint64_t * op1, vuint64m8_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei8_v_i32m1))) +vint32m1_t vamoxorei8(int32_t * op0, vuint8mf4_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei8_v_i32m1_m))) +vint32m1_t vamoxorei8(vbool32_t op0, int32_t * op1, vuint8mf4_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei8_v_i32m2))) +vint32m2_t vamoxorei8(int32_t * op0, vuint8mf2_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei8_v_i32m2_m))) +vint32m2_t vamoxorei8(vbool16_t op0, int32_t * op1, vuint8mf2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei8_v_i32m4))) +vint32m4_t vamoxorei8(int32_t * op0, vuint8m1_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei8_v_i32m4_m))) +vint32m4_t vamoxorei8(vbool8_t op0, int32_t * op1, vuint8m1_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei8_v_i32m8))) +vint32m8_t vamoxorei8(int32_t * op0, vuint8m2_t op1, vint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei8_v_i32m8_m))) +vint32m8_t vamoxorei8(vbool4_t op0, int32_t * op1, vuint8m2_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei8_v_i32mf2))) +vint32mf2_t vamoxorei8(int32_t * op0, vuint8mf8_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei8_v_i32mf2_m))) +vint32mf2_t vamoxorei8(vbool64_t op0, int32_t * op1, vuint8mf8_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei8_v_u32m1))) +vuint32m1_t vamoxorei8(uint32_t * op0, vuint8mf4_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei8_v_u32m1_m))) +vuint32m1_t vamoxorei8(vbool32_t op0, uint32_t * op1, vuint8mf4_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei8_v_u32m2))) +vuint32m2_t vamoxorei8(uint32_t * op0, vuint8mf2_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei8_v_u32m2_m))) +vuint32m2_t vamoxorei8(vbool16_t op0, uint32_t * op1, vuint8mf2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei8_v_u32m4))) +vuint32m4_t vamoxorei8(uint32_t * op0, vuint8m1_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei8_v_u32m4_m))) +vuint32m4_t vamoxorei8(vbool8_t op0, uint32_t * op1, vuint8m1_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei8_v_u32m8))) +vuint32m8_t vamoxorei8(uint32_t * op0, vuint8m2_t op1, vuint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei8_v_u32m8_m))) +vuint32m8_t vamoxorei8(vbool4_t op0, uint32_t * op1, vuint8m2_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei8_v_u32mf2))) +vuint32mf2_t vamoxorei8(uint32_t * op0, vuint8mf8_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei8_v_u32mf2_m))) +vuint32mf2_t vamoxorei8(vbool64_t op0, uint32_t * op1, vuint8mf8_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei16_v_i32m1))) +vint32m1_t vamoxorei16(int32_t * op0, vuint16mf2_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei16_v_i32m1_m))) +vint32m1_t vamoxorei16(vbool32_t op0, int32_t * op1, vuint16mf2_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei16_v_i32m2))) +vint32m2_t vamoxorei16(int32_t * op0, vuint16m1_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei16_v_i32m2_m))) +vint32m2_t vamoxorei16(vbool16_t op0, int32_t * op1, vuint16m1_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei16_v_i32m4))) +vint32m4_t vamoxorei16(int32_t * op0, vuint16m2_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei16_v_i32m4_m))) +vint32m4_t vamoxorei16(vbool8_t op0, int32_t * op1, vuint16m2_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei16_v_i32m8))) +vint32m8_t vamoxorei16(int32_t * op0, vuint16m4_t op1, vint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei16_v_i32m8_m))) +vint32m8_t vamoxorei16(vbool4_t op0, int32_t * op1, vuint16m4_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei16_v_i32mf2))) +vint32mf2_t vamoxorei16(int32_t * op0, vuint16mf4_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei16_v_i32mf2_m))) +vint32mf2_t vamoxorei16(vbool64_t op0, int32_t * op1, vuint16mf4_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei16_v_u32m1))) +vuint32m1_t vamoxorei16(uint32_t * op0, vuint16mf2_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei16_v_u32m1_m))) +vuint32m1_t vamoxorei16(vbool32_t op0, uint32_t * op1, vuint16mf2_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei16_v_u32m2))) +vuint32m2_t vamoxorei16(uint32_t * op0, vuint16m1_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei16_v_u32m2_m))) +vuint32m2_t vamoxorei16(vbool16_t op0, uint32_t * op1, vuint16m1_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei16_v_u32m4))) +vuint32m4_t vamoxorei16(uint32_t * op0, vuint16m2_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei16_v_u32m4_m))) +vuint32m4_t vamoxorei16(vbool8_t op0, uint32_t * op1, vuint16m2_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei16_v_u32m8))) +vuint32m8_t vamoxorei16(uint32_t * op0, vuint16m4_t op1, vuint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei16_v_u32m8_m))) +vuint32m8_t vamoxorei16(vbool4_t op0, uint32_t * op1, vuint16m4_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei16_v_u32mf2))) +vuint32mf2_t vamoxorei16(uint32_t * op0, vuint16mf4_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei16_v_u32mf2_m))) +vuint32mf2_t vamoxorei16(vbool64_t op0, uint32_t * op1, vuint16mf4_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei32_v_i32m1))) +vint32m1_t vamoxorei32(int32_t * op0, vuint32m1_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei32_v_i32m1_m))) +vint32m1_t vamoxorei32(vbool32_t op0, int32_t * op1, vuint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei32_v_i32m2))) +vint32m2_t vamoxorei32(int32_t * op0, vuint32m2_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei32_v_i32m2_m))) +vint32m2_t vamoxorei32(vbool16_t op0, int32_t * op1, vuint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei32_v_i32m4))) +vint32m4_t vamoxorei32(int32_t * op0, vuint32m4_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei32_v_i32m4_m))) +vint32m4_t vamoxorei32(vbool8_t op0, int32_t * op1, vuint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei32_v_i32m8))) +vint32m8_t vamoxorei32(int32_t * op0, vuint32m8_t op1, vint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei32_v_i32m8_m))) +vint32m8_t vamoxorei32(vbool4_t op0, int32_t * op1, vuint32m8_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei32_v_i32mf2))) +vint32mf2_t vamoxorei32(int32_t * op0, vuint32mf2_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei32_v_i32mf2_m))) +vint32mf2_t vamoxorei32(vbool64_t op0, int32_t * op1, vuint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei32_v_u32m1))) +vuint32m1_t vamoxorei32(uint32_t * op0, vuint32m1_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei32_v_u32m1_m))) +vuint32m1_t vamoxorei32(vbool32_t op0, uint32_t * op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei32_v_u32m2))) +vuint32m2_t vamoxorei32(uint32_t * op0, vuint32m2_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei32_v_u32m2_m))) +vuint32m2_t vamoxorei32(vbool16_t op0, uint32_t * op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei32_v_u32m4))) +vuint32m4_t vamoxorei32(uint32_t * op0, vuint32m4_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei32_v_u32m4_m))) +vuint32m4_t vamoxorei32(vbool8_t op0, uint32_t * op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei32_v_u32m8))) +vuint32m8_t vamoxorei32(uint32_t * op0, vuint32m8_t op1, vuint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei32_v_u32m8_m))) +vuint32m8_t vamoxorei32(vbool4_t op0, uint32_t * op1, vuint32m8_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei32_v_u32mf2))) +vuint32mf2_t vamoxorei32(uint32_t * op0, vuint32mf2_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei32_v_u32mf2_m))) +vuint32mf2_t vamoxorei32(vbool64_t op0, uint32_t * op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei64_v_i32m1))) +vint32m1_t vamoxorei64(int32_t * op0, vuint64m2_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei64_v_i32m1_m))) +vint32m1_t vamoxorei64(vbool32_t op0, int32_t * op1, vuint64m2_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei64_v_i32m2))) +vint32m2_t vamoxorei64(int32_t * op0, vuint64m4_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei64_v_i32m2_m))) +vint32m2_t vamoxorei64(vbool16_t op0, int32_t * op1, vuint64m4_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei64_v_i32m4))) +vint32m4_t vamoxorei64(int32_t * op0, vuint64m8_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei64_v_i32m4_m))) +vint32m4_t vamoxorei64(vbool8_t op0, int32_t * op1, vuint64m8_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei64_v_i32mf2))) +vint32mf2_t vamoxorei64(int32_t * op0, vuint64m1_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei64_v_i32mf2_m))) +vint32mf2_t vamoxorei64(vbool64_t op0, int32_t * op1, vuint64m1_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei64_v_u32m1))) +vuint32m1_t vamoxorei64(uint32_t * op0, vuint64m2_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei64_v_u32m1_m))) +vuint32m1_t vamoxorei64(vbool32_t op0, uint32_t * op1, vuint64m2_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei64_v_u32m2))) +vuint32m2_t vamoxorei64(uint32_t * op0, vuint64m4_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei64_v_u32m2_m))) +vuint32m2_t vamoxorei64(vbool16_t op0, uint32_t * op1, vuint64m4_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei64_v_u32m4))) +vuint32m4_t vamoxorei64(uint32_t * op0, vuint64m8_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei64_v_u32m4_m))) +vuint32m4_t vamoxorei64(vbool8_t op0, uint32_t * op1, vuint64m8_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei64_v_u32mf2))) +vuint32mf2_t vamoxorei64(uint32_t * op0, vuint64m1_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei64_v_u32mf2_m))) +vuint32mf2_t vamoxorei64(vbool64_t op0, uint32_t * op1, vuint64m1_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei8_v_i64m1))) +vint64m1_t vamoxorei8(int64_t * op0, vuint8mf8_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei8_v_i64m1_m))) +vint64m1_t vamoxorei8(vbool64_t op0, int64_t * op1, vuint8mf8_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei8_v_i64m2))) +vint64m2_t vamoxorei8(int64_t * op0, vuint8mf4_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei8_v_i64m2_m))) +vint64m2_t vamoxorei8(vbool32_t op0, int64_t * op1, vuint8mf4_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei8_v_i64m4))) +vint64m4_t vamoxorei8(int64_t * op0, vuint8mf2_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei8_v_i64m4_m))) +vint64m4_t vamoxorei8(vbool16_t op0, int64_t * op1, vuint8mf2_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei8_v_i64m8))) +vint64m8_t vamoxorei8(int64_t * op0, vuint8m1_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei8_v_i64m8_m))) +vint64m8_t vamoxorei8(vbool8_t op0, int64_t * op1, vuint8m1_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei8_v_u64m1))) +vuint64m1_t vamoxorei8(uint64_t * op0, vuint8mf8_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei8_v_u64m1_m))) +vuint64m1_t vamoxorei8(vbool64_t op0, uint64_t * op1, vuint8mf8_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei8_v_u64m2))) +vuint64m2_t vamoxorei8(uint64_t * op0, vuint8mf4_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei8_v_u64m2_m))) +vuint64m2_t vamoxorei8(vbool32_t op0, uint64_t * op1, vuint8mf4_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei8_v_u64m4))) +vuint64m4_t vamoxorei8(uint64_t * op0, vuint8mf2_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei8_v_u64m4_m))) +vuint64m4_t vamoxorei8(vbool16_t op0, uint64_t * op1, vuint8mf2_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei8_v_u64m8))) +vuint64m8_t vamoxorei8(uint64_t * op0, vuint8m1_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei8_v_u64m8_m))) +vuint64m8_t vamoxorei8(vbool8_t op0, uint64_t * op1, vuint8m1_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei16_v_i64m1))) +vint64m1_t vamoxorei16(int64_t * op0, vuint16mf4_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei16_v_i64m1_m))) +vint64m1_t vamoxorei16(vbool64_t op0, int64_t * op1, vuint16mf4_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei16_v_i64m2))) +vint64m2_t vamoxorei16(int64_t * op0, vuint16mf2_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei16_v_i64m2_m))) +vint64m2_t vamoxorei16(vbool32_t op0, int64_t * op1, vuint16mf2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei16_v_i64m4))) +vint64m4_t vamoxorei16(int64_t * op0, vuint16m1_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei16_v_i64m4_m))) +vint64m4_t vamoxorei16(vbool16_t op0, int64_t * op1, vuint16m1_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei16_v_i64m8))) +vint64m8_t vamoxorei16(int64_t * op0, vuint16m2_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei16_v_i64m8_m))) +vint64m8_t vamoxorei16(vbool8_t op0, int64_t * op1, vuint16m2_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei16_v_u64m1))) +vuint64m1_t vamoxorei16(uint64_t * op0, vuint16mf4_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei16_v_u64m1_m))) +vuint64m1_t vamoxorei16(vbool64_t op0, uint64_t * op1, vuint16mf4_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei16_v_u64m2))) +vuint64m2_t vamoxorei16(uint64_t * op0, vuint16mf2_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei16_v_u64m2_m))) +vuint64m2_t vamoxorei16(vbool32_t op0, uint64_t * op1, vuint16mf2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei16_v_u64m4))) +vuint64m4_t vamoxorei16(uint64_t * op0, vuint16m1_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei16_v_u64m4_m))) +vuint64m4_t vamoxorei16(vbool16_t op0, uint64_t * op1, vuint16m1_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei16_v_u64m8))) +vuint64m8_t vamoxorei16(uint64_t * op0, vuint16m2_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei16_v_u64m8_m))) +vuint64m8_t vamoxorei16(vbool8_t op0, uint64_t * op1, vuint16m2_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei32_v_i64m1))) +vint64m1_t vamoxorei32(int64_t * op0, vuint32mf2_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei32_v_i64m1_m))) +vint64m1_t vamoxorei32(vbool64_t op0, int64_t * op1, vuint32mf2_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei32_v_i64m2))) +vint64m2_t vamoxorei32(int64_t * op0, vuint32m1_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei32_v_i64m2_m))) +vint64m2_t vamoxorei32(vbool32_t op0, int64_t * op1, vuint32m1_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei32_v_i64m4))) +vint64m4_t vamoxorei32(int64_t * op0, vuint32m2_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei32_v_i64m4_m))) +vint64m4_t vamoxorei32(vbool16_t op0, int64_t * op1, vuint32m2_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei32_v_i64m8))) +vint64m8_t vamoxorei32(int64_t * op0, vuint32m4_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei32_v_i64m8_m))) +vint64m8_t vamoxorei32(vbool8_t op0, int64_t * op1, vuint32m4_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei32_v_u64m1))) +vuint64m1_t vamoxorei32(uint64_t * op0, vuint32mf2_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei32_v_u64m1_m))) +vuint64m1_t vamoxorei32(vbool64_t op0, uint64_t * op1, vuint32mf2_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei32_v_u64m2))) +vuint64m2_t vamoxorei32(uint64_t * op0, vuint32m1_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei32_v_u64m2_m))) +vuint64m2_t vamoxorei32(vbool32_t op0, uint64_t * op1, vuint32m1_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei32_v_u64m4))) +vuint64m4_t vamoxorei32(uint64_t * op0, vuint32m2_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei32_v_u64m4_m))) +vuint64m4_t vamoxorei32(vbool16_t op0, uint64_t * op1, vuint32m2_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei32_v_u64m8))) +vuint64m8_t vamoxorei32(uint64_t * op0, vuint32m4_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei32_v_u64m8_m))) +vuint64m8_t vamoxorei32(vbool8_t op0, uint64_t * op1, vuint32m4_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei64_v_i64m1))) +vint64m1_t vamoxorei64(int64_t * op0, vuint64m1_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei64_v_i64m1_m))) +vint64m1_t vamoxorei64(vbool64_t op0, int64_t * op1, vuint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei64_v_i64m2))) +vint64m2_t vamoxorei64(int64_t * op0, vuint64m2_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei64_v_i64m2_m))) +vint64m2_t vamoxorei64(vbool32_t op0, int64_t * op1, vuint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei64_v_i64m4))) +vint64m4_t vamoxorei64(int64_t * op0, vuint64m4_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei64_v_i64m4_m))) +vint64m4_t vamoxorei64(vbool16_t op0, int64_t * op1, vuint64m4_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei64_v_i64m8))) +vint64m8_t vamoxorei64(int64_t * op0, vuint64m8_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei64_v_i64m8_m))) +vint64m8_t vamoxorei64(vbool8_t op0, int64_t * op1, vuint64m8_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei64_v_u64m1))) +vuint64m1_t vamoxorei64(uint64_t * op0, vuint64m1_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei64_v_u64m1_m))) +vuint64m1_t vamoxorei64(vbool64_t op0, uint64_t * op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei64_v_u64m2))) +vuint64m2_t vamoxorei64(uint64_t * op0, vuint64m2_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei64_v_u64m2_m))) +vuint64m2_t vamoxorei64(vbool32_t op0, uint64_t * op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei64_v_u64m4))) +vuint64m4_t vamoxorei64(uint64_t * op0, vuint64m4_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei64_v_u64m4_m))) +vuint64m4_t vamoxorei64(vbool16_t op0, uint64_t * op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei64_v_u64m8))) +vuint64m8_t vamoxorei64(uint64_t * op0, vuint64m8_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoxorei64_v_u64m8_m))) +vuint64m8_t vamoxorei64(vbool8_t op0, uint64_t * op1, vuint64m8_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei8_v_i32m1))) +vint32m1_t vamoandei8(int32_t * op0, vuint8mf4_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei8_v_i32m1_m))) +vint32m1_t vamoandei8(vbool32_t op0, int32_t * op1, vuint8mf4_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei8_v_i32m2))) +vint32m2_t vamoandei8(int32_t * op0, vuint8mf2_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei8_v_i32m2_m))) +vint32m2_t vamoandei8(vbool16_t op0, int32_t * op1, vuint8mf2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei8_v_i32m4))) +vint32m4_t vamoandei8(int32_t * op0, vuint8m1_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei8_v_i32m4_m))) +vint32m4_t vamoandei8(vbool8_t op0, int32_t * op1, vuint8m1_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei8_v_i32m8))) +vint32m8_t vamoandei8(int32_t * op0, vuint8m2_t op1, vint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei8_v_i32m8_m))) +vint32m8_t vamoandei8(vbool4_t op0, int32_t * op1, vuint8m2_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei8_v_i32mf2))) +vint32mf2_t vamoandei8(int32_t * op0, vuint8mf8_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei8_v_i32mf2_m))) +vint32mf2_t vamoandei8(vbool64_t op0, int32_t * op1, vuint8mf8_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei8_v_u32m1))) +vuint32m1_t vamoandei8(uint32_t * op0, vuint8mf4_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei8_v_u32m1_m))) +vuint32m1_t vamoandei8(vbool32_t op0, uint32_t * op1, vuint8mf4_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei8_v_u32m2))) +vuint32m2_t vamoandei8(uint32_t * op0, vuint8mf2_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei8_v_u32m2_m))) +vuint32m2_t vamoandei8(vbool16_t op0, uint32_t * op1, vuint8mf2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei8_v_u32m4))) +vuint32m4_t vamoandei8(uint32_t * op0, vuint8m1_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei8_v_u32m4_m))) +vuint32m4_t vamoandei8(vbool8_t op0, uint32_t * op1, vuint8m1_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei8_v_u32m8))) +vuint32m8_t vamoandei8(uint32_t * op0, vuint8m2_t op1, vuint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei8_v_u32m8_m))) +vuint32m8_t vamoandei8(vbool4_t op0, uint32_t * op1, vuint8m2_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei8_v_u32mf2))) +vuint32mf2_t vamoandei8(uint32_t * op0, vuint8mf8_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei8_v_u32mf2_m))) +vuint32mf2_t vamoandei8(vbool64_t op0, uint32_t * op1, vuint8mf8_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei16_v_i32m1))) +vint32m1_t vamoandei16(int32_t * op0, vuint16mf2_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei16_v_i32m1_m))) +vint32m1_t vamoandei16(vbool32_t op0, int32_t * op1, vuint16mf2_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei16_v_i32m2))) +vint32m2_t vamoandei16(int32_t * op0, vuint16m1_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei16_v_i32m2_m))) +vint32m2_t vamoandei16(vbool16_t op0, int32_t * op1, vuint16m1_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei16_v_i32m4))) +vint32m4_t vamoandei16(int32_t * op0, vuint16m2_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei16_v_i32m4_m))) +vint32m4_t vamoandei16(vbool8_t op0, int32_t * op1, vuint16m2_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei16_v_i32m8))) +vint32m8_t vamoandei16(int32_t * op0, vuint16m4_t op1, vint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei16_v_i32m8_m))) +vint32m8_t vamoandei16(vbool4_t op0, int32_t * op1, vuint16m4_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei16_v_i32mf2))) +vint32mf2_t vamoandei16(int32_t * op0, vuint16mf4_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei16_v_i32mf2_m))) +vint32mf2_t vamoandei16(vbool64_t op0, int32_t * op1, vuint16mf4_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei16_v_u32m1))) +vuint32m1_t vamoandei16(uint32_t * op0, vuint16mf2_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei16_v_u32m1_m))) +vuint32m1_t vamoandei16(vbool32_t op0, uint32_t * op1, vuint16mf2_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei16_v_u32m2))) +vuint32m2_t vamoandei16(uint32_t * op0, vuint16m1_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei16_v_u32m2_m))) +vuint32m2_t vamoandei16(vbool16_t op0, uint32_t * op1, vuint16m1_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei16_v_u32m4))) +vuint32m4_t vamoandei16(uint32_t * op0, vuint16m2_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei16_v_u32m4_m))) +vuint32m4_t vamoandei16(vbool8_t op0, uint32_t * op1, vuint16m2_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei16_v_u32m8))) +vuint32m8_t vamoandei16(uint32_t * op0, vuint16m4_t op1, vuint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei16_v_u32m8_m))) +vuint32m8_t vamoandei16(vbool4_t op0, uint32_t * op1, vuint16m4_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei16_v_u32mf2))) +vuint32mf2_t vamoandei16(uint32_t * op0, vuint16mf4_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei16_v_u32mf2_m))) +vuint32mf2_t vamoandei16(vbool64_t op0, uint32_t * op1, vuint16mf4_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei32_v_i32m1))) +vint32m1_t vamoandei32(int32_t * op0, vuint32m1_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei32_v_i32m1_m))) +vint32m1_t vamoandei32(vbool32_t op0, int32_t * op1, vuint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei32_v_i32m2))) +vint32m2_t vamoandei32(int32_t * op0, vuint32m2_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei32_v_i32m2_m))) +vint32m2_t vamoandei32(vbool16_t op0, int32_t * op1, vuint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei32_v_i32m4))) +vint32m4_t vamoandei32(int32_t * op0, vuint32m4_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei32_v_i32m4_m))) +vint32m4_t vamoandei32(vbool8_t op0, int32_t * op1, vuint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei32_v_i32m8))) +vint32m8_t vamoandei32(int32_t * op0, vuint32m8_t op1, vint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei32_v_i32m8_m))) +vint32m8_t vamoandei32(vbool4_t op0, int32_t * op1, vuint32m8_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei32_v_i32mf2))) +vint32mf2_t vamoandei32(int32_t * op0, vuint32mf2_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei32_v_i32mf2_m))) +vint32mf2_t vamoandei32(vbool64_t op0, int32_t * op1, vuint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei32_v_u32m1))) +vuint32m1_t vamoandei32(uint32_t * op0, vuint32m1_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei32_v_u32m1_m))) +vuint32m1_t vamoandei32(vbool32_t op0, uint32_t * op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei32_v_u32m2))) +vuint32m2_t vamoandei32(uint32_t * op0, vuint32m2_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei32_v_u32m2_m))) +vuint32m2_t vamoandei32(vbool16_t op0, uint32_t * op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei32_v_u32m4))) +vuint32m4_t vamoandei32(uint32_t * op0, vuint32m4_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei32_v_u32m4_m))) +vuint32m4_t vamoandei32(vbool8_t op0, uint32_t * op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei32_v_u32m8))) +vuint32m8_t vamoandei32(uint32_t * op0, vuint32m8_t op1, vuint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei32_v_u32m8_m))) +vuint32m8_t vamoandei32(vbool4_t op0, uint32_t * op1, vuint32m8_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei32_v_u32mf2))) +vuint32mf2_t vamoandei32(uint32_t * op0, vuint32mf2_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei32_v_u32mf2_m))) +vuint32mf2_t vamoandei32(vbool64_t op0, uint32_t * op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei64_v_i32m1))) +vint32m1_t vamoandei64(int32_t * op0, vuint64m2_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei64_v_i32m1_m))) +vint32m1_t vamoandei64(vbool32_t op0, int32_t * op1, vuint64m2_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei64_v_i32m2))) +vint32m2_t vamoandei64(int32_t * op0, vuint64m4_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei64_v_i32m2_m))) +vint32m2_t vamoandei64(vbool16_t op0, int32_t * op1, vuint64m4_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei64_v_i32m4))) +vint32m4_t vamoandei64(int32_t * op0, vuint64m8_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei64_v_i32m4_m))) +vint32m4_t vamoandei64(vbool8_t op0, int32_t * op1, vuint64m8_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei64_v_i32mf2))) +vint32mf2_t vamoandei64(int32_t * op0, vuint64m1_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei64_v_i32mf2_m))) +vint32mf2_t vamoandei64(vbool64_t op0, int32_t * op1, vuint64m1_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei64_v_u32m1))) +vuint32m1_t vamoandei64(uint32_t * op0, vuint64m2_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei64_v_u32m1_m))) +vuint32m1_t vamoandei64(vbool32_t op0, uint32_t * op1, vuint64m2_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei64_v_u32m2))) +vuint32m2_t vamoandei64(uint32_t * op0, vuint64m4_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei64_v_u32m2_m))) +vuint32m2_t vamoandei64(vbool16_t op0, uint32_t * op1, vuint64m4_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei64_v_u32m4))) +vuint32m4_t vamoandei64(uint32_t * op0, vuint64m8_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei64_v_u32m4_m))) +vuint32m4_t vamoandei64(vbool8_t op0, uint32_t * op1, vuint64m8_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei64_v_u32mf2))) +vuint32mf2_t vamoandei64(uint32_t * op0, vuint64m1_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei64_v_u32mf2_m))) +vuint32mf2_t vamoandei64(vbool64_t op0, uint32_t * op1, vuint64m1_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei8_v_i64m1))) +vint64m1_t vamoandei8(int64_t * op0, vuint8mf8_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei8_v_i64m1_m))) +vint64m1_t vamoandei8(vbool64_t op0, int64_t * op1, vuint8mf8_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei8_v_i64m2))) +vint64m2_t vamoandei8(int64_t * op0, vuint8mf4_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei8_v_i64m2_m))) +vint64m2_t vamoandei8(vbool32_t op0, int64_t * op1, vuint8mf4_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei8_v_i64m4))) +vint64m4_t vamoandei8(int64_t * op0, vuint8mf2_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei8_v_i64m4_m))) +vint64m4_t vamoandei8(vbool16_t op0, int64_t * op1, vuint8mf2_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei8_v_i64m8))) +vint64m8_t vamoandei8(int64_t * op0, vuint8m1_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei8_v_i64m8_m))) +vint64m8_t vamoandei8(vbool8_t op0, int64_t * op1, vuint8m1_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei8_v_u64m1))) +vuint64m1_t vamoandei8(uint64_t * op0, vuint8mf8_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei8_v_u64m1_m))) +vuint64m1_t vamoandei8(vbool64_t op0, uint64_t * op1, vuint8mf8_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei8_v_u64m2))) +vuint64m2_t vamoandei8(uint64_t * op0, vuint8mf4_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei8_v_u64m2_m))) +vuint64m2_t vamoandei8(vbool32_t op0, uint64_t * op1, vuint8mf4_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei8_v_u64m4))) +vuint64m4_t vamoandei8(uint64_t * op0, vuint8mf2_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei8_v_u64m4_m))) +vuint64m4_t vamoandei8(vbool16_t op0, uint64_t * op1, vuint8mf2_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei8_v_u64m8))) +vuint64m8_t vamoandei8(uint64_t * op0, vuint8m1_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei8_v_u64m8_m))) +vuint64m8_t vamoandei8(vbool8_t op0, uint64_t * op1, vuint8m1_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei16_v_i64m1))) +vint64m1_t vamoandei16(int64_t * op0, vuint16mf4_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei16_v_i64m1_m))) +vint64m1_t vamoandei16(vbool64_t op0, int64_t * op1, vuint16mf4_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei16_v_i64m2))) +vint64m2_t vamoandei16(int64_t * op0, vuint16mf2_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei16_v_i64m2_m))) +vint64m2_t vamoandei16(vbool32_t op0, int64_t * op1, vuint16mf2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei16_v_i64m4))) +vint64m4_t vamoandei16(int64_t * op0, vuint16m1_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei16_v_i64m4_m))) +vint64m4_t vamoandei16(vbool16_t op0, int64_t * op1, vuint16m1_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei16_v_i64m8))) +vint64m8_t vamoandei16(int64_t * op0, vuint16m2_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei16_v_i64m8_m))) +vint64m8_t vamoandei16(vbool8_t op0, int64_t * op1, vuint16m2_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei16_v_u64m1))) +vuint64m1_t vamoandei16(uint64_t * op0, vuint16mf4_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei16_v_u64m1_m))) +vuint64m1_t vamoandei16(vbool64_t op0, uint64_t * op1, vuint16mf4_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei16_v_u64m2))) +vuint64m2_t vamoandei16(uint64_t * op0, vuint16mf2_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei16_v_u64m2_m))) +vuint64m2_t vamoandei16(vbool32_t op0, uint64_t * op1, vuint16mf2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei16_v_u64m4))) +vuint64m4_t vamoandei16(uint64_t * op0, vuint16m1_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei16_v_u64m4_m))) +vuint64m4_t vamoandei16(vbool16_t op0, uint64_t * op1, vuint16m1_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei16_v_u64m8))) +vuint64m8_t vamoandei16(uint64_t * op0, vuint16m2_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei16_v_u64m8_m))) +vuint64m8_t vamoandei16(vbool8_t op0, uint64_t * op1, vuint16m2_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei32_v_i64m1))) +vint64m1_t vamoandei32(int64_t * op0, vuint32mf2_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei32_v_i64m1_m))) +vint64m1_t vamoandei32(vbool64_t op0, int64_t * op1, vuint32mf2_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei32_v_i64m2))) +vint64m2_t vamoandei32(int64_t * op0, vuint32m1_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei32_v_i64m2_m))) +vint64m2_t vamoandei32(vbool32_t op0, int64_t * op1, vuint32m1_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei32_v_i64m4))) +vint64m4_t vamoandei32(int64_t * op0, vuint32m2_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei32_v_i64m4_m))) +vint64m4_t vamoandei32(vbool16_t op0, int64_t * op1, vuint32m2_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei32_v_i64m8))) +vint64m8_t vamoandei32(int64_t * op0, vuint32m4_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei32_v_i64m8_m))) +vint64m8_t vamoandei32(vbool8_t op0, int64_t * op1, vuint32m4_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei32_v_u64m1))) +vuint64m1_t vamoandei32(uint64_t * op0, vuint32mf2_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei32_v_u64m1_m))) +vuint64m1_t vamoandei32(vbool64_t op0, uint64_t * op1, vuint32mf2_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei32_v_u64m2))) +vuint64m2_t vamoandei32(uint64_t * op0, vuint32m1_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei32_v_u64m2_m))) +vuint64m2_t vamoandei32(vbool32_t op0, uint64_t * op1, vuint32m1_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei32_v_u64m4))) +vuint64m4_t vamoandei32(uint64_t * op0, vuint32m2_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei32_v_u64m4_m))) +vuint64m4_t vamoandei32(vbool16_t op0, uint64_t * op1, vuint32m2_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei32_v_u64m8))) +vuint64m8_t vamoandei32(uint64_t * op0, vuint32m4_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei32_v_u64m8_m))) +vuint64m8_t vamoandei32(vbool8_t op0, uint64_t * op1, vuint32m4_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei64_v_i64m1))) +vint64m1_t vamoandei64(int64_t * op0, vuint64m1_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei64_v_i64m1_m))) +vint64m1_t vamoandei64(vbool64_t op0, int64_t * op1, vuint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei64_v_i64m2))) +vint64m2_t vamoandei64(int64_t * op0, vuint64m2_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei64_v_i64m2_m))) +vint64m2_t vamoandei64(vbool32_t op0, int64_t * op1, vuint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei64_v_i64m4))) +vint64m4_t vamoandei64(int64_t * op0, vuint64m4_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei64_v_i64m4_m))) +vint64m4_t vamoandei64(vbool16_t op0, int64_t * op1, vuint64m4_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei64_v_i64m8))) +vint64m8_t vamoandei64(int64_t * op0, vuint64m8_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei64_v_i64m8_m))) +vint64m8_t vamoandei64(vbool8_t op0, int64_t * op1, vuint64m8_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei64_v_u64m1))) +vuint64m1_t vamoandei64(uint64_t * op0, vuint64m1_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei64_v_u64m1_m))) +vuint64m1_t vamoandei64(vbool64_t op0, uint64_t * op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei64_v_u64m2))) +vuint64m2_t vamoandei64(uint64_t * op0, vuint64m2_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei64_v_u64m2_m))) +vuint64m2_t vamoandei64(vbool32_t op0, uint64_t * op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei64_v_u64m4))) +vuint64m4_t vamoandei64(uint64_t * op0, vuint64m4_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei64_v_u64m4_m))) +vuint64m4_t vamoandei64(vbool16_t op0, uint64_t * op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei64_v_u64m8))) +vuint64m8_t vamoandei64(uint64_t * op0, vuint64m8_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoandei64_v_u64m8_m))) +vuint64m8_t vamoandei64(vbool8_t op0, uint64_t * op1, vuint64m8_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei8_v_i32m1))) +vint32m1_t vamoorei8(int32_t * op0, vuint8mf4_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei8_v_i32m1_m))) +vint32m1_t vamoorei8(vbool32_t op0, int32_t * op1, vuint8mf4_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei8_v_i32m2))) +vint32m2_t vamoorei8(int32_t * op0, vuint8mf2_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei8_v_i32m2_m))) +vint32m2_t vamoorei8(vbool16_t op0, int32_t * op1, vuint8mf2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei8_v_i32m4))) +vint32m4_t vamoorei8(int32_t * op0, vuint8m1_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei8_v_i32m4_m))) +vint32m4_t vamoorei8(vbool8_t op0, int32_t * op1, vuint8m1_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei8_v_i32m8))) +vint32m8_t vamoorei8(int32_t * op0, vuint8m2_t op1, vint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei8_v_i32m8_m))) +vint32m8_t vamoorei8(vbool4_t op0, int32_t * op1, vuint8m2_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei8_v_i32mf2))) +vint32mf2_t vamoorei8(int32_t * op0, vuint8mf8_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei8_v_i32mf2_m))) +vint32mf2_t vamoorei8(vbool64_t op0, int32_t * op1, vuint8mf8_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei8_v_u32m1))) +vuint32m1_t vamoorei8(uint32_t * op0, vuint8mf4_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei8_v_u32m1_m))) +vuint32m1_t vamoorei8(vbool32_t op0, uint32_t * op1, vuint8mf4_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei8_v_u32m2))) +vuint32m2_t vamoorei8(uint32_t * op0, vuint8mf2_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei8_v_u32m2_m))) +vuint32m2_t vamoorei8(vbool16_t op0, uint32_t * op1, vuint8mf2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei8_v_u32m4))) +vuint32m4_t vamoorei8(uint32_t * op0, vuint8m1_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei8_v_u32m4_m))) +vuint32m4_t vamoorei8(vbool8_t op0, uint32_t * op1, vuint8m1_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei8_v_u32m8))) +vuint32m8_t vamoorei8(uint32_t * op0, vuint8m2_t op1, vuint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei8_v_u32m8_m))) +vuint32m8_t vamoorei8(vbool4_t op0, uint32_t * op1, vuint8m2_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei8_v_u32mf2))) +vuint32mf2_t vamoorei8(uint32_t * op0, vuint8mf8_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei8_v_u32mf2_m))) +vuint32mf2_t vamoorei8(vbool64_t op0, uint32_t * op1, vuint8mf8_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei16_v_i32m1))) +vint32m1_t vamoorei16(int32_t * op0, vuint16mf2_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei16_v_i32m1_m))) +vint32m1_t vamoorei16(vbool32_t op0, int32_t * op1, vuint16mf2_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei16_v_i32m2))) +vint32m2_t vamoorei16(int32_t * op0, vuint16m1_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei16_v_i32m2_m))) +vint32m2_t vamoorei16(vbool16_t op0, int32_t * op1, vuint16m1_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei16_v_i32m4))) +vint32m4_t vamoorei16(int32_t * op0, vuint16m2_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei16_v_i32m4_m))) +vint32m4_t vamoorei16(vbool8_t op0, int32_t * op1, vuint16m2_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei16_v_i32m8))) +vint32m8_t vamoorei16(int32_t * op0, vuint16m4_t op1, vint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei16_v_i32m8_m))) +vint32m8_t vamoorei16(vbool4_t op0, int32_t * op1, vuint16m4_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei16_v_i32mf2))) +vint32mf2_t vamoorei16(int32_t * op0, vuint16mf4_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei16_v_i32mf2_m))) +vint32mf2_t vamoorei16(vbool64_t op0, int32_t * op1, vuint16mf4_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei16_v_u32m1))) +vuint32m1_t vamoorei16(uint32_t * op0, vuint16mf2_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei16_v_u32m1_m))) +vuint32m1_t vamoorei16(vbool32_t op0, uint32_t * op1, vuint16mf2_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei16_v_u32m2))) +vuint32m2_t vamoorei16(uint32_t * op0, vuint16m1_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei16_v_u32m2_m))) +vuint32m2_t vamoorei16(vbool16_t op0, uint32_t * op1, vuint16m1_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei16_v_u32m4))) +vuint32m4_t vamoorei16(uint32_t * op0, vuint16m2_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei16_v_u32m4_m))) +vuint32m4_t vamoorei16(vbool8_t op0, uint32_t * op1, vuint16m2_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei16_v_u32m8))) +vuint32m8_t vamoorei16(uint32_t * op0, vuint16m4_t op1, vuint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei16_v_u32m8_m))) +vuint32m8_t vamoorei16(vbool4_t op0, uint32_t * op1, vuint16m4_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei16_v_u32mf2))) +vuint32mf2_t vamoorei16(uint32_t * op0, vuint16mf4_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei16_v_u32mf2_m))) +vuint32mf2_t vamoorei16(vbool64_t op0, uint32_t * op1, vuint16mf4_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei32_v_i32m1))) +vint32m1_t vamoorei32(int32_t * op0, vuint32m1_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei32_v_i32m1_m))) +vint32m1_t vamoorei32(vbool32_t op0, int32_t * op1, vuint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei32_v_i32m2))) +vint32m2_t vamoorei32(int32_t * op0, vuint32m2_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei32_v_i32m2_m))) +vint32m2_t vamoorei32(vbool16_t op0, int32_t * op1, vuint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei32_v_i32m4))) +vint32m4_t vamoorei32(int32_t * op0, vuint32m4_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei32_v_i32m4_m))) +vint32m4_t vamoorei32(vbool8_t op0, int32_t * op1, vuint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei32_v_i32m8))) +vint32m8_t vamoorei32(int32_t * op0, vuint32m8_t op1, vint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei32_v_i32m8_m))) +vint32m8_t vamoorei32(vbool4_t op0, int32_t * op1, vuint32m8_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei32_v_i32mf2))) +vint32mf2_t vamoorei32(int32_t * op0, vuint32mf2_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei32_v_i32mf2_m))) +vint32mf2_t vamoorei32(vbool64_t op0, int32_t * op1, vuint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei32_v_u32m1))) +vuint32m1_t vamoorei32(uint32_t * op0, vuint32m1_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei32_v_u32m1_m))) +vuint32m1_t vamoorei32(vbool32_t op0, uint32_t * op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei32_v_u32m2))) +vuint32m2_t vamoorei32(uint32_t * op0, vuint32m2_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei32_v_u32m2_m))) +vuint32m2_t vamoorei32(vbool16_t op0, uint32_t * op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei32_v_u32m4))) +vuint32m4_t vamoorei32(uint32_t * op0, vuint32m4_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei32_v_u32m4_m))) +vuint32m4_t vamoorei32(vbool8_t op0, uint32_t * op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei32_v_u32m8))) +vuint32m8_t vamoorei32(uint32_t * op0, vuint32m8_t op1, vuint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei32_v_u32m8_m))) +vuint32m8_t vamoorei32(vbool4_t op0, uint32_t * op1, vuint32m8_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei32_v_u32mf2))) +vuint32mf2_t vamoorei32(uint32_t * op0, vuint32mf2_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei32_v_u32mf2_m))) +vuint32mf2_t vamoorei32(vbool64_t op0, uint32_t * op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei64_v_i32m1))) +vint32m1_t vamoorei64(int32_t * op0, vuint64m2_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei64_v_i32m1_m))) +vint32m1_t vamoorei64(vbool32_t op0, int32_t * op1, vuint64m2_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei64_v_i32m2))) +vint32m2_t vamoorei64(int32_t * op0, vuint64m4_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei64_v_i32m2_m))) +vint32m2_t vamoorei64(vbool16_t op0, int32_t * op1, vuint64m4_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei64_v_i32m4))) +vint32m4_t vamoorei64(int32_t * op0, vuint64m8_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei64_v_i32m4_m))) +vint32m4_t vamoorei64(vbool8_t op0, int32_t * op1, vuint64m8_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei64_v_i32mf2))) +vint32mf2_t vamoorei64(int32_t * op0, vuint64m1_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei64_v_i32mf2_m))) +vint32mf2_t vamoorei64(vbool64_t op0, int32_t * op1, vuint64m1_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei64_v_u32m1))) +vuint32m1_t vamoorei64(uint32_t * op0, vuint64m2_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei64_v_u32m1_m))) +vuint32m1_t vamoorei64(vbool32_t op0, uint32_t * op1, vuint64m2_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei64_v_u32m2))) +vuint32m2_t vamoorei64(uint32_t * op0, vuint64m4_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei64_v_u32m2_m))) +vuint32m2_t vamoorei64(vbool16_t op0, uint32_t * op1, vuint64m4_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei64_v_u32m4))) +vuint32m4_t vamoorei64(uint32_t * op0, vuint64m8_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei64_v_u32m4_m))) +vuint32m4_t vamoorei64(vbool8_t op0, uint32_t * op1, vuint64m8_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei64_v_u32mf2))) +vuint32mf2_t vamoorei64(uint32_t * op0, vuint64m1_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei64_v_u32mf2_m))) +vuint32mf2_t vamoorei64(vbool64_t op0, uint32_t * op1, vuint64m1_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei8_v_i64m1))) +vint64m1_t vamoorei8(int64_t * op0, vuint8mf8_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei8_v_i64m1_m))) +vint64m1_t vamoorei8(vbool64_t op0, int64_t * op1, vuint8mf8_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei8_v_i64m2))) +vint64m2_t vamoorei8(int64_t * op0, vuint8mf4_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei8_v_i64m2_m))) +vint64m2_t vamoorei8(vbool32_t op0, int64_t * op1, vuint8mf4_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei8_v_i64m4))) +vint64m4_t vamoorei8(int64_t * op0, vuint8mf2_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei8_v_i64m4_m))) +vint64m4_t vamoorei8(vbool16_t op0, int64_t * op1, vuint8mf2_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei8_v_i64m8))) +vint64m8_t vamoorei8(int64_t * op0, vuint8m1_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei8_v_i64m8_m))) +vint64m8_t vamoorei8(vbool8_t op0, int64_t * op1, vuint8m1_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei8_v_u64m1))) +vuint64m1_t vamoorei8(uint64_t * op0, vuint8mf8_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei8_v_u64m1_m))) +vuint64m1_t vamoorei8(vbool64_t op0, uint64_t * op1, vuint8mf8_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei8_v_u64m2))) +vuint64m2_t vamoorei8(uint64_t * op0, vuint8mf4_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei8_v_u64m2_m))) +vuint64m2_t vamoorei8(vbool32_t op0, uint64_t * op1, vuint8mf4_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei8_v_u64m4))) +vuint64m4_t vamoorei8(uint64_t * op0, vuint8mf2_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei8_v_u64m4_m))) +vuint64m4_t vamoorei8(vbool16_t op0, uint64_t * op1, vuint8mf2_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei8_v_u64m8))) +vuint64m8_t vamoorei8(uint64_t * op0, vuint8m1_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei8_v_u64m8_m))) +vuint64m8_t vamoorei8(vbool8_t op0, uint64_t * op1, vuint8m1_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei16_v_i64m1))) +vint64m1_t vamoorei16(int64_t * op0, vuint16mf4_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei16_v_i64m1_m))) +vint64m1_t vamoorei16(vbool64_t op0, int64_t * op1, vuint16mf4_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei16_v_i64m2))) +vint64m2_t vamoorei16(int64_t * op0, vuint16mf2_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei16_v_i64m2_m))) +vint64m2_t vamoorei16(vbool32_t op0, int64_t * op1, vuint16mf2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei16_v_i64m4))) +vint64m4_t vamoorei16(int64_t * op0, vuint16m1_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei16_v_i64m4_m))) +vint64m4_t vamoorei16(vbool16_t op0, int64_t * op1, vuint16m1_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei16_v_i64m8))) +vint64m8_t vamoorei16(int64_t * op0, vuint16m2_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei16_v_i64m8_m))) +vint64m8_t vamoorei16(vbool8_t op0, int64_t * op1, vuint16m2_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei16_v_u64m1))) +vuint64m1_t vamoorei16(uint64_t * op0, vuint16mf4_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei16_v_u64m1_m))) +vuint64m1_t vamoorei16(vbool64_t op0, uint64_t * op1, vuint16mf4_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei16_v_u64m2))) +vuint64m2_t vamoorei16(uint64_t * op0, vuint16mf2_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei16_v_u64m2_m))) +vuint64m2_t vamoorei16(vbool32_t op0, uint64_t * op1, vuint16mf2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei16_v_u64m4))) +vuint64m4_t vamoorei16(uint64_t * op0, vuint16m1_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei16_v_u64m4_m))) +vuint64m4_t vamoorei16(vbool16_t op0, uint64_t * op1, vuint16m1_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei16_v_u64m8))) +vuint64m8_t vamoorei16(uint64_t * op0, vuint16m2_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei16_v_u64m8_m))) +vuint64m8_t vamoorei16(vbool8_t op0, uint64_t * op1, vuint16m2_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei32_v_i64m1))) +vint64m1_t vamoorei32(int64_t * op0, vuint32mf2_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei32_v_i64m1_m))) +vint64m1_t vamoorei32(vbool64_t op0, int64_t * op1, vuint32mf2_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei32_v_i64m2))) +vint64m2_t vamoorei32(int64_t * op0, vuint32m1_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei32_v_i64m2_m))) +vint64m2_t vamoorei32(vbool32_t op0, int64_t * op1, vuint32m1_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei32_v_i64m4))) +vint64m4_t vamoorei32(int64_t * op0, vuint32m2_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei32_v_i64m4_m))) +vint64m4_t vamoorei32(vbool16_t op0, int64_t * op1, vuint32m2_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei32_v_i64m8))) +vint64m8_t vamoorei32(int64_t * op0, vuint32m4_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei32_v_i64m8_m))) +vint64m8_t vamoorei32(vbool8_t op0, int64_t * op1, vuint32m4_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei32_v_u64m1))) +vuint64m1_t vamoorei32(uint64_t * op0, vuint32mf2_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei32_v_u64m1_m))) +vuint64m1_t vamoorei32(vbool64_t op0, uint64_t * op1, vuint32mf2_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei32_v_u64m2))) +vuint64m2_t vamoorei32(uint64_t * op0, vuint32m1_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei32_v_u64m2_m))) +vuint64m2_t vamoorei32(vbool32_t op0, uint64_t * op1, vuint32m1_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei32_v_u64m4))) +vuint64m4_t vamoorei32(uint64_t * op0, vuint32m2_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei32_v_u64m4_m))) +vuint64m4_t vamoorei32(vbool16_t op0, uint64_t * op1, vuint32m2_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei32_v_u64m8))) +vuint64m8_t vamoorei32(uint64_t * op0, vuint32m4_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei32_v_u64m8_m))) +vuint64m8_t vamoorei32(vbool8_t op0, uint64_t * op1, vuint32m4_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei64_v_i64m1))) +vint64m1_t vamoorei64(int64_t * op0, vuint64m1_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei64_v_i64m1_m))) +vint64m1_t vamoorei64(vbool64_t op0, int64_t * op1, vuint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei64_v_i64m2))) +vint64m2_t vamoorei64(int64_t * op0, vuint64m2_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei64_v_i64m2_m))) +vint64m2_t vamoorei64(vbool32_t op0, int64_t * op1, vuint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei64_v_i64m4))) +vint64m4_t vamoorei64(int64_t * op0, vuint64m4_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei64_v_i64m4_m))) +vint64m4_t vamoorei64(vbool16_t op0, int64_t * op1, vuint64m4_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei64_v_i64m8))) +vint64m8_t vamoorei64(int64_t * op0, vuint64m8_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei64_v_i64m8_m))) +vint64m8_t vamoorei64(vbool8_t op0, int64_t * op1, vuint64m8_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei64_v_u64m1))) +vuint64m1_t vamoorei64(uint64_t * op0, vuint64m1_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei64_v_u64m1_m))) +vuint64m1_t vamoorei64(vbool64_t op0, uint64_t * op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei64_v_u64m2))) +vuint64m2_t vamoorei64(uint64_t * op0, vuint64m2_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei64_v_u64m2_m))) +vuint64m2_t vamoorei64(vbool32_t op0, uint64_t * op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei64_v_u64m4))) +vuint64m4_t vamoorei64(uint64_t * op0, vuint64m4_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei64_v_u64m4_m))) +vuint64m4_t vamoorei64(vbool16_t op0, uint64_t * op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei64_v_u64m8))) +vuint64m8_t vamoorei64(uint64_t * op0, vuint64m8_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoorei64_v_u64m8_m))) +vuint64m8_t vamoorei64(vbool8_t op0, uint64_t * op1, vuint64m8_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei8_v_i32m1))) +vint32m1_t vamominei8(int32_t * op0, vuint8mf4_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei8_v_i32m1_m))) +vint32m1_t vamominei8(vbool32_t op0, int32_t * op1, vuint8mf4_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei8_v_i32m2))) +vint32m2_t vamominei8(int32_t * op0, vuint8mf2_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei8_v_i32m2_m))) +vint32m2_t vamominei8(vbool16_t op0, int32_t * op1, vuint8mf2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei8_v_i32m4))) +vint32m4_t vamominei8(int32_t * op0, vuint8m1_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei8_v_i32m4_m))) +vint32m4_t vamominei8(vbool8_t op0, int32_t * op1, vuint8m1_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei8_v_i32m8))) +vint32m8_t vamominei8(int32_t * op0, vuint8m2_t op1, vint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei8_v_i32m8_m))) +vint32m8_t vamominei8(vbool4_t op0, int32_t * op1, vuint8m2_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei8_v_i32mf2))) +vint32mf2_t vamominei8(int32_t * op0, vuint8mf8_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei8_v_i32mf2_m))) +vint32mf2_t vamominei8(vbool64_t op0, int32_t * op1, vuint8mf8_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei16_v_i32m1))) +vint32m1_t vamominei16(int32_t * op0, vuint16mf2_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei16_v_i32m1_m))) +vint32m1_t vamominei16(vbool32_t op0, int32_t * op1, vuint16mf2_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei16_v_i32m2))) +vint32m2_t vamominei16(int32_t * op0, vuint16m1_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei16_v_i32m2_m))) +vint32m2_t vamominei16(vbool16_t op0, int32_t * op1, vuint16m1_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei16_v_i32m4))) +vint32m4_t vamominei16(int32_t * op0, vuint16m2_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei16_v_i32m4_m))) +vint32m4_t vamominei16(vbool8_t op0, int32_t * op1, vuint16m2_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei16_v_i32m8))) +vint32m8_t vamominei16(int32_t * op0, vuint16m4_t op1, vint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei16_v_i32m8_m))) +vint32m8_t vamominei16(vbool4_t op0, int32_t * op1, vuint16m4_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei16_v_i32mf2))) +vint32mf2_t vamominei16(int32_t * op0, vuint16mf4_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei16_v_i32mf2_m))) +vint32mf2_t vamominei16(vbool64_t op0, int32_t * op1, vuint16mf4_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei32_v_i32m1))) +vint32m1_t vamominei32(int32_t * op0, vuint32m1_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei32_v_i32m1_m))) +vint32m1_t vamominei32(vbool32_t op0, int32_t * op1, vuint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei32_v_i32m2))) +vint32m2_t vamominei32(int32_t * op0, vuint32m2_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei32_v_i32m2_m))) +vint32m2_t vamominei32(vbool16_t op0, int32_t * op1, vuint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei32_v_i32m4))) +vint32m4_t vamominei32(int32_t * op0, vuint32m4_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei32_v_i32m4_m))) +vint32m4_t vamominei32(vbool8_t op0, int32_t * op1, vuint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei32_v_i32m8))) +vint32m8_t vamominei32(int32_t * op0, vuint32m8_t op1, vint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei32_v_i32m8_m))) +vint32m8_t vamominei32(vbool4_t op0, int32_t * op1, vuint32m8_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei32_v_i32mf2))) +vint32mf2_t vamominei32(int32_t * op0, vuint32mf2_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei32_v_i32mf2_m))) +vint32mf2_t vamominei32(vbool64_t op0, int32_t * op1, vuint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei64_v_i32m1))) +vint32m1_t vamominei64(int32_t * op0, vuint64m2_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei64_v_i32m1_m))) +vint32m1_t vamominei64(vbool32_t op0, int32_t * op1, vuint64m2_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei64_v_i32m2))) +vint32m2_t vamominei64(int32_t * op0, vuint64m4_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei64_v_i32m2_m))) +vint32m2_t vamominei64(vbool16_t op0, int32_t * op1, vuint64m4_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei64_v_i32m4))) +vint32m4_t vamominei64(int32_t * op0, vuint64m8_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei64_v_i32m4_m))) +vint32m4_t vamominei64(vbool8_t op0, int32_t * op1, vuint64m8_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei64_v_i32mf2))) +vint32mf2_t vamominei64(int32_t * op0, vuint64m1_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei64_v_i32mf2_m))) +vint32mf2_t vamominei64(vbool64_t op0, int32_t * op1, vuint64m1_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei8_v_i64m1))) +vint64m1_t vamominei8(int64_t * op0, vuint8mf8_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei8_v_i64m1_m))) +vint64m1_t vamominei8(vbool64_t op0, int64_t * op1, vuint8mf8_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei8_v_i64m2))) +vint64m2_t vamominei8(int64_t * op0, vuint8mf4_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei8_v_i64m2_m))) +vint64m2_t vamominei8(vbool32_t op0, int64_t * op1, vuint8mf4_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei8_v_i64m4))) +vint64m4_t vamominei8(int64_t * op0, vuint8mf2_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei8_v_i64m4_m))) +vint64m4_t vamominei8(vbool16_t op0, int64_t * op1, vuint8mf2_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei8_v_i64m8))) +vint64m8_t vamominei8(int64_t * op0, vuint8m1_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei8_v_i64m8_m))) +vint64m8_t vamominei8(vbool8_t op0, int64_t * op1, vuint8m1_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei16_v_i64m1))) +vint64m1_t vamominei16(int64_t * op0, vuint16mf4_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei16_v_i64m1_m))) +vint64m1_t vamominei16(vbool64_t op0, int64_t * op1, vuint16mf4_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei16_v_i64m2))) +vint64m2_t vamominei16(int64_t * op0, vuint16mf2_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei16_v_i64m2_m))) +vint64m2_t vamominei16(vbool32_t op0, int64_t * op1, vuint16mf2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei16_v_i64m4))) +vint64m4_t vamominei16(int64_t * op0, vuint16m1_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei16_v_i64m4_m))) +vint64m4_t vamominei16(vbool16_t op0, int64_t * op1, vuint16m1_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei16_v_i64m8))) +vint64m8_t vamominei16(int64_t * op0, vuint16m2_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei16_v_i64m8_m))) +vint64m8_t vamominei16(vbool8_t op0, int64_t * op1, vuint16m2_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei32_v_i64m1))) +vint64m1_t vamominei32(int64_t * op0, vuint32mf2_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei32_v_i64m1_m))) +vint64m1_t vamominei32(vbool64_t op0, int64_t * op1, vuint32mf2_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei32_v_i64m2))) +vint64m2_t vamominei32(int64_t * op0, vuint32m1_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei32_v_i64m2_m))) +vint64m2_t vamominei32(vbool32_t op0, int64_t * op1, vuint32m1_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei32_v_i64m4))) +vint64m4_t vamominei32(int64_t * op0, vuint32m2_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei32_v_i64m4_m))) +vint64m4_t vamominei32(vbool16_t op0, int64_t * op1, vuint32m2_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei32_v_i64m8))) +vint64m8_t vamominei32(int64_t * op0, vuint32m4_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei32_v_i64m8_m))) +vint64m8_t vamominei32(vbool8_t op0, int64_t * op1, vuint32m4_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei64_v_i64m1))) +vint64m1_t vamominei64(int64_t * op0, vuint64m1_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei64_v_i64m1_m))) +vint64m1_t vamominei64(vbool64_t op0, int64_t * op1, vuint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei64_v_i64m2))) +vint64m2_t vamominei64(int64_t * op0, vuint64m2_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei64_v_i64m2_m))) +vint64m2_t vamominei64(vbool32_t op0, int64_t * op1, vuint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei64_v_i64m4))) +vint64m4_t vamominei64(int64_t * op0, vuint64m4_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei64_v_i64m4_m))) +vint64m4_t vamominei64(vbool16_t op0, int64_t * op1, vuint64m4_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei64_v_i64m8))) +vint64m8_t vamominei64(int64_t * op0, vuint64m8_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominei64_v_i64m8_m))) +vint64m8_t vamominei64(vbool8_t op0, int64_t * op1, vuint64m8_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei8_v_i32m1))) +vint32m1_t vamomaxei8(int32_t * op0, vuint8mf4_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei8_v_i32m1_m))) +vint32m1_t vamomaxei8(vbool32_t op0, int32_t * op1, vuint8mf4_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei8_v_i32m2))) +vint32m2_t vamomaxei8(int32_t * op0, vuint8mf2_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei8_v_i32m2_m))) +vint32m2_t vamomaxei8(vbool16_t op0, int32_t * op1, vuint8mf2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei8_v_i32m4))) +vint32m4_t vamomaxei8(int32_t * op0, vuint8m1_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei8_v_i32m4_m))) +vint32m4_t vamomaxei8(vbool8_t op0, int32_t * op1, vuint8m1_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei8_v_i32m8))) +vint32m8_t vamomaxei8(int32_t * op0, vuint8m2_t op1, vint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei8_v_i32m8_m))) +vint32m8_t vamomaxei8(vbool4_t op0, int32_t * op1, vuint8m2_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei8_v_i32mf2))) +vint32mf2_t vamomaxei8(int32_t * op0, vuint8mf8_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei8_v_i32mf2_m))) +vint32mf2_t vamomaxei8(vbool64_t op0, int32_t * op1, vuint8mf8_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei16_v_i32m1))) +vint32m1_t vamomaxei16(int32_t * op0, vuint16mf2_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei16_v_i32m1_m))) +vint32m1_t vamomaxei16(vbool32_t op0, int32_t * op1, vuint16mf2_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei16_v_i32m2))) +vint32m2_t vamomaxei16(int32_t * op0, vuint16m1_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei16_v_i32m2_m))) +vint32m2_t vamomaxei16(vbool16_t op0, int32_t * op1, vuint16m1_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei16_v_i32m4))) +vint32m4_t vamomaxei16(int32_t * op0, vuint16m2_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei16_v_i32m4_m))) +vint32m4_t vamomaxei16(vbool8_t op0, int32_t * op1, vuint16m2_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei16_v_i32m8))) +vint32m8_t vamomaxei16(int32_t * op0, vuint16m4_t op1, vint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei16_v_i32m8_m))) +vint32m8_t vamomaxei16(vbool4_t op0, int32_t * op1, vuint16m4_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei16_v_i32mf2))) +vint32mf2_t vamomaxei16(int32_t * op0, vuint16mf4_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei16_v_i32mf2_m))) +vint32mf2_t vamomaxei16(vbool64_t op0, int32_t * op1, vuint16mf4_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei32_v_i32m1))) +vint32m1_t vamomaxei32(int32_t * op0, vuint32m1_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei32_v_i32m1_m))) +vint32m1_t vamomaxei32(vbool32_t op0, int32_t * op1, vuint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei32_v_i32m2))) +vint32m2_t vamomaxei32(int32_t * op0, vuint32m2_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei32_v_i32m2_m))) +vint32m2_t vamomaxei32(vbool16_t op0, int32_t * op1, vuint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei32_v_i32m4))) +vint32m4_t vamomaxei32(int32_t * op0, vuint32m4_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei32_v_i32m4_m))) +vint32m4_t vamomaxei32(vbool8_t op0, int32_t * op1, vuint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei32_v_i32m8))) +vint32m8_t vamomaxei32(int32_t * op0, vuint32m8_t op1, vint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei32_v_i32m8_m))) +vint32m8_t vamomaxei32(vbool4_t op0, int32_t * op1, vuint32m8_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei32_v_i32mf2))) +vint32mf2_t vamomaxei32(int32_t * op0, vuint32mf2_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei32_v_i32mf2_m))) +vint32mf2_t vamomaxei32(vbool64_t op0, int32_t * op1, vuint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei64_v_i32m1))) +vint32m1_t vamomaxei64(int32_t * op0, vuint64m2_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei64_v_i32m1_m))) +vint32m1_t vamomaxei64(vbool32_t op0, int32_t * op1, vuint64m2_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei64_v_i32m2))) +vint32m2_t vamomaxei64(int32_t * op0, vuint64m4_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei64_v_i32m2_m))) +vint32m2_t vamomaxei64(vbool16_t op0, int32_t * op1, vuint64m4_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei64_v_i32m4))) +vint32m4_t vamomaxei64(int32_t * op0, vuint64m8_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei64_v_i32m4_m))) +vint32m4_t vamomaxei64(vbool8_t op0, int32_t * op1, vuint64m8_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei64_v_i32mf2))) +vint32mf2_t vamomaxei64(int32_t * op0, vuint64m1_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei64_v_i32mf2_m))) +vint32mf2_t vamomaxei64(vbool64_t op0, int32_t * op1, vuint64m1_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei8_v_i64m1))) +vint64m1_t vamomaxei8(int64_t * op0, vuint8mf8_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei8_v_i64m1_m))) +vint64m1_t vamomaxei8(vbool64_t op0, int64_t * op1, vuint8mf8_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei8_v_i64m2))) +vint64m2_t vamomaxei8(int64_t * op0, vuint8mf4_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei8_v_i64m2_m))) +vint64m2_t vamomaxei8(vbool32_t op0, int64_t * op1, vuint8mf4_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei8_v_i64m4))) +vint64m4_t vamomaxei8(int64_t * op0, vuint8mf2_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei8_v_i64m4_m))) +vint64m4_t vamomaxei8(vbool16_t op0, int64_t * op1, vuint8mf2_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei8_v_i64m8))) +vint64m8_t vamomaxei8(int64_t * op0, vuint8m1_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei8_v_i64m8_m))) +vint64m8_t vamomaxei8(vbool8_t op0, int64_t * op1, vuint8m1_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei16_v_i64m1))) +vint64m1_t vamomaxei16(int64_t * op0, vuint16mf4_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei16_v_i64m1_m))) +vint64m1_t vamomaxei16(vbool64_t op0, int64_t * op1, vuint16mf4_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei16_v_i64m2))) +vint64m2_t vamomaxei16(int64_t * op0, vuint16mf2_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei16_v_i64m2_m))) +vint64m2_t vamomaxei16(vbool32_t op0, int64_t * op1, vuint16mf2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei16_v_i64m4))) +vint64m4_t vamomaxei16(int64_t * op0, vuint16m1_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei16_v_i64m4_m))) +vint64m4_t vamomaxei16(vbool16_t op0, int64_t * op1, vuint16m1_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei16_v_i64m8))) +vint64m8_t vamomaxei16(int64_t * op0, vuint16m2_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei16_v_i64m8_m))) +vint64m8_t vamomaxei16(vbool8_t op0, int64_t * op1, vuint16m2_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei32_v_i64m1))) +vint64m1_t vamomaxei32(int64_t * op0, vuint32mf2_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei32_v_i64m1_m))) +vint64m1_t vamomaxei32(vbool64_t op0, int64_t * op1, vuint32mf2_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei32_v_i64m2))) +vint64m2_t vamomaxei32(int64_t * op0, vuint32m1_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei32_v_i64m2_m))) +vint64m2_t vamomaxei32(vbool32_t op0, int64_t * op1, vuint32m1_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei32_v_i64m4))) +vint64m4_t vamomaxei32(int64_t * op0, vuint32m2_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei32_v_i64m4_m))) +vint64m4_t vamomaxei32(vbool16_t op0, int64_t * op1, vuint32m2_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei32_v_i64m8))) +vint64m8_t vamomaxei32(int64_t * op0, vuint32m4_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei32_v_i64m8_m))) +vint64m8_t vamomaxei32(vbool8_t op0, int64_t * op1, vuint32m4_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei64_v_i64m1))) +vint64m1_t vamomaxei64(int64_t * op0, vuint64m1_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei64_v_i64m1_m))) +vint64m1_t vamomaxei64(vbool64_t op0, int64_t * op1, vuint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei64_v_i64m2))) +vint64m2_t vamomaxei64(int64_t * op0, vuint64m2_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei64_v_i64m2_m))) +vint64m2_t vamomaxei64(vbool32_t op0, int64_t * op1, vuint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei64_v_i64m4))) +vint64m4_t vamomaxei64(int64_t * op0, vuint64m4_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei64_v_i64m4_m))) +vint64m4_t vamomaxei64(vbool16_t op0, int64_t * op1, vuint64m4_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei64_v_i64m8))) +vint64m8_t vamomaxei64(int64_t * op0, vuint64m8_t op1, vint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxei64_v_i64m8_m))) +vint64m8_t vamomaxei64(vbool8_t op0, int64_t * op1, vuint64m8_t op2, vint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei8_v_u32m1))) +vuint32m1_t vamominuei8(uint32_t * op0, vuint8mf4_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei8_v_u32m1_m))) +vuint32m1_t vamominuei8(vbool32_t op0, uint32_t * op1, vuint8mf4_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei8_v_u32m2))) +vuint32m2_t vamominuei8(uint32_t * op0, vuint8mf2_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei8_v_u32m2_m))) +vuint32m2_t vamominuei8(vbool16_t op0, uint32_t * op1, vuint8mf2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei8_v_u32m4))) +vuint32m4_t vamominuei8(uint32_t * op0, vuint8m1_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei8_v_u32m4_m))) +vuint32m4_t vamominuei8(vbool8_t op0, uint32_t * op1, vuint8m1_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei8_v_u32m8))) +vuint32m8_t vamominuei8(uint32_t * op0, vuint8m2_t op1, vuint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei8_v_u32m8_m))) +vuint32m8_t vamominuei8(vbool4_t op0, uint32_t * op1, vuint8m2_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei8_v_u32mf2))) +vuint32mf2_t vamominuei8(uint32_t * op0, vuint8mf8_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei8_v_u32mf2_m))) +vuint32mf2_t vamominuei8(vbool64_t op0, uint32_t * op1, vuint8mf8_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei16_v_u32m1))) +vuint32m1_t vamominuei16(uint32_t * op0, vuint16mf2_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei16_v_u32m1_m))) +vuint32m1_t vamominuei16(vbool32_t op0, uint32_t * op1, vuint16mf2_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei16_v_u32m2))) +vuint32m2_t vamominuei16(uint32_t * op0, vuint16m1_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei16_v_u32m2_m))) +vuint32m2_t vamominuei16(vbool16_t op0, uint32_t * op1, vuint16m1_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei16_v_u32m4))) +vuint32m4_t vamominuei16(uint32_t * op0, vuint16m2_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei16_v_u32m4_m))) +vuint32m4_t vamominuei16(vbool8_t op0, uint32_t * op1, vuint16m2_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei16_v_u32m8))) +vuint32m8_t vamominuei16(uint32_t * op0, vuint16m4_t op1, vuint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei16_v_u32m8_m))) +vuint32m8_t vamominuei16(vbool4_t op0, uint32_t * op1, vuint16m4_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei16_v_u32mf2))) +vuint32mf2_t vamominuei16(uint32_t * op0, vuint16mf4_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei16_v_u32mf2_m))) +vuint32mf2_t vamominuei16(vbool64_t op0, uint32_t * op1, vuint16mf4_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei32_v_u32m1))) +vuint32m1_t vamominuei32(uint32_t * op0, vuint32m1_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei32_v_u32m1_m))) +vuint32m1_t vamominuei32(vbool32_t op0, uint32_t * op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei32_v_u32m2))) +vuint32m2_t vamominuei32(uint32_t * op0, vuint32m2_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei32_v_u32m2_m))) +vuint32m2_t vamominuei32(vbool16_t op0, uint32_t * op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei32_v_u32m4))) +vuint32m4_t vamominuei32(uint32_t * op0, vuint32m4_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei32_v_u32m4_m))) +vuint32m4_t vamominuei32(vbool8_t op0, uint32_t * op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei32_v_u32m8))) +vuint32m8_t vamominuei32(uint32_t * op0, vuint32m8_t op1, vuint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei32_v_u32m8_m))) +vuint32m8_t vamominuei32(vbool4_t op0, uint32_t * op1, vuint32m8_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei32_v_u32mf2))) +vuint32mf2_t vamominuei32(uint32_t * op0, vuint32mf2_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei32_v_u32mf2_m))) +vuint32mf2_t vamominuei32(vbool64_t op0, uint32_t * op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei64_v_u32m1))) +vuint32m1_t vamominuei64(uint32_t * op0, vuint64m2_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei64_v_u32m1_m))) +vuint32m1_t vamominuei64(vbool32_t op0, uint32_t * op1, vuint64m2_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei64_v_u32m2))) +vuint32m2_t vamominuei64(uint32_t * op0, vuint64m4_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei64_v_u32m2_m))) +vuint32m2_t vamominuei64(vbool16_t op0, uint32_t * op1, vuint64m4_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei64_v_u32m4))) +vuint32m4_t vamominuei64(uint32_t * op0, vuint64m8_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei64_v_u32m4_m))) +vuint32m4_t vamominuei64(vbool8_t op0, uint32_t * op1, vuint64m8_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei64_v_u32mf2))) +vuint32mf2_t vamominuei64(uint32_t * op0, vuint64m1_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei64_v_u32mf2_m))) +vuint32mf2_t vamominuei64(vbool64_t op0, uint32_t * op1, vuint64m1_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei8_v_u64m1))) +vuint64m1_t vamominuei8(uint64_t * op0, vuint8mf8_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei8_v_u64m1_m))) +vuint64m1_t vamominuei8(vbool64_t op0, uint64_t * op1, vuint8mf8_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei8_v_u64m2))) +vuint64m2_t vamominuei8(uint64_t * op0, vuint8mf4_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei8_v_u64m2_m))) +vuint64m2_t vamominuei8(vbool32_t op0, uint64_t * op1, vuint8mf4_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei8_v_u64m4))) +vuint64m4_t vamominuei8(uint64_t * op0, vuint8mf2_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei8_v_u64m4_m))) +vuint64m4_t vamominuei8(vbool16_t op0, uint64_t * op1, vuint8mf2_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei8_v_u64m8))) +vuint64m8_t vamominuei8(uint64_t * op0, vuint8m1_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei8_v_u64m8_m))) +vuint64m8_t vamominuei8(vbool8_t op0, uint64_t * op1, vuint8m1_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei16_v_u64m1))) +vuint64m1_t vamominuei16(uint64_t * op0, vuint16mf4_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei16_v_u64m1_m))) +vuint64m1_t vamominuei16(vbool64_t op0, uint64_t * op1, vuint16mf4_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei16_v_u64m2))) +vuint64m2_t vamominuei16(uint64_t * op0, vuint16mf2_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei16_v_u64m2_m))) +vuint64m2_t vamominuei16(vbool32_t op0, uint64_t * op1, vuint16mf2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei16_v_u64m4))) +vuint64m4_t vamominuei16(uint64_t * op0, vuint16m1_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei16_v_u64m4_m))) +vuint64m4_t vamominuei16(vbool16_t op0, uint64_t * op1, vuint16m1_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei16_v_u64m8))) +vuint64m8_t vamominuei16(uint64_t * op0, vuint16m2_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei16_v_u64m8_m))) +vuint64m8_t vamominuei16(vbool8_t op0, uint64_t * op1, vuint16m2_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei32_v_u64m1))) +vuint64m1_t vamominuei32(uint64_t * op0, vuint32mf2_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei32_v_u64m1_m))) +vuint64m1_t vamominuei32(vbool64_t op0, uint64_t * op1, vuint32mf2_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei32_v_u64m2))) +vuint64m2_t vamominuei32(uint64_t * op0, vuint32m1_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei32_v_u64m2_m))) +vuint64m2_t vamominuei32(vbool32_t op0, uint64_t * op1, vuint32m1_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei32_v_u64m4))) +vuint64m4_t vamominuei32(uint64_t * op0, vuint32m2_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei32_v_u64m4_m))) +vuint64m4_t vamominuei32(vbool16_t op0, uint64_t * op1, vuint32m2_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei32_v_u64m8))) +vuint64m8_t vamominuei32(uint64_t * op0, vuint32m4_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei32_v_u64m8_m))) +vuint64m8_t vamominuei32(vbool8_t op0, uint64_t * op1, vuint32m4_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei64_v_u64m1))) +vuint64m1_t vamominuei64(uint64_t * op0, vuint64m1_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei64_v_u64m1_m))) +vuint64m1_t vamominuei64(vbool64_t op0, uint64_t * op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei64_v_u64m2))) +vuint64m2_t vamominuei64(uint64_t * op0, vuint64m2_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei64_v_u64m2_m))) +vuint64m2_t vamominuei64(vbool32_t op0, uint64_t * op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei64_v_u64m4))) +vuint64m4_t vamominuei64(uint64_t * op0, vuint64m4_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei64_v_u64m4_m))) +vuint64m4_t vamominuei64(vbool16_t op0, uint64_t * op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei64_v_u64m8))) +vuint64m8_t vamominuei64(uint64_t * op0, vuint64m8_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamominuei64_v_u64m8_m))) +vuint64m8_t vamominuei64(vbool8_t op0, uint64_t * op1, vuint64m8_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei8_v_u32m1))) +vuint32m1_t vamomaxuei8(uint32_t * op0, vuint8mf4_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei8_v_u32m1_m))) +vuint32m1_t vamomaxuei8(vbool32_t op0, uint32_t * op1, vuint8mf4_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei8_v_u32m2))) +vuint32m2_t vamomaxuei8(uint32_t * op0, vuint8mf2_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei8_v_u32m2_m))) +vuint32m2_t vamomaxuei8(vbool16_t op0, uint32_t * op1, vuint8mf2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei8_v_u32m4))) +vuint32m4_t vamomaxuei8(uint32_t * op0, vuint8m1_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei8_v_u32m4_m))) +vuint32m4_t vamomaxuei8(vbool8_t op0, uint32_t * op1, vuint8m1_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei8_v_u32m8))) +vuint32m8_t vamomaxuei8(uint32_t * op0, vuint8m2_t op1, vuint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei8_v_u32m8_m))) +vuint32m8_t vamomaxuei8(vbool4_t op0, uint32_t * op1, vuint8m2_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei8_v_u32mf2))) +vuint32mf2_t vamomaxuei8(uint32_t * op0, vuint8mf8_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei8_v_u32mf2_m))) +vuint32mf2_t vamomaxuei8(vbool64_t op0, uint32_t * op1, vuint8mf8_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei16_v_u32m1))) +vuint32m1_t vamomaxuei16(uint32_t * op0, vuint16mf2_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei16_v_u32m1_m))) +vuint32m1_t vamomaxuei16(vbool32_t op0, uint32_t * op1, vuint16mf2_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei16_v_u32m2))) +vuint32m2_t vamomaxuei16(uint32_t * op0, vuint16m1_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei16_v_u32m2_m))) +vuint32m2_t vamomaxuei16(vbool16_t op0, uint32_t * op1, vuint16m1_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei16_v_u32m4))) +vuint32m4_t vamomaxuei16(uint32_t * op0, vuint16m2_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei16_v_u32m4_m))) +vuint32m4_t vamomaxuei16(vbool8_t op0, uint32_t * op1, vuint16m2_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei16_v_u32m8))) +vuint32m8_t vamomaxuei16(uint32_t * op0, vuint16m4_t op1, vuint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei16_v_u32m8_m))) +vuint32m8_t vamomaxuei16(vbool4_t op0, uint32_t * op1, vuint16m4_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei16_v_u32mf2))) +vuint32mf2_t vamomaxuei16(uint32_t * op0, vuint16mf4_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei16_v_u32mf2_m))) +vuint32mf2_t vamomaxuei16(vbool64_t op0, uint32_t * op1, vuint16mf4_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei32_v_u32m1))) +vuint32m1_t vamomaxuei32(uint32_t * op0, vuint32m1_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei32_v_u32m1_m))) +vuint32m1_t vamomaxuei32(vbool32_t op0, uint32_t * op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei32_v_u32m2))) +vuint32m2_t vamomaxuei32(uint32_t * op0, vuint32m2_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei32_v_u32m2_m))) +vuint32m2_t vamomaxuei32(vbool16_t op0, uint32_t * op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei32_v_u32m4))) +vuint32m4_t vamomaxuei32(uint32_t * op0, vuint32m4_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei32_v_u32m4_m))) +vuint32m4_t vamomaxuei32(vbool8_t op0, uint32_t * op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei32_v_u32m8))) +vuint32m8_t vamomaxuei32(uint32_t * op0, vuint32m8_t op1, vuint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei32_v_u32m8_m))) +vuint32m8_t vamomaxuei32(vbool4_t op0, uint32_t * op1, vuint32m8_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei32_v_u32mf2))) +vuint32mf2_t vamomaxuei32(uint32_t * op0, vuint32mf2_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei32_v_u32mf2_m))) +vuint32mf2_t vamomaxuei32(vbool64_t op0, uint32_t * op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei64_v_u32m1))) +vuint32m1_t vamomaxuei64(uint32_t * op0, vuint64m2_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei64_v_u32m1_m))) +vuint32m1_t vamomaxuei64(vbool32_t op0, uint32_t * op1, vuint64m2_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei64_v_u32m2))) +vuint32m2_t vamomaxuei64(uint32_t * op0, vuint64m4_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei64_v_u32m2_m))) +vuint32m2_t vamomaxuei64(vbool16_t op0, uint32_t * op1, vuint64m4_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei64_v_u32m4))) +vuint32m4_t vamomaxuei64(uint32_t * op0, vuint64m8_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei64_v_u32m4_m))) +vuint32m4_t vamomaxuei64(vbool8_t op0, uint32_t * op1, vuint64m8_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei64_v_u32mf2))) +vuint32mf2_t vamomaxuei64(uint32_t * op0, vuint64m1_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei64_v_u32mf2_m))) +vuint32mf2_t vamomaxuei64(vbool64_t op0, uint32_t * op1, vuint64m1_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei8_v_u64m1))) +vuint64m1_t vamomaxuei8(uint64_t * op0, vuint8mf8_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei8_v_u64m1_m))) +vuint64m1_t vamomaxuei8(vbool64_t op0, uint64_t * op1, vuint8mf8_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei8_v_u64m2))) +vuint64m2_t vamomaxuei8(uint64_t * op0, vuint8mf4_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei8_v_u64m2_m))) +vuint64m2_t vamomaxuei8(vbool32_t op0, uint64_t * op1, vuint8mf4_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei8_v_u64m4))) +vuint64m4_t vamomaxuei8(uint64_t * op0, vuint8mf2_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei8_v_u64m4_m))) +vuint64m4_t vamomaxuei8(vbool16_t op0, uint64_t * op1, vuint8mf2_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei8_v_u64m8))) +vuint64m8_t vamomaxuei8(uint64_t * op0, vuint8m1_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei8_v_u64m8_m))) +vuint64m8_t vamomaxuei8(vbool8_t op0, uint64_t * op1, vuint8m1_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei16_v_u64m1))) +vuint64m1_t vamomaxuei16(uint64_t * op0, vuint16mf4_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei16_v_u64m1_m))) +vuint64m1_t vamomaxuei16(vbool64_t op0, uint64_t * op1, vuint16mf4_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei16_v_u64m2))) +vuint64m2_t vamomaxuei16(uint64_t * op0, vuint16mf2_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei16_v_u64m2_m))) +vuint64m2_t vamomaxuei16(vbool32_t op0, uint64_t * op1, vuint16mf2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei16_v_u64m4))) +vuint64m4_t vamomaxuei16(uint64_t * op0, vuint16m1_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei16_v_u64m4_m))) +vuint64m4_t vamomaxuei16(vbool16_t op0, uint64_t * op1, vuint16m1_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei16_v_u64m8))) +vuint64m8_t vamomaxuei16(uint64_t * op0, vuint16m2_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei16_v_u64m8_m))) +vuint64m8_t vamomaxuei16(vbool8_t op0, uint64_t * op1, vuint16m2_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei32_v_u64m1))) +vuint64m1_t vamomaxuei32(uint64_t * op0, vuint32mf2_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei32_v_u64m1_m))) +vuint64m1_t vamomaxuei32(vbool64_t op0, uint64_t * op1, vuint32mf2_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei32_v_u64m2))) +vuint64m2_t vamomaxuei32(uint64_t * op0, vuint32m1_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei32_v_u64m2_m))) +vuint64m2_t vamomaxuei32(vbool32_t op0, uint64_t * op1, vuint32m1_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei32_v_u64m4))) +vuint64m4_t vamomaxuei32(uint64_t * op0, vuint32m2_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei32_v_u64m4_m))) +vuint64m4_t vamomaxuei32(vbool16_t op0, uint64_t * op1, vuint32m2_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei32_v_u64m8))) +vuint64m8_t vamomaxuei32(uint64_t * op0, vuint32m4_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei32_v_u64m8_m))) +vuint64m8_t vamomaxuei32(vbool8_t op0, uint64_t * op1, vuint32m4_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei64_v_u64m1))) +vuint64m1_t vamomaxuei64(uint64_t * op0, vuint64m1_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei64_v_u64m1_m))) +vuint64m1_t vamomaxuei64(vbool64_t op0, uint64_t * op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei64_v_u64m2))) +vuint64m2_t vamomaxuei64(uint64_t * op0, vuint64m2_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei64_v_u64m2_m))) +vuint64m2_t vamomaxuei64(vbool32_t op0, uint64_t * op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei64_v_u64m4))) +vuint64m4_t vamomaxuei64(uint64_t * op0, vuint64m4_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei64_v_u64m4_m))) +vuint64m4_t vamomaxuei64(vbool16_t op0, uint64_t * op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei64_v_u64m8))) +vuint64m8_t vamomaxuei64(uint64_t * op0, vuint64m8_t op1, vuint64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamomaxuei64_v_u64m8_m))) +vuint64m8_t vamomaxuei64(vbool8_t op0, uint64_t * op1, vuint64m8_t op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_i32m1))) +vint32m1_t vamoswapei8(int32_t * op0, vuint8mf4_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_i32m1_m))) +vint32m1_t vamoswapei8(vbool32_t op0, int32_t * op1, vuint8mf4_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_i32m2))) +vint32m2_t vamoswapei8(int32_t * op0, vuint8mf2_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_i32m2_m))) +vint32m2_t vamoswapei8(vbool16_t op0, int32_t * op1, vuint8mf2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_i32m4))) +vint32m4_t vamoswapei8(int32_t * op0, vuint8m1_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_i32m4_m))) +vint32m4_t vamoswapei8(vbool8_t op0, int32_t * op1, vuint8m1_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_i32m8))) +vint32m8_t vamoswapei8(int32_t * op0, vuint8m2_t op1, vint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_i32m8_m))) +vint32m8_t vamoswapei8(vbool4_t op0, int32_t * op1, vuint8m2_t op2, vint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_i32mf2))) +vint32mf2_t vamoswapei8(int32_t * op0, vuint8mf8_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_i32mf2_m))) +vint32mf2_t vamoswapei8(vbool64_t op0, int32_t * op1, vuint8mf8_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_u32m1))) +vuint32m1_t vamoswapei8(uint32_t * op0, vuint8mf4_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_u32m1_m))) +vuint32m1_t vamoswapei8(vbool32_t op0, uint32_t * op1, vuint8mf4_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_u32m2))) +vuint32m2_t vamoswapei8(uint32_t * op0, vuint8mf2_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_u32m2_m))) +vuint32m2_t vamoswapei8(vbool16_t op0, uint32_t * op1, vuint8mf2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_u32m4))) +vuint32m4_t vamoswapei8(uint32_t * op0, vuint8m1_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_u32m4_m))) +vuint32m4_t vamoswapei8(vbool8_t op0, uint32_t * op1, vuint8m1_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_u32m8))) +vuint32m8_t vamoswapei8(uint32_t * op0, vuint8m2_t op1, vuint32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_u32m8_m))) +vuint32m8_t vamoswapei8(vbool4_t op0, uint32_t * op1, vuint8m2_t op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_u32mf2))) +vuint32mf2_t vamoswapei8(uint32_t * op0, vuint8mf8_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_u32mf2_m))) +vuint32mf2_t vamoswapei8(vbool64_t op0, uint32_t * op1, vuint8mf8_t op2, vuint32mf2_t op3, size_t op4); + +#endif + +#if defined(__riscv_f) && defined(__riscv_zvamo) +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_f32m1))) +vfloat32m1_t vamoswapei8(float * op0, vuint8mf4_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_f32m1_m))) +vfloat32m1_t vamoswapei8(vbool32_t op0, float * op1, vuint8mf4_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_f32m2))) +vfloat32m2_t vamoswapei8(float * op0, vuint8mf2_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_f32m2_m))) +vfloat32m2_t vamoswapei8(vbool16_t op0, float * op1, vuint8mf2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_f32m4))) +vfloat32m4_t vamoswapei8(float * op0, vuint8m1_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_f32m4_m))) +vfloat32m4_t vamoswapei8(vbool8_t op0, float * op1, vuint8m1_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_f32m8))) +vfloat32m8_t vamoswapei8(float * op0, vuint8m2_t op1, vfloat32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_f32m8_m))) +vfloat32m8_t vamoswapei8(vbool4_t op0, float * op1, vuint8m2_t op2, vfloat32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_f32mf2))) +vfloat32mf2_t vamoswapei8(float * op0, vuint8mf8_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_f32mf2_m))) +vfloat32mf2_t vamoswapei8(vbool64_t op0, float * op1, vuint8mf8_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_f32m1))) +vfloat32m1_t vamoswapei16(float * op0, vuint16mf2_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_f32m1_m))) +vfloat32m1_t vamoswapei16(vbool32_t op0, float * op1, vuint16mf2_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_f32m2))) +vfloat32m2_t vamoswapei16(float * op0, vuint16m1_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_f32m2_m))) +vfloat32m2_t vamoswapei16(vbool16_t op0, float * op1, vuint16m1_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_f32m4))) +vfloat32m4_t vamoswapei16(float * op0, vuint16m2_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_f32m4_m))) +vfloat32m4_t vamoswapei16(vbool8_t op0, float * op1, vuint16m2_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_f32m8))) +vfloat32m8_t vamoswapei16(float * op0, vuint16m4_t op1, vfloat32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_f32m8_m))) +vfloat32m8_t vamoswapei16(vbool4_t op0, float * op1, vuint16m4_t op2, vfloat32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_f32mf2))) +vfloat32mf2_t vamoswapei16(float * op0, vuint16mf4_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_f32mf2_m))) +vfloat32mf2_t vamoswapei16(vbool64_t op0, float * op1, vuint16mf4_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_f32m1))) +vfloat32m1_t vamoswapei32(float * op0, vuint32m1_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_f32m1_m))) +vfloat32m1_t vamoswapei32(vbool32_t op0, float * op1, vuint32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_f32m2))) +vfloat32m2_t vamoswapei32(float * op0, vuint32m2_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_f32m2_m))) +vfloat32m2_t vamoswapei32(vbool16_t op0, float * op1, vuint32m2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_f32m4))) +vfloat32m4_t vamoswapei32(float * op0, vuint32m4_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_f32m4_m))) +vfloat32m4_t vamoswapei32(vbool8_t op0, float * op1, vuint32m4_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_f32m8))) +vfloat32m8_t vamoswapei32(float * op0, vuint32m8_t op1, vfloat32m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_f32m8_m))) +vfloat32m8_t vamoswapei32(vbool4_t op0, float * op1, vuint32m8_t op2, vfloat32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_f32mf2))) +vfloat32mf2_t vamoswapei32(float * op0, vuint32mf2_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_f32mf2_m))) +vfloat32mf2_t vamoswapei32(vbool64_t op0, float * op1, vuint32mf2_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_f32m1))) +vfloat32m1_t vamoswapei64(float * op0, vuint64m2_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_f32m1_m))) +vfloat32m1_t vamoswapei64(vbool32_t op0, float * op1, vuint64m2_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_f32m2))) +vfloat32m2_t vamoswapei64(float * op0, vuint64m4_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_f32m2_m))) +vfloat32m2_t vamoswapei64(vbool16_t op0, float * op1, vuint64m4_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_f32m4))) +vfloat32m4_t vamoswapei64(float * op0, vuint64m8_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_f32m4_m))) +vfloat32m4_t vamoswapei64(vbool8_t op0, float * op1, vuint64m8_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_f32mf2))) +vfloat32mf2_t vamoswapei64(float * op0, vuint64m1_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_f32mf2_m))) +vfloat32mf2_t vamoswapei64(vbool64_t op0, float * op1, vuint64m1_t op2, vfloat32mf2_t op3, size_t op4); + +#endif + +#if defined(__riscv_d) && defined(__riscv_zvamo) +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_f64m1))) +vfloat64m1_t vamoswapei8(double * op0, vuint8mf8_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_f64m1_m))) +vfloat64m1_t vamoswapei8(vbool64_t op0, double * op1, vuint8mf8_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_f64m2))) +vfloat64m2_t vamoswapei8(double * op0, vuint8mf4_t op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_f64m2_m))) +vfloat64m2_t vamoswapei8(vbool32_t op0, double * op1, vuint8mf4_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_f64m4))) +vfloat64m4_t vamoswapei8(double * op0, vuint8mf2_t op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_f64m4_m))) +vfloat64m4_t vamoswapei8(vbool16_t op0, double * op1, vuint8mf2_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_f64m8))) +vfloat64m8_t vamoswapei8(double * op0, vuint8m1_t op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei8_v_f64m8_m))) +vfloat64m8_t vamoswapei8(vbool8_t op0, double * op1, vuint8m1_t op2, vfloat64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_f64m1))) +vfloat64m1_t vamoswapei16(double * op0, vuint16mf4_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_f64m1_m))) +vfloat64m1_t vamoswapei16(vbool64_t op0, double * op1, vuint16mf4_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_f64m2))) +vfloat64m2_t vamoswapei16(double * op0, vuint16mf2_t op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_f64m2_m))) +vfloat64m2_t vamoswapei16(vbool32_t op0, double * op1, vuint16mf2_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_f64m4))) +vfloat64m4_t vamoswapei16(double * op0, vuint16m1_t op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_f64m4_m))) +vfloat64m4_t vamoswapei16(vbool16_t op0, double * op1, vuint16m1_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_f64m8))) +vfloat64m8_t vamoswapei16(double * op0, vuint16m2_t op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei16_v_f64m8_m))) +vfloat64m8_t vamoswapei16(vbool8_t op0, double * op1, vuint16m2_t op2, vfloat64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_f64m1))) +vfloat64m1_t vamoswapei32(double * op0, vuint32mf2_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_f64m1_m))) +vfloat64m1_t vamoswapei32(vbool64_t op0, double * op1, vuint32mf2_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_f64m2))) +vfloat64m2_t vamoswapei32(double * op0, vuint32m1_t op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_f64m2_m))) +vfloat64m2_t vamoswapei32(vbool32_t op0, double * op1, vuint32m1_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_f64m4))) +vfloat64m4_t vamoswapei32(double * op0, vuint32m2_t op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_f64m4_m))) +vfloat64m4_t vamoswapei32(vbool16_t op0, double * op1, vuint32m2_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_f64m8))) +vfloat64m8_t vamoswapei32(double * op0, vuint32m4_t op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei32_v_f64m8_m))) +vfloat64m8_t vamoswapei32(vbool8_t op0, double * op1, vuint32m4_t op2, vfloat64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_f64m1))) +vfloat64m1_t vamoswapei64(double * op0, vuint64m1_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_f64m1_m))) +vfloat64m1_t vamoswapei64(vbool64_t op0, double * op1, vuint64m1_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_f64m2))) +vfloat64m2_t vamoswapei64(double * op0, vuint64m2_t op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_f64m2_m))) +vfloat64m2_t vamoswapei64(vbool32_t op0, double * op1, vuint64m2_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_f64m4))) +vfloat64m4_t vamoswapei64(double * op0, vuint64m4_t op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_f64m4_m))) +vfloat64m4_t vamoswapei64(vbool16_t op0, double * op1, vuint64m4_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_f64m8))) +vfloat64m8_t vamoswapei64(double * op0, vuint64m8_t op1, vfloat64m8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vamoswapei64_v_f64m8_m))) +vfloat64m8_t vamoswapei64(vbool8_t op0, double * op1, vuint64m8_t op2, vfloat64m8_t op3, size_t op4); + +#endif + +#if defined(__riscv_zvlsseg) +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_i64m1))) +void vloxseg2ei8(vint64m1_t * op0, vint64m1_t * op1, const int64_t * op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_i64m1_m))) +void vloxseg2ei8(vint64m1_t * op0, vint64m1_t * op1, vbool64_t op2, vint64m1_t op3, vint64m1_t op4, const int64_t * op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_i64m2))) +void vloxseg2ei8(vint64m2_t * op0, vint64m2_t * op1, const int64_t * op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_i64m2_m))) +void vloxseg2ei8(vint64m2_t * op0, vint64m2_t * op1, vbool32_t op2, vint64m2_t op3, vint64m2_t op4, const int64_t * op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_i64m4))) +void vloxseg2ei8(vint64m4_t * op0, vint64m4_t * op1, const int64_t * op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_i64m4_m))) +void vloxseg2ei8(vint64m4_t * op0, vint64m4_t * op1, vbool16_t op2, vint64m4_t op3, vint64m4_t op4, const int64_t * op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_u64m1))) +void vloxseg2ei8(vuint64m1_t * op0, vuint64m1_t * op1, const uint64_t * op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_u64m1_m))) +void vloxseg2ei8(vuint64m1_t * op0, vuint64m1_t * op1, vbool64_t op2, vuint64m1_t op3, vuint64m1_t op4, const uint64_t * op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_u64m2))) +void vloxseg2ei8(vuint64m2_t * op0, vuint64m2_t * op1, const uint64_t * op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_u64m2_m))) +void vloxseg2ei8(vuint64m2_t * op0, vuint64m2_t * op1, vbool32_t op2, vuint64m2_t op3, vuint64m2_t op4, const uint64_t * op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_u64m4))) +void vloxseg2ei8(vuint64m4_t * op0, vuint64m4_t * op1, const uint64_t * op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_u64m4_m))) +void vloxseg2ei8(vuint64m4_t * op0, vuint64m4_t * op1, vbool16_t op2, vuint64m4_t op3, vuint64m4_t op4, const uint64_t * op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_i64m1))) +void vloxseg3ei8(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, const int64_t * op3, vuint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_i64m1_m))) +void vloxseg3ei8(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vbool64_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, const int64_t * op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_i64m2))) +void vloxseg3ei8(vint64m2_t * op0, vint64m2_t * op1, vint64m2_t * op2, const int64_t * op3, vuint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_i64m2_m))) +void vloxseg3ei8(vint64m2_t * op0, vint64m2_t * op1, vint64m2_t * op2, vbool32_t op3, vint64m2_t op4, vint64m2_t op5, vint64m2_t op6, const int64_t * op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_u64m1))) +void vloxseg3ei8(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, const uint64_t * op3, vuint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_u64m1_m))) +void vloxseg3ei8(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vbool64_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, const uint64_t * op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_u64m2))) +void vloxseg3ei8(vuint64m2_t * op0, vuint64m2_t * op1, vuint64m2_t * op2, const uint64_t * op3, vuint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_u64m2_m))) +void vloxseg3ei8(vuint64m2_t * op0, vuint64m2_t * op1, vuint64m2_t * op2, vbool32_t op3, vuint64m2_t op4, vuint64m2_t op5, vuint64m2_t op6, const uint64_t * op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_i64m1))) +void vloxseg4ei8(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, const int64_t * op4, vuint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_i64m1_m))) +void vloxseg4ei8(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vbool64_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, const int64_t * op9, vuint8mf8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_i64m2))) +void vloxseg4ei8(vint64m2_t * op0, vint64m2_t * op1, vint64m2_t * op2, vint64m2_t * op3, const int64_t * op4, vuint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_i64m2_m))) +void vloxseg4ei8(vint64m2_t * op0, vint64m2_t * op1, vint64m2_t * op2, vint64m2_t * op3, vbool32_t op4, vint64m2_t op5, vint64m2_t op6, vint64m2_t op7, vint64m2_t op8, const int64_t * op9, vuint8mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_u64m1))) +void vloxseg4ei8(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, const uint64_t * op4, vuint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_u64m1_m))) +void vloxseg4ei8(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vbool64_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, const uint64_t * op9, vuint8mf8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_u64m2))) +void vloxseg4ei8(vuint64m2_t * op0, vuint64m2_t * op1, vuint64m2_t * op2, vuint64m2_t * op3, const uint64_t * op4, vuint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_u64m2_m))) +void vloxseg4ei8(vuint64m2_t * op0, vuint64m2_t * op1, vuint64m2_t * op2, vuint64m2_t * op3, vbool32_t op4, vuint64m2_t op5, vuint64m2_t op6, vuint64m2_t op7, vuint64m2_t op8, const uint64_t * op9, vuint8mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_i64m1))) +void vloxseg5ei8(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, const int64_t * op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_i64m1_m))) +void vloxseg5ei8(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vbool64_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, vint64m1_t op10, const int64_t * op11, vuint8mf8_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_u64m1))) +void vloxseg5ei8(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, const uint64_t * op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_u64m1_m))) +void vloxseg5ei8(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vbool64_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, vuint64m1_t op10, const uint64_t * op11, vuint8mf8_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_i64m1))) +void vloxseg6ei8(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, const int64_t * op6, vuint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_i64m1_m))) +void vloxseg6ei8(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vbool64_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, vint64m1_t op10, vint64m1_t op11, vint64m1_t op12, const int64_t * op13, vuint8mf8_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_u64m1))) +void vloxseg6ei8(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, const uint64_t * op6, vuint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_u64m1_m))) +void vloxseg6ei8(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vbool64_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, vuint64m1_t op10, vuint64m1_t op11, vuint64m1_t op12, const uint64_t * op13, vuint8mf8_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_i64m1))) +void vloxseg7ei8(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vint64m1_t * op6, const int64_t * op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_i64m1_m))) +void vloxseg7ei8(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vint64m1_t * op6, vbool64_t op7, vint64m1_t op8, vint64m1_t op9, vint64m1_t op10, vint64m1_t op11, vint64m1_t op12, vint64m1_t op13, vint64m1_t op14, const int64_t * op15, vuint8mf8_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_u64m1))) +void vloxseg7ei8(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vuint64m1_t * op6, const uint64_t * op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_u64m1_m))) +void vloxseg7ei8(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vuint64m1_t * op6, vbool64_t op7, vuint64m1_t op8, vuint64m1_t op9, vuint64m1_t op10, vuint64m1_t op11, vuint64m1_t op12, vuint64m1_t op13, vuint64m1_t op14, const uint64_t * op15, vuint8mf8_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_i64m1))) +void vloxseg8ei8(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vint64m1_t * op6, vint64m1_t * op7, const int64_t * op8, vuint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_i64m1_m))) +void vloxseg8ei8(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vint64m1_t * op6, vint64m1_t * op7, vbool64_t op8, vint64m1_t op9, vint64m1_t op10, vint64m1_t op11, vint64m1_t op12, vint64m1_t op13, vint64m1_t op14, vint64m1_t op15, vint64m1_t op16, const int64_t * op17, vuint8mf8_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_u64m1))) +void vloxseg8ei8(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vuint64m1_t * op6, vuint64m1_t * op7, const uint64_t * op8, vuint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_u64m1_m))) +void vloxseg8ei8(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vuint64m1_t * op6, vuint64m1_t * op7, vbool64_t op8, vuint64m1_t op9, vuint64m1_t op10, vuint64m1_t op11, vuint64m1_t op12, vuint64m1_t op13, vuint64m1_t op14, vuint64m1_t op15, vuint64m1_t op16, const uint64_t * op17, vuint8mf8_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_i64m1))) +void vloxseg2ei16(vint64m1_t * op0, vint64m1_t * op1, const int64_t * op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_i64m1_m))) +void vloxseg2ei16(vint64m1_t * op0, vint64m1_t * op1, vbool64_t op2, vint64m1_t op3, vint64m1_t op4, const int64_t * op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_i64m2))) +void vloxseg2ei16(vint64m2_t * op0, vint64m2_t * op1, const int64_t * op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_i64m2_m))) +void vloxseg2ei16(vint64m2_t * op0, vint64m2_t * op1, vbool32_t op2, vint64m2_t op3, vint64m2_t op4, const int64_t * op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_i64m4))) +void vloxseg2ei16(vint64m4_t * op0, vint64m4_t * op1, const int64_t * op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_i64m4_m))) +void vloxseg2ei16(vint64m4_t * op0, vint64m4_t * op1, vbool16_t op2, vint64m4_t op3, vint64m4_t op4, const int64_t * op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_u64m1))) +void vloxseg2ei16(vuint64m1_t * op0, vuint64m1_t * op1, const uint64_t * op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_u64m1_m))) +void vloxseg2ei16(vuint64m1_t * op0, vuint64m1_t * op1, vbool64_t op2, vuint64m1_t op3, vuint64m1_t op4, const uint64_t * op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_u64m2))) +void vloxseg2ei16(vuint64m2_t * op0, vuint64m2_t * op1, const uint64_t * op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_u64m2_m))) +void vloxseg2ei16(vuint64m2_t * op0, vuint64m2_t * op1, vbool32_t op2, vuint64m2_t op3, vuint64m2_t op4, const uint64_t * op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_u64m4))) +void vloxseg2ei16(vuint64m4_t * op0, vuint64m4_t * op1, const uint64_t * op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_u64m4_m))) +void vloxseg2ei16(vuint64m4_t * op0, vuint64m4_t * op1, vbool16_t op2, vuint64m4_t op3, vuint64m4_t op4, const uint64_t * op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_i64m1))) +void vloxseg3ei16(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, const int64_t * op3, vuint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_i64m1_m))) +void vloxseg3ei16(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vbool64_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, const int64_t * op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_i64m2))) +void vloxseg3ei16(vint64m2_t * op0, vint64m2_t * op1, vint64m2_t * op2, const int64_t * op3, vuint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_i64m2_m))) +void vloxseg3ei16(vint64m2_t * op0, vint64m2_t * op1, vint64m2_t * op2, vbool32_t op3, vint64m2_t op4, vint64m2_t op5, vint64m2_t op6, const int64_t * op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_u64m1))) +void vloxseg3ei16(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, const uint64_t * op3, vuint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_u64m1_m))) +void vloxseg3ei16(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vbool64_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, const uint64_t * op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_u64m2))) +void vloxseg3ei16(vuint64m2_t * op0, vuint64m2_t * op1, vuint64m2_t * op2, const uint64_t * op3, vuint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_u64m2_m))) +void vloxseg3ei16(vuint64m2_t * op0, vuint64m2_t * op1, vuint64m2_t * op2, vbool32_t op3, vuint64m2_t op4, vuint64m2_t op5, vuint64m2_t op6, const uint64_t * op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_i64m1))) +void vloxseg4ei16(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, const int64_t * op4, vuint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_i64m1_m))) +void vloxseg4ei16(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vbool64_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, const int64_t * op9, vuint16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_i64m2))) +void vloxseg4ei16(vint64m2_t * op0, vint64m2_t * op1, vint64m2_t * op2, vint64m2_t * op3, const int64_t * op4, vuint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_i64m2_m))) +void vloxseg4ei16(vint64m2_t * op0, vint64m2_t * op1, vint64m2_t * op2, vint64m2_t * op3, vbool32_t op4, vint64m2_t op5, vint64m2_t op6, vint64m2_t op7, vint64m2_t op8, const int64_t * op9, vuint16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_u64m1))) +void vloxseg4ei16(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, const uint64_t * op4, vuint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_u64m1_m))) +void vloxseg4ei16(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vbool64_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, const uint64_t * op9, vuint16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_u64m2))) +void vloxseg4ei16(vuint64m2_t * op0, vuint64m2_t * op1, vuint64m2_t * op2, vuint64m2_t * op3, const uint64_t * op4, vuint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_u64m2_m))) +void vloxseg4ei16(vuint64m2_t * op0, vuint64m2_t * op1, vuint64m2_t * op2, vuint64m2_t * op3, vbool32_t op4, vuint64m2_t op5, vuint64m2_t op6, vuint64m2_t op7, vuint64m2_t op8, const uint64_t * op9, vuint16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_i64m1))) +void vloxseg5ei16(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, const int64_t * op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_i64m1_m))) +void vloxseg5ei16(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vbool64_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, vint64m1_t op10, const int64_t * op11, vuint16mf4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_u64m1))) +void vloxseg5ei16(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, const uint64_t * op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_u64m1_m))) +void vloxseg5ei16(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vbool64_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, vuint64m1_t op10, const uint64_t * op11, vuint16mf4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_i64m1))) +void vloxseg6ei16(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, const int64_t * op6, vuint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_i64m1_m))) +void vloxseg6ei16(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vbool64_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, vint64m1_t op10, vint64m1_t op11, vint64m1_t op12, const int64_t * op13, vuint16mf4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_u64m1))) +void vloxseg6ei16(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, const uint64_t * op6, vuint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_u64m1_m))) +void vloxseg6ei16(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vbool64_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, vuint64m1_t op10, vuint64m1_t op11, vuint64m1_t op12, const uint64_t * op13, vuint16mf4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_i64m1))) +void vloxseg7ei16(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vint64m1_t * op6, const int64_t * op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_i64m1_m))) +void vloxseg7ei16(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vint64m1_t * op6, vbool64_t op7, vint64m1_t op8, vint64m1_t op9, vint64m1_t op10, vint64m1_t op11, vint64m1_t op12, vint64m1_t op13, vint64m1_t op14, const int64_t * op15, vuint16mf4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_u64m1))) +void vloxseg7ei16(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vuint64m1_t * op6, const uint64_t * op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_u64m1_m))) +void vloxseg7ei16(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vuint64m1_t * op6, vbool64_t op7, vuint64m1_t op8, vuint64m1_t op9, vuint64m1_t op10, vuint64m1_t op11, vuint64m1_t op12, vuint64m1_t op13, vuint64m1_t op14, const uint64_t * op15, vuint16mf4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_i64m1))) +void vloxseg8ei16(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vint64m1_t * op6, vint64m1_t * op7, const int64_t * op8, vuint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_i64m1_m))) +void vloxseg8ei16(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vint64m1_t * op6, vint64m1_t * op7, vbool64_t op8, vint64m1_t op9, vint64m1_t op10, vint64m1_t op11, vint64m1_t op12, vint64m1_t op13, vint64m1_t op14, vint64m1_t op15, vint64m1_t op16, const int64_t * op17, vuint16mf4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_u64m1))) +void vloxseg8ei16(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vuint64m1_t * op6, vuint64m1_t * op7, const uint64_t * op8, vuint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_u64m1_m))) +void vloxseg8ei16(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vuint64m1_t * op6, vuint64m1_t * op7, vbool64_t op8, vuint64m1_t op9, vuint64m1_t op10, vuint64m1_t op11, vuint64m1_t op12, vuint64m1_t op13, vuint64m1_t op14, vuint64m1_t op15, vuint64m1_t op16, const uint64_t * op17, vuint16mf4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_i64m1))) +void vloxseg2ei32(vint64m1_t * op0, vint64m1_t * op1, const int64_t * op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_i64m1_m))) +void vloxseg2ei32(vint64m1_t * op0, vint64m1_t * op1, vbool64_t op2, vint64m1_t op3, vint64m1_t op4, const int64_t * op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_i64m2))) +void vloxseg2ei32(vint64m2_t * op0, vint64m2_t * op1, const int64_t * op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_i64m2_m))) +void vloxseg2ei32(vint64m2_t * op0, vint64m2_t * op1, vbool32_t op2, vint64m2_t op3, vint64m2_t op4, const int64_t * op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_i64m4))) +void vloxseg2ei32(vint64m4_t * op0, vint64m4_t * op1, const int64_t * op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_i64m4_m))) +void vloxseg2ei32(vint64m4_t * op0, vint64m4_t * op1, vbool16_t op2, vint64m4_t op3, vint64m4_t op4, const int64_t * op5, vuint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_u64m1))) +void vloxseg2ei32(vuint64m1_t * op0, vuint64m1_t * op1, const uint64_t * op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_u64m1_m))) +void vloxseg2ei32(vuint64m1_t * op0, vuint64m1_t * op1, vbool64_t op2, vuint64m1_t op3, vuint64m1_t op4, const uint64_t * op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_u64m2))) +void vloxseg2ei32(vuint64m2_t * op0, vuint64m2_t * op1, const uint64_t * op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_u64m2_m))) +void vloxseg2ei32(vuint64m2_t * op0, vuint64m2_t * op1, vbool32_t op2, vuint64m2_t op3, vuint64m2_t op4, const uint64_t * op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_u64m4))) +void vloxseg2ei32(vuint64m4_t * op0, vuint64m4_t * op1, const uint64_t * op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_u64m4_m))) +void vloxseg2ei32(vuint64m4_t * op0, vuint64m4_t * op1, vbool16_t op2, vuint64m4_t op3, vuint64m4_t op4, const uint64_t * op5, vuint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_i64m1))) +void vloxseg3ei32(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, const int64_t * op3, vuint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_i64m1_m))) +void vloxseg3ei32(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vbool64_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, const int64_t * op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_i64m2))) +void vloxseg3ei32(vint64m2_t * op0, vint64m2_t * op1, vint64m2_t * op2, const int64_t * op3, vuint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_i64m2_m))) +void vloxseg3ei32(vint64m2_t * op0, vint64m2_t * op1, vint64m2_t * op2, vbool32_t op3, vint64m2_t op4, vint64m2_t op5, vint64m2_t op6, const int64_t * op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_u64m1))) +void vloxseg3ei32(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, const uint64_t * op3, vuint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_u64m1_m))) +void vloxseg3ei32(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vbool64_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, const uint64_t * op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_u64m2))) +void vloxseg3ei32(vuint64m2_t * op0, vuint64m2_t * op1, vuint64m2_t * op2, const uint64_t * op3, vuint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_u64m2_m))) +void vloxseg3ei32(vuint64m2_t * op0, vuint64m2_t * op1, vuint64m2_t * op2, vbool32_t op3, vuint64m2_t op4, vuint64m2_t op5, vuint64m2_t op6, const uint64_t * op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_i64m1))) +void vloxseg4ei32(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, const int64_t * op4, vuint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_i64m1_m))) +void vloxseg4ei32(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vbool64_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, const int64_t * op9, vuint32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_i64m2))) +void vloxseg4ei32(vint64m2_t * op0, vint64m2_t * op1, vint64m2_t * op2, vint64m2_t * op3, const int64_t * op4, vuint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_i64m2_m))) +void vloxseg4ei32(vint64m2_t * op0, vint64m2_t * op1, vint64m2_t * op2, vint64m2_t * op3, vbool32_t op4, vint64m2_t op5, vint64m2_t op6, vint64m2_t op7, vint64m2_t op8, const int64_t * op9, vuint32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_u64m1))) +void vloxseg4ei32(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, const uint64_t * op4, vuint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_u64m1_m))) +void vloxseg4ei32(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vbool64_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, const uint64_t * op9, vuint32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_u64m2))) +void vloxseg4ei32(vuint64m2_t * op0, vuint64m2_t * op1, vuint64m2_t * op2, vuint64m2_t * op3, const uint64_t * op4, vuint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_u64m2_m))) +void vloxseg4ei32(vuint64m2_t * op0, vuint64m2_t * op1, vuint64m2_t * op2, vuint64m2_t * op3, vbool32_t op4, vuint64m2_t op5, vuint64m2_t op6, vuint64m2_t op7, vuint64m2_t op8, const uint64_t * op9, vuint32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_i64m1))) +void vloxseg5ei32(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, const int64_t * op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_i64m1_m))) +void vloxseg5ei32(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vbool64_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, vint64m1_t op10, const int64_t * op11, vuint32mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_u64m1))) +void vloxseg5ei32(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, const uint64_t * op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_u64m1_m))) +void vloxseg5ei32(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vbool64_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, vuint64m1_t op10, const uint64_t * op11, vuint32mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_i64m1))) +void vloxseg6ei32(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, const int64_t * op6, vuint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_i64m1_m))) +void vloxseg6ei32(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vbool64_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, vint64m1_t op10, vint64m1_t op11, vint64m1_t op12, const int64_t * op13, vuint32mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_u64m1))) +void vloxseg6ei32(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, const uint64_t * op6, vuint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_u64m1_m))) +void vloxseg6ei32(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vbool64_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, vuint64m1_t op10, vuint64m1_t op11, vuint64m1_t op12, const uint64_t * op13, vuint32mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_i64m1))) +void vloxseg7ei32(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vint64m1_t * op6, const int64_t * op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_i64m1_m))) +void vloxseg7ei32(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vint64m1_t * op6, vbool64_t op7, vint64m1_t op8, vint64m1_t op9, vint64m1_t op10, vint64m1_t op11, vint64m1_t op12, vint64m1_t op13, vint64m1_t op14, const int64_t * op15, vuint32mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_u64m1))) +void vloxseg7ei32(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vuint64m1_t * op6, const uint64_t * op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_u64m1_m))) +void vloxseg7ei32(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vuint64m1_t * op6, vbool64_t op7, vuint64m1_t op8, vuint64m1_t op9, vuint64m1_t op10, vuint64m1_t op11, vuint64m1_t op12, vuint64m1_t op13, vuint64m1_t op14, const uint64_t * op15, vuint32mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_i64m1))) +void vloxseg8ei32(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vint64m1_t * op6, vint64m1_t * op7, const int64_t * op8, vuint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_i64m1_m))) +void vloxseg8ei32(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vint64m1_t * op6, vint64m1_t * op7, vbool64_t op8, vint64m1_t op9, vint64m1_t op10, vint64m1_t op11, vint64m1_t op12, vint64m1_t op13, vint64m1_t op14, vint64m1_t op15, vint64m1_t op16, const int64_t * op17, vuint32mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_u64m1))) +void vloxseg8ei32(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vuint64m1_t * op6, vuint64m1_t * op7, const uint64_t * op8, vuint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_u64m1_m))) +void vloxseg8ei32(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vuint64m1_t * op6, vuint64m1_t * op7, vbool64_t op8, vuint64m1_t op9, vuint64m1_t op10, vuint64m1_t op11, vuint64m1_t op12, vuint64m1_t op13, vuint64m1_t op14, vuint64m1_t op15, vuint64m1_t op16, const uint64_t * op17, vuint32mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_i64m1))) +void vloxseg2ei64(vint64m1_t * op0, vint64m1_t * op1, const int64_t * op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_i64m1_m))) +void vloxseg2ei64(vint64m1_t * op0, vint64m1_t * op1, vbool64_t op2, vint64m1_t op3, vint64m1_t op4, const int64_t * op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_i64m2))) +void vloxseg2ei64(vint64m2_t * op0, vint64m2_t * op1, const int64_t * op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_i64m2_m))) +void vloxseg2ei64(vint64m2_t * op0, vint64m2_t * op1, vbool32_t op2, vint64m2_t op3, vint64m2_t op4, const int64_t * op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_i64m4))) +void vloxseg2ei64(vint64m4_t * op0, vint64m4_t * op1, const int64_t * op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_i64m4_m))) +void vloxseg2ei64(vint64m4_t * op0, vint64m4_t * op1, vbool16_t op2, vint64m4_t op3, vint64m4_t op4, const int64_t * op5, vuint64m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_u64m1))) +void vloxseg2ei64(vuint64m1_t * op0, vuint64m1_t * op1, const uint64_t * op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_u64m1_m))) +void vloxseg2ei64(vuint64m1_t * op0, vuint64m1_t * op1, vbool64_t op2, vuint64m1_t op3, vuint64m1_t op4, const uint64_t * op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_u64m2))) +void vloxseg2ei64(vuint64m2_t * op0, vuint64m2_t * op1, const uint64_t * op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_u64m2_m))) +void vloxseg2ei64(vuint64m2_t * op0, vuint64m2_t * op1, vbool32_t op2, vuint64m2_t op3, vuint64m2_t op4, const uint64_t * op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_u64m4))) +void vloxseg2ei64(vuint64m4_t * op0, vuint64m4_t * op1, const uint64_t * op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_u64m4_m))) +void vloxseg2ei64(vuint64m4_t * op0, vuint64m4_t * op1, vbool16_t op2, vuint64m4_t op3, vuint64m4_t op4, const uint64_t * op5, vuint64m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_i64m1))) +void vloxseg3ei64(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, const int64_t * op3, vuint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_i64m1_m))) +void vloxseg3ei64(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vbool64_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, const int64_t * op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_i64m2))) +void vloxseg3ei64(vint64m2_t * op0, vint64m2_t * op1, vint64m2_t * op2, const int64_t * op3, vuint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_i64m2_m))) +void vloxseg3ei64(vint64m2_t * op0, vint64m2_t * op1, vint64m2_t * op2, vbool32_t op3, vint64m2_t op4, vint64m2_t op5, vint64m2_t op6, const int64_t * op7, vuint64m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_u64m1))) +void vloxseg3ei64(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, const uint64_t * op3, vuint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_u64m1_m))) +void vloxseg3ei64(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vbool64_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, const uint64_t * op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_u64m2))) +void vloxseg3ei64(vuint64m2_t * op0, vuint64m2_t * op1, vuint64m2_t * op2, const uint64_t * op3, vuint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_u64m2_m))) +void vloxseg3ei64(vuint64m2_t * op0, vuint64m2_t * op1, vuint64m2_t * op2, vbool32_t op3, vuint64m2_t op4, vuint64m2_t op5, vuint64m2_t op6, const uint64_t * op7, vuint64m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_i64m1))) +void vloxseg4ei64(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, const int64_t * op4, vuint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_i64m1_m))) +void vloxseg4ei64(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vbool64_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, const int64_t * op9, vuint64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_i64m2))) +void vloxseg4ei64(vint64m2_t * op0, vint64m2_t * op1, vint64m2_t * op2, vint64m2_t * op3, const int64_t * op4, vuint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_i64m2_m))) +void vloxseg4ei64(vint64m2_t * op0, vint64m2_t * op1, vint64m2_t * op2, vint64m2_t * op3, vbool32_t op4, vint64m2_t op5, vint64m2_t op6, vint64m2_t op7, vint64m2_t op8, const int64_t * op9, vuint64m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_u64m1))) +void vloxseg4ei64(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, const uint64_t * op4, vuint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_u64m1_m))) +void vloxseg4ei64(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vbool64_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, const uint64_t * op9, vuint64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_u64m2))) +void vloxseg4ei64(vuint64m2_t * op0, vuint64m2_t * op1, vuint64m2_t * op2, vuint64m2_t * op3, const uint64_t * op4, vuint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_u64m2_m))) +void vloxseg4ei64(vuint64m2_t * op0, vuint64m2_t * op1, vuint64m2_t * op2, vuint64m2_t * op3, vbool32_t op4, vuint64m2_t op5, vuint64m2_t op6, vuint64m2_t op7, vuint64m2_t op8, const uint64_t * op9, vuint64m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_i64m1))) +void vloxseg5ei64(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, const int64_t * op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_i64m1_m))) +void vloxseg5ei64(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vbool64_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, vint64m1_t op10, const int64_t * op11, vuint64m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_u64m1))) +void vloxseg5ei64(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, const uint64_t * op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_u64m1_m))) +void vloxseg5ei64(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vbool64_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, vuint64m1_t op10, const uint64_t * op11, vuint64m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_i64m1))) +void vloxseg6ei64(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, const int64_t * op6, vuint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_i64m1_m))) +void vloxseg6ei64(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vbool64_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, vint64m1_t op10, vint64m1_t op11, vint64m1_t op12, const int64_t * op13, vuint64m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_u64m1))) +void vloxseg6ei64(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, const uint64_t * op6, vuint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_u64m1_m))) +void vloxseg6ei64(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vbool64_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, vuint64m1_t op10, vuint64m1_t op11, vuint64m1_t op12, const uint64_t * op13, vuint64m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_i64m1))) +void vloxseg7ei64(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vint64m1_t * op6, const int64_t * op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_i64m1_m))) +void vloxseg7ei64(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vint64m1_t * op6, vbool64_t op7, vint64m1_t op8, vint64m1_t op9, vint64m1_t op10, vint64m1_t op11, vint64m1_t op12, vint64m1_t op13, vint64m1_t op14, const int64_t * op15, vuint64m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_u64m1))) +void vloxseg7ei64(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vuint64m1_t * op6, const uint64_t * op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_u64m1_m))) +void vloxseg7ei64(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vuint64m1_t * op6, vbool64_t op7, vuint64m1_t op8, vuint64m1_t op9, vuint64m1_t op10, vuint64m1_t op11, vuint64m1_t op12, vuint64m1_t op13, vuint64m1_t op14, const uint64_t * op15, vuint64m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_i64m1))) +void vloxseg8ei64(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vint64m1_t * op6, vint64m1_t * op7, const int64_t * op8, vuint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_i64m1_m))) +void vloxseg8ei64(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vint64m1_t * op6, vint64m1_t * op7, vbool64_t op8, vint64m1_t op9, vint64m1_t op10, vint64m1_t op11, vint64m1_t op12, vint64m1_t op13, vint64m1_t op14, vint64m1_t op15, vint64m1_t op16, const int64_t * op17, vuint64m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_u64m1))) +void vloxseg8ei64(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vuint64m1_t * op6, vuint64m1_t * op7, const uint64_t * op8, vuint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_u64m1_m))) +void vloxseg8ei64(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vuint64m1_t * op6, vuint64m1_t * op7, vbool64_t op8, vuint64m1_t op9, vuint64m1_t op10, vuint64m1_t op11, vuint64m1_t op12, vuint64m1_t op13, vuint64m1_t op14, vuint64m1_t op15, vuint64m1_t op16, const uint64_t * op17, vuint64m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e8_v_i8m1))) +void vsseg3e8(int8_t * op0, vint8m1_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e8_v_i8m1_m))) +void vsseg3e8(vbool8_t op0, int8_t * op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e8_v_i8m2))) +void vsseg3e8(int8_t * op0, vint8m2_t op1, vint8m2_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e8_v_i8m2_m))) +void vsseg3e8(vbool4_t op0, int8_t * op1, vint8m2_t op2, vint8m2_t op3, vint8m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e8_v_i8mf2))) +void vsseg3e8(int8_t * op0, vint8mf2_t op1, vint8mf2_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e8_v_i8mf2_m))) +void vsseg3e8(vbool16_t op0, int8_t * op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e8_v_i8mf4))) +void vsseg3e8(int8_t * op0, vint8mf4_t op1, vint8mf4_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e8_v_i8mf4_m))) +void vsseg3e8(vbool32_t op0, int8_t * op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e8_v_i8mf8))) +void vsseg3e8(int8_t * op0, vint8mf8_t op1, vint8mf8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e8_v_i8mf8_m))) +void vsseg3e8(vbool64_t op0, int8_t * op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e8_v_u8m1))) +void vsseg3e8(uint8_t * op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e8_v_u8m1_m))) +void vsseg3e8(vbool8_t op0, uint8_t * op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e8_v_u8m2))) +void vsseg3e8(uint8_t * op0, vuint8m2_t op1, vuint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e8_v_u8m2_m))) +void vsseg3e8(vbool4_t op0, uint8_t * op1, vuint8m2_t op2, vuint8m2_t op3, vuint8m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e8_v_u8mf2))) +void vsseg3e8(uint8_t * op0, vuint8mf2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e8_v_u8mf2_m))) +void vsseg3e8(vbool16_t op0, uint8_t * op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e8_v_u8mf4))) +void vsseg3e8(uint8_t * op0, vuint8mf4_t op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e8_v_u8mf4_m))) +void vsseg3e8(vbool32_t op0, uint8_t * op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e8_v_u8mf8))) +void vsseg3e8(uint8_t * op0, vuint8mf8_t op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e8_v_u8mf8_m))) +void vsseg3e8(vbool64_t op0, uint8_t * op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e8_v_i8m1))) +void vsseg4e8(int8_t * op0, vint8m1_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e8_v_i8m1_m))) +void vsseg4e8(vbool8_t op0, int8_t * op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e8_v_i8m2))) +void vsseg4e8(int8_t * op0, vint8m2_t op1, vint8m2_t op2, vint8m2_t op3, vint8m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e8_v_i8m2_m))) +void vsseg4e8(vbool4_t op0, int8_t * op1, vint8m2_t op2, vint8m2_t op3, vint8m2_t op4, vint8m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e8_v_i8mf2))) +void vsseg4e8(int8_t * op0, vint8mf2_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e8_v_i8mf2_m))) +void vsseg4e8(vbool16_t op0, int8_t * op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e8_v_i8mf4))) +void vsseg4e8(int8_t * op0, vint8mf4_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e8_v_i8mf4_m))) +void vsseg4e8(vbool32_t op0, int8_t * op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e8_v_i8mf8))) +void vsseg4e8(int8_t * op0, vint8mf8_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e8_v_i8mf8_m))) +void vsseg4e8(vbool64_t op0, int8_t * op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e8_v_u8m1))) +void vsseg4e8(uint8_t * op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e8_v_u8m1_m))) +void vsseg4e8(vbool8_t op0, uint8_t * op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e8_v_u8m2))) +void vsseg4e8(uint8_t * op0, vuint8m2_t op1, vuint8m2_t op2, vuint8m2_t op3, vuint8m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e8_v_u8m2_m))) +void vsseg4e8(vbool4_t op0, uint8_t * op1, vuint8m2_t op2, vuint8m2_t op3, vuint8m2_t op4, vuint8m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e8_v_u8mf2))) +void vsseg4e8(uint8_t * op0, vuint8mf2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e8_v_u8mf2_m))) +void vsseg4e8(vbool16_t op0, uint8_t * op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e8_v_u8mf4))) +void vsseg4e8(uint8_t * op0, vuint8mf4_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e8_v_u8mf4_m))) +void vsseg4e8(vbool32_t op0, uint8_t * op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e8_v_u8mf8))) +void vsseg4e8(uint8_t * op0, vuint8mf8_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e8_v_u8mf8_m))) +void vsseg4e8(vbool64_t op0, uint8_t * op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e8_v_i8m1))) +void vsseg5e8(int8_t * op0, vint8m1_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e8_v_i8m1_m))) +void vsseg5e8(vbool8_t op0, int8_t * op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e8_v_i8mf2))) +void vsseg5e8(int8_t * op0, vint8mf2_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e8_v_i8mf2_m))) +void vsseg5e8(vbool16_t op0, int8_t * op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e8_v_i8mf4))) +void vsseg5e8(int8_t * op0, vint8mf4_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e8_v_i8mf4_m))) +void vsseg5e8(vbool32_t op0, int8_t * op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e8_v_i8mf8))) +void vsseg5e8(int8_t * op0, vint8mf8_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e8_v_i8mf8_m))) +void vsseg5e8(vbool64_t op0, int8_t * op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e8_v_u8m1))) +void vsseg5e8(uint8_t * op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e8_v_u8m1_m))) +void vsseg5e8(vbool8_t op0, uint8_t * op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e8_v_u8mf2))) +void vsseg5e8(uint8_t * op0, vuint8mf2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e8_v_u8mf2_m))) +void vsseg5e8(vbool16_t op0, uint8_t * op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e8_v_u8mf4))) +void vsseg5e8(uint8_t * op0, vuint8mf4_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e8_v_u8mf4_m))) +void vsseg5e8(vbool32_t op0, uint8_t * op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e8_v_u8mf8))) +void vsseg5e8(uint8_t * op0, vuint8mf8_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e8_v_u8mf8_m))) +void vsseg5e8(vbool64_t op0, uint8_t * op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e8_v_i8m1))) +void vsseg6e8(int8_t * op0, vint8m1_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e8_v_i8m1_m))) +void vsseg6e8(vbool8_t op0, int8_t * op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e8_v_i8mf2))) +void vsseg6e8(int8_t * op0, vint8mf2_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e8_v_i8mf2_m))) +void vsseg6e8(vbool16_t op0, int8_t * op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e8_v_i8mf4))) +void vsseg6e8(int8_t * op0, vint8mf4_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e8_v_i8mf4_m))) +void vsseg6e8(vbool32_t op0, int8_t * op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e8_v_i8mf8))) +void vsseg6e8(int8_t * op0, vint8mf8_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e8_v_i8mf8_m))) +void vsseg6e8(vbool64_t op0, int8_t * op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e8_v_u8m1))) +void vsseg6e8(uint8_t * op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e8_v_u8m1_m))) +void vsseg6e8(vbool8_t op0, uint8_t * op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e8_v_u8mf2))) +void vsseg6e8(uint8_t * op0, vuint8mf2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e8_v_u8mf2_m))) +void vsseg6e8(vbool16_t op0, uint8_t * op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e8_v_u8mf4))) +void vsseg6e8(uint8_t * op0, vuint8mf4_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e8_v_u8mf4_m))) +void vsseg6e8(vbool32_t op0, uint8_t * op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e8_v_u8mf8))) +void vsseg6e8(uint8_t * op0, vuint8mf8_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e8_v_u8mf8_m))) +void vsseg6e8(vbool64_t op0, uint8_t * op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e8_v_i8m1))) +void vsseg7e8(int8_t * op0, vint8m1_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e8_v_i8m1_m))) +void vsseg7e8(vbool8_t op0, int8_t * op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e8_v_i8mf2))) +void vsseg7e8(int8_t * op0, vint8mf2_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e8_v_i8mf2_m))) +void vsseg7e8(vbool16_t op0, int8_t * op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e8_v_i8mf4))) +void vsseg7e8(int8_t * op0, vint8mf4_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e8_v_i8mf4_m))) +void vsseg7e8(vbool32_t op0, int8_t * op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e8_v_i8mf8))) +void vsseg7e8(int8_t * op0, vint8mf8_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e8_v_i8mf8_m))) +void vsseg7e8(vbool64_t op0, int8_t * op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e8_v_u8m1))) +void vsseg7e8(uint8_t * op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e8_v_u8m1_m))) +void vsseg7e8(vbool8_t op0, uint8_t * op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e8_v_u8mf2))) +void vsseg7e8(uint8_t * op0, vuint8mf2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e8_v_u8mf2_m))) +void vsseg7e8(vbool16_t op0, uint8_t * op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e8_v_u8mf4))) +void vsseg7e8(uint8_t * op0, vuint8mf4_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e8_v_u8mf4_m))) +void vsseg7e8(vbool32_t op0, uint8_t * op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e8_v_u8mf8))) +void vsseg7e8(uint8_t * op0, vuint8mf8_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e8_v_u8mf8_m))) +void vsseg7e8(vbool64_t op0, uint8_t * op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e8_v_i8m1))) +void vsseg8e8(int8_t * op0, vint8m1_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e8_v_i8m1_m))) +void vsseg8e8(vbool8_t op0, int8_t * op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e8_v_i8mf2))) +void vsseg8e8(int8_t * op0, vint8mf2_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e8_v_i8mf2_m))) +void vsseg8e8(vbool16_t op0, int8_t * op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e8_v_i8mf4))) +void vsseg8e8(int8_t * op0, vint8mf4_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e8_v_i8mf4_m))) +void vsseg8e8(vbool32_t op0, int8_t * op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e8_v_i8mf8))) +void vsseg8e8(int8_t * op0, vint8mf8_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e8_v_i8mf8_m))) +void vsseg8e8(vbool64_t op0, int8_t * op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e8_v_u8m1))) +void vsseg8e8(uint8_t * op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e8_v_u8m1_m))) +void vsseg8e8(vbool8_t op0, uint8_t * op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e8_v_u8mf2))) +void vsseg8e8(uint8_t * op0, vuint8mf2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e8_v_u8mf2_m))) +void vsseg8e8(vbool16_t op0, uint8_t * op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e8_v_u8mf4))) +void vsseg8e8(uint8_t * op0, vuint8mf4_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e8_v_u8mf4_m))) +void vsseg8e8(vbool32_t op0, uint8_t * op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e8_v_u8mf8))) +void vsseg8e8(uint8_t * op0, vuint8mf8_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e8_v_u8mf8_m))) +void vsseg8e8(vbool64_t op0, uint8_t * op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e16_v_i16m1))) +void vsseg2e16(int16_t * op0, vint16m1_t op1, vint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e16_v_i16m1_m))) +void vsseg2e16(vbool16_t op0, int16_t * op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e16_v_i16m2))) +void vsseg2e16(int16_t * op0, vint16m2_t op1, vint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e16_v_i16m2_m))) +void vsseg2e16(vbool8_t op0, int16_t * op1, vint16m2_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e16_v_i16m4))) +void vsseg2e16(int16_t * op0, vint16m4_t op1, vint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e16_v_i16m4_m))) +void vsseg2e16(vbool4_t op0, int16_t * op1, vint16m4_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e16_v_i16mf2))) +void vsseg2e16(int16_t * op0, vint16mf2_t op1, vint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e16_v_i16mf2_m))) +void vsseg2e16(vbool32_t op0, int16_t * op1, vint16mf2_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e16_v_i16mf4))) +void vsseg2e16(int16_t * op0, vint16mf4_t op1, vint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e16_v_i16mf4_m))) +void vsseg2e16(vbool64_t op0, int16_t * op1, vint16mf4_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e16_v_u16m1))) +void vsseg2e16(uint16_t * op0, vuint16m1_t op1, vuint16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e16_v_u16m1_m))) +void vsseg2e16(vbool16_t op0, uint16_t * op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e16_v_u16m2))) +void vsseg2e16(uint16_t * op0, vuint16m2_t op1, vuint16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e16_v_u16m2_m))) +void vsseg2e16(vbool8_t op0, uint16_t * op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e16_v_u16m4))) +void vsseg2e16(uint16_t * op0, vuint16m4_t op1, vuint16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e16_v_u16m4_m))) +void vsseg2e16(vbool4_t op0, uint16_t * op1, vuint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e16_v_u16mf2))) +void vsseg2e16(uint16_t * op0, vuint16mf2_t op1, vuint16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e16_v_u16mf2_m))) +void vsseg2e16(vbool32_t op0, uint16_t * op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e16_v_u16mf4))) +void vsseg2e16(uint16_t * op0, vuint16mf4_t op1, vuint16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e16_v_u16mf4_m))) +void vsseg2e16(vbool64_t op0, uint16_t * op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e16_v_i16m1))) +void vsseg3e16(int16_t * op0, vint16m1_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e16_v_i16m1_m))) +void vsseg3e16(vbool16_t op0, int16_t * op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e16_v_i16m2))) +void vsseg3e16(int16_t * op0, vint16m2_t op1, vint16m2_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e16_v_i16m2_m))) +void vsseg3e16(vbool8_t op0, int16_t * op1, vint16m2_t op2, vint16m2_t op3, vint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e16_v_i16mf2))) +void vsseg3e16(int16_t * op0, vint16mf2_t op1, vint16mf2_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e16_v_i16mf2_m))) +void vsseg3e16(vbool32_t op0, int16_t * op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e16_v_i16mf4))) +void vsseg3e16(int16_t * op0, vint16mf4_t op1, vint16mf4_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e16_v_i16mf4_m))) +void vsseg3e16(vbool64_t op0, int16_t * op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e16_v_u16m1))) +void vsseg3e16(uint16_t * op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e16_v_u16m1_m))) +void vsseg3e16(vbool16_t op0, uint16_t * op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e16_v_u16m2))) +void vsseg3e16(uint16_t * op0, vuint16m2_t op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e16_v_u16m2_m))) +void vsseg3e16(vbool8_t op0, uint16_t * op1, vuint16m2_t op2, vuint16m2_t op3, vuint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e16_v_u16mf2))) +void vsseg3e16(uint16_t * op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e16_v_u16mf2_m))) +void vsseg3e16(vbool32_t op0, uint16_t * op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e16_v_u16mf4))) +void vsseg3e16(uint16_t * op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e16_v_u16mf4_m))) +void vsseg3e16(vbool64_t op0, uint16_t * op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e16_v_i16m1))) +void vsseg4e16(int16_t * op0, vint16m1_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e16_v_i16m1_m))) +void vsseg4e16(vbool16_t op0, int16_t * op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e16_v_i16m2))) +void vsseg4e16(int16_t * op0, vint16m2_t op1, vint16m2_t op2, vint16m2_t op3, vint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e16_v_i16m2_m))) +void vsseg4e16(vbool8_t op0, int16_t * op1, vint16m2_t op2, vint16m2_t op3, vint16m2_t op4, vint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e16_v_i16mf2))) +void vsseg4e16(int16_t * op0, vint16mf2_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e16_v_i16mf2_m))) +void vsseg4e16(vbool32_t op0, int16_t * op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e16_v_i16mf4))) +void vsseg4e16(int16_t * op0, vint16mf4_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e16_v_i16mf4_m))) +void vsseg4e16(vbool64_t op0, int16_t * op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e16_v_u16m1))) +void vsseg4e16(uint16_t * op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e16_v_u16m1_m))) +void vsseg4e16(vbool16_t op0, uint16_t * op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e16_v_u16m2))) +void vsseg4e16(uint16_t * op0, vuint16m2_t op1, vuint16m2_t op2, vuint16m2_t op3, vuint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e16_v_u16m2_m))) +void vsseg4e16(vbool8_t op0, uint16_t * op1, vuint16m2_t op2, vuint16m2_t op3, vuint16m2_t op4, vuint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e16_v_u16mf2))) +void vsseg4e16(uint16_t * op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e16_v_u16mf2_m))) +void vsseg4e16(vbool32_t op0, uint16_t * op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e16_v_u16mf4))) +void vsseg4e16(uint16_t * op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e16_v_u16mf4_m))) +void vsseg4e16(vbool64_t op0, uint16_t * op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e16_v_i16m1))) +void vsseg5e16(int16_t * op0, vint16m1_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e16_v_i16m1_m))) +void vsseg5e16(vbool16_t op0, int16_t * op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e16_v_i16mf2))) +void vsseg5e16(int16_t * op0, vint16mf2_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e16_v_i16mf2_m))) +void vsseg5e16(vbool32_t op0, int16_t * op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e16_v_i16mf4))) +void vsseg5e16(int16_t * op0, vint16mf4_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e16_v_i16mf4_m))) +void vsseg5e16(vbool64_t op0, int16_t * op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e16_v_u16m1))) +void vsseg5e16(uint16_t * op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e16_v_u16m1_m))) +void vsseg5e16(vbool16_t op0, uint16_t * op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e16_v_u16mf2))) +void vsseg5e16(uint16_t * op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e16_v_u16mf2_m))) +void vsseg5e16(vbool32_t op0, uint16_t * op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e16_v_u16mf4))) +void vsseg5e16(uint16_t * op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e16_v_u16mf4_m))) +void vsseg5e16(vbool64_t op0, uint16_t * op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e16_v_i16m1))) +void vsseg6e16(int16_t * op0, vint16m1_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e16_v_i16m1_m))) +void vsseg6e16(vbool16_t op0, int16_t * op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e16_v_i16mf2))) +void vsseg6e16(int16_t * op0, vint16mf2_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e16_v_i16mf2_m))) +void vsseg6e16(vbool32_t op0, int16_t * op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e16_v_i16mf4))) +void vsseg6e16(int16_t * op0, vint16mf4_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e16_v_i16mf4_m))) +void vsseg6e16(vbool64_t op0, int16_t * op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e16_v_u16m1))) +void vsseg6e16(uint16_t * op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e16_v_u16m1_m))) +void vsseg6e16(vbool16_t op0, uint16_t * op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e16_v_u16mf2))) +void vsseg6e16(uint16_t * op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e16_v_u16mf2_m))) +void vsseg6e16(vbool32_t op0, uint16_t * op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e16_v_u16mf4))) +void vsseg6e16(uint16_t * op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e16_v_u16mf4_m))) +void vsseg6e16(vbool64_t op0, uint16_t * op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e16_v_i16m1))) +void vsseg7e16(int16_t * op0, vint16m1_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e16_v_i16m1_m))) +void vsseg7e16(vbool16_t op0, int16_t * op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e16_v_i16mf2))) +void vsseg7e16(int16_t * op0, vint16mf2_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e16_v_i16mf2_m))) +void vsseg7e16(vbool32_t op0, int16_t * op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e16_v_i16mf4))) +void vsseg7e16(int16_t * op0, vint16mf4_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e16_v_i16mf4_m))) +void vsseg7e16(vbool64_t op0, int16_t * op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e16_v_u16m1))) +void vsseg7e16(uint16_t * op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e16_v_u16m1_m))) +void vsseg7e16(vbool16_t op0, uint16_t * op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e16_v_u16mf2))) +void vsseg7e16(uint16_t * op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e16_v_u16mf2_m))) +void vsseg7e16(vbool32_t op0, uint16_t * op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e16_v_u16mf4))) +void vsseg7e16(uint16_t * op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e16_v_u16mf4_m))) +void vsseg7e16(vbool64_t op0, uint16_t * op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e16_v_i16m1))) +void vsseg8e16(int16_t * op0, vint16m1_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e16_v_i16m1_m))) +void vsseg8e16(vbool16_t op0, int16_t * op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e16_v_i16mf2))) +void vsseg8e16(int16_t * op0, vint16mf2_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e16_v_i16mf2_m))) +void vsseg8e16(vbool32_t op0, int16_t * op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e16_v_i16mf4))) +void vsseg8e16(int16_t * op0, vint16mf4_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e16_v_i16mf4_m))) +void vsseg8e16(vbool64_t op0, int16_t * op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e16_v_u16m1))) +void vsseg8e16(uint16_t * op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e16_v_u16m1_m))) +void vsseg8e16(vbool16_t op0, uint16_t * op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e16_v_u16mf2))) +void vsseg8e16(uint16_t * op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e16_v_u16mf2_m))) +void vsseg8e16(vbool32_t op0, uint16_t * op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e16_v_u16mf4))) +void vsseg8e16(uint16_t * op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e16_v_u16mf4_m))) +void vsseg8e16(vbool64_t op0, uint16_t * op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e32_v_i32m1))) +void vsseg2e32(int32_t * op0, vint32m1_t op1, vint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e32_v_i32m1_m))) +void vsseg2e32(vbool32_t op0, int32_t * op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e32_v_i32m2))) +void vsseg2e32(int32_t * op0, vint32m2_t op1, vint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e32_v_i32m2_m))) +void vsseg2e32(vbool16_t op0, int32_t * op1, vint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e32_v_i32m4))) +void vsseg2e32(int32_t * op0, vint32m4_t op1, vint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e32_v_i32m4_m))) +void vsseg2e32(vbool8_t op0, int32_t * op1, vint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e32_v_i32mf2))) +void vsseg2e32(int32_t * op0, vint32mf2_t op1, vint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e32_v_i32mf2_m))) +void vsseg2e32(vbool64_t op0, int32_t * op1, vint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e32_v_u32m1))) +void vsseg2e32(uint32_t * op0, vuint32m1_t op1, vuint32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e32_v_u32m1_m))) +void vsseg2e32(vbool32_t op0, uint32_t * op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e32_v_u32m2))) +void vsseg2e32(uint32_t * op0, vuint32m2_t op1, vuint32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e32_v_u32m2_m))) +void vsseg2e32(vbool16_t op0, uint32_t * op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e32_v_u32m4))) +void vsseg2e32(uint32_t * op0, vuint32m4_t op1, vuint32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e32_v_u32m4_m))) +void vsseg2e32(vbool8_t op0, uint32_t * op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e32_v_u32mf2))) +void vsseg2e32(uint32_t * op0, vuint32mf2_t op1, vuint32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e32_v_u32mf2_m))) +void vsseg2e32(vbool64_t op0, uint32_t * op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e32_v_i32m1))) +void vsseg3e32(int32_t * op0, vint32m1_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e32_v_i32m1_m))) +void vsseg3e32(vbool32_t op0, int32_t * op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e32_v_i32m2))) +void vsseg3e32(int32_t * op0, vint32m2_t op1, vint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e32_v_i32m2_m))) +void vsseg3e32(vbool16_t op0, int32_t * op1, vint32m2_t op2, vint32m2_t op3, vint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e32_v_i32mf2))) +void vsseg3e32(int32_t * op0, vint32mf2_t op1, vint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e32_v_i32mf2_m))) +void vsseg3e32(vbool64_t op0, int32_t * op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e32_v_u32m1))) +void vsseg3e32(uint32_t * op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e32_v_u32m1_m))) +void vsseg3e32(vbool32_t op0, uint32_t * op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e32_v_u32m2))) +void vsseg3e32(uint32_t * op0, vuint32m2_t op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e32_v_u32m2_m))) +void vsseg3e32(vbool16_t op0, uint32_t * op1, vuint32m2_t op2, vuint32m2_t op3, vuint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e32_v_u32mf2))) +void vsseg3e32(uint32_t * op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e32_v_u32mf2_m))) +void vsseg3e32(vbool64_t op0, uint32_t * op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e32_v_i32m1))) +void vsseg4e32(int32_t * op0, vint32m1_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e32_v_i32m1_m))) +void vsseg4e32(vbool32_t op0, int32_t * op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e32_v_i32m2))) +void vsseg4e32(int32_t * op0, vint32m2_t op1, vint32m2_t op2, vint32m2_t op3, vint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e32_v_i32m2_m))) +void vsseg4e32(vbool16_t op0, int32_t * op1, vint32m2_t op2, vint32m2_t op3, vint32m2_t op4, vint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e32_v_i32mf2))) +void vsseg4e32(int32_t * op0, vint32mf2_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e32_v_i32mf2_m))) +void vsseg4e32(vbool64_t op0, int32_t * op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e32_v_u32m1))) +void vsseg4e32(uint32_t * op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e32_v_u32m1_m))) +void vsseg4e32(vbool32_t op0, uint32_t * op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e32_v_u32m2))) +void vsseg4e32(uint32_t * op0, vuint32m2_t op1, vuint32m2_t op2, vuint32m2_t op3, vuint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e32_v_u32m2_m))) +void vsseg4e32(vbool16_t op0, uint32_t * op1, vuint32m2_t op2, vuint32m2_t op3, vuint32m2_t op4, vuint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e32_v_u32mf2))) +void vsseg4e32(uint32_t * op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e32_v_u32mf2_m))) +void vsseg4e32(vbool64_t op0, uint32_t * op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e32_v_i32m1))) +void vsseg5e32(int32_t * op0, vint32m1_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e32_v_i32m1_m))) +void vsseg5e32(vbool32_t op0, int32_t * op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e32_v_i32mf2))) +void vsseg5e32(int32_t * op0, vint32mf2_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e32_v_i32mf2_m))) +void vsseg5e32(vbool64_t op0, int32_t * op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e32_v_u32m1))) +void vsseg5e32(uint32_t * op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e32_v_u32m1_m))) +void vsseg5e32(vbool32_t op0, uint32_t * op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e32_v_u32mf2))) +void vsseg5e32(uint32_t * op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e32_v_u32mf2_m))) +void vsseg5e32(vbool64_t op0, uint32_t * op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e32_v_i32m1))) +void vsseg6e32(int32_t * op0, vint32m1_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e32_v_i32m1_m))) +void vsseg6e32(vbool32_t op0, int32_t * op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e32_v_i32mf2))) +void vsseg6e32(int32_t * op0, vint32mf2_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e32_v_i32mf2_m))) +void vsseg6e32(vbool64_t op0, int32_t * op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e32_v_u32m1))) +void vsseg6e32(uint32_t * op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e32_v_u32m1_m))) +void vsseg6e32(vbool32_t op0, uint32_t * op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e32_v_u32mf2))) +void vsseg6e32(uint32_t * op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e32_v_u32mf2_m))) +void vsseg6e32(vbool64_t op0, uint32_t * op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e32_v_i32m1))) +void vsseg7e32(int32_t * op0, vint32m1_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e32_v_i32m1_m))) +void vsseg7e32(vbool32_t op0, int32_t * op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e32_v_i32mf2))) +void vsseg7e32(int32_t * op0, vint32mf2_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e32_v_i32mf2_m))) +void vsseg7e32(vbool64_t op0, int32_t * op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e32_v_u32m1))) +void vsseg7e32(uint32_t * op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e32_v_u32m1_m))) +void vsseg7e32(vbool32_t op0, uint32_t * op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e32_v_u32mf2))) +void vsseg7e32(uint32_t * op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e32_v_u32mf2_m))) +void vsseg7e32(vbool64_t op0, uint32_t * op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e32_v_i32m1))) +void vsseg8e32(int32_t * op0, vint32m1_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e32_v_i32m1_m))) +void vsseg8e32(vbool32_t op0, int32_t * op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e32_v_i32mf2))) +void vsseg8e32(int32_t * op0, vint32mf2_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e32_v_i32mf2_m))) +void vsseg8e32(vbool64_t op0, int32_t * op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e32_v_u32m1))) +void vsseg8e32(uint32_t * op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e32_v_u32m1_m))) +void vsseg8e32(vbool32_t op0, uint32_t * op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e32_v_u32mf2))) +void vsseg8e32(uint32_t * op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e32_v_u32mf2_m))) +void vsseg8e32(vbool64_t op0, uint32_t * op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e64_v_i64m1))) +void vsseg2e64(int64_t * op0, vint64m1_t op1, vint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e64_v_i64m1_m))) +void vsseg2e64(vbool64_t op0, int64_t * op1, vint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e64_v_i64m2))) +void vsseg2e64(int64_t * op0, vint64m2_t op1, vint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e64_v_i64m2_m))) +void vsseg2e64(vbool32_t op0, int64_t * op1, vint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e64_v_i64m4))) +void vsseg2e64(int64_t * op0, vint64m4_t op1, vint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e64_v_i64m4_m))) +void vsseg2e64(vbool16_t op0, int64_t * op1, vint64m4_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e64_v_u64m1))) +void vsseg2e64(uint64_t * op0, vuint64m1_t op1, vuint64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e64_v_u64m1_m))) +void vsseg2e64(vbool64_t op0, uint64_t * op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e64_v_u64m2))) +void vsseg2e64(uint64_t * op0, vuint64m2_t op1, vuint64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e64_v_u64m2_m))) +void vsseg2e64(vbool32_t op0, uint64_t * op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e64_v_u64m4))) +void vsseg2e64(uint64_t * op0, vuint64m4_t op1, vuint64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e64_v_u64m4_m))) +void vsseg2e64(vbool16_t op0, uint64_t * op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e64_v_i64m1))) +void vsseg3e64(int64_t * op0, vint64m1_t op1, vint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e64_v_i64m1_m))) +void vsseg3e64(vbool64_t op0, int64_t * op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e64_v_i64m2))) +void vsseg3e64(int64_t * op0, vint64m2_t op1, vint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e64_v_i64m2_m))) +void vsseg3e64(vbool32_t op0, int64_t * op1, vint64m2_t op2, vint64m2_t op3, vint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e64_v_u64m1))) +void vsseg3e64(uint64_t * op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e64_v_u64m1_m))) +void vsseg3e64(vbool64_t op0, uint64_t * op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e64_v_u64m2))) +void vsseg3e64(uint64_t * op0, vuint64m2_t op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e64_v_u64m2_m))) +void vsseg3e64(vbool32_t op0, uint64_t * op1, vuint64m2_t op2, vuint64m2_t op3, vuint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e64_v_i64m1))) +void vsseg4e64(int64_t * op0, vint64m1_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e64_v_i64m1_m))) +void vsseg4e64(vbool64_t op0, int64_t * op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e64_v_i64m2))) +void vsseg4e64(int64_t * op0, vint64m2_t op1, vint64m2_t op2, vint64m2_t op3, vint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e64_v_i64m2_m))) +void vsseg4e64(vbool32_t op0, int64_t * op1, vint64m2_t op2, vint64m2_t op3, vint64m2_t op4, vint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e64_v_u64m1))) +void vsseg4e64(uint64_t * op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e64_v_u64m1_m))) +void vsseg4e64(vbool64_t op0, uint64_t * op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e64_v_u64m2))) +void vsseg4e64(uint64_t * op0, vuint64m2_t op1, vuint64m2_t op2, vuint64m2_t op3, vuint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e64_v_u64m2_m))) +void vsseg4e64(vbool32_t op0, uint64_t * op1, vuint64m2_t op2, vuint64m2_t op3, vuint64m2_t op4, vuint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e64_v_i64m1))) +void vsseg5e64(int64_t * op0, vint64m1_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e64_v_i64m1_m))) +void vsseg5e64(vbool64_t op0, int64_t * op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e64_v_u64m1))) +void vsseg5e64(uint64_t * op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e64_v_u64m1_m))) +void vsseg5e64(vbool64_t op0, uint64_t * op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e64_v_i64m1))) +void vsseg6e64(int64_t * op0, vint64m1_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e64_v_i64m1_m))) +void vsseg6e64(vbool64_t op0, int64_t * op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e64_v_u64m1))) +void vsseg6e64(uint64_t * op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e64_v_u64m1_m))) +void vsseg6e64(vbool64_t op0, uint64_t * op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e64_v_i64m1))) +void vsseg7e64(int64_t * op0, vint64m1_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e64_v_i64m1_m))) +void vsseg7e64(vbool64_t op0, int64_t * op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e64_v_u64m1))) +void vsseg7e64(uint64_t * op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e64_v_u64m1_m))) +void vsseg7e64(vbool64_t op0, uint64_t * op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e64_v_i64m1))) +void vsseg8e64(int64_t * op0, vint64m1_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e64_v_i64m1_m))) +void vsseg8e64(vbool64_t op0, int64_t * op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e64_v_u64m1))) +void vsseg8e64(uint64_t * op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e64_v_u64m1_m))) +void vsseg8e64(vbool64_t op0, uint64_t * op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e8_v_i8m1))) +void vssseg3e8(int8_t * op0, ptrdiff_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e8_v_i8m1_m))) +void vssseg3e8(vbool8_t op0, int8_t * op1, ptrdiff_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e8_v_i8m2))) +void vssseg3e8(int8_t * op0, ptrdiff_t op1, vint8m2_t op2, vint8m2_t op3, vint8m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e8_v_i8m2_m))) +void vssseg3e8(vbool4_t op0, int8_t * op1, ptrdiff_t op2, vint8m2_t op3, vint8m2_t op4, vint8m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e8_v_i8mf2))) +void vssseg3e8(int8_t * op0, ptrdiff_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e8_v_i8mf2_m))) +void vssseg3e8(vbool16_t op0, int8_t * op1, ptrdiff_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e8_v_i8mf4))) +void vssseg3e8(int8_t * op0, ptrdiff_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e8_v_i8mf4_m))) +void vssseg3e8(vbool32_t op0, int8_t * op1, ptrdiff_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e8_v_i8mf8))) +void vssseg3e8(int8_t * op0, ptrdiff_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e8_v_i8mf8_m))) +void vssseg3e8(vbool64_t op0, int8_t * op1, ptrdiff_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e8_v_u8m1))) +void vssseg3e8(uint8_t * op0, ptrdiff_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e8_v_u8m1_m))) +void vssseg3e8(vbool8_t op0, uint8_t * op1, ptrdiff_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e8_v_u8m2))) +void vssseg3e8(uint8_t * op0, ptrdiff_t op1, vuint8m2_t op2, vuint8m2_t op3, vuint8m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e8_v_u8m2_m))) +void vssseg3e8(vbool4_t op0, uint8_t * op1, ptrdiff_t op2, vuint8m2_t op3, vuint8m2_t op4, vuint8m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e8_v_u8mf2))) +void vssseg3e8(uint8_t * op0, ptrdiff_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e8_v_u8mf2_m))) +void vssseg3e8(vbool16_t op0, uint8_t * op1, ptrdiff_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e8_v_u8mf4))) +void vssseg3e8(uint8_t * op0, ptrdiff_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e8_v_u8mf4_m))) +void vssseg3e8(vbool32_t op0, uint8_t * op1, ptrdiff_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e8_v_u8mf8))) +void vssseg3e8(uint8_t * op0, ptrdiff_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e8_v_u8mf8_m))) +void vssseg3e8(vbool64_t op0, uint8_t * op1, ptrdiff_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e8_v_i8m1))) +void vssseg4e8(int8_t * op0, ptrdiff_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e8_v_i8m1_m))) +void vssseg4e8(vbool8_t op0, int8_t * op1, ptrdiff_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e8_v_i8m2))) +void vssseg4e8(int8_t * op0, ptrdiff_t op1, vint8m2_t op2, vint8m2_t op3, vint8m2_t op4, vint8m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e8_v_i8m2_m))) +void vssseg4e8(vbool4_t op0, int8_t * op1, ptrdiff_t op2, vint8m2_t op3, vint8m2_t op4, vint8m2_t op5, vint8m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e8_v_i8mf2))) +void vssseg4e8(int8_t * op0, ptrdiff_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e8_v_i8mf2_m))) +void vssseg4e8(vbool16_t op0, int8_t * op1, ptrdiff_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e8_v_i8mf4))) +void vssseg4e8(int8_t * op0, ptrdiff_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e8_v_i8mf4_m))) +void vssseg4e8(vbool32_t op0, int8_t * op1, ptrdiff_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e8_v_i8mf8))) +void vssseg4e8(int8_t * op0, ptrdiff_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e8_v_i8mf8_m))) +void vssseg4e8(vbool64_t op0, int8_t * op1, ptrdiff_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e8_v_u8m1))) +void vssseg4e8(uint8_t * op0, ptrdiff_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e8_v_u8m1_m))) +void vssseg4e8(vbool8_t op0, uint8_t * op1, ptrdiff_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e8_v_u8m2))) +void vssseg4e8(uint8_t * op0, ptrdiff_t op1, vuint8m2_t op2, vuint8m2_t op3, vuint8m2_t op4, vuint8m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e8_v_u8m2_m))) +void vssseg4e8(vbool4_t op0, uint8_t * op1, ptrdiff_t op2, vuint8m2_t op3, vuint8m2_t op4, vuint8m2_t op5, vuint8m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e8_v_u8mf2))) +void vssseg4e8(uint8_t * op0, ptrdiff_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e8_v_u8mf2_m))) +void vssseg4e8(vbool16_t op0, uint8_t * op1, ptrdiff_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e8_v_u8mf4))) +void vssseg4e8(uint8_t * op0, ptrdiff_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e8_v_u8mf4_m))) +void vssseg4e8(vbool32_t op0, uint8_t * op1, ptrdiff_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e8_v_u8mf8))) +void vssseg4e8(uint8_t * op0, ptrdiff_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e8_v_u8mf8_m))) +void vssseg4e8(vbool64_t op0, uint8_t * op1, ptrdiff_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e8_v_i8m1))) +void vssseg5e8(int8_t * op0, ptrdiff_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e8_v_i8m1_m))) +void vssseg5e8(vbool8_t op0, int8_t * op1, ptrdiff_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e8_v_i8mf2))) +void vssseg5e8(int8_t * op0, ptrdiff_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e8_v_i8mf2_m))) +void vssseg5e8(vbool16_t op0, int8_t * op1, ptrdiff_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e8_v_i8mf4))) +void vssseg5e8(int8_t * op0, ptrdiff_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e8_v_i8mf4_m))) +void vssseg5e8(vbool32_t op0, int8_t * op1, ptrdiff_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e8_v_i8mf8))) +void vssseg5e8(int8_t * op0, ptrdiff_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e8_v_i8mf8_m))) +void vssseg5e8(vbool64_t op0, int8_t * op1, ptrdiff_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e8_v_u8m1))) +void vssseg5e8(uint8_t * op0, ptrdiff_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e8_v_u8m1_m))) +void vssseg5e8(vbool8_t op0, uint8_t * op1, ptrdiff_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e8_v_u8mf2))) +void vssseg5e8(uint8_t * op0, ptrdiff_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e8_v_u8mf2_m))) +void vssseg5e8(vbool16_t op0, uint8_t * op1, ptrdiff_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e8_v_u8mf4))) +void vssseg5e8(uint8_t * op0, ptrdiff_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e8_v_u8mf4_m))) +void vssseg5e8(vbool32_t op0, uint8_t * op1, ptrdiff_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e8_v_u8mf8))) +void vssseg5e8(uint8_t * op0, ptrdiff_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e8_v_u8mf8_m))) +void vssseg5e8(vbool64_t op0, uint8_t * op1, ptrdiff_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e8_v_i8m1))) +void vssseg6e8(int8_t * op0, ptrdiff_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e8_v_i8m1_m))) +void vssseg6e8(vbool8_t op0, int8_t * op1, ptrdiff_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e8_v_i8mf2))) +void vssseg6e8(int8_t * op0, ptrdiff_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e8_v_i8mf2_m))) +void vssseg6e8(vbool16_t op0, int8_t * op1, ptrdiff_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e8_v_i8mf4))) +void vssseg6e8(int8_t * op0, ptrdiff_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e8_v_i8mf4_m))) +void vssseg6e8(vbool32_t op0, int8_t * op1, ptrdiff_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e8_v_i8mf8))) +void vssseg6e8(int8_t * op0, ptrdiff_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e8_v_i8mf8_m))) +void vssseg6e8(vbool64_t op0, int8_t * op1, ptrdiff_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e8_v_u8m1))) +void vssseg6e8(uint8_t * op0, ptrdiff_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e8_v_u8m1_m))) +void vssseg6e8(vbool8_t op0, uint8_t * op1, ptrdiff_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e8_v_u8mf2))) +void vssseg6e8(uint8_t * op0, ptrdiff_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e8_v_u8mf2_m))) +void vssseg6e8(vbool16_t op0, uint8_t * op1, ptrdiff_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e8_v_u8mf4))) +void vssseg6e8(uint8_t * op0, ptrdiff_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e8_v_u8mf4_m))) +void vssseg6e8(vbool32_t op0, uint8_t * op1, ptrdiff_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e8_v_u8mf8))) +void vssseg6e8(uint8_t * op0, ptrdiff_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e8_v_u8mf8_m))) +void vssseg6e8(vbool64_t op0, uint8_t * op1, ptrdiff_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e8_v_i8m1))) +void vssseg7e8(int8_t * op0, ptrdiff_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e8_v_i8m1_m))) +void vssseg7e8(vbool8_t op0, int8_t * op1, ptrdiff_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e8_v_i8mf2))) +void vssseg7e8(int8_t * op0, ptrdiff_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e8_v_i8mf2_m))) +void vssseg7e8(vbool16_t op0, int8_t * op1, ptrdiff_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e8_v_i8mf4))) +void vssseg7e8(int8_t * op0, ptrdiff_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e8_v_i8mf4_m))) +void vssseg7e8(vbool32_t op0, int8_t * op1, ptrdiff_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e8_v_i8mf8))) +void vssseg7e8(int8_t * op0, ptrdiff_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e8_v_i8mf8_m))) +void vssseg7e8(vbool64_t op0, int8_t * op1, ptrdiff_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e8_v_u8m1))) +void vssseg7e8(uint8_t * op0, ptrdiff_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e8_v_u8m1_m))) +void vssseg7e8(vbool8_t op0, uint8_t * op1, ptrdiff_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e8_v_u8mf2))) +void vssseg7e8(uint8_t * op0, ptrdiff_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e8_v_u8mf2_m))) +void vssseg7e8(vbool16_t op0, uint8_t * op1, ptrdiff_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e8_v_u8mf4))) +void vssseg7e8(uint8_t * op0, ptrdiff_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e8_v_u8mf4_m))) +void vssseg7e8(vbool32_t op0, uint8_t * op1, ptrdiff_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e8_v_u8mf8))) +void vssseg7e8(uint8_t * op0, ptrdiff_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e8_v_u8mf8_m))) +void vssseg7e8(vbool64_t op0, uint8_t * op1, ptrdiff_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e8_v_i8m1))) +void vssseg8e8(int8_t * op0, ptrdiff_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e8_v_i8m1_m))) +void vssseg8e8(vbool8_t op0, int8_t * op1, ptrdiff_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, vint8m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e8_v_i8mf2))) +void vssseg8e8(int8_t * op0, ptrdiff_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e8_v_i8mf2_m))) +void vssseg8e8(vbool16_t op0, int8_t * op1, ptrdiff_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, vint8mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e8_v_i8mf4))) +void vssseg8e8(int8_t * op0, ptrdiff_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e8_v_i8mf4_m))) +void vssseg8e8(vbool32_t op0, int8_t * op1, ptrdiff_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, vint8mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e8_v_i8mf8))) +void vssseg8e8(int8_t * op0, ptrdiff_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e8_v_i8mf8_m))) +void vssseg8e8(vbool64_t op0, int8_t * op1, ptrdiff_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, vint8mf8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e8_v_u8m1))) +void vssseg8e8(uint8_t * op0, ptrdiff_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e8_v_u8m1_m))) +void vssseg8e8(vbool8_t op0, uint8_t * op1, ptrdiff_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, vuint8m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e8_v_u8mf2))) +void vssseg8e8(uint8_t * op0, ptrdiff_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e8_v_u8mf2_m))) +void vssseg8e8(vbool16_t op0, uint8_t * op1, ptrdiff_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, vuint8mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e8_v_u8mf4))) +void vssseg8e8(uint8_t * op0, ptrdiff_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e8_v_u8mf4_m))) +void vssseg8e8(vbool32_t op0, uint8_t * op1, ptrdiff_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, vuint8mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e8_v_u8mf8))) +void vssseg8e8(uint8_t * op0, ptrdiff_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e8_v_u8mf8_m))) +void vssseg8e8(vbool64_t op0, uint8_t * op1, ptrdiff_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, vuint8mf8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e16_v_i16m1))) +void vssseg2e16(int16_t * op0, ptrdiff_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e16_v_i16m1_m))) +void vssseg2e16(vbool16_t op0, int16_t * op1, ptrdiff_t op2, vint16m1_t op3, vint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e16_v_i16m2))) +void vssseg2e16(int16_t * op0, ptrdiff_t op1, vint16m2_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e16_v_i16m2_m))) +void vssseg2e16(vbool8_t op0, int16_t * op1, ptrdiff_t op2, vint16m2_t op3, vint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e16_v_i16m4))) +void vssseg2e16(int16_t * op0, ptrdiff_t op1, vint16m4_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e16_v_i16m4_m))) +void vssseg2e16(vbool4_t op0, int16_t * op1, ptrdiff_t op2, vint16m4_t op3, vint16m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e16_v_i16mf2))) +void vssseg2e16(int16_t * op0, ptrdiff_t op1, vint16mf2_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e16_v_i16mf2_m))) +void vssseg2e16(vbool32_t op0, int16_t * op1, ptrdiff_t op2, vint16mf2_t op3, vint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e16_v_i16mf4))) +void vssseg2e16(int16_t * op0, ptrdiff_t op1, vint16mf4_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e16_v_i16mf4_m))) +void vssseg2e16(vbool64_t op0, int16_t * op1, ptrdiff_t op2, vint16mf4_t op3, vint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e16_v_u16m1))) +void vssseg2e16(uint16_t * op0, ptrdiff_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e16_v_u16m1_m))) +void vssseg2e16(vbool16_t op0, uint16_t * op1, ptrdiff_t op2, vuint16m1_t op3, vuint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e16_v_u16m2))) +void vssseg2e16(uint16_t * op0, ptrdiff_t op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e16_v_u16m2_m))) +void vssseg2e16(vbool8_t op0, uint16_t * op1, ptrdiff_t op2, vuint16m2_t op3, vuint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e16_v_u16m4))) +void vssseg2e16(uint16_t * op0, ptrdiff_t op1, vuint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e16_v_u16m4_m))) +void vssseg2e16(vbool4_t op0, uint16_t * op1, ptrdiff_t op2, vuint16m4_t op3, vuint16m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e16_v_u16mf2))) +void vssseg2e16(uint16_t * op0, ptrdiff_t op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e16_v_u16mf2_m))) +void vssseg2e16(vbool32_t op0, uint16_t * op1, ptrdiff_t op2, vuint16mf2_t op3, vuint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e16_v_u16mf4))) +void vssseg2e16(uint16_t * op0, ptrdiff_t op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e16_v_u16mf4_m))) +void vssseg2e16(vbool64_t op0, uint16_t * op1, ptrdiff_t op2, vuint16mf4_t op3, vuint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e16_v_i16m1))) +void vssseg3e16(int16_t * op0, ptrdiff_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e16_v_i16m1_m))) +void vssseg3e16(vbool16_t op0, int16_t * op1, ptrdiff_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e16_v_i16m2))) +void vssseg3e16(int16_t * op0, ptrdiff_t op1, vint16m2_t op2, vint16m2_t op3, vint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e16_v_i16m2_m))) +void vssseg3e16(vbool8_t op0, int16_t * op1, ptrdiff_t op2, vint16m2_t op3, vint16m2_t op4, vint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e16_v_i16mf2))) +void vssseg3e16(int16_t * op0, ptrdiff_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e16_v_i16mf2_m))) +void vssseg3e16(vbool32_t op0, int16_t * op1, ptrdiff_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e16_v_i16mf4))) +void vssseg3e16(int16_t * op0, ptrdiff_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e16_v_i16mf4_m))) +void vssseg3e16(vbool64_t op0, int16_t * op1, ptrdiff_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e16_v_u16m1))) +void vssseg3e16(uint16_t * op0, ptrdiff_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e16_v_u16m1_m))) +void vssseg3e16(vbool16_t op0, uint16_t * op1, ptrdiff_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e16_v_u16m2))) +void vssseg3e16(uint16_t * op0, ptrdiff_t op1, vuint16m2_t op2, vuint16m2_t op3, vuint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e16_v_u16m2_m))) +void vssseg3e16(vbool8_t op0, uint16_t * op1, ptrdiff_t op2, vuint16m2_t op3, vuint16m2_t op4, vuint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e16_v_u16mf2))) +void vssseg3e16(uint16_t * op0, ptrdiff_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e16_v_u16mf2_m))) +void vssseg3e16(vbool32_t op0, uint16_t * op1, ptrdiff_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e16_v_u16mf4))) +void vssseg3e16(uint16_t * op0, ptrdiff_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e16_v_u16mf4_m))) +void vssseg3e16(vbool64_t op0, uint16_t * op1, ptrdiff_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e16_v_i16m1))) +void vssseg4e16(int16_t * op0, ptrdiff_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e16_v_i16m1_m))) +void vssseg4e16(vbool16_t op0, int16_t * op1, ptrdiff_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e16_v_i16m2))) +void vssseg4e16(int16_t * op0, ptrdiff_t op1, vint16m2_t op2, vint16m2_t op3, vint16m2_t op4, vint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e16_v_i16m2_m))) +void vssseg4e16(vbool8_t op0, int16_t * op1, ptrdiff_t op2, vint16m2_t op3, vint16m2_t op4, vint16m2_t op5, vint16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e16_v_i16mf2))) +void vssseg4e16(int16_t * op0, ptrdiff_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e16_v_i16mf2_m))) +void vssseg4e16(vbool32_t op0, int16_t * op1, ptrdiff_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e16_v_i16mf4))) +void vssseg4e16(int16_t * op0, ptrdiff_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e16_v_i16mf4_m))) +void vssseg4e16(vbool64_t op0, int16_t * op1, ptrdiff_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e16_v_u16m1))) +void vssseg4e16(uint16_t * op0, ptrdiff_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e16_v_u16m1_m))) +void vssseg4e16(vbool16_t op0, uint16_t * op1, ptrdiff_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e16_v_u16m2))) +void vssseg4e16(uint16_t * op0, ptrdiff_t op1, vuint16m2_t op2, vuint16m2_t op3, vuint16m2_t op4, vuint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e16_v_u16m2_m))) +void vssseg4e16(vbool8_t op0, uint16_t * op1, ptrdiff_t op2, vuint16m2_t op3, vuint16m2_t op4, vuint16m2_t op5, vuint16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e16_v_u16mf2))) +void vssseg4e16(uint16_t * op0, ptrdiff_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e16_v_u16mf2_m))) +void vssseg4e16(vbool32_t op0, uint16_t * op1, ptrdiff_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e16_v_u16mf4))) +void vssseg4e16(uint16_t * op0, ptrdiff_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e16_v_u16mf4_m))) +void vssseg4e16(vbool64_t op0, uint16_t * op1, ptrdiff_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e16_v_i16m1))) +void vssseg5e16(int16_t * op0, ptrdiff_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e16_v_i16m1_m))) +void vssseg5e16(vbool16_t op0, int16_t * op1, ptrdiff_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e16_v_i16mf2))) +void vssseg5e16(int16_t * op0, ptrdiff_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e16_v_i16mf2_m))) +void vssseg5e16(vbool32_t op0, int16_t * op1, ptrdiff_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e16_v_i16mf4))) +void vssseg5e16(int16_t * op0, ptrdiff_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e16_v_i16mf4_m))) +void vssseg5e16(vbool64_t op0, int16_t * op1, ptrdiff_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e16_v_u16m1))) +void vssseg5e16(uint16_t * op0, ptrdiff_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e16_v_u16m1_m))) +void vssseg5e16(vbool16_t op0, uint16_t * op1, ptrdiff_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e16_v_u16mf2))) +void vssseg5e16(uint16_t * op0, ptrdiff_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e16_v_u16mf2_m))) +void vssseg5e16(vbool32_t op0, uint16_t * op1, ptrdiff_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e16_v_u16mf4))) +void vssseg5e16(uint16_t * op0, ptrdiff_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e16_v_u16mf4_m))) +void vssseg5e16(vbool64_t op0, uint16_t * op1, ptrdiff_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e16_v_i16m1))) +void vssseg6e16(int16_t * op0, ptrdiff_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e16_v_i16m1_m))) +void vssseg6e16(vbool16_t op0, int16_t * op1, ptrdiff_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e16_v_i16mf2))) +void vssseg6e16(int16_t * op0, ptrdiff_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e16_v_i16mf2_m))) +void vssseg6e16(vbool32_t op0, int16_t * op1, ptrdiff_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e16_v_i16mf4))) +void vssseg6e16(int16_t * op0, ptrdiff_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e16_v_i16mf4_m))) +void vssseg6e16(vbool64_t op0, int16_t * op1, ptrdiff_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e16_v_u16m1))) +void vssseg6e16(uint16_t * op0, ptrdiff_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e16_v_u16m1_m))) +void vssseg6e16(vbool16_t op0, uint16_t * op1, ptrdiff_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e16_v_u16mf2))) +void vssseg6e16(uint16_t * op0, ptrdiff_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e16_v_u16mf2_m))) +void vssseg6e16(vbool32_t op0, uint16_t * op1, ptrdiff_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e16_v_u16mf4))) +void vssseg6e16(uint16_t * op0, ptrdiff_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e16_v_u16mf4_m))) +void vssseg6e16(vbool64_t op0, uint16_t * op1, ptrdiff_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e16_v_i16m1))) +void vssseg7e16(int16_t * op0, ptrdiff_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e16_v_i16m1_m))) +void vssseg7e16(vbool16_t op0, int16_t * op1, ptrdiff_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e16_v_i16mf2))) +void vssseg7e16(int16_t * op0, ptrdiff_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e16_v_i16mf2_m))) +void vssseg7e16(vbool32_t op0, int16_t * op1, ptrdiff_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e16_v_i16mf4))) +void vssseg7e16(int16_t * op0, ptrdiff_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e16_v_i16mf4_m))) +void vssseg7e16(vbool64_t op0, int16_t * op1, ptrdiff_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e16_v_u16m1))) +void vssseg7e16(uint16_t * op0, ptrdiff_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e16_v_u16m1_m))) +void vssseg7e16(vbool16_t op0, uint16_t * op1, ptrdiff_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e16_v_u16mf2))) +void vssseg7e16(uint16_t * op0, ptrdiff_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e16_v_u16mf2_m))) +void vssseg7e16(vbool32_t op0, uint16_t * op1, ptrdiff_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e16_v_u16mf4))) +void vssseg7e16(uint16_t * op0, ptrdiff_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e16_v_u16mf4_m))) +void vssseg7e16(vbool64_t op0, uint16_t * op1, ptrdiff_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e16_v_i16m1))) +void vssseg8e16(int16_t * op0, ptrdiff_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e16_v_i16m1_m))) +void vssseg8e16(vbool16_t op0, int16_t * op1, ptrdiff_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, vint16m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e16_v_i16mf2))) +void vssseg8e16(int16_t * op0, ptrdiff_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e16_v_i16mf2_m))) +void vssseg8e16(vbool32_t op0, int16_t * op1, ptrdiff_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, vint16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e16_v_i16mf4))) +void vssseg8e16(int16_t * op0, ptrdiff_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e16_v_i16mf4_m))) +void vssseg8e16(vbool64_t op0, int16_t * op1, ptrdiff_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, vint16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e16_v_u16m1))) +void vssseg8e16(uint16_t * op0, ptrdiff_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e16_v_u16m1_m))) +void vssseg8e16(vbool16_t op0, uint16_t * op1, ptrdiff_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, vuint16m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e16_v_u16mf2))) +void vssseg8e16(uint16_t * op0, ptrdiff_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e16_v_u16mf2_m))) +void vssseg8e16(vbool32_t op0, uint16_t * op1, ptrdiff_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, vuint16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e16_v_u16mf4))) +void vssseg8e16(uint16_t * op0, ptrdiff_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e16_v_u16mf4_m))) +void vssseg8e16(vbool64_t op0, uint16_t * op1, ptrdiff_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, vuint16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e32_v_i32m1))) +void vssseg2e32(int32_t * op0, ptrdiff_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e32_v_i32m1_m))) +void vssseg2e32(vbool32_t op0, int32_t * op1, ptrdiff_t op2, vint32m1_t op3, vint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e32_v_i32m2))) +void vssseg2e32(int32_t * op0, ptrdiff_t op1, vint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e32_v_i32m2_m))) +void vssseg2e32(vbool16_t op0, int32_t * op1, ptrdiff_t op2, vint32m2_t op3, vint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e32_v_i32m4))) +void vssseg2e32(int32_t * op0, ptrdiff_t op1, vint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e32_v_i32m4_m))) +void vssseg2e32(vbool8_t op0, int32_t * op1, ptrdiff_t op2, vint32m4_t op3, vint32m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e32_v_i32mf2))) +void vssseg2e32(int32_t * op0, ptrdiff_t op1, vint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e32_v_i32mf2_m))) +void vssseg2e32(vbool64_t op0, int32_t * op1, ptrdiff_t op2, vint32mf2_t op3, vint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e32_v_u32m1))) +void vssseg2e32(uint32_t * op0, ptrdiff_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e32_v_u32m1_m))) +void vssseg2e32(vbool32_t op0, uint32_t * op1, ptrdiff_t op2, vuint32m1_t op3, vuint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e32_v_u32m2))) +void vssseg2e32(uint32_t * op0, ptrdiff_t op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e32_v_u32m2_m))) +void vssseg2e32(vbool16_t op0, uint32_t * op1, ptrdiff_t op2, vuint32m2_t op3, vuint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e32_v_u32m4))) +void vssseg2e32(uint32_t * op0, ptrdiff_t op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e32_v_u32m4_m))) +void vssseg2e32(vbool8_t op0, uint32_t * op1, ptrdiff_t op2, vuint32m4_t op3, vuint32m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e32_v_u32mf2))) +void vssseg2e32(uint32_t * op0, ptrdiff_t op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e32_v_u32mf2_m))) +void vssseg2e32(vbool64_t op0, uint32_t * op1, ptrdiff_t op2, vuint32mf2_t op3, vuint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e32_v_i32m1))) +void vssseg3e32(int32_t * op0, ptrdiff_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e32_v_i32m1_m))) +void vssseg3e32(vbool32_t op0, int32_t * op1, ptrdiff_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e32_v_i32m2))) +void vssseg3e32(int32_t * op0, ptrdiff_t op1, vint32m2_t op2, vint32m2_t op3, vint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e32_v_i32m2_m))) +void vssseg3e32(vbool16_t op0, int32_t * op1, ptrdiff_t op2, vint32m2_t op3, vint32m2_t op4, vint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e32_v_i32mf2))) +void vssseg3e32(int32_t * op0, ptrdiff_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e32_v_i32mf2_m))) +void vssseg3e32(vbool64_t op0, int32_t * op1, ptrdiff_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e32_v_u32m1))) +void vssseg3e32(uint32_t * op0, ptrdiff_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e32_v_u32m1_m))) +void vssseg3e32(vbool32_t op0, uint32_t * op1, ptrdiff_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e32_v_u32m2))) +void vssseg3e32(uint32_t * op0, ptrdiff_t op1, vuint32m2_t op2, vuint32m2_t op3, vuint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e32_v_u32m2_m))) +void vssseg3e32(vbool16_t op0, uint32_t * op1, ptrdiff_t op2, vuint32m2_t op3, vuint32m2_t op4, vuint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e32_v_u32mf2))) +void vssseg3e32(uint32_t * op0, ptrdiff_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e32_v_u32mf2_m))) +void vssseg3e32(vbool64_t op0, uint32_t * op1, ptrdiff_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e32_v_i32m1))) +void vssseg4e32(int32_t * op0, ptrdiff_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e32_v_i32m1_m))) +void vssseg4e32(vbool32_t op0, int32_t * op1, ptrdiff_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e32_v_i32m2))) +void vssseg4e32(int32_t * op0, ptrdiff_t op1, vint32m2_t op2, vint32m2_t op3, vint32m2_t op4, vint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e32_v_i32m2_m))) +void vssseg4e32(vbool16_t op0, int32_t * op1, ptrdiff_t op2, vint32m2_t op3, vint32m2_t op4, vint32m2_t op5, vint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e32_v_i32mf2))) +void vssseg4e32(int32_t * op0, ptrdiff_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e32_v_i32mf2_m))) +void vssseg4e32(vbool64_t op0, int32_t * op1, ptrdiff_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e32_v_u32m1))) +void vssseg4e32(uint32_t * op0, ptrdiff_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e32_v_u32m1_m))) +void vssseg4e32(vbool32_t op0, uint32_t * op1, ptrdiff_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e32_v_u32m2))) +void vssseg4e32(uint32_t * op0, ptrdiff_t op1, vuint32m2_t op2, vuint32m2_t op3, vuint32m2_t op4, vuint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e32_v_u32m2_m))) +void vssseg4e32(vbool16_t op0, uint32_t * op1, ptrdiff_t op2, vuint32m2_t op3, vuint32m2_t op4, vuint32m2_t op5, vuint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e32_v_u32mf2))) +void vssseg4e32(uint32_t * op0, ptrdiff_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e32_v_u32mf2_m))) +void vssseg4e32(vbool64_t op0, uint32_t * op1, ptrdiff_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e32_v_i32m1))) +void vssseg5e32(int32_t * op0, ptrdiff_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e32_v_i32m1_m))) +void vssseg5e32(vbool32_t op0, int32_t * op1, ptrdiff_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e32_v_i32mf2))) +void vssseg5e32(int32_t * op0, ptrdiff_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e32_v_i32mf2_m))) +void vssseg5e32(vbool64_t op0, int32_t * op1, ptrdiff_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e32_v_u32m1))) +void vssseg5e32(uint32_t * op0, ptrdiff_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e32_v_u32m1_m))) +void vssseg5e32(vbool32_t op0, uint32_t * op1, ptrdiff_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e32_v_u32mf2))) +void vssseg5e32(uint32_t * op0, ptrdiff_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e32_v_u32mf2_m))) +void vssseg5e32(vbool64_t op0, uint32_t * op1, ptrdiff_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e32_v_i32m1))) +void vssseg6e32(int32_t * op0, ptrdiff_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e32_v_i32m1_m))) +void vssseg6e32(vbool32_t op0, int32_t * op1, ptrdiff_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e32_v_i32mf2))) +void vssseg6e32(int32_t * op0, ptrdiff_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e32_v_i32mf2_m))) +void vssseg6e32(vbool64_t op0, int32_t * op1, ptrdiff_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e32_v_u32m1))) +void vssseg6e32(uint32_t * op0, ptrdiff_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e32_v_u32m1_m))) +void vssseg6e32(vbool32_t op0, uint32_t * op1, ptrdiff_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e32_v_u32mf2))) +void vssseg6e32(uint32_t * op0, ptrdiff_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e32_v_u32mf2_m))) +void vssseg6e32(vbool64_t op0, uint32_t * op1, ptrdiff_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e32_v_i32m1))) +void vssseg7e32(int32_t * op0, ptrdiff_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e32_v_i32m1_m))) +void vssseg7e32(vbool32_t op0, int32_t * op1, ptrdiff_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e32_v_i32mf2))) +void vssseg7e32(int32_t * op0, ptrdiff_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e32_v_i32mf2_m))) +void vssseg7e32(vbool64_t op0, int32_t * op1, ptrdiff_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e32_v_u32m1))) +void vssseg7e32(uint32_t * op0, ptrdiff_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e32_v_u32m1_m))) +void vssseg7e32(vbool32_t op0, uint32_t * op1, ptrdiff_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e32_v_u32mf2))) +void vssseg7e32(uint32_t * op0, ptrdiff_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e32_v_u32mf2_m))) +void vssseg7e32(vbool64_t op0, uint32_t * op1, ptrdiff_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e32_v_i32m1))) +void vssseg8e32(int32_t * op0, ptrdiff_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e32_v_i32m1_m))) +void vssseg8e32(vbool32_t op0, int32_t * op1, ptrdiff_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, vint32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e32_v_i32mf2))) +void vssseg8e32(int32_t * op0, ptrdiff_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e32_v_i32mf2_m))) +void vssseg8e32(vbool64_t op0, int32_t * op1, ptrdiff_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, vint32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e32_v_u32m1))) +void vssseg8e32(uint32_t * op0, ptrdiff_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e32_v_u32m1_m))) +void vssseg8e32(vbool32_t op0, uint32_t * op1, ptrdiff_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, vuint32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e32_v_u32mf2))) +void vssseg8e32(uint32_t * op0, ptrdiff_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e32_v_u32mf2_m))) +void vssseg8e32(vbool64_t op0, uint32_t * op1, ptrdiff_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, vuint32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e64_v_i64m1))) +void vssseg2e64(int64_t * op0, ptrdiff_t op1, vint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e64_v_i64m1_m))) +void vssseg2e64(vbool64_t op0, int64_t * op1, ptrdiff_t op2, vint64m1_t op3, vint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e64_v_i64m2))) +void vssseg2e64(int64_t * op0, ptrdiff_t op1, vint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e64_v_i64m2_m))) +void vssseg2e64(vbool32_t op0, int64_t * op1, ptrdiff_t op2, vint64m2_t op3, vint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e64_v_i64m4))) +void vssseg2e64(int64_t * op0, ptrdiff_t op1, vint64m4_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e64_v_i64m4_m))) +void vssseg2e64(vbool16_t op0, int64_t * op1, ptrdiff_t op2, vint64m4_t op3, vint64m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e64_v_u64m1))) +void vssseg2e64(uint64_t * op0, ptrdiff_t op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e64_v_u64m1_m))) +void vssseg2e64(vbool64_t op0, uint64_t * op1, ptrdiff_t op2, vuint64m1_t op3, vuint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e64_v_u64m2))) +void vssseg2e64(uint64_t * op0, ptrdiff_t op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e64_v_u64m2_m))) +void vssseg2e64(vbool32_t op0, uint64_t * op1, ptrdiff_t op2, vuint64m2_t op3, vuint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e64_v_u64m4))) +void vssseg2e64(uint64_t * op0, ptrdiff_t op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e64_v_u64m4_m))) +void vssseg2e64(vbool16_t op0, uint64_t * op1, ptrdiff_t op2, vuint64m4_t op3, vuint64m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e64_v_i64m1))) +void vssseg3e64(int64_t * op0, ptrdiff_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e64_v_i64m1_m))) +void vssseg3e64(vbool64_t op0, int64_t * op1, ptrdiff_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e64_v_i64m2))) +void vssseg3e64(int64_t * op0, ptrdiff_t op1, vint64m2_t op2, vint64m2_t op3, vint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e64_v_i64m2_m))) +void vssseg3e64(vbool32_t op0, int64_t * op1, ptrdiff_t op2, vint64m2_t op3, vint64m2_t op4, vint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e64_v_u64m1))) +void vssseg3e64(uint64_t * op0, ptrdiff_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e64_v_u64m1_m))) +void vssseg3e64(vbool64_t op0, uint64_t * op1, ptrdiff_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e64_v_u64m2))) +void vssseg3e64(uint64_t * op0, ptrdiff_t op1, vuint64m2_t op2, vuint64m2_t op3, vuint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e64_v_u64m2_m))) +void vssseg3e64(vbool32_t op0, uint64_t * op1, ptrdiff_t op2, vuint64m2_t op3, vuint64m2_t op4, vuint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e64_v_i64m1))) +void vssseg4e64(int64_t * op0, ptrdiff_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e64_v_i64m1_m))) +void vssseg4e64(vbool64_t op0, int64_t * op1, ptrdiff_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e64_v_i64m2))) +void vssseg4e64(int64_t * op0, ptrdiff_t op1, vint64m2_t op2, vint64m2_t op3, vint64m2_t op4, vint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e64_v_i64m2_m))) +void vssseg4e64(vbool32_t op0, int64_t * op1, ptrdiff_t op2, vint64m2_t op3, vint64m2_t op4, vint64m2_t op5, vint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e64_v_u64m1))) +void vssseg4e64(uint64_t * op0, ptrdiff_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e64_v_u64m1_m))) +void vssseg4e64(vbool64_t op0, uint64_t * op1, ptrdiff_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e64_v_u64m2))) +void vssseg4e64(uint64_t * op0, ptrdiff_t op1, vuint64m2_t op2, vuint64m2_t op3, vuint64m2_t op4, vuint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e64_v_u64m2_m))) +void vssseg4e64(vbool32_t op0, uint64_t * op1, ptrdiff_t op2, vuint64m2_t op3, vuint64m2_t op4, vuint64m2_t op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e64_v_i64m1))) +void vssseg5e64(int64_t * op0, ptrdiff_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e64_v_i64m1_m))) +void vssseg5e64(vbool64_t op0, int64_t * op1, ptrdiff_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e64_v_u64m1))) +void vssseg5e64(uint64_t * op0, ptrdiff_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e64_v_u64m1_m))) +void vssseg5e64(vbool64_t op0, uint64_t * op1, ptrdiff_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e64_v_i64m1))) +void vssseg6e64(int64_t * op0, ptrdiff_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e64_v_i64m1_m))) +void vssseg6e64(vbool64_t op0, int64_t * op1, ptrdiff_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e64_v_u64m1))) +void vssseg6e64(uint64_t * op0, ptrdiff_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e64_v_u64m1_m))) +void vssseg6e64(vbool64_t op0, uint64_t * op1, ptrdiff_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e64_v_i64m1))) +void vssseg7e64(int64_t * op0, ptrdiff_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e64_v_i64m1_m))) +void vssseg7e64(vbool64_t op0, int64_t * op1, ptrdiff_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e64_v_u64m1))) +void vssseg7e64(uint64_t * op0, ptrdiff_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e64_v_u64m1_m))) +void vssseg7e64(vbool64_t op0, uint64_t * op1, ptrdiff_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e64_v_i64m1))) +void vssseg8e64(int64_t * op0, ptrdiff_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e64_v_i64m1_m))) +void vssseg8e64(vbool64_t op0, int64_t * op1, ptrdiff_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, vint64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e64_v_u64m1))) +void vssseg8e64(uint64_t * op0, ptrdiff_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e64_v_u64m1_m))) +void vssseg8e64(vbool64_t op0, uint64_t * op1, ptrdiff_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, vuint64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_i8m1))) +void vsuxseg3ei8(int8_t * op0, vuint8m1_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_i8m1_m))) +void vsuxseg3ei8(vbool8_t op0, int8_t * op1, vuint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_i8m2))) +void vsuxseg3ei8(int8_t * op0, vuint8m2_t op1, vint8m2_t op2, vint8m2_t op3, vint8m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_i8m2_m))) +void vsuxseg3ei8(vbool4_t op0, int8_t * op1, vuint8m2_t op2, vint8m2_t op3, vint8m2_t op4, vint8m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_i8mf2))) +void vsuxseg3ei8(int8_t * op0, vuint8mf2_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_i8mf2_m))) +void vsuxseg3ei8(vbool16_t op0, int8_t * op1, vuint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_i8mf4))) +void vsuxseg3ei8(int8_t * op0, vuint8mf4_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_i8mf4_m))) +void vsuxseg3ei8(vbool32_t op0, int8_t * op1, vuint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_i8mf8))) +void vsuxseg3ei8(int8_t * op0, vuint8mf8_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_i8mf8_m))) +void vsuxseg3ei8(vbool64_t op0, int8_t * op1, vuint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_u8m1))) +void vsuxseg3ei8(uint8_t * op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_u8m1_m))) +void vsuxseg3ei8(vbool8_t op0, uint8_t * op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_u8m2))) +void vsuxseg3ei8(uint8_t * op0, vuint8m2_t op1, vuint8m2_t op2, vuint8m2_t op3, vuint8m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_u8m2_m))) +void vsuxseg3ei8(vbool4_t op0, uint8_t * op1, vuint8m2_t op2, vuint8m2_t op3, vuint8m2_t op4, vuint8m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_u8mf2))) +void vsuxseg3ei8(uint8_t * op0, vuint8mf2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_u8mf2_m))) +void vsuxseg3ei8(vbool16_t op0, uint8_t * op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_u8mf4))) +void vsuxseg3ei8(uint8_t * op0, vuint8mf4_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_u8mf4_m))) +void vsuxseg3ei8(vbool32_t op0, uint8_t * op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_u8mf8))) +void vsuxseg3ei8(uint8_t * op0, vuint8mf8_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_u8mf8_m))) +void vsuxseg3ei8(vbool64_t op0, uint8_t * op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_i8m1))) +void vsuxseg4ei8(int8_t * op0, vuint8m1_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_i8m1_m))) +void vsuxseg4ei8(vbool8_t op0, int8_t * op1, vuint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_i8m2))) +void vsuxseg4ei8(int8_t * op0, vuint8m2_t op1, vint8m2_t op2, vint8m2_t op3, vint8m2_t op4, vint8m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_i8m2_m))) +void vsuxseg4ei8(vbool4_t op0, int8_t * op1, vuint8m2_t op2, vint8m2_t op3, vint8m2_t op4, vint8m2_t op5, vint8m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_i8mf2))) +void vsuxseg4ei8(int8_t * op0, vuint8mf2_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_i8mf2_m))) +void vsuxseg4ei8(vbool16_t op0, int8_t * op1, vuint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_i8mf4))) +void vsuxseg4ei8(int8_t * op0, vuint8mf4_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_i8mf4_m))) +void vsuxseg4ei8(vbool32_t op0, int8_t * op1, vuint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_i8mf8))) +void vsuxseg4ei8(int8_t * op0, vuint8mf8_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_i8mf8_m))) +void vsuxseg4ei8(vbool64_t op0, int8_t * op1, vuint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_u8m1))) +void vsuxseg4ei8(uint8_t * op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_u8m1_m))) +void vsuxseg4ei8(vbool8_t op0, uint8_t * op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_u8m2))) +void vsuxseg4ei8(uint8_t * op0, vuint8m2_t op1, vuint8m2_t op2, vuint8m2_t op3, vuint8m2_t op4, vuint8m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_u8m2_m))) +void vsuxseg4ei8(vbool4_t op0, uint8_t * op1, vuint8m2_t op2, vuint8m2_t op3, vuint8m2_t op4, vuint8m2_t op5, vuint8m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_u8mf2))) +void vsuxseg4ei8(uint8_t * op0, vuint8mf2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_u8mf2_m))) +void vsuxseg4ei8(vbool16_t op0, uint8_t * op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_u8mf4))) +void vsuxseg4ei8(uint8_t * op0, vuint8mf4_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_u8mf4_m))) +void vsuxseg4ei8(vbool32_t op0, uint8_t * op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_u8mf8))) +void vsuxseg4ei8(uint8_t * op0, vuint8mf8_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_u8mf8_m))) +void vsuxseg4ei8(vbool64_t op0, uint8_t * op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_i8m1))) +void vsuxseg5ei8(int8_t * op0, vuint8m1_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_i8m1_m))) +void vsuxseg5ei8(vbool8_t op0, int8_t * op1, vuint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_i8mf2))) +void vsuxseg5ei8(int8_t * op0, vuint8mf2_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_i8mf2_m))) +void vsuxseg5ei8(vbool16_t op0, int8_t * op1, vuint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_i8mf4))) +void vsuxseg5ei8(int8_t * op0, vuint8mf4_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_i8mf4_m))) +void vsuxseg5ei8(vbool32_t op0, int8_t * op1, vuint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_i8mf8))) +void vsuxseg5ei8(int8_t * op0, vuint8mf8_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_i8mf8_m))) +void vsuxseg5ei8(vbool64_t op0, int8_t * op1, vuint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_u8m1))) +void vsuxseg5ei8(uint8_t * op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_u8m1_m))) +void vsuxseg5ei8(vbool8_t op0, uint8_t * op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_u8mf2))) +void vsuxseg5ei8(uint8_t * op0, vuint8mf2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_u8mf2_m))) +void vsuxseg5ei8(vbool16_t op0, uint8_t * op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_u8mf4))) +void vsuxseg5ei8(uint8_t * op0, vuint8mf4_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_u8mf4_m))) +void vsuxseg5ei8(vbool32_t op0, uint8_t * op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_u8mf8))) +void vsuxseg5ei8(uint8_t * op0, vuint8mf8_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_u8mf8_m))) +void vsuxseg5ei8(vbool64_t op0, uint8_t * op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_i8m1))) +void vsuxseg6ei8(int8_t * op0, vuint8m1_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_i8m1_m))) +void vsuxseg6ei8(vbool8_t op0, int8_t * op1, vuint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_i8mf2))) +void vsuxseg6ei8(int8_t * op0, vuint8mf2_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_i8mf2_m))) +void vsuxseg6ei8(vbool16_t op0, int8_t * op1, vuint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_i8mf4))) +void vsuxseg6ei8(int8_t * op0, vuint8mf4_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_i8mf4_m))) +void vsuxseg6ei8(vbool32_t op0, int8_t * op1, vuint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_i8mf8))) +void vsuxseg6ei8(int8_t * op0, vuint8mf8_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_i8mf8_m))) +void vsuxseg6ei8(vbool64_t op0, int8_t * op1, vuint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_u8m1))) +void vsuxseg6ei8(uint8_t * op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_u8m1_m))) +void vsuxseg6ei8(vbool8_t op0, uint8_t * op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_u8mf2))) +void vsuxseg6ei8(uint8_t * op0, vuint8mf2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_u8mf2_m))) +void vsuxseg6ei8(vbool16_t op0, uint8_t * op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_u8mf4))) +void vsuxseg6ei8(uint8_t * op0, vuint8mf4_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_u8mf4_m))) +void vsuxseg6ei8(vbool32_t op0, uint8_t * op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_u8mf8))) +void vsuxseg6ei8(uint8_t * op0, vuint8mf8_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_u8mf8_m))) +void vsuxseg6ei8(vbool64_t op0, uint8_t * op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_i8m1))) +void vsuxseg7ei8(int8_t * op0, vuint8m1_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_i8m1_m))) +void vsuxseg7ei8(vbool8_t op0, int8_t * op1, vuint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_i8mf2))) +void vsuxseg7ei8(int8_t * op0, vuint8mf2_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_i8mf2_m))) +void vsuxseg7ei8(vbool16_t op0, int8_t * op1, vuint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_i8mf4))) +void vsuxseg7ei8(int8_t * op0, vuint8mf4_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_i8mf4_m))) +void vsuxseg7ei8(vbool32_t op0, int8_t * op1, vuint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_i8mf8))) +void vsuxseg7ei8(int8_t * op0, vuint8mf8_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_i8mf8_m))) +void vsuxseg7ei8(vbool64_t op0, int8_t * op1, vuint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_u8m1))) +void vsuxseg7ei8(uint8_t * op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_u8m1_m))) +void vsuxseg7ei8(vbool8_t op0, uint8_t * op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_u8mf2))) +void vsuxseg7ei8(uint8_t * op0, vuint8mf2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_u8mf2_m))) +void vsuxseg7ei8(vbool16_t op0, uint8_t * op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_u8mf4))) +void vsuxseg7ei8(uint8_t * op0, vuint8mf4_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_u8mf4_m))) +void vsuxseg7ei8(vbool32_t op0, uint8_t * op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_u8mf8))) +void vsuxseg7ei8(uint8_t * op0, vuint8mf8_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_u8mf8_m))) +void vsuxseg7ei8(vbool64_t op0, uint8_t * op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_i8m1))) +void vsuxseg8ei8(int8_t * op0, vuint8m1_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_i8m1_m))) +void vsuxseg8ei8(vbool8_t op0, int8_t * op1, vuint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, vint8m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_i8mf2))) +void vsuxseg8ei8(int8_t * op0, vuint8mf2_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_i8mf2_m))) +void vsuxseg8ei8(vbool16_t op0, int8_t * op1, vuint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, vint8mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_i8mf4))) +void vsuxseg8ei8(int8_t * op0, vuint8mf4_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_i8mf4_m))) +void vsuxseg8ei8(vbool32_t op0, int8_t * op1, vuint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, vint8mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_i8mf8))) +void vsuxseg8ei8(int8_t * op0, vuint8mf8_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_i8mf8_m))) +void vsuxseg8ei8(vbool64_t op0, int8_t * op1, vuint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, vint8mf8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_u8m1))) +void vsuxseg8ei8(uint8_t * op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_u8m1_m))) +void vsuxseg8ei8(vbool8_t op0, uint8_t * op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, vuint8m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_u8mf2))) +void vsuxseg8ei8(uint8_t * op0, vuint8mf2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_u8mf2_m))) +void vsuxseg8ei8(vbool16_t op0, uint8_t * op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, vuint8mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_u8mf4))) +void vsuxseg8ei8(uint8_t * op0, vuint8mf4_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_u8mf4_m))) +void vsuxseg8ei8(vbool32_t op0, uint8_t * op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, vuint8mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_u8mf8))) +void vsuxseg8ei8(uint8_t * op0, vuint8mf8_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_u8mf8_m))) +void vsuxseg8ei8(vbool64_t op0, uint8_t * op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, vuint8mf8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_i8m1))) +void vsuxseg2ei16(int8_t * op0, vuint16m2_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_i8m1_m))) +void vsuxseg2ei16(vbool8_t op0, int8_t * op1, vuint16m2_t op2, vint8m1_t op3, vint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_i8m2))) +void vsuxseg2ei16(int8_t * op0, vuint16m4_t op1, vint8m2_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_i8m2_m))) +void vsuxseg2ei16(vbool4_t op0, int8_t * op1, vuint16m4_t op2, vint8m2_t op3, vint8m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_i8m4))) +void vsuxseg2ei16(int8_t * op0, vuint16m8_t op1, vint8m4_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_i8m4_m))) +void vsuxseg2ei16(vbool2_t op0, int8_t * op1, vuint16m8_t op2, vint8m4_t op3, vint8m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_i8mf2))) +void vsuxseg2ei16(int8_t * op0, vuint16m1_t op1, vint8mf2_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_i8mf2_m))) +void vsuxseg2ei16(vbool16_t op0, int8_t * op1, vuint16m1_t op2, vint8mf2_t op3, vint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_i8mf4))) +void vsuxseg2ei16(int8_t * op0, vuint16mf2_t op1, vint8mf4_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_i8mf4_m))) +void vsuxseg2ei16(vbool32_t op0, int8_t * op1, vuint16mf2_t op2, vint8mf4_t op3, vint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_i8mf8))) +void vsuxseg2ei16(int8_t * op0, vuint16mf4_t op1, vint8mf8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_i8mf8_m))) +void vsuxseg2ei16(vbool64_t op0, int8_t * op1, vuint16mf4_t op2, vint8mf8_t op3, vint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_u8m1))) +void vsuxseg2ei16(uint8_t * op0, vuint16m2_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_u8m1_m))) +void vsuxseg2ei16(vbool8_t op0, uint8_t * op1, vuint16m2_t op2, vuint8m1_t op3, vuint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_u8m2))) +void vsuxseg2ei16(uint8_t * op0, vuint16m4_t op1, vuint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_u8m2_m))) +void vsuxseg2ei16(vbool4_t op0, uint8_t * op1, vuint16m4_t op2, vuint8m2_t op3, vuint8m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_u8m4))) +void vsuxseg2ei16(uint8_t * op0, vuint16m8_t op1, vuint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_u8m4_m))) +void vsuxseg2ei16(vbool2_t op0, uint8_t * op1, vuint16m8_t op2, vuint8m4_t op3, vuint8m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_u8mf2))) +void vsuxseg2ei16(uint8_t * op0, vuint16m1_t op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_u8mf2_m))) +void vsuxseg2ei16(vbool16_t op0, uint8_t * op1, vuint16m1_t op2, vuint8mf2_t op3, vuint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_u8mf4))) +void vsuxseg2ei16(uint8_t * op0, vuint16mf2_t op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_u8mf4_m))) +void vsuxseg2ei16(vbool32_t op0, uint8_t * op1, vuint16mf2_t op2, vuint8mf4_t op3, vuint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_u8mf8))) +void vsuxseg2ei16(uint8_t * op0, vuint16mf4_t op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_u8mf8_m))) +void vsuxseg2ei16(vbool64_t op0, uint8_t * op1, vuint16mf4_t op2, vuint8mf8_t op3, vuint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_i8m1))) +void vsuxseg3ei16(int8_t * op0, vuint16m2_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_i8m1_m))) +void vsuxseg3ei16(vbool8_t op0, int8_t * op1, vuint16m2_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_i8m2))) +void vsuxseg3ei16(int8_t * op0, vuint16m4_t op1, vint8m2_t op2, vint8m2_t op3, vint8m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_i8m2_m))) +void vsuxseg3ei16(vbool4_t op0, int8_t * op1, vuint16m4_t op2, vint8m2_t op3, vint8m2_t op4, vint8m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_i8mf2))) +void vsuxseg3ei16(int8_t * op0, vuint16m1_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_i8mf2_m))) +void vsuxseg3ei16(vbool16_t op0, int8_t * op1, vuint16m1_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_i8mf4))) +void vsuxseg3ei16(int8_t * op0, vuint16mf2_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_i8mf4_m))) +void vsuxseg3ei16(vbool32_t op0, int8_t * op1, vuint16mf2_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_i8mf8))) +void vsuxseg3ei16(int8_t * op0, vuint16mf4_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_i8mf8_m))) +void vsuxseg3ei16(vbool64_t op0, int8_t * op1, vuint16mf4_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_u8m1))) +void vsuxseg3ei16(uint8_t * op0, vuint16m2_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_u8m1_m))) +void vsuxseg3ei16(vbool8_t op0, uint8_t * op1, vuint16m2_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_u8m2))) +void vsuxseg3ei16(uint8_t * op0, vuint16m4_t op1, vuint8m2_t op2, vuint8m2_t op3, vuint8m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_u8m2_m))) +void vsuxseg3ei16(vbool4_t op0, uint8_t * op1, vuint16m4_t op2, vuint8m2_t op3, vuint8m2_t op4, vuint8m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_u8mf2))) +void vsuxseg3ei16(uint8_t * op0, vuint16m1_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_u8mf2_m))) +void vsuxseg3ei16(vbool16_t op0, uint8_t * op1, vuint16m1_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_u8mf4))) +void vsuxseg3ei16(uint8_t * op0, vuint16mf2_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_u8mf4_m))) +void vsuxseg3ei16(vbool32_t op0, uint8_t * op1, vuint16mf2_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_u8mf8))) +void vsuxseg3ei16(uint8_t * op0, vuint16mf4_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_u8mf8_m))) +void vsuxseg3ei16(vbool64_t op0, uint8_t * op1, vuint16mf4_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_i8m1))) +void vsuxseg4ei16(int8_t * op0, vuint16m2_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_i8m1_m))) +void vsuxseg4ei16(vbool8_t op0, int8_t * op1, vuint16m2_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_i8m2))) +void vsuxseg4ei16(int8_t * op0, vuint16m4_t op1, vint8m2_t op2, vint8m2_t op3, vint8m2_t op4, vint8m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_i8m2_m))) +void vsuxseg4ei16(vbool4_t op0, int8_t * op1, vuint16m4_t op2, vint8m2_t op3, vint8m2_t op4, vint8m2_t op5, vint8m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_i8mf2))) +void vsuxseg4ei16(int8_t * op0, vuint16m1_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_i8mf2_m))) +void vsuxseg4ei16(vbool16_t op0, int8_t * op1, vuint16m1_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_i8mf4))) +void vsuxseg4ei16(int8_t * op0, vuint16mf2_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_i8mf4_m))) +void vsuxseg4ei16(vbool32_t op0, int8_t * op1, vuint16mf2_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_i8mf8))) +void vsuxseg4ei16(int8_t * op0, vuint16mf4_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_i8mf8_m))) +void vsuxseg4ei16(vbool64_t op0, int8_t * op1, vuint16mf4_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_u8m1))) +void vsuxseg4ei16(uint8_t * op0, vuint16m2_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_u8m1_m))) +void vsuxseg4ei16(vbool8_t op0, uint8_t * op1, vuint16m2_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_u8m2))) +void vsuxseg4ei16(uint8_t * op0, vuint16m4_t op1, vuint8m2_t op2, vuint8m2_t op3, vuint8m2_t op4, vuint8m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_u8m2_m))) +void vsuxseg4ei16(vbool4_t op0, uint8_t * op1, vuint16m4_t op2, vuint8m2_t op3, vuint8m2_t op4, vuint8m2_t op5, vuint8m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_u8mf2))) +void vsuxseg4ei16(uint8_t * op0, vuint16m1_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_u8mf2_m))) +void vsuxseg4ei16(vbool16_t op0, uint8_t * op1, vuint16m1_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_u8mf4))) +void vsuxseg4ei16(uint8_t * op0, vuint16mf2_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_u8mf4_m))) +void vsuxseg4ei16(vbool32_t op0, uint8_t * op1, vuint16mf2_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_u8mf8))) +void vsuxseg4ei16(uint8_t * op0, vuint16mf4_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_u8mf8_m))) +void vsuxseg4ei16(vbool64_t op0, uint8_t * op1, vuint16mf4_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_i8m1))) +void vsuxseg5ei16(int8_t * op0, vuint16m2_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_i8m1_m))) +void vsuxseg5ei16(vbool8_t op0, int8_t * op1, vuint16m2_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_i8mf2))) +void vsuxseg5ei16(int8_t * op0, vuint16m1_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_i8mf2_m))) +void vsuxseg5ei16(vbool16_t op0, int8_t * op1, vuint16m1_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_i8mf4))) +void vsuxseg5ei16(int8_t * op0, vuint16mf2_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_i8mf4_m))) +void vsuxseg5ei16(vbool32_t op0, int8_t * op1, vuint16mf2_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_i8mf8))) +void vsuxseg5ei16(int8_t * op0, vuint16mf4_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_i8mf8_m))) +void vsuxseg5ei16(vbool64_t op0, int8_t * op1, vuint16mf4_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_u8m1))) +void vsuxseg5ei16(uint8_t * op0, vuint16m2_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_u8m1_m))) +void vsuxseg5ei16(vbool8_t op0, uint8_t * op1, vuint16m2_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_u8mf2))) +void vsuxseg5ei16(uint8_t * op0, vuint16m1_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_u8mf2_m))) +void vsuxseg5ei16(vbool16_t op0, uint8_t * op1, vuint16m1_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_u8mf4))) +void vsuxseg5ei16(uint8_t * op0, vuint16mf2_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_u8mf4_m))) +void vsuxseg5ei16(vbool32_t op0, uint8_t * op1, vuint16mf2_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_u8mf8))) +void vsuxseg5ei16(uint8_t * op0, vuint16mf4_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_u8mf8_m))) +void vsuxseg5ei16(vbool64_t op0, uint8_t * op1, vuint16mf4_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_i8m1))) +void vsuxseg6ei16(int8_t * op0, vuint16m2_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_i8m1_m))) +void vsuxseg6ei16(vbool8_t op0, int8_t * op1, vuint16m2_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_i8mf2))) +void vsuxseg6ei16(int8_t * op0, vuint16m1_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_i8mf2_m))) +void vsuxseg6ei16(vbool16_t op0, int8_t * op1, vuint16m1_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_i8mf4))) +void vsuxseg6ei16(int8_t * op0, vuint16mf2_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_i8mf4_m))) +void vsuxseg6ei16(vbool32_t op0, int8_t * op1, vuint16mf2_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_i8mf8))) +void vsuxseg6ei16(int8_t * op0, vuint16mf4_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_i8mf8_m))) +void vsuxseg6ei16(vbool64_t op0, int8_t * op1, vuint16mf4_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_u8m1))) +void vsuxseg6ei16(uint8_t * op0, vuint16m2_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_u8m1_m))) +void vsuxseg6ei16(vbool8_t op0, uint8_t * op1, vuint16m2_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_u8mf2))) +void vsuxseg6ei16(uint8_t * op0, vuint16m1_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_u8mf2_m))) +void vsuxseg6ei16(vbool16_t op0, uint8_t * op1, vuint16m1_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_u8mf4))) +void vsuxseg6ei16(uint8_t * op0, vuint16mf2_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_u8mf4_m))) +void vsuxseg6ei16(vbool32_t op0, uint8_t * op1, vuint16mf2_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_u8mf8))) +void vsuxseg6ei16(uint8_t * op0, vuint16mf4_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_u8mf8_m))) +void vsuxseg6ei16(vbool64_t op0, uint8_t * op1, vuint16mf4_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_i8m1))) +void vsuxseg7ei16(int8_t * op0, vuint16m2_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_i8m1_m))) +void vsuxseg7ei16(vbool8_t op0, int8_t * op1, vuint16m2_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_i8mf2))) +void vsuxseg7ei16(int8_t * op0, vuint16m1_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_i8mf2_m))) +void vsuxseg7ei16(vbool16_t op0, int8_t * op1, vuint16m1_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_i8mf4))) +void vsuxseg7ei16(int8_t * op0, vuint16mf2_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_i8mf4_m))) +void vsuxseg7ei16(vbool32_t op0, int8_t * op1, vuint16mf2_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_i8mf8))) +void vsuxseg7ei16(int8_t * op0, vuint16mf4_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_i8mf8_m))) +void vsuxseg7ei16(vbool64_t op0, int8_t * op1, vuint16mf4_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_u8m1))) +void vsuxseg7ei16(uint8_t * op0, vuint16m2_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_u8m1_m))) +void vsuxseg7ei16(vbool8_t op0, uint8_t * op1, vuint16m2_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_u8mf2))) +void vsuxseg7ei16(uint8_t * op0, vuint16m1_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_u8mf2_m))) +void vsuxseg7ei16(vbool16_t op0, uint8_t * op1, vuint16m1_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_u8mf4))) +void vsuxseg7ei16(uint8_t * op0, vuint16mf2_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_u8mf4_m))) +void vsuxseg7ei16(vbool32_t op0, uint8_t * op1, vuint16mf2_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_u8mf8))) +void vsuxseg7ei16(uint8_t * op0, vuint16mf4_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_u8mf8_m))) +void vsuxseg7ei16(vbool64_t op0, uint8_t * op1, vuint16mf4_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_i8m1))) +void vsuxseg8ei16(int8_t * op0, vuint16m2_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_i8m1_m))) +void vsuxseg8ei16(vbool8_t op0, int8_t * op1, vuint16m2_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, vint8m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_i8mf2))) +void vsuxseg8ei16(int8_t * op0, vuint16m1_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_i8mf2_m))) +void vsuxseg8ei16(vbool16_t op0, int8_t * op1, vuint16m1_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, vint8mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_i8mf4))) +void vsuxseg8ei16(int8_t * op0, vuint16mf2_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_i8mf4_m))) +void vsuxseg8ei16(vbool32_t op0, int8_t * op1, vuint16mf2_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, vint8mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_i8mf8))) +void vsuxseg8ei16(int8_t * op0, vuint16mf4_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_i8mf8_m))) +void vsuxseg8ei16(vbool64_t op0, int8_t * op1, vuint16mf4_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, vint8mf8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_u8m1))) +void vsuxseg8ei16(uint8_t * op0, vuint16m2_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_u8m1_m))) +void vsuxseg8ei16(vbool8_t op0, uint8_t * op1, vuint16m2_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, vuint8m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_u8mf2))) +void vsuxseg8ei16(uint8_t * op0, vuint16m1_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_u8mf2_m))) +void vsuxseg8ei16(vbool16_t op0, uint8_t * op1, vuint16m1_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, vuint8mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_u8mf4))) +void vsuxseg8ei16(uint8_t * op0, vuint16mf2_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_u8mf4_m))) +void vsuxseg8ei16(vbool32_t op0, uint8_t * op1, vuint16mf2_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, vuint8mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_u8mf8))) +void vsuxseg8ei16(uint8_t * op0, vuint16mf4_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_u8mf8_m))) +void vsuxseg8ei16(vbool64_t op0, uint8_t * op1, vuint16mf4_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, vuint8mf8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_i8m1))) +void vsuxseg2ei32(int8_t * op0, vuint32m4_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_i8m1_m))) +void vsuxseg2ei32(vbool8_t op0, int8_t * op1, vuint32m4_t op2, vint8m1_t op3, vint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_i8m2))) +void vsuxseg2ei32(int8_t * op0, vuint32m8_t op1, vint8m2_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_i8m2_m))) +void vsuxseg2ei32(vbool4_t op0, int8_t * op1, vuint32m8_t op2, vint8m2_t op3, vint8m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_i8mf2))) +void vsuxseg2ei32(int8_t * op0, vuint32m2_t op1, vint8mf2_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_i8mf2_m))) +void vsuxseg2ei32(vbool16_t op0, int8_t * op1, vuint32m2_t op2, vint8mf2_t op3, vint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_i8mf4))) +void vsuxseg2ei32(int8_t * op0, vuint32m1_t op1, vint8mf4_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_i8mf4_m))) +void vsuxseg2ei32(vbool32_t op0, int8_t * op1, vuint32m1_t op2, vint8mf4_t op3, vint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_i8mf8))) +void vsuxseg2ei32(int8_t * op0, vuint32mf2_t op1, vint8mf8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_i8mf8_m))) +void vsuxseg2ei32(vbool64_t op0, int8_t * op1, vuint32mf2_t op2, vint8mf8_t op3, vint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_u8m1))) +void vsuxseg2ei32(uint8_t * op0, vuint32m4_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_u8m1_m))) +void vsuxseg2ei32(vbool8_t op0, uint8_t * op1, vuint32m4_t op2, vuint8m1_t op3, vuint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_u8m2))) +void vsuxseg2ei32(uint8_t * op0, vuint32m8_t op1, vuint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_u8m2_m))) +void vsuxseg2ei32(vbool4_t op0, uint8_t * op1, vuint32m8_t op2, vuint8m2_t op3, vuint8m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_u8mf2))) +void vsuxseg2ei32(uint8_t * op0, vuint32m2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_u8mf2_m))) +void vsuxseg2ei32(vbool16_t op0, uint8_t * op1, vuint32m2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_u8mf4))) +void vsuxseg2ei32(uint8_t * op0, vuint32m1_t op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_u8mf4_m))) +void vsuxseg2ei32(vbool32_t op0, uint8_t * op1, vuint32m1_t op2, vuint8mf4_t op3, vuint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_u8mf8))) +void vsuxseg2ei32(uint8_t * op0, vuint32mf2_t op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_u8mf8_m))) +void vsuxseg2ei32(vbool64_t op0, uint8_t * op1, vuint32mf2_t op2, vuint8mf8_t op3, vuint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_i8m1))) +void vsuxseg3ei32(int8_t * op0, vuint32m4_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_i8m1_m))) +void vsuxseg3ei32(vbool8_t op0, int8_t * op1, vuint32m4_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_i8m2))) +void vsuxseg3ei32(int8_t * op0, vuint32m8_t op1, vint8m2_t op2, vint8m2_t op3, vint8m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_i8m2_m))) +void vsuxseg3ei32(vbool4_t op0, int8_t * op1, vuint32m8_t op2, vint8m2_t op3, vint8m2_t op4, vint8m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_i8mf2))) +void vsuxseg3ei32(int8_t * op0, vuint32m2_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_i8mf2_m))) +void vsuxseg3ei32(vbool16_t op0, int8_t * op1, vuint32m2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_i8mf4))) +void vsuxseg3ei32(int8_t * op0, vuint32m1_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_i8mf4_m))) +void vsuxseg3ei32(vbool32_t op0, int8_t * op1, vuint32m1_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_i8mf8))) +void vsuxseg3ei32(int8_t * op0, vuint32mf2_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_i8mf8_m))) +void vsuxseg3ei32(vbool64_t op0, int8_t * op1, vuint32mf2_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_u8m1))) +void vsuxseg3ei32(uint8_t * op0, vuint32m4_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_u8m1_m))) +void vsuxseg3ei32(vbool8_t op0, uint8_t * op1, vuint32m4_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_u8m2))) +void vsuxseg3ei32(uint8_t * op0, vuint32m8_t op1, vuint8m2_t op2, vuint8m2_t op3, vuint8m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_u8m2_m))) +void vsuxseg3ei32(vbool4_t op0, uint8_t * op1, vuint32m8_t op2, vuint8m2_t op3, vuint8m2_t op4, vuint8m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_u8mf2))) +void vsuxseg3ei32(uint8_t * op0, vuint32m2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_u8mf2_m))) +void vsuxseg3ei32(vbool16_t op0, uint8_t * op1, vuint32m2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_u8mf4))) +void vsuxseg3ei32(uint8_t * op0, vuint32m1_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_u8mf4_m))) +void vsuxseg3ei32(vbool32_t op0, uint8_t * op1, vuint32m1_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_u8mf8))) +void vsuxseg3ei32(uint8_t * op0, vuint32mf2_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_u8mf8_m))) +void vsuxseg3ei32(vbool64_t op0, uint8_t * op1, vuint32mf2_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_i8m1))) +void vsuxseg4ei32(int8_t * op0, vuint32m4_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_i8m1_m))) +void vsuxseg4ei32(vbool8_t op0, int8_t * op1, vuint32m4_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_i8m2))) +void vsuxseg4ei32(int8_t * op0, vuint32m8_t op1, vint8m2_t op2, vint8m2_t op3, vint8m2_t op4, vint8m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_i8m2_m))) +void vsuxseg4ei32(vbool4_t op0, int8_t * op1, vuint32m8_t op2, vint8m2_t op3, vint8m2_t op4, vint8m2_t op5, vint8m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_i8mf2))) +void vsuxseg4ei32(int8_t * op0, vuint32m2_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_i8mf2_m))) +void vsuxseg4ei32(vbool16_t op0, int8_t * op1, vuint32m2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_i8mf4))) +void vsuxseg4ei32(int8_t * op0, vuint32m1_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_i8mf4_m))) +void vsuxseg4ei32(vbool32_t op0, int8_t * op1, vuint32m1_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_i8mf8))) +void vsuxseg4ei32(int8_t * op0, vuint32mf2_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_i8mf8_m))) +void vsuxseg4ei32(vbool64_t op0, int8_t * op1, vuint32mf2_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_u8m1))) +void vsuxseg4ei32(uint8_t * op0, vuint32m4_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_u8m1_m))) +void vsuxseg4ei32(vbool8_t op0, uint8_t * op1, vuint32m4_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_u8m2))) +void vsuxseg4ei32(uint8_t * op0, vuint32m8_t op1, vuint8m2_t op2, vuint8m2_t op3, vuint8m2_t op4, vuint8m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_u8m2_m))) +void vsuxseg4ei32(vbool4_t op0, uint8_t * op1, vuint32m8_t op2, vuint8m2_t op3, vuint8m2_t op4, vuint8m2_t op5, vuint8m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_u8mf2))) +void vsuxseg4ei32(uint8_t * op0, vuint32m2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_u8mf2_m))) +void vsuxseg4ei32(vbool16_t op0, uint8_t * op1, vuint32m2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_u8mf4))) +void vsuxseg4ei32(uint8_t * op0, vuint32m1_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_u8mf4_m))) +void vsuxseg4ei32(vbool32_t op0, uint8_t * op1, vuint32m1_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_u8mf8))) +void vsuxseg4ei32(uint8_t * op0, vuint32mf2_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_u8mf8_m))) +void vsuxseg4ei32(vbool64_t op0, uint8_t * op1, vuint32mf2_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_i8m1))) +void vsuxseg5ei32(int8_t * op0, vuint32m4_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_i8m1_m))) +void vsuxseg5ei32(vbool8_t op0, int8_t * op1, vuint32m4_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_i8mf2))) +void vsuxseg5ei32(int8_t * op0, vuint32m2_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_i8mf2_m))) +void vsuxseg5ei32(vbool16_t op0, int8_t * op1, vuint32m2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_i8mf4))) +void vsuxseg5ei32(int8_t * op0, vuint32m1_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_i8mf4_m))) +void vsuxseg5ei32(vbool32_t op0, int8_t * op1, vuint32m1_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_i8mf8))) +void vsuxseg5ei32(int8_t * op0, vuint32mf2_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_i8mf8_m))) +void vsuxseg5ei32(vbool64_t op0, int8_t * op1, vuint32mf2_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_u8m1))) +void vsuxseg5ei32(uint8_t * op0, vuint32m4_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_u8m1_m))) +void vsuxseg5ei32(vbool8_t op0, uint8_t * op1, vuint32m4_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_u8mf2))) +void vsuxseg5ei32(uint8_t * op0, vuint32m2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_u8mf2_m))) +void vsuxseg5ei32(vbool16_t op0, uint8_t * op1, vuint32m2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_u8mf4))) +void vsuxseg5ei32(uint8_t * op0, vuint32m1_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_u8mf4_m))) +void vsuxseg5ei32(vbool32_t op0, uint8_t * op1, vuint32m1_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_u8mf8))) +void vsuxseg5ei32(uint8_t * op0, vuint32mf2_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_u8mf8_m))) +void vsuxseg5ei32(vbool64_t op0, uint8_t * op1, vuint32mf2_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_i8m1))) +void vsuxseg6ei32(int8_t * op0, vuint32m4_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_i8m1_m))) +void vsuxseg6ei32(vbool8_t op0, int8_t * op1, vuint32m4_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_i8mf2))) +void vsuxseg6ei32(int8_t * op0, vuint32m2_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_i8mf2_m))) +void vsuxseg6ei32(vbool16_t op0, int8_t * op1, vuint32m2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_i8mf4))) +void vsuxseg6ei32(int8_t * op0, vuint32m1_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_i8mf4_m))) +void vsuxseg6ei32(vbool32_t op0, int8_t * op1, vuint32m1_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_i8mf8))) +void vsuxseg6ei32(int8_t * op0, vuint32mf2_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_i8mf8_m))) +void vsuxseg6ei32(vbool64_t op0, int8_t * op1, vuint32mf2_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_u8m1))) +void vsuxseg6ei32(uint8_t * op0, vuint32m4_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_u8m1_m))) +void vsuxseg6ei32(vbool8_t op0, uint8_t * op1, vuint32m4_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_u8mf2))) +void vsuxseg6ei32(uint8_t * op0, vuint32m2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_u8mf2_m))) +void vsuxseg6ei32(vbool16_t op0, uint8_t * op1, vuint32m2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_u8mf4))) +void vsuxseg6ei32(uint8_t * op0, vuint32m1_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_u8mf4_m))) +void vsuxseg6ei32(vbool32_t op0, uint8_t * op1, vuint32m1_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_u8mf8))) +void vsuxseg6ei32(uint8_t * op0, vuint32mf2_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_u8mf8_m))) +void vsuxseg6ei32(vbool64_t op0, uint8_t * op1, vuint32mf2_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_i8m1))) +void vsuxseg7ei32(int8_t * op0, vuint32m4_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_i8m1_m))) +void vsuxseg7ei32(vbool8_t op0, int8_t * op1, vuint32m4_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_i8mf2))) +void vsuxseg7ei32(int8_t * op0, vuint32m2_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_i8mf2_m))) +void vsuxseg7ei32(vbool16_t op0, int8_t * op1, vuint32m2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_i8mf4))) +void vsuxseg7ei32(int8_t * op0, vuint32m1_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_i8mf4_m))) +void vsuxseg7ei32(vbool32_t op0, int8_t * op1, vuint32m1_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_i8mf8))) +void vsuxseg7ei32(int8_t * op0, vuint32mf2_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_i8mf8_m))) +void vsuxseg7ei32(vbool64_t op0, int8_t * op1, vuint32mf2_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_u8m1))) +void vsuxseg7ei32(uint8_t * op0, vuint32m4_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_u8m1_m))) +void vsuxseg7ei32(vbool8_t op0, uint8_t * op1, vuint32m4_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_u8mf2))) +void vsuxseg7ei32(uint8_t * op0, vuint32m2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_u8mf2_m))) +void vsuxseg7ei32(vbool16_t op0, uint8_t * op1, vuint32m2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_u8mf4))) +void vsuxseg7ei32(uint8_t * op0, vuint32m1_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_u8mf4_m))) +void vsuxseg7ei32(vbool32_t op0, uint8_t * op1, vuint32m1_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_u8mf8))) +void vsuxseg7ei32(uint8_t * op0, vuint32mf2_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_u8mf8_m))) +void vsuxseg7ei32(vbool64_t op0, uint8_t * op1, vuint32mf2_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_i8m1))) +void vsuxseg8ei32(int8_t * op0, vuint32m4_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_i8m1_m))) +void vsuxseg8ei32(vbool8_t op0, int8_t * op1, vuint32m4_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, vint8m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_i8mf2))) +void vsuxseg8ei32(int8_t * op0, vuint32m2_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_i8mf2_m))) +void vsuxseg8ei32(vbool16_t op0, int8_t * op1, vuint32m2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, vint8mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_i8mf4))) +void vsuxseg8ei32(int8_t * op0, vuint32m1_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_i8mf4_m))) +void vsuxseg8ei32(vbool32_t op0, int8_t * op1, vuint32m1_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, vint8mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_i8mf8))) +void vsuxseg8ei32(int8_t * op0, vuint32mf2_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_i8mf8_m))) +void vsuxseg8ei32(vbool64_t op0, int8_t * op1, vuint32mf2_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, vint8mf8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_u8m1))) +void vsuxseg8ei32(uint8_t * op0, vuint32m4_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_u8m1_m))) +void vsuxseg8ei32(vbool8_t op0, uint8_t * op1, vuint32m4_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, vuint8m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_u8mf2))) +void vsuxseg8ei32(uint8_t * op0, vuint32m2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_u8mf2_m))) +void vsuxseg8ei32(vbool16_t op0, uint8_t * op1, vuint32m2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, vuint8mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_u8mf4))) +void vsuxseg8ei32(uint8_t * op0, vuint32m1_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_u8mf4_m))) +void vsuxseg8ei32(vbool32_t op0, uint8_t * op1, vuint32m1_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, vuint8mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_u8mf8))) +void vsuxseg8ei32(uint8_t * op0, vuint32mf2_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_u8mf8_m))) +void vsuxseg8ei32(vbool64_t op0, uint8_t * op1, vuint32mf2_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, vuint8mf8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_i8m1))) +void vsuxseg2ei64(int8_t * op0, vuint64m8_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_i8m1_m))) +void vsuxseg2ei64(vbool8_t op0, int8_t * op1, vuint64m8_t op2, vint8m1_t op3, vint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_i8mf2))) +void vsuxseg2ei64(int8_t * op0, vuint64m4_t op1, vint8mf2_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_i8mf2_m))) +void vsuxseg2ei64(vbool16_t op0, int8_t * op1, vuint64m4_t op2, vint8mf2_t op3, vint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_i8mf4))) +void vsuxseg2ei64(int8_t * op0, vuint64m2_t op1, vint8mf4_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_i8mf4_m))) +void vsuxseg2ei64(vbool32_t op0, int8_t * op1, vuint64m2_t op2, vint8mf4_t op3, vint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_i8mf8))) +void vsuxseg2ei64(int8_t * op0, vuint64m1_t op1, vint8mf8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_i8mf8_m))) +void vsuxseg2ei64(vbool64_t op0, int8_t * op1, vuint64m1_t op2, vint8mf8_t op3, vint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_u8m1))) +void vsuxseg2ei64(uint8_t * op0, vuint64m8_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_u8m1_m))) +void vsuxseg2ei64(vbool8_t op0, uint8_t * op1, vuint64m8_t op2, vuint8m1_t op3, vuint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_u8mf2))) +void vsuxseg2ei64(uint8_t * op0, vuint64m4_t op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_u8mf2_m))) +void vsuxseg2ei64(vbool16_t op0, uint8_t * op1, vuint64m4_t op2, vuint8mf2_t op3, vuint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_u8mf4))) +void vsuxseg2ei64(uint8_t * op0, vuint64m2_t op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_u8mf4_m))) +void vsuxseg2ei64(vbool32_t op0, uint8_t * op1, vuint64m2_t op2, vuint8mf4_t op3, vuint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_u8mf8))) +void vsuxseg2ei64(uint8_t * op0, vuint64m1_t op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_u8mf8_m))) +void vsuxseg2ei64(vbool64_t op0, uint8_t * op1, vuint64m1_t op2, vuint8mf8_t op3, vuint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_i8m1))) +void vsuxseg3ei64(int8_t * op0, vuint64m8_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_i8m1_m))) +void vsuxseg3ei64(vbool8_t op0, int8_t * op1, vuint64m8_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_i8mf2))) +void vsuxseg3ei64(int8_t * op0, vuint64m4_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_i8mf2_m))) +void vsuxseg3ei64(vbool16_t op0, int8_t * op1, vuint64m4_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_i8mf4))) +void vsuxseg3ei64(int8_t * op0, vuint64m2_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_i8mf4_m))) +void vsuxseg3ei64(vbool32_t op0, int8_t * op1, vuint64m2_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_i8mf8))) +void vsuxseg3ei64(int8_t * op0, vuint64m1_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_i8mf8_m))) +void vsuxseg3ei64(vbool64_t op0, int8_t * op1, vuint64m1_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_u8m1))) +void vsuxseg3ei64(uint8_t * op0, vuint64m8_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_u8m1_m))) +void vsuxseg3ei64(vbool8_t op0, uint8_t * op1, vuint64m8_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_u8mf2))) +void vsuxseg3ei64(uint8_t * op0, vuint64m4_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_u8mf2_m))) +void vsuxseg3ei64(vbool16_t op0, uint8_t * op1, vuint64m4_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_u8mf4))) +void vsuxseg3ei64(uint8_t * op0, vuint64m2_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_u8mf4_m))) +void vsuxseg3ei64(vbool32_t op0, uint8_t * op1, vuint64m2_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_u8mf8))) +void vsuxseg3ei64(uint8_t * op0, vuint64m1_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_u8mf8_m))) +void vsuxseg3ei64(vbool64_t op0, uint8_t * op1, vuint64m1_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_i8m1))) +void vsuxseg4ei64(int8_t * op0, vuint64m8_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_i8m1_m))) +void vsuxseg4ei64(vbool8_t op0, int8_t * op1, vuint64m8_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_i8mf2))) +void vsuxseg4ei64(int8_t * op0, vuint64m4_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_i8mf2_m))) +void vsuxseg4ei64(vbool16_t op0, int8_t * op1, vuint64m4_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_i8mf4))) +void vsuxseg4ei64(int8_t * op0, vuint64m2_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_i8mf4_m))) +void vsuxseg4ei64(vbool32_t op0, int8_t * op1, vuint64m2_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_i8mf8))) +void vsuxseg4ei64(int8_t * op0, vuint64m1_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_i8mf8_m))) +void vsuxseg4ei64(vbool64_t op0, int8_t * op1, vuint64m1_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_u8m1))) +void vsuxseg4ei64(uint8_t * op0, vuint64m8_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_u8m1_m))) +void vsuxseg4ei64(vbool8_t op0, uint8_t * op1, vuint64m8_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_u8mf2))) +void vsuxseg4ei64(uint8_t * op0, vuint64m4_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_u8mf2_m))) +void vsuxseg4ei64(vbool16_t op0, uint8_t * op1, vuint64m4_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_u8mf4))) +void vsuxseg4ei64(uint8_t * op0, vuint64m2_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_u8mf4_m))) +void vsuxseg4ei64(vbool32_t op0, uint8_t * op1, vuint64m2_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_u8mf8))) +void vsuxseg4ei64(uint8_t * op0, vuint64m1_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_u8mf8_m))) +void vsuxseg4ei64(vbool64_t op0, uint8_t * op1, vuint64m1_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_i8m1))) +void vsuxseg5ei64(int8_t * op0, vuint64m8_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_i8m1_m))) +void vsuxseg5ei64(vbool8_t op0, int8_t * op1, vuint64m8_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_i8mf2))) +void vsuxseg5ei64(int8_t * op0, vuint64m4_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_i8mf2_m))) +void vsuxseg5ei64(vbool16_t op0, int8_t * op1, vuint64m4_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_i8mf4))) +void vsuxseg5ei64(int8_t * op0, vuint64m2_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_i8mf4_m))) +void vsuxseg5ei64(vbool32_t op0, int8_t * op1, vuint64m2_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_i8mf8))) +void vsuxseg5ei64(int8_t * op0, vuint64m1_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_i8mf8_m))) +void vsuxseg5ei64(vbool64_t op0, int8_t * op1, vuint64m1_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_u8m1))) +void vsuxseg5ei64(uint8_t * op0, vuint64m8_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_u8m1_m))) +void vsuxseg5ei64(vbool8_t op0, uint8_t * op1, vuint64m8_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_u8mf2))) +void vsuxseg5ei64(uint8_t * op0, vuint64m4_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_u8mf2_m))) +void vsuxseg5ei64(vbool16_t op0, uint8_t * op1, vuint64m4_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_u8mf4))) +void vsuxseg5ei64(uint8_t * op0, vuint64m2_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_u8mf4_m))) +void vsuxseg5ei64(vbool32_t op0, uint8_t * op1, vuint64m2_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_u8mf8))) +void vsuxseg5ei64(uint8_t * op0, vuint64m1_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_u8mf8_m))) +void vsuxseg5ei64(vbool64_t op0, uint8_t * op1, vuint64m1_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_i8m1))) +void vsuxseg6ei64(int8_t * op0, vuint64m8_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_i8m1_m))) +void vsuxseg6ei64(vbool8_t op0, int8_t * op1, vuint64m8_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_i8mf2))) +void vsuxseg6ei64(int8_t * op0, vuint64m4_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_i8mf2_m))) +void vsuxseg6ei64(vbool16_t op0, int8_t * op1, vuint64m4_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_i8mf4))) +void vsuxseg6ei64(int8_t * op0, vuint64m2_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_i8mf4_m))) +void vsuxseg6ei64(vbool32_t op0, int8_t * op1, vuint64m2_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_i8mf8))) +void vsuxseg6ei64(int8_t * op0, vuint64m1_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_i8mf8_m))) +void vsuxseg6ei64(vbool64_t op0, int8_t * op1, vuint64m1_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_u8m1))) +void vsuxseg6ei64(uint8_t * op0, vuint64m8_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_u8m1_m))) +void vsuxseg6ei64(vbool8_t op0, uint8_t * op1, vuint64m8_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_u8mf2))) +void vsuxseg6ei64(uint8_t * op0, vuint64m4_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_u8mf2_m))) +void vsuxseg6ei64(vbool16_t op0, uint8_t * op1, vuint64m4_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_u8mf4))) +void vsuxseg6ei64(uint8_t * op0, vuint64m2_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_u8mf4_m))) +void vsuxseg6ei64(vbool32_t op0, uint8_t * op1, vuint64m2_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_u8mf8))) +void vsuxseg6ei64(uint8_t * op0, vuint64m1_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_u8mf8_m))) +void vsuxseg6ei64(vbool64_t op0, uint8_t * op1, vuint64m1_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_i8m1))) +void vsuxseg7ei64(int8_t * op0, vuint64m8_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_i8m1_m))) +void vsuxseg7ei64(vbool8_t op0, int8_t * op1, vuint64m8_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_i8mf2))) +void vsuxseg7ei64(int8_t * op0, vuint64m4_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_i8mf2_m))) +void vsuxseg7ei64(vbool16_t op0, int8_t * op1, vuint64m4_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_i8mf4))) +void vsuxseg7ei64(int8_t * op0, vuint64m2_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_i8mf4_m))) +void vsuxseg7ei64(vbool32_t op0, int8_t * op1, vuint64m2_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_i8mf8))) +void vsuxseg7ei64(int8_t * op0, vuint64m1_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_i8mf8_m))) +void vsuxseg7ei64(vbool64_t op0, int8_t * op1, vuint64m1_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_u8m1))) +void vsuxseg7ei64(uint8_t * op0, vuint64m8_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_u8m1_m))) +void vsuxseg7ei64(vbool8_t op0, uint8_t * op1, vuint64m8_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_u8mf2))) +void vsuxseg7ei64(uint8_t * op0, vuint64m4_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_u8mf2_m))) +void vsuxseg7ei64(vbool16_t op0, uint8_t * op1, vuint64m4_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_u8mf4))) +void vsuxseg7ei64(uint8_t * op0, vuint64m2_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_u8mf4_m))) +void vsuxseg7ei64(vbool32_t op0, uint8_t * op1, vuint64m2_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_u8mf8))) +void vsuxseg7ei64(uint8_t * op0, vuint64m1_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_u8mf8_m))) +void vsuxseg7ei64(vbool64_t op0, uint8_t * op1, vuint64m1_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_i8m1))) +void vsuxseg8ei64(int8_t * op0, vuint64m8_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_i8m1_m))) +void vsuxseg8ei64(vbool8_t op0, int8_t * op1, vuint64m8_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, vint8m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_i8mf2))) +void vsuxseg8ei64(int8_t * op0, vuint64m4_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_i8mf2_m))) +void vsuxseg8ei64(vbool16_t op0, int8_t * op1, vuint64m4_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, vint8mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_i8mf4))) +void vsuxseg8ei64(int8_t * op0, vuint64m2_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_i8mf4_m))) +void vsuxseg8ei64(vbool32_t op0, int8_t * op1, vuint64m2_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, vint8mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_i8mf8))) +void vsuxseg8ei64(int8_t * op0, vuint64m1_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_i8mf8_m))) +void vsuxseg8ei64(vbool64_t op0, int8_t * op1, vuint64m1_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, vint8mf8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_u8m1))) +void vsuxseg8ei64(uint8_t * op0, vuint64m8_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_u8m1_m))) +void vsuxseg8ei64(vbool8_t op0, uint8_t * op1, vuint64m8_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, vuint8m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_u8mf2))) +void vsuxseg8ei64(uint8_t * op0, vuint64m4_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_u8mf2_m))) +void vsuxseg8ei64(vbool16_t op0, uint8_t * op1, vuint64m4_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, vuint8mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_u8mf4))) +void vsuxseg8ei64(uint8_t * op0, vuint64m2_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_u8mf4_m))) +void vsuxseg8ei64(vbool32_t op0, uint8_t * op1, vuint64m2_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, vuint8mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_u8mf8))) +void vsuxseg8ei64(uint8_t * op0, vuint64m1_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_u8mf8_m))) +void vsuxseg8ei64(vbool64_t op0, uint8_t * op1, vuint64m1_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, vuint8mf8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_i16m1))) +void vsuxseg2ei8(int16_t * op0, vuint8mf2_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_i16m1_m))) +void vsuxseg2ei8(vbool16_t op0, int16_t * op1, vuint8mf2_t op2, vint16m1_t op3, vint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_i16m2))) +void vsuxseg2ei8(int16_t * op0, vuint8m1_t op1, vint16m2_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_i16m2_m))) +void vsuxseg2ei8(vbool8_t op0, int16_t * op1, vuint8m1_t op2, vint16m2_t op3, vint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_i16m4))) +void vsuxseg2ei8(int16_t * op0, vuint8m2_t op1, vint16m4_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_i16m4_m))) +void vsuxseg2ei8(vbool4_t op0, int16_t * op1, vuint8m2_t op2, vint16m4_t op3, vint16m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_i16mf2))) +void vsuxseg2ei8(int16_t * op0, vuint8mf4_t op1, vint16mf2_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_i16mf2_m))) +void vsuxseg2ei8(vbool32_t op0, int16_t * op1, vuint8mf4_t op2, vint16mf2_t op3, vint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_i16mf4))) +void vsuxseg2ei8(int16_t * op0, vuint8mf8_t op1, vint16mf4_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_i16mf4_m))) +void vsuxseg2ei8(vbool64_t op0, int16_t * op1, vuint8mf8_t op2, vint16mf4_t op3, vint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_u16m1))) +void vsuxseg2ei8(uint16_t * op0, vuint8mf2_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_u16m1_m))) +void vsuxseg2ei8(vbool16_t op0, uint16_t * op1, vuint8mf2_t op2, vuint16m1_t op3, vuint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_u16m2))) +void vsuxseg2ei8(uint16_t * op0, vuint8m1_t op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_u16m2_m))) +void vsuxseg2ei8(vbool8_t op0, uint16_t * op1, vuint8m1_t op2, vuint16m2_t op3, vuint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_u16m4))) +void vsuxseg2ei8(uint16_t * op0, vuint8m2_t op1, vuint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_u16m4_m))) +void vsuxseg2ei8(vbool4_t op0, uint16_t * op1, vuint8m2_t op2, vuint16m4_t op3, vuint16m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_u16mf2))) +void vsuxseg2ei8(uint16_t * op0, vuint8mf4_t op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_u16mf2_m))) +void vsuxseg2ei8(vbool32_t op0, uint16_t * op1, vuint8mf4_t op2, vuint16mf2_t op3, vuint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_u16mf4))) +void vsuxseg2ei8(uint16_t * op0, vuint8mf8_t op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_u16mf4_m))) +void vsuxseg2ei8(vbool64_t op0, uint16_t * op1, vuint8mf8_t op2, vuint16mf4_t op3, vuint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_i16m1))) +void vsuxseg3ei8(int16_t * op0, vuint8mf2_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_i16m1_m))) +void vsuxseg3ei8(vbool16_t op0, int16_t * op1, vuint8mf2_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_i16m2))) +void vsuxseg3ei8(int16_t * op0, vuint8m1_t op1, vint16m2_t op2, vint16m2_t op3, vint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_i16m2_m))) +void vsuxseg3ei8(vbool8_t op0, int16_t * op1, vuint8m1_t op2, vint16m2_t op3, vint16m2_t op4, vint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_i16mf2))) +void vsuxseg3ei8(int16_t * op0, vuint8mf4_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_i16mf2_m))) +void vsuxseg3ei8(vbool32_t op0, int16_t * op1, vuint8mf4_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_i16mf4))) +void vsuxseg3ei8(int16_t * op0, vuint8mf8_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_i16mf4_m))) +void vsuxseg3ei8(vbool64_t op0, int16_t * op1, vuint8mf8_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_u16m1))) +void vsuxseg3ei8(uint16_t * op0, vuint8mf2_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_u16m1_m))) +void vsuxseg3ei8(vbool16_t op0, uint16_t * op1, vuint8mf2_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_u16m2))) +void vsuxseg3ei8(uint16_t * op0, vuint8m1_t op1, vuint16m2_t op2, vuint16m2_t op3, vuint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_u16m2_m))) +void vsuxseg3ei8(vbool8_t op0, uint16_t * op1, vuint8m1_t op2, vuint16m2_t op3, vuint16m2_t op4, vuint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_u16mf2))) +void vsuxseg3ei8(uint16_t * op0, vuint8mf4_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_u16mf2_m))) +void vsuxseg3ei8(vbool32_t op0, uint16_t * op1, vuint8mf4_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_u16mf4))) +void vsuxseg3ei8(uint16_t * op0, vuint8mf8_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_u16mf4_m))) +void vsuxseg3ei8(vbool64_t op0, uint16_t * op1, vuint8mf8_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_i16m1))) +void vsuxseg4ei8(int16_t * op0, vuint8mf2_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_i16m1_m))) +void vsuxseg4ei8(vbool16_t op0, int16_t * op1, vuint8mf2_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_i16m2))) +void vsuxseg4ei8(int16_t * op0, vuint8m1_t op1, vint16m2_t op2, vint16m2_t op3, vint16m2_t op4, vint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_i16m2_m))) +void vsuxseg4ei8(vbool8_t op0, int16_t * op1, vuint8m1_t op2, vint16m2_t op3, vint16m2_t op4, vint16m2_t op5, vint16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_i16mf2))) +void vsuxseg4ei8(int16_t * op0, vuint8mf4_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_i16mf2_m))) +void vsuxseg4ei8(vbool32_t op0, int16_t * op1, vuint8mf4_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_i16mf4))) +void vsuxseg4ei8(int16_t * op0, vuint8mf8_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_i16mf4_m))) +void vsuxseg4ei8(vbool64_t op0, int16_t * op1, vuint8mf8_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_u16m1))) +void vsuxseg4ei8(uint16_t * op0, vuint8mf2_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_u16m1_m))) +void vsuxseg4ei8(vbool16_t op0, uint16_t * op1, vuint8mf2_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_u16m2))) +void vsuxseg4ei8(uint16_t * op0, vuint8m1_t op1, vuint16m2_t op2, vuint16m2_t op3, vuint16m2_t op4, vuint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_u16m2_m))) +void vsuxseg4ei8(vbool8_t op0, uint16_t * op1, vuint8m1_t op2, vuint16m2_t op3, vuint16m2_t op4, vuint16m2_t op5, vuint16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_u16mf2))) +void vsuxseg4ei8(uint16_t * op0, vuint8mf4_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_u16mf2_m))) +void vsuxseg4ei8(vbool32_t op0, uint16_t * op1, vuint8mf4_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_u16mf4))) +void vsuxseg4ei8(uint16_t * op0, vuint8mf8_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_u16mf4_m))) +void vsuxseg4ei8(vbool64_t op0, uint16_t * op1, vuint8mf8_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_i16m1))) +void vsuxseg5ei8(int16_t * op0, vuint8mf2_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_i16m1_m))) +void vsuxseg5ei8(vbool16_t op0, int16_t * op1, vuint8mf2_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_i16mf2))) +void vsuxseg5ei8(int16_t * op0, vuint8mf4_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_i16mf2_m))) +void vsuxseg5ei8(vbool32_t op0, int16_t * op1, vuint8mf4_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_i16mf4))) +void vsuxseg5ei8(int16_t * op0, vuint8mf8_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_i16mf4_m))) +void vsuxseg5ei8(vbool64_t op0, int16_t * op1, vuint8mf8_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_u16m1))) +void vsuxseg5ei8(uint16_t * op0, vuint8mf2_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_u16m1_m))) +void vsuxseg5ei8(vbool16_t op0, uint16_t * op1, vuint8mf2_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_u16mf2))) +void vsuxseg5ei8(uint16_t * op0, vuint8mf4_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_u16mf2_m))) +void vsuxseg5ei8(vbool32_t op0, uint16_t * op1, vuint8mf4_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_u16mf4))) +void vsuxseg5ei8(uint16_t * op0, vuint8mf8_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_u16mf4_m))) +void vsuxseg5ei8(vbool64_t op0, uint16_t * op1, vuint8mf8_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_i16m1))) +void vsuxseg6ei8(int16_t * op0, vuint8mf2_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_i16m1_m))) +void vsuxseg6ei8(vbool16_t op0, int16_t * op1, vuint8mf2_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_i16mf2))) +void vsuxseg6ei8(int16_t * op0, vuint8mf4_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_i16mf2_m))) +void vsuxseg6ei8(vbool32_t op0, int16_t * op1, vuint8mf4_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_i16mf4))) +void vsuxseg6ei8(int16_t * op0, vuint8mf8_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_i16mf4_m))) +void vsuxseg6ei8(vbool64_t op0, int16_t * op1, vuint8mf8_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_u16m1))) +void vsuxseg6ei8(uint16_t * op0, vuint8mf2_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_u16m1_m))) +void vsuxseg6ei8(vbool16_t op0, uint16_t * op1, vuint8mf2_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_u16mf2))) +void vsuxseg6ei8(uint16_t * op0, vuint8mf4_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_u16mf2_m))) +void vsuxseg6ei8(vbool32_t op0, uint16_t * op1, vuint8mf4_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_u16mf4))) +void vsuxseg6ei8(uint16_t * op0, vuint8mf8_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_u16mf4_m))) +void vsuxseg6ei8(vbool64_t op0, uint16_t * op1, vuint8mf8_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_i16m1))) +void vsuxseg7ei8(int16_t * op0, vuint8mf2_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_i16m1_m))) +void vsuxseg7ei8(vbool16_t op0, int16_t * op1, vuint8mf2_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_i16mf2))) +void vsuxseg7ei8(int16_t * op0, vuint8mf4_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_i16mf2_m))) +void vsuxseg7ei8(vbool32_t op0, int16_t * op1, vuint8mf4_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_i16mf4))) +void vsuxseg7ei8(int16_t * op0, vuint8mf8_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_i16mf4_m))) +void vsuxseg7ei8(vbool64_t op0, int16_t * op1, vuint8mf8_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_u16m1))) +void vsuxseg7ei8(uint16_t * op0, vuint8mf2_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_u16m1_m))) +void vsuxseg7ei8(vbool16_t op0, uint16_t * op1, vuint8mf2_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_u16mf2))) +void vsuxseg7ei8(uint16_t * op0, vuint8mf4_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_u16mf2_m))) +void vsuxseg7ei8(vbool32_t op0, uint16_t * op1, vuint8mf4_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_u16mf4))) +void vsuxseg7ei8(uint16_t * op0, vuint8mf8_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_u16mf4_m))) +void vsuxseg7ei8(vbool64_t op0, uint16_t * op1, vuint8mf8_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_i16m1))) +void vsuxseg8ei8(int16_t * op0, vuint8mf2_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_i16m1_m))) +void vsuxseg8ei8(vbool16_t op0, int16_t * op1, vuint8mf2_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, vint16m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_i16mf2))) +void vsuxseg8ei8(int16_t * op0, vuint8mf4_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_i16mf2_m))) +void vsuxseg8ei8(vbool32_t op0, int16_t * op1, vuint8mf4_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, vint16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_i16mf4))) +void vsuxseg8ei8(int16_t * op0, vuint8mf8_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_i16mf4_m))) +void vsuxseg8ei8(vbool64_t op0, int16_t * op1, vuint8mf8_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, vint16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_u16m1))) +void vsuxseg8ei8(uint16_t * op0, vuint8mf2_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_u16m1_m))) +void vsuxseg8ei8(vbool16_t op0, uint16_t * op1, vuint8mf2_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, vuint16m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_u16mf2))) +void vsuxseg8ei8(uint16_t * op0, vuint8mf4_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_u16mf2_m))) +void vsuxseg8ei8(vbool32_t op0, uint16_t * op1, vuint8mf4_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, vuint16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_u16mf4))) +void vsuxseg8ei8(uint16_t * op0, vuint8mf8_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_u16mf4_m))) +void vsuxseg8ei8(vbool64_t op0, uint16_t * op1, vuint8mf8_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, vuint16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_i16m1))) +void vsuxseg2ei16(int16_t * op0, vuint16m1_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_i16m1_m))) +void vsuxseg2ei16(vbool16_t op0, int16_t * op1, vuint16m1_t op2, vint16m1_t op3, vint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_i16m2))) +void vsuxseg2ei16(int16_t * op0, vuint16m2_t op1, vint16m2_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_i16m2_m))) +void vsuxseg2ei16(vbool8_t op0, int16_t * op1, vuint16m2_t op2, vint16m2_t op3, vint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_i16m4))) +void vsuxseg2ei16(int16_t * op0, vuint16m4_t op1, vint16m4_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_i16m4_m))) +void vsuxseg2ei16(vbool4_t op0, int16_t * op1, vuint16m4_t op2, vint16m4_t op3, vint16m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_i16mf2))) +void vsuxseg2ei16(int16_t * op0, vuint16mf2_t op1, vint16mf2_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_i16mf2_m))) +void vsuxseg2ei16(vbool32_t op0, int16_t * op1, vuint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_i16mf4))) +void vsuxseg2ei16(int16_t * op0, vuint16mf4_t op1, vint16mf4_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_i16mf4_m))) +void vsuxseg2ei16(vbool64_t op0, int16_t * op1, vuint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_u16m1))) +void vsuxseg2ei16(uint16_t * op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_u16m1_m))) +void vsuxseg2ei16(vbool16_t op0, uint16_t * op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_u16m2))) +void vsuxseg2ei16(uint16_t * op0, vuint16m2_t op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_u16m2_m))) +void vsuxseg2ei16(vbool8_t op0, uint16_t * op1, vuint16m2_t op2, vuint16m2_t op3, vuint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_u16m4))) +void vsuxseg2ei16(uint16_t * op0, vuint16m4_t op1, vuint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_u16m4_m))) +void vsuxseg2ei16(vbool4_t op0, uint16_t * op1, vuint16m4_t op2, vuint16m4_t op3, vuint16m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_u16mf2))) +void vsuxseg2ei16(uint16_t * op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_u16mf2_m))) +void vsuxseg2ei16(vbool32_t op0, uint16_t * op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_u16mf4))) +void vsuxseg2ei16(uint16_t * op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_u16mf4_m))) +void vsuxseg2ei16(vbool64_t op0, uint16_t * op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_i16m1))) +void vsuxseg3ei16(int16_t * op0, vuint16m1_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_i16m1_m))) +void vsuxseg3ei16(vbool16_t op0, int16_t * op1, vuint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_i16m2))) +void vsuxseg3ei16(int16_t * op0, vuint16m2_t op1, vint16m2_t op2, vint16m2_t op3, vint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_i16m2_m))) +void vsuxseg3ei16(vbool8_t op0, int16_t * op1, vuint16m2_t op2, vint16m2_t op3, vint16m2_t op4, vint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_i16mf2))) +void vsuxseg3ei16(int16_t * op0, vuint16mf2_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_i16mf2_m))) +void vsuxseg3ei16(vbool32_t op0, int16_t * op1, vuint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_i16mf4))) +void vsuxseg3ei16(int16_t * op0, vuint16mf4_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_i16mf4_m))) +void vsuxseg3ei16(vbool64_t op0, int16_t * op1, vuint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_u16m1))) +void vsuxseg3ei16(uint16_t * op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_u16m1_m))) +void vsuxseg3ei16(vbool16_t op0, uint16_t * op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_u16m2))) +void vsuxseg3ei16(uint16_t * op0, vuint16m2_t op1, vuint16m2_t op2, vuint16m2_t op3, vuint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_u16m2_m))) +void vsuxseg3ei16(vbool8_t op0, uint16_t * op1, vuint16m2_t op2, vuint16m2_t op3, vuint16m2_t op4, vuint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_u16mf2))) +void vsuxseg3ei16(uint16_t * op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_u16mf2_m))) +void vsuxseg3ei16(vbool32_t op0, uint16_t * op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_u16mf4))) +void vsuxseg3ei16(uint16_t * op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_u16mf4_m))) +void vsuxseg3ei16(vbool64_t op0, uint16_t * op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_i16m1))) +void vsuxseg4ei16(int16_t * op0, vuint16m1_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_i16m1_m))) +void vsuxseg4ei16(vbool16_t op0, int16_t * op1, vuint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_i16m2))) +void vsuxseg4ei16(int16_t * op0, vuint16m2_t op1, vint16m2_t op2, vint16m2_t op3, vint16m2_t op4, vint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_i16m2_m))) +void vsuxseg4ei16(vbool8_t op0, int16_t * op1, vuint16m2_t op2, vint16m2_t op3, vint16m2_t op4, vint16m2_t op5, vint16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_i16mf2))) +void vsuxseg4ei16(int16_t * op0, vuint16mf2_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_i16mf2_m))) +void vsuxseg4ei16(vbool32_t op0, int16_t * op1, vuint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_i16mf4))) +void vsuxseg4ei16(int16_t * op0, vuint16mf4_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_i16mf4_m))) +void vsuxseg4ei16(vbool64_t op0, int16_t * op1, vuint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_u16m1))) +void vsuxseg4ei16(uint16_t * op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_u16m1_m))) +void vsuxseg4ei16(vbool16_t op0, uint16_t * op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_u16m2))) +void vsuxseg4ei16(uint16_t * op0, vuint16m2_t op1, vuint16m2_t op2, vuint16m2_t op3, vuint16m2_t op4, vuint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_u16m2_m))) +void vsuxseg4ei16(vbool8_t op0, uint16_t * op1, vuint16m2_t op2, vuint16m2_t op3, vuint16m2_t op4, vuint16m2_t op5, vuint16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_u16mf2))) +void vsuxseg4ei16(uint16_t * op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_u16mf2_m))) +void vsuxseg4ei16(vbool32_t op0, uint16_t * op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_u16mf4))) +void vsuxseg4ei16(uint16_t * op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_u16mf4_m))) +void vsuxseg4ei16(vbool64_t op0, uint16_t * op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_i16m1))) +void vsuxseg5ei16(int16_t * op0, vuint16m1_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_i16m1_m))) +void vsuxseg5ei16(vbool16_t op0, int16_t * op1, vuint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_i16mf2))) +void vsuxseg5ei16(int16_t * op0, vuint16mf2_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_i16mf2_m))) +void vsuxseg5ei16(vbool32_t op0, int16_t * op1, vuint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_i16mf4))) +void vsuxseg5ei16(int16_t * op0, vuint16mf4_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_i16mf4_m))) +void vsuxseg5ei16(vbool64_t op0, int16_t * op1, vuint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_u16m1))) +void vsuxseg5ei16(uint16_t * op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_u16m1_m))) +void vsuxseg5ei16(vbool16_t op0, uint16_t * op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_u16mf2))) +void vsuxseg5ei16(uint16_t * op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_u16mf2_m))) +void vsuxseg5ei16(vbool32_t op0, uint16_t * op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_u16mf4))) +void vsuxseg5ei16(uint16_t * op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_u16mf4_m))) +void vsuxseg5ei16(vbool64_t op0, uint16_t * op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_i16m1))) +void vsuxseg6ei16(int16_t * op0, vuint16m1_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_i16m1_m))) +void vsuxseg6ei16(vbool16_t op0, int16_t * op1, vuint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_i16mf2))) +void vsuxseg6ei16(int16_t * op0, vuint16mf2_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_i16mf2_m))) +void vsuxseg6ei16(vbool32_t op0, int16_t * op1, vuint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_i16mf4))) +void vsuxseg6ei16(int16_t * op0, vuint16mf4_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_i16mf4_m))) +void vsuxseg6ei16(vbool64_t op0, int16_t * op1, vuint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_u16m1))) +void vsuxseg6ei16(uint16_t * op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_u16m1_m))) +void vsuxseg6ei16(vbool16_t op0, uint16_t * op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_u16mf2))) +void vsuxseg6ei16(uint16_t * op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_u16mf2_m))) +void vsuxseg6ei16(vbool32_t op0, uint16_t * op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_u16mf4))) +void vsuxseg6ei16(uint16_t * op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_u16mf4_m))) +void vsuxseg6ei16(vbool64_t op0, uint16_t * op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_i16m1))) +void vsuxseg7ei16(int16_t * op0, vuint16m1_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_i16m1_m))) +void vsuxseg7ei16(vbool16_t op0, int16_t * op1, vuint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_i16mf2))) +void vsuxseg7ei16(int16_t * op0, vuint16mf2_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_i16mf2_m))) +void vsuxseg7ei16(vbool32_t op0, int16_t * op1, vuint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_i16mf4))) +void vsuxseg7ei16(int16_t * op0, vuint16mf4_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_i16mf4_m))) +void vsuxseg7ei16(vbool64_t op0, int16_t * op1, vuint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_u16m1))) +void vsuxseg7ei16(uint16_t * op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_u16m1_m))) +void vsuxseg7ei16(vbool16_t op0, uint16_t * op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_u16mf2))) +void vsuxseg7ei16(uint16_t * op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_u16mf2_m))) +void vsuxseg7ei16(vbool32_t op0, uint16_t * op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_u16mf4))) +void vsuxseg7ei16(uint16_t * op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_u16mf4_m))) +void vsuxseg7ei16(vbool64_t op0, uint16_t * op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_i16m1))) +void vsuxseg8ei16(int16_t * op0, vuint16m1_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_i16m1_m))) +void vsuxseg8ei16(vbool16_t op0, int16_t * op1, vuint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, vint16m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_i16mf2))) +void vsuxseg8ei16(int16_t * op0, vuint16mf2_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_i16mf2_m))) +void vsuxseg8ei16(vbool32_t op0, int16_t * op1, vuint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, vint16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_i16mf4))) +void vsuxseg8ei16(int16_t * op0, vuint16mf4_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_i16mf4_m))) +void vsuxseg8ei16(vbool64_t op0, int16_t * op1, vuint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, vint16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_u16m1))) +void vsuxseg8ei16(uint16_t * op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_u16m1_m))) +void vsuxseg8ei16(vbool16_t op0, uint16_t * op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, vuint16m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_u16mf2))) +void vsuxseg8ei16(uint16_t * op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_u16mf2_m))) +void vsuxseg8ei16(vbool32_t op0, uint16_t * op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, vuint16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_u16mf4))) +void vsuxseg8ei16(uint16_t * op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_u16mf4_m))) +void vsuxseg8ei16(vbool64_t op0, uint16_t * op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, vuint16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_i16m1))) +void vsuxseg2ei32(int16_t * op0, vuint32m2_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_i16m1_m))) +void vsuxseg2ei32(vbool16_t op0, int16_t * op1, vuint32m2_t op2, vint16m1_t op3, vint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_i16m2))) +void vsuxseg2ei32(int16_t * op0, vuint32m4_t op1, vint16m2_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_i16m2_m))) +void vsuxseg2ei32(vbool8_t op0, int16_t * op1, vuint32m4_t op2, vint16m2_t op3, vint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_i16m4))) +void vsuxseg2ei32(int16_t * op0, vuint32m8_t op1, vint16m4_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_i16m4_m))) +void vsuxseg2ei32(vbool4_t op0, int16_t * op1, vuint32m8_t op2, vint16m4_t op3, vint16m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_i16mf2))) +void vsuxseg2ei32(int16_t * op0, vuint32m1_t op1, vint16mf2_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_i16mf2_m))) +void vsuxseg2ei32(vbool32_t op0, int16_t * op1, vuint32m1_t op2, vint16mf2_t op3, vint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_i16mf4))) +void vsuxseg2ei32(int16_t * op0, vuint32mf2_t op1, vint16mf4_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_i16mf4_m))) +void vsuxseg2ei32(vbool64_t op0, int16_t * op1, vuint32mf2_t op2, vint16mf4_t op3, vint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_u16m1))) +void vsuxseg2ei32(uint16_t * op0, vuint32m2_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_u16m1_m))) +void vsuxseg2ei32(vbool16_t op0, uint16_t * op1, vuint32m2_t op2, vuint16m1_t op3, vuint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_u16m2))) +void vsuxseg2ei32(uint16_t * op0, vuint32m4_t op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_u16m2_m))) +void vsuxseg2ei32(vbool8_t op0, uint16_t * op1, vuint32m4_t op2, vuint16m2_t op3, vuint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_u16m4))) +void vsuxseg2ei32(uint16_t * op0, vuint32m8_t op1, vuint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_u16m4_m))) +void vsuxseg2ei32(vbool4_t op0, uint16_t * op1, vuint32m8_t op2, vuint16m4_t op3, vuint16m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_u16mf2))) +void vsuxseg2ei32(uint16_t * op0, vuint32m1_t op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_u16mf2_m))) +void vsuxseg2ei32(vbool32_t op0, uint16_t * op1, vuint32m1_t op2, vuint16mf2_t op3, vuint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_u16mf4))) +void vsuxseg2ei32(uint16_t * op0, vuint32mf2_t op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_u16mf4_m))) +void vsuxseg2ei32(vbool64_t op0, uint16_t * op1, vuint32mf2_t op2, vuint16mf4_t op3, vuint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_i16m1))) +void vsuxseg3ei32(int16_t * op0, vuint32m2_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_i16m1_m))) +void vsuxseg3ei32(vbool16_t op0, int16_t * op1, vuint32m2_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_i16m2))) +void vsuxseg3ei32(int16_t * op0, vuint32m4_t op1, vint16m2_t op2, vint16m2_t op3, vint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_i16m2_m))) +void vsuxseg3ei32(vbool8_t op0, int16_t * op1, vuint32m4_t op2, vint16m2_t op3, vint16m2_t op4, vint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_i16mf2))) +void vsuxseg3ei32(int16_t * op0, vuint32m1_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_i16mf2_m))) +void vsuxseg3ei32(vbool32_t op0, int16_t * op1, vuint32m1_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_i16mf4))) +void vsuxseg3ei32(int16_t * op0, vuint32mf2_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_i16mf4_m))) +void vsuxseg3ei32(vbool64_t op0, int16_t * op1, vuint32mf2_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_u16m1))) +void vsuxseg3ei32(uint16_t * op0, vuint32m2_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_u16m1_m))) +void vsuxseg3ei32(vbool16_t op0, uint16_t * op1, vuint32m2_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_u16m2))) +void vsuxseg3ei32(uint16_t * op0, vuint32m4_t op1, vuint16m2_t op2, vuint16m2_t op3, vuint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_u16m2_m))) +void vsuxseg3ei32(vbool8_t op0, uint16_t * op1, vuint32m4_t op2, vuint16m2_t op3, vuint16m2_t op4, vuint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_u16mf2))) +void vsuxseg3ei32(uint16_t * op0, vuint32m1_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_u16mf2_m))) +void vsuxseg3ei32(vbool32_t op0, uint16_t * op1, vuint32m1_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_u16mf4))) +void vsuxseg3ei32(uint16_t * op0, vuint32mf2_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_u16mf4_m))) +void vsuxseg3ei32(vbool64_t op0, uint16_t * op1, vuint32mf2_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_i16m1))) +void vsuxseg4ei32(int16_t * op0, vuint32m2_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_i16m1_m))) +void vsuxseg4ei32(vbool16_t op0, int16_t * op1, vuint32m2_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_i16m2))) +void vsuxseg4ei32(int16_t * op0, vuint32m4_t op1, vint16m2_t op2, vint16m2_t op3, vint16m2_t op4, vint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_i16m2_m))) +void vsuxseg4ei32(vbool8_t op0, int16_t * op1, vuint32m4_t op2, vint16m2_t op3, vint16m2_t op4, vint16m2_t op5, vint16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_i16mf2))) +void vsuxseg4ei32(int16_t * op0, vuint32m1_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_i16mf2_m))) +void vsuxseg4ei32(vbool32_t op0, int16_t * op1, vuint32m1_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_i16mf4))) +void vsuxseg4ei32(int16_t * op0, vuint32mf2_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_i16mf4_m))) +void vsuxseg4ei32(vbool64_t op0, int16_t * op1, vuint32mf2_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_u16m1))) +void vsuxseg4ei32(uint16_t * op0, vuint32m2_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_u16m1_m))) +void vsuxseg4ei32(vbool16_t op0, uint16_t * op1, vuint32m2_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_u16m2))) +void vsuxseg4ei32(uint16_t * op0, vuint32m4_t op1, vuint16m2_t op2, vuint16m2_t op3, vuint16m2_t op4, vuint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_u16m2_m))) +void vsuxseg4ei32(vbool8_t op0, uint16_t * op1, vuint32m4_t op2, vuint16m2_t op3, vuint16m2_t op4, vuint16m2_t op5, vuint16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_u16mf2))) +void vsuxseg4ei32(uint16_t * op0, vuint32m1_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_u16mf2_m))) +void vsuxseg4ei32(vbool32_t op0, uint16_t * op1, vuint32m1_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_u16mf4))) +void vsuxseg4ei32(uint16_t * op0, vuint32mf2_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_u16mf4_m))) +void vsuxseg4ei32(vbool64_t op0, uint16_t * op1, vuint32mf2_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_i16m1))) +void vsuxseg5ei32(int16_t * op0, vuint32m2_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_i16m1_m))) +void vsuxseg5ei32(vbool16_t op0, int16_t * op1, vuint32m2_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_i16mf2))) +void vsuxseg5ei32(int16_t * op0, vuint32m1_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_i16mf2_m))) +void vsuxseg5ei32(vbool32_t op0, int16_t * op1, vuint32m1_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_i16mf4))) +void vsuxseg5ei32(int16_t * op0, vuint32mf2_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_i16mf4_m))) +void vsuxseg5ei32(vbool64_t op0, int16_t * op1, vuint32mf2_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_u16m1))) +void vsuxseg5ei32(uint16_t * op0, vuint32m2_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_u16m1_m))) +void vsuxseg5ei32(vbool16_t op0, uint16_t * op1, vuint32m2_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_u16mf2))) +void vsuxseg5ei32(uint16_t * op0, vuint32m1_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_u16mf2_m))) +void vsuxseg5ei32(vbool32_t op0, uint16_t * op1, vuint32m1_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_u16mf4))) +void vsuxseg5ei32(uint16_t * op0, vuint32mf2_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_u16mf4_m))) +void vsuxseg5ei32(vbool64_t op0, uint16_t * op1, vuint32mf2_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_i16m1))) +void vsuxseg6ei32(int16_t * op0, vuint32m2_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_i16m1_m))) +void vsuxseg6ei32(vbool16_t op0, int16_t * op1, vuint32m2_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_i16mf2))) +void vsuxseg6ei32(int16_t * op0, vuint32m1_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_i16mf2_m))) +void vsuxseg6ei32(vbool32_t op0, int16_t * op1, vuint32m1_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_i16mf4))) +void vsuxseg6ei32(int16_t * op0, vuint32mf2_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_i16mf4_m))) +void vsuxseg6ei32(vbool64_t op0, int16_t * op1, vuint32mf2_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_u16m1))) +void vsuxseg6ei32(uint16_t * op0, vuint32m2_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_u16m1_m))) +void vsuxseg6ei32(vbool16_t op0, uint16_t * op1, vuint32m2_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_u16mf2))) +void vsuxseg6ei32(uint16_t * op0, vuint32m1_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_u16mf2_m))) +void vsuxseg6ei32(vbool32_t op0, uint16_t * op1, vuint32m1_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_u16mf4))) +void vsuxseg6ei32(uint16_t * op0, vuint32mf2_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_u16mf4_m))) +void vsuxseg6ei32(vbool64_t op0, uint16_t * op1, vuint32mf2_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_i16m1))) +void vsuxseg7ei32(int16_t * op0, vuint32m2_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_i16m1_m))) +void vsuxseg7ei32(vbool16_t op0, int16_t * op1, vuint32m2_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_i16mf2))) +void vsuxseg7ei32(int16_t * op0, vuint32m1_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_i16mf2_m))) +void vsuxseg7ei32(vbool32_t op0, int16_t * op1, vuint32m1_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_i16mf4))) +void vsuxseg7ei32(int16_t * op0, vuint32mf2_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_i16mf4_m))) +void vsuxseg7ei32(vbool64_t op0, int16_t * op1, vuint32mf2_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_u16m1))) +void vsuxseg7ei32(uint16_t * op0, vuint32m2_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_u16m1_m))) +void vsuxseg7ei32(vbool16_t op0, uint16_t * op1, vuint32m2_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_u16mf2))) +void vsuxseg7ei32(uint16_t * op0, vuint32m1_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_u16mf2_m))) +void vsuxseg7ei32(vbool32_t op0, uint16_t * op1, vuint32m1_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_u16mf4))) +void vsuxseg7ei32(uint16_t * op0, vuint32mf2_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_u16mf4_m))) +void vsuxseg7ei32(vbool64_t op0, uint16_t * op1, vuint32mf2_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_i16m1))) +void vsuxseg8ei32(int16_t * op0, vuint32m2_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_i16m1_m))) +void vsuxseg8ei32(vbool16_t op0, int16_t * op1, vuint32m2_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, vint16m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_i16mf2))) +void vsuxseg8ei32(int16_t * op0, vuint32m1_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_i16mf2_m))) +void vsuxseg8ei32(vbool32_t op0, int16_t * op1, vuint32m1_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, vint16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_i16mf4))) +void vsuxseg8ei32(int16_t * op0, vuint32mf2_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_i16mf4_m))) +void vsuxseg8ei32(vbool64_t op0, int16_t * op1, vuint32mf2_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, vint16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_u16m1))) +void vsuxseg8ei32(uint16_t * op0, vuint32m2_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_u16m1_m))) +void vsuxseg8ei32(vbool16_t op0, uint16_t * op1, vuint32m2_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, vuint16m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_u16mf2))) +void vsuxseg8ei32(uint16_t * op0, vuint32m1_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_u16mf2_m))) +void vsuxseg8ei32(vbool32_t op0, uint16_t * op1, vuint32m1_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, vuint16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_u16mf4))) +void vsuxseg8ei32(uint16_t * op0, vuint32mf2_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_u16mf4_m))) +void vsuxseg8ei32(vbool64_t op0, uint16_t * op1, vuint32mf2_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, vuint16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_i16m1))) +void vsuxseg2ei64(int16_t * op0, vuint64m4_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_i16m1_m))) +void vsuxseg2ei64(vbool16_t op0, int16_t * op1, vuint64m4_t op2, vint16m1_t op3, vint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_i16m2))) +void vsuxseg2ei64(int16_t * op0, vuint64m8_t op1, vint16m2_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_i16m2_m))) +void vsuxseg2ei64(vbool8_t op0, int16_t * op1, vuint64m8_t op2, vint16m2_t op3, vint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_i16mf2))) +void vsuxseg2ei64(int16_t * op0, vuint64m2_t op1, vint16mf2_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_i16mf2_m))) +void vsuxseg2ei64(vbool32_t op0, int16_t * op1, vuint64m2_t op2, vint16mf2_t op3, vint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_i16mf4))) +void vsuxseg2ei64(int16_t * op0, vuint64m1_t op1, vint16mf4_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_i16mf4_m))) +void vsuxseg2ei64(vbool64_t op0, int16_t * op1, vuint64m1_t op2, vint16mf4_t op3, vint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_u16m1))) +void vsuxseg2ei64(uint16_t * op0, vuint64m4_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_u16m1_m))) +void vsuxseg2ei64(vbool16_t op0, uint16_t * op1, vuint64m4_t op2, vuint16m1_t op3, vuint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_u16m2))) +void vsuxseg2ei64(uint16_t * op0, vuint64m8_t op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_u16m2_m))) +void vsuxseg2ei64(vbool8_t op0, uint16_t * op1, vuint64m8_t op2, vuint16m2_t op3, vuint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_u16mf2))) +void vsuxseg2ei64(uint16_t * op0, vuint64m2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_u16mf2_m))) +void vsuxseg2ei64(vbool32_t op0, uint16_t * op1, vuint64m2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_u16mf4))) +void vsuxseg2ei64(uint16_t * op0, vuint64m1_t op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_u16mf4_m))) +void vsuxseg2ei64(vbool64_t op0, uint16_t * op1, vuint64m1_t op2, vuint16mf4_t op3, vuint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_i16m1))) +void vsuxseg3ei64(int16_t * op0, vuint64m4_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_i16m1_m))) +void vsuxseg3ei64(vbool16_t op0, int16_t * op1, vuint64m4_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_i16m2))) +void vsuxseg3ei64(int16_t * op0, vuint64m8_t op1, vint16m2_t op2, vint16m2_t op3, vint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_i16m2_m))) +void vsuxseg3ei64(vbool8_t op0, int16_t * op1, vuint64m8_t op2, vint16m2_t op3, vint16m2_t op4, vint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_i16mf2))) +void vsuxseg3ei64(int16_t * op0, vuint64m2_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_i16mf2_m))) +void vsuxseg3ei64(vbool32_t op0, int16_t * op1, vuint64m2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_i16mf4))) +void vsuxseg3ei64(int16_t * op0, vuint64m1_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_i16mf4_m))) +void vsuxseg3ei64(vbool64_t op0, int16_t * op1, vuint64m1_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_u16m1))) +void vsuxseg3ei64(uint16_t * op0, vuint64m4_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_u16m1_m))) +void vsuxseg3ei64(vbool16_t op0, uint16_t * op1, vuint64m4_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_u16m2))) +void vsuxseg3ei64(uint16_t * op0, vuint64m8_t op1, vuint16m2_t op2, vuint16m2_t op3, vuint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_u16m2_m))) +void vsuxseg3ei64(vbool8_t op0, uint16_t * op1, vuint64m8_t op2, vuint16m2_t op3, vuint16m2_t op4, vuint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_u16mf2))) +void vsuxseg3ei64(uint16_t * op0, vuint64m2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_u16mf2_m))) +void vsuxseg3ei64(vbool32_t op0, uint16_t * op1, vuint64m2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_u16mf4))) +void vsuxseg3ei64(uint16_t * op0, vuint64m1_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_u16mf4_m))) +void vsuxseg3ei64(vbool64_t op0, uint16_t * op1, vuint64m1_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_i16m1))) +void vsuxseg4ei64(int16_t * op0, vuint64m4_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_i16m1_m))) +void vsuxseg4ei64(vbool16_t op0, int16_t * op1, vuint64m4_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_i16m2))) +void vsuxseg4ei64(int16_t * op0, vuint64m8_t op1, vint16m2_t op2, vint16m2_t op3, vint16m2_t op4, vint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_i16m2_m))) +void vsuxseg4ei64(vbool8_t op0, int16_t * op1, vuint64m8_t op2, vint16m2_t op3, vint16m2_t op4, vint16m2_t op5, vint16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_i16mf2))) +void vsuxseg4ei64(int16_t * op0, vuint64m2_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_i16mf2_m))) +void vsuxseg4ei64(vbool32_t op0, int16_t * op1, vuint64m2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_i16mf4))) +void vsuxseg4ei64(int16_t * op0, vuint64m1_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_i16mf4_m))) +void vsuxseg4ei64(vbool64_t op0, int16_t * op1, vuint64m1_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_u16m1))) +void vsuxseg4ei64(uint16_t * op0, vuint64m4_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_u16m1_m))) +void vsuxseg4ei64(vbool16_t op0, uint16_t * op1, vuint64m4_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_u16m2))) +void vsuxseg4ei64(uint16_t * op0, vuint64m8_t op1, vuint16m2_t op2, vuint16m2_t op3, vuint16m2_t op4, vuint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_u16m2_m))) +void vsuxseg4ei64(vbool8_t op0, uint16_t * op1, vuint64m8_t op2, vuint16m2_t op3, vuint16m2_t op4, vuint16m2_t op5, vuint16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_u16mf2))) +void vsuxseg4ei64(uint16_t * op0, vuint64m2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_u16mf2_m))) +void vsuxseg4ei64(vbool32_t op0, uint16_t * op1, vuint64m2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_u16mf4))) +void vsuxseg4ei64(uint16_t * op0, vuint64m1_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_u16mf4_m))) +void vsuxseg4ei64(vbool64_t op0, uint16_t * op1, vuint64m1_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_i16m1))) +void vsuxseg5ei64(int16_t * op0, vuint64m4_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_i16m1_m))) +void vsuxseg5ei64(vbool16_t op0, int16_t * op1, vuint64m4_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_i16mf2))) +void vsuxseg5ei64(int16_t * op0, vuint64m2_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_i16mf2_m))) +void vsuxseg5ei64(vbool32_t op0, int16_t * op1, vuint64m2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_i16mf4))) +void vsuxseg5ei64(int16_t * op0, vuint64m1_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_i16mf4_m))) +void vsuxseg5ei64(vbool64_t op0, int16_t * op1, vuint64m1_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_u16m1))) +void vsuxseg5ei64(uint16_t * op0, vuint64m4_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_u16m1_m))) +void vsuxseg5ei64(vbool16_t op0, uint16_t * op1, vuint64m4_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_u16mf2))) +void vsuxseg5ei64(uint16_t * op0, vuint64m2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_u16mf2_m))) +void vsuxseg5ei64(vbool32_t op0, uint16_t * op1, vuint64m2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_u16mf4))) +void vsuxseg5ei64(uint16_t * op0, vuint64m1_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_u16mf4_m))) +void vsuxseg5ei64(vbool64_t op0, uint16_t * op1, vuint64m1_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_i16m1))) +void vsuxseg6ei64(int16_t * op0, vuint64m4_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_i16m1_m))) +void vsuxseg6ei64(vbool16_t op0, int16_t * op1, vuint64m4_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_i16mf2))) +void vsuxseg6ei64(int16_t * op0, vuint64m2_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_i16mf2_m))) +void vsuxseg6ei64(vbool32_t op0, int16_t * op1, vuint64m2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_i16mf4))) +void vsuxseg6ei64(int16_t * op0, vuint64m1_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_i16mf4_m))) +void vsuxseg6ei64(vbool64_t op0, int16_t * op1, vuint64m1_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_u16m1))) +void vsuxseg6ei64(uint16_t * op0, vuint64m4_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_u16m1_m))) +void vsuxseg6ei64(vbool16_t op0, uint16_t * op1, vuint64m4_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_u16mf2))) +void vsuxseg6ei64(uint16_t * op0, vuint64m2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_u16mf2_m))) +void vsuxseg6ei64(vbool32_t op0, uint16_t * op1, vuint64m2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_u16mf4))) +void vsuxseg6ei64(uint16_t * op0, vuint64m1_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_u16mf4_m))) +void vsuxseg6ei64(vbool64_t op0, uint16_t * op1, vuint64m1_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_i16m1))) +void vsuxseg7ei64(int16_t * op0, vuint64m4_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_i16m1_m))) +void vsuxseg7ei64(vbool16_t op0, int16_t * op1, vuint64m4_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_i16mf2))) +void vsuxseg7ei64(int16_t * op0, vuint64m2_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_i16mf2_m))) +void vsuxseg7ei64(vbool32_t op0, int16_t * op1, vuint64m2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_i16mf4))) +void vsuxseg7ei64(int16_t * op0, vuint64m1_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_i16mf4_m))) +void vsuxseg7ei64(vbool64_t op0, int16_t * op1, vuint64m1_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_u16m1))) +void vsuxseg7ei64(uint16_t * op0, vuint64m4_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_u16m1_m))) +void vsuxseg7ei64(vbool16_t op0, uint16_t * op1, vuint64m4_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_u16mf2))) +void vsuxseg7ei64(uint16_t * op0, vuint64m2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_u16mf2_m))) +void vsuxseg7ei64(vbool32_t op0, uint16_t * op1, vuint64m2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_u16mf4))) +void vsuxseg7ei64(uint16_t * op0, vuint64m1_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_u16mf4_m))) +void vsuxseg7ei64(vbool64_t op0, uint16_t * op1, vuint64m1_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_i16m1))) +void vsuxseg8ei64(int16_t * op0, vuint64m4_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_i16m1_m))) +void vsuxseg8ei64(vbool16_t op0, int16_t * op1, vuint64m4_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, vint16m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_i16mf2))) +void vsuxseg8ei64(int16_t * op0, vuint64m2_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_i16mf2_m))) +void vsuxseg8ei64(vbool32_t op0, int16_t * op1, vuint64m2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, vint16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_i16mf4))) +void vsuxseg8ei64(int16_t * op0, vuint64m1_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_i16mf4_m))) +void vsuxseg8ei64(vbool64_t op0, int16_t * op1, vuint64m1_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, vint16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_u16m1))) +void vsuxseg8ei64(uint16_t * op0, vuint64m4_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_u16m1_m))) +void vsuxseg8ei64(vbool16_t op0, uint16_t * op1, vuint64m4_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, vuint16m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_u16mf2))) +void vsuxseg8ei64(uint16_t * op0, vuint64m2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_u16mf2_m))) +void vsuxseg8ei64(vbool32_t op0, uint16_t * op1, vuint64m2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, vuint16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_u16mf4))) +void vsuxseg8ei64(uint16_t * op0, vuint64m1_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_u16mf4_m))) +void vsuxseg8ei64(vbool64_t op0, uint16_t * op1, vuint64m1_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, vuint16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_i32m1))) +void vsuxseg2ei8(int32_t * op0, vuint8mf4_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_i32m1_m))) +void vsuxseg2ei8(vbool32_t op0, int32_t * op1, vuint8mf4_t op2, vint32m1_t op3, vint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_i32m2))) +void vsuxseg2ei8(int32_t * op0, vuint8mf2_t op1, vint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_i32m2_m))) +void vsuxseg2ei8(vbool16_t op0, int32_t * op1, vuint8mf2_t op2, vint32m2_t op3, vint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_i32m4))) +void vsuxseg2ei8(int32_t * op0, vuint8m1_t op1, vint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_i32m4_m))) +void vsuxseg2ei8(vbool8_t op0, int32_t * op1, vuint8m1_t op2, vint32m4_t op3, vint32m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_i32mf2))) +void vsuxseg2ei8(int32_t * op0, vuint8mf8_t op1, vint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_i32mf2_m))) +void vsuxseg2ei8(vbool64_t op0, int32_t * op1, vuint8mf8_t op2, vint32mf2_t op3, vint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_u32m1))) +void vsuxseg2ei8(uint32_t * op0, vuint8mf4_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_u32m1_m))) +void vsuxseg2ei8(vbool32_t op0, uint32_t * op1, vuint8mf4_t op2, vuint32m1_t op3, vuint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_u32m2))) +void vsuxseg2ei8(uint32_t * op0, vuint8mf2_t op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_u32m2_m))) +void vsuxseg2ei8(vbool16_t op0, uint32_t * op1, vuint8mf2_t op2, vuint32m2_t op3, vuint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_u32m4))) +void vsuxseg2ei8(uint32_t * op0, vuint8m1_t op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_u32m4_m))) +void vsuxseg2ei8(vbool8_t op0, uint32_t * op1, vuint8m1_t op2, vuint32m4_t op3, vuint32m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_u32mf2))) +void vsuxseg2ei8(uint32_t * op0, vuint8mf8_t op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_u32mf2_m))) +void vsuxseg2ei8(vbool64_t op0, uint32_t * op1, vuint8mf8_t op2, vuint32mf2_t op3, vuint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_i32m1))) +void vsuxseg3ei8(int32_t * op0, vuint8mf4_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_i32m1_m))) +void vsuxseg3ei8(vbool32_t op0, int32_t * op1, vuint8mf4_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_i32m2))) +void vsuxseg3ei8(int32_t * op0, vuint8mf2_t op1, vint32m2_t op2, vint32m2_t op3, vint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_i32m2_m))) +void vsuxseg3ei8(vbool16_t op0, int32_t * op1, vuint8mf2_t op2, vint32m2_t op3, vint32m2_t op4, vint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_i32mf2))) +void vsuxseg3ei8(int32_t * op0, vuint8mf8_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_i32mf2_m))) +void vsuxseg3ei8(vbool64_t op0, int32_t * op1, vuint8mf8_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_u32m1))) +void vsuxseg3ei8(uint32_t * op0, vuint8mf4_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_u32m1_m))) +void vsuxseg3ei8(vbool32_t op0, uint32_t * op1, vuint8mf4_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_u32m2))) +void vsuxseg3ei8(uint32_t * op0, vuint8mf2_t op1, vuint32m2_t op2, vuint32m2_t op3, vuint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_u32m2_m))) +void vsuxseg3ei8(vbool16_t op0, uint32_t * op1, vuint8mf2_t op2, vuint32m2_t op3, vuint32m2_t op4, vuint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_u32mf2))) +void vsuxseg3ei8(uint32_t * op0, vuint8mf8_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_u32mf2_m))) +void vsuxseg3ei8(vbool64_t op0, uint32_t * op1, vuint8mf8_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_i32m1))) +void vsuxseg4ei8(int32_t * op0, vuint8mf4_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_i32m1_m))) +void vsuxseg4ei8(vbool32_t op0, int32_t * op1, vuint8mf4_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_i32m2))) +void vsuxseg4ei8(int32_t * op0, vuint8mf2_t op1, vint32m2_t op2, vint32m2_t op3, vint32m2_t op4, vint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_i32m2_m))) +void vsuxseg4ei8(vbool16_t op0, int32_t * op1, vuint8mf2_t op2, vint32m2_t op3, vint32m2_t op4, vint32m2_t op5, vint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_i32mf2))) +void vsuxseg4ei8(int32_t * op0, vuint8mf8_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_i32mf2_m))) +void vsuxseg4ei8(vbool64_t op0, int32_t * op1, vuint8mf8_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_u32m1))) +void vsuxseg4ei8(uint32_t * op0, vuint8mf4_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_u32m1_m))) +void vsuxseg4ei8(vbool32_t op0, uint32_t * op1, vuint8mf4_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_u32m2))) +void vsuxseg4ei8(uint32_t * op0, vuint8mf2_t op1, vuint32m2_t op2, vuint32m2_t op3, vuint32m2_t op4, vuint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_u32m2_m))) +void vsuxseg4ei8(vbool16_t op0, uint32_t * op1, vuint8mf2_t op2, vuint32m2_t op3, vuint32m2_t op4, vuint32m2_t op5, vuint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_u32mf2))) +void vsuxseg4ei8(uint32_t * op0, vuint8mf8_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_u32mf2_m))) +void vsuxseg4ei8(vbool64_t op0, uint32_t * op1, vuint8mf8_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_i32m1))) +void vsuxseg5ei8(int32_t * op0, vuint8mf4_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_i32m1_m))) +void vsuxseg5ei8(vbool32_t op0, int32_t * op1, vuint8mf4_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_i32mf2))) +void vsuxseg5ei8(int32_t * op0, vuint8mf8_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_i32mf2_m))) +void vsuxseg5ei8(vbool64_t op0, int32_t * op1, vuint8mf8_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_u32m1))) +void vsuxseg5ei8(uint32_t * op0, vuint8mf4_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_u32m1_m))) +void vsuxseg5ei8(vbool32_t op0, uint32_t * op1, vuint8mf4_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_u32mf2))) +void vsuxseg5ei8(uint32_t * op0, vuint8mf8_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_u32mf2_m))) +void vsuxseg5ei8(vbool64_t op0, uint32_t * op1, vuint8mf8_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_i32m1))) +void vsuxseg6ei8(int32_t * op0, vuint8mf4_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_i32m1_m))) +void vsuxseg6ei8(vbool32_t op0, int32_t * op1, vuint8mf4_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_i32mf2))) +void vsuxseg6ei8(int32_t * op0, vuint8mf8_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_i32mf2_m))) +void vsuxseg6ei8(vbool64_t op0, int32_t * op1, vuint8mf8_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_u32m1))) +void vsuxseg6ei8(uint32_t * op0, vuint8mf4_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_u32m1_m))) +void vsuxseg6ei8(vbool32_t op0, uint32_t * op1, vuint8mf4_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_u32mf2))) +void vsuxseg6ei8(uint32_t * op0, vuint8mf8_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_u32mf2_m))) +void vsuxseg6ei8(vbool64_t op0, uint32_t * op1, vuint8mf8_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_i32m1))) +void vsuxseg7ei8(int32_t * op0, vuint8mf4_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_i32m1_m))) +void vsuxseg7ei8(vbool32_t op0, int32_t * op1, vuint8mf4_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_i32mf2))) +void vsuxseg7ei8(int32_t * op0, vuint8mf8_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_i32mf2_m))) +void vsuxseg7ei8(vbool64_t op0, int32_t * op1, vuint8mf8_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_u32m1))) +void vsuxseg7ei8(uint32_t * op0, vuint8mf4_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_u32m1_m))) +void vsuxseg7ei8(vbool32_t op0, uint32_t * op1, vuint8mf4_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_u32mf2))) +void vsuxseg7ei8(uint32_t * op0, vuint8mf8_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_u32mf2_m))) +void vsuxseg7ei8(vbool64_t op0, uint32_t * op1, vuint8mf8_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_i32m1))) +void vsuxseg8ei8(int32_t * op0, vuint8mf4_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_i32m1_m))) +void vsuxseg8ei8(vbool32_t op0, int32_t * op1, vuint8mf4_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, vint32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_i32mf2))) +void vsuxseg8ei8(int32_t * op0, vuint8mf8_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_i32mf2_m))) +void vsuxseg8ei8(vbool64_t op0, int32_t * op1, vuint8mf8_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, vint32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_u32m1))) +void vsuxseg8ei8(uint32_t * op0, vuint8mf4_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_u32m1_m))) +void vsuxseg8ei8(vbool32_t op0, uint32_t * op1, vuint8mf4_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, vuint32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_u32mf2))) +void vsuxseg8ei8(uint32_t * op0, vuint8mf8_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_u32mf2_m))) +void vsuxseg8ei8(vbool64_t op0, uint32_t * op1, vuint8mf8_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, vuint32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_i32m1))) +void vsuxseg2ei16(int32_t * op0, vuint16mf2_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_i32m1_m))) +void vsuxseg2ei16(vbool32_t op0, int32_t * op1, vuint16mf2_t op2, vint32m1_t op3, vint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_i32m2))) +void vsuxseg2ei16(int32_t * op0, vuint16m1_t op1, vint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_i32m2_m))) +void vsuxseg2ei16(vbool16_t op0, int32_t * op1, vuint16m1_t op2, vint32m2_t op3, vint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_i32m4))) +void vsuxseg2ei16(int32_t * op0, vuint16m2_t op1, vint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_i32m4_m))) +void vsuxseg2ei16(vbool8_t op0, int32_t * op1, vuint16m2_t op2, vint32m4_t op3, vint32m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_i32mf2))) +void vsuxseg2ei16(int32_t * op0, vuint16mf4_t op1, vint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_i32mf2_m))) +void vsuxseg2ei16(vbool64_t op0, int32_t * op1, vuint16mf4_t op2, vint32mf2_t op3, vint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_u32m1))) +void vsuxseg2ei16(uint32_t * op0, vuint16mf2_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_u32m1_m))) +void vsuxseg2ei16(vbool32_t op0, uint32_t * op1, vuint16mf2_t op2, vuint32m1_t op3, vuint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_u32m2))) +void vsuxseg2ei16(uint32_t * op0, vuint16m1_t op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_u32m2_m))) +void vsuxseg2ei16(vbool16_t op0, uint32_t * op1, vuint16m1_t op2, vuint32m2_t op3, vuint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_u32m4))) +void vsuxseg2ei16(uint32_t * op0, vuint16m2_t op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_u32m4_m))) +void vsuxseg2ei16(vbool8_t op0, uint32_t * op1, vuint16m2_t op2, vuint32m4_t op3, vuint32m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_u32mf2))) +void vsuxseg2ei16(uint32_t * op0, vuint16mf4_t op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_u32mf2_m))) +void vsuxseg2ei16(vbool64_t op0, uint32_t * op1, vuint16mf4_t op2, vuint32mf2_t op3, vuint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_i32m1))) +void vsuxseg3ei16(int32_t * op0, vuint16mf2_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_i32m1_m))) +void vsuxseg3ei16(vbool32_t op0, int32_t * op1, vuint16mf2_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_i32m2))) +void vsuxseg3ei16(int32_t * op0, vuint16m1_t op1, vint32m2_t op2, vint32m2_t op3, vint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_i32m2_m))) +void vsuxseg3ei16(vbool16_t op0, int32_t * op1, vuint16m1_t op2, vint32m2_t op3, vint32m2_t op4, vint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_i32mf2))) +void vsuxseg3ei16(int32_t * op0, vuint16mf4_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_i32mf2_m))) +void vsuxseg3ei16(vbool64_t op0, int32_t * op1, vuint16mf4_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_u32m1))) +void vsuxseg3ei16(uint32_t * op0, vuint16mf2_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_u32m1_m))) +void vsuxseg3ei16(vbool32_t op0, uint32_t * op1, vuint16mf2_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_u32m2))) +void vsuxseg3ei16(uint32_t * op0, vuint16m1_t op1, vuint32m2_t op2, vuint32m2_t op3, vuint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_u32m2_m))) +void vsuxseg3ei16(vbool16_t op0, uint32_t * op1, vuint16m1_t op2, vuint32m2_t op3, vuint32m2_t op4, vuint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_u32mf2))) +void vsuxseg3ei16(uint32_t * op0, vuint16mf4_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_u32mf2_m))) +void vsuxseg3ei16(vbool64_t op0, uint32_t * op1, vuint16mf4_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_i32m1))) +void vsuxseg4ei16(int32_t * op0, vuint16mf2_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_i32m1_m))) +void vsuxseg4ei16(vbool32_t op0, int32_t * op1, vuint16mf2_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_i32m2))) +void vsuxseg4ei16(int32_t * op0, vuint16m1_t op1, vint32m2_t op2, vint32m2_t op3, vint32m2_t op4, vint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_i32m2_m))) +void vsuxseg4ei16(vbool16_t op0, int32_t * op1, vuint16m1_t op2, vint32m2_t op3, vint32m2_t op4, vint32m2_t op5, vint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_i32mf2))) +void vsuxseg4ei16(int32_t * op0, vuint16mf4_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_i32mf2_m))) +void vsuxseg4ei16(vbool64_t op0, int32_t * op1, vuint16mf4_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_u32m1))) +void vsuxseg4ei16(uint32_t * op0, vuint16mf2_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_u32m1_m))) +void vsuxseg4ei16(vbool32_t op0, uint32_t * op1, vuint16mf2_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_u32m2))) +void vsuxseg4ei16(uint32_t * op0, vuint16m1_t op1, vuint32m2_t op2, vuint32m2_t op3, vuint32m2_t op4, vuint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_u32m2_m))) +void vsuxseg4ei16(vbool16_t op0, uint32_t * op1, vuint16m1_t op2, vuint32m2_t op3, vuint32m2_t op4, vuint32m2_t op5, vuint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_u32mf2))) +void vsuxseg4ei16(uint32_t * op0, vuint16mf4_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_u32mf2_m))) +void vsuxseg4ei16(vbool64_t op0, uint32_t * op1, vuint16mf4_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_i32m1))) +void vsuxseg5ei16(int32_t * op0, vuint16mf2_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_i32m1_m))) +void vsuxseg5ei16(vbool32_t op0, int32_t * op1, vuint16mf2_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_i32mf2))) +void vsuxseg5ei16(int32_t * op0, vuint16mf4_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_i32mf2_m))) +void vsuxseg5ei16(vbool64_t op0, int32_t * op1, vuint16mf4_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_u32m1))) +void vsuxseg5ei16(uint32_t * op0, vuint16mf2_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_u32m1_m))) +void vsuxseg5ei16(vbool32_t op0, uint32_t * op1, vuint16mf2_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_u32mf2))) +void vsuxseg5ei16(uint32_t * op0, vuint16mf4_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_u32mf2_m))) +void vsuxseg5ei16(vbool64_t op0, uint32_t * op1, vuint16mf4_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_i32m1))) +void vsuxseg6ei16(int32_t * op0, vuint16mf2_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_i32m1_m))) +void vsuxseg6ei16(vbool32_t op0, int32_t * op1, vuint16mf2_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_i32mf2))) +void vsuxseg6ei16(int32_t * op0, vuint16mf4_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_i32mf2_m))) +void vsuxseg6ei16(vbool64_t op0, int32_t * op1, vuint16mf4_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_u32m1))) +void vsuxseg6ei16(uint32_t * op0, vuint16mf2_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_u32m1_m))) +void vsuxseg6ei16(vbool32_t op0, uint32_t * op1, vuint16mf2_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_u32mf2))) +void vsuxseg6ei16(uint32_t * op0, vuint16mf4_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_u32mf2_m))) +void vsuxseg6ei16(vbool64_t op0, uint32_t * op1, vuint16mf4_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_i32m1))) +void vsuxseg7ei16(int32_t * op0, vuint16mf2_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_i32m1_m))) +void vsuxseg7ei16(vbool32_t op0, int32_t * op1, vuint16mf2_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_i32mf2))) +void vsuxseg7ei16(int32_t * op0, vuint16mf4_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_i32mf2_m))) +void vsuxseg7ei16(vbool64_t op0, int32_t * op1, vuint16mf4_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_u32m1))) +void vsuxseg7ei16(uint32_t * op0, vuint16mf2_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_u32m1_m))) +void vsuxseg7ei16(vbool32_t op0, uint32_t * op1, vuint16mf2_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_u32mf2))) +void vsuxseg7ei16(uint32_t * op0, vuint16mf4_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_u32mf2_m))) +void vsuxseg7ei16(vbool64_t op0, uint32_t * op1, vuint16mf4_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_i32m1))) +void vsuxseg8ei16(int32_t * op0, vuint16mf2_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_i32m1_m))) +void vsuxseg8ei16(vbool32_t op0, int32_t * op1, vuint16mf2_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, vint32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_i32mf2))) +void vsuxseg8ei16(int32_t * op0, vuint16mf4_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_i32mf2_m))) +void vsuxseg8ei16(vbool64_t op0, int32_t * op1, vuint16mf4_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, vint32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_u32m1))) +void vsuxseg8ei16(uint32_t * op0, vuint16mf2_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_u32m1_m))) +void vsuxseg8ei16(vbool32_t op0, uint32_t * op1, vuint16mf2_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, vuint32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_u32mf2))) +void vsuxseg8ei16(uint32_t * op0, vuint16mf4_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_u32mf2_m))) +void vsuxseg8ei16(vbool64_t op0, uint32_t * op1, vuint16mf4_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, vuint32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_i32m1))) +void vsuxseg2ei32(int32_t * op0, vuint32m1_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_i32m1_m))) +void vsuxseg2ei32(vbool32_t op0, int32_t * op1, vuint32m1_t op2, vint32m1_t op3, vint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_i32m2))) +void vsuxseg2ei32(int32_t * op0, vuint32m2_t op1, vint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_i32m2_m))) +void vsuxseg2ei32(vbool16_t op0, int32_t * op1, vuint32m2_t op2, vint32m2_t op3, vint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_i32m4))) +void vsuxseg2ei32(int32_t * op0, vuint32m4_t op1, vint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_i32m4_m))) +void vsuxseg2ei32(vbool8_t op0, int32_t * op1, vuint32m4_t op2, vint32m4_t op3, vint32m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_i32mf2))) +void vsuxseg2ei32(int32_t * op0, vuint32mf2_t op1, vint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_i32mf2_m))) +void vsuxseg2ei32(vbool64_t op0, int32_t * op1, vuint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_u32m1))) +void vsuxseg2ei32(uint32_t * op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_u32m1_m))) +void vsuxseg2ei32(vbool32_t op0, uint32_t * op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_u32m2))) +void vsuxseg2ei32(uint32_t * op0, vuint32m2_t op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_u32m2_m))) +void vsuxseg2ei32(vbool16_t op0, uint32_t * op1, vuint32m2_t op2, vuint32m2_t op3, vuint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_u32m4))) +void vsuxseg2ei32(uint32_t * op0, vuint32m4_t op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_u32m4_m))) +void vsuxseg2ei32(vbool8_t op0, uint32_t * op1, vuint32m4_t op2, vuint32m4_t op3, vuint32m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_u32mf2))) +void vsuxseg2ei32(uint32_t * op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_u32mf2_m))) +void vsuxseg2ei32(vbool64_t op0, uint32_t * op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_i32m1))) +void vsuxseg3ei32(int32_t * op0, vuint32m1_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_i32m1_m))) +void vsuxseg3ei32(vbool32_t op0, int32_t * op1, vuint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_i32m2))) +void vsuxseg3ei32(int32_t * op0, vuint32m2_t op1, vint32m2_t op2, vint32m2_t op3, vint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_i32m2_m))) +void vsuxseg3ei32(vbool16_t op0, int32_t * op1, vuint32m2_t op2, vint32m2_t op3, vint32m2_t op4, vint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_i32mf2))) +void vsuxseg3ei32(int32_t * op0, vuint32mf2_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_i32mf2_m))) +void vsuxseg3ei32(vbool64_t op0, int32_t * op1, vuint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_u32m1))) +void vsuxseg3ei32(uint32_t * op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_u32m1_m))) +void vsuxseg3ei32(vbool32_t op0, uint32_t * op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_u32m2))) +void vsuxseg3ei32(uint32_t * op0, vuint32m2_t op1, vuint32m2_t op2, vuint32m2_t op3, vuint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_u32m2_m))) +void vsuxseg3ei32(vbool16_t op0, uint32_t * op1, vuint32m2_t op2, vuint32m2_t op3, vuint32m2_t op4, vuint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_u32mf2))) +void vsuxseg3ei32(uint32_t * op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_u32mf2_m))) +void vsuxseg3ei32(vbool64_t op0, uint32_t * op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_i32m1))) +void vsuxseg4ei32(int32_t * op0, vuint32m1_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_i32m1_m))) +void vsuxseg4ei32(vbool32_t op0, int32_t * op1, vuint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_i32m2))) +void vsuxseg4ei32(int32_t * op0, vuint32m2_t op1, vint32m2_t op2, vint32m2_t op3, vint32m2_t op4, vint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_i32m2_m))) +void vsuxseg4ei32(vbool16_t op0, int32_t * op1, vuint32m2_t op2, vint32m2_t op3, vint32m2_t op4, vint32m2_t op5, vint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_i32mf2))) +void vsuxseg4ei32(int32_t * op0, vuint32mf2_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_i32mf2_m))) +void vsuxseg4ei32(vbool64_t op0, int32_t * op1, vuint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_u32m1))) +void vsuxseg4ei32(uint32_t * op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_u32m1_m))) +void vsuxseg4ei32(vbool32_t op0, uint32_t * op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_u32m2))) +void vsuxseg4ei32(uint32_t * op0, vuint32m2_t op1, vuint32m2_t op2, vuint32m2_t op3, vuint32m2_t op4, vuint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_u32m2_m))) +void vsuxseg4ei32(vbool16_t op0, uint32_t * op1, vuint32m2_t op2, vuint32m2_t op3, vuint32m2_t op4, vuint32m2_t op5, vuint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_u32mf2))) +void vsuxseg4ei32(uint32_t * op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_u32mf2_m))) +void vsuxseg4ei32(vbool64_t op0, uint32_t * op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_i32m1))) +void vsuxseg5ei32(int32_t * op0, vuint32m1_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_i32m1_m))) +void vsuxseg5ei32(vbool32_t op0, int32_t * op1, vuint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_i32mf2))) +void vsuxseg5ei32(int32_t * op0, vuint32mf2_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_i32mf2_m))) +void vsuxseg5ei32(vbool64_t op0, int32_t * op1, vuint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_u32m1))) +void vsuxseg5ei32(uint32_t * op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_u32m1_m))) +void vsuxseg5ei32(vbool32_t op0, uint32_t * op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_u32mf2))) +void vsuxseg5ei32(uint32_t * op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_u32mf2_m))) +void vsuxseg5ei32(vbool64_t op0, uint32_t * op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_i32m1))) +void vsuxseg6ei32(int32_t * op0, vuint32m1_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_i32m1_m))) +void vsuxseg6ei32(vbool32_t op0, int32_t * op1, vuint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_i32mf2))) +void vsuxseg6ei32(int32_t * op0, vuint32mf2_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_i32mf2_m))) +void vsuxseg6ei32(vbool64_t op0, int32_t * op1, vuint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_u32m1))) +void vsuxseg6ei32(uint32_t * op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_u32m1_m))) +void vsuxseg6ei32(vbool32_t op0, uint32_t * op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_u32mf2))) +void vsuxseg6ei32(uint32_t * op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_u32mf2_m))) +void vsuxseg6ei32(vbool64_t op0, uint32_t * op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_i32m1))) +void vsuxseg7ei32(int32_t * op0, vuint32m1_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_i32m1_m))) +void vsuxseg7ei32(vbool32_t op0, int32_t * op1, vuint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_i32mf2))) +void vsuxseg7ei32(int32_t * op0, vuint32mf2_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_i32mf2_m))) +void vsuxseg7ei32(vbool64_t op0, int32_t * op1, vuint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_u32m1))) +void vsuxseg7ei32(uint32_t * op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_u32m1_m))) +void vsuxseg7ei32(vbool32_t op0, uint32_t * op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_u32mf2))) +void vsuxseg7ei32(uint32_t * op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_u32mf2_m))) +void vsuxseg7ei32(vbool64_t op0, uint32_t * op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_i32m1))) +void vsuxseg8ei32(int32_t * op0, vuint32m1_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_i32m1_m))) +void vsuxseg8ei32(vbool32_t op0, int32_t * op1, vuint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, vint32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_i32mf2))) +void vsuxseg8ei32(int32_t * op0, vuint32mf2_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_i32mf2_m))) +void vsuxseg8ei32(vbool64_t op0, int32_t * op1, vuint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, vint32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_u32m1))) +void vsuxseg8ei32(uint32_t * op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_u32m1_m))) +void vsuxseg8ei32(vbool32_t op0, uint32_t * op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, vuint32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_u32mf2))) +void vsuxseg8ei32(uint32_t * op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_u32mf2_m))) +void vsuxseg8ei32(vbool64_t op0, uint32_t * op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, vuint32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_i32m1))) +void vsuxseg2ei64(int32_t * op0, vuint64m2_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_i32m1_m))) +void vsuxseg2ei64(vbool32_t op0, int32_t * op1, vuint64m2_t op2, vint32m1_t op3, vint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_i32m2))) +void vsuxseg2ei64(int32_t * op0, vuint64m4_t op1, vint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_i32m2_m))) +void vsuxseg2ei64(vbool16_t op0, int32_t * op1, vuint64m4_t op2, vint32m2_t op3, vint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_i32m4))) +void vsuxseg2ei64(int32_t * op0, vuint64m8_t op1, vint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_i32m4_m))) +void vsuxseg2ei64(vbool8_t op0, int32_t * op1, vuint64m8_t op2, vint32m4_t op3, vint32m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_i32mf2))) +void vsuxseg2ei64(int32_t * op0, vuint64m1_t op1, vint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_i32mf2_m))) +void vsuxseg2ei64(vbool64_t op0, int32_t * op1, vuint64m1_t op2, vint32mf2_t op3, vint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_u32m1))) +void vsuxseg2ei64(uint32_t * op0, vuint64m2_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_u32m1_m))) +void vsuxseg2ei64(vbool32_t op0, uint32_t * op1, vuint64m2_t op2, vuint32m1_t op3, vuint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_u32m2))) +void vsuxseg2ei64(uint32_t * op0, vuint64m4_t op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_u32m2_m))) +void vsuxseg2ei64(vbool16_t op0, uint32_t * op1, vuint64m4_t op2, vuint32m2_t op3, vuint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_u32m4))) +void vsuxseg2ei64(uint32_t * op0, vuint64m8_t op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_u32m4_m))) +void vsuxseg2ei64(vbool8_t op0, uint32_t * op1, vuint64m8_t op2, vuint32m4_t op3, vuint32m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_u32mf2))) +void vsuxseg2ei64(uint32_t * op0, vuint64m1_t op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_u32mf2_m))) +void vsuxseg2ei64(vbool64_t op0, uint32_t * op1, vuint64m1_t op2, vuint32mf2_t op3, vuint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_i32m1))) +void vsuxseg3ei64(int32_t * op0, vuint64m2_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_i32m1_m))) +void vsuxseg3ei64(vbool32_t op0, int32_t * op1, vuint64m2_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_i32m2))) +void vsuxseg3ei64(int32_t * op0, vuint64m4_t op1, vint32m2_t op2, vint32m2_t op3, vint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_i32m2_m))) +void vsuxseg3ei64(vbool16_t op0, int32_t * op1, vuint64m4_t op2, vint32m2_t op3, vint32m2_t op4, vint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_i32mf2))) +void vsuxseg3ei64(int32_t * op0, vuint64m1_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_i32mf2_m))) +void vsuxseg3ei64(vbool64_t op0, int32_t * op1, vuint64m1_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_u32m1))) +void vsuxseg3ei64(uint32_t * op0, vuint64m2_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_u32m1_m))) +void vsuxseg3ei64(vbool32_t op0, uint32_t * op1, vuint64m2_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_u32m2))) +void vsuxseg3ei64(uint32_t * op0, vuint64m4_t op1, vuint32m2_t op2, vuint32m2_t op3, vuint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_u32m2_m))) +void vsuxseg3ei64(vbool16_t op0, uint32_t * op1, vuint64m4_t op2, vuint32m2_t op3, vuint32m2_t op4, vuint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_u32mf2))) +void vsuxseg3ei64(uint32_t * op0, vuint64m1_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_u32mf2_m))) +void vsuxseg3ei64(vbool64_t op0, uint32_t * op1, vuint64m1_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_i32m1))) +void vsuxseg4ei64(int32_t * op0, vuint64m2_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_i32m1_m))) +void vsuxseg4ei64(vbool32_t op0, int32_t * op1, vuint64m2_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_i32m2))) +void vsuxseg4ei64(int32_t * op0, vuint64m4_t op1, vint32m2_t op2, vint32m2_t op3, vint32m2_t op4, vint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_i32m2_m))) +void vsuxseg4ei64(vbool16_t op0, int32_t * op1, vuint64m4_t op2, vint32m2_t op3, vint32m2_t op4, vint32m2_t op5, vint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_i32mf2))) +void vsuxseg4ei64(int32_t * op0, vuint64m1_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_i32mf2_m))) +void vsuxseg4ei64(vbool64_t op0, int32_t * op1, vuint64m1_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_u32m1))) +void vsuxseg4ei64(uint32_t * op0, vuint64m2_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_u32m1_m))) +void vsuxseg4ei64(vbool32_t op0, uint32_t * op1, vuint64m2_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_u32m2))) +void vsuxseg4ei64(uint32_t * op0, vuint64m4_t op1, vuint32m2_t op2, vuint32m2_t op3, vuint32m2_t op4, vuint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_u32m2_m))) +void vsuxseg4ei64(vbool16_t op0, uint32_t * op1, vuint64m4_t op2, vuint32m2_t op3, vuint32m2_t op4, vuint32m2_t op5, vuint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_u32mf2))) +void vsuxseg4ei64(uint32_t * op0, vuint64m1_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_u32mf2_m))) +void vsuxseg4ei64(vbool64_t op0, uint32_t * op1, vuint64m1_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_i32m1))) +void vsuxseg5ei64(int32_t * op0, vuint64m2_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_i32m1_m))) +void vsuxseg5ei64(vbool32_t op0, int32_t * op1, vuint64m2_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_i32mf2))) +void vsuxseg5ei64(int32_t * op0, vuint64m1_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_i32mf2_m))) +void vsuxseg5ei64(vbool64_t op0, int32_t * op1, vuint64m1_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_u32m1))) +void vsuxseg5ei64(uint32_t * op0, vuint64m2_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_u32m1_m))) +void vsuxseg5ei64(vbool32_t op0, uint32_t * op1, vuint64m2_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_u32mf2))) +void vsuxseg5ei64(uint32_t * op0, vuint64m1_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_u32mf2_m))) +void vsuxseg5ei64(vbool64_t op0, uint32_t * op1, vuint64m1_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_i32m1))) +void vsuxseg6ei64(int32_t * op0, vuint64m2_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_i32m1_m))) +void vsuxseg6ei64(vbool32_t op0, int32_t * op1, vuint64m2_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_i32mf2))) +void vsuxseg6ei64(int32_t * op0, vuint64m1_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_i32mf2_m))) +void vsuxseg6ei64(vbool64_t op0, int32_t * op1, vuint64m1_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_u32m1))) +void vsuxseg6ei64(uint32_t * op0, vuint64m2_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_u32m1_m))) +void vsuxseg6ei64(vbool32_t op0, uint32_t * op1, vuint64m2_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_u32mf2))) +void vsuxseg6ei64(uint32_t * op0, vuint64m1_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_u32mf2_m))) +void vsuxseg6ei64(vbool64_t op0, uint32_t * op1, vuint64m1_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_i32m1))) +void vsuxseg7ei64(int32_t * op0, vuint64m2_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_i32m1_m))) +void vsuxseg7ei64(vbool32_t op0, int32_t * op1, vuint64m2_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_i32mf2))) +void vsuxseg7ei64(int32_t * op0, vuint64m1_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_i32mf2_m))) +void vsuxseg7ei64(vbool64_t op0, int32_t * op1, vuint64m1_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_u32m1))) +void vsuxseg7ei64(uint32_t * op0, vuint64m2_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_u32m1_m))) +void vsuxseg7ei64(vbool32_t op0, uint32_t * op1, vuint64m2_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_u32mf2))) +void vsuxseg7ei64(uint32_t * op0, vuint64m1_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_u32mf2_m))) +void vsuxseg7ei64(vbool64_t op0, uint32_t * op1, vuint64m1_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_i32m1))) +void vsuxseg8ei64(int32_t * op0, vuint64m2_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_i32m1_m))) +void vsuxseg8ei64(vbool32_t op0, int32_t * op1, vuint64m2_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, vint32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_i32mf2))) +void vsuxseg8ei64(int32_t * op0, vuint64m1_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_i32mf2_m))) +void vsuxseg8ei64(vbool64_t op0, int32_t * op1, vuint64m1_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, vint32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_u32m1))) +void vsuxseg8ei64(uint32_t * op0, vuint64m2_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_u32m1_m))) +void vsuxseg8ei64(vbool32_t op0, uint32_t * op1, vuint64m2_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, vuint32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_u32mf2))) +void vsuxseg8ei64(uint32_t * op0, vuint64m1_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_u32mf2_m))) +void vsuxseg8ei64(vbool64_t op0, uint32_t * op1, vuint64m1_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, vuint32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_i64m1))) +void vsuxseg2ei8(int64_t * op0, vuint8mf8_t op1, vint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_i64m1_m))) +void vsuxseg2ei8(vbool64_t op0, int64_t * op1, vuint8mf8_t op2, vint64m1_t op3, vint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_i64m2))) +void vsuxseg2ei8(int64_t * op0, vuint8mf4_t op1, vint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_i64m2_m))) +void vsuxseg2ei8(vbool32_t op0, int64_t * op1, vuint8mf4_t op2, vint64m2_t op3, vint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_i64m4))) +void vsuxseg2ei8(int64_t * op0, vuint8mf2_t op1, vint64m4_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_i64m4_m))) +void vsuxseg2ei8(vbool16_t op0, int64_t * op1, vuint8mf2_t op2, vint64m4_t op3, vint64m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_u64m1))) +void vsuxseg2ei8(uint64_t * op0, vuint8mf8_t op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_u64m1_m))) +void vsuxseg2ei8(vbool64_t op0, uint64_t * op1, vuint8mf8_t op2, vuint64m1_t op3, vuint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_u64m2))) +void vsuxseg2ei8(uint64_t * op0, vuint8mf4_t op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_u64m2_m))) +void vsuxseg2ei8(vbool32_t op0, uint64_t * op1, vuint8mf4_t op2, vuint64m2_t op3, vuint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_u64m4))) +void vsuxseg2ei8(uint64_t * op0, vuint8mf2_t op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_u64m4_m))) +void vsuxseg2ei8(vbool16_t op0, uint64_t * op1, vuint8mf2_t op2, vuint64m4_t op3, vuint64m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_i64m1))) +void vsuxseg3ei8(int64_t * op0, vuint8mf8_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_i64m1_m))) +void vsuxseg3ei8(vbool64_t op0, int64_t * op1, vuint8mf8_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_i64m2))) +void vsuxseg3ei8(int64_t * op0, vuint8mf4_t op1, vint64m2_t op2, vint64m2_t op3, vint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_i64m2_m))) +void vsuxseg3ei8(vbool32_t op0, int64_t * op1, vuint8mf4_t op2, vint64m2_t op3, vint64m2_t op4, vint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_u64m1))) +void vsuxseg3ei8(uint64_t * op0, vuint8mf8_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_u64m1_m))) +void vsuxseg3ei8(vbool64_t op0, uint64_t * op1, vuint8mf8_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_u64m2))) +void vsuxseg3ei8(uint64_t * op0, vuint8mf4_t op1, vuint64m2_t op2, vuint64m2_t op3, vuint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_u64m2_m))) +void vsuxseg3ei8(vbool32_t op0, uint64_t * op1, vuint8mf4_t op2, vuint64m2_t op3, vuint64m2_t op4, vuint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_i64m1))) +void vsuxseg4ei8(int64_t * op0, vuint8mf8_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_i64m1_m))) +void vsuxseg4ei8(vbool64_t op0, int64_t * op1, vuint8mf8_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_i64m2))) +void vsuxseg4ei8(int64_t * op0, vuint8mf4_t op1, vint64m2_t op2, vint64m2_t op3, vint64m2_t op4, vint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_i64m2_m))) +void vsuxseg4ei8(vbool32_t op0, int64_t * op1, vuint8mf4_t op2, vint64m2_t op3, vint64m2_t op4, vint64m2_t op5, vint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_u64m1))) +void vsuxseg4ei8(uint64_t * op0, vuint8mf8_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_u64m1_m))) +void vsuxseg4ei8(vbool64_t op0, uint64_t * op1, vuint8mf8_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_u64m2))) +void vsuxseg4ei8(uint64_t * op0, vuint8mf4_t op1, vuint64m2_t op2, vuint64m2_t op3, vuint64m2_t op4, vuint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_u64m2_m))) +void vsuxseg4ei8(vbool32_t op0, uint64_t * op1, vuint8mf4_t op2, vuint64m2_t op3, vuint64m2_t op4, vuint64m2_t op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_i64m1))) +void vsuxseg5ei8(int64_t * op0, vuint8mf8_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_i64m1_m))) +void vsuxseg5ei8(vbool64_t op0, int64_t * op1, vuint8mf8_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_u64m1))) +void vsuxseg5ei8(uint64_t * op0, vuint8mf8_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_u64m1_m))) +void vsuxseg5ei8(vbool64_t op0, uint64_t * op1, vuint8mf8_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_i64m1))) +void vsuxseg6ei8(int64_t * op0, vuint8mf8_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_i64m1_m))) +void vsuxseg6ei8(vbool64_t op0, int64_t * op1, vuint8mf8_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_u64m1))) +void vsuxseg6ei8(uint64_t * op0, vuint8mf8_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_u64m1_m))) +void vsuxseg6ei8(vbool64_t op0, uint64_t * op1, vuint8mf8_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_i64m1))) +void vsuxseg7ei8(int64_t * op0, vuint8mf8_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_i64m1_m))) +void vsuxseg7ei8(vbool64_t op0, int64_t * op1, vuint8mf8_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_u64m1))) +void vsuxseg7ei8(uint64_t * op0, vuint8mf8_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_u64m1_m))) +void vsuxseg7ei8(vbool64_t op0, uint64_t * op1, vuint8mf8_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_i64m1))) +void vsuxseg8ei8(int64_t * op0, vuint8mf8_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_i64m1_m))) +void vsuxseg8ei8(vbool64_t op0, int64_t * op1, vuint8mf8_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, vint64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_u64m1))) +void vsuxseg8ei8(uint64_t * op0, vuint8mf8_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_u64m1_m))) +void vsuxseg8ei8(vbool64_t op0, uint64_t * op1, vuint8mf8_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, vuint64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_i64m1))) +void vsuxseg2ei16(int64_t * op0, vuint16mf4_t op1, vint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_i64m1_m))) +void vsuxseg2ei16(vbool64_t op0, int64_t * op1, vuint16mf4_t op2, vint64m1_t op3, vint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_i64m2))) +void vsuxseg2ei16(int64_t * op0, vuint16mf2_t op1, vint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_i64m2_m))) +void vsuxseg2ei16(vbool32_t op0, int64_t * op1, vuint16mf2_t op2, vint64m2_t op3, vint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_i64m4))) +void vsuxseg2ei16(int64_t * op0, vuint16m1_t op1, vint64m4_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_i64m4_m))) +void vsuxseg2ei16(vbool16_t op0, int64_t * op1, vuint16m1_t op2, vint64m4_t op3, vint64m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_u64m1))) +void vsuxseg2ei16(uint64_t * op0, vuint16mf4_t op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_u64m1_m))) +void vsuxseg2ei16(vbool64_t op0, uint64_t * op1, vuint16mf4_t op2, vuint64m1_t op3, vuint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_u64m2))) +void vsuxseg2ei16(uint64_t * op0, vuint16mf2_t op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_u64m2_m))) +void vsuxseg2ei16(vbool32_t op0, uint64_t * op1, vuint16mf2_t op2, vuint64m2_t op3, vuint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_u64m4))) +void vsuxseg2ei16(uint64_t * op0, vuint16m1_t op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_u64m4_m))) +void vsuxseg2ei16(vbool16_t op0, uint64_t * op1, vuint16m1_t op2, vuint64m4_t op3, vuint64m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_i64m1))) +void vsuxseg3ei16(int64_t * op0, vuint16mf4_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_i64m1_m))) +void vsuxseg3ei16(vbool64_t op0, int64_t * op1, vuint16mf4_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_i64m2))) +void vsuxseg3ei16(int64_t * op0, vuint16mf2_t op1, vint64m2_t op2, vint64m2_t op3, vint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_i64m2_m))) +void vsuxseg3ei16(vbool32_t op0, int64_t * op1, vuint16mf2_t op2, vint64m2_t op3, vint64m2_t op4, vint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_u64m1))) +void vsuxseg3ei16(uint64_t * op0, vuint16mf4_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_u64m1_m))) +void vsuxseg3ei16(vbool64_t op0, uint64_t * op1, vuint16mf4_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_u64m2))) +void vsuxseg3ei16(uint64_t * op0, vuint16mf2_t op1, vuint64m2_t op2, vuint64m2_t op3, vuint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_u64m2_m))) +void vsuxseg3ei16(vbool32_t op0, uint64_t * op1, vuint16mf2_t op2, vuint64m2_t op3, vuint64m2_t op4, vuint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_i64m1))) +void vsuxseg4ei16(int64_t * op0, vuint16mf4_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_i64m1_m))) +void vsuxseg4ei16(vbool64_t op0, int64_t * op1, vuint16mf4_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_i64m2))) +void vsuxseg4ei16(int64_t * op0, vuint16mf2_t op1, vint64m2_t op2, vint64m2_t op3, vint64m2_t op4, vint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_i64m2_m))) +void vsuxseg4ei16(vbool32_t op0, int64_t * op1, vuint16mf2_t op2, vint64m2_t op3, vint64m2_t op4, vint64m2_t op5, vint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_u64m1))) +void vsuxseg4ei16(uint64_t * op0, vuint16mf4_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_u64m1_m))) +void vsuxseg4ei16(vbool64_t op0, uint64_t * op1, vuint16mf4_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_u64m2))) +void vsuxseg4ei16(uint64_t * op0, vuint16mf2_t op1, vuint64m2_t op2, vuint64m2_t op3, vuint64m2_t op4, vuint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_u64m2_m))) +void vsuxseg4ei16(vbool32_t op0, uint64_t * op1, vuint16mf2_t op2, vuint64m2_t op3, vuint64m2_t op4, vuint64m2_t op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_i64m1))) +void vsuxseg5ei16(int64_t * op0, vuint16mf4_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_i64m1_m))) +void vsuxseg5ei16(vbool64_t op0, int64_t * op1, vuint16mf4_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_u64m1))) +void vsuxseg5ei16(uint64_t * op0, vuint16mf4_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_u64m1_m))) +void vsuxseg5ei16(vbool64_t op0, uint64_t * op1, vuint16mf4_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_i64m1))) +void vsuxseg6ei16(int64_t * op0, vuint16mf4_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_i64m1_m))) +void vsuxseg6ei16(vbool64_t op0, int64_t * op1, vuint16mf4_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_u64m1))) +void vsuxseg6ei16(uint64_t * op0, vuint16mf4_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_u64m1_m))) +void vsuxseg6ei16(vbool64_t op0, uint64_t * op1, vuint16mf4_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_i64m1))) +void vsuxseg7ei16(int64_t * op0, vuint16mf4_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_i64m1_m))) +void vsuxseg7ei16(vbool64_t op0, int64_t * op1, vuint16mf4_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_u64m1))) +void vsuxseg7ei16(uint64_t * op0, vuint16mf4_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_u64m1_m))) +void vsuxseg7ei16(vbool64_t op0, uint64_t * op1, vuint16mf4_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_i64m1))) +void vsuxseg8ei16(int64_t * op0, vuint16mf4_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_i64m1_m))) +void vsuxseg8ei16(vbool64_t op0, int64_t * op1, vuint16mf4_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, vint64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_u64m1))) +void vsuxseg8ei16(uint64_t * op0, vuint16mf4_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_u64m1_m))) +void vsuxseg8ei16(vbool64_t op0, uint64_t * op1, vuint16mf4_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, vuint64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_i64m1))) +void vsuxseg2ei32(int64_t * op0, vuint32mf2_t op1, vint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_i64m1_m))) +void vsuxseg2ei32(vbool64_t op0, int64_t * op1, vuint32mf2_t op2, vint64m1_t op3, vint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_i64m2))) +void vsuxseg2ei32(int64_t * op0, vuint32m1_t op1, vint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_i64m2_m))) +void vsuxseg2ei32(vbool32_t op0, int64_t * op1, vuint32m1_t op2, vint64m2_t op3, vint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_i64m4))) +void vsuxseg2ei32(int64_t * op0, vuint32m2_t op1, vint64m4_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_i64m4_m))) +void vsuxseg2ei32(vbool16_t op0, int64_t * op1, vuint32m2_t op2, vint64m4_t op3, vint64m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_u64m1))) +void vsuxseg2ei32(uint64_t * op0, vuint32mf2_t op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_u64m1_m))) +void vsuxseg2ei32(vbool64_t op0, uint64_t * op1, vuint32mf2_t op2, vuint64m1_t op3, vuint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_u64m2))) +void vsuxseg2ei32(uint64_t * op0, vuint32m1_t op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_u64m2_m))) +void vsuxseg2ei32(vbool32_t op0, uint64_t * op1, vuint32m1_t op2, vuint64m2_t op3, vuint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_u64m4))) +void vsuxseg2ei32(uint64_t * op0, vuint32m2_t op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_u64m4_m))) +void vsuxseg2ei32(vbool16_t op0, uint64_t * op1, vuint32m2_t op2, vuint64m4_t op3, vuint64m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_i64m1))) +void vsuxseg3ei32(int64_t * op0, vuint32mf2_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_i64m1_m))) +void vsuxseg3ei32(vbool64_t op0, int64_t * op1, vuint32mf2_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_i64m2))) +void vsuxseg3ei32(int64_t * op0, vuint32m1_t op1, vint64m2_t op2, vint64m2_t op3, vint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_i64m2_m))) +void vsuxseg3ei32(vbool32_t op0, int64_t * op1, vuint32m1_t op2, vint64m2_t op3, vint64m2_t op4, vint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_u64m1))) +void vsuxseg3ei32(uint64_t * op0, vuint32mf2_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_u64m1_m))) +void vsuxseg3ei32(vbool64_t op0, uint64_t * op1, vuint32mf2_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_u64m2))) +void vsuxseg3ei32(uint64_t * op0, vuint32m1_t op1, vuint64m2_t op2, vuint64m2_t op3, vuint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_u64m2_m))) +void vsuxseg3ei32(vbool32_t op0, uint64_t * op1, vuint32m1_t op2, vuint64m2_t op3, vuint64m2_t op4, vuint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_i64m1))) +void vsuxseg4ei32(int64_t * op0, vuint32mf2_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_i64m1_m))) +void vsuxseg4ei32(vbool64_t op0, int64_t * op1, vuint32mf2_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_i64m2))) +void vsuxseg4ei32(int64_t * op0, vuint32m1_t op1, vint64m2_t op2, vint64m2_t op3, vint64m2_t op4, vint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_i64m2_m))) +void vsuxseg4ei32(vbool32_t op0, int64_t * op1, vuint32m1_t op2, vint64m2_t op3, vint64m2_t op4, vint64m2_t op5, vint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_u64m1))) +void vsuxseg4ei32(uint64_t * op0, vuint32mf2_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_u64m1_m))) +void vsuxseg4ei32(vbool64_t op0, uint64_t * op1, vuint32mf2_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_u64m2))) +void vsuxseg4ei32(uint64_t * op0, vuint32m1_t op1, vuint64m2_t op2, vuint64m2_t op3, vuint64m2_t op4, vuint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_u64m2_m))) +void vsuxseg4ei32(vbool32_t op0, uint64_t * op1, vuint32m1_t op2, vuint64m2_t op3, vuint64m2_t op4, vuint64m2_t op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_i64m1))) +void vsuxseg5ei32(int64_t * op0, vuint32mf2_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_i64m1_m))) +void vsuxseg5ei32(vbool64_t op0, int64_t * op1, vuint32mf2_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_u64m1))) +void vsuxseg5ei32(uint64_t * op0, vuint32mf2_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_u64m1_m))) +void vsuxseg5ei32(vbool64_t op0, uint64_t * op1, vuint32mf2_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_i64m1))) +void vsuxseg6ei32(int64_t * op0, vuint32mf2_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_i64m1_m))) +void vsuxseg6ei32(vbool64_t op0, int64_t * op1, vuint32mf2_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_u64m1))) +void vsuxseg6ei32(uint64_t * op0, vuint32mf2_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_u64m1_m))) +void vsuxseg6ei32(vbool64_t op0, uint64_t * op1, vuint32mf2_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_i64m1))) +void vsuxseg7ei32(int64_t * op0, vuint32mf2_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_i64m1_m))) +void vsuxseg7ei32(vbool64_t op0, int64_t * op1, vuint32mf2_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_u64m1))) +void vsuxseg7ei32(uint64_t * op0, vuint32mf2_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_u64m1_m))) +void vsuxseg7ei32(vbool64_t op0, uint64_t * op1, vuint32mf2_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_i64m1))) +void vsuxseg8ei32(int64_t * op0, vuint32mf2_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_i64m1_m))) +void vsuxseg8ei32(vbool64_t op0, int64_t * op1, vuint32mf2_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, vint64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_u64m1))) +void vsuxseg8ei32(uint64_t * op0, vuint32mf2_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_u64m1_m))) +void vsuxseg8ei32(vbool64_t op0, uint64_t * op1, vuint32mf2_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, vuint64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_i64m1))) +void vsuxseg2ei64(int64_t * op0, vuint64m1_t op1, vint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_i64m1_m))) +void vsuxseg2ei64(vbool64_t op0, int64_t * op1, vuint64m1_t op2, vint64m1_t op3, vint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_i64m2))) +void vsuxseg2ei64(int64_t * op0, vuint64m2_t op1, vint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_i64m2_m))) +void vsuxseg2ei64(vbool32_t op0, int64_t * op1, vuint64m2_t op2, vint64m2_t op3, vint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_i64m4))) +void vsuxseg2ei64(int64_t * op0, vuint64m4_t op1, vint64m4_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_i64m4_m))) +void vsuxseg2ei64(vbool16_t op0, int64_t * op1, vuint64m4_t op2, vint64m4_t op3, vint64m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_u64m1))) +void vsuxseg2ei64(uint64_t * op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_u64m1_m))) +void vsuxseg2ei64(vbool64_t op0, uint64_t * op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_u64m2))) +void vsuxseg2ei64(uint64_t * op0, vuint64m2_t op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_u64m2_m))) +void vsuxseg2ei64(vbool32_t op0, uint64_t * op1, vuint64m2_t op2, vuint64m2_t op3, vuint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_u64m4))) +void vsuxseg2ei64(uint64_t * op0, vuint64m4_t op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_u64m4_m))) +void vsuxseg2ei64(vbool16_t op0, uint64_t * op1, vuint64m4_t op2, vuint64m4_t op3, vuint64m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_i64m1))) +void vsuxseg3ei64(int64_t * op0, vuint64m1_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_i64m1_m))) +void vsuxseg3ei64(vbool64_t op0, int64_t * op1, vuint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_i64m2))) +void vsuxseg3ei64(int64_t * op0, vuint64m2_t op1, vint64m2_t op2, vint64m2_t op3, vint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_i64m2_m))) +void vsuxseg3ei64(vbool32_t op0, int64_t * op1, vuint64m2_t op2, vint64m2_t op3, vint64m2_t op4, vint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_u64m1))) +void vsuxseg3ei64(uint64_t * op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_u64m1_m))) +void vsuxseg3ei64(vbool64_t op0, uint64_t * op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_u64m2))) +void vsuxseg3ei64(uint64_t * op0, vuint64m2_t op1, vuint64m2_t op2, vuint64m2_t op3, vuint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_u64m2_m))) +void vsuxseg3ei64(vbool32_t op0, uint64_t * op1, vuint64m2_t op2, vuint64m2_t op3, vuint64m2_t op4, vuint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_i64m1))) +void vsuxseg4ei64(int64_t * op0, vuint64m1_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_i64m1_m))) +void vsuxseg4ei64(vbool64_t op0, int64_t * op1, vuint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_i64m2))) +void vsuxseg4ei64(int64_t * op0, vuint64m2_t op1, vint64m2_t op2, vint64m2_t op3, vint64m2_t op4, vint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_i64m2_m))) +void vsuxseg4ei64(vbool32_t op0, int64_t * op1, vuint64m2_t op2, vint64m2_t op3, vint64m2_t op4, vint64m2_t op5, vint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_u64m1))) +void vsuxseg4ei64(uint64_t * op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_u64m1_m))) +void vsuxseg4ei64(vbool64_t op0, uint64_t * op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_u64m2))) +void vsuxseg4ei64(uint64_t * op0, vuint64m2_t op1, vuint64m2_t op2, vuint64m2_t op3, vuint64m2_t op4, vuint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_u64m2_m))) +void vsuxseg4ei64(vbool32_t op0, uint64_t * op1, vuint64m2_t op2, vuint64m2_t op3, vuint64m2_t op4, vuint64m2_t op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_i64m1))) +void vsuxseg5ei64(int64_t * op0, vuint64m1_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_i64m1_m))) +void vsuxseg5ei64(vbool64_t op0, int64_t * op1, vuint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_u64m1))) +void vsuxseg5ei64(uint64_t * op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_u64m1_m))) +void vsuxseg5ei64(vbool64_t op0, uint64_t * op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_i64m1))) +void vsuxseg6ei64(int64_t * op0, vuint64m1_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_i64m1_m))) +void vsuxseg6ei64(vbool64_t op0, int64_t * op1, vuint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_u64m1))) +void vsuxseg6ei64(uint64_t * op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_u64m1_m))) +void vsuxseg6ei64(vbool64_t op0, uint64_t * op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_i64m1))) +void vsuxseg7ei64(int64_t * op0, vuint64m1_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_i64m1_m))) +void vsuxseg7ei64(vbool64_t op0, int64_t * op1, vuint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_u64m1))) +void vsuxseg7ei64(uint64_t * op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_u64m1_m))) +void vsuxseg7ei64(vbool64_t op0, uint64_t * op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_i64m1))) +void vsuxseg8ei64(int64_t * op0, vuint64m1_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_i64m1_m))) +void vsuxseg8ei64(vbool64_t op0, int64_t * op1, vuint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, vint64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_u64m1))) +void vsuxseg8ei64(uint64_t * op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_u64m1_m))) +void vsuxseg8ei64(vbool64_t op0, uint64_t * op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, vuint64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_i8m1))) +void vsoxseg2ei8(int8_t * op0, vuint8m1_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_i8m1_m))) +void vsoxseg2ei8(vbool8_t op0, int8_t * op1, vuint8m1_t op2, vint8m1_t op3, vint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_i8m2))) +void vsoxseg2ei8(int8_t * op0, vuint8m2_t op1, vint8m2_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_i8m2_m))) +void vsoxseg2ei8(vbool4_t op0, int8_t * op1, vuint8m2_t op2, vint8m2_t op3, vint8m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_i8m4))) +void vsoxseg2ei8(int8_t * op0, vuint8m4_t op1, vint8m4_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_i8m4_m))) +void vsoxseg2ei8(vbool2_t op0, int8_t * op1, vuint8m4_t op2, vint8m4_t op3, vint8m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_i8mf2))) +void vsoxseg2ei8(int8_t * op0, vuint8mf2_t op1, vint8mf2_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_i8mf2_m))) +void vsoxseg2ei8(vbool16_t op0, int8_t * op1, vuint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_i8mf4))) +void vsoxseg2ei8(int8_t * op0, vuint8mf4_t op1, vint8mf4_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_i8mf4_m))) +void vsoxseg2ei8(vbool32_t op0, int8_t * op1, vuint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_i8mf8))) +void vsoxseg2ei8(int8_t * op0, vuint8mf8_t op1, vint8mf8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_i8mf8_m))) +void vsoxseg2ei8(vbool64_t op0, int8_t * op1, vuint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_u8m1))) +void vsoxseg2ei8(uint8_t * op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_u8m1_m))) +void vsoxseg2ei8(vbool8_t op0, uint8_t * op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_u8m2))) +void vsoxseg2ei8(uint8_t * op0, vuint8m2_t op1, vuint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_u8m2_m))) +void vsoxseg2ei8(vbool4_t op0, uint8_t * op1, vuint8m2_t op2, vuint8m2_t op3, vuint8m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_u8m4))) +void vsoxseg2ei8(uint8_t * op0, vuint8m4_t op1, vuint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_u8m4_m))) +void vsoxseg2ei8(vbool2_t op0, uint8_t * op1, vuint8m4_t op2, vuint8m4_t op3, vuint8m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_u8mf2))) +void vsoxseg2ei8(uint8_t * op0, vuint8mf2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_u8mf2_m))) +void vsoxseg2ei8(vbool16_t op0, uint8_t * op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_u8mf4))) +void vsoxseg2ei8(uint8_t * op0, vuint8mf4_t op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_u8mf4_m))) +void vsoxseg2ei8(vbool32_t op0, uint8_t * op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_u8mf8))) +void vsoxseg2ei8(uint8_t * op0, vuint8mf8_t op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_u8mf8_m))) +void vsoxseg2ei8(vbool64_t op0, uint8_t * op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_i8m1))) +void vsoxseg3ei8(int8_t * op0, vuint8m1_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_i8m1_m))) +void vsoxseg3ei8(vbool8_t op0, int8_t * op1, vuint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_i8m2))) +void vsoxseg3ei8(int8_t * op0, vuint8m2_t op1, vint8m2_t op2, vint8m2_t op3, vint8m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_i8m2_m))) +void vsoxseg3ei8(vbool4_t op0, int8_t * op1, vuint8m2_t op2, vint8m2_t op3, vint8m2_t op4, vint8m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_i8mf2))) +void vsoxseg3ei8(int8_t * op0, vuint8mf2_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_i8mf2_m))) +void vsoxseg3ei8(vbool16_t op0, int8_t * op1, vuint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_i8mf4))) +void vsoxseg3ei8(int8_t * op0, vuint8mf4_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_i8mf4_m))) +void vsoxseg3ei8(vbool32_t op0, int8_t * op1, vuint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_i8mf8))) +void vsoxseg3ei8(int8_t * op0, vuint8mf8_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_i8mf8_m))) +void vsoxseg3ei8(vbool64_t op0, int8_t * op1, vuint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_u8m1))) +void vsoxseg3ei8(uint8_t * op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_u8m1_m))) +void vsoxseg3ei8(vbool8_t op0, uint8_t * op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_u8m2))) +void vsoxseg3ei8(uint8_t * op0, vuint8m2_t op1, vuint8m2_t op2, vuint8m2_t op3, vuint8m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_u8m2_m))) +void vsoxseg3ei8(vbool4_t op0, uint8_t * op1, vuint8m2_t op2, vuint8m2_t op3, vuint8m2_t op4, vuint8m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_u8mf2))) +void vsoxseg3ei8(uint8_t * op0, vuint8mf2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_u8mf2_m))) +void vsoxseg3ei8(vbool16_t op0, uint8_t * op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_u8mf4))) +void vsoxseg3ei8(uint8_t * op0, vuint8mf4_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_u8mf4_m))) +void vsoxseg3ei8(vbool32_t op0, uint8_t * op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_u8mf8))) +void vsoxseg3ei8(uint8_t * op0, vuint8mf8_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_u8mf8_m))) +void vsoxseg3ei8(vbool64_t op0, uint8_t * op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_i8m1))) +void vsoxseg4ei8(int8_t * op0, vuint8m1_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_i8m1_m))) +void vsoxseg4ei8(vbool8_t op0, int8_t * op1, vuint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_i8m2))) +void vsoxseg4ei8(int8_t * op0, vuint8m2_t op1, vint8m2_t op2, vint8m2_t op3, vint8m2_t op4, vint8m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_i8m2_m))) +void vsoxseg4ei8(vbool4_t op0, int8_t * op1, vuint8m2_t op2, vint8m2_t op3, vint8m2_t op4, vint8m2_t op5, vint8m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_i8mf2))) +void vsoxseg4ei8(int8_t * op0, vuint8mf2_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_i8mf2_m))) +void vsoxseg4ei8(vbool16_t op0, int8_t * op1, vuint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_i8mf4))) +void vsoxseg4ei8(int8_t * op0, vuint8mf4_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_i8mf4_m))) +void vsoxseg4ei8(vbool32_t op0, int8_t * op1, vuint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_i8mf8))) +void vsoxseg4ei8(int8_t * op0, vuint8mf8_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_i8mf8_m))) +void vsoxseg4ei8(vbool64_t op0, int8_t * op1, vuint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_u8m1))) +void vsoxseg4ei8(uint8_t * op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_u8m1_m))) +void vsoxseg4ei8(vbool8_t op0, uint8_t * op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_u8m2))) +void vsoxseg4ei8(uint8_t * op0, vuint8m2_t op1, vuint8m2_t op2, vuint8m2_t op3, vuint8m2_t op4, vuint8m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_u8m2_m))) +void vsoxseg4ei8(vbool4_t op0, uint8_t * op1, vuint8m2_t op2, vuint8m2_t op3, vuint8m2_t op4, vuint8m2_t op5, vuint8m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_u8mf2))) +void vsoxseg4ei8(uint8_t * op0, vuint8mf2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_u8mf2_m))) +void vsoxseg4ei8(vbool16_t op0, uint8_t * op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_u8mf4))) +void vsoxseg4ei8(uint8_t * op0, vuint8mf4_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_u8mf4_m))) +void vsoxseg4ei8(vbool32_t op0, uint8_t * op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_u8mf8))) +void vsoxseg4ei8(uint8_t * op0, vuint8mf8_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_u8mf8_m))) +void vsoxseg4ei8(vbool64_t op0, uint8_t * op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_i8m1))) +void vsoxseg5ei8(int8_t * op0, vuint8m1_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_i8m1_m))) +void vsoxseg5ei8(vbool8_t op0, int8_t * op1, vuint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_i8mf2))) +void vsoxseg5ei8(int8_t * op0, vuint8mf2_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_i8mf2_m))) +void vsoxseg5ei8(vbool16_t op0, int8_t * op1, vuint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_i8mf4))) +void vsoxseg5ei8(int8_t * op0, vuint8mf4_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_i8mf4_m))) +void vsoxseg5ei8(vbool32_t op0, int8_t * op1, vuint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_i8mf8))) +void vsoxseg5ei8(int8_t * op0, vuint8mf8_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_i8mf8_m))) +void vsoxseg5ei8(vbool64_t op0, int8_t * op1, vuint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_u8m1))) +void vsoxseg5ei8(uint8_t * op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_u8m1_m))) +void vsoxseg5ei8(vbool8_t op0, uint8_t * op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_u8mf2))) +void vsoxseg5ei8(uint8_t * op0, vuint8mf2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_u8mf2_m))) +void vsoxseg5ei8(vbool16_t op0, uint8_t * op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_u8mf4))) +void vsoxseg5ei8(uint8_t * op0, vuint8mf4_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_u8mf4_m))) +void vsoxseg5ei8(vbool32_t op0, uint8_t * op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_u8mf8))) +void vsoxseg5ei8(uint8_t * op0, vuint8mf8_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_u8mf8_m))) +void vsoxseg5ei8(vbool64_t op0, uint8_t * op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_i8m1))) +void vsoxseg6ei8(int8_t * op0, vuint8m1_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_i8m1_m))) +void vsoxseg6ei8(vbool8_t op0, int8_t * op1, vuint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_i8mf2))) +void vsoxseg6ei8(int8_t * op0, vuint8mf2_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_i8mf2_m))) +void vsoxseg6ei8(vbool16_t op0, int8_t * op1, vuint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_i8mf4))) +void vsoxseg6ei8(int8_t * op0, vuint8mf4_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_i8mf4_m))) +void vsoxseg6ei8(vbool32_t op0, int8_t * op1, vuint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_i8mf8))) +void vsoxseg6ei8(int8_t * op0, vuint8mf8_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_i8mf8_m))) +void vsoxseg6ei8(vbool64_t op0, int8_t * op1, vuint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_u8m1))) +void vsoxseg6ei8(uint8_t * op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_u8m1_m))) +void vsoxseg6ei8(vbool8_t op0, uint8_t * op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_u8mf2))) +void vsoxseg6ei8(uint8_t * op0, vuint8mf2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_u8mf2_m))) +void vsoxseg6ei8(vbool16_t op0, uint8_t * op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_u8mf4))) +void vsoxseg6ei8(uint8_t * op0, vuint8mf4_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_u8mf4_m))) +void vsoxseg6ei8(vbool32_t op0, uint8_t * op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_u8mf8))) +void vsoxseg6ei8(uint8_t * op0, vuint8mf8_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_u8mf8_m))) +void vsoxseg6ei8(vbool64_t op0, uint8_t * op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_i8m1))) +void vsoxseg7ei8(int8_t * op0, vuint8m1_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_i8m1_m))) +void vsoxseg7ei8(vbool8_t op0, int8_t * op1, vuint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_i8mf2))) +void vsoxseg7ei8(int8_t * op0, vuint8mf2_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_i8mf2_m))) +void vsoxseg7ei8(vbool16_t op0, int8_t * op1, vuint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_i8mf4))) +void vsoxseg7ei8(int8_t * op0, vuint8mf4_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_i8mf4_m))) +void vsoxseg7ei8(vbool32_t op0, int8_t * op1, vuint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_i8mf8))) +void vsoxseg7ei8(int8_t * op0, vuint8mf8_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_i8mf8_m))) +void vsoxseg7ei8(vbool64_t op0, int8_t * op1, vuint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_u8m1))) +void vsoxseg7ei8(uint8_t * op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_u8m1_m))) +void vsoxseg7ei8(vbool8_t op0, uint8_t * op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_u8mf2))) +void vsoxseg7ei8(uint8_t * op0, vuint8mf2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_u8mf2_m))) +void vsoxseg7ei8(vbool16_t op0, uint8_t * op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_u8mf4))) +void vsoxseg7ei8(uint8_t * op0, vuint8mf4_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_u8mf4_m))) +void vsoxseg7ei8(vbool32_t op0, uint8_t * op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_u8mf8))) +void vsoxseg7ei8(uint8_t * op0, vuint8mf8_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_u8mf8_m))) +void vsoxseg7ei8(vbool64_t op0, uint8_t * op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_i8m1))) +void vsoxseg8ei8(int8_t * op0, vuint8m1_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_i8m1_m))) +void vsoxseg8ei8(vbool8_t op0, int8_t * op1, vuint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, vint8m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_i8mf2))) +void vsoxseg8ei8(int8_t * op0, vuint8mf2_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_i8mf2_m))) +void vsoxseg8ei8(vbool16_t op0, int8_t * op1, vuint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, vint8mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_i8mf4))) +void vsoxseg8ei8(int8_t * op0, vuint8mf4_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_i8mf4_m))) +void vsoxseg8ei8(vbool32_t op0, int8_t * op1, vuint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, vint8mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_i8mf8))) +void vsoxseg8ei8(int8_t * op0, vuint8mf8_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_i8mf8_m))) +void vsoxseg8ei8(vbool64_t op0, int8_t * op1, vuint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, vint8mf8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_u8m1))) +void vsoxseg8ei8(uint8_t * op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_u8m1_m))) +void vsoxseg8ei8(vbool8_t op0, uint8_t * op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, vuint8m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_u8mf2))) +void vsoxseg8ei8(uint8_t * op0, vuint8mf2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_u8mf2_m))) +void vsoxseg8ei8(vbool16_t op0, uint8_t * op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, vuint8mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_u8mf4))) +void vsoxseg8ei8(uint8_t * op0, vuint8mf4_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_u8mf4_m))) +void vsoxseg8ei8(vbool32_t op0, uint8_t * op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, vuint8mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_u8mf8))) +void vsoxseg8ei8(uint8_t * op0, vuint8mf8_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_u8mf8_m))) +void vsoxseg8ei8(vbool64_t op0, uint8_t * op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, vuint8mf8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_i8m1))) +void vsoxseg2ei16(int8_t * op0, vuint16m2_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_i8m1_m))) +void vsoxseg2ei16(vbool8_t op0, int8_t * op1, vuint16m2_t op2, vint8m1_t op3, vint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_i8m2))) +void vsoxseg2ei16(int8_t * op0, vuint16m4_t op1, vint8m2_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_i8m2_m))) +void vsoxseg2ei16(vbool4_t op0, int8_t * op1, vuint16m4_t op2, vint8m2_t op3, vint8m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_i8m4))) +void vsoxseg2ei16(int8_t * op0, vuint16m8_t op1, vint8m4_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_i8m4_m))) +void vsoxseg2ei16(vbool2_t op0, int8_t * op1, vuint16m8_t op2, vint8m4_t op3, vint8m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_i8mf2))) +void vsoxseg2ei16(int8_t * op0, vuint16m1_t op1, vint8mf2_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_i8mf2_m))) +void vsoxseg2ei16(vbool16_t op0, int8_t * op1, vuint16m1_t op2, vint8mf2_t op3, vint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_i8mf4))) +void vsoxseg2ei16(int8_t * op0, vuint16mf2_t op1, vint8mf4_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_i8mf4_m))) +void vsoxseg2ei16(vbool32_t op0, int8_t * op1, vuint16mf2_t op2, vint8mf4_t op3, vint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_i8mf8))) +void vsoxseg2ei16(int8_t * op0, vuint16mf4_t op1, vint8mf8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_i8mf8_m))) +void vsoxseg2ei16(vbool64_t op0, int8_t * op1, vuint16mf4_t op2, vint8mf8_t op3, vint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_u8m1))) +void vsoxseg2ei16(uint8_t * op0, vuint16m2_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_u8m1_m))) +void vsoxseg2ei16(vbool8_t op0, uint8_t * op1, vuint16m2_t op2, vuint8m1_t op3, vuint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_u8m2))) +void vsoxseg2ei16(uint8_t * op0, vuint16m4_t op1, vuint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_u8m2_m))) +void vsoxseg2ei16(vbool4_t op0, uint8_t * op1, vuint16m4_t op2, vuint8m2_t op3, vuint8m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_u8m4))) +void vsoxseg2ei16(uint8_t * op0, vuint16m8_t op1, vuint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_u8m4_m))) +void vsoxseg2ei16(vbool2_t op0, uint8_t * op1, vuint16m8_t op2, vuint8m4_t op3, vuint8m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_u8mf2))) +void vsoxseg2ei16(uint8_t * op0, vuint16m1_t op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_u8mf2_m))) +void vsoxseg2ei16(vbool16_t op0, uint8_t * op1, vuint16m1_t op2, vuint8mf2_t op3, vuint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_u8mf4))) +void vsoxseg2ei16(uint8_t * op0, vuint16mf2_t op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_u8mf4_m))) +void vsoxseg2ei16(vbool32_t op0, uint8_t * op1, vuint16mf2_t op2, vuint8mf4_t op3, vuint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_u8mf8))) +void vsoxseg2ei16(uint8_t * op0, vuint16mf4_t op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_u8mf8_m))) +void vsoxseg2ei16(vbool64_t op0, uint8_t * op1, vuint16mf4_t op2, vuint8mf8_t op3, vuint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_i8m1))) +void vsoxseg3ei16(int8_t * op0, vuint16m2_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_i8m1_m))) +void vsoxseg3ei16(vbool8_t op0, int8_t * op1, vuint16m2_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_i8m2))) +void vsoxseg3ei16(int8_t * op0, vuint16m4_t op1, vint8m2_t op2, vint8m2_t op3, vint8m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_i8m2_m))) +void vsoxseg3ei16(vbool4_t op0, int8_t * op1, vuint16m4_t op2, vint8m2_t op3, vint8m2_t op4, vint8m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_i8mf2))) +void vsoxseg3ei16(int8_t * op0, vuint16m1_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_i8mf2_m))) +void vsoxseg3ei16(vbool16_t op0, int8_t * op1, vuint16m1_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_i8mf4))) +void vsoxseg3ei16(int8_t * op0, vuint16mf2_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_i8mf4_m))) +void vsoxseg3ei16(vbool32_t op0, int8_t * op1, vuint16mf2_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_i8mf8))) +void vsoxseg3ei16(int8_t * op0, vuint16mf4_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_i8mf8_m))) +void vsoxseg3ei16(vbool64_t op0, int8_t * op1, vuint16mf4_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_u8m1))) +void vsoxseg3ei16(uint8_t * op0, vuint16m2_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_u8m1_m))) +void vsoxseg3ei16(vbool8_t op0, uint8_t * op1, vuint16m2_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_u8m2))) +void vsoxseg3ei16(uint8_t * op0, vuint16m4_t op1, vuint8m2_t op2, vuint8m2_t op3, vuint8m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_u8m2_m))) +void vsoxseg3ei16(vbool4_t op0, uint8_t * op1, vuint16m4_t op2, vuint8m2_t op3, vuint8m2_t op4, vuint8m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_u8mf2))) +void vsoxseg3ei16(uint8_t * op0, vuint16m1_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_u8mf2_m))) +void vsoxseg3ei16(vbool16_t op0, uint8_t * op1, vuint16m1_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_u8mf4))) +void vsoxseg3ei16(uint8_t * op0, vuint16mf2_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_u8mf4_m))) +void vsoxseg3ei16(vbool32_t op0, uint8_t * op1, vuint16mf2_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_u8mf8))) +void vsoxseg3ei16(uint8_t * op0, vuint16mf4_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_u8mf8_m))) +void vsoxseg3ei16(vbool64_t op0, uint8_t * op1, vuint16mf4_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_i8m1))) +void vsoxseg4ei16(int8_t * op0, vuint16m2_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_i8m1_m))) +void vsoxseg4ei16(vbool8_t op0, int8_t * op1, vuint16m2_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_i8m2))) +void vsoxseg4ei16(int8_t * op0, vuint16m4_t op1, vint8m2_t op2, vint8m2_t op3, vint8m2_t op4, vint8m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_i8m2_m))) +void vsoxseg4ei16(vbool4_t op0, int8_t * op1, vuint16m4_t op2, vint8m2_t op3, vint8m2_t op4, vint8m2_t op5, vint8m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_i8mf2))) +void vsoxseg4ei16(int8_t * op0, vuint16m1_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_i8mf2_m))) +void vsoxseg4ei16(vbool16_t op0, int8_t * op1, vuint16m1_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_i8mf4))) +void vsoxseg4ei16(int8_t * op0, vuint16mf2_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_i8mf4_m))) +void vsoxseg4ei16(vbool32_t op0, int8_t * op1, vuint16mf2_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_i8mf8))) +void vsoxseg4ei16(int8_t * op0, vuint16mf4_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_i8mf8_m))) +void vsoxseg4ei16(vbool64_t op0, int8_t * op1, vuint16mf4_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_u8m1))) +void vsoxseg4ei16(uint8_t * op0, vuint16m2_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_u8m1_m))) +void vsoxseg4ei16(vbool8_t op0, uint8_t * op1, vuint16m2_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_u8m2))) +void vsoxseg4ei16(uint8_t * op0, vuint16m4_t op1, vuint8m2_t op2, vuint8m2_t op3, vuint8m2_t op4, vuint8m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_u8m2_m))) +void vsoxseg4ei16(vbool4_t op0, uint8_t * op1, vuint16m4_t op2, vuint8m2_t op3, vuint8m2_t op4, vuint8m2_t op5, vuint8m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_u8mf2))) +void vsoxseg4ei16(uint8_t * op0, vuint16m1_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_u8mf2_m))) +void vsoxseg4ei16(vbool16_t op0, uint8_t * op1, vuint16m1_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_u8mf4))) +void vsoxseg4ei16(uint8_t * op0, vuint16mf2_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_u8mf4_m))) +void vsoxseg4ei16(vbool32_t op0, uint8_t * op1, vuint16mf2_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_u8mf8))) +void vsoxseg4ei16(uint8_t * op0, vuint16mf4_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_u8mf8_m))) +void vsoxseg4ei16(vbool64_t op0, uint8_t * op1, vuint16mf4_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_i8m1))) +void vsoxseg5ei16(int8_t * op0, vuint16m2_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_i8m1_m))) +void vsoxseg5ei16(vbool8_t op0, int8_t * op1, vuint16m2_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_i8mf2))) +void vsoxseg5ei16(int8_t * op0, vuint16m1_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_i8mf2_m))) +void vsoxseg5ei16(vbool16_t op0, int8_t * op1, vuint16m1_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_i8mf4))) +void vsoxseg5ei16(int8_t * op0, vuint16mf2_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_i8mf4_m))) +void vsoxseg5ei16(vbool32_t op0, int8_t * op1, vuint16mf2_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_i8mf8))) +void vsoxseg5ei16(int8_t * op0, vuint16mf4_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_i8mf8_m))) +void vsoxseg5ei16(vbool64_t op0, int8_t * op1, vuint16mf4_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_u8m1))) +void vsoxseg5ei16(uint8_t * op0, vuint16m2_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_u8m1_m))) +void vsoxseg5ei16(vbool8_t op0, uint8_t * op1, vuint16m2_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_u8mf2))) +void vsoxseg5ei16(uint8_t * op0, vuint16m1_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_u8mf2_m))) +void vsoxseg5ei16(vbool16_t op0, uint8_t * op1, vuint16m1_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_u8mf4))) +void vsoxseg5ei16(uint8_t * op0, vuint16mf2_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_u8mf4_m))) +void vsoxseg5ei16(vbool32_t op0, uint8_t * op1, vuint16mf2_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_u8mf8))) +void vsoxseg5ei16(uint8_t * op0, vuint16mf4_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_u8mf8_m))) +void vsoxseg5ei16(vbool64_t op0, uint8_t * op1, vuint16mf4_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_i8m1))) +void vsoxseg6ei16(int8_t * op0, vuint16m2_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_i8m1_m))) +void vsoxseg6ei16(vbool8_t op0, int8_t * op1, vuint16m2_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_i8mf2))) +void vsoxseg6ei16(int8_t * op0, vuint16m1_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_i8mf2_m))) +void vsoxseg6ei16(vbool16_t op0, int8_t * op1, vuint16m1_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_i8mf4))) +void vsoxseg6ei16(int8_t * op0, vuint16mf2_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_i8mf4_m))) +void vsoxseg6ei16(vbool32_t op0, int8_t * op1, vuint16mf2_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_i8mf8))) +void vsoxseg6ei16(int8_t * op0, vuint16mf4_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_i8mf8_m))) +void vsoxseg6ei16(vbool64_t op0, int8_t * op1, vuint16mf4_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_u8m1))) +void vsoxseg6ei16(uint8_t * op0, vuint16m2_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_u8m1_m))) +void vsoxseg6ei16(vbool8_t op0, uint8_t * op1, vuint16m2_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_u8mf2))) +void vsoxseg6ei16(uint8_t * op0, vuint16m1_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_u8mf2_m))) +void vsoxseg6ei16(vbool16_t op0, uint8_t * op1, vuint16m1_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_u8mf4))) +void vsoxseg6ei16(uint8_t * op0, vuint16mf2_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_u8mf4_m))) +void vsoxseg6ei16(vbool32_t op0, uint8_t * op1, vuint16mf2_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_u8mf8))) +void vsoxseg6ei16(uint8_t * op0, vuint16mf4_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_u8mf8_m))) +void vsoxseg6ei16(vbool64_t op0, uint8_t * op1, vuint16mf4_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_i8m1))) +void vsoxseg7ei16(int8_t * op0, vuint16m2_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_i8m1_m))) +void vsoxseg7ei16(vbool8_t op0, int8_t * op1, vuint16m2_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_i8mf2))) +void vsoxseg7ei16(int8_t * op0, vuint16m1_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_i8mf2_m))) +void vsoxseg7ei16(vbool16_t op0, int8_t * op1, vuint16m1_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_i8mf4))) +void vsoxseg7ei16(int8_t * op0, vuint16mf2_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_i8mf4_m))) +void vsoxseg7ei16(vbool32_t op0, int8_t * op1, vuint16mf2_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_i8mf8))) +void vsoxseg7ei16(int8_t * op0, vuint16mf4_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_i8mf8_m))) +void vsoxseg7ei16(vbool64_t op0, int8_t * op1, vuint16mf4_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_u8m1))) +void vsoxseg7ei16(uint8_t * op0, vuint16m2_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_u8m1_m))) +void vsoxseg7ei16(vbool8_t op0, uint8_t * op1, vuint16m2_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_u8mf2))) +void vsoxseg7ei16(uint8_t * op0, vuint16m1_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_u8mf2_m))) +void vsoxseg7ei16(vbool16_t op0, uint8_t * op1, vuint16m1_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_u8mf4))) +void vsoxseg7ei16(uint8_t * op0, vuint16mf2_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_u8mf4_m))) +void vsoxseg7ei16(vbool32_t op0, uint8_t * op1, vuint16mf2_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_u8mf8))) +void vsoxseg7ei16(uint8_t * op0, vuint16mf4_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_u8mf8_m))) +void vsoxseg7ei16(vbool64_t op0, uint8_t * op1, vuint16mf4_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_i8m1))) +void vsoxseg8ei16(int8_t * op0, vuint16m2_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_i8m1_m))) +void vsoxseg8ei16(vbool8_t op0, int8_t * op1, vuint16m2_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, vint8m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_i8mf2))) +void vsoxseg8ei16(int8_t * op0, vuint16m1_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_i8mf2_m))) +void vsoxseg8ei16(vbool16_t op0, int8_t * op1, vuint16m1_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, vint8mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_i8mf4))) +void vsoxseg8ei16(int8_t * op0, vuint16mf2_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_i8mf4_m))) +void vsoxseg8ei16(vbool32_t op0, int8_t * op1, vuint16mf2_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, vint8mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_i8mf8))) +void vsoxseg8ei16(int8_t * op0, vuint16mf4_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_i8mf8_m))) +void vsoxseg8ei16(vbool64_t op0, int8_t * op1, vuint16mf4_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, vint8mf8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_u8m1))) +void vsoxseg8ei16(uint8_t * op0, vuint16m2_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_u8m1_m))) +void vsoxseg8ei16(vbool8_t op0, uint8_t * op1, vuint16m2_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, vuint8m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_u8mf2))) +void vsoxseg8ei16(uint8_t * op0, vuint16m1_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_u8mf2_m))) +void vsoxseg8ei16(vbool16_t op0, uint8_t * op1, vuint16m1_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, vuint8mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_u8mf4))) +void vsoxseg8ei16(uint8_t * op0, vuint16mf2_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_u8mf4_m))) +void vsoxseg8ei16(vbool32_t op0, uint8_t * op1, vuint16mf2_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, vuint8mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_u8mf8))) +void vsoxseg8ei16(uint8_t * op0, vuint16mf4_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_u8mf8_m))) +void vsoxseg8ei16(vbool64_t op0, uint8_t * op1, vuint16mf4_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, vuint8mf8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_i8m1))) +void vsoxseg2ei32(int8_t * op0, vuint32m4_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_i8m1_m))) +void vsoxseg2ei32(vbool8_t op0, int8_t * op1, vuint32m4_t op2, vint8m1_t op3, vint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_i8m2))) +void vsoxseg2ei32(int8_t * op0, vuint32m8_t op1, vint8m2_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_i8m2_m))) +void vsoxseg2ei32(vbool4_t op0, int8_t * op1, vuint32m8_t op2, vint8m2_t op3, vint8m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_i8mf2))) +void vsoxseg2ei32(int8_t * op0, vuint32m2_t op1, vint8mf2_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_i8mf2_m))) +void vsoxseg2ei32(vbool16_t op0, int8_t * op1, vuint32m2_t op2, vint8mf2_t op3, vint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_i8mf4))) +void vsoxseg2ei32(int8_t * op0, vuint32m1_t op1, vint8mf4_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_i8mf4_m))) +void vsoxseg2ei32(vbool32_t op0, int8_t * op1, vuint32m1_t op2, vint8mf4_t op3, vint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_i8mf8))) +void vsoxseg2ei32(int8_t * op0, vuint32mf2_t op1, vint8mf8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_i8mf8_m))) +void vsoxseg2ei32(vbool64_t op0, int8_t * op1, vuint32mf2_t op2, vint8mf8_t op3, vint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_u8m1))) +void vsoxseg2ei32(uint8_t * op0, vuint32m4_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_u8m1_m))) +void vsoxseg2ei32(vbool8_t op0, uint8_t * op1, vuint32m4_t op2, vuint8m1_t op3, vuint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_u8m2))) +void vsoxseg2ei32(uint8_t * op0, vuint32m8_t op1, vuint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_u8m2_m))) +void vsoxseg2ei32(vbool4_t op0, uint8_t * op1, vuint32m8_t op2, vuint8m2_t op3, vuint8m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_u8mf2))) +void vsoxseg2ei32(uint8_t * op0, vuint32m2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_u8mf2_m))) +void vsoxseg2ei32(vbool16_t op0, uint8_t * op1, vuint32m2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_u8mf4))) +void vsoxseg2ei32(uint8_t * op0, vuint32m1_t op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_u8mf4_m))) +void vsoxseg2ei32(vbool32_t op0, uint8_t * op1, vuint32m1_t op2, vuint8mf4_t op3, vuint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_u8mf8))) +void vsoxseg2ei32(uint8_t * op0, vuint32mf2_t op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_u8mf8_m))) +void vsoxseg2ei32(vbool64_t op0, uint8_t * op1, vuint32mf2_t op2, vuint8mf8_t op3, vuint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_i8m1))) +void vsoxseg3ei32(int8_t * op0, vuint32m4_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_i8m1_m))) +void vsoxseg3ei32(vbool8_t op0, int8_t * op1, vuint32m4_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_i8m2))) +void vsoxseg3ei32(int8_t * op0, vuint32m8_t op1, vint8m2_t op2, vint8m2_t op3, vint8m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_i8m2_m))) +void vsoxseg3ei32(vbool4_t op0, int8_t * op1, vuint32m8_t op2, vint8m2_t op3, vint8m2_t op4, vint8m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_i8mf2))) +void vsoxseg3ei32(int8_t * op0, vuint32m2_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_i8mf2_m))) +void vsoxseg3ei32(vbool16_t op0, int8_t * op1, vuint32m2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_i8mf4))) +void vsoxseg3ei32(int8_t * op0, vuint32m1_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_i8mf4_m))) +void vsoxseg3ei32(vbool32_t op0, int8_t * op1, vuint32m1_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_i8mf8))) +void vsoxseg3ei32(int8_t * op0, vuint32mf2_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_i8mf8_m))) +void vsoxseg3ei32(vbool64_t op0, int8_t * op1, vuint32mf2_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_u8m1))) +void vsoxseg3ei32(uint8_t * op0, vuint32m4_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_u8m1_m))) +void vsoxseg3ei32(vbool8_t op0, uint8_t * op1, vuint32m4_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_u8m2))) +void vsoxseg3ei32(uint8_t * op0, vuint32m8_t op1, vuint8m2_t op2, vuint8m2_t op3, vuint8m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_u8m2_m))) +void vsoxseg3ei32(vbool4_t op0, uint8_t * op1, vuint32m8_t op2, vuint8m2_t op3, vuint8m2_t op4, vuint8m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_u8mf2))) +void vsoxseg3ei32(uint8_t * op0, vuint32m2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_u8mf2_m))) +void vsoxseg3ei32(vbool16_t op0, uint8_t * op1, vuint32m2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_u8mf4))) +void vsoxseg3ei32(uint8_t * op0, vuint32m1_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_u8mf4_m))) +void vsoxseg3ei32(vbool32_t op0, uint8_t * op1, vuint32m1_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_u8mf8))) +void vsoxseg3ei32(uint8_t * op0, vuint32mf2_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_u8mf8_m))) +void vsoxseg3ei32(vbool64_t op0, uint8_t * op1, vuint32mf2_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_i8m1))) +void vsoxseg4ei32(int8_t * op0, vuint32m4_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_i8m1_m))) +void vsoxseg4ei32(vbool8_t op0, int8_t * op1, vuint32m4_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_i8m2))) +void vsoxseg4ei32(int8_t * op0, vuint32m8_t op1, vint8m2_t op2, vint8m2_t op3, vint8m2_t op4, vint8m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_i8m2_m))) +void vsoxseg4ei32(vbool4_t op0, int8_t * op1, vuint32m8_t op2, vint8m2_t op3, vint8m2_t op4, vint8m2_t op5, vint8m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_i8mf2))) +void vsoxseg4ei32(int8_t * op0, vuint32m2_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_i8mf2_m))) +void vsoxseg4ei32(vbool16_t op0, int8_t * op1, vuint32m2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_i8mf4))) +void vsoxseg4ei32(int8_t * op0, vuint32m1_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_i8mf4_m))) +void vsoxseg4ei32(vbool32_t op0, int8_t * op1, vuint32m1_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_i8mf8))) +void vsoxseg4ei32(int8_t * op0, vuint32mf2_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_i8mf8_m))) +void vsoxseg4ei32(vbool64_t op0, int8_t * op1, vuint32mf2_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_u8m1))) +void vsoxseg4ei32(uint8_t * op0, vuint32m4_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_u8m1_m))) +void vsoxseg4ei32(vbool8_t op0, uint8_t * op1, vuint32m4_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_u8m2))) +void vsoxseg4ei32(uint8_t * op0, vuint32m8_t op1, vuint8m2_t op2, vuint8m2_t op3, vuint8m2_t op4, vuint8m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_u8m2_m))) +void vsoxseg4ei32(vbool4_t op0, uint8_t * op1, vuint32m8_t op2, vuint8m2_t op3, vuint8m2_t op4, vuint8m2_t op5, vuint8m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_u8mf2))) +void vsoxseg4ei32(uint8_t * op0, vuint32m2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_u8mf2_m))) +void vsoxseg4ei32(vbool16_t op0, uint8_t * op1, vuint32m2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_u8mf4))) +void vsoxseg4ei32(uint8_t * op0, vuint32m1_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_u8mf4_m))) +void vsoxseg4ei32(vbool32_t op0, uint8_t * op1, vuint32m1_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_u8mf8))) +void vsoxseg4ei32(uint8_t * op0, vuint32mf2_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_u8mf8_m))) +void vsoxseg4ei32(vbool64_t op0, uint8_t * op1, vuint32mf2_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_i8m1))) +void vsoxseg5ei32(int8_t * op0, vuint32m4_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_i8m1_m))) +void vsoxseg5ei32(vbool8_t op0, int8_t * op1, vuint32m4_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_i8mf2))) +void vsoxseg5ei32(int8_t * op0, vuint32m2_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_i8mf2_m))) +void vsoxseg5ei32(vbool16_t op0, int8_t * op1, vuint32m2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_i8mf4))) +void vsoxseg5ei32(int8_t * op0, vuint32m1_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_i8mf4_m))) +void vsoxseg5ei32(vbool32_t op0, int8_t * op1, vuint32m1_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_i8mf8))) +void vsoxseg5ei32(int8_t * op0, vuint32mf2_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_i8mf8_m))) +void vsoxseg5ei32(vbool64_t op0, int8_t * op1, vuint32mf2_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_u8m1))) +void vsoxseg5ei32(uint8_t * op0, vuint32m4_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_u8m1_m))) +void vsoxseg5ei32(vbool8_t op0, uint8_t * op1, vuint32m4_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_u8mf2))) +void vsoxseg5ei32(uint8_t * op0, vuint32m2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_u8mf2_m))) +void vsoxseg5ei32(vbool16_t op0, uint8_t * op1, vuint32m2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_u8mf4))) +void vsoxseg5ei32(uint8_t * op0, vuint32m1_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_u8mf4_m))) +void vsoxseg5ei32(vbool32_t op0, uint8_t * op1, vuint32m1_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_u8mf8))) +void vsoxseg5ei32(uint8_t * op0, vuint32mf2_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_u8mf8_m))) +void vsoxseg5ei32(vbool64_t op0, uint8_t * op1, vuint32mf2_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_i8m1))) +void vsoxseg6ei32(int8_t * op0, vuint32m4_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_i8m1_m))) +void vsoxseg6ei32(vbool8_t op0, int8_t * op1, vuint32m4_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_i8mf2))) +void vsoxseg6ei32(int8_t * op0, vuint32m2_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_i8mf2_m))) +void vsoxseg6ei32(vbool16_t op0, int8_t * op1, vuint32m2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_i8mf4))) +void vsoxseg6ei32(int8_t * op0, vuint32m1_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_i8mf4_m))) +void vsoxseg6ei32(vbool32_t op0, int8_t * op1, vuint32m1_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_i8mf8))) +void vsoxseg6ei32(int8_t * op0, vuint32mf2_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_i8mf8_m))) +void vsoxseg6ei32(vbool64_t op0, int8_t * op1, vuint32mf2_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_u8m1))) +void vsoxseg6ei32(uint8_t * op0, vuint32m4_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_u8m1_m))) +void vsoxseg6ei32(vbool8_t op0, uint8_t * op1, vuint32m4_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_u8mf2))) +void vsoxseg6ei32(uint8_t * op0, vuint32m2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_u8mf2_m))) +void vsoxseg6ei32(vbool16_t op0, uint8_t * op1, vuint32m2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_u8mf4))) +void vsoxseg6ei32(uint8_t * op0, vuint32m1_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_u8mf4_m))) +void vsoxseg6ei32(vbool32_t op0, uint8_t * op1, vuint32m1_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_u8mf8))) +void vsoxseg6ei32(uint8_t * op0, vuint32mf2_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_u8mf8_m))) +void vsoxseg6ei32(vbool64_t op0, uint8_t * op1, vuint32mf2_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_i8m1))) +void vsoxseg7ei32(int8_t * op0, vuint32m4_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_i8m1_m))) +void vsoxseg7ei32(vbool8_t op0, int8_t * op1, vuint32m4_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_i8mf2))) +void vsoxseg7ei32(int8_t * op0, vuint32m2_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_i8mf2_m))) +void vsoxseg7ei32(vbool16_t op0, int8_t * op1, vuint32m2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_i8mf4))) +void vsoxseg7ei32(int8_t * op0, vuint32m1_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_i8mf4_m))) +void vsoxseg7ei32(vbool32_t op0, int8_t * op1, vuint32m1_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_i8mf8))) +void vsoxseg7ei32(int8_t * op0, vuint32mf2_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_i8mf8_m))) +void vsoxseg7ei32(vbool64_t op0, int8_t * op1, vuint32mf2_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_u8m1))) +void vsoxseg7ei32(uint8_t * op0, vuint32m4_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_u8m1_m))) +void vsoxseg7ei32(vbool8_t op0, uint8_t * op1, vuint32m4_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_u8mf2))) +void vsoxseg7ei32(uint8_t * op0, vuint32m2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_u8mf2_m))) +void vsoxseg7ei32(vbool16_t op0, uint8_t * op1, vuint32m2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_u8mf4))) +void vsoxseg7ei32(uint8_t * op0, vuint32m1_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_u8mf4_m))) +void vsoxseg7ei32(vbool32_t op0, uint8_t * op1, vuint32m1_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_u8mf8))) +void vsoxseg7ei32(uint8_t * op0, vuint32mf2_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_u8mf8_m))) +void vsoxseg7ei32(vbool64_t op0, uint8_t * op1, vuint32mf2_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_i8m1))) +void vsoxseg8ei32(int8_t * op0, vuint32m4_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_i8m1_m))) +void vsoxseg8ei32(vbool8_t op0, int8_t * op1, vuint32m4_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, vint8m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_i8mf2))) +void vsoxseg8ei32(int8_t * op0, vuint32m2_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_i8mf2_m))) +void vsoxseg8ei32(vbool16_t op0, int8_t * op1, vuint32m2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, vint8mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_i8mf4))) +void vsoxseg8ei32(int8_t * op0, vuint32m1_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_i8mf4_m))) +void vsoxseg8ei32(vbool32_t op0, int8_t * op1, vuint32m1_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, vint8mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_i8mf8))) +void vsoxseg8ei32(int8_t * op0, vuint32mf2_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_i8mf8_m))) +void vsoxseg8ei32(vbool64_t op0, int8_t * op1, vuint32mf2_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, vint8mf8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_u8m1))) +void vsoxseg8ei32(uint8_t * op0, vuint32m4_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_u8m1_m))) +void vsoxseg8ei32(vbool8_t op0, uint8_t * op1, vuint32m4_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, vuint8m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_u8mf2))) +void vsoxseg8ei32(uint8_t * op0, vuint32m2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_u8mf2_m))) +void vsoxseg8ei32(vbool16_t op0, uint8_t * op1, vuint32m2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, vuint8mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_u8mf4))) +void vsoxseg8ei32(uint8_t * op0, vuint32m1_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_u8mf4_m))) +void vsoxseg8ei32(vbool32_t op0, uint8_t * op1, vuint32m1_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, vuint8mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_u8mf8))) +void vsoxseg8ei32(uint8_t * op0, vuint32mf2_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_u8mf8_m))) +void vsoxseg8ei32(vbool64_t op0, uint8_t * op1, vuint32mf2_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, vuint8mf8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_i8m1))) +void vsoxseg2ei64(int8_t * op0, vuint64m8_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_i8m1_m))) +void vsoxseg2ei64(vbool8_t op0, int8_t * op1, vuint64m8_t op2, vint8m1_t op3, vint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_i8mf2))) +void vsoxseg2ei64(int8_t * op0, vuint64m4_t op1, vint8mf2_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_i8mf2_m))) +void vsoxseg2ei64(vbool16_t op0, int8_t * op1, vuint64m4_t op2, vint8mf2_t op3, vint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_i8mf4))) +void vsoxseg2ei64(int8_t * op0, vuint64m2_t op1, vint8mf4_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_i8mf4_m))) +void vsoxseg2ei64(vbool32_t op0, int8_t * op1, vuint64m2_t op2, vint8mf4_t op3, vint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_i8mf8))) +void vsoxseg2ei64(int8_t * op0, vuint64m1_t op1, vint8mf8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_i8mf8_m))) +void vsoxseg2ei64(vbool64_t op0, int8_t * op1, vuint64m1_t op2, vint8mf8_t op3, vint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_u8m1))) +void vsoxseg2ei64(uint8_t * op0, vuint64m8_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_u8m1_m))) +void vsoxseg2ei64(vbool8_t op0, uint8_t * op1, vuint64m8_t op2, vuint8m1_t op3, vuint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_u8mf2))) +void vsoxseg2ei64(uint8_t * op0, vuint64m4_t op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_u8mf2_m))) +void vsoxseg2ei64(vbool16_t op0, uint8_t * op1, vuint64m4_t op2, vuint8mf2_t op3, vuint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_u8mf4))) +void vsoxseg2ei64(uint8_t * op0, vuint64m2_t op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_u8mf4_m))) +void vsoxseg2ei64(vbool32_t op0, uint8_t * op1, vuint64m2_t op2, vuint8mf4_t op3, vuint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_u8mf8))) +void vsoxseg2ei64(uint8_t * op0, vuint64m1_t op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_u8mf8_m))) +void vsoxseg2ei64(vbool64_t op0, uint8_t * op1, vuint64m1_t op2, vuint8mf8_t op3, vuint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_i8m1))) +void vsoxseg3ei64(int8_t * op0, vuint64m8_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_i8m1_m))) +void vsoxseg3ei64(vbool8_t op0, int8_t * op1, vuint64m8_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_i8mf2))) +void vsoxseg3ei64(int8_t * op0, vuint64m4_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_i8mf2_m))) +void vsoxseg3ei64(vbool16_t op0, int8_t * op1, vuint64m4_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_i8mf4))) +void vsoxseg3ei64(int8_t * op0, vuint64m2_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_i8mf4_m))) +void vsoxseg3ei64(vbool32_t op0, int8_t * op1, vuint64m2_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_i8mf8))) +void vsoxseg3ei64(int8_t * op0, vuint64m1_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_i8mf8_m))) +void vsoxseg3ei64(vbool64_t op0, int8_t * op1, vuint64m1_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_u8m1))) +void vsoxseg3ei64(uint8_t * op0, vuint64m8_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_u8m1_m))) +void vsoxseg3ei64(vbool8_t op0, uint8_t * op1, vuint64m8_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_u8mf2))) +void vsoxseg3ei64(uint8_t * op0, vuint64m4_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_u8mf2_m))) +void vsoxseg3ei64(vbool16_t op0, uint8_t * op1, vuint64m4_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_u8mf4))) +void vsoxseg3ei64(uint8_t * op0, vuint64m2_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_u8mf4_m))) +void vsoxseg3ei64(vbool32_t op0, uint8_t * op1, vuint64m2_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_u8mf8))) +void vsoxseg3ei64(uint8_t * op0, vuint64m1_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_u8mf8_m))) +void vsoxseg3ei64(vbool64_t op0, uint8_t * op1, vuint64m1_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_i8m1))) +void vsoxseg4ei64(int8_t * op0, vuint64m8_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_i8m1_m))) +void vsoxseg4ei64(vbool8_t op0, int8_t * op1, vuint64m8_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_i8mf2))) +void vsoxseg4ei64(int8_t * op0, vuint64m4_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_i8mf2_m))) +void vsoxseg4ei64(vbool16_t op0, int8_t * op1, vuint64m4_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_i8mf4))) +void vsoxseg4ei64(int8_t * op0, vuint64m2_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_i8mf4_m))) +void vsoxseg4ei64(vbool32_t op0, int8_t * op1, vuint64m2_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_i8mf8))) +void vsoxseg4ei64(int8_t * op0, vuint64m1_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_i8mf8_m))) +void vsoxseg4ei64(vbool64_t op0, int8_t * op1, vuint64m1_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_u8m1))) +void vsoxseg4ei64(uint8_t * op0, vuint64m8_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_u8m1_m))) +void vsoxseg4ei64(vbool8_t op0, uint8_t * op1, vuint64m8_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_u8mf2))) +void vsoxseg4ei64(uint8_t * op0, vuint64m4_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_u8mf2_m))) +void vsoxseg4ei64(vbool16_t op0, uint8_t * op1, vuint64m4_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_u8mf4))) +void vsoxseg4ei64(uint8_t * op0, vuint64m2_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_u8mf4_m))) +void vsoxseg4ei64(vbool32_t op0, uint8_t * op1, vuint64m2_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_u8mf8))) +void vsoxseg4ei64(uint8_t * op0, vuint64m1_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_u8mf8_m))) +void vsoxseg4ei64(vbool64_t op0, uint8_t * op1, vuint64m1_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_i8m1))) +void vsoxseg5ei64(int8_t * op0, vuint64m8_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_i8m1_m))) +void vsoxseg5ei64(vbool8_t op0, int8_t * op1, vuint64m8_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_i8mf2))) +void vsoxseg5ei64(int8_t * op0, vuint64m4_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_i8mf2_m))) +void vsoxseg5ei64(vbool16_t op0, int8_t * op1, vuint64m4_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_i8mf4))) +void vsoxseg5ei64(int8_t * op0, vuint64m2_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_i8mf4_m))) +void vsoxseg5ei64(vbool32_t op0, int8_t * op1, vuint64m2_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_i8mf8))) +void vsoxseg5ei64(int8_t * op0, vuint64m1_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_i8mf8_m))) +void vsoxseg5ei64(vbool64_t op0, int8_t * op1, vuint64m1_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_u8m1))) +void vsoxseg5ei64(uint8_t * op0, vuint64m8_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_u8m1_m))) +void vsoxseg5ei64(vbool8_t op0, uint8_t * op1, vuint64m8_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_u8mf2))) +void vsoxseg5ei64(uint8_t * op0, vuint64m4_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_u8mf2_m))) +void vsoxseg5ei64(vbool16_t op0, uint8_t * op1, vuint64m4_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_u8mf4))) +void vsoxseg5ei64(uint8_t * op0, vuint64m2_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_u8mf4_m))) +void vsoxseg5ei64(vbool32_t op0, uint8_t * op1, vuint64m2_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_u8mf8))) +void vsoxseg5ei64(uint8_t * op0, vuint64m1_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_u8mf8_m))) +void vsoxseg5ei64(vbool64_t op0, uint8_t * op1, vuint64m1_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_i8m1))) +void vsoxseg6ei64(int8_t * op0, vuint64m8_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_i8m1_m))) +void vsoxseg6ei64(vbool8_t op0, int8_t * op1, vuint64m8_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_i8mf2))) +void vsoxseg6ei64(int8_t * op0, vuint64m4_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_i8mf2_m))) +void vsoxseg6ei64(vbool16_t op0, int8_t * op1, vuint64m4_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_i8mf4))) +void vsoxseg6ei64(int8_t * op0, vuint64m2_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_i8mf4_m))) +void vsoxseg6ei64(vbool32_t op0, int8_t * op1, vuint64m2_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_i8mf8))) +void vsoxseg6ei64(int8_t * op0, vuint64m1_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_i8mf8_m))) +void vsoxseg6ei64(vbool64_t op0, int8_t * op1, vuint64m1_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_u8m1))) +void vsoxseg6ei64(uint8_t * op0, vuint64m8_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_u8m1_m))) +void vsoxseg6ei64(vbool8_t op0, uint8_t * op1, vuint64m8_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_u8mf2))) +void vsoxseg6ei64(uint8_t * op0, vuint64m4_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_u8mf2_m))) +void vsoxseg6ei64(vbool16_t op0, uint8_t * op1, vuint64m4_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_u8mf4))) +void vsoxseg6ei64(uint8_t * op0, vuint64m2_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_u8mf4_m))) +void vsoxseg6ei64(vbool32_t op0, uint8_t * op1, vuint64m2_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_u8mf8))) +void vsoxseg6ei64(uint8_t * op0, vuint64m1_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_u8mf8_m))) +void vsoxseg6ei64(vbool64_t op0, uint8_t * op1, vuint64m1_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_i8m1))) +void vsoxseg7ei64(int8_t * op0, vuint64m8_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_i8m1_m))) +void vsoxseg7ei64(vbool8_t op0, int8_t * op1, vuint64m8_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_i8mf2))) +void vsoxseg7ei64(int8_t * op0, vuint64m4_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_i8mf2_m))) +void vsoxseg7ei64(vbool16_t op0, int8_t * op1, vuint64m4_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_i8mf4))) +void vsoxseg7ei64(int8_t * op0, vuint64m2_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_i8mf4_m))) +void vsoxseg7ei64(vbool32_t op0, int8_t * op1, vuint64m2_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_i8mf8))) +void vsoxseg7ei64(int8_t * op0, vuint64m1_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_i8mf8_m))) +void vsoxseg7ei64(vbool64_t op0, int8_t * op1, vuint64m1_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_u8m1))) +void vsoxseg7ei64(uint8_t * op0, vuint64m8_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_u8m1_m))) +void vsoxseg7ei64(vbool8_t op0, uint8_t * op1, vuint64m8_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_u8mf2))) +void vsoxseg7ei64(uint8_t * op0, vuint64m4_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_u8mf2_m))) +void vsoxseg7ei64(vbool16_t op0, uint8_t * op1, vuint64m4_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_u8mf4))) +void vsoxseg7ei64(uint8_t * op0, vuint64m2_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_u8mf4_m))) +void vsoxseg7ei64(vbool32_t op0, uint8_t * op1, vuint64m2_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_u8mf8))) +void vsoxseg7ei64(uint8_t * op0, vuint64m1_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_u8mf8_m))) +void vsoxseg7ei64(vbool64_t op0, uint8_t * op1, vuint64m1_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_i8m1))) +void vsoxseg8ei64(int8_t * op0, vuint64m8_t op1, vint8m1_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_i8m1_m))) +void vsoxseg8ei64(vbool8_t op0, int8_t * op1, vuint64m8_t op2, vint8m1_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, vint8m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_i8mf2))) +void vsoxseg8ei64(int8_t * op0, vuint64m4_t op1, vint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_i8mf2_m))) +void vsoxseg8ei64(vbool16_t op0, int8_t * op1, vuint64m4_t op2, vint8mf2_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, vint8mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_i8mf4))) +void vsoxseg8ei64(int8_t * op0, vuint64m2_t op1, vint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_i8mf4_m))) +void vsoxseg8ei64(vbool32_t op0, int8_t * op1, vuint64m2_t op2, vint8mf4_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, vint8mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_i8mf8))) +void vsoxseg8ei64(int8_t * op0, vuint64m1_t op1, vint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_i8mf8_m))) +void vsoxseg8ei64(vbool64_t op0, int8_t * op1, vuint64m1_t op2, vint8mf8_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, vint8mf8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_u8m1))) +void vsoxseg8ei64(uint8_t * op0, vuint64m8_t op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_u8m1_m))) +void vsoxseg8ei64(vbool8_t op0, uint8_t * op1, vuint64m8_t op2, vuint8m1_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, vuint8m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_u8mf2))) +void vsoxseg8ei64(uint8_t * op0, vuint64m4_t op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_u8mf2_m))) +void vsoxseg8ei64(vbool16_t op0, uint8_t * op1, vuint64m4_t op2, vuint8mf2_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, vuint8mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_u8mf4))) +void vsoxseg8ei64(uint8_t * op0, vuint64m2_t op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_u8mf4_m))) +void vsoxseg8ei64(vbool32_t op0, uint8_t * op1, vuint64m2_t op2, vuint8mf4_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, vuint8mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_u8mf8))) +void vsoxseg8ei64(uint8_t * op0, vuint64m1_t op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_u8mf8_m))) +void vsoxseg8ei64(vbool64_t op0, uint8_t * op1, vuint64m1_t op2, vuint8mf8_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, vuint8mf8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_i16m1))) +void vsoxseg2ei8(int16_t * op0, vuint8mf2_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_i16m1_m))) +void vsoxseg2ei8(vbool16_t op0, int16_t * op1, vuint8mf2_t op2, vint16m1_t op3, vint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_i16m2))) +void vsoxseg2ei8(int16_t * op0, vuint8m1_t op1, vint16m2_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_i16m2_m))) +void vsoxseg2ei8(vbool8_t op0, int16_t * op1, vuint8m1_t op2, vint16m2_t op3, vint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_i16m4))) +void vsoxseg2ei8(int16_t * op0, vuint8m2_t op1, vint16m4_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_i16m4_m))) +void vsoxseg2ei8(vbool4_t op0, int16_t * op1, vuint8m2_t op2, vint16m4_t op3, vint16m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_i16mf2))) +void vsoxseg2ei8(int16_t * op0, vuint8mf4_t op1, vint16mf2_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_i16mf2_m))) +void vsoxseg2ei8(vbool32_t op0, int16_t * op1, vuint8mf4_t op2, vint16mf2_t op3, vint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_i16mf4))) +void vsoxseg2ei8(int16_t * op0, vuint8mf8_t op1, vint16mf4_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_i16mf4_m))) +void vsoxseg2ei8(vbool64_t op0, int16_t * op1, vuint8mf8_t op2, vint16mf4_t op3, vint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_u16m1))) +void vsoxseg2ei8(uint16_t * op0, vuint8mf2_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_u16m1_m))) +void vsoxseg2ei8(vbool16_t op0, uint16_t * op1, vuint8mf2_t op2, vuint16m1_t op3, vuint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_u16m2))) +void vsoxseg2ei8(uint16_t * op0, vuint8m1_t op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_u16m2_m))) +void vsoxseg2ei8(vbool8_t op0, uint16_t * op1, vuint8m1_t op2, vuint16m2_t op3, vuint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_u16m4))) +void vsoxseg2ei8(uint16_t * op0, vuint8m2_t op1, vuint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_u16m4_m))) +void vsoxseg2ei8(vbool4_t op0, uint16_t * op1, vuint8m2_t op2, vuint16m4_t op3, vuint16m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_u16mf2))) +void vsoxseg2ei8(uint16_t * op0, vuint8mf4_t op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_u16mf2_m))) +void vsoxseg2ei8(vbool32_t op0, uint16_t * op1, vuint8mf4_t op2, vuint16mf2_t op3, vuint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_u16mf4))) +void vsoxseg2ei8(uint16_t * op0, vuint8mf8_t op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_u16mf4_m))) +void vsoxseg2ei8(vbool64_t op0, uint16_t * op1, vuint8mf8_t op2, vuint16mf4_t op3, vuint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_i16m1))) +void vsoxseg3ei8(int16_t * op0, vuint8mf2_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_i16m1_m))) +void vsoxseg3ei8(vbool16_t op0, int16_t * op1, vuint8mf2_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_i16m2))) +void vsoxseg3ei8(int16_t * op0, vuint8m1_t op1, vint16m2_t op2, vint16m2_t op3, vint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_i16m2_m))) +void vsoxseg3ei8(vbool8_t op0, int16_t * op1, vuint8m1_t op2, vint16m2_t op3, vint16m2_t op4, vint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_i16mf2))) +void vsoxseg3ei8(int16_t * op0, vuint8mf4_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_i16mf2_m))) +void vsoxseg3ei8(vbool32_t op0, int16_t * op1, vuint8mf4_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_i16mf4))) +void vsoxseg3ei8(int16_t * op0, vuint8mf8_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_i16mf4_m))) +void vsoxseg3ei8(vbool64_t op0, int16_t * op1, vuint8mf8_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_u16m1))) +void vsoxseg3ei8(uint16_t * op0, vuint8mf2_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_u16m1_m))) +void vsoxseg3ei8(vbool16_t op0, uint16_t * op1, vuint8mf2_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_u16m2))) +void vsoxseg3ei8(uint16_t * op0, vuint8m1_t op1, vuint16m2_t op2, vuint16m2_t op3, vuint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_u16m2_m))) +void vsoxseg3ei8(vbool8_t op0, uint16_t * op1, vuint8m1_t op2, vuint16m2_t op3, vuint16m2_t op4, vuint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_u16mf2))) +void vsoxseg3ei8(uint16_t * op0, vuint8mf4_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_u16mf2_m))) +void vsoxseg3ei8(vbool32_t op0, uint16_t * op1, vuint8mf4_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_u16mf4))) +void vsoxseg3ei8(uint16_t * op0, vuint8mf8_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_u16mf4_m))) +void vsoxseg3ei8(vbool64_t op0, uint16_t * op1, vuint8mf8_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_i16m1))) +void vsoxseg4ei8(int16_t * op0, vuint8mf2_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_i16m1_m))) +void vsoxseg4ei8(vbool16_t op0, int16_t * op1, vuint8mf2_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_i16m2))) +void vsoxseg4ei8(int16_t * op0, vuint8m1_t op1, vint16m2_t op2, vint16m2_t op3, vint16m2_t op4, vint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_i16m2_m))) +void vsoxseg4ei8(vbool8_t op0, int16_t * op1, vuint8m1_t op2, vint16m2_t op3, vint16m2_t op4, vint16m2_t op5, vint16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_i16mf2))) +void vsoxseg4ei8(int16_t * op0, vuint8mf4_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_i16mf2_m))) +void vsoxseg4ei8(vbool32_t op0, int16_t * op1, vuint8mf4_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_i16mf4))) +void vsoxseg4ei8(int16_t * op0, vuint8mf8_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_i16mf4_m))) +void vsoxseg4ei8(vbool64_t op0, int16_t * op1, vuint8mf8_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_u16m1))) +void vsoxseg4ei8(uint16_t * op0, vuint8mf2_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_u16m1_m))) +void vsoxseg4ei8(vbool16_t op0, uint16_t * op1, vuint8mf2_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_u16m2))) +void vsoxseg4ei8(uint16_t * op0, vuint8m1_t op1, vuint16m2_t op2, vuint16m2_t op3, vuint16m2_t op4, vuint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_u16m2_m))) +void vsoxseg4ei8(vbool8_t op0, uint16_t * op1, vuint8m1_t op2, vuint16m2_t op3, vuint16m2_t op4, vuint16m2_t op5, vuint16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_u16mf2))) +void vsoxseg4ei8(uint16_t * op0, vuint8mf4_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_u16mf2_m))) +void vsoxseg4ei8(vbool32_t op0, uint16_t * op1, vuint8mf4_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_u16mf4))) +void vsoxseg4ei8(uint16_t * op0, vuint8mf8_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_u16mf4_m))) +void vsoxseg4ei8(vbool64_t op0, uint16_t * op1, vuint8mf8_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_i16m1))) +void vsoxseg5ei8(int16_t * op0, vuint8mf2_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_i16m1_m))) +void vsoxseg5ei8(vbool16_t op0, int16_t * op1, vuint8mf2_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_i16mf2))) +void vsoxseg5ei8(int16_t * op0, vuint8mf4_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_i16mf2_m))) +void vsoxseg5ei8(vbool32_t op0, int16_t * op1, vuint8mf4_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_i16mf4))) +void vsoxseg5ei8(int16_t * op0, vuint8mf8_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_i16mf4_m))) +void vsoxseg5ei8(vbool64_t op0, int16_t * op1, vuint8mf8_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_u16m1))) +void vsoxseg5ei8(uint16_t * op0, vuint8mf2_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_u16m1_m))) +void vsoxseg5ei8(vbool16_t op0, uint16_t * op1, vuint8mf2_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_u16mf2))) +void vsoxseg5ei8(uint16_t * op0, vuint8mf4_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_u16mf2_m))) +void vsoxseg5ei8(vbool32_t op0, uint16_t * op1, vuint8mf4_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_u16mf4))) +void vsoxseg5ei8(uint16_t * op0, vuint8mf8_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_u16mf4_m))) +void vsoxseg5ei8(vbool64_t op0, uint16_t * op1, vuint8mf8_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_i16m1))) +void vsoxseg6ei8(int16_t * op0, vuint8mf2_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_i16m1_m))) +void vsoxseg6ei8(vbool16_t op0, int16_t * op1, vuint8mf2_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_i16mf2))) +void vsoxseg6ei8(int16_t * op0, vuint8mf4_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_i16mf2_m))) +void vsoxseg6ei8(vbool32_t op0, int16_t * op1, vuint8mf4_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_i16mf4))) +void vsoxseg6ei8(int16_t * op0, vuint8mf8_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_i16mf4_m))) +void vsoxseg6ei8(vbool64_t op0, int16_t * op1, vuint8mf8_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_u16m1))) +void vsoxseg6ei8(uint16_t * op0, vuint8mf2_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_u16m1_m))) +void vsoxseg6ei8(vbool16_t op0, uint16_t * op1, vuint8mf2_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_u16mf2))) +void vsoxseg6ei8(uint16_t * op0, vuint8mf4_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_u16mf2_m))) +void vsoxseg6ei8(vbool32_t op0, uint16_t * op1, vuint8mf4_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_u16mf4))) +void vsoxseg6ei8(uint16_t * op0, vuint8mf8_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_u16mf4_m))) +void vsoxseg6ei8(vbool64_t op0, uint16_t * op1, vuint8mf8_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_i16m1))) +void vsoxseg7ei8(int16_t * op0, vuint8mf2_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_i16m1_m))) +void vsoxseg7ei8(vbool16_t op0, int16_t * op1, vuint8mf2_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_i16mf2))) +void vsoxseg7ei8(int16_t * op0, vuint8mf4_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_i16mf2_m))) +void vsoxseg7ei8(vbool32_t op0, int16_t * op1, vuint8mf4_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_i16mf4))) +void vsoxseg7ei8(int16_t * op0, vuint8mf8_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_i16mf4_m))) +void vsoxseg7ei8(vbool64_t op0, int16_t * op1, vuint8mf8_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_u16m1))) +void vsoxseg7ei8(uint16_t * op0, vuint8mf2_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_u16m1_m))) +void vsoxseg7ei8(vbool16_t op0, uint16_t * op1, vuint8mf2_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_u16mf2))) +void vsoxseg7ei8(uint16_t * op0, vuint8mf4_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_u16mf2_m))) +void vsoxseg7ei8(vbool32_t op0, uint16_t * op1, vuint8mf4_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_u16mf4))) +void vsoxseg7ei8(uint16_t * op0, vuint8mf8_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_u16mf4_m))) +void vsoxseg7ei8(vbool64_t op0, uint16_t * op1, vuint8mf8_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_i16m1))) +void vsoxseg8ei8(int16_t * op0, vuint8mf2_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_i16m1_m))) +void vsoxseg8ei8(vbool16_t op0, int16_t * op1, vuint8mf2_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, vint16m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_i16mf2))) +void vsoxseg8ei8(int16_t * op0, vuint8mf4_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_i16mf2_m))) +void vsoxseg8ei8(vbool32_t op0, int16_t * op1, vuint8mf4_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, vint16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_i16mf4))) +void vsoxseg8ei8(int16_t * op0, vuint8mf8_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_i16mf4_m))) +void vsoxseg8ei8(vbool64_t op0, int16_t * op1, vuint8mf8_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, vint16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_u16m1))) +void vsoxseg8ei8(uint16_t * op0, vuint8mf2_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_u16m1_m))) +void vsoxseg8ei8(vbool16_t op0, uint16_t * op1, vuint8mf2_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, vuint16m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_u16mf2))) +void vsoxseg8ei8(uint16_t * op0, vuint8mf4_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_u16mf2_m))) +void vsoxseg8ei8(vbool32_t op0, uint16_t * op1, vuint8mf4_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, vuint16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_u16mf4))) +void vsoxseg8ei8(uint16_t * op0, vuint8mf8_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_u16mf4_m))) +void vsoxseg8ei8(vbool64_t op0, uint16_t * op1, vuint8mf8_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, vuint16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_i16m1))) +void vsoxseg2ei16(int16_t * op0, vuint16m1_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_i16m1_m))) +void vsoxseg2ei16(vbool16_t op0, int16_t * op1, vuint16m1_t op2, vint16m1_t op3, vint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_i16m2))) +void vsoxseg2ei16(int16_t * op0, vuint16m2_t op1, vint16m2_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_i16m2_m))) +void vsoxseg2ei16(vbool8_t op0, int16_t * op1, vuint16m2_t op2, vint16m2_t op3, vint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_i16m4))) +void vsoxseg2ei16(int16_t * op0, vuint16m4_t op1, vint16m4_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_i16m4_m))) +void vsoxseg2ei16(vbool4_t op0, int16_t * op1, vuint16m4_t op2, vint16m4_t op3, vint16m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_i16mf2))) +void vsoxseg2ei16(int16_t * op0, vuint16mf2_t op1, vint16mf2_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_i16mf2_m))) +void vsoxseg2ei16(vbool32_t op0, int16_t * op1, vuint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_i16mf4))) +void vsoxseg2ei16(int16_t * op0, vuint16mf4_t op1, vint16mf4_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_i16mf4_m))) +void vsoxseg2ei16(vbool64_t op0, int16_t * op1, vuint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_u16m1))) +void vsoxseg2ei16(uint16_t * op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_u16m1_m))) +void vsoxseg2ei16(vbool16_t op0, uint16_t * op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_u16m2))) +void vsoxseg2ei16(uint16_t * op0, vuint16m2_t op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_u16m2_m))) +void vsoxseg2ei16(vbool8_t op0, uint16_t * op1, vuint16m2_t op2, vuint16m2_t op3, vuint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_u16m4))) +void vsoxseg2ei16(uint16_t * op0, vuint16m4_t op1, vuint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_u16m4_m))) +void vsoxseg2ei16(vbool4_t op0, uint16_t * op1, vuint16m4_t op2, vuint16m4_t op3, vuint16m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_u16mf2))) +void vsoxseg2ei16(uint16_t * op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_u16mf2_m))) +void vsoxseg2ei16(vbool32_t op0, uint16_t * op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_u16mf4))) +void vsoxseg2ei16(uint16_t * op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_u16mf4_m))) +void vsoxseg2ei16(vbool64_t op0, uint16_t * op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_i16m1))) +void vsoxseg3ei16(int16_t * op0, vuint16m1_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_i16m1_m))) +void vsoxseg3ei16(vbool16_t op0, int16_t * op1, vuint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_i16m2))) +void vsoxseg3ei16(int16_t * op0, vuint16m2_t op1, vint16m2_t op2, vint16m2_t op3, vint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_i16m2_m))) +void vsoxseg3ei16(vbool8_t op0, int16_t * op1, vuint16m2_t op2, vint16m2_t op3, vint16m2_t op4, vint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_i16mf2))) +void vsoxseg3ei16(int16_t * op0, vuint16mf2_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_i16mf2_m))) +void vsoxseg3ei16(vbool32_t op0, int16_t * op1, vuint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_i16mf4))) +void vsoxseg3ei16(int16_t * op0, vuint16mf4_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_i16mf4_m))) +void vsoxseg3ei16(vbool64_t op0, int16_t * op1, vuint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_u16m1))) +void vsoxseg3ei16(uint16_t * op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_u16m1_m))) +void vsoxseg3ei16(vbool16_t op0, uint16_t * op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_u16m2))) +void vsoxseg3ei16(uint16_t * op0, vuint16m2_t op1, vuint16m2_t op2, vuint16m2_t op3, vuint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_u16m2_m))) +void vsoxseg3ei16(vbool8_t op0, uint16_t * op1, vuint16m2_t op2, vuint16m2_t op3, vuint16m2_t op4, vuint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_u16mf2))) +void vsoxseg3ei16(uint16_t * op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_u16mf2_m))) +void vsoxseg3ei16(vbool32_t op0, uint16_t * op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_u16mf4))) +void vsoxseg3ei16(uint16_t * op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_u16mf4_m))) +void vsoxseg3ei16(vbool64_t op0, uint16_t * op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_i16m1))) +void vsoxseg4ei16(int16_t * op0, vuint16m1_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_i16m1_m))) +void vsoxseg4ei16(vbool16_t op0, int16_t * op1, vuint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_i16m2))) +void vsoxseg4ei16(int16_t * op0, vuint16m2_t op1, vint16m2_t op2, vint16m2_t op3, vint16m2_t op4, vint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_i16m2_m))) +void vsoxseg4ei16(vbool8_t op0, int16_t * op1, vuint16m2_t op2, vint16m2_t op3, vint16m2_t op4, vint16m2_t op5, vint16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_i16mf2))) +void vsoxseg4ei16(int16_t * op0, vuint16mf2_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_i16mf2_m))) +void vsoxseg4ei16(vbool32_t op0, int16_t * op1, vuint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_i16mf4))) +void vsoxseg4ei16(int16_t * op0, vuint16mf4_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_i16mf4_m))) +void vsoxseg4ei16(vbool64_t op0, int16_t * op1, vuint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_u16m1))) +void vsoxseg4ei16(uint16_t * op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_u16m1_m))) +void vsoxseg4ei16(vbool16_t op0, uint16_t * op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_u16m2))) +void vsoxseg4ei16(uint16_t * op0, vuint16m2_t op1, vuint16m2_t op2, vuint16m2_t op3, vuint16m2_t op4, vuint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_u16m2_m))) +void vsoxseg4ei16(vbool8_t op0, uint16_t * op1, vuint16m2_t op2, vuint16m2_t op3, vuint16m2_t op4, vuint16m2_t op5, vuint16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_u16mf2))) +void vsoxseg4ei16(uint16_t * op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_u16mf2_m))) +void vsoxseg4ei16(vbool32_t op0, uint16_t * op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_u16mf4))) +void vsoxseg4ei16(uint16_t * op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_u16mf4_m))) +void vsoxseg4ei16(vbool64_t op0, uint16_t * op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_i16m1))) +void vsoxseg5ei16(int16_t * op0, vuint16m1_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_i16m1_m))) +void vsoxseg5ei16(vbool16_t op0, int16_t * op1, vuint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_i16mf2))) +void vsoxseg5ei16(int16_t * op0, vuint16mf2_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_i16mf2_m))) +void vsoxseg5ei16(vbool32_t op0, int16_t * op1, vuint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_i16mf4))) +void vsoxseg5ei16(int16_t * op0, vuint16mf4_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_i16mf4_m))) +void vsoxseg5ei16(vbool64_t op0, int16_t * op1, vuint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_u16m1))) +void vsoxseg5ei16(uint16_t * op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_u16m1_m))) +void vsoxseg5ei16(vbool16_t op0, uint16_t * op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_u16mf2))) +void vsoxseg5ei16(uint16_t * op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_u16mf2_m))) +void vsoxseg5ei16(vbool32_t op0, uint16_t * op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_u16mf4))) +void vsoxseg5ei16(uint16_t * op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_u16mf4_m))) +void vsoxseg5ei16(vbool64_t op0, uint16_t * op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_i16m1))) +void vsoxseg6ei16(int16_t * op0, vuint16m1_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_i16m1_m))) +void vsoxseg6ei16(vbool16_t op0, int16_t * op1, vuint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_i16mf2))) +void vsoxseg6ei16(int16_t * op0, vuint16mf2_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_i16mf2_m))) +void vsoxseg6ei16(vbool32_t op0, int16_t * op1, vuint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_i16mf4))) +void vsoxseg6ei16(int16_t * op0, vuint16mf4_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_i16mf4_m))) +void vsoxseg6ei16(vbool64_t op0, int16_t * op1, vuint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_u16m1))) +void vsoxseg6ei16(uint16_t * op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_u16m1_m))) +void vsoxseg6ei16(vbool16_t op0, uint16_t * op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_u16mf2))) +void vsoxseg6ei16(uint16_t * op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_u16mf2_m))) +void vsoxseg6ei16(vbool32_t op0, uint16_t * op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_u16mf4))) +void vsoxseg6ei16(uint16_t * op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_u16mf4_m))) +void vsoxseg6ei16(vbool64_t op0, uint16_t * op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_i16m1))) +void vsoxseg7ei16(int16_t * op0, vuint16m1_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_i16m1_m))) +void vsoxseg7ei16(vbool16_t op0, int16_t * op1, vuint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_i16mf2))) +void vsoxseg7ei16(int16_t * op0, vuint16mf2_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_i16mf2_m))) +void vsoxseg7ei16(vbool32_t op0, int16_t * op1, vuint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_i16mf4))) +void vsoxseg7ei16(int16_t * op0, vuint16mf4_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_i16mf4_m))) +void vsoxseg7ei16(vbool64_t op0, int16_t * op1, vuint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_u16m1))) +void vsoxseg7ei16(uint16_t * op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_u16m1_m))) +void vsoxseg7ei16(vbool16_t op0, uint16_t * op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_u16mf2))) +void vsoxseg7ei16(uint16_t * op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_u16mf2_m))) +void vsoxseg7ei16(vbool32_t op0, uint16_t * op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_u16mf4))) +void vsoxseg7ei16(uint16_t * op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_u16mf4_m))) +void vsoxseg7ei16(vbool64_t op0, uint16_t * op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_i16m1))) +void vsoxseg8ei16(int16_t * op0, vuint16m1_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_i16m1_m))) +void vsoxseg8ei16(vbool16_t op0, int16_t * op1, vuint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, vint16m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_i16mf2))) +void vsoxseg8ei16(int16_t * op0, vuint16mf2_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_i16mf2_m))) +void vsoxseg8ei16(vbool32_t op0, int16_t * op1, vuint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, vint16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_i16mf4))) +void vsoxseg8ei16(int16_t * op0, vuint16mf4_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_i16mf4_m))) +void vsoxseg8ei16(vbool64_t op0, int16_t * op1, vuint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, vint16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_u16m1))) +void vsoxseg8ei16(uint16_t * op0, vuint16m1_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_u16m1_m))) +void vsoxseg8ei16(vbool16_t op0, uint16_t * op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, vuint16m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_u16mf2))) +void vsoxseg8ei16(uint16_t * op0, vuint16mf2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_u16mf2_m))) +void vsoxseg8ei16(vbool32_t op0, uint16_t * op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, vuint16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_u16mf4))) +void vsoxseg8ei16(uint16_t * op0, vuint16mf4_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_u16mf4_m))) +void vsoxseg8ei16(vbool64_t op0, uint16_t * op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, vuint16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_i16m1))) +void vsoxseg2ei32(int16_t * op0, vuint32m2_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_i16m1_m))) +void vsoxseg2ei32(vbool16_t op0, int16_t * op1, vuint32m2_t op2, vint16m1_t op3, vint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_i16m2))) +void vsoxseg2ei32(int16_t * op0, vuint32m4_t op1, vint16m2_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_i16m2_m))) +void vsoxseg2ei32(vbool8_t op0, int16_t * op1, vuint32m4_t op2, vint16m2_t op3, vint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_i16m4))) +void vsoxseg2ei32(int16_t * op0, vuint32m8_t op1, vint16m4_t op2, vint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_i16m4_m))) +void vsoxseg2ei32(vbool4_t op0, int16_t * op1, vuint32m8_t op2, vint16m4_t op3, vint16m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_i16mf2))) +void vsoxseg2ei32(int16_t * op0, vuint32m1_t op1, vint16mf2_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_i16mf2_m))) +void vsoxseg2ei32(vbool32_t op0, int16_t * op1, vuint32m1_t op2, vint16mf2_t op3, vint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_i16mf4))) +void vsoxseg2ei32(int16_t * op0, vuint32mf2_t op1, vint16mf4_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_i16mf4_m))) +void vsoxseg2ei32(vbool64_t op0, int16_t * op1, vuint32mf2_t op2, vint16mf4_t op3, vint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_u16m1))) +void vsoxseg2ei32(uint16_t * op0, vuint32m2_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_u16m1_m))) +void vsoxseg2ei32(vbool16_t op0, uint16_t * op1, vuint32m2_t op2, vuint16m1_t op3, vuint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_u16m2))) +void vsoxseg2ei32(uint16_t * op0, vuint32m4_t op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_u16m2_m))) +void vsoxseg2ei32(vbool8_t op0, uint16_t * op1, vuint32m4_t op2, vuint16m2_t op3, vuint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_u16m4))) +void vsoxseg2ei32(uint16_t * op0, vuint32m8_t op1, vuint16m4_t op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_u16m4_m))) +void vsoxseg2ei32(vbool4_t op0, uint16_t * op1, vuint32m8_t op2, vuint16m4_t op3, vuint16m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_u16mf2))) +void vsoxseg2ei32(uint16_t * op0, vuint32m1_t op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_u16mf2_m))) +void vsoxseg2ei32(vbool32_t op0, uint16_t * op1, vuint32m1_t op2, vuint16mf2_t op3, vuint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_u16mf4))) +void vsoxseg2ei32(uint16_t * op0, vuint32mf2_t op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_u16mf4_m))) +void vsoxseg2ei32(vbool64_t op0, uint16_t * op1, vuint32mf2_t op2, vuint16mf4_t op3, vuint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_i16m1))) +void vsoxseg3ei32(int16_t * op0, vuint32m2_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_i16m1_m))) +void vsoxseg3ei32(vbool16_t op0, int16_t * op1, vuint32m2_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_i16m2))) +void vsoxseg3ei32(int16_t * op0, vuint32m4_t op1, vint16m2_t op2, vint16m2_t op3, vint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_i16m2_m))) +void vsoxseg3ei32(vbool8_t op0, int16_t * op1, vuint32m4_t op2, vint16m2_t op3, vint16m2_t op4, vint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_i16mf2))) +void vsoxseg3ei32(int16_t * op0, vuint32m1_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_i16mf2_m))) +void vsoxseg3ei32(vbool32_t op0, int16_t * op1, vuint32m1_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_i16mf4))) +void vsoxseg3ei32(int16_t * op0, vuint32mf2_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_i16mf4_m))) +void vsoxseg3ei32(vbool64_t op0, int16_t * op1, vuint32mf2_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_u16m1))) +void vsoxseg3ei32(uint16_t * op0, vuint32m2_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_u16m1_m))) +void vsoxseg3ei32(vbool16_t op0, uint16_t * op1, vuint32m2_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_u16m2))) +void vsoxseg3ei32(uint16_t * op0, vuint32m4_t op1, vuint16m2_t op2, vuint16m2_t op3, vuint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_u16m2_m))) +void vsoxseg3ei32(vbool8_t op0, uint16_t * op1, vuint32m4_t op2, vuint16m2_t op3, vuint16m2_t op4, vuint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_u16mf2))) +void vsoxseg3ei32(uint16_t * op0, vuint32m1_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_u16mf2_m))) +void vsoxseg3ei32(vbool32_t op0, uint16_t * op1, vuint32m1_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_u16mf4))) +void vsoxseg3ei32(uint16_t * op0, vuint32mf2_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_u16mf4_m))) +void vsoxseg3ei32(vbool64_t op0, uint16_t * op1, vuint32mf2_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_i16m1))) +void vsoxseg4ei32(int16_t * op0, vuint32m2_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_i16m1_m))) +void vsoxseg4ei32(vbool16_t op0, int16_t * op1, vuint32m2_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_i16m2))) +void vsoxseg4ei32(int16_t * op0, vuint32m4_t op1, vint16m2_t op2, vint16m2_t op3, vint16m2_t op4, vint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_i16m2_m))) +void vsoxseg4ei32(vbool8_t op0, int16_t * op1, vuint32m4_t op2, vint16m2_t op3, vint16m2_t op4, vint16m2_t op5, vint16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_i16mf2))) +void vsoxseg4ei32(int16_t * op0, vuint32m1_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_i16mf2_m))) +void vsoxseg4ei32(vbool32_t op0, int16_t * op1, vuint32m1_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_i16mf4))) +void vsoxseg4ei32(int16_t * op0, vuint32mf2_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_i16mf4_m))) +void vsoxseg4ei32(vbool64_t op0, int16_t * op1, vuint32mf2_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_u16m1))) +void vsoxseg4ei32(uint16_t * op0, vuint32m2_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_u16m1_m))) +void vsoxseg4ei32(vbool16_t op0, uint16_t * op1, vuint32m2_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_u16m2))) +void vsoxseg4ei32(uint16_t * op0, vuint32m4_t op1, vuint16m2_t op2, vuint16m2_t op3, vuint16m2_t op4, vuint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_u16m2_m))) +void vsoxseg4ei32(vbool8_t op0, uint16_t * op1, vuint32m4_t op2, vuint16m2_t op3, vuint16m2_t op4, vuint16m2_t op5, vuint16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_u16mf2))) +void vsoxseg4ei32(uint16_t * op0, vuint32m1_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_u16mf2_m))) +void vsoxseg4ei32(vbool32_t op0, uint16_t * op1, vuint32m1_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_u16mf4))) +void vsoxseg4ei32(uint16_t * op0, vuint32mf2_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_u16mf4_m))) +void vsoxseg4ei32(vbool64_t op0, uint16_t * op1, vuint32mf2_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_i16m1))) +void vsoxseg5ei32(int16_t * op0, vuint32m2_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_i16m1_m))) +void vsoxseg5ei32(vbool16_t op0, int16_t * op1, vuint32m2_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_i16mf2))) +void vsoxseg5ei32(int16_t * op0, vuint32m1_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_i16mf2_m))) +void vsoxseg5ei32(vbool32_t op0, int16_t * op1, vuint32m1_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_i16mf4))) +void vsoxseg5ei32(int16_t * op0, vuint32mf2_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_i16mf4_m))) +void vsoxseg5ei32(vbool64_t op0, int16_t * op1, vuint32mf2_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_u16m1))) +void vsoxseg5ei32(uint16_t * op0, vuint32m2_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_u16m1_m))) +void vsoxseg5ei32(vbool16_t op0, uint16_t * op1, vuint32m2_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_u16mf2))) +void vsoxseg5ei32(uint16_t * op0, vuint32m1_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_u16mf2_m))) +void vsoxseg5ei32(vbool32_t op0, uint16_t * op1, vuint32m1_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_u16mf4))) +void vsoxseg5ei32(uint16_t * op0, vuint32mf2_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_u16mf4_m))) +void vsoxseg5ei32(vbool64_t op0, uint16_t * op1, vuint32mf2_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_i16m1))) +void vsoxseg6ei32(int16_t * op0, vuint32m2_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_i16m1_m))) +void vsoxseg6ei32(vbool16_t op0, int16_t * op1, vuint32m2_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_i16mf2))) +void vsoxseg6ei32(int16_t * op0, vuint32m1_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_i16mf2_m))) +void vsoxseg6ei32(vbool32_t op0, int16_t * op1, vuint32m1_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_i16mf4))) +void vsoxseg6ei32(int16_t * op0, vuint32mf2_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_i16mf4_m))) +void vsoxseg6ei32(vbool64_t op0, int16_t * op1, vuint32mf2_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_u16m1))) +void vsoxseg6ei32(uint16_t * op0, vuint32m2_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_u16m1_m))) +void vsoxseg6ei32(vbool16_t op0, uint16_t * op1, vuint32m2_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_u16mf2))) +void vsoxseg6ei32(uint16_t * op0, vuint32m1_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_u16mf2_m))) +void vsoxseg6ei32(vbool32_t op0, uint16_t * op1, vuint32m1_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_u16mf4))) +void vsoxseg6ei32(uint16_t * op0, vuint32mf2_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_u16mf4_m))) +void vsoxseg6ei32(vbool64_t op0, uint16_t * op1, vuint32mf2_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_i16m1))) +void vsoxseg7ei32(int16_t * op0, vuint32m2_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_i16m1_m))) +void vsoxseg7ei32(vbool16_t op0, int16_t * op1, vuint32m2_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_i16mf2))) +void vsoxseg7ei32(int16_t * op0, vuint32m1_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_i16mf2_m))) +void vsoxseg7ei32(vbool32_t op0, int16_t * op1, vuint32m1_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_i16mf4))) +void vsoxseg7ei32(int16_t * op0, vuint32mf2_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_i16mf4_m))) +void vsoxseg7ei32(vbool64_t op0, int16_t * op1, vuint32mf2_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_u16m1))) +void vsoxseg7ei32(uint16_t * op0, vuint32m2_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_u16m1_m))) +void vsoxseg7ei32(vbool16_t op0, uint16_t * op1, vuint32m2_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_u16mf2))) +void vsoxseg7ei32(uint16_t * op0, vuint32m1_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_u16mf2_m))) +void vsoxseg7ei32(vbool32_t op0, uint16_t * op1, vuint32m1_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_u16mf4))) +void vsoxseg7ei32(uint16_t * op0, vuint32mf2_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_u16mf4_m))) +void vsoxseg7ei32(vbool64_t op0, uint16_t * op1, vuint32mf2_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_i16m1))) +void vsoxseg8ei32(int16_t * op0, vuint32m2_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_i16m1_m))) +void vsoxseg8ei32(vbool16_t op0, int16_t * op1, vuint32m2_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, vint16m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_i16mf2))) +void vsoxseg8ei32(int16_t * op0, vuint32m1_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_i16mf2_m))) +void vsoxseg8ei32(vbool32_t op0, int16_t * op1, vuint32m1_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, vint16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_i16mf4))) +void vsoxseg8ei32(int16_t * op0, vuint32mf2_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_i16mf4_m))) +void vsoxseg8ei32(vbool64_t op0, int16_t * op1, vuint32mf2_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, vint16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_u16m1))) +void vsoxseg8ei32(uint16_t * op0, vuint32m2_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_u16m1_m))) +void vsoxseg8ei32(vbool16_t op0, uint16_t * op1, vuint32m2_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, vuint16m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_u16mf2))) +void vsoxseg8ei32(uint16_t * op0, vuint32m1_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_u16mf2_m))) +void vsoxseg8ei32(vbool32_t op0, uint16_t * op1, vuint32m1_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, vuint16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_u16mf4))) +void vsoxseg8ei32(uint16_t * op0, vuint32mf2_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_u16mf4_m))) +void vsoxseg8ei32(vbool64_t op0, uint16_t * op1, vuint32mf2_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, vuint16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_i16m1))) +void vsoxseg2ei64(int16_t * op0, vuint64m4_t op1, vint16m1_t op2, vint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_i16m1_m))) +void vsoxseg2ei64(vbool16_t op0, int16_t * op1, vuint64m4_t op2, vint16m1_t op3, vint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_i16m2))) +void vsoxseg2ei64(int16_t * op0, vuint64m8_t op1, vint16m2_t op2, vint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_i16m2_m))) +void vsoxseg2ei64(vbool8_t op0, int16_t * op1, vuint64m8_t op2, vint16m2_t op3, vint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_i16mf2))) +void vsoxseg2ei64(int16_t * op0, vuint64m2_t op1, vint16mf2_t op2, vint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_i16mf2_m))) +void vsoxseg2ei64(vbool32_t op0, int16_t * op1, vuint64m2_t op2, vint16mf2_t op3, vint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_i16mf4))) +void vsoxseg2ei64(int16_t * op0, vuint64m1_t op1, vint16mf4_t op2, vint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_i16mf4_m))) +void vsoxseg2ei64(vbool64_t op0, int16_t * op1, vuint64m1_t op2, vint16mf4_t op3, vint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_u16m1))) +void vsoxseg2ei64(uint16_t * op0, vuint64m4_t op1, vuint16m1_t op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_u16m1_m))) +void vsoxseg2ei64(vbool16_t op0, uint16_t * op1, vuint64m4_t op2, vuint16m1_t op3, vuint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_u16m2))) +void vsoxseg2ei64(uint16_t * op0, vuint64m8_t op1, vuint16m2_t op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_u16m2_m))) +void vsoxseg2ei64(vbool8_t op0, uint16_t * op1, vuint64m8_t op2, vuint16m2_t op3, vuint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_u16mf2))) +void vsoxseg2ei64(uint16_t * op0, vuint64m2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_u16mf2_m))) +void vsoxseg2ei64(vbool32_t op0, uint16_t * op1, vuint64m2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_u16mf4))) +void vsoxseg2ei64(uint16_t * op0, vuint64m1_t op1, vuint16mf4_t op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_u16mf4_m))) +void vsoxseg2ei64(vbool64_t op0, uint16_t * op1, vuint64m1_t op2, vuint16mf4_t op3, vuint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_i16m1))) +void vsoxseg3ei64(int16_t * op0, vuint64m4_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_i16m1_m))) +void vsoxseg3ei64(vbool16_t op0, int16_t * op1, vuint64m4_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_i16m2))) +void vsoxseg3ei64(int16_t * op0, vuint64m8_t op1, vint16m2_t op2, vint16m2_t op3, vint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_i16m2_m))) +void vsoxseg3ei64(vbool8_t op0, int16_t * op1, vuint64m8_t op2, vint16m2_t op3, vint16m2_t op4, vint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_i16mf2))) +void vsoxseg3ei64(int16_t * op0, vuint64m2_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_i16mf2_m))) +void vsoxseg3ei64(vbool32_t op0, int16_t * op1, vuint64m2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_i16mf4))) +void vsoxseg3ei64(int16_t * op0, vuint64m1_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_i16mf4_m))) +void vsoxseg3ei64(vbool64_t op0, int16_t * op1, vuint64m1_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_u16m1))) +void vsoxseg3ei64(uint16_t * op0, vuint64m4_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_u16m1_m))) +void vsoxseg3ei64(vbool16_t op0, uint16_t * op1, vuint64m4_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_u16m2))) +void vsoxseg3ei64(uint16_t * op0, vuint64m8_t op1, vuint16m2_t op2, vuint16m2_t op3, vuint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_u16m2_m))) +void vsoxseg3ei64(vbool8_t op0, uint16_t * op1, vuint64m8_t op2, vuint16m2_t op3, vuint16m2_t op4, vuint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_u16mf2))) +void vsoxseg3ei64(uint16_t * op0, vuint64m2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_u16mf2_m))) +void vsoxseg3ei64(vbool32_t op0, uint16_t * op1, vuint64m2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_u16mf4))) +void vsoxseg3ei64(uint16_t * op0, vuint64m1_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_u16mf4_m))) +void vsoxseg3ei64(vbool64_t op0, uint16_t * op1, vuint64m1_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_i16m1))) +void vsoxseg4ei64(int16_t * op0, vuint64m4_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_i16m1_m))) +void vsoxseg4ei64(vbool16_t op0, int16_t * op1, vuint64m4_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_i16m2))) +void vsoxseg4ei64(int16_t * op0, vuint64m8_t op1, vint16m2_t op2, vint16m2_t op3, vint16m2_t op4, vint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_i16m2_m))) +void vsoxseg4ei64(vbool8_t op0, int16_t * op1, vuint64m8_t op2, vint16m2_t op3, vint16m2_t op4, vint16m2_t op5, vint16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_i16mf2))) +void vsoxseg4ei64(int16_t * op0, vuint64m2_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_i16mf2_m))) +void vsoxseg4ei64(vbool32_t op0, int16_t * op1, vuint64m2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_i16mf4))) +void vsoxseg4ei64(int16_t * op0, vuint64m1_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_i16mf4_m))) +void vsoxseg4ei64(vbool64_t op0, int16_t * op1, vuint64m1_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_u16m1))) +void vsoxseg4ei64(uint16_t * op0, vuint64m4_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_u16m1_m))) +void vsoxseg4ei64(vbool16_t op0, uint16_t * op1, vuint64m4_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_u16m2))) +void vsoxseg4ei64(uint16_t * op0, vuint64m8_t op1, vuint16m2_t op2, vuint16m2_t op3, vuint16m2_t op4, vuint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_u16m2_m))) +void vsoxseg4ei64(vbool8_t op0, uint16_t * op1, vuint64m8_t op2, vuint16m2_t op3, vuint16m2_t op4, vuint16m2_t op5, vuint16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_u16mf2))) +void vsoxseg4ei64(uint16_t * op0, vuint64m2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_u16mf2_m))) +void vsoxseg4ei64(vbool32_t op0, uint16_t * op1, vuint64m2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_u16mf4))) +void vsoxseg4ei64(uint16_t * op0, vuint64m1_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_u16mf4_m))) +void vsoxseg4ei64(vbool64_t op0, uint16_t * op1, vuint64m1_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_i16m1))) +void vsoxseg5ei64(int16_t * op0, vuint64m4_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_i16m1_m))) +void vsoxseg5ei64(vbool16_t op0, int16_t * op1, vuint64m4_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_i16mf2))) +void vsoxseg5ei64(int16_t * op0, vuint64m2_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_i16mf2_m))) +void vsoxseg5ei64(vbool32_t op0, int16_t * op1, vuint64m2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_i16mf4))) +void vsoxseg5ei64(int16_t * op0, vuint64m1_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_i16mf4_m))) +void vsoxseg5ei64(vbool64_t op0, int16_t * op1, vuint64m1_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_u16m1))) +void vsoxseg5ei64(uint16_t * op0, vuint64m4_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_u16m1_m))) +void vsoxseg5ei64(vbool16_t op0, uint16_t * op1, vuint64m4_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_u16mf2))) +void vsoxseg5ei64(uint16_t * op0, vuint64m2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_u16mf2_m))) +void vsoxseg5ei64(vbool32_t op0, uint16_t * op1, vuint64m2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_u16mf4))) +void vsoxseg5ei64(uint16_t * op0, vuint64m1_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_u16mf4_m))) +void vsoxseg5ei64(vbool64_t op0, uint16_t * op1, vuint64m1_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_i16m1))) +void vsoxseg6ei64(int16_t * op0, vuint64m4_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_i16m1_m))) +void vsoxseg6ei64(vbool16_t op0, int16_t * op1, vuint64m4_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_i16mf2))) +void vsoxseg6ei64(int16_t * op0, vuint64m2_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_i16mf2_m))) +void vsoxseg6ei64(vbool32_t op0, int16_t * op1, vuint64m2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_i16mf4))) +void vsoxseg6ei64(int16_t * op0, vuint64m1_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_i16mf4_m))) +void vsoxseg6ei64(vbool64_t op0, int16_t * op1, vuint64m1_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_u16m1))) +void vsoxseg6ei64(uint16_t * op0, vuint64m4_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_u16m1_m))) +void vsoxseg6ei64(vbool16_t op0, uint16_t * op1, vuint64m4_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_u16mf2))) +void vsoxseg6ei64(uint16_t * op0, vuint64m2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_u16mf2_m))) +void vsoxseg6ei64(vbool32_t op0, uint16_t * op1, vuint64m2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_u16mf4))) +void vsoxseg6ei64(uint16_t * op0, vuint64m1_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_u16mf4_m))) +void vsoxseg6ei64(vbool64_t op0, uint16_t * op1, vuint64m1_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_i16m1))) +void vsoxseg7ei64(int16_t * op0, vuint64m4_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_i16m1_m))) +void vsoxseg7ei64(vbool16_t op0, int16_t * op1, vuint64m4_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_i16mf2))) +void vsoxseg7ei64(int16_t * op0, vuint64m2_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_i16mf2_m))) +void vsoxseg7ei64(vbool32_t op0, int16_t * op1, vuint64m2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_i16mf4))) +void vsoxseg7ei64(int16_t * op0, vuint64m1_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_i16mf4_m))) +void vsoxseg7ei64(vbool64_t op0, int16_t * op1, vuint64m1_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_u16m1))) +void vsoxseg7ei64(uint16_t * op0, vuint64m4_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_u16m1_m))) +void vsoxseg7ei64(vbool16_t op0, uint16_t * op1, vuint64m4_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_u16mf2))) +void vsoxseg7ei64(uint16_t * op0, vuint64m2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_u16mf2_m))) +void vsoxseg7ei64(vbool32_t op0, uint16_t * op1, vuint64m2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_u16mf4))) +void vsoxseg7ei64(uint16_t * op0, vuint64m1_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_u16mf4_m))) +void vsoxseg7ei64(vbool64_t op0, uint16_t * op1, vuint64m1_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_i16m1))) +void vsoxseg8ei64(int16_t * op0, vuint64m4_t op1, vint16m1_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_i16m1_m))) +void vsoxseg8ei64(vbool16_t op0, int16_t * op1, vuint64m4_t op2, vint16m1_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, vint16m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_i16mf2))) +void vsoxseg8ei64(int16_t * op0, vuint64m2_t op1, vint16mf2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_i16mf2_m))) +void vsoxseg8ei64(vbool32_t op0, int16_t * op1, vuint64m2_t op2, vint16mf2_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, vint16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_i16mf4))) +void vsoxseg8ei64(int16_t * op0, vuint64m1_t op1, vint16mf4_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_i16mf4_m))) +void vsoxseg8ei64(vbool64_t op0, int16_t * op1, vuint64m1_t op2, vint16mf4_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, vint16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_u16m1))) +void vsoxseg8ei64(uint16_t * op0, vuint64m4_t op1, vuint16m1_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_u16m1_m))) +void vsoxseg8ei64(vbool16_t op0, uint16_t * op1, vuint64m4_t op2, vuint16m1_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, vuint16m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_u16mf2))) +void vsoxseg8ei64(uint16_t * op0, vuint64m2_t op1, vuint16mf2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_u16mf2_m))) +void vsoxseg8ei64(vbool32_t op0, uint16_t * op1, vuint64m2_t op2, vuint16mf2_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, vuint16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_u16mf4))) +void vsoxseg8ei64(uint16_t * op0, vuint64m1_t op1, vuint16mf4_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_u16mf4_m))) +void vsoxseg8ei64(vbool64_t op0, uint16_t * op1, vuint64m1_t op2, vuint16mf4_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, vuint16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_i32m1))) +void vsoxseg2ei8(int32_t * op0, vuint8mf4_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_i32m1_m))) +void vsoxseg2ei8(vbool32_t op0, int32_t * op1, vuint8mf4_t op2, vint32m1_t op3, vint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_i32m2))) +void vsoxseg2ei8(int32_t * op0, vuint8mf2_t op1, vint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_i32m2_m))) +void vsoxseg2ei8(vbool16_t op0, int32_t * op1, vuint8mf2_t op2, vint32m2_t op3, vint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_i32m4))) +void vsoxseg2ei8(int32_t * op0, vuint8m1_t op1, vint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_i32m4_m))) +void vsoxseg2ei8(vbool8_t op0, int32_t * op1, vuint8m1_t op2, vint32m4_t op3, vint32m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_i32mf2))) +void vsoxseg2ei8(int32_t * op0, vuint8mf8_t op1, vint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_i32mf2_m))) +void vsoxseg2ei8(vbool64_t op0, int32_t * op1, vuint8mf8_t op2, vint32mf2_t op3, vint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_u32m1))) +void vsoxseg2ei8(uint32_t * op0, vuint8mf4_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_u32m1_m))) +void vsoxseg2ei8(vbool32_t op0, uint32_t * op1, vuint8mf4_t op2, vuint32m1_t op3, vuint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_u32m2))) +void vsoxseg2ei8(uint32_t * op0, vuint8mf2_t op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_u32m2_m))) +void vsoxseg2ei8(vbool16_t op0, uint32_t * op1, vuint8mf2_t op2, vuint32m2_t op3, vuint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_u32m4))) +void vsoxseg2ei8(uint32_t * op0, vuint8m1_t op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_u32m4_m))) +void vsoxseg2ei8(vbool8_t op0, uint32_t * op1, vuint8m1_t op2, vuint32m4_t op3, vuint32m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_u32mf2))) +void vsoxseg2ei8(uint32_t * op0, vuint8mf8_t op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_u32mf2_m))) +void vsoxseg2ei8(vbool64_t op0, uint32_t * op1, vuint8mf8_t op2, vuint32mf2_t op3, vuint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_i32m1))) +void vsoxseg3ei8(int32_t * op0, vuint8mf4_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_i32m1_m))) +void vsoxseg3ei8(vbool32_t op0, int32_t * op1, vuint8mf4_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_i32m2))) +void vsoxseg3ei8(int32_t * op0, vuint8mf2_t op1, vint32m2_t op2, vint32m2_t op3, vint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_i32m2_m))) +void vsoxseg3ei8(vbool16_t op0, int32_t * op1, vuint8mf2_t op2, vint32m2_t op3, vint32m2_t op4, vint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_i32mf2))) +void vsoxseg3ei8(int32_t * op0, vuint8mf8_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_i32mf2_m))) +void vsoxseg3ei8(vbool64_t op0, int32_t * op1, vuint8mf8_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_u32m1))) +void vsoxseg3ei8(uint32_t * op0, vuint8mf4_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_u32m1_m))) +void vsoxseg3ei8(vbool32_t op0, uint32_t * op1, vuint8mf4_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_u32m2))) +void vsoxseg3ei8(uint32_t * op0, vuint8mf2_t op1, vuint32m2_t op2, vuint32m2_t op3, vuint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_u32m2_m))) +void vsoxseg3ei8(vbool16_t op0, uint32_t * op1, vuint8mf2_t op2, vuint32m2_t op3, vuint32m2_t op4, vuint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_u32mf2))) +void vsoxseg3ei8(uint32_t * op0, vuint8mf8_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_u32mf2_m))) +void vsoxseg3ei8(vbool64_t op0, uint32_t * op1, vuint8mf8_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_i32m1))) +void vsoxseg4ei8(int32_t * op0, vuint8mf4_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_i32m1_m))) +void vsoxseg4ei8(vbool32_t op0, int32_t * op1, vuint8mf4_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_i32m2))) +void vsoxseg4ei8(int32_t * op0, vuint8mf2_t op1, vint32m2_t op2, vint32m2_t op3, vint32m2_t op4, vint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_i32m2_m))) +void vsoxseg4ei8(vbool16_t op0, int32_t * op1, vuint8mf2_t op2, vint32m2_t op3, vint32m2_t op4, vint32m2_t op5, vint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_i32mf2))) +void vsoxseg4ei8(int32_t * op0, vuint8mf8_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_i32mf2_m))) +void vsoxseg4ei8(vbool64_t op0, int32_t * op1, vuint8mf8_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_u32m1))) +void vsoxseg4ei8(uint32_t * op0, vuint8mf4_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_u32m1_m))) +void vsoxseg4ei8(vbool32_t op0, uint32_t * op1, vuint8mf4_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_u32m2))) +void vsoxseg4ei8(uint32_t * op0, vuint8mf2_t op1, vuint32m2_t op2, vuint32m2_t op3, vuint32m2_t op4, vuint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_u32m2_m))) +void vsoxseg4ei8(vbool16_t op0, uint32_t * op1, vuint8mf2_t op2, vuint32m2_t op3, vuint32m2_t op4, vuint32m2_t op5, vuint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_u32mf2))) +void vsoxseg4ei8(uint32_t * op0, vuint8mf8_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_u32mf2_m))) +void vsoxseg4ei8(vbool64_t op0, uint32_t * op1, vuint8mf8_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_i32m1))) +void vsoxseg5ei8(int32_t * op0, vuint8mf4_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_i32m1_m))) +void vsoxseg5ei8(vbool32_t op0, int32_t * op1, vuint8mf4_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_i32mf2))) +void vsoxseg5ei8(int32_t * op0, vuint8mf8_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_i32mf2_m))) +void vsoxseg5ei8(vbool64_t op0, int32_t * op1, vuint8mf8_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_u32m1))) +void vsoxseg5ei8(uint32_t * op0, vuint8mf4_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_u32m1_m))) +void vsoxseg5ei8(vbool32_t op0, uint32_t * op1, vuint8mf4_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_u32mf2))) +void vsoxseg5ei8(uint32_t * op0, vuint8mf8_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_u32mf2_m))) +void vsoxseg5ei8(vbool64_t op0, uint32_t * op1, vuint8mf8_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_i32m1))) +void vsoxseg6ei8(int32_t * op0, vuint8mf4_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_i32m1_m))) +void vsoxseg6ei8(vbool32_t op0, int32_t * op1, vuint8mf4_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_i32mf2))) +void vsoxseg6ei8(int32_t * op0, vuint8mf8_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_i32mf2_m))) +void vsoxseg6ei8(vbool64_t op0, int32_t * op1, vuint8mf8_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_u32m1))) +void vsoxseg6ei8(uint32_t * op0, vuint8mf4_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_u32m1_m))) +void vsoxseg6ei8(vbool32_t op0, uint32_t * op1, vuint8mf4_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_u32mf2))) +void vsoxseg6ei8(uint32_t * op0, vuint8mf8_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_u32mf2_m))) +void vsoxseg6ei8(vbool64_t op0, uint32_t * op1, vuint8mf8_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_i32m1))) +void vsoxseg7ei8(int32_t * op0, vuint8mf4_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_i32m1_m))) +void vsoxseg7ei8(vbool32_t op0, int32_t * op1, vuint8mf4_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_i32mf2))) +void vsoxseg7ei8(int32_t * op0, vuint8mf8_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_i32mf2_m))) +void vsoxseg7ei8(vbool64_t op0, int32_t * op1, vuint8mf8_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_u32m1))) +void vsoxseg7ei8(uint32_t * op0, vuint8mf4_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_u32m1_m))) +void vsoxseg7ei8(vbool32_t op0, uint32_t * op1, vuint8mf4_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_u32mf2))) +void vsoxseg7ei8(uint32_t * op0, vuint8mf8_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_u32mf2_m))) +void vsoxseg7ei8(vbool64_t op0, uint32_t * op1, vuint8mf8_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_i32m1))) +void vsoxseg8ei8(int32_t * op0, vuint8mf4_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_i32m1_m))) +void vsoxseg8ei8(vbool32_t op0, int32_t * op1, vuint8mf4_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, vint32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_i32mf2))) +void vsoxseg8ei8(int32_t * op0, vuint8mf8_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_i32mf2_m))) +void vsoxseg8ei8(vbool64_t op0, int32_t * op1, vuint8mf8_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, vint32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_u32m1))) +void vsoxseg8ei8(uint32_t * op0, vuint8mf4_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_u32m1_m))) +void vsoxseg8ei8(vbool32_t op0, uint32_t * op1, vuint8mf4_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, vuint32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_u32mf2))) +void vsoxseg8ei8(uint32_t * op0, vuint8mf8_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_u32mf2_m))) +void vsoxseg8ei8(vbool64_t op0, uint32_t * op1, vuint8mf8_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, vuint32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_i32m1))) +void vsoxseg2ei16(int32_t * op0, vuint16mf2_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_i32m1_m))) +void vsoxseg2ei16(vbool32_t op0, int32_t * op1, vuint16mf2_t op2, vint32m1_t op3, vint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_i32m2))) +void vsoxseg2ei16(int32_t * op0, vuint16m1_t op1, vint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_i32m2_m))) +void vsoxseg2ei16(vbool16_t op0, int32_t * op1, vuint16m1_t op2, vint32m2_t op3, vint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_i32m4))) +void vsoxseg2ei16(int32_t * op0, vuint16m2_t op1, vint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_i32m4_m))) +void vsoxseg2ei16(vbool8_t op0, int32_t * op1, vuint16m2_t op2, vint32m4_t op3, vint32m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_i32mf2))) +void vsoxseg2ei16(int32_t * op0, vuint16mf4_t op1, vint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_i32mf2_m))) +void vsoxseg2ei16(vbool64_t op0, int32_t * op1, vuint16mf4_t op2, vint32mf2_t op3, vint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_u32m1))) +void vsoxseg2ei16(uint32_t * op0, vuint16mf2_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_u32m1_m))) +void vsoxseg2ei16(vbool32_t op0, uint32_t * op1, vuint16mf2_t op2, vuint32m1_t op3, vuint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_u32m2))) +void vsoxseg2ei16(uint32_t * op0, vuint16m1_t op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_u32m2_m))) +void vsoxseg2ei16(vbool16_t op0, uint32_t * op1, vuint16m1_t op2, vuint32m2_t op3, vuint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_u32m4))) +void vsoxseg2ei16(uint32_t * op0, vuint16m2_t op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_u32m4_m))) +void vsoxseg2ei16(vbool8_t op0, uint32_t * op1, vuint16m2_t op2, vuint32m4_t op3, vuint32m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_u32mf2))) +void vsoxseg2ei16(uint32_t * op0, vuint16mf4_t op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_u32mf2_m))) +void vsoxseg2ei16(vbool64_t op0, uint32_t * op1, vuint16mf4_t op2, vuint32mf2_t op3, vuint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_i32m1))) +void vsoxseg3ei16(int32_t * op0, vuint16mf2_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_i32m1_m))) +void vsoxseg3ei16(vbool32_t op0, int32_t * op1, vuint16mf2_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_i32m2))) +void vsoxseg3ei16(int32_t * op0, vuint16m1_t op1, vint32m2_t op2, vint32m2_t op3, vint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_i32m2_m))) +void vsoxseg3ei16(vbool16_t op0, int32_t * op1, vuint16m1_t op2, vint32m2_t op3, vint32m2_t op4, vint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_i32mf2))) +void vsoxseg3ei16(int32_t * op0, vuint16mf4_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_i32mf2_m))) +void vsoxseg3ei16(vbool64_t op0, int32_t * op1, vuint16mf4_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_u32m1))) +void vsoxseg3ei16(uint32_t * op0, vuint16mf2_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_u32m1_m))) +void vsoxseg3ei16(vbool32_t op0, uint32_t * op1, vuint16mf2_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_u32m2))) +void vsoxseg3ei16(uint32_t * op0, vuint16m1_t op1, vuint32m2_t op2, vuint32m2_t op3, vuint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_u32m2_m))) +void vsoxseg3ei16(vbool16_t op0, uint32_t * op1, vuint16m1_t op2, vuint32m2_t op3, vuint32m2_t op4, vuint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_u32mf2))) +void vsoxseg3ei16(uint32_t * op0, vuint16mf4_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_u32mf2_m))) +void vsoxseg3ei16(vbool64_t op0, uint32_t * op1, vuint16mf4_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_i32m1))) +void vsoxseg4ei16(int32_t * op0, vuint16mf2_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_i32m1_m))) +void vsoxseg4ei16(vbool32_t op0, int32_t * op1, vuint16mf2_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_i32m2))) +void vsoxseg4ei16(int32_t * op0, vuint16m1_t op1, vint32m2_t op2, vint32m2_t op3, vint32m2_t op4, vint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_i32m2_m))) +void vsoxseg4ei16(vbool16_t op0, int32_t * op1, vuint16m1_t op2, vint32m2_t op3, vint32m2_t op4, vint32m2_t op5, vint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_i32mf2))) +void vsoxseg4ei16(int32_t * op0, vuint16mf4_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_i32mf2_m))) +void vsoxseg4ei16(vbool64_t op0, int32_t * op1, vuint16mf4_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_u32m1))) +void vsoxseg4ei16(uint32_t * op0, vuint16mf2_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_u32m1_m))) +void vsoxseg4ei16(vbool32_t op0, uint32_t * op1, vuint16mf2_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_u32m2))) +void vsoxseg4ei16(uint32_t * op0, vuint16m1_t op1, vuint32m2_t op2, vuint32m2_t op3, vuint32m2_t op4, vuint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_u32m2_m))) +void vsoxseg4ei16(vbool16_t op0, uint32_t * op1, vuint16m1_t op2, vuint32m2_t op3, vuint32m2_t op4, vuint32m2_t op5, vuint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_u32mf2))) +void vsoxseg4ei16(uint32_t * op0, vuint16mf4_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_u32mf2_m))) +void vsoxseg4ei16(vbool64_t op0, uint32_t * op1, vuint16mf4_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_i32m1))) +void vsoxseg5ei16(int32_t * op0, vuint16mf2_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_i32m1_m))) +void vsoxseg5ei16(vbool32_t op0, int32_t * op1, vuint16mf2_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_i32mf2))) +void vsoxseg5ei16(int32_t * op0, vuint16mf4_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_i32mf2_m))) +void vsoxseg5ei16(vbool64_t op0, int32_t * op1, vuint16mf4_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_u32m1))) +void vsoxseg5ei16(uint32_t * op0, vuint16mf2_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_u32m1_m))) +void vsoxseg5ei16(vbool32_t op0, uint32_t * op1, vuint16mf2_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_u32mf2))) +void vsoxseg5ei16(uint32_t * op0, vuint16mf4_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_u32mf2_m))) +void vsoxseg5ei16(vbool64_t op0, uint32_t * op1, vuint16mf4_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_i32m1))) +void vsoxseg6ei16(int32_t * op0, vuint16mf2_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_i32m1_m))) +void vsoxseg6ei16(vbool32_t op0, int32_t * op1, vuint16mf2_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_i32mf2))) +void vsoxseg6ei16(int32_t * op0, vuint16mf4_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_i32mf2_m))) +void vsoxseg6ei16(vbool64_t op0, int32_t * op1, vuint16mf4_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_u32m1))) +void vsoxseg6ei16(uint32_t * op0, vuint16mf2_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_u32m1_m))) +void vsoxseg6ei16(vbool32_t op0, uint32_t * op1, vuint16mf2_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_u32mf2))) +void vsoxseg6ei16(uint32_t * op0, vuint16mf4_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_u32mf2_m))) +void vsoxseg6ei16(vbool64_t op0, uint32_t * op1, vuint16mf4_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_i32m1))) +void vsoxseg7ei16(int32_t * op0, vuint16mf2_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_i32m1_m))) +void vsoxseg7ei16(vbool32_t op0, int32_t * op1, vuint16mf2_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_i32mf2))) +void vsoxseg7ei16(int32_t * op0, vuint16mf4_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_i32mf2_m))) +void vsoxseg7ei16(vbool64_t op0, int32_t * op1, vuint16mf4_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_u32m1))) +void vsoxseg7ei16(uint32_t * op0, vuint16mf2_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_u32m1_m))) +void vsoxseg7ei16(vbool32_t op0, uint32_t * op1, vuint16mf2_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_u32mf2))) +void vsoxseg7ei16(uint32_t * op0, vuint16mf4_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_u32mf2_m))) +void vsoxseg7ei16(vbool64_t op0, uint32_t * op1, vuint16mf4_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_i32m1))) +void vsoxseg8ei16(int32_t * op0, vuint16mf2_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_i32m1_m))) +void vsoxseg8ei16(vbool32_t op0, int32_t * op1, vuint16mf2_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, vint32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_i32mf2))) +void vsoxseg8ei16(int32_t * op0, vuint16mf4_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_i32mf2_m))) +void vsoxseg8ei16(vbool64_t op0, int32_t * op1, vuint16mf4_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, vint32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_u32m1))) +void vsoxseg8ei16(uint32_t * op0, vuint16mf2_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_u32m1_m))) +void vsoxseg8ei16(vbool32_t op0, uint32_t * op1, vuint16mf2_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, vuint32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_u32mf2))) +void vsoxseg8ei16(uint32_t * op0, vuint16mf4_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_u32mf2_m))) +void vsoxseg8ei16(vbool64_t op0, uint32_t * op1, vuint16mf4_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, vuint32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_i32m1))) +void vsoxseg2ei32(int32_t * op0, vuint32m1_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_i32m1_m))) +void vsoxseg2ei32(vbool32_t op0, int32_t * op1, vuint32m1_t op2, vint32m1_t op3, vint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_i32m2))) +void vsoxseg2ei32(int32_t * op0, vuint32m2_t op1, vint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_i32m2_m))) +void vsoxseg2ei32(vbool16_t op0, int32_t * op1, vuint32m2_t op2, vint32m2_t op3, vint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_i32m4))) +void vsoxseg2ei32(int32_t * op0, vuint32m4_t op1, vint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_i32m4_m))) +void vsoxseg2ei32(vbool8_t op0, int32_t * op1, vuint32m4_t op2, vint32m4_t op3, vint32m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_i32mf2))) +void vsoxseg2ei32(int32_t * op0, vuint32mf2_t op1, vint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_i32mf2_m))) +void vsoxseg2ei32(vbool64_t op0, int32_t * op1, vuint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_u32m1))) +void vsoxseg2ei32(uint32_t * op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_u32m1_m))) +void vsoxseg2ei32(vbool32_t op0, uint32_t * op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_u32m2))) +void vsoxseg2ei32(uint32_t * op0, vuint32m2_t op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_u32m2_m))) +void vsoxseg2ei32(vbool16_t op0, uint32_t * op1, vuint32m2_t op2, vuint32m2_t op3, vuint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_u32m4))) +void vsoxseg2ei32(uint32_t * op0, vuint32m4_t op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_u32m4_m))) +void vsoxseg2ei32(vbool8_t op0, uint32_t * op1, vuint32m4_t op2, vuint32m4_t op3, vuint32m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_u32mf2))) +void vsoxseg2ei32(uint32_t * op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_u32mf2_m))) +void vsoxseg2ei32(vbool64_t op0, uint32_t * op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_i32m1))) +void vsoxseg3ei32(int32_t * op0, vuint32m1_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_i32m1_m))) +void vsoxseg3ei32(vbool32_t op0, int32_t * op1, vuint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_i32m2))) +void vsoxseg3ei32(int32_t * op0, vuint32m2_t op1, vint32m2_t op2, vint32m2_t op3, vint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_i32m2_m))) +void vsoxseg3ei32(vbool16_t op0, int32_t * op1, vuint32m2_t op2, vint32m2_t op3, vint32m2_t op4, vint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_i32mf2))) +void vsoxseg3ei32(int32_t * op0, vuint32mf2_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_i32mf2_m))) +void vsoxseg3ei32(vbool64_t op0, int32_t * op1, vuint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_u32m1))) +void vsoxseg3ei32(uint32_t * op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_u32m1_m))) +void vsoxseg3ei32(vbool32_t op0, uint32_t * op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_u32m2))) +void vsoxseg3ei32(uint32_t * op0, vuint32m2_t op1, vuint32m2_t op2, vuint32m2_t op3, vuint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_u32m2_m))) +void vsoxseg3ei32(vbool16_t op0, uint32_t * op1, vuint32m2_t op2, vuint32m2_t op3, vuint32m2_t op4, vuint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_u32mf2))) +void vsoxseg3ei32(uint32_t * op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_u32mf2_m))) +void vsoxseg3ei32(vbool64_t op0, uint32_t * op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_i32m1))) +void vsoxseg4ei32(int32_t * op0, vuint32m1_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_i32m1_m))) +void vsoxseg4ei32(vbool32_t op0, int32_t * op1, vuint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_i32m2))) +void vsoxseg4ei32(int32_t * op0, vuint32m2_t op1, vint32m2_t op2, vint32m2_t op3, vint32m2_t op4, vint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_i32m2_m))) +void vsoxseg4ei32(vbool16_t op0, int32_t * op1, vuint32m2_t op2, vint32m2_t op3, vint32m2_t op4, vint32m2_t op5, vint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_i32mf2))) +void vsoxseg4ei32(int32_t * op0, vuint32mf2_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_i32mf2_m))) +void vsoxseg4ei32(vbool64_t op0, int32_t * op1, vuint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_u32m1))) +void vsoxseg4ei32(uint32_t * op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_u32m1_m))) +void vsoxseg4ei32(vbool32_t op0, uint32_t * op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_u32m2))) +void vsoxseg4ei32(uint32_t * op0, vuint32m2_t op1, vuint32m2_t op2, vuint32m2_t op3, vuint32m2_t op4, vuint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_u32m2_m))) +void vsoxseg4ei32(vbool16_t op0, uint32_t * op1, vuint32m2_t op2, vuint32m2_t op3, vuint32m2_t op4, vuint32m2_t op5, vuint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_u32mf2))) +void vsoxseg4ei32(uint32_t * op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_u32mf2_m))) +void vsoxseg4ei32(vbool64_t op0, uint32_t * op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_i32m1))) +void vsoxseg5ei32(int32_t * op0, vuint32m1_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_i32m1_m))) +void vsoxseg5ei32(vbool32_t op0, int32_t * op1, vuint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_i32mf2))) +void vsoxseg5ei32(int32_t * op0, vuint32mf2_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_i32mf2_m))) +void vsoxseg5ei32(vbool64_t op0, int32_t * op1, vuint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_u32m1))) +void vsoxseg5ei32(uint32_t * op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_u32m1_m))) +void vsoxseg5ei32(vbool32_t op0, uint32_t * op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_u32mf2))) +void vsoxseg5ei32(uint32_t * op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_u32mf2_m))) +void vsoxseg5ei32(vbool64_t op0, uint32_t * op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_i32m1))) +void vsoxseg6ei32(int32_t * op0, vuint32m1_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_i32m1_m))) +void vsoxseg6ei32(vbool32_t op0, int32_t * op1, vuint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_i32mf2))) +void vsoxseg6ei32(int32_t * op0, vuint32mf2_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_i32mf2_m))) +void vsoxseg6ei32(vbool64_t op0, int32_t * op1, vuint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_u32m1))) +void vsoxseg6ei32(uint32_t * op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_u32m1_m))) +void vsoxseg6ei32(vbool32_t op0, uint32_t * op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_u32mf2))) +void vsoxseg6ei32(uint32_t * op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_u32mf2_m))) +void vsoxseg6ei32(vbool64_t op0, uint32_t * op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_i32m1))) +void vsoxseg7ei32(int32_t * op0, vuint32m1_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_i32m1_m))) +void vsoxseg7ei32(vbool32_t op0, int32_t * op1, vuint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_i32mf2))) +void vsoxseg7ei32(int32_t * op0, vuint32mf2_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_i32mf2_m))) +void vsoxseg7ei32(vbool64_t op0, int32_t * op1, vuint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_u32m1))) +void vsoxseg7ei32(uint32_t * op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_u32m1_m))) +void vsoxseg7ei32(vbool32_t op0, uint32_t * op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_u32mf2))) +void vsoxseg7ei32(uint32_t * op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_u32mf2_m))) +void vsoxseg7ei32(vbool64_t op0, uint32_t * op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_i32m1))) +void vsoxseg8ei32(int32_t * op0, vuint32m1_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_i32m1_m))) +void vsoxseg8ei32(vbool32_t op0, int32_t * op1, vuint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, vint32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_i32mf2))) +void vsoxseg8ei32(int32_t * op0, vuint32mf2_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_i32mf2_m))) +void vsoxseg8ei32(vbool64_t op0, int32_t * op1, vuint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, vint32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_u32m1))) +void vsoxseg8ei32(uint32_t * op0, vuint32m1_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_u32m1_m))) +void vsoxseg8ei32(vbool32_t op0, uint32_t * op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, vuint32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_u32mf2))) +void vsoxseg8ei32(uint32_t * op0, vuint32mf2_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_u32mf2_m))) +void vsoxseg8ei32(vbool64_t op0, uint32_t * op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, vuint32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_i32m1))) +void vsoxseg2ei64(int32_t * op0, vuint64m2_t op1, vint32m1_t op2, vint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_i32m1_m))) +void vsoxseg2ei64(vbool32_t op0, int32_t * op1, vuint64m2_t op2, vint32m1_t op3, vint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_i32m2))) +void vsoxseg2ei64(int32_t * op0, vuint64m4_t op1, vint32m2_t op2, vint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_i32m2_m))) +void vsoxseg2ei64(vbool16_t op0, int32_t * op1, vuint64m4_t op2, vint32m2_t op3, vint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_i32m4))) +void vsoxseg2ei64(int32_t * op0, vuint64m8_t op1, vint32m4_t op2, vint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_i32m4_m))) +void vsoxseg2ei64(vbool8_t op0, int32_t * op1, vuint64m8_t op2, vint32m4_t op3, vint32m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_i32mf2))) +void vsoxseg2ei64(int32_t * op0, vuint64m1_t op1, vint32mf2_t op2, vint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_i32mf2_m))) +void vsoxseg2ei64(vbool64_t op0, int32_t * op1, vuint64m1_t op2, vint32mf2_t op3, vint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_u32m1))) +void vsoxseg2ei64(uint32_t * op0, vuint64m2_t op1, vuint32m1_t op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_u32m1_m))) +void vsoxseg2ei64(vbool32_t op0, uint32_t * op1, vuint64m2_t op2, vuint32m1_t op3, vuint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_u32m2))) +void vsoxseg2ei64(uint32_t * op0, vuint64m4_t op1, vuint32m2_t op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_u32m2_m))) +void vsoxseg2ei64(vbool16_t op0, uint32_t * op1, vuint64m4_t op2, vuint32m2_t op3, vuint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_u32m4))) +void vsoxseg2ei64(uint32_t * op0, vuint64m8_t op1, vuint32m4_t op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_u32m4_m))) +void vsoxseg2ei64(vbool8_t op0, uint32_t * op1, vuint64m8_t op2, vuint32m4_t op3, vuint32m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_u32mf2))) +void vsoxseg2ei64(uint32_t * op0, vuint64m1_t op1, vuint32mf2_t op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_u32mf2_m))) +void vsoxseg2ei64(vbool64_t op0, uint32_t * op1, vuint64m1_t op2, vuint32mf2_t op3, vuint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_i32m1))) +void vsoxseg3ei64(int32_t * op0, vuint64m2_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_i32m1_m))) +void vsoxseg3ei64(vbool32_t op0, int32_t * op1, vuint64m2_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_i32m2))) +void vsoxseg3ei64(int32_t * op0, vuint64m4_t op1, vint32m2_t op2, vint32m2_t op3, vint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_i32m2_m))) +void vsoxseg3ei64(vbool16_t op0, int32_t * op1, vuint64m4_t op2, vint32m2_t op3, vint32m2_t op4, vint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_i32mf2))) +void vsoxseg3ei64(int32_t * op0, vuint64m1_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_i32mf2_m))) +void vsoxseg3ei64(vbool64_t op0, int32_t * op1, vuint64m1_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_u32m1))) +void vsoxseg3ei64(uint32_t * op0, vuint64m2_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_u32m1_m))) +void vsoxseg3ei64(vbool32_t op0, uint32_t * op1, vuint64m2_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_u32m2))) +void vsoxseg3ei64(uint32_t * op0, vuint64m4_t op1, vuint32m2_t op2, vuint32m2_t op3, vuint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_u32m2_m))) +void vsoxseg3ei64(vbool16_t op0, uint32_t * op1, vuint64m4_t op2, vuint32m2_t op3, vuint32m2_t op4, vuint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_u32mf2))) +void vsoxseg3ei64(uint32_t * op0, vuint64m1_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_u32mf2_m))) +void vsoxseg3ei64(vbool64_t op0, uint32_t * op1, vuint64m1_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_i32m1))) +void vsoxseg4ei64(int32_t * op0, vuint64m2_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_i32m1_m))) +void vsoxseg4ei64(vbool32_t op0, int32_t * op1, vuint64m2_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_i32m2))) +void vsoxseg4ei64(int32_t * op0, vuint64m4_t op1, vint32m2_t op2, vint32m2_t op3, vint32m2_t op4, vint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_i32m2_m))) +void vsoxseg4ei64(vbool16_t op0, int32_t * op1, vuint64m4_t op2, vint32m2_t op3, vint32m2_t op4, vint32m2_t op5, vint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_i32mf2))) +void vsoxseg4ei64(int32_t * op0, vuint64m1_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_i32mf2_m))) +void vsoxseg4ei64(vbool64_t op0, int32_t * op1, vuint64m1_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_u32m1))) +void vsoxseg4ei64(uint32_t * op0, vuint64m2_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_u32m1_m))) +void vsoxseg4ei64(vbool32_t op0, uint32_t * op1, vuint64m2_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_u32m2))) +void vsoxseg4ei64(uint32_t * op0, vuint64m4_t op1, vuint32m2_t op2, vuint32m2_t op3, vuint32m2_t op4, vuint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_u32m2_m))) +void vsoxseg4ei64(vbool16_t op0, uint32_t * op1, vuint64m4_t op2, vuint32m2_t op3, vuint32m2_t op4, vuint32m2_t op5, vuint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_u32mf2))) +void vsoxseg4ei64(uint32_t * op0, vuint64m1_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_u32mf2_m))) +void vsoxseg4ei64(vbool64_t op0, uint32_t * op1, vuint64m1_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_i32m1))) +void vsoxseg5ei64(int32_t * op0, vuint64m2_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_i32m1_m))) +void vsoxseg5ei64(vbool32_t op0, int32_t * op1, vuint64m2_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_i32mf2))) +void vsoxseg5ei64(int32_t * op0, vuint64m1_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_i32mf2_m))) +void vsoxseg5ei64(vbool64_t op0, int32_t * op1, vuint64m1_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_u32m1))) +void vsoxseg5ei64(uint32_t * op0, vuint64m2_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_u32m1_m))) +void vsoxseg5ei64(vbool32_t op0, uint32_t * op1, vuint64m2_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_u32mf2))) +void vsoxseg5ei64(uint32_t * op0, vuint64m1_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_u32mf2_m))) +void vsoxseg5ei64(vbool64_t op0, uint32_t * op1, vuint64m1_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_i32m1))) +void vsoxseg6ei64(int32_t * op0, vuint64m2_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_i32m1_m))) +void vsoxseg6ei64(vbool32_t op0, int32_t * op1, vuint64m2_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_i32mf2))) +void vsoxseg6ei64(int32_t * op0, vuint64m1_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_i32mf2_m))) +void vsoxseg6ei64(vbool64_t op0, int32_t * op1, vuint64m1_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_u32m1))) +void vsoxseg6ei64(uint32_t * op0, vuint64m2_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_u32m1_m))) +void vsoxseg6ei64(vbool32_t op0, uint32_t * op1, vuint64m2_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_u32mf2))) +void vsoxseg6ei64(uint32_t * op0, vuint64m1_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_u32mf2_m))) +void vsoxseg6ei64(vbool64_t op0, uint32_t * op1, vuint64m1_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_i32m1))) +void vsoxseg7ei64(int32_t * op0, vuint64m2_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_i32m1_m))) +void vsoxseg7ei64(vbool32_t op0, int32_t * op1, vuint64m2_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_i32mf2))) +void vsoxseg7ei64(int32_t * op0, vuint64m1_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_i32mf2_m))) +void vsoxseg7ei64(vbool64_t op0, int32_t * op1, vuint64m1_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_u32m1))) +void vsoxseg7ei64(uint32_t * op0, vuint64m2_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_u32m1_m))) +void vsoxseg7ei64(vbool32_t op0, uint32_t * op1, vuint64m2_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_u32mf2))) +void vsoxseg7ei64(uint32_t * op0, vuint64m1_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_u32mf2_m))) +void vsoxseg7ei64(vbool64_t op0, uint32_t * op1, vuint64m1_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_i32m1))) +void vsoxseg8ei64(int32_t * op0, vuint64m2_t op1, vint32m1_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_i32m1_m))) +void vsoxseg8ei64(vbool32_t op0, int32_t * op1, vuint64m2_t op2, vint32m1_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, vint32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_i32mf2))) +void vsoxseg8ei64(int32_t * op0, vuint64m1_t op1, vint32mf2_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_i32mf2_m))) +void vsoxseg8ei64(vbool64_t op0, int32_t * op1, vuint64m1_t op2, vint32mf2_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, vint32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_u32m1))) +void vsoxseg8ei64(uint32_t * op0, vuint64m2_t op1, vuint32m1_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_u32m1_m))) +void vsoxseg8ei64(vbool32_t op0, uint32_t * op1, vuint64m2_t op2, vuint32m1_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, vuint32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_u32mf2))) +void vsoxseg8ei64(uint32_t * op0, vuint64m1_t op1, vuint32mf2_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_u32mf2_m))) +void vsoxseg8ei64(vbool64_t op0, uint32_t * op1, vuint64m1_t op2, vuint32mf2_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, vuint32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_i64m1))) +void vsoxseg2ei8(int64_t * op0, vuint8mf8_t op1, vint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_i64m1_m))) +void vsoxseg2ei8(vbool64_t op0, int64_t * op1, vuint8mf8_t op2, vint64m1_t op3, vint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_i64m2))) +void vsoxseg2ei8(int64_t * op0, vuint8mf4_t op1, vint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_i64m2_m))) +void vsoxseg2ei8(vbool32_t op0, int64_t * op1, vuint8mf4_t op2, vint64m2_t op3, vint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_i64m4))) +void vsoxseg2ei8(int64_t * op0, vuint8mf2_t op1, vint64m4_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_i64m4_m))) +void vsoxseg2ei8(vbool16_t op0, int64_t * op1, vuint8mf2_t op2, vint64m4_t op3, vint64m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_u64m1))) +void vsoxseg2ei8(uint64_t * op0, vuint8mf8_t op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_u64m1_m))) +void vsoxseg2ei8(vbool64_t op0, uint64_t * op1, vuint8mf8_t op2, vuint64m1_t op3, vuint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_u64m2))) +void vsoxseg2ei8(uint64_t * op0, vuint8mf4_t op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_u64m2_m))) +void vsoxseg2ei8(vbool32_t op0, uint64_t * op1, vuint8mf4_t op2, vuint64m2_t op3, vuint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_u64m4))) +void vsoxseg2ei8(uint64_t * op0, vuint8mf2_t op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_u64m4_m))) +void vsoxseg2ei8(vbool16_t op0, uint64_t * op1, vuint8mf2_t op2, vuint64m4_t op3, vuint64m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_i64m1))) +void vsoxseg3ei8(int64_t * op0, vuint8mf8_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_i64m1_m))) +void vsoxseg3ei8(vbool64_t op0, int64_t * op1, vuint8mf8_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_i64m2))) +void vsoxseg3ei8(int64_t * op0, vuint8mf4_t op1, vint64m2_t op2, vint64m2_t op3, vint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_i64m2_m))) +void vsoxseg3ei8(vbool32_t op0, int64_t * op1, vuint8mf4_t op2, vint64m2_t op3, vint64m2_t op4, vint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_u64m1))) +void vsoxseg3ei8(uint64_t * op0, vuint8mf8_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_u64m1_m))) +void vsoxseg3ei8(vbool64_t op0, uint64_t * op1, vuint8mf8_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_u64m2))) +void vsoxseg3ei8(uint64_t * op0, vuint8mf4_t op1, vuint64m2_t op2, vuint64m2_t op3, vuint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_u64m2_m))) +void vsoxseg3ei8(vbool32_t op0, uint64_t * op1, vuint8mf4_t op2, vuint64m2_t op3, vuint64m2_t op4, vuint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_i64m1))) +void vsoxseg4ei8(int64_t * op0, vuint8mf8_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_i64m1_m))) +void vsoxseg4ei8(vbool64_t op0, int64_t * op1, vuint8mf8_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_i64m2))) +void vsoxseg4ei8(int64_t * op0, vuint8mf4_t op1, vint64m2_t op2, vint64m2_t op3, vint64m2_t op4, vint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_i64m2_m))) +void vsoxseg4ei8(vbool32_t op0, int64_t * op1, vuint8mf4_t op2, vint64m2_t op3, vint64m2_t op4, vint64m2_t op5, vint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_u64m1))) +void vsoxseg4ei8(uint64_t * op0, vuint8mf8_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_u64m1_m))) +void vsoxseg4ei8(vbool64_t op0, uint64_t * op1, vuint8mf8_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_u64m2))) +void vsoxseg4ei8(uint64_t * op0, vuint8mf4_t op1, vuint64m2_t op2, vuint64m2_t op3, vuint64m2_t op4, vuint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_u64m2_m))) +void vsoxseg4ei8(vbool32_t op0, uint64_t * op1, vuint8mf4_t op2, vuint64m2_t op3, vuint64m2_t op4, vuint64m2_t op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_i64m1))) +void vsoxseg5ei8(int64_t * op0, vuint8mf8_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_i64m1_m))) +void vsoxseg5ei8(vbool64_t op0, int64_t * op1, vuint8mf8_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_u64m1))) +void vsoxseg5ei8(uint64_t * op0, vuint8mf8_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_u64m1_m))) +void vsoxseg5ei8(vbool64_t op0, uint64_t * op1, vuint8mf8_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_i64m1))) +void vsoxseg6ei8(int64_t * op0, vuint8mf8_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_i64m1_m))) +void vsoxseg6ei8(vbool64_t op0, int64_t * op1, vuint8mf8_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_u64m1))) +void vsoxseg6ei8(uint64_t * op0, vuint8mf8_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_u64m1_m))) +void vsoxseg6ei8(vbool64_t op0, uint64_t * op1, vuint8mf8_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_i64m1))) +void vsoxseg7ei8(int64_t * op0, vuint8mf8_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_i64m1_m))) +void vsoxseg7ei8(vbool64_t op0, int64_t * op1, vuint8mf8_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_u64m1))) +void vsoxseg7ei8(uint64_t * op0, vuint8mf8_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_u64m1_m))) +void vsoxseg7ei8(vbool64_t op0, uint64_t * op1, vuint8mf8_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_i64m1))) +void vsoxseg8ei8(int64_t * op0, vuint8mf8_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_i64m1_m))) +void vsoxseg8ei8(vbool64_t op0, int64_t * op1, vuint8mf8_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, vint64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_u64m1))) +void vsoxseg8ei8(uint64_t * op0, vuint8mf8_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_u64m1_m))) +void vsoxseg8ei8(vbool64_t op0, uint64_t * op1, vuint8mf8_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, vuint64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_i64m1))) +void vsoxseg2ei16(int64_t * op0, vuint16mf4_t op1, vint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_i64m1_m))) +void vsoxseg2ei16(vbool64_t op0, int64_t * op1, vuint16mf4_t op2, vint64m1_t op3, vint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_i64m2))) +void vsoxseg2ei16(int64_t * op0, vuint16mf2_t op1, vint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_i64m2_m))) +void vsoxseg2ei16(vbool32_t op0, int64_t * op1, vuint16mf2_t op2, vint64m2_t op3, vint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_i64m4))) +void vsoxseg2ei16(int64_t * op0, vuint16m1_t op1, vint64m4_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_i64m4_m))) +void vsoxseg2ei16(vbool16_t op0, int64_t * op1, vuint16m1_t op2, vint64m4_t op3, vint64m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_u64m1))) +void vsoxseg2ei16(uint64_t * op0, vuint16mf4_t op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_u64m1_m))) +void vsoxseg2ei16(vbool64_t op0, uint64_t * op1, vuint16mf4_t op2, vuint64m1_t op3, vuint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_u64m2))) +void vsoxseg2ei16(uint64_t * op0, vuint16mf2_t op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_u64m2_m))) +void vsoxseg2ei16(vbool32_t op0, uint64_t * op1, vuint16mf2_t op2, vuint64m2_t op3, vuint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_u64m4))) +void vsoxseg2ei16(uint64_t * op0, vuint16m1_t op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_u64m4_m))) +void vsoxseg2ei16(vbool16_t op0, uint64_t * op1, vuint16m1_t op2, vuint64m4_t op3, vuint64m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_i64m1))) +void vsoxseg3ei16(int64_t * op0, vuint16mf4_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_i64m1_m))) +void vsoxseg3ei16(vbool64_t op0, int64_t * op1, vuint16mf4_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_i64m2))) +void vsoxseg3ei16(int64_t * op0, vuint16mf2_t op1, vint64m2_t op2, vint64m2_t op3, vint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_i64m2_m))) +void vsoxseg3ei16(vbool32_t op0, int64_t * op1, vuint16mf2_t op2, vint64m2_t op3, vint64m2_t op4, vint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_u64m1))) +void vsoxseg3ei16(uint64_t * op0, vuint16mf4_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_u64m1_m))) +void vsoxseg3ei16(vbool64_t op0, uint64_t * op1, vuint16mf4_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_u64m2))) +void vsoxseg3ei16(uint64_t * op0, vuint16mf2_t op1, vuint64m2_t op2, vuint64m2_t op3, vuint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_u64m2_m))) +void vsoxseg3ei16(vbool32_t op0, uint64_t * op1, vuint16mf2_t op2, vuint64m2_t op3, vuint64m2_t op4, vuint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_i64m1))) +void vsoxseg4ei16(int64_t * op0, vuint16mf4_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_i64m1_m))) +void vsoxseg4ei16(vbool64_t op0, int64_t * op1, vuint16mf4_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_i64m2))) +void vsoxseg4ei16(int64_t * op0, vuint16mf2_t op1, vint64m2_t op2, vint64m2_t op3, vint64m2_t op4, vint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_i64m2_m))) +void vsoxseg4ei16(vbool32_t op0, int64_t * op1, vuint16mf2_t op2, vint64m2_t op3, vint64m2_t op4, vint64m2_t op5, vint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_u64m1))) +void vsoxseg4ei16(uint64_t * op0, vuint16mf4_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_u64m1_m))) +void vsoxseg4ei16(vbool64_t op0, uint64_t * op1, vuint16mf4_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_u64m2))) +void vsoxseg4ei16(uint64_t * op0, vuint16mf2_t op1, vuint64m2_t op2, vuint64m2_t op3, vuint64m2_t op4, vuint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_u64m2_m))) +void vsoxseg4ei16(vbool32_t op0, uint64_t * op1, vuint16mf2_t op2, vuint64m2_t op3, vuint64m2_t op4, vuint64m2_t op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_i64m1))) +void vsoxseg5ei16(int64_t * op0, vuint16mf4_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_i64m1_m))) +void vsoxseg5ei16(vbool64_t op0, int64_t * op1, vuint16mf4_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_u64m1))) +void vsoxseg5ei16(uint64_t * op0, vuint16mf4_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_u64m1_m))) +void vsoxseg5ei16(vbool64_t op0, uint64_t * op1, vuint16mf4_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_i64m1))) +void vsoxseg6ei16(int64_t * op0, vuint16mf4_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_i64m1_m))) +void vsoxseg6ei16(vbool64_t op0, int64_t * op1, vuint16mf4_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_u64m1))) +void vsoxseg6ei16(uint64_t * op0, vuint16mf4_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_u64m1_m))) +void vsoxseg6ei16(vbool64_t op0, uint64_t * op1, vuint16mf4_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_i64m1))) +void vsoxseg7ei16(int64_t * op0, vuint16mf4_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_i64m1_m))) +void vsoxseg7ei16(vbool64_t op0, int64_t * op1, vuint16mf4_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_u64m1))) +void vsoxseg7ei16(uint64_t * op0, vuint16mf4_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_u64m1_m))) +void vsoxseg7ei16(vbool64_t op0, uint64_t * op1, vuint16mf4_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_i64m1))) +void vsoxseg8ei16(int64_t * op0, vuint16mf4_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_i64m1_m))) +void vsoxseg8ei16(vbool64_t op0, int64_t * op1, vuint16mf4_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, vint64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_u64m1))) +void vsoxseg8ei16(uint64_t * op0, vuint16mf4_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_u64m1_m))) +void vsoxseg8ei16(vbool64_t op0, uint64_t * op1, vuint16mf4_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, vuint64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_i64m1))) +void vsoxseg2ei32(int64_t * op0, vuint32mf2_t op1, vint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_i64m1_m))) +void vsoxseg2ei32(vbool64_t op0, int64_t * op1, vuint32mf2_t op2, vint64m1_t op3, vint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_i64m2))) +void vsoxseg2ei32(int64_t * op0, vuint32m1_t op1, vint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_i64m2_m))) +void vsoxseg2ei32(vbool32_t op0, int64_t * op1, vuint32m1_t op2, vint64m2_t op3, vint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_i64m4))) +void vsoxseg2ei32(int64_t * op0, vuint32m2_t op1, vint64m4_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_i64m4_m))) +void vsoxseg2ei32(vbool16_t op0, int64_t * op1, vuint32m2_t op2, vint64m4_t op3, vint64m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_u64m1))) +void vsoxseg2ei32(uint64_t * op0, vuint32mf2_t op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_u64m1_m))) +void vsoxseg2ei32(vbool64_t op0, uint64_t * op1, vuint32mf2_t op2, vuint64m1_t op3, vuint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_u64m2))) +void vsoxseg2ei32(uint64_t * op0, vuint32m1_t op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_u64m2_m))) +void vsoxseg2ei32(vbool32_t op0, uint64_t * op1, vuint32m1_t op2, vuint64m2_t op3, vuint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_u64m4))) +void vsoxseg2ei32(uint64_t * op0, vuint32m2_t op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_u64m4_m))) +void vsoxseg2ei32(vbool16_t op0, uint64_t * op1, vuint32m2_t op2, vuint64m4_t op3, vuint64m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_i64m1))) +void vsoxseg3ei32(int64_t * op0, vuint32mf2_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_i64m1_m))) +void vsoxseg3ei32(vbool64_t op0, int64_t * op1, vuint32mf2_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_i64m2))) +void vsoxseg3ei32(int64_t * op0, vuint32m1_t op1, vint64m2_t op2, vint64m2_t op3, vint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_i64m2_m))) +void vsoxseg3ei32(vbool32_t op0, int64_t * op1, vuint32m1_t op2, vint64m2_t op3, vint64m2_t op4, vint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_u64m1))) +void vsoxseg3ei32(uint64_t * op0, vuint32mf2_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_u64m1_m))) +void vsoxseg3ei32(vbool64_t op0, uint64_t * op1, vuint32mf2_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_u64m2))) +void vsoxseg3ei32(uint64_t * op0, vuint32m1_t op1, vuint64m2_t op2, vuint64m2_t op3, vuint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_u64m2_m))) +void vsoxseg3ei32(vbool32_t op0, uint64_t * op1, vuint32m1_t op2, vuint64m2_t op3, vuint64m2_t op4, vuint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_i64m1))) +void vsoxseg4ei32(int64_t * op0, vuint32mf2_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_i64m1_m))) +void vsoxseg4ei32(vbool64_t op0, int64_t * op1, vuint32mf2_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_i64m2))) +void vsoxseg4ei32(int64_t * op0, vuint32m1_t op1, vint64m2_t op2, vint64m2_t op3, vint64m2_t op4, vint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_i64m2_m))) +void vsoxseg4ei32(vbool32_t op0, int64_t * op1, vuint32m1_t op2, vint64m2_t op3, vint64m2_t op4, vint64m2_t op5, vint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_u64m1))) +void vsoxseg4ei32(uint64_t * op0, vuint32mf2_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_u64m1_m))) +void vsoxseg4ei32(vbool64_t op0, uint64_t * op1, vuint32mf2_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_u64m2))) +void vsoxseg4ei32(uint64_t * op0, vuint32m1_t op1, vuint64m2_t op2, vuint64m2_t op3, vuint64m2_t op4, vuint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_u64m2_m))) +void vsoxseg4ei32(vbool32_t op0, uint64_t * op1, vuint32m1_t op2, vuint64m2_t op3, vuint64m2_t op4, vuint64m2_t op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_i64m1))) +void vsoxseg5ei32(int64_t * op0, vuint32mf2_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_i64m1_m))) +void vsoxseg5ei32(vbool64_t op0, int64_t * op1, vuint32mf2_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_u64m1))) +void vsoxseg5ei32(uint64_t * op0, vuint32mf2_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_u64m1_m))) +void vsoxseg5ei32(vbool64_t op0, uint64_t * op1, vuint32mf2_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_i64m1))) +void vsoxseg6ei32(int64_t * op0, vuint32mf2_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_i64m1_m))) +void vsoxseg6ei32(vbool64_t op0, int64_t * op1, vuint32mf2_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_u64m1))) +void vsoxseg6ei32(uint64_t * op0, vuint32mf2_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_u64m1_m))) +void vsoxseg6ei32(vbool64_t op0, uint64_t * op1, vuint32mf2_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_i64m1))) +void vsoxseg7ei32(int64_t * op0, vuint32mf2_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_i64m1_m))) +void vsoxseg7ei32(vbool64_t op0, int64_t * op1, vuint32mf2_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_u64m1))) +void vsoxseg7ei32(uint64_t * op0, vuint32mf2_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_u64m1_m))) +void vsoxseg7ei32(vbool64_t op0, uint64_t * op1, vuint32mf2_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_i64m1))) +void vsoxseg8ei32(int64_t * op0, vuint32mf2_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_i64m1_m))) +void vsoxseg8ei32(vbool64_t op0, int64_t * op1, vuint32mf2_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, vint64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_u64m1))) +void vsoxseg8ei32(uint64_t * op0, vuint32mf2_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_u64m1_m))) +void vsoxseg8ei32(vbool64_t op0, uint64_t * op1, vuint32mf2_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, vuint64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_i64m1))) +void vsoxseg2ei64(int64_t * op0, vuint64m1_t op1, vint64m1_t op2, vint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_i64m1_m))) +void vsoxseg2ei64(vbool64_t op0, int64_t * op1, vuint64m1_t op2, vint64m1_t op3, vint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_i64m2))) +void vsoxseg2ei64(int64_t * op0, vuint64m2_t op1, vint64m2_t op2, vint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_i64m2_m))) +void vsoxseg2ei64(vbool32_t op0, int64_t * op1, vuint64m2_t op2, vint64m2_t op3, vint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_i64m4))) +void vsoxseg2ei64(int64_t * op0, vuint64m4_t op1, vint64m4_t op2, vint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_i64m4_m))) +void vsoxseg2ei64(vbool16_t op0, int64_t * op1, vuint64m4_t op2, vint64m4_t op3, vint64m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_u64m1))) +void vsoxseg2ei64(uint64_t * op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_u64m1_m))) +void vsoxseg2ei64(vbool64_t op0, uint64_t * op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_u64m2))) +void vsoxseg2ei64(uint64_t * op0, vuint64m2_t op1, vuint64m2_t op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_u64m2_m))) +void vsoxseg2ei64(vbool32_t op0, uint64_t * op1, vuint64m2_t op2, vuint64m2_t op3, vuint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_u64m4))) +void vsoxseg2ei64(uint64_t * op0, vuint64m4_t op1, vuint64m4_t op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_u64m4_m))) +void vsoxseg2ei64(vbool16_t op0, uint64_t * op1, vuint64m4_t op2, vuint64m4_t op3, vuint64m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_i64m1))) +void vsoxseg3ei64(int64_t * op0, vuint64m1_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_i64m1_m))) +void vsoxseg3ei64(vbool64_t op0, int64_t * op1, vuint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_i64m2))) +void vsoxseg3ei64(int64_t * op0, vuint64m2_t op1, vint64m2_t op2, vint64m2_t op3, vint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_i64m2_m))) +void vsoxseg3ei64(vbool32_t op0, int64_t * op1, vuint64m2_t op2, vint64m2_t op3, vint64m2_t op4, vint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_u64m1))) +void vsoxseg3ei64(uint64_t * op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_u64m1_m))) +void vsoxseg3ei64(vbool64_t op0, uint64_t * op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_u64m2))) +void vsoxseg3ei64(uint64_t * op0, vuint64m2_t op1, vuint64m2_t op2, vuint64m2_t op3, vuint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_u64m2_m))) +void vsoxseg3ei64(vbool32_t op0, uint64_t * op1, vuint64m2_t op2, vuint64m2_t op3, vuint64m2_t op4, vuint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_i64m1))) +void vsoxseg4ei64(int64_t * op0, vuint64m1_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_i64m1_m))) +void vsoxseg4ei64(vbool64_t op0, int64_t * op1, vuint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_i64m2))) +void vsoxseg4ei64(int64_t * op0, vuint64m2_t op1, vint64m2_t op2, vint64m2_t op3, vint64m2_t op4, vint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_i64m2_m))) +void vsoxseg4ei64(vbool32_t op0, int64_t * op1, vuint64m2_t op2, vint64m2_t op3, vint64m2_t op4, vint64m2_t op5, vint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_u64m1))) +void vsoxseg4ei64(uint64_t * op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_u64m1_m))) +void vsoxseg4ei64(vbool64_t op0, uint64_t * op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_u64m2))) +void vsoxseg4ei64(uint64_t * op0, vuint64m2_t op1, vuint64m2_t op2, vuint64m2_t op3, vuint64m2_t op4, vuint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_u64m2_m))) +void vsoxseg4ei64(vbool32_t op0, uint64_t * op1, vuint64m2_t op2, vuint64m2_t op3, vuint64m2_t op4, vuint64m2_t op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_i64m1))) +void vsoxseg5ei64(int64_t * op0, vuint64m1_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_i64m1_m))) +void vsoxseg5ei64(vbool64_t op0, int64_t * op1, vuint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_u64m1))) +void vsoxseg5ei64(uint64_t * op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_u64m1_m))) +void vsoxseg5ei64(vbool64_t op0, uint64_t * op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_i64m1))) +void vsoxseg6ei64(int64_t * op0, vuint64m1_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_i64m1_m))) +void vsoxseg6ei64(vbool64_t op0, int64_t * op1, vuint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_u64m1))) +void vsoxseg6ei64(uint64_t * op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_u64m1_m))) +void vsoxseg6ei64(vbool64_t op0, uint64_t * op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_i64m1))) +void vsoxseg7ei64(int64_t * op0, vuint64m1_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_i64m1_m))) +void vsoxseg7ei64(vbool64_t op0, int64_t * op1, vuint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_u64m1))) +void vsoxseg7ei64(uint64_t * op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_u64m1_m))) +void vsoxseg7ei64(vbool64_t op0, uint64_t * op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_i64m1))) +void vsoxseg8ei64(int64_t * op0, vuint64m1_t op1, vint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_i64m1_m))) +void vsoxseg8ei64(vbool64_t op0, int64_t * op1, vuint64m1_t op2, vint64m1_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, vint64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_u64m1))) +void vsoxseg8ei64(uint64_t * op0, vuint64m1_t op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_u64m1_m))) +void vsoxseg8ei64(vbool64_t op0, uint64_t * op1, vuint64m1_t op2, vuint64m1_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, vuint64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e8_v_i8m1_m))) +void vlseg2e8(vint8m1_t * op0, vint8m1_t * op1, vbool8_t op2, vint8m1_t op3, vint8m1_t op4, const int8_t * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e8_v_i8m2_m))) +void vlseg2e8(vint8m2_t * op0, vint8m2_t * op1, vbool4_t op2, vint8m2_t op3, vint8m2_t op4, const int8_t * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e8_v_i8m4_m))) +void vlseg2e8(vint8m4_t * op0, vint8m4_t * op1, vbool2_t op2, vint8m4_t op3, vint8m4_t op4, const int8_t * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e8_v_i8mf2_m))) +void vlseg2e8(vint8mf2_t * op0, vint8mf2_t * op1, vbool16_t op2, vint8mf2_t op3, vint8mf2_t op4, const int8_t * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e8_v_i8mf4_m))) +void vlseg2e8(vint8mf4_t * op0, vint8mf4_t * op1, vbool32_t op2, vint8mf4_t op3, vint8mf4_t op4, const int8_t * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e8_v_i8mf8_m))) +void vlseg2e8(vint8mf8_t * op0, vint8mf8_t * op1, vbool64_t op2, vint8mf8_t op3, vint8mf8_t op4, const int8_t * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e8_v_u8m1_m))) +void vlseg2e8(vuint8m1_t * op0, vuint8m1_t * op1, vbool8_t op2, vuint8m1_t op3, vuint8m1_t op4, const uint8_t * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e8_v_u8m2_m))) +void vlseg2e8(vuint8m2_t * op0, vuint8m2_t * op1, vbool4_t op2, vuint8m2_t op3, vuint8m2_t op4, const uint8_t * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e8_v_u8m4_m))) +void vlseg2e8(vuint8m4_t * op0, vuint8m4_t * op1, vbool2_t op2, vuint8m4_t op3, vuint8m4_t op4, const uint8_t * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e8_v_u8mf2_m))) +void vlseg2e8(vuint8mf2_t * op0, vuint8mf2_t * op1, vbool16_t op2, vuint8mf2_t op3, vuint8mf2_t op4, const uint8_t * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e8_v_u8mf4_m))) +void vlseg2e8(vuint8mf4_t * op0, vuint8mf4_t * op1, vbool32_t op2, vuint8mf4_t op3, vuint8mf4_t op4, const uint8_t * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e8_v_u8mf8_m))) +void vlseg2e8(vuint8mf8_t * op0, vuint8mf8_t * op1, vbool64_t op2, vuint8mf8_t op3, vuint8mf8_t op4, const uint8_t * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e8_v_i8m1_m))) +void vlseg3e8(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vbool8_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, const int8_t * op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e8_v_i8m2_m))) +void vlseg3e8(vint8m2_t * op0, vint8m2_t * op1, vint8m2_t * op2, vbool4_t op3, vint8m2_t op4, vint8m2_t op5, vint8m2_t op6, const int8_t * op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e8_v_i8mf2_m))) +void vlseg3e8(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vbool16_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, const int8_t * op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e8_v_i8mf4_m))) +void vlseg3e8(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vbool32_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, const int8_t * op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e8_v_i8mf8_m))) +void vlseg3e8(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vbool64_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, const int8_t * op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e8_v_u8m1_m))) +void vlseg3e8(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vbool8_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, const uint8_t * op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e8_v_u8m2_m))) +void vlseg3e8(vuint8m2_t * op0, vuint8m2_t * op1, vuint8m2_t * op2, vbool4_t op3, vuint8m2_t op4, vuint8m2_t op5, vuint8m2_t op6, const uint8_t * op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e8_v_u8mf2_m))) +void vlseg3e8(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vbool16_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, const uint8_t * op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e8_v_u8mf4_m))) +void vlseg3e8(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vbool32_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, const uint8_t * op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e8_v_u8mf8_m))) +void vlseg3e8(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vbool64_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, const uint8_t * op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e8_v_i8m1_m))) +void vlseg4e8(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vbool8_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, const int8_t * op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e8_v_i8m2_m))) +void vlseg4e8(vint8m2_t * op0, vint8m2_t * op1, vint8m2_t * op2, vint8m2_t * op3, vbool4_t op4, vint8m2_t op5, vint8m2_t op6, vint8m2_t op7, vint8m2_t op8, const int8_t * op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e8_v_i8mf2_m))) +void vlseg4e8(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vbool16_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, const int8_t * op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e8_v_i8mf4_m))) +void vlseg4e8(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vbool32_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, const int8_t * op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e8_v_i8mf8_m))) +void vlseg4e8(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vbool64_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, const int8_t * op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e8_v_u8m1_m))) +void vlseg4e8(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vbool8_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, const uint8_t * op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e8_v_u8m2_m))) +void vlseg4e8(vuint8m2_t * op0, vuint8m2_t * op1, vuint8m2_t * op2, vuint8m2_t * op3, vbool4_t op4, vuint8m2_t op5, vuint8m2_t op6, vuint8m2_t op7, vuint8m2_t op8, const uint8_t * op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e8_v_u8mf2_m))) +void vlseg4e8(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vbool16_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, const uint8_t * op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e8_v_u8mf4_m))) +void vlseg4e8(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vbool32_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, const uint8_t * op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e8_v_u8mf8_m))) +void vlseg4e8(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vbool64_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, const uint8_t * op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e8_v_i8m1_m))) +void vlseg5e8(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vbool8_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, vint8m1_t op10, const int8_t * op11, size_t op12); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e8_v_i8mf2_m))) +void vlseg5e8(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vbool16_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, vint8mf2_t op10, const int8_t * op11, size_t op12); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e8_v_i8mf4_m))) +void vlseg5e8(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vbool32_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, vint8mf4_t op10, const int8_t * op11, size_t op12); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e8_v_i8mf8_m))) +void vlseg5e8(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vbool64_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, vint8mf8_t op10, const int8_t * op11, size_t op12); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e8_v_u8m1_m))) +void vlseg5e8(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vbool8_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, vuint8m1_t op10, const uint8_t * op11, size_t op12); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e8_v_u8mf2_m))) +void vlseg5e8(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vbool16_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, vuint8mf2_t op10, const uint8_t * op11, size_t op12); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e8_v_u8mf4_m))) +void vlseg5e8(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vbool32_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, vuint8mf4_t op10, const uint8_t * op11, size_t op12); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e8_v_u8mf8_m))) +void vlseg5e8(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vbool64_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, vuint8mf8_t op10, const uint8_t * op11, size_t op12); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e8_v_i8m1_m))) +void vlseg6e8(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vbool8_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, vint8m1_t op10, vint8m1_t op11, vint8m1_t op12, const int8_t * op13, size_t op14); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e8_v_i8mf2_m))) +void vlseg6e8(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vbool16_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, vint8mf2_t op10, vint8mf2_t op11, vint8mf2_t op12, const int8_t * op13, size_t op14); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e8_v_i8mf4_m))) +void vlseg6e8(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vbool32_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, vint8mf4_t op10, vint8mf4_t op11, vint8mf4_t op12, const int8_t * op13, size_t op14); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e8_v_i8mf8_m))) +void vlseg6e8(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vbool64_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, vint8mf8_t op10, vint8mf8_t op11, vint8mf8_t op12, const int8_t * op13, size_t op14); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e8_v_u8m1_m))) +void vlseg6e8(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vbool8_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, vuint8m1_t op10, vuint8m1_t op11, vuint8m1_t op12, const uint8_t * op13, size_t op14); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e8_v_u8mf2_m))) +void vlseg6e8(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vbool16_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, vuint8mf2_t op10, vuint8mf2_t op11, vuint8mf2_t op12, const uint8_t * op13, size_t op14); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e8_v_u8mf4_m))) +void vlseg6e8(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vbool32_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, vuint8mf4_t op10, vuint8mf4_t op11, vuint8mf4_t op12, const uint8_t * op13, size_t op14); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e8_v_u8mf8_m))) +void vlseg6e8(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vbool64_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, vuint8mf8_t op10, vuint8mf8_t op11, vuint8mf8_t op12, const uint8_t * op13, size_t op14); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e8_v_i8m1_m))) +void vlseg7e8(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vint8m1_t * op6, vbool8_t op7, vint8m1_t op8, vint8m1_t op9, vint8m1_t op10, vint8m1_t op11, vint8m1_t op12, vint8m1_t op13, vint8m1_t op14, const int8_t * op15, size_t op16); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e8_v_i8mf2_m))) +void vlseg7e8(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vint8mf2_t * op6, vbool16_t op7, vint8mf2_t op8, vint8mf2_t op9, vint8mf2_t op10, vint8mf2_t op11, vint8mf2_t op12, vint8mf2_t op13, vint8mf2_t op14, const int8_t * op15, size_t op16); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e8_v_i8mf4_m))) +void vlseg7e8(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vint8mf4_t * op6, vbool32_t op7, vint8mf4_t op8, vint8mf4_t op9, vint8mf4_t op10, vint8mf4_t op11, vint8mf4_t op12, vint8mf4_t op13, vint8mf4_t op14, const int8_t * op15, size_t op16); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e8_v_i8mf8_m))) +void vlseg7e8(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vint8mf8_t * op6, vbool64_t op7, vint8mf8_t op8, vint8mf8_t op9, vint8mf8_t op10, vint8mf8_t op11, vint8mf8_t op12, vint8mf8_t op13, vint8mf8_t op14, const int8_t * op15, size_t op16); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e8_v_u8m1_m))) +void vlseg7e8(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vuint8m1_t * op6, vbool8_t op7, vuint8m1_t op8, vuint8m1_t op9, vuint8m1_t op10, vuint8m1_t op11, vuint8m1_t op12, vuint8m1_t op13, vuint8m1_t op14, const uint8_t * op15, size_t op16); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e8_v_u8mf2_m))) +void vlseg7e8(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vuint8mf2_t * op6, vbool16_t op7, vuint8mf2_t op8, vuint8mf2_t op9, vuint8mf2_t op10, vuint8mf2_t op11, vuint8mf2_t op12, vuint8mf2_t op13, vuint8mf2_t op14, const uint8_t * op15, size_t op16); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e8_v_u8mf4_m))) +void vlseg7e8(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vuint8mf4_t * op6, vbool32_t op7, vuint8mf4_t op8, vuint8mf4_t op9, vuint8mf4_t op10, vuint8mf4_t op11, vuint8mf4_t op12, vuint8mf4_t op13, vuint8mf4_t op14, const uint8_t * op15, size_t op16); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e8_v_u8mf8_m))) +void vlseg7e8(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vuint8mf8_t * op6, vbool64_t op7, vuint8mf8_t op8, vuint8mf8_t op9, vuint8mf8_t op10, vuint8mf8_t op11, vuint8mf8_t op12, vuint8mf8_t op13, vuint8mf8_t op14, const uint8_t * op15, size_t op16); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e8_v_i8m1_m))) +void vlseg8e8(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vint8m1_t * op6, vint8m1_t * op7, vbool8_t op8, vint8m1_t op9, vint8m1_t op10, vint8m1_t op11, vint8m1_t op12, vint8m1_t op13, vint8m1_t op14, vint8m1_t op15, vint8m1_t op16, const int8_t * op17, size_t op18); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e8_v_i8mf2_m))) +void vlseg8e8(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vint8mf2_t * op6, vint8mf2_t * op7, vbool16_t op8, vint8mf2_t op9, vint8mf2_t op10, vint8mf2_t op11, vint8mf2_t op12, vint8mf2_t op13, vint8mf2_t op14, vint8mf2_t op15, vint8mf2_t op16, const int8_t * op17, size_t op18); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e8_v_i8mf4_m))) +void vlseg8e8(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vint8mf4_t * op6, vint8mf4_t * op7, vbool32_t op8, vint8mf4_t op9, vint8mf4_t op10, vint8mf4_t op11, vint8mf4_t op12, vint8mf4_t op13, vint8mf4_t op14, vint8mf4_t op15, vint8mf4_t op16, const int8_t * op17, size_t op18); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e8_v_i8mf8_m))) +void vlseg8e8(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vint8mf8_t * op6, vint8mf8_t * op7, vbool64_t op8, vint8mf8_t op9, vint8mf8_t op10, vint8mf8_t op11, vint8mf8_t op12, vint8mf8_t op13, vint8mf8_t op14, vint8mf8_t op15, vint8mf8_t op16, const int8_t * op17, size_t op18); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e8_v_u8m1_m))) +void vlseg8e8(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vuint8m1_t * op6, vuint8m1_t * op7, vbool8_t op8, vuint8m1_t op9, vuint8m1_t op10, vuint8m1_t op11, vuint8m1_t op12, vuint8m1_t op13, vuint8m1_t op14, vuint8m1_t op15, vuint8m1_t op16, const uint8_t * op17, size_t op18); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e8_v_u8mf2_m))) +void vlseg8e8(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vuint8mf2_t * op6, vuint8mf2_t * op7, vbool16_t op8, vuint8mf2_t op9, vuint8mf2_t op10, vuint8mf2_t op11, vuint8mf2_t op12, vuint8mf2_t op13, vuint8mf2_t op14, vuint8mf2_t op15, vuint8mf2_t op16, const uint8_t * op17, size_t op18); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e8_v_u8mf4_m))) +void vlseg8e8(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vuint8mf4_t * op6, vuint8mf4_t * op7, vbool32_t op8, vuint8mf4_t op9, vuint8mf4_t op10, vuint8mf4_t op11, vuint8mf4_t op12, vuint8mf4_t op13, vuint8mf4_t op14, vuint8mf4_t op15, vuint8mf4_t op16, const uint8_t * op17, size_t op18); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e8_v_u8mf8_m))) +void vlseg8e8(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vuint8mf8_t * op6, vuint8mf8_t * op7, vbool64_t op8, vuint8mf8_t op9, vuint8mf8_t op10, vuint8mf8_t op11, vuint8mf8_t op12, vuint8mf8_t op13, vuint8mf8_t op14, vuint8mf8_t op15, vuint8mf8_t op16, const uint8_t * op17, size_t op18); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e16_v_i16m1_m))) +void vlseg2e16(vint16m1_t * op0, vint16m1_t * op1, vbool16_t op2, vint16m1_t op3, vint16m1_t op4, const int16_t * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e16_v_i16m2_m))) +void vlseg2e16(vint16m2_t * op0, vint16m2_t * op1, vbool8_t op2, vint16m2_t op3, vint16m2_t op4, const int16_t * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e16_v_i16m4_m))) +void vlseg2e16(vint16m4_t * op0, vint16m4_t * op1, vbool4_t op2, vint16m4_t op3, vint16m4_t op4, const int16_t * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e16_v_i16mf2_m))) +void vlseg2e16(vint16mf2_t * op0, vint16mf2_t * op1, vbool32_t op2, vint16mf2_t op3, vint16mf2_t op4, const int16_t * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e16_v_i16mf4_m))) +void vlseg2e16(vint16mf4_t * op0, vint16mf4_t * op1, vbool64_t op2, vint16mf4_t op3, vint16mf4_t op4, const int16_t * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e16_v_u16m1_m))) +void vlseg2e16(vuint16m1_t * op0, vuint16m1_t * op1, vbool16_t op2, vuint16m1_t op3, vuint16m1_t op4, const uint16_t * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e16_v_u16m2_m))) +void vlseg2e16(vuint16m2_t * op0, vuint16m2_t * op1, vbool8_t op2, vuint16m2_t op3, vuint16m2_t op4, const uint16_t * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e16_v_u16m4_m))) +void vlseg2e16(vuint16m4_t * op0, vuint16m4_t * op1, vbool4_t op2, vuint16m4_t op3, vuint16m4_t op4, const uint16_t * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e16_v_u16mf2_m))) +void vlseg2e16(vuint16mf2_t * op0, vuint16mf2_t * op1, vbool32_t op2, vuint16mf2_t op3, vuint16mf2_t op4, const uint16_t * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e16_v_u16mf4_m))) +void vlseg2e16(vuint16mf4_t * op0, vuint16mf4_t * op1, vbool64_t op2, vuint16mf4_t op3, vuint16mf4_t op4, const uint16_t * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e16_v_i16m1_m))) +void vlseg3e16(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vbool16_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, const int16_t * op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e16_v_i16m2_m))) +void vlseg3e16(vint16m2_t * op0, vint16m2_t * op1, vint16m2_t * op2, vbool8_t op3, vint16m2_t op4, vint16m2_t op5, vint16m2_t op6, const int16_t * op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e16_v_i16mf2_m))) +void vlseg3e16(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vbool32_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, const int16_t * op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e16_v_i16mf4_m))) +void vlseg3e16(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vbool64_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, const int16_t * op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e16_v_u16m1_m))) +void vlseg3e16(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vbool16_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, const uint16_t * op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e16_v_u16m2_m))) +void vlseg3e16(vuint16m2_t * op0, vuint16m2_t * op1, vuint16m2_t * op2, vbool8_t op3, vuint16m2_t op4, vuint16m2_t op5, vuint16m2_t op6, const uint16_t * op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e16_v_u16mf2_m))) +void vlseg3e16(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vbool32_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, const uint16_t * op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e16_v_u16mf4_m))) +void vlseg3e16(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vbool64_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, const uint16_t * op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e16_v_i16m1_m))) +void vlseg4e16(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vbool16_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, const int16_t * op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e16_v_i16m2_m))) +void vlseg4e16(vint16m2_t * op0, vint16m2_t * op1, vint16m2_t * op2, vint16m2_t * op3, vbool8_t op4, vint16m2_t op5, vint16m2_t op6, vint16m2_t op7, vint16m2_t op8, const int16_t * op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e16_v_i16mf2_m))) +void vlseg4e16(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vbool32_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, const int16_t * op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e16_v_i16mf4_m))) +void vlseg4e16(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vbool64_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, const int16_t * op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e16_v_u16m1_m))) +void vlseg4e16(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vbool16_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, const uint16_t * op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e16_v_u16m2_m))) +void vlseg4e16(vuint16m2_t * op0, vuint16m2_t * op1, vuint16m2_t * op2, vuint16m2_t * op3, vbool8_t op4, vuint16m2_t op5, vuint16m2_t op6, vuint16m2_t op7, vuint16m2_t op8, const uint16_t * op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e16_v_u16mf2_m))) +void vlseg4e16(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vbool32_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, const uint16_t * op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e16_v_u16mf4_m))) +void vlseg4e16(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vbool64_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, const uint16_t * op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e16_v_i16m1_m))) +void vlseg5e16(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vbool16_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, vint16m1_t op10, const int16_t * op11, size_t op12); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e16_v_i16mf2_m))) +void vlseg5e16(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vbool32_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, vint16mf2_t op10, const int16_t * op11, size_t op12); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e16_v_i16mf4_m))) +void vlseg5e16(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vbool64_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, vint16mf4_t op10, const int16_t * op11, size_t op12); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e16_v_u16m1_m))) +void vlseg5e16(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vbool16_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, vuint16m1_t op10, const uint16_t * op11, size_t op12); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e16_v_u16mf2_m))) +void vlseg5e16(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vbool32_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, vuint16mf2_t op10, const uint16_t * op11, size_t op12); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e16_v_u16mf4_m))) +void vlseg5e16(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vbool64_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, vuint16mf4_t op10, const uint16_t * op11, size_t op12); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e16_v_i16m1_m))) +void vlseg6e16(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vbool16_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, vint16m1_t op10, vint16m1_t op11, vint16m1_t op12, const int16_t * op13, size_t op14); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e16_v_i16mf2_m))) +void vlseg6e16(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vbool32_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, vint16mf2_t op10, vint16mf2_t op11, vint16mf2_t op12, const int16_t * op13, size_t op14); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e16_v_i16mf4_m))) +void vlseg6e16(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vbool64_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, vint16mf4_t op10, vint16mf4_t op11, vint16mf4_t op12, const int16_t * op13, size_t op14); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e16_v_u16m1_m))) +void vlseg6e16(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vbool16_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, vuint16m1_t op10, vuint16m1_t op11, vuint16m1_t op12, const uint16_t * op13, size_t op14); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e16_v_u16mf2_m))) +void vlseg6e16(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vbool32_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, vuint16mf2_t op10, vuint16mf2_t op11, vuint16mf2_t op12, const uint16_t * op13, size_t op14); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e16_v_u16mf4_m))) +void vlseg6e16(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vbool64_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, vuint16mf4_t op10, vuint16mf4_t op11, vuint16mf4_t op12, const uint16_t * op13, size_t op14); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e16_v_i16m1_m))) +void vlseg7e16(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vint16m1_t * op6, vbool16_t op7, vint16m1_t op8, vint16m1_t op9, vint16m1_t op10, vint16m1_t op11, vint16m1_t op12, vint16m1_t op13, vint16m1_t op14, const int16_t * op15, size_t op16); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e16_v_i16mf2_m))) +void vlseg7e16(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vint16mf2_t * op6, vbool32_t op7, vint16mf2_t op8, vint16mf2_t op9, vint16mf2_t op10, vint16mf2_t op11, vint16mf2_t op12, vint16mf2_t op13, vint16mf2_t op14, const int16_t * op15, size_t op16); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e16_v_i16mf4_m))) +void vlseg7e16(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vint16mf4_t * op6, vbool64_t op7, vint16mf4_t op8, vint16mf4_t op9, vint16mf4_t op10, vint16mf4_t op11, vint16mf4_t op12, vint16mf4_t op13, vint16mf4_t op14, const int16_t * op15, size_t op16); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e16_v_u16m1_m))) +void vlseg7e16(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vuint16m1_t * op6, vbool16_t op7, vuint16m1_t op8, vuint16m1_t op9, vuint16m1_t op10, vuint16m1_t op11, vuint16m1_t op12, vuint16m1_t op13, vuint16m1_t op14, const uint16_t * op15, size_t op16); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e16_v_u16mf2_m))) +void vlseg7e16(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vuint16mf2_t * op6, vbool32_t op7, vuint16mf2_t op8, vuint16mf2_t op9, vuint16mf2_t op10, vuint16mf2_t op11, vuint16mf2_t op12, vuint16mf2_t op13, vuint16mf2_t op14, const uint16_t * op15, size_t op16); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e16_v_u16mf4_m))) +void vlseg7e16(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vuint16mf4_t * op6, vbool64_t op7, vuint16mf4_t op8, vuint16mf4_t op9, vuint16mf4_t op10, vuint16mf4_t op11, vuint16mf4_t op12, vuint16mf4_t op13, vuint16mf4_t op14, const uint16_t * op15, size_t op16); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e16_v_i16m1_m))) +void vlseg8e16(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vint16m1_t * op6, vint16m1_t * op7, vbool16_t op8, vint16m1_t op9, vint16m1_t op10, vint16m1_t op11, vint16m1_t op12, vint16m1_t op13, vint16m1_t op14, vint16m1_t op15, vint16m1_t op16, const int16_t * op17, size_t op18); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e16_v_i16mf2_m))) +void vlseg8e16(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vint16mf2_t * op6, vint16mf2_t * op7, vbool32_t op8, vint16mf2_t op9, vint16mf2_t op10, vint16mf2_t op11, vint16mf2_t op12, vint16mf2_t op13, vint16mf2_t op14, vint16mf2_t op15, vint16mf2_t op16, const int16_t * op17, size_t op18); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e16_v_i16mf4_m))) +void vlseg8e16(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vint16mf4_t * op6, vint16mf4_t * op7, vbool64_t op8, vint16mf4_t op9, vint16mf4_t op10, vint16mf4_t op11, vint16mf4_t op12, vint16mf4_t op13, vint16mf4_t op14, vint16mf4_t op15, vint16mf4_t op16, const int16_t * op17, size_t op18); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e16_v_u16m1_m))) +void vlseg8e16(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vuint16m1_t * op6, vuint16m1_t * op7, vbool16_t op8, vuint16m1_t op9, vuint16m1_t op10, vuint16m1_t op11, vuint16m1_t op12, vuint16m1_t op13, vuint16m1_t op14, vuint16m1_t op15, vuint16m1_t op16, const uint16_t * op17, size_t op18); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e16_v_u16mf2_m))) +void vlseg8e16(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vuint16mf2_t * op6, vuint16mf2_t * op7, vbool32_t op8, vuint16mf2_t op9, vuint16mf2_t op10, vuint16mf2_t op11, vuint16mf2_t op12, vuint16mf2_t op13, vuint16mf2_t op14, vuint16mf2_t op15, vuint16mf2_t op16, const uint16_t * op17, size_t op18); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e16_v_u16mf4_m))) +void vlseg8e16(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vuint16mf4_t * op6, vuint16mf4_t * op7, vbool64_t op8, vuint16mf4_t op9, vuint16mf4_t op10, vuint16mf4_t op11, vuint16mf4_t op12, vuint16mf4_t op13, vuint16mf4_t op14, vuint16mf4_t op15, vuint16mf4_t op16, const uint16_t * op17, size_t op18); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e32_v_i32m1_m))) +void vlseg2e32(vint32m1_t * op0, vint32m1_t * op1, vbool32_t op2, vint32m1_t op3, vint32m1_t op4, const int32_t * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e32_v_i32m2_m))) +void vlseg2e32(vint32m2_t * op0, vint32m2_t * op1, vbool16_t op2, vint32m2_t op3, vint32m2_t op4, const int32_t * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e32_v_i32m4_m))) +void vlseg2e32(vint32m4_t * op0, vint32m4_t * op1, vbool8_t op2, vint32m4_t op3, vint32m4_t op4, const int32_t * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e32_v_i32mf2_m))) +void vlseg2e32(vint32mf2_t * op0, vint32mf2_t * op1, vbool64_t op2, vint32mf2_t op3, vint32mf2_t op4, const int32_t * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e32_v_u32m1_m))) +void vlseg2e32(vuint32m1_t * op0, vuint32m1_t * op1, vbool32_t op2, vuint32m1_t op3, vuint32m1_t op4, const uint32_t * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e32_v_u32m2_m))) +void vlseg2e32(vuint32m2_t * op0, vuint32m2_t * op1, vbool16_t op2, vuint32m2_t op3, vuint32m2_t op4, const uint32_t * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e32_v_u32m4_m))) +void vlseg2e32(vuint32m4_t * op0, vuint32m4_t * op1, vbool8_t op2, vuint32m4_t op3, vuint32m4_t op4, const uint32_t * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e32_v_u32mf2_m))) +void vlseg2e32(vuint32mf2_t * op0, vuint32mf2_t * op1, vbool64_t op2, vuint32mf2_t op3, vuint32mf2_t op4, const uint32_t * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e32_v_i32m1_m))) +void vlseg3e32(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vbool32_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, const int32_t * op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e32_v_i32m2_m))) +void vlseg3e32(vint32m2_t * op0, vint32m2_t * op1, vint32m2_t * op2, vbool16_t op3, vint32m2_t op4, vint32m2_t op5, vint32m2_t op6, const int32_t * op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e32_v_i32mf2_m))) +void vlseg3e32(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vbool64_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, const int32_t * op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e32_v_u32m1_m))) +void vlseg3e32(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vbool32_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, const uint32_t * op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e32_v_u32m2_m))) +void vlseg3e32(vuint32m2_t * op0, vuint32m2_t * op1, vuint32m2_t * op2, vbool16_t op3, vuint32m2_t op4, vuint32m2_t op5, vuint32m2_t op6, const uint32_t * op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e32_v_u32mf2_m))) +void vlseg3e32(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vbool64_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, const uint32_t * op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e32_v_i32m1_m))) +void vlseg4e32(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vbool32_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, const int32_t * op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e32_v_i32m2_m))) +void vlseg4e32(vint32m2_t * op0, vint32m2_t * op1, vint32m2_t * op2, vint32m2_t * op3, vbool16_t op4, vint32m2_t op5, vint32m2_t op6, vint32m2_t op7, vint32m2_t op8, const int32_t * op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e32_v_i32mf2_m))) +void vlseg4e32(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vbool64_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, const int32_t * op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e32_v_u32m1_m))) +void vlseg4e32(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vbool32_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, const uint32_t * op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e32_v_u32m2_m))) +void vlseg4e32(vuint32m2_t * op0, vuint32m2_t * op1, vuint32m2_t * op2, vuint32m2_t * op3, vbool16_t op4, vuint32m2_t op5, vuint32m2_t op6, vuint32m2_t op7, vuint32m2_t op8, const uint32_t * op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e32_v_u32mf2_m))) +void vlseg4e32(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vbool64_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, const uint32_t * op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e32_v_i32m1_m))) +void vlseg5e32(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vbool32_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, vint32m1_t op10, const int32_t * op11, size_t op12); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e32_v_i32mf2_m))) +void vlseg5e32(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vbool64_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, vint32mf2_t op10, const int32_t * op11, size_t op12); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e32_v_u32m1_m))) +void vlseg5e32(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vbool32_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, vuint32m1_t op10, const uint32_t * op11, size_t op12); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e32_v_u32mf2_m))) +void vlseg5e32(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vbool64_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, vuint32mf2_t op10, const uint32_t * op11, size_t op12); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e32_v_i32m1_m))) +void vlseg6e32(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vbool32_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, vint32m1_t op10, vint32m1_t op11, vint32m1_t op12, const int32_t * op13, size_t op14); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e32_v_i32mf2_m))) +void vlseg6e32(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vbool64_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, vint32mf2_t op10, vint32mf2_t op11, vint32mf2_t op12, const int32_t * op13, size_t op14); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e32_v_u32m1_m))) +void vlseg6e32(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vbool32_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, vuint32m1_t op10, vuint32m1_t op11, vuint32m1_t op12, const uint32_t * op13, size_t op14); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e32_v_u32mf2_m))) +void vlseg6e32(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vbool64_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, vuint32mf2_t op10, vuint32mf2_t op11, vuint32mf2_t op12, const uint32_t * op13, size_t op14); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e32_v_i32m1_m))) +void vlseg7e32(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vint32m1_t * op6, vbool32_t op7, vint32m1_t op8, vint32m1_t op9, vint32m1_t op10, vint32m1_t op11, vint32m1_t op12, vint32m1_t op13, vint32m1_t op14, const int32_t * op15, size_t op16); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e32_v_i32mf2_m))) +void vlseg7e32(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vint32mf2_t * op6, vbool64_t op7, vint32mf2_t op8, vint32mf2_t op9, vint32mf2_t op10, vint32mf2_t op11, vint32mf2_t op12, vint32mf2_t op13, vint32mf2_t op14, const int32_t * op15, size_t op16); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e32_v_u32m1_m))) +void vlseg7e32(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vuint32m1_t * op6, vbool32_t op7, vuint32m1_t op8, vuint32m1_t op9, vuint32m1_t op10, vuint32m1_t op11, vuint32m1_t op12, vuint32m1_t op13, vuint32m1_t op14, const uint32_t * op15, size_t op16); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e32_v_u32mf2_m))) +void vlseg7e32(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vuint32mf2_t * op6, vbool64_t op7, vuint32mf2_t op8, vuint32mf2_t op9, vuint32mf2_t op10, vuint32mf2_t op11, vuint32mf2_t op12, vuint32mf2_t op13, vuint32mf2_t op14, const uint32_t * op15, size_t op16); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e32_v_i32m1_m))) +void vlseg8e32(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vint32m1_t * op6, vint32m1_t * op7, vbool32_t op8, vint32m1_t op9, vint32m1_t op10, vint32m1_t op11, vint32m1_t op12, vint32m1_t op13, vint32m1_t op14, vint32m1_t op15, vint32m1_t op16, const int32_t * op17, size_t op18); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e32_v_i32mf2_m))) +void vlseg8e32(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vint32mf2_t * op6, vint32mf2_t * op7, vbool64_t op8, vint32mf2_t op9, vint32mf2_t op10, vint32mf2_t op11, vint32mf2_t op12, vint32mf2_t op13, vint32mf2_t op14, vint32mf2_t op15, vint32mf2_t op16, const int32_t * op17, size_t op18); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e32_v_u32m1_m))) +void vlseg8e32(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vuint32m1_t * op6, vuint32m1_t * op7, vbool32_t op8, vuint32m1_t op9, vuint32m1_t op10, vuint32m1_t op11, vuint32m1_t op12, vuint32m1_t op13, vuint32m1_t op14, vuint32m1_t op15, vuint32m1_t op16, const uint32_t * op17, size_t op18); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e32_v_u32mf2_m))) +void vlseg8e32(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vuint32mf2_t * op6, vuint32mf2_t * op7, vbool64_t op8, vuint32mf2_t op9, vuint32mf2_t op10, vuint32mf2_t op11, vuint32mf2_t op12, vuint32mf2_t op13, vuint32mf2_t op14, vuint32mf2_t op15, vuint32mf2_t op16, const uint32_t * op17, size_t op18); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e64_v_i64m1_m))) +void vlseg2e64(vint64m1_t * op0, vint64m1_t * op1, vbool64_t op2, vint64m1_t op3, vint64m1_t op4, const int64_t * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e64_v_i64m2_m))) +void vlseg2e64(vint64m2_t * op0, vint64m2_t * op1, vbool32_t op2, vint64m2_t op3, vint64m2_t op4, const int64_t * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e64_v_i64m4_m))) +void vlseg2e64(vint64m4_t * op0, vint64m4_t * op1, vbool16_t op2, vint64m4_t op3, vint64m4_t op4, const int64_t * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e64_v_u64m1_m))) +void vlseg2e64(vuint64m1_t * op0, vuint64m1_t * op1, vbool64_t op2, vuint64m1_t op3, vuint64m1_t op4, const uint64_t * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e64_v_u64m2_m))) +void vlseg2e64(vuint64m2_t * op0, vuint64m2_t * op1, vbool32_t op2, vuint64m2_t op3, vuint64m2_t op4, const uint64_t * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e64_v_u64m4_m))) +void vlseg2e64(vuint64m4_t * op0, vuint64m4_t * op1, vbool16_t op2, vuint64m4_t op3, vuint64m4_t op4, const uint64_t * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e64_v_i64m1_m))) +void vlseg3e64(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vbool64_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, const int64_t * op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e64_v_i64m2_m))) +void vlseg3e64(vint64m2_t * op0, vint64m2_t * op1, vint64m2_t * op2, vbool32_t op3, vint64m2_t op4, vint64m2_t op5, vint64m2_t op6, const int64_t * op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e64_v_u64m1_m))) +void vlseg3e64(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vbool64_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, const uint64_t * op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e64_v_u64m2_m))) +void vlseg3e64(vuint64m2_t * op0, vuint64m2_t * op1, vuint64m2_t * op2, vbool32_t op3, vuint64m2_t op4, vuint64m2_t op5, vuint64m2_t op6, const uint64_t * op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e64_v_i64m1_m))) +void vlseg4e64(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vbool64_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, const int64_t * op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e64_v_i64m2_m))) +void vlseg4e64(vint64m2_t * op0, vint64m2_t * op1, vint64m2_t * op2, vint64m2_t * op3, vbool32_t op4, vint64m2_t op5, vint64m2_t op6, vint64m2_t op7, vint64m2_t op8, const int64_t * op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e64_v_u64m1_m))) +void vlseg4e64(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vbool64_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, const uint64_t * op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e64_v_u64m2_m))) +void vlseg4e64(vuint64m2_t * op0, vuint64m2_t * op1, vuint64m2_t * op2, vuint64m2_t * op3, vbool32_t op4, vuint64m2_t op5, vuint64m2_t op6, vuint64m2_t op7, vuint64m2_t op8, const uint64_t * op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e64_v_i64m1_m))) +void vlseg5e64(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vbool64_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, vint64m1_t op10, const int64_t * op11, size_t op12); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e64_v_u64m1_m))) +void vlseg5e64(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vbool64_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, vuint64m1_t op10, const uint64_t * op11, size_t op12); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e64_v_i64m1_m))) +void vlseg6e64(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vbool64_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, vint64m1_t op10, vint64m1_t op11, vint64m1_t op12, const int64_t * op13, size_t op14); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e64_v_u64m1_m))) +void vlseg6e64(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vbool64_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, vuint64m1_t op10, vuint64m1_t op11, vuint64m1_t op12, const uint64_t * op13, size_t op14); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e64_v_i64m1_m))) +void vlseg7e64(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vint64m1_t * op6, vbool64_t op7, vint64m1_t op8, vint64m1_t op9, vint64m1_t op10, vint64m1_t op11, vint64m1_t op12, vint64m1_t op13, vint64m1_t op14, const int64_t * op15, size_t op16); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e64_v_u64m1_m))) +void vlseg7e64(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vuint64m1_t * op6, vbool64_t op7, vuint64m1_t op8, vuint64m1_t op9, vuint64m1_t op10, vuint64m1_t op11, vuint64m1_t op12, vuint64m1_t op13, vuint64m1_t op14, const uint64_t * op15, size_t op16); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e64_v_i64m1_m))) +void vlseg8e64(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vint64m1_t * op6, vint64m1_t * op7, vbool64_t op8, vint64m1_t op9, vint64m1_t op10, vint64m1_t op11, vint64m1_t op12, vint64m1_t op13, vint64m1_t op14, vint64m1_t op15, vint64m1_t op16, const int64_t * op17, size_t op18); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e64_v_u64m1_m))) +void vlseg8e64(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vuint64m1_t * op6, vuint64m1_t * op7, vbool64_t op8, vuint64m1_t op9, vuint64m1_t op10, vuint64m1_t op11, vuint64m1_t op12, vuint64m1_t op13, vuint64m1_t op14, vuint64m1_t op15, vuint64m1_t op16, const uint64_t * op17, size_t op18); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e8ff_v_i8m1_m))) +void vlseg3e8ff(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vbool8_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, const int8_t * op7, size_t * op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e8ff_v_i8m2_m))) +void vlseg3e8ff(vint8m2_t * op0, vint8m2_t * op1, vint8m2_t * op2, vbool4_t op3, vint8m2_t op4, vint8m2_t op5, vint8m2_t op6, const int8_t * op7, size_t * op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e8ff_v_i8mf2_m))) +void vlseg3e8ff(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vbool16_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, const int8_t * op7, size_t * op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e8ff_v_i8mf4_m))) +void vlseg3e8ff(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vbool32_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, const int8_t * op7, size_t * op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e8ff_v_i8mf8_m))) +void vlseg3e8ff(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vbool64_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, const int8_t * op7, size_t * op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e8ff_v_u8m1_m))) +void vlseg3e8ff(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vbool8_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, const uint8_t * op7, size_t * op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e8ff_v_u8m2_m))) +void vlseg3e8ff(vuint8m2_t * op0, vuint8m2_t * op1, vuint8m2_t * op2, vbool4_t op3, vuint8m2_t op4, vuint8m2_t op5, vuint8m2_t op6, const uint8_t * op7, size_t * op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e8ff_v_u8mf2_m))) +void vlseg3e8ff(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vbool16_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, const uint8_t * op7, size_t * op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e8ff_v_u8mf4_m))) +void vlseg3e8ff(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vbool32_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, const uint8_t * op7, size_t * op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e8ff_v_u8mf8_m))) +void vlseg3e8ff(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vbool64_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, const uint8_t * op7, size_t * op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e8ff_v_i8m1_m))) +void vlseg4e8ff(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vbool8_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, const int8_t * op9, size_t * op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e8ff_v_i8m2_m))) +void vlseg4e8ff(vint8m2_t * op0, vint8m2_t * op1, vint8m2_t * op2, vint8m2_t * op3, vbool4_t op4, vint8m2_t op5, vint8m2_t op6, vint8m2_t op7, vint8m2_t op8, const int8_t * op9, size_t * op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e8ff_v_i8mf2_m))) +void vlseg4e8ff(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vbool16_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, const int8_t * op9, size_t * op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e8ff_v_i8mf4_m))) +void vlseg4e8ff(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vbool32_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, const int8_t * op9, size_t * op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e8ff_v_i8mf8_m))) +void vlseg4e8ff(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vbool64_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, const int8_t * op9, size_t * op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e8ff_v_u8m1_m))) +void vlseg4e8ff(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vbool8_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, const uint8_t * op9, size_t * op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e8ff_v_u8m2_m))) +void vlseg4e8ff(vuint8m2_t * op0, vuint8m2_t * op1, vuint8m2_t * op2, vuint8m2_t * op3, vbool4_t op4, vuint8m2_t op5, vuint8m2_t op6, vuint8m2_t op7, vuint8m2_t op8, const uint8_t * op9, size_t * op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e8ff_v_u8mf2_m))) +void vlseg4e8ff(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vbool16_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, const uint8_t * op9, size_t * op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e8ff_v_u8mf4_m))) +void vlseg4e8ff(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vbool32_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, const uint8_t * op9, size_t * op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e8ff_v_u8mf8_m))) +void vlseg4e8ff(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vbool64_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, const uint8_t * op9, size_t * op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e8ff_v_i8m1_m))) +void vlseg5e8ff(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vbool8_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, vint8m1_t op10, const int8_t * op11, size_t * op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e8ff_v_i8mf2_m))) +void vlseg5e8ff(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vbool16_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, vint8mf2_t op10, const int8_t * op11, size_t * op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e8ff_v_i8mf4_m))) +void vlseg5e8ff(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vbool32_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, vint8mf4_t op10, const int8_t * op11, size_t * op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e8ff_v_i8mf8_m))) +void vlseg5e8ff(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vbool64_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, vint8mf8_t op10, const int8_t * op11, size_t * op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e8ff_v_u8m1_m))) +void vlseg5e8ff(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vbool8_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, vuint8m1_t op10, const uint8_t * op11, size_t * op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e8ff_v_u8mf2_m))) +void vlseg5e8ff(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vbool16_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, vuint8mf2_t op10, const uint8_t * op11, size_t * op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e8ff_v_u8mf4_m))) +void vlseg5e8ff(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vbool32_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, vuint8mf4_t op10, const uint8_t * op11, size_t * op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e8ff_v_u8mf8_m))) +void vlseg5e8ff(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vbool64_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, vuint8mf8_t op10, const uint8_t * op11, size_t * op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e8ff_v_i8m1_m))) +void vlseg6e8ff(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vbool8_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, vint8m1_t op10, vint8m1_t op11, vint8m1_t op12, const int8_t * op13, size_t * op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e8ff_v_i8mf2_m))) +void vlseg6e8ff(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vbool16_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, vint8mf2_t op10, vint8mf2_t op11, vint8mf2_t op12, const int8_t * op13, size_t * op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e8ff_v_i8mf4_m))) +void vlseg6e8ff(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vbool32_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, vint8mf4_t op10, vint8mf4_t op11, vint8mf4_t op12, const int8_t * op13, size_t * op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e8ff_v_i8mf8_m))) +void vlseg6e8ff(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vbool64_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, vint8mf8_t op10, vint8mf8_t op11, vint8mf8_t op12, const int8_t * op13, size_t * op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e8ff_v_u8m1_m))) +void vlseg6e8ff(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vbool8_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, vuint8m1_t op10, vuint8m1_t op11, vuint8m1_t op12, const uint8_t * op13, size_t * op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e8ff_v_u8mf2_m))) +void vlseg6e8ff(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vbool16_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, vuint8mf2_t op10, vuint8mf2_t op11, vuint8mf2_t op12, const uint8_t * op13, size_t * op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e8ff_v_u8mf4_m))) +void vlseg6e8ff(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vbool32_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, vuint8mf4_t op10, vuint8mf4_t op11, vuint8mf4_t op12, const uint8_t * op13, size_t * op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e8ff_v_u8mf8_m))) +void vlseg6e8ff(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vbool64_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, vuint8mf8_t op10, vuint8mf8_t op11, vuint8mf8_t op12, const uint8_t * op13, size_t * op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e8ff_v_i8m1_m))) +void vlseg7e8ff(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vint8m1_t * op6, vbool8_t op7, vint8m1_t op8, vint8m1_t op9, vint8m1_t op10, vint8m1_t op11, vint8m1_t op12, vint8m1_t op13, vint8m1_t op14, const int8_t * op15, size_t * op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e8ff_v_i8mf2_m))) +void vlseg7e8ff(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vint8mf2_t * op6, vbool16_t op7, vint8mf2_t op8, vint8mf2_t op9, vint8mf2_t op10, vint8mf2_t op11, vint8mf2_t op12, vint8mf2_t op13, vint8mf2_t op14, const int8_t * op15, size_t * op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e8ff_v_i8mf4_m))) +void vlseg7e8ff(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vint8mf4_t * op6, vbool32_t op7, vint8mf4_t op8, vint8mf4_t op9, vint8mf4_t op10, vint8mf4_t op11, vint8mf4_t op12, vint8mf4_t op13, vint8mf4_t op14, const int8_t * op15, size_t * op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e8ff_v_i8mf8_m))) +void vlseg7e8ff(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vint8mf8_t * op6, vbool64_t op7, vint8mf8_t op8, vint8mf8_t op9, vint8mf8_t op10, vint8mf8_t op11, vint8mf8_t op12, vint8mf8_t op13, vint8mf8_t op14, const int8_t * op15, size_t * op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e8ff_v_u8m1_m))) +void vlseg7e8ff(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vuint8m1_t * op6, vbool8_t op7, vuint8m1_t op8, vuint8m1_t op9, vuint8m1_t op10, vuint8m1_t op11, vuint8m1_t op12, vuint8m1_t op13, vuint8m1_t op14, const uint8_t * op15, size_t * op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e8ff_v_u8mf2_m))) +void vlseg7e8ff(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vuint8mf2_t * op6, vbool16_t op7, vuint8mf2_t op8, vuint8mf2_t op9, vuint8mf2_t op10, vuint8mf2_t op11, vuint8mf2_t op12, vuint8mf2_t op13, vuint8mf2_t op14, const uint8_t * op15, size_t * op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e8ff_v_u8mf4_m))) +void vlseg7e8ff(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vuint8mf4_t * op6, vbool32_t op7, vuint8mf4_t op8, vuint8mf4_t op9, vuint8mf4_t op10, vuint8mf4_t op11, vuint8mf4_t op12, vuint8mf4_t op13, vuint8mf4_t op14, const uint8_t * op15, size_t * op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e8ff_v_u8mf8_m))) +void vlseg7e8ff(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vuint8mf8_t * op6, vbool64_t op7, vuint8mf8_t op8, vuint8mf8_t op9, vuint8mf8_t op10, vuint8mf8_t op11, vuint8mf8_t op12, vuint8mf8_t op13, vuint8mf8_t op14, const uint8_t * op15, size_t * op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e8ff_v_i8m1_m))) +void vlseg8e8ff(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vint8m1_t * op6, vint8m1_t * op7, vbool8_t op8, vint8m1_t op9, vint8m1_t op10, vint8m1_t op11, vint8m1_t op12, vint8m1_t op13, vint8m1_t op14, vint8m1_t op15, vint8m1_t op16, const int8_t * op17, size_t * op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e8ff_v_i8mf2_m))) +void vlseg8e8ff(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vint8mf2_t * op6, vint8mf2_t * op7, vbool16_t op8, vint8mf2_t op9, vint8mf2_t op10, vint8mf2_t op11, vint8mf2_t op12, vint8mf2_t op13, vint8mf2_t op14, vint8mf2_t op15, vint8mf2_t op16, const int8_t * op17, size_t * op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e8ff_v_i8mf4_m))) +void vlseg8e8ff(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vint8mf4_t * op6, vint8mf4_t * op7, vbool32_t op8, vint8mf4_t op9, vint8mf4_t op10, vint8mf4_t op11, vint8mf4_t op12, vint8mf4_t op13, vint8mf4_t op14, vint8mf4_t op15, vint8mf4_t op16, const int8_t * op17, size_t * op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e8ff_v_i8mf8_m))) +void vlseg8e8ff(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vint8mf8_t * op6, vint8mf8_t * op7, vbool64_t op8, vint8mf8_t op9, vint8mf8_t op10, vint8mf8_t op11, vint8mf8_t op12, vint8mf8_t op13, vint8mf8_t op14, vint8mf8_t op15, vint8mf8_t op16, const int8_t * op17, size_t * op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e8ff_v_u8m1_m))) +void vlseg8e8ff(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vuint8m1_t * op6, vuint8m1_t * op7, vbool8_t op8, vuint8m1_t op9, vuint8m1_t op10, vuint8m1_t op11, vuint8m1_t op12, vuint8m1_t op13, vuint8m1_t op14, vuint8m1_t op15, vuint8m1_t op16, const uint8_t * op17, size_t * op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e8ff_v_u8mf2_m))) +void vlseg8e8ff(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vuint8mf2_t * op6, vuint8mf2_t * op7, vbool16_t op8, vuint8mf2_t op9, vuint8mf2_t op10, vuint8mf2_t op11, vuint8mf2_t op12, vuint8mf2_t op13, vuint8mf2_t op14, vuint8mf2_t op15, vuint8mf2_t op16, const uint8_t * op17, size_t * op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e8ff_v_u8mf4_m))) +void vlseg8e8ff(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vuint8mf4_t * op6, vuint8mf4_t * op7, vbool32_t op8, vuint8mf4_t op9, vuint8mf4_t op10, vuint8mf4_t op11, vuint8mf4_t op12, vuint8mf4_t op13, vuint8mf4_t op14, vuint8mf4_t op15, vuint8mf4_t op16, const uint8_t * op17, size_t * op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e8ff_v_u8mf8_m))) +void vlseg8e8ff(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vuint8mf8_t * op6, vuint8mf8_t * op7, vbool64_t op8, vuint8mf8_t op9, vuint8mf8_t op10, vuint8mf8_t op11, vuint8mf8_t op12, vuint8mf8_t op13, vuint8mf8_t op14, vuint8mf8_t op15, vuint8mf8_t op16, const uint8_t * op17, size_t * op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e16ff_v_i16m1_m))) +void vlseg2e16ff(vint16m1_t * op0, vint16m1_t * op1, vbool16_t op2, vint16m1_t op3, vint16m1_t op4, const int16_t * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e16ff_v_i16m2_m))) +void vlseg2e16ff(vint16m2_t * op0, vint16m2_t * op1, vbool8_t op2, vint16m2_t op3, vint16m2_t op4, const int16_t * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e16ff_v_i16m4_m))) +void vlseg2e16ff(vint16m4_t * op0, vint16m4_t * op1, vbool4_t op2, vint16m4_t op3, vint16m4_t op4, const int16_t * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e16ff_v_i16mf2_m))) +void vlseg2e16ff(vint16mf2_t * op0, vint16mf2_t * op1, vbool32_t op2, vint16mf2_t op3, vint16mf2_t op4, const int16_t * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e16ff_v_i16mf4_m))) +void vlseg2e16ff(vint16mf4_t * op0, vint16mf4_t * op1, vbool64_t op2, vint16mf4_t op3, vint16mf4_t op4, const int16_t * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e16ff_v_u16m1_m))) +void vlseg2e16ff(vuint16m1_t * op0, vuint16m1_t * op1, vbool16_t op2, vuint16m1_t op3, vuint16m1_t op4, const uint16_t * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e16ff_v_u16m2_m))) +void vlseg2e16ff(vuint16m2_t * op0, vuint16m2_t * op1, vbool8_t op2, vuint16m2_t op3, vuint16m2_t op4, const uint16_t * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e16ff_v_u16m4_m))) +void vlseg2e16ff(vuint16m4_t * op0, vuint16m4_t * op1, vbool4_t op2, vuint16m4_t op3, vuint16m4_t op4, const uint16_t * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e16ff_v_u16mf2_m))) +void vlseg2e16ff(vuint16mf2_t * op0, vuint16mf2_t * op1, vbool32_t op2, vuint16mf2_t op3, vuint16mf2_t op4, const uint16_t * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e16ff_v_u16mf4_m))) +void vlseg2e16ff(vuint16mf4_t * op0, vuint16mf4_t * op1, vbool64_t op2, vuint16mf4_t op3, vuint16mf4_t op4, const uint16_t * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e16ff_v_i16m1_m))) +void vlseg3e16ff(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vbool16_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, const int16_t * op7, size_t * op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e16ff_v_i16m2_m))) +void vlseg3e16ff(vint16m2_t * op0, vint16m2_t * op1, vint16m2_t * op2, vbool8_t op3, vint16m2_t op4, vint16m2_t op5, vint16m2_t op6, const int16_t * op7, size_t * op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e16ff_v_i16mf2_m))) +void vlseg3e16ff(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vbool32_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, const int16_t * op7, size_t * op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e16ff_v_i16mf4_m))) +void vlseg3e16ff(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vbool64_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, const int16_t * op7, size_t * op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e16ff_v_u16m1_m))) +void vlseg3e16ff(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vbool16_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, const uint16_t * op7, size_t * op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e16ff_v_u16m2_m))) +void vlseg3e16ff(vuint16m2_t * op0, vuint16m2_t * op1, vuint16m2_t * op2, vbool8_t op3, vuint16m2_t op4, vuint16m2_t op5, vuint16m2_t op6, const uint16_t * op7, size_t * op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e16ff_v_u16mf2_m))) +void vlseg3e16ff(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vbool32_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, const uint16_t * op7, size_t * op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e16ff_v_u16mf4_m))) +void vlseg3e16ff(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vbool64_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, const uint16_t * op7, size_t * op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e16ff_v_i16m1_m))) +void vlseg4e16ff(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vbool16_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, const int16_t * op9, size_t * op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e16ff_v_i16m2_m))) +void vlseg4e16ff(vint16m2_t * op0, vint16m2_t * op1, vint16m2_t * op2, vint16m2_t * op3, vbool8_t op4, vint16m2_t op5, vint16m2_t op6, vint16m2_t op7, vint16m2_t op8, const int16_t * op9, size_t * op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e16ff_v_i16mf2_m))) +void vlseg4e16ff(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vbool32_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, const int16_t * op9, size_t * op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e16ff_v_i16mf4_m))) +void vlseg4e16ff(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vbool64_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, const int16_t * op9, size_t * op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e16ff_v_u16m1_m))) +void vlseg4e16ff(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vbool16_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, const uint16_t * op9, size_t * op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e16ff_v_u16m2_m))) +void vlseg4e16ff(vuint16m2_t * op0, vuint16m2_t * op1, vuint16m2_t * op2, vuint16m2_t * op3, vbool8_t op4, vuint16m2_t op5, vuint16m2_t op6, vuint16m2_t op7, vuint16m2_t op8, const uint16_t * op9, size_t * op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e16ff_v_u16mf2_m))) +void vlseg4e16ff(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vbool32_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, const uint16_t * op9, size_t * op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e16ff_v_u16mf4_m))) +void vlseg4e16ff(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vbool64_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, const uint16_t * op9, size_t * op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e16ff_v_i16m1_m))) +void vlseg5e16ff(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vbool16_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, vint16m1_t op10, const int16_t * op11, size_t * op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e16ff_v_i16mf2_m))) +void vlseg5e16ff(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vbool32_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, vint16mf2_t op10, const int16_t * op11, size_t * op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e16ff_v_i16mf4_m))) +void vlseg5e16ff(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vbool64_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, vint16mf4_t op10, const int16_t * op11, size_t * op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e16ff_v_u16m1_m))) +void vlseg5e16ff(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vbool16_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, vuint16m1_t op10, const uint16_t * op11, size_t * op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e16ff_v_u16mf2_m))) +void vlseg5e16ff(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vbool32_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, vuint16mf2_t op10, const uint16_t * op11, size_t * op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e16ff_v_u16mf4_m))) +void vlseg5e16ff(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vbool64_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, vuint16mf4_t op10, const uint16_t * op11, size_t * op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e16ff_v_i16m1_m))) +void vlseg6e16ff(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vbool16_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, vint16m1_t op10, vint16m1_t op11, vint16m1_t op12, const int16_t * op13, size_t * op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e16ff_v_i16mf2_m))) +void vlseg6e16ff(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vbool32_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, vint16mf2_t op10, vint16mf2_t op11, vint16mf2_t op12, const int16_t * op13, size_t * op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e16ff_v_i16mf4_m))) +void vlseg6e16ff(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vbool64_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, vint16mf4_t op10, vint16mf4_t op11, vint16mf4_t op12, const int16_t * op13, size_t * op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e16ff_v_u16m1_m))) +void vlseg6e16ff(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vbool16_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, vuint16m1_t op10, vuint16m1_t op11, vuint16m1_t op12, const uint16_t * op13, size_t * op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e16ff_v_u16mf2_m))) +void vlseg6e16ff(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vbool32_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, vuint16mf2_t op10, vuint16mf2_t op11, vuint16mf2_t op12, const uint16_t * op13, size_t * op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e16ff_v_u16mf4_m))) +void vlseg6e16ff(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vbool64_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, vuint16mf4_t op10, vuint16mf4_t op11, vuint16mf4_t op12, const uint16_t * op13, size_t * op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e16ff_v_i16m1_m))) +void vlseg7e16ff(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vint16m1_t * op6, vbool16_t op7, vint16m1_t op8, vint16m1_t op9, vint16m1_t op10, vint16m1_t op11, vint16m1_t op12, vint16m1_t op13, vint16m1_t op14, const int16_t * op15, size_t * op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e16ff_v_i16mf2_m))) +void vlseg7e16ff(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vint16mf2_t * op6, vbool32_t op7, vint16mf2_t op8, vint16mf2_t op9, vint16mf2_t op10, vint16mf2_t op11, vint16mf2_t op12, vint16mf2_t op13, vint16mf2_t op14, const int16_t * op15, size_t * op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e16ff_v_i16mf4_m))) +void vlseg7e16ff(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vint16mf4_t * op6, vbool64_t op7, vint16mf4_t op8, vint16mf4_t op9, vint16mf4_t op10, vint16mf4_t op11, vint16mf4_t op12, vint16mf4_t op13, vint16mf4_t op14, const int16_t * op15, size_t * op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e16ff_v_u16m1_m))) +void vlseg7e16ff(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vuint16m1_t * op6, vbool16_t op7, vuint16m1_t op8, vuint16m1_t op9, vuint16m1_t op10, vuint16m1_t op11, vuint16m1_t op12, vuint16m1_t op13, vuint16m1_t op14, const uint16_t * op15, size_t * op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e16ff_v_u16mf2_m))) +void vlseg7e16ff(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vuint16mf2_t * op6, vbool32_t op7, vuint16mf2_t op8, vuint16mf2_t op9, vuint16mf2_t op10, vuint16mf2_t op11, vuint16mf2_t op12, vuint16mf2_t op13, vuint16mf2_t op14, const uint16_t * op15, size_t * op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e16ff_v_u16mf4_m))) +void vlseg7e16ff(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vuint16mf4_t * op6, vbool64_t op7, vuint16mf4_t op8, vuint16mf4_t op9, vuint16mf4_t op10, vuint16mf4_t op11, vuint16mf4_t op12, vuint16mf4_t op13, vuint16mf4_t op14, const uint16_t * op15, size_t * op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e16ff_v_i16m1_m))) +void vlseg8e16ff(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vint16m1_t * op6, vint16m1_t * op7, vbool16_t op8, vint16m1_t op9, vint16m1_t op10, vint16m1_t op11, vint16m1_t op12, vint16m1_t op13, vint16m1_t op14, vint16m1_t op15, vint16m1_t op16, const int16_t * op17, size_t * op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e16ff_v_i16mf2_m))) +void vlseg8e16ff(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vint16mf2_t * op6, vint16mf2_t * op7, vbool32_t op8, vint16mf2_t op9, vint16mf2_t op10, vint16mf2_t op11, vint16mf2_t op12, vint16mf2_t op13, vint16mf2_t op14, vint16mf2_t op15, vint16mf2_t op16, const int16_t * op17, size_t * op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e16ff_v_i16mf4_m))) +void vlseg8e16ff(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vint16mf4_t * op6, vint16mf4_t * op7, vbool64_t op8, vint16mf4_t op9, vint16mf4_t op10, vint16mf4_t op11, vint16mf4_t op12, vint16mf4_t op13, vint16mf4_t op14, vint16mf4_t op15, vint16mf4_t op16, const int16_t * op17, size_t * op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e16ff_v_u16m1_m))) +void vlseg8e16ff(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vuint16m1_t * op6, vuint16m1_t * op7, vbool16_t op8, vuint16m1_t op9, vuint16m1_t op10, vuint16m1_t op11, vuint16m1_t op12, vuint16m1_t op13, vuint16m1_t op14, vuint16m1_t op15, vuint16m1_t op16, const uint16_t * op17, size_t * op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e16ff_v_u16mf2_m))) +void vlseg8e16ff(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vuint16mf2_t * op6, vuint16mf2_t * op7, vbool32_t op8, vuint16mf2_t op9, vuint16mf2_t op10, vuint16mf2_t op11, vuint16mf2_t op12, vuint16mf2_t op13, vuint16mf2_t op14, vuint16mf2_t op15, vuint16mf2_t op16, const uint16_t * op17, size_t * op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e16ff_v_u16mf4_m))) +void vlseg8e16ff(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vuint16mf4_t * op6, vuint16mf4_t * op7, vbool64_t op8, vuint16mf4_t op9, vuint16mf4_t op10, vuint16mf4_t op11, vuint16mf4_t op12, vuint16mf4_t op13, vuint16mf4_t op14, vuint16mf4_t op15, vuint16mf4_t op16, const uint16_t * op17, size_t * op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e32ff_v_i32m1_m))) +void vlseg2e32ff(vint32m1_t * op0, vint32m1_t * op1, vbool32_t op2, vint32m1_t op3, vint32m1_t op4, const int32_t * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e32ff_v_i32m2_m))) +void vlseg2e32ff(vint32m2_t * op0, vint32m2_t * op1, vbool16_t op2, vint32m2_t op3, vint32m2_t op4, const int32_t * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e32ff_v_i32m4_m))) +void vlseg2e32ff(vint32m4_t * op0, vint32m4_t * op1, vbool8_t op2, vint32m4_t op3, vint32m4_t op4, const int32_t * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e32ff_v_i32mf2_m))) +void vlseg2e32ff(vint32mf2_t * op0, vint32mf2_t * op1, vbool64_t op2, vint32mf2_t op3, vint32mf2_t op4, const int32_t * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e8ff_v_i8m1_m))) +void vlseg2e8ff(vint8m1_t * op0, vint8m1_t * op1, vbool8_t op2, vint8m1_t op3, vint8m1_t op4, const int8_t * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e8ff_v_i8m2_m))) +void vlseg2e8ff(vint8m2_t * op0, vint8m2_t * op1, vbool4_t op2, vint8m2_t op3, vint8m2_t op4, const int8_t * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e8ff_v_i8m4_m))) +void vlseg2e8ff(vint8m4_t * op0, vint8m4_t * op1, vbool2_t op2, vint8m4_t op3, vint8m4_t op4, const int8_t * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e8ff_v_i8mf2_m))) +void vlseg2e8ff(vint8mf2_t * op0, vint8mf2_t * op1, vbool16_t op2, vint8mf2_t op3, vint8mf2_t op4, const int8_t * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e8ff_v_i8mf4_m))) +void vlseg2e8ff(vint8mf4_t * op0, vint8mf4_t * op1, vbool32_t op2, vint8mf4_t op3, vint8mf4_t op4, const int8_t * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e8ff_v_i8mf8_m))) +void vlseg2e8ff(vint8mf8_t * op0, vint8mf8_t * op1, vbool64_t op2, vint8mf8_t op3, vint8mf8_t op4, const int8_t * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e32ff_v_u32m1_m))) +void vlseg2e32ff(vuint32m1_t * op0, vuint32m1_t * op1, vbool32_t op2, vuint32m1_t op3, vuint32m1_t op4, const uint32_t * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e32ff_v_u32m2_m))) +void vlseg2e32ff(vuint32m2_t * op0, vuint32m2_t * op1, vbool16_t op2, vuint32m2_t op3, vuint32m2_t op4, const uint32_t * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e32ff_v_u32m4_m))) +void vlseg2e32ff(vuint32m4_t * op0, vuint32m4_t * op1, vbool8_t op2, vuint32m4_t op3, vuint32m4_t op4, const uint32_t * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e32ff_v_u32mf2_m))) +void vlseg2e32ff(vuint32mf2_t * op0, vuint32mf2_t * op1, vbool64_t op2, vuint32mf2_t op3, vuint32mf2_t op4, const uint32_t * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e32ff_v_i32m1_m))) +void vlseg3e32ff(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vbool32_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, const int32_t * op7, size_t * op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e32ff_v_i32m2_m))) +void vlseg3e32ff(vint32m2_t * op0, vint32m2_t * op1, vint32m2_t * op2, vbool16_t op3, vint32m2_t op4, vint32m2_t op5, vint32m2_t op6, const int32_t * op7, size_t * op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e32ff_v_i32mf2_m))) +void vlseg3e32ff(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vbool64_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, const int32_t * op7, size_t * op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e32ff_v_u32m1_m))) +void vlseg3e32ff(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vbool32_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, const uint32_t * op7, size_t * op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e32ff_v_u32m2_m))) +void vlseg3e32ff(vuint32m2_t * op0, vuint32m2_t * op1, vuint32m2_t * op2, vbool16_t op3, vuint32m2_t op4, vuint32m2_t op5, vuint32m2_t op6, const uint32_t * op7, size_t * op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e32ff_v_u32mf2_m))) +void vlseg3e32ff(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vbool64_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, const uint32_t * op7, size_t * op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e32ff_v_i32m1_m))) +void vlseg4e32ff(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vbool32_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, const int32_t * op9, size_t * op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e32ff_v_i32m2_m))) +void vlseg4e32ff(vint32m2_t * op0, vint32m2_t * op1, vint32m2_t * op2, vint32m2_t * op3, vbool16_t op4, vint32m2_t op5, vint32m2_t op6, vint32m2_t op7, vint32m2_t op8, const int32_t * op9, size_t * op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e32ff_v_i32mf2_m))) +void vlseg4e32ff(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vbool64_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, const int32_t * op9, size_t * op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e32ff_v_u32m1_m))) +void vlseg4e32ff(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vbool32_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, const uint32_t * op9, size_t * op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e32ff_v_u32m2_m))) +void vlseg4e32ff(vuint32m2_t * op0, vuint32m2_t * op1, vuint32m2_t * op2, vuint32m2_t * op3, vbool16_t op4, vuint32m2_t op5, vuint32m2_t op6, vuint32m2_t op7, vuint32m2_t op8, const uint32_t * op9, size_t * op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e32ff_v_u32mf2_m))) +void vlseg4e32ff(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vbool64_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, const uint32_t * op9, size_t * op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e32ff_v_i32m1_m))) +void vlseg5e32ff(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vbool32_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, vint32m1_t op10, const int32_t * op11, size_t * op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e32ff_v_i32mf2_m))) +void vlseg5e32ff(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vbool64_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, vint32mf2_t op10, const int32_t * op11, size_t * op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e32ff_v_u32m1_m))) +void vlseg5e32ff(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vbool32_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, vuint32m1_t op10, const uint32_t * op11, size_t * op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e32ff_v_u32mf2_m))) +void vlseg5e32ff(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vbool64_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, vuint32mf2_t op10, const uint32_t * op11, size_t * op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e32ff_v_i32m1_m))) +void vlseg6e32ff(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vbool32_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, vint32m1_t op10, vint32m1_t op11, vint32m1_t op12, const int32_t * op13, size_t * op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e32ff_v_i32mf2_m))) +void vlseg6e32ff(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vbool64_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, vint32mf2_t op10, vint32mf2_t op11, vint32mf2_t op12, const int32_t * op13, size_t * op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e32ff_v_u32m1_m))) +void vlseg6e32ff(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vbool32_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, vuint32m1_t op10, vuint32m1_t op11, vuint32m1_t op12, const uint32_t * op13, size_t * op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e32ff_v_u32mf2_m))) +void vlseg6e32ff(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vbool64_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, vuint32mf2_t op10, vuint32mf2_t op11, vuint32mf2_t op12, const uint32_t * op13, size_t * op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e32ff_v_i32m1_m))) +void vlseg7e32ff(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vint32m1_t * op6, vbool32_t op7, vint32m1_t op8, vint32m1_t op9, vint32m1_t op10, vint32m1_t op11, vint32m1_t op12, vint32m1_t op13, vint32m1_t op14, const int32_t * op15, size_t * op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e32ff_v_i32mf2_m))) +void vlseg7e32ff(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vint32mf2_t * op6, vbool64_t op7, vint32mf2_t op8, vint32mf2_t op9, vint32mf2_t op10, vint32mf2_t op11, vint32mf2_t op12, vint32mf2_t op13, vint32mf2_t op14, const int32_t * op15, size_t * op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e8ff_v_u8m1_m))) +void vlseg2e8ff(vuint8m1_t * op0, vuint8m1_t * op1, vbool8_t op2, vuint8m1_t op3, vuint8m1_t op4, const uint8_t * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e8ff_v_u8m2_m))) +void vlseg2e8ff(vuint8m2_t * op0, vuint8m2_t * op1, vbool4_t op2, vuint8m2_t op3, vuint8m2_t op4, const uint8_t * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e8ff_v_u8m4_m))) +void vlseg2e8ff(vuint8m4_t * op0, vuint8m4_t * op1, vbool2_t op2, vuint8m4_t op3, vuint8m4_t op4, const uint8_t * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e8ff_v_u8mf2_m))) +void vlseg2e8ff(vuint8mf2_t * op0, vuint8mf2_t * op1, vbool16_t op2, vuint8mf2_t op3, vuint8mf2_t op4, const uint8_t * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e8ff_v_u8mf4_m))) +void vlseg2e8ff(vuint8mf4_t * op0, vuint8mf4_t * op1, vbool32_t op2, vuint8mf4_t op3, vuint8mf4_t op4, const uint8_t * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e8ff_v_u8mf8_m))) +void vlseg2e8ff(vuint8mf8_t * op0, vuint8mf8_t * op1, vbool64_t op2, vuint8mf8_t op3, vuint8mf8_t op4, const uint8_t * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e32ff_v_u32m1_m))) +void vlseg7e32ff(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vuint32m1_t * op6, vbool32_t op7, vuint32m1_t op8, vuint32m1_t op9, vuint32m1_t op10, vuint32m1_t op11, vuint32m1_t op12, vuint32m1_t op13, vuint32m1_t op14, const uint32_t * op15, size_t * op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e32ff_v_u32mf2_m))) +void vlseg7e32ff(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vuint32mf2_t * op6, vbool64_t op7, vuint32mf2_t op8, vuint32mf2_t op9, vuint32mf2_t op10, vuint32mf2_t op11, vuint32mf2_t op12, vuint32mf2_t op13, vuint32mf2_t op14, const uint32_t * op15, size_t * op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e32ff_v_i32m1_m))) +void vlseg8e32ff(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vint32m1_t * op6, vint32m1_t * op7, vbool32_t op8, vint32m1_t op9, vint32m1_t op10, vint32m1_t op11, vint32m1_t op12, vint32m1_t op13, vint32m1_t op14, vint32m1_t op15, vint32m1_t op16, const int32_t * op17, size_t * op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e32ff_v_i32mf2_m))) +void vlseg8e32ff(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vint32mf2_t * op6, vint32mf2_t * op7, vbool64_t op8, vint32mf2_t op9, vint32mf2_t op10, vint32mf2_t op11, vint32mf2_t op12, vint32mf2_t op13, vint32mf2_t op14, vint32mf2_t op15, vint32mf2_t op16, const int32_t * op17, size_t * op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e32ff_v_u32m1_m))) +void vlseg8e32ff(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vuint32m1_t * op6, vuint32m1_t * op7, vbool32_t op8, vuint32m1_t op9, vuint32m1_t op10, vuint32m1_t op11, vuint32m1_t op12, vuint32m1_t op13, vuint32m1_t op14, vuint32m1_t op15, vuint32m1_t op16, const uint32_t * op17, size_t * op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e32ff_v_u32mf2_m))) +void vlseg8e32ff(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vuint32mf2_t * op6, vuint32mf2_t * op7, vbool64_t op8, vuint32mf2_t op9, vuint32mf2_t op10, vuint32mf2_t op11, vuint32mf2_t op12, vuint32mf2_t op13, vuint32mf2_t op14, vuint32mf2_t op15, vuint32mf2_t op16, const uint32_t * op17, size_t * op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e64ff_v_i64m1_m))) +void vlseg2e64ff(vint64m1_t * op0, vint64m1_t * op1, vbool64_t op2, vint64m1_t op3, vint64m1_t op4, const int64_t * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e64ff_v_i64m2_m))) +void vlseg2e64ff(vint64m2_t * op0, vint64m2_t * op1, vbool32_t op2, vint64m2_t op3, vint64m2_t op4, const int64_t * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e64ff_v_i64m4_m))) +void vlseg2e64ff(vint64m4_t * op0, vint64m4_t * op1, vbool16_t op2, vint64m4_t op3, vint64m4_t op4, const int64_t * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e64ff_v_u64m1_m))) +void vlseg2e64ff(vuint64m1_t * op0, vuint64m1_t * op1, vbool64_t op2, vuint64m1_t op3, vuint64m1_t op4, const uint64_t * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e64ff_v_u64m2_m))) +void vlseg2e64ff(vuint64m2_t * op0, vuint64m2_t * op1, vbool32_t op2, vuint64m2_t op3, vuint64m2_t op4, const uint64_t * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e64ff_v_u64m4_m))) +void vlseg2e64ff(vuint64m4_t * op0, vuint64m4_t * op1, vbool16_t op2, vuint64m4_t op3, vuint64m4_t op4, const uint64_t * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e64ff_v_i64m1_m))) +void vlseg3e64ff(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vbool64_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, const int64_t * op7, size_t * op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e64ff_v_i64m2_m))) +void vlseg3e64ff(vint64m2_t * op0, vint64m2_t * op1, vint64m2_t * op2, vbool32_t op3, vint64m2_t op4, vint64m2_t op5, vint64m2_t op6, const int64_t * op7, size_t * op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e64ff_v_u64m1_m))) +void vlseg3e64ff(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vbool64_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, const uint64_t * op7, size_t * op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e64ff_v_u64m2_m))) +void vlseg3e64ff(vuint64m2_t * op0, vuint64m2_t * op1, vuint64m2_t * op2, vbool32_t op3, vuint64m2_t op4, vuint64m2_t op5, vuint64m2_t op6, const uint64_t * op7, size_t * op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e64ff_v_i64m1_m))) +void vlseg4e64ff(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vbool64_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, const int64_t * op9, size_t * op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e64ff_v_i64m2_m))) +void vlseg4e64ff(vint64m2_t * op0, vint64m2_t * op1, vint64m2_t * op2, vint64m2_t * op3, vbool32_t op4, vint64m2_t op5, vint64m2_t op6, vint64m2_t op7, vint64m2_t op8, const int64_t * op9, size_t * op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e64ff_v_u64m1_m))) +void vlseg4e64ff(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vbool64_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, const uint64_t * op9, size_t * op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e64ff_v_u64m2_m))) +void vlseg4e64ff(vuint64m2_t * op0, vuint64m2_t * op1, vuint64m2_t * op2, vuint64m2_t * op3, vbool32_t op4, vuint64m2_t op5, vuint64m2_t op6, vuint64m2_t op7, vuint64m2_t op8, const uint64_t * op9, size_t * op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e64ff_v_i64m1_m))) +void vlseg5e64ff(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vbool64_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, vint64m1_t op10, const int64_t * op11, size_t * op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e8_v_i8m1_m))) +void vlsseg2e8(vint8m1_t * op0, vint8m1_t * op1, vbool8_t op2, vint8m1_t op3, vint8m1_t op4, const int8_t * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e8_v_i8m2_m))) +void vlsseg2e8(vint8m2_t * op0, vint8m2_t * op1, vbool4_t op2, vint8m2_t op3, vint8m2_t op4, const int8_t * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e8_v_i8m4_m))) +void vlsseg2e8(vint8m4_t * op0, vint8m4_t * op1, vbool2_t op2, vint8m4_t op3, vint8m4_t op4, const int8_t * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e8_v_i8mf2_m))) +void vlsseg2e8(vint8mf2_t * op0, vint8mf2_t * op1, vbool16_t op2, vint8mf2_t op3, vint8mf2_t op4, const int8_t * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e8_v_i8mf4_m))) +void vlsseg2e8(vint8mf4_t * op0, vint8mf4_t * op1, vbool32_t op2, vint8mf4_t op3, vint8mf4_t op4, const int8_t * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e8_v_i8mf8_m))) +void vlsseg2e8(vint8mf8_t * op0, vint8mf8_t * op1, vbool64_t op2, vint8mf8_t op3, vint8mf8_t op4, const int8_t * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e64ff_v_u64m1_m))) +void vlseg5e64ff(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vbool64_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, vuint64m1_t op10, const uint64_t * op11, size_t * op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e64ff_v_i64m1_m))) +void vlseg6e64ff(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vbool64_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, vint64m1_t op10, vint64m1_t op11, vint64m1_t op12, const int64_t * op13, size_t * op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e64ff_v_u64m1_m))) +void vlseg6e64ff(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vbool64_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, vuint64m1_t op10, vuint64m1_t op11, vuint64m1_t op12, const uint64_t * op13, size_t * op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e64ff_v_i64m1_m))) +void vlseg7e64ff(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vint64m1_t * op6, vbool64_t op7, vint64m1_t op8, vint64m1_t op9, vint64m1_t op10, vint64m1_t op11, vint64m1_t op12, vint64m1_t op13, vint64m1_t op14, const int64_t * op15, size_t * op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e64ff_v_u64m1_m))) +void vlseg7e64ff(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vuint64m1_t * op6, vbool64_t op7, vuint64m1_t op8, vuint64m1_t op9, vuint64m1_t op10, vuint64m1_t op11, vuint64m1_t op12, vuint64m1_t op13, vuint64m1_t op14, const uint64_t * op15, size_t * op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e64ff_v_i64m1_m))) +void vlseg8e64ff(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vint64m1_t * op6, vint64m1_t * op7, vbool64_t op8, vint64m1_t op9, vint64m1_t op10, vint64m1_t op11, vint64m1_t op12, vint64m1_t op13, vint64m1_t op14, vint64m1_t op15, vint64m1_t op16, const int64_t * op17, size_t * op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e64ff_v_u64m1_m))) +void vlseg8e64ff(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vuint64m1_t * op6, vuint64m1_t * op7, vbool64_t op8, vuint64m1_t op9, vuint64m1_t op10, vuint64m1_t op11, vuint64m1_t op12, vuint64m1_t op13, vuint64m1_t op14, vuint64m1_t op15, vuint64m1_t op16, const uint64_t * op17, size_t * op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e8_v_u8m1_m))) +void vlsseg2e8(vuint8m1_t * op0, vuint8m1_t * op1, vbool8_t op2, vuint8m1_t op3, vuint8m1_t op4, const uint8_t * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e8_v_u8m2_m))) +void vlsseg2e8(vuint8m2_t * op0, vuint8m2_t * op1, vbool4_t op2, vuint8m2_t op3, vuint8m2_t op4, const uint8_t * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e8_v_u8m4_m))) +void vlsseg2e8(vuint8m4_t * op0, vuint8m4_t * op1, vbool2_t op2, vuint8m4_t op3, vuint8m4_t op4, const uint8_t * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e8_v_u8mf2_m))) +void vlsseg2e8(vuint8mf2_t * op0, vuint8mf2_t * op1, vbool16_t op2, vuint8mf2_t op3, vuint8mf2_t op4, const uint8_t * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e8_v_u8mf4_m))) +void vlsseg2e8(vuint8mf4_t * op0, vuint8mf4_t * op1, vbool32_t op2, vuint8mf4_t op3, vuint8mf4_t op4, const uint8_t * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e8_v_u8mf8_m))) +void vlsseg2e8(vuint8mf8_t * op0, vuint8mf8_t * op1, vbool64_t op2, vuint8mf8_t op3, vuint8mf8_t op4, const uint8_t * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_i8m1))) +void vluxseg2ei8(vint8m1_t * op0, vint8m1_t * op1, const int8_t * op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_i8m1_m))) +void vluxseg2ei8(vint8m1_t * op0, vint8m1_t * op1, vbool8_t op2, vint8m1_t op3, vint8m1_t op4, const int8_t * op5, vuint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_i8m2))) +void vluxseg2ei8(vint8m2_t * op0, vint8m2_t * op1, const int8_t * op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_i8m2_m))) +void vluxseg2ei8(vint8m2_t * op0, vint8m2_t * op1, vbool4_t op2, vint8m2_t op3, vint8m2_t op4, const int8_t * op5, vuint8m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_i8m4))) +void vluxseg2ei8(vint8m4_t * op0, vint8m4_t * op1, const int8_t * op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_i8m4_m))) +void vluxseg2ei8(vint8m4_t * op0, vint8m4_t * op1, vbool2_t op2, vint8m4_t op3, vint8m4_t op4, const int8_t * op5, vuint8m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_i8mf2))) +void vluxseg2ei8(vint8mf2_t * op0, vint8mf2_t * op1, const int8_t * op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_i8mf2_m))) +void vluxseg2ei8(vint8mf2_t * op0, vint8mf2_t * op1, vbool16_t op2, vint8mf2_t op3, vint8mf2_t op4, const int8_t * op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_i8mf4))) +void vluxseg2ei8(vint8mf4_t * op0, vint8mf4_t * op1, const int8_t * op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_i8mf4_m))) +void vluxseg2ei8(vint8mf4_t * op0, vint8mf4_t * op1, vbool32_t op2, vint8mf4_t op3, vint8mf4_t op4, const int8_t * op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_i8mf8))) +void vluxseg2ei8(vint8mf8_t * op0, vint8mf8_t * op1, const int8_t * op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_i8mf8_m))) +void vluxseg2ei8(vint8mf8_t * op0, vint8mf8_t * op1, vbool64_t op2, vint8mf8_t op3, vint8mf8_t op4, const int8_t * op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg3e8_v_i8m1_m))) +void vlsseg3e8(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vbool8_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, const int8_t * op7, ptrdiff_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg3e8_v_i8m2_m))) +void vlsseg3e8(vint8m2_t * op0, vint8m2_t * op1, vint8m2_t * op2, vbool4_t op3, vint8m2_t op4, vint8m2_t op5, vint8m2_t op6, const int8_t * op7, ptrdiff_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg3e8_v_i8mf2_m))) +void vlsseg3e8(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vbool16_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, const int8_t * op7, ptrdiff_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg3e8_v_i8mf4_m))) +void vlsseg3e8(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vbool32_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, const int8_t * op7, ptrdiff_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg3e8_v_i8mf8_m))) +void vlsseg3e8(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vbool64_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, const int8_t * op7, ptrdiff_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_u8m1))) +void vluxseg2ei8(vuint8m1_t * op0, vuint8m1_t * op1, const uint8_t * op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_u8m1_m))) +void vluxseg2ei8(vuint8m1_t * op0, vuint8m1_t * op1, vbool8_t op2, vuint8m1_t op3, vuint8m1_t op4, const uint8_t * op5, vuint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_u8m2))) +void vluxseg2ei8(vuint8m2_t * op0, vuint8m2_t * op1, const uint8_t * op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_u8m2_m))) +void vluxseg2ei8(vuint8m2_t * op0, vuint8m2_t * op1, vbool4_t op2, vuint8m2_t op3, vuint8m2_t op4, const uint8_t * op5, vuint8m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_u8m4))) +void vluxseg2ei8(vuint8m4_t * op0, vuint8m4_t * op1, const uint8_t * op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_u8m4_m))) +void vluxseg2ei8(vuint8m4_t * op0, vuint8m4_t * op1, vbool2_t op2, vuint8m4_t op3, vuint8m4_t op4, const uint8_t * op5, vuint8m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_u8mf2))) +void vluxseg2ei8(vuint8mf2_t * op0, vuint8mf2_t * op1, const uint8_t * op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_u8mf2_m))) +void vluxseg2ei8(vuint8mf2_t * op0, vuint8mf2_t * op1, vbool16_t op2, vuint8mf2_t op3, vuint8mf2_t op4, const uint8_t * op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_u8mf4))) +void vluxseg2ei8(vuint8mf4_t * op0, vuint8mf4_t * op1, const uint8_t * op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_u8mf4_m))) +void vluxseg2ei8(vuint8mf4_t * op0, vuint8mf4_t * op1, vbool32_t op2, vuint8mf4_t op3, vuint8mf4_t op4, const uint8_t * op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_u8mf8))) +void vluxseg2ei8(vuint8mf8_t * op0, vuint8mf8_t * op1, const uint8_t * op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_u8mf8_m))) +void vluxseg2ei8(vuint8mf8_t * op0, vuint8mf8_t * op1, vbool64_t op2, vuint8mf8_t op3, vuint8mf8_t op4, const uint8_t * op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg3e8_v_u8m1_m))) +void vlsseg3e8(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vbool8_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, const uint8_t * op7, ptrdiff_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg3e8_v_u8m2_m))) +void vlsseg3e8(vuint8m2_t * op0, vuint8m2_t * op1, vuint8m2_t * op2, vbool4_t op3, vuint8m2_t op4, vuint8m2_t op5, vuint8m2_t op6, const uint8_t * op7, ptrdiff_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg3e8_v_u8mf2_m))) +void vlsseg3e8(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vbool16_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, const uint8_t * op7, ptrdiff_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg3e8_v_u8mf4_m))) +void vlsseg3e8(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vbool32_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, const uint8_t * op7, ptrdiff_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg3e8_v_u8mf8_m))) +void vlsseg3e8(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vbool64_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, const uint8_t * op7, ptrdiff_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg4e8_v_i8m1_m))) +void vlsseg4e8(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vbool8_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, const int8_t * op9, ptrdiff_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg4e8_v_i8m2_m))) +void vlsseg4e8(vint8m2_t * op0, vint8m2_t * op1, vint8m2_t * op2, vint8m2_t * op3, vbool4_t op4, vint8m2_t op5, vint8m2_t op6, vint8m2_t op7, vint8m2_t op8, const int8_t * op9, ptrdiff_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg4e8_v_i8mf2_m))) +void vlsseg4e8(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vbool16_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, const int8_t * op9, ptrdiff_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg4e8_v_i8mf4_m))) +void vlsseg4e8(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vbool32_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, const int8_t * op9, ptrdiff_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg4e8_v_i8mf8_m))) +void vlsseg4e8(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vbool64_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, const int8_t * op9, ptrdiff_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg4e8_v_u8m1_m))) +void vlsseg4e8(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vbool8_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, const uint8_t * op9, ptrdiff_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg4e8_v_u8m2_m))) +void vlsseg4e8(vuint8m2_t * op0, vuint8m2_t * op1, vuint8m2_t * op2, vuint8m2_t * op3, vbool4_t op4, vuint8m2_t op5, vuint8m2_t op6, vuint8m2_t op7, vuint8m2_t op8, const uint8_t * op9, ptrdiff_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg4e8_v_u8mf2_m))) +void vlsseg4e8(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vbool16_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, const uint8_t * op9, ptrdiff_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg4e8_v_u8mf4_m))) +void vlsseg4e8(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vbool32_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, const uint8_t * op9, ptrdiff_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg4e8_v_u8mf8_m))) +void vlsseg4e8(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vbool64_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, const uint8_t * op9, ptrdiff_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg5e8_v_i8m1_m))) +void vlsseg5e8(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vbool8_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, vint8m1_t op10, const int8_t * op11, ptrdiff_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg5e8_v_i8mf2_m))) +void vlsseg5e8(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vbool16_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, vint8mf2_t op10, const int8_t * op11, ptrdiff_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg5e8_v_i8mf4_m))) +void vlsseg5e8(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vbool32_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, vint8mf4_t op10, const int8_t * op11, ptrdiff_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg5e8_v_i8mf8_m))) +void vlsseg5e8(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vbool64_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, vint8mf8_t op10, const int8_t * op11, ptrdiff_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg5e8_v_u8m1_m))) +void vlsseg5e8(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vbool8_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, vuint8m1_t op10, const uint8_t * op11, ptrdiff_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg5e8_v_u8mf2_m))) +void vlsseg5e8(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vbool16_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, vuint8mf2_t op10, const uint8_t * op11, ptrdiff_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg5e8_v_u8mf4_m))) +void vlsseg5e8(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vbool32_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, vuint8mf4_t op10, const uint8_t * op11, ptrdiff_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg5e8_v_u8mf8_m))) +void vlsseg5e8(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vbool64_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, vuint8mf8_t op10, const uint8_t * op11, ptrdiff_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg6e8_v_i8m1_m))) +void vlsseg6e8(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vbool8_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, vint8m1_t op10, vint8m1_t op11, vint8m1_t op12, const int8_t * op13, ptrdiff_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg6e8_v_i8mf2_m))) +void vlsseg6e8(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vbool16_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, vint8mf2_t op10, vint8mf2_t op11, vint8mf2_t op12, const int8_t * op13, ptrdiff_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg6e8_v_i8mf4_m))) +void vlsseg6e8(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vbool32_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, vint8mf4_t op10, vint8mf4_t op11, vint8mf4_t op12, const int8_t * op13, ptrdiff_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg6e8_v_i8mf8_m))) +void vlsseg6e8(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vbool64_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, vint8mf8_t op10, vint8mf8_t op11, vint8mf8_t op12, const int8_t * op13, ptrdiff_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg6e8_v_u8m1_m))) +void vlsseg6e8(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vbool8_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, vuint8m1_t op10, vuint8m1_t op11, vuint8m1_t op12, const uint8_t * op13, ptrdiff_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg6e8_v_u8mf2_m))) +void vlsseg6e8(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vbool16_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, vuint8mf2_t op10, vuint8mf2_t op11, vuint8mf2_t op12, const uint8_t * op13, ptrdiff_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg6e8_v_u8mf4_m))) +void vlsseg6e8(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vbool32_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, vuint8mf4_t op10, vuint8mf4_t op11, vuint8mf4_t op12, const uint8_t * op13, ptrdiff_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg6e8_v_u8mf8_m))) +void vlsseg6e8(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vbool64_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, vuint8mf8_t op10, vuint8mf8_t op11, vuint8mf8_t op12, const uint8_t * op13, ptrdiff_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg7e8_v_i8m1_m))) +void vlsseg7e8(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vint8m1_t * op6, vbool8_t op7, vint8m1_t op8, vint8m1_t op9, vint8m1_t op10, vint8m1_t op11, vint8m1_t op12, vint8m1_t op13, vint8m1_t op14, const int8_t * op15, ptrdiff_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg7e8_v_i8mf2_m))) +void vlsseg7e8(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vint8mf2_t * op6, vbool16_t op7, vint8mf2_t op8, vint8mf2_t op9, vint8mf2_t op10, vint8mf2_t op11, vint8mf2_t op12, vint8mf2_t op13, vint8mf2_t op14, const int8_t * op15, ptrdiff_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg7e8_v_i8mf4_m))) +void vlsseg7e8(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vint8mf4_t * op6, vbool32_t op7, vint8mf4_t op8, vint8mf4_t op9, vint8mf4_t op10, vint8mf4_t op11, vint8mf4_t op12, vint8mf4_t op13, vint8mf4_t op14, const int8_t * op15, ptrdiff_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg7e8_v_i8mf8_m))) +void vlsseg7e8(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vint8mf8_t * op6, vbool64_t op7, vint8mf8_t op8, vint8mf8_t op9, vint8mf8_t op10, vint8mf8_t op11, vint8mf8_t op12, vint8mf8_t op13, vint8mf8_t op14, const int8_t * op15, ptrdiff_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg7e8_v_u8m1_m))) +void vlsseg7e8(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vuint8m1_t * op6, vbool8_t op7, vuint8m1_t op8, vuint8m1_t op9, vuint8m1_t op10, vuint8m1_t op11, vuint8m1_t op12, vuint8m1_t op13, vuint8m1_t op14, const uint8_t * op15, ptrdiff_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg7e8_v_u8mf2_m))) +void vlsseg7e8(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vuint8mf2_t * op6, vbool16_t op7, vuint8mf2_t op8, vuint8mf2_t op9, vuint8mf2_t op10, vuint8mf2_t op11, vuint8mf2_t op12, vuint8mf2_t op13, vuint8mf2_t op14, const uint8_t * op15, ptrdiff_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg7e8_v_u8mf4_m))) +void vlsseg7e8(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vuint8mf4_t * op6, vbool32_t op7, vuint8mf4_t op8, vuint8mf4_t op9, vuint8mf4_t op10, vuint8mf4_t op11, vuint8mf4_t op12, vuint8mf4_t op13, vuint8mf4_t op14, const uint8_t * op15, ptrdiff_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg7e8_v_u8mf8_m))) +void vlsseg7e8(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vuint8mf8_t * op6, vbool64_t op7, vuint8mf8_t op8, vuint8mf8_t op9, vuint8mf8_t op10, vuint8mf8_t op11, vuint8mf8_t op12, vuint8mf8_t op13, vuint8mf8_t op14, const uint8_t * op15, ptrdiff_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg8e8_v_i8m1_m))) +void vlsseg8e8(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vint8m1_t * op6, vint8m1_t * op7, vbool8_t op8, vint8m1_t op9, vint8m1_t op10, vint8m1_t op11, vint8m1_t op12, vint8m1_t op13, vint8m1_t op14, vint8m1_t op15, vint8m1_t op16, const int8_t * op17, ptrdiff_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg8e8_v_i8mf2_m))) +void vlsseg8e8(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vint8mf2_t * op6, vint8mf2_t * op7, vbool16_t op8, vint8mf2_t op9, vint8mf2_t op10, vint8mf2_t op11, vint8mf2_t op12, vint8mf2_t op13, vint8mf2_t op14, vint8mf2_t op15, vint8mf2_t op16, const int8_t * op17, ptrdiff_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg8e8_v_i8mf4_m))) +void vlsseg8e8(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vint8mf4_t * op6, vint8mf4_t * op7, vbool32_t op8, vint8mf4_t op9, vint8mf4_t op10, vint8mf4_t op11, vint8mf4_t op12, vint8mf4_t op13, vint8mf4_t op14, vint8mf4_t op15, vint8mf4_t op16, const int8_t * op17, ptrdiff_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg8e8_v_i8mf8_m))) +void vlsseg8e8(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vint8mf8_t * op6, vint8mf8_t * op7, vbool64_t op8, vint8mf8_t op9, vint8mf8_t op10, vint8mf8_t op11, vint8mf8_t op12, vint8mf8_t op13, vint8mf8_t op14, vint8mf8_t op15, vint8mf8_t op16, const int8_t * op17, ptrdiff_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e8_v_i8m1))) +void vsseg2e8(int8_t * op0, vint8m1_t op1, vint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e8_v_i8m1_m))) +void vsseg2e8(vbool8_t op0, int8_t * op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e8_v_i8m2))) +void vsseg2e8(int8_t * op0, vint8m2_t op1, vint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e8_v_i8m2_m))) +void vsseg2e8(vbool4_t op0, int8_t * op1, vint8m2_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e8_v_i8m4))) +void vsseg2e8(int8_t * op0, vint8m4_t op1, vint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e8_v_i8m4_m))) +void vsseg2e8(vbool2_t op0, int8_t * op1, vint8m4_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e8_v_i8mf2))) +void vsseg2e8(int8_t * op0, vint8mf2_t op1, vint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e8_v_i8mf2_m))) +void vsseg2e8(vbool16_t op0, int8_t * op1, vint8mf2_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e8_v_i8mf4))) +void vsseg2e8(int8_t * op0, vint8mf4_t op1, vint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e8_v_i8mf4_m))) +void vsseg2e8(vbool32_t op0, int8_t * op1, vint8mf4_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e8_v_i8mf8))) +void vsseg2e8(int8_t * op0, vint8mf8_t op1, vint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e8_v_i8mf8_m))) +void vsseg2e8(vbool64_t op0, int8_t * op1, vint8mf8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg8e8_v_u8m1_m))) +void vlsseg8e8(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vuint8m1_t * op6, vuint8m1_t * op7, vbool8_t op8, vuint8m1_t op9, vuint8m1_t op10, vuint8m1_t op11, vuint8m1_t op12, vuint8m1_t op13, vuint8m1_t op14, vuint8m1_t op15, vuint8m1_t op16, const uint8_t * op17, ptrdiff_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg8e8_v_u8mf2_m))) +void vlsseg8e8(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vuint8mf2_t * op6, vuint8mf2_t * op7, vbool16_t op8, vuint8mf2_t op9, vuint8mf2_t op10, vuint8mf2_t op11, vuint8mf2_t op12, vuint8mf2_t op13, vuint8mf2_t op14, vuint8mf2_t op15, vuint8mf2_t op16, const uint8_t * op17, ptrdiff_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg8e8_v_u8mf4_m))) +void vlsseg8e8(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vuint8mf4_t * op6, vuint8mf4_t * op7, vbool32_t op8, vuint8mf4_t op9, vuint8mf4_t op10, vuint8mf4_t op11, vuint8mf4_t op12, vuint8mf4_t op13, vuint8mf4_t op14, vuint8mf4_t op15, vuint8mf4_t op16, const uint8_t * op17, ptrdiff_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg8e8_v_u8mf8_m))) +void vlsseg8e8(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vuint8mf8_t * op6, vuint8mf8_t * op7, vbool64_t op8, vuint8mf8_t op9, vuint8mf8_t op10, vuint8mf8_t op11, vuint8mf8_t op12, vuint8mf8_t op13, vuint8mf8_t op14, vuint8mf8_t op15, vuint8mf8_t op16, const uint8_t * op17, ptrdiff_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e16_v_i16m1_m))) +void vlsseg2e16(vint16m1_t * op0, vint16m1_t * op1, vbool16_t op2, vint16m1_t op3, vint16m1_t op4, const int16_t * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e16_v_i16m2_m))) +void vlsseg2e16(vint16m2_t * op0, vint16m2_t * op1, vbool8_t op2, vint16m2_t op3, vint16m2_t op4, const int16_t * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e16_v_i16m4_m))) +void vlsseg2e16(vint16m4_t * op0, vint16m4_t * op1, vbool4_t op2, vint16m4_t op3, vint16m4_t op4, const int16_t * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e16_v_i16mf2_m))) +void vlsseg2e16(vint16mf2_t * op0, vint16mf2_t * op1, vbool32_t op2, vint16mf2_t op3, vint16mf2_t op4, const int16_t * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e16_v_i16mf4_m))) +void vlsseg2e16(vint16mf4_t * op0, vint16mf4_t * op1, vbool64_t op2, vint16mf4_t op3, vint16mf4_t op4, const int16_t * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e16_v_u16m1_m))) +void vlsseg2e16(vuint16m1_t * op0, vuint16m1_t * op1, vbool16_t op2, vuint16m1_t op3, vuint16m1_t op4, const uint16_t * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e16_v_u16m2_m))) +void vlsseg2e16(vuint16m2_t * op0, vuint16m2_t * op1, vbool8_t op2, vuint16m2_t op3, vuint16m2_t op4, const uint16_t * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e16_v_u16m4_m))) +void vlsseg2e16(vuint16m4_t * op0, vuint16m4_t * op1, vbool4_t op2, vuint16m4_t op3, vuint16m4_t op4, const uint16_t * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e16_v_u16mf2_m))) +void vlsseg2e16(vuint16mf2_t * op0, vuint16mf2_t * op1, vbool32_t op2, vuint16mf2_t op3, vuint16mf2_t op4, const uint16_t * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e16_v_u16mf4_m))) +void vlsseg2e16(vuint16mf4_t * op0, vuint16mf4_t * op1, vbool64_t op2, vuint16mf4_t op3, vuint16mf4_t op4, const uint16_t * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg3e16_v_i16m1_m))) +void vlsseg3e16(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vbool16_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, const int16_t * op7, ptrdiff_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg3e16_v_i16m2_m))) +void vlsseg3e16(vint16m2_t * op0, vint16m2_t * op1, vint16m2_t * op2, vbool8_t op3, vint16m2_t op4, vint16m2_t op5, vint16m2_t op6, const int16_t * op7, ptrdiff_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg3e16_v_i16mf2_m))) +void vlsseg3e16(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vbool32_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, const int16_t * op7, ptrdiff_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg3e16_v_i16mf4_m))) +void vlsseg3e16(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vbool64_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, const int16_t * op7, ptrdiff_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg3e16_v_u16m1_m))) +void vlsseg3e16(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vbool16_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, const uint16_t * op7, ptrdiff_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg3e16_v_u16m2_m))) +void vlsseg3e16(vuint16m2_t * op0, vuint16m2_t * op1, vuint16m2_t * op2, vbool8_t op3, vuint16m2_t op4, vuint16m2_t op5, vuint16m2_t op6, const uint16_t * op7, ptrdiff_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg3e16_v_u16mf2_m))) +void vlsseg3e16(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vbool32_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, const uint16_t * op7, ptrdiff_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg3e16_v_u16mf4_m))) +void vlsseg3e16(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vbool64_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, const uint16_t * op7, ptrdiff_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg4e16_v_i16m1_m))) +void vlsseg4e16(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vbool16_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, const int16_t * op9, ptrdiff_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg4e16_v_i16m2_m))) +void vlsseg4e16(vint16m2_t * op0, vint16m2_t * op1, vint16m2_t * op2, vint16m2_t * op3, vbool8_t op4, vint16m2_t op5, vint16m2_t op6, vint16m2_t op7, vint16m2_t op8, const int16_t * op9, ptrdiff_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg4e16_v_i16mf2_m))) +void vlsseg4e16(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vbool32_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, const int16_t * op9, ptrdiff_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg4e16_v_i16mf4_m))) +void vlsseg4e16(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vbool64_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, const int16_t * op9, ptrdiff_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg4e16_v_u16m1_m))) +void vlsseg4e16(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vbool16_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, const uint16_t * op9, ptrdiff_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg4e16_v_u16m2_m))) +void vlsseg4e16(vuint16m2_t * op0, vuint16m2_t * op1, vuint16m2_t * op2, vuint16m2_t * op3, vbool8_t op4, vuint16m2_t op5, vuint16m2_t op6, vuint16m2_t op7, vuint16m2_t op8, const uint16_t * op9, ptrdiff_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg4e16_v_u16mf2_m))) +void vlsseg4e16(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vbool32_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, const uint16_t * op9, ptrdiff_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg4e16_v_u16mf4_m))) +void vlsseg4e16(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vbool64_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, const uint16_t * op9, ptrdiff_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg5e16_v_i16m1_m))) +void vlsseg5e16(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vbool16_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, vint16m1_t op10, const int16_t * op11, ptrdiff_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg5e16_v_i16mf2_m))) +void vlsseg5e16(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vbool32_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, vint16mf2_t op10, const int16_t * op11, ptrdiff_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg5e16_v_i16mf4_m))) +void vlsseg5e16(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vbool64_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, vint16mf4_t op10, const int16_t * op11, ptrdiff_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg5e16_v_u16m1_m))) +void vlsseg5e16(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vbool16_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, vuint16m1_t op10, const uint16_t * op11, ptrdiff_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg5e16_v_u16mf2_m))) +void vlsseg5e16(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vbool32_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, vuint16mf2_t op10, const uint16_t * op11, ptrdiff_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg5e16_v_u16mf4_m))) +void vlsseg5e16(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vbool64_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, vuint16mf4_t op10, const uint16_t * op11, ptrdiff_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg6e16_v_i16m1_m))) +void vlsseg6e16(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vbool16_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, vint16m1_t op10, vint16m1_t op11, vint16m1_t op12, const int16_t * op13, ptrdiff_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg6e16_v_i16mf2_m))) +void vlsseg6e16(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vbool32_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, vint16mf2_t op10, vint16mf2_t op11, vint16mf2_t op12, const int16_t * op13, ptrdiff_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg6e16_v_i16mf4_m))) +void vlsseg6e16(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vbool64_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, vint16mf4_t op10, vint16mf4_t op11, vint16mf4_t op12, const int16_t * op13, ptrdiff_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e8_v_u8m1))) +void vsseg2e8(uint8_t * op0, vuint8m1_t op1, vuint8m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e8_v_u8m1_m))) +void vsseg2e8(vbool8_t op0, uint8_t * op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e8_v_u8m2))) +void vsseg2e8(uint8_t * op0, vuint8m2_t op1, vuint8m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e8_v_u8m2_m))) +void vsseg2e8(vbool4_t op0, uint8_t * op1, vuint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e8_v_u8m4))) +void vsseg2e8(uint8_t * op0, vuint8m4_t op1, vuint8m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e8_v_u8m4_m))) +void vsseg2e8(vbool2_t op0, uint8_t * op1, vuint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e8_v_u8mf2))) +void vsseg2e8(uint8_t * op0, vuint8mf2_t op1, vuint8mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e8_v_u8mf2_m))) +void vsseg2e8(vbool16_t op0, uint8_t * op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e8_v_u8mf4))) +void vsseg2e8(uint8_t * op0, vuint8mf4_t op1, vuint8mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e8_v_u8mf4_m))) +void vsseg2e8(vbool32_t op0, uint8_t * op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e8_v_u8mf8))) +void vsseg2e8(uint8_t * op0, vuint8mf8_t op1, vuint8mf8_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e8_v_u8mf8_m))) +void vsseg2e8(vbool64_t op0, uint8_t * op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg6e16_v_u16m1_m))) +void vlsseg6e16(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vbool16_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, vuint16m1_t op10, vuint16m1_t op11, vuint16m1_t op12, const uint16_t * op13, ptrdiff_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg6e16_v_u16mf2_m))) +void vlsseg6e16(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vbool32_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, vuint16mf2_t op10, vuint16mf2_t op11, vuint16mf2_t op12, const uint16_t * op13, ptrdiff_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg6e16_v_u16mf4_m))) +void vlsseg6e16(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vbool64_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, vuint16mf4_t op10, vuint16mf4_t op11, vuint16mf4_t op12, const uint16_t * op13, ptrdiff_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg7e16_v_i16m1_m))) +void vlsseg7e16(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vint16m1_t * op6, vbool16_t op7, vint16m1_t op8, vint16m1_t op9, vint16m1_t op10, vint16m1_t op11, vint16m1_t op12, vint16m1_t op13, vint16m1_t op14, const int16_t * op15, ptrdiff_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg7e16_v_i16mf2_m))) +void vlsseg7e16(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vint16mf2_t * op6, vbool32_t op7, vint16mf2_t op8, vint16mf2_t op9, vint16mf2_t op10, vint16mf2_t op11, vint16mf2_t op12, vint16mf2_t op13, vint16mf2_t op14, const int16_t * op15, ptrdiff_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg7e16_v_i16mf4_m))) +void vlsseg7e16(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vint16mf4_t * op6, vbool64_t op7, vint16mf4_t op8, vint16mf4_t op9, vint16mf4_t op10, vint16mf4_t op11, vint16mf4_t op12, vint16mf4_t op13, vint16mf4_t op14, const int16_t * op15, ptrdiff_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg7e16_v_u16m1_m))) +void vlsseg7e16(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vuint16m1_t * op6, vbool16_t op7, vuint16m1_t op8, vuint16m1_t op9, vuint16m1_t op10, vuint16m1_t op11, vuint16m1_t op12, vuint16m1_t op13, vuint16m1_t op14, const uint16_t * op15, ptrdiff_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg7e16_v_u16mf2_m))) +void vlsseg7e16(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vuint16mf2_t * op6, vbool32_t op7, vuint16mf2_t op8, vuint16mf2_t op9, vuint16mf2_t op10, vuint16mf2_t op11, vuint16mf2_t op12, vuint16mf2_t op13, vuint16mf2_t op14, const uint16_t * op15, ptrdiff_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg7e16_v_u16mf4_m))) +void vlsseg7e16(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vuint16mf4_t * op6, vbool64_t op7, vuint16mf4_t op8, vuint16mf4_t op9, vuint16mf4_t op10, vuint16mf4_t op11, vuint16mf4_t op12, vuint16mf4_t op13, vuint16mf4_t op14, const uint16_t * op15, ptrdiff_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg8e16_v_i16m1_m))) +void vlsseg8e16(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vint16m1_t * op6, vint16m1_t * op7, vbool16_t op8, vint16m1_t op9, vint16m1_t op10, vint16m1_t op11, vint16m1_t op12, vint16m1_t op13, vint16m1_t op14, vint16m1_t op15, vint16m1_t op16, const int16_t * op17, ptrdiff_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg8e16_v_i16mf2_m))) +void vlsseg8e16(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vint16mf2_t * op6, vint16mf2_t * op7, vbool32_t op8, vint16mf2_t op9, vint16mf2_t op10, vint16mf2_t op11, vint16mf2_t op12, vint16mf2_t op13, vint16mf2_t op14, vint16mf2_t op15, vint16mf2_t op16, const int16_t * op17, ptrdiff_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg8e16_v_i16mf4_m))) +void vlsseg8e16(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vint16mf4_t * op6, vint16mf4_t * op7, vbool64_t op8, vint16mf4_t op9, vint16mf4_t op10, vint16mf4_t op11, vint16mf4_t op12, vint16mf4_t op13, vint16mf4_t op14, vint16mf4_t op15, vint16mf4_t op16, const int16_t * op17, ptrdiff_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg8e16_v_u16m1_m))) +void vlsseg8e16(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vuint16m1_t * op6, vuint16m1_t * op7, vbool16_t op8, vuint16m1_t op9, vuint16m1_t op10, vuint16m1_t op11, vuint16m1_t op12, vuint16m1_t op13, vuint16m1_t op14, vuint16m1_t op15, vuint16m1_t op16, const uint16_t * op17, ptrdiff_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg8e16_v_u16mf2_m))) +void vlsseg8e16(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vuint16mf2_t * op6, vuint16mf2_t * op7, vbool32_t op8, vuint16mf2_t op9, vuint16mf2_t op10, vuint16mf2_t op11, vuint16mf2_t op12, vuint16mf2_t op13, vuint16mf2_t op14, vuint16mf2_t op15, vuint16mf2_t op16, const uint16_t * op17, ptrdiff_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg8e16_v_u16mf4_m))) +void vlsseg8e16(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vuint16mf4_t * op6, vuint16mf4_t * op7, vbool64_t op8, vuint16mf4_t op9, vuint16mf4_t op10, vuint16mf4_t op11, vuint16mf4_t op12, vuint16mf4_t op13, vuint16mf4_t op14, vuint16mf4_t op15, vuint16mf4_t op16, const uint16_t * op17, ptrdiff_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e32_v_i32m1_m))) +void vlsseg2e32(vint32m1_t * op0, vint32m1_t * op1, vbool32_t op2, vint32m1_t op3, vint32m1_t op4, const int32_t * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e32_v_i32m2_m))) +void vlsseg2e32(vint32m2_t * op0, vint32m2_t * op1, vbool16_t op2, vint32m2_t op3, vint32m2_t op4, const int32_t * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e32_v_i32m4_m))) +void vlsseg2e32(vint32m4_t * op0, vint32m4_t * op1, vbool8_t op2, vint32m4_t op3, vint32m4_t op4, const int32_t * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e32_v_i32mf2_m))) +void vlsseg2e32(vint32mf2_t * op0, vint32mf2_t * op1, vbool64_t op2, vint32mf2_t op3, vint32mf2_t op4, const int32_t * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e32_v_u32m1_m))) +void vlsseg2e32(vuint32m1_t * op0, vuint32m1_t * op1, vbool32_t op2, vuint32m1_t op3, vuint32m1_t op4, const uint32_t * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e32_v_u32m2_m))) +void vlsseg2e32(vuint32m2_t * op0, vuint32m2_t * op1, vbool16_t op2, vuint32m2_t op3, vuint32m2_t op4, const uint32_t * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e32_v_u32m4_m))) +void vlsseg2e32(vuint32m4_t * op0, vuint32m4_t * op1, vbool8_t op2, vuint32m4_t op3, vuint32m4_t op4, const uint32_t * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e32_v_u32mf2_m))) +void vlsseg2e32(vuint32mf2_t * op0, vuint32mf2_t * op1, vbool64_t op2, vuint32mf2_t op3, vuint32mf2_t op4, const uint32_t * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg3e32_v_i32m1_m))) +void vlsseg3e32(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vbool32_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, const int32_t * op7, ptrdiff_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg3e32_v_i32m2_m))) +void vlsseg3e32(vint32m2_t * op0, vint32m2_t * op1, vint32m2_t * op2, vbool16_t op3, vint32m2_t op4, vint32m2_t op5, vint32m2_t op6, const int32_t * op7, ptrdiff_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg3e32_v_i32mf2_m))) +void vlsseg3e32(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vbool64_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, const int32_t * op7, ptrdiff_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg3e32_v_u32m1_m))) +void vlsseg3e32(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vbool32_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, const uint32_t * op7, ptrdiff_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg3e32_v_u32m2_m))) +void vlsseg3e32(vuint32m2_t * op0, vuint32m2_t * op1, vuint32m2_t * op2, vbool16_t op3, vuint32m2_t op4, vuint32m2_t op5, vuint32m2_t op6, const uint32_t * op7, ptrdiff_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg3e32_v_u32mf2_m))) +void vlsseg3e32(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vbool64_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, const uint32_t * op7, ptrdiff_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg4e32_v_i32m1_m))) +void vlsseg4e32(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vbool32_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, const int32_t * op9, ptrdiff_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg4e32_v_i32m2_m))) +void vlsseg4e32(vint32m2_t * op0, vint32m2_t * op1, vint32m2_t * op2, vint32m2_t * op3, vbool16_t op4, vint32m2_t op5, vint32m2_t op6, vint32m2_t op7, vint32m2_t op8, const int32_t * op9, ptrdiff_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg4e32_v_i32mf2_m))) +void vlsseg4e32(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vbool64_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, const int32_t * op9, ptrdiff_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg4e32_v_u32m1_m))) +void vlsseg4e32(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vbool32_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, const uint32_t * op9, ptrdiff_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg4e32_v_u32m2_m))) +void vlsseg4e32(vuint32m2_t * op0, vuint32m2_t * op1, vuint32m2_t * op2, vuint32m2_t * op3, vbool16_t op4, vuint32m2_t op5, vuint32m2_t op6, vuint32m2_t op7, vuint32m2_t op8, const uint32_t * op9, ptrdiff_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg4e32_v_u32mf2_m))) +void vlsseg4e32(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vbool64_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, const uint32_t * op9, ptrdiff_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg5e32_v_i32m1_m))) +void vlsseg5e32(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vbool32_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, vint32m1_t op10, const int32_t * op11, ptrdiff_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg5e32_v_i32mf2_m))) +void vlsseg5e32(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vbool64_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, vint32mf2_t op10, const int32_t * op11, ptrdiff_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg5e32_v_u32m1_m))) +void vlsseg5e32(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vbool32_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, vuint32m1_t op10, const uint32_t * op11, ptrdiff_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg5e32_v_u32mf2_m))) +void vlsseg5e32(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vbool64_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, vuint32mf2_t op10, const uint32_t * op11, ptrdiff_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg6e32_v_i32m1_m))) +void vlsseg6e32(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vbool32_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, vint32m1_t op10, vint32m1_t op11, vint32m1_t op12, const int32_t * op13, ptrdiff_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg6e32_v_i32mf2_m))) +void vlsseg6e32(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vbool64_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, vint32mf2_t op10, vint32mf2_t op11, vint32mf2_t op12, const int32_t * op13, ptrdiff_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg6e32_v_u32m1_m))) +void vlsseg6e32(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vbool32_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, vuint32m1_t op10, vuint32m1_t op11, vuint32m1_t op12, const uint32_t * op13, ptrdiff_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg6e32_v_u32mf2_m))) +void vlsseg6e32(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vbool64_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, vuint32mf2_t op10, vuint32mf2_t op11, vuint32mf2_t op12, const uint32_t * op13, ptrdiff_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg7e32_v_i32m1_m))) +void vlsseg7e32(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vint32m1_t * op6, vbool32_t op7, vint32m1_t op8, vint32m1_t op9, vint32m1_t op10, vint32m1_t op11, vint32m1_t op12, vint32m1_t op13, vint32m1_t op14, const int32_t * op15, ptrdiff_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg7e32_v_i32mf2_m))) +void vlsseg7e32(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vint32mf2_t * op6, vbool64_t op7, vint32mf2_t op8, vint32mf2_t op9, vint32mf2_t op10, vint32mf2_t op11, vint32mf2_t op12, vint32mf2_t op13, vint32mf2_t op14, const int32_t * op15, ptrdiff_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg7e32_v_u32m1_m))) +void vlsseg7e32(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vuint32m1_t * op6, vbool32_t op7, vuint32m1_t op8, vuint32m1_t op9, vuint32m1_t op10, vuint32m1_t op11, vuint32m1_t op12, vuint32m1_t op13, vuint32m1_t op14, const uint32_t * op15, ptrdiff_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg7e32_v_u32mf2_m))) +void vlsseg7e32(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vuint32mf2_t * op6, vbool64_t op7, vuint32mf2_t op8, vuint32mf2_t op9, vuint32mf2_t op10, vuint32mf2_t op11, vuint32mf2_t op12, vuint32mf2_t op13, vuint32mf2_t op14, const uint32_t * op15, ptrdiff_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg8e32_v_i32m1_m))) +void vlsseg8e32(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vint32m1_t * op6, vint32m1_t * op7, vbool32_t op8, vint32m1_t op9, vint32m1_t op10, vint32m1_t op11, vint32m1_t op12, vint32m1_t op13, vint32m1_t op14, vint32m1_t op15, vint32m1_t op16, const int32_t * op17, ptrdiff_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg8e32_v_i32mf2_m))) +void vlsseg8e32(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vint32mf2_t * op6, vint32mf2_t * op7, vbool64_t op8, vint32mf2_t op9, vint32mf2_t op10, vint32mf2_t op11, vint32mf2_t op12, vint32mf2_t op13, vint32mf2_t op14, vint32mf2_t op15, vint32mf2_t op16, const int32_t * op17, ptrdiff_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg8e32_v_u32m1_m))) +void vlsseg8e32(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vuint32m1_t * op6, vuint32m1_t * op7, vbool32_t op8, vuint32m1_t op9, vuint32m1_t op10, vuint32m1_t op11, vuint32m1_t op12, vuint32m1_t op13, vuint32m1_t op14, vuint32m1_t op15, vuint32m1_t op16, const uint32_t * op17, ptrdiff_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg8e32_v_u32mf2_m))) +void vlsseg8e32(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vuint32mf2_t * op6, vuint32mf2_t * op7, vbool64_t op8, vuint32mf2_t op9, vuint32mf2_t op10, vuint32mf2_t op11, vuint32mf2_t op12, vuint32mf2_t op13, vuint32mf2_t op14, vuint32mf2_t op15, vuint32mf2_t op16, const uint32_t * op17, ptrdiff_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e64_v_i64m1_m))) +void vlsseg2e64(vint64m1_t * op0, vint64m1_t * op1, vbool64_t op2, vint64m1_t op3, vint64m1_t op4, const int64_t * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e64_v_i64m2_m))) +void vlsseg2e64(vint64m2_t * op0, vint64m2_t * op1, vbool32_t op2, vint64m2_t op3, vint64m2_t op4, const int64_t * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e64_v_i64m4_m))) +void vlsseg2e64(vint64m4_t * op0, vint64m4_t * op1, vbool16_t op2, vint64m4_t op3, vint64m4_t op4, const int64_t * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e64_v_u64m1_m))) +void vlsseg2e64(vuint64m1_t * op0, vuint64m1_t * op1, vbool64_t op2, vuint64m1_t op3, vuint64m1_t op4, const uint64_t * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e64_v_u64m2_m))) +void vlsseg2e64(vuint64m2_t * op0, vuint64m2_t * op1, vbool32_t op2, vuint64m2_t op3, vuint64m2_t op4, const uint64_t * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e64_v_u64m4_m))) +void vlsseg2e64(vuint64m4_t * op0, vuint64m4_t * op1, vbool16_t op2, vuint64m4_t op3, vuint64m4_t op4, const uint64_t * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg3e64_v_i64m1_m))) +void vlsseg3e64(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vbool64_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, const int64_t * op7, ptrdiff_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg3e64_v_i64m2_m))) +void vlsseg3e64(vint64m2_t * op0, vint64m2_t * op1, vint64m2_t * op2, vbool32_t op3, vint64m2_t op4, vint64m2_t op5, vint64m2_t op6, const int64_t * op7, ptrdiff_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg3e64_v_u64m1_m))) +void vlsseg3e64(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vbool64_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, const uint64_t * op7, ptrdiff_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg3e64_v_u64m2_m))) +void vlsseg3e64(vuint64m2_t * op0, vuint64m2_t * op1, vuint64m2_t * op2, vbool32_t op3, vuint64m2_t op4, vuint64m2_t op5, vuint64m2_t op6, const uint64_t * op7, ptrdiff_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg4e64_v_i64m1_m))) +void vlsseg4e64(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vbool64_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, const int64_t * op9, ptrdiff_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg4e64_v_i64m2_m))) +void vlsseg4e64(vint64m2_t * op0, vint64m2_t * op1, vint64m2_t * op2, vint64m2_t * op3, vbool32_t op4, vint64m2_t op5, vint64m2_t op6, vint64m2_t op7, vint64m2_t op8, const int64_t * op9, ptrdiff_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg4e64_v_u64m1_m))) +void vlsseg4e64(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vbool64_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, const uint64_t * op9, ptrdiff_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg4e64_v_u64m2_m))) +void vlsseg4e64(vuint64m2_t * op0, vuint64m2_t * op1, vuint64m2_t * op2, vuint64m2_t * op3, vbool32_t op4, vuint64m2_t op5, vuint64m2_t op6, vuint64m2_t op7, vuint64m2_t op8, const uint64_t * op9, ptrdiff_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg5e64_v_i64m1_m))) +void vlsseg5e64(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vbool64_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, vint64m1_t op10, const int64_t * op11, ptrdiff_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg5e64_v_u64m1_m))) +void vlsseg5e64(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vbool64_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, vuint64m1_t op10, const uint64_t * op11, ptrdiff_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg6e64_v_i64m1_m))) +void vlsseg6e64(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vbool64_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, vint64m1_t op10, vint64m1_t op11, vint64m1_t op12, const int64_t * op13, ptrdiff_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg6e64_v_u64m1_m))) +void vlsseg6e64(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vbool64_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, vuint64m1_t op10, vuint64m1_t op11, vuint64m1_t op12, const uint64_t * op13, ptrdiff_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg7e64_v_i64m1_m))) +void vlsseg7e64(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vint64m1_t * op6, vbool64_t op7, vint64m1_t op8, vint64m1_t op9, vint64m1_t op10, vint64m1_t op11, vint64m1_t op12, vint64m1_t op13, vint64m1_t op14, const int64_t * op15, ptrdiff_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg7e64_v_u64m1_m))) +void vlsseg7e64(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vuint64m1_t * op6, vbool64_t op7, vuint64m1_t op8, vuint64m1_t op9, vuint64m1_t op10, vuint64m1_t op11, vuint64m1_t op12, vuint64m1_t op13, vuint64m1_t op14, const uint64_t * op15, ptrdiff_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg8e64_v_i64m1_m))) +void vlsseg8e64(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vint64m1_t * op6, vint64m1_t * op7, vbool64_t op8, vint64m1_t op9, vint64m1_t op10, vint64m1_t op11, vint64m1_t op12, vint64m1_t op13, vint64m1_t op14, vint64m1_t op15, vint64m1_t op16, const int64_t * op17, ptrdiff_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg8e64_v_u64m1_m))) +void vlsseg8e64(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vuint64m1_t * op6, vuint64m1_t * op7, vbool64_t op8, vuint64m1_t op9, vuint64m1_t op10, vuint64m1_t op11, vuint64m1_t op12, vuint64m1_t op13, vuint64m1_t op14, vuint64m1_t op15, vuint64m1_t op16, const uint64_t * op17, ptrdiff_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_i8m1))) +void vluxseg3ei8(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, const int8_t * op3, vuint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_i8m1_m))) +void vluxseg3ei8(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vbool8_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, const int8_t * op7, vuint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_i8m2))) +void vluxseg3ei8(vint8m2_t * op0, vint8m2_t * op1, vint8m2_t * op2, const int8_t * op3, vuint8m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_i8m2_m))) +void vluxseg3ei8(vint8m2_t * op0, vint8m2_t * op1, vint8m2_t * op2, vbool4_t op3, vint8m2_t op4, vint8m2_t op5, vint8m2_t op6, const int8_t * op7, vuint8m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_i8mf2))) +void vluxseg3ei8(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, const int8_t * op3, vuint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_i8mf2_m))) +void vluxseg3ei8(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vbool16_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, const int8_t * op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_i8mf4))) +void vluxseg3ei8(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, const int8_t * op3, vuint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_i8mf4_m))) +void vluxseg3ei8(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vbool32_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, const int8_t * op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_i8mf8))) +void vluxseg3ei8(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, const int8_t * op3, vuint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_i8mf8_m))) +void vluxseg3ei8(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vbool64_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, const int8_t * op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_u8m1))) +void vluxseg3ei8(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, const uint8_t * op3, vuint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_u8m1_m))) +void vluxseg3ei8(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vbool8_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, const uint8_t * op7, vuint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_u8m2))) +void vluxseg3ei8(vuint8m2_t * op0, vuint8m2_t * op1, vuint8m2_t * op2, const uint8_t * op3, vuint8m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_u8m2_m))) +void vluxseg3ei8(vuint8m2_t * op0, vuint8m2_t * op1, vuint8m2_t * op2, vbool4_t op3, vuint8m2_t op4, vuint8m2_t op5, vuint8m2_t op6, const uint8_t * op7, vuint8m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_u8mf2))) +void vluxseg3ei8(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, const uint8_t * op3, vuint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_u8mf2_m))) +void vluxseg3ei8(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vbool16_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, const uint8_t * op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_u8mf4))) +void vluxseg3ei8(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, const uint8_t * op3, vuint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_u8mf4_m))) +void vluxseg3ei8(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vbool32_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, const uint8_t * op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_u8mf8))) +void vluxseg3ei8(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, const uint8_t * op3, vuint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_u8mf8_m))) +void vluxseg3ei8(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vbool64_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, const uint8_t * op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_i8m1))) +void vluxseg4ei8(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, const int8_t * op4, vuint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_i8m1_m))) +void vluxseg4ei8(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vbool8_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, const int8_t * op9, vuint8m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_i8m2))) +void vluxseg4ei8(vint8m2_t * op0, vint8m2_t * op1, vint8m2_t * op2, vint8m2_t * op3, const int8_t * op4, vuint8m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_i8m2_m))) +void vluxseg4ei8(vint8m2_t * op0, vint8m2_t * op1, vint8m2_t * op2, vint8m2_t * op3, vbool4_t op4, vint8m2_t op5, vint8m2_t op6, vint8m2_t op7, vint8m2_t op8, const int8_t * op9, vuint8m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_i8mf2))) +void vluxseg4ei8(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, const int8_t * op4, vuint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_i8mf2_m))) +void vluxseg4ei8(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vbool16_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, const int8_t * op9, vuint8mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_i8mf4))) +void vluxseg4ei8(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, const int8_t * op4, vuint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_i8mf4_m))) +void vluxseg4ei8(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vbool32_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, const int8_t * op9, vuint8mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_i8mf8))) +void vluxseg4ei8(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, const int8_t * op4, vuint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_i8mf8_m))) +void vluxseg4ei8(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vbool64_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, const int8_t * op9, vuint8mf8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_u8m1))) +void vluxseg4ei8(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, const uint8_t * op4, vuint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_u8m1_m))) +void vluxseg4ei8(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vbool8_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, const uint8_t * op9, vuint8m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_u8m2))) +void vluxseg4ei8(vuint8m2_t * op0, vuint8m2_t * op1, vuint8m2_t * op2, vuint8m2_t * op3, const uint8_t * op4, vuint8m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_u8m2_m))) +void vluxseg4ei8(vuint8m2_t * op0, vuint8m2_t * op1, vuint8m2_t * op2, vuint8m2_t * op3, vbool4_t op4, vuint8m2_t op5, vuint8m2_t op6, vuint8m2_t op7, vuint8m2_t op8, const uint8_t * op9, vuint8m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_u8mf2))) +void vluxseg4ei8(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, const uint8_t * op4, vuint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_u8mf2_m))) +void vluxseg4ei8(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vbool16_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, const uint8_t * op9, vuint8mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_u8mf4))) +void vluxseg4ei8(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, const uint8_t * op4, vuint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_u8mf4_m))) +void vluxseg4ei8(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vbool32_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, const uint8_t * op9, vuint8mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_u8mf8))) +void vluxseg4ei8(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, const uint8_t * op4, vuint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_u8mf8_m))) +void vluxseg4ei8(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vbool64_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, const uint8_t * op9, vuint8mf8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_i8m1))) +void vluxseg5ei8(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, const int8_t * op5, vuint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_i8m1_m))) +void vluxseg5ei8(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vbool8_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, vint8m1_t op10, const int8_t * op11, vuint8m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_i8mf2))) +void vluxseg5ei8(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, const int8_t * op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_i8mf2_m))) +void vluxseg5ei8(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vbool16_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, vint8mf2_t op10, const int8_t * op11, vuint8mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_i8mf4))) +void vluxseg5ei8(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, const int8_t * op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_i8mf4_m))) +void vluxseg5ei8(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vbool32_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, vint8mf4_t op10, const int8_t * op11, vuint8mf4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_i8mf8))) +void vluxseg5ei8(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, const int8_t * op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_i8mf8_m))) +void vluxseg5ei8(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vbool64_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, vint8mf8_t op10, const int8_t * op11, vuint8mf8_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_u8m1))) +void vluxseg5ei8(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, const uint8_t * op5, vuint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_u8m1_m))) +void vluxseg5ei8(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vbool8_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, vuint8m1_t op10, const uint8_t * op11, vuint8m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_u8mf2))) +void vluxseg5ei8(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, const uint8_t * op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_u8mf2_m))) +void vluxseg5ei8(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vbool16_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, vuint8mf2_t op10, const uint8_t * op11, vuint8mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_u8mf4))) +void vluxseg5ei8(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, const uint8_t * op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_u8mf4_m))) +void vluxseg5ei8(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vbool32_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, vuint8mf4_t op10, const uint8_t * op11, vuint8mf4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_u8mf8))) +void vluxseg5ei8(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, const uint8_t * op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_u8mf8_m))) +void vluxseg5ei8(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vbool64_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, vuint8mf8_t op10, const uint8_t * op11, vuint8mf8_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_i8m1))) +void vluxseg6ei8(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, const int8_t * op6, vuint8m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_i8m1_m))) +void vluxseg6ei8(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vbool8_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, vint8m1_t op10, vint8m1_t op11, vint8m1_t op12, const int8_t * op13, vuint8m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_i8mf2))) +void vluxseg6ei8(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, const int8_t * op6, vuint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_i8mf2_m))) +void vluxseg6ei8(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vbool16_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, vint8mf2_t op10, vint8mf2_t op11, vint8mf2_t op12, const int8_t * op13, vuint8mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_i8mf4))) +void vluxseg6ei8(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, const int8_t * op6, vuint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_i8mf4_m))) +void vluxseg6ei8(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vbool32_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, vint8mf4_t op10, vint8mf4_t op11, vint8mf4_t op12, const int8_t * op13, vuint8mf4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_i8mf8))) +void vluxseg6ei8(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, const int8_t * op6, vuint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_i8mf8_m))) +void vluxseg6ei8(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vbool64_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, vint8mf8_t op10, vint8mf8_t op11, vint8mf8_t op12, const int8_t * op13, vuint8mf8_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_u8m1))) +void vluxseg6ei8(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, const uint8_t * op6, vuint8m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_u8m1_m))) +void vluxseg6ei8(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vbool8_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, vuint8m1_t op10, vuint8m1_t op11, vuint8m1_t op12, const uint8_t * op13, vuint8m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_u8mf2))) +void vluxseg6ei8(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, const uint8_t * op6, vuint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_u8mf2_m))) +void vluxseg6ei8(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vbool16_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, vuint8mf2_t op10, vuint8mf2_t op11, vuint8mf2_t op12, const uint8_t * op13, vuint8mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_u8mf4))) +void vluxseg6ei8(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, const uint8_t * op6, vuint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_u8mf4_m))) +void vluxseg6ei8(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vbool32_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, vuint8mf4_t op10, vuint8mf4_t op11, vuint8mf4_t op12, const uint8_t * op13, vuint8mf4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_u8mf8))) +void vluxseg6ei8(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, const uint8_t * op6, vuint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_u8mf8_m))) +void vluxseg6ei8(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vbool64_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, vuint8mf8_t op10, vuint8mf8_t op11, vuint8mf8_t op12, const uint8_t * op13, vuint8mf8_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_i8m1))) +void vluxseg7ei8(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vint8m1_t * op6, const int8_t * op7, vuint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_i8m1_m))) +void vluxseg7ei8(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vint8m1_t * op6, vbool8_t op7, vint8m1_t op8, vint8m1_t op9, vint8m1_t op10, vint8m1_t op11, vint8m1_t op12, vint8m1_t op13, vint8m1_t op14, const int8_t * op15, vuint8m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_i8mf2))) +void vluxseg7ei8(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vint8mf2_t * op6, const int8_t * op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_i8mf2_m))) +void vluxseg7ei8(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vint8mf2_t * op6, vbool16_t op7, vint8mf2_t op8, vint8mf2_t op9, vint8mf2_t op10, vint8mf2_t op11, vint8mf2_t op12, vint8mf2_t op13, vint8mf2_t op14, const int8_t * op15, vuint8mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_i8mf4))) +void vluxseg7ei8(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vint8mf4_t * op6, const int8_t * op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_i8mf4_m))) +void vluxseg7ei8(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vint8mf4_t * op6, vbool32_t op7, vint8mf4_t op8, vint8mf4_t op9, vint8mf4_t op10, vint8mf4_t op11, vint8mf4_t op12, vint8mf4_t op13, vint8mf4_t op14, const int8_t * op15, vuint8mf4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_i8mf8))) +void vluxseg7ei8(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vint8mf8_t * op6, const int8_t * op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_i8mf8_m))) +void vluxseg7ei8(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vint8mf8_t * op6, vbool64_t op7, vint8mf8_t op8, vint8mf8_t op9, vint8mf8_t op10, vint8mf8_t op11, vint8mf8_t op12, vint8mf8_t op13, vint8mf8_t op14, const int8_t * op15, vuint8mf8_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_u8m1))) +void vluxseg7ei8(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vuint8m1_t * op6, const uint8_t * op7, vuint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_u8m1_m))) +void vluxseg7ei8(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vuint8m1_t * op6, vbool8_t op7, vuint8m1_t op8, vuint8m1_t op9, vuint8m1_t op10, vuint8m1_t op11, vuint8m1_t op12, vuint8m1_t op13, vuint8m1_t op14, const uint8_t * op15, vuint8m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_u8mf2))) +void vluxseg7ei8(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vuint8mf2_t * op6, const uint8_t * op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_u8mf2_m))) +void vluxseg7ei8(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vuint8mf2_t * op6, vbool16_t op7, vuint8mf2_t op8, vuint8mf2_t op9, vuint8mf2_t op10, vuint8mf2_t op11, vuint8mf2_t op12, vuint8mf2_t op13, vuint8mf2_t op14, const uint8_t * op15, vuint8mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_u8mf4))) +void vluxseg7ei8(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vuint8mf4_t * op6, const uint8_t * op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_u8mf4_m))) +void vluxseg7ei8(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vuint8mf4_t * op6, vbool32_t op7, vuint8mf4_t op8, vuint8mf4_t op9, vuint8mf4_t op10, vuint8mf4_t op11, vuint8mf4_t op12, vuint8mf4_t op13, vuint8mf4_t op14, const uint8_t * op15, vuint8mf4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_u8mf8))) +void vluxseg7ei8(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vuint8mf8_t * op6, const uint8_t * op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_u8mf8_m))) +void vluxseg7ei8(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vuint8mf8_t * op6, vbool64_t op7, vuint8mf8_t op8, vuint8mf8_t op9, vuint8mf8_t op10, vuint8mf8_t op11, vuint8mf8_t op12, vuint8mf8_t op13, vuint8mf8_t op14, const uint8_t * op15, vuint8mf8_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_i8m1))) +void vluxseg8ei8(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vint8m1_t * op6, vint8m1_t * op7, const int8_t * op8, vuint8m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_i8m1_m))) +void vluxseg8ei8(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vint8m1_t * op6, vint8m1_t * op7, vbool8_t op8, vint8m1_t op9, vint8m1_t op10, vint8m1_t op11, vint8m1_t op12, vint8m1_t op13, vint8m1_t op14, vint8m1_t op15, vint8m1_t op16, const int8_t * op17, vuint8m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_i8mf2))) +void vluxseg8ei8(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vint8mf2_t * op6, vint8mf2_t * op7, const int8_t * op8, vuint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_i8mf2_m))) +void vluxseg8ei8(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vint8mf2_t * op6, vint8mf2_t * op7, vbool16_t op8, vint8mf2_t op9, vint8mf2_t op10, vint8mf2_t op11, vint8mf2_t op12, vint8mf2_t op13, vint8mf2_t op14, vint8mf2_t op15, vint8mf2_t op16, const int8_t * op17, vuint8mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_i8mf4))) +void vluxseg8ei8(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vint8mf4_t * op6, vint8mf4_t * op7, const int8_t * op8, vuint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_i8mf4_m))) +void vluxseg8ei8(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vint8mf4_t * op6, vint8mf4_t * op7, vbool32_t op8, vint8mf4_t op9, vint8mf4_t op10, vint8mf4_t op11, vint8mf4_t op12, vint8mf4_t op13, vint8mf4_t op14, vint8mf4_t op15, vint8mf4_t op16, const int8_t * op17, vuint8mf4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_i8mf8))) +void vluxseg8ei8(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vint8mf8_t * op6, vint8mf8_t * op7, const int8_t * op8, vuint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_i8mf8_m))) +void vluxseg8ei8(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vint8mf8_t * op6, vint8mf8_t * op7, vbool64_t op8, vint8mf8_t op9, vint8mf8_t op10, vint8mf8_t op11, vint8mf8_t op12, vint8mf8_t op13, vint8mf8_t op14, vint8mf8_t op15, vint8mf8_t op16, const int8_t * op17, vuint8mf8_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_u8m1))) +void vluxseg8ei8(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vuint8m1_t * op6, vuint8m1_t * op7, const uint8_t * op8, vuint8m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_u8m1_m))) +void vluxseg8ei8(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vuint8m1_t * op6, vuint8m1_t * op7, vbool8_t op8, vuint8m1_t op9, vuint8m1_t op10, vuint8m1_t op11, vuint8m1_t op12, vuint8m1_t op13, vuint8m1_t op14, vuint8m1_t op15, vuint8m1_t op16, const uint8_t * op17, vuint8m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_u8mf2))) +void vluxseg8ei8(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vuint8mf2_t * op6, vuint8mf2_t * op7, const uint8_t * op8, vuint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_u8mf2_m))) +void vluxseg8ei8(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vuint8mf2_t * op6, vuint8mf2_t * op7, vbool16_t op8, vuint8mf2_t op9, vuint8mf2_t op10, vuint8mf2_t op11, vuint8mf2_t op12, vuint8mf2_t op13, vuint8mf2_t op14, vuint8mf2_t op15, vuint8mf2_t op16, const uint8_t * op17, vuint8mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_u8mf4))) +void vluxseg8ei8(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vuint8mf4_t * op6, vuint8mf4_t * op7, const uint8_t * op8, vuint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_u8mf4_m))) +void vluxseg8ei8(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vuint8mf4_t * op6, vuint8mf4_t * op7, vbool32_t op8, vuint8mf4_t op9, vuint8mf4_t op10, vuint8mf4_t op11, vuint8mf4_t op12, vuint8mf4_t op13, vuint8mf4_t op14, vuint8mf4_t op15, vuint8mf4_t op16, const uint8_t * op17, vuint8mf4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_u8mf8))) +void vluxseg8ei8(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vuint8mf8_t * op6, vuint8mf8_t * op7, const uint8_t * op8, vuint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_u8mf8_m))) +void vluxseg8ei8(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vuint8mf8_t * op6, vuint8mf8_t * op7, vbool64_t op8, vuint8mf8_t op9, vuint8mf8_t op10, vuint8mf8_t op11, vuint8mf8_t op12, vuint8mf8_t op13, vuint8mf8_t op14, vuint8mf8_t op15, vuint8mf8_t op16, const uint8_t * op17, vuint8mf8_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_i8m1))) +void vluxseg2ei16(vint8m1_t * op0, vint8m1_t * op1, const int8_t * op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_i8m1_m))) +void vluxseg2ei16(vint8m1_t * op0, vint8m1_t * op1, vbool8_t op2, vint8m1_t op3, vint8m1_t op4, const int8_t * op5, vuint16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_i8m2))) +void vluxseg2ei16(vint8m2_t * op0, vint8m2_t * op1, const int8_t * op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_i8m2_m))) +void vluxseg2ei16(vint8m2_t * op0, vint8m2_t * op1, vbool4_t op2, vint8m2_t op3, vint8m2_t op4, const int8_t * op5, vuint16m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_i8m4))) +void vluxseg2ei16(vint8m4_t * op0, vint8m4_t * op1, const int8_t * op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_i8m4_m))) +void vluxseg2ei16(vint8m4_t * op0, vint8m4_t * op1, vbool2_t op2, vint8m4_t op3, vint8m4_t op4, const int8_t * op5, vuint16m8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_i8mf2))) +void vluxseg2ei16(vint8mf2_t * op0, vint8mf2_t * op1, const int8_t * op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_i8mf2_m))) +void vluxseg2ei16(vint8mf2_t * op0, vint8mf2_t * op1, vbool16_t op2, vint8mf2_t op3, vint8mf2_t op4, const int8_t * op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_i8mf4))) +void vluxseg2ei16(vint8mf4_t * op0, vint8mf4_t * op1, const int8_t * op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_i8mf4_m))) +void vluxseg2ei16(vint8mf4_t * op0, vint8mf4_t * op1, vbool32_t op2, vint8mf4_t op3, vint8mf4_t op4, const int8_t * op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_i8mf8))) +void vluxseg2ei16(vint8mf8_t * op0, vint8mf8_t * op1, const int8_t * op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_i8mf8_m))) +void vluxseg2ei16(vint8mf8_t * op0, vint8mf8_t * op1, vbool64_t op2, vint8mf8_t op3, vint8mf8_t op4, const int8_t * op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_u8m1))) +void vluxseg2ei16(vuint8m1_t * op0, vuint8m1_t * op1, const uint8_t * op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_u8m1_m))) +void vluxseg2ei16(vuint8m1_t * op0, vuint8m1_t * op1, vbool8_t op2, vuint8m1_t op3, vuint8m1_t op4, const uint8_t * op5, vuint16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_u8m2))) +void vluxseg2ei16(vuint8m2_t * op0, vuint8m2_t * op1, const uint8_t * op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_u8m2_m))) +void vluxseg2ei16(vuint8m2_t * op0, vuint8m2_t * op1, vbool4_t op2, vuint8m2_t op3, vuint8m2_t op4, const uint8_t * op5, vuint16m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_u8m4))) +void vluxseg2ei16(vuint8m4_t * op0, vuint8m4_t * op1, const uint8_t * op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_u8m4_m))) +void vluxseg2ei16(vuint8m4_t * op0, vuint8m4_t * op1, vbool2_t op2, vuint8m4_t op3, vuint8m4_t op4, const uint8_t * op5, vuint16m8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_u8mf2))) +void vluxseg2ei16(vuint8mf2_t * op0, vuint8mf2_t * op1, const uint8_t * op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_u8mf2_m))) +void vluxseg2ei16(vuint8mf2_t * op0, vuint8mf2_t * op1, vbool16_t op2, vuint8mf2_t op3, vuint8mf2_t op4, const uint8_t * op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_u8mf4))) +void vluxseg2ei16(vuint8mf4_t * op0, vuint8mf4_t * op1, const uint8_t * op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_u8mf4_m))) +void vluxseg2ei16(vuint8mf4_t * op0, vuint8mf4_t * op1, vbool32_t op2, vuint8mf4_t op3, vuint8mf4_t op4, const uint8_t * op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_u8mf8))) +void vluxseg2ei16(vuint8mf8_t * op0, vuint8mf8_t * op1, const uint8_t * op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_u8mf8_m))) +void vluxseg2ei16(vuint8mf8_t * op0, vuint8mf8_t * op1, vbool64_t op2, vuint8mf8_t op3, vuint8mf8_t op4, const uint8_t * op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_i8m1))) +void vluxseg3ei16(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, const int8_t * op3, vuint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_i8m1_m))) +void vluxseg3ei16(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vbool8_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, const int8_t * op7, vuint16m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_i8m2))) +void vluxseg3ei16(vint8m2_t * op0, vint8m2_t * op1, vint8m2_t * op2, const int8_t * op3, vuint16m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_i8m2_m))) +void vluxseg3ei16(vint8m2_t * op0, vint8m2_t * op1, vint8m2_t * op2, vbool4_t op3, vint8m2_t op4, vint8m2_t op5, vint8m2_t op6, const int8_t * op7, vuint16m4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_i8mf2))) +void vluxseg3ei16(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, const int8_t * op3, vuint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_i8mf2_m))) +void vluxseg3ei16(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vbool16_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, const int8_t * op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_i8mf4))) +void vluxseg3ei16(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, const int8_t * op3, vuint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_i8mf4_m))) +void vluxseg3ei16(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vbool32_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, const int8_t * op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_i8mf8))) +void vluxseg3ei16(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, const int8_t * op3, vuint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_i8mf8_m))) +void vluxseg3ei16(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vbool64_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, const int8_t * op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_u8m1))) +void vluxseg3ei16(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, const uint8_t * op3, vuint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_u8m1_m))) +void vluxseg3ei16(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vbool8_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, const uint8_t * op7, vuint16m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_u8m2))) +void vluxseg3ei16(vuint8m2_t * op0, vuint8m2_t * op1, vuint8m2_t * op2, const uint8_t * op3, vuint16m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_u8m2_m))) +void vluxseg3ei16(vuint8m2_t * op0, vuint8m2_t * op1, vuint8m2_t * op2, vbool4_t op3, vuint8m2_t op4, vuint8m2_t op5, vuint8m2_t op6, const uint8_t * op7, vuint16m4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_u8mf2))) +void vluxseg3ei16(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, const uint8_t * op3, vuint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_u8mf2_m))) +void vluxseg3ei16(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vbool16_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, const uint8_t * op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_u8mf4))) +void vluxseg3ei16(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, const uint8_t * op3, vuint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_u8mf4_m))) +void vluxseg3ei16(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vbool32_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, const uint8_t * op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_u8mf8))) +void vluxseg3ei16(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, const uint8_t * op3, vuint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_u8mf8_m))) +void vluxseg3ei16(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vbool64_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, const uint8_t * op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_i8m1))) +void vluxseg4ei16(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, const int8_t * op4, vuint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_i8m1_m))) +void vluxseg4ei16(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vbool8_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, const int8_t * op9, vuint16m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_i8m2))) +void vluxseg4ei16(vint8m2_t * op0, vint8m2_t * op1, vint8m2_t * op2, vint8m2_t * op3, const int8_t * op4, vuint16m4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_i8m2_m))) +void vluxseg4ei16(vint8m2_t * op0, vint8m2_t * op1, vint8m2_t * op2, vint8m2_t * op3, vbool4_t op4, vint8m2_t op5, vint8m2_t op6, vint8m2_t op7, vint8m2_t op8, const int8_t * op9, vuint16m4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_i8mf2))) +void vluxseg4ei16(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, const int8_t * op4, vuint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_i8mf2_m))) +void vluxseg4ei16(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vbool16_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, const int8_t * op9, vuint16m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_i8mf4))) +void vluxseg4ei16(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, const int8_t * op4, vuint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_i8mf4_m))) +void vluxseg4ei16(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vbool32_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, const int8_t * op9, vuint16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_i8mf8))) +void vluxseg4ei16(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, const int8_t * op4, vuint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_i8mf8_m))) +void vluxseg4ei16(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vbool64_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, const int8_t * op9, vuint16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_u8m1))) +void vluxseg4ei16(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, const uint8_t * op4, vuint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_u8m1_m))) +void vluxseg4ei16(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vbool8_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, const uint8_t * op9, vuint16m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_u8m2))) +void vluxseg4ei16(vuint8m2_t * op0, vuint8m2_t * op1, vuint8m2_t * op2, vuint8m2_t * op3, const uint8_t * op4, vuint16m4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_u8m2_m))) +void vluxseg4ei16(vuint8m2_t * op0, vuint8m2_t * op1, vuint8m2_t * op2, vuint8m2_t * op3, vbool4_t op4, vuint8m2_t op5, vuint8m2_t op6, vuint8m2_t op7, vuint8m2_t op8, const uint8_t * op9, vuint16m4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_u8mf2))) +void vluxseg4ei16(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, const uint8_t * op4, vuint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_u8mf2_m))) +void vluxseg4ei16(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vbool16_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, const uint8_t * op9, vuint16m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_u8mf4))) +void vluxseg4ei16(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, const uint8_t * op4, vuint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_u8mf4_m))) +void vluxseg4ei16(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vbool32_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, const uint8_t * op9, vuint16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_u8mf8))) +void vluxseg4ei16(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, const uint8_t * op4, vuint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_u8mf8_m))) +void vluxseg4ei16(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vbool64_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, const uint8_t * op9, vuint16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_i8m1))) +void vluxseg5ei16(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, const int8_t * op5, vuint16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_i8m1_m))) +void vluxseg5ei16(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vbool8_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, vint8m1_t op10, const int8_t * op11, vuint16m2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_i8mf2))) +void vluxseg5ei16(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, const int8_t * op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_i8mf2_m))) +void vluxseg5ei16(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vbool16_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, vint8mf2_t op10, const int8_t * op11, vuint16m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_i8mf4))) +void vluxseg5ei16(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, const int8_t * op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_i8mf4_m))) +void vluxseg5ei16(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vbool32_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, vint8mf4_t op10, const int8_t * op11, vuint16mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_i8mf8))) +void vluxseg5ei16(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, const int8_t * op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_i8mf8_m))) +void vluxseg5ei16(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vbool64_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, vint8mf8_t op10, const int8_t * op11, vuint16mf4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_u8m1))) +void vluxseg5ei16(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, const uint8_t * op5, vuint16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_u8m1_m))) +void vluxseg5ei16(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vbool8_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, vuint8m1_t op10, const uint8_t * op11, vuint16m2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_u8mf2))) +void vluxseg5ei16(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, const uint8_t * op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_u8mf2_m))) +void vluxseg5ei16(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vbool16_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, vuint8mf2_t op10, const uint8_t * op11, vuint16m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_u8mf4))) +void vluxseg5ei16(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, const uint8_t * op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_u8mf4_m))) +void vluxseg5ei16(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vbool32_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, vuint8mf4_t op10, const uint8_t * op11, vuint16mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_u8mf8))) +void vluxseg5ei16(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, const uint8_t * op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_u8mf8_m))) +void vluxseg5ei16(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vbool64_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, vuint8mf8_t op10, const uint8_t * op11, vuint16mf4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_i8m1))) +void vluxseg6ei16(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, const int8_t * op6, vuint16m2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_i8m1_m))) +void vluxseg6ei16(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vbool8_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, vint8m1_t op10, vint8m1_t op11, vint8m1_t op12, const int8_t * op13, vuint16m2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_i8mf2))) +void vluxseg6ei16(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, const int8_t * op6, vuint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_i8mf2_m))) +void vluxseg6ei16(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vbool16_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, vint8mf2_t op10, vint8mf2_t op11, vint8mf2_t op12, const int8_t * op13, vuint16m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_i8mf4))) +void vluxseg6ei16(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, const int8_t * op6, vuint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_i8mf4_m))) +void vluxseg6ei16(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vbool32_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, vint8mf4_t op10, vint8mf4_t op11, vint8mf4_t op12, const int8_t * op13, vuint16mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_i8mf8))) +void vluxseg6ei16(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, const int8_t * op6, vuint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_i8mf8_m))) +void vluxseg6ei16(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vbool64_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, vint8mf8_t op10, vint8mf8_t op11, vint8mf8_t op12, const int8_t * op13, vuint16mf4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_u8m1))) +void vluxseg6ei16(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, const uint8_t * op6, vuint16m2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_u8m1_m))) +void vluxseg6ei16(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vbool8_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, vuint8m1_t op10, vuint8m1_t op11, vuint8m1_t op12, const uint8_t * op13, vuint16m2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_u8mf2))) +void vluxseg6ei16(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, const uint8_t * op6, vuint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_u8mf2_m))) +void vluxseg6ei16(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vbool16_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, vuint8mf2_t op10, vuint8mf2_t op11, vuint8mf2_t op12, const uint8_t * op13, vuint16m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_u8mf4))) +void vluxseg6ei16(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, const uint8_t * op6, vuint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_u8mf4_m))) +void vluxseg6ei16(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vbool32_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, vuint8mf4_t op10, vuint8mf4_t op11, vuint8mf4_t op12, const uint8_t * op13, vuint16mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_u8mf8))) +void vluxseg6ei16(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, const uint8_t * op6, vuint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_u8mf8_m))) +void vluxseg6ei16(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vbool64_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, vuint8mf8_t op10, vuint8mf8_t op11, vuint8mf8_t op12, const uint8_t * op13, vuint16mf4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_i8m1))) +void vluxseg7ei16(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vint8m1_t * op6, const int8_t * op7, vuint16m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_i8m1_m))) +void vluxseg7ei16(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vint8m1_t * op6, vbool8_t op7, vint8m1_t op8, vint8m1_t op9, vint8m1_t op10, vint8m1_t op11, vint8m1_t op12, vint8m1_t op13, vint8m1_t op14, const int8_t * op15, vuint16m2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_i8mf2))) +void vluxseg7ei16(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vint8mf2_t * op6, const int8_t * op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_i8mf2_m))) +void vluxseg7ei16(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vint8mf2_t * op6, vbool16_t op7, vint8mf2_t op8, vint8mf2_t op9, vint8mf2_t op10, vint8mf2_t op11, vint8mf2_t op12, vint8mf2_t op13, vint8mf2_t op14, const int8_t * op15, vuint16m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_i8mf4))) +void vluxseg7ei16(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vint8mf4_t * op6, const int8_t * op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_i8mf4_m))) +void vluxseg7ei16(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vint8mf4_t * op6, vbool32_t op7, vint8mf4_t op8, vint8mf4_t op9, vint8mf4_t op10, vint8mf4_t op11, vint8mf4_t op12, vint8mf4_t op13, vint8mf4_t op14, const int8_t * op15, vuint16mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_i8mf8))) +void vluxseg7ei16(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vint8mf8_t * op6, const int8_t * op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_i8mf8_m))) +void vluxseg7ei16(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vint8mf8_t * op6, vbool64_t op7, vint8mf8_t op8, vint8mf8_t op9, vint8mf8_t op10, vint8mf8_t op11, vint8mf8_t op12, vint8mf8_t op13, vint8mf8_t op14, const int8_t * op15, vuint16mf4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_u8m1))) +void vluxseg7ei16(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vuint8m1_t * op6, const uint8_t * op7, vuint16m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_u8m1_m))) +void vluxseg7ei16(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vuint8m1_t * op6, vbool8_t op7, vuint8m1_t op8, vuint8m1_t op9, vuint8m1_t op10, vuint8m1_t op11, vuint8m1_t op12, vuint8m1_t op13, vuint8m1_t op14, const uint8_t * op15, vuint16m2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_u8mf2))) +void vluxseg7ei16(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vuint8mf2_t * op6, const uint8_t * op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_u8mf2_m))) +void vluxseg7ei16(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vuint8mf2_t * op6, vbool16_t op7, vuint8mf2_t op8, vuint8mf2_t op9, vuint8mf2_t op10, vuint8mf2_t op11, vuint8mf2_t op12, vuint8mf2_t op13, vuint8mf2_t op14, const uint8_t * op15, vuint16m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_u8mf4))) +void vluxseg7ei16(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vuint8mf4_t * op6, const uint8_t * op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_u8mf4_m))) +void vluxseg7ei16(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vuint8mf4_t * op6, vbool32_t op7, vuint8mf4_t op8, vuint8mf4_t op9, vuint8mf4_t op10, vuint8mf4_t op11, vuint8mf4_t op12, vuint8mf4_t op13, vuint8mf4_t op14, const uint8_t * op15, vuint16mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_u8mf8))) +void vluxseg7ei16(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vuint8mf8_t * op6, const uint8_t * op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_u8mf8_m))) +void vluxseg7ei16(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vuint8mf8_t * op6, vbool64_t op7, vuint8mf8_t op8, vuint8mf8_t op9, vuint8mf8_t op10, vuint8mf8_t op11, vuint8mf8_t op12, vuint8mf8_t op13, vuint8mf8_t op14, const uint8_t * op15, vuint16mf4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_i8m1))) +void vluxseg8ei16(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vint8m1_t * op6, vint8m1_t * op7, const int8_t * op8, vuint16m2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_i8m1_m))) +void vluxseg8ei16(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vint8m1_t * op6, vint8m1_t * op7, vbool8_t op8, vint8m1_t op9, vint8m1_t op10, vint8m1_t op11, vint8m1_t op12, vint8m1_t op13, vint8m1_t op14, vint8m1_t op15, vint8m1_t op16, const int8_t * op17, vuint16m2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_i8mf2))) +void vluxseg8ei16(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vint8mf2_t * op6, vint8mf2_t * op7, const int8_t * op8, vuint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_i8mf2_m))) +void vluxseg8ei16(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vint8mf2_t * op6, vint8mf2_t * op7, vbool16_t op8, vint8mf2_t op9, vint8mf2_t op10, vint8mf2_t op11, vint8mf2_t op12, vint8mf2_t op13, vint8mf2_t op14, vint8mf2_t op15, vint8mf2_t op16, const int8_t * op17, vuint16m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_i8mf4))) +void vluxseg8ei16(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vint8mf4_t * op6, vint8mf4_t * op7, const int8_t * op8, vuint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_i8mf4_m))) +void vluxseg8ei16(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vint8mf4_t * op6, vint8mf4_t * op7, vbool32_t op8, vint8mf4_t op9, vint8mf4_t op10, vint8mf4_t op11, vint8mf4_t op12, vint8mf4_t op13, vint8mf4_t op14, vint8mf4_t op15, vint8mf4_t op16, const int8_t * op17, vuint16mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_i8mf8))) +void vluxseg8ei16(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vint8mf8_t * op6, vint8mf8_t * op7, const int8_t * op8, vuint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_i8mf8_m))) +void vluxseg8ei16(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vint8mf8_t * op6, vint8mf8_t * op7, vbool64_t op8, vint8mf8_t op9, vint8mf8_t op10, vint8mf8_t op11, vint8mf8_t op12, vint8mf8_t op13, vint8mf8_t op14, vint8mf8_t op15, vint8mf8_t op16, const int8_t * op17, vuint16mf4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_u8m1))) +void vluxseg8ei16(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vuint8m1_t * op6, vuint8m1_t * op7, const uint8_t * op8, vuint16m2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_u8m1_m))) +void vluxseg8ei16(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vuint8m1_t * op6, vuint8m1_t * op7, vbool8_t op8, vuint8m1_t op9, vuint8m1_t op10, vuint8m1_t op11, vuint8m1_t op12, vuint8m1_t op13, vuint8m1_t op14, vuint8m1_t op15, vuint8m1_t op16, const uint8_t * op17, vuint16m2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_u8mf2))) +void vluxseg8ei16(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vuint8mf2_t * op6, vuint8mf2_t * op7, const uint8_t * op8, vuint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_u8mf2_m))) +void vluxseg8ei16(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vuint8mf2_t * op6, vuint8mf2_t * op7, vbool16_t op8, vuint8mf2_t op9, vuint8mf2_t op10, vuint8mf2_t op11, vuint8mf2_t op12, vuint8mf2_t op13, vuint8mf2_t op14, vuint8mf2_t op15, vuint8mf2_t op16, const uint8_t * op17, vuint16m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_u8mf4))) +void vluxseg8ei16(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vuint8mf4_t * op6, vuint8mf4_t * op7, const uint8_t * op8, vuint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_u8mf4_m))) +void vluxseg8ei16(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vuint8mf4_t * op6, vuint8mf4_t * op7, vbool32_t op8, vuint8mf4_t op9, vuint8mf4_t op10, vuint8mf4_t op11, vuint8mf4_t op12, vuint8mf4_t op13, vuint8mf4_t op14, vuint8mf4_t op15, vuint8mf4_t op16, const uint8_t * op17, vuint16mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_u8mf8))) +void vluxseg8ei16(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vuint8mf8_t * op6, vuint8mf8_t * op7, const uint8_t * op8, vuint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_u8mf8_m))) +void vluxseg8ei16(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vuint8mf8_t * op6, vuint8mf8_t * op7, vbool64_t op8, vuint8mf8_t op9, vuint8mf8_t op10, vuint8mf8_t op11, vuint8mf8_t op12, vuint8mf8_t op13, vuint8mf8_t op14, vuint8mf8_t op15, vuint8mf8_t op16, const uint8_t * op17, vuint16mf4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_i8m1))) +void vluxseg2ei32(vint8m1_t * op0, vint8m1_t * op1, const int8_t * op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_i8m1_m))) +void vluxseg2ei32(vint8m1_t * op0, vint8m1_t * op1, vbool8_t op2, vint8m1_t op3, vint8m1_t op4, const int8_t * op5, vuint32m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_i8m2))) +void vluxseg2ei32(vint8m2_t * op0, vint8m2_t * op1, const int8_t * op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_i8m2_m))) +void vluxseg2ei32(vint8m2_t * op0, vint8m2_t * op1, vbool4_t op2, vint8m2_t op3, vint8m2_t op4, const int8_t * op5, vuint32m8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_i8mf2))) +void vluxseg2ei32(vint8mf2_t * op0, vint8mf2_t * op1, const int8_t * op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_i8mf2_m))) +void vluxseg2ei32(vint8mf2_t * op0, vint8mf2_t * op1, vbool16_t op2, vint8mf2_t op3, vint8mf2_t op4, const int8_t * op5, vuint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_i8mf4))) +void vluxseg2ei32(vint8mf4_t * op0, vint8mf4_t * op1, const int8_t * op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_i8mf4_m))) +void vluxseg2ei32(vint8mf4_t * op0, vint8mf4_t * op1, vbool32_t op2, vint8mf4_t op3, vint8mf4_t op4, const int8_t * op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_i8mf8))) +void vluxseg2ei32(vint8mf8_t * op0, vint8mf8_t * op1, const int8_t * op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_i8mf8_m))) +void vluxseg2ei32(vint8mf8_t * op0, vint8mf8_t * op1, vbool64_t op2, vint8mf8_t op3, vint8mf8_t op4, const int8_t * op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_u8m1))) +void vluxseg2ei32(vuint8m1_t * op0, vuint8m1_t * op1, const uint8_t * op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_u8m1_m))) +void vluxseg2ei32(vuint8m1_t * op0, vuint8m1_t * op1, vbool8_t op2, vuint8m1_t op3, vuint8m1_t op4, const uint8_t * op5, vuint32m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_u8m2))) +void vluxseg2ei32(vuint8m2_t * op0, vuint8m2_t * op1, const uint8_t * op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_u8m2_m))) +void vluxseg2ei32(vuint8m2_t * op0, vuint8m2_t * op1, vbool4_t op2, vuint8m2_t op3, vuint8m2_t op4, const uint8_t * op5, vuint32m8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_u8mf2))) +void vluxseg2ei32(vuint8mf2_t * op0, vuint8mf2_t * op1, const uint8_t * op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_u8mf2_m))) +void vluxseg2ei32(vuint8mf2_t * op0, vuint8mf2_t * op1, vbool16_t op2, vuint8mf2_t op3, vuint8mf2_t op4, const uint8_t * op5, vuint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_u8mf4))) +void vluxseg2ei32(vuint8mf4_t * op0, vuint8mf4_t * op1, const uint8_t * op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_u8mf4_m))) +void vluxseg2ei32(vuint8mf4_t * op0, vuint8mf4_t * op1, vbool32_t op2, vuint8mf4_t op3, vuint8mf4_t op4, const uint8_t * op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_u8mf8))) +void vluxseg2ei32(vuint8mf8_t * op0, vuint8mf8_t * op1, const uint8_t * op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_u8mf8_m))) +void vluxseg2ei32(vuint8mf8_t * op0, vuint8mf8_t * op1, vbool64_t op2, vuint8mf8_t op3, vuint8mf8_t op4, const uint8_t * op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_i8m1))) +void vluxseg3ei32(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, const int8_t * op3, vuint32m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_i8m1_m))) +void vluxseg3ei32(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vbool8_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, const int8_t * op7, vuint32m4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_i8m2))) +void vluxseg3ei32(vint8m2_t * op0, vint8m2_t * op1, vint8m2_t * op2, const int8_t * op3, vuint32m8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_i8m2_m))) +void vluxseg3ei32(vint8m2_t * op0, vint8m2_t * op1, vint8m2_t * op2, vbool4_t op3, vint8m2_t op4, vint8m2_t op5, vint8m2_t op6, const int8_t * op7, vuint32m8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_i8mf2))) +void vluxseg3ei32(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, const int8_t * op3, vuint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_i8mf2_m))) +void vluxseg3ei32(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vbool16_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, const int8_t * op7, vuint32m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_i8mf4))) +void vluxseg3ei32(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, const int8_t * op3, vuint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_i8mf4_m))) +void vluxseg3ei32(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vbool32_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, const int8_t * op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_i8mf8))) +void vluxseg3ei32(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, const int8_t * op3, vuint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_i8mf8_m))) +void vluxseg3ei32(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vbool64_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, const int8_t * op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_u8m1))) +void vluxseg3ei32(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, const uint8_t * op3, vuint32m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_u8m1_m))) +void vluxseg3ei32(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vbool8_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, const uint8_t * op7, vuint32m4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_u8m2))) +void vluxseg3ei32(vuint8m2_t * op0, vuint8m2_t * op1, vuint8m2_t * op2, const uint8_t * op3, vuint32m8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_u8m2_m))) +void vluxseg3ei32(vuint8m2_t * op0, vuint8m2_t * op1, vuint8m2_t * op2, vbool4_t op3, vuint8m2_t op4, vuint8m2_t op5, vuint8m2_t op6, const uint8_t * op7, vuint32m8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_u8mf2))) +void vluxseg3ei32(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, const uint8_t * op3, vuint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_u8mf2_m))) +void vluxseg3ei32(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vbool16_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, const uint8_t * op7, vuint32m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_u8mf4))) +void vluxseg3ei32(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, const uint8_t * op3, vuint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_u8mf4_m))) +void vluxseg3ei32(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vbool32_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, const uint8_t * op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_u8mf8))) +void vluxseg3ei32(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, const uint8_t * op3, vuint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_u8mf8_m))) +void vluxseg3ei32(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vbool64_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, const uint8_t * op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_i8m1))) +void vluxseg4ei32(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, const int8_t * op4, vuint32m4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_i8m1_m))) +void vluxseg4ei32(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vbool8_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, const int8_t * op9, vuint32m4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_i8m2))) +void vluxseg4ei32(vint8m2_t * op0, vint8m2_t * op1, vint8m2_t * op2, vint8m2_t * op3, const int8_t * op4, vuint32m8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_i8m2_m))) +void vluxseg4ei32(vint8m2_t * op0, vint8m2_t * op1, vint8m2_t * op2, vint8m2_t * op3, vbool4_t op4, vint8m2_t op5, vint8m2_t op6, vint8m2_t op7, vint8m2_t op8, const int8_t * op9, vuint32m8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_i8mf2))) +void vluxseg4ei32(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, const int8_t * op4, vuint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_i8mf2_m))) +void vluxseg4ei32(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vbool16_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, const int8_t * op9, vuint32m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_i8mf4))) +void vluxseg4ei32(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, const int8_t * op4, vuint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_i8mf4_m))) +void vluxseg4ei32(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vbool32_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, const int8_t * op9, vuint32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_i8mf8))) +void vluxseg4ei32(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, const int8_t * op4, vuint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_i8mf8_m))) +void vluxseg4ei32(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vbool64_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, const int8_t * op9, vuint32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_u8m1))) +void vluxseg4ei32(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, const uint8_t * op4, vuint32m4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_u8m1_m))) +void vluxseg4ei32(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vbool8_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, const uint8_t * op9, vuint32m4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_u8m2))) +void vluxseg4ei32(vuint8m2_t * op0, vuint8m2_t * op1, vuint8m2_t * op2, vuint8m2_t * op3, const uint8_t * op4, vuint32m8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_u8m2_m))) +void vluxseg4ei32(vuint8m2_t * op0, vuint8m2_t * op1, vuint8m2_t * op2, vuint8m2_t * op3, vbool4_t op4, vuint8m2_t op5, vuint8m2_t op6, vuint8m2_t op7, vuint8m2_t op8, const uint8_t * op9, vuint32m8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_u8mf2))) +void vluxseg4ei32(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, const uint8_t * op4, vuint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_u8mf2_m))) +void vluxseg4ei32(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vbool16_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, const uint8_t * op9, vuint32m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_u8mf4))) +void vluxseg4ei32(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, const uint8_t * op4, vuint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_u8mf4_m))) +void vluxseg4ei32(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vbool32_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, const uint8_t * op9, vuint32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_u8mf8))) +void vluxseg4ei32(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, const uint8_t * op4, vuint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_u8mf8_m))) +void vluxseg4ei32(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vbool64_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, const uint8_t * op9, vuint32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_i8m1))) +void vluxseg5ei32(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, const int8_t * op5, vuint32m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_i8m1_m))) +void vluxseg5ei32(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vbool8_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, vint8m1_t op10, const int8_t * op11, vuint32m4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_i8mf2))) +void vluxseg5ei32(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, const int8_t * op5, vuint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_i8mf2_m))) +void vluxseg5ei32(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vbool16_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, vint8mf2_t op10, const int8_t * op11, vuint32m2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_i8mf4))) +void vluxseg5ei32(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, const int8_t * op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_i8mf4_m))) +void vluxseg5ei32(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vbool32_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, vint8mf4_t op10, const int8_t * op11, vuint32m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_i8mf8))) +void vluxseg5ei32(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, const int8_t * op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_i8mf8_m))) +void vluxseg5ei32(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vbool64_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, vint8mf8_t op10, const int8_t * op11, vuint32mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_u8m1))) +void vluxseg5ei32(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, const uint8_t * op5, vuint32m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_u8m1_m))) +void vluxseg5ei32(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vbool8_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, vuint8m1_t op10, const uint8_t * op11, vuint32m4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_u8mf2))) +void vluxseg5ei32(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, const uint8_t * op5, vuint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_u8mf2_m))) +void vluxseg5ei32(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vbool16_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, vuint8mf2_t op10, const uint8_t * op11, vuint32m2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_u8mf4))) +void vluxseg5ei32(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, const uint8_t * op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_u8mf4_m))) +void vluxseg5ei32(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vbool32_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, vuint8mf4_t op10, const uint8_t * op11, vuint32m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_u8mf8))) +void vluxseg5ei32(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, const uint8_t * op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_u8mf8_m))) +void vluxseg5ei32(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vbool64_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, vuint8mf8_t op10, const uint8_t * op11, vuint32mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_i8m1))) +void vluxseg6ei32(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, const int8_t * op6, vuint32m4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_i8m1_m))) +void vluxseg6ei32(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vbool8_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, vint8m1_t op10, vint8m1_t op11, vint8m1_t op12, const int8_t * op13, vuint32m4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_i8mf2))) +void vluxseg6ei32(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, const int8_t * op6, vuint32m2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_i8mf2_m))) +void vluxseg6ei32(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vbool16_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, vint8mf2_t op10, vint8mf2_t op11, vint8mf2_t op12, const int8_t * op13, vuint32m2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_i8mf4))) +void vluxseg6ei32(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, const int8_t * op6, vuint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_i8mf4_m))) +void vluxseg6ei32(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vbool32_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, vint8mf4_t op10, vint8mf4_t op11, vint8mf4_t op12, const int8_t * op13, vuint32m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_i8mf8))) +void vluxseg6ei32(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, const int8_t * op6, vuint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_i8mf8_m))) +void vluxseg6ei32(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vbool64_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, vint8mf8_t op10, vint8mf8_t op11, vint8mf8_t op12, const int8_t * op13, vuint32mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_u8m1))) +void vluxseg6ei32(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, const uint8_t * op6, vuint32m4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_u8m1_m))) +void vluxseg6ei32(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vbool8_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, vuint8m1_t op10, vuint8m1_t op11, vuint8m1_t op12, const uint8_t * op13, vuint32m4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_u8mf2))) +void vluxseg6ei32(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, const uint8_t * op6, vuint32m2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_u8mf2_m))) +void vluxseg6ei32(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vbool16_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, vuint8mf2_t op10, vuint8mf2_t op11, vuint8mf2_t op12, const uint8_t * op13, vuint32m2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_u8mf4))) +void vluxseg6ei32(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, const uint8_t * op6, vuint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_u8mf4_m))) +void vluxseg6ei32(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vbool32_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, vuint8mf4_t op10, vuint8mf4_t op11, vuint8mf4_t op12, const uint8_t * op13, vuint32m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_u8mf8))) +void vluxseg6ei32(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, const uint8_t * op6, vuint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_u8mf8_m))) +void vluxseg6ei32(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vbool64_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, vuint8mf8_t op10, vuint8mf8_t op11, vuint8mf8_t op12, const uint8_t * op13, vuint32mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_i8m1))) +void vluxseg7ei32(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vint8m1_t * op6, const int8_t * op7, vuint32m4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_i8m1_m))) +void vluxseg7ei32(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vint8m1_t * op6, vbool8_t op7, vint8m1_t op8, vint8m1_t op9, vint8m1_t op10, vint8m1_t op11, vint8m1_t op12, vint8m1_t op13, vint8m1_t op14, const int8_t * op15, vuint32m4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_i8mf2))) +void vluxseg7ei32(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vint8mf2_t * op6, const int8_t * op7, vuint32m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_i8mf2_m))) +void vluxseg7ei32(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vint8mf2_t * op6, vbool16_t op7, vint8mf2_t op8, vint8mf2_t op9, vint8mf2_t op10, vint8mf2_t op11, vint8mf2_t op12, vint8mf2_t op13, vint8mf2_t op14, const int8_t * op15, vuint32m2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_i8mf4))) +void vluxseg7ei32(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vint8mf4_t * op6, const int8_t * op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_i8mf4_m))) +void vluxseg7ei32(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vint8mf4_t * op6, vbool32_t op7, vint8mf4_t op8, vint8mf4_t op9, vint8mf4_t op10, vint8mf4_t op11, vint8mf4_t op12, vint8mf4_t op13, vint8mf4_t op14, const int8_t * op15, vuint32m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_i8mf8))) +void vluxseg7ei32(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vint8mf8_t * op6, const int8_t * op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_i8mf8_m))) +void vluxseg7ei32(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vint8mf8_t * op6, vbool64_t op7, vint8mf8_t op8, vint8mf8_t op9, vint8mf8_t op10, vint8mf8_t op11, vint8mf8_t op12, vint8mf8_t op13, vint8mf8_t op14, const int8_t * op15, vuint32mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_u8m1))) +void vluxseg7ei32(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vuint8m1_t * op6, const uint8_t * op7, vuint32m4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_u8m1_m))) +void vluxseg7ei32(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vuint8m1_t * op6, vbool8_t op7, vuint8m1_t op8, vuint8m1_t op9, vuint8m1_t op10, vuint8m1_t op11, vuint8m1_t op12, vuint8m1_t op13, vuint8m1_t op14, const uint8_t * op15, vuint32m4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_u8mf2))) +void vluxseg7ei32(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vuint8mf2_t * op6, const uint8_t * op7, vuint32m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_u8mf2_m))) +void vluxseg7ei32(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vuint8mf2_t * op6, vbool16_t op7, vuint8mf2_t op8, vuint8mf2_t op9, vuint8mf2_t op10, vuint8mf2_t op11, vuint8mf2_t op12, vuint8mf2_t op13, vuint8mf2_t op14, const uint8_t * op15, vuint32m2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_u8mf4))) +void vluxseg7ei32(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vuint8mf4_t * op6, const uint8_t * op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_u8mf4_m))) +void vluxseg7ei32(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vuint8mf4_t * op6, vbool32_t op7, vuint8mf4_t op8, vuint8mf4_t op9, vuint8mf4_t op10, vuint8mf4_t op11, vuint8mf4_t op12, vuint8mf4_t op13, vuint8mf4_t op14, const uint8_t * op15, vuint32m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_u8mf8))) +void vluxseg7ei32(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vuint8mf8_t * op6, const uint8_t * op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_u8mf8_m))) +void vluxseg7ei32(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vuint8mf8_t * op6, vbool64_t op7, vuint8mf8_t op8, vuint8mf8_t op9, vuint8mf8_t op10, vuint8mf8_t op11, vuint8mf8_t op12, vuint8mf8_t op13, vuint8mf8_t op14, const uint8_t * op15, vuint32mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_i8m1))) +void vluxseg8ei32(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vint8m1_t * op6, vint8m1_t * op7, const int8_t * op8, vuint32m4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_i8m1_m))) +void vluxseg8ei32(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vint8m1_t * op6, vint8m1_t * op7, vbool8_t op8, vint8m1_t op9, vint8m1_t op10, vint8m1_t op11, vint8m1_t op12, vint8m1_t op13, vint8m1_t op14, vint8m1_t op15, vint8m1_t op16, const int8_t * op17, vuint32m4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_i8mf2))) +void vluxseg8ei32(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vint8mf2_t * op6, vint8mf2_t * op7, const int8_t * op8, vuint32m2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_i8mf2_m))) +void vluxseg8ei32(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vint8mf2_t * op6, vint8mf2_t * op7, vbool16_t op8, vint8mf2_t op9, vint8mf2_t op10, vint8mf2_t op11, vint8mf2_t op12, vint8mf2_t op13, vint8mf2_t op14, vint8mf2_t op15, vint8mf2_t op16, const int8_t * op17, vuint32m2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_i8mf4))) +void vluxseg8ei32(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vint8mf4_t * op6, vint8mf4_t * op7, const int8_t * op8, vuint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_i8mf4_m))) +void vluxseg8ei32(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vint8mf4_t * op6, vint8mf4_t * op7, vbool32_t op8, vint8mf4_t op9, vint8mf4_t op10, vint8mf4_t op11, vint8mf4_t op12, vint8mf4_t op13, vint8mf4_t op14, vint8mf4_t op15, vint8mf4_t op16, const int8_t * op17, vuint32m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_i8mf8))) +void vluxseg8ei32(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vint8mf8_t * op6, vint8mf8_t * op7, const int8_t * op8, vuint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_i8mf8_m))) +void vluxseg8ei32(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vint8mf8_t * op6, vint8mf8_t * op7, vbool64_t op8, vint8mf8_t op9, vint8mf8_t op10, vint8mf8_t op11, vint8mf8_t op12, vint8mf8_t op13, vint8mf8_t op14, vint8mf8_t op15, vint8mf8_t op16, const int8_t * op17, vuint32mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_u8m1))) +void vluxseg8ei32(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vuint8m1_t * op6, vuint8m1_t * op7, const uint8_t * op8, vuint32m4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_u8m1_m))) +void vluxseg8ei32(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vuint8m1_t * op6, vuint8m1_t * op7, vbool8_t op8, vuint8m1_t op9, vuint8m1_t op10, vuint8m1_t op11, vuint8m1_t op12, vuint8m1_t op13, vuint8m1_t op14, vuint8m1_t op15, vuint8m1_t op16, const uint8_t * op17, vuint32m4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_u8mf2))) +void vluxseg8ei32(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vuint8mf2_t * op6, vuint8mf2_t * op7, const uint8_t * op8, vuint32m2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_u8mf2_m))) +void vluxseg8ei32(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vuint8mf2_t * op6, vuint8mf2_t * op7, vbool16_t op8, vuint8mf2_t op9, vuint8mf2_t op10, vuint8mf2_t op11, vuint8mf2_t op12, vuint8mf2_t op13, vuint8mf2_t op14, vuint8mf2_t op15, vuint8mf2_t op16, const uint8_t * op17, vuint32m2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_u8mf4))) +void vluxseg8ei32(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vuint8mf4_t * op6, vuint8mf4_t * op7, const uint8_t * op8, vuint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_u8mf4_m))) +void vluxseg8ei32(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vuint8mf4_t * op6, vuint8mf4_t * op7, vbool32_t op8, vuint8mf4_t op9, vuint8mf4_t op10, vuint8mf4_t op11, vuint8mf4_t op12, vuint8mf4_t op13, vuint8mf4_t op14, vuint8mf4_t op15, vuint8mf4_t op16, const uint8_t * op17, vuint32m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_u8mf8))) +void vluxseg8ei32(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vuint8mf8_t * op6, vuint8mf8_t * op7, const uint8_t * op8, vuint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_u8mf8_m))) +void vluxseg8ei32(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vuint8mf8_t * op6, vuint8mf8_t * op7, vbool64_t op8, vuint8mf8_t op9, vuint8mf8_t op10, vuint8mf8_t op11, vuint8mf8_t op12, vuint8mf8_t op13, vuint8mf8_t op14, vuint8mf8_t op15, vuint8mf8_t op16, const uint8_t * op17, vuint32mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_i8m1))) +void vluxseg2ei64(vint8m1_t * op0, vint8m1_t * op1, const int8_t * op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_i8m1_m))) +void vluxseg2ei64(vint8m1_t * op0, vint8m1_t * op1, vbool8_t op2, vint8m1_t op3, vint8m1_t op4, const int8_t * op5, vuint64m8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_i8mf2))) +void vluxseg2ei64(vint8mf2_t * op0, vint8mf2_t * op1, const int8_t * op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_i8mf2_m))) +void vluxseg2ei64(vint8mf2_t * op0, vint8mf2_t * op1, vbool16_t op2, vint8mf2_t op3, vint8mf2_t op4, const int8_t * op5, vuint64m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_i8mf4))) +void vluxseg2ei64(vint8mf4_t * op0, vint8mf4_t * op1, const int8_t * op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_i8mf4_m))) +void vluxseg2ei64(vint8mf4_t * op0, vint8mf4_t * op1, vbool32_t op2, vint8mf4_t op3, vint8mf4_t op4, const int8_t * op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_i8mf8))) +void vluxseg2ei64(vint8mf8_t * op0, vint8mf8_t * op1, const int8_t * op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_i8mf8_m))) +void vluxseg2ei64(vint8mf8_t * op0, vint8mf8_t * op1, vbool64_t op2, vint8mf8_t op3, vint8mf8_t op4, const int8_t * op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_u8m1))) +void vluxseg2ei64(vuint8m1_t * op0, vuint8m1_t * op1, const uint8_t * op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_u8m1_m))) +void vluxseg2ei64(vuint8m1_t * op0, vuint8m1_t * op1, vbool8_t op2, vuint8m1_t op3, vuint8m1_t op4, const uint8_t * op5, vuint64m8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_u8mf2))) +void vluxseg2ei64(vuint8mf2_t * op0, vuint8mf2_t * op1, const uint8_t * op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_u8mf2_m))) +void vluxseg2ei64(vuint8mf2_t * op0, vuint8mf2_t * op1, vbool16_t op2, vuint8mf2_t op3, vuint8mf2_t op4, const uint8_t * op5, vuint64m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_u8mf4))) +void vluxseg2ei64(vuint8mf4_t * op0, vuint8mf4_t * op1, const uint8_t * op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_u8mf4_m))) +void vluxseg2ei64(vuint8mf4_t * op0, vuint8mf4_t * op1, vbool32_t op2, vuint8mf4_t op3, vuint8mf4_t op4, const uint8_t * op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_u8mf8))) +void vluxseg2ei64(vuint8mf8_t * op0, vuint8mf8_t * op1, const uint8_t * op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_u8mf8_m))) +void vluxseg2ei64(vuint8mf8_t * op0, vuint8mf8_t * op1, vbool64_t op2, vuint8mf8_t op3, vuint8mf8_t op4, const uint8_t * op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_i8m1))) +void vluxseg3ei64(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, const int8_t * op3, vuint64m8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_i8m1_m))) +void vluxseg3ei64(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vbool8_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, const int8_t * op7, vuint64m8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_i8mf2))) +void vluxseg3ei64(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, const int8_t * op3, vuint64m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_i8mf2_m))) +void vluxseg3ei64(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vbool16_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, const int8_t * op7, vuint64m4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_i8mf4))) +void vluxseg3ei64(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, const int8_t * op3, vuint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_i8mf4_m))) +void vluxseg3ei64(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vbool32_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, const int8_t * op7, vuint64m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_i8mf8))) +void vluxseg3ei64(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, const int8_t * op3, vuint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_i8mf8_m))) +void vluxseg3ei64(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vbool64_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, const int8_t * op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_u8m1))) +void vluxseg3ei64(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, const uint8_t * op3, vuint64m8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_u8m1_m))) +void vluxseg3ei64(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vbool8_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, const uint8_t * op7, vuint64m8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_u8mf2))) +void vluxseg3ei64(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, const uint8_t * op3, vuint64m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_u8mf2_m))) +void vluxseg3ei64(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vbool16_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, const uint8_t * op7, vuint64m4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_u8mf4))) +void vluxseg3ei64(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, const uint8_t * op3, vuint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_u8mf4_m))) +void vluxseg3ei64(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vbool32_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, const uint8_t * op7, vuint64m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_u8mf8))) +void vluxseg3ei64(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, const uint8_t * op3, vuint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_u8mf8_m))) +void vluxseg3ei64(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vbool64_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, const uint8_t * op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_i8m1))) +void vluxseg4ei64(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, const int8_t * op4, vuint64m8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_i8m1_m))) +void vluxseg4ei64(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vbool8_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, const int8_t * op9, vuint64m8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_i8mf2))) +void vluxseg4ei64(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, const int8_t * op4, vuint64m4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_i8mf2_m))) +void vluxseg4ei64(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vbool16_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, const int8_t * op9, vuint64m4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_i8mf4))) +void vluxseg4ei64(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, const int8_t * op4, vuint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_i8mf4_m))) +void vluxseg4ei64(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vbool32_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, const int8_t * op9, vuint64m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_i8mf8))) +void vluxseg4ei64(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, const int8_t * op4, vuint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_i8mf8_m))) +void vluxseg4ei64(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vbool64_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, const int8_t * op9, vuint64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_u8m1))) +void vluxseg4ei64(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, const uint8_t * op4, vuint64m8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_u8m1_m))) +void vluxseg4ei64(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vbool8_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, const uint8_t * op9, vuint64m8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_u8mf2))) +void vluxseg4ei64(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, const uint8_t * op4, vuint64m4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_u8mf2_m))) +void vluxseg4ei64(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vbool16_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, const uint8_t * op9, vuint64m4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_u8mf4))) +void vluxseg4ei64(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, const uint8_t * op4, vuint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_u8mf4_m))) +void vluxseg4ei64(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vbool32_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, const uint8_t * op9, vuint64m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_u8mf8))) +void vluxseg4ei64(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, const uint8_t * op4, vuint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_u8mf8_m))) +void vluxseg4ei64(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vbool64_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, const uint8_t * op9, vuint64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_i8m1))) +void vluxseg5ei64(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, const int8_t * op5, vuint64m8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_i8m1_m))) +void vluxseg5ei64(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vbool8_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, vint8m1_t op10, const int8_t * op11, vuint64m8_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_i8mf2))) +void vluxseg5ei64(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, const int8_t * op5, vuint64m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_i8mf2_m))) +void vluxseg5ei64(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vbool16_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, vint8mf2_t op10, const int8_t * op11, vuint64m4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_i8mf4))) +void vluxseg5ei64(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, const int8_t * op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_i8mf4_m))) +void vluxseg5ei64(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vbool32_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, vint8mf4_t op10, const int8_t * op11, vuint64m2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_i8mf8))) +void vluxseg5ei64(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, const int8_t * op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_i8mf8_m))) +void vluxseg5ei64(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vbool64_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, vint8mf8_t op10, const int8_t * op11, vuint64m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_u8m1))) +void vluxseg5ei64(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, const uint8_t * op5, vuint64m8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_u8m1_m))) +void vluxseg5ei64(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vbool8_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, vuint8m1_t op10, const uint8_t * op11, vuint64m8_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_u8mf2))) +void vluxseg5ei64(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, const uint8_t * op5, vuint64m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_u8mf2_m))) +void vluxseg5ei64(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vbool16_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, vuint8mf2_t op10, const uint8_t * op11, vuint64m4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_u8mf4))) +void vluxseg5ei64(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, const uint8_t * op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_u8mf4_m))) +void vluxseg5ei64(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vbool32_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, vuint8mf4_t op10, const uint8_t * op11, vuint64m2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_u8mf8))) +void vluxseg5ei64(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, const uint8_t * op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_u8mf8_m))) +void vluxseg5ei64(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vbool64_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, vuint8mf8_t op10, const uint8_t * op11, vuint64m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_i8m1))) +void vluxseg6ei64(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, const int8_t * op6, vuint64m8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_i8m1_m))) +void vluxseg6ei64(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vbool8_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, vint8m1_t op10, vint8m1_t op11, vint8m1_t op12, const int8_t * op13, vuint64m8_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_i8mf2))) +void vluxseg6ei64(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, const int8_t * op6, vuint64m4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_i8mf2_m))) +void vluxseg6ei64(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vbool16_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, vint8mf2_t op10, vint8mf2_t op11, vint8mf2_t op12, const int8_t * op13, vuint64m4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_i8mf4))) +void vluxseg6ei64(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, const int8_t * op6, vuint64m2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_i8mf4_m))) +void vluxseg6ei64(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vbool32_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, vint8mf4_t op10, vint8mf4_t op11, vint8mf4_t op12, const int8_t * op13, vuint64m2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_i8mf8))) +void vluxseg6ei64(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, const int8_t * op6, vuint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_i8mf8_m))) +void vluxseg6ei64(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vbool64_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, vint8mf8_t op10, vint8mf8_t op11, vint8mf8_t op12, const int8_t * op13, vuint64m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_u8m1))) +void vluxseg6ei64(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, const uint8_t * op6, vuint64m8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_u8m1_m))) +void vluxseg6ei64(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vbool8_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, vuint8m1_t op10, vuint8m1_t op11, vuint8m1_t op12, const uint8_t * op13, vuint64m8_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_u8mf2))) +void vluxseg6ei64(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, const uint8_t * op6, vuint64m4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_u8mf2_m))) +void vluxseg6ei64(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vbool16_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, vuint8mf2_t op10, vuint8mf2_t op11, vuint8mf2_t op12, const uint8_t * op13, vuint64m4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_u8mf4))) +void vluxseg6ei64(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, const uint8_t * op6, vuint64m2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_u8mf4_m))) +void vluxseg6ei64(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vbool32_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, vuint8mf4_t op10, vuint8mf4_t op11, vuint8mf4_t op12, const uint8_t * op13, vuint64m2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_u8mf8))) +void vluxseg6ei64(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, const uint8_t * op6, vuint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_u8mf8_m))) +void vluxseg6ei64(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vbool64_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, vuint8mf8_t op10, vuint8mf8_t op11, vuint8mf8_t op12, const uint8_t * op13, vuint64m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_i8m1))) +void vluxseg7ei64(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vint8m1_t * op6, const int8_t * op7, vuint64m8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_i8m1_m))) +void vluxseg7ei64(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vint8m1_t * op6, vbool8_t op7, vint8m1_t op8, vint8m1_t op9, vint8m1_t op10, vint8m1_t op11, vint8m1_t op12, vint8m1_t op13, vint8m1_t op14, const int8_t * op15, vuint64m8_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_i8mf2))) +void vluxseg7ei64(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vint8mf2_t * op6, const int8_t * op7, vuint64m4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_i8mf2_m))) +void vluxseg7ei64(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vint8mf2_t * op6, vbool16_t op7, vint8mf2_t op8, vint8mf2_t op9, vint8mf2_t op10, vint8mf2_t op11, vint8mf2_t op12, vint8mf2_t op13, vint8mf2_t op14, const int8_t * op15, vuint64m4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_i8mf4))) +void vluxseg7ei64(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vint8mf4_t * op6, const int8_t * op7, vuint64m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_i8mf4_m))) +void vluxseg7ei64(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vint8mf4_t * op6, vbool32_t op7, vint8mf4_t op8, vint8mf4_t op9, vint8mf4_t op10, vint8mf4_t op11, vint8mf4_t op12, vint8mf4_t op13, vint8mf4_t op14, const int8_t * op15, vuint64m2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_i8mf8))) +void vluxseg7ei64(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vint8mf8_t * op6, const int8_t * op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_i8mf8_m))) +void vluxseg7ei64(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vint8mf8_t * op6, vbool64_t op7, vint8mf8_t op8, vint8mf8_t op9, vint8mf8_t op10, vint8mf8_t op11, vint8mf8_t op12, vint8mf8_t op13, vint8mf8_t op14, const int8_t * op15, vuint64m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_u8m1))) +void vluxseg7ei64(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vuint8m1_t * op6, const uint8_t * op7, vuint64m8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_u8m1_m))) +void vluxseg7ei64(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vuint8m1_t * op6, vbool8_t op7, vuint8m1_t op8, vuint8m1_t op9, vuint8m1_t op10, vuint8m1_t op11, vuint8m1_t op12, vuint8m1_t op13, vuint8m1_t op14, const uint8_t * op15, vuint64m8_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_u8mf2))) +void vluxseg7ei64(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vuint8mf2_t * op6, const uint8_t * op7, vuint64m4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_u8mf2_m))) +void vluxseg7ei64(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vuint8mf2_t * op6, vbool16_t op7, vuint8mf2_t op8, vuint8mf2_t op9, vuint8mf2_t op10, vuint8mf2_t op11, vuint8mf2_t op12, vuint8mf2_t op13, vuint8mf2_t op14, const uint8_t * op15, vuint64m4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_u8mf4))) +void vluxseg7ei64(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vuint8mf4_t * op6, const uint8_t * op7, vuint64m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_u8mf4_m))) +void vluxseg7ei64(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vuint8mf4_t * op6, vbool32_t op7, vuint8mf4_t op8, vuint8mf4_t op9, vuint8mf4_t op10, vuint8mf4_t op11, vuint8mf4_t op12, vuint8mf4_t op13, vuint8mf4_t op14, const uint8_t * op15, vuint64m2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_u8mf8))) +void vluxseg7ei64(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vuint8mf8_t * op6, const uint8_t * op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_u8mf8_m))) +void vluxseg7ei64(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vuint8mf8_t * op6, vbool64_t op7, vuint8mf8_t op8, vuint8mf8_t op9, vuint8mf8_t op10, vuint8mf8_t op11, vuint8mf8_t op12, vuint8mf8_t op13, vuint8mf8_t op14, const uint8_t * op15, vuint64m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_i8m1))) +void vluxseg8ei64(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vint8m1_t * op6, vint8m1_t * op7, const int8_t * op8, vuint64m8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_i8m1_m))) +void vluxseg8ei64(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vint8m1_t * op6, vint8m1_t * op7, vbool8_t op8, vint8m1_t op9, vint8m1_t op10, vint8m1_t op11, vint8m1_t op12, vint8m1_t op13, vint8m1_t op14, vint8m1_t op15, vint8m1_t op16, const int8_t * op17, vuint64m8_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_i8mf2))) +void vluxseg8ei64(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vint8mf2_t * op6, vint8mf2_t * op7, const int8_t * op8, vuint64m4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_i8mf2_m))) +void vluxseg8ei64(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vint8mf2_t * op6, vint8mf2_t * op7, vbool16_t op8, vint8mf2_t op9, vint8mf2_t op10, vint8mf2_t op11, vint8mf2_t op12, vint8mf2_t op13, vint8mf2_t op14, vint8mf2_t op15, vint8mf2_t op16, const int8_t * op17, vuint64m4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_i8mf4))) +void vluxseg8ei64(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vint8mf4_t * op6, vint8mf4_t * op7, const int8_t * op8, vuint64m2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_i8mf4_m))) +void vluxseg8ei64(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vint8mf4_t * op6, vint8mf4_t * op7, vbool32_t op8, vint8mf4_t op9, vint8mf4_t op10, vint8mf4_t op11, vint8mf4_t op12, vint8mf4_t op13, vint8mf4_t op14, vint8mf4_t op15, vint8mf4_t op16, const int8_t * op17, vuint64m2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_i8mf8))) +void vluxseg8ei64(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vint8mf8_t * op6, vint8mf8_t * op7, const int8_t * op8, vuint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_i8mf8_m))) +void vluxseg8ei64(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vint8mf8_t * op6, vint8mf8_t * op7, vbool64_t op8, vint8mf8_t op9, vint8mf8_t op10, vint8mf8_t op11, vint8mf8_t op12, vint8mf8_t op13, vint8mf8_t op14, vint8mf8_t op15, vint8mf8_t op16, const int8_t * op17, vuint64m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_u8m1))) +void vluxseg8ei64(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vuint8m1_t * op6, vuint8m1_t * op7, const uint8_t * op8, vuint64m8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_u8m1_m))) +void vluxseg8ei64(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vuint8m1_t * op6, vuint8m1_t * op7, vbool8_t op8, vuint8m1_t op9, vuint8m1_t op10, vuint8m1_t op11, vuint8m1_t op12, vuint8m1_t op13, vuint8m1_t op14, vuint8m1_t op15, vuint8m1_t op16, const uint8_t * op17, vuint64m8_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_u8mf2))) +void vluxseg8ei64(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vuint8mf2_t * op6, vuint8mf2_t * op7, const uint8_t * op8, vuint64m4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_u8mf2_m))) +void vluxseg8ei64(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vuint8mf2_t * op6, vuint8mf2_t * op7, vbool16_t op8, vuint8mf2_t op9, vuint8mf2_t op10, vuint8mf2_t op11, vuint8mf2_t op12, vuint8mf2_t op13, vuint8mf2_t op14, vuint8mf2_t op15, vuint8mf2_t op16, const uint8_t * op17, vuint64m4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_u8mf4))) +void vluxseg8ei64(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vuint8mf4_t * op6, vuint8mf4_t * op7, const uint8_t * op8, vuint64m2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_u8mf4_m))) +void vluxseg8ei64(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vuint8mf4_t * op6, vuint8mf4_t * op7, vbool32_t op8, vuint8mf4_t op9, vuint8mf4_t op10, vuint8mf4_t op11, vuint8mf4_t op12, vuint8mf4_t op13, vuint8mf4_t op14, vuint8mf4_t op15, vuint8mf4_t op16, const uint8_t * op17, vuint64m2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_u8mf8))) +void vluxseg8ei64(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vuint8mf8_t * op6, vuint8mf8_t * op7, const uint8_t * op8, vuint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_u8mf8_m))) +void vluxseg8ei64(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vuint8mf8_t * op6, vuint8mf8_t * op7, vbool64_t op8, vuint8mf8_t op9, vuint8mf8_t op10, vuint8mf8_t op11, vuint8mf8_t op12, vuint8mf8_t op13, vuint8mf8_t op14, vuint8mf8_t op15, vuint8mf8_t op16, const uint8_t * op17, vuint64m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_i16m1))) +void vluxseg2ei8(vint16m1_t * op0, vint16m1_t * op1, const int16_t * op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_i16m1_m))) +void vluxseg2ei8(vint16m1_t * op0, vint16m1_t * op1, vbool16_t op2, vint16m1_t op3, vint16m1_t op4, const int16_t * op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_i16m2))) +void vluxseg2ei8(vint16m2_t * op0, vint16m2_t * op1, const int16_t * op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_i16m2_m))) +void vluxseg2ei8(vint16m2_t * op0, vint16m2_t * op1, vbool8_t op2, vint16m2_t op3, vint16m2_t op4, const int16_t * op5, vuint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_i16m4))) +void vluxseg2ei8(vint16m4_t * op0, vint16m4_t * op1, const int16_t * op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_i16m4_m))) +void vluxseg2ei8(vint16m4_t * op0, vint16m4_t * op1, vbool4_t op2, vint16m4_t op3, vint16m4_t op4, const int16_t * op5, vuint8m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_i16mf2))) +void vluxseg2ei8(vint16mf2_t * op0, vint16mf2_t * op1, const int16_t * op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_i16mf2_m))) +void vluxseg2ei8(vint16mf2_t * op0, vint16mf2_t * op1, vbool32_t op2, vint16mf2_t op3, vint16mf2_t op4, const int16_t * op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_i16mf4))) +void vluxseg2ei8(vint16mf4_t * op0, vint16mf4_t * op1, const int16_t * op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_i16mf4_m))) +void vluxseg2ei8(vint16mf4_t * op0, vint16mf4_t * op1, vbool64_t op2, vint16mf4_t op3, vint16mf4_t op4, const int16_t * op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_u16m1))) +void vluxseg2ei8(vuint16m1_t * op0, vuint16m1_t * op1, const uint16_t * op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_u16m1_m))) +void vluxseg2ei8(vuint16m1_t * op0, vuint16m1_t * op1, vbool16_t op2, vuint16m1_t op3, vuint16m1_t op4, const uint16_t * op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_u16m2))) +void vluxseg2ei8(vuint16m2_t * op0, vuint16m2_t * op1, const uint16_t * op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_u16m2_m))) +void vluxseg2ei8(vuint16m2_t * op0, vuint16m2_t * op1, vbool8_t op2, vuint16m2_t op3, vuint16m2_t op4, const uint16_t * op5, vuint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_u16m4))) +void vluxseg2ei8(vuint16m4_t * op0, vuint16m4_t * op1, const uint16_t * op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_u16m4_m))) +void vluxseg2ei8(vuint16m4_t * op0, vuint16m4_t * op1, vbool4_t op2, vuint16m4_t op3, vuint16m4_t op4, const uint16_t * op5, vuint8m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_u16mf2))) +void vluxseg2ei8(vuint16mf2_t * op0, vuint16mf2_t * op1, const uint16_t * op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_u16mf2_m))) +void vluxseg2ei8(vuint16mf2_t * op0, vuint16mf2_t * op1, vbool32_t op2, vuint16mf2_t op3, vuint16mf2_t op4, const uint16_t * op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_u16mf4))) +void vluxseg2ei8(vuint16mf4_t * op0, vuint16mf4_t * op1, const uint16_t * op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_u16mf4_m))) +void vluxseg2ei8(vuint16mf4_t * op0, vuint16mf4_t * op1, vbool64_t op2, vuint16mf4_t op3, vuint16mf4_t op4, const uint16_t * op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_i16m1))) +void vluxseg3ei8(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, const int16_t * op3, vuint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_i16m1_m))) +void vluxseg3ei8(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vbool16_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, const int16_t * op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_i16m2))) +void vluxseg3ei8(vint16m2_t * op0, vint16m2_t * op1, vint16m2_t * op2, const int16_t * op3, vuint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_i16m2_m))) +void vluxseg3ei8(vint16m2_t * op0, vint16m2_t * op1, vint16m2_t * op2, vbool8_t op3, vint16m2_t op4, vint16m2_t op5, vint16m2_t op6, const int16_t * op7, vuint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_i16mf2))) +void vluxseg3ei8(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, const int16_t * op3, vuint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_i16mf2_m))) +void vluxseg3ei8(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vbool32_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, const int16_t * op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_i16mf4))) +void vluxseg3ei8(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, const int16_t * op3, vuint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_i16mf4_m))) +void vluxseg3ei8(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vbool64_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, const int16_t * op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_u16m1))) +void vluxseg3ei8(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, const uint16_t * op3, vuint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_u16m1_m))) +void vluxseg3ei8(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vbool16_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, const uint16_t * op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_u16m2))) +void vluxseg3ei8(vuint16m2_t * op0, vuint16m2_t * op1, vuint16m2_t * op2, const uint16_t * op3, vuint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_u16m2_m))) +void vluxseg3ei8(vuint16m2_t * op0, vuint16m2_t * op1, vuint16m2_t * op2, vbool8_t op3, vuint16m2_t op4, vuint16m2_t op5, vuint16m2_t op6, const uint16_t * op7, vuint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_u16mf2))) +void vluxseg3ei8(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, const uint16_t * op3, vuint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_u16mf2_m))) +void vluxseg3ei8(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vbool32_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, const uint16_t * op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_u16mf4))) +void vluxseg3ei8(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, const uint16_t * op3, vuint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_u16mf4_m))) +void vluxseg3ei8(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vbool64_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, const uint16_t * op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_i16m1))) +void vluxseg4ei8(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, const int16_t * op4, vuint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_i16m1_m))) +void vluxseg4ei8(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vbool16_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, const int16_t * op9, vuint8mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_i16m2))) +void vluxseg4ei8(vint16m2_t * op0, vint16m2_t * op1, vint16m2_t * op2, vint16m2_t * op3, const int16_t * op4, vuint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_i16m2_m))) +void vluxseg4ei8(vint16m2_t * op0, vint16m2_t * op1, vint16m2_t * op2, vint16m2_t * op3, vbool8_t op4, vint16m2_t op5, vint16m2_t op6, vint16m2_t op7, vint16m2_t op8, const int16_t * op9, vuint8m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_i16mf2))) +void vluxseg4ei8(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, const int16_t * op4, vuint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_i16mf2_m))) +void vluxseg4ei8(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vbool32_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, const int16_t * op9, vuint8mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_i16mf4))) +void vluxseg4ei8(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, const int16_t * op4, vuint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_i16mf4_m))) +void vluxseg4ei8(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vbool64_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, const int16_t * op9, vuint8mf8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_u16m1))) +void vluxseg4ei8(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, const uint16_t * op4, vuint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_u16m1_m))) +void vluxseg4ei8(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vbool16_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, const uint16_t * op9, vuint8mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_u16m2))) +void vluxseg4ei8(vuint16m2_t * op0, vuint16m2_t * op1, vuint16m2_t * op2, vuint16m2_t * op3, const uint16_t * op4, vuint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_u16m2_m))) +void vluxseg4ei8(vuint16m2_t * op0, vuint16m2_t * op1, vuint16m2_t * op2, vuint16m2_t * op3, vbool8_t op4, vuint16m2_t op5, vuint16m2_t op6, vuint16m2_t op7, vuint16m2_t op8, const uint16_t * op9, vuint8m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_u16mf2))) +void vluxseg4ei8(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, const uint16_t * op4, vuint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_u16mf2_m))) +void vluxseg4ei8(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vbool32_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, const uint16_t * op9, vuint8mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_u16mf4))) +void vluxseg4ei8(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, const uint16_t * op4, vuint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_u16mf4_m))) +void vluxseg4ei8(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vbool64_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, const uint16_t * op9, vuint8mf8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_i16m1))) +void vluxseg5ei8(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, const int16_t * op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_i16m1_m))) +void vluxseg5ei8(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vbool16_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, vint16m1_t op10, const int16_t * op11, vuint8mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_i16mf2))) +void vluxseg5ei8(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, const int16_t * op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_i16mf2_m))) +void vluxseg5ei8(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vbool32_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, vint16mf2_t op10, const int16_t * op11, vuint8mf4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_i16mf4))) +void vluxseg5ei8(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, const int16_t * op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_i16mf4_m))) +void vluxseg5ei8(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vbool64_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, vint16mf4_t op10, const int16_t * op11, vuint8mf8_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_u16m1))) +void vluxseg5ei8(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, const uint16_t * op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_u16m1_m))) +void vluxseg5ei8(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vbool16_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, vuint16m1_t op10, const uint16_t * op11, vuint8mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_u16mf2))) +void vluxseg5ei8(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, const uint16_t * op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_u16mf2_m))) +void vluxseg5ei8(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vbool32_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, vuint16mf2_t op10, const uint16_t * op11, vuint8mf4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_u16mf4))) +void vluxseg5ei8(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, const uint16_t * op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_u16mf4_m))) +void vluxseg5ei8(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vbool64_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, vuint16mf4_t op10, const uint16_t * op11, vuint8mf8_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_i16m1))) +void vluxseg6ei8(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, const int16_t * op6, vuint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_i16m1_m))) +void vluxseg6ei8(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vbool16_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, vint16m1_t op10, vint16m1_t op11, vint16m1_t op12, const int16_t * op13, vuint8mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_i16mf2))) +void vluxseg6ei8(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, const int16_t * op6, vuint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_i16mf2_m))) +void vluxseg6ei8(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vbool32_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, vint16mf2_t op10, vint16mf2_t op11, vint16mf2_t op12, const int16_t * op13, vuint8mf4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_i16mf4))) +void vluxseg6ei8(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, const int16_t * op6, vuint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_i16mf4_m))) +void vluxseg6ei8(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vbool64_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, vint16mf4_t op10, vint16mf4_t op11, vint16mf4_t op12, const int16_t * op13, vuint8mf8_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_u16m1))) +void vluxseg6ei8(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, const uint16_t * op6, vuint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_u16m1_m))) +void vluxseg6ei8(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vbool16_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, vuint16m1_t op10, vuint16m1_t op11, vuint16m1_t op12, const uint16_t * op13, vuint8mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_u16mf2))) +void vluxseg6ei8(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, const uint16_t * op6, vuint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_u16mf2_m))) +void vluxseg6ei8(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vbool32_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, vuint16mf2_t op10, vuint16mf2_t op11, vuint16mf2_t op12, const uint16_t * op13, vuint8mf4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_u16mf4))) +void vluxseg6ei8(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, const uint16_t * op6, vuint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_u16mf4_m))) +void vluxseg6ei8(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vbool64_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, vuint16mf4_t op10, vuint16mf4_t op11, vuint16mf4_t op12, const uint16_t * op13, vuint8mf8_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_i16m1))) +void vluxseg7ei8(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vint16m1_t * op6, const int16_t * op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_i16m1_m))) +void vluxseg7ei8(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vint16m1_t * op6, vbool16_t op7, vint16m1_t op8, vint16m1_t op9, vint16m1_t op10, vint16m1_t op11, vint16m1_t op12, vint16m1_t op13, vint16m1_t op14, const int16_t * op15, vuint8mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_i16mf2))) +void vluxseg7ei8(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vint16mf2_t * op6, const int16_t * op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_i16mf2_m))) +void vluxseg7ei8(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vint16mf2_t * op6, vbool32_t op7, vint16mf2_t op8, vint16mf2_t op9, vint16mf2_t op10, vint16mf2_t op11, vint16mf2_t op12, vint16mf2_t op13, vint16mf2_t op14, const int16_t * op15, vuint8mf4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_i16mf4))) +void vluxseg7ei8(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vint16mf4_t * op6, const int16_t * op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_i16mf4_m))) +void vluxseg7ei8(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vint16mf4_t * op6, vbool64_t op7, vint16mf4_t op8, vint16mf4_t op9, vint16mf4_t op10, vint16mf4_t op11, vint16mf4_t op12, vint16mf4_t op13, vint16mf4_t op14, const int16_t * op15, vuint8mf8_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_u16m1))) +void vluxseg7ei8(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vuint16m1_t * op6, const uint16_t * op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_u16m1_m))) +void vluxseg7ei8(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vuint16m1_t * op6, vbool16_t op7, vuint16m1_t op8, vuint16m1_t op9, vuint16m1_t op10, vuint16m1_t op11, vuint16m1_t op12, vuint16m1_t op13, vuint16m1_t op14, const uint16_t * op15, vuint8mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_u16mf2))) +void vluxseg7ei8(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vuint16mf2_t * op6, const uint16_t * op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_u16mf2_m))) +void vluxseg7ei8(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vuint16mf2_t * op6, vbool32_t op7, vuint16mf2_t op8, vuint16mf2_t op9, vuint16mf2_t op10, vuint16mf2_t op11, vuint16mf2_t op12, vuint16mf2_t op13, vuint16mf2_t op14, const uint16_t * op15, vuint8mf4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_u16mf4))) +void vluxseg7ei8(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vuint16mf4_t * op6, const uint16_t * op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_u16mf4_m))) +void vluxseg7ei8(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vuint16mf4_t * op6, vbool64_t op7, vuint16mf4_t op8, vuint16mf4_t op9, vuint16mf4_t op10, vuint16mf4_t op11, vuint16mf4_t op12, vuint16mf4_t op13, vuint16mf4_t op14, const uint16_t * op15, vuint8mf8_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_i16m1))) +void vluxseg8ei8(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vint16m1_t * op6, vint16m1_t * op7, const int16_t * op8, vuint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_i16m1_m))) +void vluxseg8ei8(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vint16m1_t * op6, vint16m1_t * op7, vbool16_t op8, vint16m1_t op9, vint16m1_t op10, vint16m1_t op11, vint16m1_t op12, vint16m1_t op13, vint16m1_t op14, vint16m1_t op15, vint16m1_t op16, const int16_t * op17, vuint8mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_i16mf2))) +void vluxseg8ei8(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vint16mf2_t * op6, vint16mf2_t * op7, const int16_t * op8, vuint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_i16mf2_m))) +void vluxseg8ei8(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vint16mf2_t * op6, vint16mf2_t * op7, vbool32_t op8, vint16mf2_t op9, vint16mf2_t op10, vint16mf2_t op11, vint16mf2_t op12, vint16mf2_t op13, vint16mf2_t op14, vint16mf2_t op15, vint16mf2_t op16, const int16_t * op17, vuint8mf4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_i16mf4))) +void vluxseg8ei8(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vint16mf4_t * op6, vint16mf4_t * op7, const int16_t * op8, vuint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_i16mf4_m))) +void vluxseg8ei8(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vint16mf4_t * op6, vint16mf4_t * op7, vbool64_t op8, vint16mf4_t op9, vint16mf4_t op10, vint16mf4_t op11, vint16mf4_t op12, vint16mf4_t op13, vint16mf4_t op14, vint16mf4_t op15, vint16mf4_t op16, const int16_t * op17, vuint8mf8_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_u16m1))) +void vluxseg8ei8(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vuint16m1_t * op6, vuint16m1_t * op7, const uint16_t * op8, vuint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_u16m1_m))) +void vluxseg8ei8(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vuint16m1_t * op6, vuint16m1_t * op7, vbool16_t op8, vuint16m1_t op9, vuint16m1_t op10, vuint16m1_t op11, vuint16m1_t op12, vuint16m1_t op13, vuint16m1_t op14, vuint16m1_t op15, vuint16m1_t op16, const uint16_t * op17, vuint8mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_u16mf2))) +void vluxseg8ei8(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vuint16mf2_t * op6, vuint16mf2_t * op7, const uint16_t * op8, vuint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_u16mf2_m))) +void vluxseg8ei8(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vuint16mf2_t * op6, vuint16mf2_t * op7, vbool32_t op8, vuint16mf2_t op9, vuint16mf2_t op10, vuint16mf2_t op11, vuint16mf2_t op12, vuint16mf2_t op13, vuint16mf2_t op14, vuint16mf2_t op15, vuint16mf2_t op16, const uint16_t * op17, vuint8mf4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_u16mf4))) +void vluxseg8ei8(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vuint16mf4_t * op6, vuint16mf4_t * op7, const uint16_t * op8, vuint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_u16mf4_m))) +void vluxseg8ei8(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vuint16mf4_t * op6, vuint16mf4_t * op7, vbool64_t op8, vuint16mf4_t op9, vuint16mf4_t op10, vuint16mf4_t op11, vuint16mf4_t op12, vuint16mf4_t op13, vuint16mf4_t op14, vuint16mf4_t op15, vuint16mf4_t op16, const uint16_t * op17, vuint8mf8_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_i16m1))) +void vluxseg2ei16(vint16m1_t * op0, vint16m1_t * op1, const int16_t * op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_i16m1_m))) +void vluxseg2ei16(vint16m1_t * op0, vint16m1_t * op1, vbool16_t op2, vint16m1_t op3, vint16m1_t op4, const int16_t * op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_i16m2))) +void vluxseg2ei16(vint16m2_t * op0, vint16m2_t * op1, const int16_t * op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_i16m2_m))) +void vluxseg2ei16(vint16m2_t * op0, vint16m2_t * op1, vbool8_t op2, vint16m2_t op3, vint16m2_t op4, const int16_t * op5, vuint16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_i16m4))) +void vluxseg2ei16(vint16m4_t * op0, vint16m4_t * op1, const int16_t * op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_i16m4_m))) +void vluxseg2ei16(vint16m4_t * op0, vint16m4_t * op1, vbool4_t op2, vint16m4_t op3, vint16m4_t op4, const int16_t * op5, vuint16m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_i16mf2))) +void vluxseg2ei16(vint16mf2_t * op0, vint16mf2_t * op1, const int16_t * op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_i16mf2_m))) +void vluxseg2ei16(vint16mf2_t * op0, vint16mf2_t * op1, vbool32_t op2, vint16mf2_t op3, vint16mf2_t op4, const int16_t * op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_i16mf4))) +void vluxseg2ei16(vint16mf4_t * op0, vint16mf4_t * op1, const int16_t * op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_i16mf4_m))) +void vluxseg2ei16(vint16mf4_t * op0, vint16mf4_t * op1, vbool64_t op2, vint16mf4_t op3, vint16mf4_t op4, const int16_t * op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_u16m1))) +void vluxseg2ei16(vuint16m1_t * op0, vuint16m1_t * op1, const uint16_t * op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_u16m1_m))) +void vluxseg2ei16(vuint16m1_t * op0, vuint16m1_t * op1, vbool16_t op2, vuint16m1_t op3, vuint16m1_t op4, const uint16_t * op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_u16m2))) +void vluxseg2ei16(vuint16m2_t * op0, vuint16m2_t * op1, const uint16_t * op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_u16m2_m))) +void vluxseg2ei16(vuint16m2_t * op0, vuint16m2_t * op1, vbool8_t op2, vuint16m2_t op3, vuint16m2_t op4, const uint16_t * op5, vuint16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_u16m4))) +void vluxseg2ei16(vuint16m4_t * op0, vuint16m4_t * op1, const uint16_t * op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_u16m4_m))) +void vluxseg2ei16(vuint16m4_t * op0, vuint16m4_t * op1, vbool4_t op2, vuint16m4_t op3, vuint16m4_t op4, const uint16_t * op5, vuint16m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_u16mf2))) +void vluxseg2ei16(vuint16mf2_t * op0, vuint16mf2_t * op1, const uint16_t * op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_u16mf2_m))) +void vluxseg2ei16(vuint16mf2_t * op0, vuint16mf2_t * op1, vbool32_t op2, vuint16mf2_t op3, vuint16mf2_t op4, const uint16_t * op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_u16mf4))) +void vluxseg2ei16(vuint16mf4_t * op0, vuint16mf4_t * op1, const uint16_t * op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_u16mf4_m))) +void vluxseg2ei16(vuint16mf4_t * op0, vuint16mf4_t * op1, vbool64_t op2, vuint16mf4_t op3, vuint16mf4_t op4, const uint16_t * op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_i16m1))) +void vluxseg3ei16(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, const int16_t * op3, vuint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_i16m1_m))) +void vluxseg3ei16(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vbool16_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, const int16_t * op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_i16m2))) +void vluxseg3ei16(vint16m2_t * op0, vint16m2_t * op1, vint16m2_t * op2, const int16_t * op3, vuint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_i16m2_m))) +void vluxseg3ei16(vint16m2_t * op0, vint16m2_t * op1, vint16m2_t * op2, vbool8_t op3, vint16m2_t op4, vint16m2_t op5, vint16m2_t op6, const int16_t * op7, vuint16m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_i16mf2))) +void vluxseg3ei16(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, const int16_t * op3, vuint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_i16mf2_m))) +void vluxseg3ei16(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vbool32_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, const int16_t * op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_i16mf4))) +void vluxseg3ei16(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, const int16_t * op3, vuint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_i16mf4_m))) +void vluxseg3ei16(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vbool64_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, const int16_t * op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_u16m1))) +void vluxseg3ei16(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, const uint16_t * op3, vuint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_u16m1_m))) +void vluxseg3ei16(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vbool16_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, const uint16_t * op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_u16m2))) +void vluxseg3ei16(vuint16m2_t * op0, vuint16m2_t * op1, vuint16m2_t * op2, const uint16_t * op3, vuint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_u16m2_m))) +void vluxseg3ei16(vuint16m2_t * op0, vuint16m2_t * op1, vuint16m2_t * op2, vbool8_t op3, vuint16m2_t op4, vuint16m2_t op5, vuint16m2_t op6, const uint16_t * op7, vuint16m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_u16mf2))) +void vluxseg3ei16(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, const uint16_t * op3, vuint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_u16mf2_m))) +void vluxseg3ei16(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vbool32_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, const uint16_t * op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_u16mf4))) +void vluxseg3ei16(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, const uint16_t * op3, vuint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_u16mf4_m))) +void vluxseg3ei16(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vbool64_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, const uint16_t * op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_i16m1))) +void vluxseg4ei16(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, const int16_t * op4, vuint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_i16m1_m))) +void vluxseg4ei16(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vbool16_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, const int16_t * op9, vuint16m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_i16m2))) +void vluxseg4ei16(vint16m2_t * op0, vint16m2_t * op1, vint16m2_t * op2, vint16m2_t * op3, const int16_t * op4, vuint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_i16m2_m))) +void vluxseg4ei16(vint16m2_t * op0, vint16m2_t * op1, vint16m2_t * op2, vint16m2_t * op3, vbool8_t op4, vint16m2_t op5, vint16m2_t op6, vint16m2_t op7, vint16m2_t op8, const int16_t * op9, vuint16m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_i16mf2))) +void vluxseg4ei16(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, const int16_t * op4, vuint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_i16mf2_m))) +void vluxseg4ei16(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vbool32_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, const int16_t * op9, vuint16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_i16mf4))) +void vluxseg4ei16(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, const int16_t * op4, vuint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_i16mf4_m))) +void vluxseg4ei16(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vbool64_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, const int16_t * op9, vuint16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_u16m1))) +void vluxseg4ei16(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, const uint16_t * op4, vuint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_u16m1_m))) +void vluxseg4ei16(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vbool16_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, const uint16_t * op9, vuint16m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_u16m2))) +void vluxseg4ei16(vuint16m2_t * op0, vuint16m2_t * op1, vuint16m2_t * op2, vuint16m2_t * op3, const uint16_t * op4, vuint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_u16m2_m))) +void vluxseg4ei16(vuint16m2_t * op0, vuint16m2_t * op1, vuint16m2_t * op2, vuint16m2_t * op3, vbool8_t op4, vuint16m2_t op5, vuint16m2_t op6, vuint16m2_t op7, vuint16m2_t op8, const uint16_t * op9, vuint16m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_u16mf2))) +void vluxseg4ei16(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, const uint16_t * op4, vuint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_u16mf2_m))) +void vluxseg4ei16(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vbool32_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, const uint16_t * op9, vuint16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_u16mf4))) +void vluxseg4ei16(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, const uint16_t * op4, vuint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_u16mf4_m))) +void vluxseg4ei16(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vbool64_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, const uint16_t * op9, vuint16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_i16m1))) +void vluxseg5ei16(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, const int16_t * op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_i16m1_m))) +void vluxseg5ei16(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vbool16_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, vint16m1_t op10, const int16_t * op11, vuint16m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_i16mf2))) +void vluxseg5ei16(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, const int16_t * op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_i16mf2_m))) +void vluxseg5ei16(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vbool32_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, vint16mf2_t op10, const int16_t * op11, vuint16mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_i16mf4))) +void vluxseg5ei16(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, const int16_t * op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_i16mf4_m))) +void vluxseg5ei16(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vbool64_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, vint16mf4_t op10, const int16_t * op11, vuint16mf4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_u16m1))) +void vluxseg5ei16(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, const uint16_t * op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_u16m1_m))) +void vluxseg5ei16(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vbool16_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, vuint16m1_t op10, const uint16_t * op11, vuint16m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_u16mf2))) +void vluxseg5ei16(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, const uint16_t * op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_u16mf2_m))) +void vluxseg5ei16(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vbool32_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, vuint16mf2_t op10, const uint16_t * op11, vuint16mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_u16mf4))) +void vluxseg5ei16(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, const uint16_t * op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_u16mf4_m))) +void vluxseg5ei16(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vbool64_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, vuint16mf4_t op10, const uint16_t * op11, vuint16mf4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_i16m1))) +void vluxseg6ei16(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, const int16_t * op6, vuint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_i16m1_m))) +void vluxseg6ei16(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vbool16_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, vint16m1_t op10, vint16m1_t op11, vint16m1_t op12, const int16_t * op13, vuint16m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_i16mf2))) +void vluxseg6ei16(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, const int16_t * op6, vuint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_i16mf2_m))) +void vluxseg6ei16(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vbool32_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, vint16mf2_t op10, vint16mf2_t op11, vint16mf2_t op12, const int16_t * op13, vuint16mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_i16mf4))) +void vluxseg6ei16(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, const int16_t * op6, vuint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_i16mf4_m))) +void vluxseg6ei16(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vbool64_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, vint16mf4_t op10, vint16mf4_t op11, vint16mf4_t op12, const int16_t * op13, vuint16mf4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_u16m1))) +void vluxseg6ei16(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, const uint16_t * op6, vuint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_u16m1_m))) +void vluxseg6ei16(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vbool16_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, vuint16m1_t op10, vuint16m1_t op11, vuint16m1_t op12, const uint16_t * op13, vuint16m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_u16mf2))) +void vluxseg6ei16(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, const uint16_t * op6, vuint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_u16mf2_m))) +void vluxseg6ei16(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vbool32_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, vuint16mf2_t op10, vuint16mf2_t op11, vuint16mf2_t op12, const uint16_t * op13, vuint16mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_u16mf4))) +void vluxseg6ei16(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, const uint16_t * op6, vuint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_u16mf4_m))) +void vluxseg6ei16(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vbool64_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, vuint16mf4_t op10, vuint16mf4_t op11, vuint16mf4_t op12, const uint16_t * op13, vuint16mf4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_i16m1))) +void vluxseg7ei16(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vint16m1_t * op6, const int16_t * op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_i16m1_m))) +void vluxseg7ei16(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vint16m1_t * op6, vbool16_t op7, vint16m1_t op8, vint16m1_t op9, vint16m1_t op10, vint16m1_t op11, vint16m1_t op12, vint16m1_t op13, vint16m1_t op14, const int16_t * op15, vuint16m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_i16mf2))) +void vluxseg7ei16(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vint16mf2_t * op6, const int16_t * op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_i16mf2_m))) +void vluxseg7ei16(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vint16mf2_t * op6, vbool32_t op7, vint16mf2_t op8, vint16mf2_t op9, vint16mf2_t op10, vint16mf2_t op11, vint16mf2_t op12, vint16mf2_t op13, vint16mf2_t op14, const int16_t * op15, vuint16mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_i16mf4))) +void vluxseg7ei16(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vint16mf4_t * op6, const int16_t * op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_i16mf4_m))) +void vluxseg7ei16(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vint16mf4_t * op6, vbool64_t op7, vint16mf4_t op8, vint16mf4_t op9, vint16mf4_t op10, vint16mf4_t op11, vint16mf4_t op12, vint16mf4_t op13, vint16mf4_t op14, const int16_t * op15, vuint16mf4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_u16m1))) +void vluxseg7ei16(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vuint16m1_t * op6, const uint16_t * op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_u16m1_m))) +void vluxseg7ei16(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vuint16m1_t * op6, vbool16_t op7, vuint16m1_t op8, vuint16m1_t op9, vuint16m1_t op10, vuint16m1_t op11, vuint16m1_t op12, vuint16m1_t op13, vuint16m1_t op14, const uint16_t * op15, vuint16m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_u16mf2))) +void vluxseg7ei16(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vuint16mf2_t * op6, const uint16_t * op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_u16mf2_m))) +void vluxseg7ei16(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vuint16mf2_t * op6, vbool32_t op7, vuint16mf2_t op8, vuint16mf2_t op9, vuint16mf2_t op10, vuint16mf2_t op11, vuint16mf2_t op12, vuint16mf2_t op13, vuint16mf2_t op14, const uint16_t * op15, vuint16mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_u16mf4))) +void vluxseg7ei16(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vuint16mf4_t * op6, const uint16_t * op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_u16mf4_m))) +void vluxseg7ei16(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vuint16mf4_t * op6, vbool64_t op7, vuint16mf4_t op8, vuint16mf4_t op9, vuint16mf4_t op10, vuint16mf4_t op11, vuint16mf4_t op12, vuint16mf4_t op13, vuint16mf4_t op14, const uint16_t * op15, vuint16mf4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_i16m1))) +void vluxseg8ei16(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vint16m1_t * op6, vint16m1_t * op7, const int16_t * op8, vuint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_i16m1_m))) +void vluxseg8ei16(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vint16m1_t * op6, vint16m1_t * op7, vbool16_t op8, vint16m1_t op9, vint16m1_t op10, vint16m1_t op11, vint16m1_t op12, vint16m1_t op13, vint16m1_t op14, vint16m1_t op15, vint16m1_t op16, const int16_t * op17, vuint16m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_i16mf2))) +void vluxseg8ei16(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vint16mf2_t * op6, vint16mf2_t * op7, const int16_t * op8, vuint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_i16mf2_m))) +void vluxseg8ei16(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vint16mf2_t * op6, vint16mf2_t * op7, vbool32_t op8, vint16mf2_t op9, vint16mf2_t op10, vint16mf2_t op11, vint16mf2_t op12, vint16mf2_t op13, vint16mf2_t op14, vint16mf2_t op15, vint16mf2_t op16, const int16_t * op17, vuint16mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_i16mf4))) +void vluxseg8ei16(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vint16mf4_t * op6, vint16mf4_t * op7, const int16_t * op8, vuint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_i16mf4_m))) +void vluxseg8ei16(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vint16mf4_t * op6, vint16mf4_t * op7, vbool64_t op8, vint16mf4_t op9, vint16mf4_t op10, vint16mf4_t op11, vint16mf4_t op12, vint16mf4_t op13, vint16mf4_t op14, vint16mf4_t op15, vint16mf4_t op16, const int16_t * op17, vuint16mf4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_u16m1))) +void vluxseg8ei16(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vuint16m1_t * op6, vuint16m1_t * op7, const uint16_t * op8, vuint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_u16m1_m))) +void vluxseg8ei16(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vuint16m1_t * op6, vuint16m1_t * op7, vbool16_t op8, vuint16m1_t op9, vuint16m1_t op10, vuint16m1_t op11, vuint16m1_t op12, vuint16m1_t op13, vuint16m1_t op14, vuint16m1_t op15, vuint16m1_t op16, const uint16_t * op17, vuint16m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_u16mf2))) +void vluxseg8ei16(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vuint16mf2_t * op6, vuint16mf2_t * op7, const uint16_t * op8, vuint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_u16mf2_m))) +void vluxseg8ei16(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vuint16mf2_t * op6, vuint16mf2_t * op7, vbool32_t op8, vuint16mf2_t op9, vuint16mf2_t op10, vuint16mf2_t op11, vuint16mf2_t op12, vuint16mf2_t op13, vuint16mf2_t op14, vuint16mf2_t op15, vuint16mf2_t op16, const uint16_t * op17, vuint16mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_u16mf4))) +void vluxseg8ei16(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vuint16mf4_t * op6, vuint16mf4_t * op7, const uint16_t * op8, vuint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_u16mf4_m))) +void vluxseg8ei16(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vuint16mf4_t * op6, vuint16mf4_t * op7, vbool64_t op8, vuint16mf4_t op9, vuint16mf4_t op10, vuint16mf4_t op11, vuint16mf4_t op12, vuint16mf4_t op13, vuint16mf4_t op14, vuint16mf4_t op15, vuint16mf4_t op16, const uint16_t * op17, vuint16mf4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_i16m1))) +void vluxseg2ei32(vint16m1_t * op0, vint16m1_t * op1, const int16_t * op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_i16m1_m))) +void vluxseg2ei32(vint16m1_t * op0, vint16m1_t * op1, vbool16_t op2, vint16m1_t op3, vint16m1_t op4, const int16_t * op5, vuint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_i16m2))) +void vluxseg2ei32(vint16m2_t * op0, vint16m2_t * op1, const int16_t * op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_i16m2_m))) +void vluxseg2ei32(vint16m2_t * op0, vint16m2_t * op1, vbool8_t op2, vint16m2_t op3, vint16m2_t op4, const int16_t * op5, vuint32m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_i16m4))) +void vluxseg2ei32(vint16m4_t * op0, vint16m4_t * op1, const int16_t * op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_i16m4_m))) +void vluxseg2ei32(vint16m4_t * op0, vint16m4_t * op1, vbool4_t op2, vint16m4_t op3, vint16m4_t op4, const int16_t * op5, vuint32m8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_i16mf2))) +void vluxseg2ei32(vint16mf2_t * op0, vint16mf2_t * op1, const int16_t * op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_i16mf2_m))) +void vluxseg2ei32(vint16mf2_t * op0, vint16mf2_t * op1, vbool32_t op2, vint16mf2_t op3, vint16mf2_t op4, const int16_t * op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_i16mf4))) +void vluxseg2ei32(vint16mf4_t * op0, vint16mf4_t * op1, const int16_t * op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_i16mf4_m))) +void vluxseg2ei32(vint16mf4_t * op0, vint16mf4_t * op1, vbool64_t op2, vint16mf4_t op3, vint16mf4_t op4, const int16_t * op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_u16m1))) +void vluxseg2ei32(vuint16m1_t * op0, vuint16m1_t * op1, const uint16_t * op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_u16m1_m))) +void vluxseg2ei32(vuint16m1_t * op0, vuint16m1_t * op1, vbool16_t op2, vuint16m1_t op3, vuint16m1_t op4, const uint16_t * op5, vuint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_u16m2))) +void vluxseg2ei32(vuint16m2_t * op0, vuint16m2_t * op1, const uint16_t * op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_u16m2_m))) +void vluxseg2ei32(vuint16m2_t * op0, vuint16m2_t * op1, vbool8_t op2, vuint16m2_t op3, vuint16m2_t op4, const uint16_t * op5, vuint32m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_u16m4))) +void vluxseg2ei32(vuint16m4_t * op0, vuint16m4_t * op1, const uint16_t * op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_u16m4_m))) +void vluxseg2ei32(vuint16m4_t * op0, vuint16m4_t * op1, vbool4_t op2, vuint16m4_t op3, vuint16m4_t op4, const uint16_t * op5, vuint32m8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_u16mf2))) +void vluxseg2ei32(vuint16mf2_t * op0, vuint16mf2_t * op1, const uint16_t * op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_u16mf2_m))) +void vluxseg2ei32(vuint16mf2_t * op0, vuint16mf2_t * op1, vbool32_t op2, vuint16mf2_t op3, vuint16mf2_t op4, const uint16_t * op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_u16mf4))) +void vluxseg2ei32(vuint16mf4_t * op0, vuint16mf4_t * op1, const uint16_t * op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_u16mf4_m))) +void vluxseg2ei32(vuint16mf4_t * op0, vuint16mf4_t * op1, vbool64_t op2, vuint16mf4_t op3, vuint16mf4_t op4, const uint16_t * op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_i16m1))) +void vluxseg3ei32(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, const int16_t * op3, vuint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_i16m1_m))) +void vluxseg3ei32(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vbool16_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, const int16_t * op7, vuint32m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_i16m2))) +void vluxseg3ei32(vint16m2_t * op0, vint16m2_t * op1, vint16m2_t * op2, const int16_t * op3, vuint32m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_i16m2_m))) +void vluxseg3ei32(vint16m2_t * op0, vint16m2_t * op1, vint16m2_t * op2, vbool8_t op3, vint16m2_t op4, vint16m2_t op5, vint16m2_t op6, const int16_t * op7, vuint32m4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_i16mf2))) +void vluxseg3ei32(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, const int16_t * op3, vuint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_i16mf2_m))) +void vluxseg3ei32(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vbool32_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, const int16_t * op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_i16mf4))) +void vluxseg3ei32(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, const int16_t * op3, vuint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_i16mf4_m))) +void vluxseg3ei32(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vbool64_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, const int16_t * op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_u16m1))) +void vluxseg3ei32(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, const uint16_t * op3, vuint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_u16m1_m))) +void vluxseg3ei32(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vbool16_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, const uint16_t * op7, vuint32m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_u16m2))) +void vluxseg3ei32(vuint16m2_t * op0, vuint16m2_t * op1, vuint16m2_t * op2, const uint16_t * op3, vuint32m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_u16m2_m))) +void vluxseg3ei32(vuint16m2_t * op0, vuint16m2_t * op1, vuint16m2_t * op2, vbool8_t op3, vuint16m2_t op4, vuint16m2_t op5, vuint16m2_t op6, const uint16_t * op7, vuint32m4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_u16mf2))) +void vluxseg3ei32(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, const uint16_t * op3, vuint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_u16mf2_m))) +void vluxseg3ei32(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vbool32_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, const uint16_t * op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_u16mf4))) +void vluxseg3ei32(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, const uint16_t * op3, vuint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_u16mf4_m))) +void vluxseg3ei32(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vbool64_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, const uint16_t * op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_i16m1))) +void vluxseg4ei32(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, const int16_t * op4, vuint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_i16m1_m))) +void vluxseg4ei32(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vbool16_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, const int16_t * op9, vuint32m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_i16m2))) +void vluxseg4ei32(vint16m2_t * op0, vint16m2_t * op1, vint16m2_t * op2, vint16m2_t * op3, const int16_t * op4, vuint32m4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_i16m2_m))) +void vluxseg4ei32(vint16m2_t * op0, vint16m2_t * op1, vint16m2_t * op2, vint16m2_t * op3, vbool8_t op4, vint16m2_t op5, vint16m2_t op6, vint16m2_t op7, vint16m2_t op8, const int16_t * op9, vuint32m4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_i16mf2))) +void vluxseg4ei32(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, const int16_t * op4, vuint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_i16mf2_m))) +void vluxseg4ei32(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vbool32_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, const int16_t * op9, vuint32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_i16mf4))) +void vluxseg4ei32(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, const int16_t * op4, vuint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_i16mf4_m))) +void vluxseg4ei32(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vbool64_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, const int16_t * op9, vuint32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_u16m1))) +void vluxseg4ei32(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, const uint16_t * op4, vuint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_u16m1_m))) +void vluxseg4ei32(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vbool16_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, const uint16_t * op9, vuint32m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_u16m2))) +void vluxseg4ei32(vuint16m2_t * op0, vuint16m2_t * op1, vuint16m2_t * op2, vuint16m2_t * op3, const uint16_t * op4, vuint32m4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_u16m2_m))) +void vluxseg4ei32(vuint16m2_t * op0, vuint16m2_t * op1, vuint16m2_t * op2, vuint16m2_t * op3, vbool8_t op4, vuint16m2_t op5, vuint16m2_t op6, vuint16m2_t op7, vuint16m2_t op8, const uint16_t * op9, vuint32m4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_u16mf2))) +void vluxseg4ei32(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, const uint16_t * op4, vuint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_u16mf2_m))) +void vluxseg4ei32(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vbool32_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, const uint16_t * op9, vuint32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_u16mf4))) +void vluxseg4ei32(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, const uint16_t * op4, vuint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_u16mf4_m))) +void vluxseg4ei32(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vbool64_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, const uint16_t * op9, vuint32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_i16m1))) +void vluxseg5ei32(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, const int16_t * op5, vuint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_i16m1_m))) +void vluxseg5ei32(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vbool16_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, vint16m1_t op10, const int16_t * op11, vuint32m2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_i16mf2))) +void vluxseg5ei32(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, const int16_t * op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_i16mf2_m))) +void vluxseg5ei32(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vbool32_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, vint16mf2_t op10, const int16_t * op11, vuint32m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_i16mf4))) +void vluxseg5ei32(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, const int16_t * op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_i16mf4_m))) +void vluxseg5ei32(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vbool64_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, vint16mf4_t op10, const int16_t * op11, vuint32mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_u16m1))) +void vluxseg5ei32(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, const uint16_t * op5, vuint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_u16m1_m))) +void vluxseg5ei32(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vbool16_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, vuint16m1_t op10, const uint16_t * op11, vuint32m2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_u16mf2))) +void vluxseg5ei32(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, const uint16_t * op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_u16mf2_m))) +void vluxseg5ei32(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vbool32_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, vuint16mf2_t op10, const uint16_t * op11, vuint32m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_u16mf4))) +void vluxseg5ei32(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, const uint16_t * op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_u16mf4_m))) +void vluxseg5ei32(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vbool64_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, vuint16mf4_t op10, const uint16_t * op11, vuint32mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_i16m1))) +void vluxseg6ei32(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, const int16_t * op6, vuint32m2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_i16m1_m))) +void vluxseg6ei32(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vbool16_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, vint16m1_t op10, vint16m1_t op11, vint16m1_t op12, const int16_t * op13, vuint32m2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_i16mf2))) +void vluxseg6ei32(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, const int16_t * op6, vuint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_i16mf2_m))) +void vluxseg6ei32(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vbool32_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, vint16mf2_t op10, vint16mf2_t op11, vint16mf2_t op12, const int16_t * op13, vuint32m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_i16mf4))) +void vluxseg6ei32(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, const int16_t * op6, vuint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_i16mf4_m))) +void vluxseg6ei32(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vbool64_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, vint16mf4_t op10, vint16mf4_t op11, vint16mf4_t op12, const int16_t * op13, vuint32mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_u16m1))) +void vluxseg6ei32(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, const uint16_t * op6, vuint32m2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_u16m1_m))) +void vluxseg6ei32(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vbool16_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, vuint16m1_t op10, vuint16m1_t op11, vuint16m1_t op12, const uint16_t * op13, vuint32m2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_u16mf2))) +void vluxseg6ei32(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, const uint16_t * op6, vuint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_u16mf2_m))) +void vluxseg6ei32(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vbool32_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, vuint16mf2_t op10, vuint16mf2_t op11, vuint16mf2_t op12, const uint16_t * op13, vuint32m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_u16mf4))) +void vluxseg6ei32(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, const uint16_t * op6, vuint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_u16mf4_m))) +void vluxseg6ei32(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vbool64_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, vuint16mf4_t op10, vuint16mf4_t op11, vuint16mf4_t op12, const uint16_t * op13, vuint32mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_i16m1))) +void vluxseg7ei32(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vint16m1_t * op6, const int16_t * op7, vuint32m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_i16m1_m))) +void vluxseg7ei32(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vint16m1_t * op6, vbool16_t op7, vint16m1_t op8, vint16m1_t op9, vint16m1_t op10, vint16m1_t op11, vint16m1_t op12, vint16m1_t op13, vint16m1_t op14, const int16_t * op15, vuint32m2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_i16mf2))) +void vluxseg7ei32(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vint16mf2_t * op6, const int16_t * op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_i16mf2_m))) +void vluxseg7ei32(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vint16mf2_t * op6, vbool32_t op7, vint16mf2_t op8, vint16mf2_t op9, vint16mf2_t op10, vint16mf2_t op11, vint16mf2_t op12, vint16mf2_t op13, vint16mf2_t op14, const int16_t * op15, vuint32m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_i16mf4))) +void vluxseg7ei32(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vint16mf4_t * op6, const int16_t * op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_i16mf4_m))) +void vluxseg7ei32(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vint16mf4_t * op6, vbool64_t op7, vint16mf4_t op8, vint16mf4_t op9, vint16mf4_t op10, vint16mf4_t op11, vint16mf4_t op12, vint16mf4_t op13, vint16mf4_t op14, const int16_t * op15, vuint32mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_u16m1))) +void vluxseg7ei32(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vuint16m1_t * op6, const uint16_t * op7, vuint32m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_u16m1_m))) +void vluxseg7ei32(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vuint16m1_t * op6, vbool16_t op7, vuint16m1_t op8, vuint16m1_t op9, vuint16m1_t op10, vuint16m1_t op11, vuint16m1_t op12, vuint16m1_t op13, vuint16m1_t op14, const uint16_t * op15, vuint32m2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_u16mf2))) +void vluxseg7ei32(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vuint16mf2_t * op6, const uint16_t * op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_u16mf2_m))) +void vluxseg7ei32(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vuint16mf2_t * op6, vbool32_t op7, vuint16mf2_t op8, vuint16mf2_t op9, vuint16mf2_t op10, vuint16mf2_t op11, vuint16mf2_t op12, vuint16mf2_t op13, vuint16mf2_t op14, const uint16_t * op15, vuint32m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_u16mf4))) +void vluxseg7ei32(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vuint16mf4_t * op6, const uint16_t * op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_u16mf4_m))) +void vluxseg7ei32(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vuint16mf4_t * op6, vbool64_t op7, vuint16mf4_t op8, vuint16mf4_t op9, vuint16mf4_t op10, vuint16mf4_t op11, vuint16mf4_t op12, vuint16mf4_t op13, vuint16mf4_t op14, const uint16_t * op15, vuint32mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_i16m1))) +void vluxseg8ei32(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vint16m1_t * op6, vint16m1_t * op7, const int16_t * op8, vuint32m2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_i16m1_m))) +void vluxseg8ei32(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vint16m1_t * op6, vint16m1_t * op7, vbool16_t op8, vint16m1_t op9, vint16m1_t op10, vint16m1_t op11, vint16m1_t op12, vint16m1_t op13, vint16m1_t op14, vint16m1_t op15, vint16m1_t op16, const int16_t * op17, vuint32m2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_i16mf2))) +void vluxseg8ei32(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vint16mf2_t * op6, vint16mf2_t * op7, const int16_t * op8, vuint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_i16mf2_m))) +void vluxseg8ei32(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vint16mf2_t * op6, vint16mf2_t * op7, vbool32_t op8, vint16mf2_t op9, vint16mf2_t op10, vint16mf2_t op11, vint16mf2_t op12, vint16mf2_t op13, vint16mf2_t op14, vint16mf2_t op15, vint16mf2_t op16, const int16_t * op17, vuint32m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_i16mf4))) +void vluxseg8ei32(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vint16mf4_t * op6, vint16mf4_t * op7, const int16_t * op8, vuint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_i16mf4_m))) +void vluxseg8ei32(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vint16mf4_t * op6, vint16mf4_t * op7, vbool64_t op8, vint16mf4_t op9, vint16mf4_t op10, vint16mf4_t op11, vint16mf4_t op12, vint16mf4_t op13, vint16mf4_t op14, vint16mf4_t op15, vint16mf4_t op16, const int16_t * op17, vuint32mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e8_v_i8m1))) +void vssseg2e8(int8_t * op0, ptrdiff_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e8_v_i8m1_m))) +void vssseg2e8(vbool8_t op0, int8_t * op1, ptrdiff_t op2, vint8m1_t op3, vint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e8_v_i8m2))) +void vssseg2e8(int8_t * op0, ptrdiff_t op1, vint8m2_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e8_v_i8m2_m))) +void vssseg2e8(vbool4_t op0, int8_t * op1, ptrdiff_t op2, vint8m2_t op3, vint8m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e8_v_i8m4))) +void vssseg2e8(int8_t * op0, ptrdiff_t op1, vint8m4_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e8_v_i8m4_m))) +void vssseg2e8(vbool2_t op0, int8_t * op1, ptrdiff_t op2, vint8m4_t op3, vint8m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e8_v_i8mf2))) +void vssseg2e8(int8_t * op0, ptrdiff_t op1, vint8mf2_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e8_v_i8mf2_m))) +void vssseg2e8(vbool16_t op0, int8_t * op1, ptrdiff_t op2, vint8mf2_t op3, vint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e8_v_i8mf4))) +void vssseg2e8(int8_t * op0, ptrdiff_t op1, vint8mf4_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e8_v_i8mf4_m))) +void vssseg2e8(vbool32_t op0, int8_t * op1, ptrdiff_t op2, vint8mf4_t op3, vint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e8_v_i8mf8))) +void vssseg2e8(int8_t * op0, ptrdiff_t op1, vint8mf8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e8_v_i8mf8_m))) +void vssseg2e8(vbool64_t op0, int8_t * op1, ptrdiff_t op2, vint8mf8_t op3, vint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_u16m1))) +void vluxseg8ei32(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vuint16m1_t * op6, vuint16m1_t * op7, const uint16_t * op8, vuint32m2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_u16m1_m))) +void vluxseg8ei32(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vuint16m1_t * op6, vuint16m1_t * op7, vbool16_t op8, vuint16m1_t op9, vuint16m1_t op10, vuint16m1_t op11, vuint16m1_t op12, vuint16m1_t op13, vuint16m1_t op14, vuint16m1_t op15, vuint16m1_t op16, const uint16_t * op17, vuint32m2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_u16mf2))) +void vluxseg8ei32(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vuint16mf2_t * op6, vuint16mf2_t * op7, const uint16_t * op8, vuint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_u16mf2_m))) +void vluxseg8ei32(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vuint16mf2_t * op6, vuint16mf2_t * op7, vbool32_t op8, vuint16mf2_t op9, vuint16mf2_t op10, vuint16mf2_t op11, vuint16mf2_t op12, vuint16mf2_t op13, vuint16mf2_t op14, vuint16mf2_t op15, vuint16mf2_t op16, const uint16_t * op17, vuint32m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_u16mf4))) +void vluxseg8ei32(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vuint16mf4_t * op6, vuint16mf4_t * op7, const uint16_t * op8, vuint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_u16mf4_m))) +void vluxseg8ei32(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vuint16mf4_t * op6, vuint16mf4_t * op7, vbool64_t op8, vuint16mf4_t op9, vuint16mf4_t op10, vuint16mf4_t op11, vuint16mf4_t op12, vuint16mf4_t op13, vuint16mf4_t op14, vuint16mf4_t op15, vuint16mf4_t op16, const uint16_t * op17, vuint32mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_i16m1))) +void vluxseg2ei64(vint16m1_t * op0, vint16m1_t * op1, const int16_t * op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_i16m1_m))) +void vluxseg2ei64(vint16m1_t * op0, vint16m1_t * op1, vbool16_t op2, vint16m1_t op3, vint16m1_t op4, const int16_t * op5, vuint64m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_i16m2))) +void vluxseg2ei64(vint16m2_t * op0, vint16m2_t * op1, const int16_t * op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_i16m2_m))) +void vluxseg2ei64(vint16m2_t * op0, vint16m2_t * op1, vbool8_t op2, vint16m2_t op3, vint16m2_t op4, const int16_t * op5, vuint64m8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_i16mf2))) +void vluxseg2ei64(vint16mf2_t * op0, vint16mf2_t * op1, const int16_t * op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_i16mf2_m))) +void vluxseg2ei64(vint16mf2_t * op0, vint16mf2_t * op1, vbool32_t op2, vint16mf2_t op3, vint16mf2_t op4, const int16_t * op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_i16mf4))) +void vluxseg2ei64(vint16mf4_t * op0, vint16mf4_t * op1, const int16_t * op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_i16mf4_m))) +void vluxseg2ei64(vint16mf4_t * op0, vint16mf4_t * op1, vbool64_t op2, vint16mf4_t op3, vint16mf4_t op4, const int16_t * op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_u16m1))) +void vluxseg2ei64(vuint16m1_t * op0, vuint16m1_t * op1, const uint16_t * op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_u16m1_m))) +void vluxseg2ei64(vuint16m1_t * op0, vuint16m1_t * op1, vbool16_t op2, vuint16m1_t op3, vuint16m1_t op4, const uint16_t * op5, vuint64m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_u16m2))) +void vluxseg2ei64(vuint16m2_t * op0, vuint16m2_t * op1, const uint16_t * op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_u16m2_m))) +void vluxseg2ei64(vuint16m2_t * op0, vuint16m2_t * op1, vbool8_t op2, vuint16m2_t op3, vuint16m2_t op4, const uint16_t * op5, vuint64m8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_u16mf2))) +void vluxseg2ei64(vuint16mf2_t * op0, vuint16mf2_t * op1, const uint16_t * op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_u16mf2_m))) +void vluxseg2ei64(vuint16mf2_t * op0, vuint16mf2_t * op1, vbool32_t op2, vuint16mf2_t op3, vuint16mf2_t op4, const uint16_t * op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_u16mf4))) +void vluxseg2ei64(vuint16mf4_t * op0, vuint16mf4_t * op1, const uint16_t * op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_u16mf4_m))) +void vluxseg2ei64(vuint16mf4_t * op0, vuint16mf4_t * op1, vbool64_t op2, vuint16mf4_t op3, vuint16mf4_t op4, const uint16_t * op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_i16m1))) +void vluxseg3ei64(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, const int16_t * op3, vuint64m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_i16m1_m))) +void vluxseg3ei64(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vbool16_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, const int16_t * op7, vuint64m4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_i16m2))) +void vluxseg3ei64(vint16m2_t * op0, vint16m2_t * op1, vint16m2_t * op2, const int16_t * op3, vuint64m8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_i16m2_m))) +void vluxseg3ei64(vint16m2_t * op0, vint16m2_t * op1, vint16m2_t * op2, vbool8_t op3, vint16m2_t op4, vint16m2_t op5, vint16m2_t op6, const int16_t * op7, vuint64m8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_i16mf2))) +void vluxseg3ei64(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, const int16_t * op3, vuint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_i16mf2_m))) +void vluxseg3ei64(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vbool32_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, const int16_t * op7, vuint64m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_i16mf4))) +void vluxseg3ei64(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, const int16_t * op3, vuint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_i16mf4_m))) +void vluxseg3ei64(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vbool64_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, const int16_t * op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_u16m1))) +void vluxseg3ei64(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, const uint16_t * op3, vuint64m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_u16m1_m))) +void vluxseg3ei64(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vbool16_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, const uint16_t * op7, vuint64m4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_u16m2))) +void vluxseg3ei64(vuint16m2_t * op0, vuint16m2_t * op1, vuint16m2_t * op2, const uint16_t * op3, vuint64m8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_u16m2_m))) +void vluxseg3ei64(vuint16m2_t * op0, vuint16m2_t * op1, vuint16m2_t * op2, vbool8_t op3, vuint16m2_t op4, vuint16m2_t op5, vuint16m2_t op6, const uint16_t * op7, vuint64m8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_u16mf2))) +void vluxseg3ei64(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, const uint16_t * op3, vuint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_u16mf2_m))) +void vluxseg3ei64(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vbool32_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, const uint16_t * op7, vuint64m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_u16mf4))) +void vluxseg3ei64(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, const uint16_t * op3, vuint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_u16mf4_m))) +void vluxseg3ei64(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vbool64_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, const uint16_t * op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_i16m1))) +void vluxseg4ei64(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, const int16_t * op4, vuint64m4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_i16m1_m))) +void vluxseg4ei64(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vbool16_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, const int16_t * op9, vuint64m4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_i16m2))) +void vluxseg4ei64(vint16m2_t * op0, vint16m2_t * op1, vint16m2_t * op2, vint16m2_t * op3, const int16_t * op4, vuint64m8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_i16m2_m))) +void vluxseg4ei64(vint16m2_t * op0, vint16m2_t * op1, vint16m2_t * op2, vint16m2_t * op3, vbool8_t op4, vint16m2_t op5, vint16m2_t op6, vint16m2_t op7, vint16m2_t op8, const int16_t * op9, vuint64m8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_i16mf2))) +void vluxseg4ei64(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, const int16_t * op4, vuint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_i16mf2_m))) +void vluxseg4ei64(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vbool32_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, const int16_t * op9, vuint64m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_i16mf4))) +void vluxseg4ei64(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, const int16_t * op4, vuint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_i16mf4_m))) +void vluxseg4ei64(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vbool64_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, const int16_t * op9, vuint64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_u16m1))) +void vluxseg4ei64(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, const uint16_t * op4, vuint64m4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_u16m1_m))) +void vluxseg4ei64(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vbool16_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, const uint16_t * op9, vuint64m4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_u16m2))) +void vluxseg4ei64(vuint16m2_t * op0, vuint16m2_t * op1, vuint16m2_t * op2, vuint16m2_t * op3, const uint16_t * op4, vuint64m8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_u16m2_m))) +void vluxseg4ei64(vuint16m2_t * op0, vuint16m2_t * op1, vuint16m2_t * op2, vuint16m2_t * op3, vbool8_t op4, vuint16m2_t op5, vuint16m2_t op6, vuint16m2_t op7, vuint16m2_t op8, const uint16_t * op9, vuint64m8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_u16mf2))) +void vluxseg4ei64(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, const uint16_t * op4, vuint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_u16mf2_m))) +void vluxseg4ei64(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vbool32_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, const uint16_t * op9, vuint64m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_u16mf4))) +void vluxseg4ei64(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, const uint16_t * op4, vuint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_u16mf4_m))) +void vluxseg4ei64(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vbool64_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, const uint16_t * op9, vuint64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_i16m1))) +void vluxseg5ei64(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, const int16_t * op5, vuint64m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_i16m1_m))) +void vluxseg5ei64(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vbool16_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, vint16m1_t op10, const int16_t * op11, vuint64m4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_i16mf2))) +void vluxseg5ei64(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, const int16_t * op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_i16mf2_m))) +void vluxseg5ei64(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vbool32_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, vint16mf2_t op10, const int16_t * op11, vuint64m2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_i16mf4))) +void vluxseg5ei64(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, const int16_t * op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_i16mf4_m))) +void vluxseg5ei64(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vbool64_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, vint16mf4_t op10, const int16_t * op11, vuint64m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_u16m1))) +void vluxseg5ei64(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, const uint16_t * op5, vuint64m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_u16m1_m))) +void vluxseg5ei64(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vbool16_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, vuint16m1_t op10, const uint16_t * op11, vuint64m4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_u16mf2))) +void vluxseg5ei64(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, const uint16_t * op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_u16mf2_m))) +void vluxseg5ei64(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vbool32_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, vuint16mf2_t op10, const uint16_t * op11, vuint64m2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_u16mf4))) +void vluxseg5ei64(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, const uint16_t * op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_u16mf4_m))) +void vluxseg5ei64(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vbool64_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, vuint16mf4_t op10, const uint16_t * op11, vuint64m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_i16m1))) +void vluxseg6ei64(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, const int16_t * op6, vuint64m4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_i16m1_m))) +void vluxseg6ei64(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vbool16_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, vint16m1_t op10, vint16m1_t op11, vint16m1_t op12, const int16_t * op13, vuint64m4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_i16mf2))) +void vluxseg6ei64(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, const int16_t * op6, vuint64m2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_i16mf2_m))) +void vluxseg6ei64(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vbool32_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, vint16mf2_t op10, vint16mf2_t op11, vint16mf2_t op12, const int16_t * op13, vuint64m2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_i16mf4))) +void vluxseg6ei64(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, const int16_t * op6, vuint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_i16mf4_m))) +void vluxseg6ei64(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vbool64_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, vint16mf4_t op10, vint16mf4_t op11, vint16mf4_t op12, const int16_t * op13, vuint64m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e8_v_u8m1))) +void vssseg2e8(uint8_t * op0, ptrdiff_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e8_v_u8m1_m))) +void vssseg2e8(vbool8_t op0, uint8_t * op1, ptrdiff_t op2, vuint8m1_t op3, vuint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e8_v_u8m2))) +void vssseg2e8(uint8_t * op0, ptrdiff_t op1, vuint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e8_v_u8m2_m))) +void vssseg2e8(vbool4_t op0, uint8_t * op1, ptrdiff_t op2, vuint8m2_t op3, vuint8m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e8_v_u8m4))) +void vssseg2e8(uint8_t * op0, ptrdiff_t op1, vuint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e8_v_u8m4_m))) +void vssseg2e8(vbool2_t op0, uint8_t * op1, ptrdiff_t op2, vuint8m4_t op3, vuint8m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e8_v_u8mf2))) +void vssseg2e8(uint8_t * op0, ptrdiff_t op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e8_v_u8mf2_m))) +void vssseg2e8(vbool16_t op0, uint8_t * op1, ptrdiff_t op2, vuint8mf2_t op3, vuint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e8_v_u8mf4))) +void vssseg2e8(uint8_t * op0, ptrdiff_t op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e8_v_u8mf4_m))) +void vssseg2e8(vbool32_t op0, uint8_t * op1, ptrdiff_t op2, vuint8mf4_t op3, vuint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e8_v_u8mf8))) +void vssseg2e8(uint8_t * op0, ptrdiff_t op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e8_v_u8mf8_m))) +void vssseg2e8(vbool64_t op0, uint8_t * op1, ptrdiff_t op2, vuint8mf8_t op3, vuint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_u16m1))) +void vluxseg6ei64(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, const uint16_t * op6, vuint64m4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_u16m1_m))) +void vluxseg6ei64(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vbool16_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, vuint16m1_t op10, vuint16m1_t op11, vuint16m1_t op12, const uint16_t * op13, vuint64m4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_u16mf2))) +void vluxseg6ei64(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, const uint16_t * op6, vuint64m2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_u16mf2_m))) +void vluxseg6ei64(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vbool32_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, vuint16mf2_t op10, vuint16mf2_t op11, vuint16mf2_t op12, const uint16_t * op13, vuint64m2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_u16mf4))) +void vluxseg6ei64(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, const uint16_t * op6, vuint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_u16mf4_m))) +void vluxseg6ei64(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vbool64_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, vuint16mf4_t op10, vuint16mf4_t op11, vuint16mf4_t op12, const uint16_t * op13, vuint64m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_i16m1))) +void vluxseg7ei64(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vint16m1_t * op6, const int16_t * op7, vuint64m4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_i16m1_m))) +void vluxseg7ei64(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vint16m1_t * op6, vbool16_t op7, vint16m1_t op8, vint16m1_t op9, vint16m1_t op10, vint16m1_t op11, vint16m1_t op12, vint16m1_t op13, vint16m1_t op14, const int16_t * op15, vuint64m4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_i16mf2))) +void vluxseg7ei64(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vint16mf2_t * op6, const int16_t * op7, vuint64m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_i16mf2_m))) +void vluxseg7ei64(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vint16mf2_t * op6, vbool32_t op7, vint16mf2_t op8, vint16mf2_t op9, vint16mf2_t op10, vint16mf2_t op11, vint16mf2_t op12, vint16mf2_t op13, vint16mf2_t op14, const int16_t * op15, vuint64m2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_i16mf4))) +void vluxseg7ei64(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vint16mf4_t * op6, const int16_t * op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_i16mf4_m))) +void vluxseg7ei64(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vint16mf4_t * op6, vbool64_t op7, vint16mf4_t op8, vint16mf4_t op9, vint16mf4_t op10, vint16mf4_t op11, vint16mf4_t op12, vint16mf4_t op13, vint16mf4_t op14, const int16_t * op15, vuint64m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_u16m1))) +void vluxseg7ei64(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vuint16m1_t * op6, const uint16_t * op7, vuint64m4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_u16m1_m))) +void vluxseg7ei64(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vuint16m1_t * op6, vbool16_t op7, vuint16m1_t op8, vuint16m1_t op9, vuint16m1_t op10, vuint16m1_t op11, vuint16m1_t op12, vuint16m1_t op13, vuint16m1_t op14, const uint16_t * op15, vuint64m4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_u16mf2))) +void vluxseg7ei64(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vuint16mf2_t * op6, const uint16_t * op7, vuint64m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_u16mf2_m))) +void vluxseg7ei64(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vuint16mf2_t * op6, vbool32_t op7, vuint16mf2_t op8, vuint16mf2_t op9, vuint16mf2_t op10, vuint16mf2_t op11, vuint16mf2_t op12, vuint16mf2_t op13, vuint16mf2_t op14, const uint16_t * op15, vuint64m2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_u16mf4))) +void vluxseg7ei64(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vuint16mf4_t * op6, const uint16_t * op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_u16mf4_m))) +void vluxseg7ei64(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vuint16mf4_t * op6, vbool64_t op7, vuint16mf4_t op8, vuint16mf4_t op9, vuint16mf4_t op10, vuint16mf4_t op11, vuint16mf4_t op12, vuint16mf4_t op13, vuint16mf4_t op14, const uint16_t * op15, vuint64m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_i16m1))) +void vluxseg8ei64(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vint16m1_t * op6, vint16m1_t * op7, const int16_t * op8, vuint64m4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_i16m1_m))) +void vluxseg8ei64(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vint16m1_t * op6, vint16m1_t * op7, vbool16_t op8, vint16m1_t op9, vint16m1_t op10, vint16m1_t op11, vint16m1_t op12, vint16m1_t op13, vint16m1_t op14, vint16m1_t op15, vint16m1_t op16, const int16_t * op17, vuint64m4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_i16mf2))) +void vluxseg8ei64(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vint16mf2_t * op6, vint16mf2_t * op7, const int16_t * op8, vuint64m2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_i16mf2_m))) +void vluxseg8ei64(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vint16mf2_t * op6, vint16mf2_t * op7, vbool32_t op8, vint16mf2_t op9, vint16mf2_t op10, vint16mf2_t op11, vint16mf2_t op12, vint16mf2_t op13, vint16mf2_t op14, vint16mf2_t op15, vint16mf2_t op16, const int16_t * op17, vuint64m2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_i16mf4))) +void vluxseg8ei64(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vint16mf4_t * op6, vint16mf4_t * op7, const int16_t * op8, vuint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_i16mf4_m))) +void vluxseg8ei64(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vint16mf4_t * op6, vint16mf4_t * op7, vbool64_t op8, vint16mf4_t op9, vint16mf4_t op10, vint16mf4_t op11, vint16mf4_t op12, vint16mf4_t op13, vint16mf4_t op14, vint16mf4_t op15, vint16mf4_t op16, const int16_t * op17, vuint64m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_u16m1))) +void vluxseg8ei64(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vuint16m1_t * op6, vuint16m1_t * op7, const uint16_t * op8, vuint64m4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_u16m1_m))) +void vluxseg8ei64(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vuint16m1_t * op6, vuint16m1_t * op7, vbool16_t op8, vuint16m1_t op9, vuint16m1_t op10, vuint16m1_t op11, vuint16m1_t op12, vuint16m1_t op13, vuint16m1_t op14, vuint16m1_t op15, vuint16m1_t op16, const uint16_t * op17, vuint64m4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_u16mf2))) +void vluxseg8ei64(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vuint16mf2_t * op6, vuint16mf2_t * op7, const uint16_t * op8, vuint64m2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_u16mf2_m))) +void vluxseg8ei64(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vuint16mf2_t * op6, vuint16mf2_t * op7, vbool32_t op8, vuint16mf2_t op9, vuint16mf2_t op10, vuint16mf2_t op11, vuint16mf2_t op12, vuint16mf2_t op13, vuint16mf2_t op14, vuint16mf2_t op15, vuint16mf2_t op16, const uint16_t * op17, vuint64m2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_u16mf4))) +void vluxseg8ei64(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vuint16mf4_t * op6, vuint16mf4_t * op7, const uint16_t * op8, vuint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_u16mf4_m))) +void vluxseg8ei64(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vuint16mf4_t * op6, vuint16mf4_t * op7, vbool64_t op8, vuint16mf4_t op9, vuint16mf4_t op10, vuint16mf4_t op11, vuint16mf4_t op12, vuint16mf4_t op13, vuint16mf4_t op14, vuint16mf4_t op15, vuint16mf4_t op16, const uint16_t * op17, vuint64m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_i32m1))) +void vluxseg2ei8(vint32m1_t * op0, vint32m1_t * op1, const int32_t * op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_i32m1_m))) +void vluxseg2ei8(vint32m1_t * op0, vint32m1_t * op1, vbool32_t op2, vint32m1_t op3, vint32m1_t op4, const int32_t * op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_i32m2))) +void vluxseg2ei8(vint32m2_t * op0, vint32m2_t * op1, const int32_t * op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_i32m2_m))) +void vluxseg2ei8(vint32m2_t * op0, vint32m2_t * op1, vbool16_t op2, vint32m2_t op3, vint32m2_t op4, const int32_t * op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_i32m4))) +void vluxseg2ei8(vint32m4_t * op0, vint32m4_t * op1, const int32_t * op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_i32m4_m))) +void vluxseg2ei8(vint32m4_t * op0, vint32m4_t * op1, vbool8_t op2, vint32m4_t op3, vint32m4_t op4, const int32_t * op5, vuint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_i32mf2))) +void vluxseg2ei8(vint32mf2_t * op0, vint32mf2_t * op1, const int32_t * op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_i32mf2_m))) +void vluxseg2ei8(vint32mf2_t * op0, vint32mf2_t * op1, vbool64_t op2, vint32mf2_t op3, vint32mf2_t op4, const int32_t * op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_u32m1))) +void vluxseg2ei8(vuint32m1_t * op0, vuint32m1_t * op1, const uint32_t * op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_u32m1_m))) +void vluxseg2ei8(vuint32m1_t * op0, vuint32m1_t * op1, vbool32_t op2, vuint32m1_t op3, vuint32m1_t op4, const uint32_t * op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_u32m2))) +void vluxseg2ei8(vuint32m2_t * op0, vuint32m2_t * op1, const uint32_t * op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_u32m2_m))) +void vluxseg2ei8(vuint32m2_t * op0, vuint32m2_t * op1, vbool16_t op2, vuint32m2_t op3, vuint32m2_t op4, const uint32_t * op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_u32m4))) +void vluxseg2ei8(vuint32m4_t * op0, vuint32m4_t * op1, const uint32_t * op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_u32m4_m))) +void vluxseg2ei8(vuint32m4_t * op0, vuint32m4_t * op1, vbool8_t op2, vuint32m4_t op3, vuint32m4_t op4, const uint32_t * op5, vuint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_u32mf2))) +void vluxseg2ei8(vuint32mf2_t * op0, vuint32mf2_t * op1, const uint32_t * op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_u32mf2_m))) +void vluxseg2ei8(vuint32mf2_t * op0, vuint32mf2_t * op1, vbool64_t op2, vuint32mf2_t op3, vuint32mf2_t op4, const uint32_t * op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_i32m1))) +void vluxseg3ei8(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, const int32_t * op3, vuint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_i32m1_m))) +void vluxseg3ei8(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vbool32_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, const int32_t * op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_i32m2))) +void vluxseg3ei8(vint32m2_t * op0, vint32m2_t * op1, vint32m2_t * op2, const int32_t * op3, vuint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_i32m2_m))) +void vluxseg3ei8(vint32m2_t * op0, vint32m2_t * op1, vint32m2_t * op2, vbool16_t op3, vint32m2_t op4, vint32m2_t op5, vint32m2_t op6, const int32_t * op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_i32mf2))) +void vluxseg3ei8(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, const int32_t * op3, vuint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_i32mf2_m))) +void vluxseg3ei8(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vbool64_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, const int32_t * op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_u32m1))) +void vluxseg3ei8(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, const uint32_t * op3, vuint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_u32m1_m))) +void vluxseg3ei8(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vbool32_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, const uint32_t * op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_u32m2))) +void vluxseg3ei8(vuint32m2_t * op0, vuint32m2_t * op1, vuint32m2_t * op2, const uint32_t * op3, vuint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_u32m2_m))) +void vluxseg3ei8(vuint32m2_t * op0, vuint32m2_t * op1, vuint32m2_t * op2, vbool16_t op3, vuint32m2_t op4, vuint32m2_t op5, vuint32m2_t op6, const uint32_t * op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_u32mf2))) +void vluxseg3ei8(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, const uint32_t * op3, vuint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_u32mf2_m))) +void vluxseg3ei8(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vbool64_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, const uint32_t * op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_i32m1))) +void vluxseg4ei8(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, const int32_t * op4, vuint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_i32m1_m))) +void vluxseg4ei8(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vbool32_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, const int32_t * op9, vuint8mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_i32m2))) +void vluxseg4ei8(vint32m2_t * op0, vint32m2_t * op1, vint32m2_t * op2, vint32m2_t * op3, const int32_t * op4, vuint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_i32m2_m))) +void vluxseg4ei8(vint32m2_t * op0, vint32m2_t * op1, vint32m2_t * op2, vint32m2_t * op3, vbool16_t op4, vint32m2_t op5, vint32m2_t op6, vint32m2_t op7, vint32m2_t op8, const int32_t * op9, vuint8mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_i32mf2))) +void vluxseg4ei8(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, const int32_t * op4, vuint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_i32mf2_m))) +void vluxseg4ei8(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vbool64_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, const int32_t * op9, vuint8mf8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_i8m1))) +void vsuxseg2ei8(int8_t * op0, vuint8m1_t op1, vint8m1_t op2, vint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_i8m1_m))) +void vsuxseg2ei8(vbool8_t op0, int8_t * op1, vuint8m1_t op2, vint8m1_t op3, vint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_i8m2))) +void vsuxseg2ei8(int8_t * op0, vuint8m2_t op1, vint8m2_t op2, vint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_i8m2_m))) +void vsuxseg2ei8(vbool4_t op0, int8_t * op1, vuint8m2_t op2, vint8m2_t op3, vint8m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_i8m4))) +void vsuxseg2ei8(int8_t * op0, vuint8m4_t op1, vint8m4_t op2, vint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_i8m4_m))) +void vsuxseg2ei8(vbool2_t op0, int8_t * op1, vuint8m4_t op2, vint8m4_t op3, vint8m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_i8mf2))) +void vsuxseg2ei8(int8_t * op0, vuint8mf2_t op1, vint8mf2_t op2, vint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_i8mf2_m))) +void vsuxseg2ei8(vbool16_t op0, int8_t * op1, vuint8mf2_t op2, vint8mf2_t op3, vint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_i8mf4))) +void vsuxseg2ei8(int8_t * op0, vuint8mf4_t op1, vint8mf4_t op2, vint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_i8mf4_m))) +void vsuxseg2ei8(vbool32_t op0, int8_t * op1, vuint8mf4_t op2, vint8mf4_t op3, vint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_i8mf8))) +void vsuxseg2ei8(int8_t * op0, vuint8mf8_t op1, vint8mf8_t op2, vint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_i8mf8_m))) +void vsuxseg2ei8(vbool64_t op0, int8_t * op1, vuint8mf8_t op2, vint8mf8_t op3, vint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_u32m1))) +void vluxseg4ei8(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, const uint32_t * op4, vuint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_u32m1_m))) +void vluxseg4ei8(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vbool32_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, const uint32_t * op9, vuint8mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_u32m2))) +void vluxseg4ei8(vuint32m2_t * op0, vuint32m2_t * op1, vuint32m2_t * op2, vuint32m2_t * op3, const uint32_t * op4, vuint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_u32m2_m))) +void vluxseg4ei8(vuint32m2_t * op0, vuint32m2_t * op1, vuint32m2_t * op2, vuint32m2_t * op3, vbool16_t op4, vuint32m2_t op5, vuint32m2_t op6, vuint32m2_t op7, vuint32m2_t op8, const uint32_t * op9, vuint8mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_u32mf2))) +void vluxseg4ei8(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, const uint32_t * op4, vuint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_u32mf2_m))) +void vluxseg4ei8(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vbool64_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, const uint32_t * op9, vuint8mf8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_i32m1))) +void vluxseg5ei8(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, const int32_t * op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_i32m1_m))) +void vluxseg5ei8(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vbool32_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, vint32m1_t op10, const int32_t * op11, vuint8mf4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_i32mf2))) +void vluxseg5ei8(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, const int32_t * op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_i32mf2_m))) +void vluxseg5ei8(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vbool64_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, vint32mf2_t op10, const int32_t * op11, vuint8mf8_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_u32m1))) +void vluxseg5ei8(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, const uint32_t * op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_u32m1_m))) +void vluxseg5ei8(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vbool32_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, vuint32m1_t op10, const uint32_t * op11, vuint8mf4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_u32mf2))) +void vluxseg5ei8(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, const uint32_t * op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_u32mf2_m))) +void vluxseg5ei8(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vbool64_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, vuint32mf2_t op10, const uint32_t * op11, vuint8mf8_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_i32m1))) +void vluxseg6ei8(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, const int32_t * op6, vuint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_i32m1_m))) +void vluxseg6ei8(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vbool32_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, vint32m1_t op10, vint32m1_t op11, vint32m1_t op12, const int32_t * op13, vuint8mf4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_i32mf2))) +void vluxseg6ei8(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, const int32_t * op6, vuint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_i32mf2_m))) +void vluxseg6ei8(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vbool64_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, vint32mf2_t op10, vint32mf2_t op11, vint32mf2_t op12, const int32_t * op13, vuint8mf8_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_u32m1))) +void vluxseg6ei8(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, const uint32_t * op6, vuint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_u32m1_m))) +void vluxseg6ei8(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vbool32_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, vuint32m1_t op10, vuint32m1_t op11, vuint32m1_t op12, const uint32_t * op13, vuint8mf4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_u32mf2))) +void vluxseg6ei8(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, const uint32_t * op6, vuint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_u32mf2_m))) +void vluxseg6ei8(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vbool64_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, vuint32mf2_t op10, vuint32mf2_t op11, vuint32mf2_t op12, const uint32_t * op13, vuint8mf8_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_i32m1))) +void vluxseg7ei8(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vint32m1_t * op6, const int32_t * op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_i32m1_m))) +void vluxseg7ei8(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vint32m1_t * op6, vbool32_t op7, vint32m1_t op8, vint32m1_t op9, vint32m1_t op10, vint32m1_t op11, vint32m1_t op12, vint32m1_t op13, vint32m1_t op14, const int32_t * op15, vuint8mf4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_i32mf2))) +void vluxseg7ei8(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vint32mf2_t * op6, const int32_t * op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_i32mf2_m))) +void vluxseg7ei8(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vint32mf2_t * op6, vbool64_t op7, vint32mf2_t op8, vint32mf2_t op9, vint32mf2_t op10, vint32mf2_t op11, vint32mf2_t op12, vint32mf2_t op13, vint32mf2_t op14, const int32_t * op15, vuint8mf8_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_u32m1))) +void vluxseg7ei8(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vuint32m1_t * op6, const uint32_t * op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_u32m1_m))) +void vluxseg7ei8(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vuint32m1_t * op6, vbool32_t op7, vuint32m1_t op8, vuint32m1_t op9, vuint32m1_t op10, vuint32m1_t op11, vuint32m1_t op12, vuint32m1_t op13, vuint32m1_t op14, const uint32_t * op15, vuint8mf4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_u32mf2))) +void vluxseg7ei8(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vuint32mf2_t * op6, const uint32_t * op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_u32mf2_m))) +void vluxseg7ei8(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vuint32mf2_t * op6, vbool64_t op7, vuint32mf2_t op8, vuint32mf2_t op9, vuint32mf2_t op10, vuint32mf2_t op11, vuint32mf2_t op12, vuint32mf2_t op13, vuint32mf2_t op14, const uint32_t * op15, vuint8mf8_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_i32m1))) +void vluxseg8ei8(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vint32m1_t * op6, vint32m1_t * op7, const int32_t * op8, vuint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_i32m1_m))) +void vluxseg8ei8(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vint32m1_t * op6, vint32m1_t * op7, vbool32_t op8, vint32m1_t op9, vint32m1_t op10, vint32m1_t op11, vint32m1_t op12, vint32m1_t op13, vint32m1_t op14, vint32m1_t op15, vint32m1_t op16, const int32_t * op17, vuint8mf4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_i32mf2))) +void vluxseg8ei8(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vint32mf2_t * op6, vint32mf2_t * op7, const int32_t * op8, vuint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_i32mf2_m))) +void vluxseg8ei8(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vint32mf2_t * op6, vint32mf2_t * op7, vbool64_t op8, vint32mf2_t op9, vint32mf2_t op10, vint32mf2_t op11, vint32mf2_t op12, vint32mf2_t op13, vint32mf2_t op14, vint32mf2_t op15, vint32mf2_t op16, const int32_t * op17, vuint8mf8_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_u32m1))) +void vluxseg8ei8(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vuint32m1_t * op6, vuint32m1_t * op7, const uint32_t * op8, vuint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_u32m1_m))) +void vluxseg8ei8(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vuint32m1_t * op6, vuint32m1_t * op7, vbool32_t op8, vuint32m1_t op9, vuint32m1_t op10, vuint32m1_t op11, vuint32m1_t op12, vuint32m1_t op13, vuint32m1_t op14, vuint32m1_t op15, vuint32m1_t op16, const uint32_t * op17, vuint8mf4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_u32mf2))) +void vluxseg8ei8(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vuint32mf2_t * op6, vuint32mf2_t * op7, const uint32_t * op8, vuint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_u32mf2_m))) +void vluxseg8ei8(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vuint32mf2_t * op6, vuint32mf2_t * op7, vbool64_t op8, vuint32mf2_t op9, vuint32mf2_t op10, vuint32mf2_t op11, vuint32mf2_t op12, vuint32mf2_t op13, vuint32mf2_t op14, vuint32mf2_t op15, vuint32mf2_t op16, const uint32_t * op17, vuint8mf8_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_i32m1))) +void vluxseg2ei16(vint32m1_t * op0, vint32m1_t * op1, const int32_t * op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_i32m1_m))) +void vluxseg2ei16(vint32m1_t * op0, vint32m1_t * op1, vbool32_t op2, vint32m1_t op3, vint32m1_t op4, const int32_t * op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_i32m2))) +void vluxseg2ei16(vint32m2_t * op0, vint32m2_t * op1, const int32_t * op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_i32m2_m))) +void vluxseg2ei16(vint32m2_t * op0, vint32m2_t * op1, vbool16_t op2, vint32m2_t op3, vint32m2_t op4, const int32_t * op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_i32m4))) +void vluxseg2ei16(vint32m4_t * op0, vint32m4_t * op1, const int32_t * op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_i32m4_m))) +void vluxseg2ei16(vint32m4_t * op0, vint32m4_t * op1, vbool8_t op2, vint32m4_t op3, vint32m4_t op4, const int32_t * op5, vuint16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_i32mf2))) +void vluxseg2ei16(vint32mf2_t * op0, vint32mf2_t * op1, const int32_t * op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_i32mf2_m))) +void vluxseg2ei16(vint32mf2_t * op0, vint32mf2_t * op1, vbool64_t op2, vint32mf2_t op3, vint32mf2_t op4, const int32_t * op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_u8m1))) +void vsuxseg2ei8(uint8_t * op0, vuint8m1_t op1, vuint8m1_t op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_u8m1_m))) +void vsuxseg2ei8(vbool8_t op0, uint8_t * op1, vuint8m1_t op2, vuint8m1_t op3, vuint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_u8m2))) +void vsuxseg2ei8(uint8_t * op0, vuint8m2_t op1, vuint8m2_t op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_u8m2_m))) +void vsuxseg2ei8(vbool4_t op0, uint8_t * op1, vuint8m2_t op2, vuint8m2_t op3, vuint8m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_u8m4))) +void vsuxseg2ei8(uint8_t * op0, vuint8m4_t op1, vuint8m4_t op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_u8m4_m))) +void vsuxseg2ei8(vbool2_t op0, uint8_t * op1, vuint8m4_t op2, vuint8m4_t op3, vuint8m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_u8mf2))) +void vsuxseg2ei8(uint8_t * op0, vuint8mf2_t op1, vuint8mf2_t op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_u8mf2_m))) +void vsuxseg2ei8(vbool16_t op0, uint8_t * op1, vuint8mf2_t op2, vuint8mf2_t op3, vuint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_u8mf4))) +void vsuxseg2ei8(uint8_t * op0, vuint8mf4_t op1, vuint8mf4_t op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_u8mf4_m))) +void vsuxseg2ei8(vbool32_t op0, uint8_t * op1, vuint8mf4_t op2, vuint8mf4_t op3, vuint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_u8mf8))) +void vsuxseg2ei8(uint8_t * op0, vuint8mf8_t op1, vuint8mf8_t op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_u8mf8_m))) +void vsuxseg2ei8(vbool64_t op0, uint8_t * op1, vuint8mf8_t op2, vuint8mf8_t op3, vuint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_u32m1))) +void vluxseg2ei16(vuint32m1_t * op0, vuint32m1_t * op1, const uint32_t * op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_u32m1_m))) +void vluxseg2ei16(vuint32m1_t * op0, vuint32m1_t * op1, vbool32_t op2, vuint32m1_t op3, vuint32m1_t op4, const uint32_t * op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_u32m2))) +void vluxseg2ei16(vuint32m2_t * op0, vuint32m2_t * op1, const uint32_t * op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_u32m2_m))) +void vluxseg2ei16(vuint32m2_t * op0, vuint32m2_t * op1, vbool16_t op2, vuint32m2_t op3, vuint32m2_t op4, const uint32_t * op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_u32m4))) +void vluxseg2ei16(vuint32m4_t * op0, vuint32m4_t * op1, const uint32_t * op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_u32m4_m))) +void vluxseg2ei16(vuint32m4_t * op0, vuint32m4_t * op1, vbool8_t op2, vuint32m4_t op3, vuint32m4_t op4, const uint32_t * op5, vuint16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_u32mf2))) +void vluxseg2ei16(vuint32mf2_t * op0, vuint32mf2_t * op1, const uint32_t * op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_u32mf2_m))) +void vluxseg2ei16(vuint32mf2_t * op0, vuint32mf2_t * op1, vbool64_t op2, vuint32mf2_t op3, vuint32mf2_t op4, const uint32_t * op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_i32m1))) +void vluxseg3ei16(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, const int32_t * op3, vuint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_i32m1_m))) +void vluxseg3ei16(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vbool32_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, const int32_t * op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_i32m2))) +void vluxseg3ei16(vint32m2_t * op0, vint32m2_t * op1, vint32m2_t * op2, const int32_t * op3, vuint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_i32m2_m))) +void vluxseg3ei16(vint32m2_t * op0, vint32m2_t * op1, vint32m2_t * op2, vbool16_t op3, vint32m2_t op4, vint32m2_t op5, vint32m2_t op6, const int32_t * op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_i32mf2))) +void vluxseg3ei16(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, const int32_t * op3, vuint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_i32mf2_m))) +void vluxseg3ei16(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vbool64_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, const int32_t * op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_u32m1))) +void vluxseg3ei16(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, const uint32_t * op3, vuint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_u32m1_m))) +void vluxseg3ei16(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vbool32_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, const uint32_t * op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_u32m2))) +void vluxseg3ei16(vuint32m2_t * op0, vuint32m2_t * op1, vuint32m2_t * op2, const uint32_t * op3, vuint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_u32m2_m))) +void vluxseg3ei16(vuint32m2_t * op0, vuint32m2_t * op1, vuint32m2_t * op2, vbool16_t op3, vuint32m2_t op4, vuint32m2_t op5, vuint32m2_t op6, const uint32_t * op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_u32mf2))) +void vluxseg3ei16(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, const uint32_t * op3, vuint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_u32mf2_m))) +void vluxseg3ei16(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vbool64_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, const uint32_t * op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_i32m1))) +void vluxseg4ei16(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, const int32_t * op4, vuint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_i32m1_m))) +void vluxseg4ei16(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vbool32_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, const int32_t * op9, vuint16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_i32m2))) +void vluxseg4ei16(vint32m2_t * op0, vint32m2_t * op1, vint32m2_t * op2, vint32m2_t * op3, const int32_t * op4, vuint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_i32m2_m))) +void vluxseg4ei16(vint32m2_t * op0, vint32m2_t * op1, vint32m2_t * op2, vint32m2_t * op3, vbool16_t op4, vint32m2_t op5, vint32m2_t op6, vint32m2_t op7, vint32m2_t op8, const int32_t * op9, vuint16m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_i32mf2))) +void vluxseg4ei16(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, const int32_t * op4, vuint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_i32mf2_m))) +void vluxseg4ei16(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vbool64_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, const int32_t * op9, vuint16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_u32m1))) +void vluxseg4ei16(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, const uint32_t * op4, vuint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_u32m1_m))) +void vluxseg4ei16(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vbool32_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, const uint32_t * op9, vuint16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_u32m2))) +void vluxseg4ei16(vuint32m2_t * op0, vuint32m2_t * op1, vuint32m2_t * op2, vuint32m2_t * op3, const uint32_t * op4, vuint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_u32m2_m))) +void vluxseg4ei16(vuint32m2_t * op0, vuint32m2_t * op1, vuint32m2_t * op2, vuint32m2_t * op3, vbool16_t op4, vuint32m2_t op5, vuint32m2_t op6, vuint32m2_t op7, vuint32m2_t op8, const uint32_t * op9, vuint16m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_u32mf2))) +void vluxseg4ei16(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, const uint32_t * op4, vuint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_u32mf2_m))) +void vluxseg4ei16(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vbool64_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, const uint32_t * op9, vuint16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_i32m1))) +void vluxseg5ei16(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, const int32_t * op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_i32m1_m))) +void vluxseg5ei16(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vbool32_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, vint32m1_t op10, const int32_t * op11, vuint16mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_i32mf2))) +void vluxseg5ei16(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, const int32_t * op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_i32mf2_m))) +void vluxseg5ei16(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vbool64_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, vint32mf2_t op10, const int32_t * op11, vuint16mf4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_u32m1))) +void vluxseg5ei16(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, const uint32_t * op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_u32m1_m))) +void vluxseg5ei16(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vbool32_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, vuint32m1_t op10, const uint32_t * op11, vuint16mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_u32mf2))) +void vluxseg5ei16(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, const uint32_t * op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_u32mf2_m))) +void vluxseg5ei16(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vbool64_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, vuint32mf2_t op10, const uint32_t * op11, vuint16mf4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_i32m1))) +void vluxseg6ei16(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, const int32_t * op6, vuint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_i32m1_m))) +void vluxseg6ei16(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vbool32_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, vint32m1_t op10, vint32m1_t op11, vint32m1_t op12, const int32_t * op13, vuint16mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_i32mf2))) +void vluxseg6ei16(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, const int32_t * op6, vuint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_i32mf2_m))) +void vluxseg6ei16(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vbool64_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, vint32mf2_t op10, vint32mf2_t op11, vint32mf2_t op12, const int32_t * op13, vuint16mf4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_u32m1))) +void vluxseg6ei16(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, const uint32_t * op6, vuint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_u32m1_m))) +void vluxseg6ei16(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vbool32_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, vuint32m1_t op10, vuint32m1_t op11, vuint32m1_t op12, const uint32_t * op13, vuint16mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_u32mf2))) +void vluxseg6ei16(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, const uint32_t * op6, vuint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_u32mf2_m))) +void vluxseg6ei16(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vbool64_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, vuint32mf2_t op10, vuint32mf2_t op11, vuint32mf2_t op12, const uint32_t * op13, vuint16mf4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_i32m1))) +void vluxseg7ei16(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vint32m1_t * op6, const int32_t * op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_i32m1_m))) +void vluxseg7ei16(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vint32m1_t * op6, vbool32_t op7, vint32m1_t op8, vint32m1_t op9, vint32m1_t op10, vint32m1_t op11, vint32m1_t op12, vint32m1_t op13, vint32m1_t op14, const int32_t * op15, vuint16mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_i32mf2))) +void vluxseg7ei16(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vint32mf2_t * op6, const int32_t * op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_i32mf2_m))) +void vluxseg7ei16(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vint32mf2_t * op6, vbool64_t op7, vint32mf2_t op8, vint32mf2_t op9, vint32mf2_t op10, vint32mf2_t op11, vint32mf2_t op12, vint32mf2_t op13, vint32mf2_t op14, const int32_t * op15, vuint16mf4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_u32m1))) +void vluxseg7ei16(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vuint32m1_t * op6, const uint32_t * op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_u32m1_m))) +void vluxseg7ei16(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vuint32m1_t * op6, vbool32_t op7, vuint32m1_t op8, vuint32m1_t op9, vuint32m1_t op10, vuint32m1_t op11, vuint32m1_t op12, vuint32m1_t op13, vuint32m1_t op14, const uint32_t * op15, vuint16mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_u32mf2))) +void vluxseg7ei16(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vuint32mf2_t * op6, const uint32_t * op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_u32mf2_m))) +void vluxseg7ei16(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vuint32mf2_t * op6, vbool64_t op7, vuint32mf2_t op8, vuint32mf2_t op9, vuint32mf2_t op10, vuint32mf2_t op11, vuint32mf2_t op12, vuint32mf2_t op13, vuint32mf2_t op14, const uint32_t * op15, vuint16mf4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_i32m1))) +void vluxseg8ei16(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vint32m1_t * op6, vint32m1_t * op7, const int32_t * op8, vuint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_i32m1_m))) +void vluxseg8ei16(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vint32m1_t * op6, vint32m1_t * op7, vbool32_t op8, vint32m1_t op9, vint32m1_t op10, vint32m1_t op11, vint32m1_t op12, vint32m1_t op13, vint32m1_t op14, vint32m1_t op15, vint32m1_t op16, const int32_t * op17, vuint16mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_i32mf2))) +void vluxseg8ei16(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vint32mf2_t * op6, vint32mf2_t * op7, const int32_t * op8, vuint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_i32mf2_m))) +void vluxseg8ei16(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vint32mf2_t * op6, vint32mf2_t * op7, vbool64_t op8, vint32mf2_t op9, vint32mf2_t op10, vint32mf2_t op11, vint32mf2_t op12, vint32mf2_t op13, vint32mf2_t op14, vint32mf2_t op15, vint32mf2_t op16, const int32_t * op17, vuint16mf4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_u32m1))) +void vluxseg8ei16(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vuint32m1_t * op6, vuint32m1_t * op7, const uint32_t * op8, vuint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_u32m1_m))) +void vluxseg8ei16(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vuint32m1_t * op6, vuint32m1_t * op7, vbool32_t op8, vuint32m1_t op9, vuint32m1_t op10, vuint32m1_t op11, vuint32m1_t op12, vuint32m1_t op13, vuint32m1_t op14, vuint32m1_t op15, vuint32m1_t op16, const uint32_t * op17, vuint16mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_u32mf2))) +void vluxseg8ei16(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vuint32mf2_t * op6, vuint32mf2_t * op7, const uint32_t * op8, vuint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_u32mf2_m))) +void vluxseg8ei16(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vuint32mf2_t * op6, vuint32mf2_t * op7, vbool64_t op8, vuint32mf2_t op9, vuint32mf2_t op10, vuint32mf2_t op11, vuint32mf2_t op12, vuint32mf2_t op13, vuint32mf2_t op14, vuint32mf2_t op15, vuint32mf2_t op16, const uint32_t * op17, vuint16mf4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_i32m1))) +void vluxseg2ei32(vint32m1_t * op0, vint32m1_t * op1, const int32_t * op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_i32m1_m))) +void vluxseg2ei32(vint32m1_t * op0, vint32m1_t * op1, vbool32_t op2, vint32m1_t op3, vint32m1_t op4, const int32_t * op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_i32m2))) +void vluxseg2ei32(vint32m2_t * op0, vint32m2_t * op1, const int32_t * op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_i32m2_m))) +void vluxseg2ei32(vint32m2_t * op0, vint32m2_t * op1, vbool16_t op2, vint32m2_t op3, vint32m2_t op4, const int32_t * op5, vuint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_i32m4))) +void vluxseg2ei32(vint32m4_t * op0, vint32m4_t * op1, const int32_t * op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_i32m4_m))) +void vluxseg2ei32(vint32m4_t * op0, vint32m4_t * op1, vbool8_t op2, vint32m4_t op3, vint32m4_t op4, const int32_t * op5, vuint32m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_i32mf2))) +void vluxseg2ei32(vint32mf2_t * op0, vint32mf2_t * op1, const int32_t * op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_i32mf2_m))) +void vluxseg2ei32(vint32mf2_t * op0, vint32mf2_t * op1, vbool64_t op2, vint32mf2_t op3, vint32mf2_t op4, const int32_t * op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_u32m1))) +void vluxseg2ei32(vuint32m1_t * op0, vuint32m1_t * op1, const uint32_t * op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_u32m1_m))) +void vluxseg2ei32(vuint32m1_t * op0, vuint32m1_t * op1, vbool32_t op2, vuint32m1_t op3, vuint32m1_t op4, const uint32_t * op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_u32m2))) +void vluxseg2ei32(vuint32m2_t * op0, vuint32m2_t * op1, const uint32_t * op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_u32m2_m))) +void vluxseg2ei32(vuint32m2_t * op0, vuint32m2_t * op1, vbool16_t op2, vuint32m2_t op3, vuint32m2_t op4, const uint32_t * op5, vuint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_u32m4))) +void vluxseg2ei32(vuint32m4_t * op0, vuint32m4_t * op1, const uint32_t * op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_u32m4_m))) +void vluxseg2ei32(vuint32m4_t * op0, vuint32m4_t * op1, vbool8_t op2, vuint32m4_t op3, vuint32m4_t op4, const uint32_t * op5, vuint32m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_u32mf2))) +void vluxseg2ei32(vuint32mf2_t * op0, vuint32mf2_t * op1, const uint32_t * op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_u32mf2_m))) +void vluxseg2ei32(vuint32mf2_t * op0, vuint32mf2_t * op1, vbool64_t op2, vuint32mf2_t op3, vuint32mf2_t op4, const uint32_t * op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_i32m1))) +void vluxseg3ei32(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, const int32_t * op3, vuint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_i32m1_m))) +void vluxseg3ei32(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vbool32_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, const int32_t * op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_i32m2))) +void vluxseg3ei32(vint32m2_t * op0, vint32m2_t * op1, vint32m2_t * op2, const int32_t * op3, vuint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_i32m2_m))) +void vluxseg3ei32(vint32m2_t * op0, vint32m2_t * op1, vint32m2_t * op2, vbool16_t op3, vint32m2_t op4, vint32m2_t op5, vint32m2_t op6, const int32_t * op7, vuint32m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_i32mf2))) +void vluxseg3ei32(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, const int32_t * op3, vuint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_i32mf2_m))) +void vluxseg3ei32(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vbool64_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, const int32_t * op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_u32m1))) +void vluxseg3ei32(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, const uint32_t * op3, vuint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_u32m1_m))) +void vluxseg3ei32(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vbool32_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, const uint32_t * op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_u32m2))) +void vluxseg3ei32(vuint32m2_t * op0, vuint32m2_t * op1, vuint32m2_t * op2, const uint32_t * op3, vuint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_u32m2_m))) +void vluxseg3ei32(vuint32m2_t * op0, vuint32m2_t * op1, vuint32m2_t * op2, vbool16_t op3, vuint32m2_t op4, vuint32m2_t op5, vuint32m2_t op6, const uint32_t * op7, vuint32m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_u32mf2))) +void vluxseg3ei32(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, const uint32_t * op3, vuint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_u32mf2_m))) +void vluxseg3ei32(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vbool64_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, const uint32_t * op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_i32m1))) +void vluxseg4ei32(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, const int32_t * op4, vuint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_i32m1_m))) +void vluxseg4ei32(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vbool32_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, const int32_t * op9, vuint32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_i32m2))) +void vluxseg4ei32(vint32m2_t * op0, vint32m2_t * op1, vint32m2_t * op2, vint32m2_t * op3, const int32_t * op4, vuint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_i32m2_m))) +void vluxseg4ei32(vint32m2_t * op0, vint32m2_t * op1, vint32m2_t * op2, vint32m2_t * op3, vbool16_t op4, vint32m2_t op5, vint32m2_t op6, vint32m2_t op7, vint32m2_t op8, const int32_t * op9, vuint32m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_i32mf2))) +void vluxseg4ei32(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, const int32_t * op4, vuint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_i32mf2_m))) +void vluxseg4ei32(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vbool64_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, const int32_t * op9, vuint32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_u32m1))) +void vluxseg4ei32(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, const uint32_t * op4, vuint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_u32m1_m))) +void vluxseg4ei32(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vbool32_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, const uint32_t * op9, vuint32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_u32m2))) +void vluxseg4ei32(vuint32m2_t * op0, vuint32m2_t * op1, vuint32m2_t * op2, vuint32m2_t * op3, const uint32_t * op4, vuint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_u32m2_m))) +void vluxseg4ei32(vuint32m2_t * op0, vuint32m2_t * op1, vuint32m2_t * op2, vuint32m2_t * op3, vbool16_t op4, vuint32m2_t op5, vuint32m2_t op6, vuint32m2_t op7, vuint32m2_t op8, const uint32_t * op9, vuint32m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_u32mf2))) +void vluxseg4ei32(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, const uint32_t * op4, vuint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_u32mf2_m))) +void vluxseg4ei32(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vbool64_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, const uint32_t * op9, vuint32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_i32m1))) +void vluxseg5ei32(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, const int32_t * op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_i32m1_m))) +void vluxseg5ei32(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vbool32_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, vint32m1_t op10, const int32_t * op11, vuint32m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_i32mf2))) +void vluxseg5ei32(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, const int32_t * op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_i32mf2_m))) +void vluxseg5ei32(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vbool64_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, vint32mf2_t op10, const int32_t * op11, vuint32mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_u32m1))) +void vluxseg5ei32(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, const uint32_t * op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_u32m1_m))) +void vluxseg5ei32(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vbool32_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, vuint32m1_t op10, const uint32_t * op11, vuint32m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_u32mf2))) +void vluxseg5ei32(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, const uint32_t * op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_u32mf2_m))) +void vluxseg5ei32(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vbool64_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, vuint32mf2_t op10, const uint32_t * op11, vuint32mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_i32m1))) +void vluxseg6ei32(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, const int32_t * op6, vuint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_i32m1_m))) +void vluxseg6ei32(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vbool32_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, vint32m1_t op10, vint32m1_t op11, vint32m1_t op12, const int32_t * op13, vuint32m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_i32mf2))) +void vluxseg6ei32(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, const int32_t * op6, vuint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_i32mf2_m))) +void vluxseg6ei32(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vbool64_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, vint32mf2_t op10, vint32mf2_t op11, vint32mf2_t op12, const int32_t * op13, vuint32mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_u32m1))) +void vluxseg6ei32(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, const uint32_t * op6, vuint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_u32m1_m))) +void vluxseg6ei32(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vbool32_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, vuint32m1_t op10, vuint32m1_t op11, vuint32m1_t op12, const uint32_t * op13, vuint32m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_u32mf2))) +void vluxseg6ei32(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, const uint32_t * op6, vuint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_u32mf2_m))) +void vluxseg6ei32(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vbool64_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, vuint32mf2_t op10, vuint32mf2_t op11, vuint32mf2_t op12, const uint32_t * op13, vuint32mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_i32m1))) +void vluxseg7ei32(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vint32m1_t * op6, const int32_t * op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_i32m1_m))) +void vluxseg7ei32(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vint32m1_t * op6, vbool32_t op7, vint32m1_t op8, vint32m1_t op9, vint32m1_t op10, vint32m1_t op11, vint32m1_t op12, vint32m1_t op13, vint32m1_t op14, const int32_t * op15, vuint32m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_i32mf2))) +void vluxseg7ei32(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vint32mf2_t * op6, const int32_t * op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_i32mf2_m))) +void vluxseg7ei32(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vint32mf2_t * op6, vbool64_t op7, vint32mf2_t op8, vint32mf2_t op9, vint32mf2_t op10, vint32mf2_t op11, vint32mf2_t op12, vint32mf2_t op13, vint32mf2_t op14, const int32_t * op15, vuint32mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_u32m1))) +void vluxseg7ei32(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vuint32m1_t * op6, const uint32_t * op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_u32m1_m))) +void vluxseg7ei32(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vuint32m1_t * op6, vbool32_t op7, vuint32m1_t op8, vuint32m1_t op9, vuint32m1_t op10, vuint32m1_t op11, vuint32m1_t op12, vuint32m1_t op13, vuint32m1_t op14, const uint32_t * op15, vuint32m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_u32mf2))) +void vluxseg7ei32(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vuint32mf2_t * op6, const uint32_t * op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_u32mf2_m))) +void vluxseg7ei32(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vuint32mf2_t * op6, vbool64_t op7, vuint32mf2_t op8, vuint32mf2_t op9, vuint32mf2_t op10, vuint32mf2_t op11, vuint32mf2_t op12, vuint32mf2_t op13, vuint32mf2_t op14, const uint32_t * op15, vuint32mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_i32m1))) +void vluxseg8ei32(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vint32m1_t * op6, vint32m1_t * op7, const int32_t * op8, vuint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_i32m1_m))) +void vluxseg8ei32(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vint32m1_t * op6, vint32m1_t * op7, vbool32_t op8, vint32m1_t op9, vint32m1_t op10, vint32m1_t op11, vint32m1_t op12, vint32m1_t op13, vint32m1_t op14, vint32m1_t op15, vint32m1_t op16, const int32_t * op17, vuint32m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_i32mf2))) +void vluxseg8ei32(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vint32mf2_t * op6, vint32mf2_t * op7, const int32_t * op8, vuint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_i32mf2_m))) +void vluxseg8ei32(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vint32mf2_t * op6, vint32mf2_t * op7, vbool64_t op8, vint32mf2_t op9, vint32mf2_t op10, vint32mf2_t op11, vint32mf2_t op12, vint32mf2_t op13, vint32mf2_t op14, vint32mf2_t op15, vint32mf2_t op16, const int32_t * op17, vuint32mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_u32m1))) +void vluxseg8ei32(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vuint32m1_t * op6, vuint32m1_t * op7, const uint32_t * op8, vuint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_u32m1_m))) +void vluxseg8ei32(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vuint32m1_t * op6, vuint32m1_t * op7, vbool32_t op8, vuint32m1_t op9, vuint32m1_t op10, vuint32m1_t op11, vuint32m1_t op12, vuint32m1_t op13, vuint32m1_t op14, vuint32m1_t op15, vuint32m1_t op16, const uint32_t * op17, vuint32m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_u32mf2))) +void vluxseg8ei32(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vuint32mf2_t * op6, vuint32mf2_t * op7, const uint32_t * op8, vuint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_u32mf2_m))) +void vluxseg8ei32(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vuint32mf2_t * op6, vuint32mf2_t * op7, vbool64_t op8, vuint32mf2_t op9, vuint32mf2_t op10, vuint32mf2_t op11, vuint32mf2_t op12, vuint32mf2_t op13, vuint32mf2_t op14, vuint32mf2_t op15, vuint32mf2_t op16, const uint32_t * op17, vuint32mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_i32m1))) +void vluxseg2ei64(vint32m1_t * op0, vint32m1_t * op1, const int32_t * op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_i32m1_m))) +void vluxseg2ei64(vint32m1_t * op0, vint32m1_t * op1, vbool32_t op2, vint32m1_t op3, vint32m1_t op4, const int32_t * op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_i32m2))) +void vluxseg2ei64(vint32m2_t * op0, vint32m2_t * op1, const int32_t * op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_i32m2_m))) +void vluxseg2ei64(vint32m2_t * op0, vint32m2_t * op1, vbool16_t op2, vint32m2_t op3, vint32m2_t op4, const int32_t * op5, vuint64m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_i32m4))) +void vluxseg2ei64(vint32m4_t * op0, vint32m4_t * op1, const int32_t * op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_i32m4_m))) +void vluxseg2ei64(vint32m4_t * op0, vint32m4_t * op1, vbool8_t op2, vint32m4_t op3, vint32m4_t op4, const int32_t * op5, vuint64m8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_i32mf2))) +void vluxseg2ei64(vint32mf2_t * op0, vint32mf2_t * op1, const int32_t * op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_i32mf2_m))) +void vluxseg2ei64(vint32mf2_t * op0, vint32mf2_t * op1, vbool64_t op2, vint32mf2_t op3, vint32mf2_t op4, const int32_t * op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_u32m1))) +void vluxseg2ei64(vuint32m1_t * op0, vuint32m1_t * op1, const uint32_t * op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_u32m1_m))) +void vluxseg2ei64(vuint32m1_t * op0, vuint32m1_t * op1, vbool32_t op2, vuint32m1_t op3, vuint32m1_t op4, const uint32_t * op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_u32m2))) +void vluxseg2ei64(vuint32m2_t * op0, vuint32m2_t * op1, const uint32_t * op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_u32m2_m))) +void vluxseg2ei64(vuint32m2_t * op0, vuint32m2_t * op1, vbool16_t op2, vuint32m2_t op3, vuint32m2_t op4, const uint32_t * op5, vuint64m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_u32m4))) +void vluxseg2ei64(vuint32m4_t * op0, vuint32m4_t * op1, const uint32_t * op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_u32m4_m))) +void vluxseg2ei64(vuint32m4_t * op0, vuint32m4_t * op1, vbool8_t op2, vuint32m4_t op3, vuint32m4_t op4, const uint32_t * op5, vuint64m8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_u32mf2))) +void vluxseg2ei64(vuint32mf2_t * op0, vuint32mf2_t * op1, const uint32_t * op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_u32mf2_m))) +void vluxseg2ei64(vuint32mf2_t * op0, vuint32mf2_t * op1, vbool64_t op2, vuint32mf2_t op3, vuint32mf2_t op4, const uint32_t * op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_i32m1))) +void vluxseg3ei64(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, const int32_t * op3, vuint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_i32m1_m))) +void vluxseg3ei64(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vbool32_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, const int32_t * op7, vuint64m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_i32m2))) +void vluxseg3ei64(vint32m2_t * op0, vint32m2_t * op1, vint32m2_t * op2, const int32_t * op3, vuint64m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_i32m2_m))) +void vluxseg3ei64(vint32m2_t * op0, vint32m2_t * op1, vint32m2_t * op2, vbool16_t op3, vint32m2_t op4, vint32m2_t op5, vint32m2_t op6, const int32_t * op7, vuint64m4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_i32mf2))) +void vluxseg3ei64(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, const int32_t * op3, vuint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_i32mf2_m))) +void vluxseg3ei64(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vbool64_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, const int32_t * op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_u32m1))) +void vluxseg3ei64(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, const uint32_t * op3, vuint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_u32m1_m))) +void vluxseg3ei64(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vbool32_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, const uint32_t * op7, vuint64m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_u32m2))) +void vluxseg3ei64(vuint32m2_t * op0, vuint32m2_t * op1, vuint32m2_t * op2, const uint32_t * op3, vuint64m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_u32m2_m))) +void vluxseg3ei64(vuint32m2_t * op0, vuint32m2_t * op1, vuint32m2_t * op2, vbool16_t op3, vuint32m2_t op4, vuint32m2_t op5, vuint32m2_t op6, const uint32_t * op7, vuint64m4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_u32mf2))) +void vluxseg3ei64(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, const uint32_t * op3, vuint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_u32mf2_m))) +void vluxseg3ei64(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vbool64_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, const uint32_t * op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_i32m1))) +void vluxseg4ei64(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, const int32_t * op4, vuint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_i32m1_m))) +void vluxseg4ei64(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vbool32_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, const int32_t * op9, vuint64m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_i32m2))) +void vluxseg4ei64(vint32m2_t * op0, vint32m2_t * op1, vint32m2_t * op2, vint32m2_t * op3, const int32_t * op4, vuint64m4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_i32m2_m))) +void vluxseg4ei64(vint32m2_t * op0, vint32m2_t * op1, vint32m2_t * op2, vint32m2_t * op3, vbool16_t op4, vint32m2_t op5, vint32m2_t op6, vint32m2_t op7, vint32m2_t op8, const int32_t * op9, vuint64m4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_i32mf2))) +void vluxseg4ei64(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, const int32_t * op4, vuint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_i32mf2_m))) +void vluxseg4ei64(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vbool64_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, const int32_t * op9, vuint64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_u32m1))) +void vluxseg4ei64(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, const uint32_t * op4, vuint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_u32m1_m))) +void vluxseg4ei64(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vbool32_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, const uint32_t * op9, vuint64m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_u32m2))) +void vluxseg4ei64(vuint32m2_t * op0, vuint32m2_t * op1, vuint32m2_t * op2, vuint32m2_t * op3, const uint32_t * op4, vuint64m4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_u32m2_m))) +void vluxseg4ei64(vuint32m2_t * op0, vuint32m2_t * op1, vuint32m2_t * op2, vuint32m2_t * op3, vbool16_t op4, vuint32m2_t op5, vuint32m2_t op6, vuint32m2_t op7, vuint32m2_t op8, const uint32_t * op9, vuint64m4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_u32mf2))) +void vluxseg4ei64(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, const uint32_t * op4, vuint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_u32mf2_m))) +void vluxseg4ei64(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vbool64_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, const uint32_t * op9, vuint64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_i32m1))) +void vluxseg5ei64(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, const int32_t * op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_i32m1_m))) +void vluxseg5ei64(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vbool32_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, vint32m1_t op10, const int32_t * op11, vuint64m2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_i32mf2))) +void vluxseg5ei64(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, const int32_t * op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_i32mf2_m))) +void vluxseg5ei64(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vbool64_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, vint32mf2_t op10, const int32_t * op11, vuint64m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_u32m1))) +void vluxseg5ei64(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, const uint32_t * op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_u32m1_m))) +void vluxseg5ei64(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vbool32_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, vuint32m1_t op10, const uint32_t * op11, vuint64m2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_u32mf2))) +void vluxseg5ei64(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, const uint32_t * op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_u32mf2_m))) +void vluxseg5ei64(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vbool64_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, vuint32mf2_t op10, const uint32_t * op11, vuint64m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_i32m1))) +void vluxseg6ei64(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, const int32_t * op6, vuint64m2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_i32m1_m))) +void vluxseg6ei64(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vbool32_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, vint32m1_t op10, vint32m1_t op11, vint32m1_t op12, const int32_t * op13, vuint64m2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_i32mf2))) +void vluxseg6ei64(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, const int32_t * op6, vuint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_i32mf2_m))) +void vluxseg6ei64(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vbool64_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, vint32mf2_t op10, vint32mf2_t op11, vint32mf2_t op12, const int32_t * op13, vuint64m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_u32m1))) +void vluxseg6ei64(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, const uint32_t * op6, vuint64m2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_u32m1_m))) +void vluxseg6ei64(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vbool32_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, vuint32m1_t op10, vuint32m1_t op11, vuint32m1_t op12, const uint32_t * op13, vuint64m2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_u32mf2))) +void vluxseg6ei64(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, const uint32_t * op6, vuint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_u32mf2_m))) +void vluxseg6ei64(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vbool64_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, vuint32mf2_t op10, vuint32mf2_t op11, vuint32mf2_t op12, const uint32_t * op13, vuint64m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_i32m1))) +void vluxseg7ei64(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vint32m1_t * op6, const int32_t * op7, vuint64m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_i32m1_m))) +void vluxseg7ei64(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vint32m1_t * op6, vbool32_t op7, vint32m1_t op8, vint32m1_t op9, vint32m1_t op10, vint32m1_t op11, vint32m1_t op12, vint32m1_t op13, vint32m1_t op14, const int32_t * op15, vuint64m2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_i32mf2))) +void vluxseg7ei64(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vint32mf2_t * op6, const int32_t * op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_i32mf2_m))) +void vluxseg7ei64(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vint32mf2_t * op6, vbool64_t op7, vint32mf2_t op8, vint32mf2_t op9, vint32mf2_t op10, vint32mf2_t op11, vint32mf2_t op12, vint32mf2_t op13, vint32mf2_t op14, const int32_t * op15, vuint64m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_u32m1))) +void vluxseg7ei64(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vuint32m1_t * op6, const uint32_t * op7, vuint64m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_u32m1_m))) +void vluxseg7ei64(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vuint32m1_t * op6, vbool32_t op7, vuint32m1_t op8, vuint32m1_t op9, vuint32m1_t op10, vuint32m1_t op11, vuint32m1_t op12, vuint32m1_t op13, vuint32m1_t op14, const uint32_t * op15, vuint64m2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_u32mf2))) +void vluxseg7ei64(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vuint32mf2_t * op6, const uint32_t * op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_u32mf2_m))) +void vluxseg7ei64(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vuint32mf2_t * op6, vbool64_t op7, vuint32mf2_t op8, vuint32mf2_t op9, vuint32mf2_t op10, vuint32mf2_t op11, vuint32mf2_t op12, vuint32mf2_t op13, vuint32mf2_t op14, const uint32_t * op15, vuint64m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_i32m1))) +void vluxseg8ei64(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vint32m1_t * op6, vint32m1_t * op7, const int32_t * op8, vuint64m2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_i32m1_m))) +void vluxseg8ei64(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vint32m1_t * op6, vint32m1_t * op7, vbool32_t op8, vint32m1_t op9, vint32m1_t op10, vint32m1_t op11, vint32m1_t op12, vint32m1_t op13, vint32m1_t op14, vint32m1_t op15, vint32m1_t op16, const int32_t * op17, vuint64m2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_i32mf2))) +void vluxseg8ei64(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vint32mf2_t * op6, vint32mf2_t * op7, const int32_t * op8, vuint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_i32mf2_m))) +void vluxseg8ei64(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vint32mf2_t * op6, vint32mf2_t * op7, vbool64_t op8, vint32mf2_t op9, vint32mf2_t op10, vint32mf2_t op11, vint32mf2_t op12, vint32mf2_t op13, vint32mf2_t op14, vint32mf2_t op15, vint32mf2_t op16, const int32_t * op17, vuint64m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_u32m1))) +void vluxseg8ei64(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vuint32m1_t * op6, vuint32m1_t * op7, const uint32_t * op8, vuint64m2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_u32m1_m))) +void vluxseg8ei64(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vuint32m1_t * op6, vuint32m1_t * op7, vbool32_t op8, vuint32m1_t op9, vuint32m1_t op10, vuint32m1_t op11, vuint32m1_t op12, vuint32m1_t op13, vuint32m1_t op14, vuint32m1_t op15, vuint32m1_t op16, const uint32_t * op17, vuint64m2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_u32mf2))) +void vluxseg8ei64(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vuint32mf2_t * op6, vuint32mf2_t * op7, const uint32_t * op8, vuint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_u32mf2_m))) +void vluxseg8ei64(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vuint32mf2_t * op6, vuint32mf2_t * op7, vbool64_t op8, vuint32mf2_t op9, vuint32mf2_t op10, vuint32mf2_t op11, vuint32mf2_t op12, vuint32mf2_t op13, vuint32mf2_t op14, vuint32mf2_t op15, vuint32mf2_t op16, const uint32_t * op17, vuint64m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_i64m1))) +void vluxseg2ei8(vint64m1_t * op0, vint64m1_t * op1, const int64_t * op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_i64m1_m))) +void vluxseg2ei8(vint64m1_t * op0, vint64m1_t * op1, vbool64_t op2, vint64m1_t op3, vint64m1_t op4, const int64_t * op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_i64m2))) +void vluxseg2ei8(vint64m2_t * op0, vint64m2_t * op1, const int64_t * op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_i64m2_m))) +void vluxseg2ei8(vint64m2_t * op0, vint64m2_t * op1, vbool32_t op2, vint64m2_t op3, vint64m2_t op4, const int64_t * op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_i64m4))) +void vluxseg2ei8(vint64m4_t * op0, vint64m4_t * op1, const int64_t * op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_i64m4_m))) +void vluxseg2ei8(vint64m4_t * op0, vint64m4_t * op1, vbool16_t op2, vint64m4_t op3, vint64m4_t op4, const int64_t * op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_u64m1))) +void vluxseg2ei8(vuint64m1_t * op0, vuint64m1_t * op1, const uint64_t * op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_u64m1_m))) +void vluxseg2ei8(vuint64m1_t * op0, vuint64m1_t * op1, vbool64_t op2, vuint64m1_t op3, vuint64m1_t op4, const uint64_t * op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_u64m2))) +void vluxseg2ei8(vuint64m2_t * op0, vuint64m2_t * op1, const uint64_t * op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_u64m2_m))) +void vluxseg2ei8(vuint64m2_t * op0, vuint64m2_t * op1, vbool32_t op2, vuint64m2_t op3, vuint64m2_t op4, const uint64_t * op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_u64m4))) +void vluxseg2ei8(vuint64m4_t * op0, vuint64m4_t * op1, const uint64_t * op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_u64m4_m))) +void vluxseg2ei8(vuint64m4_t * op0, vuint64m4_t * op1, vbool16_t op2, vuint64m4_t op3, vuint64m4_t op4, const uint64_t * op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_i64m1))) +void vluxseg3ei8(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, const int64_t * op3, vuint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_i64m1_m))) +void vluxseg3ei8(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vbool64_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, const int64_t * op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_i64m2))) +void vluxseg3ei8(vint64m2_t * op0, vint64m2_t * op1, vint64m2_t * op2, const int64_t * op3, vuint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_i64m2_m))) +void vluxseg3ei8(vint64m2_t * op0, vint64m2_t * op1, vint64m2_t * op2, vbool32_t op3, vint64m2_t op4, vint64m2_t op5, vint64m2_t op6, const int64_t * op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_u64m1))) +void vluxseg3ei8(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, const uint64_t * op3, vuint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_u64m1_m))) +void vluxseg3ei8(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vbool64_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, const uint64_t * op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_u64m2))) +void vluxseg3ei8(vuint64m2_t * op0, vuint64m2_t * op1, vuint64m2_t * op2, const uint64_t * op3, vuint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_u64m2_m))) +void vluxseg3ei8(vuint64m2_t * op0, vuint64m2_t * op1, vuint64m2_t * op2, vbool32_t op3, vuint64m2_t op4, vuint64m2_t op5, vuint64m2_t op6, const uint64_t * op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_i64m1))) +void vluxseg4ei8(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, const int64_t * op4, vuint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_i64m1_m))) +void vluxseg4ei8(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vbool64_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, const int64_t * op9, vuint8mf8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_i64m2))) +void vluxseg4ei8(vint64m2_t * op0, vint64m2_t * op1, vint64m2_t * op2, vint64m2_t * op3, const int64_t * op4, vuint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_i64m2_m))) +void vluxseg4ei8(vint64m2_t * op0, vint64m2_t * op1, vint64m2_t * op2, vint64m2_t * op3, vbool32_t op4, vint64m2_t op5, vint64m2_t op6, vint64m2_t op7, vint64m2_t op8, const int64_t * op9, vuint8mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_u64m1))) +void vluxseg4ei8(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, const uint64_t * op4, vuint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_u64m1_m))) +void vluxseg4ei8(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vbool64_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, const uint64_t * op9, vuint8mf8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_u64m2))) +void vluxseg4ei8(vuint64m2_t * op0, vuint64m2_t * op1, vuint64m2_t * op2, vuint64m2_t * op3, const uint64_t * op4, vuint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_u64m2_m))) +void vluxseg4ei8(vuint64m2_t * op0, vuint64m2_t * op1, vuint64m2_t * op2, vuint64m2_t * op3, vbool32_t op4, vuint64m2_t op5, vuint64m2_t op6, vuint64m2_t op7, vuint64m2_t op8, const uint64_t * op9, vuint8mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_i64m1))) +void vluxseg5ei8(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, const int64_t * op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_i64m1_m))) +void vluxseg5ei8(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vbool64_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, vint64m1_t op10, const int64_t * op11, vuint8mf8_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_u64m1))) +void vluxseg5ei8(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, const uint64_t * op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_u64m1_m))) +void vluxseg5ei8(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vbool64_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, vuint64m1_t op10, const uint64_t * op11, vuint8mf8_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_i64m1))) +void vluxseg6ei8(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, const int64_t * op6, vuint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_i64m1_m))) +void vluxseg6ei8(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vbool64_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, vint64m1_t op10, vint64m1_t op11, vint64m1_t op12, const int64_t * op13, vuint8mf8_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_u64m1))) +void vluxseg6ei8(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, const uint64_t * op6, vuint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_u64m1_m))) +void vluxseg6ei8(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vbool64_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, vuint64m1_t op10, vuint64m1_t op11, vuint64m1_t op12, const uint64_t * op13, vuint8mf8_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_i64m1))) +void vluxseg7ei8(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vint64m1_t * op6, const int64_t * op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_i64m1_m))) +void vluxseg7ei8(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vint64m1_t * op6, vbool64_t op7, vint64m1_t op8, vint64m1_t op9, vint64m1_t op10, vint64m1_t op11, vint64m1_t op12, vint64m1_t op13, vint64m1_t op14, const int64_t * op15, vuint8mf8_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_u64m1))) +void vluxseg7ei8(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vuint64m1_t * op6, const uint64_t * op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_u64m1_m))) +void vluxseg7ei8(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vuint64m1_t * op6, vbool64_t op7, vuint64m1_t op8, vuint64m1_t op9, vuint64m1_t op10, vuint64m1_t op11, vuint64m1_t op12, vuint64m1_t op13, vuint64m1_t op14, const uint64_t * op15, vuint8mf8_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_i64m1))) +void vluxseg8ei8(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vint64m1_t * op6, vint64m1_t * op7, const int64_t * op8, vuint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_i64m1_m))) +void vluxseg8ei8(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vint64m1_t * op6, vint64m1_t * op7, vbool64_t op8, vint64m1_t op9, vint64m1_t op10, vint64m1_t op11, vint64m1_t op12, vint64m1_t op13, vint64m1_t op14, vint64m1_t op15, vint64m1_t op16, const int64_t * op17, vuint8mf8_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_u64m1))) +void vluxseg8ei8(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vuint64m1_t * op6, vuint64m1_t * op7, const uint64_t * op8, vuint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_u64m1_m))) +void vluxseg8ei8(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vuint64m1_t * op6, vuint64m1_t * op7, vbool64_t op8, vuint64m1_t op9, vuint64m1_t op10, vuint64m1_t op11, vuint64m1_t op12, vuint64m1_t op13, vuint64m1_t op14, vuint64m1_t op15, vuint64m1_t op16, const uint64_t * op17, vuint8mf8_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_i64m1))) +void vluxseg2ei16(vint64m1_t * op0, vint64m1_t * op1, const int64_t * op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_i64m1_m))) +void vluxseg2ei16(vint64m1_t * op0, vint64m1_t * op1, vbool64_t op2, vint64m1_t op3, vint64m1_t op4, const int64_t * op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_i64m2))) +void vluxseg2ei16(vint64m2_t * op0, vint64m2_t * op1, const int64_t * op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_i64m2_m))) +void vluxseg2ei16(vint64m2_t * op0, vint64m2_t * op1, vbool32_t op2, vint64m2_t op3, vint64m2_t op4, const int64_t * op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_i64m4))) +void vluxseg2ei16(vint64m4_t * op0, vint64m4_t * op1, const int64_t * op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_i64m4_m))) +void vluxseg2ei16(vint64m4_t * op0, vint64m4_t * op1, vbool16_t op2, vint64m4_t op3, vint64m4_t op4, const int64_t * op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_u64m1))) +void vluxseg2ei16(vuint64m1_t * op0, vuint64m1_t * op1, const uint64_t * op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_u64m1_m))) +void vluxseg2ei16(vuint64m1_t * op0, vuint64m1_t * op1, vbool64_t op2, vuint64m1_t op3, vuint64m1_t op4, const uint64_t * op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_u64m2))) +void vluxseg2ei16(vuint64m2_t * op0, vuint64m2_t * op1, const uint64_t * op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_u64m2_m))) +void vluxseg2ei16(vuint64m2_t * op0, vuint64m2_t * op1, vbool32_t op2, vuint64m2_t op3, vuint64m2_t op4, const uint64_t * op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_u64m4))) +void vluxseg2ei16(vuint64m4_t * op0, vuint64m4_t * op1, const uint64_t * op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_u64m4_m))) +void vluxseg2ei16(vuint64m4_t * op0, vuint64m4_t * op1, vbool16_t op2, vuint64m4_t op3, vuint64m4_t op4, const uint64_t * op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_i64m1))) +void vluxseg3ei16(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, const int64_t * op3, vuint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_i64m1_m))) +void vluxseg3ei16(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vbool64_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, const int64_t * op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_i64m2))) +void vluxseg3ei16(vint64m2_t * op0, vint64m2_t * op1, vint64m2_t * op2, const int64_t * op3, vuint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_i64m2_m))) +void vluxseg3ei16(vint64m2_t * op0, vint64m2_t * op1, vint64m2_t * op2, vbool32_t op3, vint64m2_t op4, vint64m2_t op5, vint64m2_t op6, const int64_t * op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_u64m1))) +void vluxseg3ei16(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, const uint64_t * op3, vuint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_u64m1_m))) +void vluxseg3ei16(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vbool64_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, const uint64_t * op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_u64m2))) +void vluxseg3ei16(vuint64m2_t * op0, vuint64m2_t * op1, vuint64m2_t * op2, const uint64_t * op3, vuint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_u64m2_m))) +void vluxseg3ei16(vuint64m2_t * op0, vuint64m2_t * op1, vuint64m2_t * op2, vbool32_t op3, vuint64m2_t op4, vuint64m2_t op5, vuint64m2_t op6, const uint64_t * op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_i64m1))) +void vluxseg4ei16(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, const int64_t * op4, vuint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_i64m1_m))) +void vluxseg4ei16(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vbool64_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, const int64_t * op9, vuint16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_i64m2))) +void vluxseg4ei16(vint64m2_t * op0, vint64m2_t * op1, vint64m2_t * op2, vint64m2_t * op3, const int64_t * op4, vuint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_i64m2_m))) +void vluxseg4ei16(vint64m2_t * op0, vint64m2_t * op1, vint64m2_t * op2, vint64m2_t * op3, vbool32_t op4, vint64m2_t op5, vint64m2_t op6, vint64m2_t op7, vint64m2_t op8, const int64_t * op9, vuint16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_u64m1))) +void vluxseg4ei16(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, const uint64_t * op4, vuint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_u64m1_m))) +void vluxseg4ei16(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vbool64_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, const uint64_t * op9, vuint16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_u64m2))) +void vluxseg4ei16(vuint64m2_t * op0, vuint64m2_t * op1, vuint64m2_t * op2, vuint64m2_t * op3, const uint64_t * op4, vuint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_u64m2_m))) +void vluxseg4ei16(vuint64m2_t * op0, vuint64m2_t * op1, vuint64m2_t * op2, vuint64m2_t * op3, vbool32_t op4, vuint64m2_t op5, vuint64m2_t op6, vuint64m2_t op7, vuint64m2_t op8, const uint64_t * op9, vuint16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_i64m1))) +void vluxseg5ei16(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, const int64_t * op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_i64m1_m))) +void vluxseg5ei16(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vbool64_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, vint64m1_t op10, const int64_t * op11, vuint16mf4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_u64m1))) +void vluxseg5ei16(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, const uint64_t * op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_u64m1_m))) +void vluxseg5ei16(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vbool64_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, vuint64m1_t op10, const uint64_t * op11, vuint16mf4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_i64m1))) +void vluxseg6ei16(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, const int64_t * op6, vuint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_i64m1_m))) +void vluxseg6ei16(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vbool64_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, vint64m1_t op10, vint64m1_t op11, vint64m1_t op12, const int64_t * op13, vuint16mf4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_u64m1))) +void vluxseg6ei16(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, const uint64_t * op6, vuint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_u64m1_m))) +void vluxseg6ei16(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vbool64_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, vuint64m1_t op10, vuint64m1_t op11, vuint64m1_t op12, const uint64_t * op13, vuint16mf4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_i64m1))) +void vluxseg7ei16(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vint64m1_t * op6, const int64_t * op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_i64m1_m))) +void vluxseg7ei16(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vint64m1_t * op6, vbool64_t op7, vint64m1_t op8, vint64m1_t op9, vint64m1_t op10, vint64m1_t op11, vint64m1_t op12, vint64m1_t op13, vint64m1_t op14, const int64_t * op15, vuint16mf4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_u64m1))) +void vluxseg7ei16(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vuint64m1_t * op6, const uint64_t * op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_u64m1_m))) +void vluxseg7ei16(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vuint64m1_t * op6, vbool64_t op7, vuint64m1_t op8, vuint64m1_t op9, vuint64m1_t op10, vuint64m1_t op11, vuint64m1_t op12, vuint64m1_t op13, vuint64m1_t op14, const uint64_t * op15, vuint16mf4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_i64m1))) +void vluxseg8ei16(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vint64m1_t * op6, vint64m1_t * op7, const int64_t * op8, vuint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_i64m1_m))) +void vluxseg8ei16(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vint64m1_t * op6, vint64m1_t * op7, vbool64_t op8, vint64m1_t op9, vint64m1_t op10, vint64m1_t op11, vint64m1_t op12, vint64m1_t op13, vint64m1_t op14, vint64m1_t op15, vint64m1_t op16, const int64_t * op17, vuint16mf4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_u64m1))) +void vluxseg8ei16(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vuint64m1_t * op6, vuint64m1_t * op7, const uint64_t * op8, vuint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_u64m1_m))) +void vluxseg8ei16(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vuint64m1_t * op6, vuint64m1_t * op7, vbool64_t op8, vuint64m1_t op9, vuint64m1_t op10, vuint64m1_t op11, vuint64m1_t op12, vuint64m1_t op13, vuint64m1_t op14, vuint64m1_t op15, vuint64m1_t op16, const uint64_t * op17, vuint16mf4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_i64m1))) +void vluxseg2ei32(vint64m1_t * op0, vint64m1_t * op1, const int64_t * op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_i64m1_m))) +void vluxseg2ei32(vint64m1_t * op0, vint64m1_t * op1, vbool64_t op2, vint64m1_t op3, vint64m1_t op4, const int64_t * op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_i64m2))) +void vluxseg2ei32(vint64m2_t * op0, vint64m2_t * op1, const int64_t * op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_i64m2_m))) +void vluxseg2ei32(vint64m2_t * op0, vint64m2_t * op1, vbool32_t op2, vint64m2_t op3, vint64m2_t op4, const int64_t * op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_i64m4))) +void vluxseg2ei32(vint64m4_t * op0, vint64m4_t * op1, const int64_t * op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_i64m4_m))) +void vluxseg2ei32(vint64m4_t * op0, vint64m4_t * op1, vbool16_t op2, vint64m4_t op3, vint64m4_t op4, const int64_t * op5, vuint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_u64m1))) +void vluxseg2ei32(vuint64m1_t * op0, vuint64m1_t * op1, const uint64_t * op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_u64m1_m))) +void vluxseg2ei32(vuint64m1_t * op0, vuint64m1_t * op1, vbool64_t op2, vuint64m1_t op3, vuint64m1_t op4, const uint64_t * op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_u64m2))) +void vluxseg2ei32(vuint64m2_t * op0, vuint64m2_t * op1, const uint64_t * op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_u64m2_m))) +void vluxseg2ei32(vuint64m2_t * op0, vuint64m2_t * op1, vbool32_t op2, vuint64m2_t op3, vuint64m2_t op4, const uint64_t * op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_u64m4))) +void vluxseg2ei32(vuint64m4_t * op0, vuint64m4_t * op1, const uint64_t * op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_u64m4_m))) +void vluxseg2ei32(vuint64m4_t * op0, vuint64m4_t * op1, vbool16_t op2, vuint64m4_t op3, vuint64m4_t op4, const uint64_t * op5, vuint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_i64m1))) +void vluxseg3ei32(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, const int64_t * op3, vuint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_i64m1_m))) +void vluxseg3ei32(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vbool64_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, const int64_t * op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_i64m2))) +void vluxseg3ei32(vint64m2_t * op0, vint64m2_t * op1, vint64m2_t * op2, const int64_t * op3, vuint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_i64m2_m))) +void vluxseg3ei32(vint64m2_t * op0, vint64m2_t * op1, vint64m2_t * op2, vbool32_t op3, vint64m2_t op4, vint64m2_t op5, vint64m2_t op6, const int64_t * op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_u64m1))) +void vluxseg3ei32(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, const uint64_t * op3, vuint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_u64m1_m))) +void vluxseg3ei32(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vbool64_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, const uint64_t * op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_u64m2))) +void vluxseg3ei32(vuint64m2_t * op0, vuint64m2_t * op1, vuint64m2_t * op2, const uint64_t * op3, vuint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_u64m2_m))) +void vluxseg3ei32(vuint64m2_t * op0, vuint64m2_t * op1, vuint64m2_t * op2, vbool32_t op3, vuint64m2_t op4, vuint64m2_t op5, vuint64m2_t op6, const uint64_t * op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_i64m1))) +void vluxseg4ei32(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, const int64_t * op4, vuint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_i64m1_m))) +void vluxseg4ei32(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vbool64_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, const int64_t * op9, vuint32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_i64m2))) +void vluxseg4ei32(vint64m2_t * op0, vint64m2_t * op1, vint64m2_t * op2, vint64m2_t * op3, const int64_t * op4, vuint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_i64m2_m))) +void vluxseg4ei32(vint64m2_t * op0, vint64m2_t * op1, vint64m2_t * op2, vint64m2_t * op3, vbool32_t op4, vint64m2_t op5, vint64m2_t op6, vint64m2_t op7, vint64m2_t op8, const int64_t * op9, vuint32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_u64m1))) +void vluxseg4ei32(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, const uint64_t * op4, vuint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_u64m1_m))) +void vluxseg4ei32(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vbool64_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, const uint64_t * op9, vuint32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_u64m2))) +void vluxseg4ei32(vuint64m2_t * op0, vuint64m2_t * op1, vuint64m2_t * op2, vuint64m2_t * op3, const uint64_t * op4, vuint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_u64m2_m))) +void vluxseg4ei32(vuint64m2_t * op0, vuint64m2_t * op1, vuint64m2_t * op2, vuint64m2_t * op3, vbool32_t op4, vuint64m2_t op5, vuint64m2_t op6, vuint64m2_t op7, vuint64m2_t op8, const uint64_t * op9, vuint32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_i64m1))) +void vluxseg5ei32(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, const int64_t * op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_i64m1_m))) +void vluxseg5ei32(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vbool64_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, vint64m1_t op10, const int64_t * op11, vuint32mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_u64m1))) +void vluxseg5ei32(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, const uint64_t * op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_u64m1_m))) +void vluxseg5ei32(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vbool64_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, vuint64m1_t op10, const uint64_t * op11, vuint32mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_i64m1))) +void vluxseg6ei32(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, const int64_t * op6, vuint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_i64m1_m))) +void vluxseg6ei32(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vbool64_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, vint64m1_t op10, vint64m1_t op11, vint64m1_t op12, const int64_t * op13, vuint32mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_u64m1))) +void vluxseg6ei32(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, const uint64_t * op6, vuint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_u64m1_m))) +void vluxseg6ei32(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vbool64_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, vuint64m1_t op10, vuint64m1_t op11, vuint64m1_t op12, const uint64_t * op13, vuint32mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_i64m1))) +void vluxseg7ei32(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vint64m1_t * op6, const int64_t * op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_i64m1_m))) +void vluxseg7ei32(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vint64m1_t * op6, vbool64_t op7, vint64m1_t op8, vint64m1_t op9, vint64m1_t op10, vint64m1_t op11, vint64m1_t op12, vint64m1_t op13, vint64m1_t op14, const int64_t * op15, vuint32mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_u64m1))) +void vluxseg7ei32(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vuint64m1_t * op6, const uint64_t * op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_u64m1_m))) +void vluxseg7ei32(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vuint64m1_t * op6, vbool64_t op7, vuint64m1_t op8, vuint64m1_t op9, vuint64m1_t op10, vuint64m1_t op11, vuint64m1_t op12, vuint64m1_t op13, vuint64m1_t op14, const uint64_t * op15, vuint32mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_i64m1))) +void vluxseg8ei32(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vint64m1_t * op6, vint64m1_t * op7, const int64_t * op8, vuint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_i64m1_m))) +void vluxseg8ei32(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vint64m1_t * op6, vint64m1_t * op7, vbool64_t op8, vint64m1_t op9, vint64m1_t op10, vint64m1_t op11, vint64m1_t op12, vint64m1_t op13, vint64m1_t op14, vint64m1_t op15, vint64m1_t op16, const int64_t * op17, vuint32mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_u64m1))) +void vluxseg8ei32(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vuint64m1_t * op6, vuint64m1_t * op7, const uint64_t * op8, vuint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_u64m1_m))) +void vluxseg8ei32(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vuint64m1_t * op6, vuint64m1_t * op7, vbool64_t op8, vuint64m1_t op9, vuint64m1_t op10, vuint64m1_t op11, vuint64m1_t op12, vuint64m1_t op13, vuint64m1_t op14, vuint64m1_t op15, vuint64m1_t op16, const uint64_t * op17, vuint32mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_i64m1))) +void vluxseg2ei64(vint64m1_t * op0, vint64m1_t * op1, const int64_t * op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_i64m1_m))) +void vluxseg2ei64(vint64m1_t * op0, vint64m1_t * op1, vbool64_t op2, vint64m1_t op3, vint64m1_t op4, const int64_t * op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_i64m2))) +void vluxseg2ei64(vint64m2_t * op0, vint64m2_t * op1, const int64_t * op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_i64m2_m))) +void vluxseg2ei64(vint64m2_t * op0, vint64m2_t * op1, vbool32_t op2, vint64m2_t op3, vint64m2_t op4, const int64_t * op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_i64m4))) +void vluxseg2ei64(vint64m4_t * op0, vint64m4_t * op1, const int64_t * op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_i64m4_m))) +void vluxseg2ei64(vint64m4_t * op0, vint64m4_t * op1, vbool16_t op2, vint64m4_t op3, vint64m4_t op4, const int64_t * op5, vuint64m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_u64m1))) +void vluxseg2ei64(vuint64m1_t * op0, vuint64m1_t * op1, const uint64_t * op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_u64m1_m))) +void vluxseg2ei64(vuint64m1_t * op0, vuint64m1_t * op1, vbool64_t op2, vuint64m1_t op3, vuint64m1_t op4, const uint64_t * op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_u64m2))) +void vluxseg2ei64(vuint64m2_t * op0, vuint64m2_t * op1, const uint64_t * op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_u64m2_m))) +void vluxseg2ei64(vuint64m2_t * op0, vuint64m2_t * op1, vbool32_t op2, vuint64m2_t op3, vuint64m2_t op4, const uint64_t * op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_u64m4))) +void vluxseg2ei64(vuint64m4_t * op0, vuint64m4_t * op1, const uint64_t * op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_u64m4_m))) +void vluxseg2ei64(vuint64m4_t * op0, vuint64m4_t * op1, vbool16_t op2, vuint64m4_t op3, vuint64m4_t op4, const uint64_t * op5, vuint64m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_i64m1))) +void vluxseg3ei64(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, const int64_t * op3, vuint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_i64m1_m))) +void vluxseg3ei64(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vbool64_t op3, vint64m1_t op4, vint64m1_t op5, vint64m1_t op6, const int64_t * op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_i64m2))) +void vluxseg3ei64(vint64m2_t * op0, vint64m2_t * op1, vint64m2_t * op2, const int64_t * op3, vuint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_i64m2_m))) +void vluxseg3ei64(vint64m2_t * op0, vint64m2_t * op1, vint64m2_t * op2, vbool32_t op3, vint64m2_t op4, vint64m2_t op5, vint64m2_t op6, const int64_t * op7, vuint64m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_u64m1))) +void vluxseg3ei64(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, const uint64_t * op3, vuint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_u64m1_m))) +void vluxseg3ei64(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vbool64_t op3, vuint64m1_t op4, vuint64m1_t op5, vuint64m1_t op6, const uint64_t * op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_u64m2))) +void vluxseg3ei64(vuint64m2_t * op0, vuint64m2_t * op1, vuint64m2_t * op2, const uint64_t * op3, vuint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_u64m2_m))) +void vluxseg3ei64(vuint64m2_t * op0, vuint64m2_t * op1, vuint64m2_t * op2, vbool32_t op3, vuint64m2_t op4, vuint64m2_t op5, vuint64m2_t op6, const uint64_t * op7, vuint64m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_i64m1))) +void vluxseg4ei64(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, const int64_t * op4, vuint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_i64m1_m))) +void vluxseg4ei64(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vbool64_t op4, vint64m1_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, const int64_t * op9, vuint64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_i64m2))) +void vluxseg4ei64(vint64m2_t * op0, vint64m2_t * op1, vint64m2_t * op2, vint64m2_t * op3, const int64_t * op4, vuint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_i64m2_m))) +void vluxseg4ei64(vint64m2_t * op0, vint64m2_t * op1, vint64m2_t * op2, vint64m2_t * op3, vbool32_t op4, vint64m2_t op5, vint64m2_t op6, vint64m2_t op7, vint64m2_t op8, const int64_t * op9, vuint64m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_u64m1))) +void vluxseg4ei64(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, const uint64_t * op4, vuint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_u64m1_m))) +void vluxseg4ei64(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vbool64_t op4, vuint64m1_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, const uint64_t * op9, vuint64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_u64m2))) +void vluxseg4ei64(vuint64m2_t * op0, vuint64m2_t * op1, vuint64m2_t * op2, vuint64m2_t * op3, const uint64_t * op4, vuint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_u64m2_m))) +void vluxseg4ei64(vuint64m2_t * op0, vuint64m2_t * op1, vuint64m2_t * op2, vuint64m2_t * op3, vbool32_t op4, vuint64m2_t op5, vuint64m2_t op6, vuint64m2_t op7, vuint64m2_t op8, const uint64_t * op9, vuint64m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_i64m1))) +void vluxseg5ei64(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, const int64_t * op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_i64m1_m))) +void vluxseg5ei64(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vbool64_t op5, vint64m1_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, vint64m1_t op10, const int64_t * op11, vuint64m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_u64m1))) +void vluxseg5ei64(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, const uint64_t * op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_u64m1_m))) +void vluxseg5ei64(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vbool64_t op5, vuint64m1_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, vuint64m1_t op10, const uint64_t * op11, vuint64m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_i64m1))) +void vluxseg6ei64(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, const int64_t * op6, vuint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_i64m1_m))) +void vluxseg6ei64(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vbool64_t op6, vint64m1_t op7, vint64m1_t op8, vint64m1_t op9, vint64m1_t op10, vint64m1_t op11, vint64m1_t op12, const int64_t * op13, vuint64m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_u64m1))) +void vluxseg6ei64(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, const uint64_t * op6, vuint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_u64m1_m))) +void vluxseg6ei64(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vbool64_t op6, vuint64m1_t op7, vuint64m1_t op8, vuint64m1_t op9, vuint64m1_t op10, vuint64m1_t op11, vuint64m1_t op12, const uint64_t * op13, vuint64m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_i64m1))) +void vluxseg7ei64(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vint64m1_t * op6, const int64_t * op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_i64m1_m))) +void vluxseg7ei64(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vint64m1_t * op6, vbool64_t op7, vint64m1_t op8, vint64m1_t op9, vint64m1_t op10, vint64m1_t op11, vint64m1_t op12, vint64m1_t op13, vint64m1_t op14, const int64_t * op15, vuint64m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_u64m1))) +void vluxseg7ei64(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vuint64m1_t * op6, const uint64_t * op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_u64m1_m))) +void vluxseg7ei64(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vuint64m1_t * op6, vbool64_t op7, vuint64m1_t op8, vuint64m1_t op9, vuint64m1_t op10, vuint64m1_t op11, vuint64m1_t op12, vuint64m1_t op13, vuint64m1_t op14, const uint64_t * op15, vuint64m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_i64m1))) +void vluxseg8ei64(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vint64m1_t * op6, vint64m1_t * op7, const int64_t * op8, vuint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_i64m1_m))) +void vluxseg8ei64(vint64m1_t * op0, vint64m1_t * op1, vint64m1_t * op2, vint64m1_t * op3, vint64m1_t * op4, vint64m1_t * op5, vint64m1_t * op6, vint64m1_t * op7, vbool64_t op8, vint64m1_t op9, vint64m1_t op10, vint64m1_t op11, vint64m1_t op12, vint64m1_t op13, vint64m1_t op14, vint64m1_t op15, vint64m1_t op16, const int64_t * op17, vuint64m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_u64m1))) +void vluxseg8ei64(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vuint64m1_t * op6, vuint64m1_t * op7, const uint64_t * op8, vuint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_u64m1_m))) +void vluxseg8ei64(vuint64m1_t * op0, vuint64m1_t * op1, vuint64m1_t * op2, vuint64m1_t * op3, vuint64m1_t * op4, vuint64m1_t * op5, vuint64m1_t * op6, vuint64m1_t * op7, vbool64_t op8, vuint64m1_t op9, vuint64m1_t op10, vuint64m1_t op11, vuint64m1_t op12, vuint64m1_t op13, vuint64m1_t op14, vuint64m1_t op15, vuint64m1_t op16, const uint64_t * op17, vuint64m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_i8m1))) +void vloxseg2ei8(vint8m1_t * op0, vint8m1_t * op1, const int8_t * op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_i8m1_m))) +void vloxseg2ei8(vint8m1_t * op0, vint8m1_t * op1, vbool8_t op2, vint8m1_t op3, vint8m1_t op4, const int8_t * op5, vuint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_i8m2))) +void vloxseg2ei8(vint8m2_t * op0, vint8m2_t * op1, const int8_t * op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_i8m2_m))) +void vloxseg2ei8(vint8m2_t * op0, vint8m2_t * op1, vbool4_t op2, vint8m2_t op3, vint8m2_t op4, const int8_t * op5, vuint8m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_i8m4))) +void vloxseg2ei8(vint8m4_t * op0, vint8m4_t * op1, const int8_t * op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_i8m4_m))) +void vloxseg2ei8(vint8m4_t * op0, vint8m4_t * op1, vbool2_t op2, vint8m4_t op3, vint8m4_t op4, const int8_t * op5, vuint8m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_i8mf2))) +void vloxseg2ei8(vint8mf2_t * op0, vint8mf2_t * op1, const int8_t * op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_i8mf2_m))) +void vloxseg2ei8(vint8mf2_t * op0, vint8mf2_t * op1, vbool16_t op2, vint8mf2_t op3, vint8mf2_t op4, const int8_t * op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_i8mf4))) +void vloxseg2ei8(vint8mf4_t * op0, vint8mf4_t * op1, const int8_t * op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_i8mf4_m))) +void vloxseg2ei8(vint8mf4_t * op0, vint8mf4_t * op1, vbool32_t op2, vint8mf4_t op3, vint8mf4_t op4, const int8_t * op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_i8mf8))) +void vloxseg2ei8(vint8mf8_t * op0, vint8mf8_t * op1, const int8_t * op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_i8mf8_m))) +void vloxseg2ei8(vint8mf8_t * op0, vint8mf8_t * op1, vbool64_t op2, vint8mf8_t op3, vint8mf8_t op4, const int8_t * op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_u8m1))) +void vloxseg2ei8(vuint8m1_t * op0, vuint8m1_t * op1, const uint8_t * op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_u8m1_m))) +void vloxseg2ei8(vuint8m1_t * op0, vuint8m1_t * op1, vbool8_t op2, vuint8m1_t op3, vuint8m1_t op4, const uint8_t * op5, vuint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_u8m2))) +void vloxseg2ei8(vuint8m2_t * op0, vuint8m2_t * op1, const uint8_t * op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_u8m2_m))) +void vloxseg2ei8(vuint8m2_t * op0, vuint8m2_t * op1, vbool4_t op2, vuint8m2_t op3, vuint8m2_t op4, const uint8_t * op5, vuint8m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_u8m4))) +void vloxseg2ei8(vuint8m4_t * op0, vuint8m4_t * op1, const uint8_t * op2, vuint8m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_u8m4_m))) +void vloxseg2ei8(vuint8m4_t * op0, vuint8m4_t * op1, vbool2_t op2, vuint8m4_t op3, vuint8m4_t op4, const uint8_t * op5, vuint8m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_u8mf2))) +void vloxseg2ei8(vuint8mf2_t * op0, vuint8mf2_t * op1, const uint8_t * op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_u8mf2_m))) +void vloxseg2ei8(vuint8mf2_t * op0, vuint8mf2_t * op1, vbool16_t op2, vuint8mf2_t op3, vuint8mf2_t op4, const uint8_t * op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_u8mf4))) +void vloxseg2ei8(vuint8mf4_t * op0, vuint8mf4_t * op1, const uint8_t * op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_u8mf4_m))) +void vloxseg2ei8(vuint8mf4_t * op0, vuint8mf4_t * op1, vbool32_t op2, vuint8mf4_t op3, vuint8mf4_t op4, const uint8_t * op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_u8mf8))) +void vloxseg2ei8(vuint8mf8_t * op0, vuint8mf8_t * op1, const uint8_t * op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_u8mf8_m))) +void vloxseg2ei8(vuint8mf8_t * op0, vuint8mf8_t * op1, vbool64_t op2, vuint8mf8_t op3, vuint8mf8_t op4, const uint8_t * op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_i8m1))) +void vloxseg3ei8(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, const int8_t * op3, vuint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_i8m1_m))) +void vloxseg3ei8(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vbool8_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, const int8_t * op7, vuint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_i8m2))) +void vloxseg3ei8(vint8m2_t * op0, vint8m2_t * op1, vint8m2_t * op2, const int8_t * op3, vuint8m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_i8m2_m))) +void vloxseg3ei8(vint8m2_t * op0, vint8m2_t * op1, vint8m2_t * op2, vbool4_t op3, vint8m2_t op4, vint8m2_t op5, vint8m2_t op6, const int8_t * op7, vuint8m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_i8mf2))) +void vloxseg3ei8(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, const int8_t * op3, vuint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_i8mf2_m))) +void vloxseg3ei8(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vbool16_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, const int8_t * op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_i8mf4))) +void vloxseg3ei8(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, const int8_t * op3, vuint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_i8mf4_m))) +void vloxseg3ei8(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vbool32_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, const int8_t * op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_i8mf8))) +void vloxseg3ei8(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, const int8_t * op3, vuint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_i8mf8_m))) +void vloxseg3ei8(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vbool64_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, const int8_t * op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_u8m1))) +void vloxseg3ei8(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, const uint8_t * op3, vuint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_u8m1_m))) +void vloxseg3ei8(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vbool8_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, const uint8_t * op7, vuint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_u8m2))) +void vloxseg3ei8(vuint8m2_t * op0, vuint8m2_t * op1, vuint8m2_t * op2, const uint8_t * op3, vuint8m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_u8m2_m))) +void vloxseg3ei8(vuint8m2_t * op0, vuint8m2_t * op1, vuint8m2_t * op2, vbool4_t op3, vuint8m2_t op4, vuint8m2_t op5, vuint8m2_t op6, const uint8_t * op7, vuint8m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_u8mf2))) +void vloxseg3ei8(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, const uint8_t * op3, vuint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_u8mf2_m))) +void vloxseg3ei8(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vbool16_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, const uint8_t * op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_u8mf4))) +void vloxseg3ei8(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, const uint8_t * op3, vuint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_u8mf4_m))) +void vloxseg3ei8(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vbool32_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, const uint8_t * op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_u8mf8))) +void vloxseg3ei8(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, const uint8_t * op3, vuint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_u8mf8_m))) +void vloxseg3ei8(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vbool64_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, const uint8_t * op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_i8m1))) +void vloxseg4ei8(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, const int8_t * op4, vuint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_i8m1_m))) +void vloxseg4ei8(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vbool8_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, const int8_t * op9, vuint8m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_i8m2))) +void vloxseg4ei8(vint8m2_t * op0, vint8m2_t * op1, vint8m2_t * op2, vint8m2_t * op3, const int8_t * op4, vuint8m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_i8m2_m))) +void vloxseg4ei8(vint8m2_t * op0, vint8m2_t * op1, vint8m2_t * op2, vint8m2_t * op3, vbool4_t op4, vint8m2_t op5, vint8m2_t op6, vint8m2_t op7, vint8m2_t op8, const int8_t * op9, vuint8m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_i8mf2))) +void vloxseg4ei8(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, const int8_t * op4, vuint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_i8mf2_m))) +void vloxseg4ei8(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vbool16_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, const int8_t * op9, vuint8mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_i8mf4))) +void vloxseg4ei8(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, const int8_t * op4, vuint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_i8mf4_m))) +void vloxseg4ei8(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vbool32_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, const int8_t * op9, vuint8mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_i8mf8))) +void vloxseg4ei8(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, const int8_t * op4, vuint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_i8mf8_m))) +void vloxseg4ei8(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vbool64_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, const int8_t * op9, vuint8mf8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_u8m1))) +void vloxseg4ei8(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, const uint8_t * op4, vuint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_u8m1_m))) +void vloxseg4ei8(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vbool8_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, const uint8_t * op9, vuint8m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_u8m2))) +void vloxseg4ei8(vuint8m2_t * op0, vuint8m2_t * op1, vuint8m2_t * op2, vuint8m2_t * op3, const uint8_t * op4, vuint8m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_u8m2_m))) +void vloxseg4ei8(vuint8m2_t * op0, vuint8m2_t * op1, vuint8m2_t * op2, vuint8m2_t * op3, vbool4_t op4, vuint8m2_t op5, vuint8m2_t op6, vuint8m2_t op7, vuint8m2_t op8, const uint8_t * op9, vuint8m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_u8mf2))) +void vloxseg4ei8(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, const uint8_t * op4, vuint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_u8mf2_m))) +void vloxseg4ei8(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vbool16_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, const uint8_t * op9, vuint8mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_u8mf4))) +void vloxseg4ei8(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, const uint8_t * op4, vuint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_u8mf4_m))) +void vloxseg4ei8(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vbool32_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, const uint8_t * op9, vuint8mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_u8mf8))) +void vloxseg4ei8(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, const uint8_t * op4, vuint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_u8mf8_m))) +void vloxseg4ei8(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vbool64_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, const uint8_t * op9, vuint8mf8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_i8m1))) +void vloxseg5ei8(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, const int8_t * op5, vuint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_i8m1_m))) +void vloxseg5ei8(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vbool8_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, vint8m1_t op10, const int8_t * op11, vuint8m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_i8mf2))) +void vloxseg5ei8(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, const int8_t * op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_i8mf2_m))) +void vloxseg5ei8(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vbool16_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, vint8mf2_t op10, const int8_t * op11, vuint8mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_i8mf4))) +void vloxseg5ei8(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, const int8_t * op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_i8mf4_m))) +void vloxseg5ei8(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vbool32_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, vint8mf4_t op10, const int8_t * op11, vuint8mf4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_i8mf8))) +void vloxseg5ei8(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, const int8_t * op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_i8mf8_m))) +void vloxseg5ei8(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vbool64_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, vint8mf8_t op10, const int8_t * op11, vuint8mf8_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_u8m1))) +void vloxseg5ei8(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, const uint8_t * op5, vuint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_u8m1_m))) +void vloxseg5ei8(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vbool8_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, vuint8m1_t op10, const uint8_t * op11, vuint8m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_u8mf2))) +void vloxseg5ei8(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, const uint8_t * op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_u8mf2_m))) +void vloxseg5ei8(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vbool16_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, vuint8mf2_t op10, const uint8_t * op11, vuint8mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_u8mf4))) +void vloxseg5ei8(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, const uint8_t * op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_u8mf4_m))) +void vloxseg5ei8(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vbool32_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, vuint8mf4_t op10, const uint8_t * op11, vuint8mf4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_u8mf8))) +void vloxseg5ei8(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, const uint8_t * op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_u8mf8_m))) +void vloxseg5ei8(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vbool64_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, vuint8mf8_t op10, const uint8_t * op11, vuint8mf8_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_i8m1))) +void vloxseg6ei8(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, const int8_t * op6, vuint8m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_i8m1_m))) +void vloxseg6ei8(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vbool8_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, vint8m1_t op10, vint8m1_t op11, vint8m1_t op12, const int8_t * op13, vuint8m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_i8mf2))) +void vloxseg6ei8(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, const int8_t * op6, vuint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_i8mf2_m))) +void vloxseg6ei8(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vbool16_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, vint8mf2_t op10, vint8mf2_t op11, vint8mf2_t op12, const int8_t * op13, vuint8mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_i8mf4))) +void vloxseg6ei8(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, const int8_t * op6, vuint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_i8mf4_m))) +void vloxseg6ei8(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vbool32_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, vint8mf4_t op10, vint8mf4_t op11, vint8mf4_t op12, const int8_t * op13, vuint8mf4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_i8mf8))) +void vloxseg6ei8(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, const int8_t * op6, vuint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_i8mf8_m))) +void vloxseg6ei8(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vbool64_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, vint8mf8_t op10, vint8mf8_t op11, vint8mf8_t op12, const int8_t * op13, vuint8mf8_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_u8m1))) +void vloxseg6ei8(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, const uint8_t * op6, vuint8m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_u8m1_m))) +void vloxseg6ei8(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vbool8_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, vuint8m1_t op10, vuint8m1_t op11, vuint8m1_t op12, const uint8_t * op13, vuint8m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_u8mf2))) +void vloxseg6ei8(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, const uint8_t * op6, vuint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_u8mf2_m))) +void vloxseg6ei8(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vbool16_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, vuint8mf2_t op10, vuint8mf2_t op11, vuint8mf2_t op12, const uint8_t * op13, vuint8mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_u8mf4))) +void vloxseg6ei8(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, const uint8_t * op6, vuint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_u8mf4_m))) +void vloxseg6ei8(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vbool32_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, vuint8mf4_t op10, vuint8mf4_t op11, vuint8mf4_t op12, const uint8_t * op13, vuint8mf4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_u8mf8))) +void vloxseg6ei8(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, const uint8_t * op6, vuint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_u8mf8_m))) +void vloxseg6ei8(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vbool64_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, vuint8mf8_t op10, vuint8mf8_t op11, vuint8mf8_t op12, const uint8_t * op13, vuint8mf8_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_i8m1))) +void vloxseg7ei8(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vint8m1_t * op6, const int8_t * op7, vuint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_i8m1_m))) +void vloxseg7ei8(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vint8m1_t * op6, vbool8_t op7, vint8m1_t op8, vint8m1_t op9, vint8m1_t op10, vint8m1_t op11, vint8m1_t op12, vint8m1_t op13, vint8m1_t op14, const int8_t * op15, vuint8m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_i8mf2))) +void vloxseg7ei8(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vint8mf2_t * op6, const int8_t * op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_i8mf2_m))) +void vloxseg7ei8(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vint8mf2_t * op6, vbool16_t op7, vint8mf2_t op8, vint8mf2_t op9, vint8mf2_t op10, vint8mf2_t op11, vint8mf2_t op12, vint8mf2_t op13, vint8mf2_t op14, const int8_t * op15, vuint8mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_i8mf4))) +void vloxseg7ei8(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vint8mf4_t * op6, const int8_t * op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_i8mf4_m))) +void vloxseg7ei8(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vint8mf4_t * op6, vbool32_t op7, vint8mf4_t op8, vint8mf4_t op9, vint8mf4_t op10, vint8mf4_t op11, vint8mf4_t op12, vint8mf4_t op13, vint8mf4_t op14, const int8_t * op15, vuint8mf4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_i8mf8))) +void vloxseg7ei8(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vint8mf8_t * op6, const int8_t * op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_i8mf8_m))) +void vloxseg7ei8(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vint8mf8_t * op6, vbool64_t op7, vint8mf8_t op8, vint8mf8_t op9, vint8mf8_t op10, vint8mf8_t op11, vint8mf8_t op12, vint8mf8_t op13, vint8mf8_t op14, const int8_t * op15, vuint8mf8_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_u8m1))) +void vloxseg7ei8(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vuint8m1_t * op6, const uint8_t * op7, vuint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_u8m1_m))) +void vloxseg7ei8(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vuint8m1_t * op6, vbool8_t op7, vuint8m1_t op8, vuint8m1_t op9, vuint8m1_t op10, vuint8m1_t op11, vuint8m1_t op12, vuint8m1_t op13, vuint8m1_t op14, const uint8_t * op15, vuint8m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_u8mf2))) +void vloxseg7ei8(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vuint8mf2_t * op6, const uint8_t * op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_u8mf2_m))) +void vloxseg7ei8(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vuint8mf2_t * op6, vbool16_t op7, vuint8mf2_t op8, vuint8mf2_t op9, vuint8mf2_t op10, vuint8mf2_t op11, vuint8mf2_t op12, vuint8mf2_t op13, vuint8mf2_t op14, const uint8_t * op15, vuint8mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_u8mf4))) +void vloxseg7ei8(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vuint8mf4_t * op6, const uint8_t * op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_u8mf4_m))) +void vloxseg7ei8(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vuint8mf4_t * op6, vbool32_t op7, vuint8mf4_t op8, vuint8mf4_t op9, vuint8mf4_t op10, vuint8mf4_t op11, vuint8mf4_t op12, vuint8mf4_t op13, vuint8mf4_t op14, const uint8_t * op15, vuint8mf4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_u8mf8))) +void vloxseg7ei8(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vuint8mf8_t * op6, const uint8_t * op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_u8mf8_m))) +void vloxseg7ei8(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vuint8mf8_t * op6, vbool64_t op7, vuint8mf8_t op8, vuint8mf8_t op9, vuint8mf8_t op10, vuint8mf8_t op11, vuint8mf8_t op12, vuint8mf8_t op13, vuint8mf8_t op14, const uint8_t * op15, vuint8mf8_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_i8m1))) +void vloxseg8ei8(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vint8m1_t * op6, vint8m1_t * op7, const int8_t * op8, vuint8m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_i8m1_m))) +void vloxseg8ei8(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vint8m1_t * op6, vint8m1_t * op7, vbool8_t op8, vint8m1_t op9, vint8m1_t op10, vint8m1_t op11, vint8m1_t op12, vint8m1_t op13, vint8m1_t op14, vint8m1_t op15, vint8m1_t op16, const int8_t * op17, vuint8m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_i8mf2))) +void vloxseg8ei8(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vint8mf2_t * op6, vint8mf2_t * op7, const int8_t * op8, vuint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_i8mf2_m))) +void vloxseg8ei8(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vint8mf2_t * op6, vint8mf2_t * op7, vbool16_t op8, vint8mf2_t op9, vint8mf2_t op10, vint8mf2_t op11, vint8mf2_t op12, vint8mf2_t op13, vint8mf2_t op14, vint8mf2_t op15, vint8mf2_t op16, const int8_t * op17, vuint8mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_i8mf4))) +void vloxseg8ei8(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vint8mf4_t * op6, vint8mf4_t * op7, const int8_t * op8, vuint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_i8mf4_m))) +void vloxseg8ei8(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vint8mf4_t * op6, vint8mf4_t * op7, vbool32_t op8, vint8mf4_t op9, vint8mf4_t op10, vint8mf4_t op11, vint8mf4_t op12, vint8mf4_t op13, vint8mf4_t op14, vint8mf4_t op15, vint8mf4_t op16, const int8_t * op17, vuint8mf4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_i8mf8))) +void vloxseg8ei8(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vint8mf8_t * op6, vint8mf8_t * op7, const int8_t * op8, vuint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_i8mf8_m))) +void vloxseg8ei8(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vint8mf8_t * op6, vint8mf8_t * op7, vbool64_t op8, vint8mf8_t op9, vint8mf8_t op10, vint8mf8_t op11, vint8mf8_t op12, vint8mf8_t op13, vint8mf8_t op14, vint8mf8_t op15, vint8mf8_t op16, const int8_t * op17, vuint8mf8_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_u8m1))) +void vloxseg8ei8(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vuint8m1_t * op6, vuint8m1_t * op7, const uint8_t * op8, vuint8m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_u8m1_m))) +void vloxseg8ei8(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vuint8m1_t * op6, vuint8m1_t * op7, vbool8_t op8, vuint8m1_t op9, vuint8m1_t op10, vuint8m1_t op11, vuint8m1_t op12, vuint8m1_t op13, vuint8m1_t op14, vuint8m1_t op15, vuint8m1_t op16, const uint8_t * op17, vuint8m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_u8mf2))) +void vloxseg8ei8(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vuint8mf2_t * op6, vuint8mf2_t * op7, const uint8_t * op8, vuint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_u8mf2_m))) +void vloxseg8ei8(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vuint8mf2_t * op6, vuint8mf2_t * op7, vbool16_t op8, vuint8mf2_t op9, vuint8mf2_t op10, vuint8mf2_t op11, vuint8mf2_t op12, vuint8mf2_t op13, vuint8mf2_t op14, vuint8mf2_t op15, vuint8mf2_t op16, const uint8_t * op17, vuint8mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_u8mf4))) +void vloxseg8ei8(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vuint8mf4_t * op6, vuint8mf4_t * op7, const uint8_t * op8, vuint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_u8mf4_m))) +void vloxseg8ei8(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vuint8mf4_t * op6, vuint8mf4_t * op7, vbool32_t op8, vuint8mf4_t op9, vuint8mf4_t op10, vuint8mf4_t op11, vuint8mf4_t op12, vuint8mf4_t op13, vuint8mf4_t op14, vuint8mf4_t op15, vuint8mf4_t op16, const uint8_t * op17, vuint8mf4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_u8mf8))) +void vloxseg8ei8(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vuint8mf8_t * op6, vuint8mf8_t * op7, const uint8_t * op8, vuint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_u8mf8_m))) +void vloxseg8ei8(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vuint8mf8_t * op6, vuint8mf8_t * op7, vbool64_t op8, vuint8mf8_t op9, vuint8mf8_t op10, vuint8mf8_t op11, vuint8mf8_t op12, vuint8mf8_t op13, vuint8mf8_t op14, vuint8mf8_t op15, vuint8mf8_t op16, const uint8_t * op17, vuint8mf8_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_i8m1))) +void vloxseg2ei16(vint8m1_t * op0, vint8m1_t * op1, const int8_t * op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_i8m1_m))) +void vloxseg2ei16(vint8m1_t * op0, vint8m1_t * op1, vbool8_t op2, vint8m1_t op3, vint8m1_t op4, const int8_t * op5, vuint16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_i8m2))) +void vloxseg2ei16(vint8m2_t * op0, vint8m2_t * op1, const int8_t * op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_i8m2_m))) +void vloxseg2ei16(vint8m2_t * op0, vint8m2_t * op1, vbool4_t op2, vint8m2_t op3, vint8m2_t op4, const int8_t * op5, vuint16m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_i8m4))) +void vloxseg2ei16(vint8m4_t * op0, vint8m4_t * op1, const int8_t * op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_i8m4_m))) +void vloxseg2ei16(vint8m4_t * op0, vint8m4_t * op1, vbool2_t op2, vint8m4_t op3, vint8m4_t op4, const int8_t * op5, vuint16m8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_i8mf2))) +void vloxseg2ei16(vint8mf2_t * op0, vint8mf2_t * op1, const int8_t * op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_i8mf2_m))) +void vloxseg2ei16(vint8mf2_t * op0, vint8mf2_t * op1, vbool16_t op2, vint8mf2_t op3, vint8mf2_t op4, const int8_t * op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_i8mf4))) +void vloxseg2ei16(vint8mf4_t * op0, vint8mf4_t * op1, const int8_t * op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_i8mf4_m))) +void vloxseg2ei16(vint8mf4_t * op0, vint8mf4_t * op1, vbool32_t op2, vint8mf4_t op3, vint8mf4_t op4, const int8_t * op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_i8mf8))) +void vloxseg2ei16(vint8mf8_t * op0, vint8mf8_t * op1, const int8_t * op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_i8mf8_m))) +void vloxseg2ei16(vint8mf8_t * op0, vint8mf8_t * op1, vbool64_t op2, vint8mf8_t op3, vint8mf8_t op4, const int8_t * op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_u8m1))) +void vloxseg2ei16(vuint8m1_t * op0, vuint8m1_t * op1, const uint8_t * op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_u8m1_m))) +void vloxseg2ei16(vuint8m1_t * op0, vuint8m1_t * op1, vbool8_t op2, vuint8m1_t op3, vuint8m1_t op4, const uint8_t * op5, vuint16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_u8m2))) +void vloxseg2ei16(vuint8m2_t * op0, vuint8m2_t * op1, const uint8_t * op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_u8m2_m))) +void vloxseg2ei16(vuint8m2_t * op0, vuint8m2_t * op1, vbool4_t op2, vuint8m2_t op3, vuint8m2_t op4, const uint8_t * op5, vuint16m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_u8m4))) +void vloxseg2ei16(vuint8m4_t * op0, vuint8m4_t * op1, const uint8_t * op2, vuint16m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_u8m4_m))) +void vloxseg2ei16(vuint8m4_t * op0, vuint8m4_t * op1, vbool2_t op2, vuint8m4_t op3, vuint8m4_t op4, const uint8_t * op5, vuint16m8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_u8mf2))) +void vloxseg2ei16(vuint8mf2_t * op0, vuint8mf2_t * op1, const uint8_t * op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_u8mf2_m))) +void vloxseg2ei16(vuint8mf2_t * op0, vuint8mf2_t * op1, vbool16_t op2, vuint8mf2_t op3, vuint8mf2_t op4, const uint8_t * op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_u8mf4))) +void vloxseg2ei16(vuint8mf4_t * op0, vuint8mf4_t * op1, const uint8_t * op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_u8mf4_m))) +void vloxseg2ei16(vuint8mf4_t * op0, vuint8mf4_t * op1, vbool32_t op2, vuint8mf4_t op3, vuint8mf4_t op4, const uint8_t * op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_u8mf8))) +void vloxseg2ei16(vuint8mf8_t * op0, vuint8mf8_t * op1, const uint8_t * op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_u8mf8_m))) +void vloxseg2ei16(vuint8mf8_t * op0, vuint8mf8_t * op1, vbool64_t op2, vuint8mf8_t op3, vuint8mf8_t op4, const uint8_t * op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_i8m1))) +void vloxseg3ei16(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, const int8_t * op3, vuint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_i8m1_m))) +void vloxseg3ei16(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vbool8_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, const int8_t * op7, vuint16m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_i8m2))) +void vloxseg3ei16(vint8m2_t * op0, vint8m2_t * op1, vint8m2_t * op2, const int8_t * op3, vuint16m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_i8m2_m))) +void vloxseg3ei16(vint8m2_t * op0, vint8m2_t * op1, vint8m2_t * op2, vbool4_t op3, vint8m2_t op4, vint8m2_t op5, vint8m2_t op6, const int8_t * op7, vuint16m4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_i8mf2))) +void vloxseg3ei16(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, const int8_t * op3, vuint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_i8mf2_m))) +void vloxseg3ei16(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vbool16_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, const int8_t * op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_i8mf4))) +void vloxseg3ei16(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, const int8_t * op3, vuint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_i8mf4_m))) +void vloxseg3ei16(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vbool32_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, const int8_t * op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_i8mf8))) +void vloxseg3ei16(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, const int8_t * op3, vuint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_i8mf8_m))) +void vloxseg3ei16(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vbool64_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, const int8_t * op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_u8m1))) +void vloxseg3ei16(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, const uint8_t * op3, vuint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_u8m1_m))) +void vloxseg3ei16(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vbool8_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, const uint8_t * op7, vuint16m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_u8m2))) +void vloxseg3ei16(vuint8m2_t * op0, vuint8m2_t * op1, vuint8m2_t * op2, const uint8_t * op3, vuint16m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_u8m2_m))) +void vloxseg3ei16(vuint8m2_t * op0, vuint8m2_t * op1, vuint8m2_t * op2, vbool4_t op3, vuint8m2_t op4, vuint8m2_t op5, vuint8m2_t op6, const uint8_t * op7, vuint16m4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_u8mf2))) +void vloxseg3ei16(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, const uint8_t * op3, vuint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_u8mf2_m))) +void vloxseg3ei16(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vbool16_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, const uint8_t * op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_u8mf4))) +void vloxseg3ei16(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, const uint8_t * op3, vuint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_u8mf4_m))) +void vloxseg3ei16(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vbool32_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, const uint8_t * op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_u8mf8))) +void vloxseg3ei16(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, const uint8_t * op3, vuint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_u8mf8_m))) +void vloxseg3ei16(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vbool64_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, const uint8_t * op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_i8m1))) +void vloxseg4ei16(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, const int8_t * op4, vuint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_i8m1_m))) +void vloxseg4ei16(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vbool8_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, const int8_t * op9, vuint16m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_i8m2))) +void vloxseg4ei16(vint8m2_t * op0, vint8m2_t * op1, vint8m2_t * op2, vint8m2_t * op3, const int8_t * op4, vuint16m4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_i8m2_m))) +void vloxseg4ei16(vint8m2_t * op0, vint8m2_t * op1, vint8m2_t * op2, vint8m2_t * op3, vbool4_t op4, vint8m2_t op5, vint8m2_t op6, vint8m2_t op7, vint8m2_t op8, const int8_t * op9, vuint16m4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_i8mf2))) +void vloxseg4ei16(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, const int8_t * op4, vuint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_i8mf2_m))) +void vloxseg4ei16(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vbool16_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, const int8_t * op9, vuint16m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_i8mf4))) +void vloxseg4ei16(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, const int8_t * op4, vuint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_i8mf4_m))) +void vloxseg4ei16(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vbool32_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, const int8_t * op9, vuint16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_i8mf8))) +void vloxseg4ei16(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, const int8_t * op4, vuint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_i8mf8_m))) +void vloxseg4ei16(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vbool64_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, const int8_t * op9, vuint16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_u8m1))) +void vloxseg4ei16(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, const uint8_t * op4, vuint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_u8m1_m))) +void vloxseg4ei16(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vbool8_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, const uint8_t * op9, vuint16m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_u8m2))) +void vloxseg4ei16(vuint8m2_t * op0, vuint8m2_t * op1, vuint8m2_t * op2, vuint8m2_t * op3, const uint8_t * op4, vuint16m4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_u8m2_m))) +void vloxseg4ei16(vuint8m2_t * op0, vuint8m2_t * op1, vuint8m2_t * op2, vuint8m2_t * op3, vbool4_t op4, vuint8m2_t op5, vuint8m2_t op6, vuint8m2_t op7, vuint8m2_t op8, const uint8_t * op9, vuint16m4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_u8mf2))) +void vloxseg4ei16(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, const uint8_t * op4, vuint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_u8mf2_m))) +void vloxseg4ei16(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vbool16_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, const uint8_t * op9, vuint16m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_u8mf4))) +void vloxseg4ei16(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, const uint8_t * op4, vuint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_u8mf4_m))) +void vloxseg4ei16(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vbool32_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, const uint8_t * op9, vuint16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_u8mf8))) +void vloxseg4ei16(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, const uint8_t * op4, vuint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_u8mf8_m))) +void vloxseg4ei16(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vbool64_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, const uint8_t * op9, vuint16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_i8m1))) +void vloxseg5ei16(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, const int8_t * op5, vuint16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_i8m1_m))) +void vloxseg5ei16(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vbool8_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, vint8m1_t op10, const int8_t * op11, vuint16m2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_i8mf2))) +void vloxseg5ei16(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, const int8_t * op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_i8mf2_m))) +void vloxseg5ei16(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vbool16_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, vint8mf2_t op10, const int8_t * op11, vuint16m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_i8mf4))) +void vloxseg5ei16(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, const int8_t * op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_i8mf4_m))) +void vloxseg5ei16(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vbool32_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, vint8mf4_t op10, const int8_t * op11, vuint16mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_i8mf8))) +void vloxseg5ei16(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, const int8_t * op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_i8mf8_m))) +void vloxseg5ei16(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vbool64_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, vint8mf8_t op10, const int8_t * op11, vuint16mf4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_u8m1))) +void vloxseg5ei16(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, const uint8_t * op5, vuint16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_u8m1_m))) +void vloxseg5ei16(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vbool8_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, vuint8m1_t op10, const uint8_t * op11, vuint16m2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_u8mf2))) +void vloxseg5ei16(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, const uint8_t * op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_u8mf2_m))) +void vloxseg5ei16(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vbool16_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, vuint8mf2_t op10, const uint8_t * op11, vuint16m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_u8mf4))) +void vloxseg5ei16(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, const uint8_t * op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_u8mf4_m))) +void vloxseg5ei16(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vbool32_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, vuint8mf4_t op10, const uint8_t * op11, vuint16mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_u8mf8))) +void vloxseg5ei16(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, const uint8_t * op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_u8mf8_m))) +void vloxseg5ei16(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vbool64_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, vuint8mf8_t op10, const uint8_t * op11, vuint16mf4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_i8m1))) +void vloxseg6ei16(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, const int8_t * op6, vuint16m2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_i8m1_m))) +void vloxseg6ei16(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vbool8_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, vint8m1_t op10, vint8m1_t op11, vint8m1_t op12, const int8_t * op13, vuint16m2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_i8mf2))) +void vloxseg6ei16(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, const int8_t * op6, vuint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_i8mf2_m))) +void vloxseg6ei16(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vbool16_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, vint8mf2_t op10, vint8mf2_t op11, vint8mf2_t op12, const int8_t * op13, vuint16m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_i8mf4))) +void vloxseg6ei16(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, const int8_t * op6, vuint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_i8mf4_m))) +void vloxseg6ei16(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vbool32_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, vint8mf4_t op10, vint8mf4_t op11, vint8mf4_t op12, const int8_t * op13, vuint16mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_i8mf8))) +void vloxseg6ei16(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, const int8_t * op6, vuint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_i8mf8_m))) +void vloxseg6ei16(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vbool64_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, vint8mf8_t op10, vint8mf8_t op11, vint8mf8_t op12, const int8_t * op13, vuint16mf4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_u8m1))) +void vloxseg6ei16(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, const uint8_t * op6, vuint16m2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_u8m1_m))) +void vloxseg6ei16(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vbool8_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, vuint8m1_t op10, vuint8m1_t op11, vuint8m1_t op12, const uint8_t * op13, vuint16m2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_u8mf2))) +void vloxseg6ei16(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, const uint8_t * op6, vuint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_u8mf2_m))) +void vloxseg6ei16(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vbool16_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, vuint8mf2_t op10, vuint8mf2_t op11, vuint8mf2_t op12, const uint8_t * op13, vuint16m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_u8mf4))) +void vloxseg6ei16(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, const uint8_t * op6, vuint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_u8mf4_m))) +void vloxseg6ei16(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vbool32_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, vuint8mf4_t op10, vuint8mf4_t op11, vuint8mf4_t op12, const uint8_t * op13, vuint16mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_u8mf8))) +void vloxseg6ei16(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, const uint8_t * op6, vuint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_u8mf8_m))) +void vloxseg6ei16(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vbool64_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, vuint8mf8_t op10, vuint8mf8_t op11, vuint8mf8_t op12, const uint8_t * op13, vuint16mf4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_i8m1))) +void vloxseg7ei16(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vint8m1_t * op6, const int8_t * op7, vuint16m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_i8m1_m))) +void vloxseg7ei16(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vint8m1_t * op6, vbool8_t op7, vint8m1_t op8, vint8m1_t op9, vint8m1_t op10, vint8m1_t op11, vint8m1_t op12, vint8m1_t op13, vint8m1_t op14, const int8_t * op15, vuint16m2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_i8mf2))) +void vloxseg7ei16(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vint8mf2_t * op6, const int8_t * op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_i8mf2_m))) +void vloxseg7ei16(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vint8mf2_t * op6, vbool16_t op7, vint8mf2_t op8, vint8mf2_t op9, vint8mf2_t op10, vint8mf2_t op11, vint8mf2_t op12, vint8mf2_t op13, vint8mf2_t op14, const int8_t * op15, vuint16m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_i8mf4))) +void vloxseg7ei16(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vint8mf4_t * op6, const int8_t * op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_i8mf4_m))) +void vloxseg7ei16(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vint8mf4_t * op6, vbool32_t op7, vint8mf4_t op8, vint8mf4_t op9, vint8mf4_t op10, vint8mf4_t op11, vint8mf4_t op12, vint8mf4_t op13, vint8mf4_t op14, const int8_t * op15, vuint16mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_i8mf8))) +void vloxseg7ei16(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vint8mf8_t * op6, const int8_t * op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_i8mf8_m))) +void vloxseg7ei16(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vint8mf8_t * op6, vbool64_t op7, vint8mf8_t op8, vint8mf8_t op9, vint8mf8_t op10, vint8mf8_t op11, vint8mf8_t op12, vint8mf8_t op13, vint8mf8_t op14, const int8_t * op15, vuint16mf4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_u8m1))) +void vloxseg7ei16(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vuint8m1_t * op6, const uint8_t * op7, vuint16m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_u8m1_m))) +void vloxseg7ei16(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vuint8m1_t * op6, vbool8_t op7, vuint8m1_t op8, vuint8m1_t op9, vuint8m1_t op10, vuint8m1_t op11, vuint8m1_t op12, vuint8m1_t op13, vuint8m1_t op14, const uint8_t * op15, vuint16m2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_u8mf2))) +void vloxseg7ei16(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vuint8mf2_t * op6, const uint8_t * op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_u8mf2_m))) +void vloxseg7ei16(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vuint8mf2_t * op6, vbool16_t op7, vuint8mf2_t op8, vuint8mf2_t op9, vuint8mf2_t op10, vuint8mf2_t op11, vuint8mf2_t op12, vuint8mf2_t op13, vuint8mf2_t op14, const uint8_t * op15, vuint16m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_u8mf4))) +void vloxseg7ei16(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vuint8mf4_t * op6, const uint8_t * op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_u8mf4_m))) +void vloxseg7ei16(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vuint8mf4_t * op6, vbool32_t op7, vuint8mf4_t op8, vuint8mf4_t op9, vuint8mf4_t op10, vuint8mf4_t op11, vuint8mf4_t op12, vuint8mf4_t op13, vuint8mf4_t op14, const uint8_t * op15, vuint16mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_u8mf8))) +void vloxseg7ei16(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vuint8mf8_t * op6, const uint8_t * op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_u8mf8_m))) +void vloxseg7ei16(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vuint8mf8_t * op6, vbool64_t op7, vuint8mf8_t op8, vuint8mf8_t op9, vuint8mf8_t op10, vuint8mf8_t op11, vuint8mf8_t op12, vuint8mf8_t op13, vuint8mf8_t op14, const uint8_t * op15, vuint16mf4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_i8m1))) +void vloxseg8ei16(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vint8m1_t * op6, vint8m1_t * op7, const int8_t * op8, vuint16m2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_i8m1_m))) +void vloxseg8ei16(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vint8m1_t * op6, vint8m1_t * op7, vbool8_t op8, vint8m1_t op9, vint8m1_t op10, vint8m1_t op11, vint8m1_t op12, vint8m1_t op13, vint8m1_t op14, vint8m1_t op15, vint8m1_t op16, const int8_t * op17, vuint16m2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_i8mf2))) +void vloxseg8ei16(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vint8mf2_t * op6, vint8mf2_t * op7, const int8_t * op8, vuint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_i8mf2_m))) +void vloxseg8ei16(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vint8mf2_t * op6, vint8mf2_t * op7, vbool16_t op8, vint8mf2_t op9, vint8mf2_t op10, vint8mf2_t op11, vint8mf2_t op12, vint8mf2_t op13, vint8mf2_t op14, vint8mf2_t op15, vint8mf2_t op16, const int8_t * op17, vuint16m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_i8mf4))) +void vloxseg8ei16(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vint8mf4_t * op6, vint8mf4_t * op7, const int8_t * op8, vuint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_i8mf4_m))) +void vloxseg8ei16(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vint8mf4_t * op6, vint8mf4_t * op7, vbool32_t op8, vint8mf4_t op9, vint8mf4_t op10, vint8mf4_t op11, vint8mf4_t op12, vint8mf4_t op13, vint8mf4_t op14, vint8mf4_t op15, vint8mf4_t op16, const int8_t * op17, vuint16mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_i8mf8))) +void vloxseg8ei16(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vint8mf8_t * op6, vint8mf8_t * op7, const int8_t * op8, vuint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_i8mf8_m))) +void vloxseg8ei16(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vint8mf8_t * op6, vint8mf8_t * op7, vbool64_t op8, vint8mf8_t op9, vint8mf8_t op10, vint8mf8_t op11, vint8mf8_t op12, vint8mf8_t op13, vint8mf8_t op14, vint8mf8_t op15, vint8mf8_t op16, const int8_t * op17, vuint16mf4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_u8m1))) +void vloxseg8ei16(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vuint8m1_t * op6, vuint8m1_t * op7, const uint8_t * op8, vuint16m2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_u8m1_m))) +void vloxseg8ei16(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vuint8m1_t * op6, vuint8m1_t * op7, vbool8_t op8, vuint8m1_t op9, vuint8m1_t op10, vuint8m1_t op11, vuint8m1_t op12, vuint8m1_t op13, vuint8m1_t op14, vuint8m1_t op15, vuint8m1_t op16, const uint8_t * op17, vuint16m2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_u8mf2))) +void vloxseg8ei16(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vuint8mf2_t * op6, vuint8mf2_t * op7, const uint8_t * op8, vuint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_u8mf2_m))) +void vloxseg8ei16(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vuint8mf2_t * op6, vuint8mf2_t * op7, vbool16_t op8, vuint8mf2_t op9, vuint8mf2_t op10, vuint8mf2_t op11, vuint8mf2_t op12, vuint8mf2_t op13, vuint8mf2_t op14, vuint8mf2_t op15, vuint8mf2_t op16, const uint8_t * op17, vuint16m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_u8mf4))) +void vloxseg8ei16(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vuint8mf4_t * op6, vuint8mf4_t * op7, const uint8_t * op8, vuint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_u8mf4_m))) +void vloxseg8ei16(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vuint8mf4_t * op6, vuint8mf4_t * op7, vbool32_t op8, vuint8mf4_t op9, vuint8mf4_t op10, vuint8mf4_t op11, vuint8mf4_t op12, vuint8mf4_t op13, vuint8mf4_t op14, vuint8mf4_t op15, vuint8mf4_t op16, const uint8_t * op17, vuint16mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_u8mf8))) +void vloxseg8ei16(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vuint8mf8_t * op6, vuint8mf8_t * op7, const uint8_t * op8, vuint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_u8mf8_m))) +void vloxseg8ei16(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vuint8mf8_t * op6, vuint8mf8_t * op7, vbool64_t op8, vuint8mf8_t op9, vuint8mf8_t op10, vuint8mf8_t op11, vuint8mf8_t op12, vuint8mf8_t op13, vuint8mf8_t op14, vuint8mf8_t op15, vuint8mf8_t op16, const uint8_t * op17, vuint16mf4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_i8m1))) +void vloxseg2ei32(vint8m1_t * op0, vint8m1_t * op1, const int8_t * op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_i8m1_m))) +void vloxseg2ei32(vint8m1_t * op0, vint8m1_t * op1, vbool8_t op2, vint8m1_t op3, vint8m1_t op4, const int8_t * op5, vuint32m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_i8m2))) +void vloxseg2ei32(vint8m2_t * op0, vint8m2_t * op1, const int8_t * op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_i8m2_m))) +void vloxseg2ei32(vint8m2_t * op0, vint8m2_t * op1, vbool4_t op2, vint8m2_t op3, vint8m2_t op4, const int8_t * op5, vuint32m8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_i8mf2))) +void vloxseg2ei32(vint8mf2_t * op0, vint8mf2_t * op1, const int8_t * op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_i8mf2_m))) +void vloxseg2ei32(vint8mf2_t * op0, vint8mf2_t * op1, vbool16_t op2, vint8mf2_t op3, vint8mf2_t op4, const int8_t * op5, vuint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_i8mf4))) +void vloxseg2ei32(vint8mf4_t * op0, vint8mf4_t * op1, const int8_t * op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_i8mf4_m))) +void vloxseg2ei32(vint8mf4_t * op0, vint8mf4_t * op1, vbool32_t op2, vint8mf4_t op3, vint8mf4_t op4, const int8_t * op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_i8mf8))) +void vloxseg2ei32(vint8mf8_t * op0, vint8mf8_t * op1, const int8_t * op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_i8mf8_m))) +void vloxseg2ei32(vint8mf8_t * op0, vint8mf8_t * op1, vbool64_t op2, vint8mf8_t op3, vint8mf8_t op4, const int8_t * op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_u8m1))) +void vloxseg2ei32(vuint8m1_t * op0, vuint8m1_t * op1, const uint8_t * op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_u8m1_m))) +void vloxseg2ei32(vuint8m1_t * op0, vuint8m1_t * op1, vbool8_t op2, vuint8m1_t op3, vuint8m1_t op4, const uint8_t * op5, vuint32m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_u8m2))) +void vloxseg2ei32(vuint8m2_t * op0, vuint8m2_t * op1, const uint8_t * op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_u8m2_m))) +void vloxseg2ei32(vuint8m2_t * op0, vuint8m2_t * op1, vbool4_t op2, vuint8m2_t op3, vuint8m2_t op4, const uint8_t * op5, vuint32m8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_u8mf2))) +void vloxseg2ei32(vuint8mf2_t * op0, vuint8mf2_t * op1, const uint8_t * op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_u8mf2_m))) +void vloxseg2ei32(vuint8mf2_t * op0, vuint8mf2_t * op1, vbool16_t op2, vuint8mf2_t op3, vuint8mf2_t op4, const uint8_t * op5, vuint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_u8mf4))) +void vloxseg2ei32(vuint8mf4_t * op0, vuint8mf4_t * op1, const uint8_t * op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_u8mf4_m))) +void vloxseg2ei32(vuint8mf4_t * op0, vuint8mf4_t * op1, vbool32_t op2, vuint8mf4_t op3, vuint8mf4_t op4, const uint8_t * op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_u8mf8))) +void vloxseg2ei32(vuint8mf8_t * op0, vuint8mf8_t * op1, const uint8_t * op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_u8mf8_m))) +void vloxseg2ei32(vuint8mf8_t * op0, vuint8mf8_t * op1, vbool64_t op2, vuint8mf8_t op3, vuint8mf8_t op4, const uint8_t * op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_i8m1))) +void vloxseg3ei32(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, const int8_t * op3, vuint32m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_i8m1_m))) +void vloxseg3ei32(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vbool8_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, const int8_t * op7, vuint32m4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_i8m2))) +void vloxseg3ei32(vint8m2_t * op0, vint8m2_t * op1, vint8m2_t * op2, const int8_t * op3, vuint32m8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_i8m2_m))) +void vloxseg3ei32(vint8m2_t * op0, vint8m2_t * op1, vint8m2_t * op2, vbool4_t op3, vint8m2_t op4, vint8m2_t op5, vint8m2_t op6, const int8_t * op7, vuint32m8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_i8mf2))) +void vloxseg3ei32(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, const int8_t * op3, vuint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_i8mf2_m))) +void vloxseg3ei32(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vbool16_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, const int8_t * op7, vuint32m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_i8mf4))) +void vloxseg3ei32(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, const int8_t * op3, vuint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_i8mf4_m))) +void vloxseg3ei32(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vbool32_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, const int8_t * op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_i8mf8))) +void vloxseg3ei32(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, const int8_t * op3, vuint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_i8mf8_m))) +void vloxseg3ei32(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vbool64_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, const int8_t * op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_u8m1))) +void vloxseg3ei32(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, const uint8_t * op3, vuint32m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_u8m1_m))) +void vloxseg3ei32(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vbool8_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, const uint8_t * op7, vuint32m4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_u8m2))) +void vloxseg3ei32(vuint8m2_t * op0, vuint8m2_t * op1, vuint8m2_t * op2, const uint8_t * op3, vuint32m8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_u8m2_m))) +void vloxseg3ei32(vuint8m2_t * op0, vuint8m2_t * op1, vuint8m2_t * op2, vbool4_t op3, vuint8m2_t op4, vuint8m2_t op5, vuint8m2_t op6, const uint8_t * op7, vuint32m8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_u8mf2))) +void vloxseg3ei32(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, const uint8_t * op3, vuint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_u8mf2_m))) +void vloxseg3ei32(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vbool16_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, const uint8_t * op7, vuint32m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_u8mf4))) +void vloxseg3ei32(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, const uint8_t * op3, vuint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_u8mf4_m))) +void vloxseg3ei32(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vbool32_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, const uint8_t * op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_u8mf8))) +void vloxseg3ei32(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, const uint8_t * op3, vuint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_u8mf8_m))) +void vloxseg3ei32(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vbool64_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, const uint8_t * op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_i8m1))) +void vloxseg4ei32(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, const int8_t * op4, vuint32m4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_i8m1_m))) +void vloxseg4ei32(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vbool8_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, const int8_t * op9, vuint32m4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_i8m2))) +void vloxseg4ei32(vint8m2_t * op0, vint8m2_t * op1, vint8m2_t * op2, vint8m2_t * op3, const int8_t * op4, vuint32m8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_i8m2_m))) +void vloxseg4ei32(vint8m2_t * op0, vint8m2_t * op1, vint8m2_t * op2, vint8m2_t * op3, vbool4_t op4, vint8m2_t op5, vint8m2_t op6, vint8m2_t op7, vint8m2_t op8, const int8_t * op9, vuint32m8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_i8mf2))) +void vloxseg4ei32(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, const int8_t * op4, vuint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_i8mf2_m))) +void vloxseg4ei32(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vbool16_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, const int8_t * op9, vuint32m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_i8mf4))) +void vloxseg4ei32(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, const int8_t * op4, vuint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_i8mf4_m))) +void vloxseg4ei32(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vbool32_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, const int8_t * op9, vuint32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_i8mf8))) +void vloxseg4ei32(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, const int8_t * op4, vuint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_i8mf8_m))) +void vloxseg4ei32(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vbool64_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, const int8_t * op9, vuint32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_u8m1))) +void vloxseg4ei32(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, const uint8_t * op4, vuint32m4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_u8m1_m))) +void vloxseg4ei32(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vbool8_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, const uint8_t * op9, vuint32m4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_u8m2))) +void vloxseg4ei32(vuint8m2_t * op0, vuint8m2_t * op1, vuint8m2_t * op2, vuint8m2_t * op3, const uint8_t * op4, vuint32m8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_u8m2_m))) +void vloxseg4ei32(vuint8m2_t * op0, vuint8m2_t * op1, vuint8m2_t * op2, vuint8m2_t * op3, vbool4_t op4, vuint8m2_t op5, vuint8m2_t op6, vuint8m2_t op7, vuint8m2_t op8, const uint8_t * op9, vuint32m8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_u8mf2))) +void vloxseg4ei32(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, const uint8_t * op4, vuint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_u8mf2_m))) +void vloxseg4ei32(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vbool16_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, const uint8_t * op9, vuint32m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_u8mf4))) +void vloxseg4ei32(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, const uint8_t * op4, vuint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_u8mf4_m))) +void vloxseg4ei32(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vbool32_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, const uint8_t * op9, vuint32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_u8mf8))) +void vloxseg4ei32(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, const uint8_t * op4, vuint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_u8mf8_m))) +void vloxseg4ei32(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vbool64_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, const uint8_t * op9, vuint32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_i8m1))) +void vloxseg5ei32(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, const int8_t * op5, vuint32m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_i8m1_m))) +void vloxseg5ei32(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vbool8_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, vint8m1_t op10, const int8_t * op11, vuint32m4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_i8mf2))) +void vloxseg5ei32(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, const int8_t * op5, vuint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_i8mf2_m))) +void vloxseg5ei32(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vbool16_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, vint8mf2_t op10, const int8_t * op11, vuint32m2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_i8mf4))) +void vloxseg5ei32(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, const int8_t * op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_i8mf4_m))) +void vloxseg5ei32(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vbool32_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, vint8mf4_t op10, const int8_t * op11, vuint32m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_i8mf8))) +void vloxseg5ei32(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, const int8_t * op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_i8mf8_m))) +void vloxseg5ei32(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vbool64_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, vint8mf8_t op10, const int8_t * op11, vuint32mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_u8m1))) +void vloxseg5ei32(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, const uint8_t * op5, vuint32m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_u8m1_m))) +void vloxseg5ei32(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vbool8_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, vuint8m1_t op10, const uint8_t * op11, vuint32m4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_u8mf2))) +void vloxseg5ei32(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, const uint8_t * op5, vuint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_u8mf2_m))) +void vloxseg5ei32(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vbool16_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, vuint8mf2_t op10, const uint8_t * op11, vuint32m2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_u8mf4))) +void vloxseg5ei32(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, const uint8_t * op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_u8mf4_m))) +void vloxseg5ei32(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vbool32_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, vuint8mf4_t op10, const uint8_t * op11, vuint32m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_u8mf8))) +void vloxseg5ei32(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, const uint8_t * op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_u8mf8_m))) +void vloxseg5ei32(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vbool64_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, vuint8mf8_t op10, const uint8_t * op11, vuint32mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_i8m1))) +void vloxseg6ei32(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, const int8_t * op6, vuint32m4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_i8m1_m))) +void vloxseg6ei32(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vbool8_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, vint8m1_t op10, vint8m1_t op11, vint8m1_t op12, const int8_t * op13, vuint32m4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_i8mf2))) +void vloxseg6ei32(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, const int8_t * op6, vuint32m2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_i8mf2_m))) +void vloxseg6ei32(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vbool16_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, vint8mf2_t op10, vint8mf2_t op11, vint8mf2_t op12, const int8_t * op13, vuint32m2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_i8mf4))) +void vloxseg6ei32(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, const int8_t * op6, vuint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_i8mf4_m))) +void vloxseg6ei32(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vbool32_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, vint8mf4_t op10, vint8mf4_t op11, vint8mf4_t op12, const int8_t * op13, vuint32m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_i8mf8))) +void vloxseg6ei32(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, const int8_t * op6, vuint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_i8mf8_m))) +void vloxseg6ei32(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vbool64_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, vint8mf8_t op10, vint8mf8_t op11, vint8mf8_t op12, const int8_t * op13, vuint32mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_u8m1))) +void vloxseg6ei32(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, const uint8_t * op6, vuint32m4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_u8m1_m))) +void vloxseg6ei32(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vbool8_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, vuint8m1_t op10, vuint8m1_t op11, vuint8m1_t op12, const uint8_t * op13, vuint32m4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_u8mf2))) +void vloxseg6ei32(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, const uint8_t * op6, vuint32m2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_u8mf2_m))) +void vloxseg6ei32(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vbool16_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, vuint8mf2_t op10, vuint8mf2_t op11, vuint8mf2_t op12, const uint8_t * op13, vuint32m2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_u8mf4))) +void vloxseg6ei32(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, const uint8_t * op6, vuint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_u8mf4_m))) +void vloxseg6ei32(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vbool32_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, vuint8mf4_t op10, vuint8mf4_t op11, vuint8mf4_t op12, const uint8_t * op13, vuint32m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_u8mf8))) +void vloxseg6ei32(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, const uint8_t * op6, vuint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_u8mf8_m))) +void vloxseg6ei32(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vbool64_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, vuint8mf8_t op10, vuint8mf8_t op11, vuint8mf8_t op12, const uint8_t * op13, vuint32mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_i8m1))) +void vloxseg7ei32(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vint8m1_t * op6, const int8_t * op7, vuint32m4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_i8m1_m))) +void vloxseg7ei32(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vint8m1_t * op6, vbool8_t op7, vint8m1_t op8, vint8m1_t op9, vint8m1_t op10, vint8m1_t op11, vint8m1_t op12, vint8m1_t op13, vint8m1_t op14, const int8_t * op15, vuint32m4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_i8mf2))) +void vloxseg7ei32(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vint8mf2_t * op6, const int8_t * op7, vuint32m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_i8mf2_m))) +void vloxseg7ei32(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vint8mf2_t * op6, vbool16_t op7, vint8mf2_t op8, vint8mf2_t op9, vint8mf2_t op10, vint8mf2_t op11, vint8mf2_t op12, vint8mf2_t op13, vint8mf2_t op14, const int8_t * op15, vuint32m2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_i8mf4))) +void vloxseg7ei32(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vint8mf4_t * op6, const int8_t * op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_i8mf4_m))) +void vloxseg7ei32(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vint8mf4_t * op6, vbool32_t op7, vint8mf4_t op8, vint8mf4_t op9, vint8mf4_t op10, vint8mf4_t op11, vint8mf4_t op12, vint8mf4_t op13, vint8mf4_t op14, const int8_t * op15, vuint32m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_i8mf8))) +void vloxseg7ei32(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vint8mf8_t * op6, const int8_t * op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_i8mf8_m))) +void vloxseg7ei32(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vint8mf8_t * op6, vbool64_t op7, vint8mf8_t op8, vint8mf8_t op9, vint8mf8_t op10, vint8mf8_t op11, vint8mf8_t op12, vint8mf8_t op13, vint8mf8_t op14, const int8_t * op15, vuint32mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_u8m1))) +void vloxseg7ei32(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vuint8m1_t * op6, const uint8_t * op7, vuint32m4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_u8m1_m))) +void vloxseg7ei32(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vuint8m1_t * op6, vbool8_t op7, vuint8m1_t op8, vuint8m1_t op9, vuint8m1_t op10, vuint8m1_t op11, vuint8m1_t op12, vuint8m1_t op13, vuint8m1_t op14, const uint8_t * op15, vuint32m4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_u8mf2))) +void vloxseg7ei32(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vuint8mf2_t * op6, const uint8_t * op7, vuint32m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_u8mf2_m))) +void vloxseg7ei32(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vuint8mf2_t * op6, vbool16_t op7, vuint8mf2_t op8, vuint8mf2_t op9, vuint8mf2_t op10, vuint8mf2_t op11, vuint8mf2_t op12, vuint8mf2_t op13, vuint8mf2_t op14, const uint8_t * op15, vuint32m2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_u8mf4))) +void vloxseg7ei32(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vuint8mf4_t * op6, const uint8_t * op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_u8mf4_m))) +void vloxseg7ei32(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vuint8mf4_t * op6, vbool32_t op7, vuint8mf4_t op8, vuint8mf4_t op9, vuint8mf4_t op10, vuint8mf4_t op11, vuint8mf4_t op12, vuint8mf4_t op13, vuint8mf4_t op14, const uint8_t * op15, vuint32m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_u8mf8))) +void vloxseg7ei32(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vuint8mf8_t * op6, const uint8_t * op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_u8mf8_m))) +void vloxseg7ei32(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vuint8mf8_t * op6, vbool64_t op7, vuint8mf8_t op8, vuint8mf8_t op9, vuint8mf8_t op10, vuint8mf8_t op11, vuint8mf8_t op12, vuint8mf8_t op13, vuint8mf8_t op14, const uint8_t * op15, vuint32mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_i8m1))) +void vloxseg8ei32(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vint8m1_t * op6, vint8m1_t * op7, const int8_t * op8, vuint32m4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_i8m1_m))) +void vloxseg8ei32(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vint8m1_t * op6, vint8m1_t * op7, vbool8_t op8, vint8m1_t op9, vint8m1_t op10, vint8m1_t op11, vint8m1_t op12, vint8m1_t op13, vint8m1_t op14, vint8m1_t op15, vint8m1_t op16, const int8_t * op17, vuint32m4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_i8mf2))) +void vloxseg8ei32(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vint8mf2_t * op6, vint8mf2_t * op7, const int8_t * op8, vuint32m2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_i8mf2_m))) +void vloxseg8ei32(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vint8mf2_t * op6, vint8mf2_t * op7, vbool16_t op8, vint8mf2_t op9, vint8mf2_t op10, vint8mf2_t op11, vint8mf2_t op12, vint8mf2_t op13, vint8mf2_t op14, vint8mf2_t op15, vint8mf2_t op16, const int8_t * op17, vuint32m2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_i8mf4))) +void vloxseg8ei32(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vint8mf4_t * op6, vint8mf4_t * op7, const int8_t * op8, vuint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_i8mf4_m))) +void vloxseg8ei32(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vint8mf4_t * op6, vint8mf4_t * op7, vbool32_t op8, vint8mf4_t op9, vint8mf4_t op10, vint8mf4_t op11, vint8mf4_t op12, vint8mf4_t op13, vint8mf4_t op14, vint8mf4_t op15, vint8mf4_t op16, const int8_t * op17, vuint32m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_i8mf8))) +void vloxseg8ei32(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vint8mf8_t * op6, vint8mf8_t * op7, const int8_t * op8, vuint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_i8mf8_m))) +void vloxseg8ei32(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vint8mf8_t * op6, vint8mf8_t * op7, vbool64_t op8, vint8mf8_t op9, vint8mf8_t op10, vint8mf8_t op11, vint8mf8_t op12, vint8mf8_t op13, vint8mf8_t op14, vint8mf8_t op15, vint8mf8_t op16, const int8_t * op17, vuint32mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_u8m1))) +void vloxseg8ei32(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vuint8m1_t * op6, vuint8m1_t * op7, const uint8_t * op8, vuint32m4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_u8m1_m))) +void vloxseg8ei32(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vuint8m1_t * op6, vuint8m1_t * op7, vbool8_t op8, vuint8m1_t op9, vuint8m1_t op10, vuint8m1_t op11, vuint8m1_t op12, vuint8m1_t op13, vuint8m1_t op14, vuint8m1_t op15, vuint8m1_t op16, const uint8_t * op17, vuint32m4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_u8mf2))) +void vloxseg8ei32(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vuint8mf2_t * op6, vuint8mf2_t * op7, const uint8_t * op8, vuint32m2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_u8mf2_m))) +void vloxseg8ei32(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vuint8mf2_t * op6, vuint8mf2_t * op7, vbool16_t op8, vuint8mf2_t op9, vuint8mf2_t op10, vuint8mf2_t op11, vuint8mf2_t op12, vuint8mf2_t op13, vuint8mf2_t op14, vuint8mf2_t op15, vuint8mf2_t op16, const uint8_t * op17, vuint32m2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_u8mf4))) +void vloxseg8ei32(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vuint8mf4_t * op6, vuint8mf4_t * op7, const uint8_t * op8, vuint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_u8mf4_m))) +void vloxseg8ei32(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vuint8mf4_t * op6, vuint8mf4_t * op7, vbool32_t op8, vuint8mf4_t op9, vuint8mf4_t op10, vuint8mf4_t op11, vuint8mf4_t op12, vuint8mf4_t op13, vuint8mf4_t op14, vuint8mf4_t op15, vuint8mf4_t op16, const uint8_t * op17, vuint32m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_u8mf8))) +void vloxseg8ei32(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vuint8mf8_t * op6, vuint8mf8_t * op7, const uint8_t * op8, vuint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_u8mf8_m))) +void vloxseg8ei32(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vuint8mf8_t * op6, vuint8mf8_t * op7, vbool64_t op8, vuint8mf8_t op9, vuint8mf8_t op10, vuint8mf8_t op11, vuint8mf8_t op12, vuint8mf8_t op13, vuint8mf8_t op14, vuint8mf8_t op15, vuint8mf8_t op16, const uint8_t * op17, vuint32mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_i8m1))) +void vloxseg2ei64(vint8m1_t * op0, vint8m1_t * op1, const int8_t * op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_i8m1_m))) +void vloxseg2ei64(vint8m1_t * op0, vint8m1_t * op1, vbool8_t op2, vint8m1_t op3, vint8m1_t op4, const int8_t * op5, vuint64m8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_i8mf2))) +void vloxseg2ei64(vint8mf2_t * op0, vint8mf2_t * op1, const int8_t * op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_i8mf2_m))) +void vloxseg2ei64(vint8mf2_t * op0, vint8mf2_t * op1, vbool16_t op2, vint8mf2_t op3, vint8mf2_t op4, const int8_t * op5, vuint64m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_i8mf4))) +void vloxseg2ei64(vint8mf4_t * op0, vint8mf4_t * op1, const int8_t * op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_i8mf4_m))) +void vloxseg2ei64(vint8mf4_t * op0, vint8mf4_t * op1, vbool32_t op2, vint8mf4_t op3, vint8mf4_t op4, const int8_t * op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_i8mf8))) +void vloxseg2ei64(vint8mf8_t * op0, vint8mf8_t * op1, const int8_t * op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_i8mf8_m))) +void vloxseg2ei64(vint8mf8_t * op0, vint8mf8_t * op1, vbool64_t op2, vint8mf8_t op3, vint8mf8_t op4, const int8_t * op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_u8m1))) +void vloxseg2ei64(vuint8m1_t * op0, vuint8m1_t * op1, const uint8_t * op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_u8m1_m))) +void vloxseg2ei64(vuint8m1_t * op0, vuint8m1_t * op1, vbool8_t op2, vuint8m1_t op3, vuint8m1_t op4, const uint8_t * op5, vuint64m8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_u8mf2))) +void vloxseg2ei64(vuint8mf2_t * op0, vuint8mf2_t * op1, const uint8_t * op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_u8mf2_m))) +void vloxseg2ei64(vuint8mf2_t * op0, vuint8mf2_t * op1, vbool16_t op2, vuint8mf2_t op3, vuint8mf2_t op4, const uint8_t * op5, vuint64m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_u8mf4))) +void vloxseg2ei64(vuint8mf4_t * op0, vuint8mf4_t * op1, const uint8_t * op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_u8mf4_m))) +void vloxseg2ei64(vuint8mf4_t * op0, vuint8mf4_t * op1, vbool32_t op2, vuint8mf4_t op3, vuint8mf4_t op4, const uint8_t * op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_u8mf8))) +void vloxseg2ei64(vuint8mf8_t * op0, vuint8mf8_t * op1, const uint8_t * op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_u8mf8_m))) +void vloxseg2ei64(vuint8mf8_t * op0, vuint8mf8_t * op1, vbool64_t op2, vuint8mf8_t op3, vuint8mf8_t op4, const uint8_t * op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_i8m1))) +void vloxseg3ei64(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, const int8_t * op3, vuint64m8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_i8m1_m))) +void vloxseg3ei64(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vbool8_t op3, vint8m1_t op4, vint8m1_t op5, vint8m1_t op6, const int8_t * op7, vuint64m8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_i8mf2))) +void vloxseg3ei64(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, const int8_t * op3, vuint64m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_i8mf2_m))) +void vloxseg3ei64(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vbool16_t op3, vint8mf2_t op4, vint8mf2_t op5, vint8mf2_t op6, const int8_t * op7, vuint64m4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_i8mf4))) +void vloxseg3ei64(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, const int8_t * op3, vuint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_i8mf4_m))) +void vloxseg3ei64(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vbool32_t op3, vint8mf4_t op4, vint8mf4_t op5, vint8mf4_t op6, const int8_t * op7, vuint64m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_i8mf8))) +void vloxseg3ei64(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, const int8_t * op3, vuint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_i8mf8_m))) +void vloxseg3ei64(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vbool64_t op3, vint8mf8_t op4, vint8mf8_t op5, vint8mf8_t op6, const int8_t * op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_u8m1))) +void vloxseg3ei64(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, const uint8_t * op3, vuint64m8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_u8m1_m))) +void vloxseg3ei64(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vbool8_t op3, vuint8m1_t op4, vuint8m1_t op5, vuint8m1_t op6, const uint8_t * op7, vuint64m8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_u8mf2))) +void vloxseg3ei64(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, const uint8_t * op3, vuint64m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_u8mf2_m))) +void vloxseg3ei64(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vbool16_t op3, vuint8mf2_t op4, vuint8mf2_t op5, vuint8mf2_t op6, const uint8_t * op7, vuint64m4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_u8mf4))) +void vloxseg3ei64(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, const uint8_t * op3, vuint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_u8mf4_m))) +void vloxseg3ei64(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vbool32_t op3, vuint8mf4_t op4, vuint8mf4_t op5, vuint8mf4_t op6, const uint8_t * op7, vuint64m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_u8mf8))) +void vloxseg3ei64(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, const uint8_t * op3, vuint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_u8mf8_m))) +void vloxseg3ei64(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vbool64_t op3, vuint8mf8_t op4, vuint8mf8_t op5, vuint8mf8_t op6, const uint8_t * op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_i8m1))) +void vloxseg4ei64(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, const int8_t * op4, vuint64m8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_i8m1_m))) +void vloxseg4ei64(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vbool8_t op4, vint8m1_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, const int8_t * op9, vuint64m8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_i8mf2))) +void vloxseg4ei64(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, const int8_t * op4, vuint64m4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_i8mf2_m))) +void vloxseg4ei64(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vbool16_t op4, vint8mf2_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, const int8_t * op9, vuint64m4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_i8mf4))) +void vloxseg4ei64(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, const int8_t * op4, vuint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_i8mf4_m))) +void vloxseg4ei64(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vbool32_t op4, vint8mf4_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, const int8_t * op9, vuint64m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_i8mf8))) +void vloxseg4ei64(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, const int8_t * op4, vuint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_i8mf8_m))) +void vloxseg4ei64(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vbool64_t op4, vint8mf8_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, const int8_t * op9, vuint64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_u8m1))) +void vloxseg4ei64(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, const uint8_t * op4, vuint64m8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_u8m1_m))) +void vloxseg4ei64(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vbool8_t op4, vuint8m1_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, const uint8_t * op9, vuint64m8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_u8mf2))) +void vloxseg4ei64(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, const uint8_t * op4, vuint64m4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_u8mf2_m))) +void vloxseg4ei64(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vbool16_t op4, vuint8mf2_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, const uint8_t * op9, vuint64m4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_u8mf4))) +void vloxseg4ei64(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, const uint8_t * op4, vuint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_u8mf4_m))) +void vloxseg4ei64(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vbool32_t op4, vuint8mf4_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, const uint8_t * op9, vuint64m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_u8mf8))) +void vloxseg4ei64(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, const uint8_t * op4, vuint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_u8mf8_m))) +void vloxseg4ei64(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vbool64_t op4, vuint8mf8_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, const uint8_t * op9, vuint64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_i8m1))) +void vloxseg5ei64(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, const int8_t * op5, vuint64m8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_i8m1_m))) +void vloxseg5ei64(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vbool8_t op5, vint8m1_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, vint8m1_t op10, const int8_t * op11, vuint64m8_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_i8mf2))) +void vloxseg5ei64(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, const int8_t * op5, vuint64m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_i8mf2_m))) +void vloxseg5ei64(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vbool16_t op5, vint8mf2_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, vint8mf2_t op10, const int8_t * op11, vuint64m4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_i8mf4))) +void vloxseg5ei64(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, const int8_t * op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_i8mf4_m))) +void vloxseg5ei64(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vbool32_t op5, vint8mf4_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, vint8mf4_t op10, const int8_t * op11, vuint64m2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_i8mf8))) +void vloxseg5ei64(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, const int8_t * op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_i8mf8_m))) +void vloxseg5ei64(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vbool64_t op5, vint8mf8_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, vint8mf8_t op10, const int8_t * op11, vuint64m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_u8m1))) +void vloxseg5ei64(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, const uint8_t * op5, vuint64m8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_u8m1_m))) +void vloxseg5ei64(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vbool8_t op5, vuint8m1_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, vuint8m1_t op10, const uint8_t * op11, vuint64m8_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_u8mf2))) +void vloxseg5ei64(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, const uint8_t * op5, vuint64m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_u8mf2_m))) +void vloxseg5ei64(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vbool16_t op5, vuint8mf2_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, vuint8mf2_t op10, const uint8_t * op11, vuint64m4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_u8mf4))) +void vloxseg5ei64(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, const uint8_t * op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_u8mf4_m))) +void vloxseg5ei64(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vbool32_t op5, vuint8mf4_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, vuint8mf4_t op10, const uint8_t * op11, vuint64m2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_u8mf8))) +void vloxseg5ei64(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, const uint8_t * op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_u8mf8_m))) +void vloxseg5ei64(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vbool64_t op5, vuint8mf8_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, vuint8mf8_t op10, const uint8_t * op11, vuint64m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_i8m1))) +void vloxseg6ei64(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, const int8_t * op6, vuint64m8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_i8m1_m))) +void vloxseg6ei64(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vbool8_t op6, vint8m1_t op7, vint8m1_t op8, vint8m1_t op9, vint8m1_t op10, vint8m1_t op11, vint8m1_t op12, const int8_t * op13, vuint64m8_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_i8mf2))) +void vloxseg6ei64(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, const int8_t * op6, vuint64m4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_i8mf2_m))) +void vloxseg6ei64(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vbool16_t op6, vint8mf2_t op7, vint8mf2_t op8, vint8mf2_t op9, vint8mf2_t op10, vint8mf2_t op11, vint8mf2_t op12, const int8_t * op13, vuint64m4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_i8mf4))) +void vloxseg6ei64(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, const int8_t * op6, vuint64m2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_i8mf4_m))) +void vloxseg6ei64(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vbool32_t op6, vint8mf4_t op7, vint8mf4_t op8, vint8mf4_t op9, vint8mf4_t op10, vint8mf4_t op11, vint8mf4_t op12, const int8_t * op13, vuint64m2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_i8mf8))) +void vloxseg6ei64(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, const int8_t * op6, vuint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_i8mf8_m))) +void vloxseg6ei64(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vbool64_t op6, vint8mf8_t op7, vint8mf8_t op8, vint8mf8_t op9, vint8mf8_t op10, vint8mf8_t op11, vint8mf8_t op12, const int8_t * op13, vuint64m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_u8m1))) +void vloxseg6ei64(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, const uint8_t * op6, vuint64m8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_u8m1_m))) +void vloxseg6ei64(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vbool8_t op6, vuint8m1_t op7, vuint8m1_t op8, vuint8m1_t op9, vuint8m1_t op10, vuint8m1_t op11, vuint8m1_t op12, const uint8_t * op13, vuint64m8_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_u8mf2))) +void vloxseg6ei64(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, const uint8_t * op6, vuint64m4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_u8mf2_m))) +void vloxseg6ei64(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vbool16_t op6, vuint8mf2_t op7, vuint8mf2_t op8, vuint8mf2_t op9, vuint8mf2_t op10, vuint8mf2_t op11, vuint8mf2_t op12, const uint8_t * op13, vuint64m4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_u8mf4))) +void vloxseg6ei64(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, const uint8_t * op6, vuint64m2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_u8mf4_m))) +void vloxseg6ei64(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vbool32_t op6, vuint8mf4_t op7, vuint8mf4_t op8, vuint8mf4_t op9, vuint8mf4_t op10, vuint8mf4_t op11, vuint8mf4_t op12, const uint8_t * op13, vuint64m2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_u8mf8))) +void vloxseg6ei64(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, const uint8_t * op6, vuint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_u8mf8_m))) +void vloxseg6ei64(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vbool64_t op6, vuint8mf8_t op7, vuint8mf8_t op8, vuint8mf8_t op9, vuint8mf8_t op10, vuint8mf8_t op11, vuint8mf8_t op12, const uint8_t * op13, vuint64m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_i8m1))) +void vloxseg7ei64(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vint8m1_t * op6, const int8_t * op7, vuint64m8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_i8m1_m))) +void vloxseg7ei64(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vint8m1_t * op6, vbool8_t op7, vint8m1_t op8, vint8m1_t op9, vint8m1_t op10, vint8m1_t op11, vint8m1_t op12, vint8m1_t op13, vint8m1_t op14, const int8_t * op15, vuint64m8_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_i8mf2))) +void vloxseg7ei64(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vint8mf2_t * op6, const int8_t * op7, vuint64m4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_i8mf2_m))) +void vloxseg7ei64(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vint8mf2_t * op6, vbool16_t op7, vint8mf2_t op8, vint8mf2_t op9, vint8mf2_t op10, vint8mf2_t op11, vint8mf2_t op12, vint8mf2_t op13, vint8mf2_t op14, const int8_t * op15, vuint64m4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_i8mf4))) +void vloxseg7ei64(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vint8mf4_t * op6, const int8_t * op7, vuint64m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_i8mf4_m))) +void vloxseg7ei64(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vint8mf4_t * op6, vbool32_t op7, vint8mf4_t op8, vint8mf4_t op9, vint8mf4_t op10, vint8mf4_t op11, vint8mf4_t op12, vint8mf4_t op13, vint8mf4_t op14, const int8_t * op15, vuint64m2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_i8mf8))) +void vloxseg7ei64(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vint8mf8_t * op6, const int8_t * op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_i8mf8_m))) +void vloxseg7ei64(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vint8mf8_t * op6, vbool64_t op7, vint8mf8_t op8, vint8mf8_t op9, vint8mf8_t op10, vint8mf8_t op11, vint8mf8_t op12, vint8mf8_t op13, vint8mf8_t op14, const int8_t * op15, vuint64m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_u8m1))) +void vloxseg7ei64(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vuint8m1_t * op6, const uint8_t * op7, vuint64m8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_u8m1_m))) +void vloxseg7ei64(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vuint8m1_t * op6, vbool8_t op7, vuint8m1_t op8, vuint8m1_t op9, vuint8m1_t op10, vuint8m1_t op11, vuint8m1_t op12, vuint8m1_t op13, vuint8m1_t op14, const uint8_t * op15, vuint64m8_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_u8mf2))) +void vloxseg7ei64(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vuint8mf2_t * op6, const uint8_t * op7, vuint64m4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_u8mf2_m))) +void vloxseg7ei64(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vuint8mf2_t * op6, vbool16_t op7, vuint8mf2_t op8, vuint8mf2_t op9, vuint8mf2_t op10, vuint8mf2_t op11, vuint8mf2_t op12, vuint8mf2_t op13, vuint8mf2_t op14, const uint8_t * op15, vuint64m4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_u8mf4))) +void vloxseg7ei64(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vuint8mf4_t * op6, const uint8_t * op7, vuint64m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_u8mf4_m))) +void vloxseg7ei64(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vuint8mf4_t * op6, vbool32_t op7, vuint8mf4_t op8, vuint8mf4_t op9, vuint8mf4_t op10, vuint8mf4_t op11, vuint8mf4_t op12, vuint8mf4_t op13, vuint8mf4_t op14, const uint8_t * op15, vuint64m2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_u8mf8))) +void vloxseg7ei64(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vuint8mf8_t * op6, const uint8_t * op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_u8mf8_m))) +void vloxseg7ei64(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vuint8mf8_t * op6, vbool64_t op7, vuint8mf8_t op8, vuint8mf8_t op9, vuint8mf8_t op10, vuint8mf8_t op11, vuint8mf8_t op12, vuint8mf8_t op13, vuint8mf8_t op14, const uint8_t * op15, vuint64m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_i8m1))) +void vloxseg8ei64(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vint8m1_t * op6, vint8m1_t * op7, const int8_t * op8, vuint64m8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_i8m1_m))) +void vloxseg8ei64(vint8m1_t * op0, vint8m1_t * op1, vint8m1_t * op2, vint8m1_t * op3, vint8m1_t * op4, vint8m1_t * op5, vint8m1_t * op6, vint8m1_t * op7, vbool8_t op8, vint8m1_t op9, vint8m1_t op10, vint8m1_t op11, vint8m1_t op12, vint8m1_t op13, vint8m1_t op14, vint8m1_t op15, vint8m1_t op16, const int8_t * op17, vuint64m8_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_i8mf2))) +void vloxseg8ei64(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vint8mf2_t * op6, vint8mf2_t * op7, const int8_t * op8, vuint64m4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_i8mf2_m))) +void vloxseg8ei64(vint8mf2_t * op0, vint8mf2_t * op1, vint8mf2_t * op2, vint8mf2_t * op3, vint8mf2_t * op4, vint8mf2_t * op5, vint8mf2_t * op6, vint8mf2_t * op7, vbool16_t op8, vint8mf2_t op9, vint8mf2_t op10, vint8mf2_t op11, vint8mf2_t op12, vint8mf2_t op13, vint8mf2_t op14, vint8mf2_t op15, vint8mf2_t op16, const int8_t * op17, vuint64m4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_i8mf4))) +void vloxseg8ei64(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vint8mf4_t * op6, vint8mf4_t * op7, const int8_t * op8, vuint64m2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_i8mf4_m))) +void vloxseg8ei64(vint8mf4_t * op0, vint8mf4_t * op1, vint8mf4_t * op2, vint8mf4_t * op3, vint8mf4_t * op4, vint8mf4_t * op5, vint8mf4_t * op6, vint8mf4_t * op7, vbool32_t op8, vint8mf4_t op9, vint8mf4_t op10, vint8mf4_t op11, vint8mf4_t op12, vint8mf4_t op13, vint8mf4_t op14, vint8mf4_t op15, vint8mf4_t op16, const int8_t * op17, vuint64m2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_i8mf8))) +void vloxseg8ei64(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vint8mf8_t * op6, vint8mf8_t * op7, const int8_t * op8, vuint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_i8mf8_m))) +void vloxseg8ei64(vint8mf8_t * op0, vint8mf8_t * op1, vint8mf8_t * op2, vint8mf8_t * op3, vint8mf8_t * op4, vint8mf8_t * op5, vint8mf8_t * op6, vint8mf8_t * op7, vbool64_t op8, vint8mf8_t op9, vint8mf8_t op10, vint8mf8_t op11, vint8mf8_t op12, vint8mf8_t op13, vint8mf8_t op14, vint8mf8_t op15, vint8mf8_t op16, const int8_t * op17, vuint64m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_u8m1))) +void vloxseg8ei64(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vuint8m1_t * op6, vuint8m1_t * op7, const uint8_t * op8, vuint64m8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_u8m1_m))) +void vloxseg8ei64(vuint8m1_t * op0, vuint8m1_t * op1, vuint8m1_t * op2, vuint8m1_t * op3, vuint8m1_t * op4, vuint8m1_t * op5, vuint8m1_t * op6, vuint8m1_t * op7, vbool8_t op8, vuint8m1_t op9, vuint8m1_t op10, vuint8m1_t op11, vuint8m1_t op12, vuint8m1_t op13, vuint8m1_t op14, vuint8m1_t op15, vuint8m1_t op16, const uint8_t * op17, vuint64m8_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_u8mf2))) +void vloxseg8ei64(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vuint8mf2_t * op6, vuint8mf2_t * op7, const uint8_t * op8, vuint64m4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_u8mf2_m))) +void vloxseg8ei64(vuint8mf2_t * op0, vuint8mf2_t * op1, vuint8mf2_t * op2, vuint8mf2_t * op3, vuint8mf2_t * op4, vuint8mf2_t * op5, vuint8mf2_t * op6, vuint8mf2_t * op7, vbool16_t op8, vuint8mf2_t op9, vuint8mf2_t op10, vuint8mf2_t op11, vuint8mf2_t op12, vuint8mf2_t op13, vuint8mf2_t op14, vuint8mf2_t op15, vuint8mf2_t op16, const uint8_t * op17, vuint64m4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_u8mf4))) +void vloxseg8ei64(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vuint8mf4_t * op6, vuint8mf4_t * op7, const uint8_t * op8, vuint64m2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_u8mf4_m))) +void vloxseg8ei64(vuint8mf4_t * op0, vuint8mf4_t * op1, vuint8mf4_t * op2, vuint8mf4_t * op3, vuint8mf4_t * op4, vuint8mf4_t * op5, vuint8mf4_t * op6, vuint8mf4_t * op7, vbool32_t op8, vuint8mf4_t op9, vuint8mf4_t op10, vuint8mf4_t op11, vuint8mf4_t op12, vuint8mf4_t op13, vuint8mf4_t op14, vuint8mf4_t op15, vuint8mf4_t op16, const uint8_t * op17, vuint64m2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_u8mf8))) +void vloxseg8ei64(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vuint8mf8_t * op6, vuint8mf8_t * op7, const uint8_t * op8, vuint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_u8mf8_m))) +void vloxseg8ei64(vuint8mf8_t * op0, vuint8mf8_t * op1, vuint8mf8_t * op2, vuint8mf8_t * op3, vuint8mf8_t * op4, vuint8mf8_t * op5, vuint8mf8_t * op6, vuint8mf8_t * op7, vbool64_t op8, vuint8mf8_t op9, vuint8mf8_t op10, vuint8mf8_t op11, vuint8mf8_t op12, vuint8mf8_t op13, vuint8mf8_t op14, vuint8mf8_t op15, vuint8mf8_t op16, const uint8_t * op17, vuint64m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_i16m1))) +void vloxseg2ei8(vint16m1_t * op0, vint16m1_t * op1, const int16_t * op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_i16m1_m))) +void vloxseg2ei8(vint16m1_t * op0, vint16m1_t * op1, vbool16_t op2, vint16m1_t op3, vint16m1_t op4, const int16_t * op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_i16m2))) +void vloxseg2ei8(vint16m2_t * op0, vint16m2_t * op1, const int16_t * op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_i16m2_m))) +void vloxseg2ei8(vint16m2_t * op0, vint16m2_t * op1, vbool8_t op2, vint16m2_t op3, vint16m2_t op4, const int16_t * op5, vuint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_i16m4))) +void vloxseg2ei8(vint16m4_t * op0, vint16m4_t * op1, const int16_t * op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_i16m4_m))) +void vloxseg2ei8(vint16m4_t * op0, vint16m4_t * op1, vbool4_t op2, vint16m4_t op3, vint16m4_t op4, const int16_t * op5, vuint8m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_i16mf2))) +void vloxseg2ei8(vint16mf2_t * op0, vint16mf2_t * op1, const int16_t * op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_i16mf2_m))) +void vloxseg2ei8(vint16mf2_t * op0, vint16mf2_t * op1, vbool32_t op2, vint16mf2_t op3, vint16mf2_t op4, const int16_t * op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_i16mf4))) +void vloxseg2ei8(vint16mf4_t * op0, vint16mf4_t * op1, const int16_t * op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_i16mf4_m))) +void vloxseg2ei8(vint16mf4_t * op0, vint16mf4_t * op1, vbool64_t op2, vint16mf4_t op3, vint16mf4_t op4, const int16_t * op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_u16m1))) +void vloxseg2ei8(vuint16m1_t * op0, vuint16m1_t * op1, const uint16_t * op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_u16m1_m))) +void vloxseg2ei8(vuint16m1_t * op0, vuint16m1_t * op1, vbool16_t op2, vuint16m1_t op3, vuint16m1_t op4, const uint16_t * op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_u16m2))) +void vloxseg2ei8(vuint16m2_t * op0, vuint16m2_t * op1, const uint16_t * op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_u16m2_m))) +void vloxseg2ei8(vuint16m2_t * op0, vuint16m2_t * op1, vbool8_t op2, vuint16m2_t op3, vuint16m2_t op4, const uint16_t * op5, vuint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_u16m4))) +void vloxseg2ei8(vuint16m4_t * op0, vuint16m4_t * op1, const uint16_t * op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_u16m4_m))) +void vloxseg2ei8(vuint16m4_t * op0, vuint16m4_t * op1, vbool4_t op2, vuint16m4_t op3, vuint16m4_t op4, const uint16_t * op5, vuint8m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_u16mf2))) +void vloxseg2ei8(vuint16mf2_t * op0, vuint16mf2_t * op1, const uint16_t * op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_u16mf2_m))) +void vloxseg2ei8(vuint16mf2_t * op0, vuint16mf2_t * op1, vbool32_t op2, vuint16mf2_t op3, vuint16mf2_t op4, const uint16_t * op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_u16mf4))) +void vloxseg2ei8(vuint16mf4_t * op0, vuint16mf4_t * op1, const uint16_t * op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_u16mf4_m))) +void vloxseg2ei8(vuint16mf4_t * op0, vuint16mf4_t * op1, vbool64_t op2, vuint16mf4_t op3, vuint16mf4_t op4, const uint16_t * op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_i16m1))) +void vloxseg3ei8(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, const int16_t * op3, vuint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_i16m1_m))) +void vloxseg3ei8(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vbool16_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, const int16_t * op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_i16m2))) +void vloxseg3ei8(vint16m2_t * op0, vint16m2_t * op1, vint16m2_t * op2, const int16_t * op3, vuint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_i16m2_m))) +void vloxseg3ei8(vint16m2_t * op0, vint16m2_t * op1, vint16m2_t * op2, vbool8_t op3, vint16m2_t op4, vint16m2_t op5, vint16m2_t op6, const int16_t * op7, vuint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_i16mf2))) +void vloxseg3ei8(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, const int16_t * op3, vuint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_i16mf2_m))) +void vloxseg3ei8(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vbool32_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, const int16_t * op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_i16mf4))) +void vloxseg3ei8(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, const int16_t * op3, vuint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_i16mf4_m))) +void vloxseg3ei8(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vbool64_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, const int16_t * op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_u16m1))) +void vloxseg3ei8(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, const uint16_t * op3, vuint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_u16m1_m))) +void vloxseg3ei8(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vbool16_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, const uint16_t * op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_u16m2))) +void vloxseg3ei8(vuint16m2_t * op0, vuint16m2_t * op1, vuint16m2_t * op2, const uint16_t * op3, vuint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_u16m2_m))) +void vloxseg3ei8(vuint16m2_t * op0, vuint16m2_t * op1, vuint16m2_t * op2, vbool8_t op3, vuint16m2_t op4, vuint16m2_t op5, vuint16m2_t op6, const uint16_t * op7, vuint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_u16mf2))) +void vloxseg3ei8(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, const uint16_t * op3, vuint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_u16mf2_m))) +void vloxseg3ei8(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vbool32_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, const uint16_t * op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_u16mf4))) +void vloxseg3ei8(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, const uint16_t * op3, vuint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_u16mf4_m))) +void vloxseg3ei8(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vbool64_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, const uint16_t * op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_i16m1))) +void vloxseg4ei8(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, const int16_t * op4, vuint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_i16m1_m))) +void vloxseg4ei8(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vbool16_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, const int16_t * op9, vuint8mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_i16m2))) +void vloxseg4ei8(vint16m2_t * op0, vint16m2_t * op1, vint16m2_t * op2, vint16m2_t * op3, const int16_t * op4, vuint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_i16m2_m))) +void vloxseg4ei8(vint16m2_t * op0, vint16m2_t * op1, vint16m2_t * op2, vint16m2_t * op3, vbool8_t op4, vint16m2_t op5, vint16m2_t op6, vint16m2_t op7, vint16m2_t op8, const int16_t * op9, vuint8m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_i16mf2))) +void vloxseg4ei8(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, const int16_t * op4, vuint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_i16mf2_m))) +void vloxseg4ei8(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vbool32_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, const int16_t * op9, vuint8mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_i16mf4))) +void vloxseg4ei8(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, const int16_t * op4, vuint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_i16mf4_m))) +void vloxseg4ei8(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vbool64_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, const int16_t * op9, vuint8mf8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_u16m1))) +void vloxseg4ei8(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, const uint16_t * op4, vuint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_u16m1_m))) +void vloxseg4ei8(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vbool16_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, const uint16_t * op9, vuint8mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_u16m2))) +void vloxseg4ei8(vuint16m2_t * op0, vuint16m2_t * op1, vuint16m2_t * op2, vuint16m2_t * op3, const uint16_t * op4, vuint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_u16m2_m))) +void vloxseg4ei8(vuint16m2_t * op0, vuint16m2_t * op1, vuint16m2_t * op2, vuint16m2_t * op3, vbool8_t op4, vuint16m2_t op5, vuint16m2_t op6, vuint16m2_t op7, vuint16m2_t op8, const uint16_t * op9, vuint8m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_u16mf2))) +void vloxseg4ei8(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, const uint16_t * op4, vuint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_u16mf2_m))) +void vloxseg4ei8(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vbool32_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, const uint16_t * op9, vuint8mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_u16mf4))) +void vloxseg4ei8(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, const uint16_t * op4, vuint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_u16mf4_m))) +void vloxseg4ei8(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vbool64_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, const uint16_t * op9, vuint8mf8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_i16m1))) +void vloxseg5ei8(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, const int16_t * op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_i16m1_m))) +void vloxseg5ei8(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vbool16_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, vint16m1_t op10, const int16_t * op11, vuint8mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_i16mf2))) +void vloxseg5ei8(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, const int16_t * op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_i16mf2_m))) +void vloxseg5ei8(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vbool32_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, vint16mf2_t op10, const int16_t * op11, vuint8mf4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_i16mf4))) +void vloxseg5ei8(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, const int16_t * op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_i16mf4_m))) +void vloxseg5ei8(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vbool64_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, vint16mf4_t op10, const int16_t * op11, vuint8mf8_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_u16m1))) +void vloxseg5ei8(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, const uint16_t * op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_u16m1_m))) +void vloxseg5ei8(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vbool16_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, vuint16m1_t op10, const uint16_t * op11, vuint8mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_u16mf2))) +void vloxseg5ei8(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, const uint16_t * op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_u16mf2_m))) +void vloxseg5ei8(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vbool32_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, vuint16mf2_t op10, const uint16_t * op11, vuint8mf4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_u16mf4))) +void vloxseg5ei8(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, const uint16_t * op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_u16mf4_m))) +void vloxseg5ei8(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vbool64_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, vuint16mf4_t op10, const uint16_t * op11, vuint8mf8_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_i16m1))) +void vloxseg6ei8(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, const int16_t * op6, vuint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_i16m1_m))) +void vloxseg6ei8(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vbool16_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, vint16m1_t op10, vint16m1_t op11, vint16m1_t op12, const int16_t * op13, vuint8mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_i16mf2))) +void vloxseg6ei8(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, const int16_t * op6, vuint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_i16mf2_m))) +void vloxseg6ei8(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vbool32_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, vint16mf2_t op10, vint16mf2_t op11, vint16mf2_t op12, const int16_t * op13, vuint8mf4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_i16mf4))) +void vloxseg6ei8(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, const int16_t * op6, vuint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_i16mf4_m))) +void vloxseg6ei8(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vbool64_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, vint16mf4_t op10, vint16mf4_t op11, vint16mf4_t op12, const int16_t * op13, vuint8mf8_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_u16m1))) +void vloxseg6ei8(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, const uint16_t * op6, vuint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_u16m1_m))) +void vloxseg6ei8(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vbool16_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, vuint16m1_t op10, vuint16m1_t op11, vuint16m1_t op12, const uint16_t * op13, vuint8mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_u16mf2))) +void vloxseg6ei8(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, const uint16_t * op6, vuint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_u16mf2_m))) +void vloxseg6ei8(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vbool32_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, vuint16mf2_t op10, vuint16mf2_t op11, vuint16mf2_t op12, const uint16_t * op13, vuint8mf4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_u16mf4))) +void vloxseg6ei8(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, const uint16_t * op6, vuint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_u16mf4_m))) +void vloxseg6ei8(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vbool64_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, vuint16mf4_t op10, vuint16mf4_t op11, vuint16mf4_t op12, const uint16_t * op13, vuint8mf8_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_i16m1))) +void vloxseg7ei8(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vint16m1_t * op6, const int16_t * op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_i16m1_m))) +void vloxseg7ei8(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vint16m1_t * op6, vbool16_t op7, vint16m1_t op8, vint16m1_t op9, vint16m1_t op10, vint16m1_t op11, vint16m1_t op12, vint16m1_t op13, vint16m1_t op14, const int16_t * op15, vuint8mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_i16mf2))) +void vloxseg7ei8(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vint16mf2_t * op6, const int16_t * op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_i16mf2_m))) +void vloxseg7ei8(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vint16mf2_t * op6, vbool32_t op7, vint16mf2_t op8, vint16mf2_t op9, vint16mf2_t op10, vint16mf2_t op11, vint16mf2_t op12, vint16mf2_t op13, vint16mf2_t op14, const int16_t * op15, vuint8mf4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_i16mf4))) +void vloxseg7ei8(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vint16mf4_t * op6, const int16_t * op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_i16mf4_m))) +void vloxseg7ei8(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vint16mf4_t * op6, vbool64_t op7, vint16mf4_t op8, vint16mf4_t op9, vint16mf4_t op10, vint16mf4_t op11, vint16mf4_t op12, vint16mf4_t op13, vint16mf4_t op14, const int16_t * op15, vuint8mf8_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_u16m1))) +void vloxseg7ei8(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vuint16m1_t * op6, const uint16_t * op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_u16m1_m))) +void vloxseg7ei8(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vuint16m1_t * op6, vbool16_t op7, vuint16m1_t op8, vuint16m1_t op9, vuint16m1_t op10, vuint16m1_t op11, vuint16m1_t op12, vuint16m1_t op13, vuint16m1_t op14, const uint16_t * op15, vuint8mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_u16mf2))) +void vloxseg7ei8(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vuint16mf2_t * op6, const uint16_t * op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_u16mf2_m))) +void vloxseg7ei8(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vuint16mf2_t * op6, vbool32_t op7, vuint16mf2_t op8, vuint16mf2_t op9, vuint16mf2_t op10, vuint16mf2_t op11, vuint16mf2_t op12, vuint16mf2_t op13, vuint16mf2_t op14, const uint16_t * op15, vuint8mf4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_u16mf4))) +void vloxseg7ei8(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vuint16mf4_t * op6, const uint16_t * op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_u16mf4_m))) +void vloxseg7ei8(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vuint16mf4_t * op6, vbool64_t op7, vuint16mf4_t op8, vuint16mf4_t op9, vuint16mf4_t op10, vuint16mf4_t op11, vuint16mf4_t op12, vuint16mf4_t op13, vuint16mf4_t op14, const uint16_t * op15, vuint8mf8_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_i16m1))) +void vloxseg8ei8(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vint16m1_t * op6, vint16m1_t * op7, const int16_t * op8, vuint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_i16m1_m))) +void vloxseg8ei8(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vint16m1_t * op6, vint16m1_t * op7, vbool16_t op8, vint16m1_t op9, vint16m1_t op10, vint16m1_t op11, vint16m1_t op12, vint16m1_t op13, vint16m1_t op14, vint16m1_t op15, vint16m1_t op16, const int16_t * op17, vuint8mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_i16mf2))) +void vloxseg8ei8(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vint16mf2_t * op6, vint16mf2_t * op7, const int16_t * op8, vuint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_i16mf2_m))) +void vloxseg8ei8(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vint16mf2_t * op6, vint16mf2_t * op7, vbool32_t op8, vint16mf2_t op9, vint16mf2_t op10, vint16mf2_t op11, vint16mf2_t op12, vint16mf2_t op13, vint16mf2_t op14, vint16mf2_t op15, vint16mf2_t op16, const int16_t * op17, vuint8mf4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_i16mf4))) +void vloxseg8ei8(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vint16mf4_t * op6, vint16mf4_t * op7, const int16_t * op8, vuint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_i16mf4_m))) +void vloxseg8ei8(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vint16mf4_t * op6, vint16mf4_t * op7, vbool64_t op8, vint16mf4_t op9, vint16mf4_t op10, vint16mf4_t op11, vint16mf4_t op12, vint16mf4_t op13, vint16mf4_t op14, vint16mf4_t op15, vint16mf4_t op16, const int16_t * op17, vuint8mf8_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_u16m1))) +void vloxseg8ei8(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vuint16m1_t * op6, vuint16m1_t * op7, const uint16_t * op8, vuint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_u16m1_m))) +void vloxseg8ei8(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vuint16m1_t * op6, vuint16m1_t * op7, vbool16_t op8, vuint16m1_t op9, vuint16m1_t op10, vuint16m1_t op11, vuint16m1_t op12, vuint16m1_t op13, vuint16m1_t op14, vuint16m1_t op15, vuint16m1_t op16, const uint16_t * op17, vuint8mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_u16mf2))) +void vloxseg8ei8(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vuint16mf2_t * op6, vuint16mf2_t * op7, const uint16_t * op8, vuint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_u16mf2_m))) +void vloxseg8ei8(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vuint16mf2_t * op6, vuint16mf2_t * op7, vbool32_t op8, vuint16mf2_t op9, vuint16mf2_t op10, vuint16mf2_t op11, vuint16mf2_t op12, vuint16mf2_t op13, vuint16mf2_t op14, vuint16mf2_t op15, vuint16mf2_t op16, const uint16_t * op17, vuint8mf4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_u16mf4))) +void vloxseg8ei8(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vuint16mf4_t * op6, vuint16mf4_t * op7, const uint16_t * op8, vuint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_u16mf4_m))) +void vloxseg8ei8(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vuint16mf4_t * op6, vuint16mf4_t * op7, vbool64_t op8, vuint16mf4_t op9, vuint16mf4_t op10, vuint16mf4_t op11, vuint16mf4_t op12, vuint16mf4_t op13, vuint16mf4_t op14, vuint16mf4_t op15, vuint16mf4_t op16, const uint16_t * op17, vuint8mf8_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_i16m1))) +void vloxseg2ei16(vint16m1_t * op0, vint16m1_t * op1, const int16_t * op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_i16m1_m))) +void vloxseg2ei16(vint16m1_t * op0, vint16m1_t * op1, vbool16_t op2, vint16m1_t op3, vint16m1_t op4, const int16_t * op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_i16m2))) +void vloxseg2ei16(vint16m2_t * op0, vint16m2_t * op1, const int16_t * op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_i16m2_m))) +void vloxseg2ei16(vint16m2_t * op0, vint16m2_t * op1, vbool8_t op2, vint16m2_t op3, vint16m2_t op4, const int16_t * op5, vuint16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_i16m4))) +void vloxseg2ei16(vint16m4_t * op0, vint16m4_t * op1, const int16_t * op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_i16m4_m))) +void vloxseg2ei16(vint16m4_t * op0, vint16m4_t * op1, vbool4_t op2, vint16m4_t op3, vint16m4_t op4, const int16_t * op5, vuint16m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_i16mf2))) +void vloxseg2ei16(vint16mf2_t * op0, vint16mf2_t * op1, const int16_t * op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_i16mf2_m))) +void vloxseg2ei16(vint16mf2_t * op0, vint16mf2_t * op1, vbool32_t op2, vint16mf2_t op3, vint16mf2_t op4, const int16_t * op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_i16mf4))) +void vloxseg2ei16(vint16mf4_t * op0, vint16mf4_t * op1, const int16_t * op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_i16mf4_m))) +void vloxseg2ei16(vint16mf4_t * op0, vint16mf4_t * op1, vbool64_t op2, vint16mf4_t op3, vint16mf4_t op4, const int16_t * op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_u16m1))) +void vloxseg2ei16(vuint16m1_t * op0, vuint16m1_t * op1, const uint16_t * op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_u16m1_m))) +void vloxseg2ei16(vuint16m1_t * op0, vuint16m1_t * op1, vbool16_t op2, vuint16m1_t op3, vuint16m1_t op4, const uint16_t * op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_u16m2))) +void vloxseg2ei16(vuint16m2_t * op0, vuint16m2_t * op1, const uint16_t * op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_u16m2_m))) +void vloxseg2ei16(vuint16m2_t * op0, vuint16m2_t * op1, vbool8_t op2, vuint16m2_t op3, vuint16m2_t op4, const uint16_t * op5, vuint16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_u16m4))) +void vloxseg2ei16(vuint16m4_t * op0, vuint16m4_t * op1, const uint16_t * op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_u16m4_m))) +void vloxseg2ei16(vuint16m4_t * op0, vuint16m4_t * op1, vbool4_t op2, vuint16m4_t op3, vuint16m4_t op4, const uint16_t * op5, vuint16m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_u16mf2))) +void vloxseg2ei16(vuint16mf2_t * op0, vuint16mf2_t * op1, const uint16_t * op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_u16mf2_m))) +void vloxseg2ei16(vuint16mf2_t * op0, vuint16mf2_t * op1, vbool32_t op2, vuint16mf2_t op3, vuint16mf2_t op4, const uint16_t * op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_u16mf4))) +void vloxseg2ei16(vuint16mf4_t * op0, vuint16mf4_t * op1, const uint16_t * op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_u16mf4_m))) +void vloxseg2ei16(vuint16mf4_t * op0, vuint16mf4_t * op1, vbool64_t op2, vuint16mf4_t op3, vuint16mf4_t op4, const uint16_t * op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_i16m1))) +void vloxseg3ei16(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, const int16_t * op3, vuint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_i16m1_m))) +void vloxseg3ei16(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vbool16_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, const int16_t * op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_i16m2))) +void vloxseg3ei16(vint16m2_t * op0, vint16m2_t * op1, vint16m2_t * op2, const int16_t * op3, vuint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_i16m2_m))) +void vloxseg3ei16(vint16m2_t * op0, vint16m2_t * op1, vint16m2_t * op2, vbool8_t op3, vint16m2_t op4, vint16m2_t op5, vint16m2_t op6, const int16_t * op7, vuint16m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_i16mf2))) +void vloxseg3ei16(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, const int16_t * op3, vuint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_i16mf2_m))) +void vloxseg3ei16(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vbool32_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, const int16_t * op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_i16mf4))) +void vloxseg3ei16(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, const int16_t * op3, vuint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_i16mf4_m))) +void vloxseg3ei16(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vbool64_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, const int16_t * op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_u16m1))) +void vloxseg3ei16(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, const uint16_t * op3, vuint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_u16m1_m))) +void vloxseg3ei16(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vbool16_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, const uint16_t * op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_u16m2))) +void vloxseg3ei16(vuint16m2_t * op0, vuint16m2_t * op1, vuint16m2_t * op2, const uint16_t * op3, vuint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_u16m2_m))) +void vloxseg3ei16(vuint16m2_t * op0, vuint16m2_t * op1, vuint16m2_t * op2, vbool8_t op3, vuint16m2_t op4, vuint16m2_t op5, vuint16m2_t op6, const uint16_t * op7, vuint16m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_u16mf2))) +void vloxseg3ei16(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, const uint16_t * op3, vuint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_u16mf2_m))) +void vloxseg3ei16(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vbool32_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, const uint16_t * op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_u16mf4))) +void vloxseg3ei16(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, const uint16_t * op3, vuint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_u16mf4_m))) +void vloxseg3ei16(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vbool64_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, const uint16_t * op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_i16m1))) +void vloxseg4ei16(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, const int16_t * op4, vuint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_i16m1_m))) +void vloxseg4ei16(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vbool16_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, const int16_t * op9, vuint16m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_i16m2))) +void vloxseg4ei16(vint16m2_t * op0, vint16m2_t * op1, vint16m2_t * op2, vint16m2_t * op3, const int16_t * op4, vuint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_i16m2_m))) +void vloxseg4ei16(vint16m2_t * op0, vint16m2_t * op1, vint16m2_t * op2, vint16m2_t * op3, vbool8_t op4, vint16m2_t op5, vint16m2_t op6, vint16m2_t op7, vint16m2_t op8, const int16_t * op9, vuint16m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_i16mf2))) +void vloxseg4ei16(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, const int16_t * op4, vuint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_i16mf2_m))) +void vloxseg4ei16(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vbool32_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, const int16_t * op9, vuint16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_i16mf4))) +void vloxseg4ei16(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, const int16_t * op4, vuint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_i16mf4_m))) +void vloxseg4ei16(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vbool64_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, const int16_t * op9, vuint16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_u16m1))) +void vloxseg4ei16(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, const uint16_t * op4, vuint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_u16m1_m))) +void vloxseg4ei16(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vbool16_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, const uint16_t * op9, vuint16m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_u16m2))) +void vloxseg4ei16(vuint16m2_t * op0, vuint16m2_t * op1, vuint16m2_t * op2, vuint16m2_t * op3, const uint16_t * op4, vuint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_u16m2_m))) +void vloxseg4ei16(vuint16m2_t * op0, vuint16m2_t * op1, vuint16m2_t * op2, vuint16m2_t * op3, vbool8_t op4, vuint16m2_t op5, vuint16m2_t op6, vuint16m2_t op7, vuint16m2_t op8, const uint16_t * op9, vuint16m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_u16mf2))) +void vloxseg4ei16(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, const uint16_t * op4, vuint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_u16mf2_m))) +void vloxseg4ei16(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vbool32_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, const uint16_t * op9, vuint16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_u16mf4))) +void vloxseg4ei16(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, const uint16_t * op4, vuint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_u16mf4_m))) +void vloxseg4ei16(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vbool64_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, const uint16_t * op9, vuint16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_i16m1))) +void vloxseg5ei16(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, const int16_t * op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_i16m1_m))) +void vloxseg5ei16(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vbool16_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, vint16m1_t op10, const int16_t * op11, vuint16m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_i16mf2))) +void vloxseg5ei16(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, const int16_t * op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_i16mf2_m))) +void vloxseg5ei16(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vbool32_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, vint16mf2_t op10, const int16_t * op11, vuint16mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_i16mf4))) +void vloxseg5ei16(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, const int16_t * op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_i16mf4_m))) +void vloxseg5ei16(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vbool64_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, vint16mf4_t op10, const int16_t * op11, vuint16mf4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_u16m1))) +void vloxseg5ei16(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, const uint16_t * op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_u16m1_m))) +void vloxseg5ei16(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vbool16_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, vuint16m1_t op10, const uint16_t * op11, vuint16m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_u16mf2))) +void vloxseg5ei16(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, const uint16_t * op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_u16mf2_m))) +void vloxseg5ei16(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vbool32_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, vuint16mf2_t op10, const uint16_t * op11, vuint16mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_u16mf4))) +void vloxseg5ei16(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, const uint16_t * op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_u16mf4_m))) +void vloxseg5ei16(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vbool64_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, vuint16mf4_t op10, const uint16_t * op11, vuint16mf4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_i16m1))) +void vloxseg6ei16(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, const int16_t * op6, vuint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_i16m1_m))) +void vloxseg6ei16(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vbool16_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, vint16m1_t op10, vint16m1_t op11, vint16m1_t op12, const int16_t * op13, vuint16m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_i16mf2))) +void vloxseg6ei16(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, const int16_t * op6, vuint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_i16mf2_m))) +void vloxseg6ei16(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vbool32_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, vint16mf2_t op10, vint16mf2_t op11, vint16mf2_t op12, const int16_t * op13, vuint16mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_i16mf4))) +void vloxseg6ei16(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, const int16_t * op6, vuint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_i16mf4_m))) +void vloxseg6ei16(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vbool64_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, vint16mf4_t op10, vint16mf4_t op11, vint16mf4_t op12, const int16_t * op13, vuint16mf4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_u16m1))) +void vloxseg6ei16(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, const uint16_t * op6, vuint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_u16m1_m))) +void vloxseg6ei16(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vbool16_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, vuint16m1_t op10, vuint16m1_t op11, vuint16m1_t op12, const uint16_t * op13, vuint16m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_u16mf2))) +void vloxseg6ei16(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, const uint16_t * op6, vuint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_u16mf2_m))) +void vloxseg6ei16(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vbool32_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, vuint16mf2_t op10, vuint16mf2_t op11, vuint16mf2_t op12, const uint16_t * op13, vuint16mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_u16mf4))) +void vloxseg6ei16(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, const uint16_t * op6, vuint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_u16mf4_m))) +void vloxseg6ei16(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vbool64_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, vuint16mf4_t op10, vuint16mf4_t op11, vuint16mf4_t op12, const uint16_t * op13, vuint16mf4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_i16m1))) +void vloxseg7ei16(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vint16m1_t * op6, const int16_t * op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_i16m1_m))) +void vloxseg7ei16(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vint16m1_t * op6, vbool16_t op7, vint16m1_t op8, vint16m1_t op9, vint16m1_t op10, vint16m1_t op11, vint16m1_t op12, vint16m1_t op13, vint16m1_t op14, const int16_t * op15, vuint16m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_i16mf2))) +void vloxseg7ei16(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vint16mf2_t * op6, const int16_t * op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_i16mf2_m))) +void vloxseg7ei16(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vint16mf2_t * op6, vbool32_t op7, vint16mf2_t op8, vint16mf2_t op9, vint16mf2_t op10, vint16mf2_t op11, vint16mf2_t op12, vint16mf2_t op13, vint16mf2_t op14, const int16_t * op15, vuint16mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_i16mf4))) +void vloxseg7ei16(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vint16mf4_t * op6, const int16_t * op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_i16mf4_m))) +void vloxseg7ei16(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vint16mf4_t * op6, vbool64_t op7, vint16mf4_t op8, vint16mf4_t op9, vint16mf4_t op10, vint16mf4_t op11, vint16mf4_t op12, vint16mf4_t op13, vint16mf4_t op14, const int16_t * op15, vuint16mf4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_u16m1))) +void vloxseg7ei16(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vuint16m1_t * op6, const uint16_t * op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_u16m1_m))) +void vloxseg7ei16(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vuint16m1_t * op6, vbool16_t op7, vuint16m1_t op8, vuint16m1_t op9, vuint16m1_t op10, vuint16m1_t op11, vuint16m1_t op12, vuint16m1_t op13, vuint16m1_t op14, const uint16_t * op15, vuint16m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_u16mf2))) +void vloxseg7ei16(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vuint16mf2_t * op6, const uint16_t * op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_u16mf2_m))) +void vloxseg7ei16(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vuint16mf2_t * op6, vbool32_t op7, vuint16mf2_t op8, vuint16mf2_t op9, vuint16mf2_t op10, vuint16mf2_t op11, vuint16mf2_t op12, vuint16mf2_t op13, vuint16mf2_t op14, const uint16_t * op15, vuint16mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_u16mf4))) +void vloxseg7ei16(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vuint16mf4_t * op6, const uint16_t * op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_u16mf4_m))) +void vloxseg7ei16(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vuint16mf4_t * op6, vbool64_t op7, vuint16mf4_t op8, vuint16mf4_t op9, vuint16mf4_t op10, vuint16mf4_t op11, vuint16mf4_t op12, vuint16mf4_t op13, vuint16mf4_t op14, const uint16_t * op15, vuint16mf4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_i16m1))) +void vloxseg8ei16(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vint16m1_t * op6, vint16m1_t * op7, const int16_t * op8, vuint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_i16m1_m))) +void vloxseg8ei16(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vint16m1_t * op6, vint16m1_t * op7, vbool16_t op8, vint16m1_t op9, vint16m1_t op10, vint16m1_t op11, vint16m1_t op12, vint16m1_t op13, vint16m1_t op14, vint16m1_t op15, vint16m1_t op16, const int16_t * op17, vuint16m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_i16mf2))) +void vloxseg8ei16(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vint16mf2_t * op6, vint16mf2_t * op7, const int16_t * op8, vuint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_i16mf2_m))) +void vloxseg8ei16(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vint16mf2_t * op6, vint16mf2_t * op7, vbool32_t op8, vint16mf2_t op9, vint16mf2_t op10, vint16mf2_t op11, vint16mf2_t op12, vint16mf2_t op13, vint16mf2_t op14, vint16mf2_t op15, vint16mf2_t op16, const int16_t * op17, vuint16mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_i16mf4))) +void vloxseg8ei16(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vint16mf4_t * op6, vint16mf4_t * op7, const int16_t * op8, vuint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_i16mf4_m))) +void vloxseg8ei16(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vint16mf4_t * op6, vint16mf4_t * op7, vbool64_t op8, vint16mf4_t op9, vint16mf4_t op10, vint16mf4_t op11, vint16mf4_t op12, vint16mf4_t op13, vint16mf4_t op14, vint16mf4_t op15, vint16mf4_t op16, const int16_t * op17, vuint16mf4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_u16m1))) +void vloxseg8ei16(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vuint16m1_t * op6, vuint16m1_t * op7, const uint16_t * op8, vuint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_u16m1_m))) +void vloxseg8ei16(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vuint16m1_t * op6, vuint16m1_t * op7, vbool16_t op8, vuint16m1_t op9, vuint16m1_t op10, vuint16m1_t op11, vuint16m1_t op12, vuint16m1_t op13, vuint16m1_t op14, vuint16m1_t op15, vuint16m1_t op16, const uint16_t * op17, vuint16m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_u16mf2))) +void vloxseg8ei16(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vuint16mf2_t * op6, vuint16mf2_t * op7, const uint16_t * op8, vuint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_u16mf2_m))) +void vloxseg8ei16(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vuint16mf2_t * op6, vuint16mf2_t * op7, vbool32_t op8, vuint16mf2_t op9, vuint16mf2_t op10, vuint16mf2_t op11, vuint16mf2_t op12, vuint16mf2_t op13, vuint16mf2_t op14, vuint16mf2_t op15, vuint16mf2_t op16, const uint16_t * op17, vuint16mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_u16mf4))) +void vloxseg8ei16(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vuint16mf4_t * op6, vuint16mf4_t * op7, const uint16_t * op8, vuint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_u16mf4_m))) +void vloxseg8ei16(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vuint16mf4_t * op6, vuint16mf4_t * op7, vbool64_t op8, vuint16mf4_t op9, vuint16mf4_t op10, vuint16mf4_t op11, vuint16mf4_t op12, vuint16mf4_t op13, vuint16mf4_t op14, vuint16mf4_t op15, vuint16mf4_t op16, const uint16_t * op17, vuint16mf4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_i16m1))) +void vloxseg2ei32(vint16m1_t * op0, vint16m1_t * op1, const int16_t * op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_i16m1_m))) +void vloxseg2ei32(vint16m1_t * op0, vint16m1_t * op1, vbool16_t op2, vint16m1_t op3, vint16m1_t op4, const int16_t * op5, vuint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_i16m2))) +void vloxseg2ei32(vint16m2_t * op0, vint16m2_t * op1, const int16_t * op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_i16m2_m))) +void vloxseg2ei32(vint16m2_t * op0, vint16m2_t * op1, vbool8_t op2, vint16m2_t op3, vint16m2_t op4, const int16_t * op5, vuint32m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_i16m4))) +void vloxseg2ei32(vint16m4_t * op0, vint16m4_t * op1, const int16_t * op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_i16m4_m))) +void vloxseg2ei32(vint16m4_t * op0, vint16m4_t * op1, vbool4_t op2, vint16m4_t op3, vint16m4_t op4, const int16_t * op5, vuint32m8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_i16mf2))) +void vloxseg2ei32(vint16mf2_t * op0, vint16mf2_t * op1, const int16_t * op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_i16mf2_m))) +void vloxseg2ei32(vint16mf2_t * op0, vint16mf2_t * op1, vbool32_t op2, vint16mf2_t op3, vint16mf2_t op4, const int16_t * op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_i16mf4))) +void vloxseg2ei32(vint16mf4_t * op0, vint16mf4_t * op1, const int16_t * op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_i16mf4_m))) +void vloxseg2ei32(vint16mf4_t * op0, vint16mf4_t * op1, vbool64_t op2, vint16mf4_t op3, vint16mf4_t op4, const int16_t * op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_u16m1))) +void vloxseg2ei32(vuint16m1_t * op0, vuint16m1_t * op1, const uint16_t * op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_u16m1_m))) +void vloxseg2ei32(vuint16m1_t * op0, vuint16m1_t * op1, vbool16_t op2, vuint16m1_t op3, vuint16m1_t op4, const uint16_t * op5, vuint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_u16m2))) +void vloxseg2ei32(vuint16m2_t * op0, vuint16m2_t * op1, const uint16_t * op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_u16m2_m))) +void vloxseg2ei32(vuint16m2_t * op0, vuint16m2_t * op1, vbool8_t op2, vuint16m2_t op3, vuint16m2_t op4, const uint16_t * op5, vuint32m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_u16m4))) +void vloxseg2ei32(vuint16m4_t * op0, vuint16m4_t * op1, const uint16_t * op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_u16m4_m))) +void vloxseg2ei32(vuint16m4_t * op0, vuint16m4_t * op1, vbool4_t op2, vuint16m4_t op3, vuint16m4_t op4, const uint16_t * op5, vuint32m8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_u16mf2))) +void vloxseg2ei32(vuint16mf2_t * op0, vuint16mf2_t * op1, const uint16_t * op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_u16mf2_m))) +void vloxseg2ei32(vuint16mf2_t * op0, vuint16mf2_t * op1, vbool32_t op2, vuint16mf2_t op3, vuint16mf2_t op4, const uint16_t * op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_u16mf4))) +void vloxseg2ei32(vuint16mf4_t * op0, vuint16mf4_t * op1, const uint16_t * op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_u16mf4_m))) +void vloxseg2ei32(vuint16mf4_t * op0, vuint16mf4_t * op1, vbool64_t op2, vuint16mf4_t op3, vuint16mf4_t op4, const uint16_t * op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_i16m1))) +void vloxseg3ei32(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, const int16_t * op3, vuint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_i16m1_m))) +void vloxseg3ei32(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vbool16_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, const int16_t * op7, vuint32m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_i16m2))) +void vloxseg3ei32(vint16m2_t * op0, vint16m2_t * op1, vint16m2_t * op2, const int16_t * op3, vuint32m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_i16m2_m))) +void vloxseg3ei32(vint16m2_t * op0, vint16m2_t * op1, vint16m2_t * op2, vbool8_t op3, vint16m2_t op4, vint16m2_t op5, vint16m2_t op6, const int16_t * op7, vuint32m4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_i16mf2))) +void vloxseg3ei32(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, const int16_t * op3, vuint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_i16mf2_m))) +void vloxseg3ei32(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vbool32_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, const int16_t * op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_i16mf4))) +void vloxseg3ei32(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, const int16_t * op3, vuint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_i16mf4_m))) +void vloxseg3ei32(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vbool64_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, const int16_t * op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_u16m1))) +void vloxseg3ei32(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, const uint16_t * op3, vuint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_u16m1_m))) +void vloxseg3ei32(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vbool16_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, const uint16_t * op7, vuint32m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_u16m2))) +void vloxseg3ei32(vuint16m2_t * op0, vuint16m2_t * op1, vuint16m2_t * op2, const uint16_t * op3, vuint32m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_u16m2_m))) +void vloxseg3ei32(vuint16m2_t * op0, vuint16m2_t * op1, vuint16m2_t * op2, vbool8_t op3, vuint16m2_t op4, vuint16m2_t op5, vuint16m2_t op6, const uint16_t * op7, vuint32m4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_u16mf2))) +void vloxseg3ei32(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, const uint16_t * op3, vuint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_u16mf2_m))) +void vloxseg3ei32(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vbool32_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, const uint16_t * op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_u16mf4))) +void vloxseg3ei32(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, const uint16_t * op3, vuint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_u16mf4_m))) +void vloxseg3ei32(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vbool64_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, const uint16_t * op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_i16m1))) +void vloxseg4ei32(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, const int16_t * op4, vuint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_i16m1_m))) +void vloxseg4ei32(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vbool16_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, const int16_t * op9, vuint32m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_i16m2))) +void vloxseg4ei32(vint16m2_t * op0, vint16m2_t * op1, vint16m2_t * op2, vint16m2_t * op3, const int16_t * op4, vuint32m4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_i16m2_m))) +void vloxseg4ei32(vint16m2_t * op0, vint16m2_t * op1, vint16m2_t * op2, vint16m2_t * op3, vbool8_t op4, vint16m2_t op5, vint16m2_t op6, vint16m2_t op7, vint16m2_t op8, const int16_t * op9, vuint32m4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_i16mf2))) +void vloxseg4ei32(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, const int16_t * op4, vuint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_i16mf2_m))) +void vloxseg4ei32(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vbool32_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, const int16_t * op9, vuint32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_i16mf4))) +void vloxseg4ei32(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, const int16_t * op4, vuint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_i16mf4_m))) +void vloxseg4ei32(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vbool64_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, const int16_t * op9, vuint32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_u16m1))) +void vloxseg4ei32(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, const uint16_t * op4, vuint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_u16m1_m))) +void vloxseg4ei32(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vbool16_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, const uint16_t * op9, vuint32m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_u16m2))) +void vloxseg4ei32(vuint16m2_t * op0, vuint16m2_t * op1, vuint16m2_t * op2, vuint16m2_t * op3, const uint16_t * op4, vuint32m4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_u16m2_m))) +void vloxseg4ei32(vuint16m2_t * op0, vuint16m2_t * op1, vuint16m2_t * op2, vuint16m2_t * op3, vbool8_t op4, vuint16m2_t op5, vuint16m2_t op6, vuint16m2_t op7, vuint16m2_t op8, const uint16_t * op9, vuint32m4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_u16mf2))) +void vloxseg4ei32(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, const uint16_t * op4, vuint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_u16mf2_m))) +void vloxseg4ei32(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vbool32_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, const uint16_t * op9, vuint32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_u16mf4))) +void vloxseg4ei32(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, const uint16_t * op4, vuint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_u16mf4_m))) +void vloxseg4ei32(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vbool64_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, const uint16_t * op9, vuint32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_i16m1))) +void vloxseg5ei32(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, const int16_t * op5, vuint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_i16m1_m))) +void vloxseg5ei32(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vbool16_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, vint16m1_t op10, const int16_t * op11, vuint32m2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_i16mf2))) +void vloxseg5ei32(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, const int16_t * op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_i16mf2_m))) +void vloxseg5ei32(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vbool32_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, vint16mf2_t op10, const int16_t * op11, vuint32m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_i16mf4))) +void vloxseg5ei32(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, const int16_t * op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_i16mf4_m))) +void vloxseg5ei32(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vbool64_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, vint16mf4_t op10, const int16_t * op11, vuint32mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_u16m1))) +void vloxseg5ei32(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, const uint16_t * op5, vuint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_u16m1_m))) +void vloxseg5ei32(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vbool16_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, vuint16m1_t op10, const uint16_t * op11, vuint32m2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_u16mf2))) +void vloxseg5ei32(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, const uint16_t * op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_u16mf2_m))) +void vloxseg5ei32(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vbool32_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, vuint16mf2_t op10, const uint16_t * op11, vuint32m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_u16mf4))) +void vloxseg5ei32(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, const uint16_t * op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_u16mf4_m))) +void vloxseg5ei32(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vbool64_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, vuint16mf4_t op10, const uint16_t * op11, vuint32mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_i16m1))) +void vloxseg6ei32(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, const int16_t * op6, vuint32m2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_i16m1_m))) +void vloxseg6ei32(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vbool16_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, vint16m1_t op10, vint16m1_t op11, vint16m1_t op12, const int16_t * op13, vuint32m2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_i16mf2))) +void vloxseg6ei32(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, const int16_t * op6, vuint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_i16mf2_m))) +void vloxseg6ei32(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vbool32_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, vint16mf2_t op10, vint16mf2_t op11, vint16mf2_t op12, const int16_t * op13, vuint32m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_i16mf4))) +void vloxseg6ei32(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, const int16_t * op6, vuint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_i16mf4_m))) +void vloxseg6ei32(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vbool64_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, vint16mf4_t op10, vint16mf4_t op11, vint16mf4_t op12, const int16_t * op13, vuint32mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_u16m1))) +void vloxseg6ei32(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, const uint16_t * op6, vuint32m2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_u16m1_m))) +void vloxseg6ei32(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vbool16_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, vuint16m1_t op10, vuint16m1_t op11, vuint16m1_t op12, const uint16_t * op13, vuint32m2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_u16mf2))) +void vloxseg6ei32(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, const uint16_t * op6, vuint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_u16mf2_m))) +void vloxseg6ei32(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vbool32_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, vuint16mf2_t op10, vuint16mf2_t op11, vuint16mf2_t op12, const uint16_t * op13, vuint32m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_u16mf4))) +void vloxseg6ei32(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, const uint16_t * op6, vuint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_u16mf4_m))) +void vloxseg6ei32(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vbool64_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, vuint16mf4_t op10, vuint16mf4_t op11, vuint16mf4_t op12, const uint16_t * op13, vuint32mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_i16m1))) +void vloxseg7ei32(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vint16m1_t * op6, const int16_t * op7, vuint32m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_i16m1_m))) +void vloxseg7ei32(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vint16m1_t * op6, vbool16_t op7, vint16m1_t op8, vint16m1_t op9, vint16m1_t op10, vint16m1_t op11, vint16m1_t op12, vint16m1_t op13, vint16m1_t op14, const int16_t * op15, vuint32m2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_i16mf2))) +void vloxseg7ei32(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vint16mf2_t * op6, const int16_t * op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_i16mf2_m))) +void vloxseg7ei32(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vint16mf2_t * op6, vbool32_t op7, vint16mf2_t op8, vint16mf2_t op9, vint16mf2_t op10, vint16mf2_t op11, vint16mf2_t op12, vint16mf2_t op13, vint16mf2_t op14, const int16_t * op15, vuint32m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_i16mf4))) +void vloxseg7ei32(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vint16mf4_t * op6, const int16_t * op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_i16mf4_m))) +void vloxseg7ei32(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vint16mf4_t * op6, vbool64_t op7, vint16mf4_t op8, vint16mf4_t op9, vint16mf4_t op10, vint16mf4_t op11, vint16mf4_t op12, vint16mf4_t op13, vint16mf4_t op14, const int16_t * op15, vuint32mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_u16m1))) +void vloxseg7ei32(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vuint16m1_t * op6, const uint16_t * op7, vuint32m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_u16m1_m))) +void vloxseg7ei32(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vuint16m1_t * op6, vbool16_t op7, vuint16m1_t op8, vuint16m1_t op9, vuint16m1_t op10, vuint16m1_t op11, vuint16m1_t op12, vuint16m1_t op13, vuint16m1_t op14, const uint16_t * op15, vuint32m2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_u16mf2))) +void vloxseg7ei32(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vuint16mf2_t * op6, const uint16_t * op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_u16mf2_m))) +void vloxseg7ei32(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vuint16mf2_t * op6, vbool32_t op7, vuint16mf2_t op8, vuint16mf2_t op9, vuint16mf2_t op10, vuint16mf2_t op11, vuint16mf2_t op12, vuint16mf2_t op13, vuint16mf2_t op14, const uint16_t * op15, vuint32m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_u16mf4))) +void vloxseg7ei32(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vuint16mf4_t * op6, const uint16_t * op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_u16mf4_m))) +void vloxseg7ei32(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vuint16mf4_t * op6, vbool64_t op7, vuint16mf4_t op8, vuint16mf4_t op9, vuint16mf4_t op10, vuint16mf4_t op11, vuint16mf4_t op12, vuint16mf4_t op13, vuint16mf4_t op14, const uint16_t * op15, vuint32mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_i16m1))) +void vloxseg8ei32(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vint16m1_t * op6, vint16m1_t * op7, const int16_t * op8, vuint32m2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_i16m1_m))) +void vloxseg8ei32(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vint16m1_t * op6, vint16m1_t * op7, vbool16_t op8, vint16m1_t op9, vint16m1_t op10, vint16m1_t op11, vint16m1_t op12, vint16m1_t op13, vint16m1_t op14, vint16m1_t op15, vint16m1_t op16, const int16_t * op17, vuint32m2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_i16mf2))) +void vloxseg8ei32(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vint16mf2_t * op6, vint16mf2_t * op7, const int16_t * op8, vuint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_i16mf2_m))) +void vloxseg8ei32(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vint16mf2_t * op6, vint16mf2_t * op7, vbool32_t op8, vint16mf2_t op9, vint16mf2_t op10, vint16mf2_t op11, vint16mf2_t op12, vint16mf2_t op13, vint16mf2_t op14, vint16mf2_t op15, vint16mf2_t op16, const int16_t * op17, vuint32m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_i16mf4))) +void vloxseg8ei32(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vint16mf4_t * op6, vint16mf4_t * op7, const int16_t * op8, vuint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_i16mf4_m))) +void vloxseg8ei32(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vint16mf4_t * op6, vint16mf4_t * op7, vbool64_t op8, vint16mf4_t op9, vint16mf4_t op10, vint16mf4_t op11, vint16mf4_t op12, vint16mf4_t op13, vint16mf4_t op14, vint16mf4_t op15, vint16mf4_t op16, const int16_t * op17, vuint32mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_u16m1))) +void vloxseg8ei32(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vuint16m1_t * op6, vuint16m1_t * op7, const uint16_t * op8, vuint32m2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_u16m1_m))) +void vloxseg8ei32(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vuint16m1_t * op6, vuint16m1_t * op7, vbool16_t op8, vuint16m1_t op9, vuint16m1_t op10, vuint16m1_t op11, vuint16m1_t op12, vuint16m1_t op13, vuint16m1_t op14, vuint16m1_t op15, vuint16m1_t op16, const uint16_t * op17, vuint32m2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_u16mf2))) +void vloxseg8ei32(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vuint16mf2_t * op6, vuint16mf2_t * op7, const uint16_t * op8, vuint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_u16mf2_m))) +void vloxseg8ei32(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vuint16mf2_t * op6, vuint16mf2_t * op7, vbool32_t op8, vuint16mf2_t op9, vuint16mf2_t op10, vuint16mf2_t op11, vuint16mf2_t op12, vuint16mf2_t op13, vuint16mf2_t op14, vuint16mf2_t op15, vuint16mf2_t op16, const uint16_t * op17, vuint32m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_u16mf4))) +void vloxseg8ei32(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vuint16mf4_t * op6, vuint16mf4_t * op7, const uint16_t * op8, vuint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_u16mf4_m))) +void vloxseg8ei32(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vuint16mf4_t * op6, vuint16mf4_t * op7, vbool64_t op8, vuint16mf4_t op9, vuint16mf4_t op10, vuint16mf4_t op11, vuint16mf4_t op12, vuint16mf4_t op13, vuint16mf4_t op14, vuint16mf4_t op15, vuint16mf4_t op16, const uint16_t * op17, vuint32mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_i16m1))) +void vloxseg2ei64(vint16m1_t * op0, vint16m1_t * op1, const int16_t * op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_i16m1_m))) +void vloxseg2ei64(vint16m1_t * op0, vint16m1_t * op1, vbool16_t op2, vint16m1_t op3, vint16m1_t op4, const int16_t * op5, vuint64m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_i16m2))) +void vloxseg2ei64(vint16m2_t * op0, vint16m2_t * op1, const int16_t * op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_i16m2_m))) +void vloxseg2ei64(vint16m2_t * op0, vint16m2_t * op1, vbool8_t op2, vint16m2_t op3, vint16m2_t op4, const int16_t * op5, vuint64m8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_i16mf2))) +void vloxseg2ei64(vint16mf2_t * op0, vint16mf2_t * op1, const int16_t * op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_i16mf2_m))) +void vloxseg2ei64(vint16mf2_t * op0, vint16mf2_t * op1, vbool32_t op2, vint16mf2_t op3, vint16mf2_t op4, const int16_t * op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_i16mf4))) +void vloxseg2ei64(vint16mf4_t * op0, vint16mf4_t * op1, const int16_t * op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_i16mf4_m))) +void vloxseg2ei64(vint16mf4_t * op0, vint16mf4_t * op1, vbool64_t op2, vint16mf4_t op3, vint16mf4_t op4, const int16_t * op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_u16m1))) +void vloxseg2ei64(vuint16m1_t * op0, vuint16m1_t * op1, const uint16_t * op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_u16m1_m))) +void vloxseg2ei64(vuint16m1_t * op0, vuint16m1_t * op1, vbool16_t op2, vuint16m1_t op3, vuint16m1_t op4, const uint16_t * op5, vuint64m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_u16m2))) +void vloxseg2ei64(vuint16m2_t * op0, vuint16m2_t * op1, const uint16_t * op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_u16m2_m))) +void vloxseg2ei64(vuint16m2_t * op0, vuint16m2_t * op1, vbool8_t op2, vuint16m2_t op3, vuint16m2_t op4, const uint16_t * op5, vuint64m8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_u16mf2))) +void vloxseg2ei64(vuint16mf2_t * op0, vuint16mf2_t * op1, const uint16_t * op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_u16mf2_m))) +void vloxseg2ei64(vuint16mf2_t * op0, vuint16mf2_t * op1, vbool32_t op2, vuint16mf2_t op3, vuint16mf2_t op4, const uint16_t * op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_u16mf4))) +void vloxseg2ei64(vuint16mf4_t * op0, vuint16mf4_t * op1, const uint16_t * op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_u16mf4_m))) +void vloxseg2ei64(vuint16mf4_t * op0, vuint16mf4_t * op1, vbool64_t op2, vuint16mf4_t op3, vuint16mf4_t op4, const uint16_t * op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_i16m1))) +void vloxseg3ei64(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, const int16_t * op3, vuint64m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_i16m1_m))) +void vloxseg3ei64(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vbool16_t op3, vint16m1_t op4, vint16m1_t op5, vint16m1_t op6, const int16_t * op7, vuint64m4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_i16m2))) +void vloxseg3ei64(vint16m2_t * op0, vint16m2_t * op1, vint16m2_t * op2, const int16_t * op3, vuint64m8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_i16m2_m))) +void vloxseg3ei64(vint16m2_t * op0, vint16m2_t * op1, vint16m2_t * op2, vbool8_t op3, vint16m2_t op4, vint16m2_t op5, vint16m2_t op6, const int16_t * op7, vuint64m8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_i16mf2))) +void vloxseg3ei64(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, const int16_t * op3, vuint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_i16mf2_m))) +void vloxseg3ei64(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vbool32_t op3, vint16mf2_t op4, vint16mf2_t op5, vint16mf2_t op6, const int16_t * op7, vuint64m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_i16mf4))) +void vloxseg3ei64(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, const int16_t * op3, vuint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_i16mf4_m))) +void vloxseg3ei64(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vbool64_t op3, vint16mf4_t op4, vint16mf4_t op5, vint16mf4_t op6, const int16_t * op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_u16m1))) +void vloxseg3ei64(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, const uint16_t * op3, vuint64m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_u16m1_m))) +void vloxseg3ei64(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vbool16_t op3, vuint16m1_t op4, vuint16m1_t op5, vuint16m1_t op6, const uint16_t * op7, vuint64m4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_u16m2))) +void vloxseg3ei64(vuint16m2_t * op0, vuint16m2_t * op1, vuint16m2_t * op2, const uint16_t * op3, vuint64m8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_u16m2_m))) +void vloxseg3ei64(vuint16m2_t * op0, vuint16m2_t * op1, vuint16m2_t * op2, vbool8_t op3, vuint16m2_t op4, vuint16m2_t op5, vuint16m2_t op6, const uint16_t * op7, vuint64m8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_u16mf2))) +void vloxseg3ei64(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, const uint16_t * op3, vuint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_u16mf2_m))) +void vloxseg3ei64(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vbool32_t op3, vuint16mf2_t op4, vuint16mf2_t op5, vuint16mf2_t op6, const uint16_t * op7, vuint64m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_u16mf4))) +void vloxseg3ei64(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, const uint16_t * op3, vuint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_u16mf4_m))) +void vloxseg3ei64(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vbool64_t op3, vuint16mf4_t op4, vuint16mf4_t op5, vuint16mf4_t op6, const uint16_t * op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_i16m1))) +void vloxseg4ei64(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, const int16_t * op4, vuint64m4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_i16m1_m))) +void vloxseg4ei64(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vbool16_t op4, vint16m1_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, const int16_t * op9, vuint64m4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_i16m2))) +void vloxseg4ei64(vint16m2_t * op0, vint16m2_t * op1, vint16m2_t * op2, vint16m2_t * op3, const int16_t * op4, vuint64m8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_i16m2_m))) +void vloxseg4ei64(vint16m2_t * op0, vint16m2_t * op1, vint16m2_t * op2, vint16m2_t * op3, vbool8_t op4, vint16m2_t op5, vint16m2_t op6, vint16m2_t op7, vint16m2_t op8, const int16_t * op9, vuint64m8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_i16mf2))) +void vloxseg4ei64(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, const int16_t * op4, vuint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_i16mf2_m))) +void vloxseg4ei64(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vbool32_t op4, vint16mf2_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, const int16_t * op9, vuint64m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_i16mf4))) +void vloxseg4ei64(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, const int16_t * op4, vuint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_i16mf4_m))) +void vloxseg4ei64(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vbool64_t op4, vint16mf4_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, const int16_t * op9, vuint64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_u16m1))) +void vloxseg4ei64(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, const uint16_t * op4, vuint64m4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_u16m1_m))) +void vloxseg4ei64(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vbool16_t op4, vuint16m1_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, const uint16_t * op9, vuint64m4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_u16m2))) +void vloxseg4ei64(vuint16m2_t * op0, vuint16m2_t * op1, vuint16m2_t * op2, vuint16m2_t * op3, const uint16_t * op4, vuint64m8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_u16m2_m))) +void vloxseg4ei64(vuint16m2_t * op0, vuint16m2_t * op1, vuint16m2_t * op2, vuint16m2_t * op3, vbool8_t op4, vuint16m2_t op5, vuint16m2_t op6, vuint16m2_t op7, vuint16m2_t op8, const uint16_t * op9, vuint64m8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_u16mf2))) +void vloxseg4ei64(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, const uint16_t * op4, vuint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_u16mf2_m))) +void vloxseg4ei64(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vbool32_t op4, vuint16mf2_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, const uint16_t * op9, vuint64m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_u16mf4))) +void vloxseg4ei64(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, const uint16_t * op4, vuint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_u16mf4_m))) +void vloxseg4ei64(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vbool64_t op4, vuint16mf4_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, const uint16_t * op9, vuint64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_i16m1))) +void vloxseg5ei64(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, const int16_t * op5, vuint64m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_i16m1_m))) +void vloxseg5ei64(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vbool16_t op5, vint16m1_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, vint16m1_t op10, const int16_t * op11, vuint64m4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_i16mf2))) +void vloxseg5ei64(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, const int16_t * op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_i16mf2_m))) +void vloxseg5ei64(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vbool32_t op5, vint16mf2_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, vint16mf2_t op10, const int16_t * op11, vuint64m2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_i16mf4))) +void vloxseg5ei64(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, const int16_t * op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_i16mf4_m))) +void vloxseg5ei64(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vbool64_t op5, vint16mf4_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, vint16mf4_t op10, const int16_t * op11, vuint64m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_u16m1))) +void vloxseg5ei64(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, const uint16_t * op5, vuint64m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_u16m1_m))) +void vloxseg5ei64(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vbool16_t op5, vuint16m1_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, vuint16m1_t op10, const uint16_t * op11, vuint64m4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_u16mf2))) +void vloxseg5ei64(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, const uint16_t * op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_u16mf2_m))) +void vloxseg5ei64(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vbool32_t op5, vuint16mf2_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, vuint16mf2_t op10, const uint16_t * op11, vuint64m2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_u16mf4))) +void vloxseg5ei64(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, const uint16_t * op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_u16mf4_m))) +void vloxseg5ei64(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vbool64_t op5, vuint16mf4_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, vuint16mf4_t op10, const uint16_t * op11, vuint64m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_i16m1))) +void vloxseg6ei64(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, const int16_t * op6, vuint64m4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_i16m1_m))) +void vloxseg6ei64(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vbool16_t op6, vint16m1_t op7, vint16m1_t op8, vint16m1_t op9, vint16m1_t op10, vint16m1_t op11, vint16m1_t op12, const int16_t * op13, vuint64m4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_i16mf2))) +void vloxseg6ei64(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, const int16_t * op6, vuint64m2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_i16mf2_m))) +void vloxseg6ei64(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vbool32_t op6, vint16mf2_t op7, vint16mf2_t op8, vint16mf2_t op9, vint16mf2_t op10, vint16mf2_t op11, vint16mf2_t op12, const int16_t * op13, vuint64m2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_i16mf4))) +void vloxseg6ei64(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, const int16_t * op6, vuint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_i16mf4_m))) +void vloxseg6ei64(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vbool64_t op6, vint16mf4_t op7, vint16mf4_t op8, vint16mf4_t op9, vint16mf4_t op10, vint16mf4_t op11, vint16mf4_t op12, const int16_t * op13, vuint64m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_u16m1))) +void vloxseg6ei64(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, const uint16_t * op6, vuint64m4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_u16m1_m))) +void vloxseg6ei64(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vbool16_t op6, vuint16m1_t op7, vuint16m1_t op8, vuint16m1_t op9, vuint16m1_t op10, vuint16m1_t op11, vuint16m1_t op12, const uint16_t * op13, vuint64m4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_u16mf2))) +void vloxseg6ei64(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, const uint16_t * op6, vuint64m2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_u16mf2_m))) +void vloxseg6ei64(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vbool32_t op6, vuint16mf2_t op7, vuint16mf2_t op8, vuint16mf2_t op9, vuint16mf2_t op10, vuint16mf2_t op11, vuint16mf2_t op12, const uint16_t * op13, vuint64m2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_u16mf4))) +void vloxseg6ei64(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, const uint16_t * op6, vuint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_u16mf4_m))) +void vloxseg6ei64(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vbool64_t op6, vuint16mf4_t op7, vuint16mf4_t op8, vuint16mf4_t op9, vuint16mf4_t op10, vuint16mf4_t op11, vuint16mf4_t op12, const uint16_t * op13, vuint64m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_i16m1))) +void vloxseg7ei64(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vint16m1_t * op6, const int16_t * op7, vuint64m4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_i16m1_m))) +void vloxseg7ei64(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vint16m1_t * op6, vbool16_t op7, vint16m1_t op8, vint16m1_t op9, vint16m1_t op10, vint16m1_t op11, vint16m1_t op12, vint16m1_t op13, vint16m1_t op14, const int16_t * op15, vuint64m4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_i16mf2))) +void vloxseg7ei64(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vint16mf2_t * op6, const int16_t * op7, vuint64m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_i16mf2_m))) +void vloxseg7ei64(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vint16mf2_t * op6, vbool32_t op7, vint16mf2_t op8, vint16mf2_t op9, vint16mf2_t op10, vint16mf2_t op11, vint16mf2_t op12, vint16mf2_t op13, vint16mf2_t op14, const int16_t * op15, vuint64m2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_i16mf4))) +void vloxseg7ei64(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vint16mf4_t * op6, const int16_t * op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_i16mf4_m))) +void vloxseg7ei64(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vint16mf4_t * op6, vbool64_t op7, vint16mf4_t op8, vint16mf4_t op9, vint16mf4_t op10, vint16mf4_t op11, vint16mf4_t op12, vint16mf4_t op13, vint16mf4_t op14, const int16_t * op15, vuint64m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_u16m1))) +void vloxseg7ei64(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vuint16m1_t * op6, const uint16_t * op7, vuint64m4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_u16m1_m))) +void vloxseg7ei64(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vuint16m1_t * op6, vbool16_t op7, vuint16m1_t op8, vuint16m1_t op9, vuint16m1_t op10, vuint16m1_t op11, vuint16m1_t op12, vuint16m1_t op13, vuint16m1_t op14, const uint16_t * op15, vuint64m4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_u16mf2))) +void vloxseg7ei64(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vuint16mf2_t * op6, const uint16_t * op7, vuint64m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_u16mf2_m))) +void vloxseg7ei64(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vuint16mf2_t * op6, vbool32_t op7, vuint16mf2_t op8, vuint16mf2_t op9, vuint16mf2_t op10, vuint16mf2_t op11, vuint16mf2_t op12, vuint16mf2_t op13, vuint16mf2_t op14, const uint16_t * op15, vuint64m2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_u16mf4))) +void vloxseg7ei64(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vuint16mf4_t * op6, const uint16_t * op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_u16mf4_m))) +void vloxseg7ei64(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vuint16mf4_t * op6, vbool64_t op7, vuint16mf4_t op8, vuint16mf4_t op9, vuint16mf4_t op10, vuint16mf4_t op11, vuint16mf4_t op12, vuint16mf4_t op13, vuint16mf4_t op14, const uint16_t * op15, vuint64m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_i16m1))) +void vloxseg8ei64(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vint16m1_t * op6, vint16m1_t * op7, const int16_t * op8, vuint64m4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_i16m1_m))) +void vloxseg8ei64(vint16m1_t * op0, vint16m1_t * op1, vint16m1_t * op2, vint16m1_t * op3, vint16m1_t * op4, vint16m1_t * op5, vint16m1_t * op6, vint16m1_t * op7, vbool16_t op8, vint16m1_t op9, vint16m1_t op10, vint16m1_t op11, vint16m1_t op12, vint16m1_t op13, vint16m1_t op14, vint16m1_t op15, vint16m1_t op16, const int16_t * op17, vuint64m4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_i16mf2))) +void vloxseg8ei64(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vint16mf2_t * op6, vint16mf2_t * op7, const int16_t * op8, vuint64m2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_i16mf2_m))) +void vloxseg8ei64(vint16mf2_t * op0, vint16mf2_t * op1, vint16mf2_t * op2, vint16mf2_t * op3, vint16mf2_t * op4, vint16mf2_t * op5, vint16mf2_t * op6, vint16mf2_t * op7, vbool32_t op8, vint16mf2_t op9, vint16mf2_t op10, vint16mf2_t op11, vint16mf2_t op12, vint16mf2_t op13, vint16mf2_t op14, vint16mf2_t op15, vint16mf2_t op16, const int16_t * op17, vuint64m2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_i16mf4))) +void vloxseg8ei64(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vint16mf4_t * op6, vint16mf4_t * op7, const int16_t * op8, vuint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_i16mf4_m))) +void vloxseg8ei64(vint16mf4_t * op0, vint16mf4_t * op1, vint16mf4_t * op2, vint16mf4_t * op3, vint16mf4_t * op4, vint16mf4_t * op5, vint16mf4_t * op6, vint16mf4_t * op7, vbool64_t op8, vint16mf4_t op9, vint16mf4_t op10, vint16mf4_t op11, vint16mf4_t op12, vint16mf4_t op13, vint16mf4_t op14, vint16mf4_t op15, vint16mf4_t op16, const int16_t * op17, vuint64m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_u16m1))) +void vloxseg8ei64(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vuint16m1_t * op6, vuint16m1_t * op7, const uint16_t * op8, vuint64m4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_u16m1_m))) +void vloxseg8ei64(vuint16m1_t * op0, vuint16m1_t * op1, vuint16m1_t * op2, vuint16m1_t * op3, vuint16m1_t * op4, vuint16m1_t * op5, vuint16m1_t * op6, vuint16m1_t * op7, vbool16_t op8, vuint16m1_t op9, vuint16m1_t op10, vuint16m1_t op11, vuint16m1_t op12, vuint16m1_t op13, vuint16m1_t op14, vuint16m1_t op15, vuint16m1_t op16, const uint16_t * op17, vuint64m4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_u16mf2))) +void vloxseg8ei64(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vuint16mf2_t * op6, vuint16mf2_t * op7, const uint16_t * op8, vuint64m2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_u16mf2_m))) +void vloxseg8ei64(vuint16mf2_t * op0, vuint16mf2_t * op1, vuint16mf2_t * op2, vuint16mf2_t * op3, vuint16mf2_t * op4, vuint16mf2_t * op5, vuint16mf2_t * op6, vuint16mf2_t * op7, vbool32_t op8, vuint16mf2_t op9, vuint16mf2_t op10, vuint16mf2_t op11, vuint16mf2_t op12, vuint16mf2_t op13, vuint16mf2_t op14, vuint16mf2_t op15, vuint16mf2_t op16, const uint16_t * op17, vuint64m2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_u16mf4))) +void vloxseg8ei64(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vuint16mf4_t * op6, vuint16mf4_t * op7, const uint16_t * op8, vuint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_u16mf4_m))) +void vloxseg8ei64(vuint16mf4_t * op0, vuint16mf4_t * op1, vuint16mf4_t * op2, vuint16mf4_t * op3, vuint16mf4_t * op4, vuint16mf4_t * op5, vuint16mf4_t * op6, vuint16mf4_t * op7, vbool64_t op8, vuint16mf4_t op9, vuint16mf4_t op10, vuint16mf4_t op11, vuint16mf4_t op12, vuint16mf4_t op13, vuint16mf4_t op14, vuint16mf4_t op15, vuint16mf4_t op16, const uint16_t * op17, vuint64m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_i32m1))) +void vloxseg2ei8(vint32m1_t * op0, vint32m1_t * op1, const int32_t * op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_i32m1_m))) +void vloxseg2ei8(vint32m1_t * op0, vint32m1_t * op1, vbool32_t op2, vint32m1_t op3, vint32m1_t op4, const int32_t * op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_i32m2))) +void vloxseg2ei8(vint32m2_t * op0, vint32m2_t * op1, const int32_t * op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_i32m2_m))) +void vloxseg2ei8(vint32m2_t * op0, vint32m2_t * op1, vbool16_t op2, vint32m2_t op3, vint32m2_t op4, const int32_t * op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_i32m4))) +void vloxseg2ei8(vint32m4_t * op0, vint32m4_t * op1, const int32_t * op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_i32m4_m))) +void vloxseg2ei8(vint32m4_t * op0, vint32m4_t * op1, vbool8_t op2, vint32m4_t op3, vint32m4_t op4, const int32_t * op5, vuint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_i32mf2))) +void vloxseg2ei8(vint32mf2_t * op0, vint32mf2_t * op1, const int32_t * op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_i32mf2_m))) +void vloxseg2ei8(vint32mf2_t * op0, vint32mf2_t * op1, vbool64_t op2, vint32mf2_t op3, vint32mf2_t op4, const int32_t * op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_u32m1))) +void vloxseg2ei8(vuint32m1_t * op0, vuint32m1_t * op1, const uint32_t * op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_u32m1_m))) +void vloxseg2ei8(vuint32m1_t * op0, vuint32m1_t * op1, vbool32_t op2, vuint32m1_t op3, vuint32m1_t op4, const uint32_t * op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_u32m2))) +void vloxseg2ei8(vuint32m2_t * op0, vuint32m2_t * op1, const uint32_t * op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_u32m2_m))) +void vloxseg2ei8(vuint32m2_t * op0, vuint32m2_t * op1, vbool16_t op2, vuint32m2_t op3, vuint32m2_t op4, const uint32_t * op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_u32m4))) +void vloxseg2ei8(vuint32m4_t * op0, vuint32m4_t * op1, const uint32_t * op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_u32m4_m))) +void vloxseg2ei8(vuint32m4_t * op0, vuint32m4_t * op1, vbool8_t op2, vuint32m4_t op3, vuint32m4_t op4, const uint32_t * op5, vuint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_u32mf2))) +void vloxseg2ei8(vuint32mf2_t * op0, vuint32mf2_t * op1, const uint32_t * op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_u32mf2_m))) +void vloxseg2ei8(vuint32mf2_t * op0, vuint32mf2_t * op1, vbool64_t op2, vuint32mf2_t op3, vuint32mf2_t op4, const uint32_t * op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_i32m1))) +void vloxseg3ei8(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, const int32_t * op3, vuint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_i32m1_m))) +void vloxseg3ei8(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vbool32_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, const int32_t * op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_i32m2))) +void vloxseg3ei8(vint32m2_t * op0, vint32m2_t * op1, vint32m2_t * op2, const int32_t * op3, vuint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_i32m2_m))) +void vloxseg3ei8(vint32m2_t * op0, vint32m2_t * op1, vint32m2_t * op2, vbool16_t op3, vint32m2_t op4, vint32m2_t op5, vint32m2_t op6, const int32_t * op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_i32mf2))) +void vloxseg3ei8(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, const int32_t * op3, vuint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_i32mf2_m))) +void vloxseg3ei8(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vbool64_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, const int32_t * op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_u32m1))) +void vloxseg3ei8(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, const uint32_t * op3, vuint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_u32m1_m))) +void vloxseg3ei8(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vbool32_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, const uint32_t * op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_u32m2))) +void vloxseg3ei8(vuint32m2_t * op0, vuint32m2_t * op1, vuint32m2_t * op2, const uint32_t * op3, vuint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_u32m2_m))) +void vloxseg3ei8(vuint32m2_t * op0, vuint32m2_t * op1, vuint32m2_t * op2, vbool16_t op3, vuint32m2_t op4, vuint32m2_t op5, vuint32m2_t op6, const uint32_t * op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_u32mf2))) +void vloxseg3ei8(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, const uint32_t * op3, vuint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_u32mf2_m))) +void vloxseg3ei8(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vbool64_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, const uint32_t * op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_i32m1))) +void vloxseg4ei8(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, const int32_t * op4, vuint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_i32m1_m))) +void vloxseg4ei8(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vbool32_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, const int32_t * op9, vuint8mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_i32m2))) +void vloxseg4ei8(vint32m2_t * op0, vint32m2_t * op1, vint32m2_t * op2, vint32m2_t * op3, const int32_t * op4, vuint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_i32m2_m))) +void vloxseg4ei8(vint32m2_t * op0, vint32m2_t * op1, vint32m2_t * op2, vint32m2_t * op3, vbool16_t op4, vint32m2_t op5, vint32m2_t op6, vint32m2_t op7, vint32m2_t op8, const int32_t * op9, vuint8mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_i32mf2))) +void vloxseg4ei8(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, const int32_t * op4, vuint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_i32mf2_m))) +void vloxseg4ei8(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vbool64_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, const int32_t * op9, vuint8mf8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_u32m1))) +void vloxseg4ei8(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, const uint32_t * op4, vuint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_u32m1_m))) +void vloxseg4ei8(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vbool32_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, const uint32_t * op9, vuint8mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_u32m2))) +void vloxseg4ei8(vuint32m2_t * op0, vuint32m2_t * op1, vuint32m2_t * op2, vuint32m2_t * op3, const uint32_t * op4, vuint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_u32m2_m))) +void vloxseg4ei8(vuint32m2_t * op0, vuint32m2_t * op1, vuint32m2_t * op2, vuint32m2_t * op3, vbool16_t op4, vuint32m2_t op5, vuint32m2_t op6, vuint32m2_t op7, vuint32m2_t op8, const uint32_t * op9, vuint8mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_u32mf2))) +void vloxseg4ei8(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, const uint32_t * op4, vuint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_u32mf2_m))) +void vloxseg4ei8(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vbool64_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, const uint32_t * op9, vuint8mf8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_i32m1))) +void vloxseg5ei8(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, const int32_t * op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_i32m1_m))) +void vloxseg5ei8(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vbool32_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, vint32m1_t op10, const int32_t * op11, vuint8mf4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_i32mf2))) +void vloxseg5ei8(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, const int32_t * op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_i32mf2_m))) +void vloxseg5ei8(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vbool64_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, vint32mf2_t op10, const int32_t * op11, vuint8mf8_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_u32m1))) +void vloxseg5ei8(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, const uint32_t * op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_u32m1_m))) +void vloxseg5ei8(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vbool32_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, vuint32m1_t op10, const uint32_t * op11, vuint8mf4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_u32mf2))) +void vloxseg5ei8(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, const uint32_t * op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_u32mf2_m))) +void vloxseg5ei8(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vbool64_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, vuint32mf2_t op10, const uint32_t * op11, vuint8mf8_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_i32m1))) +void vloxseg6ei8(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, const int32_t * op6, vuint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_i32m1_m))) +void vloxseg6ei8(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vbool32_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, vint32m1_t op10, vint32m1_t op11, vint32m1_t op12, const int32_t * op13, vuint8mf4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_i32mf2))) +void vloxseg6ei8(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, const int32_t * op6, vuint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_i32mf2_m))) +void vloxseg6ei8(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vbool64_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, vint32mf2_t op10, vint32mf2_t op11, vint32mf2_t op12, const int32_t * op13, vuint8mf8_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_u32m1))) +void vloxseg6ei8(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, const uint32_t * op6, vuint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_u32m1_m))) +void vloxseg6ei8(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vbool32_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, vuint32m1_t op10, vuint32m1_t op11, vuint32m1_t op12, const uint32_t * op13, vuint8mf4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_u32mf2))) +void vloxseg6ei8(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, const uint32_t * op6, vuint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_u32mf2_m))) +void vloxseg6ei8(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vbool64_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, vuint32mf2_t op10, vuint32mf2_t op11, vuint32mf2_t op12, const uint32_t * op13, vuint8mf8_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_i32m1))) +void vloxseg7ei8(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vint32m1_t * op6, const int32_t * op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_i32m1_m))) +void vloxseg7ei8(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vint32m1_t * op6, vbool32_t op7, vint32m1_t op8, vint32m1_t op9, vint32m1_t op10, vint32m1_t op11, vint32m1_t op12, vint32m1_t op13, vint32m1_t op14, const int32_t * op15, vuint8mf4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_i32mf2))) +void vloxseg7ei8(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vint32mf2_t * op6, const int32_t * op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_i32mf2_m))) +void vloxseg7ei8(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vint32mf2_t * op6, vbool64_t op7, vint32mf2_t op8, vint32mf2_t op9, vint32mf2_t op10, vint32mf2_t op11, vint32mf2_t op12, vint32mf2_t op13, vint32mf2_t op14, const int32_t * op15, vuint8mf8_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_u32m1))) +void vloxseg7ei8(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vuint32m1_t * op6, const uint32_t * op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_u32m1_m))) +void vloxseg7ei8(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vuint32m1_t * op6, vbool32_t op7, vuint32m1_t op8, vuint32m1_t op9, vuint32m1_t op10, vuint32m1_t op11, vuint32m1_t op12, vuint32m1_t op13, vuint32m1_t op14, const uint32_t * op15, vuint8mf4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_u32mf2))) +void vloxseg7ei8(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vuint32mf2_t * op6, const uint32_t * op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_u32mf2_m))) +void vloxseg7ei8(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vuint32mf2_t * op6, vbool64_t op7, vuint32mf2_t op8, vuint32mf2_t op9, vuint32mf2_t op10, vuint32mf2_t op11, vuint32mf2_t op12, vuint32mf2_t op13, vuint32mf2_t op14, const uint32_t * op15, vuint8mf8_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_i32m1))) +void vloxseg8ei8(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vint32m1_t * op6, vint32m1_t * op7, const int32_t * op8, vuint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_i32m1_m))) +void vloxseg8ei8(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vint32m1_t * op6, vint32m1_t * op7, vbool32_t op8, vint32m1_t op9, vint32m1_t op10, vint32m1_t op11, vint32m1_t op12, vint32m1_t op13, vint32m1_t op14, vint32m1_t op15, vint32m1_t op16, const int32_t * op17, vuint8mf4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_i32mf2))) +void vloxseg8ei8(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vint32mf2_t * op6, vint32mf2_t * op7, const int32_t * op8, vuint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_i32mf2_m))) +void vloxseg8ei8(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vint32mf2_t * op6, vint32mf2_t * op7, vbool64_t op8, vint32mf2_t op9, vint32mf2_t op10, vint32mf2_t op11, vint32mf2_t op12, vint32mf2_t op13, vint32mf2_t op14, vint32mf2_t op15, vint32mf2_t op16, const int32_t * op17, vuint8mf8_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_u32m1))) +void vloxseg8ei8(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vuint32m1_t * op6, vuint32m1_t * op7, const uint32_t * op8, vuint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_u32m1_m))) +void vloxseg8ei8(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vuint32m1_t * op6, vuint32m1_t * op7, vbool32_t op8, vuint32m1_t op9, vuint32m1_t op10, vuint32m1_t op11, vuint32m1_t op12, vuint32m1_t op13, vuint32m1_t op14, vuint32m1_t op15, vuint32m1_t op16, const uint32_t * op17, vuint8mf4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_u32mf2))) +void vloxseg8ei8(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vuint32mf2_t * op6, vuint32mf2_t * op7, const uint32_t * op8, vuint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_u32mf2_m))) +void vloxseg8ei8(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vuint32mf2_t * op6, vuint32mf2_t * op7, vbool64_t op8, vuint32mf2_t op9, vuint32mf2_t op10, vuint32mf2_t op11, vuint32mf2_t op12, vuint32mf2_t op13, vuint32mf2_t op14, vuint32mf2_t op15, vuint32mf2_t op16, const uint32_t * op17, vuint8mf8_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_i32m1))) +void vloxseg2ei16(vint32m1_t * op0, vint32m1_t * op1, const int32_t * op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_i32m1_m))) +void vloxseg2ei16(vint32m1_t * op0, vint32m1_t * op1, vbool32_t op2, vint32m1_t op3, vint32m1_t op4, const int32_t * op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_i32m2))) +void vloxseg2ei16(vint32m2_t * op0, vint32m2_t * op1, const int32_t * op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_i32m2_m))) +void vloxseg2ei16(vint32m2_t * op0, vint32m2_t * op1, vbool16_t op2, vint32m2_t op3, vint32m2_t op4, const int32_t * op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_i32m4))) +void vloxseg2ei16(vint32m4_t * op0, vint32m4_t * op1, const int32_t * op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_i32m4_m))) +void vloxseg2ei16(vint32m4_t * op0, vint32m4_t * op1, vbool8_t op2, vint32m4_t op3, vint32m4_t op4, const int32_t * op5, vuint16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_i32mf2))) +void vloxseg2ei16(vint32mf2_t * op0, vint32mf2_t * op1, const int32_t * op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_i32mf2_m))) +void vloxseg2ei16(vint32mf2_t * op0, vint32mf2_t * op1, vbool64_t op2, vint32mf2_t op3, vint32mf2_t op4, const int32_t * op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_u32m1))) +void vloxseg2ei16(vuint32m1_t * op0, vuint32m1_t * op1, const uint32_t * op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_u32m1_m))) +void vloxseg2ei16(vuint32m1_t * op0, vuint32m1_t * op1, vbool32_t op2, vuint32m1_t op3, vuint32m1_t op4, const uint32_t * op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_u32m2))) +void vloxseg2ei16(vuint32m2_t * op0, vuint32m2_t * op1, const uint32_t * op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_u32m2_m))) +void vloxseg2ei16(vuint32m2_t * op0, vuint32m2_t * op1, vbool16_t op2, vuint32m2_t op3, vuint32m2_t op4, const uint32_t * op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_u32m4))) +void vloxseg2ei16(vuint32m4_t * op0, vuint32m4_t * op1, const uint32_t * op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_u32m4_m))) +void vloxseg2ei16(vuint32m4_t * op0, vuint32m4_t * op1, vbool8_t op2, vuint32m4_t op3, vuint32m4_t op4, const uint32_t * op5, vuint16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_u32mf2))) +void vloxseg2ei16(vuint32mf2_t * op0, vuint32mf2_t * op1, const uint32_t * op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_u32mf2_m))) +void vloxseg2ei16(vuint32mf2_t * op0, vuint32mf2_t * op1, vbool64_t op2, vuint32mf2_t op3, vuint32mf2_t op4, const uint32_t * op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_i32m1))) +void vloxseg3ei16(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, const int32_t * op3, vuint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_i32m1_m))) +void vloxseg3ei16(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vbool32_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, const int32_t * op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_i32m2))) +void vloxseg3ei16(vint32m2_t * op0, vint32m2_t * op1, vint32m2_t * op2, const int32_t * op3, vuint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_i32m2_m))) +void vloxseg3ei16(vint32m2_t * op0, vint32m2_t * op1, vint32m2_t * op2, vbool16_t op3, vint32m2_t op4, vint32m2_t op5, vint32m2_t op6, const int32_t * op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_i32mf2))) +void vloxseg3ei16(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, const int32_t * op3, vuint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_i32mf2_m))) +void vloxseg3ei16(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vbool64_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, const int32_t * op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_u32m1))) +void vloxseg3ei16(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, const uint32_t * op3, vuint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_u32m1_m))) +void vloxseg3ei16(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vbool32_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, const uint32_t * op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_u32m2))) +void vloxseg3ei16(vuint32m2_t * op0, vuint32m2_t * op1, vuint32m2_t * op2, const uint32_t * op3, vuint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_u32m2_m))) +void vloxseg3ei16(vuint32m2_t * op0, vuint32m2_t * op1, vuint32m2_t * op2, vbool16_t op3, vuint32m2_t op4, vuint32m2_t op5, vuint32m2_t op6, const uint32_t * op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_u32mf2))) +void vloxseg3ei16(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, const uint32_t * op3, vuint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_u32mf2_m))) +void vloxseg3ei16(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vbool64_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, const uint32_t * op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_i32m1))) +void vloxseg4ei16(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, const int32_t * op4, vuint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_i32m1_m))) +void vloxseg4ei16(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vbool32_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, const int32_t * op9, vuint16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_i32m2))) +void vloxseg4ei16(vint32m2_t * op0, vint32m2_t * op1, vint32m2_t * op2, vint32m2_t * op3, const int32_t * op4, vuint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_i32m2_m))) +void vloxseg4ei16(vint32m2_t * op0, vint32m2_t * op1, vint32m2_t * op2, vint32m2_t * op3, vbool16_t op4, vint32m2_t op5, vint32m2_t op6, vint32m2_t op7, vint32m2_t op8, const int32_t * op9, vuint16m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_i32mf2))) +void vloxseg4ei16(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, const int32_t * op4, vuint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_i32mf2_m))) +void vloxseg4ei16(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vbool64_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, const int32_t * op9, vuint16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_u32m1))) +void vloxseg4ei16(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, const uint32_t * op4, vuint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_u32m1_m))) +void vloxseg4ei16(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vbool32_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, const uint32_t * op9, vuint16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_u32m2))) +void vloxseg4ei16(vuint32m2_t * op0, vuint32m2_t * op1, vuint32m2_t * op2, vuint32m2_t * op3, const uint32_t * op4, vuint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_u32m2_m))) +void vloxseg4ei16(vuint32m2_t * op0, vuint32m2_t * op1, vuint32m2_t * op2, vuint32m2_t * op3, vbool16_t op4, vuint32m2_t op5, vuint32m2_t op6, vuint32m2_t op7, vuint32m2_t op8, const uint32_t * op9, vuint16m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_u32mf2))) +void vloxseg4ei16(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, const uint32_t * op4, vuint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_u32mf2_m))) +void vloxseg4ei16(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vbool64_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, const uint32_t * op9, vuint16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_i32m1))) +void vloxseg5ei16(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, const int32_t * op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_i32m1_m))) +void vloxseg5ei16(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vbool32_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, vint32m1_t op10, const int32_t * op11, vuint16mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_i32mf2))) +void vloxseg5ei16(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, const int32_t * op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_i32mf2_m))) +void vloxseg5ei16(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vbool64_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, vint32mf2_t op10, const int32_t * op11, vuint16mf4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_u32m1))) +void vloxseg5ei16(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, const uint32_t * op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_u32m1_m))) +void vloxseg5ei16(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vbool32_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, vuint32m1_t op10, const uint32_t * op11, vuint16mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_u32mf2))) +void vloxseg5ei16(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, const uint32_t * op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_u32mf2_m))) +void vloxseg5ei16(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vbool64_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, vuint32mf2_t op10, const uint32_t * op11, vuint16mf4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_i32m1))) +void vloxseg6ei16(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, const int32_t * op6, vuint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_i32m1_m))) +void vloxseg6ei16(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vbool32_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, vint32m1_t op10, vint32m1_t op11, vint32m1_t op12, const int32_t * op13, vuint16mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_i32mf2))) +void vloxseg6ei16(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, const int32_t * op6, vuint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_i32mf2_m))) +void vloxseg6ei16(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vbool64_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, vint32mf2_t op10, vint32mf2_t op11, vint32mf2_t op12, const int32_t * op13, vuint16mf4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_u32m1))) +void vloxseg6ei16(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, const uint32_t * op6, vuint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_u32m1_m))) +void vloxseg6ei16(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vbool32_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, vuint32m1_t op10, vuint32m1_t op11, vuint32m1_t op12, const uint32_t * op13, vuint16mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_u32mf2))) +void vloxseg6ei16(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, const uint32_t * op6, vuint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_u32mf2_m))) +void vloxseg6ei16(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vbool64_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, vuint32mf2_t op10, vuint32mf2_t op11, vuint32mf2_t op12, const uint32_t * op13, vuint16mf4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_i32m1))) +void vloxseg7ei16(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vint32m1_t * op6, const int32_t * op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_i32m1_m))) +void vloxseg7ei16(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vint32m1_t * op6, vbool32_t op7, vint32m1_t op8, vint32m1_t op9, vint32m1_t op10, vint32m1_t op11, vint32m1_t op12, vint32m1_t op13, vint32m1_t op14, const int32_t * op15, vuint16mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_i32mf2))) +void vloxseg7ei16(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vint32mf2_t * op6, const int32_t * op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_i32mf2_m))) +void vloxseg7ei16(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vint32mf2_t * op6, vbool64_t op7, vint32mf2_t op8, vint32mf2_t op9, vint32mf2_t op10, vint32mf2_t op11, vint32mf2_t op12, vint32mf2_t op13, vint32mf2_t op14, const int32_t * op15, vuint16mf4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_u32m1))) +void vloxseg7ei16(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vuint32m1_t * op6, const uint32_t * op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_u32m1_m))) +void vloxseg7ei16(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vuint32m1_t * op6, vbool32_t op7, vuint32m1_t op8, vuint32m1_t op9, vuint32m1_t op10, vuint32m1_t op11, vuint32m1_t op12, vuint32m1_t op13, vuint32m1_t op14, const uint32_t * op15, vuint16mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_u32mf2))) +void vloxseg7ei16(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vuint32mf2_t * op6, const uint32_t * op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_u32mf2_m))) +void vloxseg7ei16(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vuint32mf2_t * op6, vbool64_t op7, vuint32mf2_t op8, vuint32mf2_t op9, vuint32mf2_t op10, vuint32mf2_t op11, vuint32mf2_t op12, vuint32mf2_t op13, vuint32mf2_t op14, const uint32_t * op15, vuint16mf4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_i32m1))) +void vloxseg8ei16(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vint32m1_t * op6, vint32m1_t * op7, const int32_t * op8, vuint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_i32m1_m))) +void vloxseg8ei16(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vint32m1_t * op6, vint32m1_t * op7, vbool32_t op8, vint32m1_t op9, vint32m1_t op10, vint32m1_t op11, vint32m1_t op12, vint32m1_t op13, vint32m1_t op14, vint32m1_t op15, vint32m1_t op16, const int32_t * op17, vuint16mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_i32mf2))) +void vloxseg8ei16(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vint32mf2_t * op6, vint32mf2_t * op7, const int32_t * op8, vuint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_i32mf2_m))) +void vloxseg8ei16(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vint32mf2_t * op6, vint32mf2_t * op7, vbool64_t op8, vint32mf2_t op9, vint32mf2_t op10, vint32mf2_t op11, vint32mf2_t op12, vint32mf2_t op13, vint32mf2_t op14, vint32mf2_t op15, vint32mf2_t op16, const int32_t * op17, vuint16mf4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_u32m1))) +void vloxseg8ei16(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vuint32m1_t * op6, vuint32m1_t * op7, const uint32_t * op8, vuint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_u32m1_m))) +void vloxseg8ei16(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vuint32m1_t * op6, vuint32m1_t * op7, vbool32_t op8, vuint32m1_t op9, vuint32m1_t op10, vuint32m1_t op11, vuint32m1_t op12, vuint32m1_t op13, vuint32m1_t op14, vuint32m1_t op15, vuint32m1_t op16, const uint32_t * op17, vuint16mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_u32mf2))) +void vloxseg8ei16(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vuint32mf2_t * op6, vuint32mf2_t * op7, const uint32_t * op8, vuint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_u32mf2_m))) +void vloxseg8ei16(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vuint32mf2_t * op6, vuint32mf2_t * op7, vbool64_t op8, vuint32mf2_t op9, vuint32mf2_t op10, vuint32mf2_t op11, vuint32mf2_t op12, vuint32mf2_t op13, vuint32mf2_t op14, vuint32mf2_t op15, vuint32mf2_t op16, const uint32_t * op17, vuint16mf4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_i32m1))) +void vloxseg2ei32(vint32m1_t * op0, vint32m1_t * op1, const int32_t * op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_i32m1_m))) +void vloxseg2ei32(vint32m1_t * op0, vint32m1_t * op1, vbool32_t op2, vint32m1_t op3, vint32m1_t op4, const int32_t * op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_i32m2))) +void vloxseg2ei32(vint32m2_t * op0, vint32m2_t * op1, const int32_t * op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_i32m2_m))) +void vloxseg2ei32(vint32m2_t * op0, vint32m2_t * op1, vbool16_t op2, vint32m2_t op3, vint32m2_t op4, const int32_t * op5, vuint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_i32m4))) +void vloxseg2ei32(vint32m4_t * op0, vint32m4_t * op1, const int32_t * op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_i32m4_m))) +void vloxseg2ei32(vint32m4_t * op0, vint32m4_t * op1, vbool8_t op2, vint32m4_t op3, vint32m4_t op4, const int32_t * op5, vuint32m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_i32mf2))) +void vloxseg2ei32(vint32mf2_t * op0, vint32mf2_t * op1, const int32_t * op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_i32mf2_m))) +void vloxseg2ei32(vint32mf2_t * op0, vint32mf2_t * op1, vbool64_t op2, vint32mf2_t op3, vint32mf2_t op4, const int32_t * op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_u32m1))) +void vloxseg2ei32(vuint32m1_t * op0, vuint32m1_t * op1, const uint32_t * op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_u32m1_m))) +void vloxseg2ei32(vuint32m1_t * op0, vuint32m1_t * op1, vbool32_t op2, vuint32m1_t op3, vuint32m1_t op4, const uint32_t * op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_u32m2))) +void vloxseg2ei32(vuint32m2_t * op0, vuint32m2_t * op1, const uint32_t * op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_u32m2_m))) +void vloxseg2ei32(vuint32m2_t * op0, vuint32m2_t * op1, vbool16_t op2, vuint32m2_t op3, vuint32m2_t op4, const uint32_t * op5, vuint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_u32m4))) +void vloxseg2ei32(vuint32m4_t * op0, vuint32m4_t * op1, const uint32_t * op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_u32m4_m))) +void vloxseg2ei32(vuint32m4_t * op0, vuint32m4_t * op1, vbool8_t op2, vuint32m4_t op3, vuint32m4_t op4, const uint32_t * op5, vuint32m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_u32mf2))) +void vloxseg2ei32(vuint32mf2_t * op0, vuint32mf2_t * op1, const uint32_t * op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_u32mf2_m))) +void vloxseg2ei32(vuint32mf2_t * op0, vuint32mf2_t * op1, vbool64_t op2, vuint32mf2_t op3, vuint32mf2_t op4, const uint32_t * op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_i32m1))) +void vloxseg3ei32(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, const int32_t * op3, vuint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_i32m1_m))) +void vloxseg3ei32(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vbool32_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, const int32_t * op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_i32m2))) +void vloxseg3ei32(vint32m2_t * op0, vint32m2_t * op1, vint32m2_t * op2, const int32_t * op3, vuint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_i32m2_m))) +void vloxseg3ei32(vint32m2_t * op0, vint32m2_t * op1, vint32m2_t * op2, vbool16_t op3, vint32m2_t op4, vint32m2_t op5, vint32m2_t op6, const int32_t * op7, vuint32m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_i32mf2))) +void vloxseg3ei32(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, const int32_t * op3, vuint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_i32mf2_m))) +void vloxseg3ei32(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vbool64_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, const int32_t * op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_u32m1))) +void vloxseg3ei32(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, const uint32_t * op3, vuint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_u32m1_m))) +void vloxseg3ei32(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vbool32_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, const uint32_t * op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_u32m2))) +void vloxseg3ei32(vuint32m2_t * op0, vuint32m2_t * op1, vuint32m2_t * op2, const uint32_t * op3, vuint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_u32m2_m))) +void vloxseg3ei32(vuint32m2_t * op0, vuint32m2_t * op1, vuint32m2_t * op2, vbool16_t op3, vuint32m2_t op4, vuint32m2_t op5, vuint32m2_t op6, const uint32_t * op7, vuint32m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_u32mf2))) +void vloxseg3ei32(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, const uint32_t * op3, vuint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_u32mf2_m))) +void vloxseg3ei32(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vbool64_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, const uint32_t * op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_i32m1))) +void vloxseg4ei32(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, const int32_t * op4, vuint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_i32m1_m))) +void vloxseg4ei32(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vbool32_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, const int32_t * op9, vuint32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_i32m2))) +void vloxseg4ei32(vint32m2_t * op0, vint32m2_t * op1, vint32m2_t * op2, vint32m2_t * op3, const int32_t * op4, vuint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_i32m2_m))) +void vloxseg4ei32(vint32m2_t * op0, vint32m2_t * op1, vint32m2_t * op2, vint32m2_t * op3, vbool16_t op4, vint32m2_t op5, vint32m2_t op6, vint32m2_t op7, vint32m2_t op8, const int32_t * op9, vuint32m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_i32mf2))) +void vloxseg4ei32(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, const int32_t * op4, vuint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_i32mf2_m))) +void vloxseg4ei32(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vbool64_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, const int32_t * op9, vuint32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_u32m1))) +void vloxseg4ei32(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, const uint32_t * op4, vuint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_u32m1_m))) +void vloxseg4ei32(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vbool32_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, const uint32_t * op9, vuint32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_u32m2))) +void vloxseg4ei32(vuint32m2_t * op0, vuint32m2_t * op1, vuint32m2_t * op2, vuint32m2_t * op3, const uint32_t * op4, vuint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_u32m2_m))) +void vloxseg4ei32(vuint32m2_t * op0, vuint32m2_t * op1, vuint32m2_t * op2, vuint32m2_t * op3, vbool16_t op4, vuint32m2_t op5, vuint32m2_t op6, vuint32m2_t op7, vuint32m2_t op8, const uint32_t * op9, vuint32m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_u32mf2))) +void vloxseg4ei32(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, const uint32_t * op4, vuint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_u32mf2_m))) +void vloxseg4ei32(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vbool64_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, const uint32_t * op9, vuint32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_i32m1))) +void vloxseg5ei32(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, const int32_t * op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_i32m1_m))) +void vloxseg5ei32(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vbool32_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, vint32m1_t op10, const int32_t * op11, vuint32m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_i32mf2))) +void vloxseg5ei32(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, const int32_t * op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_i32mf2_m))) +void vloxseg5ei32(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vbool64_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, vint32mf2_t op10, const int32_t * op11, vuint32mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_u32m1))) +void vloxseg5ei32(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, const uint32_t * op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_u32m1_m))) +void vloxseg5ei32(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vbool32_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, vuint32m1_t op10, const uint32_t * op11, vuint32m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_u32mf2))) +void vloxseg5ei32(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, const uint32_t * op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_u32mf2_m))) +void vloxseg5ei32(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vbool64_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, vuint32mf2_t op10, const uint32_t * op11, vuint32mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_i32m1))) +void vloxseg6ei32(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, const int32_t * op6, vuint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_i32m1_m))) +void vloxseg6ei32(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vbool32_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, vint32m1_t op10, vint32m1_t op11, vint32m1_t op12, const int32_t * op13, vuint32m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_i32mf2))) +void vloxseg6ei32(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, const int32_t * op6, vuint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_i32mf2_m))) +void vloxseg6ei32(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vbool64_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, vint32mf2_t op10, vint32mf2_t op11, vint32mf2_t op12, const int32_t * op13, vuint32mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_u32m1))) +void vloxseg6ei32(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, const uint32_t * op6, vuint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_u32m1_m))) +void vloxseg6ei32(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vbool32_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, vuint32m1_t op10, vuint32m1_t op11, vuint32m1_t op12, const uint32_t * op13, vuint32m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_u32mf2))) +void vloxseg6ei32(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, const uint32_t * op6, vuint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_u32mf2_m))) +void vloxseg6ei32(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vbool64_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, vuint32mf2_t op10, vuint32mf2_t op11, vuint32mf2_t op12, const uint32_t * op13, vuint32mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_i32m1))) +void vloxseg7ei32(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vint32m1_t * op6, const int32_t * op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_i32m1_m))) +void vloxseg7ei32(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vint32m1_t * op6, vbool32_t op7, vint32m1_t op8, vint32m1_t op9, vint32m1_t op10, vint32m1_t op11, vint32m1_t op12, vint32m1_t op13, vint32m1_t op14, const int32_t * op15, vuint32m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_i32mf2))) +void vloxseg7ei32(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vint32mf2_t * op6, const int32_t * op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_i32mf2_m))) +void vloxseg7ei32(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vint32mf2_t * op6, vbool64_t op7, vint32mf2_t op8, vint32mf2_t op9, vint32mf2_t op10, vint32mf2_t op11, vint32mf2_t op12, vint32mf2_t op13, vint32mf2_t op14, const int32_t * op15, vuint32mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_u32m1))) +void vloxseg7ei32(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vuint32m1_t * op6, const uint32_t * op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_u32m1_m))) +void vloxseg7ei32(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vuint32m1_t * op6, vbool32_t op7, vuint32m1_t op8, vuint32m1_t op9, vuint32m1_t op10, vuint32m1_t op11, vuint32m1_t op12, vuint32m1_t op13, vuint32m1_t op14, const uint32_t * op15, vuint32m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_u32mf2))) +void vloxseg7ei32(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vuint32mf2_t * op6, const uint32_t * op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_u32mf2_m))) +void vloxseg7ei32(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vuint32mf2_t * op6, vbool64_t op7, vuint32mf2_t op8, vuint32mf2_t op9, vuint32mf2_t op10, vuint32mf2_t op11, vuint32mf2_t op12, vuint32mf2_t op13, vuint32mf2_t op14, const uint32_t * op15, vuint32mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_i32m1))) +void vloxseg8ei32(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vint32m1_t * op6, vint32m1_t * op7, const int32_t * op8, vuint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_i32m1_m))) +void vloxseg8ei32(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vint32m1_t * op6, vint32m1_t * op7, vbool32_t op8, vint32m1_t op9, vint32m1_t op10, vint32m1_t op11, vint32m1_t op12, vint32m1_t op13, vint32m1_t op14, vint32m1_t op15, vint32m1_t op16, const int32_t * op17, vuint32m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_i32mf2))) +void vloxseg8ei32(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vint32mf2_t * op6, vint32mf2_t * op7, const int32_t * op8, vuint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_i32mf2_m))) +void vloxseg8ei32(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vint32mf2_t * op6, vint32mf2_t * op7, vbool64_t op8, vint32mf2_t op9, vint32mf2_t op10, vint32mf2_t op11, vint32mf2_t op12, vint32mf2_t op13, vint32mf2_t op14, vint32mf2_t op15, vint32mf2_t op16, const int32_t * op17, vuint32mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_u32m1))) +void vloxseg8ei32(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vuint32m1_t * op6, vuint32m1_t * op7, const uint32_t * op8, vuint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_u32m1_m))) +void vloxseg8ei32(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vuint32m1_t * op6, vuint32m1_t * op7, vbool32_t op8, vuint32m1_t op9, vuint32m1_t op10, vuint32m1_t op11, vuint32m1_t op12, vuint32m1_t op13, vuint32m1_t op14, vuint32m1_t op15, vuint32m1_t op16, const uint32_t * op17, vuint32m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_u32mf2))) +void vloxseg8ei32(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vuint32mf2_t * op6, vuint32mf2_t * op7, const uint32_t * op8, vuint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_u32mf2_m))) +void vloxseg8ei32(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vuint32mf2_t * op6, vuint32mf2_t * op7, vbool64_t op8, vuint32mf2_t op9, vuint32mf2_t op10, vuint32mf2_t op11, vuint32mf2_t op12, vuint32mf2_t op13, vuint32mf2_t op14, vuint32mf2_t op15, vuint32mf2_t op16, const uint32_t * op17, vuint32mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_i32m1))) +void vloxseg2ei64(vint32m1_t * op0, vint32m1_t * op1, const int32_t * op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_i32m1_m))) +void vloxseg2ei64(vint32m1_t * op0, vint32m1_t * op1, vbool32_t op2, vint32m1_t op3, vint32m1_t op4, const int32_t * op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_i32m2))) +void vloxseg2ei64(vint32m2_t * op0, vint32m2_t * op1, const int32_t * op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_i32m2_m))) +void vloxseg2ei64(vint32m2_t * op0, vint32m2_t * op1, vbool16_t op2, vint32m2_t op3, vint32m2_t op4, const int32_t * op5, vuint64m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_i32m4))) +void vloxseg2ei64(vint32m4_t * op0, vint32m4_t * op1, const int32_t * op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_i32m4_m))) +void vloxseg2ei64(vint32m4_t * op0, vint32m4_t * op1, vbool8_t op2, vint32m4_t op3, vint32m4_t op4, const int32_t * op5, vuint64m8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_i32mf2))) +void vloxseg2ei64(vint32mf2_t * op0, vint32mf2_t * op1, const int32_t * op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_i32mf2_m))) +void vloxseg2ei64(vint32mf2_t * op0, vint32mf2_t * op1, vbool64_t op2, vint32mf2_t op3, vint32mf2_t op4, const int32_t * op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_u32m1))) +void vloxseg2ei64(vuint32m1_t * op0, vuint32m1_t * op1, const uint32_t * op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_u32m1_m))) +void vloxseg2ei64(vuint32m1_t * op0, vuint32m1_t * op1, vbool32_t op2, vuint32m1_t op3, vuint32m1_t op4, const uint32_t * op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_u32m2))) +void vloxseg2ei64(vuint32m2_t * op0, vuint32m2_t * op1, const uint32_t * op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_u32m2_m))) +void vloxseg2ei64(vuint32m2_t * op0, vuint32m2_t * op1, vbool16_t op2, vuint32m2_t op3, vuint32m2_t op4, const uint32_t * op5, vuint64m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_u32m4))) +void vloxseg2ei64(vuint32m4_t * op0, vuint32m4_t * op1, const uint32_t * op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_u32m4_m))) +void vloxseg2ei64(vuint32m4_t * op0, vuint32m4_t * op1, vbool8_t op2, vuint32m4_t op3, vuint32m4_t op4, const uint32_t * op5, vuint64m8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_u32mf2))) +void vloxseg2ei64(vuint32mf2_t * op0, vuint32mf2_t * op1, const uint32_t * op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_u32mf2_m))) +void vloxseg2ei64(vuint32mf2_t * op0, vuint32mf2_t * op1, vbool64_t op2, vuint32mf2_t op3, vuint32mf2_t op4, const uint32_t * op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_i32m1))) +void vloxseg3ei64(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, const int32_t * op3, vuint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_i32m1_m))) +void vloxseg3ei64(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vbool32_t op3, vint32m1_t op4, vint32m1_t op5, vint32m1_t op6, const int32_t * op7, vuint64m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_i32m2))) +void vloxseg3ei64(vint32m2_t * op0, vint32m2_t * op1, vint32m2_t * op2, const int32_t * op3, vuint64m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_i32m2_m))) +void vloxseg3ei64(vint32m2_t * op0, vint32m2_t * op1, vint32m2_t * op2, vbool16_t op3, vint32m2_t op4, vint32m2_t op5, vint32m2_t op6, const int32_t * op7, vuint64m4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_i32mf2))) +void vloxseg3ei64(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, const int32_t * op3, vuint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_i32mf2_m))) +void vloxseg3ei64(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vbool64_t op3, vint32mf2_t op4, vint32mf2_t op5, vint32mf2_t op6, const int32_t * op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_u32m1))) +void vloxseg3ei64(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, const uint32_t * op3, vuint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_u32m1_m))) +void vloxseg3ei64(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vbool32_t op3, vuint32m1_t op4, vuint32m1_t op5, vuint32m1_t op6, const uint32_t * op7, vuint64m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_u32m2))) +void vloxseg3ei64(vuint32m2_t * op0, vuint32m2_t * op1, vuint32m2_t * op2, const uint32_t * op3, vuint64m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_u32m2_m))) +void vloxseg3ei64(vuint32m2_t * op0, vuint32m2_t * op1, vuint32m2_t * op2, vbool16_t op3, vuint32m2_t op4, vuint32m2_t op5, vuint32m2_t op6, const uint32_t * op7, vuint64m4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_u32mf2))) +void vloxseg3ei64(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, const uint32_t * op3, vuint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_u32mf2_m))) +void vloxseg3ei64(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vbool64_t op3, vuint32mf2_t op4, vuint32mf2_t op5, vuint32mf2_t op6, const uint32_t * op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_i32m1))) +void vloxseg4ei64(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, const int32_t * op4, vuint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_i32m1_m))) +void vloxseg4ei64(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vbool32_t op4, vint32m1_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, const int32_t * op9, vuint64m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_i32m2))) +void vloxseg4ei64(vint32m2_t * op0, vint32m2_t * op1, vint32m2_t * op2, vint32m2_t * op3, const int32_t * op4, vuint64m4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_i32m2_m))) +void vloxseg4ei64(vint32m2_t * op0, vint32m2_t * op1, vint32m2_t * op2, vint32m2_t * op3, vbool16_t op4, vint32m2_t op5, vint32m2_t op6, vint32m2_t op7, vint32m2_t op8, const int32_t * op9, vuint64m4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_i32mf2))) +void vloxseg4ei64(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, const int32_t * op4, vuint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_i32mf2_m))) +void vloxseg4ei64(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vbool64_t op4, vint32mf2_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, const int32_t * op9, vuint64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_u32m1))) +void vloxseg4ei64(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, const uint32_t * op4, vuint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_u32m1_m))) +void vloxseg4ei64(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vbool32_t op4, vuint32m1_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, const uint32_t * op9, vuint64m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_u32m2))) +void vloxseg4ei64(vuint32m2_t * op0, vuint32m2_t * op1, vuint32m2_t * op2, vuint32m2_t * op3, const uint32_t * op4, vuint64m4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_u32m2_m))) +void vloxseg4ei64(vuint32m2_t * op0, vuint32m2_t * op1, vuint32m2_t * op2, vuint32m2_t * op3, vbool16_t op4, vuint32m2_t op5, vuint32m2_t op6, vuint32m2_t op7, vuint32m2_t op8, const uint32_t * op9, vuint64m4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_u32mf2))) +void vloxseg4ei64(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, const uint32_t * op4, vuint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_u32mf2_m))) +void vloxseg4ei64(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vbool64_t op4, vuint32mf2_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, const uint32_t * op9, vuint64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_i32m1))) +void vloxseg5ei64(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, const int32_t * op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_i32m1_m))) +void vloxseg5ei64(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vbool32_t op5, vint32m1_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, vint32m1_t op10, const int32_t * op11, vuint64m2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_i32mf2))) +void vloxseg5ei64(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, const int32_t * op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_i32mf2_m))) +void vloxseg5ei64(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vbool64_t op5, vint32mf2_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, vint32mf2_t op10, const int32_t * op11, vuint64m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_u32m1))) +void vloxseg5ei64(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, const uint32_t * op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_u32m1_m))) +void vloxseg5ei64(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vbool32_t op5, vuint32m1_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, vuint32m1_t op10, const uint32_t * op11, vuint64m2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_u32mf2))) +void vloxseg5ei64(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, const uint32_t * op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_u32mf2_m))) +void vloxseg5ei64(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vbool64_t op5, vuint32mf2_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, vuint32mf2_t op10, const uint32_t * op11, vuint64m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_i32m1))) +void vloxseg6ei64(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, const int32_t * op6, vuint64m2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_i32m1_m))) +void vloxseg6ei64(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vbool32_t op6, vint32m1_t op7, vint32m1_t op8, vint32m1_t op9, vint32m1_t op10, vint32m1_t op11, vint32m1_t op12, const int32_t * op13, vuint64m2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_i32mf2))) +void vloxseg6ei64(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, const int32_t * op6, vuint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_i32mf2_m))) +void vloxseg6ei64(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vbool64_t op6, vint32mf2_t op7, vint32mf2_t op8, vint32mf2_t op9, vint32mf2_t op10, vint32mf2_t op11, vint32mf2_t op12, const int32_t * op13, vuint64m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_u32m1))) +void vloxseg6ei64(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, const uint32_t * op6, vuint64m2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_u32m1_m))) +void vloxseg6ei64(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vbool32_t op6, vuint32m1_t op7, vuint32m1_t op8, vuint32m1_t op9, vuint32m1_t op10, vuint32m1_t op11, vuint32m1_t op12, const uint32_t * op13, vuint64m2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_u32mf2))) +void vloxseg6ei64(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, const uint32_t * op6, vuint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_u32mf2_m))) +void vloxseg6ei64(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vbool64_t op6, vuint32mf2_t op7, vuint32mf2_t op8, vuint32mf2_t op9, vuint32mf2_t op10, vuint32mf2_t op11, vuint32mf2_t op12, const uint32_t * op13, vuint64m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_i32m1))) +void vloxseg7ei64(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vint32m1_t * op6, const int32_t * op7, vuint64m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_i32m1_m))) +void vloxseg7ei64(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vint32m1_t * op6, vbool32_t op7, vint32m1_t op8, vint32m1_t op9, vint32m1_t op10, vint32m1_t op11, vint32m1_t op12, vint32m1_t op13, vint32m1_t op14, const int32_t * op15, vuint64m2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_i32mf2))) +void vloxseg7ei64(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vint32mf2_t * op6, const int32_t * op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_i32mf2_m))) +void vloxseg7ei64(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vint32mf2_t * op6, vbool64_t op7, vint32mf2_t op8, vint32mf2_t op9, vint32mf2_t op10, vint32mf2_t op11, vint32mf2_t op12, vint32mf2_t op13, vint32mf2_t op14, const int32_t * op15, vuint64m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_u32m1))) +void vloxseg7ei64(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vuint32m1_t * op6, const uint32_t * op7, vuint64m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_u32m1_m))) +void vloxseg7ei64(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vuint32m1_t * op6, vbool32_t op7, vuint32m1_t op8, vuint32m1_t op9, vuint32m1_t op10, vuint32m1_t op11, vuint32m1_t op12, vuint32m1_t op13, vuint32m1_t op14, const uint32_t * op15, vuint64m2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_u32mf2))) +void vloxseg7ei64(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vuint32mf2_t * op6, const uint32_t * op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_u32mf2_m))) +void vloxseg7ei64(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vuint32mf2_t * op6, vbool64_t op7, vuint32mf2_t op8, vuint32mf2_t op9, vuint32mf2_t op10, vuint32mf2_t op11, vuint32mf2_t op12, vuint32mf2_t op13, vuint32mf2_t op14, const uint32_t * op15, vuint64m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_i32m1))) +void vloxseg8ei64(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vint32m1_t * op6, vint32m1_t * op7, const int32_t * op8, vuint64m2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_i32m1_m))) +void vloxseg8ei64(vint32m1_t * op0, vint32m1_t * op1, vint32m1_t * op2, vint32m1_t * op3, vint32m1_t * op4, vint32m1_t * op5, vint32m1_t * op6, vint32m1_t * op7, vbool32_t op8, vint32m1_t op9, vint32m1_t op10, vint32m1_t op11, vint32m1_t op12, vint32m1_t op13, vint32m1_t op14, vint32m1_t op15, vint32m1_t op16, const int32_t * op17, vuint64m2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_i32mf2))) +void vloxseg8ei64(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vint32mf2_t * op6, vint32mf2_t * op7, const int32_t * op8, vuint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_i32mf2_m))) +void vloxseg8ei64(vint32mf2_t * op0, vint32mf2_t * op1, vint32mf2_t * op2, vint32mf2_t * op3, vint32mf2_t * op4, vint32mf2_t * op5, vint32mf2_t * op6, vint32mf2_t * op7, vbool64_t op8, vint32mf2_t op9, vint32mf2_t op10, vint32mf2_t op11, vint32mf2_t op12, vint32mf2_t op13, vint32mf2_t op14, vint32mf2_t op15, vint32mf2_t op16, const int32_t * op17, vuint64m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_u32m1))) +void vloxseg8ei64(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vuint32m1_t * op6, vuint32m1_t * op7, const uint32_t * op8, vuint64m2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_u32m1_m))) +void vloxseg8ei64(vuint32m1_t * op0, vuint32m1_t * op1, vuint32m1_t * op2, vuint32m1_t * op3, vuint32m1_t * op4, vuint32m1_t * op5, vuint32m1_t * op6, vuint32m1_t * op7, vbool32_t op8, vuint32m1_t op9, vuint32m1_t op10, vuint32m1_t op11, vuint32m1_t op12, vuint32m1_t op13, vuint32m1_t op14, vuint32m1_t op15, vuint32m1_t op16, const uint32_t * op17, vuint64m2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_u32mf2))) +void vloxseg8ei64(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vuint32mf2_t * op6, vuint32mf2_t * op7, const uint32_t * op8, vuint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_u32mf2_m))) +void vloxseg8ei64(vuint32mf2_t * op0, vuint32mf2_t * op1, vuint32mf2_t * op2, vuint32mf2_t * op3, vuint32mf2_t * op4, vuint32mf2_t * op5, vuint32mf2_t * op6, vuint32mf2_t * op7, vbool64_t op8, vuint32mf2_t op9, vuint32mf2_t op10, vuint32mf2_t op11, vuint32mf2_t op12, vuint32mf2_t op13, vuint32mf2_t op14, vuint32mf2_t op15, vuint32mf2_t op16, const uint32_t * op17, vuint64m1_t op18, size_t op19); + +#endif + +#if defined(__riscv_f) && defined(__riscv_zvlsseg) +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_f32m1))) +void vloxseg2ei8(vfloat32m1_t * op0, vfloat32m1_t * op1, const float * op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_f32m1_m))) +void vloxseg2ei8(vfloat32m1_t * op0, vfloat32m1_t * op1, vbool32_t op2, vfloat32m1_t op3, vfloat32m1_t op4, const float * op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_f32m2))) +void vloxseg2ei8(vfloat32m2_t * op0, vfloat32m2_t * op1, const float * op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_f32m2_m))) +void vloxseg2ei8(vfloat32m2_t * op0, vfloat32m2_t * op1, vbool16_t op2, vfloat32m2_t op3, vfloat32m2_t op4, const float * op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_f32m4))) +void vloxseg2ei8(vfloat32m4_t * op0, vfloat32m4_t * op1, const float * op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_f32m4_m))) +void vloxseg2ei8(vfloat32m4_t * op0, vfloat32m4_t * op1, vbool8_t op2, vfloat32m4_t op3, vfloat32m4_t op4, const float * op5, vuint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_f32mf2))) +void vloxseg2ei8(vfloat32mf2_t * op0, vfloat32mf2_t * op1, const float * op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_f32mf2_m))) +void vloxseg2ei8(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vbool64_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, const float * op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_f32m1))) +void vloxseg3ei8(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, const float * op3, vuint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_f32m1_m))) +void vloxseg3ei8(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vbool32_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, const float * op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_f32m2))) +void vloxseg3ei8(vfloat32m2_t * op0, vfloat32m2_t * op1, vfloat32m2_t * op2, const float * op3, vuint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_f32m2_m))) +void vloxseg3ei8(vfloat32m2_t * op0, vfloat32m2_t * op1, vfloat32m2_t * op2, vbool16_t op3, vfloat32m2_t op4, vfloat32m2_t op5, vfloat32m2_t op6, const float * op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_f32mf2))) +void vloxseg3ei8(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, const float * op3, vuint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_f32mf2_m))) +void vloxseg3ei8(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vbool64_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, const float * op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_f32m1))) +void vloxseg4ei8(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, const float * op4, vuint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_f32m1_m))) +void vloxseg4ei8(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vbool32_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, const float * op9, vuint8mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_f32m2))) +void vloxseg4ei8(vfloat32m2_t * op0, vfloat32m2_t * op1, vfloat32m2_t * op2, vfloat32m2_t * op3, const float * op4, vuint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_f32m2_m))) +void vloxseg4ei8(vfloat32m2_t * op0, vfloat32m2_t * op1, vfloat32m2_t * op2, vfloat32m2_t * op3, vbool16_t op4, vfloat32m2_t op5, vfloat32m2_t op6, vfloat32m2_t op7, vfloat32m2_t op8, const float * op9, vuint8mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_f32mf2))) +void vloxseg4ei8(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, const float * op4, vuint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_f32mf2_m))) +void vloxseg4ei8(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vbool64_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, const float * op9, vuint8mf8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_f32m1))) +void vloxseg5ei8(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, const float * op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_f32m1_m))) +void vloxseg5ei8(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vbool32_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, vfloat32m1_t op10, const float * op11, vuint8mf4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_f32mf2))) +void vloxseg5ei8(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, const float * op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_f32mf2_m))) +void vloxseg5ei8(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vbool64_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, const float * op11, vuint8mf8_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_f32m1))) +void vloxseg6ei8(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, const float * op6, vuint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_f32m1_m))) +void vloxseg6ei8(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vbool32_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, vfloat32m1_t op10, vfloat32m1_t op11, vfloat32m1_t op12, const float * op13, vuint8mf4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_f32mf2))) +void vloxseg6ei8(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, const float * op6, vuint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_f32mf2_m))) +void vloxseg6ei8(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vbool64_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, vfloat32mf2_t op11, vfloat32mf2_t op12, const float * op13, vuint8mf8_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_f32m1))) +void vloxseg7ei8(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vfloat32m1_t * op6, const float * op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_f32m1_m))) +void vloxseg7ei8(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vfloat32m1_t * op6, vbool32_t op7, vfloat32m1_t op8, vfloat32m1_t op9, vfloat32m1_t op10, vfloat32m1_t op11, vfloat32m1_t op12, vfloat32m1_t op13, vfloat32m1_t op14, const float * op15, vuint8mf4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_f32mf2))) +void vloxseg7ei8(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vfloat32mf2_t * op6, const float * op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_f32mf2_m))) +void vloxseg7ei8(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vfloat32mf2_t * op6, vbool64_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, vfloat32mf2_t op11, vfloat32mf2_t op12, vfloat32mf2_t op13, vfloat32mf2_t op14, const float * op15, vuint8mf8_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_f32m1))) +void vloxseg8ei8(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vfloat32m1_t * op6, vfloat32m1_t * op7, const float * op8, vuint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_f32m1_m))) +void vloxseg8ei8(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vfloat32m1_t * op6, vfloat32m1_t * op7, vbool32_t op8, vfloat32m1_t op9, vfloat32m1_t op10, vfloat32m1_t op11, vfloat32m1_t op12, vfloat32m1_t op13, vfloat32m1_t op14, vfloat32m1_t op15, vfloat32m1_t op16, const float * op17, vuint8mf4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_f32mf2))) +void vloxseg8ei8(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vfloat32mf2_t * op6, vfloat32mf2_t * op7, const float * op8, vuint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_f32mf2_m))) +void vloxseg8ei8(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vfloat32mf2_t * op6, vfloat32mf2_t * op7, vbool64_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, vfloat32mf2_t op11, vfloat32mf2_t op12, vfloat32mf2_t op13, vfloat32mf2_t op14, vfloat32mf2_t op15, vfloat32mf2_t op16, const float * op17, vuint8mf8_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_f32m1))) +void vloxseg2ei16(vfloat32m1_t * op0, vfloat32m1_t * op1, const float * op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_f32m1_m))) +void vloxseg2ei16(vfloat32m1_t * op0, vfloat32m1_t * op1, vbool32_t op2, vfloat32m1_t op3, vfloat32m1_t op4, const float * op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_f32m2))) +void vloxseg2ei16(vfloat32m2_t * op0, vfloat32m2_t * op1, const float * op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_f32m2_m))) +void vloxseg2ei16(vfloat32m2_t * op0, vfloat32m2_t * op1, vbool16_t op2, vfloat32m2_t op3, vfloat32m2_t op4, const float * op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_f32m4))) +void vloxseg2ei16(vfloat32m4_t * op0, vfloat32m4_t * op1, const float * op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_f32m4_m))) +void vloxseg2ei16(vfloat32m4_t * op0, vfloat32m4_t * op1, vbool8_t op2, vfloat32m4_t op3, vfloat32m4_t op4, const float * op5, vuint16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_f32mf2))) +void vloxseg2ei16(vfloat32mf2_t * op0, vfloat32mf2_t * op1, const float * op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_f32mf2_m))) +void vloxseg2ei16(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vbool64_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, const float * op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_f32m1))) +void vloxseg3ei16(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, const float * op3, vuint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_f32m1_m))) +void vloxseg3ei16(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vbool32_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, const float * op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_f32m2))) +void vloxseg3ei16(vfloat32m2_t * op0, vfloat32m2_t * op1, vfloat32m2_t * op2, const float * op3, vuint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_f32m2_m))) +void vloxseg3ei16(vfloat32m2_t * op0, vfloat32m2_t * op1, vfloat32m2_t * op2, vbool16_t op3, vfloat32m2_t op4, vfloat32m2_t op5, vfloat32m2_t op6, const float * op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_f32mf2))) +void vloxseg3ei16(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, const float * op3, vuint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_f32mf2_m))) +void vloxseg3ei16(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vbool64_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, const float * op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_f32m1))) +void vloxseg4ei16(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, const float * op4, vuint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_f32m1_m))) +void vloxseg4ei16(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vbool32_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, const float * op9, vuint16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_f32m2))) +void vloxseg4ei16(vfloat32m2_t * op0, vfloat32m2_t * op1, vfloat32m2_t * op2, vfloat32m2_t * op3, const float * op4, vuint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_f32m2_m))) +void vloxseg4ei16(vfloat32m2_t * op0, vfloat32m2_t * op1, vfloat32m2_t * op2, vfloat32m2_t * op3, vbool16_t op4, vfloat32m2_t op5, vfloat32m2_t op6, vfloat32m2_t op7, vfloat32m2_t op8, const float * op9, vuint16m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_f32mf2))) +void vloxseg4ei16(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, const float * op4, vuint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_f32mf2_m))) +void vloxseg4ei16(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vbool64_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, const float * op9, vuint16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_f32m1))) +void vloxseg5ei16(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, const float * op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_f32m1_m))) +void vloxseg5ei16(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vbool32_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, vfloat32m1_t op10, const float * op11, vuint16mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_f32mf2))) +void vloxseg5ei16(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, const float * op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_f32mf2_m))) +void vloxseg5ei16(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vbool64_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, const float * op11, vuint16mf4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_f32m1))) +void vloxseg6ei16(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, const float * op6, vuint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_f32m1_m))) +void vloxseg6ei16(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vbool32_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, vfloat32m1_t op10, vfloat32m1_t op11, vfloat32m1_t op12, const float * op13, vuint16mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_f32mf2))) +void vloxseg6ei16(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, const float * op6, vuint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_f32mf2_m))) +void vloxseg6ei16(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vbool64_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, vfloat32mf2_t op11, vfloat32mf2_t op12, const float * op13, vuint16mf4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_f32m1))) +void vloxseg7ei16(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vfloat32m1_t * op6, const float * op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_f32m1_m))) +void vloxseg7ei16(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vfloat32m1_t * op6, vbool32_t op7, vfloat32m1_t op8, vfloat32m1_t op9, vfloat32m1_t op10, vfloat32m1_t op11, vfloat32m1_t op12, vfloat32m1_t op13, vfloat32m1_t op14, const float * op15, vuint16mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_f32mf2))) +void vloxseg7ei16(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vfloat32mf2_t * op6, const float * op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_f32mf2_m))) +void vloxseg7ei16(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vfloat32mf2_t * op6, vbool64_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, vfloat32mf2_t op11, vfloat32mf2_t op12, vfloat32mf2_t op13, vfloat32mf2_t op14, const float * op15, vuint16mf4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_f32m1))) +void vloxseg8ei16(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vfloat32m1_t * op6, vfloat32m1_t * op7, const float * op8, vuint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_f32m1_m))) +void vloxseg8ei16(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vfloat32m1_t * op6, vfloat32m1_t * op7, vbool32_t op8, vfloat32m1_t op9, vfloat32m1_t op10, vfloat32m1_t op11, vfloat32m1_t op12, vfloat32m1_t op13, vfloat32m1_t op14, vfloat32m1_t op15, vfloat32m1_t op16, const float * op17, vuint16mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_f32mf2))) +void vloxseg8ei16(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vfloat32mf2_t * op6, vfloat32mf2_t * op7, const float * op8, vuint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_f32mf2_m))) +void vloxseg8ei16(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vfloat32mf2_t * op6, vfloat32mf2_t * op7, vbool64_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, vfloat32mf2_t op11, vfloat32mf2_t op12, vfloat32mf2_t op13, vfloat32mf2_t op14, vfloat32mf2_t op15, vfloat32mf2_t op16, const float * op17, vuint16mf4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_f32m1))) +void vloxseg2ei32(vfloat32m1_t * op0, vfloat32m1_t * op1, const float * op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_f32m1_m))) +void vloxseg2ei32(vfloat32m1_t * op0, vfloat32m1_t * op1, vbool32_t op2, vfloat32m1_t op3, vfloat32m1_t op4, const float * op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_f32m2))) +void vloxseg2ei32(vfloat32m2_t * op0, vfloat32m2_t * op1, const float * op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_f32m2_m))) +void vloxseg2ei32(vfloat32m2_t * op0, vfloat32m2_t * op1, vbool16_t op2, vfloat32m2_t op3, vfloat32m2_t op4, const float * op5, vuint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_f32m4))) +void vloxseg2ei32(vfloat32m4_t * op0, vfloat32m4_t * op1, const float * op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_f32m4_m))) +void vloxseg2ei32(vfloat32m4_t * op0, vfloat32m4_t * op1, vbool8_t op2, vfloat32m4_t op3, vfloat32m4_t op4, const float * op5, vuint32m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_f32mf2))) +void vloxseg2ei32(vfloat32mf2_t * op0, vfloat32mf2_t * op1, const float * op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_f32mf2_m))) +void vloxseg2ei32(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vbool64_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, const float * op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_f32m1))) +void vloxseg3ei32(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, const float * op3, vuint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_f32m1_m))) +void vloxseg3ei32(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vbool32_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, const float * op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_f32m2))) +void vloxseg3ei32(vfloat32m2_t * op0, vfloat32m2_t * op1, vfloat32m2_t * op2, const float * op3, vuint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_f32m2_m))) +void vloxseg3ei32(vfloat32m2_t * op0, vfloat32m2_t * op1, vfloat32m2_t * op2, vbool16_t op3, vfloat32m2_t op4, vfloat32m2_t op5, vfloat32m2_t op6, const float * op7, vuint32m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_f32mf2))) +void vloxseg3ei32(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, const float * op3, vuint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_f32mf2_m))) +void vloxseg3ei32(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vbool64_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, const float * op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_f32m1))) +void vloxseg4ei32(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, const float * op4, vuint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_f32m1_m))) +void vloxseg4ei32(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vbool32_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, const float * op9, vuint32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_f32m2))) +void vloxseg4ei32(vfloat32m2_t * op0, vfloat32m2_t * op1, vfloat32m2_t * op2, vfloat32m2_t * op3, const float * op4, vuint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_f32m2_m))) +void vloxseg4ei32(vfloat32m2_t * op0, vfloat32m2_t * op1, vfloat32m2_t * op2, vfloat32m2_t * op3, vbool16_t op4, vfloat32m2_t op5, vfloat32m2_t op6, vfloat32m2_t op7, vfloat32m2_t op8, const float * op9, vuint32m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_f32mf2))) +void vloxseg4ei32(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, const float * op4, vuint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_f32mf2_m))) +void vloxseg4ei32(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vbool64_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, const float * op9, vuint32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_f32m1))) +void vloxseg5ei32(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, const float * op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_f32m1_m))) +void vloxseg5ei32(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vbool32_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, vfloat32m1_t op10, const float * op11, vuint32m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_f32mf2))) +void vloxseg5ei32(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, const float * op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_f32mf2_m))) +void vloxseg5ei32(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vbool64_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, const float * op11, vuint32mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_f32m1))) +void vloxseg6ei32(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, const float * op6, vuint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_f32m1_m))) +void vloxseg6ei32(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vbool32_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, vfloat32m1_t op10, vfloat32m1_t op11, vfloat32m1_t op12, const float * op13, vuint32m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_f32mf2))) +void vloxseg6ei32(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, const float * op6, vuint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_f32mf2_m))) +void vloxseg6ei32(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vbool64_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, vfloat32mf2_t op11, vfloat32mf2_t op12, const float * op13, vuint32mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_f32m1))) +void vloxseg7ei32(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vfloat32m1_t * op6, const float * op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_f32m1_m))) +void vloxseg7ei32(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vfloat32m1_t * op6, vbool32_t op7, vfloat32m1_t op8, vfloat32m1_t op9, vfloat32m1_t op10, vfloat32m1_t op11, vfloat32m1_t op12, vfloat32m1_t op13, vfloat32m1_t op14, const float * op15, vuint32m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_f32mf2))) +void vloxseg7ei32(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vfloat32mf2_t * op6, const float * op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_f32mf2_m))) +void vloxseg7ei32(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vfloat32mf2_t * op6, vbool64_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, vfloat32mf2_t op11, vfloat32mf2_t op12, vfloat32mf2_t op13, vfloat32mf2_t op14, const float * op15, vuint32mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_f32m1))) +void vloxseg8ei32(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vfloat32m1_t * op6, vfloat32m1_t * op7, const float * op8, vuint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_f32m1_m))) +void vloxseg8ei32(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vfloat32m1_t * op6, vfloat32m1_t * op7, vbool32_t op8, vfloat32m1_t op9, vfloat32m1_t op10, vfloat32m1_t op11, vfloat32m1_t op12, vfloat32m1_t op13, vfloat32m1_t op14, vfloat32m1_t op15, vfloat32m1_t op16, const float * op17, vuint32m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_f32mf2))) +void vloxseg8ei32(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vfloat32mf2_t * op6, vfloat32mf2_t * op7, const float * op8, vuint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_f32mf2_m))) +void vloxseg8ei32(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vfloat32mf2_t * op6, vfloat32mf2_t * op7, vbool64_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, vfloat32mf2_t op11, vfloat32mf2_t op12, vfloat32mf2_t op13, vfloat32mf2_t op14, vfloat32mf2_t op15, vfloat32mf2_t op16, const float * op17, vuint32mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_f32m1))) +void vloxseg2ei64(vfloat32m1_t * op0, vfloat32m1_t * op1, const float * op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_f32m1_m))) +void vloxseg2ei64(vfloat32m1_t * op0, vfloat32m1_t * op1, vbool32_t op2, vfloat32m1_t op3, vfloat32m1_t op4, const float * op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_f32m2))) +void vloxseg2ei64(vfloat32m2_t * op0, vfloat32m2_t * op1, const float * op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_f32m2_m))) +void vloxseg2ei64(vfloat32m2_t * op0, vfloat32m2_t * op1, vbool16_t op2, vfloat32m2_t op3, vfloat32m2_t op4, const float * op5, vuint64m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_f32m4))) +void vloxseg2ei64(vfloat32m4_t * op0, vfloat32m4_t * op1, const float * op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_f32m4_m))) +void vloxseg2ei64(vfloat32m4_t * op0, vfloat32m4_t * op1, vbool8_t op2, vfloat32m4_t op3, vfloat32m4_t op4, const float * op5, vuint64m8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_f32mf2))) +void vloxseg2ei64(vfloat32mf2_t * op0, vfloat32mf2_t * op1, const float * op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_f32mf2_m))) +void vloxseg2ei64(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vbool64_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, const float * op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_f32m1))) +void vloxseg3ei64(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, const float * op3, vuint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_f32m1_m))) +void vloxseg3ei64(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vbool32_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, const float * op7, vuint64m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_f32m2))) +void vloxseg3ei64(vfloat32m2_t * op0, vfloat32m2_t * op1, vfloat32m2_t * op2, const float * op3, vuint64m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_f32m2_m))) +void vloxseg3ei64(vfloat32m2_t * op0, vfloat32m2_t * op1, vfloat32m2_t * op2, vbool16_t op3, vfloat32m2_t op4, vfloat32m2_t op5, vfloat32m2_t op6, const float * op7, vuint64m4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_f32mf2))) +void vloxseg3ei64(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, const float * op3, vuint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_f32mf2_m))) +void vloxseg3ei64(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vbool64_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, const float * op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_f32m1))) +void vloxseg4ei64(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, const float * op4, vuint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_f32m1_m))) +void vloxseg4ei64(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vbool32_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, const float * op9, vuint64m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_f32m2))) +void vloxseg4ei64(vfloat32m2_t * op0, vfloat32m2_t * op1, vfloat32m2_t * op2, vfloat32m2_t * op3, const float * op4, vuint64m4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_f32m2_m))) +void vloxseg4ei64(vfloat32m2_t * op0, vfloat32m2_t * op1, vfloat32m2_t * op2, vfloat32m2_t * op3, vbool16_t op4, vfloat32m2_t op5, vfloat32m2_t op6, vfloat32m2_t op7, vfloat32m2_t op8, const float * op9, vuint64m4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_f32mf2))) +void vloxseg4ei64(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, const float * op4, vuint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_f32mf2_m))) +void vloxseg4ei64(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vbool64_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, const float * op9, vuint64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_f32m1))) +void vloxseg5ei64(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, const float * op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_f32m1_m))) +void vloxseg5ei64(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vbool32_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, vfloat32m1_t op10, const float * op11, vuint64m2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_f32mf2))) +void vloxseg5ei64(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, const float * op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_f32mf2_m))) +void vloxseg5ei64(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vbool64_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, const float * op11, vuint64m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_f32m1))) +void vloxseg6ei64(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, const float * op6, vuint64m2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_f32m1_m))) +void vloxseg6ei64(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vbool32_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, vfloat32m1_t op10, vfloat32m1_t op11, vfloat32m1_t op12, const float * op13, vuint64m2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_f32mf2))) +void vloxseg6ei64(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, const float * op6, vuint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_f32mf2_m))) +void vloxseg6ei64(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vbool64_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, vfloat32mf2_t op11, vfloat32mf2_t op12, const float * op13, vuint64m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_f32m1))) +void vloxseg7ei64(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vfloat32m1_t * op6, const float * op7, vuint64m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_f32m1_m))) +void vloxseg7ei64(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vfloat32m1_t * op6, vbool32_t op7, vfloat32m1_t op8, vfloat32m1_t op9, vfloat32m1_t op10, vfloat32m1_t op11, vfloat32m1_t op12, vfloat32m1_t op13, vfloat32m1_t op14, const float * op15, vuint64m2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_f32mf2))) +void vloxseg7ei64(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vfloat32mf2_t * op6, const float * op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_f32mf2_m))) +void vloxseg7ei64(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vfloat32mf2_t * op6, vbool64_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, vfloat32mf2_t op11, vfloat32mf2_t op12, vfloat32mf2_t op13, vfloat32mf2_t op14, const float * op15, vuint64m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_f32m1))) +void vloxseg8ei64(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vfloat32m1_t * op6, vfloat32m1_t * op7, const float * op8, vuint64m2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_f32m1_m))) +void vloxseg8ei64(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vfloat32m1_t * op6, vfloat32m1_t * op7, vbool32_t op8, vfloat32m1_t op9, vfloat32m1_t op10, vfloat32m1_t op11, vfloat32m1_t op12, vfloat32m1_t op13, vfloat32m1_t op14, vfloat32m1_t op15, vfloat32m1_t op16, const float * op17, vuint64m2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_f32mf2))) +void vloxseg8ei64(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vfloat32mf2_t * op6, vfloat32mf2_t * op7, const float * op8, vuint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_f32mf2_m))) +void vloxseg8ei64(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vfloat32mf2_t * op6, vfloat32mf2_t * op7, vbool64_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, vfloat32mf2_t op11, vfloat32mf2_t op12, vfloat32mf2_t op13, vfloat32mf2_t op14, vfloat32mf2_t op15, vfloat32mf2_t op16, const float * op17, vuint64m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e32_v_f32m1))) +void vsseg2e32(float * op0, vfloat32m1_t op1, vfloat32m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e32_v_f32m1_m))) +void vsseg2e32(vbool32_t op0, float * op1, vfloat32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e32_v_f32m2))) +void vsseg2e32(float * op0, vfloat32m2_t op1, vfloat32m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e32_v_f32m2_m))) +void vsseg2e32(vbool16_t op0, float * op1, vfloat32m2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e32_v_f32m4))) +void vsseg2e32(float * op0, vfloat32m4_t op1, vfloat32m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e32_v_f32m4_m))) +void vsseg2e32(vbool8_t op0, float * op1, vfloat32m4_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e32_v_f32mf2))) +void vsseg2e32(float * op0, vfloat32mf2_t op1, vfloat32mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e32_v_f32mf2_m))) +void vsseg2e32(vbool64_t op0, float * op1, vfloat32mf2_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e32_v_f32m1))) +void vsseg3e32(float * op0, vfloat32m1_t op1, vfloat32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e32_v_f32m1_m))) +void vsseg3e32(vbool32_t op0, float * op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e32_v_f32m2))) +void vsseg3e32(float * op0, vfloat32m2_t op1, vfloat32m2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e32_v_f32m2_m))) +void vsseg3e32(vbool16_t op0, float * op1, vfloat32m2_t op2, vfloat32m2_t op3, vfloat32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e32_v_f32mf2))) +void vsseg3e32(float * op0, vfloat32mf2_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e32_v_f32mf2_m))) +void vsseg3e32(vbool64_t op0, float * op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e32_v_f32m1))) +void vsseg4e32(float * op0, vfloat32m1_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e32_v_f32m1_m))) +void vsseg4e32(vbool32_t op0, float * op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e32_v_f32m2))) +void vsseg4e32(float * op0, vfloat32m2_t op1, vfloat32m2_t op2, vfloat32m2_t op3, vfloat32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e32_v_f32m2_m))) +void vsseg4e32(vbool16_t op0, float * op1, vfloat32m2_t op2, vfloat32m2_t op3, vfloat32m2_t op4, vfloat32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e32_v_f32mf2))) +void vsseg4e32(float * op0, vfloat32mf2_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e32_v_f32mf2_m))) +void vsseg4e32(vbool64_t op0, float * op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e32_v_f32m1))) +void vsseg5e32(float * op0, vfloat32m1_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e32_v_f32m1_m))) +void vsseg5e32(vbool32_t op0, float * op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e32_v_f32mf2))) +void vsseg5e32(float * op0, vfloat32mf2_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e32_v_f32mf2_m))) +void vsseg5e32(vbool64_t op0, float * op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e32_v_f32m1))) +void vsseg6e32(float * op0, vfloat32m1_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e32_v_f32m1_m))) +void vsseg6e32(vbool32_t op0, float * op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e32_v_f32mf2))) +void vsseg6e32(float * op0, vfloat32mf2_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e32_v_f32mf2_m))) +void vsseg6e32(vbool64_t op0, float * op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e32_v_f32m1))) +void vsseg7e32(float * op0, vfloat32m1_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e32_v_f32m1_m))) +void vsseg7e32(vbool32_t op0, float * op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e32_v_f32mf2))) +void vsseg7e32(float * op0, vfloat32mf2_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e32_v_f32mf2_m))) +void vsseg7e32(vbool64_t op0, float * op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e32_v_f32m1))) +void vsseg8e32(float * op0, vfloat32m1_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e32_v_f32m1_m))) +void vsseg8e32(vbool32_t op0, float * op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e32_v_f32mf2))) +void vsseg8e32(float * op0, vfloat32mf2_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e32_v_f32mf2_m))) +void vsseg8e32(vbool64_t op0, float * op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e32_v_f32m1))) +void vssseg2e32(float * op0, ptrdiff_t op1, vfloat32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e32_v_f32m1_m))) +void vssseg2e32(vbool32_t op0, float * op1, ptrdiff_t op2, vfloat32m1_t op3, vfloat32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e32_v_f32m2))) +void vssseg2e32(float * op0, ptrdiff_t op1, vfloat32m2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e32_v_f32m2_m))) +void vssseg2e32(vbool16_t op0, float * op1, ptrdiff_t op2, vfloat32m2_t op3, vfloat32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e32_v_f32m4))) +void vssseg2e32(float * op0, ptrdiff_t op1, vfloat32m4_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e32_v_f32m4_m))) +void vssseg2e32(vbool8_t op0, float * op1, ptrdiff_t op2, vfloat32m4_t op3, vfloat32m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e32_v_f32mf2))) +void vssseg2e32(float * op0, ptrdiff_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e32_v_f32mf2_m))) +void vssseg2e32(vbool64_t op0, float * op1, ptrdiff_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e32_v_f32m1))) +void vssseg3e32(float * op0, ptrdiff_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e32_v_f32m1_m))) +void vssseg3e32(vbool32_t op0, float * op1, ptrdiff_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e32_v_f32m2))) +void vssseg3e32(float * op0, ptrdiff_t op1, vfloat32m2_t op2, vfloat32m2_t op3, vfloat32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e32_v_f32m2_m))) +void vssseg3e32(vbool16_t op0, float * op1, ptrdiff_t op2, vfloat32m2_t op3, vfloat32m2_t op4, vfloat32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e32_v_f32mf2))) +void vssseg3e32(float * op0, ptrdiff_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e32_v_f32mf2_m))) +void vssseg3e32(vbool64_t op0, float * op1, ptrdiff_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e32_v_f32m1))) +void vssseg4e32(float * op0, ptrdiff_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e32_v_f32m1_m))) +void vssseg4e32(vbool32_t op0, float * op1, ptrdiff_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e32_v_f32m2))) +void vssseg4e32(float * op0, ptrdiff_t op1, vfloat32m2_t op2, vfloat32m2_t op3, vfloat32m2_t op4, vfloat32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e32_v_f32m2_m))) +void vssseg4e32(vbool16_t op0, float * op1, ptrdiff_t op2, vfloat32m2_t op3, vfloat32m2_t op4, vfloat32m2_t op5, vfloat32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e32_v_f32mf2))) +void vssseg4e32(float * op0, ptrdiff_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e32_v_f32mf2_m))) +void vssseg4e32(vbool64_t op0, float * op1, ptrdiff_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e32_v_f32m1))) +void vssseg5e32(float * op0, ptrdiff_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e32_v_f32m1_m))) +void vssseg5e32(vbool32_t op0, float * op1, ptrdiff_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e32_v_f32mf2))) +void vssseg5e32(float * op0, ptrdiff_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e32_v_f32mf2_m))) +void vssseg5e32(vbool64_t op0, float * op1, ptrdiff_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e32_v_f32m1))) +void vssseg6e32(float * op0, ptrdiff_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e32_v_f32m1_m))) +void vssseg6e32(vbool32_t op0, float * op1, ptrdiff_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e32_v_f32mf2))) +void vssseg6e32(float * op0, ptrdiff_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e32_v_f32mf2_m))) +void vssseg6e32(vbool64_t op0, float * op1, ptrdiff_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e32_v_f32m1))) +void vssseg7e32(float * op0, ptrdiff_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e32_v_f32m1_m))) +void vssseg7e32(vbool32_t op0, float * op1, ptrdiff_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e32_v_f32mf2))) +void vssseg7e32(float * op0, ptrdiff_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e32_v_f32mf2_m))) +void vssseg7e32(vbool64_t op0, float * op1, ptrdiff_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e32_v_f32m1))) +void vssseg8e32(float * op0, ptrdiff_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e32_v_f32m1_m))) +void vssseg8e32(vbool32_t op0, float * op1, ptrdiff_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, vfloat32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e32_v_f32mf2))) +void vssseg8e32(float * op0, ptrdiff_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e32_v_f32mf2_m))) +void vssseg8e32(vbool64_t op0, float * op1, ptrdiff_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_f32m1))) +void vsuxseg2ei8(float * op0, vuint8mf4_t op1, vfloat32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_f32m1_m))) +void vsuxseg2ei8(vbool32_t op0, float * op1, vuint8mf4_t op2, vfloat32m1_t op3, vfloat32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_f32m2))) +void vsuxseg2ei8(float * op0, vuint8mf2_t op1, vfloat32m2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_f32m2_m))) +void vsuxseg2ei8(vbool16_t op0, float * op1, vuint8mf2_t op2, vfloat32m2_t op3, vfloat32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_f32m4))) +void vsuxseg2ei8(float * op0, vuint8m1_t op1, vfloat32m4_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_f32m4_m))) +void vsuxseg2ei8(vbool8_t op0, float * op1, vuint8m1_t op2, vfloat32m4_t op3, vfloat32m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_f32mf2))) +void vsuxseg2ei8(float * op0, vuint8mf8_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_f32mf2_m))) +void vsuxseg2ei8(vbool64_t op0, float * op1, vuint8mf8_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_f32m1))) +void vsuxseg3ei8(float * op0, vuint8mf4_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_f32m1_m))) +void vsuxseg3ei8(vbool32_t op0, float * op1, vuint8mf4_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_f32m2))) +void vsuxseg3ei8(float * op0, vuint8mf2_t op1, vfloat32m2_t op2, vfloat32m2_t op3, vfloat32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_f32m2_m))) +void vsuxseg3ei8(vbool16_t op0, float * op1, vuint8mf2_t op2, vfloat32m2_t op3, vfloat32m2_t op4, vfloat32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_f32mf2))) +void vsuxseg3ei8(float * op0, vuint8mf8_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_f32mf2_m))) +void vsuxseg3ei8(vbool64_t op0, float * op1, vuint8mf8_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_f32m1))) +void vsuxseg4ei8(float * op0, vuint8mf4_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_f32m1_m))) +void vsuxseg4ei8(vbool32_t op0, float * op1, vuint8mf4_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_f32m2))) +void vsuxseg4ei8(float * op0, vuint8mf2_t op1, vfloat32m2_t op2, vfloat32m2_t op3, vfloat32m2_t op4, vfloat32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_f32m2_m))) +void vsuxseg4ei8(vbool16_t op0, float * op1, vuint8mf2_t op2, vfloat32m2_t op3, vfloat32m2_t op4, vfloat32m2_t op5, vfloat32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_f32mf2))) +void vsuxseg4ei8(float * op0, vuint8mf8_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_f32mf2_m))) +void vsuxseg4ei8(vbool64_t op0, float * op1, vuint8mf8_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_f32m1))) +void vsuxseg5ei8(float * op0, vuint8mf4_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_f32m1_m))) +void vsuxseg5ei8(vbool32_t op0, float * op1, vuint8mf4_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_f32mf2))) +void vsuxseg5ei8(float * op0, vuint8mf8_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_f32mf2_m))) +void vsuxseg5ei8(vbool64_t op0, float * op1, vuint8mf8_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_f32m1))) +void vsuxseg6ei8(float * op0, vuint8mf4_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_f32m1_m))) +void vsuxseg6ei8(vbool32_t op0, float * op1, vuint8mf4_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_f32mf2))) +void vsuxseg6ei8(float * op0, vuint8mf8_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_f32mf2_m))) +void vsuxseg6ei8(vbool64_t op0, float * op1, vuint8mf8_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_f32m1))) +void vsuxseg7ei8(float * op0, vuint8mf4_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_f32m1_m))) +void vsuxseg7ei8(vbool32_t op0, float * op1, vuint8mf4_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_f32mf2))) +void vsuxseg7ei8(float * op0, vuint8mf8_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_f32mf2_m))) +void vsuxseg7ei8(vbool64_t op0, float * op1, vuint8mf8_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_f32m1))) +void vsuxseg8ei8(float * op0, vuint8mf4_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_f32m1_m))) +void vsuxseg8ei8(vbool32_t op0, float * op1, vuint8mf4_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, vfloat32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_f32mf2))) +void vsuxseg8ei8(float * op0, vuint8mf8_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_f32mf2_m))) +void vsuxseg8ei8(vbool64_t op0, float * op1, vuint8mf8_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_f32m1))) +void vsuxseg2ei16(float * op0, vuint16mf2_t op1, vfloat32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_f32m1_m))) +void vsuxseg2ei16(vbool32_t op0, float * op1, vuint16mf2_t op2, vfloat32m1_t op3, vfloat32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_f32m2))) +void vsuxseg2ei16(float * op0, vuint16m1_t op1, vfloat32m2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_f32m2_m))) +void vsuxseg2ei16(vbool16_t op0, float * op1, vuint16m1_t op2, vfloat32m2_t op3, vfloat32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_f32m4))) +void vsuxseg2ei16(float * op0, vuint16m2_t op1, vfloat32m4_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_f32m4_m))) +void vsuxseg2ei16(vbool8_t op0, float * op1, vuint16m2_t op2, vfloat32m4_t op3, vfloat32m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_f32mf2))) +void vsuxseg2ei16(float * op0, vuint16mf4_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_f32mf2_m))) +void vsuxseg2ei16(vbool64_t op0, float * op1, vuint16mf4_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_f32m1))) +void vsuxseg3ei16(float * op0, vuint16mf2_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_f32m1_m))) +void vsuxseg3ei16(vbool32_t op0, float * op1, vuint16mf2_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_f32m2))) +void vsuxseg3ei16(float * op0, vuint16m1_t op1, vfloat32m2_t op2, vfloat32m2_t op3, vfloat32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_f32m2_m))) +void vsuxseg3ei16(vbool16_t op0, float * op1, vuint16m1_t op2, vfloat32m2_t op3, vfloat32m2_t op4, vfloat32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_f32mf2))) +void vsuxseg3ei16(float * op0, vuint16mf4_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_f32mf2_m))) +void vsuxseg3ei16(vbool64_t op0, float * op1, vuint16mf4_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_f32m1))) +void vsuxseg4ei16(float * op0, vuint16mf2_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_f32m1_m))) +void vsuxseg4ei16(vbool32_t op0, float * op1, vuint16mf2_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_f32m2))) +void vsuxseg4ei16(float * op0, vuint16m1_t op1, vfloat32m2_t op2, vfloat32m2_t op3, vfloat32m2_t op4, vfloat32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_f32m2_m))) +void vsuxseg4ei16(vbool16_t op0, float * op1, vuint16m1_t op2, vfloat32m2_t op3, vfloat32m2_t op4, vfloat32m2_t op5, vfloat32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_f32mf2))) +void vsuxseg4ei16(float * op0, vuint16mf4_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_f32mf2_m))) +void vsuxseg4ei16(vbool64_t op0, float * op1, vuint16mf4_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_f32m1))) +void vsuxseg5ei16(float * op0, vuint16mf2_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_f32m1_m))) +void vsuxseg5ei16(vbool32_t op0, float * op1, vuint16mf2_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_f32mf2))) +void vsuxseg5ei16(float * op0, vuint16mf4_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_f32mf2_m))) +void vsuxseg5ei16(vbool64_t op0, float * op1, vuint16mf4_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_f32m1))) +void vsuxseg6ei16(float * op0, vuint16mf2_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_f32m1_m))) +void vsuxseg6ei16(vbool32_t op0, float * op1, vuint16mf2_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_f32mf2))) +void vsuxseg6ei16(float * op0, vuint16mf4_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_f32mf2_m))) +void vsuxseg6ei16(vbool64_t op0, float * op1, vuint16mf4_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_f32m1))) +void vsuxseg7ei16(float * op0, vuint16mf2_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_f32m1_m))) +void vsuxseg7ei16(vbool32_t op0, float * op1, vuint16mf2_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_f32mf2))) +void vsuxseg7ei16(float * op0, vuint16mf4_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_f32mf2_m))) +void vsuxseg7ei16(vbool64_t op0, float * op1, vuint16mf4_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_f32m1))) +void vsuxseg8ei16(float * op0, vuint16mf2_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_f32m1_m))) +void vsuxseg8ei16(vbool32_t op0, float * op1, vuint16mf2_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, vfloat32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_f32mf2))) +void vsuxseg8ei16(float * op0, vuint16mf4_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_f32mf2_m))) +void vsuxseg8ei16(vbool64_t op0, float * op1, vuint16mf4_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_f32m1))) +void vsuxseg2ei32(float * op0, vuint32m1_t op1, vfloat32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_f32m1_m))) +void vsuxseg2ei32(vbool32_t op0, float * op1, vuint32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_f32m2))) +void vsuxseg2ei32(float * op0, vuint32m2_t op1, vfloat32m2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_f32m2_m))) +void vsuxseg2ei32(vbool16_t op0, float * op1, vuint32m2_t op2, vfloat32m2_t op3, vfloat32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_f32m4))) +void vsuxseg2ei32(float * op0, vuint32m4_t op1, vfloat32m4_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_f32m4_m))) +void vsuxseg2ei32(vbool8_t op0, float * op1, vuint32m4_t op2, vfloat32m4_t op3, vfloat32m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_f32mf2))) +void vsuxseg2ei32(float * op0, vuint32mf2_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_f32mf2_m))) +void vsuxseg2ei32(vbool64_t op0, float * op1, vuint32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_f32m1))) +void vsuxseg3ei32(float * op0, vuint32m1_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_f32m1_m))) +void vsuxseg3ei32(vbool32_t op0, float * op1, vuint32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_f32m2))) +void vsuxseg3ei32(float * op0, vuint32m2_t op1, vfloat32m2_t op2, vfloat32m2_t op3, vfloat32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_f32m2_m))) +void vsuxseg3ei32(vbool16_t op0, float * op1, vuint32m2_t op2, vfloat32m2_t op3, vfloat32m2_t op4, vfloat32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_f32mf2))) +void vsuxseg3ei32(float * op0, vuint32mf2_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_f32mf2_m))) +void vsuxseg3ei32(vbool64_t op0, float * op1, vuint32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_f32m1))) +void vsuxseg4ei32(float * op0, vuint32m1_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_f32m1_m))) +void vsuxseg4ei32(vbool32_t op0, float * op1, vuint32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_f32m2))) +void vsuxseg4ei32(float * op0, vuint32m2_t op1, vfloat32m2_t op2, vfloat32m2_t op3, vfloat32m2_t op4, vfloat32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_f32m2_m))) +void vsuxseg4ei32(vbool16_t op0, float * op1, vuint32m2_t op2, vfloat32m2_t op3, vfloat32m2_t op4, vfloat32m2_t op5, vfloat32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_f32mf2))) +void vsuxseg4ei32(float * op0, vuint32mf2_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_f32mf2_m))) +void vsuxseg4ei32(vbool64_t op0, float * op1, vuint32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_f32m1))) +void vsuxseg5ei32(float * op0, vuint32m1_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_f32m1_m))) +void vsuxseg5ei32(vbool32_t op0, float * op1, vuint32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_f32mf2))) +void vsuxseg5ei32(float * op0, vuint32mf2_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_f32mf2_m))) +void vsuxseg5ei32(vbool64_t op0, float * op1, vuint32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_f32m1))) +void vsuxseg6ei32(float * op0, vuint32m1_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_f32m1_m))) +void vsuxseg6ei32(vbool32_t op0, float * op1, vuint32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_f32mf2))) +void vsuxseg6ei32(float * op0, vuint32mf2_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_f32mf2_m))) +void vsuxseg6ei32(vbool64_t op0, float * op1, vuint32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_f32m1))) +void vsuxseg7ei32(float * op0, vuint32m1_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_f32m1_m))) +void vsuxseg7ei32(vbool32_t op0, float * op1, vuint32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_f32mf2))) +void vsuxseg7ei32(float * op0, vuint32mf2_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_f32mf2_m))) +void vsuxseg7ei32(vbool64_t op0, float * op1, vuint32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_f32m1))) +void vsuxseg8ei32(float * op0, vuint32m1_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_f32m1_m))) +void vsuxseg8ei32(vbool32_t op0, float * op1, vuint32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, vfloat32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_f32mf2))) +void vsuxseg8ei32(float * op0, vuint32mf2_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_f32mf2_m))) +void vsuxseg8ei32(vbool64_t op0, float * op1, vuint32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_f32m1))) +void vsuxseg2ei64(float * op0, vuint64m2_t op1, vfloat32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_f32m1_m))) +void vsuxseg2ei64(vbool32_t op0, float * op1, vuint64m2_t op2, vfloat32m1_t op3, vfloat32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_f32m2))) +void vsuxseg2ei64(float * op0, vuint64m4_t op1, vfloat32m2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_f32m2_m))) +void vsuxseg2ei64(vbool16_t op0, float * op1, vuint64m4_t op2, vfloat32m2_t op3, vfloat32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_f32m4))) +void vsuxseg2ei64(float * op0, vuint64m8_t op1, vfloat32m4_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_f32m4_m))) +void vsuxseg2ei64(vbool8_t op0, float * op1, vuint64m8_t op2, vfloat32m4_t op3, vfloat32m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_f32mf2))) +void vsuxseg2ei64(float * op0, vuint64m1_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_f32mf2_m))) +void vsuxseg2ei64(vbool64_t op0, float * op1, vuint64m1_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_f32m1))) +void vsuxseg3ei64(float * op0, vuint64m2_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_f32m1_m))) +void vsuxseg3ei64(vbool32_t op0, float * op1, vuint64m2_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_f32m2))) +void vsuxseg3ei64(float * op0, vuint64m4_t op1, vfloat32m2_t op2, vfloat32m2_t op3, vfloat32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_f32m2_m))) +void vsuxseg3ei64(vbool16_t op0, float * op1, vuint64m4_t op2, vfloat32m2_t op3, vfloat32m2_t op4, vfloat32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_f32mf2))) +void vsuxseg3ei64(float * op0, vuint64m1_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_f32mf2_m))) +void vsuxseg3ei64(vbool64_t op0, float * op1, vuint64m1_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_f32m1))) +void vsuxseg4ei64(float * op0, vuint64m2_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_f32m1_m))) +void vsuxseg4ei64(vbool32_t op0, float * op1, vuint64m2_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_f32m2))) +void vsuxseg4ei64(float * op0, vuint64m4_t op1, vfloat32m2_t op2, vfloat32m2_t op3, vfloat32m2_t op4, vfloat32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_f32m2_m))) +void vsuxseg4ei64(vbool16_t op0, float * op1, vuint64m4_t op2, vfloat32m2_t op3, vfloat32m2_t op4, vfloat32m2_t op5, vfloat32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_f32mf2))) +void vsuxseg4ei64(float * op0, vuint64m1_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_f32mf2_m))) +void vsuxseg4ei64(vbool64_t op0, float * op1, vuint64m1_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_f32m1))) +void vsuxseg5ei64(float * op0, vuint64m2_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_f32m1_m))) +void vsuxseg5ei64(vbool32_t op0, float * op1, vuint64m2_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_f32mf2))) +void vsuxseg5ei64(float * op0, vuint64m1_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_f32mf2_m))) +void vsuxseg5ei64(vbool64_t op0, float * op1, vuint64m1_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_f32m1))) +void vsuxseg6ei64(float * op0, vuint64m2_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_f32m1_m))) +void vsuxseg6ei64(vbool32_t op0, float * op1, vuint64m2_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_f32mf2))) +void vsuxseg6ei64(float * op0, vuint64m1_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_f32mf2_m))) +void vsuxseg6ei64(vbool64_t op0, float * op1, vuint64m1_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_f32m1))) +void vsuxseg7ei64(float * op0, vuint64m2_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_f32m1_m))) +void vsuxseg7ei64(vbool32_t op0, float * op1, vuint64m2_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_f32mf2))) +void vsuxseg7ei64(float * op0, vuint64m1_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_f32mf2_m))) +void vsuxseg7ei64(vbool64_t op0, float * op1, vuint64m1_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_f32m1))) +void vsuxseg8ei64(float * op0, vuint64m2_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_f32m1_m))) +void vsuxseg8ei64(vbool32_t op0, float * op1, vuint64m2_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, vfloat32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_f32mf2))) +void vsuxseg8ei64(float * op0, vuint64m1_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_f32mf2_m))) +void vsuxseg8ei64(vbool64_t op0, float * op1, vuint64m1_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_f32m1))) +void vsoxseg2ei8(float * op0, vuint8mf4_t op1, vfloat32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_f32m1_m))) +void vsoxseg2ei8(vbool32_t op0, float * op1, vuint8mf4_t op2, vfloat32m1_t op3, vfloat32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_f32m2))) +void vsoxseg2ei8(float * op0, vuint8mf2_t op1, vfloat32m2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_f32m2_m))) +void vsoxseg2ei8(vbool16_t op0, float * op1, vuint8mf2_t op2, vfloat32m2_t op3, vfloat32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_f32m4))) +void vsoxseg2ei8(float * op0, vuint8m1_t op1, vfloat32m4_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_f32m4_m))) +void vsoxseg2ei8(vbool8_t op0, float * op1, vuint8m1_t op2, vfloat32m4_t op3, vfloat32m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_f32mf2))) +void vsoxseg2ei8(float * op0, vuint8mf8_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_f32mf2_m))) +void vsoxseg2ei8(vbool64_t op0, float * op1, vuint8mf8_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_f32m1))) +void vsoxseg3ei8(float * op0, vuint8mf4_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_f32m1_m))) +void vsoxseg3ei8(vbool32_t op0, float * op1, vuint8mf4_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_f32m2))) +void vsoxseg3ei8(float * op0, vuint8mf2_t op1, vfloat32m2_t op2, vfloat32m2_t op3, vfloat32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_f32m2_m))) +void vsoxseg3ei8(vbool16_t op0, float * op1, vuint8mf2_t op2, vfloat32m2_t op3, vfloat32m2_t op4, vfloat32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_f32mf2))) +void vsoxseg3ei8(float * op0, vuint8mf8_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_f32mf2_m))) +void vsoxseg3ei8(vbool64_t op0, float * op1, vuint8mf8_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_f32m1))) +void vsoxseg4ei8(float * op0, vuint8mf4_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_f32m1_m))) +void vsoxseg4ei8(vbool32_t op0, float * op1, vuint8mf4_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_f32m2))) +void vsoxseg4ei8(float * op0, vuint8mf2_t op1, vfloat32m2_t op2, vfloat32m2_t op3, vfloat32m2_t op4, vfloat32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_f32m2_m))) +void vsoxseg4ei8(vbool16_t op0, float * op1, vuint8mf2_t op2, vfloat32m2_t op3, vfloat32m2_t op4, vfloat32m2_t op5, vfloat32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_f32mf2))) +void vsoxseg4ei8(float * op0, vuint8mf8_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_f32mf2_m))) +void vsoxseg4ei8(vbool64_t op0, float * op1, vuint8mf8_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_f32m1))) +void vsoxseg5ei8(float * op0, vuint8mf4_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_f32m1_m))) +void vsoxseg5ei8(vbool32_t op0, float * op1, vuint8mf4_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_f32mf2))) +void vsoxseg5ei8(float * op0, vuint8mf8_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_f32mf2_m))) +void vsoxseg5ei8(vbool64_t op0, float * op1, vuint8mf8_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_f32m1))) +void vsoxseg6ei8(float * op0, vuint8mf4_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_f32m1_m))) +void vsoxseg6ei8(vbool32_t op0, float * op1, vuint8mf4_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_f32mf2))) +void vsoxseg6ei8(float * op0, vuint8mf8_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_f32mf2_m))) +void vsoxseg6ei8(vbool64_t op0, float * op1, vuint8mf8_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_f32m1))) +void vsoxseg7ei8(float * op0, vuint8mf4_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_f32m1_m))) +void vsoxseg7ei8(vbool32_t op0, float * op1, vuint8mf4_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_f32mf2))) +void vsoxseg7ei8(float * op0, vuint8mf8_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_f32mf2_m))) +void vsoxseg7ei8(vbool64_t op0, float * op1, vuint8mf8_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_f32m1))) +void vsoxseg8ei8(float * op0, vuint8mf4_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_f32m1_m))) +void vsoxseg8ei8(vbool32_t op0, float * op1, vuint8mf4_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, vfloat32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_f32mf2))) +void vsoxseg8ei8(float * op0, vuint8mf8_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_f32mf2_m))) +void vsoxseg8ei8(vbool64_t op0, float * op1, vuint8mf8_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_f32m1))) +void vsoxseg2ei16(float * op0, vuint16mf2_t op1, vfloat32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_f32m1_m))) +void vsoxseg2ei16(vbool32_t op0, float * op1, vuint16mf2_t op2, vfloat32m1_t op3, vfloat32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_f32m2))) +void vsoxseg2ei16(float * op0, vuint16m1_t op1, vfloat32m2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_f32m2_m))) +void vsoxseg2ei16(vbool16_t op0, float * op1, vuint16m1_t op2, vfloat32m2_t op3, vfloat32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_f32m4))) +void vsoxseg2ei16(float * op0, vuint16m2_t op1, vfloat32m4_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_f32m4_m))) +void vsoxseg2ei16(vbool8_t op0, float * op1, vuint16m2_t op2, vfloat32m4_t op3, vfloat32m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_f32mf2))) +void vsoxseg2ei16(float * op0, vuint16mf4_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_f32mf2_m))) +void vsoxseg2ei16(vbool64_t op0, float * op1, vuint16mf4_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_f32m1))) +void vsoxseg3ei16(float * op0, vuint16mf2_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_f32m1_m))) +void vsoxseg3ei16(vbool32_t op0, float * op1, vuint16mf2_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_f32m2))) +void vsoxseg3ei16(float * op0, vuint16m1_t op1, vfloat32m2_t op2, vfloat32m2_t op3, vfloat32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_f32m2_m))) +void vsoxseg3ei16(vbool16_t op0, float * op1, vuint16m1_t op2, vfloat32m2_t op3, vfloat32m2_t op4, vfloat32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_f32mf2))) +void vsoxseg3ei16(float * op0, vuint16mf4_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_f32mf2_m))) +void vsoxseg3ei16(vbool64_t op0, float * op1, vuint16mf4_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_f32m1))) +void vsoxseg4ei16(float * op0, vuint16mf2_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_f32m1_m))) +void vsoxseg4ei16(vbool32_t op0, float * op1, vuint16mf2_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_f32m2))) +void vsoxseg4ei16(float * op0, vuint16m1_t op1, vfloat32m2_t op2, vfloat32m2_t op3, vfloat32m2_t op4, vfloat32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_f32m2_m))) +void vsoxseg4ei16(vbool16_t op0, float * op1, vuint16m1_t op2, vfloat32m2_t op3, vfloat32m2_t op4, vfloat32m2_t op5, vfloat32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_f32mf2))) +void vsoxseg4ei16(float * op0, vuint16mf4_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_f32mf2_m))) +void vsoxseg4ei16(vbool64_t op0, float * op1, vuint16mf4_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_f32m1))) +void vsoxseg5ei16(float * op0, vuint16mf2_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_f32m1_m))) +void vsoxseg5ei16(vbool32_t op0, float * op1, vuint16mf2_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_f32mf2))) +void vsoxseg5ei16(float * op0, vuint16mf4_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_f32mf2_m))) +void vsoxseg5ei16(vbool64_t op0, float * op1, vuint16mf4_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_f32m1))) +void vsoxseg6ei16(float * op0, vuint16mf2_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_f32m1_m))) +void vsoxseg6ei16(vbool32_t op0, float * op1, vuint16mf2_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_f32mf2))) +void vsoxseg6ei16(float * op0, vuint16mf4_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_f32mf2_m))) +void vsoxseg6ei16(vbool64_t op0, float * op1, vuint16mf4_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_f32m1))) +void vsoxseg7ei16(float * op0, vuint16mf2_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_f32m1_m))) +void vsoxseg7ei16(vbool32_t op0, float * op1, vuint16mf2_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_f32mf2))) +void vsoxseg7ei16(float * op0, vuint16mf4_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_f32mf2_m))) +void vsoxseg7ei16(vbool64_t op0, float * op1, vuint16mf4_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_f32m1))) +void vsoxseg8ei16(float * op0, vuint16mf2_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_f32m1_m))) +void vsoxseg8ei16(vbool32_t op0, float * op1, vuint16mf2_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, vfloat32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_f32mf2))) +void vsoxseg8ei16(float * op0, vuint16mf4_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_f32mf2_m))) +void vsoxseg8ei16(vbool64_t op0, float * op1, vuint16mf4_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_f32m1))) +void vsoxseg2ei32(float * op0, vuint32m1_t op1, vfloat32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_f32m1_m))) +void vsoxseg2ei32(vbool32_t op0, float * op1, vuint32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_f32m2))) +void vsoxseg2ei32(float * op0, vuint32m2_t op1, vfloat32m2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_f32m2_m))) +void vsoxseg2ei32(vbool16_t op0, float * op1, vuint32m2_t op2, vfloat32m2_t op3, vfloat32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_f32m4))) +void vsoxseg2ei32(float * op0, vuint32m4_t op1, vfloat32m4_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_f32m4_m))) +void vsoxseg2ei32(vbool8_t op0, float * op1, vuint32m4_t op2, vfloat32m4_t op3, vfloat32m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_f32mf2))) +void vsoxseg2ei32(float * op0, vuint32mf2_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_f32mf2_m))) +void vsoxseg2ei32(vbool64_t op0, float * op1, vuint32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_f32m1))) +void vsoxseg3ei32(float * op0, vuint32m1_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_f32m1_m))) +void vsoxseg3ei32(vbool32_t op0, float * op1, vuint32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_f32m2))) +void vsoxseg3ei32(float * op0, vuint32m2_t op1, vfloat32m2_t op2, vfloat32m2_t op3, vfloat32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_f32m2_m))) +void vsoxseg3ei32(vbool16_t op0, float * op1, vuint32m2_t op2, vfloat32m2_t op3, vfloat32m2_t op4, vfloat32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_f32mf2))) +void vsoxseg3ei32(float * op0, vuint32mf2_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_f32mf2_m))) +void vsoxseg3ei32(vbool64_t op0, float * op1, vuint32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_f32m1))) +void vsoxseg4ei32(float * op0, vuint32m1_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_f32m1_m))) +void vsoxseg4ei32(vbool32_t op0, float * op1, vuint32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_f32m2))) +void vsoxseg4ei32(float * op0, vuint32m2_t op1, vfloat32m2_t op2, vfloat32m2_t op3, vfloat32m2_t op4, vfloat32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_f32m2_m))) +void vsoxseg4ei32(vbool16_t op0, float * op1, vuint32m2_t op2, vfloat32m2_t op3, vfloat32m2_t op4, vfloat32m2_t op5, vfloat32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_f32mf2))) +void vsoxseg4ei32(float * op0, vuint32mf2_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_f32mf2_m))) +void vsoxseg4ei32(vbool64_t op0, float * op1, vuint32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_f32m1))) +void vsoxseg5ei32(float * op0, vuint32m1_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_f32m1_m))) +void vsoxseg5ei32(vbool32_t op0, float * op1, vuint32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_f32mf2))) +void vsoxseg5ei32(float * op0, vuint32mf2_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_f32mf2_m))) +void vsoxseg5ei32(vbool64_t op0, float * op1, vuint32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_f32m1))) +void vsoxseg6ei32(float * op0, vuint32m1_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_f32m1_m))) +void vsoxseg6ei32(vbool32_t op0, float * op1, vuint32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_f32mf2))) +void vsoxseg6ei32(float * op0, vuint32mf2_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_f32mf2_m))) +void vsoxseg6ei32(vbool64_t op0, float * op1, vuint32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_f32m1))) +void vsoxseg7ei32(float * op0, vuint32m1_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_f32m1_m))) +void vsoxseg7ei32(vbool32_t op0, float * op1, vuint32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_f32mf2))) +void vsoxseg7ei32(float * op0, vuint32mf2_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_f32mf2_m))) +void vsoxseg7ei32(vbool64_t op0, float * op1, vuint32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_f32m1))) +void vsoxseg8ei32(float * op0, vuint32m1_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_f32m1_m))) +void vsoxseg8ei32(vbool32_t op0, float * op1, vuint32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, vfloat32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_f32mf2))) +void vsoxseg8ei32(float * op0, vuint32mf2_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_f32mf2_m))) +void vsoxseg8ei32(vbool64_t op0, float * op1, vuint32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_f32m1))) +void vsoxseg2ei64(float * op0, vuint64m2_t op1, vfloat32m1_t op2, vfloat32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_f32m1_m))) +void vsoxseg2ei64(vbool32_t op0, float * op1, vuint64m2_t op2, vfloat32m1_t op3, vfloat32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_f32m2))) +void vsoxseg2ei64(float * op0, vuint64m4_t op1, vfloat32m2_t op2, vfloat32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_f32m2_m))) +void vsoxseg2ei64(vbool16_t op0, float * op1, vuint64m4_t op2, vfloat32m2_t op3, vfloat32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_f32m4))) +void vsoxseg2ei64(float * op0, vuint64m8_t op1, vfloat32m4_t op2, vfloat32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_f32m4_m))) +void vsoxseg2ei64(vbool8_t op0, float * op1, vuint64m8_t op2, vfloat32m4_t op3, vfloat32m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_f32mf2))) +void vsoxseg2ei64(float * op0, vuint64m1_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_f32mf2_m))) +void vsoxseg2ei64(vbool64_t op0, float * op1, vuint64m1_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_f32m1))) +void vsoxseg3ei64(float * op0, vuint64m2_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_f32m1_m))) +void vsoxseg3ei64(vbool32_t op0, float * op1, vuint64m2_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_f32m2))) +void vsoxseg3ei64(float * op0, vuint64m4_t op1, vfloat32m2_t op2, vfloat32m2_t op3, vfloat32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_f32m2_m))) +void vsoxseg3ei64(vbool16_t op0, float * op1, vuint64m4_t op2, vfloat32m2_t op3, vfloat32m2_t op4, vfloat32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_f32mf2))) +void vsoxseg3ei64(float * op0, vuint64m1_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_f32mf2_m))) +void vsoxseg3ei64(vbool64_t op0, float * op1, vuint64m1_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_f32m1))) +void vsoxseg4ei64(float * op0, vuint64m2_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_f32m1_m))) +void vsoxseg4ei64(vbool32_t op0, float * op1, vuint64m2_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_f32m2))) +void vsoxseg4ei64(float * op0, vuint64m4_t op1, vfloat32m2_t op2, vfloat32m2_t op3, vfloat32m2_t op4, vfloat32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_f32m2_m))) +void vsoxseg4ei64(vbool16_t op0, float * op1, vuint64m4_t op2, vfloat32m2_t op3, vfloat32m2_t op4, vfloat32m2_t op5, vfloat32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_f32mf2))) +void vsoxseg4ei64(float * op0, vuint64m1_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_f32mf2_m))) +void vsoxseg4ei64(vbool64_t op0, float * op1, vuint64m1_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_f32m1))) +void vsoxseg5ei64(float * op0, vuint64m2_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_f32m1_m))) +void vsoxseg5ei64(vbool32_t op0, float * op1, vuint64m2_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_f32mf2))) +void vsoxseg5ei64(float * op0, vuint64m1_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_f32mf2_m))) +void vsoxseg5ei64(vbool64_t op0, float * op1, vuint64m1_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_f32m1))) +void vsoxseg6ei64(float * op0, vuint64m2_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_f32m1_m))) +void vsoxseg6ei64(vbool32_t op0, float * op1, vuint64m2_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_f32mf2))) +void vsoxseg6ei64(float * op0, vuint64m1_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_f32mf2_m))) +void vsoxseg6ei64(vbool64_t op0, float * op1, vuint64m1_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_f32m1))) +void vsoxseg7ei64(float * op0, vuint64m2_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_f32m1_m))) +void vsoxseg7ei64(vbool32_t op0, float * op1, vuint64m2_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_f32mf2))) +void vsoxseg7ei64(float * op0, vuint64m1_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_f32mf2_m))) +void vsoxseg7ei64(vbool64_t op0, float * op1, vuint64m1_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_f32m1))) +void vsoxseg8ei64(float * op0, vuint64m2_t op1, vfloat32m1_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_f32m1_m))) +void vsoxseg8ei64(vbool32_t op0, float * op1, vuint64m2_t op2, vfloat32m1_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, vfloat32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_f32mf2))) +void vsoxseg8ei64(float * op0, vuint64m1_t op1, vfloat32mf2_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_f32mf2_m))) +void vsoxseg8ei64(vbool64_t op0, float * op1, vuint64m1_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e32_v_f32m1_m))) +void vlseg2e32(vfloat32m1_t * op0, vfloat32m1_t * op1, vbool32_t op2, vfloat32m1_t op3, vfloat32m1_t op4, const float * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e32_v_f32m2_m))) +void vlseg2e32(vfloat32m2_t * op0, vfloat32m2_t * op1, vbool16_t op2, vfloat32m2_t op3, vfloat32m2_t op4, const float * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e32_v_f32m4_m))) +void vlseg2e32(vfloat32m4_t * op0, vfloat32m4_t * op1, vbool8_t op2, vfloat32m4_t op3, vfloat32m4_t op4, const float * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e32_v_f32mf2_m))) +void vlseg2e32(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vbool64_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, const float * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e32_v_f32m1_m))) +void vlseg3e32(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vbool32_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, const float * op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e32_v_f32m2_m))) +void vlseg3e32(vfloat32m2_t * op0, vfloat32m2_t * op1, vfloat32m2_t * op2, vbool16_t op3, vfloat32m2_t op4, vfloat32m2_t op5, vfloat32m2_t op6, const float * op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e32_v_f32mf2_m))) +void vlseg3e32(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vbool64_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, const float * op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e32_v_f32m1_m))) +void vlseg4e32(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vbool32_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, const float * op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e32_v_f32m2_m))) +void vlseg4e32(vfloat32m2_t * op0, vfloat32m2_t * op1, vfloat32m2_t * op2, vfloat32m2_t * op3, vbool16_t op4, vfloat32m2_t op5, vfloat32m2_t op6, vfloat32m2_t op7, vfloat32m2_t op8, const float * op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e32_v_f32mf2_m))) +void vlseg4e32(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vbool64_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, const float * op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e32_v_f32m1_m))) +void vlseg5e32(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vbool32_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, vfloat32m1_t op10, const float * op11, size_t op12); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e32_v_f32mf2_m))) +void vlseg5e32(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vbool64_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, const float * op11, size_t op12); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e32_v_f32m1_m))) +void vlseg6e32(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vbool32_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, vfloat32m1_t op10, vfloat32m1_t op11, vfloat32m1_t op12, const float * op13, size_t op14); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e32_v_f32mf2_m))) +void vlseg6e32(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vbool64_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, vfloat32mf2_t op11, vfloat32mf2_t op12, const float * op13, size_t op14); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e32_v_f32m1_m))) +void vlseg7e32(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vfloat32m1_t * op6, vbool32_t op7, vfloat32m1_t op8, vfloat32m1_t op9, vfloat32m1_t op10, vfloat32m1_t op11, vfloat32m1_t op12, vfloat32m1_t op13, vfloat32m1_t op14, const float * op15, size_t op16); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e32_v_f32mf2_m))) +void vlseg7e32(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vfloat32mf2_t * op6, vbool64_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, vfloat32mf2_t op11, vfloat32mf2_t op12, vfloat32mf2_t op13, vfloat32mf2_t op14, const float * op15, size_t op16); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e32_v_f32m1_m))) +void vlseg8e32(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vfloat32m1_t * op6, vfloat32m1_t * op7, vbool32_t op8, vfloat32m1_t op9, vfloat32m1_t op10, vfloat32m1_t op11, vfloat32m1_t op12, vfloat32m1_t op13, vfloat32m1_t op14, vfloat32m1_t op15, vfloat32m1_t op16, const float * op17, size_t op18); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e32_v_f32mf2_m))) +void vlseg8e32(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vfloat32mf2_t * op6, vfloat32mf2_t * op7, vbool64_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, vfloat32mf2_t op11, vfloat32mf2_t op12, vfloat32mf2_t op13, vfloat32mf2_t op14, vfloat32mf2_t op15, vfloat32mf2_t op16, const float * op17, size_t op18); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e32ff_v_f32m1_m))) +void vlseg2e32ff(vfloat32m1_t * op0, vfloat32m1_t * op1, vbool32_t op2, vfloat32m1_t op3, vfloat32m1_t op4, const float * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e32ff_v_f32m2_m))) +void vlseg2e32ff(vfloat32m2_t * op0, vfloat32m2_t * op1, vbool16_t op2, vfloat32m2_t op3, vfloat32m2_t op4, const float * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e32ff_v_f32m4_m))) +void vlseg2e32ff(vfloat32m4_t * op0, vfloat32m4_t * op1, vbool8_t op2, vfloat32m4_t op3, vfloat32m4_t op4, const float * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e32ff_v_f32mf2_m))) +void vlseg2e32ff(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vbool64_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, const float * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e32ff_v_f32m1_m))) +void vlseg3e32ff(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vbool32_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, const float * op7, size_t * op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e32ff_v_f32m2_m))) +void vlseg3e32ff(vfloat32m2_t * op0, vfloat32m2_t * op1, vfloat32m2_t * op2, vbool16_t op3, vfloat32m2_t op4, vfloat32m2_t op5, vfloat32m2_t op6, const float * op7, size_t * op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e32ff_v_f32mf2_m))) +void vlseg3e32ff(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vbool64_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, const float * op7, size_t * op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e32ff_v_f32m1_m))) +void vlseg4e32ff(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vbool32_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, const float * op9, size_t * op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e32ff_v_f32m2_m))) +void vlseg4e32ff(vfloat32m2_t * op0, vfloat32m2_t * op1, vfloat32m2_t * op2, vfloat32m2_t * op3, vbool16_t op4, vfloat32m2_t op5, vfloat32m2_t op6, vfloat32m2_t op7, vfloat32m2_t op8, const float * op9, size_t * op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e32ff_v_f32mf2_m))) +void vlseg4e32ff(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vbool64_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, const float * op9, size_t * op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e32ff_v_f32m1_m))) +void vlseg5e32ff(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vbool32_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, vfloat32m1_t op10, const float * op11, size_t * op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e32ff_v_f32mf2_m))) +void vlseg5e32ff(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vbool64_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, const float * op11, size_t * op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e32ff_v_f32m1_m))) +void vlseg6e32ff(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vbool32_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, vfloat32m1_t op10, vfloat32m1_t op11, vfloat32m1_t op12, const float * op13, size_t * op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e32ff_v_f32mf2_m))) +void vlseg6e32ff(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vbool64_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, vfloat32mf2_t op11, vfloat32mf2_t op12, const float * op13, size_t * op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e32ff_v_f32m1_m))) +void vlseg7e32ff(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vfloat32m1_t * op6, vbool32_t op7, vfloat32m1_t op8, vfloat32m1_t op9, vfloat32m1_t op10, vfloat32m1_t op11, vfloat32m1_t op12, vfloat32m1_t op13, vfloat32m1_t op14, const float * op15, size_t * op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e32ff_v_f32mf2_m))) +void vlseg7e32ff(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vfloat32mf2_t * op6, vbool64_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, vfloat32mf2_t op11, vfloat32mf2_t op12, vfloat32mf2_t op13, vfloat32mf2_t op14, const float * op15, size_t * op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e32ff_v_f32m1_m))) +void vlseg8e32ff(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vfloat32m1_t * op6, vfloat32m1_t * op7, vbool32_t op8, vfloat32m1_t op9, vfloat32m1_t op10, vfloat32m1_t op11, vfloat32m1_t op12, vfloat32m1_t op13, vfloat32m1_t op14, vfloat32m1_t op15, vfloat32m1_t op16, const float * op17, size_t * op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e32ff_v_f32mf2_m))) +void vlseg8e32ff(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vfloat32mf2_t * op6, vfloat32mf2_t * op7, vbool64_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, vfloat32mf2_t op11, vfloat32mf2_t op12, vfloat32mf2_t op13, vfloat32mf2_t op14, vfloat32mf2_t op15, vfloat32mf2_t op16, const float * op17, size_t * op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e32_v_f32m1_m))) +void vlsseg2e32(vfloat32m1_t * op0, vfloat32m1_t * op1, vbool32_t op2, vfloat32m1_t op3, vfloat32m1_t op4, const float * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e32_v_f32m2_m))) +void vlsseg2e32(vfloat32m2_t * op0, vfloat32m2_t * op1, vbool16_t op2, vfloat32m2_t op3, vfloat32m2_t op4, const float * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e32_v_f32m4_m))) +void vlsseg2e32(vfloat32m4_t * op0, vfloat32m4_t * op1, vbool8_t op2, vfloat32m4_t op3, vfloat32m4_t op4, const float * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e32_v_f32mf2_m))) +void vlsseg2e32(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vbool64_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, const float * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg3e32_v_f32m1_m))) +void vlsseg3e32(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vbool32_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, const float * op7, ptrdiff_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg3e32_v_f32m2_m))) +void vlsseg3e32(vfloat32m2_t * op0, vfloat32m2_t * op1, vfloat32m2_t * op2, vbool16_t op3, vfloat32m2_t op4, vfloat32m2_t op5, vfloat32m2_t op6, const float * op7, ptrdiff_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg3e32_v_f32mf2_m))) +void vlsseg3e32(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vbool64_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, const float * op7, ptrdiff_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg4e32_v_f32m1_m))) +void vlsseg4e32(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vbool32_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, const float * op9, ptrdiff_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg4e32_v_f32m2_m))) +void vlsseg4e32(vfloat32m2_t * op0, vfloat32m2_t * op1, vfloat32m2_t * op2, vfloat32m2_t * op3, vbool16_t op4, vfloat32m2_t op5, vfloat32m2_t op6, vfloat32m2_t op7, vfloat32m2_t op8, const float * op9, ptrdiff_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg4e32_v_f32mf2_m))) +void vlsseg4e32(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vbool64_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, const float * op9, ptrdiff_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg5e32_v_f32m1_m))) +void vlsseg5e32(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vbool32_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, vfloat32m1_t op10, const float * op11, ptrdiff_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg5e32_v_f32mf2_m))) +void vlsseg5e32(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vbool64_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, const float * op11, ptrdiff_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg6e32_v_f32m1_m))) +void vlsseg6e32(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vbool32_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, vfloat32m1_t op10, vfloat32m1_t op11, vfloat32m1_t op12, const float * op13, ptrdiff_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg6e32_v_f32mf2_m))) +void vlsseg6e32(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vbool64_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, vfloat32mf2_t op11, vfloat32mf2_t op12, const float * op13, ptrdiff_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg7e32_v_f32m1_m))) +void vlsseg7e32(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vfloat32m1_t * op6, vbool32_t op7, vfloat32m1_t op8, vfloat32m1_t op9, vfloat32m1_t op10, vfloat32m1_t op11, vfloat32m1_t op12, vfloat32m1_t op13, vfloat32m1_t op14, const float * op15, ptrdiff_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg7e32_v_f32mf2_m))) +void vlsseg7e32(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vfloat32mf2_t * op6, vbool64_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, vfloat32mf2_t op11, vfloat32mf2_t op12, vfloat32mf2_t op13, vfloat32mf2_t op14, const float * op15, ptrdiff_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg8e32_v_f32m1_m))) +void vlsseg8e32(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vfloat32m1_t * op6, vfloat32m1_t * op7, vbool32_t op8, vfloat32m1_t op9, vfloat32m1_t op10, vfloat32m1_t op11, vfloat32m1_t op12, vfloat32m1_t op13, vfloat32m1_t op14, vfloat32m1_t op15, vfloat32m1_t op16, const float * op17, ptrdiff_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg8e32_v_f32mf2_m))) +void vlsseg8e32(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vfloat32mf2_t * op6, vfloat32mf2_t * op7, vbool64_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, vfloat32mf2_t op11, vfloat32mf2_t op12, vfloat32mf2_t op13, vfloat32mf2_t op14, vfloat32mf2_t op15, vfloat32mf2_t op16, const float * op17, ptrdiff_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_f32m1))) +void vluxseg2ei8(vfloat32m1_t * op0, vfloat32m1_t * op1, const float * op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_f32m1_m))) +void vluxseg2ei8(vfloat32m1_t * op0, vfloat32m1_t * op1, vbool32_t op2, vfloat32m1_t op3, vfloat32m1_t op4, const float * op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_f32m2))) +void vluxseg2ei8(vfloat32m2_t * op0, vfloat32m2_t * op1, const float * op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_f32m2_m))) +void vluxseg2ei8(vfloat32m2_t * op0, vfloat32m2_t * op1, vbool16_t op2, vfloat32m2_t op3, vfloat32m2_t op4, const float * op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_f32m4))) +void vluxseg2ei8(vfloat32m4_t * op0, vfloat32m4_t * op1, const float * op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_f32m4_m))) +void vluxseg2ei8(vfloat32m4_t * op0, vfloat32m4_t * op1, vbool8_t op2, vfloat32m4_t op3, vfloat32m4_t op4, const float * op5, vuint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_f32mf2))) +void vluxseg2ei8(vfloat32mf2_t * op0, vfloat32mf2_t * op1, const float * op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_f32mf2_m))) +void vluxseg2ei8(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vbool64_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, const float * op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_f32m1))) +void vluxseg3ei8(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, const float * op3, vuint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_f32m1_m))) +void vluxseg3ei8(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vbool32_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, const float * op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_f32m2))) +void vluxseg3ei8(vfloat32m2_t * op0, vfloat32m2_t * op1, vfloat32m2_t * op2, const float * op3, vuint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_f32m2_m))) +void vluxseg3ei8(vfloat32m2_t * op0, vfloat32m2_t * op1, vfloat32m2_t * op2, vbool16_t op3, vfloat32m2_t op4, vfloat32m2_t op5, vfloat32m2_t op6, const float * op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_f32mf2))) +void vluxseg3ei8(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, const float * op3, vuint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_f32mf2_m))) +void vluxseg3ei8(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vbool64_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, const float * op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_f32m1))) +void vluxseg4ei8(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, const float * op4, vuint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_f32m1_m))) +void vluxseg4ei8(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vbool32_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, const float * op9, vuint8mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_f32m2))) +void vluxseg4ei8(vfloat32m2_t * op0, vfloat32m2_t * op1, vfloat32m2_t * op2, vfloat32m2_t * op3, const float * op4, vuint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_f32m2_m))) +void vluxseg4ei8(vfloat32m2_t * op0, vfloat32m2_t * op1, vfloat32m2_t * op2, vfloat32m2_t * op3, vbool16_t op4, vfloat32m2_t op5, vfloat32m2_t op6, vfloat32m2_t op7, vfloat32m2_t op8, const float * op9, vuint8mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_f32mf2))) +void vluxseg4ei8(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, const float * op4, vuint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_f32mf2_m))) +void vluxseg4ei8(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vbool64_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, const float * op9, vuint8mf8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_f32m1))) +void vluxseg5ei8(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, const float * op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_f32m1_m))) +void vluxseg5ei8(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vbool32_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, vfloat32m1_t op10, const float * op11, vuint8mf4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_f32mf2))) +void vluxseg5ei8(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, const float * op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_f32mf2_m))) +void vluxseg5ei8(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vbool64_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, const float * op11, vuint8mf8_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_f32m1))) +void vluxseg6ei8(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, const float * op6, vuint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_f32m1_m))) +void vluxseg6ei8(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vbool32_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, vfloat32m1_t op10, vfloat32m1_t op11, vfloat32m1_t op12, const float * op13, vuint8mf4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_f32mf2))) +void vluxseg6ei8(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, const float * op6, vuint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_f32mf2_m))) +void vluxseg6ei8(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vbool64_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, vfloat32mf2_t op11, vfloat32mf2_t op12, const float * op13, vuint8mf8_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_f32m1))) +void vluxseg7ei8(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vfloat32m1_t * op6, const float * op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_f32m1_m))) +void vluxseg7ei8(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vfloat32m1_t * op6, vbool32_t op7, vfloat32m1_t op8, vfloat32m1_t op9, vfloat32m1_t op10, vfloat32m1_t op11, vfloat32m1_t op12, vfloat32m1_t op13, vfloat32m1_t op14, const float * op15, vuint8mf4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_f32mf2))) +void vluxseg7ei8(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vfloat32mf2_t * op6, const float * op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_f32mf2_m))) +void vluxseg7ei8(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vfloat32mf2_t * op6, vbool64_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, vfloat32mf2_t op11, vfloat32mf2_t op12, vfloat32mf2_t op13, vfloat32mf2_t op14, const float * op15, vuint8mf8_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_f32m1))) +void vluxseg8ei8(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vfloat32m1_t * op6, vfloat32m1_t * op7, const float * op8, vuint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_f32m1_m))) +void vluxseg8ei8(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vfloat32m1_t * op6, vfloat32m1_t * op7, vbool32_t op8, vfloat32m1_t op9, vfloat32m1_t op10, vfloat32m1_t op11, vfloat32m1_t op12, vfloat32m1_t op13, vfloat32m1_t op14, vfloat32m1_t op15, vfloat32m1_t op16, const float * op17, vuint8mf4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_f32mf2))) +void vluxseg8ei8(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vfloat32mf2_t * op6, vfloat32mf2_t * op7, const float * op8, vuint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_f32mf2_m))) +void vluxseg8ei8(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vfloat32mf2_t * op6, vfloat32mf2_t * op7, vbool64_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, vfloat32mf2_t op11, vfloat32mf2_t op12, vfloat32mf2_t op13, vfloat32mf2_t op14, vfloat32mf2_t op15, vfloat32mf2_t op16, const float * op17, vuint8mf8_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_f32m1))) +void vluxseg2ei16(vfloat32m1_t * op0, vfloat32m1_t * op1, const float * op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_f32m1_m))) +void vluxseg2ei16(vfloat32m1_t * op0, vfloat32m1_t * op1, vbool32_t op2, vfloat32m1_t op3, vfloat32m1_t op4, const float * op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_f32m2))) +void vluxseg2ei16(vfloat32m2_t * op0, vfloat32m2_t * op1, const float * op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_f32m2_m))) +void vluxseg2ei16(vfloat32m2_t * op0, vfloat32m2_t * op1, vbool16_t op2, vfloat32m2_t op3, vfloat32m2_t op4, const float * op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_f32m4))) +void vluxseg2ei16(vfloat32m4_t * op0, vfloat32m4_t * op1, const float * op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_f32m4_m))) +void vluxseg2ei16(vfloat32m4_t * op0, vfloat32m4_t * op1, vbool8_t op2, vfloat32m4_t op3, vfloat32m4_t op4, const float * op5, vuint16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_f32mf2))) +void vluxseg2ei16(vfloat32mf2_t * op0, vfloat32mf2_t * op1, const float * op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_f32mf2_m))) +void vluxseg2ei16(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vbool64_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, const float * op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_f32m1))) +void vluxseg3ei16(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, const float * op3, vuint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_f32m1_m))) +void vluxseg3ei16(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vbool32_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, const float * op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_f32m2))) +void vluxseg3ei16(vfloat32m2_t * op0, vfloat32m2_t * op1, vfloat32m2_t * op2, const float * op3, vuint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_f32m2_m))) +void vluxseg3ei16(vfloat32m2_t * op0, vfloat32m2_t * op1, vfloat32m2_t * op2, vbool16_t op3, vfloat32m2_t op4, vfloat32m2_t op5, vfloat32m2_t op6, const float * op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_f32mf2))) +void vluxseg3ei16(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, const float * op3, vuint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_f32mf2_m))) +void vluxseg3ei16(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vbool64_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, const float * op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_f32m1))) +void vluxseg4ei16(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, const float * op4, vuint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_f32m1_m))) +void vluxseg4ei16(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vbool32_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, const float * op9, vuint16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_f32m2))) +void vluxseg4ei16(vfloat32m2_t * op0, vfloat32m2_t * op1, vfloat32m2_t * op2, vfloat32m2_t * op3, const float * op4, vuint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_f32m2_m))) +void vluxseg4ei16(vfloat32m2_t * op0, vfloat32m2_t * op1, vfloat32m2_t * op2, vfloat32m2_t * op3, vbool16_t op4, vfloat32m2_t op5, vfloat32m2_t op6, vfloat32m2_t op7, vfloat32m2_t op8, const float * op9, vuint16m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_f32mf2))) +void vluxseg4ei16(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, const float * op4, vuint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_f32mf2_m))) +void vluxseg4ei16(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vbool64_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, const float * op9, vuint16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_f32m1))) +void vluxseg5ei16(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, const float * op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_f32m1_m))) +void vluxseg5ei16(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vbool32_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, vfloat32m1_t op10, const float * op11, vuint16mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_f32mf2))) +void vluxseg5ei16(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, const float * op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_f32mf2_m))) +void vluxseg5ei16(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vbool64_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, const float * op11, vuint16mf4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_f32m1))) +void vluxseg6ei16(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, const float * op6, vuint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_f32m1_m))) +void vluxseg6ei16(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vbool32_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, vfloat32m1_t op10, vfloat32m1_t op11, vfloat32m1_t op12, const float * op13, vuint16mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_f32mf2))) +void vluxseg6ei16(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, const float * op6, vuint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_f32mf2_m))) +void vluxseg6ei16(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vbool64_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, vfloat32mf2_t op11, vfloat32mf2_t op12, const float * op13, vuint16mf4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_f32m1))) +void vluxseg7ei16(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vfloat32m1_t * op6, const float * op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_f32m1_m))) +void vluxseg7ei16(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vfloat32m1_t * op6, vbool32_t op7, vfloat32m1_t op8, vfloat32m1_t op9, vfloat32m1_t op10, vfloat32m1_t op11, vfloat32m1_t op12, vfloat32m1_t op13, vfloat32m1_t op14, const float * op15, vuint16mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_f32mf2))) +void vluxseg7ei16(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vfloat32mf2_t * op6, const float * op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_f32mf2_m))) +void vluxseg7ei16(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vfloat32mf2_t * op6, vbool64_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, vfloat32mf2_t op11, vfloat32mf2_t op12, vfloat32mf2_t op13, vfloat32mf2_t op14, const float * op15, vuint16mf4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_f32m1))) +void vluxseg8ei16(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vfloat32m1_t * op6, vfloat32m1_t * op7, const float * op8, vuint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_f32m1_m))) +void vluxseg8ei16(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vfloat32m1_t * op6, vfloat32m1_t * op7, vbool32_t op8, vfloat32m1_t op9, vfloat32m1_t op10, vfloat32m1_t op11, vfloat32m1_t op12, vfloat32m1_t op13, vfloat32m1_t op14, vfloat32m1_t op15, vfloat32m1_t op16, const float * op17, vuint16mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_f32mf2))) +void vluxseg8ei16(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vfloat32mf2_t * op6, vfloat32mf2_t * op7, const float * op8, vuint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_f32mf2_m))) +void vluxseg8ei16(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vfloat32mf2_t * op6, vfloat32mf2_t * op7, vbool64_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, vfloat32mf2_t op11, vfloat32mf2_t op12, vfloat32mf2_t op13, vfloat32mf2_t op14, vfloat32mf2_t op15, vfloat32mf2_t op16, const float * op17, vuint16mf4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_f32m1))) +void vluxseg2ei32(vfloat32m1_t * op0, vfloat32m1_t * op1, const float * op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_f32m1_m))) +void vluxseg2ei32(vfloat32m1_t * op0, vfloat32m1_t * op1, vbool32_t op2, vfloat32m1_t op3, vfloat32m1_t op4, const float * op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_f32m2))) +void vluxseg2ei32(vfloat32m2_t * op0, vfloat32m2_t * op1, const float * op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_f32m2_m))) +void vluxseg2ei32(vfloat32m2_t * op0, vfloat32m2_t * op1, vbool16_t op2, vfloat32m2_t op3, vfloat32m2_t op4, const float * op5, vuint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_f32m4))) +void vluxseg2ei32(vfloat32m4_t * op0, vfloat32m4_t * op1, const float * op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_f32m4_m))) +void vluxseg2ei32(vfloat32m4_t * op0, vfloat32m4_t * op1, vbool8_t op2, vfloat32m4_t op3, vfloat32m4_t op4, const float * op5, vuint32m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_f32mf2))) +void vluxseg2ei32(vfloat32mf2_t * op0, vfloat32mf2_t * op1, const float * op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_f32mf2_m))) +void vluxseg2ei32(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vbool64_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, const float * op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_f32m1))) +void vluxseg3ei32(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, const float * op3, vuint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_f32m1_m))) +void vluxseg3ei32(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vbool32_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, const float * op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_f32m2))) +void vluxseg3ei32(vfloat32m2_t * op0, vfloat32m2_t * op1, vfloat32m2_t * op2, const float * op3, vuint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_f32m2_m))) +void vluxseg3ei32(vfloat32m2_t * op0, vfloat32m2_t * op1, vfloat32m2_t * op2, vbool16_t op3, vfloat32m2_t op4, vfloat32m2_t op5, vfloat32m2_t op6, const float * op7, vuint32m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_f32mf2))) +void vluxseg3ei32(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, const float * op3, vuint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_f32mf2_m))) +void vluxseg3ei32(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vbool64_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, const float * op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_f32m1))) +void vluxseg4ei32(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, const float * op4, vuint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_f32m1_m))) +void vluxseg4ei32(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vbool32_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, const float * op9, vuint32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_f32m2))) +void vluxseg4ei32(vfloat32m2_t * op0, vfloat32m2_t * op1, vfloat32m2_t * op2, vfloat32m2_t * op3, const float * op4, vuint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_f32m2_m))) +void vluxseg4ei32(vfloat32m2_t * op0, vfloat32m2_t * op1, vfloat32m2_t * op2, vfloat32m2_t * op3, vbool16_t op4, vfloat32m2_t op5, vfloat32m2_t op6, vfloat32m2_t op7, vfloat32m2_t op8, const float * op9, vuint32m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_f32mf2))) +void vluxseg4ei32(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, const float * op4, vuint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_f32mf2_m))) +void vluxseg4ei32(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vbool64_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, const float * op9, vuint32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_f32m1))) +void vluxseg5ei32(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, const float * op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_f32m1_m))) +void vluxseg5ei32(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vbool32_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, vfloat32m1_t op10, const float * op11, vuint32m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_f32mf2))) +void vluxseg5ei32(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, const float * op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_f32mf2_m))) +void vluxseg5ei32(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vbool64_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, const float * op11, vuint32mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_f32m1))) +void vluxseg6ei32(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, const float * op6, vuint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_f32m1_m))) +void vluxseg6ei32(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vbool32_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, vfloat32m1_t op10, vfloat32m1_t op11, vfloat32m1_t op12, const float * op13, vuint32m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_f32mf2))) +void vluxseg6ei32(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, const float * op6, vuint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_f32mf2_m))) +void vluxseg6ei32(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vbool64_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, vfloat32mf2_t op11, vfloat32mf2_t op12, const float * op13, vuint32mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_f32m1))) +void vluxseg7ei32(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vfloat32m1_t * op6, const float * op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_f32m1_m))) +void vluxseg7ei32(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vfloat32m1_t * op6, vbool32_t op7, vfloat32m1_t op8, vfloat32m1_t op9, vfloat32m1_t op10, vfloat32m1_t op11, vfloat32m1_t op12, vfloat32m1_t op13, vfloat32m1_t op14, const float * op15, vuint32m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_f32mf2))) +void vluxseg7ei32(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vfloat32mf2_t * op6, const float * op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_f32mf2_m))) +void vluxseg7ei32(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vfloat32mf2_t * op6, vbool64_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, vfloat32mf2_t op11, vfloat32mf2_t op12, vfloat32mf2_t op13, vfloat32mf2_t op14, const float * op15, vuint32mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_f32m1))) +void vluxseg8ei32(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vfloat32m1_t * op6, vfloat32m1_t * op7, const float * op8, vuint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_f32m1_m))) +void vluxseg8ei32(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vfloat32m1_t * op6, vfloat32m1_t * op7, vbool32_t op8, vfloat32m1_t op9, vfloat32m1_t op10, vfloat32m1_t op11, vfloat32m1_t op12, vfloat32m1_t op13, vfloat32m1_t op14, vfloat32m1_t op15, vfloat32m1_t op16, const float * op17, vuint32m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_f32mf2))) +void vluxseg8ei32(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vfloat32mf2_t * op6, vfloat32mf2_t * op7, const float * op8, vuint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_f32mf2_m))) +void vluxseg8ei32(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vfloat32mf2_t * op6, vfloat32mf2_t * op7, vbool64_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, vfloat32mf2_t op11, vfloat32mf2_t op12, vfloat32mf2_t op13, vfloat32mf2_t op14, vfloat32mf2_t op15, vfloat32mf2_t op16, const float * op17, vuint32mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_f32m1))) +void vluxseg2ei64(vfloat32m1_t * op0, vfloat32m1_t * op1, const float * op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_f32m1_m))) +void vluxseg2ei64(vfloat32m1_t * op0, vfloat32m1_t * op1, vbool32_t op2, vfloat32m1_t op3, vfloat32m1_t op4, const float * op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_f32m2))) +void vluxseg2ei64(vfloat32m2_t * op0, vfloat32m2_t * op1, const float * op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_f32m2_m))) +void vluxseg2ei64(vfloat32m2_t * op0, vfloat32m2_t * op1, vbool16_t op2, vfloat32m2_t op3, vfloat32m2_t op4, const float * op5, vuint64m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_f32m4))) +void vluxseg2ei64(vfloat32m4_t * op0, vfloat32m4_t * op1, const float * op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_f32m4_m))) +void vluxseg2ei64(vfloat32m4_t * op0, vfloat32m4_t * op1, vbool8_t op2, vfloat32m4_t op3, vfloat32m4_t op4, const float * op5, vuint64m8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_f32mf2))) +void vluxseg2ei64(vfloat32mf2_t * op0, vfloat32mf2_t * op1, const float * op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_f32mf2_m))) +void vluxseg2ei64(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vbool64_t op2, vfloat32mf2_t op3, vfloat32mf2_t op4, const float * op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_f32m1))) +void vluxseg3ei64(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, const float * op3, vuint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_f32m1_m))) +void vluxseg3ei64(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vbool32_t op3, vfloat32m1_t op4, vfloat32m1_t op5, vfloat32m1_t op6, const float * op7, vuint64m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_f32m2))) +void vluxseg3ei64(vfloat32m2_t * op0, vfloat32m2_t * op1, vfloat32m2_t * op2, const float * op3, vuint64m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_f32m2_m))) +void vluxseg3ei64(vfloat32m2_t * op0, vfloat32m2_t * op1, vfloat32m2_t * op2, vbool16_t op3, vfloat32m2_t op4, vfloat32m2_t op5, vfloat32m2_t op6, const float * op7, vuint64m4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_f32mf2))) +void vluxseg3ei64(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, const float * op3, vuint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_f32mf2_m))) +void vluxseg3ei64(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vbool64_t op3, vfloat32mf2_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, const float * op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_f32m1))) +void vluxseg4ei64(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, const float * op4, vuint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_f32m1_m))) +void vluxseg4ei64(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vbool32_t op4, vfloat32m1_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, const float * op9, vuint64m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_f32m2))) +void vluxseg4ei64(vfloat32m2_t * op0, vfloat32m2_t * op1, vfloat32m2_t * op2, vfloat32m2_t * op3, const float * op4, vuint64m4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_f32m2_m))) +void vluxseg4ei64(vfloat32m2_t * op0, vfloat32m2_t * op1, vfloat32m2_t * op2, vfloat32m2_t * op3, vbool16_t op4, vfloat32m2_t op5, vfloat32m2_t op6, vfloat32m2_t op7, vfloat32m2_t op8, const float * op9, vuint64m4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_f32mf2))) +void vluxseg4ei64(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, const float * op4, vuint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_f32mf2_m))) +void vluxseg4ei64(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vbool64_t op4, vfloat32mf2_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, const float * op9, vuint64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_f32m1))) +void vluxseg5ei64(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, const float * op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_f32m1_m))) +void vluxseg5ei64(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vbool32_t op5, vfloat32m1_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, vfloat32m1_t op10, const float * op11, vuint64m2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_f32mf2))) +void vluxseg5ei64(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, const float * op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_f32mf2_m))) +void vluxseg5ei64(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vbool64_t op5, vfloat32mf2_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, const float * op11, vuint64m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_f32m1))) +void vluxseg6ei64(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, const float * op6, vuint64m2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_f32m1_m))) +void vluxseg6ei64(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vbool32_t op6, vfloat32m1_t op7, vfloat32m1_t op8, vfloat32m1_t op9, vfloat32m1_t op10, vfloat32m1_t op11, vfloat32m1_t op12, const float * op13, vuint64m2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_f32mf2))) +void vluxseg6ei64(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, const float * op6, vuint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_f32mf2_m))) +void vluxseg6ei64(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vbool64_t op6, vfloat32mf2_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, vfloat32mf2_t op11, vfloat32mf2_t op12, const float * op13, vuint64m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_f32m1))) +void vluxseg7ei64(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vfloat32m1_t * op6, const float * op7, vuint64m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_f32m1_m))) +void vluxseg7ei64(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vfloat32m1_t * op6, vbool32_t op7, vfloat32m1_t op8, vfloat32m1_t op9, vfloat32m1_t op10, vfloat32m1_t op11, vfloat32m1_t op12, vfloat32m1_t op13, vfloat32m1_t op14, const float * op15, vuint64m2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_f32mf2))) +void vluxseg7ei64(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vfloat32mf2_t * op6, const float * op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_f32mf2_m))) +void vluxseg7ei64(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vfloat32mf2_t * op6, vbool64_t op7, vfloat32mf2_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, vfloat32mf2_t op11, vfloat32mf2_t op12, vfloat32mf2_t op13, vfloat32mf2_t op14, const float * op15, vuint64m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_f32m1))) +void vluxseg8ei64(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vfloat32m1_t * op6, vfloat32m1_t * op7, const float * op8, vuint64m2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_f32m1_m))) +void vluxseg8ei64(vfloat32m1_t * op0, vfloat32m1_t * op1, vfloat32m1_t * op2, vfloat32m1_t * op3, vfloat32m1_t * op4, vfloat32m1_t * op5, vfloat32m1_t * op6, vfloat32m1_t * op7, vbool32_t op8, vfloat32m1_t op9, vfloat32m1_t op10, vfloat32m1_t op11, vfloat32m1_t op12, vfloat32m1_t op13, vfloat32m1_t op14, vfloat32m1_t op15, vfloat32m1_t op16, const float * op17, vuint64m2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_f32mf2))) +void vluxseg8ei64(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vfloat32mf2_t * op6, vfloat32mf2_t * op7, const float * op8, vuint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_f32mf2_m))) +void vluxseg8ei64(vfloat32mf2_t * op0, vfloat32mf2_t * op1, vfloat32mf2_t * op2, vfloat32mf2_t * op3, vfloat32mf2_t * op4, vfloat32mf2_t * op5, vfloat32mf2_t * op6, vfloat32mf2_t * op7, vbool64_t op8, vfloat32mf2_t op9, vfloat32mf2_t op10, vfloat32mf2_t op11, vfloat32mf2_t op12, vfloat32mf2_t op13, vfloat32mf2_t op14, vfloat32mf2_t op15, vfloat32mf2_t op16, const float * op17, vuint64m1_t op18, size_t op19); + +#endif + +#if defined(__riscv_d) && defined(__riscv_zvlsseg) +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_f64m1))) +void vloxseg2ei8(vfloat64m1_t * op0, vfloat64m1_t * op1, const double * op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_f64m1_m))) +void vloxseg2ei8(vfloat64m1_t * op0, vfloat64m1_t * op1, vbool64_t op2, vfloat64m1_t op3, vfloat64m1_t op4, const double * op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_f64m2))) +void vloxseg2ei8(vfloat64m2_t * op0, vfloat64m2_t * op1, const double * op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_f64m2_m))) +void vloxseg2ei8(vfloat64m2_t * op0, vfloat64m2_t * op1, vbool32_t op2, vfloat64m2_t op3, vfloat64m2_t op4, const double * op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_f64m4))) +void vloxseg2ei8(vfloat64m4_t * op0, vfloat64m4_t * op1, const double * op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_f64m4_m))) +void vloxseg2ei8(vfloat64m4_t * op0, vfloat64m4_t * op1, vbool16_t op2, vfloat64m4_t op3, vfloat64m4_t op4, const double * op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_f64m1))) +void vloxseg3ei8(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, const double * op3, vuint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_f64m1_m))) +void vloxseg3ei8(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vbool64_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, const double * op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_f64m2))) +void vloxseg3ei8(vfloat64m2_t * op0, vfloat64m2_t * op1, vfloat64m2_t * op2, const double * op3, vuint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_f64m2_m))) +void vloxseg3ei8(vfloat64m2_t * op0, vfloat64m2_t * op1, vfloat64m2_t * op2, vbool32_t op3, vfloat64m2_t op4, vfloat64m2_t op5, vfloat64m2_t op6, const double * op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_f64m1))) +void vloxseg4ei8(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, const double * op4, vuint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_f64m1_m))) +void vloxseg4ei8(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vbool64_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, const double * op9, vuint8mf8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_f64m2))) +void vloxseg4ei8(vfloat64m2_t * op0, vfloat64m2_t * op1, vfloat64m2_t * op2, vfloat64m2_t * op3, const double * op4, vuint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_f64m2_m))) +void vloxseg4ei8(vfloat64m2_t * op0, vfloat64m2_t * op1, vfloat64m2_t * op2, vfloat64m2_t * op3, vbool32_t op4, vfloat64m2_t op5, vfloat64m2_t op6, vfloat64m2_t op7, vfloat64m2_t op8, const double * op9, vuint8mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_f64m1))) +void vloxseg5ei8(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, const double * op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_f64m1_m))) +void vloxseg5ei8(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vbool64_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, vfloat64m1_t op10, const double * op11, vuint8mf8_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_f64m1))) +void vloxseg6ei8(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, const double * op6, vuint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_f64m1_m))) +void vloxseg6ei8(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vbool64_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, vfloat64m1_t op10, vfloat64m1_t op11, vfloat64m1_t op12, const double * op13, vuint8mf8_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_f64m1))) +void vloxseg7ei8(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vfloat64m1_t * op6, const double * op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_f64m1_m))) +void vloxseg7ei8(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vfloat64m1_t * op6, vbool64_t op7, vfloat64m1_t op8, vfloat64m1_t op9, vfloat64m1_t op10, vfloat64m1_t op11, vfloat64m1_t op12, vfloat64m1_t op13, vfloat64m1_t op14, const double * op15, vuint8mf8_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_f64m1))) +void vloxseg8ei8(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vfloat64m1_t * op6, vfloat64m1_t * op7, const double * op8, vuint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_f64m1_m))) +void vloxseg8ei8(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vfloat64m1_t * op6, vfloat64m1_t * op7, vbool64_t op8, vfloat64m1_t op9, vfloat64m1_t op10, vfloat64m1_t op11, vfloat64m1_t op12, vfloat64m1_t op13, vfloat64m1_t op14, vfloat64m1_t op15, vfloat64m1_t op16, const double * op17, vuint8mf8_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_f64m1))) +void vloxseg2ei16(vfloat64m1_t * op0, vfloat64m1_t * op1, const double * op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_f64m1_m))) +void vloxseg2ei16(vfloat64m1_t * op0, vfloat64m1_t * op1, vbool64_t op2, vfloat64m1_t op3, vfloat64m1_t op4, const double * op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_f64m2))) +void vloxseg2ei16(vfloat64m2_t * op0, vfloat64m2_t * op1, const double * op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_f64m2_m))) +void vloxseg2ei16(vfloat64m2_t * op0, vfloat64m2_t * op1, vbool32_t op2, vfloat64m2_t op3, vfloat64m2_t op4, const double * op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_f64m4))) +void vloxseg2ei16(vfloat64m4_t * op0, vfloat64m4_t * op1, const double * op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_f64m4_m))) +void vloxseg2ei16(vfloat64m4_t * op0, vfloat64m4_t * op1, vbool16_t op2, vfloat64m4_t op3, vfloat64m4_t op4, const double * op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_f64m1))) +void vloxseg3ei16(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, const double * op3, vuint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_f64m1_m))) +void vloxseg3ei16(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vbool64_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, const double * op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_f64m2))) +void vloxseg3ei16(vfloat64m2_t * op0, vfloat64m2_t * op1, vfloat64m2_t * op2, const double * op3, vuint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_f64m2_m))) +void vloxseg3ei16(vfloat64m2_t * op0, vfloat64m2_t * op1, vfloat64m2_t * op2, vbool32_t op3, vfloat64m2_t op4, vfloat64m2_t op5, vfloat64m2_t op6, const double * op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_f64m1))) +void vloxseg4ei16(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, const double * op4, vuint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_f64m1_m))) +void vloxseg4ei16(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vbool64_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, const double * op9, vuint16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_f64m2))) +void vloxseg4ei16(vfloat64m2_t * op0, vfloat64m2_t * op1, vfloat64m2_t * op2, vfloat64m2_t * op3, const double * op4, vuint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_f64m2_m))) +void vloxseg4ei16(vfloat64m2_t * op0, vfloat64m2_t * op1, vfloat64m2_t * op2, vfloat64m2_t * op3, vbool32_t op4, vfloat64m2_t op5, vfloat64m2_t op6, vfloat64m2_t op7, vfloat64m2_t op8, const double * op9, vuint16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_f64m1))) +void vloxseg5ei16(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, const double * op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_f64m1_m))) +void vloxseg5ei16(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vbool64_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, vfloat64m1_t op10, const double * op11, vuint16mf4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_f64m1))) +void vloxseg6ei16(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, const double * op6, vuint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_f64m1_m))) +void vloxseg6ei16(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vbool64_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, vfloat64m1_t op10, vfloat64m1_t op11, vfloat64m1_t op12, const double * op13, vuint16mf4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_f64m1))) +void vloxseg7ei16(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vfloat64m1_t * op6, const double * op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_f64m1_m))) +void vloxseg7ei16(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vfloat64m1_t * op6, vbool64_t op7, vfloat64m1_t op8, vfloat64m1_t op9, vfloat64m1_t op10, vfloat64m1_t op11, vfloat64m1_t op12, vfloat64m1_t op13, vfloat64m1_t op14, const double * op15, vuint16mf4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_f64m1))) +void vloxseg8ei16(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vfloat64m1_t * op6, vfloat64m1_t * op7, const double * op8, vuint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_f64m1_m))) +void vloxseg8ei16(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vfloat64m1_t * op6, vfloat64m1_t * op7, vbool64_t op8, vfloat64m1_t op9, vfloat64m1_t op10, vfloat64m1_t op11, vfloat64m1_t op12, vfloat64m1_t op13, vfloat64m1_t op14, vfloat64m1_t op15, vfloat64m1_t op16, const double * op17, vuint16mf4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_f64m1))) +void vloxseg2ei32(vfloat64m1_t * op0, vfloat64m1_t * op1, const double * op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_f64m1_m))) +void vloxseg2ei32(vfloat64m1_t * op0, vfloat64m1_t * op1, vbool64_t op2, vfloat64m1_t op3, vfloat64m1_t op4, const double * op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_f64m2))) +void vloxseg2ei32(vfloat64m2_t * op0, vfloat64m2_t * op1, const double * op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_f64m2_m))) +void vloxseg2ei32(vfloat64m2_t * op0, vfloat64m2_t * op1, vbool32_t op2, vfloat64m2_t op3, vfloat64m2_t op4, const double * op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_f64m4))) +void vloxseg2ei32(vfloat64m4_t * op0, vfloat64m4_t * op1, const double * op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_f64m4_m))) +void vloxseg2ei32(vfloat64m4_t * op0, vfloat64m4_t * op1, vbool16_t op2, vfloat64m4_t op3, vfloat64m4_t op4, const double * op5, vuint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_f64m1))) +void vloxseg3ei32(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, const double * op3, vuint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_f64m1_m))) +void vloxseg3ei32(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vbool64_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, const double * op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_f64m2))) +void vloxseg3ei32(vfloat64m2_t * op0, vfloat64m2_t * op1, vfloat64m2_t * op2, const double * op3, vuint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_f64m2_m))) +void vloxseg3ei32(vfloat64m2_t * op0, vfloat64m2_t * op1, vfloat64m2_t * op2, vbool32_t op3, vfloat64m2_t op4, vfloat64m2_t op5, vfloat64m2_t op6, const double * op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_f64m1))) +void vloxseg4ei32(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, const double * op4, vuint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_f64m1_m))) +void vloxseg4ei32(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vbool64_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, const double * op9, vuint32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_f64m2))) +void vloxseg4ei32(vfloat64m2_t * op0, vfloat64m2_t * op1, vfloat64m2_t * op2, vfloat64m2_t * op3, const double * op4, vuint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_f64m2_m))) +void vloxseg4ei32(vfloat64m2_t * op0, vfloat64m2_t * op1, vfloat64m2_t * op2, vfloat64m2_t * op3, vbool32_t op4, vfloat64m2_t op5, vfloat64m2_t op6, vfloat64m2_t op7, vfloat64m2_t op8, const double * op9, vuint32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_f64m1))) +void vloxseg5ei32(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, const double * op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_f64m1_m))) +void vloxseg5ei32(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vbool64_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, vfloat64m1_t op10, const double * op11, vuint32mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_f64m1))) +void vloxseg6ei32(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, const double * op6, vuint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_f64m1_m))) +void vloxseg6ei32(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vbool64_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, vfloat64m1_t op10, vfloat64m1_t op11, vfloat64m1_t op12, const double * op13, vuint32mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_f64m1))) +void vloxseg7ei32(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vfloat64m1_t * op6, const double * op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_f64m1_m))) +void vloxseg7ei32(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vfloat64m1_t * op6, vbool64_t op7, vfloat64m1_t op8, vfloat64m1_t op9, vfloat64m1_t op10, vfloat64m1_t op11, vfloat64m1_t op12, vfloat64m1_t op13, vfloat64m1_t op14, const double * op15, vuint32mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_f64m1))) +void vloxseg8ei32(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vfloat64m1_t * op6, vfloat64m1_t * op7, const double * op8, vuint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_f64m1_m))) +void vloxseg8ei32(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vfloat64m1_t * op6, vfloat64m1_t * op7, vbool64_t op8, vfloat64m1_t op9, vfloat64m1_t op10, vfloat64m1_t op11, vfloat64m1_t op12, vfloat64m1_t op13, vfloat64m1_t op14, vfloat64m1_t op15, vfloat64m1_t op16, const double * op17, vuint32mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_f64m1))) +void vloxseg2ei64(vfloat64m1_t * op0, vfloat64m1_t * op1, const double * op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_f64m1_m))) +void vloxseg2ei64(vfloat64m1_t * op0, vfloat64m1_t * op1, vbool64_t op2, vfloat64m1_t op3, vfloat64m1_t op4, const double * op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_f64m2))) +void vloxseg2ei64(vfloat64m2_t * op0, vfloat64m2_t * op1, const double * op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_f64m2_m))) +void vloxseg2ei64(vfloat64m2_t * op0, vfloat64m2_t * op1, vbool32_t op2, vfloat64m2_t op3, vfloat64m2_t op4, const double * op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_f64m4))) +void vloxseg2ei64(vfloat64m4_t * op0, vfloat64m4_t * op1, const double * op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_f64m4_m))) +void vloxseg2ei64(vfloat64m4_t * op0, vfloat64m4_t * op1, vbool16_t op2, vfloat64m4_t op3, vfloat64m4_t op4, const double * op5, vuint64m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_f64m1))) +void vloxseg3ei64(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, const double * op3, vuint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_f64m1_m))) +void vloxseg3ei64(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vbool64_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, const double * op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_f64m2))) +void vloxseg3ei64(vfloat64m2_t * op0, vfloat64m2_t * op1, vfloat64m2_t * op2, const double * op3, vuint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_f64m2_m))) +void vloxseg3ei64(vfloat64m2_t * op0, vfloat64m2_t * op1, vfloat64m2_t * op2, vbool32_t op3, vfloat64m2_t op4, vfloat64m2_t op5, vfloat64m2_t op6, const double * op7, vuint64m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_f64m1))) +void vloxseg4ei64(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, const double * op4, vuint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_f64m1_m))) +void vloxseg4ei64(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vbool64_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, const double * op9, vuint64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_f64m2))) +void vloxseg4ei64(vfloat64m2_t * op0, vfloat64m2_t * op1, vfloat64m2_t * op2, vfloat64m2_t * op3, const double * op4, vuint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_f64m2_m))) +void vloxseg4ei64(vfloat64m2_t * op0, vfloat64m2_t * op1, vfloat64m2_t * op2, vfloat64m2_t * op3, vbool32_t op4, vfloat64m2_t op5, vfloat64m2_t op6, vfloat64m2_t op7, vfloat64m2_t op8, const double * op9, vuint64m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_f64m1))) +void vloxseg5ei64(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, const double * op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_f64m1_m))) +void vloxseg5ei64(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vbool64_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, vfloat64m1_t op10, const double * op11, vuint64m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_f64m1))) +void vloxseg6ei64(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, const double * op6, vuint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_f64m1_m))) +void vloxseg6ei64(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vbool64_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, vfloat64m1_t op10, vfloat64m1_t op11, vfloat64m1_t op12, const double * op13, vuint64m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_f64m1))) +void vloxseg7ei64(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vfloat64m1_t * op6, const double * op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_f64m1_m))) +void vloxseg7ei64(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vfloat64m1_t * op6, vbool64_t op7, vfloat64m1_t op8, vfloat64m1_t op9, vfloat64m1_t op10, vfloat64m1_t op11, vfloat64m1_t op12, vfloat64m1_t op13, vfloat64m1_t op14, const double * op15, vuint64m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_f64m1))) +void vloxseg8ei64(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vfloat64m1_t * op6, vfloat64m1_t * op7, const double * op8, vuint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_f64m1_m))) +void vloxseg8ei64(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vfloat64m1_t * op6, vfloat64m1_t * op7, vbool64_t op8, vfloat64m1_t op9, vfloat64m1_t op10, vfloat64m1_t op11, vfloat64m1_t op12, vfloat64m1_t op13, vfloat64m1_t op14, vfloat64m1_t op15, vfloat64m1_t op16, const double * op17, vuint64m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e64_v_f64m1))) +void vsseg2e64(double * op0, vfloat64m1_t op1, vfloat64m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e64_v_f64m1_m))) +void vsseg2e64(vbool64_t op0, double * op1, vfloat64m1_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e64_v_f64m2))) +void vsseg2e64(double * op0, vfloat64m2_t op1, vfloat64m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e64_v_f64m2_m))) +void vsseg2e64(vbool32_t op0, double * op1, vfloat64m2_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e64_v_f64m4))) +void vsseg2e64(double * op0, vfloat64m4_t op1, vfloat64m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e64_v_f64m4_m))) +void vsseg2e64(vbool16_t op0, double * op1, vfloat64m4_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e64_v_f64m1))) +void vsseg3e64(double * op0, vfloat64m1_t op1, vfloat64m1_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e64_v_f64m1_m))) +void vsseg3e64(vbool64_t op0, double * op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e64_v_f64m2))) +void vsseg3e64(double * op0, vfloat64m2_t op1, vfloat64m2_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e64_v_f64m2_m))) +void vsseg3e64(vbool32_t op0, double * op1, vfloat64m2_t op2, vfloat64m2_t op3, vfloat64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e64_v_f64m1))) +void vsseg4e64(double * op0, vfloat64m1_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e64_v_f64m1_m))) +void vsseg4e64(vbool64_t op0, double * op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e64_v_f64m2))) +void vsseg4e64(double * op0, vfloat64m2_t op1, vfloat64m2_t op2, vfloat64m2_t op3, vfloat64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e64_v_f64m2_m))) +void vsseg4e64(vbool32_t op0, double * op1, vfloat64m2_t op2, vfloat64m2_t op3, vfloat64m2_t op4, vfloat64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e64_v_f64m1))) +void vsseg5e64(double * op0, vfloat64m1_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e64_v_f64m1_m))) +void vsseg5e64(vbool64_t op0, double * op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e64_v_f64m1))) +void vsseg6e64(double * op0, vfloat64m1_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e64_v_f64m1_m))) +void vsseg6e64(vbool64_t op0, double * op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e64_v_f64m1))) +void vsseg7e64(double * op0, vfloat64m1_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e64_v_f64m1_m))) +void vsseg7e64(vbool64_t op0, double * op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e64_v_f64m1))) +void vsseg8e64(double * op0, vfloat64m1_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e64_v_f64m1_m))) +void vsseg8e64(vbool64_t op0, double * op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e64_v_f64m1))) +void vssseg2e64(double * op0, ptrdiff_t op1, vfloat64m1_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e64_v_f64m1_m))) +void vssseg2e64(vbool64_t op0, double * op1, ptrdiff_t op2, vfloat64m1_t op3, vfloat64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e64_v_f64m2))) +void vssseg2e64(double * op0, ptrdiff_t op1, vfloat64m2_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e64_v_f64m2_m))) +void vssseg2e64(vbool32_t op0, double * op1, ptrdiff_t op2, vfloat64m2_t op3, vfloat64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e64_v_f64m4))) +void vssseg2e64(double * op0, ptrdiff_t op1, vfloat64m4_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e64_v_f64m4_m))) +void vssseg2e64(vbool16_t op0, double * op1, ptrdiff_t op2, vfloat64m4_t op3, vfloat64m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e64_v_f64m1))) +void vssseg3e64(double * op0, ptrdiff_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e64_v_f64m1_m))) +void vssseg3e64(vbool64_t op0, double * op1, ptrdiff_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e64_v_f64m2))) +void vssseg3e64(double * op0, ptrdiff_t op1, vfloat64m2_t op2, vfloat64m2_t op3, vfloat64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e64_v_f64m2_m))) +void vssseg3e64(vbool32_t op0, double * op1, ptrdiff_t op2, vfloat64m2_t op3, vfloat64m2_t op4, vfloat64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e64_v_f64m1))) +void vssseg4e64(double * op0, ptrdiff_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e64_v_f64m1_m))) +void vssseg4e64(vbool64_t op0, double * op1, ptrdiff_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e64_v_f64m2))) +void vssseg4e64(double * op0, ptrdiff_t op1, vfloat64m2_t op2, vfloat64m2_t op3, vfloat64m2_t op4, vfloat64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e64_v_f64m2_m))) +void vssseg4e64(vbool32_t op0, double * op1, ptrdiff_t op2, vfloat64m2_t op3, vfloat64m2_t op4, vfloat64m2_t op5, vfloat64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e64_v_f64m1))) +void vssseg5e64(double * op0, ptrdiff_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e64_v_f64m1_m))) +void vssseg5e64(vbool64_t op0, double * op1, ptrdiff_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e64_v_f64m1))) +void vssseg6e64(double * op0, ptrdiff_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e64_v_f64m1_m))) +void vssseg6e64(vbool64_t op0, double * op1, ptrdiff_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e64_v_f64m1))) +void vssseg7e64(double * op0, ptrdiff_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e64_v_f64m1_m))) +void vssseg7e64(vbool64_t op0, double * op1, ptrdiff_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e64_v_f64m1))) +void vssseg8e64(double * op0, ptrdiff_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e64_v_f64m1_m))) +void vssseg8e64(vbool64_t op0, double * op1, ptrdiff_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, vfloat64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_f64m1))) +void vsuxseg2ei8(double * op0, vuint8mf8_t op1, vfloat64m1_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_f64m1_m))) +void vsuxseg2ei8(vbool64_t op0, double * op1, vuint8mf8_t op2, vfloat64m1_t op3, vfloat64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_f64m2))) +void vsuxseg2ei8(double * op0, vuint8mf4_t op1, vfloat64m2_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_f64m2_m))) +void vsuxseg2ei8(vbool32_t op0, double * op1, vuint8mf4_t op2, vfloat64m2_t op3, vfloat64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_f64m4))) +void vsuxseg2ei8(double * op0, vuint8mf2_t op1, vfloat64m4_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_f64m4_m))) +void vsuxseg2ei8(vbool16_t op0, double * op1, vuint8mf2_t op2, vfloat64m4_t op3, vfloat64m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_f64m1))) +void vsuxseg3ei8(double * op0, vuint8mf8_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_f64m1_m))) +void vsuxseg3ei8(vbool64_t op0, double * op1, vuint8mf8_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_f64m2))) +void vsuxseg3ei8(double * op0, vuint8mf4_t op1, vfloat64m2_t op2, vfloat64m2_t op3, vfloat64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_f64m2_m))) +void vsuxseg3ei8(vbool32_t op0, double * op1, vuint8mf4_t op2, vfloat64m2_t op3, vfloat64m2_t op4, vfloat64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_f64m1))) +void vsuxseg4ei8(double * op0, vuint8mf8_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_f64m1_m))) +void vsuxseg4ei8(vbool64_t op0, double * op1, vuint8mf8_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_f64m2))) +void vsuxseg4ei8(double * op0, vuint8mf4_t op1, vfloat64m2_t op2, vfloat64m2_t op3, vfloat64m2_t op4, vfloat64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_f64m2_m))) +void vsuxseg4ei8(vbool32_t op0, double * op1, vuint8mf4_t op2, vfloat64m2_t op3, vfloat64m2_t op4, vfloat64m2_t op5, vfloat64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_f64m1))) +void vsuxseg5ei8(double * op0, vuint8mf8_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_f64m1_m))) +void vsuxseg5ei8(vbool64_t op0, double * op1, vuint8mf8_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_f64m1))) +void vsuxseg6ei8(double * op0, vuint8mf8_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_f64m1_m))) +void vsuxseg6ei8(vbool64_t op0, double * op1, vuint8mf8_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_f64m1))) +void vsuxseg7ei8(double * op0, vuint8mf8_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_f64m1_m))) +void vsuxseg7ei8(vbool64_t op0, double * op1, vuint8mf8_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_f64m1))) +void vsuxseg8ei8(double * op0, vuint8mf8_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_f64m1_m))) +void vsuxseg8ei8(vbool64_t op0, double * op1, vuint8mf8_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, vfloat64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_f64m1))) +void vsuxseg2ei16(double * op0, vuint16mf4_t op1, vfloat64m1_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_f64m1_m))) +void vsuxseg2ei16(vbool64_t op0, double * op1, vuint16mf4_t op2, vfloat64m1_t op3, vfloat64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_f64m2))) +void vsuxseg2ei16(double * op0, vuint16mf2_t op1, vfloat64m2_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_f64m2_m))) +void vsuxseg2ei16(vbool32_t op0, double * op1, vuint16mf2_t op2, vfloat64m2_t op3, vfloat64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_f64m4))) +void vsuxseg2ei16(double * op0, vuint16m1_t op1, vfloat64m4_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_f64m4_m))) +void vsuxseg2ei16(vbool16_t op0, double * op1, vuint16m1_t op2, vfloat64m4_t op3, vfloat64m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_f64m1))) +void vsuxseg3ei16(double * op0, vuint16mf4_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_f64m1_m))) +void vsuxseg3ei16(vbool64_t op0, double * op1, vuint16mf4_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_f64m2))) +void vsuxseg3ei16(double * op0, vuint16mf2_t op1, vfloat64m2_t op2, vfloat64m2_t op3, vfloat64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_f64m2_m))) +void vsuxseg3ei16(vbool32_t op0, double * op1, vuint16mf2_t op2, vfloat64m2_t op3, vfloat64m2_t op4, vfloat64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_f64m1))) +void vsuxseg4ei16(double * op0, vuint16mf4_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_f64m1_m))) +void vsuxseg4ei16(vbool64_t op0, double * op1, vuint16mf4_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_f64m2))) +void vsuxseg4ei16(double * op0, vuint16mf2_t op1, vfloat64m2_t op2, vfloat64m2_t op3, vfloat64m2_t op4, vfloat64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_f64m2_m))) +void vsuxseg4ei16(vbool32_t op0, double * op1, vuint16mf2_t op2, vfloat64m2_t op3, vfloat64m2_t op4, vfloat64m2_t op5, vfloat64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_f64m1))) +void vsuxseg5ei16(double * op0, vuint16mf4_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_f64m1_m))) +void vsuxseg5ei16(vbool64_t op0, double * op1, vuint16mf4_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_f64m1))) +void vsuxseg6ei16(double * op0, vuint16mf4_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_f64m1_m))) +void vsuxseg6ei16(vbool64_t op0, double * op1, vuint16mf4_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_f64m1))) +void vsuxseg7ei16(double * op0, vuint16mf4_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_f64m1_m))) +void vsuxseg7ei16(vbool64_t op0, double * op1, vuint16mf4_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_f64m1))) +void vsuxseg8ei16(double * op0, vuint16mf4_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_f64m1_m))) +void vsuxseg8ei16(vbool64_t op0, double * op1, vuint16mf4_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, vfloat64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_f64m1))) +void vsuxseg2ei32(double * op0, vuint32mf2_t op1, vfloat64m1_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_f64m1_m))) +void vsuxseg2ei32(vbool64_t op0, double * op1, vuint32mf2_t op2, vfloat64m1_t op3, vfloat64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_f64m2))) +void vsuxseg2ei32(double * op0, vuint32m1_t op1, vfloat64m2_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_f64m2_m))) +void vsuxseg2ei32(vbool32_t op0, double * op1, vuint32m1_t op2, vfloat64m2_t op3, vfloat64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_f64m4))) +void vsuxseg2ei32(double * op0, vuint32m2_t op1, vfloat64m4_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_f64m4_m))) +void vsuxseg2ei32(vbool16_t op0, double * op1, vuint32m2_t op2, vfloat64m4_t op3, vfloat64m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_f64m1))) +void vsuxseg3ei32(double * op0, vuint32mf2_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_f64m1_m))) +void vsuxseg3ei32(vbool64_t op0, double * op1, vuint32mf2_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_f64m2))) +void vsuxseg3ei32(double * op0, vuint32m1_t op1, vfloat64m2_t op2, vfloat64m2_t op3, vfloat64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_f64m2_m))) +void vsuxseg3ei32(vbool32_t op0, double * op1, vuint32m1_t op2, vfloat64m2_t op3, vfloat64m2_t op4, vfloat64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_f64m1))) +void vsuxseg4ei32(double * op0, vuint32mf2_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_f64m1_m))) +void vsuxseg4ei32(vbool64_t op0, double * op1, vuint32mf2_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_f64m2))) +void vsuxseg4ei32(double * op0, vuint32m1_t op1, vfloat64m2_t op2, vfloat64m2_t op3, vfloat64m2_t op4, vfloat64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_f64m2_m))) +void vsuxseg4ei32(vbool32_t op0, double * op1, vuint32m1_t op2, vfloat64m2_t op3, vfloat64m2_t op4, vfloat64m2_t op5, vfloat64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_f64m1))) +void vsuxseg5ei32(double * op0, vuint32mf2_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_f64m1_m))) +void vsuxseg5ei32(vbool64_t op0, double * op1, vuint32mf2_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_f64m1))) +void vsuxseg6ei32(double * op0, vuint32mf2_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_f64m1_m))) +void vsuxseg6ei32(vbool64_t op0, double * op1, vuint32mf2_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_f64m1))) +void vsuxseg7ei32(double * op0, vuint32mf2_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_f64m1_m))) +void vsuxseg7ei32(vbool64_t op0, double * op1, vuint32mf2_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_f64m1))) +void vsuxseg8ei32(double * op0, vuint32mf2_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_f64m1_m))) +void vsuxseg8ei32(vbool64_t op0, double * op1, vuint32mf2_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, vfloat64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_f64m1))) +void vsuxseg2ei64(double * op0, vuint64m1_t op1, vfloat64m1_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_f64m1_m))) +void vsuxseg2ei64(vbool64_t op0, double * op1, vuint64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_f64m2))) +void vsuxseg2ei64(double * op0, vuint64m2_t op1, vfloat64m2_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_f64m2_m))) +void vsuxseg2ei64(vbool32_t op0, double * op1, vuint64m2_t op2, vfloat64m2_t op3, vfloat64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_f64m4))) +void vsuxseg2ei64(double * op0, vuint64m4_t op1, vfloat64m4_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_f64m4_m))) +void vsuxseg2ei64(vbool16_t op0, double * op1, vuint64m4_t op2, vfloat64m4_t op3, vfloat64m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_f64m1))) +void vsuxseg3ei64(double * op0, vuint64m1_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_f64m1_m))) +void vsuxseg3ei64(vbool64_t op0, double * op1, vuint64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_f64m2))) +void vsuxseg3ei64(double * op0, vuint64m2_t op1, vfloat64m2_t op2, vfloat64m2_t op3, vfloat64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_f64m2_m))) +void vsuxseg3ei64(vbool32_t op0, double * op1, vuint64m2_t op2, vfloat64m2_t op3, vfloat64m2_t op4, vfloat64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_f64m1))) +void vsuxseg4ei64(double * op0, vuint64m1_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_f64m1_m))) +void vsuxseg4ei64(vbool64_t op0, double * op1, vuint64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_f64m2))) +void vsuxseg4ei64(double * op0, vuint64m2_t op1, vfloat64m2_t op2, vfloat64m2_t op3, vfloat64m2_t op4, vfloat64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_f64m2_m))) +void vsuxseg4ei64(vbool32_t op0, double * op1, vuint64m2_t op2, vfloat64m2_t op3, vfloat64m2_t op4, vfloat64m2_t op5, vfloat64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_f64m1))) +void vsuxseg5ei64(double * op0, vuint64m1_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_f64m1_m))) +void vsuxseg5ei64(vbool64_t op0, double * op1, vuint64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_f64m1))) +void vsuxseg6ei64(double * op0, vuint64m1_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_f64m1_m))) +void vsuxseg6ei64(vbool64_t op0, double * op1, vuint64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_f64m1))) +void vsuxseg7ei64(double * op0, vuint64m1_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_f64m1_m))) +void vsuxseg7ei64(vbool64_t op0, double * op1, vuint64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_f64m1))) +void vsuxseg8ei64(double * op0, vuint64m1_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_f64m1_m))) +void vsuxseg8ei64(vbool64_t op0, double * op1, vuint64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, vfloat64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_f64m1))) +void vsoxseg2ei8(double * op0, vuint8mf8_t op1, vfloat64m1_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_f64m1_m))) +void vsoxseg2ei8(vbool64_t op0, double * op1, vuint8mf8_t op2, vfloat64m1_t op3, vfloat64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_f64m2))) +void vsoxseg2ei8(double * op0, vuint8mf4_t op1, vfloat64m2_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_f64m2_m))) +void vsoxseg2ei8(vbool32_t op0, double * op1, vuint8mf4_t op2, vfloat64m2_t op3, vfloat64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_f64m4))) +void vsoxseg2ei8(double * op0, vuint8mf2_t op1, vfloat64m4_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_f64m4_m))) +void vsoxseg2ei8(vbool16_t op0, double * op1, vuint8mf2_t op2, vfloat64m4_t op3, vfloat64m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_f64m1))) +void vsoxseg3ei8(double * op0, vuint8mf8_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_f64m1_m))) +void vsoxseg3ei8(vbool64_t op0, double * op1, vuint8mf8_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_f64m2))) +void vsoxseg3ei8(double * op0, vuint8mf4_t op1, vfloat64m2_t op2, vfloat64m2_t op3, vfloat64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_f64m2_m))) +void vsoxseg3ei8(vbool32_t op0, double * op1, vuint8mf4_t op2, vfloat64m2_t op3, vfloat64m2_t op4, vfloat64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_f64m1))) +void vsoxseg4ei8(double * op0, vuint8mf8_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_f64m1_m))) +void vsoxseg4ei8(vbool64_t op0, double * op1, vuint8mf8_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_f64m2))) +void vsoxseg4ei8(double * op0, vuint8mf4_t op1, vfloat64m2_t op2, vfloat64m2_t op3, vfloat64m2_t op4, vfloat64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_f64m2_m))) +void vsoxseg4ei8(vbool32_t op0, double * op1, vuint8mf4_t op2, vfloat64m2_t op3, vfloat64m2_t op4, vfloat64m2_t op5, vfloat64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_f64m1))) +void vsoxseg5ei8(double * op0, vuint8mf8_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_f64m1_m))) +void vsoxseg5ei8(vbool64_t op0, double * op1, vuint8mf8_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_f64m1))) +void vsoxseg6ei8(double * op0, vuint8mf8_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_f64m1_m))) +void vsoxseg6ei8(vbool64_t op0, double * op1, vuint8mf8_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_f64m1))) +void vsoxseg7ei8(double * op0, vuint8mf8_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_f64m1_m))) +void vsoxseg7ei8(vbool64_t op0, double * op1, vuint8mf8_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_f64m1))) +void vsoxseg8ei8(double * op0, vuint8mf8_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_f64m1_m))) +void vsoxseg8ei8(vbool64_t op0, double * op1, vuint8mf8_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, vfloat64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_f64m1))) +void vsoxseg2ei16(double * op0, vuint16mf4_t op1, vfloat64m1_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_f64m1_m))) +void vsoxseg2ei16(vbool64_t op0, double * op1, vuint16mf4_t op2, vfloat64m1_t op3, vfloat64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_f64m2))) +void vsoxseg2ei16(double * op0, vuint16mf2_t op1, vfloat64m2_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_f64m2_m))) +void vsoxseg2ei16(vbool32_t op0, double * op1, vuint16mf2_t op2, vfloat64m2_t op3, vfloat64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_f64m4))) +void vsoxseg2ei16(double * op0, vuint16m1_t op1, vfloat64m4_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_f64m4_m))) +void vsoxseg2ei16(vbool16_t op0, double * op1, vuint16m1_t op2, vfloat64m4_t op3, vfloat64m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_f64m1))) +void vsoxseg3ei16(double * op0, vuint16mf4_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_f64m1_m))) +void vsoxseg3ei16(vbool64_t op0, double * op1, vuint16mf4_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_f64m2))) +void vsoxseg3ei16(double * op0, vuint16mf2_t op1, vfloat64m2_t op2, vfloat64m2_t op3, vfloat64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_f64m2_m))) +void vsoxseg3ei16(vbool32_t op0, double * op1, vuint16mf2_t op2, vfloat64m2_t op3, vfloat64m2_t op4, vfloat64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_f64m1))) +void vsoxseg4ei16(double * op0, vuint16mf4_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_f64m1_m))) +void vsoxseg4ei16(vbool64_t op0, double * op1, vuint16mf4_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_f64m2))) +void vsoxseg4ei16(double * op0, vuint16mf2_t op1, vfloat64m2_t op2, vfloat64m2_t op3, vfloat64m2_t op4, vfloat64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_f64m2_m))) +void vsoxseg4ei16(vbool32_t op0, double * op1, vuint16mf2_t op2, vfloat64m2_t op3, vfloat64m2_t op4, vfloat64m2_t op5, vfloat64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_f64m1))) +void vsoxseg5ei16(double * op0, vuint16mf4_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_f64m1_m))) +void vsoxseg5ei16(vbool64_t op0, double * op1, vuint16mf4_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_f64m1))) +void vsoxseg6ei16(double * op0, vuint16mf4_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_f64m1_m))) +void vsoxseg6ei16(vbool64_t op0, double * op1, vuint16mf4_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_f64m1))) +void vsoxseg7ei16(double * op0, vuint16mf4_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_f64m1_m))) +void vsoxseg7ei16(vbool64_t op0, double * op1, vuint16mf4_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_f64m1))) +void vsoxseg8ei16(double * op0, vuint16mf4_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_f64m1_m))) +void vsoxseg8ei16(vbool64_t op0, double * op1, vuint16mf4_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, vfloat64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_f64m1))) +void vsoxseg2ei32(double * op0, vuint32mf2_t op1, vfloat64m1_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_f64m1_m))) +void vsoxseg2ei32(vbool64_t op0, double * op1, vuint32mf2_t op2, vfloat64m1_t op3, vfloat64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_f64m2))) +void vsoxseg2ei32(double * op0, vuint32m1_t op1, vfloat64m2_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_f64m2_m))) +void vsoxseg2ei32(vbool32_t op0, double * op1, vuint32m1_t op2, vfloat64m2_t op3, vfloat64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_f64m4))) +void vsoxseg2ei32(double * op0, vuint32m2_t op1, vfloat64m4_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_f64m4_m))) +void vsoxseg2ei32(vbool16_t op0, double * op1, vuint32m2_t op2, vfloat64m4_t op3, vfloat64m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_f64m1))) +void vsoxseg3ei32(double * op0, vuint32mf2_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_f64m1_m))) +void vsoxseg3ei32(vbool64_t op0, double * op1, vuint32mf2_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_f64m2))) +void vsoxseg3ei32(double * op0, vuint32m1_t op1, vfloat64m2_t op2, vfloat64m2_t op3, vfloat64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_f64m2_m))) +void vsoxseg3ei32(vbool32_t op0, double * op1, vuint32m1_t op2, vfloat64m2_t op3, vfloat64m2_t op4, vfloat64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_f64m1))) +void vsoxseg4ei32(double * op0, vuint32mf2_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_f64m1_m))) +void vsoxseg4ei32(vbool64_t op0, double * op1, vuint32mf2_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_f64m2))) +void vsoxseg4ei32(double * op0, vuint32m1_t op1, vfloat64m2_t op2, vfloat64m2_t op3, vfloat64m2_t op4, vfloat64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_f64m2_m))) +void vsoxseg4ei32(vbool32_t op0, double * op1, vuint32m1_t op2, vfloat64m2_t op3, vfloat64m2_t op4, vfloat64m2_t op5, vfloat64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_f64m1))) +void vsoxseg5ei32(double * op0, vuint32mf2_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_f64m1_m))) +void vsoxseg5ei32(vbool64_t op0, double * op1, vuint32mf2_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_f64m1))) +void vsoxseg6ei32(double * op0, vuint32mf2_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_f64m1_m))) +void vsoxseg6ei32(vbool64_t op0, double * op1, vuint32mf2_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_f64m1))) +void vsoxseg7ei32(double * op0, vuint32mf2_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_f64m1_m))) +void vsoxseg7ei32(vbool64_t op0, double * op1, vuint32mf2_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_f64m1))) +void vsoxseg8ei32(double * op0, vuint32mf2_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_f64m1_m))) +void vsoxseg8ei32(vbool64_t op0, double * op1, vuint32mf2_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, vfloat64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_f64m1))) +void vsoxseg2ei64(double * op0, vuint64m1_t op1, vfloat64m1_t op2, vfloat64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_f64m1_m))) +void vsoxseg2ei64(vbool64_t op0, double * op1, vuint64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_f64m2))) +void vsoxseg2ei64(double * op0, vuint64m2_t op1, vfloat64m2_t op2, vfloat64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_f64m2_m))) +void vsoxseg2ei64(vbool32_t op0, double * op1, vuint64m2_t op2, vfloat64m2_t op3, vfloat64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_f64m4))) +void vsoxseg2ei64(double * op0, vuint64m4_t op1, vfloat64m4_t op2, vfloat64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_f64m4_m))) +void vsoxseg2ei64(vbool16_t op0, double * op1, vuint64m4_t op2, vfloat64m4_t op3, vfloat64m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_f64m1))) +void vsoxseg3ei64(double * op0, vuint64m1_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_f64m1_m))) +void vsoxseg3ei64(vbool64_t op0, double * op1, vuint64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_f64m2))) +void vsoxseg3ei64(double * op0, vuint64m2_t op1, vfloat64m2_t op2, vfloat64m2_t op3, vfloat64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_f64m2_m))) +void vsoxseg3ei64(vbool32_t op0, double * op1, vuint64m2_t op2, vfloat64m2_t op3, vfloat64m2_t op4, vfloat64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_f64m1))) +void vsoxseg4ei64(double * op0, vuint64m1_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_f64m1_m))) +void vsoxseg4ei64(vbool64_t op0, double * op1, vuint64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_f64m2))) +void vsoxseg4ei64(double * op0, vuint64m2_t op1, vfloat64m2_t op2, vfloat64m2_t op3, vfloat64m2_t op4, vfloat64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_f64m2_m))) +void vsoxseg4ei64(vbool32_t op0, double * op1, vuint64m2_t op2, vfloat64m2_t op3, vfloat64m2_t op4, vfloat64m2_t op5, vfloat64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_f64m1))) +void vsoxseg5ei64(double * op0, vuint64m1_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_f64m1_m))) +void vsoxseg5ei64(vbool64_t op0, double * op1, vuint64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_f64m1))) +void vsoxseg6ei64(double * op0, vuint64m1_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_f64m1_m))) +void vsoxseg6ei64(vbool64_t op0, double * op1, vuint64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_f64m1))) +void vsoxseg7ei64(double * op0, vuint64m1_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_f64m1_m))) +void vsoxseg7ei64(vbool64_t op0, double * op1, vuint64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_f64m1))) +void vsoxseg8ei64(double * op0, vuint64m1_t op1, vfloat64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_f64m1_m))) +void vsoxseg8ei64(vbool64_t op0, double * op1, vuint64m1_t op2, vfloat64m1_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, vfloat64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e64_v_f64m1_m))) +void vlseg2e64(vfloat64m1_t * op0, vfloat64m1_t * op1, vbool64_t op2, vfloat64m1_t op3, vfloat64m1_t op4, const double * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e64_v_f64m2_m))) +void vlseg2e64(vfloat64m2_t * op0, vfloat64m2_t * op1, vbool32_t op2, vfloat64m2_t op3, vfloat64m2_t op4, const double * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e64_v_f64m4_m))) +void vlseg2e64(vfloat64m4_t * op0, vfloat64m4_t * op1, vbool16_t op2, vfloat64m4_t op3, vfloat64m4_t op4, const double * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e64_v_f64m1_m))) +void vlseg3e64(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vbool64_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, const double * op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e64_v_f64m2_m))) +void vlseg3e64(vfloat64m2_t * op0, vfloat64m2_t * op1, vfloat64m2_t * op2, vbool32_t op3, vfloat64m2_t op4, vfloat64m2_t op5, vfloat64m2_t op6, const double * op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e64_v_f64m1_m))) +void vlseg4e64(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vbool64_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, const double * op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e64_v_f64m2_m))) +void vlseg4e64(vfloat64m2_t * op0, vfloat64m2_t * op1, vfloat64m2_t * op2, vfloat64m2_t * op3, vbool32_t op4, vfloat64m2_t op5, vfloat64m2_t op6, vfloat64m2_t op7, vfloat64m2_t op8, const double * op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e64_v_f64m1_m))) +void vlseg5e64(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vbool64_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, vfloat64m1_t op10, const double * op11, size_t op12); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e64_v_f64m1_m))) +void vlseg6e64(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vbool64_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, vfloat64m1_t op10, vfloat64m1_t op11, vfloat64m1_t op12, const double * op13, size_t op14); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e64_v_f64m1_m))) +void vlseg7e64(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vfloat64m1_t * op6, vbool64_t op7, vfloat64m1_t op8, vfloat64m1_t op9, vfloat64m1_t op10, vfloat64m1_t op11, vfloat64m1_t op12, vfloat64m1_t op13, vfloat64m1_t op14, const double * op15, size_t op16); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e64_v_f64m1_m))) +void vlseg8e64(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vfloat64m1_t * op6, vfloat64m1_t * op7, vbool64_t op8, vfloat64m1_t op9, vfloat64m1_t op10, vfloat64m1_t op11, vfloat64m1_t op12, vfloat64m1_t op13, vfloat64m1_t op14, vfloat64m1_t op15, vfloat64m1_t op16, const double * op17, size_t op18); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e64ff_v_f64m1_m))) +void vlseg2e64ff(vfloat64m1_t * op0, vfloat64m1_t * op1, vbool64_t op2, vfloat64m1_t op3, vfloat64m1_t op4, const double * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e64ff_v_f64m2_m))) +void vlseg2e64ff(vfloat64m2_t * op0, vfloat64m2_t * op1, vbool32_t op2, vfloat64m2_t op3, vfloat64m2_t op4, const double * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e64ff_v_f64m4_m))) +void vlseg2e64ff(vfloat64m4_t * op0, vfloat64m4_t * op1, vbool16_t op2, vfloat64m4_t op3, vfloat64m4_t op4, const double * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e64ff_v_f64m1_m))) +void vlseg3e64ff(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vbool64_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, const double * op7, size_t * op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e64ff_v_f64m2_m))) +void vlseg3e64ff(vfloat64m2_t * op0, vfloat64m2_t * op1, vfloat64m2_t * op2, vbool32_t op3, vfloat64m2_t op4, vfloat64m2_t op5, vfloat64m2_t op6, const double * op7, size_t * op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e64ff_v_f64m1_m))) +void vlseg4e64ff(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vbool64_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, const double * op9, size_t * op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e64ff_v_f64m2_m))) +void vlseg4e64ff(vfloat64m2_t * op0, vfloat64m2_t * op1, vfloat64m2_t * op2, vfloat64m2_t * op3, vbool32_t op4, vfloat64m2_t op5, vfloat64m2_t op6, vfloat64m2_t op7, vfloat64m2_t op8, const double * op9, size_t * op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e64ff_v_f64m1_m))) +void vlseg5e64ff(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vbool64_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, vfloat64m1_t op10, const double * op11, size_t * op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e64ff_v_f64m1_m))) +void vlseg6e64ff(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vbool64_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, vfloat64m1_t op10, vfloat64m1_t op11, vfloat64m1_t op12, const double * op13, size_t * op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e64ff_v_f64m1_m))) +void vlseg7e64ff(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vfloat64m1_t * op6, vbool64_t op7, vfloat64m1_t op8, vfloat64m1_t op9, vfloat64m1_t op10, vfloat64m1_t op11, vfloat64m1_t op12, vfloat64m1_t op13, vfloat64m1_t op14, const double * op15, size_t * op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e64ff_v_f64m1_m))) +void vlseg8e64ff(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vfloat64m1_t * op6, vfloat64m1_t * op7, vbool64_t op8, vfloat64m1_t op9, vfloat64m1_t op10, vfloat64m1_t op11, vfloat64m1_t op12, vfloat64m1_t op13, vfloat64m1_t op14, vfloat64m1_t op15, vfloat64m1_t op16, const double * op17, size_t * op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e64_v_f64m1_m))) +void vlsseg2e64(vfloat64m1_t * op0, vfloat64m1_t * op1, vbool64_t op2, vfloat64m1_t op3, vfloat64m1_t op4, const double * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e64_v_f64m2_m))) +void vlsseg2e64(vfloat64m2_t * op0, vfloat64m2_t * op1, vbool32_t op2, vfloat64m2_t op3, vfloat64m2_t op4, const double * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e64_v_f64m4_m))) +void vlsseg2e64(vfloat64m4_t * op0, vfloat64m4_t * op1, vbool16_t op2, vfloat64m4_t op3, vfloat64m4_t op4, const double * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg3e64_v_f64m1_m))) +void vlsseg3e64(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vbool64_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, const double * op7, ptrdiff_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg3e64_v_f64m2_m))) +void vlsseg3e64(vfloat64m2_t * op0, vfloat64m2_t * op1, vfloat64m2_t * op2, vbool32_t op3, vfloat64m2_t op4, vfloat64m2_t op5, vfloat64m2_t op6, const double * op7, ptrdiff_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg4e64_v_f64m1_m))) +void vlsseg4e64(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vbool64_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, const double * op9, ptrdiff_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg4e64_v_f64m2_m))) +void vlsseg4e64(vfloat64m2_t * op0, vfloat64m2_t * op1, vfloat64m2_t * op2, vfloat64m2_t * op3, vbool32_t op4, vfloat64m2_t op5, vfloat64m2_t op6, vfloat64m2_t op7, vfloat64m2_t op8, const double * op9, ptrdiff_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg5e64_v_f64m1_m))) +void vlsseg5e64(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vbool64_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, vfloat64m1_t op10, const double * op11, ptrdiff_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg6e64_v_f64m1_m))) +void vlsseg6e64(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vbool64_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, vfloat64m1_t op10, vfloat64m1_t op11, vfloat64m1_t op12, const double * op13, ptrdiff_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg7e64_v_f64m1_m))) +void vlsseg7e64(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vfloat64m1_t * op6, vbool64_t op7, vfloat64m1_t op8, vfloat64m1_t op9, vfloat64m1_t op10, vfloat64m1_t op11, vfloat64m1_t op12, vfloat64m1_t op13, vfloat64m1_t op14, const double * op15, ptrdiff_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg8e64_v_f64m1_m))) +void vlsseg8e64(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vfloat64m1_t * op6, vfloat64m1_t * op7, vbool64_t op8, vfloat64m1_t op9, vfloat64m1_t op10, vfloat64m1_t op11, vfloat64m1_t op12, vfloat64m1_t op13, vfloat64m1_t op14, vfloat64m1_t op15, vfloat64m1_t op16, const double * op17, ptrdiff_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_f64m1))) +void vluxseg2ei8(vfloat64m1_t * op0, vfloat64m1_t * op1, const double * op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_f64m1_m))) +void vluxseg2ei8(vfloat64m1_t * op0, vfloat64m1_t * op1, vbool64_t op2, vfloat64m1_t op3, vfloat64m1_t op4, const double * op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_f64m2))) +void vluxseg2ei8(vfloat64m2_t * op0, vfloat64m2_t * op1, const double * op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_f64m2_m))) +void vluxseg2ei8(vfloat64m2_t * op0, vfloat64m2_t * op1, vbool32_t op2, vfloat64m2_t op3, vfloat64m2_t op4, const double * op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_f64m4))) +void vluxseg2ei8(vfloat64m4_t * op0, vfloat64m4_t * op1, const double * op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_f64m4_m))) +void vluxseg2ei8(vfloat64m4_t * op0, vfloat64m4_t * op1, vbool16_t op2, vfloat64m4_t op3, vfloat64m4_t op4, const double * op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_f64m1))) +void vluxseg3ei8(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, const double * op3, vuint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_f64m1_m))) +void vluxseg3ei8(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vbool64_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, const double * op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_f64m2))) +void vluxseg3ei8(vfloat64m2_t * op0, vfloat64m2_t * op1, vfloat64m2_t * op2, const double * op3, vuint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_f64m2_m))) +void vluxseg3ei8(vfloat64m2_t * op0, vfloat64m2_t * op1, vfloat64m2_t * op2, vbool32_t op3, vfloat64m2_t op4, vfloat64m2_t op5, vfloat64m2_t op6, const double * op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_f64m1))) +void vluxseg4ei8(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, const double * op4, vuint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_f64m1_m))) +void vluxseg4ei8(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vbool64_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, const double * op9, vuint8mf8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_f64m2))) +void vluxseg4ei8(vfloat64m2_t * op0, vfloat64m2_t * op1, vfloat64m2_t * op2, vfloat64m2_t * op3, const double * op4, vuint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_f64m2_m))) +void vluxseg4ei8(vfloat64m2_t * op0, vfloat64m2_t * op1, vfloat64m2_t * op2, vfloat64m2_t * op3, vbool32_t op4, vfloat64m2_t op5, vfloat64m2_t op6, vfloat64m2_t op7, vfloat64m2_t op8, const double * op9, vuint8mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_f64m1))) +void vluxseg5ei8(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, const double * op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_f64m1_m))) +void vluxseg5ei8(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vbool64_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, vfloat64m1_t op10, const double * op11, vuint8mf8_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_f64m1))) +void vluxseg6ei8(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, const double * op6, vuint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_f64m1_m))) +void vluxseg6ei8(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vbool64_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, vfloat64m1_t op10, vfloat64m1_t op11, vfloat64m1_t op12, const double * op13, vuint8mf8_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_f64m1))) +void vluxseg7ei8(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vfloat64m1_t * op6, const double * op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_f64m1_m))) +void vluxseg7ei8(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vfloat64m1_t * op6, vbool64_t op7, vfloat64m1_t op8, vfloat64m1_t op9, vfloat64m1_t op10, vfloat64m1_t op11, vfloat64m1_t op12, vfloat64m1_t op13, vfloat64m1_t op14, const double * op15, vuint8mf8_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_f64m1))) +void vluxseg8ei8(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vfloat64m1_t * op6, vfloat64m1_t * op7, const double * op8, vuint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_f64m1_m))) +void vluxseg8ei8(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vfloat64m1_t * op6, vfloat64m1_t * op7, vbool64_t op8, vfloat64m1_t op9, vfloat64m1_t op10, vfloat64m1_t op11, vfloat64m1_t op12, vfloat64m1_t op13, vfloat64m1_t op14, vfloat64m1_t op15, vfloat64m1_t op16, const double * op17, vuint8mf8_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_f64m1))) +void vluxseg2ei16(vfloat64m1_t * op0, vfloat64m1_t * op1, const double * op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_f64m1_m))) +void vluxseg2ei16(vfloat64m1_t * op0, vfloat64m1_t * op1, vbool64_t op2, vfloat64m1_t op3, vfloat64m1_t op4, const double * op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_f64m2))) +void vluxseg2ei16(vfloat64m2_t * op0, vfloat64m2_t * op1, const double * op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_f64m2_m))) +void vluxseg2ei16(vfloat64m2_t * op0, vfloat64m2_t * op1, vbool32_t op2, vfloat64m2_t op3, vfloat64m2_t op4, const double * op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_f64m4))) +void vluxseg2ei16(vfloat64m4_t * op0, vfloat64m4_t * op1, const double * op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_f64m4_m))) +void vluxseg2ei16(vfloat64m4_t * op0, vfloat64m4_t * op1, vbool16_t op2, vfloat64m4_t op3, vfloat64m4_t op4, const double * op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_f64m1))) +void vluxseg3ei16(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, const double * op3, vuint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_f64m1_m))) +void vluxseg3ei16(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vbool64_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, const double * op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_f64m2))) +void vluxseg3ei16(vfloat64m2_t * op0, vfloat64m2_t * op1, vfloat64m2_t * op2, const double * op3, vuint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_f64m2_m))) +void vluxseg3ei16(vfloat64m2_t * op0, vfloat64m2_t * op1, vfloat64m2_t * op2, vbool32_t op3, vfloat64m2_t op4, vfloat64m2_t op5, vfloat64m2_t op6, const double * op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_f64m1))) +void vluxseg4ei16(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, const double * op4, vuint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_f64m1_m))) +void vluxseg4ei16(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vbool64_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, const double * op9, vuint16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_f64m2))) +void vluxseg4ei16(vfloat64m2_t * op0, vfloat64m2_t * op1, vfloat64m2_t * op2, vfloat64m2_t * op3, const double * op4, vuint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_f64m2_m))) +void vluxseg4ei16(vfloat64m2_t * op0, vfloat64m2_t * op1, vfloat64m2_t * op2, vfloat64m2_t * op3, vbool32_t op4, vfloat64m2_t op5, vfloat64m2_t op6, vfloat64m2_t op7, vfloat64m2_t op8, const double * op9, vuint16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_f64m1))) +void vluxseg5ei16(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, const double * op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_f64m1_m))) +void vluxseg5ei16(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vbool64_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, vfloat64m1_t op10, const double * op11, vuint16mf4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_f64m1))) +void vluxseg6ei16(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, const double * op6, vuint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_f64m1_m))) +void vluxseg6ei16(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vbool64_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, vfloat64m1_t op10, vfloat64m1_t op11, vfloat64m1_t op12, const double * op13, vuint16mf4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_f64m1))) +void vluxseg7ei16(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vfloat64m1_t * op6, const double * op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_f64m1_m))) +void vluxseg7ei16(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vfloat64m1_t * op6, vbool64_t op7, vfloat64m1_t op8, vfloat64m1_t op9, vfloat64m1_t op10, vfloat64m1_t op11, vfloat64m1_t op12, vfloat64m1_t op13, vfloat64m1_t op14, const double * op15, vuint16mf4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_f64m1))) +void vluxseg8ei16(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vfloat64m1_t * op6, vfloat64m1_t * op7, const double * op8, vuint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_f64m1_m))) +void vluxseg8ei16(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vfloat64m1_t * op6, vfloat64m1_t * op7, vbool64_t op8, vfloat64m1_t op9, vfloat64m1_t op10, vfloat64m1_t op11, vfloat64m1_t op12, vfloat64m1_t op13, vfloat64m1_t op14, vfloat64m1_t op15, vfloat64m1_t op16, const double * op17, vuint16mf4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_f64m1))) +void vluxseg2ei32(vfloat64m1_t * op0, vfloat64m1_t * op1, const double * op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_f64m1_m))) +void vluxseg2ei32(vfloat64m1_t * op0, vfloat64m1_t * op1, vbool64_t op2, vfloat64m1_t op3, vfloat64m1_t op4, const double * op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_f64m2))) +void vluxseg2ei32(vfloat64m2_t * op0, vfloat64m2_t * op1, const double * op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_f64m2_m))) +void vluxseg2ei32(vfloat64m2_t * op0, vfloat64m2_t * op1, vbool32_t op2, vfloat64m2_t op3, vfloat64m2_t op4, const double * op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_f64m4))) +void vluxseg2ei32(vfloat64m4_t * op0, vfloat64m4_t * op1, const double * op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_f64m4_m))) +void vluxseg2ei32(vfloat64m4_t * op0, vfloat64m4_t * op1, vbool16_t op2, vfloat64m4_t op3, vfloat64m4_t op4, const double * op5, vuint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_f64m1))) +void vluxseg3ei32(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, const double * op3, vuint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_f64m1_m))) +void vluxseg3ei32(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vbool64_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, const double * op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_f64m2))) +void vluxseg3ei32(vfloat64m2_t * op0, vfloat64m2_t * op1, vfloat64m2_t * op2, const double * op3, vuint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_f64m2_m))) +void vluxseg3ei32(vfloat64m2_t * op0, vfloat64m2_t * op1, vfloat64m2_t * op2, vbool32_t op3, vfloat64m2_t op4, vfloat64m2_t op5, vfloat64m2_t op6, const double * op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_f64m1))) +void vluxseg4ei32(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, const double * op4, vuint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_f64m1_m))) +void vluxseg4ei32(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vbool64_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, const double * op9, vuint32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_f64m2))) +void vluxseg4ei32(vfloat64m2_t * op0, vfloat64m2_t * op1, vfloat64m2_t * op2, vfloat64m2_t * op3, const double * op4, vuint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_f64m2_m))) +void vluxseg4ei32(vfloat64m2_t * op0, vfloat64m2_t * op1, vfloat64m2_t * op2, vfloat64m2_t * op3, vbool32_t op4, vfloat64m2_t op5, vfloat64m2_t op6, vfloat64m2_t op7, vfloat64m2_t op8, const double * op9, vuint32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_f64m1))) +void vluxseg5ei32(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, const double * op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_f64m1_m))) +void vluxseg5ei32(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vbool64_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, vfloat64m1_t op10, const double * op11, vuint32mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_f64m1))) +void vluxseg6ei32(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, const double * op6, vuint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_f64m1_m))) +void vluxseg6ei32(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vbool64_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, vfloat64m1_t op10, vfloat64m1_t op11, vfloat64m1_t op12, const double * op13, vuint32mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_f64m1))) +void vluxseg7ei32(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vfloat64m1_t * op6, const double * op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_f64m1_m))) +void vluxseg7ei32(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vfloat64m1_t * op6, vbool64_t op7, vfloat64m1_t op8, vfloat64m1_t op9, vfloat64m1_t op10, vfloat64m1_t op11, vfloat64m1_t op12, vfloat64m1_t op13, vfloat64m1_t op14, const double * op15, vuint32mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_f64m1))) +void vluxseg8ei32(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vfloat64m1_t * op6, vfloat64m1_t * op7, const double * op8, vuint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_f64m1_m))) +void vluxseg8ei32(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vfloat64m1_t * op6, vfloat64m1_t * op7, vbool64_t op8, vfloat64m1_t op9, vfloat64m1_t op10, vfloat64m1_t op11, vfloat64m1_t op12, vfloat64m1_t op13, vfloat64m1_t op14, vfloat64m1_t op15, vfloat64m1_t op16, const double * op17, vuint32mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_f64m1))) +void vluxseg2ei64(vfloat64m1_t * op0, vfloat64m1_t * op1, const double * op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_f64m1_m))) +void vluxseg2ei64(vfloat64m1_t * op0, vfloat64m1_t * op1, vbool64_t op2, vfloat64m1_t op3, vfloat64m1_t op4, const double * op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_f64m2))) +void vluxseg2ei64(vfloat64m2_t * op0, vfloat64m2_t * op1, const double * op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_f64m2_m))) +void vluxseg2ei64(vfloat64m2_t * op0, vfloat64m2_t * op1, vbool32_t op2, vfloat64m2_t op3, vfloat64m2_t op4, const double * op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_f64m4))) +void vluxseg2ei64(vfloat64m4_t * op0, vfloat64m4_t * op1, const double * op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_f64m4_m))) +void vluxseg2ei64(vfloat64m4_t * op0, vfloat64m4_t * op1, vbool16_t op2, vfloat64m4_t op3, vfloat64m4_t op4, const double * op5, vuint64m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_f64m1))) +void vluxseg3ei64(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, const double * op3, vuint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_f64m1_m))) +void vluxseg3ei64(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vbool64_t op3, vfloat64m1_t op4, vfloat64m1_t op5, vfloat64m1_t op6, const double * op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_f64m2))) +void vluxseg3ei64(vfloat64m2_t * op0, vfloat64m2_t * op1, vfloat64m2_t * op2, const double * op3, vuint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_f64m2_m))) +void vluxseg3ei64(vfloat64m2_t * op0, vfloat64m2_t * op1, vfloat64m2_t * op2, vbool32_t op3, vfloat64m2_t op4, vfloat64m2_t op5, vfloat64m2_t op6, const double * op7, vuint64m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_f64m1))) +void vluxseg4ei64(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, const double * op4, vuint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_f64m1_m))) +void vluxseg4ei64(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vbool64_t op4, vfloat64m1_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, const double * op9, vuint64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_f64m2))) +void vluxseg4ei64(vfloat64m2_t * op0, vfloat64m2_t * op1, vfloat64m2_t * op2, vfloat64m2_t * op3, const double * op4, vuint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_f64m2_m))) +void vluxseg4ei64(vfloat64m2_t * op0, vfloat64m2_t * op1, vfloat64m2_t * op2, vfloat64m2_t * op3, vbool32_t op4, vfloat64m2_t op5, vfloat64m2_t op6, vfloat64m2_t op7, vfloat64m2_t op8, const double * op9, vuint64m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_f64m1))) +void vluxseg5ei64(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, const double * op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_f64m1_m))) +void vluxseg5ei64(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vbool64_t op5, vfloat64m1_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, vfloat64m1_t op10, const double * op11, vuint64m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_f64m1))) +void vluxseg6ei64(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, const double * op6, vuint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_f64m1_m))) +void vluxseg6ei64(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vbool64_t op6, vfloat64m1_t op7, vfloat64m1_t op8, vfloat64m1_t op9, vfloat64m1_t op10, vfloat64m1_t op11, vfloat64m1_t op12, const double * op13, vuint64m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_f64m1))) +void vluxseg7ei64(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vfloat64m1_t * op6, const double * op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_f64m1_m))) +void vluxseg7ei64(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vfloat64m1_t * op6, vbool64_t op7, vfloat64m1_t op8, vfloat64m1_t op9, vfloat64m1_t op10, vfloat64m1_t op11, vfloat64m1_t op12, vfloat64m1_t op13, vfloat64m1_t op14, const double * op15, vuint64m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_f64m1))) +void vluxseg8ei64(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vfloat64m1_t * op6, vfloat64m1_t * op7, const double * op8, vuint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_f64m1_m))) +void vluxseg8ei64(vfloat64m1_t * op0, vfloat64m1_t * op1, vfloat64m1_t * op2, vfloat64m1_t * op3, vfloat64m1_t * op4, vfloat64m1_t * op5, vfloat64m1_t * op6, vfloat64m1_t * op7, vbool64_t op8, vfloat64m1_t op9, vfloat64m1_t op10, vfloat64m1_t op11, vfloat64m1_t op12, vfloat64m1_t op13, vfloat64m1_t op14, vfloat64m1_t op15, vfloat64m1_t op16, const double * op17, vuint64m1_t op18, size_t op19); + +#endif + +#if defined(__riscv_zfh) && defined(__riscv_zvlsseg) +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_f16m1))) +void vloxseg2ei8(vfloat16m1_t * op0, vfloat16m1_t * op1, const _Float16 * op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_f16m1_m))) +void vloxseg2ei8(vfloat16m1_t * op0, vfloat16m1_t * op1, vbool16_t op2, vfloat16m1_t op3, vfloat16m1_t op4, const _Float16 * op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_f16m2))) +void vloxseg2ei8(vfloat16m2_t * op0, vfloat16m2_t * op1, const _Float16 * op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_f16m2_m))) +void vloxseg2ei8(vfloat16m2_t * op0, vfloat16m2_t * op1, vbool8_t op2, vfloat16m2_t op3, vfloat16m2_t op4, const _Float16 * op5, vuint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_f16m4))) +void vloxseg2ei8(vfloat16m4_t * op0, vfloat16m4_t * op1, const _Float16 * op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_f16m4_m))) +void vloxseg2ei8(vfloat16m4_t * op0, vfloat16m4_t * op1, vbool4_t op2, vfloat16m4_t op3, vfloat16m4_t op4, const _Float16 * op5, vuint8m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_f16mf2))) +void vloxseg2ei8(vfloat16mf2_t * op0, vfloat16mf2_t * op1, const _Float16 * op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_f16mf2_m))) +void vloxseg2ei8(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vbool32_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, const _Float16 * op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_f16mf4))) +void vloxseg2ei8(vfloat16mf4_t * op0, vfloat16mf4_t * op1, const _Float16 * op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei8_v_f16mf4_m))) +void vloxseg2ei8(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vbool64_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, const _Float16 * op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_f16m1))) +void vloxseg3ei8(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, const _Float16 * op3, vuint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_f16m1_m))) +void vloxseg3ei8(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vbool16_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, const _Float16 * op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_f16m2))) +void vloxseg3ei8(vfloat16m2_t * op0, vfloat16m2_t * op1, vfloat16m2_t * op2, const _Float16 * op3, vuint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_f16m2_m))) +void vloxseg3ei8(vfloat16m2_t * op0, vfloat16m2_t * op1, vfloat16m2_t * op2, vbool8_t op3, vfloat16m2_t op4, vfloat16m2_t op5, vfloat16m2_t op6, const _Float16 * op7, vuint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_f16mf2))) +void vloxseg3ei8(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, const _Float16 * op3, vuint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_f16mf2_m))) +void vloxseg3ei8(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vbool32_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, const _Float16 * op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_f16mf4))) +void vloxseg3ei8(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, const _Float16 * op3, vuint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei8_v_f16mf4_m))) +void vloxseg3ei8(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vbool64_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, const _Float16 * op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_f16m1))) +void vloxseg4ei8(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, const _Float16 * op4, vuint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_f16m1_m))) +void vloxseg4ei8(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vbool16_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, const _Float16 * op9, vuint8mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_f16m2))) +void vloxseg4ei8(vfloat16m2_t * op0, vfloat16m2_t * op1, vfloat16m2_t * op2, vfloat16m2_t * op3, const _Float16 * op4, vuint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_f16m2_m))) +void vloxseg4ei8(vfloat16m2_t * op0, vfloat16m2_t * op1, vfloat16m2_t * op2, vfloat16m2_t * op3, vbool8_t op4, vfloat16m2_t op5, vfloat16m2_t op6, vfloat16m2_t op7, vfloat16m2_t op8, const _Float16 * op9, vuint8m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_f16mf2))) +void vloxseg4ei8(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, const _Float16 * op4, vuint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_f16mf2_m))) +void vloxseg4ei8(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vbool32_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, const _Float16 * op9, vuint8mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_f16mf4))) +void vloxseg4ei8(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, const _Float16 * op4, vuint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei8_v_f16mf4_m))) +void vloxseg4ei8(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vbool64_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, const _Float16 * op9, vuint8mf8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_f16m1))) +void vloxseg5ei8(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, const _Float16 * op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_f16m1_m))) +void vloxseg5ei8(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vbool16_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, vfloat16m1_t op10, const _Float16 * op11, vuint8mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_f16mf2))) +void vloxseg5ei8(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, const _Float16 * op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_f16mf2_m))) +void vloxseg5ei8(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vbool32_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, const _Float16 * op11, vuint8mf4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_f16mf4))) +void vloxseg5ei8(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, const _Float16 * op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei8_v_f16mf4_m))) +void vloxseg5ei8(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vbool64_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, const _Float16 * op11, vuint8mf8_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_f16m1))) +void vloxseg6ei8(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, const _Float16 * op6, vuint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_f16m1_m))) +void vloxseg6ei8(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vbool16_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, vfloat16m1_t op10, vfloat16m1_t op11, vfloat16m1_t op12, const _Float16 * op13, vuint8mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_f16mf2))) +void vloxseg6ei8(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, const _Float16 * op6, vuint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_f16mf2_m))) +void vloxseg6ei8(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vbool32_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, vfloat16mf2_t op11, vfloat16mf2_t op12, const _Float16 * op13, vuint8mf4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_f16mf4))) +void vloxseg6ei8(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, const _Float16 * op6, vuint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei8_v_f16mf4_m))) +void vloxseg6ei8(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vbool64_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, vfloat16mf4_t op11, vfloat16mf4_t op12, const _Float16 * op13, vuint8mf8_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_f16m1))) +void vloxseg7ei8(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vfloat16m1_t * op6, const _Float16 * op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_f16m1_m))) +void vloxseg7ei8(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vfloat16m1_t * op6, vbool16_t op7, vfloat16m1_t op8, vfloat16m1_t op9, vfloat16m1_t op10, vfloat16m1_t op11, vfloat16m1_t op12, vfloat16m1_t op13, vfloat16m1_t op14, const _Float16 * op15, vuint8mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_f16mf2))) +void vloxseg7ei8(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vfloat16mf2_t * op6, const _Float16 * op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_f16mf2_m))) +void vloxseg7ei8(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vfloat16mf2_t * op6, vbool32_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, vfloat16mf2_t op11, vfloat16mf2_t op12, vfloat16mf2_t op13, vfloat16mf2_t op14, const _Float16 * op15, vuint8mf4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_f16mf4))) +void vloxseg7ei8(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vfloat16mf4_t * op6, const _Float16 * op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei8_v_f16mf4_m))) +void vloxseg7ei8(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vfloat16mf4_t * op6, vbool64_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, vfloat16mf4_t op11, vfloat16mf4_t op12, vfloat16mf4_t op13, vfloat16mf4_t op14, const _Float16 * op15, vuint8mf8_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_f16m1))) +void vloxseg8ei8(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vfloat16m1_t * op6, vfloat16m1_t * op7, const _Float16 * op8, vuint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_f16m1_m))) +void vloxseg8ei8(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vfloat16m1_t * op6, vfloat16m1_t * op7, vbool16_t op8, vfloat16m1_t op9, vfloat16m1_t op10, vfloat16m1_t op11, vfloat16m1_t op12, vfloat16m1_t op13, vfloat16m1_t op14, vfloat16m1_t op15, vfloat16m1_t op16, const _Float16 * op17, vuint8mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_f16mf2))) +void vloxseg8ei8(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vfloat16mf2_t * op6, vfloat16mf2_t * op7, const _Float16 * op8, vuint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_f16mf2_m))) +void vloxseg8ei8(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vfloat16mf2_t * op6, vfloat16mf2_t * op7, vbool32_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, vfloat16mf2_t op11, vfloat16mf2_t op12, vfloat16mf2_t op13, vfloat16mf2_t op14, vfloat16mf2_t op15, vfloat16mf2_t op16, const _Float16 * op17, vuint8mf4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_f16mf4))) +void vloxseg8ei8(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vfloat16mf4_t * op6, vfloat16mf4_t * op7, const _Float16 * op8, vuint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei8_v_f16mf4_m))) +void vloxseg8ei8(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vfloat16mf4_t * op6, vfloat16mf4_t * op7, vbool64_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, vfloat16mf4_t op11, vfloat16mf4_t op12, vfloat16mf4_t op13, vfloat16mf4_t op14, vfloat16mf4_t op15, vfloat16mf4_t op16, const _Float16 * op17, vuint8mf8_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_f16m1))) +void vloxseg2ei16(vfloat16m1_t * op0, vfloat16m1_t * op1, const _Float16 * op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_f16m1_m))) +void vloxseg2ei16(vfloat16m1_t * op0, vfloat16m1_t * op1, vbool16_t op2, vfloat16m1_t op3, vfloat16m1_t op4, const _Float16 * op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_f16m2))) +void vloxseg2ei16(vfloat16m2_t * op0, vfloat16m2_t * op1, const _Float16 * op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_f16m2_m))) +void vloxseg2ei16(vfloat16m2_t * op0, vfloat16m2_t * op1, vbool8_t op2, vfloat16m2_t op3, vfloat16m2_t op4, const _Float16 * op5, vuint16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_f16m4))) +void vloxseg2ei16(vfloat16m4_t * op0, vfloat16m4_t * op1, const _Float16 * op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_f16m4_m))) +void vloxseg2ei16(vfloat16m4_t * op0, vfloat16m4_t * op1, vbool4_t op2, vfloat16m4_t op3, vfloat16m4_t op4, const _Float16 * op5, vuint16m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_f16mf2))) +void vloxseg2ei16(vfloat16mf2_t * op0, vfloat16mf2_t * op1, const _Float16 * op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_f16mf2_m))) +void vloxseg2ei16(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vbool32_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, const _Float16 * op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_f16mf4))) +void vloxseg2ei16(vfloat16mf4_t * op0, vfloat16mf4_t * op1, const _Float16 * op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei16_v_f16mf4_m))) +void vloxseg2ei16(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vbool64_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, const _Float16 * op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_f16m1))) +void vloxseg3ei16(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, const _Float16 * op3, vuint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_f16m1_m))) +void vloxseg3ei16(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vbool16_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, const _Float16 * op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_f16m2))) +void vloxseg3ei16(vfloat16m2_t * op0, vfloat16m2_t * op1, vfloat16m2_t * op2, const _Float16 * op3, vuint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_f16m2_m))) +void vloxseg3ei16(vfloat16m2_t * op0, vfloat16m2_t * op1, vfloat16m2_t * op2, vbool8_t op3, vfloat16m2_t op4, vfloat16m2_t op5, vfloat16m2_t op6, const _Float16 * op7, vuint16m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_f16mf2))) +void vloxseg3ei16(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, const _Float16 * op3, vuint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_f16mf2_m))) +void vloxseg3ei16(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vbool32_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, const _Float16 * op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_f16mf4))) +void vloxseg3ei16(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, const _Float16 * op3, vuint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei16_v_f16mf4_m))) +void vloxseg3ei16(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vbool64_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, const _Float16 * op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_f16m1))) +void vloxseg4ei16(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, const _Float16 * op4, vuint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_f16m1_m))) +void vloxseg4ei16(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vbool16_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, const _Float16 * op9, vuint16m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_f16m2))) +void vloxseg4ei16(vfloat16m2_t * op0, vfloat16m2_t * op1, vfloat16m2_t * op2, vfloat16m2_t * op3, const _Float16 * op4, vuint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_f16m2_m))) +void vloxseg4ei16(vfloat16m2_t * op0, vfloat16m2_t * op1, vfloat16m2_t * op2, vfloat16m2_t * op3, vbool8_t op4, vfloat16m2_t op5, vfloat16m2_t op6, vfloat16m2_t op7, vfloat16m2_t op8, const _Float16 * op9, vuint16m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_f16mf2))) +void vloxseg4ei16(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, const _Float16 * op4, vuint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_f16mf2_m))) +void vloxseg4ei16(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vbool32_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, const _Float16 * op9, vuint16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_f16mf4))) +void vloxseg4ei16(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, const _Float16 * op4, vuint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei16_v_f16mf4_m))) +void vloxseg4ei16(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vbool64_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, const _Float16 * op9, vuint16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_f16m1))) +void vloxseg5ei16(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, const _Float16 * op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_f16m1_m))) +void vloxseg5ei16(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vbool16_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, vfloat16m1_t op10, const _Float16 * op11, vuint16m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_f16mf2))) +void vloxseg5ei16(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, const _Float16 * op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_f16mf2_m))) +void vloxseg5ei16(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vbool32_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, const _Float16 * op11, vuint16mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_f16mf4))) +void vloxseg5ei16(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, const _Float16 * op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei16_v_f16mf4_m))) +void vloxseg5ei16(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vbool64_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, const _Float16 * op11, vuint16mf4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_f16m1))) +void vloxseg6ei16(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, const _Float16 * op6, vuint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_f16m1_m))) +void vloxseg6ei16(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vbool16_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, vfloat16m1_t op10, vfloat16m1_t op11, vfloat16m1_t op12, const _Float16 * op13, vuint16m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_f16mf2))) +void vloxseg6ei16(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, const _Float16 * op6, vuint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_f16mf2_m))) +void vloxseg6ei16(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vbool32_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, vfloat16mf2_t op11, vfloat16mf2_t op12, const _Float16 * op13, vuint16mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_f16mf4))) +void vloxseg6ei16(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, const _Float16 * op6, vuint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei16_v_f16mf4_m))) +void vloxseg6ei16(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vbool64_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, vfloat16mf4_t op11, vfloat16mf4_t op12, const _Float16 * op13, vuint16mf4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_f16m1))) +void vloxseg7ei16(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vfloat16m1_t * op6, const _Float16 * op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_f16m1_m))) +void vloxseg7ei16(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vfloat16m1_t * op6, vbool16_t op7, vfloat16m1_t op8, vfloat16m1_t op9, vfloat16m1_t op10, vfloat16m1_t op11, vfloat16m1_t op12, vfloat16m1_t op13, vfloat16m1_t op14, const _Float16 * op15, vuint16m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_f16mf2))) +void vloxseg7ei16(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vfloat16mf2_t * op6, const _Float16 * op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_f16mf2_m))) +void vloxseg7ei16(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vfloat16mf2_t * op6, vbool32_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, vfloat16mf2_t op11, vfloat16mf2_t op12, vfloat16mf2_t op13, vfloat16mf2_t op14, const _Float16 * op15, vuint16mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_f16mf4))) +void vloxseg7ei16(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vfloat16mf4_t * op6, const _Float16 * op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei16_v_f16mf4_m))) +void vloxseg7ei16(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vfloat16mf4_t * op6, vbool64_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, vfloat16mf4_t op11, vfloat16mf4_t op12, vfloat16mf4_t op13, vfloat16mf4_t op14, const _Float16 * op15, vuint16mf4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_f16m1))) +void vloxseg8ei16(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vfloat16m1_t * op6, vfloat16m1_t * op7, const _Float16 * op8, vuint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_f16m1_m))) +void vloxseg8ei16(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vfloat16m1_t * op6, vfloat16m1_t * op7, vbool16_t op8, vfloat16m1_t op9, vfloat16m1_t op10, vfloat16m1_t op11, vfloat16m1_t op12, vfloat16m1_t op13, vfloat16m1_t op14, vfloat16m1_t op15, vfloat16m1_t op16, const _Float16 * op17, vuint16m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_f16mf2))) +void vloxseg8ei16(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vfloat16mf2_t * op6, vfloat16mf2_t * op7, const _Float16 * op8, vuint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_f16mf2_m))) +void vloxseg8ei16(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vfloat16mf2_t * op6, vfloat16mf2_t * op7, vbool32_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, vfloat16mf2_t op11, vfloat16mf2_t op12, vfloat16mf2_t op13, vfloat16mf2_t op14, vfloat16mf2_t op15, vfloat16mf2_t op16, const _Float16 * op17, vuint16mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_f16mf4))) +void vloxseg8ei16(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vfloat16mf4_t * op6, vfloat16mf4_t * op7, const _Float16 * op8, vuint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei16_v_f16mf4_m))) +void vloxseg8ei16(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vfloat16mf4_t * op6, vfloat16mf4_t * op7, vbool64_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, vfloat16mf4_t op11, vfloat16mf4_t op12, vfloat16mf4_t op13, vfloat16mf4_t op14, vfloat16mf4_t op15, vfloat16mf4_t op16, const _Float16 * op17, vuint16mf4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_f16m1))) +void vloxseg2ei32(vfloat16m1_t * op0, vfloat16m1_t * op1, const _Float16 * op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_f16m1_m))) +void vloxseg2ei32(vfloat16m1_t * op0, vfloat16m1_t * op1, vbool16_t op2, vfloat16m1_t op3, vfloat16m1_t op4, const _Float16 * op5, vuint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_f16m2))) +void vloxseg2ei32(vfloat16m2_t * op0, vfloat16m2_t * op1, const _Float16 * op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_f16m2_m))) +void vloxseg2ei32(vfloat16m2_t * op0, vfloat16m2_t * op1, vbool8_t op2, vfloat16m2_t op3, vfloat16m2_t op4, const _Float16 * op5, vuint32m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_f16m4))) +void vloxseg2ei32(vfloat16m4_t * op0, vfloat16m4_t * op1, const _Float16 * op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_f16m4_m))) +void vloxseg2ei32(vfloat16m4_t * op0, vfloat16m4_t * op1, vbool4_t op2, vfloat16m4_t op3, vfloat16m4_t op4, const _Float16 * op5, vuint32m8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_f16mf2))) +void vloxseg2ei32(vfloat16mf2_t * op0, vfloat16mf2_t * op1, const _Float16 * op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_f16mf2_m))) +void vloxseg2ei32(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vbool32_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, const _Float16 * op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_f16mf4))) +void vloxseg2ei32(vfloat16mf4_t * op0, vfloat16mf4_t * op1, const _Float16 * op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei32_v_f16mf4_m))) +void vloxseg2ei32(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vbool64_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, const _Float16 * op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_f16m1))) +void vloxseg3ei32(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, const _Float16 * op3, vuint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_f16m1_m))) +void vloxseg3ei32(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vbool16_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, const _Float16 * op7, vuint32m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_f16m2))) +void vloxseg3ei32(vfloat16m2_t * op0, vfloat16m2_t * op1, vfloat16m2_t * op2, const _Float16 * op3, vuint32m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_f16m2_m))) +void vloxseg3ei32(vfloat16m2_t * op0, vfloat16m2_t * op1, vfloat16m2_t * op2, vbool8_t op3, vfloat16m2_t op4, vfloat16m2_t op5, vfloat16m2_t op6, const _Float16 * op7, vuint32m4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_f16mf2))) +void vloxseg3ei32(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, const _Float16 * op3, vuint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_f16mf2_m))) +void vloxseg3ei32(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vbool32_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, const _Float16 * op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_f16mf4))) +void vloxseg3ei32(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, const _Float16 * op3, vuint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei32_v_f16mf4_m))) +void vloxseg3ei32(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vbool64_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, const _Float16 * op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_f16m1))) +void vloxseg4ei32(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, const _Float16 * op4, vuint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_f16m1_m))) +void vloxseg4ei32(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vbool16_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, const _Float16 * op9, vuint32m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_f16m2))) +void vloxseg4ei32(vfloat16m2_t * op0, vfloat16m2_t * op1, vfloat16m2_t * op2, vfloat16m2_t * op3, const _Float16 * op4, vuint32m4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_f16m2_m))) +void vloxseg4ei32(vfloat16m2_t * op0, vfloat16m2_t * op1, vfloat16m2_t * op2, vfloat16m2_t * op3, vbool8_t op4, vfloat16m2_t op5, vfloat16m2_t op6, vfloat16m2_t op7, vfloat16m2_t op8, const _Float16 * op9, vuint32m4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_f16mf2))) +void vloxseg4ei32(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, const _Float16 * op4, vuint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_f16mf2_m))) +void vloxseg4ei32(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vbool32_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, const _Float16 * op9, vuint32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_f16mf4))) +void vloxseg4ei32(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, const _Float16 * op4, vuint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei32_v_f16mf4_m))) +void vloxseg4ei32(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vbool64_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, const _Float16 * op9, vuint32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_f16m1))) +void vloxseg5ei32(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, const _Float16 * op5, vuint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_f16m1_m))) +void vloxseg5ei32(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vbool16_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, vfloat16m1_t op10, const _Float16 * op11, vuint32m2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_f16mf2))) +void vloxseg5ei32(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, const _Float16 * op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_f16mf2_m))) +void vloxseg5ei32(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vbool32_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, const _Float16 * op11, vuint32m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_f16mf4))) +void vloxseg5ei32(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, const _Float16 * op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei32_v_f16mf4_m))) +void vloxseg5ei32(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vbool64_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, const _Float16 * op11, vuint32mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_f16m1))) +void vloxseg6ei32(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, const _Float16 * op6, vuint32m2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_f16m1_m))) +void vloxseg6ei32(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vbool16_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, vfloat16m1_t op10, vfloat16m1_t op11, vfloat16m1_t op12, const _Float16 * op13, vuint32m2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_f16mf2))) +void vloxseg6ei32(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, const _Float16 * op6, vuint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_f16mf2_m))) +void vloxseg6ei32(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vbool32_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, vfloat16mf2_t op11, vfloat16mf2_t op12, const _Float16 * op13, vuint32m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_f16mf4))) +void vloxseg6ei32(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, const _Float16 * op6, vuint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei32_v_f16mf4_m))) +void vloxseg6ei32(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vbool64_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, vfloat16mf4_t op11, vfloat16mf4_t op12, const _Float16 * op13, vuint32mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_f16m1))) +void vloxseg7ei32(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vfloat16m1_t * op6, const _Float16 * op7, vuint32m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_f16m1_m))) +void vloxseg7ei32(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vfloat16m1_t * op6, vbool16_t op7, vfloat16m1_t op8, vfloat16m1_t op9, vfloat16m1_t op10, vfloat16m1_t op11, vfloat16m1_t op12, vfloat16m1_t op13, vfloat16m1_t op14, const _Float16 * op15, vuint32m2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_f16mf2))) +void vloxseg7ei32(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vfloat16mf2_t * op6, const _Float16 * op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_f16mf2_m))) +void vloxseg7ei32(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vfloat16mf2_t * op6, vbool32_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, vfloat16mf2_t op11, vfloat16mf2_t op12, vfloat16mf2_t op13, vfloat16mf2_t op14, const _Float16 * op15, vuint32m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_f16mf4))) +void vloxseg7ei32(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vfloat16mf4_t * op6, const _Float16 * op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei32_v_f16mf4_m))) +void vloxseg7ei32(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vfloat16mf4_t * op6, vbool64_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, vfloat16mf4_t op11, vfloat16mf4_t op12, vfloat16mf4_t op13, vfloat16mf4_t op14, const _Float16 * op15, vuint32mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_f16m1))) +void vloxseg8ei32(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vfloat16m1_t * op6, vfloat16m1_t * op7, const _Float16 * op8, vuint32m2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_f16m1_m))) +void vloxseg8ei32(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vfloat16m1_t * op6, vfloat16m1_t * op7, vbool16_t op8, vfloat16m1_t op9, vfloat16m1_t op10, vfloat16m1_t op11, vfloat16m1_t op12, vfloat16m1_t op13, vfloat16m1_t op14, vfloat16m1_t op15, vfloat16m1_t op16, const _Float16 * op17, vuint32m2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_f16mf2))) +void vloxseg8ei32(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vfloat16mf2_t * op6, vfloat16mf2_t * op7, const _Float16 * op8, vuint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_f16mf2_m))) +void vloxseg8ei32(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vfloat16mf2_t * op6, vfloat16mf2_t * op7, vbool32_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, vfloat16mf2_t op11, vfloat16mf2_t op12, vfloat16mf2_t op13, vfloat16mf2_t op14, vfloat16mf2_t op15, vfloat16mf2_t op16, const _Float16 * op17, vuint32m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_f16mf4))) +void vloxseg8ei32(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vfloat16mf4_t * op6, vfloat16mf4_t * op7, const _Float16 * op8, vuint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei32_v_f16mf4_m))) +void vloxseg8ei32(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vfloat16mf4_t * op6, vfloat16mf4_t * op7, vbool64_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, vfloat16mf4_t op11, vfloat16mf4_t op12, vfloat16mf4_t op13, vfloat16mf4_t op14, vfloat16mf4_t op15, vfloat16mf4_t op16, const _Float16 * op17, vuint32mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_f16m1))) +void vloxseg2ei64(vfloat16m1_t * op0, vfloat16m1_t * op1, const _Float16 * op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_f16m1_m))) +void vloxseg2ei64(vfloat16m1_t * op0, vfloat16m1_t * op1, vbool16_t op2, vfloat16m1_t op3, vfloat16m1_t op4, const _Float16 * op5, vuint64m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_f16m2))) +void vloxseg2ei64(vfloat16m2_t * op0, vfloat16m2_t * op1, const _Float16 * op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_f16m2_m))) +void vloxseg2ei64(vfloat16m2_t * op0, vfloat16m2_t * op1, vbool8_t op2, vfloat16m2_t op3, vfloat16m2_t op4, const _Float16 * op5, vuint64m8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_f16mf2))) +void vloxseg2ei64(vfloat16mf2_t * op0, vfloat16mf2_t * op1, const _Float16 * op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_f16mf2_m))) +void vloxseg2ei64(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vbool32_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, const _Float16 * op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_f16mf4))) +void vloxseg2ei64(vfloat16mf4_t * op0, vfloat16mf4_t * op1, const _Float16 * op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg2ei64_v_f16mf4_m))) +void vloxseg2ei64(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vbool64_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, const _Float16 * op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_f16m1))) +void vloxseg3ei64(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, const _Float16 * op3, vuint64m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_f16m1_m))) +void vloxseg3ei64(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vbool16_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, const _Float16 * op7, vuint64m4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_f16m2))) +void vloxseg3ei64(vfloat16m2_t * op0, vfloat16m2_t * op1, vfloat16m2_t * op2, const _Float16 * op3, vuint64m8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_f16m2_m))) +void vloxseg3ei64(vfloat16m2_t * op0, vfloat16m2_t * op1, vfloat16m2_t * op2, vbool8_t op3, vfloat16m2_t op4, vfloat16m2_t op5, vfloat16m2_t op6, const _Float16 * op7, vuint64m8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_f16mf2))) +void vloxseg3ei64(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, const _Float16 * op3, vuint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_f16mf2_m))) +void vloxseg3ei64(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vbool32_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, const _Float16 * op7, vuint64m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_f16mf4))) +void vloxseg3ei64(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, const _Float16 * op3, vuint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg3ei64_v_f16mf4_m))) +void vloxseg3ei64(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vbool64_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, const _Float16 * op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_f16m1))) +void vloxseg4ei64(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, const _Float16 * op4, vuint64m4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_f16m1_m))) +void vloxseg4ei64(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vbool16_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, const _Float16 * op9, vuint64m4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_f16m2))) +void vloxseg4ei64(vfloat16m2_t * op0, vfloat16m2_t * op1, vfloat16m2_t * op2, vfloat16m2_t * op3, const _Float16 * op4, vuint64m8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_f16m2_m))) +void vloxseg4ei64(vfloat16m2_t * op0, vfloat16m2_t * op1, vfloat16m2_t * op2, vfloat16m2_t * op3, vbool8_t op4, vfloat16m2_t op5, vfloat16m2_t op6, vfloat16m2_t op7, vfloat16m2_t op8, const _Float16 * op9, vuint64m8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_f16mf2))) +void vloxseg4ei64(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, const _Float16 * op4, vuint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_f16mf2_m))) +void vloxseg4ei64(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vbool32_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, const _Float16 * op9, vuint64m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_f16mf4))) +void vloxseg4ei64(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, const _Float16 * op4, vuint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg4ei64_v_f16mf4_m))) +void vloxseg4ei64(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vbool64_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, const _Float16 * op9, vuint64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_f16m1))) +void vloxseg5ei64(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, const _Float16 * op5, vuint64m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_f16m1_m))) +void vloxseg5ei64(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vbool16_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, vfloat16m1_t op10, const _Float16 * op11, vuint64m4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_f16mf2))) +void vloxseg5ei64(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, const _Float16 * op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_f16mf2_m))) +void vloxseg5ei64(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vbool32_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, const _Float16 * op11, vuint64m2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_f16mf4))) +void vloxseg5ei64(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, const _Float16 * op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg5ei64_v_f16mf4_m))) +void vloxseg5ei64(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vbool64_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, const _Float16 * op11, vuint64m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_f16m1))) +void vloxseg6ei64(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, const _Float16 * op6, vuint64m4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_f16m1_m))) +void vloxseg6ei64(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vbool16_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, vfloat16m1_t op10, vfloat16m1_t op11, vfloat16m1_t op12, const _Float16 * op13, vuint64m4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_f16mf2))) +void vloxseg6ei64(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, const _Float16 * op6, vuint64m2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_f16mf2_m))) +void vloxseg6ei64(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vbool32_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, vfloat16mf2_t op11, vfloat16mf2_t op12, const _Float16 * op13, vuint64m2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_f16mf4))) +void vloxseg6ei64(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, const _Float16 * op6, vuint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg6ei64_v_f16mf4_m))) +void vloxseg6ei64(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vbool64_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, vfloat16mf4_t op11, vfloat16mf4_t op12, const _Float16 * op13, vuint64m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_f16m1))) +void vloxseg7ei64(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vfloat16m1_t * op6, const _Float16 * op7, vuint64m4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_f16m1_m))) +void vloxseg7ei64(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vfloat16m1_t * op6, vbool16_t op7, vfloat16m1_t op8, vfloat16m1_t op9, vfloat16m1_t op10, vfloat16m1_t op11, vfloat16m1_t op12, vfloat16m1_t op13, vfloat16m1_t op14, const _Float16 * op15, vuint64m4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_f16mf2))) +void vloxseg7ei64(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vfloat16mf2_t * op6, const _Float16 * op7, vuint64m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_f16mf2_m))) +void vloxseg7ei64(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vfloat16mf2_t * op6, vbool32_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, vfloat16mf2_t op11, vfloat16mf2_t op12, vfloat16mf2_t op13, vfloat16mf2_t op14, const _Float16 * op15, vuint64m2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_f16mf4))) +void vloxseg7ei64(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vfloat16mf4_t * op6, const _Float16 * op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg7ei64_v_f16mf4_m))) +void vloxseg7ei64(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vfloat16mf4_t * op6, vbool64_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, vfloat16mf4_t op11, vfloat16mf4_t op12, vfloat16mf4_t op13, vfloat16mf4_t op14, const _Float16 * op15, vuint64m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_f16m1))) +void vloxseg8ei64(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vfloat16m1_t * op6, vfloat16m1_t * op7, const _Float16 * op8, vuint64m4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_f16m1_m))) +void vloxseg8ei64(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vfloat16m1_t * op6, vfloat16m1_t * op7, vbool16_t op8, vfloat16m1_t op9, vfloat16m1_t op10, vfloat16m1_t op11, vfloat16m1_t op12, vfloat16m1_t op13, vfloat16m1_t op14, vfloat16m1_t op15, vfloat16m1_t op16, const _Float16 * op17, vuint64m4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_f16mf2))) +void vloxseg8ei64(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vfloat16mf2_t * op6, vfloat16mf2_t * op7, const _Float16 * op8, vuint64m2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_f16mf2_m))) +void vloxseg8ei64(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vfloat16mf2_t * op6, vfloat16mf2_t * op7, vbool32_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, vfloat16mf2_t op11, vfloat16mf2_t op12, vfloat16mf2_t op13, vfloat16mf2_t op14, vfloat16mf2_t op15, vfloat16mf2_t op16, const _Float16 * op17, vuint64m2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_f16mf4))) +void vloxseg8ei64(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vfloat16mf4_t * op6, vfloat16mf4_t * op7, const _Float16 * op8, vuint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vloxseg8ei64_v_f16mf4_m))) +void vloxseg8ei64(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vfloat16mf4_t * op6, vfloat16mf4_t * op7, vbool64_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, vfloat16mf4_t op11, vfloat16mf4_t op12, vfloat16mf4_t op13, vfloat16mf4_t op14, vfloat16mf4_t op15, vfloat16mf4_t op16, const _Float16 * op17, vuint64m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e16_v_f16m1))) +void vsseg2e16(_Float16 * op0, vfloat16m1_t op1, vfloat16m1_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e16_v_f16m1_m))) +void vsseg2e16(vbool16_t op0, _Float16 * op1, vfloat16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e16_v_f16m2))) +void vsseg2e16(_Float16 * op0, vfloat16m2_t op1, vfloat16m2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e16_v_f16m2_m))) +void vsseg2e16(vbool8_t op0, _Float16 * op1, vfloat16m2_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e16_v_f16m4))) +void vsseg2e16(_Float16 * op0, vfloat16m4_t op1, vfloat16m4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e16_v_f16m4_m))) +void vsseg2e16(vbool4_t op0, _Float16 * op1, vfloat16m4_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e16_v_f16mf2))) +void vsseg2e16(_Float16 * op0, vfloat16mf2_t op1, vfloat16mf2_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e16_v_f16mf2_m))) +void vsseg2e16(vbool32_t op0, _Float16 * op1, vfloat16mf2_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e16_v_f16mf4))) +void vsseg2e16(_Float16 * op0, vfloat16mf4_t op1, vfloat16mf4_t op2, size_t op3); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg2e16_v_f16mf4_m))) +void vsseg2e16(vbool64_t op0, _Float16 * op1, vfloat16mf4_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e16_v_f16m1))) +void vsseg3e16(_Float16 * op0, vfloat16m1_t op1, vfloat16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e16_v_f16m1_m))) +void vsseg3e16(vbool16_t op0, _Float16 * op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e16_v_f16m2))) +void vsseg3e16(_Float16 * op0, vfloat16m2_t op1, vfloat16m2_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e16_v_f16m2_m))) +void vsseg3e16(vbool8_t op0, _Float16 * op1, vfloat16m2_t op2, vfloat16m2_t op3, vfloat16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e16_v_f16mf2))) +void vsseg3e16(_Float16 * op0, vfloat16mf2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e16_v_f16mf2_m))) +void vsseg3e16(vbool32_t op0, _Float16 * op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e16_v_f16mf4))) +void vsseg3e16(_Float16 * op0, vfloat16mf4_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg3e16_v_f16mf4_m))) +void vsseg3e16(vbool64_t op0, _Float16 * op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e16_v_f16m1))) +void vsseg4e16(_Float16 * op0, vfloat16m1_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e16_v_f16m1_m))) +void vsseg4e16(vbool16_t op0, _Float16 * op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e16_v_f16m2))) +void vsseg4e16(_Float16 * op0, vfloat16m2_t op1, vfloat16m2_t op2, vfloat16m2_t op3, vfloat16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e16_v_f16m2_m))) +void vsseg4e16(vbool8_t op0, _Float16 * op1, vfloat16m2_t op2, vfloat16m2_t op3, vfloat16m2_t op4, vfloat16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e16_v_f16mf2))) +void vsseg4e16(_Float16 * op0, vfloat16mf2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e16_v_f16mf2_m))) +void vsseg4e16(vbool32_t op0, _Float16 * op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e16_v_f16mf4))) +void vsseg4e16(_Float16 * op0, vfloat16mf4_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg4e16_v_f16mf4_m))) +void vsseg4e16(vbool64_t op0, _Float16 * op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e16_v_f16m1))) +void vsseg5e16(_Float16 * op0, vfloat16m1_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e16_v_f16m1_m))) +void vsseg5e16(vbool16_t op0, _Float16 * op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e16_v_f16mf2))) +void vsseg5e16(_Float16 * op0, vfloat16mf2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e16_v_f16mf2_m))) +void vsseg5e16(vbool32_t op0, _Float16 * op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e16_v_f16mf4))) +void vsseg5e16(_Float16 * op0, vfloat16mf4_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg5e16_v_f16mf4_m))) +void vsseg5e16(vbool64_t op0, _Float16 * op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e16_v_f16m1))) +void vsseg6e16(_Float16 * op0, vfloat16m1_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e16_v_f16m1_m))) +void vsseg6e16(vbool16_t op0, _Float16 * op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e16_v_f16mf2))) +void vsseg6e16(_Float16 * op0, vfloat16mf2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e16_v_f16mf2_m))) +void vsseg6e16(vbool32_t op0, _Float16 * op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e16_v_f16mf4))) +void vsseg6e16(_Float16 * op0, vfloat16mf4_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg6e16_v_f16mf4_m))) +void vsseg6e16(vbool64_t op0, _Float16 * op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e16_v_f16m1))) +void vsseg7e16(_Float16 * op0, vfloat16m1_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e16_v_f16m1_m))) +void vsseg7e16(vbool16_t op0, _Float16 * op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e16_v_f16mf2))) +void vsseg7e16(_Float16 * op0, vfloat16mf2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e16_v_f16mf2_m))) +void vsseg7e16(vbool32_t op0, _Float16 * op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e16_v_f16mf4))) +void vsseg7e16(_Float16 * op0, vfloat16mf4_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg7e16_v_f16mf4_m))) +void vsseg7e16(vbool64_t op0, _Float16 * op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e16_v_f16m1))) +void vsseg8e16(_Float16 * op0, vfloat16m1_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e16_v_f16m1_m))) +void vsseg8e16(vbool16_t op0, _Float16 * op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e16_v_f16mf2))) +void vsseg8e16(_Float16 * op0, vfloat16mf2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e16_v_f16mf2_m))) +void vsseg8e16(vbool32_t op0, _Float16 * op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e16_v_f16mf4))) +void vsseg8e16(_Float16 * op0, vfloat16mf4_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsseg8e16_v_f16mf4_m))) +void vsseg8e16(vbool64_t op0, _Float16 * op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e16_v_f16m1))) +void vssseg2e16(_Float16 * op0, ptrdiff_t op1, vfloat16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e16_v_f16m1_m))) +void vssseg2e16(vbool16_t op0, _Float16 * op1, ptrdiff_t op2, vfloat16m1_t op3, vfloat16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e16_v_f16m2))) +void vssseg2e16(_Float16 * op0, ptrdiff_t op1, vfloat16m2_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e16_v_f16m2_m))) +void vssseg2e16(vbool8_t op0, _Float16 * op1, ptrdiff_t op2, vfloat16m2_t op3, vfloat16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e16_v_f16m4))) +void vssseg2e16(_Float16 * op0, ptrdiff_t op1, vfloat16m4_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e16_v_f16m4_m))) +void vssseg2e16(vbool4_t op0, _Float16 * op1, ptrdiff_t op2, vfloat16m4_t op3, vfloat16m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e16_v_f16mf2))) +void vssseg2e16(_Float16 * op0, ptrdiff_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e16_v_f16mf2_m))) +void vssseg2e16(vbool32_t op0, _Float16 * op1, ptrdiff_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e16_v_f16mf4))) +void vssseg2e16(_Float16 * op0, ptrdiff_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg2e16_v_f16mf4_m))) +void vssseg2e16(vbool64_t op0, _Float16 * op1, ptrdiff_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e16_v_f16m1))) +void vssseg3e16(_Float16 * op0, ptrdiff_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e16_v_f16m1_m))) +void vssseg3e16(vbool16_t op0, _Float16 * op1, ptrdiff_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e16_v_f16m2))) +void vssseg3e16(_Float16 * op0, ptrdiff_t op1, vfloat16m2_t op2, vfloat16m2_t op3, vfloat16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e16_v_f16m2_m))) +void vssseg3e16(vbool8_t op0, _Float16 * op1, ptrdiff_t op2, vfloat16m2_t op3, vfloat16m2_t op4, vfloat16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e16_v_f16mf2))) +void vssseg3e16(_Float16 * op0, ptrdiff_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e16_v_f16mf2_m))) +void vssseg3e16(vbool32_t op0, _Float16 * op1, ptrdiff_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e16_v_f16mf4))) +void vssseg3e16(_Float16 * op0, ptrdiff_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg3e16_v_f16mf4_m))) +void vssseg3e16(vbool64_t op0, _Float16 * op1, ptrdiff_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e16_v_f16m1))) +void vssseg4e16(_Float16 * op0, ptrdiff_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e16_v_f16m1_m))) +void vssseg4e16(vbool16_t op0, _Float16 * op1, ptrdiff_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e16_v_f16m2))) +void vssseg4e16(_Float16 * op0, ptrdiff_t op1, vfloat16m2_t op2, vfloat16m2_t op3, vfloat16m2_t op4, vfloat16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e16_v_f16m2_m))) +void vssseg4e16(vbool8_t op0, _Float16 * op1, ptrdiff_t op2, vfloat16m2_t op3, vfloat16m2_t op4, vfloat16m2_t op5, vfloat16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e16_v_f16mf2))) +void vssseg4e16(_Float16 * op0, ptrdiff_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e16_v_f16mf2_m))) +void vssseg4e16(vbool32_t op0, _Float16 * op1, ptrdiff_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e16_v_f16mf4))) +void vssseg4e16(_Float16 * op0, ptrdiff_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg4e16_v_f16mf4_m))) +void vssseg4e16(vbool64_t op0, _Float16 * op1, ptrdiff_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e16_v_f16m1))) +void vssseg5e16(_Float16 * op0, ptrdiff_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e16_v_f16m1_m))) +void vssseg5e16(vbool16_t op0, _Float16 * op1, ptrdiff_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e16_v_f16mf2))) +void vssseg5e16(_Float16 * op0, ptrdiff_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e16_v_f16mf2_m))) +void vssseg5e16(vbool32_t op0, _Float16 * op1, ptrdiff_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e16_v_f16mf4))) +void vssseg5e16(_Float16 * op0, ptrdiff_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg5e16_v_f16mf4_m))) +void vssseg5e16(vbool64_t op0, _Float16 * op1, ptrdiff_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e16_v_f16m1))) +void vssseg6e16(_Float16 * op0, ptrdiff_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e16_v_f16m1_m))) +void vssseg6e16(vbool16_t op0, _Float16 * op1, ptrdiff_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e16_v_f16mf2))) +void vssseg6e16(_Float16 * op0, ptrdiff_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e16_v_f16mf2_m))) +void vssseg6e16(vbool32_t op0, _Float16 * op1, ptrdiff_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e16_v_f16mf4))) +void vssseg6e16(_Float16 * op0, ptrdiff_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg6e16_v_f16mf4_m))) +void vssseg6e16(vbool64_t op0, _Float16 * op1, ptrdiff_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e16_v_f16m1))) +void vssseg7e16(_Float16 * op0, ptrdiff_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e16_v_f16m1_m))) +void vssseg7e16(vbool16_t op0, _Float16 * op1, ptrdiff_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e16_v_f16mf2))) +void vssseg7e16(_Float16 * op0, ptrdiff_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e16_v_f16mf2_m))) +void vssseg7e16(vbool32_t op0, _Float16 * op1, ptrdiff_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e16_v_f16mf4))) +void vssseg7e16(_Float16 * op0, ptrdiff_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg7e16_v_f16mf4_m))) +void vssseg7e16(vbool64_t op0, _Float16 * op1, ptrdiff_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e16_v_f16m1))) +void vssseg8e16(_Float16 * op0, ptrdiff_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e16_v_f16m1_m))) +void vssseg8e16(vbool16_t op0, _Float16 * op1, ptrdiff_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, vfloat16m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e16_v_f16mf2))) +void vssseg8e16(_Float16 * op0, ptrdiff_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e16_v_f16mf2_m))) +void vssseg8e16(vbool32_t op0, _Float16 * op1, ptrdiff_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e16_v_f16mf4))) +void vssseg8e16(_Float16 * op0, ptrdiff_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vssseg8e16_v_f16mf4_m))) +void vssseg8e16(vbool64_t op0, _Float16 * op1, ptrdiff_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_f16m1))) +void vsuxseg2ei8(_Float16 * op0, vuint8mf2_t op1, vfloat16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_f16m1_m))) +void vsuxseg2ei8(vbool16_t op0, _Float16 * op1, vuint8mf2_t op2, vfloat16m1_t op3, vfloat16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_f16m2))) +void vsuxseg2ei8(_Float16 * op0, vuint8m1_t op1, vfloat16m2_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_f16m2_m))) +void vsuxseg2ei8(vbool8_t op0, _Float16 * op1, vuint8m1_t op2, vfloat16m2_t op3, vfloat16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_f16m4))) +void vsuxseg2ei8(_Float16 * op0, vuint8m2_t op1, vfloat16m4_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_f16m4_m))) +void vsuxseg2ei8(vbool4_t op0, _Float16 * op1, vuint8m2_t op2, vfloat16m4_t op3, vfloat16m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_f16mf2))) +void vsuxseg2ei8(_Float16 * op0, vuint8mf4_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_f16mf2_m))) +void vsuxseg2ei8(vbool32_t op0, _Float16 * op1, vuint8mf4_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_f16mf4))) +void vsuxseg2ei8(_Float16 * op0, vuint8mf8_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei8_v_f16mf4_m))) +void vsuxseg2ei8(vbool64_t op0, _Float16 * op1, vuint8mf8_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_f16m1))) +void vsuxseg3ei8(_Float16 * op0, vuint8mf2_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_f16m1_m))) +void vsuxseg3ei8(vbool16_t op0, _Float16 * op1, vuint8mf2_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_f16m2))) +void vsuxseg3ei8(_Float16 * op0, vuint8m1_t op1, vfloat16m2_t op2, vfloat16m2_t op3, vfloat16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_f16m2_m))) +void vsuxseg3ei8(vbool8_t op0, _Float16 * op1, vuint8m1_t op2, vfloat16m2_t op3, vfloat16m2_t op4, vfloat16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_f16mf2))) +void vsuxseg3ei8(_Float16 * op0, vuint8mf4_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_f16mf2_m))) +void vsuxseg3ei8(vbool32_t op0, _Float16 * op1, vuint8mf4_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_f16mf4))) +void vsuxseg3ei8(_Float16 * op0, vuint8mf8_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei8_v_f16mf4_m))) +void vsuxseg3ei8(vbool64_t op0, _Float16 * op1, vuint8mf8_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_f16m1))) +void vsuxseg4ei8(_Float16 * op0, vuint8mf2_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_f16m1_m))) +void vsuxseg4ei8(vbool16_t op0, _Float16 * op1, vuint8mf2_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_f16m2))) +void vsuxseg4ei8(_Float16 * op0, vuint8m1_t op1, vfloat16m2_t op2, vfloat16m2_t op3, vfloat16m2_t op4, vfloat16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_f16m2_m))) +void vsuxseg4ei8(vbool8_t op0, _Float16 * op1, vuint8m1_t op2, vfloat16m2_t op3, vfloat16m2_t op4, vfloat16m2_t op5, vfloat16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_f16mf2))) +void vsuxseg4ei8(_Float16 * op0, vuint8mf4_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_f16mf2_m))) +void vsuxseg4ei8(vbool32_t op0, _Float16 * op1, vuint8mf4_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_f16mf4))) +void vsuxseg4ei8(_Float16 * op0, vuint8mf8_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei8_v_f16mf4_m))) +void vsuxseg4ei8(vbool64_t op0, _Float16 * op1, vuint8mf8_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_f16m1))) +void vsuxseg5ei8(_Float16 * op0, vuint8mf2_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_f16m1_m))) +void vsuxseg5ei8(vbool16_t op0, _Float16 * op1, vuint8mf2_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_f16mf2))) +void vsuxseg5ei8(_Float16 * op0, vuint8mf4_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_f16mf2_m))) +void vsuxseg5ei8(vbool32_t op0, _Float16 * op1, vuint8mf4_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_f16mf4))) +void vsuxseg5ei8(_Float16 * op0, vuint8mf8_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei8_v_f16mf4_m))) +void vsuxseg5ei8(vbool64_t op0, _Float16 * op1, vuint8mf8_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_f16m1))) +void vsuxseg6ei8(_Float16 * op0, vuint8mf2_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_f16m1_m))) +void vsuxseg6ei8(vbool16_t op0, _Float16 * op1, vuint8mf2_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_f16mf2))) +void vsuxseg6ei8(_Float16 * op0, vuint8mf4_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_f16mf2_m))) +void vsuxseg6ei8(vbool32_t op0, _Float16 * op1, vuint8mf4_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_f16mf4))) +void vsuxseg6ei8(_Float16 * op0, vuint8mf8_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei8_v_f16mf4_m))) +void vsuxseg6ei8(vbool64_t op0, _Float16 * op1, vuint8mf8_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_f16m1))) +void vsuxseg7ei8(_Float16 * op0, vuint8mf2_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_f16m1_m))) +void vsuxseg7ei8(vbool16_t op0, _Float16 * op1, vuint8mf2_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_f16mf2))) +void vsuxseg7ei8(_Float16 * op0, vuint8mf4_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_f16mf2_m))) +void vsuxseg7ei8(vbool32_t op0, _Float16 * op1, vuint8mf4_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_f16mf4))) +void vsuxseg7ei8(_Float16 * op0, vuint8mf8_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei8_v_f16mf4_m))) +void vsuxseg7ei8(vbool64_t op0, _Float16 * op1, vuint8mf8_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_f16m1))) +void vsuxseg8ei8(_Float16 * op0, vuint8mf2_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_f16m1_m))) +void vsuxseg8ei8(vbool16_t op0, _Float16 * op1, vuint8mf2_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, vfloat16m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_f16mf2))) +void vsuxseg8ei8(_Float16 * op0, vuint8mf4_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_f16mf2_m))) +void vsuxseg8ei8(vbool32_t op0, _Float16 * op1, vuint8mf4_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_f16mf4))) +void vsuxseg8ei8(_Float16 * op0, vuint8mf8_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei8_v_f16mf4_m))) +void vsuxseg8ei8(vbool64_t op0, _Float16 * op1, vuint8mf8_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_f16m1))) +void vsuxseg2ei16(_Float16 * op0, vuint16m1_t op1, vfloat16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_f16m1_m))) +void vsuxseg2ei16(vbool16_t op0, _Float16 * op1, vuint16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_f16m2))) +void vsuxseg2ei16(_Float16 * op0, vuint16m2_t op1, vfloat16m2_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_f16m2_m))) +void vsuxseg2ei16(vbool8_t op0, _Float16 * op1, vuint16m2_t op2, vfloat16m2_t op3, vfloat16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_f16m4))) +void vsuxseg2ei16(_Float16 * op0, vuint16m4_t op1, vfloat16m4_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_f16m4_m))) +void vsuxseg2ei16(vbool4_t op0, _Float16 * op1, vuint16m4_t op2, vfloat16m4_t op3, vfloat16m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_f16mf2))) +void vsuxseg2ei16(_Float16 * op0, vuint16mf2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_f16mf2_m))) +void vsuxseg2ei16(vbool32_t op0, _Float16 * op1, vuint16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_f16mf4))) +void vsuxseg2ei16(_Float16 * op0, vuint16mf4_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei16_v_f16mf4_m))) +void vsuxseg2ei16(vbool64_t op0, _Float16 * op1, vuint16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_f16m1))) +void vsuxseg3ei16(_Float16 * op0, vuint16m1_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_f16m1_m))) +void vsuxseg3ei16(vbool16_t op0, _Float16 * op1, vuint16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_f16m2))) +void vsuxseg3ei16(_Float16 * op0, vuint16m2_t op1, vfloat16m2_t op2, vfloat16m2_t op3, vfloat16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_f16m2_m))) +void vsuxseg3ei16(vbool8_t op0, _Float16 * op1, vuint16m2_t op2, vfloat16m2_t op3, vfloat16m2_t op4, vfloat16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_f16mf2))) +void vsuxseg3ei16(_Float16 * op0, vuint16mf2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_f16mf2_m))) +void vsuxseg3ei16(vbool32_t op0, _Float16 * op1, vuint16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_f16mf4))) +void vsuxseg3ei16(_Float16 * op0, vuint16mf4_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei16_v_f16mf4_m))) +void vsuxseg3ei16(vbool64_t op0, _Float16 * op1, vuint16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_f16m1))) +void vsuxseg4ei16(_Float16 * op0, vuint16m1_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_f16m1_m))) +void vsuxseg4ei16(vbool16_t op0, _Float16 * op1, vuint16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_f16m2))) +void vsuxseg4ei16(_Float16 * op0, vuint16m2_t op1, vfloat16m2_t op2, vfloat16m2_t op3, vfloat16m2_t op4, vfloat16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_f16m2_m))) +void vsuxseg4ei16(vbool8_t op0, _Float16 * op1, vuint16m2_t op2, vfloat16m2_t op3, vfloat16m2_t op4, vfloat16m2_t op5, vfloat16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_f16mf2))) +void vsuxseg4ei16(_Float16 * op0, vuint16mf2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_f16mf2_m))) +void vsuxseg4ei16(vbool32_t op0, _Float16 * op1, vuint16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_f16mf4))) +void vsuxseg4ei16(_Float16 * op0, vuint16mf4_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei16_v_f16mf4_m))) +void vsuxseg4ei16(vbool64_t op0, _Float16 * op1, vuint16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_f16m1))) +void vsuxseg5ei16(_Float16 * op0, vuint16m1_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_f16m1_m))) +void vsuxseg5ei16(vbool16_t op0, _Float16 * op1, vuint16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_f16mf2))) +void vsuxseg5ei16(_Float16 * op0, vuint16mf2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_f16mf2_m))) +void vsuxseg5ei16(vbool32_t op0, _Float16 * op1, vuint16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_f16mf4))) +void vsuxseg5ei16(_Float16 * op0, vuint16mf4_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei16_v_f16mf4_m))) +void vsuxseg5ei16(vbool64_t op0, _Float16 * op1, vuint16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_f16m1))) +void vsuxseg6ei16(_Float16 * op0, vuint16m1_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_f16m1_m))) +void vsuxseg6ei16(vbool16_t op0, _Float16 * op1, vuint16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_f16mf2))) +void vsuxseg6ei16(_Float16 * op0, vuint16mf2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_f16mf2_m))) +void vsuxseg6ei16(vbool32_t op0, _Float16 * op1, vuint16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_f16mf4))) +void vsuxseg6ei16(_Float16 * op0, vuint16mf4_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei16_v_f16mf4_m))) +void vsuxseg6ei16(vbool64_t op0, _Float16 * op1, vuint16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_f16m1))) +void vsuxseg7ei16(_Float16 * op0, vuint16m1_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_f16m1_m))) +void vsuxseg7ei16(vbool16_t op0, _Float16 * op1, vuint16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_f16mf2))) +void vsuxseg7ei16(_Float16 * op0, vuint16mf2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_f16mf2_m))) +void vsuxseg7ei16(vbool32_t op0, _Float16 * op1, vuint16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_f16mf4))) +void vsuxseg7ei16(_Float16 * op0, vuint16mf4_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei16_v_f16mf4_m))) +void vsuxseg7ei16(vbool64_t op0, _Float16 * op1, vuint16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_f16m1))) +void vsuxseg8ei16(_Float16 * op0, vuint16m1_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_f16m1_m))) +void vsuxseg8ei16(vbool16_t op0, _Float16 * op1, vuint16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, vfloat16m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_f16mf2))) +void vsuxseg8ei16(_Float16 * op0, vuint16mf2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_f16mf2_m))) +void vsuxseg8ei16(vbool32_t op0, _Float16 * op1, vuint16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_f16mf4))) +void vsuxseg8ei16(_Float16 * op0, vuint16mf4_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei16_v_f16mf4_m))) +void vsuxseg8ei16(vbool64_t op0, _Float16 * op1, vuint16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_f16m1))) +void vsuxseg2ei32(_Float16 * op0, vuint32m2_t op1, vfloat16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_f16m1_m))) +void vsuxseg2ei32(vbool16_t op0, _Float16 * op1, vuint32m2_t op2, vfloat16m1_t op3, vfloat16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_f16m2))) +void vsuxseg2ei32(_Float16 * op0, vuint32m4_t op1, vfloat16m2_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_f16m2_m))) +void vsuxseg2ei32(vbool8_t op0, _Float16 * op1, vuint32m4_t op2, vfloat16m2_t op3, vfloat16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_f16m4))) +void vsuxseg2ei32(_Float16 * op0, vuint32m8_t op1, vfloat16m4_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_f16m4_m))) +void vsuxseg2ei32(vbool4_t op0, _Float16 * op1, vuint32m8_t op2, vfloat16m4_t op3, vfloat16m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_f16mf2))) +void vsuxseg2ei32(_Float16 * op0, vuint32m1_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_f16mf2_m))) +void vsuxseg2ei32(vbool32_t op0, _Float16 * op1, vuint32m1_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_f16mf4))) +void vsuxseg2ei32(_Float16 * op0, vuint32mf2_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei32_v_f16mf4_m))) +void vsuxseg2ei32(vbool64_t op0, _Float16 * op1, vuint32mf2_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_f16m1))) +void vsuxseg3ei32(_Float16 * op0, vuint32m2_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_f16m1_m))) +void vsuxseg3ei32(vbool16_t op0, _Float16 * op1, vuint32m2_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_f16m2))) +void vsuxseg3ei32(_Float16 * op0, vuint32m4_t op1, vfloat16m2_t op2, vfloat16m2_t op3, vfloat16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_f16m2_m))) +void vsuxseg3ei32(vbool8_t op0, _Float16 * op1, vuint32m4_t op2, vfloat16m2_t op3, vfloat16m2_t op4, vfloat16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_f16mf2))) +void vsuxseg3ei32(_Float16 * op0, vuint32m1_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_f16mf2_m))) +void vsuxseg3ei32(vbool32_t op0, _Float16 * op1, vuint32m1_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_f16mf4))) +void vsuxseg3ei32(_Float16 * op0, vuint32mf2_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei32_v_f16mf4_m))) +void vsuxseg3ei32(vbool64_t op0, _Float16 * op1, vuint32mf2_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_f16m1))) +void vsuxseg4ei32(_Float16 * op0, vuint32m2_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_f16m1_m))) +void vsuxseg4ei32(vbool16_t op0, _Float16 * op1, vuint32m2_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_f16m2))) +void vsuxseg4ei32(_Float16 * op0, vuint32m4_t op1, vfloat16m2_t op2, vfloat16m2_t op3, vfloat16m2_t op4, vfloat16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_f16m2_m))) +void vsuxseg4ei32(vbool8_t op0, _Float16 * op1, vuint32m4_t op2, vfloat16m2_t op3, vfloat16m2_t op4, vfloat16m2_t op5, vfloat16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_f16mf2))) +void vsuxseg4ei32(_Float16 * op0, vuint32m1_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_f16mf2_m))) +void vsuxseg4ei32(vbool32_t op0, _Float16 * op1, vuint32m1_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_f16mf4))) +void vsuxseg4ei32(_Float16 * op0, vuint32mf2_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei32_v_f16mf4_m))) +void vsuxseg4ei32(vbool64_t op0, _Float16 * op1, vuint32mf2_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_f16m1))) +void vsuxseg5ei32(_Float16 * op0, vuint32m2_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_f16m1_m))) +void vsuxseg5ei32(vbool16_t op0, _Float16 * op1, vuint32m2_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_f16mf2))) +void vsuxseg5ei32(_Float16 * op0, vuint32m1_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_f16mf2_m))) +void vsuxseg5ei32(vbool32_t op0, _Float16 * op1, vuint32m1_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_f16mf4))) +void vsuxseg5ei32(_Float16 * op0, vuint32mf2_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei32_v_f16mf4_m))) +void vsuxseg5ei32(vbool64_t op0, _Float16 * op1, vuint32mf2_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_f16m1))) +void vsuxseg6ei32(_Float16 * op0, vuint32m2_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_f16m1_m))) +void vsuxseg6ei32(vbool16_t op0, _Float16 * op1, vuint32m2_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_f16mf2))) +void vsuxseg6ei32(_Float16 * op0, vuint32m1_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_f16mf2_m))) +void vsuxseg6ei32(vbool32_t op0, _Float16 * op1, vuint32m1_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_f16mf4))) +void vsuxseg6ei32(_Float16 * op0, vuint32mf2_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei32_v_f16mf4_m))) +void vsuxseg6ei32(vbool64_t op0, _Float16 * op1, vuint32mf2_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_f16m1))) +void vsuxseg7ei32(_Float16 * op0, vuint32m2_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_f16m1_m))) +void vsuxseg7ei32(vbool16_t op0, _Float16 * op1, vuint32m2_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_f16mf2))) +void vsuxseg7ei32(_Float16 * op0, vuint32m1_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_f16mf2_m))) +void vsuxseg7ei32(vbool32_t op0, _Float16 * op1, vuint32m1_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_f16mf4))) +void vsuxseg7ei32(_Float16 * op0, vuint32mf2_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei32_v_f16mf4_m))) +void vsuxseg7ei32(vbool64_t op0, _Float16 * op1, vuint32mf2_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_f16m1))) +void vsuxseg8ei32(_Float16 * op0, vuint32m2_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_f16m1_m))) +void vsuxseg8ei32(vbool16_t op0, _Float16 * op1, vuint32m2_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, vfloat16m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_f16mf2))) +void vsuxseg8ei32(_Float16 * op0, vuint32m1_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_f16mf2_m))) +void vsuxseg8ei32(vbool32_t op0, _Float16 * op1, vuint32m1_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_f16mf4))) +void vsuxseg8ei32(_Float16 * op0, vuint32mf2_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei32_v_f16mf4_m))) +void vsuxseg8ei32(vbool64_t op0, _Float16 * op1, vuint32mf2_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_f16m1))) +void vsuxseg2ei64(_Float16 * op0, vuint64m4_t op1, vfloat16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_f16m1_m))) +void vsuxseg2ei64(vbool16_t op0, _Float16 * op1, vuint64m4_t op2, vfloat16m1_t op3, vfloat16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_f16m2))) +void vsuxseg2ei64(_Float16 * op0, vuint64m8_t op1, vfloat16m2_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_f16m2_m))) +void vsuxseg2ei64(vbool8_t op0, _Float16 * op1, vuint64m8_t op2, vfloat16m2_t op3, vfloat16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_f16mf2))) +void vsuxseg2ei64(_Float16 * op0, vuint64m2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_f16mf2_m))) +void vsuxseg2ei64(vbool32_t op0, _Float16 * op1, vuint64m2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_f16mf4))) +void vsuxseg2ei64(_Float16 * op0, vuint64m1_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg2ei64_v_f16mf4_m))) +void vsuxseg2ei64(vbool64_t op0, _Float16 * op1, vuint64m1_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_f16m1))) +void vsuxseg3ei64(_Float16 * op0, vuint64m4_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_f16m1_m))) +void vsuxseg3ei64(vbool16_t op0, _Float16 * op1, vuint64m4_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_f16m2))) +void vsuxseg3ei64(_Float16 * op0, vuint64m8_t op1, vfloat16m2_t op2, vfloat16m2_t op3, vfloat16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_f16m2_m))) +void vsuxseg3ei64(vbool8_t op0, _Float16 * op1, vuint64m8_t op2, vfloat16m2_t op3, vfloat16m2_t op4, vfloat16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_f16mf2))) +void vsuxseg3ei64(_Float16 * op0, vuint64m2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_f16mf2_m))) +void vsuxseg3ei64(vbool32_t op0, _Float16 * op1, vuint64m2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_f16mf4))) +void vsuxseg3ei64(_Float16 * op0, vuint64m1_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg3ei64_v_f16mf4_m))) +void vsuxseg3ei64(vbool64_t op0, _Float16 * op1, vuint64m1_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_f16m1))) +void vsuxseg4ei64(_Float16 * op0, vuint64m4_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_f16m1_m))) +void vsuxseg4ei64(vbool16_t op0, _Float16 * op1, vuint64m4_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_f16m2))) +void vsuxseg4ei64(_Float16 * op0, vuint64m8_t op1, vfloat16m2_t op2, vfloat16m2_t op3, vfloat16m2_t op4, vfloat16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_f16m2_m))) +void vsuxseg4ei64(vbool8_t op0, _Float16 * op1, vuint64m8_t op2, vfloat16m2_t op3, vfloat16m2_t op4, vfloat16m2_t op5, vfloat16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_f16mf2))) +void vsuxseg4ei64(_Float16 * op0, vuint64m2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_f16mf2_m))) +void vsuxseg4ei64(vbool32_t op0, _Float16 * op1, vuint64m2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_f16mf4))) +void vsuxseg4ei64(_Float16 * op0, vuint64m1_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg4ei64_v_f16mf4_m))) +void vsuxseg4ei64(vbool64_t op0, _Float16 * op1, vuint64m1_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_f16m1))) +void vsuxseg5ei64(_Float16 * op0, vuint64m4_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_f16m1_m))) +void vsuxseg5ei64(vbool16_t op0, _Float16 * op1, vuint64m4_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_f16mf2))) +void vsuxseg5ei64(_Float16 * op0, vuint64m2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_f16mf2_m))) +void vsuxseg5ei64(vbool32_t op0, _Float16 * op1, vuint64m2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_f16mf4))) +void vsuxseg5ei64(_Float16 * op0, vuint64m1_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg5ei64_v_f16mf4_m))) +void vsuxseg5ei64(vbool64_t op0, _Float16 * op1, vuint64m1_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_f16m1))) +void vsuxseg6ei64(_Float16 * op0, vuint64m4_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_f16m1_m))) +void vsuxseg6ei64(vbool16_t op0, _Float16 * op1, vuint64m4_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_f16mf2))) +void vsuxseg6ei64(_Float16 * op0, vuint64m2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_f16mf2_m))) +void vsuxseg6ei64(vbool32_t op0, _Float16 * op1, vuint64m2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_f16mf4))) +void vsuxseg6ei64(_Float16 * op0, vuint64m1_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg6ei64_v_f16mf4_m))) +void vsuxseg6ei64(vbool64_t op0, _Float16 * op1, vuint64m1_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_f16m1))) +void vsuxseg7ei64(_Float16 * op0, vuint64m4_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_f16m1_m))) +void vsuxseg7ei64(vbool16_t op0, _Float16 * op1, vuint64m4_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_f16mf2))) +void vsuxseg7ei64(_Float16 * op0, vuint64m2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_f16mf2_m))) +void vsuxseg7ei64(vbool32_t op0, _Float16 * op1, vuint64m2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_f16mf4))) +void vsuxseg7ei64(_Float16 * op0, vuint64m1_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg7ei64_v_f16mf4_m))) +void vsuxseg7ei64(vbool64_t op0, _Float16 * op1, vuint64m1_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_f16m1))) +void vsuxseg8ei64(_Float16 * op0, vuint64m4_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_f16m1_m))) +void vsuxseg8ei64(vbool16_t op0, _Float16 * op1, vuint64m4_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, vfloat16m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_f16mf2))) +void vsuxseg8ei64(_Float16 * op0, vuint64m2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_f16mf2_m))) +void vsuxseg8ei64(vbool32_t op0, _Float16 * op1, vuint64m2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_f16mf4))) +void vsuxseg8ei64(_Float16 * op0, vuint64m1_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsuxseg8ei64_v_f16mf4_m))) +void vsuxseg8ei64(vbool64_t op0, _Float16 * op1, vuint64m1_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_f16m1))) +void vsoxseg2ei8(_Float16 * op0, vuint8mf2_t op1, vfloat16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_f16m1_m))) +void vsoxseg2ei8(vbool16_t op0, _Float16 * op1, vuint8mf2_t op2, vfloat16m1_t op3, vfloat16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_f16m2))) +void vsoxseg2ei8(_Float16 * op0, vuint8m1_t op1, vfloat16m2_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_f16m2_m))) +void vsoxseg2ei8(vbool8_t op0, _Float16 * op1, vuint8m1_t op2, vfloat16m2_t op3, vfloat16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_f16m4))) +void vsoxseg2ei8(_Float16 * op0, vuint8m2_t op1, vfloat16m4_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_f16m4_m))) +void vsoxseg2ei8(vbool4_t op0, _Float16 * op1, vuint8m2_t op2, vfloat16m4_t op3, vfloat16m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_f16mf2))) +void vsoxseg2ei8(_Float16 * op0, vuint8mf4_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_f16mf2_m))) +void vsoxseg2ei8(vbool32_t op0, _Float16 * op1, vuint8mf4_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_f16mf4))) +void vsoxseg2ei8(_Float16 * op0, vuint8mf8_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei8_v_f16mf4_m))) +void vsoxseg2ei8(vbool64_t op0, _Float16 * op1, vuint8mf8_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_f16m1))) +void vsoxseg3ei8(_Float16 * op0, vuint8mf2_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_f16m1_m))) +void vsoxseg3ei8(vbool16_t op0, _Float16 * op1, vuint8mf2_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_f16m2))) +void vsoxseg3ei8(_Float16 * op0, vuint8m1_t op1, vfloat16m2_t op2, vfloat16m2_t op3, vfloat16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_f16m2_m))) +void vsoxseg3ei8(vbool8_t op0, _Float16 * op1, vuint8m1_t op2, vfloat16m2_t op3, vfloat16m2_t op4, vfloat16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_f16mf2))) +void vsoxseg3ei8(_Float16 * op0, vuint8mf4_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_f16mf2_m))) +void vsoxseg3ei8(vbool32_t op0, _Float16 * op1, vuint8mf4_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_f16mf4))) +void vsoxseg3ei8(_Float16 * op0, vuint8mf8_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei8_v_f16mf4_m))) +void vsoxseg3ei8(vbool64_t op0, _Float16 * op1, vuint8mf8_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_f16m1))) +void vsoxseg4ei8(_Float16 * op0, vuint8mf2_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_f16m1_m))) +void vsoxseg4ei8(vbool16_t op0, _Float16 * op1, vuint8mf2_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_f16m2))) +void vsoxseg4ei8(_Float16 * op0, vuint8m1_t op1, vfloat16m2_t op2, vfloat16m2_t op3, vfloat16m2_t op4, vfloat16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_f16m2_m))) +void vsoxseg4ei8(vbool8_t op0, _Float16 * op1, vuint8m1_t op2, vfloat16m2_t op3, vfloat16m2_t op4, vfloat16m2_t op5, vfloat16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_f16mf2))) +void vsoxseg4ei8(_Float16 * op0, vuint8mf4_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_f16mf2_m))) +void vsoxseg4ei8(vbool32_t op0, _Float16 * op1, vuint8mf4_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_f16mf4))) +void vsoxseg4ei8(_Float16 * op0, vuint8mf8_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei8_v_f16mf4_m))) +void vsoxseg4ei8(vbool64_t op0, _Float16 * op1, vuint8mf8_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_f16m1))) +void vsoxseg5ei8(_Float16 * op0, vuint8mf2_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_f16m1_m))) +void vsoxseg5ei8(vbool16_t op0, _Float16 * op1, vuint8mf2_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_f16mf2))) +void vsoxseg5ei8(_Float16 * op0, vuint8mf4_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_f16mf2_m))) +void vsoxseg5ei8(vbool32_t op0, _Float16 * op1, vuint8mf4_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_f16mf4))) +void vsoxseg5ei8(_Float16 * op0, vuint8mf8_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei8_v_f16mf4_m))) +void vsoxseg5ei8(vbool64_t op0, _Float16 * op1, vuint8mf8_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_f16m1))) +void vsoxseg6ei8(_Float16 * op0, vuint8mf2_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_f16m1_m))) +void vsoxseg6ei8(vbool16_t op0, _Float16 * op1, vuint8mf2_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_f16mf2))) +void vsoxseg6ei8(_Float16 * op0, vuint8mf4_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_f16mf2_m))) +void vsoxseg6ei8(vbool32_t op0, _Float16 * op1, vuint8mf4_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_f16mf4))) +void vsoxseg6ei8(_Float16 * op0, vuint8mf8_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei8_v_f16mf4_m))) +void vsoxseg6ei8(vbool64_t op0, _Float16 * op1, vuint8mf8_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_f16m1))) +void vsoxseg7ei8(_Float16 * op0, vuint8mf2_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_f16m1_m))) +void vsoxseg7ei8(vbool16_t op0, _Float16 * op1, vuint8mf2_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_f16mf2))) +void vsoxseg7ei8(_Float16 * op0, vuint8mf4_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_f16mf2_m))) +void vsoxseg7ei8(vbool32_t op0, _Float16 * op1, vuint8mf4_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_f16mf4))) +void vsoxseg7ei8(_Float16 * op0, vuint8mf8_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei8_v_f16mf4_m))) +void vsoxseg7ei8(vbool64_t op0, _Float16 * op1, vuint8mf8_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_f16m1))) +void vsoxseg8ei8(_Float16 * op0, vuint8mf2_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_f16m1_m))) +void vsoxseg8ei8(vbool16_t op0, _Float16 * op1, vuint8mf2_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, vfloat16m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_f16mf2))) +void vsoxseg8ei8(_Float16 * op0, vuint8mf4_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_f16mf2_m))) +void vsoxseg8ei8(vbool32_t op0, _Float16 * op1, vuint8mf4_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_f16mf4))) +void vsoxseg8ei8(_Float16 * op0, vuint8mf8_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei8_v_f16mf4_m))) +void vsoxseg8ei8(vbool64_t op0, _Float16 * op1, vuint8mf8_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_f16m1))) +void vsoxseg2ei16(_Float16 * op0, vuint16m1_t op1, vfloat16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_f16m1_m))) +void vsoxseg2ei16(vbool16_t op0, _Float16 * op1, vuint16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_f16m2))) +void vsoxseg2ei16(_Float16 * op0, vuint16m2_t op1, vfloat16m2_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_f16m2_m))) +void vsoxseg2ei16(vbool8_t op0, _Float16 * op1, vuint16m2_t op2, vfloat16m2_t op3, vfloat16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_f16m4))) +void vsoxseg2ei16(_Float16 * op0, vuint16m4_t op1, vfloat16m4_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_f16m4_m))) +void vsoxseg2ei16(vbool4_t op0, _Float16 * op1, vuint16m4_t op2, vfloat16m4_t op3, vfloat16m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_f16mf2))) +void vsoxseg2ei16(_Float16 * op0, vuint16mf2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_f16mf2_m))) +void vsoxseg2ei16(vbool32_t op0, _Float16 * op1, vuint16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_f16mf4))) +void vsoxseg2ei16(_Float16 * op0, vuint16mf4_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei16_v_f16mf4_m))) +void vsoxseg2ei16(vbool64_t op0, _Float16 * op1, vuint16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_f16m1))) +void vsoxseg3ei16(_Float16 * op0, vuint16m1_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_f16m1_m))) +void vsoxseg3ei16(vbool16_t op0, _Float16 * op1, vuint16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_f16m2))) +void vsoxseg3ei16(_Float16 * op0, vuint16m2_t op1, vfloat16m2_t op2, vfloat16m2_t op3, vfloat16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_f16m2_m))) +void vsoxseg3ei16(vbool8_t op0, _Float16 * op1, vuint16m2_t op2, vfloat16m2_t op3, vfloat16m2_t op4, vfloat16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_f16mf2))) +void vsoxseg3ei16(_Float16 * op0, vuint16mf2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_f16mf2_m))) +void vsoxseg3ei16(vbool32_t op0, _Float16 * op1, vuint16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_f16mf4))) +void vsoxseg3ei16(_Float16 * op0, vuint16mf4_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei16_v_f16mf4_m))) +void vsoxseg3ei16(vbool64_t op0, _Float16 * op1, vuint16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_f16m1))) +void vsoxseg4ei16(_Float16 * op0, vuint16m1_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_f16m1_m))) +void vsoxseg4ei16(vbool16_t op0, _Float16 * op1, vuint16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_f16m2))) +void vsoxseg4ei16(_Float16 * op0, vuint16m2_t op1, vfloat16m2_t op2, vfloat16m2_t op3, vfloat16m2_t op4, vfloat16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_f16m2_m))) +void vsoxseg4ei16(vbool8_t op0, _Float16 * op1, vuint16m2_t op2, vfloat16m2_t op3, vfloat16m2_t op4, vfloat16m2_t op5, vfloat16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_f16mf2))) +void vsoxseg4ei16(_Float16 * op0, vuint16mf2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_f16mf2_m))) +void vsoxseg4ei16(vbool32_t op0, _Float16 * op1, vuint16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_f16mf4))) +void vsoxseg4ei16(_Float16 * op0, vuint16mf4_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei16_v_f16mf4_m))) +void vsoxseg4ei16(vbool64_t op0, _Float16 * op1, vuint16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_f16m1))) +void vsoxseg5ei16(_Float16 * op0, vuint16m1_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_f16m1_m))) +void vsoxseg5ei16(vbool16_t op0, _Float16 * op1, vuint16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_f16mf2))) +void vsoxseg5ei16(_Float16 * op0, vuint16mf2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_f16mf2_m))) +void vsoxseg5ei16(vbool32_t op0, _Float16 * op1, vuint16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_f16mf4))) +void vsoxseg5ei16(_Float16 * op0, vuint16mf4_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei16_v_f16mf4_m))) +void vsoxseg5ei16(vbool64_t op0, _Float16 * op1, vuint16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_f16m1))) +void vsoxseg6ei16(_Float16 * op0, vuint16m1_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_f16m1_m))) +void vsoxseg6ei16(vbool16_t op0, _Float16 * op1, vuint16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_f16mf2))) +void vsoxseg6ei16(_Float16 * op0, vuint16mf2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_f16mf2_m))) +void vsoxseg6ei16(vbool32_t op0, _Float16 * op1, vuint16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_f16mf4))) +void vsoxseg6ei16(_Float16 * op0, vuint16mf4_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei16_v_f16mf4_m))) +void vsoxseg6ei16(vbool64_t op0, _Float16 * op1, vuint16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_f16m1))) +void vsoxseg7ei16(_Float16 * op0, vuint16m1_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_f16m1_m))) +void vsoxseg7ei16(vbool16_t op0, _Float16 * op1, vuint16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_f16mf2))) +void vsoxseg7ei16(_Float16 * op0, vuint16mf2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_f16mf2_m))) +void vsoxseg7ei16(vbool32_t op0, _Float16 * op1, vuint16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_f16mf4))) +void vsoxseg7ei16(_Float16 * op0, vuint16mf4_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei16_v_f16mf4_m))) +void vsoxseg7ei16(vbool64_t op0, _Float16 * op1, vuint16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_f16m1))) +void vsoxseg8ei16(_Float16 * op0, vuint16m1_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_f16m1_m))) +void vsoxseg8ei16(vbool16_t op0, _Float16 * op1, vuint16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, vfloat16m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_f16mf2))) +void vsoxseg8ei16(_Float16 * op0, vuint16mf2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_f16mf2_m))) +void vsoxseg8ei16(vbool32_t op0, _Float16 * op1, vuint16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_f16mf4))) +void vsoxseg8ei16(_Float16 * op0, vuint16mf4_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei16_v_f16mf4_m))) +void vsoxseg8ei16(vbool64_t op0, _Float16 * op1, vuint16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_f16m1))) +void vsoxseg2ei32(_Float16 * op0, vuint32m2_t op1, vfloat16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_f16m1_m))) +void vsoxseg2ei32(vbool16_t op0, _Float16 * op1, vuint32m2_t op2, vfloat16m1_t op3, vfloat16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_f16m2))) +void vsoxseg2ei32(_Float16 * op0, vuint32m4_t op1, vfloat16m2_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_f16m2_m))) +void vsoxseg2ei32(vbool8_t op0, _Float16 * op1, vuint32m4_t op2, vfloat16m2_t op3, vfloat16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_f16m4))) +void vsoxseg2ei32(_Float16 * op0, vuint32m8_t op1, vfloat16m4_t op2, vfloat16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_f16m4_m))) +void vsoxseg2ei32(vbool4_t op0, _Float16 * op1, vuint32m8_t op2, vfloat16m4_t op3, vfloat16m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_f16mf2))) +void vsoxseg2ei32(_Float16 * op0, vuint32m1_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_f16mf2_m))) +void vsoxseg2ei32(vbool32_t op0, _Float16 * op1, vuint32m1_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_f16mf4))) +void vsoxseg2ei32(_Float16 * op0, vuint32mf2_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei32_v_f16mf4_m))) +void vsoxseg2ei32(vbool64_t op0, _Float16 * op1, vuint32mf2_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_f16m1))) +void vsoxseg3ei32(_Float16 * op0, vuint32m2_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_f16m1_m))) +void vsoxseg3ei32(vbool16_t op0, _Float16 * op1, vuint32m2_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_f16m2))) +void vsoxseg3ei32(_Float16 * op0, vuint32m4_t op1, vfloat16m2_t op2, vfloat16m2_t op3, vfloat16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_f16m2_m))) +void vsoxseg3ei32(vbool8_t op0, _Float16 * op1, vuint32m4_t op2, vfloat16m2_t op3, vfloat16m2_t op4, vfloat16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_f16mf2))) +void vsoxseg3ei32(_Float16 * op0, vuint32m1_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_f16mf2_m))) +void vsoxseg3ei32(vbool32_t op0, _Float16 * op1, vuint32m1_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_f16mf4))) +void vsoxseg3ei32(_Float16 * op0, vuint32mf2_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei32_v_f16mf4_m))) +void vsoxseg3ei32(vbool64_t op0, _Float16 * op1, vuint32mf2_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_f16m1))) +void vsoxseg4ei32(_Float16 * op0, vuint32m2_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_f16m1_m))) +void vsoxseg4ei32(vbool16_t op0, _Float16 * op1, vuint32m2_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_f16m2))) +void vsoxseg4ei32(_Float16 * op0, vuint32m4_t op1, vfloat16m2_t op2, vfloat16m2_t op3, vfloat16m2_t op4, vfloat16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_f16m2_m))) +void vsoxseg4ei32(vbool8_t op0, _Float16 * op1, vuint32m4_t op2, vfloat16m2_t op3, vfloat16m2_t op4, vfloat16m2_t op5, vfloat16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_f16mf2))) +void vsoxseg4ei32(_Float16 * op0, vuint32m1_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_f16mf2_m))) +void vsoxseg4ei32(vbool32_t op0, _Float16 * op1, vuint32m1_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_f16mf4))) +void vsoxseg4ei32(_Float16 * op0, vuint32mf2_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei32_v_f16mf4_m))) +void vsoxseg4ei32(vbool64_t op0, _Float16 * op1, vuint32mf2_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_f16m1))) +void vsoxseg5ei32(_Float16 * op0, vuint32m2_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_f16m1_m))) +void vsoxseg5ei32(vbool16_t op0, _Float16 * op1, vuint32m2_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_f16mf2))) +void vsoxseg5ei32(_Float16 * op0, vuint32m1_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_f16mf2_m))) +void vsoxseg5ei32(vbool32_t op0, _Float16 * op1, vuint32m1_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_f16mf4))) +void vsoxseg5ei32(_Float16 * op0, vuint32mf2_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei32_v_f16mf4_m))) +void vsoxseg5ei32(vbool64_t op0, _Float16 * op1, vuint32mf2_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_f16m1))) +void vsoxseg6ei32(_Float16 * op0, vuint32m2_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_f16m1_m))) +void vsoxseg6ei32(vbool16_t op0, _Float16 * op1, vuint32m2_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_f16mf2))) +void vsoxseg6ei32(_Float16 * op0, vuint32m1_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_f16mf2_m))) +void vsoxseg6ei32(vbool32_t op0, _Float16 * op1, vuint32m1_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_f16mf4))) +void vsoxseg6ei32(_Float16 * op0, vuint32mf2_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei32_v_f16mf4_m))) +void vsoxseg6ei32(vbool64_t op0, _Float16 * op1, vuint32mf2_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_f16m1))) +void vsoxseg7ei32(_Float16 * op0, vuint32m2_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_f16m1_m))) +void vsoxseg7ei32(vbool16_t op0, _Float16 * op1, vuint32m2_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_f16mf2))) +void vsoxseg7ei32(_Float16 * op0, vuint32m1_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_f16mf2_m))) +void vsoxseg7ei32(vbool32_t op0, _Float16 * op1, vuint32m1_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_f16mf4))) +void vsoxseg7ei32(_Float16 * op0, vuint32mf2_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei32_v_f16mf4_m))) +void vsoxseg7ei32(vbool64_t op0, _Float16 * op1, vuint32mf2_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_f16m1))) +void vsoxseg8ei32(_Float16 * op0, vuint32m2_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_f16m1_m))) +void vsoxseg8ei32(vbool16_t op0, _Float16 * op1, vuint32m2_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, vfloat16m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_f16mf2))) +void vsoxseg8ei32(_Float16 * op0, vuint32m1_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_f16mf2_m))) +void vsoxseg8ei32(vbool32_t op0, _Float16 * op1, vuint32m1_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_f16mf4))) +void vsoxseg8ei32(_Float16 * op0, vuint32mf2_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei32_v_f16mf4_m))) +void vsoxseg8ei32(vbool64_t op0, _Float16 * op1, vuint32mf2_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_f16m1))) +void vsoxseg2ei64(_Float16 * op0, vuint64m4_t op1, vfloat16m1_t op2, vfloat16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_f16m1_m))) +void vsoxseg2ei64(vbool16_t op0, _Float16 * op1, vuint64m4_t op2, vfloat16m1_t op3, vfloat16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_f16m2))) +void vsoxseg2ei64(_Float16 * op0, vuint64m8_t op1, vfloat16m2_t op2, vfloat16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_f16m2_m))) +void vsoxseg2ei64(vbool8_t op0, _Float16 * op1, vuint64m8_t op2, vfloat16m2_t op3, vfloat16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_f16mf2))) +void vsoxseg2ei64(_Float16 * op0, vuint64m2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_f16mf2_m))) +void vsoxseg2ei64(vbool32_t op0, _Float16 * op1, vuint64m2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_f16mf4))) +void vsoxseg2ei64(_Float16 * op0, vuint64m1_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg2ei64_v_f16mf4_m))) +void vsoxseg2ei64(vbool64_t op0, _Float16 * op1, vuint64m1_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_f16m1))) +void vsoxseg3ei64(_Float16 * op0, vuint64m4_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_f16m1_m))) +void vsoxseg3ei64(vbool16_t op0, _Float16 * op1, vuint64m4_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_f16m2))) +void vsoxseg3ei64(_Float16 * op0, vuint64m8_t op1, vfloat16m2_t op2, vfloat16m2_t op3, vfloat16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_f16m2_m))) +void vsoxseg3ei64(vbool8_t op0, _Float16 * op1, vuint64m8_t op2, vfloat16m2_t op3, vfloat16m2_t op4, vfloat16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_f16mf2))) +void vsoxseg3ei64(_Float16 * op0, vuint64m2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_f16mf2_m))) +void vsoxseg3ei64(vbool32_t op0, _Float16 * op1, vuint64m2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_f16mf4))) +void vsoxseg3ei64(_Float16 * op0, vuint64m1_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg3ei64_v_f16mf4_m))) +void vsoxseg3ei64(vbool64_t op0, _Float16 * op1, vuint64m1_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_f16m1))) +void vsoxseg4ei64(_Float16 * op0, vuint64m4_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_f16m1_m))) +void vsoxseg4ei64(vbool16_t op0, _Float16 * op1, vuint64m4_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_f16m2))) +void vsoxseg4ei64(_Float16 * op0, vuint64m8_t op1, vfloat16m2_t op2, vfloat16m2_t op3, vfloat16m2_t op4, vfloat16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_f16m2_m))) +void vsoxseg4ei64(vbool8_t op0, _Float16 * op1, vuint64m8_t op2, vfloat16m2_t op3, vfloat16m2_t op4, vfloat16m2_t op5, vfloat16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_f16mf2))) +void vsoxseg4ei64(_Float16 * op0, vuint64m2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_f16mf2_m))) +void vsoxseg4ei64(vbool32_t op0, _Float16 * op1, vuint64m2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_f16mf4))) +void vsoxseg4ei64(_Float16 * op0, vuint64m1_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg4ei64_v_f16mf4_m))) +void vsoxseg4ei64(vbool64_t op0, _Float16 * op1, vuint64m1_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_f16m1))) +void vsoxseg5ei64(_Float16 * op0, vuint64m4_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_f16m1_m))) +void vsoxseg5ei64(vbool16_t op0, _Float16 * op1, vuint64m4_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_f16mf2))) +void vsoxseg5ei64(_Float16 * op0, vuint64m2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_f16mf2_m))) +void vsoxseg5ei64(vbool32_t op0, _Float16 * op1, vuint64m2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_f16mf4))) +void vsoxseg5ei64(_Float16 * op0, vuint64m1_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg5ei64_v_f16mf4_m))) +void vsoxseg5ei64(vbool64_t op0, _Float16 * op1, vuint64m1_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_f16m1))) +void vsoxseg6ei64(_Float16 * op0, vuint64m4_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_f16m1_m))) +void vsoxseg6ei64(vbool16_t op0, _Float16 * op1, vuint64m4_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_f16mf2))) +void vsoxseg6ei64(_Float16 * op0, vuint64m2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_f16mf2_m))) +void vsoxseg6ei64(vbool32_t op0, _Float16 * op1, vuint64m2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_f16mf4))) +void vsoxseg6ei64(_Float16 * op0, vuint64m1_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg6ei64_v_f16mf4_m))) +void vsoxseg6ei64(vbool64_t op0, _Float16 * op1, vuint64m1_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_f16m1))) +void vsoxseg7ei64(_Float16 * op0, vuint64m4_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_f16m1_m))) +void vsoxseg7ei64(vbool16_t op0, _Float16 * op1, vuint64m4_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_f16mf2))) +void vsoxseg7ei64(_Float16 * op0, vuint64m2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_f16mf2_m))) +void vsoxseg7ei64(vbool32_t op0, _Float16 * op1, vuint64m2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_f16mf4))) +void vsoxseg7ei64(_Float16 * op0, vuint64m1_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg7ei64_v_f16mf4_m))) +void vsoxseg7ei64(vbool64_t op0, _Float16 * op1, vuint64m1_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_f16m1))) +void vsoxseg8ei64(_Float16 * op0, vuint64m4_t op1, vfloat16m1_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_f16m1_m))) +void vsoxseg8ei64(vbool16_t op0, _Float16 * op1, vuint64m4_t op2, vfloat16m1_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, vfloat16m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_f16mf2))) +void vsoxseg8ei64(_Float16 * op0, vuint64m2_t op1, vfloat16mf2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_f16mf2_m))) +void vsoxseg8ei64(vbool32_t op0, _Float16 * op1, vuint64m2_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_f16mf4))) +void vsoxseg8ei64(_Float16 * op0, vuint64m1_t op1, vfloat16mf4_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vsoxseg8ei64_v_f16mf4_m))) +void vsoxseg8ei64(vbool64_t op0, _Float16 * op1, vuint64m1_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e16_v_f16m1_m))) +void vlseg2e16(vfloat16m1_t * op0, vfloat16m1_t * op1, vbool16_t op2, vfloat16m1_t op3, vfloat16m1_t op4, const _Float16 * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e16_v_f16m2_m))) +void vlseg2e16(vfloat16m2_t * op0, vfloat16m2_t * op1, vbool8_t op2, vfloat16m2_t op3, vfloat16m2_t op4, const _Float16 * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e16_v_f16m4_m))) +void vlseg2e16(vfloat16m4_t * op0, vfloat16m4_t * op1, vbool4_t op2, vfloat16m4_t op3, vfloat16m4_t op4, const _Float16 * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e16_v_f16mf2_m))) +void vlseg2e16(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vbool32_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, const _Float16 * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e16_v_f16mf4_m))) +void vlseg2e16(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vbool64_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, const _Float16 * op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e16_v_f16m1_m))) +void vlseg3e16(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vbool16_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, const _Float16 * op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e16_v_f16m2_m))) +void vlseg3e16(vfloat16m2_t * op0, vfloat16m2_t * op1, vfloat16m2_t * op2, vbool8_t op3, vfloat16m2_t op4, vfloat16m2_t op5, vfloat16m2_t op6, const _Float16 * op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e16_v_f16mf2_m))) +void vlseg3e16(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vbool32_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, const _Float16 * op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e16_v_f16mf4_m))) +void vlseg3e16(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vbool64_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, const _Float16 * op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e16_v_f16m1_m))) +void vlseg4e16(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vbool16_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, const _Float16 * op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e16_v_f16m2_m))) +void vlseg4e16(vfloat16m2_t * op0, vfloat16m2_t * op1, vfloat16m2_t * op2, vfloat16m2_t * op3, vbool8_t op4, vfloat16m2_t op5, vfloat16m2_t op6, vfloat16m2_t op7, vfloat16m2_t op8, const _Float16 * op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e16_v_f16mf2_m))) +void vlseg4e16(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vbool32_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, const _Float16 * op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e16_v_f16mf4_m))) +void vlseg4e16(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vbool64_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, const _Float16 * op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e16_v_f16m1_m))) +void vlseg5e16(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vbool16_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, vfloat16m1_t op10, const _Float16 * op11, size_t op12); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e16_v_f16mf2_m))) +void vlseg5e16(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vbool32_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, const _Float16 * op11, size_t op12); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e16_v_f16mf4_m))) +void vlseg5e16(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vbool64_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, const _Float16 * op11, size_t op12); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e16_v_f16m1_m))) +void vlseg6e16(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vbool16_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, vfloat16m1_t op10, vfloat16m1_t op11, vfloat16m1_t op12, const _Float16 * op13, size_t op14); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e16_v_f16mf2_m))) +void vlseg6e16(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vbool32_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, vfloat16mf2_t op11, vfloat16mf2_t op12, const _Float16 * op13, size_t op14); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e16_v_f16mf4_m))) +void vlseg6e16(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vbool64_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, vfloat16mf4_t op11, vfloat16mf4_t op12, const _Float16 * op13, size_t op14); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e16_v_f16m1_m))) +void vlseg7e16(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vfloat16m1_t * op6, vbool16_t op7, vfloat16m1_t op8, vfloat16m1_t op9, vfloat16m1_t op10, vfloat16m1_t op11, vfloat16m1_t op12, vfloat16m1_t op13, vfloat16m1_t op14, const _Float16 * op15, size_t op16); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e16_v_f16mf2_m))) +void vlseg7e16(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vfloat16mf2_t * op6, vbool32_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, vfloat16mf2_t op11, vfloat16mf2_t op12, vfloat16mf2_t op13, vfloat16mf2_t op14, const _Float16 * op15, size_t op16); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e16_v_f16mf4_m))) +void vlseg7e16(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vfloat16mf4_t * op6, vbool64_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, vfloat16mf4_t op11, vfloat16mf4_t op12, vfloat16mf4_t op13, vfloat16mf4_t op14, const _Float16 * op15, size_t op16); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e16_v_f16m1_m))) +void vlseg8e16(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vfloat16m1_t * op6, vfloat16m1_t * op7, vbool16_t op8, vfloat16m1_t op9, vfloat16m1_t op10, vfloat16m1_t op11, vfloat16m1_t op12, vfloat16m1_t op13, vfloat16m1_t op14, vfloat16m1_t op15, vfloat16m1_t op16, const _Float16 * op17, size_t op18); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e16_v_f16mf2_m))) +void vlseg8e16(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vfloat16mf2_t * op6, vfloat16mf2_t * op7, vbool32_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, vfloat16mf2_t op11, vfloat16mf2_t op12, vfloat16mf2_t op13, vfloat16mf2_t op14, vfloat16mf2_t op15, vfloat16mf2_t op16, const _Float16 * op17, size_t op18); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e16_v_f16mf4_m))) +void vlseg8e16(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vfloat16mf4_t * op6, vfloat16mf4_t * op7, vbool64_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, vfloat16mf4_t op11, vfloat16mf4_t op12, vfloat16mf4_t op13, vfloat16mf4_t op14, vfloat16mf4_t op15, vfloat16mf4_t op16, const _Float16 * op17, size_t op18); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e16ff_v_f16m1_m))) +void vlseg2e16ff(vfloat16m1_t * op0, vfloat16m1_t * op1, vbool16_t op2, vfloat16m1_t op3, vfloat16m1_t op4, const _Float16 * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e16ff_v_f16m2_m))) +void vlseg2e16ff(vfloat16m2_t * op0, vfloat16m2_t * op1, vbool8_t op2, vfloat16m2_t op3, vfloat16m2_t op4, const _Float16 * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e16ff_v_f16m4_m))) +void vlseg2e16ff(vfloat16m4_t * op0, vfloat16m4_t * op1, vbool4_t op2, vfloat16m4_t op3, vfloat16m4_t op4, const _Float16 * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e16ff_v_f16mf2_m))) +void vlseg2e16ff(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vbool32_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, const _Float16 * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg2e16ff_v_f16mf4_m))) +void vlseg2e16ff(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vbool64_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, const _Float16 * op5, size_t * op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e16ff_v_f16m1_m))) +void vlseg3e16ff(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vbool16_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, const _Float16 * op7, size_t * op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e16ff_v_f16m2_m))) +void vlseg3e16ff(vfloat16m2_t * op0, vfloat16m2_t * op1, vfloat16m2_t * op2, vbool8_t op3, vfloat16m2_t op4, vfloat16m2_t op5, vfloat16m2_t op6, const _Float16 * op7, size_t * op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e16ff_v_f16mf2_m))) +void vlseg3e16ff(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vbool32_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, const _Float16 * op7, size_t * op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg3e16ff_v_f16mf4_m))) +void vlseg3e16ff(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vbool64_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, const _Float16 * op7, size_t * op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e16ff_v_f16m1_m))) +void vlseg4e16ff(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vbool16_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, const _Float16 * op9, size_t * op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e16ff_v_f16m2_m))) +void vlseg4e16ff(vfloat16m2_t * op0, vfloat16m2_t * op1, vfloat16m2_t * op2, vfloat16m2_t * op3, vbool8_t op4, vfloat16m2_t op5, vfloat16m2_t op6, vfloat16m2_t op7, vfloat16m2_t op8, const _Float16 * op9, size_t * op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e16ff_v_f16mf2_m))) +void vlseg4e16ff(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vbool32_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, const _Float16 * op9, size_t * op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg4e16ff_v_f16mf4_m))) +void vlseg4e16ff(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vbool64_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, const _Float16 * op9, size_t * op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e16ff_v_f16m1_m))) +void vlseg5e16ff(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vbool16_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, vfloat16m1_t op10, const _Float16 * op11, size_t * op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e16ff_v_f16mf2_m))) +void vlseg5e16ff(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vbool32_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, const _Float16 * op11, size_t * op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg5e16ff_v_f16mf4_m))) +void vlseg5e16ff(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vbool64_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, const _Float16 * op11, size_t * op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e16ff_v_f16m1_m))) +void vlseg6e16ff(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vbool16_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, vfloat16m1_t op10, vfloat16m1_t op11, vfloat16m1_t op12, const _Float16 * op13, size_t * op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e16ff_v_f16mf2_m))) +void vlseg6e16ff(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vbool32_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, vfloat16mf2_t op11, vfloat16mf2_t op12, const _Float16 * op13, size_t * op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg6e16ff_v_f16mf4_m))) +void vlseg6e16ff(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vbool64_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, vfloat16mf4_t op11, vfloat16mf4_t op12, const _Float16 * op13, size_t * op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e16ff_v_f16m1_m))) +void vlseg7e16ff(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vfloat16m1_t * op6, vbool16_t op7, vfloat16m1_t op8, vfloat16m1_t op9, vfloat16m1_t op10, vfloat16m1_t op11, vfloat16m1_t op12, vfloat16m1_t op13, vfloat16m1_t op14, const _Float16 * op15, size_t * op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e16ff_v_f16mf2_m))) +void vlseg7e16ff(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vfloat16mf2_t * op6, vbool32_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, vfloat16mf2_t op11, vfloat16mf2_t op12, vfloat16mf2_t op13, vfloat16mf2_t op14, const _Float16 * op15, size_t * op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg7e16ff_v_f16mf4_m))) +void vlseg7e16ff(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vfloat16mf4_t * op6, vbool64_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, vfloat16mf4_t op11, vfloat16mf4_t op12, vfloat16mf4_t op13, vfloat16mf4_t op14, const _Float16 * op15, size_t * op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e16ff_v_f16m1_m))) +void vlseg8e16ff(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vfloat16m1_t * op6, vfloat16m1_t * op7, vbool16_t op8, vfloat16m1_t op9, vfloat16m1_t op10, vfloat16m1_t op11, vfloat16m1_t op12, vfloat16m1_t op13, vfloat16m1_t op14, vfloat16m1_t op15, vfloat16m1_t op16, const _Float16 * op17, size_t * op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e16ff_v_f16mf2_m))) +void vlseg8e16ff(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vfloat16mf2_t * op6, vfloat16mf2_t * op7, vbool32_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, vfloat16mf2_t op11, vfloat16mf2_t op12, vfloat16mf2_t op13, vfloat16mf2_t op14, vfloat16mf2_t op15, vfloat16mf2_t op16, const _Float16 * op17, size_t * op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlseg8e16ff_v_f16mf4_m))) +void vlseg8e16ff(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vfloat16mf4_t * op6, vfloat16mf4_t * op7, vbool64_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, vfloat16mf4_t op11, vfloat16mf4_t op12, vfloat16mf4_t op13, vfloat16mf4_t op14, vfloat16mf4_t op15, vfloat16mf4_t op16, const _Float16 * op17, size_t * op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e16_v_f16m1_m))) +void vlsseg2e16(vfloat16m1_t * op0, vfloat16m1_t * op1, vbool16_t op2, vfloat16m1_t op3, vfloat16m1_t op4, const _Float16 * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e16_v_f16m2_m))) +void vlsseg2e16(vfloat16m2_t * op0, vfloat16m2_t * op1, vbool8_t op2, vfloat16m2_t op3, vfloat16m2_t op4, const _Float16 * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e16_v_f16m4_m))) +void vlsseg2e16(vfloat16m4_t * op0, vfloat16m4_t * op1, vbool4_t op2, vfloat16m4_t op3, vfloat16m4_t op4, const _Float16 * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e16_v_f16mf2_m))) +void vlsseg2e16(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vbool32_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, const _Float16 * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg2e16_v_f16mf4_m))) +void vlsseg2e16(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vbool64_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, const _Float16 * op5, ptrdiff_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg3e16_v_f16m1_m))) +void vlsseg3e16(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vbool16_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, const _Float16 * op7, ptrdiff_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg3e16_v_f16m2_m))) +void vlsseg3e16(vfloat16m2_t * op0, vfloat16m2_t * op1, vfloat16m2_t * op2, vbool8_t op3, vfloat16m2_t op4, vfloat16m2_t op5, vfloat16m2_t op6, const _Float16 * op7, ptrdiff_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg3e16_v_f16mf2_m))) +void vlsseg3e16(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vbool32_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, const _Float16 * op7, ptrdiff_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg3e16_v_f16mf4_m))) +void vlsseg3e16(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vbool64_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, const _Float16 * op7, ptrdiff_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg4e16_v_f16m1_m))) +void vlsseg4e16(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vbool16_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, const _Float16 * op9, ptrdiff_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg4e16_v_f16m2_m))) +void vlsseg4e16(vfloat16m2_t * op0, vfloat16m2_t * op1, vfloat16m2_t * op2, vfloat16m2_t * op3, vbool8_t op4, vfloat16m2_t op5, vfloat16m2_t op6, vfloat16m2_t op7, vfloat16m2_t op8, const _Float16 * op9, ptrdiff_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg4e16_v_f16mf2_m))) +void vlsseg4e16(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vbool32_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, const _Float16 * op9, ptrdiff_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg4e16_v_f16mf4_m))) +void vlsseg4e16(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vbool64_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, const _Float16 * op9, ptrdiff_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg5e16_v_f16m1_m))) +void vlsseg5e16(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vbool16_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, vfloat16m1_t op10, const _Float16 * op11, ptrdiff_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg5e16_v_f16mf2_m))) +void vlsseg5e16(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vbool32_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, const _Float16 * op11, ptrdiff_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg5e16_v_f16mf4_m))) +void vlsseg5e16(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vbool64_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, const _Float16 * op11, ptrdiff_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg6e16_v_f16m1_m))) +void vlsseg6e16(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vbool16_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, vfloat16m1_t op10, vfloat16m1_t op11, vfloat16m1_t op12, const _Float16 * op13, ptrdiff_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg6e16_v_f16mf2_m))) +void vlsseg6e16(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vbool32_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, vfloat16mf2_t op11, vfloat16mf2_t op12, const _Float16 * op13, ptrdiff_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg6e16_v_f16mf4_m))) +void vlsseg6e16(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vbool64_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, vfloat16mf4_t op11, vfloat16mf4_t op12, const _Float16 * op13, ptrdiff_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg7e16_v_f16m1_m))) +void vlsseg7e16(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vfloat16m1_t * op6, vbool16_t op7, vfloat16m1_t op8, vfloat16m1_t op9, vfloat16m1_t op10, vfloat16m1_t op11, vfloat16m1_t op12, vfloat16m1_t op13, vfloat16m1_t op14, const _Float16 * op15, ptrdiff_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg7e16_v_f16mf2_m))) +void vlsseg7e16(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vfloat16mf2_t * op6, vbool32_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, vfloat16mf2_t op11, vfloat16mf2_t op12, vfloat16mf2_t op13, vfloat16mf2_t op14, const _Float16 * op15, ptrdiff_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg7e16_v_f16mf4_m))) +void vlsseg7e16(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vfloat16mf4_t * op6, vbool64_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, vfloat16mf4_t op11, vfloat16mf4_t op12, vfloat16mf4_t op13, vfloat16mf4_t op14, const _Float16 * op15, ptrdiff_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg8e16_v_f16m1_m))) +void vlsseg8e16(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vfloat16m1_t * op6, vfloat16m1_t * op7, vbool16_t op8, vfloat16m1_t op9, vfloat16m1_t op10, vfloat16m1_t op11, vfloat16m1_t op12, vfloat16m1_t op13, vfloat16m1_t op14, vfloat16m1_t op15, vfloat16m1_t op16, const _Float16 * op17, ptrdiff_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg8e16_v_f16mf2_m))) +void vlsseg8e16(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vfloat16mf2_t * op6, vfloat16mf2_t * op7, vbool32_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, vfloat16mf2_t op11, vfloat16mf2_t op12, vfloat16mf2_t op13, vfloat16mf2_t op14, vfloat16mf2_t op15, vfloat16mf2_t op16, const _Float16 * op17, ptrdiff_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vlsseg8e16_v_f16mf4_m))) +void vlsseg8e16(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vfloat16mf4_t * op6, vfloat16mf4_t * op7, vbool64_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, vfloat16mf4_t op11, vfloat16mf4_t op12, vfloat16mf4_t op13, vfloat16mf4_t op14, vfloat16mf4_t op15, vfloat16mf4_t op16, const _Float16 * op17, ptrdiff_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_f16m1))) +void vluxseg2ei8(vfloat16m1_t * op0, vfloat16m1_t * op1, const _Float16 * op2, vuint8mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_f16m1_m))) +void vluxseg2ei8(vfloat16m1_t * op0, vfloat16m1_t * op1, vbool16_t op2, vfloat16m1_t op3, vfloat16m1_t op4, const _Float16 * op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_f16m2))) +void vluxseg2ei8(vfloat16m2_t * op0, vfloat16m2_t * op1, const _Float16 * op2, vuint8m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_f16m2_m))) +void vluxseg2ei8(vfloat16m2_t * op0, vfloat16m2_t * op1, vbool8_t op2, vfloat16m2_t op3, vfloat16m2_t op4, const _Float16 * op5, vuint8m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_f16m4))) +void vluxseg2ei8(vfloat16m4_t * op0, vfloat16m4_t * op1, const _Float16 * op2, vuint8m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_f16m4_m))) +void vluxseg2ei8(vfloat16m4_t * op0, vfloat16m4_t * op1, vbool4_t op2, vfloat16m4_t op3, vfloat16m4_t op4, const _Float16 * op5, vuint8m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_f16mf2))) +void vluxseg2ei8(vfloat16mf2_t * op0, vfloat16mf2_t * op1, const _Float16 * op2, vuint8mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_f16mf2_m))) +void vluxseg2ei8(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vbool32_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, const _Float16 * op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_f16mf4))) +void vluxseg2ei8(vfloat16mf4_t * op0, vfloat16mf4_t * op1, const _Float16 * op2, vuint8mf8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei8_v_f16mf4_m))) +void vluxseg2ei8(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vbool64_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, const _Float16 * op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_f16m1))) +void vluxseg3ei8(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, const _Float16 * op3, vuint8mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_f16m1_m))) +void vluxseg3ei8(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vbool16_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, const _Float16 * op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_f16m2))) +void vluxseg3ei8(vfloat16m2_t * op0, vfloat16m2_t * op1, vfloat16m2_t * op2, const _Float16 * op3, vuint8m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_f16m2_m))) +void vluxseg3ei8(vfloat16m2_t * op0, vfloat16m2_t * op1, vfloat16m2_t * op2, vbool8_t op3, vfloat16m2_t op4, vfloat16m2_t op5, vfloat16m2_t op6, const _Float16 * op7, vuint8m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_f16mf2))) +void vluxseg3ei8(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, const _Float16 * op3, vuint8mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_f16mf2_m))) +void vluxseg3ei8(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vbool32_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, const _Float16 * op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_f16mf4))) +void vluxseg3ei8(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, const _Float16 * op3, vuint8mf8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei8_v_f16mf4_m))) +void vluxseg3ei8(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vbool64_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, const _Float16 * op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_f16m1))) +void vluxseg4ei8(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, const _Float16 * op4, vuint8mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_f16m1_m))) +void vluxseg4ei8(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vbool16_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, const _Float16 * op9, vuint8mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_f16m2))) +void vluxseg4ei8(vfloat16m2_t * op0, vfloat16m2_t * op1, vfloat16m2_t * op2, vfloat16m2_t * op3, const _Float16 * op4, vuint8m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_f16m2_m))) +void vluxseg4ei8(vfloat16m2_t * op0, vfloat16m2_t * op1, vfloat16m2_t * op2, vfloat16m2_t * op3, vbool8_t op4, vfloat16m2_t op5, vfloat16m2_t op6, vfloat16m2_t op7, vfloat16m2_t op8, const _Float16 * op9, vuint8m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_f16mf2))) +void vluxseg4ei8(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, const _Float16 * op4, vuint8mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_f16mf2_m))) +void vluxseg4ei8(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vbool32_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, const _Float16 * op9, vuint8mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_f16mf4))) +void vluxseg4ei8(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, const _Float16 * op4, vuint8mf8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei8_v_f16mf4_m))) +void vluxseg4ei8(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vbool64_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, const _Float16 * op9, vuint8mf8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_f16m1))) +void vluxseg5ei8(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, const _Float16 * op5, vuint8mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_f16m1_m))) +void vluxseg5ei8(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vbool16_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, vfloat16m1_t op10, const _Float16 * op11, vuint8mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_f16mf2))) +void vluxseg5ei8(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, const _Float16 * op5, vuint8mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_f16mf2_m))) +void vluxseg5ei8(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vbool32_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, const _Float16 * op11, vuint8mf4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_f16mf4))) +void vluxseg5ei8(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, const _Float16 * op5, vuint8mf8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei8_v_f16mf4_m))) +void vluxseg5ei8(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vbool64_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, const _Float16 * op11, vuint8mf8_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_f16m1))) +void vluxseg6ei8(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, const _Float16 * op6, vuint8mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_f16m1_m))) +void vluxseg6ei8(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vbool16_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, vfloat16m1_t op10, vfloat16m1_t op11, vfloat16m1_t op12, const _Float16 * op13, vuint8mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_f16mf2))) +void vluxseg6ei8(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, const _Float16 * op6, vuint8mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_f16mf2_m))) +void vluxseg6ei8(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vbool32_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, vfloat16mf2_t op11, vfloat16mf2_t op12, const _Float16 * op13, vuint8mf4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_f16mf4))) +void vluxseg6ei8(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, const _Float16 * op6, vuint8mf8_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei8_v_f16mf4_m))) +void vluxseg6ei8(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vbool64_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, vfloat16mf4_t op11, vfloat16mf4_t op12, const _Float16 * op13, vuint8mf8_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_f16m1))) +void vluxseg7ei8(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vfloat16m1_t * op6, const _Float16 * op7, vuint8mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_f16m1_m))) +void vluxseg7ei8(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vfloat16m1_t * op6, vbool16_t op7, vfloat16m1_t op8, vfloat16m1_t op9, vfloat16m1_t op10, vfloat16m1_t op11, vfloat16m1_t op12, vfloat16m1_t op13, vfloat16m1_t op14, const _Float16 * op15, vuint8mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_f16mf2))) +void vluxseg7ei8(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vfloat16mf2_t * op6, const _Float16 * op7, vuint8mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_f16mf2_m))) +void vluxseg7ei8(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vfloat16mf2_t * op6, vbool32_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, vfloat16mf2_t op11, vfloat16mf2_t op12, vfloat16mf2_t op13, vfloat16mf2_t op14, const _Float16 * op15, vuint8mf4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_f16mf4))) +void vluxseg7ei8(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vfloat16mf4_t * op6, const _Float16 * op7, vuint8mf8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei8_v_f16mf4_m))) +void vluxseg7ei8(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vfloat16mf4_t * op6, vbool64_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, vfloat16mf4_t op11, vfloat16mf4_t op12, vfloat16mf4_t op13, vfloat16mf4_t op14, const _Float16 * op15, vuint8mf8_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_f16m1))) +void vluxseg8ei8(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vfloat16m1_t * op6, vfloat16m1_t * op7, const _Float16 * op8, vuint8mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_f16m1_m))) +void vluxseg8ei8(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vfloat16m1_t * op6, vfloat16m1_t * op7, vbool16_t op8, vfloat16m1_t op9, vfloat16m1_t op10, vfloat16m1_t op11, vfloat16m1_t op12, vfloat16m1_t op13, vfloat16m1_t op14, vfloat16m1_t op15, vfloat16m1_t op16, const _Float16 * op17, vuint8mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_f16mf2))) +void vluxseg8ei8(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vfloat16mf2_t * op6, vfloat16mf2_t * op7, const _Float16 * op8, vuint8mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_f16mf2_m))) +void vluxseg8ei8(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vfloat16mf2_t * op6, vfloat16mf2_t * op7, vbool32_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, vfloat16mf2_t op11, vfloat16mf2_t op12, vfloat16mf2_t op13, vfloat16mf2_t op14, vfloat16mf2_t op15, vfloat16mf2_t op16, const _Float16 * op17, vuint8mf4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_f16mf4))) +void vluxseg8ei8(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vfloat16mf4_t * op6, vfloat16mf4_t * op7, const _Float16 * op8, vuint8mf8_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei8_v_f16mf4_m))) +void vluxseg8ei8(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vfloat16mf4_t * op6, vfloat16mf4_t * op7, vbool64_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, vfloat16mf4_t op11, vfloat16mf4_t op12, vfloat16mf4_t op13, vfloat16mf4_t op14, vfloat16mf4_t op15, vfloat16mf4_t op16, const _Float16 * op17, vuint8mf8_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_f16m1))) +void vluxseg2ei16(vfloat16m1_t * op0, vfloat16m1_t * op1, const _Float16 * op2, vuint16m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_f16m1_m))) +void vluxseg2ei16(vfloat16m1_t * op0, vfloat16m1_t * op1, vbool16_t op2, vfloat16m1_t op3, vfloat16m1_t op4, const _Float16 * op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_f16m2))) +void vluxseg2ei16(vfloat16m2_t * op0, vfloat16m2_t * op1, const _Float16 * op2, vuint16m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_f16m2_m))) +void vluxseg2ei16(vfloat16m2_t * op0, vfloat16m2_t * op1, vbool8_t op2, vfloat16m2_t op3, vfloat16m2_t op4, const _Float16 * op5, vuint16m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_f16m4))) +void vluxseg2ei16(vfloat16m4_t * op0, vfloat16m4_t * op1, const _Float16 * op2, vuint16m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_f16m4_m))) +void vluxseg2ei16(vfloat16m4_t * op0, vfloat16m4_t * op1, vbool4_t op2, vfloat16m4_t op3, vfloat16m4_t op4, const _Float16 * op5, vuint16m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_f16mf2))) +void vluxseg2ei16(vfloat16mf2_t * op0, vfloat16mf2_t * op1, const _Float16 * op2, vuint16mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_f16mf2_m))) +void vluxseg2ei16(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vbool32_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, const _Float16 * op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_f16mf4))) +void vluxseg2ei16(vfloat16mf4_t * op0, vfloat16mf4_t * op1, const _Float16 * op2, vuint16mf4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei16_v_f16mf4_m))) +void vluxseg2ei16(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vbool64_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, const _Float16 * op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_f16m1))) +void vluxseg3ei16(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, const _Float16 * op3, vuint16m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_f16m1_m))) +void vluxseg3ei16(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vbool16_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, const _Float16 * op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_f16m2))) +void vluxseg3ei16(vfloat16m2_t * op0, vfloat16m2_t * op1, vfloat16m2_t * op2, const _Float16 * op3, vuint16m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_f16m2_m))) +void vluxseg3ei16(vfloat16m2_t * op0, vfloat16m2_t * op1, vfloat16m2_t * op2, vbool8_t op3, vfloat16m2_t op4, vfloat16m2_t op5, vfloat16m2_t op6, const _Float16 * op7, vuint16m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_f16mf2))) +void vluxseg3ei16(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, const _Float16 * op3, vuint16mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_f16mf2_m))) +void vluxseg3ei16(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vbool32_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, const _Float16 * op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_f16mf4))) +void vluxseg3ei16(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, const _Float16 * op3, vuint16mf4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei16_v_f16mf4_m))) +void vluxseg3ei16(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vbool64_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, const _Float16 * op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_f16m1))) +void vluxseg4ei16(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, const _Float16 * op4, vuint16m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_f16m1_m))) +void vluxseg4ei16(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vbool16_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, const _Float16 * op9, vuint16m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_f16m2))) +void vluxseg4ei16(vfloat16m2_t * op0, vfloat16m2_t * op1, vfloat16m2_t * op2, vfloat16m2_t * op3, const _Float16 * op4, vuint16m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_f16m2_m))) +void vluxseg4ei16(vfloat16m2_t * op0, vfloat16m2_t * op1, vfloat16m2_t * op2, vfloat16m2_t * op3, vbool8_t op4, vfloat16m2_t op5, vfloat16m2_t op6, vfloat16m2_t op7, vfloat16m2_t op8, const _Float16 * op9, vuint16m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_f16mf2))) +void vluxseg4ei16(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, const _Float16 * op4, vuint16mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_f16mf2_m))) +void vluxseg4ei16(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vbool32_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, const _Float16 * op9, vuint16mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_f16mf4))) +void vluxseg4ei16(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, const _Float16 * op4, vuint16mf4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei16_v_f16mf4_m))) +void vluxseg4ei16(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vbool64_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, const _Float16 * op9, vuint16mf4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_f16m1))) +void vluxseg5ei16(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, const _Float16 * op5, vuint16m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_f16m1_m))) +void vluxseg5ei16(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vbool16_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, vfloat16m1_t op10, const _Float16 * op11, vuint16m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_f16mf2))) +void vluxseg5ei16(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, const _Float16 * op5, vuint16mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_f16mf2_m))) +void vluxseg5ei16(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vbool32_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, const _Float16 * op11, vuint16mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_f16mf4))) +void vluxseg5ei16(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, const _Float16 * op5, vuint16mf4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei16_v_f16mf4_m))) +void vluxseg5ei16(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vbool64_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, const _Float16 * op11, vuint16mf4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_f16m1))) +void vluxseg6ei16(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, const _Float16 * op6, vuint16m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_f16m1_m))) +void vluxseg6ei16(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vbool16_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, vfloat16m1_t op10, vfloat16m1_t op11, vfloat16m1_t op12, const _Float16 * op13, vuint16m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_f16mf2))) +void vluxseg6ei16(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, const _Float16 * op6, vuint16mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_f16mf2_m))) +void vluxseg6ei16(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vbool32_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, vfloat16mf2_t op11, vfloat16mf2_t op12, const _Float16 * op13, vuint16mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_f16mf4))) +void vluxseg6ei16(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, const _Float16 * op6, vuint16mf4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei16_v_f16mf4_m))) +void vluxseg6ei16(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vbool64_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, vfloat16mf4_t op11, vfloat16mf4_t op12, const _Float16 * op13, vuint16mf4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_f16m1))) +void vluxseg7ei16(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vfloat16m1_t * op6, const _Float16 * op7, vuint16m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_f16m1_m))) +void vluxseg7ei16(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vfloat16m1_t * op6, vbool16_t op7, vfloat16m1_t op8, vfloat16m1_t op9, vfloat16m1_t op10, vfloat16m1_t op11, vfloat16m1_t op12, vfloat16m1_t op13, vfloat16m1_t op14, const _Float16 * op15, vuint16m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_f16mf2))) +void vluxseg7ei16(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vfloat16mf2_t * op6, const _Float16 * op7, vuint16mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_f16mf2_m))) +void vluxseg7ei16(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vfloat16mf2_t * op6, vbool32_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, vfloat16mf2_t op11, vfloat16mf2_t op12, vfloat16mf2_t op13, vfloat16mf2_t op14, const _Float16 * op15, vuint16mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_f16mf4))) +void vluxseg7ei16(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vfloat16mf4_t * op6, const _Float16 * op7, vuint16mf4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei16_v_f16mf4_m))) +void vluxseg7ei16(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vfloat16mf4_t * op6, vbool64_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, vfloat16mf4_t op11, vfloat16mf4_t op12, vfloat16mf4_t op13, vfloat16mf4_t op14, const _Float16 * op15, vuint16mf4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_f16m1))) +void vluxseg8ei16(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vfloat16m1_t * op6, vfloat16m1_t * op7, const _Float16 * op8, vuint16m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_f16m1_m))) +void vluxseg8ei16(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vfloat16m1_t * op6, vfloat16m1_t * op7, vbool16_t op8, vfloat16m1_t op9, vfloat16m1_t op10, vfloat16m1_t op11, vfloat16m1_t op12, vfloat16m1_t op13, vfloat16m1_t op14, vfloat16m1_t op15, vfloat16m1_t op16, const _Float16 * op17, vuint16m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_f16mf2))) +void vluxseg8ei16(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vfloat16mf2_t * op6, vfloat16mf2_t * op7, const _Float16 * op8, vuint16mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_f16mf2_m))) +void vluxseg8ei16(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vfloat16mf2_t * op6, vfloat16mf2_t * op7, vbool32_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, vfloat16mf2_t op11, vfloat16mf2_t op12, vfloat16mf2_t op13, vfloat16mf2_t op14, vfloat16mf2_t op15, vfloat16mf2_t op16, const _Float16 * op17, vuint16mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_f16mf4))) +void vluxseg8ei16(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vfloat16mf4_t * op6, vfloat16mf4_t * op7, const _Float16 * op8, vuint16mf4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei16_v_f16mf4_m))) +void vluxseg8ei16(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vfloat16mf4_t * op6, vfloat16mf4_t * op7, vbool64_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, vfloat16mf4_t op11, vfloat16mf4_t op12, vfloat16mf4_t op13, vfloat16mf4_t op14, vfloat16mf4_t op15, vfloat16mf4_t op16, const _Float16 * op17, vuint16mf4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_f16m1))) +void vluxseg2ei32(vfloat16m1_t * op0, vfloat16m1_t * op1, const _Float16 * op2, vuint32m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_f16m1_m))) +void vluxseg2ei32(vfloat16m1_t * op0, vfloat16m1_t * op1, vbool16_t op2, vfloat16m1_t op3, vfloat16m1_t op4, const _Float16 * op5, vuint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_f16m2))) +void vluxseg2ei32(vfloat16m2_t * op0, vfloat16m2_t * op1, const _Float16 * op2, vuint32m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_f16m2_m))) +void vluxseg2ei32(vfloat16m2_t * op0, vfloat16m2_t * op1, vbool8_t op2, vfloat16m2_t op3, vfloat16m2_t op4, const _Float16 * op5, vuint32m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_f16m4))) +void vluxseg2ei32(vfloat16m4_t * op0, vfloat16m4_t * op1, const _Float16 * op2, vuint32m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_f16m4_m))) +void vluxseg2ei32(vfloat16m4_t * op0, vfloat16m4_t * op1, vbool4_t op2, vfloat16m4_t op3, vfloat16m4_t op4, const _Float16 * op5, vuint32m8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_f16mf2))) +void vluxseg2ei32(vfloat16mf2_t * op0, vfloat16mf2_t * op1, const _Float16 * op2, vuint32m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_f16mf2_m))) +void vluxseg2ei32(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vbool32_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, const _Float16 * op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_f16mf4))) +void vluxseg2ei32(vfloat16mf4_t * op0, vfloat16mf4_t * op1, const _Float16 * op2, vuint32mf2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei32_v_f16mf4_m))) +void vluxseg2ei32(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vbool64_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, const _Float16 * op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_f16m1))) +void vluxseg3ei32(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, const _Float16 * op3, vuint32m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_f16m1_m))) +void vluxseg3ei32(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vbool16_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, const _Float16 * op7, vuint32m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_f16m2))) +void vluxseg3ei32(vfloat16m2_t * op0, vfloat16m2_t * op1, vfloat16m2_t * op2, const _Float16 * op3, vuint32m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_f16m2_m))) +void vluxseg3ei32(vfloat16m2_t * op0, vfloat16m2_t * op1, vfloat16m2_t * op2, vbool8_t op3, vfloat16m2_t op4, vfloat16m2_t op5, vfloat16m2_t op6, const _Float16 * op7, vuint32m4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_f16mf2))) +void vluxseg3ei32(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, const _Float16 * op3, vuint32m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_f16mf2_m))) +void vluxseg3ei32(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vbool32_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, const _Float16 * op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_f16mf4))) +void vluxseg3ei32(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, const _Float16 * op3, vuint32mf2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei32_v_f16mf4_m))) +void vluxseg3ei32(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vbool64_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, const _Float16 * op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_f16m1))) +void vluxseg4ei32(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, const _Float16 * op4, vuint32m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_f16m1_m))) +void vluxseg4ei32(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vbool16_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, const _Float16 * op9, vuint32m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_f16m2))) +void vluxseg4ei32(vfloat16m2_t * op0, vfloat16m2_t * op1, vfloat16m2_t * op2, vfloat16m2_t * op3, const _Float16 * op4, vuint32m4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_f16m2_m))) +void vluxseg4ei32(vfloat16m2_t * op0, vfloat16m2_t * op1, vfloat16m2_t * op2, vfloat16m2_t * op3, vbool8_t op4, vfloat16m2_t op5, vfloat16m2_t op6, vfloat16m2_t op7, vfloat16m2_t op8, const _Float16 * op9, vuint32m4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_f16mf2))) +void vluxseg4ei32(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, const _Float16 * op4, vuint32m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_f16mf2_m))) +void vluxseg4ei32(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vbool32_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, const _Float16 * op9, vuint32m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_f16mf4))) +void vluxseg4ei32(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, const _Float16 * op4, vuint32mf2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei32_v_f16mf4_m))) +void vluxseg4ei32(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vbool64_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, const _Float16 * op9, vuint32mf2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_f16m1))) +void vluxseg5ei32(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, const _Float16 * op5, vuint32m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_f16m1_m))) +void vluxseg5ei32(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vbool16_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, vfloat16m1_t op10, const _Float16 * op11, vuint32m2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_f16mf2))) +void vluxseg5ei32(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, const _Float16 * op5, vuint32m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_f16mf2_m))) +void vluxseg5ei32(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vbool32_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, const _Float16 * op11, vuint32m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_f16mf4))) +void vluxseg5ei32(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, const _Float16 * op5, vuint32mf2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei32_v_f16mf4_m))) +void vluxseg5ei32(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vbool64_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, const _Float16 * op11, vuint32mf2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_f16m1))) +void vluxseg6ei32(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, const _Float16 * op6, vuint32m2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_f16m1_m))) +void vluxseg6ei32(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vbool16_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, vfloat16m1_t op10, vfloat16m1_t op11, vfloat16m1_t op12, const _Float16 * op13, vuint32m2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_f16mf2))) +void vluxseg6ei32(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, const _Float16 * op6, vuint32m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_f16mf2_m))) +void vluxseg6ei32(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vbool32_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, vfloat16mf2_t op11, vfloat16mf2_t op12, const _Float16 * op13, vuint32m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_f16mf4))) +void vluxseg6ei32(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, const _Float16 * op6, vuint32mf2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei32_v_f16mf4_m))) +void vluxseg6ei32(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vbool64_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, vfloat16mf4_t op11, vfloat16mf4_t op12, const _Float16 * op13, vuint32mf2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_f16m1))) +void vluxseg7ei32(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vfloat16m1_t * op6, const _Float16 * op7, vuint32m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_f16m1_m))) +void vluxseg7ei32(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vfloat16m1_t * op6, vbool16_t op7, vfloat16m1_t op8, vfloat16m1_t op9, vfloat16m1_t op10, vfloat16m1_t op11, vfloat16m1_t op12, vfloat16m1_t op13, vfloat16m1_t op14, const _Float16 * op15, vuint32m2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_f16mf2))) +void vluxseg7ei32(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vfloat16mf2_t * op6, const _Float16 * op7, vuint32m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_f16mf2_m))) +void vluxseg7ei32(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vfloat16mf2_t * op6, vbool32_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, vfloat16mf2_t op11, vfloat16mf2_t op12, vfloat16mf2_t op13, vfloat16mf2_t op14, const _Float16 * op15, vuint32m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_f16mf4))) +void vluxseg7ei32(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vfloat16mf4_t * op6, const _Float16 * op7, vuint32mf2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei32_v_f16mf4_m))) +void vluxseg7ei32(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vfloat16mf4_t * op6, vbool64_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, vfloat16mf4_t op11, vfloat16mf4_t op12, vfloat16mf4_t op13, vfloat16mf4_t op14, const _Float16 * op15, vuint32mf2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_f16m1))) +void vluxseg8ei32(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vfloat16m1_t * op6, vfloat16m1_t * op7, const _Float16 * op8, vuint32m2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_f16m1_m))) +void vluxseg8ei32(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vfloat16m1_t * op6, vfloat16m1_t * op7, vbool16_t op8, vfloat16m1_t op9, vfloat16m1_t op10, vfloat16m1_t op11, vfloat16m1_t op12, vfloat16m1_t op13, vfloat16m1_t op14, vfloat16m1_t op15, vfloat16m1_t op16, const _Float16 * op17, vuint32m2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_f16mf2))) +void vluxseg8ei32(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vfloat16mf2_t * op6, vfloat16mf2_t * op7, const _Float16 * op8, vuint32m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_f16mf2_m))) +void vluxseg8ei32(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vfloat16mf2_t * op6, vfloat16mf2_t * op7, vbool32_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, vfloat16mf2_t op11, vfloat16mf2_t op12, vfloat16mf2_t op13, vfloat16mf2_t op14, vfloat16mf2_t op15, vfloat16mf2_t op16, const _Float16 * op17, vuint32m1_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_f16mf4))) +void vluxseg8ei32(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vfloat16mf4_t * op6, vfloat16mf4_t * op7, const _Float16 * op8, vuint32mf2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei32_v_f16mf4_m))) +void vluxseg8ei32(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vfloat16mf4_t * op6, vfloat16mf4_t * op7, vbool64_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, vfloat16mf4_t op11, vfloat16mf4_t op12, vfloat16mf4_t op13, vfloat16mf4_t op14, vfloat16mf4_t op15, vfloat16mf4_t op16, const _Float16 * op17, vuint32mf2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_f16m1))) +void vluxseg2ei64(vfloat16m1_t * op0, vfloat16m1_t * op1, const _Float16 * op2, vuint64m4_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_f16m1_m))) +void vluxseg2ei64(vfloat16m1_t * op0, vfloat16m1_t * op1, vbool16_t op2, vfloat16m1_t op3, vfloat16m1_t op4, const _Float16 * op5, vuint64m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_f16m2))) +void vluxseg2ei64(vfloat16m2_t * op0, vfloat16m2_t * op1, const _Float16 * op2, vuint64m8_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_f16m2_m))) +void vluxseg2ei64(vfloat16m2_t * op0, vfloat16m2_t * op1, vbool8_t op2, vfloat16m2_t op3, vfloat16m2_t op4, const _Float16 * op5, vuint64m8_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_f16mf2))) +void vluxseg2ei64(vfloat16mf2_t * op0, vfloat16mf2_t * op1, const _Float16 * op2, vuint64m2_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_f16mf2_m))) +void vluxseg2ei64(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vbool32_t op2, vfloat16mf2_t op3, vfloat16mf2_t op4, const _Float16 * op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_f16mf4))) +void vluxseg2ei64(vfloat16mf4_t * op0, vfloat16mf4_t * op1, const _Float16 * op2, vuint64m1_t op3, size_t op4); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg2ei64_v_f16mf4_m))) +void vluxseg2ei64(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vbool64_t op2, vfloat16mf4_t op3, vfloat16mf4_t op4, const _Float16 * op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_f16m1))) +void vluxseg3ei64(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, const _Float16 * op3, vuint64m4_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_f16m1_m))) +void vluxseg3ei64(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vbool16_t op3, vfloat16m1_t op4, vfloat16m1_t op5, vfloat16m1_t op6, const _Float16 * op7, vuint64m4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_f16m2))) +void vluxseg3ei64(vfloat16m2_t * op0, vfloat16m2_t * op1, vfloat16m2_t * op2, const _Float16 * op3, vuint64m8_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_f16m2_m))) +void vluxseg3ei64(vfloat16m2_t * op0, vfloat16m2_t * op1, vfloat16m2_t * op2, vbool8_t op3, vfloat16m2_t op4, vfloat16m2_t op5, vfloat16m2_t op6, const _Float16 * op7, vuint64m8_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_f16mf2))) +void vluxseg3ei64(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, const _Float16 * op3, vuint64m2_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_f16mf2_m))) +void vluxseg3ei64(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vbool32_t op3, vfloat16mf2_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, const _Float16 * op7, vuint64m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_f16mf4))) +void vluxseg3ei64(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, const _Float16 * op3, vuint64m1_t op4, size_t op5); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg3ei64_v_f16mf4_m))) +void vluxseg3ei64(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vbool64_t op3, vfloat16mf4_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, const _Float16 * op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_f16m1))) +void vluxseg4ei64(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, const _Float16 * op4, vuint64m4_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_f16m1_m))) +void vluxseg4ei64(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vbool16_t op4, vfloat16m1_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, const _Float16 * op9, vuint64m4_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_f16m2))) +void vluxseg4ei64(vfloat16m2_t * op0, vfloat16m2_t * op1, vfloat16m2_t * op2, vfloat16m2_t * op3, const _Float16 * op4, vuint64m8_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_f16m2_m))) +void vluxseg4ei64(vfloat16m2_t * op0, vfloat16m2_t * op1, vfloat16m2_t * op2, vfloat16m2_t * op3, vbool8_t op4, vfloat16m2_t op5, vfloat16m2_t op6, vfloat16m2_t op7, vfloat16m2_t op8, const _Float16 * op9, vuint64m8_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_f16mf2))) +void vluxseg4ei64(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, const _Float16 * op4, vuint64m2_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_f16mf2_m))) +void vluxseg4ei64(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vbool32_t op4, vfloat16mf2_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, const _Float16 * op9, vuint64m2_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_f16mf4))) +void vluxseg4ei64(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, const _Float16 * op4, vuint64m1_t op5, size_t op6); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg4ei64_v_f16mf4_m))) +void vluxseg4ei64(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vbool64_t op4, vfloat16mf4_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, const _Float16 * op9, vuint64m1_t op10, size_t op11); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_f16m1))) +void vluxseg5ei64(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, const _Float16 * op5, vuint64m4_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_f16m1_m))) +void vluxseg5ei64(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vbool16_t op5, vfloat16m1_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, vfloat16m1_t op10, const _Float16 * op11, vuint64m4_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_f16mf2))) +void vluxseg5ei64(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, const _Float16 * op5, vuint64m2_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_f16mf2_m))) +void vluxseg5ei64(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vbool32_t op5, vfloat16mf2_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, const _Float16 * op11, vuint64m2_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_f16mf4))) +void vluxseg5ei64(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, const _Float16 * op5, vuint64m1_t op6, size_t op7); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg5ei64_v_f16mf4_m))) +void vluxseg5ei64(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vbool64_t op5, vfloat16mf4_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, const _Float16 * op11, vuint64m1_t op12, size_t op13); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_f16m1))) +void vluxseg6ei64(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, const _Float16 * op6, vuint64m4_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_f16m1_m))) +void vluxseg6ei64(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vbool16_t op6, vfloat16m1_t op7, vfloat16m1_t op8, vfloat16m1_t op9, vfloat16m1_t op10, vfloat16m1_t op11, vfloat16m1_t op12, const _Float16 * op13, vuint64m4_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_f16mf2))) +void vluxseg6ei64(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, const _Float16 * op6, vuint64m2_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_f16mf2_m))) +void vluxseg6ei64(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vbool32_t op6, vfloat16mf2_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, vfloat16mf2_t op11, vfloat16mf2_t op12, const _Float16 * op13, vuint64m2_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_f16mf4))) +void vluxseg6ei64(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, const _Float16 * op6, vuint64m1_t op7, size_t op8); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg6ei64_v_f16mf4_m))) +void vluxseg6ei64(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vbool64_t op6, vfloat16mf4_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, vfloat16mf4_t op11, vfloat16mf4_t op12, const _Float16 * op13, vuint64m1_t op14, size_t op15); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_f16m1))) +void vluxseg7ei64(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vfloat16m1_t * op6, const _Float16 * op7, vuint64m4_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_f16m1_m))) +void vluxseg7ei64(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vfloat16m1_t * op6, vbool16_t op7, vfloat16m1_t op8, vfloat16m1_t op9, vfloat16m1_t op10, vfloat16m1_t op11, vfloat16m1_t op12, vfloat16m1_t op13, vfloat16m1_t op14, const _Float16 * op15, vuint64m4_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_f16mf2))) +void vluxseg7ei64(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vfloat16mf2_t * op6, const _Float16 * op7, vuint64m2_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_f16mf2_m))) +void vluxseg7ei64(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vfloat16mf2_t * op6, vbool32_t op7, vfloat16mf2_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, vfloat16mf2_t op11, vfloat16mf2_t op12, vfloat16mf2_t op13, vfloat16mf2_t op14, const _Float16 * op15, vuint64m2_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_f16mf4))) +void vluxseg7ei64(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vfloat16mf4_t * op6, const _Float16 * op7, vuint64m1_t op8, size_t op9); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg7ei64_v_f16mf4_m))) +void vluxseg7ei64(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vfloat16mf4_t * op6, vbool64_t op7, vfloat16mf4_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, vfloat16mf4_t op11, vfloat16mf4_t op12, vfloat16mf4_t op13, vfloat16mf4_t op14, const _Float16 * op15, vuint64m1_t op16, size_t op17); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_f16m1))) +void vluxseg8ei64(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vfloat16m1_t * op6, vfloat16m1_t * op7, const _Float16 * op8, vuint64m4_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_f16m1_m))) +void vluxseg8ei64(vfloat16m1_t * op0, vfloat16m1_t * op1, vfloat16m1_t * op2, vfloat16m1_t * op3, vfloat16m1_t * op4, vfloat16m1_t * op5, vfloat16m1_t * op6, vfloat16m1_t * op7, vbool16_t op8, vfloat16m1_t op9, vfloat16m1_t op10, vfloat16m1_t op11, vfloat16m1_t op12, vfloat16m1_t op13, vfloat16m1_t op14, vfloat16m1_t op15, vfloat16m1_t op16, const _Float16 * op17, vuint64m4_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_f16mf2))) +void vluxseg8ei64(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vfloat16mf2_t * op6, vfloat16mf2_t * op7, const _Float16 * op8, vuint64m2_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_f16mf2_m))) +void vluxseg8ei64(vfloat16mf2_t * op0, vfloat16mf2_t * op1, vfloat16mf2_t * op2, vfloat16mf2_t * op3, vfloat16mf2_t * op4, vfloat16mf2_t * op5, vfloat16mf2_t * op6, vfloat16mf2_t * op7, vbool32_t op8, vfloat16mf2_t op9, vfloat16mf2_t op10, vfloat16mf2_t op11, vfloat16mf2_t op12, vfloat16mf2_t op13, vfloat16mf2_t op14, vfloat16mf2_t op15, vfloat16mf2_t op16, const _Float16 * op17, vuint64m2_t op18, size_t op19); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_f16mf4))) +void vluxseg8ei64(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vfloat16mf4_t * op6, vfloat16mf4_t * op7, const _Float16 * op8, vuint64m1_t op9, size_t op10); + +__rvv_overloaded __attribute__((clang_builtin_alias(__builtin_rvv_vluxseg8ei64_v_f16mf4_m))) +void vluxseg8ei64(vfloat16mf4_t * op0, vfloat16mf4_t * op1, vfloat16mf4_t * op2, vfloat16mf4_t * op3, vfloat16mf4_t * op4, vfloat16mf4_t * op5, vfloat16mf4_t * op6, vfloat16mf4_t * op7, vbool64_t op8, vfloat16mf4_t op9, vfloat16mf4_t op10, vfloat16mf4_t op11, vfloat16mf4_t op12, vfloat16mf4_t op13, vfloat16mf4_t op14, vfloat16mf4_t op15, vfloat16mf4_t op16, const _Float16 * op17, vuint64m1_t op18, size_t op19); + +#endif + + +#ifdef __cplusplus +} +#endif // __riscv_vector +#endif // __RISCV_VECTOR_H diff --git a/lib/include/uintrintrin.h b/lib/include/uintrintrin.h index 78aa8779c..e3839dceb 100644 --- a/lib/include/uintrintrin.h +++ b/lib/include/uintrintrin.h @@ -20,6 +20,13 @@ #ifdef __x86_64__ +struct __uintr_frame +{ + unsigned long long rip; + unsigned long long rflags; + unsigned long long rsp; +}; + /// Clears the user interrupt flag (UIF). Its effect takes place immediately: a /// user interrupt cannot be delivered on the instruction boundary following /// CLUI. Can be executed only if CR4.UINT = 1, the logical processor is in diff --git a/lib/include/vaesintrin.h b/lib/include/vaesintrin.h index c4d5c3e75..f3c0807bb 100644 --- a/lib/include/vaesintrin.h +++ b/lib/include/vaesintrin.h @@ -28,13 +28,6 @@ static __inline__ __m256i __DEFAULT_FN_ATTRS (__v4di) __B); } -static __inline__ __m512i __DEFAULT_FN_ATTRS_F - _mm512_aesenc_epi128(__m512i __A, __m512i __B) -{ - return (__m512i) __builtin_ia32_aesenc512((__v8di) __A, - (__v8di) __B); -} - static __inline__ __m256i __DEFAULT_FN_ATTRS _mm256_aesdec_epi128(__m256i __A, __m256i __B) { @@ -42,32 +35,40 @@ static __inline__ __m256i __DEFAULT_FN_ATTRS (__v4di) __B); } -static __inline__ __m512i __DEFAULT_FN_ATTRS_F - _mm512_aesdec_epi128(__m512i __A, __m512i __B) +static __inline__ __m256i __DEFAULT_FN_ATTRS + _mm256_aesenclast_epi128(__m256i __A, __m256i __B) { - return (__m512i) __builtin_ia32_aesdec512((__v8di) __A, - (__v8di) __B); + return (__m256i) __builtin_ia32_aesenclast256((__v4di) __A, + (__v4di) __B); } static __inline__ __m256i __DEFAULT_FN_ATTRS - _mm256_aesenclast_epi128(__m256i __A, __m256i __B) + _mm256_aesdeclast_epi128(__m256i __A, __m256i __B) { - return (__m256i) __builtin_ia32_aesenclast256((__v4di) __A, + return (__m256i) __builtin_ia32_aesdeclast256((__v4di) __A, (__v4di) __B); } +#ifdef __AVX512FINTRIN_H static __inline__ __m512i __DEFAULT_FN_ATTRS_F - _mm512_aesenclast_epi128(__m512i __A, __m512i __B) + _mm512_aesenc_epi128(__m512i __A, __m512i __B) { - return (__m512i) __builtin_ia32_aesenclast512((__v8di) __A, + return (__m512i) __builtin_ia32_aesenc512((__v8di) __A, (__v8di) __B); } -static __inline__ __m256i __DEFAULT_FN_ATTRS - _mm256_aesdeclast_epi128(__m256i __A, __m256i __B) +static __inline__ __m512i __DEFAULT_FN_ATTRS_F + _mm512_aesdec_epi128(__m512i __A, __m512i __B) { - return (__m256i) __builtin_ia32_aesdeclast256((__v4di) __A, - (__v4di) __B); + return (__m512i) __builtin_ia32_aesdec512((__v8di) __A, + (__v8di) __B); +} + +static __inline__ __m512i __DEFAULT_FN_ATTRS_F + _mm512_aesenclast_epi128(__m512i __A, __m512i __B) +{ + return (__m512i) __builtin_ia32_aesenclast512((__v8di) __A, + (__v8di) __B); } static __inline__ __m512i __DEFAULT_FN_ATTRS_F @@ -76,7 +77,7 @@ static __inline__ __m512i __DEFAULT_FN_ATTRS_F return (__m512i) __builtin_ia32_aesdeclast512((__v8di) __A, (__v8di) __B); } - +#endif // __AVX512FINTRIN_H #undef __DEFAULT_FN_ATTRS #undef __DEFAULT_FN_ATTRS_F diff --git a/lib/include/vecintrin.h b/lib/include/vecintrin.h index e58c9769e..ec1dbfd01 100644 --- a/lib/include/vecintrin.h +++ b/lib/include/vecintrin.h @@ -1016,64 +1016,84 @@ vec_scatter_element(__vector double __vec, static inline __ATTRS_o_ai __vector signed char vec_xl(long __offset, const signed char *__ptr) { - return *(const __vector signed char *) - ((const char *)__ptr + __offset); + __vector signed char V; + __builtin_memcpy(&V, ((const char *)__ptr + __offset), + sizeof(__vector signed char)); + return V; } static inline __ATTRS_o_ai __vector unsigned char vec_xl(long __offset, const unsigned char *__ptr) { - return *(const __vector unsigned char *) - ((const char *)__ptr + __offset); + __vector unsigned char V; + __builtin_memcpy(&V, ((const char *)__ptr + __offset), + sizeof(__vector unsigned char)); + return V; } static inline __ATTRS_o_ai __vector signed short vec_xl(long __offset, const signed short *__ptr) { - return *(const __vector signed short *) - ((const char *)__ptr + __offset); + __vector signed short V; + __builtin_memcpy(&V, ((const char *)__ptr + __offset), + sizeof(__vector signed short)); + return V; } static inline __ATTRS_o_ai __vector unsigned short vec_xl(long __offset, const unsigned short *__ptr) { - return *(const __vector unsigned short *) - ((const char *)__ptr + __offset); + __vector unsigned short V; + __builtin_memcpy(&V, ((const char *)__ptr + __offset), + sizeof(__vector unsigned short)); + return V; } static inline __ATTRS_o_ai __vector signed int vec_xl(long __offset, const signed int *__ptr) { - return *(const __vector signed int *) - ((const char *)__ptr + __offset); + __vector signed int V; + __builtin_memcpy(&V, ((const char *)__ptr + __offset), + sizeof(__vector signed int)); + return V; } static inline __ATTRS_o_ai __vector unsigned int vec_xl(long __offset, const unsigned int *__ptr) { - return *(const __vector unsigned int *) - ((const char *)__ptr + __offset); + __vector unsigned int V; + __builtin_memcpy(&V, ((const char *)__ptr + __offset), + sizeof(__vector unsigned int)); + return V; } static inline __ATTRS_o_ai __vector signed long long vec_xl(long __offset, const signed long long *__ptr) { - return *(const __vector signed long long *) - ((const char *)__ptr + __offset); + __vector signed long long V; + __builtin_memcpy(&V, ((const char *)__ptr + __offset), + sizeof(__vector signed long long)); + return V; } static inline __ATTRS_o_ai __vector unsigned long long vec_xl(long __offset, const unsigned long long *__ptr) { - return *(const __vector unsigned long long *) - ((const char *)__ptr + __offset); + __vector unsigned long long V; + __builtin_memcpy(&V, ((const char *)__ptr + __offset), + sizeof(__vector unsigned long long)); + return V; } #if __ARCH__ >= 12 static inline __ATTRS_o_ai __vector float vec_xl(long __offset, const float *__ptr) { - return *(const __vector float *) - ((const char *)__ptr + __offset); + __vector float V; + __builtin_memcpy(&V, ((const char *)__ptr + __offset), + sizeof(__vector float)); + return V; } #endif static inline __ATTRS_o_ai __vector double vec_xl(long __offset, const double *__ptr) { - return *(const __vector double *) - ((const char *)__ptr + __offset); + __vector double V; + __builtin_memcpy(&V, ((const char *)__ptr + __offset), + sizeof(__vector double)); + return V; } /*-- vec_xld2 ---------------------------------------------------------------*/ @@ -1081,64 +1101,82 @@ vec_xl(long __offset, const double *__ptr) { // This prototype is deprecated. static inline __ATTRS_o_ai __vector signed char vec_xld2(long __offset, const signed char *__ptr) { - return *(const __vector signed char *) - ((const char *)__ptr + __offset); + __vector signed char V; + __builtin_memcpy(&V, ((const char *)__ptr + __offset), + sizeof(__vector signed char)); + return V; } // This prototype is deprecated. static inline __ATTRS_o_ai __vector unsigned char vec_xld2(long __offset, const unsigned char *__ptr) { - return *(const __vector unsigned char *) - ((const char *)__ptr + __offset); + __vector unsigned char V; + __builtin_memcpy(&V, ((const char *)__ptr + __offset), + sizeof(__vector unsigned char)); + return V; } // This prototype is deprecated. static inline __ATTRS_o_ai __vector signed short vec_xld2(long __offset, const signed short *__ptr) { - return *(const __vector signed short *) - ((const char *)__ptr + __offset); + __vector signed short V; + __builtin_memcpy(&V, ((const char *)__ptr + __offset), + sizeof(__vector signed short)); + return V; } // This prototype is deprecated. static inline __ATTRS_o_ai __vector unsigned short vec_xld2(long __offset, const unsigned short *__ptr) { - return *(const __vector unsigned short *) - ((const char *)__ptr + __offset); + __vector unsigned short V; + __builtin_memcpy(&V, ((const char *)__ptr + __offset), + sizeof(__vector unsigned short)); + return V; } // This prototype is deprecated. static inline __ATTRS_o_ai __vector signed int vec_xld2(long __offset, const signed int *__ptr) { - return *(const __vector signed int *) - ((const char *)__ptr + __offset); + __vector signed int V; + __builtin_memcpy(&V, ((const char *)__ptr + __offset), + sizeof(__vector signed int)); + return V; } // This prototype is deprecated. static inline __ATTRS_o_ai __vector unsigned int vec_xld2(long __offset, const unsigned int *__ptr) { - return *(const __vector unsigned int *) - ((const char *)__ptr + __offset); + __vector unsigned int V; + __builtin_memcpy(&V, ((const char *)__ptr + __offset), + sizeof(__vector unsigned int)); + return V; } // This prototype is deprecated. static inline __ATTRS_o_ai __vector signed long long vec_xld2(long __offset, const signed long long *__ptr) { - return *(const __vector signed long long *) - ((const char *)__ptr + __offset); + __vector signed long long V; + __builtin_memcpy(&V, ((const char *)__ptr + __offset), + sizeof(__vector signed long long)); + return V; } // This prototype is deprecated. static inline __ATTRS_o_ai __vector unsigned long long vec_xld2(long __offset, const unsigned long long *__ptr) { - return *(const __vector unsigned long long *) - ((const char *)__ptr + __offset); + __vector unsigned long long V; + __builtin_memcpy(&V, ((const char *)__ptr + __offset), + sizeof(__vector unsigned long long)); + return V; } // This prototype is deprecated. static inline __ATTRS_o_ai __vector double vec_xld2(long __offset, const double *__ptr) { - return *(const __vector double *) - ((const char *)__ptr + __offset); + __vector double V; + __builtin_memcpy(&V, ((const char *)__ptr + __offset), + sizeof(__vector double)); + return V; } /*-- vec_xlw4 ---------------------------------------------------------------*/ @@ -1146,99 +1184,128 @@ vec_xld2(long __offset, const double *__ptr) { // This prototype is deprecated. static inline __ATTRS_o_ai __vector signed char vec_xlw4(long __offset, const signed char *__ptr) { - return *(const __vector signed char *) - ((const char *)__ptr + __offset); + __vector signed char V; + __builtin_memcpy(&V, ((const char *)__ptr + __offset), + sizeof(__vector signed char)); + return V; } // This prototype is deprecated. static inline __ATTRS_o_ai __vector unsigned char vec_xlw4(long __offset, const unsigned char *__ptr) { - return *(const __vector unsigned char *) - ((const char *)__ptr + __offset); + __vector unsigned char V; + __builtin_memcpy(&V, ((const char *)__ptr + __offset), + sizeof(__vector unsigned char)); + return V; } // This prototype is deprecated. static inline __ATTRS_o_ai __vector signed short vec_xlw4(long __offset, const signed short *__ptr) { - return *(const __vector signed short *) - ((const char *)__ptr + __offset); + __vector signed short V; + __builtin_memcpy(&V, ((const char *)__ptr + __offset), + sizeof(__vector signed short)); + return V; } // This prototype is deprecated. static inline __ATTRS_o_ai __vector unsigned short vec_xlw4(long __offset, const unsigned short *__ptr) { - return *(const __vector unsigned short *) - ((const char *)__ptr + __offset); + __vector unsigned short V; + __builtin_memcpy(&V, ((const char *)__ptr + __offset), + sizeof(__vector unsigned short)); + return V; } // This prototype is deprecated. static inline __ATTRS_o_ai __vector signed int vec_xlw4(long __offset, const signed int *__ptr) { - return *(const __vector signed int *) - ((const char *)__ptr + __offset); + __vector signed int V; + __builtin_memcpy(&V, ((const char *)__ptr + __offset), + sizeof(__vector signed int)); + return V; } // This prototype is deprecated. static inline __ATTRS_o_ai __vector unsigned int vec_xlw4(long __offset, const unsigned int *__ptr) { - return *(const __vector unsigned int *) - ((const char *)__ptr + __offset); + __vector unsigned int V; + __builtin_memcpy(&V, ((const char *)__ptr + __offset), + sizeof(__vector unsigned int)); + return V; } /*-- vec_xst ----------------------------------------------------------------*/ static inline __ATTRS_o_ai void vec_xst(__vector signed char __vec, long __offset, signed char *__ptr) { - *(__vector signed char *)((char *)__ptr + __offset) = __vec; + __vector signed char V = __vec; + __builtin_memcpy(((char *)__ptr + __offset), &V, + sizeof(__vector signed char)); } static inline __ATTRS_o_ai void vec_xst(__vector unsigned char __vec, long __offset, unsigned char *__ptr) { - *(__vector unsigned char *)((char *)__ptr + __offset) = __vec; + __vector unsigned char V = __vec; + __builtin_memcpy(((char *)__ptr + __offset), &V, + sizeof(__vector unsigned char)); } static inline __ATTRS_o_ai void vec_xst(__vector signed short __vec, long __offset, signed short *__ptr) { - *(__vector signed short *)((char *)__ptr + __offset) = __vec; + __vector signed short V = __vec; + __builtin_memcpy(((char *)__ptr + __offset), &V, + sizeof(__vector signed short)); } static inline __ATTRS_o_ai void vec_xst(__vector unsigned short __vec, long __offset, unsigned short *__ptr) { - *(__vector unsigned short *)((char *)__ptr + __offset) = __vec; + __vector unsigned short V = __vec; + __builtin_memcpy(((char *)__ptr + __offset), &V, + sizeof(__vector unsigned short)); } static inline __ATTRS_o_ai void vec_xst(__vector signed int __vec, long __offset, signed int *__ptr) { - *(__vector signed int *)((char *)__ptr + __offset) = __vec; + __vector signed int V = __vec; + __builtin_memcpy(((char *)__ptr + __offset), &V, sizeof(__vector signed int)); } static inline __ATTRS_o_ai void vec_xst(__vector unsigned int __vec, long __offset, unsigned int *__ptr) { - *(__vector unsigned int *)((char *)__ptr + __offset) = __vec; + __vector unsigned int V = __vec; + __builtin_memcpy(((char *)__ptr + __offset), &V, + sizeof(__vector unsigned int)); } static inline __ATTRS_o_ai void vec_xst(__vector signed long long __vec, long __offset, - signed long long *__ptr) { - *(__vector signed long long *)((char *)__ptr + __offset) = __vec; + signed long long *__ptr) { + __vector signed long long V = __vec; + __builtin_memcpy(((char *)__ptr + __offset), &V, + sizeof(__vector signed long long)); } static inline __ATTRS_o_ai void vec_xst(__vector unsigned long long __vec, long __offset, - unsigned long long *__ptr) { - *(__vector unsigned long long *)((char *)__ptr + __offset) = __vec; + unsigned long long *__ptr) { + __vector unsigned long long V = __vec; + __builtin_memcpy(((char *)__ptr + __offset), &V, + sizeof(__vector unsigned long long)); } #if __ARCH__ >= 12 static inline __ATTRS_o_ai void vec_xst(__vector float __vec, long __offset, float *__ptr) { - *(__vector float *)((char *)__ptr + __offset) = __vec; + __vector float V = __vec; + __builtin_memcpy(((char *)__ptr + __offset), &V, sizeof(__vector float)); } #endif static inline __ATTRS_o_ai void vec_xst(__vector double __vec, long __offset, double *__ptr) { - *(__vector double *)((char *)__ptr + __offset) = __vec; + __vector double V = __vec; + __builtin_memcpy(((char *)__ptr + __offset), &V, sizeof(__vector double)); } /*-- vec_xstd2 --------------------------------------------------------------*/ @@ -1246,57 +1313,73 @@ vec_xst(__vector double __vec, long __offset, double *__ptr) { // This prototype is deprecated. static inline __ATTRS_o_ai void vec_xstd2(__vector signed char __vec, long __offset, signed char *__ptr) { - *(__vector signed char *)((char *)__ptr + __offset) = __vec; + __vector signed char V = __vec; + __builtin_memcpy(((char *)__ptr + __offset), &V, + sizeof(__vector signed char)); } // This prototype is deprecated. static inline __ATTRS_o_ai void vec_xstd2(__vector unsigned char __vec, long __offset, unsigned char *__ptr) { - *(__vector unsigned char *)((char *)__ptr + __offset) = __vec; + __vector unsigned char V = __vec; + __builtin_memcpy(((char *)__ptr + __offset), &V, + sizeof(__vector unsigned char)); } // This prototype is deprecated. static inline __ATTRS_o_ai void vec_xstd2(__vector signed short __vec, long __offset, signed short *__ptr) { - *(__vector signed short *)((char *)__ptr + __offset) = __vec; + __vector signed short V = __vec; + __builtin_memcpy(((char *)__ptr + __offset), &V, + sizeof(__vector signed short)); } // This prototype is deprecated. static inline __ATTRS_o_ai void vec_xstd2(__vector unsigned short __vec, long __offset, unsigned short *__ptr) { - *(__vector unsigned short *)((char *)__ptr + __offset) = __vec; + __vector unsigned short V = __vec; + __builtin_memcpy(((char *)__ptr + __offset), &V, + sizeof(__vector unsigned short)); } // This prototype is deprecated. static inline __ATTRS_o_ai void vec_xstd2(__vector signed int __vec, long __offset, signed int *__ptr) { - *(__vector signed int *)((char *)__ptr + __offset) = __vec; + __vector signed int V = __vec; + __builtin_memcpy(((char *)__ptr + __offset), &V, sizeof(__vector signed int)); } // This prototype is deprecated. static inline __ATTRS_o_ai void vec_xstd2(__vector unsigned int __vec, long __offset, unsigned int *__ptr) { - *(__vector unsigned int *)((char *)__ptr + __offset) = __vec; + __vector unsigned int V = __vec; + __builtin_memcpy(((char *)__ptr + __offset), &V, + sizeof(__vector unsigned int)); } // This prototype is deprecated. static inline __ATTRS_o_ai void vec_xstd2(__vector signed long long __vec, long __offset, signed long long *__ptr) { - *(__vector signed long long *)((char *)__ptr + __offset) = __vec; + __vector signed long long V = __vec; + __builtin_memcpy(((char *)__ptr + __offset), &V, + sizeof(__vector signed long long)); } // This prototype is deprecated. static inline __ATTRS_o_ai void vec_xstd2(__vector unsigned long long __vec, long __offset, unsigned long long *__ptr) { - *(__vector unsigned long long *)((char *)__ptr + __offset) = __vec; + __vector unsigned long long V = __vec; + __builtin_memcpy(((char *)__ptr + __offset), &V, + sizeof(__vector unsigned long long)); } // This prototype is deprecated. static inline __ATTRS_o_ai void vec_xstd2(__vector double __vec, long __offset, double *__ptr) { - *(__vector double *)((char *)__ptr + __offset) = __vec; + __vector double V = __vec; + __builtin_memcpy(((char *)__ptr + __offset), &V, sizeof(__vector double)); } /*-- vec_xstw4 --------------------------------------------------------------*/ @@ -1304,37 +1387,48 @@ vec_xstd2(__vector double __vec, long __offset, double *__ptr) { // This prototype is deprecated. static inline __ATTRS_o_ai void vec_xstw4(__vector signed char __vec, long __offset, signed char *__ptr) { - *(__vector signed char *)((char *)__ptr + __offset) = __vec; + __vector signed char V = __vec; + __builtin_memcpy(((char *)__ptr + __offset), &V, + sizeof(__vector signed char)); } // This prototype is deprecated. static inline __ATTRS_o_ai void vec_xstw4(__vector unsigned char __vec, long __offset, unsigned char *__ptr) { - *(__vector unsigned char *)((char *)__ptr + __offset) = __vec; + __vector unsigned char V = __vec; + __builtin_memcpy(((char *)__ptr + __offset), &V, + sizeof(__vector unsigned char)); } // This prototype is deprecated. static inline __ATTRS_o_ai void vec_xstw4(__vector signed short __vec, long __offset, signed short *__ptr) { - *(__vector signed short *)((char *)__ptr + __offset) = __vec; + __vector signed short V = __vec; + __builtin_memcpy(((char *)__ptr + __offset), &V, + sizeof(__vector signed short)); } // This prototype is deprecated. static inline __ATTRS_o_ai void vec_xstw4(__vector unsigned short __vec, long __offset, unsigned short *__ptr) { - *(__vector unsigned short *)((char *)__ptr + __offset) = __vec; + __vector unsigned short V = __vec; + __builtin_memcpy(((char *)__ptr + __offset), &V, + sizeof(__vector unsigned short)); } // This prototype is deprecated. static inline __ATTRS_o_ai void vec_xstw4(__vector signed int __vec, long __offset, signed int *__ptr) { - *(__vector signed int *)((char *)__ptr + __offset) = __vec; + __vector signed int V = __vec; + __builtin_memcpy(((char *)__ptr + __offset), &V, sizeof(__vector signed int)); } // This prototype is deprecated. static inline __ATTRS_o_ai void vec_xstw4(__vector unsigned int __vec, long __offset, unsigned int *__ptr) { - *(__vector unsigned int *)((char *)__ptr + __offset) = __vec; + __vector unsigned int V = __vec; + __builtin_memcpy(((char *)__ptr + __offset), &V, + sizeof(__vector unsigned int)); } /*-- vec_load_bndry ---------------------------------------------------------*/ @@ -9259,6 +9353,41 @@ vec_fp_test_data_class(__vector double __a, int __b, int *__c) __VEC_CLASS_FP_ZERO | \ __VEC_CLASS_FP_INFINITY) +/*-- vec_extend_to_fp32_hi --------------------------------------------------*/ + +#if __ARCH__ >= 14 +#define vec_extend_to_fp32_hi(X, W) \ + ((__vector float)__builtin_s390_vclfnhs((X), (W))); +#endif + +/*-- vec_extend_to_fp32_hi --------------------------------------------------*/ + +#if __ARCH__ >= 14 +#define vec_extend_to_fp32_lo(X, W) \ + ((__vector float)__builtin_s390_vclfnls((X), (W))); +#endif + +/*-- vec_round_from_fp32 ----------------------------------------------------*/ + +#if __ARCH__ >= 14 +#define vec_round_from_fp32(X, Y, W) \ + ((__vector unsigned short)__builtin_s390_vcrnfs((X), (Y), (W))); +#endif + +/*-- vec_convert_to_fp16 ----------------------------------------------------*/ + +#if __ARCH__ >= 14 +#define vec_convert_to_fp16(X, W) \ + ((__vector unsigned short)__builtin_s390_vcfn((X), (W))); +#endif + +/*-- vec_convert_from_fp16 --------------------------------------------------*/ + +#if __ARCH__ >= 14 +#define vec_convert_from_fp16(X, W) \ + ((__vector unsigned short)__builtin_s390_vcnf((X), (W))); +#endif + /*-- vec_cp_until_zero ------------------------------------------------------*/ static inline __ATTRS_o_ai __vector signed char diff --git a/lib/include/vpclmulqdqintrin.h b/lib/include/vpclmulqdqintrin.h index 470d83254..44daadb07 100644 --- a/lib/include/vpclmulqdqintrin.h +++ b/lib/include/vpclmulqdqintrin.h @@ -19,10 +19,12 @@ (__v4di)(__m256i)(B), \ (char)(I)) +#ifdef __AVX512FINTRIN_H #define _mm512_clmulepi64_epi128(A, B, I) \ (__m512i)__builtin_ia32_pclmulqdq512((__v8di)(__m512i)(A), \ (__v8di)(__m512i)(B), \ (char)(I)) +#endif // __AVX512FINTRIN_H #endif /* __VPCLMULQDQINTRIN_H */ diff --git a/lib/include/wasm_simd128.h b/lib/include/wasm_simd128.h index ac88516ac..712fa0378 100644 --- a/lib/include/wasm_simd128.h +++ b/lib/include/wasm_simd128.h @@ -40,13 +40,18 @@ typedef unsigned char __u8x8 typedef short __i16x4 __attribute__((__vector_size__(8), __aligned__(8))); typedef unsigned short __u16x4 __attribute__((__vector_size__(8), __aligned__(8))); +typedef int __i32x2 __attribute__((__vector_size__(8), __aligned__(8))); +typedef unsigned int __u32x2 + __attribute__((__vector_size__(8), __aligned__(8))); +typedef float __f32x2 __attribute__((__vector_size__(8), __aligned__(8))); #define __DEFAULT_FN_ATTRS \ __attribute__((__always_inline__, __nodebug__, __target__("simd128"), \ __min_vector_width__(128))) -#define __REQUIRE_CONSTANT(e) \ - _Static_assert(__builtin_constant_p(e), "Expected constant") +#define __REQUIRE_CONSTANT(c) \ + __attribute__((__diagnose_if__(!__builtin_constant_p(c), \ + #c " must be constant", "error"))) static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_v128_load(const void *__mem) { // UB-free unaligned access copied from xmmintrin.h @@ -57,102 +62,158 @@ static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_v128_load(const void *__mem) { } static __inline__ v128_t __DEFAULT_FN_ATTRS -wasm_v8x16_load_splat(const void *__mem) { - struct __wasm_v8x16_load_splat_struct { +wasm_v128_load8_splat(const void *__mem) { + struct __wasm_v128_load8_splat_struct { uint8_t __v; } __attribute__((__packed__, __may_alias__)); - uint8_t __v = ((const struct __wasm_v8x16_load_splat_struct *)__mem)->__v; + uint8_t __v = ((const struct __wasm_v128_load8_splat_struct *)__mem)->__v; return (v128_t)(__u8x16){__v, __v, __v, __v, __v, __v, __v, __v, __v, __v, __v, __v, __v, __v, __v, __v}; } static __inline__ v128_t __DEFAULT_FN_ATTRS -wasm_v16x8_load_splat(const void *__mem) { - struct __wasm_v16x8_load_splat_struct { +wasm_v128_load16_splat(const void *__mem) { + struct __wasm_v128_load16_splat_struct { uint16_t __v; } __attribute__((__packed__, __may_alias__)); - uint16_t __v = ((const struct __wasm_v16x8_load_splat_struct *)__mem)->__v; + uint16_t __v = ((const struct __wasm_v128_load16_splat_struct *)__mem)->__v; return (v128_t)(__u16x8){__v, __v, __v, __v, __v, __v, __v, __v}; } static __inline__ v128_t __DEFAULT_FN_ATTRS -wasm_v32x4_load_splat(const void *__mem) { - struct __wasm_v32x4_load_splat_struct { +wasm_v128_load32_splat(const void *__mem) { + struct __wasm_v128_load32_splat_struct { uint32_t __v; } __attribute__((__packed__, __may_alias__)); - uint32_t __v = ((const struct __wasm_v32x4_load_splat_struct *)__mem)->__v; + uint32_t __v = ((const struct __wasm_v128_load32_splat_struct *)__mem)->__v; return (v128_t)(__u32x4){__v, __v, __v, __v}; } static __inline__ v128_t __DEFAULT_FN_ATTRS -wasm_v64x2_load_splat(const void *__mem) { - struct __wasm_v64x2_load_splat_struct { +wasm_v128_load64_splat(const void *__mem) { + struct __wasm_v128_load64_splat_struct { uint64_t __v; } __attribute__((__packed__, __may_alias__)); - uint64_t __v = ((const struct __wasm_v64x2_load_splat_struct *)__mem)->__v; + uint64_t __v = ((const struct __wasm_v128_load64_splat_struct *)__mem)->__v; return (v128_t)(__u64x2){__v, __v}; } static __inline__ v128_t __DEFAULT_FN_ATTRS -wasm_i16x8_load_8x8(const void *__mem) { - typedef int8_t __i8x8 __attribute__((__vector_size__(8), __aligned__(8))); - struct __wasm_i16x8_load_8x8_struct { +wasm_i16x8_load8x8(const void *__mem) { + struct __wasm_i16x8_load8x8_struct { __i8x8 __v; } __attribute__((__packed__, __may_alias__)); - __i8x8 __v = ((const struct __wasm_i16x8_load_8x8_struct *)__mem)->__v; + __i8x8 __v = ((const struct __wasm_i16x8_load8x8_struct *)__mem)->__v; return (v128_t) __builtin_convertvector(__v, __i16x8); } static __inline__ v128_t __DEFAULT_FN_ATTRS -wasm_u16x8_load_8x8(const void *__mem) { - typedef uint8_t __u8x8 __attribute__((__vector_size__(8), __aligned__(8))); - struct __wasm_u16x8_load_8x8_struct { +wasm_u16x8_load8x8(const void *__mem) { + struct __wasm_u16x8_load8x8_struct { __u8x8 __v; } __attribute__((__packed__, __may_alias__)); - __u8x8 __v = ((const struct __wasm_u16x8_load_8x8_struct *)__mem)->__v; + __u8x8 __v = ((const struct __wasm_u16x8_load8x8_struct *)__mem)->__v; return (v128_t) __builtin_convertvector(__v, __u16x8); } static __inline__ v128_t __DEFAULT_FN_ATTRS -wasm_i32x4_load_16x4(const void *__mem) { - typedef int16_t __i16x4 __attribute__((__vector_size__(8), __aligned__(8))); - struct __wasm_i32x4_load_16x4_struct { +wasm_i32x4_load16x4(const void *__mem) { + struct __wasm_i32x4_load16x4_struct { __i16x4 __v; } __attribute__((__packed__, __may_alias__)); - __i16x4 __v = ((const struct __wasm_i32x4_load_16x4_struct *)__mem)->__v; + __i16x4 __v = ((const struct __wasm_i32x4_load16x4_struct *)__mem)->__v; return (v128_t) __builtin_convertvector(__v, __i32x4); } static __inline__ v128_t __DEFAULT_FN_ATTRS -wasm_u32x4_load_16x4(const void *__mem) { - typedef uint16_t __u16x4 __attribute__((__vector_size__(8), __aligned__(8))); - struct __wasm_u32x4_load_16x4_struct { +wasm_u32x4_load16x4(const void *__mem) { + struct __wasm_u32x4_load16x4_struct { __u16x4 __v; } __attribute__((__packed__, __may_alias__)); - __u16x4 __v = ((const struct __wasm_u32x4_load_16x4_struct *)__mem)->__v; + __u16x4 __v = ((const struct __wasm_u32x4_load16x4_struct *)__mem)->__v; return (v128_t) __builtin_convertvector(__v, __u32x4); } static __inline__ v128_t __DEFAULT_FN_ATTRS -wasm_i64x2_load_32x2(const void *__mem) { - typedef int32_t __i32x2 __attribute__((__vector_size__(8), __aligned__(8))); - struct __wasm_i64x2_load_32x2_struct { +wasm_i64x2_load32x2(const void *__mem) { + struct __wasm_i64x2_load32x2_struct { __i32x2 __v; } __attribute__((__packed__, __may_alias__)); - __i32x2 __v = ((const struct __wasm_i64x2_load_32x2_struct *)__mem)->__v; + __i32x2 __v = ((const struct __wasm_i64x2_load32x2_struct *)__mem)->__v; return (v128_t) __builtin_convertvector(__v, __i64x2); } static __inline__ v128_t __DEFAULT_FN_ATTRS -wasm_u64x2_load_32x2(const void *__mem) { - typedef uint32_t __u32x2 __attribute__((__vector_size__(8), __aligned__(8))); - struct __wasm_u64x2_load_32x2_struct { +wasm_u64x2_load32x2(const void *__mem) { + struct __wasm_u64x2_load32x2_struct { __u32x2 __v; } __attribute__((__packed__, __may_alias__)); - __u32x2 __v = ((const struct __wasm_u64x2_load_32x2_struct *)__mem)->__v; + __u32x2 __v = ((const struct __wasm_u64x2_load32x2_struct *)__mem)->__v; return (v128_t) __builtin_convertvector(__v, __u64x2); } +static __inline__ v128_t __DEFAULT_FN_ATTRS +wasm_v128_load32_zero(const void *__mem) { + struct __wasm_v128_load32_zero_struct { + int32_t __v; + } __attribute__((__packed__, __may_alias__)); + int32_t __v = ((const struct __wasm_v128_load32_zero_struct *)__mem)->__v; + return (v128_t)(__i32x4){__v, 0, 0, 0}; +} + +static __inline__ v128_t __DEFAULT_FN_ATTRS +wasm_v128_load64_zero(const void *__mem) { + struct __wasm_v128_load64_zero_struct { + int64_t __v; + } __attribute__((__packed__, __may_alias__)); + int64_t __v = ((const struct __wasm_v128_load64_zero_struct *)__mem)->__v; + return (v128_t)(__i64x2){__v, 0}; +} + +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_v128_load8_lane( + const void *__mem, v128_t __vec, int __i) __REQUIRE_CONSTANT(__i) { + struct __wasm_v128_load8_lane_struct { + int8_t __v; + } __attribute__((__packed__, __may_alias__)); + int8_t __v = ((const struct __wasm_v128_load8_lane_struct *)__mem)->__v; + __i8x16 __ret = (__i8x16)__vec; + __ret[__i] = __v; + return (v128_t)__ret; +} + +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_v128_load16_lane( + const void *__mem, v128_t __vec, int __i) __REQUIRE_CONSTANT(__i) { + struct __wasm_v128_load16_lane_struct { + int16_t __v; + } __attribute__((__packed__, __may_alias__)); + int16_t __v = ((const struct __wasm_v128_load16_lane_struct *)__mem)->__v; + __i16x8 __ret = (__i16x8)__vec; + __ret[__i] = __v; + return (v128_t)__ret; +} + +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_v128_load32_lane( + const void *__mem, v128_t __vec, int __i) __REQUIRE_CONSTANT(__i) { + struct __wasm_v128_load32_lane_struct { + int32_t __v; + } __attribute__((__packed__, __may_alias__)); + int32_t __v = ((const struct __wasm_v128_load32_lane_struct *)__mem)->__v; + __i32x4 __ret = (__i32x4)__vec; + __ret[__i] = __v; + return (v128_t)__ret; +} + +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_v128_load64_lane( + const void *__mem, v128_t __vec, int __i) __REQUIRE_CONSTANT(__i) { + struct __wasm_v128_load64_lane_struct { + int64_t __v; + } __attribute__((__packed__, __may_alias__)); + int64_t __v = ((const struct __wasm_v128_load64_lane_struct *)__mem)->__v; + __i64x2 __ret = (__i64x2)__vec; + __ret[__i] = __v; + return (v128_t)__ret; +} + static __inline__ void __DEFAULT_FN_ATTRS wasm_v128_store(void *__mem, v128_t __a) { // UB-free unaligned access copied from xmmintrin.h @@ -162,6 +223,49 @@ static __inline__ void __DEFAULT_FN_ATTRS wasm_v128_store(void *__mem, ((struct __wasm_v128_store_struct *)__mem)->__v = __a; } +static __inline__ void __DEFAULT_FN_ATTRS wasm_v128_store8_lane(void *__mem, + v128_t __vec, + int __i) + __REQUIRE_CONSTANT(__i) { + struct __wasm_v128_store8_lane_struct { + int8_t __v; + } __attribute__((__packed__, __may_alias__)); + ((struct __wasm_v128_store8_lane_struct *)__mem)->__v = ((__i8x16)__vec)[__i]; +} + +static __inline__ void __DEFAULT_FN_ATTRS wasm_v128_store16_lane(void *__mem, + v128_t __vec, + int __i) + __REQUIRE_CONSTANT(__i) { + struct __wasm_v128_store16_lane_struct { + int16_t __v; + } __attribute__((__packed__, __may_alias__)); + ((struct __wasm_v128_store16_lane_struct *)__mem)->__v = + ((__i16x8)__vec)[__i]; +} + +static __inline__ void __DEFAULT_FN_ATTRS wasm_v128_store32_lane(void *__mem, + v128_t __vec, + int __i) + __REQUIRE_CONSTANT(__i) { + struct __wasm_v128_store32_lane_struct { + int32_t __v; + } __attribute__((__packed__, __may_alias__)); + ((struct __wasm_v128_store32_lane_struct *)__mem)->__v = + ((__i32x4)__vec)[__i]; +} + +static __inline__ void __DEFAULT_FN_ATTRS wasm_v128_store64_lane(void *__mem, + v128_t __vec, + int __i) + __REQUIRE_CONSTANT(__i) { + struct __wasm_v128_store64_lane_struct { + int64_t __v; + } __attribute__((__packed__, __may_alias__)); + ((struct __wasm_v128_store64_lane_struct *)__mem)->__v = + ((__i64x2)__vec)[__i]; +} + static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i8x16_make(int8_t __c0, int8_t __c1, int8_t __c2, int8_t __c3, int8_t __c4, int8_t __c5, int8_t __c6, int8_t __c7, int8_t __c8, int8_t __c9, @@ -185,6 +289,11 @@ static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i32x4_make(int32_t __c0, return (v128_t)(__i32x4){__c0, __c1, __c2, __c3}; } +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i64x2_make(int64_t __c0, + int64_t __c1) { + return (v128_t)(__i64x2){__c0, __c1}; +} + static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f32x4_make(float __c0, float __c1, float __c2, @@ -192,150 +301,221 @@ static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f32x4_make(float __c0, return (v128_t)(__f32x4){__c0, __c1, __c2, __c3}; } -static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i64x2_make(int64_t __c0, - int64_t __c1) { +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f64x2_make(double __c0, + double __c1) { + return (v128_t)(__f64x2){__c0, __c1}; +} + +static __inline__ v128_t __DEFAULT_FN_ATTRS +wasm_i8x16_const(int8_t __c0, int8_t __c1, int8_t __c2, int8_t __c3, + int8_t __c4, int8_t __c5, int8_t __c6, int8_t __c7, + int8_t __c8, int8_t __c9, int8_t __c10, int8_t __c11, + int8_t __c12, int8_t __c13, int8_t __c14, int8_t __c15) + __REQUIRE_CONSTANT(__c0) __REQUIRE_CONSTANT(__c1) __REQUIRE_CONSTANT(__c2) + __REQUIRE_CONSTANT(__c3) __REQUIRE_CONSTANT(__c4) + __REQUIRE_CONSTANT(__c5) __REQUIRE_CONSTANT(__c6) + __REQUIRE_CONSTANT(__c7) __REQUIRE_CONSTANT(__c8) + __REQUIRE_CONSTANT(__c9) __REQUIRE_CONSTANT(__c10) + __REQUIRE_CONSTANT(__c11) __REQUIRE_CONSTANT(__c12) + __REQUIRE_CONSTANT(__c13) __REQUIRE_CONSTANT(__c14) + __REQUIRE_CONSTANT(__c15) { + return (v128_t)(__i8x16){__c0, __c1, __c2, __c3, __c4, __c5, + __c6, __c7, __c8, __c9, __c10, __c11, + __c12, __c13, __c14, __c15}; +} + +static __inline__ v128_t __DEFAULT_FN_ATTRS +wasm_i16x8_const(int16_t __c0, int16_t __c1, int16_t __c2, int16_t __c3, + int16_t __c4, int16_t __c5, int16_t __c6, int16_t __c7) + __REQUIRE_CONSTANT(__c0) __REQUIRE_CONSTANT(__c1) __REQUIRE_CONSTANT(__c2) + __REQUIRE_CONSTANT(__c3) __REQUIRE_CONSTANT(__c4) + __REQUIRE_CONSTANT(__c5) __REQUIRE_CONSTANT(__c6) + __REQUIRE_CONSTANT(__c7) { + return (v128_t)(__i16x8){__c0, __c1, __c2, __c3, __c4, __c5, __c6, __c7}; +} + +static __inline__ v128_t __DEFAULT_FN_ATTRS +wasm_i32x4_const(int32_t __c0, int32_t __c1, int32_t __c2, int32_t __c3) + __REQUIRE_CONSTANT(__c0) __REQUIRE_CONSTANT(__c1) __REQUIRE_CONSTANT(__c2) + __REQUIRE_CONSTANT(__c3) { + return (v128_t)(__i32x4){__c0, __c1, __c2, __c3}; +} + +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i64x2_const(int64_t __c0, + int64_t __c1) + __REQUIRE_CONSTANT(__c0) __REQUIRE_CONSTANT(__c1) { return (v128_t)(__i64x2){__c0, __c1}; } -static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f64x2_make(double __c0, - double __c1) { +static __inline__ v128_t __DEFAULT_FN_ATTRS +wasm_f32x4_const(float __c0, float __c1, float __c2, float __c3) + __REQUIRE_CONSTANT(__c0) __REQUIRE_CONSTANT(__c1) __REQUIRE_CONSTANT(__c2) + __REQUIRE_CONSTANT(__c3) { + return (v128_t)(__f32x4){__c0, __c1, __c2, __c3}; +} + +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f64x2_const(double __c0, + double __c1) + __REQUIRE_CONSTANT(__c0) __REQUIRE_CONSTANT(__c1) { return (v128_t)(__f64x2){__c0, __c1}; } -#define wasm_i8x16_const(__c0, __c1, __c2, __c3, __c4, __c5, __c6, __c7, __c8, \ - __c9, __c10, __c11, __c12, __c13, __c14, __c15) \ - __extension__({ \ - __REQUIRE_CONSTANT(__c0); \ - __REQUIRE_CONSTANT(__c1); \ - __REQUIRE_CONSTANT(__c2); \ - __REQUIRE_CONSTANT(__c3); \ - __REQUIRE_CONSTANT(__c4); \ - __REQUIRE_CONSTANT(__c5); \ - __REQUIRE_CONSTANT(__c6); \ - __REQUIRE_CONSTANT(__c7); \ - __REQUIRE_CONSTANT(__c8); \ - __REQUIRE_CONSTANT(__c9); \ - __REQUIRE_CONSTANT(__c10); \ - __REQUIRE_CONSTANT(__c11); \ - __REQUIRE_CONSTANT(__c12); \ - __REQUIRE_CONSTANT(__c13); \ - __REQUIRE_CONSTANT(__c14); \ - __REQUIRE_CONSTANT(__c15); \ - (v128_t)(__i8x16){__c0, __c1, __c2, __c3, __c4, __c5, __c6, __c7, \ - __c8, __c9, __c10, __c11, __c12, __c13, __c14, __c15}; \ - }) - -#define wasm_i16x8_const(__c0, __c1, __c2, __c3, __c4, __c5, __c6, __c7) \ - __extension__({ \ - __REQUIRE_CONSTANT(__c0); \ - __REQUIRE_CONSTANT(__c1); \ - __REQUIRE_CONSTANT(__c2); \ - __REQUIRE_CONSTANT(__c3); \ - __REQUIRE_CONSTANT(__c4); \ - __REQUIRE_CONSTANT(__c5); \ - __REQUIRE_CONSTANT(__c6); \ - __REQUIRE_CONSTANT(__c7); \ - (v128_t)(__i16x8){__c0, __c1, __c2, __c3, __c4, __c5, __c6, __c7}; \ - }) - -#define wasm_i32x4_const(__c0, __c1, __c2, __c3) \ - __extension__({ \ - __REQUIRE_CONSTANT(__c0); \ - __REQUIRE_CONSTANT(__c1); \ - __REQUIRE_CONSTANT(__c2); \ - __REQUIRE_CONSTANT(__c3); \ - (v128_t)(__i32x4){__c0, __c1, __c2, __c3}; \ - }) - -#define wasm_f32x4_const(__c0, __c1, __c2, __c3) \ - __extension__({ \ - __REQUIRE_CONSTANT(__c0); \ - __REQUIRE_CONSTANT(__c1); \ - __REQUIRE_CONSTANT(__c2); \ - __REQUIRE_CONSTANT(__c3); \ - (v128_t)(__f32x4){__c0, __c1, __c2, __c3}; \ - }) - -#define wasm_i64x2_const(__c0, __c1) \ - __extension__({ \ - __REQUIRE_CONSTANT(__c0); \ - __REQUIRE_CONSTANT(__c1); \ - (v128_t)(__i64x2){__c0, __c1}; \ - }) - -#define wasm_f64x2_const(__c0, __c1) \ - __extension__({ \ - __REQUIRE_CONSTANT(__c0); \ - __REQUIRE_CONSTANT(__c1); \ - (v128_t)(__f64x2){__c0, __c1}; \ - }) +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i8x16_const_splat(int8_t __c) + __REQUIRE_CONSTANT(__c) { + return (v128_t)(__i8x16){__c, __c, __c, __c, __c, __c, __c, __c, + __c, __c, __c, __c, __c, __c, __c, __c}; +} + +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i16x8_const_splat(int16_t __c) + __REQUIRE_CONSTANT(__c) { + return (v128_t)(__i16x8){__c, __c, __c, __c, __c, __c, __c, __c}; +} + +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i32x4_const_splat(int32_t __c) + __REQUIRE_CONSTANT(__c) { + return (v128_t)(__i32x4){__c, __c, __c, __c}; +} + +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i64x2_const_splat(int64_t __c) + __REQUIRE_CONSTANT(__c) { + return (v128_t)(__i64x2){__c, __c}; +} + +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f32x4_const_splat(float __c) + __REQUIRE_CONSTANT(__c) { + return (v128_t)(__f32x4){__c, __c, __c, __c}; +} + +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f64x2_const_splat(double __c) + __REQUIRE_CONSTANT(__c) { + return (v128_t)(__f64x2){__c, __c}; +} static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i8x16_splat(int8_t __a) { return (v128_t)(__i8x16){__a, __a, __a, __a, __a, __a, __a, __a, __a, __a, __a, __a, __a, __a, __a, __a}; } -#define wasm_i8x16_extract_lane(__a, __i) \ - (__builtin_wasm_extract_lane_s_i8x16((__i8x16)(__a), __i)) +static __inline__ int8_t __DEFAULT_FN_ATTRS wasm_i8x16_extract_lane(v128_t __a, + int __i) + __REQUIRE_CONSTANT(__i) { + return ((__i8x16)__a)[__i]; +} -#define wasm_u8x16_extract_lane(__a, __i) \ - (__builtin_wasm_extract_lane_u_i8x16((__u8x16)(__a), __i)) +static __inline__ uint8_t __DEFAULT_FN_ATTRS wasm_u8x16_extract_lane(v128_t __a, + int __i) + __REQUIRE_CONSTANT(__i) { + return ((__u8x16)__a)[__i]; +} -#define wasm_i8x16_replace_lane(__a, __i, __b) \ - ((v128_t)__builtin_wasm_replace_lane_i8x16((__i8x16)(__a), __i, __b)) +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i8x16_replace_lane(v128_t __a, + int __i, + int8_t __b) + __REQUIRE_CONSTANT(__i) { + __i8x16 __v = (__i8x16)__a; + __v[__i] = __b; + return (v128_t)__v; +} static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i16x8_splat(int16_t __a) { return (v128_t)(__i16x8){__a, __a, __a, __a, __a, __a, __a, __a}; } -#define wasm_i16x8_extract_lane(__a, __i) \ - (__builtin_wasm_extract_lane_s_i16x8((__i16x8)(__a), __i)) +static __inline__ int16_t __DEFAULT_FN_ATTRS wasm_i16x8_extract_lane(v128_t __a, + int __i) + __REQUIRE_CONSTANT(__i) { + return ((__i16x8)__a)[__i]; +} -#define wasm_u16x8_extract_lane(__a, __i) \ - (__builtin_wasm_extract_lane_u_i16x8((__u16x8)(__a), __i)) +static __inline__ uint16_t __DEFAULT_FN_ATTRS +wasm_u16x8_extract_lane(v128_t __a, int __i) __REQUIRE_CONSTANT(__i) { + return ((__u16x8)__a)[__i]; +} -#define wasm_i16x8_replace_lane(__a, __i, __b) \ - ((v128_t)__builtin_wasm_replace_lane_i16x8((__i16x8)(__a), __i, __b)) +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i16x8_replace_lane(v128_t __a, + int __i, + int16_t __b) + __REQUIRE_CONSTANT(__i) { + __i16x8 __v = (__i16x8)__a; + __v[__i] = __b; + return (v128_t)__v; +} static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i32x4_splat(int32_t __a) { return (v128_t)(__i32x4){__a, __a, __a, __a}; } -#define wasm_i32x4_extract_lane(__a, __i) \ - (__builtin_wasm_extract_lane_i32x4((__i32x4)(__a), __i)) +static __inline__ int32_t __DEFAULT_FN_ATTRS wasm_i32x4_extract_lane(v128_t __a, + int __i) + __REQUIRE_CONSTANT(__i) { + return ((__i32x4)__a)[__i]; +} -#define wasm_i32x4_replace_lane(__a, __i, __b) \ - ((v128_t)__builtin_wasm_replace_lane_i32x4((__i32x4)(__a), __i, __b)) +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i32x4_replace_lane(v128_t __a, + int __i, + int32_t __b) + __REQUIRE_CONSTANT(__i) { + __i32x4 __v = (__i32x4)__a; + __v[__i] = __b; + return (v128_t)__v; +} static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i64x2_splat(int64_t __a) { return (v128_t)(__i64x2){__a, __a}; } -#define wasm_i64x2_extract_lane(__a, __i) \ - (__builtin_wasm_extract_lane_i64x2((__i64x2)(__a), __i)) +static __inline__ int64_t __DEFAULT_FN_ATTRS wasm_i64x2_extract_lane(v128_t __a, + int __i) + __REQUIRE_CONSTANT(__i) { + return ((__i64x2)__a)[__i]; +} -#define wasm_i64x2_replace_lane(__a, __i, __b) \ - ((v128_t)__builtin_wasm_replace_lane_i64x2((__i64x2)(__a), __i, __b)) +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i64x2_replace_lane(v128_t __a, + int __i, + int64_t __b) + __REQUIRE_CONSTANT(__i) { + __i64x2 __v = (__i64x2)__a; + __v[__i] = __b; + return (v128_t)__v; +} static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f32x4_splat(float __a) { return (v128_t)(__f32x4){__a, __a, __a, __a}; } -#define wasm_f32x4_extract_lane(__a, __i) \ - (__builtin_wasm_extract_lane_f32x4((__f32x4)(__a), __i)) +static __inline__ float __DEFAULT_FN_ATTRS wasm_f32x4_extract_lane(v128_t __a, + int __i) + __REQUIRE_CONSTANT(__i) { + return ((__f32x4)__a)[__i]; +} -#define wasm_f32x4_replace_lane(__a, __i, __b) \ - ((v128_t)__builtin_wasm_replace_lane_f32x4((__f32x4)(__a), __i, __b)) +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f32x4_replace_lane(v128_t __a, + int __i, + float __b) + __REQUIRE_CONSTANT(__i) { + __f32x4 __v = (__f32x4)__a; + __v[__i] = __b; + return (v128_t)__v; +} static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f64x2_splat(double __a) { return (v128_t)(__f64x2){__a, __a}; } -#define wasm_f64x2_extract_lane(__a, __i) \ - (__builtin_wasm_extract_lane_f64x2((__f64x2)(__a), __i)) +static __inline__ double __DEFAULT_FN_ATTRS wasm_f64x2_extract_lane(v128_t __a, + int __i) + __REQUIRE_CONSTANT(__i) { + return ((__f64x2)__a)[__i]; +} -#define wasm_f64x2_replace_lane(__a, __i, __b) \ - ((v128_t)__builtin_wasm_replace_lane_f64x2((__f64x2)(__a), __i, __b)) +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f64x2_replace_lane(v128_t __a, + int __i, + double __b) + __REQUIRE_CONSTANT(__i) { + __f64x2 __v = (__f64x2)__a; + __v[__i] = __b; + return (v128_t)__v; +} static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i8x16_eq(v128_t __a, v128_t __b) { @@ -487,6 +667,36 @@ static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_u32x4_ge(v128_t __a, return (v128_t)((__u32x4)__a >= (__u32x4)__b); } +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i64x2_eq(v128_t __a, + v128_t __b) { + return (v128_t)((__i64x2)__a == (__i64x2)__b); +} + +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i64x2_ne(v128_t __a, + v128_t __b) { + return (v128_t)((__i64x2)__a != (__i64x2)__b); +} + +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i64x2_lt(v128_t __a, + v128_t __b) { + return (v128_t)((__i64x2)__a < (__i64x2)__b); +} + +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i64x2_gt(v128_t __a, + v128_t __b) { + return (v128_t)((__i64x2)__a > (__i64x2)__b); +} + +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i64x2_le(v128_t __a, + v128_t __b) { + return (v128_t)((__i64x2)__a <= (__i64x2)__b); +} + +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i64x2_ge(v128_t __a, + v128_t __b) { + return (v128_t)((__i64x2)__a >= (__i64x2)__b); +} + static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f32x4_eq(v128_t __a, v128_t __b) { return (v128_t)((__f32x4)__a == (__f32x4)__b); @@ -571,6 +781,10 @@ static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_v128_andnot(v128_t __a, return __a & ~__b; } +static __inline__ bool __DEFAULT_FN_ATTRS wasm_v128_any_true(v128_t __a) { + return __builtin_wasm_any_true_v128((__i8x16)__a); +} + static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_v128_bitselect(v128_t __a, v128_t __b, v128_t __mask) { @@ -586,14 +800,18 @@ static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i8x16_neg(v128_t __a) { return (v128_t)(-(__u8x16)__a); } -static __inline__ bool __DEFAULT_FN_ATTRS wasm_i8x16_any_true(v128_t __a) { - return __builtin_wasm_any_true_i8x16((__i8x16)__a); -} - static __inline__ bool __DEFAULT_FN_ATTRS wasm_i8x16_all_true(v128_t __a) { return __builtin_wasm_all_true_i8x16((__i8x16)__a); } +static __inline__ int32_t __DEFAULT_FN_ATTRS wasm_i8x16_bitmask(v128_t __a) { + return __builtin_wasm_bitmask_i8x16((__i8x16)__a); +} + +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i8x16_popcnt(v128_t __a) { + return (v128_t)__builtin_wasm_popcnt_i8x16((__i8x16)__a); +} + static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i8x16_shl(v128_t __a, int32_t __b) { return (v128_t)((__i8x16)__a << __b); @@ -614,16 +832,14 @@ static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i8x16_add(v128_t __a, return (v128_t)((__u8x16)__a + (__u8x16)__b); } -static __inline__ v128_t __DEFAULT_FN_ATTRS -wasm_i8x16_add_saturate(v128_t __a, v128_t __b) { - return (v128_t)__builtin_wasm_add_saturate_s_i8x16((__i8x16)__a, - (__i8x16)__b); +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i8x16_add_sat(v128_t __a, + v128_t __b) { + return (v128_t)__builtin_wasm_add_sat_s_i8x16((__i8x16)__a, (__i8x16)__b); } -static __inline__ v128_t __DEFAULT_FN_ATTRS -wasm_u8x16_add_saturate(v128_t __a, v128_t __b) { - return (v128_t)__builtin_wasm_add_saturate_u_i8x16((__u8x16)__a, - (__u8x16)__b); +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_u8x16_add_sat(v128_t __a, + v128_t __b) { + return (v128_t)__builtin_wasm_add_sat_u_i8x16((__u8x16)__a, (__u8x16)__b); } static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i8x16_sub(v128_t __a, @@ -631,16 +847,14 @@ static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i8x16_sub(v128_t __a, return (v128_t)((__u8x16)__a - (__u8x16)__b); } -static __inline__ v128_t __DEFAULT_FN_ATTRS -wasm_i8x16_sub_saturate(v128_t __a, v128_t __b) { - return (v128_t)__builtin_wasm_sub_saturate_s_i8x16((__i8x16)__a, - (__i8x16)__b); +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i8x16_sub_sat(v128_t __a, + v128_t __b) { + return (v128_t)__builtin_wasm_sub_sat_s_i8x16((__i8x16)__a, (__i8x16)__b); } -static __inline__ v128_t __DEFAULT_FN_ATTRS -wasm_u8x16_sub_saturate(v128_t __a, v128_t __b) { - return (v128_t)__builtin_wasm_sub_saturate_u_i8x16((__u8x16)__a, - (__u8x16)__b); +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_u8x16_sub_sat(v128_t __a, + v128_t __b) { + return (v128_t)__builtin_wasm_sub_sat_u_i8x16((__u8x16)__a, (__u8x16)__b); } static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i8x16_min(v128_t __a, @@ -676,14 +890,14 @@ static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i16x8_neg(v128_t __a) { return (v128_t)(-(__u16x8)__a); } -static __inline__ bool __DEFAULT_FN_ATTRS wasm_i16x8_any_true(v128_t __a) { - return __builtin_wasm_any_true_i16x8((__i16x8)__a); -} - static __inline__ bool __DEFAULT_FN_ATTRS wasm_i16x8_all_true(v128_t __a) { return __builtin_wasm_all_true_i16x8((__i16x8)__a); } +static __inline__ int32_t __DEFAULT_FN_ATTRS wasm_i16x8_bitmask(v128_t __a) { + return __builtin_wasm_bitmask_i16x8((__i16x8)__a); +} + static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i16x8_shl(v128_t __a, int32_t __b) { return (v128_t)((__i16x8)__a << __b); @@ -704,16 +918,14 @@ static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i16x8_add(v128_t __a, return (v128_t)((__u16x8)__a + (__u16x8)__b); } -static __inline__ v128_t __DEFAULT_FN_ATTRS -wasm_i16x8_add_saturate(v128_t __a, v128_t __b) { - return (v128_t)__builtin_wasm_add_saturate_s_i16x8((__i16x8)__a, - (__i16x8)__b); +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i16x8_add_sat(v128_t __a, + v128_t __b) { + return (v128_t)__builtin_wasm_add_sat_s_i16x8((__i16x8)__a, (__i16x8)__b); } -static __inline__ v128_t __DEFAULT_FN_ATTRS -wasm_u16x8_add_saturate(v128_t __a, v128_t __b) { - return (v128_t)__builtin_wasm_add_saturate_u_i16x8((__u16x8)__a, - (__u16x8)__b); +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_u16x8_add_sat(v128_t __a, + v128_t __b) { + return (v128_t)__builtin_wasm_add_sat_u_i16x8((__u16x8)__a, (__u16x8)__b); } static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i16x8_sub(v128_t __a, @@ -721,16 +933,14 @@ static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i16x8_sub(v128_t __a, return (v128_t)((__i16x8)__a - (__i16x8)__b); } -static __inline__ v128_t __DEFAULT_FN_ATTRS -wasm_i16x8_sub_saturate(v128_t __a, v128_t __b) { - return (v128_t)__builtin_wasm_sub_saturate_s_i16x8((__i16x8)__a, - (__i16x8)__b); +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i16x8_sub_sat(v128_t __a, + v128_t __b) { + return (v128_t)__builtin_wasm_sub_sat_s_i16x8((__i16x8)__a, (__i16x8)__b); } -static __inline__ v128_t __DEFAULT_FN_ATTRS -wasm_u16x8_sub_saturate(v128_t __a, v128_t __b) { - return (v128_t)__builtin_wasm_sub_saturate_u_i16x8((__u16x8)__a, - (__u16x8)__b); +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_u16x8_sub_sat(v128_t __a, + v128_t __b) { + return (v128_t)__builtin_wasm_sub_sat_u_i16x8((__u16x8)__a, (__u16x8)__b); } static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i16x8_mul(v128_t __a, @@ -771,14 +981,14 @@ static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i32x4_neg(v128_t __a) { return (v128_t)(-(__u32x4)__a); } -static __inline__ bool __DEFAULT_FN_ATTRS wasm_i32x4_any_true(v128_t __a) { - return __builtin_wasm_any_true_i32x4((__i32x4)__a); -} - static __inline__ bool __DEFAULT_FN_ATTRS wasm_i32x4_all_true(v128_t __a) { return __builtin_wasm_all_true_i32x4((__i32x4)__a); } +static __inline__ int32_t __DEFAULT_FN_ATTRS wasm_i32x4_bitmask(v128_t __a) { + return __builtin_wasm_bitmask_i32x4((__i32x4)__a); +} + static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i32x4_shl(v128_t __a, int32_t __b) { return (v128_t)((__i32x4)__a << __b); @@ -829,21 +1039,26 @@ static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_u32x4_max(v128_t __a, return (v128_t)__builtin_wasm_max_u_i32x4((__u32x4)__a, (__u32x4)__b); } -static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i64x2_neg(v128_t __a) { - return (v128_t)(-(__u64x2)__a); +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i32x4_dot_i16x8(v128_t __a, + v128_t __b) { + return (v128_t)__builtin_wasm_dot_s_i32x4_i16x8((__i16x8)__a, (__i16x8)__b); } -#ifdef __wasm_unimplemented_simd128__ +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i64x2_abs(v128_t __a) { + return (v128_t)__builtin_wasm_abs_i64x2((__i64x2)__a); +} -static __inline__ bool __DEFAULT_FN_ATTRS wasm_i64x2_any_true(v128_t __a) { - return __builtin_wasm_any_true_i64x2((__i64x2)__a); +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i64x2_neg(v128_t __a) { + return (v128_t)(-(__u64x2)__a); } static __inline__ bool __DEFAULT_FN_ATTRS wasm_i64x2_all_true(v128_t __a) { return __builtin_wasm_all_true_i64x2((__i64x2)__a); } -#endif // __wasm_unimplemented_simd128__ +static __inline__ int32_t __DEFAULT_FN_ATTRS wasm_i64x2_bitmask(v128_t __a) { + return __builtin_wasm_bitmask_i64x2((__i64x2)__a); +} static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i64x2_shl(v128_t __a, int32_t __b) { @@ -887,23 +1102,21 @@ static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f32x4_sqrt(v128_t __a) { return (v128_t)__builtin_wasm_sqrt_f32x4((__f32x4)__a); } -#ifdef __wasm_unimplemented_simd128__ +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f32x4_ceil(v128_t __a) { + return (v128_t)__builtin_wasm_ceil_f32x4((__f32x4)__a); +} -static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f32x4_qfma(v128_t __a, - v128_t __b, - v128_t __c) { - return (v128_t)__builtin_wasm_qfma_f32x4((__f32x4)__a, (__f32x4)__b, - (__f32x4)__c); +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f32x4_floor(v128_t __a) { + return (v128_t)__builtin_wasm_floor_f32x4((__f32x4)__a); } -static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f32x4_qfms(v128_t __a, - v128_t __b, - v128_t __c) { - return (v128_t)__builtin_wasm_qfms_f32x4((__f32x4)__a, (__f32x4)__b, - (__f32x4)__c); +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f32x4_trunc(v128_t __a) { + return (v128_t)__builtin_wasm_trunc_f32x4((__f32x4)__a); } -#endif // __wasm_unimplemented_simd128__ +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f32x4_nearest(v128_t __a) { + return (v128_t)__builtin_wasm_nearest_f32x4((__f32x4)__a); +} static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f32x4_add(v128_t __a, v128_t __b) { @@ -937,12 +1150,14 @@ static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f32x4_max(v128_t __a, static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f32x4_pmin(v128_t __a, v128_t __b) { - return (v128_t)__builtin_wasm_pmin_f32x4((__f32x4)__a, (__f32x4)__b); + __i32x4 __mask = (__i32x4)((__f32x4)__b < (__f32x4)__a); + return (v128_t)((((__i32x4)__b) & __mask) | (((__i32x4)__a) & ~__mask)); } static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f32x4_pmax(v128_t __a, v128_t __b) { - return (v128_t)__builtin_wasm_pmax_f32x4((__f32x4)__a, (__f32x4)__b); + __i32x4 __mask = (__i32x4)((__f32x4)__a < (__f32x4)__b); + return (v128_t)((((__i32x4)__b) & __mask) | (((__i32x4)__a) & ~__mask)); } static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f64x2_abs(v128_t __a) { @@ -957,23 +1172,21 @@ static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f64x2_sqrt(v128_t __a) { return (v128_t)__builtin_wasm_sqrt_f64x2((__f64x2)__a); } -#ifdef __wasm_unimplemented_simd128__ +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f64x2_ceil(v128_t __a) { + return (v128_t)__builtin_wasm_ceil_f64x2((__f64x2)__a); +} -static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f64x2_qfma(v128_t __a, - v128_t __b, - v128_t __c) { - return (v128_t)__builtin_wasm_qfma_f64x2((__f64x2)__a, (__f64x2)__b, - (__f64x2)__c); +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f64x2_floor(v128_t __a) { + return (v128_t)__builtin_wasm_floor_f64x2((__f64x2)__a); } -static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f64x2_qfms(v128_t __a, - v128_t __b, - v128_t __c) { - return (v128_t)__builtin_wasm_qfms_f64x2((__f64x2)__a, (__f64x2)__b, - (__f64x2)__c); +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f64x2_trunc(v128_t __a) { + return (v128_t)__builtin_wasm_trunc_f64x2((__f64x2)__a); } -#endif // __wasm_unimplemented_simd128__ +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f64x2_nearest(v128_t __a) { + return (v128_t)__builtin_wasm_nearest_f64x2((__f64x2)__a); +} static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f64x2_add(v128_t __a, v128_t __b) { @@ -1007,21 +1220,23 @@ static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f64x2_max(v128_t __a, static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f64x2_pmin(v128_t __a, v128_t __b) { - return (v128_t)__builtin_wasm_pmin_f64x2((__f64x2)__a, (__f64x2)__b); + __i64x2 __mask = (__i64x2)((__f64x2)__b < (__f64x2)__a); + return (v128_t)((((__i64x2)__b) & __mask) | (((__i64x2)__a) & ~__mask)); } static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_f64x2_pmax(v128_t __a, v128_t __b) { - return (v128_t)__builtin_wasm_pmax_f64x2((__f64x2)__a, (__f64x2)__b); + __i64x2 __mask = (__i64x2)((__f64x2)__a < (__f64x2)__b); + return (v128_t)((((__i64x2)__b) & __mask) | (((__i64x2)__a) & ~__mask)); } static __inline__ v128_t __DEFAULT_FN_ATTRS -wasm_i32x4_trunc_saturate_f32x4(v128_t __a) { +wasm_i32x4_trunc_sat_f32x4(v128_t __a) { return (v128_t)__builtin_wasm_trunc_saturate_s_i32x4_f32x4((__f32x4)__a); } static __inline__ v128_t __DEFAULT_FN_ATTRS -wasm_u32x4_trunc_saturate_f32x4(v128_t __a) { +wasm_u32x4_trunc_sat_f32x4(v128_t __a) { return (v128_t)__builtin_wasm_trunc_saturate_u_i32x4_f32x4((__f32x4)__a); } @@ -1035,38 +1250,71 @@ wasm_f32x4_convert_u32x4(v128_t __a) { return (v128_t) __builtin_convertvector((__u32x4)__a, __f32x4); } -#define wasm_v8x16_shuffle(__a, __b, __c0, __c1, __c2, __c3, __c4, __c5, __c6, \ +static __inline__ v128_t __DEFAULT_FN_ATTRS +wasm_f64x2_convert_low_i32x4(v128_t __a) { + return (v128_t) __builtin_convertvector((__i32x2){__a[0], __a[1]}, __f64x2); +} + +static __inline__ v128_t __DEFAULT_FN_ATTRS +wasm_f64x2_convert_low_u32x4(v128_t __a) { + return (v128_t) __builtin_convertvector((__u32x2){__a[0], __a[1]}, __f64x2); +} + +static __inline__ v128_t __DEFAULT_FN_ATTRS +wasm_i32x4_trunc_sat_f64x2_zero(v128_t __a) { + return (v128_t)__builtin_wasm_trunc_sat_zero_s_f64x2_i32x4((__f64x2)__a); +} + +static __inline__ v128_t __DEFAULT_FN_ATTRS +wasm_u32x4_trunc_sat_f64x2_zero(v128_t __a) { + return (v128_t)__builtin_wasm_trunc_sat_zero_u_f64x2_i32x4((__f64x2)__a); +} + +static __inline__ v128_t __DEFAULT_FN_ATTRS +wasm_f32x4_demote_f64x2_zero(v128_t __a) { + return (v128_t) __builtin_convertvector( + __builtin_shufflevector((__f64x2)__a, (__f64x2){0, 0}, 0, 1, 2, 3), + __f32x4); +} + +static __inline__ v128_t __DEFAULT_FN_ATTRS +wasm_f64x2_promote_low_f32x4(v128_t __a) { + return (v128_t) __builtin_convertvector( + (__f32x2){((__f32x4)__a)[0], ((__f32x4)__a)[1]}, __f64x2); +} + +#define wasm_i8x16_shuffle(__a, __b, __c0, __c1, __c2, __c3, __c4, __c5, __c6, \ __c7, __c8, __c9, __c10, __c11, __c12, __c13, \ __c14, __c15) \ - ((v128_t)__builtin_wasm_shuffle_v8x16( \ + ((v128_t)__builtin_wasm_shuffle_i8x16( \ (__i8x16)(__a), (__i8x16)(__b), __c0, __c1, __c2, __c3, __c4, __c5, \ __c6, __c7, __c8, __c9, __c10, __c11, __c12, __c13, __c14, __c15)) -#define wasm_v16x8_shuffle(__a, __b, __c0, __c1, __c2, __c3, __c4, __c5, __c6, \ +#define wasm_i16x8_shuffle(__a, __b, __c0, __c1, __c2, __c3, __c4, __c5, __c6, \ __c7) \ - ((v128_t)__builtin_wasm_shuffle_v8x16( \ + ((v128_t)__builtin_wasm_shuffle_i8x16( \ (__i8x16)(__a), (__i8x16)(__b), (__c0)*2, (__c0)*2 + 1, (__c1)*2, \ (__c1)*2 + 1, (__c2)*2, (__c2)*2 + 1, (__c3)*2, (__c3)*2 + 1, (__c4)*2, \ (__c4)*2 + 1, (__c5)*2, (__c5)*2 + 1, (__c6)*2, (__c6)*2 + 1, (__c7)*2, \ (__c7)*2 + 1)) -#define wasm_v32x4_shuffle(__a, __b, __c0, __c1, __c2, __c3) \ - ((v128_t)__builtin_wasm_shuffle_v8x16( \ +#define wasm_i32x4_shuffle(__a, __b, __c0, __c1, __c2, __c3) \ + ((v128_t)__builtin_wasm_shuffle_i8x16( \ (__i8x16)(__a), (__i8x16)(__b), (__c0)*4, (__c0)*4 + 1, (__c0)*4 + 2, \ (__c0)*4 + 3, (__c1)*4, (__c1)*4 + 1, (__c1)*4 + 2, (__c1)*4 + 3, \ (__c2)*4, (__c2)*4 + 1, (__c2)*4 + 2, (__c2)*4 + 3, (__c3)*4, \ (__c3)*4 + 1, (__c3)*4 + 2, (__c3)*4 + 3)) -#define wasm_v64x2_shuffle(__a, __b, __c0, __c1) \ - ((v128_t)__builtin_wasm_shuffle_v8x16( \ +#define wasm_i64x2_shuffle(__a, __b, __c0, __c1) \ + ((v128_t)__builtin_wasm_shuffle_i8x16( \ (__i8x16)(__a), (__i8x16)(__b), (__c0)*8, (__c0)*8 + 1, (__c0)*8 + 2, \ (__c0)*8 + 3, (__c0)*8 + 4, (__c0)*8 + 5, (__c0)*8 + 6, (__c0)*8 + 7, \ (__c1)*8, (__c1)*8 + 1, (__c1)*8 + 2, (__c1)*8 + 3, (__c1)*8 + 4, \ (__c1)*8 + 5, (__c1)*8 + 6, (__c1)*8 + 7)) -static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_v8x16_swizzle(v128_t __a, +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i8x16_swizzle(v128_t __a, v128_t __b) { - return (v128_t)__builtin_wasm_swizzle_v8x16((__i8x16)__a, (__i8x16)__b); + return (v128_t)__builtin_wasm_swizzle_i8x16((__i8x16)__a, (__i8x16)__b); } static __inline__ v128_t __DEFAULT_FN_ATTRS @@ -1077,8 +1325,8 @@ wasm_i8x16_narrow_i16x8(v128_t __a, v128_t __b) { static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_u8x16_narrow_i16x8(v128_t __a, v128_t __b) { - return (v128_t)__builtin_wasm_narrow_u_i8x16_i16x8((__u16x8)__a, - (__u16x8)__b); + return (v128_t)__builtin_wasm_narrow_u_i8x16_i16x8((__i16x8)__a, + (__i16x8)__b); } static __inline__ v128_t __DEFAULT_FN_ATTRS @@ -1089,12 +1337,12 @@ wasm_i16x8_narrow_i32x4(v128_t __a, v128_t __b) { static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_u16x8_narrow_i32x4(v128_t __a, v128_t __b) { - return (v128_t)__builtin_wasm_narrow_u_i16x8_i32x4((__u32x4)__a, - (__u32x4)__b); + return (v128_t)__builtin_wasm_narrow_u_i16x8_i32x4((__i32x4)__a, + (__i32x4)__b); } static __inline__ v128_t __DEFAULT_FN_ATTRS -wasm_i16x8_widen_low_i8x16(v128_t __a) { +wasm_i16x8_extend_low_i8x16(v128_t __a) { return (v128_t) __builtin_convertvector( (__i8x8){((__i8x16)__a)[0], ((__i8x16)__a)[1], ((__i8x16)__a)[2], ((__i8x16)__a)[3], ((__i8x16)__a)[4], ((__i8x16)__a)[5], @@ -1103,7 +1351,7 @@ wasm_i16x8_widen_low_i8x16(v128_t __a) { } static __inline__ v128_t __DEFAULT_FN_ATTRS -wasm_i16x8_widen_high_i8x16(v128_t __a) { +wasm_i16x8_extend_high_i8x16(v128_t __a) { return (v128_t) __builtin_convertvector( (__i8x8){((__i8x16)__a)[8], ((__i8x16)__a)[9], ((__i8x16)__a)[10], ((__i8x16)__a)[11], ((__i8x16)__a)[12], ((__i8x16)__a)[13], @@ -1112,7 +1360,7 @@ wasm_i16x8_widen_high_i8x16(v128_t __a) { } static __inline__ v128_t __DEFAULT_FN_ATTRS -wasm_i16x8_widen_low_u8x16(v128_t __a) { +wasm_u16x8_extend_low_u8x16(v128_t __a) { return (v128_t) __builtin_convertvector( (__u8x8){((__u8x16)__a)[0], ((__u8x16)__a)[1], ((__u8x16)__a)[2], ((__u8x16)__a)[3], ((__u8x16)__a)[4], ((__u8x16)__a)[5], @@ -1121,7 +1369,7 @@ wasm_i16x8_widen_low_u8x16(v128_t __a) { } static __inline__ v128_t __DEFAULT_FN_ATTRS -wasm_i16x8_widen_high_u8x16(v128_t __a) { +wasm_u16x8_extend_high_u8x16(v128_t __a) { return (v128_t) __builtin_convertvector( (__u8x8){((__u8x16)__a)[8], ((__u8x16)__a)[9], ((__u8x16)__a)[10], ((__u8x16)__a)[11], ((__u8x16)__a)[12], ((__u8x16)__a)[13], @@ -1130,7 +1378,7 @@ wasm_i16x8_widen_high_u8x16(v128_t __a) { } static __inline__ v128_t __DEFAULT_FN_ATTRS -wasm_i32x4_widen_low_i16x8(v128_t __a) { +wasm_i32x4_extend_low_i16x8(v128_t __a) { return (v128_t) __builtin_convertvector( (__i16x4){((__i16x8)__a)[0], ((__i16x8)__a)[1], ((__i16x8)__a)[2], ((__i16x8)__a)[3]}, @@ -1138,7 +1386,7 @@ wasm_i32x4_widen_low_i16x8(v128_t __a) { } static __inline__ v128_t __DEFAULT_FN_ATTRS -wasm_i32x4_widen_high_i16x8(v128_t __a) { +wasm_i32x4_extend_high_i16x8(v128_t __a) { return (v128_t) __builtin_convertvector( (__i16x4){((__i16x8)__a)[4], ((__i16x8)__a)[5], ((__i16x8)__a)[6], ((__i16x8)__a)[7]}, @@ -1146,7 +1394,7 @@ wasm_i32x4_widen_high_i16x8(v128_t __a) { } static __inline__ v128_t __DEFAULT_FN_ATTRS -wasm_i32x4_widen_low_u16x8(v128_t __a) { +wasm_u32x4_extend_low_u16x8(v128_t __a) { return (v128_t) __builtin_convertvector( (__u16x4){((__u16x8)__a)[0], ((__u16x8)__a)[1], ((__u16x8)__a)[2], ((__u16x8)__a)[3]}, @@ -1154,14 +1402,333 @@ wasm_i32x4_widen_low_u16x8(v128_t __a) { } static __inline__ v128_t __DEFAULT_FN_ATTRS -wasm_i32x4_widen_high_u16x8(v128_t __a) { +wasm_u32x4_extend_high_u16x8(v128_t __a) { return (v128_t) __builtin_convertvector( (__u16x4){((__u16x8)__a)[4], ((__u16x8)__a)[5], ((__u16x8)__a)[6], ((__u16x8)__a)[7]}, __u32x4); } +static __inline__ v128_t __DEFAULT_FN_ATTRS +wasm_i64x2_extend_low_i32x4(v128_t __a) { + return (v128_t) __builtin_convertvector( + (__i32x2){((__i32x4)__a)[0], ((__i32x4)__a)[1]}, __i64x2); +} + +static __inline__ v128_t __DEFAULT_FN_ATTRS +wasm_i64x2_extend_high_i32x4(v128_t __a) { + return (v128_t) __builtin_convertvector( + (__i32x2){((__i32x4)__a)[2], ((__i32x4)__a)[3]}, __i64x2); +} + +static __inline__ v128_t __DEFAULT_FN_ATTRS +wasm_u64x2_extend_low_u32x4(v128_t __a) { + return (v128_t) __builtin_convertvector( + (__u32x2){((__u32x4)__a)[0], ((__u32x4)__a)[1]}, __u64x2); +} + +static __inline__ v128_t __DEFAULT_FN_ATTRS +wasm_u64x2_extend_high_u32x4(v128_t __a) { + return (v128_t) __builtin_convertvector( + (__u32x2){((__u32x4)__a)[2], ((__u32x4)__a)[3]}, __u64x2); +} + +static __inline__ v128_t __DEFAULT_FN_ATTRS +wasm_i16x8_extadd_pairwise_i8x16(v128_t __a) { + return (v128_t)__builtin_wasm_extadd_pairwise_i8x16_s_i16x8((__i8x16)__a); +} + +static __inline__ v128_t __DEFAULT_FN_ATTRS +wasm_u16x8_extadd_pairwise_u8x16(v128_t __a) { + return (v128_t)__builtin_wasm_extadd_pairwise_i8x16_u_i16x8((__u8x16)__a); +} + +static __inline__ v128_t __DEFAULT_FN_ATTRS +wasm_i32x4_extadd_pairwise_i16x8(v128_t __a) { + return (v128_t)__builtin_wasm_extadd_pairwise_i16x8_s_i32x4((__i16x8)__a); +} + +static __inline__ v128_t __DEFAULT_FN_ATTRS +wasm_u32x4_extadd_pairwise_u16x8(v128_t __a) { + return (v128_t)__builtin_wasm_extadd_pairwise_i16x8_u_i32x4((__u16x8)__a); +} + +static __inline__ v128_t __DEFAULT_FN_ATTRS +wasm_i16x8_extmul_low_i8x16(v128_t __a, v128_t __b) { + return (v128_t)((__i16x8)wasm_i16x8_extend_low_i8x16(__a) * + (__i16x8)wasm_i16x8_extend_low_i8x16(__b)); +} + +static __inline__ v128_t __DEFAULT_FN_ATTRS +wasm_i16x8_extmul_high_i8x16(v128_t __a, v128_t __b) { + return (v128_t)((__i16x8)wasm_i16x8_extend_high_i8x16(__a) * + (__i16x8)wasm_i16x8_extend_high_i8x16(__b)); +} + +static __inline__ v128_t __DEFAULT_FN_ATTRS +wasm_u16x8_extmul_low_u8x16(v128_t __a, v128_t __b) { + return (v128_t)((__u16x8)wasm_u16x8_extend_low_u8x16(__a) * + (__u16x8)wasm_u16x8_extend_low_u8x16(__b)); +} + +static __inline__ v128_t __DEFAULT_FN_ATTRS +wasm_u16x8_extmul_high_u8x16(v128_t __a, v128_t __b) { + return (v128_t)((__u16x8)wasm_u16x8_extend_high_u8x16(__a) * + (__u16x8)wasm_u16x8_extend_high_u8x16(__b)); +} + +static __inline__ v128_t __DEFAULT_FN_ATTRS +wasm_i32x4_extmul_low_i16x8(v128_t __a, v128_t __b) { + return (v128_t)((__i32x4)wasm_i32x4_extend_low_i16x8(__a) * + (__i32x4)wasm_i32x4_extend_low_i16x8(__b)); +} + +static __inline__ v128_t __DEFAULT_FN_ATTRS +wasm_i32x4_extmul_high_i16x8(v128_t __a, v128_t __b) { + return (v128_t)((__i32x4)wasm_i32x4_extend_high_i16x8(__a) * + (__i32x4)wasm_i32x4_extend_high_i16x8(__b)); +} + +static __inline__ v128_t __DEFAULT_FN_ATTRS +wasm_u32x4_extmul_low_u16x8(v128_t __a, v128_t __b) { + return (v128_t)((__u32x4)wasm_u32x4_extend_low_u16x8(__a) * + (__u32x4)wasm_u32x4_extend_low_u16x8(__b)); +} + +static __inline__ v128_t __DEFAULT_FN_ATTRS +wasm_u32x4_extmul_high_u16x8(v128_t __a, v128_t __b) { + return (v128_t)((__u32x4)wasm_u32x4_extend_high_u16x8(__a) * + (__u32x4)wasm_u32x4_extend_high_u16x8(__b)); +} + +static __inline__ v128_t __DEFAULT_FN_ATTRS +wasm_i64x2_extmul_low_i32x4(v128_t __a, v128_t __b) { + return (v128_t)((__i64x2)wasm_i64x2_extend_low_i32x4(__a) * + (__i64x2)wasm_i64x2_extend_low_i32x4(__b)); +} + +static __inline__ v128_t __DEFAULT_FN_ATTRS +wasm_i64x2_extmul_high_i32x4(v128_t __a, v128_t __b) { + return (v128_t)((__i64x2)wasm_i64x2_extend_high_i32x4(__a) * + (__i64x2)wasm_i64x2_extend_high_i32x4(__b)); +} + +static __inline__ v128_t __DEFAULT_FN_ATTRS +wasm_u64x2_extmul_low_u32x4(v128_t __a, v128_t __b) { + return (v128_t)((__u64x2)wasm_u64x2_extend_low_u32x4(__a) * + (__u64x2)wasm_u64x2_extend_low_u32x4(__b)); +} + +static __inline__ v128_t __DEFAULT_FN_ATTRS +wasm_u64x2_extmul_high_u32x4(v128_t __a, v128_t __b) { + return (v128_t)((__u64x2)wasm_u64x2_extend_high_u32x4(__a) * + (__u64x2)wasm_u64x2_extend_high_u32x4(__b)); +} + +static __inline__ v128_t __DEFAULT_FN_ATTRS wasm_i16x8_q15mulr_sat(v128_t __a, + v128_t __b) { + return (v128_t)__builtin_wasm_q15mulr_sat_s_i16x8((__i16x8)__a, (__i16x8)__b); +} + +// Old intrinsic names supported to ease transitioning to the standard names. Do +// not use these; they will be removed in the near future. + +#define __DEPRECATED_FN_ATTRS(__replacement) \ + __DEFAULT_FN_ATTRS __attribute__( \ + (deprecated("use " __replacement " instead", __replacement))) + +#define __WASM_STR(X) #X + +#ifdef __DEPRECATED +#define __DEPRECATED_WASM_MACRO(__name, __replacement) \ + _Pragma(__WASM_STR(GCC warning( \ + "'" __name "' is deprecated: use '" __replacement "' instead"))) +#else +#define __DEPRECATED_WASM_MACRO(__name, __replacement) +#endif + +static __inline__ v128_t __DEPRECATED_FN_ATTRS("wasm_v128_load8_splat") +wasm_v8x16_load_splat(const void *__mem) { + return wasm_v128_load8_splat(__mem); +} + +static __inline__ v128_t __DEPRECATED_FN_ATTRS("wasm_v128_load16_splat") +wasm_v16x8_load_splat(const void *__mem) { + return wasm_v128_load16_splat(__mem); +} + +static __inline__ v128_t __DEPRECATED_FN_ATTRS("wasm_v128_load32_splat") +wasm_v32x4_load_splat(const void *__mem) { + return wasm_v128_load32_splat(__mem); +} + +static __inline__ v128_t __DEPRECATED_FN_ATTRS("wasm_v128_load64_splat") +wasm_v64x2_load_splat(const void *__mem) { + return wasm_v128_load64_splat(__mem); +} + +static __inline__ v128_t __DEPRECATED_FN_ATTRS("wasm_i16x8_load8x8") +wasm_i16x8_load_8x8(const void *__mem) { + return wasm_i16x8_load8x8(__mem); +} + +static __inline__ v128_t __DEPRECATED_FN_ATTRS("wasm_u16x8_load8x8") +wasm_u16x8_load_8x8(const void *__mem) { + return wasm_u16x8_load8x8(__mem); +} + +static __inline__ v128_t __DEPRECATED_FN_ATTRS("wasm_i32x4_load16x4") +wasm_i32x4_load_16x4(const void *__mem) { + return wasm_i32x4_load16x4(__mem); +} + +static __inline__ v128_t __DEPRECATED_FN_ATTRS("wasm_u32x4_load16x4") +wasm_u32x4_load_16x4(const void *__mem) { + return wasm_u32x4_load16x4(__mem); +} + +static __inline__ v128_t __DEPRECATED_FN_ATTRS("wasm_i64x2_load32x2") +wasm_i64x2_load_32x2(const void *__mem) { + return wasm_i64x2_load32x2(__mem); +} + +static __inline__ v128_t __DEPRECATED_FN_ATTRS("wasm_u64x2_load32x2") +wasm_u64x2_load_32x2(const void *__mem) { + return wasm_u64x2_load32x2(__mem); +} + +#define wasm_v8x16_shuffle(__a, __b, __c0, __c1, __c2, __c3, __c4, __c5, __c6, \ + __c7, __c8, __c9, __c10, __c11, __c12, __c13, \ + __c14, __c15) \ + __DEPRECATED_WASM_MACRO("wasm_v8x16_shuffle", "wasm_i8x16_shuffle") \ + wasm_i8x16_shuffle(__a, __b, __c0, __c1, __c2, __c3, __c4, __c5, __c6, __c7, \ + __c8, __c9, __c10, __c11, __c12, __c13, __c14, __c15) + +#define wasm_v16x8_shuffle(__a, __b, __c0, __c1, __c2, __c3, __c4, __c5, __c6, \ + __c7) \ + __DEPRECATED_WASM_MACRO("wasm_v16x8_shuffle", "wasm_i16x8_shuffle") \ + wasm_i16x8_shuffle(__a, __b, __c0, __c1, __c2, __c3, __c4, __c5, __c6, __c7) + +#define wasm_v32x4_shuffle(__a, __b, __c0, __c1, __c2, __c3) \ + __DEPRECATED_WASM_MACRO("wasm_v32x4_shuffle", "wasm_i32x4_shuffle") \ + wasm_i32x4_shuffle(__a, __b, __c0, __c1, __c2, __c3) + +#define wasm_v64x2_shuffle(__a, __b, __c0, __c1) \ + __DEPRECATED_WASM_MACRO("wasm_v64x2_shuffle", "wasm_i64x2_shuffle") \ + wasm_i64x2_shuffle(__a, __b, __c0, __c1) + +static __inline__ v128_t __DEPRECATED_FN_ATTRS("wasm_i8x16_swizzle") +wasm_v8x16_swizzle(v128_t __a, v128_t __b) { + return wasm_i8x16_swizzle(__a, __b); +} + +static __inline__ bool __DEPRECATED_FN_ATTRS("wasm_v128_any_true") +wasm_i8x16_any_true(v128_t __a) { + return wasm_v128_any_true(__a); +} + +static __inline__ bool __DEPRECATED_FN_ATTRS("wasm_v128_any_true") +wasm_i16x8_any_true(v128_t __a) { + return wasm_v128_any_true(__a); +} + +static __inline__ bool __DEPRECATED_FN_ATTRS("wasm_v128_any_true") +wasm_i32x4_any_true(v128_t __a) { + return wasm_v128_any_true(__a); +} + +static __inline__ v128_t __DEPRECATED_FN_ATTRS("wasm_i8x16_add_sat") +wasm_i8x16_add_saturate(v128_t __a, v128_t __b) { + return wasm_i8x16_add_sat(__a, __b); +} + +static __inline__ v128_t __DEPRECATED_FN_ATTRS("wasm_u8x16_add_sat") +wasm_u8x16_add_saturate(v128_t __a, v128_t __b) { + return wasm_u8x16_add_sat(__a, __b); +} + +static __inline__ v128_t __DEPRECATED_FN_ATTRS("wasm_i8x16_sub_sat") +wasm_i8x16_sub_saturate(v128_t __a, v128_t __b) { + return wasm_i8x16_sub_sat(__a, __b); +} + +static __inline__ v128_t __DEPRECATED_FN_ATTRS("wasm_u8x16_sub_sat") +wasm_u8x16_sub_saturate(v128_t __a, v128_t __b) { + return wasm_u8x16_sub_sat(__a, __b); +} + +static __inline__ v128_t __DEPRECATED_FN_ATTRS("wasm_i16x8_add_sat") +wasm_i16x8_add_saturate(v128_t __a, v128_t __b) { + return wasm_i16x8_add_sat(__a, __b); +} + +static __inline__ v128_t __DEPRECATED_FN_ATTRS("wasm_u16x8_add_sat") +wasm_u16x8_add_saturate(v128_t __a, v128_t __b) { + return wasm_u16x8_add_sat(__a, __b); +} + +static __inline__ v128_t __DEPRECATED_FN_ATTRS("wasm_i16x8_sub_sat") +wasm_i16x8_sub_saturate(v128_t __a, v128_t __b) { + return wasm_i16x8_sub_sat(__a, __b); +} + +static __inline__ v128_t __DEPRECATED_FN_ATTRS("wasm_u16x8_sub_sat") +wasm_u16x8_sub_saturate(v128_t __a, v128_t __b) { + return wasm_u16x8_sub_sat(__a, __b); +} + +static __inline__ v128_t __DEPRECATED_FN_ATTRS("wasm_i16x8_extend_low_i8x16") +wasm_i16x8_widen_low_i8x16(v128_t __a) { + return wasm_i16x8_extend_low_i8x16(__a); +} + +static __inline__ v128_t __DEPRECATED_FN_ATTRS("wasm_i16x8_extend_high_i8x16") +wasm_i16x8_widen_high_i8x16(v128_t __a) { + return wasm_i16x8_extend_high_i8x16(__a); +} + +static __inline__ v128_t __DEPRECATED_FN_ATTRS("wasm_u16x8_extend_low_u8x16") +wasm_i16x8_widen_low_u8x16(v128_t __a) { + return wasm_u16x8_extend_low_u8x16(__a); +} + +static __inline__ v128_t __DEPRECATED_FN_ATTRS("wasm_u16x8_extend_high_u8x16") +wasm_i16x8_widen_high_u8x16(v128_t __a) { + return wasm_u16x8_extend_high_u8x16(__a); +} + +static __inline__ v128_t __DEPRECATED_FN_ATTRS("wasm_i32x4_extend_low_i16x8") +wasm_i32x4_widen_low_i16x8(v128_t __a) { + return wasm_i32x4_extend_low_i16x8(__a); +} + +static __inline__ v128_t __DEPRECATED_FN_ATTRS("wasm_i32x4_extend_high_i16x8") +wasm_i32x4_widen_high_i16x8(v128_t __a) { + return wasm_i32x4_extend_high_i16x8(__a); +} + +static __inline__ v128_t __DEPRECATED_FN_ATTRS("wasm_u32x4_extend_low_u16x8") +wasm_i32x4_widen_low_u16x8(v128_t __a) { + return wasm_u32x4_extend_low_u16x8(__a); +} + +static __inline__ v128_t __DEPRECATED_FN_ATTRS("wasm_u32x4_extend_high_u16x8") +wasm_i32x4_widen_high_u16x8(v128_t __a) { + return wasm_u32x4_extend_high_u16x8(__a); +} + +static __inline__ v128_t __DEPRECATED_FN_ATTRS("wasm_i32x4_trunc_sat_f32x4") +wasm_i32x4_trunc_saturate_f32x4(v128_t __a) { + return wasm_i32x4_trunc_sat_f32x4(__a); +} + +static __inline__ v128_t __DEPRECATED_FN_ATTRS("wasm_u32x4_trunc_sat_f32x4") +wasm_u32x4_trunc_saturate_f32x4(v128_t __a) { + return wasm_u32x4_trunc_sat_f32x4(__a); +} + // Undefine helper macros #undef __DEFAULT_FN_ATTRS +#undef __DEPRECATED_FN_ATTRS #endif // __WASM_SIMD128_H diff --git a/lib/libc/include/aarch64-linux-gnu/bits/floatn.h b/lib/libc/include/aarch64-linux-gnu/bits/floatn.h new file mode 100644 index 000000000..0b1b5d444 --- /dev/null +++ b/lib/libc/include/aarch64-linux-gnu/bits/floatn.h @@ -0,0 +1,97 @@ +/* Macros to control TS 18661-3 glibc features on ldbl-128 platforms. + Copyright (C) 2017-2021 Free Software Foundation, Inc. + This file is part of the GNU C Library. + + The GNU C Library is free software; you can redistribute it and/or + modify it under the terms of the GNU Lesser General Public + License as published by the Free Software Foundation; either + version 2.1 of the License, or (at your option) any later version. + + The GNU C Library is distributed in the hope that it will be useful, + but WITHOUT ANY WARRANTY; without even the implied warranty of + MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU + Lesser General Public License for more details. + + You should have received a copy of the GNU Lesser General Public + License along with the GNU C Library; if not, see + . */ + +#ifndef _BITS_FLOATN_H +#define _BITS_FLOATN_H + +#include +#include + +/* Defined to 1 if the current compiler invocation provides a + floating-point type with the IEEE 754 binary128 format, and this + glibc includes corresponding *f128 interfaces for it. */ +#ifndef __NO_LONG_DOUBLE_MATH +# define __HAVE_FLOAT128 1 +#else +/* glibc does not support _Float128 for platforms where long double is + normally binary128 when building with long double as binary64. + GCC's default for supported scalar modes does not support it either + in that case. */ +# define __HAVE_FLOAT128 0 +#endif + +/* Defined to 1 if __HAVE_FLOAT128 is 1 and the type is ABI-distinct + from the default float, double and long double types in this glibc. */ +#define __HAVE_DISTINCT_FLOAT128 0 + +/* Defined to 1 if the current compiler invocation provides a + floating-point type with the right format for _Float64x, and this + glibc includes corresponding *f64x interfaces for it. */ +#define __HAVE_FLOAT64X __HAVE_FLOAT128 + +/* Defined to 1 if __HAVE_FLOAT64X is 1 and _Float64x has the format + of long double. Otherwise, if __HAVE_FLOAT64X is 1, _Float64x has + the format of _Float128, which must be different from that of long + double. */ +#define __HAVE_FLOAT64X_LONG_DOUBLE __HAVE_FLOAT128 + +#ifndef __ASSEMBLER__ + +/* Defined to concatenate the literal suffix to be used with _Float128 + types, if __HAVE_FLOAT128 is 1. */ +# if __HAVE_FLOAT128 +# if !__GNUC_PREREQ (7, 0) || defined __cplusplus +/* The literal suffix f128 exists only since GCC 7.0. */ +# define __f128(x) x##l +# else +# define __f128(x) x##f128 +# endif +# endif + +/* Defined to a complex binary128 type if __HAVE_FLOAT128 is 1. */ +# if __HAVE_FLOAT128 +# if !__GNUC_PREREQ (7, 0) || defined __cplusplus +# define __CFLOAT128 _Complex long double +# else +# define __CFLOAT128 _Complex _Float128 +# endif +# endif + +/* The remaining of this file provides support for older compilers. */ +# if __HAVE_FLOAT128 + +/* The type _Float128 exists only since GCC 7.0. */ +# if !__GNUC_PREREQ (7, 0) || defined __cplusplus +typedef long double _Float128; +# endif + +/* Various built-in functions do not exist before GCC 7.0. */ +# if !__GNUC_PREREQ (7, 0) +# define __builtin_huge_valf128() (__builtin_huge_vall ()) +# define __builtin_inff128() (__builtin_infl ()) +# define __builtin_nanf128(x) (__builtin_nanl (x)) +# define __builtin_nansf128(x) (__builtin_nansl (x)) +# endif + +# endif + +#endif /* !__ASSEMBLER__. */ + +#include + +#endif /* _BITS_FLOATN_H */ \ No newline at end of file diff --git a/lib/libc/include/aarch64-linux-musl/bits/posix.h b/lib/libc/include/aarch64-linux-musl/bits/posix.h new file mode 100644 index 000000000..9fcc205cc --- /dev/null +++ b/lib/libc/include/aarch64-linux-musl/bits/posix.h @@ -0,0 +1,2 @@ +#define _POSIX_V6_LP64_OFF64 1 +#define _POSIX_V7_LP64_OFF64 1 \ No newline at end of file diff --git a/lib/libc/include/aarch64_be-linux-gnu/bits/floatn.h b/lib/libc/include/aarch64_be-linux-gnu/bits/floatn.h new file mode 100644 index 000000000..0b1b5d444 --- /dev/null +++ b/lib/libc/include/aarch64_be-linux-gnu/bits/floatn.h @@ -0,0 +1,97 @@ +/* Macros to control TS 18661-3 glibc features on ldbl-128 platforms. + Copyright (C) 2017-2021 Free Software Foundation, Inc. + This file is part of the GNU C Library. + + The GNU C Library is free software; you can redistribute it and/or + modify it under the terms of the GNU Lesser General Public + License as published by the Free Software Foundation; either + version 2.1 of the License, or (at your option) any later version. + + The GNU C Library is distributed in the hope that it will be useful, + but WITHOUT ANY WARRANTY; without even the implied warranty of + MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU + Lesser General Public License for more details. + + You should have received a copy of the GNU Lesser General Public + License along with the GNU C Library; if not, see + . */ + +#ifndef _BITS_FLOATN_H +#define _BITS_FLOATN_H + +#include +#include + +/* Defined to 1 if the current compiler invocation provides a + floating-point type with the IEEE 754 binary128 format, and this + glibc includes corresponding *f128 interfaces for it. */ +#ifndef __NO_LONG_DOUBLE_MATH +# define __HAVE_FLOAT128 1 +#else +/* glibc does not support _Float128 for platforms where long double is + normally binary128 when building with long double as binary64. + GCC's default for supported scalar modes does not support it either + in that case. */ +# define __HAVE_FLOAT128 0 +#endif + +/* Defined to 1 if __HAVE_FLOAT128 is 1 and the type is ABI-distinct + from the default float, double and long double types in this glibc. */ +#define __HAVE_DISTINCT_FLOAT128 0 + +/* Defined to 1 if the current compiler invocation provides a + floating-point type with the right format for _Float64x, and this + glibc includes corresponding *f64x interfaces for it. */ +#define __HAVE_FLOAT64X __HAVE_FLOAT128 + +/* Defined to 1 if __HAVE_FLOAT64X is 1 and _Float64x has the format + of long double. Otherwise, if __HAVE_FLOAT64X is 1, _Float64x has + the format of _Float128, which must be different from that of long + double. */ +#define __HAVE_FLOAT64X_LONG_DOUBLE __HAVE_FLOAT128 + +#ifndef __ASSEMBLER__ + +/* Defined to concatenate the literal suffix to be used with _Float128 + types, if __HAVE_FLOAT128 is 1. */ +# if __HAVE_FLOAT128 +# if !__GNUC_PREREQ (7, 0) || defined __cplusplus +/* The literal suffix f128 exists only since GCC 7.0. */ +# define __f128(x) x##l +# else +# define __f128(x) x##f128 +# endif +# endif + +/* Defined to a complex binary128 type if __HAVE_FLOAT128 is 1. */ +# if __HAVE_FLOAT128 +# if !__GNUC_PREREQ (7, 0) || defined __cplusplus +# define __CFLOAT128 _Complex long double +# else +# define __CFLOAT128 _Complex _Float128 +# endif +# endif + +/* The remaining of this file provides support for older compilers. */ +# if __HAVE_FLOAT128 + +/* The type _Float128 exists only since GCC 7.0. */ +# if !__GNUC_PREREQ (7, 0) || defined __cplusplus +typedef long double _Float128; +# endif + +/* Various built-in functions do not exist before GCC 7.0. */ +# if !__GNUC_PREREQ (7, 0) +# define __builtin_huge_valf128() (__builtin_huge_vall ()) +# define __builtin_inff128() (__builtin_infl ()) +# define __builtin_nanf128(x) (__builtin_nanl (x)) +# define __builtin_nansf128(x) (__builtin_nansl (x)) +# endif + +# endif + +#endif /* !__ASSEMBLER__. */ + +#include + +#endif /* _BITS_FLOATN_H */ \ No newline at end of file diff --git a/lib/libc/include/arm-linux-gnueabi/bits/floatn.h b/lib/libc/include/arm-linux-gnueabi/bits/floatn.h deleted file mode 100644 index 18018fa9f..000000000 --- a/lib/libc/include/arm-linux-gnueabi/bits/floatn.h +++ /dev/null @@ -1,52 +0,0 @@ -/* Macros to control TS 18661-3 glibc features. - Copyright (C) 2017-2021 Free Software Foundation, Inc. - This file is part of the GNU C Library. - - The GNU C Library is free software; you can redistribute it and/or - modify it under the terms of the GNU Lesser General Public - License as published by the Free Software Foundation; either - version 2.1 of the License, or (at your option) any later version. - - The GNU C Library is distributed in the hope that it will be useful, - but WITHOUT ANY WARRANTY; without even the implied warranty of - MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU - Lesser General Public License for more details. - - You should have received a copy of the GNU Lesser General Public - License along with the GNU C Library; if not, see - . */ - -/* Defined to 1 if the current compiler invocation provides a - floating-point type with the IEEE 754 binary128 format, and this glibc - includes corresponding *f128 interfaces for it. */ -#define __HAVE_FLOAT128 0 - -/* Defined to 1 if __HAVE_FLOAT128 is 1 and the type is ABI-distinct - from the default float, double and long double types in this glibc. */ -#define __HAVE_DISTINCT_FLOAT128 0 - -/* Defined to 1 if the current compiler invocation provides a - floating-point type with the right format for _Float64x, and this - glibc includes corresponding *f64x interfaces for it. */ -#define __HAVE_FLOAT64X 0 - -/* Defined to 1 if __HAVE_FLOAT64X is 1 and _Float64x has the format - of long double. Otherwise, if __HAVE_FLOAT64X is 1, _Float64x has - the format of _Float128, which must be different from that of long - double. */ -#define __HAVE_FLOAT64X_LONG_DOUBLE 0 - -#ifndef __ASSEMBLER__ - -/* Defined to concatenate the literal suffix to be used with _Float128 - types, if __HAVE_FLOAT128 is 1. - E.g.: #define __f128(x) x##f128. */ -# undef __f128 - -/* Defined to a complex binary128 type if __HAVE_FLOAT128 is 1. - E.g.: #define __CFLOAT128 _Complex _Float128. */ -# undef __CFLOAT128 - -#endif /* !__ASSEMBLER__. */ - -#include \ No newline at end of file diff --git a/lib/libc/include/arm-linux-gnueabihf/bits/floatn.h b/lib/libc/include/arm-linux-gnueabihf/bits/floatn.h deleted file mode 100644 index 18018fa9f..000000000 --- a/lib/libc/include/arm-linux-gnueabihf/bits/floatn.h +++ /dev/null @@ -1,52 +0,0 @@ -/* Macros to control TS 18661-3 glibc features. - Copyright (C) 2017-2021 Free Software Foundation, Inc. - This file is part of the GNU C Library. - - The GNU C Library is free software; you can redistribute it and/or - modify it under the terms of the GNU Lesser General Public - License as published by the Free Software Foundation; either - version 2.1 of the License, or (at your option) any later version. - - The GNU C Library is distributed in the hope that it will be useful, - but WITHOUT ANY WARRANTY; without even the implied warranty of - MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU - Lesser General Public License for more details. - - You should have received a copy of the GNU Lesser General Public - License along with the GNU C Library; if not, see - . */ - -/* Defined to 1 if the current compiler invocation provides a - floating-point type with the IEEE 754 binary128 format, and this glibc - includes corresponding *f128 interfaces for it. */ -#define __HAVE_FLOAT128 0 - -/* Defined to 1 if __HAVE_FLOAT128 is 1 and the type is ABI-distinct - from the default float, double and long double types in this glibc. */ -#define __HAVE_DISTINCT_FLOAT128 0 - -/* Defined to 1 if the current compiler invocation provides a - floating-point type with the right format for _Float64x, and this - glibc includes corresponding *f64x interfaces for it. */ -#define __HAVE_FLOAT64X 0 - -/* Defined to 1 if __HAVE_FLOAT64X is 1 and _Float64x has the format - of long double. Otherwise, if __HAVE_FLOAT64X is 1, _Float64x has - the format of _Float128, which must be different from that of long - double. */ -#define __HAVE_FLOAT64X_LONG_DOUBLE 0 - -#ifndef __ASSEMBLER__ - -/* Defined to concatenate the literal suffix to be used with _Float128 - types, if __HAVE_FLOAT128 is 1. - E.g.: #define __f128(x) x##f128. */ -# undef __f128 - -/* Defined to a complex binary128 type if __HAVE_FLOAT128 is 1. - E.g.: #define __CFLOAT128 _Complex _Float128. */ -# undef __CFLOAT128 - -#endif /* !__ASSEMBLER__. */ - -#include \ No newline at end of file diff --git a/lib/libc/include/arm-linux-musl/bits/posix.h b/lib/libc/include/arm-linux-musl/bits/posix.h deleted file mode 100644 index b0aba3e4a..000000000 --- a/lib/libc/include/arm-linux-musl/bits/posix.h +++ /dev/null @@ -1,2 +0,0 @@ -#define _POSIX_V6_ILP32_OFFBIG 1 -#define _POSIX_V7_ILP32_OFFBIG 1 \ No newline at end of file diff --git a/lib/libc/include/armeb-linux-gnueabi/bits/floatn.h b/lib/libc/include/armeb-linux-gnueabi/bits/floatn.h deleted file mode 100644 index 18018fa9f..000000000 --- a/lib/libc/include/armeb-linux-gnueabi/bits/floatn.h +++ /dev/null @@ -1,52 +0,0 @@ -/* Macros to control TS 18661-3 glibc features. - Copyright (C) 2017-2021 Free Software Foundation, Inc. - This file is part of the GNU C Library. - - The GNU C Library is free software; you can redistribute it and/or - modify it under the terms of the GNU Lesser General Public - License as published by the Free Software Foundation; either - version 2.1 of the License, or (at your option) any later version. - - The GNU C Library is distributed in the hope that it will be useful, - but WITHOUT ANY WARRANTY; without even the implied warranty of - MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU - Lesser General Public License for more details. - - You should have received a copy of the GNU Lesser General Public - License along with the GNU C Library; if not, see - . */ - -/* Defined to 1 if the current compiler invocation provides a - floating-point type with the IEEE 754 binary128 format, and this glibc - includes corresponding *f128 interfaces for it. */ -#define __HAVE_FLOAT128 0 - -/* Defined to 1 if __HAVE_FLOAT128 is 1 and the type is ABI-distinct - from the default float, double and long double types in this glibc. */ -#define __HAVE_DISTINCT_FLOAT128 0 - -/* Defined to 1 if the current compiler invocation provides a - floating-point type with the right format for _Float64x, and this - glibc includes corresponding *f64x interfaces for it. */ -#define __HAVE_FLOAT64X 0 - -/* Defined to 1 if __HAVE_FLOAT64X is 1 and _Float64x has the format - of long double. Otherwise, if __HAVE_FLOAT64X is 1, _Float64x has - the format of _Float128, which must be different from that of long - double. */ -#define __HAVE_FLOAT64X_LONG_DOUBLE 0 - -#ifndef __ASSEMBLER__ - -/* Defined to concatenate the literal suffix to be used with _Float128 - types, if __HAVE_FLOAT128 is 1. - E.g.: #define __f128(x) x##f128. */ -# undef __f128 - -/* Defined to a complex binary128 type if __HAVE_FLOAT128 is 1. - E.g.: #define __CFLOAT128 _Complex _Float128. */ -# undef __CFLOAT128 - -#endif /* !__ASSEMBLER__. */ - -#include \ No newline at end of file diff --git a/lib/libc/include/armeb-linux-gnueabihf/bits/floatn.h b/lib/libc/include/armeb-linux-gnueabihf/bits/floatn.h deleted file mode 100644 index 18018fa9f..000000000 --- a/lib/libc/include/armeb-linux-gnueabihf/bits/floatn.h +++ /dev/null @@ -1,52 +0,0 @@ -/* Macros to control TS 18661-3 glibc features. - Copyright (C) 2017-2021 Free Software Foundation, Inc. - This file is part of the GNU C Library. - - The GNU C Library is free software; you can redistribute it and/or - modify it under the terms of the GNU Lesser General Public - License as published by the Free Software Foundation; either - version 2.1 of the License, or (at your option) any later version. - - The GNU C Library is distributed in the hope that it will be useful, - but WITHOUT ANY WARRANTY; without even the implied warranty of - MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU - Lesser General Public License for more details. - - You should have received a copy of the GNU Lesser General Public - License along with the GNU C Library; if not, see - . */ - -/* Defined to 1 if the current compiler invocation provides a - floating-point type with the IEEE 754 binary128 format, and this glibc - includes corresponding *f128 interfaces for it. */ -#define __HAVE_FLOAT128 0 - -/* Defined to 1 if __HAVE_FLOAT128 is 1 and the type is ABI-distinct - from the default float, double and long double types in this glibc. */ -#define __HAVE_DISTINCT_FLOAT128 0 - -/* Defined to 1 if the current compiler invocation provides a - floating-point type with the right format for _Float64x, and this - glibc includes corresponding *f64x interfaces for it. */ -#define __HAVE_FLOAT64X 0 - -/* Defined to 1 if __HAVE_FLOAT64X is 1 and _Float64x has the format - of long double. Otherwise, if __HAVE_FLOAT64X is 1, _Float64x has - the format of _Float128, which must be different from that of long - double. */ -#define __HAVE_FLOAT64X_LONG_DOUBLE 0 - -#ifndef __ASSEMBLER__ - -/* Defined to concatenate the literal suffix to be used with _Float128 - types, if __HAVE_FLOAT128 is 1. - E.g.: #define __f128(x) x##f128. */ -# undef __f128 - -/* Defined to a complex binary128 type if __HAVE_FLOAT128 is 1. - E.g.: #define __CFLOAT128 _Complex _Float128. */ -# undef __CFLOAT128 - -#endif /* !__ASSEMBLER__. */ - -#include \ No newline at end of file diff --git a/lib/libc/include/csky-linux-gnueabi/bits/floatn.h b/lib/libc/include/csky-linux-gnueabi/bits/floatn.h deleted file mode 100644 index 18018fa9f..000000000 --- a/lib/libc/include/csky-linux-gnueabi/bits/floatn.h +++ /dev/null @@ -1,52 +0,0 @@ -/* Macros to control TS 18661-3 glibc features. - Copyright (C) 2017-2021 Free Software Foundation, Inc. - This file is part of the GNU C Library. - - The GNU C Library is free software; you can redistribute it and/or - modify it under the terms of the GNU Lesser General Public - License as published by the Free Software Foundation; either - version 2.1 of the License, or (at your option) any later version. - - The GNU C Library is distributed in the hope that it will be useful, - but WITHOUT ANY WARRANTY; without even the implied warranty of - MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU - Lesser General Public License for more details. - - You should have received a copy of the GNU Lesser General Public - License along with the GNU C Library; if not, see - . */ - -/* Defined to 1 if the current compiler invocation provides a - floating-point type with the IEEE 754 binary128 format, and this glibc - includes corresponding *f128 interfaces for it. */ -#define __HAVE_FLOAT128 0 - -/* Defined to 1 if __HAVE_FLOAT128 is 1 and the type is ABI-distinct - from the default float, double and long double types in this glibc. */ -#define __HAVE_DISTINCT_FLOAT128 0 - -/* Defined to 1 if the current compiler invocation provides a - floating-point type with the right format for _Float64x, and this - glibc includes corresponding *f64x interfaces for it. */ -#define __HAVE_FLOAT64X 0 - -/* Defined to 1 if __HAVE_FLOAT64X is 1 and _Float64x has the format - of long double. Otherwise, if __HAVE_FLOAT64X is 1, _Float64x has - the format of _Float128, which must be different from that of long - double. */ -#define __HAVE_FLOAT64X_LONG_DOUBLE 0 - -#ifndef __ASSEMBLER__ - -/* Defined to concatenate the literal suffix to be used with _Float128 - types, if __HAVE_FLOAT128 is 1. - E.g.: #define __f128(x) x##f128. */ -# undef __f128 - -/* Defined to a complex binary128 type if __HAVE_FLOAT128 is 1. - E.g.: #define __CFLOAT128 _Complex _Float128. */ -# undef __CFLOAT128 - -#endif /* !__ASSEMBLER__. */ - -#include \ No newline at end of file diff --git a/lib/libc/include/csky-linux-gnueabi/bits/struct_rwlock.h b/lib/libc/include/csky-linux-gnueabi/bits/struct_rwlock.h deleted file mode 100644 index 7431c5597..000000000 --- a/lib/libc/include/csky-linux-gnueabi/bits/struct_rwlock.h +++ /dev/null @@ -1,61 +0,0 @@ -/* Default read-write lock implementation struct definitions. - Copyright (C) 2019-2021 Free Software Foundation, Inc. - This file is part of the GNU C Library. - - The GNU C Library is free software; you can redistribute it and/or - modify it under the terms of the GNU Lesser General Public - License as published by the Free Software Foundation; either - version 2.1 of the License, or (at your option) any later version. - - The GNU C Library is distributed in the hope that it will be useful, - but WITHOUT ANY WARRANTY; without even the implied warranty of - MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU - Lesser General Public License for more details. - - You should have received a copy of the GNU Lesser General Public - License along with the GNU C Library; if not, see - . */ - -#ifndef __RWLOCK_INTERNAL_H -#define __RWLOCK_INTERNAL_H - -#include - -/* Generic struct for both POSIX read-write lock. New ports are expected - to use the default layout, however archictetures can redefine it to add - arch-specific extensions (such as lock-elision). The struct have a size - of 32 bytes on both LP32 and LP64 architectures. */ - -struct __pthread_rwlock_arch_t -{ - unsigned int __readers; - unsigned int __writers; - unsigned int __wrphase_futex; - unsigned int __writers_futex; - unsigned int __pad3; - unsigned int __pad4; - /* FLAGS must stay at its position in the structure to maintain - binary compatibility. */ -#if __BYTE_ORDER == __BIG_ENDIAN - unsigned char __pad1; - unsigned char __pad2; - unsigned char __shared; - unsigned char __flags; -#else - unsigned char __flags; - unsigned char __shared; - unsigned char __pad1; - unsigned char __pad2; -#endif - int __cur_writer; -}; - -#if __BYTE_ORDER == __BIG_ENDIAN -# define __PTHREAD_RWLOCK_INITIALIZER(__flags) \ - 0, 0, 0, 0, 0, 0, 0, 0, 0, __flags, 0 -#else -# define __PTHREAD_RWLOCK_INITIALIZER(__flags) \ - 0, 0, 0, 0, 0, 0, __flags, 0, 0, 0, 0 -#endif - -#endif \ No newline at end of file diff --git a/lib/libc/include/csky-linux-gnueabi/bits/wordsize.h b/lib/libc/include/csky-linux-gnueabi/bits/wordsize.h deleted file mode 100644 index 6561e924f..000000000 --- a/lib/libc/include/csky-linux-gnueabi/bits/wordsize.h +++ /dev/null @@ -1,21 +0,0 @@ -/* Copyright (C) 1999-2021 Free Software Foundation, Inc. - This file is part of the GNU C Library. - - The GNU C Library is free software; you can redistribute it and/or - modify it under the terms of the GNU Lesser General Public - License as published by the Free Software Foundation; either - version 2.1 of the License, or (at your option) any later version. - - The GNU C Library is distributed in the hope that it will be useful, - but WITHOUT ANY WARRANTY; without even the implied warranty of - MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU - Lesser General Public License for more details. - - You should have received a copy of the GNU Lesser General Public - License along with the GNU C Library; if not, see - . */ - -#define __WORDSIZE 32 -#define __WORDSIZE_TIME64_COMPAT32 0 -#define __WORDSIZE32_SIZE_ULONG 0 -#define __WORDSIZE32_PTRDIFF_LONG 0 \ No newline at end of file diff --git a/lib/libc/include/csky-linux-gnueabihf/bits/floatn.h b/lib/libc/include/csky-linux-gnueabihf/bits/floatn.h deleted file mode 100644 index 18018fa9f..000000000 --- a/lib/libc/include/csky-linux-gnueabihf/bits/floatn.h +++ /dev/null @@ -1,52 +0,0 @@ -/* Macros to control TS 18661-3 glibc features. - Copyright (C) 2017-2021 Free Software Foundation, Inc. - This file is part of the GNU C Library. - - The GNU C Library is free software; you can redistribute it and/or - modify it under the terms of the GNU Lesser General Public - License as published by the Free Software Foundation; either - version 2.1 of the License, or (at your option) any later version. - - The GNU C Library is distributed in the hope that it will be useful, - but WITHOUT ANY WARRANTY; without even the implied warranty of - MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU - Lesser General Public License for more details. - - You should have received a copy of the GNU Lesser General Public - License along with the GNU C Library; if not, see - . */ - -/* Defined to 1 if the current compiler invocation provides a - floating-point type with the IEEE 754 binary128 format, and this glibc - includes corresponding *f128 interfaces for it. */ -#define __HAVE_FLOAT128 0 - -/* Defined to 1 if __HAVE_FLOAT128 is 1 and the type is ABI-distinct - from the default float, double and long double types in this glibc. */ -#define __HAVE_DISTINCT_FLOAT128 0 - -/* Defined to 1 if the current compiler invocation provides a - floating-point type with the right format for _Float64x, and this - glibc includes corresponding *f64x interfaces for it. */ -#define __HAVE_FLOAT64X 0 - -/* Defined to 1 if __HAVE_FLOAT64X is 1 and _Float64x has the format - of long double. Otherwise, if __HAVE_FLOAT64X is 1, _Float64x has - the format of _Float128, which must be different from that of long - double. */ -#define __HAVE_FLOAT64X_LONG_DOUBLE 0 - -#ifndef __ASSEMBLER__ - -/* Defined to concatenate the literal suffix to be used with _Float128 - types, if __HAVE_FLOAT128 is 1. - E.g.: #define __f128(x) x##f128. */ -# undef __f128 - -/* Defined to a complex binary128 type if __HAVE_FLOAT128 is 1. - E.g.: #define __CFLOAT128 _Complex _Float128. */ -# undef __CFLOAT128 - -#endif /* !__ASSEMBLER__. */ - -#include \ No newline at end of file diff --git a/lib/libc/include/csky-linux-gnueabihf/bits/struct_rwlock.h b/lib/libc/include/csky-linux-gnueabihf/bits/struct_rwlock.h deleted file mode 100644 index 7431c5597..000000000 --- a/lib/libc/include/csky-linux-gnueabihf/bits/struct_rwlock.h +++ /dev/null @@ -1,61 +0,0 @@ -/* Default read-write lock implementation struct definitions. - Copyright (C) 2019-2021 Free Software Foundation, Inc. - This file is part of the GNU C Library. - - The GNU C Library is free software; you can redistribute it and/or - modify it under the terms of the GNU Lesser General Public - License as published by the Free Software Foundation; either - version 2.1 of the License, or (at your option) any later version. - - The GNU C Library is distributed in the hope that it will be useful, - but WITHOUT ANY WARRANTY; without even the implied warranty of - MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU - Lesser General Public License for more details. - - You should have received a copy of the GNU Lesser General Public - License along with the GNU C Library; if not, see - . */ - -#ifndef __RWLOCK_INTERNAL_H -#define __RWLOCK_INTERNAL_H - -#include - -/* Generic struct for both POSIX read-write lock. New ports are expected - to use the default layout, however archictetures can redefine it to add - arch-specific extensions (such as lock-elision). The struct have a size - of 32 bytes on both LP32 and LP64 architectures. */ - -struct __pthread_rwlock_arch_t -{ - unsigned int __readers; - unsigned int __writers; - unsigned int __wrphase_futex; - unsigned int __writers_futex; - unsigned int __pad3; - unsigned int __pad4; - /* FLAGS must stay at its position in the structure to maintain - binary compatibility. */ -#if __BYTE_ORDER == __BIG_ENDIAN - unsigned char __pad1; - unsigned char __pad2; - unsigned char __shared; - unsigned char __flags; -#else - unsigned char __flags; - unsigned char __shared; - unsigned char __pad1; - unsigned char __pad2; -#endif - int __cur_writer; -}; - -#if __BYTE_ORDER == __BIG_ENDIAN -# define __PTHREAD_RWLOCK_INITIALIZER(__flags) \ - 0, 0, 0, 0, 0, 0, 0, 0, 0, __flags, 0 -#else -# define __PTHREAD_RWLOCK_INITIALIZER(__flags) \ - 0, 0, 0, 0, 0, 0, __flags, 0, 0, 0, 0 -#endif - -#endif \ No newline at end of file diff --git a/lib/libc/include/csky-linux-gnueabihf/bits/wordsize.h b/lib/libc/include/csky-linux-gnueabihf/bits/wordsize.h deleted file mode 100644 index 6561e924f..000000000 --- a/lib/libc/include/csky-linux-gnueabihf/bits/wordsize.h +++ /dev/null @@ -1,21 +0,0 @@ -/* Copyright (C) 1999-2021 Free Software Foundation, Inc. - This file is part of the GNU C Library. - - The GNU C Library is free software; you can redistribute it and/or - modify it under the terms of the GNU Lesser General Public - License as published by the Free Software Foundation; either - version 2.1 of the License, or (at your option) any later version. - - The GNU C Library is distributed in the hope that it will be useful, - but WITHOUT ANY WARRANTY; without even the implied warranty of - MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU - Lesser General Public License for more details. - - You should have received a copy of the GNU Lesser General Public - License along with the GNU C Library; if not, see - . */ - -#define __WORDSIZE 32 -#define __WORDSIZE_TIME64_COMPAT32 0 -#define __WORDSIZE32_SIZE_ULONG 0 -#define __WORDSIZE32_PTRDIFF_LONG 0 \ No newline at end of file diff --git a/lib/libc/include/generic-glibc/bits/floatn.h b/lib/libc/include/generic-glibc/bits/floatn.h index 0b1b5d444..18018fa9f 100644 --- a/lib/libc/include/generic-glibc/bits/floatn.h +++ b/lib/libc/include/generic-glibc/bits/floatn.h @@ -1,4 +1,4 @@ -/* Macros to control TS 18661-3 glibc features on ldbl-128 platforms. +/* Macros to control TS 18661-3 glibc features. Copyright (C) 2017-2021 Free Software Foundation, Inc. This file is part of the GNU C Library. @@ -16,24 +16,10 @@ License along with the GNU C Library; if not, see . */ -#ifndef _BITS_FLOATN_H -#define _BITS_FLOATN_H - -#include -#include - /* Defined to 1 if the current compiler invocation provides a - floating-point type with the IEEE 754 binary128 format, and this - glibc includes corresponding *f128 interfaces for it. */ -#ifndef __NO_LONG_DOUBLE_MATH -# define __HAVE_FLOAT128 1 -#else -/* glibc does not support _Float128 for platforms where long double is - normally binary128 when building with long double as binary64. - GCC's default for supported scalar modes does not support it either - in that case. */ -# define __HAVE_FLOAT128 0 -#endif + floating-point type with the IEEE 754 binary128 format, and this glibc + includes corresponding *f128 interfaces for it. */ +#define __HAVE_FLOAT128 0 /* Defined to 1 if __HAVE_FLOAT128 is 1 and the type is ABI-distinct from the default float, double and long double types in this glibc. */ @@ -42,56 +28,25 @@ /* Defined to 1 if the current compiler invocation provides a floating-point type with the right format for _Float64x, and this glibc includes corresponding *f64x interfaces for it. */ -#define __HAVE_FLOAT64X __HAVE_FLOAT128 +#define __HAVE_FLOAT64X 0 /* Defined to 1 if __HAVE_FLOAT64X is 1 and _Float64x has the format of long double. Otherwise, if __HAVE_FLOAT64X is 1, _Float64x has the format of _Float128, which must be different from that of long double. */ -#define __HAVE_FLOAT64X_LONG_DOUBLE __HAVE_FLOAT128 +#define __HAVE_FLOAT64X_LONG_DOUBLE 0 #ifndef __ASSEMBLER__ /* Defined to concatenate the literal suffix to be used with _Float128 - types, if __HAVE_FLOAT128 is 1. */ -# if __HAVE_FLOAT128 -# if !__GNUC_PREREQ (7, 0) || defined __cplusplus -/* The literal suffix f128 exists only since GCC 7.0. */ -# define __f128(x) x##l -# else -# define __f128(x) x##f128 -# endif -# endif - -/* Defined to a complex binary128 type if __HAVE_FLOAT128 is 1. */ -# if __HAVE_FLOAT128 -# if !__GNUC_PREREQ (7, 0) || defined __cplusplus -# define __CFLOAT128 _Complex long double -# else -# define __CFLOAT128 _Complex _Float128 -# endif -# endif - -/* The remaining of this file provides support for older compilers. */ -# if __HAVE_FLOAT128 + types, if __HAVE_FLOAT128 is 1. + E.g.: #define __f128(x) x##f128. */ +# undef __f128 -/* The type _Float128 exists only since GCC 7.0. */ -# if !__GNUC_PREREQ (7, 0) || defined __cplusplus -typedef long double _Float128; -# endif - -/* Various built-in functions do not exist before GCC 7.0. */ -# if !__GNUC_PREREQ (7, 0) -# define __builtin_huge_valf128() (__builtin_huge_vall ()) -# define __builtin_inff128() (__builtin_infl ()) -# define __builtin_nanf128(x) (__builtin_nanl (x)) -# define __builtin_nansf128(x) (__builtin_nansl (x)) -# endif - -# endif +/* Defined to a complex binary128 type if __HAVE_FLOAT128 is 1. + E.g.: #define __CFLOAT128 _Complex _Float128. */ +# undef __CFLOAT128 #endif /* !__ASSEMBLER__. */ -#include - -#endif /* _BITS_FLOATN_H */ \ No newline at end of file +#include \ No newline at end of file diff --git a/lib/libc/include/generic-glibc/bits/struct_rwlock.h b/lib/libc/include/generic-glibc/bits/struct_rwlock.h index 6355f539b..7431c5597 100644 --- a/lib/libc/include/generic-glibc/bits/struct_rwlock.h +++ b/lib/libc/include/generic-glibc/bits/struct_rwlock.h @@ -1,4 +1,4 @@ -/* MIPS internal rwlock struct definitions. +/* Default read-write lock implementation struct definitions. Copyright (C) 2019-2021 Free Software Foundation, Inc. This file is part of the GNU C Library. @@ -16,8 +16,15 @@ License along with the GNU C Library; if not, see . */ -#ifndef _RWLOCK_INTERNAL_H -#define _RWLOCK_INTERNAL_H +#ifndef __RWLOCK_INTERNAL_H +#define __RWLOCK_INTERNAL_H + +#include + +/* Generic struct for both POSIX read-write lock. New ports are expected + to use the default layout, however archictetures can redefine it to add + arch-specific extensions (such as lock-elision). The struct have a size + of 32 bytes on both LP32 and LP64 architectures. */ struct __pthread_rwlock_arch_t { @@ -27,45 +34,28 @@ struct __pthread_rwlock_arch_t unsigned int __writers_futex; unsigned int __pad3; unsigned int __pad4; -#if _MIPS_SIM == _ABI64 - int __cur_writer; - int __shared; - unsigned long int __pad1; - unsigned long int __pad2; - /* FLAGS must stay at this position in the structure to maintain + /* FLAGS must stay at its position in the structure to maintain binary compatibility. */ - unsigned int __flags; -# else -# if __BYTE_ORDER == __BIG_ENDIAN +#if __BYTE_ORDER == __BIG_ENDIAN unsigned char __pad1; unsigned char __pad2; unsigned char __shared; - /* FLAGS must stay at this position in the structure to maintain - binary compatibility. */ unsigned char __flags; -# else - /* FLAGS must stay at this position in the structure to maintain - binary compatibility. */ +#else unsigned char __flags; unsigned char __shared; unsigned char __pad1; unsigned char __pad2; -# endif - int __cur_writer; #endif + int __cur_writer; }; -#if _MIPS_SIM == _ABI64 +#if __BYTE_ORDER == __BIG_ENDIAN # define __PTHREAD_RWLOCK_INITIALIZER(__flags) \ - 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, __flags -#else -# if __BYTE_ORDER == __BIG_ENDIAN -# define __PTHREAD_RWLOCK_INITIALIZER(__flags) \ 0, 0, 0, 0, 0, 0, 0, 0, 0, __flags, 0 -# else -# define __PTHREAD_RWLOCK_INITIALIZER(__flags) \ +#else +# define __PTHREAD_RWLOCK_INITIALIZER(__flags) \ 0, 0, 0, 0, 0, 0, __flags, 0, 0, 0, 0 -# endif #endif #endif \ No newline at end of file diff --git a/lib/libc/include/generic-glibc/bits/wordsize.h b/lib/libc/include/generic-glibc/bits/wordsize.h index 9e405e4e4..6561e924f 100644 --- a/lib/libc/include/generic-glibc/bits/wordsize.h +++ b/lib/libc/include/generic-glibc/bits/wordsize.h @@ -1,4 +1,4 @@ -/* Copyright (C) 2002-2021 Free Software Foundation, Inc. +/* Copyright (C) 1999-2021 Free Software Foundation, Inc. This file is part of the GNU C Library. The GNU C Library is free software; you can redistribute it and/or @@ -12,20 +12,10 @@ Lesser General Public License for more details. You should have received a copy of the GNU Lesser General Public - License along with the GNU C Library. If not, see + License along with the GNU C Library; if not, see . */ -#include - -#define __WORDSIZE _MIPS_SZPTR - -#if _MIPS_SIM == _ABI64 -# define __WORDSIZE_TIME64_COMPAT32 1 -#else -# define __WORDSIZE_TIME64_COMPAT32 0 -#endif - -#if __WORDSIZE == 32 +#define __WORDSIZE 32 +#define __WORDSIZE_TIME64_COMPAT32 0 #define __WORDSIZE32_SIZE_ULONG 0 -#define __WORDSIZE32_PTRDIFF_LONG 0 -#endif \ No newline at end of file +#define __WORDSIZE32_PTRDIFF_LONG 0 \ No newline at end of file diff --git a/lib/libc/include/generic-musl/bits/posix.h b/lib/libc/include/generic-musl/bits/posix.h index 9fcc205cc..b0aba3e4a 100644 --- a/lib/libc/include/generic-musl/bits/posix.h +++ b/lib/libc/include/generic-musl/bits/posix.h @@ -1,2 +1,2 @@ -#define _POSIX_V6_LP64_OFF64 1 -#define _POSIX_V7_LP64_OFF64 1 \ No newline at end of file +#define _POSIX_V6_ILP32_OFFBIG 1 +#define _POSIX_V7_ILP32_OFFBIG 1 \ No newline at end of file diff --git a/lib/libc/include/i386-linux-musl/bits/posix.h b/lib/libc/include/i386-linux-musl/bits/posix.h deleted file mode 100644 index b0aba3e4a..000000000 --- a/lib/libc/include/i386-linux-musl/bits/posix.h +++ /dev/null @@ -1,2 +0,0 @@ -#define _POSIX_V6_ILP32_OFFBIG 1 -#define _POSIX_V7_ILP32_OFFBIG 1 \ No newline at end of file diff --git a/lib/libc/include/m68k-linux-gnu/bits/a.out.h b/lib/libc/include/m68k-linux-gnu/bits/a.out.h new file mode 100644 index 000000000..398f4c27a --- /dev/null +++ b/lib/libc/include/m68k-linux-gnu/bits/a.out.h @@ -0,0 +1,3 @@ +#ifndef __A_OUT_GNU_H__ +# error "Never use directly; include instead." +#endif \ No newline at end of file diff --git a/lib/libc/include/m68k-linux-gnu/bits/endianness.h b/lib/libc/include/m68k-linux-gnu/bits/endianness.h new file mode 100644 index 000000000..f0dc4ef15 --- /dev/null +++ b/lib/libc/include/m68k-linux-gnu/bits/endianness.h @@ -0,0 +1,11 @@ +#ifndef _BITS_ENDIANNESS_H +#define _BITS_ENDIANNESS_H 1 + +#ifndef _BITS_ENDIAN_H +# error "Never use directly; include instead." +#endif + +/* m68k is big-endian. */ +#define __BYTE_ORDER __BIG_ENDIAN + +#endif /* bits/endianness.h */ \ No newline at end of file diff --git a/lib/libc/include/m68k-linux-gnu/bits/fcntl.h b/lib/libc/include/m68k-linux-gnu/bits/fcntl.h new file mode 100644 index 000000000..f4f6d57c1 --- /dev/null +++ b/lib/libc/include/m68k-linux-gnu/bits/fcntl.h @@ -0,0 +1,54 @@ +/* O_*, F_*, FD_* bit values for Linux. + Copyright (C) 2000-2021 Free Software Foundation, Inc. + This file is part of the GNU C Library. + + The GNU C Library is free software; you can redistribute it and/or + modify it under the terms of the GNU Lesser General Public + License as published by the Free Software Foundation; either + version 2.1 of the License, or (at your option) any later version. + + The GNU C Library is distributed in the hope that it will be useful, + but WITHOUT ANY WARRANTY; without even the implied warranty of + MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU + Lesser General Public License for more details. + + You should have received a copy of the GNU Lesser General Public + License along with the GNU C Library. If not, see + . */ + +#ifndef _FCNTL_H +# error "Never use directly; include instead." +#endif + +#define __O_DIRECTORY 040000 /* Must be a directory. */ +#define __O_NOFOLLOW 0100000 /* Do not follow links. */ +#define __O_DIRECT 0200000 /* Direct disk access. */ +#define __O_LARGEFILE 0400000 + +struct flock + { + short int l_type; /* Type of lock: F_RDLCK, F_WRLCK, or F_UNLCK. */ + short int l_whence; /* Where `l_start' is relative to (like `lseek'). */ +#ifndef __USE_FILE_OFFSET64 + __off_t l_start; /* Offset where the lock begins. */ + __off_t l_len; /* Size of the locked area; zero means until EOF. */ +#else + __off64_t l_start; /* Offset where the lock begins. */ + __off64_t l_len; /* Size of the locked area; zero means until EOF. */ +#endif + __pid_t l_pid; /* Process holding the lock. */ + }; + +#ifdef __USE_LARGEFILE64 +struct flock64 + { + short int l_type; /* Type of lock: F_RDLCK, F_WRLCK, or F_UNLCK. */ + short int l_whence; /* Where `l_start' is relative to (like `lseek'). */ + __off64_t l_start; /* Offset where the lock begins. */ + __off64_t l_len; /* Size of the locked area; zero means until EOF. */ + __pid_t l_pid; /* Process holding the lock. */ + }; +#endif + +/* Include generic Linux declarations. */ +#include \ No newline at end of file diff --git a/lib/libc/include/m68k-linux-gnu/bits/fenv.h b/lib/libc/include/m68k-linux-gnu/bits/fenv.h new file mode 100644 index 000000000..1bd8601ce --- /dev/null +++ b/lib/libc/include/m68k-linux-gnu/bits/fenv.h @@ -0,0 +1,131 @@ +/* Copyright (C) 1997-2021 Free Software Foundation, Inc. + This file is part of the GNU C Library. + + The GNU C Library is free software; you can redistribute it and/or + modify it under the terms of the GNU Lesser General Public + License as published by the Free Software Foundation; either + version 2.1 of the License, or (at your option) any later version. + + The GNU C Library is distributed in the hope that it will be useful, + but WITHOUT ANY WARRANTY; without even the implied warranty of + MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU + Lesser General Public License for more details. + + You should have received a copy of the GNU Lesser General Public + License along with the GNU C Library. If not, see + . */ + +#ifndef _FENV_H +# error "Never use directly; include instead." +#endif + + +#if defined __HAVE_68881__ || defined __HAVE_FPU__ || defined __mcffpu__ + +/* Define bits representing the exception. We use the bit positions of + the appropriate bits in the FPSR Accrued Exception Byte. */ +enum + { + FE_INEXACT = +# define FE_INEXACT (1 << 3) + FE_INEXACT, + FE_DIVBYZERO = +# define FE_DIVBYZERO (1 << 4) + FE_DIVBYZERO, + FE_UNDERFLOW = +# define FE_UNDERFLOW (1 << 5) + FE_UNDERFLOW, + FE_OVERFLOW = +# define FE_OVERFLOW (1 << 6) + FE_OVERFLOW, + FE_INVALID = +# define FE_INVALID (1 << 7) + FE_INVALID + }; + +# define FE_ALL_EXCEPT \ + (FE_INEXACT | FE_DIVBYZERO | FE_UNDERFLOW | FE_OVERFLOW | FE_INVALID) + +/* The m68k FPU supports all of the four defined rounding modes. We use + the bit positions in the FPCR Mode Control Byte as the values for the + appropriate macros. */ +enum + { + FE_TONEAREST = +# define FE_TONEAREST 0 + FE_TONEAREST, + FE_TOWARDZERO = +# define FE_TOWARDZERO (1 << 4) + FE_TOWARDZERO, + FE_DOWNWARD = +# define FE_DOWNWARD (2 << 4) + FE_DOWNWARD, + FE_UPWARD = +# define FE_UPWARD (3 << 4) + FE_UPWARD + }; + +#else + +/* In the soft-float case, only rounding to nearest is supported, with + no exceptions. */ + +# define FE_ALL_EXCEPT 0 + +enum + { + __FE_UNDEFINED = -1, + + FE_TONEAREST = +# define FE_TONEAREST 0 + FE_TONEAREST + }; + +#endif + + +/* Type representing exception flags. */ +typedef unsigned int fexcept_t; + + +#if defined __HAVE_68881__ || defined __HAVE_FPU__ || defined __mcffpu__ + +/* Type representing floating-point environment. This structure + corresponds to the layout of the block written by `fmovem'. */ +typedef struct + { + unsigned int __control_register; + unsigned int __status_register; + unsigned int __instruction_address; + } +fenv_t; + +#else + +/* Keep ABI compatibility with the type used in the generic + bits/fenv.h, formerly used for no-FPU ColdFire. */ +typedef struct + { + fexcept_t __excepts; + } +fenv_t; + +#endif + +/* If the default argument is used we use this value. */ +#define FE_DFL_ENV ((const fenv_t *) -1) + +#if defined __USE_GNU && (defined __HAVE_68881__ \ + || defined __HAVE_FPU__ \ + || defined __mcffpu__) +/* Floating-point environment where none of the exceptions are masked. */ +# define FE_NOMASK_ENV ((const fenv_t *) -2) +#endif + +#if __GLIBC_USE (IEC_60559_BFP_EXT_C2X) +/* Type representing floating-point control modes. */ +typedef unsigned int femode_t; + +/* Default floating-point control modes. */ +# define FE_DFL_MODE ((const femode_t *) -1L) +#endif \ No newline at end of file diff --git a/lib/libc/include/m68k-linux-gnu/bits/flt-eval-method.h b/lib/libc/include/m68k-linux-gnu/bits/flt-eval-method.h new file mode 100644 index 000000000..33644b4dc --- /dev/null +++ b/lib/libc/include/m68k-linux-gnu/bits/flt-eval-method.h @@ -0,0 +1,25 @@ +/* Define __GLIBC_FLT_EVAL_METHOD. M68K version. + Copyright (C) 2016-2021 Free Software Foundation, Inc. + This file is part of the GNU C Library. + + The GNU C Library is free software; you can redistribute it and/or + modify it under the terms of the GNU Lesser General Public + License as published by the Free Software Foundation; either + version 2.1 of the License, or (at your option) any later version. + + The GNU C Library is distributed in the hope that it will be useful, + but WITHOUT ANY WARRANTY; without even the implied warranty of + MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU + Lesser General Public License for more details. + + You should have received a copy of the GNU Lesser General Public + License along with the GNU C Library; if not, see + . */ + +#ifndef _MATH_H +# error "Never use directly; include instead." +#endif + +/* The m68k FPUs evaluate all values in the 96-bit floating-point + format which is also available for the user as 'long double'. */ +#define __GLIBC_FLT_EVAL_METHOD 2 \ No newline at end of file diff --git a/lib/libc/include/armeb-linux-gnueabi/bits/wordsize.h b/lib/libc/include/m68k-linux-gnu/bits/fp-logb.h similarity index 72% rename from lib/libc/include/armeb-linux-gnueabi/bits/wordsize.h rename to lib/libc/include/m68k-linux-gnu/bits/fp-logb.h index 6561e924f..1f184bd5d 100644 --- a/lib/libc/include/armeb-linux-gnueabi/bits/wordsize.h +++ b/lib/libc/include/m68k-linux-gnu/bits/fp-logb.h @@ -1,4 +1,5 @@ -/* Copyright (C) 1999-2021 Free Software Foundation, Inc. +/* Define __FP_LOGB0_IS_MIN and __FP_LOGBNAN_IS_MIN. M68K version. + Copyright (C) 2016-2021 Free Software Foundation, Inc. This file is part of the GNU C Library. The GNU C Library is free software; you can redistribute it and/or @@ -15,7 +16,9 @@ License along with the GNU C Library; if not, see . */ -#define __WORDSIZE 32 -#define __WORDSIZE_TIME64_COMPAT32 0 -#define __WORDSIZE32_SIZE_ULONG 0 -#define __WORDSIZE32_PTRDIFF_LONG 0 \ No newline at end of file +#ifndef _MATH_H +# error "Never use directly; include instead." +#endif + +#define __FP_LOGB0_IS_MIN 1 +#define __FP_LOGBNAN_IS_MIN 0 \ No newline at end of file diff --git a/lib/libc/include/m68k-linux-gnu/bits/iscanonical.h b/lib/libc/include/m68k-linux-gnu/bits/iscanonical.h new file mode 100644 index 000000000..8e6c0d05f --- /dev/null +++ b/lib/libc/include/m68k-linux-gnu/bits/iscanonical.h @@ -0,0 +1,54 @@ +/* Define iscanonical macro. ldbl-96 version. + Copyright (C) 2016-2021 Free Software Foundation, Inc. + This file is part of the GNU C Library. + + The GNU C Library is free software; you can redistribute it and/or + modify it under the terms of the GNU Lesser General Public + License as published by the Free Software Foundation; either + version 2.1 of the License, or (at your option) any later version. + + The GNU C Library is distributed in the hope that it will be useful, + but WITHOUT ANY WARRANTY; without even the implied warranty of + MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU + Lesser General Public License for more details. + + You should have received a copy of the GNU Lesser General Public + License along with the GNU C Library; if not, see + . */ + +#ifndef _MATH_H +# error "Never use directly; include instead." +#endif + +extern int __iscanonicall (long double __x) + __THROW __attribute__ ((__const__)); +#define __iscanonicalf(x) ((void) (__typeof (x)) (x), 1) +#define __iscanonical(x) ((void) (__typeof (x)) (x), 1) +#if __HAVE_DISTINCT_FLOAT128 +# define __iscanonicalf128(x) ((void) (__typeof (x)) (x), 1) +#endif + +/* Return nonzero value if X is canonical. In IEEE interchange binary + formats, all values are canonical, but the argument must still be + converted to its semantic type for any exceptions arising from the + conversion, before being discarded; in extended precision, there + are encodings that are not consistently handled as corresponding to + any particular value of the type, and we return 0 for those. */ +#ifndef __cplusplus +# define iscanonical(x) __MATH_TG ((x), __iscanonical, (x)) +#else +/* In C++ mode, __MATH_TG cannot be used, because it relies on + __builtin_types_compatible_p, which is a C-only builtin. On the + other hand, overloading provides the means to distinguish between + the floating-point types. The overloading resolution will match + the correct parameter (regardless of type qualifiers (i.e.: const + and volatile)). */ +extern "C++" { +inline int iscanonical (float __val) { return __iscanonicalf (__val); } +inline int iscanonical (double __val) { return __iscanonical (__val); } +inline int iscanonical (long double __val) { return __iscanonicall (__val); } +# if __HAVE_DISTINCT_FLOAT128 +inline int iscanonical (_Float128 __val) { return __iscanonicalf128 (__val); } +# endif +} +#endif /* __cplusplus */ \ No newline at end of file diff --git a/lib/libc/include/m68k-linux-gnu/bits/link.h b/lib/libc/include/m68k-linux-gnu/bits/link.h new file mode 100644 index 000000000..55d203f37 --- /dev/null +++ b/lib/libc/include/m68k-linux-gnu/bits/link.h @@ -0,0 +1,57 @@ +/* Copyright (C) 2005-2021 Free Software Foundation, Inc. + This file is part of the GNU C Library. + + The GNU C Library is free software; you can redistribute it and/or + modify it under the terms of the GNU Lesser General Public + License as published by the Free Software Foundation; either + version 2.1 of the License, or (at your option) any later version. + + The GNU C Library is distributed in the hope that it will be useful, + but WITHOUT ANY WARRANTY; without even the implied warranty of + MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU + Lesser General Public License for more details. + + You should have received a copy of the GNU Lesser General Public + License along with the GNU C Library. If not, see + . */ + +#ifndef _LINK_H +# error "Never include directly; use instead." +#endif + + +/* Registers for entry into PLT on M68K. */ +typedef struct La_m68k_regs +{ + uint32_t lr_a0; + uint32_t lr_a1; + uint32_t lr_sp; +} La_m68k_regs; + +/* Return values for calls from PLT on M68K. */ +typedef struct La_m68k_retval +{ + uint32_t lrv_d0; + uint32_t lrv_d1; + uint32_t lrv_a0; + long double lrv_fp0; +} La_m68k_retval; + + +__BEGIN_DECLS + +extern Elf32_Addr la_m68k_gnu_pltenter (Elf32_Sym *__sym, unsigned int __ndx, + uintptr_t *__refcook, + uintptr_t *__defcook, + La_m68k_regs *__regs, + unsigned int *__flags, + const char *__symname, + long int *__framesizep); +extern unsigned int la_m68k_gnu_pltexit (Elf32_Sym *__sym, unsigned int __ndx, + uintptr_t *__refcook, + uintptr_t *__defcook, + const La_m68k_regs *__inregs, + La_m68k_retval *__outregs, + const char *__symname); + +__END_DECLS \ No newline at end of file diff --git a/lib/libc/include/arm-linux-gnueabihf/bits/wordsize.h b/lib/libc/include/m68k-linux-gnu/bits/long-double.h similarity index 68% rename from lib/libc/include/arm-linux-gnueabihf/bits/wordsize.h rename to lib/libc/include/m68k-linux-gnu/bits/long-double.h index 6561e924f..ef3832bc8 100644 --- a/lib/libc/include/arm-linux-gnueabihf/bits/wordsize.h +++ b/lib/libc/include/m68k-linux-gnu/bits/long-double.h @@ -1,9 +1,10 @@ -/* Copyright (C) 1999-2021 Free Software Foundation, Inc. +/* Properties of long double type. ldbl-96 version. + Copyright (C) 2016-2021 Free Software Foundation, Inc. This file is part of the GNU C Library. The GNU C Library is free software; you can redistribute it and/or modify it under the terms of the GNU Lesser General Public - License as published by the Free Software Foundation; either + License published by the Free Software Foundation; either version 2.1 of the License, or (at your option) any later version. The GNU C Library is distributed in the hope that it will be useful, @@ -15,7 +16,6 @@ License along with the GNU C Library; if not, see . */ -#define __WORDSIZE 32 -#define __WORDSIZE_TIME64_COMPAT32 0 -#define __WORDSIZE32_SIZE_ULONG 0 -#define __WORDSIZE32_PTRDIFF_LONG 0 \ No newline at end of file +/* long double is distinct from double, so there is nothing to + define here. */ +#define __LDOUBLE_REDIRECTS_TO_FLOAT128_ABI 0 \ No newline at end of file diff --git a/lib/libc/include/m68k-linux-gnu/bits/poll.h b/lib/libc/include/m68k-linux-gnu/bits/poll.h new file mode 100644 index 000000000..fd12b7092 --- /dev/null +++ b/lib/libc/include/m68k-linux-gnu/bits/poll.h @@ -0,0 +1,49 @@ +/* Copyright (C) 1997-2021 Free Software Foundation, Inc. + This file is part of the GNU C Library. + + The GNU C Library is free software; you can redistribute it and/or + modify it under the terms of the GNU Lesser General Public + License as published by the Free Software Foundation; either + version 2.1 of the License, or (at your option) any later version. + + The GNU C Library is distributed in the hope that it will be useful, + but WITHOUT ANY WARRANTY; without even the implied warranty of + MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU + Lesser General Public License for more details. + + You should have received a copy of the GNU Lesser General Public + License along with the GNU C Library. If not, see + . */ + +#ifndef _SYS_POLL_H +# error "Never use directly; include instead." +#endif + +/* Event types that can be polled for. These bits may be set in `events' + to indicate the interesting event types; they will appear in `revents' + to indicate the status of the file descriptor. */ +#define POLLIN 0x001 /* There is data to read. */ +#define POLLPRI 0x002 /* There is urgent data to read. */ +#define POLLOUT 0x004 /* Writing now will not block. */ + +#if defined __USE_XOPEN || defined __USE_XOPEN2K8 +/* These values are defined in XPG4.2. */ +# define POLLRDNORM 0x040 /* Normal data may be read. */ +# define POLLRDBAND 0x080 /* Priority data may be read. */ +# define POLLWRNORM POLLOUT /* Writing now will not block. */ +# define POLLWRBAND 0x100 /* Priority data may be written. */ +#endif + +#ifdef __USE_GNU +/* These are extensions for Linux. */ +# define POLLMSG 0x400 +# define POLLREMOVE 0x1000 +# define POLLRDHUP 0x2000 +#endif + +/* Event types always implicitly polled for. These bits need not be set in + `events', but they will appear in `revents' to indicate the status of + the file descriptor. */ +#define POLLERR 0x008 /* Error condition. */ +#define POLLHUP 0x010 /* Hung up. */ +#define POLLNVAL 0x020 /* Invalid polling request. */ \ No newline at end of file diff --git a/lib/libc/include/armeb-linux-gnueabihf/bits/wordsize.h b/lib/libc/include/m68k-linux-gnu/bits/procfs-id.h similarity index 69% rename from lib/libc/include/armeb-linux-gnueabihf/bits/wordsize.h rename to lib/libc/include/m68k-linux-gnu/bits/procfs-id.h index 6561e924f..9ac1a07e1 100644 --- a/lib/libc/include/armeb-linux-gnueabihf/bits/wordsize.h +++ b/lib/libc/include/m68k-linux-gnu/bits/procfs-id.h @@ -1,4 +1,6 @@ -/* Copyright (C) 1999-2021 Free Software Foundation, Inc. +/* Types of pr_uid and pr_gid in struct elf_prpsinfo. M68K version. + Copyright (C) 2018-2021 Free Software Foundation, Inc. + This file is part of the GNU C Library. The GNU C Library is free software; you can redistribute it and/or @@ -15,7 +17,9 @@ License along with the GNU C Library; if not, see . */ -#define __WORDSIZE 32 -#define __WORDSIZE_TIME64_COMPAT32 0 -#define __WORDSIZE32_SIZE_ULONG 0 -#define __WORDSIZE32_PTRDIFF_LONG 0 \ No newline at end of file +#ifndef _SYS_PROCFS_H +# error "Never include directly; use instead." +#endif + +typedef unsigned short int __pr_uid_t; +typedef unsigned short int __pr_gid_t; \ No newline at end of file diff --git a/lib/libc/include/m68k-linux-gnu/bits/procfs.h b/lib/libc/include/m68k-linux-gnu/bits/procfs.h new file mode 100644 index 000000000..65a067f77 --- /dev/null +++ b/lib/libc/include/m68k-linux-gnu/bits/procfs.h @@ -0,0 +1,34 @@ +/* Types for registers for sys/procfs.h. M68K version. + Copyright (C) 1996-2021 Free Software Foundation, Inc. + This file is part of the GNU C Library. + + The GNU C Library is free software; you can redistribute it and/or + modify it under the terms of the GNU Lesser General Public + License as published by the Free Software Foundation; either + version 2.1 of the License, or (at your option) any later version. + + The GNU C Library is distributed in the hope that it will be useful, + but WITHOUT ANY WARRANTY; without even the implied warranty of + MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU + Lesser General Public License for more details. + + You should have received a copy of the GNU Lesser General Public + License along with the GNU C Library. If not, see + . */ + +#ifndef _SYS_PROCFS_H +# error "Never include directly; use instead." +#endif + +/* Type for a general-purpose register. */ +typedef unsigned long elf_greg_t; + +/* And the whole bunch of them. We could have used `struct + user_regs_struct' directly in the typedef, but tradition says that + the register set is an array, which does have some peculiar + semantics, so leave it that way. */ +#define ELF_NGREG (sizeof (struct user_regs_struct) / sizeof (elf_greg_t)) +typedef elf_greg_t elf_gregset_t[ELF_NGREG]; + +/* Register set for the floating-point registers. */ +typedef struct user_m68kfp_struct elf_fpregset_t; \ No newline at end of file diff --git a/lib/libc/include/m68k-linux-gnu/bits/pthreadtypes-arch.h b/lib/libc/include/m68k-linux-gnu/bits/pthreadtypes-arch.h new file mode 100644 index 000000000..262ca65c8 --- /dev/null +++ b/lib/libc/include/m68k-linux-gnu/bits/pthreadtypes-arch.h @@ -0,0 +1,37 @@ +/* Copyright (C) 2010-2021 Free Software Foundation, Inc. + This file is part of the GNU C Library. + Contributed by Maxim Kuvyrkov , 2010. + + The GNU C Library is free software; you can redistribute it and/or + modify it under the terms of the GNU Lesser General Public + License as published by the Free Software Foundation; either + version 2.1 of the License, or (at your option) any later version. + + The GNU C Library is distributed in the hope that it will be useful, + but WITHOUT ANY WARRANTY; without even the implied warranty of + MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU + Lesser General Public License for more details. + + You should have received a copy of the GNU Lesser General Public + License along with the GNU C Library. If not, see + . */ + +#ifndef _BITS_PTHREADTYPES_ARCH_H +#define _BITS_PTHREADTYPES_ARCH_H 1 + +#include + +#define __SIZEOF_PTHREAD_ATTR_T 36 +#define __SIZEOF_PTHREAD_MUTEX_T 24 +#define __SIZEOF_PTHREAD_MUTEXATTR_T 4 +#define __SIZEOF_PTHREAD_COND_T 48 +#define __SIZEOF_PTHREAD_CONDATTR_T 4 +#define __SIZEOF_PTHREAD_RWLOCK_T 32 +#define __SIZEOF_PTHREAD_RWLOCKATTR_T 8 +#define __SIZEOF_PTHREAD_BARRIER_T 20 +#define __SIZEOF_PTHREAD_BARRIERATTR_T 4 + +#define __LOCK_ALIGNMENT __attribute__ ((__aligned__ (4))) +#define __ONCE_ALIGNMENT __attribute__ ((__aligned__ (4))) + +#endif /* bits/pthreadtypes.h */ \ No newline at end of file diff --git a/lib/libc/include/m68k-linux-gnu/bits/semaphore.h b/lib/libc/include/m68k-linux-gnu/bits/semaphore.h new file mode 100644 index 000000000..d0ef954dd --- /dev/null +++ b/lib/libc/include/m68k-linux-gnu/bits/semaphore.h @@ -0,0 +1,35 @@ +/* Copyright (C) 2010-2021 Free Software Foundation, Inc. + This file is part of the GNU C Library. + Contributed by Maxim Kuvyrkov , 2010. + + The GNU C Library is free software; you can redistribute it and/or + modify it under the terms of the GNU Lesser General Public + License as published by the Free Software Foundation; either + version 2.1 of the License, or (at your option) any later version. + + The GNU C Library is distributed in the hope that it will be useful, + but WITHOUT ANY WARRANTY; without even the implied warranty of + MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU + Lesser General Public License for more details. + + You should have received a copy of the GNU Lesser General Public + License along with the GNU C Library. If not, see + . */ + +#ifndef _SEMAPHORE_H +# error "Never use directly; include instead." +#endif + + +#define __SIZEOF_SEM_T 16 + + +/* Value returned if `sem_open' failed. */ +#define SEM_FAILED ((sem_t *) 0) + + +typedef union +{ + char __size[__SIZEOF_SEM_T]; + long int __align __attribute__ ((__aligned__ (4))); +} sem_t; \ No newline at end of file diff --git a/lib/libc/include/m68k-linux-gnu/bits/setjmp.h b/lib/libc/include/m68k-linux-gnu/bits/setjmp.h new file mode 100644 index 000000000..3cd6dc3e6 --- /dev/null +++ b/lib/libc/include/m68k-linux-gnu/bits/setjmp.h @@ -0,0 +1,46 @@ +/* Copyright (C) 1997-2021 Free Software Foundation, Inc. + This file is part of the GNU C Library. + + The GNU C Library is free software; you can redistribute it and/or + modify it under the terms of the GNU Lesser General Public + License as published by the Free Software Foundation; either + version 2.1 of the License, or (at your option) any later version. + + The GNU C Library is distributed in the hope that it will be useful, + but WITHOUT ANY WARRANTY; without even the implied warranty of + MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU + Lesser General Public License for more details. + + You should have received a copy of the GNU Lesser General Public + License along with the GNU C Library. If not, see + . */ + +/* Define the machine-dependent type `jmp_buf'. m68k version. */ +#ifndef _BITS_SETJMP_H +#define _BITS_SETJMP_H 1 + +#if !defined _SETJMP_H && !defined _PTHREAD_H +# error "Never include directly; use instead." +#endif + +typedef struct __jmp_buf_internal_tag + { + /* There are eight 4-byte data registers, but D0 is not saved. */ + long int __dregs[7]; + + /* There are six 4-byte address registers, plus the FP and SP. */ + int *__aregs[6]; + int *__fp; + int *__sp; + +#if defined __HAVE_68881__ || defined __HAVE_FPU__ + /* There are eight floating point registers which + are saved in IEEE 96-bit extended format. */ + char __fpregs[8 * (96 / 8)]; +#elif defined __mcffpu__ + char __fpregs[8 * (64 / 8)]; +#endif + + } __jmp_buf[1]; + +#endif /* bits/setjmp.h */ \ No newline at end of file diff --git a/lib/libc/include/m68k-linux-gnu/bits/sockaddr.h b/lib/libc/include/m68k-linux-gnu/bits/sockaddr.h new file mode 100644 index 000000000..fe1093bd0 --- /dev/null +++ b/lib/libc/include/m68k-linux-gnu/bits/sockaddr.h @@ -0,0 +1,42 @@ +/* Definition of struct sockaddr_* members and sizes, Linux/m68k version. + Copyright (C) 1995-2021 Free Software Foundation, Inc. + This file is part of the GNU C Library. + + The GNU C Library is free software; you can redistribute it and/or + modify it under the terms of the GNU Lesser General Public + License as published by the Free Software Foundation; either + version 2.1 of the License, or (at your option) any later version. + + The GNU C Library is distributed in the hope that it will be useful, + but WITHOUT ANY WARRANTY; without even the implied warranty of + MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU + Lesser General Public License for more details. + + You should have received a copy of the GNU Lesser General Public + License along with the GNU C Library; if not, see + . */ + +/* + * Never include this file directly; use instead. + */ + +#ifndef _BITS_SOCKADDR_H +#define _BITS_SOCKADDR_H 1 + + +/* POSIX.1g specifies this type name for the `sa_family' member. */ +typedef unsigned short int sa_family_t; + +/* This macro is used to declare the initial common members + of the data types used for socket addresses, `struct sockaddr', + `struct sockaddr_in', `struct sockaddr_un', etc. */ + +#define __SOCKADDR_COMMON(sa_prefix) \ + sa_family_t sa_prefix##family + +#define __SOCKADDR_COMMON_SIZE (sizeof (unsigned short int)) + +/* Size of struct sockaddr_storage. */ +#define _SS_SIZE 126 + +#endif /* bits/sockaddr.h */ \ No newline at end of file diff --git a/lib/libc/include/m68k-linux-gnu/bits/struct_stat.h b/lib/libc/include/m68k-linux-gnu/bits/struct_stat.h new file mode 100644 index 000000000..e85f20f65 --- /dev/null +++ b/lib/libc/include/m68k-linux-gnu/bits/struct_stat.h @@ -0,0 +1,127 @@ +/* Definition for struct stat. + Copyright (C) 2020-2021 Free Software Foundation, Inc. + This file is part of the GNU C Library. + + The GNU C Library is free software; you can redistribute it and/or + modify it under the terms of the GNU Lesser General Public + License as published by the Free Software Foundation; either + version 2.1 of the License, or (at your option) any later version. + + The GNU C Library is distributed in the hope that it will be useful, + but WITHOUT ANY WARRANTY; without even the implied warranty of + MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU + Lesser General Public License for more details. + + You should have received a copy of the GNU Lesser General Public + License along with the GNU C Library. If not, see + . */ + +#if !defined _SYS_STAT_H && !defined _FCNTL_H +# error "Never include directly; use instead." +#endif + +#ifndef _BITS_STRUCT_STAT_H +#define _BITS_STRUCT_STAT_H 1 + +struct stat + { + __dev_t st_dev; /* Device. */ + unsigned short int __pad1; +#ifndef __USE_FILE_OFFSET64 + __ino_t st_ino; /* File serial number. */ +#else + __ino_t __st_ino; /* 32bit file serial number. */ +#endif + __mode_t st_mode; /* File mode. */ + __nlink_t st_nlink; /* Link count. */ + __uid_t st_uid; /* User ID of the file's owner. */ + __gid_t st_gid; /* Group ID of the file's group.*/ + __dev_t st_rdev; /* Device number, if device. */ + unsigned short int __pad2; +#ifndef __USE_FILE_OFFSET64 + __off_t st_size; /* Size of file, in bytes. */ +#else + __off64_t st_size; /* Size of file, in bytes. */ +#endif + __blksize_t st_blksize; /* Optimal block size for I/O. */ + +#ifndef __USE_FILE_OFFSET64 + __blkcnt_t st_blocks; /* Number 512-byte blocks allocated. */ +#else + __blkcnt64_t st_blocks; /* Number 512-byte blocks allocated. */ +#endif +#ifdef __USE_XOPEN2K8 + /* Nanosecond resolution timestamps are stored in a format + equivalent to 'struct timespec'. This is the type used + whenever possible but the Unix namespace rules do not allow the + identifier 'timespec' to appear in the header. + Therefore we have to handle the use of this header in strictly + standard-compliant sources special. */ + struct timespec st_atim; /* Time of last access. */ + struct timespec st_mtim; /* Time of last modification. */ + struct timespec st_ctim; /* Time of last status change. */ +# define st_atime st_atim.tv_sec /* Backward compatibility. */ +# define st_mtime st_mtim.tv_sec +# define st_ctime st_ctim.tv_sec +#else + __time_t st_atime; /* Time of last access. */ + unsigned long int st_atimensec; /* Nscecs of last access. */ + __time_t st_mtime; /* Time of last modification. */ + unsigned long int st_mtimensec; /* Nsecs of last modification. */ + __time_t st_ctime; /* Time of last status change. */ + unsigned long int st_ctimensec; /* Nsecs of last status change. */ +#endif +#ifndef __USE_FILE_OFFSET64 + unsigned long int __glibc_reserved4; + unsigned long int __glibc_reserved5; +#else + __ino64_t st_ino; /* File serial number. */ +#endif + }; + +#ifdef __USE_LARGEFILE64 +struct stat64 + { + __dev_t st_dev; /* Device. */ + unsigned short int __pad1; + + __ino_t __st_ino; /* 32bit file serial number. */ + __mode_t st_mode; /* File mode. */ + __nlink_t st_nlink; /* Link count. */ + __uid_t st_uid; /* User ID of the file's owner. */ + __gid_t st_gid; /* Group ID of the file's group.*/ + __dev_t st_rdev; /* Device number, if device. */ + unsigned short int __pad2; + __off64_t st_size; /* Size of file, in bytes. */ + __blksize_t st_blksize; /* Optimal block size for I/O. */ + + __blkcnt64_t st_blocks; /* Number 512-byte blocks allocated. */ +# ifdef __USE_XOPEN2K8 + /* Nanosecond resolution timestamps are stored in a format + equivalent to 'struct timespec'. This is the type used + whenever possible but the Unix namespace rules do not allow the + identifier 'timespec' to appear in the header. + Therefore we have to handle the use of this header in strictly + standard-compliant sources special. */ + struct timespec st_atim; /* Time of last access. */ + struct timespec st_mtim; /* Time of last modification. */ + struct timespec st_ctim; /* Time of last status change. */ +# else + __time_t st_atime; /* Time of last access. */ + unsigned long int st_atimensec; /* Nscecs of last access. */ + __time_t st_mtime; /* Time of last modification. */ + unsigned long int st_mtimensec; /* Nsecs of last modification. */ + __time_t st_ctime; /* Time of last status change. */ + unsigned long int st_ctimensec; /* Nsecs of last status change. */ +# endif + __ino64_t st_ino; /* File serial number. */ + }; +#endif + +/* Tell code we have these members. */ +#define _STATBUF_ST_BLKSIZE +#define _STATBUF_ST_RDEV +/* Nanosecond resolution time values are supported. */ +#define _STATBUF_ST_NSEC + +#endif /* _BITS_STRUCT_STAT_H */ \ No newline at end of file diff --git a/lib/libc/include/m68k-linux-gnu/fpu_control.h b/lib/libc/include/m68k-linux-gnu/fpu_control.h new file mode 100644 index 000000000..70ed49f09 --- /dev/null +++ b/lib/libc/include/m68k-linux-gnu/fpu_control.h @@ -0,0 +1,118 @@ +/* 68k FPU control word definitions. + Copyright (C) 1996-2021 Free Software Foundation, Inc. + This file is part of the GNU C Library. + + The GNU C Library is free software; you can redistribute it and/or + modify it under the terms of the GNU Lesser General Public + License as published by the Free Software Foundation; either + version 2.1 of the License, or (at your option) any later version. + + The GNU C Library is distributed in the hope that it will be useful, + but WITHOUT ANY WARRANTY; without even the implied warranty of + MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU + Lesser General Public License for more details. + + You should have received a copy of the GNU Lesser General Public + License along with the GNU C Library. If not, see + . */ + +#ifndef _FPU_CONTROL_H +#define _FPU_CONTROL_H + +/* + * Motorola floating point control register bits. + * + * 31-16 -> reserved (read as 0, ignored on write) + * 15 -> enable trap for BSUN exception + * 14 -> enable trap for SNAN exception + * 13 -> enable trap for OPERR exception + * 12 -> enable trap for OVFL exception + * 11 -> enable trap for UNFL exception + * 10 -> enable trap for DZ exception + * 9 -> enable trap for INEX2 exception (INEX on Coldfire) + * 8 -> enable trap for INEX1 exception (IDE on Coldfire) + * 7-6 -> Precision Control (only bit 6 is used on Coldfire) + * 5-4 -> Rounding Control + * 3-0 -> zero (read as 0, write as 0) + * + * + * Precision Control: + * 00 - round to extended precision + * 01 - round to single precision + * 10 - round to double precision + * 11 - undefined + * + * Rounding Control: + * 00 - rounding to nearest (RN) + * 01 - rounding toward zero (RZ) + * 10 - rounding (down)toward minus infinity (RM) + * 11 - rounding (up) toward plus infinity (RP) + * + * The hardware default is 0x0000. I choose 0x5400. + */ + +#include + +#if defined (__mcoldfire__) && !defined (__mcffpu__) + +# define _FPU_RESERVED 0xffffffff +# define _FPU_DEFAULT 0x00000000 +# define _FPU_GETCW(cw) ((cw) = 0) +# define _FPU_SETCW(cw) ((void) (cw)) + +#else + +/* masking of interrupts */ +# define _FPU_MASK_BSUN 0x8000 +# define _FPU_MASK_SNAN 0x4000 +# define _FPU_MASK_OPERR 0x2000 +# define _FPU_MASK_OVFL 0x1000 +# define _FPU_MASK_UNFL 0x0800 +# define _FPU_MASK_DZ 0x0400 +# define _FPU_MASK_INEX1 0x0200 +# define _FPU_MASK_INEX2 0x0100 + +/* precision control */ +# ifdef __mcoldfire__ +# define _FPU_DOUBLE 0x00 +# else +# define _FPU_EXTENDED 0x00 /* RECOMMENDED */ +# define _FPU_DOUBLE 0x80 +# endif +# define _FPU_SINGLE 0x40 /* DO NOT USE */ + +/* rounding control */ +# define _FPU_RC_NEAREST 0x00 /* RECOMMENDED */ +# define _FPU_RC_ZERO 0x10 +# define _FPU_RC_DOWN 0x20 +# define _FPU_RC_UP 0x30 + +# ifdef __mcoldfire__ +# define _FPU_RESERVED 0xFFFF800F +# else +# define _FPU_RESERVED 0xFFFF000F /* Reserved bits in fpucr */ +# endif + + +/* Now two recommended fpucr */ + +/* The fdlibm code requires no interrupts for exceptions. Don't + change the rounding mode, it would break long double I/O! */ +# define _FPU_DEFAULT 0x00000000 + +/* IEEE: same as above, but exceptions. We must make it non-zero so + that __setfpucw works. This bit will be ignored. */ +# define _FPU_IEEE 0x00000001 + +/* Macros for accessing the hardware control word. */ +# define _FPU_GETCW(cw) __asm__ ("fmove%.l %!, %0" : "=dm" (cw)) +# define _FPU_SETCW(cw) __asm__ volatile ("fmove%.l %0, %!" : : "dm" (cw)) +#endif + +/* Type of the control word. */ +typedef unsigned int fpu_control_t __attribute__ ((__mode__ (__SI__))); + +/* Default control word set at startup. */ +extern fpu_control_t __fpu_control; + +#endif /* _M68K_FPU_CONTROL_H */ \ No newline at end of file diff --git a/lib/libc/include/m68k-linux-gnu/gnu/lib-names.h b/lib/libc/include/m68k-linux-gnu/gnu/lib-names.h new file mode 100644 index 000000000..8d51e1156 --- /dev/null +++ b/lib/libc/include/m68k-linux-gnu/gnu/lib-names.h @@ -0,0 +1,31 @@ +/* This file is automatically generated. + It defines macros to allow user program to find the shared + library files which come as part of GNU libc. */ +#ifndef __GNU_LIB_NAMES_H +#define __GNU_LIB_NAMES_H 1 + +#define LD_SO "ld.so.1" +#define LIBANL_SO "libanl.so.1" +#define LIBBROKENLOCALE_SO "libBrokenLocale.so.1" +#define LIBCRYPT_SO "libcrypt.so.1" +#define LIBC_SO "libc.so.6" +#define LIBDL_SO "libdl.so.2" +#define LIBGCC_S_SO "libgcc_s.so.2" +#define LIBMVEC_SO "libmvec.so.1" +#define LIBM_SO "libm.so.6" +#define LIBNSL_SO "libnsl.so.1" +#define LIBNSS_COMPAT_SO "libnss_compat.so.2" +#define LIBNSS_DB_SO "libnss_db.so.2" +#define LIBNSS_DNS_SO "libnss_dns.so.2" +#define LIBNSS_FILES_SO "libnss_files.so.2" +#define LIBNSS_HESIOD_SO "libnss_hesiod.so.2" +#define LIBNSS_LDAP_SO "libnss_ldap.so.2" +#define LIBNSS_TEST1_SO "libnss_test1.so.2" +#define LIBNSS_TEST2_SO "libnss_test2.so.2" +#define LIBPTHREAD_SO "libpthread.so.0" +#define LIBRESOLV_SO "libresolv.so.2" +#define LIBRT_SO "librt.so.1" +#define LIBTHREAD_DB_SO "libthread_db.so.1" +#define LIBUTIL_SO "libutil.so.1" + +#endif /* gnu/lib-names.h */ \ No newline at end of file diff --git a/lib/libc/include/m68k-linux-gnu/gnu/stubs.h b/lib/libc/include/m68k-linux-gnu/gnu/stubs.h new file mode 100644 index 000000000..636dc7328 --- /dev/null +++ b/lib/libc/include/m68k-linux-gnu/gnu/stubs.h @@ -0,0 +1,16 @@ +/* This file is automatically generated. + It defines a symbol `__stub_FUNCTION' for each function + in the C library which is a stub, meaning it will fail + every time called, usually setting errno to ENOSYS. */ + +#ifdef _LIBC + #error Applications may not define the macro _LIBC +#endif + +#define __stub_chflags +#define __stub_fchflags +#define __stub_gtty +#define __stub_revoke +#define __stub_setlogin +#define __stub_sigreturn +#define __stub_stty \ No newline at end of file diff --git a/lib/libc/include/m68k-linux-gnu/sys/reg.h b/lib/libc/include/m68k-linux-gnu/sys/reg.h new file mode 100644 index 000000000..53b46a16a --- /dev/null +++ b/lib/libc/include/m68k-linux-gnu/sys/reg.h @@ -0,0 +1,101 @@ +/* Copyright (C) 1998-2021 Free Software Foundation, Inc. + This file is part of the GNU C Library. + + The GNU C Library is free software; you can redistribute it and/or + modify it under the terms of the GNU Lesser General Public + License as published by the Free Software Foundation; either + version 2.1 of the License, or (at your option) any later version. + + The GNU C Library is distributed in the hope that it will be useful, + but WITHOUT ANY WARRANTY; without even the implied warranty of + MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU + Lesser General Public License for more details. + + You should have received a copy of the GNU Lesser General Public + License along with the GNU C Library. If not, see + . */ + +#ifndef _SYS_REG_H +#define _SYS_REG_H 1 + +/* Index into an array of 4 byte integers returned from ptrace for + location of the users' stored general purpose registers. */ + +enum +{ + PT_D1 = 0, +#define PT_D1 PT_D1 + PT_D2 = 1, +#define PT_D2 PT_D2 + PT_D3 = 2, +#define PT_D3 PT_D3 + PT_D4 = 3, +#define PT_D4 PT_D4 + PT_D5 = 4, +#define PT_D5 PT_D5 + PT_D6 = 5, +#define PT_D6 PT_D6 + PT_D7 = 6, +#define PT_D7 PT_D7 + PT_A0 = 7, +#define PT_A0 PT_A0 + PT_A1 = 8, +#define PT_A1 PT_A1 + PT_A2 = 9, +#define PT_A2 PT_A2 + PT_A3 = 10, +#define PT_A3 PT_A3 + PT_A4 = 11, +#define PT_A4 PT_A4 + PT_A5 = 12, +#define PT_A5 PT_A5 + PT_A6 = 13, +#define PT_A6 PT_A6 + PT_D0 = 14, +#define PT_D0 PT_D0 + PT_USP = 15, +#define PT_USP PT_USP + PT_ORIG_D0 = 16, +#define PT_ORIG_D0 PT_ORIG_D0 + PT_SR = 17, +#define PT_SR PT_SR + PT_PC = 18, +#define PT_PC PT_PC + +#ifdef __mcoldfire__ + PT_FP0 = 21, + PT_FP1 = 23, + PT_FP2 = 25, + PT_FP3 = 27, + PT_FP4 = 29, + PT_FP5 = 31, + PT_FP6 = 33, + PT_FP7 = 35, +#else + PT_FP0 = 21, + PT_FP1 = 24, + PT_FP2 = 27, + PT_FP3 = 30, + PT_FP4 = 33, + PT_FP5 = 36, + PT_FP6 = 39, + PT_FP7 = 42, +#endif +#define PT_FP0 PT_FP0 +#define PT_FP1 PT_FP1 +#define PT_FP2 PT_FP2 +#define PT_FP3 PT_FP3 +#define PT_FP4 PT_FP4 +#define PT_FP5 PT_FP5 +#define PT_FP6 PT_FP6 +#define PT_FP7 PT_FP7 + + PT_FPCR = 45, +#define PT_FPCR PT_FPCR + PT_FPSR = 46, +#define PT_FPSR PT_FPSR + PT_FPIAR = 47 +#define PT_FPIAR PT_FPIAR +}; + +#endif /* _SYS_REG_H */ \ No newline at end of file diff --git a/lib/libc/include/m68k-linux-gnu/sys/ucontext.h b/lib/libc/include/m68k-linux-gnu/sys/ucontext.h new file mode 100644 index 000000000..6ec3ab57d --- /dev/null +++ b/lib/libc/include/m68k-linux-gnu/sys/ucontext.h @@ -0,0 +1,130 @@ +/* Copyright (C) 1997-2021 Free Software Foundation, Inc. + This file is part of the GNU C Library. + + The GNU C Library is free software; you can redistribute it and/or + modify it under the terms of the GNU Lesser General Public + License as published by the Free Software Foundation; either + version 2.1 of the License, or (at your option) any later version. + + The GNU C Library is distributed in the hope that it will be useful, + but WITHOUT ANY WARRANTY; without even the implied warranty of + MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU + Lesser General Public License for more details. + + You should have received a copy of the GNU Lesser General Public + License along with the GNU C Library. If not, see + . */ + +/* System V/m68k ABI compliant context switching support. */ + +#ifndef _SYS_UCONTEXT_H +#define _SYS_UCONTEXT_H 1 + +#include + +#include +#include + + +/* Type for general register. */ +typedef int greg_t; + +/* Number of general registers. */ +#define __NGREG 18 +#ifdef __USE_MISC +# define NGREG __NGREG +#endif + +/* Container for all general registers. */ +typedef greg_t gregset_t[__NGREG]; + +#ifdef __USE_MISC +/* Number of each register is the `gregset_t' array. */ +enum +{ + R_D0 = 0, +# define R_D0 R_D0 + R_D1 = 1, +# define R_D1 R_D1 + R_D2 = 2, +# define R_D2 R_D2 + R_D3 = 3, +# define R_D3 R_D3 + R_D4 = 4, +# define R_D4 R_D4 + R_D5 = 5, +# define R_D5 R_D5 + R_D6 = 6, +# define R_D6 R_D6 + R_D7 = 7, +# define R_D7 R_D7 + R_A0 = 8, +# define R_A0 R_A0 + R_A1 = 9, +# define R_A1 R_A1 + R_A2 = 10, +# define R_A2 R_A2 + R_A3 = 11, +# define R_A3 R_A3 + R_A4 = 12, +# define R_A4 R_A4 + R_A5 = 13, +# define R_A5 R_A5 + R_A6 = 14, +# define R_A6 R_A6 + R_A7 = 15, +# define R_A7 R_A7 + R_SP = 15, +# define R_SP R_SP + R_PC = 16, +# define R_PC R_PC + R_PS = 17 +# define R_PS R_PS +}; +#endif + +#ifdef __USE_MISC +# define __ctx(fld) fld +#else +# define __ctx(fld) __ ## fld +#endif + +/* Structure to describe FPU registers. */ +typedef struct +{ + int __ctx(f_pcr); + int __ctx(f_psr); + int __ctx(f_fpiaddr); +#ifdef __mcoldfire__ + int __ctx(f_fpregs)[8][2]; +#else + int __ctx(f_fpregs)[8][3]; +#endif +} fpregset_t; + +/* Context to describe whole processor state. */ +typedef struct +{ + int __ctx(version); + gregset_t __ctx(gregs); + fpregset_t __ctx(fpregs); +} mcontext_t; + +#ifdef __USE_MISC +# define MCONTEXT_VERSION 2 +#endif + +/* Userlevel context. */ +typedef struct ucontext_t +{ + unsigned long __ctx(uc_flags); + struct ucontext_t *uc_link; + stack_t uc_stack; + mcontext_t uc_mcontext; + unsigned long __glibc_reserved1[80]; + sigset_t uc_sigmask; +} ucontext_t; + +#undef __ctx + +#endif /* sys/ucontext.h */ \ No newline at end of file diff --git a/lib/libc/include/m68k-linux-gnu/sys/user.h b/lib/libc/include/m68k-linux-gnu/sys/user.h new file mode 100644 index 000000000..7e210f85b --- /dev/null +++ b/lib/libc/include/m68k-linux-gnu/sys/user.h @@ -0,0 +1,65 @@ +/* Copyright (C) 2008-2021 Free Software Foundation, Inc. + This file is part of the GNU C Library. + + The GNU C Library is free software; you can redistribute it and/or + modify it under the terms of the GNU Lesser General Public + License as published by the Free Software Foundation; either + version 2.1 of the License, or (at your option) any later version. + + The GNU C Library is distributed in the hope that it will be useful, + but WITHOUT ANY WARRANTY; without even the implied warranty of + MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU + Lesser General Public License for more details. + + You should have received a copy of the GNU Lesser General Public + License along with the GNU C Library. If not, see + . */ + +#ifndef _SYS_USER_H +#define _SYS_USER_H 1 + +/* The whole purpose of this file is for GDB and GDB only. Don't read + too much into it. Don't use it for anything other than GDB unless + you know what you are doing. */ + +struct user_m68kfp_struct { + unsigned long fpregs[8*3]; + unsigned long fpcntl[3]; +}; + +struct user_regs_struct { + long d1, d2, d3, d4, d5, d6, d7; + long a0, a1, a2, a3, a4, a5, a6; + long d0; + long usp; + long orig_d0; + short stkadj; + short sr; + long pc; + short fmtvec; + short __fill; +}; + +struct user { + struct user_regs_struct regs; + int u_fpvalid; + struct user_m68kfp_struct m68kfp; + unsigned long int u_tsize; + unsigned long int u_dsize; + unsigned long int u_ssize; + unsigned long start_code; + unsigned long start_stack; + long int signal; + int reserved; + unsigned long u_ar0; + struct user_m68kfp_struct *u_fpstate; + unsigned long magic; + char u_comm[32]; +}; + +#define NBPG 4096 +#define UPAGES 1 +#define HOST_TEXT_START_ADDR u.start_code +#define HOST_STACK_END_ADDR (u.start_stack + u.u_ssize * NBPG) + +#endif \ No newline at end of file diff --git a/lib/libc/include/m68k-linux-musl/bits/alltypes.h b/lib/libc/include/m68k-linux-musl/bits/alltypes.h new file mode 100644 index 000000000..12744ba59 --- /dev/null +++ b/lib/libc/include/m68k-linux-musl/bits/alltypes.h @@ -0,0 +1,424 @@ +#define _REDIR_TIME64 1 +#define _Addr int +#define _Int64 long long +#define _Reg int + +#define __BYTE_ORDER 4321 +#define __LONG_MAX 0x7fffffffL + +#ifndef __cplusplus +#ifdef __WCHAR_TYPE__ +#if defined(__NEED_wchar_t) && !defined(__DEFINED_wchar_t) +typedef __WCHAR_TYPE__ wchar_t; +#define __DEFINED_wchar_t +#endif + +#else +#if defined(__NEED_wchar_t) && !defined(__DEFINED_wchar_t) +typedef long wchar_t; +#define __DEFINED_wchar_t +#endif + +#endif +#endif + +#if __mcffpu__ +#if defined(__NEED_float_t) && !defined(__DEFINED_float_t) +typedef float float_t; +#define __DEFINED_float_t +#endif + +#if defined(__NEED_double_t) && !defined(__DEFINED_double_t) +typedef double double_t; +#define __DEFINED_double_t +#endif + +#else +#if defined(__NEED_float_t) && !defined(__DEFINED_float_t) +typedef long double float_t; +#define __DEFINED_float_t +#endif + +#if defined(__NEED_double_t) && !defined(__DEFINED_double_t) +typedef long double double_t; +#define __DEFINED_double_t +#endif + +#endif + +#if defined(__NEED_max_align_t) && !defined(__DEFINED_max_align_t) +typedef struct { long long __ll; long double __ld; } max_align_t; +#define __DEFINED_max_align_t +#endif + +#define __LITTLE_ENDIAN 1234 +#define __BIG_ENDIAN 4321 +#define __USE_TIME_BITS64 1 + +#if defined(__NEED_size_t) && !defined(__DEFINED_size_t) +typedef unsigned _Addr size_t; +#define __DEFINED_size_t +#endif + +#if defined(__NEED_uintptr_t) && !defined(__DEFINED_uintptr_t) +typedef unsigned _Addr uintptr_t; +#define __DEFINED_uintptr_t +#endif + +#if defined(__NEED_ptrdiff_t) && !defined(__DEFINED_ptrdiff_t) +typedef _Addr ptrdiff_t; +#define __DEFINED_ptrdiff_t +#endif + +#if defined(__NEED_ssize_t) && !defined(__DEFINED_ssize_t) +typedef _Addr ssize_t; +#define __DEFINED_ssize_t +#endif + +#if defined(__NEED_intptr_t) && !defined(__DEFINED_intptr_t) +typedef _Addr intptr_t; +#define __DEFINED_intptr_t +#endif + +#if defined(__NEED_regoff_t) && !defined(__DEFINED_regoff_t) +typedef _Addr regoff_t; +#define __DEFINED_regoff_t +#endif + +#if defined(__NEED_register_t) && !defined(__DEFINED_register_t) +typedef _Reg register_t; +#define __DEFINED_register_t +#endif + +#if defined(__NEED_time_t) && !defined(__DEFINED_time_t) +typedef _Int64 time_t; +#define __DEFINED_time_t +#endif + +#if defined(__NEED_suseconds_t) && !defined(__DEFINED_suseconds_t) +typedef _Int64 suseconds_t; +#define __DEFINED_suseconds_t +#endif + + +#if defined(__NEED_int8_t) && !defined(__DEFINED_int8_t) +typedef signed char int8_t; +#define __DEFINED_int8_t +#endif + +#if defined(__NEED_int16_t) && !defined(__DEFINED_int16_t) +typedef signed short int16_t; +#define __DEFINED_int16_t +#endif + +#if defined(__NEED_int32_t) && !defined(__DEFINED_int32_t) +typedef signed int int32_t; +#define __DEFINED_int32_t +#endif + +#if defined(__NEED_int64_t) && !defined(__DEFINED_int64_t) +typedef signed _Int64 int64_t; +#define __DEFINED_int64_t +#endif + +#if defined(__NEED_intmax_t) && !defined(__DEFINED_intmax_t) +typedef signed _Int64 intmax_t; +#define __DEFINED_intmax_t +#endif + +#if defined(__NEED_uint8_t) && !defined(__DEFINED_uint8_t) +typedef unsigned char uint8_t; +#define __DEFINED_uint8_t +#endif + +#if defined(__NEED_uint16_t) && !defined(__DEFINED_uint16_t) +typedef unsigned short uint16_t; +#define __DEFINED_uint16_t +#endif + +#if defined(__NEED_uint32_t) && !defined(__DEFINED_uint32_t) +typedef unsigned int uint32_t; +#define __DEFINED_uint32_t +#endif + +#if defined(__NEED_uint64_t) && !defined(__DEFINED_uint64_t) +typedef unsigned _Int64 uint64_t; +#define __DEFINED_uint64_t +#endif + +#if defined(__NEED_u_int64_t) && !defined(__DEFINED_u_int64_t) +typedef unsigned _Int64 u_int64_t; +#define __DEFINED_u_int64_t +#endif + +#if defined(__NEED_uintmax_t) && !defined(__DEFINED_uintmax_t) +typedef unsigned _Int64 uintmax_t; +#define __DEFINED_uintmax_t +#endif + + +#if defined(__NEED_mode_t) && !defined(__DEFINED_mode_t) +typedef unsigned mode_t; +#define __DEFINED_mode_t +#endif + +#if defined(__NEED_nlink_t) && !defined(__DEFINED_nlink_t) +typedef unsigned _Reg nlink_t; +#define __DEFINED_nlink_t +#endif + +#if defined(__NEED_off_t) && !defined(__DEFINED_off_t) +typedef _Int64 off_t; +#define __DEFINED_off_t +#endif + +#if defined(__NEED_ino_t) && !defined(__DEFINED_ino_t) +typedef unsigned _Int64 ino_t; +#define __DEFINED_ino_t +#endif + +#if defined(__NEED_dev_t) && !defined(__DEFINED_dev_t) +typedef unsigned _Int64 dev_t; +#define __DEFINED_dev_t +#endif + +#if defined(__NEED_blksize_t) && !defined(__DEFINED_blksize_t) +typedef long blksize_t; +#define __DEFINED_blksize_t +#endif + +#if defined(__NEED_blkcnt_t) && !defined(__DEFINED_blkcnt_t) +typedef _Int64 blkcnt_t; +#define __DEFINED_blkcnt_t +#endif + +#if defined(__NEED_fsblkcnt_t) && !defined(__DEFINED_fsblkcnt_t) +typedef unsigned _Int64 fsblkcnt_t; +#define __DEFINED_fsblkcnt_t +#endif + +#if defined(__NEED_fsfilcnt_t) && !defined(__DEFINED_fsfilcnt_t) +typedef unsigned _Int64 fsfilcnt_t; +#define __DEFINED_fsfilcnt_t +#endif + + +#if defined(__NEED_wint_t) && !defined(__DEFINED_wint_t) +typedef unsigned wint_t; +#define __DEFINED_wint_t +#endif + +#if defined(__NEED_wctype_t) && !defined(__DEFINED_wctype_t) +typedef unsigned long wctype_t; +#define __DEFINED_wctype_t +#endif + + +#if defined(__NEED_timer_t) && !defined(__DEFINED_timer_t) +typedef void * timer_t; +#define __DEFINED_timer_t +#endif + +#if defined(__NEED_clockid_t) && !defined(__DEFINED_clockid_t) +typedef int clockid_t; +#define __DEFINED_clockid_t +#endif + +#if defined(__NEED_clock_t) && !defined(__DEFINED_clock_t) +typedef long clock_t; +#define __DEFINED_clock_t +#endif + +#if defined(__NEED_struct_timeval) && !defined(__DEFINED_struct_timeval) +struct timeval { time_t tv_sec; suseconds_t tv_usec; }; +#define __DEFINED_struct_timeval +#endif + +#if defined(__NEED_struct_timespec) && !defined(__DEFINED_struct_timespec) +struct timespec { time_t tv_sec; int :8*(sizeof(time_t)-sizeof(long))*(__BYTE_ORDER==4321); long tv_nsec; int :8*(sizeof(time_t)-sizeof(long))*(__BYTE_ORDER!=4321); }; +#define __DEFINED_struct_timespec +#endif + + +#if defined(__NEED_pid_t) && !defined(__DEFINED_pid_t) +typedef int pid_t; +#define __DEFINED_pid_t +#endif + +#if defined(__NEED_id_t) && !defined(__DEFINED_id_t) +typedef unsigned id_t; +#define __DEFINED_id_t +#endif + +#if defined(__NEED_uid_t) && !defined(__DEFINED_uid_t) +typedef unsigned uid_t; +#define __DEFINED_uid_t +#endif + +#if defined(__NEED_gid_t) && !defined(__DEFINED_gid_t) +typedef unsigned gid_t; +#define __DEFINED_gid_t +#endif + +#if defined(__NEED_key_t) && !defined(__DEFINED_key_t) +typedef int key_t; +#define __DEFINED_key_t +#endif + +#if defined(__NEED_useconds_t) && !defined(__DEFINED_useconds_t) +typedef unsigned useconds_t; +#define __DEFINED_useconds_t +#endif + + +#ifdef __cplusplus +#if defined(__NEED_pthread_t) && !defined(__DEFINED_pthread_t) +typedef unsigned long pthread_t; +#define __DEFINED_pthread_t +#endif + +#else +#if defined(__NEED_pthread_t) && !defined(__DEFINED_pthread_t) +typedef struct __pthread * pthread_t; +#define __DEFINED_pthread_t +#endif + +#endif +#if defined(__NEED_pthread_once_t) && !defined(__DEFINED_pthread_once_t) +typedef int pthread_once_t; +#define __DEFINED_pthread_once_t +#endif + +#if defined(__NEED_pthread_key_t) && !defined(__DEFINED_pthread_key_t) +typedef unsigned pthread_key_t; +#define __DEFINED_pthread_key_t +#endif + +#if defined(__NEED_pthread_spinlock_t) && !defined(__DEFINED_pthread_spinlock_t) +typedef int pthread_spinlock_t; +#define __DEFINED_pthread_spinlock_t +#endif + +#if defined(__NEED_pthread_mutexattr_t) && !defined(__DEFINED_pthread_mutexattr_t) +typedef struct { unsigned __attr; } pthread_mutexattr_t; +#define __DEFINED_pthread_mutexattr_t +#endif + +#if defined(__NEED_pthread_condattr_t) && !defined(__DEFINED_pthread_condattr_t) +typedef struct { unsigned __attr; } pthread_condattr_t; +#define __DEFINED_pthread_condattr_t +#endif + +#if defined(__NEED_pthread_barrierattr_t) && !defined(__DEFINED_pthread_barrierattr_t) +typedef struct { unsigned __attr; } pthread_barrierattr_t; +#define __DEFINED_pthread_barrierattr_t +#endif + +#if defined(__NEED_pthread_rwlockattr_t) && !defined(__DEFINED_pthread_rwlockattr_t) +typedef struct { unsigned __attr[2]; } pthread_rwlockattr_t; +#define __DEFINED_pthread_rwlockattr_t +#endif + + +#if defined(__NEED_struct__IO_FILE) && !defined(__DEFINED_struct__IO_FILE) +struct _IO_FILE { char __x; }; +#define __DEFINED_struct__IO_FILE +#endif + +#if defined(__NEED_FILE) && !defined(__DEFINED_FILE) +typedef struct _IO_FILE FILE; +#define __DEFINED_FILE +#endif + + +#if defined(__NEED_va_list) && !defined(__DEFINED_va_list) +typedef __builtin_va_list va_list; +#define __DEFINED_va_list +#endif + +#if defined(__NEED___isoc_va_list) && !defined(__DEFINED___isoc_va_list) +typedef __builtin_va_list __isoc_va_list; +#define __DEFINED___isoc_va_list +#endif + + +#if defined(__NEED_mbstate_t) && !defined(__DEFINED_mbstate_t) +typedef struct __mbstate_t { unsigned __opaque1, __opaque2; } mbstate_t; +#define __DEFINED_mbstate_t +#endif + + +#if defined(__NEED_locale_t) && !defined(__DEFINED_locale_t) +typedef struct __locale_struct * locale_t; +#define __DEFINED_locale_t +#endif + + +#if defined(__NEED_sigset_t) && !defined(__DEFINED_sigset_t) +typedef struct __sigset_t { unsigned long __bits[128/sizeof(long)]; } sigset_t; +#define __DEFINED_sigset_t +#endif + + +#if defined(__NEED_struct_iovec) && !defined(__DEFINED_struct_iovec) +struct iovec { void *iov_base; size_t iov_len; }; +#define __DEFINED_struct_iovec +#endif + + +#if defined(__NEED_struct_winsize) && !defined(__DEFINED_struct_winsize) +struct winsize { unsigned short ws_row, ws_col, ws_xpixel, ws_ypixel; }; +#define __DEFINED_struct_winsize +#endif + + +#if defined(__NEED_socklen_t) && !defined(__DEFINED_socklen_t) +typedef unsigned socklen_t; +#define __DEFINED_socklen_t +#endif + +#if defined(__NEED_sa_family_t) && !defined(__DEFINED_sa_family_t) +typedef unsigned short sa_family_t; +#define __DEFINED_sa_family_t +#endif + + +#if defined(__NEED_pthread_attr_t) && !defined(__DEFINED_pthread_attr_t) +typedef struct { union { int __i[sizeof(long)==8?14:9]; volatile int __vi[sizeof(long)==8?14:9]; unsigned long __s[sizeof(long)==8?7:9]; } __u; } pthread_attr_t; +#define __DEFINED_pthread_attr_t +#endif + +#if defined(__NEED_pthread_mutex_t) && !defined(__DEFINED_pthread_mutex_t) +typedef struct { union { int __i[sizeof(long)==8?10:6]; volatile int __vi[sizeof(long)==8?10:6]; volatile void *volatile __p[sizeof(long)==8?5:6]; } __u; } pthread_mutex_t; +#define __DEFINED_pthread_mutex_t +#endif + +#if defined(__NEED_mtx_t) && !defined(__DEFINED_mtx_t) +typedef struct { union { int __i[sizeof(long)==8?10:6]; volatile int __vi[sizeof(long)==8?10:6]; volatile void *volatile __p[sizeof(long)==8?5:6]; } __u; } mtx_t; +#define __DEFINED_mtx_t +#endif + +#if defined(__NEED_pthread_cond_t) && !defined(__DEFINED_pthread_cond_t) +typedef struct { union { int __i[12]; volatile int __vi[12]; void *__p[12*sizeof(int)/sizeof(void*)]; } __u; } pthread_cond_t; +#define __DEFINED_pthread_cond_t +#endif + +#if defined(__NEED_cnd_t) && !defined(__DEFINED_cnd_t) +typedef struct { union { int __i[12]; volatile int __vi[12]; void *__p[12*sizeof(int)/sizeof(void*)]; } __u; } cnd_t; +#define __DEFINED_cnd_t +#endif + +#if defined(__NEED_pthread_rwlock_t) && !defined(__DEFINED_pthread_rwlock_t) +typedef struct { union { int __i[sizeof(long)==8?14:8]; volatile int __vi[sizeof(long)==8?14:8]; void *__p[sizeof(long)==8?7:8]; } __u; } pthread_rwlock_t; +#define __DEFINED_pthread_rwlock_t +#endif + +#if defined(__NEED_pthread_barrier_t) && !defined(__DEFINED_pthread_barrier_t) +typedef struct { union { int __i[sizeof(long)==8?8:5]; volatile int __vi[sizeof(long)==8?8:5]; void *__p[sizeof(long)==8?4:5]; } __u; } pthread_barrier_t; +#define __DEFINED_pthread_barrier_t +#endif + + +#undef _Addr +#undef _Int64 +#undef _Reg \ No newline at end of file diff --git a/lib/libc/include/i386-linux-musl/bits/fcntl.h b/lib/libc/include/m68k-linux-musl/bits/fcntl.h similarity index 83% rename from lib/libc/include/i386-linux-musl/bits/fcntl.h rename to lib/libc/include/m68k-linux-musl/bits/fcntl.h index 2514dfc23..5896a00fd 100644 --- a/lib/libc/include/i386-linux-musl/bits/fcntl.h +++ b/lib/libc/include/m68k-linux-musl/bits/fcntl.h @@ -7,13 +7,13 @@ #define O_DSYNC 010000 #define O_SYNC 04010000 #define O_RSYNC 04010000 -#define O_DIRECTORY 0200000 -#define O_NOFOLLOW 0400000 +#define O_DIRECTORY 040000 +#define O_NOFOLLOW 0100000 #define O_CLOEXEC 02000000 #define O_ASYNC 020000 -#define O_DIRECT 040000 -#define O_LARGEFILE 0100000 +#define O_DIRECT 0200000 +#define O_LARGEFILE 0400000 #define O_NOATIME 01000000 #define O_PATH 010000000 #define O_TMPFILE 020200000 diff --git a/lib/libc/include/m68k-linux-musl/bits/fenv.h b/lib/libc/include/m68k-linux-musl/bits/fenv.h new file mode 100644 index 000000000..6c9bbca08 --- /dev/null +++ b/lib/libc/include/m68k-linux-musl/bits/fenv.h @@ -0,0 +1,29 @@ +#if __HAVE_68881__ || __mcffpu__ + +#define FE_INEXACT 8 +#define FE_DIVBYZERO 16 +#define FE_UNDERFLOW 32 +#define FE_OVERFLOW 64 +#define FE_INVALID 128 + +#define FE_ALL_EXCEPT 0xf8 + +#define FE_TONEAREST 0 +#define FE_TOWARDZERO 16 +#define FE_DOWNWARD 32 +#define FE_UPWARD 48 + +#else + +#define FE_ALL_EXCEPT 0 +#define FE_TONEAREST 0 + +#endif + +typedef unsigned fexcept_t; + +typedef struct { + unsigned __control_register, __status_register, __instruction_address; +} fenv_t; + +#define FE_DFL_ENV ((const fenv_t *) -1) \ No newline at end of file diff --git a/lib/libc/include/m68k-linux-musl/bits/float.h b/lib/libc/include/m68k-linux-musl/bits/float.h new file mode 100644 index 000000000..e1e768927 --- /dev/null +++ b/lib/libc/include/m68k-linux-musl/bits/float.h @@ -0,0 +1,39 @@ +#if !__mcffpu__ + +#define FLT_EVAL_METHOD 2 + +#define LDBL_TRUE_MIN 3.6451995318824746025e-4951L +#define LDBL_MIN 1.68105157155604675313e-4932L +#define LDBL_MAX 1.1897314953572317650e+4932L +#define LDBL_EPSILON 1.0842021724855044340e-19L + +#define LDBL_MANT_DIG 64 +#define LDBL_MIN_EXP (-16382) +#define LDBL_MAX_EXP 16384 + +#define LDBL_DIG 18 +#define LDBL_MIN_10_EXP (-4931) +#define LDBL_MAX_10_EXP 4932 + +#define DECIMAL_DIG 21 + +#else + +#define FLT_EVAL_METHOD 0 + +#define LDBL_TRUE_MIN 4.94065645841246544177e-324L +#define LDBL_MIN 2.22507385850720138309e-308L +#define LDBL_MAX 1.79769313486231570815e+308L +#define LDBL_EPSILON 2.22044604925031308085e-16L + +#define LDBL_MANT_DIG 53 +#define LDBL_MIN_EXP (-1021) +#define LDBL_MAX_EXP 1024 + +#define LDBL_DIG 15 +#define LDBL_MIN_10_EXP (-307) +#define LDBL_MAX_10_EXP 308 + +#define DECIMAL_DIG 17 + +#endif \ No newline at end of file diff --git a/lib/libc/include/m68k-linux-musl/bits/ipcstat.h b/lib/libc/include/m68k-linux-musl/bits/ipcstat.h new file mode 100644 index 000000000..d055f3e83 --- /dev/null +++ b/lib/libc/include/m68k-linux-musl/bits/ipcstat.h @@ -0,0 +1 @@ +#define IPC_STAT 0x102 \ No newline at end of file diff --git a/lib/libc/include/m68k-linux-musl/bits/msg.h b/lib/libc/include/m68k-linux-musl/bits/msg.h new file mode 100644 index 000000000..037fd956d --- /dev/null +++ b/lib/libc/include/m68k-linux-musl/bits/msg.h @@ -0,0 +1,18 @@ +struct msqid_ds { + struct ipc_perm msg_perm; + unsigned long __msg_stime_lo; + unsigned long __msg_stime_hi; + unsigned long __msg_rtime_lo; + unsigned long __msg_rtime_hi; + unsigned long __msg_ctime_lo; + unsigned long __msg_ctime_hi; + unsigned long msg_cbytes; + msgqnum_t msg_qnum; + msglen_t msg_qbytes; + pid_t msg_lspid; + pid_t msg_lrpid; + unsigned long __unused[2]; + time_t msg_stime; + time_t msg_rtime; + time_t msg_ctime; +}; \ No newline at end of file diff --git a/lib/libc/include/m68k-linux-musl/bits/ptrace.h b/lib/libc/include/m68k-linux-musl/bits/ptrace.h new file mode 100644 index 000000000..aa2719745 --- /dev/null +++ b/lib/libc/include/m68k-linux-musl/bits/ptrace.h @@ -0,0 +1,2 @@ +#define PTRACE_GET_THREAD_AREA 25 +#define PTRACE_SINGLEBLOCK 33 \ No newline at end of file diff --git a/lib/libc/include/m68k-linux-musl/bits/reg.h b/lib/libc/include/m68k-linux-musl/bits/reg.h new file mode 100644 index 000000000..b1d525541 --- /dev/null +++ b/lib/libc/include/m68k-linux-musl/bits/reg.h @@ -0,0 +1,45 @@ +#undef __WORDSIZE +#define __WORDSIZE 32 +#define PT_D1 0 +#define PT_D2 1 +#define PT_D3 2 +#define PT_D4 3 +#define PT_D5 4 +#define PT_D6 5 +#define PT_D7 6 +#define PT_A0 7 +#define PT_A1 8 +#define PT_A2 9 +#define PT_A3 10 +#define PT_A4 11 +#define PT_A5 12 +#define PT_A6 13 +#define PT_D0 14 +#define PT_USP 15 +#define PT_ORIG_D0 16 +#define PT_SR 17 +#define PT_PC 18 + +#if __mcffpu__ +#define PT_FP0 21 +#define PT_FP1 23 +#define PT_FP2 25 +#define PT_FP3 27 +#define PT_FP4 29 +#define PT_FP5 31 +#define PT_FP6 33 +#define PT_FP7 35 +#else +#define PT_FP0 21 +#define PT_FP1 24 +#define PT_FP2 27 +#define PT_FP3 30 +#define PT_FP4 33 +#define PT_FP5 36 +#define PT_FP6 39 +#define PT_FP7 42 +#endif + +#define PT_FPCR 45 +#define PT_FPSR 46 +#define PT_FPIAR 47 \ No newline at end of file diff --git a/lib/libc/include/m68k-linux-musl/bits/sem.h b/lib/libc/include/m68k-linux-musl/bits/sem.h new file mode 100644 index 000000000..97738aaff --- /dev/null +++ b/lib/libc/include/m68k-linux-musl/bits/sem.h @@ -0,0 +1,13 @@ +struct semid_ds { + struct ipc_perm sem_perm; + unsigned long __sem_otime_lo; + unsigned long __sem_otime_hi; + unsigned long __sem_ctime_lo; + unsigned long __sem_ctime_hi; + char __sem_nsems_pad[sizeof(long)-sizeof(short)]; + unsigned short sem_nsems; + long __unused3; + long __unused4; + time_t sem_otime; + time_t sem_ctime; +}; \ No newline at end of file diff --git a/lib/libc/include/m68k-linux-musl/bits/setjmp.h b/lib/libc/include/m68k-linux-musl/bits/setjmp.h new file mode 100644 index 000000000..f4491c930 --- /dev/null +++ b/lib/libc/include/m68k-linux-musl/bits/setjmp.h @@ -0,0 +1 @@ +typedef unsigned long __jmp_buf[39]; \ No newline at end of file diff --git a/lib/libc/include/m68k-linux-musl/bits/shm.h b/lib/libc/include/m68k-linux-musl/bits/shm.h new file mode 100644 index 000000000..1ecb03ca0 --- /dev/null +++ b/lib/libc/include/m68k-linux-musl/bits/shm.h @@ -0,0 +1,31 @@ +#define SHMLBA 4096 + +struct shmid_ds { + struct ipc_perm shm_perm; + size_t shm_segsz; + unsigned long __shm_atime_lo; + unsigned long __shm_atime_hi; + unsigned long __shm_dtime_lo; + unsigned long __shm_dtime_hi; + unsigned long __shm_ctime_lo; + unsigned long __shm_ctime_hi; + pid_t shm_cpid; + pid_t shm_lpid; + unsigned long shm_nattch; + unsigned long __pad1; + unsigned long __pad2; + unsigned long __pad3; + time_t shm_atime; + time_t shm_dtime; + time_t shm_ctime; +}; + +struct shminfo { + unsigned long shmmax, shmmin, shmmni, shmseg, shmall, __unused[4]; +}; + +struct shm_info { + int __used_ids; + unsigned long shm_tot, shm_rss, shm_swp; + unsigned long __swap_attempts, __swap_successes; +}; \ No newline at end of file diff --git a/lib/libc/include/m68k-linux-musl/bits/signal.h b/lib/libc/include/m68k-linux-musl/bits/signal.h new file mode 100644 index 000000000..17cd3ad43 --- /dev/null +++ b/lib/libc/include/m68k-linux-musl/bits/signal.h @@ -0,0 +1,140 @@ +#if defined(_POSIX_SOURCE) || defined(_POSIX_C_SOURCE) \ + || defined(_XOPEN_SOURCE) || defined(_GNU_SOURCE) || defined(_BSD_SOURCE) + +#if defined(_XOPEN_SOURCE) || defined(_GNU_SOURCE) || defined(_BSD_SOURCE) +#define MINSIGSTKSZ 2048 +#define SIGSTKSZ 8192 +#endif + +#ifdef _GNU_SOURCE +enum { R_D0 = 0 }; +#define R_D0 R_D0 +enum { R_D1 = 1 }; +#define R_D1 R_D1 +enum { R_D2 = 2 }; +#define R_D2 R_D2 +enum { R_D3 = 3 }; +#define R_D3 R_D3 +enum { R_D4 = 4 }; +#define R_D4 R_D4 +enum { R_D5 = 5 }; +#define R_D5 R_D5 +enum { R_D6 = 6 }; +#define R_D6 R_D6 +enum { R_D7 = 7 }; +#define R_D7 R_D7 +enum { R_A0 = 8 }; +#define R_A0 R_A0 +enum { R_A1 = 9 }; +#define R_A1 R_A1 +enum { R_A2 = 10 }; +#define R_A2 R_A2 +enum { R_A3 = 11 }; +#define R_A3 R_A3 +enum { R_A4 = 12 }; +#define R_A4 R_A4 +enum { R_A5 = 13 }; +#define R_A5 R_A5 +enum { R_A6 = 14 }; +#define R_A6 R_A6 +enum { R_A7 = 15 }; +#define R_A7 R_A7 +enum { R_SP = 15 }; +#define R_SP R_SP +enum { R_PC = 16 }; +#define R_PC R_PC +enum { R_PS = 17 }; +#define R_PS R_PS +#endif + +#if defined(_GNU_SOURCE) || defined(_BSD_SOURCE) + +struct sigcontext { + unsigned long sc_mask, sc_usp, sc_d0, sc_d1, sc_a0, sc_a1; + unsigned short sc_sr; + unsigned long sc_pc; + unsigned short sc_formatvec; + unsigned long sc_fpregs[6], sc_fpcntl[3]; + unsigned char sc_fpstate[216]; +}; + +typedef int greg_t, gregset_t[18]; +typedef struct { + int f_pcr, f_psr, f_fpiaddr, f_fpregs[8][3]; +} fpregset_t; + +typedef struct { + int version; + gregset_t gregs; + fpregset_t fpregs; +} mcontext_t; +#else +typedef struct { + int __version; + int __gregs[18]; + int __fpregs[27]; +} mcontext_t; +#endif + +struct sigaltstack { + void *ss_sp; + int ss_flags; + size_t ss_size; +}; + +typedef struct __ucontext { + unsigned long uc_flags; + struct __ucontext *uc_link; + stack_t uc_stack; + mcontext_t uc_mcontext; + long __reserved[80]; + sigset_t uc_sigmask; +} ucontext_t; + +#define SA_NOCLDSTOP 1 +#define SA_NOCLDWAIT 2 +#define SA_SIGINFO 4 +#define SA_ONSTACK 0x08000000 +#define SA_RESTART 0x10000000 +#define SA_NODEFER 0x40000000 +#define SA_RESETHAND 0x80000000 +#define SA_RESTORER 0x04000000 + +#endif + +#define SIGHUP 1 +#define SIGINT 2 +#define SIGQUIT 3 +#define SIGILL 4 +#define SIGTRAP 5 +#define SIGABRT 6 +#define SIGIOT SIGABRT +#define SIGBUS 7 +#define SIGFPE 8 +#define SIGKILL 9 +#define SIGUSR1 10 +#define SIGSEGV 11 +#define SIGUSR2 12 +#define SIGPIPE 13 +#define SIGALRM 14 +#define SIGTERM 15 +#define SIGSTKFLT 16 +#define SIGCHLD 17 +#define SIGCONT 18 +#define SIGSTOP 19 +#define SIGTSTP 20 +#define SIGTTIN 21 +#define SIGTTOU 22 +#define SIGURG 23 +#define SIGXCPU 24 +#define SIGXFSZ 25 +#define SIGVTALRM 26 +#define SIGPROF 27 +#define SIGWINCH 28 +#define SIGIO 29 +#define SIGPOLL 29 +#define SIGPWR 30 +#define SIGSYS 31 +#define SIGUNUSED SIGSYS + +#define _NSIG 65 \ No newline at end of file diff --git a/lib/libc/include/m68k-linux-musl/bits/stat.h b/lib/libc/include/m68k-linux-musl/bits/stat.h new file mode 100644 index 000000000..9a0d732f7 --- /dev/null +++ b/lib/libc/include/m68k-linux-musl/bits/stat.h @@ -0,0 +1,25 @@ +/* copied from kernel definition, but with padding replaced + * by the corresponding correctly-sized userspace types. */ + +struct stat { + dev_t st_dev; + short __st_dev_padding; + long __st_ino_truncated; + mode_t st_mode; + nlink_t st_nlink; + uid_t st_uid; + gid_t st_gid; + dev_t st_rdev; + short __st_rdev_padding; + off_t st_size; + blksize_t st_blksize; + blkcnt_t st_blocks; + struct { + long tv_sec; + long tv_nsec; + } __st_atim32, __st_mtim32, __st_ctim32; + ino_t st_ino; + struct timespec st_atim; + struct timespec st_mtim; + struct timespec st_ctim; +}; \ No newline at end of file diff --git a/lib/libc/include/m68k-linux-musl/bits/stdint.h b/lib/libc/include/m68k-linux-musl/bits/stdint.h new file mode 100644 index 000000000..b70a87dc9 --- /dev/null +++ b/lib/libc/include/m68k-linux-musl/bits/stdint.h @@ -0,0 +1,20 @@ +typedef int32_t int_fast16_t; +typedef int32_t int_fast32_t; +typedef uint32_t uint_fast16_t; +typedef uint32_t uint_fast32_t; + +#define INT_FAST16_MIN INT32_MIN +#define INT_FAST32_MIN INT32_MIN + +#define INT_FAST16_MAX INT32_MAX +#define INT_FAST32_MAX INT32_MAX + +#define UINT_FAST16_MAX UINT32_MAX +#define UINT_FAST32_MAX UINT32_MAX + +#define INTPTR_MIN INT32_MIN +#define INTPTR_MAX INT32_MAX +#define UINTPTR_MAX UINT32_MAX +#define PTRDIFF_MIN INT32_MIN +#define PTRDIFF_MAX INT32_MAX +#define SIZE_MAX UINT32_MAX \ No newline at end of file diff --git a/lib/libc/include/m68k-linux-musl/bits/syscall.h b/lib/libc/include/m68k-linux-musl/bits/syscall.h new file mode 100644 index 000000000..01c5178ed --- /dev/null +++ b/lib/libc/include/m68k-linux-musl/bits/syscall.h @@ -0,0 +1,824 @@ +#define __NR_restart_syscall 0 +#define __NR_exit 1 +#define __NR_fork 2 +#define __NR_read 3 +#define __NR_write 4 +#define __NR_open 5 +#define __NR_close 6 +#define __NR_waitpid 7 +#define __NR_creat 8 +#define __NR_link 9 +#define __NR_unlink 10 +#define __NR_execve 11 +#define __NR_chdir 12 +#define __NR_time 13 +#define __NR_mknod 14 +#define __NR_chmod 15 +#define __NR_chown 16 +#define __NR_oldstat 18 +#define __NR_lseek 19 +#define __NR_getpid 20 +#define __NR_mount 21 +#define __NR_umount 22 +#define __NR_setuid 23 +#define __NR_getuid 24 +#define __NR_stime 25 +#define __NR_ptrace 26 +#define __NR_alarm 27 +#define __NR_oldfstat 28 +#define __NR_pause 29 +#define __NR_utime 30 +#define __NR_access 33 +#define __NR_nice 34 +#define __NR_sync 36 +#define __NR_kill 37 +#define __NR_rename 38 +#define __NR_mkdir 39 +#define __NR_rmdir 40 +#define __NR_dup 41 +#define __NR_pipe 42 +#define __NR_times 43 +#define __NR_brk 45 +#define __NR_setgid 46 +#define __NR_getgid 47 +#define __NR_signal 48 +#define __NR_geteuid 49 +#define __NR_getegid 50 +#define __NR_acct 51 +#define __NR_umount2 52 +#define __NR_ioctl 54 +#define __NR_fcntl 55 +#define __NR_setpgid 57 +#define __NR_umask 60 +#define __NR_chroot 61 +#define __NR_ustat 62 +#define __NR_dup2 63 +#define __NR_getppid 64 +#define __NR_getpgrp 65 +#define __NR_setsid 66 +#define __NR_sigaction 67 +#define __NR_sgetmask 68 +#define __NR_ssetmask 69 +#define __NR_setreuid 70 +#define __NR_setregid 71 +#define __NR_sigsuspend 72 +#define __NR_sigpending 73 +#define __NR_sethostname 74 +#define __NR_setrlimit 75 +#define __NR_getrlimit 76 +#define __NR_getrusage 77 +#define __NR_gettimeofday_time32 78 +#define __NR_settimeofday_time32 79 +#define __NR_getgroups 80 +#define __NR_setgroups 81 +#define __NR_select 82 +#define __NR_symlink 83 +#define __NR_oldlstat 84 +#define __NR_readlink 85 +#define __NR_uselib 86 +#define __NR_swapon 87 +#define __NR_reboot 88 +#define __NR_readdir 89 +#define __NR_mmap 90 +#define __NR_munmap 91 +#define __NR_truncate 92 +#define __NR_ftruncate 93 +#define __NR_fchmod 94 +#define __NR_fchown 95 +#define __NR_getpriority 96 +#define __NR_setpriority 97 +#define __NR_statfs 99 +#define __NR_fstatfs 100 +#define __NR_socketcall 102 +#define __NR_syslog 103 +#define __NR_setitimer 104 +#define __NR_getitimer 105 +#define __NR_stat 106 +#define __NR_lstat 107 +#define __NR_fstat 108 +#define __NR_vhangup 111 +#define __NR_wait4 114 +#define __NR_swapoff 115 +#define __NR_sysinfo 116 +#define __NR_ipc 117 +#define __NR_fsync 118 +#define __NR_sigreturn 119 +#define __NR_clone 120 +#define __NR_setdomainname 121 +#define __NR_uname 122 +#define __NR_cacheflush 123 +#define __NR_adjtimex 124 +#define __NR_mprotect 125 +#define __NR_sigprocmask 126 +#define __NR_create_module 127 +#define __NR_init_module 128 +#define __NR_delete_module 129 +#define __NR_get_kernel_syms 130 +#define __NR_quotactl 131 +#define __NR_getpgid 132 +#define __NR_fchdir 133 +#define __NR_bdflush 134 +#define __NR_sysfs 135 +#define __NR_personality 136 +#define __NR_setfsuid 138 +#define __NR_setfsgid 139 +#define __NR__llseek 140 +#define __NR_getdents 141 +#define __NR__newselect 142 +#define __NR_flock 143 +#define __NR_msync 144 +#define __NR_readv 145 +#define __NR_writev 146 +#define __NR_getsid 147 +#define __NR_fdatasync 148 +#define __NR__sysctl 149 +#define __NR_mlock 150 +#define __NR_munlock 151 +#define __NR_mlockall 152 +#define __NR_munlockall 153 +#define __NR_sched_setparam 154 +#define __NR_sched_getparam 155 +#define __NR_sched_setscheduler 156 +#define __NR_sched_getscheduler 157 +#define __NR_sched_yield 158 +#define __NR_sched_get_priority_max 159 +#define __NR_sched_get_priority_min 160 +#define __NR_sched_rr_get_interval 161 +#define __NR_nanosleep 162 +#define __NR_mremap 163 +#define __NR_setresuid 164 +#define __NR_getresuid 165 +#define __NR_getpagesize 166 +#define __NR_query_module 167 +#define __NR_poll 168 +#define __NR_nfsservctl 169 +#define __NR_setresgid 170 +#define __NR_getresgid 171 +#define __NR_prctl 172 +#define __NR_rt_sigreturn 173 +#define __NR_rt_sigaction 174 +#define __NR_rt_sigprocmask 175 +#define __NR_rt_sigpending 176 +#define __NR_rt_sigtimedwait 177 +#define __NR_rt_sigqueueinfo 178 +#define __NR_rt_sigsuspend 179 +#define __NR_pread64 180 +#define __NR_pwrite64 181 +#define __NR_lchown 182 +#define __NR_getcwd 183 +#define __NR_capget 184 +#define __NR_capset 185 +#define __NR_sigaltstack 186 +#define __NR_sendfile 187 +#define __NR_getpmsg 188 +#define __NR_putpmsg 189 +#define __NR_vfork 190 +#define __NR_ugetrlimit 191 +#define __NR_mmap2 192 +#define __NR_truncate64 193 +#define __NR_ftruncate64 194 +#define __NR_stat64 195 +#define __NR_lstat64 196 +#define __NR_fstat64 197 +#define __NR_chown32 198 +#define __NR_getuid32 199 +#define __NR_getgid32 200 +#define __NR_geteuid32 201 +#define __NR_getegid32 202 +#define __NR_setreuid32 203 +#define __NR_setregid32 204 +#define __NR_getgroups32 205 +#define __NR_setgroups32 206 +#define __NR_fchown32 207 +#define __NR_setresuid32 208 +#define __NR_getresuid32 209 +#define __NR_setresgid32 210 +#define __NR_getresgid32 211 +#define __NR_lchown32 212 +#define __NR_setuid32 213 +#define __NR_setgid32 214 +#define __NR_setfsuid32 215 +#define __NR_setfsgid32 216 +#define __NR_pivot_root 217 +#define __NR_getdents64 220 +#define __NR_gettid 221 +#define __NR_tkill 222 +#define __NR_setxattr 223 +#define __NR_lsetxattr 224 +#define __NR_fsetxattr 225 +#define __NR_getxattr 226 +#define __NR_lgetxattr 227 +#define __NR_fgetxattr 228 +#define __NR_listxattr 229 +#define __NR_llistxattr 230 +#define __NR_flistxattr 231 +#define __NR_removexattr 232 +#define __NR_lremovexattr 233 +#define __NR_fremovexattr 234 +#define __NR_futex 235 +#define __NR_sendfile64 236 +#define __NR_mincore 237 +#define __NR_madvise 238 +#define __NR_fcntl64 239 +#define __NR_readahead 240 +#define __NR_io_setup 241 +#define __NR_io_destroy 242 +#define __NR_io_getevents 243 +#define __NR_io_submit 244 +#define __NR_io_cancel 245 +#define __NR_fadvise64 246 +#define __NR_exit_group 247 +#define __NR_lookup_dcookie 248 +#define __NR_epoll_create 249 +#define __NR_epoll_ctl 250 +#define __NR_epoll_wait 251 +#define __NR_remap_file_pages 252 +#define __NR_set_tid_address 253 +#define __NR_timer_create 254 +#define __NR_timer_settime32 255 +#define __NR_timer_gettime32 256 +#define __NR_timer_getoverrun 257 +#define __NR_timer_delete 258 +#define __NR_clock_settime32 259 +#define __NR_clock_gettime32 260 +#define __NR_clock_getres_time32 261 +#define __NR_clock_nanosleep_time32 262 +#define __NR_statfs64 263 +#define __NR_fstatfs64 264 +#define __NR_tgkill 265 +#define __NR_utimes 266 +#define __NR_fadvise64_64 267 +#define __NR_mbind 268 +#define __NR_get_mempolicy 269 +#define __NR_set_mempolicy 270 +#define __NR_mq_open 271 +#define __NR_mq_unlink 272 +#define __NR_mq_timedsend 273 +#define __NR_mq_timedreceive 274 +#define __NR_mq_notify 275 +#define __NR_mq_getsetattr 276 +#define __NR_waitid 277 +#define __NR_add_key 279 +#define __NR_request_key 280 +#define __NR_keyctl 281 +#define __NR_ioprio_set 282 +#define __NR_ioprio_get 283 +#define __NR_inotify_init 284 +#define __NR_inotify_add_watch 285 +#define __NR_inotify_rm_watch 286 +#define __NR_migrate_pages 287 +#define __NR_openat 288 +#define __NR_mkdirat 289 +#define __NR_mknodat 290 +#define __NR_fchownat 291 +#define __NR_futimesat 292 +#define __NR_fstatat64 293 +#define __NR_unlinkat 294 +#define __NR_renameat 295 +#define __NR_linkat 296 +#define __NR_symlinkat 297 +#define __NR_readlinkat 298 +#define __NR_fchmodat 299 +#define __NR_faccessat 300 +#define __NR_pselect6 301 +#define __NR_ppoll 302 +#define __NR_unshare 303 +#define __NR_set_robust_list 304 +#define __NR_get_robust_list 305 +#define __NR_splice 306 +#define __NR_sync_file_range 307 +#define __NR_tee 308 +#define __NR_vmsplice 309 +#define __NR_move_pages 310 +#define __NR_sched_setaffinity 311 +#define __NR_sched_getaffinity 312 +#define __NR_kexec_load 313 +#define __NR_getcpu 314 +#define __NR_epoll_pwait 315 +#define __NR_utimensat 316 +#define __NR_signalfd 317 +#define __NR_timerfd_create 318 +#define __NR_eventfd 319 +#define __NR_fallocate 320 +#define __NR_timerfd_settime32 321 +#define __NR_timerfd_gettime32 322 +#define __NR_signalfd4 323 +#define __NR_eventfd2 324 +#define __NR_epoll_create1 325 +#define __NR_dup3 326 +#define __NR_pipe2 327 +#define __NR_inotify_init1 328 +#define __NR_preadv 329 +#define __NR_pwritev 330 +#define __NR_rt_tgsigqueueinfo 331 +#define __NR_perf_event_open 332 +#define __NR_get_thread_area 333 +#define __NR_set_thread_area 334 +#define __NR_atomic_cmpxchg_32 335 +#define __NR_atomic_barrier 336 +#define __NR_fanotify_init 337 +#define __NR_fanotify_mark 338 +#define __NR_prlimit64 339 +#define __NR_name_to_handle_at 340 +#define __NR_open_by_handle_at 341 +#define __NR_clock_adjtime 342 +#define __NR_syncfs 343 +#define __NR_setns 344 +#define __NR_process_vm_readv 345 +#define __NR_process_vm_writev 346 +#define __NR_kcmp 347 +#define __NR_finit_module 348 +#define __NR_sched_setattr 349 +#define __NR_sched_getattr 350 +#define __NR_renameat2 351 +#define __NR_getrandom 352 +#define __NR_memfd_create 353 +#define __NR_bpf 354 +#define __NR_execveat 355 +#define __NR_socket 356 +#define __NR_socketpair 357 +#define __NR_bind 358 +#define __NR_connect 359 +#define __NR_listen 360 +#define __NR_accept4 361 +#define __NR_getsockopt 362 +#define __NR_setsockopt 363 +#define __NR_getsockname 364 +#define __NR_getpeername 365 +#define __NR_sendto 366 +#define __NR_sendmsg 367 +#define __NR_recvfrom 368 +#define __NR_recvmsg 369 +#define __NR_shutdown 370 +#define __NR_recvmmsg 371 +#define __NR_sendmmsg 372 +#define __NR_userfaultfd 373 +#define __NR_membarrier 374 +#define __NR_mlock2 375 +#define __NR_copy_file_range 376 +#define __NR_preadv2 377 +#define __NR_pwritev2 378 +#define __NR_statx 379 +#define __NR_seccomp 380 +#define __NR_pkey_mprotect 381 +#define __NR_pkey_alloc 382 +#define __NR_pkey_free 383 +#define __NR_rseq 384 +#define __NR_semget 393 +#define __NR_semctl 394 +#define __NR_shmget 395 +#define __NR_shmctl 396 +#define __NR_shmat 397 +#define __NR_shmdt 398 +#define __NR_msgget 399 +#define __NR_msgsnd 400 +#define __NR_msgrcv 401 +#define __NR_msgctl 402 +#define __NR_clock_gettime64 403 +#define __NR_clock_settime64 404 +#define __NR_clock_adjtime64 405 +#define __NR_clock_getres_time64 406 +#define __NR_clock_nanosleep_time64 407 +#define __NR_timer_gettime64 408 +#define __NR_timer_settime64 409 +#define __NR_timerfd_gettime64 410 +#define __NR_timerfd_settime64 411 +#define __NR_utimensat_time64 412 +#define __NR_pselect6_time64 413 +#define __NR_ppoll_time64 414 +#define __NR_io_pgetevents_time64 416 +#define __NR_recvmmsg_time64 417 +#define __NR_mq_timedsend_time64 418 +#define __NR_mq_timedreceive_time64 419 +#define __NR_semtimedop_time64 420 +#define __NR_rt_sigtimedwait_time64 421 +#define __NR_futex_time64 422 +#define __NR_sched_rr_get_interval_time64 423 +#define __NR_pidfd_send_signal 424 +#define __NR_io_uring_setup 425 +#define __NR_io_uring_enter 426 +#define __NR_io_uring_register 427 +#define __NR_open_tree 428 +#define __NR_move_mount 429 +#define __NR_fsopen 430 +#define __NR_fsconfig 431 +#define __NR_fsmount 432 +#define __NR_fspick 433 +#define __NR_pidfd_open 434 +#define __NR_clone3 435 +#define __NR_close_range 436 +#define __NR_openat2 437 +#define __NR_pidfd_getfd 438 +#define __NR_faccessat2 439 +#define SYS_restart_syscall 0 +#define SYS_exit 1 +#define SYS_fork 2 +#define SYS_read 3 +#define SYS_write 4 +#define SYS_open 5 +#define SYS_close 6 +#define SYS_waitpid 7 +#define SYS_creat 8 +#define SYS_link 9 +#define SYS_unlink 10 +#define SYS_execve 11 +#define SYS_chdir 12 +#define SYS_time 13 +#define SYS_mknod 14 +#define SYS_chmod 15 +#define SYS_chown 16 +#define SYS_oldstat 18 +#define SYS_lseek 19 +#define SYS_getpid 20 +#define SYS_mount 21 +#define SYS_umount 22 +#define SYS_setuid 23 +#define SYS_getuid 24 +#define SYS_stime 25 +#define SYS_ptrace 26 +#define SYS_alarm 27 +#define SYS_oldfstat 28 +#define SYS_pause 29 +#define SYS_utime 30 +#define SYS_access 33 +#define SYS_nice 34 +#define SYS_sync 36 +#define SYS_kill 37 +#define SYS_rename 38 +#define SYS_mkdir 39 +#define SYS_rmdir 40 +#define SYS_dup 41 +#define SYS_pipe 42 +#define SYS_times 43 +#define SYS_brk 45 +#define SYS_setgid 46 +#define SYS_getgid 47 +#define SYS_signal 48 +#define SYS_geteuid 49 +#define SYS_getegid 50 +#define SYS_acct 51 +#define SYS_umount2 52 +#define SYS_ioctl 54 +#define SYS_fcntl 55 +#define SYS_setpgid 57 +#define SYS_umask 60 +#define SYS_chroot 61 +#define SYS_ustat 62 +#define SYS_dup2 63 +#define SYS_getppid 64 +#define SYS_getpgrp 65 +#define SYS_setsid 66 +#define SYS_sigaction 67 +#define SYS_sgetmask 68 +#define SYS_ssetmask 69 +#define SYS_setreuid 70 +#define SYS_setregid 71 +#define SYS_sigsuspend 72 +#define SYS_sigpending 73 +#define SYS_sethostname 74 +#define SYS_setrlimit 75 +#define SYS_getrlimit 76 +#define SYS_getrusage 77 +#define SYS_gettimeofday_time32 78 +#define SYS_settimeofday_time32 79 +#define SYS_getgroups 80 +#define SYS_setgroups 81 +#define SYS_select 82 +#define SYS_symlink 83 +#define SYS_oldlstat 84 +#define SYS_readlink 85 +#define SYS_uselib 86 +#define SYS_swapon 87 +#define SYS_reboot 88 +#define SYS_readdir 89 +#define SYS_mmap 90 +#define SYS_munmap 91 +#define SYS_truncate 92 +#define SYS_ftruncate 93 +#define SYS_fchmod 94 +#define SYS_fchown 95 +#define SYS_getpriority 96 +#define SYS_setpriority 97 +#define SYS_statfs 99 +#define SYS_fstatfs 100 +#define SYS_socketcall 102 +#define SYS_syslog 103 +#define SYS_setitimer 104 +#define SYS_getitimer 105 +#define SYS_stat 106 +#define SYS_lstat 107 +#define SYS_fstat 108 +#define SYS_vhangup 111 +#define SYS_wait4 114 +#define SYS_swapoff 115 +#define SYS_sysinfo 116 +#define SYS_ipc 117 +#define SYS_fsync 118 +#define SYS_sigreturn 119 +#define SYS_clone 120 +#define SYS_setdomainname 121 +#define SYS_uname 122 +#define SYS_cacheflush 123 +#define SYS_adjtimex 124 +#define SYS_mprotect 125 +#define SYS_sigprocmask 126 +#define SYS_create_module 127 +#define SYS_init_module 128 +#define SYS_delete_module 129 +#define SYS_get_kernel_syms 130 +#define SYS_quotactl 131 +#define SYS_getpgid 132 +#define SYS_fchdir 133 +#define SYS_bdflush 134 +#define SYS_sysfs 135 +#define SYS_personality 136 +#define SYS_setfsuid 138 +#define SYS_setfsgid 139 +#define SYS__llseek 140 +#define SYS_getdents 141 +#define SYS__newselect 142 +#define SYS_flock 143 +#define SYS_msync 144 +#define SYS_readv 145 +#define SYS_writev 146 +#define SYS_getsid 147 +#define SYS_fdatasync 148 +#define SYS__sysctl 149 +#define SYS_mlock 150 +#define SYS_munlock 151 +#define SYS_mlockall 152 +#define SYS_munlockall 153 +#define SYS_sched_setparam 154 +#define SYS_sched_getparam 155 +#define SYS_sched_setscheduler 156 +#define SYS_sched_getscheduler 157 +#define SYS_sched_yield 158 +#define SYS_sched_get_priority_max 159 +#define SYS_sched_get_priority_min 160 +#define SYS_sched_rr_get_interval 161 +#define SYS_nanosleep 162 +#define SYS_mremap 163 +#define SYS_setresuid 164 +#define SYS_getresuid 165 +#define SYS_getpagesize 166 +#define SYS_query_module 167 +#define SYS_poll 168 +#define SYS_nfsservctl 169 +#define SYS_setresgid 170 +#define SYS_getresgid 171 +#define SYS_prctl 172 +#define SYS_rt_sigreturn 173 +#define SYS_rt_sigaction 174 +#define SYS_rt_sigprocmask 175 +#define SYS_rt_sigpending 176 +#define SYS_rt_sigtimedwait 177 +#define SYS_rt_sigqueueinfo 178 +#define SYS_rt_sigsuspend 179 +#define SYS_pread64 180 +#define SYS_pwrite64 181 +#define SYS_lchown 182 +#define SYS_getcwd 183 +#define SYS_capget 184 +#define SYS_capset 185 +#define SYS_sigaltstack 186 +#define SYS_sendfile 187 +#define SYS_getpmsg 188 +#define SYS_putpmsg 189 +#define SYS_vfork 190 +#define SYS_ugetrlimit 191 +#define SYS_mmap2 192 +#define SYS_truncate64 193 +#define SYS_ftruncate64 194 +#define SYS_stat64 195 +#define SYS_lstat64 196 +#define SYS_fstat64 197 +#define SYS_chown32 198 +#define SYS_getuid32 199 +#define SYS_getgid32 200 +#define SYS_geteuid32 201 +#define SYS_getegid32 202 +#define SYS_setreuid32 203 +#define SYS_setregid32 204 +#define SYS_getgroups32 205 +#define SYS_setgroups32 206 +#define SYS_fchown32 207 +#define SYS_setresuid32 208 +#define SYS_getresuid32 209 +#define SYS_setresgid32 210 +#define SYS_getresgid32 211 +#define SYS_lchown32 212 +#define SYS_setuid32 213 +#define SYS_setgid32 214 +#define SYS_setfsuid32 215 +#define SYS_setfsgid32 216 +#define SYS_pivot_root 217 +#define SYS_getdents64 220 +#define SYS_gettid 221 +#define SYS_tkill 222 +#define SYS_setxattr 223 +#define SYS_lsetxattr 224 +#define SYS_fsetxattr 225 +#define SYS_getxattr 226 +#define SYS_lgetxattr 227 +#define SYS_fgetxattr 228 +#define SYS_listxattr 229 +#define SYS_llistxattr 230 +#define SYS_flistxattr 231 +#define SYS_removexattr 232 +#define SYS_lremovexattr 233 +#define SYS_fremovexattr 234 +#define SYS_futex 235 +#define SYS_sendfile64 236 +#define SYS_mincore 237 +#define SYS_madvise 238 +#define SYS_fcntl64 239 +#define SYS_readahead 240 +#define SYS_io_setup 241 +#define SYS_io_destroy 242 +#define SYS_io_getevents 243 +#define SYS_io_submit 244 +#define SYS_io_cancel 245 +#define SYS_fadvise64 246 +#define SYS_exit_group 247 +#define SYS_lookup_dcookie 248 +#define SYS_epoll_create 249 +#define SYS_epoll_ctl 250 +#define SYS_epoll_wait 251 +#define SYS_remap_file_pages 252 +#define SYS_set_tid_address 253 +#define SYS_timer_create 254 +#define SYS_timer_settime32 255 +#define SYS_timer_gettime32 256 +#define SYS_timer_getoverrun 257 +#define SYS_timer_delete 258 +#define SYS_clock_settime32 259 +#define SYS_clock_gettime32 260 +#define SYS_clock_getres_time32 261 +#define SYS_clock_nanosleep_time32 262 +#define SYS_statfs64 263 +#define SYS_fstatfs64 264 +#define SYS_tgkill 265 +#define SYS_utimes 266 +#define SYS_fadvise64_64 267 +#define SYS_mbind 268 +#define SYS_get_mempolicy 269 +#define SYS_set_mempolicy 270 +#define SYS_mq_open 271 +#define SYS_mq_unlink 272 +#define SYS_mq_timedsend 273 +#define SYS_mq_timedreceive 274 +#define SYS_mq_notify 275 +#define SYS_mq_getsetattr 276 +#define SYS_waitid 277 +#define SYS_add_key 279 +#define SYS_request_key 280 +#define SYS_keyctl 281 +#define SYS_ioprio_set 282 +#define SYS_ioprio_get 283 +#define SYS_inotify_init 284 +#define SYS_inotify_add_watch 285 +#define SYS_inotify_rm_watch 286 +#define SYS_migrate_pages 287 +#define SYS_openat 288 +#define SYS_mkdirat 289 +#define SYS_mknodat 290 +#define SYS_fchownat 291 +#define SYS_futimesat 292 +#define SYS_fstatat64 293 +#define SYS_unlinkat 294 +#define SYS_renameat 295 +#define SYS_linkat 296 +#define SYS_symlinkat 297 +#define SYS_readlinkat 298 +#define SYS_fchmodat 299 +#define SYS_faccessat 300 +#define SYS_pselect6 301 +#define SYS_ppoll 302 +#define SYS_unshare 303 +#define SYS_set_robust_list 304 +#define SYS_get_robust_list 305 +#define SYS_splice 306 +#define SYS_sync_file_range 307 +#define SYS_tee 308 +#define SYS_vmsplice 309 +#define SYS_move_pages 310 +#define SYS_sched_setaffinity 311 +#define SYS_sched_getaffinity 312 +#define SYS_kexec_load 313 +#define SYS_getcpu 314 +#define SYS_epoll_pwait 315 +#define SYS_utimensat 316 +#define SYS_signalfd 317 +#define SYS_timerfd_create 318 +#define SYS_eventfd 319 +#define SYS_fallocate 320 +#define SYS_timerfd_settime32 321 +#define SYS_timerfd_gettime32 322 +#define SYS_signalfd4 323 +#define SYS_eventfd2 324 +#define SYS_epoll_create1 325 +#define SYS_dup3 326 +#define SYS_pipe2 327 +#define SYS_inotify_init1 328 +#define SYS_preadv 329 +#define SYS_pwritev 330 +#define SYS_rt_tgsigqueueinfo 331 +#define SYS_perf_event_open 332 +#define SYS_get_thread_area 333 +#define SYS_set_thread_area 334 +#define SYS_atomic_cmpxchg_32 335 +#define SYS_atomic_barrier 336 +#define SYS_fanotify_init 337 +#define SYS_fanotify_mark 338 +#define SYS_prlimit64 339 +#define SYS_name_to_handle_at 340 +#define SYS_open_by_handle_at 341 +#define SYS_clock_adjtime 342 +#define SYS_syncfs 343 +#define SYS_setns 344 +#define SYS_process_vm_readv 345 +#define SYS_process_vm_writev 346 +#define SYS_kcmp 347 +#define SYS_finit_module 348 +#define SYS_sched_setattr 349 +#define SYS_sched_getattr 350 +#define SYS_renameat2 351 +#define SYS_getrandom 352 +#define SYS_memfd_create 353 +#define SYS_bpf 354 +#define SYS_execveat 355 +#define SYS_socket 356 +#define SYS_socketpair 357 +#define SYS_bind 358 +#define SYS_connect 359 +#define SYS_listen 360 +#define SYS_accept4 361 +#define SYS_getsockopt 362 +#define SYS_setsockopt 363 +#define SYS_getsockname 364 +#define SYS_getpeername 365 +#define SYS_sendto 366 +#define SYS_sendmsg 367 +#define SYS_recvfrom 368 +#define SYS_recvmsg 369 +#define SYS_shutdown 370 +#define SYS_recvmmsg 371 +#define SYS_sendmmsg 372 +#define SYS_userfaultfd 373 +#define SYS_membarrier 374 +#define SYS_mlock2 375 +#define SYS_copy_file_range 376 +#define SYS_preadv2 377 +#define SYS_pwritev2 378 +#define SYS_statx 379 +#define SYS_seccomp 380 +#define SYS_pkey_mprotect 381 +#define SYS_pkey_alloc 382 +#define SYS_pkey_free 383 +#define SYS_rseq 384 +#define SYS_semget 393 +#define SYS_semctl 394 +#define SYS_shmget 395 +#define SYS_shmctl 396 +#define SYS_shmat 397 +#define SYS_shmdt 398 +#define SYS_msgget 399 +#define SYS_msgsnd 400 +#define SYS_msgrcv 401 +#define SYS_msgctl 402 +#define SYS_clock_gettime64 403 +#define SYS_clock_settime64 404 +#define SYS_clock_adjtime64 405 +#define SYS_clock_getres_time64 406 +#define SYS_clock_nanosleep_time64 407 +#define SYS_timer_gettime64 408 +#define SYS_timer_settime64 409 +#define SYS_timerfd_gettime64 410 +#define SYS_timerfd_settime64 411 +#define SYS_utimensat_time64 412 +#define SYS_pselect6_time64 413 +#define SYS_ppoll_time64 414 +#define SYS_io_pgetevents_time64 416 +#define SYS_recvmmsg_time64 417 +#define SYS_mq_timedsend_time64 418 +#define SYS_mq_timedreceive_time64 419 +#define SYS_semtimedop_time64 420 +#define SYS_rt_sigtimedwait_time64 421 +#define SYS_futex_time64 422 +#define SYS_sched_rr_get_interval_time64 423 +#define SYS_pidfd_send_signal 424 +#define SYS_io_uring_setup 425 +#define SYS_io_uring_enter 426 +#define SYS_io_uring_register 427 +#define SYS_open_tree 428 +#define SYS_move_mount 429 +#define SYS_fsopen 430 +#define SYS_fsconfig 431 +#define SYS_fsmount 432 +#define SYS_fspick 433 +#define SYS_pidfd_open 434 +#define SYS_clone3 435 +#define SYS_close_range 436 +#define SYS_openat2 437 +#define SYS_pidfd_getfd 438 +#define SYS_faccessat2 439 \ No newline at end of file diff --git a/lib/libc/include/m68k-linux-musl/bits/user.h b/lib/libc/include/m68k-linux-musl/bits/user.h new file mode 100644 index 000000000..a18c06ced --- /dev/null +++ b/lib/libc/include/m68k-linux-musl/bits/user.h @@ -0,0 +1,38 @@ +#undef __WORDSIZE +#define __WORDSIZE 32 + +struct user_m68kfp_struct { + unsigned long fpregs[24], fpcntl[3]; +}; + +struct user_regs_struct { + long d1, d2, d3, d4, d5, d6, d7; + long a0, a1, a2, a3, a4, a5, a6; + long d0, usp, orig_d0; + short stkadj, sr; + long pc; + short fmtvec, __pad; +}; + +struct user { + struct user_regs_struct regs; + int u_fpvalid; + struct user_m68kfp_struct m68kfp; + unsigned long u_tsize, u_dsize, u_ssize, start_code, start_stack; + long signal; + int reserved; + unsigned long u_ar0; + struct user_m68kfp_struct *u_fpstate; + unsigned long magic; + char u_comm[32]; +}; + +#define ELF_NGREG 20 +typedef unsigned long elf_greg_t; +typedef elf_greg_t elf_gregset_t[ELF_NGREG]; +typedef struct user_m68kfp_struct elf_fpregset_t; + +#define NBPG 4096 +#define UPAGES 1 +#define HOST_TEXT_START_ADDR (u.start_code) +#define HOST_STACK_END_ADDR (u.start_stack + u.u_ssize * NBPG) \ No newline at end of file diff --git a/lib/libc/include/armeb-linux-gnueabi/bits/struct_rwlock.h b/lib/libc/include/mips-linux-gnu/bits/struct_rwlock.h similarity index 64% rename from lib/libc/include/armeb-linux-gnueabi/bits/struct_rwlock.h rename to lib/libc/include/mips-linux-gnu/bits/struct_rwlock.h index 7431c5597..6355f539b 100644 --- a/lib/libc/include/armeb-linux-gnueabi/bits/struct_rwlock.h +++ b/lib/libc/include/mips-linux-gnu/bits/struct_rwlock.h @@ -1,4 +1,4 @@ -/* Default read-write lock implementation struct definitions. +/* MIPS internal rwlock struct definitions. Copyright (C) 2019-2021 Free Software Foundation, Inc. This file is part of the GNU C Library. @@ -16,15 +16,8 @@ License along with the GNU C Library; if not, see . */ -#ifndef __RWLOCK_INTERNAL_H -#define __RWLOCK_INTERNAL_H - -#include - -/* Generic struct for both POSIX read-write lock. New ports are expected - to use the default layout, however archictetures can redefine it to add - arch-specific extensions (such as lock-elision). The struct have a size - of 32 bytes on both LP32 and LP64 architectures. */ +#ifndef _RWLOCK_INTERNAL_H +#define _RWLOCK_INTERNAL_H struct __pthread_rwlock_arch_t { @@ -34,28 +27,45 @@ struct __pthread_rwlock_arch_t unsigned int __writers_futex; unsigned int __pad3; unsigned int __pad4; - /* FLAGS must stay at its position in the structure to maintain +#if _MIPS_SIM == _ABI64 + int __cur_writer; + int __shared; + unsigned long int __pad1; + unsigned long int __pad2; + /* FLAGS must stay at this position in the structure to maintain binary compatibility. */ -#if __BYTE_ORDER == __BIG_ENDIAN + unsigned int __flags; +# else +# if __BYTE_ORDER == __BIG_ENDIAN unsigned char __pad1; unsigned char __pad2; unsigned char __shared; + /* FLAGS must stay at this position in the structure to maintain + binary compatibility. */ unsigned char __flags; -#else +# else + /* FLAGS must stay at this position in the structure to maintain + binary compatibility. */ unsigned char __flags; unsigned char __shared; unsigned char __pad1; unsigned char __pad2; -#endif +# endif int __cur_writer; +#endif }; -#if __BYTE_ORDER == __BIG_ENDIAN +#if _MIPS_SIM == _ABI64 # define __PTHREAD_RWLOCK_INITIALIZER(__flags) \ - 0, 0, 0, 0, 0, 0, 0, 0, 0, __flags, 0 + 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, __flags #else -# define __PTHREAD_RWLOCK_INITIALIZER(__flags) \ +# if __BYTE_ORDER == __BIG_ENDIAN +# define __PTHREAD_RWLOCK_INITIALIZER(__flags) \ + 0, 0, 0, 0, 0, 0, 0, 0, 0, __flags, 0 +# else +# define __PTHREAD_RWLOCK_INITIALIZER(__flags) \ 0, 0, 0, 0, 0, 0, __flags, 0, 0, 0, 0 +# endif #endif #endif \ No newline at end of file diff --git a/lib/libc/include/arm-linux-gnueabi/bits/wordsize.h b/lib/libc/include/mips-linux-gnu/bits/wordsize.h similarity index 67% rename from lib/libc/include/arm-linux-gnueabi/bits/wordsize.h rename to lib/libc/include/mips-linux-gnu/bits/wordsize.h index 6561e924f..9e405e4e4 100644 --- a/lib/libc/include/arm-linux-gnueabi/bits/wordsize.h +++ b/lib/libc/include/mips-linux-gnu/bits/wordsize.h @@ -1,4 +1,4 @@ -/* Copyright (C) 1999-2021 Free Software Foundation, Inc. +/* Copyright (C) 2002-2021 Free Software Foundation, Inc. This file is part of the GNU C Library. The GNU C Library is free software; you can redistribute it and/or @@ -12,10 +12,20 @@ Lesser General Public License for more details. You should have received a copy of the GNU Lesser General Public - License along with the GNU C Library; if not, see + License along with the GNU C Library. If not, see . */ -#define __WORDSIZE 32 -#define __WORDSIZE_TIME64_COMPAT32 0 +#include + +#define __WORDSIZE _MIPS_SZPTR + +#if _MIPS_SIM == _ABI64 +# define __WORDSIZE_TIME64_COMPAT32 1 +#else +# define __WORDSIZE_TIME64_COMPAT32 0 +#endif + +#if __WORDSIZE == 32 #define __WORDSIZE32_SIZE_ULONG 0 -#define __WORDSIZE32_PTRDIFF_LONG 0 \ No newline at end of file +#define __WORDSIZE32_PTRDIFF_LONG 0 +#endif \ No newline at end of file diff --git a/lib/libc/include/mips-linux-musl/bits/posix.h b/lib/libc/include/mips-linux-musl/bits/posix.h deleted file mode 100644 index b0aba3e4a..000000000 --- a/lib/libc/include/mips-linux-musl/bits/posix.h +++ /dev/null @@ -1,2 +0,0 @@ -#define _POSIX_V6_ILP32_OFFBIG 1 -#define _POSIX_V7_ILP32_OFFBIG 1 \ No newline at end of file diff --git a/lib/libc/include/arm-linux-gnueabi/bits/struct_rwlock.h b/lib/libc/include/mips64-linux-gnuabi64/bits/struct_rwlock.h similarity index 64% rename from lib/libc/include/arm-linux-gnueabi/bits/struct_rwlock.h rename to lib/libc/include/mips64-linux-gnuabi64/bits/struct_rwlock.h index 7431c5597..6355f539b 100644 --- a/lib/libc/include/arm-linux-gnueabi/bits/struct_rwlock.h +++ b/lib/libc/include/mips64-linux-gnuabi64/bits/struct_rwlock.h @@ -1,4 +1,4 @@ -/* Default read-write lock implementation struct definitions. +/* MIPS internal rwlock struct definitions. Copyright (C) 2019-2021 Free Software Foundation, Inc. This file is part of the GNU C Library. @@ -16,15 +16,8 @@ License along with the GNU C Library; if not, see . */ -#ifndef __RWLOCK_INTERNAL_H -#define __RWLOCK_INTERNAL_H - -#include - -/* Generic struct for both POSIX read-write lock. New ports are expected - to use the default layout, however archictetures can redefine it to add - arch-specific extensions (such as lock-elision). The struct have a size - of 32 bytes on both LP32 and LP64 architectures. */ +#ifndef _RWLOCK_INTERNAL_H +#define _RWLOCK_INTERNAL_H struct __pthread_rwlock_arch_t { @@ -34,28 +27,45 @@ struct __pthread_rwlock_arch_t unsigned int __writers_futex; unsigned int __pad3; unsigned int __pad4; - /* FLAGS must stay at its position in the structure to maintain +#if _MIPS_SIM == _ABI64 + int __cur_writer; + int __shared; + unsigned long int __pad1; + unsigned long int __pad2; + /* FLAGS must stay at this position in the structure to maintain binary compatibility. */ -#if __BYTE_ORDER == __BIG_ENDIAN + unsigned int __flags; +# else +# if __BYTE_ORDER == __BIG_ENDIAN unsigned char __pad1; unsigned char __pad2; unsigned char __shared; + /* FLAGS must stay at this position in the structure to maintain + binary compatibility. */ unsigned char __flags; -#else +# else + /* FLAGS must stay at this position in the structure to maintain + binary compatibility. */ unsigned char __flags; unsigned char __shared; unsigned char __pad1; unsigned char __pad2; -#endif +# endif int __cur_writer; +#endif }; -#if __BYTE_ORDER == __BIG_ENDIAN +#if _MIPS_SIM == _ABI64 # define __PTHREAD_RWLOCK_INITIALIZER(__flags) \ - 0, 0, 0, 0, 0, 0, 0, 0, 0, __flags, 0 + 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, __flags #else -# define __PTHREAD_RWLOCK_INITIALIZER(__flags) \ +# if __BYTE_ORDER == __BIG_ENDIAN +# define __PTHREAD_RWLOCK_INITIALIZER(__flags) \ + 0, 0, 0, 0, 0, 0, 0, 0, 0, __flags, 0 +# else +# define __PTHREAD_RWLOCK_INITIALIZER(__flags) \ 0, 0, 0, 0, 0, 0, __flags, 0, 0, 0, 0 +# endif #endif #endif \ No newline at end of file diff --git a/lib/libc/include/mips64-linux-gnuabi64/bits/wordsize.h b/lib/libc/include/mips64-linux-gnuabi64/bits/wordsize.h new file mode 100644 index 000000000..9e405e4e4 --- /dev/null +++ b/lib/libc/include/mips64-linux-gnuabi64/bits/wordsize.h @@ -0,0 +1,31 @@ +/* Copyright (C) 2002-2021 Free Software Foundation, Inc. + This file is part of the GNU C Library. + + The GNU C Library is free software; you can redistribute it and/or + modify it under the terms of the GNU Lesser General Public + License as published by the Free Software Foundation; either + version 2.1 of the License, or (at your option) any later version. + + The GNU C Library is distributed in the hope that it will be useful, + but WITHOUT ANY WARRANTY; without even the implied warranty of + MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU + Lesser General Public License for more details. + + You should have received a copy of the GNU Lesser General Public + License along with the GNU C Library. If not, see + . */ + +#include + +#define __WORDSIZE _MIPS_SZPTR + +#if _MIPS_SIM == _ABI64 +# define __WORDSIZE_TIME64_COMPAT32 1 +#else +# define __WORDSIZE_TIME64_COMPAT32 0 +#endif + +#if __WORDSIZE == 32 +#define __WORDSIZE32_SIZE_ULONG 0 +#define __WORDSIZE32_PTRDIFF_LONG 0 +#endif \ No newline at end of file diff --git a/lib/libc/include/arm-linux-gnueabihf/bits/struct_rwlock.h b/lib/libc/include/mips64-linux-gnuabin32/bits/struct_rwlock.h similarity index 64% rename from lib/libc/include/arm-linux-gnueabihf/bits/struct_rwlock.h rename to lib/libc/include/mips64-linux-gnuabin32/bits/struct_rwlock.h index 7431c5597..6355f539b 100644 --- a/lib/libc/include/arm-linux-gnueabihf/bits/struct_rwlock.h +++ b/lib/libc/include/mips64-linux-gnuabin32/bits/struct_rwlock.h @@ -1,4 +1,4 @@ -/* Default read-write lock implementation struct definitions. +/* MIPS internal rwlock struct definitions. Copyright (C) 2019-2021 Free Software Foundation, Inc. This file is part of the GNU C Library. @@ -16,15 +16,8 @@ License along with the GNU C Library; if not, see . */ -#ifndef __RWLOCK_INTERNAL_H -#define __RWLOCK_INTERNAL_H - -#include - -/* Generic struct for both POSIX read-write lock. New ports are expected - to use the default layout, however archictetures can redefine it to add - arch-specific extensions (such as lock-elision). The struct have a size - of 32 bytes on both LP32 and LP64 architectures. */ +#ifndef _RWLOCK_INTERNAL_H +#define _RWLOCK_INTERNAL_H struct __pthread_rwlock_arch_t { @@ -34,28 +27,45 @@ struct __pthread_rwlock_arch_t unsigned int __writers_futex; unsigned int __pad3; unsigned int __pad4; - /* FLAGS must stay at its position in the structure to maintain +#if _MIPS_SIM == _ABI64 + int __cur_writer; + int __shared; + unsigned long int __pad1; + unsigned long int __pad2; + /* FLAGS must stay at this position in the structure to maintain binary compatibility. */ -#if __BYTE_ORDER == __BIG_ENDIAN + unsigned int __flags; +# else +# if __BYTE_ORDER == __BIG_ENDIAN unsigned char __pad1; unsigned char __pad2; unsigned char __shared; + /* FLAGS must stay at this position in the structure to maintain + binary compatibility. */ unsigned char __flags; -#else +# else + /* FLAGS must stay at this position in the structure to maintain + binary compatibility. */ unsigned char __flags; unsigned char __shared; unsigned char __pad1; unsigned char __pad2; -#endif +# endif int __cur_writer; +#endif }; -#if __BYTE_ORDER == __BIG_ENDIAN +#if _MIPS_SIM == _ABI64 # define __PTHREAD_RWLOCK_INITIALIZER(__flags) \ - 0, 0, 0, 0, 0, 0, 0, 0, 0, __flags, 0 + 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, __flags #else -# define __PTHREAD_RWLOCK_INITIALIZER(__flags) \ +# if __BYTE_ORDER == __BIG_ENDIAN +# define __PTHREAD_RWLOCK_INITIALIZER(__flags) \ + 0, 0, 0, 0, 0, 0, 0, 0, 0, __flags, 0 +# else +# define __PTHREAD_RWLOCK_INITIALIZER(__flags) \ 0, 0, 0, 0, 0, 0, __flags, 0, 0, 0, 0 +# endif #endif #endif \ No newline at end of file diff --git a/lib/libc/include/mips64-linux-gnuabin32/bits/wordsize.h b/lib/libc/include/mips64-linux-gnuabin32/bits/wordsize.h new file mode 100644 index 000000000..9e405e4e4 --- /dev/null +++ b/lib/libc/include/mips64-linux-gnuabin32/bits/wordsize.h @@ -0,0 +1,31 @@ +/* Copyright (C) 2002-2021 Free Software Foundation, Inc. + This file is part of the GNU C Library. + + The GNU C Library is free software; you can redistribute it and/or + modify it under the terms of the GNU Lesser General Public + License as published by the Free Software Foundation; either + version 2.1 of the License, or (at your option) any later version. + + The GNU C Library is distributed in the hope that it will be useful, + but WITHOUT ANY WARRANTY; without even the implied warranty of + MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU + Lesser General Public License for more details. + + You should have received a copy of the GNU Lesser General Public + License along with the GNU C Library. If not, see + . */ + +#include + +#define __WORDSIZE _MIPS_SZPTR + +#if _MIPS_SIM == _ABI64 +# define __WORDSIZE_TIME64_COMPAT32 1 +#else +# define __WORDSIZE_TIME64_COMPAT32 0 +#endif + +#if __WORDSIZE == 32 +#define __WORDSIZE32_SIZE_ULONG 0 +#define __WORDSIZE32_PTRDIFF_LONG 0 +#endif \ No newline at end of file diff --git a/lib/libc/include/armeb-linux-gnueabihf/bits/struct_rwlock.h b/lib/libc/include/mips64el-linux-gnuabi64/bits/struct_rwlock.h similarity index 64% rename from lib/libc/include/armeb-linux-gnueabihf/bits/struct_rwlock.h rename to lib/libc/include/mips64el-linux-gnuabi64/bits/struct_rwlock.h index 7431c5597..6355f539b 100644 --- a/lib/libc/include/armeb-linux-gnueabihf/bits/struct_rwlock.h +++ b/lib/libc/include/mips64el-linux-gnuabi64/bits/struct_rwlock.h @@ -1,4 +1,4 @@ -/* Default read-write lock implementation struct definitions. +/* MIPS internal rwlock struct definitions. Copyright (C) 2019-2021 Free Software Foundation, Inc. This file is part of the GNU C Library. @@ -16,15 +16,8 @@ License along with the GNU C Library; if not, see . */ -#ifndef __RWLOCK_INTERNAL_H -#define __RWLOCK_INTERNAL_H - -#include - -/* Generic struct for both POSIX read-write lock. New ports are expected - to use the default layout, however archictetures can redefine it to add - arch-specific extensions (such as lock-elision). The struct have a size - of 32 bytes on both LP32 and LP64 architectures. */ +#ifndef _RWLOCK_INTERNAL_H +#define _RWLOCK_INTERNAL_H struct __pthread_rwlock_arch_t { @@ -34,28 +27,45 @@ struct __pthread_rwlock_arch_t unsigned int __writers_futex; unsigned int __pad3; unsigned int __pad4; - /* FLAGS must stay at its position in the structure to maintain +#if _MIPS_SIM == _ABI64 + int __cur_writer; + int __shared; + unsigned long int __pad1; + unsigned long int __pad2; + /* FLAGS must stay at this position in the structure to maintain binary compatibility. */ -#if __BYTE_ORDER == __BIG_ENDIAN + unsigned int __flags; +# else +# if __BYTE_ORDER == __BIG_ENDIAN unsigned char __pad1; unsigned char __pad2; unsigned char __shared; + /* FLAGS must stay at this position in the structure to maintain + binary compatibility. */ unsigned char __flags; -#else +# else + /* FLAGS must stay at this position in the structure to maintain + binary compatibility. */ unsigned char __flags; unsigned char __shared; unsigned char __pad1; unsigned char __pad2; -#endif +# endif int __cur_writer; +#endif }; -#if __BYTE_ORDER == __BIG_ENDIAN +#if _MIPS_SIM == _ABI64 # define __PTHREAD_RWLOCK_INITIALIZER(__flags) \ - 0, 0, 0, 0, 0, 0, 0, 0, 0, __flags, 0 + 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, __flags #else -# define __PTHREAD_RWLOCK_INITIALIZER(__flags) \ +# if __BYTE_ORDER == __BIG_ENDIAN +# define __PTHREAD_RWLOCK_INITIALIZER(__flags) \ + 0, 0, 0, 0, 0, 0, 0, 0, 0, __flags, 0 +# else +# define __PTHREAD_RWLOCK_INITIALIZER(__flags) \ 0, 0, 0, 0, 0, 0, __flags, 0, 0, 0, 0 +# endif #endif #endif \ No newline at end of file diff --git a/lib/libc/include/mips64el-linux-gnuabi64/bits/wordsize.h b/lib/libc/include/mips64el-linux-gnuabi64/bits/wordsize.h new file mode 100644 index 000000000..9e405e4e4 --- /dev/null +++ b/lib/libc/include/mips64el-linux-gnuabi64/bits/wordsize.h @@ -0,0 +1,31 @@ +/* Copyright (C) 2002-2021 Free Software Foundation, Inc. + This file is part of the GNU C Library. + + The GNU C Library is free software; you can redistribute it and/or + modify it under the terms of the GNU Lesser General Public + License as published by the Free Software Foundation; either + version 2.1 of the License, or (at your option) any later version. + + The GNU C Library is distributed in the hope that it will be useful, + but WITHOUT ANY WARRANTY; without even the implied warranty of + MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU + Lesser General Public License for more details. + + You should have received a copy of the GNU Lesser General Public + License along with the GNU C Library. If not, see + . */ + +#include + +#define __WORDSIZE _MIPS_SZPTR + +#if _MIPS_SIM == _ABI64 +# define __WORDSIZE_TIME64_COMPAT32 1 +#else +# define __WORDSIZE_TIME64_COMPAT32 0 +#endif + +#if __WORDSIZE == 32 +#define __WORDSIZE32_SIZE_ULONG 0 +#define __WORDSIZE32_PTRDIFF_LONG 0 +#endif \ No newline at end of file diff --git a/lib/libc/include/mips64el-linux-gnuabin32/bits/struct_rwlock.h b/lib/libc/include/mips64el-linux-gnuabin32/bits/struct_rwlock.h new file mode 100644 index 000000000..6355f539b --- /dev/null +++ b/lib/libc/include/mips64el-linux-gnuabin32/bits/struct_rwlock.h @@ -0,0 +1,71 @@ +/* MIPS internal rwlock struct definitions. + Copyright (C) 2019-2021 Free Software Foundation, Inc. + This file is part of the GNU C Library. + + The GNU C Library is free software; you can redistribute it and/or + modify it under the terms of the GNU Lesser General Public + License as published by the Free Software Foundation; either + version 2.1 of the License, or (at your option) any later version. + + The GNU C Library is distributed in the hope that it will be useful, + but WITHOUT ANY WARRANTY; without even the implied warranty of + MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU + Lesser General Public License for more details. + + You should have received a copy of the GNU Lesser General Public + License along with the GNU C Library; if not, see + . */ + +#ifndef _RWLOCK_INTERNAL_H +#define _RWLOCK_INTERNAL_H + +struct __pthread_rwlock_arch_t +{ + unsigned int __readers; + unsigned int __writers; + unsigned int __wrphase_futex; + unsigned int __writers_futex; + unsigned int __pad3; + unsigned int __pad4; +#if _MIPS_SIM == _ABI64 + int __cur_writer; + int __shared; + unsigned long int __pad1; + unsigned long int __pad2; + /* FLAGS must stay at this position in the structure to maintain + binary compatibility. */ + unsigned int __flags; +# else +# if __BYTE_ORDER == __BIG_ENDIAN + unsigned char __pad1; + unsigned char __pad2; + unsigned char __shared; + /* FLAGS must stay at this position in the structure to maintain + binary compatibility. */ + unsigned char __flags; +# else + /* FLAGS must stay at this position in the structure to maintain + binary compatibility. */ + unsigned char __flags; + unsigned char __shared; + unsigned char __pad1; + unsigned char __pad2; +# endif + int __cur_writer; +#endif +}; + +#if _MIPS_SIM == _ABI64 +# define __PTHREAD_RWLOCK_INITIALIZER(__flags) \ + 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, __flags +#else +# if __BYTE_ORDER == __BIG_ENDIAN +# define __PTHREAD_RWLOCK_INITIALIZER(__flags) \ + 0, 0, 0, 0, 0, 0, 0, 0, 0, __flags, 0 +# else +# define __PTHREAD_RWLOCK_INITIALIZER(__flags) \ + 0, 0, 0, 0, 0, 0, __flags, 0, 0, 0, 0 +# endif +#endif + +#endif \ No newline at end of file diff --git a/lib/libc/include/mips64el-linux-gnuabin32/bits/wordsize.h b/lib/libc/include/mips64el-linux-gnuabin32/bits/wordsize.h new file mode 100644 index 000000000..9e405e4e4 --- /dev/null +++ b/lib/libc/include/mips64el-linux-gnuabin32/bits/wordsize.h @@ -0,0 +1,31 @@ +/* Copyright (C) 2002-2021 Free Software Foundation, Inc. + This file is part of the GNU C Library. + + The GNU C Library is free software; you can redistribute it and/or + modify it under the terms of the GNU Lesser General Public + License as published by the Free Software Foundation; either + version 2.1 of the License, or (at your option) any later version. + + The GNU C Library is distributed in the hope that it will be useful, + but WITHOUT ANY WARRANTY; without even the implied warranty of + MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU + Lesser General Public License for more details. + + You should have received a copy of the GNU Lesser General Public + License along with the GNU C Library. If not, see + . */ + +#include + +#define __WORDSIZE _MIPS_SZPTR + +#if _MIPS_SIM == _ABI64 +# define __WORDSIZE_TIME64_COMPAT32 1 +#else +# define __WORDSIZE_TIME64_COMPAT32 0 +#endif + +#if __WORDSIZE == 32 +#define __WORDSIZE32_SIZE_ULONG 0 +#define __WORDSIZE32_PTRDIFF_LONG 0 +#endif \ No newline at end of file diff --git a/lib/libc/include/mipsel-linux-gnu/bits/struct_rwlock.h b/lib/libc/include/mipsel-linux-gnu/bits/struct_rwlock.h new file mode 100644 index 000000000..6355f539b --- /dev/null +++ b/lib/libc/include/mipsel-linux-gnu/bits/struct_rwlock.h @@ -0,0 +1,71 @@ +/* MIPS internal rwlock struct definitions. + Copyright (C) 2019-2021 Free Software Foundation, Inc. + This file is part of the GNU C Library. + + The GNU C Library is free software; you can redistribute it and/or + modify it under the terms of the GNU Lesser General Public + License as published by the Free Software Foundation; either + version 2.1 of the License, or (at your option) any later version. + + The GNU C Library is distributed in the hope that it will be useful, + but WITHOUT ANY WARRANTY; without even the implied warranty of + MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU + Lesser General Public License for more details. + + You should have received a copy of the GNU Lesser General Public + License along with the GNU C Library; if not, see + . */ + +#ifndef _RWLOCK_INTERNAL_H +#define _RWLOCK_INTERNAL_H + +struct __pthread_rwlock_arch_t +{ + unsigned int __readers; + unsigned int __writers; + unsigned int __wrphase_futex; + unsigned int __writers_futex; + unsigned int __pad3; + unsigned int __pad4; +#if _MIPS_SIM == _ABI64 + int __cur_writer; + int __shared; + unsigned long int __pad1; + unsigned long int __pad2; + /* FLAGS must stay at this position in the structure to maintain + binary compatibility. */ + unsigned int __flags; +# else +# if __BYTE_ORDER == __BIG_ENDIAN + unsigned char __pad1; + unsigned char __pad2; + unsigned char __shared; + /* FLAGS must stay at this position in the structure to maintain + binary compatibility. */ + unsigned char __flags; +# else + /* FLAGS must stay at this position in the structure to maintain + binary compatibility. */ + unsigned char __flags; + unsigned char __shared; + unsigned char __pad1; + unsigned char __pad2; +# endif + int __cur_writer; +#endif +}; + +#if _MIPS_SIM == _ABI64 +# define __PTHREAD_RWLOCK_INITIALIZER(__flags) \ + 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, __flags +#else +# if __BYTE_ORDER == __BIG_ENDIAN +# define __PTHREAD_RWLOCK_INITIALIZER(__flags) \ + 0, 0, 0, 0, 0, 0, 0, 0, 0, __flags, 0 +# else +# define __PTHREAD_RWLOCK_INITIALIZER(__flags) \ + 0, 0, 0, 0, 0, 0, __flags, 0, 0, 0, 0 +# endif +#endif + +#endif \ No newline at end of file diff --git a/lib/libc/include/mipsel-linux-gnu/bits/wordsize.h b/lib/libc/include/mipsel-linux-gnu/bits/wordsize.h new file mode 100644 index 000000000..9e405e4e4 --- /dev/null +++ b/lib/libc/include/mipsel-linux-gnu/bits/wordsize.h @@ -0,0 +1,31 @@ +/* Copyright (C) 2002-2021 Free Software Foundation, Inc. + This file is part of the GNU C Library. + + The GNU C Library is free software; you can redistribute it and/or + modify it under the terms of the GNU Lesser General Public + License as published by the Free Software Foundation; either + version 2.1 of the License, or (at your option) any later version. + + The GNU C Library is distributed in the hope that it will be useful, + but WITHOUT ANY WARRANTY; without even the implied warranty of + MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU + Lesser General Public License for more details. + + You should have received a copy of the GNU Lesser General Public + License along with the GNU C Library. If not, see + . */ + +#include + +#define __WORDSIZE _MIPS_SZPTR + +#if _MIPS_SIM == _ABI64 +# define __WORDSIZE_TIME64_COMPAT32 1 +#else +# define __WORDSIZE_TIME64_COMPAT32 0 +#endif + +#if __WORDSIZE == 32 +#define __WORDSIZE32_SIZE_ULONG 0 +#define __WORDSIZE32_PTRDIFF_LONG 0 +#endif \ No newline at end of file diff --git a/lib/libc/include/powerpc-linux-musl/bits/posix.h b/lib/libc/include/powerpc-linux-musl/bits/posix.h deleted file mode 100644 index b0aba3e4a..000000000 --- a/lib/libc/include/powerpc-linux-musl/bits/posix.h +++ /dev/null @@ -1,2 +0,0 @@ -#define _POSIX_V6_ILP32_OFFBIG 1 -#define _POSIX_V7_ILP32_OFFBIG 1 \ No newline at end of file diff --git a/lib/libc/include/powerpc64-linux-musl/bits/posix.h b/lib/libc/include/powerpc64-linux-musl/bits/posix.h new file mode 100644 index 000000000..9fcc205cc --- /dev/null +++ b/lib/libc/include/powerpc64-linux-musl/bits/posix.h @@ -0,0 +1,2 @@ +#define _POSIX_V6_LP64_OFF64 1 +#define _POSIX_V7_LP64_OFF64 1 \ No newline at end of file diff --git a/lib/libc/include/riscv64-linux-gnu/bits/floatn.h b/lib/libc/include/riscv64-linux-gnu/bits/floatn.h new file mode 100644 index 000000000..0b1b5d444 --- /dev/null +++ b/lib/libc/include/riscv64-linux-gnu/bits/floatn.h @@ -0,0 +1,97 @@ +/* Macros to control TS 18661-3 glibc features on ldbl-128 platforms. + Copyright (C) 2017-2021 Free Software Foundation, Inc. + This file is part of the GNU C Library. + + The GNU C Library is free software; you can redistribute it and/or + modify it under the terms of the GNU Lesser General Public + License as published by the Free Software Foundation; either + version 2.1 of the License, or (at your option) any later version. + + The GNU C Library is distributed in the hope that it will be useful, + but WITHOUT ANY WARRANTY; without even the implied warranty of + MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU + Lesser General Public License for more details. + + You should have received a copy of the GNU Lesser General Public + License along with the GNU C Library; if not, see + . */ + +#ifndef _BITS_FLOATN_H +#define _BITS_FLOATN_H + +#include +#include + +/* Defined to 1 if the current compiler invocation provides a + floating-point type with the IEEE 754 binary128 format, and this + glibc includes corresponding *f128 interfaces for it. */ +#ifndef __NO_LONG_DOUBLE_MATH +# define __HAVE_FLOAT128 1 +#else +/* glibc does not support _Float128 for platforms where long double is + normally binary128 when building with long double as binary64. + GCC's default for supported scalar modes does not support it either + in that case. */ +# define __HAVE_FLOAT128 0 +#endif + +/* Defined to 1 if __HAVE_FLOAT128 is 1 and the type is ABI-distinct + from the default float, double and long double types in this glibc. */ +#define __HAVE_DISTINCT_FLOAT128 0 + +/* Defined to 1 if the current compiler invocation provides a + floating-point type with the right format for _Float64x, and this + glibc includes corresponding *f64x interfaces for it. */ +#define __HAVE_FLOAT64X __HAVE_FLOAT128 + +/* Defined to 1 if __HAVE_FLOAT64X is 1 and _Float64x has the format + of long double. Otherwise, if __HAVE_FLOAT64X is 1, _Float64x has + the format of _Float128, which must be different from that of long + double. */ +#define __HAVE_FLOAT64X_LONG_DOUBLE __HAVE_FLOAT128 + +#ifndef __ASSEMBLER__ + +/* Defined to concatenate the literal suffix to be used with _Float128 + types, if __HAVE_FLOAT128 is 1. */ +# if __HAVE_FLOAT128 +# if !__GNUC_PREREQ (7, 0) || defined __cplusplus +/* The literal suffix f128 exists only since GCC 7.0. */ +# define __f128(x) x##l +# else +# define __f128(x) x##f128 +# endif +# endif + +/* Defined to a complex binary128 type if __HAVE_FLOAT128 is 1. */ +# if __HAVE_FLOAT128 +# if !__GNUC_PREREQ (7, 0) || defined __cplusplus +# define __CFLOAT128 _Complex long double +# else +# define __CFLOAT128 _Complex _Float128 +# endif +# endif + +/* The remaining of this file provides support for older compilers. */ +# if __HAVE_FLOAT128 + +/* The type _Float128 exists only since GCC 7.0. */ +# if !__GNUC_PREREQ (7, 0) || defined __cplusplus +typedef long double _Float128; +# endif + +/* Various built-in functions do not exist before GCC 7.0. */ +# if !__GNUC_PREREQ (7, 0) +# define __builtin_huge_valf128() (__builtin_huge_vall ()) +# define __builtin_inff128() (__builtin_infl ()) +# define __builtin_nanf128(x) (__builtin_nanl (x)) +# define __builtin_nansf128(x) (__builtin_nansl (x)) +# endif + +# endif + +#endif /* !__ASSEMBLER__. */ + +#include + +#endif /* _BITS_FLOATN_H */ \ No newline at end of file diff --git a/lib/libc/include/riscv64-linux-musl/bits/fcntl.h b/lib/libc/include/riscv64-linux-musl/bits/fcntl.h deleted file mode 100644 index 4d410ad75..000000000 --- a/lib/libc/include/riscv64-linux-musl/bits/fcntl.h +++ /dev/null @@ -1,38 +0,0 @@ -#define O_CREAT 0100 -#define O_EXCL 0200 -#define O_NOCTTY 0400 -#define O_TRUNC 01000 -#define O_APPEND 02000 -#define O_NONBLOCK 04000 -#define O_DSYNC 010000 -#define O_SYNC 04010000 -#define O_RSYNC 04010000 -#define O_DIRECTORY 0200000 -#define O_NOFOLLOW 0400000 -#define O_CLOEXEC 02000000 - -#define O_ASYNC 020000 -#define O_DIRECT 040000 -#define O_LARGEFILE 0100000 -#define O_NOATIME 01000000 -#define O_PATH 010000000 -#define O_TMPFILE 020200000 -#define O_NDELAY O_NONBLOCK - -#define F_DUPFD 0 -#define F_GETFD 1 -#define F_SETFD 2 -#define F_GETFL 3 -#define F_SETFL 4 -#define F_GETLK 5 -#define F_SETLK 6 -#define F_SETLKW 7 -#define F_SETOWN 8 -#define F_GETOWN 9 -#define F_SETSIG 10 -#define F_GETSIG 11 - -#define F_SETOWN_EX 15 -#define F_GETOWN_EX 16 - -#define F_GETOWNER_UIDS 17 \ No newline at end of file diff --git a/lib/libc/include/s390x-linux-gnu/bits/floatn.h b/lib/libc/include/s390x-linux-gnu/bits/floatn.h new file mode 100644 index 000000000..0b1b5d444 --- /dev/null +++ b/lib/libc/include/s390x-linux-gnu/bits/floatn.h @@ -0,0 +1,97 @@ +/* Macros to control TS 18661-3 glibc features on ldbl-128 platforms. + Copyright (C) 2017-2021 Free Software Foundation, Inc. + This file is part of the GNU C Library. + + The GNU C Library is free software; you can redistribute it and/or + modify it under the terms of the GNU Lesser General Public + License as published by the Free Software Foundation; either + version 2.1 of the License, or (at your option) any later version. + + The GNU C Library is distributed in the hope that it will be useful, + but WITHOUT ANY WARRANTY; without even the implied warranty of + MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU + Lesser General Public License for more details. + + You should have received a copy of the GNU Lesser General Public + License along with the GNU C Library; if not, see + . */ + +#ifndef _BITS_FLOATN_H +#define _BITS_FLOATN_H + +#include +#include + +/* Defined to 1 if the current compiler invocation provides a + floating-point type with the IEEE 754 binary128 format, and this + glibc includes corresponding *f128 interfaces for it. */ +#ifndef __NO_LONG_DOUBLE_MATH +# define __HAVE_FLOAT128 1 +#else +/* glibc does not support _Float128 for platforms where long double is + normally binary128 when building with long double as binary64. + GCC's default for supported scalar modes does not support it either + in that case. */ +# define __HAVE_FLOAT128 0 +#endif + +/* Defined to 1 if __HAVE_FLOAT128 is 1 and the type is ABI-distinct + from the default float, double and long double types in this glibc. */ +#define __HAVE_DISTINCT_FLOAT128 0 + +/* Defined to 1 if the current compiler invocation provides a + floating-point type with the right format for _Float64x, and this + glibc includes corresponding *f64x interfaces for it. */ +#define __HAVE_FLOAT64X __HAVE_FLOAT128 + +/* Defined to 1 if __HAVE_FLOAT64X is 1 and _Float64x has the format + of long double. Otherwise, if __HAVE_FLOAT64X is 1, _Float64x has + the format of _Float128, which must be different from that of long + double. */ +#define __HAVE_FLOAT64X_LONG_DOUBLE __HAVE_FLOAT128 + +#ifndef __ASSEMBLER__ + +/* Defined to concatenate the literal suffix to be used with _Float128 + types, if __HAVE_FLOAT128 is 1. */ +# if __HAVE_FLOAT128 +# if !__GNUC_PREREQ (7, 0) || defined __cplusplus +/* The literal suffix f128 exists only since GCC 7.0. */ +# define __f128(x) x##l +# else +# define __f128(x) x##f128 +# endif +# endif + +/* Defined to a complex binary128 type if __HAVE_FLOAT128 is 1. */ +# if __HAVE_FLOAT128 +# if !__GNUC_PREREQ (7, 0) || defined __cplusplus +# define __CFLOAT128 _Complex long double +# else +# define __CFLOAT128 _Complex _Float128 +# endif +# endif + +/* The remaining of this file provides support for older compilers. */ +# if __HAVE_FLOAT128 + +/* The type _Float128 exists only since GCC 7.0. */ +# if !__GNUC_PREREQ (7, 0) || defined __cplusplus +typedef long double _Float128; +# endif + +/* Various built-in functions do not exist before GCC 7.0. */ +# if !__GNUC_PREREQ (7, 0) +# define __builtin_huge_valf128() (__builtin_huge_vall ()) +# define __builtin_inff128() (__builtin_infl ()) +# define __builtin_nanf128(x) (__builtin_nanl (x)) +# define __builtin_nansf128(x) (__builtin_nansl (x)) +# endif + +# endif + +#endif /* !__ASSEMBLER__. */ + +#include + +#endif /* _BITS_FLOATN_H */ \ No newline at end of file diff --git a/lib/libc/include/s390x-linux-musl/bits/posix.h b/lib/libc/include/s390x-linux-musl/bits/posix.h new file mode 100644 index 000000000..9fcc205cc --- /dev/null +++ b/lib/libc/include/s390x-linux-musl/bits/posix.h @@ -0,0 +1,2 @@ +#define _POSIX_V6_LP64_OFF64 1 +#define _POSIX_V7_LP64_OFF64 1 \ No newline at end of file diff --git a/lib/libc/include/sparc-linux-gnu/bits/floatn.h b/lib/libc/include/sparc-linux-gnu/bits/floatn.h new file mode 100644 index 000000000..0b1b5d444 --- /dev/null +++ b/lib/libc/include/sparc-linux-gnu/bits/floatn.h @@ -0,0 +1,97 @@ +/* Macros to control TS 18661-3 glibc features on ldbl-128 platforms. + Copyright (C) 2017-2021 Free Software Foundation, Inc. + This file is part of the GNU C Library. + + The GNU C Library is free software; you can redistribute it and/or + modify it under the terms of the GNU Lesser General Public + License as published by the Free Software Foundation; either + version 2.1 of the License, or (at your option) any later version. + + The GNU C Library is distributed in the hope that it will be useful, + but WITHOUT ANY WARRANTY; without even the implied warranty of + MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU + Lesser General Public License for more details. + + You should have received a copy of the GNU Lesser General Public + License along with the GNU C Library; if not, see + . */ + +#ifndef _BITS_FLOATN_H +#define _BITS_FLOATN_H + +#include +#include + +/* Defined to 1 if the current compiler invocation provides a + floating-point type with the IEEE 754 binary128 format, and this + glibc includes corresponding *f128 interfaces for it. */ +#ifndef __NO_LONG_DOUBLE_MATH +# define __HAVE_FLOAT128 1 +#else +/* glibc does not support _Float128 for platforms where long double is + normally binary128 when building with long double as binary64. + GCC's default for supported scalar modes does not support it either + in that case. */ +# define __HAVE_FLOAT128 0 +#endif + +/* Defined to 1 if __HAVE_FLOAT128 is 1 and the type is ABI-distinct + from the default float, double and long double types in this glibc. */ +#define __HAVE_DISTINCT_FLOAT128 0 + +/* Defined to 1 if the current compiler invocation provides a + floating-point type with the right format for _Float64x, and this + glibc includes corresponding *f64x interfaces for it. */ +#define __HAVE_FLOAT64X __HAVE_FLOAT128 + +/* Defined to 1 if __HAVE_FLOAT64X is 1 and _Float64x has the format + of long double. Otherwise, if __HAVE_FLOAT64X is 1, _Float64x has + the format of _Float128, which must be different from that of long + double. */ +#define __HAVE_FLOAT64X_LONG_DOUBLE __HAVE_FLOAT128 + +#ifndef __ASSEMBLER__ + +/* Defined to concatenate the literal suffix to be used with _Float128 + types, if __HAVE_FLOAT128 is 1. */ +# if __HAVE_FLOAT128 +# if !__GNUC_PREREQ (7, 0) || defined __cplusplus +/* The literal suffix f128 exists only since GCC 7.0. */ +# define __f128(x) x##l +# else +# define __f128(x) x##f128 +# endif +# endif + +/* Defined to a complex binary128 type if __HAVE_FLOAT128 is 1. */ +# if __HAVE_FLOAT128 +# if !__GNUC_PREREQ (7, 0) || defined __cplusplus +# define __CFLOAT128 _Complex long double +# else +# define __CFLOAT128 _Complex _Float128 +# endif +# endif + +/* The remaining of this file provides support for older compilers. */ +# if __HAVE_FLOAT128 + +/* The type _Float128 exists only since GCC 7.0. */ +# if !__GNUC_PREREQ (7, 0) || defined __cplusplus +typedef long double _Float128; +# endif + +/* Various built-in functions do not exist before GCC 7.0. */ +# if !__GNUC_PREREQ (7, 0) +# define __builtin_huge_valf128() (__builtin_huge_vall ()) +# define __builtin_inff128() (__builtin_infl ()) +# define __builtin_nanf128(x) (__builtin_nanl (x)) +# define __builtin_nansf128(x) (__builtin_nansl (x)) +# endif + +# endif + +#endif /* !__ASSEMBLER__. */ + +#include + +#endif /* _BITS_FLOATN_H */ \ No newline at end of file diff --git a/lib/libc/include/sparcv9-linux-gnu/bits/floatn.h b/lib/libc/include/sparcv9-linux-gnu/bits/floatn.h new file mode 100644 index 000000000..0b1b5d444 --- /dev/null +++ b/lib/libc/include/sparcv9-linux-gnu/bits/floatn.h @@ -0,0 +1,97 @@ +/* Macros to control TS 18661-3 glibc features on ldbl-128 platforms. + Copyright (C) 2017-2021 Free Software Foundation, Inc. + This file is part of the GNU C Library. + + The GNU C Library is free software; you can redistribute it and/or + modify it under the terms of the GNU Lesser General Public + License as published by the Free Software Foundation; either + version 2.1 of the License, or (at your option) any later version. + + The GNU C Library is distributed in the hope that it will be useful, + but WITHOUT ANY WARRANTY; without even the implied warranty of + MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU + Lesser General Public License for more details. + + You should have received a copy of the GNU Lesser General Public + License along with the GNU C Library; if not, see + . */ + +#ifndef _BITS_FLOATN_H +#define _BITS_FLOATN_H + +#include +#include + +/* Defined to 1 if the current compiler invocation provides a + floating-point type with the IEEE 754 binary128 format, and this + glibc includes corresponding *f128 interfaces for it. */ +#ifndef __NO_LONG_DOUBLE_MATH +# define __HAVE_FLOAT128 1 +#else +/* glibc does not support _Float128 for platforms where long double is + normally binary128 when building with long double as binary64. + GCC's default for supported scalar modes does not support it either + in that case. */ +# define __HAVE_FLOAT128 0 +#endif + +/* Defined to 1 if __HAVE_FLOAT128 is 1 and the type is ABI-distinct + from the default float, double and long double types in this glibc. */ +#define __HAVE_DISTINCT_FLOAT128 0 + +/* Defined to 1 if the current compiler invocation provides a + floating-point type with the right format for _Float64x, and this + glibc includes corresponding *f64x interfaces for it. */ +#define __HAVE_FLOAT64X __HAVE_FLOAT128 + +/* Defined to 1 if __HAVE_FLOAT64X is 1 and _Float64x has the format + of long double. Otherwise, if __HAVE_FLOAT64X is 1, _Float64x has + the format of _Float128, which must be different from that of long + double. */ +#define __HAVE_FLOAT64X_LONG_DOUBLE __HAVE_FLOAT128 + +#ifndef __ASSEMBLER__ + +/* Defined to concatenate the literal suffix to be used with _Float128 + types, if __HAVE_FLOAT128 is 1. */ +# if __HAVE_FLOAT128 +# if !__GNUC_PREREQ (7, 0) || defined __cplusplus +/* The literal suffix f128 exists only since GCC 7.0. */ +# define __f128(x) x##l +# else +# define __f128(x) x##f128 +# endif +# endif + +/* Defined to a complex binary128 type if __HAVE_FLOAT128 is 1. */ +# if __HAVE_FLOAT128 +# if !__GNUC_PREREQ (7, 0) || defined __cplusplus +# define __CFLOAT128 _Complex long double +# else +# define __CFLOAT128 _Complex _Float128 +# endif +# endif + +/* The remaining of this file provides support for older compilers. */ +# if __HAVE_FLOAT128 + +/* The type _Float128 exists only since GCC 7.0. */ +# if !__GNUC_PREREQ (7, 0) || defined __cplusplus +typedef long double _Float128; +# endif + +/* Various built-in functions do not exist before GCC 7.0. */ +# if !__GNUC_PREREQ (7, 0) +# define __builtin_huge_valf128() (__builtin_huge_vall ()) +# define __builtin_inff128() (__builtin_infl ()) +# define __builtin_nanf128(x) (__builtin_nanl (x)) +# define __builtin_nansf128(x) (__builtin_nansl (x)) +# endif + +# endif + +#endif /* !__ASSEMBLER__. */ + +#include + +#endif /* _BITS_FLOATN_H */ \ No newline at end of file diff --git a/lib/libc/include/x86_64-linux-musl/bits/fcntl.h b/lib/libc/include/x86_64-linux-musl/bits/fcntl.h deleted file mode 100644 index 94aa51c8c..000000000 --- a/lib/libc/include/x86_64-linux-musl/bits/fcntl.h +++ /dev/null @@ -1,40 +0,0 @@ -#define O_CREAT 0100 -#define O_EXCL 0200 -#define O_NOCTTY 0400 -#define O_TRUNC 01000 -#define O_APPEND 02000 -#define O_NONBLOCK 04000 -#define O_DSYNC 010000 -#define O_SYNC 04010000 -#define O_RSYNC 04010000 -#define O_DIRECTORY 0200000 -#define O_NOFOLLOW 0400000 -#define O_CLOEXEC 02000000 - -#define O_ASYNC 020000 -#define O_DIRECT 040000 -#define O_LARGEFILE 0 -#define O_NOATIME 01000000 -#define O_PATH 010000000 -#define O_TMPFILE 020200000 -#define O_NDELAY O_NONBLOCK - -#define F_DUPFD 0 -#define F_GETFD 1 -#define F_SETFD 2 -#define F_GETFL 3 -#define F_SETFL 4 - -#define F_SETOWN 8 -#define F_GETOWN 9 -#define F_SETSIG 10 -#define F_GETSIG 11 - -#define F_GETLK 5 -#define F_SETLK 6 -#define F_SETLKW 7 - -#define F_SETOWN_EX 15 -#define F_GETOWN_EX 16 - -#define F_GETOWNER_UIDS 17 \ No newline at end of file diff --git a/lib/libc/include/x86_64-linux-musl/bits/posix.h b/lib/libc/include/x86_64-linux-musl/bits/posix.h new file mode 100644 index 000000000..9fcc205cc --- /dev/null +++ b/lib/libc/include/x86_64-linux-musl/bits/posix.h @@ -0,0 +1,2 @@ +#define _POSIX_V6_LP64_OFF64 1 +#define _POSIX_V7_LP64_OFF64 1 \ No newline at end of file diff --git a/lib/libcxx/include/__algorithm/adjacent_find.h b/lib/libcxx/include/__algorithm/adjacent_find.h new file mode 100644 index 000000000..0a2aa055c --- /dev/null +++ b/lib/libcxx/include/__algorithm/adjacent_find.h @@ -0,0 +1,51 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_ADJACENT_FIND_H +#define _LIBCPP___ALGORITHM_ADJACENT_FIND_H + +#include <__config> +#include <__algorithm/comp.h> +#include <__iterator/iterator_traits.h> + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +_LIBCPP_NODISCARD_EXT inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 _ForwardIterator +adjacent_find(_ForwardIterator __first, _ForwardIterator __last, _BinaryPredicate __pred) { + if (__first != __last) { + _ForwardIterator __i = __first; + while (++__i != __last) { + if (__pred(*__first, *__i)) + return __first; + __first = __i; + } + } + return __last; +} + +template +_LIBCPP_NODISCARD_EXT inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 _ForwardIterator +adjacent_find(_ForwardIterator __first, _ForwardIterator __last) { + typedef typename iterator_traits<_ForwardIterator>::value_type __v; + return _VSTD::adjacent_find(__first, __last, __equal_to<__v>()); +} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_ADJACENT_FIND_H diff --git a/lib/libcxx/include/__algorithm/all_of.h b/lib/libcxx/include/__algorithm/all_of.h new file mode 100644 index 000000000..7d6ed5085 --- /dev/null +++ b/lib/libcxx/include/__algorithm/all_of.h @@ -0,0 +1,37 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_ALL_OF_H +#define _LIBCPP___ALGORITHM_ALL_OF_H + +#include <__config> + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +_LIBCPP_NODISCARD_EXT inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 bool +all_of(_InputIterator __first, _InputIterator __last, _Predicate __pred) { + for (; __first != __last; ++__first) + if (!__pred(*__first)) + return false; + return true; +} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_ALL_OF_H diff --git a/lib/libcxx/include/__algorithm/any_of.h b/lib/libcxx/include/__algorithm/any_of.h new file mode 100644 index 000000000..d5a6c094a --- /dev/null +++ b/lib/libcxx/include/__algorithm/any_of.h @@ -0,0 +1,37 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_ANY_OF_H +#define _LIBCPP___ALGORITHM_ANY_OF_H + +#include <__config> + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +_LIBCPP_NODISCARD_EXT inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 bool +any_of(_InputIterator __first, _InputIterator __last, _Predicate __pred) { + for (; __first != __last; ++__first) + if (__pred(*__first)) + return true; + return false; +} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_ANY_OF_H diff --git a/lib/libcxx/include/__algorithm/binary_search.h b/lib/libcxx/include/__algorithm/binary_search.h new file mode 100644 index 000000000..766f5da16 --- /dev/null +++ b/lib/libcxx/include/__algorithm/binary_search.h @@ -0,0 +1,61 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_BINARY_SEARCH_H +#define _LIBCPP___ALGORITHM_BINARY_SEARCH_H + +#include <__config> +#include <__algorithm/comp.h> +#include <__algorithm/lower_bound.h> +#include <__algorithm/comp_ref_type.h> +#include <__iterator/iterator_traits.h> + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +bool +__binary_search(_ForwardIterator __first, _ForwardIterator __last, const _Tp& __value_, _Compare __comp) +{ + __first = _VSTD::__lower_bound<_Compare>(__first, __last, __value_, __comp); + return __first != __last && !__comp(__value_, *__first); +} + +template +_LIBCPP_NODISCARD_EXT inline +_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +bool +binary_search(_ForwardIterator __first, _ForwardIterator __last, const _Tp& __value_, _Compare __comp) +{ + typedef typename __comp_ref_type<_Compare>::type _Comp_ref; + return _VSTD::__binary_search<_Comp_ref>(__first, __last, __value_, __comp); +} + +template +_LIBCPP_NODISCARD_EXT inline +_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +bool +binary_search(_ForwardIterator __first, _ForwardIterator __last, const _Tp& __value_) +{ + return _VSTD::binary_search(__first, __last, __value_, + __less::value_type, _Tp>()); +} + + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_BINARY_SEARCH_H diff --git a/lib/libcxx/include/__algorithm/clamp.h b/lib/libcxx/include/__algorithm/clamp.h new file mode 100644 index 000000000..f8eba03cf --- /dev/null +++ b/lib/libcxx/include/__algorithm/clamp.h @@ -0,0 +1,52 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_CLAMP_H +#define _LIBCPP___ALGORITHM_CLAMP_H + +#include <__config> +#include <__debug> +#include <__algorithm/comp.h> + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +#if _LIBCPP_STD_VER > 14 +// clamp +template +_LIBCPP_NODISCARD_EXT inline +_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR +const _Tp& +clamp(const _Tp& __v, const _Tp& __lo, const _Tp& __hi, _Compare __comp) +{ + _LIBCPP_ASSERT(!__comp(__hi, __lo), "Bad bounds passed to std::clamp"); + return __comp(__v, __lo) ? __lo : __comp(__hi, __v) ? __hi : __v; + +} + +template +_LIBCPP_NODISCARD_EXT inline +_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR +const _Tp& +clamp(const _Tp& __v, const _Tp& __lo, const _Tp& __hi) +{ + return _VSTD::clamp(__v, __lo, __hi, __less<_Tp>()); +} +#endif + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_CLAMP_H diff --git a/lib/libcxx/include/__algorithm/comp.h b/lib/libcxx/include/__algorithm/comp.h new file mode 100644 index 000000000..2039cf7b3 --- /dev/null +++ b/lib/libcxx/include/__algorithm/comp.h @@ -0,0 +1,97 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_COMP_H +#define _LIBCPP___ALGORITHM_COMP_H + +#include <__config> + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +// I'd like to replace these with _VSTD::equal_to, but can't because: +// * That only works with C++14 and later, and +// * We haven't included here. +template +struct __equal_to +{ + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 bool operator()(const _T1& __x, const _T1& __y) const {return __x == __y;} + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 bool operator()(const _T1& __x, const _T2& __y) const {return __x == __y;} + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 bool operator()(const _T2& __x, const _T1& __y) const {return __x == __y;} + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 bool operator()(const _T2& __x, const _T2& __y) const {return __x == __y;} +}; + +template +struct __equal_to<_T1, _T1> +{ + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 + bool operator()(const _T1& __x, const _T1& __y) const {return __x == __y;} +}; + +template +struct __equal_to +{ + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 + bool operator()(const _T1& __x, const _T1& __y) const {return __x == __y;} +}; + +template +struct __equal_to<_T1, const _T1> +{ + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 + bool operator()(const _T1& __x, const _T1& __y) const {return __x == __y;} +}; + +template +struct __less +{ + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 + bool operator()(const _T1& __x, const _T1& __y) const {return __x < __y;} + + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 + bool operator()(const _T1& __x, const _T2& __y) const {return __x < __y;} + + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 + bool operator()(const _T2& __x, const _T1& __y) const {return __x < __y;} + + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 + bool operator()(const _T2& __x, const _T2& __y) const {return __x < __y;} +}; + +template +struct __less<_T1, _T1> +{ + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 + bool operator()(const _T1& __x, const _T1& __y) const {return __x < __y;} +}; + +template +struct __less +{ + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 + bool operator()(const _T1& __x, const _T1& __y) const {return __x < __y;} +}; + +template +struct __less<_T1, const _T1> +{ + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 + bool operator()(const _T1& __x, const _T1& __y) const {return __x < __y;} +}; + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_COMP_H diff --git a/lib/libcxx/include/__algorithm/comp_ref_type.h b/lib/libcxx/include/__algorithm/comp_ref_type.h new file mode 100644 index 000000000..b3bca82c0 --- /dev/null +++ b/lib/libcxx/include/__algorithm/comp_ref_type.h @@ -0,0 +1,87 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_COMP_REF_TYPE_H +#define _LIBCPP___ALGORITHM_COMP_REF_TYPE_H + +#include <__config> +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +#ifdef _LIBCPP_DEBUG + +template +struct __debug_less +{ + _Compare &__comp_; + _LIBCPP_CONSTEXPR_AFTER_CXX17 + __debug_less(_Compare& __c) : __comp_(__c) {} + + template + _LIBCPP_CONSTEXPR_AFTER_CXX17 + bool operator()(const _Tp& __x, const _Up& __y) + { + bool __r = __comp_(__x, __y); + if (__r) + __do_compare_assert(0, __y, __x); + return __r; + } + + template + _LIBCPP_CONSTEXPR_AFTER_CXX17 + bool operator()(_Tp& __x, _Up& __y) + { + bool __r = __comp_(__x, __y); + if (__r) + __do_compare_assert(0, __y, __x); + return __r; + } + + template + _LIBCPP_CONSTEXPR_AFTER_CXX17 + inline _LIBCPP_INLINE_VISIBILITY + decltype((void)declval<_Compare&>()( + declval<_LHS &>(), declval<_RHS &>())) + __do_compare_assert(int, _LHS & __l, _RHS & __r) { + _LIBCPP_ASSERT(!__comp_(__l, __r), + "Comparator does not induce a strict weak ordering"); + } + + template + _LIBCPP_CONSTEXPR_AFTER_CXX17 + inline _LIBCPP_INLINE_VISIBILITY + void __do_compare_assert(long, _LHS &, _RHS &) {} +}; + +#endif // _LIBCPP_DEBUG + +template +struct __comp_ref_type { + // Pass the comparator by lvalue reference. Or in debug mode, using a + // debugging wrapper that stores a reference. +#ifndef _LIBCPP_DEBUG + typedef typename add_lvalue_reference<_Comp>::type type; +#else + typedef __debug_less<_Comp> type; +#endif +}; + + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_COMP_REF_TYPE_H diff --git a/lib/libcxx/include/__algorithm/copy.h b/lib/libcxx/include/__algorithm/copy.h new file mode 100644 index 000000000..9db7434c0 --- /dev/null +++ b/lib/libcxx/include/__algorithm/copy.h @@ -0,0 +1,82 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_COPY_H +#define _LIBCPP___ALGORITHM_COPY_H + +#include <__config> +#include <__algorithm/unwrap_iter.h> +#include <__iterator/iterator_traits.h> +#include +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +// copy + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +_OutputIterator +__copy_constexpr(_InputIterator __first, _InputIterator __last, _OutputIterator __result) +{ + for (; __first != __last; ++__first, (void) ++__result) + *__result = *__first; + return __result; +} + +template +inline _LIBCPP_INLINE_VISIBILITY +_OutputIterator +__copy(_InputIterator __first, _InputIterator __last, _OutputIterator __result) +{ + return _VSTD::__copy_constexpr(__first, __last, __result); +} + +template +inline _LIBCPP_INLINE_VISIBILITY +typename enable_if +< + is_same::type, _Up>::value && + is_trivially_copy_assignable<_Up>::value, + _Up* +>::type +__copy(_Tp* __first, _Tp* __last, _Up* __result) +{ + const size_t __n = static_cast(__last - __first); + if (__n > 0) + _VSTD::memmove(__result, __first, __n * sizeof(_Up)); + return __result + __n; +} + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +_OutputIterator +copy(_InputIterator __first, _InputIterator __last, _OutputIterator __result) +{ + if (__libcpp_is_constant_evaluated()) { + return _VSTD::__copy_constexpr(__first, __last, __result); + } else { + return _VSTD::__rewrap_iter(__result, + _VSTD::__copy(_VSTD::__unwrap_iter(__first), + _VSTD::__unwrap_iter(__last), + _VSTD::__unwrap_iter(__result))); + } +} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_COPY_H diff --git a/lib/libcxx/include/__algorithm/copy_backward.h b/lib/libcxx/include/__algorithm/copy_backward.h new file mode 100644 index 000000000..03a9c5f2d --- /dev/null +++ b/lib/libcxx/include/__algorithm/copy_backward.h @@ -0,0 +1,84 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_COPY_BACKWARD_H +#define _LIBCPP___ALGORITHM_COPY_BACKWARD_H + +#include <__config> +#include <__algorithm/unwrap_iter.h> +#include <__iterator/iterator_traits.h> +#include +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +_OutputIterator +__copy_backward_constexpr(_BidirectionalIterator __first, _BidirectionalIterator __last, _OutputIterator __result) +{ + while (__first != __last) + *--__result = *--__last; + return __result; +} + +template +inline _LIBCPP_INLINE_VISIBILITY +_OutputIterator +__copy_backward(_BidirectionalIterator __first, _BidirectionalIterator __last, _OutputIterator __result) +{ + return _VSTD::__copy_backward_constexpr(__first, __last, __result); +} + +template +inline _LIBCPP_INLINE_VISIBILITY +typename enable_if +< + is_same::type, _Up>::value && + is_trivially_copy_assignable<_Up>::value, + _Up* +>::type +__copy_backward(_Tp* __first, _Tp* __last, _Up* __result) +{ + const size_t __n = static_cast(__last - __first); + if (__n > 0) + { + __result -= __n; + _VSTD::memmove(__result, __first, __n * sizeof(_Up)); + } + return __result; +} + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +_BidirectionalIterator2 +copy_backward(_BidirectionalIterator1 __first, _BidirectionalIterator1 __last, + _BidirectionalIterator2 __result) +{ + if (__libcpp_is_constant_evaluated()) { + return _VSTD::__copy_backward_constexpr(__first, __last, __result); + } else { + return _VSTD::__rewrap_iter(__result, + _VSTD::__copy_backward(_VSTD::__unwrap_iter(__first), + _VSTD::__unwrap_iter(__last), + _VSTD::__unwrap_iter(__result))); + } +} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_COPY_BACKWARD_H diff --git a/lib/libcxx/include/__algorithm/copy_if.h b/lib/libcxx/include/__algorithm/copy_if.h new file mode 100644 index 000000000..153304c7a --- /dev/null +++ b/lib/libcxx/include/__algorithm/copy_if.h @@ -0,0 +1,48 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_COPY_IF_H +#define _LIBCPP___ALGORITHM_COPY_IF_H + +#include <__config> +#include <__algorithm/unwrap_iter.h> +#include <__iterator/iterator_traits.h> +#include +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +_OutputIterator +copy_if(_InputIterator __first, _InputIterator __last, + _OutputIterator __result, _Predicate __pred) +{ + for (; __first != __last; ++__first) + { + if (__pred(*__first)) + { + *__result = *__first; + ++__result; + } + } + return __result; +} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_COPY_IF_H diff --git a/lib/libcxx/include/__algorithm/copy_n.h b/lib/libcxx/include/__algorithm/copy_n.h new file mode 100644 index 000000000..bbfeb8686 --- /dev/null +++ b/lib/libcxx/include/__algorithm/copy_n.h @@ -0,0 +1,72 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_COPY_N_H +#define _LIBCPP___ALGORITHM_COPY_N_H + +#include <__config> +#include <__algorithm/copy.h> +#include <__algorithm/unwrap_iter.h> +#include <__iterator/iterator_traits.h> +#include +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +typename enable_if +< + __is_cpp17_input_iterator<_InputIterator>::value && + !__is_cpp17_random_access_iterator<_InputIterator>::value, + _OutputIterator +>::type +copy_n(_InputIterator __first, _Size __orig_n, _OutputIterator __result) +{ + typedef decltype(_VSTD::__convert_to_integral(__orig_n)) _IntegralSize; + _IntegralSize __n = __orig_n; + if (__n > 0) + { + *__result = *__first; + ++__result; + for (--__n; __n > 0; --__n) + { + ++__first; + *__result = *__first; + ++__result; + } + } + return __result; +} + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +typename enable_if +< + __is_cpp17_random_access_iterator<_InputIterator>::value, + _OutputIterator +>::type +copy_n(_InputIterator __first, _Size __orig_n, _OutputIterator __result) +{ + typedef decltype(_VSTD::__convert_to_integral(__orig_n)) _IntegralSize; + _IntegralSize __n = __orig_n; + return _VSTD::copy(__first, __first + __n, __result); +} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_COPY_N_H diff --git a/lib/libcxx/include/__algorithm/count.h b/lib/libcxx/include/__algorithm/count.h new file mode 100644 index 000000000..7f2d1954c --- /dev/null +++ b/lib/libcxx/include/__algorithm/count.h @@ -0,0 +1,40 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_COUNT_H +#define _LIBCPP___ALGORITHM_COUNT_H + +#include <__config> +#include <__iterator/iterator_traits.h> + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +_LIBCPP_NODISCARD_EXT inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 + typename iterator_traits<_InputIterator>::difference_type + count(_InputIterator __first, _InputIterator __last, const _Tp& __value_) { + typename iterator_traits<_InputIterator>::difference_type __r(0); + for (; __first != __last; ++__first) + if (*__first == __value_) + ++__r; + return __r; +} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_COUNT_H diff --git a/lib/libcxx/include/__algorithm/count_if.h b/lib/libcxx/include/__algorithm/count_if.h new file mode 100644 index 000000000..a5efffb92 --- /dev/null +++ b/lib/libcxx/include/__algorithm/count_if.h @@ -0,0 +1,40 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_COUNT_IF_H +#define _LIBCPP___ALGORITHM_COUNT_IF_H + +#include <__config> +#include <__iterator/iterator_traits.h> + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +_LIBCPP_NODISCARD_EXT inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 + typename iterator_traits<_InputIterator>::difference_type + count_if(_InputIterator __first, _InputIterator __last, _Predicate __pred) { + typename iterator_traits<_InputIterator>::difference_type __r(0); + for (; __first != __last; ++__first) + if (__pred(*__first)) + ++__r; + return __r; +} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_COUNT_IF_H diff --git a/lib/libcxx/include/__algorithm/equal.h b/lib/libcxx/include/__algorithm/equal.h new file mode 100644 index 000000000..bc6755981 --- /dev/null +++ b/lib/libcxx/include/__algorithm/equal.h @@ -0,0 +1,90 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_EQUAL_H +#define _LIBCPP___ALGORITHM_EQUAL_H + +#include <__config> +#include <__algorithm/comp.h> +#include <__iterator/iterator_traits.h> +#include // FIXME: replace with <__iterator/distance.h> when it lands + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +_LIBCPP_NODISCARD_EXT inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 bool +equal(_InputIterator1 __first1, _InputIterator1 __last1, _InputIterator2 __first2, _BinaryPredicate __pred) { + for (; __first1 != __last1; ++__first1, (void)++__first2) + if (!__pred(*__first1, *__first2)) + return false; + return true; +} + +template +_LIBCPP_NODISCARD_EXT inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 bool +equal(_InputIterator1 __first1, _InputIterator1 __last1, _InputIterator2 __first2) { + typedef typename iterator_traits<_InputIterator1>::value_type __v1; + typedef typename iterator_traits<_InputIterator2>::value_type __v2; + return _VSTD::equal(__first1, __last1, __first2, __equal_to<__v1, __v2>()); +} + +#if _LIBCPP_STD_VER > 11 +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 bool +__equal(_InputIterator1 __first1, _InputIterator1 __last1, _InputIterator2 __first2, _InputIterator2 __last2, + _BinaryPredicate __pred, input_iterator_tag, input_iterator_tag) { + for (; __first1 != __last1 && __first2 != __last2; ++__first1, (void)++__first2) + if (!__pred(*__first1, *__first2)) + return false; + return __first1 == __last1 && __first2 == __last2; +} + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 bool +__equal(_RandomAccessIterator1 __first1, _RandomAccessIterator1 __last1, _RandomAccessIterator2 __first2, + _RandomAccessIterator2 __last2, _BinaryPredicate __pred, random_access_iterator_tag, + random_access_iterator_tag) { + if (_VSTD::distance(__first1, __last1) != _VSTD::distance(__first2, __last2)) + return false; + return _VSTD::equal<_RandomAccessIterator1, _RandomAccessIterator2, + typename add_lvalue_reference<_BinaryPredicate>::type>(__first1, __last1, __first2, __pred); +} + +template +_LIBCPP_NODISCARD_EXT inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 bool +equal(_InputIterator1 __first1, _InputIterator1 __last1, _InputIterator2 __first2, _InputIterator2 __last2, + _BinaryPredicate __pred) { + return _VSTD::__equal::type>( + __first1, __last1, __first2, __last2, __pred, typename iterator_traits<_InputIterator1>::iterator_category(), + typename iterator_traits<_InputIterator2>::iterator_category()); +} + +template +_LIBCPP_NODISCARD_EXT inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 bool +equal(_InputIterator1 __first1, _InputIterator1 __last1, _InputIterator2 __first2, _InputIterator2 __last2) { + typedef typename iterator_traits<_InputIterator1>::value_type __v1; + typedef typename iterator_traits<_InputIterator2>::value_type __v2; + return _VSTD::__equal(__first1, __last1, __first2, __last2, __equal_to<__v1, __v2>(), + typename iterator_traits<_InputIterator1>::iterator_category(), + typename iterator_traits<_InputIterator2>::iterator_category()); +} +#endif + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_EQUAL_H diff --git a/lib/libcxx/include/__algorithm/equal_range.h b/lib/libcxx/include/__algorithm/equal_range.h new file mode 100644 index 000000000..9694dae95 --- /dev/null +++ b/lib/libcxx/include/__algorithm/equal_range.h @@ -0,0 +1,87 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_EQUAL_RANGE_H +#define _LIBCPP___ALGORITHM_EQUAL_RANGE_H + +#include <__config> +#include <__algorithm/comp.h> +#include <__algorithm/comp_ref_type.h> +#include <__algorithm/half_positive.h> +#include <__algorithm/lower_bound.h> +#include <__algorithm/upper_bound.h> +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +_LIBCPP_CONSTEXPR_AFTER_CXX17 pair<_ForwardIterator, _ForwardIterator> +__equal_range(_ForwardIterator __first, _ForwardIterator __last, const _Tp& __value_, _Compare __comp) +{ + typedef typename iterator_traits<_ForwardIterator>::difference_type difference_type; + difference_type __len = _VSTD::distance(__first, __last); + while (__len != 0) + { + difference_type __l2 = _VSTD::__half_positive(__len); + _ForwardIterator __m = __first; + _VSTD::advance(__m, __l2); + if (__comp(*__m, __value_)) + { + __first = ++__m; + __len -= __l2 + 1; + } + else if (__comp(__value_, *__m)) + { + __last = __m; + __len = __l2; + } + else + { + _ForwardIterator __mp1 = __m; + return pair<_ForwardIterator, _ForwardIterator> + ( + _VSTD::__lower_bound<_Compare>(__first, __m, __value_, __comp), + _VSTD::__upper_bound<_Compare>(++__mp1, __last, __value_, __comp) + ); + } + } + return pair<_ForwardIterator, _ForwardIterator>(__first, __first); +} + +template +_LIBCPP_NODISCARD_EXT inline +_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +pair<_ForwardIterator, _ForwardIterator> +equal_range(_ForwardIterator __first, _ForwardIterator __last, const _Tp& __value_, _Compare __comp) +{ + typedef typename __comp_ref_type<_Compare>::type _Comp_ref; + return _VSTD::__equal_range<_Comp_ref>(__first, __last, __value_, __comp); +} + +template +_LIBCPP_NODISCARD_EXT inline +_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +pair<_ForwardIterator, _ForwardIterator> +equal_range(_ForwardIterator __first, _ForwardIterator __last, const _Tp& __value_) +{ + return _VSTD::equal_range(__first, __last, __value_, + __less::value_type, _Tp>()); +} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_EQUAL_RANGE_H diff --git a/lib/libcxx/include/__algorithm/fill.h b/lib/libcxx/include/__algorithm/fill.h new file mode 100644 index 000000000..4fefe8653 --- /dev/null +++ b/lib/libcxx/include/__algorithm/fill.h @@ -0,0 +1,55 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_FILL_H +#define _LIBCPP___ALGORITHM_FILL_H + +#include <__config> +#include <__algorithm/fill_n.h> +#include <__iterator/iterator_traits.h> +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +void +__fill(_ForwardIterator __first, _ForwardIterator __last, const _Tp& __value_, forward_iterator_tag) +{ + for (; __first != __last; ++__first) + *__first = __value_; +} + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +void +__fill(_RandomAccessIterator __first, _RandomAccessIterator __last, const _Tp& __value_, random_access_iterator_tag) +{ + _VSTD::fill_n(__first, __last - __first, __value_); +} + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +void +fill(_ForwardIterator __first, _ForwardIterator __last, const _Tp& __value_) +{ + _VSTD::__fill(__first, __last, __value_, typename iterator_traits<_ForwardIterator>::iterator_category()); +} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_FILL_H diff --git a/lib/libcxx/include/__algorithm/fill_n.h b/lib/libcxx/include/__algorithm/fill_n.h new file mode 100644 index 000000000..34a245edb --- /dev/null +++ b/lib/libcxx/include/__algorithm/fill_n.h @@ -0,0 +1,47 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_FILL_N_H +#define _LIBCPP___ALGORITHM_FILL_N_H + +#include <__config> +#include <__iterator/iterator_traits.h> +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +_OutputIterator +__fill_n(_OutputIterator __first, _Size __n, const _Tp& __value_) +{ + for (; __n > 0; ++__first, (void) --__n) + *__first = __value_; + return __first; +} + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +_OutputIterator +fill_n(_OutputIterator __first, _Size __n, const _Tp& __value_) +{ + return _VSTD::__fill_n(__first, _VSTD::__convert_to_integral(__n), __value_); +} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_FILL_N_H diff --git a/lib/libcxx/include/__algorithm/find.h b/lib/libcxx/include/__algorithm/find.h new file mode 100644 index 000000000..bc593dc54 --- /dev/null +++ b/lib/libcxx/include/__algorithm/find.h @@ -0,0 +1,37 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_FIND_H +#define _LIBCPP___ALGORITHM_FIND_H + +#include <__config> + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +_LIBCPP_NODISCARD_EXT inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 _InputIterator +find(_InputIterator __first, _InputIterator __last, const _Tp& __value_) { + for (; __first != __last; ++__first) + if (*__first == __value_) + break; + return __first; +} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_FIND_H diff --git a/lib/libcxx/include/__algorithm/find_end.h b/lib/libcxx/include/__algorithm/find_end.h new file mode 100644 index 000000000..f4277f003 --- /dev/null +++ b/lib/libcxx/include/__algorithm/find_end.h @@ -0,0 +1,154 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_FIND_END_OF_H +#define _LIBCPP___ALGORITHM_FIND_END_OF_H + +#include <__config> +#include <__algorithm/comp.h> +#include <__iterator/iterator_traits.h> +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +_LIBCPP_CONSTEXPR_AFTER_CXX17 _ForwardIterator1 __find_end(_ForwardIterator1 __first1, _ForwardIterator1 __last1, + _ForwardIterator2 __first2, _ForwardIterator2 __last2, + _BinaryPredicate __pred, forward_iterator_tag, + forward_iterator_tag) { + // modeled after search algorithm + _ForwardIterator1 __r = __last1; // __last1 is the "default" answer + if (__first2 == __last2) + return __r; + while (true) { + while (true) { + if (__first1 == __last1) // if source exhausted return last correct answer + return __r; // (or __last1 if never found) + if (__pred(*__first1, *__first2)) + break; + ++__first1; + } + // *__first1 matches *__first2, now match elements after here + _ForwardIterator1 __m1 = __first1; + _ForwardIterator2 __m2 = __first2; + while (true) { + if (++__m2 == __last2) { // Pattern exhaused, record answer and search for another one + __r = __first1; + ++__first1; + break; + } + if (++__m1 == __last1) // Source exhausted, return last answer + return __r; + if (!__pred(*__m1, *__m2)) // mismatch, restart with a new __first + { + ++__first1; + break; + } // else there is a match, check next elements + } + } +} + +template +_LIBCPP_CONSTEXPR_AFTER_CXX17 _BidirectionalIterator1 __find_end( + _BidirectionalIterator1 __first1, _BidirectionalIterator1 __last1, _BidirectionalIterator2 __first2, + _BidirectionalIterator2 __last2, _BinaryPredicate __pred, bidirectional_iterator_tag, bidirectional_iterator_tag) { + // modeled after search algorithm (in reverse) + if (__first2 == __last2) + return __last1; // Everything matches an empty sequence + _BidirectionalIterator1 __l1 = __last1; + _BidirectionalIterator2 __l2 = __last2; + --__l2; + while (true) { + // Find last element in sequence 1 that matchs *(__last2-1), with a mininum of loop checks + while (true) { + if (__first1 == __l1) // return __last1 if no element matches *__first2 + return __last1; + if (__pred(*--__l1, *__l2)) + break; + } + // *__l1 matches *__l2, now match elements before here + _BidirectionalIterator1 __m1 = __l1; + _BidirectionalIterator2 __m2 = __l2; + while (true) { + if (__m2 == __first2) // If pattern exhausted, __m1 is the answer (works for 1 element pattern) + return __m1; + if (__m1 == __first1) // Otherwise if source exhaused, pattern not found + return __last1; + if (!__pred(*--__m1, *--__m2)) // if there is a mismatch, restart with a new __l1 + { + break; + } // else there is a match, check next elements + } + } +} + +template +_LIBCPP_CONSTEXPR_AFTER_CXX11 _RandomAccessIterator1 __find_end( + _RandomAccessIterator1 __first1, _RandomAccessIterator1 __last1, _RandomAccessIterator2 __first2, + _RandomAccessIterator2 __last2, _BinaryPredicate __pred, random_access_iterator_tag, random_access_iterator_tag) { + // Take advantage of knowing source and pattern lengths. Stop short when source is smaller than pattern + typename iterator_traits<_RandomAccessIterator2>::difference_type __len2 = __last2 - __first2; + if (__len2 == 0) + return __last1; + typename iterator_traits<_RandomAccessIterator1>::difference_type __len1 = __last1 - __first1; + if (__len1 < __len2) + return __last1; + const _RandomAccessIterator1 __s = __first1 + (__len2 - 1); // End of pattern match can't go before here + _RandomAccessIterator1 __l1 = __last1; + _RandomAccessIterator2 __l2 = __last2; + --__l2; + while (true) { + while (true) { + if (__s == __l1) + return __last1; + if (__pred(*--__l1, *__l2)) + break; + } + _RandomAccessIterator1 __m1 = __l1; + _RandomAccessIterator2 __m2 = __l2; + while (true) { + if (__m2 == __first2) + return __m1; + // no need to check range on __m1 because __s guarantees we have enough source + if (!__pred(*--__m1, *--__m2)) { + break; + } + } + } +} + +template +_LIBCPP_NODISCARD_EXT inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 _ForwardIterator1 +find_end(_ForwardIterator1 __first1, _ForwardIterator1 __last1, _ForwardIterator2 __first2, _ForwardIterator2 __last2, + _BinaryPredicate __pred) { + return _VSTD::__find_end::type>( + __first1, __last1, __first2, __last2, __pred, typename iterator_traits<_ForwardIterator1>::iterator_category(), + typename iterator_traits<_ForwardIterator2>::iterator_category()); +} + +template +_LIBCPP_NODISCARD_EXT inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 _ForwardIterator1 +find_end(_ForwardIterator1 __first1, _ForwardIterator1 __last1, _ForwardIterator2 __first2, _ForwardIterator2 __last2) { + typedef typename iterator_traits<_ForwardIterator1>::value_type __v1; + typedef typename iterator_traits<_ForwardIterator2>::value_type __v2; + return _VSTD::find_end(__first1, __last1, __first2, __last2, __equal_to<__v1, __v2>()); +} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_FIND_END_OF_H diff --git a/lib/libcxx/include/__algorithm/find_first_of.h b/lib/libcxx/include/__algorithm/find_first_of.h new file mode 100644 index 000000000..d956c8dda --- /dev/null +++ b/lib/libcxx/include/__algorithm/find_first_of.h @@ -0,0 +1,57 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_FIND_FIRST_OF_H +#define _LIBCPP___ALGORITHM_FIND_FIRST_OF_H + +#include <__config> +#include <__algorithm/comp.h> +#include <__iterator/iterator_traits.h> + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +_LIBCPP_CONSTEXPR_AFTER_CXX11 _ForwardIterator1 __find_first_of_ce(_ForwardIterator1 __first1, + _ForwardIterator1 __last1, + _ForwardIterator2 __first2, + _ForwardIterator2 __last2, _BinaryPredicate __pred) { + for (; __first1 != __last1; ++__first1) + for (_ForwardIterator2 __j = __first2; __j != __last2; ++__j) + if (__pred(*__first1, *__j)) + return __first1; + return __last1; +} + +template +_LIBCPP_NODISCARD_EXT inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 _ForwardIterator1 +find_first_of(_ForwardIterator1 __first1, _ForwardIterator1 __last1, _ForwardIterator2 __first2, + _ForwardIterator2 __last2, _BinaryPredicate __pred) { + return _VSTD::__find_first_of_ce(__first1, __last1, __first2, __last2, __pred); +} + +template +_LIBCPP_NODISCARD_EXT inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 _ForwardIterator1 find_first_of( + _ForwardIterator1 __first1, _ForwardIterator1 __last1, _ForwardIterator2 __first2, _ForwardIterator2 __last2) { + typedef typename iterator_traits<_ForwardIterator1>::value_type __v1; + typedef typename iterator_traits<_ForwardIterator2>::value_type __v2; + return _VSTD::__find_first_of_ce(__first1, __last1, __first2, __last2, __equal_to<__v1, __v2>()); +} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_FIND_FIRST_OF_H diff --git a/lib/libcxx/include/__algorithm/find_if.h b/lib/libcxx/include/__algorithm/find_if.h new file mode 100644 index 000000000..456cc5b0a --- /dev/null +++ b/lib/libcxx/include/__algorithm/find_if.h @@ -0,0 +1,37 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_FIND_IF_H +#define _LIBCPP___ALGORITHM_FIND_IF_H + +#include <__config> + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +_LIBCPP_NODISCARD_EXT inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 _InputIterator +find_if(_InputIterator __first, _InputIterator __last, _Predicate __pred) { + for (; __first != __last; ++__first) + if (__pred(*__first)) + break; + return __first; +} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_FIND_IF_H diff --git a/lib/libcxx/include/__algorithm/find_if_not.h b/lib/libcxx/include/__algorithm/find_if_not.h new file mode 100644 index 000000000..d7d25160c --- /dev/null +++ b/lib/libcxx/include/__algorithm/find_if_not.h @@ -0,0 +1,37 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_FIND_IF_NOT_H +#define _LIBCPP___ALGORITHM_FIND_IF_NOT_H + +#include <__config> + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +_LIBCPP_NODISCARD_EXT inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 _InputIterator +find_if_not(_InputIterator __first, _InputIterator __last, _Predicate __pred) { + for (; __first != __last; ++__first) + if (!__pred(*__first)) + break; + return __first; +} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_FIND_IF_NOT_H diff --git a/lib/libcxx/include/__algorithm/for_each.h b/lib/libcxx/include/__algorithm/for_each.h new file mode 100644 index 000000000..e71a36a1d --- /dev/null +++ b/lib/libcxx/include/__algorithm/for_each.h @@ -0,0 +1,37 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_FOR_EACH_H +#define _LIBCPP___ALGORITHM_FOR_EACH_H + +#include <__config> + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 _Function for_each(_InputIterator __first, + _InputIterator __last, + _Function __f) { + for (; __first != __last; ++__first) + __f(*__first); + return __f; +} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_FOR_EACH_H diff --git a/lib/libcxx/include/__algorithm/for_each_n.h b/lib/libcxx/include/__algorithm/for_each_n.h new file mode 100644 index 000000000..77f6c8621 --- /dev/null +++ b/lib/libcxx/include/__algorithm/for_each_n.h @@ -0,0 +1,47 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_FOR_EACH_N_H +#define _LIBCPP___ALGORITHM_FOR_EACH_N_H + +#include <__config> +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +#if _LIBCPP_STD_VER > 14 + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 _InputIterator for_each_n(_InputIterator __first, + _Size __orig_n, + _Function __f) { + typedef decltype(_VSTD::__convert_to_integral(__orig_n)) _IntegralSize; + _IntegralSize __n = __orig_n; + while (__n > 0) { + __f(*__first); + ++__first; + --__n; + } + return __first; +} + +#endif + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_FOR_EACH_N_H diff --git a/lib/libcxx/include/__algorithm/generate.h b/lib/libcxx/include/__algorithm/generate.h new file mode 100644 index 000000000..d3e1133e2 --- /dev/null +++ b/lib/libcxx/include/__algorithm/generate.h @@ -0,0 +1,36 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_GENERATE_H +#define _LIBCPP___ALGORITHM_GENERATE_H + +#include <__config> + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +void +generate(_ForwardIterator __first, _ForwardIterator __last, _Generator __gen) +{ + for (; __first != __last; ++__first) + *__first = __gen(); +} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_GENERATE_H diff --git a/lib/libcxx/include/__algorithm/generate_n.h b/lib/libcxx/include/__algorithm/generate_n.h new file mode 100644 index 000000000..c31259858 --- /dev/null +++ b/lib/libcxx/include/__algorithm/generate_n.h @@ -0,0 +1,40 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_GENERATE_N_H +#define _LIBCPP___ALGORITHM_GENERATE_N_H + +#include <__config> +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +_OutputIterator +generate_n(_OutputIterator __first, _Size __orig_n, _Generator __gen) +{ + typedef decltype(_VSTD::__convert_to_integral(__orig_n)) _IntegralSize; + _IntegralSize __n = __orig_n; + for (; __n > 0; ++__first, (void) --__n) + *__first = __gen(); + return __first; +} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_GENERATE_N_H diff --git a/lib/libcxx/include/__algorithm/half_positive.h b/lib/libcxx/include/__algorithm/half_positive.h new file mode 100644 index 000000000..b03efc495 --- /dev/null +++ b/lib/libcxx/include/__algorithm/half_positive.h @@ -0,0 +1,54 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_HALF_POSITIVE_H +#define _LIBCPP___ALGORITHM_HALF_POSITIVE_H + +#include <__config> +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +// Perform division by two quickly for positive integers (llvm.org/PR39129) + +template +_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR +typename enable_if +< + is_integral<_Integral>::value, + _Integral +>::type +__half_positive(_Integral __value) +{ + return static_cast<_Integral>(static_cast::type>(__value) / 2); +} + +template +_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR +typename enable_if +< + !is_integral<_Tp>::value, + _Tp +>::type +__half_positive(_Tp __value) +{ + return __value / 2; +} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_HALF_POSITIVE_H diff --git a/lib/libcxx/include/__algorithm/includes.h b/lib/libcxx/include/__algorithm/includes.h new file mode 100644 index 000000000..ff298a511 --- /dev/null +++ b/lib/libcxx/include/__algorithm/includes.h @@ -0,0 +1,67 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_INCLUDES_H +#define _LIBCPP___ALGORITHM_INCLUDES_H + +#include <__config> +#include <__algorithm/comp.h> +#include <__algorithm/comp_ref_type.h> +#include <__iterator/iterator_traits.h> + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +_LIBCPP_CONSTEXPR_AFTER_CXX17 bool +__includes(_InputIterator1 __first1, _InputIterator1 __last1, _InputIterator2 __first2, _InputIterator2 __last2, + _Compare __comp) +{ + for (; __first2 != __last2; ++__first1) + { + if (__first1 == __last1 || __comp(*__first2, *__first1)) + return false; + if (!__comp(*__first1, *__first2)) + ++__first2; + } + return true; +} + +template +_LIBCPP_NODISCARD_EXT inline +_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +bool +includes(_InputIterator1 __first1, _InputIterator1 __last1, _InputIterator2 __first2, _InputIterator2 __last2, + _Compare __comp) +{ + typedef typename __comp_ref_type<_Compare>::type _Comp_ref; + return _VSTD::__includes<_Comp_ref>(__first1, __last1, __first2, __last2, __comp); +} + +template +_LIBCPP_NODISCARD_EXT inline +_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +bool +includes(_InputIterator1 __first1, _InputIterator1 __last1, _InputIterator2 __first2, _InputIterator2 __last2) +{ + return _VSTD::includes(__first1, __last1, __first2, __last2, + __less::value_type, + typename iterator_traits<_InputIterator2>::value_type>()); +} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_INCLUDES_H diff --git a/lib/libcxx/include/__algorithm/inplace_merge.h b/lib/libcxx/include/__algorithm/inplace_merge.h new file mode 100644 index 000000000..c74633a74 --- /dev/null +++ b/lib/libcxx/include/__algorithm/inplace_merge.h @@ -0,0 +1,231 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_INPLACE_MERGE_H +#define _LIBCPP___ALGORITHM_INPLACE_MERGE_H + +#include <__config> +#include <__algorithm/comp_ref_type.h> +#include <__algorithm/comp.h> +#include <__algorithm/lower_bound.h> +#include <__algorithm/min.h> +#include <__algorithm/move.h> +#include <__algorithm/rotate.h> +#include <__algorithm/upper_bound.h> +#include <__iterator/iterator_traits.h> +#include <__utility/swap.h> +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +class __invert // invert the sense of a comparison +{ +private: + _Predicate __p_; +public: + _LIBCPP_INLINE_VISIBILITY __invert() {} + + _LIBCPP_INLINE_VISIBILITY + explicit __invert(_Predicate __p) : __p_(__p) {} + + template + _LIBCPP_INLINE_VISIBILITY + bool operator()(const _T1& __x) {return !__p_(__x);} + + template + _LIBCPP_INLINE_VISIBILITY + bool operator()(const _T1& __x, const _T2& __y) {return __p_(__y, __x);} +}; + +template +void __half_inplace_merge(_InputIterator1 __first1, _InputIterator1 __last1, + _InputIterator2 __first2, _InputIterator2 __last2, + _OutputIterator __result, _Compare __comp) +{ + for (; __first1 != __last1; ++__result) + { + if (__first2 == __last2) + { + _VSTD::move(__first1, __last1, __result); + return; + } + + if (__comp(*__first2, *__first1)) + { + *__result = _VSTD::move(*__first2); + ++__first2; + } + else + { + *__result = _VSTD::move(*__first1); + ++__first1; + } + } + // __first2 through __last2 are already in the right spot. +} + +template +void +__buffered_inplace_merge(_BidirectionalIterator __first, _BidirectionalIterator __middle, _BidirectionalIterator __last, + _Compare __comp, typename iterator_traits<_BidirectionalIterator>::difference_type __len1, + typename iterator_traits<_BidirectionalIterator>::difference_type __len2, + typename iterator_traits<_BidirectionalIterator>::value_type* __buff) +{ + typedef typename iterator_traits<_BidirectionalIterator>::value_type value_type; + __destruct_n __d(0); + unique_ptr __h2(__buff, __d); + if (__len1 <= __len2) + { + value_type* __p = __buff; + for (_BidirectionalIterator __i = __first; __i != __middle; __d.template __incr(), (void) ++__i, (void) ++__p) + ::new ((void*)__p) value_type(_VSTD::move(*__i)); + _VSTD::__half_inplace_merge<_Compare>(__buff, __p, __middle, __last, __first, __comp); + } + else + { + value_type* __p = __buff; + for (_BidirectionalIterator __i = __middle; __i != __last; __d.template __incr(), (void) ++__i, (void) ++__p) + ::new ((void*)__p) value_type(_VSTD::move(*__i)); + typedef reverse_iterator<_BidirectionalIterator> _RBi; + typedef reverse_iterator _Rv; + typedef __invert<_Compare> _Inverted; + _VSTD::__half_inplace_merge<_Inverted>(_Rv(__p), _Rv(__buff), + _RBi(__middle), _RBi(__first), + _RBi(__last), _Inverted(__comp)); + } +} + +template +void +__inplace_merge(_BidirectionalIterator __first, _BidirectionalIterator __middle, _BidirectionalIterator __last, + _Compare __comp, typename iterator_traits<_BidirectionalIterator>::difference_type __len1, + typename iterator_traits<_BidirectionalIterator>::difference_type __len2, + typename iterator_traits<_BidirectionalIterator>::value_type* __buff, ptrdiff_t __buff_size) +{ + typedef typename iterator_traits<_BidirectionalIterator>::difference_type difference_type; + while (true) + { + // if __middle == __last, we're done + if (__len2 == 0) + return; + if (__len1 <= __buff_size || __len2 <= __buff_size) + return _VSTD::__buffered_inplace_merge<_Compare> + (__first, __middle, __last, __comp, __len1, __len2, __buff); + // shrink [__first, __middle) as much as possible (with no moves), returning if it shrinks to 0 + for (; true; ++__first, (void) --__len1) + { + if (__len1 == 0) + return; + if (__comp(*__middle, *__first)) + break; + } + // __first < __middle < __last + // *__first > *__middle + // partition [__first, __m1) [__m1, __middle) [__middle, __m2) [__m2, __last) such that + // all elements in: + // [__first, __m1) <= [__middle, __m2) + // [__middle, __m2) < [__m1, __middle) + // [__m1, __middle) <= [__m2, __last) + // and __m1 or __m2 is in the middle of its range + _BidirectionalIterator __m1; // "median" of [__first, __middle) + _BidirectionalIterator __m2; // "median" of [__middle, __last) + difference_type __len11; // distance(__first, __m1) + difference_type __len21; // distance(__middle, __m2) + // binary search smaller range + if (__len1 < __len2) + { // __len >= 1, __len2 >= 2 + __len21 = __len2 / 2; + __m2 = __middle; + _VSTD::advance(__m2, __len21); + __m1 = _VSTD::__upper_bound<_Compare>(__first, __middle, *__m2, __comp); + __len11 = _VSTD::distance(__first, __m1); + } + else + { + if (__len1 == 1) + { // __len1 >= __len2 && __len2 > 0, therefore __len2 == 1 + // It is known *__first > *__middle + swap(*__first, *__middle); + return; + } + // __len1 >= 2, __len2 >= 1 + __len11 = __len1 / 2; + __m1 = __first; + _VSTD::advance(__m1, __len11); + __m2 = _VSTD::__lower_bound<_Compare>(__middle, __last, *__m1, __comp); + __len21 = _VSTD::distance(__middle, __m2); + } + difference_type __len12 = __len1 - __len11; // distance(__m1, __middle) + difference_type __len22 = __len2 - __len21; // distance(__m2, __last) + // [__first, __m1) [__m1, __middle) [__middle, __m2) [__m2, __last) + // swap middle two partitions + __middle = _VSTD::rotate(__m1, __middle, __m2); + // __len12 and __len21 now have swapped meanings + // merge smaller range with recursive call and larger with tail recursion elimination + if (__len11 + __len21 < __len12 + __len22) + { + _VSTD::__inplace_merge<_Compare>(__first, __m1, __middle, __comp, __len11, __len21, __buff, __buff_size); +// _VSTD::__inplace_merge<_Compare>(__middle, __m2, __last, __comp, __len12, __len22, __buff, __buff_size); + __first = __middle; + __middle = __m2; + __len1 = __len12; + __len2 = __len22; + } + else + { + _VSTD::__inplace_merge<_Compare>(__middle, __m2, __last, __comp, __len12, __len22, __buff, __buff_size); +// _VSTD::__inplace_merge<_Compare>(__first, __m1, __middle, __comp, __len11, __len21, __buff, __buff_size); + __last = __middle; + __middle = __m1; + __len1 = __len11; + __len2 = __len21; + } + } +} + +template +inline _LIBCPP_INLINE_VISIBILITY +void +inplace_merge(_BidirectionalIterator __first, _BidirectionalIterator __middle, _BidirectionalIterator __last, + _Compare __comp) +{ + typedef typename iterator_traits<_BidirectionalIterator>::value_type value_type; + typedef typename iterator_traits<_BidirectionalIterator>::difference_type difference_type; + difference_type __len1 = _VSTD::distance(__first, __middle); + difference_type __len2 = _VSTD::distance(__middle, __last); + difference_type __buf_size = _VSTD::min(__len1, __len2); + pair __buf = _VSTD::get_temporary_buffer(__buf_size); + unique_ptr __h(__buf.first); + typedef typename __comp_ref_type<_Compare>::type _Comp_ref; + return _VSTD::__inplace_merge<_Comp_ref>(__first, __middle, __last, __comp, __len1, __len2, + __buf.first, __buf.second); +} + +template +inline _LIBCPP_INLINE_VISIBILITY +void +inplace_merge(_BidirectionalIterator __first, _BidirectionalIterator __middle, _BidirectionalIterator __last) +{ + _VSTD::inplace_merge(__first, __middle, __last, + __less::value_type>()); +} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_INPLACE_MERGE_H diff --git a/lib/libcxx/include/__algorithm/is_heap.h b/lib/libcxx/include/__algorithm/is_heap.h new file mode 100644 index 000000000..bc3682d47 --- /dev/null +++ b/lib/libcxx/include/__algorithm/is_heap.h @@ -0,0 +1,48 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_IS_HEAP_H +#define _LIBCPP___ALGORITHM_IS_HEAP_H + +#include <__config> +#include <__algorithm/comp.h> +#include <__algorithm/is_heap_until.h> +#include <__iterator/iterator_traits.h> + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +_LIBCPP_NODISCARD_EXT inline +_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +bool +is_heap(_RandomAccessIterator __first, _RandomAccessIterator __last, _Compare __comp) +{ + return _VSTD::is_heap_until(__first, __last, __comp) == __last; +} + +template +_LIBCPP_NODISCARD_EXT inline +_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +bool +is_heap(_RandomAccessIterator __first, _RandomAccessIterator __last) +{ + return _VSTD::is_heap(__first, __last, __less::value_type>()); +} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_IS_HEAP_H diff --git a/lib/libcxx/include/__algorithm/is_heap_until.h b/lib/libcxx/include/__algorithm/is_heap_until.h new file mode 100644 index 000000000..8c52edb7d --- /dev/null +++ b/lib/libcxx/include/__algorithm/is_heap_until.h @@ -0,0 +1,65 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_IS_HEAP_UNTIL_H +#define _LIBCPP___ALGORITHM_IS_HEAP_UNTIL_H + +#include <__config> +#include <__algorithm/comp.h> +#include <__iterator/iterator_traits.h> + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +_LIBCPP_NODISCARD_EXT _LIBCPP_CONSTEXPR_AFTER_CXX17 _RandomAccessIterator +is_heap_until(_RandomAccessIterator __first, _RandomAccessIterator __last, _Compare __comp) +{ + typedef typename iterator_traits<_RandomAccessIterator>::difference_type difference_type; + difference_type __len = __last - __first; + difference_type __p = 0; + difference_type __c = 1; + _RandomAccessIterator __pp = __first; + while (__c < __len) + { + _RandomAccessIterator __cp = __first + __c; + if (__comp(*__pp, *__cp)) + return __cp; + ++__c; + ++__cp; + if (__c == __len) + return __last; + if (__comp(*__pp, *__cp)) + return __cp; + ++__p; + ++__pp; + __c = 2 * __p + 1; + } + return __last; +} + +template +_LIBCPP_NODISCARD_EXT inline +_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +_RandomAccessIterator +is_heap_until(_RandomAccessIterator __first, _RandomAccessIterator __last) +{ + return _VSTD::is_heap_until(__first, __last, __less::value_type>()); +} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_IS_HEAP_UNTIL_H diff --git a/lib/libcxx/include/__algorithm/is_partitioned.h b/lib/libcxx/include/__algorithm/is_partitioned.h new file mode 100644 index 000000000..43de6650c --- /dev/null +++ b/lib/libcxx/include/__algorithm/is_partitioned.h @@ -0,0 +1,43 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_IS_PARTITIONED_H +#define _LIBCPP___ALGORITHM_IS_PARTITIONED_H + +#include <__config> + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +_LIBCPP_NODISCARD_EXT _LIBCPP_CONSTEXPR_AFTER_CXX17 bool +is_partitioned(_InputIterator __first, _InputIterator __last, _Predicate __pred) +{ + for (; __first != __last; ++__first) + if (!__pred(*__first)) + break; + if ( __first == __last ) + return true; + ++__first; + for (; __first != __last; ++__first) + if (__pred(*__first)) + return false; + return true; +} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_IS_PARTITIONED_H diff --git a/lib/libcxx/include/__algorithm/is_permutation.h b/lib/libcxx/include/__algorithm/is_permutation.h new file mode 100644 index 000000000..0545eb763 --- /dev/null +++ b/lib/libcxx/include/__algorithm/is_permutation.h @@ -0,0 +1,168 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_IS_PERMUTATION_H +#define _LIBCPP___ALGORITHM_IS_PERMUTATION_H + +#include <__algorithm/comp.h> +#include <__config> +#include <__iterator/iterator_traits.h> +#include <__iterator/next.h> +#include // FIXME: replace with <__iterator/distance.h> when it lands + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +_LIBCPP_NODISCARD_EXT _LIBCPP_CONSTEXPR_AFTER_CXX17 bool +is_permutation(_ForwardIterator1 __first1, _ForwardIterator1 __last1, _ForwardIterator2 __first2, + _BinaryPredicate __pred) { + // shorten sequences as much as possible by lopping of any equal prefix + for (; __first1 != __last1; ++__first1, (void)++__first2) + if (!__pred(*__first1, *__first2)) + break; + if (__first1 == __last1) + return true; + + // __first1 != __last1 && *__first1 != *__first2 + typedef typename iterator_traits<_ForwardIterator1>::difference_type _D1; + _D1 __l1 = _VSTD::distance(__first1, __last1); + if (__l1 == _D1(1)) + return false; + _ForwardIterator2 __last2 = _VSTD::next(__first2, __l1); + // For each element in [f1, l1) see if there are the same number of + // equal elements in [f2, l2) + for (_ForwardIterator1 __i = __first1; __i != __last1; ++__i) { + // Have we already counted the number of *__i in [f1, l1)? + _ForwardIterator1 __match = __first1; + for (; __match != __i; ++__match) + if (__pred(*__match, *__i)) + break; + if (__match == __i) { + // Count number of *__i in [f2, l2) + _D1 __c2 = 0; + for (_ForwardIterator2 __j = __first2; __j != __last2; ++__j) + if (__pred(*__i, *__j)) + ++__c2; + if (__c2 == 0) + return false; + // Count number of *__i in [__i, l1) (we can start with 1) + _D1 __c1 = 1; + for (_ForwardIterator1 __j = _VSTD::next(__i); __j != __last1; ++__j) + if (__pred(*__i, *__j)) + ++__c1; + if (__c1 != __c2) + return false; + } + } + return true; +} + +template +_LIBCPP_NODISCARD_EXT inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 bool +is_permutation(_ForwardIterator1 __first1, _ForwardIterator1 __last1, _ForwardIterator2 __first2) { + typedef typename iterator_traits<_ForwardIterator1>::value_type __v1; + typedef typename iterator_traits<_ForwardIterator2>::value_type __v2; + return _VSTD::is_permutation(__first1, __last1, __first2, __equal_to<__v1, __v2>()); +} + +#if _LIBCPP_STD_VER > 11 +template +_LIBCPP_CONSTEXPR_AFTER_CXX17 bool +__is_permutation(_ForwardIterator1 __first1, _ForwardIterator1 __last1, _ForwardIterator2 __first2, + _ForwardIterator2 __last2, _BinaryPredicate __pred, forward_iterator_tag, forward_iterator_tag) { + // shorten sequences as much as possible by lopping of any equal prefix + for (; __first1 != __last1 && __first2 != __last2; ++__first1, (void)++__first2) + if (!__pred(*__first1, *__first2)) + break; + if (__first1 == __last1) + return __first2 == __last2; + else if (__first2 == __last2) + return false; + + typedef typename iterator_traits<_ForwardIterator1>::difference_type _D1; + _D1 __l1 = _VSTD::distance(__first1, __last1); + + typedef typename iterator_traits<_ForwardIterator2>::difference_type _D2; + _D2 __l2 = _VSTD::distance(__first2, __last2); + if (__l1 != __l2) + return false; + + // For each element in [f1, l1) see if there are the same number of + // equal elements in [f2, l2) + for (_ForwardIterator1 __i = __first1; __i != __last1; ++__i) { + // Have we already counted the number of *__i in [f1, l1)? + _ForwardIterator1 __match = __first1; + for (; __match != __i; ++__match) + if (__pred(*__match, *__i)) + break; + if (__match == __i) { + // Count number of *__i in [f2, l2) + _D1 __c2 = 0; + for (_ForwardIterator2 __j = __first2; __j != __last2; ++__j) + if (__pred(*__i, *__j)) + ++__c2; + if (__c2 == 0) + return false; + // Count number of *__i in [__i, l1) (we can start with 1) + _D1 __c1 = 1; + for (_ForwardIterator1 __j = _VSTD::next(__i); __j != __last1; ++__j) + if (__pred(*__i, *__j)) + ++__c1; + if (__c1 != __c2) + return false; + } + } + return true; +} + +template +_LIBCPP_CONSTEXPR_AFTER_CXX17 bool __is_permutation(_RandomAccessIterator1 __first1, _RandomAccessIterator2 __last1, + _RandomAccessIterator1 __first2, _RandomAccessIterator2 __last2, + _BinaryPredicate __pred, random_access_iterator_tag, + random_access_iterator_tag) { + if (_VSTD::distance(__first1, __last1) != _VSTD::distance(__first2, __last2)) + return false; + return _VSTD::is_permutation<_RandomAccessIterator1, _RandomAccessIterator2, + typename add_lvalue_reference<_BinaryPredicate>::type>(__first1, __last1, __first2, + __pred); +} + +template +_LIBCPP_NODISCARD_EXT inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 bool +is_permutation(_ForwardIterator1 __first1, _ForwardIterator1 __last1, _ForwardIterator2 __first2, + _ForwardIterator2 __last2, _BinaryPredicate __pred) { + return _VSTD::__is_permutation::type>( + __first1, __last1, __first2, __last2, __pred, typename iterator_traits<_ForwardIterator1>::iterator_category(), + typename iterator_traits<_ForwardIterator2>::iterator_category()); +} + +template +_LIBCPP_NODISCARD_EXT inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 bool +is_permutation(_ForwardIterator1 __first1, _ForwardIterator1 __last1, _ForwardIterator2 __first2, + _ForwardIterator2 __last2) { + typedef typename iterator_traits<_ForwardIterator1>::value_type __v1; + typedef typename iterator_traits<_ForwardIterator2>::value_type __v2; + return _VSTD::__is_permutation(__first1, __last1, __first2, __last2, __equal_to<__v1, __v2>(), + typename iterator_traits<_ForwardIterator1>::iterator_category(), + typename iterator_traits<_ForwardIterator2>::iterator_category()); +} +#endif + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_IS_PERMUTATION_H diff --git a/lib/libcxx/include/__algorithm/is_sorted.h b/lib/libcxx/include/__algorithm/is_sorted.h new file mode 100644 index 000000000..30d8da049 --- /dev/null +++ b/lib/libcxx/include/__algorithm/is_sorted.h @@ -0,0 +1,48 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_IS_SORTED_H +#define _LIBCPP___ALGORITHM_IS_SORTED_H + +#include <__algorithm/comp.h> +#include <__algorithm/is_sorted_until.h> +#include <__config> +#include <__iterator/iterator_traits.h> + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +_LIBCPP_NODISCARD_EXT inline +_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +bool +is_sorted(_ForwardIterator __first, _ForwardIterator __last, _Compare __comp) +{ + return _VSTD::is_sorted_until(__first, __last, __comp) == __last; +} + +template +_LIBCPP_NODISCARD_EXT inline +_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +bool +is_sorted(_ForwardIterator __first, _ForwardIterator __last) +{ + return _VSTD::is_sorted(__first, __last, __less::value_type>()); +} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_IS_SORTED_H diff --git a/lib/libcxx/include/__algorithm/is_sorted_until.h b/lib/libcxx/include/__algorithm/is_sorted_until.h new file mode 100644 index 000000000..a914b5a56 --- /dev/null +++ b/lib/libcxx/include/__algorithm/is_sorted_until.h @@ -0,0 +1,55 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_IS_SORTED_UNTIL_H +#define _LIBCPP___ALGORITHM_IS_SORTED_UNTIL_H + +#include <__config> +#include <__algorithm/comp.h> +#include <__iterator/iterator_traits.h> + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +_LIBCPP_NODISCARD_EXT _LIBCPP_CONSTEXPR_AFTER_CXX17 _ForwardIterator +is_sorted_until(_ForwardIterator __first, _ForwardIterator __last, _Compare __comp) +{ + if (__first != __last) + { + _ForwardIterator __i = __first; + while (++__i != __last) + { + if (__comp(*__i, *__first)) + return __i; + __first = __i; + } + } + return __last; +} + +template +_LIBCPP_NODISCARD_EXT inline +_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +_ForwardIterator +is_sorted_until(_ForwardIterator __first, _ForwardIterator __last) +{ + return _VSTD::is_sorted_until(__first, __last, __less::value_type>()); +} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_IS_SORTED_UNTIL_H diff --git a/lib/libcxx/include/__algorithm/iter_swap.h b/lib/libcxx/include/__algorithm/iter_swap.h new file mode 100644 index 000000000..b63bce62c --- /dev/null +++ b/lib/libcxx/include/__algorithm/iter_swap.h @@ -0,0 +1,37 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_ITER_SWAP_H +#define _LIBCPP___ALGORITHM_ITER_SWAP_H + +#include <__config> +#include <__utility/declval.h> +#include <__utility/swap.h> + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 void iter_swap(_ForwardIterator1 __a, + _ForwardIterator2 __b) + // _NOEXCEPT_(_NOEXCEPT_(swap(*__a, *__b))) + _NOEXCEPT_(_NOEXCEPT_(swap(*declval<_ForwardIterator1>(), *declval<_ForwardIterator2>()))) { + swap(*__a, *__b); +} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_ITER_SWAP_H diff --git a/lib/libcxx/include/__algorithm/lexicographical_compare.h b/lib/libcxx/include/__algorithm/lexicographical_compare.h new file mode 100644 index 000000000..6e4a90bdc --- /dev/null +++ b/lib/libcxx/include/__algorithm/lexicographical_compare.h @@ -0,0 +1,68 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_LEXICOGRAPHICAL_COMPARE_H +#define _LIBCPP___ALGORITHM_LEXICOGRAPHICAL_COMPARE_H + +#include <__config> +#include <__algorithm/comp.h> +#include <__algorithm/comp_ref_type.h> +#include <__iterator/iterator_traits.h> + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +_LIBCPP_CONSTEXPR_AFTER_CXX17 bool +__lexicographical_compare(_InputIterator1 __first1, _InputIterator1 __last1, + _InputIterator2 __first2, _InputIterator2 __last2, _Compare __comp) +{ + for (; __first2 != __last2; ++__first1, (void) ++__first2) + { + if (__first1 == __last1 || __comp(*__first1, *__first2)) + return true; + if (__comp(*__first2, *__first1)) + return false; + } + return false; +} + +template +_LIBCPP_NODISCARD_EXT inline +_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +bool +lexicographical_compare(_InputIterator1 __first1, _InputIterator1 __last1, + _InputIterator2 __first2, _InputIterator2 __last2, _Compare __comp) +{ + typedef typename __comp_ref_type<_Compare>::type _Comp_ref; + return _VSTD::__lexicographical_compare<_Comp_ref>(__first1, __last1, __first2, __last2, __comp); +} + +template +_LIBCPP_NODISCARD_EXT inline +_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +bool +lexicographical_compare(_InputIterator1 __first1, _InputIterator1 __last1, + _InputIterator2 __first2, _InputIterator2 __last2) +{ + return _VSTD::lexicographical_compare(__first1, __last1, __first2, __last2, + __less::value_type, + typename iterator_traits<_InputIterator2>::value_type>()); +} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_LEXICOGRAPHICAL_COMPARE_H diff --git a/lib/libcxx/include/__algorithm/lower_bound.h b/lib/libcxx/include/__algorithm/lower_bound.h new file mode 100644 index 000000000..1448c8963 --- /dev/null +++ b/lib/libcxx/include/__algorithm/lower_bound.h @@ -0,0 +1,72 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_LOWER_BOUND_H +#define _LIBCPP___ALGORITHM_LOWER_BOUND_H + +#include <__config> +#include <__algorithm/comp.h> +#include <__algorithm/half_positive.h> +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +_LIBCPP_CONSTEXPR_AFTER_CXX17 _ForwardIterator +__lower_bound(_ForwardIterator __first, _ForwardIterator __last, const _Tp& __value_, _Compare __comp) +{ + typedef typename iterator_traits<_ForwardIterator>::difference_type difference_type; + difference_type __len = _VSTD::distance(__first, __last); + while (__len != 0) + { + difference_type __l2 = _VSTD::__half_positive(__len); + _ForwardIterator __m = __first; + _VSTD::advance(__m, __l2); + if (__comp(*__m, __value_)) + { + __first = ++__m; + __len -= __l2 + 1; + } + else + __len = __l2; + } + return __first; +} + +template +_LIBCPP_NODISCARD_EXT inline +_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +_ForwardIterator +lower_bound(_ForwardIterator __first, _ForwardIterator __last, const _Tp& __value_, _Compare __comp) +{ + typedef typename add_lvalue_reference<_Compare>::type _Comp_ref; + return _VSTD::__lower_bound<_Comp_ref>(__first, __last, __value_, __comp); +} + +template +_LIBCPP_NODISCARD_EXT inline +_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +_ForwardIterator +lower_bound(_ForwardIterator __first, _ForwardIterator __last, const _Tp& __value_) +{ + return _VSTD::lower_bound(__first, __last, __value_, + __less::value_type, _Tp>()); +} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_LOWER_BOUND_H diff --git a/lib/libcxx/include/__algorithm/make_heap.h b/lib/libcxx/include/__algorithm/make_heap.h new file mode 100644 index 000000000..eca4013cb --- /dev/null +++ b/lib/libcxx/include/__algorithm/make_heap.h @@ -0,0 +1,64 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_MAKE_HEAP_H +#define _LIBCPP___ALGORITHM_MAKE_HEAP_H + +#include <__config> +#include <__algorithm/comp.h> +#include <__algorithm/comp_ref_type.h> +#include <__algorithm/sift_down.h> +#include <__iterator/iterator_traits.h> + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +_LIBCPP_CONSTEXPR_AFTER_CXX11 void +__make_heap(_RandomAccessIterator __first, _RandomAccessIterator __last, _Compare __comp) +{ + typedef typename iterator_traits<_RandomAccessIterator>::difference_type difference_type; + difference_type __n = __last - __first; + if (__n > 1) + { + // start from the first parent, there is no need to consider children + for (difference_type __start = (__n - 2) / 2; __start >= 0; --__start) + { + _VSTD::__sift_down<_Compare>(__first, __last, __comp, __n, __first + __start); + } + } +} + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +void +make_heap(_RandomAccessIterator __first, _RandomAccessIterator __last, _Compare __comp) +{ + typedef typename __comp_ref_type<_Compare>::type _Comp_ref; + _VSTD::__make_heap<_Comp_ref>(__first, __last, __comp); +} + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +void +make_heap(_RandomAccessIterator __first, _RandomAccessIterator __last) +{ + _VSTD::make_heap(__first, __last, __less::value_type>()); +} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_MAKE_HEAP_H diff --git a/lib/libcxx/include/__algorithm/max.h b/lib/libcxx/include/__algorithm/max.h new file mode 100644 index 000000000..2fa97cad8 --- /dev/null +++ b/lib/libcxx/include/__algorithm/max.h @@ -0,0 +1,70 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_MAX_H +#define _LIBCPP___ALGORITHM_MAX_H + +#include <__config> +#include <__algorithm/comp.h> +#include <__algorithm/max_element.h> +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +_LIBCPP_NODISCARD_EXT inline +_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 +const _Tp& +max(const _Tp& __a, const _Tp& __b, _Compare __comp) +{ + return __comp(__a, __b) ? __b : __a; +} + +template +_LIBCPP_NODISCARD_EXT inline +_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 +const _Tp& +max(const _Tp& __a, const _Tp& __b) +{ + return _VSTD::max(__a, __b, __less<_Tp>()); +} + +#ifndef _LIBCPP_CXX03_LANG + +template +_LIBCPP_NODISCARD_EXT inline +_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 +_Tp +max(initializer_list<_Tp> __t, _Compare __comp) +{ + return *_VSTD::max_element(__t.begin(), __t.end(), __comp); +} + +template +_LIBCPP_NODISCARD_EXT inline +_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 +_Tp +max(initializer_list<_Tp> __t) +{ + return *_VSTD::max_element(__t.begin(), __t.end(), __less<_Tp>()); +} + +#endif // _LIBCPP_CXX03_LANG + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_MAX_H diff --git a/lib/libcxx/include/__algorithm/max_element.h b/lib/libcxx/include/__algorithm/max_element.h new file mode 100644 index 000000000..b93b67eac --- /dev/null +++ b/lib/libcxx/include/__algorithm/max_element.h @@ -0,0 +1,58 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_MAX_ELEMENT_H +#define _LIBCPP___ALGORITHM_MAX_ELEMENT_H + +#include <__config> +#include <__algorithm/comp.h> +#include <__iterator/iterator_traits.h> + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +_LIBCPP_NODISCARD_EXT inline +_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 +_ForwardIterator +max_element(_ForwardIterator __first, _ForwardIterator __last, _Compare __comp) +{ + static_assert(__is_cpp17_forward_iterator<_ForwardIterator>::value, + "std::max_element requires a ForwardIterator"); + if (__first != __last) + { + _ForwardIterator __i = __first; + while (++__i != __last) + if (__comp(*__first, *__i)) + __first = __i; + } + return __first; +} + + +template +_LIBCPP_NODISCARD_EXT inline +_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 +_ForwardIterator +max_element(_ForwardIterator __first, _ForwardIterator __last) +{ + return _VSTD::max_element(__first, __last, + __less::value_type>()); +} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_MAX_ELEMENT_H diff --git a/lib/libcxx/include/__algorithm/merge.h b/lib/libcxx/include/__algorithm/merge.h new file mode 100644 index 000000000..ea53ad68f --- /dev/null +++ b/lib/libcxx/include/__algorithm/merge.h @@ -0,0 +1,76 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_MERGE_H +#define _LIBCPP___ALGORITHM_MERGE_H + +#include <__config> +#include <__algorithm/comp.h> +#include <__algorithm/comp_ref_type.h> +#include <__algorithm/copy.h> +#include <__iterator/iterator_traits.h> + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +_LIBCPP_CONSTEXPR_AFTER_CXX17 +_OutputIterator +__merge(_InputIterator1 __first1, _InputIterator1 __last1, + _InputIterator2 __first2, _InputIterator2 __last2, _OutputIterator __result, _Compare __comp) +{ + for (; __first1 != __last1; ++__result) + { + if (__first2 == __last2) + return _VSTD::copy(__first1, __last1, __result); + if (__comp(*__first2, *__first1)) + { + *__result = *__first2; + ++__first2; + } + else + { + *__result = *__first1; + ++__first1; + } + } + return _VSTD::copy(__first2, __last2, __result); +} + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +_OutputIterator +merge(_InputIterator1 __first1, _InputIterator1 __last1, + _InputIterator2 __first2, _InputIterator2 __last2, _OutputIterator __result, _Compare __comp) +{ + typedef typename __comp_ref_type<_Compare>::type _Comp_ref; + return _VSTD::__merge<_Comp_ref>(__first1, __last1, __first2, __last2, __result, __comp); +} + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +_OutputIterator +merge(_InputIterator1 __first1, _InputIterator1 __last1, + _InputIterator2 __first2, _InputIterator2 __last2, _OutputIterator __result) +{ + typedef typename iterator_traits<_InputIterator1>::value_type __v1; + typedef typename iterator_traits<_InputIterator2>::value_type __v2; + return _VSTD::merge(__first1, __last1, __first2, __last2, __result, __less<__v1, __v2>()); +} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_MERGE_H diff --git a/lib/libcxx/include/__algorithm/min.h b/lib/libcxx/include/__algorithm/min.h new file mode 100644 index 000000000..9fea7f70a --- /dev/null +++ b/lib/libcxx/include/__algorithm/min.h @@ -0,0 +1,70 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_MIN_H +#define _LIBCPP___ALGORITHM_MIN_H + +#include <__config> +#include <__algorithm/comp.h> +#include <__algorithm/min_element.h> +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +_LIBCPP_NODISCARD_EXT inline +_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 +const _Tp& +min(const _Tp& __a, const _Tp& __b, _Compare __comp) +{ + return __comp(__b, __a) ? __b : __a; +} + +template +_LIBCPP_NODISCARD_EXT inline +_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 +const _Tp& +min(const _Tp& __a, const _Tp& __b) +{ + return _VSTD::min(__a, __b, __less<_Tp>()); +} + +#ifndef _LIBCPP_CXX03_LANG + +template +_LIBCPP_NODISCARD_EXT inline +_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 +_Tp +min(initializer_list<_Tp> __t, _Compare __comp) +{ + return *_VSTD::min_element(__t.begin(), __t.end(), __comp); +} + +template +_LIBCPP_NODISCARD_EXT inline +_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 +_Tp +min(initializer_list<_Tp> __t) +{ + return *_VSTD::min_element(__t.begin(), __t.end(), __less<_Tp>()); +} + +#endif // _LIBCPP_CXX03_LANG + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_MIN_H diff --git a/lib/libcxx/include/__algorithm/min_element.h b/lib/libcxx/include/__algorithm/min_element.h new file mode 100644 index 000000000..6bff140e5 --- /dev/null +++ b/lib/libcxx/include/__algorithm/min_element.h @@ -0,0 +1,57 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_MIN_ELEMENT_H +#define _LIBCPP___ALGORITHM_MIN_ELEMENT_H + +#include <__config> +#include <__algorithm/comp.h> +#include <__iterator/iterator_traits.h> + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +_LIBCPP_NODISCARD_EXT inline +_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 +_ForwardIterator +min_element(_ForwardIterator __first, _ForwardIterator __last, _Compare __comp) +{ + static_assert(__is_cpp17_forward_iterator<_ForwardIterator>::value, + "std::min_element requires a ForwardIterator"); + if (__first != __last) + { + _ForwardIterator __i = __first; + while (++__i != __last) + if (__comp(*__i, *__first)) + __first = __i; + } + return __first; +} + +template +_LIBCPP_NODISCARD_EXT inline +_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 +_ForwardIterator +min_element(_ForwardIterator __first, _ForwardIterator __last) +{ + return _VSTD::min_element(__first, __last, + __less::value_type>()); +} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_MIN_ELEMENT_H diff --git a/lib/libcxx/include/__algorithm/minmax.h b/lib/libcxx/include/__algorithm/minmax.h new file mode 100644 index 000000000..63753f2fe --- /dev/null +++ b/lib/libcxx/include/__algorithm/minmax.h @@ -0,0 +1,101 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_MINMAX_H +#define _LIBCPP___ALGORITHM_MINMAX_H + +#include <__config> +#include <__algorithm/comp.h> +#include +#include + + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +_LIBCPP_NODISCARD_EXT inline +_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 +pair +minmax(const _Tp& __a, const _Tp& __b, _Compare __comp) +{ + return __comp(__b, __a) ? pair(__b, __a) : + pair(__a, __b); +} + +template +_LIBCPP_NODISCARD_EXT inline +_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 +pair +minmax(const _Tp& __a, const _Tp& __b) +{ + return _VSTD::minmax(__a, __b, __less<_Tp>()); +} + +#ifndef _LIBCPP_CXX03_LANG + +template +_LIBCPP_NODISCARD_EXT inline +_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 +pair<_Tp, _Tp> +minmax(initializer_list<_Tp> __t, _Compare __comp) +{ + typedef typename initializer_list<_Tp>::const_iterator _Iter; + _Iter __first = __t.begin(); + _Iter __last = __t.end(); + pair<_Tp, _Tp> __result(*__first, *__first); + + ++__first; + if (__t.size() % 2 == 0) + { + if (__comp(*__first, __result.first)) + __result.first = *__first; + else + __result.second = *__first; + ++__first; + } + + while (__first != __last) + { + _Tp __prev = *__first++; + if (__comp(*__first, __prev)) { + if ( __comp(*__first, __result.first)) __result.first = *__first; + if (!__comp(__prev, __result.second)) __result.second = __prev; + } + else { + if ( __comp(__prev, __result.first)) __result.first = __prev; + if (!__comp(*__first, __result.second)) __result.second = *__first; + } + + __first++; + } + return __result; +} + +template +_LIBCPP_NODISCARD_EXT inline +_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 +pair<_Tp, _Tp> +minmax(initializer_list<_Tp> __t) +{ + return _VSTD::minmax(__t, __less<_Tp>()); +} + +#endif // _LIBCPP_CXX03_LANG + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_MINMAX_H diff --git a/lib/libcxx/include/__algorithm/minmax_element.h b/lib/libcxx/include/__algorithm/minmax_element.h new file mode 100644 index 000000000..1eba006de --- /dev/null +++ b/lib/libcxx/include/__algorithm/minmax_element.h @@ -0,0 +1,90 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_MINMAX_ELEMENT_H +#define _LIBCPP___ALGORITHM_MINMAX_ELEMENT_H + +#include <__config> +#include <__algorithm/comp.h> +#include <__iterator/iterator_traits.h> +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +_LIBCPP_NODISCARD_EXT _LIBCPP_CONSTEXPR_AFTER_CXX11 +pair<_ForwardIterator, _ForwardIterator> +minmax_element(_ForwardIterator __first, _ForwardIterator __last, _Compare __comp) +{ + static_assert(__is_cpp17_forward_iterator<_ForwardIterator>::value, + "std::minmax_element requires a ForwardIterator"); + pair<_ForwardIterator, _ForwardIterator> __result(__first, __first); + if (__first != __last) + { + if (++__first != __last) + { + if (__comp(*__first, *__result.first)) + __result.first = __first; + else + __result.second = __first; + while (++__first != __last) + { + _ForwardIterator __i = __first; + if (++__first == __last) + { + if (__comp(*__i, *__result.first)) + __result.first = __i; + else if (!__comp(*__i, *__result.second)) + __result.second = __i; + break; + } + else + { + if (__comp(*__first, *__i)) + { + if (__comp(*__first, *__result.first)) + __result.first = __first; + if (!__comp(*__i, *__result.second)) + __result.second = __i; + } + else + { + if (__comp(*__i, *__result.first)) + __result.first = __i; + if (!__comp(*__first, *__result.second)) + __result.second = __first; + } + } + } + } + } + return __result; +} + +template +_LIBCPP_NODISCARD_EXT inline +_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 +pair<_ForwardIterator, _ForwardIterator> +minmax_element(_ForwardIterator __first, _ForwardIterator __last) +{ + return _VSTD::minmax_element(__first, __last, + __less::value_type>()); +} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_MINMAX_ELEMENT_H diff --git a/lib/libcxx/include/__algorithm/mismatch.h b/lib/libcxx/include/__algorithm/mismatch.h new file mode 100644 index 000000000..fdd2bc8c0 --- /dev/null +++ b/lib/libcxx/include/__algorithm/mismatch.h @@ -0,0 +1,72 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_MISMATCH_H +#define _LIBCPP___ALGORITHM_MISMATCH_H + +#include <__config> +#include <__algorithm/comp.h> +#include <__iterator/iterator_traits.h> +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +_LIBCPP_NODISCARD_EXT inline _LIBCPP_INLINE_VISIBILITY + _LIBCPP_CONSTEXPR_AFTER_CXX17 pair<_InputIterator1, _InputIterator2> + mismatch(_InputIterator1 __first1, _InputIterator1 __last1, _InputIterator2 __first2, _BinaryPredicate __pred) { + for (; __first1 != __last1; ++__first1, (void)++__first2) + if (!__pred(*__first1, *__first2)) + break; + return pair<_InputIterator1, _InputIterator2>(__first1, __first2); +} + +template +_LIBCPP_NODISCARD_EXT inline _LIBCPP_INLINE_VISIBILITY + _LIBCPP_CONSTEXPR_AFTER_CXX17 pair<_InputIterator1, _InputIterator2> + mismatch(_InputIterator1 __first1, _InputIterator1 __last1, _InputIterator2 __first2) { + typedef typename iterator_traits<_InputIterator1>::value_type __v1; + typedef typename iterator_traits<_InputIterator2>::value_type __v2; + return _VSTD::mismatch(__first1, __last1, __first2, __equal_to<__v1, __v2>()); +} + +#if _LIBCPP_STD_VER > 11 +template +_LIBCPP_NODISCARD_EXT inline _LIBCPP_INLINE_VISIBILITY + _LIBCPP_CONSTEXPR_AFTER_CXX17 pair<_InputIterator1, _InputIterator2> + mismatch(_InputIterator1 __first1, _InputIterator1 __last1, _InputIterator2 __first2, _InputIterator2 __last2, + _BinaryPredicate __pred) { + for (; __first1 != __last1 && __first2 != __last2; ++__first1, (void)++__first2) + if (!__pred(*__first1, *__first2)) + break; + return pair<_InputIterator1, _InputIterator2>(__first1, __first2); +} + +template +_LIBCPP_NODISCARD_EXT inline _LIBCPP_INLINE_VISIBILITY + _LIBCPP_CONSTEXPR_AFTER_CXX17 pair<_InputIterator1, _InputIterator2> + mismatch(_InputIterator1 __first1, _InputIterator1 __last1, _InputIterator2 __first2, _InputIterator2 __last2) { + typedef typename iterator_traits<_InputIterator1>::value_type __v1; + typedef typename iterator_traits<_InputIterator2>::value_type __v2; + return _VSTD::mismatch(__first1, __last1, __first2, __last2, __equal_to<__v1, __v2>()); +} +#endif + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_MISMATCH_H diff --git a/lib/libcxx/include/__algorithm/move.h b/lib/libcxx/include/__algorithm/move.h new file mode 100644 index 000000000..f5fc74854 --- /dev/null +++ b/lib/libcxx/include/__algorithm/move.h @@ -0,0 +1,83 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_MOVE_H +#define _LIBCPP___ALGORITHM_MOVE_H + +#include <__config> +#include <__algorithm/unwrap_iter.h> +#include <__utility/move.h> +#include +#include +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +// move + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 +_OutputIterator +__move_constexpr(_InputIterator __first, _InputIterator __last, _OutputIterator __result) +{ + for (; __first != __last; ++__first, (void) ++__result) + *__result = _VSTD::move(*__first); + return __result; +} + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 +_OutputIterator +__move(_InputIterator __first, _InputIterator __last, _OutputIterator __result) +{ + return _VSTD::__move_constexpr(__first, __last, __result); +} + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 +typename enable_if +< + is_same::type, _Up>::value && + is_trivially_move_assignable<_Up>::value, + _Up* +>::type +__move(_Tp* __first, _Tp* __last, _Up* __result) +{ + const size_t __n = static_cast(__last - __first); + if (__n > 0) + _VSTD::memmove(__result, __first, __n * sizeof(_Up)); + return __result + __n; +} + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +_OutputIterator +move(_InputIterator __first, _InputIterator __last, _OutputIterator __result) +{ + if (__libcpp_is_constant_evaluated()) { + return _VSTD::__move_constexpr(__first, __last, __result); + } else { + return _VSTD::__rewrap_iter(__result, + _VSTD::__move(_VSTD::__unwrap_iter(__first), + _VSTD::__unwrap_iter(__last), + _VSTD::__unwrap_iter(__result))); + } +} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_MOVE_H diff --git a/lib/libcxx/include/__algorithm/move_backward.h b/lib/libcxx/include/__algorithm/move_backward.h new file mode 100644 index 000000000..1c93b9819 --- /dev/null +++ b/lib/libcxx/include/__algorithm/move_backward.h @@ -0,0 +1,84 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_MOVE_BACKWARD_H +#define _LIBCPP___ALGORITHM_MOVE_BACKWARD_H + +#include <__config> +#include <__algorithm/unwrap_iter.h> +#include +#include +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 +_OutputIterator +__move_backward_constexpr(_InputIterator __first, _InputIterator __last, _OutputIterator __result) +{ + while (__first != __last) + *--__result = _VSTD::move(*--__last); + return __result; +} + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 +_OutputIterator +__move_backward(_InputIterator __first, _InputIterator __last, _OutputIterator __result) +{ + return _VSTD::__move_backward_constexpr(__first, __last, __result); +} + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 +typename enable_if +< + is_same::type, _Up>::value && + is_trivially_move_assignable<_Up>::value, + _Up* +>::type +__move_backward(_Tp* __first, _Tp* __last, _Up* __result) +{ + const size_t __n = static_cast(__last - __first); + if (__n > 0) + { + __result -= __n; + _VSTD::memmove(__result, __first, __n * sizeof(_Up)); + } + return __result; +} + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +_BidirectionalIterator2 +move_backward(_BidirectionalIterator1 __first, _BidirectionalIterator1 __last, + _BidirectionalIterator2 __result) +{ + if (__libcpp_is_constant_evaluated()) { + return _VSTD::__move_backward_constexpr(__first, __last, __result); + } else { + return _VSTD::__rewrap_iter(__result, + _VSTD::__move_backward(_VSTD::__unwrap_iter(__first), + _VSTD::__unwrap_iter(__last), + _VSTD::__unwrap_iter(__result))); + } +} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_MOVE_BACKWARD_H diff --git a/lib/libcxx/include/__algorithm/next_permutation.h b/lib/libcxx/include/__algorithm/next_permutation.h new file mode 100644 index 000000000..a337e5efa --- /dev/null +++ b/lib/libcxx/include/__algorithm/next_permutation.h @@ -0,0 +1,77 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_NEXT_PERMUTATION_H +#define _LIBCPP___ALGORITHM_NEXT_PERMUTATION_H + +#include <__config> +#include <__algorithm/comp.h> +#include <__algorithm/comp_ref_type.h> +#include <__algorithm/reverse.h> +#include <__iterator/iterator_traits.h> +#include <__utility/swap.h> + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +_LIBCPP_CONSTEXPR_AFTER_CXX17 bool +__next_permutation(_BidirectionalIterator __first, _BidirectionalIterator __last, _Compare __comp) +{ + _BidirectionalIterator __i = __last; + if (__first == __last || __first == --__i) + return false; + while (true) + { + _BidirectionalIterator __ip1 = __i; + if (__comp(*--__i, *__ip1)) + { + _BidirectionalIterator __j = __last; + while (!__comp(*__i, *--__j)) + ; + swap(*__i, *__j); + _VSTD::reverse(__ip1, __last); + return true; + } + if (__i == __first) + { + _VSTD::reverse(__first, __last); + return false; + } + } +} + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +bool +next_permutation(_BidirectionalIterator __first, _BidirectionalIterator __last, _Compare __comp) +{ + typedef typename __comp_ref_type<_Compare>::type _Comp_ref; + return _VSTD::__next_permutation<_Comp_ref>(__first, __last, __comp); +} + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +bool +next_permutation(_BidirectionalIterator __first, _BidirectionalIterator __last) +{ + return _VSTD::next_permutation(__first, __last, + __less::value_type>()); +} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_NEXT_PERMUTATION_H diff --git a/lib/libcxx/include/__algorithm/none_of.h b/lib/libcxx/include/__algorithm/none_of.h new file mode 100644 index 000000000..285691520 --- /dev/null +++ b/lib/libcxx/include/__algorithm/none_of.h @@ -0,0 +1,37 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_NONE_OF_H +#define _LIBCPP___ALGORITHM_NONE_OF_H + +#include <__config> + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +_LIBCPP_NODISCARD_EXT inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 bool +none_of(_InputIterator __first, _InputIterator __last, _Predicate __pred) { + for (; __first != __last; ++__first) + if (__pred(*__first)) + return false; + return true; +} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_NONE_OF_H diff --git a/lib/libcxx/include/__algorithm/nth_element.h b/lib/libcxx/include/__algorithm/nth_element.h new file mode 100644 index 000000000..67a03cfc3 --- /dev/null +++ b/lib/libcxx/include/__algorithm/nth_element.h @@ -0,0 +1,244 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_NTH_ELEMENT_H +#define _LIBCPP___ALGORITHM_NTH_ELEMENT_H + +#include <__config> +#include <__algorithm/comp.h> +#include <__algorithm/comp_ref_type.h> +#include <__algorithm/sort.h> +#include <__iterator/iterator_traits.h> +#include <__utility/swap.h> + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +_LIBCPP_CONSTEXPR_AFTER_CXX11 bool +__nth_element_find_guard(_RandomAccessIterator& __i, _RandomAccessIterator& __j, + _RandomAccessIterator __m, _Compare __comp) +{ + // manually guard downward moving __j against __i + while (true) { + if (__i == --__j) { + return false; + } + if (__comp(*__j, *__m)) { + return true; // found guard for downward moving __j, now use unguarded partition + } + } +} + +template +_LIBCPP_CONSTEXPR_AFTER_CXX11 void +__nth_element(_RandomAccessIterator __first, _RandomAccessIterator __nth, _RandomAccessIterator __last, _Compare __comp) +{ + // _Compare is known to be a reference type + typedef typename iterator_traits<_RandomAccessIterator>::difference_type difference_type; + const difference_type __limit = 7; + while (true) + { + if (__nth == __last) + return; + difference_type __len = __last - __first; + switch (__len) + { + case 0: + case 1: + return; + case 2: + if (__comp(*--__last, *__first)) + swap(*__first, *__last); + return; + case 3: + { + _RandomAccessIterator __m = __first; + _VSTD::__sort3<_Compare>(__first, ++__m, --__last, __comp); + return; + } + } + if (__len <= __limit) + { + _VSTD::__selection_sort<_Compare>(__first, __last, __comp); + return; + } + // __len > __limit >= 3 + _RandomAccessIterator __m = __first + __len/2; + _RandomAccessIterator __lm1 = __last; + unsigned __n_swaps = _VSTD::__sort3<_Compare>(__first, __m, --__lm1, __comp); + // *__m is median + // partition [__first, __m) < *__m and *__m <= [__m, __last) + // (this inhibits tossing elements equivalent to __m around unnecessarily) + _RandomAccessIterator __i = __first; + _RandomAccessIterator __j = __lm1; + // j points beyond range to be tested, *__lm1 is known to be <= *__m + // The search going up is known to be guarded but the search coming down isn't. + // Prime the downward search with a guard. + if (!__comp(*__i, *__m)) // if *__first == *__m + { + // *__first == *__m, *__first doesn't go in first part + if (_VSTD::__nth_element_find_guard<_Compare>(__i, __j, __m, __comp)) { + swap(*__i, *__j); + ++__n_swaps; + } else { + // *__first == *__m, *__m <= all other elements + // Partition instead into [__first, __i) == *__first and *__first < [__i, __last) + ++__i; // __first + 1 + __j = __last; + if (!__comp(*__first, *--__j)) { // we need a guard if *__first == *(__last-1) + while (true) { + if (__i == __j) { + return; // [__first, __last) all equivalent elements + } else if (__comp(*__first, *__i)) { + swap(*__i, *__j); + ++__n_swaps; + ++__i; + break; + } + ++__i; + } + } + // [__first, __i) == *__first and *__first < [__j, __last) and __j == __last - 1 + if (__i == __j) { + return; + } + while (true) { + while (!__comp(*__first, *__i)) + ++__i; + while (__comp(*__first, *--__j)) + ; + if (__i >= __j) + break; + swap(*__i, *__j); + ++__n_swaps; + ++__i; + } + // [__first, __i) == *__first and *__first < [__i, __last) + // The first part is sorted, + if (__nth < __i) { + return; + } + // __nth_element the second part + // _VSTD::__nth_element<_Compare>(__i, __nth, __last, __comp); + __first = __i; + continue; + } + } + ++__i; + // j points beyond range to be tested, *__lm1 is known to be <= *__m + // if not yet partitioned... + if (__i < __j) + { + // known that *(__i - 1) < *__m + while (true) + { + // __m still guards upward moving __i + while (__comp(*__i, *__m)) + ++__i; + // It is now known that a guard exists for downward moving __j + while (!__comp(*--__j, *__m)) + ; + if (__i >= __j) + break; + swap(*__i, *__j); + ++__n_swaps; + // It is known that __m != __j + // If __m just moved, follow it + if (__m == __i) + __m = __j; + ++__i; + } + } + // [__first, __i) < *__m and *__m <= [__i, __last) + if (__i != __m && __comp(*__m, *__i)) + { + swap(*__i, *__m); + ++__n_swaps; + } + // [__first, __i) < *__i and *__i <= [__i+1, __last) + if (__nth == __i) + return; + if (__n_swaps == 0) + { + // We were given a perfectly partitioned sequence. Coincidence? + if (__nth < __i) + { + // Check for [__first, __i) already sorted + __j = __m = __first; + while (true) { + if (++__j == __i) { + // [__first, __i) sorted + return; + } + if (__comp(*__j, *__m)) { + // not yet sorted, so sort + break; + } + __m = __j; + } + } + else + { + // Check for [__i, __last) already sorted + __j = __m = __i; + while (true) { + if (++__j == __last) { + // [__i, __last) sorted + return; + } + if (__comp(*__j, *__m)) { + // not yet sorted, so sort + break; + } + __m = __j; + } + } + } + // __nth_element on range containing __nth + if (__nth < __i) + { + // _VSTD::__nth_element<_Compare>(__first, __nth, __i, __comp); + __last = __i; + } + else + { + // _VSTD::__nth_element<_Compare>(__i+1, __nth, __last, __comp); + __first = ++__i; + } + } +} + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +void +nth_element(_RandomAccessIterator __first, _RandomAccessIterator __nth, _RandomAccessIterator __last, _Compare __comp) +{ + typedef typename __comp_ref_type<_Compare>::type _Comp_ref; + _VSTD::__nth_element<_Comp_ref>(__first, __nth, __last, __comp); +} + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +void +nth_element(_RandomAccessIterator __first, _RandomAccessIterator __nth, _RandomAccessIterator __last) +{ + _VSTD::nth_element(__first, __nth, __last, __less::value_type>()); +} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_NTH_ELEMENT_H diff --git a/lib/libcxx/include/__algorithm/partial_sort.h b/lib/libcxx/include/__algorithm/partial_sort.h new file mode 100644 index 000000000..4f9872c49 --- /dev/null +++ b/lib/libcxx/include/__algorithm/partial_sort.h @@ -0,0 +1,71 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_PARTIAL_SORT_H +#define _LIBCPP___ALGORITHM_PARTIAL_SORT_H + +#include <__config> +#include <__algorithm/comp.h> +#include <__algorithm/comp_ref_type.h> +#include <__algorithm/make_heap.h> +#include <__algorithm/sift_down.h> +#include <__algorithm/sort_heap.h> +#include <__iterator/iterator_traits.h> +#include <__utility/swap.h> + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +_LIBCPP_CONSTEXPR_AFTER_CXX17 void +__partial_sort(_RandomAccessIterator __first, _RandomAccessIterator __middle, _RandomAccessIterator __last, + _Compare __comp) +{ + _VSTD::__make_heap<_Compare>(__first, __middle, __comp); + typename iterator_traits<_RandomAccessIterator>::difference_type __len = __middle - __first; + for (_RandomAccessIterator __i = __middle; __i != __last; ++__i) + { + if (__comp(*__i, *__first)) + { + swap(*__i, *__first); + _VSTD::__sift_down<_Compare>(__first, __middle, __comp, __len, __first); + } + } + _VSTD::__sort_heap<_Compare>(__first, __middle, __comp); +} + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +void +partial_sort(_RandomAccessIterator __first, _RandomAccessIterator __middle, _RandomAccessIterator __last, + _Compare __comp) +{ + typedef typename __comp_ref_type<_Compare>::type _Comp_ref; + _VSTD::__partial_sort<_Comp_ref>(__first, __middle, __last, __comp); +} + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +void +partial_sort(_RandomAccessIterator __first, _RandomAccessIterator __middle, _RandomAccessIterator __last) +{ + _VSTD::partial_sort(__first, __middle, __last, + __less::value_type>()); +} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_PARTIAL_SORT_H diff --git a/lib/libcxx/include/__algorithm/partial_sort_copy.h b/lib/libcxx/include/__algorithm/partial_sort_copy.h new file mode 100644 index 000000000..31a12615f --- /dev/null +++ b/lib/libcxx/include/__algorithm/partial_sort_copy.h @@ -0,0 +1,77 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_PARTIAL_SORT_COPY_H +#define _LIBCPP___ALGORITHM_PARTIAL_SORT_COPY_H + +#include <__config> +#include <__algorithm/comp.h> +#include <__algorithm/comp_ref_type.h> +#include <__algorithm/make_heap.h> +#include <__algorithm/sift_down.h> +#include <__algorithm/sort_heap.h> +#include <__iterator/iterator_traits.h> +#include // swap + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +_LIBCPP_CONSTEXPR_AFTER_CXX17 _RandomAccessIterator +__partial_sort_copy(_InputIterator __first, _InputIterator __last, + _RandomAccessIterator __result_first, _RandomAccessIterator __result_last, _Compare __comp) +{ + _RandomAccessIterator __r = __result_first; + if (__r != __result_last) + { + for (; __first != __last && __r != __result_last; ++__first, (void) ++__r) + *__r = *__first; + _VSTD::__make_heap<_Compare>(__result_first, __r, __comp); + typename iterator_traits<_RandomAccessIterator>::difference_type __len = __r - __result_first; + for (; __first != __last; ++__first) + if (__comp(*__first, *__result_first)) + { + *__result_first = *__first; + _VSTD::__sift_down<_Compare>(__result_first, __r, __comp, __len, __result_first); + } + _VSTD::__sort_heap<_Compare>(__result_first, __r, __comp); + } + return __r; +} + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +_RandomAccessIterator +partial_sort_copy(_InputIterator __first, _InputIterator __last, + _RandomAccessIterator __result_first, _RandomAccessIterator __result_last, _Compare __comp) +{ + typedef typename __comp_ref_type<_Compare>::type _Comp_ref; + return _VSTD::__partial_sort_copy<_Comp_ref>(__first, __last, __result_first, __result_last, __comp); +} + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +_RandomAccessIterator +partial_sort_copy(_InputIterator __first, _InputIterator __last, + _RandomAccessIterator __result_first, _RandomAccessIterator __result_last) +{ + return _VSTD::partial_sort_copy(__first, __last, __result_first, __result_last, + __less::value_type>()); +} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_PARTIAL_SORT_COPY_H diff --git a/lib/libcxx/include/__algorithm/partition.h b/lib/libcxx/include/__algorithm/partition.h new file mode 100644 index 000000000..c859eaca2 --- /dev/null +++ b/lib/libcxx/include/__algorithm/partition.h @@ -0,0 +1,88 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_PARTITION_H +#define _LIBCPP___ALGORITHM_PARTITION_H + +#include <__config> +#include <__iterator/iterator_traits.h> +#include <__utility/swap.h> +#include // pair +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +_LIBCPP_CONSTEXPR_AFTER_CXX17 _ForwardIterator +__partition(_ForwardIterator __first, _ForwardIterator __last, _Predicate __pred, forward_iterator_tag) +{ + while (true) + { + if (__first == __last) + return __first; + if (!__pred(*__first)) + break; + ++__first; + } + for (_ForwardIterator __p = __first; ++__p != __last;) + { + if (__pred(*__p)) + { + swap(*__first, *__p); + ++__first; + } + } + return __first; +} + +template +_LIBCPP_CONSTEXPR_AFTER_CXX17 _BidirectionalIterator +__partition(_BidirectionalIterator __first, _BidirectionalIterator __last, _Predicate __pred, + bidirectional_iterator_tag) +{ + while (true) + { + while (true) + { + if (__first == __last) + return __first; + if (!__pred(*__first)) + break; + ++__first; + } + do + { + if (__first == --__last) + return __first; + } while (!__pred(*__last)); + swap(*__first, *__last); + ++__first; + } +} + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +_ForwardIterator +partition(_ForwardIterator __first, _ForwardIterator __last, _Predicate __pred) +{ + return _VSTD::__partition::type> + (__first, __last, __pred, typename iterator_traits<_ForwardIterator>::iterator_category()); +} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_PARTITION_H diff --git a/lib/libcxx/include/__algorithm/partition_copy.h b/lib/libcxx/include/__algorithm/partition_copy.h new file mode 100644 index 000000000..445eacdfb --- /dev/null +++ b/lib/libcxx/include/__algorithm/partition_copy.h @@ -0,0 +1,52 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_PARTITION_COPY_H +#define _LIBCPP___ALGORITHM_PARTITION_COPY_H + +#include <__config> +#include <__iterator/iterator_traits.h> +#include // pair + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +_LIBCPP_CONSTEXPR_AFTER_CXX17 pair<_OutputIterator1, _OutputIterator2> +partition_copy(_InputIterator __first, _InputIterator __last, + _OutputIterator1 __out_true, _OutputIterator2 __out_false, + _Predicate __pred) +{ + for (; __first != __last; ++__first) + { + if (__pred(*__first)) + { + *__out_true = *__first; + ++__out_true; + } + else + { + *__out_false = *__first; + ++__out_false; + } + } + return pair<_OutputIterator1, _OutputIterator2>(__out_true, __out_false); +} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_PARTITION_COPY_H diff --git a/lib/libcxx/include/__algorithm/partition_point.h b/lib/libcxx/include/__algorithm/partition_point.h new file mode 100644 index 000000000..12ddacf9d --- /dev/null +++ b/lib/libcxx/include/__algorithm/partition_point.h @@ -0,0 +1,51 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_PARTITION_POINT_H +#define _LIBCPP___ALGORITHM_PARTITION_POINT_H + +#include <__config> +#include <__algorithm/half_positive.h> +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +_LIBCPP_CONSTEXPR_AFTER_CXX17 _ForwardIterator +partition_point(_ForwardIterator __first, _ForwardIterator __last, _Predicate __pred) +{ + typedef typename iterator_traits<_ForwardIterator>::difference_type difference_type; + difference_type __len = _VSTD::distance(__first, __last); + while (__len != 0) + { + difference_type __l2 = _VSTD::__half_positive(__len); + _ForwardIterator __m = __first; + _VSTD::advance(__m, __l2); + if (__pred(*__m)) + { + __first = ++__m; + __len -= __l2 + 1; + } + else + __len = __l2; + } + return __first; +} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_PARTITION_POINT_H diff --git a/lib/libcxx/include/__algorithm/pop_heap.h b/lib/libcxx/include/__algorithm/pop_heap.h new file mode 100644 index 000000000..7ebbef25e --- /dev/null +++ b/lib/libcxx/include/__algorithm/pop_heap.h @@ -0,0 +1,62 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_POP_HEAP_H +#define _LIBCPP___ALGORITHM_POP_HEAP_H + +#include <__config> +#include <__algorithm/comp.h> +#include <__algorithm/comp_ref_type.h> +#include <__algorithm/sift_down.h> +#include <__iterator/iterator_traits.h> +#include <__utility/swap.h> + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +void +__pop_heap(_RandomAccessIterator __first, _RandomAccessIterator __last, _Compare __comp, + typename iterator_traits<_RandomAccessIterator>::difference_type __len) +{ + if (__len > 1) + { + swap(*__first, *--__last); + _VSTD::__sift_down<_Compare>(__first, __last, __comp, __len - 1, __first); + } +} + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +void +pop_heap(_RandomAccessIterator __first, _RandomAccessIterator __last, _Compare __comp) +{ + typedef typename __comp_ref_type<_Compare>::type _Comp_ref; + _VSTD::__pop_heap<_Comp_ref>(__first, __last, __comp, __last - __first); +} + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +void +pop_heap(_RandomAccessIterator __first, _RandomAccessIterator __last) +{ + _VSTD::pop_heap(__first, __last, __less::value_type>()); +} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_POP_HEAP_H diff --git a/lib/libcxx/include/__algorithm/prev_permutation.h b/lib/libcxx/include/__algorithm/prev_permutation.h new file mode 100644 index 000000000..d6daa73ad --- /dev/null +++ b/lib/libcxx/include/__algorithm/prev_permutation.h @@ -0,0 +1,77 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_PREV_PERMUTATION_H +#define _LIBCPP___ALGORITHM_PREV_PERMUTATION_H + +#include <__config> +#include <__algorithm/comp.h> +#include <__algorithm/comp_ref_type.h> +#include <__algorithm/reverse.h> +#include <__iterator/iterator_traits.h> +#include <__utility/swap.h> + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +_LIBCPP_CONSTEXPR_AFTER_CXX17 bool +__prev_permutation(_BidirectionalIterator __first, _BidirectionalIterator __last, _Compare __comp) +{ + _BidirectionalIterator __i = __last; + if (__first == __last || __first == --__i) + return false; + while (true) + { + _BidirectionalIterator __ip1 = __i; + if (__comp(*__ip1, *--__i)) + { + _BidirectionalIterator __j = __last; + while (!__comp(*--__j, *__i)) + ; + swap(*__i, *__j); + _VSTD::reverse(__ip1, __last); + return true; + } + if (__i == __first) + { + _VSTD::reverse(__first, __last); + return false; + } + } +} + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +bool +prev_permutation(_BidirectionalIterator __first, _BidirectionalIterator __last, _Compare __comp) +{ + typedef typename __comp_ref_type<_Compare>::type _Comp_ref; + return _VSTD::__prev_permutation<_Comp_ref>(__first, __last, __comp); +} + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +bool +prev_permutation(_BidirectionalIterator __first, _BidirectionalIterator __last) +{ + return _VSTD::prev_permutation(__first, __last, + __less::value_type>()); +} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_PREV_PERMUTATION_H diff --git a/lib/libcxx/include/__algorithm/push_heap.h b/lib/libcxx/include/__algorithm/push_heap.h new file mode 100644 index 000000000..82a7c123d --- /dev/null +++ b/lib/libcxx/include/__algorithm/push_heap.h @@ -0,0 +1,75 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_PUSH_HEAP_H +#define _LIBCPP___ALGORITHM_PUSH_HEAP_H + +#include <__config> +#include <__algorithm/comp.h> +#include <__algorithm/comp_ref_type.h> +#include <__iterator/iterator_traits.h> +#include <__utility/move.h> + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +_LIBCPP_CONSTEXPR_AFTER_CXX11 void +__sift_up(_RandomAccessIterator __first, _RandomAccessIterator __last, _Compare __comp, + typename iterator_traits<_RandomAccessIterator>::difference_type __len) +{ + typedef typename iterator_traits<_RandomAccessIterator>::value_type value_type; + if (__len > 1) + { + __len = (__len - 2) / 2; + _RandomAccessIterator __ptr = __first + __len; + if (__comp(*__ptr, *--__last)) + { + value_type __t(_VSTD::move(*__last)); + do + { + *__last = _VSTD::move(*__ptr); + __last = __ptr; + if (__len == 0) + break; + __len = (__len - 1) / 2; + __ptr = __first + __len; + } while (__comp(*__ptr, __t)); + *__last = _VSTD::move(__t); + } + } +} + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +void +push_heap(_RandomAccessIterator __first, _RandomAccessIterator __last, _Compare __comp) +{ + typedef typename __comp_ref_type<_Compare>::type _Comp_ref; + _VSTD::__sift_up<_Comp_ref>(__first, __last, __comp, __last - __first); +} + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +void +push_heap(_RandomAccessIterator __first, _RandomAccessIterator __last) +{ + _VSTD::push_heap(__first, __last, __less::value_type>()); +} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_PUSH_HEAP_H diff --git a/lib/libcxx/include/__algorithm/remove.h b/lib/libcxx/include/__algorithm/remove.h new file mode 100644 index 000000000..4717d7d97 --- /dev/null +++ b/lib/libcxx/include/__algorithm/remove.h @@ -0,0 +1,50 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_REMOVE_H +#define _LIBCPP___ALGORITHM_REMOVE_H + +#include <__config> +#include <__algorithm/find.h> +#include <__algorithm/find_if.h> +#include <__utility/move.h> + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +_LIBCPP_NODISCARD_EXT _LIBCPP_CONSTEXPR_AFTER_CXX17 _ForwardIterator +remove(_ForwardIterator __first, _ForwardIterator __last, const _Tp& __value_) +{ + __first = _VSTD::find(__first, __last, __value_); + if (__first != __last) + { + _ForwardIterator __i = __first; + while (++__i != __last) + { + if (!(*__i == __value_)) + { + *__first = _VSTD::move(*__i); + ++__first; + } + } + } + return __first; +} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_REMOVE_H diff --git a/lib/libcxx/include/__algorithm/remove_copy.h b/lib/libcxx/include/__algorithm/remove_copy.h new file mode 100644 index 000000000..5d2b64034 --- /dev/null +++ b/lib/libcxx/include/__algorithm/remove_copy.h @@ -0,0 +1,43 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_REMOVE_COPY_H +#define _LIBCPP___ALGORITHM_REMOVE_COPY_H + +#include <__config> + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +_OutputIterator +remove_copy(_InputIterator __first, _InputIterator __last, _OutputIterator __result, const _Tp& __value_) +{ + for (; __first != __last; ++__first) + { + if (!(*__first == __value_)) + { + *__result = *__first; + ++__result; + } + } + return __result; +} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_REMOVE_COPY_H diff --git a/lib/libcxx/include/__algorithm/remove_copy_if.h b/lib/libcxx/include/__algorithm/remove_copy_if.h new file mode 100644 index 000000000..44822564a --- /dev/null +++ b/lib/libcxx/include/__algorithm/remove_copy_if.h @@ -0,0 +1,43 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_REMOVE_COPY_IF_H +#define _LIBCPP___ALGORITHM_REMOVE_COPY_IF_H + +#include <__config> + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +_OutputIterator +remove_copy_if(_InputIterator __first, _InputIterator __last, _OutputIterator __result, _Predicate __pred) +{ + for (; __first != __last; ++__first) + { + if (!__pred(*__first)) + { + *__result = *__first; + ++__result; + } + } + return __result; +} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_REMOVE_COPY_IF_H diff --git a/lib/libcxx/include/__algorithm/remove_if.h b/lib/libcxx/include/__algorithm/remove_if.h new file mode 100644 index 000000000..e506b4c67 --- /dev/null +++ b/lib/libcxx/include/__algorithm/remove_if.h @@ -0,0 +1,51 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_REMOVE_IF_H +#define _LIBCPP___ALGORITHM_REMOVE_IF_H + +#include <__config> +#include <__algorithm/find_if.h> +#include +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +_LIBCPP_NODISCARD_EXT _LIBCPP_CONSTEXPR_AFTER_CXX17 _ForwardIterator +remove_if(_ForwardIterator __first, _ForwardIterator __last, _Predicate __pred) +{ + __first = _VSTD::find_if<_ForwardIterator, typename add_lvalue_reference<_Predicate>::type> + (__first, __last, __pred); + if (__first != __last) + { + _ForwardIterator __i = __first; + while (++__i != __last) + { + if (!__pred(*__i)) + { + *__first = _VSTD::move(*__i); + ++__first; + } + } + } + return __first; +} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_REMOVE_IF_H diff --git a/lib/libcxx/include/__algorithm/replace.h b/lib/libcxx/include/__algorithm/replace.h new file mode 100644 index 000000000..b723ffeea --- /dev/null +++ b/lib/libcxx/include/__algorithm/replace.h @@ -0,0 +1,37 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_REPLACE_H +#define _LIBCPP___ALGORITHM_REPLACE_H + +#include <__config> + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +void +replace(_ForwardIterator __first, _ForwardIterator __last, const _Tp& __old_value, const _Tp& __new_value) +{ + for (; __first != __last; ++__first) + if (*__first == __old_value) + *__first = __new_value; +} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_REPLACE_H diff --git a/lib/libcxx/include/__algorithm/replace_copy.h b/lib/libcxx/include/__algorithm/replace_copy.h new file mode 100644 index 000000000..1923a57e4 --- /dev/null +++ b/lib/libcxx/include/__algorithm/replace_copy.h @@ -0,0 +1,41 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_REPLACE_COPY_H +#define _LIBCPP___ALGORITHM_REPLACE_COPY_H + +#include <__config> + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +_OutputIterator +replace_copy(_InputIterator __first, _InputIterator __last, _OutputIterator __result, + const _Tp& __old_value, const _Tp& __new_value) +{ + for (; __first != __last; ++__first, (void) ++__result) + if (*__first == __old_value) + *__result = __new_value; + else + *__result = *__first; + return __result; +} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_REPLACE_COPY_H diff --git a/lib/libcxx/include/__algorithm/replace_copy_if.h b/lib/libcxx/include/__algorithm/replace_copy_if.h new file mode 100644 index 000000000..72b6f7369 --- /dev/null +++ b/lib/libcxx/include/__algorithm/replace_copy_if.h @@ -0,0 +1,41 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_REPLACE_COPY_IF_H +#define _LIBCPP___ALGORITHM_REPLACE_COPY_IF_H + +#include <__config> + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +_OutputIterator +replace_copy_if(_InputIterator __first, _InputIterator __last, _OutputIterator __result, + _Predicate __pred, const _Tp& __new_value) +{ + for (; __first != __last; ++__first, (void) ++__result) + if (__pred(*__first)) + *__result = __new_value; + else + *__result = *__first; + return __result; +} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_REPLACE_COPY_IF_H diff --git a/lib/libcxx/include/__algorithm/replace_if.h b/lib/libcxx/include/__algorithm/replace_if.h new file mode 100644 index 000000000..49101a5ce --- /dev/null +++ b/lib/libcxx/include/__algorithm/replace_if.h @@ -0,0 +1,37 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_REPLACE_IF_H +#define _LIBCPP___ALGORITHM_REPLACE_IF_H + +#include <__config> + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +void +replace_if(_ForwardIterator __first, _ForwardIterator __last, _Predicate __pred, const _Tp& __new_value) +{ + for (; __first != __last; ++__first) + if (__pred(*__first)) + *__first = __new_value; +} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_REPLACE_IF_H diff --git a/lib/libcxx/include/__algorithm/reverse.h b/lib/libcxx/include/__algorithm/reverse.h new file mode 100644 index 000000000..e538de114 --- /dev/null +++ b/lib/libcxx/include/__algorithm/reverse.h @@ -0,0 +1,61 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_REVERSE_H +#define _LIBCPP___ALGORITHM_REVERSE_H + +#include <__config> +#include <__algorithm/iter_swap.h> +#include <__iterator/iterator_traits.h> + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +void +__reverse(_BidirectionalIterator __first, _BidirectionalIterator __last, bidirectional_iterator_tag) +{ + while (__first != __last) + { + if (__first == --__last) + break; + _VSTD::iter_swap(__first, __last); + ++__first; + } +} + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +void +__reverse(_RandomAccessIterator __first, _RandomAccessIterator __last, random_access_iterator_tag) +{ + if (__first != __last) + for (; __first < --__last; ++__first) + _VSTD::iter_swap(__first, __last); +} + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +void +reverse(_BidirectionalIterator __first, _BidirectionalIterator __last) +{ + _VSTD::__reverse(__first, __last, typename iterator_traits<_BidirectionalIterator>::iterator_category()); +} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_REVERSE_H diff --git a/lib/libcxx/include/__algorithm/reverse_copy.h b/lib/libcxx/include/__algorithm/reverse_copy.h new file mode 100644 index 000000000..48ce60cf8 --- /dev/null +++ b/lib/libcxx/include/__algorithm/reverse_copy.h @@ -0,0 +1,37 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_REVERSE_COPY_H +#define _LIBCPP___ALGORITHM_REVERSE_COPY_H + +#include <__config> + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +_OutputIterator +reverse_copy(_BidirectionalIterator __first, _BidirectionalIterator __last, _OutputIterator __result) +{ + for (; __first != __last; ++__result) + *__result = *--__last; + return __result; +} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_REVERSE_COPY_H diff --git a/lib/libcxx/include/__algorithm/rotate.h b/lib/libcxx/include/__algorithm/rotate.h new file mode 100644 index 000000000..0c9ccd7bf --- /dev/null +++ b/lib/libcxx/include/__algorithm/rotate.h @@ -0,0 +1,205 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_ROTATE_H +#define _LIBCPP___ALGORITHM_ROTATE_H + +#include <__algorithm/move.h> +#include <__algorithm/move_backward.h> +#include <__algorithm/swap_ranges.h> +#include <__config> +#include <__iterator/iterator_traits.h> +#include <__iterator/next.h> +#include <__iterator/prev.h> +#include <__utility/swap.h> +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +_LIBCPP_CONSTEXPR_AFTER_CXX11 _ForwardIterator +__rotate_left(_ForwardIterator __first, _ForwardIterator __last) +{ + typedef typename iterator_traits<_ForwardIterator>::value_type value_type; + value_type __tmp = _VSTD::move(*__first); + _ForwardIterator __lm1 = _VSTD::move(_VSTD::next(__first), __last, __first); + *__lm1 = _VSTD::move(__tmp); + return __lm1; +} + +template +_LIBCPP_CONSTEXPR_AFTER_CXX11 _BidirectionalIterator +__rotate_right(_BidirectionalIterator __first, _BidirectionalIterator __last) +{ + typedef typename iterator_traits<_BidirectionalIterator>::value_type value_type; + _BidirectionalIterator __lm1 = _VSTD::prev(__last); + value_type __tmp = _VSTD::move(*__lm1); + _BidirectionalIterator __fp1 = _VSTD::move_backward(__first, __lm1, __last); + *__first = _VSTD::move(__tmp); + return __fp1; +} + +template +_LIBCPP_CONSTEXPR_AFTER_CXX14 _ForwardIterator +__rotate_forward(_ForwardIterator __first, _ForwardIterator __middle, _ForwardIterator __last) +{ + _ForwardIterator __i = __middle; + while (true) + { + swap(*__first, *__i); + ++__first; + if (++__i == __last) + break; + if (__first == __middle) + __middle = __i; + } + _ForwardIterator __r = __first; + if (__first != __middle) + { + __i = __middle; + while (true) + { + swap(*__first, *__i); + ++__first; + if (++__i == __last) + { + if (__first == __middle) + break; + __i = __middle; + } + else if (__first == __middle) + __middle = __i; + } + } + return __r; +} + +template +inline _LIBCPP_INLINE_VISIBILITY +_LIBCPP_CONSTEXPR_AFTER_CXX14 _Integral +__algo_gcd(_Integral __x, _Integral __y) +{ + do + { + _Integral __t = __x % __y; + __x = __y; + __y = __t; + } while (__y); + return __x; +} + +template +_LIBCPP_CONSTEXPR_AFTER_CXX14 _RandomAccessIterator +__rotate_gcd(_RandomAccessIterator __first, _RandomAccessIterator __middle, _RandomAccessIterator __last) +{ + typedef typename iterator_traits<_RandomAccessIterator>::difference_type difference_type; + typedef typename iterator_traits<_RandomAccessIterator>::value_type value_type; + + const difference_type __m1 = __middle - __first; + const difference_type __m2 = __last - __middle; + if (__m1 == __m2) + { + _VSTD::swap_ranges(__first, __middle, __middle); + return __middle; + } + const difference_type __g = _VSTD::__algo_gcd(__m1, __m2); + for (_RandomAccessIterator __p = __first + __g; __p != __first;) + { + value_type __t(_VSTD::move(*--__p)); + _RandomAccessIterator __p1 = __p; + _RandomAccessIterator __p2 = __p1 + __m1; + do + { + *__p1 = _VSTD::move(*__p2); + __p1 = __p2; + const difference_type __d = __last - __p2; + if (__m1 < __d) + __p2 += __m1; + else + __p2 = __first + (__m1 - __d); + } while (__p2 != __p); + *__p1 = _VSTD::move(__t); + } + return __first + __m2; +} + +template +inline _LIBCPP_INLINE_VISIBILITY +_LIBCPP_CONSTEXPR_AFTER_CXX11 _ForwardIterator +__rotate(_ForwardIterator __first, _ForwardIterator __middle, _ForwardIterator __last, + _VSTD::forward_iterator_tag) +{ + typedef typename iterator_traits<_ForwardIterator>::value_type value_type; + if (is_trivially_move_assignable::value) + { + if (_VSTD::next(__first) == __middle) + return _VSTD::__rotate_left(__first, __last); + } + return _VSTD::__rotate_forward(__first, __middle, __last); +} + +template +inline _LIBCPP_INLINE_VISIBILITY +_LIBCPP_CONSTEXPR_AFTER_CXX11 _BidirectionalIterator +__rotate(_BidirectionalIterator __first, _BidirectionalIterator __middle, _BidirectionalIterator __last, + bidirectional_iterator_tag) +{ + typedef typename iterator_traits<_BidirectionalIterator>::value_type value_type; + if (is_trivially_move_assignable::value) + { + if (_VSTD::next(__first) == __middle) + return _VSTD::__rotate_left(__first, __last); + if (_VSTD::next(__middle) == __last) + return _VSTD::__rotate_right(__first, __last); + } + return _VSTD::__rotate_forward(__first, __middle, __last); +} + +template +inline _LIBCPP_INLINE_VISIBILITY +_LIBCPP_CONSTEXPR_AFTER_CXX11 _RandomAccessIterator +__rotate(_RandomAccessIterator __first, _RandomAccessIterator __middle, _RandomAccessIterator __last, + random_access_iterator_tag) +{ + typedef typename iterator_traits<_RandomAccessIterator>::value_type value_type; + if (is_trivially_move_assignable::value) + { + if (_VSTD::next(__first) == __middle) + return _VSTD::__rotate_left(__first, __last); + if (_VSTD::next(__middle) == __last) + return _VSTD::__rotate_right(__first, __last); + return _VSTD::__rotate_gcd(__first, __middle, __last); + } + return _VSTD::__rotate_forward(__first, __middle, __last); +} + +template +inline _LIBCPP_INLINE_VISIBILITY +_LIBCPP_CONSTEXPR_AFTER_CXX17 _ForwardIterator +rotate(_ForwardIterator __first, _ForwardIterator __middle, _ForwardIterator __last) +{ + if (__first == __middle) + return __last; + if (__middle == __last) + return __first; + return _VSTD::__rotate(__first, __middle, __last, + typename iterator_traits<_ForwardIterator>::iterator_category()); +} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_ROTATE_H diff --git a/lib/libcxx/include/__algorithm/rotate_copy.h b/lib/libcxx/include/__algorithm/rotate_copy.h new file mode 100644 index 000000000..d5ab7d3b5 --- /dev/null +++ b/lib/libcxx/include/__algorithm/rotate_copy.h @@ -0,0 +1,38 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_ROTATE_COPY_H +#define _LIBCPP___ALGORITHM_ROTATE_COPY_H + +#include <__config> +#include <__algorithm/copy.h> +#include +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +_OutputIterator +rotate_copy(_ForwardIterator __first, _ForwardIterator __middle, _ForwardIterator __last, _OutputIterator __result) +{ + return _VSTD::copy(__first, __middle, _VSTD::copy(__middle, __last, __result)); +} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_ROTATE_COPY_H diff --git a/lib/libcxx/include/__algorithm/sample.h b/lib/libcxx/include/__algorithm/sample.h new file mode 100644 index 000000000..2aac6ffa8 --- /dev/null +++ b/lib/libcxx/include/__algorithm/sample.h @@ -0,0 +1,101 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_SAMPLE_H +#define _LIBCPP___ALGORITHM_SAMPLE_H + +#include <__config> +#include <__algorithm/min.h> +#include <__random/uniform_int_distribution.h> +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +_LIBCPP_INLINE_VISIBILITY +_SampleIterator __sample(_PopulationIterator __first, + _PopulationIterator __last, _SampleIterator __output_iter, + _Distance __n, + _UniformRandomNumberGenerator & __g, + input_iterator_tag) { + + _Distance __k = 0; + for (; __first != __last && __k < __n; ++__first, (void) ++__k) + __output_iter[__k] = *__first; + _Distance __sz = __k; + for (; __first != __last; ++__first, (void) ++__k) { + _Distance __r = uniform_int_distribution<_Distance>(0, __k)(__g); + if (__r < __sz) + __output_iter[__r] = *__first; + } + return __output_iter + _VSTD::min(__n, __k); +} + +template +_LIBCPP_INLINE_VISIBILITY +_SampleIterator __sample(_PopulationIterator __first, + _PopulationIterator __last, _SampleIterator __output_iter, + _Distance __n, + _UniformRandomNumberGenerator& __g, + forward_iterator_tag) { + _Distance __unsampled_sz = _VSTD::distance(__first, __last); + for (__n = _VSTD::min(__n, __unsampled_sz); __n != 0; ++__first) { + _Distance __r = uniform_int_distribution<_Distance>(0, --__unsampled_sz)(__g); + if (__r < __n) { + *__output_iter++ = *__first; + --__n; + } + } + return __output_iter; +} + +template +_LIBCPP_INLINE_VISIBILITY +_SampleIterator __sample(_PopulationIterator __first, + _PopulationIterator __last, _SampleIterator __output_iter, + _Distance __n, _UniformRandomNumberGenerator& __g) { + typedef typename iterator_traits<_PopulationIterator>::iterator_category + _PopCategory; + typedef typename iterator_traits<_PopulationIterator>::difference_type + _Difference; + static_assert(__is_cpp17_forward_iterator<_PopulationIterator>::value || + __is_cpp17_random_access_iterator<_SampleIterator>::value, + "SampleIterator must meet the requirements of RandomAccessIterator"); + typedef typename common_type<_Distance, _Difference>::type _CommonType; + _LIBCPP_ASSERT(__n >= 0, "N must be a positive number."); + return _VSTD::__sample( + __first, __last, __output_iter, _CommonType(__n), + __g, _PopCategory()); +} + +#if _LIBCPP_STD_VER > 14 +template +inline _LIBCPP_INLINE_VISIBILITY +_SampleIterator sample(_PopulationIterator __first, + _PopulationIterator __last, _SampleIterator __output_iter, + _Distance __n, _UniformRandomNumberGenerator&& __g) { + return _VSTD::__sample(__first, __last, __output_iter, __n, __g); +} +#endif // _LIBCPP_STD_VER > 14 + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_SAMPLE_H diff --git a/lib/libcxx/include/__algorithm/search.h b/lib/libcxx/include/__algorithm/search.h new file mode 100644 index 000000000..008b8ebb0 --- /dev/null +++ b/lib/libcxx/include/__algorithm/search.h @@ -0,0 +1,131 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_SEARCH_H +#define _LIBCPP___ALGORITHM_SEARCH_H + +#include <__algorithm/comp.h> +#include <__config> +#include <__iterator/iterator_traits.h> +#include +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +pair<_ForwardIterator1, _ForwardIterator1> + _LIBCPP_CONSTEXPR_AFTER_CXX11 __search(_ForwardIterator1 __first1, _ForwardIterator1 __last1, + _ForwardIterator2 __first2, _ForwardIterator2 __last2, + _BinaryPredicate __pred, forward_iterator_tag, forward_iterator_tag) { + if (__first2 == __last2) + return _VSTD::make_pair(__first1, __first1); // Everything matches an empty sequence + while (true) { + // Find first element in sequence 1 that matchs *__first2, with a mininum of loop checks + while (true) { + if (__first1 == __last1) // return __last1 if no element matches *__first2 + return _VSTD::make_pair(__last1, __last1); + if (__pred(*__first1, *__first2)) + break; + ++__first1; + } + // *__first1 matches *__first2, now match elements after here + _ForwardIterator1 __m1 = __first1; + _ForwardIterator2 __m2 = __first2; + while (true) { + if (++__m2 == __last2) // If pattern exhausted, __first1 is the answer (works for 1 element pattern) + return _VSTD::make_pair(__first1, __m1); + if (++__m1 == __last1) // Otherwise if source exhaused, pattern not found + return _VSTD::make_pair(__last1, __last1); + if (!__pred(*__m1, *__m2)) // if there is a mismatch, restart with a new __first1 + { + ++__first1; + break; + } // else there is a match, check next elements + } + } +} + +template +_LIBCPP_CONSTEXPR_AFTER_CXX11 pair<_RandomAccessIterator1, _RandomAccessIterator1> +__search(_RandomAccessIterator1 __first1, _RandomAccessIterator1 __last1, _RandomAccessIterator2 __first2, + _RandomAccessIterator2 __last2, _BinaryPredicate __pred, random_access_iterator_tag, + random_access_iterator_tag) { + typedef typename iterator_traits<_RandomAccessIterator1>::difference_type _D1; + typedef typename iterator_traits<_RandomAccessIterator2>::difference_type _D2; + // Take advantage of knowing source and pattern lengths. Stop short when source is smaller than pattern + const _D2 __len2 = __last2 - __first2; + if (__len2 == 0) + return _VSTD::make_pair(__first1, __first1); + const _D1 __len1 = __last1 - __first1; + if (__len1 < __len2) + return _VSTD::make_pair(__last1, __last1); + const _RandomAccessIterator1 __s = __last1 - (__len2 - 1); // Start of pattern match can't go beyond here + + while (true) { + while (true) { + if (__first1 == __s) + return _VSTD::make_pair(__last1, __last1); + if (__pred(*__first1, *__first2)) + break; + ++__first1; + } + + _RandomAccessIterator1 __m1 = __first1; + _RandomAccessIterator2 __m2 = __first2; + while (true) { + if (++__m2 == __last2) + return _VSTD::make_pair(__first1, __first1 + __len2); + ++__m1; // no need to check range on __m1 because __s guarantees we have enough source + if (!__pred(*__m1, *__m2)) { + ++__first1; + break; + } + } + } +} + +template +_LIBCPP_NODISCARD_EXT inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 _ForwardIterator1 +search(_ForwardIterator1 __first1, _ForwardIterator1 __last1, _ForwardIterator2 __first2, _ForwardIterator2 __last2, + _BinaryPredicate __pred) { + return _VSTD::__search::type>( + __first1, __last1, __first2, __last2, __pred, + typename iterator_traits<_ForwardIterator1>::iterator_category(), + typename iterator_traits<_ForwardIterator2>::iterator_category()).first; +} + +template +_LIBCPP_NODISCARD_EXT inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 _ForwardIterator1 +search(_ForwardIterator1 __first1, _ForwardIterator1 __last1, _ForwardIterator2 __first2, _ForwardIterator2 __last2) { + typedef typename iterator_traits<_ForwardIterator1>::value_type __v1; + typedef typename iterator_traits<_ForwardIterator2>::value_type __v2; + return _VSTD::search(__first1, __last1, __first2, __last2, __equal_to<__v1, __v2>()); +} + +#if _LIBCPP_STD_VER > 14 +template +_LIBCPP_NODISCARD_EXT _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 _ForwardIterator +search(_ForwardIterator __f, _ForwardIterator __l, const _Searcher& __s) { + return __s(__f, __l).first; +} + +#endif + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_SEARCH_H diff --git a/lib/libcxx/include/__algorithm/search_n.h b/lib/libcxx/include/__algorithm/search_n.h new file mode 100644 index 000000000..1584e8e61 --- /dev/null +++ b/lib/libcxx/include/__algorithm/search_n.h @@ -0,0 +1,116 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_SEARCH_N_H +#define _LIBCPP___ALGORITHM_SEARCH_N_H + +#include <__config> +#include <__algorithm/comp.h> +#include <__iterator/iterator_traits.h> +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +_LIBCPP_CONSTEXPR_AFTER_CXX17 _ForwardIterator __search_n(_ForwardIterator __first, _ForwardIterator __last, + _Size __count, const _Tp& __value_, _BinaryPredicate __pred, + forward_iterator_tag) { + if (__count <= 0) + return __first; + while (true) { + // Find first element in sequence that matchs __value_, with a mininum of loop checks + while (true) { + if (__first == __last) // return __last if no element matches __value_ + return __last; + if (__pred(*__first, __value_)) + break; + ++__first; + } + // *__first matches __value_, now match elements after here + _ForwardIterator __m = __first; + _Size __c(0); + while (true) { + if (++__c == __count) // If pattern exhausted, __first is the answer (works for 1 element pattern) + return __first; + if (++__m == __last) // Otherwise if source exhaused, pattern not found + return __last; + if (!__pred(*__m, __value_)) // if there is a mismatch, restart with a new __first + { + __first = __m; + ++__first; + break; + } // else there is a match, check next elements + } + } +} + +template +_LIBCPP_CONSTEXPR_AFTER_CXX17 _RandomAccessIterator __search_n(_RandomAccessIterator __first, + _RandomAccessIterator __last, _Size __count, + const _Tp& __value_, _BinaryPredicate __pred, + random_access_iterator_tag) { + if (__count <= 0) + return __first; + _Size __len = static_cast<_Size>(__last - __first); + if (__len < __count) + return __last; + const _RandomAccessIterator __s = __last - (__count - 1); // Start of pattern match can't go beyond here + while (true) { + // Find first element in sequence that matchs __value_, with a mininum of loop checks + while (true) { + if (__first >= __s) // return __last if no element matches __value_ + return __last; + if (__pred(*__first, __value_)) + break; + ++__first; + } + // *__first matches __value_, now match elements after here + _RandomAccessIterator __m = __first; + _Size __c(0); + while (true) { + if (++__c == __count) // If pattern exhausted, __first is the answer (works for 1 element pattern) + return __first; + ++__m; // no need to check range on __m because __s guarantees we have enough source + if (!__pred(*__m, __value_)) // if there is a mismatch, restart with a new __first + { + __first = __m; + ++__first; + break; + } // else there is a match, check next elements + } + } +} + +template +_LIBCPP_NODISCARD_EXT inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 _ForwardIterator search_n( + _ForwardIterator __first, _ForwardIterator __last, _Size __count, const _Tp& __value_, _BinaryPredicate __pred) { + return _VSTD::__search_n::type>( + __first, __last, _VSTD::__convert_to_integral(__count), __value_, __pred, + typename iterator_traits<_ForwardIterator>::iterator_category()); +} + +template +_LIBCPP_NODISCARD_EXT inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 _ForwardIterator +search_n(_ForwardIterator __first, _ForwardIterator __last, _Size __count, const _Tp& __value_) { + typedef typename iterator_traits<_ForwardIterator>::value_type __v; + return _VSTD::search_n(__first, __last, _VSTD::__convert_to_integral(__count), __value_, __equal_to<__v, _Tp>()); +} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_SEARCH_N_H diff --git a/lib/libcxx/include/__algorithm/set_difference.h b/lib/libcxx/include/__algorithm/set_difference.h new file mode 100644 index 000000000..f4c985d97 --- /dev/null +++ b/lib/libcxx/include/__algorithm/set_difference.h @@ -0,0 +1,77 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_SET_DIFFERENCE_H +#define _LIBCPP___ALGORITHM_SET_DIFFERENCE_H + +#include <__config> +#include <__algorithm/comp.h> +#include <__algorithm/comp_ref_type.h> +#include <__algorithm/copy.h> +#include <__iterator/iterator_traits.h> + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +_LIBCPP_CONSTEXPR_AFTER_CXX17 _OutputIterator +__set_difference(_InputIterator1 __first1, _InputIterator1 __last1, + _InputIterator2 __first2, _InputIterator2 __last2, _OutputIterator __result, _Compare __comp) +{ + while (__first1 != __last1) + { + if (__first2 == __last2) + return _VSTD::copy(__first1, __last1, __result); + if (__comp(*__first1, *__first2)) + { + *__result = *__first1; + ++__result; + ++__first1; + } + else + { + if (!__comp(*__first2, *__first1)) + ++__first1; + ++__first2; + } + } + return __result; +} + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +_OutputIterator +set_difference(_InputIterator1 __first1, _InputIterator1 __last1, + _InputIterator2 __first2, _InputIterator2 __last2, _OutputIterator __result, _Compare __comp) +{ + typedef typename __comp_ref_type<_Compare>::type _Comp_ref; + return _VSTD::__set_difference<_Comp_ref>(__first1, __last1, __first2, __last2, __result, __comp); +} + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +_OutputIterator +set_difference(_InputIterator1 __first1, _InputIterator1 __last1, + _InputIterator2 __first2, _InputIterator2 __last2, _OutputIterator __result) +{ + return _VSTD::set_difference(__first1, __last1, __first2, __last2, __result, + __less::value_type, + typename iterator_traits<_InputIterator2>::value_type>()); +} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_SET_DIFFERENCE_H diff --git a/lib/libcxx/include/__algorithm/set_intersection.h b/lib/libcxx/include/__algorithm/set_intersection.h new file mode 100644 index 000000000..9d34b66c2 --- /dev/null +++ b/lib/libcxx/include/__algorithm/set_intersection.h @@ -0,0 +1,74 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_SET_INTERSECTION_H +#define _LIBCPP___ALGORITHM_SET_INTERSECTION_H + +#include <__config> +#include <__algorithm/comp.h> +#include <__algorithm/comp_ref_type.h> +#include <__iterator/iterator_traits.h> + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +_LIBCPP_CONSTEXPR_AFTER_CXX17 _OutputIterator +__set_intersection(_InputIterator1 __first1, _InputIterator1 __last1, + _InputIterator2 __first2, _InputIterator2 __last2, _OutputIterator __result, _Compare __comp) +{ + while (__first1 != __last1 && __first2 != __last2) + { + if (__comp(*__first1, *__first2)) + ++__first1; + else + { + if (!__comp(*__first2, *__first1)) + { + *__result = *__first1; + ++__result; + ++__first1; + } + ++__first2; + } + } + return __result; +} + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +_OutputIterator +set_intersection(_InputIterator1 __first1, _InputIterator1 __last1, + _InputIterator2 __first2, _InputIterator2 __last2, _OutputIterator __result, _Compare __comp) +{ + typedef typename __comp_ref_type<_Compare>::type _Comp_ref; + return _VSTD::__set_intersection<_Comp_ref>(__first1, __last1, __first2, __last2, __result, __comp); +} + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +_OutputIterator +set_intersection(_InputIterator1 __first1, _InputIterator1 __last1, + _InputIterator2 __first2, _InputIterator2 __last2, _OutputIterator __result) +{ + return _VSTD::set_intersection(__first1, __last1, __first2, __last2, __result, + __less::value_type, + typename iterator_traits<_InputIterator2>::value_type>()); +} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_SET_INTERSECTION_H diff --git a/lib/libcxx/include/__algorithm/set_symmetric_difference.h b/lib/libcxx/include/__algorithm/set_symmetric_difference.h new file mode 100644 index 000000000..5650b836a --- /dev/null +++ b/lib/libcxx/include/__algorithm/set_symmetric_difference.h @@ -0,0 +1,82 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_SET_SYMMETRIC_DIFFERENCE_H +#define _LIBCPP___ALGORITHM_SET_SYMMETRIC_DIFFERENCE_H + +#include <__config> +#include <__algorithm/comp.h> +#include <__algorithm/comp_ref_type.h> +#include <__algorithm/copy.h> +#include <__iterator/iterator_traits.h> + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +_LIBCPP_CONSTEXPR_AFTER_CXX17 _OutputIterator +__set_symmetric_difference(_InputIterator1 __first1, _InputIterator1 __last1, + _InputIterator2 __first2, _InputIterator2 __last2, _OutputIterator __result, _Compare __comp) +{ + while (__first1 != __last1) + { + if (__first2 == __last2) + return _VSTD::copy(__first1, __last1, __result); + if (__comp(*__first1, *__first2)) + { + *__result = *__first1; + ++__result; + ++__first1; + } + else + { + if (__comp(*__first2, *__first1)) + { + *__result = *__first2; + ++__result; + } + else + ++__first1; + ++__first2; + } + } + return _VSTD::copy(__first2, __last2, __result); +} + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +_OutputIterator +set_symmetric_difference(_InputIterator1 __first1, _InputIterator1 __last1, + _InputIterator2 __first2, _InputIterator2 __last2, _OutputIterator __result, _Compare __comp) +{ + typedef typename __comp_ref_type<_Compare>::type _Comp_ref; + return _VSTD::__set_symmetric_difference<_Comp_ref>(__first1, __last1, __first2, __last2, __result, __comp); +} + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +_OutputIterator +set_symmetric_difference(_InputIterator1 __first1, _InputIterator1 __last1, + _InputIterator2 __first2, _InputIterator2 __last2, _OutputIterator __result) +{ + return _VSTD::set_symmetric_difference(__first1, __last1, __first2, __last2, __result, + __less::value_type, + typename iterator_traits<_InputIterator2>::value_type>()); +} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_SET_SYMMETRIC_DIFFERENCE_H diff --git a/lib/libcxx/include/__algorithm/set_union.h b/lib/libcxx/include/__algorithm/set_union.h new file mode 100644 index 000000000..c0874e957 --- /dev/null +++ b/lib/libcxx/include/__algorithm/set_union.h @@ -0,0 +1,77 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_SET_UNION_H +#define _LIBCPP___ALGORITHM_SET_UNION_H + +#include <__config> +#include <__algorithm/comp.h> +#include <__algorithm/comp_ref_type.h> +#include <__algorithm/copy.h> +#include <__iterator/iterator_traits.h> + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +_LIBCPP_CONSTEXPR_AFTER_CXX17 _OutputIterator +__set_union(_InputIterator1 __first1, _InputIterator1 __last1, + _InputIterator2 __first2, _InputIterator2 __last2, _OutputIterator __result, _Compare __comp) +{ + for (; __first1 != __last1; ++__result) + { + if (__first2 == __last2) + return _VSTD::copy(__first1, __last1, __result); + if (__comp(*__first2, *__first1)) + { + *__result = *__first2; + ++__first2; + } + else + { + if (!__comp(*__first1, *__first2)) + ++__first2; + *__result = *__first1; + ++__first1; + } + } + return _VSTD::copy(__first2, __last2, __result); +} + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +_OutputIterator +set_union(_InputIterator1 __first1, _InputIterator1 __last1, + _InputIterator2 __first2, _InputIterator2 __last2, _OutputIterator __result, _Compare __comp) +{ + typedef typename __comp_ref_type<_Compare>::type _Comp_ref; + return _VSTD::__set_union<_Comp_ref>(__first1, __last1, __first2, __last2, __result, __comp); +} + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +_OutputIterator +set_union(_InputIterator1 __first1, _InputIterator1 __last1, + _InputIterator2 __first2, _InputIterator2 __last2, _OutputIterator __result) +{ + return _VSTD::set_union(__first1, __last1, __first2, __last2, __result, + __less::value_type, + typename iterator_traits<_InputIterator2>::value_type>()); +} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_SET_UNION_H diff --git a/lib/libcxx/include/__algorithm/shift_left.h b/lib/libcxx/include/__algorithm/shift_left.h new file mode 100644 index 000000000..961b89cb0 --- /dev/null +++ b/lib/libcxx/include/__algorithm/shift_left.h @@ -0,0 +1,61 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_SHIFT_LEFT_H +#define _LIBCPP___ALGORITHM_SHIFT_LEFT_H + +#include <__config> +#include <__algorithm/move.h> +#include <__iterator/iterator_traits.h> +#include // swap + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +#if _LIBCPP_STD_VER > 17 + +template +inline _LIBCPP_INLINE_VISIBILITY constexpr +_ForwardIterator +shift_left(_ForwardIterator __first, _ForwardIterator __last, + typename iterator_traits<_ForwardIterator>::difference_type __n) +{ + if (__n == 0) { + return __last; + } + + _ForwardIterator __m = __first; + if constexpr (__is_cpp17_random_access_iterator<_ForwardIterator>::value) { + if (__n >= __last - __first) { + return __first; + } + __m += __n; + } else { + for (; __n > 0; --__n) { + if (__m == __last) { + return __first; + } + ++__m; + } + } + return _VSTD::move(__m, __last, __first); +} + +#endif // _LIBCPP_STD_VER > 17 + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_SHIFT_LEFT_H diff --git a/lib/libcxx/include/__algorithm/shift_right.h b/lib/libcxx/include/__algorithm/shift_right.h new file mode 100644 index 000000000..5cb419506 --- /dev/null +++ b/lib/libcxx/include/__algorithm/shift_right.h @@ -0,0 +1,106 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_SHIFT_RIGHT_H +#define _LIBCPP___ALGORITHM_SHIFT_RIGHT_H + +#include <__config> +#include <__algorithm/move.h> +#include <__algorithm/move_backward.h> +#include <__algorithm/swap_ranges.h> +#include <__iterator/iterator_traits.h> +#include // swap + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +#if _LIBCPP_STD_VER > 17 + +template +inline _LIBCPP_INLINE_VISIBILITY constexpr +_ForwardIterator +shift_right(_ForwardIterator __first, _ForwardIterator __last, + typename iterator_traits<_ForwardIterator>::difference_type __n) +{ + if (__n == 0) { + return __first; + } + + if constexpr (__is_cpp17_random_access_iterator<_ForwardIterator>::value) { + decltype(__n) __d = __last - __first; + if (__n >= __d) { + return __last; + } + _ForwardIterator __m = __first + (__d - __n); + return _VSTD::move_backward(__first, __m, __last); + } else if constexpr (__is_cpp17_bidirectional_iterator<_ForwardIterator>::value) { + _ForwardIterator __m = __last; + for (; __n > 0; --__n) { + if (__m == __first) { + return __last; + } + --__m; + } + return _VSTD::move_backward(__first, __m, __last); + } else { + _ForwardIterator __ret = __first; + for (; __n > 0; --__n) { + if (__ret == __last) { + return __last; + } + ++__ret; + } + + // We have an __n-element scratch space from __first to __ret. + // Slide an __n-element window [__trail, __lead) from left to right. + // We're essentially doing swap_ranges(__first, __ret, __trail, __lead) + // over and over; but once __lead reaches __last we needn't bother + // to save the values of elements [__trail, __last). + + auto __trail = __first; + auto __lead = __ret; + while (__trail != __ret) { + if (__lead == __last) { + _VSTD::move(__first, __trail, __ret); + return __ret; + } + ++__trail; + ++__lead; + } + + _ForwardIterator __mid = __first; + while (true) { + if (__lead == __last) { + __trail = _VSTD::move(__mid, __ret, __trail); + _VSTD::move(__first, __mid, __trail); + return __ret; + } + swap(*__mid, *__trail); + ++__mid; + ++__trail; + ++__lead; + if (__mid == __ret) { + __mid = __first; + } + } + } +} + +#endif // _LIBCPP_STD_VER > 17 + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_SHIFT_RIGHT_H diff --git a/lib/libcxx/include/__algorithm/shuffle.h b/lib/libcxx/include/__algorithm/shuffle.h new file mode 100644 index 000000000..637fca538 --- /dev/null +++ b/lib/libcxx/include/__algorithm/shuffle.h @@ -0,0 +1,127 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_SHUFFLE_H +#define _LIBCPP___ALGORITHM_SHUFFLE_H + +#include <__config> +#include <__iterator/iterator_traits.h> +#include <__random/uniform_int_distribution.h> +#include <__utility/swap.h> +#include +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + + +#if _LIBCPP_STD_VER <= 14 || defined(_LIBCPP_ENABLE_CXX17_REMOVED_RANDOM_SHUFFLE) \ + || defined(_LIBCPP_BUILDING_LIBRARY) +class _LIBCPP_TYPE_VIS __rs_default; + +_LIBCPP_FUNC_VIS __rs_default __rs_get(); + +class _LIBCPP_TYPE_VIS __rs_default +{ + static unsigned __c_; + + __rs_default(); +public: + typedef uint_fast32_t result_type; + + static const result_type _Min = 0; + static const result_type _Max = 0xFFFFFFFF; + + __rs_default(const __rs_default&); + ~__rs_default(); + + result_type operator()(); + + static _LIBCPP_CONSTEXPR result_type min() {return _Min;} + static _LIBCPP_CONSTEXPR result_type max() {return _Max;} + + friend _LIBCPP_FUNC_VIS __rs_default __rs_get(); +}; + +_LIBCPP_FUNC_VIS __rs_default __rs_get(); + +template +_LIBCPP_DEPRECATED_IN_CXX14 void +random_shuffle(_RandomAccessIterator __first, _RandomAccessIterator __last) +{ + typedef typename iterator_traits<_RandomAccessIterator>::difference_type difference_type; + typedef uniform_int_distribution _Dp; + typedef typename _Dp::param_type _Pp; + difference_type __d = __last - __first; + if (__d > 1) + { + _Dp __uid; + __rs_default __g = __rs_get(); + for (--__last, (void) --__d; __first < __last; ++__first, (void) --__d) + { + difference_type __i = __uid(__g, _Pp(0, __d)); + if (__i != difference_type(0)) + swap(*__first, *(__first + __i)); + } + } +} + +template +_LIBCPP_DEPRECATED_IN_CXX14 void +random_shuffle(_RandomAccessIterator __first, _RandomAccessIterator __last, +#ifndef _LIBCPP_CXX03_LANG + _RandomNumberGenerator&& __rand) +#else + _RandomNumberGenerator& __rand) +#endif +{ + typedef typename iterator_traits<_RandomAccessIterator>::difference_type difference_type; + difference_type __d = __last - __first; + if (__d > 1) + { + for (--__last; __first < __last; ++__first, (void) --__d) + { + difference_type __i = __rand(__d); + if (__i != difference_type(0)) + swap(*__first, *(__first + __i)); + } + } +} +#endif + +template + void shuffle(_RandomAccessIterator __first, _RandomAccessIterator __last, + _UniformRandomNumberGenerator&& __g) +{ + typedef typename iterator_traits<_RandomAccessIterator>::difference_type difference_type; + typedef uniform_int_distribution _Dp; + typedef typename _Dp::param_type _Pp; + difference_type __d = __last - __first; + if (__d > 1) + { + _Dp __uid; + for (--__last, (void) --__d; __first < __last; ++__first, (void) --__d) + { + difference_type __i = __uid(__g, _Pp(0, __d)); + if (__i != difference_type(0)) + swap(*__first, *(__first + __i)); + } + } +} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_SHUFFLE_H diff --git a/lib/libcxx/include/__algorithm/sift_down.h b/lib/libcxx/include/__algorithm/sift_down.h new file mode 100644 index 000000000..dd4b54ed8 --- /dev/null +++ b/lib/libcxx/include/__algorithm/sift_down.h @@ -0,0 +1,84 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_SIFT_DOWN_H +#define _LIBCPP___ALGORITHM_SIFT_DOWN_H + +#include <__config> +#include <__iterator/iterator_traits.h> +#include <__utility/move.h> + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +_LIBCPP_CONSTEXPR_AFTER_CXX11 void +__sift_down(_RandomAccessIterator __first, _RandomAccessIterator /*__last*/, + _Compare __comp, + typename iterator_traits<_RandomAccessIterator>::difference_type __len, + _RandomAccessIterator __start) +{ + typedef typename iterator_traits<_RandomAccessIterator>::difference_type difference_type; + typedef typename iterator_traits<_RandomAccessIterator>::value_type value_type; + // left-child of __start is at 2 * __start + 1 + // right-child of __start is at 2 * __start + 2 + difference_type __child = __start - __first; + + if (__len < 2 || (__len - 2) / 2 < __child) + return; + + __child = 2 * __child + 1; + _RandomAccessIterator __child_i = __first + __child; + + if ((__child + 1) < __len && __comp(*__child_i, *(__child_i + 1))) { + // right-child exists and is greater than left-child + ++__child_i; + ++__child; + } + + // check if we are in heap-order + if (__comp(*__child_i, *__start)) + // we are, __start is larger than it's largest child + return; + + value_type __top(_VSTD::move(*__start)); + do + { + // we are not in heap-order, swap the parent with its largest child + *__start = _VSTD::move(*__child_i); + __start = __child_i; + + if ((__len - 2) / 2 < __child) + break; + + // recompute the child based off of the updated parent + __child = 2 * __child + 1; + __child_i = __first + __child; + + if ((__child + 1) < __len && __comp(*__child_i, *(__child_i + 1))) { + // right-child exists and is greater than left-child + ++__child_i; + ++__child; + } + + // check if we are in heap-order + } while (!__comp(*__child_i, __top)); + *__start = _VSTD::move(__top); +} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_SIFT_DOWN_H diff --git a/lib/libcxx/include/__algorithm/sort.h b/lib/libcxx/include/__algorithm/sort.h new file mode 100644 index 000000000..39ec21302 --- /dev/null +++ b/lib/libcxx/include/__algorithm/sort.h @@ -0,0 +1,530 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_SORT_H +#define _LIBCPP___ALGORITHM_SORT_H + +#include <__config> +#include <__algorithm/comp.h> +#include <__algorithm/comp_ref_type.h> +#include <__algorithm/min_element.h> +#include <__algorithm/partial_sort.h> +#include <__algorithm/unwrap_iter.h> +#include <__utility/swap.h> +#include +#include // swap + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +// stable, 2-3 compares, 0-2 swaps + +template +_LIBCPP_CONSTEXPR_AFTER_CXX11 unsigned +__sort3(_ForwardIterator __x, _ForwardIterator __y, _ForwardIterator __z, _Compare __c) +{ + unsigned __r = 0; + if (!__c(*__y, *__x)) // if x <= y + { + if (!__c(*__z, *__y)) // if y <= z + return __r; // x <= y && y <= z + // x <= y && y > z + swap(*__y, *__z); // x <= z && y < z + __r = 1; + if (__c(*__y, *__x)) // if x > y + { + swap(*__x, *__y); // x < y && y <= z + __r = 2; + } + return __r; // x <= y && y < z + } + if (__c(*__z, *__y)) // x > y, if y > z + { + swap(*__x, *__z); // x < y && y < z + __r = 1; + return __r; + } + swap(*__x, *__y); // x > y && y <= z + __r = 1; // x < y && x <= z + if (__c(*__z, *__y)) // if y > z + { + swap(*__y, *__z); // x <= y && y < z + __r = 2; + } + return __r; +} // x <= y && y <= z + +// stable, 3-6 compares, 0-5 swaps + +template +unsigned +__sort4(_ForwardIterator __x1, _ForwardIterator __x2, _ForwardIterator __x3, + _ForwardIterator __x4, _Compare __c) +{ + unsigned __r = _VSTD::__sort3<_Compare>(__x1, __x2, __x3, __c); + if (__c(*__x4, *__x3)) + { + swap(*__x3, *__x4); + ++__r; + if (__c(*__x3, *__x2)) + { + swap(*__x2, *__x3); + ++__r; + if (__c(*__x2, *__x1)) + { + swap(*__x1, *__x2); + ++__r; + } + } + } + return __r; +} + +// stable, 4-10 compares, 0-9 swaps + +template +_LIBCPP_HIDDEN +unsigned +__sort5(_ForwardIterator __x1, _ForwardIterator __x2, _ForwardIterator __x3, + _ForwardIterator __x4, _ForwardIterator __x5, _Compare __c) +{ + unsigned __r = _VSTD::__sort4<_Compare>(__x1, __x2, __x3, __x4, __c); + if (__c(*__x5, *__x4)) + { + swap(*__x4, *__x5); + ++__r; + if (__c(*__x4, *__x3)) + { + swap(*__x3, *__x4); + ++__r; + if (__c(*__x3, *__x2)) + { + swap(*__x2, *__x3); + ++__r; + if (__c(*__x2, *__x1)) + { + swap(*__x1, *__x2); + ++__r; + } + } + } + } + return __r; +} + +// Assumes size > 0 +template +_LIBCPP_CONSTEXPR_AFTER_CXX11 void +__selection_sort(_BidirectionalIterator __first, _BidirectionalIterator __last, _Compare __comp) +{ + _BidirectionalIterator __lm1 = __last; + for (--__lm1; __first != __lm1; ++__first) + { + _BidirectionalIterator __i = _VSTD::min_element<_BidirectionalIterator, + typename add_lvalue_reference<_Compare>::type> + (__first, __last, __comp); + if (__i != __first) + swap(*__first, *__i); + } +} + +template +void +__insertion_sort(_BidirectionalIterator __first, _BidirectionalIterator __last, _Compare __comp) +{ + typedef typename iterator_traits<_BidirectionalIterator>::value_type value_type; + if (__first != __last) + { + _BidirectionalIterator __i = __first; + for (++__i; __i != __last; ++__i) + { + _BidirectionalIterator __j = __i; + value_type __t(_VSTD::move(*__j)); + for (_BidirectionalIterator __k = __i; __k != __first && __comp(__t, *--__k); --__j) + *__j = _VSTD::move(*__k); + *__j = _VSTD::move(__t); + } + } +} + +template +void +__insertion_sort_3(_RandomAccessIterator __first, _RandomAccessIterator __last, _Compare __comp) +{ + typedef typename iterator_traits<_RandomAccessIterator>::value_type value_type; + _RandomAccessIterator __j = __first+2; + _VSTD::__sort3<_Compare>(__first, __first+1, __j, __comp); + for (_RandomAccessIterator __i = __j+1; __i != __last; ++__i) + { + if (__comp(*__i, *__j)) + { + value_type __t(_VSTD::move(*__i)); + _RandomAccessIterator __k = __j; + __j = __i; + do + { + *__j = _VSTD::move(*__k); + __j = __k; + } while (__j != __first && __comp(__t, *--__k)); + *__j = _VSTD::move(__t); + } + __j = __i; + } +} + +template +bool +__insertion_sort_incomplete(_RandomAccessIterator __first, _RandomAccessIterator __last, _Compare __comp) +{ + switch (__last - __first) + { + case 0: + case 1: + return true; + case 2: + if (__comp(*--__last, *__first)) + swap(*__first, *__last); + return true; + case 3: + _VSTD::__sort3<_Compare>(__first, __first+1, --__last, __comp); + return true; + case 4: + _VSTD::__sort4<_Compare>(__first, __first+1, __first+2, --__last, __comp); + return true; + case 5: + _VSTD::__sort5<_Compare>(__first, __first+1, __first+2, __first+3, --__last, __comp); + return true; + } + typedef typename iterator_traits<_RandomAccessIterator>::value_type value_type; + _RandomAccessIterator __j = __first+2; + _VSTD::__sort3<_Compare>(__first, __first+1, __j, __comp); + const unsigned __limit = 8; + unsigned __count = 0; + for (_RandomAccessIterator __i = __j+1; __i != __last; ++__i) + { + if (__comp(*__i, *__j)) + { + value_type __t(_VSTD::move(*__i)); + _RandomAccessIterator __k = __j; + __j = __i; + do + { + *__j = _VSTD::move(*__k); + __j = __k; + } while (__j != __first && __comp(__t, *--__k)); + *__j = _VSTD::move(__t); + if (++__count == __limit) + return ++__i == __last; + } + __j = __i; + } + return true; +} + +template +void +__insertion_sort_move(_BidirectionalIterator __first1, _BidirectionalIterator __last1, + typename iterator_traits<_BidirectionalIterator>::value_type* __first2, _Compare __comp) +{ + typedef typename iterator_traits<_BidirectionalIterator>::value_type value_type; + if (__first1 != __last1) + { + __destruct_n __d(0); + unique_ptr __h(__first2, __d); + value_type* __last2 = __first2; + ::new ((void*)__last2) value_type(_VSTD::move(*__first1)); + __d.template __incr(); + for (++__last2; ++__first1 != __last1; ++__last2) + { + value_type* __j2 = __last2; + value_type* __i2 = __j2; + if (__comp(*__first1, *--__i2)) + { + ::new ((void*)__j2) value_type(_VSTD::move(*__i2)); + __d.template __incr(); + for (--__j2; __i2 != __first2 && __comp(*__first1, *--__i2); --__j2) + *__j2 = _VSTD::move(*__i2); + *__j2 = _VSTD::move(*__first1); + } + else + { + ::new ((void*)__j2) value_type(_VSTD::move(*__first1)); + __d.template __incr(); + } + } + __h.release(); + } +} + +template +void +__sort(_RandomAccessIterator __first, _RandomAccessIterator __last, _Compare __comp) +{ + typedef typename iterator_traits<_RandomAccessIterator>::difference_type difference_type; + typedef typename iterator_traits<_RandomAccessIterator>::value_type value_type; + const difference_type __limit = is_trivially_copy_constructible::value && + is_trivially_copy_assignable::value ? 30 : 6; + while (true) + { + __restart: + difference_type __len = __last - __first; + switch (__len) + { + case 0: + case 1: + return; + case 2: + if (__comp(*--__last, *__first)) + swap(*__first, *__last); + return; + case 3: + _VSTD::__sort3<_Compare>(__first, __first+1, --__last, __comp); + return; + case 4: + _VSTD::__sort4<_Compare>(__first, __first+1, __first+2, --__last, __comp); + return; + case 5: + _VSTD::__sort5<_Compare>(__first, __first+1, __first+2, __first+3, --__last, __comp); + return; + } + if (__len <= __limit) + { + _VSTD::__insertion_sort_3<_Compare>(__first, __last, __comp); + return; + } + // __len > 5 + _RandomAccessIterator __m = __first; + _RandomAccessIterator __lm1 = __last; + --__lm1; + unsigned __n_swaps; + { + difference_type __delta; + if (__len >= 1000) + { + __delta = __len/2; + __m += __delta; + __delta /= 2; + __n_swaps = _VSTD::__sort5<_Compare>(__first, __first + __delta, __m, __m+__delta, __lm1, __comp); + } + else + { + __delta = __len/2; + __m += __delta; + __n_swaps = _VSTD::__sort3<_Compare>(__first, __m, __lm1, __comp); + } + } + // *__m is median + // partition [__first, __m) < *__m and *__m <= [__m, __last) + // (this inhibits tossing elements equivalent to __m around unnecessarily) + _RandomAccessIterator __i = __first; + _RandomAccessIterator __j = __lm1; + // j points beyond range to be tested, *__m is known to be <= *__lm1 + // The search going up is known to be guarded but the search coming down isn't. + // Prime the downward search with a guard. + if (!__comp(*__i, *__m)) // if *__first == *__m + { + // *__first == *__m, *__first doesn't go in first part + // manually guard downward moving __j against __i + while (true) + { + if (__i == --__j) + { + // *__first == *__m, *__m <= all other elements + // Parition instead into [__first, __i) == *__first and *__first < [__i, __last) + ++__i; // __first + 1 + __j = __last; + if (!__comp(*__first, *--__j)) // we need a guard if *__first == *(__last-1) + { + while (true) + { + if (__i == __j) + return; // [__first, __last) all equivalent elements + if (__comp(*__first, *__i)) + { + swap(*__i, *__j); + ++__n_swaps; + ++__i; + break; + } + ++__i; + } + } + // [__first, __i) == *__first and *__first < [__j, __last) and __j == __last - 1 + if (__i == __j) + return; + while (true) + { + while (!__comp(*__first, *__i)) + ++__i; + while (__comp(*__first, *--__j)) + ; + if (__i >= __j) + break; + swap(*__i, *__j); + ++__n_swaps; + ++__i; + } + // [__first, __i) == *__first and *__first < [__i, __last) + // The first part is sorted, sort the second part + // _VSTD::__sort<_Compare>(__i, __last, __comp); + __first = __i; + goto __restart; + } + if (__comp(*__j, *__m)) + { + swap(*__i, *__j); + ++__n_swaps; + break; // found guard for downward moving __j, now use unguarded partition + } + } + } + // It is known that *__i < *__m + ++__i; + // j points beyond range to be tested, *__m is known to be <= *__lm1 + // if not yet partitioned... + if (__i < __j) + { + // known that *(__i - 1) < *__m + // known that __i <= __m + while (true) + { + // __m still guards upward moving __i + while (__comp(*__i, *__m)) + ++__i; + // It is now known that a guard exists for downward moving __j + while (!__comp(*--__j, *__m)) + ; + if (__i > __j) + break; + swap(*__i, *__j); + ++__n_swaps; + // It is known that __m != __j + // If __m just moved, follow it + if (__m == __i) + __m = __j; + ++__i; + } + } + // [__first, __i) < *__m and *__m <= [__i, __last) + if (__i != __m && __comp(*__m, *__i)) + { + swap(*__i, *__m); + ++__n_swaps; + } + // [__first, __i) < *__i and *__i <= [__i+1, __last) + // If we were given a perfect partition, see if insertion sort is quick... + if (__n_swaps == 0) + { + bool __fs = _VSTD::__insertion_sort_incomplete<_Compare>(__first, __i, __comp); + if (_VSTD::__insertion_sort_incomplete<_Compare>(__i+1, __last, __comp)) + { + if (__fs) + return; + __last = __i; + continue; + } + else + { + if (__fs) + { + __first = ++__i; + continue; + } + } + } + // sort smaller range with recursive call and larger with tail recursion elimination + if (__i - __first < __last - __i) + { + _VSTD::__sort<_Compare>(__first, __i, __comp); + // _VSTD::__sort<_Compare>(__i+1, __last, __comp); + __first = ++__i; + } + else + { + _VSTD::__sort<_Compare>(__i+1, __last, __comp); + // _VSTD::__sort<_Compare>(__first, __i, __comp); + __last = __i; + } + } +} + +template +inline _LIBCPP_INLINE_VISIBILITY +void +__sort(_Tp** __first, _Tp** __last, __less<_Tp*>&) +{ + __less __comp; + _VSTD::__sort<__less&, uintptr_t*>((uintptr_t*)__first, (uintptr_t*)__last, __comp); +} + +_LIBCPP_EXTERN_TEMPLATE(_LIBCPP_FUNC_VIS void __sort<__less&, char*>(char*, char*, __less&)) +_LIBCPP_EXTERN_TEMPLATE(_LIBCPP_FUNC_VIS void __sort<__less&, wchar_t*>(wchar_t*, wchar_t*, __less&)) +_LIBCPP_EXTERN_TEMPLATE(_LIBCPP_FUNC_VIS void __sort<__less&, signed char*>(signed char*, signed char*, __less&)) +_LIBCPP_EXTERN_TEMPLATE(_LIBCPP_FUNC_VIS void __sort<__less&, unsigned char*>(unsigned char*, unsigned char*, __less&)) +_LIBCPP_EXTERN_TEMPLATE(_LIBCPP_FUNC_VIS void __sort<__less&, short*>(short*, short*, __less&)) +_LIBCPP_EXTERN_TEMPLATE(_LIBCPP_FUNC_VIS void __sort<__less&, unsigned short*>(unsigned short*, unsigned short*, __less&)) +_LIBCPP_EXTERN_TEMPLATE(_LIBCPP_FUNC_VIS void __sort<__less&, int*>(int*, int*, __less&)) +_LIBCPP_EXTERN_TEMPLATE(_LIBCPP_FUNC_VIS void __sort<__less&, unsigned*>(unsigned*, unsigned*, __less&)) +_LIBCPP_EXTERN_TEMPLATE(_LIBCPP_FUNC_VIS void __sort<__less&, long*>(long*, long*, __less&)) +_LIBCPP_EXTERN_TEMPLATE(_LIBCPP_FUNC_VIS void __sort<__less&, unsigned long*>(unsigned long*, unsigned long*, __less&)) +_LIBCPP_EXTERN_TEMPLATE(_LIBCPP_FUNC_VIS void __sort<__less&, long long*>(long long*, long long*, __less&)) +_LIBCPP_EXTERN_TEMPLATE(_LIBCPP_FUNC_VIS void __sort<__less&, unsigned long long*>(unsigned long long*, unsigned long long*, __less&)) +_LIBCPP_EXTERN_TEMPLATE(_LIBCPP_FUNC_VIS void __sort<__less&, float*>(float*, float*, __less&)) +_LIBCPP_EXTERN_TEMPLATE(_LIBCPP_FUNC_VIS void __sort<__less&, double*>(double*, double*, __less&)) +_LIBCPP_EXTERN_TEMPLATE(_LIBCPP_FUNC_VIS void __sort<__less&, long double*>(long double*, long double*, __less&)) + +_LIBCPP_EXTERN_TEMPLATE(_LIBCPP_FUNC_VIS bool __insertion_sort_incomplete<__less&, char*>(char*, char*, __less&)) +_LIBCPP_EXTERN_TEMPLATE(_LIBCPP_FUNC_VIS bool __insertion_sort_incomplete<__less&, wchar_t*>(wchar_t*, wchar_t*, __less&)) +_LIBCPP_EXTERN_TEMPLATE(_LIBCPP_FUNC_VIS bool __insertion_sort_incomplete<__less&, signed char*>(signed char*, signed char*, __less&)) +_LIBCPP_EXTERN_TEMPLATE(_LIBCPP_FUNC_VIS bool __insertion_sort_incomplete<__less&, unsigned char*>(unsigned char*, unsigned char*, __less&)) +_LIBCPP_EXTERN_TEMPLATE(_LIBCPP_FUNC_VIS bool __insertion_sort_incomplete<__less&, short*>(short*, short*, __less&)) +_LIBCPP_EXTERN_TEMPLATE(_LIBCPP_FUNC_VIS bool __insertion_sort_incomplete<__less&, unsigned short*>(unsigned short*, unsigned short*, __less&)) +_LIBCPP_EXTERN_TEMPLATE(_LIBCPP_FUNC_VIS bool __insertion_sort_incomplete<__less&, int*>(int*, int*, __less&)) +_LIBCPP_EXTERN_TEMPLATE(_LIBCPP_FUNC_VIS bool __insertion_sort_incomplete<__less&, unsigned*>(unsigned*, unsigned*, __less&)) +_LIBCPP_EXTERN_TEMPLATE(_LIBCPP_FUNC_VIS bool __insertion_sort_incomplete<__less&, long*>(long*, long*, __less&)) +_LIBCPP_EXTERN_TEMPLATE(_LIBCPP_FUNC_VIS bool __insertion_sort_incomplete<__less&, unsigned long*>(unsigned long*, unsigned long*, __less&)) +_LIBCPP_EXTERN_TEMPLATE(_LIBCPP_FUNC_VIS bool __insertion_sort_incomplete<__less&, long long*>(long long*, long long*, __less&)) +_LIBCPP_EXTERN_TEMPLATE(_LIBCPP_FUNC_VIS bool __insertion_sort_incomplete<__less&, unsigned long long*>(unsigned long long*, unsigned long long*, __less&)) +_LIBCPP_EXTERN_TEMPLATE(_LIBCPP_FUNC_VIS bool __insertion_sort_incomplete<__less&, float*>(float*, float*, __less&)) +_LIBCPP_EXTERN_TEMPLATE(_LIBCPP_FUNC_VIS bool __insertion_sort_incomplete<__less&, double*>(double*, double*, __less&)) +_LIBCPP_EXTERN_TEMPLATE(_LIBCPP_FUNC_VIS bool __insertion_sort_incomplete<__less&, long double*>(long double*, long double*, __less&)) + +_LIBCPP_EXTERN_TEMPLATE(_LIBCPP_FUNC_VIS unsigned __sort5<__less&, long double*>(long double*, long double*, long double*, long double*, long double*, __less&)) + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +void +sort(_RandomAccessIterator __first, _RandomAccessIterator __last, _Compare __comp) +{ + typedef typename __comp_ref_type<_Compare>::type _Comp_ref; + if (__libcpp_is_constant_evaluated()) { + _VSTD::__partial_sort<_Comp_ref>(__first, __last, __last, _Comp_ref(__comp)); + } else { + _VSTD::__sort<_Comp_ref>(_VSTD::__unwrap_iter(__first), _VSTD::__unwrap_iter(__last), _Comp_ref(__comp)); + } +} + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +void +sort(_RandomAccessIterator __first, _RandomAccessIterator __last) +{ + _VSTD::sort(__first, __last, __less::value_type>()); +} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_SORT_H diff --git a/lib/libcxx/include/__algorithm/sort_heap.h b/lib/libcxx/include/__algorithm/sort_heap.h new file mode 100644 index 000000000..aa8ef762a --- /dev/null +++ b/lib/libcxx/include/__algorithm/sort_heap.h @@ -0,0 +1,58 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_SORT_HEAP_H +#define _LIBCPP___ALGORITHM_SORT_HEAP_H + +#include <__config> +#include <__algorithm/comp.h> +#include <__algorithm/comp_ref_type.h> +#include <__algorithm/pop_heap.h> +#include <__iterator/iterator_traits.h> +#include // swap + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +_LIBCPP_CONSTEXPR_AFTER_CXX17 void +__sort_heap(_RandomAccessIterator __first, _RandomAccessIterator __last, _Compare __comp) +{ + typedef typename iterator_traits<_RandomAccessIterator>::difference_type difference_type; + for (difference_type __n = __last - __first; __n > 1; --__last, (void) --__n) + _VSTD::__pop_heap<_Compare>(__first, __last, __comp, __n); +} + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +void +sort_heap(_RandomAccessIterator __first, _RandomAccessIterator __last, _Compare __comp) +{ + typedef typename __comp_ref_type<_Compare>::type _Comp_ref; + _VSTD::__sort_heap<_Comp_ref>(__first, __last, __comp); +} + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +void +sort_heap(_RandomAccessIterator __first, _RandomAccessIterator __last) +{ + _VSTD::sort_heap(__first, __last, __less::value_type>()); +} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_SORT_HEAP_H diff --git a/lib/libcxx/include/__algorithm/stable_partition.h b/lib/libcxx/include/__algorithm/stable_partition.h new file mode 100644 index 000000000..931335f44 --- /dev/null +++ b/lib/libcxx/include/__algorithm/stable_partition.h @@ -0,0 +1,305 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_STABLE_PARTITION_H +#define _LIBCPP___ALGORITHM_STABLE_PARTITION_H + +#include <__config> +#include <__algorithm/rotate.h> +#include <__iterator/iterator_traits.h> +#include <__utility/swap.h> +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +_ForwardIterator +__stable_partition(_ForwardIterator __first, _ForwardIterator __last, _Predicate __pred, + _Distance __len, _Pair __p, forward_iterator_tag __fit) +{ + // *__first is known to be false + // __len >= 1 + if (__len == 1) + return __first; + if (__len == 2) + { + _ForwardIterator __m = __first; + if (__pred(*++__m)) + { + swap(*__first, *__m); + return __m; + } + return __first; + } + if (__len <= __p.second) + { // The buffer is big enough to use + typedef typename iterator_traits<_ForwardIterator>::value_type value_type; + __destruct_n __d(0); + unique_ptr __h(__p.first, __d); + // Move the falses into the temporary buffer, and the trues to the front of the line + // Update __first to always point to the end of the trues + value_type* __t = __p.first; + ::new ((void*)__t) value_type(_VSTD::move(*__first)); + __d.template __incr(); + ++__t; + _ForwardIterator __i = __first; + while (++__i != __last) + { + if (__pred(*__i)) + { + *__first = _VSTD::move(*__i); + ++__first; + } + else + { + ::new ((void*)__t) value_type(_VSTD::move(*__i)); + __d.template __incr(); + ++__t; + } + } + // All trues now at start of range, all falses in buffer + // Move falses back into range, but don't mess up __first which points to first false + __i = __first; + for (value_type* __t2 = __p.first; __t2 < __t; ++__t2, (void) ++__i) + *__i = _VSTD::move(*__t2); + // __h destructs moved-from values out of the temp buffer, but doesn't deallocate buffer + return __first; + } + // Else not enough buffer, do in place + // __len >= 3 + _ForwardIterator __m = __first; + _Distance __len2 = __len / 2; // __len2 >= 2 + _VSTD::advance(__m, __len2); + // recurse on [__first, __m), *__first know to be false + // F????????????????? + // f m l + typedef typename add_lvalue_reference<_Predicate>::type _PredRef; + _ForwardIterator __first_false = _VSTD::__stable_partition<_PredRef>(__first, __m, __pred, __len2, __p, __fit); + // TTTFFFFF?????????? + // f ff m l + // recurse on [__m, __last], except increase __m until *(__m) is false, *__last know to be true + _ForwardIterator __m1 = __m; + _ForwardIterator __second_false = __last; + _Distance __len_half = __len - __len2; + while (__pred(*__m1)) + { + if (++__m1 == __last) + goto __second_half_done; + --__len_half; + } + // TTTFFFFFTTTF?????? + // f ff m m1 l + __second_false = _VSTD::__stable_partition<_PredRef>(__m1, __last, __pred, __len_half, __p, __fit); +__second_half_done: + // TTTFFFFFTTTTTFFFFF + // f ff m sf l + return _VSTD::rotate(__first_false, __m, __second_false); + // TTTTTTTTFFFFFFFFFF + // | +} + +template +_ForwardIterator +__stable_partition(_ForwardIterator __first, _ForwardIterator __last, _Predicate __pred, + forward_iterator_tag) +{ + const unsigned __alloc_limit = 3; // might want to make this a function of trivial assignment + // Either prove all true and return __first or point to first false + while (true) + { + if (__first == __last) + return __first; + if (!__pred(*__first)) + break; + ++__first; + } + // We now have a reduced range [__first, __last) + // *__first is known to be false + typedef typename iterator_traits<_ForwardIterator>::difference_type difference_type; + typedef typename iterator_traits<_ForwardIterator>::value_type value_type; + difference_type __len = _VSTD::distance(__first, __last); + pair __p(0, 0); + unique_ptr __h; + if (__len >= __alloc_limit) + { + __p = _VSTD::get_temporary_buffer(__len); + __h.reset(__p.first); + } + return _VSTD::__stable_partition::type> + (__first, __last, __pred, __len, __p, forward_iterator_tag()); +} + +template +_BidirectionalIterator +__stable_partition(_BidirectionalIterator __first, _BidirectionalIterator __last, _Predicate __pred, + _Distance __len, _Pair __p, bidirectional_iterator_tag __bit) +{ + // *__first is known to be false + // *__last is known to be true + // __len >= 2 + if (__len == 2) + { + swap(*__first, *__last); + return __last; + } + if (__len == 3) + { + _BidirectionalIterator __m = __first; + if (__pred(*++__m)) + { + swap(*__first, *__m); + swap(*__m, *__last); + return __last; + } + swap(*__m, *__last); + swap(*__first, *__m); + return __m; + } + if (__len <= __p.second) + { // The buffer is big enough to use + typedef typename iterator_traits<_BidirectionalIterator>::value_type value_type; + __destruct_n __d(0); + unique_ptr __h(__p.first, __d); + // Move the falses into the temporary buffer, and the trues to the front of the line + // Update __first to always point to the end of the trues + value_type* __t = __p.first; + ::new ((void*)__t) value_type(_VSTD::move(*__first)); + __d.template __incr(); + ++__t; + _BidirectionalIterator __i = __first; + while (++__i != __last) + { + if (__pred(*__i)) + { + *__first = _VSTD::move(*__i); + ++__first; + } + else + { + ::new ((void*)__t) value_type(_VSTD::move(*__i)); + __d.template __incr(); + ++__t; + } + } + // move *__last, known to be true + *__first = _VSTD::move(*__i); + __i = ++__first; + // All trues now at start of range, all falses in buffer + // Move falses back into range, but don't mess up __first which points to first false + for (value_type* __t2 = __p.first; __t2 < __t; ++__t2, (void) ++__i) + *__i = _VSTD::move(*__t2); + // __h destructs moved-from values out of the temp buffer, but doesn't deallocate buffer + return __first; + } + // Else not enough buffer, do in place + // __len >= 4 + _BidirectionalIterator __m = __first; + _Distance __len2 = __len / 2; // __len2 >= 2 + _VSTD::advance(__m, __len2); + // recurse on [__first, __m-1], except reduce __m-1 until *(__m-1) is true, *__first know to be false + // F????????????????T + // f m l + _BidirectionalIterator __m1 = __m; + _BidirectionalIterator __first_false = __first; + _Distance __len_half = __len2; + while (!__pred(*--__m1)) + { + if (__m1 == __first) + goto __first_half_done; + --__len_half; + } + // F???TFFF?????????T + // f m1 m l + typedef typename add_lvalue_reference<_Predicate>::type _PredRef; + __first_false = _VSTD::__stable_partition<_PredRef>(__first, __m1, __pred, __len_half, __p, __bit); +__first_half_done: + // TTTFFFFF?????????T + // f ff m l + // recurse on [__m, __last], except increase __m until *(__m) is false, *__last know to be true + __m1 = __m; + _BidirectionalIterator __second_false = __last; + ++__second_false; + __len_half = __len - __len2; + while (__pred(*__m1)) + { + if (++__m1 == __last) + goto __second_half_done; + --__len_half; + } + // TTTFFFFFTTTF?????T + // f ff m m1 l + __second_false = _VSTD::__stable_partition<_PredRef>(__m1, __last, __pred, __len_half, __p, __bit); +__second_half_done: + // TTTFFFFFTTTTTFFFFF + // f ff m sf l + return _VSTD::rotate(__first_false, __m, __second_false); + // TTTTTTTTFFFFFFFFFF + // | +} + +template +_BidirectionalIterator +__stable_partition(_BidirectionalIterator __first, _BidirectionalIterator __last, _Predicate __pred, + bidirectional_iterator_tag) +{ + typedef typename iterator_traits<_BidirectionalIterator>::difference_type difference_type; + typedef typename iterator_traits<_BidirectionalIterator>::value_type value_type; + const difference_type __alloc_limit = 4; // might want to make this a function of trivial assignment + // Either prove all true and return __first or point to first false + while (true) + { + if (__first == __last) + return __first; + if (!__pred(*__first)) + break; + ++__first; + } + // __first points to first false, everything prior to __first is already set. + // Either prove [__first, __last) is all false and return __first, or point __last to last true + do + { + if (__first == --__last) + return __first; + } while (!__pred(*__last)); + // We now have a reduced range [__first, __last] + // *__first is known to be false + // *__last is known to be true + // __len >= 2 + difference_type __len = _VSTD::distance(__first, __last) + 1; + pair __p(0, 0); + unique_ptr __h; + if (__len >= __alloc_limit) + { + __p = _VSTD::get_temporary_buffer(__len); + __h.reset(__p.first); + } + return _VSTD::__stable_partition::type> + (__first, __last, __pred, __len, __p, bidirectional_iterator_tag()); +} + +template +inline _LIBCPP_INLINE_VISIBILITY +_ForwardIterator +stable_partition(_ForwardIterator __first, _ForwardIterator __last, _Predicate __pred) +{ + return _VSTD::__stable_partition::type> + (__first, __last, __pred, typename iterator_traits<_ForwardIterator>::iterator_category()); +} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_STABLE_PARTITION_H diff --git a/lib/libcxx/include/__algorithm/stable_sort.h b/lib/libcxx/include/__algorithm/stable_sort.h new file mode 100644 index 000000000..32b239a0d --- /dev/null +++ b/lib/libcxx/include/__algorithm/stable_sort.h @@ -0,0 +1,235 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_STABLE_SORT_H +#define _LIBCPP___ALGORITHM_STABLE_SORT_H + +#include <__config> +#include <__algorithm/comp.h> +#include <__algorithm/comp_ref_type.h> +#include <__algorithm/inplace_merge.h> +#include <__algorithm/sort.h> +#include <__iterator/iterator_traits.h> +#include <__utility/swap.h> +#include +#include // swap + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +void +__merge_move_construct(_InputIterator1 __first1, _InputIterator1 __last1, + _InputIterator2 __first2, _InputIterator2 __last2, + typename iterator_traits<_InputIterator1>::value_type* __result, _Compare __comp) +{ + typedef typename iterator_traits<_InputIterator1>::value_type value_type; + __destruct_n __d(0); + unique_ptr __h(__result, __d); + for (; true; ++__result) + { + if (__first1 == __last1) + { + for (; __first2 != __last2; ++__first2, ++__result, (void)__d.template __incr()) + ::new ((void*)__result) value_type(_VSTD::move(*__first2)); + __h.release(); + return; + } + if (__first2 == __last2) + { + for (; __first1 != __last1; ++__first1, ++__result, (void)__d.template __incr()) + ::new ((void*)__result) value_type(_VSTD::move(*__first1)); + __h.release(); + return; + } + if (__comp(*__first2, *__first1)) + { + ::new ((void*)__result) value_type(_VSTD::move(*__first2)); + __d.template __incr(); + ++__first2; + } + else + { + ::new ((void*)__result) value_type(_VSTD::move(*__first1)); + __d.template __incr(); + ++__first1; + } + } +} + +template +void +__merge_move_assign(_InputIterator1 __first1, _InputIterator1 __last1, + _InputIterator2 __first2, _InputIterator2 __last2, + _OutputIterator __result, _Compare __comp) +{ + for (; __first1 != __last1; ++__result) + { + if (__first2 == __last2) + { + for (; __first1 != __last1; ++__first1, (void) ++__result) + *__result = _VSTD::move(*__first1); + return; + } + if (__comp(*__first2, *__first1)) + { + *__result = _VSTD::move(*__first2); + ++__first2; + } + else + { + *__result = _VSTD::move(*__first1); + ++__first1; + } + } + for (; __first2 != __last2; ++__first2, (void) ++__result) + *__result = _VSTD::move(*__first2); +} + +template +void +__stable_sort(_RandomAccessIterator __first, _RandomAccessIterator __last, _Compare __comp, + typename iterator_traits<_RandomAccessIterator>::difference_type __len, + typename iterator_traits<_RandomAccessIterator>::value_type* __buff, ptrdiff_t __buff_size); + +template +void +__stable_sort_move(_RandomAccessIterator __first1, _RandomAccessIterator __last1, _Compare __comp, + typename iterator_traits<_RandomAccessIterator>::difference_type __len, + typename iterator_traits<_RandomAccessIterator>::value_type* __first2) +{ + typedef typename iterator_traits<_RandomAccessIterator>::value_type value_type; + switch (__len) + { + case 0: + return; + case 1: + ::new ((void*)__first2) value_type(_VSTD::move(*__first1)); + return; + case 2: + __destruct_n __d(0); + unique_ptr __h2(__first2, __d); + if (__comp(*--__last1, *__first1)) + { + ::new ((void*)__first2) value_type(_VSTD::move(*__last1)); + __d.template __incr(); + ++__first2; + ::new ((void*)__first2) value_type(_VSTD::move(*__first1)); + } + else + { + ::new ((void*)__first2) value_type(_VSTD::move(*__first1)); + __d.template __incr(); + ++__first2; + ::new ((void*)__first2) value_type(_VSTD::move(*__last1)); + } + __h2.release(); + return; + } + if (__len <= 8) + { + _VSTD::__insertion_sort_move<_Compare>(__first1, __last1, __first2, __comp); + return; + } + typename iterator_traits<_RandomAccessIterator>::difference_type __l2 = __len / 2; + _RandomAccessIterator __m = __first1 + __l2; + _VSTD::__stable_sort<_Compare>(__first1, __m, __comp, __l2, __first2, __l2); + _VSTD::__stable_sort<_Compare>(__m, __last1, __comp, __len - __l2, __first2 + __l2, __len - __l2); + _VSTD::__merge_move_construct<_Compare>(__first1, __m, __m, __last1, __first2, __comp); +} + +template +struct __stable_sort_switch +{ + static const unsigned value = 128*is_trivially_copy_assignable<_Tp>::value; +}; + +template +void +__stable_sort(_RandomAccessIterator __first, _RandomAccessIterator __last, _Compare __comp, + typename iterator_traits<_RandomAccessIterator>::difference_type __len, + typename iterator_traits<_RandomAccessIterator>::value_type* __buff, ptrdiff_t __buff_size) +{ + typedef typename iterator_traits<_RandomAccessIterator>::value_type value_type; + typedef typename iterator_traits<_RandomAccessIterator>::difference_type difference_type; + switch (__len) + { + case 0: + case 1: + return; + case 2: + if (__comp(*--__last, *__first)) + swap(*__first, *__last); + return; + } + if (__len <= static_cast(__stable_sort_switch::value)) + { + _VSTD::__insertion_sort<_Compare>(__first, __last, __comp); + return; + } + typename iterator_traits<_RandomAccessIterator>::difference_type __l2 = __len / 2; + _RandomAccessIterator __m = __first + __l2; + if (__len <= __buff_size) + { + __destruct_n __d(0); + unique_ptr __h2(__buff, __d); + _VSTD::__stable_sort_move<_Compare>(__first, __m, __comp, __l2, __buff); + __d.__set(__l2, (value_type*)nullptr); + _VSTD::__stable_sort_move<_Compare>(__m, __last, __comp, __len - __l2, __buff + __l2); + __d.__set(__len, (value_type*)nullptr); + _VSTD::__merge_move_assign<_Compare>(__buff, __buff + __l2, __buff + __l2, __buff + __len, __first, __comp); +// _VSTD::__merge<_Compare>(move_iterator(__buff), +// move_iterator(__buff + __l2), +// move_iterator<_RandomAccessIterator>(__buff + __l2), +// move_iterator<_RandomAccessIterator>(__buff + __len), +// __first, __comp); + return; + } + _VSTD::__stable_sort<_Compare>(__first, __m, __comp, __l2, __buff, __buff_size); + _VSTD::__stable_sort<_Compare>(__m, __last, __comp, __len - __l2, __buff, __buff_size); + _VSTD::__inplace_merge<_Compare>(__first, __m, __last, __comp, __l2, __len - __l2, __buff, __buff_size); +} + +template +inline _LIBCPP_INLINE_VISIBILITY +void +stable_sort(_RandomAccessIterator __first, _RandomAccessIterator __last, _Compare __comp) +{ + typedef typename iterator_traits<_RandomAccessIterator>::value_type value_type; + typedef typename iterator_traits<_RandomAccessIterator>::difference_type difference_type; + difference_type __len = __last - __first; + pair __buf(0, 0); + unique_ptr __h; + if (__len > static_cast(__stable_sort_switch::value)) + { + __buf = _VSTD::get_temporary_buffer(__len); + __h.reset(__buf.first); + } + typedef typename __comp_ref_type<_Compare>::type _Comp_ref; + _VSTD::__stable_sort<_Comp_ref>(__first, __last, __comp, __len, __buf.first, __buf.second); +} + +template +inline _LIBCPP_INLINE_VISIBILITY +void +stable_sort(_RandomAccessIterator __first, _RandomAccessIterator __last) +{ + _VSTD::stable_sort(__first, __last, __less::value_type>()); +} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_STABLE_SORT_H diff --git a/lib/libcxx/include/__algorithm/swap_ranges.h b/lib/libcxx/include/__algorithm/swap_ranges.h new file mode 100644 index 000000000..3c72dbd24 --- /dev/null +++ b/lib/libcxx/include/__algorithm/swap_ranges.h @@ -0,0 +1,37 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_SWAP_RANGES_H +#define _LIBCPP___ALGORITHM_SWAP_RANGES_H + +#include <__config> +#include <__utility/swap.h> +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 _ForwardIterator2 +swap_ranges(_ForwardIterator1 __first1, _ForwardIterator1 __last1, _ForwardIterator2 __first2) { + for (; __first1 != __last1; ++__first1, (void)++__first2) + swap(*__first1, *__first2); + return __first2; +} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_SWAP_RANGES_H diff --git a/lib/libcxx/include/__algorithm/transform.h b/lib/libcxx/include/__algorithm/transform.h new file mode 100644 index 000000000..218f0f12e --- /dev/null +++ b/lib/libcxx/include/__algorithm/transform.h @@ -0,0 +1,48 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_TRANSFORM_H +#define _LIBCPP___ALGORITHM_TRANSFORM_H + +#include <__config> + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +_OutputIterator +transform(_InputIterator __first, _InputIterator __last, _OutputIterator __result, _UnaryOperation __op) +{ + for (; __first != __last; ++__first, (void) ++__result) + *__result = __op(*__first); + return __result; +} + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +_OutputIterator +transform(_InputIterator1 __first1, _InputIterator1 __last1, _InputIterator2 __first2, + _OutputIterator __result, _BinaryOperation __binary_op) +{ + for (; __first1 != __last1; ++__first1, (void) ++__first2, ++__result) + *__result = __binary_op(*__first1, *__first2); + return __result; +} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_TRANSFORM_H diff --git a/lib/libcxx/include/__algorithm/unique.h b/lib/libcxx/include/__algorithm/unique.h new file mode 100644 index 000000000..fb6251a39 --- /dev/null +++ b/lib/libcxx/include/__algorithm/unique.h @@ -0,0 +1,63 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_UNIQUE_H +#define _LIBCPP___ALGORITHM_UNIQUE_H + +#include <__config> +#include <__algorithm/comp.h> +#include <__algorithm/adjacent_find.h> +#include <__iterator/iterator_traits.h> +#include <__utility/move.h> +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +// unique + +template +_LIBCPP_NODISCARD_EXT _LIBCPP_CONSTEXPR_AFTER_CXX17 _ForwardIterator +unique(_ForwardIterator __first, _ForwardIterator __last, _BinaryPredicate __pred) +{ + __first = _VSTD::adjacent_find<_ForwardIterator, typename add_lvalue_reference<_BinaryPredicate>::type> + (__first, __last, __pred); + if (__first != __last) + { + // ... a a ? ... + // f i + _ForwardIterator __i = __first; + for (++__i; ++__i != __last;) + if (!__pred(*__first, *__i)) + *++__first = _VSTD::move(*__i); + ++__first; + } + return __first; +} + +template +_LIBCPP_NODISCARD_EXT inline +_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +_ForwardIterator +unique(_ForwardIterator __first, _ForwardIterator __last) +{ + typedef typename iterator_traits<_ForwardIterator>::value_type __v; + return _VSTD::unique(__first, __last, __equal_to<__v>()); +} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_UNIQUE_H diff --git a/lib/libcxx/include/__algorithm/unique_copy.h b/lib/libcxx/include/__algorithm/unique_copy.h new file mode 100644 index 000000000..974a7c4df --- /dev/null +++ b/lib/libcxx/include/__algorithm/unique_copy.h @@ -0,0 +1,114 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_UNIQUE_COPY_H +#define _LIBCPP___ALGORITHM_UNIQUE_COPY_H + +#include <__config> +#include <__algorithm/comp.h> +#include <__iterator/iterator_traits.h> +#include +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +_LIBCPP_CONSTEXPR_AFTER_CXX17 _OutputIterator +__unique_copy(_InputIterator __first, _InputIterator __last, _OutputIterator __result, _BinaryPredicate __pred, + input_iterator_tag, output_iterator_tag) +{ + if (__first != __last) + { + typename iterator_traits<_InputIterator>::value_type __t(*__first); + *__result = __t; + ++__result; + while (++__first != __last) + { + if (!__pred(__t, *__first)) + { + __t = *__first; + *__result = __t; + ++__result; + } + } + } + return __result; +} + +template +_LIBCPP_CONSTEXPR_AFTER_CXX17 _OutputIterator +__unique_copy(_ForwardIterator __first, _ForwardIterator __last, _OutputIterator __result, _BinaryPredicate __pred, + forward_iterator_tag, output_iterator_tag) +{ + if (__first != __last) + { + _ForwardIterator __i = __first; + *__result = *__i; + ++__result; + while (++__first != __last) + { + if (!__pred(*__i, *__first)) + { + *__result = *__first; + ++__result; + __i = __first; + } + } + } + return __result; +} + +template +_LIBCPP_CONSTEXPR_AFTER_CXX17 _ForwardIterator +__unique_copy(_InputIterator __first, _InputIterator __last, _ForwardIterator __result, _BinaryPredicate __pred, + input_iterator_tag, forward_iterator_tag) +{ + if (__first != __last) + { + *__result = *__first; + while (++__first != __last) + if (!__pred(*__result, *__first)) + *++__result = *__first; + ++__result; + } + return __result; +} + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +_OutputIterator +unique_copy(_InputIterator __first, _InputIterator __last, _OutputIterator __result, _BinaryPredicate __pred) +{ + return _VSTD::__unique_copy::type> + (__first, __last, __result, __pred, + typename iterator_traits<_InputIterator>::iterator_category(), + typename iterator_traits<_OutputIterator>::iterator_category()); +} + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +_OutputIterator +unique_copy(_InputIterator __first, _InputIterator __last, _OutputIterator __result) +{ + typedef typename iterator_traits<_InputIterator>::value_type __v; + return _VSTD::unique_copy(__first, __last, __result, __equal_to<__v>()); +} + + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_UNIQUE_COPY_H diff --git a/lib/libcxx/include/__algorithm/unwrap_iter.h b/lib/libcxx/include/__algorithm/unwrap_iter.h new file mode 100644 index 000000000..a45d45cdd --- /dev/null +++ b/lib/libcxx/include/__algorithm/unwrap_iter.h @@ -0,0 +1,87 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_UNWRAP_ITER_H +#define _LIBCPP___ALGORITHM_UNWRAP_ITER_H + +#include <__config> +#include +#include <__memory/pointer_traits.h> +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +// The job of __unwrap_iter is to lower contiguous iterators (such as +// vector::iterator) into pointers, to reduce the number of template +// instantiations and to enable pointer-based optimizations e.g. in std::copy. +// For iterators that are not contiguous, it must be a no-op. +// In debug mode, we don't do this. +// +// __unwrap_iter is non-constexpr for user-defined iterators whose +// `to_address` and/or `operator->` is non-constexpr. This is okay; but we +// try to avoid doing __unwrap_iter in constant-evaluated contexts anyway. +// +// Some algorithms (e.g. std::copy, but not std::sort) need to convert an +// "unwrapped" result back into a contiguous iterator. Since contiguous iterators +// are random-access, we can do this portably using iterator arithmetic; this +// is the job of __rewrap_iter. + +template ::value> +struct __unwrap_iter_impl { + static _LIBCPP_CONSTEXPR _Iter + __apply(_Iter __i) _NOEXCEPT { + return __i; + } +}; + +#if _LIBCPP_DEBUG_LEVEL < 2 + +template +struct __unwrap_iter_impl<_Iter, true> { + static _LIBCPP_CONSTEXPR decltype(_VSTD::__to_address(declval<_Iter>())) + __apply(_Iter __i) _NOEXCEPT { + return _VSTD::__to_address(__i); + } +}; + +#endif // _LIBCPP_DEBUG_LEVEL < 2 + +template > +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR +decltype(_Impl::__apply(declval<_Iter>())) +__unwrap_iter(_Iter __i) _NOEXCEPT +{ + return _Impl::__apply(__i); +} + +template +_OrigIter __rewrap_iter(_OrigIter, _OrigIter __result) +{ + return __result; +} + +template +_OrigIter __rewrap_iter(_OrigIter __first, _UnwrappedIter __result) +{ + // Precondition: __result is reachable from __first + // Precondition: _OrigIter is a contiguous iterator + return __first + (__result - _VSTD::__unwrap_iter(__first)); +} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_UNWRAP_ITER_H diff --git a/lib/libcxx/include/__algorithm/upper_bound.h b/lib/libcxx/include/__algorithm/upper_bound.h new file mode 100644 index 000000000..7be607f82 --- /dev/null +++ b/lib/libcxx/include/__algorithm/upper_bound.h @@ -0,0 +1,72 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ALGORITHM_UPPER_BOUND_H +#define _LIBCPP___ALGORITHM_UPPER_BOUND_H + +#include <__config> +#include <__algorithm/comp.h> +#include <__algorithm/half_positive.h> +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +_LIBCPP_CONSTEXPR_AFTER_CXX17 _ForwardIterator +__upper_bound(_ForwardIterator __first, _ForwardIterator __last, const _Tp& __value_, _Compare __comp) +{ + typedef typename iterator_traits<_ForwardIterator>::difference_type difference_type; + difference_type __len = _VSTD::distance(__first, __last); + while (__len != 0) + { + difference_type __l2 = _VSTD::__half_positive(__len); + _ForwardIterator __m = __first; + _VSTD::advance(__m, __l2); + if (__comp(__value_, *__m)) + __len = __l2; + else + { + __first = ++__m; + __len -= __l2 + 1; + } + } + return __first; +} + +template +_LIBCPP_NODISCARD_EXT inline +_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +_ForwardIterator +upper_bound(_ForwardIterator __first, _ForwardIterator __last, const _Tp& __value_, _Compare __comp) +{ + typedef typename add_lvalue_reference<_Compare>::type _Comp_ref; + return _VSTD::__upper_bound<_Comp_ref>(__first, __last, __value_, __comp); +} + +template +_LIBCPP_NODISCARD_EXT inline +_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +_ForwardIterator +upper_bound(_ForwardIterator __first, _ForwardIterator __last, const _Tp& __value_) +{ + return _VSTD::upper_bound(__first, __last, __value_, + __less<_Tp, typename iterator_traits<_ForwardIterator>::value_type>()); +} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ALGORITHM_UPPER_BOUND_H diff --git a/lib/libcxx/include/__availability b/lib/libcxx/include/__availability index db2267c8e..13d11950f 100644 --- a/lib/libcxx/include/__availability +++ b/lib/libcxx/include/__availability @@ -43,6 +43,14 @@ // as unavailable. When vendors decide to ship the feature as part of their // shared library, they can update the markup appropriately. // +// Furthermore, many features in the standard library have corresponding +// feature-test macros. When a feature is made unavailable on some deployment +// target, a macro should be defined to signal that it is unavailable. That +// macro can then be picked up when feature-test macros are generated (see +// generate_feature_test_macro_components.py) to make sure that feature-test +// macros don't announce a feature as being implemented if it has been marked +// as unavailable. +// // Note that this mechanism is disabled by default in the "upstream" libc++. // Availability annotations are only meaningful when shipping libc++ inside // a platform (i.e. as a system library), and so vendors that want them should @@ -76,6 +84,8 @@ // This controls the availability of std::shared_mutex and std::shared_timed_mutex, // which were added to the dylib later. # define _LIBCPP_AVAILABILITY_SHARED_MUTEX +// # define _LIBCPP_AVAILABILITY_DISABLE_FTM___cpp_lib_shared_mutex +// # define _LIBCPP_AVAILABILITY_DISABLE_FTM___cpp_lib_shared_timed_mutex // These macros control the availability of std::bad_optional_access and // other exception types. These were put in the shared library to prevent @@ -114,6 +124,7 @@ # define _LIBCPP_AVAILABILITY_FILESYSTEM # define _LIBCPP_AVAILABILITY_FILESYSTEM_PUSH # define _LIBCPP_AVAILABILITY_FILESYSTEM_POP +// # define _LIBCPP_AVAILABILITY_DISABLE_FTM___cpp_lib_filesystem // This controls the availability of std::to_chars. # define _LIBCPP_AVAILABILITY_TO_CHARS @@ -122,6 +133,17 @@ // which requires shared library support for various operations // (see libcxx/src/atomic.cpp). # define _LIBCPP_AVAILABILITY_SYNC +// # define _LIBCPP_AVAILABILITY_DISABLE_FTM___cpp_lib_atomic_wait +// # define _LIBCPP_AVAILABILITY_DISABLE_FTM___cpp_lib_barrier +// # define _LIBCPP_AVAILABILITY_DISABLE_FTM___cpp_lib_latch +// # define _LIBCPP_AVAILABILITY_DISABLE_FTM___cpp_lib_semaphore + + // This controls the availability of the C++20 format library. + // The library is in development and not ABI stable yet. Currently + // P2216 is aiming to be retroactively accepted in C++20. This paper + // contains ABI breaking changes. +# define _LIBCPP_AVAILABILITY_FORMAT +// # define _LIBCPP_AVAILABILITY_DISABLE_FTM___cpp_lib_format #elif defined(__APPLE__) @@ -130,6 +152,14 @@ __attribute__((availability(ios,strict,introduced=10.0))) \ __attribute__((availability(tvos,strict,introduced=10.0))) \ __attribute__((availability(watchos,strict,introduced=3.0))) +# if (defined(__ENVIRONMENT_MAC_OS_X_VERSION_MIN_REQUIRED__) && __ENVIRONMENT_MAC_OS_X_VERSION_MIN_REQUIRED__ < 101200) || \ + (defined(__ENVIRONMENT_IPHONE_OS_VERSION_MIN_REQUIRED__) && __ENVIRONMENT_IPHONE_OS_VERSION_MIN_REQUIRED__ < 100000) || \ + (defined(__ENVIRONMENT_TV_OS_VERSION_MIN_REQUIRED__) && __ENVIRONMENT_TV_OS_VERSION_MIN_REQUIRED__ < 100000) || \ + (defined(__ENVIRONMENT_WATCH_OS_VERSION_MIN_REQUIRED__) && __ENVIRONMENT_WATCH_OS_VERSION_MIN_REQUIRED__ < 30000) +# define _LIBCPP_AVAILABILITY_DISABLE_FTM___cpp_lib_shared_mutex +# define _LIBCPP_AVAILABILITY_DISABLE_FTM___cpp_lib_shared_timed_mutex +# endif + # define _LIBCPP_AVAILABILITY_BAD_OPTIONAL_ACCESS \ __attribute__((availability(macosx,strict,introduced=10.13))) \ __attribute__((availability(ios,strict,introduced=11.0))) \ @@ -139,27 +169,34 @@ _LIBCPP_AVAILABILITY_BAD_OPTIONAL_ACCESS # define _LIBCPP_AVAILABILITY_BAD_ANY_CAST \ _LIBCPP_AVAILABILITY_BAD_OPTIONAL_ACCESS + # define _LIBCPP_AVAILABILITY_UNCAUGHT_EXCEPTIONS \ __attribute__((availability(macosx,strict,introduced=10.12))) \ __attribute__((availability(ios,strict,introduced=10.0))) \ __attribute__((availability(tvos,strict,introduced=10.0))) \ __attribute__((availability(watchos,strict,introduced=3.0))) + # define _LIBCPP_AVAILABILITY_SIZED_NEW_DELETE \ __attribute__((availability(macosx,strict,introduced=10.12))) \ __attribute__((availability(ios,strict,introduced=10.0))) \ __attribute__((availability(tvos,strict,introduced=10.0))) \ __attribute__((availability(watchos,strict,introduced=3.0))) + # define _LIBCPP_AVAILABILITY_FUTURE_ERROR \ __attribute__((availability(ios,strict,introduced=6.0))) + # define _LIBCPP_AVAILABILITY_TYPEINFO_VTABLE \ __attribute__((availability(macosx,strict,introduced=10.9))) \ __attribute__((availability(ios,strict,introduced=7.0))) + # define _LIBCPP_AVAILABILITY_LOCALE_CATEGORY \ __attribute__((availability(macosx,strict,introduced=10.9))) \ __attribute__((availability(ios,strict,introduced=7.0))) + # define _LIBCPP_AVAILABILITY_ATOMIC_SHARED_PTR \ __attribute__((availability(macosx,strict,introduced=10.9))) \ __attribute__((availability(ios,strict,introduced=7.0))) + # define _LIBCPP_AVAILABILITY_FILESYSTEM \ __attribute__((availability(macosx,strict,introduced=10.15))) \ __attribute__((availability(ios,strict,introduced=13.0))) \ @@ -175,11 +212,38 @@ _Pragma("clang attribute pop") \ _Pragma("clang attribute pop") \ _Pragma("clang attribute pop") +# if (defined(__ENVIRONMENT_MAC_OS_X_VERSION_MIN_REQUIRED__) && __ENVIRONMENT_MAC_OS_X_VERSION_MIN_REQUIRED__ < 101500) || \ + (defined(__ENVIRONMENT_IPHONE_OS_VERSION_MIN_REQUIRED__) && __ENVIRONMENT_IPHONE_OS_VERSION_MIN_REQUIRED__ < 130000) || \ + (defined(__ENVIRONMENT_TV_OS_VERSION_MIN_REQUIRED__) && __ENVIRONMENT_TV_OS_VERSION_MIN_REQUIRED__ < 130000) || \ + (defined(__ENVIRONMENT_WATCH_OS_VERSION_MIN_REQUIRED__) && __ENVIRONMENT_WATCH_OS_VERSION_MIN_REQUIRED__ < 60000) +# define _LIBCPP_AVAILABILITY_DISABLE_FTM___cpp_lib_filesystem +# endif + # define _LIBCPP_AVAILABILITY_TO_CHARS \ _LIBCPP_AVAILABILITY_FILESYSTEM + # define _LIBCPP_AVAILABILITY_SYNC \ - __attribute__((unavailable)) + __attribute__((availability(macosx,strict,introduced=11.0))) \ + __attribute__((availability(ios,strict,introduced=14.0))) \ + __attribute__((availability(tvos,strict,introduced=14.0))) \ + __attribute__((availability(watchos,strict,introduced=7.0))) +# if (defined(__ENVIRONMENT_MAC_OS_X_VERSION_MIN_REQUIRED__) && __ENVIRONMENT_MAC_OS_X_VERSION_MIN_REQUIRED__ < 110000) || \ + (defined(__ENVIRONMENT_IPHONE_OS_VERSION_MIN_REQUIRED__) && __ENVIRONMENT_IPHONE_OS_VERSION_MIN_REQUIRED__ < 140000) || \ + (defined(__ENVIRONMENT_TV_OS_VERSION_MIN_REQUIRED__) && __ENVIRONMENT_TV_OS_VERSION_MIN_REQUIRED__ < 140000) || \ + (defined(__ENVIRONMENT_WATCH_OS_VERSION_MIN_REQUIRED__) && __ENVIRONMENT_WATCH_OS_VERSION_MIN_REQUIRED__ < 70000) +# define _LIBCPP_AVAILABILITY_DISABLE_FTM___cpp_lib_atomic_wait +# define _LIBCPP_AVAILABILITY_DISABLE_FTM___cpp_lib_barrier +# define _LIBCPP_AVAILABILITY_DISABLE_FTM___cpp_lib_latch +# define _LIBCPP_AVAILABILITY_DISABLE_FTM___cpp_lib_semaphore +# endif + // This controls the availability of the C++20 format library. + // The library is in development and not ABI stable yet. Currently + // P2216 is aiming to be retroactively accepted in C++20. This paper + // contains ABI breaking changes. +# define _LIBCPP_AVAILABILITY_FORMAT \ + __attribute__((unavailable)) +# define _LIBCPP_AVAILABILITY_DISABLE_FTM___cpp_lib_format #else // ...New vendors can add availability markup here... @@ -203,4 +267,4 @@ # define _LIBCPP_AVAILABILITY_THROW_BAD_VARIANT_ACCESS _LIBCPP_AVAILABILITY_BAD_VARIANT_ACCESS #endif -#endif // _LIBCPP___AVAILABILITY +#endif // _LIBCPP___AVAILABILITY diff --git a/lib/libcxx/include/__bit_reference b/lib/libcxx/include/__bit_reference index 9cfb4b84e..a02492c07 100644 --- a/lib/libcxx/include/__bit_reference +++ b/lib/libcxx/include/__bit_reference @@ -1114,28 +1114,26 @@ public: #endif {} - // avoid re-declaring a copy constructor for the non-const version. - using __type_for_copy_to_const = - _If<_IsConst, __bit_iterator<_Cp, false>, struct __private_nat>; - + // When _IsConst=false, this is the copy constructor. + // It is non-trivial. Making it trivial would break ABI. + // When _IsConst=true, this is a converting constructor; + // the copy and move constructors are implicitly generated + // and trivial. _LIBCPP_INLINE_VISIBILITY - __bit_iterator(const __type_for_copy_to_const& __it) _NOEXCEPT + __bit_iterator(const __bit_iterator<_Cp, false>& __it) _NOEXCEPT : __seg_(__it.__seg_), __ctz_(__it.__ctz_) {} - // The non-const __bit_iterator has historically had a non-trivial - // copy constructor (as a quirk of its construction). We need to maintain - // this for ABI purposes. - using __type_for_abi_non_trivial_copy_ctor = - _If; - - _LIBCPP_INLINE_VISIBILITY - __bit_iterator(__type_for_abi_non_trivial_copy_ctor const& __it) _NOEXCEPT - : __seg_(__it.__seg_), __ctz_(__it.__ctz_) {} - - // Always declare the copy assignment operator since the implicit declaration - // is deprecated. + // When _IsConst=false, we have a user-provided copy constructor, + // so we must also provide a copy assignment operator because + // the implicit generation of a defaulted one is deprecated. + // When _IsConst=true, the assignment operators are + // implicitly generated and trivial. _LIBCPP_INLINE_VISIBILITY - __bit_iterator& operator=(__bit_iterator const&) = default; + __bit_iterator& operator=(const _If<_IsConst, struct __private_nat, __bit_iterator>& __it) { + __seg_ = __it.__seg_; + __ctz_ = __it.__ctz_; + return *this; + } _LIBCPP_INLINE_VISIBILITY reference operator*() const _NOEXCEPT {return reference(__seg_, __storage_type(1) << __ctz_);} @@ -1302,4 +1300,4 @@ _LIBCPP_END_NAMESPACE_STD _LIBCPP_POP_MACROS -#endif // _LIBCPP___BIT_REFERENCE +#endif // _LIBCPP___BIT_REFERENCE diff --git a/lib/libcxx/include/__bits b/lib/libcxx/include/__bits index 0d321da92..b565a7823 100644 --- a/lib/libcxx/include/__bits +++ b/lib/libcxx/include/__bits @@ -76,7 +76,6 @@ inline _LIBCPP_INLINE_VISIBILITY int __libcpp_ctz(unsigned long long __x) { unsigned long __where; #if defined(_LIBCPP_HAS_BITSCAN64) - (defined(_M_AMD64) || defined(__x86_64__)) if (_BitScanForward64(&__where, __x)) return static_cast(__where); #else @@ -143,4 +142,4 @@ _LIBCPP_END_NAMESPACE_STD _LIBCPP_POP_MACROS -#endif // _LIBCPP__BITS +#endif // _LIBCPP___BITS diff --git a/lib/libcxx/include/__bsd_locale_fallbacks.h b/lib/libcxx/include/__bsd_locale_fallbacks.h index a807fe039..ed0eabf60 100644 --- a/lib/libcxx/include/__bsd_locale_fallbacks.h +++ b/lib/libcxx/include/__bsd_locale_fallbacks.h @@ -13,9 +13,9 @@ #ifndef _LIBCPP_BSD_LOCALE_FALLBACKS_DEFAULTS_H #define _LIBCPP_BSD_LOCALE_FALLBACKS_DEFAULTS_H -#include -#include #include +#include +#include #if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) #pragma GCC system_header diff --git a/lib/libcxx/include/__config b/lib/libcxx/include/__config index a3838c89e..a945bd9ab 100644 --- a/lib/libcxx/include/__config +++ b/lib/libcxx/include/__config @@ -32,7 +32,7 @@ # define _GNUC_VER_NEW 0 #endif -#define _LIBCPP_VERSION 12000 +#define _LIBCPP_VERSION 13000 #ifndef _LIBCPP_ABI_VERSION # define _LIBCPP_ABI_VERSION 1 @@ -54,7 +54,7 @@ # else # define _LIBCPP_STD_VER 21 // current year, or date of c++2b ratification # endif -#endif // _LIBCPP_STD_VER +#endif // _LIBCPP_STD_VER #if defined(__ELF__) # define _LIBCPP_OBJECT_FORMAT_ELF 1 @@ -86,17 +86,18 @@ // provided under the alternate keyword __nullptr, which changes the mangling // of nullptr_t. This option is ABI incompatible with GCC in C++03 mode. # define _LIBCPP_ABI_ALWAYS_USE_CXX11_NULLPTR -// Define the `pointer_safety` enum as a C++11 strongly typed enumeration -// instead of as a class simulating an enum. If this option is enabled -// `pointer_safety` and `get_pointer_safety()` will no longer be available -// in C++03. -# define _LIBCPP_ABI_POINTER_SAFETY_ENUM_TYPE // Define a key function for `bad_function_call` in the library, to centralize // its vtable and typeinfo to libc++ rather than having all other libraries // using that class define their own copies. # define _LIBCPP_ABI_BAD_FUNCTION_CALL_KEY_FUNCTION // Enable optimized version of __do_get_(un)signed which avoids redundant copies. # define _LIBCPP_ABI_OPTIMIZED_LOCALE_NUM_GET +// In C++20 and later, don't derive std::plus from std::binary_function, +// nor std::negate from std::unary_function. +# define _LIBCPP_ABI_NO_BINDER_BASES +// Give reverse_iterator one data member of type T, not two. +// Also, in C++17 and later, don't derive iterator types from std::iterator. +# define _LIBCPP_ABI_NO_ITERATOR_BASES // Use the smallest possible integer type to represent the index of the variant. // Previously libc++ used "unsigned int" exclusively. # define _LIBCPP_ABI_VARIANT_INDEX_TYPE_OPTIMIZATION @@ -104,6 +105,8 @@ # define _LIBCPP_ABI_OPTIMIZED_FUNCTION // All the regex constants must be distinct and nonzero. # define _LIBCPP_ABI_REGEX_CONSTANTS_NONZERO +// Use raw pointers, not wrapped ones, for std::span's iterator type. +# define _LIBCPP_ABI_SPAN_POINTER_ITERATORS // Re-worked external template instantiations for std::string with a focus on // performance and fast-path inlining. # define _LIBCPP_ABI_STRING_OPTIMIZED_EXTERNAL_INSTANTIATION @@ -181,11 +184,12 @@ #define __has_include(...) 0 #endif -#if defined(__clang__) -# define _LIBCPP_COMPILER_CLANG -# ifndef __apple_build_version__ -# define _LIBCPP_CLANG_VER (__clang_major__ * 100 + __clang_minor__) -# endif +#if defined(__apple_build_version__) +# define _LIBCPP_COMPILER_CLANG_BASED +# define _LIBCPP_APPLE_CLANG_VER (__apple_build_version__ / 10000) +#elif defined(__clang__) +# define _LIBCPP_COMPILER_CLANG_BASED +# define _LIBCPP_CLANG_VER (__clang_major__ * 100 + __clang_minor__) #elif defined(__GNUC__) # define _LIBCPP_COMPILER_GCC #elif defined(_MSC_VER) @@ -235,13 +239,13 @@ # if __LITTLE_ENDIAN__ # define _LIBCPP_LITTLE_ENDIAN # endif // __LITTLE_ENDIAN__ -#endif // __LITTLE_ENDIAN__ +#endif // __LITTLE_ENDIAN__ #ifdef __BIG_ENDIAN__ # if __BIG_ENDIAN__ # define _LIBCPP_BIG_ENDIAN # endif // __BIG_ENDIAN__ -#endif // __BIG_ENDIAN__ +#endif // __BIG_ENDIAN__ #ifdef __BYTE_ORDER__ # if __BYTE_ORDER__ == __ORDER_LITTLE_ENDIAN__ @@ -262,7 +266,7 @@ # ifndef __LONG_LONG_SUPPORTED # define _LIBCPP_HAS_NO_LONG_LONG # endif // __LONG_LONG_SUPPORTED -#endif // __FreeBSD__ +#endif // __FreeBSD__ #if defined(__NetBSD__) || defined(__OpenBSD__) # include @@ -271,7 +275,7 @@ # else // _BYTE_ORDER == _LITTLE_ENDIAN # define _LIBCPP_BIG_ENDIAN # endif // _BYTE_ORDER == _LITTLE_ENDIAN -#endif // defined(__NetBSD__) || defined(__OpenBSD__) +#endif // defined(__NetBSD__) || defined(__OpenBSD__) #if defined(_WIN32) # define _LIBCPP_WIN32API @@ -340,7 +344,7 @@ # else // __BYTE_ORDER == __BIG_ENDIAN # error unable to determine endian # endif -#endif // !defined(_LIBCPP_LITTLE_ENDIAN) && !defined(_LIBCPP_BIG_ENDIAN) +#endif // !defined(_LIBCPP_LITTLE_ENDIAN) && !defined(_LIBCPP_BIG_ENDIAN) #if __has_attribute(__no_sanitize__) && !defined(_LIBCPP_COMPILER_GCC) # define _LIBCPP_NO_CFI __attribute__((__no_sanitize__("cfi"))) @@ -348,7 +352,17 @@ # define _LIBCPP_NO_CFI #endif -#if __ISO_C_VISIBLE >= 2011 || __cplusplus >= 201103L +// If the compiler supports using_if_exists, pretend we have those functions and they'll +// be picked up if the C library provides them. +// +// TODO: Once we drop support for Clang 12, we can assume the compiler supports using_if_exists +// for platforms that don't have a conforming C11 library, so we can drop this whole thing. +#if __has_attribute(using_if_exists) +# define _LIBCPP_HAS_TIMESPEC_GET +# define _LIBCPP_HAS_QUICK_EXIT +# define _LIBCPP_HAS_ALIGNED_ALLOC +#else +#if (defined(__ISO_C_VISIBLE) && (__ISO_C_VISIBLE >= 2011)) || __cplusplus >= 201103L # if defined(__FreeBSD__) # define _LIBCPP_HAS_ALIGNED_ALLOC # define _LIBCPP_HAS_QUICK_EXIT @@ -387,22 +401,26 @@ # define _LIBCPP_HAS_QUICK_EXIT # define _LIBCPP_HAS_TIMESPEC_GET # endif +# elif defined(_LIBCPP_MSVCRT) + // Using Microsoft's C Runtime library, not MinGW +# define _LIBCPP_HAS_TIMESPEC_GET # elif defined(__APPLE__) // timespec_get and aligned_alloc were introduced in macOS 10.15 and // aligned releases -# if (__ENVIRONMENT_MAC_OS_X_VERSION_MIN_REQUIRED__ >= 101500 || \ - __ENVIRONMENT_IPHONE_OS_VERSION_MIN_REQUIRED__ >= 130000 || \ - __ENVIRONMENT_TV_OS_VERSION_MIN_REQUIRED__ >= 130000 || \ - __ENVIRONMENT_WATCH_OS_VERSION_MIN_REQUIRED__ >= 60000) +# if ((defined(__ENVIRONMENT_MAC_OS_X_VERSION_MIN_REQUIRED__) && __ENVIRONMENT_MAC_OS_X_VERSION_MIN_REQUIRED__ >= 101500) || \ + (defined(__ENVIRONMENT_IPHONE_OS_VERSION_MIN_REQUIRED__) && __ENVIRONMENT_IPHONE_OS_VERSION_MIN_REQUIRED__ >= 130000) || \ + (defined(__ENVIRONMENT_TV_OS_VERSION_MIN_REQUIRED__) && __ENVIRONMENT_TV_OS_VERSION_MIN_REQUIRED__ >= 130000) || \ + (defined(__ENVIRONMENT_WATCH_OS_VERSION_MIN_REQUIRED__) && __ENVIRONMENT_WATCH_OS_VERSION_MIN_REQUIRED__ >= 60000)) # define _LIBCPP_HAS_ALIGNED_ALLOC # define _LIBCPP_HAS_TIMESPEC_GET # endif # endif // __APPLE__ #endif +#endif // __has_attribute(using_if_exists) #ifndef _LIBCPP_CXX03_LANG # define _LIBCPP_ALIGNOF(_Tp) alignof(_Tp) -#elif defined(_LIBCPP_COMPILER_CLANG) +#elif defined(_LIBCPP_COMPILER_CLANG_BASED) # define _LIBCPP_ALIGNOF(_Tp) _Alignof(_Tp) #else # error "We don't know a correct way to implement alignof(T) in C++03 outside of Clang" @@ -410,7 +428,7 @@ #define _LIBCPP_PREFERRED_ALIGNOF(_Tp) __alignof(_Tp) -#if defined(_LIBCPP_COMPILER_CLANG) +#if defined(_LIBCPP_COMPILER_CLANG_BASED) #if defined(_LIBCPP_ALTERNATE_STRING_LAYOUT) # error _LIBCPP_ALTERNATE_STRING_LAYOUT is deprecated, please use _LIBCPP_ABI_ALTERNATE_STRING_LAYOUT instead @@ -484,12 +502,12 @@ typedef __char32_t char32_t; #define _LIBCPP_HAS_NO_NOEXCEPT #endif -#if !defined(_LIBCPP_HAS_NO_ASAN) && !__has_feature(address_sanitizer) +#if !__has_feature(address_sanitizer) #define _LIBCPP_HAS_NO_ASAN #endif // Allow for build-time disabling of unsigned integer sanitization -#if !defined(_LIBCPP_DISABLE_UBSAN_UNSIGNED_INTEGER_CHECK) && __has_attribute(no_sanitize) +#if __has_attribute(no_sanitize) #define _LIBCPP_DISABLE_UBSAN_UNSIGNED_INTEGER_CHECK __attribute__((__no_sanitize__("unsigned-integer-overflow"))) #endif @@ -508,8 +526,8 @@ typedef __char32_t char32_t; #define _LIBCPP_ALWAYS_INLINE __attribute__ ((__always_inline__)) // Literal operators ""d and ""y are supported starting with LLVM Clang 8 and AppleClang 10.0.1 -#if (defined(_LIBCPP_CLANG_VER) && _LIBCPP_CLANG_VER < 800) || \ - (defined(__apple_build_version__) && __apple_build_version__ < 10010000) +#if (defined(_LIBCPP_CLANG_VER) && _LIBCPP_CLANG_VER < 800) || \ + (defined(_LIBCPP_APPLE_CLANG_VER) && _LIBCPP_APPLE_CLANG_VER < 1001) #define _LIBCPP_HAS_NO_CXX20_CHRONO_LITERALS #endif @@ -522,7 +540,7 @@ typedef __char32_t char32_t; #define _LIBCPP_NORETURN __attribute__((noreturn)) -#if !__EXCEPTIONS +#if !defined(__EXCEPTIONS) # define _LIBCPP_NO_EXCEPTIONS #endif @@ -536,7 +554,7 @@ typedef __char32_t char32_t; #define _LIBCPP_HAS_NO_VARIABLE_TEMPLATES #endif -#if !defined(_LIBCPP_HAS_NO_ASAN) && !defined(__SANITIZE_ADDRESS__) +#if !defined(__SANITIZE_ADDRESS__) #define _LIBCPP_HAS_NO_ASAN #endif @@ -640,6 +658,7 @@ typedef __char32_t char32_t; #define _LIBCPP_HIDDEN #define _LIBCPP_METHOD_TEMPLATE_IMPLICIT_INSTANTIATION_VIS #define _LIBCPP_TEMPLATE_VIS +#define _LIBCPP_TEMPLATE_DATA_VIS #define _LIBCPP_ENUM_VIS #endif // defined(_LIBCPP_OBJECT_FORMAT_COFF) @@ -689,6 +708,14 @@ typedef __char32_t char32_t; # endif #endif +#ifndef _LIBCPP_TEMPLATE_DATA_VIS +# if !defined(_LIBCPP_DISABLE_VISIBILITY_ANNOTATIONS) +# define _LIBCPP_TEMPLATE_DATA_VIS __attribute__ ((__visibility__("default"))) +# else +# define _LIBCPP_TEMPLATE_DATA_VIS +# endif +#endif + #ifndef _LIBCPP_EXPORTED_FROM_ABI # if !defined(_LIBCPP_DISABLE_VISIBILITY_ANNOTATIONS) # define _LIBCPP_EXPORTED_FROM_ABI __attribute__((__visibility__("default"))) @@ -792,10 +819,8 @@ _LIBCPP_BEGIN_NAMESPACE_STD _LIBCPP_END_NAMESPACE_STD #define _VSTD_FS _VSTD::__fs::filesystem -#ifndef _LIBCPP_PREFERRED_OVERLOAD -# if __has_attribute(__enable_if__) -# define _LIBCPP_PREFERRED_OVERLOAD __attribute__ ((__enable_if__(true, ""))) -# endif +#if __has_attribute(__enable_if__) +# define _LIBCPP_PREFERRED_OVERLOAD __attribute__ ((__enable_if__(true, ""))) #endif #ifndef _LIBCPP_HAS_NO_NOEXCEPT @@ -809,7 +834,7 @@ _LIBCPP_BEGIN_NAMESPACE_STD _LIBCPP_END_NAMESPACE_STD #ifdef _LIBCPP_HAS_NO_UNICODE_CHARS typedef unsigned short char16_t; typedef unsigned int char32_t; -#endif // _LIBCPP_HAS_NO_UNICODE_CHARS +#endif // _LIBCPP_HAS_NO_UNICODE_CHARS #ifndef __SIZEOF_INT128__ #define _LIBCPP_HAS_NO_INT128 @@ -818,7 +843,7 @@ typedef unsigned int char32_t; #ifdef _LIBCPP_CXX03_LANG # define static_assert(...) _Static_assert(__VA_ARGS__) # define decltype(...) __decltype(__VA_ARGS__) -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG #ifdef _LIBCPP_CXX03_LANG # define _LIBCPP_CONSTEXPR @@ -832,6 +857,14 @@ typedef unsigned int char32_t; # define _LIBCPP_CONSTEVAL consteval #endif +#if !defined(__cpp_concepts) || __cpp_concepts < 201907L +#define _LIBCPP_HAS_NO_CONCEPTS +#endif + +#if _LIBCPP_STD_VER <= 17 || defined(_LIBCPP_HAS_NO_CONCEPTS) +#define _LIBCPP_HAS_NO_RANGES +#endif + #ifdef _LIBCPP_CXX03_LANG # define _LIBCPP_DEFAULT {} #else @@ -850,11 +883,10 @@ typedef unsigned int char32_t; # define _LIBCPP_NOALIAS #endif -#if __has_feature(cxx_explicit_conversions) || defined(__IBMCPP__) || \ - (!defined(_LIBCPP_CXX03_LANG) && defined(__GNUC__)) // All supported GCC versions -# define _LIBCPP_EXPLICIT explicit +#if __has_attribute(using_if_exists) +# define _LIBCPP_USING_IF_EXISTS __attribute__((using_if_exists)) #else -# define _LIBCPP_EXPLICIT +# define _LIBCPP_USING_IF_EXISTS #endif #ifdef _LIBCPP_HAS_NO_STRONG_ENUMS @@ -868,7 +900,7 @@ typedef unsigned int char32_t; #else // _LIBCPP_HAS_NO_STRONG_ENUMS # define _LIBCPP_DECLARE_STRONG_ENUM(x) enum class _LIBCPP_ENUM_VIS x # define _LIBCPP_DECLARE_STRONG_ENUM_EPILOG(x) -#endif // _LIBCPP_HAS_NO_STRONG_ENUMS +#endif // _LIBCPP_HAS_NO_STRONG_ENUMS // _LIBCPP_DEBUG potential values: // - undefined: No assertions. This is the default. @@ -884,33 +916,25 @@ typedef unsigned int char32_t; # error Supported values for _LIBCPP_DEBUG are 0 and 1 #endif -// _LIBCPP_DEBUG_LEVEL is always defined to one of [0, 1, 2] at this point -#if _LIBCPP_DEBUG_LEVEL >= 1 && !defined(_LIBCPP_DISABLE_EXTERN_TEMPLATE) -# define _LIBCPP_EXTERN_TEMPLATE(...) -#endif - -#ifdef _LIBCPP_DISABLE_EXTERN_TEMPLATE -# define _LIBCPP_EXTERN_TEMPLATE(...) -# define _LIBCPP_EXTERN_TEMPLATE_EVEN_IN_DEBUG_MODE(...) -#endif - -#ifndef _LIBCPP_EXTERN_TEMPLATE -#define _LIBCPP_EXTERN_TEMPLATE(...) extern template __VA_ARGS__; -#endif - -// When the Debug mode is enabled, we disable extern declarations because we -// don't want to use the functions compiled in the library, which might not -// have had the debug mode enabled when built. However, some extern declarations -// need to be used, because code correctness depends on it (several instances -// in the ). Those special declarations are declared with -// _LIBCPP_EXTERN_TEMPLATE_EVEN_IN_DEBUG_MODE, which is enabled even -// when the debug mode is enabled. -#ifndef _LIBCPP_EXTERN_TEMPLATE_EVEN_IN_DEBUG_MODE -# define _LIBCPP_EXTERN_TEMPLATE_EVEN_IN_DEBUG_MODE(...) extern template __VA_ARGS__; -#endif - -#ifndef _LIBCPP_EXTERN_TEMPLATE_DEFINE -#define _LIBCPP_EXTERN_TEMPLATE_DEFINE(...) template __VA_ARGS__; +// Libc++ allows disabling extern template instantiation declarations by +// means of users defining _LIBCPP_DISABLE_EXTERN_TEMPLATE. +// +// Furthermore, when the Debug mode is enabled, we disable extern declarations +// when building user code because we don't want to use the functions compiled +// in the library, which might not have had the debug mode enabled when built. +// However, some extern declarations need to be used, because code correctness +// depends on it (several instances in ). Those special declarations +// are declared with _LIBCPP_EXTERN_TEMPLATE_EVEN_IN_DEBUG_MODE, which is enabled +// even when the debug mode is enabled. +#if defined(_LIBCPP_DISABLE_EXTERN_TEMPLATE) +# define _LIBCPP_EXTERN_TEMPLATE(...) /* nothing */ +# define _LIBCPP_EXTERN_TEMPLATE_EVEN_IN_DEBUG_MODE(...) /* nothing */ +#elif _LIBCPP_DEBUG_LEVEL >= 1 && !defined(_LIBCPP_BUILDING_LIBRARY) +# define _LIBCPP_EXTERN_TEMPLATE(...) /* nothing */ +# define _LIBCPP_EXTERN_TEMPLATE_EVEN_IN_DEBUG_MODE(...) extern template __VA_ARGS__; +#else +# define _LIBCPP_EXTERN_TEMPLATE(...) extern template __VA_ARGS__; +# define _LIBCPP_EXTERN_TEMPLATE_EVEN_IN_DEBUG_MODE(...) extern template __VA_ARGS__; #endif #if defined(__APPLE__) || defined(__FreeBSD__) || defined(_LIBCPP_MSVCRT_LIKE) || \ @@ -918,14 +942,7 @@ typedef unsigned int char32_t; #define _LIBCPP_LOCALE__L_EXTENSIONS 1 #endif -#if defined(__unix__) || (defined(__APPLE__) && defined(__MACH__)) -// Most unix variants have catopen. These are the specific ones that don't. -# if !defined(__BIONIC__) && !defined(_NEWLIB_VERSION) -# define _LIBCPP_HAS_CATOPEN 1 -# endif -#endif - -#ifdef __FreeBSD__ +#if defined(__FreeBSD__) || defined(__DragonFly__) #define _DECLARE_C99_LDBL_MATH 1 #endif @@ -948,9 +965,8 @@ typedef unsigned int char32_t; # endif #endif // defined(__APPLE__) -#if !defined(_LIBCPP_HAS_NO_ALIGNED_ALLOCATION) && \ - (defined(_LIBCPP_HAS_NO_LIBRARY_ALIGNED_ALLOCATION) || \ - (!defined(__cpp_aligned_new) || __cpp_aligned_new < 201606)) +#if defined(_LIBCPP_HAS_NO_LIBRARY_ALIGNED_ALLOCATION) || \ + (!defined(__cpp_aligned_new) || __cpp_aligned_new < 201606) # define _LIBCPP_HAS_NO_ALIGNED_ALLOCATION #endif @@ -963,7 +979,7 @@ typedef unsigned int char32_t; #endif #if _LIBCPP_STD_VER <= 17 || !defined(__cpp_char8_t) -#define _LIBCPP_NO_HAS_CHAR8_T +#define _LIBCPP_HAS_NO_CHAR8_T #endif // Deprecation macros. @@ -1006,24 +1022,23 @@ typedef unsigned int char32_t; # define _LIBCPP_DEPRECATED_IN_CXX20 #endif -#if !defined(_LIBCPP_NO_HAS_CHAR8_T) +#if !defined(_LIBCPP_HAS_NO_CHAR8_T) # define _LIBCPP_DEPRECATED_WITH_CHAR8_T _LIBCPP_DEPRECATED #else # define _LIBCPP_DEPRECATED_WITH_CHAR8_T #endif // Macros to enter and leave a state where deprecation warnings are suppressed. -#if !defined(_LIBCPP_SUPPRESS_DEPRECATED_PUSH) && \ - (defined(_LIBCPP_COMPILER_CLANG) || defined(_LIBCPP_COMPILER_GCC)) -# define _LIBCPP_SUPPRESS_DEPRECATED_PUSH \ - _Pragma("GCC diagnostic push") \ - _Pragma("GCC diagnostic ignored \"-Wdeprecated\"") -# define _LIBCPP_SUPPRESS_DEPRECATED_POP \ - _Pragma("GCC diagnostic pop") -#endif -#if !defined(_LIBCPP_SUPPRESS_DEPRECATED_PUSH) -# define _LIBCPP_SUPPRESS_DEPRECATED_PUSH -# define _LIBCPP_SUPPRESS_DEPRECATED_POP +#if defined(_LIBCPP_COMPILER_CLANG_BASED) || defined(_LIBCPP_COMPILER_GCC) +# define _LIBCPP_SUPPRESS_DEPRECATED_PUSH \ + _Pragma("GCC diagnostic push") \ + _Pragma("GCC diagnostic ignored \"-Wdeprecated\"") \ + _Pragma("GCC diagnostic ignored \"-Wdeprecated-declarations\"") +# define _LIBCPP_SUPPRESS_DEPRECATED_POP \ + _Pragma("GCC diagnostic pop") +#else +# define _LIBCPP_SUPPRESS_DEPRECATED_PUSH +# define _LIBCPP_SUPPRESS_DEPRECATED_POP #endif #if _LIBCPP_STD_VER <= 11 @@ -1054,7 +1069,7 @@ typedef unsigned int char32_t; // NODISCARD macros to the correct attribute. #if __has_cpp_attribute(nodiscard) || defined(_LIBCPP_COMPILER_MSVC) # define _LIBCPP_NODISCARD_ATTRIBUTE [[nodiscard]] -#elif defined(_LIBCPP_COMPILER_CLANG) && !defined(_LIBCPP_CXX03_LANG) +#elif defined(_LIBCPP_COMPILER_CLANG_BASED) && !defined(_LIBCPP_CXX03_LANG) # define _LIBCPP_NODISCARD_ATTRIBUTE [[clang::warn_unused_result]] #else // We can't use GCC's [[gnu::warn_unused_result]] and @@ -1084,12 +1099,10 @@ typedef unsigned int char32_t; # define _LIBCPP_INLINE_VAR #endif -#ifndef _LIBCPP_CONSTEXPR_IF_NODEBUG #if defined(_LIBCPP_DEBUG) || defined(_LIBCPP_HAS_NO_CXX14_CONSTEXPR) -#define _LIBCPP_CONSTEXPR_IF_NODEBUG +# define _LIBCPP_CONSTEXPR_IF_NODEBUG #else -#define _LIBCPP_CONSTEXPR_IF_NODEBUG constexpr -#endif +# define _LIBCPP_CONSTEXPR_IF_NODEBUG constexpr #endif #if __has_attribute(no_destroy) @@ -1104,7 +1117,7 @@ extern "C" _LIBCPP_FUNC_VIS void __sanitizer_annotate_contiguous_container( #endif // Try to find out if RTTI is disabled. -#if defined(_LIBCPP_COMPILER_CLANG) && !__has_feature(cxx_rtti) +#if defined(_LIBCPP_COMPILER_CLANG_BASED) && !__has_feature(cxx_rtti) # define _LIBCPP_NO_RTTI #elif defined(__GNUC__) && !defined(__GXX_RTTI) # define _LIBCPP_NO_RTTI @@ -1132,6 +1145,7 @@ extern "C" _LIBCPP_FUNC_VIS void __sanitizer_annotate_contiguous_container( defined(__CloudABI__) || \ defined(__sun__) || \ defined(__MVS__) || \ + defined(_AIX) || \ (defined(__MINGW32__) && __has_include()) # define _LIBCPP_HAS_THREAD_API_PTHREAD # elif defined(__Fuchsia__) @@ -1218,12 +1232,10 @@ extern "C" _LIBCPP_FUNC_VIS void __sanitizer_annotate_contiguous_container( #endif // Some systems do not provide gets() in their C library, for security reasons. -#ifndef _LIBCPP_C_HAS_NO_GETS -# if defined(_LIBCPP_MSVCRT) || \ - (defined(__FreeBSD_version) && __FreeBSD_version >= 1300043) || \ - defined(__OpenBSD__) -# define _LIBCPP_C_HAS_NO_GETS -# endif +#if defined(_LIBCPP_MSVCRT) || \ + (defined(__FreeBSD_version) && __FreeBSD_version >= 1300043) || \ + defined(__OpenBSD__) +# define _LIBCPP_C_HAS_NO_GETS #endif #if defined(__BIONIC__) || defined(__CloudABI__) || defined(__NuttX__) || \ @@ -1273,13 +1285,11 @@ extern "C" _LIBCPP_FUNC_VIS void __sanitizer_annotate_contiguous_container( # endif #endif -#ifndef _LIBCPP_THREAD_SAFETY_ANNOTATION -# ifdef _LIBCPP_HAS_THREAD_SAFETY_ANNOTATIONS -# define _LIBCPP_THREAD_SAFETY_ANNOTATION(x) __attribute__((x)) -# else -# define _LIBCPP_THREAD_SAFETY_ANNOTATION(x) -# endif -#endif // _LIBCPP_THREAD_SAFETY_ANNOTATION +#ifdef _LIBCPP_HAS_THREAD_SAFETY_ANNOTATIONS +# define _LIBCPP_THREAD_SAFETY_ANNOTATION(x) __attribute__((x)) +#else +# define _LIBCPP_THREAD_SAFETY_ANNOTATION(x) +#endif #if __has_attribute(require_constant_initialization) # define _LIBCPP_SAFE_STATIC __attribute__((__require_constant_initialization__)) @@ -1295,12 +1305,6 @@ extern "C" _LIBCPP_FUNC_VIS void __sanitizer_annotate_contiguous_container( #define _LIBCPP_HAS_NO_BUILTIN_IS_CONSTANT_EVALUATED #endif -#if !defined(_LIBCPP_HAS_NO_OFF_T_FUNCTIONS) -# if defined(_LIBCPP_MSVCRT) || defined(_NEWLIB_VERSION) -# define _LIBCPP_HAS_NO_OFF_T_FUNCTIONS -# endif -#endif - #if __has_attribute(diagnose_if) && !defined(_LIBCPP_DISABLE_ADDITIONAL_DIAGNOSTICS) # define _LIBCPP_DIAGNOSE_WARNING(...) \ __attribute__((diagnose_if(__VA_ARGS__, "warning"))) @@ -1328,14 +1332,17 @@ extern "C" _LIBCPP_FUNC_VIS void __sanitizer_annotate_contiguous_container( #define _LIBCPP_NODEBUG #endif -#ifndef _LIBCPP_NODEBUG_TYPE -#if __has_attribute(__nodebug__) && \ - (defined(_LIBCPP_CLANG_VER) && _LIBCPP_CLANG_VER >= 900) -#define _LIBCPP_NODEBUG_TYPE __attribute__((nodebug)) +#if __has_attribute(__nodebug__) && (defined(_LIBCPP_CLANG_VER) && _LIBCPP_CLANG_VER >= 900) +# define _LIBCPP_NODEBUG_TYPE __attribute__((nodebug)) #else -#define _LIBCPP_NODEBUG_TYPE +# define _LIBCPP_NODEBUG_TYPE +#endif + +#if __has_attribute(__standalone_debug__) +#define _LIBCPP_STANDALONE_DEBUG __attribute__((__standalone_debug__)) +#else +#define _LIBCPP_STANDALONE_DEBUG #endif -#endif // !defined(_LIBCPP_NODEBUG_TYPE) #if __has_attribute(__preferred_name__) #define _LIBCPP_PREFERRED_NAME(x) __attribute__((__preferred_name__(x))) @@ -1352,11 +1359,19 @@ extern "C" _LIBCPP_FUNC_VIS void __sanitizer_annotate_contiguous_container( #if defined(_LIBCPP_ENABLE_CXX17_REMOVED_FEATURES) #define _LIBCPP_ENABLE_CXX17_REMOVED_AUTO_PTR -#define _LIBCPP_ENABLE_CXX17_REMOVED_UNEXPECTED_FUNCTIONS -#define _LIBCPP_ENABLE_CXX17_REMOVED_RANDOM_SHUFFLE #define _LIBCPP_ENABLE_CXX17_REMOVED_BINDERS +#define _LIBCPP_ENABLE_CXX17_REMOVED_RANDOM_SHUFFLE +#define _LIBCPP_ENABLE_CXX17_REMOVED_UNEXPECTED_FUNCTIONS #endif // _LIBCPP_ENABLE_CXX17_REMOVED_FEATURES +#if defined(_LIBCPP_ENABLE_CXX20_REMOVED_FEATURES) +#define _LIBCPP_ENABLE_CXX20_REMOVED_ALLOCATOR_MEMBERS +#define _LIBCPP_ENABLE_CXX20_REMOVED_BINDER_TYPEDEFS +#define _LIBCPP_ENABLE_CXX20_REMOVED_NEGATORS +#define _LIBCPP_ENABLE_CXX20_REMOVED_RAW_STORAGE_ITERATOR +#define _LIBCPP_ENABLE_CXX20_REMOVED_TYPE_TRAITS +#endif // _LIBCPP_ENABLE_CXX20_REMOVED_FEATURES + #if !defined(__cpp_deduction_guides) || __cpp_deduction_guides < 201611 #define _LIBCPP_HAS_NO_DEDUCTION_GUIDES #endif @@ -1405,7 +1420,7 @@ extern "C" _LIBCPP_FUNC_VIS void __sanitizer_annotate_contiguous_container( #ifndef _LIBCPP_NO_AUTO_LINK # if defined(_LIBCPP_ABI_MICROSOFT) && !defined(_LIBCPP_BUILDING_LIBRARY) -# if defined(_DLL) +# if !defined(_LIBCPP_DISABLE_VISIBILITY_ANNOTATIONS) # pragma comment(lib, "c++.lib") # else # pragma comment(lib, "libc++.lib") @@ -1413,8 +1428,6 @@ extern "C" _LIBCPP_FUNC_VIS void __sanitizer_annotate_contiguous_container( # endif // defined(_LIBCPP_ABI_MICROSOFT) && !defined(_LIBCPP_BUILDING_LIBRARY) #endif // _LIBCPP_NO_AUTO_LINK -#define _LIBCPP_UNUSED_VAR(x) ((void)(x)) - // Configures the fopen close-on-exec mode character, if any. This string will // be appended to any mode string used by fstream for fopen/fdopen. // @@ -1445,6 +1458,13 @@ extern "C" _LIBCPP_FUNC_VIS void __sanitizer_annotate_contiguous_container( # define _LIBCPP_INIT_PRIORITY_MAX #endif +#if defined(__GNUC__) || defined(__clang__) +#define _LIBCPP_FORMAT_PRINTF(a, b) \ + __attribute__((__format__(__printf__, a, b))) +#else +#define _LIBCPP_FORMAT_PRINTF(a, b) +#endif + #endif // __cplusplus #endif // _LIBCPP_CONFIG diff --git a/lib/libcxx/include/__config_site.in b/lib/libcxx/include/__config_site.in deleted file mode 100644 index ec4d410bb..000000000 --- a/lib/libcxx/include/__config_site.in +++ /dev/null @@ -1,40 +0,0 @@ -//===----------------------------------------------------------------------===// -// -// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. -// See https://llvm.org/LICENSE.txt for license information. -// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception -// -//===----------------------------------------------------------------------===// - -#ifndef _LIBCPP_CONFIG_SITE -#define _LIBCPP_CONFIG_SITE - -#cmakedefine _LIBCPP_ABI_VERSION @_LIBCPP_ABI_VERSION@ -#cmakedefine _LIBCPP_ABI_UNSTABLE -#cmakedefine _LIBCPP_ABI_FORCE_ITANIUM -#cmakedefine _LIBCPP_ABI_FORCE_MICROSOFT -#cmakedefine _LIBCPP_HIDE_FROM_ABI_PER_TU_BY_DEFAULT -#cmakedefine _LIBCPP_HAS_NO_GLOBAL_FILESYSTEM_NAMESPACE -#cmakedefine _LIBCPP_HAS_NO_STDIN -#cmakedefine _LIBCPP_HAS_NO_STDOUT -#cmakedefine _LIBCPP_HAS_NO_THREADS -#cmakedefine _LIBCPP_HAS_NO_MONOTONIC_CLOCK -#cmakedefine _LIBCPP_HAS_NO_THREAD_UNSAFE_C_FUNCTIONS -#cmakedefine _LIBCPP_HAS_MUSL_LIBC -#cmakedefine _LIBCPP_HAS_THREAD_API_PTHREAD -#cmakedefine _LIBCPP_HAS_THREAD_API_EXTERNAL -#cmakedefine _LIBCPP_HAS_THREAD_API_WIN32 -#cmakedefine _LIBCPP_HAS_THREAD_LIBRARY_EXTERNAL -#cmakedefine _LIBCPP_DISABLE_VISIBILITY_ANNOTATIONS -#cmakedefine _LIBCPP_HAS_NO_VENDOR_AVAILABILITY_ANNOTATIONS -#cmakedefine _LIBCPP_NO_VCRUNTIME -#cmakedefine _LIBCPP_TYPEINFO_COMPARISON_IMPLEMENTATION @_LIBCPP_TYPEINFO_COMPARISON_IMPLEMENTATION@ -#cmakedefine _LIBCPP_ABI_NAMESPACE @_LIBCPP_ABI_NAMESPACE@ -#cmakedefine _LIBCPP_HAS_NO_FILESYSTEM_LIBRARY -#cmakedefine _LIBCPP_HAS_PARALLEL_ALGORITHMS -#cmakedefine _LIBCPP_HAS_NO_RANDOM_DEVICE -#cmakedefine _LIBCPP_HAS_NO_LOCALIZATION - -@_LIBCPP_ABI_DEFINES@ - -#endif // _LIBCPP_CONFIG_SITE diff --git a/lib/libcxx/include/__debug b/lib/libcxx/include/__debug index 7b5bfb3f8..771e43163 100644 --- a/lib/libcxx/include/__debug +++ b/lib/libcxx/include/__debug @@ -270,4 +270,4 @@ _LIBCPP_FUNC_VIS const __libcpp_db* __get_const_db(); _LIBCPP_END_NAMESPACE_STD -#endif // _LIBCPP_DEBUG_H +#endif // _LIBCPP_DEBUG_H diff --git a/lib/libcxx/include/__errc b/lib/libcxx/include/__errc index a8ad29f36..81da2e197 100644 --- a/lib/libcxx/include/__errc +++ b/lib/libcxx/include/__errc @@ -214,4 +214,4 @@ _LIBCPP_DECLARE_STRONG_ENUM_EPILOG(errc) _LIBCPP_END_NAMESPACE_STD -#endif // _LIBCPP___ERRC +#endif // _LIBCPP___ERRC diff --git a/lib/libcxx/include/__format/format_error.h b/lib/libcxx/include/__format/format_error.h new file mode 100644 index 000000000..f983d0ce4 --- /dev/null +++ b/lib/libcxx/include/__format/format_error.h @@ -0,0 +1,56 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___FORMAT_FORMAT_ERROR_H +#define _LIBCPP___FORMAT_FORMAT_ERROR_H + +#include <__config> +#include + +#ifdef _LIBCPP_NO_EXCEPTIONS +#include +#endif + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +#if _LIBCPP_STD_VER > 17 + +class _LIBCPP_EXCEPTION_ABI format_error : public runtime_error { +public: + _LIBCPP_HIDE_FROM_ABI explicit format_error(const string& __s) + : runtime_error(__s) {} + _LIBCPP_HIDE_FROM_ABI explicit format_error(const char* __s) + : runtime_error(__s) {} + virtual ~format_error() noexcept; +}; + +_LIBCPP_NORETURN inline _LIBCPP_HIDE_FROM_ABI void +__throw_format_error(const char* __s) { +#ifndef _LIBCPP_NO_EXCEPTIONS + throw format_error(__s); +#else + (void)__s; + _VSTD::abort(); +#endif +} + +#endif //_LIBCPP_STD_VER > 17 + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___FORMAT_FORMAT_ERROR_H diff --git a/lib/libcxx/include/__format/format_parse_context.h b/lib/libcxx/include/__format/format_parse_context.h new file mode 100644 index 000000000..db39c1b54 --- /dev/null +++ b/lib/libcxx/include/__format/format_parse_context.h @@ -0,0 +1,113 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___FORMAT_FORMAT_PARSE_CONTEXT_H +#define _LIBCPP___FORMAT_FORMAT_PARSE_CONTEXT_H + +#include <__config> +#include <__format/format_error.h> +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +#if _LIBCPP_STD_VER > 17 + +// TODO FMT Remove this once we require compilers with proper C++20 support. +// If the compiler has no concepts support, the format header will be disabled. +// Without concepts support enable_if needs to be used and that too much effort +// to support compilers with partial C++20 support. +#if !defined(_LIBCPP_HAS_NO_CONCEPTS) && \ + !defined(_LIBCPP_HAS_NO_BUILTIN_IS_CONSTANT_EVALUATED) + +template +class _LIBCPP_TEMPLATE_VIS _LIBCPP_AVAILABILITY_FORMAT basic_format_parse_context { +public: + using char_type = _CharT; + using const_iterator = typename basic_string_view<_CharT>::const_iterator; + using iterator = const_iterator; + + _LIBCPP_HIDE_FROM_ABI + constexpr explicit basic_format_parse_context(basic_string_view<_CharT> __fmt, + size_t __num_args = 0) noexcept + : __begin_(__fmt.begin()), + __end_(__fmt.end()), + __indexing_(__unknown), + __next_arg_id_(0), + __num_args_(__num_args) {} + + basic_format_parse_context(const basic_format_parse_context&) = delete; + basic_format_parse_context& + operator=(const basic_format_parse_context&) = delete; + + _LIBCPP_HIDE_FROM_ABI constexpr const_iterator begin() const noexcept { + return __begin_; + } + _LIBCPP_HIDE_FROM_ABI constexpr const_iterator end() const noexcept { + return __end_; + } + _LIBCPP_HIDE_FROM_ABI constexpr void advance_to(const_iterator __it) { + __begin_ = __it; + } + + _LIBCPP_HIDE_FROM_ABI constexpr size_t next_arg_id() { + if (__indexing_ == __manual) + __throw_format_error("Using automatic argument numbering in manual " + "argument numbering mode"); + + if (__indexing_ == __unknown) + __indexing_ = __automatic; + return __next_arg_id_++; + } + _LIBCPP_HIDE_FROM_ABI constexpr void check_arg_id(size_t __id) { + if (__indexing_ == __automatic) + __throw_format_error("Using manual argument numbering in automatic " + "argument numbering mode"); + + if (__indexing_ == __unknown) + __indexing_ = __manual; + + // Throws an exception to make the expression a non core constant + // expression as required by: + // [format.parse.ctx]/11 + // Remarks: Call expressions where id >= num_args_ are not core constant + // expressions ([expr.const]). + // Note: the Throws clause [format.parse.ctx]/10 doesn't specify the + // behavior when id >= num_args_. + if (is_constant_evaluated() && __id >= __num_args_) + __throw_format_error("Argument index outside the valid range"); + } + +private: + iterator __begin_; + iterator __end_; + enum _Indexing { __unknown, __manual, __automatic }; + _Indexing __indexing_; + size_t __next_arg_id_; + size_t __num_args_; +}; + +using format_parse_context = basic_format_parse_context; +using wformat_parse_context = basic_format_parse_context; + +#endif // !defined(_LIBCPP_HAS_NO_CONCEPTS) && !defined(_LIBCPP_HAS_NO_BUILTIN_IS_CONSTANT_EVALUATED) + +#endif //_LIBCPP_STD_VER > 17 + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___FORMAT_FORMAT_PARSE_CONTEXT_H diff --git a/lib/libcxx/include/__function_like.h b/lib/libcxx/include/__function_like.h new file mode 100644 index 000000000..8a3597bac --- /dev/null +++ b/lib/libcxx/include/__function_like.h @@ -0,0 +1,56 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ITERATOR_FUNCTION_LIKE_H +#define _LIBCPP___ITERATOR_FUNCTION_LIKE_H + +#include <__config> + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +#if !defined(_LIBCPP_HAS_NO_RANGES) + +namespace ranges { +// Per [range.iter.ops.general] and [algorithms.requirements], functions in namespace std::ranges +// can't be found by ADL and inhibit ADL when found by unqualified lookup. The easiest way to +// facilitate this is to use function objects. +// +// Since these are still standard library functions, we use `__function_like` to eliminate most of +// the properties that function objects get by default (e.g. semiregularity, addressability), to +// limit the surface area of the unintended public interface, so as to curb the effect of Hyrum's +// law. +struct __function_like { + __function_like() = delete; + __function_like(__function_like const&) = delete; + __function_like& operator=(__function_like const&) = delete; + + void operator&() const = delete; + + struct __tag { }; + +protected: + constexpr explicit __function_like(__tag) noexcept {} + ~__function_like() = default; +}; +} // namespace ranges + +#endif // !defined(_LIBCPP_HAS_NO_RANGES) + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ITERATOR_FUNCTION_LIKE_H diff --git a/lib/libcxx/include/__functional/binary_function.h b/lib/libcxx/include/__functional/binary_function.h new file mode 100644 index 000000000..8ca7b0666 --- /dev/null +++ b/lib/libcxx/include/__functional/binary_function.h @@ -0,0 +1,31 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___FUNCTIONAL_BINARY_FUNCTION_H +#define _LIBCPP___FUNCTIONAL_BINARY_FUNCTION_H + +#include <__config> + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +struct _LIBCPP_TEMPLATE_VIS binary_function +{ + typedef _Arg1 first_argument_type; + typedef _Arg2 second_argument_type; + typedef _Result result_type; +}; + +_LIBCPP_END_NAMESPACE_STD + +#endif // _LIBCPP___FUNCTIONAL_BINARY_FUNCTION_H diff --git a/lib/libcxx/include/__functional/binary_negate.h b/lib/libcxx/include/__functional/binary_negate.h new file mode 100644 index 000000000..4fc3f1ba2 --- /dev/null +++ b/lib/libcxx/include/__functional/binary_negate.h @@ -0,0 +1,50 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___FUNCTIONAL_BINARY_NEGATE_H +#define _LIBCPP___FUNCTIONAL_BINARY_NEGATE_H + +#include <__config> +#include <__functional/binary_function.h> + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_BEGIN_NAMESPACE_STD + +#if _LIBCPP_STD_VER <= 17 || defined(_LIBCPP_ENABLE_CXX20_REMOVED_NEGATORS) + +template +class _LIBCPP_TEMPLATE_VIS _LIBCPP_DEPRECATED_IN_CXX17 binary_negate + : public binary_function +{ + _Predicate __pred_; +public: + _LIBCPP_INLINE_VISIBILITY explicit _LIBCPP_CONSTEXPR_AFTER_CXX11 + binary_negate(const _Predicate& __pred) : __pred_(__pred) {} + + _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY + bool operator()(const typename _Predicate::first_argument_type& __x, + const typename _Predicate::second_argument_type& __y) const + {return !__pred_(__x, __y);} +}; + +template +_LIBCPP_DEPRECATED_IN_CXX17 inline _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY +binary_negate<_Predicate> +not2(const _Predicate& __pred) {return binary_negate<_Predicate>(__pred);} + +#endif // _LIBCPP_STD_VER <= 17 || defined(_LIBCPP_ENABLE_CXX20_REMOVED_NEGATORS) + +_LIBCPP_END_NAMESPACE_STD + +#endif // _LIBCPP___FUNCTIONAL_BINARY_NEGATE_H diff --git a/lib/libcxx/include/__functional/bind.h b/lib/libcxx/include/__functional/bind.h new file mode 100644 index 000000000..79dfad723 --- /dev/null +++ b/lib/libcxx/include/__functional/bind.h @@ -0,0 +1,386 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___FUNCTIONAL_BIND_H +#define _LIBCPP___FUNCTIONAL_BIND_H + +#include <__config> +#include <__functional/weak_result_type.h> +#include <__functional/invoke.h> +#include +#include +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_BEGIN_NAMESPACE_STD + +template struct __is_bind_expression : public false_type {}; +template struct _LIBCPP_TEMPLATE_VIS is_bind_expression + : public __is_bind_expression::type> {}; + +#if _LIBCPP_STD_VER > 14 +template +_LIBCPP_INLINE_VAR constexpr size_t is_bind_expression_v = is_bind_expression<_Tp>::value; +#endif + +template struct __is_placeholder : public integral_constant {}; +template struct _LIBCPP_TEMPLATE_VIS is_placeholder + : public __is_placeholder::type> {}; + +#if _LIBCPP_STD_VER > 14 +template +_LIBCPP_INLINE_VAR constexpr size_t is_placeholder_v = is_placeholder<_Tp>::value; +#endif + +namespace placeholders +{ + +template struct __ph {}; + +#if defined(_LIBCPP_CXX03_LANG) || defined(_LIBCPP_BUILDING_LIBRARY) +_LIBCPP_FUNC_VIS extern const __ph<1> _1; +_LIBCPP_FUNC_VIS extern const __ph<2> _2; +_LIBCPP_FUNC_VIS extern const __ph<3> _3; +_LIBCPP_FUNC_VIS extern const __ph<4> _4; +_LIBCPP_FUNC_VIS extern const __ph<5> _5; +_LIBCPP_FUNC_VIS extern const __ph<6> _6; +_LIBCPP_FUNC_VIS extern const __ph<7> _7; +_LIBCPP_FUNC_VIS extern const __ph<8> _8; +_LIBCPP_FUNC_VIS extern const __ph<9> _9; +_LIBCPP_FUNC_VIS extern const __ph<10> _10; +#else +/* _LIBCPP_INLINE_VAR */ constexpr __ph<1> _1{}; +/* _LIBCPP_INLINE_VAR */ constexpr __ph<2> _2{}; +/* _LIBCPP_INLINE_VAR */ constexpr __ph<3> _3{}; +/* _LIBCPP_INLINE_VAR */ constexpr __ph<4> _4{}; +/* _LIBCPP_INLINE_VAR */ constexpr __ph<5> _5{}; +/* _LIBCPP_INLINE_VAR */ constexpr __ph<6> _6{}; +/* _LIBCPP_INLINE_VAR */ constexpr __ph<7> _7{}; +/* _LIBCPP_INLINE_VAR */ constexpr __ph<8> _8{}; +/* _LIBCPP_INLINE_VAR */ constexpr __ph<9> _9{}; +/* _LIBCPP_INLINE_VAR */ constexpr __ph<10> _10{}; +#endif // defined(_LIBCPP_CXX03_LANG) || defined(_LIBCPP_BUILDING_LIBRARY) + +} // placeholders + +template +struct __is_placeholder > + : public integral_constant {}; + + +#ifndef _LIBCPP_CXX03_LANG + +template +inline _LIBCPP_INLINE_VISIBILITY +_Tp& +__mu(reference_wrapper<_Tp> __t, _Uj&) +{ + return __t.get(); +} + +template +inline _LIBCPP_INLINE_VISIBILITY +typename __invoke_of<_Ti&, _Uj...>::type +__mu_expand(_Ti& __ti, tuple<_Uj...>& __uj, __tuple_indices<_Indx...>) +{ + return __ti(_VSTD::forward<_Uj>(_VSTD::get<_Indx>(__uj))...); +} + +template +inline _LIBCPP_INLINE_VISIBILITY +typename _EnableIf +< + is_bind_expression<_Ti>::value, + __invoke_of<_Ti&, _Uj...> +>::type +__mu(_Ti& __ti, tuple<_Uj...>& __uj) +{ + typedef typename __make_tuple_indices::type __indices; + return _VSTD::__mu_expand(__ti, __uj, __indices()); +} + +template +struct __mu_return2 {}; + +template +struct __mu_return2 +{ + typedef typename tuple_element::value - 1, _Uj>::type type; +}; + +template +inline _LIBCPP_INLINE_VISIBILITY +typename enable_if +< + 0 < is_placeholder<_Ti>::value, + typename __mu_return2<0 < is_placeholder<_Ti>::value, _Ti, _Uj>::type +>::type +__mu(_Ti&, _Uj& __uj) +{ + const size_t _Indx = is_placeholder<_Ti>::value - 1; + return _VSTD::forward::type>(_VSTD::get<_Indx>(__uj)); +} + +template +inline _LIBCPP_INLINE_VISIBILITY +typename enable_if +< + !is_bind_expression<_Ti>::value && + is_placeholder<_Ti>::value == 0 && + !__is_reference_wrapper<_Ti>::value, + _Ti& +>::type +__mu(_Ti& __ti, _Uj&) +{ + return __ti; +} + +template +struct __mu_return_impl; + +template +struct __mu_return_invokable // false +{ + typedef __nat type; +}; + +template +struct __mu_return_invokable +{ + typedef typename __invoke_of<_Ti&, _Uj...>::type type; +}; + +template +struct __mu_return_impl<_Ti, false, true, false, tuple<_Uj...> > + : public __mu_return_invokable<__invokable<_Ti&, _Uj...>::value, _Ti, _Uj...> +{ +}; + +template +struct __mu_return_impl<_Ti, false, false, true, _TupleUj> +{ + typedef typename tuple_element::value - 1, + _TupleUj>::type&& type; +}; + +template +struct __mu_return_impl<_Ti, true, false, false, _TupleUj> +{ + typedef typename _Ti::type& type; +}; + +template +struct __mu_return_impl<_Ti, false, false, false, _TupleUj> +{ + typedef _Ti& type; +}; + +template +struct __mu_return + : public __mu_return_impl<_Ti, + __is_reference_wrapper<_Ti>::value, + is_bind_expression<_Ti>::value, + 0 < is_placeholder<_Ti>::value && + is_placeholder<_Ti>::value <= tuple_size<_TupleUj>::value, + _TupleUj> +{ +}; + +template +struct __is_valid_bind_return +{ + static const bool value = false; +}; + +template +struct __is_valid_bind_return<_Fp, tuple<_BoundArgs...>, _TupleUj> +{ + static const bool value = __invokable<_Fp, + typename __mu_return<_BoundArgs, _TupleUj>::type...>::value; +}; + +template +struct __is_valid_bind_return<_Fp, const tuple<_BoundArgs...>, _TupleUj> +{ + static const bool value = __invokable<_Fp, + typename __mu_return::type...>::value; +}; + +template ::value> +struct __bind_return; + +template +struct __bind_return<_Fp, tuple<_BoundArgs...>, _TupleUj, true> +{ + typedef typename __invoke_of + < + _Fp&, + typename __mu_return + < + _BoundArgs, + _TupleUj + >::type... + >::type type; +}; + +template +struct __bind_return<_Fp, const tuple<_BoundArgs...>, _TupleUj, true> +{ + typedef typename __invoke_of + < + _Fp&, + typename __mu_return + < + const _BoundArgs, + _TupleUj + >::type... + >::type type; +}; + +template +inline _LIBCPP_INLINE_VISIBILITY +typename __bind_return<_Fp, _BoundArgs, _Args>::type +__apply_functor(_Fp& __f, _BoundArgs& __bound_args, __tuple_indices<_Indx...>, + _Args&& __args) +{ + return _VSTD::__invoke(__f, _VSTD::__mu(_VSTD::get<_Indx>(__bound_args), __args)...); +} + +template +class __bind +#if _LIBCPP_STD_VER <= 17 || !defined(_LIBCPP_ABI_NO_BINDER_BASES) + : public __weak_result_type::type> +#endif +{ +protected: + typedef typename decay<_Fp>::type _Fd; + typedef tuple::type...> _Td; +private: + _Fd __f_; + _Td __bound_args_; + + typedef typename __make_tuple_indices::type __indices; +public: + template ::value && + !is_same::type, + __bind>::value + >::type> + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 + explicit __bind(_Gp&& __f, _BA&& ...__bound_args) + : __f_(_VSTD::forward<_Gp>(__f)), + __bound_args_(_VSTD::forward<_BA>(__bound_args)...) {} + + template + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 + typename __bind_return<_Fd, _Td, tuple<_Args&&...> >::type + operator()(_Args&& ...__args) + { + return _VSTD::__apply_functor(__f_, __bound_args_, __indices(), + tuple<_Args&&...>(_VSTD::forward<_Args>(__args)...)); + } + + template + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 + typename __bind_return >::type + operator()(_Args&& ...__args) const + { + return _VSTD::__apply_functor(__f_, __bound_args_, __indices(), + tuple<_Args&&...>(_VSTD::forward<_Args>(__args)...)); + } +}; + +template +struct __is_bind_expression<__bind<_Fp, _BoundArgs...> > : public true_type {}; + +template +class __bind_r + : public __bind<_Fp, _BoundArgs...> +{ + typedef __bind<_Fp, _BoundArgs...> base; + typedef typename base::_Fd _Fd; + typedef typename base::_Td _Td; +public: + typedef _Rp result_type; + + + template ::value && + !is_same::type, + __bind_r>::value + >::type> + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 + explicit __bind_r(_Gp&& __f, _BA&& ...__bound_args) + : base(_VSTD::forward<_Gp>(__f), + _VSTD::forward<_BA>(__bound_args)...) {} + + template + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 + typename enable_if + < + is_convertible >::type, + result_type>::value || is_void<_Rp>::value, + result_type + >::type + operator()(_Args&& ...__args) + { + typedef __invoke_void_return_wrapper<_Rp> _Invoker; + return _Invoker::__call(static_cast(*this), _VSTD::forward<_Args>(__args)...); + } + + template + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 + typename enable_if + < + is_convertible >::type, + result_type>::value || is_void<_Rp>::value, + result_type + >::type + operator()(_Args&& ...__args) const + { + typedef __invoke_void_return_wrapper<_Rp> _Invoker; + return _Invoker::__call(static_cast(*this), _VSTD::forward<_Args>(__args)...); + } +}; + +template +struct __is_bind_expression<__bind_r<_Rp, _Fp, _BoundArgs...> > : public true_type {}; + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +__bind<_Fp, _BoundArgs...> +bind(_Fp&& __f, _BoundArgs&&... __bound_args) +{ + typedef __bind<_Fp, _BoundArgs...> type; + return type(_VSTD::forward<_Fp>(__f), _VSTD::forward<_BoundArgs>(__bound_args)...); +} + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +__bind_r<_Rp, _Fp, _BoundArgs...> +bind(_Fp&& __f, _BoundArgs&&... __bound_args) +{ + typedef __bind_r<_Rp, _Fp, _BoundArgs...> type; + return type(_VSTD::forward<_Fp>(__f), _VSTD::forward<_BoundArgs>(__bound_args)...); +} + +#endif // _LIBCPP_CXX03_LANG + +_LIBCPP_END_NAMESPACE_STD + +#endif // _LIBCPP___FUNCTIONAL_BIND_H diff --git a/lib/libcxx/include/__functional/bind_front.h b/lib/libcxx/include/__functional/bind_front.h new file mode 100644 index 000000000..8690499f2 --- /dev/null +++ b/lib/libcxx/include/__functional/bind_front.h @@ -0,0 +1,52 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___FUNCTIONAL_BIND_FRONT_H +#define _LIBCPP___FUNCTIONAL_BIND_FRONT_H + +#include <__config> +#include <__functional/perfect_forward.h> +#include <__functional/invoke.h> +#include +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_BEGIN_NAMESPACE_STD + +#if _LIBCPP_STD_VER > 17 + +struct __bind_front_op +{ + template + constexpr static auto __call(_Args&&... __args) + noexcept(noexcept(_VSTD::invoke(_VSTD::forward<_Args>(__args)...))) + -> decltype( _VSTD::invoke(_VSTD::forward<_Args>(__args)...)) + { return _VSTD::invoke(_VSTD::forward<_Args>(__args)...); } +}; + +template, _Fn>, + is_move_constructible>, + is_constructible, _Args>..., + is_move_constructible>... + >::value>> +constexpr auto bind_front(_Fn&& __f, _Args&&... __args) +{ + return __perfect_forward<__bind_front_op, _Fn, _Args...>(_VSTD::forward<_Fn>(__f), + _VSTD::forward<_Args>(__args)...); +} + +#endif // _LIBCPP_STD_VER > 17 + +_LIBCPP_END_NAMESPACE_STD + +#endif // _LIBCPP___FUNCTIONAL_BIND_FRONT_H diff --git a/lib/libcxx/include/__functional/binder1st.h b/lib/libcxx/include/__functional/binder1st.h new file mode 100644 index 000000000..5dd8f5cf0 --- /dev/null +++ b/lib/libcxx/include/__functional/binder1st.h @@ -0,0 +1,54 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___FUNCTIONAL_BINDER1ST_H +#define _LIBCPP___FUNCTIONAL_BINDER1ST_H + +#include <__config> +#include <__functional/unary_function.h> + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_BEGIN_NAMESPACE_STD + +#if _LIBCPP_STD_VER <= 14 || defined(_LIBCPP_ENABLE_CXX17_REMOVED_BINDERS) + +template +class _LIBCPP_TEMPLATE_VIS _LIBCPP_DEPRECATED_IN_CXX11 binder1st + : public unary_function +{ +protected: + __Operation op; + typename __Operation::first_argument_type value; +public: + _LIBCPP_INLINE_VISIBILITY binder1st(const __Operation& __x, + const typename __Operation::first_argument_type __y) + : op(__x), value(__y) {} + _LIBCPP_INLINE_VISIBILITY typename __Operation::result_type operator() + (typename __Operation::second_argument_type& __x) const + {return op(value, __x);} + _LIBCPP_INLINE_VISIBILITY typename __Operation::result_type operator() + (const typename __Operation::second_argument_type& __x) const + {return op(value, __x);} +}; + +template +_LIBCPP_DEPRECATED_IN_CXX11 inline _LIBCPP_INLINE_VISIBILITY +binder1st<__Operation> +bind1st(const __Operation& __op, const _Tp& __x) + {return binder1st<__Operation>(__op, __x);} + +#endif // _LIBCPP_STD_VER <= 14 || defined(_LIBCPP_ENABLE_CXX17_REMOVED_BINDERS) + +_LIBCPP_END_NAMESPACE_STD + +#endif // _LIBCPP___FUNCTIONAL_BINDER1ST_H diff --git a/lib/libcxx/include/__functional/binder2nd.h b/lib/libcxx/include/__functional/binder2nd.h new file mode 100644 index 000000000..3ed5f5bf4 --- /dev/null +++ b/lib/libcxx/include/__functional/binder2nd.h @@ -0,0 +1,54 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___FUNCTIONAL_BINDER2ND_H +#define _LIBCPP___FUNCTIONAL_BINDER2ND_H + +#include <__config> +#include <__functional/unary_function.h> + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_BEGIN_NAMESPACE_STD + +#if _LIBCPP_STD_VER <= 14 || defined(_LIBCPP_ENABLE_CXX17_REMOVED_BINDERS) + +template +class _LIBCPP_TEMPLATE_VIS _LIBCPP_DEPRECATED_IN_CXX11 binder2nd + : public unary_function +{ +protected: + __Operation op; + typename __Operation::second_argument_type value; +public: + _LIBCPP_INLINE_VISIBILITY + binder2nd(const __Operation& __x, const typename __Operation::second_argument_type __y) + : op(__x), value(__y) {} + _LIBCPP_INLINE_VISIBILITY typename __Operation::result_type operator() + ( typename __Operation::first_argument_type& __x) const + {return op(__x, value);} + _LIBCPP_INLINE_VISIBILITY typename __Operation::result_type operator() + (const typename __Operation::first_argument_type& __x) const + {return op(__x, value);} +}; + +template +_LIBCPP_DEPRECATED_IN_CXX11 inline _LIBCPP_INLINE_VISIBILITY +binder2nd<__Operation> +bind2nd(const __Operation& __op, const _Tp& __x) + {return binder2nd<__Operation>(__op, __x);} + +#endif // _LIBCPP_STD_VER <= 14 || defined(_LIBCPP_ENABLE_CXX17_REMOVED_BINDERS) + +_LIBCPP_END_NAMESPACE_STD + +#endif // _LIBCPP___FUNCTIONAL_BINDER2ND_H diff --git a/lib/libcxx/include/__functional/default_searcher.h b/lib/libcxx/include/__functional/default_searcher.h new file mode 100644 index 000000000..1acbc1883 --- /dev/null +++ b/lib/libcxx/include/__functional/default_searcher.h @@ -0,0 +1,56 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___FUNCTIONAL_DEFAULT_SEARCHER_H +#define _LIBCPP___FUNCTIONAL_DEFAULT_SEARCHER_H + +#include <__algorithm/search.h> +#include <__config> +#include <__functional/operations.h> +#include <__iterator/iterator_traits.h> +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_BEGIN_NAMESPACE_STD + +#if _LIBCPP_STD_VER > 14 + +// default searcher +template> +class _LIBCPP_TEMPLATE_VIS default_searcher { +public: + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 + default_searcher(_ForwardIterator __f, _ForwardIterator __l, + _BinaryPredicate __p = _BinaryPredicate()) + : __first_(__f), __last_(__l), __pred_(__p) {} + + template + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 + pair<_ForwardIterator2, _ForwardIterator2> + operator () (_ForwardIterator2 __f, _ForwardIterator2 __l) const + { + return _VSTD::__search(__f, __l, __first_, __last_, __pred_, + typename iterator_traits<_ForwardIterator>::iterator_category(), + typename iterator_traits<_ForwardIterator2>::iterator_category()); + } + +private: + _ForwardIterator __first_; + _ForwardIterator __last_; + _BinaryPredicate __pred_; + }; + +#endif // _LIBCPP_STD_VER > 14 + +_LIBCPP_END_NAMESPACE_STD + +#endif // _LIBCPP___FUNCTIONAL_DEFAULT_SEARCHER_H diff --git a/lib/libcxx/include/__functional_03 b/lib/libcxx/include/__functional/function.h similarity index 52% rename from lib/libcxx/include/__functional_03 rename to lib/libcxx/include/__functional/function.h index 961648061..ba629e1d1 100644 --- a/lib/libcxx/include/__functional_03 +++ b/lib/libcxx/include/__functional/function.h @@ -7,15 +7,1229 @@ // //===----------------------------------------------------------------------===// -#ifndef _LIBCPP_FUNCTIONAL_03 -#define _LIBCPP_FUNCTIONAL_03 - -// manual variadic expansion for +#ifndef _LIBCPP___FUNCTIONAL_FUNCTION_H +#define _LIBCPP___FUNCTIONAL_FUNCTION_H + +#include <__config> +#include <__functional/binary_function.h> +#include <__functional/invoke.h> +#include <__functional/unary_function.h> +#include <__iterator/iterator_traits.h> +#include <__memory/allocator_traits.h> +#include <__memory/compressed_pair.h> +#include <__memory/shared_ptr.h> +#include +#include // TODO: replace with <__memory/__builtin_new_allocator.h> +#include +#include #if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) #pragma GCC system_header #endif +_LIBCPP_BEGIN_NAMESPACE_STD + +// bad_function_call + +class _LIBCPP_EXCEPTION_ABI bad_function_call + : public exception +{ +#ifdef _LIBCPP_ABI_BAD_FUNCTION_CALL_KEY_FUNCTION +public: + virtual ~bad_function_call() _NOEXCEPT; + + virtual const char* what() const _NOEXCEPT; +#endif +}; + +_LIBCPP_NORETURN inline _LIBCPP_INLINE_VISIBILITY +void __throw_bad_function_call() +{ +#ifndef _LIBCPP_NO_EXCEPTIONS + throw bad_function_call(); +#else + _VSTD::abort(); +#endif +} + +#if defined(_LIBCPP_CXX03_LANG) && !defined(_LIBCPP_DISABLE_DEPRECATION_WARNINGS) && __has_attribute(deprecated) +# define _LIBCPP_DEPRECATED_CXX03_FUNCTION \ + __attribute__((deprecated("Using std::function in C++03 is not supported anymore. Please upgrade to C++11 or later, or use a different type"))) +#else +# define _LIBCPP_DEPRECATED_CXX03_FUNCTION /* nothing */ +#endif + +template class _LIBCPP_DEPRECATED_CXX03_FUNCTION _LIBCPP_TEMPLATE_VIS function; // undefined + +namespace __function +{ + +template +struct __maybe_derive_from_unary_function +{ +}; + +template +struct __maybe_derive_from_unary_function<_Rp(_A1)> + : public unary_function<_A1, _Rp> +{ +}; + +template +struct __maybe_derive_from_binary_function +{ +}; + +template +struct __maybe_derive_from_binary_function<_Rp(_A1, _A2)> + : public binary_function<_A1, _A2, _Rp> +{ +}; + +template +_LIBCPP_INLINE_VISIBILITY +bool __not_null(_Fp const&) { return true; } + +template +_LIBCPP_INLINE_VISIBILITY +bool __not_null(_Fp* __ptr) { return __ptr; } + +template +_LIBCPP_INLINE_VISIBILITY +bool __not_null(_Ret _Class::*__ptr) { return __ptr; } + +template +_LIBCPP_INLINE_VISIBILITY +bool __not_null(function<_Fp> const& __f) { return !!__f; } + +#ifdef _LIBCPP_HAS_EXTENSION_BLOCKS +template +_LIBCPP_INLINE_VISIBILITY +bool __not_null(_Rp (^__p)(_Args...)) { return __p; } +#endif + +} // namespace __function + +#ifndef _LIBCPP_CXX03_LANG + +namespace __function { + +// __alloc_func holds a functor and an allocator. + +template class __alloc_func; +template +class __default_alloc_func; + +template +class __alloc_func<_Fp, _Ap, _Rp(_ArgTypes...)> +{ + __compressed_pair<_Fp, _Ap> __f_; + + public: + typedef _LIBCPP_NODEBUG_TYPE _Fp _Target; + typedef _LIBCPP_NODEBUG_TYPE _Ap _Alloc; + + _LIBCPP_INLINE_VISIBILITY + const _Target& __target() const { return __f_.first(); } + + // WIN32 APIs may define __allocator, so use __get_allocator instead. + _LIBCPP_INLINE_VISIBILITY + const _Alloc& __get_allocator() const { return __f_.second(); } + + _LIBCPP_INLINE_VISIBILITY + explicit __alloc_func(_Target&& __f) + : __f_(piecewise_construct, _VSTD::forward_as_tuple(_VSTD::move(__f)), + _VSTD::forward_as_tuple()) + { + } + + _LIBCPP_INLINE_VISIBILITY + explicit __alloc_func(const _Target& __f, const _Alloc& __a) + : __f_(piecewise_construct, _VSTD::forward_as_tuple(__f), + _VSTD::forward_as_tuple(__a)) + { + } + + _LIBCPP_INLINE_VISIBILITY + explicit __alloc_func(const _Target& __f, _Alloc&& __a) + : __f_(piecewise_construct, _VSTD::forward_as_tuple(__f), + _VSTD::forward_as_tuple(_VSTD::move(__a))) + { + } + + _LIBCPP_INLINE_VISIBILITY + explicit __alloc_func(_Target&& __f, _Alloc&& __a) + : __f_(piecewise_construct, _VSTD::forward_as_tuple(_VSTD::move(__f)), + _VSTD::forward_as_tuple(_VSTD::move(__a))) + { + } + + _LIBCPP_INLINE_VISIBILITY + _Rp operator()(_ArgTypes&&... __arg) + { + typedef __invoke_void_return_wrapper<_Rp> _Invoker; + return _Invoker::__call(__f_.first(), + _VSTD::forward<_ArgTypes>(__arg)...); + } + + _LIBCPP_INLINE_VISIBILITY + __alloc_func* __clone() const + { + typedef allocator_traits<_Alloc> __alloc_traits; + typedef + typename __rebind_alloc_helper<__alloc_traits, __alloc_func>::type + _AA; + _AA __a(__f_.second()); + typedef __allocator_destructor<_AA> _Dp; + unique_ptr<__alloc_func, _Dp> __hold(__a.allocate(1), _Dp(__a, 1)); + ::new ((void*)__hold.get()) __alloc_func(__f_.first(), _Alloc(__a)); + return __hold.release(); + } + + _LIBCPP_INLINE_VISIBILITY + void destroy() _NOEXCEPT { __f_.~__compressed_pair<_Target, _Alloc>(); } + + static void __destroy_and_delete(__alloc_func* __f) { + typedef allocator_traits<_Alloc> __alloc_traits; + typedef typename __rebind_alloc_helper<__alloc_traits, __alloc_func>::type + _FunAlloc; + _FunAlloc __a(__f->__get_allocator()); + __f->destroy(); + __a.deallocate(__f, 1); + } +}; + +template +class __default_alloc_func<_Fp, _Rp(_ArgTypes...)> { + _Fp __f_; + +public: + typedef _LIBCPP_NODEBUG_TYPE _Fp _Target; + + _LIBCPP_INLINE_VISIBILITY + const _Target& __target() const { return __f_; } + + _LIBCPP_INLINE_VISIBILITY + explicit __default_alloc_func(_Target&& __f) : __f_(_VSTD::move(__f)) {} + + _LIBCPP_INLINE_VISIBILITY + explicit __default_alloc_func(const _Target& __f) : __f_(__f) {} + + _LIBCPP_INLINE_VISIBILITY + _Rp operator()(_ArgTypes&&... __arg) { + typedef __invoke_void_return_wrapper<_Rp> _Invoker; + return _Invoker::__call(__f_, _VSTD::forward<_ArgTypes>(__arg)...); + } + + _LIBCPP_INLINE_VISIBILITY + __default_alloc_func* __clone() const { + __builtin_new_allocator::__holder_t __hold = + __builtin_new_allocator::__allocate_type<__default_alloc_func>(1); + __default_alloc_func* __res = + ::new ((void*)__hold.get()) __default_alloc_func(__f_); + (void)__hold.release(); + return __res; + } + + _LIBCPP_INLINE_VISIBILITY + void destroy() _NOEXCEPT { __f_.~_Target(); } + + static void __destroy_and_delete(__default_alloc_func* __f) { + __f->destroy(); + __builtin_new_allocator::__deallocate_type<__default_alloc_func>(__f, 1); + } +}; + +// __base provides an abstract interface for copyable functors. + +template class _LIBCPP_TEMPLATE_VIS __base; + +template +class __base<_Rp(_ArgTypes...)> +{ + __base(const __base&); + __base& operator=(const __base&); +public: + _LIBCPP_INLINE_VISIBILITY __base() {} + _LIBCPP_INLINE_VISIBILITY virtual ~__base() {} + virtual __base* __clone() const = 0; + virtual void __clone(__base*) const = 0; + virtual void destroy() _NOEXCEPT = 0; + virtual void destroy_deallocate() _NOEXCEPT = 0; + virtual _Rp operator()(_ArgTypes&& ...) = 0; +#ifndef _LIBCPP_NO_RTTI + virtual const void* target(const type_info&) const _NOEXCEPT = 0; + virtual const std::type_info& target_type() const _NOEXCEPT = 0; +#endif // _LIBCPP_NO_RTTI +}; + +// __func implements __base for a given functor type. + +template class __func; + +template +class __func<_Fp, _Alloc, _Rp(_ArgTypes...)> + : public __base<_Rp(_ArgTypes...)> +{ + __alloc_func<_Fp, _Alloc, _Rp(_ArgTypes...)> __f_; +public: + _LIBCPP_INLINE_VISIBILITY + explicit __func(_Fp&& __f) + : __f_(_VSTD::move(__f)) {} + + _LIBCPP_INLINE_VISIBILITY + explicit __func(const _Fp& __f, const _Alloc& __a) + : __f_(__f, __a) {} + + _LIBCPP_INLINE_VISIBILITY + explicit __func(const _Fp& __f, _Alloc&& __a) + : __f_(__f, _VSTD::move(__a)) {} + + _LIBCPP_INLINE_VISIBILITY + explicit __func(_Fp&& __f, _Alloc&& __a) + : __f_(_VSTD::move(__f), _VSTD::move(__a)) {} + + virtual __base<_Rp(_ArgTypes...)>* __clone() const; + virtual void __clone(__base<_Rp(_ArgTypes...)>*) const; + virtual void destroy() _NOEXCEPT; + virtual void destroy_deallocate() _NOEXCEPT; + virtual _Rp operator()(_ArgTypes&&... __arg); +#ifndef _LIBCPP_NO_RTTI + virtual const void* target(const type_info&) const _NOEXCEPT; + virtual const std::type_info& target_type() const _NOEXCEPT; +#endif // _LIBCPP_NO_RTTI +}; + +template +__base<_Rp(_ArgTypes...)>* +__func<_Fp, _Alloc, _Rp(_ArgTypes...)>::__clone() const +{ + typedef allocator_traits<_Alloc> __alloc_traits; + typedef typename __rebind_alloc_helper<__alloc_traits, __func>::type _Ap; + _Ap __a(__f_.__get_allocator()); + typedef __allocator_destructor<_Ap> _Dp; + unique_ptr<__func, _Dp> __hold(__a.allocate(1), _Dp(__a, 1)); + ::new ((void*)__hold.get()) __func(__f_.__target(), _Alloc(__a)); + return __hold.release(); +} + +template +void +__func<_Fp, _Alloc, _Rp(_ArgTypes...)>::__clone(__base<_Rp(_ArgTypes...)>* __p) const +{ + ::new ((void*)__p) __func(__f_.__target(), __f_.__get_allocator()); +} + +template +void +__func<_Fp, _Alloc, _Rp(_ArgTypes...)>::destroy() _NOEXCEPT +{ + __f_.destroy(); +} + +template +void +__func<_Fp, _Alloc, _Rp(_ArgTypes...)>::destroy_deallocate() _NOEXCEPT +{ + typedef allocator_traits<_Alloc> __alloc_traits; + typedef typename __rebind_alloc_helper<__alloc_traits, __func>::type _Ap; + _Ap __a(__f_.__get_allocator()); + __f_.destroy(); + __a.deallocate(this, 1); +} + +template +_Rp +__func<_Fp, _Alloc, _Rp(_ArgTypes...)>::operator()(_ArgTypes&& ... __arg) +{ + return __f_(_VSTD::forward<_ArgTypes>(__arg)...); +} + +#ifndef _LIBCPP_NO_RTTI + +template +const void* +__func<_Fp, _Alloc, _Rp(_ArgTypes...)>::target(const type_info& __ti) const _NOEXCEPT +{ + if (__ti == typeid(_Fp)) + return &__f_.__target(); + return nullptr; +} + +template +const std::type_info& +__func<_Fp, _Alloc, _Rp(_ArgTypes...)>::target_type() const _NOEXCEPT +{ + return typeid(_Fp); +} + +#endif // _LIBCPP_NO_RTTI + +// __value_func creates a value-type from a __func. + +template class __value_func; + +template class __value_func<_Rp(_ArgTypes...)> +{ + typename aligned_storage<3 * sizeof(void*)>::type __buf_; + + typedef __base<_Rp(_ArgTypes...)> __func; + __func* __f_; + + _LIBCPP_NO_CFI static __func* __as_base(void* p) + { + return reinterpret_cast<__func*>(p); + } + + public: + _LIBCPP_INLINE_VISIBILITY + __value_func() _NOEXCEPT : __f_(nullptr) {} + + template + _LIBCPP_INLINE_VISIBILITY __value_func(_Fp&& __f, const _Alloc& __a) + : __f_(nullptr) + { + typedef allocator_traits<_Alloc> __alloc_traits; + typedef __function::__func<_Fp, _Alloc, _Rp(_ArgTypes...)> _Fun; + typedef typename __rebind_alloc_helper<__alloc_traits, _Fun>::type + _FunAlloc; + + if (__function::__not_null(__f)) + { + _FunAlloc __af(__a); + if (sizeof(_Fun) <= sizeof(__buf_) && + is_nothrow_copy_constructible<_Fp>::value && + is_nothrow_copy_constructible<_FunAlloc>::value) + { + __f_ = + ::new ((void*)&__buf_) _Fun(_VSTD::move(__f), _Alloc(__af)); + } + else + { + typedef __allocator_destructor<_FunAlloc> _Dp; + unique_ptr<__func, _Dp> __hold(__af.allocate(1), _Dp(__af, 1)); + ::new ((void*)__hold.get()) _Fun(_VSTD::move(__f), _Alloc(__a)); + __f_ = __hold.release(); + } + } + } + + template ::type, __value_func>::value>::type> + _LIBCPP_INLINE_VISIBILITY explicit __value_func(_Fp&& __f) + : __value_func(_VSTD::forward<_Fp>(__f), allocator<_Fp>()) {} + + _LIBCPP_INLINE_VISIBILITY + __value_func(const __value_func& __f) + { + if (__f.__f_ == nullptr) + __f_ = nullptr; + else if ((void*)__f.__f_ == &__f.__buf_) + { + __f_ = __as_base(&__buf_); + __f.__f_->__clone(__f_); + } + else + __f_ = __f.__f_->__clone(); + } + + _LIBCPP_INLINE_VISIBILITY + __value_func(__value_func&& __f) _NOEXCEPT + { + if (__f.__f_ == nullptr) + __f_ = nullptr; + else if ((void*)__f.__f_ == &__f.__buf_) + { + __f_ = __as_base(&__buf_); + __f.__f_->__clone(__f_); + } + else + { + __f_ = __f.__f_; + __f.__f_ = nullptr; + } + } + + _LIBCPP_INLINE_VISIBILITY + ~__value_func() + { + if ((void*)__f_ == &__buf_) + __f_->destroy(); + else if (__f_) + __f_->destroy_deallocate(); + } + + _LIBCPP_INLINE_VISIBILITY + __value_func& operator=(__value_func&& __f) + { + *this = nullptr; + if (__f.__f_ == nullptr) + __f_ = nullptr; + else if ((void*)__f.__f_ == &__f.__buf_) + { + __f_ = __as_base(&__buf_); + __f.__f_->__clone(__f_); + } + else + { + __f_ = __f.__f_; + __f.__f_ = nullptr; + } + return *this; + } + + _LIBCPP_INLINE_VISIBILITY + __value_func& operator=(nullptr_t) + { + __func* __f = __f_; + __f_ = nullptr; + if ((void*)__f == &__buf_) + __f->destroy(); + else if (__f) + __f->destroy_deallocate(); + return *this; + } + + _LIBCPP_INLINE_VISIBILITY + _Rp operator()(_ArgTypes&&... __args) const + { + if (__f_ == nullptr) + __throw_bad_function_call(); + return (*__f_)(_VSTD::forward<_ArgTypes>(__args)...); + } + + _LIBCPP_INLINE_VISIBILITY + void swap(__value_func& __f) _NOEXCEPT + { + if (&__f == this) + return; + if ((void*)__f_ == &__buf_ && (void*)__f.__f_ == &__f.__buf_) + { + typename aligned_storage::type __tempbuf; + __func* __t = __as_base(&__tempbuf); + __f_->__clone(__t); + __f_->destroy(); + __f_ = nullptr; + __f.__f_->__clone(__as_base(&__buf_)); + __f.__f_->destroy(); + __f.__f_ = nullptr; + __f_ = __as_base(&__buf_); + __t->__clone(__as_base(&__f.__buf_)); + __t->destroy(); + __f.__f_ = __as_base(&__f.__buf_); + } + else if ((void*)__f_ == &__buf_) + { + __f_->__clone(__as_base(&__f.__buf_)); + __f_->destroy(); + __f_ = __f.__f_; + __f.__f_ = __as_base(&__f.__buf_); + } + else if ((void*)__f.__f_ == &__f.__buf_) + { + __f.__f_->__clone(__as_base(&__buf_)); + __f.__f_->destroy(); + __f.__f_ = __f_; + __f_ = __as_base(&__buf_); + } + else + _VSTD::swap(__f_, __f.__f_); + } + + _LIBCPP_INLINE_VISIBILITY + explicit operator bool() const _NOEXCEPT { return __f_ != nullptr; } + +#ifndef _LIBCPP_NO_RTTI + _LIBCPP_INLINE_VISIBILITY + const std::type_info& target_type() const _NOEXCEPT + { + if (__f_ == nullptr) + return typeid(void); + return __f_->target_type(); + } + + template + _LIBCPP_INLINE_VISIBILITY const _Tp* target() const _NOEXCEPT + { + if (__f_ == nullptr) + return nullptr; + return (const _Tp*)__f_->target(typeid(_Tp)); + } +#endif // _LIBCPP_NO_RTTI +}; + +// Storage for a functor object, to be used with __policy to manage copy and +// destruction. +union __policy_storage +{ + mutable char __small[sizeof(void*) * 2]; + void* __large; +}; + +// True if _Fun can safely be held in __policy_storage.__small. +template +struct __use_small_storage + : public integral_constant< + bool, sizeof(_Fun) <= sizeof(__policy_storage) && + _LIBCPP_ALIGNOF(_Fun) <= _LIBCPP_ALIGNOF(__policy_storage) && + is_trivially_copy_constructible<_Fun>::value && + is_trivially_destructible<_Fun>::value> {}; + +// Policy contains information about how to copy, destroy, and move the +// underlying functor. You can think of it as a vtable of sorts. +struct __policy +{ + // Used to copy or destroy __large values. null for trivial objects. + void* (*const __clone)(const void*); + void (*const __destroy)(void*); + + // True if this is the null policy (no value). + const bool __is_null; + + // The target type. May be null if RTTI is disabled. + const std::type_info* const __type_info; + + // Returns a pointer to a static policy object suitable for the functor + // type. + template + _LIBCPP_INLINE_VISIBILITY static const __policy* __create() + { + return __choose_policy<_Fun>(__use_small_storage<_Fun>()); + } + + _LIBCPP_INLINE_VISIBILITY + static const __policy* __create_empty() + { + static const _LIBCPP_CONSTEXPR __policy __policy_ = {nullptr, nullptr, + true, +#ifndef _LIBCPP_NO_RTTI + &typeid(void) +#else + nullptr +#endif + }; + return &__policy_; + } + + private: + template static void* __large_clone(const void* __s) + { + const _Fun* __f = static_cast(__s); + return __f->__clone(); + } + + template + static void __large_destroy(void* __s) { + _Fun::__destroy_and_delete(static_cast<_Fun*>(__s)); + } + + template + _LIBCPP_INLINE_VISIBILITY static const __policy* + __choose_policy(/* is_small = */ false_type) { + static const _LIBCPP_CONSTEXPR __policy __policy_ = { + &__large_clone<_Fun>, &__large_destroy<_Fun>, false, +#ifndef _LIBCPP_NO_RTTI + &typeid(typename _Fun::_Target) +#else + nullptr +#endif + }; + return &__policy_; + } + + template + _LIBCPP_INLINE_VISIBILITY static const __policy* + __choose_policy(/* is_small = */ true_type) + { + static const _LIBCPP_CONSTEXPR __policy __policy_ = { + nullptr, nullptr, false, +#ifndef _LIBCPP_NO_RTTI + &typeid(typename _Fun::_Target) +#else + nullptr +#endif + }; + return &__policy_; + } +}; + +// Used to choose between perfect forwarding or pass-by-value. Pass-by-value is +// faster for types that can be passed in registers. +template +using __fast_forward = + typename conditional::value, _Tp, _Tp&&>::type; + +// __policy_invoker calls an instance of __alloc_func held in __policy_storage. + +template struct __policy_invoker; + +template +struct __policy_invoker<_Rp(_ArgTypes...)> +{ + typedef _Rp (*__Call)(const __policy_storage*, + __fast_forward<_ArgTypes>...); + + __Call __call_; + + // Creates an invoker that throws bad_function_call. + _LIBCPP_INLINE_VISIBILITY + __policy_invoker() : __call_(&__call_empty) {} + + // Creates an invoker that calls the given instance of __func. + template + _LIBCPP_INLINE_VISIBILITY static __policy_invoker __create() + { + return __policy_invoker(&__call_impl<_Fun>); + } + + private: + _LIBCPP_INLINE_VISIBILITY + explicit __policy_invoker(__Call __c) : __call_(__c) {} + + static _Rp __call_empty(const __policy_storage*, + __fast_forward<_ArgTypes>...) + { + __throw_bad_function_call(); + } + + template + static _Rp __call_impl(const __policy_storage* __buf, + __fast_forward<_ArgTypes>... __args) + { + _Fun* __f = reinterpret_cast<_Fun*>(__use_small_storage<_Fun>::value + ? &__buf->__small + : __buf->__large); + return (*__f)(_VSTD::forward<_ArgTypes>(__args)...); + } +}; + +// __policy_func uses a __policy and __policy_invoker to create a type-erased, +// copyable functor. + +template class __policy_func; + +template class __policy_func<_Rp(_ArgTypes...)> +{ + // Inline storage for small objects. + __policy_storage __buf_; + + // Calls the value stored in __buf_. This could technically be part of + // policy, but storing it here eliminates a level of indirection inside + // operator(). + typedef __function::__policy_invoker<_Rp(_ArgTypes...)> __invoker; + __invoker __invoker_; + + // The policy that describes how to move / copy / destroy __buf_. Never + // null, even if the function is empty. + const __policy* __policy_; + + public: + _LIBCPP_INLINE_VISIBILITY + __policy_func() : __policy_(__policy::__create_empty()) {} + + template + _LIBCPP_INLINE_VISIBILITY __policy_func(_Fp&& __f, const _Alloc& __a) + : __policy_(__policy::__create_empty()) + { + typedef __alloc_func<_Fp, _Alloc, _Rp(_ArgTypes...)> _Fun; + typedef allocator_traits<_Alloc> __alloc_traits; + typedef typename __rebind_alloc_helper<__alloc_traits, _Fun>::type + _FunAlloc; + + if (__function::__not_null(__f)) + { + __invoker_ = __invoker::template __create<_Fun>(); + __policy_ = __policy::__create<_Fun>(); + + _FunAlloc __af(__a); + if (__use_small_storage<_Fun>()) + { + ::new ((void*)&__buf_.__small) + _Fun(_VSTD::move(__f), _Alloc(__af)); + } + else + { + typedef __allocator_destructor<_FunAlloc> _Dp; + unique_ptr<_Fun, _Dp> __hold(__af.allocate(1), _Dp(__af, 1)); + ::new ((void*)__hold.get()) + _Fun(_VSTD::move(__f), _Alloc(__af)); + __buf_.__large = __hold.release(); + } + } + } + + template ::type, __policy_func>::value>::type> + _LIBCPP_INLINE_VISIBILITY explicit __policy_func(_Fp&& __f) + : __policy_(__policy::__create_empty()) { + typedef __default_alloc_func<_Fp, _Rp(_ArgTypes...)> _Fun; + + if (__function::__not_null(__f)) { + __invoker_ = __invoker::template __create<_Fun>(); + __policy_ = __policy::__create<_Fun>(); + if (__use_small_storage<_Fun>()) { + ::new ((void*)&__buf_.__small) _Fun(_VSTD::move(__f)); + } else { + __builtin_new_allocator::__holder_t __hold = + __builtin_new_allocator::__allocate_type<_Fun>(1); + __buf_.__large = ::new ((void*)__hold.get()) _Fun(_VSTD::move(__f)); + (void)__hold.release(); + } + } + } + + _LIBCPP_INLINE_VISIBILITY + __policy_func(const __policy_func& __f) + : __buf_(__f.__buf_), __invoker_(__f.__invoker_), + __policy_(__f.__policy_) + { + if (__policy_->__clone) + __buf_.__large = __policy_->__clone(__f.__buf_.__large); + } + + _LIBCPP_INLINE_VISIBILITY + __policy_func(__policy_func&& __f) + : __buf_(__f.__buf_), __invoker_(__f.__invoker_), + __policy_(__f.__policy_) + { + if (__policy_->__destroy) + { + __f.__policy_ = __policy::__create_empty(); + __f.__invoker_ = __invoker(); + } + } + + _LIBCPP_INLINE_VISIBILITY + ~__policy_func() + { + if (__policy_->__destroy) + __policy_->__destroy(__buf_.__large); + } + + _LIBCPP_INLINE_VISIBILITY + __policy_func& operator=(__policy_func&& __f) + { + *this = nullptr; + __buf_ = __f.__buf_; + __invoker_ = __f.__invoker_; + __policy_ = __f.__policy_; + __f.__policy_ = __policy::__create_empty(); + __f.__invoker_ = __invoker(); + return *this; + } + + _LIBCPP_INLINE_VISIBILITY + __policy_func& operator=(nullptr_t) + { + const __policy* __p = __policy_; + __policy_ = __policy::__create_empty(); + __invoker_ = __invoker(); + if (__p->__destroy) + __p->__destroy(__buf_.__large); + return *this; + } + + _LIBCPP_INLINE_VISIBILITY + _Rp operator()(_ArgTypes&&... __args) const + { + return __invoker_.__call_(_VSTD::addressof(__buf_), + _VSTD::forward<_ArgTypes>(__args)...); + } + + _LIBCPP_INLINE_VISIBILITY + void swap(__policy_func& __f) + { + _VSTD::swap(__invoker_, __f.__invoker_); + _VSTD::swap(__policy_, __f.__policy_); + _VSTD::swap(__buf_, __f.__buf_); + } + + _LIBCPP_INLINE_VISIBILITY + explicit operator bool() const _NOEXCEPT + { + return !__policy_->__is_null; + } + +#ifndef _LIBCPP_NO_RTTI + _LIBCPP_INLINE_VISIBILITY + const std::type_info& target_type() const _NOEXCEPT + { + return *__policy_->__type_info; + } + + template + _LIBCPP_INLINE_VISIBILITY const _Tp* target() const _NOEXCEPT + { + if (__policy_->__is_null || typeid(_Tp) != *__policy_->__type_info) + return nullptr; + if (__policy_->__clone) // Out of line storage. + return reinterpret_cast(__buf_.__large); + else + return reinterpret_cast(&__buf_.__small); + } +#endif // _LIBCPP_NO_RTTI +}; + +#if defined(_LIBCPP_HAS_BLOCKS_RUNTIME) && !defined(_LIBCPP_HAS_OBJC_ARC) + +extern "C" void *_Block_copy(const void *); +extern "C" void _Block_release(const void *); + +template +class __func<_Rp1(^)(_ArgTypes1...), _Alloc, _Rp(_ArgTypes...)> + : public __base<_Rp(_ArgTypes...)> +{ + typedef _Rp1(^__block_type)(_ArgTypes1...); + __block_type __f_; + +public: + _LIBCPP_INLINE_VISIBILITY + explicit __func(__block_type const& __f) + : __f_(reinterpret_cast<__block_type>(__f ? _Block_copy(__f) : nullptr)) + { } + + // [TODO] add && to save on a retain + + _LIBCPP_INLINE_VISIBILITY + explicit __func(__block_type __f, const _Alloc& /* unused */) + : __f_(reinterpret_cast<__block_type>(__f ? _Block_copy(__f) : nullptr)) + { } + + virtual __base<_Rp(_ArgTypes...)>* __clone() const { + _LIBCPP_ASSERT(false, + "Block pointers are just pointers, so they should always fit into " + "std::function's small buffer optimization. This function should " + "never be invoked."); + return nullptr; + } + + virtual void __clone(__base<_Rp(_ArgTypes...)>* __p) const { + ::new ((void*)__p) __func(__f_); + } + + virtual void destroy() _NOEXCEPT { + if (__f_) + _Block_release(__f_); + __f_ = 0; + } + + virtual void destroy_deallocate() _NOEXCEPT { + _LIBCPP_ASSERT(false, + "Block pointers are just pointers, so they should always fit into " + "std::function's small buffer optimization. This function should " + "never be invoked."); + } + + virtual _Rp operator()(_ArgTypes&& ... __arg) { + return _VSTD::__invoke(__f_, _VSTD::forward<_ArgTypes>(__arg)...); + } + +#ifndef _LIBCPP_NO_RTTI + virtual const void* target(type_info const& __ti) const _NOEXCEPT { + if (__ti == typeid(__func::__block_type)) + return &__f_; + return (const void*)nullptr; + } + + virtual const std::type_info& target_type() const _NOEXCEPT { + return typeid(__func::__block_type); + } +#endif // _LIBCPP_NO_RTTI +}; + +#endif // _LIBCPP_HAS_EXTENSION_BLOCKS && !_LIBCPP_HAS_OBJC_ARC + +} // __function + +template +class _LIBCPP_TEMPLATE_VIS function<_Rp(_ArgTypes...)> +#if _LIBCPP_STD_VER <= 17 || !defined(_LIBCPP_ABI_NO_BINDER_BASES) + : public __function::__maybe_derive_from_unary_function<_Rp(_ArgTypes...)>, + public __function::__maybe_derive_from_binary_function<_Rp(_ArgTypes...)> +#endif +{ +#ifndef _LIBCPP_ABI_OPTIMIZED_FUNCTION + typedef __function::__value_func<_Rp(_ArgTypes...)> __func; +#else + typedef __function::__policy_func<_Rp(_ArgTypes...)> __func; +#endif + + __func __f_; + + template , function>, + __invokable<_Fp, _ArgTypes...> + >::value> + struct __callable; + template + struct __callable<_Fp, true> + { + static const bool value = is_void<_Rp>::value || + __is_core_convertible::type, + _Rp>::value; + }; + template + struct __callable<_Fp, false> + { + static const bool value = false; + }; + + template + using _EnableIfLValueCallable = typename enable_if<__callable<_Fp&>::value>::type; +public: + typedef _Rp result_type; + + // construct/copy/destroy: + _LIBCPP_INLINE_VISIBILITY + function() _NOEXCEPT { } + _LIBCPP_INLINE_VISIBILITY + function(nullptr_t) _NOEXCEPT {} + function(const function&); + function(function&&) _NOEXCEPT; + template> + function(_Fp); + +#if _LIBCPP_STD_VER <= 14 + template + _LIBCPP_INLINE_VISIBILITY + function(allocator_arg_t, const _Alloc&) _NOEXCEPT {} + template + _LIBCPP_INLINE_VISIBILITY + function(allocator_arg_t, const _Alloc&, nullptr_t) _NOEXCEPT {} + template + function(allocator_arg_t, const _Alloc&, const function&); + template + function(allocator_arg_t, const _Alloc&, function&&); + template> + function(allocator_arg_t, const _Alloc& __a, _Fp __f); +#endif + + function& operator=(const function&); + function& operator=(function&&) _NOEXCEPT; + function& operator=(nullptr_t) _NOEXCEPT; + template::type>> + function& operator=(_Fp&&); + + ~function(); + + // function modifiers: + void swap(function&) _NOEXCEPT; + +#if _LIBCPP_STD_VER <= 14 + template + _LIBCPP_INLINE_VISIBILITY + void assign(_Fp&& __f, const _Alloc& __a) + {function(allocator_arg, __a, _VSTD::forward<_Fp>(__f)).swap(*this);} +#endif + + // function capacity: + _LIBCPP_INLINE_VISIBILITY + explicit operator bool() const _NOEXCEPT { + return static_cast(__f_); + } + + // deleted overloads close possible hole in the type system + template + bool operator==(const function<_R2(_ArgTypes2...)>&) const = delete; + template + bool operator!=(const function<_R2(_ArgTypes2...)>&) const = delete; +public: + // function invocation: + _Rp operator()(_ArgTypes...) const; + +#ifndef _LIBCPP_NO_RTTI + // function target access: + const std::type_info& target_type() const _NOEXCEPT; + template _Tp* target() _NOEXCEPT; + template const _Tp* target() const _NOEXCEPT; +#endif // _LIBCPP_NO_RTTI +}; + +#ifndef _LIBCPP_HAS_NO_DEDUCTION_GUIDES +template +function(_Rp(*)(_Ap...)) -> function<_Rp(_Ap...)>; + +template +struct __strip_signature; + +template +struct __strip_signature<_Rp (_Gp::*) (_Ap...)> { using type = _Rp(_Ap...); }; +template +struct __strip_signature<_Rp (_Gp::*) (_Ap...) const> { using type = _Rp(_Ap...); }; +template +struct __strip_signature<_Rp (_Gp::*) (_Ap...) volatile> { using type = _Rp(_Ap...); }; +template +struct __strip_signature<_Rp (_Gp::*) (_Ap...) const volatile> { using type = _Rp(_Ap...); }; + +template +struct __strip_signature<_Rp (_Gp::*) (_Ap...) &> { using type = _Rp(_Ap...); }; +template +struct __strip_signature<_Rp (_Gp::*) (_Ap...) const &> { using type = _Rp(_Ap...); }; +template +struct __strip_signature<_Rp (_Gp::*) (_Ap...) volatile &> { using type = _Rp(_Ap...); }; +template +struct __strip_signature<_Rp (_Gp::*) (_Ap...) const volatile &> { using type = _Rp(_Ap...); }; + +template +struct __strip_signature<_Rp (_Gp::*) (_Ap...) noexcept> { using type = _Rp(_Ap...); }; +template +struct __strip_signature<_Rp (_Gp::*) (_Ap...) const noexcept> { using type = _Rp(_Ap...); }; +template +struct __strip_signature<_Rp (_Gp::*) (_Ap...) volatile noexcept> { using type = _Rp(_Ap...); }; +template +struct __strip_signature<_Rp (_Gp::*) (_Ap...) const volatile noexcept> { using type = _Rp(_Ap...); }; + +template +struct __strip_signature<_Rp (_Gp::*) (_Ap...) & noexcept> { using type = _Rp(_Ap...); }; +template +struct __strip_signature<_Rp (_Gp::*) (_Ap...) const & noexcept> { using type = _Rp(_Ap...); }; +template +struct __strip_signature<_Rp (_Gp::*) (_Ap...) volatile & noexcept> { using type = _Rp(_Ap...); }; +template +struct __strip_signature<_Rp (_Gp::*) (_Ap...) const volatile & noexcept> { using type = _Rp(_Ap...); }; + +template::type> +function(_Fp) -> function<_Stripped>; +#endif // !_LIBCPP_HAS_NO_DEDUCTION_GUIDES + +template +function<_Rp(_ArgTypes...)>::function(const function& __f) : __f_(__f.__f_) {} + +#if _LIBCPP_STD_VER <= 14 +template +template +function<_Rp(_ArgTypes...)>::function(allocator_arg_t, const _Alloc&, + const function& __f) : __f_(__f.__f_) {} +#endif + +template +function<_Rp(_ArgTypes...)>::function(function&& __f) _NOEXCEPT + : __f_(_VSTD::move(__f.__f_)) {} + +#if _LIBCPP_STD_VER <= 14 +template +template +function<_Rp(_ArgTypes...)>::function(allocator_arg_t, const _Alloc&, + function&& __f) + : __f_(_VSTD::move(__f.__f_)) {} +#endif + +template +template +function<_Rp(_ArgTypes...)>::function(_Fp __f) : __f_(_VSTD::move(__f)) {} + +#if _LIBCPP_STD_VER <= 14 +template +template +function<_Rp(_ArgTypes...)>::function(allocator_arg_t, const _Alloc& __a, + _Fp __f) + : __f_(_VSTD::move(__f), __a) {} +#endif + +template +function<_Rp(_ArgTypes...)>& +function<_Rp(_ArgTypes...)>::operator=(const function& __f) +{ + function(__f).swap(*this); + return *this; +} + +template +function<_Rp(_ArgTypes...)>& +function<_Rp(_ArgTypes...)>::operator=(function&& __f) _NOEXCEPT +{ + __f_ = _VSTD::move(__f.__f_); + return *this; +} + +template +function<_Rp(_ArgTypes...)>& +function<_Rp(_ArgTypes...)>::operator=(nullptr_t) _NOEXCEPT +{ + __f_ = nullptr; + return *this; +} + +template +template +function<_Rp(_ArgTypes...)>& +function<_Rp(_ArgTypes...)>::operator=(_Fp&& __f) +{ + function(_VSTD::forward<_Fp>(__f)).swap(*this); + return *this; +} + +template +function<_Rp(_ArgTypes...)>::~function() {} + +template +void +function<_Rp(_ArgTypes...)>::swap(function& __f) _NOEXCEPT +{ + __f_.swap(__f.__f_); +} + +template +_Rp +function<_Rp(_ArgTypes...)>::operator()(_ArgTypes... __arg) const +{ + return __f_(_VSTD::forward<_ArgTypes>(__arg)...); +} + +#ifndef _LIBCPP_NO_RTTI + +template +const std::type_info& +function<_Rp(_ArgTypes...)>::target_type() const _NOEXCEPT +{ + return __f_.target_type(); +} + +template +template +_Tp* +function<_Rp(_ArgTypes...)>::target() _NOEXCEPT +{ + return (_Tp*)(__f_.template target<_Tp>()); +} + +template +template +const _Tp* +function<_Rp(_ArgTypes...)>::target() const _NOEXCEPT +{ + return __f_.template target<_Tp>(); +} + +#endif // _LIBCPP_NO_RTTI + +template +inline _LIBCPP_INLINE_VISIBILITY +bool +operator==(const function<_Rp(_ArgTypes...)>& __f, nullptr_t) _NOEXCEPT {return !__f;} + +template +inline _LIBCPP_INLINE_VISIBILITY +bool +operator==(nullptr_t, const function<_Rp(_ArgTypes...)>& __f) _NOEXCEPT {return !__f;} + +template +inline _LIBCPP_INLINE_VISIBILITY +bool +operator!=(const function<_Rp(_ArgTypes...)>& __f, nullptr_t) _NOEXCEPT {return (bool)__f;} + +template +inline _LIBCPP_INLINE_VISIBILITY +bool +operator!=(nullptr_t, const function<_Rp(_ArgTypes...)>& __f) _NOEXCEPT {return (bool)__f;} + +template +inline _LIBCPP_INLINE_VISIBILITY +void +swap(function<_Rp(_ArgTypes...)>& __x, function<_Rp(_ArgTypes...)>& __y) _NOEXCEPT +{return __x.swap(__y);} + +#else // _LIBCPP_CXX03_LANG + namespace __function { template class __base; @@ -36,7 +1250,7 @@ public: #ifndef _LIBCPP_NO_RTTI virtual const void* target(const type_info&) const = 0; virtual const std::type_info& target_type() const = 0; -#endif // _LIBCPP_NO_RTTI +#endif // _LIBCPP_NO_RTTI }; template @@ -55,7 +1269,7 @@ public: #ifndef _LIBCPP_NO_RTTI virtual const void* target(const type_info&) const = 0; virtual const std::type_info& target_type() const = 0; -#endif // _LIBCPP_NO_RTTI +#endif // _LIBCPP_NO_RTTI }; template @@ -74,7 +1288,7 @@ public: #ifndef _LIBCPP_NO_RTTI virtual const void* target(const type_info&) const = 0; virtual const std::type_info& target_type() const = 0; -#endif // _LIBCPP_NO_RTTI +#endif // _LIBCPP_NO_RTTI }; template @@ -93,7 +1307,7 @@ public: #ifndef _LIBCPP_NO_RTTI virtual const void* target(const type_info&) const = 0; virtual const std::type_info& target_type() const = 0; -#endif // _LIBCPP_NO_RTTI +#endif // _LIBCPP_NO_RTTI }; template class __func; @@ -114,7 +1328,7 @@ public: #ifndef _LIBCPP_NO_RTTI virtual const void* target(const type_info&) const; virtual const std::type_info& target_type() const; -#endif // _LIBCPP_NO_RTTI +#endif // _LIBCPP_NO_RTTI }; template @@ -181,7 +1395,7 @@ __func<_Fp, _Alloc, _Rp()>::target_type() const return typeid(_Fp); } -#endif // _LIBCPP_NO_RTTI +#endif // _LIBCPP_NO_RTTI template class __func<_Fp, _Alloc, _Rp(_A0)> @@ -200,7 +1414,7 @@ public: #ifndef _LIBCPP_NO_RTTI virtual const void* target(const type_info&) const; virtual const std::type_info& target_type() const; -#endif // _LIBCPP_NO_RTTI +#endif // _LIBCPP_NO_RTTI }; template @@ -267,7 +1481,7 @@ __func<_Fp, _Alloc, _Rp(_A0)>::target_type() const return typeid(_Fp); } -#endif // _LIBCPP_NO_RTTI +#endif // _LIBCPP_NO_RTTI template class __func<_Fp, _Alloc, _Rp(_A0, _A1)> @@ -286,7 +1500,7 @@ public: #ifndef _LIBCPP_NO_RTTI virtual const void* target(const type_info&) const; virtual const std::type_info& target_type() const; -#endif // _LIBCPP_NO_RTTI +#endif // _LIBCPP_NO_RTTI }; template @@ -353,7 +1567,7 @@ __func<_Fp, _Alloc, _Rp(_A0, _A1)>::target_type() const return typeid(_Fp); } -#endif // _LIBCPP_NO_RTTI +#endif // _LIBCPP_NO_RTTI template class __func<_Fp, _Alloc, _Rp(_A0, _A1, _A2)> @@ -372,7 +1586,7 @@ public: #ifndef _LIBCPP_NO_RTTI virtual const void* target(const type_info&) const; virtual const std::type_info& target_type() const; -#endif // _LIBCPP_NO_RTTI +#endif // _LIBCPP_NO_RTTI }; template @@ -439,7 +1653,7 @@ __func<_Fp, _Alloc, _Rp(_A0, _A1, _A2)>::target_type() const return typeid(_Fp); } -#endif // _LIBCPP_NO_RTTI +#endif // _LIBCPP_NO_RTTI } // __function @@ -493,7 +1707,7 @@ public: {function(allocator_arg, __a, __f).swap(*this);} // 20.7.16.2.3, function capacity: - _LIBCPP_INLINE_VISIBILITY operator bool() const {return __f_;} + _LIBCPP_INLINE_VISIBILITY explicit operator bool() const {return __f_;} private: // deleted overloads close possible hole in the type system @@ -510,7 +1724,7 @@ public: const std::type_info& target_type() const; template _Tp* target(); template const _Tp* target() const; -#endif // _LIBCPP_NO_RTTI +#endif // _LIBCPP_NO_RTTI }; template @@ -720,7 +1934,7 @@ function<_Rp()>::target() const return (const _Tp*)__f_->target(typeid(_Tp)); } -#endif // _LIBCPP_NO_RTTI +#endif // _LIBCPP_NO_RTTI template class _LIBCPP_TEMPLATE_VIS function<_Rp(_A0)> @@ -773,7 +1987,7 @@ public: {function(allocator_arg, __a, __f).swap(*this);} // 20.7.16.2.3, function capacity: - _LIBCPP_INLINE_VISIBILITY operator bool() const {return __f_;} + _LIBCPP_INLINE_VISIBILITY explicit operator bool() const {return __f_;} private: // deleted overloads close possible hole in the type system @@ -790,7 +2004,7 @@ public: const std::type_info& target_type() const; template _Tp* target(); template const _Tp* target() const; -#endif // _LIBCPP_NO_RTTI +#endif // _LIBCPP_NO_RTTI }; template @@ -1000,7 +2214,7 @@ function<_Rp(_A0)>::target() const return (const _Tp*)__f_->target(typeid(_Tp)); } -#endif // _LIBCPP_NO_RTTI +#endif // _LIBCPP_NO_RTTI template class _LIBCPP_TEMPLATE_VIS function<_Rp(_A0, _A1)> @@ -1053,7 +2267,7 @@ public: {function(allocator_arg, __a, __f).swap(*this);} // 20.7.16.2.3, function capacity: - operator bool() const {return __f_;} + _LIBCPP_INLINE_VISIBILITY explicit operator bool() const {return __f_;} private: // deleted overloads close possible hole in the type system @@ -1070,7 +2284,7 @@ public: const std::type_info& target_type() const; template _Tp* target(); template const _Tp* target() const; -#endif // _LIBCPP_NO_RTTI +#endif // _LIBCPP_NO_RTTI }; template @@ -1280,7 +2494,7 @@ function<_Rp(_A0, _A1)>::target() const return (const _Tp*)__f_->target(typeid(_Tp)); } -#endif // _LIBCPP_NO_RTTI +#endif // _LIBCPP_NO_RTTI template class _LIBCPP_TEMPLATE_VIS function<_Rp(_A0, _A1, _A2)> @@ -1332,7 +2546,7 @@ public: {function(allocator_arg, __a, __f).swap(*this);} // 20.7.16.2.3, function capacity: - _LIBCPP_INLINE_VISIBILITY operator bool() const {return __f_;} + _LIBCPP_INLINE_VISIBILITY explicit operator bool() const {return __f_;} private: // deleted overloads close possible hole in the type system @@ -1349,7 +2563,7 @@ public: const std::type_info& target_type() const; template _Tp* target(); template const _Tp* target() const; -#endif // _LIBCPP_NO_RTTI +#endif // _LIBCPP_NO_RTTI }; template @@ -1560,7 +2774,7 @@ function<_Rp(_A0, _A1, _A2)>::target() const return (const _Tp*)__f_->target(typeid(_Tp)); } -#endif // _LIBCPP_NO_RTTI +#endif // _LIBCPP_NO_RTTI template inline _LIBCPP_INLINE_VISIBILITY @@ -1588,4 +2802,8 @@ void swap(function<_Fp>& __x, function<_Fp>& __y) {return __x.swap(__y);} -#endif // _LIBCPP_FUNCTIONAL_03 +#endif + +_LIBCPP_END_NAMESPACE_STD + +#endif // _LIBCPP___FUNCTIONAL_FUNCTION_H diff --git a/lib/libcxx/include/__functional/hash.h b/lib/libcxx/include/__functional/hash.h new file mode 100644 index 000000000..ebcbbad13 --- /dev/null +++ b/lib/libcxx/include/__functional/hash.h @@ -0,0 +1,873 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___FUNCTIONAL_HASH_H +#define _LIBCPP___FUNCTIONAL_HASH_H + +#include <__config> +#include <__functional/unary_function.h> +#include <__tuple> +#include <__utility/forward.h> +#include <__utility/move.h> +#include <__utility/pair.h> +#include <__utility/swap.h> +#include +#include +#include +#include +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +inline _LIBCPP_INLINE_VISIBILITY +_Size +__loadword(const void* __p) +{ + _Size __r; + _VSTD::memcpy(&__r, __p, sizeof(__r)); + return __r; +} + +// We use murmur2 when size_t is 32 bits, and cityhash64 when size_t +// is 64 bits. This is because cityhash64 uses 64bit x 64bit +// multiplication, which can be very slow on 32-bit systems. +template +struct __murmur2_or_cityhash; + +template +struct __murmur2_or_cityhash<_Size, 32> +{ + inline _Size operator()(const void* __key, _Size __len) + _LIBCPP_DISABLE_UBSAN_UNSIGNED_INTEGER_CHECK; +}; + +// murmur2 +template +_Size +__murmur2_or_cityhash<_Size, 32>::operator()(const void* __key, _Size __len) +{ + const _Size __m = 0x5bd1e995; + const _Size __r = 24; + _Size __h = __len; + const unsigned char* __data = static_cast(__key); + for (; __len >= 4; __data += 4, __len -= 4) + { + _Size __k = __loadword<_Size>(__data); + __k *= __m; + __k ^= __k >> __r; + __k *= __m; + __h *= __m; + __h ^= __k; + } + switch (__len) + { + case 3: + __h ^= static_cast<_Size>(__data[2] << 16); + _LIBCPP_FALLTHROUGH(); + case 2: + __h ^= static_cast<_Size>(__data[1] << 8); + _LIBCPP_FALLTHROUGH(); + case 1: + __h ^= __data[0]; + __h *= __m; + } + __h ^= __h >> 13; + __h *= __m; + __h ^= __h >> 15; + return __h; +} + +template +struct __murmur2_or_cityhash<_Size, 64> +{ + inline _Size operator()(const void* __key, _Size __len) _LIBCPP_DISABLE_UBSAN_UNSIGNED_INTEGER_CHECK; + + private: + // Some primes between 2^63 and 2^64. + static const _Size __k0 = 0xc3a5c85c97cb3127ULL; + static const _Size __k1 = 0xb492b66fbe98f273ULL; + static const _Size __k2 = 0x9ae16a3b2f90404fULL; + static const _Size __k3 = 0xc949d7c7509e6557ULL; + + static _Size __rotate(_Size __val, int __shift) { + return __shift == 0 ? __val : ((__val >> __shift) | (__val << (64 - __shift))); + } + + static _Size __rotate_by_at_least_1(_Size __val, int __shift) { + return (__val >> __shift) | (__val << (64 - __shift)); + } + + static _Size __shift_mix(_Size __val) { + return __val ^ (__val >> 47); + } + + static _Size __hash_len_16(_Size __u, _Size __v) + _LIBCPP_DISABLE_UBSAN_UNSIGNED_INTEGER_CHECK + { + const _Size __mul = 0x9ddfea08eb382d69ULL; + _Size __a = (__u ^ __v) * __mul; + __a ^= (__a >> 47); + _Size __b = (__v ^ __a) * __mul; + __b ^= (__b >> 47); + __b *= __mul; + return __b; + } + + static _Size __hash_len_0_to_16(const char* __s, _Size __len) + _LIBCPP_DISABLE_UBSAN_UNSIGNED_INTEGER_CHECK + { + if (__len > 8) { + const _Size __a = __loadword<_Size>(__s); + const _Size __b = __loadword<_Size>(__s + __len - 8); + return __hash_len_16(__a, __rotate_by_at_least_1(__b + __len, __len)) ^ __b; + } + if (__len >= 4) { + const uint32_t __a = __loadword(__s); + const uint32_t __b = __loadword(__s + __len - 4); + return __hash_len_16(__len + (__a << 3), __b); + } + if (__len > 0) { + const unsigned char __a = static_cast(__s[0]); + const unsigned char __b = static_cast(__s[__len >> 1]); + const unsigned char __c = static_cast(__s[__len - 1]); + const uint32_t __y = static_cast(__a) + + (static_cast(__b) << 8); + const uint32_t __z = __len + (static_cast(__c) << 2); + return __shift_mix(__y * __k2 ^ __z * __k3) * __k2; + } + return __k2; + } + + static _Size __hash_len_17_to_32(const char *__s, _Size __len) + _LIBCPP_DISABLE_UBSAN_UNSIGNED_INTEGER_CHECK + { + const _Size __a = __loadword<_Size>(__s) * __k1; + const _Size __b = __loadword<_Size>(__s + 8); + const _Size __c = __loadword<_Size>(__s + __len - 8) * __k2; + const _Size __d = __loadword<_Size>(__s + __len - 16) * __k0; + return __hash_len_16(__rotate(__a - __b, 43) + __rotate(__c, 30) + __d, + __a + __rotate(__b ^ __k3, 20) - __c + __len); + } + + // Return a 16-byte hash for 48 bytes. Quick and dirty. + // Callers do best to use "random-looking" values for a and b. + static pair<_Size, _Size> __weak_hash_len_32_with_seeds( + _Size __w, _Size __x, _Size __y, _Size __z, _Size __a, _Size __b) + _LIBCPP_DISABLE_UBSAN_UNSIGNED_INTEGER_CHECK + { + __a += __w; + __b = __rotate(__b + __a + __z, 21); + const _Size __c = __a; + __a += __x; + __a += __y; + __b += __rotate(__a, 44); + return pair<_Size, _Size>(__a + __z, __b + __c); + } + + // Return a 16-byte hash for s[0] ... s[31], a, and b. Quick and dirty. + static pair<_Size, _Size> __weak_hash_len_32_with_seeds( + const char* __s, _Size __a, _Size __b) + _LIBCPP_DISABLE_UBSAN_UNSIGNED_INTEGER_CHECK + { + return __weak_hash_len_32_with_seeds(__loadword<_Size>(__s), + __loadword<_Size>(__s + 8), + __loadword<_Size>(__s + 16), + __loadword<_Size>(__s + 24), + __a, + __b); + } + + // Return an 8-byte hash for 33 to 64 bytes. + static _Size __hash_len_33_to_64(const char *__s, size_t __len) + _LIBCPP_DISABLE_UBSAN_UNSIGNED_INTEGER_CHECK + { + _Size __z = __loadword<_Size>(__s + 24); + _Size __a = __loadword<_Size>(__s) + + (__len + __loadword<_Size>(__s + __len - 16)) * __k0; + _Size __b = __rotate(__a + __z, 52); + _Size __c = __rotate(__a, 37); + __a += __loadword<_Size>(__s + 8); + __c += __rotate(__a, 7); + __a += __loadword<_Size>(__s + 16); + _Size __vf = __a + __z; + _Size __vs = __b + __rotate(__a, 31) + __c; + __a = __loadword<_Size>(__s + 16) + __loadword<_Size>(__s + __len - 32); + __z += __loadword<_Size>(__s + __len - 8); + __b = __rotate(__a + __z, 52); + __c = __rotate(__a, 37); + __a += __loadword<_Size>(__s + __len - 24); + __c += __rotate(__a, 7); + __a += __loadword<_Size>(__s + __len - 16); + _Size __wf = __a + __z; + _Size __ws = __b + __rotate(__a, 31) + __c; + _Size __r = __shift_mix((__vf + __ws) * __k2 + (__wf + __vs) * __k0); + return __shift_mix(__r * __k0 + __vs) * __k2; + } +}; + +// cityhash64 +template +_Size +__murmur2_or_cityhash<_Size, 64>::operator()(const void* __key, _Size __len) +{ + const char* __s = static_cast(__key); + if (__len <= 32) { + if (__len <= 16) { + return __hash_len_0_to_16(__s, __len); + } else { + return __hash_len_17_to_32(__s, __len); + } + } else if (__len <= 64) { + return __hash_len_33_to_64(__s, __len); + } + + // For strings over 64 bytes we hash the end first, and then as we + // loop we keep 56 bytes of state: v, w, x, y, and z. + _Size __x = __loadword<_Size>(__s + __len - 40); + _Size __y = __loadword<_Size>(__s + __len - 16) + + __loadword<_Size>(__s + __len - 56); + _Size __z = __hash_len_16(__loadword<_Size>(__s + __len - 48) + __len, + __loadword<_Size>(__s + __len - 24)); + pair<_Size, _Size> __v = __weak_hash_len_32_with_seeds(__s + __len - 64, __len, __z); + pair<_Size, _Size> __w = __weak_hash_len_32_with_seeds(__s + __len - 32, __y + __k1, __x); + __x = __x * __k1 + __loadword<_Size>(__s); + + // Decrease len to the nearest multiple of 64, and operate on 64-byte chunks. + __len = (__len - 1) & ~static_cast<_Size>(63); + do { + __x = __rotate(__x + __y + __v.first + __loadword<_Size>(__s + 8), 37) * __k1; + __y = __rotate(__y + __v.second + __loadword<_Size>(__s + 48), 42) * __k1; + __x ^= __w.second; + __y += __v.first + __loadword<_Size>(__s + 40); + __z = __rotate(__z + __w.first, 33) * __k1; + __v = __weak_hash_len_32_with_seeds(__s, __v.second * __k1, __x + __w.first); + __w = __weak_hash_len_32_with_seeds(__s + 32, __z + __w.second, + __y + __loadword<_Size>(__s + 16)); + _VSTD::swap(__z, __x); + __s += 64; + __len -= 64; + } while (__len != 0); + return __hash_len_16( + __hash_len_16(__v.first, __w.first) + __shift_mix(__y) * __k1 + __z, + __hash_len_16(__v.second, __w.second) + __x); +} + +template +struct __scalar_hash; + +_LIBCPP_SUPPRESS_DEPRECATED_PUSH +template +struct __scalar_hash<_Tp, 0> +#if !defined(_LIBCPP_ABI_NO_BINDER_BASES) + : public unary_function<_Tp, size_t> +#endif +{ +_LIBCPP_SUPPRESS_DEPRECATED_POP +#if _LIBCPP_STD_VER <= 17 || defined(_LIBCPP_ENABLE_CXX20_REMOVED_BINDER_TYPEDEFS) + _LIBCPP_DEPRECATED_IN_CXX17 typedef size_t result_type; + _LIBCPP_DEPRECATED_IN_CXX17 typedef _Tp argument_type; +#endif + _LIBCPP_INLINE_VISIBILITY + size_t operator()(_Tp __v) const _NOEXCEPT + { + union + { + _Tp __t; + size_t __a; + } __u; + __u.__a = 0; + __u.__t = __v; + return __u.__a; + } +}; + +_LIBCPP_SUPPRESS_DEPRECATED_PUSH +template +struct __scalar_hash<_Tp, 1> +#if !defined(_LIBCPP_ABI_NO_BINDER_BASES) + : public unary_function<_Tp, size_t> +#endif +{ +_LIBCPP_SUPPRESS_DEPRECATED_POP +#if _LIBCPP_STD_VER <= 17 || defined(_LIBCPP_ENABLE_CXX20_REMOVED_BINDER_TYPEDEFS) + _LIBCPP_DEPRECATED_IN_CXX17 typedef size_t result_type; + _LIBCPP_DEPRECATED_IN_CXX17 typedef _Tp argument_type; +#endif + _LIBCPP_INLINE_VISIBILITY + size_t operator()(_Tp __v) const _NOEXCEPT + { + union + { + _Tp __t; + size_t __a; + } __u; + __u.__t = __v; + return __u.__a; + } +}; + +_LIBCPP_SUPPRESS_DEPRECATED_PUSH +template +struct __scalar_hash<_Tp, 2> +#if !defined(_LIBCPP_ABI_NO_BINDER_BASES) + : public unary_function<_Tp, size_t> +#endif +{ +_LIBCPP_SUPPRESS_DEPRECATED_POP +#if _LIBCPP_STD_VER <= 17 || defined(_LIBCPP_ENABLE_CXX20_REMOVED_BINDER_TYPEDEFS) + _LIBCPP_DEPRECATED_IN_CXX17 typedef size_t result_type; + _LIBCPP_DEPRECATED_IN_CXX17 typedef _Tp argument_type; +#endif + _LIBCPP_INLINE_VISIBILITY + size_t operator()(_Tp __v) const _NOEXCEPT + { + union + { + _Tp __t; + struct + { + size_t __a; + size_t __b; + } __s; + } __u; + __u.__t = __v; + return __murmur2_or_cityhash()(&__u, sizeof(__u)); + } +}; + +_LIBCPP_SUPPRESS_DEPRECATED_PUSH +template +struct __scalar_hash<_Tp, 3> +#if !defined(_LIBCPP_ABI_NO_BINDER_BASES) + : public unary_function<_Tp, size_t> +#endif +{ +_LIBCPP_SUPPRESS_DEPRECATED_POP +#if _LIBCPP_STD_VER <= 17 || defined(_LIBCPP_ENABLE_CXX20_REMOVED_BINDER_TYPEDEFS) + _LIBCPP_DEPRECATED_IN_CXX17 typedef size_t result_type; + _LIBCPP_DEPRECATED_IN_CXX17 typedef _Tp argument_type; +#endif + _LIBCPP_INLINE_VISIBILITY + size_t operator()(_Tp __v) const _NOEXCEPT + { + union + { + _Tp __t; + struct + { + size_t __a; + size_t __b; + size_t __c; + } __s; + } __u; + __u.__t = __v; + return __murmur2_or_cityhash()(&__u, sizeof(__u)); + } +}; + +_LIBCPP_SUPPRESS_DEPRECATED_PUSH +template +struct __scalar_hash<_Tp, 4> +#if !defined(_LIBCPP_ABI_NO_BINDER_BASES) + : public unary_function<_Tp, size_t> +#endif +{ +_LIBCPP_SUPPRESS_DEPRECATED_POP +#if _LIBCPP_STD_VER <= 17 || defined(_LIBCPP_ENABLE_CXX20_REMOVED_BINDER_TYPEDEFS) + _LIBCPP_DEPRECATED_IN_CXX17 typedef size_t result_type; + _LIBCPP_DEPRECATED_IN_CXX17 typedef _Tp argument_type; +#endif + _LIBCPP_INLINE_VISIBILITY + size_t operator()(_Tp __v) const _NOEXCEPT + { + union + { + _Tp __t; + struct + { + size_t __a; + size_t __b; + size_t __c; + size_t __d; + } __s; + } __u; + __u.__t = __v; + return __murmur2_or_cityhash()(&__u, sizeof(__u)); + } +}; + +struct _PairT { + size_t first; + size_t second; +}; + +_LIBCPP_INLINE_VISIBILITY +inline size_t __hash_combine(size_t __lhs, size_t __rhs) _NOEXCEPT { + typedef __scalar_hash<_PairT> _HashT; + const _PairT __p = {__lhs, __rhs}; + return _HashT()(__p); +} + +_LIBCPP_SUPPRESS_DEPRECATED_PUSH +template +struct _LIBCPP_TEMPLATE_VIS hash<_Tp*> +#if !defined(_LIBCPP_ABI_NO_BINDER_BASES) + : public unary_function<_Tp*, size_t> +#endif +{ +_LIBCPP_SUPPRESS_DEPRECATED_POP +#if _LIBCPP_STD_VER <= 17 || defined(_LIBCPP_ENABLE_CXX20_REMOVED_BINDER_TYPEDEFS) + _LIBCPP_DEPRECATED_IN_CXX17 typedef size_t result_type; + _LIBCPP_DEPRECATED_IN_CXX17 typedef _Tp* argument_type; +#endif + _LIBCPP_INLINE_VISIBILITY + size_t operator()(_Tp* __v) const _NOEXCEPT + { + union + { + _Tp* __t; + size_t __a; + } __u; + __u.__t = __v; + return __murmur2_or_cityhash()(&__u, sizeof(__u)); + } +}; + +_LIBCPP_SUPPRESS_DEPRECATED_PUSH +template <> +struct _LIBCPP_TEMPLATE_VIS hash +#if !defined(_LIBCPP_ABI_NO_BINDER_BASES) + : public unary_function +#endif +{ +_LIBCPP_SUPPRESS_DEPRECATED_POP +#if _LIBCPP_STD_VER <= 17 || defined(_LIBCPP_ENABLE_CXX20_REMOVED_BINDER_TYPEDEFS) + _LIBCPP_DEPRECATED_IN_CXX17 typedef size_t result_type; + _LIBCPP_DEPRECATED_IN_CXX17 typedef bool argument_type; +#endif + _LIBCPP_INLINE_VISIBILITY + size_t operator()(bool __v) const _NOEXCEPT {return static_cast(__v);} +}; + +_LIBCPP_SUPPRESS_DEPRECATED_PUSH +template <> +struct _LIBCPP_TEMPLATE_VIS hash +#if !defined(_LIBCPP_ABI_NO_BINDER_BASES) + : public unary_function +#endif +{ +_LIBCPP_SUPPRESS_DEPRECATED_POP +#if _LIBCPP_STD_VER <= 17 || defined(_LIBCPP_ENABLE_CXX20_REMOVED_BINDER_TYPEDEFS) + _LIBCPP_DEPRECATED_IN_CXX17 typedef size_t result_type; + _LIBCPP_DEPRECATED_IN_CXX17 typedef char argument_type; +#endif + _LIBCPP_INLINE_VISIBILITY + size_t operator()(char __v) const _NOEXCEPT {return static_cast(__v);} +}; + +_LIBCPP_SUPPRESS_DEPRECATED_PUSH +template <> +struct _LIBCPP_TEMPLATE_VIS hash +#if !defined(_LIBCPP_ABI_NO_BINDER_BASES) + : public unary_function +#endif +{ +_LIBCPP_SUPPRESS_DEPRECATED_POP +#if _LIBCPP_STD_VER <= 17 || defined(_LIBCPP_ENABLE_CXX20_REMOVED_BINDER_TYPEDEFS) + _LIBCPP_DEPRECATED_IN_CXX17 typedef size_t result_type; + _LIBCPP_DEPRECATED_IN_CXX17 typedef signed char argument_type; +#endif + _LIBCPP_INLINE_VISIBILITY + size_t operator()(signed char __v) const _NOEXCEPT {return static_cast(__v);} +}; + +_LIBCPP_SUPPRESS_DEPRECATED_PUSH +template <> +struct _LIBCPP_TEMPLATE_VIS hash +#if !defined(_LIBCPP_ABI_NO_BINDER_BASES) + : public unary_function +#endif +{ +_LIBCPP_SUPPRESS_DEPRECATED_POP +#if _LIBCPP_STD_VER <= 17 || defined(_LIBCPP_ENABLE_CXX20_REMOVED_BINDER_TYPEDEFS) + _LIBCPP_DEPRECATED_IN_CXX17 typedef size_t result_type; + _LIBCPP_DEPRECATED_IN_CXX17 typedef unsigned char argument_type; +#endif + _LIBCPP_INLINE_VISIBILITY + size_t operator()(unsigned char __v) const _NOEXCEPT {return static_cast(__v);} +}; + +#ifndef _LIBCPP_HAS_NO_CHAR8_T +_LIBCPP_SUPPRESS_DEPRECATED_PUSH +template <> +struct _LIBCPP_TEMPLATE_VIS hash +#if !defined(_LIBCPP_ABI_NO_BINDER_BASES) + : public unary_function +#endif +{ +_LIBCPP_SUPPRESS_DEPRECATED_POP +#if _LIBCPP_STD_VER <= 17 || defined(_LIBCPP_ENABLE_CXX20_REMOVED_BINDER_TYPEDEFS) + _LIBCPP_DEPRECATED_IN_CXX17 typedef size_t result_type; + _LIBCPP_DEPRECATED_IN_CXX17 typedef char8_t argument_type; +#endif + _LIBCPP_INLINE_VISIBILITY + size_t operator()(char8_t __v) const _NOEXCEPT {return static_cast(__v);} +}; +#endif // !_LIBCPP_HAS_NO_CHAR8_T + +#ifndef _LIBCPP_HAS_NO_UNICODE_CHARS + +_LIBCPP_SUPPRESS_DEPRECATED_PUSH +template <> +struct _LIBCPP_TEMPLATE_VIS hash +#if !defined(_LIBCPP_ABI_NO_BINDER_BASES) + : public unary_function +#endif +{ +_LIBCPP_SUPPRESS_DEPRECATED_POP +#if _LIBCPP_STD_VER <= 17 || defined(_LIBCPP_ENABLE_CXX20_REMOVED_BINDER_TYPEDEFS) + _LIBCPP_DEPRECATED_IN_CXX17 typedef size_t result_type; + _LIBCPP_DEPRECATED_IN_CXX17 typedef char16_t argument_type; +#endif + _LIBCPP_INLINE_VISIBILITY + size_t operator()(char16_t __v) const _NOEXCEPT {return static_cast(__v);} +}; + +_LIBCPP_SUPPRESS_DEPRECATED_PUSH +template <> +struct _LIBCPP_TEMPLATE_VIS hash +#if !defined(_LIBCPP_ABI_NO_BINDER_BASES) + : public unary_function +#endif +{ +_LIBCPP_SUPPRESS_DEPRECATED_POP +#if _LIBCPP_STD_VER <= 17 || defined(_LIBCPP_ENABLE_CXX20_REMOVED_BINDER_TYPEDEFS) + _LIBCPP_DEPRECATED_IN_CXX17 typedef size_t result_type; + _LIBCPP_DEPRECATED_IN_CXX17 typedef char32_t argument_type; +#endif + _LIBCPP_INLINE_VISIBILITY + size_t operator()(char32_t __v) const _NOEXCEPT {return static_cast(__v);} +}; + +#endif // _LIBCPP_HAS_NO_UNICODE_CHARS + +_LIBCPP_SUPPRESS_DEPRECATED_PUSH +template <> +struct _LIBCPP_TEMPLATE_VIS hash +#if !defined(_LIBCPP_ABI_NO_BINDER_BASES) + : public unary_function +#endif +{ +_LIBCPP_SUPPRESS_DEPRECATED_POP +#if _LIBCPP_STD_VER <= 17 || defined(_LIBCPP_ENABLE_CXX20_REMOVED_BINDER_TYPEDEFS) + _LIBCPP_DEPRECATED_IN_CXX17 typedef size_t result_type; + _LIBCPP_DEPRECATED_IN_CXX17 typedef wchar_t argument_type; +#endif + _LIBCPP_INLINE_VISIBILITY + size_t operator()(wchar_t __v) const _NOEXCEPT {return static_cast(__v);} +}; + +_LIBCPP_SUPPRESS_DEPRECATED_PUSH +template <> +struct _LIBCPP_TEMPLATE_VIS hash +#if !defined(_LIBCPP_ABI_NO_BINDER_BASES) + : public unary_function +#endif +{ +_LIBCPP_SUPPRESS_DEPRECATED_POP +#if _LIBCPP_STD_VER <= 17 || defined(_LIBCPP_ENABLE_CXX20_REMOVED_BINDER_TYPEDEFS) + _LIBCPP_DEPRECATED_IN_CXX17 typedef size_t result_type; + _LIBCPP_DEPRECATED_IN_CXX17 typedef short argument_type; +#endif + _LIBCPP_INLINE_VISIBILITY + size_t operator()(short __v) const _NOEXCEPT {return static_cast(__v);} +}; + +_LIBCPP_SUPPRESS_DEPRECATED_PUSH +template <> +struct _LIBCPP_TEMPLATE_VIS hash +#if !defined(_LIBCPP_ABI_NO_BINDER_BASES) + : public unary_function +#endif +{ +_LIBCPP_SUPPRESS_DEPRECATED_POP +#if _LIBCPP_STD_VER <= 17 || defined(_LIBCPP_ENABLE_CXX20_REMOVED_BINDER_TYPEDEFS) + _LIBCPP_DEPRECATED_IN_CXX17 typedef size_t result_type; + _LIBCPP_DEPRECATED_IN_CXX17 typedef unsigned short argument_type; +#endif + _LIBCPP_INLINE_VISIBILITY + size_t operator()(unsigned short __v) const _NOEXCEPT {return static_cast(__v);} +}; + +_LIBCPP_SUPPRESS_DEPRECATED_PUSH +template <> +struct _LIBCPP_TEMPLATE_VIS hash +#if !defined(_LIBCPP_ABI_NO_BINDER_BASES) + : public unary_function +#endif +{ +_LIBCPP_SUPPRESS_DEPRECATED_POP +#if _LIBCPP_STD_VER <= 17 || defined(_LIBCPP_ENABLE_CXX20_REMOVED_BINDER_TYPEDEFS) + _LIBCPP_DEPRECATED_IN_CXX17 typedef size_t result_type; + _LIBCPP_DEPRECATED_IN_CXX17 typedef int argument_type; +#endif + _LIBCPP_INLINE_VISIBILITY + size_t operator()(int __v) const _NOEXCEPT {return static_cast(__v);} +}; + +_LIBCPP_SUPPRESS_DEPRECATED_PUSH +template <> +struct _LIBCPP_TEMPLATE_VIS hash +#if !defined(_LIBCPP_ABI_NO_BINDER_BASES) + : public unary_function +#endif +{ +_LIBCPP_SUPPRESS_DEPRECATED_POP +#if _LIBCPP_STD_VER <= 17 || defined(_LIBCPP_ENABLE_CXX20_REMOVED_BINDER_TYPEDEFS) + _LIBCPP_DEPRECATED_IN_CXX17 typedef size_t result_type; + _LIBCPP_DEPRECATED_IN_CXX17 typedef unsigned int argument_type; +#endif + _LIBCPP_INLINE_VISIBILITY + size_t operator()(unsigned int __v) const _NOEXCEPT {return static_cast(__v);} +}; + +_LIBCPP_SUPPRESS_DEPRECATED_PUSH +template <> +struct _LIBCPP_TEMPLATE_VIS hash +#if !defined(_LIBCPP_ABI_NO_BINDER_BASES) + : public unary_function +#endif +{ +_LIBCPP_SUPPRESS_DEPRECATED_POP +#if _LIBCPP_STD_VER <= 17 || defined(_LIBCPP_ENABLE_CXX20_REMOVED_BINDER_TYPEDEFS) + _LIBCPP_DEPRECATED_IN_CXX17 typedef size_t result_type; + _LIBCPP_DEPRECATED_IN_CXX17 typedef long argument_type; +#endif + _LIBCPP_INLINE_VISIBILITY + size_t operator()(long __v) const _NOEXCEPT {return static_cast(__v);} +}; + +_LIBCPP_SUPPRESS_DEPRECATED_PUSH +template <> +struct _LIBCPP_TEMPLATE_VIS hash +#if !defined(_LIBCPP_ABI_NO_BINDER_BASES) + : public unary_function +#endif +{ +_LIBCPP_SUPPRESS_DEPRECATED_POP +#if _LIBCPP_STD_VER <= 17 || defined(_LIBCPP_ENABLE_CXX20_REMOVED_BINDER_TYPEDEFS) + _LIBCPP_DEPRECATED_IN_CXX17 typedef size_t result_type; + _LIBCPP_DEPRECATED_IN_CXX17 typedef unsigned long argument_type; +#endif + _LIBCPP_INLINE_VISIBILITY + size_t operator()(unsigned long __v) const _NOEXCEPT {return static_cast(__v);} +}; + +template <> +struct _LIBCPP_TEMPLATE_VIS hash + : public __scalar_hash +{ +}; + +template <> +struct _LIBCPP_TEMPLATE_VIS hash + : public __scalar_hash +{ +}; + +#ifndef _LIBCPP_HAS_NO_INT128 + +template <> +struct _LIBCPP_TEMPLATE_VIS hash<__int128_t> + : public __scalar_hash<__int128_t> +{ +}; + +template <> +struct _LIBCPP_TEMPLATE_VIS hash<__uint128_t> + : public __scalar_hash<__uint128_t> +{ +}; + +#endif + +template <> +struct _LIBCPP_TEMPLATE_VIS hash + : public __scalar_hash +{ + _LIBCPP_INLINE_VISIBILITY + size_t operator()(float __v) const _NOEXCEPT + { + // -0.0 and 0.0 should return same hash + if (__v == 0.0f) + return 0; + return __scalar_hash::operator()(__v); + } +}; + +template <> +struct _LIBCPP_TEMPLATE_VIS hash + : public __scalar_hash +{ + _LIBCPP_INLINE_VISIBILITY + size_t operator()(double __v) const _NOEXCEPT + { + // -0.0 and 0.0 should return same hash + if (__v == 0.0) + return 0; + return __scalar_hash::operator()(__v); + } +}; + +template <> +struct _LIBCPP_TEMPLATE_VIS hash + : public __scalar_hash +{ + _LIBCPP_INLINE_VISIBILITY + size_t operator()(long double __v) const _NOEXCEPT + { + // -0.0 and 0.0 should return same hash + if (__v == 0.0L) + return 0; +#if defined(__i386__) || (defined(__x86_64__) && defined(__ILP32__)) + // Zero out padding bits + union + { + long double __t; + struct + { + size_t __a; + size_t __b; + size_t __c; + size_t __d; + } __s; + } __u; + __u.__s.__a = 0; + __u.__s.__b = 0; + __u.__s.__c = 0; + __u.__s.__d = 0; + __u.__t = __v; + return __u.__s.__a ^ __u.__s.__b ^ __u.__s.__c ^ __u.__s.__d; +#elif defined(__x86_64__) + // Zero out padding bits + union + { + long double __t; + struct + { + size_t __a; + size_t __b; + } __s; + } __u; + __u.__s.__a = 0; + __u.__s.__b = 0; + __u.__t = __v; + return __u.__s.__a ^ __u.__s.__b; +#else + return __scalar_hash::operator()(__v); +#endif + } +}; + +#if _LIBCPP_STD_VER > 11 + +_LIBCPP_SUPPRESS_DEPRECATED_PUSH +template ::value> +struct _LIBCPP_TEMPLATE_VIS __enum_hash +#if !defined(_LIBCPP_ABI_NO_BINDER_BASES) + : public unary_function<_Tp, size_t> +#endif +{ +_LIBCPP_SUPPRESS_DEPRECATED_POP +#if _LIBCPP_STD_VER <= 17 || defined(_LIBCPP_ENABLE_CXX20_REMOVED_BINDER_TYPEDEFS) + _LIBCPP_DEPRECATED_IN_CXX17 typedef size_t result_type; + _LIBCPP_DEPRECATED_IN_CXX17 typedef _Tp argument_type; +#endif + _LIBCPP_INLINE_VISIBILITY + size_t operator()(_Tp __v) const _NOEXCEPT + { + typedef typename underlying_type<_Tp>::type type; + return hash{}(static_cast(__v)); + } +}; +template +struct _LIBCPP_TEMPLATE_VIS __enum_hash<_Tp, false> { + __enum_hash() = delete; + __enum_hash(__enum_hash const&) = delete; + __enum_hash& operator=(__enum_hash const&) = delete; +}; + +template +struct _LIBCPP_TEMPLATE_VIS hash : public __enum_hash<_Tp> +{ +}; +#endif + +#if _LIBCPP_STD_VER > 14 + +_LIBCPP_SUPPRESS_DEPRECATED_PUSH +template <> +struct _LIBCPP_TEMPLATE_VIS hash +#if !defined(_LIBCPP_ABI_NO_BINDER_BASES) + : public unary_function +#endif +{ +_LIBCPP_SUPPRESS_DEPRECATED_POP +#if _LIBCPP_STD_VER <= 17 || defined(_LIBCPP_ENABLE_CXX20_REMOVED_BINDER_TYPEDEFS) + _LIBCPP_DEPRECATED_IN_CXX17 typedef size_t result_type; + _LIBCPP_DEPRECATED_IN_CXX17 typedef nullptr_t argument_type; +#endif + _LIBCPP_INLINE_VISIBILITY + size_t operator()(nullptr_t) const _NOEXCEPT { + return 662607004ull; + } +}; +#endif + +#ifndef _LIBCPP_CXX03_LANG +template +using __check_hash_requirements _LIBCPP_NODEBUG_TYPE = integral_constant::value && + is_move_constructible<_Hash>::value && + __invokable_r::value +>; + +template > +using __has_enabled_hash _LIBCPP_NODEBUG_TYPE = integral_constant::value && + is_default_constructible<_Hash>::value +>; + +#if _LIBCPP_STD_VER > 14 +template +using __enable_hash_helper_imp _LIBCPP_NODEBUG_TYPE = _Type; + +template +using __enable_hash_helper _LIBCPP_NODEBUG_TYPE = __enable_hash_helper_imp<_Type, + typename enable_if<__all<__has_enabled_hash<_Keys>::value...>::value>::type +>; +#else +template +using __enable_hash_helper _LIBCPP_NODEBUG_TYPE = _Type; +#endif + +#endif // !_LIBCPP_CXX03_LANG + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___FUNCTIONAL_HASH_H diff --git a/lib/libcxx/include/__functional/identity.h b/lib/libcxx/include/__functional/identity.h new file mode 100644 index 000000000..6b8346b3b --- /dev/null +++ b/lib/libcxx/include/__functional/identity.h @@ -0,0 +1,37 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___FUNCTIONAL_IDENTITY_H +#define _LIBCPP___FUNCTIONAL_IDENTITY_H + +#include <__config> +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_BEGIN_NAMESPACE_STD + +#if _LIBCPP_STD_VER > 17 + +struct identity { + template + _LIBCPP_NODISCARD_EXT constexpr _Tp&& operator()(_Tp&& __t) const noexcept + { + return _VSTD::forward<_Tp>(__t); + } + + using is_transparent = void; +}; +#endif // _LIBCPP_STD_VER > 17 + +_LIBCPP_END_NAMESPACE_STD + +#endif // _LIBCPP___FUNCTIONAL_IDENTITY_H diff --git a/lib/libcxx/include/__functional/invoke.h b/lib/libcxx/include/__functional/invoke.h new file mode 100644 index 000000000..0e167c75d --- /dev/null +++ b/lib/libcxx/include/__functional/invoke.h @@ -0,0 +1,100 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___FUNCTIONAL_INVOKE_H +#define _LIBCPP___FUNCTIONAL_INVOKE_H + +#include <__config> +#include <__functional/weak_result_type.h> +#include <__utility/forward.h> +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_BEGIN_NAMESPACE_STD + +template ::value> +struct __invoke_void_return_wrapper +{ +#ifndef _LIBCPP_CXX03_LANG + template + static _Ret __call(_Args&&... __args) { + return _VSTD::__invoke(_VSTD::forward<_Args>(__args)...); + } +#else + template + static _Ret __call(_Fn __f) { + return _VSTD::__invoke(__f); + } + + template + static _Ret __call(_Fn __f, _A0& __a0) { + return _VSTD::__invoke(__f, __a0); + } + + template + static _Ret __call(_Fn __f, _A0& __a0, _A1& __a1) { + return _VSTD::__invoke(__f, __a0, __a1); + } + + template + static _Ret __call(_Fn __f, _A0& __a0, _A1& __a1, _A2& __a2){ + return _VSTD::__invoke(__f, __a0, __a1, __a2); + } +#endif +}; + +template +struct __invoke_void_return_wrapper<_Ret, true> +{ +#ifndef _LIBCPP_CXX03_LANG + template + static void __call(_Args&&... __args) { + _VSTD::__invoke(_VSTD::forward<_Args>(__args)...); + } +#else + template + static void __call(_Fn __f) { + _VSTD::__invoke(__f); + } + + template + static void __call(_Fn __f, _A0& __a0) { + _VSTD::__invoke(__f, __a0); + } + + template + static void __call(_Fn __f, _A0& __a0, _A1& __a1) { + _VSTD::__invoke(__f, __a0, __a1); + } + + template + static void __call(_Fn __f, _A0& __a0, _A1& __a1, _A2& __a2) { + _VSTD::__invoke(__f, __a0, __a1, __a2); + } +#endif +}; + +#if _LIBCPP_STD_VER > 14 + +template +_LIBCPP_CONSTEXPR_AFTER_CXX17 invoke_result_t<_Fn, _Args...> +invoke(_Fn&& __f, _Args&&... __args) + noexcept(is_nothrow_invocable_v<_Fn, _Args...>) +{ + return _VSTD::__invoke(_VSTD::forward<_Fn>(__f), _VSTD::forward<_Args>(__args)...); +} + +#endif // _LIBCPP_STD_VER > 14 + +_LIBCPP_END_NAMESPACE_STD + +#endif // _LIBCPP___FUNCTIONAL_INVOKE_H diff --git a/lib/libcxx/include/__functional/is_transparent.h b/lib/libcxx/include/__functional/is_transparent.h new file mode 100644 index 000000000..4a72aa8e2 --- /dev/null +++ b/lib/libcxx/include/__functional/is_transparent.h @@ -0,0 +1,36 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___FUNCTIONAL_IS_TRANSPARENT +#define _LIBCPP___FUNCTIONAL_IS_TRANSPARENT + +#include <__config> +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_BEGIN_NAMESPACE_STD + +#if _LIBCPP_STD_VER > 11 + +template +struct __is_transparent : false_type {}; + +template +struct __is_transparent<_Tp, _Up, + typename __void_t::type> + : true_type {}; + +#endif + +_LIBCPP_END_NAMESPACE_STD + +#endif // _LIBCPP___FUNCTIONAL_IS_TRANSPARENT diff --git a/lib/libcxx/include/__functional/mem_fn.h b/lib/libcxx/include/__functional/mem_fn.h new file mode 100644 index 000000000..1fa070a42 --- /dev/null +++ b/lib/libcxx/include/__functional/mem_fn.h @@ -0,0 +1,161 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___FUNCTIONAL_MEM_FN_H +#define _LIBCPP___FUNCTIONAL_MEM_FN_H + +#include <__config> +#include <__functional/weak_result_type.h> +#include <__functional/binary_function.h> +#include <__functional/invoke.h> +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +class __mem_fn +#if _LIBCPP_STD_VER <= 17 || !defined(_LIBCPP_ABI_NO_BINDER_BASES) + : public __weak_result_type<_Tp> +#endif +{ +public: + // types + typedef _Tp type; +private: + type __f_; + +public: + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 + __mem_fn(type __f) _NOEXCEPT : __f_(__f) {} + +#ifndef _LIBCPP_CXX03_LANG + // invoke + template + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 + typename __invoke_return::type + operator() (_ArgTypes&&... __args) const { + return _VSTD::__invoke(__f_, _VSTD::forward<_ArgTypes>(__args)...); + } +#else + + template + _LIBCPP_INLINE_VISIBILITY + typename __invoke_return0::type + operator() (_A0& __a0) const { + return _VSTD::__invoke(__f_, __a0); + } + + template + _LIBCPP_INLINE_VISIBILITY + typename __invoke_return0::type + operator() (_A0 const& __a0) const { + return _VSTD::__invoke(__f_, __a0); + } + + template + _LIBCPP_INLINE_VISIBILITY + typename __invoke_return1::type + operator() (_A0& __a0, _A1& __a1) const { + return _VSTD::__invoke(__f_, __a0, __a1); + } + + template + _LIBCPP_INLINE_VISIBILITY + typename __invoke_return1::type + operator() (_A0 const& __a0, _A1& __a1) const { + return _VSTD::__invoke(__f_, __a0, __a1); + } + + template + _LIBCPP_INLINE_VISIBILITY + typename __invoke_return1::type + operator() (_A0& __a0, _A1 const& __a1) const { + return _VSTD::__invoke(__f_, __a0, __a1); + } + + template + _LIBCPP_INLINE_VISIBILITY + typename __invoke_return1::type + operator() (_A0 const& __a0, _A1 const& __a1) const { + return _VSTD::__invoke(__f_, __a0, __a1); + } + + template + _LIBCPP_INLINE_VISIBILITY + typename __invoke_return2::type + operator() (_A0& __a0, _A1& __a1, _A2& __a2) const { + return _VSTD::__invoke(__f_, __a0, __a1, __a2); + } + + template + _LIBCPP_INLINE_VISIBILITY + typename __invoke_return2::type + operator() (_A0 const& __a0, _A1& __a1, _A2& __a2) const { + return _VSTD::__invoke(__f_, __a0, __a1, __a2); + } + + template + _LIBCPP_INLINE_VISIBILITY + typename __invoke_return2::type + operator() (_A0& __a0, _A1 const& __a1, _A2& __a2) const { + return _VSTD::__invoke(__f_, __a0, __a1, __a2); + } + + template + _LIBCPP_INLINE_VISIBILITY + typename __invoke_return2::type + operator() (_A0& __a0, _A1& __a1, _A2 const& __a2) const { + return _VSTD::__invoke(__f_, __a0, __a1, __a2); + } + + template + _LIBCPP_INLINE_VISIBILITY + typename __invoke_return2::type + operator() (_A0 const& __a0, _A1 const& __a1, _A2& __a2) const { + return _VSTD::__invoke(__f_, __a0, __a1, __a2); + } + + template + _LIBCPP_INLINE_VISIBILITY + typename __invoke_return2::type + operator() (_A0 const& __a0, _A1& __a1, _A2 const& __a2) const { + return _VSTD::__invoke(__f_, __a0, __a1, __a2); + } + + template + _LIBCPP_INLINE_VISIBILITY + typename __invoke_return2::type + operator() (_A0& __a0, _A1 const& __a1, _A2 const& __a2) const { + return _VSTD::__invoke(__f_, __a0, __a1, __a2); + } + + template + _LIBCPP_INLINE_VISIBILITY + typename __invoke_return2::type + operator() (_A0 const& __a0, _A1 const& __a1, _A2 const& __a2) const { + return _VSTD::__invoke(__f_, __a0, __a1, __a2); + } +#endif +}; + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +__mem_fn<_Rp _Tp::*> +mem_fn(_Rp _Tp::* __pm) _NOEXCEPT +{ + return __mem_fn<_Rp _Tp::*>(__pm); +} + +_LIBCPP_END_NAMESPACE_STD + +#endif // _LIBCPP___FUNCTIONAL_MEM_FN_H diff --git a/lib/libcxx/include/__functional/mem_fun_ref.h b/lib/libcxx/include/__functional/mem_fun_ref.h new file mode 100644 index 000000000..4616da0b0 --- /dev/null +++ b/lib/libcxx/include/__functional/mem_fun_ref.h @@ -0,0 +1,173 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___FUNCTIONAL_MEM_FUN_REF_H +#define _LIBCPP___FUNCTIONAL_MEM_FUN_REF_H + +#include <__config> +#include <__functional/unary_function.h> +#include <__functional/binary_function.h> + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_BEGIN_NAMESPACE_STD + +#if _LIBCPP_STD_VER <= 14 || defined(_LIBCPP_ENABLE_CXX17_REMOVED_BINDERS) + +template +class _LIBCPP_TEMPLATE_VIS _LIBCPP_DEPRECATED_IN_CXX11 mem_fun_t + : public unary_function<_Tp*, _Sp> +{ + _Sp (_Tp::*__p_)(); +public: + _LIBCPP_INLINE_VISIBILITY explicit mem_fun_t(_Sp (_Tp::*__p)()) + : __p_(__p) {} + _LIBCPP_INLINE_VISIBILITY _Sp operator()(_Tp* __p) const + {return (__p->*__p_)();} +}; + +template +class _LIBCPP_TEMPLATE_VIS _LIBCPP_DEPRECATED_IN_CXX11 mem_fun1_t + : public binary_function<_Tp*, _Ap, _Sp> +{ + _Sp (_Tp::*__p_)(_Ap); +public: + _LIBCPP_INLINE_VISIBILITY explicit mem_fun1_t(_Sp (_Tp::*__p)(_Ap)) + : __p_(__p) {} + _LIBCPP_INLINE_VISIBILITY _Sp operator()(_Tp* __p, _Ap __x) const + {return (__p->*__p_)(__x);} +}; + +template +_LIBCPP_DEPRECATED_IN_CXX11 inline _LIBCPP_INLINE_VISIBILITY +mem_fun_t<_Sp,_Tp> +mem_fun(_Sp (_Tp::*__f)()) + {return mem_fun_t<_Sp,_Tp>(__f);} + +template +_LIBCPP_DEPRECATED_IN_CXX11 inline _LIBCPP_INLINE_VISIBILITY +mem_fun1_t<_Sp,_Tp,_Ap> +mem_fun(_Sp (_Tp::*__f)(_Ap)) + {return mem_fun1_t<_Sp,_Tp,_Ap>(__f);} + +template +class _LIBCPP_TEMPLATE_VIS _LIBCPP_DEPRECATED_IN_CXX11 mem_fun_ref_t + : public unary_function<_Tp, _Sp> +{ + _Sp (_Tp::*__p_)(); +public: + _LIBCPP_INLINE_VISIBILITY explicit mem_fun_ref_t(_Sp (_Tp::*__p)()) + : __p_(__p) {} + _LIBCPP_INLINE_VISIBILITY _Sp operator()(_Tp& __p) const + {return (__p.*__p_)();} +}; + +template +class _LIBCPP_TEMPLATE_VIS _LIBCPP_DEPRECATED_IN_CXX11 mem_fun1_ref_t + : public binary_function<_Tp, _Ap, _Sp> +{ + _Sp (_Tp::*__p_)(_Ap); +public: + _LIBCPP_INLINE_VISIBILITY explicit mem_fun1_ref_t(_Sp (_Tp::*__p)(_Ap)) + : __p_(__p) {} + _LIBCPP_INLINE_VISIBILITY _Sp operator()(_Tp& __p, _Ap __x) const + {return (__p.*__p_)(__x);} +}; + +template +_LIBCPP_DEPRECATED_IN_CXX11 inline _LIBCPP_INLINE_VISIBILITY +mem_fun_ref_t<_Sp,_Tp> +mem_fun_ref(_Sp (_Tp::*__f)()) + {return mem_fun_ref_t<_Sp,_Tp>(__f);} + +template +_LIBCPP_DEPRECATED_IN_CXX11 inline _LIBCPP_INLINE_VISIBILITY +mem_fun1_ref_t<_Sp,_Tp,_Ap> +mem_fun_ref(_Sp (_Tp::*__f)(_Ap)) + {return mem_fun1_ref_t<_Sp,_Tp,_Ap>(__f);} + +template +class _LIBCPP_TEMPLATE_VIS _LIBCPP_DEPRECATED_IN_CXX11 const_mem_fun_t + : public unary_function +{ + _Sp (_Tp::*__p_)() const; +public: + _LIBCPP_INLINE_VISIBILITY explicit const_mem_fun_t(_Sp (_Tp::*__p)() const) + : __p_(__p) {} + _LIBCPP_INLINE_VISIBILITY _Sp operator()(const _Tp* __p) const + {return (__p->*__p_)();} +}; + +template +class _LIBCPP_TEMPLATE_VIS _LIBCPP_DEPRECATED_IN_CXX11 const_mem_fun1_t + : public binary_function +{ + _Sp (_Tp::*__p_)(_Ap) const; +public: + _LIBCPP_INLINE_VISIBILITY explicit const_mem_fun1_t(_Sp (_Tp::*__p)(_Ap) const) + : __p_(__p) {} + _LIBCPP_INLINE_VISIBILITY _Sp operator()(const _Tp* __p, _Ap __x) const + {return (__p->*__p_)(__x);} +}; + +template +_LIBCPP_DEPRECATED_IN_CXX11 inline _LIBCPP_INLINE_VISIBILITY +const_mem_fun_t<_Sp,_Tp> +mem_fun(_Sp (_Tp::*__f)() const) + {return const_mem_fun_t<_Sp,_Tp>(__f);} + +template +_LIBCPP_DEPRECATED_IN_CXX11 inline _LIBCPP_INLINE_VISIBILITY +const_mem_fun1_t<_Sp,_Tp,_Ap> +mem_fun(_Sp (_Tp::*__f)(_Ap) const) + {return const_mem_fun1_t<_Sp,_Tp,_Ap>(__f);} + +template +class _LIBCPP_TEMPLATE_VIS _LIBCPP_DEPRECATED_IN_CXX11 const_mem_fun_ref_t + : public unary_function<_Tp, _Sp> +{ + _Sp (_Tp::*__p_)() const; +public: + _LIBCPP_INLINE_VISIBILITY explicit const_mem_fun_ref_t(_Sp (_Tp::*__p)() const) + : __p_(__p) {} + _LIBCPP_INLINE_VISIBILITY _Sp operator()(const _Tp& __p) const + {return (__p.*__p_)();} +}; + +template +class _LIBCPP_TEMPLATE_VIS _LIBCPP_DEPRECATED_IN_CXX11 const_mem_fun1_ref_t + : public binary_function<_Tp, _Ap, _Sp> +{ + _Sp (_Tp::*__p_)(_Ap) const; +public: + _LIBCPP_INLINE_VISIBILITY explicit const_mem_fun1_ref_t(_Sp (_Tp::*__p)(_Ap) const) + : __p_(__p) {} + _LIBCPP_INLINE_VISIBILITY _Sp operator()(const _Tp& __p, _Ap __x) const + {return (__p.*__p_)(__x);} +}; + +template +_LIBCPP_DEPRECATED_IN_CXX11 inline _LIBCPP_INLINE_VISIBILITY +const_mem_fun_ref_t<_Sp,_Tp> +mem_fun_ref(_Sp (_Tp::*__f)() const) + {return const_mem_fun_ref_t<_Sp,_Tp>(__f);} + +template +_LIBCPP_DEPRECATED_IN_CXX11 inline _LIBCPP_INLINE_VISIBILITY +const_mem_fun1_ref_t<_Sp,_Tp,_Ap> +mem_fun_ref(_Sp (_Tp::*__f)(_Ap) const) + {return const_mem_fun1_ref_t<_Sp,_Tp,_Ap>(__f);} + +#endif // _LIBCPP_STD_VER <= 14 || defined(_LIBCPP_ENABLE_CXX17_REMOVED_BINDERS) + +_LIBCPP_END_NAMESPACE_STD + +#endif // _LIBCPP___FUNCTIONAL_MEM_FUN_REF_H diff --git a/lib/libcxx/include/__functional/not_fn.h b/lib/libcxx/include/__functional/not_fn.h new file mode 100644 index 000000000..632be5ff0 --- /dev/null +++ b/lib/libcxx/include/__functional/not_fn.h @@ -0,0 +1,47 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___FUNCTIONAL_NOT_FN_H +#define _LIBCPP___FUNCTIONAL_NOT_FN_H + +#include <__config> +#include <__functional/perfect_forward.h> +#include <__functional/invoke.h> +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_BEGIN_NAMESPACE_STD + +#if _LIBCPP_STD_VER > 14 + +struct __not_fn_op +{ + template + static _LIBCPP_CONSTEXPR_AFTER_CXX17 auto __call(_Args&&... __args) + noexcept(noexcept(!_VSTD::invoke(_VSTD::forward<_Args>(__args)...))) + -> decltype( !_VSTD::invoke(_VSTD::forward<_Args>(__args)...)) + { return !_VSTD::invoke(_VSTD::forward<_Args>(__args)...); } +}; + +template, _Fn> && + is_move_constructible_v<_Fn>>> +_LIBCPP_CONSTEXPR_AFTER_CXX17 auto not_fn(_Fn&& __f) +{ + return __perfect_forward<__not_fn_op, _Fn>(_VSTD::forward<_Fn>(__f)); +} + +#endif // _LIBCPP_STD_VER > 14 + +_LIBCPP_END_NAMESPACE_STD + +#endif // _LIBCPP___FUNCTIONAL_NOT_FN_H diff --git a/lib/libcxx/include/__functional/operations.h b/lib/libcxx/include/__functional/operations.h new file mode 100644 index 000000000..667d17988 --- /dev/null +++ b/lib/libcxx/include/__functional/operations.h @@ -0,0 +1,729 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___FUNCTIONAL_OPERATIONS_H +#define _LIBCPP___FUNCTIONAL_OPERATIONS_H + +#include <__config> +#include <__functional/binary_function.h> +#include <__functional/unary_function.h> +#include <__utility/forward.h> + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_BEGIN_NAMESPACE_STD + +// Arithmetic operations + +_LIBCPP_SUPPRESS_DEPRECATED_PUSH +#if _LIBCPP_STD_VER > 11 +template +#else +template +#endif +struct _LIBCPP_TEMPLATE_VIS plus +#if !defined(_LIBCPP_ABI_NO_BINDER_BASES) + : binary_function<_Tp, _Tp, _Tp> +#endif +{ +_LIBCPP_SUPPRESS_DEPRECATED_POP + typedef _Tp __result_type; // used by valarray +#if _LIBCPP_STD_VER <= 17 || defined(_LIBCPP_ENABLE_CXX20_REMOVED_BINDER_TYPEDEFS) + _LIBCPP_DEPRECATED_IN_CXX17 typedef _Tp result_type; + _LIBCPP_DEPRECATED_IN_CXX17 typedef _Tp first_argument_type; + _LIBCPP_DEPRECATED_IN_CXX17 typedef _Tp second_argument_type; +#endif + _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY + _Tp operator()(const _Tp& __x, const _Tp& __y) const + {return __x + __y;} +}; + +#if _LIBCPP_STD_VER > 11 +template <> +struct _LIBCPP_TEMPLATE_VIS plus +{ + template + _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY + auto operator()(_T1&& __t, _T2&& __u) const + _NOEXCEPT_(noexcept(_VSTD::forward<_T1>(__t) + _VSTD::forward<_T2>(__u))) + -> decltype (_VSTD::forward<_T1>(__t) + _VSTD::forward<_T2>(__u)) + { return _VSTD::forward<_T1>(__t) + _VSTD::forward<_T2>(__u); } + typedef void is_transparent; +}; +#endif + +_LIBCPP_SUPPRESS_DEPRECATED_PUSH +#if _LIBCPP_STD_VER > 11 +template +#else +template +#endif +struct _LIBCPP_TEMPLATE_VIS minus +#if !defined(_LIBCPP_ABI_NO_BINDER_BASES) + : binary_function<_Tp, _Tp, _Tp> +#endif +{ +_LIBCPP_SUPPRESS_DEPRECATED_POP + typedef _Tp __result_type; // used by valarray +#if _LIBCPP_STD_VER <= 17 || defined(_LIBCPP_ENABLE_CXX20_REMOVED_BINDER_TYPEDEFS) + _LIBCPP_DEPRECATED_IN_CXX17 typedef _Tp result_type; + _LIBCPP_DEPRECATED_IN_CXX17 typedef _Tp first_argument_type; + _LIBCPP_DEPRECATED_IN_CXX17 typedef _Tp second_argument_type; +#endif + _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY + _Tp operator()(const _Tp& __x, const _Tp& __y) const + {return __x - __y;} +}; + +#if _LIBCPP_STD_VER > 11 +template <> +struct _LIBCPP_TEMPLATE_VIS minus +{ + template + _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY + auto operator()(_T1&& __t, _T2&& __u) const + _NOEXCEPT_(noexcept(_VSTD::forward<_T1>(__t) - _VSTD::forward<_T2>(__u))) + -> decltype (_VSTD::forward<_T1>(__t) - _VSTD::forward<_T2>(__u)) + { return _VSTD::forward<_T1>(__t) - _VSTD::forward<_T2>(__u); } + typedef void is_transparent; +}; +#endif + +_LIBCPP_SUPPRESS_DEPRECATED_PUSH +#if _LIBCPP_STD_VER > 11 +template +#else +template +#endif +struct _LIBCPP_TEMPLATE_VIS multiplies +#if !defined(_LIBCPP_ABI_NO_BINDER_BASES) + : binary_function<_Tp, _Tp, _Tp> +#endif +{ +_LIBCPP_SUPPRESS_DEPRECATED_POP + typedef _Tp __result_type; // used by valarray +#if _LIBCPP_STD_VER <= 17 || defined(_LIBCPP_ENABLE_CXX20_REMOVED_BINDER_TYPEDEFS) + _LIBCPP_DEPRECATED_IN_CXX17 typedef _Tp result_type; + _LIBCPP_DEPRECATED_IN_CXX17 typedef _Tp first_argument_type; + _LIBCPP_DEPRECATED_IN_CXX17 typedef _Tp second_argument_type; +#endif + _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY + _Tp operator()(const _Tp& __x, const _Tp& __y) const + {return __x * __y;} +}; + +#if _LIBCPP_STD_VER > 11 +template <> +struct _LIBCPP_TEMPLATE_VIS multiplies +{ + template + _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY + auto operator()(_T1&& __t, _T2&& __u) const + _NOEXCEPT_(noexcept(_VSTD::forward<_T1>(__t) * _VSTD::forward<_T2>(__u))) + -> decltype (_VSTD::forward<_T1>(__t) * _VSTD::forward<_T2>(__u)) + { return _VSTD::forward<_T1>(__t) * _VSTD::forward<_T2>(__u); } + typedef void is_transparent; +}; +#endif + +_LIBCPP_SUPPRESS_DEPRECATED_PUSH +#if _LIBCPP_STD_VER > 11 +template +#else +template +#endif +struct _LIBCPP_TEMPLATE_VIS divides +#if !defined(_LIBCPP_ABI_NO_BINDER_BASES) + : binary_function<_Tp, _Tp, _Tp> +#endif +{ +_LIBCPP_SUPPRESS_DEPRECATED_POP + typedef _Tp __result_type; // used by valarray +#if _LIBCPP_STD_VER <= 17 || defined(_LIBCPP_ENABLE_CXX20_REMOVED_BINDER_TYPEDEFS) + _LIBCPP_DEPRECATED_IN_CXX17 typedef _Tp result_type; + _LIBCPP_DEPRECATED_IN_CXX17 typedef _Tp first_argument_type; + _LIBCPP_DEPRECATED_IN_CXX17 typedef _Tp second_argument_type; +#endif + _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY + _Tp operator()(const _Tp& __x, const _Tp& __y) const + {return __x / __y;} +}; + +#if _LIBCPP_STD_VER > 11 +template <> +struct _LIBCPP_TEMPLATE_VIS divides +{ + template + _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY + auto operator()(_T1&& __t, _T2&& __u) const + _NOEXCEPT_(noexcept(_VSTD::forward<_T1>(__t) / _VSTD::forward<_T2>(__u))) + -> decltype (_VSTD::forward<_T1>(__t) / _VSTD::forward<_T2>(__u)) + { return _VSTD::forward<_T1>(__t) / _VSTD::forward<_T2>(__u); } + typedef void is_transparent; +}; +#endif + +_LIBCPP_SUPPRESS_DEPRECATED_PUSH +#if _LIBCPP_STD_VER > 11 +template +#else +template +#endif +struct _LIBCPP_TEMPLATE_VIS modulus +#if !defined(_LIBCPP_ABI_NO_BINDER_BASES) + : binary_function<_Tp, _Tp, _Tp> +#endif +{ +_LIBCPP_SUPPRESS_DEPRECATED_POP + typedef _Tp __result_type; // used by valarray +#if _LIBCPP_STD_VER <= 17 || defined(_LIBCPP_ENABLE_CXX20_REMOVED_BINDER_TYPEDEFS) + _LIBCPP_DEPRECATED_IN_CXX17 typedef _Tp result_type; + _LIBCPP_DEPRECATED_IN_CXX17 typedef _Tp first_argument_type; + _LIBCPP_DEPRECATED_IN_CXX17 typedef _Tp second_argument_type; +#endif + _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY + _Tp operator()(const _Tp& __x, const _Tp& __y) const + {return __x % __y;} +}; + +#if _LIBCPP_STD_VER > 11 +template <> +struct _LIBCPP_TEMPLATE_VIS modulus +{ + template + _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY + auto operator()(_T1&& __t, _T2&& __u) const + _NOEXCEPT_(noexcept(_VSTD::forward<_T1>(__t) % _VSTD::forward<_T2>(__u))) + -> decltype (_VSTD::forward<_T1>(__t) % _VSTD::forward<_T2>(__u)) + { return _VSTD::forward<_T1>(__t) % _VSTD::forward<_T2>(__u); } + typedef void is_transparent; +}; +#endif + +_LIBCPP_SUPPRESS_DEPRECATED_PUSH +#if _LIBCPP_STD_VER > 11 +template +#else +template +#endif +struct _LIBCPP_TEMPLATE_VIS negate +#if !defined(_LIBCPP_ABI_NO_BINDER_BASES) + : unary_function<_Tp, _Tp> +#endif +{ +_LIBCPP_SUPPRESS_DEPRECATED_POP + typedef _Tp __result_type; // used by valarray +#if _LIBCPP_STD_VER <= 17 || defined(_LIBCPP_ENABLE_CXX20_REMOVED_BINDER_TYPEDEFS) + _LIBCPP_DEPRECATED_IN_CXX17 typedef _Tp result_type; + _LIBCPP_DEPRECATED_IN_CXX17 typedef _Tp argument_type; +#endif + _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY + _Tp operator()(const _Tp& __x) const + {return -__x;} +}; + +#if _LIBCPP_STD_VER > 11 +template <> +struct _LIBCPP_TEMPLATE_VIS negate +{ + template + _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY + auto operator()(_Tp&& __x) const + _NOEXCEPT_(noexcept(- _VSTD::forward<_Tp>(__x))) + -> decltype (- _VSTD::forward<_Tp>(__x)) + { return - _VSTD::forward<_Tp>(__x); } + typedef void is_transparent; +}; +#endif + +// Bitwise operations + +_LIBCPP_SUPPRESS_DEPRECATED_PUSH +#if _LIBCPP_STD_VER > 11 +template +#else +template +#endif +struct _LIBCPP_TEMPLATE_VIS bit_and +#if !defined(_LIBCPP_ABI_NO_BINDER_BASES) + : binary_function<_Tp, _Tp, _Tp> +#endif +{ +_LIBCPP_SUPPRESS_DEPRECATED_POP + typedef _Tp __result_type; // used by valarray +#if _LIBCPP_STD_VER <= 17 || defined(_LIBCPP_ENABLE_CXX20_REMOVED_BINDER_TYPEDEFS) + _LIBCPP_DEPRECATED_IN_CXX17 typedef _Tp result_type; + _LIBCPP_DEPRECATED_IN_CXX17 typedef _Tp first_argument_type; + _LIBCPP_DEPRECATED_IN_CXX17 typedef _Tp second_argument_type; +#endif + _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY + _Tp operator()(const _Tp& __x, const _Tp& __y) const + {return __x & __y;} +}; + +#if _LIBCPP_STD_VER > 11 +template <> +struct _LIBCPP_TEMPLATE_VIS bit_and +{ + template + _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY + auto operator()(_T1&& __t, _T2&& __u) const + _NOEXCEPT_(noexcept(_VSTD::forward<_T1>(__t) & _VSTD::forward<_T2>(__u))) + -> decltype (_VSTD::forward<_T1>(__t) & _VSTD::forward<_T2>(__u)) + { return _VSTD::forward<_T1>(__t) & _VSTD::forward<_T2>(__u); } + typedef void is_transparent; +}; +#endif + +#if _LIBCPP_STD_VER > 11 +_LIBCPP_SUPPRESS_DEPRECATED_PUSH +template +struct _LIBCPP_TEMPLATE_VIS bit_not +#if !defined(_LIBCPP_ABI_NO_BINDER_BASES) + : unary_function<_Tp, _Tp> +#endif +{ +_LIBCPP_SUPPRESS_DEPRECATED_POP +#if _LIBCPP_STD_VER <= 17 || defined(_LIBCPP_ENABLE_CXX20_REMOVED_BINDER_TYPEDEFS) + _LIBCPP_DEPRECATED_IN_CXX17 typedef _Tp result_type; + _LIBCPP_DEPRECATED_IN_CXX17 typedef _Tp argument_type; +#endif + _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY + _Tp operator()(const _Tp& __x) const + {return ~__x;} +}; + +template <> +struct _LIBCPP_TEMPLATE_VIS bit_not +{ + template + _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY + auto operator()(_Tp&& __x) const + _NOEXCEPT_(noexcept(~_VSTD::forward<_Tp>(__x))) + -> decltype (~_VSTD::forward<_Tp>(__x)) + { return ~_VSTD::forward<_Tp>(__x); } + typedef void is_transparent; +}; +#endif + +_LIBCPP_SUPPRESS_DEPRECATED_PUSH +#if _LIBCPP_STD_VER > 11 +template +#else +template +#endif +struct _LIBCPP_TEMPLATE_VIS bit_or +#if !defined(_LIBCPP_ABI_NO_BINDER_BASES) + : binary_function<_Tp, _Tp, _Tp> +#endif +{ +_LIBCPP_SUPPRESS_DEPRECATED_POP + typedef _Tp __result_type; // used by valarray +#if _LIBCPP_STD_VER <= 17 || defined(_LIBCPP_ENABLE_CXX20_REMOVED_BINDER_TYPEDEFS) + _LIBCPP_DEPRECATED_IN_CXX17 typedef _Tp result_type; + _LIBCPP_DEPRECATED_IN_CXX17 typedef _Tp first_argument_type; + _LIBCPP_DEPRECATED_IN_CXX17 typedef _Tp second_argument_type; +#endif + _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY + _Tp operator()(const _Tp& __x, const _Tp& __y) const + {return __x | __y;} +}; + +#if _LIBCPP_STD_VER > 11 +template <> +struct _LIBCPP_TEMPLATE_VIS bit_or +{ + template + _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY + auto operator()(_T1&& __t, _T2&& __u) const + _NOEXCEPT_(noexcept(_VSTD::forward<_T1>(__t) | _VSTD::forward<_T2>(__u))) + -> decltype (_VSTD::forward<_T1>(__t) | _VSTD::forward<_T2>(__u)) + { return _VSTD::forward<_T1>(__t) | _VSTD::forward<_T2>(__u); } + typedef void is_transparent; +}; +#endif + +_LIBCPP_SUPPRESS_DEPRECATED_PUSH +#if _LIBCPP_STD_VER > 11 +template +#else +template +#endif +struct _LIBCPP_TEMPLATE_VIS bit_xor +#if !defined(_LIBCPP_ABI_NO_BINDER_BASES) + : binary_function<_Tp, _Tp, _Tp> +#endif +{ +_LIBCPP_SUPPRESS_DEPRECATED_POP + typedef _Tp __result_type; // used by valarray +#if _LIBCPP_STD_VER <= 17 || defined(_LIBCPP_ENABLE_CXX20_REMOVED_BINDER_TYPEDEFS) + _LIBCPP_DEPRECATED_IN_CXX17 typedef _Tp result_type; + _LIBCPP_DEPRECATED_IN_CXX17 typedef _Tp first_argument_type; + _LIBCPP_DEPRECATED_IN_CXX17 typedef _Tp second_argument_type; +#endif + _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY + _Tp operator()(const _Tp& __x, const _Tp& __y) const + {return __x ^ __y;} +}; + +#if _LIBCPP_STD_VER > 11 +template <> +struct _LIBCPP_TEMPLATE_VIS bit_xor +{ + template + _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY + auto operator()(_T1&& __t, _T2&& __u) const + _NOEXCEPT_(noexcept(_VSTD::forward<_T1>(__t) ^ _VSTD::forward<_T2>(__u))) + -> decltype (_VSTD::forward<_T1>(__t) ^ _VSTD::forward<_T2>(__u)) + { return _VSTD::forward<_T1>(__t) ^ _VSTD::forward<_T2>(__u); } + typedef void is_transparent; +}; +#endif + +// Comparison operations + +_LIBCPP_SUPPRESS_DEPRECATED_PUSH +#if _LIBCPP_STD_VER > 11 +template +#else +template +#endif +struct _LIBCPP_TEMPLATE_VIS equal_to +#if !defined(_LIBCPP_ABI_NO_BINDER_BASES) + : binary_function<_Tp, _Tp, bool> +#endif +{ +_LIBCPP_SUPPRESS_DEPRECATED_POP + typedef bool __result_type; // used by valarray +#if _LIBCPP_STD_VER <= 17 || defined(_LIBCPP_ENABLE_CXX20_REMOVED_BINDER_TYPEDEFS) + _LIBCPP_DEPRECATED_IN_CXX17 typedef bool result_type; + _LIBCPP_DEPRECATED_IN_CXX17 typedef _Tp first_argument_type; + _LIBCPP_DEPRECATED_IN_CXX17 typedef _Tp second_argument_type; +#endif + _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY + bool operator()(const _Tp& __x, const _Tp& __y) const + {return __x == __y;} +}; + +#if _LIBCPP_STD_VER > 11 +template <> +struct _LIBCPP_TEMPLATE_VIS equal_to +{ + template + _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY + auto operator()(_T1&& __t, _T2&& __u) const + _NOEXCEPT_(noexcept(_VSTD::forward<_T1>(__t) == _VSTD::forward<_T2>(__u))) + -> decltype (_VSTD::forward<_T1>(__t) == _VSTD::forward<_T2>(__u)) + { return _VSTD::forward<_T1>(__t) == _VSTD::forward<_T2>(__u); } + typedef void is_transparent; +}; +#endif + +_LIBCPP_SUPPRESS_DEPRECATED_PUSH +#if _LIBCPP_STD_VER > 11 +template +#else +template +#endif +struct _LIBCPP_TEMPLATE_VIS not_equal_to +#if !defined(_LIBCPP_ABI_NO_BINDER_BASES) + : binary_function<_Tp, _Tp, bool> +#endif +{ +_LIBCPP_SUPPRESS_DEPRECATED_POP + typedef bool __result_type; // used by valarray +#if _LIBCPP_STD_VER <= 17 || defined(_LIBCPP_ENABLE_CXX20_REMOVED_BINDER_TYPEDEFS) + _LIBCPP_DEPRECATED_IN_CXX17 typedef bool result_type; + _LIBCPP_DEPRECATED_IN_CXX17 typedef _Tp first_argument_type; + _LIBCPP_DEPRECATED_IN_CXX17 typedef _Tp second_argument_type; +#endif + _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY + bool operator()(const _Tp& __x, const _Tp& __y) const + {return __x != __y;} +}; + +#if _LIBCPP_STD_VER > 11 +template <> +struct _LIBCPP_TEMPLATE_VIS not_equal_to +{ + template + _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY + auto operator()(_T1&& __t, _T2&& __u) const + _NOEXCEPT_(noexcept(_VSTD::forward<_T1>(__t) != _VSTD::forward<_T2>(__u))) + -> decltype (_VSTD::forward<_T1>(__t) != _VSTD::forward<_T2>(__u)) + { return _VSTD::forward<_T1>(__t) != _VSTD::forward<_T2>(__u); } + typedef void is_transparent; +}; +#endif + +_LIBCPP_SUPPRESS_DEPRECATED_PUSH +#if _LIBCPP_STD_VER > 11 +template +#else +template +#endif +struct _LIBCPP_TEMPLATE_VIS less +#if !defined(_LIBCPP_ABI_NO_BINDER_BASES) + : binary_function<_Tp, _Tp, bool> +#endif +{ +_LIBCPP_SUPPRESS_DEPRECATED_POP + typedef bool __result_type; // used by valarray +#if _LIBCPP_STD_VER <= 17 || defined(_LIBCPP_ENABLE_CXX20_REMOVED_BINDER_TYPEDEFS) + _LIBCPP_DEPRECATED_IN_CXX17 typedef bool result_type; + _LIBCPP_DEPRECATED_IN_CXX17 typedef _Tp first_argument_type; + _LIBCPP_DEPRECATED_IN_CXX17 typedef _Tp second_argument_type; +#endif + _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY + bool operator()(const _Tp& __x, const _Tp& __y) const + {return __x < __y;} +}; + +#if _LIBCPP_STD_VER > 11 +template <> +struct _LIBCPP_TEMPLATE_VIS less +{ + template + _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY + auto operator()(_T1&& __t, _T2&& __u) const + _NOEXCEPT_(noexcept(_VSTD::forward<_T1>(__t) < _VSTD::forward<_T2>(__u))) + -> decltype (_VSTD::forward<_T1>(__t) < _VSTD::forward<_T2>(__u)) + { return _VSTD::forward<_T1>(__t) < _VSTD::forward<_T2>(__u); } + typedef void is_transparent; +}; +#endif + +_LIBCPP_SUPPRESS_DEPRECATED_PUSH +#if _LIBCPP_STD_VER > 11 +template +#else +template +#endif +struct _LIBCPP_TEMPLATE_VIS less_equal +#if !defined(_LIBCPP_ABI_NO_BINDER_BASES) + : binary_function<_Tp, _Tp, bool> +#endif +{ +_LIBCPP_SUPPRESS_DEPRECATED_POP + typedef bool __result_type; // used by valarray +#if _LIBCPP_STD_VER <= 17 || defined(_LIBCPP_ENABLE_CXX20_REMOVED_BINDER_TYPEDEFS) + _LIBCPP_DEPRECATED_IN_CXX17 typedef bool result_type; + _LIBCPP_DEPRECATED_IN_CXX17 typedef _Tp first_argument_type; + _LIBCPP_DEPRECATED_IN_CXX17 typedef _Tp second_argument_type; +#endif + _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY + bool operator()(const _Tp& __x, const _Tp& __y) const + {return __x <= __y;} +}; + +#if _LIBCPP_STD_VER > 11 +template <> +struct _LIBCPP_TEMPLATE_VIS less_equal +{ + template + _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY + auto operator()(_T1&& __t, _T2&& __u) const + _NOEXCEPT_(noexcept(_VSTD::forward<_T1>(__t) <= _VSTD::forward<_T2>(__u))) + -> decltype (_VSTD::forward<_T1>(__t) <= _VSTD::forward<_T2>(__u)) + { return _VSTD::forward<_T1>(__t) <= _VSTD::forward<_T2>(__u); } + typedef void is_transparent; +}; +#endif + +_LIBCPP_SUPPRESS_DEPRECATED_PUSH +#if _LIBCPP_STD_VER > 11 +template +#else +template +#endif +struct _LIBCPP_TEMPLATE_VIS greater_equal +#if !defined(_LIBCPP_ABI_NO_BINDER_BASES) + : binary_function<_Tp, _Tp, bool> +#endif +{ +_LIBCPP_SUPPRESS_DEPRECATED_POP + typedef bool __result_type; // used by valarray +#if _LIBCPP_STD_VER <= 17 || defined(_LIBCPP_ENABLE_CXX20_REMOVED_BINDER_TYPEDEFS) + _LIBCPP_DEPRECATED_IN_CXX17 typedef bool result_type; + _LIBCPP_DEPRECATED_IN_CXX17 typedef _Tp first_argument_type; + _LIBCPP_DEPRECATED_IN_CXX17 typedef _Tp second_argument_type; +#endif + _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY + bool operator()(const _Tp& __x, const _Tp& __y) const + {return __x >= __y;} +}; + +#if _LIBCPP_STD_VER > 11 +template <> +struct _LIBCPP_TEMPLATE_VIS greater_equal +{ + template + _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY + auto operator()(_T1&& __t, _T2&& __u) const + _NOEXCEPT_(noexcept(_VSTD::forward<_T1>(__t) >= _VSTD::forward<_T2>(__u))) + -> decltype (_VSTD::forward<_T1>(__t) >= _VSTD::forward<_T2>(__u)) + { return _VSTD::forward<_T1>(__t) >= _VSTD::forward<_T2>(__u); } + typedef void is_transparent; +}; +#endif + +_LIBCPP_SUPPRESS_DEPRECATED_PUSH +#if _LIBCPP_STD_VER > 11 +template +#else +template +#endif +struct _LIBCPP_TEMPLATE_VIS greater +#if !defined(_LIBCPP_ABI_NO_BINDER_BASES) + : binary_function<_Tp, _Tp, bool> +#endif +{ +_LIBCPP_SUPPRESS_DEPRECATED_POP + typedef bool __result_type; // used by valarray +#if _LIBCPP_STD_VER <= 17 || defined(_LIBCPP_ENABLE_CXX20_REMOVED_BINDER_TYPEDEFS) + _LIBCPP_DEPRECATED_IN_CXX17 typedef bool result_type; + _LIBCPP_DEPRECATED_IN_CXX17 typedef _Tp first_argument_type; + _LIBCPP_DEPRECATED_IN_CXX17 typedef _Tp second_argument_type; +#endif + _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY + bool operator()(const _Tp& __x, const _Tp& __y) const + {return __x > __y;} +}; + +#if _LIBCPP_STD_VER > 11 +template <> +struct _LIBCPP_TEMPLATE_VIS greater +{ + template + _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY + auto operator()(_T1&& __t, _T2&& __u) const + _NOEXCEPT_(noexcept(_VSTD::forward<_T1>(__t) > _VSTD::forward<_T2>(__u))) + -> decltype (_VSTD::forward<_T1>(__t) > _VSTD::forward<_T2>(__u)) + { return _VSTD::forward<_T1>(__t) > _VSTD::forward<_T2>(__u); } + typedef void is_transparent; +}; +#endif + +// Logical operations + +_LIBCPP_SUPPRESS_DEPRECATED_PUSH +#if _LIBCPP_STD_VER > 11 +template +#else +template +#endif +struct _LIBCPP_TEMPLATE_VIS logical_and +#if !defined(_LIBCPP_ABI_NO_BINDER_BASES) + : binary_function<_Tp, _Tp, bool> +#endif +{ +_LIBCPP_SUPPRESS_DEPRECATED_POP + typedef bool __result_type; // used by valarray +#if _LIBCPP_STD_VER <= 17 || defined(_LIBCPP_ENABLE_CXX20_REMOVED_BINDER_TYPEDEFS) + _LIBCPP_DEPRECATED_IN_CXX17 typedef bool result_type; + _LIBCPP_DEPRECATED_IN_CXX17 typedef _Tp first_argument_type; + _LIBCPP_DEPRECATED_IN_CXX17 typedef _Tp second_argument_type; +#endif + _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY + bool operator()(const _Tp& __x, const _Tp& __y) const + {return __x && __y;} +}; + +#if _LIBCPP_STD_VER > 11 +template <> +struct _LIBCPP_TEMPLATE_VIS logical_and +{ + template + _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY + auto operator()(_T1&& __t, _T2&& __u) const + _NOEXCEPT_(noexcept(_VSTD::forward<_T1>(__t) && _VSTD::forward<_T2>(__u))) + -> decltype (_VSTD::forward<_T1>(__t) && _VSTD::forward<_T2>(__u)) + { return _VSTD::forward<_T1>(__t) && _VSTD::forward<_T2>(__u); } + typedef void is_transparent; +}; +#endif + +_LIBCPP_SUPPRESS_DEPRECATED_PUSH +#if _LIBCPP_STD_VER > 11 +template +#else +template +#endif +struct _LIBCPP_TEMPLATE_VIS logical_not +#if !defined(_LIBCPP_ABI_NO_BINDER_BASES) + : unary_function<_Tp, bool> +#endif +{ +_LIBCPP_SUPPRESS_DEPRECATED_POP + typedef bool __result_type; // used by valarray +#if _LIBCPP_STD_VER <= 17 || defined(_LIBCPP_ENABLE_CXX20_REMOVED_BINDER_TYPEDEFS) + _LIBCPP_DEPRECATED_IN_CXX17 typedef bool result_type; + _LIBCPP_DEPRECATED_IN_CXX17 typedef _Tp argument_type; +#endif + _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY + bool operator()(const _Tp& __x) const + {return !__x;} +}; + +#if _LIBCPP_STD_VER > 11 +template <> +struct _LIBCPP_TEMPLATE_VIS logical_not +{ + template + _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY + auto operator()(_Tp&& __x) const + _NOEXCEPT_(noexcept(!_VSTD::forward<_Tp>(__x))) + -> decltype (!_VSTD::forward<_Tp>(__x)) + { return !_VSTD::forward<_Tp>(__x); } + typedef void is_transparent; +}; +#endif + +_LIBCPP_SUPPRESS_DEPRECATED_PUSH +#if _LIBCPP_STD_VER > 11 +template +#else +template +#endif +struct _LIBCPP_TEMPLATE_VIS logical_or +#if !defined(_LIBCPP_ABI_NO_BINDER_BASES) + : binary_function<_Tp, _Tp, bool> +#endif +{ +_LIBCPP_SUPPRESS_DEPRECATED_POP + typedef bool __result_type; // used by valarray +#if _LIBCPP_STD_VER <= 17 || defined(_LIBCPP_ENABLE_CXX20_REMOVED_BINDER_TYPEDEFS) + _LIBCPP_DEPRECATED_IN_CXX17 typedef bool result_type; + _LIBCPP_DEPRECATED_IN_CXX17 typedef _Tp first_argument_type; + _LIBCPP_DEPRECATED_IN_CXX17 typedef _Tp second_argument_type; +#endif + _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY + bool operator()(const _Tp& __x, const _Tp& __y) const + {return __x || __y;} +}; + +#if _LIBCPP_STD_VER > 11 +template <> +struct _LIBCPP_TEMPLATE_VIS logical_or +{ + template + _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY + auto operator()(_T1&& __t, _T2&& __u) const + _NOEXCEPT_(noexcept(_VSTD::forward<_T1>(__t) || _VSTD::forward<_T2>(__u))) + -> decltype (_VSTD::forward<_T1>(__t) || _VSTD::forward<_T2>(__u)) + { return _VSTD::forward<_T1>(__t) || _VSTD::forward<_T2>(__u); } + typedef void is_transparent; +}; +#endif + +_LIBCPP_END_NAMESPACE_STD + +#endif // _LIBCPP___FUNCTIONAL_OPERATIONS_H diff --git a/lib/libcxx/include/__functional/perfect_forward.h b/lib/libcxx/include/__functional/perfect_forward.h new file mode 100644 index 000000000..a5678e159 --- /dev/null +++ b/lib/libcxx/include/__functional/perfect_forward.h @@ -0,0 +1,88 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___FUNCTIONAL_PERFECT_FORWARD_H +#define _LIBCPP___FUNCTIONAL_PERFECT_FORWARD_H + +#include <__config> +#include +#include +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_BEGIN_NAMESPACE_STD + +#if _LIBCPP_STD_VER > 14 + +template::value>::type> +struct __perfect_forward_impl; + +template +struct __perfect_forward_impl<_Op, __tuple_types<_Bound...>, __tuple_indices<_Idxs...>> +{ + tuple<_Bound...> __bound_; + + template + _LIBCPP_INLINE_VISIBILITY constexpr auto operator()(_Args&&... __args) & + noexcept(noexcept(_Op::__call(_VSTD::get<_Idxs>(__bound_)..., _VSTD::forward<_Args>(__args)...))) + -> decltype( _Op::__call(_VSTD::get<_Idxs>(__bound_)..., _VSTD::forward<_Args>(__args)...)) + {return _Op::__call(_VSTD::get<_Idxs>(__bound_)..., _VSTD::forward<_Args>(__args)...);} + + template + _LIBCPP_INLINE_VISIBILITY constexpr auto operator()(_Args&&... __args) const& + noexcept(noexcept(_Op::__call(_VSTD::get<_Idxs>(__bound_)..., _VSTD::forward<_Args>(__args)...))) + -> decltype( _Op::__call(_VSTD::get<_Idxs>(__bound_)..., _VSTD::forward<_Args>(__args)...)) + {return _Op::__call(_VSTD::get<_Idxs>(__bound_)..., _VSTD::forward<_Args>(__args)...);} + + template + _LIBCPP_INLINE_VISIBILITY constexpr auto operator()(_Args&&... __args) && + noexcept(noexcept(_Op::__call(_VSTD::get<_Idxs>(_VSTD::move(__bound_))..., + _VSTD::forward<_Args>(__args)...))) + -> decltype( _Op::__call(_VSTD::get<_Idxs>(_VSTD::move(__bound_))..., + _VSTD::forward<_Args>(__args)...)) + {return _Op::__call(_VSTD::get<_Idxs>(_VSTD::move(__bound_))..., + _VSTD::forward<_Args>(__args)...);} + + template + _LIBCPP_INLINE_VISIBILITY constexpr auto operator()(_Args&&... __args) const&& + noexcept(noexcept(_Op::__call(_VSTD::get<_Idxs>(_VSTD::move(__bound_))..., + _VSTD::forward<_Args>(__args)...))) + -> decltype( _Op::__call(_VSTD::get<_Idxs>(_VSTD::move(__bound_))..., + _VSTD::forward<_Args>(__args)...)) + {return _Op::__call(_VSTD::get<_Idxs>(_VSTD::move(__bound_))..., + _VSTD::forward<_Args>(__args)...);} + + template>::type, + class = _EnableIf>> + constexpr __perfect_forward_impl(__perfect_forward_impl const& __other) + : __bound_(__other.__bound_) {} + + template>::type, + class = _EnableIf>> + constexpr __perfect_forward_impl(__perfect_forward_impl && __other) + : __bound_(_VSTD::move(__other.__bound_)) {} + + template + explicit constexpr __perfect_forward_impl(_BoundArgs&&... __bound) : + __bound_(_VSTD::forward<_BoundArgs>(__bound)...) { } +}; + +template +using __perfect_forward = + __perfect_forward_impl<_Op, __tuple_types...>>; + +#endif // _LIBCPP_STD_VER > 14 + +_LIBCPP_END_NAMESPACE_STD + +#endif // _LIBCPP___FUNCTIONAL_PERFECT_FORWARD_H diff --git a/lib/libcxx/include/__functional/pointer_to_binary_function.h b/lib/libcxx/include/__functional/pointer_to_binary_function.h new file mode 100644 index 000000000..d4a6c1674 --- /dev/null +++ b/lib/libcxx/include/__functional/pointer_to_binary_function.h @@ -0,0 +1,46 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___FUNCTIONAL_POINTER_TO_BINARY_FUNCTION_H +#define _LIBCPP___FUNCTIONAL_POINTER_TO_BINARY_FUNCTION_H + +#include <__config> +#include <__functional/binary_function.h> + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_BEGIN_NAMESPACE_STD + +#if _LIBCPP_STD_VER <= 14 || defined(_LIBCPP_ENABLE_CXX17_REMOVED_BINDERS) + +template +class _LIBCPP_TEMPLATE_VIS _LIBCPP_DEPRECATED_IN_CXX11 pointer_to_binary_function + : public binary_function<_Arg1, _Arg2, _Result> +{ + _Result (*__f_)(_Arg1, _Arg2); +public: + _LIBCPP_INLINE_VISIBILITY explicit pointer_to_binary_function(_Result (*__f)(_Arg1, _Arg2)) + : __f_(__f) {} + _LIBCPP_INLINE_VISIBILITY _Result operator()(_Arg1 __x, _Arg2 __y) const + {return __f_(__x, __y);} +}; + +template +_LIBCPP_DEPRECATED_IN_CXX11 inline _LIBCPP_INLINE_VISIBILITY +pointer_to_binary_function<_Arg1,_Arg2,_Result> +ptr_fun(_Result (*__f)(_Arg1,_Arg2)) + {return pointer_to_binary_function<_Arg1,_Arg2,_Result>(__f);} + +#endif + +_LIBCPP_END_NAMESPACE_STD + +#endif // _LIBCPP___FUNCTIONAL_POINTER_TO_BINARY_FUNCTION_H diff --git a/lib/libcxx/include/__functional/pointer_to_unary_function.h b/lib/libcxx/include/__functional/pointer_to_unary_function.h new file mode 100644 index 000000000..0ac4561cc --- /dev/null +++ b/lib/libcxx/include/__functional/pointer_to_unary_function.h @@ -0,0 +1,46 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___FUNCTIONAL_POINTER_TO_UNARY_FUNCTION_H +#define _LIBCPP___FUNCTIONAL_POINTER_TO_UNARY_FUNCTION_H + +#include <__config> +#include <__functional/unary_function.h> + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_BEGIN_NAMESPACE_STD + +#if _LIBCPP_STD_VER <= 14 || defined(_LIBCPP_ENABLE_CXX17_REMOVED_BINDERS) + +template +class _LIBCPP_TEMPLATE_VIS _LIBCPP_DEPRECATED_IN_CXX11 pointer_to_unary_function + : public unary_function<_Arg, _Result> +{ + _Result (*__f_)(_Arg); +public: + _LIBCPP_INLINE_VISIBILITY explicit pointer_to_unary_function(_Result (*__f)(_Arg)) + : __f_(__f) {} + _LIBCPP_INLINE_VISIBILITY _Result operator()(_Arg __x) const + {return __f_(__x);} +}; + +template +_LIBCPP_DEPRECATED_IN_CXX11 inline _LIBCPP_INLINE_VISIBILITY +pointer_to_unary_function<_Arg,_Result> +ptr_fun(_Result (*__f)(_Arg)) + {return pointer_to_unary_function<_Arg,_Result>(__f);} + +#endif // _LIBCPP_STD_VER <= 14 || defined(_LIBCPP_ENABLE_CXX17_REMOVED_BINDERS) + +_LIBCPP_END_NAMESPACE_STD + +#endif // _LIBCPP___FUNCTIONAL_POINTER_TO_UNARY_FUNCTION_H diff --git a/lib/libcxx/include/__functional/ranges_operations.h b/lib/libcxx/include/__functional/ranges_operations.h new file mode 100644 index 000000000..777c53525 --- /dev/null +++ b/lib/libcxx/include/__functional/ranges_operations.h @@ -0,0 +1,97 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___FUNCTIONAL_RANGES_OPERATIONS_H +#define _LIBCPP___FUNCTIONAL_RANGES_OPERATIONS_H + +#include <__config> +#include +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_BEGIN_NAMESPACE_STD + +#if !defined(_LIBCPP_HAS_NO_RANGES) +namespace ranges { + +struct equal_to { + template + requires equality_comparable_with<_Tp, _Up> + [[nodiscard]] constexpr bool operator()(_Tp &&__t, _Up &&__u) const + noexcept(noexcept(bool(_VSTD::forward<_Tp>(__t) == _VSTD::forward<_Up>(__u)))) { + return _VSTD::forward<_Tp>(__t) == _VSTD::forward<_Up>(__u); + } + + using is_transparent = void; +}; + +struct not_equal_to { + template + requires equality_comparable_with<_Tp, _Up> + [[nodiscard]] constexpr bool operator()(_Tp &&__t, _Up &&__u) const + noexcept(noexcept(bool(!(_VSTD::forward<_Tp>(__t) == _VSTD::forward<_Up>(__u))))) { + return !(_VSTD::forward<_Tp>(__t) == _VSTD::forward<_Up>(__u)); + } + + using is_transparent = void; +}; + +struct less { + template + requires totally_ordered_with<_Tp, _Up> + [[nodiscard]] constexpr bool operator()(_Tp &&__t, _Up &&__u) const + noexcept(noexcept(bool(_VSTD::forward<_Tp>(__t) < _VSTD::forward<_Up>(__u)))) { + return _VSTD::forward<_Tp>(__t) < _VSTD::forward<_Up>(__u); + } + + using is_transparent = void; +}; + +struct less_equal { + template + requires totally_ordered_with<_Tp, _Up> + [[nodiscard]] constexpr bool operator()(_Tp &&__t, _Up &&__u) const + noexcept(noexcept(bool(!(_VSTD::forward<_Up>(__u) < _VSTD::forward<_Tp>(__t))))) { + return !(_VSTD::forward<_Up>(__u) < _VSTD::forward<_Tp>(__t)); + } + + using is_transparent = void; +}; + +struct greater { + template + requires totally_ordered_with<_Tp, _Up> + [[nodiscard]] constexpr bool operator()(_Tp &&__t, _Up &&__u) const + noexcept(noexcept(bool(_VSTD::forward<_Up>(__u) < _VSTD::forward<_Tp>(__t)))) { + return _VSTD::forward<_Up>(__u) < _VSTD::forward<_Tp>(__t); + } + + using is_transparent = void; +}; + +struct greater_equal { + template + requires totally_ordered_with<_Tp, _Up> + [[nodiscard]] constexpr bool operator()(_Tp &&__t, _Up &&__u) const + noexcept(noexcept(bool(!(_VSTD::forward<_Tp>(__t) < _VSTD::forward<_Up>(__u))))) { + return !(_VSTD::forward<_Tp>(__t) < _VSTD::forward<_Up>(__u)); + } + + using is_transparent = void; +}; + +} // namespace ranges +#endif // !defined(_LIBCPP_HAS_NO_RANGES) + +_LIBCPP_END_NAMESPACE_STD + +#endif // _LIBCPP___FUNCTIONAL_RANGES_OPERATIONS_H diff --git a/lib/libcxx/include/__functional/reference_wrapper.h b/lib/libcxx/include/__functional/reference_wrapper.h new file mode 100644 index 000000000..09f4a6494 --- /dev/null +++ b/lib/libcxx/include/__functional/reference_wrapper.h @@ -0,0 +1,223 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___FUNCTIONAL_REFERENCE_WRAPPER_H +#define _LIBCPP___FUNCTIONAL_REFERENCE_WRAPPER_H + +#include <__config> +#include <__functional/weak_result_type.h> +#include <__memory/addressof.h> +#include <__utility/forward.h> +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +class _LIBCPP_TEMPLATE_VIS reference_wrapper +#if _LIBCPP_STD_VER <= 17 || !defined(_LIBCPP_ABI_NO_BINDER_BASES) + : public __weak_result_type<_Tp> +#endif +{ +public: + // types + typedef _Tp type; +private: + type* __f_; + +#ifndef _LIBCPP_CXX03_LANG + static void __fun(_Tp&) _NOEXCEPT; + static void __fun(_Tp&&) = delete; +#endif + +public: + // construct/copy/destroy +#ifdef _LIBCPP_CXX03_LANG + _LIBCPP_INLINE_VISIBILITY + reference_wrapper(type& __f) _NOEXCEPT + : __f_(_VSTD::addressof(__f)) {} +#else + template ::value, decltype(__fun(declval<_Up>())) >> + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 + reference_wrapper(_Up&& __u) _NOEXCEPT_(noexcept(__fun(declval<_Up>()))) { + type& __f = static_cast<_Up&&>(__u); + __f_ = _VSTD::addressof(__f); + } +#endif + + // access + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 + operator type&() const _NOEXCEPT {return *__f_;} + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 + type& get() const _NOEXCEPT {return *__f_;} + +#ifndef _LIBCPP_CXX03_LANG + // invoke + template + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 + typename __invoke_of::type + operator() (_ArgTypes&&... __args) const { + return _VSTD::__invoke(get(), _VSTD::forward<_ArgTypes>(__args)...); + } +#else + + _LIBCPP_INLINE_VISIBILITY + typename __invoke_return::type + operator() () const { + return _VSTD::__invoke(get()); + } + + template + _LIBCPP_INLINE_VISIBILITY + typename __invoke_return0::type + operator() (_A0& __a0) const { + return _VSTD::__invoke(get(), __a0); + } + + template + _LIBCPP_INLINE_VISIBILITY + typename __invoke_return0::type + operator() (_A0 const& __a0) const { + return _VSTD::__invoke(get(), __a0); + } + + template + _LIBCPP_INLINE_VISIBILITY + typename __invoke_return1::type + operator() (_A0& __a0, _A1& __a1) const { + return _VSTD::__invoke(get(), __a0, __a1); + } + + template + _LIBCPP_INLINE_VISIBILITY + typename __invoke_return1::type + operator() (_A0 const& __a0, _A1& __a1) const { + return _VSTD::__invoke(get(), __a0, __a1); + } + + template + _LIBCPP_INLINE_VISIBILITY + typename __invoke_return1::type + operator() (_A0& __a0, _A1 const& __a1) const { + return _VSTD::__invoke(get(), __a0, __a1); + } + + template + _LIBCPP_INLINE_VISIBILITY + typename __invoke_return1::type + operator() (_A0 const& __a0, _A1 const& __a1) const { + return _VSTD::__invoke(get(), __a0, __a1); + } + + template + _LIBCPP_INLINE_VISIBILITY + typename __invoke_return2::type + operator() (_A0& __a0, _A1& __a1, _A2& __a2) const { + return _VSTD::__invoke(get(), __a0, __a1, __a2); + } + + template + _LIBCPP_INLINE_VISIBILITY + typename __invoke_return2::type + operator() (_A0 const& __a0, _A1& __a1, _A2& __a2) const { + return _VSTD::__invoke(get(), __a0, __a1, __a2); + } + + template + _LIBCPP_INLINE_VISIBILITY + typename __invoke_return2::type + operator() (_A0& __a0, _A1 const& __a1, _A2& __a2) const { + return _VSTD::__invoke(get(), __a0, __a1, __a2); + } + + template + _LIBCPP_INLINE_VISIBILITY + typename __invoke_return2::type + operator() (_A0& __a0, _A1& __a1, _A2 const& __a2) const { + return _VSTD::__invoke(get(), __a0, __a1, __a2); + } + + template + _LIBCPP_INLINE_VISIBILITY + typename __invoke_return2::type + operator() (_A0 const& __a0, _A1 const& __a1, _A2& __a2) const { + return _VSTD::__invoke(get(), __a0, __a1, __a2); + } + + template + _LIBCPP_INLINE_VISIBILITY + typename __invoke_return2::type + operator() (_A0 const& __a0, _A1& __a1, _A2 const& __a2) const { + return _VSTD::__invoke(get(), __a0, __a1, __a2); + } + + template + _LIBCPP_INLINE_VISIBILITY + typename __invoke_return2::type + operator() (_A0& __a0, _A1 const& __a1, _A2 const& __a2) const { + return _VSTD::__invoke(get(), __a0, __a1, __a2); + } + + template + _LIBCPP_INLINE_VISIBILITY + typename __invoke_return2::type + operator() (_A0 const& __a0, _A1 const& __a1, _A2 const& __a2) const { + return _VSTD::__invoke(get(), __a0, __a1, __a2); + } +#endif // _LIBCPP_CXX03_LANG +}; + +#ifndef _LIBCPP_HAS_NO_DEDUCTION_GUIDES +template +reference_wrapper(_Tp&) -> reference_wrapper<_Tp>; +#endif + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +reference_wrapper<_Tp> +ref(_Tp& __t) _NOEXCEPT +{ + return reference_wrapper<_Tp>(__t); +} + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +reference_wrapper<_Tp> +ref(reference_wrapper<_Tp> __t) _NOEXCEPT +{ + return _VSTD::ref(__t.get()); +} + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +reference_wrapper +cref(const _Tp& __t) _NOEXCEPT +{ + return reference_wrapper(__t); +} + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +reference_wrapper +cref(reference_wrapper<_Tp> __t) _NOEXCEPT +{ + return _VSTD::cref(__t.get()); +} + +#ifndef _LIBCPP_CXX03_LANG +template void ref(const _Tp&&) = delete; +template void cref(const _Tp&&) = delete; +#endif + +_LIBCPP_END_NAMESPACE_STD + +#endif // _LIBCPP___FUNCTIONAL_REFERENCE_WRAPPER_H diff --git a/lib/libcxx/include/__functional/unary_function.h b/lib/libcxx/include/__functional/unary_function.h new file mode 100644 index 000000000..8084ef4b0 --- /dev/null +++ b/lib/libcxx/include/__functional/unary_function.h @@ -0,0 +1,34 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___FUNCTIONAL_UNARY_FUNCTION_H +#define _LIBCPP___FUNCTIONAL_UNARY_FUNCTION_H + +#include <__config> + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +struct _LIBCPP_TEMPLATE_VIS unary_function +{ + typedef _Arg argument_type; + typedef _Result result_type; +}; + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___FUNCTIONAL_UNARY_FUNCTION_H diff --git a/lib/libcxx/include/__functional/unary_negate.h b/lib/libcxx/include/__functional/unary_negate.h new file mode 100644 index 000000000..71257cf40 --- /dev/null +++ b/lib/libcxx/include/__functional/unary_negate.h @@ -0,0 +1,47 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___FUNCTIONAL_UNARY_NEGATE_H +#define _LIBCPP___FUNCTIONAL_UNARY_NEGATE_H + +#include <__config> +#include <__functional/unary_function.h> + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_BEGIN_NAMESPACE_STD + +#if _LIBCPP_STD_VER <= 17 || defined(_LIBCPP_ENABLE_CXX20_REMOVED_NEGATORS) + +template +class _LIBCPP_TEMPLATE_VIS _LIBCPP_DEPRECATED_IN_CXX17 unary_negate + : public unary_function +{ + _Predicate __pred_; +public: + _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY + explicit unary_negate(const _Predicate& __pred) + : __pred_(__pred) {} + _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY + bool operator()(const typename _Predicate::argument_type& __x) const + {return !__pred_(__x);} +}; + +template +_LIBCPP_DEPRECATED_IN_CXX17 inline _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY +unary_negate<_Predicate> +not1(const _Predicate& __pred) {return unary_negate<_Predicate>(__pred);} + +#endif // _LIBCPP_STD_VER <= 17 || defined(_LIBCPP_ENABLE_CXX20_REMOVED_NEGATORS) + +_LIBCPP_END_NAMESPACE_STD + +#endif // _LIBCPP___FUNCTIONAL_UNARY_NEGATE_H diff --git a/lib/libcxx/include/__functional/unwrap_ref.h b/lib/libcxx/include/__functional/unwrap_ref.h new file mode 100644 index 000000000..4d091ec35 --- /dev/null +++ b/lib/libcxx/include/__functional/unwrap_ref.h @@ -0,0 +1,62 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___FUNCTIONAL_UNWRAP_REF_H +#define _LIBCPP___FUNCTIONAL_UNWRAP_REF_H + +#include <__config> + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +struct __unwrap_reference { typedef _LIBCPP_NODEBUG_TYPE _Tp type; }; + +template +class reference_wrapper; + +template +struct __unwrap_reference > { typedef _LIBCPP_NODEBUG_TYPE _Tp& type; }; + +template +struct decay; + +#if _LIBCPP_STD_VER > 17 +template +struct unwrap_reference : __unwrap_reference<_Tp> { }; + +template +using unwrap_reference_t = typename unwrap_reference<_Tp>::type; + +template +struct unwrap_ref_decay : unwrap_reference::type> { }; + +template +using unwrap_ref_decay_t = typename unwrap_ref_decay<_Tp>::type; +#endif // > C++17 + +template +struct __unwrap_ref_decay +#if _LIBCPP_STD_VER > 17 + : unwrap_ref_decay<_Tp> +#else + : __unwrap_reference::type> +#endif +{ }; + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___FUNCTIONAL_UNWRAP_REF_H diff --git a/lib/libcxx/include/__functional/weak_result_type.h b/lib/libcxx/include/__functional/weak_result_type.h new file mode 100644 index 000000000..2ee85acf1 --- /dev/null +++ b/lib/libcxx/include/__functional/weak_result_type.h @@ -0,0 +1,481 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___FUNCTIONAL_WEAK_RESULT_TYPE_H +#define _LIBCPP___FUNCTIONAL_WEAK_RESULT_TYPE_H + +#include <__config> +#include <__functional/binary_function.h> +#include <__functional/unary_function.h> +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +struct __has_result_type +{ +private: + struct __two {char __lx; char __lxx;}; + template static __two __test(...); + template static char __test(typename _Up::result_type* = 0); +public: + static const bool value = sizeof(__test<_Tp>(0)) == 1; +}; + +// __weak_result_type + +template +struct __derives_from_unary_function +{ +private: + struct __two {char __lx; char __lxx;}; + static __two __test(...); + template + static unary_function<_Ap, _Rp> + __test(const volatile unary_function<_Ap, _Rp>*); +public: + static const bool value = !is_same::value; + typedef decltype(__test((_Tp*)0)) type; +}; + +template +struct __derives_from_binary_function +{ +private: + struct __two {char __lx; char __lxx;}; + static __two __test(...); + template + static binary_function<_A1, _A2, _Rp> + __test(const volatile binary_function<_A1, _A2, _Rp>*); +public: + static const bool value = !is_same::value; + typedef decltype(__test((_Tp*)0)) type; +}; + +template ::value> +struct __maybe_derive_from_unary_function // bool is true + : public __derives_from_unary_function<_Tp>::type +{ +}; + +template +struct __maybe_derive_from_unary_function<_Tp, false> +{ +}; + +template ::value> +struct __maybe_derive_from_binary_function // bool is true + : public __derives_from_binary_function<_Tp>::type +{ +}; + +template +struct __maybe_derive_from_binary_function<_Tp, false> +{ +}; + +template ::value> +struct __weak_result_type_imp // bool is true + : public __maybe_derive_from_unary_function<_Tp>, + public __maybe_derive_from_binary_function<_Tp> +{ + typedef _LIBCPP_NODEBUG_TYPE typename _Tp::result_type result_type; +}; + +template +struct __weak_result_type_imp<_Tp, false> + : public __maybe_derive_from_unary_function<_Tp>, + public __maybe_derive_from_binary_function<_Tp> +{ +}; + +template +struct __weak_result_type + : public __weak_result_type_imp<_Tp> +{ +}; + +// 0 argument case + +template +struct __weak_result_type<_Rp ()> +{ + typedef _LIBCPP_NODEBUG_TYPE _Rp result_type; +}; + +template +struct __weak_result_type<_Rp (&)()> +{ + typedef _LIBCPP_NODEBUG_TYPE _Rp result_type; +}; + +template +struct __weak_result_type<_Rp (*)()> +{ + typedef _LIBCPP_NODEBUG_TYPE _Rp result_type; +}; + +// 1 argument case + +template +struct __weak_result_type<_Rp (_A1)> + : public unary_function<_A1, _Rp> +{ +}; + +template +struct __weak_result_type<_Rp (&)(_A1)> + : public unary_function<_A1, _Rp> +{ +}; + +template +struct __weak_result_type<_Rp (*)(_A1)> + : public unary_function<_A1, _Rp> +{ +}; + +template +struct __weak_result_type<_Rp (_Cp::*)()> + : public unary_function<_Cp*, _Rp> +{ +}; + +template +struct __weak_result_type<_Rp (_Cp::*)() const> + : public unary_function +{ +}; + +template +struct __weak_result_type<_Rp (_Cp::*)() volatile> + : public unary_function +{ +}; + +template +struct __weak_result_type<_Rp (_Cp::*)() const volatile> + : public unary_function +{ +}; + +// 2 argument case + +template +struct __weak_result_type<_Rp (_A1, _A2)> + : public binary_function<_A1, _A2, _Rp> +{ +}; + +template +struct __weak_result_type<_Rp (*)(_A1, _A2)> + : public binary_function<_A1, _A2, _Rp> +{ +}; + +template +struct __weak_result_type<_Rp (&)(_A1, _A2)> + : public binary_function<_A1, _A2, _Rp> +{ +}; + +template +struct __weak_result_type<_Rp (_Cp::*)(_A1)> + : public binary_function<_Cp*, _A1, _Rp> +{ +}; + +template +struct __weak_result_type<_Rp (_Cp::*)(_A1) const> + : public binary_function +{ +}; + +template +struct __weak_result_type<_Rp (_Cp::*)(_A1) volatile> + : public binary_function +{ +}; + +template +struct __weak_result_type<_Rp (_Cp::*)(_A1) const volatile> + : public binary_function +{ +}; + + +#ifndef _LIBCPP_CXX03_LANG +// 3 or more arguments + +template +struct __weak_result_type<_Rp (_A1, _A2, _A3, _A4...)> +{ + typedef _Rp result_type; +}; + +template +struct __weak_result_type<_Rp (&)(_A1, _A2, _A3, _A4...)> +{ + typedef _Rp result_type; +}; + +template +struct __weak_result_type<_Rp (*)(_A1, _A2, _A3, _A4...)> +{ + typedef _Rp result_type; +}; + +template +struct __weak_result_type<_Rp (_Cp::*)(_A1, _A2, _A3...)> +{ + typedef _Rp result_type; +}; + +template +struct __weak_result_type<_Rp (_Cp::*)(_A1, _A2, _A3...) const> +{ + typedef _Rp result_type; +}; + +template +struct __weak_result_type<_Rp (_Cp::*)(_A1, _A2, _A3...) volatile> +{ + typedef _Rp result_type; +}; + +template +struct __weak_result_type<_Rp (_Cp::*)(_A1, _A2, _A3...) const volatile> +{ + typedef _Rp result_type; +}; + +template +struct __invoke_return +{ + typedef decltype(_VSTD::__invoke(declval<_Tp>(), declval<_Args>()...)) type; +}; + +#else // defined(_LIBCPP_CXX03_LANG) + +template +struct __enable_invoke_imp; + +template +struct __enable_invoke_imp<_Ret, _T1, true, true> { + typedef _Ret _Bullet1; + typedef _Bullet1 type; +}; + +template +struct __enable_invoke_imp<_Ret, _T1, true, false> { + typedef _Ret _Bullet2; + typedef _Bullet2 type; +}; + +template +struct __enable_invoke_imp<_Ret, _T1, false, true> { + typedef typename add_lvalue_reference< + typename __apply_cv<_T1, _Ret>::type + >::type _Bullet3; + typedef _Bullet3 type; +}; + +template +struct __enable_invoke_imp<_Ret, _T1, false, false> { + typedef typename add_lvalue_reference< + typename __apply_cv()), _Ret>::type + >::type _Bullet4; + typedef _Bullet4 type; +}; + +template +struct __enable_invoke_imp<_Ret, _T1*, false, false> { + typedef typename add_lvalue_reference< + typename __apply_cv<_T1, _Ret>::type + >::type _Bullet4; + typedef _Bullet4 type; +}; + +template , + class _Ret = typename _Traits::_ReturnType, + class _Class = typename _Traits::_ClassType> +struct __enable_invoke : __enable_invoke_imp< + _Ret, _T1, + is_member_function_pointer<_Fn>::value, + is_base_of<_Class, typename remove_reference<_T1>::type>::value> +{ +}; + +__nat __invoke(__any, ...); + +// first bullet + +template +inline _LIBCPP_INLINE_VISIBILITY +typename __enable_invoke<_Fn, _T1>::_Bullet1 +__invoke(_Fn __f, _T1& __t1) { + return (__t1.*__f)(); +} + +template +inline _LIBCPP_INLINE_VISIBILITY +typename __enable_invoke<_Fn, _T1>::_Bullet1 +__invoke(_Fn __f, _T1& __t1, _A0& __a0) { + return (__t1.*__f)(__a0); +} + +template +inline _LIBCPP_INLINE_VISIBILITY +typename __enable_invoke<_Fn, _T1>::_Bullet1 +__invoke(_Fn __f, _T1& __t1, _A0& __a0, _A1& __a1) { + return (__t1.*__f)(__a0, __a1); +} + +template +inline _LIBCPP_INLINE_VISIBILITY +typename __enable_invoke<_Fn, _T1>::_Bullet1 +__invoke(_Fn __f, _T1& __t1, _A0& __a0, _A1& __a1, _A2& __a2) { + return (__t1.*__f)(__a0, __a1, __a2); +} + +template +inline _LIBCPP_INLINE_VISIBILITY +typename __enable_invoke<_Fn, _T1>::_Bullet2 +__invoke(_Fn __f, _T1& __t1) { + return ((*__t1).*__f)(); +} + +template +inline _LIBCPP_INLINE_VISIBILITY +typename __enable_invoke<_Fn, _T1>::_Bullet2 +__invoke(_Fn __f, _T1& __t1, _A0& __a0) { + return ((*__t1).*__f)(__a0); +} + +template +inline _LIBCPP_INLINE_VISIBILITY +typename __enable_invoke<_Fn, _T1>::_Bullet2 +__invoke(_Fn __f, _T1& __t1, _A0& __a0, _A1& __a1) { + return ((*__t1).*__f)(__a0, __a1); +} + +template +inline _LIBCPP_INLINE_VISIBILITY +typename __enable_invoke<_Fn, _T1>::_Bullet2 +__invoke(_Fn __f, _T1& __t1, _A0& __a0, _A1& __a1, _A2& __a2) { + return ((*__t1).*__f)(__a0, __a1, __a2); +} + +template +inline _LIBCPP_INLINE_VISIBILITY +typename __enable_invoke<_Fn, _T1>::_Bullet3 +__invoke(_Fn __f, _T1& __t1) { + return __t1.*__f; +} + +template +inline _LIBCPP_INLINE_VISIBILITY +typename __enable_invoke<_Fn, _T1>::_Bullet4 +__invoke(_Fn __f, _T1& __t1) { + return (*__t1).*__f; +} + +// fifth bullet + +template +inline _LIBCPP_INLINE_VISIBILITY +decltype(declval<_Fp&>()()) +__invoke(_Fp& __f) +{ + return __f(); +} + +template +inline _LIBCPP_INLINE_VISIBILITY +decltype(declval<_Fp&>()(declval<_A0&>())) +__invoke(_Fp& __f, _A0& __a0) +{ + return __f(__a0); +} + +template +inline _LIBCPP_INLINE_VISIBILITY +decltype(declval<_Fp&>()(declval<_A0&>(), declval<_A1&>())) +__invoke(_Fp& __f, _A0& __a0, _A1& __a1) +{ + return __f(__a0, __a1); +} + +template +inline _LIBCPP_INLINE_VISIBILITY +decltype(declval<_Fp&>()(declval<_A0&>(), declval<_A1&>(), declval<_A2&>())) +__invoke(_Fp& __f, _A0& __a0, _A1& __a1, _A2& __a2) +{ + return __f(__a0, __a1, __a2); +} + +template >::value> +struct __invoke_return +{ + typedef typename __weak_result_type<_Fp>::result_type type; +}; + +template +struct __invoke_return<_Fp, false> +{ + typedef decltype(_VSTD::__invoke(declval<_Fp&>())) type; +}; + +template +struct __invoke_return0 +{ + typedef decltype(_VSTD::__invoke(declval<_Tp&>(), declval<_A0&>())) type; +}; + +template +struct __invoke_return0<_Rp _Tp::*, _A0> +{ + typedef typename __enable_invoke<_Rp _Tp::*, _A0>::type type; +}; + +template +struct __invoke_return1 +{ + typedef decltype(_VSTD::__invoke(declval<_Tp&>(), declval<_A0&>(), + declval<_A1&>())) type; +}; + +template +struct __invoke_return1<_Rp _Class::*, _A0, _A1> { + typedef typename __enable_invoke<_Rp _Class::*, _A0>::type type; +}; + +template +struct __invoke_return2 +{ + typedef decltype(_VSTD::__invoke(declval<_Tp&>(), declval<_A0&>(), + declval<_A1&>(), + declval<_A2&>())) type; +}; + +template +struct __invoke_return2<_Ret _Class::*, _A0, _A1, _A2> { + typedef typename __enable_invoke<_Ret _Class::*, _A0>::type type; +}; + +#endif // !defined(_LIBCPP_CXX03_LANG) + +_LIBCPP_END_NAMESPACE_STD + +#endif // _LIBCPP___FUNCTIONAL_WEAK_RESULT_TYPE_H diff --git a/lib/libcxx/include/__functional_base b/lib/libcxx/include/__functional_base index 1c02e960d..ccc3f3a58 100644 --- a/lib/libcxx/include/__functional_base +++ b/lib/libcxx/include/__functional_base @@ -11,645 +11,22 @@ #define _LIBCPP_FUNCTIONAL_BASE #include <__config> -#include -#include +#include <__functional/binary_function.h> +#include <__functional/invoke.h> +#include <__functional/operations.h> +#include <__functional/reference_wrapper.h> +#include <__functional/unary_function.h> +#include <__functional/weak_result_type.h> +#include <__memory/allocator_arg_t.h> +#include <__memory/uses_allocator.h> #include #include +#include +#include #include #if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) #pragma GCC system_header #endif -_LIBCPP_BEGIN_NAMESPACE_STD - -template -struct _LIBCPP_TEMPLATE_VIS binary_function -{ - typedef _Arg1 first_argument_type; - typedef _Arg2 second_argument_type; - typedef _Result result_type; -}; - -template -struct __has_result_type -{ -private: - struct __two {char __lx; char __lxx;}; - template static __two __test(...); - template static char __test(typename _Up::result_type* = 0); -public: - static const bool value = sizeof(__test<_Tp>(0)) == 1; -}; - -#if _LIBCPP_STD_VER > 11 -template -#else -template -#endif -struct _LIBCPP_TEMPLATE_VIS less : binary_function<_Tp, _Tp, bool> -{ - _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY - bool operator()(const _Tp& __x, const _Tp& __y) const - {return __x < __y;} -}; - -#if _LIBCPP_STD_VER > 11 -template <> -struct _LIBCPP_TEMPLATE_VIS less -{ - template - _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY - auto operator()(_T1&& __t, _T2&& __u) const - _NOEXCEPT_(noexcept(_VSTD::forward<_T1>(__t) < _VSTD::forward<_T2>(__u))) - -> decltype (_VSTD::forward<_T1>(__t) < _VSTD::forward<_T2>(__u)) - { return _VSTD::forward<_T1>(__t) < _VSTD::forward<_T2>(__u); } - typedef void is_transparent; -}; -#endif - -// __weak_result_type - -template -struct __derives_from_unary_function -{ -private: - struct __two {char __lx; char __lxx;}; - static __two __test(...); - template - static unary_function<_Ap, _Rp> - __test(const volatile unary_function<_Ap, _Rp>*); -public: - static const bool value = !is_same::value; - typedef decltype(__test((_Tp*)0)) type; -}; - -template -struct __derives_from_binary_function -{ -private: - struct __two {char __lx; char __lxx;}; - static __two __test(...); - template - static binary_function<_A1, _A2, _Rp> - __test(const volatile binary_function<_A1, _A2, _Rp>*); -public: - static const bool value = !is_same::value; - typedef decltype(__test((_Tp*)0)) type; -}; - -template ::value> -struct __maybe_derive_from_unary_function // bool is true - : public __derives_from_unary_function<_Tp>::type -{ -}; - -template -struct __maybe_derive_from_unary_function<_Tp, false> -{ -}; - -template ::value> -struct __maybe_derive_from_binary_function // bool is true - : public __derives_from_binary_function<_Tp>::type -{ -}; - -template -struct __maybe_derive_from_binary_function<_Tp, false> -{ -}; - -template ::value> -struct __weak_result_type_imp // bool is true - : public __maybe_derive_from_unary_function<_Tp>, - public __maybe_derive_from_binary_function<_Tp> -{ - typedef _LIBCPP_NODEBUG_TYPE typename _Tp::result_type result_type; -}; - -template -struct __weak_result_type_imp<_Tp, false> - : public __maybe_derive_from_unary_function<_Tp>, - public __maybe_derive_from_binary_function<_Tp> -{ -}; - -template -struct __weak_result_type - : public __weak_result_type_imp<_Tp> -{ -}; - -// 0 argument case - -template -struct __weak_result_type<_Rp ()> -{ - typedef _LIBCPP_NODEBUG_TYPE _Rp result_type; -}; - -template -struct __weak_result_type<_Rp (&)()> -{ - typedef _LIBCPP_NODEBUG_TYPE _Rp result_type; -}; - -template -struct __weak_result_type<_Rp (*)()> -{ - typedef _LIBCPP_NODEBUG_TYPE _Rp result_type; -}; - -// 1 argument case - -template -struct __weak_result_type<_Rp (_A1)> - : public unary_function<_A1, _Rp> -{ -}; - -template -struct __weak_result_type<_Rp (&)(_A1)> - : public unary_function<_A1, _Rp> -{ -}; - -template -struct __weak_result_type<_Rp (*)(_A1)> - : public unary_function<_A1, _Rp> -{ -}; - -template -struct __weak_result_type<_Rp (_Cp::*)()> - : public unary_function<_Cp*, _Rp> -{ -}; - -template -struct __weak_result_type<_Rp (_Cp::*)() const> - : public unary_function -{ -}; - -template -struct __weak_result_type<_Rp (_Cp::*)() volatile> - : public unary_function -{ -}; - -template -struct __weak_result_type<_Rp (_Cp::*)() const volatile> - : public unary_function -{ -}; - -// 2 argument case - -template -struct __weak_result_type<_Rp (_A1, _A2)> - : public binary_function<_A1, _A2, _Rp> -{ -}; - -template -struct __weak_result_type<_Rp (*)(_A1, _A2)> - : public binary_function<_A1, _A2, _Rp> -{ -}; - -template -struct __weak_result_type<_Rp (&)(_A1, _A2)> - : public binary_function<_A1, _A2, _Rp> -{ -}; - -template -struct __weak_result_type<_Rp (_Cp::*)(_A1)> - : public binary_function<_Cp*, _A1, _Rp> -{ -}; - -template -struct __weak_result_type<_Rp (_Cp::*)(_A1) const> - : public binary_function -{ -}; - -template -struct __weak_result_type<_Rp (_Cp::*)(_A1) volatile> - : public binary_function -{ -}; - -template -struct __weak_result_type<_Rp (_Cp::*)(_A1) const volatile> - : public binary_function -{ -}; - - -#ifndef _LIBCPP_CXX03_LANG -// 3 or more arguments - -template -struct __weak_result_type<_Rp (_A1, _A2, _A3, _A4...)> -{ - typedef _Rp result_type; -}; - -template -struct __weak_result_type<_Rp (&)(_A1, _A2, _A3, _A4...)> -{ - typedef _Rp result_type; -}; - -template -struct __weak_result_type<_Rp (*)(_A1, _A2, _A3, _A4...)> -{ - typedef _Rp result_type; -}; - -template -struct __weak_result_type<_Rp (_Cp::*)(_A1, _A2, _A3...)> -{ - typedef _Rp result_type; -}; - -template -struct __weak_result_type<_Rp (_Cp::*)(_A1, _A2, _A3...) const> -{ - typedef _Rp result_type; -}; - -template -struct __weak_result_type<_Rp (_Cp::*)(_A1, _A2, _A3...) volatile> -{ - typedef _Rp result_type; -}; - -template -struct __weak_result_type<_Rp (_Cp::*)(_A1, _A2, _A3...) const volatile> -{ - typedef _Rp result_type; -}; - -template -struct __invoke_return -{ - typedef decltype(_VSTD::__invoke(declval<_Tp>(), declval<_Args>()...)) type; -}; - -#else // defined(_LIBCPP_CXX03_LANG) - -#include <__functional_base_03> - -#endif // !defined(_LIBCPP_CXX03_LANG) - - -template ::value> -struct __invoke_void_return_wrapper -{ -#ifndef _LIBCPP_CXX03_LANG - template - static _Ret __call(_Args&&... __args) { - return _VSTD::__invoke(_VSTD::forward<_Args>(__args)...); - } -#else - template - static _Ret __call(_Fn __f) { - return _VSTD::__invoke(__f); - } - - template - static _Ret __call(_Fn __f, _A0& __a0) { - return _VSTD::__invoke(__f, __a0); - } - - template - static _Ret __call(_Fn __f, _A0& __a0, _A1& __a1) { - return _VSTD::__invoke(__f, __a0, __a1); - } - - template - static _Ret __call(_Fn __f, _A0& __a0, _A1& __a1, _A2& __a2){ - return _VSTD::__invoke(__f, __a0, __a1, __a2); - } -#endif -}; - -template -struct __invoke_void_return_wrapper<_Ret, true> -{ -#ifndef _LIBCPP_CXX03_LANG - template - static void __call(_Args&&... __args) { - _VSTD::__invoke(_VSTD::forward<_Args>(__args)...); - } -#else - template - static void __call(_Fn __f) { - _VSTD::__invoke(__f); - } - - template - static void __call(_Fn __f, _A0& __a0) { - _VSTD::__invoke(__f, __a0); - } - - template - static void __call(_Fn __f, _A0& __a0, _A1& __a1) { - _VSTD::__invoke(__f, __a0, __a1); - } - - template - static void __call(_Fn __f, _A0& __a0, _A1& __a1, _A2& __a2) { - _VSTD::__invoke(__f, __a0, __a1, __a2); - } -#endif -}; - -template -class _LIBCPP_TEMPLATE_VIS reference_wrapper - : public __weak_result_type<_Tp> -{ -public: - // types - typedef _Tp type; -private: - type* __f_; - -public: - // construct/copy/destroy - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 - reference_wrapper(type& __f) _NOEXCEPT - : __f_(_VSTD::addressof(__f)) {} -#ifndef _LIBCPP_CXX03_LANG - private: reference_wrapper(type&&); public: // = delete; // do not bind to temps -#endif - - // access - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 - operator type&() const _NOEXCEPT {return *__f_;} - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 - type& get() const _NOEXCEPT {return *__f_;} - -#ifndef _LIBCPP_CXX03_LANG - // invoke - template - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 - typename __invoke_of::type - operator() (_ArgTypes&&... __args) const { - return _VSTD::__invoke(get(), _VSTD::forward<_ArgTypes>(__args)...); - } -#else - - _LIBCPP_INLINE_VISIBILITY - typename __invoke_return::type - operator() () const { - return _VSTD::__invoke(get()); - } - - template - _LIBCPP_INLINE_VISIBILITY - typename __invoke_return0::type - operator() (_A0& __a0) const { - return _VSTD::__invoke(get(), __a0); - } - - template - _LIBCPP_INLINE_VISIBILITY - typename __invoke_return0::type - operator() (_A0 const& __a0) const { - return _VSTD::__invoke(get(), __a0); - } - - template - _LIBCPP_INLINE_VISIBILITY - typename __invoke_return1::type - operator() (_A0& __a0, _A1& __a1) const { - return _VSTD::__invoke(get(), __a0, __a1); - } - - template - _LIBCPP_INLINE_VISIBILITY - typename __invoke_return1::type - operator() (_A0 const& __a0, _A1& __a1) const { - return _VSTD::__invoke(get(), __a0, __a1); - } - - template - _LIBCPP_INLINE_VISIBILITY - typename __invoke_return1::type - operator() (_A0& __a0, _A1 const& __a1) const { - return _VSTD::__invoke(get(), __a0, __a1); - } - - template - _LIBCPP_INLINE_VISIBILITY - typename __invoke_return1::type - operator() (_A0 const& __a0, _A1 const& __a1) const { - return _VSTD::__invoke(get(), __a0, __a1); - } - - template - _LIBCPP_INLINE_VISIBILITY - typename __invoke_return2::type - operator() (_A0& __a0, _A1& __a1, _A2& __a2) const { - return _VSTD::__invoke(get(), __a0, __a1, __a2); - } - - template - _LIBCPP_INLINE_VISIBILITY - typename __invoke_return2::type - operator() (_A0 const& __a0, _A1& __a1, _A2& __a2) const { - return _VSTD::__invoke(get(), __a0, __a1, __a2); - } - - template - _LIBCPP_INLINE_VISIBILITY - typename __invoke_return2::type - operator() (_A0& __a0, _A1 const& __a1, _A2& __a2) const { - return _VSTD::__invoke(get(), __a0, __a1, __a2); - } - - template - _LIBCPP_INLINE_VISIBILITY - typename __invoke_return2::type - operator() (_A0& __a0, _A1& __a1, _A2 const& __a2) const { - return _VSTD::__invoke(get(), __a0, __a1, __a2); - } - - template - _LIBCPP_INLINE_VISIBILITY - typename __invoke_return2::type - operator() (_A0 const& __a0, _A1 const& __a1, _A2& __a2) const { - return _VSTD::__invoke(get(), __a0, __a1, __a2); - } - - template - _LIBCPP_INLINE_VISIBILITY - typename __invoke_return2::type - operator() (_A0 const& __a0, _A1& __a1, _A2 const& __a2) const { - return _VSTD::__invoke(get(), __a0, __a1, __a2); - } - - template - _LIBCPP_INLINE_VISIBILITY - typename __invoke_return2::type - operator() (_A0& __a0, _A1 const& __a1, _A2 const& __a2) const { - return _VSTD::__invoke(get(), __a0, __a1, __a2); - } - - template - _LIBCPP_INLINE_VISIBILITY - typename __invoke_return2::type - operator() (_A0 const& __a0, _A1 const& __a1, _A2 const& __a2) const { - return _VSTD::__invoke(get(), __a0, __a1, __a2); - } -#endif // _LIBCPP_CXX03_LANG -}; - - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -reference_wrapper<_Tp> -ref(_Tp& __t) _NOEXCEPT -{ - return reference_wrapper<_Tp>(__t); -} - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -reference_wrapper<_Tp> -ref(reference_wrapper<_Tp> __t) _NOEXCEPT -{ - return _VSTD::ref(__t.get()); -} - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -reference_wrapper -cref(const _Tp& __t) _NOEXCEPT -{ - return reference_wrapper(__t); -} - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -reference_wrapper -cref(reference_wrapper<_Tp> __t) _NOEXCEPT -{ - return _VSTD::cref(__t.get()); -} - -#ifndef _LIBCPP_CXX03_LANG -template void ref(const _Tp&&) = delete; -template void cref(const _Tp&&) = delete; -#endif - -#if _LIBCPP_STD_VER > 11 -template -struct __is_transparent : false_type {}; - -template -struct __is_transparent<_Tp, _Up, - typename __void_t::type> - : true_type {}; -#endif - -// allocator_arg_t - -struct _LIBCPP_TEMPLATE_VIS allocator_arg_t { explicit allocator_arg_t() = default; }; - -#if defined(_LIBCPP_CXX03_LANG) || defined(_LIBCPP_BUILDING_LIBRARY) -extern _LIBCPP_EXPORTED_FROM_ABI const allocator_arg_t allocator_arg; -#else -/* _LIBCPP_INLINE_VAR */ constexpr allocator_arg_t allocator_arg = allocator_arg_t(); -#endif - -// uses_allocator - -template -struct __has_allocator_type -{ -private: - struct __two {char __lx; char __lxx;}; - template static __two __test(...); - template static char __test(typename _Up::allocator_type* = 0); -public: - static const bool value = sizeof(__test<_Tp>(0)) == 1; -}; - -template ::value> -struct __uses_allocator - : public integral_constant::value> -{ -}; - -template -struct __uses_allocator<_Tp, _Alloc, false> - : public false_type -{ -}; - -template -struct _LIBCPP_TEMPLATE_VIS uses_allocator - : public __uses_allocator<_Tp, _Alloc> -{ -}; - -#if _LIBCPP_STD_VER > 14 -template -_LIBCPP_INLINE_VAR constexpr size_t uses_allocator_v = uses_allocator<_Tp, _Alloc>::value; -#endif - -#ifndef _LIBCPP_CXX03_LANG - -// allocator construction - -template -struct __uses_alloc_ctor_imp -{ - typedef _LIBCPP_NODEBUG_TYPE typename __uncvref<_Alloc>::type _RawAlloc; - static const bool __ua = uses_allocator<_Tp, _RawAlloc>::value; - static const bool __ic = - is_constructible<_Tp, allocator_arg_t, _Alloc, _Args...>::value; - static const int value = __ua ? 2 - __ic : 0; -}; - -template -struct __uses_alloc_ctor - : integral_constant::value> - {}; - -template -inline _LIBCPP_INLINE_VISIBILITY -void __user_alloc_construct_impl (integral_constant, _Tp *__storage, const _Allocator &, _Args &&... __args ) -{ - new (__storage) _Tp (_VSTD::forward<_Args>(__args)...); -} - -// FIXME: This should have a version which takes a non-const alloc. -template -inline _LIBCPP_INLINE_VISIBILITY -void __user_alloc_construct_impl (integral_constant, _Tp *__storage, const _Allocator &__a, _Args &&... __args ) -{ - new (__storage) _Tp (allocator_arg, __a, _VSTD::forward<_Args>(__args)...); -} - -// FIXME: This should have a version which takes a non-const alloc. -template -inline _LIBCPP_INLINE_VISIBILITY -void __user_alloc_construct_impl (integral_constant, _Tp *__storage, const _Allocator &__a, _Args &&... __args ) -{ - new (__storage) _Tp (_VSTD::forward<_Args>(__args)..., __a); -} - -#endif // _LIBCPP_CXX03_LANG - -_LIBCPP_END_NAMESPACE_STD - -#endif // _LIBCPP_FUNCTIONAL_BASE +#endif // _LIBCPP_FUNCTIONAL_BASE diff --git a/lib/libcxx/include/__functional_base_03 b/lib/libcxx/include/__functional_base_03 deleted file mode 100644 index 9b08bd26a..000000000 --- a/lib/libcxx/include/__functional_base_03 +++ /dev/null @@ -1,223 +0,0 @@ -// -*- C++ -*- -//===----------------------------------------------------------------------===// -// -// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. -// See https://llvm.org/LICENSE.txt for license information. -// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception -// -//===----------------------------------------------------------------------===// - -#ifndef _LIBCPP_FUNCTIONAL_BASE_03 -#define _LIBCPP_FUNCTIONAL_BASE_03 - -// manual variadic expansion for - -// __invoke - -template -struct __enable_invoke_imp; - -template -struct __enable_invoke_imp<_Ret, _T1, true, true> { - typedef _Ret _Bullet1; - typedef _Bullet1 type; -}; - -template -struct __enable_invoke_imp<_Ret, _T1, true, false> { - typedef _Ret _Bullet2; - typedef _Bullet2 type; -}; - -template -struct __enable_invoke_imp<_Ret, _T1, false, true> { - typedef typename add_lvalue_reference< - typename __apply_cv<_T1, _Ret>::type - >::type _Bullet3; - typedef _Bullet3 type; -}; - -template -struct __enable_invoke_imp<_Ret, _T1, false, false> { - typedef typename add_lvalue_reference< - typename __apply_cv()), _Ret>::type - >::type _Bullet4; - typedef _Bullet4 type; -}; - -template -struct __enable_invoke_imp<_Ret, _T1*, false, false> { - typedef typename add_lvalue_reference< - typename __apply_cv<_T1, _Ret>::type - >::type _Bullet4; - typedef _Bullet4 type; -}; - -template , - class _Ret = typename _Traits::_ReturnType, - class _Class = typename _Traits::_ClassType> -struct __enable_invoke : __enable_invoke_imp< - _Ret, _T1, - is_member_function_pointer<_Fn>::value, - is_base_of<_Class, typename remove_reference<_T1>::type>::value> -{ -}; - -__nat __invoke(__any, ...); - -// first bullet - -template -inline _LIBCPP_INLINE_VISIBILITY -typename __enable_invoke<_Fn, _T1>::_Bullet1 -__invoke(_Fn __f, _T1& __t1) { - return (__t1.*__f)(); -} - -template -inline _LIBCPP_INLINE_VISIBILITY -typename __enable_invoke<_Fn, _T1>::_Bullet1 -__invoke(_Fn __f, _T1& __t1, _A0& __a0) { - return (__t1.*__f)(__a0); -} - -template -inline _LIBCPP_INLINE_VISIBILITY -typename __enable_invoke<_Fn, _T1>::_Bullet1 -__invoke(_Fn __f, _T1& __t1, _A0& __a0, _A1& __a1) { - return (__t1.*__f)(__a0, __a1); -} - -template -inline _LIBCPP_INLINE_VISIBILITY -typename __enable_invoke<_Fn, _T1>::_Bullet1 -__invoke(_Fn __f, _T1& __t1, _A0& __a0, _A1& __a1, _A2& __a2) { - return (__t1.*__f)(__a0, __a1, __a2); -} - -template -inline _LIBCPP_INLINE_VISIBILITY -typename __enable_invoke<_Fn, _T1>::_Bullet2 -__invoke(_Fn __f, _T1& __t1) { - return ((*__t1).*__f)(); -} - -template -inline _LIBCPP_INLINE_VISIBILITY -typename __enable_invoke<_Fn, _T1>::_Bullet2 -__invoke(_Fn __f, _T1& __t1, _A0& __a0) { - return ((*__t1).*__f)(__a0); -} - -template -inline _LIBCPP_INLINE_VISIBILITY -typename __enable_invoke<_Fn, _T1>::_Bullet2 -__invoke(_Fn __f, _T1& __t1, _A0& __a0, _A1& __a1) { - return ((*__t1).*__f)(__a0, __a1); -} - -template -inline _LIBCPP_INLINE_VISIBILITY -typename __enable_invoke<_Fn, _T1>::_Bullet2 -__invoke(_Fn __f, _T1& __t1, _A0& __a0, _A1& __a1, _A2& __a2) { - return ((*__t1).*__f)(__a0, __a1, __a2); -} - -template -inline _LIBCPP_INLINE_VISIBILITY -typename __enable_invoke<_Fn, _T1>::_Bullet3 -__invoke(_Fn __f, _T1& __t1) { - return __t1.*__f; -} - -template -inline _LIBCPP_INLINE_VISIBILITY -typename __enable_invoke<_Fn, _T1>::_Bullet4 -__invoke(_Fn __f, _T1& __t1) { - return (*__t1).*__f; -} - -// fifth bullet - -template -inline _LIBCPP_INLINE_VISIBILITY -decltype(declval<_Fp&>()()) -__invoke(_Fp& __f) -{ - return __f(); -} - -template -inline _LIBCPP_INLINE_VISIBILITY -decltype(declval<_Fp&>()(declval<_A0&>())) -__invoke(_Fp& __f, _A0& __a0) -{ - return __f(__a0); -} - -template -inline _LIBCPP_INLINE_VISIBILITY -decltype(declval<_Fp&>()(declval<_A0&>(), declval<_A1&>())) -__invoke(_Fp& __f, _A0& __a0, _A1& __a1) -{ - return __f(__a0, __a1); -} - -template -inline _LIBCPP_INLINE_VISIBILITY -decltype(declval<_Fp&>()(declval<_A0&>(), declval<_A1&>(), declval<_A2&>())) -__invoke(_Fp& __f, _A0& __a0, _A1& __a1, _A2& __a2) -{ - return __f(__a0, __a1, __a2); -} - -template >::value> -struct __invoke_return -{ - typedef typename __weak_result_type<_Fp>::result_type type; -}; - -template -struct __invoke_return<_Fp, false> -{ - typedef decltype(_VSTD::__invoke(declval<_Fp&>())) type; -}; - -template -struct __invoke_return0 -{ - typedef decltype(_VSTD::__invoke(declval<_Tp&>(), declval<_A0&>())) type; -}; - -template -struct __invoke_return0<_Rp _Tp::*, _A0> -{ - typedef typename __enable_invoke<_Rp _Tp::*, _A0>::type type; -}; - -template -struct __invoke_return1 -{ - typedef decltype(_VSTD::__invoke(declval<_Tp&>(), declval<_A0&>(), - declval<_A1&>())) type; -}; - -template -struct __invoke_return1<_Rp _Class::*, _A0, _A1> { - typedef typename __enable_invoke<_Rp _Class::*, _A0>::type type; -}; - -template -struct __invoke_return2 -{ - typedef decltype(_VSTD::__invoke(declval<_Tp&>(), declval<_A0&>(), - declval<_A1&>(), - declval<_A2&>())) type; -}; - -template -struct __invoke_return2<_Ret _Class::*, _A0, _A1, _A2> { - typedef typename __enable_invoke<_Ret _Class::*, _A0>::type type; -}; -#endif // _LIBCPP_FUNCTIONAL_BASE_03 diff --git a/lib/libcxx/include/__hash_table b/lib/libcxx/include/__hash_table index 521ebbf2c..df0f7c80d 100644 --- a/lib/libcxx/include/__hash_table +++ b/lib/libcxx/include/__hash_table @@ -10,16 +10,16 @@ #ifndef _LIBCPP__HASH_TABLE #define _LIBCPP__HASH_TABLE +#include <__bits> // __libcpp_clz #include <__config> -#include -#include -#include +#include <__debug> #include #include -#include +#include +#include +#include #include - -#include <__debug> +#include #if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) #pragma GCC system_header @@ -89,7 +89,7 @@ struct __hash_node_base }; template -struct __hash_node +struct _LIBCPP_STANDALONE_DEBUG __hash_node : public __hash_node_base < typename __rebind_pointer<_VoidPtr, __hash_node<_Tp, _VoidPtr> >::type @@ -317,7 +317,7 @@ public: } return *this; } -#endif // _LIBCPP_DEBUG_LEVEL == 2 +#endif // _LIBCPP_DEBUG_LEVEL == 2 _LIBCPP_INLINE_VISIBILITY reference operator*() const { @@ -336,7 +336,7 @@ public: _LIBCPP_INLINE_VISIBILITY __hash_iterator& operator++() { _LIBCPP_DEBUG_ASSERT(__get_const_db()->__dereferenceable(this), - "Attempted to increment non-incrementable unordered container iterator"); + "Attempted to increment a non-incrementable unordered container iterator"); __node_ = __node_->__next_; return *this; } @@ -438,7 +438,7 @@ public: } return *this; } -#endif // _LIBCPP_DEBUG_LEVEL == 2 +#endif // _LIBCPP_DEBUG_LEVEL == 2 _LIBCPP_INLINE_VISIBILITY reference operator*() const { @@ -456,7 +456,7 @@ public: _LIBCPP_INLINE_VISIBILITY __hash_const_iterator& operator++() { _LIBCPP_DEBUG_ASSERT(__get_const_db()->__dereferenceable(this), - "Attempted to increment non-incrementable unordered container const_iterator"); + "Attempted to increment a non-incrementable unordered container const_iterator"); __node_ = __node_->__next_; return *this; } @@ -550,7 +550,7 @@ public: } return *this; } -#endif // _LIBCPP_DEBUG_LEVEL == 2 +#endif // _LIBCPP_DEBUG_LEVEL == 2 _LIBCPP_INLINE_VISIBILITY reference operator*() const { @@ -569,7 +569,7 @@ public: _LIBCPP_INLINE_VISIBILITY __hash_local_iterator& operator++() { _LIBCPP_DEBUG_ASSERT(__get_const_db()->__dereferenceable(this), - "Attempted to increment non-incrementable unordered container local_iterator"); + "Attempted to increment a non-incrementable unordered container local_iterator"); __node_ = __node_->__next_; if (__node_ != nullptr && __constrain_hash(__node_->__hash(), __bucket_count_) != __bucket_) __node_ = nullptr; @@ -695,7 +695,7 @@ public: } return *this; } -#endif // _LIBCPP_DEBUG_LEVEL == 2 +#endif // _LIBCPP_DEBUG_LEVEL == 2 _LIBCPP_INLINE_VISIBILITY reference operator*() const { @@ -714,7 +714,7 @@ public: _LIBCPP_INLINE_VISIBILITY __hash_const_local_iterator& operator++() { _LIBCPP_DEBUG_ASSERT(__get_const_db()->__dereferenceable(this), - "Attempted to increment non-incrementable unordered container const_local_iterator"); + "Attempted to increment a non-incrementable unordered container const_local_iterator"); __node_ = __node_->__next_; if (__node_ != nullptr && __constrain_hash(__node_->__hash(), __bucket_count_) != __bucket_) __node_ = nullptr; @@ -741,9 +741,9 @@ public: private: #if _LIBCPP_DEBUG_LEVEL == 2 _LIBCPP_INLINE_VISIBILITY - __hash_const_local_iterator(__next_pointer __node, size_t __bucket, + __hash_const_local_iterator(__next_pointer __node_ptr, size_t __bucket, size_t __bucket_count, const void* __c) _NOEXCEPT - : __node_(__node), + : __node_(__node_ptr), __bucket_(__bucket), __bucket_count_(__bucket_count) { @@ -753,9 +753,9 @@ private: } #else _LIBCPP_INLINE_VISIBILITY - __hash_const_local_iterator(__next_pointer __node, size_t __bucket, + __hash_const_local_iterator(__next_pointer __node_ptr, size_t __bucket, size_t __bucket_count) _NOEXCEPT - : __node_(__node), + : __node_(__node_ptr), __bucket_(__bucket), __bucket_count_(__bucket_count) { @@ -1337,7 +1337,7 @@ public: bool __addable(const const_iterator* __i, ptrdiff_t __n) const; bool __subscriptable(const const_iterator* __i, ptrdiff_t __n) const; -#endif // _LIBCPP_DEBUG_LEVEL == 2 +#endif // _LIBCPP_DEBUG_LEVEL == 2 private: void __rehash(size_type __n); @@ -1645,7 +1645,7 @@ __hash_table<_Tp, _Hash, _Equal, _Alloc>::__move_assign( #ifndef _LIBCPP_NO_EXCEPTIONS try { -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS const_iterator __i = __u.begin(); while (__cache != nullptr && __u.size() != 0) { @@ -1662,7 +1662,7 @@ __hash_table<_Tp, _Hash, _Equal, _Alloc>::__move_assign( __deallocate_node(__cache); throw; } -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS __deallocate_node(__cache); } const_iterator __i = __u.begin(); @@ -1707,7 +1707,7 @@ __hash_table<_Tp, _Hash, _Equal, _Alloc>::__assign_unique(_InputIterator __first #ifndef _LIBCPP_NO_EXCEPTIONS try { -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS for (; __cache != nullptr && __first != __last; ++__first) { __cache->__upcast()->__value_ = *__first; @@ -1722,7 +1722,7 @@ __hash_table<_Tp, _Hash, _Equal, _Alloc>::__assign_unique(_InputIterator __first __deallocate_node(__cache); throw; } -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS __deallocate_node(__cache); } for (; __first != __last; ++__first) @@ -1747,7 +1747,7 @@ __hash_table<_Tp, _Hash, _Equal, _Alloc>::__assign_multi(_InputIterator __first, #ifndef _LIBCPP_NO_EXCEPTIONS try { -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS for (; __cache != nullptr && __first != __last; ++__first) { __cache->__upcast()->__value_ = *__first; @@ -1762,7 +1762,7 @@ __hash_table<_Tp, _Hash, _Equal, _Alloc>::__assign_multi(_InputIterator __first, __deallocate_node(__cache); throw; } -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS __deallocate_node(__cache); } for (; __first != __last; ++__first) @@ -2299,7 +2299,7 @@ __hash_table<_Tp, _Hash, _Equal, _Alloc>::__node_handle_merge_multi( __node_insert_multi_perform(__src_ptr, __pn); } } -#endif // _LIBCPP_STD_VER > 14 +#endif // _LIBCPP_STD_VER > 14 template void @@ -2506,11 +2506,11 @@ __hash_table<_Tp, _Hash, _Equal, _Alloc>::erase(const_iterator __first, { #if _LIBCPP_DEBUG_LEVEL == 2 _LIBCPP_ASSERT(__get_const_db()->__find_c_from_i(&__first) == this, - "unodered container::erase(iterator, iterator) called with an iterator not" - " referring to this unodered container"); + "unordered container::erase(iterator, iterator) called with an iterator not" + " referring to this container"); _LIBCPP_ASSERT(__get_const_db()->__find_c_from_i(&__last) == this, - "unodered container::erase(iterator, iterator) called with an iterator not" - " referring to this unodered container"); + "unordered container::erase(iterator, iterator) called with an iterator not" + " referring to this container"); #endif for (const_iterator __p = __first; __first != __last; __p = __first) { @@ -2804,10 +2804,10 @@ __hash_table<_Tp, _Hash, _Equal, _Alloc>::__subscriptable(const const_iterator*, return false; } -#endif // _LIBCPP_DEBUG_LEVEL == 2 +#endif // _LIBCPP_DEBUG_LEVEL == 2 _LIBCPP_END_NAMESPACE_STD _LIBCPP_POP_MACROS -#endif // _LIBCPP__HASH_TABLE +#endif // _LIBCPP__HASH_TABLE diff --git a/lib/libcxx/include/__iterator/access.h b/lib/libcxx/include/__iterator/access.h new file mode 100644 index 000000000..c0576b459 --- /dev/null +++ b/lib/libcxx/include/__iterator/access.h @@ -0,0 +1,134 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ITERATOR_ACCESS_H +#define _LIBCPP___ITERATOR_ACCESS_H + +#include <__config> +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 +_Tp* +begin(_Tp (&__array)[_Np]) +{ + return __array; +} + +template +_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 +_Tp* +end(_Tp (&__array)[_Np]) +{ + return __array + _Np; +} + +#if !defined(_LIBCPP_CXX03_LANG) + +template +_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 +auto +begin(_Cp& __c) -> decltype(__c.begin()) +{ + return __c.begin(); +} + +template +_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 +auto +begin(const _Cp& __c) -> decltype(__c.begin()) +{ + return __c.begin(); +} + +template +_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 +auto +end(_Cp& __c) -> decltype(__c.end()) +{ + return __c.end(); +} + +template +_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 +auto +end(const _Cp& __c) -> decltype(__c.end()) +{ + return __c.end(); +} + +#if _LIBCPP_STD_VER > 11 + +template +_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 +auto cbegin(const _Cp& __c) -> decltype(_VSTD::begin(__c)) +{ + return _VSTD::begin(__c); +} + +template +_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 +auto cend(const _Cp& __c) -> decltype(_VSTD::end(__c)) +{ + return _VSTD::end(__c); +} + +#endif + + +#else // defined(_LIBCPP_CXX03_LANG) + +template +_LIBCPP_INLINE_VISIBILITY +typename _Cp::iterator +begin(_Cp& __c) +{ + return __c.begin(); +} + +template +_LIBCPP_INLINE_VISIBILITY +typename _Cp::const_iterator +begin(const _Cp& __c) +{ + return __c.begin(); +} + +template +_LIBCPP_INLINE_VISIBILITY +typename _Cp::iterator +end(_Cp& __c) +{ + return __c.end(); +} + +template +_LIBCPP_INLINE_VISIBILITY +typename _Cp::const_iterator +end(const _Cp& __c) +{ + return __c.end(); +} + +#endif // !defined(_LIBCPP_CXX03_LANG) + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ITERATOR_ACCESS_H diff --git a/lib/libcxx/include/__iterator/advance.h b/lib/libcxx/include/__iterator/advance.h new file mode 100644 index 000000000..47bce1ddf --- /dev/null +++ b/lib/libcxx/include/__iterator/advance.h @@ -0,0 +1,200 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ITERATOR_ADVANCE_H +#define _LIBCPP___ITERATOR_ADVANCE_H + +#include <__config> +#include <__debug> +#include <__function_like.h> +#include <__iterator/concepts.h> +#include <__iterator/incrementable_traits.h> +#include <__iterator/iterator_traits.h> +#include <__utility/move.h> +#include +#include +#include +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +_LIBCPP_HIDE_FROM_ABI _LIBCPP_CONSTEXPR_AFTER_CXX14 +void __advance(_InputIter& __i, typename iterator_traits<_InputIter>::difference_type __n, input_iterator_tag) { + for (; __n > 0; --__n) + ++__i; +} + +template +_LIBCPP_HIDE_FROM_ABI _LIBCPP_CONSTEXPR_AFTER_CXX14 +void __advance(_BiDirIter& __i, typename iterator_traits<_BiDirIter>::difference_type __n, bidirectional_iterator_tag) { + if (__n >= 0) + for (; __n > 0; --__n) + ++__i; + else + for (; __n < 0; ++__n) + --__i; +} + +template +_LIBCPP_HIDE_FROM_ABI _LIBCPP_CONSTEXPR_AFTER_CXX14 +void __advance(_RandIter& __i, typename iterator_traits<_RandIter>::difference_type __n, random_access_iterator_tag) { + __i += __n; +} + +template < + class _InputIter, class _Distance, + class _IntegralDistance = decltype(_VSTD::__convert_to_integral(declval<_Distance>())), + class = _EnableIf::value> > +_LIBCPP_HIDE_FROM_ABI _LIBCPP_CONSTEXPR_AFTER_CXX14 +void advance(_InputIter& __i, _Distance __orig_n) { + typedef typename iterator_traits<_InputIter>::difference_type _Difference; + _Difference __n = static_cast<_Difference>(_VSTD::__convert_to_integral(__orig_n)); + _LIBCPP_ASSERT(__n >= 0 || __is_cpp17_bidirectional_iterator<_InputIter>::value, + "Attempt to advance(it, n) with negative n on a non-bidirectional iterator"); + _VSTD::__advance(__i, __n, typename iterator_traits<_InputIter>::iterator_category()); +} + +#if !defined(_LIBCPP_HAS_NO_RANGES) + +namespace ranges { +// [range.iter.op.advance] +struct __advance_fn final : private __function_like { +private: + template + _LIBCPP_HIDE_FROM_ABI + static constexpr _Tp __magnitude_geq(_Tp __a, _Tp __b) noexcept { + return __a < 0 ? (__a <= __b) : (__a >= __b); + } + + template + _LIBCPP_HIDE_FROM_ABI + static constexpr void __advance_forward(_Ip& __i, iter_difference_t<_Ip> __n) { + while (__n > 0) { + --__n; + ++__i; + } + } + + template + _LIBCPP_HIDE_FROM_ABI + static constexpr void __advance_backward(_Ip& __i, iter_difference_t<_Ip> __n) { + while (__n < 0) { + ++__n; + --__i; + } + } + +public: + constexpr explicit __advance_fn(__tag __x) noexcept : __function_like(__x) {} + + // Preconditions: If `I` does not model `bidirectional_iterator`, `n` is not negative. + template + _LIBCPP_HIDE_FROM_ABI + constexpr void operator()(_Ip& __i, iter_difference_t<_Ip> __n) const { + _LIBCPP_ASSERT(__n >= 0 || bidirectional_iterator<_Ip>, + "If `n < 0`, then `bidirectional_iterator` must be true."); + + // If `I` models `random_access_iterator`, equivalent to `i += n`. + if constexpr (random_access_iterator<_Ip>) { + __i += __n; + return; + } else if constexpr (bidirectional_iterator<_Ip>) { + // Otherwise, if `n` is non-negative, increments `i` by `n`. + __advance_forward(__i, __n); + // Otherwise, decrements `i` by `-n`. + __advance_backward(__i, __n); + return; + } else { + // Otherwise, if `n` is non-negative, increments `i` by `n`. + __advance_forward(__i, __n); + return; + } + } + + // Preconditions: Either `assignable_from || sized_sentinel_for` is modeled, or [i, bound) denotes a range. + template _Sp> + _LIBCPP_HIDE_FROM_ABI + constexpr void operator()(_Ip& __i, _Sp __bound) const { + // If `I` and `S` model `assignable_from`, equivalent to `i = std::move(bound)`. + if constexpr (assignable_from<_Ip&, _Sp>) { + __i = _VSTD::move(__bound); + } + // Otherwise, if `S` and `I` model `sized_sentinel_for`, equivalent to `ranges::advance(i, bound - i)`. + else if constexpr (sized_sentinel_for<_Sp, _Ip>) { + (*this)(__i, __bound - __i); + } + // Otherwise, while `bool(i != bound)` is true, increments `i`. + else { + while (__i != __bound) { + ++__i; + } + } + } + + // Preconditions: + // * If `n > 0`, [i, bound) denotes a range. + // * If `n == 0`, [i, bound) or [bound, i) denotes a range. + // * If `n < 0`, [bound, i) denotes a range, `I` models `bidirectional_iterator`, and `I` and `S` model `same_as`. + // Returns: `n - M`, where `M` is the difference between the the ending and starting position. + template _Sp> + _LIBCPP_HIDE_FROM_ABI + constexpr iter_difference_t<_Ip> operator()(_Ip& __i, iter_difference_t<_Ip> __n, _Sp __bound) const { + _LIBCPP_ASSERT((__n >= 0) || (bidirectional_iterator<_Ip> && same_as<_Ip, _Sp>), + "If `n < 0`, then `bidirectional_iterator && same_as` must be true."); + // If `S` and `I` model `sized_sentinel_for`: + if constexpr (sized_sentinel_for<_Sp, _Ip>) { + // If |n| >= |bound - i|, equivalent to `ranges::advance(i, bound)`. + if (const auto __M = __bound - __i; __magnitude_geq(__n, __M)) { + (*this)(__i, __bound); + return __n - __M; + } + + // Otherwise, equivalent to `ranges::advance(i, n)`. + (*this)(__i, __n); + return 0; + } else { + // Otherwise, if `n` is non-negative, while `bool(i != bound)` is true, increments `i` but at + // most `n` times. + while (__i != __bound && __n > 0) { + ++__i; + --__n; + } + + // Otherwise, while `bool(i != bound)` is true, decrements `i` but at most `-n` times. + if constexpr (bidirectional_iterator<_Ip> && same_as<_Ip, _Sp>) { + while (__i != __bound && __n < 0) { + --__i; + ++__n; + } + } + return __n; + } + + _LIBCPP_UNREACHABLE(); + } +}; + +inline constexpr auto advance = __advance_fn(__function_like::__tag()); +} // namespace ranges + +#endif // !defined(_LIBCPP_HAS_NO_RANGES) + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ITERATOR_ADVANCE_H diff --git a/lib/libcxx/include/__iterator/back_insert_iterator.h b/lib/libcxx/include/__iterator/back_insert_iterator.h new file mode 100644 index 000000000..f34cb863b --- /dev/null +++ b/lib/libcxx/include/__iterator/back_insert_iterator.h @@ -0,0 +1,75 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ITERATOR_BACK_INSERT_ITERATOR_H +#define _LIBCPP___ITERATOR_BACK_INSERT_ITERATOR_H + +#include <__config> +#include <__iterator/iterator.h> +#include <__iterator/iterator_traits.h> +#include <__memory/addressof.h> +#include <__utility/move.h> +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +_LIBCPP_SUPPRESS_DEPRECATED_PUSH +template +class _LIBCPP_TEMPLATE_VIS back_insert_iterator +#if _LIBCPP_STD_VER <= 14 || !defined(_LIBCPP_ABI_NO_ITERATOR_BASES) + : public iterator +#endif +{ +_LIBCPP_SUPPRESS_DEPRECATED_POP +protected: + _Container* container; +public: + typedef output_iterator_tag iterator_category; + typedef void value_type; +#if _LIBCPP_STD_VER > 17 + typedef ptrdiff_t difference_type; +#else + typedef void difference_type; +#endif + typedef void pointer; + typedef void reference; + typedef _Container container_type; + + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 explicit back_insert_iterator(_Container& __x) : container(_VSTD::addressof(__x)) {} + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 back_insert_iterator& operator=(const typename _Container::value_type& __value_) + {container->push_back(__value_); return *this;} +#ifndef _LIBCPP_CXX03_LANG + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 back_insert_iterator& operator=(typename _Container::value_type&& __value_) + {container->push_back(_VSTD::move(__value_)); return *this;} +#endif // _LIBCPP_CXX03_LANG + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 back_insert_iterator& operator*() {return *this;} + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 back_insert_iterator& operator++() {return *this;} + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 back_insert_iterator operator++(int) {return *this;} +}; + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +back_insert_iterator<_Container> +back_inserter(_Container& __x) +{ + return back_insert_iterator<_Container>(__x); +} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ITERATOR_BACK_INSERT_ITERATOR_H diff --git a/lib/libcxx/include/__iterator/common_iterator.h b/lib/libcxx/include/__iterator/common_iterator.h new file mode 100644 index 000000000..fb01d8bd4 --- /dev/null +++ b/lib/libcxx/include/__iterator/common_iterator.h @@ -0,0 +1,301 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ITERATOR_COMMON_ITERATOR_H +#define _LIBCPP___ITERATOR_COMMON_ITERATOR_H + +#include <__config> +#include <__debug> +#include <__iterator/concepts.h> +#include <__iterator/incrementable_traits.h> +#include <__iterator/iter_move.h> +#include <__iterator/iter_swap.h> +#include <__iterator/iterator_traits.h> +#include <__iterator/readable_traits.h> +#include +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +#if !defined(_LIBCPP_HAS_NO_RANGES) + +template _Sent> + requires (!same_as<_Iter, _Sent> && copyable<_Iter>) +class common_iterator { + class __proxy { + friend common_iterator; + + iter_value_t<_Iter> __value; + // We can move __x because the only caller verifies that __x is not a reference. + constexpr __proxy(iter_reference_t<_Iter>&& __x) + : __value(_VSTD::move(__x)) {} + + public: + const iter_value_t<_Iter>* operator->() const { + return _VSTD::addressof(__value); + } + }; + + class __postfix_proxy { + friend common_iterator; + + iter_value_t<_Iter> __value; + constexpr __postfix_proxy(iter_reference_t<_Iter>&& __x) + : __value(_VSTD::forward>(__x)) {} + + public: + constexpr static bool __valid_for_iter = + constructible_from, iter_reference_t<_Iter>> && + move_constructible>; + + const iter_value_t<_Iter>& operator*() const { + return __value; + } + }; + +public: + variant<_Iter, _Sent> __hold_; + + common_iterator() requires default_initializable<_Iter> = default; + + constexpr common_iterator(_Iter __i) : __hold_(in_place_type<_Iter>, _VSTD::move(__i)) {} + constexpr common_iterator(_Sent __s) : __hold_(in_place_type<_Sent>, _VSTD::move(__s)) {} + + template + requires convertible_to && convertible_to + constexpr common_iterator(const common_iterator<_I2, _S2>& __other) + : __hold_([&]() -> variant<_Iter, _Sent> { + _LIBCPP_ASSERT(!__other.__hold_.valueless_by_exception(), "Constructed from valueless iterator."); + if (__other.__hold_.index() == 0) + return variant<_Iter, _Sent>{in_place_index<0>, _VSTD::__unchecked_get<0>(__other.__hold_)}; + return variant<_Iter, _Sent>{in_place_index<1>, _VSTD::__unchecked_get<1>(__other.__hold_)}; + }()) {} + + template + requires convertible_to && convertible_to && + assignable_from<_Iter&, const _I2&> && assignable_from<_Sent&, const _S2&> + common_iterator& operator=(const common_iterator<_I2, _S2>& __other) { + _LIBCPP_ASSERT(!__other.__hold_.valueless_by_exception(), "Assigned from valueless iterator."); + + auto __idx = __hold_.index(); + auto __other_idx = __other.__hold_.index(); + + // If they're the same index, just assign. + if (__idx == 0 && __other_idx == 0) + _VSTD::__unchecked_get<0>(__hold_) = _VSTD::__unchecked_get<0>(__other.__hold_); + else if (__idx == 1 && __other_idx == 1) + _VSTD::__unchecked_get<1>(__hold_) = _VSTD::__unchecked_get<1>(__other.__hold_); + + // Otherwise replace with the oposite element. + else if (__other_idx == 1) + __hold_.template emplace<1>(_VSTD::__unchecked_get<1>(__other.__hold_)); + else if (__other_idx == 0) + __hold_.template emplace<0>(_VSTD::__unchecked_get<0>(__other.__hold_)); + + return *this; + } + + decltype(auto) operator*() + { + _LIBCPP_ASSERT(holds_alternative<_Iter>(__hold_), + "Cannot dereference sentinel. Common iterator not holding an iterator."); + return *_VSTD::__unchecked_get<_Iter>(__hold_); + } + + decltype(auto) operator*() const + requires __dereferenceable + { + _LIBCPP_ASSERT(holds_alternative<_Iter>(__hold_), + "Cannot dereference sentinel. Common iterator not holding an iterator."); + return *_VSTD::__unchecked_get<_Iter>(__hold_); + } + + template + decltype(auto) operator->() const + requires indirectly_readable && + (requires(const _I2& __i) { __i.operator->(); } || + is_reference_v> || + constructible_from, iter_reference_t<_I2>>) + { + _LIBCPP_ASSERT(holds_alternative<_Iter>(__hold_), + "Cannot dereference sentinel. Common iterator not holding an iterator."); + + if constexpr (is_pointer_v<_Iter> || requires(const _Iter& __i) { __i.operator->(); }) { + return _VSTD::__unchecked_get<_Iter>(__hold_); + } else if constexpr (is_reference_v>) { + auto&& __tmp = *_VSTD::__unchecked_get<_Iter>(__hold_); + return _VSTD::addressof(__tmp); + } else { + return __proxy(*_VSTD::__unchecked_get<_Iter>(__hold_)); + } + } + + common_iterator& operator++() { + _LIBCPP_ASSERT(holds_alternative<_Iter>(__hold_), + "Cannot increment sentinel. Common iterator not holding an iterator."); + ++_VSTD::__unchecked_get<_Iter>(__hold_); return *this; + } + + decltype(auto) operator++(int) { + _LIBCPP_ASSERT(holds_alternative<_Iter>(__hold_), + "Cannot increment sentinel. Common iterator not holding an iterator."); + + if constexpr (forward_iterator<_Iter>) { + auto __tmp = *this; + ++*this; + return __tmp; + } else if constexpr (requires (_Iter& __i) { { *__i++ } -> __referenceable; } || + !__postfix_proxy::__valid_for_iter) { + return _VSTD::__unchecked_get<_Iter>(__hold_)++; + } else { + __postfix_proxy __p(**this); + ++*this; + return __p; + } + } + + template _S2> + requires sentinel_for<_Sent, _I2> + friend bool operator==(const common_iterator& __x, const common_iterator<_I2, _S2>& __y) { + _LIBCPP_ASSERT(!__x.__hold_.valueless_by_exception() && + !__y.__hold_.valueless_by_exception(), + "One or both common_iterators are valueless. (Cannot compare valueless iterators.)"); + + auto __x_index = __x.__hold_.index(); + auto __y_index = __y.__hold_.index(); + + if (__x_index == __y_index) + return true; + + if (__x_index == 0) + return _VSTD::__unchecked_get<_Iter>(__x.__hold_) == _VSTD::__unchecked_get<_S2>(__y.__hold_); + + return _VSTD::__unchecked_get<_Sent>(__x.__hold_) == _VSTD::__unchecked_get<_I2>(__y.__hold_); + } + + template _S2> + requires sentinel_for<_Sent, _I2> && equality_comparable_with<_Iter, _I2> + friend bool operator==(const common_iterator& __x, const common_iterator<_I2, _S2>& __y) { + _LIBCPP_ASSERT(!__x.__hold_.valueless_by_exception() && + !__y.__hold_.valueless_by_exception(), + "One or both common_iterators are valueless. (Cannot compare valueless iterators.)"); + + auto __x_index = __x.__hold_.index(); + auto __y_index = __y.__hold_.index(); + + if (__x_index == 1 && __y_index == 1) + return true; + + if (__x_index == 0 && __y_index == 0) + return _VSTD::__unchecked_get<_Iter>(__x.__hold_) == _VSTD::__unchecked_get<_I2>(__y.__hold_); + + if (__x_index == 0) + return _VSTD::__unchecked_get<_Iter>(__x.__hold_) == _VSTD::__unchecked_get<_S2>(__y.__hold_); + + return _VSTD::__unchecked_get<_Sent>(__x.__hold_) == _VSTD::__unchecked_get<_I2>(__y.__hold_); + } + + template _I2, sized_sentinel_for<_Iter> _S2> + requires sized_sentinel_for<_Sent, _I2> + friend iter_difference_t<_I2> operator-(const common_iterator& __x, const common_iterator<_I2, _S2>& __y) { + _LIBCPP_ASSERT(!__x.__hold_.valueless_by_exception() && + !__y.__hold_.valueless_by_exception(), + "One or both common_iterators are valueless. (Cannot subtract valueless iterators.)"); + + auto __x_index = __x.__hold_.index(); + auto __y_index = __y.__hold_.index(); + + if (__x_index == 1 && __y_index == 1) + return 0; + + if (__x_index == 0 && __y_index == 0) + return _VSTD::__unchecked_get<_Iter>(__x.__hold_) - _VSTD::__unchecked_get<_I2>(__y.__hold_); + + if (__x_index == 0) + return _VSTD::__unchecked_get<_Iter>(__x.__hold_) - _VSTD::__unchecked_get<_S2>(__y.__hold_); + + return _VSTD::__unchecked_get<_Sent>(__x.__hold_) - _VSTD::__unchecked_get<_I2>(__y.__hold_); + } + + friend iter_rvalue_reference_t<_Iter> iter_move(const common_iterator& __i) + noexcept(noexcept(ranges::iter_move(declval()))) + requires input_iterator<_Iter> + { + _LIBCPP_ASSERT(holds_alternative<_Iter>(__i.__hold_), + "Cannot iter_move a sentinel. Common iterator not holding an iterator."); + return ranges::iter_move( _VSTD::__unchecked_get<_Iter>(__i.__hold_)); + } + + template _I2, class _S2> + friend void iter_swap(const common_iterator& __x, const common_iterator<_I2, _S2>& __y) + noexcept(noexcept(ranges::iter_swap(declval(), declval()))) + { + _LIBCPP_ASSERT(holds_alternative<_Iter>(__x.__hold_), + "Cannot swap __y with a sentinel. Common iterator (__x) not holding an iterator."); + _LIBCPP_ASSERT(holds_alternative<_Iter>(__y.__hold_), + "Cannot swap __x with a sentinel. Common iterator (__y) not holding an iterator."); + return ranges::iter_swap( _VSTD::__unchecked_get<_Iter>(__x.__hold_), _VSTD::__unchecked_get<_Iter>(__y.__hold_)); + } +}; + +template +struct incrementable_traits> { + using difference_type = iter_difference_t<_Iter>; +}; + +template +concept __denotes_forward_iter = + requires { typename iterator_traits<_Iter>::iterator_category; } && + derived_from::iterator_category, forward_iterator_tag>; + +template +concept __common_iter_has_ptr_op = requires(const common_iterator<_Iter, _Sent>& __a) { + __a.operator->(); +}; + +template +struct __arrow_type_or_void { + using type = void; +}; + +template + requires __common_iter_has_ptr_op<_Iter, _Sent> +struct __arrow_type_or_void<_Iter, _Sent> { + using type = decltype(declval>().operator->()); +}; + +template +struct iterator_traits> { + using iterator_concept = _If, + forward_iterator_tag, + input_iterator_tag>; + using iterator_category = _If<__denotes_forward_iter<_Iter>, + forward_iterator_tag, + input_iterator_tag>; + using pointer = typename __arrow_type_or_void<_Iter, _Sent>::type; + using value_type = iter_value_t<_Iter>; + using difference_type = iter_difference_t<_Iter>; + using reference = iter_reference_t<_Iter>; +}; + + +#endif // !defined(_LIBCPP_HAS_NO_RANGES) + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ITERATOR_COMMON_ITERATOR_H diff --git a/lib/libcxx/include/__iterator/concepts.h b/lib/libcxx/include/__iterator/concepts.h new file mode 100644 index 000000000..6eb4aef10 --- /dev/null +++ b/lib/libcxx/include/__iterator/concepts.h @@ -0,0 +1,272 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ITERATOR_CONCEPTS_H +#define _LIBCPP___ITERATOR_CONCEPTS_H + +#include <__config> +#include <__iterator/incrementable_traits.h> +#include <__iterator/iter_move.h> +#include <__iterator/iterator_traits.h> +#include <__iterator/readable_traits.h> +#include <__memory/pointer_traits.h> +#include <__utility/forward.h> +#include +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +#if !defined(_LIBCPP_HAS_NO_RANGES) + +// clang-format off + +// [iterator.concept.readable] +template +concept __indirectly_readable_impl = + requires(const _In __i) { + typename iter_value_t<_In>; + typename iter_reference_t<_In>; + typename iter_rvalue_reference_t<_In>; + { *__i } -> same_as>; + { ranges::iter_move(__i) } -> same_as>; + } && + common_reference_with&&, iter_value_t<_In>&> && + common_reference_with&&, iter_rvalue_reference_t<_In>&&> && + common_reference_with&&, const iter_value_t<_In>&>; + +template +concept indirectly_readable = __indirectly_readable_impl>; + +template +using iter_common_reference_t = common_reference_t, iter_value_t<_Tp>&>; + +// [iterator.concept.writable] +template +concept indirectly_writable = + requires(_Out&& __o, _Tp&& __t) { + *__o = _VSTD::forward<_Tp>(__t); // not required to be equality-preserving + *_VSTD::forward<_Out>(__o) = _VSTD::forward<_Tp>(__t); // not required to be equality-preserving + const_cast&&>(*__o) = _VSTD::forward<_Tp>(__t); // not required to be equality-preserving + const_cast&&>(*_VSTD::forward<_Out>(__o)) = _VSTD::forward<_Tp>(__t); // not required to be equality-preserving + }; + +// [iterator.concept.winc] +template +concept __integer_like = integral<_Tp> && !same_as<_Tp, bool>; + +template +concept __signed_integer_like = signed_integral<_Tp>; + +template +concept weakly_incrementable = + movable<_Ip> && + requires(_Ip __i) { + typename iter_difference_t<_Ip>; + requires __signed_integer_like>; + { ++__i } -> same_as<_Ip&>; // not required to be equality-preserving + __i++; // not required to be equality-preserving + }; + +// [iterator.concept.inc] +template +concept incrementable = + regular<_Ip> && + weakly_incrementable<_Ip> && + requires(_Ip __i) { + { __i++ } -> same_as<_Ip>; + }; + +// [iterator.concept.iterator] +template +concept input_or_output_iterator = + requires(_Ip __i) { + { *__i } -> __referenceable; + } && + weakly_incrementable<_Ip>; + +// [iterator.concept.sentinel] +template +concept sentinel_for = + semiregular<_Sp> && + input_or_output_iterator<_Ip> && + __weakly_equality_comparable_with<_Sp, _Ip>; + +template +inline constexpr bool disable_sized_sentinel_for = false; + +template +concept sized_sentinel_for = + sentinel_for<_Sp, _Ip> && + !disable_sized_sentinel_for, remove_cv_t<_Ip>> && + requires(const _Ip& __i, const _Sp& __s) { + { __s - __i } -> same_as>; + { __i - __s } -> same_as>; + }; + +// [iterator.concept.input] +template +concept input_iterator = + input_or_output_iterator<_Ip> && + indirectly_readable<_Ip> && + requires { typename _ITER_CONCEPT<_Ip>; } && + derived_from<_ITER_CONCEPT<_Ip>, input_iterator_tag>; + +// [iterator.concept.output] +template +concept output_iterator = + input_or_output_iterator<_Ip> && + indirectly_writable<_Ip, _Tp> && + requires (_Ip __it, _Tp&& __t) { + *__it++ = _VSTD::forward<_Tp>(__t); // not required to be equality-preserving + }; + +// [iterator.concept.forward] +template +concept forward_iterator = + input_iterator<_Ip> && + derived_from<_ITER_CONCEPT<_Ip>, forward_iterator_tag> && + incrementable<_Ip> && + sentinel_for<_Ip, _Ip>; + +// [iterator.concept.bidir] +template +concept bidirectional_iterator = + forward_iterator<_Ip> && + derived_from<_ITER_CONCEPT<_Ip>, bidirectional_iterator_tag> && + requires(_Ip __i) { + { --__i } -> same_as<_Ip&>; + { __i-- } -> same_as<_Ip>; + }; + +template +concept random_access_iterator = + bidirectional_iterator<_Ip> && + derived_from<_ITER_CONCEPT<_Ip>, random_access_iterator_tag> && + totally_ordered<_Ip> && + sized_sentinel_for<_Ip, _Ip> && + requires(_Ip __i, const _Ip __j, const iter_difference_t<_Ip> __n) { + { __i += __n } -> same_as<_Ip&>; + { __j + __n } -> same_as<_Ip>; + { __n + __j } -> same_as<_Ip>; + { __i -= __n } -> same_as<_Ip&>; + { __j - __n } -> same_as<_Ip>; + { __j[__n] } -> same_as>; + }; + +template +concept contiguous_iterator = + random_access_iterator<_Ip> && + derived_from<_ITER_CONCEPT<_Ip>, contiguous_iterator_tag> && + is_lvalue_reference_v> && + same_as, remove_cvref_t>> && + (is_pointer_v<_Ip> || requires { sizeof(__pointer_traits_element_type<_Ip>); }) && + requires(const _Ip& __i) { + { _VSTD::to_address(__i) } -> same_as>>; + }; + +template +concept __has_arrow = input_iterator<_Ip> && (is_pointer_v<_Ip> || requires(_Ip __i) { __i.operator->(); }); + +// [indirectcallable.indirectinvocable] +template +concept indirectly_unary_invocable = + indirectly_readable<_It> && + copy_constructible<_Fp> && + invocable<_Fp&, iter_value_t<_It>&> && + invocable<_Fp&, iter_reference_t<_It>> && + invocable<_Fp&, iter_common_reference_t<_It>> && + common_reference_with< + invoke_result_t<_Fp&, iter_value_t<_It>&>, + invoke_result_t<_Fp&, iter_reference_t<_It>>>; + +template +concept indirectly_regular_unary_invocable = + indirectly_readable<_It> && + copy_constructible<_Fp> && + regular_invocable<_Fp&, iter_value_t<_It>&> && + regular_invocable<_Fp&, iter_reference_t<_It>> && + regular_invocable<_Fp&, iter_common_reference_t<_It>> && + common_reference_with< + invoke_result_t<_Fp&, iter_value_t<_It>&>, + invoke_result_t<_Fp&, iter_reference_t<_It>>>; + +template +concept indirect_unary_predicate = + indirectly_readable<_It> && + copy_constructible<_Fp> && + predicate<_Fp&, iter_value_t<_It>&> && + predicate<_Fp&, iter_reference_t<_It>> && + predicate<_Fp&, iter_common_reference_t<_It>>; + +template +concept indirect_binary_predicate = + indirectly_readable<_It1> && indirectly_readable<_It2> && + copy_constructible<_Fp> && + predicate<_Fp&, iter_value_t<_It1>&, iter_value_t<_It2>&> && + predicate<_Fp&, iter_value_t<_It1>&, iter_reference_t<_It2>> && + predicate<_Fp&, iter_reference_t<_It1>, iter_value_t<_It2>&> && + predicate<_Fp&, iter_reference_t<_It1>, iter_reference_t<_It2>> && + predicate<_Fp&, iter_common_reference_t<_It1>, iter_common_reference_t<_It2>>; + +template +concept indirect_equivalence_relation = + indirectly_readable<_It1> && indirectly_readable<_It2> && + copy_constructible<_Fp> && + equivalence_relation<_Fp&, iter_value_t<_It1>&, iter_value_t<_It2>&> && + equivalence_relation<_Fp&, iter_value_t<_It1>&, iter_reference_t<_It2>> && + equivalence_relation<_Fp&, iter_reference_t<_It1>, iter_value_t<_It2>&> && + equivalence_relation<_Fp&, iter_reference_t<_It1>, iter_reference_t<_It2>> && + equivalence_relation<_Fp&, iter_common_reference_t<_It1>, iter_common_reference_t<_It2>>; + +template +concept indirect_strict_weak_order = + indirectly_readable<_It1> && indirectly_readable<_It2> && + copy_constructible<_Fp> && + strict_weak_order<_Fp&, iter_value_t<_It1>&, iter_value_t<_It2>&> && + strict_weak_order<_Fp&, iter_value_t<_It1>&, iter_reference_t<_It2>> && + strict_weak_order<_Fp&, iter_reference_t<_It1>, iter_value_t<_It2>&> && + strict_weak_order<_Fp&, iter_reference_t<_It1>, iter_reference_t<_It2>> && + strict_weak_order<_Fp&, iter_common_reference_t<_It1>, iter_common_reference_t<_It2>>; + +template + requires (indirectly_readable<_Its> && ...) && invocable<_Fp, iter_reference_t<_Its>...> +using indirect_result_t = invoke_result_t<_Fp, iter_reference_t<_Its>...>; + +template +concept indirectly_movable = + indirectly_readable<_In> && + indirectly_writable<_Out, iter_rvalue_reference_t<_In>>; + +template +concept indirectly_movable_storable = + indirectly_movable<_In, _Out> && + indirectly_writable<_Out, iter_value_t<_In>> && + movable> && + constructible_from, iter_rvalue_reference_t<_In>> && + assignable_from&, iter_rvalue_reference_t<_In>>; + +// Note: indirectly_swappable is located in iter_swap.h to prevent a dependency cycle +// (both iter_swap and indirectly_swappable require indirectly_readable). + +// clang-format on + +#endif // !defined(_LIBCPP_HAS_NO_RANGES) + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ITERATOR_CONCEPTS_H diff --git a/lib/libcxx/include/__iterator/counted_iterator.h b/lib/libcxx/include/__iterator/counted_iterator.h new file mode 100644 index 000000000..7136aaf02 --- /dev/null +++ b/lib/libcxx/include/__iterator/counted_iterator.h @@ -0,0 +1,306 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// +#ifndef _LIBCPP___ITERATOR_COUNTED_ITERATOR_H +#define _LIBCPP___ITERATOR_COUNTED_ITERATOR_H + +#include <__config> +#include <__debug> +#include <__iterator/concepts.h> +#include <__iterator/default_sentinel.h> +#include <__iterator/iter_move.h> +#include <__iterator/iter_swap.h> +#include <__iterator/incrementable_traits.h> +#include <__iterator/iterator_traits.h> +#include <__iterator/readable_traits.h> +#include <__memory/pointer_traits.h> +#include +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +#if !defined(_LIBCPP_HAS_NO_RANGES) + +template +struct __counted_iterator_concept {}; + +template + requires requires { typename _Iter::iterator_concept; } +struct __counted_iterator_concept<_Iter> { + using iterator_concept = typename _Iter::iterator_concept; +}; + +template +struct __counted_iterator_category {}; + +template + requires requires { typename _Iter::iterator_category; } +struct __counted_iterator_category<_Iter> { + using iterator_category = typename _Iter::iterator_category; +}; + +template +struct __counted_iterator_value_type {}; + +template +struct __counted_iterator_value_type<_Iter> { + using value_type = iter_value_t<_Iter>; +}; + +template +class counted_iterator + : public __counted_iterator_concept<_Iter> + , public __counted_iterator_category<_Iter> + , public __counted_iterator_value_type<_Iter> +{ +public: + [[no_unique_address]] _Iter __current_ = _Iter(); + iter_difference_t<_Iter> __count_ = 0; + + using iterator_type = _Iter; + using difference_type = iter_difference_t<_Iter>; + + _LIBCPP_HIDE_FROM_ABI + constexpr counted_iterator() requires default_initializable<_Iter> = default; + + _LIBCPP_HIDE_FROM_ABI + constexpr counted_iterator(_Iter __iter, iter_difference_t<_Iter> __n) + : __current_(_VSTD::move(__iter)), __count_(__n) { + _LIBCPP_ASSERT(__n >= 0, "__n must not be negative."); + } + + template + requires convertible_to + _LIBCPP_HIDE_FROM_ABI + constexpr counted_iterator(const counted_iterator<_I2>& __other) + : __current_(__other.__current_), __count_(__other.__count_) {} + + template + requires assignable_from<_Iter&, const _I2&> + _LIBCPP_HIDE_FROM_ABI + constexpr counted_iterator& operator=(const counted_iterator<_I2>& __other) { + __current_ = __other.__current_; + __count_ = __other.__count_; + return *this; + } + + _LIBCPP_HIDE_FROM_ABI + constexpr const _Iter& base() const& { return __current_; } + + _LIBCPP_HIDE_FROM_ABI + constexpr _Iter base() && { return _VSTD::move(__current_); } + + _LIBCPP_HIDE_FROM_ABI + constexpr iter_difference_t<_Iter> count() const noexcept { return __count_; } + + _LIBCPP_HIDE_FROM_ABI + constexpr decltype(auto) operator*() { + _LIBCPP_ASSERT(__count_ > 0, "Iterator is equal to or past end."); + return *__current_; + } + + _LIBCPP_HIDE_FROM_ABI + constexpr decltype(auto) operator*() const + requires __dereferenceable + { + _LIBCPP_ASSERT(__count_ > 0, "Iterator is equal to or past end."); + return *__current_; + } + + _LIBCPP_HIDE_FROM_ABI + constexpr auto operator->() const noexcept + requires contiguous_iterator<_Iter> + { + return _VSTD::to_address(__current_); + } + + _LIBCPP_HIDE_FROM_ABI + constexpr counted_iterator& operator++() { + _LIBCPP_ASSERT(__count_ > 0, "Iterator already at or past end."); + ++__current_; + --__count_; + return *this; + } + + _LIBCPP_HIDE_FROM_ABI + decltype(auto) operator++(int) { + _LIBCPP_ASSERT(__count_ > 0, "Iterator already at or past end."); + --__count_; +#ifndef _LIBCPP_NO_EXCEPTIONS + try { return __current_++; } + catch(...) { ++__count_; throw; } +#else + return __current_++; +#endif // _LIBCPP_NO_EXCEPTIONS + } + + _LIBCPP_HIDE_FROM_ABI + constexpr counted_iterator operator++(int) + requires forward_iterator<_Iter> + { + _LIBCPP_ASSERT(__count_ > 0, "Iterator already at or past end."); + counted_iterator __tmp = *this; + ++*this; + return __tmp; + } + + _LIBCPP_HIDE_FROM_ABI + constexpr counted_iterator& operator--() + requires bidirectional_iterator<_Iter> + { + --__current_; + ++__count_; + return *this; + } + + _LIBCPP_HIDE_FROM_ABI + constexpr counted_iterator operator--(int) + requires bidirectional_iterator<_Iter> + { + counted_iterator __tmp = *this; + --*this; + return __tmp; + } + + _LIBCPP_HIDE_FROM_ABI + constexpr counted_iterator operator+(iter_difference_t<_Iter> __n) const + requires random_access_iterator<_Iter> + { + return counted_iterator(__current_ + __n, __count_ - __n); + } + + _LIBCPP_HIDE_FROM_ABI + friend constexpr counted_iterator operator+( + iter_difference_t<_Iter> __n, const counted_iterator& __x) + requires random_access_iterator<_Iter> + { + return __x + __n; + } + + _LIBCPP_HIDE_FROM_ABI + constexpr counted_iterator& operator+=(iter_difference_t<_Iter> __n) + requires random_access_iterator<_Iter> + { + _LIBCPP_ASSERT(__n <= __count_, "Cannot advance iterator past end."); + __current_ += __n; + __count_ -= __n; + return *this; + } + + _LIBCPP_HIDE_FROM_ABI + constexpr counted_iterator operator-(iter_difference_t<_Iter> __n) const + requires random_access_iterator<_Iter> + { + return counted_iterator(__current_ - __n, __count_ + __n); + } + + template _I2> + _LIBCPP_HIDE_FROM_ABI + friend constexpr iter_difference_t<_I2> operator-( + const counted_iterator& __lhs, const counted_iterator<_I2>& __rhs) + { + return __rhs.__count_ - __lhs.__count_; + } + + _LIBCPP_HIDE_FROM_ABI + friend constexpr iter_difference_t<_Iter> operator-( + const counted_iterator& __lhs, default_sentinel_t) + { + return -__lhs.__count_; + } + + _LIBCPP_HIDE_FROM_ABI + friend constexpr iter_difference_t<_Iter> operator-( + default_sentinel_t, const counted_iterator& __rhs) + { + return __rhs.__count_; + } + + _LIBCPP_HIDE_FROM_ABI + constexpr counted_iterator& operator-=(iter_difference_t<_Iter> __n) + requires random_access_iterator<_Iter> + { + _LIBCPP_ASSERT(-__n <= __count_, "Attempt to subtract too large of a size: " + "counted_iterator would be decremented before the " + "first element of its range."); + __current_ -= __n; + __count_ += __n; + return *this; + } + + _LIBCPP_HIDE_FROM_ABI + constexpr decltype(auto) operator[](iter_difference_t<_Iter> __n) const + requires random_access_iterator<_Iter> + { + _LIBCPP_ASSERT(__n < __count_, "Subscript argument must be less than size."); + return __current_[__n]; + } + + template _I2> + _LIBCPP_HIDE_FROM_ABI + friend constexpr bool operator==( + const counted_iterator& __lhs, const counted_iterator<_I2>& __rhs) + { + return __lhs.__count_ == __rhs.__count_; + } + + _LIBCPP_HIDE_FROM_ABI + friend constexpr bool operator==( + const counted_iterator& __lhs, default_sentinel_t) + { + return __lhs.__count_ == 0; + } + + template _I2> + friend constexpr strong_ordering operator<=>( + const counted_iterator& __lhs, const counted_iterator<_I2>& __rhs) + { + return __rhs.__count_ <=> __lhs.__count_; + } + + _LIBCPP_HIDE_FROM_ABI + friend constexpr iter_rvalue_reference_t<_Iter> iter_move(const counted_iterator& __i) + noexcept(noexcept(ranges::iter_move(__i.__current_))) + requires input_iterator<_Iter> + { + _LIBCPP_ASSERT(__i.__count_ > 0, "Iterator must not be past end of range."); + return ranges::iter_move(__i.__current_); + } + + template _I2> + _LIBCPP_HIDE_FROM_ABI + friend constexpr void iter_swap(const counted_iterator& __x, const counted_iterator<_I2>& __y) + noexcept(noexcept(ranges::iter_swap(__x.__current_, __y.__current_))) + { + _LIBCPP_ASSERT(__x.__count_ > 0 && __y.__count_ > 0, + "Iterators must not be past end of range."); + return ranges::iter_swap(__x.__current_, __y.__current_); + } +}; + +template + requires same_as<_ITER_TRAITS<_Iter>, iterator_traits<_Iter>> +struct iterator_traits> : iterator_traits<_Iter> { + using pointer = conditional_t, + add_pointer_t>, void>; +}; + +#endif // !defined(_LIBCPP_HAS_NO_RANGES) + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ITERATOR_COUNTED_ITERATOR_H diff --git a/lib/libcxx/include/__iterator/data.h b/lib/libcxx/include/__iterator/data.h new file mode 100644 index 000000000..cd8e37b96 --- /dev/null +++ b/lib/libcxx/include/__iterator/data.h @@ -0,0 +1,56 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ITERATOR_DATA_H +#define _LIBCPP___ITERATOR_DATA_H + +#include <__config> +#include +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +#if _LIBCPP_STD_VER > 14 + +template constexpr +_LIBCPP_INLINE_VISIBILITY +auto data(_Cont& __c) +_NOEXCEPT_(noexcept(__c.data())) +-> decltype (__c.data()) +{ return __c.data(); } + +template constexpr +_LIBCPP_INLINE_VISIBILITY +auto data(const _Cont& __c) +_NOEXCEPT_(noexcept(__c.data())) +-> decltype (__c.data()) +{ return __c.data(); } + +template +_LIBCPP_INLINE_VISIBILITY +constexpr _Tp* data(_Tp (&__array)[_Sz]) noexcept { return __array; } + +template +_LIBCPP_INLINE_VISIBILITY +constexpr const _Ep* data(initializer_list<_Ep> __il) noexcept { return __il.begin(); } + +#endif + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ITERATOR_DATA_H diff --git a/lib/libcxx/include/__iterator/default_sentinel.h b/lib/libcxx/include/__iterator/default_sentinel.h new file mode 100644 index 000000000..934a56fd9 --- /dev/null +++ b/lib/libcxx/include/__iterator/default_sentinel.h @@ -0,0 +1,35 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ITERATOR_DEFAULT_SENTINEL_H +#define _LIBCPP___ITERATOR_DEFAULT_SENTINEL_H + +#include <__config> + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +#if !defined(_LIBCPP_HAS_NO_RANGES) + +struct default_sentinel_t { }; +inline constexpr default_sentinel_t default_sentinel{}; + +#endif // !defined(_LIBCPP_HAS_NO_RANGES) + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ITERATOR_DEFAULT_SENTINEL_H diff --git a/lib/libcxx/include/__iterator/distance.h b/lib/libcxx/include/__iterator/distance.h new file mode 100644 index 000000000..33e4af84d --- /dev/null +++ b/lib/libcxx/include/__iterator/distance.h @@ -0,0 +1,56 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ITERATOR_DISTANCE_H +#define _LIBCPP___ITERATOR_DISTANCE_H + +#include <__config> +#include <__iterator/iterator_traits.h> + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 +typename iterator_traits<_InputIter>::difference_type +__distance(_InputIter __first, _InputIter __last, input_iterator_tag) +{ + typename iterator_traits<_InputIter>::difference_type __r(0); + for (; __first != __last; ++__first) + ++__r; + return __r; +} + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 +typename iterator_traits<_RandIter>::difference_type +__distance(_RandIter __first, _RandIter __last, random_access_iterator_tag) +{ + return __last - __first; +} + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 +typename iterator_traits<_InputIter>::difference_type +distance(_InputIter __first, _InputIter __last) +{ + return _VSTD::__distance(__first, __last, typename iterator_traits<_InputIter>::iterator_category()); +} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ITERATOR_DISTANCE_H diff --git a/lib/libcxx/include/__iterator/empty.h b/lib/libcxx/include/__iterator/empty.h new file mode 100644 index 000000000..4dd59f5cc --- /dev/null +++ b/lib/libcxx/include/__iterator/empty.h @@ -0,0 +1,49 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ITERATOR_EMPTY_H +#define _LIBCPP___ITERATOR_EMPTY_H + +#include <__config> +#include +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +#if _LIBCPP_STD_VER > 14 + +template +_LIBCPP_NODISCARD_AFTER_CXX17 _LIBCPP_INLINE_VISIBILITY +constexpr auto empty(const _Cont& __c) +_NOEXCEPT_(noexcept(__c.empty())) +-> decltype (__c.empty()) +{ return __c.empty(); } + +template +_LIBCPP_NODISCARD_AFTER_CXX17 _LIBCPP_INLINE_VISIBILITY +constexpr bool empty(const _Tp (&)[_Sz]) noexcept { return false; } + +template +_LIBCPP_NODISCARD_AFTER_CXX17 _LIBCPP_INLINE_VISIBILITY +constexpr bool empty(initializer_list<_Ep> __il) noexcept { return __il.size() == 0; } + +#endif // _LIBCPP_STD_VER > 14 + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ITERATOR_EMPTY_H diff --git a/lib/libcxx/include/__iterator/erase_if_container.h b/lib/libcxx/include/__iterator/erase_if_container.h new file mode 100644 index 000000000..a5dfd0720 --- /dev/null +++ b/lib/libcxx/include/__iterator/erase_if_container.h @@ -0,0 +1,45 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ITERATOR_ERASE_IF_CONTAINER_H +#define _LIBCPP___ITERATOR_ERASE_IF_CONTAINER_H + +#include <__config> + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +_LIBCPP_HIDE_FROM_ABI +typename _Container::size_type +__libcpp_erase_if_container(_Container& __c, _Predicate& __pred) { + typename _Container::size_type __old_size = __c.size(); + + const typename _Container::iterator __last = __c.end(); + for (typename _Container::iterator __iter = __c.begin(); __iter != __last;) { + if (__pred(*__iter)) + __iter = __c.erase(__iter); + else + ++__iter; + } + + return __old_size - __c.size(); +} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ITERATOR_ERASE_IF_CONTAINER_H diff --git a/lib/libcxx/include/__iterator/front_insert_iterator.h b/lib/libcxx/include/__iterator/front_insert_iterator.h new file mode 100644 index 000000000..0421dd5c4 --- /dev/null +++ b/lib/libcxx/include/__iterator/front_insert_iterator.h @@ -0,0 +1,75 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ITERATOR_FRONT_INSERT_ITERATOR_H +#define _LIBCPP___ITERATOR_FRONT_INSERT_ITERATOR_H + +#include <__config> +#include <__iterator/iterator.h> +#include <__iterator/iterator_traits.h> +#include <__memory/addressof.h> +#include <__utility/move.h> +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +_LIBCPP_SUPPRESS_DEPRECATED_PUSH +template +class _LIBCPP_TEMPLATE_VIS front_insert_iterator +#if _LIBCPP_STD_VER <= 14 || !defined(_LIBCPP_ABI_NO_ITERATOR_BASES) + : public iterator +#endif +{ +_LIBCPP_SUPPRESS_DEPRECATED_POP +protected: + _Container* container; +public: + typedef output_iterator_tag iterator_category; + typedef void value_type; +#if _LIBCPP_STD_VER > 17 + typedef ptrdiff_t difference_type; +#else + typedef void difference_type; +#endif + typedef void pointer; + typedef void reference; + typedef _Container container_type; + + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 explicit front_insert_iterator(_Container& __x) : container(_VSTD::addressof(__x)) {} + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 front_insert_iterator& operator=(const typename _Container::value_type& __value_) + {container->push_front(__value_); return *this;} +#ifndef _LIBCPP_CXX03_LANG + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 front_insert_iterator& operator=(typename _Container::value_type&& __value_) + {container->push_front(_VSTD::move(__value_)); return *this;} +#endif // _LIBCPP_CXX03_LANG + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 front_insert_iterator& operator*() {return *this;} + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 front_insert_iterator& operator++() {return *this;} + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 front_insert_iterator operator++(int) {return *this;} +}; + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +front_insert_iterator<_Container> +front_inserter(_Container& __x) +{ + return front_insert_iterator<_Container>(__x); +} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ITERATOR_FRONT_INSERT_ITERATOR_H diff --git a/lib/libcxx/include/__iterator/incrementable_traits.h b/lib/libcxx/include/__iterator/incrementable_traits.h new file mode 100644 index 000000000..5a433982b --- /dev/null +++ b/lib/libcxx/include/__iterator/incrementable_traits.h @@ -0,0 +1,77 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ITERATOR_INCREMENTABLE_TRAITS_H +#define _LIBCPP___ITERATOR_INCREMENTABLE_TRAITS_H + +#include <__config> +#include +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +#if !defined(_LIBCPP_HAS_NO_RANGES) + +// [incrementable.traits] +template struct incrementable_traits {}; + +template +requires is_object_v<_Tp> +struct incrementable_traits<_Tp*> { + using difference_type = ptrdiff_t; +}; + +template +struct incrementable_traits : incrementable_traits<_Ip> {}; + +template +concept __has_member_difference_type = requires { typename _Tp::difference_type; }; + +template<__has_member_difference_type _Tp> +struct incrementable_traits<_Tp> { + using difference_type = typename _Tp::difference_type; +}; + +template +concept __has_integral_minus = + requires(const _Tp& __x, const _Tp& __y) { + { __x - __y } -> integral; + }; + +template<__has_integral_minus _Tp> +requires (!__has_member_difference_type<_Tp>) +struct incrementable_traits<_Tp> { + using difference_type = make_signed_t() - declval<_Tp>())>; +}; + +template +struct iterator_traits; + +// Let `RI` be `remove_cvref_t`. The type `iter_difference_t` denotes +// `incrementable_traits::difference_type` if `iterator_traits` names a specialization +// generated from the primary template, and `iterator_traits::difference_type` otherwise. +template +using iter_difference_t = typename conditional_t<__is_primary_template > >::value, + incrementable_traits >, + iterator_traits > >::difference_type; + +#endif // !defined(_LIBCPP_HAS_NO_RANGES) + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ITERATOR_INCREMENTABLE_TRAITS_H diff --git a/lib/libcxx/include/__iterator/insert_iterator.h b/lib/libcxx/include/__iterator/insert_iterator.h new file mode 100644 index 000000000..265814182 --- /dev/null +++ b/lib/libcxx/include/__iterator/insert_iterator.h @@ -0,0 +1,77 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ITERATOR_INSERT_ITERATOR_H +#define _LIBCPP___ITERATOR_INSERT_ITERATOR_H + +#include <__config> +#include <__iterator/iterator.h> +#include <__iterator/iterator_traits.h> +#include <__memory/addressof.h> +#include <__utility/move.h> +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +_LIBCPP_SUPPRESS_DEPRECATED_PUSH +template +class _LIBCPP_TEMPLATE_VIS insert_iterator +#if _LIBCPP_STD_VER <= 14 || !defined(_LIBCPP_ABI_NO_ITERATOR_BASES) + : public iterator +#endif +{ +_LIBCPP_SUPPRESS_DEPRECATED_POP +protected: + _Container* container; + typename _Container::iterator iter; // FIXME: `ranges::iterator_t` in C++20 mode +public: + typedef output_iterator_tag iterator_category; + typedef void value_type; +#if _LIBCPP_STD_VER > 17 + typedef ptrdiff_t difference_type; +#else + typedef void difference_type; +#endif + typedef void pointer; + typedef void reference; + typedef _Container container_type; + + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 insert_iterator(_Container& __x, typename _Container::iterator __i) + : container(_VSTD::addressof(__x)), iter(__i) {} + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 insert_iterator& operator=(const typename _Container::value_type& __value_) + {iter = container->insert(iter, __value_); ++iter; return *this;} +#ifndef _LIBCPP_CXX03_LANG + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 insert_iterator& operator=(typename _Container::value_type&& __value_) + {iter = container->insert(iter, _VSTD::move(__value_)); ++iter; return *this;} +#endif // _LIBCPP_CXX03_LANG + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 insert_iterator& operator*() {return *this;} + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 insert_iterator& operator++() {return *this;} + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 insert_iterator& operator++(int) {return *this;} +}; + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +insert_iterator<_Container> +inserter(_Container& __x, typename _Container::iterator __i) +{ + return insert_iterator<_Container>(__x, __i); +} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ITERATOR_INSERT_ITERATOR_H diff --git a/lib/libcxx/include/__iterator/istream_iterator.h b/lib/libcxx/include/__iterator/istream_iterator.h new file mode 100644 index 000000000..f39faa6d5 --- /dev/null +++ b/lib/libcxx/include/__iterator/istream_iterator.h @@ -0,0 +1,103 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ITERATOR_ISTREAM_ITERATOR_H +#define _LIBCPP___ITERATOR_ISTREAM_ITERATOR_H + +#include <__config> +#include <__iterator/iterator.h> +#include <__iterator/iterator_traits.h> +#include <__memory/addressof.h> +#include // for forward declarations of char_traits and basic_istream + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +_LIBCPP_SUPPRESS_DEPRECATED_PUSH +template , class _Distance = ptrdiff_t> +class _LIBCPP_TEMPLATE_VIS istream_iterator +#if _LIBCPP_STD_VER <= 14 || !defined(_LIBCPP_ABI_NO_ITERATOR_BASES) + : public iterator +#endif +{ +_LIBCPP_SUPPRESS_DEPRECATED_POP +public: + typedef input_iterator_tag iterator_category; + typedef _Tp value_type; + typedef _Distance difference_type; + typedef const _Tp* pointer; + typedef const _Tp& reference; + typedef _CharT char_type; + typedef _Traits traits_type; + typedef basic_istream<_CharT,_Traits> istream_type; +private: + istream_type* __in_stream_; + _Tp __value_; +public: + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR istream_iterator() : __in_stream_(nullptr), __value_() {} + _LIBCPP_INLINE_VISIBILITY istream_iterator(istream_type& __s) : __in_stream_(_VSTD::addressof(__s)) + { + if (!(*__in_stream_ >> __value_)) + __in_stream_ = nullptr; + } + + _LIBCPP_INLINE_VISIBILITY const _Tp& operator*() const {return __value_;} + _LIBCPP_INLINE_VISIBILITY const _Tp* operator->() const {return _VSTD::addressof((operator*()));} + _LIBCPP_INLINE_VISIBILITY istream_iterator& operator++() + { + if (!(*__in_stream_ >> __value_)) + __in_stream_ = nullptr; + return *this; + } + _LIBCPP_INLINE_VISIBILITY istream_iterator operator++(int) + {istream_iterator __t(*this); ++(*this); return __t;} + + template + friend _LIBCPP_INLINE_VISIBILITY + bool + operator==(const istream_iterator<_Up, _CharU, _TraitsU, _DistanceU>& __x, + const istream_iterator<_Up, _CharU, _TraitsU, _DistanceU>& __y); + + template + friend _LIBCPP_INLINE_VISIBILITY + bool + operator==(const istream_iterator<_Up, _CharU, _TraitsU, _DistanceU>& __x, + const istream_iterator<_Up, _CharU, _TraitsU, _DistanceU>& __y); +}; + +template +inline _LIBCPP_INLINE_VISIBILITY +bool +operator==(const istream_iterator<_Tp, _CharT, _Traits, _Distance>& __x, + const istream_iterator<_Tp, _CharT, _Traits, _Distance>& __y) +{ + return __x.__in_stream_ == __y.__in_stream_; +} + +template +inline _LIBCPP_INLINE_VISIBILITY +bool +operator!=(const istream_iterator<_Tp, _CharT, _Traits, _Distance>& __x, + const istream_iterator<_Tp, _CharT, _Traits, _Distance>& __y) +{ + return !(__x == __y); +} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ITERATOR_ISTREAM_ITERATOR_H diff --git a/lib/libcxx/include/__iterator/istreambuf_iterator.h b/lib/libcxx/include/__iterator/istreambuf_iterator.h new file mode 100644 index 000000000..119698d54 --- /dev/null +++ b/lib/libcxx/include/__iterator/istreambuf_iterator.h @@ -0,0 +1,110 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ITERATOR_ISTREAMBUF_ITERATOR_H +#define _LIBCPP___ITERATOR_ISTREAMBUF_ITERATOR_H + +#include <__config> +#include <__iterator/iterator.h> +#include <__iterator/iterator_traits.h> +#include // for forward declaration of basic_streambuf + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +_LIBCPP_SUPPRESS_DEPRECATED_PUSH +template +class _LIBCPP_TEMPLATE_VIS istreambuf_iterator +#if _LIBCPP_STD_VER <= 14 || !defined(_LIBCPP_ABI_NO_ITERATOR_BASES) + : public iterator +#endif +{ +_LIBCPP_SUPPRESS_DEPRECATED_POP +public: + typedef input_iterator_tag iterator_category; + typedef _CharT value_type; + typedef typename _Traits::off_type difference_type; + typedef _CharT* pointer; + typedef _CharT reference; + typedef _CharT char_type; + typedef _Traits traits_type; + typedef typename _Traits::int_type int_type; + typedef basic_streambuf<_CharT,_Traits> streambuf_type; + typedef basic_istream<_CharT,_Traits> istream_type; +private: + mutable streambuf_type* __sbuf_; + + class __proxy + { + char_type __keep_; + streambuf_type* __sbuf_; + _LIBCPP_INLINE_VISIBILITY __proxy(char_type __c, streambuf_type* __s) + : __keep_(__c), __sbuf_(__s) {} + friend class istreambuf_iterator; + public: + _LIBCPP_INLINE_VISIBILITY char_type operator*() const {return __keep_;} + }; + + _LIBCPP_INLINE_VISIBILITY + bool __test_for_eof() const + { + if (__sbuf_ && traits_type::eq_int_type(__sbuf_->sgetc(), traits_type::eof())) + __sbuf_ = nullptr; + return __sbuf_ == nullptr; + } +public: + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR istreambuf_iterator() _NOEXCEPT : __sbuf_(nullptr) {} + _LIBCPP_INLINE_VISIBILITY istreambuf_iterator(istream_type& __s) _NOEXCEPT + : __sbuf_(__s.rdbuf()) {} + _LIBCPP_INLINE_VISIBILITY istreambuf_iterator(streambuf_type* __s) _NOEXCEPT + : __sbuf_(__s) {} + _LIBCPP_INLINE_VISIBILITY istreambuf_iterator(const __proxy& __p) _NOEXCEPT + : __sbuf_(__p.__sbuf_) {} + + _LIBCPP_INLINE_VISIBILITY char_type operator*() const + {return static_cast(__sbuf_->sgetc());} + _LIBCPP_INLINE_VISIBILITY istreambuf_iterator& operator++() + { + __sbuf_->sbumpc(); + return *this; + } + _LIBCPP_INLINE_VISIBILITY __proxy operator++(int) + { + return __proxy(__sbuf_->sbumpc(), __sbuf_); + } + + _LIBCPP_INLINE_VISIBILITY bool equal(const istreambuf_iterator& __b) const + {return __test_for_eof() == __b.__test_for_eof();} +}; + +template +inline _LIBCPP_INLINE_VISIBILITY +bool operator==(const istreambuf_iterator<_CharT,_Traits>& __a, + const istreambuf_iterator<_CharT,_Traits>& __b) + {return __a.equal(__b);} + +template +inline _LIBCPP_INLINE_VISIBILITY +bool operator!=(const istreambuf_iterator<_CharT,_Traits>& __a, + const istreambuf_iterator<_CharT,_Traits>& __b) + {return !__a.equal(__b);} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ITERATOR_ISTREAMBUF_ITERATOR_H diff --git a/lib/libcxx/include/__iterator/iter_move.h b/lib/libcxx/include/__iterator/iter_move.h new file mode 100644 index 000000000..5540799e1 --- /dev/null +++ b/lib/libcxx/include/__iterator/iter_move.h @@ -0,0 +1,91 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ITERATOR_ITER_MOVE_H +#define _LIBCPP___ITERATOR_ITER_MOVE_H + +#include <__config> +#include <__iterator/iterator_traits.h> +#include <__utility/forward.h> +#include // __class_or_enum +#include +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +#if !defined(_LIBCPP_HAS_NO_RANGES) + +namespace ranges::__iter_move { +void iter_move(); + +template +concept __unqualified_iter_move = requires(_Ip&& __i) { + iter_move(_VSTD::forward<_Ip>(__i)); +}; + +// [iterator.cust.move]/1 +// The name ranges::iter_move denotes a customization point object. +// The expression ranges::iter_move(E) for a subexpression E is +// expression-equivalent to: +struct __fn { + // [iterator.cust.move]/1.1 + // iter_move(E), if E has class or enumeration type and iter_move(E) is a + // well-formed expression when treated as an unevaluated operand, [...] + template + requires __class_or_enum> && __unqualified_iter_move<_Ip> + [[nodiscard]] _LIBCPP_HIDE_FROM_ABI constexpr decltype(auto) operator()(_Ip&& __i) const + noexcept(noexcept(iter_move(_VSTD::forward<_Ip>(__i)))) + { + return iter_move(_VSTD::forward<_Ip>(__i)); + } + + // [iterator.cust.move]/1.2 + // Otherwise, if the expression *E is well-formed: + // 1.2.1 if *E is an lvalue, std::move(*E); + // 1.2.2 otherwise, *E. + template + requires (!(__class_or_enum> && __unqualified_iter_move<_Ip>)) && + requires(_Ip&& __i) { *_VSTD::forward<_Ip>(__i); } + [[nodiscard]] _LIBCPP_HIDE_FROM_ABI constexpr decltype(auto) operator()(_Ip&& __i) const + noexcept(noexcept(*_VSTD::forward<_Ip>(__i))) + { + if constexpr (is_lvalue_reference_v(__i))>) { + return _VSTD::move(*_VSTD::forward<_Ip>(__i)); + } else { + return *_VSTD::forward<_Ip>(__i); + } + } + + // [iterator.cust.move]/1.3 + // Otherwise, ranges::iter_move(E) is ill-formed. +}; +} // namespace ranges::__iter_move + +namespace ranges::inline __cpo { + inline constexpr auto iter_move = __iter_move::__fn{}; +} + +template<__dereferenceable _Tp> +requires requires(_Tp& __t) { { ranges::iter_move(__t) } -> __referenceable; } +using iter_rvalue_reference_t = decltype(ranges::iter_move(declval<_Tp&>())); + +#endif // !_LIBCPP_HAS_NO_RANGES + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ITERATOR_ITER_MOVE_H diff --git a/lib/libcxx/include/__iterator/iter_swap.h b/lib/libcxx/include/__iterator/iter_swap.h new file mode 100644 index 000000000..d70da09b4 --- /dev/null +++ b/lib/libcxx/include/__iterator/iter_swap.h @@ -0,0 +1,107 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// +#ifndef _LIBCPP___ITERATOR_ITER_SWAP_H +#define _LIBCPP___ITERATOR_ITER_SWAP_H + +#include <__config> +#include <__iterator/concepts.h> +#include <__iterator/iter_move.h> +#include <__iterator/iterator_traits.h> +#include <__iterator/readable_traits.h> +#include <__ranges/access.h> +#include +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +#if !defined(_LIBCPP_HAS_NO_RANGES) + +namespace ranges { +namespace __iter_swap { + template + void iter_swap(_I1, _I2) = delete; + + template + concept __unqualified_iter_swap = requires(_T1&& __x, _T2&& __y) { + iter_swap(_VSTD::forward<_T1>(__x), _VSTD::forward<_T2>(__y)); + }; + + template + concept __readable_swappable = + indirectly_readable<_T1> && indirectly_readable<_T2> && + swappable_with, iter_reference_t<_T2>>; + + struct __fn { + template + requires __unqualified_iter_swap<_T1, _T2> + _LIBCPP_HIDE_FROM_ABI + constexpr void operator()(_T1&& __x, _T2&& __y) const + noexcept(noexcept(iter_swap(_VSTD::forward<_T1>(__x), _VSTD::forward<_T2>(__y)))) + { + (void)iter_swap(_VSTD::forward<_T1>(__x), _VSTD::forward<_T2>(__y)); + } + + template + requires (!__unqualified_iter_swap<_T1, _T2>) && + __readable_swappable<_T1, _T2> + _LIBCPP_HIDE_FROM_ABI + constexpr void operator()(_T1&& __x, _T2&& __y) const + noexcept(noexcept(ranges::swap(*_VSTD::forward<_T1>(__x), *_VSTD::forward<_T2>(__y)))) + { + ranges::swap(*_VSTD::forward<_T1>(__x), *_VSTD::forward<_T2>(__y)); + } + + template + requires (!__unqualified_iter_swap<_T1, _T2> && + !__readable_swappable<_T1, _T2>) && + indirectly_movable_storable<_T1, _T2> && + indirectly_movable_storable<_T2, _T1> + _LIBCPP_HIDE_FROM_ABI + constexpr void operator()(_T1&& __x, _T2&& __y) const + noexcept(noexcept(iter_value_t<_T2>(ranges::iter_move(__y))) && + noexcept(*__y = ranges::iter_move(__x)) && + noexcept(*_VSTD::forward<_T1>(__x) = declval>())) + { + iter_value_t<_T2> __old(ranges::iter_move(__y)); + *__y = ranges::iter_move(__x); + *_VSTD::forward<_T1>(__x) = _VSTD::move(__old); + } + }; +} // end namespace __iter_swap + +inline namespace __cpo { + inline constexpr auto iter_swap = __iter_swap::__fn{}; +} // namespace __cpo + +} // namespace ranges + +template +concept indirectly_swappable = + indirectly_readable<_I1> && indirectly_readable<_I2> && + requires(const _I1 __i1, const _I2 __i2) { + ranges::iter_swap(__i1, __i1); + ranges::iter_swap(__i2, __i2); + ranges::iter_swap(__i1, __i2); + ranges::iter_swap(__i2, __i1); + }; + +#endif // !defined(_LIBCPP_HAS_NO_RANGES) + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ITERATOR_ITER_SWAP_H diff --git a/lib/libcxx/include/__iterator/iterator.h b/lib/libcxx/include/__iterator/iterator.h new file mode 100644 index 000000000..dfd481e35 --- /dev/null +++ b/lib/libcxx/include/__iterator/iterator.h @@ -0,0 +1,40 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ITERATOR_ITERATOR_H +#define _LIBCPP___ITERATOR_ITERATOR_H + +#include <__config> +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +struct _LIBCPP_TEMPLATE_VIS _LIBCPP_DEPRECATED_IN_CXX17 iterator +{ + typedef _Tp value_type; + typedef _Distance difference_type; + typedef _Pointer pointer; + typedef _Reference reference; + typedef _Category iterator_category; +}; + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ITERATOR_ITERATOR_H diff --git a/lib/libcxx/include/__iterator/iterator_traits.h b/lib/libcxx/include/__iterator/iterator_traits.h new file mode 100644 index 000000000..5275705c1 --- /dev/null +++ b/lib/libcxx/include/__iterator/iterator_traits.h @@ -0,0 +1,500 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ITERATOR_ITERATOR_TRAITS_H +#define _LIBCPP___ITERATOR_ITERATOR_TRAITS_H + +#include <__config> +#include <__iterator/incrementable_traits.h> +#include <__iterator/readable_traits.h> +#include +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +#if !defined(_LIBCPP_HAS_NO_RANGES) + +template +using __with_reference = _Tp&; + +template +concept __referenceable = requires { + typename __with_reference<_Tp>; +}; + +template +concept __dereferenceable = requires(_Tp& __t) { + { *__t } -> __referenceable; // not required to be equality-preserving +}; + +// [iterator.traits] +template<__dereferenceable _Tp> +using iter_reference_t = decltype(*declval<_Tp&>()); + +#endif // !defined(_LIBCPP_HAS_NO_RANGES) + +template +struct _LIBCPP_TEMPLATE_VIS iterator_traits; + +struct _LIBCPP_TEMPLATE_VIS input_iterator_tag {}; +struct _LIBCPP_TEMPLATE_VIS output_iterator_tag {}; +struct _LIBCPP_TEMPLATE_VIS forward_iterator_tag : public input_iterator_tag {}; +struct _LIBCPP_TEMPLATE_VIS bidirectional_iterator_tag : public forward_iterator_tag {}; +struct _LIBCPP_TEMPLATE_VIS random_access_iterator_tag : public bidirectional_iterator_tag {}; +#if _LIBCPP_STD_VER > 17 +struct _LIBCPP_TEMPLATE_VIS contiguous_iterator_tag : public random_access_iterator_tag {}; +#endif + +template +struct __iter_traits_cache { + using type = _If< + __is_primary_template >::value, + _Iter, + iterator_traits<_Iter> + >; +}; +template +using _ITER_TRAITS = typename __iter_traits_cache<_Iter>::type; + +struct __iter_concept_concept_test { + template + using _Apply = typename _ITER_TRAITS<_Iter>::iterator_concept; +}; +struct __iter_concept_category_test { + template + using _Apply = typename _ITER_TRAITS<_Iter>::iterator_category; +}; +struct __iter_concept_random_fallback { + template + using _Apply = _EnableIf< + __is_primary_template >::value, + random_access_iterator_tag + >; +}; + +template struct __test_iter_concept + : _IsValidExpansion<_Tester::template _Apply, _Iter>, + _Tester +{ +}; + +template +struct __iter_concept_cache { + using type = _Or< + __test_iter_concept<_Iter, __iter_concept_concept_test>, + __test_iter_concept<_Iter, __iter_concept_category_test>, + __test_iter_concept<_Iter, __iter_concept_random_fallback> + >; +}; + +template +using _ITER_CONCEPT = typename __iter_concept_cache<_Iter>::type::template _Apply<_Iter>; + + +template +struct __has_iterator_typedefs +{ +private: + struct __two {char __lx; char __lxx;}; + template static __two __test(...); + template static char __test(typename __void_t::type* = 0, + typename __void_t::type* = 0, + typename __void_t::type* = 0, + typename __void_t::type* = 0, + typename __void_t::type* = 0); +public: + static const bool value = sizeof(__test<_Tp>(0,0,0,0,0)) == 1; +}; + + +template +struct __has_iterator_category +{ +private: + struct __two {char __lx; char __lxx;}; + template static __two __test(...); + template static char __test(typename _Up::iterator_category* = nullptr); +public: + static const bool value = sizeof(__test<_Tp>(nullptr)) == 1; +}; + +template +struct __has_iterator_concept +{ +private: + struct __two {char __lx; char __lxx;}; + template static __two __test(...); + template static char __test(typename _Up::iterator_concept* = nullptr); +public: + static const bool value = sizeof(__test<_Tp>(nullptr)) == 1; +}; + +#if !defined(_LIBCPP_HAS_NO_RANGES) + +// The `cpp17-*-iterator` exposition-only concepts are easily confused with the Cpp17*Iterator tables, +// so they've been banished to a namespace that makes it obvious they have a niche use-case. +namespace __iterator_traits_detail { +template +concept __cpp17_iterator = + requires(_Ip __i) { + { *__i } -> __referenceable; + { ++__i } -> same_as<_Ip&>; + { *__i++ } -> __referenceable; + } && + copyable<_Ip>; + +template +concept __cpp17_input_iterator = + __cpp17_iterator<_Ip> && + equality_comparable<_Ip> && + requires(_Ip __i) { + typename incrementable_traits<_Ip>::difference_type; + typename indirectly_readable_traits<_Ip>::value_type; + typename common_reference_t&&, + typename indirectly_readable_traits<_Ip>::value_type&>; + typename common_reference_t::value_type&>; + requires signed_integral::difference_type>; + }; + +template +concept __cpp17_forward_iterator = + __cpp17_input_iterator<_Ip> && + constructible_from<_Ip> && + is_lvalue_reference_v> && + same_as>, + typename indirectly_readable_traits<_Ip>::value_type> && + requires(_Ip __i) { + { __i++ } -> convertible_to<_Ip const&>; + { *__i++ } -> same_as>; + }; + +template +concept __cpp17_bidirectional_iterator = + __cpp17_forward_iterator<_Ip> && + requires(_Ip __i) { + { --__i } -> same_as<_Ip&>; + { __i-- } -> convertible_to<_Ip const&>; + { *__i-- } -> same_as>; + }; + +template +concept __cpp17_random_access_iterator = + __cpp17_bidirectional_iterator<_Ip> && + totally_ordered<_Ip> && + requires(_Ip __i, typename incrementable_traits<_Ip>::difference_type __n) { + { __i += __n } -> same_as<_Ip&>; + { __i -= __n } -> same_as<_Ip&>; + { __i + __n } -> same_as<_Ip>; + { __n + __i } -> same_as<_Ip>; + { __i - __n } -> same_as<_Ip>; + { __i - __i } -> same_as; + { __i[__n] } -> convertible_to>; + }; +} // namespace __iterator_traits_detail + +template +concept __has_member_reference = requires { typename _Ip::reference; }; + +template +concept __has_member_pointer = requires { typename _Ip::pointer; }; + +template +concept __has_member_iterator_category = requires { typename _Ip::iterator_category; }; + +template +concept __specifies_members = requires { + typename _Ip::value_type; + typename _Ip::difference_type; + requires __has_member_reference<_Ip>; + requires __has_member_iterator_category<_Ip>; + }; + +template +struct __iterator_traits_member_pointer_or_void { + using type = void; +}; + +template<__has_member_pointer _Tp> +struct __iterator_traits_member_pointer_or_void<_Tp> { + using type = typename _Tp::pointer; +}; + +template +concept __cpp17_iterator_missing_members = + !__specifies_members<_Tp> && + __iterator_traits_detail::__cpp17_iterator<_Tp>; + +template +concept __cpp17_input_iterator_missing_members = + __cpp17_iterator_missing_members<_Tp> && + __iterator_traits_detail::__cpp17_input_iterator<_Tp>; + +// Otherwise, `pointer` names `void`. +template +struct __iterator_traits_member_pointer_or_arrow_or_void { using type = void; }; + +// [iterator.traits]/3.2.1 +// If the qualified-id `I::pointer` is valid and denotes a type, `pointer` names that type. +template<__has_member_pointer _Ip> +struct __iterator_traits_member_pointer_or_arrow_or_void<_Ip> { using type = typename _Ip::pointer; }; + +// Otherwise, if `decltype(declval().operator->())` is well-formed, then `pointer` names that +// type. +template + requires requires(_Ip& __i) { __i.operator->(); } && (!__has_member_pointer<_Ip>) +struct __iterator_traits_member_pointer_or_arrow_or_void<_Ip> { + using type = decltype(declval<_Ip&>().operator->()); +}; + +// Otherwise, `reference` names `iter-reference-t`. +template +struct __iterator_traits_member_reference { using type = iter_reference_t<_Ip>; }; + +// [iterator.traits]/3.2.2 +// If the qualified-id `I::reference` is valid and denotes a type, `reference` names that type. +template<__has_member_reference _Ip> +struct __iterator_traits_member_reference<_Ip> { using type = typename _Ip::reference; }; + +// [iterator.traits]/3.2.3.4 +// input_iterator_tag +template +struct __deduce_iterator_category { + using type = input_iterator_tag; +}; + +// [iterator.traits]/3.2.3.1 +// `random_access_iterator_tag` if `I` satisfies `cpp17-random-access-iterator`, or otherwise +template<__iterator_traits_detail::__cpp17_random_access_iterator _Ip> +struct __deduce_iterator_category<_Ip> { + using type = random_access_iterator_tag; +}; + +// [iterator.traits]/3.2.3.2 +// `bidirectional_iterator_tag` if `I` satisfies `cpp17-bidirectional-iterator`, or otherwise +template<__iterator_traits_detail::__cpp17_bidirectional_iterator _Ip> +struct __deduce_iterator_category<_Ip> { + using type = bidirectional_iterator_tag; +}; + +// [iterator.traits]/3.2.3.3 +// `forward_iterator_tag` if `I` satisfies `cpp17-forward-iterator`, or otherwise +template<__iterator_traits_detail::__cpp17_forward_iterator _Ip> +struct __deduce_iterator_category<_Ip> { + using type = forward_iterator_tag; +}; + +template +struct __iterator_traits_iterator_category : __deduce_iterator_category<_Ip> {}; + +// [iterator.traits]/3.2.3 +// If the qualified-id `I::iterator-category` is valid and denotes a type, `iterator-category` names +// that type. +template<__has_member_iterator_category _Ip> +struct __iterator_traits_iterator_category<_Ip> { + using type = typename _Ip::iterator_category; +}; + +// otherwise, it names void. +template +struct __iterator_traits_difference_type { using type = void; }; + +// If the qualified-id `incrementable_traits::difference_type` is valid and denotes a type, then +// `difference_type` names that type; +template +requires requires { typename incrementable_traits<_Ip>::difference_type; } +struct __iterator_traits_difference_type<_Ip> { + using type = typename incrementable_traits<_Ip>::difference_type; +}; + +// [iterator.traits]/3.4 +// Otherwise, `iterator_traits` has no members by any of the above names. +template +struct __iterator_traits {}; + +// [iterator.traits]/3.1 +// If `I` has valid ([temp.deduct]) member types `difference-type`, `value-type`, `reference`, and +// `iterator-category`, then `iterator-traits` has the following publicly accessible members: +template<__specifies_members _Ip> +struct __iterator_traits<_Ip> { + using iterator_category = typename _Ip::iterator_category; + using value_type = typename _Ip::value_type; + using difference_type = typename _Ip::difference_type; + using pointer = typename __iterator_traits_member_pointer_or_void<_Ip>::type; + using reference = typename _Ip::reference; +}; + +// [iterator.traits]/3.2 +// Otherwise, if `I` satisfies the exposition-only concept `cpp17-input-iterator`, +// `iterator-traits` has the following publicly accessible members: +template<__cpp17_input_iterator_missing_members _Ip> +struct __iterator_traits<_Ip> { + using iterator_category = typename __iterator_traits_iterator_category<_Ip>::type; + using value_type = typename indirectly_readable_traits<_Ip>::value_type; + using difference_type = typename incrementable_traits<_Ip>::difference_type; + using pointer = typename __iterator_traits_member_pointer_or_arrow_or_void<_Ip>::type; + using reference = typename __iterator_traits_member_reference<_Ip>::type; +}; + +// Otherwise, if `I` satisfies the exposition-only concept `cpp17-iterator`, then +// `iterator_traits` has the following publicly accessible members: +template<__cpp17_iterator_missing_members _Ip> +struct __iterator_traits<_Ip> { + using iterator_category = output_iterator_tag; + using value_type = void; + using difference_type = typename __iterator_traits_difference_type<_Ip>::type; + using pointer = void; + using reference = void; +}; + +template +struct iterator_traits : __iterator_traits<_Ip> { + using __primary_template = iterator_traits; +}; + +#else // !defined(_LIBCPP_HAS_NO_RANGES) + +template struct __iterator_traits {}; + +template struct __iterator_traits_impl {}; + +template +struct __iterator_traits_impl<_Iter, true> +{ + typedef typename _Iter::difference_type difference_type; + typedef typename _Iter::value_type value_type; + typedef typename _Iter::pointer pointer; + typedef typename _Iter::reference reference; + typedef typename _Iter::iterator_category iterator_category; +}; + +template +struct __iterator_traits<_Iter, true> + : __iterator_traits_impl + < + _Iter, + is_convertible::value || + is_convertible::value + > +{}; + +// iterator_traits will only have the nested types if Iterator::iterator_category +// exists. Else iterator_traits will be an empty class. This is a +// conforming extension which allows some programs to compile and behave as +// the client expects instead of failing at compile time. + +template +struct _LIBCPP_TEMPLATE_VIS iterator_traits + : __iterator_traits<_Iter, __has_iterator_typedefs<_Iter>::value> { + + using __primary_template = iterator_traits; +}; +#endif // !defined(_LIBCPP_HAS_NO_RANGES) + +template +#if !defined(_LIBCPP_HAS_NO_RANGES) +requires is_object_v<_Tp> +#endif +struct _LIBCPP_TEMPLATE_VIS iterator_traits<_Tp*> +{ + typedef ptrdiff_t difference_type; + typedef typename remove_cv<_Tp>::type value_type; + typedef _Tp* pointer; + typedef _Tp& reference; + typedef random_access_iterator_tag iterator_category; +#if _LIBCPP_STD_VER > 17 + typedef contiguous_iterator_tag iterator_concept; +#endif +}; + +template >::value> +struct __has_iterator_category_convertible_to + : is_convertible::iterator_category, _Up> +{}; + +template +struct __has_iterator_category_convertible_to<_Tp, _Up, false> : false_type {}; + +template ::value> +struct __has_iterator_concept_convertible_to + : is_convertible +{}; + +template +struct __has_iterator_concept_convertible_to<_Tp, _Up, false> : false_type {}; + +template +struct __is_cpp17_input_iterator : public __has_iterator_category_convertible_to<_Tp, input_iterator_tag> {}; + +template +struct __is_cpp17_forward_iterator : public __has_iterator_category_convertible_to<_Tp, forward_iterator_tag> {}; + +template +struct __is_cpp17_bidirectional_iterator : public __has_iterator_category_convertible_to<_Tp, bidirectional_iterator_tag> {}; + +template +struct __is_cpp17_random_access_iterator : public __has_iterator_category_convertible_to<_Tp, random_access_iterator_tag> {}; + +// __is_cpp17_contiguous_iterator determines if an iterator is known by +// libc++ to be contiguous, either because it advertises itself as such +// (in C++20) or because it is a pointer type or a known trivial wrapper +// around a (possibly fancy) pointer type, such as __wrap_iter. +// Such iterators receive special "contiguous" optimizations in +// std::copy and std::sort. +// +#if _LIBCPP_STD_VER > 17 +template +struct __is_cpp17_contiguous_iterator : _Or< + __has_iterator_category_convertible_to<_Tp, contiguous_iterator_tag>, + __has_iterator_concept_convertible_to<_Tp, contiguous_iterator_tag> +> {}; +#else +template +struct __is_cpp17_contiguous_iterator : false_type {}; +#endif + +// Any native pointer which is an iterator is also a contiguous iterator. +template +struct __is_cpp17_contiguous_iterator<_Up*> : true_type {}; + + +template +struct __is_exactly_cpp17_input_iterator + : public integral_constant::value && + !__has_iterator_category_convertible_to<_Tp, forward_iterator_tag>::value> {}; + +#ifndef _LIBCPP_HAS_NO_DEDUCTION_GUIDES +template +using __iter_value_type = typename iterator_traits<_InputIterator>::value_type; + +template +using __iter_key_type = remove_const_t::value_type::first_type>; + +template +using __iter_mapped_type = typename iterator_traits<_InputIterator>::value_type::second_type; + +template +using __iter_to_alloc_type = pair< + add_const_t::value_type::first_type>, + typename iterator_traits<_InputIterator>::value_type::second_type>; +#endif // _LIBCPP_HAS_NO_DEDUCTION_GUIDES + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ITERATOR_ITERATOR_TRAITS_H diff --git a/lib/libcxx/include/__iterator/move_iterator.h b/lib/libcxx/include/__iterator/move_iterator.h new file mode 100644 index 000000000..7819743bd --- /dev/null +++ b/lib/libcxx/include/__iterator/move_iterator.h @@ -0,0 +1,189 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ITERATOR_MOVE_ITERATOR_H +#define _LIBCPP___ITERATOR_MOVE_ITERATOR_H + +#include <__config> +#include <__iterator/iterator_traits.h> +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +class _LIBCPP_TEMPLATE_VIS move_iterator +{ +private: + _Iter __i; +public: + typedef _Iter iterator_type; + typedef typename iterator_traits::value_type value_type; + typedef typename iterator_traits::difference_type difference_type; + typedef iterator_type pointer; + typedef _If<__is_cpp17_random_access_iterator<_Iter>::value, + random_access_iterator_tag, + typename iterator_traits<_Iter>::iterator_category> iterator_category; +#if _LIBCPP_STD_VER > 17 + typedef input_iterator_tag iterator_concept; +#endif + +#ifndef _LIBCPP_CXX03_LANG + typedef typename iterator_traits::reference __reference; + typedef typename conditional< + is_reference<__reference>::value, + typename remove_reference<__reference>::type&&, + __reference + >::type reference; +#else + typedef typename iterator_traits::reference reference; +#endif + + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 + move_iterator() : __i() {} + + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 + explicit move_iterator(_Iter __x) : __i(__x) {} + + template ::value && is_convertible<_Up const&, _Iter>::value + > > + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 + move_iterator(const move_iterator<_Up>& __u) : __i(__u.base()) {} + + template ::value && + is_convertible<_Up const&, _Iter>::value && + is_assignable<_Iter&, _Up const&>::value + > > + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 + move_iterator& operator=(const move_iterator<_Up>& __u) { + __i = __u.base(); + return *this; + } + + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 _Iter base() const {return __i;} + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 + reference operator*() const { return static_cast(*__i); } + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 + pointer operator->() const { return __i;} + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 + move_iterator& operator++() {++__i; return *this;} + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 + move_iterator operator++(int) {move_iterator __tmp(*this); ++__i; return __tmp;} + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 + move_iterator& operator--() {--__i; return *this;} + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 + move_iterator operator--(int) {move_iterator __tmp(*this); --__i; return __tmp;} + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 + move_iterator operator+ (difference_type __n) const {return move_iterator(__i + __n);} + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 + move_iterator& operator+=(difference_type __n) {__i += __n; return *this;} + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 + move_iterator operator- (difference_type __n) const {return move_iterator(__i - __n);} + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 + move_iterator& operator-=(difference_type __n) {__i -= __n; return *this;} + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 + reference operator[](difference_type __n) const { return static_cast(__i[__n]); } +}; + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 +bool +operator==(const move_iterator<_Iter1>& __x, const move_iterator<_Iter2>& __y) +{ + return __x.base() == __y.base(); +} + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 +bool +operator<(const move_iterator<_Iter1>& __x, const move_iterator<_Iter2>& __y) +{ + return __x.base() < __y.base(); +} + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 +bool +operator!=(const move_iterator<_Iter1>& __x, const move_iterator<_Iter2>& __y) +{ + return __x.base() != __y.base(); +} + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 +bool +operator>(const move_iterator<_Iter1>& __x, const move_iterator<_Iter2>& __y) +{ + return __x.base() > __y.base(); +} + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 +bool +operator>=(const move_iterator<_Iter1>& __x, const move_iterator<_Iter2>& __y) +{ + return __x.base() >= __y.base(); +} + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 +bool +operator<=(const move_iterator<_Iter1>& __x, const move_iterator<_Iter2>& __y) +{ + return __x.base() <= __y.base(); +} + +#ifndef _LIBCPP_CXX03_LANG +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 +auto +operator-(const move_iterator<_Iter1>& __x, const move_iterator<_Iter2>& __y) +-> decltype(__x.base() - __y.base()) +{ + return __x.base() - __y.base(); +} +#else +template +inline _LIBCPP_INLINE_VISIBILITY +typename move_iterator<_Iter1>::difference_type +operator-(const move_iterator<_Iter1>& __x, const move_iterator<_Iter2>& __y) +{ + return __x.base() - __y.base(); +} +#endif + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 +move_iterator<_Iter> +operator+(typename move_iterator<_Iter>::difference_type __n, const move_iterator<_Iter>& __x) +{ + return move_iterator<_Iter>(__x.base() + __n); +} + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 +move_iterator<_Iter> +make_move_iterator(_Iter __i) +{ + return move_iterator<_Iter>(__i); +} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ITERATOR_MOVE_ITERATOR_H diff --git a/lib/libcxx/include/__iterator/next.h b/lib/libcxx/include/__iterator/next.h new file mode 100644 index 000000000..1eecaa975 --- /dev/null +++ b/lib/libcxx/include/__iterator/next.h @@ -0,0 +1,87 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ITERATOR_NEXT_H +#define _LIBCPP___ITERATOR_NEXT_H + +#include <__config> +#include <__debug> +#include <__function_like.h> +#include <__iterator/advance.h> +#include <__iterator/concepts.h> +#include <__iterator/incrementable_traits.h> +#include <__iterator/iterator_traits.h> +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 + typename enable_if<__is_cpp17_input_iterator<_InputIter>::value, _InputIter>::type + next(_InputIter __x, typename iterator_traits<_InputIter>::difference_type __n = 1) { + _LIBCPP_ASSERT(__n >= 0 || __is_cpp17_bidirectional_iterator<_InputIter>::value, + "Attempt to next(it, n) with negative n on a non-bidirectional iterator"); + + _VSTD::advance(__x, __n); + return __x; +} + +#if !defined(_LIBCPP_HAS_NO_RANGES) + +namespace ranges { +struct __next_fn final : private __function_like { + _LIBCPP_HIDE_FROM_ABI + constexpr explicit __next_fn(__tag __x) noexcept : __function_like(__x) {} + + template + _LIBCPP_HIDE_FROM_ABI + constexpr _Ip operator()(_Ip __x) const { + ++__x; + return __x; + } + + template + _LIBCPP_HIDE_FROM_ABI + constexpr _Ip operator()(_Ip __x, iter_difference_t<_Ip> __n) const { + ranges::advance(__x, __n); + return __x; + } + + template _Sp> + _LIBCPP_HIDE_FROM_ABI + constexpr _Ip operator()(_Ip __x, _Sp __bound) const { + ranges::advance(__x, __bound); + return __x; + } + + template _Sp> + _LIBCPP_HIDE_FROM_ABI + constexpr _Ip operator()(_Ip __x, iter_difference_t<_Ip> __n, _Sp __bound) const { + ranges::advance(__x, __n, __bound); + return __x; + } +}; + +inline constexpr auto next = __next_fn(__function_like::__tag()); +} // namespace ranges + +#endif // !defined(_LIBCPP_HAS_NO_RANGES) + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ITERATOR_PRIMITIVES_H diff --git a/lib/libcxx/include/__iterator/ostream_iterator.h b/lib/libcxx/include/__iterator/ostream_iterator.h new file mode 100644 index 000000000..5b4466c86 --- /dev/null +++ b/lib/libcxx/include/__iterator/ostream_iterator.h @@ -0,0 +1,75 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ITERATOR_OSTREAM_ITERATOR_H +#define _LIBCPP___ITERATOR_OSTREAM_ITERATOR_H + +#include <__config> +#include <__iterator/iterator.h> +#include <__iterator/iterator_traits.h> +#include <__memory/addressof.h> +#include // for forward declarations of char_traits and basic_ostream + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +_LIBCPP_SUPPRESS_DEPRECATED_PUSH +template > +class _LIBCPP_TEMPLATE_VIS ostream_iterator +#if _LIBCPP_STD_VER <= 14 || !defined(_LIBCPP_ABI_NO_ITERATOR_BASES) + : public iterator +#endif +{ +_LIBCPP_SUPPRESS_DEPRECATED_POP +public: + typedef output_iterator_tag iterator_category; + typedef void value_type; +#if _LIBCPP_STD_VER > 17 + typedef ptrdiff_t difference_type; +#else + typedef void difference_type; +#endif + typedef void pointer; + typedef void reference; + typedef _CharT char_type; + typedef _Traits traits_type; + typedef basic_ostream<_CharT, _Traits> ostream_type; + +private: + ostream_type* __out_stream_; + const char_type* __delim_; +public: + _LIBCPP_INLINE_VISIBILITY ostream_iterator(ostream_type& __s) _NOEXCEPT + : __out_stream_(_VSTD::addressof(__s)), __delim_(nullptr) {} + _LIBCPP_INLINE_VISIBILITY ostream_iterator(ostream_type& __s, const _CharT* __delimiter) _NOEXCEPT + : __out_stream_(_VSTD::addressof(__s)), __delim_(__delimiter) {} + _LIBCPP_INLINE_VISIBILITY ostream_iterator& operator=(const _Tp& __value_) + { + *__out_stream_ << __value_; + if (__delim_) + *__out_stream_ << __delim_; + return *this; + } + + _LIBCPP_INLINE_VISIBILITY ostream_iterator& operator*() {return *this;} + _LIBCPP_INLINE_VISIBILITY ostream_iterator& operator++() {return *this;} + _LIBCPP_INLINE_VISIBILITY ostream_iterator& operator++(int) {return *this;} +}; + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ITERATOR_OSTREAM_ITERATOR_H diff --git a/lib/libcxx/include/__iterator/ostreambuf_iterator.h b/lib/libcxx/include/__iterator/ostreambuf_iterator.h new file mode 100644 index 000000000..90309dacd --- /dev/null +++ b/lib/libcxx/include/__iterator/ostreambuf_iterator.h @@ -0,0 +1,81 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ITERATOR_OSTREAMBUF_ITERATOR_H +#define _LIBCPP___ITERATOR_OSTREAMBUF_ITERATOR_H + +#include <__config> +#include <__iterator/iterator.h> +#include <__iterator/iterator_traits.h> +#include // for forward declaration of basic_streambuf + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +_LIBCPP_SUPPRESS_DEPRECATED_PUSH +template +class _LIBCPP_TEMPLATE_VIS ostreambuf_iterator +#if _LIBCPP_STD_VER <= 14 || !defined(_LIBCPP_ABI_NO_ITERATOR_BASES) + : public iterator +#endif +{ +_LIBCPP_SUPPRESS_DEPRECATED_POP +public: + typedef output_iterator_tag iterator_category; + typedef void value_type; +#if _LIBCPP_STD_VER > 17 + typedef ptrdiff_t difference_type; +#else + typedef void difference_type; +#endif + typedef void pointer; + typedef void reference; + typedef _CharT char_type; + typedef _Traits traits_type; + typedef basic_streambuf<_CharT, _Traits> streambuf_type; + typedef basic_ostream<_CharT, _Traits> ostream_type; + +private: + streambuf_type* __sbuf_; +public: + _LIBCPP_INLINE_VISIBILITY ostreambuf_iterator(ostream_type& __s) _NOEXCEPT + : __sbuf_(__s.rdbuf()) {} + _LIBCPP_INLINE_VISIBILITY ostreambuf_iterator(streambuf_type* __s) _NOEXCEPT + : __sbuf_(__s) {} + _LIBCPP_INLINE_VISIBILITY ostreambuf_iterator& operator=(_CharT __c) + { + if (__sbuf_ && traits_type::eq_int_type(__sbuf_->sputc(__c), traits_type::eof())) + __sbuf_ = nullptr; + return *this; + } + _LIBCPP_INLINE_VISIBILITY ostreambuf_iterator& operator*() {return *this;} + _LIBCPP_INLINE_VISIBILITY ostreambuf_iterator& operator++() {return *this;} + _LIBCPP_INLINE_VISIBILITY ostreambuf_iterator& operator++(int) {return *this;} + _LIBCPP_INLINE_VISIBILITY bool failed() const _NOEXCEPT {return __sbuf_ == nullptr;} + + template + friend + _LIBCPP_HIDDEN + ostreambuf_iterator<_Ch, _Tr> + __pad_and_output(ostreambuf_iterator<_Ch, _Tr> __s, + const _Ch* __ob, const _Ch* __op, const _Ch* __oe, + ios_base& __iob, _Ch __fl); +}; + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ITERATOR_OSTREAMBUF_ITERATOR_H diff --git a/lib/libcxx/include/__iterator/prev.h b/lib/libcxx/include/__iterator/prev.h new file mode 100644 index 000000000..cb8a57135 --- /dev/null +++ b/lib/libcxx/include/__iterator/prev.h @@ -0,0 +1,79 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ITERATOR_PREV_H +#define _LIBCPP___ITERATOR_PREV_H + +#include <__config> +#include <__debug> +#include <__function_like.h> +#include <__iterator/advance.h> +#include <__iterator/concepts.h> +#include <__iterator/incrementable_traits.h> +#include <__iterator/iterator_traits.h> +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 + typename enable_if<__is_cpp17_input_iterator<_InputIter>::value, _InputIter>::type + prev(_InputIter __x, typename iterator_traits<_InputIter>::difference_type __n = 1) { + _LIBCPP_ASSERT(__n <= 0 || __is_cpp17_bidirectional_iterator<_InputIter>::value, + "Attempt to prev(it, n) with a positive n on a non-bidirectional iterator"); + _VSTD::advance(__x, -__n); + return __x; +} + +#if !defined(_LIBCPP_HAS_NO_RANGES) + +namespace ranges { +struct __prev_fn final : private __function_like { + _LIBCPP_HIDE_FROM_ABI + constexpr explicit __prev_fn(__tag __x) noexcept : __function_like(__x) {} + + template + _LIBCPP_HIDE_FROM_ABI + constexpr _Ip operator()(_Ip __x) const { + --__x; + return __x; + } + + template + _LIBCPP_HIDE_FROM_ABI + constexpr _Ip operator()(_Ip __x, iter_difference_t<_Ip> __n) const { + ranges::advance(__x, -__n); + return __x; + } + + template + _LIBCPP_HIDE_FROM_ABI + constexpr _Ip operator()(_Ip __x, iter_difference_t<_Ip> __n, _Ip __bound) const { + ranges::advance(__x, -__n, __bound); + return __x; + } +}; + +inline constexpr auto prev = __prev_fn(__function_like::__tag()); +} // namespace ranges + +#endif // !defined(_LIBCPP_HAS_NO_RANGES) + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ITERATOR_PREV_H diff --git a/lib/libcxx/include/__iterator/projected.h b/lib/libcxx/include/__iterator/projected.h new file mode 100644 index 000000000..7064a5eb9 --- /dev/null +++ b/lib/libcxx/include/__iterator/projected.h @@ -0,0 +1,45 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// +#ifndef _LIBCPP___ITERATOR_PROJECTED_H +#define _LIBCPP___ITERATOR_PROJECTED_H + +#include <__config> +#include <__iterator/concepts.h> +#include <__iterator/incrementable_traits.h> +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +#if !defined(_LIBCPP_HAS_NO_RANGES) + +template _Proj> +struct projected { + using value_type = remove_cvref_t>; + indirect_result_t<_Proj&, _It> operator*() const; // not defined +}; + +template +struct incrementable_traits> { + using difference_type = iter_difference_t<_It>; +}; + +#endif // !defined(_LIBCPP_HAS_NO_RANGES) + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ITERATOR_PROJECTED_H diff --git a/lib/libcxx/include/__iterator/readable_traits.h b/lib/libcxx/include/__iterator/readable_traits.h new file mode 100644 index 000000000..fbad106e4 --- /dev/null +++ b/lib/libcxx/include/__iterator/readable_traits.h @@ -0,0 +1,91 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ITERATOR_READABLE_TRAITS_H +#define _LIBCPP___ITERATOR_READABLE_TRAITS_H + +#include <__config> +#include +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +#if !defined(_LIBCPP_HAS_NO_RANGES) + +// [readable.traits] +template struct __cond_value_type {}; + +template +requires is_object_v<_Tp> +struct __cond_value_type<_Tp> { using value_type = remove_cv_t<_Tp>; }; + +template +concept __has_member_value_type = requires { typename _Tp::value_type; }; + +template +concept __has_member_element_type = requires { typename _Tp::element_type; }; + +template struct indirectly_readable_traits {}; + +template +requires is_array_v<_Ip> +struct indirectly_readable_traits<_Ip> { + using value_type = remove_cv_t>; +}; + +template +struct indirectly_readable_traits : indirectly_readable_traits<_Ip> {}; + +template +struct indirectly_readable_traits<_Tp*> : __cond_value_type<_Tp> {}; + +template<__has_member_value_type _Tp> +struct indirectly_readable_traits<_Tp> + : __cond_value_type {}; + +template<__has_member_element_type _Tp> +struct indirectly_readable_traits<_Tp> + : __cond_value_type {}; + +// Pre-emptively applies LWG3541 +template<__has_member_value_type _Tp> +requires __has_member_element_type<_Tp> +struct indirectly_readable_traits<_Tp> {}; +template<__has_member_value_type _Tp> +requires __has_member_element_type<_Tp> && + same_as, + remove_cv_t> +struct indirectly_readable_traits<_Tp> + : __cond_value_type {}; + +template +struct iterator_traits; + +// Let `RI` be `remove_cvref_t`. The type `iter_value_t` denotes +// `indirectly_readable_traits::value_type` if `iterator_traits` names a specialization +// generated from the primary template, and `iterator_traits::value_type` otherwise. +template +using iter_value_t = typename conditional_t<__is_primary_template > >::value, + indirectly_readable_traits >, + iterator_traits > >::value_type; + +#endif // !defined(_LIBCPP_HAS_NO_RANGES) + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ITERATOR_READABLE_TRAITS_H diff --git a/lib/libcxx/include/__iterator/reverse_access.h b/lib/libcxx/include/__iterator/reverse_access.h new file mode 100644 index 000000000..66cc3568c --- /dev/null +++ b/lib/libcxx/include/__iterator/reverse_access.h @@ -0,0 +1,109 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ITERATOR_REVERSE_ACCESS_H +#define _LIBCPP___ITERATOR_REVERSE_ACCESS_H + +#include <__config> +#include <__iterator/reverse_iterator.h> +#include +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +#if !defined(_LIBCPP_CXX03_LANG) + +#if _LIBCPP_STD_VER > 11 + +template +_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 +reverse_iterator<_Tp*> rbegin(_Tp (&__array)[_Np]) +{ + return reverse_iterator<_Tp*>(__array + _Np); +} + +template +_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 +reverse_iterator<_Tp*> rend(_Tp (&__array)[_Np]) +{ + return reverse_iterator<_Tp*>(__array); +} + +template +_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 +reverse_iterator rbegin(initializer_list<_Ep> __il) +{ + return reverse_iterator(__il.end()); +} + +template +_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 +reverse_iterator rend(initializer_list<_Ep> __il) +{ + return reverse_iterator(__il.begin()); +} + +template +_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 +auto rbegin(_Cp& __c) -> decltype(__c.rbegin()) +{ + return __c.rbegin(); +} + +template +_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 +auto rbegin(const _Cp& __c) -> decltype(__c.rbegin()) +{ + return __c.rbegin(); +} + +template +_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 +auto rend(_Cp& __c) -> decltype(__c.rend()) +{ + return __c.rend(); +} + +template +_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 +auto rend(const _Cp& __c) -> decltype(__c.rend()) +{ + return __c.rend(); +} + +template +_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 +auto crbegin(const _Cp& __c) -> decltype(_VSTD::rbegin(__c)) +{ + return _VSTD::rbegin(__c); +} + +template +_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 +auto crend(const _Cp& __c) -> decltype(_VSTD::rend(__c)) +{ + return _VSTD::rend(__c); +} + +#endif + +#endif // !defined(_LIBCPP_CXX03_LANG) + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ITERATOR_REVERSE_ACCESS_H diff --git a/lib/libcxx/include/__iterator/reverse_iterator.h b/lib/libcxx/include/__iterator/reverse_iterator.h new file mode 100644 index 000000000..76424a89a --- /dev/null +++ b/lib/libcxx/include/__iterator/reverse_iterator.h @@ -0,0 +1,239 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ITERATOR_REVERSE_ITERATOR_H +#define _LIBCPP___ITERATOR_REVERSE_ITERATOR_H + +#include <__config> +#include <__iterator/iterator.h> +#include <__iterator/iterator_traits.h> +#include <__memory/addressof.h> +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +struct __is_stashing_iterator : false_type {}; + +template +struct __is_stashing_iterator<_Tp, typename __void_t::type> + : true_type {}; + +_LIBCPP_SUPPRESS_DEPRECATED_PUSH +template +class _LIBCPP_TEMPLATE_VIS reverse_iterator +#if _LIBCPP_STD_VER <= 14 || !defined(_LIBCPP_ABI_NO_ITERATOR_BASES) + : public iterator::iterator_category, + typename iterator_traits<_Iter>::value_type, + typename iterator_traits<_Iter>::difference_type, + typename iterator_traits<_Iter>::pointer, + typename iterator_traits<_Iter>::reference> +#endif +{ +_LIBCPP_SUPPRESS_DEPRECATED_POP +private: +#ifndef _LIBCPP_ABI_NO_ITERATOR_BASES + _Iter __t; // no longer used as of LWG #2360, not removed due to ABI break +#endif + + static_assert(!__is_stashing_iterator<_Iter>::value, + "The specified iterator type cannot be used with reverse_iterator; " + "Using stashing iterators with reverse_iterator causes undefined behavior"); + +protected: + _Iter current; +public: + typedef _Iter iterator_type; + typedef typename iterator_traits<_Iter>::difference_type difference_type; + typedef typename iterator_traits<_Iter>::reference reference; + typedef typename iterator_traits<_Iter>::pointer pointer; + typedef _If<__is_cpp17_random_access_iterator<_Iter>::value, + random_access_iterator_tag, + typename iterator_traits<_Iter>::iterator_category> iterator_category; + typedef typename iterator_traits<_Iter>::value_type value_type; + +#if _LIBCPP_STD_VER > 17 + typedef _If<__is_cpp17_random_access_iterator<_Iter>::value, + random_access_iterator_tag, + bidirectional_iterator_tag> iterator_concept; +#endif + +#ifndef _LIBCPP_ABI_NO_ITERATOR_BASES + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 + reverse_iterator() : __t(), current() {} + + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 + explicit reverse_iterator(_Iter __x) : __t(__x), current(__x) {} + + template ::value && is_convertible<_Up const&, _Iter>::value + > > + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 + reverse_iterator(const reverse_iterator<_Up>& __u) + : __t(__u.base()), current(__u.base()) + { } + + template ::value && + is_convertible<_Up const&, _Iter>::value && + is_assignable<_Up const&, _Iter>::value + > > + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 + reverse_iterator& operator=(const reverse_iterator<_Up>& __u) { + __t = current = __u.base(); + return *this; + } +#else + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 + reverse_iterator() : current() {} + + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 + explicit reverse_iterator(_Iter __x) : current(__x) {} + + template ::value && is_convertible<_Up const&, _Iter>::value + > > + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 + reverse_iterator(const reverse_iterator<_Up>& __u) + : current(__u.base()) + { } + + template ::value && + is_convertible<_Up const&, _Iter>::value && + is_assignable<_Up const&, _Iter>::value + > > + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 + reverse_iterator& operator=(const reverse_iterator<_Up>& __u) { + current = __u.base(); + return *this; + } +#endif + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 + _Iter base() const {return current;} + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 + reference operator*() const {_Iter __tmp = current; return *--__tmp;} + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 + pointer operator->() const {return _VSTD::addressof(operator*());} + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 + reverse_iterator& operator++() {--current; return *this;} + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 + reverse_iterator operator++(int) {reverse_iterator __tmp(*this); --current; return __tmp;} + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 + reverse_iterator& operator--() {++current; return *this;} + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 + reverse_iterator operator--(int) {reverse_iterator __tmp(*this); ++current; return __tmp;} + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 + reverse_iterator operator+ (difference_type __n) const {return reverse_iterator(current - __n);} + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 + reverse_iterator& operator+=(difference_type __n) {current -= __n; return *this;} + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 + reverse_iterator operator- (difference_type __n) const {return reverse_iterator(current + __n);} + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 + reverse_iterator& operator-=(difference_type __n) {current += __n; return *this;} + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 + reference operator[](difference_type __n) const {return *(*this + __n);} +}; + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 +bool +operator==(const reverse_iterator<_Iter1>& __x, const reverse_iterator<_Iter2>& __y) +{ + return __x.base() == __y.base(); +} + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 +bool +operator<(const reverse_iterator<_Iter1>& __x, const reverse_iterator<_Iter2>& __y) +{ + return __x.base() > __y.base(); +} + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 +bool +operator!=(const reverse_iterator<_Iter1>& __x, const reverse_iterator<_Iter2>& __y) +{ + return __x.base() != __y.base(); +} + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 +bool +operator>(const reverse_iterator<_Iter1>& __x, const reverse_iterator<_Iter2>& __y) +{ + return __x.base() < __y.base(); +} + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 +bool +operator>=(const reverse_iterator<_Iter1>& __x, const reverse_iterator<_Iter2>& __y) +{ + return __x.base() <= __y.base(); +} + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 +bool +operator<=(const reverse_iterator<_Iter1>& __x, const reverse_iterator<_Iter2>& __y) +{ + return __x.base() >= __y.base(); +} + +#ifndef _LIBCPP_CXX03_LANG +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 +auto +operator-(const reverse_iterator<_Iter1>& __x, const reverse_iterator<_Iter2>& __y) +-> decltype(__y.base() - __x.base()) +{ + return __y.base() - __x.base(); +} +#else +template +inline _LIBCPP_INLINE_VISIBILITY +typename reverse_iterator<_Iter1>::difference_type +operator-(const reverse_iterator<_Iter1>& __x, const reverse_iterator<_Iter2>& __y) +{ + return __y.base() - __x.base(); +} +#endif + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 +reverse_iterator<_Iter> +operator+(typename reverse_iterator<_Iter>::difference_type __n, const reverse_iterator<_Iter>& __x) +{ + return reverse_iterator<_Iter>(__x.base() - __n); +} + +#if _LIBCPP_STD_VER > 11 +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 +reverse_iterator<_Iter> make_reverse_iterator(_Iter __i) +{ + return reverse_iterator<_Iter>(__i); +} +#endif + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ITERATOR_REVERSE_ITERATOR_H diff --git a/lib/libcxx/include/__iterator/size.h b/lib/libcxx/include/__iterator/size.h new file mode 100644 index 000000000..259424f1d --- /dev/null +++ b/lib/libcxx/include/__iterator/size.h @@ -0,0 +1,58 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ITERATOR_SIZE_H +#define _LIBCPP___ITERATOR_SIZE_H + +#include <__config> +#include +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +#if _LIBCPP_STD_VER > 14 + +template +_LIBCPP_INLINE_VISIBILITY +constexpr auto size(const _Cont& __c) +_NOEXCEPT_(noexcept(__c.size())) +-> decltype (__c.size()) +{ return __c.size(); } + +template +_LIBCPP_INLINE_VISIBILITY +constexpr size_t size(const _Tp (&)[_Sz]) noexcept { return _Sz; } + +#if _LIBCPP_STD_VER > 17 +template +_LIBCPP_INLINE_VISIBILITY +constexpr auto ssize(const _Cont& __c) +_NOEXCEPT_(noexcept(static_cast>>(__c.size()))) +-> common_type_t> +{ return static_cast>>(__c.size()); } + +template +_LIBCPP_INLINE_VISIBILITY +constexpr ptrdiff_t ssize(const _Tp (&)[_Sz]) noexcept { return _Sz; } +#endif + +#endif // _LIBCPP_STD_VER > 14 + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ITERATOR_SIZE_H diff --git a/lib/libcxx/include/__iterator/wrap_iter.h b/lib/libcxx/include/__iterator/wrap_iter.h new file mode 100644 index 000000000..e35a372b4 --- /dev/null +++ b/lib/libcxx/include/__iterator/wrap_iter.h @@ -0,0 +1,300 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___ITERATOR_WRAP_ITER_H +#define _LIBCPP___ITERATOR_WRAP_ITER_H + +#include <__config> +#include <__debug> +#include <__iterator/iterator_traits.h> +#include <__memory/pointer_traits.h> // __to_address +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +class __wrap_iter +{ +public: + typedef _Iter iterator_type; + typedef typename iterator_traits::value_type value_type; + typedef typename iterator_traits::difference_type difference_type; + typedef typename iterator_traits::pointer pointer; + typedef typename iterator_traits::reference reference; + typedef typename iterator_traits::iterator_category iterator_category; +#if _LIBCPP_STD_VER > 17 + typedef contiguous_iterator_tag iterator_concept; +#endif + +private: + iterator_type __i; +public: + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_IF_NODEBUG __wrap_iter() _NOEXCEPT +#if _LIBCPP_STD_VER > 11 + : __i{} +#endif + { +#if _LIBCPP_DEBUG_LEVEL == 2 + __get_db()->__insert_i(this); +#endif + } + template _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_IF_NODEBUG + __wrap_iter(const __wrap_iter<_Up>& __u, + typename enable_if::value>::type* = nullptr) _NOEXCEPT + : __i(__u.base()) + { +#if _LIBCPP_DEBUG_LEVEL == 2 + __get_db()->__iterator_copy(this, &__u); +#endif + } +#if _LIBCPP_DEBUG_LEVEL == 2 + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_IF_NODEBUG + __wrap_iter(const __wrap_iter& __x) + : __i(__x.base()) + { + __get_db()->__iterator_copy(this, &__x); + } + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_IF_NODEBUG + __wrap_iter& operator=(const __wrap_iter& __x) + { + if (this != &__x) + { + __get_db()->__iterator_copy(this, &__x); + __i = __x.__i; + } + return *this; + } + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_IF_NODEBUG + ~__wrap_iter() + { + __get_db()->__erase_i(this); + } +#endif + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_IF_NODEBUG reference operator*() const _NOEXCEPT + { +#if _LIBCPP_DEBUG_LEVEL == 2 + _LIBCPP_ASSERT(__get_const_db()->__dereferenceable(this), + "Attempted to dereference a non-dereferenceable iterator"); +#endif + return *__i; + } + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_IF_NODEBUG pointer operator->() const _NOEXCEPT + { +#if _LIBCPP_DEBUG_LEVEL == 2 + _LIBCPP_ASSERT(__get_const_db()->__dereferenceable(this), + "Attempted to dereference a non-dereferenceable iterator"); +#endif + return _VSTD::__to_address(__i); + } + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_IF_NODEBUG __wrap_iter& operator++() _NOEXCEPT + { +#if _LIBCPP_DEBUG_LEVEL == 2 + _LIBCPP_ASSERT(__get_const_db()->__dereferenceable(this), + "Attempted to increment a non-incrementable iterator"); +#endif + ++__i; + return *this; + } + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_IF_NODEBUG __wrap_iter operator++(int) _NOEXCEPT + {__wrap_iter __tmp(*this); ++(*this); return __tmp;} + + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_IF_NODEBUG __wrap_iter& operator--() _NOEXCEPT + { +#if _LIBCPP_DEBUG_LEVEL == 2 + _LIBCPP_ASSERT(__get_const_db()->__decrementable(this), + "Attempted to decrement a non-decrementable iterator"); +#endif + --__i; + return *this; + } + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_IF_NODEBUG __wrap_iter operator--(int) _NOEXCEPT + {__wrap_iter __tmp(*this); --(*this); return __tmp;} + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_IF_NODEBUG __wrap_iter operator+ (difference_type __n) const _NOEXCEPT + {__wrap_iter __w(*this); __w += __n; return __w;} + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_IF_NODEBUG __wrap_iter& operator+=(difference_type __n) _NOEXCEPT + { +#if _LIBCPP_DEBUG_LEVEL == 2 + _LIBCPP_ASSERT(__get_const_db()->__addable(this, __n), + "Attempted to add/subtract an iterator outside its valid range"); +#endif + __i += __n; + return *this; + } + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_IF_NODEBUG __wrap_iter operator- (difference_type __n) const _NOEXCEPT + {return *this + (-__n);} + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_IF_NODEBUG __wrap_iter& operator-=(difference_type __n) _NOEXCEPT + {*this += -__n; return *this;} + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_IF_NODEBUG reference operator[](difference_type __n) const _NOEXCEPT + { +#if _LIBCPP_DEBUG_LEVEL == 2 + _LIBCPP_ASSERT(__get_const_db()->__subscriptable(this, __n), + "Attempted to subscript an iterator outside its valid range"); +#endif + return __i[__n]; + } + + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_IF_NODEBUG iterator_type base() const _NOEXCEPT {return __i;} + +private: +#if _LIBCPP_DEBUG_LEVEL == 2 + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_IF_NODEBUG __wrap_iter(const void* __p, iterator_type __x) : __i(__x) + { + __get_db()->__insert_ic(this, __p); + } +#else + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_IF_NODEBUG __wrap_iter(iterator_type __x) _NOEXCEPT : __i(__x) {} +#endif + + template friend class __wrap_iter; + template friend class basic_string; + template friend class _LIBCPP_TEMPLATE_VIS vector; + template friend class _LIBCPP_TEMPLATE_VIS span; +}; + +template +_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_IF_NODEBUG +bool operator==(const __wrap_iter<_Iter1>& __x, const __wrap_iter<_Iter1>& __y) _NOEXCEPT +{ + return __x.base() == __y.base(); +} + +template +_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_IF_NODEBUG +bool operator==(const __wrap_iter<_Iter1>& __x, const __wrap_iter<_Iter2>& __y) _NOEXCEPT +{ + return __x.base() == __y.base(); +} + +template +_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_IF_NODEBUG +bool operator<(const __wrap_iter<_Iter1>& __x, const __wrap_iter<_Iter1>& __y) _NOEXCEPT +{ +#if _LIBCPP_DEBUG_LEVEL == 2 + _LIBCPP_ASSERT(__get_const_db()->__less_than_comparable(&__x, &__y), + "Attempted to compare incomparable iterators"); +#endif + return __x.base() < __y.base(); +} + +template +_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_IF_NODEBUG +bool operator<(const __wrap_iter<_Iter1>& __x, const __wrap_iter<_Iter2>& __y) _NOEXCEPT +{ +#if _LIBCPP_DEBUG_LEVEL == 2 + _LIBCPP_ASSERT(__get_const_db()->__less_than_comparable(&__x, &__y), + "Attempted to compare incomparable iterators"); +#endif + return __x.base() < __y.base(); +} + +template +_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_IF_NODEBUG +bool operator!=(const __wrap_iter<_Iter1>& __x, const __wrap_iter<_Iter1>& __y) _NOEXCEPT +{ + return !(__x == __y); +} + +template +_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_IF_NODEBUG +bool operator!=(const __wrap_iter<_Iter1>& __x, const __wrap_iter<_Iter2>& __y) _NOEXCEPT +{ + return !(__x == __y); +} + +template +_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_IF_NODEBUG +bool operator>(const __wrap_iter<_Iter1>& __x, const __wrap_iter<_Iter1>& __y) _NOEXCEPT +{ + return __y < __x; +} + +template +_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_IF_NODEBUG +bool operator>(const __wrap_iter<_Iter1>& __x, const __wrap_iter<_Iter2>& __y) _NOEXCEPT +{ + return __y < __x; +} + +template +_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_IF_NODEBUG +bool operator>=(const __wrap_iter<_Iter1>& __x, const __wrap_iter<_Iter1>& __y) _NOEXCEPT +{ + return !(__x < __y); +} + +template +_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_IF_NODEBUG +bool operator>=(const __wrap_iter<_Iter1>& __x, const __wrap_iter<_Iter2>& __y) _NOEXCEPT +{ + return !(__x < __y); +} + +template +_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_IF_NODEBUG +bool operator<=(const __wrap_iter<_Iter1>& __x, const __wrap_iter<_Iter1>& __y) _NOEXCEPT +{ + return !(__y < __x); +} + +template +_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_IF_NODEBUG +bool operator<=(const __wrap_iter<_Iter1>& __x, const __wrap_iter<_Iter2>& __y) _NOEXCEPT +{ + return !(__y < __x); +} + +template +_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_IF_NODEBUG +#ifndef _LIBCPP_CXX03_LANG +auto operator-(const __wrap_iter<_Iter1>& __x, const __wrap_iter<_Iter2>& __y) _NOEXCEPT + -> decltype(__x.base() - __y.base()) +#else +typename __wrap_iter<_Iter1>::difference_type +operator-(const __wrap_iter<_Iter1>& __x, const __wrap_iter<_Iter2>& __y) _NOEXCEPT +#endif // C++03 +{ +#if _LIBCPP_DEBUG_LEVEL == 2 + _LIBCPP_ASSERT(__get_const_db()->__less_than_comparable(&__x, &__y), + "Attempted to subtract incompatible iterators"); +#endif + return __x.base() - __y.base(); +} + +template +_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_IF_NODEBUG +__wrap_iter<_Iter1> operator+(typename __wrap_iter<_Iter1>::difference_type __n, __wrap_iter<_Iter1> __x) _NOEXCEPT +{ + __x += __n; + return __x; +} + +#if _LIBCPP_STD_VER <= 17 +template +struct __is_cpp17_contiguous_iterator<__wrap_iter<_It> > : true_type {}; +#endif + +template +_LIBCPP_CONSTEXPR +decltype(_VSTD::__to_address(declval<_Iter>())) +__to_address(__wrap_iter<_Iter> __w) _NOEXCEPT { + return _VSTD::__to_address(__w.base()); +} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___ITERATOR_WRAP_ITER_H diff --git a/lib/libcxx/include/__libcpp_version b/lib/libcxx/include/__libcpp_version index e334181b4..09514aa4d 100644 --- a/lib/libcxx/include/__libcpp_version +++ b/lib/libcxx/include/__libcpp_version @@ -1 +1 @@ -12000 +13000 diff --git a/lib/libcxx/include/__locale b/lib/libcxx/include/__locale index 77e5faab2..ad742997d 100644 --- a/lib/libcxx/include/__locale +++ b/lib/libcxx/include/__locale @@ -10,8 +10,8 @@ #ifndef _LIBCPP___LOCALE #define _LIBCPP___LOCALE -#include <__config> #include <__availability> +#include <__config> #include #include #include @@ -1161,7 +1161,7 @@ protected: virtual int do_max_length() const _NOEXCEPT; }; -#ifndef _LIBCPP_NO_HAS_CHAR8_T +#ifndef _LIBCPP_HAS_NO_CHAR8_T // template <> class codecvt // C++20 @@ -1337,7 +1337,7 @@ protected: virtual int do_max_length() const _NOEXCEPT; }; -#ifndef _LIBCPP_NO_HAS_CHAR8_T +#ifndef _LIBCPP_HAS_NO_CHAR8_T // template <> class codecvt // C++20 @@ -1455,7 +1455,7 @@ _LIBCPP_EXTERN_TEMPLATE_EVEN_IN_DEBUG_MODE(class _LIBCPP_EXTERN_TEMPLATE_TYPE_VI _LIBCPP_EXTERN_TEMPLATE_EVEN_IN_DEBUG_MODE(class _LIBCPP_EXTERN_TEMPLATE_TYPE_VIS codecvt_byname) _LIBCPP_EXTERN_TEMPLATE_EVEN_IN_DEBUG_MODE(class _LIBCPP_DEPRECATED_IN_CXX20 _LIBCPP_EXTERN_TEMPLATE_TYPE_VIS codecvt_byname) // deprecated in C++20 _LIBCPP_EXTERN_TEMPLATE_EVEN_IN_DEBUG_MODE(class _LIBCPP_DEPRECATED_IN_CXX20 _LIBCPP_EXTERN_TEMPLATE_TYPE_VIS codecvt_byname) // deprecated in C++20 -#ifndef _LIBCPP_NO_HAS_CHAR8_T +#ifndef _LIBCPP_HAS_NO_CHAR8_T _LIBCPP_EXTERN_TEMPLATE_EVEN_IN_DEBUG_MODE(class _LIBCPP_EXTERN_TEMPLATE_TYPE_VIS codecvt_byname) // C++20 _LIBCPP_EXTERN_TEMPLATE_EVEN_IN_DEBUG_MODE(class _LIBCPP_EXTERN_TEMPLATE_TYPE_VIS codecvt_byname) // C++20 #endif @@ -1484,14 +1484,14 @@ struct __narrow_to_utf8<8> _LIBCPP_SUPPRESS_DEPRECATED_PUSH template <> -struct _LIBCPP_TEMPLATE_VIS __narrow_to_utf8<16> +struct _LIBCPP_TYPE_VIS __narrow_to_utf8<16> : public codecvt { _LIBCPP_INLINE_VISIBILITY __narrow_to_utf8() : codecvt(1) {} _LIBCPP_SUPPRESS_DEPRECATED_POP - _LIBCPP_EXPORTED_FROM_ABI ~__narrow_to_utf8(); + ~__narrow_to_utf8(); template _LIBCPP_INLINE_VISIBILITY @@ -1520,14 +1520,14 @@ _LIBCPP_SUPPRESS_DEPRECATED_POP _LIBCPP_SUPPRESS_DEPRECATED_PUSH template <> -struct _LIBCPP_TEMPLATE_VIS __narrow_to_utf8<32> +struct _LIBCPP_TYPE_VIS __narrow_to_utf8<32> : public codecvt { _LIBCPP_INLINE_VISIBILITY __narrow_to_utf8() : codecvt(1) {} _LIBCPP_SUPPRESS_DEPRECATED_POP - _LIBCPP_EXPORTED_FROM_ABI ~__narrow_to_utf8(); + ~__narrow_to_utf8(); template _LIBCPP_INLINE_VISIBILITY @@ -1578,14 +1578,14 @@ struct __widen_from_utf8<8> _LIBCPP_SUPPRESS_DEPRECATED_PUSH template <> -struct _LIBCPP_TEMPLATE_VIS __widen_from_utf8<16> +struct _LIBCPP_TYPE_VIS __widen_from_utf8<16> : public codecvt { _LIBCPP_INLINE_VISIBILITY __widen_from_utf8() : codecvt(1) {} _LIBCPP_SUPPRESS_DEPRECATED_POP - _LIBCPP_EXPORTED_FROM_ABI ~__widen_from_utf8(); + ~__widen_from_utf8(); template _LIBCPP_INLINE_VISIBILITY @@ -1614,14 +1614,14 @@ _LIBCPP_SUPPRESS_DEPRECATED_POP _LIBCPP_SUPPRESS_DEPRECATED_PUSH template <> -struct _LIBCPP_TEMPLATE_VIS __widen_from_utf8<32> +struct _LIBCPP_TYPE_VIS __widen_from_utf8<32> : public codecvt { _LIBCPP_INLINE_VISIBILITY __widen_from_utf8() : codecvt(1) {} _LIBCPP_SUPPRESS_DEPRECATED_POP - _LIBCPP_EXPORTED_FROM_ABI ~__widen_from_utf8(); + ~__widen_from_utf8(); template _LIBCPP_INLINE_VISIBILITY @@ -1756,4 +1756,4 @@ private: _LIBCPP_END_NAMESPACE_STD -#endif // _LIBCPP___LOCALE +#endif // _LIBCPP___LOCALE diff --git a/lib/libcxx/include/__memory/base.h b/lib/libcxx/include/__memory/addressof.h similarity index 72% rename from lib/libcxx/include/__memory/base.h rename to lib/libcxx/include/__memory/addressof.h index 70728bd7f..5efdb5878 100644 --- a/lib/libcxx/include/__memory/base.h +++ b/lib/libcxx/include/__memory/addressof.h @@ -7,12 +7,10 @@ // //===----------------------------------------------------------------------===// -#ifndef _LIBCPP___MEMORY_BASE_H -#define _LIBCPP___MEMORY_BASE_H +#ifndef _LIBCPP___MEMORY_ADDRESSOF_H +#define _LIBCPP___MEMORY_ADDRESSOF_H #include <__config> -#include <__debug> -#include #if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) #pragma GCC system_header @@ -23,7 +21,6 @@ _LIBCPP_PUSH_MACROS _LIBCPP_BEGIN_NAMESPACE_STD -// addressof #ifndef _LIBCPP_HAS_NO_BUILTIN_ADDRESSOF template @@ -92,36 +89,8 @@ addressof(__unsafe_unretained _Tp& __x) _NOEXCEPT template _Tp* addressof(const _Tp&&) noexcept = delete; #endif -// construct_at - -#if _LIBCPP_STD_VER > 17 - -template()) _Tp(_VSTD::declval<_Args>()...) -)> -_LIBCPP_INLINE_VISIBILITY -constexpr _Tp* construct_at(_Tp* __location, _Args&& ...__args) { - _LIBCPP_ASSERT(__location, "null pointer given to construct_at"); - return ::new ((void*)__location) _Tp(_VSTD::forward<_Args>(__args)...); -} - -#endif - -// destroy_at - -#if _LIBCPP_STD_VER > 14 - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -void destroy_at(_Tp* __loc) { - _LIBCPP_ASSERT(__loc, "null pointer given to destroy_at"); - __loc->~_Tp(); -} - -#endif - _LIBCPP_END_NAMESPACE_STD _LIBCPP_POP_MACROS -#endif // _LIBCPP___MEMORY_BASE_H +#endif // _LIBCPP___MEMORY_ADDRESSOF_H diff --git a/lib/libcxx/include/__memory/utilities.h b/lib/libcxx/include/__memory/allocation_guard.h similarity index 94% rename from lib/libcxx/include/__memory/utilities.h rename to lib/libcxx/include/__memory/allocation_guard.h index aac3d11ca..4987af293 100644 --- a/lib/libcxx/include/__memory/utilities.h +++ b/lib/libcxx/include/__memory/allocation_guard.h @@ -7,12 +7,13 @@ // //===----------------------------------------------------------------------===// -#ifndef _LIBCPP___MEMORY_UTILITIES_H -#define _LIBCPP___MEMORY_UTILITIES_H +#ifndef _LIBCPP___MEMORY_ALLOCATION_GUARD_H +#define _LIBCPP___MEMORY_ALLOCATION_GUARD_H #include <__config> #include <__memory/allocator_traits.h> #include +#include #if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) #pragma GCC system_header @@ -85,4 +86,4 @@ _LIBCPP_END_NAMESPACE_STD _LIBCPP_POP_MACROS -#endif // _LIBCPP___MEMORY_UTILITIES_H +#endif // _LIBCPP___MEMORY_ALLOCATION_GUARD_H diff --git a/lib/libcxx/include/__memory/allocator.h b/lib/libcxx/include/__memory/allocator.h new file mode 100644 index 000000000..2c21a16e7 --- /dev/null +++ b/lib/libcxx/include/__memory/allocator.h @@ -0,0 +1,254 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___MEMORY_ALLOCATOR_H +#define _LIBCPP___MEMORY_ALLOCATOR_H + +#include <__config> +#include <__memory/allocator_traits.h> +#include <__utility/forward.h> +#include +#include +#include +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template class allocator; + +#if _LIBCPP_STD_VER <= 17 +template <> +class _LIBCPP_TEMPLATE_VIS allocator +{ +public: + _LIBCPP_DEPRECATED_IN_CXX17 typedef void* pointer; + _LIBCPP_DEPRECATED_IN_CXX17 typedef const void* const_pointer; + _LIBCPP_DEPRECATED_IN_CXX17 typedef void value_type; + + template struct _LIBCPP_DEPRECATED_IN_CXX17 rebind {typedef allocator<_Up> other;}; +}; + +template <> +class _LIBCPP_TEMPLATE_VIS allocator +{ +public: + _LIBCPP_DEPRECATED_IN_CXX17 typedef const void* pointer; + _LIBCPP_DEPRECATED_IN_CXX17 typedef const void* const_pointer; + _LIBCPP_DEPRECATED_IN_CXX17 typedef const void value_type; + + template struct _LIBCPP_DEPRECATED_IN_CXX17 rebind {typedef allocator<_Up> other;}; +}; +#endif + +// This class provides a non-trivial default constructor to the class that derives from it +// if the condition is satisfied. +// +// The second template parameter exists to allow giving a unique type to __non_trivial_if, +// which makes it possible to avoid breaking the ABI when making this a base class of an +// existing class. Without that, imagine we have classes D1 and D2, both of which used to +// have no base classes, but which now derive from __non_trivial_if. The layout of a class +// that inherits from both D1 and D2 will change because the two __non_trivial_if base +// classes are not allowed to share the same address. +// +// By making those __non_trivial_if base classes unique, we work around this problem and +// it is safe to start deriving from __non_trivial_if in existing classes. +template +struct __non_trivial_if { }; + +template +struct __non_trivial_if { + _LIBCPP_INLINE_VISIBILITY + _LIBCPP_CONSTEXPR __non_trivial_if() _NOEXCEPT { } +}; + +// allocator +// +// Note: For ABI compatibility between C++20 and previous standards, we make +// allocator trivial in C++20. + +template +class _LIBCPP_TEMPLATE_VIS allocator + : private __non_trivial_if::value, allocator<_Tp> > +{ +public: + typedef size_t size_type; + typedef ptrdiff_t difference_type; + typedef _Tp value_type; + typedef true_type propagate_on_container_move_assignment; + typedef true_type is_always_equal; + + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 + allocator() _NOEXCEPT _LIBCPP_DEFAULT + + template + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 + allocator(const allocator<_Up>&) _NOEXCEPT { } + + _LIBCPP_NODISCARD_AFTER_CXX17 _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 + _Tp* allocate(size_t __n) { + if (__n > allocator_traits::max_size(*this)) + __throw_length_error("allocator::allocate(size_t n)" + " 'n' exceeds maximum supported size"); + if (__libcpp_is_constant_evaluated()) { + return static_cast<_Tp*>(::operator new(__n * sizeof(_Tp))); + } else { + return static_cast<_Tp*>(_VSTD::__libcpp_allocate(__n * sizeof(_Tp), _LIBCPP_ALIGNOF(_Tp))); + } + } + + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 + void deallocate(_Tp* __p, size_t __n) _NOEXCEPT { + if (__libcpp_is_constant_evaluated()) { + ::operator delete(__p); + } else { + _VSTD::__libcpp_deallocate((void*)__p, __n * sizeof(_Tp), _LIBCPP_ALIGNOF(_Tp)); + } + } + + // C++20 Removed members +#if _LIBCPP_STD_VER <= 17 || defined(_LIBCPP_ENABLE_CXX20_REMOVED_ALLOCATOR_MEMBERS) + _LIBCPP_DEPRECATED_IN_CXX17 typedef _Tp* pointer; + _LIBCPP_DEPRECATED_IN_CXX17 typedef const _Tp* const_pointer; + _LIBCPP_DEPRECATED_IN_CXX17 typedef _Tp& reference; + _LIBCPP_DEPRECATED_IN_CXX17 typedef const _Tp& const_reference; + + template + struct _LIBCPP_DEPRECATED_IN_CXX17 rebind { + typedef allocator<_Up> other; + }; + + _LIBCPP_DEPRECATED_IN_CXX17 _LIBCPP_INLINE_VISIBILITY + pointer address(reference __x) const _NOEXCEPT { + return _VSTD::addressof(__x); + } + _LIBCPP_DEPRECATED_IN_CXX17 _LIBCPP_INLINE_VISIBILITY + const_pointer address(const_reference __x) const _NOEXCEPT { + return _VSTD::addressof(__x); + } + + _LIBCPP_NODISCARD_AFTER_CXX17 _LIBCPP_INLINE_VISIBILITY _LIBCPP_DEPRECATED_IN_CXX17 + _Tp* allocate(size_t __n, const void*) { + return allocate(__n); + } + + _LIBCPP_DEPRECATED_IN_CXX17 _LIBCPP_INLINE_VISIBILITY size_type max_size() const _NOEXCEPT { + return size_type(~0) / sizeof(_Tp); + } + + template + _LIBCPP_DEPRECATED_IN_CXX17 _LIBCPP_INLINE_VISIBILITY + void construct(_Up* __p, _Args&&... __args) { + ::new ((void*)__p) _Up(_VSTD::forward<_Args>(__args)...); + } + + _LIBCPP_DEPRECATED_IN_CXX17 _LIBCPP_INLINE_VISIBILITY + void destroy(pointer __p) { + __p->~_Tp(); + } +#endif +}; + +template +class _LIBCPP_TEMPLATE_VIS allocator + : private __non_trivial_if::value, allocator > +{ +public: + typedef size_t size_type; + typedef ptrdiff_t difference_type; + typedef const _Tp value_type; + typedef true_type propagate_on_container_move_assignment; + typedef true_type is_always_equal; + + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 + allocator() _NOEXCEPT _LIBCPP_DEFAULT + + template + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 + allocator(const allocator<_Up>&) _NOEXCEPT { } + + _LIBCPP_NODISCARD_AFTER_CXX17 _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 + const _Tp* allocate(size_t __n) { + if (__n > allocator_traits::max_size(*this)) + __throw_length_error("allocator::allocate(size_t n)" + " 'n' exceeds maximum supported size"); + if (__libcpp_is_constant_evaluated()) { + return static_cast(::operator new(__n * sizeof(_Tp))); + } else { + return static_cast(_VSTD::__libcpp_allocate(__n * sizeof(_Tp), _LIBCPP_ALIGNOF(_Tp))); + } + } + + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 + void deallocate(const _Tp* __p, size_t __n) { + if (__libcpp_is_constant_evaluated()) { + ::operator delete(const_cast<_Tp*>(__p)); + } else { + _VSTD::__libcpp_deallocate((void*) const_cast<_Tp *>(__p), __n * sizeof(_Tp), _LIBCPP_ALIGNOF(_Tp)); + } + } + + // C++20 Removed members +#if _LIBCPP_STD_VER <= 17 || defined(_LIBCPP_ENABLE_CXX20_REMOVED_ALLOCATOR_MEMBERS) + _LIBCPP_DEPRECATED_IN_CXX17 typedef const _Tp* pointer; + _LIBCPP_DEPRECATED_IN_CXX17 typedef const _Tp* const_pointer; + _LIBCPP_DEPRECATED_IN_CXX17 typedef const _Tp& reference; + _LIBCPP_DEPRECATED_IN_CXX17 typedef const _Tp& const_reference; + + template + struct _LIBCPP_DEPRECATED_IN_CXX17 rebind { + typedef allocator<_Up> other; + }; + + _LIBCPP_DEPRECATED_IN_CXX17 _LIBCPP_INLINE_VISIBILITY + const_pointer address(const_reference __x) const _NOEXCEPT { + return _VSTD::addressof(__x); + } + + _LIBCPP_NODISCARD_AFTER_CXX17 _LIBCPP_INLINE_VISIBILITY _LIBCPP_DEPRECATED_IN_CXX17 + const _Tp* allocate(size_t __n, const void*) { + return allocate(__n); + } + + _LIBCPP_DEPRECATED_IN_CXX17 _LIBCPP_INLINE_VISIBILITY size_type max_size() const _NOEXCEPT { + return size_type(~0) / sizeof(_Tp); + } + + template + _LIBCPP_DEPRECATED_IN_CXX17 _LIBCPP_INLINE_VISIBILITY + void construct(_Up* __p, _Args&&... __args) { + ::new ((void*)__p) _Up(_VSTD::forward<_Args>(__args)...); + } + + _LIBCPP_DEPRECATED_IN_CXX17 _LIBCPP_INLINE_VISIBILITY + void destroy(pointer __p) { + __p->~_Tp(); + } +#endif +}; + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +bool operator==(const allocator<_Tp>&, const allocator<_Up>&) _NOEXCEPT {return true;} + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +bool operator!=(const allocator<_Tp>&, const allocator<_Up>&) _NOEXCEPT {return false;} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___MEMORY_ALLOCATOR_H diff --git a/lib/libcxx/include/__memory/allocator_arg_t.h b/lib/libcxx/include/__memory/allocator_arg_t.h new file mode 100644 index 000000000..830c6b814 --- /dev/null +++ b/lib/libcxx/include/__memory/allocator_arg_t.h @@ -0,0 +1,78 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___FUNCTIONAL___ALLOCATOR_ARG_T_H +#define _LIBCPP___FUNCTIONAL___ALLOCATOR_ARG_T_H + +#include <__config> +#include <__memory/uses_allocator.h> +#include <__utility/forward.h> +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_BEGIN_NAMESPACE_STD + +struct _LIBCPP_TEMPLATE_VIS allocator_arg_t { explicit allocator_arg_t() = default; }; + +#if defined(_LIBCPP_CXX03_LANG) || defined(_LIBCPP_BUILDING_LIBRARY) +extern _LIBCPP_EXPORTED_FROM_ABI const allocator_arg_t allocator_arg; +#else +/* _LIBCPP_INLINE_VAR */ constexpr allocator_arg_t allocator_arg = allocator_arg_t(); +#endif + +#ifndef _LIBCPP_CXX03_LANG + +// allocator construction + +template +struct __uses_alloc_ctor_imp +{ + typedef _LIBCPP_NODEBUG_TYPE typename __uncvref<_Alloc>::type _RawAlloc; + static const bool __ua = uses_allocator<_Tp, _RawAlloc>::value; + static const bool __ic = + is_constructible<_Tp, allocator_arg_t, _Alloc, _Args...>::value; + static const int value = __ua ? 2 - __ic : 0; +}; + +template +struct __uses_alloc_ctor + : integral_constant::value> + {}; + +template +inline _LIBCPP_INLINE_VISIBILITY +void __user_alloc_construct_impl (integral_constant, _Tp *__storage, const _Allocator &, _Args &&... __args ) +{ + new (__storage) _Tp (_VSTD::forward<_Args>(__args)...); +} + +// FIXME: This should have a version which takes a non-const alloc. +template +inline _LIBCPP_INLINE_VISIBILITY +void __user_alloc_construct_impl (integral_constant, _Tp *__storage, const _Allocator &__a, _Args &&... __args ) +{ + new (__storage) _Tp (allocator_arg, __a, _VSTD::forward<_Args>(__args)...); +} + +// FIXME: This should have a version which takes a non-const alloc. +template +inline _LIBCPP_INLINE_VISIBILITY +void __user_alloc_construct_impl (integral_constant, _Tp *__storage, const _Allocator &__a, _Args &&... __args ) +{ + new (__storage) _Tp (_VSTD::forward<_Args>(__args)..., __a); +} + +#endif // _LIBCPP_CXX03_LANG + +_LIBCPP_END_NAMESPACE_STD + +#endif // _LIBCPP___FUNCTIONAL___ALLOCATOR_ARG_T_H diff --git a/lib/libcxx/include/__memory/allocator_traits.h b/lib/libcxx/include/__memory/allocator_traits.h index 9443f61b7..a02af0dea 100644 --- a/lib/libcxx/include/__memory/allocator_traits.h +++ b/lib/libcxx/include/__memory/allocator_traits.h @@ -11,8 +11,10 @@ #define _LIBCPP___MEMORY_ALLOCATOR_TRAITS_H #include <__config> -#include <__memory/base.h> +#include <__memory/construct_at.h> #include <__memory/pointer_traits.h> +#include <__utility/forward.h> +#include #include #if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) @@ -253,7 +255,7 @@ struct _LIBCPP_TEMPLATE_VIS allocator_traits struct rebind_traits { using other = allocator_traits::other>; }; -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG _LIBCPP_NODISCARD_AFTER_CXX17 _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 static pointer allocate(allocator_type& __a, size_type __n) { @@ -360,8 +362,10 @@ struct __rebind_alloc_helper { template struct __is_default_allocator : false_type { }; +template class allocator; + template -struct __is_default_allocator<_VSTD::allocator<_Tp> > : true_type { }; +struct __is_default_allocator > : true_type { }; // __is_cpp17_move_insertable template @@ -398,4 +402,4 @@ _LIBCPP_END_NAMESPACE_STD _LIBCPP_POP_MACROS -#endif // _LIBCPP___MEMORY_ALLOCATOR_TRAITS_H +#endif // _LIBCPP___MEMORY_ALLOCATOR_TRAITS_H diff --git a/lib/libcxx/include/__memory/auto_ptr.h b/lib/libcxx/include/__memory/auto_ptr.h new file mode 100644 index 000000000..f8d2b507b --- /dev/null +++ b/lib/libcxx/include/__memory/auto_ptr.h @@ -0,0 +1,86 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___MEMORY_AUTO_PTR_H +#define _LIBCPP___MEMORY_AUTO_PTR_H + +#include <__config> +#include <__nullptr> + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +struct _LIBCPP_DEPRECATED_IN_CXX11 auto_ptr_ref +{ + _Tp* __ptr_; +}; + +template +class _LIBCPP_TEMPLATE_VIS _LIBCPP_DEPRECATED_IN_CXX11 auto_ptr +{ +private: + _Tp* __ptr_; +public: + typedef _Tp element_type; + + _LIBCPP_INLINE_VISIBILITY explicit auto_ptr(_Tp* __p = 0) _NOEXCEPT : __ptr_(__p) {} + _LIBCPP_INLINE_VISIBILITY auto_ptr(auto_ptr& __p) _NOEXCEPT : __ptr_(__p.release()) {} + template _LIBCPP_INLINE_VISIBILITY auto_ptr(auto_ptr<_Up>& __p) _NOEXCEPT + : __ptr_(__p.release()) {} + _LIBCPP_INLINE_VISIBILITY auto_ptr& operator=(auto_ptr& __p) _NOEXCEPT + {reset(__p.release()); return *this;} + template _LIBCPP_INLINE_VISIBILITY auto_ptr& operator=(auto_ptr<_Up>& __p) _NOEXCEPT + {reset(__p.release()); return *this;} + _LIBCPP_INLINE_VISIBILITY auto_ptr& operator=(auto_ptr_ref<_Tp> __p) _NOEXCEPT + {reset(__p.__ptr_); return *this;} + _LIBCPP_INLINE_VISIBILITY ~auto_ptr() _NOEXCEPT {delete __ptr_;} + + _LIBCPP_INLINE_VISIBILITY _Tp& operator*() const _NOEXCEPT + {return *__ptr_;} + _LIBCPP_INLINE_VISIBILITY _Tp* operator->() const _NOEXCEPT {return __ptr_;} + _LIBCPP_INLINE_VISIBILITY _Tp* get() const _NOEXCEPT {return __ptr_;} + _LIBCPP_INLINE_VISIBILITY _Tp* release() _NOEXCEPT + { + _Tp* __t = __ptr_; + __ptr_ = nullptr; + return __t; + } + _LIBCPP_INLINE_VISIBILITY void reset(_Tp* __p = 0) _NOEXCEPT + { + if (__ptr_ != __p) + delete __ptr_; + __ptr_ = __p; + } + + _LIBCPP_INLINE_VISIBILITY auto_ptr(auto_ptr_ref<_Tp> __p) _NOEXCEPT : __ptr_(__p.__ptr_) {} + template _LIBCPP_INLINE_VISIBILITY operator auto_ptr_ref<_Up>() _NOEXCEPT + {auto_ptr_ref<_Up> __t; __t.__ptr_ = release(); return __t;} + template _LIBCPP_INLINE_VISIBILITY operator auto_ptr<_Up>() _NOEXCEPT + {return auto_ptr<_Up>(release());} +}; + +template <> +class _LIBCPP_TEMPLATE_VIS _LIBCPP_DEPRECATED_IN_CXX11 auto_ptr +{ +public: + typedef void element_type; +}; + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___MEMORY_AUTO_PTR_H diff --git a/lib/libcxx/include/__memory/compressed_pair.h b/lib/libcxx/include/__memory/compressed_pair.h new file mode 100644 index 000000000..08f0318a1 --- /dev/null +++ b/lib/libcxx/include/__memory/compressed_pair.h @@ -0,0 +1,201 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___MEMORY_COMPRESSED_PAIR_H +#define _LIBCPP___MEMORY_COMPRESSED_PAIR_H + +#include <__config> +#include <__utility/forward.h> +#include // needed in c++03 for some constructors +#include +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +// Tag used to default initialize one or both of the pair's elements. +struct __default_init_tag {}; +struct __value_init_tag {}; + +template ::value && !__libcpp_is_final<_Tp>::value> +struct __compressed_pair_elem { + typedef _Tp _ParamT; + typedef _Tp& reference; + typedef const _Tp& const_reference; + + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR + __compressed_pair_elem(__default_init_tag) {} + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR + __compressed_pair_elem(__value_init_tag) : __value_() {} + + template ::type>::value + >::type> + _LIBCPP_INLINE_VISIBILITY + _LIBCPP_CONSTEXPR explicit + __compressed_pair_elem(_Up&& __u) + : __value_(_VSTD::forward<_Up>(__u)) + { + } + + +#ifndef _LIBCPP_CXX03_LANG + template + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 + __compressed_pair_elem(piecewise_construct_t, tuple<_Args...> __args, + __tuple_indices<_Indexes...>) + : __value_(_VSTD::forward<_Args>(_VSTD::get<_Indexes>(__args))...) {} +#endif + + + _LIBCPP_INLINE_VISIBILITY reference __get() _NOEXCEPT { return __value_; } + _LIBCPP_INLINE_VISIBILITY + const_reference __get() const _NOEXCEPT { return __value_; } + +private: + _Tp __value_; +}; + +template +struct __compressed_pair_elem<_Tp, _Idx, true> : private _Tp { + typedef _Tp _ParamT; + typedef _Tp& reference; + typedef const _Tp& const_reference; + typedef _Tp __value_type; + + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR __compressed_pair_elem() = default; + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR + __compressed_pair_elem(__default_init_tag) {} + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR + __compressed_pair_elem(__value_init_tag) : __value_type() {} + + template ::type>::value + >::type> + _LIBCPP_INLINE_VISIBILITY + _LIBCPP_CONSTEXPR explicit + __compressed_pair_elem(_Up&& __u) + : __value_type(_VSTD::forward<_Up>(__u)) + {} + +#ifndef _LIBCPP_CXX03_LANG + template + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 + __compressed_pair_elem(piecewise_construct_t, tuple<_Args...> __args, + __tuple_indices<_Indexes...>) + : __value_type(_VSTD::forward<_Args>(_VSTD::get<_Indexes>(__args))...) {} +#endif + + _LIBCPP_INLINE_VISIBILITY reference __get() _NOEXCEPT { return *this; } + _LIBCPP_INLINE_VISIBILITY + const_reference __get() const _NOEXCEPT { return *this; } +}; + +template +class __compressed_pair : private __compressed_pair_elem<_T1, 0>, + private __compressed_pair_elem<_T2, 1> { +public: + // NOTE: This static assert should never fire because __compressed_pair + // is *almost never* used in a scenario where it's possible for T1 == T2. + // (The exception is std::function where it is possible that the function + // object and the allocator have the same type). + static_assert((!is_same<_T1, _T2>::value), + "__compressed_pair cannot be instantiated when T1 and T2 are the same type; " + "The current implementation is NOT ABI-compatible with the previous " + "implementation for this configuration"); + + typedef _LIBCPP_NODEBUG_TYPE __compressed_pair_elem<_T1, 0> _Base1; + typedef _LIBCPP_NODEBUG_TYPE __compressed_pair_elem<_T2, 1> _Base2; + + template , _Dummy>::value && + __dependent_type, _Dummy>::value + >::type + > + _LIBCPP_INLINE_VISIBILITY + _LIBCPP_CONSTEXPR __compressed_pair() : _Base1(__value_init_tag()), _Base2(__value_init_tag()) {} + + template + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR + __compressed_pair(_U1&& __t1, _U2&& __t2) + : _Base1(_VSTD::forward<_U1>(__t1)), _Base2(_VSTD::forward<_U2>(__t2)) {} + +#ifndef _LIBCPP_CXX03_LANG + template + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 + __compressed_pair(piecewise_construct_t __pc, tuple<_Args1...> __first_args, + tuple<_Args2...> __second_args) + : _Base1(__pc, _VSTD::move(__first_args), + typename __make_tuple_indices::type()), + _Base2(__pc, _VSTD::move(__second_args), + typename __make_tuple_indices::type()) {} +#endif + + _LIBCPP_INLINE_VISIBILITY + typename _Base1::reference first() _NOEXCEPT { + return static_cast<_Base1&>(*this).__get(); + } + + _LIBCPP_INLINE_VISIBILITY + typename _Base1::const_reference first() const _NOEXCEPT { + return static_cast<_Base1 const&>(*this).__get(); + } + + _LIBCPP_INLINE_VISIBILITY + typename _Base2::reference second() _NOEXCEPT { + return static_cast<_Base2&>(*this).__get(); + } + + _LIBCPP_INLINE_VISIBILITY + typename _Base2::const_reference second() const _NOEXCEPT { + return static_cast<_Base2 const&>(*this).__get(); + } + + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR + static _Base1* __get_first_base(__compressed_pair* __pair) _NOEXCEPT { + return static_cast<_Base1*>(__pair); + } + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR + static _Base2* __get_second_base(__compressed_pair* __pair) _NOEXCEPT { + return static_cast<_Base2*>(__pair); + } + + _LIBCPP_INLINE_VISIBILITY + void swap(__compressed_pair& __x) + _NOEXCEPT_(__is_nothrow_swappable<_T1>::value && + __is_nothrow_swappable<_T2>::value) + { + using _VSTD::swap; + swap(first(), __x.first()); + swap(second(), __x.second()); + } +}; + +template +inline _LIBCPP_INLINE_VISIBILITY +void swap(__compressed_pair<_T1, _T2>& __x, __compressed_pair<_T1, _T2>& __y) + _NOEXCEPT_(__is_nothrow_swappable<_T1>::value && + __is_nothrow_swappable<_T2>::value) { + __x.swap(__y); +} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___MEMORY_COMPRESSED_PAIR_H diff --git a/lib/libcxx/include/__memory/construct_at.h b/lib/libcxx/include/__memory/construct_at.h new file mode 100644 index 000000000..7ab19315d --- /dev/null +++ b/lib/libcxx/include/__memory/construct_at.h @@ -0,0 +1,59 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___MEMORY_CONSTRUCT_AT_H +#define _LIBCPP___MEMORY_CONSTRUCT_AT_H + +#include <__config> +#include <__debug> +#include <__utility/forward.h> +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +// construct_at + +#if _LIBCPP_STD_VER > 17 + +template()) _Tp(declval<_Args>()...) +)> +_LIBCPP_INLINE_VISIBILITY +constexpr _Tp* construct_at(_Tp* __location, _Args&& ...__args) { + _LIBCPP_ASSERT(__location, "null pointer given to construct_at"); + return ::new ((void*)__location) _Tp(_VSTD::forward<_Args>(__args)...); +} + +#endif + +// destroy_at + +#if _LIBCPP_STD_VER > 14 + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +void destroy_at(_Tp* __loc) { + _LIBCPP_ASSERT(__loc, "null pointer given to destroy_at"); + __loc->~_Tp(); +} + +#endif + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___MEMORY_CONSTRUCT_AT_H diff --git a/lib/libcxx/include/__memory/pointer_safety.h b/lib/libcxx/include/__memory/pointer_safety.h new file mode 100644 index 000000000..87a6a9659 --- /dev/null +++ b/lib/libcxx/include/__memory/pointer_safety.h @@ -0,0 +1,57 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___MEMORY_POINTER_SAFETY_H +#define _LIBCPP___MEMORY_POINTER_SAFETY_H + +#include <__config> +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +#if !defined(_LIBCPP_CXX03_LANG) + +enum class pointer_safety : unsigned char { + relaxed, + preferred, + strict +}; + +inline _LIBCPP_INLINE_VISIBILITY +pointer_safety get_pointer_safety() _NOEXCEPT { + return pointer_safety::relaxed; +} + +_LIBCPP_FUNC_VIS void declare_reachable(void* __p); +_LIBCPP_FUNC_VIS void declare_no_pointers(char* __p, size_t __n); +_LIBCPP_FUNC_VIS void undeclare_no_pointers(char* __p, size_t __n); +_LIBCPP_FUNC_VIS void* __undeclare_reachable(void* __p); + +template +inline _LIBCPP_INLINE_VISIBILITY +_Tp* +undeclare_reachable(_Tp* __p) +{ + return static_cast<_Tp*>(__undeclare_reachable(__p)); +} + +#endif // !C++03 + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___MEMORY_POINTER_SAFETY_H diff --git a/lib/libcxx/include/__memory/pointer_traits.h b/lib/libcxx/include/__memory/pointer_traits.h index b2c5d34cb..d5442b84e 100644 --- a/lib/libcxx/include/__memory/pointer_traits.h +++ b/lib/libcxx/include/__memory/pointer_traits.h @@ -11,6 +11,7 @@ #define _LIBCPP___MEMORY_POINTER_TRAITS_H #include <__config> +#include <__memory/addressof.h> #include #if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) @@ -120,7 +121,7 @@ struct _LIBCPP_TEMPLATE_VIS pointer_traits #else template struct rebind {typedef typename __pointer_traits_rebind::type other;}; -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG private: struct __nat {}; @@ -162,8 +163,54 @@ struct __rebind_pointer { #endif }; +// to_address + +template +struct __to_address_helper; + +template +_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR +_Tp* __to_address(_Tp* __p) _NOEXCEPT { + static_assert(!is_function<_Tp>::value, "_Tp is a function type"); + return __p; +} + +// enable_if is needed here to avoid instantiating checks for fancy pointers on raw pointers +template ::value> > +_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR +typename decay::__call(declval()))>::type +__to_address(const _Pointer& __p) _NOEXCEPT { + return __to_address_helper<_Pointer>::__call(__p); +} + +template +struct __to_address_helper { + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR + static decltype(_VSTD::__to_address(declval().operator->())) + __call(const _Pointer&__p) _NOEXCEPT { + return _VSTD::__to_address(__p.operator->()); + } +}; + +template +struct __to_address_helper<_Pointer, decltype((void)pointer_traits<_Pointer>::to_address(declval()))> { + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR + static decltype(pointer_traits<_Pointer>::to_address(declval())) + __call(const _Pointer&__p) _NOEXCEPT { + return pointer_traits<_Pointer>::to_address(__p); + } +}; + +#if _LIBCPP_STD_VER > 17 +template +inline _LIBCPP_INLINE_VISIBILITY constexpr +auto to_address(const _Pointer& __p) noexcept { + return _VSTD::__to_address(__p); +} +#endif + _LIBCPP_END_NAMESPACE_STD _LIBCPP_POP_MACROS -#endif // _LIBCPP___MEMORY_POINTER_TRAITS_H +#endif // _LIBCPP___MEMORY_POINTER_TRAITS_H diff --git a/lib/libcxx/include/__memory/raw_storage_iterator.h b/lib/libcxx/include/__memory/raw_storage_iterator.h new file mode 100644 index 000000000..e8f82b219 --- /dev/null +++ b/lib/libcxx/include/__memory/raw_storage_iterator.h @@ -0,0 +1,73 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___MEMORY_RAW_STORAGE_ITERATOR_H +#define _LIBCPP___MEMORY_RAW_STORAGE_ITERATOR_H + +#include <__config> +#include <__memory/addressof.h> +#include +#include +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +#if _LIBCPP_STD_VER <= 17 || defined(_LIBCPP_ENABLE_CXX20_REMOVED_RAW_STORAGE_ITERATOR) + +_LIBCPP_SUPPRESS_DEPRECATED_PUSH +template +class _LIBCPP_TEMPLATE_VIS _LIBCPP_DEPRECATED_IN_CXX17 raw_storage_iterator +#if _LIBCPP_STD_VER <= 14 || !defined(_LIBCPP_ABI_NO_ITERATOR_BASES) + : public iterator +#endif +{ +_LIBCPP_SUPPRESS_DEPRECATED_POP +private: + _OutputIterator __x_; +public: + typedef output_iterator_tag iterator_category; + typedef void value_type; +#if _LIBCPP_STD_VER > 17 + typedef ptrdiff_t difference_type; +#else + typedef void difference_type; +#endif + typedef void pointer; + typedef void reference; + + _LIBCPP_INLINE_VISIBILITY explicit raw_storage_iterator(_OutputIterator __x) : __x_(__x) {} + _LIBCPP_INLINE_VISIBILITY raw_storage_iterator& operator*() {return *this;} + _LIBCPP_INLINE_VISIBILITY raw_storage_iterator& operator=(const _Tp& __element) + {::new ((void*)_VSTD::addressof(*__x_)) _Tp(__element); return *this;} +#if _LIBCPP_STD_VER >= 14 + _LIBCPP_INLINE_VISIBILITY raw_storage_iterator& operator=(_Tp&& __element) + {::new ((void*)_VSTD::addressof(*__x_)) _Tp(_VSTD::move(__element)); return *this;} +#endif + _LIBCPP_INLINE_VISIBILITY raw_storage_iterator& operator++() {++__x_; return *this;} + _LIBCPP_INLINE_VISIBILITY raw_storage_iterator operator++(int) + {raw_storage_iterator __t(*this); ++__x_; return __t;} +#if _LIBCPP_STD_VER >= 14 + _LIBCPP_INLINE_VISIBILITY _OutputIterator base() const { return __x_; } +#endif +}; + +#endif // _LIBCPP_STD_VER <= 17 || defined(_LIBCPP_ENABLE_CXX20_REMOVED_RAW_STORAGE_ITERATOR) + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___MEMORY_RAW_STORAGE_ITERATOR_H diff --git a/lib/libcxx/include/__memory/shared_ptr.h b/lib/libcxx/include/__memory/shared_ptr.h new file mode 100644 index 000000000..04161c4b7 --- /dev/null +++ b/lib/libcxx/include/__memory/shared_ptr.h @@ -0,0 +1,1879 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___MEMORY_SHARED_PTR_H +#define _LIBCPP___MEMORY_SHARED_PTR_H + +#include <__availability> +#include <__config> +#include <__functional_base> +#include <__functional/binary_function.h> +#include <__functional/operations.h> +#include <__functional/reference_wrapper.h> +#include <__memory/addressof.h> +#include <__memory/allocation_guard.h> +#include <__memory/allocator_traits.h> +#include <__memory/allocator.h> +#include <__memory/compressed_pair.h> +#include <__memory/pointer_traits.h> +#include <__memory/unique_ptr.h> +#include <__utility/forward.h> +#include +#include // abort +#include +#include +#include +#include +#include +#if !defined(_LIBCPP_HAS_NO_ATOMIC_HEADER) +# include +#endif + +#if _LIBCPP_STD_VER <= 14 || defined(_LIBCPP_ENABLE_CXX17_REMOVED_AUTO_PTR) +# include <__memory/auto_ptr.h> +#endif + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +class __allocator_destructor +{ + typedef _LIBCPP_NODEBUG_TYPE allocator_traits<_Alloc> __alloc_traits; +public: + typedef _LIBCPP_NODEBUG_TYPE typename __alloc_traits::pointer pointer; + typedef _LIBCPP_NODEBUG_TYPE typename __alloc_traits::size_type size_type; +private: + _Alloc& __alloc_; + size_type __s_; +public: + _LIBCPP_INLINE_VISIBILITY __allocator_destructor(_Alloc& __a, size_type __s) + _NOEXCEPT + : __alloc_(__a), __s_(__s) {} + _LIBCPP_INLINE_VISIBILITY + void operator()(pointer __p) _NOEXCEPT + {__alloc_traits::deallocate(__alloc_, __p, __s_);} +}; + +// NOTE: Relaxed and acq/rel atomics (for increment and decrement respectively) +// should be sufficient for thread safety. +// See https://llvm.org/PR22803 +#if defined(__clang__) && __has_builtin(__atomic_add_fetch) \ + && defined(__ATOMIC_RELAXED) \ + && defined(__ATOMIC_ACQ_REL) +# define _LIBCPP_HAS_BUILTIN_ATOMIC_SUPPORT +#elif defined(_LIBCPP_COMPILER_GCC) +# define _LIBCPP_HAS_BUILTIN_ATOMIC_SUPPORT +#endif + +template +inline _LIBCPP_INLINE_VISIBILITY +_ValueType __libcpp_relaxed_load(_ValueType const* __value) { +#if !defined(_LIBCPP_HAS_NO_THREADS) && \ + defined(__ATOMIC_RELAXED) && \ + (__has_builtin(__atomic_load_n) || defined(_LIBCPP_COMPILER_GCC)) + return __atomic_load_n(__value, __ATOMIC_RELAXED); +#else + return *__value; +#endif +} + +template +inline _LIBCPP_INLINE_VISIBILITY +_ValueType __libcpp_acquire_load(_ValueType const* __value) { +#if !defined(_LIBCPP_HAS_NO_THREADS) && \ + defined(__ATOMIC_ACQUIRE) && \ + (__has_builtin(__atomic_load_n) || defined(_LIBCPP_COMPILER_GCC)) + return __atomic_load_n(__value, __ATOMIC_ACQUIRE); +#else + return *__value; +#endif +} + +template +inline _LIBCPP_INLINE_VISIBILITY _Tp +__libcpp_atomic_refcount_increment(_Tp& __t) _NOEXCEPT +{ +#if defined(_LIBCPP_HAS_BUILTIN_ATOMIC_SUPPORT) && !defined(_LIBCPP_HAS_NO_THREADS) + return __atomic_add_fetch(&__t, 1, __ATOMIC_RELAXED); +#else + return __t += 1; +#endif +} + +template +inline _LIBCPP_INLINE_VISIBILITY _Tp +__libcpp_atomic_refcount_decrement(_Tp& __t) _NOEXCEPT +{ +#if defined(_LIBCPP_HAS_BUILTIN_ATOMIC_SUPPORT) && !defined(_LIBCPP_HAS_NO_THREADS) + return __atomic_add_fetch(&__t, -1, __ATOMIC_ACQ_REL); +#else + return __t -= 1; +#endif +} + +class _LIBCPP_EXCEPTION_ABI bad_weak_ptr + : public std::exception +{ +public: + bad_weak_ptr() _NOEXCEPT = default; + bad_weak_ptr(const bad_weak_ptr&) _NOEXCEPT = default; + virtual ~bad_weak_ptr() _NOEXCEPT; + virtual const char* what() const _NOEXCEPT; +}; + +_LIBCPP_NORETURN inline _LIBCPP_INLINE_VISIBILITY +void __throw_bad_weak_ptr() +{ +#ifndef _LIBCPP_NO_EXCEPTIONS + throw bad_weak_ptr(); +#else + _VSTD::abort(); +#endif +} + +template class _LIBCPP_TEMPLATE_VIS weak_ptr; + +class _LIBCPP_TYPE_VIS __shared_count +{ + __shared_count(const __shared_count&); + __shared_count& operator=(const __shared_count&); + +protected: + long __shared_owners_; + virtual ~__shared_count(); +private: + virtual void __on_zero_shared() _NOEXCEPT = 0; + +public: + _LIBCPP_INLINE_VISIBILITY + explicit __shared_count(long __refs = 0) _NOEXCEPT + : __shared_owners_(__refs) {} + +#if defined(_LIBCPP_BUILDING_LIBRARY) && \ + defined(_LIBCPP_DEPRECATED_ABI_LEGACY_LIBRARY_DEFINITIONS_FOR_INLINE_FUNCTIONS) + void __add_shared() _NOEXCEPT; + bool __release_shared() _NOEXCEPT; +#else + _LIBCPP_INLINE_VISIBILITY + void __add_shared() _NOEXCEPT { + __libcpp_atomic_refcount_increment(__shared_owners_); + } + _LIBCPP_INLINE_VISIBILITY + bool __release_shared() _NOEXCEPT { + if (__libcpp_atomic_refcount_decrement(__shared_owners_) == -1) { + __on_zero_shared(); + return true; + } + return false; + } +#endif + _LIBCPP_INLINE_VISIBILITY + long use_count() const _NOEXCEPT { + return __libcpp_relaxed_load(&__shared_owners_) + 1; + } +}; + +class _LIBCPP_TYPE_VIS __shared_weak_count + : private __shared_count +{ + long __shared_weak_owners_; + +public: + _LIBCPP_INLINE_VISIBILITY + explicit __shared_weak_count(long __refs = 0) _NOEXCEPT + : __shared_count(__refs), + __shared_weak_owners_(__refs) {} +protected: + virtual ~__shared_weak_count(); + +public: +#if defined(_LIBCPP_BUILDING_LIBRARY) && \ + defined(_LIBCPP_DEPRECATED_ABI_LEGACY_LIBRARY_DEFINITIONS_FOR_INLINE_FUNCTIONS) + void __add_shared() _NOEXCEPT; + void __add_weak() _NOEXCEPT; + void __release_shared() _NOEXCEPT; +#else + _LIBCPP_INLINE_VISIBILITY + void __add_shared() _NOEXCEPT { + __shared_count::__add_shared(); + } + _LIBCPP_INLINE_VISIBILITY + void __add_weak() _NOEXCEPT { + __libcpp_atomic_refcount_increment(__shared_weak_owners_); + } + _LIBCPP_INLINE_VISIBILITY + void __release_shared() _NOEXCEPT { + if (__shared_count::__release_shared()) + __release_weak(); + } +#endif + void __release_weak() _NOEXCEPT; + _LIBCPP_INLINE_VISIBILITY + long use_count() const _NOEXCEPT {return __shared_count::use_count();} + __shared_weak_count* lock() _NOEXCEPT; + + virtual const void* __get_deleter(const type_info&) const _NOEXCEPT; +private: + virtual void __on_zero_shared_weak() _NOEXCEPT = 0; +}; + +template +class __shared_ptr_pointer + : public __shared_weak_count +{ + __compressed_pair<__compressed_pair<_Tp, _Dp>, _Alloc> __data_; +public: + _LIBCPP_INLINE_VISIBILITY + __shared_ptr_pointer(_Tp __p, _Dp __d, _Alloc __a) + : __data_(__compressed_pair<_Tp, _Dp>(__p, _VSTD::move(__d)), _VSTD::move(__a)) {} + +#ifndef _LIBCPP_NO_RTTI + virtual const void* __get_deleter(const type_info&) const _NOEXCEPT; +#endif + +private: + virtual void __on_zero_shared() _NOEXCEPT; + virtual void __on_zero_shared_weak() _NOEXCEPT; +}; + +#ifndef _LIBCPP_NO_RTTI + +template +const void* +__shared_ptr_pointer<_Tp, _Dp, _Alloc>::__get_deleter(const type_info& __t) const _NOEXCEPT +{ + return __t == typeid(_Dp) ? _VSTD::addressof(__data_.first().second()) : nullptr; +} + +#endif // _LIBCPP_NO_RTTI + +template +void +__shared_ptr_pointer<_Tp, _Dp, _Alloc>::__on_zero_shared() _NOEXCEPT +{ + __data_.first().second()(__data_.first().first()); + __data_.first().second().~_Dp(); +} + +template +void +__shared_ptr_pointer<_Tp, _Dp, _Alloc>::__on_zero_shared_weak() _NOEXCEPT +{ + typedef typename __allocator_traits_rebind<_Alloc, __shared_ptr_pointer>::type _Al; + typedef allocator_traits<_Al> _ATraits; + typedef pointer_traits _PTraits; + + _Al __a(__data_.second()); + __data_.second().~_Alloc(); + __a.deallocate(_PTraits::pointer_to(*this), 1); +} + +template +struct __shared_ptr_emplace + : __shared_weak_count +{ + template + _LIBCPP_HIDE_FROM_ABI + explicit __shared_ptr_emplace(_Alloc __a, _Args&& ...__args) + : __storage_(_VSTD::move(__a)) + { +#if _LIBCPP_STD_VER > 17 + using _TpAlloc = typename __allocator_traits_rebind<_Alloc, _Tp>::type; + _TpAlloc __tmp(*__get_alloc()); + allocator_traits<_TpAlloc>::construct(__tmp, __get_elem(), _VSTD::forward<_Args>(__args)...); +#else + ::new ((void*)__get_elem()) _Tp(_VSTD::forward<_Args>(__args)...); +#endif + } + + _LIBCPP_HIDE_FROM_ABI + _Alloc* __get_alloc() _NOEXCEPT { return __storage_.__get_alloc(); } + + _LIBCPP_HIDE_FROM_ABI + _Tp* __get_elem() _NOEXCEPT { return __storage_.__get_elem(); } + +private: + virtual void __on_zero_shared() _NOEXCEPT { +#if _LIBCPP_STD_VER > 17 + using _TpAlloc = typename __allocator_traits_rebind<_Alloc, _Tp>::type; + _TpAlloc __tmp(*__get_alloc()); + allocator_traits<_TpAlloc>::destroy(__tmp, __get_elem()); +#else + __get_elem()->~_Tp(); +#endif + } + + virtual void __on_zero_shared_weak() _NOEXCEPT { + using _ControlBlockAlloc = typename __allocator_traits_rebind<_Alloc, __shared_ptr_emplace>::type; + using _ControlBlockPointer = typename allocator_traits<_ControlBlockAlloc>::pointer; + _ControlBlockAlloc __tmp(*__get_alloc()); + __storage_.~_Storage(); + allocator_traits<_ControlBlockAlloc>::deallocate(__tmp, + pointer_traits<_ControlBlockPointer>::pointer_to(*this), 1); + } + + // This class implements the control block for non-array shared pointers created + // through `std::allocate_shared` and `std::make_shared`. + // + // In previous versions of the library, we used a compressed pair to store + // both the _Alloc and the _Tp. This implies using EBO, which is incompatible + // with Allocator construction for _Tp. To allow implementing P0674 in C++20, + // we now use a properly aligned char buffer while making sure that we maintain + // the same layout that we had when we used a compressed pair. + using _CompressedPair = __compressed_pair<_Alloc, _Tp>; + struct _ALIGNAS_TYPE(_CompressedPair) _Storage { + char __blob_[sizeof(_CompressedPair)]; + + _LIBCPP_HIDE_FROM_ABI explicit _Storage(_Alloc&& __a) { + ::new ((void*)__get_alloc()) _Alloc(_VSTD::move(__a)); + } + _LIBCPP_HIDE_FROM_ABI ~_Storage() { + __get_alloc()->~_Alloc(); + } + _Alloc* __get_alloc() _NOEXCEPT { + _CompressedPair *__as_pair = reinterpret_cast<_CompressedPair*>(__blob_); + typename _CompressedPair::_Base1* __first = _CompressedPair::__get_first_base(__as_pair); + _Alloc *__alloc = reinterpret_cast<_Alloc*>(__first); + return __alloc; + } + _LIBCPP_NO_CFI _Tp* __get_elem() _NOEXCEPT { + _CompressedPair *__as_pair = reinterpret_cast<_CompressedPair*>(__blob_); + typename _CompressedPair::_Base2* __second = _CompressedPair::__get_second_base(__as_pair); + _Tp *__elem = reinterpret_cast<_Tp*>(__second); + return __elem; + } + }; + + static_assert(_LIBCPP_ALIGNOF(_Storage) == _LIBCPP_ALIGNOF(_CompressedPair), ""); + static_assert(sizeof(_Storage) == sizeof(_CompressedPair), ""); + _Storage __storage_; +}; + +struct __shared_ptr_dummy_rebind_allocator_type; +template <> +class _LIBCPP_TEMPLATE_VIS allocator<__shared_ptr_dummy_rebind_allocator_type> +{ +public: + template + struct rebind + { + typedef allocator<_Other> other; + }; +}; + +template class _LIBCPP_TEMPLATE_VIS enable_shared_from_this; + +template +struct __compatible_with +#if _LIBCPP_STD_VER > 14 + : is_convertible*, remove_extent_t<_Up>*> {}; +#else + : is_convertible<_Tp*, _Up*> {}; +#endif // _LIBCPP_STD_VER > 14 + +template +struct __is_deletable : false_type { }; +template +struct __is_deletable<_Ptr, decltype(delete declval<_Ptr>())> : true_type { }; + +template +struct __is_array_deletable : false_type { }; +template +struct __is_array_deletable<_Ptr, decltype(delete[] declval<_Ptr>())> : true_type { }; + +template ()(declval<_Pt>()))> +static true_type __well_formed_deleter_test(int); + +template +static false_type __well_formed_deleter_test(...); + +template +struct __well_formed_deleter : decltype(__well_formed_deleter_test<_Dp, _Pt>(0)) {}; + +template +struct __shared_ptr_deleter_ctor_reqs +{ + static const bool value = __compatible_with<_Tp, _Yp>::value && + is_move_constructible<_Dp>::value && + __well_formed_deleter<_Dp, _Tp*>::value; +}; + +#if defined(_LIBCPP_ABI_ENABLE_SHARED_PTR_TRIVIAL_ABI) +# define _LIBCPP_SHARED_PTR_TRIVIAL_ABI __attribute__((trivial_abi)) +#else +# define _LIBCPP_SHARED_PTR_TRIVIAL_ABI +#endif + +template +class _LIBCPP_SHARED_PTR_TRIVIAL_ABI _LIBCPP_TEMPLATE_VIS shared_ptr +{ +public: +#if _LIBCPP_STD_VER > 14 + typedef weak_ptr<_Tp> weak_type; + typedef remove_extent_t<_Tp> element_type; +#else + typedef _Tp element_type; +#endif + +private: + element_type* __ptr_; + __shared_weak_count* __cntrl_; + + struct __nat {int __for_bool_;}; +public: + _LIBCPP_INLINE_VISIBILITY + _LIBCPP_CONSTEXPR shared_ptr() _NOEXCEPT; + _LIBCPP_INLINE_VISIBILITY + _LIBCPP_CONSTEXPR shared_ptr(nullptr_t) _NOEXCEPT; + + template + // In C++03 we get errors when trying to do SFINAE with the + // delete operator, so we always pretend that it's deletable. + // The same happens on GCC. +#if !defined(_LIBCPP_CXX03_LANG) && !defined(_LIBCPP_COMPILER_GCC) + , _If::value, __is_array_deletable<_Yp*>, __is_deletable<_Yp*> > +#endif + >::value + > > + explicit shared_ptr(_Yp* __p) : __ptr_(__p) { + unique_ptr<_Yp> __hold(__p); + typedef typename __shared_ptr_default_allocator<_Yp>::type _AllocT; + typedef __shared_ptr_pointer<_Yp*, __shared_ptr_default_delete<_Tp, _Yp>, _AllocT > _CntrlBlk; + __cntrl_ = new _CntrlBlk(__p, __shared_ptr_default_delete<_Tp, _Yp>(), _AllocT()); + __hold.release(); + __enable_weak_this(__p, __p); + } + + template + shared_ptr(_Yp* __p, _Dp __d, + typename enable_if<__shared_ptr_deleter_ctor_reqs<_Dp, _Yp, element_type>::value, __nat>::type = __nat()); + template + shared_ptr(_Yp* __p, _Dp __d, _Alloc __a, + typename enable_if<__shared_ptr_deleter_ctor_reqs<_Dp, _Yp, element_type>::value, __nat>::type = __nat()); + template shared_ptr(nullptr_t __p, _Dp __d); + template shared_ptr(nullptr_t __p, _Dp __d, _Alloc __a); + template _LIBCPP_INLINE_VISIBILITY shared_ptr(const shared_ptr<_Yp>& __r, element_type* __p) _NOEXCEPT; + _LIBCPP_INLINE_VISIBILITY + shared_ptr(const shared_ptr& __r) _NOEXCEPT; + template + _LIBCPP_INLINE_VISIBILITY + shared_ptr(const shared_ptr<_Yp>& __r, + typename enable_if<__compatible_with<_Yp, element_type>::value, __nat>::type = __nat()) + _NOEXCEPT; + _LIBCPP_INLINE_VISIBILITY + shared_ptr(shared_ptr&& __r) _NOEXCEPT; + template _LIBCPP_INLINE_VISIBILITY shared_ptr(shared_ptr<_Yp>&& __r, + typename enable_if<__compatible_with<_Yp, element_type>::value, __nat>::type = __nat()) + _NOEXCEPT; + template explicit shared_ptr(const weak_ptr<_Yp>& __r, + typename enable_if::value, __nat>::type= __nat()); +#if _LIBCPP_STD_VER <= 14 || defined(_LIBCPP_ENABLE_CXX17_REMOVED_AUTO_PTR) + template + shared_ptr(auto_ptr<_Yp>&& __r, + typename enable_if::value, __nat>::type = __nat()); +#endif + template + shared_ptr(unique_ptr<_Yp, _Dp>&&, + typename enable_if + < + !is_lvalue_reference<_Dp>::value && + is_convertible::pointer, element_type*>::value, + __nat + >::type = __nat()); + template + shared_ptr(unique_ptr<_Yp, _Dp>&&, + typename enable_if + < + is_lvalue_reference<_Dp>::value && + is_convertible::pointer, element_type*>::value, + __nat + >::type = __nat()); + + ~shared_ptr(); + + _LIBCPP_INLINE_VISIBILITY + shared_ptr& operator=(const shared_ptr& __r) _NOEXCEPT; + template + typename enable_if + < + __compatible_with<_Yp, element_type>::value, + shared_ptr& + >::type + _LIBCPP_INLINE_VISIBILITY + operator=(const shared_ptr<_Yp>& __r) _NOEXCEPT; + _LIBCPP_INLINE_VISIBILITY + shared_ptr& operator=(shared_ptr&& __r) _NOEXCEPT; + template + typename enable_if + < + __compatible_with<_Yp, element_type>::value, + shared_ptr& + >::type + _LIBCPP_INLINE_VISIBILITY + operator=(shared_ptr<_Yp>&& __r); +#if _LIBCPP_STD_VER <= 14 || defined(_LIBCPP_ENABLE_CXX17_REMOVED_AUTO_PTR) + template + _LIBCPP_INLINE_VISIBILITY + typename enable_if + < + !is_array<_Yp>::value && + is_convertible<_Yp*, element_type*>::value, + shared_ptr + >::type& + operator=(auto_ptr<_Yp>&& __r); +#endif + template + typename enable_if + < + is_convertible::pointer, element_type*>::value, + shared_ptr& + >::type + _LIBCPP_INLINE_VISIBILITY + operator=(unique_ptr<_Yp, _Dp>&& __r); + + _LIBCPP_INLINE_VISIBILITY + void swap(shared_ptr& __r) _NOEXCEPT; + _LIBCPP_INLINE_VISIBILITY + void reset() _NOEXCEPT; + template + typename enable_if + < + __compatible_with<_Yp, element_type>::value, + void + >::type + _LIBCPP_INLINE_VISIBILITY + reset(_Yp* __p); + template + typename enable_if + < + __compatible_with<_Yp, element_type>::value, + void + >::type + _LIBCPP_INLINE_VISIBILITY + reset(_Yp* __p, _Dp __d); + template + typename enable_if + < + __compatible_with<_Yp, element_type>::value, + void + >::type + _LIBCPP_INLINE_VISIBILITY + reset(_Yp* __p, _Dp __d, _Alloc __a); + + _LIBCPP_INLINE_VISIBILITY + element_type* get() const _NOEXCEPT {return __ptr_;} + _LIBCPP_INLINE_VISIBILITY + typename add_lvalue_reference::type operator*() const _NOEXCEPT + {return *__ptr_;} + _LIBCPP_INLINE_VISIBILITY + element_type* operator->() const _NOEXCEPT + { + static_assert(!is_array<_Tp>::value, + "std::shared_ptr::operator-> is only valid when T is not an array type."); + return __ptr_; + } + _LIBCPP_INLINE_VISIBILITY + long use_count() const _NOEXCEPT {return __cntrl_ ? __cntrl_->use_count() : 0;} + _LIBCPP_INLINE_VISIBILITY + bool unique() const _NOEXCEPT {return use_count() == 1;} + _LIBCPP_INLINE_VISIBILITY + explicit operator bool() const _NOEXCEPT {return get() != nullptr;} + template + _LIBCPP_INLINE_VISIBILITY + bool owner_before(shared_ptr<_Up> const& __p) const _NOEXCEPT + {return __cntrl_ < __p.__cntrl_;} + template + _LIBCPP_INLINE_VISIBILITY + bool owner_before(weak_ptr<_Up> const& __p) const _NOEXCEPT + {return __cntrl_ < __p.__cntrl_;} + _LIBCPP_INLINE_VISIBILITY + bool + __owner_equivalent(const shared_ptr& __p) const + {return __cntrl_ == __p.__cntrl_;} + +#if _LIBCPP_STD_VER > 14 + typename add_lvalue_reference::type + _LIBCPP_INLINE_VISIBILITY + operator[](ptrdiff_t __i) const + { + static_assert(is_array<_Tp>::value, + "std::shared_ptr::operator[] is only valid when T is an array type."); + return __ptr_[__i]; + } +#endif + +#ifndef _LIBCPP_NO_RTTI + template + _LIBCPP_INLINE_VISIBILITY + _Dp* __get_deleter() const _NOEXCEPT + {return static_cast<_Dp*>(__cntrl_ + ? const_cast(__cntrl_->__get_deleter(typeid(_Dp))) + : nullptr);} +#endif // _LIBCPP_NO_RTTI + + template + static shared_ptr<_Tp> + __create_with_control_block(_Yp* __p, _CntrlBlk* __cntrl) _NOEXCEPT + { + shared_ptr<_Tp> __r; + __r.__ptr_ = __p; + __r.__cntrl_ = __cntrl; + __r.__enable_weak_this(__r.__ptr_, __r.__ptr_); + return __r; + } + +private: + template ::value> + struct __shared_ptr_default_allocator + { + typedef allocator<_Yp> type; + }; + + template + struct __shared_ptr_default_allocator<_Yp, true> + { + typedef allocator<__shared_ptr_dummy_rebind_allocator_type> type; + }; + + template + _LIBCPP_INLINE_VISIBILITY + typename enable_if* + >::value, + void>::type + __enable_weak_this(const enable_shared_from_this<_Yp>* __e, + _OrigPtr* __ptr) _NOEXCEPT + { + typedef typename remove_cv<_Yp>::type _RawYp; + if (__e && __e->__weak_this_.expired()) + { + __e->__weak_this_ = shared_ptr<_RawYp>(*this, + const_cast<_RawYp*>(static_cast(__ptr))); + } + } + + _LIBCPP_INLINE_VISIBILITY void __enable_weak_this(...) _NOEXCEPT {} + + template + struct __shared_ptr_default_delete + : default_delete<_Yp> {}; + + template + struct __shared_ptr_default_delete<_Yp[_Sz], _Un> + : default_delete<_Yp[]> {}; + + template + struct __shared_ptr_default_delete<_Yp[], _Un> + : default_delete<_Yp[]> {}; + + template friend class _LIBCPP_TEMPLATE_VIS shared_ptr; + template friend class _LIBCPP_TEMPLATE_VIS weak_ptr; +}; + +#ifndef _LIBCPP_HAS_NO_DEDUCTION_GUIDES +template +shared_ptr(weak_ptr<_Tp>) -> shared_ptr<_Tp>; +template +shared_ptr(unique_ptr<_Tp, _Dp>) -> shared_ptr<_Tp>; +#endif + +template +inline +_LIBCPP_CONSTEXPR +shared_ptr<_Tp>::shared_ptr() _NOEXCEPT + : __ptr_(nullptr), + __cntrl_(nullptr) +{ +} + +template +inline +_LIBCPP_CONSTEXPR +shared_ptr<_Tp>::shared_ptr(nullptr_t) _NOEXCEPT + : __ptr_(nullptr), + __cntrl_(nullptr) +{ +} + +template +template +shared_ptr<_Tp>::shared_ptr(_Yp* __p, _Dp __d, + typename enable_if<__shared_ptr_deleter_ctor_reqs<_Dp, _Yp, element_type>::value, __nat>::type) + : __ptr_(__p) +{ +#ifndef _LIBCPP_NO_EXCEPTIONS + try + { +#endif // _LIBCPP_NO_EXCEPTIONS + typedef typename __shared_ptr_default_allocator<_Yp>::type _AllocT; + typedef __shared_ptr_pointer<_Yp*, _Dp, _AllocT > _CntrlBlk; +#ifndef _LIBCPP_CXX03_LANG + __cntrl_ = new _CntrlBlk(__p, _VSTD::move(__d), _AllocT()); +#else + __cntrl_ = new _CntrlBlk(__p, __d, _AllocT()); +#endif // not _LIBCPP_CXX03_LANG + __enable_weak_this(__p, __p); +#ifndef _LIBCPP_NO_EXCEPTIONS + } + catch (...) + { + __d(__p); + throw; + } +#endif // _LIBCPP_NO_EXCEPTIONS +} + +template +template +shared_ptr<_Tp>::shared_ptr(nullptr_t __p, _Dp __d) + : __ptr_(nullptr) +{ +#ifndef _LIBCPP_NO_EXCEPTIONS + try + { +#endif // _LIBCPP_NO_EXCEPTIONS + typedef typename __shared_ptr_default_allocator<_Tp>::type _AllocT; + typedef __shared_ptr_pointer _CntrlBlk; +#ifndef _LIBCPP_CXX03_LANG + __cntrl_ = new _CntrlBlk(__p, _VSTD::move(__d), _AllocT()); +#else + __cntrl_ = new _CntrlBlk(__p, __d, _AllocT()); +#endif // not _LIBCPP_CXX03_LANG +#ifndef _LIBCPP_NO_EXCEPTIONS + } + catch (...) + { + __d(__p); + throw; + } +#endif // _LIBCPP_NO_EXCEPTIONS +} + +template +template +shared_ptr<_Tp>::shared_ptr(_Yp* __p, _Dp __d, _Alloc __a, + typename enable_if<__shared_ptr_deleter_ctor_reqs<_Dp, _Yp, element_type>::value, __nat>::type) + : __ptr_(__p) +{ +#ifndef _LIBCPP_NO_EXCEPTIONS + try + { +#endif // _LIBCPP_NO_EXCEPTIONS + typedef __shared_ptr_pointer<_Yp*, _Dp, _Alloc> _CntrlBlk; + typedef typename __allocator_traits_rebind<_Alloc, _CntrlBlk>::type _A2; + typedef __allocator_destructor<_A2> _D2; + _A2 __a2(__a); + unique_ptr<_CntrlBlk, _D2> __hold2(__a2.allocate(1), _D2(__a2, 1)); + ::new ((void*)_VSTD::addressof(*__hold2.get())) +#ifndef _LIBCPP_CXX03_LANG + _CntrlBlk(__p, _VSTD::move(__d), __a); +#else + _CntrlBlk(__p, __d, __a); +#endif // not _LIBCPP_CXX03_LANG + __cntrl_ = _VSTD::addressof(*__hold2.release()); + __enable_weak_this(__p, __p); +#ifndef _LIBCPP_NO_EXCEPTIONS + } + catch (...) + { + __d(__p); + throw; + } +#endif // _LIBCPP_NO_EXCEPTIONS +} + +template +template +shared_ptr<_Tp>::shared_ptr(nullptr_t __p, _Dp __d, _Alloc __a) + : __ptr_(nullptr) +{ +#ifndef _LIBCPP_NO_EXCEPTIONS + try + { +#endif // _LIBCPP_NO_EXCEPTIONS + typedef __shared_ptr_pointer _CntrlBlk; + typedef typename __allocator_traits_rebind<_Alloc, _CntrlBlk>::type _A2; + typedef __allocator_destructor<_A2> _D2; + _A2 __a2(__a); + unique_ptr<_CntrlBlk, _D2> __hold2(__a2.allocate(1), _D2(__a2, 1)); + ::new ((void*)_VSTD::addressof(*__hold2.get())) +#ifndef _LIBCPP_CXX03_LANG + _CntrlBlk(__p, _VSTD::move(__d), __a); +#else + _CntrlBlk(__p, __d, __a); +#endif // not _LIBCPP_CXX03_LANG + __cntrl_ = _VSTD::addressof(*__hold2.release()); +#ifndef _LIBCPP_NO_EXCEPTIONS + } + catch (...) + { + __d(__p); + throw; + } +#endif // _LIBCPP_NO_EXCEPTIONS +} + +template +template +inline +shared_ptr<_Tp>::shared_ptr(const shared_ptr<_Yp>& __r, element_type *__p) _NOEXCEPT + : __ptr_(__p), + __cntrl_(__r.__cntrl_) +{ + if (__cntrl_) + __cntrl_->__add_shared(); +} + +template +inline +shared_ptr<_Tp>::shared_ptr(const shared_ptr& __r) _NOEXCEPT + : __ptr_(__r.__ptr_), + __cntrl_(__r.__cntrl_) +{ + if (__cntrl_) + __cntrl_->__add_shared(); +} + +template +template +inline +shared_ptr<_Tp>::shared_ptr(const shared_ptr<_Yp>& __r, + typename enable_if<__compatible_with<_Yp, element_type>::value, __nat>::type) + _NOEXCEPT + : __ptr_(__r.__ptr_), + __cntrl_(__r.__cntrl_) +{ + if (__cntrl_) + __cntrl_->__add_shared(); +} + +template +inline +shared_ptr<_Tp>::shared_ptr(shared_ptr&& __r) _NOEXCEPT + : __ptr_(__r.__ptr_), + __cntrl_(__r.__cntrl_) +{ + __r.__ptr_ = nullptr; + __r.__cntrl_ = nullptr; +} + +template +template +inline +shared_ptr<_Tp>::shared_ptr(shared_ptr<_Yp>&& __r, + typename enable_if<__compatible_with<_Yp, element_type>::value, __nat>::type) + _NOEXCEPT + : __ptr_(__r.__ptr_), + __cntrl_(__r.__cntrl_) +{ + __r.__ptr_ = nullptr; + __r.__cntrl_ = nullptr; +} + +#if _LIBCPP_STD_VER <= 14 || defined(_LIBCPP_ENABLE_CXX17_REMOVED_AUTO_PTR) +template +template +shared_ptr<_Tp>::shared_ptr(auto_ptr<_Yp>&& __r, + typename enable_if::value, __nat>::type) + : __ptr_(__r.get()) +{ + typedef __shared_ptr_pointer<_Yp*, default_delete<_Yp>, allocator<_Yp> > _CntrlBlk; + __cntrl_ = new _CntrlBlk(__r.get(), default_delete<_Yp>(), allocator<_Yp>()); + __enable_weak_this(__r.get(), __r.get()); + __r.release(); +} +#endif + +template +template +shared_ptr<_Tp>::shared_ptr(unique_ptr<_Yp, _Dp>&& __r, + typename enable_if + < + !is_lvalue_reference<_Dp>::value && + is_convertible::pointer, element_type*>::value, + __nat + >::type) + : __ptr_(__r.get()) +{ +#if _LIBCPP_STD_VER > 11 + if (__ptr_ == nullptr) + __cntrl_ = nullptr; + else +#endif + { + typedef typename __shared_ptr_default_allocator<_Yp>::type _AllocT; + typedef __shared_ptr_pointer::pointer, _Dp, _AllocT > _CntrlBlk; + __cntrl_ = new _CntrlBlk(__r.get(), __r.get_deleter(), _AllocT()); + __enable_weak_this(__r.get(), __r.get()); + } + __r.release(); +} + +template +template +shared_ptr<_Tp>::shared_ptr(unique_ptr<_Yp, _Dp>&& __r, + typename enable_if + < + is_lvalue_reference<_Dp>::value && + is_convertible::pointer, element_type*>::value, + __nat + >::type) + : __ptr_(__r.get()) +{ +#if _LIBCPP_STD_VER > 11 + if (__ptr_ == nullptr) + __cntrl_ = nullptr; + else +#endif + { + typedef typename __shared_ptr_default_allocator<_Yp>::type _AllocT; + typedef __shared_ptr_pointer::pointer, + reference_wrapper::type>, + _AllocT > _CntrlBlk; + __cntrl_ = new _CntrlBlk(__r.get(), _VSTD::ref(__r.get_deleter()), _AllocT()); + __enable_weak_this(__r.get(), __r.get()); + } + __r.release(); +} + +template +shared_ptr<_Tp>::~shared_ptr() +{ + if (__cntrl_) + __cntrl_->__release_shared(); +} + +template +inline +shared_ptr<_Tp>& +shared_ptr<_Tp>::operator=(const shared_ptr& __r) _NOEXCEPT +{ + shared_ptr(__r).swap(*this); + return *this; +} + +template +template +inline +typename enable_if +< + __compatible_with<_Yp, typename shared_ptr<_Tp>::element_type>::value, + shared_ptr<_Tp>& +>::type +shared_ptr<_Tp>::operator=(const shared_ptr<_Yp>& __r) _NOEXCEPT +{ + shared_ptr(__r).swap(*this); + return *this; +} + +template +inline +shared_ptr<_Tp>& +shared_ptr<_Tp>::operator=(shared_ptr&& __r) _NOEXCEPT +{ + shared_ptr(_VSTD::move(__r)).swap(*this); + return *this; +} + +template +template +inline +typename enable_if +< + __compatible_with<_Yp, typename shared_ptr<_Tp>::element_type>::value, + shared_ptr<_Tp>& +>::type +shared_ptr<_Tp>::operator=(shared_ptr<_Yp>&& __r) +{ + shared_ptr(_VSTD::move(__r)).swap(*this); + return *this; +} + +#if _LIBCPP_STD_VER <= 14 || defined(_LIBCPP_ENABLE_CXX17_REMOVED_AUTO_PTR) +template +template +inline +typename enable_if +< + !is_array<_Yp>::value && + is_convertible<_Yp*, typename shared_ptr<_Tp>::element_type*>::value, + shared_ptr<_Tp> +>::type& +shared_ptr<_Tp>::operator=(auto_ptr<_Yp>&& __r) +{ + shared_ptr(_VSTD::move(__r)).swap(*this); + return *this; +} +#endif + +template +template +inline +typename enable_if +< + is_convertible::pointer, + typename shared_ptr<_Tp>::element_type*>::value, + shared_ptr<_Tp>& +>::type +shared_ptr<_Tp>::operator=(unique_ptr<_Yp, _Dp>&& __r) +{ + shared_ptr(_VSTD::move(__r)).swap(*this); + return *this; +} + +template +inline +void +shared_ptr<_Tp>::swap(shared_ptr& __r) _NOEXCEPT +{ + _VSTD::swap(__ptr_, __r.__ptr_); + _VSTD::swap(__cntrl_, __r.__cntrl_); +} + +template +inline +void +shared_ptr<_Tp>::reset() _NOEXCEPT +{ + shared_ptr().swap(*this); +} + +template +template +inline +typename enable_if +< + __compatible_with<_Yp, typename shared_ptr<_Tp>::element_type>::value, + void +>::type +shared_ptr<_Tp>::reset(_Yp* __p) +{ + shared_ptr(__p).swap(*this); +} + +template +template +inline +typename enable_if +< + __compatible_with<_Yp, typename shared_ptr<_Tp>::element_type>::value, + void +>::type +shared_ptr<_Tp>::reset(_Yp* __p, _Dp __d) +{ + shared_ptr(__p, __d).swap(*this); +} + +template +template +inline +typename enable_if +< + __compatible_with<_Yp, typename shared_ptr<_Tp>::element_type>::value, + void +>::type +shared_ptr<_Tp>::reset(_Yp* __p, _Dp __d, _Alloc __a) +{ + shared_ptr(__p, __d, __a).swap(*this); +} + +// +// std::allocate_shared and std::make_shared +// +template::value> > +_LIBCPP_HIDE_FROM_ABI +shared_ptr<_Tp> allocate_shared(const _Alloc& __a, _Args&& ...__args) +{ + using _ControlBlock = __shared_ptr_emplace<_Tp, _Alloc>; + using _ControlBlockAllocator = typename __allocator_traits_rebind<_Alloc, _ControlBlock>::type; + __allocation_guard<_ControlBlockAllocator> __guard(__a, 1); + ::new ((void*)_VSTD::addressof(*__guard.__get())) _ControlBlock(__a, _VSTD::forward<_Args>(__args)...); + auto __control_block = __guard.__release_ptr(); + return shared_ptr<_Tp>::__create_with_control_block((*__control_block).__get_elem(), _VSTD::addressof(*__control_block)); +} + +template::value> > +_LIBCPP_HIDE_FROM_ABI +shared_ptr<_Tp> make_shared(_Args&& ...__args) +{ + return _VSTD::allocate_shared<_Tp>(allocator<_Tp>(), _VSTD::forward<_Args>(__args)...); +} + +template +inline _LIBCPP_INLINE_VISIBILITY +bool +operator==(const shared_ptr<_Tp>& __x, const shared_ptr<_Up>& __y) _NOEXCEPT +{ + return __x.get() == __y.get(); +} + +template +inline _LIBCPP_INLINE_VISIBILITY +bool +operator!=(const shared_ptr<_Tp>& __x, const shared_ptr<_Up>& __y) _NOEXCEPT +{ + return !(__x == __y); +} + +template +inline _LIBCPP_INLINE_VISIBILITY +bool +operator<(const shared_ptr<_Tp>& __x, const shared_ptr<_Up>& __y) _NOEXCEPT +{ +#if _LIBCPP_STD_VER <= 11 + typedef typename common_type<_Tp*, _Up*>::type _Vp; + return less<_Vp>()(__x.get(), __y.get()); +#else + return less<>()(__x.get(), __y.get()); +#endif + +} + +template +inline _LIBCPP_INLINE_VISIBILITY +bool +operator>(const shared_ptr<_Tp>& __x, const shared_ptr<_Up>& __y) _NOEXCEPT +{ + return __y < __x; +} + +template +inline _LIBCPP_INLINE_VISIBILITY +bool +operator<=(const shared_ptr<_Tp>& __x, const shared_ptr<_Up>& __y) _NOEXCEPT +{ + return !(__y < __x); +} + +template +inline _LIBCPP_INLINE_VISIBILITY +bool +operator>=(const shared_ptr<_Tp>& __x, const shared_ptr<_Up>& __y) _NOEXCEPT +{ + return !(__x < __y); +} + +template +inline _LIBCPP_INLINE_VISIBILITY +bool +operator==(const shared_ptr<_Tp>& __x, nullptr_t) _NOEXCEPT +{ + return !__x; +} + +template +inline _LIBCPP_INLINE_VISIBILITY +bool +operator==(nullptr_t, const shared_ptr<_Tp>& __x) _NOEXCEPT +{ + return !__x; +} + +template +inline _LIBCPP_INLINE_VISIBILITY +bool +operator!=(const shared_ptr<_Tp>& __x, nullptr_t) _NOEXCEPT +{ + return static_cast(__x); +} + +template +inline _LIBCPP_INLINE_VISIBILITY +bool +operator!=(nullptr_t, const shared_ptr<_Tp>& __x) _NOEXCEPT +{ + return static_cast(__x); +} + +template +inline _LIBCPP_INLINE_VISIBILITY +bool +operator<(const shared_ptr<_Tp>& __x, nullptr_t) _NOEXCEPT +{ + return less<_Tp*>()(__x.get(), nullptr); +} + +template +inline _LIBCPP_INLINE_VISIBILITY +bool +operator<(nullptr_t, const shared_ptr<_Tp>& __x) _NOEXCEPT +{ + return less<_Tp*>()(nullptr, __x.get()); +} + +template +inline _LIBCPP_INLINE_VISIBILITY +bool +operator>(const shared_ptr<_Tp>& __x, nullptr_t) _NOEXCEPT +{ + return nullptr < __x; +} + +template +inline _LIBCPP_INLINE_VISIBILITY +bool +operator>(nullptr_t, const shared_ptr<_Tp>& __x) _NOEXCEPT +{ + return __x < nullptr; +} + +template +inline _LIBCPP_INLINE_VISIBILITY +bool +operator<=(const shared_ptr<_Tp>& __x, nullptr_t) _NOEXCEPT +{ + return !(nullptr < __x); +} + +template +inline _LIBCPP_INLINE_VISIBILITY +bool +operator<=(nullptr_t, const shared_ptr<_Tp>& __x) _NOEXCEPT +{ + return !(__x < nullptr); +} + +template +inline _LIBCPP_INLINE_VISIBILITY +bool +operator>=(const shared_ptr<_Tp>& __x, nullptr_t) _NOEXCEPT +{ + return !(__x < nullptr); +} + +template +inline _LIBCPP_INLINE_VISIBILITY +bool +operator>=(nullptr_t, const shared_ptr<_Tp>& __x) _NOEXCEPT +{ + return !(nullptr < __x); +} + +template +inline _LIBCPP_INLINE_VISIBILITY +void +swap(shared_ptr<_Tp>& __x, shared_ptr<_Tp>& __y) _NOEXCEPT +{ + __x.swap(__y); +} + +template +inline _LIBCPP_INLINE_VISIBILITY +shared_ptr<_Tp> +static_pointer_cast(const shared_ptr<_Up>& __r) _NOEXCEPT +{ + return shared_ptr<_Tp>(__r, + static_cast< + typename shared_ptr<_Tp>::element_type*>(__r.get())); +} + +template +inline _LIBCPP_INLINE_VISIBILITY +shared_ptr<_Tp> +dynamic_pointer_cast(const shared_ptr<_Up>& __r) _NOEXCEPT +{ + typedef typename shared_ptr<_Tp>::element_type _ET; + _ET* __p = dynamic_cast<_ET*>(__r.get()); + return __p ? shared_ptr<_Tp>(__r, __p) : shared_ptr<_Tp>(); +} + +template +shared_ptr<_Tp> +const_pointer_cast(const shared_ptr<_Up>& __r) _NOEXCEPT +{ + typedef typename shared_ptr<_Tp>::element_type _RTp; + return shared_ptr<_Tp>(__r, const_cast<_RTp*>(__r.get())); +} + +template +shared_ptr<_Tp> +reinterpret_pointer_cast(const shared_ptr<_Up>& __r) _NOEXCEPT +{ + return shared_ptr<_Tp>(__r, + reinterpret_cast< + typename shared_ptr<_Tp>::element_type*>(__r.get())); +} + +#ifndef _LIBCPP_NO_RTTI + +template +inline _LIBCPP_INLINE_VISIBILITY +_Dp* +get_deleter(const shared_ptr<_Tp>& __p) _NOEXCEPT +{ + return __p.template __get_deleter<_Dp>(); +} + +#endif // _LIBCPP_NO_RTTI + +template +class _LIBCPP_SHARED_PTR_TRIVIAL_ABI _LIBCPP_TEMPLATE_VIS weak_ptr +{ +public: + typedef _Tp element_type; +private: + element_type* __ptr_; + __shared_weak_count* __cntrl_; + +public: + _LIBCPP_INLINE_VISIBILITY + _LIBCPP_CONSTEXPR weak_ptr() _NOEXCEPT; + template _LIBCPP_INLINE_VISIBILITY weak_ptr(shared_ptr<_Yp> const& __r, + typename enable_if::value, __nat*>::type = 0) + _NOEXCEPT; + _LIBCPP_INLINE_VISIBILITY + weak_ptr(weak_ptr const& __r) _NOEXCEPT; + template _LIBCPP_INLINE_VISIBILITY weak_ptr(weak_ptr<_Yp> const& __r, + typename enable_if::value, __nat*>::type = 0) + _NOEXCEPT; + + _LIBCPP_INLINE_VISIBILITY + weak_ptr(weak_ptr&& __r) _NOEXCEPT; + template _LIBCPP_INLINE_VISIBILITY weak_ptr(weak_ptr<_Yp>&& __r, + typename enable_if::value, __nat*>::type = 0) + _NOEXCEPT; + ~weak_ptr(); + + _LIBCPP_INLINE_VISIBILITY + weak_ptr& operator=(weak_ptr const& __r) _NOEXCEPT; + template + typename enable_if + < + is_convertible<_Yp*, element_type*>::value, + weak_ptr& + >::type + _LIBCPP_INLINE_VISIBILITY + operator=(weak_ptr<_Yp> const& __r) _NOEXCEPT; + + _LIBCPP_INLINE_VISIBILITY + weak_ptr& operator=(weak_ptr&& __r) _NOEXCEPT; + template + typename enable_if + < + is_convertible<_Yp*, element_type*>::value, + weak_ptr& + >::type + _LIBCPP_INLINE_VISIBILITY + operator=(weak_ptr<_Yp>&& __r) _NOEXCEPT; + + template + typename enable_if + < + is_convertible<_Yp*, element_type*>::value, + weak_ptr& + >::type + _LIBCPP_INLINE_VISIBILITY + operator=(shared_ptr<_Yp> const& __r) _NOEXCEPT; + + _LIBCPP_INLINE_VISIBILITY + void swap(weak_ptr& __r) _NOEXCEPT; + _LIBCPP_INLINE_VISIBILITY + void reset() _NOEXCEPT; + + _LIBCPP_INLINE_VISIBILITY + long use_count() const _NOEXCEPT + {return __cntrl_ ? __cntrl_->use_count() : 0;} + _LIBCPP_INLINE_VISIBILITY + bool expired() const _NOEXCEPT + {return __cntrl_ == nullptr || __cntrl_->use_count() == 0;} + shared_ptr<_Tp> lock() const _NOEXCEPT; + template + _LIBCPP_INLINE_VISIBILITY + bool owner_before(const shared_ptr<_Up>& __r) const _NOEXCEPT + {return __cntrl_ < __r.__cntrl_;} + template + _LIBCPP_INLINE_VISIBILITY + bool owner_before(const weak_ptr<_Up>& __r) const _NOEXCEPT + {return __cntrl_ < __r.__cntrl_;} + + template friend class _LIBCPP_TEMPLATE_VIS weak_ptr; + template friend class _LIBCPP_TEMPLATE_VIS shared_ptr; +}; + +#ifndef _LIBCPP_HAS_NO_DEDUCTION_GUIDES +template +weak_ptr(shared_ptr<_Tp>) -> weak_ptr<_Tp>; +#endif + +template +inline +_LIBCPP_CONSTEXPR +weak_ptr<_Tp>::weak_ptr() _NOEXCEPT + : __ptr_(nullptr), + __cntrl_(nullptr) +{ +} + +template +inline +weak_ptr<_Tp>::weak_ptr(weak_ptr const& __r) _NOEXCEPT + : __ptr_(__r.__ptr_), + __cntrl_(__r.__cntrl_) +{ + if (__cntrl_) + __cntrl_->__add_weak(); +} + +template +template +inline +weak_ptr<_Tp>::weak_ptr(shared_ptr<_Yp> const& __r, + typename enable_if::value, __nat*>::type) + _NOEXCEPT + : __ptr_(__r.__ptr_), + __cntrl_(__r.__cntrl_) +{ + if (__cntrl_) + __cntrl_->__add_weak(); +} + +template +template +inline +weak_ptr<_Tp>::weak_ptr(weak_ptr<_Yp> const& __r, + typename enable_if::value, __nat*>::type) + _NOEXCEPT + : __ptr_(__r.__ptr_), + __cntrl_(__r.__cntrl_) +{ + if (__cntrl_) + __cntrl_->__add_weak(); +} + +template +inline +weak_ptr<_Tp>::weak_ptr(weak_ptr&& __r) _NOEXCEPT + : __ptr_(__r.__ptr_), + __cntrl_(__r.__cntrl_) +{ + __r.__ptr_ = nullptr; + __r.__cntrl_ = nullptr; +} + +template +template +inline +weak_ptr<_Tp>::weak_ptr(weak_ptr<_Yp>&& __r, + typename enable_if::value, __nat*>::type) + _NOEXCEPT + : __ptr_(__r.__ptr_), + __cntrl_(__r.__cntrl_) +{ + __r.__ptr_ = nullptr; + __r.__cntrl_ = nullptr; +} + +template +weak_ptr<_Tp>::~weak_ptr() +{ + if (__cntrl_) + __cntrl_->__release_weak(); +} + +template +inline +weak_ptr<_Tp>& +weak_ptr<_Tp>::operator=(weak_ptr const& __r) _NOEXCEPT +{ + weak_ptr(__r).swap(*this); + return *this; +} + +template +template +inline +typename enable_if +< + is_convertible<_Yp*, _Tp*>::value, + weak_ptr<_Tp>& +>::type +weak_ptr<_Tp>::operator=(weak_ptr<_Yp> const& __r) _NOEXCEPT +{ + weak_ptr(__r).swap(*this); + return *this; +} + +template +inline +weak_ptr<_Tp>& +weak_ptr<_Tp>::operator=(weak_ptr&& __r) _NOEXCEPT +{ + weak_ptr(_VSTD::move(__r)).swap(*this); + return *this; +} + +template +template +inline +typename enable_if +< + is_convertible<_Yp*, _Tp*>::value, + weak_ptr<_Tp>& +>::type +weak_ptr<_Tp>::operator=(weak_ptr<_Yp>&& __r) _NOEXCEPT +{ + weak_ptr(_VSTD::move(__r)).swap(*this); + return *this; +} + +template +template +inline +typename enable_if +< + is_convertible<_Yp*, _Tp*>::value, + weak_ptr<_Tp>& +>::type +weak_ptr<_Tp>::operator=(shared_ptr<_Yp> const& __r) _NOEXCEPT +{ + weak_ptr(__r).swap(*this); + return *this; +} + +template +inline +void +weak_ptr<_Tp>::swap(weak_ptr& __r) _NOEXCEPT +{ + _VSTD::swap(__ptr_, __r.__ptr_); + _VSTD::swap(__cntrl_, __r.__cntrl_); +} + +template +inline _LIBCPP_INLINE_VISIBILITY +void +swap(weak_ptr<_Tp>& __x, weak_ptr<_Tp>& __y) _NOEXCEPT +{ + __x.swap(__y); +} + +template +inline +void +weak_ptr<_Tp>::reset() _NOEXCEPT +{ + weak_ptr().swap(*this); +} + +template +template +shared_ptr<_Tp>::shared_ptr(const weak_ptr<_Yp>& __r, + typename enable_if::value, __nat>::type) + : __ptr_(__r.__ptr_), + __cntrl_(__r.__cntrl_ ? __r.__cntrl_->lock() : __r.__cntrl_) +{ + if (__cntrl_ == nullptr) + __throw_bad_weak_ptr(); +} + +template +shared_ptr<_Tp> +weak_ptr<_Tp>::lock() const _NOEXCEPT +{ + shared_ptr<_Tp> __r; + __r.__cntrl_ = __cntrl_ ? __cntrl_->lock() : __cntrl_; + if (__r.__cntrl_) + __r.__ptr_ = __ptr_; + return __r; +} + +#if _LIBCPP_STD_VER > 14 +template struct owner_less; +#else +template struct owner_less; +#endif + + +_LIBCPP_SUPPRESS_DEPRECATED_PUSH +template +struct _LIBCPP_TEMPLATE_VIS owner_less > +#if !defined(_LIBCPP_ABI_NO_BINDER_BASES) + : binary_function, shared_ptr<_Tp>, bool> +#endif +{ +_LIBCPP_SUPPRESS_DEPRECATED_POP +#if _LIBCPP_STD_VER <= 17 || defined(_LIBCPP_ENABLE_CXX20_REMOVED_BINDER_TYPEDEFS) + _LIBCPP_DEPRECATED_IN_CXX17 typedef bool result_type; + _LIBCPP_DEPRECATED_IN_CXX17 typedef shared_ptr<_Tp> first_argument_type; + _LIBCPP_DEPRECATED_IN_CXX17 typedef shared_ptr<_Tp> second_argument_type; +#endif + _LIBCPP_INLINE_VISIBILITY + bool operator()(shared_ptr<_Tp> const& __x, shared_ptr<_Tp> const& __y) const _NOEXCEPT + {return __x.owner_before(__y);} + _LIBCPP_INLINE_VISIBILITY + bool operator()(shared_ptr<_Tp> const& __x, weak_ptr<_Tp> const& __y) const _NOEXCEPT + {return __x.owner_before(__y);} + _LIBCPP_INLINE_VISIBILITY + bool operator()( weak_ptr<_Tp> const& __x, shared_ptr<_Tp> const& __y) const _NOEXCEPT + {return __x.owner_before(__y);} +}; + +_LIBCPP_SUPPRESS_DEPRECATED_PUSH +template +struct _LIBCPP_TEMPLATE_VIS owner_less > +#if !defined(_LIBCPP_ABI_NO_BINDER_BASES) + : binary_function, weak_ptr<_Tp>, bool> +#endif +{ +_LIBCPP_SUPPRESS_DEPRECATED_POP +#if _LIBCPP_STD_VER <= 17 || defined(_LIBCPP_ENABLE_CXX20_REMOVED_BINDER_TYPEDEFS) + _LIBCPP_DEPRECATED_IN_CXX17 typedef bool result_type; + _LIBCPP_DEPRECATED_IN_CXX17 typedef weak_ptr<_Tp> first_argument_type; + _LIBCPP_DEPRECATED_IN_CXX17 typedef weak_ptr<_Tp> second_argument_type; +#endif + _LIBCPP_INLINE_VISIBILITY + bool operator()( weak_ptr<_Tp> const& __x, weak_ptr<_Tp> const& __y) const _NOEXCEPT + {return __x.owner_before(__y);} + _LIBCPP_INLINE_VISIBILITY + bool operator()(shared_ptr<_Tp> const& __x, weak_ptr<_Tp> const& __y) const _NOEXCEPT + {return __x.owner_before(__y);} + _LIBCPP_INLINE_VISIBILITY + bool operator()( weak_ptr<_Tp> const& __x, shared_ptr<_Tp> const& __y) const _NOEXCEPT + {return __x.owner_before(__y);} +}; + +#if _LIBCPP_STD_VER > 14 +template <> +struct _LIBCPP_TEMPLATE_VIS owner_less +{ + template + _LIBCPP_INLINE_VISIBILITY + bool operator()( shared_ptr<_Tp> const& __x, shared_ptr<_Up> const& __y) const _NOEXCEPT + {return __x.owner_before(__y);} + template + _LIBCPP_INLINE_VISIBILITY + bool operator()( shared_ptr<_Tp> const& __x, weak_ptr<_Up> const& __y) const _NOEXCEPT + {return __x.owner_before(__y);} + template + _LIBCPP_INLINE_VISIBILITY + bool operator()( weak_ptr<_Tp> const& __x, shared_ptr<_Up> const& __y) const _NOEXCEPT + {return __x.owner_before(__y);} + template + _LIBCPP_INLINE_VISIBILITY + bool operator()( weak_ptr<_Tp> const& __x, weak_ptr<_Up> const& __y) const _NOEXCEPT + {return __x.owner_before(__y);} + typedef void is_transparent; +}; +#endif + +template +class _LIBCPP_TEMPLATE_VIS enable_shared_from_this +{ + mutable weak_ptr<_Tp> __weak_this_; +protected: + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR + enable_shared_from_this() _NOEXCEPT {} + _LIBCPP_INLINE_VISIBILITY + enable_shared_from_this(enable_shared_from_this const&) _NOEXCEPT {} + _LIBCPP_INLINE_VISIBILITY + enable_shared_from_this& operator=(enable_shared_from_this const&) _NOEXCEPT + {return *this;} + _LIBCPP_INLINE_VISIBILITY + ~enable_shared_from_this() {} +public: + _LIBCPP_INLINE_VISIBILITY + shared_ptr<_Tp> shared_from_this() + {return shared_ptr<_Tp>(__weak_this_);} + _LIBCPP_INLINE_VISIBILITY + shared_ptr<_Tp const> shared_from_this() const + {return shared_ptr(__weak_this_);} + +#if _LIBCPP_STD_VER > 14 + _LIBCPP_INLINE_VISIBILITY + weak_ptr<_Tp> weak_from_this() _NOEXCEPT + { return __weak_this_; } + + _LIBCPP_INLINE_VISIBILITY + weak_ptr weak_from_this() const _NOEXCEPT + { return __weak_this_; } +#endif // _LIBCPP_STD_VER > 14 + + template friend class shared_ptr; +}; + +template struct _LIBCPP_TEMPLATE_VIS hash; + +template +struct _LIBCPP_TEMPLATE_VIS hash > +{ +#if _LIBCPP_STD_VER <= 17 || defined(_LIBCPP_ENABLE_CXX20_REMOVED_BINDER_TYPEDEFS) + _LIBCPP_DEPRECATED_IN_CXX17 typedef shared_ptr<_Tp> argument_type; + _LIBCPP_DEPRECATED_IN_CXX17 typedef size_t result_type; +#endif + + _LIBCPP_INLINE_VISIBILITY + size_t operator()(const shared_ptr<_Tp>& __ptr) const _NOEXCEPT + { + return hash::element_type*>()(__ptr.get()); + } +}; + +template +inline _LIBCPP_INLINE_VISIBILITY +basic_ostream<_CharT, _Traits>& +operator<<(basic_ostream<_CharT, _Traits>& __os, shared_ptr<_Yp> const& __p); + + +#if !defined(_LIBCPP_HAS_NO_ATOMIC_HEADER) + +class _LIBCPP_TYPE_VIS __sp_mut +{ + void* __lx; +public: + void lock() _NOEXCEPT; + void unlock() _NOEXCEPT; + +private: + _LIBCPP_CONSTEXPR __sp_mut(void*) _NOEXCEPT; + __sp_mut(const __sp_mut&); + __sp_mut& operator=(const __sp_mut&); + + friend _LIBCPP_FUNC_VIS __sp_mut& __get_sp_mut(const void*); +}; + +_LIBCPP_FUNC_VIS _LIBCPP_AVAILABILITY_ATOMIC_SHARED_PTR +__sp_mut& __get_sp_mut(const void*); + +template +inline _LIBCPP_INLINE_VISIBILITY +bool +atomic_is_lock_free(const shared_ptr<_Tp>*) +{ + return false; +} + +template +_LIBCPP_AVAILABILITY_ATOMIC_SHARED_PTR +shared_ptr<_Tp> +atomic_load(const shared_ptr<_Tp>* __p) +{ + __sp_mut& __m = __get_sp_mut(__p); + __m.lock(); + shared_ptr<_Tp> __q = *__p; + __m.unlock(); + return __q; +} + +template +inline _LIBCPP_INLINE_VISIBILITY +_LIBCPP_AVAILABILITY_ATOMIC_SHARED_PTR +shared_ptr<_Tp> +atomic_load_explicit(const shared_ptr<_Tp>* __p, memory_order) +{ + return atomic_load(__p); +} + +template +_LIBCPP_AVAILABILITY_ATOMIC_SHARED_PTR +void +atomic_store(shared_ptr<_Tp>* __p, shared_ptr<_Tp> __r) +{ + __sp_mut& __m = __get_sp_mut(__p); + __m.lock(); + __p->swap(__r); + __m.unlock(); +} + +template +inline _LIBCPP_INLINE_VISIBILITY +_LIBCPP_AVAILABILITY_ATOMIC_SHARED_PTR +void +atomic_store_explicit(shared_ptr<_Tp>* __p, shared_ptr<_Tp> __r, memory_order) +{ + atomic_store(__p, __r); +} + +template +_LIBCPP_AVAILABILITY_ATOMIC_SHARED_PTR +shared_ptr<_Tp> +atomic_exchange(shared_ptr<_Tp>* __p, shared_ptr<_Tp> __r) +{ + __sp_mut& __m = __get_sp_mut(__p); + __m.lock(); + __p->swap(__r); + __m.unlock(); + return __r; +} + +template +inline _LIBCPP_INLINE_VISIBILITY +_LIBCPP_AVAILABILITY_ATOMIC_SHARED_PTR +shared_ptr<_Tp> +atomic_exchange_explicit(shared_ptr<_Tp>* __p, shared_ptr<_Tp> __r, memory_order) +{ + return atomic_exchange(__p, __r); +} + +template +_LIBCPP_AVAILABILITY_ATOMIC_SHARED_PTR +bool +atomic_compare_exchange_strong(shared_ptr<_Tp>* __p, shared_ptr<_Tp>* __v, shared_ptr<_Tp> __w) +{ + shared_ptr<_Tp> __temp; + __sp_mut& __m = __get_sp_mut(__p); + __m.lock(); + if (__p->__owner_equivalent(*__v)) + { + _VSTD::swap(__temp, *__p); + *__p = __w; + __m.unlock(); + return true; + } + _VSTD::swap(__temp, *__v); + *__v = *__p; + __m.unlock(); + return false; +} + +template +inline _LIBCPP_INLINE_VISIBILITY +_LIBCPP_AVAILABILITY_ATOMIC_SHARED_PTR +bool +atomic_compare_exchange_weak(shared_ptr<_Tp>* __p, shared_ptr<_Tp>* __v, shared_ptr<_Tp> __w) +{ + return atomic_compare_exchange_strong(__p, __v, __w); +} + +template +inline _LIBCPP_INLINE_VISIBILITY +_LIBCPP_AVAILABILITY_ATOMIC_SHARED_PTR +bool +atomic_compare_exchange_strong_explicit(shared_ptr<_Tp>* __p, shared_ptr<_Tp>* __v, + shared_ptr<_Tp> __w, memory_order, memory_order) +{ + return atomic_compare_exchange_strong(__p, __v, __w); +} + +template +inline _LIBCPP_INLINE_VISIBILITY +_LIBCPP_AVAILABILITY_ATOMIC_SHARED_PTR +bool +atomic_compare_exchange_weak_explicit(shared_ptr<_Tp>* __p, shared_ptr<_Tp>* __v, + shared_ptr<_Tp> __w, memory_order, memory_order) +{ + return atomic_compare_exchange_weak(__p, __v, __w); +} + +#endif // !defined(_LIBCPP_HAS_NO_ATOMIC_HEADER) + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___MEMORY_SHARED_PTR_H diff --git a/lib/libcxx/include/__memory/temporary_buffer.h b/lib/libcxx/include/__memory/temporary_buffer.h new file mode 100644 index 000000000..6d1884f72 --- /dev/null +++ b/lib/libcxx/include/__memory/temporary_buffer.h @@ -0,0 +1,89 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___MEMORY_TEMPORARY_BUFFER_H +#define _LIBCPP___MEMORY_TEMPORARY_BUFFER_H + +#include <__config> +#include +#include +#include // pair + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +_LIBCPP_NODISCARD_EXT _LIBCPP_NO_CFI +pair<_Tp*, ptrdiff_t> +get_temporary_buffer(ptrdiff_t __n) _NOEXCEPT +{ + pair<_Tp*, ptrdiff_t> __r(0, 0); + const ptrdiff_t __m = (~ptrdiff_t(0) ^ + ptrdiff_t(ptrdiff_t(1) << (sizeof(ptrdiff_t) * __CHAR_BIT__ - 1))) + / sizeof(_Tp); + if (__n > __m) + __n = __m; + while (__n > 0) + { +#if !defined(_LIBCPP_HAS_NO_ALIGNED_ALLOCATION) + if (__is_overaligned_for_new(_LIBCPP_ALIGNOF(_Tp))) + { + align_val_t __al = + align_val_t(alignment_of<_Tp>::value); + __r.first = static_cast<_Tp*>(::operator new( + __n * sizeof(_Tp), __al, nothrow)); + } else { + __r.first = static_cast<_Tp*>(::operator new( + __n * sizeof(_Tp), nothrow)); + } +#else + if (__is_overaligned_for_new(_LIBCPP_ALIGNOF(_Tp))) + { + // Since aligned operator new is unavailable, return an empty + // buffer rather than one with invalid alignment. + return __r; + } + + __r.first = static_cast<_Tp*>(::operator new(__n * sizeof(_Tp), nothrow)); +#endif + + if (__r.first) + { + __r.second = __n; + break; + } + __n /= 2; + } + return __r; +} + +template +inline _LIBCPP_INLINE_VISIBILITY +void return_temporary_buffer(_Tp* __p) _NOEXCEPT +{ + _VSTD::__libcpp_deallocate_unsized((void*)__p, _LIBCPP_ALIGNOF(_Tp)); +} + +struct __return_temporary_buffer +{ + template + _LIBCPP_INLINE_VISIBILITY void operator()(_Tp* __p) const {_VSTD::return_temporary_buffer(__p);} +}; + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___MEMORY_TEMPORARY_BUFFER_H diff --git a/lib/libcxx/include/__memory/uninitialized_algorithms.h b/lib/libcxx/include/__memory/uninitialized_algorithms.h new file mode 100644 index 000000000..39edabbd8 --- /dev/null +++ b/lib/libcxx/include/__memory/uninitialized_algorithms.h @@ -0,0 +1,261 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___MEMORY_UNINITIALIZED_ALGORITHMS_H +#define _LIBCPP___MEMORY_UNINITIALIZED_ALGORITHMS_H + +#include <__config> +#include <__memory/addressof.h> +#include <__memory/construct_at.h> +#include +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +_ForwardIterator +uninitialized_copy(_InputIterator __f, _InputIterator __l, _ForwardIterator __r) +{ + typedef typename iterator_traits<_ForwardIterator>::value_type value_type; +#ifndef _LIBCPP_NO_EXCEPTIONS + _ForwardIterator __s = __r; + try + { +#endif + for (; __f != __l; ++__f, (void) ++__r) + ::new ((void*)_VSTD::addressof(*__r)) value_type(*__f); +#ifndef _LIBCPP_NO_EXCEPTIONS + } + catch (...) + { + for (; __s != __r; ++__s) + __s->~value_type(); + throw; + } +#endif + return __r; +} + +template +_ForwardIterator +uninitialized_copy_n(_InputIterator __f, _Size __n, _ForwardIterator __r) +{ + typedef typename iterator_traits<_ForwardIterator>::value_type value_type; +#ifndef _LIBCPP_NO_EXCEPTIONS + _ForwardIterator __s = __r; + try + { +#endif + for (; __n > 0; ++__f, (void) ++__r, (void) --__n) + ::new ((void*)_VSTD::addressof(*__r)) value_type(*__f); +#ifndef _LIBCPP_NO_EXCEPTIONS + } + catch (...) + { + for (; __s != __r; ++__s) + __s->~value_type(); + throw; + } +#endif + return __r; +} + +template +void +uninitialized_fill(_ForwardIterator __f, _ForwardIterator __l, const _Tp& __x) +{ + typedef typename iterator_traits<_ForwardIterator>::value_type value_type; +#ifndef _LIBCPP_NO_EXCEPTIONS + _ForwardIterator __s = __f; + try + { +#endif + for (; __f != __l; ++__f) + ::new ((void*)_VSTD::addressof(*__f)) value_type(__x); +#ifndef _LIBCPP_NO_EXCEPTIONS + } + catch (...) + { + for (; __s != __f; ++__s) + __s->~value_type(); + throw; + } +#endif +} + +template +_ForwardIterator +uninitialized_fill_n(_ForwardIterator __f, _Size __n, const _Tp& __x) +{ + typedef typename iterator_traits<_ForwardIterator>::value_type value_type; +#ifndef _LIBCPP_NO_EXCEPTIONS + _ForwardIterator __s = __f; + try + { +#endif + for (; __n > 0; ++__f, (void) --__n) + ::new ((void*)_VSTD::addressof(*__f)) value_type(__x); +#ifndef _LIBCPP_NO_EXCEPTIONS + } + catch (...) + { + for (; __s != __f; ++__s) + __s->~value_type(); + throw; + } +#endif + return __f; +} + +#if _LIBCPP_STD_VER > 14 + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +void destroy(_ForwardIterator __first, _ForwardIterator __last) { + for (; __first != __last; ++__first) + _VSTD::destroy_at(_VSTD::addressof(*__first)); +} + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +_ForwardIterator destroy_n(_ForwardIterator __first, _Size __n) { + for (; __n > 0; (void)++__first, --__n) + _VSTD::destroy_at(_VSTD::addressof(*__first)); + return __first; +} + +template +inline _LIBCPP_INLINE_VISIBILITY +void uninitialized_default_construct(_ForwardIterator __first, _ForwardIterator __last) { + using _Vt = typename iterator_traits<_ForwardIterator>::value_type; + auto __idx = __first; +#ifndef _LIBCPP_NO_EXCEPTIONS + try { +#endif + for (; __idx != __last; ++__idx) + ::new ((void*)_VSTD::addressof(*__idx)) _Vt; +#ifndef _LIBCPP_NO_EXCEPTIONS + } catch (...) { + _VSTD::destroy(__first, __idx); + throw; + } +#endif +} + +template +inline _LIBCPP_INLINE_VISIBILITY +_ForwardIterator uninitialized_default_construct_n(_ForwardIterator __first, _Size __n) { + using _Vt = typename iterator_traits<_ForwardIterator>::value_type; + auto __idx = __first; +#ifndef _LIBCPP_NO_EXCEPTIONS + try { +#endif + for (; __n > 0; (void)++__idx, --__n) + ::new ((void*)_VSTD::addressof(*__idx)) _Vt; + return __idx; +#ifndef _LIBCPP_NO_EXCEPTIONS + } catch (...) { + _VSTD::destroy(__first, __idx); + throw; + } +#endif +} + + +template +inline _LIBCPP_INLINE_VISIBILITY +void uninitialized_value_construct(_ForwardIterator __first, _ForwardIterator __last) { + using _Vt = typename iterator_traits<_ForwardIterator>::value_type; + auto __idx = __first; +#ifndef _LIBCPP_NO_EXCEPTIONS + try { +#endif + for (; __idx != __last; ++__idx) + ::new ((void*)_VSTD::addressof(*__idx)) _Vt(); +#ifndef _LIBCPP_NO_EXCEPTIONS + } catch (...) { + _VSTD::destroy(__first, __idx); + throw; + } +#endif +} + +template +inline _LIBCPP_INLINE_VISIBILITY +_ForwardIterator uninitialized_value_construct_n(_ForwardIterator __first, _Size __n) { + using _Vt = typename iterator_traits<_ForwardIterator>::value_type; + auto __idx = __first; +#ifndef _LIBCPP_NO_EXCEPTIONS + try { +#endif + for (; __n > 0; (void)++__idx, --__n) + ::new ((void*)_VSTD::addressof(*__idx)) _Vt(); + return __idx; +#ifndef _LIBCPP_NO_EXCEPTIONS + } catch (...) { + _VSTD::destroy(__first, __idx); + throw; + } +#endif +} + + +template +inline _LIBCPP_INLINE_VISIBILITY +_ForwardIt uninitialized_move(_InputIt __first, _InputIt __last, _ForwardIt __first_res) { + using _Vt = typename iterator_traits<_ForwardIt>::value_type; + auto __idx = __first_res; +#ifndef _LIBCPP_NO_EXCEPTIONS + try { +#endif + for (; __first != __last; (void)++__idx, ++__first) + ::new ((void*)_VSTD::addressof(*__idx)) _Vt(_VSTD::move(*__first)); + return __idx; +#ifndef _LIBCPP_NO_EXCEPTIONS + } catch (...) { + _VSTD::destroy(__first_res, __idx); + throw; + } +#endif +} + +template +inline _LIBCPP_INLINE_VISIBILITY +pair<_InputIt, _ForwardIt> +uninitialized_move_n(_InputIt __first, _Size __n, _ForwardIt __first_res) { + using _Vt = typename iterator_traits<_ForwardIt>::value_type; + auto __idx = __first_res; +#ifndef _LIBCPP_NO_EXCEPTIONS + try { +#endif + for (; __n > 0; ++__idx, (void)++__first, --__n) + ::new ((void*)_VSTD::addressof(*__idx)) _Vt(_VSTD::move(*__first)); + return {__first, __idx}; +#ifndef _LIBCPP_NO_EXCEPTIONS + } catch (...) { + _VSTD::destroy(__first_res, __idx); + throw; + } +#endif +} + +#endif // _LIBCPP_STD_VER > 14 + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___MEMORY_UNINITIALIZED_ALGORITHMS_H diff --git a/lib/libcxx/include/__memory/unique_ptr.h b/lib/libcxx/include/__memory/unique_ptr.h new file mode 100644 index 000000000..083e0a8c2 --- /dev/null +++ b/lib/libcxx/include/__memory/unique_ptr.h @@ -0,0 +1,773 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___MEMORY_UNIQUE_PTR_H +#define _LIBCPP___MEMORY_UNIQUE_PTR_H + +#include <__config> +#include <__functional_base> +#include <__functional/hash.h> +#include <__functional/operations.h> +#include <__memory/allocator_traits.h> // __pointer +#include <__memory/compressed_pair.h> +#include <__utility/forward.h> +#include +#include +#include + +#if _LIBCPP_STD_VER <= 14 || defined(_LIBCPP_ENABLE_CXX17_REMOVED_AUTO_PTR) +# include <__memory/auto_ptr.h> +#endif + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +struct _LIBCPP_TEMPLATE_VIS default_delete { + static_assert(!is_function<_Tp>::value, + "default_delete cannot be instantiated for function types"); +#ifndef _LIBCPP_CXX03_LANG + _LIBCPP_INLINE_VISIBILITY constexpr default_delete() _NOEXCEPT = default; +#else + _LIBCPP_INLINE_VISIBILITY default_delete() {} +#endif + template + _LIBCPP_INLINE_VISIBILITY + default_delete(const default_delete<_Up>&, + typename enable_if::value>::type* = + 0) _NOEXCEPT {} + + _LIBCPP_INLINE_VISIBILITY void operator()(_Tp* __ptr) const _NOEXCEPT { + static_assert(sizeof(_Tp) > 0, + "default_delete can not delete incomplete type"); + static_assert(!is_void<_Tp>::value, + "default_delete can not delete incomplete type"); + delete __ptr; + } +}; + +template +struct _LIBCPP_TEMPLATE_VIS default_delete<_Tp[]> { +private: + template + struct _EnableIfConvertible + : enable_if::value> {}; + +public: +#ifndef _LIBCPP_CXX03_LANG + _LIBCPP_INLINE_VISIBILITY constexpr default_delete() _NOEXCEPT = default; +#else + _LIBCPP_INLINE_VISIBILITY default_delete() {} +#endif + + template + _LIBCPP_INLINE_VISIBILITY + default_delete(const default_delete<_Up[]>&, + typename _EnableIfConvertible<_Up>::type* = 0) _NOEXCEPT {} + + template + _LIBCPP_INLINE_VISIBILITY + typename _EnableIfConvertible<_Up>::type + operator()(_Up* __ptr) const _NOEXCEPT { + static_assert(sizeof(_Tp) > 0, + "default_delete can not delete incomplete type"); + static_assert(!is_void<_Tp>::value, + "default_delete can not delete void type"); + delete[] __ptr; + } +}; + +template +struct __unique_ptr_deleter_sfinae { + static_assert(!is_reference<_Deleter>::value, "incorrect specialization"); + typedef const _Deleter& __lval_ref_type; + typedef _Deleter&& __good_rval_ref_type; + typedef true_type __enable_rval_overload; +}; + +template +struct __unique_ptr_deleter_sfinae<_Deleter const&> { + typedef const _Deleter& __lval_ref_type; + typedef const _Deleter&& __bad_rval_ref_type; + typedef false_type __enable_rval_overload; +}; + +template +struct __unique_ptr_deleter_sfinae<_Deleter&> { + typedef _Deleter& __lval_ref_type; + typedef _Deleter&& __bad_rval_ref_type; + typedef false_type __enable_rval_overload; +}; + +#if defined(_LIBCPP_ABI_ENABLE_UNIQUE_PTR_TRIVIAL_ABI) +# define _LIBCPP_UNIQUE_PTR_TRIVIAL_ABI __attribute__((trivial_abi)) +#else +# define _LIBCPP_UNIQUE_PTR_TRIVIAL_ABI +#endif + +template > +class _LIBCPP_UNIQUE_PTR_TRIVIAL_ABI _LIBCPP_TEMPLATE_VIS unique_ptr { +public: + typedef _Tp element_type; + typedef _Dp deleter_type; + typedef _LIBCPP_NODEBUG_TYPE typename __pointer<_Tp, deleter_type>::type pointer; + + static_assert(!is_rvalue_reference::value, + "the specified deleter type cannot be an rvalue reference"); + +private: + __compressed_pair __ptr_; + + struct __nat { int __for_bool_; }; + + typedef _LIBCPP_NODEBUG_TYPE __unique_ptr_deleter_sfinae<_Dp> _DeleterSFINAE; + + template + using _LValRefType _LIBCPP_NODEBUG_TYPE = + typename __dependent_type<_DeleterSFINAE, _Dummy>::__lval_ref_type; + + template + using _GoodRValRefType _LIBCPP_NODEBUG_TYPE = + typename __dependent_type<_DeleterSFINAE, _Dummy>::__good_rval_ref_type; + + template + using _BadRValRefType _LIBCPP_NODEBUG_TYPE = + typename __dependent_type<_DeleterSFINAE, _Dummy>::__bad_rval_ref_type; + + template , _Dummy>::type> + using _EnableIfDeleterDefaultConstructible _LIBCPP_NODEBUG_TYPE = + typename enable_if::value && + !is_pointer<_Deleter>::value>::type; + + template + using _EnableIfDeleterConstructible _LIBCPP_NODEBUG_TYPE = + typename enable_if::value>::type; + + template + using _EnableIfMoveConvertible _LIBCPP_NODEBUG_TYPE = typename enable_if< + is_convertible::value && + !is_array<_Up>::value + >::type; + + template + using _EnableIfDeleterConvertible _LIBCPP_NODEBUG_TYPE = typename enable_if< + (is_reference<_Dp>::value && is_same<_Dp, _UDel>::value) || + (!is_reference<_Dp>::value && is_convertible<_UDel, _Dp>::value) + >::type; + + template + using _EnableIfDeleterAssignable = typename enable_if< + is_assignable<_Dp&, _UDel&&>::value + >::type; + +public: + template > + _LIBCPP_INLINE_VISIBILITY + _LIBCPP_CONSTEXPR unique_ptr() _NOEXCEPT : __ptr_(pointer(), __default_init_tag()) {} + + template > + _LIBCPP_INLINE_VISIBILITY + _LIBCPP_CONSTEXPR unique_ptr(nullptr_t) _NOEXCEPT : __ptr_(pointer(), __default_init_tag()) {} + + template > + _LIBCPP_INLINE_VISIBILITY + explicit unique_ptr(pointer __p) _NOEXCEPT : __ptr_(__p, __default_init_tag()) {} + + template > > + _LIBCPP_INLINE_VISIBILITY + unique_ptr(pointer __p, _LValRefType<_Dummy> __d) _NOEXCEPT + : __ptr_(__p, __d) {} + + template > > + _LIBCPP_INLINE_VISIBILITY + unique_ptr(pointer __p, _GoodRValRefType<_Dummy> __d) _NOEXCEPT + : __ptr_(__p, _VSTD::move(__d)) { + static_assert(!is_reference::value, + "rvalue deleter bound to reference"); + } + + template > > + _LIBCPP_INLINE_VISIBILITY + unique_ptr(pointer __p, _BadRValRefType<_Dummy> __d) = delete; + + _LIBCPP_INLINE_VISIBILITY + unique_ptr(unique_ptr&& __u) _NOEXCEPT + : __ptr_(__u.release(), _VSTD::forward(__u.get_deleter())) { + } + + template , _Up>, + class = _EnableIfDeleterConvertible<_Ep> + > + _LIBCPP_INLINE_VISIBILITY + unique_ptr(unique_ptr<_Up, _Ep>&& __u) _NOEXCEPT + : __ptr_(__u.release(), _VSTD::forward<_Ep>(__u.get_deleter())) {} + +#if _LIBCPP_STD_VER <= 14 || defined(_LIBCPP_ENABLE_CXX17_REMOVED_AUTO_PTR) + template + _LIBCPP_INLINE_VISIBILITY + unique_ptr(auto_ptr<_Up>&& __p, + typename enable_if::value && + is_same<_Dp, default_delete<_Tp> >::value, + __nat>::type = __nat()) _NOEXCEPT + : __ptr_(__p.release(), __default_init_tag()) {} +#endif + + _LIBCPP_INLINE_VISIBILITY + unique_ptr& operator=(unique_ptr&& __u) _NOEXCEPT { + reset(__u.release()); + __ptr_.second() = _VSTD::forward(__u.get_deleter()); + return *this; + } + + template , _Up>, + class = _EnableIfDeleterAssignable<_Ep> + > + _LIBCPP_INLINE_VISIBILITY + unique_ptr& operator=(unique_ptr<_Up, _Ep>&& __u) _NOEXCEPT { + reset(__u.release()); + __ptr_.second() = _VSTD::forward<_Ep>(__u.get_deleter()); + return *this; + } + +#if _LIBCPP_STD_VER <= 14 || defined(_LIBCPP_ENABLE_CXX17_REMOVED_AUTO_PTR) + template + _LIBCPP_INLINE_VISIBILITY + typename enable_if::value && + is_same<_Dp, default_delete<_Tp> >::value, + unique_ptr&>::type + operator=(auto_ptr<_Up> __p) { + reset(__p.release()); + return *this; + } +#endif + +#ifdef _LIBCPP_CXX03_LANG + unique_ptr(unique_ptr const&) = delete; + unique_ptr& operator=(unique_ptr const&) = delete; +#endif + + + _LIBCPP_INLINE_VISIBILITY + ~unique_ptr() { reset(); } + + _LIBCPP_INLINE_VISIBILITY + unique_ptr& operator=(nullptr_t) _NOEXCEPT { + reset(); + return *this; + } + + _LIBCPP_INLINE_VISIBILITY + typename add_lvalue_reference<_Tp>::type + operator*() const { + return *__ptr_.first(); + } + _LIBCPP_INLINE_VISIBILITY + pointer operator->() const _NOEXCEPT { + return __ptr_.first(); + } + _LIBCPP_INLINE_VISIBILITY + pointer get() const _NOEXCEPT { + return __ptr_.first(); + } + _LIBCPP_INLINE_VISIBILITY + deleter_type& get_deleter() _NOEXCEPT { + return __ptr_.second(); + } + _LIBCPP_INLINE_VISIBILITY + const deleter_type& get_deleter() const _NOEXCEPT { + return __ptr_.second(); + } + _LIBCPP_INLINE_VISIBILITY + explicit operator bool() const _NOEXCEPT { + return __ptr_.first() != nullptr; + } + + _LIBCPP_INLINE_VISIBILITY + pointer release() _NOEXCEPT { + pointer __t = __ptr_.first(); + __ptr_.first() = pointer(); + return __t; + } + + _LIBCPP_INLINE_VISIBILITY + void reset(pointer __p = pointer()) _NOEXCEPT { + pointer __tmp = __ptr_.first(); + __ptr_.first() = __p; + if (__tmp) + __ptr_.second()(__tmp); + } + + _LIBCPP_INLINE_VISIBILITY + void swap(unique_ptr& __u) _NOEXCEPT { + __ptr_.swap(__u.__ptr_); + } +}; + + +template +class _LIBCPP_UNIQUE_PTR_TRIVIAL_ABI _LIBCPP_TEMPLATE_VIS unique_ptr<_Tp[], _Dp> { +public: + typedef _Tp element_type; + typedef _Dp deleter_type; + typedef typename __pointer<_Tp, deleter_type>::type pointer; + +private: + __compressed_pair __ptr_; + + template + struct _CheckArrayPointerConversion : is_same<_From, pointer> {}; + + template + struct _CheckArrayPointerConversion<_FromElem*> + : integral_constant::value || + (is_same::value && + is_convertible<_FromElem(*)[], element_type(*)[]>::value) + > + {}; + + typedef __unique_ptr_deleter_sfinae<_Dp> _DeleterSFINAE; + + template + using _LValRefType _LIBCPP_NODEBUG_TYPE = + typename __dependent_type<_DeleterSFINAE, _Dummy>::__lval_ref_type; + + template + using _GoodRValRefType _LIBCPP_NODEBUG_TYPE = + typename __dependent_type<_DeleterSFINAE, _Dummy>::__good_rval_ref_type; + + template + using _BadRValRefType _LIBCPP_NODEBUG_TYPE = + typename __dependent_type<_DeleterSFINAE, _Dummy>::__bad_rval_ref_type; + + template , _Dummy>::type> + using _EnableIfDeleterDefaultConstructible _LIBCPP_NODEBUG_TYPE = + typename enable_if::value && + !is_pointer<_Deleter>::value>::type; + + template + using _EnableIfDeleterConstructible _LIBCPP_NODEBUG_TYPE = + typename enable_if::value>::type; + + template + using _EnableIfPointerConvertible _LIBCPP_NODEBUG_TYPE = typename enable_if< + _CheckArrayPointerConversion<_Pp>::value + >::type; + + template + using _EnableIfMoveConvertible _LIBCPP_NODEBUG_TYPE = typename enable_if< + is_array<_Up>::value && + is_same::value && + is_same::value && + is_convertible<_ElemT(*)[], element_type(*)[]>::value + >::type; + + template + using _EnableIfDeleterConvertible _LIBCPP_NODEBUG_TYPE = typename enable_if< + (is_reference<_Dp>::value && is_same<_Dp, _UDel>::value) || + (!is_reference<_Dp>::value && is_convertible<_UDel, _Dp>::value) + >::type; + + template + using _EnableIfDeleterAssignable _LIBCPP_NODEBUG_TYPE = typename enable_if< + is_assignable<_Dp&, _UDel&&>::value + >::type; + +public: + template > + _LIBCPP_INLINE_VISIBILITY + _LIBCPP_CONSTEXPR unique_ptr() _NOEXCEPT : __ptr_(pointer(), __default_init_tag()) {} + + template > + _LIBCPP_INLINE_VISIBILITY + _LIBCPP_CONSTEXPR unique_ptr(nullptr_t) _NOEXCEPT : __ptr_(pointer(), __default_init_tag()) {} + + template , + class = _EnableIfPointerConvertible<_Pp> > + _LIBCPP_INLINE_VISIBILITY + explicit unique_ptr(_Pp __p) _NOEXCEPT + : __ptr_(__p, __default_init_tag()) {} + + template >, + class = _EnableIfPointerConvertible<_Pp> > + _LIBCPP_INLINE_VISIBILITY + unique_ptr(_Pp __p, _LValRefType<_Dummy> __d) _NOEXCEPT + : __ptr_(__p, __d) {} + + template > > + _LIBCPP_INLINE_VISIBILITY + unique_ptr(nullptr_t, _LValRefType<_Dummy> __d) _NOEXCEPT + : __ptr_(nullptr, __d) {} + + template >, + class = _EnableIfPointerConvertible<_Pp> > + _LIBCPP_INLINE_VISIBILITY + unique_ptr(_Pp __p, _GoodRValRefType<_Dummy> __d) _NOEXCEPT + : __ptr_(__p, _VSTD::move(__d)) { + static_assert(!is_reference::value, + "rvalue deleter bound to reference"); + } + + template > > + _LIBCPP_INLINE_VISIBILITY + unique_ptr(nullptr_t, _GoodRValRefType<_Dummy> __d) _NOEXCEPT + : __ptr_(nullptr, _VSTD::move(__d)) { + static_assert(!is_reference::value, + "rvalue deleter bound to reference"); + } + + template >, + class = _EnableIfPointerConvertible<_Pp> > + _LIBCPP_INLINE_VISIBILITY + unique_ptr(_Pp __p, _BadRValRefType<_Dummy> __d) = delete; + + _LIBCPP_INLINE_VISIBILITY + unique_ptr(unique_ptr&& __u) _NOEXCEPT + : __ptr_(__u.release(), _VSTD::forward(__u.get_deleter())) { + } + + _LIBCPP_INLINE_VISIBILITY + unique_ptr& operator=(unique_ptr&& __u) _NOEXCEPT { + reset(__u.release()); + __ptr_.second() = _VSTD::forward(__u.get_deleter()); + return *this; + } + + template , _Up>, + class = _EnableIfDeleterConvertible<_Ep> + > + _LIBCPP_INLINE_VISIBILITY + unique_ptr(unique_ptr<_Up, _Ep>&& __u) _NOEXCEPT + : __ptr_(__u.release(), _VSTD::forward<_Ep>(__u.get_deleter())) { + } + + template , _Up>, + class = _EnableIfDeleterAssignable<_Ep> + > + _LIBCPP_INLINE_VISIBILITY + unique_ptr& + operator=(unique_ptr<_Up, _Ep>&& __u) _NOEXCEPT { + reset(__u.release()); + __ptr_.second() = _VSTD::forward<_Ep>(__u.get_deleter()); + return *this; + } + +#ifdef _LIBCPP_CXX03_LANG + unique_ptr(unique_ptr const&) = delete; + unique_ptr& operator=(unique_ptr const&) = delete; +#endif + +public: + _LIBCPP_INLINE_VISIBILITY + ~unique_ptr() { reset(); } + + _LIBCPP_INLINE_VISIBILITY + unique_ptr& operator=(nullptr_t) _NOEXCEPT { + reset(); + return *this; + } + + _LIBCPP_INLINE_VISIBILITY + typename add_lvalue_reference<_Tp>::type + operator[](size_t __i) const { + return __ptr_.first()[__i]; + } + _LIBCPP_INLINE_VISIBILITY + pointer get() const _NOEXCEPT { + return __ptr_.first(); + } + + _LIBCPP_INLINE_VISIBILITY + deleter_type& get_deleter() _NOEXCEPT { + return __ptr_.second(); + } + + _LIBCPP_INLINE_VISIBILITY + const deleter_type& get_deleter() const _NOEXCEPT { + return __ptr_.second(); + } + _LIBCPP_INLINE_VISIBILITY + explicit operator bool() const _NOEXCEPT { + return __ptr_.first() != nullptr; + } + + _LIBCPP_INLINE_VISIBILITY + pointer release() _NOEXCEPT { + pointer __t = __ptr_.first(); + __ptr_.first() = pointer(); + return __t; + } + + template + _LIBCPP_INLINE_VISIBILITY + typename enable_if< + _CheckArrayPointerConversion<_Pp>::value + >::type + reset(_Pp __p) _NOEXCEPT { + pointer __tmp = __ptr_.first(); + __ptr_.first() = __p; + if (__tmp) + __ptr_.second()(__tmp); + } + + _LIBCPP_INLINE_VISIBILITY + void reset(nullptr_t = nullptr) _NOEXCEPT { + pointer __tmp = __ptr_.first(); + __ptr_.first() = nullptr; + if (__tmp) + __ptr_.second()(__tmp); + } + + _LIBCPP_INLINE_VISIBILITY + void swap(unique_ptr& __u) _NOEXCEPT { + __ptr_.swap(__u.__ptr_); + } + +}; + +template +inline _LIBCPP_INLINE_VISIBILITY +typename enable_if< + __is_swappable<_Dp>::value, + void +>::type +swap(unique_ptr<_Tp, _Dp>& __x, unique_ptr<_Tp, _Dp>& __y) _NOEXCEPT {__x.swap(__y);} + +template +inline _LIBCPP_INLINE_VISIBILITY +bool +operator==(const unique_ptr<_T1, _D1>& __x, const unique_ptr<_T2, _D2>& __y) {return __x.get() == __y.get();} + +template +inline _LIBCPP_INLINE_VISIBILITY +bool +operator!=(const unique_ptr<_T1, _D1>& __x, const unique_ptr<_T2, _D2>& __y) {return !(__x == __y);} + +template +inline _LIBCPP_INLINE_VISIBILITY +bool +operator< (const unique_ptr<_T1, _D1>& __x, const unique_ptr<_T2, _D2>& __y) +{ + typedef typename unique_ptr<_T1, _D1>::pointer _P1; + typedef typename unique_ptr<_T2, _D2>::pointer _P2; + typedef typename common_type<_P1, _P2>::type _Vp; + return less<_Vp>()(__x.get(), __y.get()); +} + +template +inline _LIBCPP_INLINE_VISIBILITY +bool +operator> (const unique_ptr<_T1, _D1>& __x, const unique_ptr<_T2, _D2>& __y) {return __y < __x;} + +template +inline _LIBCPP_INLINE_VISIBILITY +bool +operator<=(const unique_ptr<_T1, _D1>& __x, const unique_ptr<_T2, _D2>& __y) {return !(__y < __x);} + +template +inline _LIBCPP_INLINE_VISIBILITY +bool +operator>=(const unique_ptr<_T1, _D1>& __x, const unique_ptr<_T2, _D2>& __y) {return !(__x < __y);} + +template +inline _LIBCPP_INLINE_VISIBILITY +bool +operator==(const unique_ptr<_T1, _D1>& __x, nullptr_t) _NOEXCEPT +{ + return !__x; +} + +template +inline _LIBCPP_INLINE_VISIBILITY +bool +operator==(nullptr_t, const unique_ptr<_T1, _D1>& __x) _NOEXCEPT +{ + return !__x; +} + +template +inline _LIBCPP_INLINE_VISIBILITY +bool +operator!=(const unique_ptr<_T1, _D1>& __x, nullptr_t) _NOEXCEPT +{ + return static_cast(__x); +} + +template +inline _LIBCPP_INLINE_VISIBILITY +bool +operator!=(nullptr_t, const unique_ptr<_T1, _D1>& __x) _NOEXCEPT +{ + return static_cast(__x); +} + +template +inline _LIBCPP_INLINE_VISIBILITY +bool +operator<(const unique_ptr<_T1, _D1>& __x, nullptr_t) +{ + typedef typename unique_ptr<_T1, _D1>::pointer _P1; + return less<_P1>()(__x.get(), nullptr); +} + +template +inline _LIBCPP_INLINE_VISIBILITY +bool +operator<(nullptr_t, const unique_ptr<_T1, _D1>& __x) +{ + typedef typename unique_ptr<_T1, _D1>::pointer _P1; + return less<_P1>()(nullptr, __x.get()); +} + +template +inline _LIBCPP_INLINE_VISIBILITY +bool +operator>(const unique_ptr<_T1, _D1>& __x, nullptr_t) +{ + return nullptr < __x; +} + +template +inline _LIBCPP_INLINE_VISIBILITY +bool +operator>(nullptr_t, const unique_ptr<_T1, _D1>& __x) +{ + return __x < nullptr; +} + +template +inline _LIBCPP_INLINE_VISIBILITY +bool +operator<=(const unique_ptr<_T1, _D1>& __x, nullptr_t) +{ + return !(nullptr < __x); +} + +template +inline _LIBCPP_INLINE_VISIBILITY +bool +operator<=(nullptr_t, const unique_ptr<_T1, _D1>& __x) +{ + return !(__x < nullptr); +} + +template +inline _LIBCPP_INLINE_VISIBILITY +bool +operator>=(const unique_ptr<_T1, _D1>& __x, nullptr_t) +{ + return !(__x < nullptr); +} + +template +inline _LIBCPP_INLINE_VISIBILITY +bool +operator>=(nullptr_t, const unique_ptr<_T1, _D1>& __x) +{ + return !(nullptr < __x); +} + +#if _LIBCPP_STD_VER > 11 + +template +struct __unique_if +{ + typedef unique_ptr<_Tp> __unique_single; +}; + +template +struct __unique_if<_Tp[]> +{ + typedef unique_ptr<_Tp[]> __unique_array_unknown_bound; +}; + +template +struct __unique_if<_Tp[_Np]> +{ + typedef void __unique_array_known_bound; +}; + +template +inline _LIBCPP_INLINE_VISIBILITY +typename __unique_if<_Tp>::__unique_single +make_unique(_Args&&... __args) +{ + return unique_ptr<_Tp>(new _Tp(_VSTD::forward<_Args>(__args)...)); +} + +template +inline _LIBCPP_INLINE_VISIBILITY +typename __unique_if<_Tp>::__unique_array_unknown_bound +make_unique(size_t __n) +{ + typedef typename remove_extent<_Tp>::type _Up; + return unique_ptr<_Tp>(new _Up[__n]()); +} + +template + typename __unique_if<_Tp>::__unique_array_known_bound + make_unique(_Args&&...) = delete; + +#endif // _LIBCPP_STD_VER > 11 + +template struct _LIBCPP_TEMPLATE_VIS hash; + +template +#ifdef _LIBCPP_CXX03_LANG +struct _LIBCPP_TEMPLATE_VIS hash > +#else +struct _LIBCPP_TEMPLATE_VIS hash<__enable_hash_helper< + unique_ptr<_Tp, _Dp>, typename unique_ptr<_Tp, _Dp>::pointer> > +#endif +{ +#if _LIBCPP_STD_VER <= 17 || defined(_LIBCPP_ENABLE_CXX20_REMOVED_BINDER_TYPEDEFS) + _LIBCPP_DEPRECATED_IN_CXX17 typedef unique_ptr<_Tp, _Dp> argument_type; + _LIBCPP_DEPRECATED_IN_CXX17 typedef size_t result_type; +#endif + + _LIBCPP_INLINE_VISIBILITY + size_t operator()(const unique_ptr<_Tp, _Dp>& __ptr) const + { + typedef typename unique_ptr<_Tp, _Dp>::pointer pointer; + return hash()(__ptr.get()); + } +}; + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___MEMORY_UNIQUE_PTR_H diff --git a/lib/libcxx/include/__memory/uses_allocator.h b/lib/libcxx/include/__memory/uses_allocator.h new file mode 100644 index 000000000..36e752057 --- /dev/null +++ b/lib/libcxx/include/__memory/uses_allocator.h @@ -0,0 +1,60 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___MEMORY_USES_ALLOCATOR_H +#define _LIBCPP___MEMORY_USES_ALLOCATOR_H + +#include <__config> +#include +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +struct __has_allocator_type +{ +private: + struct __two {char __lx; char __lxx;}; + template static __two __test(...); + template static char __test(typename _Up::allocator_type* = 0); +public: + static const bool value = sizeof(__test<_Tp>(0)) == 1; +}; + +template ::value> +struct __uses_allocator + : public integral_constant::value> +{ +}; + +template +struct __uses_allocator<_Tp, _Alloc, false> + : public false_type +{ +}; + +template +struct _LIBCPP_TEMPLATE_VIS uses_allocator + : public __uses_allocator<_Tp, _Alloc> +{ +}; + +#if _LIBCPP_STD_VER > 14 +template +_LIBCPP_INLINE_VAR constexpr size_t uses_allocator_v = uses_allocator<_Tp, _Alloc>::value; +#endif + +_LIBCPP_END_NAMESPACE_STD + +#endif // _LIBCPP___MEMORY_USES_ALLOCATOR_H diff --git a/lib/libcxx/include/__mutex_base b/lib/libcxx/include/__mutex_base index 96454ace9..77590a8fd 100644 --- a/lib/libcxx/include/__mutex_base +++ b/lib/libcxx/include/__mutex_base @@ -11,10 +11,9 @@ #define _LIBCPP___MUTEX_BASE #include <__config> +#include <__threading_support> #include #include -#include <__threading_support> - #include #if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) @@ -190,8 +189,7 @@ public: _LIBCPP_INLINE_VISIBILITY bool owns_lock() const _NOEXCEPT {return __owns_;} _LIBCPP_INLINE_VISIBILITY - _LIBCPP_EXPLICIT - operator bool () const _NOEXCEPT {return __owns_;} + explicit operator bool() const _NOEXCEPT {return __owns_;} _LIBCPP_INLINE_VISIBILITY mutex_type* mutex() const _NOEXCEPT {return __m_;} }; @@ -526,4 +524,4 @@ _LIBCPP_END_NAMESPACE_STD _LIBCPP_POP_MACROS -#endif // _LIBCPP___MUTEX_BASE +#endif // _LIBCPP___MUTEX_BASE diff --git a/lib/libcxx/include/__node_handle b/lib/libcxx/include/__node_handle index be1fe17a8..f3ffa5e3a 100644 --- a/lib/libcxx/include/__node_handle +++ b/lib/libcxx/include/__node_handle @@ -11,6 +11,7 @@ #define _LIBCPP___NODE_HANDLE #include <__config> +#include <__debug> #include #include @@ -205,4 +206,4 @@ struct _LIBCPP_TEMPLATE_VIS __insert_return_type _LIBCPP_END_NAMESPACE_STD _LIBCPP_POP_MACROS -#endif +#endif // _LIBCPP___NODE_HANDLE diff --git a/lib/libcxx/include/__nullptr b/lib/libcxx/include/__nullptr index 45529a710..e14751164 100644 --- a/lib/libcxx/include/__nullptr +++ b/lib/libcxx/include/__nullptr @@ -56,6 +56,6 @@ namespace std typedef decltype(nullptr) nullptr_t; } -#endif // _LIBCPP_HAS_NO_NULLPTR +#endif // _LIBCPP_HAS_NO_NULLPTR -#endif // _LIBCPP_NULLPTR +#endif // _LIBCPP_NULLPTR diff --git a/lib/libcxx/include/__random/uniform_int_distribution.h b/lib/libcxx/include/__random/uniform_int_distribution.h new file mode 100644 index 000000000..a7cfa1ec7 --- /dev/null +++ b/lib/libcxx/include/__random/uniform_int_distribution.h @@ -0,0 +1,316 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___RANDOM_UNIFORM_INT_DISTRIBUTION_H +#define _LIBCPP___RANDOM_UNIFORM_INT_DISTRIBUTION_H + +#include <__bits> +#include <__config> +#include +#include +#include +#include +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +// __independent_bits_engine + +template +struct __log2_imp +{ + static const size_t value = _Xp & ((unsigned long long)(1) << _Rp) ? _Rp + : __log2_imp<_Xp, _Rp - 1>::value; +}; + +template +struct __log2_imp<_Xp, 0> +{ + static const size_t value = 0; +}; + +template +struct __log2_imp<0, _Rp> +{ + static const size_t value = _Rp + 1; +}; + +template +struct __log2 +{ + static const size_t value = __log2_imp<_Xp, + sizeof(_UIntType) * __CHAR_BIT__ - 1>::value; +}; + +template +class __independent_bits_engine +{ +public: + // types + typedef _UIntType result_type; + +private: + typedef typename _Engine::result_type _Engine_result_type; + typedef typename conditional + < + sizeof(_Engine_result_type) <= sizeof(result_type), + result_type, + _Engine_result_type + >::type _Working_result_type; + + _Engine& __e_; + size_t __w_; + size_t __w0_; + size_t __n_; + size_t __n0_; + _Working_result_type __y0_; + _Working_result_type __y1_; + _Engine_result_type __mask0_; + _Engine_result_type __mask1_; + +#ifdef _LIBCPP_CXX03_LANG + static const _Working_result_type _Rp = _Engine::_Max - _Engine::_Min + + _Working_result_type(1); +#else + static _LIBCPP_CONSTEXPR const _Working_result_type _Rp = _Engine::max() - _Engine::min() + + _Working_result_type(1); +#endif + static _LIBCPP_CONSTEXPR const size_t __m = __log2<_Working_result_type, _Rp>::value; + static _LIBCPP_CONSTEXPR const size_t _WDt = numeric_limits<_Working_result_type>::digits; + static _LIBCPP_CONSTEXPR const size_t _EDt = numeric_limits<_Engine_result_type>::digits; + +public: + // constructors and seeding functions + __independent_bits_engine(_Engine& __e, size_t __w); + + // generating functions + result_type operator()() {return __eval(integral_constant());} + +private: + result_type __eval(false_type); + result_type __eval(true_type); +}; + +template +__independent_bits_engine<_Engine, _UIntType> + ::__independent_bits_engine(_Engine& __e, size_t __w) + : __e_(__e), + __w_(__w) +{ + __n_ = __w_ / __m + (__w_ % __m != 0); + __w0_ = __w_ / __n_; + if (_Rp == 0) + __y0_ = _Rp; + else if (__w0_ < _WDt) + __y0_ = (_Rp >> __w0_) << __w0_; + else + __y0_ = 0; + if (_Rp - __y0_ > __y0_ / __n_) + { + ++__n_; + __w0_ = __w_ / __n_; + if (__w0_ < _WDt) + __y0_ = (_Rp >> __w0_) << __w0_; + else + __y0_ = 0; + } + __n0_ = __n_ - __w_ % __n_; + if (__w0_ < _WDt - 1) + __y1_ = (_Rp >> (__w0_ + 1)) << (__w0_ + 1); + else + __y1_ = 0; + __mask0_ = __w0_ > 0 ? _Engine_result_type(~0) >> (_EDt - __w0_) : + _Engine_result_type(0); + __mask1_ = __w0_ < _EDt - 1 ? + _Engine_result_type(~0) >> (_EDt - (__w0_ + 1)) : + _Engine_result_type(~0); +} + +template +inline +_UIntType +__independent_bits_engine<_Engine, _UIntType>::__eval(false_type) +{ + return static_cast(__e_() & __mask0_); +} + +template +_UIntType +__independent_bits_engine<_Engine, _UIntType>::__eval(true_type) +{ + const size_t _WRt = numeric_limits::digits; + result_type _Sp = 0; + for (size_t __k = 0; __k < __n0_; ++__k) + { + _Engine_result_type __u; + do + { + __u = __e_() - _Engine::min(); + } while (__u >= __y0_); + if (__w0_ < _WRt) + _Sp <<= __w0_; + else + _Sp = 0; + _Sp += __u & __mask0_; + } + for (size_t __k = __n0_; __k < __n_; ++__k) + { + _Engine_result_type __u; + do + { + __u = __e_() - _Engine::min(); + } while (__u >= __y1_); + if (__w0_ < _WRt - 1) + _Sp <<= __w0_ + 1; + else + _Sp = 0; + _Sp += __u & __mask1_; + } + return _Sp; +} + +template +class uniform_int_distribution +{ +public: + // types + typedef _IntType result_type; + + class param_type + { + result_type __a_; + result_type __b_; + public: + typedef uniform_int_distribution distribution_type; + + explicit param_type(result_type __a = 0, + result_type __b = numeric_limits::max()) + : __a_(__a), __b_(__b) {} + + result_type a() const {return __a_;} + result_type b() const {return __b_;} + + friend bool operator==(const param_type& __x, const param_type& __y) + {return __x.__a_ == __y.__a_ && __x.__b_ == __y.__b_;} + friend bool operator!=(const param_type& __x, const param_type& __y) + {return !(__x == __y);} + }; + +private: + param_type __p_; + +public: + // constructors and reset functions +#ifndef _LIBCPP_CXX03_LANG + uniform_int_distribution() : uniform_int_distribution(0) {} + explicit uniform_int_distribution( + result_type __a, result_type __b = numeric_limits::max()) + : __p_(param_type(__a, __b)) {} +#else + explicit uniform_int_distribution( + result_type __a = 0, + result_type __b = numeric_limits::max()) + : __p_(param_type(__a, __b)) {} +#endif + explicit uniform_int_distribution(const param_type& __p) : __p_(__p) {} + void reset() {} + + // generating functions + template result_type operator()(_URNG& __g) + {return (*this)(__g, __p_);} + template result_type operator()(_URNG& __g, const param_type& __p); + + // property functions + result_type a() const {return __p_.a();} + result_type b() const {return __p_.b();} + + param_type param() const {return __p_;} + void param(const param_type& __p) {__p_ = __p;} + + result_type min() const {return a();} + result_type max() const {return b();} + + friend bool operator==(const uniform_int_distribution& __x, + const uniform_int_distribution& __y) + {return __x.__p_ == __y.__p_;} + friend bool operator!=(const uniform_int_distribution& __x, + const uniform_int_distribution& __y) + {return !(__x == __y);} +}; + +template +template +typename uniform_int_distribution<_IntType>::result_type +uniform_int_distribution<_IntType>::operator()(_URNG& __g, const param_type& __p) +_LIBCPP_DISABLE_UBSAN_UNSIGNED_INTEGER_CHECK +{ + typedef typename conditional::type _UIntType; + const _UIntType _Rp = _UIntType(__p.b()) - _UIntType(__p.a()) + _UIntType(1); + if (_Rp == 1) + return __p.a(); + const size_t _Dt = numeric_limits<_UIntType>::digits; + typedef __independent_bits_engine<_URNG, _UIntType> _Eng; + if (_Rp == 0) + return static_cast(_Eng(__g, _Dt)()); + size_t __w = _Dt - __libcpp_clz(_Rp) - 1; + if ((_Rp & (numeric_limits<_UIntType>::max() >> (_Dt - __w))) != 0) + ++__w; + _Eng __e(__g, __w); + _UIntType __u; + do + { + __u = __e(); + } while (__u >= _Rp); + return static_cast(__u + __p.a()); +} + +template +basic_ostream<_CharT, _Traits>& +operator<<(basic_ostream<_CharT, _Traits>& __os, + const uniform_int_distribution<_IT>& __x) +{ + __save_flags<_CharT, _Traits> __lx(__os); + typedef basic_ostream<_CharT, _Traits> _Ostream; + __os.flags(_Ostream::dec | _Ostream::left); + _CharT __sp = __os.widen(' '); + __os.fill(__sp); + return __os << __x.a() << __sp << __x.b(); +} + +template +basic_istream<_CharT, _Traits>& +operator>>(basic_istream<_CharT, _Traits>& __is, + uniform_int_distribution<_IT>& __x) +{ + typedef uniform_int_distribution<_IT> _Eng; + typedef typename _Eng::result_type result_type; + typedef typename _Eng::param_type param_type; + __save_flags<_CharT, _Traits> __lx(__is); + typedef basic_istream<_CharT, _Traits> _Istream; + __is.flags(_Istream::dec | _Istream::skipws); + result_type __a; + result_type __b; + __is >> __a >> __b; + if (!__is.fail()) + __x.param(param_type(__a, __b)); + return __is; +} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___RANDOM_UNIFORM_INT_DISTRIBUTION_H diff --git a/lib/libcxx/include/__ranges/access.h b/lib/libcxx/include/__ranges/access.h new file mode 100644 index 000000000..add848887 --- /dev/null +++ b/lib/libcxx/include/__ranges/access.h @@ -0,0 +1,222 @@ +// -*- C++ -*- +//===------------------------ __ranges/access.h ---------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// +#ifndef _LIBCPP___RANGES_ACCESS_H +#define _LIBCPP___RANGES_ACCESS_H + +#include <__config> +#include <__iterator/concepts.h> +#include <__iterator/readable_traits.h> +#include <__ranges/enable_borrowed_range.h> +#include <__utility/__decay_copy.h> +#include <__utility/forward.h> +#include +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +#if !defined(_LIBCPP_HAS_NO_RANGES) + +// clang-format off + +namespace ranges { + template + concept __can_borrow = + is_lvalue_reference_v<_Tp> || enable_borrowed_range >; + + template + concept __is_complete = requires { sizeof(_Tp); }; +} // namespace ranges + +// [range.access.begin] +namespace ranges::__begin { + template + concept __member_begin = + __can_borrow<_Tp> && + requires(_Tp&& __t) { + { _VSTD::__decay_copy(__t.begin()) } -> input_or_output_iterator; + }; + + void begin(auto&) = delete; + void begin(const auto&) = delete; + + template + concept __unqualified_begin = + !__member_begin<_Tp> && + __can_borrow<_Tp> && + __class_or_enum > && + requires(_Tp && __t) { + { _VSTD::__decay_copy(begin(__t)) } -> input_or_output_iterator; + }; + + struct __fn { + template + requires is_array_v> + [[nodiscard]] _LIBCPP_HIDE_FROM_ABI constexpr auto operator()(_Tp& __t) const noexcept { + constexpr bool __complete = __is_complete >; + if constexpr (__complete) { // used to disable cryptic diagnostic + return __t + 0; + } + else { + static_assert(__complete, "`std::ranges::begin` is SFINAE-unfriendly on arrays of an incomplete type."); + } + } + + template + requires __member_begin<_Tp> + [[nodiscard]] _LIBCPP_HIDE_FROM_ABI constexpr auto operator()(_Tp&& __t) const + noexcept(noexcept(_VSTD::__decay_copy(__t.begin()))) + { + return __t.begin(); + } + + template + requires __unqualified_begin<_Tp> + [[nodiscard]] _LIBCPP_HIDE_FROM_ABI constexpr auto operator()(_Tp&& __t) const + noexcept(noexcept(_VSTD::__decay_copy(begin(__t)))) + { + return begin(__t); + } + + void operator()(auto&&) const = delete; + }; +} // namespace ranges::__begin + +namespace ranges { + inline namespace __cpo { + inline constexpr auto begin = __begin::__fn{}; + } // namespace __cpo + + template + using iterator_t = decltype(ranges::begin(declval<_Tp&>())); +} // namespace ranges + +// [range.access.end] +namespace ranges::__end { + template + concept __member_end = + __can_borrow<_Tp> && + requires(_Tp&& __t) { + typename iterator_t<_Tp>; + { _VSTD::__decay_copy(_VSTD::forward<_Tp>(__t).end()) } -> sentinel_for >; + }; + + void end(auto&) = delete; + void end(const auto&) = delete; + + template + concept __unqualified_end = + !__member_end<_Tp> && + __can_borrow<_Tp> && + __class_or_enum > && + requires(_Tp && __t) { + typename iterator_t<_Tp>; + { _VSTD::__decay_copy(end(_VSTD::forward<_Tp>(__t))) } -> sentinel_for >; + }; + + class __fn { + public: + template + [[nodiscard]] _LIBCPP_HIDE_FROM_ABI constexpr auto operator()(_Tp (&__t)[_Np]) const noexcept { + constexpr bool __complete = __is_complete >; + if constexpr (__complete) { // used to disable cryptic diagnostic + return __t + _Np; + } + else { + static_assert(__complete, "`std::ranges::end` is SFINAE-unfriendly on arrays of an incomplete type."); + } + } + + template + requires __member_end<_Tp> + [[nodiscard]] _LIBCPP_HIDE_FROM_ABI constexpr auto operator()(_Tp&& __t) const + noexcept(noexcept(_VSTD::__decay_copy(__t.end()))) + { + return _VSTD::forward<_Tp>(__t).end(); + } + + template + requires __unqualified_end<_Tp> + [[nodiscard]] _LIBCPP_HIDE_FROM_ABI constexpr auto operator()(_Tp&& __t) const + noexcept(noexcept(_VSTD::__decay_copy(end(__t)))) + { + return end(__t); + } + + void operator()(auto&&) const = delete; + }; +} // namespace ranges::__end + +namespace ranges::inline __cpo { + inline constexpr auto end = __end::__fn{}; +} // namespace ranges::__cpo + +namespace ranges::__cbegin { + struct __fn { + template + requires invocable + [[nodiscard]] _LIBCPP_HIDE_FROM_ABI constexpr auto operator()(_Tp& __t) const + noexcept(noexcept(ranges::begin(_VSTD::as_const(__t)))) + { + return ranges::begin(_VSTD::as_const(__t)); + } + + template + requires is_rvalue_reference_v<_Tp> && invocable + [[nodiscard]] _LIBCPP_HIDE_FROM_ABI constexpr auto operator()(_Tp&& __t) const + noexcept(noexcept(ranges::begin(static_cast<_Tp const&&>(__t)))) + { + return ranges::begin(static_cast<_Tp const&&>(__t)); + } + }; +} // namespace ranges::__cbegin + +namespace ranges::inline __cpo { + inline constexpr auto cbegin = __cbegin::__fn{}; +} // namespace ranges::__cpo + +namespace ranges::__cend { + struct __fn { + template + requires invocable + [[nodiscard]] _LIBCPP_HIDE_FROM_ABI constexpr auto operator()(_Tp& __t) const + noexcept(noexcept(ranges::end(_VSTD::as_const(__t)))) + { + return ranges::end(_VSTD::as_const(__t)); + } + + template + requires is_rvalue_reference_v<_Tp> && invocable + [[nodiscard]] _LIBCPP_HIDE_FROM_ABI constexpr auto operator()(_Tp&& __t) const + noexcept(noexcept(ranges::end(static_cast<_Tp const&&>(__t)))) + { + return ranges::end(static_cast<_Tp const&&>(__t)); + } + }; +} // namespace ranges::__cend + +namespace ranges::inline __cpo { + inline constexpr auto cend = __cend::__fn{}; +} // namespace ranges::__cpo + +// clang-format off + +#endif // !defined(_LIBCPP_HAS_NO_RANGES) + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___RANGES_ACCESS_H diff --git a/lib/libcxx/include/__ranges/all.h b/lib/libcxx/include/__ranges/all.h new file mode 100644 index 000000000..d678d3e5d --- /dev/null +++ b/lib/libcxx/include/__ranges/all.h @@ -0,0 +1,86 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// +#ifndef _LIBCPP___RANGES_ALL_H +#define _LIBCPP___RANGES_ALL_H + +#include <__config> +#include <__iterator/concepts.h> +#include <__iterator/iterator_traits.h> +#include <__ranges/access.h> +#include <__ranges/concepts.h> +#include <__ranges/ref_view.h> +#include <__ranges/subrange.h> +#include <__utility/__decay_copy.h> +#include <__utility/declval.h> +#include <__utility/forward.h> +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +#if !defined(_LIBCPP_HAS_NO_RANGES) + +namespace views { + +namespace __all { + struct __fn { + template + requires ranges::view> + _LIBCPP_HIDE_FROM_ABI + constexpr auto operator()(_Tp&& __t) const + noexcept(noexcept(_VSTD::__decay_copy(_VSTD::forward<_Tp>(__t)))) + { + return _VSTD::forward<_Tp>(__t); + } + + template + requires (!ranges::view>) && + requires (_Tp&& __t) { ranges::ref_view{_VSTD::forward<_Tp>(__t)}; } + _LIBCPP_HIDE_FROM_ABI + constexpr auto operator()(_Tp&& __t) const + noexcept(noexcept(ranges::ref_view{_VSTD::forward<_Tp>(__t)})) + { + return ranges::ref_view{_VSTD::forward<_Tp>(__t)}; + } + + template + requires (!ranges::view> && + !requires (_Tp&& __t) { ranges::ref_view{_VSTD::forward<_Tp>(__t)}; } && + requires (_Tp&& __t) { ranges::subrange{_VSTD::forward<_Tp>(__t)}; }) + _LIBCPP_HIDE_FROM_ABI + constexpr auto operator()(_Tp&& __t) const + noexcept(noexcept(ranges::subrange{_VSTD::forward<_Tp>(__t)})) + { + return ranges::subrange{_VSTD::forward<_Tp>(__t)}; + } + }; +} + +inline namespace __cpo { + inline constexpr auto all = __all::__fn{}; +} // namespace __cpo + +template +using all_t = decltype(views::all(declval<_Range>())); + +} // namespace views + +#endif // !defined(_LIBCPP_HAS_NO_RANGES) + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___RANGES_ALL_H diff --git a/lib/libcxx/include/__ranges/common_view.h b/lib/libcxx/include/__ranges/common_view.h new file mode 100644 index 000000000..dab82602f --- /dev/null +++ b/lib/libcxx/include/__ranges/common_view.h @@ -0,0 +1,113 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// +#ifndef _LIBCPP___RANGES_COMMON_VIEW_H +#define _LIBCPP___RANGES_COMMON_VIEW_H + +#include <__config> +#include <__iterator/common_iterator.h> +#include <__iterator/iterator_traits.h> +#include <__ranges/access.h> +#include <__ranges/all.h> +#include <__ranges/concepts.h> +#include <__ranges/enable_borrowed_range.h> +#include <__ranges/size.h> +#include <__ranges/view_interface.h> +#include +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +#if !defined(_LIBCPP_HAS_NO_RANGES) + +namespace ranges { + +template + requires (!common_range<_View> && copyable>) +class common_view : public view_interface> { + _View __base_ = _View(); + +public: + _LIBCPP_HIDE_FROM_ABI + common_view() requires default_initializable<_View> = default; + + _LIBCPP_HIDE_FROM_ABI + constexpr explicit common_view(_View __v) : __base_(_VSTD::move(__v)) { } + + _LIBCPP_HIDE_FROM_ABI + constexpr _View base() const& requires copy_constructible<_View> { return __base_; } + + _LIBCPP_HIDE_FROM_ABI + constexpr _View base() && { return _VSTD::move(__base_); } + + _LIBCPP_HIDE_FROM_ABI + constexpr auto begin() { + if constexpr (random_access_range<_View> && sized_range<_View>) + return ranges::begin(__base_); + else + return common_iterator, sentinel_t<_View>>(ranges::begin(__base_)); + } + + _LIBCPP_HIDE_FROM_ABI + constexpr auto begin() const requires range { + if constexpr (random_access_range && sized_range) + return ranges::begin(__base_); + else + return common_iterator, sentinel_t>(ranges::begin(__base_)); + } + + _LIBCPP_HIDE_FROM_ABI + constexpr auto end() { + if constexpr (random_access_range<_View> && sized_range<_View>) + return ranges::begin(__base_) + ranges::size(__base_); + else + return common_iterator, sentinel_t<_View>>(ranges::end(__base_)); + } + + _LIBCPP_HIDE_FROM_ABI + constexpr auto end() const requires range { + if constexpr (random_access_range && sized_range) + return ranges::begin(__base_) + ranges::size(__base_); + else + return common_iterator, sentinel_t>(ranges::end(__base_)); + } + + _LIBCPP_HIDE_FROM_ABI + constexpr auto size() requires sized_range<_View> { + return ranges::size(__base_); + } + + _LIBCPP_HIDE_FROM_ABI + constexpr auto size() const requires sized_range { + return ranges::size(__base_); + } +}; + +template +common_view(_Range&&) + -> common_view>; + +template +inline constexpr bool enable_borrowed_range> = enable_borrowed_range<_View>; + +} // namespace ranges + +#endif // !defined(_LIBCPP_HAS_NO_RANGES) + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___RANGES_COMMON_VIEW_H diff --git a/lib/libcxx/include/__ranges/concepts.h b/lib/libcxx/include/__ranges/concepts.h new file mode 100644 index 000000000..2f912c284 --- /dev/null +++ b/lib/libcxx/include/__ranges/concepts.h @@ -0,0 +1,138 @@ +// -*- C++ -*- +//===--------------------- __ranges/concepts.h ----------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// +#ifndef _LIBCPP___RANGES_CONCEPTS_H +#define _LIBCPP___RANGES_CONCEPTS_H + +#include <__config> +#include <__iterator/concepts.h> +#include <__iterator/incrementable_traits.h> +#include <__iterator/iter_move.h> +#include <__iterator/iterator_traits.h> +#include <__iterator/readable_traits.h> +#include <__ranges/access.h> +#include <__ranges/enable_borrowed_range.h> +#include <__ranges/data.h> +#include <__ranges/enable_view.h> +#include <__ranges/size.h> +#include +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +// clang-format off + +#if !defined(_LIBCPP_HAS_NO_RANGES) + +namespace ranges { + // [range.range] + template + concept range = requires(_Tp& __t) { + ranges::begin(__t); // sometimes equality-preserving + ranges::end(__t); + }; + + template + concept borrowed_range = range<_Range> && + (is_lvalue_reference_v<_Range> || enable_borrowed_range>); + + // `iterator_t` defined in <__ranges/access.h> + + template + using sentinel_t = decltype(ranges::end(declval<_Rp&>())); + + template + using range_difference_t = iter_difference_t>; + + template + using range_value_t = iter_value_t>; + + template + using range_reference_t = iter_reference_t>; + + template + using range_rvalue_reference_t = iter_rvalue_reference_t>; + + // [range.sized] + template + concept sized_range = range<_Tp> && requires(_Tp& __t) { ranges::size(__t); }; + + template + using range_size_t = decltype(ranges::size(declval<_Rp&>())); + + // `disable_sized_range` defined in `<__ranges/size.h>` + + // [range.view], views + + // `enable_view` defined in <__ranges/enable_view.h> + // `view_base` defined in <__ranges/enable_view.h> + + template + concept view = + range<_Tp> && + movable<_Tp> && + enable_view<_Tp>; + + template + concept __simple_view = + view<_Range> && range && + same_as, iterator_t> && + same_as, iterator_t>; + + // [range.refinements], other range refinements + template + concept output_range = range<_Rp> && output_iterator, _Tp>; + + template + concept input_range = range<_Tp> && input_iterator>; + + template + concept forward_range = input_range<_Tp> && forward_iterator>; + + template + concept bidirectional_range = forward_range<_Tp> && bidirectional_iterator>; + + template + concept random_access_range = + bidirectional_range<_Tp> && random_access_iterator>; + + template + concept contiguous_range = + random_access_range<_Tp> && + contiguous_iterator> && + requires(_Tp& __t) { + { ranges::data(__t) } -> same_as>>; + }; + + template + concept common_range = range<_Tp> && same_as, sentinel_t<_Tp>>; + + template + concept viewable_range = + range<_Tp> && ( + (view> && constructible_from, _Tp>) || + (!view> && borrowed_range<_Tp>) + ); +} // namespace ranges + +#endif // !defined(_LIBCPP_HAS_NO_RANGES) + +// clang-format on + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___RANGES_CONCEPTS_H diff --git a/lib/libcxx/include/__ranges/copyable_box.h b/lib/libcxx/include/__ranges/copyable_box.h new file mode 100644 index 000000000..f2d3843f7 --- /dev/null +++ b/lib/libcxx/include/__ranges/copyable_box.h @@ -0,0 +1,175 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___RANGES_COPYABLE_BOX_H +#define _LIBCPP___RANGES_COPYABLE_BOX_H + +#include <__config> +#include <__memory/addressof.h> +#include <__memory/construct_at.h> +#include <__utility/move.h> +#include +#include +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +#if !defined(_LIBCPP_HAS_NO_RANGES) + +// __copyable_box allows turning a type that is copy-constructible (but maybe not copy-assignable) into +// a type that is both copy-constructible and copy-assignable. It does that by introducing an empty state +// and basically doing destroy-then-copy-construct in the assignment operator. The empty state is necessary +// to handle the case where the copy construction fails after destroying the object. +// +// In some cases, we can completely avoid the use of an empty state; we provide a specialization of +// __copyable_box that does this, see below for the details. + +template +concept __copy_constructible_object = copy_constructible<_Tp> && is_object_v<_Tp>; + +namespace ranges { + // Primary template - uses std::optional and introduces an empty state in case assignment fails. + template<__copy_constructible_object _Tp> + class __copyable_box { + [[no_unique_address]] optional<_Tp> __val_; + + public: + template + requires is_constructible_v<_Tp, _Args...> + _LIBCPP_HIDE_FROM_ABI + constexpr explicit __copyable_box(in_place_t, _Args&& ...__args) + noexcept(is_nothrow_constructible_v<_Tp, _Args...>) + : __val_(in_place, _VSTD::forward<_Args>(__args)...) + { } + + _LIBCPP_HIDE_FROM_ABI + constexpr __copyable_box() noexcept(is_nothrow_default_constructible_v<_Tp>) + requires default_initializable<_Tp> + : __val_(in_place) + { } + + _LIBCPP_HIDE_FROM_ABI __copyable_box(__copyable_box const&) = default; + _LIBCPP_HIDE_FROM_ABI __copyable_box(__copyable_box&&) = default; + + _LIBCPP_HIDE_FROM_ABI + constexpr __copyable_box& operator=(__copyable_box const& __other) + noexcept(is_nothrow_copy_constructible_v<_Tp>) + { + if (this != _VSTD::addressof(__other)) { + if (__other.__has_value()) __val_.emplace(*__other); + else __val_.reset(); + } + return *this; + } + + _LIBCPP_HIDE_FROM_ABI + __copyable_box& operator=(__copyable_box&&) requires movable<_Tp> = default; + + _LIBCPP_HIDE_FROM_ABI + constexpr __copyable_box& operator=(__copyable_box&& __other) + noexcept(is_nothrow_move_constructible_v<_Tp>) + { + if (this != _VSTD::addressof(__other)) { + if (__other.__has_value()) __val_.emplace(_VSTD::move(*__other)); + else __val_.reset(); + } + return *this; + } + + _LIBCPP_HIDE_FROM_ABI constexpr _Tp const& operator*() const noexcept { return *__val_; } + _LIBCPP_HIDE_FROM_ABI constexpr _Tp& operator*() noexcept { return *__val_; } + _LIBCPP_HIDE_FROM_ABI constexpr bool __has_value() const noexcept { return __val_.has_value(); } + }; + + // This partial specialization implements an optimization for when we know we don't need to store + // an empty state to represent failure to perform an assignment. For copy-assignment, this happens: + // + // 1. If the type is copyable (which includes copy-assignment), we can use the type's own assignment operator + // directly and avoid using std::optional. + // 2. If the type is not copyable, but it is nothrow-copy-constructible, then we can implement assignment as + // destroy-and-then-construct and we know it will never fail, so we don't need an empty state. + // + // The exact same reasoning can be applied for move-assignment, with copyable replaced by movable and + // nothrow-copy-constructible replaced by nothrow-move-constructible. This specialization is enabled + // whenever we can apply any of these optimizations for both the copy assignment and the move assignment + // operator. + template + concept __doesnt_need_empty_state_for_copy = copyable<_Tp> || is_nothrow_copy_constructible_v<_Tp>; + + template + concept __doesnt_need_empty_state_for_move = movable<_Tp> || is_nothrow_move_constructible_v<_Tp>; + + template<__copy_constructible_object _Tp> + requires __doesnt_need_empty_state_for_copy<_Tp> && __doesnt_need_empty_state_for_move<_Tp> + class __copyable_box<_Tp> { + [[no_unique_address]] _Tp __val_; + + public: + template + requires is_constructible_v<_Tp, _Args...> + _LIBCPP_HIDE_FROM_ABI + constexpr explicit __copyable_box(in_place_t, _Args&& ...__args) + noexcept(is_nothrow_constructible_v<_Tp, _Args...>) + : __val_(_VSTD::forward<_Args>(__args)...) + { } + + _LIBCPP_HIDE_FROM_ABI + constexpr __copyable_box() noexcept(is_nothrow_default_constructible_v<_Tp>) + requires default_initializable<_Tp> + : __val_() + { } + + _LIBCPP_HIDE_FROM_ABI __copyable_box(__copyable_box const&) = default; + _LIBCPP_HIDE_FROM_ABI __copyable_box(__copyable_box&&) = default; + + // Implementation of assignment operators in case we perform optimization (1) + _LIBCPP_HIDE_FROM_ABI __copyable_box& operator=(__copyable_box const&) requires copyable<_Tp> = default; + _LIBCPP_HIDE_FROM_ABI __copyable_box& operator=(__copyable_box&&) requires movable<_Tp> = default; + + // Implementation of assignment operators in case we perform optimization (2) + _LIBCPP_HIDE_FROM_ABI + constexpr __copyable_box& operator=(__copyable_box const& __other) noexcept { + static_assert(is_nothrow_copy_constructible_v<_Tp>); + if (this != _VSTD::addressof(__other)) { + _VSTD::destroy_at(_VSTD::addressof(__val_)); + _VSTD::construct_at(_VSTD::addressof(__val_), __other.__val_); + } + return *this; + } + + _LIBCPP_HIDE_FROM_ABI + constexpr __copyable_box& operator=(__copyable_box&& __other) noexcept { + static_assert(is_nothrow_move_constructible_v<_Tp>); + if (this != _VSTD::addressof(__other)) { + _VSTD::destroy_at(_VSTD::addressof(__val_)); + _VSTD::construct_at(_VSTD::addressof(__val_), _VSTD::move(__other.__val_)); + } + return *this; + } + + _LIBCPP_HIDE_FROM_ABI constexpr _Tp const& operator*() const noexcept { return __val_; } + _LIBCPP_HIDE_FROM_ABI constexpr _Tp& operator*() noexcept { return __val_; } + _LIBCPP_HIDE_FROM_ABI constexpr bool __has_value() const noexcept { return true; } + }; +} // namespace ranges + +#endif // !defined(_LIBCPP_HAS_NO_RANGES) + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___RANGES_COPYABLE_BOX_H diff --git a/lib/libcxx/include/__ranges/dangling.h b/lib/libcxx/include/__ranges/dangling.h new file mode 100644 index 000000000..deb02a1a4 --- /dev/null +++ b/lib/libcxx/include/__ranges/dangling.h @@ -0,0 +1,47 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___RANGES_DANGLING_H +#define _LIBCPP___RANGES_DANGLING_H + +#include <__config> +#include <__ranges/access.h> +#include <__ranges/concepts.h> +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +#if !defined(_LIBCPP_HAS_NO_RANGES) + +namespace ranges { +struct dangling { + dangling() = default; + _LIBCPP_HIDE_FROM_ABI constexpr dangling(auto&&...) noexcept {} +}; + +template +using borrowed_iterator_t = _If, iterator_t<_Rp>, dangling>; + +// borrowed_subrange_t defined in <__ranges/subrange.h> +} // namespace ranges + +#endif // !_LIBCPP_HAS_NO_RANGES + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___RANGES_DANGLING_H diff --git a/lib/libcxx/include/__ranges/data.h b/lib/libcxx/include/__ranges/data.h new file mode 100644 index 000000000..dae30984d --- /dev/null +++ b/lib/libcxx/include/__ranges/data.h @@ -0,0 +1,86 @@ +// -*- C++ -*- +//===------------------------ __ranges/data.h ------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// +#ifndef _LIBCPP___RANGES_DATA_H +#define _LIBCPP___RANGES_DATA_H + +#include <__config> +#include <__iterator/concepts.h> +#include <__iterator/iterator_traits.h> +#include <__memory/pointer_traits.h> +#include <__ranges/access.h> +#include <__utility/forward.h> +#include +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +#if !defined(_LIBCPP_HAS_NO_RANGES) + +// clang-format off +namespace ranges { +// [range.prim.data] +namespace __data { + template + concept __ptr_to_object = is_pointer_v<_Tp> && is_object_v>; + + template + concept __member_data = + requires(_Tp&& __t) { + { _VSTD::forward<_Tp>(__t) } -> __can_borrow; + { __t.data() } -> __ptr_to_object; + }; + + template + concept __ranges_begin_invocable = + !__member_data<_Tp> && + requires(_Tp&& __t) { + { _VSTD::forward<_Tp>(__t) } -> __can_borrow; + { ranges::begin(_VSTD::forward<_Tp>(__t)) } -> contiguous_iterator; + }; + + struct __fn { + template <__member_data _Tp> + requires __can_borrow<_Tp> + _LIBCPP_HIDE_FROM_ABI + constexpr __ptr_to_object auto operator()(_Tp&& __t) const + noexcept(noexcept(__t.data())) { + return __t.data(); + } + + template<__ranges_begin_invocable _Tp> + requires __can_borrow<_Tp> + _LIBCPP_HIDE_FROM_ABI + constexpr __ptr_to_object auto operator()(_Tp&& __t) const + noexcept(noexcept(_VSTD::to_address(ranges::begin(_VSTD::forward<_Tp>(__t))))) { + return _VSTD::to_address(ranges::begin(_VSTD::forward<_Tp>(__t))); + } + }; +} // end namespace __data + +inline namespace __cpo { + inline constexpr const auto data = __data::__fn{}; +} // namespace __cpo +} // namespace ranges + +// clang-format off + +#endif // !defined(_LIBCPP_HAS_NO_RANGES) + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___RANGES_DATA_H diff --git a/lib/libcxx/include/__ranges/drop_view.h b/lib/libcxx/include/__ranges/drop_view.h new file mode 100644 index 000000000..099fd2271 --- /dev/null +++ b/lib/libcxx/include/__ranges/drop_view.h @@ -0,0 +1,131 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// +#ifndef _LIBCPP___RANGES_DROP_VIEW_H +#define _LIBCPP___RANGES_DROP_VIEW_H + +#include <__config> +#include <__iterator/concepts.h> +#include <__iterator/iterator_traits.h> +#include <__iterator/next.h> +#include <__ranges/access.h> +#include <__ranges/all.h> +#include <__ranges/concepts.h> +#include <__ranges/enable_borrowed_range.h> +#include <__ranges/non_propagating_cache.h> +#include <__ranges/size.h> +#include <__ranges/view_interface.h> +#include <__utility/move.h> +#include +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +#if !defined(_LIBCPP_HAS_NO_RANGES) + +namespace ranges { + template + class drop_view + : public view_interface> + { + // We cache begin() whenever ranges::next is not guaranteed O(1) to provide an + // amortized O(1) begin() method. If this is an input_range, then we cannot cache + // begin because begin is not equality preserving. + // Note: drop_view::begin() is still trivially amortized O(1) because + // one can't call begin() on it more than once. + static constexpr bool _UseCache = forward_range<_View> && !(random_access_range<_View> && sized_range<_View>); + using _Cache = _If<_UseCache, __non_propagating_cache>, __empty_cache>; + [[no_unique_address]] _Cache __cached_begin_ = _Cache(); + range_difference_t<_View> __count_ = 0; + _View __base_ = _View(); + +public: + drop_view() requires default_initializable<_View> = default; + + _LIBCPP_HIDE_FROM_ABI + constexpr drop_view(_View __base, range_difference_t<_View> __count) + : __count_(__count) + , __base_(_VSTD::move(__base)) + { + _LIBCPP_ASSERT(__count_ >= 0, "count must be greater than or equal to zero."); + } + + _LIBCPP_HIDE_FROM_ABI constexpr _View base() const& requires copy_constructible<_View> { return __base_; } + _LIBCPP_HIDE_FROM_ABI constexpr _View base() && { return _VSTD::move(__base_); } + + _LIBCPP_HIDE_FROM_ABI + constexpr auto begin() + requires (!(__simple_view<_View> && + random_access_range && sized_range)) + { + if constexpr (_UseCache) + if (__cached_begin_.__has_value()) + return *__cached_begin_; + + auto __tmp = ranges::next(ranges::begin(__base_), __count_, ranges::end(__base_)); + if constexpr (_UseCache) + __cached_begin_.__set(__tmp); + return __tmp; + } + + _LIBCPP_HIDE_FROM_ABI + constexpr auto begin() const + requires random_access_range && sized_range + { + return ranges::next(ranges::begin(__base_), __count_, ranges::end(__base_)); + } + + _LIBCPP_HIDE_FROM_ABI + constexpr auto end() + requires (!__simple_view<_View>) + { return ranges::end(__base_); } + + _LIBCPP_HIDE_FROM_ABI + constexpr auto end() const + requires range + { return ranges::end(__base_); } + + _LIBCPP_HIDE_FROM_ABI + static constexpr auto __size(auto& __self) { + const auto __s = ranges::size(__self.__base_); + const auto __c = static_cast(__self.__count_); + return __s < __c ? 0 : __s - __c; + } + + _LIBCPP_HIDE_FROM_ABI + constexpr auto size() + requires sized_range<_View> + { return __size(*this); } + + _LIBCPP_HIDE_FROM_ABI + constexpr auto size() const + requires sized_range + { return __size(*this); } + }; + + template + drop_view(_Range&&, range_difference_t<_Range>) -> drop_view>; + + template + inline constexpr bool enable_borrowed_range> = enable_borrowed_range<_Tp>; +} // namespace ranges + +#endif // !defined(_LIBCPP_HAS_NO_RANGES) + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___RANGES_DROP_VIEW_H diff --git a/lib/libcxx/include/__ranges/empty.h b/lib/libcxx/include/__ranges/empty.h new file mode 100644 index 000000000..73892a8c1 --- /dev/null +++ b/lib/libcxx/include/__ranges/empty.h @@ -0,0 +1,86 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// +#ifndef _LIBCPP___RANGES_EMPTY_H +#define _LIBCPP___RANGES_EMPTY_H + +#include <__config> +#include <__iterator/concepts.h> +#include <__ranges/access.h> +#include <__ranges/size.h> +#include <__utility/forward.h> +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +#if !defined(_LIBCPP_HAS_NO_RANGES) + +// clang-format off +namespace ranges { +// [range.prim.empty] +namespace __empty { + template + concept __member_empty = requires(_Tp&& __t) { + bool(_VSTD::forward<_Tp>(__t).empty()); + }; + + template + concept __can_invoke_size = + !__member_empty<_Tp> && + requires(_Tp&& __t) { ranges::size(_VSTD::forward<_Tp>(__t)); }; + + template + concept __can_compare_begin_end = + !__member_empty<_Tp> && + !__can_invoke_size<_Tp> && + requires(_Tp&& __t) { + bool(ranges::begin(__t) == ranges::end(__t)); + { ranges::begin(__t) } -> forward_iterator; + }; + + struct __fn { + template <__member_empty _Tp> + [[nodiscard]] _LIBCPP_HIDE_FROM_ABI constexpr bool operator()(_Tp&& __t) const + noexcept(noexcept(bool(__t.empty()))) { + return __t.empty(); + } + + template <__can_invoke_size _Tp> + [[nodiscard]] _LIBCPP_HIDE_FROM_ABI constexpr bool operator()(_Tp&& __t) const + noexcept(noexcept(ranges::size(_VSTD::forward<_Tp>(__t)))) { + return ranges::size(_VSTD::forward<_Tp>(__t)) == 0; + } + + template<__can_compare_begin_end _Tp> + [[nodiscard]] _LIBCPP_HIDE_FROM_ABI constexpr bool operator()(_Tp&& __t) const + noexcept(noexcept(bool(ranges::begin(__t) == ranges::end(__t)))) { + return ranges::begin(__t) == ranges::end(__t); + } + }; +} + +inline namespace __cpo { + inline constexpr auto empty = __empty::__fn{}; +} // namespace __cpo +} // namespace ranges +// clang-format off + +#endif // !defined(_LIBCPP_HAS_NO_RANGES) + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___RANGES_EMPTY_H diff --git a/lib/libcxx/include/__ranges/empty_view.h b/lib/libcxx/include/__ranges/empty_view.h new file mode 100644 index 000000000..7c0f307c8 --- /dev/null +++ b/lib/libcxx/include/__ranges/empty_view.h @@ -0,0 +1,46 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// +#ifndef _LIBCPP___RANGES_EMPTY_VIEW_H +#define _LIBCPP___RANGES_EMPTY_VIEW_H + +#include <__config> +#include <__ranges/view_interface.h> +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +#if !defined(_LIBCPP_HAS_NO_RANGES) + +namespace ranges { + template + requires is_object_v<_Tp> + class empty_view : public view_interface> { + public: + _LIBCPP_HIDE_FROM_ABI static constexpr _Tp* begin() noexcept { return nullptr; } + _LIBCPP_HIDE_FROM_ABI static constexpr _Tp* end() noexcept { return nullptr; } + _LIBCPP_HIDE_FROM_ABI static constexpr _Tp* data() noexcept { return nullptr; } + _LIBCPP_HIDE_FROM_ABI static constexpr size_t size() noexcept { return 0; } + _LIBCPP_HIDE_FROM_ABI static constexpr bool empty() noexcept { return true; } + }; +} // namespace ranges + +#endif // !defined(_LIBCPP_HAS_NO_RANGES) + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___RANGES_EMPTY_VIEW_H diff --git a/lib/libcxx/include/__ranges/enable_borrowed_range.h b/lib/libcxx/include/__ranges/enable_borrowed_range.h new file mode 100644 index 000000000..618b2223c --- /dev/null +++ b/lib/libcxx/include/__ranges/enable_borrowed_range.h @@ -0,0 +1,46 @@ +// -*- C++ -*- +//===------------------ __ranges/enable_borrowed_range.h ------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___RANGES_ENABLE_BORROWED_RANGE_H +#define _LIBCPP___RANGES_ENABLE_BORROWED_RANGE_H + +// These customization variables are used in and . The +// separate header is used to avoid including the entire header in +// and . + +#include <__config> + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +#if _LIBCPP_STD_VER > 17 && !defined(_LIBCPP_HAS_NO_RANGES) + +namespace ranges +{ + +// [range.range], ranges + +template +inline constexpr bool enable_borrowed_range = false; + +} // namespace ranges + +#endif // _LIBCPP_STD_VER > 17 && !defined(_LIBCPP_HAS_NO_RANGES) + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___RANGES_ENABLE_BORROWED_RANGE_H diff --git a/lib/libcxx/include/__ranges/enable_view.h b/lib/libcxx/include/__ranges/enable_view.h new file mode 100644 index 000000000..2628d51ce --- /dev/null +++ b/lib/libcxx/include/__ranges/enable_view.h @@ -0,0 +1,42 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___RANGES_ENABLE_VIEW_H +#define _LIBCPP___RANGES_ENABLE_VIEW_H + +#include <__config> +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +#if !defined(_LIBCPP_HAS_NO_RANGES) + +namespace ranges { + +struct view_base { }; + +template +inline constexpr bool enable_view = derived_from<_Tp, view_base>; + +} // end namespace ranges + +#endif // !_LIBCPP_HAS_NO_RANGES + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___RANGES_ENABLE_VIEW_H diff --git a/lib/libcxx/include/__ranges/non_propagating_cache.h b/lib/libcxx/include/__ranges/non_propagating_cache.h new file mode 100644 index 000000000..878f7070a --- /dev/null +++ b/lib/libcxx/include/__ranges/non_propagating_cache.h @@ -0,0 +1,99 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// +#ifndef _LIBCPP___RANGES_NON_PROPAGATING_CACHE_H +#define _LIBCPP___RANGES_NON_PROPAGATING_CACHE_H + +#include <__config> +#include <__iterator/concepts.h> // indirectly_readable +#include <__iterator/iterator_traits.h> // iter_reference_t +#include <__memory/addressof.h> +#include // constructible_from +#include +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +// clang-format off + +#if !defined(_LIBCPP_HAS_NO_RANGES) + +namespace ranges { + // __non_propagating_cache is a helper type that allows storing an optional value in it, + // but which does not copy the source's value when it is copy constructed/assigned to, + // and which resets the source's value when it is moved-from. + // + // This type is used as an implementation detail of some views that need to cache the + // result of `begin()` in order to provide an amortized O(1) begin() method. Typically, + // we don't want to propagate the value of the cache upon copy because the cached iterator + // may refer to internal details of the source view. + template + requires is_object_v<_Tp> + class _LIBCPP_TEMPLATE_VIS __non_propagating_cache { + optional<_Tp> __value_ = nullopt; + + public: + _LIBCPP_HIDE_FROM_ABI __non_propagating_cache() = default; + + _LIBCPP_HIDE_FROM_ABI + constexpr __non_propagating_cache(__non_propagating_cache const&) noexcept + : __value_(nullopt) + { } + + _LIBCPP_HIDE_FROM_ABI + constexpr __non_propagating_cache(__non_propagating_cache&& __other) noexcept + : __value_(nullopt) + { + __other.__value_.reset(); + } + + _LIBCPP_HIDE_FROM_ABI + constexpr __non_propagating_cache& operator=(__non_propagating_cache const& __other) noexcept { + if (this != _VSTD::addressof(__other)) { + __value_.reset(); + } + return *this; + } + + _LIBCPP_HIDE_FROM_ABI + constexpr __non_propagating_cache& operator=(__non_propagating_cache&& __other) noexcept { + __value_.reset(); + __other.__value_.reset(); + return *this; + } + + _LIBCPP_HIDE_FROM_ABI + constexpr _Tp& operator*() { return *__value_; } + _LIBCPP_HIDE_FROM_ABI + constexpr _Tp const& operator*() const { return *__value_; } + + _LIBCPP_HIDE_FROM_ABI + constexpr bool __has_value() const { return __value_.has_value(); } + _LIBCPP_HIDE_FROM_ABI + constexpr void __set(_Tp const& __value) { __value_.emplace(__value); } + _LIBCPP_HIDE_FROM_ABI + constexpr void __set(_Tp&& __value) { __value_.emplace(_VSTD::move(__value)); } + }; + + struct __empty_cache { }; +} // namespace ranges + +#endif // _LIBCPP_STD_VER > 17 && !defined(_LIBCPP_HAS_NO_RANGES) + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___RANGES_NON_PROPAGATING_CACHE_H diff --git a/lib/libcxx/include/__ranges/ref_view.h b/lib/libcxx/include/__ranges/ref_view.h new file mode 100644 index 000000000..fb45a3598 --- /dev/null +++ b/lib/libcxx/include/__ranges/ref_view.h @@ -0,0 +1,87 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// +#ifndef _LIBCPP___RANGES_REF_VIEW_H +#define _LIBCPP___RANGES_REF_VIEW_H + +#include <__config> +#include <__iterator/concepts.h> +#include <__iterator/incrementable_traits.h> +#include <__iterator/iterator_traits.h> +#include <__memory/addressof.h> +#include <__ranges/access.h> +#include <__ranges/concepts.h> +#include <__ranges/data.h> +#include <__ranges/empty.h> +#include <__ranges/size.h> +#include <__ranges/view_interface.h> +#include +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +#if !defined(_LIBCPP_HAS_NO_RANGES) + +namespace ranges { + template + requires is_object_v<_Range> + class ref_view : public view_interface> { + _Range *__range_; + + static void __fun(_Range&); + static void __fun(_Range&&) = delete; + +public: + template + requires __different_from<_Tp, ref_view> && + convertible_to<_Tp, _Range&> && requires { __fun(declval<_Tp>()); } + _LIBCPP_HIDE_FROM_ABI + constexpr ref_view(_Tp&& __t) + : __range_(_VSTD::addressof(static_cast<_Range&>(_VSTD::forward<_Tp>(__t)))) + {} + + _LIBCPP_HIDE_FROM_ABI constexpr _Range& base() const { return *__range_; } + + _LIBCPP_HIDE_FROM_ABI constexpr iterator_t<_Range> begin() const { return ranges::begin(*__range_); } + _LIBCPP_HIDE_FROM_ABI constexpr sentinel_t<_Range> end() const { return ranges::end(*__range_); } + + _LIBCPP_HIDE_FROM_ABI + constexpr bool empty() const + requires requires { ranges::empty(*__range_); } + { return ranges::empty(*__range_); } + + _LIBCPP_HIDE_FROM_ABI + constexpr auto size() const + requires sized_range<_Range> + { return ranges::size(*__range_); } + + _LIBCPP_HIDE_FROM_ABI + constexpr auto data() const + requires contiguous_range<_Range> + { return ranges::data(*__range_); } + }; + + template + ref_view(_Range&) -> ref_view<_Range>; + +} // namespace ranges + +#endif // !defined(_LIBCPP_HAS_NO_RANGES) + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___RANGES_REF_VIEW_H diff --git a/lib/libcxx/include/__ranges/size.h b/lib/libcxx/include/__ranges/size.h new file mode 100644 index 000000000..ce7183e15 --- /dev/null +++ b/lib/libcxx/include/__ranges/size.h @@ -0,0 +1,132 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// +#ifndef _LIBCPP___RANGES_SIZE_H +#define _LIBCPP___RANGES_SIZE_H + +#include <__config> +#include <__iterator/concepts.h> +#include <__iterator/iterator_traits.h> +#include <__ranges/access.h> +#include <__utility/__decay_copy.h> +#include <__utility/forward.h> +#include +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +#if !defined(_LIBCPP_HAS_NO_RANGES) + +// clang-format off +namespace ranges { +template +inline constexpr bool disable_sized_range = false; + +// [range.prim.size] +namespace __size { + void size(auto&) = delete; + void size(const auto&) = delete; + + template + concept __size_enabled = !disable_sized_range>; + + template + concept __member_size = __size_enabled<_Tp> && requires(_Tp&& __t) { + { _VSTD::__decay_copy(_VSTD::forward<_Tp>(__t).size()) } -> __integer_like; + }; + + template + concept __unqualified_size = + __size_enabled<_Tp> && + !__member_size<_Tp> && + __class_or_enum> && + requires(_Tp&& __t) { + { _VSTD::__decay_copy(size(_VSTD::forward<_Tp>(__t))) } -> __integer_like; + }; + + template + concept __difference = + !__member_size<_Tp> && + !__unqualified_size<_Tp> && + __class_or_enum> && + requires(_Tp&& __t) { + { ranges::begin(__t) } -> forward_iterator; + { ranges::end(__t) } -> sized_sentinel_for()))>; + }; + + struct __fn { + template + [[nodiscard]] _LIBCPP_HIDE_FROM_ABI constexpr size_t operator()(_Tp (&&)[_Sz]) const noexcept { + return _Sz; + } + + template + [[nodiscard]] _LIBCPP_HIDE_FROM_ABI constexpr size_t operator()(_Tp (&)[_Sz]) const noexcept { + return _Sz; + } + + template <__member_size _Tp> + [[nodiscard]] _LIBCPP_HIDE_FROM_ABI constexpr __integer_like auto operator()(_Tp&& __t) const + noexcept(noexcept(_VSTD::forward<_Tp>(__t).size())) { + return _VSTD::forward<_Tp>(__t).size(); + } + + template <__unqualified_size _Tp> + [[nodiscard]] _LIBCPP_HIDE_FROM_ABI constexpr __integer_like auto operator()(_Tp&& __t) const + noexcept(noexcept(size(_VSTD::forward<_Tp>(__t)))) { + return size(_VSTD::forward<_Tp>(__t)); + } + + template<__difference _Tp> + [[nodiscard]] _LIBCPP_HIDE_FROM_ABI constexpr __integer_like auto operator()(_Tp&& __t) const + noexcept(noexcept(ranges::end(__t) - ranges::begin(__t))) { + return _VSTD::__to_unsigned_like(ranges::end(__t) - ranges::begin(__t)); + } + }; +} // end namespace __size + +inline namespace __cpo { + inline constexpr auto size = __size::__fn{}; +} // namespace __cpo + +namespace __ssize { + struct __fn { + template + requires requires (_Tp&& __t) { ranges::size(__t); } + [[nodiscard]] _LIBCPP_HIDE_FROM_ABI constexpr integral auto operator()(_Tp&& __t) const + noexcept(noexcept(ranges::size(__t))) { + using _Signed = make_signed_t; + if constexpr (sizeof(ptrdiff_t) > sizeof(_Signed)) + return static_cast(ranges::size(__t)); + else + return static_cast<_Signed>(ranges::size(__t)); + } + }; +} + +inline namespace __cpo { + inline constexpr const auto ssize = __ssize::__fn{}; +} // namespace __cpo +} // namespace ranges + +// clang-format off + +#endif // !defined(_LIBCPP_HAS_NO_RANGES) + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___RANGES_SIZE_H diff --git a/lib/libcxx/include/__ranges/subrange.h b/lib/libcxx/include/__ranges/subrange.h new file mode 100644 index 000000000..25d333db0 --- /dev/null +++ b/lib/libcxx/include/__ranges/subrange.h @@ -0,0 +1,267 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// +#ifndef _LIBCPP___RANGES_SUBRANGE_H +#define _LIBCPP___RANGES_SUBRANGE_H + +#include <__config> +#include <__iterator/concepts.h> +#include <__iterator/incrementable_traits.h> +#include <__iterator/iterator_traits.h> +#include <__iterator/advance.h> +#include <__ranges/access.h> +#include <__ranges/concepts.h> +#include <__ranges/dangling.h> +#include <__ranges/enable_borrowed_range.h> +#include <__ranges/size.h> +#include <__ranges/view_interface.h> +#include +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +#if !defined(_LIBCPP_HAS_NO_RANGES) + +// clang-format off +namespace ranges { + template + concept __convertible_to_non_slicing = + convertible_to<_From, _To> && + // If they're both pointers, they must have the same element type. + !(is_pointer_v> && + is_pointer_v> && + __different_from>, remove_pointer_t>>); + + template + concept __pair_like = + !is_reference_v<_Tp> && requires(_Tp __t) { + typename tuple_size<_Tp>::type; // Ensures `tuple_size` is complete. + requires derived_from, integral_constant>; + typename tuple_element_t<0, remove_const_t<_Tp>>; + typename tuple_element_t<1, remove_const_t<_Tp>>; + { _VSTD::get<0>(__t) } -> convertible_to&>; + { _VSTD::get<1>(__t) } -> convertible_to&>; + }; + + template + concept __pair_like_convertible_from = + !range<_Pair> && __pair_like<_Pair> && + constructible_from<_Pair, _Iter, _Sent> && + __convertible_to_non_slicing<_Iter, tuple_element_t<0, _Pair>> && + convertible_to<_Sent, tuple_element_t<1, _Pair>>; + + enum class _LIBCPP_ENUM_VIS subrange_kind : bool { unsized, sized }; + + template + struct __subrange_base { + static constexpr bool __store_size = false; + _Iter __begin_ = _Iter(); + _Sent __end_ = _Sent(); + + _LIBCPP_HIDE_FROM_ABI + constexpr __subrange_base() = default; + + _LIBCPP_HIDE_FROM_ABI + constexpr __subrange_base(_Iter __iter, _Sent __sent, make_unsigned_t> = 0) + : __begin_(_VSTD::move(__iter)), __end_(__sent) { } + }; + + template + struct __subrange_base<_Iter, _Sent, true> { + static constexpr bool __store_size = true; + _Iter __begin_ = _Iter(); + _Sent __end_ = _Sent(); + make_unsigned_t> __size_ = 0; + + _LIBCPP_HIDE_FROM_ABI + constexpr __subrange_base() = default; + + _LIBCPP_HIDE_FROM_ABI + constexpr __subrange_base(_Iter __iter, _Sent __sent, decltype(__size_) __size) + : __begin_(_VSTD::move(__iter)), __end_(__sent), __size_(__size) { } + }; + + template _Sent = _Iter, + subrange_kind _Kind = sized_sentinel_for<_Sent, _Iter> + ? subrange_kind::sized + : subrange_kind::unsized> + requires (_Kind == subrange_kind::sized || !sized_sentinel_for<_Sent, _Iter>) + struct _LIBCPP_TEMPLATE_VIS subrange + : public view_interface>, + private __subrange_base<_Iter, _Sent, _Kind == subrange_kind::sized && !sized_sentinel_for<_Sent, _Iter>> { + + using _Base = __subrange_base<_Iter, _Sent, _Kind == subrange_kind::sized && !sized_sentinel_for<_Sent, _Iter>>; + + _LIBCPP_HIDE_FROM_ABI + subrange() requires default_initializable<_Iter> = default; + + _LIBCPP_HIDE_FROM_ABI + constexpr subrange(__convertible_to_non_slicing<_Iter> auto __iter, _Sent __sent) + requires (!_Base::__store_size) + : _Base(_VSTD::move(__iter), __sent) {} + + _LIBCPP_HIDE_FROM_ABI + constexpr subrange(__convertible_to_non_slicing<_Iter> auto __iter, _Sent __sent, + make_unsigned_t> __n) + requires (_Kind == subrange_kind::sized) + : _Base(_VSTD::move(__iter), __sent, __n) { } + + template<__different_from _Range> + requires borrowed_range<_Range> && + __convertible_to_non_slicing, _Iter> && + convertible_to, _Sent> + _LIBCPP_HIDE_FROM_ABI + constexpr subrange(_Range&& __range) + requires (!_Base::__store_size) + : subrange(ranges::begin(__range), ranges::end(__range)) { } + + template<__different_from _Range> + requires borrowed_range<_Range> && + __convertible_to_non_slicing, _Iter> && + convertible_to, _Sent> + _LIBCPP_HIDE_FROM_ABI + constexpr subrange(_Range&& __range) + requires _Base::__store_size && sized_range<_Range> + : subrange(__range, ranges::size(__range)) { } + + + template + requires __convertible_to_non_slicing, _Iter> && + convertible_to, _Sent> + _LIBCPP_HIDE_FROM_ABI + constexpr subrange(_Range&& __range, make_unsigned_t> __n) + requires (_Kind == subrange_kind::sized) + : subrange(ranges::begin(__range), ranges::end(__range), __n) { } + + template<__different_from _Pair> + requires __pair_like_convertible_from<_Pair, const _Iter&, const _Sent&> + _LIBCPP_HIDE_FROM_ABI + constexpr operator _Pair() const { return _Pair(this->__begin_, this->__end_); } + + _LIBCPP_HIDE_FROM_ABI + constexpr _Iter begin() const requires copyable<_Iter> { + return this->__begin_; + } + + [[nodiscard]] _LIBCPP_HIDE_FROM_ABI constexpr _Iter begin() requires (!copyable<_Iter>) { + return _VSTD::move(this->__begin_); + } + + _LIBCPP_HIDE_FROM_ABI + constexpr _Sent end() const { return this->__end_; } + + [[nodiscard]] _LIBCPP_HIDE_FROM_ABI constexpr bool empty() const { return this->__begin_ == this->__end_; } + + _LIBCPP_HIDE_FROM_ABI + constexpr make_unsigned_t> size() const + requires (_Kind == subrange_kind::sized) + { + if constexpr (_Base::__store_size) + return this->__size_; + else + return __to_unsigned_like(this->__end_ - this->__begin_); + } + + [[nodiscard]] _LIBCPP_HIDE_FROM_ABI constexpr subrange next(iter_difference_t<_Iter> __n = 1) const& + requires forward_iterator<_Iter> { + auto __tmp = *this; + __tmp.advance(__n); + return __tmp; + } + + [[nodiscard]] _LIBCPP_HIDE_FROM_ABI constexpr subrange next(iter_difference_t<_Iter> __n = 1) && { + advance(__n); + return _VSTD::move(*this); + } + + [[nodiscard]] _LIBCPP_HIDE_FROM_ABI constexpr subrange prev(iter_difference_t<_Iter> __n = 1) const + requires bidirectional_iterator<_Iter> { + auto __tmp = *this; + __tmp.advance(-__n); + return __tmp; + } + + _LIBCPP_HIDE_FROM_ABI + constexpr subrange& advance(iter_difference_t<_Iter> __n) { + if constexpr (bidirectional_iterator<_Iter>) { + if (__n < 0) { + ranges::advance(this->__begin_, __n); + if constexpr (_Base::__store_size) + this->__size_ += _VSTD::__to_unsigned_like(-__n); + return *this; + } + } + + auto __d = __n - ranges::advance(this->__begin_, __n, this->__end_); + if constexpr (_Base::__store_size) + this->__size_ -= _VSTD::__to_unsigned_like(__d); + return *this; + } + }; + + template _Sent> + subrange(_Iter, _Sent) -> subrange<_Iter, _Sent>; + + template _Sent> + subrange(_Iter, _Sent, make_unsigned_t>) + -> subrange<_Iter, _Sent, subrange_kind::sized>; + + template + subrange(_Range&&) -> subrange, sentinel_t<_Range>, + (sized_range<_Range> || sized_sentinel_for, iterator_t<_Range>>) + ? subrange_kind::sized : subrange_kind::unsized>; + + template + subrange(_Range&&, make_unsigned_t>) + -> subrange, sentinel_t<_Range>, subrange_kind::sized>; + + template + requires (_Index < 2) + _LIBCPP_HIDE_FROM_ABI + constexpr auto get(const subrange<_Iter, _Sent, _Kind>& __subrange) { + if constexpr (_Index == 0) + return __subrange.begin(); + else + return __subrange.end(); + } + + template + requires (_Index < 2) + _LIBCPP_HIDE_FROM_ABI + constexpr auto get(subrange<_Iter, _Sent, _Kind>&& __subrange) { + if constexpr (_Index == 0) + return __subrange.begin(); + else + return __subrange.end(); + } + + template + inline constexpr bool enable_borrowed_range> = true; + + template + using borrowed_subrange_t = _If, subrange >, dangling>; +} // namespace ranges + +using ranges::get; + +// clang-format off + +#endif // !defined(_LIBCPP_HAS_NO_RANGES) + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___RANGES_SUBRANGE_H diff --git a/lib/libcxx/include/__ranges/transform_view.h b/lib/libcxx/include/__ranges/transform_view.h new file mode 100644 index 000000000..4243dc036 --- /dev/null +++ b/lib/libcxx/include/__ranges/transform_view.h @@ -0,0 +1,408 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// +#ifndef _LIBCPP___RANGES_TRANSFORM_VIEW_H +#define _LIBCPP___RANGES_TRANSFORM_VIEW_H + +#include <__config> +#include <__iterator/concepts.h> +#include <__iterator/iter_swap.h> +#include <__iterator/iterator_traits.h> +#include <__ranges/access.h> +#include <__ranges/all.h> +#include <__ranges/concepts.h> +#include <__ranges/copyable_box.h> +#include <__ranges/empty.h> +#include <__ranges/size.h> +#include <__ranges/view_interface.h> +#include +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +#if !defined(_LIBCPP_HAS_NO_RANGES) + +namespace ranges { + +template +concept __transform_view_constraints = + view<_View> && is_object_v<_Fn> && + regular_invocable<_Fn&, range_reference_t<_View>> && + __referenceable>>; + +template + requires __transform_view_constraints<_View, _Fn> +class transform_view : public view_interface> { + template class __iterator; + template class __sentinel; + + [[no_unique_address]] __copyable_box<_Fn> __func_; + [[no_unique_address]] _View __base_ = _View(); + +public: + _LIBCPP_HIDE_FROM_ABI + transform_view() + requires default_initializable<_View> && default_initializable<_Fn> = default; + + _LIBCPP_HIDE_FROM_ABI + constexpr transform_view(_View __base, _Fn __func) + : __func_(_VSTD::in_place, _VSTD::move(__func)), __base_(_VSTD::move(__base)) {} + + _LIBCPP_HIDE_FROM_ABI + constexpr _View base() const& requires copy_constructible<_View> { return __base_; } + _LIBCPP_HIDE_FROM_ABI + constexpr _View base() && { return _VSTD::move(__base_); } + + _LIBCPP_HIDE_FROM_ABI + constexpr __iterator begin() { + return __iterator{*this, ranges::begin(__base_)}; + } + _LIBCPP_HIDE_FROM_ABI + constexpr __iterator begin() const + requires range && + regular_invocable> + { + return __iterator(*this, ranges::begin(__base_)); + } + + _LIBCPP_HIDE_FROM_ABI + constexpr __sentinel end() { + return __sentinel(ranges::end(__base_)); + } + _LIBCPP_HIDE_FROM_ABI + constexpr __iterator end() + requires common_range<_View> + { + return __iterator(*this, ranges::end(__base_)); + } + _LIBCPP_HIDE_FROM_ABI + constexpr __sentinel end() const + requires range && + regular_invocable> + { + return __sentinel(ranges::end(__base_)); + } + _LIBCPP_HIDE_FROM_ABI + constexpr __iterator end() const + requires common_range && + regular_invocable> + { + return __iterator(*this, ranges::end(__base_)); + } + + _LIBCPP_HIDE_FROM_ABI + constexpr auto size() requires sized_range<_View> { return ranges::size(__base_); } + _LIBCPP_HIDE_FROM_ABI + constexpr auto size() const requires sized_range { return ranges::size(__base_); } +}; + +template +transform_view(_Range&&, _Fn) -> transform_view, _Fn>; + +template +struct __transform_view_iterator_concept { using type = input_iterator_tag; }; + +template +struct __transform_view_iterator_concept<_View> { using type = random_access_iterator_tag; }; + +template +struct __transform_view_iterator_concept<_View> { using type = bidirectional_iterator_tag; }; + +template +struct __transform_view_iterator_concept<_View> { using type = forward_iterator_tag; }; + +template +struct __transform_view_iterator_category_base {}; + +template +struct __transform_view_iterator_category_base<_View, _Fn> { + using _Cat = typename iterator_traits>::iterator_category; + + using iterator_category = conditional_t< + is_lvalue_reference_v>>, + conditional_t< + derived_from<_Cat, contiguous_iterator_tag>, + random_access_iterator_tag, + _Cat + >, + input_iterator_tag + >; +}; + +template + requires __transform_view_constraints<_View, _Fn> +template +class transform_view<_View, _Fn>::__iterator + : public __transform_view_iterator_category_base<_View, _Fn> { + + using _Parent = __maybe_const<_Const, transform_view>; + using _Base = __maybe_const<_Const, _View>; + + _Parent *__parent_ = nullptr; + + template + friend class transform_view<_View, _Fn>::__iterator; + + template + friend class transform_view<_View, _Fn>::__sentinel; + +public: + iterator_t<_Base> __current_ = iterator_t<_Base>(); + + using iterator_concept = typename __transform_view_iterator_concept<_View>::type; + using value_type = remove_cvref_t>>; + using difference_type = range_difference_t<_Base>; + + _LIBCPP_HIDE_FROM_ABI + __iterator() requires default_initializable> = default; + + _LIBCPP_HIDE_FROM_ABI + constexpr __iterator(_Parent& __parent, iterator_t<_Base> __current) + : __parent_(_VSTD::addressof(__parent)), __current_(_VSTD::move(__current)) {} + + // Note: `__i` should always be `__iterator`, but directly using + // `__iterator` is ill-formed when `_Const` is false + // (see http://wg21.link/class.copy.ctor#5). + _LIBCPP_HIDE_FROM_ABI + constexpr __iterator(__iterator __i) + requires _Const && convertible_to, iterator_t<_Base>> + : __parent_(__i.__parent_), __current_(_VSTD::move(__i.__current_)) {} + + _LIBCPP_HIDE_FROM_ABI + constexpr iterator_t<_Base> base() const& + requires copyable> + { + return __current_; + } + + _LIBCPP_HIDE_FROM_ABI + constexpr iterator_t<_Base> base() && { + return _VSTD::move(__current_); + } + + _LIBCPP_HIDE_FROM_ABI + constexpr decltype(auto) operator*() const + noexcept(noexcept(_VSTD::invoke(*__parent_->__func_, *__current_))) + { + return _VSTD::invoke(*__parent_->__func_, *__current_); + } + + _LIBCPP_HIDE_FROM_ABI + constexpr __iterator& operator++() { + ++__current_; + return *this; + } + + _LIBCPP_HIDE_FROM_ABI + constexpr void operator++(int) { ++__current_; } + + _LIBCPP_HIDE_FROM_ABI + constexpr __iterator operator++(int) + requires forward_range<_Base> + { + auto __tmp = *this; + ++*this; + return __tmp; + } + + _LIBCPP_HIDE_FROM_ABI + constexpr __iterator& operator--() + requires bidirectional_range<_Base> + { + --__current_; + return *this; + } + + _LIBCPP_HIDE_FROM_ABI + constexpr __iterator operator--(int) + requires bidirectional_range<_Base> + { + auto __tmp = *this; + --*this; + return __tmp; + } + + _LIBCPP_HIDE_FROM_ABI + constexpr __iterator& operator+=(difference_type __n) + requires random_access_range<_Base> + { + __current_ += __n; + return *this; + } + + _LIBCPP_HIDE_FROM_ABI + constexpr __iterator& operator-=(difference_type __n) + requires random_access_range<_Base> + { + __current_ -= __n; + return *this; + } + + _LIBCPP_HIDE_FROM_ABI + constexpr decltype(auto) operator[](difference_type __n) const + noexcept(noexcept(_VSTD::invoke(*__parent_->__func_, __current_[__n]))) + requires random_access_range<_Base> + { + return _VSTD::invoke(*__parent_->__func_, __current_[__n]); + } + + _LIBCPP_HIDE_FROM_ABI + friend constexpr bool operator==(const __iterator& __x, const __iterator& __y) + requires equality_comparable> + { + return __x.__current_ == __y.__current_; + } + + _LIBCPP_HIDE_FROM_ABI + friend constexpr bool operator<(const __iterator& __x, const __iterator& __y) + requires random_access_range<_Base> + { + return __x.__current_ < __y.__current_; + } + + _LIBCPP_HIDE_FROM_ABI + friend constexpr bool operator>(const __iterator& __x, const __iterator& __y) + requires random_access_range<_Base> + { + return __x.__current_ > __y.__current_; + } + + _LIBCPP_HIDE_FROM_ABI + friend constexpr bool operator<=(const __iterator& __x, const __iterator& __y) + requires random_access_range<_Base> + { + return __x.__current_ <= __y.__current_; + } + + _LIBCPP_HIDE_FROM_ABI + friend constexpr bool operator>=(const __iterator& __x, const __iterator& __y) + requires random_access_range<_Base> + { + return __x.__current_ >= __y.__current_; + } + +// TODO: Fix this as soon as soon as three_way_comparable is implemented. +// _LIBCPP_HIDE_FROM_ABI +// friend constexpr auto operator<=>(const __iterator& __x, const __iterator& __y) +// requires random_access_range<_Base> && three_way_comparable> +// { +// return __x.__current_ <=> __y.__current_; +// } + + _LIBCPP_HIDE_FROM_ABI + friend constexpr __iterator operator+(__iterator __i, difference_type __n) + requires random_access_range<_Base> + { + return __iterator{*__i.__parent_, __i.__current_ + __n}; + } + + _LIBCPP_HIDE_FROM_ABI + friend constexpr __iterator operator+(difference_type __n, __iterator __i) + requires random_access_range<_Base> + { + return __iterator{*__i.__parent_, __i.__current_ + __n}; + } + + _LIBCPP_HIDE_FROM_ABI + friend constexpr __iterator operator-(__iterator __i, difference_type __n) + requires random_access_range<_Base> + { + return __iterator{*__i.__parent_, __i.__current_ - __n}; + } + + _LIBCPP_HIDE_FROM_ABI + friend constexpr difference_type operator-(const __iterator& __x, const __iterator& __y) + requires sized_sentinel_for, iterator_t<_Base>> + { + return __x.__current_ - __y.__current_; + } + + _LIBCPP_HIDE_FROM_ABI + friend constexpr decltype(auto) iter_move(const __iterator& __i) + noexcept(noexcept(*__i)) + { + if constexpr (is_lvalue_reference_v) + return _VSTD::move(*__i); + else + return *__i; + } +}; + +template + requires __transform_view_constraints<_View, _Fn> +template +class transform_view<_View, _Fn>::__sentinel { + using _Parent = __maybe_const<_Const, transform_view>; + using _Base = __maybe_const<_Const, _View>; + + sentinel_t<_Base> __end_ = sentinel_t<_Base>(); + + template + friend class transform_view<_View, _Fn>::__iterator; + + template + friend class transform_view<_View, _Fn>::__sentinel; + +public: + _LIBCPP_HIDE_FROM_ABI + __sentinel() = default; + + _LIBCPP_HIDE_FROM_ABI + constexpr explicit __sentinel(sentinel_t<_Base> __end) : __end_(__end) {} + + // Note: `__i` should always be `__sentinel`, but directly using + // `__sentinel` is ill-formed when `_Const` is false + // (see http://wg21.link/class.copy.ctor#5). + _LIBCPP_HIDE_FROM_ABI + constexpr __sentinel(__sentinel __i) + requires _Const && convertible_to, sentinel_t<_Base>> + : __end_(_VSTD::move(__i.__end_)) {} + + _LIBCPP_HIDE_FROM_ABI + constexpr sentinel_t<_Base> base() const { return __end_; } + + template + requires sentinel_for, iterator_t<__maybe_const<_OtherConst, _View>>> + _LIBCPP_HIDE_FROM_ABI + friend constexpr bool operator==(const __iterator<_OtherConst>& __x, const __sentinel& __y) { + return __x.__current_ == __y.__end_; + } + + template + requires sized_sentinel_for, iterator_t<__maybe_const<_OtherConst, _View>>> + _LIBCPP_HIDE_FROM_ABI + friend constexpr range_difference_t<__maybe_const<_OtherConst, _View>> + operator-(const __iterator<_OtherConst>& __x, const __sentinel& __y) { + return __x.__current_ - __y.__end_; + } + + template + requires sized_sentinel_for, iterator_t<__maybe_const<_OtherConst, _View>>> + _LIBCPP_HIDE_FROM_ABI + friend constexpr range_difference_t<__maybe_const<_OtherConst, _View>> + operator-(const __sentinel& __x, const __iterator<_OtherConst>& __y) { + return __x.__end_ - __y.__current_; + } +}; + +} // namespace ranges + +#endif // !defined(_LIBCPP_HAS_NO_RANGES) + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___RANGES_TRANSFORM_VIEW_H diff --git a/lib/libcxx/include/__ranges/view_interface.h b/lib/libcxx/include/__ranges/view_interface.h new file mode 100644 index 000000000..62cc5fd24 --- /dev/null +++ b/lib/libcxx/include/__ranges/view_interface.h @@ -0,0 +1,198 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// +#ifndef _LIBCPP___RANGES_VIEW_INTERFACE_H +#define _LIBCPP___RANGES_VIEW_INTERFACE_H + +#include <__config> +#include <__iterator/concepts.h> +#include <__iterator/iterator_traits.h> +#include <__iterator/prev.h> +#include <__memory/pointer_traits.h> +#include <__ranges/access.h> +#include <__ranges/concepts.h> +#include <__ranges/empty.h> +#include <__ranges/enable_view.h> +#include +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +#if !defined(_LIBCPP_HAS_NO_RANGES) + +namespace ranges { + +template +concept __can_empty = requires(_Tp __t) { ranges::empty(__t); }; + +template +void __implicitly_convert_to(type_identity_t<_Tp>) noexcept; + +template + requires is_class_v<_Derived> && same_as<_Derived, remove_cv_t<_Derived>> +class view_interface : public view_base { + _LIBCPP_HIDE_FROM_ABI + constexpr _Derived& __derived() noexcept { + return static_cast<_Derived&>(*this); + } + + _LIBCPP_HIDE_FROM_ABI + constexpr _Derived const& __derived() const noexcept { + return static_cast<_Derived const&>(*this); + } + +public: + template + [[nodiscard]] _LIBCPP_HIDE_FROM_ABI constexpr bool empty() + noexcept(noexcept(__implicitly_convert_to(ranges::begin(__derived()) == ranges::end(__derived())))) + requires forward_range<_D2> + { + return ranges::begin(__derived()) == ranges::end(__derived()); + } + + template + [[nodiscard]] _LIBCPP_HIDE_FROM_ABI constexpr bool empty() const + noexcept(noexcept(__implicitly_convert_to(ranges::begin(__derived()) == ranges::end(__derived())))) + requires forward_range + { + return ranges::begin(__derived()) == ranges::end(__derived()); + } + + template + _LIBCPP_HIDE_FROM_ABI + constexpr explicit operator bool() + noexcept(noexcept(ranges::empty(declval<_D2>()))) + requires __can_empty<_D2> + { + return !ranges::empty(__derived()); + } + + template + _LIBCPP_HIDE_FROM_ABI + constexpr explicit operator bool() const + noexcept(noexcept(ranges::empty(declval()))) + requires __can_empty + { + return !ranges::empty(__derived()); + } + + template + _LIBCPP_HIDE_FROM_ABI + constexpr auto data() + noexcept(noexcept(_VSTD::to_address(ranges::begin(__derived())))) + requires contiguous_iterator> + { + return _VSTD::to_address(ranges::begin(__derived())); + } + + template + _LIBCPP_HIDE_FROM_ABI + constexpr auto data() const + noexcept(noexcept(_VSTD::to_address(ranges::begin(__derived())))) + requires range && contiguous_iterator> + { + return _VSTD::to_address(ranges::begin(__derived())); + } + + template + _LIBCPP_HIDE_FROM_ABI + constexpr auto size() + noexcept(noexcept(ranges::end(__derived()) - ranges::begin(__derived()))) + requires forward_range<_D2> + && sized_sentinel_for, iterator_t<_D2>> + { + return ranges::end(__derived()) - ranges::begin(__derived()); + } + + template + _LIBCPP_HIDE_FROM_ABI + constexpr auto size() const + noexcept(noexcept(ranges::end(__derived()) - ranges::begin(__derived()))) + requires forward_range + && sized_sentinel_for, iterator_t> + { + return ranges::end(__derived()) - ranges::begin(__derived()); + } + + template + _LIBCPP_HIDE_FROM_ABI + constexpr decltype(auto) front() + noexcept(noexcept(*ranges::begin(__derived()))) + requires forward_range<_D2> + { + _LIBCPP_ASSERT(!empty(), + "Precondition `!empty()` not satisfied. `.front()` called on an empty view."); + return *ranges::begin(__derived()); + } + + template + _LIBCPP_HIDE_FROM_ABI + constexpr decltype(auto) front() const + noexcept(noexcept(*ranges::begin(__derived()))) + requires forward_range + { + _LIBCPP_ASSERT(!empty(), + "Precondition `!empty()` not satisfied. `.front()` called on an empty view."); + return *ranges::begin(__derived()); + } + + template + _LIBCPP_HIDE_FROM_ABI + constexpr decltype(auto) back() + noexcept(noexcept(*ranges::prev(ranges::end(__derived())))) + requires bidirectional_range<_D2> && common_range<_D2> + { + _LIBCPP_ASSERT(!empty(), + "Precondition `!empty()` not satisfied. `.back()` called on an empty view."); + return *ranges::prev(ranges::end(__derived())); + } + + template + _LIBCPP_HIDE_FROM_ABI + constexpr decltype(auto) back() const + noexcept(noexcept(*ranges::prev(ranges::end(__derived())))) + requires bidirectional_range && common_range + { + _LIBCPP_ASSERT(!empty(), + "Precondition `!empty()` not satisfied. `.back()` called on an empty view."); + return *ranges::prev(ranges::end(__derived())); + } + + template + _LIBCPP_HIDE_FROM_ABI + constexpr decltype(auto) operator[](range_difference_t<_RARange> __index) + noexcept(noexcept(ranges::begin(__derived())[__index])) + { + return ranges::begin(__derived())[__index]; + } + + template + _LIBCPP_HIDE_FROM_ABI + constexpr decltype(auto) operator[](range_difference_t<_RARange> __index) const + noexcept(noexcept(ranges::begin(__derived())[__index])) + { + return ranges::begin(__derived())[__index]; + } +}; + +} + +#endif // !defined(_LIBCPP_HAS_NO_RANGES) + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___RANGES_VIEW_INTERFACE_H diff --git a/lib/libcxx/include/__split_buffer b/lib/libcxx/include/__split_buffer index 20480d19d..901c0374a 100644 --- a/lib/libcxx/include/__split_buffer +++ b/lib/libcxx/include/__split_buffer @@ -3,8 +3,9 @@ #define _LIBCPP_SPLIT_BUFFER #include <__config> -#include +#include <__utility/forward.h> #include +#include #if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) #pragma GCC system_header @@ -20,8 +21,8 @@ template class __split_buffer_common { protected: - void __throw_length_error() const; - void __throw_out_of_range() const; + _LIBCPP_NORETURN void __throw_length_error() const; + _LIBCPP_NORETURN void __throw_out_of_range() const; }; template > @@ -444,7 +445,7 @@ __split_buffer<_Tp, _Allocator>::shrink_to_fit() _NOEXCEPT #ifndef _LIBCPP_NO_EXCEPTIONS try { -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS __split_buffer __t(size(), 0, __alloc()); __t.__construct_at_end(move_iterator(__begin_), move_iterator(__end_)); @@ -458,7 +459,7 @@ __split_buffer<_Tp, _Allocator>::shrink_to_fit() _NOEXCEPT catch (...) { } -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS } } @@ -625,4 +626,4 @@ _LIBCPP_END_NAMESPACE_STD _LIBCPP_POP_MACROS -#endif // _LIBCPP_SPLIT_BUFFER +#endif // _LIBCPP_SPLIT_BUFFER diff --git a/lib/libcxx/include/__std_stream b/lib/libcxx/include/__std_stream index 5a9a470a9..65e90b789 100644 --- a/lib/libcxx/include/__std_stream +++ b/lib/libcxx/include/__std_stream @@ -11,10 +11,10 @@ #define _LIBCPP___STD_STREAM #include <__config> -#include -#include #include <__locale> #include +#include +#include #if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) #pragma GCC system_header @@ -358,4 +358,4 @@ _LIBCPP_END_NAMESPACE_STD _LIBCPP_POP_MACROS -#endif // _LIBCPP___STD_STREAM +#endif // _LIBCPP___STD_STREAM diff --git a/lib/libcxx/include/__string b/lib/libcxx/include/__string index d8b672e4c..b77a7fb4f 100644 --- a/lib/libcxx/include/__string +++ b/lib/libcxx/include/__string @@ -10,55 +10,21 @@ #ifndef _LIBCPP___STRING #define _LIBCPP___STRING -/* - string synopsis - -namespace std -{ - -template -struct char_traits -{ - typedef charT char_type; - typedef ... int_type; - typedef streamoff off_type; - typedef streampos pos_type; - typedef mbstate_t state_type; - - static constexpr void assign(char_type& c1, const char_type& c2) noexcept; - static constexpr bool eq(char_type c1, char_type c2) noexcept; - static constexpr bool lt(char_type c1, char_type c2) noexcept; - - static constexpr int compare(const char_type* s1, const char_type* s2, size_t n); - static constexpr size_t length(const char_type* s); - static constexpr const char_type* - find(const char_type* s, size_t n, const char_type& a); - - static constexpr char_type* move(char_type* s1, const char_type* s2, size_t n); // constexpr in C++20 - static constexpr char_type* copy(char_type* s1, const char_type* s2, size_t n); // constexpr in C++20 - static constexpr char_type* assign(char_type* s, size_t n, char_type a); // constexpr in C++20 - - static constexpr int_type not_eof(int_type c) noexcept; - static constexpr char_type to_char_type(int_type c) noexcept; - static constexpr int_type to_int_type(char_type c) noexcept; - static constexpr bool eq_int_type(int_type c1, int_type c2) noexcept; - static constexpr int_type eof() noexcept; -}; - -template <> struct char_traits; -template <> struct char_traits; -template <> struct char_traits; // c++20 - -} // std - -*/ - #include <__config> -#include // for search and min -#include // for EOF -#include // for memcpy -#include // for wmemcpy -#include // for __murmur2_or_cityhash +#include <__algorithm/copy.h> +#include <__algorithm/copy_backward.h> +#include <__algorithm/copy_n.h> +#include <__algorithm/fill_n.h> +#include <__algorithm/find_first_of.h> +#include <__algorithm/find_end.h> +#include <__algorithm/min.h> +#include <__functional/hash.h> // for __murmur2_or_cityhash +#include <__iterator/iterator_traits.h> +#include // for EOF +#include // for uint_least16_t +#include // for memcpy +#include // for wmemcpy +#include // for __libcpp_is_constant_evaluated #include <__debug> @@ -581,7 +547,7 @@ char_traits::find(const char_type* __s, size_t __n, const char_type& __ } -#ifndef _LIBCPP_NO_HAS_CHAR8_T +#ifndef _LIBCPP_HAS_NO_CHAR8_T template <> struct _LIBCPP_TEMPLATE_VIS char_traits @@ -688,7 +654,7 @@ char_traits::find(const char_type* __s, size_t __n, const char_type& __ return nullptr; } -#endif // #_LIBCPP_NO_HAS_CHAR8_T +#endif // #_LIBCPP_HAS_NO_CHAR8_T #ifndef _LIBCPP_HAS_NO_UNICODE_CHARS @@ -932,7 +898,7 @@ char_traits::assign(char_type* __s, size_t __n, char_type __a) _NOEXCE return __r; } -#endif // _LIBCPP_HAS_NO_UNICODE_CHARS +#endif // _LIBCPP_HAS_NO_UNICODE_CHARS // helper fns for basic_string and string_view @@ -953,7 +919,7 @@ __str_find(const _CharT *__p, _SizeT __sz, template inline _LIBCPP_CONSTEXPR_AFTER_CXX11 const _CharT * __search_substring(const _CharT *__first1, const _CharT *__last1, - const _CharT *__first2, const _CharT *__last2) { + const _CharT *__first2, const _CharT *__last2) _NOEXCEPT { // Take advantage of knowing source and pattern lengths. // Stop short when source is smaller than pattern. const ptrdiff_t __len2 = __last2 - __first2; @@ -1177,4 +1143,4 @@ _LIBCPP_END_NAMESPACE_STD _LIBCPP_POP_MACROS -#endif // _LIBCPP___STRING +#endif // _LIBCPP___STRING diff --git a/lib/libcxx/include/__support/ibm/gettod_zos.h b/lib/libcxx/include/__support/ibm/gettod_zos.h new file mode 100644 index 000000000..46e02a6c8 --- /dev/null +++ b/lib/libcxx/include/__support/ibm/gettod_zos.h @@ -0,0 +1,53 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP_SUPPORT_IBM_GETTOD_ZOS_H +#define _LIBCPP_SUPPORT_IBM_GETTOD_ZOS_H + +#include + +static inline int gettimeofdayMonotonic(struct timespec64* Output) { + + // The POSIX gettimeofday() function is not available on z/OS. Therefore, + // we will call stcke and other hardware instructions in implement equivalent. + // Note that nanoseconds alone will overflow when reaching new epoch in 2042. + + struct _t { + uint64_t Hi; + uint64_t Lo; + }; + struct _t Value = {0, 0}; + uint64_t CC = 0; + asm(" stcke %0\n" + " ipm %1\n" + " srlg %1,%1,28\n" + : "=m"(Value), "+r"(CC)::); + + if (CC != 0) { + errno = EMVSTODNOTSET; + return CC; + } + uint64_t us = (Value.Hi >> 4); + uint64_t ns = ((Value.Hi & 0x0F) << 8) + (Value.Lo >> 56); + ns = (ns * 1000) >> 12; + us = us - 2208988800000000; + + register uint64_t DivPair0 asm("r0"); // dividend (upper half), remainder + DivPair0 = 0; + register uint64_t DivPair1 asm("r1"); // dividend (lower half), quotient + DivPair1 = us; + uint64_t Divisor = 1000000; + asm(" dlgr %0,%2" : "+r"(DivPair0), "+r"(DivPair1) : "r"(Divisor) :); + + Output->tv_sec = DivPair1; + Output->tv_nsec = DivPair0 * 1000 + ns; + return 0; +} + +#endif // _LIBCPP_SUPPORT_IBM_GETTOD_ZOS_H diff --git a/lib/libcxx/include/__support/ibm/locale_mgmt_zos.h b/lib/libcxx/include/__support/ibm/locale_mgmt_zos.h new file mode 100644 index 000000000..90ad2c2c8 --- /dev/null +++ b/lib/libcxx/include/__support/ibm/locale_mgmt_zos.h @@ -0,0 +1,53 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP_SUPPORT_IBM_LOCALE_MGMT_ZOS_H +#define _LIBCPP_SUPPORT_IBM_LOCALE_MGMT_ZOS_H + +#if defined(__MVS__) +#include +#include + +#ifdef __cplusplus +extern "C" { +#endif + +#define _LC_MAX LC_MESSAGES /* highest real category */ +#define _NCAT (_LC_MAX + 1) /* maximum + 1 */ + +#define _CATMASK(n) (1 << (n)) +#define LC_COLLATE_MASK _CATMASK(LC_COLLATE) +#define LC_CTYPE_MASK _CATMASK(LC_CTYPE) +#define LC_MONETARY_MASK _CATMASK(LC_MONETARY) +#define LC_NUMERIC_MASK _CATMASK(LC_NUMERIC) +#define LC_TIME_MASK _CATMASK(LC_TIME) +#define LC_MESSAGES_MASK _CATMASK(LC_MESSAGES) +#define LC_ALL_MASK (_CATMASK(_NCAT) - 1) + +typedef struct locale_struct { + int category_mask; + std::string lc_collate; + std::string lc_ctype; + std::string lc_monetary; + std::string lc_numeric; + std::string lc_time; + std::string lc_messages; +} * locale_t; + +// z/OS does not have newlocale, freelocale and uselocale. +// The functions below are workarounds in single thread mode. +locale_t newlocale(int category_mask, const char* locale, locale_t base); +void freelocale(locale_t locobj); +locale_t uselocale(locale_t newloc); + +#ifdef __cplusplus +} +#endif +#endif // defined(__MVS__) +#endif // _LIBCPP_SUPPORT_IBM_LOCALE_MGMT_ZOS_H diff --git a/lib/libcxx/include/__support/ibm/nanosleep.h b/lib/libcxx/include/__support/ibm/nanosleep.h index c82f4eb0c..ec6379b8a 100644 --- a/lib/libcxx/include/__support/ibm/nanosleep.h +++ b/lib/libcxx/include/__support/ibm/nanosleep.h @@ -12,27 +12,45 @@ #include -inline int nanosleep(const struct timespec* req, struct timespec* rem) -{ - // The nanosleep() function is not available on z/OS. Therefore, we will call - // sleep() to sleep for whole seconds and usleep() to sleep for any remaining - // fraction of a second. Any remaining nanoseconds will round up to the next - // microsecond. - - useconds_t __micro_sec = (rem->tv_nsec + 999) / 1000; - if (__micro_sec > 999999) - { - ++rem->tv_sec; - __micro_sec -= 1000000; - } - while (rem->tv_sec) - rem->tv_sec = sleep(rem->tv_sec); - if (__micro_sec) { - rem->tv_nsec = __micro_sec * 1000; - return usleep(__micro_sec); - } - rem->tv_nsec = 0; - return 0; +inline int nanosleep(const struct timespec* __req, struct timespec* __rem) { + // The nanosleep() function is not available on z/OS. Therefore, we will call + // sleep() to sleep for whole seconds and usleep() to sleep for any remaining + // fraction of a second. Any remaining nanoseconds will round up to the next + // microsecond. + if (__req->tv_sec < 0 || __req->tv_nsec < 0 || __req->tv_nsec > 999999999) { + errno = EINVAL; + return -1; + } + useconds_t __micro_sec = + static_cast((__req->tv_nsec + 999) / 1000); + time_t __sec = __req->tv_sec; + if (__micro_sec > 999999) { + ++__sec; + __micro_sec -= 1000000; + } + __sec = sleep(static_cast(__sec)); + if (__sec) { + if (__rem) { + // Updating the remaining time to sleep in case of unsuccessful call to sleep(). + __rem->tv_sec = __sec; + __rem->tv_nsec = __micro_sec * 1000; + } + errno = EINTR; + return -1; + } + if (__micro_sec) { + int __rt = usleep(__micro_sec); + if (__rt != 0 && __rem) { + // The usleep() does not provide the amount of remaining time upon its failure, + // so the time slept will be ignored. + __rem->tv_sec = 0; + __rem->tv_nsec = __micro_sec * 1000; + // The errno is already set. + return -1; + } + return __rt; + } + return 0; } #endif // _LIBCPP_SUPPORT_IBM_NANOSLEEP_H diff --git a/lib/libcxx/include/__support/ibm/xlocale.h b/lib/libcxx/include/__support/ibm/xlocale.h index ad07a255f..58bdc67af 100644 --- a/lib/libcxx/include/__support/ibm/xlocale.h +++ b/lib/libcxx/include/__support/ibm/xlocale.h @@ -11,6 +11,8 @@ #define _LIBCPP_SUPPORT_IBM_XLOCALE_H #include <__support/ibm/locale_mgmt_aix.h> +#include <__support/ibm/locale_mgmt_zos.h> +#include #include "cstdlib" @@ -210,11 +212,13 @@ size_t wcsxfrm_l(wchar_t *__ws1, const wchar_t *__ws2, size_t __n, // strftime_l() is defined by POSIX. However, AIX 7.1 and z/OS do not have it // implemented yet. z/OS retrieves it from the POSIX fallbacks. +#if !defined(_AIX72) static inline size_t strftime_l(char *__s, size_t __size, const char *__fmt, const struct tm *__tm, locale_t locale) { return __xstrftime(locale, __s, __size, __fmt, __tm); } +#endif #elif defined(__MVS__) #include @@ -222,47 +226,101 @@ size_t strftime_l(char *__s, size_t __size, const char *__fmt, #include <__support/xlocale/__posix_l_fallback.h> #endif // defined(__MVS__) +namespace { + +struct __setAndRestore { + explicit __setAndRestore(locale_t locale) { + if (locale == (locale_t)0) { + __cloc = newlocale(LC_ALL_MASK, "C", /* base */ (locale_t)0); + __stored = uselocale(__cloc); + } else { + __stored = uselocale(locale); + } + } + + ~__setAndRestore() { + uselocale(__stored); + if (__cloc) + freelocale(__cloc); + } + +private: + locale_t __stored = (locale_t)0; + locale_t __cloc = (locale_t)0; +}; + +} // namespace + // The following are not POSIX routines. These are quick-and-dirty hacks // to make things pretend to work static inline long long strtoll_l(const char *__nptr, char **__endptr, int __base, locale_t locale) { + __setAndRestore __newloc(locale); return strtoll(__nptr, __endptr, __base); } + static inline long strtol_l(const char *__nptr, char **__endptr, int __base, locale_t locale) { + __setAndRestore __newloc(locale); return strtol(__nptr, __endptr, __base); } + +static inline +double strtod_l(const char *__nptr, char **__endptr, + locale_t locale) { + __setAndRestore __newloc(locale); + return strtod(__nptr, __endptr); +} + +static inline +float strtof_l(const char *__nptr, char **__endptr, + locale_t locale) { + __setAndRestore __newloc(locale); + return strtof(__nptr, __endptr); +} + static inline long double strtold_l(const char *__nptr, char **__endptr, locale_t locale) { + __setAndRestore __newloc(locale); return strtold(__nptr, __endptr); } + static inline unsigned long long strtoull_l(const char *__nptr, char **__endptr, int __base, locale_t locale) { + __setAndRestore __newloc(locale); return strtoull(__nptr, __endptr, __base); } + static inline unsigned long strtoul_l(const char *__nptr, char **__endptr, int __base, locale_t locale) { + __setAndRestore __newloc(locale); return strtoul(__nptr, __endptr, __base); } static inline -int vasprintf(char **strp, const char *fmt, va_list ap) -{ +int vasprintf(char **strp, const char *fmt, va_list ap) { const size_t buff_size = 256; - int str_size; - if ((*strp = (char *)malloc(buff_size)) == NULL) - { + if ((*strp = (char *)malloc(buff_size)) == NULL) { return -1; } - if ((str_size = vsnprintf(*strp, buff_size, fmt, ap)) >= buff_size) - { - if ((*strp = (char *)realloc(*strp, str_size + 1)) == NULL) - { + + va_list ap_copy; + // va_copy may not be provided by the C library in C++ 03 mode. +#if defined(_LIBCPP_CXX03_LANG) && __has_builtin(__builtin_va_copy) + __builtin_va_copy(ap_copy, ap); +#else + va_copy(ap_copy, ap); +#endif + int str_size = vsnprintf(*strp, buff_size, fmt, ap_copy); + va_end(ap_copy); + + if ((size_t) str_size >= buff_size) { + if ((*strp = (char *)realloc(*strp, str_size + 1)) == NULL) { return -1; } str_size = vsnprintf(*strp, str_size + 1, fmt, ap); diff --git a/lib/libcxx/include/__support/openbsd/xlocale.h b/lib/libcxx/include/__support/openbsd/xlocale.h index 1136fa327..49d66fde1 100644 --- a/lib/libcxx/include/__support/openbsd/xlocale.h +++ b/lib/libcxx/include/__support/openbsd/xlocale.h @@ -10,10 +10,10 @@ #ifndef _LIBCPP_SUPPORT_OPENBSD_XLOCALE_H #define _LIBCPP_SUPPORT_OPENBSD_XLOCALE_H -#include +#include <__support/xlocale/__strtonum_fallback.h> #include -#include +#include #include -#include <__support/xlocale/__strtonum_fallback.h> +#include #endif diff --git a/lib/libcxx/include/__support/win32/limits_msvc_win32.h b/lib/libcxx/include/__support/win32/limits_msvc_win32.h index 758d24647..c80df6693 100644 --- a/lib/libcxx/include/__support/win32/limits_msvc_win32.h +++ b/lib/libcxx/include/__support/win32/limits_msvc_win32.h @@ -17,8 +17,8 @@ #error "This header should only be included when using Microsofts C1XX frontend" #endif -#include // CHAR_BIT #include // limit constants +#include // CHAR_BIT #include // HUGE_VAL #include // internal MSVC header providing the needed functionality diff --git a/lib/libcxx/include/__support/win32/locale_win32.h b/lib/libcxx/include/__support/win32/locale_win32.h index d32a7a8ad..dd8a6001d 100644 --- a/lib/libcxx/include/__support/win32/locale_win32.h +++ b/lib/libcxx/include/__support/win32/locale_win32.h @@ -11,9 +11,28 @@ #define _LIBCPP_SUPPORT_WIN32_LOCALE_WIN32_H #include <__config> -#include -#include // _locale_t #include <__nullptr> +#include // _locale_t +#include + +#define _X_ALL LC_ALL +#define _X_COLLATE LC_COLLATE +#define _X_CTYPE LC_CTYPE +#define _X_MONETARY LC_MONETARY +#define _X_NUMERIC LC_NUMERIC +#define _X_TIME LC_TIME +#define _X_MAX LC_MAX +#define _X_MESSAGES 6 +#define _NCAT (_X_MESSAGES + 1) + +#define _CATMASK(n) ((1 << (n)) >> 1) +#define _M_COLLATE _CATMASK(_X_COLLATE) +#define _M_CTYPE _CATMASK(_X_CTYPE) +#define _M_MONETARY _CATMASK(_X_MONETARY) +#define _M_NUMERIC _CATMASK(_X_NUMERIC) +#define _M_TIME _CATMASK(_X_TIME) +#define _M_MESSAGES _CATMASK(_X_MESSAGES) +#define _M_ALL (_CATMASK(_NCAT) - 1) #define LC_COLLATE_MASK _M_COLLATE #define LC_CTYPE_MASK _M_CTYPE diff --git a/lib/libcxx/include/__threading_support b/lib/libcxx/include/__threading_support index de572f3ff..4d867167c 100644 --- a/lib/libcxx/include/__threading_support +++ b/lib/libcxx/include/__threading_support @@ -10,11 +10,12 @@ #ifndef _LIBCPP_THREADING_SUPPORT #define _LIBCPP_THREADING_SUPPORT -#include <__config> #include <__availability> +#include <__config> #include -#include #include +#include +#include #ifdef __MVS__ # include <__support/ibm/nanosleep.h> @@ -28,14 +29,15 @@ # include <__external_threading> #elif !defined(_LIBCPP_HAS_NO_THREADS) +#if defined(__APPLE__) || defined(__MVS__) +# define _LIBCPP_NO_NATIVE_SEMAPHORES +#endif + #if defined(_LIBCPP_HAS_THREAD_API_PTHREAD) # include # include -# if defined(__APPLE__) || defined(__MVS__) -# define _LIBCPP_NO_NATIVE_SEMAPHORES -# endif # ifndef _LIBCPP_NO_NATIVE_SEMAPHORES -# include +# include # endif #elif defined(_LIBCPP_HAS_THREAD_API_C11) # include @@ -149,6 +151,9 @@ typedef void* __libcpp_condvar_t; // Semaphore typedef void* __libcpp_semaphore_t; +#if defined(_LIBCPP_HAS_THREAD_API_WIN32) +# define _LIBCPP_SEMAPHORE_MAX (::std::numeric_limits::max()) +#endif // Execute Once typedef void* __libcpp_exec_once_flag; @@ -390,7 +395,7 @@ bool __libcpp_recursive_mutex_trylock(__libcpp_recursive_mutex_t *__m) return pthread_mutex_trylock(__m) == 0; } -int __libcpp_recursive_mutex_unlock(__libcpp_mutex_t *__m) +int __libcpp_recursive_mutex_unlock(__libcpp_recursive_mutex_t *__m) { return pthread_mutex_unlock(__m); } @@ -500,7 +505,7 @@ bool __libcpp_thread_id_less(__libcpp_thread_id t1, __libcpp_thread_id t2) // Thread bool __libcpp_thread_isnull(const __libcpp_thread_t *__t) { - return *__t == __libcpp_thread_t(); + return __libcpp_thread_get_id(__t) == 0; } int __libcpp_thread_create(__libcpp_thread_t *__t, void *(*__func)(void *), @@ -578,7 +583,7 @@ bool __libcpp_recursive_mutex_trylock(__libcpp_recursive_mutex_t *__m) return mtx_trylock(__m) == thrd_success; } -int __libcpp_recursive_mutex_unlock(__libcpp_mutex_t *__m) +int __libcpp_recursive_mutex_unlock(__libcpp_recursive_mutex_t *__m) { return mtx_unlock(__m) == thrd_success ? 0 : EINVAL; } diff --git a/lib/libcxx/include/__tree b/lib/libcxx/include/__tree index 0f6e4ec37..6113322ed 100644 --- a/lib/libcxx/include/__tree +++ b/lib/libcxx/include/__tree @@ -11,10 +11,12 @@ #define _LIBCPP___TREE #include <__config> +#include <__utility/forward.h> +#include #include +#include #include #include -#include #if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) #pragma GCC system_header @@ -714,7 +716,7 @@ public: }; template -class __tree_node_base +class _LIBCPP_STANDALONE_DEBUG __tree_node_base : public __tree_node_base_types<_VoidPtr>::__end_node_type { typedef __tree_node_base_types<_VoidPtr> _NodeBaseTypes; @@ -742,7 +744,7 @@ private: }; template -class __tree_node +class _LIBCPP_STANDALONE_DEBUG __tree_node : public __tree_node_base<_VoidPtr> { public: @@ -2410,7 +2412,7 @@ __tree<_Tp, _Compare, _Allocator>::__node_handle_merge_multi(_Tree& __source) } } -#endif // _LIBCPP_STD_VER > 14 +#endif // _LIBCPP_STD_VER > 14 template typename __tree<_Tp, _Compare, _Allocator>::iterator @@ -2743,4 +2745,4 @@ _LIBCPP_END_NAMESPACE_STD _LIBCPP_POP_MACROS -#endif // _LIBCPP___TREE +#endif // _LIBCPP___TREE diff --git a/lib/libcxx/include/__tuple b/lib/libcxx/include/__tuple index 4da9ec55f..082ec869e 100644 --- a/lib/libcxx/include/__tuple +++ b/lib/libcxx/include/__tuple @@ -134,7 +134,7 @@ template<> struct __parity<7> { template struct __pmake : __repeat @@ -548,4 +548,4 @@ struct __sfinae_assign_base { _LIBCPP_END_NAMESPACE_STD -#endif // _LIBCPP___TUPLE +#endif // _LIBCPP___TUPLE diff --git a/lib/libcxx/include/__utility/__decay_copy.h b/lib/libcxx/include/__utility/__decay_copy.h new file mode 100644 index 000000000..eda8db6a6 --- /dev/null +++ b/lib/libcxx/include/__utility/__decay_copy.h @@ -0,0 +1,39 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___TYPE_TRAITS_DECAY_COPY_H +#define _LIBCPP___TYPE_TRAITS_DECAY_COPY_H + +#include <__config> +#include <__utility/forward.h> +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +inline _LIBCPP_INLINE_VISIBILITY typename decay<_Tp>::type __decay_copy(_Tp&& __t) +#if _LIBCPP_STD_VER > 17 + noexcept(is_nothrow_convertible_v<_Tp, remove_reference_t<_Tp> >) +#endif +{ + return _VSTD::forward<_Tp>(__t); +} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___TYPE_TRAITS_DECAY_COPY_H diff --git a/lib/libcxx/include/__utility/as_const.h b/lib/libcxx/include/__utility/as_const.h new file mode 100644 index 000000000..2f23eb431 --- /dev/null +++ b/lib/libcxx/include/__utility/as_const.h @@ -0,0 +1,38 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___UTILITY_AS_CONST_H +#define _LIBCPP___UTILITY_AS_CONST_H + +#include <__config> +#include <__utility/forward.h> +#include <__utility/move.h> +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +#if _LIBCPP_STD_VER > 14 +template +_LIBCPP_NODISCARD_EXT constexpr add_const_t<_Tp>& as_const(_Tp& __t) noexcept { return __t; } + +template +void as_const(const _Tp&&) = delete; +#endif + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___UTILITY_AS_CONST_H diff --git a/lib/libcxx/include/__utility/cmp.h b/lib/libcxx/include/__utility/cmp.h new file mode 100644 index 000000000..a14e373c3 --- /dev/null +++ b/lib/libcxx/include/__utility/cmp.h @@ -0,0 +1,107 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___UTILITY_CMP_H +#define _LIBCPP___UTILITY_CMP_H + +#include <__config> +#include <__utility/forward.h> +#include <__utility/move.h> +#include +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +#if _LIBCPP_STD_VER > 17 && !defined(_LIBCPP_HAS_NO_CONCEPTS) +template +struct _IsSameAsAny : _Or<_IsSame<_Tp, _Up>...> {}; + +template +concept __is_safe_integral_cmp = is_integral_v<_Tp> && + !_IsSameAsAny<_Tp, bool, char, +#ifndef _LIBCPP_HAS_NO_CHAR8_T + char8_t, +#endif +#ifndef _LIBCPP_HAS_NO_UNICODE_CHARS + char16_t, char32_t, +#endif + wchar_t>::value; + +template<__is_safe_integral_cmp _Tp, __is_safe_integral_cmp _Up> +_LIBCPP_INLINE_VISIBILITY constexpr +bool cmp_equal(_Tp __t, _Up __u) noexcept +{ + if constexpr (is_signed_v<_Tp> == is_signed_v<_Up>) + return __t == __u; + else if constexpr (is_signed_v<_Tp>) + return __t < 0 ? false : make_unsigned_t<_Tp>(__t) == __u; + else + return __u < 0 ? false : __t == make_unsigned_t<_Up>(__u); +} + +template<__is_safe_integral_cmp _Tp, __is_safe_integral_cmp _Up> +_LIBCPP_INLINE_VISIBILITY constexpr +bool cmp_not_equal(_Tp __t, _Up __u) noexcept +{ + return !_VSTD::cmp_equal(__t, __u); +} + +template<__is_safe_integral_cmp _Tp, __is_safe_integral_cmp _Up> +_LIBCPP_INLINE_VISIBILITY constexpr +bool cmp_less(_Tp __t, _Up __u) noexcept +{ + if constexpr (is_signed_v<_Tp> == is_signed_v<_Up>) + return __t < __u; + else if constexpr (is_signed_v<_Tp>) + return __t < 0 ? true : make_unsigned_t<_Tp>(__t) < __u; + else + return __u < 0 ? false : __t < make_unsigned_t<_Up>(__u); +} + +template<__is_safe_integral_cmp _Tp, __is_safe_integral_cmp _Up> +_LIBCPP_INLINE_VISIBILITY constexpr +bool cmp_greater(_Tp __t, _Up __u) noexcept +{ + return _VSTD::cmp_less(__u, __t); +} + +template<__is_safe_integral_cmp _Tp, __is_safe_integral_cmp _Up> +_LIBCPP_INLINE_VISIBILITY constexpr +bool cmp_less_equal(_Tp __t, _Up __u) noexcept +{ + return !_VSTD::cmp_greater(__t, __u); +} + +template<__is_safe_integral_cmp _Tp, __is_safe_integral_cmp _Up> +_LIBCPP_INLINE_VISIBILITY constexpr +bool cmp_greater_equal(_Tp __t, _Up __u) noexcept +{ + return !_VSTD::cmp_less(__t, __u); +} + +template<__is_safe_integral_cmp _Tp, __is_safe_integral_cmp _Up> +_LIBCPP_INLINE_VISIBILITY constexpr +bool in_range(_Up __u) noexcept +{ + return _VSTD::cmp_less_equal(__u, numeric_limits<_Tp>::max()) && + _VSTD::cmp_greater_equal(__u, numeric_limits<_Tp>::min()); +} +#endif + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___UTILITY_CMP_H diff --git a/lib/libcxx/include/__utility/declval.h b/lib/libcxx/include/__utility/declval.h new file mode 100644 index 000000000..185527cc6 --- /dev/null +++ b/lib/libcxx/include/__utility/declval.h @@ -0,0 +1,39 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___UTILITY_DECLVAL_H +#define _LIBCPP___UTILITY_DECLVAL_H + +#include <__config> + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +// Suppress deprecation notice for volatile-qualified return type resulting +// from volatile-qualified types _Tp. +_LIBCPP_SUPPRESS_DEPRECATED_PUSH +template +_Tp&& __declval(int); +template +_Tp __declval(long); +_LIBCPP_SUPPRESS_DEPRECATED_POP + +template +decltype(__declval<_Tp>(0)) declval() _NOEXCEPT; + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___UTILITY_DECLVAL_H diff --git a/lib/libcxx/include/__utility/exchange.h b/lib/libcxx/include/__utility/exchange.h new file mode 100644 index 000000000..4d5211d94 --- /dev/null +++ b/lib/libcxx/include/__utility/exchange.h @@ -0,0 +1,40 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___UTILITY_EXCHANGE_H +#define _LIBCPP___UTILITY_EXCHANGE_H + +#include <__config> +#include <__utility/forward.h> +#include <__utility/move.h> + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +#if _LIBCPP_STD_VER > 11 +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +_T1 exchange(_T1& __obj, _T2 && __new_value) +{ + _T1 __old_value = _VSTD::move(__obj); + __obj = _VSTD::forward<_T2>(__new_value); + return __old_value; +} +#endif // _LIBCPP_STD_VER > 11 + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___UTILITY_EXCHANGE_H diff --git a/lib/libcxx/include/__utility/forward.h b/lib/libcxx/include/__utility/forward.h new file mode 100644 index 000000000..c994f1290 --- /dev/null +++ b/lib/libcxx/include/__utility/forward.h @@ -0,0 +1,42 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___UTILITY_FORWARD_H +#define _LIBCPP___UTILITY_FORWARD_H + +#include <__config> +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +_LIBCPP_NODISCARD_EXT inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR _Tp&& +forward(typename remove_reference<_Tp>::type& __t) _NOEXCEPT { + return static_cast<_Tp&&>(__t); +} + +template +_LIBCPP_NODISCARD_EXT inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR _Tp&& +forward(typename remove_reference<_Tp>::type&& __t) _NOEXCEPT { + static_assert(!is_lvalue_reference<_Tp>::value, "cannot forward an rvalue as an lvalue"); + return static_cast<_Tp&&>(__t); +} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___UTILITY_FORWARD_H diff --git a/lib/libcxx/include/__utility/in_place.h b/lib/libcxx/include/__utility/in_place.h new file mode 100644 index 000000000..964d08332 --- /dev/null +++ b/lib/libcxx/include/__utility/in_place.h @@ -0,0 +1,63 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___UTILITY_IN_PLACE_H +#define _LIBCPP___UTILITY_IN_PLACE_H + +#include <__config> +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +#if _LIBCPP_STD_VER > 14 + +struct _LIBCPP_TYPE_VIS in_place_t { + explicit in_place_t() = default; +}; +_LIBCPP_INLINE_VAR constexpr in_place_t in_place{}; + +template +struct _LIBCPP_TEMPLATE_VIS in_place_type_t { + explicit in_place_type_t() = default; +}; +template +_LIBCPP_INLINE_VAR constexpr in_place_type_t<_Tp> in_place_type{}; + +template +struct _LIBCPP_TEMPLATE_VIS in_place_index_t { + explicit in_place_index_t() = default; +}; +template +_LIBCPP_INLINE_VAR constexpr in_place_index_t<_Idx> in_place_index{}; + +template struct __is_inplace_type_imp : false_type {}; +template struct __is_inplace_type_imp> : true_type {}; + +template +using __is_inplace_type = __is_inplace_type_imp<__uncvref_t<_Tp>>; + +template struct __is_inplace_index_imp : false_type {}; +template struct __is_inplace_index_imp> : true_type {}; + +template +using __is_inplace_index = __is_inplace_index_imp<__uncvref_t<_Tp>>; + +#endif // _LIBCPP_STD_VER > 14 + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___UTILITY_IN_PLACE_H diff --git a/lib/libcxx/include/__utility/integer_sequence.h b/lib/libcxx/include/__utility/integer_sequence.h new file mode 100644 index 000000000..963c4a967 --- /dev/null +++ b/lib/libcxx/include/__utility/integer_sequence.h @@ -0,0 +1,83 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___UTILITY_INTEGER_SEQUENCE_H +#define _LIBCPP___UTILITY_INTEGER_SEQUENCE_H + +#include <__config> +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +#if _LIBCPP_STD_VER > 11 + +template +struct _LIBCPP_TEMPLATE_VIS integer_sequence +{ + typedef _Tp value_type; + static_assert( is_integral<_Tp>::value, + "std::integer_sequence can only be instantiated with an integral type" ); + static + _LIBCPP_INLINE_VISIBILITY + constexpr + size_t + size() noexcept { return sizeof...(_Ip); } +}; + +template + using index_sequence = integer_sequence; + +#if __has_builtin(__make_integer_seq) && !defined(_LIBCPP_TESTING_FALLBACK_MAKE_INTEGER_SEQUENCE) + +template +using __make_integer_sequence _LIBCPP_NODEBUG_TYPE = __make_integer_seq; + +#else + +template using __make_integer_sequence_unchecked _LIBCPP_NODEBUG_TYPE = + typename __detail::__make<_Np>::type::template __convert; + +template +struct __make_integer_sequence_checked +{ + static_assert(is_integral<_Tp>::value, + "std::make_integer_sequence can only be instantiated with an integral type" ); + static_assert(0 <= _Ep, "std::make_integer_sequence must have a non-negative sequence length"); + // Workaround GCC bug by preventing bad installations when 0 <= _Ep + // https://gcc.gnu.org/bugzilla/show_bug.cgi?id=68929 + typedef _LIBCPP_NODEBUG_TYPE __make_integer_sequence_unchecked<_Tp, 0 <= _Ep ? _Ep : 0> type; +}; + +template +using __make_integer_sequence _LIBCPP_NODEBUG_TYPE = typename __make_integer_sequence_checked<_Tp, _Ep>::type; + +#endif + +template + using make_integer_sequence = __make_integer_sequence<_Tp, _Np>; + +template + using make_index_sequence = make_integer_sequence; + +template + using index_sequence_for = make_index_sequence; + +#endif // _LIBCPP_STD_VER > 11 + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___UTILITY_INTEGER_SEQUENCE_H diff --git a/lib/libcxx/include/__utility/move.h b/lib/libcxx/include/__utility/move.h new file mode 100644 index 000000000..d3c56f93e --- /dev/null +++ b/lib/libcxx/include/__utility/move.h @@ -0,0 +1,52 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___UTILITY_MOVE_H +#define _LIBCPP___UTILITY_MOVE_H + +#include <__config> +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +template +_LIBCPP_NODISCARD_EXT inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR typename remove_reference<_Tp>::type&& +move(_Tp&& __t) _NOEXCEPT { + typedef _LIBCPP_NODEBUG_TYPE typename remove_reference<_Tp>::type _Up; + return static_cast<_Up&&>(__t); +} + +#ifndef _LIBCPP_CXX03_LANG +template +using __move_if_noexcept_result_t = + typename conditional::value && is_copy_constructible<_Tp>::value, const _Tp&, + _Tp&&>::type; +#else // _LIBCPP_CXX03_LANG +template +using __move_if_noexcept_result_t = const _Tp&; +#endif + +template +_LIBCPP_NODISCARD_EXT inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 __move_if_noexcept_result_t<_Tp> +move_if_noexcept(_Tp& __x) _NOEXCEPT { + return _VSTD::move(__x); +} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___UTILITY_MOVE_H diff --git a/lib/libcxx/include/__utility/pair.h b/lib/libcxx/include/__utility/pair.h new file mode 100644 index 000000000..e0216f340 --- /dev/null +++ b/lib/libcxx/include/__utility/pair.h @@ -0,0 +1,585 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___UTILITY_PAIR_H +#define _LIBCPP___UTILITY_PAIR_H + +#include <__config> +#include <__functional/unwrap_ref.h> +#include <__tuple> +#include <__utility/forward.h> +#include <__utility/move.h> +#include <__utility/piecewise_construct.h> +#include +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + + +#if defined(_LIBCPP_DEPRECATED_ABI_DISABLE_PAIR_TRIVIAL_COPY_CTOR) +template +struct __non_trivially_copyable_base { + _LIBCPP_CONSTEXPR _LIBCPP_INLINE_VISIBILITY + __non_trivially_copyable_base() _NOEXCEPT {} + _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY + __non_trivially_copyable_base(__non_trivially_copyable_base const&) _NOEXCEPT {} +}; +#endif + +template +struct _LIBCPP_TEMPLATE_VIS pair +#if defined(_LIBCPP_DEPRECATED_ABI_DISABLE_PAIR_TRIVIAL_COPY_CTOR) +: private __non_trivially_copyable_base<_T1, _T2> +#endif +{ + typedef _T1 first_type; + typedef _T2 second_type; + + _T1 first; + _T2 second; + +#if !defined(_LIBCPP_CXX03_LANG) + pair(pair const&) = default; + pair(pair&&) = default; +#else + // Use the implicitly declared copy constructor in C++03 +#endif + +#ifdef _LIBCPP_CXX03_LANG + _LIBCPP_INLINE_VISIBILITY + pair() : first(), second() {} + + _LIBCPP_INLINE_VISIBILITY + pair(_T1 const& __t1, _T2 const& __t2) : first(__t1), second(__t2) {} + + template + _LIBCPP_INLINE_VISIBILITY + pair(const pair<_U1, _U2>& __p) : first(__p.first), second(__p.second) {} + + _LIBCPP_INLINE_VISIBILITY + pair& operator=(pair const& __p) { + first = __p.first; + second = __p.second; + return *this; + } +#else + template + using _EnableB _LIBCPP_NODEBUG_TYPE = typename enable_if<_Val, bool>::type; + + struct _CheckArgs { + template + static constexpr bool __enable_explicit_default() { + return is_default_constructible<_T1>::value + && is_default_constructible<_T2>::value + && !__enable_implicit_default<>(); + } + + template + static constexpr bool __enable_implicit_default() { + return __is_implicitly_default_constructible<_T1>::value + && __is_implicitly_default_constructible<_T2>::value; + } + + template + static constexpr bool __enable_explicit() { + return is_constructible::value + && is_constructible::value + && (!is_convertible<_U1, first_type>::value + || !is_convertible<_U2, second_type>::value); + } + + template + static constexpr bool __enable_implicit() { + return is_constructible::value + && is_constructible::value + && is_convertible<_U1, first_type>::value + && is_convertible<_U2, second_type>::value; + } + }; + + template + using _CheckArgsDep _LIBCPP_NODEBUG_TYPE = typename conditional< + _MaybeEnable, _CheckArgs, __check_tuple_constructor_fail>::type; + + struct _CheckTupleLikeConstructor { + template + static constexpr bool __enable_implicit() { + return __tuple_convertible<_Tuple, pair>::value; + } + + template + static constexpr bool __enable_explicit() { + return __tuple_constructible<_Tuple, pair>::value + && !__tuple_convertible<_Tuple, pair>::value; + } + + template + static constexpr bool __enable_assign() { + return __tuple_assignable<_Tuple, pair>::value; + } + }; + + template + using _CheckTLC _LIBCPP_NODEBUG_TYPE = typename conditional< + __tuple_like_with_size<_Tuple, 2>::value + && !is_same::type, pair>::value, + _CheckTupleLikeConstructor, + __check_tuple_constructor_fail + >::type; + + template::__enable_explicit_default() + > = false> + explicit _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR + pair() _NOEXCEPT_(is_nothrow_default_constructible::value && + is_nothrow_default_constructible::value) + : first(), second() {} + + template::__enable_implicit_default() + > = false> + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR + pair() _NOEXCEPT_(is_nothrow_default_constructible::value && + is_nothrow_default_constructible::value) + : first(), second() {} + + template ::template __enable_explicit<_T1 const&, _T2 const&>() + > = false> + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 + explicit pair(_T1 const& __t1, _T2 const& __t2) + _NOEXCEPT_(is_nothrow_copy_constructible::value && + is_nothrow_copy_constructible::value) + : first(__t1), second(__t2) {} + + template::template __enable_implicit<_T1 const&, _T2 const&>() + > = false> + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 + pair(_T1 const& __t1, _T2 const& __t2) + _NOEXCEPT_(is_nothrow_copy_constructible::value && + is_nothrow_copy_constructible::value) + : first(__t1), second(__t2) {} + + template() + > = false> + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 + explicit pair(_U1&& __u1, _U2&& __u2) + _NOEXCEPT_((is_nothrow_constructible::value && + is_nothrow_constructible::value)) + : first(_VSTD::forward<_U1>(__u1)), second(_VSTD::forward<_U2>(__u2)) {} + + template() + > = false> + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 + pair(_U1&& __u1, _U2&& __u2) + _NOEXCEPT_((is_nothrow_constructible::value && + is_nothrow_constructible::value)) + : first(_VSTD::forward<_U1>(__u1)), second(_VSTD::forward<_U2>(__u2)) {} + + template() + > = false> + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 + explicit pair(pair<_U1, _U2> const& __p) + _NOEXCEPT_((is_nothrow_constructible::value && + is_nothrow_constructible::value)) + : first(__p.first), second(__p.second) {} + + template() + > = false> + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 + pair(pair<_U1, _U2> const& __p) + _NOEXCEPT_((is_nothrow_constructible::value && + is_nothrow_constructible::value)) + : first(__p.first), second(__p.second) {} + + template() + > = false> + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 + explicit pair(pair<_U1, _U2>&&__p) + _NOEXCEPT_((is_nothrow_constructible::value && + is_nothrow_constructible::value)) + : first(_VSTD::forward<_U1>(__p.first)), second(_VSTD::forward<_U2>(__p.second)) {} + + template() + > = false> + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 + pair(pair<_U1, _U2>&& __p) + _NOEXCEPT_((is_nothrow_constructible::value && + is_nothrow_constructible::value)) + : first(_VSTD::forward<_U1>(__p.first)), second(_VSTD::forward<_U2>(__p.second)) {} + + template::template __enable_explicit<_Tuple>() + > = false> + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 + explicit pair(_Tuple&& __p) + : first(_VSTD::get<0>(_VSTD::forward<_Tuple>(__p))), + second(_VSTD::get<1>(_VSTD::forward<_Tuple>(__p))) {} + + template::template __enable_implicit<_Tuple>() + > = false> + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 + pair(_Tuple&& __p) + : first(_VSTD::get<0>(_VSTD::forward<_Tuple>(__p))), + second(_VSTD::get<1>(_VSTD::forward<_Tuple>(__p))) {} + + template + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 + pair(piecewise_construct_t __pc, + tuple<_Args1...> __first_args, tuple<_Args2...> __second_args) + _NOEXCEPT_((is_nothrow_constructible::value && + is_nothrow_constructible::value)) + : pair(__pc, __first_args, __second_args, + typename __make_tuple_indices::type(), + typename __make_tuple_indices::type()) {} + + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 + pair& operator=(typename conditional< + is_copy_assignable::value && + is_copy_assignable::value, + pair, __nat>::type const& __p) + _NOEXCEPT_(is_nothrow_copy_assignable::value && + is_nothrow_copy_assignable::value) + { + first = __p.first; + second = __p.second; + return *this; + } + + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 + pair& operator=(typename conditional< + is_move_assignable::value && + is_move_assignable::value, + pair, __nat>::type&& __p) + _NOEXCEPT_(is_nothrow_move_assignable::value && + is_nothrow_move_assignable::value) + { + first = _VSTD::forward(__p.first); + second = _VSTD::forward(__p.second); + return *this; + } + + template ::template __enable_assign<_Tuple>() + > = false> + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 + pair& operator=(_Tuple&& __p) { + first = _VSTD::get<0>(_VSTD::forward<_Tuple>(__p)); + second = _VSTD::get<1>(_VSTD::forward<_Tuple>(__p)); + return *this; + } +#endif + + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 + void + swap(pair& __p) _NOEXCEPT_(__is_nothrow_swappable::value && + __is_nothrow_swappable::value) + { + using _VSTD::swap; + swap(first, __p.first); + swap(second, __p.second); + } +private: + +#ifndef _LIBCPP_CXX03_LANG + template + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 + pair(piecewise_construct_t, + tuple<_Args1...>& __first_args, tuple<_Args2...>& __second_args, + __tuple_indices<_I1...>, __tuple_indices<_I2...>); +#endif +}; + +#ifndef _LIBCPP_HAS_NO_DEDUCTION_GUIDES +template +pair(_T1, _T2) -> pair<_T1, _T2>; +#endif // _LIBCPP_HAS_NO_DEDUCTION_GUIDES + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 +bool +operator==(const pair<_T1,_T2>& __x, const pair<_T1,_T2>& __y) +{ + return __x.first == __y.first && __x.second == __y.second; +} + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 +bool +operator!=(const pair<_T1,_T2>& __x, const pair<_T1,_T2>& __y) +{ + return !(__x == __y); +} + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 +bool +operator< (const pair<_T1,_T2>& __x, const pair<_T1,_T2>& __y) +{ + return __x.first < __y.first || (!(__y.first < __x.first) && __x.second < __y.second); +} + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 +bool +operator> (const pair<_T1,_T2>& __x, const pair<_T1,_T2>& __y) +{ + return __y < __x; +} + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 +bool +operator>=(const pair<_T1,_T2>& __x, const pair<_T1,_T2>& __y) +{ + return !(__x < __y); +} + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 +bool +operator<=(const pair<_T1,_T2>& __x, const pair<_T1,_T2>& __y) +{ + return !(__y < __x); +} + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 +typename enable_if +< + __is_swappable<_T1>::value && + __is_swappable<_T2>::value, + void +>::type +swap(pair<_T1, _T2>& __x, pair<_T1, _T2>& __y) + _NOEXCEPT_((__is_nothrow_swappable<_T1>::value && + __is_nothrow_swappable<_T2>::value)) +{ + __x.swap(__y); +} + +#ifndef _LIBCPP_CXX03_LANG + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 +pair::type, typename __unwrap_ref_decay<_T2>::type> +make_pair(_T1&& __t1, _T2&& __t2) +{ + return pair::type, typename __unwrap_ref_decay<_T2>::type> + (_VSTD::forward<_T1>(__t1), _VSTD::forward<_T2>(__t2)); +} + +#else // _LIBCPP_CXX03_LANG + +template +inline _LIBCPP_INLINE_VISIBILITY +pair<_T1,_T2> +make_pair(_T1 __x, _T2 __y) +{ + return pair<_T1, _T2>(__x, __y); +} + +#endif // _LIBCPP_CXX03_LANG + +template + struct _LIBCPP_TEMPLATE_VIS tuple_size > + : public integral_constant {}; + +template +struct _LIBCPP_TEMPLATE_VIS tuple_element<_Ip, pair<_T1, _T2> > +{ + static_assert(_Ip < 2, "Index out of bounds in std::tuple_element>"); +}; + +template +struct _LIBCPP_TEMPLATE_VIS tuple_element<0, pair<_T1, _T2> > +{ + typedef _LIBCPP_NODEBUG_TYPE _T1 type; +}; + +template +struct _LIBCPP_TEMPLATE_VIS tuple_element<1, pair<_T1, _T2> > +{ + typedef _LIBCPP_NODEBUG_TYPE _T2 type; +}; + +template struct __get_pair; + +template <> +struct __get_pair<0> +{ + template + static + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 + _T1& + get(pair<_T1, _T2>& __p) _NOEXCEPT {return __p.first;} + + template + static + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 + const _T1& + get(const pair<_T1, _T2>& __p) _NOEXCEPT {return __p.first;} + +#ifndef _LIBCPP_CXX03_LANG + template + static + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 + _T1&& + get(pair<_T1, _T2>&& __p) _NOEXCEPT {return _VSTD::forward<_T1>(__p.first);} + + template + static + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 + const _T1&& + get(const pair<_T1, _T2>&& __p) _NOEXCEPT {return _VSTD::forward(__p.first);} +#endif // _LIBCPP_CXX03_LANG +}; + +template <> +struct __get_pair<1> +{ + template + static + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 + _T2& + get(pair<_T1, _T2>& __p) _NOEXCEPT {return __p.second;} + + template + static + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 + const _T2& + get(const pair<_T1, _T2>& __p) _NOEXCEPT {return __p.second;} + +#ifndef _LIBCPP_CXX03_LANG + template + static + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 + _T2&& + get(pair<_T1, _T2>&& __p) _NOEXCEPT {return _VSTD::forward<_T2>(__p.second);} + + template + static + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 + const _T2&& + get(const pair<_T1, _T2>&& __p) _NOEXCEPT {return _VSTD::forward(__p.second);} +#endif // _LIBCPP_CXX03_LANG +}; + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 +typename tuple_element<_Ip, pair<_T1, _T2> >::type& +get(pair<_T1, _T2>& __p) _NOEXCEPT +{ + return __get_pair<_Ip>::get(__p); +} + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 +const typename tuple_element<_Ip, pair<_T1, _T2> >::type& +get(const pair<_T1, _T2>& __p) _NOEXCEPT +{ + return __get_pair<_Ip>::get(__p); +} + +#ifndef _LIBCPP_CXX03_LANG +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 +typename tuple_element<_Ip, pair<_T1, _T2> >::type&& +get(pair<_T1, _T2>&& __p) _NOEXCEPT +{ + return __get_pair<_Ip>::get(_VSTD::move(__p)); +} + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 +const typename tuple_element<_Ip, pair<_T1, _T2> >::type&& +get(const pair<_T1, _T2>&& __p) _NOEXCEPT +{ + return __get_pair<_Ip>::get(_VSTD::move(__p)); +} +#endif // _LIBCPP_CXX03_LANG + +#if _LIBCPP_STD_VER > 11 +template +inline _LIBCPP_INLINE_VISIBILITY +constexpr _T1 & get(pair<_T1, _T2>& __p) _NOEXCEPT +{ + return __get_pair<0>::get(__p); +} + +template +inline _LIBCPP_INLINE_VISIBILITY +constexpr _T1 const & get(pair<_T1, _T2> const& __p) _NOEXCEPT +{ + return __get_pair<0>::get(__p); +} + +template +inline _LIBCPP_INLINE_VISIBILITY +constexpr _T1 && get(pair<_T1, _T2>&& __p) _NOEXCEPT +{ + return __get_pair<0>::get(_VSTD::move(__p)); +} + +template +inline _LIBCPP_INLINE_VISIBILITY +constexpr _T1 const && get(pair<_T1, _T2> const&& __p) _NOEXCEPT +{ + return __get_pair<0>::get(_VSTD::move(__p)); +} + +template +inline _LIBCPP_INLINE_VISIBILITY +constexpr _T1 & get(pair<_T2, _T1>& __p) _NOEXCEPT +{ + return __get_pair<1>::get(__p); +} + +template +inline _LIBCPP_INLINE_VISIBILITY +constexpr _T1 const & get(pair<_T2, _T1> const& __p) _NOEXCEPT +{ + return __get_pair<1>::get(__p); +} + +template +inline _LIBCPP_INLINE_VISIBILITY +constexpr _T1 && get(pair<_T2, _T1>&& __p) _NOEXCEPT +{ + return __get_pair<1>::get(_VSTD::move(__p)); +} + +template +inline _LIBCPP_INLINE_VISIBILITY +constexpr _T1 const && get(pair<_T2, _T1> const&& __p) _NOEXCEPT +{ + return __get_pair<1>::get(_VSTD::move(__p)); +} + +#endif + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___UTILITY_PAIR_H diff --git a/lib/libcxx/include/__utility/piecewise_construct.h b/lib/libcxx/include/__utility/piecewise_construct.h new file mode 100644 index 000000000..8bef01c9f --- /dev/null +++ b/lib/libcxx/include/__utility/piecewise_construct.h @@ -0,0 +1,34 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___UTILITY_PIECEWISE_CONSTRUCT_H +#define _LIBCPP___UTILITY_PIECEWISE_CONSTRUCT_H + +#include <__config> + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +struct _LIBCPP_TEMPLATE_VIS piecewise_construct_t { explicit piecewise_construct_t() = default; }; +#if defined(_LIBCPP_CXX03_LANG) || defined(_LIBCPP_BUILDING_LIBRARY) +extern _LIBCPP_EXPORTED_FROM_ABI const piecewise_construct_t piecewise_construct;// = piecewise_construct_t(); +#else +/* _LIBCPP_INLINE_VAR */ constexpr piecewise_construct_t piecewise_construct = piecewise_construct_t(); +#endif + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___UTILITY_PIECEWISE_CONSTRUCT_H diff --git a/lib/libcxx/include/__utility/rel_ops.h b/lib/libcxx/include/__utility/rel_ops.h new file mode 100644 index 000000000..b900da80f --- /dev/null +++ b/lib/libcxx/include/__utility/rel_ops.h @@ -0,0 +1,67 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___UTILITY_REL_OPS_H +#define _LIBCPP___UTILITY_REL_OPS_H + +#include <__config> +#include <__utility/forward.h> +#include <__utility/move.h> +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +namespace rel_ops +{ + +template +inline _LIBCPP_INLINE_VISIBILITY +bool +operator!=(const _Tp& __x, const _Tp& __y) +{ + return !(__x == __y); +} + +template +inline _LIBCPP_INLINE_VISIBILITY +bool +operator> (const _Tp& __x, const _Tp& __y) +{ + return __y < __x; +} + +template +inline _LIBCPP_INLINE_VISIBILITY +bool +operator<=(const _Tp& __x, const _Tp& __y) +{ + return !(__y < __x); +} + +template +inline _LIBCPP_INLINE_VISIBILITY +bool +operator>=(const _Tp& __x, const _Tp& __y) +{ + return !(__x < __y); +} + +} // rel_ops + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___UTILITY_REL_OPS_H diff --git a/lib/libcxx/include/__utility/swap.h b/lib/libcxx/include/__utility/swap.h new file mode 100644 index 000000000..8af83a9a0 --- /dev/null +++ b/lib/libcxx/include/__utility/swap.h @@ -0,0 +1,55 @@ +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___UTILITY_SWAP_H +#define _LIBCPP___UTILITY_SWAP_H + +#include <__config> +#include <__utility/declval.h> +#include <__utility/move.h> +#include +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +#ifndef _LIBCPP_CXX03_LANG +template +using __swap_result_t = typename enable_if::value && is_move_assignable<_Tp>::value>::type; +#else +template +using __swap_result_t = void; +#endif + +template +inline _LIBCPP_INLINE_VISIBILITY __swap_result_t<_Tp> _LIBCPP_CONSTEXPR_AFTER_CXX17 swap(_Tp& __x, _Tp& __y) + _NOEXCEPT_(is_nothrow_move_constructible<_Tp>::value&& is_nothrow_move_assignable<_Tp>::value) { + _Tp __t(_VSTD::move(__x)); + __x = _VSTD::move(__y); + __y = _VSTD::move(__t); +} + +template +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 typename enable_if<__is_swappable<_Tp>::value>::type +swap(_Tp (&__a)[_Np], _Tp (&__b)[_Np]) _NOEXCEPT_(__is_nothrow_swappable<_Tp>::value) { + for (size_t __i = 0; __i != _Np; ++__i) { + swap(__a[__i], __b[__i]); + } +} + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___UTILITY_SWAP_H diff --git a/lib/libcxx/include/__utility/to_underlying.h b/lib/libcxx/include/__utility/to_underlying.h new file mode 100644 index 000000000..fd22f89ff --- /dev/null +++ b/lib/libcxx/include/__utility/to_underlying.h @@ -0,0 +1,45 @@ +// -*- C++ -*- +//===----------------- __utility/to_underlying.h --------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___UTILITY_TO_UNDERLYING_H +#define _LIBCPP___UTILITY_TO_UNDERLYING_H + +#include <__config> +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +#ifndef _LIBCPP_CXX03_LANG +template +_LIBCPP_INLINE_VISIBILITY constexpr typename underlying_type<_Tp>::type +__to_underlying(_Tp __val) noexcept { + return static_cast::type>(__val); +} +#endif // !_LIBCPP_CXX03_LANG + +#if _LIBCPP_STD_VER > 20 +template +_LIBCPP_NODISCARD_EXT _LIBCPP_INLINE_VISIBILITY constexpr underlying_type_t<_Tp> +to_underlying(_Tp __val) noexcept { + return _VSTD::__to_underlying(__val); +} +#endif + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___UTILITY_TO_UNDERLYING_H diff --git a/lib/libcxx/include/__variant/monostate.h b/lib/libcxx/include/__variant/monostate.h new file mode 100644 index 000000000..36e3eead1 --- /dev/null +++ b/lib/libcxx/include/__variant/monostate.h @@ -0,0 +1,65 @@ +// -*- C++ -*- +//===----------------------------------------------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP___VARIANT_MONOSTATE_H +#define _LIBCPP___VARIANT_MONOSTATE_H + +#include <__config> +#include <__functional/hash.h> +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +#if _LIBCPP_STD_VER > 14 + +struct _LIBCPP_TEMPLATE_VIS monostate {}; + +inline _LIBCPP_INLINE_VISIBILITY +constexpr bool operator<(monostate, monostate) noexcept { return false; } + +inline _LIBCPP_INLINE_VISIBILITY +constexpr bool operator>(monostate, monostate) noexcept { return false; } + +inline _LIBCPP_INLINE_VISIBILITY +constexpr bool operator<=(monostate, monostate) noexcept { return true; } + +inline _LIBCPP_INLINE_VISIBILITY +constexpr bool operator>=(monostate, monostate) noexcept { return true; } + +inline _LIBCPP_INLINE_VISIBILITY +constexpr bool operator==(monostate, monostate) noexcept { return true; } + +inline _LIBCPP_INLINE_VISIBILITY +constexpr bool operator!=(monostate, monostate) noexcept { return false; } + +template <> +struct _LIBCPP_TEMPLATE_VIS hash { + using argument_type = monostate; + using result_type = size_t; + + inline _LIBCPP_INLINE_VISIBILITY + result_type operator()(const argument_type&) const _NOEXCEPT { + return 66740831; // return a fundamentally attractive random value. + } +}; + +#endif // _LIBCPP_STD_VER > 14 + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // _LIBCPP___VARIANT_MONOSTATE_H diff --git a/lib/libcxx/include/algorithm b/lib/libcxx/include/algorithm index f7fb2013a..849302a7d 100644 --- a/lib/libcxx/include/algorithm +++ b/lib/libcxx/include/algorithm @@ -351,11 +351,11 @@ template is_sorted_until(ForwardIterator first, ForwardIterator last, Compare comp); template - void + constexpr void // constexpr in C++20 sort(RandomAccessIterator first, RandomAccessIterator last); template - void + constexpr void // constexpr in C++20 sort(RandomAccessIterator first, RandomAccessIterator last, Compare comp); template @@ -367,29 +367,29 @@ template stable_sort(RandomAccessIterator first, RandomAccessIterator last, Compare comp); template - void + constexpr void // constexpr in C++20 partial_sort(RandomAccessIterator first, RandomAccessIterator middle, RandomAccessIterator last); template - void + constexpr void // constexpr in C++20 partial_sort(RandomAccessIterator first, RandomAccessIterator middle, RandomAccessIterator last, Compare comp); template - RandomAccessIterator + constexpr RandomAccessIterator // constexpr in C++20 partial_sort_copy(InputIterator first, InputIterator last, RandomAccessIterator result_first, RandomAccessIterator result_last); template - RandomAccessIterator + constexpr RandomAccessIterator // constexpr in C++20 partial_sort_copy(InputIterator first, InputIterator last, RandomAccessIterator result_first, RandomAccessIterator result_last, Compare comp); template - void + constexpr void // constexpr in C++20 nth_element(RandomAccessIterator first, RandomAccessIterator nth, RandomAccessIterator last); template - void + constexpr void // constexpr in C++20 nth_element(RandomAccessIterator first, RandomAccessIterator nth, RandomAccessIterator last, Compare comp); template @@ -491,35 +491,35 @@ template - void + constexpr void // constexpr in C++20 push_heap(RandomAccessIterator first, RandomAccessIterator last); template - void + constexpr void // constexpr in C++20 push_heap(RandomAccessIterator first, RandomAccessIterator last, Compare comp); template - void + constexpr void // constexpr in C++20 pop_heap(RandomAccessIterator first, RandomAccessIterator last); template - void + constexpr void // constexpr in C++20 pop_heap(RandomAccessIterator first, RandomAccessIterator last, Compare comp); template - void + constexpr void // constexpr in C++20 make_heap(RandomAccessIterator first, RandomAccessIterator last); template - void + constexpr void // constexpr in C++20 make_heap(RandomAccessIterator first, RandomAccessIterator last, Compare comp); template - void + constexpr void // constexpr in C++20 sort_heap(RandomAccessIterator first, RandomAccessIterator last); template - void + constexpr void // constexpr in C++20 sort_heap(RandomAccessIterator first, RandomAccessIterator last, Compare comp); template @@ -646,18 +646,113 @@ template */ #include <__config> -#include -#include +#include <__debug> +#include <__bits> // __libcpp_clz +#include #include +#include +#include #include // needed to provide swap_ranges. #include -#include #include -#include -#include +#include +#include +#include // swap_ranges #include -#include <__debug> +#include <__algorithm/adjacent_find.h> +#include <__algorithm/all_of.h> +#include <__algorithm/any_of.h> +#include <__algorithm/binary_search.h> +#include <__algorithm/clamp.h> +#include <__algorithm/comp.h> +#include <__algorithm/comp_ref_type.h> +#include <__algorithm/copy.h> +#include <__algorithm/copy_backward.h> +#include <__algorithm/copy_if.h> +#include <__algorithm/copy_n.h> +#include <__algorithm/count.h> +#include <__algorithm/count_if.h> +#include <__algorithm/equal.h> +#include <__algorithm/equal_range.h> +#include <__algorithm/fill_n.h> +#include <__algorithm/fill.h> +#include <__algorithm/find.h> +#include <__algorithm/find_end.h> +#include <__algorithm/find_first_of.h> +#include <__algorithm/find_if.h> +#include <__algorithm/find_if_not.h> +#include <__algorithm/for_each.h> +#include <__algorithm/for_each_n.h> +#include <__algorithm/generate_n.h> +#include <__algorithm/generate.h> +#include <__algorithm/half_positive.h> +#include <__algorithm/includes.h> +#include <__algorithm/inplace_merge.h> +#include <__algorithm/is_heap.h> +#include <__algorithm/is_heap_until.h> +#include <__algorithm/is_partitioned.h> +#include <__algorithm/is_permutation.h> +#include <__algorithm/is_sorted.h> +#include <__algorithm/is_sorted_until.h> +#include <__algorithm/iter_swap.h> +#include <__algorithm/lexicographical_compare.h> +#include <__algorithm/lower_bound.h> +#include <__algorithm/make_heap.h> +#include <__algorithm/max.h> +#include <__algorithm/max_element.h> +#include <__algorithm/merge.h> +#include <__algorithm/min.h> +#include <__algorithm/min_element.h> +#include <__algorithm/minmax.h> +#include <__algorithm/minmax_element.h> +#include <__algorithm/mismatch.h> +#include <__algorithm/move.h> +#include <__algorithm/move_backward.h> +#include <__algorithm/next_permutation.h> +#include <__algorithm/none_of.h> +#include <__algorithm/nth_element.h> +#include <__algorithm/partial_sort.h> +#include <__algorithm/partial_sort_copy.h> +#include <__algorithm/partition.h> +#include <__algorithm/partition_copy.h> +#include <__algorithm/partition_point.h> +#include <__algorithm/pop_heap.h> +#include <__algorithm/prev_permutation.h> +#include <__algorithm/push_heap.h> +#include <__algorithm/remove.h> +#include <__algorithm/remove_copy.h> +#include <__algorithm/remove_copy_if.h> +#include <__algorithm/remove_if.h> +#include <__algorithm/replace.h> +#include <__algorithm/replace_copy.h> +#include <__algorithm/replace_copy_if.h> +#include <__algorithm/replace_if.h> +#include <__algorithm/reverse.h> +#include <__algorithm/reverse_copy.h> +#include <__algorithm/rotate.h> +#include <__algorithm/rotate_copy.h> +#include <__algorithm/sample.h> +#include <__algorithm/search.h> +#include <__algorithm/search_n.h> +#include <__algorithm/set_difference.h> +#include <__algorithm/set_intersection.h> +#include <__algorithm/set_symmetric_difference.h> +#include <__algorithm/set_union.h> +#include <__algorithm/shift_left.h> +#include <__algorithm/shift_right.h> +#include <__algorithm/shuffle.h> +#include <__algorithm/sift_down.h> +#include <__algorithm/sort.h> +#include <__algorithm/sort_heap.h> +#include <__algorithm/stable_partition.h> +#include <__algorithm/stable_sort.h> +#include <__algorithm/swap_ranges.h> +#include <__algorithm/transform.h> +#include <__algorithm/unique_copy.h> +#include <__algorithm/unique.h> +#include <__algorithm/unwrap_iter.h> +#include <__algorithm/upper_bound.h> #if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) #pragma GCC system_header @@ -666,5197 +761,10 @@ template _LIBCPP_PUSH_MACROS #include <__undef_macros> - -_LIBCPP_BEGIN_NAMESPACE_STD - -// I'd like to replace these with _VSTD::equal_to, but can't because: -// * That only works with C++14 and later, and -// * We haven't included here. -template -struct __equal_to -{ - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 bool operator()(const _T1& __x, const _T1& __y) const {return __x == __y;} - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 bool operator()(const _T1& __x, const _T2& __y) const {return __x == __y;} - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 bool operator()(const _T2& __x, const _T1& __y) const {return __x == __y;} - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 bool operator()(const _T2& __x, const _T2& __y) const {return __x == __y;} -}; - -template -struct __equal_to<_T1, _T1> -{ - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 - bool operator()(const _T1& __x, const _T1& __y) const {return __x == __y;} -}; - -template -struct __equal_to -{ - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 - bool operator()(const _T1& __x, const _T1& __y) const {return __x == __y;} -}; - -template -struct __equal_to<_T1, const _T1> -{ - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 - bool operator()(const _T1& __x, const _T1& __y) const {return __x == __y;} -}; - -template -struct __less -{ - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 - bool operator()(const _T1& __x, const _T1& __y) const {return __x < __y;} - - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 - bool operator()(const _T1& __x, const _T2& __y) const {return __x < __y;} - - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 - bool operator()(const _T2& __x, const _T1& __y) const {return __x < __y;} - - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 - bool operator()(const _T2& __x, const _T2& __y) const {return __x < __y;} -}; - -template -struct __less<_T1, _T1> -{ - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 - bool operator()(const _T1& __x, const _T1& __y) const {return __x < __y;} -}; - -template -struct __less -{ - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 - bool operator()(const _T1& __x, const _T1& __y) const {return __x < __y;} -}; - -template -struct __less<_T1, const _T1> -{ - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 - bool operator()(const _T1& __x, const _T1& __y) const {return __x < __y;} -}; - -template -class __invert // invert the sense of a comparison -{ -private: - _Predicate __p_; -public: - _LIBCPP_INLINE_VISIBILITY __invert() {} - - _LIBCPP_INLINE_VISIBILITY - explicit __invert(_Predicate __p) : __p_(__p) {} - - template - _LIBCPP_INLINE_VISIBILITY - bool operator()(const _T1& __x) {return !__p_(__x);} - - template - _LIBCPP_INLINE_VISIBILITY - bool operator()(const _T1& __x, const _T2& __y) {return __p_(__y, __x);} -}; - -// Perform division by two quickly for positive integers (llvm.org/PR39129) - -template -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR -typename enable_if -< - is_integral<_Integral>::value, - _Integral ->::type -__half_positive(_Integral __value) -{ - return static_cast<_Integral>(static_cast::type>(__value) / 2); -} - -template -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR -typename enable_if -< - !is_integral<_Tp>::value, - _Tp ->::type -__half_positive(_Tp __value) -{ - return __value / 2; -} - -#ifdef _LIBCPP_DEBUG - -template -struct __debug_less -{ - _Compare &__comp_; - _LIBCPP_CONSTEXPR_AFTER_CXX17 - __debug_less(_Compare& __c) : __comp_(__c) {} - - template - _LIBCPP_CONSTEXPR_AFTER_CXX17 - bool operator()(const _Tp& __x, const _Up& __y) - { - bool __r = __comp_(__x, __y); - if (__r) - __do_compare_assert(0, __y, __x); - return __r; - } - - template - _LIBCPP_CONSTEXPR_AFTER_CXX17 - bool operator()(_Tp& __x, _Up& __y) - { - bool __r = __comp_(__x, __y); - if (__r) - __do_compare_assert(0, __y, __x); - return __r; - } - - template - _LIBCPP_CONSTEXPR_AFTER_CXX17 - inline _LIBCPP_INLINE_VISIBILITY - decltype((void)_VSTD::declval<_Compare&>()( - _VSTD::declval<_LHS &>(), _VSTD::declval<_RHS &>())) - __do_compare_assert(int, _LHS & __l, _RHS & __r) { - _LIBCPP_ASSERT(!__comp_(__l, __r), - "Comparator does not induce a strict weak ordering"); - } - - template - _LIBCPP_CONSTEXPR_AFTER_CXX17 - inline _LIBCPP_INLINE_VISIBILITY - void __do_compare_assert(long, _LHS &, _RHS &) {} -}; - -#endif // _LIBCPP_DEBUG - -template -struct __comp_ref_type { - // Pass the comparator by lvalue reference. Or in debug mode, using a - // debugging wrapper that stores a reference. -#ifndef _LIBCPP_DEBUG - typedef typename add_lvalue_reference<_Comp>::type type; -#else - typedef __debug_less<_Comp> type; -#endif -}; - -// all_of - -template -_LIBCPP_NODISCARD_EXT inline -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -bool -all_of(_InputIterator __first, _InputIterator __last, _Predicate __pred) -{ - for (; __first != __last; ++__first) - if (!__pred(*__first)) - return false; - return true; -} - -// any_of - -template -_LIBCPP_NODISCARD_EXT inline -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -bool -any_of(_InputIterator __first, _InputIterator __last, _Predicate __pred) -{ - for (; __first != __last; ++__first) - if (__pred(*__first)) - return true; - return false; -} - -// none_of - -template -_LIBCPP_NODISCARD_EXT inline -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -bool -none_of(_InputIterator __first, _InputIterator __last, _Predicate __pred) -{ - for (; __first != __last; ++__first) - if (__pred(*__first)) - return false; - return true; -} - -// for_each - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -_Function -for_each(_InputIterator __first, _InputIterator __last, _Function __f) -{ - for (; __first != __last; ++__first) - __f(*__first); - return __f; -} - -#if _LIBCPP_STD_VER > 14 -// for_each_n - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -_InputIterator -for_each_n(_InputIterator __first, _Size __orig_n, _Function __f) -{ - typedef decltype(_VSTD::__convert_to_integral(__orig_n)) _IntegralSize; - _IntegralSize __n = __orig_n; - while (__n > 0) - { - __f(*__first); - ++__first; - --__n; - } - return __first; -} -#endif - -// find - -template -_LIBCPP_NODISCARD_EXT inline -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -_InputIterator -find(_InputIterator __first, _InputIterator __last, const _Tp& __value_) -{ - for (; __first != __last; ++__first) - if (*__first == __value_) - break; - return __first; -} - -// find_if - -template -_LIBCPP_NODISCARD_EXT inline -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -_InputIterator -find_if(_InputIterator __first, _InputIterator __last, _Predicate __pred) -{ - for (; __first != __last; ++__first) - if (__pred(*__first)) - break; - return __first; -} - -// find_if_not - -template -_LIBCPP_NODISCARD_EXT inline -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -_InputIterator -find_if_not(_InputIterator __first, _InputIterator __last, _Predicate __pred) -{ - for (; __first != __last; ++__first) - if (!__pred(*__first)) - break; - return __first; -} - -// find_end - -template -_LIBCPP_CONSTEXPR_AFTER_CXX17 _ForwardIterator1 -__find_end(_ForwardIterator1 __first1, _ForwardIterator1 __last1, - _ForwardIterator2 __first2, _ForwardIterator2 __last2, _BinaryPredicate __pred, - forward_iterator_tag, forward_iterator_tag) -{ - // modeled after search algorithm - _ForwardIterator1 __r = __last1; // __last1 is the "default" answer - if (__first2 == __last2) - return __r; - while (true) - { - while (true) - { - if (__first1 == __last1) // if source exhausted return last correct answer - return __r; // (or __last1 if never found) - if (__pred(*__first1, *__first2)) - break; - ++__first1; - } - // *__first1 matches *__first2, now match elements after here - _ForwardIterator1 __m1 = __first1; - _ForwardIterator2 __m2 = __first2; - while (true) - { - if (++__m2 == __last2) - { // Pattern exhaused, record answer and search for another one - __r = __first1; - ++__first1; - break; - } - if (++__m1 == __last1) // Source exhausted, return last answer - return __r; - if (!__pred(*__m1, *__m2)) // mismatch, restart with a new __first - { - ++__first1; - break; - } // else there is a match, check next elements - } - } -} - -template -_LIBCPP_CONSTEXPR_AFTER_CXX17 _BidirectionalIterator1 -__find_end(_BidirectionalIterator1 __first1, _BidirectionalIterator1 __last1, - _BidirectionalIterator2 __first2, _BidirectionalIterator2 __last2, _BinaryPredicate __pred, - bidirectional_iterator_tag, bidirectional_iterator_tag) -{ - // modeled after search algorithm (in reverse) - if (__first2 == __last2) - return __last1; // Everything matches an empty sequence - _BidirectionalIterator1 __l1 = __last1; - _BidirectionalIterator2 __l2 = __last2; - --__l2; - while (true) - { - // Find last element in sequence 1 that matchs *(__last2-1), with a mininum of loop checks - while (true) - { - if (__first1 == __l1) // return __last1 if no element matches *__first2 - return __last1; - if (__pred(*--__l1, *__l2)) - break; - } - // *__l1 matches *__l2, now match elements before here - _BidirectionalIterator1 __m1 = __l1; - _BidirectionalIterator2 __m2 = __l2; - while (true) - { - if (__m2 == __first2) // If pattern exhausted, __m1 is the answer (works for 1 element pattern) - return __m1; - if (__m1 == __first1) // Otherwise if source exhaused, pattern not found - return __last1; - if (!__pred(*--__m1, *--__m2)) // if there is a mismatch, restart with a new __l1 - { - break; - } // else there is a match, check next elements - } - } -} - -template -_LIBCPP_CONSTEXPR_AFTER_CXX11 _RandomAccessIterator1 -__find_end(_RandomAccessIterator1 __first1, _RandomAccessIterator1 __last1, - _RandomAccessIterator2 __first2, _RandomAccessIterator2 __last2, _BinaryPredicate __pred, - random_access_iterator_tag, random_access_iterator_tag) -{ - // Take advantage of knowing source and pattern lengths. Stop short when source is smaller than pattern - typename iterator_traits<_RandomAccessIterator2>::difference_type __len2 = __last2 - __first2; - if (__len2 == 0) - return __last1; - typename iterator_traits<_RandomAccessIterator1>::difference_type __len1 = __last1 - __first1; - if (__len1 < __len2) - return __last1; - const _RandomAccessIterator1 __s = __first1 + (__len2 - 1); // End of pattern match can't go before here - _RandomAccessIterator1 __l1 = __last1; - _RandomAccessIterator2 __l2 = __last2; - --__l2; - while (true) - { - while (true) - { - if (__s == __l1) - return __last1; - if (__pred(*--__l1, *__l2)) - break; - } - _RandomAccessIterator1 __m1 = __l1; - _RandomAccessIterator2 __m2 = __l2; - while (true) - { - if (__m2 == __first2) - return __m1; - // no need to check range on __m1 because __s guarantees we have enough source - if (!__pred(*--__m1, *--__m2)) - { - break; - } - } - } -} - -template -_LIBCPP_NODISCARD_EXT inline -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -_ForwardIterator1 -find_end(_ForwardIterator1 __first1, _ForwardIterator1 __last1, - _ForwardIterator2 __first2, _ForwardIterator2 __last2, _BinaryPredicate __pred) -{ - return _VSTD::__find_end::type> - (__first1, __last1, __first2, __last2, __pred, - typename iterator_traits<_ForwardIterator1>::iterator_category(), - typename iterator_traits<_ForwardIterator2>::iterator_category()); -} - -template -_LIBCPP_NODISCARD_EXT inline -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -_ForwardIterator1 -find_end(_ForwardIterator1 __first1, _ForwardIterator1 __last1, - _ForwardIterator2 __first2, _ForwardIterator2 __last2) -{ - typedef typename iterator_traits<_ForwardIterator1>::value_type __v1; - typedef typename iterator_traits<_ForwardIterator2>::value_type __v2; - return _VSTD::find_end(__first1, __last1, __first2, __last2, __equal_to<__v1, __v2>()); -} - -// find_first_of - -template -_LIBCPP_CONSTEXPR_AFTER_CXX11 _ForwardIterator1 -__find_first_of_ce(_ForwardIterator1 __first1, _ForwardIterator1 __last1, - _ForwardIterator2 __first2, _ForwardIterator2 __last2, _BinaryPredicate __pred) -{ - for (; __first1 != __last1; ++__first1) - for (_ForwardIterator2 __j = __first2; __j != __last2; ++__j) - if (__pred(*__first1, *__j)) - return __first1; - return __last1; -} - - -template -_LIBCPP_NODISCARD_EXT inline -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -_ForwardIterator1 -find_first_of(_ForwardIterator1 __first1, _ForwardIterator1 __last1, - _ForwardIterator2 __first2, _ForwardIterator2 __last2, _BinaryPredicate __pred) -{ - return _VSTD::__find_first_of_ce(__first1, __last1, __first2, __last2, __pred); -} - -template -_LIBCPP_NODISCARD_EXT inline -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -_ForwardIterator1 -find_first_of(_ForwardIterator1 __first1, _ForwardIterator1 __last1, - _ForwardIterator2 __first2, _ForwardIterator2 __last2) -{ - typedef typename iterator_traits<_ForwardIterator1>::value_type __v1; - typedef typename iterator_traits<_ForwardIterator2>::value_type __v2; - return _VSTD::__find_first_of_ce(__first1, __last1, __first2, __last2, __equal_to<__v1, __v2>()); -} - -// adjacent_find - -template -_LIBCPP_NODISCARD_EXT inline -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -_ForwardIterator -adjacent_find(_ForwardIterator __first, _ForwardIterator __last, _BinaryPredicate __pred) -{ - if (__first != __last) - { - _ForwardIterator __i = __first; - while (++__i != __last) - { - if (__pred(*__first, *__i)) - return __first; - __first = __i; - } - } - return __last; -} - -template -_LIBCPP_NODISCARD_EXT inline -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -_ForwardIterator -adjacent_find(_ForwardIterator __first, _ForwardIterator __last) -{ - typedef typename iterator_traits<_ForwardIterator>::value_type __v; - return _VSTD::adjacent_find(__first, __last, __equal_to<__v>()); -} - -// count - -template -_LIBCPP_NODISCARD_EXT inline -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -typename iterator_traits<_InputIterator>::difference_type -count(_InputIterator __first, _InputIterator __last, const _Tp& __value_) -{ - typename iterator_traits<_InputIterator>::difference_type __r(0); - for (; __first != __last; ++__first) - if (*__first == __value_) - ++__r; - return __r; -} - -// count_if - -template -_LIBCPP_NODISCARD_EXT inline -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -typename iterator_traits<_InputIterator>::difference_type -count_if(_InputIterator __first, _InputIterator __last, _Predicate __pred) -{ - typename iterator_traits<_InputIterator>::difference_type __r(0); - for (; __first != __last; ++__first) - if (__pred(*__first)) - ++__r; - return __r; -} - -// mismatch - -template -_LIBCPP_NODISCARD_EXT inline -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -pair<_InputIterator1, _InputIterator2> -mismatch(_InputIterator1 __first1, _InputIterator1 __last1, - _InputIterator2 __first2, _BinaryPredicate __pred) -{ - for (; __first1 != __last1; ++__first1, (void) ++__first2) - if (!__pred(*__first1, *__first2)) - break; - return pair<_InputIterator1, _InputIterator2>(__first1, __first2); -} - -template -_LIBCPP_NODISCARD_EXT inline -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -pair<_InputIterator1, _InputIterator2> -mismatch(_InputIterator1 __first1, _InputIterator1 __last1, _InputIterator2 __first2) -{ - typedef typename iterator_traits<_InputIterator1>::value_type __v1; - typedef typename iterator_traits<_InputIterator2>::value_type __v2; - return _VSTD::mismatch(__first1, __last1, __first2, __equal_to<__v1, __v2>()); -} - -#if _LIBCPP_STD_VER > 11 -template -_LIBCPP_NODISCARD_EXT inline -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -pair<_InputIterator1, _InputIterator2> -mismatch(_InputIterator1 __first1, _InputIterator1 __last1, - _InputIterator2 __first2, _InputIterator2 __last2, - _BinaryPredicate __pred) -{ - for (; __first1 != __last1 && __first2 != __last2; ++__first1, (void) ++__first2) - if (!__pred(*__first1, *__first2)) - break; - return pair<_InputIterator1, _InputIterator2>(__first1, __first2); -} - -template -_LIBCPP_NODISCARD_EXT inline -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -pair<_InputIterator1, _InputIterator2> -mismatch(_InputIterator1 __first1, _InputIterator1 __last1, - _InputIterator2 __first2, _InputIterator2 __last2) -{ - typedef typename iterator_traits<_InputIterator1>::value_type __v1; - typedef typename iterator_traits<_InputIterator2>::value_type __v2; - return _VSTD::mismatch(__first1, __last1, __first2, __last2, __equal_to<__v1, __v2>()); -} -#endif - -// equal - -template -_LIBCPP_NODISCARD_EXT inline -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -bool -equal(_InputIterator1 __first1, _InputIterator1 __last1, _InputIterator2 __first2, _BinaryPredicate __pred) -{ - for (; __first1 != __last1; ++__first1, (void) ++__first2) - if (!__pred(*__first1, *__first2)) - return false; - return true; -} - -template -_LIBCPP_NODISCARD_EXT inline -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -bool -equal(_InputIterator1 __first1, _InputIterator1 __last1, _InputIterator2 __first2) -{ - typedef typename iterator_traits<_InputIterator1>::value_type __v1; - typedef typename iterator_traits<_InputIterator2>::value_type __v2; - return _VSTD::equal(__first1, __last1, __first2, __equal_to<__v1, __v2>()); -} - -#if _LIBCPP_STD_VER > 11 -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -bool -__equal(_InputIterator1 __first1, _InputIterator1 __last1, - _InputIterator2 __first2, _InputIterator2 __last2, _BinaryPredicate __pred, - input_iterator_tag, input_iterator_tag ) -{ - for (; __first1 != __last1 && __first2 != __last2; ++__first1, (void) ++__first2) - if (!__pred(*__first1, *__first2)) - return false; - return __first1 == __last1 && __first2 == __last2; -} - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -bool -__equal(_RandomAccessIterator1 __first1, _RandomAccessIterator1 __last1, - _RandomAccessIterator2 __first2, _RandomAccessIterator2 __last2, _BinaryPredicate __pred, - random_access_iterator_tag, random_access_iterator_tag ) -{ - if ( _VSTD::distance(__first1, __last1) != _VSTD::distance(__first2, __last2)) - return false; - return _VSTD::equal<_RandomAccessIterator1, _RandomAccessIterator2, - typename add_lvalue_reference<_BinaryPredicate>::type> - (__first1, __last1, __first2, __pred ); -} - -template -_LIBCPP_NODISCARD_EXT inline -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -bool -equal(_InputIterator1 __first1, _InputIterator1 __last1, - _InputIterator2 __first2, _InputIterator2 __last2, _BinaryPredicate __pred ) -{ - return _VSTD::__equal::type> - (__first1, __last1, __first2, __last2, __pred, - typename iterator_traits<_InputIterator1>::iterator_category(), - typename iterator_traits<_InputIterator2>::iterator_category()); -} - -template -_LIBCPP_NODISCARD_EXT inline -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -bool -equal(_InputIterator1 __first1, _InputIterator1 __last1, - _InputIterator2 __first2, _InputIterator2 __last2) -{ - typedef typename iterator_traits<_InputIterator1>::value_type __v1; - typedef typename iterator_traits<_InputIterator2>::value_type __v2; - return _VSTD::__equal(__first1, __last1, __first2, __last2, __equal_to<__v1, __v2>(), - typename iterator_traits<_InputIterator1>::iterator_category(), - typename iterator_traits<_InputIterator2>::iterator_category()); -} -#endif - -// is_permutation - -template -_LIBCPP_NODISCARD_EXT _LIBCPP_CONSTEXPR_AFTER_CXX17 bool -is_permutation(_ForwardIterator1 __first1, _ForwardIterator1 __last1, - _ForwardIterator2 __first2, _BinaryPredicate __pred) -{ -// shorten sequences as much as possible by lopping of any equal prefix - for (; __first1 != __last1; ++__first1, (void) ++__first2) - if (!__pred(*__first1, *__first2)) - break; - if (__first1 == __last1) - return true; - -// __first1 != __last1 && *__first1 != *__first2 - typedef typename iterator_traits<_ForwardIterator1>::difference_type _D1; - _D1 __l1 = _VSTD::distance(__first1, __last1); - if (__l1 == _D1(1)) - return false; - _ForwardIterator2 __last2 = _VSTD::next(__first2, __l1); - // For each element in [f1, l1) see if there are the same number of - // equal elements in [f2, l2) - for (_ForwardIterator1 __i = __first1; __i != __last1; ++__i) - { - // Have we already counted the number of *__i in [f1, l1)? - _ForwardIterator1 __match = __first1; - for (; __match != __i; ++__match) - if (__pred(*__match, *__i)) - break; - if (__match == __i) { - // Count number of *__i in [f2, l2) - _D1 __c2 = 0; - for (_ForwardIterator2 __j = __first2; __j != __last2; ++__j) - if (__pred(*__i, *__j)) - ++__c2; - if (__c2 == 0) - return false; - // Count number of *__i in [__i, l1) (we can start with 1) - _D1 __c1 = 1; - for (_ForwardIterator1 __j = _VSTD::next(__i); __j != __last1; ++__j) - if (__pred(*__i, *__j)) - ++__c1; - if (__c1 != __c2) - return false; - } - } - return true; -} - -template -_LIBCPP_NODISCARD_EXT inline -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -bool -is_permutation(_ForwardIterator1 __first1, _ForwardIterator1 __last1, - _ForwardIterator2 __first2) -{ - typedef typename iterator_traits<_ForwardIterator1>::value_type __v1; - typedef typename iterator_traits<_ForwardIterator2>::value_type __v2; - return _VSTD::is_permutation(__first1, __last1, __first2, __equal_to<__v1, __v2>()); -} - -#if _LIBCPP_STD_VER > 11 -template -_LIBCPP_CONSTEXPR_AFTER_CXX17 bool -__is_permutation(_ForwardIterator1 __first1, _ForwardIterator1 __last1, - _ForwardIterator2 __first2, _ForwardIterator2 __last2, - _BinaryPredicate __pred, - forward_iterator_tag, forward_iterator_tag ) -{ -// shorten sequences as much as possible by lopping of any equal prefix - for (; __first1 != __last1 && __first2 != __last2; ++__first1, (void) ++__first2) - if (!__pred(*__first1, *__first2)) - break; - if (__first1 == __last1) - return __first2 == __last2; - else if (__first2 == __last2) - return false; - - typedef typename iterator_traits<_ForwardIterator1>::difference_type _D1; - _D1 __l1 = _VSTD::distance(__first1, __last1); - - typedef typename iterator_traits<_ForwardIterator2>::difference_type _D2; - _D2 __l2 = _VSTD::distance(__first2, __last2); - if (__l1 != __l2) - return false; - - // For each element in [f1, l1) see if there are the same number of - // equal elements in [f2, l2) - for (_ForwardIterator1 __i = __first1; __i != __last1; ++__i) - { - // Have we already counted the number of *__i in [f1, l1)? - _ForwardIterator1 __match = __first1; - for (; __match != __i; ++__match) - if (__pred(*__match, *__i)) - break; - if (__match == __i) { - // Count number of *__i in [f2, l2) - _D1 __c2 = 0; - for (_ForwardIterator2 __j = __first2; __j != __last2; ++__j) - if (__pred(*__i, *__j)) - ++__c2; - if (__c2 == 0) - return false; - // Count number of *__i in [__i, l1) (we can start with 1) - _D1 __c1 = 1; - for (_ForwardIterator1 __j = _VSTD::next(__i); __j != __last1; ++__j) - if (__pred(*__i, *__j)) - ++__c1; - if (__c1 != __c2) - return false; - } - } - return true; -} - -template -_LIBCPP_CONSTEXPR_AFTER_CXX17 bool -__is_permutation(_RandomAccessIterator1 __first1, _RandomAccessIterator2 __last1, - _RandomAccessIterator1 __first2, _RandomAccessIterator2 __last2, - _BinaryPredicate __pred, - random_access_iterator_tag, random_access_iterator_tag ) -{ - if ( _VSTD::distance(__first1, __last1) != _VSTD::distance(__first2, __last2)) - return false; - return _VSTD::is_permutation<_RandomAccessIterator1, _RandomAccessIterator2, - typename add_lvalue_reference<_BinaryPredicate>::type> - (__first1, __last1, __first2, __pred ); -} - -template -_LIBCPP_NODISCARD_EXT inline -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -bool -is_permutation(_ForwardIterator1 __first1, _ForwardIterator1 __last1, - _ForwardIterator2 __first2, _ForwardIterator2 __last2, - _BinaryPredicate __pred ) -{ - return _VSTD::__is_permutation::type> - (__first1, __last1, __first2, __last2, __pred, - typename iterator_traits<_ForwardIterator1>::iterator_category(), - typename iterator_traits<_ForwardIterator2>::iterator_category()); -} - -template -_LIBCPP_NODISCARD_EXT inline -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -bool -is_permutation(_ForwardIterator1 __first1, _ForwardIterator1 __last1, - _ForwardIterator2 __first2, _ForwardIterator2 __last2) -{ - typedef typename iterator_traits<_ForwardIterator1>::value_type __v1; - typedef typename iterator_traits<_ForwardIterator2>::value_type __v2; - return _VSTD::__is_permutation(__first1, __last1, __first2, __last2, - __equal_to<__v1, __v2>(), - typename iterator_traits<_ForwardIterator1>::iterator_category(), - typename iterator_traits<_ForwardIterator2>::iterator_category()); -} -#endif - -// search -// __search is in - -template -_LIBCPP_NODISCARD_EXT inline -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -_ForwardIterator1 -search(_ForwardIterator1 __first1, _ForwardIterator1 __last1, - _ForwardIterator2 __first2, _ForwardIterator2 __last2, _BinaryPredicate __pred) -{ - return _VSTD::__search::type> - (__first1, __last1, __first2, __last2, __pred, - typename iterator_traits<_ForwardIterator1>::iterator_category(), - typename iterator_traits<_ForwardIterator2>::iterator_category()) - .first; -} - -template -_LIBCPP_NODISCARD_EXT inline -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -_ForwardIterator1 -search(_ForwardIterator1 __first1, _ForwardIterator1 __last1, - _ForwardIterator2 __first2, _ForwardIterator2 __last2) -{ - typedef typename iterator_traits<_ForwardIterator1>::value_type __v1; - typedef typename iterator_traits<_ForwardIterator2>::value_type __v2; - return _VSTD::search(__first1, __last1, __first2, __last2, __equal_to<__v1, __v2>()); -} - - -#if _LIBCPP_STD_VER > 14 -template -_LIBCPP_NODISCARD_EXT _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -_ForwardIterator search(_ForwardIterator __f, _ForwardIterator __l, const _Searcher &__s) -{ return __s(__f, __l).first; } -#endif - -// search_n - -template -_LIBCPP_CONSTEXPR_AFTER_CXX17 _ForwardIterator -__search_n(_ForwardIterator __first, _ForwardIterator __last, - _Size __count, const _Tp& __value_, _BinaryPredicate __pred, forward_iterator_tag) -{ - if (__count <= 0) - return __first; - while (true) - { - // Find first element in sequence that matchs __value_, with a mininum of loop checks - while (true) - { - if (__first == __last) // return __last if no element matches __value_ - return __last; - if (__pred(*__first, __value_)) - break; - ++__first; - } - // *__first matches __value_, now match elements after here - _ForwardIterator __m = __first; - _Size __c(0); - while (true) - { - if (++__c == __count) // If pattern exhausted, __first is the answer (works for 1 element pattern) - return __first; - if (++__m == __last) // Otherwise if source exhaused, pattern not found - return __last; - if (!__pred(*__m, __value_)) // if there is a mismatch, restart with a new __first - { - __first = __m; - ++__first; - break; - } // else there is a match, check next elements - } - } -} - -template -_LIBCPP_CONSTEXPR_AFTER_CXX17 _RandomAccessIterator -__search_n(_RandomAccessIterator __first, _RandomAccessIterator __last, - _Size __count, const _Tp& __value_, _BinaryPredicate __pred, random_access_iterator_tag) -{ - if (__count <= 0) - return __first; - _Size __len = static_cast<_Size>(__last - __first); - if (__len < __count) - return __last; - const _RandomAccessIterator __s = __last - (__count - 1); // Start of pattern match can't go beyond here - while (true) - { - // Find first element in sequence that matchs __value_, with a mininum of loop checks - while (true) - { - if (__first >= __s) // return __last if no element matches __value_ - return __last; - if (__pred(*__first, __value_)) - break; - ++__first; - } - // *__first matches __value_, now match elements after here - _RandomAccessIterator __m = __first; - _Size __c(0); - while (true) - { - if (++__c == __count) // If pattern exhausted, __first is the answer (works for 1 element pattern) - return __first; - ++__m; // no need to check range on __m because __s guarantees we have enough source - if (!__pred(*__m, __value_)) // if there is a mismatch, restart with a new __first - { - __first = __m; - ++__first; - break; - } // else there is a match, check next elements - } - } -} - -template -_LIBCPP_NODISCARD_EXT inline -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -_ForwardIterator -search_n(_ForwardIterator __first, _ForwardIterator __last, - _Size __count, const _Tp& __value_, _BinaryPredicate __pred) -{ - return _VSTD::__search_n::type> - (__first, __last, _VSTD::__convert_to_integral(__count), __value_, __pred, - typename iterator_traits<_ForwardIterator>::iterator_category()); -} - -template -_LIBCPP_NODISCARD_EXT inline -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -_ForwardIterator -search_n(_ForwardIterator __first, _ForwardIterator __last, _Size __count, const _Tp& __value_) -{ - typedef typename iterator_traits<_ForwardIterator>::value_type __v; - return _VSTD::search_n(__first, __last, _VSTD::__convert_to_integral(__count), - __value_, __equal_to<__v, _Tp>()); -} - -// copy -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR -_Iter -__unwrap_iter(_Iter __i) -{ - return __i; -} - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR -typename enable_if -< - is_trivially_copy_assignable<_Tp>::value, - _Tp* ->::type -__unwrap_iter(move_iterator<_Tp*> __i) -{ - return __i.base(); -} - -#if _LIBCPP_DEBUG_LEVEL < 2 - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR -typename enable_if -< - is_trivially_copy_assignable<_Tp>::value, - _Tp* ->::type -__unwrap_iter(__wrap_iter<_Tp*> __i) -{ - return __i.base(); -} - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR -typename enable_if -< - is_trivially_copy_assignable<_Tp>::value, - const _Tp* ->::type -__unwrap_iter(__wrap_iter __i) -{ - return __i.base(); -} - -#else - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR -typename enable_if -< - is_trivially_copy_assignable<_Tp>::value, - __wrap_iter<_Tp*> ->::type -__unwrap_iter(__wrap_iter<_Tp*> __i) -{ - return __i; -} - -#endif // _LIBCPP_DEBUG_LEVEL < 2 - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -_OutputIterator -__copy_constexpr(_InputIterator __first, _InputIterator __last, _OutputIterator __result) -{ - for (; __first != __last; ++__first, (void) ++__result) - *__result = *__first; - return __result; -} - -template -inline _LIBCPP_INLINE_VISIBILITY -_OutputIterator -__copy(_InputIterator __first, _InputIterator __last, _OutputIterator __result) -{ - return _VSTD::__copy_constexpr(__first, __last, __result); -} - -template -inline _LIBCPP_INLINE_VISIBILITY -typename enable_if -< - is_same::type, _Up>::value && - is_trivially_copy_assignable<_Up>::value, - _Up* ->::type -__copy(_Tp* __first, _Tp* __last, _Up* __result) -{ - const size_t __n = static_cast(__last - __first); - if (__n > 0) - _VSTD::memmove(__result, __first, __n * sizeof(_Up)); - return __result + __n; -} - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -_OutputIterator -copy(_InputIterator __first, _InputIterator __last, _OutputIterator __result) -{ - if (__libcpp_is_constant_evaluated()) { - return _VSTD::__copy_constexpr( - _VSTD::__unwrap_iter(__first), _VSTD::__unwrap_iter(__last), _VSTD::__unwrap_iter(__result)); - } else { - return _VSTD::__copy( - _VSTD::__unwrap_iter(__first), _VSTD::__unwrap_iter(__last), _VSTD::__unwrap_iter(__result)); - } -} - -// copy_backward - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -_OutputIterator -__copy_backward_constexpr(_BidirectionalIterator __first, _BidirectionalIterator __last, _OutputIterator __result) -{ - while (__first != __last) - *--__result = *--__last; - return __result; -} - -template -inline _LIBCPP_INLINE_VISIBILITY -_OutputIterator -__copy_backward(_BidirectionalIterator __first, _BidirectionalIterator __last, _OutputIterator __result) -{ - return _VSTD::__copy_backward_constexpr(__first, __last, __result); -} - -template -inline _LIBCPP_INLINE_VISIBILITY -typename enable_if -< - is_same::type, _Up>::value && - is_trivially_copy_assignable<_Up>::value, - _Up* ->::type -__copy_backward(_Tp* __first, _Tp* __last, _Up* __result) -{ - const size_t __n = static_cast(__last - __first); - if (__n > 0) - { - __result -= __n; - _VSTD::memmove(__result, __first, __n * sizeof(_Up)); - } - return __result; -} - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -_BidirectionalIterator2 -copy_backward(_BidirectionalIterator1 __first, _BidirectionalIterator1 __last, - _BidirectionalIterator2 __result) -{ - if (__libcpp_is_constant_evaluated()) { - return _VSTD::__copy_backward_constexpr(_VSTD::__unwrap_iter(__first), - _VSTD::__unwrap_iter(__last), - _VSTD::__unwrap_iter(__result)); - } else { - return _VSTD::__copy_backward(_VSTD::__unwrap_iter(__first), - _VSTD::__unwrap_iter(__last), - _VSTD::__unwrap_iter(__result)); - } -} - -// copy_if - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -_OutputIterator -copy_if(_InputIterator __first, _InputIterator __last, - _OutputIterator __result, _Predicate __pred) -{ - for (; __first != __last; ++__first) - { - if (__pred(*__first)) - { - *__result = *__first; - ++__result; - } - } - return __result; -} - -// copy_n - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -typename enable_if -< - __is_cpp17_input_iterator<_InputIterator>::value && - !__is_cpp17_random_access_iterator<_InputIterator>::value, - _OutputIterator ->::type -copy_n(_InputIterator __first, _Size __orig_n, _OutputIterator __result) -{ - typedef decltype(_VSTD::__convert_to_integral(__orig_n)) _IntegralSize; - _IntegralSize __n = __orig_n; - if (__n > 0) - { - *__result = *__first; - ++__result; - for (--__n; __n > 0; --__n) - { - ++__first; - *__result = *__first; - ++__result; - } - } - return __result; -} - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -typename enable_if -< - __is_cpp17_random_access_iterator<_InputIterator>::value, - _OutputIterator ->::type -copy_n(_InputIterator __first, _Size __orig_n, _OutputIterator __result) -{ - typedef decltype(_VSTD::__convert_to_integral(__orig_n)) _IntegralSize; - _IntegralSize __n = __orig_n; - return _VSTD::copy(__first, __first + __n, __result); -} - -// move - -// __move_constexpr exists so that __move doesn't call itself when delegating to the constexpr -// version of __move. -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 -_OutputIterator -__move_constexpr(_InputIterator __first, _InputIterator __last, _OutputIterator __result) -{ - for (; __first != __last; ++__first, (void) ++__result) - *__result = _VSTD::move(*__first); - return __result; -} - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 -_OutputIterator -__move(_InputIterator __first, _InputIterator __last, _OutputIterator __result) -{ - return _VSTD::__move_constexpr(__first, __last, __result); -} - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 -typename enable_if -< - is_same::type, _Up>::value && - is_trivially_copy_assignable<_Up>::value, - _Up* ->::type -__move(_Tp* __first, _Tp* __last, _Up* __result) -{ - if (__libcpp_is_constant_evaluated()) - return _VSTD::__move_constexpr(__first, __last, __result); - const size_t __n = static_cast(__last - __first); - if (__n > 0) - _VSTD::memmove(__result, __first, __n * sizeof(_Up)); - return __result + __n; -} - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -_OutputIterator -move(_InputIterator __first, _InputIterator __last, _OutputIterator __result) -{ - return _VSTD::__move(_VSTD::__unwrap_iter(__first), _VSTD::__unwrap_iter(__last), _VSTD::__unwrap_iter(__result)); -} - -// move_backward - -// __move_backward_constexpr exists so that __move_backward doesn't call itself when delegating to -// the constexpr version of __move_backward. -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 -_OutputIterator -__move_backward_constexpr(_InputIterator __first, _InputIterator __last, _OutputIterator __result) -{ - while (__first != __last) - *--__result = _VSTD::move(*--__last); - return __result; -} - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 -_OutputIterator -__move_backward(_InputIterator __first, _InputIterator __last, _OutputIterator __result) -{ - return _VSTD::__move_backward_constexpr(__first, __last, __result); -} - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 -typename enable_if -< - is_same::type, _Up>::value && - is_trivially_copy_assignable<_Up>::value, - _Up* ->::type -__move_backward(_Tp* __first, _Tp* __last, _Up* __result) -{ - if (__libcpp_is_constant_evaluated()) - return _VSTD::__move_backward_constexpr(__first, __last, __result); - const size_t __n = static_cast(__last - __first); - if (__n > 0) - { - __result -= __n; - _VSTD::memmove(__result, __first, __n * sizeof(_Up)); - } - return __result; -} - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -_BidirectionalIterator2 -move_backward(_BidirectionalIterator1 __first, _BidirectionalIterator1 __last, - _BidirectionalIterator2 __result) -{ - return _VSTD::__move_backward(_VSTD::__unwrap_iter(__first), _VSTD::__unwrap_iter(__last), _VSTD::__unwrap_iter(__result)); -} - -// iter_swap - -// moved to for better swap / noexcept support - -// transform - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -_OutputIterator -transform(_InputIterator __first, _InputIterator __last, _OutputIterator __result, _UnaryOperation __op) -{ - for (; __first != __last; ++__first, (void) ++__result) - *__result = __op(*__first); - return __result; -} - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -_OutputIterator -transform(_InputIterator1 __first1, _InputIterator1 __last1, _InputIterator2 __first2, - _OutputIterator __result, _BinaryOperation __binary_op) -{ - for (; __first1 != __last1; ++__first1, (void) ++__first2, ++__result) - *__result = __binary_op(*__first1, *__first2); - return __result; -} - -// replace - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -void -replace(_ForwardIterator __first, _ForwardIterator __last, const _Tp& __old_value, const _Tp& __new_value) -{ - for (; __first != __last; ++__first) - if (*__first == __old_value) - *__first = __new_value; -} - -// replace_if - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -void -replace_if(_ForwardIterator __first, _ForwardIterator __last, _Predicate __pred, const _Tp& __new_value) -{ - for (; __first != __last; ++__first) - if (__pred(*__first)) - *__first = __new_value; -} - -// replace_copy - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -_OutputIterator -replace_copy(_InputIterator __first, _InputIterator __last, _OutputIterator __result, - const _Tp& __old_value, const _Tp& __new_value) -{ - for (; __first != __last; ++__first, (void) ++__result) - if (*__first == __old_value) - *__result = __new_value; - else - *__result = *__first; - return __result; -} - -// replace_copy_if - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -_OutputIterator -replace_copy_if(_InputIterator __first, _InputIterator __last, _OutputIterator __result, - _Predicate __pred, const _Tp& __new_value) -{ - for (; __first != __last; ++__first, (void) ++__result) - if (__pred(*__first)) - *__result = __new_value; - else - *__result = *__first; - return __result; -} - -// fill_n - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -_OutputIterator -__fill_n(_OutputIterator __first, _Size __n, const _Tp& __value_) -{ - for (; __n > 0; ++__first, (void) --__n) - *__first = __value_; - return __first; -} - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -_OutputIterator -fill_n(_OutputIterator __first, _Size __n, const _Tp& __value_) -{ - return _VSTD::__fill_n(__first, _VSTD::__convert_to_integral(__n), __value_); -} - -// fill - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -void -__fill(_ForwardIterator __first, _ForwardIterator __last, const _Tp& __value_, forward_iterator_tag) -{ - for (; __first != __last; ++__first) - *__first = __value_; -} - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -void -__fill(_RandomAccessIterator __first, _RandomAccessIterator __last, const _Tp& __value_, random_access_iterator_tag) -{ - _VSTD::fill_n(__first, __last - __first, __value_); -} - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -void -fill(_ForwardIterator __first, _ForwardIterator __last, const _Tp& __value_) -{ - _VSTD::__fill(__first, __last, __value_, typename iterator_traits<_ForwardIterator>::iterator_category()); -} - -// generate - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -void -generate(_ForwardIterator __first, _ForwardIterator __last, _Generator __gen) -{ - for (; __first != __last; ++__first) - *__first = __gen(); -} - -// generate_n - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -_OutputIterator -generate_n(_OutputIterator __first, _Size __orig_n, _Generator __gen) -{ - typedef decltype(_VSTD::__convert_to_integral(__orig_n)) _IntegralSize; - _IntegralSize __n = __orig_n; - for (; __n > 0; ++__first, (void) --__n) - *__first = __gen(); - return __first; -} - -// remove - -template -_LIBCPP_NODISCARD_EXT _LIBCPP_CONSTEXPR_AFTER_CXX17 _ForwardIterator -remove(_ForwardIterator __first, _ForwardIterator __last, const _Tp& __value_) -{ - __first = _VSTD::find(__first, __last, __value_); - if (__first != __last) - { - _ForwardIterator __i = __first; - while (++__i != __last) - { - if (!(*__i == __value_)) - { - *__first = _VSTD::move(*__i); - ++__first; - } - } - } - return __first; -} - -// remove_if - -template -_LIBCPP_NODISCARD_EXT _LIBCPP_CONSTEXPR_AFTER_CXX17 _ForwardIterator -remove_if(_ForwardIterator __first, _ForwardIterator __last, _Predicate __pred) -{ - __first = _VSTD::find_if<_ForwardIterator, typename add_lvalue_reference<_Predicate>::type> - (__first, __last, __pred); - if (__first != __last) - { - _ForwardIterator __i = __first; - while (++__i != __last) - { - if (!__pred(*__i)) - { - *__first = _VSTD::move(*__i); - ++__first; - } - } - } - return __first; -} - -// remove_copy - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -_OutputIterator -remove_copy(_InputIterator __first, _InputIterator __last, _OutputIterator __result, const _Tp& __value_) -{ - for (; __first != __last; ++__first) - { - if (!(*__first == __value_)) - { - *__result = *__first; - ++__result; - } - } - return __result; -} - -// remove_copy_if - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -_OutputIterator -remove_copy_if(_InputIterator __first, _InputIterator __last, _OutputIterator __result, _Predicate __pred) -{ - for (; __first != __last; ++__first) - { - if (!__pred(*__first)) - { - *__result = *__first; - ++__result; - } - } - return __result; -} - -// unique - -template -_LIBCPP_NODISCARD_EXT _LIBCPP_CONSTEXPR_AFTER_CXX17 _ForwardIterator -unique(_ForwardIterator __first, _ForwardIterator __last, _BinaryPredicate __pred) -{ - __first = _VSTD::adjacent_find<_ForwardIterator, typename add_lvalue_reference<_BinaryPredicate>::type> - (__first, __last, __pred); - if (__first != __last) - { - // ... a a ? ... - // f i - _ForwardIterator __i = __first; - for (++__i; ++__i != __last;) - if (!__pred(*__first, *__i)) - *++__first = _VSTD::move(*__i); - ++__first; - } - return __first; -} - -template -_LIBCPP_NODISCARD_EXT inline -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -_ForwardIterator -unique(_ForwardIterator __first, _ForwardIterator __last) -{ - typedef typename iterator_traits<_ForwardIterator>::value_type __v; - return _VSTD::unique(__first, __last, __equal_to<__v>()); -} - -// unique_copy - -template -_LIBCPP_CONSTEXPR_AFTER_CXX17 _OutputIterator -__unique_copy(_InputIterator __first, _InputIterator __last, _OutputIterator __result, _BinaryPredicate __pred, - input_iterator_tag, output_iterator_tag) -{ - if (__first != __last) - { - typename iterator_traits<_InputIterator>::value_type __t(*__first); - *__result = __t; - ++__result; - while (++__first != __last) - { - if (!__pred(__t, *__first)) - { - __t = *__first; - *__result = __t; - ++__result; - } - } - } - return __result; -} - -template -_LIBCPP_CONSTEXPR_AFTER_CXX17 _OutputIterator -__unique_copy(_ForwardIterator __first, _ForwardIterator __last, _OutputIterator __result, _BinaryPredicate __pred, - forward_iterator_tag, output_iterator_tag) -{ - if (__first != __last) - { - _ForwardIterator __i = __first; - *__result = *__i; - ++__result; - while (++__first != __last) - { - if (!__pred(*__i, *__first)) - { - *__result = *__first; - ++__result; - __i = __first; - } - } - } - return __result; -} - -template -_LIBCPP_CONSTEXPR_AFTER_CXX17 _ForwardIterator -__unique_copy(_InputIterator __first, _InputIterator __last, _ForwardIterator __result, _BinaryPredicate __pred, - input_iterator_tag, forward_iterator_tag) -{ - if (__first != __last) - { - *__result = *__first; - while (++__first != __last) - if (!__pred(*__result, *__first)) - *++__result = *__first; - ++__result; - } - return __result; -} - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -_OutputIterator -unique_copy(_InputIterator __first, _InputIterator __last, _OutputIterator __result, _BinaryPredicate __pred) -{ - return _VSTD::__unique_copy::type> - (__first, __last, __result, __pred, - typename iterator_traits<_InputIterator>::iterator_category(), - typename iterator_traits<_OutputIterator>::iterator_category()); -} - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -_OutputIterator -unique_copy(_InputIterator __first, _InputIterator __last, _OutputIterator __result) -{ - typedef typename iterator_traits<_InputIterator>::value_type __v; - return _VSTD::unique_copy(__first, __last, __result, __equal_to<__v>()); -} - -// reverse - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -void -__reverse(_BidirectionalIterator __first, _BidirectionalIterator __last, bidirectional_iterator_tag) -{ - while (__first != __last) - { - if (__first == --__last) - break; - _VSTD::iter_swap(__first, __last); - ++__first; - } -} - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -void -__reverse(_RandomAccessIterator __first, _RandomAccessIterator __last, random_access_iterator_tag) -{ - if (__first != __last) - for (; __first < --__last; ++__first) - _VSTD::iter_swap(__first, __last); -} - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -void -reverse(_BidirectionalIterator __first, _BidirectionalIterator __last) -{ - _VSTD::__reverse(__first, __last, typename iterator_traits<_BidirectionalIterator>::iterator_category()); -} - -// reverse_copy - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -_OutputIterator -reverse_copy(_BidirectionalIterator __first, _BidirectionalIterator __last, _OutputIterator __result) -{ - for (; __first != __last; ++__result) - *__result = *--__last; - return __result; -} - -// rotate - -template -_LIBCPP_CONSTEXPR_AFTER_CXX11 _ForwardIterator -__rotate_left(_ForwardIterator __first, _ForwardIterator __last) -{ - typedef typename iterator_traits<_ForwardIterator>::value_type value_type; - value_type __tmp = _VSTD::move(*__first); - _ForwardIterator __lm1 = _VSTD::move(_VSTD::next(__first), __last, __first); - *__lm1 = _VSTD::move(__tmp); - return __lm1; -} - -template -_LIBCPP_CONSTEXPR_AFTER_CXX11 _BidirectionalIterator -__rotate_right(_BidirectionalIterator __first, _BidirectionalIterator __last) -{ - typedef typename iterator_traits<_BidirectionalIterator>::value_type value_type; - _BidirectionalIterator __lm1 = _VSTD::prev(__last); - value_type __tmp = _VSTD::move(*__lm1); - _BidirectionalIterator __fp1 = _VSTD::move_backward(__first, __lm1, __last); - *__first = _VSTD::move(__tmp); - return __fp1; -} - -template -_LIBCPP_CONSTEXPR_AFTER_CXX14 _ForwardIterator -__rotate_forward(_ForwardIterator __first, _ForwardIterator __middle, _ForwardIterator __last) -{ - _ForwardIterator __i = __middle; - while (true) - { - swap(*__first, *__i); - ++__first; - if (++__i == __last) - break; - if (__first == __middle) - __middle = __i; - } - _ForwardIterator __r = __first; - if (__first != __middle) - { - __i = __middle; - while (true) - { - swap(*__first, *__i); - ++__first; - if (++__i == __last) - { - if (__first == __middle) - break; - __i = __middle; - } - else if (__first == __middle) - __middle = __i; - } - } - return __r; -} - -template -inline _LIBCPP_INLINE_VISIBILITY -_LIBCPP_CONSTEXPR_AFTER_CXX14 _Integral -__algo_gcd(_Integral __x, _Integral __y) -{ - do - { - _Integral __t = __x % __y; - __x = __y; - __y = __t; - } while (__y); - return __x; -} - -template -_LIBCPP_CONSTEXPR_AFTER_CXX14 _RandomAccessIterator -__rotate_gcd(_RandomAccessIterator __first, _RandomAccessIterator __middle, _RandomAccessIterator __last) -{ - typedef typename iterator_traits<_RandomAccessIterator>::difference_type difference_type; - typedef typename iterator_traits<_RandomAccessIterator>::value_type value_type; - - const difference_type __m1 = __middle - __first; - const difference_type __m2 = __last - __middle; - if (__m1 == __m2) - { - _VSTD::swap_ranges(__first, __middle, __middle); - return __middle; - } - const difference_type __g = _VSTD::__algo_gcd(__m1, __m2); - for (_RandomAccessIterator __p = __first + __g; __p != __first;) - { - value_type __t(_VSTD::move(*--__p)); - _RandomAccessIterator __p1 = __p; - _RandomAccessIterator __p2 = __p1 + __m1; - do - { - *__p1 = _VSTD::move(*__p2); - __p1 = __p2; - const difference_type __d = __last - __p2; - if (__m1 < __d) - __p2 += __m1; - else - __p2 = __first + (__m1 - __d); - } while (__p2 != __p); - *__p1 = _VSTD::move(__t); - } - return __first + __m2; -} - -template -inline _LIBCPP_INLINE_VISIBILITY -_LIBCPP_CONSTEXPR_AFTER_CXX11 _ForwardIterator -__rotate(_ForwardIterator __first, _ForwardIterator __middle, _ForwardIterator __last, - _VSTD::forward_iterator_tag) -{ - typedef typename _VSTD::iterator_traits<_ForwardIterator>::value_type value_type; - if (_VSTD::is_trivially_move_assignable::value) - { - if (_VSTD::next(__first) == __middle) - return _VSTD::__rotate_left(__first, __last); - } - return _VSTD::__rotate_forward(__first, __middle, __last); -} - -template -inline _LIBCPP_INLINE_VISIBILITY -_LIBCPP_CONSTEXPR_AFTER_CXX11 _BidirectionalIterator -__rotate(_BidirectionalIterator __first, _BidirectionalIterator __middle, _BidirectionalIterator __last, - _VSTD::bidirectional_iterator_tag) -{ - typedef typename _VSTD::iterator_traits<_BidirectionalIterator>::value_type value_type; - if (_VSTD::is_trivially_move_assignable::value) - { - if (_VSTD::next(__first) == __middle) - return _VSTD::__rotate_left(__first, __last); - if (_VSTD::next(__middle) == __last) - return _VSTD::__rotate_right(__first, __last); - } - return _VSTD::__rotate_forward(__first, __middle, __last); -} - -template -inline _LIBCPP_INLINE_VISIBILITY -_LIBCPP_CONSTEXPR_AFTER_CXX11 _RandomAccessIterator -__rotate(_RandomAccessIterator __first, _RandomAccessIterator __middle, _RandomAccessIterator __last, - _VSTD::random_access_iterator_tag) -{ - typedef typename _VSTD::iterator_traits<_RandomAccessIterator>::value_type value_type; - if (_VSTD::is_trivially_move_assignable::value) - { - if (_VSTD::next(__first) == __middle) - return _VSTD::__rotate_left(__first, __last); - if (_VSTD::next(__middle) == __last) - return _VSTD::__rotate_right(__first, __last); - return _VSTD::__rotate_gcd(__first, __middle, __last); - } - return _VSTD::__rotate_forward(__first, __middle, __last); -} - -template -inline _LIBCPP_INLINE_VISIBILITY -_LIBCPP_CONSTEXPR_AFTER_CXX17 _ForwardIterator -rotate(_ForwardIterator __first, _ForwardIterator __middle, _ForwardIterator __last) -{ - if (__first == __middle) - return __last; - if (__middle == __last) - return __first; - return _VSTD::__rotate(__first, __middle, __last, - typename _VSTD::iterator_traits<_ForwardIterator>::iterator_category()); -} - -// rotate_copy - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -_OutputIterator -rotate_copy(_ForwardIterator __first, _ForwardIterator __middle, _ForwardIterator __last, _OutputIterator __result) -{ - return _VSTD::copy(__first, __middle, _VSTD::copy(__middle, __last, __result)); -} - -// min_element - -template -_LIBCPP_NODISCARD_EXT inline -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 -_ForwardIterator -min_element(_ForwardIterator __first, _ForwardIterator __last, _Compare __comp) -{ - static_assert(__is_cpp17_forward_iterator<_ForwardIterator>::value, - "std::min_element requires a ForwardIterator"); - if (__first != __last) - { - _ForwardIterator __i = __first; - while (++__i != __last) - if (__comp(*__i, *__first)) - __first = __i; - } - return __first; -} - -template -_LIBCPP_NODISCARD_EXT inline -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 -_ForwardIterator -min_element(_ForwardIterator __first, _ForwardIterator __last) -{ - return _VSTD::min_element(__first, __last, - __less::value_type>()); -} - -// min - -template -_LIBCPP_NODISCARD_EXT inline -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 -const _Tp& -min(const _Tp& __a, const _Tp& __b, _Compare __comp) -{ - return __comp(__b, __a) ? __b : __a; -} - -template -_LIBCPP_NODISCARD_EXT inline -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 -const _Tp& -min(const _Tp& __a, const _Tp& __b) -{ - return _VSTD::min(__a, __b, __less<_Tp>()); -} - -#ifndef _LIBCPP_CXX03_LANG - -template -_LIBCPP_NODISCARD_EXT inline -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 -_Tp -min(initializer_list<_Tp> __t, _Compare __comp) -{ - return *_VSTD::min_element(__t.begin(), __t.end(), __comp); -} - -template -_LIBCPP_NODISCARD_EXT inline -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 -_Tp -min(initializer_list<_Tp> __t) -{ - return *_VSTD::min_element(__t.begin(), __t.end(), __less<_Tp>()); -} - -#endif // _LIBCPP_CXX03_LANG - -// max_element - -template -_LIBCPP_NODISCARD_EXT inline -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 -_ForwardIterator -max_element(_ForwardIterator __first, _ForwardIterator __last, _Compare __comp) -{ - static_assert(__is_cpp17_forward_iterator<_ForwardIterator>::value, - "std::max_element requires a ForwardIterator"); - if (__first != __last) - { - _ForwardIterator __i = __first; - while (++__i != __last) - if (__comp(*__first, *__i)) - __first = __i; - } - return __first; -} - - -template -_LIBCPP_NODISCARD_EXT inline -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 -_ForwardIterator -max_element(_ForwardIterator __first, _ForwardIterator __last) -{ - return _VSTD::max_element(__first, __last, - __less::value_type>()); -} - -// max - -template -_LIBCPP_NODISCARD_EXT inline -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 -const _Tp& -max(const _Tp& __a, const _Tp& __b, _Compare __comp) -{ - return __comp(__a, __b) ? __b : __a; -} - -template -_LIBCPP_NODISCARD_EXT inline -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 -const _Tp& -max(const _Tp& __a, const _Tp& __b) -{ - return _VSTD::max(__a, __b, __less<_Tp>()); -} - -#ifndef _LIBCPP_CXX03_LANG - -template -_LIBCPP_NODISCARD_EXT inline -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 -_Tp -max(initializer_list<_Tp> __t, _Compare __comp) -{ - return *_VSTD::max_element(__t.begin(), __t.end(), __comp); -} - -template -_LIBCPP_NODISCARD_EXT inline -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 -_Tp -max(initializer_list<_Tp> __t) -{ - return *_VSTD::max_element(__t.begin(), __t.end(), __less<_Tp>()); -} - -#endif // _LIBCPP_CXX03_LANG - -#if _LIBCPP_STD_VER > 14 -// clamp -template -_LIBCPP_NODISCARD_EXT inline -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR -const _Tp& -clamp(const _Tp& __v, const _Tp& __lo, const _Tp& __hi, _Compare __comp) -{ - _LIBCPP_ASSERT(!__comp(__hi, __lo), "Bad bounds passed to std::clamp"); - return __comp(__v, __lo) ? __lo : __comp(__hi, __v) ? __hi : __v; - -} - -template -_LIBCPP_NODISCARD_EXT inline -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR -const _Tp& -clamp(const _Tp& __v, const _Tp& __lo, const _Tp& __hi) -{ - return _VSTD::clamp(__v, __lo, __hi, __less<_Tp>()); -} -#endif - -// minmax_element - -template -_LIBCPP_NODISCARD_EXT _LIBCPP_CONSTEXPR_AFTER_CXX11 -pair<_ForwardIterator, _ForwardIterator> -minmax_element(_ForwardIterator __first, _ForwardIterator __last, _Compare __comp) -{ - static_assert(__is_cpp17_forward_iterator<_ForwardIterator>::value, - "std::minmax_element requires a ForwardIterator"); - pair<_ForwardIterator, _ForwardIterator> __result(__first, __first); - if (__first != __last) - { - if (++__first != __last) - { - if (__comp(*__first, *__result.first)) - __result.first = __first; - else - __result.second = __first; - while (++__first != __last) - { - _ForwardIterator __i = __first; - if (++__first == __last) - { - if (__comp(*__i, *__result.first)) - __result.first = __i; - else if (!__comp(*__i, *__result.second)) - __result.second = __i; - break; - } - else - { - if (__comp(*__first, *__i)) - { - if (__comp(*__first, *__result.first)) - __result.first = __first; - if (!__comp(*__i, *__result.second)) - __result.second = __i; - } - else - { - if (__comp(*__i, *__result.first)) - __result.first = __i; - if (!__comp(*__first, *__result.second)) - __result.second = __first; - } - } - } - } - } - return __result; -} - -template -_LIBCPP_NODISCARD_EXT inline -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 -pair<_ForwardIterator, _ForwardIterator> -minmax_element(_ForwardIterator __first, _ForwardIterator __last) -{ - return _VSTD::minmax_element(__first, __last, - __less::value_type>()); -} - -// minmax - -template -_LIBCPP_NODISCARD_EXT inline -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 -pair -minmax(const _Tp& __a, const _Tp& __b, _Compare __comp) -{ - return __comp(__b, __a) ? pair(__b, __a) : - pair(__a, __b); -} - -template -_LIBCPP_NODISCARD_EXT inline -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 -pair -minmax(const _Tp& __a, const _Tp& __b) -{ - return _VSTD::minmax(__a, __b, __less<_Tp>()); -} - -#ifndef _LIBCPP_CXX03_LANG - -template -_LIBCPP_NODISCARD_EXT inline -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 -pair<_Tp, _Tp> -minmax(initializer_list<_Tp> __t, _Compare __comp) -{ - typedef typename initializer_list<_Tp>::const_iterator _Iter; - _Iter __first = __t.begin(); - _Iter __last = __t.end(); - pair<_Tp, _Tp> __result(*__first, *__first); - - ++__first; - if (__t.size() % 2 == 0) - { - if (__comp(*__first, __result.first)) - __result.first = *__first; - else - __result.second = *__first; - ++__first; - } - - while (__first != __last) - { - _Tp __prev = *__first++; - if (__comp(*__first, __prev)) { - if ( __comp(*__first, __result.first)) __result.first = *__first; - if (!__comp(__prev, __result.second)) __result.second = __prev; - } - else { - if ( __comp(__prev, __result.first)) __result.first = __prev; - if (!__comp(*__first, __result.second)) __result.second = *__first; - } - - __first++; - } - return __result; -} - -template -_LIBCPP_NODISCARD_EXT inline -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 -pair<_Tp, _Tp> -minmax(initializer_list<_Tp> __t) -{ - return _VSTD::minmax(__t, __less<_Tp>()); -} - -#endif // _LIBCPP_CXX03_LANG - -// random_shuffle - -// __independent_bits_engine - -template -struct __log2_imp -{ - static const size_t value = _Xp & ((unsigned long long)(1) << _Rp) ? _Rp - : __log2_imp<_Xp, _Rp - 1>::value; -}; - -template -struct __log2_imp<_Xp, 0> -{ - static const size_t value = 0; -}; - -template -struct __log2_imp<0, _Rp> -{ - static const size_t value = _Rp + 1; -}; - -template -struct __log2 -{ - static const size_t value = __log2_imp<_Xp, - sizeof(_UIntType) * __CHAR_BIT__ - 1>::value; -}; - -template -class __independent_bits_engine -{ -public: - // types - typedef _UIntType result_type; - -private: - typedef typename _Engine::result_type _Engine_result_type; - typedef typename conditional - < - sizeof(_Engine_result_type) <= sizeof(result_type), - result_type, - _Engine_result_type - >::type _Working_result_type; - - _Engine& __e_; - size_t __w_; - size_t __w0_; - size_t __n_; - size_t __n0_; - _Working_result_type __y0_; - _Working_result_type __y1_; - _Engine_result_type __mask0_; - _Engine_result_type __mask1_; - -#ifdef _LIBCPP_CXX03_LANG - static const _Working_result_type _Rp = _Engine::_Max - _Engine::_Min - + _Working_result_type(1); -#else - static _LIBCPP_CONSTEXPR const _Working_result_type _Rp = _Engine::max() - _Engine::min() - + _Working_result_type(1); -#endif - static _LIBCPP_CONSTEXPR const size_t __m = __log2<_Working_result_type, _Rp>::value; - static _LIBCPP_CONSTEXPR const size_t _WDt = numeric_limits<_Working_result_type>::digits; - static _LIBCPP_CONSTEXPR const size_t _EDt = numeric_limits<_Engine_result_type>::digits; - -public: - // constructors and seeding functions - __independent_bits_engine(_Engine& __e, size_t __w); - - // generating functions - result_type operator()() {return __eval(integral_constant());} - -private: - result_type __eval(false_type); - result_type __eval(true_type); -}; - -template -__independent_bits_engine<_Engine, _UIntType> - ::__independent_bits_engine(_Engine& __e, size_t __w) - : __e_(__e), - __w_(__w) -{ - __n_ = __w_ / __m + (__w_ % __m != 0); - __w0_ = __w_ / __n_; - if (_Rp == 0) - __y0_ = _Rp; - else if (__w0_ < _WDt) - __y0_ = (_Rp >> __w0_) << __w0_; - else - __y0_ = 0; - if (_Rp - __y0_ > __y0_ / __n_) - { - ++__n_; - __w0_ = __w_ / __n_; - if (__w0_ < _WDt) - __y0_ = (_Rp >> __w0_) << __w0_; - else - __y0_ = 0; - } - __n0_ = __n_ - __w_ % __n_; - if (__w0_ < _WDt - 1) - __y1_ = (_Rp >> (__w0_ + 1)) << (__w0_ + 1); - else - __y1_ = 0; - __mask0_ = __w0_ > 0 ? _Engine_result_type(~0) >> (_EDt - __w0_) : - _Engine_result_type(0); - __mask1_ = __w0_ < _EDt - 1 ? - _Engine_result_type(~0) >> (_EDt - (__w0_ + 1)) : - _Engine_result_type(~0); -} - -template -inline -_UIntType -__independent_bits_engine<_Engine, _UIntType>::__eval(false_type) -{ - return static_cast(__e_() & __mask0_); -} - -template -_UIntType -__independent_bits_engine<_Engine, _UIntType>::__eval(true_type) -{ - const size_t _WRt = numeric_limits::digits; - result_type _Sp = 0; - for (size_t __k = 0; __k < __n0_; ++__k) - { - _Engine_result_type __u; - do - { - __u = __e_() - _Engine::min(); - } while (__u >= __y0_); - if (__w0_ < _WRt) - _Sp <<= __w0_; - else - _Sp = 0; - _Sp += __u & __mask0_; - } - for (size_t __k = __n0_; __k < __n_; ++__k) - { - _Engine_result_type __u; - do - { - __u = __e_() - _Engine::min(); - } while (__u >= __y1_); - if (__w0_ < _WRt - 1) - _Sp <<= __w0_ + 1; - else - _Sp = 0; - _Sp += __u & __mask1_; - } - return _Sp; -} - -// uniform_int_distribution - -template -class uniform_int_distribution -{ -public: - // types - typedef _IntType result_type; - - class param_type - { - result_type __a_; - result_type __b_; - public: - typedef uniform_int_distribution distribution_type; - - explicit param_type(result_type __a = 0, - result_type __b = numeric_limits::max()) - : __a_(__a), __b_(__b) {} - - result_type a() const {return __a_;} - result_type b() const {return __b_;} - - friend bool operator==(const param_type& __x, const param_type& __y) - {return __x.__a_ == __y.__a_ && __x.__b_ == __y.__b_;} - friend bool operator!=(const param_type& __x, const param_type& __y) - {return !(__x == __y);} - }; - -private: - param_type __p_; - -public: - // constructors and reset functions -#ifndef _LIBCPP_CXX03_LANG - uniform_int_distribution() : uniform_int_distribution(0) {} - explicit uniform_int_distribution( - result_type __a, result_type __b = numeric_limits::max()) - : __p_(param_type(__a, __b)) {} -#else - explicit uniform_int_distribution( - result_type __a = 0, - result_type __b = numeric_limits::max()) - : __p_(param_type(__a, __b)) {} -#endif - explicit uniform_int_distribution(const param_type& __p) : __p_(__p) {} - void reset() {} - - // generating functions - template result_type operator()(_URNG& __g) - {return (*this)(__g, __p_);} - template result_type operator()(_URNG& __g, const param_type& __p); - - // property functions - result_type a() const {return __p_.a();} - result_type b() const {return __p_.b();} - - param_type param() const {return __p_;} - void param(const param_type& __p) {__p_ = __p;} - - result_type min() const {return a();} - result_type max() const {return b();} - - friend bool operator==(const uniform_int_distribution& __x, - const uniform_int_distribution& __y) - {return __x.__p_ == __y.__p_;} - friend bool operator!=(const uniform_int_distribution& __x, - const uniform_int_distribution& __y) - {return !(__x == __y);} -}; - -template -template -typename uniform_int_distribution<_IntType>::result_type -uniform_int_distribution<_IntType>::operator()(_URNG& __g, const param_type& __p) -_LIBCPP_DISABLE_UBSAN_UNSIGNED_INTEGER_CHECK -{ - typedef typename conditional::type _UIntType; - const _UIntType _Rp = _UIntType(__p.b()) - _UIntType(__p.a()) + _UIntType(1); - if (_Rp == 1) - return __p.a(); - const size_t _Dt = numeric_limits<_UIntType>::digits; - typedef __independent_bits_engine<_URNG, _UIntType> _Eng; - if (_Rp == 0) - return static_cast(_Eng(__g, _Dt)()); - size_t __w = _Dt - __libcpp_clz(_Rp) - 1; - if ((_Rp & (numeric_limits<_UIntType>::max() >> (_Dt - __w))) != 0) - ++__w; - _Eng __e(__g, __w); - _UIntType __u; - do - { - __u = __e(); - } while (__u >= _Rp); - return static_cast(__u + __p.a()); -} - -#if _LIBCPP_STD_VER <= 14 || defined(_LIBCPP_ENABLE_CXX17_REMOVED_RANDOM_SHUFFLE) \ - || defined(_LIBCPP_BUILDING_LIBRARY) -class _LIBCPP_TYPE_VIS __rs_default; - -_LIBCPP_FUNC_VIS __rs_default __rs_get(); - -class _LIBCPP_TYPE_VIS __rs_default -{ - static unsigned __c_; - - __rs_default(); -public: - typedef uint_fast32_t result_type; - - static const result_type _Min = 0; - static const result_type _Max = 0xFFFFFFFF; - - __rs_default(const __rs_default&); - ~__rs_default(); - - result_type operator()(); - - static _LIBCPP_CONSTEXPR result_type min() {return _Min;} - static _LIBCPP_CONSTEXPR result_type max() {return _Max;} - - friend _LIBCPP_FUNC_VIS __rs_default __rs_get(); -}; - -_LIBCPP_FUNC_VIS __rs_default __rs_get(); - -template -_LIBCPP_DEPRECATED_IN_CXX14 void -random_shuffle(_RandomAccessIterator __first, _RandomAccessIterator __last) -{ - typedef typename iterator_traits<_RandomAccessIterator>::difference_type difference_type; - typedef uniform_int_distribution _Dp; - typedef typename _Dp::param_type _Pp; - difference_type __d = __last - __first; - if (__d > 1) - { - _Dp __uid; - __rs_default __g = __rs_get(); - for (--__last, (void) --__d; __first < __last; ++__first, (void) --__d) - { - difference_type __i = __uid(__g, _Pp(0, __d)); - if (__i != difference_type(0)) - swap(*__first, *(__first + __i)); - } - } -} - -template -_LIBCPP_DEPRECATED_IN_CXX14 void -random_shuffle(_RandomAccessIterator __first, _RandomAccessIterator __last, -#ifndef _LIBCPP_CXX03_LANG - _RandomNumberGenerator&& __rand) -#else - _RandomNumberGenerator& __rand) -#endif -{ - typedef typename iterator_traits<_RandomAccessIterator>::difference_type difference_type; - difference_type __d = __last - __first; - if (__d > 1) - { - for (--__last; __first < __last; ++__first, (void) --__d) - { - difference_type __i = __rand(__d); - if (__i != difference_type(0)) - swap(*__first, *(__first + __i)); - } - } -} -#endif - -template -_LIBCPP_INLINE_VISIBILITY -_SampleIterator __sample(_PopulationIterator __first, - _PopulationIterator __last, _SampleIterator __output_iter, - _Distance __n, - _UniformRandomNumberGenerator & __g, - input_iterator_tag) { - - _Distance __k = 0; - for (; __first != __last && __k < __n; ++__first, (void) ++__k) - __output_iter[__k] = *__first; - _Distance __sz = __k; - for (; __first != __last; ++__first, (void) ++__k) { - _Distance __r = _VSTD::uniform_int_distribution<_Distance>(0, __k)(__g); - if (__r < __sz) - __output_iter[__r] = *__first; - } - return __output_iter + _VSTD::min(__n, __k); -} - -template -_LIBCPP_INLINE_VISIBILITY -_SampleIterator __sample(_PopulationIterator __first, - _PopulationIterator __last, _SampleIterator __output_iter, - _Distance __n, - _UniformRandomNumberGenerator& __g, - forward_iterator_tag) { - _Distance __unsampled_sz = _VSTD::distance(__first, __last); - for (__n = _VSTD::min(__n, __unsampled_sz); __n != 0; ++__first) { - _Distance __r = - _VSTD::uniform_int_distribution<_Distance>(0, --__unsampled_sz)(__g); - if (__r < __n) { - *__output_iter++ = *__first; - --__n; - } - } - return __output_iter; -} - -template -_LIBCPP_INLINE_VISIBILITY -_SampleIterator __sample(_PopulationIterator __first, - _PopulationIterator __last, _SampleIterator __output_iter, - _Distance __n, _UniformRandomNumberGenerator& __g) { - typedef typename iterator_traits<_PopulationIterator>::iterator_category - _PopCategory; - typedef typename iterator_traits<_PopulationIterator>::difference_type - _Difference; - static_assert(__is_cpp17_forward_iterator<_PopulationIterator>::value || - __is_cpp17_random_access_iterator<_SampleIterator>::value, - "SampleIterator must meet the requirements of RandomAccessIterator"); - typedef typename common_type<_Distance, _Difference>::type _CommonType; - _LIBCPP_ASSERT(__n >= 0, "N must be a positive number."); - return _VSTD::__sample( - __first, __last, __output_iter, _CommonType(__n), - __g, _PopCategory()); -} - -#if _LIBCPP_STD_VER > 14 -template -inline _LIBCPP_INLINE_VISIBILITY -_SampleIterator sample(_PopulationIterator __first, - _PopulationIterator __last, _SampleIterator __output_iter, - _Distance __n, _UniformRandomNumberGenerator&& __g) { - return _VSTD::__sample(__first, __last, __output_iter, __n, __g); -} -#endif // _LIBCPP_STD_VER > 14 - -template - void shuffle(_RandomAccessIterator __first, _RandomAccessIterator __last, - _UniformRandomNumberGenerator&& __g) -{ - typedef typename iterator_traits<_RandomAccessIterator>::difference_type difference_type; - typedef uniform_int_distribution _Dp; - typedef typename _Dp::param_type _Pp; - difference_type __d = __last - __first; - if (__d > 1) - { - _Dp __uid; - for (--__last, (void) --__d; __first < __last; ++__first, (void) --__d) - { - difference_type __i = __uid(__g, _Pp(0, __d)); - if (__i != difference_type(0)) - swap(*__first, *(__first + __i)); - } - } -} - -#if _LIBCPP_STD_VER > 17 - -// shift_left, shift_right - -template -inline _LIBCPP_INLINE_VISIBILITY constexpr -_ForwardIterator -shift_left(_ForwardIterator __first, _ForwardIterator __last, - typename iterator_traits<_ForwardIterator>::difference_type __n) -{ - if (__n == 0) { - return __last; - } - - _ForwardIterator __m = __first; - if constexpr (__is_cpp17_random_access_iterator<_ForwardIterator>::value) { - if (__n >= __last - __first) { - return __first; - } - __m += __n; - } else { - for (; __n > 0; --__n) { - if (__m == __last) { - return __first; - } - ++__m; - } - } - return _VSTD::move(__m, __last, __first); -} - -template -inline _LIBCPP_INLINE_VISIBILITY constexpr -_ForwardIterator -shift_right(_ForwardIterator __first, _ForwardIterator __last, - typename iterator_traits<_ForwardIterator>::difference_type __n) -{ - if (__n == 0) { - return __first; - } - - if constexpr (__is_cpp17_random_access_iterator<_ForwardIterator>::value) { - decltype(__n) __d = __last - __first; - if (__n >= __d) { - return __last; - } - _ForwardIterator __m = __first + (__d - __n); - return _VSTD::move_backward(__first, __m, __last); - } else if constexpr (__is_cpp17_bidirectional_iterator<_ForwardIterator>::value) { - _ForwardIterator __m = __last; - for (; __n > 0; --__n) { - if (__m == __first) { - return __last; - } - --__m; - } - return _VSTD::move_backward(__first, __m, __last); - } else { - _ForwardIterator __ret = __first; - for (; __n > 0; --__n) { - if (__ret == __last) { - return __last; - } - ++__ret; - } - - // We have an __n-element scratch space from __first to __ret. - // Slide an __n-element window [__trail, __lead) from left to right. - // We're essentially doing swap_ranges(__first, __ret, __trail, __lead) - // over and over; but once __lead reaches __last we needn't bother - // to save the values of elements [__trail, __last). - - auto __trail = __first; - auto __lead = __ret; - while (__trail != __ret) { - if (__lead == __last) { - _VSTD::move(__first, __trail, __ret); - return __ret; - } - ++__trail; - ++__lead; - } - - _ForwardIterator __mid = __first; - while (true) { - if (__lead == __last) { - __trail = _VSTD::move(__mid, __ret, __trail); - _VSTD::move(__first, __mid, __trail); - return __ret; - } - swap(*__mid, *__trail); - ++__mid; - ++__trail; - ++__lead; - if (__mid == __ret) { - __mid = __first; - } - } - } -} - -#endif // _LIBCPP_STD_VER > 17 - -// is_partitioned - -template -_LIBCPP_NODISCARD_EXT _LIBCPP_CONSTEXPR_AFTER_CXX17 bool -is_partitioned(_InputIterator __first, _InputIterator __last, _Predicate __pred) -{ - for (; __first != __last; ++__first) - if (!__pred(*__first)) - break; - if ( __first == __last ) - return true; - ++__first; - for (; __first != __last; ++__first) - if (__pred(*__first)) - return false; - return true; -} - -// partition - -template -_LIBCPP_CONSTEXPR_AFTER_CXX17 _ForwardIterator -__partition(_ForwardIterator __first, _ForwardIterator __last, _Predicate __pred, forward_iterator_tag) -{ - while (true) - { - if (__first == __last) - return __first; - if (!__pred(*__first)) - break; - ++__first; - } - for (_ForwardIterator __p = __first; ++__p != __last;) - { - if (__pred(*__p)) - { - swap(*__first, *__p); - ++__first; - } - } - return __first; -} - -template -_LIBCPP_CONSTEXPR_AFTER_CXX17 _BidirectionalIterator -__partition(_BidirectionalIterator __first, _BidirectionalIterator __last, _Predicate __pred, - bidirectional_iterator_tag) -{ - while (true) - { - while (true) - { - if (__first == __last) - return __first; - if (!__pred(*__first)) - break; - ++__first; - } - do - { - if (__first == --__last) - return __first; - } while (!__pred(*__last)); - swap(*__first, *__last); - ++__first; - } -} - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -_ForwardIterator -partition(_ForwardIterator __first, _ForwardIterator __last, _Predicate __pred) -{ - return _VSTD::__partition::type> - (__first, __last, __pred, typename iterator_traits<_ForwardIterator>::iterator_category()); -} - -// partition_copy - -template -_LIBCPP_CONSTEXPR_AFTER_CXX17 pair<_OutputIterator1, _OutputIterator2> -partition_copy(_InputIterator __first, _InputIterator __last, - _OutputIterator1 __out_true, _OutputIterator2 __out_false, - _Predicate __pred) -{ - for (; __first != __last; ++__first) - { - if (__pred(*__first)) - { - *__out_true = *__first; - ++__out_true; - } - else - { - *__out_false = *__first; - ++__out_false; - } - } - return pair<_OutputIterator1, _OutputIterator2>(__out_true, __out_false); -} - -// partition_point - -template -_LIBCPP_CONSTEXPR_AFTER_CXX17 _ForwardIterator -partition_point(_ForwardIterator __first, _ForwardIterator __last, _Predicate __pred) -{ - typedef typename iterator_traits<_ForwardIterator>::difference_type difference_type; - difference_type __len = _VSTD::distance(__first, __last); - while (__len != 0) - { - difference_type __l2 = _VSTD::__half_positive(__len); - _ForwardIterator __m = __first; - _VSTD::advance(__m, __l2); - if (__pred(*__m)) - { - __first = ++__m; - __len -= __l2 + 1; - } - else - __len = __l2; - } - return __first; -} - -// stable_partition - -template -_ForwardIterator -__stable_partition(_ForwardIterator __first, _ForwardIterator __last, _Predicate __pred, - _Distance __len, _Pair __p, forward_iterator_tag __fit) -{ - // *__first is known to be false - // __len >= 1 - if (__len == 1) - return __first; - if (__len == 2) - { - _ForwardIterator __m = __first; - if (__pred(*++__m)) - { - swap(*__first, *__m); - return __m; - } - return __first; - } - if (__len <= __p.second) - { // The buffer is big enough to use - typedef typename iterator_traits<_ForwardIterator>::value_type value_type; - __destruct_n __d(0); - unique_ptr __h(__p.first, __d); - // Move the falses into the temporary buffer, and the trues to the front of the line - // Update __first to always point to the end of the trues - value_type* __t = __p.first; - ::new ((void*)__t) value_type(_VSTD::move(*__first)); - __d.template __incr(); - ++__t; - _ForwardIterator __i = __first; - while (++__i != __last) - { - if (__pred(*__i)) - { - *__first = _VSTD::move(*__i); - ++__first; - } - else - { - ::new ((void*)__t) value_type(_VSTD::move(*__i)); - __d.template __incr(); - ++__t; - } - } - // All trues now at start of range, all falses in buffer - // Move falses back into range, but don't mess up __first which points to first false - __i = __first; - for (value_type* __t2 = __p.first; __t2 < __t; ++__t2, (void) ++__i) - *__i = _VSTD::move(*__t2); - // __h destructs moved-from values out of the temp buffer, but doesn't deallocate buffer - return __first; - } - // Else not enough buffer, do in place - // __len >= 3 - _ForwardIterator __m = __first; - _Distance __len2 = __len / 2; // __len2 >= 2 - _VSTD::advance(__m, __len2); - // recurse on [__first, __m), *__first know to be false - // F????????????????? - // f m l - typedef typename add_lvalue_reference<_Predicate>::type _PredRef; - _ForwardIterator __first_false = _VSTD::__stable_partition<_PredRef>(__first, __m, __pred, __len2, __p, __fit); - // TTTFFFFF?????????? - // f ff m l - // recurse on [__m, __last], except increase __m until *(__m) is false, *__last know to be true - _ForwardIterator __m1 = __m; - _ForwardIterator __second_false = __last; - _Distance __len_half = __len - __len2; - while (__pred(*__m1)) - { - if (++__m1 == __last) - goto __second_half_done; - --__len_half; - } - // TTTFFFFFTTTF?????? - // f ff m m1 l - __second_false = _VSTD::__stable_partition<_PredRef>(__m1, __last, __pred, __len_half, __p, __fit); -__second_half_done: - // TTTFFFFFTTTTTFFFFF - // f ff m sf l - return _VSTD::rotate(__first_false, __m, __second_false); - // TTTTTTTTFFFFFFFFFF - // | -} - -struct __return_temporary_buffer -{ - template - _LIBCPP_INLINE_VISIBILITY void operator()(_Tp* __p) const {_VSTD::return_temporary_buffer(__p);} -}; - -template -_ForwardIterator -__stable_partition(_ForwardIterator __first, _ForwardIterator __last, _Predicate __pred, - forward_iterator_tag) -{ - const unsigned __alloc_limit = 3; // might want to make this a function of trivial assignment - // Either prove all true and return __first or point to first false - while (true) - { - if (__first == __last) - return __first; - if (!__pred(*__first)) - break; - ++__first; - } - // We now have a reduced range [__first, __last) - // *__first is known to be false - typedef typename iterator_traits<_ForwardIterator>::difference_type difference_type; - typedef typename iterator_traits<_ForwardIterator>::value_type value_type; - difference_type __len = _VSTD::distance(__first, __last); - pair __p(0, 0); - unique_ptr __h; - if (__len >= __alloc_limit) - { - __p = _VSTD::get_temporary_buffer(__len); - __h.reset(__p.first); - } - return _VSTD::__stable_partition::type> - (__first, __last, __pred, __len, __p, forward_iterator_tag()); -} - -template -_BidirectionalIterator -__stable_partition(_BidirectionalIterator __first, _BidirectionalIterator __last, _Predicate __pred, - _Distance __len, _Pair __p, bidirectional_iterator_tag __bit) -{ - // *__first is known to be false - // *__last is known to be true - // __len >= 2 - if (__len == 2) - { - swap(*__first, *__last); - return __last; - } - if (__len == 3) - { - _BidirectionalIterator __m = __first; - if (__pred(*++__m)) - { - swap(*__first, *__m); - swap(*__m, *__last); - return __last; - } - swap(*__m, *__last); - swap(*__first, *__m); - return __m; - } - if (__len <= __p.second) - { // The buffer is big enough to use - typedef typename iterator_traits<_BidirectionalIterator>::value_type value_type; - __destruct_n __d(0); - unique_ptr __h(__p.first, __d); - // Move the falses into the temporary buffer, and the trues to the front of the line - // Update __first to always point to the end of the trues - value_type* __t = __p.first; - ::new ((void*)__t) value_type(_VSTD::move(*__first)); - __d.template __incr(); - ++__t; - _BidirectionalIterator __i = __first; - while (++__i != __last) - { - if (__pred(*__i)) - { - *__first = _VSTD::move(*__i); - ++__first; - } - else - { - ::new ((void*)__t) value_type(_VSTD::move(*__i)); - __d.template __incr(); - ++__t; - } - } - // move *__last, known to be true - *__first = _VSTD::move(*__i); - __i = ++__first; - // All trues now at start of range, all falses in buffer - // Move falses back into range, but don't mess up __first which points to first false - for (value_type* __t2 = __p.first; __t2 < __t; ++__t2, (void) ++__i) - *__i = _VSTD::move(*__t2); - // __h destructs moved-from values out of the temp buffer, but doesn't deallocate buffer - return __first; - } - // Else not enough buffer, do in place - // __len >= 4 - _BidirectionalIterator __m = __first; - _Distance __len2 = __len / 2; // __len2 >= 2 - _VSTD::advance(__m, __len2); - // recurse on [__first, __m-1], except reduce __m-1 until *(__m-1) is true, *__first know to be false - // F????????????????T - // f m l - _BidirectionalIterator __m1 = __m; - _BidirectionalIterator __first_false = __first; - _Distance __len_half = __len2; - while (!__pred(*--__m1)) - { - if (__m1 == __first) - goto __first_half_done; - --__len_half; - } - // F???TFFF?????????T - // f m1 m l - typedef typename add_lvalue_reference<_Predicate>::type _PredRef; - __first_false = _VSTD::__stable_partition<_PredRef>(__first, __m1, __pred, __len_half, __p, __bit); -__first_half_done: - // TTTFFFFF?????????T - // f ff m l - // recurse on [__m, __last], except increase __m until *(__m) is false, *__last know to be true - __m1 = __m; - _BidirectionalIterator __second_false = __last; - ++__second_false; - __len_half = __len - __len2; - while (__pred(*__m1)) - { - if (++__m1 == __last) - goto __second_half_done; - --__len_half; - } - // TTTFFFFFTTTF?????T - // f ff m m1 l - __second_false = _VSTD::__stable_partition<_PredRef>(__m1, __last, __pred, __len_half, __p, __bit); -__second_half_done: - // TTTFFFFFTTTTTFFFFF - // f ff m sf l - return _VSTD::rotate(__first_false, __m, __second_false); - // TTTTTTTTFFFFFFFFFF - // | -} - -template -_BidirectionalIterator -__stable_partition(_BidirectionalIterator __first, _BidirectionalIterator __last, _Predicate __pred, - bidirectional_iterator_tag) -{ - typedef typename iterator_traits<_BidirectionalIterator>::difference_type difference_type; - typedef typename iterator_traits<_BidirectionalIterator>::value_type value_type; - const difference_type __alloc_limit = 4; // might want to make this a function of trivial assignment - // Either prove all true and return __first or point to first false - while (true) - { - if (__first == __last) - return __first; - if (!__pred(*__first)) - break; - ++__first; - } - // __first points to first false, everything prior to __first is already set. - // Either prove [__first, __last) is all false and return __first, or point __last to last true - do - { - if (__first == --__last) - return __first; - } while (!__pred(*__last)); - // We now have a reduced range [__first, __last] - // *__first is known to be false - // *__last is known to be true - // __len >= 2 - difference_type __len = _VSTD::distance(__first, __last) + 1; - pair __p(0, 0); - unique_ptr __h; - if (__len >= __alloc_limit) - { - __p = _VSTD::get_temporary_buffer(__len); - __h.reset(__p.first); - } - return _VSTD::__stable_partition::type> - (__first, __last, __pred, __len, __p, bidirectional_iterator_tag()); -} - -template -inline _LIBCPP_INLINE_VISIBILITY -_ForwardIterator -stable_partition(_ForwardIterator __first, _ForwardIterator __last, _Predicate __pred) -{ - return _VSTD::__stable_partition::type> - (__first, __last, __pred, typename iterator_traits<_ForwardIterator>::iterator_category()); -} - -// is_sorted_until - -template -_LIBCPP_NODISCARD_EXT _LIBCPP_CONSTEXPR_AFTER_CXX17 _ForwardIterator -is_sorted_until(_ForwardIterator __first, _ForwardIterator __last, _Compare __comp) -{ - if (__first != __last) - { - _ForwardIterator __i = __first; - while (++__i != __last) - { - if (__comp(*__i, *__first)) - return __i; - __first = __i; - } - } - return __last; -} - -template -_LIBCPP_NODISCARD_EXT inline -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -_ForwardIterator -is_sorted_until(_ForwardIterator __first, _ForwardIterator __last) -{ - return _VSTD::is_sorted_until(__first, __last, __less::value_type>()); -} - -// is_sorted - -template -_LIBCPP_NODISCARD_EXT inline -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -bool -is_sorted(_ForwardIterator __first, _ForwardIterator __last, _Compare __comp) -{ - return _VSTD::is_sorted_until(__first, __last, __comp) == __last; -} - -template -_LIBCPP_NODISCARD_EXT inline -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -bool -is_sorted(_ForwardIterator __first, _ForwardIterator __last) -{ - return _VSTD::is_sorted(__first, __last, __less::value_type>()); -} - -// sort - -// stable, 2-3 compares, 0-2 swaps - -template -unsigned -__sort3(_ForwardIterator __x, _ForwardIterator __y, _ForwardIterator __z, _Compare __c) -{ - unsigned __r = 0; - if (!__c(*__y, *__x)) // if x <= y - { - if (!__c(*__z, *__y)) // if y <= z - return __r; // x <= y && y <= z - // x <= y && y > z - swap(*__y, *__z); // x <= z && y < z - __r = 1; - if (__c(*__y, *__x)) // if x > y - { - swap(*__x, *__y); // x < y && y <= z - __r = 2; - } - return __r; // x <= y && y < z - } - if (__c(*__z, *__y)) // x > y, if y > z - { - swap(*__x, *__z); // x < y && y < z - __r = 1; - return __r; - } - swap(*__x, *__y); // x > y && y <= z - __r = 1; // x < y && x <= z - if (__c(*__z, *__y)) // if y > z - { - swap(*__y, *__z); // x <= y && y < z - __r = 2; - } - return __r; -} // x <= y && y <= z - -// stable, 3-6 compares, 0-5 swaps - -template -unsigned -__sort4(_ForwardIterator __x1, _ForwardIterator __x2, _ForwardIterator __x3, - _ForwardIterator __x4, _Compare __c) -{ - unsigned __r = _VSTD::__sort3<_Compare>(__x1, __x2, __x3, __c); - if (__c(*__x4, *__x3)) - { - swap(*__x3, *__x4); - ++__r; - if (__c(*__x3, *__x2)) - { - swap(*__x2, *__x3); - ++__r; - if (__c(*__x2, *__x1)) - { - swap(*__x1, *__x2); - ++__r; - } - } - } - return __r; -} - -// stable, 4-10 compares, 0-9 swaps - -template -_LIBCPP_HIDDEN -unsigned -__sort5(_ForwardIterator __x1, _ForwardIterator __x2, _ForwardIterator __x3, - _ForwardIterator __x4, _ForwardIterator __x5, _Compare __c) -{ - unsigned __r = _VSTD::__sort4<_Compare>(__x1, __x2, __x3, __x4, __c); - if (__c(*__x5, *__x4)) - { - swap(*__x4, *__x5); - ++__r; - if (__c(*__x4, *__x3)) - { - swap(*__x3, *__x4); - ++__r; - if (__c(*__x3, *__x2)) - { - swap(*__x2, *__x3); - ++__r; - if (__c(*__x2, *__x1)) - { - swap(*__x1, *__x2); - ++__r; - } - } - } - } - return __r; -} - -// Assumes size > 0 -template -void -__selection_sort(_BidirectionalIterator __first, _BidirectionalIterator __last, _Compare __comp) -{ - _BidirectionalIterator __lm1 = __last; - for (--__lm1; __first != __lm1; ++__first) - { - _BidirectionalIterator __i = _VSTD::min_element<_BidirectionalIterator, - typename add_lvalue_reference<_Compare>::type> - (__first, __last, __comp); - if (__i != __first) - swap(*__first, *__i); - } -} - -template -void -__insertion_sort(_BidirectionalIterator __first, _BidirectionalIterator __last, _Compare __comp) -{ - typedef typename iterator_traits<_BidirectionalIterator>::value_type value_type; - if (__first != __last) - { - _BidirectionalIterator __i = __first; - for (++__i; __i != __last; ++__i) - { - _BidirectionalIterator __j = __i; - value_type __t(_VSTD::move(*__j)); - for (_BidirectionalIterator __k = __i; __k != __first && __comp(__t, *--__k); --__j) - *__j = _VSTD::move(*__k); - *__j = _VSTD::move(__t); - } - } -} - -template -void -__insertion_sort_3(_RandomAccessIterator __first, _RandomAccessIterator __last, _Compare __comp) -{ - typedef typename iterator_traits<_RandomAccessIterator>::value_type value_type; - _RandomAccessIterator __j = __first+2; - _VSTD::__sort3<_Compare>(__first, __first+1, __j, __comp); - for (_RandomAccessIterator __i = __j+1; __i != __last; ++__i) - { - if (__comp(*__i, *__j)) - { - value_type __t(_VSTD::move(*__i)); - _RandomAccessIterator __k = __j; - __j = __i; - do - { - *__j = _VSTD::move(*__k); - __j = __k; - } while (__j != __first && __comp(__t, *--__k)); - *__j = _VSTD::move(__t); - } - __j = __i; - } -} - -template -bool -__insertion_sort_incomplete(_RandomAccessIterator __first, _RandomAccessIterator __last, _Compare __comp) -{ - switch (__last - __first) - { - case 0: - case 1: - return true; - case 2: - if (__comp(*--__last, *__first)) - swap(*__first, *__last); - return true; - case 3: - _VSTD::__sort3<_Compare>(__first, __first+1, --__last, __comp); - return true; - case 4: - _VSTD::__sort4<_Compare>(__first, __first+1, __first+2, --__last, __comp); - return true; - case 5: - _VSTD::__sort5<_Compare>(__first, __first+1, __first+2, __first+3, --__last, __comp); - return true; - } - typedef typename iterator_traits<_RandomAccessIterator>::value_type value_type; - _RandomAccessIterator __j = __first+2; - _VSTD::__sort3<_Compare>(__first, __first+1, __j, __comp); - const unsigned __limit = 8; - unsigned __count = 0; - for (_RandomAccessIterator __i = __j+1; __i != __last; ++__i) - { - if (__comp(*__i, *__j)) - { - value_type __t(_VSTD::move(*__i)); - _RandomAccessIterator __k = __j; - __j = __i; - do - { - *__j = _VSTD::move(*__k); - __j = __k; - } while (__j != __first && __comp(__t, *--__k)); - *__j = _VSTD::move(__t); - if (++__count == __limit) - return ++__i == __last; - } - __j = __i; - } - return true; -} - -template -void -__insertion_sort_move(_BidirectionalIterator __first1, _BidirectionalIterator __last1, - typename iterator_traits<_BidirectionalIterator>::value_type* __first2, _Compare __comp) -{ - typedef typename iterator_traits<_BidirectionalIterator>::value_type value_type; - if (__first1 != __last1) - { - __destruct_n __d(0); - unique_ptr __h(__first2, __d); - value_type* __last2 = __first2; - ::new ((void*)__last2) value_type(_VSTD::move(*__first1)); - __d.template __incr(); - for (++__last2; ++__first1 != __last1; ++__last2) - { - value_type* __j2 = __last2; - value_type* __i2 = __j2; - if (__comp(*__first1, *--__i2)) - { - ::new ((void*)__j2) value_type(_VSTD::move(*__i2)); - __d.template __incr(); - for (--__j2; __i2 != __first2 && __comp(*__first1, *--__i2); --__j2) - *__j2 = _VSTD::move(*__i2); - *__j2 = _VSTD::move(*__first1); - } - else - { - ::new ((void*)__j2) value_type(_VSTD::move(*__first1)); - __d.template __incr(); - } - } - __h.release(); - } -} - -template -void -__sort(_RandomAccessIterator __first, _RandomAccessIterator __last, _Compare __comp) -{ - // _Compare is known to be a reference type - typedef typename iterator_traits<_RandomAccessIterator>::difference_type difference_type; - typedef typename iterator_traits<_RandomAccessIterator>::value_type value_type; - const difference_type __limit = is_trivially_copy_constructible::value && - is_trivially_copy_assignable::value ? 30 : 6; - while (true) - { - __restart: - difference_type __len = __last - __first; - switch (__len) - { - case 0: - case 1: - return; - case 2: - if (__comp(*--__last, *__first)) - swap(*__first, *__last); - return; - case 3: - _VSTD::__sort3<_Compare>(__first, __first+1, --__last, __comp); - return; - case 4: - _VSTD::__sort4<_Compare>(__first, __first+1, __first+2, --__last, __comp); - return; - case 5: - _VSTD::__sort5<_Compare>(__first, __first+1, __first+2, __first+3, --__last, __comp); - return; - } - if (__len <= __limit) - { - _VSTD::__insertion_sort_3<_Compare>(__first, __last, __comp); - return; - } - // __len > 5 - _RandomAccessIterator __m = __first; - _RandomAccessIterator __lm1 = __last; - --__lm1; - unsigned __n_swaps; - { - difference_type __delta; - if (__len >= 1000) - { - __delta = __len/2; - __m += __delta; - __delta /= 2; - __n_swaps = _VSTD::__sort5<_Compare>(__first, __first + __delta, __m, __m+__delta, __lm1, __comp); - } - else - { - __delta = __len/2; - __m += __delta; - __n_swaps = _VSTD::__sort3<_Compare>(__first, __m, __lm1, __comp); - } - } - // *__m is median - // partition [__first, __m) < *__m and *__m <= [__m, __last) - // (this inhibits tossing elements equivalent to __m around unnecessarily) - _RandomAccessIterator __i = __first; - _RandomAccessIterator __j = __lm1; - // j points beyond range to be tested, *__m is known to be <= *__lm1 - // The search going up is known to be guarded but the search coming down isn't. - // Prime the downward search with a guard. - if (!__comp(*__i, *__m)) // if *__first == *__m - { - // *__first == *__m, *__first doesn't go in first part - // manually guard downward moving __j against __i - while (true) - { - if (__i == --__j) - { - // *__first == *__m, *__m <= all other elements - // Parition instead into [__first, __i) == *__first and *__first < [__i, __last) - ++__i; // __first + 1 - __j = __last; - if (!__comp(*__first, *--__j)) // we need a guard if *__first == *(__last-1) - { - while (true) - { - if (__i == __j) - return; // [__first, __last) all equivalent elements - if (__comp(*__first, *__i)) - { - swap(*__i, *__j); - ++__n_swaps; - ++__i; - break; - } - ++__i; - } - } - // [__first, __i) == *__first and *__first < [__j, __last) and __j == __last - 1 - if (__i == __j) - return; - while (true) - { - while (!__comp(*__first, *__i)) - ++__i; - while (__comp(*__first, *--__j)) - ; - if (__i >= __j) - break; - swap(*__i, *__j); - ++__n_swaps; - ++__i; - } - // [__first, __i) == *__first and *__first < [__i, __last) - // The first part is sorted, sort the second part - // _VSTD::__sort<_Compare>(__i, __last, __comp); - __first = __i; - goto __restart; - } - if (__comp(*__j, *__m)) - { - swap(*__i, *__j); - ++__n_swaps; - break; // found guard for downward moving __j, now use unguarded partition - } - } - } - // It is known that *__i < *__m - ++__i; - // j points beyond range to be tested, *__m is known to be <= *__lm1 - // if not yet partitioned... - if (__i < __j) - { - // known that *(__i - 1) < *__m - // known that __i <= __m - while (true) - { - // __m still guards upward moving __i - while (__comp(*__i, *__m)) - ++__i; - // It is now known that a guard exists for downward moving __j - while (!__comp(*--__j, *__m)) - ; - if (__i > __j) - break; - swap(*__i, *__j); - ++__n_swaps; - // It is known that __m != __j - // If __m just moved, follow it - if (__m == __i) - __m = __j; - ++__i; - } - } - // [__first, __i) < *__m and *__m <= [__i, __last) - if (__i != __m && __comp(*__m, *__i)) - { - swap(*__i, *__m); - ++__n_swaps; - } - // [__first, __i) < *__i and *__i <= [__i+1, __last) - // If we were given a perfect partition, see if insertion sort is quick... - if (__n_swaps == 0) - { - bool __fs = _VSTD::__insertion_sort_incomplete<_Compare>(__first, __i, __comp); - if (_VSTD::__insertion_sort_incomplete<_Compare>(__i+1, __last, __comp)) - { - if (__fs) - return; - __last = __i; - continue; - } - else - { - if (__fs) - { - __first = ++__i; - continue; - } - } - } - // sort smaller range with recursive call and larger with tail recursion elimination - if (__i - __first < __last - __i) - { - _VSTD::__sort<_Compare>(__first, __i, __comp); - // _VSTD::__sort<_Compare>(__i+1, __last, __comp); - __first = ++__i; - } - else - { - _VSTD::__sort<_Compare>(__i+1, __last, __comp); - // _VSTD::__sort<_Compare>(__first, __i, __comp); - __last = __i; - } - } -} - -// This forwarder keeps the top call and the recursive calls using the same instantiation, forcing a reference _Compare -template -inline _LIBCPP_INLINE_VISIBILITY -void -sort(_RandomAccessIterator __first, _RandomAccessIterator __last, _Compare __comp) -{ - typedef typename __comp_ref_type<_Compare>::type _Comp_ref; - _VSTD::__sort<_Comp_ref>(__first, __last, _Comp_ref(__comp)); -} - -template -inline _LIBCPP_INLINE_VISIBILITY -void -sort(_RandomAccessIterator __first, _RandomAccessIterator __last) -{ - _VSTD::sort(__first, __last, __less::value_type>()); -} - -template -inline _LIBCPP_INLINE_VISIBILITY -void -sort(_Tp** __first, _Tp** __last) -{ - _VSTD::sort((uintptr_t*)__first, (uintptr_t*)__last, __less()); -} - -template -inline _LIBCPP_INLINE_VISIBILITY -void -sort(__wrap_iter<_Tp*> __first, __wrap_iter<_Tp*> __last) -{ - _VSTD::sort(__first.base(), __last.base()); -} - -template -inline _LIBCPP_INLINE_VISIBILITY -void -sort(__wrap_iter<_Tp*> __first, __wrap_iter<_Tp*> __last, _Compare __comp) -{ - typedef typename add_lvalue_reference<_Compare>::type _Comp_ref; - _VSTD::sort<_Tp*, _Comp_ref>(__first.base(), __last.base(), __comp); -} - -_LIBCPP_EXTERN_TEMPLATE(_LIBCPP_FUNC_VIS void __sort<__less&, char*>(char*, char*, __less&)) -_LIBCPP_EXTERN_TEMPLATE(_LIBCPP_FUNC_VIS void __sort<__less&, wchar_t*>(wchar_t*, wchar_t*, __less&)) -_LIBCPP_EXTERN_TEMPLATE(_LIBCPP_FUNC_VIS void __sort<__less&, signed char*>(signed char*, signed char*, __less&)) -_LIBCPP_EXTERN_TEMPLATE(_LIBCPP_FUNC_VIS void __sort<__less&, unsigned char*>(unsigned char*, unsigned char*, __less&)) -_LIBCPP_EXTERN_TEMPLATE(_LIBCPP_FUNC_VIS void __sort<__less&, short*>(short*, short*, __less&)) -_LIBCPP_EXTERN_TEMPLATE(_LIBCPP_FUNC_VIS void __sort<__less&, unsigned short*>(unsigned short*, unsigned short*, __less&)) -_LIBCPP_EXTERN_TEMPLATE(_LIBCPP_FUNC_VIS void __sort<__less&, int*>(int*, int*, __less&)) -_LIBCPP_EXTERN_TEMPLATE(_LIBCPP_FUNC_VIS void __sort<__less&, unsigned*>(unsigned*, unsigned*, __less&)) -_LIBCPP_EXTERN_TEMPLATE(_LIBCPP_FUNC_VIS void __sort<__less&, long*>(long*, long*, __less&)) -_LIBCPP_EXTERN_TEMPLATE(_LIBCPP_FUNC_VIS void __sort<__less&, unsigned long*>(unsigned long*, unsigned long*, __less&)) -_LIBCPP_EXTERN_TEMPLATE(_LIBCPP_FUNC_VIS void __sort<__less&, long long*>(long long*, long long*, __less&)) -_LIBCPP_EXTERN_TEMPLATE(_LIBCPP_FUNC_VIS void __sort<__less&, unsigned long long*>(unsigned long long*, unsigned long long*, __less&)) -_LIBCPP_EXTERN_TEMPLATE(_LIBCPP_FUNC_VIS void __sort<__less&, float*>(float*, float*, __less&)) -_LIBCPP_EXTERN_TEMPLATE(_LIBCPP_FUNC_VIS void __sort<__less&, double*>(double*, double*, __less&)) -_LIBCPP_EXTERN_TEMPLATE(_LIBCPP_FUNC_VIS void __sort<__less&, long double*>(long double*, long double*, __less&)) - -_LIBCPP_EXTERN_TEMPLATE(_LIBCPP_FUNC_VIS bool __insertion_sort_incomplete<__less&, char*>(char*, char*, __less&)) -_LIBCPP_EXTERN_TEMPLATE(_LIBCPP_FUNC_VIS bool __insertion_sort_incomplete<__less&, wchar_t*>(wchar_t*, wchar_t*, __less&)) -_LIBCPP_EXTERN_TEMPLATE(_LIBCPP_FUNC_VIS bool __insertion_sort_incomplete<__less&, signed char*>(signed char*, signed char*, __less&)) -_LIBCPP_EXTERN_TEMPLATE(_LIBCPP_FUNC_VIS bool __insertion_sort_incomplete<__less&, unsigned char*>(unsigned char*, unsigned char*, __less&)) -_LIBCPP_EXTERN_TEMPLATE(_LIBCPP_FUNC_VIS bool __insertion_sort_incomplete<__less&, short*>(short*, short*, __less&)) -_LIBCPP_EXTERN_TEMPLATE(_LIBCPP_FUNC_VIS bool __insertion_sort_incomplete<__less&, unsigned short*>(unsigned short*, unsigned short*, __less&)) -_LIBCPP_EXTERN_TEMPLATE(_LIBCPP_FUNC_VIS bool __insertion_sort_incomplete<__less&, int*>(int*, int*, __less&)) -_LIBCPP_EXTERN_TEMPLATE(_LIBCPP_FUNC_VIS bool __insertion_sort_incomplete<__less&, unsigned*>(unsigned*, unsigned*, __less&)) -_LIBCPP_EXTERN_TEMPLATE(_LIBCPP_FUNC_VIS bool __insertion_sort_incomplete<__less&, long*>(long*, long*, __less&)) -_LIBCPP_EXTERN_TEMPLATE(_LIBCPP_FUNC_VIS bool __insertion_sort_incomplete<__less&, unsigned long*>(unsigned long*, unsigned long*, __less&)) -_LIBCPP_EXTERN_TEMPLATE(_LIBCPP_FUNC_VIS bool __insertion_sort_incomplete<__less&, long long*>(long long*, long long*, __less&)) -_LIBCPP_EXTERN_TEMPLATE(_LIBCPP_FUNC_VIS bool __insertion_sort_incomplete<__less&, unsigned long long*>(unsigned long long*, unsigned long long*, __less&)) -_LIBCPP_EXTERN_TEMPLATE(_LIBCPP_FUNC_VIS bool __insertion_sort_incomplete<__less&, float*>(float*, float*, __less&)) -_LIBCPP_EXTERN_TEMPLATE(_LIBCPP_FUNC_VIS bool __insertion_sort_incomplete<__less&, double*>(double*, double*, __less&)) -_LIBCPP_EXTERN_TEMPLATE(_LIBCPP_FUNC_VIS bool __insertion_sort_incomplete<__less&, long double*>(long double*, long double*, __less&)) - -_LIBCPP_EXTERN_TEMPLATE(_LIBCPP_FUNC_VIS unsigned __sort5<__less&, long double*>(long double*, long double*, long double*, long double*, long double*, __less&)) - -// lower_bound - -template -_LIBCPP_CONSTEXPR_AFTER_CXX17 _ForwardIterator -__lower_bound(_ForwardIterator __first, _ForwardIterator __last, const _Tp& __value_, _Compare __comp) -{ - typedef typename iterator_traits<_ForwardIterator>::difference_type difference_type; - difference_type __len = _VSTD::distance(__first, __last); - while (__len != 0) - { - difference_type __l2 = _VSTD::__half_positive(__len); - _ForwardIterator __m = __first; - _VSTD::advance(__m, __l2); - if (__comp(*__m, __value_)) - { - __first = ++__m; - __len -= __l2 + 1; - } - else - __len = __l2; - } - return __first; -} - -template -_LIBCPP_NODISCARD_EXT inline -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -_ForwardIterator -lower_bound(_ForwardIterator __first, _ForwardIterator __last, const _Tp& __value_, _Compare __comp) -{ - typedef typename add_lvalue_reference<_Compare>::type _Comp_ref; - return _VSTD::__lower_bound<_Comp_ref>(__first, __last, __value_, __comp); -} - -template -_LIBCPP_NODISCARD_EXT inline -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -_ForwardIterator -lower_bound(_ForwardIterator __first, _ForwardIterator __last, const _Tp& __value_) -{ - return _VSTD::lower_bound(__first, __last, __value_, - __less::value_type, _Tp>()); -} - -// upper_bound - -template -_LIBCPP_CONSTEXPR_AFTER_CXX17 _ForwardIterator -__upper_bound(_ForwardIterator __first, _ForwardIterator __last, const _Tp& __value_, _Compare __comp) -{ - typedef typename iterator_traits<_ForwardIterator>::difference_type difference_type; - difference_type __len = _VSTD::distance(__first, __last); - while (__len != 0) - { - difference_type __l2 = _VSTD::__half_positive(__len); - _ForwardIterator __m = __first; - _VSTD::advance(__m, __l2); - if (__comp(__value_, *__m)) - __len = __l2; - else - { - __first = ++__m; - __len -= __l2 + 1; - } - } - return __first; -} - -template -_LIBCPP_NODISCARD_EXT inline -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -_ForwardIterator -upper_bound(_ForwardIterator __first, _ForwardIterator __last, const _Tp& __value_, _Compare __comp) -{ - typedef typename add_lvalue_reference<_Compare>::type _Comp_ref; - return _VSTD::__upper_bound<_Comp_ref>(__first, __last, __value_, __comp); -} - -template -_LIBCPP_NODISCARD_EXT inline -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -_ForwardIterator -upper_bound(_ForwardIterator __first, _ForwardIterator __last, const _Tp& __value_) -{ - return _VSTD::upper_bound(__first, __last, __value_, - __less<_Tp, typename iterator_traits<_ForwardIterator>::value_type>()); -} - -// equal_range - -template -_LIBCPP_CONSTEXPR_AFTER_CXX17 pair<_ForwardIterator, _ForwardIterator> -__equal_range(_ForwardIterator __first, _ForwardIterator __last, const _Tp& __value_, _Compare __comp) -{ - typedef typename iterator_traits<_ForwardIterator>::difference_type difference_type; - difference_type __len = _VSTD::distance(__first, __last); - while (__len != 0) - { - difference_type __l2 = _VSTD::__half_positive(__len); - _ForwardIterator __m = __first; - _VSTD::advance(__m, __l2); - if (__comp(*__m, __value_)) - { - __first = ++__m; - __len -= __l2 + 1; - } - else if (__comp(__value_, *__m)) - { - __last = __m; - __len = __l2; - } - else - { - _ForwardIterator __mp1 = __m; - return pair<_ForwardIterator, _ForwardIterator> - ( - _VSTD::__lower_bound<_Compare>(__first, __m, __value_, __comp), - _VSTD::__upper_bound<_Compare>(++__mp1, __last, __value_, __comp) - ); - } - } - return pair<_ForwardIterator, _ForwardIterator>(__first, __first); -} - -template -_LIBCPP_NODISCARD_EXT inline -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -pair<_ForwardIterator, _ForwardIterator> -equal_range(_ForwardIterator __first, _ForwardIterator __last, const _Tp& __value_, _Compare __comp) -{ - typedef typename __comp_ref_type<_Compare>::type _Comp_ref; - return _VSTD::__equal_range<_Comp_ref>(__first, __last, __value_, __comp); -} - -template -_LIBCPP_NODISCARD_EXT inline -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -pair<_ForwardIterator, _ForwardIterator> -equal_range(_ForwardIterator __first, _ForwardIterator __last, const _Tp& __value_) -{ - return _VSTD::equal_range(__first, __last, __value_, - __less::value_type, _Tp>()); -} - -// binary_search - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -bool -__binary_search(_ForwardIterator __first, _ForwardIterator __last, const _Tp& __value_, _Compare __comp) -{ - __first = _VSTD::__lower_bound<_Compare>(__first, __last, __value_, __comp); - return __first != __last && !__comp(__value_, *__first); -} - -template -_LIBCPP_NODISCARD_EXT inline -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -bool -binary_search(_ForwardIterator __first, _ForwardIterator __last, const _Tp& __value_, _Compare __comp) -{ - typedef typename __comp_ref_type<_Compare>::type _Comp_ref; - return _VSTD::__binary_search<_Comp_ref>(__first, __last, __value_, __comp); -} - -template -_LIBCPP_NODISCARD_EXT inline -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -bool -binary_search(_ForwardIterator __first, _ForwardIterator __last, const _Tp& __value_) -{ - return _VSTD::binary_search(__first, __last, __value_, - __less::value_type, _Tp>()); -} - -// merge - -template -_LIBCPP_CONSTEXPR_AFTER_CXX17 -_OutputIterator -__merge(_InputIterator1 __first1, _InputIterator1 __last1, - _InputIterator2 __first2, _InputIterator2 __last2, _OutputIterator __result, _Compare __comp) -{ - for (; __first1 != __last1; ++__result) - { - if (__first2 == __last2) - return _VSTD::copy(__first1, __last1, __result); - if (__comp(*__first2, *__first1)) - { - *__result = *__first2; - ++__first2; - } - else - { - *__result = *__first1; - ++__first1; - } - } - return _VSTD::copy(__first2, __last2, __result); -} - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -_OutputIterator -merge(_InputIterator1 __first1, _InputIterator1 __last1, - _InputIterator2 __first2, _InputIterator2 __last2, _OutputIterator __result, _Compare __comp) -{ - typedef typename __comp_ref_type<_Compare>::type _Comp_ref; - return _VSTD::__merge<_Comp_ref>(__first1, __last1, __first2, __last2, __result, __comp); -} - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -_OutputIterator -merge(_InputIterator1 __first1, _InputIterator1 __last1, - _InputIterator2 __first2, _InputIterator2 __last2, _OutputIterator __result) -{ - typedef typename iterator_traits<_InputIterator1>::value_type __v1; - typedef typename iterator_traits<_InputIterator2>::value_type __v2; - return _VSTD::merge(__first1, __last1, __first2, __last2, __result, __less<__v1, __v2>()); -} - -// inplace_merge - -template -void __half_inplace_merge(_InputIterator1 __first1, _InputIterator1 __last1, - _InputIterator2 __first2, _InputIterator2 __last2, - _OutputIterator __result, _Compare __comp) -{ - for (; __first1 != __last1; ++__result) - { - if (__first2 == __last2) - { - _VSTD::move(__first1, __last1, __result); - return; - } - - if (__comp(*__first2, *__first1)) - { - *__result = _VSTD::move(*__first2); - ++__first2; - } - else - { - *__result = _VSTD::move(*__first1); - ++__first1; - } - } - // __first2 through __last2 are already in the right spot. -} - -template -void -__buffered_inplace_merge(_BidirectionalIterator __first, _BidirectionalIterator __middle, _BidirectionalIterator __last, - _Compare __comp, typename iterator_traits<_BidirectionalIterator>::difference_type __len1, - typename iterator_traits<_BidirectionalIterator>::difference_type __len2, - typename iterator_traits<_BidirectionalIterator>::value_type* __buff) -{ - typedef typename iterator_traits<_BidirectionalIterator>::value_type value_type; - __destruct_n __d(0); - unique_ptr __h2(__buff, __d); - if (__len1 <= __len2) - { - value_type* __p = __buff; - for (_BidirectionalIterator __i = __first; __i != __middle; __d.template __incr(), (void) ++__i, (void) ++__p) - ::new ((void*)__p) value_type(_VSTD::move(*__i)); - _VSTD::__half_inplace_merge<_Compare>(__buff, __p, __middle, __last, __first, __comp); - } - else - { - value_type* __p = __buff; - for (_BidirectionalIterator __i = __middle; __i != __last; __d.template __incr(), (void) ++__i, (void) ++__p) - ::new ((void*)__p) value_type(_VSTD::move(*__i)); - typedef reverse_iterator<_BidirectionalIterator> _RBi; - typedef reverse_iterator _Rv; - typedef __invert<_Compare> _Inverted; - _VSTD::__half_inplace_merge<_Inverted>(_Rv(__p), _Rv(__buff), - _RBi(__middle), _RBi(__first), - _RBi(__last), _Inverted(__comp)); - } -} - -template -void -__inplace_merge(_BidirectionalIterator __first, _BidirectionalIterator __middle, _BidirectionalIterator __last, - _Compare __comp, typename iterator_traits<_BidirectionalIterator>::difference_type __len1, - typename iterator_traits<_BidirectionalIterator>::difference_type __len2, - typename iterator_traits<_BidirectionalIterator>::value_type* __buff, ptrdiff_t __buff_size) -{ - typedef typename iterator_traits<_BidirectionalIterator>::difference_type difference_type; - while (true) - { - // if __middle == __last, we're done - if (__len2 == 0) - return; - if (__len1 <= __buff_size || __len2 <= __buff_size) - return _VSTD::__buffered_inplace_merge<_Compare> - (__first, __middle, __last, __comp, __len1, __len2, __buff); - // shrink [__first, __middle) as much as possible (with no moves), returning if it shrinks to 0 - for (; true; ++__first, (void) --__len1) - { - if (__len1 == 0) - return; - if (__comp(*__middle, *__first)) - break; - } - // __first < __middle < __last - // *__first > *__middle - // partition [__first, __m1) [__m1, __middle) [__middle, __m2) [__m2, __last) such that - // all elements in: - // [__first, __m1) <= [__middle, __m2) - // [__middle, __m2) < [__m1, __middle) - // [__m1, __middle) <= [__m2, __last) - // and __m1 or __m2 is in the middle of its range - _BidirectionalIterator __m1; // "median" of [__first, __middle) - _BidirectionalIterator __m2; // "median" of [__middle, __last) - difference_type __len11; // distance(__first, __m1) - difference_type __len21; // distance(__middle, __m2) - // binary search smaller range - if (__len1 < __len2) - { // __len >= 1, __len2 >= 2 - __len21 = __len2 / 2; - __m2 = __middle; - _VSTD::advance(__m2, __len21); - __m1 = _VSTD::__upper_bound<_Compare>(__first, __middle, *__m2, __comp); - __len11 = _VSTD::distance(__first, __m1); - } - else - { - if (__len1 == 1) - { // __len1 >= __len2 && __len2 > 0, therefore __len2 == 1 - // It is known *__first > *__middle - swap(*__first, *__middle); - return; - } - // __len1 >= 2, __len2 >= 1 - __len11 = __len1 / 2; - __m1 = __first; - _VSTD::advance(__m1, __len11); - __m2 = _VSTD::__lower_bound<_Compare>(__middle, __last, *__m1, __comp); - __len21 = _VSTD::distance(__middle, __m2); - } - difference_type __len12 = __len1 - __len11; // distance(__m1, __middle) - difference_type __len22 = __len2 - __len21; // distance(__m2, __last) - // [__first, __m1) [__m1, __middle) [__middle, __m2) [__m2, __last) - // swap middle two partitions - __middle = _VSTD::rotate(__m1, __middle, __m2); - // __len12 and __len21 now have swapped meanings - // merge smaller range with recursive call and larger with tail recursion elimination - if (__len11 + __len21 < __len12 + __len22) - { - _VSTD::__inplace_merge<_Compare>(__first, __m1, __middle, __comp, __len11, __len21, __buff, __buff_size); -// _VSTD::__inplace_merge<_Compare>(__middle, __m2, __last, __comp, __len12, __len22, __buff, __buff_size); - __first = __middle; - __middle = __m2; - __len1 = __len12; - __len2 = __len22; - } - else - { - _VSTD::__inplace_merge<_Compare>(__middle, __m2, __last, __comp, __len12, __len22, __buff, __buff_size); -// _VSTD::__inplace_merge<_Compare>(__first, __m1, __middle, __comp, __len11, __len21, __buff, __buff_size); - __last = __middle; - __middle = __m1; - __len1 = __len11; - __len2 = __len21; - } - } -} - -template -inline _LIBCPP_INLINE_VISIBILITY -void -inplace_merge(_BidirectionalIterator __first, _BidirectionalIterator __middle, _BidirectionalIterator __last, - _Compare __comp) -{ - typedef typename iterator_traits<_BidirectionalIterator>::value_type value_type; - typedef typename iterator_traits<_BidirectionalIterator>::difference_type difference_type; - difference_type __len1 = _VSTD::distance(__first, __middle); - difference_type __len2 = _VSTD::distance(__middle, __last); - difference_type __buf_size = _VSTD::min(__len1, __len2); - pair __buf = _VSTD::get_temporary_buffer(__buf_size); - unique_ptr __h(__buf.first); - typedef typename __comp_ref_type<_Compare>::type _Comp_ref; - return _VSTD::__inplace_merge<_Comp_ref>(__first, __middle, __last, __comp, __len1, __len2, - __buf.first, __buf.second); -} - -template -inline _LIBCPP_INLINE_VISIBILITY -void -inplace_merge(_BidirectionalIterator __first, _BidirectionalIterator __middle, _BidirectionalIterator __last) -{ - _VSTD::inplace_merge(__first, __middle, __last, - __less::value_type>()); -} - -// stable_sort - -template -void -__merge_move_construct(_InputIterator1 __first1, _InputIterator1 __last1, - _InputIterator2 __first2, _InputIterator2 __last2, - typename iterator_traits<_InputIterator1>::value_type* __result, _Compare __comp) -{ - typedef typename iterator_traits<_InputIterator1>::value_type value_type; - __destruct_n __d(0); - unique_ptr __h(__result, __d); - for (; true; ++__result) - { - if (__first1 == __last1) - { - for (; __first2 != __last2; ++__first2, ++__result, (void)__d.template __incr()) - ::new ((void*)__result) value_type(_VSTD::move(*__first2)); - __h.release(); - return; - } - if (__first2 == __last2) - { - for (; __first1 != __last1; ++__first1, ++__result, (void)__d.template __incr()) - ::new ((void*)__result) value_type(_VSTD::move(*__first1)); - __h.release(); - return; - } - if (__comp(*__first2, *__first1)) - { - ::new ((void*)__result) value_type(_VSTD::move(*__first2)); - __d.template __incr(); - ++__first2; - } - else - { - ::new ((void*)__result) value_type(_VSTD::move(*__first1)); - __d.template __incr(); - ++__first1; - } - } -} - -template -void -__merge_move_assign(_InputIterator1 __first1, _InputIterator1 __last1, - _InputIterator2 __first2, _InputIterator2 __last2, - _OutputIterator __result, _Compare __comp) -{ - for (; __first1 != __last1; ++__result) - { - if (__first2 == __last2) - { - for (; __first1 != __last1; ++__first1, (void) ++__result) - *__result = _VSTD::move(*__first1); - return; - } - if (__comp(*__first2, *__first1)) - { - *__result = _VSTD::move(*__first2); - ++__first2; - } - else - { - *__result = _VSTD::move(*__first1); - ++__first1; - } - } - for (; __first2 != __last2; ++__first2, (void) ++__result) - *__result = _VSTD::move(*__first2); -} - -template -void -__stable_sort(_RandomAccessIterator __first, _RandomAccessIterator __last, _Compare __comp, - typename iterator_traits<_RandomAccessIterator>::difference_type __len, - typename iterator_traits<_RandomAccessIterator>::value_type* __buff, ptrdiff_t __buff_size); - -template -void -__stable_sort_move(_RandomAccessIterator __first1, _RandomAccessIterator __last1, _Compare __comp, - typename iterator_traits<_RandomAccessIterator>::difference_type __len, - typename iterator_traits<_RandomAccessIterator>::value_type* __first2) -{ - typedef typename iterator_traits<_RandomAccessIterator>::value_type value_type; - switch (__len) - { - case 0: - return; - case 1: - ::new ((void*)__first2) value_type(_VSTD::move(*__first1)); - return; - case 2: - __destruct_n __d(0); - unique_ptr __h2(__first2, __d); - if (__comp(*--__last1, *__first1)) - { - ::new ((void*)__first2) value_type(_VSTD::move(*__last1)); - __d.template __incr(); - ++__first2; - ::new ((void*)__first2) value_type(_VSTD::move(*__first1)); - } - else - { - ::new ((void*)__first2) value_type(_VSTD::move(*__first1)); - __d.template __incr(); - ++__first2; - ::new ((void*)__first2) value_type(_VSTD::move(*__last1)); - } - __h2.release(); - return; - } - if (__len <= 8) - { - _VSTD::__insertion_sort_move<_Compare>(__first1, __last1, __first2, __comp); - return; - } - typename iterator_traits<_RandomAccessIterator>::difference_type __l2 = __len / 2; - _RandomAccessIterator __m = __first1 + __l2; - _VSTD::__stable_sort<_Compare>(__first1, __m, __comp, __l2, __first2, __l2); - _VSTD::__stable_sort<_Compare>(__m, __last1, __comp, __len - __l2, __first2 + __l2, __len - __l2); - _VSTD::__merge_move_construct<_Compare>(__first1, __m, __m, __last1, __first2, __comp); -} - -template -struct __stable_sort_switch -{ - static const unsigned value = 128*is_trivially_copy_assignable<_Tp>::value; -}; - -template -void -__stable_sort(_RandomAccessIterator __first, _RandomAccessIterator __last, _Compare __comp, - typename iterator_traits<_RandomAccessIterator>::difference_type __len, - typename iterator_traits<_RandomAccessIterator>::value_type* __buff, ptrdiff_t __buff_size) -{ - typedef typename iterator_traits<_RandomAccessIterator>::value_type value_type; - typedef typename iterator_traits<_RandomAccessIterator>::difference_type difference_type; - switch (__len) - { - case 0: - case 1: - return; - case 2: - if (__comp(*--__last, *__first)) - swap(*__first, *__last); - return; - } - if (__len <= static_cast(__stable_sort_switch::value)) - { - _VSTD::__insertion_sort<_Compare>(__first, __last, __comp); - return; - } - typename iterator_traits<_RandomAccessIterator>::difference_type __l2 = __len / 2; - _RandomAccessIterator __m = __first + __l2; - if (__len <= __buff_size) - { - __destruct_n __d(0); - unique_ptr __h2(__buff, __d); - _VSTD::__stable_sort_move<_Compare>(__first, __m, __comp, __l2, __buff); - __d.__set(__l2, (value_type*)nullptr); - _VSTD::__stable_sort_move<_Compare>(__m, __last, __comp, __len - __l2, __buff + __l2); - __d.__set(__len, (value_type*)nullptr); - _VSTD::__merge_move_assign<_Compare>(__buff, __buff + __l2, __buff + __l2, __buff + __len, __first, __comp); -// _VSTD::__merge<_Compare>(move_iterator(__buff), -// move_iterator(__buff + __l2), -// move_iterator<_RandomAccessIterator>(__buff + __l2), -// move_iterator<_RandomAccessIterator>(__buff + __len), -// __first, __comp); - return; - } - _VSTD::__stable_sort<_Compare>(__first, __m, __comp, __l2, __buff, __buff_size); - _VSTD::__stable_sort<_Compare>(__m, __last, __comp, __len - __l2, __buff, __buff_size); - _VSTD::__inplace_merge<_Compare>(__first, __m, __last, __comp, __l2, __len - __l2, __buff, __buff_size); -} - -template -inline _LIBCPP_INLINE_VISIBILITY -void -stable_sort(_RandomAccessIterator __first, _RandomAccessIterator __last, _Compare __comp) -{ - typedef typename iterator_traits<_RandomAccessIterator>::value_type value_type; - typedef typename iterator_traits<_RandomAccessIterator>::difference_type difference_type; - difference_type __len = __last - __first; - pair __buf(0, 0); - unique_ptr __h; - if (__len > static_cast(__stable_sort_switch::value)) - { - __buf = _VSTD::get_temporary_buffer(__len); - __h.reset(__buf.first); - } - typedef typename __comp_ref_type<_Compare>::type _Comp_ref; - _VSTD::__stable_sort<_Comp_ref>(__first, __last, __comp, __len, __buf.first, __buf.second); -} - -template -inline _LIBCPP_INLINE_VISIBILITY -void -stable_sort(_RandomAccessIterator __first, _RandomAccessIterator __last) -{ - _VSTD::stable_sort(__first, __last, __less::value_type>()); -} - -// is_heap_until - -template -_LIBCPP_NODISCARD_EXT _LIBCPP_CONSTEXPR_AFTER_CXX17 _RandomAccessIterator -is_heap_until(_RandomAccessIterator __first, _RandomAccessIterator __last, _Compare __comp) -{ - typedef typename _VSTD::iterator_traits<_RandomAccessIterator>::difference_type difference_type; - difference_type __len = __last - __first; - difference_type __p = 0; - difference_type __c = 1; - _RandomAccessIterator __pp = __first; - while (__c < __len) - { - _RandomAccessIterator __cp = __first + __c; - if (__comp(*__pp, *__cp)) - return __cp; - ++__c; - ++__cp; - if (__c == __len) - return __last; - if (__comp(*__pp, *__cp)) - return __cp; - ++__p; - ++__pp; - __c = 2 * __p + 1; - } - return __last; -} - -template -_LIBCPP_NODISCARD_EXT inline -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -_RandomAccessIterator -is_heap_until(_RandomAccessIterator __first, _RandomAccessIterator __last) -{ - return _VSTD::is_heap_until(__first, __last, __less::value_type>()); -} - -// is_heap - -template -_LIBCPP_NODISCARD_EXT inline -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -bool -is_heap(_RandomAccessIterator __first, _RandomAccessIterator __last, _Compare __comp) -{ - return _VSTD::is_heap_until(__first, __last, __comp) == __last; -} - -template -_LIBCPP_NODISCARD_EXT inline -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -bool -is_heap(_RandomAccessIterator __first, _RandomAccessIterator __last) -{ - return _VSTD::is_heap(__first, __last, __less::value_type>()); -} - -// push_heap - -template -void -__sift_up(_RandomAccessIterator __first, _RandomAccessIterator __last, _Compare __comp, - typename iterator_traits<_RandomAccessIterator>::difference_type __len) -{ - typedef typename iterator_traits<_RandomAccessIterator>::value_type value_type; - if (__len > 1) - { - __len = (__len - 2) / 2; - _RandomAccessIterator __ptr = __first + __len; - if (__comp(*__ptr, *--__last)) - { - value_type __t(_VSTD::move(*__last)); - do - { - *__last = _VSTD::move(*__ptr); - __last = __ptr; - if (__len == 0) - break; - __len = (__len - 1) / 2; - __ptr = __first + __len; - } while (__comp(*__ptr, __t)); - *__last = _VSTD::move(__t); - } - } -} - -template -inline _LIBCPP_INLINE_VISIBILITY -void -push_heap(_RandomAccessIterator __first, _RandomAccessIterator __last, _Compare __comp) -{ - typedef typename __comp_ref_type<_Compare>::type _Comp_ref; - _VSTD::__sift_up<_Comp_ref>(__first, __last, __comp, __last - __first); -} - -template -inline _LIBCPP_INLINE_VISIBILITY -void -push_heap(_RandomAccessIterator __first, _RandomAccessIterator __last) -{ - _VSTD::push_heap(__first, __last, __less::value_type>()); -} - -// pop_heap - -template -void -__sift_down(_RandomAccessIterator __first, _RandomAccessIterator /*__last*/, - _Compare __comp, - typename iterator_traits<_RandomAccessIterator>::difference_type __len, - _RandomAccessIterator __start) -{ - typedef typename iterator_traits<_RandomAccessIterator>::difference_type difference_type; - typedef typename iterator_traits<_RandomAccessIterator>::value_type value_type; - // left-child of __start is at 2 * __start + 1 - // right-child of __start is at 2 * __start + 2 - difference_type __child = __start - __first; - - if (__len < 2 || (__len - 2) / 2 < __child) - return; - - __child = 2 * __child + 1; - _RandomAccessIterator __child_i = __first + __child; - - if ((__child + 1) < __len && __comp(*__child_i, *(__child_i + 1))) { - // right-child exists and is greater than left-child - ++__child_i; - ++__child; - } - - // check if we are in heap-order - if (__comp(*__child_i, *__start)) - // we are, __start is larger than it's largest child - return; - - value_type __top(_VSTD::move(*__start)); - do - { - // we are not in heap-order, swap the parent with it's largest child - *__start = _VSTD::move(*__child_i); - __start = __child_i; - - if ((__len - 2) / 2 < __child) - break; - - // recompute the child based off of the updated parent - __child = 2 * __child + 1; - __child_i = __first + __child; - - if ((__child + 1) < __len && __comp(*__child_i, *(__child_i + 1))) { - // right-child exists and is greater than left-child - ++__child_i; - ++__child; - } - - // check if we are in heap-order - } while (!__comp(*__child_i, __top)); - *__start = _VSTD::move(__top); -} - -template -inline _LIBCPP_INLINE_VISIBILITY -void -__pop_heap(_RandomAccessIterator __first, _RandomAccessIterator __last, _Compare __comp, - typename iterator_traits<_RandomAccessIterator>::difference_type __len) -{ - if (__len > 1) - { - swap(*__first, *--__last); - _VSTD::__sift_down<_Compare>(__first, __last, __comp, __len - 1, __first); - } -} - -template -inline _LIBCPP_INLINE_VISIBILITY -void -pop_heap(_RandomAccessIterator __first, _RandomAccessIterator __last, _Compare __comp) -{ - typedef typename __comp_ref_type<_Compare>::type _Comp_ref; - _VSTD::__pop_heap<_Comp_ref>(__first, __last, __comp, __last - __first); -} - -template -inline _LIBCPP_INLINE_VISIBILITY -void -pop_heap(_RandomAccessIterator __first, _RandomAccessIterator __last) -{ - _VSTD::pop_heap(__first, __last, __less::value_type>()); -} - -// make_heap - -template -void -__make_heap(_RandomAccessIterator __first, _RandomAccessIterator __last, _Compare __comp) -{ - typedef typename iterator_traits<_RandomAccessIterator>::difference_type difference_type; - difference_type __n = __last - __first; - if (__n > 1) - { - // start from the first parent, there is no need to consider children - for (difference_type __start = (__n - 2) / 2; __start >= 0; --__start) - { - _VSTD::__sift_down<_Compare>(__first, __last, __comp, __n, __first + __start); - } - } -} - -template -inline _LIBCPP_INLINE_VISIBILITY -void -make_heap(_RandomAccessIterator __first, _RandomAccessIterator __last, _Compare __comp) -{ - typedef typename __comp_ref_type<_Compare>::type _Comp_ref; - _VSTD::__make_heap<_Comp_ref>(__first, __last, __comp); -} - -template -inline _LIBCPP_INLINE_VISIBILITY -void -make_heap(_RandomAccessIterator __first, _RandomAccessIterator __last) -{ - _VSTD::make_heap(__first, __last, __less::value_type>()); -} - -// sort_heap - -template -void -__sort_heap(_RandomAccessIterator __first, _RandomAccessIterator __last, _Compare __comp) -{ - typedef typename iterator_traits<_RandomAccessIterator>::difference_type difference_type; - for (difference_type __n = __last - __first; __n > 1; --__last, (void) --__n) - _VSTD::__pop_heap<_Compare>(__first, __last, __comp, __n); -} - -template -inline _LIBCPP_INLINE_VISIBILITY -void -sort_heap(_RandomAccessIterator __first, _RandomAccessIterator __last, _Compare __comp) -{ - typedef typename __comp_ref_type<_Compare>::type _Comp_ref; - _VSTD::__sort_heap<_Comp_ref>(__first, __last, __comp); -} - -template -inline _LIBCPP_INLINE_VISIBILITY -void -sort_heap(_RandomAccessIterator __first, _RandomAccessIterator __last) -{ - _VSTD::sort_heap(__first, __last, __less::value_type>()); -} - -// partial_sort - -template -void -__partial_sort(_RandomAccessIterator __first, _RandomAccessIterator __middle, _RandomAccessIterator __last, - _Compare __comp) -{ - _VSTD::__make_heap<_Compare>(__first, __middle, __comp); - typename iterator_traits<_RandomAccessIterator>::difference_type __len = __middle - __first; - for (_RandomAccessIterator __i = __middle; __i != __last; ++__i) - { - if (__comp(*__i, *__first)) - { - swap(*__i, *__first); - _VSTD::__sift_down<_Compare>(__first, __middle, __comp, __len, __first); - } - } - _VSTD::__sort_heap<_Compare>(__first, __middle, __comp); -} - -template -inline _LIBCPP_INLINE_VISIBILITY -void -partial_sort(_RandomAccessIterator __first, _RandomAccessIterator __middle, _RandomAccessIterator __last, - _Compare __comp) -{ - typedef typename __comp_ref_type<_Compare>::type _Comp_ref; - _VSTD::__partial_sort<_Comp_ref>(__first, __middle, __last, __comp); -} - -template -inline _LIBCPP_INLINE_VISIBILITY -void -partial_sort(_RandomAccessIterator __first, _RandomAccessIterator __middle, _RandomAccessIterator __last) -{ - _VSTD::partial_sort(__first, __middle, __last, - __less::value_type>()); -} - -// partial_sort_copy - -template -_RandomAccessIterator -__partial_sort_copy(_InputIterator __first, _InputIterator __last, - _RandomAccessIterator __result_first, _RandomAccessIterator __result_last, _Compare __comp) -{ - _RandomAccessIterator __r = __result_first; - if (__r != __result_last) - { - for (; __first != __last && __r != __result_last; ++__first, (void) ++__r) - *__r = *__first; - _VSTD::__make_heap<_Compare>(__result_first, __r, __comp); - typename iterator_traits<_RandomAccessIterator>::difference_type __len = __r - __result_first; - for (; __first != __last; ++__first) - if (__comp(*__first, *__result_first)) - { - *__result_first = *__first; - _VSTD::__sift_down<_Compare>(__result_first, __r, __comp, __len, __result_first); - } - _VSTD::__sort_heap<_Compare>(__result_first, __r, __comp); - } - return __r; -} - -template -inline _LIBCPP_INLINE_VISIBILITY -_RandomAccessIterator -partial_sort_copy(_InputIterator __first, _InputIterator __last, - _RandomAccessIterator __result_first, _RandomAccessIterator __result_last, _Compare __comp) -{ - typedef typename __comp_ref_type<_Compare>::type _Comp_ref; - return _VSTD::__partial_sort_copy<_Comp_ref>(__first, __last, __result_first, __result_last, __comp); -} - -template -inline _LIBCPP_INLINE_VISIBILITY -_RandomAccessIterator -partial_sort_copy(_InputIterator __first, _InputIterator __last, - _RandomAccessIterator __result_first, _RandomAccessIterator __result_last) -{ - return _VSTD::partial_sort_copy(__first, __last, __result_first, __result_last, - __less::value_type>()); -} - -// nth_element - -template -void -__nth_element(_RandomAccessIterator __first, _RandomAccessIterator __nth, _RandomAccessIterator __last, _Compare __comp) -{ - // _Compare is known to be a reference type - typedef typename iterator_traits<_RandomAccessIterator>::difference_type difference_type; - const difference_type __limit = 7; - while (true) - { - __restart: - if (__nth == __last) - return; - difference_type __len = __last - __first; - switch (__len) - { - case 0: - case 1: - return; - case 2: - if (__comp(*--__last, *__first)) - swap(*__first, *__last); - return; - case 3: - { - _RandomAccessIterator __m = __first; - _VSTD::__sort3<_Compare>(__first, ++__m, --__last, __comp); - return; - } - } - if (__len <= __limit) - { - _VSTD::__selection_sort<_Compare>(__first, __last, __comp); - return; - } - // __len > __limit >= 3 - _RandomAccessIterator __m = __first + __len/2; - _RandomAccessIterator __lm1 = __last; - unsigned __n_swaps = _VSTD::__sort3<_Compare>(__first, __m, --__lm1, __comp); - // *__m is median - // partition [__first, __m) < *__m and *__m <= [__m, __last) - // (this inhibits tossing elements equivalent to __m around unnecessarily) - _RandomAccessIterator __i = __first; - _RandomAccessIterator __j = __lm1; - // j points beyond range to be tested, *__lm1 is known to be <= *__m - // The search going up is known to be guarded but the search coming down isn't. - // Prime the downward search with a guard. - if (!__comp(*__i, *__m)) // if *__first == *__m - { - // *__first == *__m, *__first doesn't go in first part - // manually guard downward moving __j against __i - while (true) - { - if (__i == --__j) - { - // *__first == *__m, *__m <= all other elements - // Partition instead into [__first, __i) == *__first and *__first < [__i, __last) - ++__i; // __first + 1 - __j = __last; - if (!__comp(*__first, *--__j)) // we need a guard if *__first == *(__last-1) - { - while (true) - { - if (__i == __j) - return; // [__first, __last) all equivalent elements - if (__comp(*__first, *__i)) - { - swap(*__i, *__j); - ++__n_swaps; - ++__i; - break; - } - ++__i; - } - } - // [__first, __i) == *__first and *__first < [__j, __last) and __j == __last - 1 - if (__i == __j) - return; - while (true) - { - while (!__comp(*__first, *__i)) - ++__i; - while (__comp(*__first, *--__j)) - ; - if (__i >= __j) - break; - swap(*__i, *__j); - ++__n_swaps; - ++__i; - } - // [__first, __i) == *__first and *__first < [__i, __last) - // The first part is sorted, - if (__nth < __i) - return; - // __nth_element the second part - // _VSTD::__nth_element<_Compare>(__i, __nth, __last, __comp); - __first = __i; - goto __restart; - } - if (__comp(*__j, *__m)) - { - swap(*__i, *__j); - ++__n_swaps; - break; // found guard for downward moving __j, now use unguarded partition - } - } - } - ++__i; - // j points beyond range to be tested, *__lm1 is known to be <= *__m - // if not yet partitioned... - if (__i < __j) - { - // known that *(__i - 1) < *__m - while (true) - { - // __m still guards upward moving __i - while (__comp(*__i, *__m)) - ++__i; - // It is now known that a guard exists for downward moving __j - while (!__comp(*--__j, *__m)) - ; - if (__i >= __j) - break; - swap(*__i, *__j); - ++__n_swaps; - // It is known that __m != __j - // If __m just moved, follow it - if (__m == __i) - __m = __j; - ++__i; - } - } - // [__first, __i) < *__m and *__m <= [__i, __last) - if (__i != __m && __comp(*__m, *__i)) - { - swap(*__i, *__m); - ++__n_swaps; - } - // [__first, __i) < *__i and *__i <= [__i+1, __last) - if (__nth == __i) - return; - if (__n_swaps == 0) - { - // We were given a perfectly partitioned sequence. Coincidence? - if (__nth < __i) - { - // Check for [__first, __i) already sorted - __j = __m = __first; - while (++__j != __i) - { - if (__comp(*__j, *__m)) - // not yet sorted, so sort - goto not_sorted; - __m = __j; - } - // [__first, __i) sorted - return; - } - else - { - // Check for [__i, __last) already sorted - __j = __m = __i; - while (++__j != __last) - { - if (__comp(*__j, *__m)) - // not yet sorted, so sort - goto not_sorted; - __m = __j; - } - // [__i, __last) sorted - return; - } - } -not_sorted: - // __nth_element on range containing __nth - if (__nth < __i) - { - // _VSTD::__nth_element<_Compare>(__first, __nth, __i, __comp); - __last = __i; - } - else - { - // _VSTD::__nth_element<_Compare>(__i+1, __nth, __last, __comp); - __first = ++__i; - } - } -} - -template -inline _LIBCPP_INLINE_VISIBILITY -void -nth_element(_RandomAccessIterator __first, _RandomAccessIterator __nth, _RandomAccessIterator __last, _Compare __comp) -{ - typedef typename __comp_ref_type<_Compare>::type _Comp_ref; - _VSTD::__nth_element<_Comp_ref>(__first, __nth, __last, __comp); -} - -template -inline _LIBCPP_INLINE_VISIBILITY -void -nth_element(_RandomAccessIterator __first, _RandomAccessIterator __nth, _RandomAccessIterator __last) -{ - _VSTD::nth_element(__first, __nth, __last, __less::value_type>()); -} - -// includes - -template -_LIBCPP_CONSTEXPR_AFTER_CXX17 bool -__includes(_InputIterator1 __first1, _InputIterator1 __last1, _InputIterator2 __first2, _InputIterator2 __last2, - _Compare __comp) -{ - for (; __first2 != __last2; ++__first1) - { - if (__first1 == __last1 || __comp(*__first2, *__first1)) - return false; - if (!__comp(*__first1, *__first2)) - ++__first2; - } - return true; -} - -template -_LIBCPP_NODISCARD_EXT inline -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -bool -includes(_InputIterator1 __first1, _InputIterator1 __last1, _InputIterator2 __first2, _InputIterator2 __last2, - _Compare __comp) -{ - typedef typename __comp_ref_type<_Compare>::type _Comp_ref; - return _VSTD::__includes<_Comp_ref>(__first1, __last1, __first2, __last2, __comp); -} - -template -_LIBCPP_NODISCARD_EXT inline -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -bool -includes(_InputIterator1 __first1, _InputIterator1 __last1, _InputIterator2 __first2, _InputIterator2 __last2) -{ - return _VSTD::includes(__first1, __last1, __first2, __last2, - __less::value_type, - typename iterator_traits<_InputIterator2>::value_type>()); -} - -// set_union - -template -_LIBCPP_CONSTEXPR_AFTER_CXX17 _OutputIterator -__set_union(_InputIterator1 __first1, _InputIterator1 __last1, - _InputIterator2 __first2, _InputIterator2 __last2, _OutputIterator __result, _Compare __comp) -{ - for (; __first1 != __last1; ++__result) - { - if (__first2 == __last2) - return _VSTD::copy(__first1, __last1, __result); - if (__comp(*__first2, *__first1)) - { - *__result = *__first2; - ++__first2; - } - else - { - if (!__comp(*__first1, *__first2)) - ++__first2; - *__result = *__first1; - ++__first1; - } - } - return _VSTD::copy(__first2, __last2, __result); -} - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -_OutputIterator -set_union(_InputIterator1 __first1, _InputIterator1 __last1, - _InputIterator2 __first2, _InputIterator2 __last2, _OutputIterator __result, _Compare __comp) -{ - typedef typename __comp_ref_type<_Compare>::type _Comp_ref; - return _VSTD::__set_union<_Comp_ref>(__first1, __last1, __first2, __last2, __result, __comp); -} - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -_OutputIterator -set_union(_InputIterator1 __first1, _InputIterator1 __last1, - _InputIterator2 __first2, _InputIterator2 __last2, _OutputIterator __result) -{ - return _VSTD::set_union(__first1, __last1, __first2, __last2, __result, - __less::value_type, - typename iterator_traits<_InputIterator2>::value_type>()); -} - -// set_intersection - -template -_LIBCPP_CONSTEXPR_AFTER_CXX17 _OutputIterator -__set_intersection(_InputIterator1 __first1, _InputIterator1 __last1, - _InputIterator2 __first2, _InputIterator2 __last2, _OutputIterator __result, _Compare __comp) -{ - while (__first1 != __last1 && __first2 != __last2) - { - if (__comp(*__first1, *__first2)) - ++__first1; - else - { - if (!__comp(*__first2, *__first1)) - { - *__result = *__first1; - ++__result; - ++__first1; - } - ++__first2; - } - } - return __result; -} - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -_OutputIterator -set_intersection(_InputIterator1 __first1, _InputIterator1 __last1, - _InputIterator2 __first2, _InputIterator2 __last2, _OutputIterator __result, _Compare __comp) -{ - typedef typename __comp_ref_type<_Compare>::type _Comp_ref; - return _VSTD::__set_intersection<_Comp_ref>(__first1, __last1, __first2, __last2, __result, __comp); -} - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -_OutputIterator -set_intersection(_InputIterator1 __first1, _InputIterator1 __last1, - _InputIterator2 __first2, _InputIterator2 __last2, _OutputIterator __result) -{ - return _VSTD::set_intersection(__first1, __last1, __first2, __last2, __result, - __less::value_type, - typename iterator_traits<_InputIterator2>::value_type>()); -} - -// set_difference - -template -_LIBCPP_CONSTEXPR_AFTER_CXX17 _OutputIterator -__set_difference(_InputIterator1 __first1, _InputIterator1 __last1, - _InputIterator2 __first2, _InputIterator2 __last2, _OutputIterator __result, _Compare __comp) -{ - while (__first1 != __last1) - { - if (__first2 == __last2) - return _VSTD::copy(__first1, __last1, __result); - if (__comp(*__first1, *__first2)) - { - *__result = *__first1; - ++__result; - ++__first1; - } - else - { - if (!__comp(*__first2, *__first1)) - ++__first1; - ++__first2; - } - } - return __result; -} - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -_OutputIterator -set_difference(_InputIterator1 __first1, _InputIterator1 __last1, - _InputIterator2 __first2, _InputIterator2 __last2, _OutputIterator __result, _Compare __comp) -{ - typedef typename __comp_ref_type<_Compare>::type _Comp_ref; - return _VSTD::__set_difference<_Comp_ref>(__first1, __last1, __first2, __last2, __result, __comp); -} - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -_OutputIterator -set_difference(_InputIterator1 __first1, _InputIterator1 __last1, - _InputIterator2 __first2, _InputIterator2 __last2, _OutputIterator __result) -{ - return _VSTD::set_difference(__first1, __last1, __first2, __last2, __result, - __less::value_type, - typename iterator_traits<_InputIterator2>::value_type>()); -} - -// set_symmetric_difference - -template -_LIBCPP_CONSTEXPR_AFTER_CXX17 _OutputIterator -__set_symmetric_difference(_InputIterator1 __first1, _InputIterator1 __last1, - _InputIterator2 __first2, _InputIterator2 __last2, _OutputIterator __result, _Compare __comp) -{ - while (__first1 != __last1) - { - if (__first2 == __last2) - return _VSTD::copy(__first1, __last1, __result); - if (__comp(*__first1, *__first2)) - { - *__result = *__first1; - ++__result; - ++__first1; - } - else - { - if (__comp(*__first2, *__first1)) - { - *__result = *__first2; - ++__result; - } - else - ++__first1; - ++__first2; - } - } - return _VSTD::copy(__first2, __last2, __result); -} - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -_OutputIterator -set_symmetric_difference(_InputIterator1 __first1, _InputIterator1 __last1, - _InputIterator2 __first2, _InputIterator2 __last2, _OutputIterator __result, _Compare __comp) -{ - typedef typename __comp_ref_type<_Compare>::type _Comp_ref; - return _VSTD::__set_symmetric_difference<_Comp_ref>(__first1, __last1, __first2, __last2, __result, __comp); -} - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -_OutputIterator -set_symmetric_difference(_InputIterator1 __first1, _InputIterator1 __last1, - _InputIterator2 __first2, _InputIterator2 __last2, _OutputIterator __result) -{ - return _VSTD::set_symmetric_difference(__first1, __last1, __first2, __last2, __result, - __less::value_type, - typename iterator_traits<_InputIterator2>::value_type>()); -} - -// lexicographical_compare - -template -_LIBCPP_CONSTEXPR_AFTER_CXX17 bool -__lexicographical_compare(_InputIterator1 __first1, _InputIterator1 __last1, - _InputIterator2 __first2, _InputIterator2 __last2, _Compare __comp) -{ - for (; __first2 != __last2; ++__first1, (void) ++__first2) - { - if (__first1 == __last1 || __comp(*__first1, *__first2)) - return true; - if (__comp(*__first2, *__first1)) - return false; - } - return false; -} - -template -_LIBCPP_NODISCARD_EXT inline -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -bool -lexicographical_compare(_InputIterator1 __first1, _InputIterator1 __last1, - _InputIterator2 __first2, _InputIterator2 __last2, _Compare __comp) -{ - typedef typename __comp_ref_type<_Compare>::type _Comp_ref; - return _VSTD::__lexicographical_compare<_Comp_ref>(__first1, __last1, __first2, __last2, __comp); -} - -template -_LIBCPP_NODISCARD_EXT inline -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -bool -lexicographical_compare(_InputIterator1 __first1, _InputIterator1 __last1, - _InputIterator2 __first2, _InputIterator2 __last2) -{ - return _VSTD::lexicographical_compare(__first1, __last1, __first2, __last2, - __less::value_type, - typename iterator_traits<_InputIterator2>::value_type>()); -} - -// next_permutation - -template -_LIBCPP_CONSTEXPR_AFTER_CXX17 bool -__next_permutation(_BidirectionalIterator __first, _BidirectionalIterator __last, _Compare __comp) -{ - _BidirectionalIterator __i = __last; - if (__first == __last || __first == --__i) - return false; - while (true) - { - _BidirectionalIterator __ip1 = __i; - if (__comp(*--__i, *__ip1)) - { - _BidirectionalIterator __j = __last; - while (!__comp(*__i, *--__j)) - ; - swap(*__i, *__j); - _VSTD::reverse(__ip1, __last); - return true; - } - if (__i == __first) - { - _VSTD::reverse(__first, __last); - return false; - } - } -} - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -bool -next_permutation(_BidirectionalIterator __first, _BidirectionalIterator __last, _Compare __comp) -{ - typedef typename __comp_ref_type<_Compare>::type _Comp_ref; - return _VSTD::__next_permutation<_Comp_ref>(__first, __last, __comp); -} - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -bool -next_permutation(_BidirectionalIterator __first, _BidirectionalIterator __last) -{ - return _VSTD::next_permutation(__first, __last, - __less::value_type>()); -} - -// prev_permutation - -template -_LIBCPP_CONSTEXPR_AFTER_CXX17 bool -__prev_permutation(_BidirectionalIterator __first, _BidirectionalIterator __last, _Compare __comp) -{ - _BidirectionalIterator __i = __last; - if (__first == __last || __first == --__i) - return false; - while (true) - { - _BidirectionalIterator __ip1 = __i; - if (__comp(*__ip1, *--__i)) - { - _BidirectionalIterator __j = __last; - while (!__comp(*--__j, *__i)) - ; - swap(*__i, *__j); - _VSTD::reverse(__ip1, __last); - return true; - } - if (__i == __first) - { - _VSTD::reverse(__first, __last); - return false; - } - } -} - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -bool -prev_permutation(_BidirectionalIterator __first, _BidirectionalIterator __last, _Compare __comp) -{ - typedef typename __comp_ref_type<_Compare>::type _Comp_ref; - return _VSTD::__prev_permutation<_Comp_ref>(__first, __last, __comp); -} - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -bool -prev_permutation(_BidirectionalIterator __first, _BidirectionalIterator __last) -{ - return _VSTD::prev_permutation(__first, __last, - __less::value_type>()); -} - -_LIBCPP_END_NAMESPACE_STD - _LIBCPP_POP_MACROS #if defined(_LIBCPP_HAS_PARALLEL_ALGORITHMS) && _LIBCPP_STD_VER >= 17 # include <__pstl_algorithm> #endif -#endif // _LIBCPP_ALGORITHM +#endif // _LIBCPP_ALGORITHM diff --git a/lib/libcxx/include/any b/lib/libcxx/include/any index 968c9769e..3a826c4d5 100644 --- a/lib/libcxx/include/any +++ b/lib/libcxx/include/any @@ -80,12 +80,13 @@ namespace std { */ -#include #include <__availability> +#include <__config> +#include <__utility/forward.h> +#include #include -#include #include -#include +#include #include #if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) diff --git a/lib/libcxx/include/array b/lib/libcxx/include/array index 9a479f7cd..022172b86 100644 --- a/lib/libcxx/include/array +++ b/lib/libcxx/include/array @@ -109,25 +109,22 @@ template const T&& get(const array&&) noexce */ #include <__config> +#include <__debug> #include <__tuple> -#include -#include -#include #include -#include #include // for _LIBCPP_UNREACHABLE +#include +#include +#include +#include #include -#include <__debug> #if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) #pragma GCC system_header #endif - - _LIBCPP_BEGIN_NAMESPACE_STD - template struct _LIBCPP_TEMPLATE_VIS array { @@ -520,4 +517,4 @@ to_array(_Tp(&&__arr)[_Size]) noexcept(is_nothrow_move_constructible_v<_Tp>) { _LIBCPP_END_NAMESPACE_STD -#endif // _LIBCPP_ARRAY +#endif // _LIBCPP_ARRAY diff --git a/lib/libcxx/include/atomic b/lib/libcxx/include/atomic index 0fc799a24..0f6aee83c 100644 --- a/lib/libcxx/include/atomic +++ b/lib/libcxx/include/atomic @@ -67,7 +67,8 @@ struct atomic bool is_lock_free() const volatile noexcept; bool is_lock_free() const noexcept; - atomic() noexcept = default; + atomic() noexcept = default; // until C++20 + constexpr atomic() noexcept(is_nothrow_default_constructible_v); // since C++20 constexpr atomic(T desr) noexcept; atomic(const atomic&) = delete; atomic& operator=(const atomic&) = delete; @@ -201,7 +202,8 @@ struct atomic bool is_lock_free() const volatile noexcept; bool is_lock_free() const noexcept; - atomic() noexcept = default; + atomic() noexcept = default; // until C++20 + constexpr atomic() noexcept; // since C++20 constexpr atomic(T* desr) noexcept; atomic(const atomic&) = delete; atomic& operator=(const atomic&) = delete; @@ -509,7 +511,8 @@ typedef atomic atomic_uintmax_t; typedef struct atomic_flag { - atomic_flag() noexcept = default; + atomic_flag() noexcept = default; // until C++20 + constexpr atomic_flag() noexcept; // since C++20 atomic_flag(const atomic_flag&) = delete; atomic_flag& operator=(const atomic_flag&) = delete; atomic_flag& operator=(const atomic_flag&) volatile = delete; @@ -574,8 +577,8 @@ template */ -#include <__config> #include <__availability> +#include <__config> #include <__threading_support> #include #include @@ -669,7 +672,7 @@ static_assert((is_same::type, __memory_order_under "unexpected underlying type for std::memory_order"); #if defined(_LIBCPP_HAS_GCC_ATOMIC_IMP) || \ - defined(_LIBCPP_ATOMIC_ONLY_USE_BUILTINS) + defined(_LIBCPP_ATOMIC_ONLY_USE_BUILTINS) // [atomics.types.generic]p1 guarantees _Tp is trivially copyable. Because // the default operator= in an object is not volatile, a byte-by-byte copy @@ -1017,26 +1020,33 @@ _Tp __cxx_atomic_exchange(__cxx_atomic_base_impl<_Tp> * __a, _Tp __value, memory return __c11_atomic_exchange(&__a->__a_value, __value, static_cast<__memory_order_underlying_t>(__order)); } +_LIBCPP_INLINE_VISIBILITY inline _LIBCPP_CONSTEXPR memory_order __to_failure_order(memory_order __order) { + // Avoid switch statement to make this a constexpr. + return __order == memory_order_release ? memory_order_relaxed: + (__order == memory_order_acq_rel ? memory_order_acquire: + __order); +} + template _LIBCPP_INLINE_VISIBILITY bool __cxx_atomic_compare_exchange_strong(__cxx_atomic_base_impl<_Tp> volatile* __a, _Tp* __expected, _Tp __value, memory_order __success, memory_order __failure) _NOEXCEPT { - return __c11_atomic_compare_exchange_strong(&__a->__a_value, __expected, __value, static_cast<__memory_order_underlying_t>(__success), static_cast<__memory_order_underlying_t>(__failure)); + return __c11_atomic_compare_exchange_strong(&__a->__a_value, __expected, __value, static_cast<__memory_order_underlying_t>(__success), static_cast<__memory_order_underlying_t>(__to_failure_order(__failure))); } template _LIBCPP_INLINE_VISIBILITY bool __cxx_atomic_compare_exchange_strong(__cxx_atomic_base_impl<_Tp> * __a, _Tp* __expected, _Tp __value, memory_order __success, memory_order __failure) _NOEXCEPT { - return __c11_atomic_compare_exchange_strong(&__a->__a_value, __expected, __value, static_cast<__memory_order_underlying_t>(__success), static_cast<__memory_order_underlying_t>(__failure)); + return __c11_atomic_compare_exchange_strong(&__a->__a_value, __expected, __value, static_cast<__memory_order_underlying_t>(__success), static_cast<__memory_order_underlying_t>(__to_failure_order(__failure))); } template _LIBCPP_INLINE_VISIBILITY bool __cxx_atomic_compare_exchange_weak(__cxx_atomic_base_impl<_Tp> volatile* __a, _Tp* __expected, _Tp __value, memory_order __success, memory_order __failure) _NOEXCEPT { - return __c11_atomic_compare_exchange_weak(&__a->__a_value, __expected, __value, static_cast<__memory_order_underlying_t>(__success), static_cast<__memory_order_underlying_t>(__failure)); + return __c11_atomic_compare_exchange_weak(&__a->__a_value, __expected, __value, static_cast<__memory_order_underlying_t>(__success), static_cast<__memory_order_underlying_t>(__to_failure_order(__failure))); } template _LIBCPP_INLINE_VISIBILITY bool __cxx_atomic_compare_exchange_weak(__cxx_atomic_base_impl<_Tp> * __a, _Tp* __expected, _Tp __value, memory_order __success, memory_order __failure) _NOEXCEPT { - return __c11_atomic_compare_exchange_weak(&__a->__a_value, __expected, __value, static_cast<__memory_order_underlying_t>(__success), static_cast<__memory_order_underlying_t>(__failure)); + return __c11_atomic_compare_exchange_weak(&__a->__a_value, __expected, __value, static_cast<__memory_order_underlying_t>(__success), static_cast<__memory_order_underlying_t>(__to_failure_order(__failure))); } template @@ -1127,7 +1137,7 @@ _Tp kill_dependency(_Tp __y) _NOEXCEPT #if defined(__CLANG_ATOMIC_BOOL_LOCK_FREE) # define ATOMIC_BOOL_LOCK_FREE __CLANG_ATOMIC_BOOL_LOCK_FREE # define ATOMIC_CHAR_LOCK_FREE __CLANG_ATOMIC_CHAR_LOCK_FREE -#ifndef _LIBCPP_NO_HAS_CHAR8_T +#ifndef _LIBCPP_HAS_NO_CHAR8_T # define ATOMIC_CHAR8_T_LOCK_FREE __CLANG_ATOMIC_CHAR8_T_LOCK_FREE #endif # define ATOMIC_CHAR16_T_LOCK_FREE __CLANG_ATOMIC_CHAR16_T_LOCK_FREE @@ -1141,7 +1151,7 @@ _Tp kill_dependency(_Tp __y) _NOEXCEPT #elif defined(__GCC_ATOMIC_BOOL_LOCK_FREE) # define ATOMIC_BOOL_LOCK_FREE __GCC_ATOMIC_BOOL_LOCK_FREE # define ATOMIC_CHAR_LOCK_FREE __GCC_ATOMIC_CHAR_LOCK_FREE -#ifndef _LIBCPP_NO_HAS_CHAR8_T +#ifndef _LIBCPP_HAS_NO_CHAR8_T # define ATOMIC_CHAR8_T_LOCK_FREE __GCC_ATOMIC_CHAR8_T_LOCK_FREE #endif # define ATOMIC_CHAR16_T_LOCK_FREE __GCC_ATOMIC_CHAR16_T_LOCK_FREE @@ -1458,7 +1468,7 @@ template<> struct __cxx_is_always_lock_free { enum { __value = 2 == ATOMIC template<> struct __cxx_is_always_lock_free { enum { __value = 2 == ATOMIC_CHAR_LOCK_FREE }; }; template<> struct __cxx_is_always_lock_free { enum { __value = 2 == ATOMIC_CHAR_LOCK_FREE }; }; template<> struct __cxx_is_always_lock_free { enum { __value = 2 == ATOMIC_CHAR_LOCK_FREE }; }; -#ifndef _LIBCPP_NO_HAS_CHAR8_T +#ifndef _LIBCPP_HAS_NO_CHAR8_T template<> struct __cxx_is_always_lock_free { enum { __value = 2 == ATOMIC_CHAR8_T_LOCK_FREE }; }; #endif template<> struct __cxx_is_always_lock_free { enum { __value = 2 == ATOMIC_CHAR16_T_LOCK_FREE }; }; @@ -1673,24 +1683,23 @@ struct __atomic_base // false _LIBCPP_AVAILABILITY_SYNC _LIBCPP_INLINE_VISIBILITY void notify_all() _NOEXCEPT {__cxx_atomic_notify_all(&__a_);} +#if _LIBCPP_STD_VER > 17 + _LIBCPP_INLINE_VISIBILITY constexpr + __atomic_base() noexcept(is_nothrow_default_constructible_v<_Tp>) : __a_(_Tp()) {} +#else _LIBCPP_INLINE_VISIBILITY __atomic_base() _NOEXCEPT _LIBCPP_DEFAULT +#endif _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR __atomic_base(_Tp __d) _NOEXCEPT : __a_(__d) {} #ifndef _LIBCPP_CXX03_LANG __atomic_base(const __atomic_base&) = delete; - __atomic_base& operator=(const __atomic_base&) = delete; - __atomic_base& operator=(const __atomic_base&) volatile = delete; #else private: _LIBCPP_INLINE_VISIBILITY __atomic_base(const __atomic_base&); - _LIBCPP_INLINE_VISIBILITY - __atomic_base& operator=(const __atomic_base&); - _LIBCPP_INLINE_VISIBILITY - __atomic_base& operator=(const __atomic_base&) volatile; #endif }; @@ -1706,8 +1715,10 @@ struct __atomic_base<_Tp, true> : public __atomic_base<_Tp, false> { typedef __atomic_base<_Tp, false> __base; - _LIBCPP_INLINE_VISIBILITY + + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 __atomic_base() _NOEXCEPT _LIBCPP_DEFAULT + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR __atomic_base(_Tp __d) _NOEXCEPT : __base(__d) {} @@ -1789,8 +1800,15 @@ struct atomic typedef __atomic_base<_Tp> __base; typedef _Tp value_type; typedef value_type difference_type; + +#if _LIBCPP_STD_VER > 17 + _LIBCPP_INLINE_VISIBILITY + atomic() = default; +#else _LIBCPP_INLINE_VISIBILITY atomic() _NOEXCEPT _LIBCPP_DEFAULT +#endif + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR atomic(_Tp __d) _NOEXCEPT : __base(__d) {} @@ -1800,6 +1818,9 @@ struct atomic _LIBCPP_INLINE_VISIBILITY _Tp operator=(_Tp __d) _NOEXCEPT {__base::store(__d); return __d;} + + atomic& operator=(const atomic&) = delete; + atomic& operator=(const atomic&) volatile = delete; }; // atomic @@ -1811,8 +1832,10 @@ struct atomic<_Tp*> typedef __atomic_base<_Tp*> __base; typedef _Tp* value_type; typedef ptrdiff_t difference_type; + _LIBCPP_INLINE_VISIBILITY atomic() _NOEXCEPT _LIBCPP_DEFAULT + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR atomic(_Tp* __d) _NOEXCEPT : __base(__d) {} @@ -1862,6 +1885,9 @@ struct atomic<_Tp*> _Tp* operator-=(ptrdiff_t __op) volatile _NOEXCEPT {return fetch_sub(__op) - __op;} _LIBCPP_INLINE_VISIBILITY _Tp* operator-=(ptrdiff_t __op) _NOEXCEPT {return fetch_sub(__op) - __op;} + + atomic& operator=(const atomic&) = delete; + atomic& operator=(const atomic&) volatile = delete; }; // atomic_is_lock_free @@ -1885,7 +1911,7 @@ atomic_is_lock_free(const atomic<_Tp>* __o) _NOEXCEPT // atomic_init template -_LIBCPP_INLINE_VISIBILITY +_LIBCPP_DEPRECATED_IN_CXX20 _LIBCPP_INLINE_VISIBILITY void atomic_init(volatile atomic<_Tp>* __o, typename atomic<_Tp>::value_type __d) _NOEXCEPT { @@ -1893,7 +1919,7 @@ atomic_init(volatile atomic<_Tp>* __o, typename atomic<_Tp>::value_type __d) _NO } template -_LIBCPP_INLINE_VISIBILITY +_LIBCPP_DEPRECATED_IN_CXX20 _LIBCPP_INLINE_VISIBILITY void atomic_init(atomic<_Tp>* __o, typename atomic<_Tp>::value_type __d) _NOEXCEPT { @@ -2534,8 +2560,13 @@ typedef struct atomic_flag void notify_all() _NOEXCEPT {__cxx_atomic_notify_all(&__a_);} +#if _LIBCPP_STD_VER > 17 + _LIBCPP_INLINE_VISIBILITY constexpr + atomic_flag() _NOEXCEPT : __a_(false) {} +#else _LIBCPP_INLINE_VISIBILITY atomic_flag() _NOEXCEPT _LIBCPP_DEFAULT +#endif _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR atomic_flag(bool __b) _NOEXCEPT : __a_(__b) {} // EXTENSION @@ -2728,7 +2759,7 @@ typedef atomic atomic_long; typedef atomic atomic_ulong; typedef atomic atomic_llong; typedef atomic atomic_ullong; -#ifndef _LIBCPP_NO_HAS_CHAR8_T +#ifndef _LIBCPP_HAS_NO_CHAR8_T typedef atomic atomic_char8_t; #endif typedef atomic atomic_char16_t; @@ -2801,4 +2832,4 @@ typedef atomic<__libcpp_unsigned_lock_free> atomic_unsigned_lock_free; _LIBCPP_END_NAMESPACE_STD -#endif // _LIBCPP_ATOMIC +#endif // _LIBCPP_ATOMIC diff --git a/lib/libcxx/include/barrier b/lib/libcxx/include/barrier index be213a689..e26dba653 100644 --- a/lib/libcxx/include/barrier +++ b/lib/libcxx/include/barrier @@ -45,8 +45,8 @@ namespace std */ -#include <__config> #include <__availability> +#include <__config> #include #ifndef _LIBCPP_HAS_NO_TREE_BARRIER # include @@ -107,7 +107,6 @@ void __destroy_barrier_algorithm_base(__barrier_algorithm_base* __barrier); template class __barrier_base { - ptrdiff_t __expected; unique_ptr<__barrier_algorithm_base, void (*)(__barrier_algorithm_base*)> __base; @@ -146,7 +145,7 @@ public: _LIBCPP_AVAILABILITY_SYNC _LIBCPP_INLINE_VISIBILITY void wait(arrival_token&& __old_phase) const { - auto const __test_fn = [=]() -> bool { + auto const __test_fn = [this, __old_phase]() -> bool { return __phase.load(memory_order_acquire) != __old_phase; }; __libcpp_thread_poll_with_backoff(__test_fn, __libcpp_timed_backoff_policy()); @@ -309,11 +308,11 @@ public: { __b.wait(_VSTD::move(__phase)); } - _LIBCPP_AVAILABILITY_SYNC _LIBCPP_INLINE_VISIBILITY + _LIBCPP_AVAILABILITY_SYNC _LIBCPP_INLINE_VISIBILITY void arrive_and_wait() { wait(arrive()); - } + } _LIBCPP_AVAILABILITY_SYNC _LIBCPP_INLINE_VISIBILITY void arrive_and_drop() { diff --git a/lib/libcxx/include/bit b/lib/libcxx/include/bit index f8c37c3d6..c64e45c5f 100644 --- a/lib/libcxx/include/bit +++ b/lib/libcxx/include/bit @@ -55,11 +55,11 @@ namespace std { */ #include <__config> -#include <__bits> +#include <__bits> // __libcpp_clz +#include <__debug> #include #include #include -#include <__debug> #if defined(__IBMCPP__) #include "__support/ibm/support.h" @@ -77,49 +77,33 @@ _LIBCPP_PUSH_MACROS _LIBCPP_BEGIN_NAMESPACE_STD - -template -using __bitop_unsigned_integer _LIBCPP_NODEBUG_TYPE = integral_constant::value && - is_unsigned<_Tp>::value && - _IsNotSame::type, bool>::value && - _IsNotSame::type, signed char>::value && - _IsNotSame::type, wchar_t>::value && - _IsNotSame::type, char16_t>::value && - _IsNotSame::type, char32_t>::value - >; - - template _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 _Tp __rotl(_Tp __t, unsigned int __cnt) _NOEXCEPT { - static_assert(__bitop_unsigned_integer<_Tp>::value, "__rotl requires unsigned"); + static_assert(__libcpp_is_unsigned_integer<_Tp>::value, "__rotl requires an unsigned integer type"); const unsigned int __dig = numeric_limits<_Tp>::digits; if ((__cnt % __dig) == 0) return __t; return (__t << (__cnt % __dig)) | (__t >> (__dig - (__cnt % __dig))); } - template _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 _Tp __rotr(_Tp __t, unsigned int __cnt) _NOEXCEPT { - static_assert(__bitop_unsigned_integer<_Tp>::value, "__rotr requires unsigned"); + static_assert(__libcpp_is_unsigned_integer<_Tp>::value, "__rotr requires an unsigned integer type"); const unsigned int __dig = numeric_limits<_Tp>::digits; if ((__cnt % __dig) == 0) return __t; return (__t >> (__cnt % __dig)) | (__t << (__dig - (__cnt % __dig))); } - - template _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 int __countr_zero(_Tp __t) _NOEXCEPT { - static_assert(__bitop_unsigned_integer<_Tp>::value, "__countr_zero requires unsigned"); + static_assert(__libcpp_is_unsigned_integer<_Tp>::value, "__countr_zero requires an unsigned integer type"); if (__t == 0) return numeric_limits<_Tp>::digits; @@ -132,14 +116,13 @@ int __countr_zero(_Tp __t) _NOEXCEPT else { int __ret = 0; - int __iter = 0; const unsigned int __ulldigits = numeric_limits::digits; - while ((__iter = __libcpp_ctz(static_cast(__t))) == __ulldigits) + while (static_cast(__t) == 0uLL) { - __ret += __iter; + __ret += __ulldigits; __t >>= __ulldigits; } - return __ret + __iter; + return __ret + __libcpp_ctz(static_cast(__t)); } } @@ -147,7 +130,7 @@ template _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 int __countl_zero(_Tp __t) _NOEXCEPT { - static_assert(__bitop_unsigned_integer<_Tp>::value, "__countl_zero requires unsigned"); + static_assert(__libcpp_is_unsigned_integer<_Tp>::value, "__countl_zero requires an unsigned integer type"); if (__t == 0) return numeric_limits<_Tp>::digits; @@ -179,30 +162,27 @@ template _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 int __countl_one(_Tp __t) _NOEXCEPT { - static_assert(__bitop_unsigned_integer<_Tp>::value, "__countl_one requires unsigned"); + static_assert(__libcpp_is_unsigned_integer<_Tp>::value, "__countl_one requires an unsigned integer type"); return __t != numeric_limits<_Tp>::max() ? __countl_zero(static_cast<_Tp>(~__t)) : numeric_limits<_Tp>::digits; } - template _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 int __countr_one(_Tp __t) _NOEXCEPT { - static_assert(__bitop_unsigned_integer<_Tp>::value, "__countr_one requires unsigned"); + static_assert(__libcpp_is_unsigned_integer<_Tp>::value, "__countr_one requires an unsigned integer type"); return __t != numeric_limits<_Tp>::max() ? __countr_zero(static_cast<_Tp>(~__t)) : numeric_limits<_Tp>::digits; } - template _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 -int -__popcount(_Tp __t) _NOEXCEPT +int __popcount(_Tp __t) _NOEXCEPT { - static_assert(__bitop_unsigned_integer<_Tp>::value, "__libcpp_popcount requires unsigned"); + static_assert(__libcpp_is_unsigned_integer<_Tp>::value, "__popcount requires an unsigned integer type"); if (sizeof(_Tp) <= sizeof(unsigned int)) return __libcpp_popcount(static_cast(__t)); else if (sizeof(_Tp) <= sizeof(unsigned long)) @@ -221,13 +201,12 @@ __popcount(_Tp __t) _NOEXCEPT } } - // integral log base 2 template _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 unsigned __bit_log2(_Tp __t) _NOEXCEPT { - static_assert(__bitop_unsigned_integer<_Tp>::value, "__bit_log2 requires unsigned"); + static_assert(__libcpp_is_unsigned_integer<_Tp>::value, "__bit_log2 requires an unsigned integer type"); return numeric_limits<_Tp>::digits - 1 - __countl_zero(__t); } @@ -235,80 +214,71 @@ template _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR bool __has_single_bit(_Tp __t) _NOEXCEPT { - static_assert(__bitop_unsigned_integer<_Tp>::value, "__has_single_bit requires unsigned"); + static_assert(__libcpp_is_unsigned_integer<_Tp>::value, "__has_single_bit requires an unsigned integer type"); return __t != 0 && (((__t & (__t - 1)) == 0)); } - #if _LIBCPP_STD_VER > 17 template _LIBCPP_INLINE_VISIBILITY constexpr -enable_if_t<__bitop_unsigned_integer<_Tp>::value, _Tp> +_EnableIf<__libcpp_is_unsigned_integer<_Tp>::value, _Tp> rotl(_Tp __t, unsigned int __cnt) noexcept { return __rotl(__t, __cnt); } - -// rotr template _LIBCPP_INLINE_VISIBILITY constexpr -enable_if_t<__bitop_unsigned_integer<_Tp>::value, _Tp> +_EnableIf<__libcpp_is_unsigned_integer<_Tp>::value, _Tp> rotr(_Tp __t, unsigned int __cnt) noexcept { return __rotr(__t, __cnt); } - template _LIBCPP_INLINE_VISIBILITY constexpr -enable_if_t<__bitop_unsigned_integer<_Tp>::value, int> +_EnableIf<__libcpp_is_unsigned_integer<_Tp>::value, int> countl_zero(_Tp __t) noexcept { return __countl_zero(__t); } - template _LIBCPP_INLINE_VISIBILITY constexpr -enable_if_t<__bitop_unsigned_integer<_Tp>::value, int> +_EnableIf<__libcpp_is_unsigned_integer<_Tp>::value, int> countl_one(_Tp __t) noexcept { return __countl_one(__t); } - template _LIBCPP_INLINE_VISIBILITY constexpr -enable_if_t<__bitop_unsigned_integer<_Tp>::value, int> +_EnableIf<__libcpp_is_unsigned_integer<_Tp>::value, int> countr_zero(_Tp __t) noexcept { return __countr_zero(__t); } - template _LIBCPP_INLINE_VISIBILITY constexpr -enable_if_t<__bitop_unsigned_integer<_Tp>::value, int> +_EnableIf<__libcpp_is_unsigned_integer<_Tp>::value, int> countr_one(_Tp __t) noexcept { return __countr_one(__t); } - template _LIBCPP_INLINE_VISIBILITY constexpr -enable_if_t<__bitop_unsigned_integer<_Tp>::value, int> +_EnableIf<__libcpp_is_unsigned_integer<_Tp>::value, int> popcount(_Tp __t) noexcept { return __popcount(__t); } - template _LIBCPP_INLINE_VISIBILITY constexpr -enable_if_t<__bitop_unsigned_integer<_Tp>::value, bool> +_EnableIf<__libcpp_is_unsigned_integer<_Tp>::value, bool> has_single_bit(_Tp __t) noexcept { return __has_single_bit(__t); @@ -316,7 +286,7 @@ has_single_bit(_Tp __t) noexcept template _LIBCPP_INLINE_VISIBILITY constexpr -enable_if_t<__bitop_unsigned_integer<_Tp>::value, _Tp> +_EnableIf<__libcpp_is_unsigned_integer<_Tp>::value, _Tp> bit_floor(_Tp __t) noexcept { return __t == 0 ? 0 : _Tp{1} << __bit_log2(__t); @@ -324,7 +294,7 @@ bit_floor(_Tp __t) noexcept template _LIBCPP_INLINE_VISIBILITY constexpr -enable_if_t<__bitop_unsigned_integer<_Tp>::value, _Tp> +_EnableIf<__libcpp_is_unsigned_integer<_Tp>::value, _Tp> bit_ceil(_Tp __t) noexcept { if (__t < 2) return 1; @@ -343,7 +313,7 @@ bit_ceil(_Tp __t) noexcept template _LIBCPP_INLINE_VISIBILITY constexpr -enable_if_t<__bitop_unsigned_integer<_Tp>::value, _Tp> +_EnableIf<__libcpp_is_unsigned_integer<_Tp>::value, _Tp> bit_width(_Tp __t) noexcept { return __t == 0 ? 0 : __bit_log2(__t) + 1; diff --git a/lib/libcxx/include/bitset b/lib/libcxx/include/bitset index 00503fe1c..4b8827e77 100644 --- a/lib/libcxx/include/bitset +++ b/lib/libcxx/include/bitset @@ -114,12 +114,12 @@ template struct hash>; #include <__config> #include <__bit_reference> -#include +#include <__functional_base> #include -#include -#include +#include #include -#include <__functional_base> +#include +#include #if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) #pragma GCC system_header @@ -202,7 +202,7 @@ private: void __init(unsigned long long __v, false_type) _NOEXCEPT; _LIBCPP_INLINE_VISIBILITY void __init(unsigned long long __v, true_type) _NOEXCEPT; -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG unsigned long to_ulong(false_type) const; _LIBCPP_INLINE_VISIBILITY unsigned long to_ulong(true_type) const; @@ -258,7 +258,7 @@ __bitset<_N_words, _Size>::__init(unsigned long long __v, true_type) _NOEXCEPT _VSTD::fill(__first_ + 1, __first_ + sizeof(__first_)/sizeof(__first_[0]), __storage_type(0)); } -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG template inline @@ -775,10 +775,7 @@ bitset<_Size>::bitset(const _CharT* __str, for (; __i < _Mp; ++__i) { _CharT __c = __str[_Mp - 1 - __i]; - if (__c == __zero) - (*this)[__i] = false; - else - (*this)[__i] = true; + (*this)[__i] = (__c == __one); } _VSTD::fill(base::__make_iter(__i), base::__make_iter(_Size), false); } @@ -803,10 +800,7 @@ bitset<_Size>::bitset(const basic_string<_CharT,_Traits,_Allocator>& __str, for (; __i < _Mp; ++__i) { _CharT __c = __str[__pos + _Mp - 1 - __i]; - if (_Traits::eq(__c, __zero)) - (*this)[__i] = false; - else - (*this)[__i] = true; + (*this)[__i] = _Traits::eq(__c, __one); } _VSTD::fill(base::__make_iter(__i), base::__make_iter(_Size), false); } @@ -1106,4 +1100,4 @@ _LIBCPP_END_NAMESPACE_STD _LIBCPP_POP_MACROS -#endif // _LIBCPP_BITSET +#endif // _LIBCPP_BITSET diff --git a/lib/libcxx/include/ccomplex b/lib/libcxx/include/ccomplex index 0d2e0f5a0..cea450912 100644 --- a/lib/libcxx/include/ccomplex +++ b/lib/libcxx/include/ccomplex @@ -25,4 +25,4 @@ // hh 080623 Created -#endif // _LIBCPP_CCOMPLEX +#endif // _LIBCPP_CCOMPLEX diff --git a/lib/libcxx/include/cctype b/lib/libcxx/include/cctype index 55fc9ebc6..b07805686 100644 --- a/lib/libcxx/include/cctype +++ b/lib/libcxx/include/cctype @@ -100,21 +100,21 @@ _LIBCPP_BEGIN_NAMESPACE_STD #endif -using ::isalnum; -using ::isalpha; -using ::isblank; -using ::iscntrl; -using ::isdigit; -using ::isgraph; -using ::islower; -using ::isprint; -using ::ispunct; -using ::isspace; -using ::isupper; -using ::isxdigit; -using ::tolower; -using ::toupper; +using ::isalnum _LIBCPP_USING_IF_EXISTS; +using ::isalpha _LIBCPP_USING_IF_EXISTS; +using ::isblank _LIBCPP_USING_IF_EXISTS; +using ::iscntrl _LIBCPP_USING_IF_EXISTS; +using ::isdigit _LIBCPP_USING_IF_EXISTS; +using ::isgraph _LIBCPP_USING_IF_EXISTS; +using ::islower _LIBCPP_USING_IF_EXISTS; +using ::isprint _LIBCPP_USING_IF_EXISTS; +using ::ispunct _LIBCPP_USING_IF_EXISTS; +using ::isspace _LIBCPP_USING_IF_EXISTS; +using ::isupper _LIBCPP_USING_IF_EXISTS; +using ::isxdigit _LIBCPP_USING_IF_EXISTS; +using ::tolower _LIBCPP_USING_IF_EXISTS; +using ::toupper _LIBCPP_USING_IF_EXISTS; _LIBCPP_END_NAMESPACE_STD -#endif // _LIBCPP_CCTYPE +#endif // _LIBCPP_CCTYPE diff --git a/lib/libcxx/include/cerrno b/lib/libcxx/include/cerrno index a9268a281..1388d7eac 100644 --- a/lib/libcxx/include/cerrno +++ b/lib/libcxx/include/cerrno @@ -29,4 +29,4 @@ Macros: #pragma GCC system_header #endif -#endif // _LIBCPP_CERRNO +#endif // _LIBCPP_CERRNO diff --git a/lib/libcxx/include/cfenv b/lib/libcxx/include/cfenv index 6cd91db5a..05b55ee36 100644 --- a/lib/libcxx/include/cfenv +++ b/lib/libcxx/include/cfenv @@ -61,21 +61,21 @@ int feupdateenv(const fenv_t* envp); _LIBCPP_BEGIN_NAMESPACE_STD -using ::fenv_t; -using ::fexcept_t; - -using ::feclearexcept; -using ::fegetexceptflag; -using ::feraiseexcept; -using ::fesetexceptflag; -using ::fetestexcept; -using ::fegetround; -using ::fesetround; -using ::fegetenv; -using ::feholdexcept; -using ::fesetenv; -using ::feupdateenv; +using ::fenv_t _LIBCPP_USING_IF_EXISTS; +using ::fexcept_t _LIBCPP_USING_IF_EXISTS; + +using ::feclearexcept _LIBCPP_USING_IF_EXISTS; +using ::fegetexceptflag _LIBCPP_USING_IF_EXISTS; +using ::feraiseexcept _LIBCPP_USING_IF_EXISTS; +using ::fesetexceptflag _LIBCPP_USING_IF_EXISTS; +using ::fetestexcept _LIBCPP_USING_IF_EXISTS; +using ::fegetround _LIBCPP_USING_IF_EXISTS; +using ::fesetround _LIBCPP_USING_IF_EXISTS; +using ::fegetenv _LIBCPP_USING_IF_EXISTS; +using ::feholdexcept _LIBCPP_USING_IF_EXISTS; +using ::fesetenv _LIBCPP_USING_IF_EXISTS; +using ::feupdateenv _LIBCPP_USING_IF_EXISTS; _LIBCPP_END_NAMESPACE_STD -#endif // _LIBCPP_CFENV +#endif // _LIBCPP_CFENV diff --git a/lib/libcxx/include/cfloat b/lib/libcxx/include/cfloat index da22c6f60..77ff5261e 100644 --- a/lib/libcxx/include/cfloat +++ b/lib/libcxx/include/cfloat @@ -76,4 +76,4 @@ Macros: #pragma GCC system_header #endif -#endif // _LIBCPP_CFLOAT +#endif // _LIBCPP_CFLOAT diff --git a/lib/libcxx/include/charconv b/lib/libcxx/include/charconv index 4666c5c51..de34112d8 100644 --- a/lib/libcxx/include/charconv +++ b/lib/libcxx/include/charconv @@ -73,11 +73,13 @@ namespace std { */ -#include <__config> #include <__availability> +#include <__config> #include <__errc> +#include <__utility/to_underlying.h> #include // for log2f #include +#include // for _LIBCPP_UNREACHABLE #include #include #include @@ -108,6 +110,47 @@ enum class _LIBCPP_ENUM_VIS chars_format general = fixed | scientific }; +inline _LIBCPP_INLINE_VISIBILITY constexpr chars_format +operator~(chars_format __x) { + return chars_format(~_VSTD::__to_underlying(__x)); +} + +inline _LIBCPP_INLINE_VISIBILITY constexpr chars_format +operator&(chars_format __x, chars_format __y) { + return chars_format(_VSTD::__to_underlying(__x) & + _VSTD::__to_underlying(__y)); +} + +inline _LIBCPP_INLINE_VISIBILITY constexpr chars_format +operator|(chars_format __x, chars_format __y) { + return chars_format(_VSTD::__to_underlying(__x) | + _VSTD::__to_underlying(__y)); +} + +inline _LIBCPP_INLINE_VISIBILITY constexpr chars_format +operator^(chars_format __x, chars_format __y) { + return chars_format(_VSTD::__to_underlying(__x) ^ + _VSTD::__to_underlying(__y)); +} + +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 chars_format& +operator&=(chars_format& __x, chars_format __y) { + __x = __x & __y; + return __x; +} + +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 chars_format& +operator|=(chars_format& __x, chars_format __y) { + __x = __x | __y; + return __x; +} + +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 chars_format& +operator^=(chars_format& __x, chars_format __y) { + __x = __x ^ __y; + return __x; +} + struct _LIBCPP_TYPE_VIS to_chars_result { char* ptr; @@ -288,19 +331,12 @@ __complement(_Tp __x) return _Tp(~__x + 1); } -template -inline _LIBCPP_INLINE_VISIBILITY typename make_unsigned<_Tp>::type -__to_unsigned(_Tp __x) -{ - return static_cast::type>(__x); -} - template _LIBCPP_AVAILABILITY_TO_CHARS inline _LIBCPP_INLINE_VISIBILITY to_chars_result __to_chars_itoa(char* __first, char* __last, _Tp __value, true_type) { - auto __x = __to_unsigned(__value); + auto __x = __to_unsigned_like(__value); if (__value < 0 && __first != __last) { *__first++ = '-'; @@ -348,7 +384,7 @@ inline _LIBCPP_INLINE_VISIBILITY to_chars_result __to_chars_integral(char* __first, char* __last, _Tp __value, int __base, true_type) { - auto __x = __to_unsigned(__value); + auto __x = __to_unsigned_like(__value); if (__value < 0 && __first != __last) { *__first++ = '-'; @@ -358,33 +394,54 @@ __to_chars_integral(char* __first, char* __last, _Tp __value, int __base, return __to_chars_integral(__first, __last, __x, __base, false_type()); } +template +_LIBCPP_AVAILABILITY_TO_CHARS _LIBCPP_INLINE_VISIBILITY int __to_chars_integral_width(_Tp __value, unsigned __base) { + _LIBCPP_ASSERT(__value >= 0, "The function requires a non-negative value."); + + unsigned __base_2 = __base * __base; + unsigned __base_3 = __base_2 * __base; + unsigned __base_4 = __base_2 * __base_2; + + int __r = 0; + while (true) { + if (__value < __base) + return __r + 1; + if (__value < __base_2) + return __r + 2; + if (__value < __base_3) + return __r + 3; + if (__value < __base_4) + return __r + 4; + + __value /= __base_4; + __r += 4; + } + + _LIBCPP_UNREACHABLE(); +} + template _LIBCPP_AVAILABILITY_TO_CHARS inline _LIBCPP_INLINE_VISIBILITY to_chars_result __to_chars_integral(char* __first, char* __last, _Tp __value, int __base, false_type) { - if (__base == 10) - return __to_chars_itoa(__first, __last, __value, false_type()); - - auto __p = __last; - while (__p != __first) - { - auto __c = __value % __base; - __value /= __base; - *--__p = "0123456789abcdefghijklmnopqrstuvwxyz"[__c]; - if (__value == 0) - break; - } - - auto __len = __last - __p; - if (__value != 0 || !__len) - return {__last, errc::value_too_large}; - else - { - _VSTD::memmove(__first, __p, __len); - return {__first + __len, {}}; - } + if (__base == 10) + return __to_chars_itoa(__first, __last, __value, false_type()); + + ptrdiff_t __cap = __last - __first; + int __n = __to_chars_integral_width(__value, __base); + if (__n > __cap) + return {__last, errc::value_too_large}; + + __last = __first + __n; + char* __p = __last; + do { + unsigned __c = __value % __base; + __value /= __base; + *--__p = "0123456789abcdefghijklmnopqrstuvwxyz"[__c]; + } while (__value != 0); + return {__last, errc(0)}; } template ::value, int>::type = 0> @@ -410,7 +467,7 @@ inline _LIBCPP_INLINE_VISIBILITY from_chars_result __sign_combinator(_It __first, _It __last, _Tp& __value, _Fn __f, _Ts... __args) { using __tl = numeric_limits<_Tp>; - decltype(__to_unsigned(__value)) __x; + decltype(__to_unsigned_like(__value)) __x; bool __neg = (__first != __last && *__first == '-'); auto __r = __f(__neg ? __first + 1 : __first, __last, __x, __args...); @@ -426,7 +483,7 @@ __sign_combinator(_It __first, _It __last, _Tp& __value, _Fn __f, _Ts... __args) if (__neg) { - if (__x <= __complement(__to_unsigned(__tl::min()))) + if (__x <= __complement(__to_unsigned_like(__tl::min()))) { __x = __complement(__x); _VSTD::memcpy(&__value, &__x, sizeof(__x)); @@ -541,7 +598,7 @@ template ::value, int>::type = 0 inline _LIBCPP_INLINE_VISIBILITY from_chars_result __from_chars_atoi(const char* __first, const char* __last, _Tp& __value) { - using __t = decltype(__to_unsigned(__value)); + using __t = decltype(__to_unsigned_like(__value)); return __sign_combinator(__first, __last, __value, __from_chars_atoi<__t>); } @@ -555,13 +612,13 @@ __from_chars_integral(const char* __first, const char* __last, _Tp& __value, return __subject_seq_combinator( __first, __last, __value, - [](const char* __p, const char* __last, _Tp& __value, + [](const char* __p, const char* __lastx, _Tp& __value, int __base) -> from_chars_result { using __tl = numeric_limits<_Tp>; auto __digits = __tl::digits / log2f(float(__base)); _Tp __a = __in_pattern(*__p++, __base).__val, __b = 0; - for (int __i = 1; __p != __last; ++__i, ++__p) + for (int __i = 1; __p != __lastx; ++__i, ++__p) { if (auto __c = __in_pattern(*__p, __base)) { @@ -579,7 +636,7 @@ __from_chars_integral(const char* __first, const char* __last, _Tp& __value, break; } - if (__p == __last || !__in_pattern(*__p, __base)) + if (__p == __lastx || !__in_pattern(*__p, __base)) { if (__tl::max() - __a >= __b) { @@ -597,7 +654,7 @@ inline _LIBCPP_INLINE_VISIBILITY from_chars_result __from_chars_integral(const char* __first, const char* __last, _Tp& __value, int __base) { - using __t = decltype(__to_unsigned(__value)); + using __t = decltype(__to_unsigned_like(__value)); return __sign_combinator(__first, __last, __value, __from_chars_integral<__t>, __base); } @@ -617,10 +674,10 @@ from_chars(const char* __first, const char* __last, _Tp& __value, int __base) return __from_chars_integral(__first, __last, __value, __base); } -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG _LIBCPP_END_NAMESPACE_STD _LIBCPP_POP_MACROS -#endif // _LIBCPP_CHARCONV +#endif // _LIBCPP_CHARCONV diff --git a/lib/libcxx/include/chrono b/lib/libcxx/include/chrono index 53e454601..6bdb92912 100644 --- a/lib/libcxx/include/chrono +++ b/lib/libcxx/include/chrono @@ -823,12 +823,13 @@ constexpr chrono::year operator ""y(unsigned lo } // std */ -#include <__config> #include <__availability> +#include <__config> +#include #include -#include -#include #include +#include +#include #include #if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) @@ -1091,7 +1092,7 @@ public: (__no_overflow<_Period2, period>::type::den == 1 && !treat_as_floating_point<_Rep2>::value)) >::type* = nullptr) - : __rep_(_VSTD::chrono::duration_cast(__d).count()) {} + : __rep_(chrono::duration_cast(__d).count()) {} // observer @@ -1410,7 +1411,7 @@ inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 time_point<_Clock, _ToDuration> time_point_cast(const time_point<_Clock, _Duration>& __t) { - return time_point<_Clock, _ToDuration>(_VSTD::chrono::duration_cast<_ToDuration>(__t.time_since_epoch())); + return time_point<_Clock, _ToDuration>(chrono::duration_cast<_ToDuration>(__t.time_since_epoch())); } #if _LIBCPP_STD_VER > 14 @@ -1926,13 +1927,13 @@ inline constexpr weekday& weekday::operator-=(const days& __dd) noexcept class weekday_indexed { private: - _VSTD::chrono::weekday __wd; + chrono::weekday __wd; unsigned char __idx; public: weekday_indexed() = default; - inline constexpr weekday_indexed(const _VSTD::chrono::weekday& __wdval, unsigned __idxval) noexcept + inline constexpr weekday_indexed(const chrono::weekday& __wdval, unsigned __idxval) noexcept : __wd{__wdval}, __idx(__idxval) {} - inline constexpr _VSTD::chrono::weekday weekday() const noexcept { return __wd; } + inline constexpr chrono::weekday weekday() const noexcept { return __wd; } inline constexpr unsigned index() const noexcept { return __idx; } inline constexpr bool ok() const noexcept { return __wd.ok() && __idx >= 1 && __idx <= 5; } }; @@ -1948,11 +1949,11 @@ bool operator!=(const weekday_indexed& __lhs, const weekday_indexed& __rhs) noex class weekday_last { private: - _VSTD::chrono::weekday __wd; + chrono::weekday __wd; public: - explicit constexpr weekday_last(const _VSTD::chrono::weekday& __val) noexcept + explicit constexpr weekday_last(const chrono::weekday& __val) noexcept : __wd{__val} {} - constexpr _VSTD::chrono::weekday weekday() const noexcept { return __wd; } + constexpr chrono::weekday weekday() const noexcept { return __wd; } constexpr bool ok() const noexcept { return __wd.ok(); } }; @@ -2308,8 +2309,8 @@ inline constexpr year_month_day year_month_day::__from_days(days __d) noexcept { - static_assert(std::numeric_limits::digits >= 18, ""); - static_assert(std::numeric_limits::digits >= 20 , ""); + static_assert(numeric_limits::digits >= 18, ""); + static_assert(numeric_limits::digits >= 20 , ""); const int __z = __d.count() + 719468; const int __era = (__z >= 0 ? __z : __z - 146096) / 146097; const unsigned __doe = static_cast(__z - __era * 146097); // [0, 146096] @@ -2325,8 +2326,8 @@ year_month_day::__from_days(days __d) noexcept // https://howardhinnant.github.io/date_algorithms.html#days_from_civil inline constexpr days year_month_day::__to_days() const noexcept { - static_assert(std::numeric_limits::digits >= 18, ""); - static_assert(std::numeric_limits::digits >= 20 , ""); + static_assert(numeric_limits::digits >= 18, ""); + static_assert(numeric_limits::digits >= 20 , ""); const int __yr = static_cast(__y) - (__m <= February); const unsigned __mth = static_cast(__m); @@ -2964,4 +2965,4 @@ _LIBCPP_END_NAMESPACE_FILESYSTEM _LIBCPP_POP_MACROS -#endif // _LIBCPP_CHRONO +#endif // _LIBCPP_CHRONO diff --git a/lib/libcxx/include/cinttypes b/lib/libcxx/include/cinttypes index 55af85cc3..7ce4a8ad5 100644 --- a/lib/libcxx/include/cinttypes +++ b/lib/libcxx/include/cinttypes @@ -244,14 +244,14 @@ uintmax_t wcstoumax(const wchar_t* restrict nptr, wchar_t** restrict endptr, int _LIBCPP_BEGIN_NAMESPACE_STD -using::imaxdiv_t; -using::imaxabs; -using::imaxdiv; -using::strtoimax; -using::strtoumax; -using::wcstoimax; -using::wcstoumax; +using ::imaxdiv_t _LIBCPP_USING_IF_EXISTS; +using ::imaxabs _LIBCPP_USING_IF_EXISTS; +using ::imaxdiv _LIBCPP_USING_IF_EXISTS; +using ::strtoimax _LIBCPP_USING_IF_EXISTS; +using ::strtoumax _LIBCPP_USING_IF_EXISTS; +using ::wcstoimax _LIBCPP_USING_IF_EXISTS; +using ::wcstoumax _LIBCPP_USING_IF_EXISTS; _LIBCPP_END_NAMESPACE_STD -#endif // _LIBCPP_CINTTYPES +#endif // _LIBCPP_CINTTYPES diff --git a/lib/libcxx/include/ciso646 b/lib/libcxx/include/ciso646 index 172f1676a..c37f63797 100644 --- a/lib/libcxx/include/ciso646 +++ b/lib/libcxx/include/ciso646 @@ -21,4 +21,4 @@ #pragma GCC system_header #endif -#endif // _LIBCPP_CISO646 +#endif // _LIBCPP_CISO646 diff --git a/lib/libcxx/include/climits b/lib/libcxx/include/climits index 43eb2d3f1..217ec6286 100644 --- a/lib/libcxx/include/climits +++ b/lib/libcxx/include/climits @@ -44,4 +44,4 @@ Macros: #pragma GCC system_header #endif -#endif // _LIBCPP_CLIMITS +#endif // _LIBCPP_CLIMITS diff --git a/lib/libcxx/include/clocale b/lib/libcxx/include/clocale index bff4e92f9..8fe91fc5a 100644 --- a/lib/libcxx/include/clocale +++ b/lib/libcxx/include/clocale @@ -43,12 +43,12 @@ lconv* localeconv(); _LIBCPP_BEGIN_NAMESPACE_STD -using ::lconv; +using ::lconv _LIBCPP_USING_IF_EXISTS; #ifndef _LIBCPP_HAS_NO_THREAD_UNSAFE_C_FUNCTIONS -using ::setlocale; +using ::setlocale _LIBCPP_USING_IF_EXISTS; #endif -using ::localeconv; +using ::localeconv _LIBCPP_USING_IF_EXISTS; _LIBCPP_END_NAMESPACE_STD -#endif // _LIBCPP_CLOCALE +#endif // _LIBCPP_CLOCALE diff --git a/lib/libcxx/include/cmath b/lib/libcxx/include/cmath index 138ae6f99..adf83c2b0 100644 --- a/lib/libcxx/include/cmath +++ b/lib/libcxx/include/cmath @@ -318,217 +318,215 @@ _LIBCPP_PUSH_MACROS _LIBCPP_BEGIN_NAMESPACE_STD -using ::signbit; -using ::fpclassify; -using ::isfinite; -using ::isinf; -using ::isnan; -using ::isnormal; -using ::isgreater; -using ::isgreaterequal; -using ::isless; -using ::islessequal; -using ::islessgreater; -using ::isunordered; -using ::isunordered; - -using ::float_t; -using ::double_t; - -#ifndef _AIX -using ::abs; -#endif - -using ::acos; -using ::acosf; -using ::asin; -using ::asinf; -using ::atan; -using ::atanf; -using ::atan2; -using ::atan2f; -using ::ceil; -using ::ceilf; -using ::cos; -using ::cosf; -using ::cosh; -using ::coshf; - -using ::exp; -using ::expf; - -using ::fabs; -using ::fabsf; -using ::floor; -using ::floorf; - -using ::fmod; -using ::fmodf; - -using ::frexp; -using ::frexpf; -using ::ldexp; -using ::ldexpf; - -using ::log; -using ::logf; - -using ::log10; -using ::log10f; -using ::modf; -using ::modff; - -using ::pow; -using ::powf; - -using ::sin; -using ::sinf; -using ::sinh; -using ::sinhf; - -using ::sqrt; -using ::sqrtf; -using ::tan; -using ::tanf; - -using ::tanh; -using ::tanhf; - -using ::acosh; -using ::acoshf; -using ::asinh; -using ::asinhf; -using ::atanh; -using ::atanhf; -using ::cbrt; -using ::cbrtf; - -using ::copysign; -using ::copysignf; - -using ::erf; -using ::erff; -using ::erfc; -using ::erfcf; -using ::exp2; -using ::exp2f; -using ::expm1; -using ::expm1f; -using ::fdim; -using ::fdimf; -using ::fmaf; -using ::fma; -using ::fmax; -using ::fmaxf; -using ::fmin; -using ::fminf; -using ::hypot; -using ::hypotf; -using ::ilogb; -using ::ilogbf; -using ::lgamma; -using ::lgammaf; -using ::llrint; -using ::llrintf; -using ::llround; -using ::llroundf; -using ::log1p; -using ::log1pf; -using ::log2; -using ::log2f; -using ::logb; -using ::logbf; -using ::lrint; -using ::lrintf; -using ::lround; -using ::lroundf; - -using ::nan; -using ::nanf; - -using ::nearbyint; -using ::nearbyintf; -using ::nextafter; -using ::nextafterf; -using ::nexttoward; -using ::nexttowardf; -using ::remainder; -using ::remainderf; -using ::remquo; -using ::remquof; -using ::rint; -using ::rintf; -using ::round; -using ::roundf; -using ::scalbln; -using ::scalblnf; -using ::scalbn; -using ::scalbnf; -using ::tgamma; -using ::tgammaf; -using ::trunc; -using ::truncf; - -using ::acosl; -using ::asinl; -using ::atanl; -using ::atan2l; -using ::ceill; -using ::cosl; -using ::coshl; -using ::expl; -using ::fabsl; -using ::floorl; -using ::fmodl; -using ::frexpl; -using ::ldexpl; -using ::logl; -using ::log10l; -using ::modfl; -using ::powl; -using ::sinl; -using ::sinhl; -using ::sqrtl; -using ::tanl; - -using ::tanhl; -using ::acoshl; -using ::asinhl; -using ::atanhl; -using ::cbrtl; - -using ::copysignl; - -using ::erfl; -using ::erfcl; -using ::exp2l; -using ::expm1l; -using ::fdiml; -using ::fmal; -using ::fmaxl; -using ::fminl; -using ::hypotl; -using ::ilogbl; -using ::lgammal; -using ::llrintl; -using ::llroundl; -using ::log1pl; -using ::log2l; -using ::logbl; -using ::lrintl; -using ::lroundl; -using ::nanl; -using ::nearbyintl; -using ::nextafterl; -using ::nexttowardl; -using ::remainderl; -using ::remquol; -using ::rintl; -using ::roundl; -using ::scalblnl; -using ::scalbnl; -using ::tgammal; -using ::truncl; +using ::signbit _LIBCPP_USING_IF_EXISTS; +using ::fpclassify _LIBCPP_USING_IF_EXISTS; +using ::isfinite _LIBCPP_USING_IF_EXISTS; +using ::isinf _LIBCPP_USING_IF_EXISTS; +using ::isnan _LIBCPP_USING_IF_EXISTS; +using ::isnormal _LIBCPP_USING_IF_EXISTS; +using ::isgreater _LIBCPP_USING_IF_EXISTS; +using ::isgreaterequal _LIBCPP_USING_IF_EXISTS; +using ::isless _LIBCPP_USING_IF_EXISTS; +using ::islessequal _LIBCPP_USING_IF_EXISTS; +using ::islessgreater _LIBCPP_USING_IF_EXISTS; +using ::isunordered _LIBCPP_USING_IF_EXISTS; +using ::isunordered _LIBCPP_USING_IF_EXISTS; + +using ::float_t _LIBCPP_USING_IF_EXISTS; +using ::double_t _LIBCPP_USING_IF_EXISTS; + +using ::abs _LIBCPP_USING_IF_EXISTS; + +using ::acos _LIBCPP_USING_IF_EXISTS; +using ::acosf _LIBCPP_USING_IF_EXISTS; +using ::asin _LIBCPP_USING_IF_EXISTS; +using ::asinf _LIBCPP_USING_IF_EXISTS; +using ::atan _LIBCPP_USING_IF_EXISTS; +using ::atanf _LIBCPP_USING_IF_EXISTS; +using ::atan2 _LIBCPP_USING_IF_EXISTS; +using ::atan2f _LIBCPP_USING_IF_EXISTS; +using ::ceil _LIBCPP_USING_IF_EXISTS; +using ::ceilf _LIBCPP_USING_IF_EXISTS; +using ::cos _LIBCPP_USING_IF_EXISTS; +using ::cosf _LIBCPP_USING_IF_EXISTS; +using ::cosh _LIBCPP_USING_IF_EXISTS; +using ::coshf _LIBCPP_USING_IF_EXISTS; + +using ::exp _LIBCPP_USING_IF_EXISTS; +using ::expf _LIBCPP_USING_IF_EXISTS; + +using ::fabs _LIBCPP_USING_IF_EXISTS; +using ::fabsf _LIBCPP_USING_IF_EXISTS; +using ::floor _LIBCPP_USING_IF_EXISTS; +using ::floorf _LIBCPP_USING_IF_EXISTS; + +using ::fmod _LIBCPP_USING_IF_EXISTS; +using ::fmodf _LIBCPP_USING_IF_EXISTS; + +using ::frexp _LIBCPP_USING_IF_EXISTS; +using ::frexpf _LIBCPP_USING_IF_EXISTS; +using ::ldexp _LIBCPP_USING_IF_EXISTS; +using ::ldexpf _LIBCPP_USING_IF_EXISTS; + +using ::log _LIBCPP_USING_IF_EXISTS; +using ::logf _LIBCPP_USING_IF_EXISTS; + +using ::log10 _LIBCPP_USING_IF_EXISTS; +using ::log10f _LIBCPP_USING_IF_EXISTS; +using ::modf _LIBCPP_USING_IF_EXISTS; +using ::modff _LIBCPP_USING_IF_EXISTS; + +using ::pow _LIBCPP_USING_IF_EXISTS; +using ::powf _LIBCPP_USING_IF_EXISTS; + +using ::sin _LIBCPP_USING_IF_EXISTS; +using ::sinf _LIBCPP_USING_IF_EXISTS; +using ::sinh _LIBCPP_USING_IF_EXISTS; +using ::sinhf _LIBCPP_USING_IF_EXISTS; + +using ::sqrt _LIBCPP_USING_IF_EXISTS; +using ::sqrtf _LIBCPP_USING_IF_EXISTS; +using ::tan _LIBCPP_USING_IF_EXISTS; +using ::tanf _LIBCPP_USING_IF_EXISTS; + +using ::tanh _LIBCPP_USING_IF_EXISTS; +using ::tanhf _LIBCPP_USING_IF_EXISTS; + +using ::acosh _LIBCPP_USING_IF_EXISTS; +using ::acoshf _LIBCPP_USING_IF_EXISTS; +using ::asinh _LIBCPP_USING_IF_EXISTS; +using ::asinhf _LIBCPP_USING_IF_EXISTS; +using ::atanh _LIBCPP_USING_IF_EXISTS; +using ::atanhf _LIBCPP_USING_IF_EXISTS; +using ::cbrt _LIBCPP_USING_IF_EXISTS; +using ::cbrtf _LIBCPP_USING_IF_EXISTS; + +using ::copysign _LIBCPP_USING_IF_EXISTS; +using ::copysignf _LIBCPP_USING_IF_EXISTS; + +using ::erf _LIBCPP_USING_IF_EXISTS; +using ::erff _LIBCPP_USING_IF_EXISTS; +using ::erfc _LIBCPP_USING_IF_EXISTS; +using ::erfcf _LIBCPP_USING_IF_EXISTS; +using ::exp2 _LIBCPP_USING_IF_EXISTS; +using ::exp2f _LIBCPP_USING_IF_EXISTS; +using ::expm1 _LIBCPP_USING_IF_EXISTS; +using ::expm1f _LIBCPP_USING_IF_EXISTS; +using ::fdim _LIBCPP_USING_IF_EXISTS; +using ::fdimf _LIBCPP_USING_IF_EXISTS; +using ::fmaf _LIBCPP_USING_IF_EXISTS; +using ::fma _LIBCPP_USING_IF_EXISTS; +using ::fmax _LIBCPP_USING_IF_EXISTS; +using ::fmaxf _LIBCPP_USING_IF_EXISTS; +using ::fmin _LIBCPP_USING_IF_EXISTS; +using ::fminf _LIBCPP_USING_IF_EXISTS; +using ::hypot _LIBCPP_USING_IF_EXISTS; +using ::hypotf _LIBCPP_USING_IF_EXISTS; +using ::ilogb _LIBCPP_USING_IF_EXISTS; +using ::ilogbf _LIBCPP_USING_IF_EXISTS; +using ::lgamma _LIBCPP_USING_IF_EXISTS; +using ::lgammaf _LIBCPP_USING_IF_EXISTS; +using ::llrint _LIBCPP_USING_IF_EXISTS; +using ::llrintf _LIBCPP_USING_IF_EXISTS; +using ::llround _LIBCPP_USING_IF_EXISTS; +using ::llroundf _LIBCPP_USING_IF_EXISTS; +using ::log1p _LIBCPP_USING_IF_EXISTS; +using ::log1pf _LIBCPP_USING_IF_EXISTS; +using ::log2 _LIBCPP_USING_IF_EXISTS; +using ::log2f _LIBCPP_USING_IF_EXISTS; +using ::logb _LIBCPP_USING_IF_EXISTS; +using ::logbf _LIBCPP_USING_IF_EXISTS; +using ::lrint _LIBCPP_USING_IF_EXISTS; +using ::lrintf _LIBCPP_USING_IF_EXISTS; +using ::lround _LIBCPP_USING_IF_EXISTS; +using ::lroundf _LIBCPP_USING_IF_EXISTS; + +using ::nan _LIBCPP_USING_IF_EXISTS; +using ::nanf _LIBCPP_USING_IF_EXISTS; + +using ::nearbyint _LIBCPP_USING_IF_EXISTS; +using ::nearbyintf _LIBCPP_USING_IF_EXISTS; +using ::nextafter _LIBCPP_USING_IF_EXISTS; +using ::nextafterf _LIBCPP_USING_IF_EXISTS; +using ::nexttoward _LIBCPP_USING_IF_EXISTS; +using ::nexttowardf _LIBCPP_USING_IF_EXISTS; +using ::remainder _LIBCPP_USING_IF_EXISTS; +using ::remainderf _LIBCPP_USING_IF_EXISTS; +using ::remquo _LIBCPP_USING_IF_EXISTS; +using ::remquof _LIBCPP_USING_IF_EXISTS; +using ::rint _LIBCPP_USING_IF_EXISTS; +using ::rintf _LIBCPP_USING_IF_EXISTS; +using ::round _LIBCPP_USING_IF_EXISTS; +using ::roundf _LIBCPP_USING_IF_EXISTS; +using ::scalbln _LIBCPP_USING_IF_EXISTS; +using ::scalblnf _LIBCPP_USING_IF_EXISTS; +using ::scalbn _LIBCPP_USING_IF_EXISTS; +using ::scalbnf _LIBCPP_USING_IF_EXISTS; +using ::tgamma _LIBCPP_USING_IF_EXISTS; +using ::tgammaf _LIBCPP_USING_IF_EXISTS; +using ::trunc _LIBCPP_USING_IF_EXISTS; +using ::truncf _LIBCPP_USING_IF_EXISTS; + +using ::acosl _LIBCPP_USING_IF_EXISTS; +using ::asinl _LIBCPP_USING_IF_EXISTS; +using ::atanl _LIBCPP_USING_IF_EXISTS; +using ::atan2l _LIBCPP_USING_IF_EXISTS; +using ::ceill _LIBCPP_USING_IF_EXISTS; +using ::cosl _LIBCPP_USING_IF_EXISTS; +using ::coshl _LIBCPP_USING_IF_EXISTS; +using ::expl _LIBCPP_USING_IF_EXISTS; +using ::fabsl _LIBCPP_USING_IF_EXISTS; +using ::floorl _LIBCPP_USING_IF_EXISTS; +using ::fmodl _LIBCPP_USING_IF_EXISTS; +using ::frexpl _LIBCPP_USING_IF_EXISTS; +using ::ldexpl _LIBCPP_USING_IF_EXISTS; +using ::logl _LIBCPP_USING_IF_EXISTS; +using ::log10l _LIBCPP_USING_IF_EXISTS; +using ::modfl _LIBCPP_USING_IF_EXISTS; +using ::powl _LIBCPP_USING_IF_EXISTS; +using ::sinl _LIBCPP_USING_IF_EXISTS; +using ::sinhl _LIBCPP_USING_IF_EXISTS; +using ::sqrtl _LIBCPP_USING_IF_EXISTS; +using ::tanl _LIBCPP_USING_IF_EXISTS; + +using ::tanhl _LIBCPP_USING_IF_EXISTS; +using ::acoshl _LIBCPP_USING_IF_EXISTS; +using ::asinhl _LIBCPP_USING_IF_EXISTS; +using ::atanhl _LIBCPP_USING_IF_EXISTS; +using ::cbrtl _LIBCPP_USING_IF_EXISTS; + +using ::copysignl _LIBCPP_USING_IF_EXISTS; + +using ::erfl _LIBCPP_USING_IF_EXISTS; +using ::erfcl _LIBCPP_USING_IF_EXISTS; +using ::exp2l _LIBCPP_USING_IF_EXISTS; +using ::expm1l _LIBCPP_USING_IF_EXISTS; +using ::fdiml _LIBCPP_USING_IF_EXISTS; +using ::fmal _LIBCPP_USING_IF_EXISTS; +using ::fmaxl _LIBCPP_USING_IF_EXISTS; +using ::fminl _LIBCPP_USING_IF_EXISTS; +using ::hypotl _LIBCPP_USING_IF_EXISTS; +using ::ilogbl _LIBCPP_USING_IF_EXISTS; +using ::lgammal _LIBCPP_USING_IF_EXISTS; +using ::llrintl _LIBCPP_USING_IF_EXISTS; +using ::llroundl _LIBCPP_USING_IF_EXISTS; +using ::log1pl _LIBCPP_USING_IF_EXISTS; +using ::log2l _LIBCPP_USING_IF_EXISTS; +using ::logbl _LIBCPP_USING_IF_EXISTS; +using ::lrintl _LIBCPP_USING_IF_EXISTS; +using ::lroundl _LIBCPP_USING_IF_EXISTS; +using ::nanl _LIBCPP_USING_IF_EXISTS; +using ::nearbyintl _LIBCPP_USING_IF_EXISTS; +using ::nextafterl _LIBCPP_USING_IF_EXISTS; +using ::nexttowardl _LIBCPP_USING_IF_EXISTS; +using ::remainderl _LIBCPP_USING_IF_EXISTS; +using ::remquol _LIBCPP_USING_IF_EXISTS; +using ::rintl _LIBCPP_USING_IF_EXISTS; +using ::roundl _LIBCPP_USING_IF_EXISTS; +using ::scalblnl _LIBCPP_USING_IF_EXISTS; +using ::scalbnl _LIBCPP_USING_IF_EXISTS; +using ::tgammal _LIBCPP_USING_IF_EXISTS; +using ::truncl _LIBCPP_USING_IF_EXISTS; #if _LIBCPP_STD_VER > 14 inline _LIBCPP_INLINE_VISIBILITY float hypot( float x, float y, float z ) { return sqrt(x*x + y*y + z*z); } @@ -623,10 +621,10 @@ _Fp __lerp(_Fp __a, _Fp __b, _Fp __t) noexcept { if (__t == 1) return __b; const _Fp __x = __a + __t * (__b - __a); - if (__t > 1 == __b > __a) - return __b < __x ? __x : __b; + if ((__t > 1) == (__b > __a)) + return __b < __x ? __x : __b; else - return __x < __b ? __x : __b; + return __x < __b ? __x : __b; } constexpr float @@ -674,4 +672,4 @@ _LIBCPP_END_NAMESPACE_STD _LIBCPP_POP_MACROS -#endif // _LIBCPP_CMATH +#endif // _LIBCPP_CMATH diff --git a/lib/libcxx/include/codecvt b/lib/libcxx/include/codecvt index 2befa1b0a..ec7d4a780 100644 --- a/lib/libcxx/include/codecvt +++ b/lib/libcxx/include/codecvt @@ -570,4 +570,4 @@ public: _LIBCPP_END_NAMESPACE_STD -#endif // _LIBCPP_CODECVT +#endif // _LIBCPP_CODECVT diff --git a/lib/libcxx/include/compare b/lib/libcxx/include/compare index 048f4821d..e924bffed 100644 --- a/lib/libcxx/include/compare +++ b/lib/libcxx/include/compare @@ -15,15 +15,13 @@ namespace std { // [cmp.categories], comparison category types - class weak_equality; - class strong_equality; class partial_ordering; class weak_ordering; class strong_ordering; // named comparison functions - constexpr bool is_eq (weak_equality cmp) noexcept { return cmp == 0; } - constexpr bool is_neq (weak_equality cmp) noexcept { return cmp != 0; } + constexpr bool is_eq (partial_ordering cmp) noexcept { return cmp == 0; } + constexpr bool is_neq (partial_ordering cmp) noexcept { return cmp != 0; } constexpr bool is_lt (partial_ordering cmp) noexcept { return cmp < 0; } constexpr bool is_lteq(partial_ordering cmp) noexcept { return cmp <= 0; } constexpr bool is_gt (partial_ordering cmp) noexcept { return cmp > 0; } @@ -41,8 +39,6 @@ namespace std { template constexpr strong_ordering strong_order(const T& a, const T& b); template constexpr weak_ordering weak_order(const T& a, const T& b); template constexpr partial_ordering partial_order(const T& a, const T& b); - template constexpr strong_equality strong_equal(const T& a, const T& b); - template constexpr weak_equality weak_equal(const T& a, const T& b); // [cmp.partialord], Class partial_ordering class partial_ordering { @@ -126,7 +122,6 @@ namespace std { #include <__config> #include -#include #ifndef _LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER #pragma GCC system_header @@ -134,8 +129,7 @@ namespace std { _LIBCPP_BEGIN_NAMESPACE_STD -#if _LIBCPP_STD_VER > 17 - +#if _LIBCPP_STD_VER > 17 && !defined(_LIBCPP_HAS_NO_SPACESHIP_OPERATOR) // exposition only enum class _LIBCPP_ENUM_VIS _EqResult : unsigned char { __zero = 0, @@ -154,138 +148,21 @@ enum class _LIBCPP_ENUM_VIS _NCmpResult : signed char { __unordered = -127 }; +class partial_ordering; +class weak_ordering; +class strong_ordering; + +template +inline constexpr bool __one_of_v = (is_same_v<_Tp, _Args> || ...); + struct _CmpUnspecifiedParam { _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEVAL _CmpUnspecifiedParam(int _CmpUnspecifiedParam::*) noexcept {} - template>> + template>> _CmpUnspecifiedParam(_Tp) = delete; }; -class weak_equality { - _LIBCPP_INLINE_VISIBILITY - constexpr explicit weak_equality(_EqResult __val) noexcept : __value_(__val) {} - -public: - static const weak_equality equivalent; - static const weak_equality nonequivalent; - - _LIBCPP_INLINE_VISIBILITY friend constexpr bool operator==(weak_equality __v, _CmpUnspecifiedParam) noexcept; - _LIBCPP_INLINE_VISIBILITY friend constexpr bool operator==(_CmpUnspecifiedParam, weak_equality __v) noexcept; - _LIBCPP_INLINE_VISIBILITY friend constexpr bool operator!=(weak_equality __v, _CmpUnspecifiedParam) noexcept; - _LIBCPP_INLINE_VISIBILITY friend constexpr bool operator!=(_CmpUnspecifiedParam, weak_equality __v) noexcept; - -#ifndef _LIBCPP_HAS_NO_SPACESHIP_OPERATOR - _LIBCPP_INLINE_VISIBILITY friend constexpr weak_equality operator<=>(weak_equality __v, _CmpUnspecifiedParam) noexcept; - _LIBCPP_INLINE_VISIBILITY friend constexpr weak_equality operator<=>(_CmpUnspecifiedParam, weak_equality __v) noexcept; -#endif - -private: - _EqResult __value_; -}; - -_LIBCPP_INLINE_VAR constexpr weak_equality weak_equality::equivalent(_EqResult::__equiv); -_LIBCPP_INLINE_VAR constexpr weak_equality weak_equality::nonequivalent(_EqResult::__nonequiv); - -_LIBCPP_INLINE_VISIBILITY -inline constexpr bool operator==(weak_equality __v, _CmpUnspecifiedParam) noexcept { - return __v.__value_ == _EqResult::__zero; -} - -_LIBCPP_INLINE_VISIBILITY -inline constexpr bool operator==(_CmpUnspecifiedParam, weak_equality __v) noexcept { - return __v.__value_ == _EqResult::__zero; -} - -_LIBCPP_INLINE_VISIBILITY -inline constexpr bool operator!=(weak_equality __v, _CmpUnspecifiedParam) noexcept { - return __v.__value_ != _EqResult::__zero; -} - -_LIBCPP_INLINE_VISIBILITY -inline constexpr bool operator!=(_CmpUnspecifiedParam, weak_equality __v) noexcept { - return __v.__value_ != _EqResult::__zero; -} - -#ifndef _LIBCPP_HAS_NO_SPACESHIP_OPERATOR -_LIBCPP_INLINE_VISIBILITY -inline constexpr weak_equality operator<=>(weak_equality __v, _CmpUnspecifiedParam) noexcept { - return __v; -} - -_LIBCPP_INLINE_VISIBILITY -inline constexpr weak_equality operator<=>(_CmpUnspecifiedParam, weak_equality __v) noexcept { - return __v; -} -#endif - -class strong_equality { - _LIBCPP_INLINE_VISIBILITY - explicit constexpr strong_equality(_EqResult __val) noexcept : __value_(__val) {} - -public: - static const strong_equality equal; - static const strong_equality nonequal; - static const strong_equality equivalent; - static const strong_equality nonequivalent; - - // conversion - _LIBCPP_INLINE_VISIBILITY constexpr operator weak_equality() const noexcept { - return __value_ == _EqResult::__zero ? weak_equality::equivalent - : weak_equality::nonequivalent; - } - - // comparisons - _LIBCPP_INLINE_VISIBILITY friend constexpr bool operator==(strong_equality __v, _CmpUnspecifiedParam) noexcept; - _LIBCPP_INLINE_VISIBILITY friend constexpr bool operator!=(strong_equality __v, _CmpUnspecifiedParam) noexcept; - _LIBCPP_INLINE_VISIBILITY friend constexpr bool operator==(_CmpUnspecifiedParam, strong_equality __v) noexcept; - _LIBCPP_INLINE_VISIBILITY friend constexpr bool operator!=(_CmpUnspecifiedParam, strong_equality __v) noexcept; - -#ifndef _LIBCPP_HAS_NO_SPACESHIP_OPERATOR - _LIBCPP_INLINE_VISIBILITY friend constexpr strong_equality operator<=>(strong_equality __v, _CmpUnspecifiedParam) noexcept; - _LIBCPP_INLINE_VISIBILITY friend constexpr strong_equality operator<=>(_CmpUnspecifiedParam, strong_equality __v) noexcept; -#endif -private: - _EqResult __value_; -}; - -_LIBCPP_INLINE_VAR constexpr strong_equality strong_equality::equal(_EqResult::__equal); -_LIBCPP_INLINE_VAR constexpr strong_equality strong_equality::nonequal(_EqResult::__nonequal); -_LIBCPP_INLINE_VAR constexpr strong_equality strong_equality::equivalent(_EqResult::__equiv); -_LIBCPP_INLINE_VAR constexpr strong_equality strong_equality::nonequivalent(_EqResult::__nonequiv); - -_LIBCPP_INLINE_VISIBILITY -constexpr bool operator==(strong_equality __v, _CmpUnspecifiedParam) noexcept { - return __v.__value_ == _EqResult::__zero; -} - -_LIBCPP_INLINE_VISIBILITY -constexpr bool operator==(_CmpUnspecifiedParam, strong_equality __v) noexcept { - return __v.__value_ == _EqResult::__zero; -} - -_LIBCPP_INLINE_VISIBILITY -constexpr bool operator!=(strong_equality __v, _CmpUnspecifiedParam) noexcept { - return __v.__value_ != _EqResult::__zero; -} - -_LIBCPP_INLINE_VISIBILITY -constexpr bool operator!=(_CmpUnspecifiedParam, strong_equality __v) noexcept { - return __v.__value_ != _EqResult::__zero; -} - -#ifndef _LIBCPP_HAS_NO_SPACESHIP_OPERATOR -_LIBCPP_INLINE_VISIBILITY -constexpr strong_equality operator<=>(strong_equality __v, _CmpUnspecifiedParam) noexcept { - return __v; -} - -_LIBCPP_INLINE_VISIBILITY -constexpr strong_equality operator<=>(_CmpUnspecifiedParam, strong_equality __v) noexcept { - return __v; -} -#endif // _LIBCPP_HAS_NO_SPACESHIP_OPERATOR - class partial_ordering { using _ValueT = signed char; @@ -311,32 +188,52 @@ public: static const partial_ordering greater; static const partial_ordering unordered; - // conversion - constexpr operator weak_equality() const noexcept { - return __value_ == 0 ? weak_equality::equivalent : weak_equality::nonequivalent; - } - // comparisons - _LIBCPP_INLINE_VISIBILITY friend constexpr bool operator==(partial_ordering __v, _CmpUnspecifiedParam) noexcept; - _LIBCPP_INLINE_VISIBILITY friend constexpr bool operator!=(partial_ordering __v, _CmpUnspecifiedParam) noexcept; - _LIBCPP_INLINE_VISIBILITY friend constexpr bool operator< (partial_ordering __v, _CmpUnspecifiedParam) noexcept; - _LIBCPP_INLINE_VISIBILITY friend constexpr bool operator<=(partial_ordering __v, _CmpUnspecifiedParam) noexcept; - _LIBCPP_INLINE_VISIBILITY friend constexpr bool operator> (partial_ordering __v, _CmpUnspecifiedParam) noexcept; - _LIBCPP_INLINE_VISIBILITY friend constexpr bool operator>=(partial_ordering __v, _CmpUnspecifiedParam) noexcept; - _LIBCPP_INLINE_VISIBILITY friend constexpr bool operator==(_CmpUnspecifiedParam, partial_ordering __v) noexcept; - _LIBCPP_INLINE_VISIBILITY friend constexpr bool operator!=(_CmpUnspecifiedParam, partial_ordering __v) noexcept; - _LIBCPP_INLINE_VISIBILITY friend constexpr bool operator< (_CmpUnspecifiedParam, partial_ordering __v) noexcept; - _LIBCPP_INLINE_VISIBILITY friend constexpr bool operator<=(_CmpUnspecifiedParam, partial_ordering __v) noexcept; - _LIBCPP_INLINE_VISIBILITY friend constexpr bool operator> (_CmpUnspecifiedParam, partial_ordering __v) noexcept; - _LIBCPP_INLINE_VISIBILITY friend constexpr bool operator>=(_CmpUnspecifiedParam, partial_ordering __v) noexcept; - -#ifndef _LIBCPP_HAS_NO_SPACESHIP_OPERATOR _LIBCPP_INLINE_VISIBILITY friend constexpr bool operator==(partial_ordering, partial_ordering) noexcept = default; - _LIBCPP_INLINE_VISIBILITY friend constexpr partial_ordering operator<=>(partial_ordering __v, _CmpUnspecifiedParam) noexcept; - _LIBCPP_INLINE_VISIBILITY friend constexpr partial_ordering operator<=>(_CmpUnspecifiedParam, partial_ordering __v) noexcept; -#endif + _LIBCPP_INLINE_VISIBILITY friend constexpr bool operator==(partial_ordering __v, _CmpUnspecifiedParam) noexcept { + return __v.__is_ordered() && __v.__value_ == 0; + } + + _LIBCPP_INLINE_VISIBILITY friend constexpr bool operator< (partial_ordering __v, _CmpUnspecifiedParam) noexcept { + return __v.__is_ordered() && __v.__value_ < 0; + } + + _LIBCPP_INLINE_VISIBILITY friend constexpr bool operator<=(partial_ordering __v, _CmpUnspecifiedParam) noexcept { + return __v.__is_ordered() && __v.__value_ <= 0; + } + + _LIBCPP_INLINE_VISIBILITY friend constexpr bool operator> (partial_ordering __v, _CmpUnspecifiedParam) noexcept { + return __v.__is_ordered() && __v.__value_ > 0; + } + + _LIBCPP_INLINE_VISIBILITY friend constexpr bool operator>=(partial_ordering __v, _CmpUnspecifiedParam) noexcept { + return __v.__is_ordered() && __v.__value_ >= 0; + } + _LIBCPP_INLINE_VISIBILITY friend constexpr bool operator< (_CmpUnspecifiedParam, partial_ordering __v) noexcept { + return __v.__is_ordered() && 0 < __v.__value_; + } + + _LIBCPP_INLINE_VISIBILITY friend constexpr bool operator<=(_CmpUnspecifiedParam, partial_ordering __v) noexcept { + return __v.__is_ordered() && 0 <= __v.__value_; + } + + _LIBCPP_INLINE_VISIBILITY friend constexpr bool operator> (_CmpUnspecifiedParam, partial_ordering __v) noexcept { + return __v.__is_ordered() && 0 > __v.__value_; + } + + _LIBCPP_INLINE_VISIBILITY friend constexpr bool operator>=(_CmpUnspecifiedParam, partial_ordering __v) noexcept { + return __v.__is_ordered() && 0 >= __v.__value_; + } + + _LIBCPP_INLINE_VISIBILITY friend constexpr partial_ordering operator<=>(partial_ordering __v, _CmpUnspecifiedParam) noexcept { + return __v; + } + + _LIBCPP_INLINE_VISIBILITY friend constexpr partial_ordering operator<=>(_CmpUnspecifiedParam, partial_ordering __v) noexcept { + return __v < 0 ? partial_ordering::greater : (__v > 0 ? partial_ordering::less : __v); + } private: _ValueT __value_; }; @@ -346,68 +243,6 @@ _LIBCPP_INLINE_VAR constexpr partial_ordering partial_ordering::equivalent(_EqRe _LIBCPP_INLINE_VAR constexpr partial_ordering partial_ordering::greater(_OrdResult::__greater); _LIBCPP_INLINE_VAR constexpr partial_ordering partial_ordering::unordered(_NCmpResult ::__unordered); -_LIBCPP_INLINE_VISIBILITY -constexpr bool operator==(partial_ordering __v, _CmpUnspecifiedParam) noexcept { - return __v.__is_ordered() && __v.__value_ == 0; -} -_LIBCPP_INLINE_VISIBILITY -constexpr bool operator< (partial_ordering __v, _CmpUnspecifiedParam) noexcept { - return __v.__is_ordered() && __v.__value_ < 0; -} -_LIBCPP_INLINE_VISIBILITY -constexpr bool operator<=(partial_ordering __v, _CmpUnspecifiedParam) noexcept { - return __v.__is_ordered() && __v.__value_ <= 0; -} -_LIBCPP_INLINE_VISIBILITY -constexpr bool operator> (partial_ordering __v, _CmpUnspecifiedParam) noexcept { - return __v.__is_ordered() && __v.__value_ > 0; -} -_LIBCPP_INLINE_VISIBILITY -constexpr bool operator>=(partial_ordering __v, _CmpUnspecifiedParam) noexcept { - return __v.__is_ordered() && __v.__value_ >= 0; -} - -_LIBCPP_INLINE_VISIBILITY -constexpr bool operator==(_CmpUnspecifiedParam, partial_ordering __v) noexcept { - return __v.__is_ordered() && 0 == __v.__value_; -} -_LIBCPP_INLINE_VISIBILITY -constexpr bool operator< (_CmpUnspecifiedParam, partial_ordering __v) noexcept { - return __v.__is_ordered() && 0 < __v.__value_; -} -_LIBCPP_INLINE_VISIBILITY -constexpr bool operator<=(_CmpUnspecifiedParam, partial_ordering __v) noexcept { - return __v.__is_ordered() && 0 <= __v.__value_; -} -_LIBCPP_INLINE_VISIBILITY -constexpr bool operator> (_CmpUnspecifiedParam, partial_ordering __v) noexcept { - return __v.__is_ordered() && 0 > __v.__value_; -} -_LIBCPP_INLINE_VISIBILITY -constexpr bool operator>=(_CmpUnspecifiedParam, partial_ordering __v) noexcept { - return __v.__is_ordered() && 0 >= __v.__value_; -} - -_LIBCPP_INLINE_VISIBILITY -constexpr bool operator!=(partial_ordering __v, _CmpUnspecifiedParam) noexcept { - return !__v.__is_ordered() || __v.__value_ != 0; -} -_LIBCPP_INLINE_VISIBILITY -constexpr bool operator!=(_CmpUnspecifiedParam, partial_ordering __v) noexcept { - return !__v.__is_ordered() || __v.__value_ != 0; -} - -#ifndef _LIBCPP_HAS_NO_SPACESHIP_OPERATOR -_LIBCPP_INLINE_VISIBILITY -constexpr partial_ordering operator<=>(partial_ordering __v, _CmpUnspecifiedParam) noexcept { - return __v; -} -_LIBCPP_INLINE_VISIBILITY -constexpr partial_ordering operator<=>(_CmpUnspecifiedParam, partial_ordering __v) noexcept { - return __v < 0 ? partial_ordering::greater : (__v > 0 ? partial_ordering::less : __v); -} -#endif // _LIBCPP_HAS_NO_SPACESHIP_OPERATOR - class weak_ordering { using _ValueT = signed char; @@ -421,13 +256,6 @@ public: static const weak_ordering equivalent; static const weak_ordering greater; - // conversions - _LIBCPP_INLINE_VISIBILITY - constexpr operator weak_equality() const noexcept { - return __value_ == 0 ? weak_equality::equivalent - : weak_equality::nonequivalent; - } - _LIBCPP_INLINE_VISIBILITY constexpr operator partial_ordering() const noexcept { return __value_ == 0 ? partial_ordering::equivalent @@ -435,25 +263,51 @@ public: } // comparisons - _LIBCPP_INLINE_VISIBILITY friend constexpr bool operator==(weak_ordering __v, _CmpUnspecifiedParam) noexcept; - _LIBCPP_INLINE_VISIBILITY friend constexpr bool operator!=(weak_ordering __v, _CmpUnspecifiedParam) noexcept; - _LIBCPP_INLINE_VISIBILITY friend constexpr bool operator< (weak_ordering __v, _CmpUnspecifiedParam) noexcept; - _LIBCPP_INLINE_VISIBILITY friend constexpr bool operator<=(weak_ordering __v, _CmpUnspecifiedParam) noexcept; - _LIBCPP_INLINE_VISIBILITY friend constexpr bool operator> (weak_ordering __v, _CmpUnspecifiedParam) noexcept; - _LIBCPP_INLINE_VISIBILITY friend constexpr bool operator>=(weak_ordering __v, _CmpUnspecifiedParam) noexcept; - _LIBCPP_INLINE_VISIBILITY friend constexpr bool operator==(_CmpUnspecifiedParam, weak_ordering __v) noexcept; - _LIBCPP_INLINE_VISIBILITY friend constexpr bool operator!=(_CmpUnspecifiedParam, weak_ordering __v) noexcept; - _LIBCPP_INLINE_VISIBILITY friend constexpr bool operator< (_CmpUnspecifiedParam, weak_ordering __v) noexcept; - _LIBCPP_INLINE_VISIBILITY friend constexpr bool operator<=(_CmpUnspecifiedParam, weak_ordering __v) noexcept; - _LIBCPP_INLINE_VISIBILITY friend constexpr bool operator> (_CmpUnspecifiedParam, weak_ordering __v) noexcept; - _LIBCPP_INLINE_VISIBILITY friend constexpr bool operator>=(_CmpUnspecifiedParam, weak_ordering __v) noexcept; - -#ifndef _LIBCPP_HAS_NO_SPACESHIP_OPERATOR _LIBCPP_INLINE_VISIBILITY friend constexpr bool operator==(weak_ordering, weak_ordering) noexcept = default; - _LIBCPP_INLINE_VISIBILITY friend constexpr weak_ordering operator<=>(weak_ordering __v, _CmpUnspecifiedParam) noexcept; - _LIBCPP_INLINE_VISIBILITY friend constexpr weak_ordering operator<=>(_CmpUnspecifiedParam, weak_ordering __v) noexcept; -#endif + _LIBCPP_INLINE_VISIBILITY friend constexpr bool operator==(weak_ordering __v, _CmpUnspecifiedParam) noexcept { + return __v.__value_ == 0; + } + + _LIBCPP_INLINE_VISIBILITY friend constexpr bool operator< (weak_ordering __v, _CmpUnspecifiedParam) noexcept { + return __v.__value_ < 0; + } + + _LIBCPP_INLINE_VISIBILITY friend constexpr bool operator<=(weak_ordering __v, _CmpUnspecifiedParam) noexcept { + return __v.__value_ <= 0; + } + + _LIBCPP_INLINE_VISIBILITY friend constexpr bool operator> (weak_ordering __v, _CmpUnspecifiedParam) noexcept { + return __v.__value_ > 0; + } + + _LIBCPP_INLINE_VISIBILITY friend constexpr bool operator>=(weak_ordering __v, _CmpUnspecifiedParam) noexcept { + return __v.__value_ >= 0; + } + + _LIBCPP_INLINE_VISIBILITY friend constexpr bool operator< (_CmpUnspecifiedParam, weak_ordering __v) noexcept { + return 0 < __v.__value_; + } + + _LIBCPP_INLINE_VISIBILITY friend constexpr bool operator<=(_CmpUnspecifiedParam, weak_ordering __v) noexcept { + return 0 <= __v.__value_; + } + + _LIBCPP_INLINE_VISIBILITY friend constexpr bool operator> (_CmpUnspecifiedParam, weak_ordering __v) noexcept { + return 0 > __v.__value_; + } + + _LIBCPP_INLINE_VISIBILITY friend constexpr bool operator>=(_CmpUnspecifiedParam, weak_ordering __v) noexcept { + return 0 >= __v.__value_; + } + + _LIBCPP_INLINE_VISIBILITY friend constexpr weak_ordering operator<=>(weak_ordering __v, _CmpUnspecifiedParam) noexcept { + return __v; + } + + _LIBCPP_INLINE_VISIBILITY friend constexpr weak_ordering operator<=>(_CmpUnspecifiedParam, weak_ordering __v) noexcept { + return __v < 0 ? weak_ordering::greater : (__v > 0 ? weak_ordering::less : __v); + } private: _ValueT __value_; @@ -462,67 +316,6 @@ private: _LIBCPP_INLINE_VAR constexpr weak_ordering weak_ordering::less(_OrdResult::__less); _LIBCPP_INLINE_VAR constexpr weak_ordering weak_ordering::equivalent(_EqResult::__equiv); _LIBCPP_INLINE_VAR constexpr weak_ordering weak_ordering::greater(_OrdResult::__greater); - -_LIBCPP_INLINE_VISIBILITY -constexpr bool operator==(weak_ordering __v, _CmpUnspecifiedParam) noexcept { - return __v.__value_ == 0; -} -_LIBCPP_INLINE_VISIBILITY -constexpr bool operator!=(weak_ordering __v, _CmpUnspecifiedParam) noexcept { - return __v.__value_ != 0; -} -_LIBCPP_INLINE_VISIBILITY -constexpr bool operator< (weak_ordering __v, _CmpUnspecifiedParam) noexcept { - return __v.__value_ < 0; -} -_LIBCPP_INLINE_VISIBILITY -constexpr bool operator<=(weak_ordering __v, _CmpUnspecifiedParam) noexcept { - return __v.__value_ <= 0; -} -_LIBCPP_INLINE_VISIBILITY -constexpr bool operator> (weak_ordering __v, _CmpUnspecifiedParam) noexcept { - return __v.__value_ > 0; -} -_LIBCPP_INLINE_VISIBILITY -constexpr bool operator>=(weak_ordering __v, _CmpUnspecifiedParam) noexcept { - return __v.__value_ >= 0; -} -_LIBCPP_INLINE_VISIBILITY -constexpr bool operator==(_CmpUnspecifiedParam, weak_ordering __v) noexcept { - return 0 == __v.__value_; -} -_LIBCPP_INLINE_VISIBILITY -constexpr bool operator!=(_CmpUnspecifiedParam, weak_ordering __v) noexcept { - return 0 != __v.__value_; -} -_LIBCPP_INLINE_VISIBILITY -constexpr bool operator< (_CmpUnspecifiedParam, weak_ordering __v) noexcept { - return 0 < __v.__value_; -} -_LIBCPP_INLINE_VISIBILITY -constexpr bool operator<=(_CmpUnspecifiedParam, weak_ordering __v) noexcept { - return 0 <= __v.__value_; -} -_LIBCPP_INLINE_VISIBILITY -constexpr bool operator> (_CmpUnspecifiedParam, weak_ordering __v) noexcept { - return 0 > __v.__value_; -} -_LIBCPP_INLINE_VISIBILITY -constexpr bool operator>=(_CmpUnspecifiedParam, weak_ordering __v) noexcept { - return 0 >= __v.__value_; -} - -#ifndef _LIBCPP_HAS_NO_SPACESHIP_OPERATOR -_LIBCPP_INLINE_VISIBILITY -constexpr weak_ordering operator<=>(weak_ordering __v, _CmpUnspecifiedParam) noexcept { - return __v; -} -_LIBCPP_INLINE_VISIBILITY -constexpr weak_ordering operator<=>(_CmpUnspecifiedParam, weak_ordering __v) noexcept { - return __v < 0 ? weak_ordering::greater : (__v > 0 ? weak_ordering::less : __v); -} -#endif // _LIBCPP_HAS_NO_SPACESHIP_OPERATOR - class strong_ordering { using _ValueT = signed char; @@ -538,18 +331,6 @@ public: static const strong_ordering greater; // conversions - _LIBCPP_INLINE_VISIBILITY - constexpr operator weak_equality() const noexcept { - return __value_ == 0 ? weak_equality::equivalent - : weak_equality::nonequivalent; - } - - _LIBCPP_INLINE_VISIBILITY - constexpr operator strong_equality() const noexcept { - return __value_ == 0 ? strong_equality::equal - : strong_equality::nonequal; - } - _LIBCPP_INLINE_VISIBILITY constexpr operator partial_ordering() const noexcept { return __value_ == 0 ? partial_ordering::equivalent @@ -563,25 +344,51 @@ public: } // comparisons - _LIBCPP_INLINE_VISIBILITY friend constexpr bool operator==(strong_ordering __v, _CmpUnspecifiedParam) noexcept; - _LIBCPP_INLINE_VISIBILITY friend constexpr bool operator!=(strong_ordering __v, _CmpUnspecifiedParam) noexcept; - _LIBCPP_INLINE_VISIBILITY friend constexpr bool operator< (strong_ordering __v, _CmpUnspecifiedParam) noexcept; - _LIBCPP_INLINE_VISIBILITY friend constexpr bool operator<=(strong_ordering __v, _CmpUnspecifiedParam) noexcept; - _LIBCPP_INLINE_VISIBILITY friend constexpr bool operator> (strong_ordering __v, _CmpUnspecifiedParam) noexcept; - _LIBCPP_INLINE_VISIBILITY friend constexpr bool operator>=(strong_ordering __v, _CmpUnspecifiedParam) noexcept; - _LIBCPP_INLINE_VISIBILITY friend constexpr bool operator==(_CmpUnspecifiedParam, strong_ordering __v) noexcept; - _LIBCPP_INLINE_VISIBILITY friend constexpr bool operator!=(_CmpUnspecifiedParam, strong_ordering __v) noexcept; - _LIBCPP_INLINE_VISIBILITY friend constexpr bool operator< (_CmpUnspecifiedParam, strong_ordering __v) noexcept; - _LIBCPP_INLINE_VISIBILITY friend constexpr bool operator<=(_CmpUnspecifiedParam, strong_ordering __v) noexcept; - _LIBCPP_INLINE_VISIBILITY friend constexpr bool operator> (_CmpUnspecifiedParam, strong_ordering __v) noexcept; - _LIBCPP_INLINE_VISIBILITY friend constexpr bool operator>=(_CmpUnspecifiedParam, strong_ordering __v) noexcept; - -#ifndef _LIBCPP_HAS_NO_SPACESHIP_OPERATOR _LIBCPP_INLINE_VISIBILITY friend constexpr bool operator==(strong_ordering, strong_ordering) noexcept = default; - _LIBCPP_INLINE_VISIBILITY friend constexpr strong_ordering operator<=>(strong_ordering __v, _CmpUnspecifiedParam) noexcept; - _LIBCPP_INLINE_VISIBILITY friend constexpr strong_ordering operator<=>(_CmpUnspecifiedParam, strong_ordering __v) noexcept; -#endif + _LIBCPP_INLINE_VISIBILITY friend constexpr bool operator==(strong_ordering __v, _CmpUnspecifiedParam) noexcept { + return __v.__value_ == 0; + } + + _LIBCPP_INLINE_VISIBILITY friend constexpr bool operator< (strong_ordering __v, _CmpUnspecifiedParam) noexcept { + return __v.__value_ < 0; + } + + _LIBCPP_INLINE_VISIBILITY friend constexpr bool operator<=(strong_ordering __v, _CmpUnspecifiedParam) noexcept { + return __v.__value_ <= 0; + } + + _LIBCPP_INLINE_VISIBILITY friend constexpr bool operator> (strong_ordering __v, _CmpUnspecifiedParam) noexcept { + return __v.__value_ > 0; + } + + _LIBCPP_INLINE_VISIBILITY friend constexpr bool operator>=(strong_ordering __v, _CmpUnspecifiedParam) noexcept { + return __v.__value_ >= 0; + } + + _LIBCPP_INLINE_VISIBILITY friend constexpr bool operator< (_CmpUnspecifiedParam, strong_ordering __v) noexcept { + return 0 < __v.__value_; + } + + _LIBCPP_INLINE_VISIBILITY friend constexpr bool operator<=(_CmpUnspecifiedParam, strong_ordering __v) noexcept { + return 0 <= __v.__value_; + } + + _LIBCPP_INLINE_VISIBILITY friend constexpr bool operator> (_CmpUnspecifiedParam, strong_ordering __v) noexcept { + return 0 > __v.__value_; + } + + _LIBCPP_INLINE_VISIBILITY friend constexpr bool operator>=(_CmpUnspecifiedParam, strong_ordering __v) noexcept { + return 0 >= __v.__value_; + } + + _LIBCPP_INLINE_VISIBILITY friend constexpr strong_ordering operator<=>(strong_ordering __v, _CmpUnspecifiedParam) noexcept { + return __v; + } + + _LIBCPP_INLINE_VISIBILITY friend constexpr strong_ordering operator<=>(_CmpUnspecifiedParam, strong_ordering __v) noexcept { + return __v < 0 ? strong_ordering::greater : (__v > 0 ? strong_ordering::less : __v); + } private: _ValueT __value_; @@ -592,73 +399,7 @@ _LIBCPP_INLINE_VAR constexpr strong_ordering strong_ordering::equal(_EqResult::_ _LIBCPP_INLINE_VAR constexpr strong_ordering strong_ordering::equivalent(_EqResult::__equiv); _LIBCPP_INLINE_VAR constexpr strong_ordering strong_ordering::greater(_OrdResult::__greater); -_LIBCPP_INLINE_VISIBILITY -constexpr bool operator==(strong_ordering __v, _CmpUnspecifiedParam) noexcept { - return __v.__value_ == 0; -} -_LIBCPP_INLINE_VISIBILITY -constexpr bool operator!=(strong_ordering __v, _CmpUnspecifiedParam) noexcept { - return __v.__value_ != 0; -} -_LIBCPP_INLINE_VISIBILITY -constexpr bool operator< (strong_ordering __v, _CmpUnspecifiedParam) noexcept { - return __v.__value_ < 0; -} -_LIBCPP_INLINE_VISIBILITY -constexpr bool operator<=(strong_ordering __v, _CmpUnspecifiedParam) noexcept { - return __v.__value_ <= 0; -} -_LIBCPP_INLINE_VISIBILITY -constexpr bool operator> (strong_ordering __v, _CmpUnspecifiedParam) noexcept { - return __v.__value_ > 0; -} -_LIBCPP_INLINE_VISIBILITY -constexpr bool operator>=(strong_ordering __v, _CmpUnspecifiedParam) noexcept { - return __v.__value_ >= 0; -} -_LIBCPP_INLINE_VISIBILITY -constexpr bool operator==(_CmpUnspecifiedParam, strong_ordering __v) noexcept { - return 0 == __v.__value_; -} -_LIBCPP_INLINE_VISIBILITY -constexpr bool operator!=(_CmpUnspecifiedParam, strong_ordering __v) noexcept { - return 0 != __v.__value_; -} -_LIBCPP_INLINE_VISIBILITY -constexpr bool operator< (_CmpUnspecifiedParam, strong_ordering __v) noexcept { - return 0 < __v.__value_; -} -_LIBCPP_INLINE_VISIBILITY -constexpr bool operator<=(_CmpUnspecifiedParam, strong_ordering __v) noexcept { - return 0 <= __v.__value_; -} -_LIBCPP_INLINE_VISIBILITY -constexpr bool operator> (_CmpUnspecifiedParam, strong_ordering __v) noexcept { - return 0 > __v.__value_; -} -_LIBCPP_INLINE_VISIBILITY -constexpr bool operator>=(_CmpUnspecifiedParam, strong_ordering __v) noexcept { - return 0 >= __v.__value_; -} - -#ifndef _LIBCPP_HAS_NO_SPACESHIP_OPERATOR -_LIBCPP_INLINE_VISIBILITY -constexpr strong_ordering operator<=>(strong_ordering __v, _CmpUnspecifiedParam) noexcept { - return __v; -} -_LIBCPP_INLINE_VISIBILITY -constexpr strong_ordering operator<=>(_CmpUnspecifiedParam, strong_ordering __v) noexcept { - return __v < 0 ? strong_ordering::greater : (__v > 0 ? strong_ordering::less : __v); -} -#endif // _LIBCPP_HAS_NO_SPACESHIP_OPERATOR - // named comparison functions -_LIBCPP_INLINE_VISIBILITY -constexpr bool is_eq(weak_equality __cmp) noexcept { return __cmp == 0; } - -_LIBCPP_INLINE_VISIBILITY -constexpr bool is_neq(weak_equality __cmp) noexcept { return __cmp != 0; } - _LIBCPP_INLINE_VISIBILITY constexpr bool is_lt(partial_ordering __cmp) noexcept { return __cmp < 0; } @@ -675,8 +416,6 @@ namespace __comp_detail { enum _ClassifyCompCategory : unsigned{ _None, - _WeakEq, - _StrongEq, _PartialOrd, _WeakOrd, _StrongOrd, @@ -686,10 +425,6 @@ enum _ClassifyCompCategory : unsigned{ template _LIBCPP_INLINE_VISIBILITY constexpr _ClassifyCompCategory __type_to_enum() noexcept { - if (is_same_v<_Tp, weak_equality>) - return _WeakEq; - if (is_same_v<_Tp, strong_equality>) - return _StrongEq; if (is_same_v<_Tp, partial_ordering>) return _PartialOrd; if (is_same_v<_Tp, weak_ordering>) @@ -701,18 +436,12 @@ constexpr _ClassifyCompCategory __type_to_enum() noexcept { template constexpr _ClassifyCompCategory -__compute_comp_type(array<_ClassifyCompCategory, _Size> __types) { - array __seen = {}; +__compute_comp_type(const _ClassifyCompCategory (&__types)[_Size]) { + int __seen[_CCC_Size] = {}; for (auto __type : __types) ++__seen[__type]; if (__seen[_None]) return _None; - if (__seen[_WeakEq]) - return _WeakEq; - if (__seen[_StrongEq] && (__seen[_PartialOrd] || __seen[_WeakOrd])) - return _WeakEq; - if (__seen[_StrongEq]) - return _StrongEq; if (__seen[_PartialOrd]) return _PartialOrd; if (__seen[_WeakOrd]) @@ -720,18 +449,13 @@ __compute_comp_type(array<_ClassifyCompCategory, _Size> __types) { return _StrongOrd; } -template +template constexpr auto __get_comp_type() { using _CCC = _ClassifyCompCategory; - constexpr array<_CCC, sizeof...(_Ts)> __type_kinds{{__comp_detail::__type_to_enum<_Ts>()...}}; - constexpr _CCC _Cat = sizeof...(_Ts) == 0 ? _StrongOrd - : __compute_comp_type(__type_kinds); + constexpr _CCC __type_kinds[] = {_StrongOrd, __type_to_enum<_Ts>()...}; + constexpr _CCC _Cat = __compute_comp_type(__type_kinds); if constexpr (_Cat == _None) return void(); - else if constexpr (_Cat == _WeakEq) - return weak_equality::equivalent; - else if constexpr (_Cat == _StrongEq) - return strong_equality::equivalent; else if constexpr (_Cat == _PartialOrd) return partial_ordering::equivalent; else if constexpr (_Cat == _WeakOrd) @@ -739,7 +463,7 @@ constexpr auto __get_comp_type() { else if constexpr (_Cat == _StrongOrd) return strong_ordering::equivalent; else - static_assert(_Cat != _Cat, "unhandled case"); + static_assert(_False, "unhandled case"); } } // namespace __comp_detail @@ -757,10 +481,8 @@ using common_comparison_category_t = typename common_comparison_category<_Ts...> template constexpr strong_ordering strong_order(const _Tp& __lhs, const _Tp& __rhs); template constexpr weak_ordering weak_order(const _Tp& __lhs, const _Tp& __rhs); template constexpr partial_ordering partial_order(const _Tp& __lhs, const _Tp& __rhs); -template constexpr strong_equality strong_equal(const _Tp& __lhs, const _Tp& __rhs); -template constexpr weak_equality weak_equal(const _Tp& __lhs, const _Tp& __rhs); -#endif // _LIBCPP_STD_VER > 17 +#endif // _LIBCPP_STD_VER > 17 && !defined(_LIBCPP_HAS_NO_SPACESHIP_OPERATOR) _LIBCPP_END_NAMESPACE_STD diff --git a/lib/libcxx/include/complex b/lib/libcxx/include/complex index 93b7bb5dd..fc52697f7 100644 --- a/lib/libcxx/include/complex +++ b/lib/libcxx/include/complex @@ -232,10 +232,10 @@ template complex tanh (const complex&); */ #include <__config> -#include -#include #include #include +#include +#include #include #if !defined(_LIBCPP_HAS_NO_LOCALIZATION) @@ -1490,4 +1490,4 @@ inline namespace literals _LIBCPP_END_NAMESPACE_STD -#endif // _LIBCPP_COMPLEX +#endif // _LIBCPP_COMPLEX diff --git a/lib/libcxx/include/complex.h b/lib/libcxx/include/complex.h index b78733b80..cdf97ac46 100644 --- a/lib/libcxx/include/complex.h +++ b/lib/libcxx/include/complex.h @@ -31,6 +31,6 @@ #include_next -#endif // __cplusplus +#endif // __cplusplus -#endif // _LIBCPP_COMPLEX_H +#endif // _LIBCPP_COMPLEX_H diff --git a/lib/libcxx/include/concepts b/lib/libcxx/include/concepts index cf5f9d639..3dec9b527 100644 --- a/lib/libcxx/include/concepts +++ b/lib/libcxx/include/concepts @@ -67,9 +67,9 @@ namespace std { template concept constructible_from = see below; - // [concept.defaultconstructible], concept default_constructible + // [concept.default.init], concept default_initializable template - concept default_constructible = see below; + concept default_initializable = see below; // [concept.moveconstructible], concept move_constructible template @@ -79,11 +79,6 @@ namespace std { template concept copy_constructible = see below; - // [concepts.compare], comparison concepts - // [concept.boolean], concept boolean - template - concept boolean = see below; - // [concept.equalitycomparable], concept equality_comparable template concept equality_comparable = see below; @@ -135,7 +130,10 @@ namespace std { */ #include <__config> +#include <__functional/invoke.h> +#include <__functional_base> #include +#include #include #if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) @@ -147,7 +145,7 @@ _LIBCPP_PUSH_MACROS _LIBCPP_BEGIN_NAMESPACE_STD -#if _LIBCPP_STD_VER > 17 && defined(__cpp_concepts) && __cpp_concepts >= 201811L +#if _LIBCPP_STD_VER > 17 && !defined(_LIBCPP_HAS_NO_CONCEPTS) // [concept.same] @@ -157,12 +155,298 @@ concept __same_as_impl = _VSTD::_IsSame<_Tp, _Up>::value; template concept same_as = __same_as_impl<_Tp, _Up> && __same_as_impl<_Up, _Tp>; +// [concept.derived] +template +concept derived_from = + is_base_of_v<_Bp, _Dp> && + is_convertible_v; + +// [concept.convertible] +template +concept convertible_to = + is_convertible_v<_From, _To> && + requires(add_rvalue_reference_t<_From> (&__f)()) { + static_cast<_To>(__f()); + }; + +// [concept.commonref] +template +concept common_reference_with = + same_as, common_reference_t<_Up, _Tp>> && + convertible_to<_Tp, common_reference_t<_Tp, _Up>> && + convertible_to<_Up, common_reference_t<_Tp, _Up>>; + +// [concept.common] +template +concept common_with = + same_as, common_type_t<_Up, _Tp>> && + requires { + static_cast>(declval<_Tp>()); + static_cast>(declval<_Up>()); + } && + common_reference_with< + add_lvalue_reference_t, + add_lvalue_reference_t> && + common_reference_with< + add_lvalue_reference_t>, + common_reference_t< + add_lvalue_reference_t, + add_lvalue_reference_t>>; + +// [concepts.arithmetic], arithmetic concepts +template +concept integral = is_integral_v<_Tp>; + +template +concept signed_integral = integral<_Tp> && is_signed_v<_Tp>; + +template +concept unsigned_integral = integral<_Tp> && !signed_integral<_Tp>; + +template +concept floating_point = is_floating_point_v<_Tp>; + +// [concept.assignable] +template +concept assignable_from = + is_lvalue_reference_v<_Lhs> && + common_reference_with<__make_const_lvalue_ref<_Lhs>, __make_const_lvalue_ref<_Rhs>> && + requires (_Lhs __lhs, _Rhs&& __rhs) { + { __lhs = _VSTD::forward<_Rhs>(__rhs) } -> same_as<_Lhs>; + }; + // [concept.destructible] template -concept destructible = _VSTD::is_nothrow_destructible_v<_Tp>; +concept destructible = is_nothrow_destructible_v<_Tp>; + +// [concept.constructible] +template +concept constructible_from = + destructible<_Tp> && is_constructible_v<_Tp, _Args...>; + +// [concept.default.init] + +template +concept __default_initializable = requires { ::new _Tp; }; + +template +concept default_initializable = constructible_from<_Tp> && + requires { _Tp{}; } && __default_initializable<_Tp>; + +// [concept.moveconstructible] +template +concept move_constructible = + constructible_from<_Tp, _Tp> && convertible_to<_Tp, _Tp>; + +// [concept.copyconstructible] +template +concept copy_constructible = + move_constructible<_Tp> && + constructible_from<_Tp, _Tp&> && convertible_to<_Tp&, _Tp> && + constructible_from<_Tp, const _Tp&> && convertible_to && + constructible_from<_Tp, const _Tp> && convertible_to; + +// Whether a type is a class type or enumeration type according to the Core wording. +template +concept __class_or_enum = is_class_v<_Tp> || is_union_v<_Tp> || is_enum_v<_Tp>; + +// [concept.swappable] +namespace ranges::__swap { + // Deleted to inhibit ADL + template + void swap(_Tp&, _Tp&) = delete; + + + // [1] + template + concept __unqualified_swappable_with = + (__class_or_enum> || __class_or_enum>) && + requires(_Tp&& __t, _Up&& __u) { + swap(_VSTD::forward<_Tp>(__t), _VSTD::forward<_Up>(__u)); + }; + + struct __fn; + + template + concept __swappable_arrays = + !__unqualified_swappable_with<_Tp(&)[_Size], _Up(&)[_Size]> && + extent_v<_Tp> == extent_v<_Up> && + requires(_Tp(& __t)[_Size], _Up(& __u)[_Size], const __fn& __swap) { + __swap(__t[0], __u[0]); + }; + + template + concept __exchangeable = + !__unqualified_swappable_with<_Tp&, _Tp&> && + move_constructible<_Tp> && + assignable_from<_Tp&, _Tp>; + + struct __fn { + // 2.1 `S` is `(void)swap(E1, E2)`* if `E1` or `E2` has class or enumeration type and... + // *The name `swap` is used here unqualified. + template + requires __unqualified_swappable_with<_Tp, _Up> + constexpr void operator()(_Tp&& __t, _Up&& __u) const + noexcept(noexcept(swap(_VSTD::forward<_Tp>(__t), _VSTD::forward<_Up>(__u)))) + { + swap(_VSTD::forward<_Tp>(__t), _VSTD::forward<_Up>(__u)); + } + + // 2.2 Otherwise, if `E1` and `E2` are lvalues of array types with equal extent and... + template + requires __swappable_arrays<_Tp, _Up, _Size> + constexpr void operator()(_Tp(& __t)[_Size], _Up(& __u)[_Size]) const + noexcept(noexcept((*this)(*__t, *__u))) + { + // TODO(cjdb): replace with `ranges::swap_ranges`. + for (size_t __i = 0; __i < _Size; ++__i) { + (*this)(__t[__i], __u[__i]); + } + } + + // 2.3 Otherwise, if `E1` and `E2` are lvalues of the same type `T` that models... + template<__exchangeable _Tp> + constexpr void operator()(_Tp& __x, _Tp& __y) const + noexcept(is_nothrow_move_constructible_v<_Tp> && is_nothrow_move_assignable_v<_Tp>) + { + __y = _VSTD::exchange(__x, _VSTD::move(__y)); + } + }; +} // namespace ranges::__swap + +namespace ranges::inline __cpo { + inline constexpr auto swap = __swap::__fn{}; +} // namespace ranges::__cpo + +template +concept swappable = requires(_Tp& __a, _Tp& __b) { ranges::swap(__a, __b); }; + +template +concept swappable_with = + common_reference_with<_Tp, _Up> && + requires(_Tp&& __t, _Up&& __u) { + ranges::swap(_VSTD::forward<_Tp>(__t), _VSTD::forward<_Tp>(__t)); + ranges::swap(_VSTD::forward<_Up>(__u), _VSTD::forward<_Up>(__u)); + ranges::swap(_VSTD::forward<_Tp>(__t), _VSTD::forward<_Up>(__u)); + ranges::swap(_VSTD::forward<_Up>(__u), _VSTD::forward<_Tp>(__t)); + }; + +// [concept.booleantestable] +template +concept __boolean_testable_impl = convertible_to<_Tp, bool>; + +template +concept __boolean_testable = __boolean_testable_impl<_Tp> && requires(_Tp&& __t) { + { !std::forward<_Tp>(__t) } -> __boolean_testable_impl; +}; + +// [concept.equalitycomparable] +template +concept __weakly_equality_comparable_with = + requires(__make_const_lvalue_ref<_Tp> __t, __make_const_lvalue_ref<_Up> __u) { + { __t == __u } -> __boolean_testable; + { __t != __u } -> __boolean_testable; + { __u == __t } -> __boolean_testable; + { __u != __t } -> __boolean_testable; + }; + +template +concept equality_comparable = __weakly_equality_comparable_with<_Tp, _Tp>; + +template +concept equality_comparable_with = + equality_comparable<_Tp> && equality_comparable<_Up> && + common_reference_with<__make_const_lvalue_ref<_Tp>, __make_const_lvalue_ref<_Up>> && + equality_comparable< + common_reference_t< + __make_const_lvalue_ref<_Tp>, + __make_const_lvalue_ref<_Up>>> && + __weakly_equality_comparable_with<_Tp, _Up>; + +// [concept.totallyordered] + +template +concept __partially_ordered_with = + requires(__make_const_lvalue_ref<_Tp> __t, __make_const_lvalue_ref<_Up> __u) { + { __t < __u } -> __boolean_testable; + { __t > __u } -> __boolean_testable; + { __t <= __u } -> __boolean_testable; + { __t >= __u } -> __boolean_testable; + { __u < __t } -> __boolean_testable; + { __u > __t } -> __boolean_testable; + { __u <= __t } -> __boolean_testable; + { __u >= __t } -> __boolean_testable; + }; + +template +concept totally_ordered = equality_comparable<_Tp> && __partially_ordered_with<_Tp, _Tp>; + +template +concept totally_ordered_with = + totally_ordered<_Tp> && totally_ordered<_Up> && + equality_comparable_with<_Tp, _Up> && + totally_ordered< + common_reference_t< + __make_const_lvalue_ref<_Tp>, + __make_const_lvalue_ref<_Up>>> && + __partially_ordered_with<_Tp, _Up>; + +// [concepts.object] +template +concept movable = + is_object_v<_Tp> && + move_constructible<_Tp> && + assignable_from<_Tp&, _Tp> && + swappable<_Tp>; + +template +concept copyable = + copy_constructible<_Tp> && + movable<_Tp> && + assignable_from<_Tp&, _Tp&> && + assignable_from<_Tp&, const _Tp&> && + assignable_from<_Tp&, const _Tp>; + +template +concept semiregular = copyable<_Tp> && default_initializable<_Tp>; + +template +concept regular = semiregular<_Tp> && equality_comparable<_Tp>; + +// [concept.invocable] +template +concept invocable = requires(_Fn&& __fn, _Args&&... __args) { + _VSTD::invoke(_VSTD::forward<_Fn>(__fn), _VSTD::forward<_Args>(__args)...); // not required to be equality preserving +}; + +// [concept.regular.invocable] +template +concept regular_invocable = invocable<_Fn, _Args...>; + +// [concept.predicate] +template +concept predicate = + regular_invocable<_Fn, _Args...> && __boolean_testable>; + +// [concept.relation] +template +concept relation = + predicate<_Rp, _Tp, _Tp> && predicate<_Rp, _Up, _Up> && + predicate<_Rp, _Tp, _Up> && predicate<_Rp, _Up, _Tp>; + +// [concept.equiv] +template +concept equivalence_relation = relation<_Rp, _Tp, _Up>; + +// [concept.strictweakorder] +template +concept strict_weak_order = relation<_Rp, _Tp, _Up>; + +template +concept __different_from = !same_as, remove_cvref_t<_Up>>; -#endif //_LIBCPP_STD_VER > 17 && defined(__cpp_concepts) && __cpp_concepts >= 201811L +#endif // _LIBCPP_STD_VER > 17 && !defined(_LIBCPP_HAS_NO_CONCEPTS) _LIBCPP_END_NAMESPACE_STD diff --git a/lib/libcxx/include/condition_variable b/lib/libcxx/include/condition_variable index 8c733448e..a33250c67 100644 --- a/lib/libcxx/include/condition_variable +++ b/lib/libcxx/include/condition_variable @@ -265,4 +265,4 @@ _LIBCPP_END_NAMESPACE_STD #endif // !_LIBCPP_HAS_NO_THREADS -#endif // _LIBCPP_CONDITION_VARIABLE +#endif // _LIBCPP_CONDITION_VARIABLE diff --git a/lib/libcxx/include/csetjmp b/lib/libcxx/include/csetjmp index ed94b50d1..41902f0e4 100644 --- a/lib/libcxx/include/csetjmp +++ b/lib/libcxx/include/csetjmp @@ -39,9 +39,9 @@ void longjmp(jmp_buf env, int val); _LIBCPP_BEGIN_NAMESPACE_STD -using ::jmp_buf; -using ::longjmp; +using ::jmp_buf _LIBCPP_USING_IF_EXISTS; +using ::longjmp _LIBCPP_USING_IF_EXISTS; _LIBCPP_END_NAMESPACE_STD -#endif // _LIBCPP_CSETJMP +#endif // _LIBCPP_CSETJMP diff --git a/lib/libcxx/include/csignal b/lib/libcxx/include/csignal index 99abd02de..3b262b561 100644 --- a/lib/libcxx/include/csignal +++ b/lib/libcxx/include/csignal @@ -48,10 +48,10 @@ int raise(int sig); _LIBCPP_BEGIN_NAMESPACE_STD -using ::sig_atomic_t; -using ::signal; -using ::raise; +using ::sig_atomic_t _LIBCPP_USING_IF_EXISTS; +using ::signal _LIBCPP_USING_IF_EXISTS; +using ::raise _LIBCPP_USING_IF_EXISTS; _LIBCPP_END_NAMESPACE_STD -#endif // _LIBCPP_CSIGNAL +#endif // _LIBCPP_CSIGNAL diff --git a/lib/libcxx/include/cstdarg b/lib/libcxx/include/cstdarg index e8147d496..352db8b8c 100644 --- a/lib/libcxx/include/cstdarg +++ b/lib/libcxx/include/cstdarg @@ -40,8 +40,8 @@ Types: _LIBCPP_BEGIN_NAMESPACE_STD -using ::va_list; +using ::va_list _LIBCPP_USING_IF_EXISTS; _LIBCPP_END_NAMESPACE_STD -#endif // _LIBCPP_CSTDARG +#endif // _LIBCPP_CSTDARG diff --git a/lib/libcxx/include/cstdbool b/lib/libcxx/include/cstdbool index fad471414..7708537ca 100644 --- a/lib/libcxx/include/cstdbool +++ b/lib/libcxx/include/cstdbool @@ -28,4 +28,4 @@ Macros: #undef __bool_true_false_are_defined #define __bool_true_false_are_defined 1 -#endif // _LIBCPP_CSTDBOOL +#endif // _LIBCPP_CSTDBOOL diff --git a/lib/libcxx/include/cstddef b/lib/libcxx/include/cstddef index 2a0bfeb6e..1b54e7b2e 100644 --- a/lib/libcxx/include/cstddef +++ b/lib/libcxx/include/cstddef @@ -46,11 +46,11 @@ Types: _LIBCPP_BEGIN_NAMESPACE_STD -using ::ptrdiff_t; -using ::size_t; +using ::ptrdiff_t _LIBCPP_USING_IF_EXISTS; +using ::size_t _LIBCPP_USING_IF_EXISTS; #if !defined(_LIBCPP_CXX03_LANG) -using ::max_align_t; +using ::max_align_t _LIBCPP_USING_IF_EXISTS; #endif template struct __libcpp_is_integral { enum { value = 0 }; }; @@ -59,13 +59,13 @@ template <> struct __libcpp_is_integral { enum { va template <> struct __libcpp_is_integral { enum { value = 1 }; }; template <> struct __libcpp_is_integral { enum { value = 1 }; }; template <> struct __libcpp_is_integral { enum { value = 1 }; }; -#ifndef _LIBCPP_NO_HAS_CHAR8_T +#ifndef _LIBCPP_HAS_NO_CHAR8_T template <> struct __libcpp_is_integral { enum { value = 1 }; }; #endif #ifndef _LIBCPP_HAS_NO_UNICODE_CHARS template <> struct __libcpp_is_integral { enum { value = 1 }; }; template <> struct __libcpp_is_integral { enum { value = 1 }; }; -#endif // _LIBCPP_HAS_NO_UNICODE_CHARS +#endif // _LIBCPP_HAS_NO_UNICODE_CHARS template <> struct __libcpp_is_integral { enum { value = 1 }; }; template <> struct __libcpp_is_integral { enum { value = 1 }; }; template <> struct __libcpp_is_integral { enum { value = 1 }; }; @@ -152,10 +152,10 @@ template { return static_cast(static_cast(static_cast(__lhs) >> __shift)); } template > - constexpr _Integer + _LIBCPP_NODISCARD_EXT constexpr _Integer to_integer(byte __b) noexcept { return static_cast<_Integer>(__b); } } #endif -#endif // _LIBCPP_CSTDDEF +#endif // _LIBCPP_CSTDDEF diff --git a/lib/libcxx/include/cstdint b/lib/libcxx/include/cstdint index f72fa0673..aa7c8b5cb 100644 --- a/lib/libcxx/include/cstdint +++ b/lib/libcxx/include/cstdint @@ -149,42 +149,42 @@ Types: _LIBCPP_BEGIN_NAMESPACE_STD -using::int8_t; -using::int16_t; -using::int32_t; -using::int64_t; - -using::uint8_t; -using::uint16_t; -using::uint32_t; -using::uint64_t; - -using::int_least8_t; -using::int_least16_t; -using::int_least32_t; -using::int_least64_t; - -using::uint_least8_t; -using::uint_least16_t; -using::uint_least32_t; -using::uint_least64_t; - -using::int_fast8_t; -using::int_fast16_t; -using::int_fast32_t; -using::int_fast64_t; - -using::uint_fast8_t; -using::uint_fast16_t; -using::uint_fast32_t; -using::uint_fast64_t; - -using::intptr_t; -using::uintptr_t; - -using::intmax_t; -using::uintmax_t; +using ::int8_t _LIBCPP_USING_IF_EXISTS; +using ::int16_t _LIBCPP_USING_IF_EXISTS; +using ::int32_t _LIBCPP_USING_IF_EXISTS; +using ::int64_t _LIBCPP_USING_IF_EXISTS; + +using ::uint8_t _LIBCPP_USING_IF_EXISTS; +using ::uint16_t _LIBCPP_USING_IF_EXISTS; +using ::uint32_t _LIBCPP_USING_IF_EXISTS; +using ::uint64_t _LIBCPP_USING_IF_EXISTS; + +using ::int_least8_t _LIBCPP_USING_IF_EXISTS; +using ::int_least16_t _LIBCPP_USING_IF_EXISTS; +using ::int_least32_t _LIBCPP_USING_IF_EXISTS; +using ::int_least64_t _LIBCPP_USING_IF_EXISTS; + +using ::uint_least8_t _LIBCPP_USING_IF_EXISTS; +using ::uint_least16_t _LIBCPP_USING_IF_EXISTS; +using ::uint_least32_t _LIBCPP_USING_IF_EXISTS; +using ::uint_least64_t _LIBCPP_USING_IF_EXISTS; + +using ::int_fast8_t _LIBCPP_USING_IF_EXISTS; +using ::int_fast16_t _LIBCPP_USING_IF_EXISTS; +using ::int_fast32_t _LIBCPP_USING_IF_EXISTS; +using ::int_fast64_t _LIBCPP_USING_IF_EXISTS; + +using ::uint_fast8_t _LIBCPP_USING_IF_EXISTS; +using ::uint_fast16_t _LIBCPP_USING_IF_EXISTS; +using ::uint_fast32_t _LIBCPP_USING_IF_EXISTS; +using ::uint_fast64_t _LIBCPP_USING_IF_EXISTS; + +using ::intptr_t _LIBCPP_USING_IF_EXISTS; +using ::uintptr_t _LIBCPP_USING_IF_EXISTS; + +using ::intmax_t _LIBCPP_USING_IF_EXISTS; +using ::uintmax_t _LIBCPP_USING_IF_EXISTS; _LIBCPP_END_NAMESPACE_STD -#endif // _LIBCPP_CSTDINT +#endif // _LIBCPP_CSTDINT diff --git a/lib/libcxx/include/cstdio b/lib/libcxx/include/cstdio index d0492a083..b480f800a 100644 --- a/lib/libcxx/include/cstdio +++ b/lib/libcxx/include/cstdio @@ -104,72 +104,72 @@ void perror(const char* s); _LIBCPP_BEGIN_NAMESPACE_STD -using ::FILE; -using ::fpos_t; -using ::size_t; - -using ::fclose; -using ::fflush; -using ::setbuf; -using ::setvbuf; -using ::fprintf; -using ::fscanf; -using ::snprintf; -using ::sprintf; -using ::sscanf; -using ::vfprintf; -using ::vfscanf; -using ::vsscanf; -using ::vsnprintf; -using ::vsprintf; -using ::fgetc; -using ::fgets; -using ::fputc; -using ::fputs; -using ::getc; -using ::putc; -using ::ungetc; -using ::fread; -using ::fwrite; +using ::FILE _LIBCPP_USING_IF_EXISTS; +using ::fpos_t _LIBCPP_USING_IF_EXISTS; +using ::size_t _LIBCPP_USING_IF_EXISTS; + +using ::fclose _LIBCPP_USING_IF_EXISTS; +using ::fflush _LIBCPP_USING_IF_EXISTS; +using ::setbuf _LIBCPP_USING_IF_EXISTS; +using ::setvbuf _LIBCPP_USING_IF_EXISTS; +using ::fprintf _LIBCPP_USING_IF_EXISTS; +using ::fscanf _LIBCPP_USING_IF_EXISTS; +using ::snprintf _LIBCPP_USING_IF_EXISTS; +using ::sprintf _LIBCPP_USING_IF_EXISTS; +using ::sscanf _LIBCPP_USING_IF_EXISTS; +using ::vfprintf _LIBCPP_USING_IF_EXISTS; +using ::vfscanf _LIBCPP_USING_IF_EXISTS; +using ::vsscanf _LIBCPP_USING_IF_EXISTS; +using ::vsnprintf _LIBCPP_USING_IF_EXISTS; +using ::vsprintf _LIBCPP_USING_IF_EXISTS; +using ::fgetc _LIBCPP_USING_IF_EXISTS; +using ::fgets _LIBCPP_USING_IF_EXISTS; +using ::fputc _LIBCPP_USING_IF_EXISTS; +using ::fputs _LIBCPP_USING_IF_EXISTS; +using ::getc _LIBCPP_USING_IF_EXISTS; +using ::putc _LIBCPP_USING_IF_EXISTS; +using ::ungetc _LIBCPP_USING_IF_EXISTS; +using ::fread _LIBCPP_USING_IF_EXISTS; +using ::fwrite _LIBCPP_USING_IF_EXISTS; #ifndef _LIBCPP_HAS_NO_FGETPOS_FSETPOS -using ::fgetpos; +using ::fgetpos _LIBCPP_USING_IF_EXISTS; #endif -using ::fseek; +using ::fseek _LIBCPP_USING_IF_EXISTS; #ifndef _LIBCPP_HAS_NO_FGETPOS_FSETPOS -using ::fsetpos; +using ::fsetpos _LIBCPP_USING_IF_EXISTS; #endif -using ::ftell; -using ::rewind; -using ::clearerr; -using ::feof; -using ::ferror; -using ::perror; +using ::ftell _LIBCPP_USING_IF_EXISTS; +using ::rewind _LIBCPP_USING_IF_EXISTS; +using ::clearerr _LIBCPP_USING_IF_EXISTS; +using ::feof _LIBCPP_USING_IF_EXISTS; +using ::ferror _LIBCPP_USING_IF_EXISTS; +using ::perror _LIBCPP_USING_IF_EXISTS; #ifndef _LIBCPP_HAS_NO_GLOBAL_FILESYSTEM_NAMESPACE -using ::fopen; -using ::freopen; -using ::remove; -using ::rename; -using ::tmpfile; -using ::tmpnam; +using ::fopen _LIBCPP_USING_IF_EXISTS; +using ::freopen _LIBCPP_USING_IF_EXISTS; +using ::remove _LIBCPP_USING_IF_EXISTS; +using ::rename _LIBCPP_USING_IF_EXISTS; +using ::tmpfile _LIBCPP_USING_IF_EXISTS; +using ::tmpnam _LIBCPP_USING_IF_EXISTS; #endif #ifndef _LIBCPP_HAS_NO_STDIN -using ::getchar; +using ::getchar _LIBCPP_USING_IF_EXISTS; #if _LIBCPP_STD_VER <= 11 && !defined(_LIBCPP_C_HAS_NO_GETS) -using ::gets; +using ::gets _LIBCPP_USING_IF_EXISTS; #endif -using ::scanf; -using ::vscanf; +using ::scanf _LIBCPP_USING_IF_EXISTS; +using ::vscanf _LIBCPP_USING_IF_EXISTS; #endif #ifndef _LIBCPP_HAS_NO_STDOUT -using ::printf; -using ::putchar; -using ::puts; -using ::vprintf; +using ::printf _LIBCPP_USING_IF_EXISTS; +using ::putchar _LIBCPP_USING_IF_EXISTS; +using ::puts _LIBCPP_USING_IF_EXISTS; +using ::vprintf _LIBCPP_USING_IF_EXISTS; #endif _LIBCPP_END_NAMESPACE_STD -#endif // _LIBCPP_CSTDIO +#endif // _LIBCPP_CSTDIO diff --git a/lib/libcxx/include/cstdlib b/lib/libcxx/include/cstdlib index 675a12d9e..ced0321aa 100644 --- a/lib/libcxx/include/cstdlib +++ b/lib/libcxx/include/cstdlib @@ -96,68 +96,68 @@ void *aligned_alloc(size_t alignment, size_t size); // C11 _LIBCPP_BEGIN_NAMESPACE_STD -using ::size_t; -using ::div_t; -using ::ldiv_t; +using ::size_t _LIBCPP_USING_IF_EXISTS; +using ::div_t _LIBCPP_USING_IF_EXISTS; +using ::ldiv_t _LIBCPP_USING_IF_EXISTS; #ifndef _LIBCPP_HAS_NO_LONG_LONG -using ::lldiv_t; +using ::lldiv_t _LIBCPP_USING_IF_EXISTS; #endif // _LIBCPP_HAS_NO_LONG_LONG -using ::atof; -using ::atoi; -using ::atol; +using ::atof _LIBCPP_USING_IF_EXISTS; +using ::atoi _LIBCPP_USING_IF_EXISTS; +using ::atol _LIBCPP_USING_IF_EXISTS; #ifndef _LIBCPP_HAS_NO_LONG_LONG -using ::atoll; +using ::atoll _LIBCPP_USING_IF_EXISTS; #endif // _LIBCPP_HAS_NO_LONG_LONG -using ::strtod; -using ::strtof; -using ::strtold; -using ::strtol; +using ::strtod _LIBCPP_USING_IF_EXISTS; +using ::strtof _LIBCPP_USING_IF_EXISTS; +using ::strtold _LIBCPP_USING_IF_EXISTS; +using ::strtol _LIBCPP_USING_IF_EXISTS; #ifndef _LIBCPP_HAS_NO_LONG_LONG -using ::strtoll; +using ::strtoll _LIBCPP_USING_IF_EXISTS; #endif // _LIBCPP_HAS_NO_LONG_LONG -using ::strtoul; +using ::strtoul _LIBCPP_USING_IF_EXISTS; #ifndef _LIBCPP_HAS_NO_LONG_LONG -using ::strtoull; +using ::strtoull _LIBCPP_USING_IF_EXISTS; #endif // _LIBCPP_HAS_NO_LONG_LONG -using ::rand; -using ::srand; -using ::calloc; -using ::free; -using ::malloc; -using ::realloc; -using ::abort; -using ::atexit; -using ::exit; -using ::_Exit; +using ::rand _LIBCPP_USING_IF_EXISTS; +using ::srand _LIBCPP_USING_IF_EXISTS; +using ::calloc _LIBCPP_USING_IF_EXISTS; +using ::free _LIBCPP_USING_IF_EXISTS; +using ::malloc _LIBCPP_USING_IF_EXISTS; +using ::realloc _LIBCPP_USING_IF_EXISTS; +using ::abort _LIBCPP_USING_IF_EXISTS; +using ::atexit _LIBCPP_USING_IF_EXISTS; +using ::exit _LIBCPP_USING_IF_EXISTS; +using ::_Exit _LIBCPP_USING_IF_EXISTS; #ifndef _LIBCPP_WINDOWS_STORE_APP -using ::getenv; -using ::system; +using ::getenv _LIBCPP_USING_IF_EXISTS; +using ::system _LIBCPP_USING_IF_EXISTS; #endif -using ::bsearch; -using ::qsort; -using ::abs; -using ::labs; +using ::bsearch _LIBCPP_USING_IF_EXISTS; +using ::qsort _LIBCPP_USING_IF_EXISTS; +using ::abs _LIBCPP_USING_IF_EXISTS; +using ::labs _LIBCPP_USING_IF_EXISTS; #ifndef _LIBCPP_HAS_NO_LONG_LONG -using ::llabs; +using ::llabs _LIBCPP_USING_IF_EXISTS; #endif // _LIBCPP_HAS_NO_LONG_LONG -using ::div; -using ::ldiv; +using ::div _LIBCPP_USING_IF_EXISTS; +using ::ldiv _LIBCPP_USING_IF_EXISTS; #ifndef _LIBCPP_HAS_NO_LONG_LONG -using ::lldiv; +using ::lldiv _LIBCPP_USING_IF_EXISTS; #endif // _LIBCPP_HAS_NO_LONG_LONG -using ::mblen; -using ::mbtowc; -using ::wctomb; -using ::mbstowcs; -using ::wcstombs; +using ::mblen _LIBCPP_USING_IF_EXISTS; +using ::mbtowc _LIBCPP_USING_IF_EXISTS; +using ::wctomb _LIBCPP_USING_IF_EXISTS; +using ::mbstowcs _LIBCPP_USING_IF_EXISTS; +using ::wcstombs _LIBCPP_USING_IF_EXISTS; #if !defined(_LIBCPP_CXX03_LANG) && defined(_LIBCPP_HAS_QUICK_EXIT) -using ::at_quick_exit; -using ::quick_exit; +using ::at_quick_exit _LIBCPP_USING_IF_EXISTS; +using ::quick_exit _LIBCPP_USING_IF_EXISTS; #endif #if _LIBCPP_STD_VER > 14 && defined(_LIBCPP_HAS_ALIGNED_ALLOC) -using ::aligned_alloc; +using ::aligned_alloc _LIBCPP_USING_IF_EXISTS; #endif _LIBCPP_END_NAMESPACE_STD -#endif // _LIBCPP_CSTDLIB +#endif // _LIBCPP_CSTDLIB diff --git a/lib/libcxx/include/cstring b/lib/libcxx/include/cstring index 8bc96a023..34449ab33 100644 --- a/lib/libcxx/include/cstring +++ b/lib/libcxx/include/cstring @@ -65,32 +65,32 @@ size_t strlen(const char* s); _LIBCPP_BEGIN_NAMESPACE_STD -using ::size_t; -using ::memcpy; -using ::memmove; -using ::strcpy; -using ::strncpy; -using ::strcat; -using ::strncat; -using ::memcmp; -using ::strcmp; -using ::strncmp; -using ::strcoll; -using ::strxfrm; -using ::memchr; -using ::strchr; -using ::strcspn; -using ::strpbrk; -using ::strrchr; -using ::strspn; -using ::strstr; +using ::size_t _LIBCPP_USING_IF_EXISTS; +using ::memcpy _LIBCPP_USING_IF_EXISTS; +using ::memmove _LIBCPP_USING_IF_EXISTS; +using ::strcpy _LIBCPP_USING_IF_EXISTS; +using ::strncpy _LIBCPP_USING_IF_EXISTS; +using ::strcat _LIBCPP_USING_IF_EXISTS; +using ::strncat _LIBCPP_USING_IF_EXISTS; +using ::memcmp _LIBCPP_USING_IF_EXISTS; +using ::strcmp _LIBCPP_USING_IF_EXISTS; +using ::strncmp _LIBCPP_USING_IF_EXISTS; +using ::strcoll _LIBCPP_USING_IF_EXISTS; +using ::strxfrm _LIBCPP_USING_IF_EXISTS; +using ::memchr _LIBCPP_USING_IF_EXISTS; +using ::strchr _LIBCPP_USING_IF_EXISTS; +using ::strcspn _LIBCPP_USING_IF_EXISTS; +using ::strpbrk _LIBCPP_USING_IF_EXISTS; +using ::strrchr _LIBCPP_USING_IF_EXISTS; +using ::strspn _LIBCPP_USING_IF_EXISTS; +using ::strstr _LIBCPP_USING_IF_EXISTS; #ifndef _LIBCPP_HAS_NO_THREAD_UNSAFE_C_FUNCTIONS -using ::strtok; +using ::strtok _LIBCPP_USING_IF_EXISTS; #endif -using ::memset; -using ::strerror; -using ::strlen; +using ::memset _LIBCPP_USING_IF_EXISTS; +using ::strerror _LIBCPP_USING_IF_EXISTS; +using ::strlen _LIBCPP_USING_IF_EXISTS; _LIBCPP_END_NAMESPACE_STD -#endif // _LIBCPP_CSTRING +#endif // _LIBCPP_CSTRING diff --git a/lib/libcxx/include/ctgmath b/lib/libcxx/include/ctgmath index ba1eeeac9..41f7f0a17 100644 --- a/lib/libcxx/include/ctgmath +++ b/lib/libcxx/include/ctgmath @@ -25,4 +25,4 @@ #pragma GCC system_header #endif -#endif // _LIBCPP_CTGMATH +#endif // _LIBCPP_CTGMATH diff --git a/lib/libcxx/include/ctime b/lib/libcxx/include/ctime index b0e6c65af..2a3fdd12e 100644 --- a/lib/libcxx/include/ctime +++ b/lib/libcxx/include/ctime @@ -59,7 +59,7 @@ int timespec_get( struct timespec *ts, int base); // C++17 // we're detecting this here instead of in <__config> because we can't include // system headers from <__config>, since it leads to circular module dependencies. // This is also meant to be a very temporary workaround until the SDKs are fixed. -#if defined(__APPLE__) +#if defined(__APPLE__) && !__has_attribute(using_if_exists) # include # if defined(_LIBCPP_HAS_TIMESPEC_GET) && (__DARWIN_C_LEVEL < __DARWIN_C_FULL) # define _LIBCPP_HAS_TIMESPEC_GET_NOT_ACTUALLY_PROVIDED @@ -68,28 +68,28 @@ int timespec_get( struct timespec *ts, int base); // C++17 _LIBCPP_BEGIN_NAMESPACE_STD -using ::clock_t; -using ::size_t; -using ::time_t; -using ::tm; +using ::clock_t _LIBCPP_USING_IF_EXISTS; +using ::size_t _LIBCPP_USING_IF_EXISTS; +using ::time_t _LIBCPP_USING_IF_EXISTS; +using ::tm _LIBCPP_USING_IF_EXISTS; #if _LIBCPP_STD_VER > 14 && defined(_LIBCPP_HAS_TIMESPEC_GET) -using ::timespec; +using ::timespec _LIBCPP_USING_IF_EXISTS; #endif -using ::clock; -using ::difftime; -using ::mktime; -using ::time; +using ::clock _LIBCPP_USING_IF_EXISTS; +using ::difftime _LIBCPP_USING_IF_EXISTS; +using ::mktime _LIBCPP_USING_IF_EXISTS; +using ::time _LIBCPP_USING_IF_EXISTS; #ifndef _LIBCPP_HAS_NO_THREAD_UNSAFE_C_FUNCTIONS -using ::asctime; -using ::ctime; -using ::gmtime; -using ::localtime; +using ::asctime _LIBCPP_USING_IF_EXISTS; +using ::ctime _LIBCPP_USING_IF_EXISTS; +using ::gmtime _LIBCPP_USING_IF_EXISTS; +using ::localtime _LIBCPP_USING_IF_EXISTS; #endif -using ::strftime; +using ::strftime _LIBCPP_USING_IF_EXISTS; #if _LIBCPP_STD_VER > 14 && defined(_LIBCPP_HAS_TIMESPEC_GET) && !defined(_LIBCPP_HAS_TIMESPEC_GET_NOT_ACTUALLY_PROVIDED) -using ::timespec_get; +using ::timespec_get _LIBCPP_USING_IF_EXISTS; #endif _LIBCPP_END_NAMESPACE_STD -#endif // _LIBCPP_CTIME +#endif // _LIBCPP_CTIME diff --git a/lib/libcxx/include/ctype.h b/lib/libcxx/include/ctype.h index dcc7935a2..967b6be1e 100644 --- a/lib/libcxx/include/ctype.h +++ b/lib/libcxx/include/ctype.h @@ -56,4 +56,4 @@ int toupper(int c); #endif -#endif // _LIBCPP_CTYPE_H +#endif // _LIBCPP_CTYPE_H diff --git a/lib/libcxx/include/cwchar b/lib/libcxx/include/cwchar index 451c621f9..f39046f0a 100644 --- a/lib/libcxx/include/cwchar +++ b/lib/libcxx/include/cwchar @@ -112,81 +112,81 @@ size_t wcsrtombs(char* restrict dst, const wchar_t** restrict src, size_t len, _LIBCPP_BEGIN_NAMESPACE_STD -using ::mbstate_t; -using ::size_t; -using ::tm; -using ::wint_t; -using ::FILE; -using ::fwprintf; -using ::fwscanf; -using ::swprintf; -using ::vfwprintf; -using ::vswprintf; -using ::swscanf; -using ::vfwscanf; -using ::vswscanf; -using ::fgetwc; -using ::fgetws; -using ::fputwc; -using ::fputws; -using ::fwide; -using ::getwc; -using ::putwc; -using ::ungetwc; -using ::wcstod; -using ::wcstof; -using ::wcstold; -using ::wcstol; +using ::mbstate_t _LIBCPP_USING_IF_EXISTS; +using ::size_t _LIBCPP_USING_IF_EXISTS; +using ::tm _LIBCPP_USING_IF_EXISTS; +using ::wint_t _LIBCPP_USING_IF_EXISTS; +using ::FILE _LIBCPP_USING_IF_EXISTS; +using ::fwprintf _LIBCPP_USING_IF_EXISTS; +using ::fwscanf _LIBCPP_USING_IF_EXISTS; +using ::swprintf _LIBCPP_USING_IF_EXISTS; +using ::vfwprintf _LIBCPP_USING_IF_EXISTS; +using ::vswprintf _LIBCPP_USING_IF_EXISTS; +using ::swscanf _LIBCPP_USING_IF_EXISTS; +using ::vfwscanf _LIBCPP_USING_IF_EXISTS; +using ::vswscanf _LIBCPP_USING_IF_EXISTS; +using ::fgetwc _LIBCPP_USING_IF_EXISTS; +using ::fgetws _LIBCPP_USING_IF_EXISTS; +using ::fputwc _LIBCPP_USING_IF_EXISTS; +using ::fputws _LIBCPP_USING_IF_EXISTS; +using ::fwide _LIBCPP_USING_IF_EXISTS; +using ::getwc _LIBCPP_USING_IF_EXISTS; +using ::putwc _LIBCPP_USING_IF_EXISTS; +using ::ungetwc _LIBCPP_USING_IF_EXISTS; +using ::wcstod _LIBCPP_USING_IF_EXISTS; +using ::wcstof _LIBCPP_USING_IF_EXISTS; +using ::wcstold _LIBCPP_USING_IF_EXISTS; +using ::wcstol _LIBCPP_USING_IF_EXISTS; #ifndef _LIBCPP_HAS_NO_LONG_LONG -using ::wcstoll; +using ::wcstoll _LIBCPP_USING_IF_EXISTS; #endif // _LIBCPP_HAS_NO_LONG_LONG -using ::wcstoul; +using ::wcstoul _LIBCPP_USING_IF_EXISTS; #ifndef _LIBCPP_HAS_NO_LONG_LONG -using ::wcstoull; +using ::wcstoull _LIBCPP_USING_IF_EXISTS; #endif // _LIBCPP_HAS_NO_LONG_LONG -using ::wcscpy; -using ::wcsncpy; -using ::wcscat; -using ::wcsncat; -using ::wcscmp; -using ::wcscoll; -using ::wcsncmp; -using ::wcsxfrm; -using ::wcschr; -using ::wcspbrk; -using ::wcsrchr; -using ::wcsstr; -using ::wmemchr; -using ::wcscspn; -using ::wcslen; -using ::wcsspn; -using ::wcstok; -using ::wmemcmp; -using ::wmemcpy; -using ::wmemmove; -using ::wmemset; -using ::wcsftime; -using ::btowc; -using ::wctob; -using ::mbsinit; -using ::mbrlen; -using ::mbrtowc; -using ::wcrtomb; -using ::mbsrtowcs; -using ::wcsrtombs; +using ::wcscpy _LIBCPP_USING_IF_EXISTS; +using ::wcsncpy _LIBCPP_USING_IF_EXISTS; +using ::wcscat _LIBCPP_USING_IF_EXISTS; +using ::wcsncat _LIBCPP_USING_IF_EXISTS; +using ::wcscmp _LIBCPP_USING_IF_EXISTS; +using ::wcscoll _LIBCPP_USING_IF_EXISTS; +using ::wcsncmp _LIBCPP_USING_IF_EXISTS; +using ::wcsxfrm _LIBCPP_USING_IF_EXISTS; +using ::wcschr _LIBCPP_USING_IF_EXISTS; +using ::wcspbrk _LIBCPP_USING_IF_EXISTS; +using ::wcsrchr _LIBCPP_USING_IF_EXISTS; +using ::wcsstr _LIBCPP_USING_IF_EXISTS; +using ::wmemchr _LIBCPP_USING_IF_EXISTS; +using ::wcscspn _LIBCPP_USING_IF_EXISTS; +using ::wcslen _LIBCPP_USING_IF_EXISTS; +using ::wcsspn _LIBCPP_USING_IF_EXISTS; +using ::wcstok _LIBCPP_USING_IF_EXISTS; +using ::wmemcmp _LIBCPP_USING_IF_EXISTS; +using ::wmemcpy _LIBCPP_USING_IF_EXISTS; +using ::wmemmove _LIBCPP_USING_IF_EXISTS; +using ::wmemset _LIBCPP_USING_IF_EXISTS; +using ::wcsftime _LIBCPP_USING_IF_EXISTS; +using ::btowc _LIBCPP_USING_IF_EXISTS; +using ::wctob _LIBCPP_USING_IF_EXISTS; +using ::mbsinit _LIBCPP_USING_IF_EXISTS; +using ::mbrlen _LIBCPP_USING_IF_EXISTS; +using ::mbrtowc _LIBCPP_USING_IF_EXISTS; +using ::wcrtomb _LIBCPP_USING_IF_EXISTS; +using ::mbsrtowcs _LIBCPP_USING_IF_EXISTS; +using ::wcsrtombs _LIBCPP_USING_IF_EXISTS; #ifndef _LIBCPP_HAS_NO_STDIN -using ::getwchar; -using ::vwscanf; -using ::wscanf; +using ::getwchar _LIBCPP_USING_IF_EXISTS; +using ::vwscanf _LIBCPP_USING_IF_EXISTS; +using ::wscanf _LIBCPP_USING_IF_EXISTS; #endif #ifndef _LIBCPP_HAS_NO_STDOUT -using ::putwchar; -using ::vwprintf; -using ::wprintf; +using ::putwchar _LIBCPP_USING_IF_EXISTS; +using ::vwprintf _LIBCPP_USING_IF_EXISTS; +using ::wprintf _LIBCPP_USING_IF_EXISTS; #endif _LIBCPP_END_NAMESPACE_STD -#endif // _LIBCPP_CWCHAR +#endif // _LIBCPP_CWCHAR diff --git a/lib/libcxx/include/cwctype b/lib/libcxx/include/cwctype index 575fd5661..27eea2f15 100644 --- a/lib/libcxx/include/cwctype +++ b/lib/libcxx/include/cwctype @@ -59,28 +59,30 @@ wctrans_t wctrans(const char* property); _LIBCPP_BEGIN_NAMESPACE_STD -using ::wint_t; -using ::wctrans_t; -using ::wctype_t; -using ::iswalnum; -using ::iswalpha; -using ::iswblank; -using ::iswcntrl; -using ::iswdigit; -using ::iswgraph; -using ::iswlower; -using ::iswprint; -using ::iswpunct; -using ::iswspace; -using ::iswupper; -using ::iswxdigit; -using ::iswctype; -using ::wctype; -using ::towlower; -using ::towupper; -using ::towctrans; -using ::wctrans; +#if defined(_LIBCPP_INCLUDED_C_LIBRARY_WCTYPE_H) +using ::wint_t _LIBCPP_USING_IF_EXISTS; +using ::wctrans_t _LIBCPP_USING_IF_EXISTS; +using ::wctype_t _LIBCPP_USING_IF_EXISTS; +using ::iswalnum _LIBCPP_USING_IF_EXISTS; +using ::iswalpha _LIBCPP_USING_IF_EXISTS; +using ::iswblank _LIBCPP_USING_IF_EXISTS; +using ::iswcntrl _LIBCPP_USING_IF_EXISTS; +using ::iswdigit _LIBCPP_USING_IF_EXISTS; +using ::iswgraph _LIBCPP_USING_IF_EXISTS; +using ::iswlower _LIBCPP_USING_IF_EXISTS; +using ::iswprint _LIBCPP_USING_IF_EXISTS; +using ::iswpunct _LIBCPP_USING_IF_EXISTS; +using ::iswspace _LIBCPP_USING_IF_EXISTS; +using ::iswupper _LIBCPP_USING_IF_EXISTS; +using ::iswxdigit _LIBCPP_USING_IF_EXISTS; +using ::iswctype _LIBCPP_USING_IF_EXISTS; +using ::wctype _LIBCPP_USING_IF_EXISTS; +using ::towlower _LIBCPP_USING_IF_EXISTS; +using ::towupper _LIBCPP_USING_IF_EXISTS; +using ::towctrans _LIBCPP_USING_IF_EXISTS; +using ::wctrans _LIBCPP_USING_IF_EXISTS; +#endif // _LIBCPP_INCLUDED_C_LIBRARY_WCTYPE_H _LIBCPP_END_NAMESPACE_STD -#endif // _LIBCPP_CWCTYPE +#endif // _LIBCPP_CWCTYPE diff --git a/lib/libcxx/include/deque b/lib/libcxx/include/deque index c6517d170..aff93deb1 100644 --- a/lib/libcxx/include/deque +++ b/lib/libcxx/include/deque @@ -161,12 +161,16 @@ template */ #include <__config> +#include <__debug> #include <__split_buffer> -#include +#include <__utility/forward.h> +#include +#include #include #include -#include +#include #include +#include #include #if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) @@ -1055,7 +1059,7 @@ public: __deque_base(__deque_base&& __c) _NOEXCEPT_(is_nothrow_move_constructible::value); __deque_base(__deque_base&& __c, const allocator_type& __a); -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG void swap(__deque_base& __c) #if _LIBCPP_STD_VER >= 14 @@ -1222,7 +1226,7 @@ __deque_base<_Tp, _Allocator>::__deque_base(__deque_base&& __c, const allocator_ } } -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG template void @@ -1315,7 +1319,7 @@ public: deque(_InputIter __f, _InputIter __l, const allocator_type& __a, typename enable_if<__is_cpp17_input_iterator<_InputIter>::value>::type* = 0); deque(const deque& __c); - deque(const deque& __c, const allocator_type& __a); + deque(const deque& __c, const __identity_t& __a); deque& operator=(const deque& __c); @@ -1329,7 +1333,7 @@ public: _LIBCPP_INLINE_VISIBILITY deque(deque&& __c) _NOEXCEPT_(is_nothrow_move_constructible<__base>::value); _LIBCPP_INLINE_VISIBILITY - deque(deque&& __c, const allocator_type& __a); + deque(deque&& __c, const __identity_t& __a); _LIBCPP_INLINE_VISIBILITY deque& operator=(deque&& __c) _NOEXCEPT_(__alloc_traits::propagate_on_container_move_assignment::value && @@ -1337,7 +1341,7 @@ public: _LIBCPP_INLINE_VISIBILITY void assign(initializer_list __il) {assign(__il.begin(), __il.end());} -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG template void assign(_InputIter __f, _InputIter __l, @@ -1440,7 +1444,7 @@ public: _LIBCPP_INLINE_VISIBILITY iterator insert(const_iterator __p, initializer_list __il) {return insert(__p, __il.begin(), __il.end());} -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG iterator insert(const_iterator __p, const value_type& __v); iterator insert(const_iterator __p, size_type __n, const value_type& __v); template @@ -1586,18 +1590,18 @@ public: #ifndef _LIBCPP_HAS_NO_DEDUCTION_GUIDES template::value_type>, - class = typename enable_if<__is_allocator<_Alloc>::value, void>::type + class _Alloc = allocator<__iter_value_type<_InputIterator>>, + class = _EnableIf<__is_allocator<_Alloc>::value> > deque(_InputIterator, _InputIterator) - -> deque::value_type, _Alloc>; + -> deque<__iter_value_type<_InputIterator>, _Alloc>; template::value, void>::type + class = _EnableIf<__is_allocator<_Alloc>::value> > deque(_InputIterator, _InputIterator, _Alloc) - -> deque::value_type, _Alloc>; + -> deque<__iter_value_type<_InputIterator>, _Alloc>; #endif @@ -1658,7 +1662,7 @@ deque<_Tp, _Allocator>::deque(const deque& __c) } template -deque<_Tp, _Allocator>::deque(const deque& __c, const allocator_type& __a) +deque<_Tp, _Allocator>::deque(const deque& __c, const __identity_t& __a) : __base(__a) { __append(__c.begin(), __c.end()); @@ -1701,7 +1705,7 @@ deque<_Tp, _Allocator>::deque(deque&& __c) template inline -deque<_Tp, _Allocator>::deque(deque&& __c, const allocator_type& __a) +deque<_Tp, _Allocator>::deque(deque&& __c, const __identity_t& __a) : __base(_VSTD::move(__c), __a) { if (__a != __c.__alloc()) @@ -1746,7 +1750,7 @@ deque<_Tp, _Allocator>::__move_assign(deque& __c, true_type) __base::__move_assign(__c); } -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG template template @@ -2128,7 +2132,7 @@ deque<_Tp, _Allocator>::emplace(const_iterator __p, _Args&&... __args) return __base::begin() + __pos; } -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG template @@ -2532,7 +2536,7 @@ deque<_Tp, _Allocator>::__add_front_capacity(size_type __n) #ifndef _LIBCPP_NO_EXCEPTIONS try { -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS for (; __nb > 0; --__nb) __buf.push_back(__alloc_traits::allocate(__a, __base::__block_size)); #ifndef _LIBCPP_NO_EXCEPTIONS @@ -2544,7 +2548,7 @@ deque<_Tp, _Allocator>::__add_front_capacity(size_type __n) __alloc_traits::deallocate(__a, *__i, __base::__block_size); throw; } -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS for (; __back_capacity > 0; --__back_capacity) { __buf.push_back(__base::__map_.back()); @@ -2674,7 +2678,7 @@ deque<_Tp, _Allocator>::__add_back_capacity(size_type __n) #ifndef _LIBCPP_NO_EXCEPTIONS try { -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS for (; __nb > 0; --__nb) __buf.push_back(__alloc_traits::allocate(__a, __base::__block_size)); #ifndef _LIBCPP_NO_EXCEPTIONS @@ -2686,7 +2690,7 @@ deque<_Tp, _Allocator>::__add_back_capacity(size_type __n) __alloc_traits::deallocate(__a, *__i, __base::__block_size); throw; } -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS for (; __front_capacity > 0; --__front_capacity) { __buf.push_back(__base::__map_.front()); @@ -2720,7 +2724,7 @@ template void deque<_Tp, _Allocator>::pop_back() { - _LIBCPP_ASSERT(!empty(), "deque::pop_back called for empty deque"); + _LIBCPP_ASSERT(!empty(), "deque::pop_back called on an empty deque"); allocator_type& __a = __base::__alloc(); size_type __p = __base::size() + __base::__start_ - 1; __alloc_traits::destroy(__a, _VSTD::__to_address(*(__base::__map_.begin() + @@ -3044,4 +3048,4 @@ _LIBCPP_END_NAMESPACE_STD _LIBCPP_POP_MACROS -#endif // _LIBCPP_DEQUE +#endif // _LIBCPP_DEQUE diff --git a/lib/libcxx/include/errno.h b/lib/libcxx/include/errno.h index 447319eff..59950c24c 100644 --- a/lib/libcxx/include/errno.h +++ b/lib/libcxx/include/errno.h @@ -72,9 +72,9 @@ static const int __elast2 = 105; #define ELAST ENOTRECOVERABLE #endif -#endif // defined(EOWNERDEAD) +#endif // defined(EOWNERDEAD) -#endif // !defined(EOWNERDEAD) || !defined(ENOTRECOVERABLE) +#endif // !defined(EOWNERDEAD) || !defined(ENOTRECOVERABLE) // supply errno values likely to be missing, particularly on Windows @@ -394,4 +394,4 @@ static const int __elast2 = 105; #endif // __cplusplus -#endif // _LIBCPP_ERRNO_H +#endif // _LIBCPP_ERRNO_H diff --git a/lib/libcxx/include/exception b/lib/libcxx/include/exception index 4bf4049f3..816f259f8 100644 --- a/lib/libcxx/include/exception +++ b/lib/libcxx/include/exception @@ -76,9 +76,9 @@ template void rethrow_if_nested(const E& e); */ -#include <__config> #include <__availability> -#include <__memory/base.h> +#include <__config> +#include <__memory/addressof.h> #include #include #include @@ -151,7 +151,7 @@ public: exception_ptr& operator=(const exception_ptr&) _NOEXCEPT; ~exception_ptr() _NOEXCEPT; - _LIBCPP_INLINE_VISIBILITY _LIBCPP_EXPLICIT operator bool() const _NOEXCEPT + _LIBCPP_INLINE_VISIBILITY explicit operator bool() const _NOEXCEPT {return __ptr_ != nullptr;} friend _LIBCPP_INLINE_VISIBILITY @@ -205,7 +205,7 @@ public: exception_ptr& operator=(const exception_ptr& __other) _NOEXCEPT; exception_ptr& operator=(nullptr_t) _NOEXCEPT; ~exception_ptr() _NOEXCEPT; - _LIBCPP_EXPLICIT operator bool() const _NOEXCEPT; + explicit operator bool() const _NOEXCEPT; }; _LIBCPP_FUNC_VIS @@ -266,7 +266,7 @@ struct __throw_with_nested<_Tp, _Up, true> { _LIBCPP_NORETURN static inline _LIBCPP_INLINE_VISIBILITY void __do_throw(_Tp&& __t) { - throw __nested<_Up>(_VSTD::forward<_Tp>(__t)); + throw __nested<_Up>(static_cast<_Tp&&>(__t)); } }; @@ -277,9 +277,9 @@ struct __throw_with_nested<_Tp, _Up, false> { __do_throw(_Tp&& __t) #else __do_throw (_Tp& __t) -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG { - throw _VSTD::forward<_Tp>(__t); + throw static_cast<_Tp&&>(__t); } }; #endif @@ -296,7 +296,7 @@ throw_with_nested(_Tp&& __t) is_class<_Up>::value && !is_base_of::value && !__libcpp_is_final<_Up>::value>:: - __do_throw(_VSTD::forward<_Tp>(__t)); + __do_throw(static_cast<_Tp&&>(__t)); #else ((void)__t); // FIXME: Make this abort @@ -330,4 +330,4 @@ rethrow_if_nested(const _Ep&, } // std -#endif // _LIBCPP_EXCEPTION +#endif // _LIBCPP_EXCEPTION diff --git a/lib/libcxx/include/experimental/__config b/lib/libcxx/include/experimental/__config index 375dc3743..f85426d05 100644 --- a/lib/libcxx/include/experimental/__config +++ b/lib/libcxx/include/experimental/__config @@ -32,10 +32,6 @@ #define _LIBCPP_END_NAMESPACE_LFTS_PMR _LIBCPP_END_NAMESPACE_LFTS } #define _VSTD_LFTS_PMR _VSTD_LFTS::pmr -#define _LIBCPP_BEGIN_NAMESPACE_CHRONO_LFTS _LIBCPP_BEGIN_NAMESPACE_STD \ - namespace chrono { namespace experimental { inline namespace fundamentals_v1 { -#define _LIBCPP_END_NAMESPACE_CHRONO_LFTS _LIBCPP_END_NAMESPACE_STD } } } - #if defined(_LIBCPP_NO_EXPERIMENTAL_DEPRECATION_WARNING_FILESYSTEM) # define _LIBCPP_DEPRECATED_EXPERIMENTAL_FILESYSTEM /* nothing */ #else diff --git a/lib/libcxx/include/experimental/__memory b/lib/libcxx/include/experimental/__memory index 4cf897846..b38b664b3 100644 --- a/lib/libcxx/include/experimental/__memory +++ b/lib/libcxx/include/experimental/__memory @@ -10,6 +10,8 @@ #ifndef _LIBCPP_EXPERIMENTAL___MEMORY #define _LIBCPP_EXPERIMENTAL___MEMORY +#include <__memory/allocator_arg_t.h> +#include <__memory/uses_allocator.h> #include #include // for erased_type #include <__functional_base> @@ -73,12 +75,35 @@ struct __lfts_uses_alloc_ctor > {}; +template +inline _LIBCPP_INLINE_VISIBILITY +void __user_alloc_construct_impl (integral_constant, _Tp *__storage, const _Allocator &, _Args &&... __args ) +{ + new (__storage) _Tp (_VSTD::forward<_Args>(__args)...); +} + +// FIXME: This should have a version which takes a non-const alloc. +template +inline _LIBCPP_INLINE_VISIBILITY +void __user_alloc_construct_impl (integral_constant, _Tp *__storage, const _Allocator &__a, _Args &&... __args ) +{ + new (__storage) _Tp (allocator_arg, __a, _VSTD::forward<_Args>(__args)...); +} + +// FIXME: This should have a version which takes a non-const alloc. +template +inline _LIBCPP_INLINE_VISIBILITY +void __user_alloc_construct_impl (integral_constant, _Tp *__storage, const _Allocator &__a, _Args &&... __args ) +{ + new (__storage) _Tp (_VSTD::forward<_Args>(__args)..., __a); +} + template inline _LIBCPP_INLINE_VISIBILITY void __lfts_user_alloc_construct( _Tp * __store, const _Alloc & __a, _Args &&... __args) { - _VSTD::__user_alloc_construct_impl( + ::std::experimental::fundamentals_v1::__user_alloc_construct_impl( typename __lfts_uses_alloc_ctor<_Tp, _Alloc, _Args...>::type() , __store, __a, _VSTD::forward<_Args>(__args)... ); diff --git a/lib/libcxx/include/experimental/functional b/lib/libcxx/include/experimental/functional index c7dda2254..e3220e16c 100644 --- a/lib/libcxx/include/experimental/functional +++ b/lib/libcxx/include/experimental/functional @@ -86,6 +86,7 @@ inline namespace fundamentals_v1 { */ +#include <__memory/uses_allocator.h> #include #include #include @@ -109,7 +110,7 @@ _LIBCPP_BEGIN_NAMESPACE_LFTS #if _LIBCPP_STD_VER > 11 // default searcher template> -class _LIBCPP_TYPE_VIS default_searcher { +class _LIBCPP_TEMPLATE_VIS default_searcher { public: _LIBCPP_INLINE_VISIBILITY default_searcher(_ForwardIterator __f, _ForwardIterator __l, @@ -122,8 +123,8 @@ public: operator () (_ForwardIterator2 __f, _ForwardIterator2 __l) const { return _VSTD::__search(__f, __l, __first_, __last_, __pred_, - typename _VSTD::iterator_traits<_ForwardIterator>::iterator_category(), - typename _VSTD::iterator_traits<_ForwardIterator2>::iterator_category()); + typename iterator_traits<_ForwardIterator>::iterator_category(), + typename iterator_traits<_ForwardIterator2>::iterator_category()); } private: @@ -154,7 +155,7 @@ public: // TODO private: public: _LIBCPP_INLINE_VISIBILITY - _BMSkipTable(std::size_t __sz, _Value __default, _Hash __hf, _BinaryPredicate __pred) + _BMSkipTable(size_t __sz, _Value __default, _Hash __hf, _BinaryPredicate __pred) : __default_value_(__default), __table(__sz, __hf, __pred) {} _LIBCPP_INLINE_VISIBILITY @@ -179,13 +180,13 @@ private: typedef _Value value_type; typedef _Key key_type; - typedef typename std::make_unsigned::type unsigned_key_type; - typedef std::array::max()> skip_map; + typedef typename make_unsigned::type unsigned_key_type; + typedef std::array::max()> skip_map; skip_map __table; public: _LIBCPP_INLINE_VISIBILITY - _BMSkipTable(std::size_t /*__sz*/, _Value __default, _Hash /*__hf*/, _BinaryPredicate /*__pred*/) + _BMSkipTable(size_t /*__sz*/, _Value __default, _Hash /*__hf*/, _BinaryPredicate /*__pred*/) { std::fill_n(__table.begin(), __table.size(), __default); } @@ -207,12 +208,12 @@ public: template ::value_type>, class _BinaryPredicate = equal_to<>> -class _LIBCPP_TYPE_VIS boyer_moore_searcher { +class _LIBCPP_TEMPLATE_VIS boyer_moore_searcher { private: typedef typename std::iterator_traits<_RandomAccessIterator1>::difference_type difference_type; typedef typename std::iterator_traits<_RandomAccessIterator1>::value_type value_type; typedef _BMSkipTable::value && // what about enums? + is_integral::value && // what about enums? sizeof(value_type) == 1 && is_same<_Hash, hash>::value && is_same<_BinaryPredicate, equal_to<>>::value @@ -247,7 +248,7 @@ public: if (__first_ == __last_) return make_pair(__f, __f); // empty pattern // If the pattern is larger than the corpus, we can't find it! - if ( __pattern_length_ > _VSTD::distance (__f, __l)) + if ( __pattern_length_ > _VSTD::distance(__f, __l)) return make_pair(__l, __l); // Do the search @@ -299,11 +300,11 @@ public: // TODO private: template void __compute_bm_prefix ( _Iterator __f, _Iterator __l, _BinaryPredicate __pred, _Container &__prefix ) { - const std::size_t __count = _VSTD::distance(__f, __l); + const size_t __count = _VSTD::distance(__f, __l); __prefix[0] = 0; - std::size_t __k = 0; - for ( std::size_t __i = 1; __i < __count; ++__i ) + size_t __k = 0; + for ( size_t __i = 1; __i < __count; ++__i ) { while ( __k > 0 && !__pred ( __f[__k], __f[__i] )) __k = __prefix [ __k - 1 ]; @@ -317,22 +318,22 @@ public: // TODO private: void __build_suffix_table(_RandomAccessIterator1 __f, _RandomAccessIterator1 __l, _BinaryPredicate __pred) { - const std::size_t __count = _VSTD::distance(__f, __l); + const size_t __count = _VSTD::distance(__f, __l); vector & __suffix = *__suffix_.get(); if (__count > 0) { - _VSTD::vector __scratch(__count); + vector __scratch(__count); __compute_bm_prefix(__f, __l, __pred, __scratch); - for ( std::size_t __i = 0; __i <= __count; __i++ ) + for ( size_t __i = 0; __i <= __count; __i++ ) __suffix[__i] = __count - __scratch[__count-1]; - typedef _VSTD::reverse_iterator<_RandomAccessIterator1> _RevIter; + typedef reverse_iterator<_RandomAccessIterator1> _RevIter; __compute_bm_prefix(_RevIter(__l), _RevIter(__f), __pred, __scratch); - for ( std::size_t __i = 0; __i < __count; __i++ ) + for ( size_t __i = 0; __i < __count; __i++ ) { - const std::size_t __j = __count - __scratch[__i]; + const size_t __j = __count - __scratch[__i]; const difference_type __k = __i - __scratch[__i] + 1; if (__suffix[__j] > __k) @@ -358,12 +359,12 @@ make_boyer_moore_searcher( _RandomAccessIterator __f, _RandomAccessIterator __l, template ::value_type>, class _BinaryPredicate = equal_to<>> -class _LIBCPP_TYPE_VIS boyer_moore_horspool_searcher { +class _LIBCPP_TEMPLATE_VIS boyer_moore_horspool_searcher { private: typedef typename std::iterator_traits<_RandomAccessIterator1>::difference_type difference_type; typedef typename std::iterator_traits<_RandomAccessIterator1>::value_type value_type; typedef _BMSkipTable::value && // what about enums? + is_integral::value && // what about enums? sizeof(value_type) == 1 && is_same<_Hash, hash>::value && is_same<_BinaryPredicate, equal_to<>>::value @@ -399,7 +400,7 @@ public: if (__first_ == __last_) return make_pair(__f, __f); // empty pattern // If the pattern is larger than the corpus, we can't find it! - if ( __pattern_length_ > _VSTD::distance (__f, __l)) + if ( __pattern_length_ > _VSTD::distance(__f, __l)) return make_pair(__l, __l); // Do the search diff --git a/lib/libcxx/include/experimental/iterator b/lib/libcxx/include/experimental/iterator index 10b059913..09ea2cbcc 100644 --- a/lib/libcxx/include/experimental/iterator +++ b/lib/libcxx/include/experimental/iterator @@ -56,6 +56,9 @@ namespace std { #if _LIBCPP_STD_VER > 11 +#include <__memory/addressof.h> +#include <__utility/move.h> +#include <__utility/forward.h> #include _LIBCPP_BEGIN_NAMESPACE_LFTS diff --git a/lib/libcxx/include/experimental/propagate_const b/lib/libcxx/include/experimental/propagate_const index 59f50c5e5..ce4b879b7 100644 --- a/lib/libcxx/include/experimental/propagate_const +++ b/lib/libcxx/include/experimental/propagate_const @@ -135,7 +135,7 @@ template class propagate_const { public: - typedef remove_reference_t())> element_type; + typedef remove_reference_t())> element_type; static_assert(!is_array<_Tp>::value, "Instantiation of propagate_const with an array type is ill-formed."); diff --git a/lib/libcxx/include/experimental/simd b/lib/libcxx/include/experimental/simd index 41f8f799a..d1aaf5504 100644 --- a/lib/libcxx/include/experimental/simd +++ b/lib/libcxx/include/experimental/simd @@ -725,12 +725,12 @@ constexpr size_t __ceil_pow_of_2(size_t __val) { template struct __vec_ext_traits { -#if !defined(_LIBCPP_COMPILER_CLANG) +#if !defined(_LIBCPP_COMPILER_CLANG_BASED) typedef _Tp type __attribute__((vector_size(__ceil_pow_of_2(__bytes)))); #endif }; -#if defined(_LIBCPP_COMPILER_CLANG) +#if defined(_LIBCPP_COMPILER_CLANG_BASED) #define _LIBCPP_SPECIALIZE_VEC_EXT(_TYPE, _NUM_ELEMENT) \ template <> \ struct __vec_ext_traits<_TYPE, sizeof(_TYPE) * _NUM_ELEMENT> { \ diff --git a/lib/libcxx/include/experimental/type_traits b/lib/libcxx/include/experimental/type_traits index 7ab097226..ea1335f96 100644 --- a/lib/libcxx/include/experimental/type_traits +++ b/lib/libcxx/include/experimental/type_traits @@ -105,7 +105,7 @@ using raw_invocation_type_t = typename raw_invocation_type<_Tp>::type; // 3.3.4, Detection idiom template using void_t = void; -struct nonesuch : private _VSTD::__nat { // make nonesuch "not an aggregate" +struct nonesuch : private __nat { // make nonesuch "not an aggregate" ~nonesuch() = delete; nonesuch (nonesuch const&) = delete; void operator=(nonesuch const&) = delete; diff --git a/lib/libcxx/include/ext/__hash b/lib/libcxx/include/ext/__hash index d03f101e9..fbeddf03a 100644 --- a/lib/libcxx/include/ext/__hash +++ b/lib/libcxx/include/ext/__hash @@ -12,6 +12,7 @@ #pragma GCC system_header +#include <__string> #include #include @@ -130,4 +131,4 @@ template <> struct _LIBCPP_TEMPLATE_VIS hash }; } -#endif // _LIBCPP_EXT_HASH +#endif // _LIBCPP_EXT_HASH diff --git a/lib/libcxx/include/ext/hash_map b/lib/libcxx/include/ext/hash_map index 2d6024cb9..60e32b09e 100644 --- a/lib/libcxx/include/ext/hash_map +++ b/lib/libcxx/include/ext/hash_map @@ -208,7 +208,7 @@ template #include #include -#if __DEPRECATED +#if defined(__DEPRECATED) && __DEPRECATED #if defined(_LIBCPP_WARNING) _LIBCPP_WARNING("Use of the header is deprecated. Migrate to ") #else @@ -350,7 +350,7 @@ public: { const_cast(__x.__value_constructed) = false; } -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG _LIBCPP_INLINE_VISIBILITY void operator()(pointer __p) @@ -981,4 +981,4 @@ operator!=(const hash_multimap<_Key, _Tp, _Hash, _Pred, _Alloc>& __x, } // __gnu_cxx -#endif // _LIBCPP_HASH_MAP +#endif // _LIBCPP_HASH_MAP diff --git a/lib/libcxx/include/ext/hash_set b/lib/libcxx/include/ext/hash_set index f0ba8d8de..af3f9c5de 100644 --- a/lib/libcxx/include/ext/hash_set +++ b/lib/libcxx/include/ext/hash_set @@ -197,7 +197,7 @@ template #include #include -#if __DEPRECATED +#if defined(__DEPRECATED) && __DEPRECATED #if defined(_LIBCPP_WARNING) _LIBCPP_WARNING("Use of the header is deprecated. Migrate to ") #else @@ -656,4 +656,4 @@ operator!=(const hash_multiset<_Value, _Hash, _Pred, _Alloc>& __x, } // __gnu_cxx -#endif // _LIBCPP_HASH_SET +#endif // _LIBCPP_HASH_SET diff --git a/lib/libcxx/include/filesystem b/lib/libcxx/include/filesystem index 92e37e183..61d6c659c 100644 --- a/lib/libcxx/include/filesystem +++ b/lib/libcxx/include/filesystem @@ -229,19 +229,22 @@ */ -#include <__config> #include <__availability> +#include <__config> +#include <__debug> +#include <__utility/forward.h> +#include +#include #include #include -#include -#include #include +#include #include #include #include +#include #include #include -#include #include #if !defined(_LIBCPP_HAS_NO_LOCALIZATION) @@ -249,8 +252,6 @@ # include // for quoted #endif -#include <__debug> - #if defined(_LIBCPP_HAS_NO_FILESYSTEM_LIBRARY) # error "The Filesystem library is not supported by this configuration of libc++" #endif @@ -276,6 +277,8 @@ struct _LIBCPP_TYPE_VIS space_info { uintmax_t available; }; +// On Windows, the library never identifies files as block, character, fifo +// or socket. enum class _LIBCPP_ENUM_VIS file_type : signed char { none = 0, not_found = -1, @@ -289,6 +292,10 @@ enum class _LIBCPP_ENUM_VIS file_type : signed char { unknown = 8 }; +// On Windows, these permission bits map to one single readonly flag per +// file, and the executable bit is always returned as set. When setting +// permissions, as long as the write bit is set for either owner, group or +// others, the readonly flag is cleared. enum class _LIBCPP_ENUM_VIS perms : unsigned { none = 0, @@ -551,7 +558,7 @@ struct __can_convert_char { static const bool value = true; using __char_type = wchar_t; }; -#ifndef _LIBCPP_NO_HAS_CHAR8_T +#ifndef _LIBCPP_HAS_NO_CHAR8_T template <> struct __can_convert_char { static const bool value = true; @@ -579,7 +586,7 @@ __is_separator(_ECharT __e) { #endif } -#ifndef _LIBCPP_NO_HAS_CHAR8_T +#ifndef _LIBCPP_HAS_NO_CHAR8_T typedef u8string __u8_string; #else typedef string __u8_string; @@ -785,7 +792,7 @@ struct _PathCVT<__path_value> { template static typename enable_if<__is_cpp17_forward_iterator<_Iter>::value>::type __append_range(__path_string& __dest, _Iter __b, _Iter __e) { - __dest.__append_forward_unsafe(__b, __e); + __dest.append(__b, __e); } template @@ -886,7 +893,7 @@ struct _PathExport { } }; -#ifndef _LIBCPP_NO_HAS_CHAR8_T +#ifndef _LIBCPP_HAS_NO_CHAR8_T template <> struct _PathExport { typedef __narrow_to_utf8 _Narrower; @@ -896,7 +903,7 @@ struct _PathExport { _Narrower()(back_inserter(__dest), __src.data(), __src.data() + __src.size()); } }; -#endif /* !_LIBCPP_NO_HAS_CHAR8_T */ +#endif /* !_LIBCPP_HAS_NO_CHAR8_T */ #endif /* _LIBCPP_WIN32API */ class _LIBCPP_TYPE_VIS path { @@ -921,7 +928,7 @@ public: typedef basic_string string_type; typedef basic_string_view __string_view; - enum class _LIBCPP_ENUM_VIS format : unsigned char { + enum _LIBCPP_ENUM_VIS format : unsigned char { auto_format, native_format, generic_format @@ -973,8 +980,8 @@ public: return *this; } - template - _LIBCPP_INLINE_VISIBILITY path& operator=(string_type&& __s) noexcept { + _LIBCPP_INLINE_VISIBILITY + path& operator=(string_type&& __s) noexcept { __pn_ = _VSTD::move(__s); return *this; } @@ -1006,14 +1013,44 @@ public: return *this; } -private: - template - static bool __source_is_absolute(_ECharT __first_or_null) { - return __is_separator(__first_or_null); - } - public: // appends +#if defined(_LIBCPP_WIN32API) + path& operator/=(const path& __p) { + auto __p_root_name = __p.__root_name(); + auto __p_root_name_size = __p_root_name.size(); + if (__p.is_absolute() || + (!__p_root_name.empty() && __p_root_name != root_name())) { + __pn_ = __p.__pn_; + return *this; + } + if (__p.has_root_directory()) { + path __root_name_str = root_name(); + __pn_ = __root_name_str.native(); + __pn_ += __string_view(__p.__pn_).substr(__p_root_name_size); + return *this; + } + if (has_filename() || (!has_root_directory() && is_absolute())) + __pn_ += preferred_separator; + __pn_ += __string_view(__p.__pn_).substr(__p_root_name_size); + return *this; + } + template + _LIBCPP_INLINE_VISIBILITY _EnableIfPathable<_Source> + operator/=(const _Source& __src) { + return operator/=(path(__src)); + } + + template + _EnableIfPathable<_Source> append(const _Source& __src) { + return operator/=(path(__src)); + } + + template + path& append(_InputIt __first, _InputIt __last) { + return operator/=(path(__first, __last)); + } +#else path& operator/=(const path& __p) { if (__p.is_absolute()) { __pn_ = __p.__pn_; @@ -1038,7 +1075,8 @@ public: _EnableIfPathable<_Source> append(const _Source& __src) { using _Traits = __is_pathable<_Source>; using _CVT = _PathCVT<_SourceChar<_Source> >; - if (__source_is_absolute(_Traits::__first_or_null(__src))) + bool __source_is_absolute = __is_separator(_Traits::__first_or_null(__src)); + if (__source_is_absolute) __pn_.clear(); else if (has_filename()) __pn_ += preferred_separator; @@ -1051,13 +1089,14 @@ public: typedef typename iterator_traits<_InputIt>::value_type _ItVal; static_assert(__can_convert_char<_ItVal>::value, "Must convertible"); using _CVT = _PathCVT<_ItVal>; - if (__first != __last && __source_is_absolute(*__first)) + if (__first != __last && __is_separator(*__first)) __pn_.clear(); else if (has_filename()) __pn_ += preferred_separator; _CVT::__append_range(__pn_, __first, __last); return *this; } +#endif // concatenation _LIBCPP_INLINE_VISIBILITY @@ -1161,7 +1200,12 @@ public: #if defined(_LIBCPP_WIN32API) _LIBCPP_INLINE_VISIBILITY _VSTD::wstring wstring() const { return __pn_; } - _VSTD::wstring generic_wstring() const { return __pn_; } + _VSTD::wstring generic_wstring() const { + _VSTD::wstring __s; + __s.resize(__pn_.size()); + _VSTD::replace_copy(__pn_.begin(), __pn_.end(), __s.begin(), '\\', '/'); + return __s; + } #if !defined(_LIBCPP_HAS_NO_LOCALIZATION) template , @@ -1198,18 +1242,29 @@ public: class _Allocator = allocator<_ECharT> > basic_string<_ECharT, _Traits, _Allocator> generic_string(const _Allocator& __a = _Allocator()) const { - return string<_ECharT, _Traits, _Allocator>(__a); + using _Str = basic_string<_ECharT, _Traits, _Allocator>; + _Str __s = string<_ECharT, _Traits, _Allocator>(__a); + // Note: This (and generic_u8string below) is slightly suboptimal as + // it iterates twice over the string; once to convert it to the right + // character type, and once to replace path delimiters. + _VSTD::replace(__s.begin(), __s.end(), + static_cast<_ECharT>('\\'), static_cast<_ECharT>('/')); + return __s; } _VSTD::string generic_string() const { return generic_string(); } _VSTD::u16string generic_u16string() const { return generic_string(); } _VSTD::u32string generic_u32string() const { return generic_string(); } - __u8_string generic_u8string() const { return u8string(); } + __u8_string generic_u8string() const { + __u8_string __s = u8string(); + _VSTD::replace(__s.begin(), __s.end(), '\\', '/'); + return __s; + } #endif /* !_LIBCPP_HAS_NO_LOCALIZATION */ #else /* _LIBCPP_WIN32API */ _LIBCPP_INLINE_VISIBILITY _VSTD::string string() const { return __pn_; } -#ifndef _LIBCPP_NO_HAS_CHAR8_T +#ifndef _LIBCPP_HAS_NO_CHAR8_T _LIBCPP_INLINE_VISIBILITY _VSTD::u8string u8string() const { return _VSTD::u8string(__pn_.begin(), __pn_.end()); } #else _LIBCPP_INLINE_VISIBILITY _VSTD::string u8string() const { return __pn_; } @@ -1241,7 +1296,7 @@ public: // generic format observers _VSTD::string generic_string() const { return __pn_; } -#ifndef _LIBCPP_NO_HAS_CHAR8_T +#ifndef _LIBCPP_HAS_NO_CHAR8_T _VSTD::u8string generic_u8string() const { return _VSTD::u8string(__pn_.begin(), __pn_.end()); } #else _VSTD::string generic_u8string() const { return __pn_; } @@ -1295,7 +1350,11 @@ public: return string_type(__root_directory()); } _LIBCPP_INLINE_VISIBILITY path root_path() const { +#if defined(_LIBCPP_WIN32API) + return string_type(__root_path_raw()); +#else return root_name().append(string_type(__root_directory())); +#endif } _LIBCPP_INLINE_VISIBILITY path relative_path() const { return string_type(__relative_path()); @@ -1341,7 +1400,28 @@ public: } _LIBCPP_INLINE_VISIBILITY bool is_absolute() const { +#if defined(_LIBCPP_WIN32API) + __string_view __root_name_str = __root_name(); + __string_view __root_dir = __root_directory(); + if (__root_name_str.size() == 2 && __root_name_str[1] == ':') { + // A drive letter with no root directory is relative, e.g. x:example. + return !__root_dir.empty(); + } + // If no root name, it's relative, e.g. \example is relative to the current drive + if (__root_name_str.empty()) + return false; + if (__root_name_str.size() < 3) + return false; + // A server root name, like \\server, is always absolute + if (__root_name_str[0] != '/' && __root_name_str[0] != '\\') + return false; + if (__root_name_str[1] != '/' && __root_name_str[1] != '\\') + return false; + // Seems to be a server root name + return true; +#else return has_root_directory(); +#endif } _LIBCPP_INLINE_VISIBILITY bool is_relative() const { return !is_absolute(); } @@ -1440,7 +1520,7 @@ _LIBCPP_INLINE_VISIBILITY _LIBCPP_DEPRECATED_WITH_CHAR8_T typename enable_if<__is_pathable<_InputIt>::value, path>::type u8path(_InputIt __f, _InputIt __l) { static_assert( -#ifndef _LIBCPP_NO_HAS_CHAR8_T +#ifndef _LIBCPP_HAS_NO_CHAR8_T is_same::__char_type, char8_t>::value || #endif is_same::__char_type, char>::value, @@ -1464,7 +1544,7 @@ _LIBCPP_INLINE_VISIBILITY _LIBCPP_DEPRECATED_WITH_CHAR8_T typename enable_if<__is_pathable<_InputIt>::value, path>::type u8path(_InputIt __f, _NullSentinel) { static_assert( -#ifndef _LIBCPP_NO_HAS_CHAR8_T +#ifndef _LIBCPP_HAS_NO_CHAR8_T is_same::__char_type, char8_t>::value || #endif is_same::__char_type, char>::value, @@ -1487,7 +1567,7 @@ _LIBCPP_INLINE_VISIBILITY _LIBCPP_DEPRECATED_WITH_CHAR8_T typename enable_if<__is_pathable<_Source>::value, path>::type u8path(const _Source& __s) { static_assert( -#ifndef _LIBCPP_NO_HAS_CHAR8_T +#ifndef _LIBCPP_HAS_NO_CHAR8_T is_same::__char_type, char8_t>::value || #endif is_same::__char_type, char>::value, @@ -1495,7 +1575,7 @@ _LIBCPP_INLINE_VISIBILITY _LIBCPP_DEPRECATED_WITH_CHAR8_T "'char' or 'char8_t'"); #if defined(_LIBCPP_WIN32API) using _Traits = __is_pathable<_Source>; - return u8path(__unwrap_iter(_Traits::__range_begin(__s)), __unwrap_iter(_Traits::__range_end(__s))); + return u8path(_VSTD::__unwrap_iter(_Traits::__range_begin(__s)), _VSTD::__unwrap_iter(_Traits::__range_end(__s))); #else return path(__s); #endif diff --git a/lib/libcxx/include/float.h b/lib/libcxx/include/float.h index 5c1e1db79..399ab4f11 100644 --- a/lib/libcxx/include/float.h +++ b/lib/libcxx/include/float.h @@ -90,4 +90,4 @@ Macros: #endif // __cplusplus -#endif // _LIBCPP_FLOAT_H +#endif // _LIBCPP_FLOAT_H diff --git a/lib/libcxx/include/format b/lib/libcxx/include/format new file mode 100644 index 000000000..cfd851aa9 --- /dev/null +++ b/lib/libcxx/include/format @@ -0,0 +1,86 @@ +// -*- C++ -*- +//===--------------------------- format -----------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP_FORMAT +#define _LIBCPP_FORMAT + +/* + +namespace std { + // [format.error], class format_error + class format_error : public runtime_error { + public: + explicit format_error(const string& what_arg); + explicit format_error(const char* what_arg); + }; + + // [format.parse.ctx], class template basic_format_parse_context + template + class basic_format_parse_context { + public: + using char_type = charT; + using const_iterator = typename basic_string_view::const_iterator; + using iterator = const_iterator; + + private: + iterator begin_; // exposition only + iterator end_; // exposition only + enum indexing { unknown, manual, automatic }; // exposition only + indexing indexing_; // exposition only + size_t next_arg_id_; // exposition only + size_t num_args_; // exposition only + + public: + constexpr explicit basic_format_parse_context(basic_string_view fmt, + size_t num_args = 0) noexcept; + basic_format_parse_context(const basic_format_parse_context&) = delete; + basic_format_parse_context& operator=(const basic_format_parse_context&) = delete; + + constexpr const_iterator begin() const noexcept; + constexpr const_iterator end() const noexcept; + constexpr void advance_to(const_iterator it); + + constexpr size_t next_arg_id(); + constexpr void check_arg_id(size_t id); + }; + using format_parse_context = basic_format_parse_context; + using wformat_parse_context = basic_format_parse_context; +} + +*/ + +// Make sure all feature tests macros are always available. +#include +// Only enable the contents of the header when libc++ was build with LIBCXX_ENABLE_INCOMPLETE_FEATURES enabled +#if !defined(_LIBCPP_HAS_NO_INCOMPLETE_FORMAT) + +#include <__config> +#include <__format/format_error.h> +#include <__format/format_parse_context.h> + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +# pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +#if _LIBCPP_STD_VER > 17 + +#endif //_LIBCPP_STD_VER > 17 + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // !defined(_LIBCPP_HAS_NO_INCOMPLETE_FORMAT) + +#endif // _LIBCPP_FORMAT diff --git a/lib/libcxx/include/forward_list b/lib/libcxx/include/forward_list index d3d6b8238..0ae8f19ec 100644 --- a/lib/libcxx/include/forward_list +++ b/lib/libcxx/include/forward_list @@ -180,11 +180,12 @@ template */ #include <__config> +#include <__utility/forward.h> +#include #include -#include -#include #include -#include +#include +#include #include #if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) @@ -272,7 +273,7 @@ struct _LIBCPP_HIDDEN __begin_node_of }; template -struct __forward_list_node +struct _LIBCPP_STANDALONE_DEBUG __forward_list_node : public __begin_node_of<_Tp, _VoidPtr>::type { typedef _Tp value_type; @@ -506,7 +507,7 @@ public: _NOEXCEPT_(is_nothrow_move_constructible<__node_allocator>::value); _LIBCPP_INLINE_VISIBILITY __forward_list_base(__forward_list_base&& __x, const allocator_type& __a); -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG private: __forward_list_base(const __forward_list_base&); @@ -534,7 +535,7 @@ public: #if _LIBCPP_STD_VER >= 14 _NOEXCEPT; #else - _NOEXCEPT_(!__node_traits::propagate_on_container_move_assignment::value || + _NOEXCEPT_(!__node_traits::propagate_on_container_swap::value || __is_nothrow_swappable<__node_allocator>::value); #endif protected: @@ -584,7 +585,7 @@ __forward_list_base<_Tp, _Alloc>::__forward_list_base(__forward_list_base&& __x, } } -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG template __forward_list_base<_Tp, _Alloc>::~__forward_list_base() @@ -599,7 +600,7 @@ __forward_list_base<_Tp, _Alloc>::swap(__forward_list_base& __x) #if _LIBCPP_STD_VER >= 14 _NOEXCEPT #else - _NOEXCEPT_(!__node_traits::propagate_on_container_move_assignment::value || + _NOEXCEPT_(!__node_traits::propagate_on_container_swap::value || __is_nothrow_swappable<__node_allocator>::value) #endif { @@ -681,7 +682,7 @@ public: __is_cpp17_input_iterator<_InputIterator>::value >::type* = nullptr); forward_list(const forward_list& __x); - forward_list(const forward_list& __x, const allocator_type& __a); + forward_list(const forward_list& __x, const __identity_t& __a); forward_list& operator=(const forward_list& __x); @@ -690,7 +691,7 @@ public: forward_list(forward_list&& __x) _NOEXCEPT_(is_nothrow_move_constructible::value) : base(_VSTD::move(__x)) {} - forward_list(forward_list&& __x, const allocator_type& __a); + forward_list(forward_list&& __x, const __identity_t& __a); forward_list(initializer_list __il); forward_list(initializer_list __il, const allocator_type& __a); @@ -706,7 +707,7 @@ public: _LIBCPP_INLINE_VISIBILITY void assign(initializer_list __il); -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG // ~forward_list() = default; @@ -775,7 +776,7 @@ public: template void emplace_front(_Args&&... __args); #endif void push_front(value_type&& __v); -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG void push_front(const value_type& __v); void pop_front(); @@ -787,7 +788,7 @@ public: iterator insert_after(const_iterator __p, value_type&& __v); iterator insert_after(const_iterator __p, initializer_list __il) {return insert_after(__p, __il.begin(), __il.end());} -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG iterator insert_after(const_iterator __p, const value_type& __v); iterator insert_after(const_iterator __p, size_type __n, const value_type& __v); template @@ -840,7 +841,7 @@ public: _LIBCPP_INLINE_VISIBILITY void merge(forward_list&& __x, _Compare __comp) {merge(__x, _VSTD::move(__comp));} -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG _LIBCPP_INLINE_VISIBILITY void merge(forward_list& __x) {merge(__x, __less());} template void merge(forward_list& __x, _Compare __comp); @@ -855,7 +856,7 @@ private: void __move_assign(forward_list& __x, true_type) _NOEXCEPT_(is_nothrow_move_assignable::value); void __move_assign(forward_list& __x, false_type); -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG template static @@ -871,18 +872,18 @@ private: #ifndef _LIBCPP_HAS_NO_DEDUCTION_GUIDES template::value_type>, - class = typename enable_if<__is_allocator<_Alloc>::value, void>::type + class _Alloc = allocator<__iter_value_type<_InputIterator>>, + class = _EnableIf<__is_allocator<_Alloc>::value> > forward_list(_InputIterator, _InputIterator) - -> forward_list::value_type, _Alloc>; + -> forward_list<__iter_value_type<_InputIterator>, _Alloc>; template::value, void>::type + class = _EnableIf<__is_allocator<_Alloc>::value> > forward_list(_InputIterator, _InputIterator, _Alloc) - -> forward_list::value_type, _Alloc>; + -> forward_list<__iter_value_type<_InputIterator>, _Alloc>; #endif template @@ -979,7 +980,7 @@ forward_list<_Tp, _Alloc>::forward_list(const forward_list& __x) template forward_list<_Tp, _Alloc>::forward_list(const forward_list& __x, - const allocator_type& __a) + const __identity_t& __a) : base(__a) { insert_after(cbefore_begin(), __x.begin(), __x.end()); @@ -1000,7 +1001,7 @@ forward_list<_Tp, _Alloc>::operator=(const forward_list& __x) #ifndef _LIBCPP_CXX03_LANG template forward_list<_Tp, _Alloc>::forward_list(forward_list&& __x, - const allocator_type& __a) + const __identity_t& __a) : base(_VSTD::move(__x), __a) { if (base::__alloc() != __x.__alloc()) @@ -1070,7 +1071,7 @@ forward_list<_Tp, _Alloc>::operator=(initializer_list __il) return *this; } -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG template template @@ -1150,7 +1151,7 @@ forward_list<_Tp, _Alloc>::push_front(value_type&& __v) base::__before_begin()->__next_ = __h.release(); } -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG template void @@ -1207,7 +1208,7 @@ forward_list<_Tp, _Alloc>::insert_after(const_iterator __p, value_type&& __v) return iterator(__r->__next_); } -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG template typename forward_list<_Tp, _Alloc>::iterator @@ -1240,7 +1241,7 @@ forward_list<_Tp, _Alloc>::insert_after(const_iterator __p, size_type __n, #ifndef _LIBCPP_NO_EXCEPTIONS try { -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS for (--__n; __n != 0; --__n, __last = __last->__next_) { __h.reset(__node_traits::allocate(__a, 1)); @@ -1260,7 +1261,7 @@ forward_list<_Tp, _Alloc>::insert_after(const_iterator __p, size_type __n, } throw; } -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS __last->__next_ = __r->__next_; __r->__next_ = __first; __r = static_cast<__begin_node_pointer>(__last); @@ -1290,7 +1291,7 @@ forward_list<_Tp, _Alloc>::insert_after(const_iterator __p, #ifndef _LIBCPP_NO_EXCEPTIONS try { -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS for (++__f; __f != __l; ++__f, ((void)(__last = __last->__next_))) { __h.reset(__node_traits::allocate(__a, 1)); @@ -1310,7 +1311,7 @@ forward_list<_Tp, _Alloc>::insert_after(const_iterator __p, } throw; } -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS __last->__next_ = __r->__next_; __r->__next_ = __first; __r = static_cast<__begin_node_pointer>(__last); @@ -1784,4 +1785,4 @@ _LIBCPP_END_NAMESPACE_STD _LIBCPP_POP_MACROS -#endif // _LIBCPP_FORWARD_LIST +#endif // _LIBCPP_FORWARD_LIST diff --git a/lib/libcxx/include/fstream b/lib/libcxx/include/fstream index d7d6b46c3..c522b8ab1 100644 --- a/lib/libcxx/include/fstream +++ b/lib/libcxx/include/fstream @@ -179,13 +179,14 @@ typedef basic_fstream wfstream; */ -#include <__config> #include <__availability> -#include -#include +#include <__config> +#include <__debug> #include <__locale> #include #include +#include +#include #if !defined(_LIBCPP_HAS_NO_FILESYSTEM_LIBRARY) # include @@ -198,6 +199,9 @@ typedef basic_fstream wfstream; _LIBCPP_PUSH_MACROS #include <__undef_macros> +#if defined(_LIBCPP_MSVCRT) || defined(_NEWLIB_VERSION) +# define _LIBCPP_HAS_NO_OFF_T_FUNCTIONS +#endif _LIBCPP_BEGIN_NAMESPACE_STD @@ -215,16 +219,12 @@ public: // 27.9.1.2 Constructors/destructor: basic_filebuf(); -#ifndef _LIBCPP_CXX03_LANG basic_filebuf(basic_filebuf&& __rhs); -#endif virtual ~basic_filebuf(); // 27.9.1.3 Assign/swap: -#ifndef _LIBCPP_CXX03_LANG _LIBCPP_INLINE_VISIBILITY basic_filebuf& operator=(basic_filebuf&& __rhs); -#endif void swap(basic_filebuf& __rhs); // 27.9.1.4 Members: @@ -244,7 +244,7 @@ public: return open(__p.c_str(), __mode); } #endif - inline _LIBCPP_INLINE_VISIBILITY + _LIBCPP_INLINE_VISIBILITY basic_filebuf* __open(int __fd, ios_base::openmode __mode); #endif basic_filebuf* close(); @@ -314,8 +314,6 @@ basic_filebuf<_CharT, _Traits>::basic_filebuf() setbuf(nullptr, 4096); } -#ifndef _LIBCPP_CXX03_LANG - template basic_filebuf<_CharT, _Traits>::basic_filebuf(basic_filebuf&& __rhs) : basic_streambuf<_CharT, _Traits>(__rhs) @@ -390,22 +388,20 @@ basic_filebuf<_CharT, _Traits>::operator=(basic_filebuf&& __rhs) return *this; } -#endif // _LIBCPP_CXX03_LANG - template basic_filebuf<_CharT, _Traits>::~basic_filebuf() { #ifndef _LIBCPP_NO_EXCEPTIONS try { -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS close(); #ifndef _LIBCPP_NO_EXCEPTIONS } catch (...) { } -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS if (__owns_eb_) delete [] __extbuf_; if (__owns_ib_) @@ -574,7 +570,7 @@ basic_filebuf<_CharT, _Traits>::open(const char* __s, ios_base::openmode __mode) } template -inline _LIBCPP_INLINE_VISIBILITY +inline basic_filebuf<_CharT, _Traits>* basic_filebuf<_CharT, _Traits>::__open(int __fd, ios_base::openmode __mode) { basic_filebuf<_CharT, _Traits>* __rt = nullptr; @@ -1160,13 +1156,10 @@ public: : basic_ifstream(__p.c_str(), __mode) {} #endif // _LIBCPP_STD_VER >= 17 #endif -#ifndef _LIBCPP_CXX03_LANG _LIBCPP_INLINE_VISIBILITY basic_ifstream(basic_ifstream&& __rhs); - _LIBCPP_INLINE_VISIBILITY basic_ifstream& operator=(basic_ifstream&& __rhs); -#endif _LIBCPP_INLINE_VISIBILITY void swap(basic_ifstream& __rhs); @@ -1236,8 +1229,6 @@ basic_ifstream<_CharT, _Traits>::basic_ifstream(const string& __s, ios_base::ope } #endif -#ifndef _LIBCPP_CXX03_LANG - template inline basic_ifstream<_CharT, _Traits>::basic_ifstream(basic_ifstream&& __rhs) @@ -1257,8 +1248,6 @@ basic_ifstream<_CharT, _Traits>::operator=(basic_ifstream&& __rhs) return *this; } -#endif // _LIBCPP_CXX03_LANG - template inline void @@ -1326,6 +1315,7 @@ basic_ifstream<_CharT, _Traits>::open(const string& __s, ios_base::openmode __mo } template +inline void basic_ifstream<_CharT, _Traits>::__open(int __fd, ios_base::openmode __mode) { if (__sb_.__open(__fd, __mode | ios_base::in)) @@ -1374,13 +1364,10 @@ public: : basic_ofstream(__p.c_str(), __mode) {} #endif // _LIBCPP_STD_VER >= 17 -#ifndef _LIBCPP_CXX03_LANG _LIBCPP_INLINE_VISIBILITY basic_ofstream(basic_ofstream&& __rhs); - _LIBCPP_INLINE_VISIBILITY basic_ofstream& operator=(basic_ofstream&& __rhs); -#endif _LIBCPP_INLINE_VISIBILITY void swap(basic_ofstream& __rhs); @@ -1449,8 +1436,6 @@ basic_ofstream<_CharT, _Traits>::basic_ofstream(const string& __s, ios_base::ope } #endif -#ifndef _LIBCPP_CXX03_LANG - template inline basic_ofstream<_CharT, _Traits>::basic_ofstream(basic_ofstream&& __rhs) @@ -1470,8 +1455,6 @@ basic_ofstream<_CharT, _Traits>::operator=(basic_ofstream&& __rhs) return *this; } -#endif // _LIBCPP_CXX03_LANG - template inline void @@ -1539,6 +1522,7 @@ basic_ofstream<_CharT, _Traits>::open(const string& __s, ios_base::openmode __mo } template +inline void basic_ofstream<_CharT, _Traits>::__open(int __fd, ios_base::openmode __mode) { if (__sb_.__open(__fd, __mode | ios_base::out)) @@ -1589,13 +1573,12 @@ public: #endif // _LIBCPP_STD_VER >= 17 #endif -#ifndef _LIBCPP_CXX03_LANG _LIBCPP_INLINE_VISIBILITY basic_fstream(basic_fstream&& __rhs); _LIBCPP_INLINE_VISIBILITY basic_fstream& operator=(basic_fstream&& __rhs); -#endif + _LIBCPP_INLINE_VISIBILITY void swap(basic_fstream& __rhs); @@ -1662,8 +1645,6 @@ basic_fstream<_CharT, _Traits>::basic_fstream(const string& __s, ios_base::openm } #endif -#ifndef _LIBCPP_CXX03_LANG - template inline basic_fstream<_CharT, _Traits>::basic_fstream(basic_fstream&& __rhs) @@ -1683,8 +1664,6 @@ basic_fstream<_CharT, _Traits>::operator=(basic_fstream&& __rhs) return *this; } -#endif // _LIBCPP_CXX03_LANG - template inline void @@ -1771,4 +1750,4 @@ _LIBCPP_END_NAMESPACE_STD _LIBCPP_POP_MACROS -#endif // _LIBCPP_FSTREAM +#endif // _LIBCPP_FSTREAM diff --git a/lib/libcxx/include/functional b/lib/libcxx/include/functional index 67baa5bd4..ecbc5667a 100644 --- a/lib/libcxx/include/functional +++ b/lib/libcxx/include/functional @@ -1,5 +1,5 @@ // -*- C++ -*- -//===------------------------ functional ----------------------------------===// +//===----------------------------------------------------------------------===// // // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. // See https://llvm.org/LICENSE.txt for license information. @@ -42,8 +42,8 @@ public: typedef see below result_type; // Not always defined // construct/copy/destroy - reference_wrapper(T&) noexcept; - reference_wrapper(T&&) = delete; // do not bind to temps + template + reference_wrapper(U&&); reference_wrapper(const reference_wrapper& x) noexcept; // assignment @@ -59,6 +59,9 @@ public: operator() (ArgTypes&&...) const; }; +template + reference_wrapper(T&) -> reference_wrapper; + template reference_wrapper ref(T& t) noexcept; template void ref(const T&& t) = delete; template reference_wrapper ref(reference_wrappert) noexcept; @@ -73,121 +76,104 @@ template using unwrap_reference_t = typename unwrap_reference::type template using unwrap_ref_decay_t = typename unwrap_ref_decay::type; // since C++20 template // in C++14 -struct plus : binary_function -{ +struct plus { T operator()(const T& x, const T& y) const; }; template // in C++14 -struct minus : binary_function -{ +struct minus { T operator()(const T& x, const T& y) const; }; template // in C++14 -struct multiplies : binary_function -{ +struct multiplies { T operator()(const T& x, const T& y) const; }; template // in C++14 -struct divides : binary_function -{ +struct divides { T operator()(const T& x, const T& y) const; }; template // in C++14 -struct modulus : binary_function -{ +struct modulus { T operator()(const T& x, const T& y) const; }; template // in C++14 -struct negate : unary_function -{ +struct negate { T operator()(const T& x) const; }; template // in C++14 -struct equal_to : binary_function -{ +struct equal_to { bool operator()(const T& x, const T& y) const; }; template // in C++14 -struct not_equal_to : binary_function -{ +struct not_equal_to { bool operator()(const T& x, const T& y) const; }; template // in C++14 -struct greater : binary_function -{ +struct greater { bool operator()(const T& x, const T& y) const; }; template // in C++14 -struct less : binary_function -{ +struct less { bool operator()(const T& x, const T& y) const; }; template // in C++14 -struct greater_equal : binary_function -{ +struct greater_equal { bool operator()(const T& x, const T& y) const; }; template // in C++14 -struct less_equal : binary_function -{ +struct less_equal { bool operator()(const T& x, const T& y) const; }; template // in C++14 -struct logical_and : binary_function -{ +struct logical_and { bool operator()(const T& x, const T& y) const; }; template // in C++14 -struct logical_or : binary_function -{ +struct logical_or { bool operator()(const T& x, const T& y) const; }; template // in C++14 -struct logical_not : unary_function -{ +struct logical_not { bool operator()(const T& x) const; }; template // in C++14 -struct bit_and : unary_function -{ - bool operator()(const T& x, const T& y) const; +struct bit_and { + T operator()(const T& x, const T& y) const; }; template // in C++14 -struct bit_or : unary_function -{ - bool operator()(const T& x, const T& y) const; +struct bit_or { + T operator()(const T& x, const T& y) const; }; template // in C++14 -struct bit_xor : unary_function -{ - bool operator()(const T& x, const T& y) const; +struct bit_xor { + T operator()(const T& x, const T& y) const; }; template // C++14 -struct bit_xor : unary_function -{ - bool operator()(const T& x) const; +struct bit_not { + T operator()(const T& x) const; }; +struct identity; // C++20 + template -class unary_negate // deprecated in C++17 +class unary_negate // deprecated in C++17, removed in C++20 : public unary_function { public: @@ -195,11 +181,11 @@ public: bool operator()(const typename Predicate::argument_type& x) const; }; -template // deprecated in C++17 +template // deprecated in C++17, removed in C++20 unary_negate not1(const Predicate& pred); template -class binary_negate // deprecated in C++17 +class binary_negate // deprecated in C++17, removed in C++20 : public binary_function @@ -210,7 +196,7 @@ public: const typename Predicate::second_argument_type& y) const; }; -template // deprecated in C++17 +template // deprecated in C++17, removed in C++20 binary_negate not2(const Predicate& pred); template @@ -501,2687 +487,43 @@ POLICY: For non-variadic implementations, the number of arguments is limited */ +#include <__algorithm/search.h> #include <__config> -#include -#include +#include <__debug> +#include <__functional/binary_function.h> // TODO: deprecate +#include <__functional/binary_negate.h> +#include <__functional/bind_front.h> +#include <__functional/bind.h> +#include <__functional/binder1st.h> +#include <__functional/binder2nd.h> +#include <__functional/default_searcher.h> +#include <__functional/function.h> +#include <__functional/hash.h> +#include <__functional/identity.h> +#include <__functional/invoke.h> +#include <__functional/mem_fn.h> // TODO: deprecate +#include <__functional/mem_fun_ref.h> +#include <__functional/not_fn.h> +#include <__functional/operations.h> +#include <__functional/pointer_to_binary_function.h> +#include <__functional/pointer_to_unary_function.h> +#include <__functional/ranges_operations.h> +#include <__functional/reference_wrapper.h> +#include <__functional/unary_function.h> // TODO: deprecate +#include <__functional/unary_negate.h> +#include <__functional/unwrap_ref.h> +#include <__utility/forward.h> +#include #include #include #include +#include +#include #include #include -#include <__functional_base> - #if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) #pragma GCC system_header #endif -_LIBCPP_BEGIN_NAMESPACE_STD - -#if _LIBCPP_STD_VER > 11 -template -#else -template -#endif -struct _LIBCPP_TEMPLATE_VIS plus : binary_function<_Tp, _Tp, _Tp> -{ - _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY - _Tp operator()(const _Tp& __x, const _Tp& __y) const - {return __x + __y;} -}; - -#if _LIBCPP_STD_VER > 11 -template <> -struct _LIBCPP_TEMPLATE_VIS plus -{ - template - _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY - auto operator()(_T1&& __t, _T2&& __u) const - _NOEXCEPT_(noexcept(_VSTD::forward<_T1>(__t) + _VSTD::forward<_T2>(__u))) - -> decltype (_VSTD::forward<_T1>(__t) + _VSTD::forward<_T2>(__u)) - { return _VSTD::forward<_T1>(__t) + _VSTD::forward<_T2>(__u); } - typedef void is_transparent; -}; -#endif - - -#if _LIBCPP_STD_VER > 11 -template -#else -template -#endif -struct _LIBCPP_TEMPLATE_VIS minus : binary_function<_Tp, _Tp, _Tp> -{ - _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY - _Tp operator()(const _Tp& __x, const _Tp& __y) const - {return __x - __y;} -}; - -#if _LIBCPP_STD_VER > 11 -template <> -struct _LIBCPP_TEMPLATE_VIS minus -{ - template - _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY - auto operator()(_T1&& __t, _T2&& __u) const - _NOEXCEPT_(noexcept(_VSTD::forward<_T1>(__t) - _VSTD::forward<_T2>(__u))) - -> decltype (_VSTD::forward<_T1>(__t) - _VSTD::forward<_T2>(__u)) - { return _VSTD::forward<_T1>(__t) - _VSTD::forward<_T2>(__u); } - typedef void is_transparent; -}; -#endif - - -#if _LIBCPP_STD_VER > 11 -template -#else -template -#endif -struct _LIBCPP_TEMPLATE_VIS multiplies : binary_function<_Tp, _Tp, _Tp> -{ - _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY - _Tp operator()(const _Tp& __x, const _Tp& __y) const - {return __x * __y;} -}; - -#if _LIBCPP_STD_VER > 11 -template <> -struct _LIBCPP_TEMPLATE_VIS multiplies -{ - template - _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY - auto operator()(_T1&& __t, _T2&& __u) const - _NOEXCEPT_(noexcept(_VSTD::forward<_T1>(__t) * _VSTD::forward<_T2>(__u))) - -> decltype (_VSTD::forward<_T1>(__t) * _VSTD::forward<_T2>(__u)) - { return _VSTD::forward<_T1>(__t) * _VSTD::forward<_T2>(__u); } - typedef void is_transparent; -}; -#endif - - -#if _LIBCPP_STD_VER > 11 -template -#else -template -#endif -struct _LIBCPP_TEMPLATE_VIS divides : binary_function<_Tp, _Tp, _Tp> -{ - _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY - _Tp operator()(const _Tp& __x, const _Tp& __y) const - {return __x / __y;} -}; - -#if _LIBCPP_STD_VER > 11 -template <> -struct _LIBCPP_TEMPLATE_VIS divides -{ - template - _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY - auto operator()(_T1&& __t, _T2&& __u) const - _NOEXCEPT_(noexcept(_VSTD::forward<_T1>(__t) / _VSTD::forward<_T2>(__u))) - -> decltype (_VSTD::forward<_T1>(__t) / _VSTD::forward<_T2>(__u)) - { return _VSTD::forward<_T1>(__t) / _VSTD::forward<_T2>(__u); } - typedef void is_transparent; -}; -#endif - - -#if _LIBCPP_STD_VER > 11 -template -#else -template -#endif -struct _LIBCPP_TEMPLATE_VIS modulus : binary_function<_Tp, _Tp, _Tp> -{ - _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY - _Tp operator()(const _Tp& __x, const _Tp& __y) const - {return __x % __y;} -}; - -#if _LIBCPP_STD_VER > 11 -template <> -struct _LIBCPP_TEMPLATE_VIS modulus -{ - template - _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY - auto operator()(_T1&& __t, _T2&& __u) const - _NOEXCEPT_(noexcept(_VSTD::forward<_T1>(__t) % _VSTD::forward<_T2>(__u))) - -> decltype (_VSTD::forward<_T1>(__t) % _VSTD::forward<_T2>(__u)) - { return _VSTD::forward<_T1>(__t) % _VSTD::forward<_T2>(__u); } - typedef void is_transparent; -}; -#endif - - -#if _LIBCPP_STD_VER > 11 -template -#else -template -#endif -struct _LIBCPP_TEMPLATE_VIS negate : unary_function<_Tp, _Tp> -{ - _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY - _Tp operator()(const _Tp& __x) const - {return -__x;} -}; - -#if _LIBCPP_STD_VER > 11 -template <> -struct _LIBCPP_TEMPLATE_VIS negate -{ - template - _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY - auto operator()(_Tp&& __x) const - _NOEXCEPT_(noexcept(- _VSTD::forward<_Tp>(__x))) - -> decltype (- _VSTD::forward<_Tp>(__x)) - { return - _VSTD::forward<_Tp>(__x); } - typedef void is_transparent; -}; -#endif - - -#if _LIBCPP_STD_VER > 11 -template -#else -template -#endif -struct _LIBCPP_TEMPLATE_VIS equal_to : binary_function<_Tp, _Tp, bool> -{ - _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY - bool operator()(const _Tp& __x, const _Tp& __y) const - {return __x == __y;} -}; - -#if _LIBCPP_STD_VER > 11 -template <> -struct _LIBCPP_TEMPLATE_VIS equal_to -{ - template - _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY - auto operator()(_T1&& __t, _T2&& __u) const - _NOEXCEPT_(noexcept(_VSTD::forward<_T1>(__t) == _VSTD::forward<_T2>(__u))) - -> decltype (_VSTD::forward<_T1>(__t) == _VSTD::forward<_T2>(__u)) - { return _VSTD::forward<_T1>(__t) == _VSTD::forward<_T2>(__u); } - typedef void is_transparent; -}; -#endif - - -#if _LIBCPP_STD_VER > 11 -template -#else -template -#endif -struct _LIBCPP_TEMPLATE_VIS not_equal_to : binary_function<_Tp, _Tp, bool> -{ - _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY - bool operator()(const _Tp& __x, const _Tp& __y) const - {return __x != __y;} -}; - -#if _LIBCPP_STD_VER > 11 -template <> -struct _LIBCPP_TEMPLATE_VIS not_equal_to -{ - template - _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY - auto operator()(_T1&& __t, _T2&& __u) const - _NOEXCEPT_(noexcept(_VSTD::forward<_T1>(__t) != _VSTD::forward<_T2>(__u))) - -> decltype (_VSTD::forward<_T1>(__t) != _VSTD::forward<_T2>(__u)) - { return _VSTD::forward<_T1>(__t) != _VSTD::forward<_T2>(__u); } - typedef void is_transparent; -}; -#endif - - -#if _LIBCPP_STD_VER > 11 -template -#else -template -#endif -struct _LIBCPP_TEMPLATE_VIS greater : binary_function<_Tp, _Tp, bool> -{ - _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY - bool operator()(const _Tp& __x, const _Tp& __y) const - {return __x > __y;} -}; - -#if _LIBCPP_STD_VER > 11 -template <> -struct _LIBCPP_TEMPLATE_VIS greater -{ - template - _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY - auto operator()(_T1&& __t, _T2&& __u) const - _NOEXCEPT_(noexcept(_VSTD::forward<_T1>(__t) > _VSTD::forward<_T2>(__u))) - -> decltype (_VSTD::forward<_T1>(__t) > _VSTD::forward<_T2>(__u)) - { return _VSTD::forward<_T1>(__t) > _VSTD::forward<_T2>(__u); } - typedef void is_transparent; -}; -#endif - - -// less in <__functional_base> - -#if _LIBCPP_STD_VER > 11 -template -#else -template -#endif -struct _LIBCPP_TEMPLATE_VIS greater_equal : binary_function<_Tp, _Tp, bool> -{ - _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY - bool operator()(const _Tp& __x, const _Tp& __y) const - {return __x >= __y;} -}; - -#if _LIBCPP_STD_VER > 11 -template <> -struct _LIBCPP_TEMPLATE_VIS greater_equal -{ - template - _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY - auto operator()(_T1&& __t, _T2&& __u) const - _NOEXCEPT_(noexcept(_VSTD::forward<_T1>(__t) >= _VSTD::forward<_T2>(__u))) - -> decltype (_VSTD::forward<_T1>(__t) >= _VSTD::forward<_T2>(__u)) - { return _VSTD::forward<_T1>(__t) >= _VSTD::forward<_T2>(__u); } - typedef void is_transparent; -}; -#endif - - -#if _LIBCPP_STD_VER > 11 -template -#else -template -#endif -struct _LIBCPP_TEMPLATE_VIS less_equal : binary_function<_Tp, _Tp, bool> -{ - _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY - bool operator()(const _Tp& __x, const _Tp& __y) const - {return __x <= __y;} -}; - -#if _LIBCPP_STD_VER > 11 -template <> -struct _LIBCPP_TEMPLATE_VIS less_equal -{ - template - _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY - auto operator()(_T1&& __t, _T2&& __u) const - _NOEXCEPT_(noexcept(_VSTD::forward<_T1>(__t) <= _VSTD::forward<_T2>(__u))) - -> decltype (_VSTD::forward<_T1>(__t) <= _VSTD::forward<_T2>(__u)) - { return _VSTD::forward<_T1>(__t) <= _VSTD::forward<_T2>(__u); } - typedef void is_transparent; -}; -#endif - - -#if _LIBCPP_STD_VER > 11 -template -#else -template -#endif -struct _LIBCPP_TEMPLATE_VIS logical_and : binary_function<_Tp, _Tp, bool> -{ - _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY - bool operator()(const _Tp& __x, const _Tp& __y) const - {return __x && __y;} -}; - -#if _LIBCPP_STD_VER > 11 -template <> -struct _LIBCPP_TEMPLATE_VIS logical_and -{ - template - _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY - auto operator()(_T1&& __t, _T2&& __u) const - _NOEXCEPT_(noexcept(_VSTD::forward<_T1>(__t) && _VSTD::forward<_T2>(__u))) - -> decltype (_VSTD::forward<_T1>(__t) && _VSTD::forward<_T2>(__u)) - { return _VSTD::forward<_T1>(__t) && _VSTD::forward<_T2>(__u); } - typedef void is_transparent; -}; -#endif - - -#if _LIBCPP_STD_VER > 11 -template -#else -template -#endif -struct _LIBCPP_TEMPLATE_VIS logical_or : binary_function<_Tp, _Tp, bool> -{ - _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY - bool operator()(const _Tp& __x, const _Tp& __y) const - {return __x || __y;} -}; - -#if _LIBCPP_STD_VER > 11 -template <> -struct _LIBCPP_TEMPLATE_VIS logical_or -{ - template - _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY - auto operator()(_T1&& __t, _T2&& __u) const - _NOEXCEPT_(noexcept(_VSTD::forward<_T1>(__t) || _VSTD::forward<_T2>(__u))) - -> decltype (_VSTD::forward<_T1>(__t) || _VSTD::forward<_T2>(__u)) - { return _VSTD::forward<_T1>(__t) || _VSTD::forward<_T2>(__u); } - typedef void is_transparent; -}; -#endif - - -#if _LIBCPP_STD_VER > 11 -template -#else -template -#endif -struct _LIBCPP_TEMPLATE_VIS logical_not : unary_function<_Tp, bool> -{ - _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY - bool operator()(const _Tp& __x) const - {return !__x;} -}; - -#if _LIBCPP_STD_VER > 11 -template <> -struct _LIBCPP_TEMPLATE_VIS logical_not -{ - template - _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY - auto operator()(_Tp&& __x) const - _NOEXCEPT_(noexcept(!_VSTD::forward<_Tp>(__x))) - -> decltype (!_VSTD::forward<_Tp>(__x)) - { return !_VSTD::forward<_Tp>(__x); } - typedef void is_transparent; -}; -#endif - - -#if _LIBCPP_STD_VER > 11 -template -#else -template -#endif -struct _LIBCPP_TEMPLATE_VIS bit_and : binary_function<_Tp, _Tp, _Tp> -{ - _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY - _Tp operator()(const _Tp& __x, const _Tp& __y) const - {return __x & __y;} -}; - -#if _LIBCPP_STD_VER > 11 -template <> -struct _LIBCPP_TEMPLATE_VIS bit_and -{ - template - _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY - auto operator()(_T1&& __t, _T2&& __u) const - _NOEXCEPT_(noexcept(_VSTD::forward<_T1>(__t) & _VSTD::forward<_T2>(__u))) - -> decltype (_VSTD::forward<_T1>(__t) & _VSTD::forward<_T2>(__u)) - { return _VSTD::forward<_T1>(__t) & _VSTD::forward<_T2>(__u); } - typedef void is_transparent; -}; -#endif - - -#if _LIBCPP_STD_VER > 11 -template -#else -template -#endif -struct _LIBCPP_TEMPLATE_VIS bit_or : binary_function<_Tp, _Tp, _Tp> -{ - _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY - _Tp operator()(const _Tp& __x, const _Tp& __y) const - {return __x | __y;} -}; - -#if _LIBCPP_STD_VER > 11 -template <> -struct _LIBCPP_TEMPLATE_VIS bit_or -{ - template - _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY - auto operator()(_T1&& __t, _T2&& __u) const - _NOEXCEPT_(noexcept(_VSTD::forward<_T1>(__t) | _VSTD::forward<_T2>(__u))) - -> decltype (_VSTD::forward<_T1>(__t) | _VSTD::forward<_T2>(__u)) - { return _VSTD::forward<_T1>(__t) | _VSTD::forward<_T2>(__u); } - typedef void is_transparent; -}; -#endif - - -#if _LIBCPP_STD_VER > 11 -template -#else -template -#endif -struct _LIBCPP_TEMPLATE_VIS bit_xor : binary_function<_Tp, _Tp, _Tp> -{ - _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY - _Tp operator()(const _Tp& __x, const _Tp& __y) const - {return __x ^ __y;} -}; - -#if _LIBCPP_STD_VER > 11 -template <> -struct _LIBCPP_TEMPLATE_VIS bit_xor -{ - template - _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY - auto operator()(_T1&& __t, _T2&& __u) const - _NOEXCEPT_(noexcept(_VSTD::forward<_T1>(__t) ^ _VSTD::forward<_T2>(__u))) - -> decltype (_VSTD::forward<_T1>(__t) ^ _VSTD::forward<_T2>(__u)) - { return _VSTD::forward<_T1>(__t) ^ _VSTD::forward<_T2>(__u); } - typedef void is_transparent; -}; -#endif - - -#if _LIBCPP_STD_VER > 11 -template -struct _LIBCPP_TEMPLATE_VIS bit_not : unary_function<_Tp, _Tp> -{ - _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY - _Tp operator()(const _Tp& __x) const - {return ~__x;} -}; - -template <> -struct _LIBCPP_TEMPLATE_VIS bit_not -{ - template - _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY - auto operator()(_Tp&& __x) const - _NOEXCEPT_(noexcept(~_VSTD::forward<_Tp>(__x))) - -> decltype (~_VSTD::forward<_Tp>(__x)) - { return ~_VSTD::forward<_Tp>(__x); } - typedef void is_transparent; -}; -#endif - -template -class _LIBCPP_TEMPLATE_VIS _LIBCPP_DEPRECATED_IN_CXX17 unary_negate - : public unary_function -{ - _Predicate __pred_; -public: - _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY - explicit unary_negate(const _Predicate& __pred) - : __pred_(__pred) {} - _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY - bool operator()(const typename _Predicate::argument_type& __x) const - {return !__pred_(__x);} -}; - -template -_LIBCPP_DEPRECATED_IN_CXX17 inline _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY -unary_negate<_Predicate> -not1(const _Predicate& __pred) {return unary_negate<_Predicate>(__pred);} - -template -class _LIBCPP_TEMPLATE_VIS _LIBCPP_DEPRECATED_IN_CXX17 binary_negate - : public binary_function -{ - _Predicate __pred_; -public: - _LIBCPP_INLINE_VISIBILITY explicit _LIBCPP_CONSTEXPR_AFTER_CXX11 - binary_negate(const _Predicate& __pred) : __pred_(__pred) {} - - _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY - bool operator()(const typename _Predicate::first_argument_type& __x, - const typename _Predicate::second_argument_type& __y) const - {return !__pred_(__x, __y);} -}; - -template -_LIBCPP_DEPRECATED_IN_CXX17 inline _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY -binary_negate<_Predicate> -not2(const _Predicate& __pred) {return binary_negate<_Predicate>(__pred);} - -#if _LIBCPP_STD_VER <= 14 || defined(_LIBCPP_ENABLE_CXX17_REMOVED_BINDERS) -template -class _LIBCPP_TEMPLATE_VIS _LIBCPP_DEPRECATED_IN_CXX11 binder1st - : public unary_function -{ -protected: - __Operation op; - typename __Operation::first_argument_type value; -public: - _LIBCPP_INLINE_VISIBILITY binder1st(const __Operation& __x, - const typename __Operation::first_argument_type __y) - : op(__x), value(__y) {} - _LIBCPP_INLINE_VISIBILITY typename __Operation::result_type operator() - (typename __Operation::second_argument_type& __x) const - {return op(value, __x);} - _LIBCPP_INLINE_VISIBILITY typename __Operation::result_type operator() - (const typename __Operation::second_argument_type& __x) const - {return op(value, __x);} -}; - -template -_LIBCPP_DEPRECATED_IN_CXX11 inline _LIBCPP_INLINE_VISIBILITY -binder1st<__Operation> -bind1st(const __Operation& __op, const _Tp& __x) - {return binder1st<__Operation>(__op, __x);} - -template -class _LIBCPP_TEMPLATE_VIS _LIBCPP_DEPRECATED_IN_CXX11 binder2nd - : public unary_function -{ -protected: - __Operation op; - typename __Operation::second_argument_type value; -public: - _LIBCPP_INLINE_VISIBILITY - binder2nd(const __Operation& __x, const typename __Operation::second_argument_type __y) - : op(__x), value(__y) {} - _LIBCPP_INLINE_VISIBILITY typename __Operation::result_type operator() - ( typename __Operation::first_argument_type& __x) const - {return op(__x, value);} - _LIBCPP_INLINE_VISIBILITY typename __Operation::result_type operator() - (const typename __Operation::first_argument_type& __x) const - {return op(__x, value);} -}; - -template -_LIBCPP_DEPRECATED_IN_CXX11 inline _LIBCPP_INLINE_VISIBILITY -binder2nd<__Operation> -bind2nd(const __Operation& __op, const _Tp& __x) - {return binder2nd<__Operation>(__op, __x);} - -template -class _LIBCPP_TEMPLATE_VIS _LIBCPP_DEPRECATED_IN_CXX11 pointer_to_unary_function - : public unary_function<_Arg, _Result> -{ - _Result (*__f_)(_Arg); -public: - _LIBCPP_INLINE_VISIBILITY explicit pointer_to_unary_function(_Result (*__f)(_Arg)) - : __f_(__f) {} - _LIBCPP_INLINE_VISIBILITY _Result operator()(_Arg __x) const - {return __f_(__x);} -}; - -template -_LIBCPP_DEPRECATED_IN_CXX11 inline _LIBCPP_INLINE_VISIBILITY -pointer_to_unary_function<_Arg,_Result> -ptr_fun(_Result (*__f)(_Arg)) - {return pointer_to_unary_function<_Arg,_Result>(__f);} - -template -class _LIBCPP_TEMPLATE_VIS _LIBCPP_DEPRECATED_IN_CXX11 pointer_to_binary_function - : public binary_function<_Arg1, _Arg2, _Result> -{ - _Result (*__f_)(_Arg1, _Arg2); -public: - _LIBCPP_INLINE_VISIBILITY explicit pointer_to_binary_function(_Result (*__f)(_Arg1, _Arg2)) - : __f_(__f) {} - _LIBCPP_INLINE_VISIBILITY _Result operator()(_Arg1 __x, _Arg2 __y) const - {return __f_(__x, __y);} -}; - -template -_LIBCPP_DEPRECATED_IN_CXX11 inline _LIBCPP_INLINE_VISIBILITY -pointer_to_binary_function<_Arg1,_Arg2,_Result> -ptr_fun(_Result (*__f)(_Arg1,_Arg2)) - {return pointer_to_binary_function<_Arg1,_Arg2,_Result>(__f);} - -template -class _LIBCPP_TEMPLATE_VIS _LIBCPP_DEPRECATED_IN_CXX11 mem_fun_t - : public unary_function<_Tp*, _Sp> -{ - _Sp (_Tp::*__p_)(); -public: - _LIBCPP_INLINE_VISIBILITY explicit mem_fun_t(_Sp (_Tp::*__p)()) - : __p_(__p) {} - _LIBCPP_INLINE_VISIBILITY _Sp operator()(_Tp* __p) const - {return (__p->*__p_)();} -}; - -template -class _LIBCPP_TEMPLATE_VIS _LIBCPP_DEPRECATED_IN_CXX11 mem_fun1_t - : public binary_function<_Tp*, _Ap, _Sp> -{ - _Sp (_Tp::*__p_)(_Ap); -public: - _LIBCPP_INLINE_VISIBILITY explicit mem_fun1_t(_Sp (_Tp::*__p)(_Ap)) - : __p_(__p) {} - _LIBCPP_INLINE_VISIBILITY _Sp operator()(_Tp* __p, _Ap __x) const - {return (__p->*__p_)(__x);} -}; - -template -_LIBCPP_DEPRECATED_IN_CXX11 inline _LIBCPP_INLINE_VISIBILITY -mem_fun_t<_Sp,_Tp> -mem_fun(_Sp (_Tp::*__f)()) - {return mem_fun_t<_Sp,_Tp>(__f);} - -template -_LIBCPP_DEPRECATED_IN_CXX11 inline _LIBCPP_INLINE_VISIBILITY -mem_fun1_t<_Sp,_Tp,_Ap> -mem_fun(_Sp (_Tp::*__f)(_Ap)) - {return mem_fun1_t<_Sp,_Tp,_Ap>(__f);} - -template -class _LIBCPP_TEMPLATE_VIS _LIBCPP_DEPRECATED_IN_CXX11 mem_fun_ref_t - : public unary_function<_Tp, _Sp> -{ - _Sp (_Tp::*__p_)(); -public: - _LIBCPP_INLINE_VISIBILITY explicit mem_fun_ref_t(_Sp (_Tp::*__p)()) - : __p_(__p) {} - _LIBCPP_INLINE_VISIBILITY _Sp operator()(_Tp& __p) const - {return (__p.*__p_)();} -}; - -template -class _LIBCPP_TEMPLATE_VIS _LIBCPP_DEPRECATED_IN_CXX11 mem_fun1_ref_t - : public binary_function<_Tp, _Ap, _Sp> -{ - _Sp (_Tp::*__p_)(_Ap); -public: - _LIBCPP_INLINE_VISIBILITY explicit mem_fun1_ref_t(_Sp (_Tp::*__p)(_Ap)) - : __p_(__p) {} - _LIBCPP_INLINE_VISIBILITY _Sp operator()(_Tp& __p, _Ap __x) const - {return (__p.*__p_)(__x);} -}; - -template -_LIBCPP_DEPRECATED_IN_CXX11 inline _LIBCPP_INLINE_VISIBILITY -mem_fun_ref_t<_Sp,_Tp> -mem_fun_ref(_Sp (_Tp::*__f)()) - {return mem_fun_ref_t<_Sp,_Tp>(__f);} - -template -_LIBCPP_DEPRECATED_IN_CXX11 inline _LIBCPP_INLINE_VISIBILITY -mem_fun1_ref_t<_Sp,_Tp,_Ap> -mem_fun_ref(_Sp (_Tp::*__f)(_Ap)) - {return mem_fun1_ref_t<_Sp,_Tp,_Ap>(__f);} - -template -class _LIBCPP_TEMPLATE_VIS _LIBCPP_DEPRECATED_IN_CXX11 const_mem_fun_t - : public unary_function -{ - _Sp (_Tp::*__p_)() const; -public: - _LIBCPP_INLINE_VISIBILITY explicit const_mem_fun_t(_Sp (_Tp::*__p)() const) - : __p_(__p) {} - _LIBCPP_INLINE_VISIBILITY _Sp operator()(const _Tp* __p) const - {return (__p->*__p_)();} -}; - -template -class _LIBCPP_TEMPLATE_VIS _LIBCPP_DEPRECATED_IN_CXX11 const_mem_fun1_t - : public binary_function -{ - _Sp (_Tp::*__p_)(_Ap) const; -public: - _LIBCPP_INLINE_VISIBILITY explicit const_mem_fun1_t(_Sp (_Tp::*__p)(_Ap) const) - : __p_(__p) {} - _LIBCPP_INLINE_VISIBILITY _Sp operator()(const _Tp* __p, _Ap __x) const - {return (__p->*__p_)(__x);} -}; - -template -_LIBCPP_DEPRECATED_IN_CXX11 inline _LIBCPP_INLINE_VISIBILITY -const_mem_fun_t<_Sp,_Tp> -mem_fun(_Sp (_Tp::*__f)() const) - {return const_mem_fun_t<_Sp,_Tp>(__f);} - -template -_LIBCPP_DEPRECATED_IN_CXX11 inline _LIBCPP_INLINE_VISIBILITY -const_mem_fun1_t<_Sp,_Tp,_Ap> -mem_fun(_Sp (_Tp::*__f)(_Ap) const) - {return const_mem_fun1_t<_Sp,_Tp,_Ap>(__f);} - -template -class _LIBCPP_TEMPLATE_VIS _LIBCPP_DEPRECATED_IN_CXX11 const_mem_fun_ref_t - : public unary_function<_Tp, _Sp> -{ - _Sp (_Tp::*__p_)() const; -public: - _LIBCPP_INLINE_VISIBILITY explicit const_mem_fun_ref_t(_Sp (_Tp::*__p)() const) - : __p_(__p) {} - _LIBCPP_INLINE_VISIBILITY _Sp operator()(const _Tp& __p) const - {return (__p.*__p_)();} -}; - -template -class _LIBCPP_TEMPLATE_VIS _LIBCPP_DEPRECATED_IN_CXX11 const_mem_fun1_ref_t - : public binary_function<_Tp, _Ap, _Sp> -{ - _Sp (_Tp::*__p_)(_Ap) const; -public: - _LIBCPP_INLINE_VISIBILITY explicit const_mem_fun1_ref_t(_Sp (_Tp::*__p)(_Ap) const) - : __p_(__p) {} - _LIBCPP_INLINE_VISIBILITY _Sp operator()(const _Tp& __p, _Ap __x) const - {return (__p.*__p_)(__x);} -}; - -template -_LIBCPP_DEPRECATED_IN_CXX11 inline _LIBCPP_INLINE_VISIBILITY -const_mem_fun_ref_t<_Sp,_Tp> -mem_fun_ref(_Sp (_Tp::*__f)() const) - {return const_mem_fun_ref_t<_Sp,_Tp>(__f);} - -template -_LIBCPP_DEPRECATED_IN_CXX11 inline _LIBCPP_INLINE_VISIBILITY -const_mem_fun1_ref_t<_Sp,_Tp,_Ap> -mem_fun_ref(_Sp (_Tp::*__f)(_Ap) const) - {return const_mem_fun1_ref_t<_Sp,_Tp,_Ap>(__f);} -#endif - -//////////////////////////////////////////////////////////////////////////////// -// MEMFUN -//============================================================================== - -template -class __mem_fn - : public __weak_result_type<_Tp> -{ -public: - // types - typedef _Tp type; -private: - type __f_; - -public: - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 - __mem_fn(type __f) _NOEXCEPT : __f_(__f) {} - -#ifndef _LIBCPP_CXX03_LANG - // invoke - template - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 - typename __invoke_return::type - operator() (_ArgTypes&&... __args) const { - return _VSTD::__invoke(__f_, _VSTD::forward<_ArgTypes>(__args)...); - } -#else - - template - _LIBCPP_INLINE_VISIBILITY - typename __invoke_return0::type - operator() (_A0& __a0) const { - return _VSTD::__invoke(__f_, __a0); - } - - template - _LIBCPP_INLINE_VISIBILITY - typename __invoke_return0::type - operator() (_A0 const& __a0) const { - return _VSTD::__invoke(__f_, __a0); - } - - template - _LIBCPP_INLINE_VISIBILITY - typename __invoke_return1::type - operator() (_A0& __a0, _A1& __a1) const { - return _VSTD::__invoke(__f_, __a0, __a1); - } - - template - _LIBCPP_INLINE_VISIBILITY - typename __invoke_return1::type - operator() (_A0 const& __a0, _A1& __a1) const { - return _VSTD::__invoke(__f_, __a0, __a1); - } - - template - _LIBCPP_INLINE_VISIBILITY - typename __invoke_return1::type - operator() (_A0& __a0, _A1 const& __a1) const { - return _VSTD::__invoke(__f_, __a0, __a1); - } - - template - _LIBCPP_INLINE_VISIBILITY - typename __invoke_return1::type - operator() (_A0 const& __a0, _A1 const& __a1) const { - return _VSTD::__invoke(__f_, __a0, __a1); - } - - template - _LIBCPP_INLINE_VISIBILITY - typename __invoke_return2::type - operator() (_A0& __a0, _A1& __a1, _A2& __a2) const { - return _VSTD::__invoke(__f_, __a0, __a1, __a2); - } - - template - _LIBCPP_INLINE_VISIBILITY - typename __invoke_return2::type - operator() (_A0 const& __a0, _A1& __a1, _A2& __a2) const { - return _VSTD::__invoke(__f_, __a0, __a1, __a2); - } - - template - _LIBCPP_INLINE_VISIBILITY - typename __invoke_return2::type - operator() (_A0& __a0, _A1 const& __a1, _A2& __a2) const { - return _VSTD::__invoke(__f_, __a0, __a1, __a2); - } - - template - _LIBCPP_INLINE_VISIBILITY - typename __invoke_return2::type - operator() (_A0& __a0, _A1& __a1, _A2 const& __a2) const { - return _VSTD::__invoke(__f_, __a0, __a1, __a2); - } - - template - _LIBCPP_INLINE_VISIBILITY - typename __invoke_return2::type - operator() (_A0 const& __a0, _A1 const& __a1, _A2& __a2) const { - return _VSTD::__invoke(__f_, __a0, __a1, __a2); - } - - template - _LIBCPP_INLINE_VISIBILITY - typename __invoke_return2::type - operator() (_A0 const& __a0, _A1& __a1, _A2 const& __a2) const { - return _VSTD::__invoke(__f_, __a0, __a1, __a2); - } - - template - _LIBCPP_INLINE_VISIBILITY - typename __invoke_return2::type - operator() (_A0& __a0, _A1 const& __a1, _A2 const& __a2) const { - return _VSTD::__invoke(__f_, __a0, __a1, __a2); - } - - template - _LIBCPP_INLINE_VISIBILITY - typename __invoke_return2::type - operator() (_A0 const& __a0, _A1 const& __a1, _A2 const& __a2) const { - return _VSTD::__invoke(__f_, __a0, __a1, __a2); - } -#endif -}; - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -__mem_fn<_Rp _Tp::*> -mem_fn(_Rp _Tp::* __pm) _NOEXCEPT -{ - return __mem_fn<_Rp _Tp::*>(__pm); -} - -//////////////////////////////////////////////////////////////////////////////// -// FUNCTION -//============================================================================== - -// bad_function_call - -class _LIBCPP_EXCEPTION_ABI bad_function_call - : public exception -{ -#ifdef _LIBCPP_ABI_BAD_FUNCTION_CALL_KEY_FUNCTION -public: - virtual ~bad_function_call() _NOEXCEPT; - - virtual const char* what() const _NOEXCEPT; -#endif -}; - -_LIBCPP_NORETURN inline _LIBCPP_INLINE_VISIBILITY -void __throw_bad_function_call() -{ -#ifndef _LIBCPP_NO_EXCEPTIONS - throw bad_function_call(); -#else - _VSTD::abort(); -#endif -} - -#if defined(_LIBCPP_CXX03_LANG) && !defined(_LIBCPP_DISABLE_DEPRECATION_WARNINGS) && __has_attribute(deprecated) -# define _LIBCPP_DEPRECATED_CXX03_FUNCTION \ - __attribute__((deprecated("Using std::function in C++03 is not supported anymore. Please upgrade to C++11 or later, or use a different type"))) -#else -# define _LIBCPP_DEPRECATED_CXX03_FUNCTION /* nothing */ -#endif - -template class _LIBCPP_DEPRECATED_CXX03_FUNCTION _LIBCPP_TEMPLATE_VIS function; // undefined - -namespace __function -{ - -template -struct __maybe_derive_from_unary_function -{ -}; - -template -struct __maybe_derive_from_unary_function<_Rp(_A1)> - : public unary_function<_A1, _Rp> -{ -}; - -template -struct __maybe_derive_from_binary_function -{ -}; - -template -struct __maybe_derive_from_binary_function<_Rp(_A1, _A2)> - : public binary_function<_A1, _A2, _Rp> -{ -}; - -template -_LIBCPP_INLINE_VISIBILITY -bool __not_null(_Fp const&) { return true; } - -template -_LIBCPP_INLINE_VISIBILITY -bool __not_null(_Fp* __ptr) { return __ptr; } - -template -_LIBCPP_INLINE_VISIBILITY -bool __not_null(_Ret _Class::*__ptr) { return __ptr; } - -template -_LIBCPP_INLINE_VISIBILITY -bool __not_null(function<_Fp> const& __f) { return !!__f; } - -#ifdef _LIBCPP_HAS_EXTENSION_BLOCKS -template -_LIBCPP_INLINE_VISIBILITY -bool __not_null(_Rp (^__p)(_Args...)) { return __p; } -#endif - -} // namespace __function - -#ifndef _LIBCPP_CXX03_LANG - -namespace __function { - -// __alloc_func holds a functor and an allocator. - -template class __alloc_func; -template -class __default_alloc_func; - -template -class __alloc_func<_Fp, _Ap, _Rp(_ArgTypes...)> -{ - __compressed_pair<_Fp, _Ap> __f_; - - public: - typedef _LIBCPP_NODEBUG_TYPE _Fp _Target; - typedef _LIBCPP_NODEBUG_TYPE _Ap _Alloc; - - _LIBCPP_INLINE_VISIBILITY - const _Target& __target() const { return __f_.first(); } - - // WIN32 APIs may define __allocator, so use __get_allocator instead. - _LIBCPP_INLINE_VISIBILITY - const _Alloc& __get_allocator() const { return __f_.second(); } - - _LIBCPP_INLINE_VISIBILITY - explicit __alloc_func(_Target&& __f) - : __f_(piecewise_construct, _VSTD::forward_as_tuple(_VSTD::move(__f)), - _VSTD::forward_as_tuple()) - { - } - - _LIBCPP_INLINE_VISIBILITY - explicit __alloc_func(const _Target& __f, const _Alloc& __a) - : __f_(piecewise_construct, _VSTD::forward_as_tuple(__f), - _VSTD::forward_as_tuple(__a)) - { - } - - _LIBCPP_INLINE_VISIBILITY - explicit __alloc_func(const _Target& __f, _Alloc&& __a) - : __f_(piecewise_construct, _VSTD::forward_as_tuple(__f), - _VSTD::forward_as_tuple(_VSTD::move(__a))) - { - } - - _LIBCPP_INLINE_VISIBILITY - explicit __alloc_func(_Target&& __f, _Alloc&& __a) - : __f_(piecewise_construct, _VSTD::forward_as_tuple(_VSTD::move(__f)), - _VSTD::forward_as_tuple(_VSTD::move(__a))) - { - } - - _LIBCPP_INLINE_VISIBILITY - _Rp operator()(_ArgTypes&&... __arg) - { - typedef __invoke_void_return_wrapper<_Rp> _Invoker; - return _Invoker::__call(__f_.first(), - _VSTD::forward<_ArgTypes>(__arg)...); - } - - _LIBCPP_INLINE_VISIBILITY - __alloc_func* __clone() const - { - typedef allocator_traits<_Alloc> __alloc_traits; - typedef - typename __rebind_alloc_helper<__alloc_traits, __alloc_func>::type - _AA; - _AA __a(__f_.second()); - typedef __allocator_destructor<_AA> _Dp; - unique_ptr<__alloc_func, _Dp> __hold(__a.allocate(1), _Dp(__a, 1)); - ::new ((void*)__hold.get()) __alloc_func(__f_.first(), _Alloc(__a)); - return __hold.release(); - } - - _LIBCPP_INLINE_VISIBILITY - void destroy() _NOEXCEPT { __f_.~__compressed_pair<_Target, _Alloc>(); } - - static void __destroy_and_delete(__alloc_func* __f) { - typedef allocator_traits<_Alloc> __alloc_traits; - typedef typename __rebind_alloc_helper<__alloc_traits, __alloc_func>::type - _FunAlloc; - _FunAlloc __a(__f->__get_allocator()); - __f->destroy(); - __a.deallocate(__f, 1); - } -}; - -template -class __default_alloc_func<_Fp, _Rp(_ArgTypes...)> { - _Fp __f_; - -public: - typedef _LIBCPP_NODEBUG_TYPE _Fp _Target; - - _LIBCPP_INLINE_VISIBILITY - const _Target& __target() const { return __f_; } - - _LIBCPP_INLINE_VISIBILITY - explicit __default_alloc_func(_Target&& __f) : __f_(_VSTD::move(__f)) {} - - _LIBCPP_INLINE_VISIBILITY - explicit __default_alloc_func(const _Target& __f) : __f_(__f) {} - - _LIBCPP_INLINE_VISIBILITY - _Rp operator()(_ArgTypes&&... __arg) { - typedef __invoke_void_return_wrapper<_Rp> _Invoker; - return _Invoker::__call(__f_, _VSTD::forward<_ArgTypes>(__arg)...); - } - - _LIBCPP_INLINE_VISIBILITY - __default_alloc_func* __clone() const { - __builtin_new_allocator::__holder_t __hold = - __builtin_new_allocator::__allocate_type<__default_alloc_func>(1); - __default_alloc_func* __res = - ::new ((void*)__hold.get()) __default_alloc_func(__f_); - (void)__hold.release(); - return __res; - } - - _LIBCPP_INLINE_VISIBILITY - void destroy() _NOEXCEPT { __f_.~_Target(); } - - static void __destroy_and_delete(__default_alloc_func* __f) { - __f->destroy(); - __builtin_new_allocator::__deallocate_type<__default_alloc_func>(__f, 1); - } -}; - -// __base provides an abstract interface for copyable functors. - -template class _LIBCPP_TEMPLATE_VIS __base; - -template -class __base<_Rp(_ArgTypes...)> -{ - __base(const __base&); - __base& operator=(const __base&); -public: - _LIBCPP_INLINE_VISIBILITY __base() {} - _LIBCPP_INLINE_VISIBILITY virtual ~__base() {} - virtual __base* __clone() const = 0; - virtual void __clone(__base*) const = 0; - virtual void destroy() _NOEXCEPT = 0; - virtual void destroy_deallocate() _NOEXCEPT = 0; - virtual _Rp operator()(_ArgTypes&& ...) = 0; -#ifndef _LIBCPP_NO_RTTI - virtual const void* target(const type_info&) const _NOEXCEPT = 0; - virtual const std::type_info& target_type() const _NOEXCEPT = 0; -#endif // _LIBCPP_NO_RTTI -}; - -// __func implements __base for a given functor type. - -template class __func; - -template -class __func<_Fp, _Alloc, _Rp(_ArgTypes...)> - : public __base<_Rp(_ArgTypes...)> -{ - __alloc_func<_Fp, _Alloc, _Rp(_ArgTypes...)> __f_; -public: - _LIBCPP_INLINE_VISIBILITY - explicit __func(_Fp&& __f) - : __f_(_VSTD::move(__f)) {} - - _LIBCPP_INLINE_VISIBILITY - explicit __func(const _Fp& __f, const _Alloc& __a) - : __f_(__f, __a) {} - - _LIBCPP_INLINE_VISIBILITY - explicit __func(const _Fp& __f, _Alloc&& __a) - : __f_(__f, _VSTD::move(__a)) {} - - _LIBCPP_INLINE_VISIBILITY - explicit __func(_Fp&& __f, _Alloc&& __a) - : __f_(_VSTD::move(__f), _VSTD::move(__a)) {} - - virtual __base<_Rp(_ArgTypes...)>* __clone() const; - virtual void __clone(__base<_Rp(_ArgTypes...)>*) const; - virtual void destroy() _NOEXCEPT; - virtual void destroy_deallocate() _NOEXCEPT; - virtual _Rp operator()(_ArgTypes&&... __arg); -#ifndef _LIBCPP_NO_RTTI - virtual const void* target(const type_info&) const _NOEXCEPT; - virtual const std::type_info& target_type() const _NOEXCEPT; -#endif // _LIBCPP_NO_RTTI -}; - -template -__base<_Rp(_ArgTypes...)>* -__func<_Fp, _Alloc, _Rp(_ArgTypes...)>::__clone() const -{ - typedef allocator_traits<_Alloc> __alloc_traits; - typedef typename __rebind_alloc_helper<__alloc_traits, __func>::type _Ap; - _Ap __a(__f_.__get_allocator()); - typedef __allocator_destructor<_Ap> _Dp; - unique_ptr<__func, _Dp> __hold(__a.allocate(1), _Dp(__a, 1)); - ::new ((void*)__hold.get()) __func(__f_.__target(), _Alloc(__a)); - return __hold.release(); -} - -template -void -__func<_Fp, _Alloc, _Rp(_ArgTypes...)>::__clone(__base<_Rp(_ArgTypes...)>* __p) const -{ - ::new ((void*)__p) __func(__f_.__target(), __f_.__get_allocator()); -} - -template -void -__func<_Fp, _Alloc, _Rp(_ArgTypes...)>::destroy() _NOEXCEPT -{ - __f_.destroy(); -} - -template -void -__func<_Fp, _Alloc, _Rp(_ArgTypes...)>::destroy_deallocate() _NOEXCEPT -{ - typedef allocator_traits<_Alloc> __alloc_traits; - typedef typename __rebind_alloc_helper<__alloc_traits, __func>::type _Ap; - _Ap __a(__f_.__get_allocator()); - __f_.destroy(); - __a.deallocate(this, 1); -} - -template -_Rp -__func<_Fp, _Alloc, _Rp(_ArgTypes...)>::operator()(_ArgTypes&& ... __arg) -{ - return __f_(_VSTD::forward<_ArgTypes>(__arg)...); -} - -#ifndef _LIBCPP_NO_RTTI - -template -const void* -__func<_Fp, _Alloc, _Rp(_ArgTypes...)>::target(const type_info& __ti) const _NOEXCEPT -{ - if (__ti == typeid(_Fp)) - return &__f_.__target(); - return nullptr; -} - -template -const std::type_info& -__func<_Fp, _Alloc, _Rp(_ArgTypes...)>::target_type() const _NOEXCEPT -{ - return typeid(_Fp); -} - -#endif // _LIBCPP_NO_RTTI - -// __value_func creates a value-type from a __func. - -template class __value_func; - -template class __value_func<_Rp(_ArgTypes...)> -{ - typename aligned_storage<3 * sizeof(void*)>::type __buf_; - - typedef __base<_Rp(_ArgTypes...)> __func; - __func* __f_; - - _LIBCPP_NO_CFI static __func* __as_base(void* p) - { - return reinterpret_cast<__func*>(p); - } - - public: - _LIBCPP_INLINE_VISIBILITY - __value_func() _NOEXCEPT : __f_(nullptr) {} - - template - _LIBCPP_INLINE_VISIBILITY __value_func(_Fp&& __f, const _Alloc& __a) - : __f_(nullptr) - { - typedef allocator_traits<_Alloc> __alloc_traits; - typedef __function::__func<_Fp, _Alloc, _Rp(_ArgTypes...)> _Fun; - typedef typename __rebind_alloc_helper<__alloc_traits, _Fun>::type - _FunAlloc; - - if (__function::__not_null(__f)) - { - _FunAlloc __af(__a); - if (sizeof(_Fun) <= sizeof(__buf_) && - is_nothrow_copy_constructible<_Fp>::value && - is_nothrow_copy_constructible<_FunAlloc>::value) - { - __f_ = - ::new ((void*)&__buf_) _Fun(_VSTD::move(__f), _Alloc(__af)); - } - else - { - typedef __allocator_destructor<_FunAlloc> _Dp; - unique_ptr<__func, _Dp> __hold(__af.allocate(1), _Dp(__af, 1)); - ::new ((void*)__hold.get()) _Fun(_VSTD::move(__f), _Alloc(__a)); - __f_ = __hold.release(); - } - } - } - - template ::type, __value_func>::value>::type> - _LIBCPP_INLINE_VISIBILITY explicit __value_func(_Fp&& __f) - : __value_func(_VSTD::forward<_Fp>(__f), allocator<_Fp>()) {} - - _LIBCPP_INLINE_VISIBILITY - __value_func(const __value_func& __f) - { - if (__f.__f_ == nullptr) - __f_ = nullptr; - else if ((void*)__f.__f_ == &__f.__buf_) - { - __f_ = __as_base(&__buf_); - __f.__f_->__clone(__f_); - } - else - __f_ = __f.__f_->__clone(); - } - - _LIBCPP_INLINE_VISIBILITY - __value_func(__value_func&& __f) _NOEXCEPT - { - if (__f.__f_ == nullptr) - __f_ = nullptr; - else if ((void*)__f.__f_ == &__f.__buf_) - { - __f_ = __as_base(&__buf_); - __f.__f_->__clone(__f_); - } - else - { - __f_ = __f.__f_; - __f.__f_ = nullptr; - } - } - - _LIBCPP_INLINE_VISIBILITY - ~__value_func() - { - if ((void*)__f_ == &__buf_) - __f_->destroy(); - else if (__f_) - __f_->destroy_deallocate(); - } - - _LIBCPP_INLINE_VISIBILITY - __value_func& operator=(__value_func&& __f) - { - *this = nullptr; - if (__f.__f_ == nullptr) - __f_ = nullptr; - else if ((void*)__f.__f_ == &__f.__buf_) - { - __f_ = __as_base(&__buf_); - __f.__f_->__clone(__f_); - } - else - { - __f_ = __f.__f_; - __f.__f_ = nullptr; - } - return *this; - } - - _LIBCPP_INLINE_VISIBILITY - __value_func& operator=(nullptr_t) - { - __func* __f = __f_; - __f_ = nullptr; - if ((void*)__f == &__buf_) - __f->destroy(); - else if (__f) - __f->destroy_deallocate(); - return *this; - } - - _LIBCPP_INLINE_VISIBILITY - _Rp operator()(_ArgTypes&&... __args) const - { - if (__f_ == nullptr) - __throw_bad_function_call(); - return (*__f_)(_VSTD::forward<_ArgTypes>(__args)...); - } - - _LIBCPP_INLINE_VISIBILITY - void swap(__value_func& __f) _NOEXCEPT - { - if (&__f == this) - return; - if ((void*)__f_ == &__buf_ && (void*)__f.__f_ == &__f.__buf_) - { - typename aligned_storage::type __tempbuf; - __func* __t = __as_base(&__tempbuf); - __f_->__clone(__t); - __f_->destroy(); - __f_ = nullptr; - __f.__f_->__clone(__as_base(&__buf_)); - __f.__f_->destroy(); - __f.__f_ = nullptr; - __f_ = __as_base(&__buf_); - __t->__clone(__as_base(&__f.__buf_)); - __t->destroy(); - __f.__f_ = __as_base(&__f.__buf_); - } - else if ((void*)__f_ == &__buf_) - { - __f_->__clone(__as_base(&__f.__buf_)); - __f_->destroy(); - __f_ = __f.__f_; - __f.__f_ = __as_base(&__f.__buf_); - } - else if ((void*)__f.__f_ == &__f.__buf_) - { - __f.__f_->__clone(__as_base(&__buf_)); - __f.__f_->destroy(); - __f.__f_ = __f_; - __f_ = __as_base(&__buf_); - } - else - _VSTD::swap(__f_, __f.__f_); - } - - _LIBCPP_INLINE_VISIBILITY - _LIBCPP_EXPLICIT operator bool() const _NOEXCEPT { return __f_ != nullptr; } - -#ifndef _LIBCPP_NO_RTTI - _LIBCPP_INLINE_VISIBILITY - const std::type_info& target_type() const _NOEXCEPT - { - if (__f_ == nullptr) - return typeid(void); - return __f_->target_type(); - } - - template - _LIBCPP_INLINE_VISIBILITY const _Tp* target() const _NOEXCEPT - { - if (__f_ == nullptr) - return nullptr; - return (const _Tp*)__f_->target(typeid(_Tp)); - } -#endif // _LIBCPP_NO_RTTI -}; - -// Storage for a functor object, to be used with __policy to manage copy and -// destruction. -union __policy_storage -{ - mutable char __small[sizeof(void*) * 2]; - void* __large; -}; - -// True if _Fun can safely be held in __policy_storage.__small. -template -struct __use_small_storage - : public _VSTD::integral_constant< - bool, sizeof(_Fun) <= sizeof(__policy_storage) && - _LIBCPP_ALIGNOF(_Fun) <= _LIBCPP_ALIGNOF(__policy_storage) && - _VSTD::is_trivially_copy_constructible<_Fun>::value && - _VSTD::is_trivially_destructible<_Fun>::value> {}; - -// Policy contains information about how to copy, destroy, and move the -// underlying functor. You can think of it as a vtable of sorts. -struct __policy -{ - // Used to copy or destroy __large values. null for trivial objects. - void* (*const __clone)(const void*); - void (*const __destroy)(void*); - - // True if this is the null policy (no value). - const bool __is_null; - - // The target type. May be null if RTTI is disabled. - const std::type_info* const __type_info; - - // Returns a pointer to a static policy object suitable for the functor - // type. - template - _LIBCPP_INLINE_VISIBILITY static const __policy* __create() - { - return __choose_policy<_Fun>(__use_small_storage<_Fun>()); - } - - _LIBCPP_INLINE_VISIBILITY - static const __policy* __create_empty() - { - static const _LIBCPP_CONSTEXPR __policy __policy_ = {nullptr, nullptr, - true, -#ifndef _LIBCPP_NO_RTTI - &typeid(void) -#else - nullptr -#endif - }; - return &__policy_; - } - - private: - template static void* __large_clone(const void* __s) - { - const _Fun* __f = static_cast(__s); - return __f->__clone(); - } - - template - static void __large_destroy(void* __s) { - _Fun::__destroy_and_delete(static_cast<_Fun*>(__s)); - } - - template - _LIBCPP_INLINE_VISIBILITY static const __policy* - __choose_policy(/* is_small = */ false_type) { - static const _LIBCPP_CONSTEXPR __policy __policy_ = { - &__large_clone<_Fun>, &__large_destroy<_Fun>, false, -#ifndef _LIBCPP_NO_RTTI - &typeid(typename _Fun::_Target) -#else - nullptr -#endif - }; - return &__policy_; - } - - template - _LIBCPP_INLINE_VISIBILITY static const __policy* - __choose_policy(/* is_small = */ true_type) - { - static const _LIBCPP_CONSTEXPR __policy __policy_ = { - nullptr, nullptr, false, -#ifndef _LIBCPP_NO_RTTI - &typeid(typename _Fun::_Target) -#else - nullptr -#endif - }; - return &__policy_; - } -}; - -// Used to choose between perfect forwarding or pass-by-value. Pass-by-value is -// faster for types that can be passed in registers. -template -using __fast_forward = - typename _VSTD::conditional<_VSTD::is_scalar<_Tp>::value, _Tp, _Tp&&>::type; - -// __policy_invoker calls an instance of __alloc_func held in __policy_storage. - -template struct __policy_invoker; - -template -struct __policy_invoker<_Rp(_ArgTypes...)> -{ - typedef _Rp (*__Call)(const __policy_storage*, - __fast_forward<_ArgTypes>...); - - __Call __call_; - - // Creates an invoker that throws bad_function_call. - _LIBCPP_INLINE_VISIBILITY - __policy_invoker() : __call_(&__call_empty) {} - - // Creates an invoker that calls the given instance of __func. - template - _LIBCPP_INLINE_VISIBILITY static __policy_invoker __create() - { - return __policy_invoker(&__call_impl<_Fun>); - } - - private: - _LIBCPP_INLINE_VISIBILITY - explicit __policy_invoker(__Call __c) : __call_(__c) {} - - static _Rp __call_empty(const __policy_storage*, - __fast_forward<_ArgTypes>...) - { - __throw_bad_function_call(); - } - - template - static _Rp __call_impl(const __policy_storage* __buf, - __fast_forward<_ArgTypes>... __args) - { - _Fun* __f = reinterpret_cast<_Fun*>(__use_small_storage<_Fun>::value - ? &__buf->__small - : __buf->__large); - return (*__f)(_VSTD::forward<_ArgTypes>(__args)...); - } -}; - -// __policy_func uses a __policy and __policy_invoker to create a type-erased, -// copyable functor. - -template class __policy_func; - -template class __policy_func<_Rp(_ArgTypes...)> -{ - // Inline storage for small objects. - __policy_storage __buf_; - - // Calls the value stored in __buf_. This could technically be part of - // policy, but storing it here eliminates a level of indirection inside - // operator(). - typedef __function::__policy_invoker<_Rp(_ArgTypes...)> __invoker; - __invoker __invoker_; - - // The policy that describes how to move / copy / destroy __buf_. Never - // null, even if the function is empty. - const __policy* __policy_; - - public: - _LIBCPP_INLINE_VISIBILITY - __policy_func() : __policy_(__policy::__create_empty()) {} - - template - _LIBCPP_INLINE_VISIBILITY __policy_func(_Fp&& __f, const _Alloc& __a) - : __policy_(__policy::__create_empty()) - { - typedef __alloc_func<_Fp, _Alloc, _Rp(_ArgTypes...)> _Fun; - typedef allocator_traits<_Alloc> __alloc_traits; - typedef typename __rebind_alloc_helper<__alloc_traits, _Fun>::type - _FunAlloc; - - if (__function::__not_null(__f)) - { - __invoker_ = __invoker::template __create<_Fun>(); - __policy_ = __policy::__create<_Fun>(); - - _FunAlloc __af(__a); - if (__use_small_storage<_Fun>()) - { - ::new ((void*)&__buf_.__small) - _Fun(_VSTD::move(__f), _Alloc(__af)); - } - else - { - typedef __allocator_destructor<_FunAlloc> _Dp; - unique_ptr<_Fun, _Dp> __hold(__af.allocate(1), _Dp(__af, 1)); - ::new ((void*)__hold.get()) - _Fun(_VSTD::move(__f), _Alloc(__af)); - __buf_.__large = __hold.release(); - } - } - } - - template ::type, __policy_func>::value>::type> - _LIBCPP_INLINE_VISIBILITY explicit __policy_func(_Fp&& __f) - : __policy_(__policy::__create_empty()) { - typedef __default_alloc_func<_Fp, _Rp(_ArgTypes...)> _Fun; - - if (__function::__not_null(__f)) { - __invoker_ = __invoker::template __create<_Fun>(); - __policy_ = __policy::__create<_Fun>(); - if (__use_small_storage<_Fun>()) { - ::new ((void*)&__buf_.__small) _Fun(_VSTD::move(__f)); - } else { - __builtin_new_allocator::__holder_t __hold = - __builtin_new_allocator::__allocate_type<_Fun>(1); - __buf_.__large = ::new ((void*)__hold.get()) _Fun(_VSTD::move(__f)); - (void)__hold.release(); - } - } - } - - _LIBCPP_INLINE_VISIBILITY - __policy_func(const __policy_func& __f) - : __buf_(__f.__buf_), __invoker_(__f.__invoker_), - __policy_(__f.__policy_) - { - if (__policy_->__clone) - __buf_.__large = __policy_->__clone(__f.__buf_.__large); - } - - _LIBCPP_INLINE_VISIBILITY - __policy_func(__policy_func&& __f) - : __buf_(__f.__buf_), __invoker_(__f.__invoker_), - __policy_(__f.__policy_) - { - if (__policy_->__destroy) - { - __f.__policy_ = __policy::__create_empty(); - __f.__invoker_ = __invoker(); - } - } - - _LIBCPP_INLINE_VISIBILITY - ~__policy_func() - { - if (__policy_->__destroy) - __policy_->__destroy(__buf_.__large); - } - - _LIBCPP_INLINE_VISIBILITY - __policy_func& operator=(__policy_func&& __f) - { - *this = nullptr; - __buf_ = __f.__buf_; - __invoker_ = __f.__invoker_; - __policy_ = __f.__policy_; - __f.__policy_ = __policy::__create_empty(); - __f.__invoker_ = __invoker(); - return *this; - } - - _LIBCPP_INLINE_VISIBILITY - __policy_func& operator=(nullptr_t) - { - const __policy* __p = __policy_; - __policy_ = __policy::__create_empty(); - __invoker_ = __invoker(); - if (__p->__destroy) - __p->__destroy(__buf_.__large); - return *this; - } - - _LIBCPP_INLINE_VISIBILITY - _Rp operator()(_ArgTypes&&... __args) const - { - return __invoker_.__call_(_VSTD::addressof(__buf_), - _VSTD::forward<_ArgTypes>(__args)...); - } - - _LIBCPP_INLINE_VISIBILITY - void swap(__policy_func& __f) - { - _VSTD::swap(__invoker_, __f.__invoker_); - _VSTD::swap(__policy_, __f.__policy_); - _VSTD::swap(__buf_, __f.__buf_); - } - - _LIBCPP_INLINE_VISIBILITY - explicit operator bool() const _NOEXCEPT - { - return !__policy_->__is_null; - } - -#ifndef _LIBCPP_NO_RTTI - _LIBCPP_INLINE_VISIBILITY - const std::type_info& target_type() const _NOEXCEPT - { - return *__policy_->__type_info; - } - - template - _LIBCPP_INLINE_VISIBILITY const _Tp* target() const _NOEXCEPT - { - if (__policy_->__is_null || typeid(_Tp) != *__policy_->__type_info) - return nullptr; - if (__policy_->__clone) // Out of line storage. - return reinterpret_cast(__buf_.__large); - else - return reinterpret_cast(&__buf_.__small); - } -#endif // _LIBCPP_NO_RTTI -}; - -#if defined(_LIBCPP_HAS_BLOCKS_RUNTIME) && !defined(_LIBCPP_HAS_OBJC_ARC) - -extern "C" void *_Block_copy(const void *); -extern "C" void _Block_release(const void *); - -template -class __func<_Rp1(^)(_ArgTypes1...), _Alloc, _Rp(_ArgTypes...)> - : public __base<_Rp(_ArgTypes...)> -{ - typedef _Rp1(^__block_type)(_ArgTypes1...); - __block_type __f_; - -public: - _LIBCPP_INLINE_VISIBILITY - explicit __func(__block_type const& __f) - : __f_(reinterpret_cast<__block_type>(__f ? _Block_copy(__f) : nullptr)) - { } - - // [TODO] add && to save on a retain - - _LIBCPP_INLINE_VISIBILITY - explicit __func(__block_type __f, const _Alloc& /* unused */) - : __f_(reinterpret_cast<__block_type>(__f ? _Block_copy(__f) : nullptr)) - { } - - virtual __base<_Rp(_ArgTypes...)>* __clone() const { - _LIBCPP_ASSERT(false, - "Block pointers are just pointers, so they should always fit into " - "std::function's small buffer optimization. This function should " - "never be invoked."); - return nullptr; - } - - virtual void __clone(__base<_Rp(_ArgTypes...)>* __p) const { - ::new ((void*)__p) __func(__f_); - } - - virtual void destroy() _NOEXCEPT { - if (__f_) - _Block_release(__f_); - __f_ = 0; - } - - virtual void destroy_deallocate() _NOEXCEPT { - _LIBCPP_ASSERT(false, - "Block pointers are just pointers, so they should always fit into " - "std::function's small buffer optimization. This function should " - "never be invoked."); - } - - virtual _Rp operator()(_ArgTypes&& ... __arg) { - return _VSTD::__invoke(__f_, _VSTD::forward<_ArgTypes>(__arg)...); - } - -#ifndef _LIBCPP_NO_RTTI - virtual const void* target(type_info const& __ti) const _NOEXCEPT { - if (__ti == typeid(__func::__block_type)) - return &__f_; - return (const void*)nullptr; - } - - virtual const std::type_info& target_type() const _NOEXCEPT { - return typeid(__func::__block_type); - } -#endif // _LIBCPP_NO_RTTI -}; - -#endif // _LIBCPP_HAS_EXTENSION_BLOCKS && !_LIBCPP_HAS_OBJC_ARC - -} // __function - -template -class _LIBCPP_TEMPLATE_VIS function<_Rp(_ArgTypes...)> - : public __function::__maybe_derive_from_unary_function<_Rp(_ArgTypes...)>, - public __function::__maybe_derive_from_binary_function<_Rp(_ArgTypes...)> -{ -#ifndef _LIBCPP_ABI_OPTIMIZED_FUNCTION - typedef __function::__value_func<_Rp(_ArgTypes...)> __func; -#else - typedef __function::__policy_func<_Rp(_ArgTypes...)> __func; -#endif - - __func __f_; - - template , function>, - __invokable<_Fp, _ArgTypes...> - >::value> - struct __callable; - template - struct __callable<_Fp, true> - { - static const bool value = is_void<_Rp>::value || - __is_core_convertible::type, - _Rp>::value; - }; - template - struct __callable<_Fp, false> - { - static const bool value = false; - }; - - template - using _EnableIfLValueCallable = typename enable_if<__callable<_Fp&>::value>::type; -public: - typedef _Rp result_type; - - // construct/copy/destroy: - _LIBCPP_INLINE_VISIBILITY - function() _NOEXCEPT { } - _LIBCPP_INLINE_VISIBILITY - function(nullptr_t) _NOEXCEPT {} - function(const function&); - function(function&&) _NOEXCEPT; - template> - function(_Fp); - -#if _LIBCPP_STD_VER <= 14 - template - _LIBCPP_INLINE_VISIBILITY - function(allocator_arg_t, const _Alloc&) _NOEXCEPT {} - template - _LIBCPP_INLINE_VISIBILITY - function(allocator_arg_t, const _Alloc&, nullptr_t) _NOEXCEPT {} - template - function(allocator_arg_t, const _Alloc&, const function&); - template - function(allocator_arg_t, const _Alloc&, function&&); - template> - function(allocator_arg_t, const _Alloc& __a, _Fp __f); -#endif - - function& operator=(const function&); - function& operator=(function&&) _NOEXCEPT; - function& operator=(nullptr_t) _NOEXCEPT; - template::type>> - function& operator=(_Fp&&); - - ~function(); - - // function modifiers: - void swap(function&) _NOEXCEPT; - -#if _LIBCPP_STD_VER <= 14 - template - _LIBCPP_INLINE_VISIBILITY - void assign(_Fp&& __f, const _Alloc& __a) - {function(allocator_arg, __a, _VSTD::forward<_Fp>(__f)).swap(*this);} -#endif - - // function capacity: - _LIBCPP_INLINE_VISIBILITY - _LIBCPP_EXPLICIT operator bool() const _NOEXCEPT { - return static_cast(__f_); - } - - // deleted overloads close possible hole in the type system - template - bool operator==(const function<_R2(_ArgTypes2...)>&) const = delete; - template - bool operator!=(const function<_R2(_ArgTypes2...)>&) const = delete; -public: - // function invocation: - _Rp operator()(_ArgTypes...) const; - -#ifndef _LIBCPP_NO_RTTI - // function target access: - const std::type_info& target_type() const _NOEXCEPT; - template _Tp* target() _NOEXCEPT; - template const _Tp* target() const _NOEXCEPT; -#endif // _LIBCPP_NO_RTTI -}; - -#ifndef _LIBCPP_HAS_NO_DEDUCTION_GUIDES -template -function(_Rp(*)(_Ap...)) -> function<_Rp(_Ap...)>; - -template -struct __strip_signature; - -template -struct __strip_signature<_Rp (_Gp::*) (_Ap...)> { using type = _Rp(_Ap...); }; -template -struct __strip_signature<_Rp (_Gp::*) (_Ap...) const> { using type = _Rp(_Ap...); }; -template -struct __strip_signature<_Rp (_Gp::*) (_Ap...) volatile> { using type = _Rp(_Ap...); }; -template -struct __strip_signature<_Rp (_Gp::*) (_Ap...) const volatile> { using type = _Rp(_Ap...); }; - -template -struct __strip_signature<_Rp (_Gp::*) (_Ap...) &> { using type = _Rp(_Ap...); }; -template -struct __strip_signature<_Rp (_Gp::*) (_Ap...) const &> { using type = _Rp(_Ap...); }; -template -struct __strip_signature<_Rp (_Gp::*) (_Ap...) volatile &> { using type = _Rp(_Ap...); }; -template -struct __strip_signature<_Rp (_Gp::*) (_Ap...) const volatile &> { using type = _Rp(_Ap...); }; - -template -struct __strip_signature<_Rp (_Gp::*) (_Ap...) noexcept> { using type = _Rp(_Ap...); }; -template -struct __strip_signature<_Rp (_Gp::*) (_Ap...) const noexcept> { using type = _Rp(_Ap...); }; -template -struct __strip_signature<_Rp (_Gp::*) (_Ap...) volatile noexcept> { using type = _Rp(_Ap...); }; -template -struct __strip_signature<_Rp (_Gp::*) (_Ap...) const volatile noexcept> { using type = _Rp(_Ap...); }; - -template -struct __strip_signature<_Rp (_Gp::*) (_Ap...) & noexcept> { using type = _Rp(_Ap...); }; -template -struct __strip_signature<_Rp (_Gp::*) (_Ap...) const & noexcept> { using type = _Rp(_Ap...); }; -template -struct __strip_signature<_Rp (_Gp::*) (_Ap...) volatile & noexcept> { using type = _Rp(_Ap...); }; -template -struct __strip_signature<_Rp (_Gp::*) (_Ap...) const volatile & noexcept> { using type = _Rp(_Ap...); }; - -template::type> -function(_Fp) -> function<_Stripped>; -#endif // !_LIBCPP_HAS_NO_DEDUCTION_GUIDES - -template -function<_Rp(_ArgTypes...)>::function(const function& __f) : __f_(__f.__f_) {} - -#if _LIBCPP_STD_VER <= 14 -template -template -function<_Rp(_ArgTypes...)>::function(allocator_arg_t, const _Alloc&, - const function& __f) : __f_(__f.__f_) {} -#endif - -template -function<_Rp(_ArgTypes...)>::function(function&& __f) _NOEXCEPT - : __f_(_VSTD::move(__f.__f_)) {} - -#if _LIBCPP_STD_VER <= 14 -template -template -function<_Rp(_ArgTypes...)>::function(allocator_arg_t, const _Alloc&, - function&& __f) - : __f_(_VSTD::move(__f.__f_)) {} -#endif - -template -template -function<_Rp(_ArgTypes...)>::function(_Fp __f) : __f_(_VSTD::move(__f)) {} - -#if _LIBCPP_STD_VER <= 14 -template -template -function<_Rp(_ArgTypes...)>::function(allocator_arg_t, const _Alloc& __a, - _Fp __f) - : __f_(_VSTD::move(__f), __a) {} -#endif - -template -function<_Rp(_ArgTypes...)>& -function<_Rp(_ArgTypes...)>::operator=(const function& __f) -{ - function(__f).swap(*this); - return *this; -} - -template -function<_Rp(_ArgTypes...)>& -function<_Rp(_ArgTypes...)>::operator=(function&& __f) _NOEXCEPT -{ - __f_ = _VSTD::move(__f.__f_); - return *this; -} - -template -function<_Rp(_ArgTypes...)>& -function<_Rp(_ArgTypes...)>::operator=(nullptr_t) _NOEXCEPT -{ - __f_ = nullptr; - return *this; -} - -template -template -function<_Rp(_ArgTypes...)>& -function<_Rp(_ArgTypes...)>::operator=(_Fp&& __f) -{ - function(_VSTD::forward<_Fp>(__f)).swap(*this); - return *this; -} - -template -function<_Rp(_ArgTypes...)>::~function() {} - -template -void -function<_Rp(_ArgTypes...)>::swap(function& __f) _NOEXCEPT -{ - __f_.swap(__f.__f_); -} - -template -_Rp -function<_Rp(_ArgTypes...)>::operator()(_ArgTypes... __arg) const -{ - return __f_(_VSTD::forward<_ArgTypes>(__arg)...); -} - -#ifndef _LIBCPP_NO_RTTI - -template -const std::type_info& -function<_Rp(_ArgTypes...)>::target_type() const _NOEXCEPT -{ - return __f_.target_type(); -} - -template -template -_Tp* -function<_Rp(_ArgTypes...)>::target() _NOEXCEPT -{ - return (_Tp*)(__f_.template target<_Tp>()); -} - -template -template -const _Tp* -function<_Rp(_ArgTypes...)>::target() const _NOEXCEPT -{ - return __f_.template target<_Tp>(); -} - -#endif // _LIBCPP_NO_RTTI - -template -inline _LIBCPP_INLINE_VISIBILITY -bool -operator==(const function<_Rp(_ArgTypes...)>& __f, nullptr_t) _NOEXCEPT {return !__f;} - -template -inline _LIBCPP_INLINE_VISIBILITY -bool -operator==(nullptr_t, const function<_Rp(_ArgTypes...)>& __f) _NOEXCEPT {return !__f;} - -template -inline _LIBCPP_INLINE_VISIBILITY -bool -operator!=(const function<_Rp(_ArgTypes...)>& __f, nullptr_t) _NOEXCEPT {return (bool)__f;} - -template -inline _LIBCPP_INLINE_VISIBILITY -bool -operator!=(nullptr_t, const function<_Rp(_ArgTypes...)>& __f) _NOEXCEPT {return (bool)__f;} - -template -inline _LIBCPP_INLINE_VISIBILITY -void -swap(function<_Rp(_ArgTypes...)>& __x, function<_Rp(_ArgTypes...)>& __y) _NOEXCEPT -{return __x.swap(__y);} - -#else // _LIBCPP_CXX03_LANG - -#include <__functional_03> - -#endif - -//////////////////////////////////////////////////////////////////////////////// -// BIND -//============================================================================== - -template struct __is_bind_expression : public false_type {}; -template struct _LIBCPP_TEMPLATE_VIS is_bind_expression - : public __is_bind_expression::type> {}; - -#if _LIBCPP_STD_VER > 14 -template -_LIBCPP_INLINE_VAR constexpr size_t is_bind_expression_v = is_bind_expression<_Tp>::value; -#endif - -template struct __is_placeholder : public integral_constant {}; -template struct _LIBCPP_TEMPLATE_VIS is_placeholder - : public __is_placeholder::type> {}; - -#if _LIBCPP_STD_VER > 14 -template -_LIBCPP_INLINE_VAR constexpr size_t is_placeholder_v = is_placeholder<_Tp>::value; -#endif - -namespace placeholders -{ - -template struct __ph {}; - -#if defined(_LIBCPP_CXX03_LANG) || defined(_LIBCPP_BUILDING_LIBRARY) -_LIBCPP_FUNC_VIS extern const __ph<1> _1; -_LIBCPP_FUNC_VIS extern const __ph<2> _2; -_LIBCPP_FUNC_VIS extern const __ph<3> _3; -_LIBCPP_FUNC_VIS extern const __ph<4> _4; -_LIBCPP_FUNC_VIS extern const __ph<5> _5; -_LIBCPP_FUNC_VIS extern const __ph<6> _6; -_LIBCPP_FUNC_VIS extern const __ph<7> _7; -_LIBCPP_FUNC_VIS extern const __ph<8> _8; -_LIBCPP_FUNC_VIS extern const __ph<9> _9; -_LIBCPP_FUNC_VIS extern const __ph<10> _10; -#else -/* _LIBCPP_INLINE_VAR */ constexpr __ph<1> _1{}; -/* _LIBCPP_INLINE_VAR */ constexpr __ph<2> _2{}; -/* _LIBCPP_INLINE_VAR */ constexpr __ph<3> _3{}; -/* _LIBCPP_INLINE_VAR */ constexpr __ph<4> _4{}; -/* _LIBCPP_INLINE_VAR */ constexpr __ph<5> _5{}; -/* _LIBCPP_INLINE_VAR */ constexpr __ph<6> _6{}; -/* _LIBCPP_INLINE_VAR */ constexpr __ph<7> _7{}; -/* _LIBCPP_INLINE_VAR */ constexpr __ph<8> _8{}; -/* _LIBCPP_INLINE_VAR */ constexpr __ph<9> _9{}; -/* _LIBCPP_INLINE_VAR */ constexpr __ph<10> _10{}; -#endif // defined(_LIBCPP_CXX03_LANG) || defined(_LIBCPP_BUILDING_LIBRARY) - -} // placeholders - -template -struct __is_placeholder > - : public integral_constant {}; - - -#ifndef _LIBCPP_CXX03_LANG - -template -inline _LIBCPP_INLINE_VISIBILITY -_Tp& -__mu(reference_wrapper<_Tp> __t, _Uj&) -{ - return __t.get(); -} - -template -inline _LIBCPP_INLINE_VISIBILITY -typename __invoke_of<_Ti&, _Uj...>::type -__mu_expand(_Ti& __ti, tuple<_Uj...>& __uj, __tuple_indices<_Indx...>) -{ - return __ti(_VSTD::forward<_Uj>(_VSTD::get<_Indx>(__uj))...); -} - -template -inline _LIBCPP_INLINE_VISIBILITY -typename _EnableIf -< - is_bind_expression<_Ti>::value, - __invoke_of<_Ti&, _Uj...> ->::type -__mu(_Ti& __ti, tuple<_Uj...>& __uj) -{ - typedef typename __make_tuple_indices::type __indices; - return _VSTD::__mu_expand(__ti, __uj, __indices()); -} - -template -struct __mu_return2 {}; - -template -struct __mu_return2 -{ - typedef typename tuple_element::value - 1, _Uj>::type type; -}; - -template -inline _LIBCPP_INLINE_VISIBILITY -typename enable_if -< - 0 < is_placeholder<_Ti>::value, - typename __mu_return2<0 < is_placeholder<_Ti>::value, _Ti, _Uj>::type ->::type -__mu(_Ti&, _Uj& __uj) -{ - const size_t _Indx = is_placeholder<_Ti>::value - 1; - return _VSTD::forward::type>(_VSTD::get<_Indx>(__uj)); -} - -template -inline _LIBCPP_INLINE_VISIBILITY -typename enable_if -< - !is_bind_expression<_Ti>::value && - is_placeholder<_Ti>::value == 0 && - !__is_reference_wrapper<_Ti>::value, - _Ti& ->::type -__mu(_Ti& __ti, _Uj&) -{ - return __ti; -} - -template -struct __mu_return_impl; - -template -struct __mu_return_invokable // false -{ - typedef __nat type; -}; - -template -struct __mu_return_invokable -{ - typedef typename __invoke_of<_Ti&, _Uj...>::type type; -}; - -template -struct __mu_return_impl<_Ti, false, true, false, tuple<_Uj...> > - : public __mu_return_invokable<__invokable<_Ti&, _Uj...>::value, _Ti, _Uj...> -{ -}; - -template -struct __mu_return_impl<_Ti, false, false, true, _TupleUj> -{ - typedef typename tuple_element::value - 1, - _TupleUj>::type&& type; -}; - -template -struct __mu_return_impl<_Ti, true, false, false, _TupleUj> -{ - typedef typename _Ti::type& type; -}; - -template -struct __mu_return_impl<_Ti, false, false, false, _TupleUj> -{ - typedef _Ti& type; -}; - -template -struct __mu_return - : public __mu_return_impl<_Ti, - __is_reference_wrapper<_Ti>::value, - is_bind_expression<_Ti>::value, - 0 < is_placeholder<_Ti>::value && - is_placeholder<_Ti>::value <= tuple_size<_TupleUj>::value, - _TupleUj> -{ -}; - -template -struct __is_valid_bind_return -{ - static const bool value = false; -}; - -template -struct __is_valid_bind_return<_Fp, tuple<_BoundArgs...>, _TupleUj> -{ - static const bool value = __invokable<_Fp, - typename __mu_return<_BoundArgs, _TupleUj>::type...>::value; -}; - -template -struct __is_valid_bind_return<_Fp, const tuple<_BoundArgs...>, _TupleUj> -{ - static const bool value = __invokable<_Fp, - typename __mu_return::type...>::value; -}; - -template ::value> -struct __bind_return; - -template -struct __bind_return<_Fp, tuple<_BoundArgs...>, _TupleUj, true> -{ - typedef typename __invoke_of - < - _Fp&, - typename __mu_return - < - _BoundArgs, - _TupleUj - >::type... - >::type type; -}; - -template -struct __bind_return<_Fp, const tuple<_BoundArgs...>, _TupleUj, true> -{ - typedef typename __invoke_of - < - _Fp&, - typename __mu_return - < - const _BoundArgs, - _TupleUj - >::type... - >::type type; -}; - -template -inline _LIBCPP_INLINE_VISIBILITY -typename __bind_return<_Fp, _BoundArgs, _Args>::type -__apply_functor(_Fp& __f, _BoundArgs& __bound_args, __tuple_indices<_Indx...>, - _Args&& __args) -{ - return _VSTD::__invoke(__f, _VSTD::__mu(_VSTD::get<_Indx>(__bound_args), __args)...); -} - -template -class __bind - : public __weak_result_type::type> -{ -protected: - typedef typename decay<_Fp>::type _Fd; - typedef tuple::type...> _Td; -private: - _Fd __f_; - _Td __bound_args_; - - typedef typename __make_tuple_indices::type __indices; -public: - template ::value && - !is_same::type, - __bind>::value - >::type> - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 - explicit __bind(_Gp&& __f, _BA&& ...__bound_args) - : __f_(_VSTD::forward<_Gp>(__f)), - __bound_args_(_VSTD::forward<_BA>(__bound_args)...) {} - - template - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 - typename __bind_return<_Fd, _Td, tuple<_Args&&...> >::type - operator()(_Args&& ...__args) - { - return _VSTD::__apply_functor(__f_, __bound_args_, __indices(), - tuple<_Args&&...>(_VSTD::forward<_Args>(__args)...)); - } - - template - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 - typename __bind_return >::type - operator()(_Args&& ...__args) const - { - return _VSTD::__apply_functor(__f_, __bound_args_, __indices(), - tuple<_Args&&...>(_VSTD::forward<_Args>(__args)...)); - } -}; - -template -struct __is_bind_expression<__bind<_Fp, _BoundArgs...> > : public true_type {}; - -template -class __bind_r - : public __bind<_Fp, _BoundArgs...> -{ - typedef __bind<_Fp, _BoundArgs...> base; - typedef typename base::_Fd _Fd; - typedef typename base::_Td _Td; -public: - typedef _Rp result_type; - - - template ::value && - !is_same::type, - __bind_r>::value - >::type> - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 - explicit __bind_r(_Gp&& __f, _BA&& ...__bound_args) - : base(_VSTD::forward<_Gp>(__f), - _VSTD::forward<_BA>(__bound_args)...) {} - - template - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 - typename enable_if - < - is_convertible >::type, - result_type>::value || is_void<_Rp>::value, - result_type - >::type - operator()(_Args&& ...__args) - { - typedef __invoke_void_return_wrapper<_Rp> _Invoker; - return _Invoker::__call(static_cast(*this), _VSTD::forward<_Args>(__args)...); - } - - template - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 - typename enable_if - < - is_convertible >::type, - result_type>::value || is_void<_Rp>::value, - result_type - >::type - operator()(_Args&& ...__args) const - { - typedef __invoke_void_return_wrapper<_Rp> _Invoker; - return _Invoker::__call(static_cast(*this), _VSTD::forward<_Args>(__args)...); - } -}; - -template -struct __is_bind_expression<__bind_r<_Rp, _Fp, _BoundArgs...> > : public true_type {}; - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -__bind<_Fp, _BoundArgs...> -bind(_Fp&& __f, _BoundArgs&&... __bound_args) -{ - typedef __bind<_Fp, _BoundArgs...> type; - return type(_VSTD::forward<_Fp>(__f), _VSTD::forward<_BoundArgs>(__bound_args)...); -} - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -__bind_r<_Rp, _Fp, _BoundArgs...> -bind(_Fp&& __f, _BoundArgs&&... __bound_args) -{ - typedef __bind_r<_Rp, _Fp, _BoundArgs...> type; - return type(_VSTD::forward<_Fp>(__f), _VSTD::forward<_BoundArgs>(__bound_args)...); -} - -#endif // _LIBCPP_CXX03_LANG - -#if _LIBCPP_STD_VER > 14 - -template -_LIBCPP_CONSTEXPR_AFTER_CXX17 invoke_result_t<_Fn, _Args...> -invoke(_Fn&& __f, _Args&&... __args) - noexcept(is_nothrow_invocable_v<_Fn, _Args...>) -{ - return _VSTD::__invoke(_VSTD::forward<_Fn>(__f), _VSTD::forward<_Args>(__args)...); -} - -template -class _LIBCPP_TEMPLATE_VIS __not_fn_imp { - _DecayFunc __fd; - -public: - __not_fn_imp() = delete; - - template - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 - auto operator()(_Args&& ...__args) & - noexcept(noexcept(!_VSTD::invoke(__fd, _VSTD::forward<_Args>(__args)...))) - -> decltype( !_VSTD::invoke(__fd, _VSTD::forward<_Args>(__args)...)) - { return !_VSTD::invoke(__fd, _VSTD::forward<_Args>(__args)...); } - - template - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 - auto operator()(_Args&& ...__args) && - noexcept(noexcept(!_VSTD::invoke(_VSTD::move(__fd), _VSTD::forward<_Args>(__args)...))) - -> decltype( !_VSTD::invoke(_VSTD::move(__fd), _VSTD::forward<_Args>(__args)...)) - { return !_VSTD::invoke(_VSTD::move(__fd), _VSTD::forward<_Args>(__args)...); } - - template - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 - auto operator()(_Args&& ...__args) const& - noexcept(noexcept(!_VSTD::invoke(__fd, _VSTD::forward<_Args>(__args)...))) - -> decltype( !_VSTD::invoke(__fd, _VSTD::forward<_Args>(__args)...)) - { return !_VSTD::invoke(__fd, _VSTD::forward<_Args>(__args)...); } - - - template - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 - auto operator()(_Args&& ...__args) const&& - noexcept(noexcept(!_VSTD::invoke(_VSTD::move(__fd), _VSTD::forward<_Args>(__args)...))) - -> decltype( !_VSTD::invoke(_VSTD::move(__fd), _VSTD::forward<_Args>(__args)...)) - { return !_VSTD::invoke(_VSTD::move(__fd), _VSTD::forward<_Args>(__args)...); } - -private: - template , __not_fn_imp>::value>> - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 - explicit __not_fn_imp(_RawFunc&& __rf) - : __fd(_VSTD::forward<_RawFunc>(__rf)) {} - - template - friend inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 - __not_fn_imp> not_fn(_RawFunc&&); -}; - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -__not_fn_imp> not_fn(_RawFunc&& __fn) { - return __not_fn_imp>(_VSTD::forward<_RawFunc>(__fn)); -} - -#endif - -// struct hash in - -template -pair<_ForwardIterator1, _ForwardIterator1> _LIBCPP_CONSTEXPR_AFTER_CXX11 -__search(_ForwardIterator1 __first1, _ForwardIterator1 __last1, - _ForwardIterator2 __first2, _ForwardIterator2 __last2, _BinaryPredicate __pred, - forward_iterator_tag, forward_iterator_tag) -{ - if (__first2 == __last2) - return _VSTD::make_pair(__first1, __first1); // Everything matches an empty sequence - while (true) - { - // Find first element in sequence 1 that matchs *__first2, with a mininum of loop checks - while (true) - { - if (__first1 == __last1) // return __last1 if no element matches *__first2 - return _VSTD::make_pair(__last1, __last1); - if (__pred(*__first1, *__first2)) - break; - ++__first1; - } - // *__first1 matches *__first2, now match elements after here - _ForwardIterator1 __m1 = __first1; - _ForwardIterator2 __m2 = __first2; - while (true) - { - if (++__m2 == __last2) // If pattern exhausted, __first1 is the answer (works for 1 element pattern) - return _VSTD::make_pair(__first1, __m1); - if (++__m1 == __last1) // Otherwise if source exhaused, pattern not found - return _VSTD::make_pair(__last1, __last1); - if (!__pred(*__m1, *__m2)) // if there is a mismatch, restart with a new __first1 - { - ++__first1; - break; - } // else there is a match, check next elements - } - } -} - -template -_LIBCPP_CONSTEXPR_AFTER_CXX11 -pair<_RandomAccessIterator1, _RandomAccessIterator1> -__search(_RandomAccessIterator1 __first1, _RandomAccessIterator1 __last1, - _RandomAccessIterator2 __first2, _RandomAccessIterator2 __last2, _BinaryPredicate __pred, - random_access_iterator_tag, random_access_iterator_tag) -{ - typedef typename iterator_traits<_RandomAccessIterator1>::difference_type _D1; - typedef typename iterator_traits<_RandomAccessIterator2>::difference_type _D2; - // Take advantage of knowing source and pattern lengths. Stop short when source is smaller than pattern - const _D2 __len2 = __last2 - __first2; - if (__len2 == 0) - return _VSTD::make_pair(__first1, __first1); - const _D1 __len1 = __last1 - __first1; - if (__len1 < __len2) - return _VSTD::make_pair(__last1, __last1); - const _RandomAccessIterator1 __s = __last1 - (__len2 - 1); // Start of pattern match can't go beyond here - - while (true) - { - while (true) - { - if (__first1 == __s) - return _VSTD::make_pair(__last1, __last1); - if (__pred(*__first1, *__first2)) - break; - ++__first1; - } - - _RandomAccessIterator1 __m1 = __first1; - _RandomAccessIterator2 __m2 = __first2; - while (true) - { - if (++__m2 == __last2) - return _VSTD::make_pair(__first1, __first1 + __len2); - ++__m1; // no need to check range on __m1 because __s guarantees we have enough source - if (!__pred(*__m1, *__m2)) - { - ++__first1; - break; - } - } - } -} - -#if _LIBCPP_STD_VER > 14 - -// default searcher -template> -class _LIBCPP_TYPE_VIS default_searcher { -public: - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 - default_searcher(_ForwardIterator __f, _ForwardIterator __l, - _BinaryPredicate __p = _BinaryPredicate()) - : __first_(__f), __last_(__l), __pred_(__p) {} - - template - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 - pair<_ForwardIterator2, _ForwardIterator2> - operator () (_ForwardIterator2 __f, _ForwardIterator2 __l) const - { - return _VSTD::__search(__f, __l, __first_, __last_, __pred_, - typename _VSTD::iterator_traits<_ForwardIterator>::iterator_category(), - typename _VSTD::iterator_traits<_ForwardIterator2>::iterator_category()); - } - -private: - _ForwardIterator __first_; - _ForwardIterator __last_; - _BinaryPredicate __pred_; - }; - -#endif // _LIBCPP_STD_VER > 14 - -#if _LIBCPP_STD_VER > 17 -template -using unwrap_reference_t = typename unwrap_reference<_Tp>::type; - -template -using unwrap_ref_decay_t = typename unwrap_ref_decay<_Tp>::type; -#endif // > C++17 - -template -inline typename _Container::size_type -__libcpp_erase_if_container(_Container& __c, _Predicate __pred) { - typename _Container::size_type __old_size = __c.size(); - - const typename _Container::iterator __last = __c.end(); - for (typename _Container::iterator __iter = __c.begin(); __iter != __last;) { - if (__pred(*__iter)) - __iter = __c.erase(__iter); - else - ++__iter; - } - - return __old_size - __c.size(); -} - -_LIBCPP_END_NAMESPACE_STD - -#endif // _LIBCPP_FUNCTIONAL +#endif // _LIBCPP_FUNCTIONAL diff --git a/lib/libcxx/include/future b/lib/libcxx/include/future index db60ab69e..349e6afc4 100644 --- a/lib/libcxx/include/future +++ b/lib/libcxx/include/future @@ -361,13 +361,18 @@ template struct uses_allocator, Alloc>; */ -#include <__config> #include <__availability> -#include -#include +#include <__config> +#include <__debug> +#include <__memory/allocator_arg_t.h> +#include <__memory/uses_allocator.h> +#include <__utility/__decay_copy.h> +#include <__utility/forward.h> #include #include +#include #include +#include #include #if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) @@ -467,7 +472,7 @@ operator^=(launch& __x, launch __y) __x = __x ^ __y; return __x; } -#endif // !_LIBCPP_HAS_NO_STRONG_ENUMS +#endif // !_LIBCPP_HAS_NO_STRONG_ENUMS //enum class future_status _LIBCPP_DECLARE_STRONG_ENUM(future_status) @@ -501,9 +506,7 @@ class _LIBCPP_EXCEPTION_ABI _LIBCPP_AVAILABILITY_FUTURE_ERROR future_error error_code __ec_; public: future_error(error_code __ec); -#if _LIBCPP_STD_VERS > 14 - explicit future_error(future_errc _Ev) : logic_error(), __ec_(make_error_code(_Ev)) {} -#endif + _LIBCPP_INLINE_VISIBILITY const error_code& code() const _NOEXCEPT {return __ec_;} @@ -862,7 +865,7 @@ __deferred_assoc_state<_Rp, _Fp>::__execute() #ifndef _LIBCPP_NO_EXCEPTIONS try { -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS this->set_value(__func_()); #ifndef _LIBCPP_NO_EXCEPTIONS } @@ -870,7 +873,7 @@ __deferred_assoc_state<_Rp, _Fp>::__execute() { this->set_exception(current_exception()); } -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS } template @@ -903,7 +906,7 @@ __deferred_assoc_state::__execute() #ifndef _LIBCPP_NO_EXCEPTIONS try { -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS __func_(); this->set_value(); #ifndef _LIBCPP_NO_EXCEPTIONS @@ -912,7 +915,7 @@ __deferred_assoc_state::__execute() { this->set_exception(current_exception()); } -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS } template @@ -945,7 +948,7 @@ __async_assoc_state<_Rp, _Fp>::__execute() #ifndef _LIBCPP_NO_EXCEPTIONS try { -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS this->set_value(__func_()); #ifndef _LIBCPP_NO_EXCEPTIONS } @@ -953,7 +956,7 @@ __async_assoc_state<_Rp, _Fp>::__execute() { this->set_exception(current_exception()); } -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS } template @@ -994,7 +997,7 @@ __async_assoc_state::__execute() #ifndef _LIBCPP_NO_EXCEPTIONS try { -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS __func_(); this->set_value(); #ifndef _LIBCPP_NO_EXCEPTIONS @@ -1003,7 +1006,7 @@ __async_assoc_state::__execute() { this->set_exception(current_exception()); } -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS } template @@ -1953,7 +1956,7 @@ packaged_task<_Rp(_ArgTypes...)>::operator()(_ArgTypes... __args) #ifndef _LIBCPP_NO_EXCEPTIONS try { -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS __p_.set_value(__f_(_VSTD::forward<_ArgTypes>(__args)...)); #ifndef _LIBCPP_NO_EXCEPTIONS } @@ -1961,7 +1964,7 @@ packaged_task<_Rp(_ArgTypes...)>::operator()(_ArgTypes... __args) { __p_.set_exception(current_exception()); } -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS } template @@ -1975,7 +1978,7 @@ packaged_task<_Rp(_ArgTypes...)>::make_ready_at_thread_exit(_ArgTypes... __args) #ifndef _LIBCPP_NO_EXCEPTIONS try { -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS __p_.set_value_at_thread_exit(__f_(_VSTD::forward<_ArgTypes>(__args)...)); #ifndef _LIBCPP_NO_EXCEPTIONS } @@ -1983,7 +1986,7 @@ packaged_task<_Rp(_ArgTypes...)>::make_ready_at_thread_exit(_ArgTypes... __args) { __p_.set_exception_at_thread_exit(current_exception()); } -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS } template @@ -2082,7 +2085,7 @@ packaged_task::operator()(_ArgTypes... __args) #ifndef _LIBCPP_NO_EXCEPTIONS try { -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS __f_(_VSTD::forward<_ArgTypes>(__args)...); __p_.set_value(); #ifndef _LIBCPP_NO_EXCEPTIONS @@ -2091,7 +2094,7 @@ packaged_task::operator()(_ArgTypes... __args) { __p_.set_exception(current_exception()); } -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS } template @@ -2105,7 +2108,7 @@ packaged_task::make_ready_at_thread_exit(_ArgTypes... __args #ifndef _LIBCPP_NO_EXCEPTIONS try { -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS __f_(_VSTD::forward<_ArgTypes>(__args)...); __p_.set_value_at_thread_exit(); #ifndef _LIBCPP_NO_EXCEPTIONS @@ -2114,7 +2117,7 @@ packaged_task::make_ready_at_thread_exit(_ArgTypes... __args { __p_.set_exception_at_thread_exit(current_exception()); } -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS } template @@ -2126,10 +2129,10 @@ packaged_task::reset() __p_ = promise(); } -template +template inline _LIBCPP_INLINE_VISIBILITY void -swap(packaged_task<_Callable>& __x, packaged_task<_Callable>& __y) _NOEXCEPT +swap(packaged_task<_Rp(_ArgTypes...)>& __x, packaged_task<_Rp(_ArgTypes...)>& __y) _NOEXCEPT { __x.swap(__y); } @@ -2456,4 +2459,4 @@ _LIBCPP_END_NAMESPACE_STD #endif // !_LIBCPP_HAS_NO_THREADS -#endif // _LIBCPP_FUTURE +#endif // _LIBCPP_FUTURE diff --git a/lib/libcxx/include/initializer_list b/lib/libcxx/include/initializer_list index 893736f57..ea1f23467 100644 --- a/lib/libcxx/include/initializer_list +++ b/lib/libcxx/include/initializer_list @@ -110,8 +110,8 @@ end(initializer_list<_Ep> __il) _NOEXCEPT return __il.end(); } -#endif // !defined(_LIBCPP_CXX03_LANG) +#endif // !defined(_LIBCPP_CXX03_LANG) } // std -#endif // _LIBCPP_INITIALIZER_LIST +#endif // _LIBCPP_INITIALIZER_LIST diff --git a/lib/libcxx/include/inttypes.h b/lib/libcxx/include/inttypes.h index e136b234a..6b7cb2648 100644 --- a/lib/libcxx/include/inttypes.h +++ b/lib/libcxx/include/inttypes.h @@ -259,4 +259,4 @@ uintmax_t wcstoumax(const wchar_t* restrict nptr, wchar_t** restrict endptr, int #endif // __cplusplus -#endif // _LIBCPP_INTTYPES_H +#endif // _LIBCPP_INTTYPES_H diff --git a/lib/libcxx/include/iomanip b/lib/libcxx/include/iomanip index 4b6418bc0..47d573b5f 100644 --- a/lib/libcxx/include/iomanip +++ b/lib/libcxx/include/iomanip @@ -304,7 +304,7 @@ operator>>(basic_istream<_CharT, _Traits>& __is, const __iom_t7<_MoneyT>& __x) #ifndef _LIBCPP_NO_EXCEPTIONS try { -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS typename basic_istream<_CharT, _Traits>::sentry __s(__is); if (__s) { @@ -321,7 +321,7 @@ operator>>(basic_istream<_CharT, _Traits>& __is, const __iom_t7<_MoneyT>& __x) { __is.__set_badbit_and_consider_rethrow(); } -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS return __is; } @@ -364,7 +364,7 @@ operator<<(basic_ostream<_CharT, _Traits>& __os, const __iom_t8<_MoneyT>& __x) #ifndef _LIBCPP_NO_EXCEPTIONS try { -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS typename basic_ostream<_CharT, _Traits>::sentry __s(__os); if (__s) { @@ -380,7 +380,7 @@ operator<<(basic_ostream<_CharT, _Traits>& __os, const __iom_t8<_MoneyT>& __x) { __os.__set_badbit_and_consider_rethrow(); } -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS return __os; } @@ -423,7 +423,7 @@ operator>>(basic_istream<_CharT, _Traits>& __is, const __iom_t9<_CharT>& __x) #ifndef _LIBCPP_NO_EXCEPTIONS try { -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS typename basic_istream<_CharT, _Traits>::sentry __s(__is); if (__s) { @@ -441,7 +441,7 @@ operator>>(basic_istream<_CharT, _Traits>& __is, const __iom_t9<_CharT>& __x) { __is.__set_badbit_and_consider_rethrow(); } -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS return __is; } @@ -484,7 +484,7 @@ operator<<(basic_ostream<_CharT, _Traits>& __os, const __iom_t10<_CharT>& __x) #ifndef _LIBCPP_NO_EXCEPTIONS try { -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS typename basic_ostream<_CharT, _Traits>::sentry __s(__os); if (__s) { @@ -501,7 +501,7 @@ operator<<(basic_ostream<_CharT, _Traits>& __os, const __iom_t10<_CharT>& __x) { __os.__set_badbit_and_consider_rethrow(); } -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS return __os; } @@ -518,7 +518,7 @@ basic_ostream<_CharT, _Traits> & __quoted_output ( basic_ostream<_CharT, _Traits> &__os, _ForwardIterator __first, _ForwardIterator __last, _CharT __delim, _CharT __escape ) { - _VSTD::basic_string<_CharT, _Traits> __str; + basic_string<_CharT, _Traits> __str; __str.push_back(__delim); for ( ; __first != __last; ++ __first ) { @@ -667,4 +667,4 @@ quoted (basic_string_view <_CharT, _Traits> __sv, _LIBCPP_END_NAMESPACE_STD -#endif // _LIBCPP_IOMANIP +#endif // _LIBCPP_IOMANIP diff --git a/lib/libcxx/include/ios b/lib/libcxx/include/ios index 653e3a95f..c9230d6a9 100644 --- a/lib/libcxx/include/ios +++ b/lib/libcxx/include/ios @@ -211,8 +211,8 @@ storage-class-specifier const error_category& iostream_category() noexcept; */ #include <__config> -#include #include <__locale> +#include #include #if !defined(_LIBCPP_HAS_NO_ATOMIC_HEADER) @@ -591,13 +591,6 @@ ios_base::exceptions(iostate __iostate) clear(__rdstate_); } -#if defined(_LIBCPP_CXX03_LANG) -struct _LIBCPP_TYPE_VIS __cxx03_bool { - typedef void (__cxx03_bool::*__bool_type)(); - void __true_value() {} -}; -#endif - template class _LIBCPP_TEMPLATE_VIS basic_ios : public ios_base @@ -614,17 +607,14 @@ public: static_assert((is_same<_CharT, typename traits_type::char_type>::value), "traits_type::char_type must be the same type as CharT"); - // __true_value will generate undefined references when linking unless - // we give it internal linkage. - -#if defined(_LIBCPP_CXX03_LANG) +#ifdef _LIBCPP_CXX03_LANG + // Preserve the ability to compare with literal 0, + // and implicitly convert to bool, but not implicitly convert to int. _LIBCPP_INLINE_VISIBILITY - operator __cxx03_bool::__bool_type() const { - return !fail() ? &__cxx03_bool::__true_value : nullptr; - } + operator void*() const {return fail() ? nullptr : (void*)this;} #else _LIBCPP_INLINE_VISIBILITY - _LIBCPP_EXPLICIT operator bool() const {return !fail();} + explicit operator bool() const {return !fail();} #endif _LIBCPP_INLINE_VISIBILITY bool operator!() const {return fail();} @@ -679,10 +669,8 @@ protected: _LIBCPP_INLINE_VISIBILITY void move(basic_ios& __rhs); -#ifndef _LIBCPP_CXX03_LANG _LIBCPP_INLINE_VISIBILITY void move(basic_ios&& __rhs) {move(__rhs);} -#endif _LIBCPP_INLINE_VISIBILITY void swap(basic_ios& __rhs) _NOEXCEPT; _LIBCPP_INLINE_VISIBILITY @@ -1037,4 +1025,4 @@ defaultfloat(ios_base& __str) _LIBCPP_END_NAMESPACE_STD -#endif // _LIBCPP_IOS +#endif // _LIBCPP_IOS diff --git a/lib/libcxx/include/iosfwd b/lib/libcxx/include/iosfwd index 0a0de99ff..499839f7d 100644 --- a/lib/libcxx/include/iosfwd +++ b/lib/libcxx/include/iosfwd @@ -84,8 +84,11 @@ typedef basic_ofstream wofstream; typedef basic_fstream wfstream; template class fpos; -typedef fpos::state_type> streampos; -typedef fpos::state_type> wstreampos; +using streampos = fpos::state_type>; +using wstreampos = fpos::state_type>; +using u8streampos = fpos::state_type>; // C++20 +using u16streampos = fpos::state_type>; +using u32streampos = fpos::state_type>; } // std @@ -104,7 +107,7 @@ class _LIBCPP_TYPE_VIS ios_base; template struct _LIBCPP_TEMPLATE_VIS char_traits; template<> struct char_traits; -#ifndef _LIBCPP_NO_HAS_CHAR8_T +#ifndef _LIBCPP_HAS_NO_CHAR8_T template<> struct char_traits; #endif template<> struct char_traits; @@ -218,13 +221,13 @@ template template class _LIBCPP_TEMPLATE_VIS fpos; typedef fpos streampos; typedef fpos wstreampos; -#ifndef _LIBCPP_NO_HAS_CHAR8_T +#ifndef _LIBCPP_HAS_NO_CHAR8_T typedef fpos u8streampos; #endif #ifndef _LIBCPP_HAS_NO_UNICODE_CHARS typedef fpos u16streampos; typedef fpos u32streampos; -#endif // _LIBCPP_HAS_NO_UNICODE_CHARS +#endif // _LIBCPP_HAS_NO_UNICODE_CHARS #if defined(_NEWLIB_VERSION) // On newlib, off_t is 'long int' @@ -276,4 +279,4 @@ public: _LIBCPP_END_NAMESPACE_STD -#endif // _LIBCPP_IOSFWD +#endif // _LIBCPP_IOSFWD diff --git a/lib/libcxx/include/iostream b/lib/libcxx/include/iostream index 595620b8a..5811fc37c 100644 --- a/lib/libcxx/include/iostream +++ b/lib/libcxx/include/iostream @@ -14,9 +14,9 @@ iostream synopsis #include -#include #include #include +#include namespace std { @@ -35,9 +35,9 @@ extern wostream wclog; #include <__config> #include -#include #include #include +#include #if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) #pragma GCC system_header @@ -60,4 +60,4 @@ extern _LIBCPP_FUNC_VIS wostream wclog; _LIBCPP_END_NAMESPACE_STD -#endif // _LIBCPP_IOSTREAM +#endif // _LIBCPP_IOSTREAM diff --git a/lib/libcxx/include/istream b/lib/libcxx/include/istream index 5e984909f..17ca68388 100644 --- a/lib/libcxx/include/istream +++ b/lib/libcxx/include/istream @@ -159,8 +159,9 @@ template */ #include <__config> -#include +#include <__utility/forward.h> #include +#include #if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) #pragma GCC system_header @@ -191,14 +192,12 @@ public: { this->init(__sb); } virtual ~basic_istream(); protected: -#ifndef _LIBCPP_CXX03_LANG inline _LIBCPP_INLINE_VISIBILITY basic_istream(basic_istream&& __rhs); // 27.7.1.1.2 Assign/swap: inline _LIBCPP_INLINE_VISIBILITY basic_istream& operator=(basic_istream&& __rhs); -#endif inline _LIBCPP_HIDE_FROM_ABI_AFTER_V1 void swap(basic_istream& __rhs) { @@ -206,10 +205,8 @@ protected: basic_ios::swap(__rhs); } -#ifndef _LIBCPP_CXX03_LANG basic_istream (const basic_istream& __rhs) = delete; basic_istream& operator=(const basic_istream& __rhs) = delete; -#endif public: // 27.7.1.1.3 Prefix/suffix: @@ -302,8 +299,7 @@ public: // ~sentry() = default; _LIBCPP_INLINE_VISIBILITY - _LIBCPP_EXPLICIT - operator bool() const {return __ok_;} + explicit operator bool() const {return __ok_;} }; template @@ -333,8 +329,6 @@ basic_istream<_CharT, _Traits>::sentry::sentry(basic_istream<_CharT, _Traits>& _ __is.setstate(ios_base::failbit); } -#ifndef _LIBCPP_CXX03_LANG - template basic_istream<_CharT, _Traits>::basic_istream(basic_istream&& __rhs) : __gc_(__rhs.__gc_) @@ -351,8 +345,6 @@ basic_istream<_CharT, _Traits>::operator=(basic_istream&& __rhs) return *this; } -#endif // _LIBCPP_CXX03_LANG - template basic_istream<_CharT, _Traits>::~basic_istream() { @@ -369,7 +361,7 @@ __input_arithmetic(basic_istream<_CharT, _Traits>& __is, _Tp& __n) { #ifndef _LIBCPP_NO_EXCEPTIONS try { -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS typedef istreambuf_iterator<_CharT, _Traits> _Ip; typedef num_get<_CharT, _Ip> _Fp; use_facet<_Fp>(__is.getloc()).get(_Ip(__is), _Ip(), __is, __state, __n); @@ -478,7 +470,7 @@ __input_arithmetic_with_numeric_limits(basic_istream<_CharT, _Traits>& __is, _Tp #ifndef _LIBCPP_NO_EXCEPTIONS try { -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS typedef istreambuf_iterator<_CharT, _Traits> _Ip; typedef num_get<_CharT, _Ip> _Fp; long __temp; @@ -508,7 +500,7 @@ __input_arithmetic_with_numeric_limits(basic_istream<_CharT, _Traits>& __is, _Tp throw; } } -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS __is.setstate(__state); } return __is; @@ -636,7 +628,7 @@ operator>>(basic_istream& __is, signed char* __s) return __is >> (char*)__s; } -#endif // _LIBCPP_STD_VER > 17 +#endif // _LIBCPP_STD_VER > 17 template basic_istream<_CharT, _Traits>& @@ -734,7 +726,7 @@ basic_istream<_CharT, _Traits>::operator>>(basic_streambuf::get(basic_streambuf& __s #ifndef _LIBCPP_NO_EXCEPTIONS try { -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS while (true) { typename traits_type::int_type __i = this->rdbuf()->sgetc(); @@ -878,7 +870,7 @@ basic_istream<_CharT, _Traits>::get(basic_streambuf& __s __state |= ios_base::badbit; // according to the spec, exceptions here are caught but not rethrown } -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS if (__gc_ == 0) __state |= ios_base::failbit; this->setstate(__state); @@ -898,7 +890,7 @@ basic_istream<_CharT, _Traits>::getline(char_type* __s, streamsize __n, char_typ #ifndef _LIBCPP_NO_EXCEPTIONS try { -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS while (true) { typename traits_type::int_type __i = this->rdbuf()->sgetc(); @@ -938,7 +930,7 @@ basic_istream<_CharT, _Traits>::getline(char_type* __s, streamsize __n, char_typ throw; } } -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS } if (__n > 0) *__s = char_type(); @@ -960,7 +952,7 @@ basic_istream<_CharT, _Traits>::ignore(streamsize __n, int_type __dlm) #ifndef _LIBCPP_NO_EXCEPTIONS try { -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS if (__n == numeric_limits::max()) { while (true) @@ -1002,7 +994,7 @@ basic_istream<_CharT, _Traits>::ignore(streamsize __n, int_type __dlm) throw; } } -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS this->setstate(__state); } return *this; @@ -1021,7 +1013,7 @@ basic_istream<_CharT, _Traits>::peek() #ifndef _LIBCPP_NO_EXCEPTIONS try { -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS __r = this->rdbuf()->sgetc(); if (traits_type::eq_int_type(__r, traits_type::eof())) __state |= ios_base::eofbit; @@ -1036,7 +1028,7 @@ basic_istream<_CharT, _Traits>::peek() throw; } } -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS this->setstate(__state); } return __r; @@ -1054,7 +1046,7 @@ basic_istream<_CharT, _Traits>::read(char_type* __s, streamsize __n) #ifndef _LIBCPP_NO_EXCEPTIONS try { -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS __gc_ = this->rdbuf()->sgetn(__s, __n); if (__gc_ != __n) __state |= ios_base::failbit | ios_base::eofbit; @@ -1069,7 +1061,7 @@ basic_istream<_CharT, _Traits>::read(char_type* __s, streamsize __n) throw; } } -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS } else { @@ -1091,7 +1083,7 @@ basic_istream<_CharT, _Traits>::readsome(char_type* __s, streamsize __n) #ifndef _LIBCPP_NO_EXCEPTIONS try { -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS streamsize __c = this->rdbuf()->in_avail(); switch (__c) { @@ -1118,7 +1110,7 @@ basic_istream<_CharT, _Traits>::readsome(char_type* __s, streamsize __n) throw; } } -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS } else { @@ -1141,7 +1133,7 @@ basic_istream<_CharT, _Traits>::putback(char_type __c) #ifndef _LIBCPP_NO_EXCEPTIONS try { -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS if (this->rdbuf() == nullptr || this->rdbuf()->sputbackc(__c) == traits_type::eof()) __state |= ios_base::badbit; #ifndef _LIBCPP_NO_EXCEPTIONS @@ -1155,7 +1147,7 @@ basic_istream<_CharT, _Traits>::putback(char_type __c) throw; } } -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS } else { @@ -1178,7 +1170,7 @@ basic_istream<_CharT, _Traits>::unget() #ifndef _LIBCPP_NO_EXCEPTIONS try { -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS if (this->rdbuf() == nullptr || this->rdbuf()->sungetc() == traits_type::eof()) __state |= ios_base::badbit; #ifndef _LIBCPP_NO_EXCEPTIONS @@ -1192,7 +1184,7 @@ basic_istream<_CharT, _Traits>::unget() throw; } } -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS } else { @@ -1214,7 +1206,7 @@ basic_istream<_CharT, _Traits>::sync() #ifndef _LIBCPP_NO_EXCEPTIONS try { -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS if (this->rdbuf() == nullptr) return -1; if (this->rdbuf()->pubsync() == -1) @@ -1233,7 +1225,7 @@ basic_istream<_CharT, _Traits>::sync() throw; } } -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS this->setstate(__state); } return __r; @@ -1251,7 +1243,7 @@ basic_istream<_CharT, _Traits>::tellg() #ifndef _LIBCPP_NO_EXCEPTIONS try { -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS __r = this->rdbuf()->pubseekoff(0, ios_base::cur, ios_base::in); #ifndef _LIBCPP_NO_EXCEPTIONS } @@ -1264,7 +1256,7 @@ basic_istream<_CharT, _Traits>::tellg() throw; } } -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS this->setstate(__state); } return __r; @@ -1282,7 +1274,7 @@ basic_istream<_CharT, _Traits>::seekg(pos_type __pos) #ifndef _LIBCPP_NO_EXCEPTIONS try { -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS if (this->rdbuf()->pubseekpos(__pos, ios_base::in) == pos_type(-1)) __state |= ios_base::failbit; #ifndef _LIBCPP_NO_EXCEPTIONS @@ -1296,7 +1288,7 @@ basic_istream<_CharT, _Traits>::seekg(pos_type __pos) throw; } } -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS this->setstate(__state); } return *this; @@ -1314,7 +1306,7 @@ basic_istream<_CharT, _Traits>::seekg(off_type __off, ios_base::seekdir __dir) #ifndef _LIBCPP_NO_EXCEPTIONS try { -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS if (this->rdbuf()->pubseekoff(__off, __dir, ios_base::in) == pos_type(-1)) __state |= ios_base::failbit; #ifndef _LIBCPP_NO_EXCEPTIONS @@ -1328,7 +1320,7 @@ basic_istream<_CharT, _Traits>::seekg(off_type __off, ios_base::seekdir __dir) throw; } } -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS this->setstate(__state); } return *this; @@ -1345,7 +1337,7 @@ ws(basic_istream<_CharT, _Traits>& __is) #ifndef _LIBCPP_NO_EXCEPTIONS try { -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS const ctype<_CharT>& __ct = use_facet >(__is.getloc()); while (true) { @@ -1370,25 +1362,23 @@ ws(basic_istream<_CharT, _Traits>& __is) throw; } } -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS __is.setstate(__state); } return __is; } -#ifndef _LIBCPP_CXX03_LANG - template struct __is_istreamable : false_type { }; template struct __is_istreamable<_Stream, _Tp, decltype( - _VSTD::declval<_Stream>() >> _VSTD::declval<_Tp>(), void() + declval<_Stream>() >> declval<_Tp>(), void() )> : true_type { }; template , - __is_istreamable<_Stream&, _Tp&&>>::value + __is_istreamable<_Stream&, _Tp&&> >::value >::type> _LIBCPP_INLINE_VISIBILITY _Stream&& operator>>(_Stream&& __is, _Tp&& __x) @@ -1397,8 +1387,6 @@ _Stream&& operator>>(_Stream&& __is, _Tp&& __x) return _VSTD::move(__is); } -#endif // _LIBCPP_CXX03_LANG - template class _LIBCPP_TEMPLATE_VIS basic_iostream : public basic_istream<_CharT, _Traits>, @@ -1420,21 +1408,18 @@ public: virtual ~basic_iostream(); protected: -#ifndef _LIBCPP_CXX03_LANG inline _LIBCPP_INLINE_VISIBILITY basic_iostream(basic_iostream&& __rhs); // assign/swap inline _LIBCPP_INLINE_VISIBILITY basic_iostream& operator=(basic_iostream&& __rhs); -#endif + inline _LIBCPP_HIDE_FROM_ABI_AFTER_V1 void swap(basic_iostream& __rhs) { basic_istream::swap(__rhs); } }; -#ifndef _LIBCPP_CXX03_LANG - template basic_iostream<_CharT, _Traits>::basic_iostream(basic_iostream&& __rhs) : basic_istream<_CharT, _Traits>(_VSTD::move(__rhs)) @@ -1449,8 +1434,6 @@ basic_iostream<_CharT, _Traits>::operator=(basic_iostream&& __rhs) return *this; } -#endif // _LIBCPP_CXX03_LANG - template basic_iostream<_CharT, _Traits>::~basic_iostream() { @@ -1574,8 +1557,6 @@ getline(basic_istream<_CharT, _Traits>& __is, return getline(__is, __str, __is.widen('\n')); } -#ifndef _LIBCPP_CXX03_LANG - template inline _LIBCPP_INLINE_VISIBILITY basic_istream<_CharT, _Traits>& @@ -1594,8 +1575,6 @@ getline(basic_istream<_CharT, _Traits>&& __is, return getline(__is, __str, __is.widen('\n')); } -#endif // _LIBCPP_CXX03_LANG - template basic_istream<_CharT, _Traits>& operator>>(basic_istream<_CharT, _Traits>& __is, bitset<_Size>& __x) @@ -1656,4 +1635,4 @@ _LIBCPP_END_NAMESPACE_STD _LIBCPP_POP_MACROS -#endif // _LIBCPP_ISTREAM +#endif // _LIBCPP_ISTREAM diff --git a/lib/libcxx/include/iterator b/lib/libcxx/include/iterator index 00a3451e7..e0b25200b 100644 --- a/lib/libcxx/include/iterator +++ b/lib/libcxx/include/iterator @@ -13,32 +13,140 @@ /* iterator synopsis +#include + namespace std { +template struct incrementable_traits; // since C++20 +template + using iter_difference_t = see below; // since C++20 + +template struct indirectly_readable_traits; // since C++20 +template + using iter_value_t = see below; // since C++20 template -struct iterator_traits -{ - typedef typename Iterator::difference_type difference_type; - typedef typename Iterator::value_type value_type; - typedef typename Iterator::pointer pointer; - typedef typename Iterator::reference reference; - typedef typename Iterator::iterator_category iterator_category; -}; +struct iterator_traits; template -struct iterator_traits -{ - typedef ptrdiff_t difference_type; - typedef T value_type; - typedef T* pointer; - typedef T& reference; - typedef random_access_iterator_tag iterator_category; -}; + requires is_object_v // since C++20 +struct iterator_traits; + +template + using iter_reference_t = decltype(*declval()); + +namespace ranges::inline unspecified { + inline constexpr unspecified iter_move = unspecified; // since C++20, nodiscard as an extension +}} + +template + requires ... +using iter_rvalue_reference_t = decltype(ranges::iter_move(declval())); // since C++20 + +// [iterator.concepts], iterator concepts +// [iterator.concept.readable], concept indirectly_readable +template + concept indirectly_readable = see below; // since C++20 + +template + using iter_common_reference_t = + common_reference_t, iter_value_t&>; // since C++20 + +// [iterator.concept.writable], concept indirectly_writable +template + concept indirectly_writable = see below; // since C++20 + +// [iterator.concept.winc], concept weakly_incrementable +template + concept weakly_incrementable = see below; // since C++20 + +// [iterator.concept.inc], concept incrementable +template + concept incrementable = see below; // since C++20 + +// [iterator.concept.iterator], concept input_or_output_iterator + template + concept input_or_output_iterator = see below; // since C++20 + +// [iterator.concept.sentinel], concept sentinel_for +template + concept sentinel_for = see below; // since C++20 + +// [iterator.concept.sizedsentinel], concept sized_sentinel_for +template + inline constexpr bool disable_sized_sentinel_for = false; + +template + concept sized_sentinel_for = see below; + +// [iterator.concept.input], concept input_iterator +template + concept input_iterator = see below; // since C++20 + +// [iterator.concept.output], concept output_iterator +template + concept output_iterator = see below; // since C++20 + +// [iterator.concept.forward], concept forward_iterator +template + concept forward_iterator = see below; // since C++20 + +// [iterator.concept.bidir], concept bidirectional_iterator +template + concept bidirectional_iterator = see below; // since C++20 + +// [iterator.concept.random.access], concept random_access_iterator +template + concept random_access_iterator = see below; // since C++20 + +// [indirectcallable] +// [indirectcallable.indirectinvocable] +template + concept indirectly_unary_invocable = see below; // since C++20 + +template + concept indirectly_regular_unary_invocable = see below; // since C++20 + +template + concept indirect_unary_predicate = see below; // since C++20 + +template + concept indirect_binary_predicate = see below; // since C++20 + +template + concept indirect_equivalence_relation = see below; // since C++20 + +template + concept indirect_strict_weak_order = see below; // since C++20 + +template + using indirect_result_t = see below; // since C++20 + +// [projected], projected +template Proj> + struct projected; // since C++20 + +template Proj> + struct incrementable_traits>; // since C++20 + +// [alg.req.ind.move], concept indirectly_movable +template + concept indirectly_movable = see below; // since C++20 + +template + concept indirectly_movable_storable = see below; // since C++20 + +// [alg.req.ind.swap], concept indirectly_swappable +template + concept indirectly_swappable = see below; // since C++20 + +template S> + requires (!same_as && copyable) +class common_iterator; // since C++20 template -struct iterator +struct iterator // deprecated in C++17 { typedef T value_type; typedef Distance difference_type; @@ -69,9 +177,20 @@ template // constexpr in C++17 constexpr BidirectionalIterator prev(BidirectionalIterator x, typename iterator_traits::difference_type n = 1); +// [range.iter.ops], range iterator operations +namespace ranges { + // [range.iter.op.advance], ranges::advance + template + constexpr void advance(I& i, iter_difference_t n); // since C++20 + template S> + constexpr void advance(I& i, S bound); // since C++20 + template S> + constexpr iter_difference_t advance(I& i, iter_difference_t n, S bound); // since C++20 +} + template class reverse_iterator - : public iterator::iterator_category, + : public iterator::iterator_category, // until C++17 typename iterator_traits::value_type, typename iterator_traits::difference_type, typename iterator_traits::pointer, @@ -142,48 +261,53 @@ constexpr reverse_iterator make_reverse_iterator(Iterator i); // C++14 template class back_insert_iterator + : public iterator // until C++17 { protected: Container* container; public: typedef Container container_type; typedef void value_type; - typedef void difference_type; + typedef void difference_type; // until C++20 + typedef ptrdiff_t difference_type; // since C++20 typedef void reference; typedef void pointer; - explicit back_insert_iterator(Container& x); - back_insert_iterator& operator=(const typename Container::value_type& value); - back_insert_iterator& operator*(); - back_insert_iterator& operator++(); - back_insert_iterator operator++(int); + explicit back_insert_iterator(Container& x); // constexpr in C++20 + back_insert_iterator& operator=(const typename Container::value_type& value); // constexpr in C++20 + back_insert_iterator& operator*(); // constexpr in C++20 + back_insert_iterator& operator++(); // constexpr in C++20 + back_insert_iterator operator++(int); // constexpr in C++20 }; -template back_insert_iterator back_inserter(Container& x); +template back_insert_iterator back_inserter(Container& x); // constexpr in C++20 template class front_insert_iterator + : public iterator // until C++17 { protected: Container* container; public: typedef Container container_type; typedef void value_type; - typedef void difference_type; + typedef void difference_type; // until C++20 + typedef ptrdiff_t difference_type; // since C++20 typedef void reference; typedef void pointer; - explicit front_insert_iterator(Container& x); - front_insert_iterator& operator=(const typename Container::value_type& value); - front_insert_iterator& operator*(); - front_insert_iterator& operator++(); - front_insert_iterator operator++(int); + explicit front_insert_iterator(Container& x); // constexpr in C++20 + front_insert_iterator& operator=(const typename Container::value_type& value); // constexpr in C++20 + front_insert_iterator& operator*(); // constexpr in C++20 + front_insert_iterator& operator++(); // constexpr in C++20 + front_insert_iterator operator++(int); // constexpr in C++20 }; -template front_insert_iterator front_inserter(Container& x); +template front_insert_iterator front_inserter(Container& x); // constexpr in C++20 template class insert_iterator + : public iterator // until C++17 { protected: Container* container; @@ -191,19 +315,20 @@ protected: public: typedef Container container_type; typedef void value_type; - typedef void difference_type; + typedef void difference_type; // until C++20 + typedef ptrdiff_t difference_type; // since C++20 typedef void reference; typedef void pointer; - insert_iterator(Container& x, typename Container::iterator i); - insert_iterator& operator=(const typename Container::value_type& value); - insert_iterator& operator*(); - insert_iterator& operator++(); - insert_iterator& operator++(int); + insert_iterator(Container& x, typename Container::iterator i); // constexpr in C++20 + insert_iterator& operator=(const typename Container::value_type& value); // constexpr in C++20 + insert_iterator& operator*(); // constexpr in C++20 + insert_iterator& operator++(); // constexpr in C++20 + insert_iterator& operator++(int); // constexpr in C++20 }; template -insert_iterator inserter(Container& x, Iterator i); +insert_iterator inserter(Container& x, Iterator i); // constexpr in C++20 template class move_iterator { @@ -274,15 +399,31 @@ constexpr move_iterator operator+( // constexpr in C++17 template // constexpr in C++17 constexpr move_iterator make_move_iterator(const Iterator& i); +// [default.sentinel], default sentinel +struct default_sentinel_t; +inline constexpr default_sentinel_t default_sentinel{}; + +// [iterators.counted], counted iterators +template class counted_iterator; + +template + requires see below + struct iterator_traits>; template , class Distance = ptrdiff_t> class istream_iterator - : public iterator + : public iterator // until C++17 { public: - typedef charT char_type; - typedef traits traits_type; - typedef basic_istream istream_type; + typedef input_iterator_tag iterator_category; + typedef T value_type; + typedef Distance difference_type; + typedef const T* pointer; + typedef const T& reference; + + typedef charT char_type; + typedef traits traits_type; + typedef basic_istream istream_type; constexpr istream_iterator(); istream_iterator(istream_type& s); @@ -304,9 +445,16 @@ bool operator!=(const istream_iterator& x, template > class ostream_iterator - : public iterator + : public iterator // until C++17 { public: + typedef output_iterator_tag iterator_category; + typedef void value_type; + typedef void difference_type; // until C++20 + typedef ptrdiff_t difference_type; // since C++20 + typedef void pointer; + typedef void reference; + typedef charT char_type; typedef traits traits_type; typedef basic_ostream ostream_type; @@ -324,16 +472,20 @@ public: template > class istreambuf_iterator - : public iterator + : public iterator // until C++17 { public: - typedef charT char_type; - typedef traits traits_type; - typedef typename traits::int_type int_type; - typedef basic_streambuf streambuf_type; - typedef basic_istream istream_type; + typedef input_iterator_tag iterator_category; + typedef charT value_type; + typedef traits::off_type difference_type; + typedef unspecified pointer; + typedef charT reference; + + typedef charT char_type; + typedef traits traits_type; + typedef traits::int_type int_type; + typedef basic_streambuf streambuf_type; + typedef basic_istream istream_type; istreambuf_iterator() noexcept; istreambuf_iterator(istream_type& s) noexcept; @@ -357,13 +509,20 @@ bool operator!=(const istreambuf_iterator& a, template > class ostreambuf_iterator - : public iterator + : public iterator // until C++17 { public: - typedef charT char_type; - typedef traits traits_type; - typedef basic_streambuf streambuf_type; - typedef basic_ostream ostream_type; + typedef output_iterator_tag iterator_category; + typedef void value_type; + typedef void difference_type; // until C++20 + typedef ptrdiff_t difference_type; // since C++20 + typedef void pointer; + typedef void reference; + + typedef charT char_type; + typedef traits traits_type; + typedef basic_streambuf streambuf_type; + typedef basic_ostream ostream_type; ostreambuf_iterator(ostream_type& s) noexcept; ostreambuf_iterator(streambuf_type* s) noexcept; @@ -399,7 +558,7 @@ template constexpr auto size(const C& c) -> decltype(c.size()); template constexpr size_t size(const T (&array)[N]) noexcept; // C++17 template constexpr auto ssize(const C& c) - -> common_type_t>; // C++20 + -> common_type_t>; // C++20 template constexpr ptrdiff_t ssize(const T (&array)[N]) noexcept; // C++20 template constexpr auto empty(const C& c) -> decltype(c.empty()); // C++17 @@ -415,1608 +574,51 @@ template constexpr const E* data(initializer_list il) noexcept; */ #include <__config> -#include // for forward declarations of vector and string. +#include <__debug> #include <__functional_base> -#include +#include <__iterator/access.h> +#include <__iterator/advance.h> +#include <__iterator/back_insert_iterator.h> +#include <__iterator/common_iterator.h> +#include <__iterator/concepts.h> +#include <__iterator/counted_iterator.h> +#include <__iterator/data.h> +#include <__iterator/default_sentinel.h> +#include <__iterator/distance.h> +#include <__iterator/empty.h> +#include <__iterator/erase_if_container.h> +#include <__iterator/front_insert_iterator.h> +#include <__iterator/incrementable_traits.h> +#include <__iterator/insert_iterator.h> +#include <__iterator/istreambuf_iterator.h> +#include <__iterator/istream_iterator.h> +#include <__iterator/iterator.h> +#include <__iterator/iterator_traits.h> +#include <__iterator/iter_move.h> +#include <__iterator/iter_swap.h> +#include <__iterator/move_iterator.h> +#include <__iterator/next.h> +#include <__iterator/ostreambuf_iterator.h> +#include <__iterator/ostream_iterator.h> +#include <__iterator/prev.h> +#include <__iterator/projected.h> +#include <__iterator/readable_traits.h> +#include <__iterator/reverse_access.h> +#include <__iterator/reverse_iterator.h> +#include <__iterator/size.h> +#include <__iterator/wrap_iter.h> +#include <__memory/addressof.h> +#include <__memory/pointer_traits.h> +#include <__utility/forward.h> +#include +#include // Mandated by the Standard. #include #include -#include <__memory/base.h> +#include #include -#include <__debug> - #if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) #pragma GCC system_header #endif -_LIBCPP_BEGIN_NAMESPACE_STD -template -struct _LIBCPP_TEMPLATE_VIS iterator_traits; - -struct _LIBCPP_TEMPLATE_VIS input_iterator_tag {}; -struct _LIBCPP_TEMPLATE_VIS output_iterator_tag {}; -struct _LIBCPP_TEMPLATE_VIS forward_iterator_tag : public input_iterator_tag {}; -struct _LIBCPP_TEMPLATE_VIS bidirectional_iterator_tag : public forward_iterator_tag {}; -struct _LIBCPP_TEMPLATE_VIS random_access_iterator_tag : public bidirectional_iterator_tag {}; -#if _LIBCPP_STD_VER > 17 -// TODO(EricWF) contiguous_iterator_tag is provided as an extension prior to -// C++20 to allow optimizations for users providing wrapped iterator types. -struct _LIBCPP_TEMPLATE_VIS contiguous_iterator_tag: public random_access_iterator_tag { }; -#endif - -template -struct __iter_traits_cache { - using type = _If< - __is_primary_template >::value, - _Iter, - iterator_traits<_Iter> - >; -}; -template -using _ITER_TRAITS = typename __iter_traits_cache<_Iter>::type; - -struct __iter_concept_concept_test { - template - using _Apply = typename _ITER_TRAITS<_Iter>::iterator_concept; -}; -struct __iter_concept_category_test { - template - using _Apply = typename _ITER_TRAITS<_Iter>::iterator_category; -}; -struct __iter_concept_random_fallback { - template - using _Apply = _EnableIf< - __is_primary_template >::value, - random_access_iterator_tag - >; -}; - -template struct __test_iter_concept - : _IsValidExpansion<_Tester::template _Apply, _Iter>, - _Tester -{ -}; - -template -struct __iter_concept_cache { - using type = _Or< - __test_iter_concept<_Iter, __iter_concept_concept_test>, - __test_iter_concept<_Iter, __iter_concept_category_test>, - __test_iter_concept<_Iter, __iter_concept_random_fallback> - >; -}; - -template -using _ITER_CONCEPT = typename __iter_concept_cache<_Iter>::type::template _Apply<_Iter>; - - -template -struct __has_iterator_typedefs -{ -private: - struct __two {char __lx; char __lxx;}; - template static __two __test(...); - template static char __test(typename __void_t::type* = 0, - typename __void_t::type* = 0, - typename __void_t::type* = 0, - typename __void_t::type* = 0, - typename __void_t::type* = 0); -public: - static const bool value = sizeof(__test<_Tp>(0,0,0,0,0)) == 1; -}; - - -template -struct __has_iterator_category -{ -private: - struct __two {char __lx; char __lxx;}; - template static __two __test(...); - template static char __test(typename _Up::iterator_category* = nullptr); -public: - static const bool value = sizeof(__test<_Tp>(nullptr)) == 1; -}; - -template struct __iterator_traits_impl {}; - -template -struct __iterator_traits_impl<_Iter, true> -{ - typedef typename _Iter::difference_type difference_type; - typedef typename _Iter::value_type value_type; - typedef typename _Iter::pointer pointer; - typedef typename _Iter::reference reference; - typedef typename _Iter::iterator_category iterator_category; -}; - -template struct __iterator_traits {}; - -template -struct __iterator_traits<_Iter, true> - : __iterator_traits_impl - < - _Iter, - is_convertible::value || - is_convertible::value - > -{}; - -// iterator_traits will only have the nested types if Iterator::iterator_category -// exists. Else iterator_traits will be an empty class. This is a -// conforming extension which allows some programs to compile and behave as -// the client expects instead of failing at compile time. - -template -struct _LIBCPP_TEMPLATE_VIS iterator_traits - : __iterator_traits<_Iter, __has_iterator_typedefs<_Iter>::value> { - - using __primary_template = iterator_traits; -}; - -template -struct _LIBCPP_TEMPLATE_VIS iterator_traits<_Tp*> -{ - typedef ptrdiff_t difference_type; - typedef typename remove_cv<_Tp>::type value_type; - typedef _Tp* pointer; - typedef _Tp& reference; - typedef random_access_iterator_tag iterator_category; -#if _LIBCPP_STD_VER > 17 - typedef contiguous_iterator_tag iterator_concept; -#endif -}; - -template >::value> -struct __has_iterator_category_convertible_to - : public integral_constant::iterator_category, _Up>::value> -{}; - -template -struct __has_iterator_category_convertible_to<_Tp, _Up, false> : public false_type {}; - -template -struct __is_cpp17_input_iterator : public __has_iterator_category_convertible_to<_Tp, input_iterator_tag> {}; - -template -struct __is_cpp17_forward_iterator : public __has_iterator_category_convertible_to<_Tp, forward_iterator_tag> {}; - -template -struct __is_cpp17_bidirectional_iterator : public __has_iterator_category_convertible_to<_Tp, bidirectional_iterator_tag> {}; - -template -struct __is_cpp17_random_access_iterator : public __has_iterator_category_convertible_to<_Tp, random_access_iterator_tag> {}; - -#if _LIBCPP_STD_VER > 17 -template -struct __is_cpp17_contiguous_iterator : public __has_iterator_category_convertible_to<_Tp, contiguous_iterator_tag> {}; -#else -template -struct __is_cpp17_contiguous_iterator : public false_type {}; -#endif - - -template -struct __is_exactly_cpp17_input_iterator - : public integral_constant::value && - !__has_iterator_category_convertible_to<_Tp, forward_iterator_tag>::value> {}; - -#ifndef _LIBCPP_HAS_NO_DEDUCTION_GUIDES -template -using __iter_value_type = typename iterator_traits<_InputIterator>::value_type; - -template -using __iter_key_type = remove_const_t::value_type::first_type>; - -template -using __iter_mapped_type = typename iterator_traits<_InputIterator>::value_type::second_type; - -template -using __iter_to_alloc_type = pair< - add_const_t::value_type::first_type>, - typename iterator_traits<_InputIterator>::value_type::second_type>; -#endif - -template -struct _LIBCPP_TEMPLATE_VIS iterator -{ - typedef _Tp value_type; - typedef _Distance difference_type; - typedef _Pointer pointer; - typedef _Reference reference; - typedef _Category iterator_category; -}; - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 -void __advance(_InputIter& __i, - typename iterator_traits<_InputIter>::difference_type __n, input_iterator_tag) -{ - for (; __n > 0; --__n) - ++__i; -} - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 -void __advance(_BiDirIter& __i, - typename iterator_traits<_BiDirIter>::difference_type __n, bidirectional_iterator_tag) -{ - if (__n >= 0) - for (; __n > 0; --__n) - ++__i; - else - for (; __n < 0; ++__n) - --__i; -} - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 -void __advance(_RandIter& __i, - typename iterator_traits<_RandIter>::difference_type __n, random_access_iterator_tag) -{ - __i += __n; -} - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 -void advance(_InputIter& __i, _Distance __orig_n) -{ - _LIBCPP_ASSERT(__orig_n >= 0 || __is_cpp17_bidirectional_iterator<_InputIter>::value, - "Attempt to advance(it, n) with negative n on a non-bidirectional iterator"); - typedef decltype(_VSTD::__convert_to_integral(__orig_n)) _IntegralSize; - _IntegralSize __n = __orig_n; - _VSTD::__advance(__i, __n, typename iterator_traits<_InputIter>::iterator_category()); -} - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 -typename iterator_traits<_InputIter>::difference_type -__distance(_InputIter __first, _InputIter __last, input_iterator_tag) -{ - typename iterator_traits<_InputIter>::difference_type __r(0); - for (; __first != __last; ++__first) - ++__r; - return __r; -} - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 -typename iterator_traits<_RandIter>::difference_type -__distance(_RandIter __first, _RandIter __last, random_access_iterator_tag) -{ - return __last - __first; -} - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 -typename iterator_traits<_InputIter>::difference_type -distance(_InputIter __first, _InputIter __last) -{ - return _VSTD::__distance(__first, __last, typename iterator_traits<_InputIter>::iterator_category()); -} - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 -typename enable_if -< - __is_cpp17_input_iterator<_InputIter>::value, - _InputIter ->::type -next(_InputIter __x, - typename iterator_traits<_InputIter>::difference_type __n = 1) -{ - _LIBCPP_ASSERT(__n >= 0 || __is_cpp17_bidirectional_iterator<_InputIter>::value, - "Attempt to next(it, n) with negative n on a non-bidirectional iterator"); - - _VSTD::advance(__x, __n); - return __x; -} - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 -typename enable_if -< - __is_cpp17_input_iterator<_InputIter>::value, - _InputIter ->::type -prev(_InputIter __x, - typename iterator_traits<_InputIter>::difference_type __n = 1) -{ - _LIBCPP_ASSERT(__n <= 0 || __is_cpp17_bidirectional_iterator<_InputIter>::value, - "Attempt to prev(it, n) with a positive n on a non-bidirectional iterator"); - _VSTD::advance(__x, -__n); - return __x; -} - - -template -struct __is_stashing_iterator : false_type {}; - -template -struct __is_stashing_iterator<_Tp, typename __void_t::type> - : true_type {}; - -template -class _LIBCPP_TEMPLATE_VIS reverse_iterator - : public iterator::iterator_category, - typename iterator_traits<_Iter>::value_type, - typename iterator_traits<_Iter>::difference_type, - typename iterator_traits<_Iter>::pointer, - typename iterator_traits<_Iter>::reference> -{ -private: - /*mutable*/ _Iter __t; // no longer used as of LWG #2360, not removed due to ABI break - - static_assert(!__is_stashing_iterator<_Iter>::value, - "The specified iterator type cannot be used with reverse_iterator; " - "Using stashing iterators with reverse_iterator causes undefined behavior"); - -protected: - _Iter current; -public: - typedef _Iter iterator_type; - typedef typename iterator_traits<_Iter>::difference_type difference_type; - typedef typename iterator_traits<_Iter>::reference reference; - typedef typename iterator_traits<_Iter>::pointer pointer; - - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 - reverse_iterator() : __t(), current() {} - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 - explicit reverse_iterator(_Iter __x) : __t(__x), current(__x) {} - template - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 - reverse_iterator(const reverse_iterator<_Up>& __u) : __t(__u.base()), current(__u.base()) {} - template - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 - reverse_iterator& operator=(const reverse_iterator<_Up>& __u) - { __t = current = __u.base(); return *this; } - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 - _Iter base() const {return current;} - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 - reference operator*() const {_Iter __tmp = current; return *--__tmp;} - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 - pointer operator->() const {return _VSTD::addressof(operator*());} - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 - reverse_iterator& operator++() {--current; return *this;} - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 - reverse_iterator operator++(int) {reverse_iterator __tmp(*this); --current; return __tmp;} - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 - reverse_iterator& operator--() {++current; return *this;} - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 - reverse_iterator operator--(int) {reverse_iterator __tmp(*this); ++current; return __tmp;} - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 - reverse_iterator operator+ (difference_type __n) const {return reverse_iterator(current - __n);} - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 - reverse_iterator& operator+=(difference_type __n) {current -= __n; return *this;} - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 - reverse_iterator operator- (difference_type __n) const {return reverse_iterator(current + __n);} - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 - reverse_iterator& operator-=(difference_type __n) {current += __n; return *this;} - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 - reference operator[](difference_type __n) const {return *(*this + __n);} -}; - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 -bool -operator==(const reverse_iterator<_Iter1>& __x, const reverse_iterator<_Iter2>& __y) -{ - return __x.base() == __y.base(); -} - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 -bool -operator<(const reverse_iterator<_Iter1>& __x, const reverse_iterator<_Iter2>& __y) -{ - return __x.base() > __y.base(); -} - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 -bool -operator!=(const reverse_iterator<_Iter1>& __x, const reverse_iterator<_Iter2>& __y) -{ - return __x.base() != __y.base(); -} - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 -bool -operator>(const reverse_iterator<_Iter1>& __x, const reverse_iterator<_Iter2>& __y) -{ - return __x.base() < __y.base(); -} - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 -bool -operator>=(const reverse_iterator<_Iter1>& __x, const reverse_iterator<_Iter2>& __y) -{ - return __x.base() <= __y.base(); -} - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 -bool -operator<=(const reverse_iterator<_Iter1>& __x, const reverse_iterator<_Iter2>& __y) -{ - return __x.base() >= __y.base(); -} - -#ifndef _LIBCPP_CXX03_LANG -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 -auto -operator-(const reverse_iterator<_Iter1>& __x, const reverse_iterator<_Iter2>& __y) --> decltype(__y.base() - __x.base()) -{ - return __y.base() - __x.base(); -} -#else -template -inline _LIBCPP_INLINE_VISIBILITY -typename reverse_iterator<_Iter1>::difference_type -operator-(const reverse_iterator<_Iter1>& __x, const reverse_iterator<_Iter2>& __y) -{ - return __y.base() - __x.base(); -} -#endif - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 -reverse_iterator<_Iter> -operator+(typename reverse_iterator<_Iter>::difference_type __n, const reverse_iterator<_Iter>& __x) -{ - return reverse_iterator<_Iter>(__x.base() - __n); -} - -#if _LIBCPP_STD_VER > 11 -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 -reverse_iterator<_Iter> make_reverse_iterator(_Iter __i) -{ - return reverse_iterator<_Iter>(__i); -} -#endif - -template -class _LIBCPP_TEMPLATE_VIS back_insert_iterator - : public iterator -{ -protected: - _Container* container; -public: - typedef _Container container_type; - - _LIBCPP_INLINE_VISIBILITY explicit back_insert_iterator(_Container& __x) : container(_VSTD::addressof(__x)) {} - _LIBCPP_INLINE_VISIBILITY back_insert_iterator& operator=(const typename _Container::value_type& __value_) - {container->push_back(__value_); return *this;} -#ifndef _LIBCPP_CXX03_LANG - _LIBCPP_INLINE_VISIBILITY back_insert_iterator& operator=(typename _Container::value_type&& __value_) - {container->push_back(_VSTD::move(__value_)); return *this;} -#endif // _LIBCPP_CXX03_LANG - _LIBCPP_INLINE_VISIBILITY back_insert_iterator& operator*() {return *this;} - _LIBCPP_INLINE_VISIBILITY back_insert_iterator& operator++() {return *this;} - _LIBCPP_INLINE_VISIBILITY back_insert_iterator operator++(int) {return *this;} -}; - -template -inline _LIBCPP_INLINE_VISIBILITY -back_insert_iterator<_Container> -back_inserter(_Container& __x) -{ - return back_insert_iterator<_Container>(__x); -} - -template -class _LIBCPP_TEMPLATE_VIS front_insert_iterator - : public iterator -{ -protected: - _Container* container; -public: - typedef _Container container_type; - - _LIBCPP_INLINE_VISIBILITY explicit front_insert_iterator(_Container& __x) : container(_VSTD::addressof(__x)) {} - _LIBCPP_INLINE_VISIBILITY front_insert_iterator& operator=(const typename _Container::value_type& __value_) - {container->push_front(__value_); return *this;} -#ifndef _LIBCPP_CXX03_LANG - _LIBCPP_INLINE_VISIBILITY front_insert_iterator& operator=(typename _Container::value_type&& __value_) - {container->push_front(_VSTD::move(__value_)); return *this;} -#endif // _LIBCPP_CXX03_LANG - _LIBCPP_INLINE_VISIBILITY front_insert_iterator& operator*() {return *this;} - _LIBCPP_INLINE_VISIBILITY front_insert_iterator& operator++() {return *this;} - _LIBCPP_INLINE_VISIBILITY front_insert_iterator operator++(int) {return *this;} -}; - -template -inline _LIBCPP_INLINE_VISIBILITY -front_insert_iterator<_Container> -front_inserter(_Container& __x) -{ - return front_insert_iterator<_Container>(__x); -} - -template -class _LIBCPP_TEMPLATE_VIS insert_iterator - : public iterator -{ -protected: - _Container* container; - typename _Container::iterator iter; -public: - typedef _Container container_type; - - _LIBCPP_INLINE_VISIBILITY insert_iterator(_Container& __x, typename _Container::iterator __i) - : container(_VSTD::addressof(__x)), iter(__i) {} - _LIBCPP_INLINE_VISIBILITY insert_iterator& operator=(const typename _Container::value_type& __value_) - {iter = container->insert(iter, __value_); ++iter; return *this;} -#ifndef _LIBCPP_CXX03_LANG - _LIBCPP_INLINE_VISIBILITY insert_iterator& operator=(typename _Container::value_type&& __value_) - {iter = container->insert(iter, _VSTD::move(__value_)); ++iter; return *this;} -#endif // _LIBCPP_CXX03_LANG - _LIBCPP_INLINE_VISIBILITY insert_iterator& operator*() {return *this;} - _LIBCPP_INLINE_VISIBILITY insert_iterator& operator++() {return *this;} - _LIBCPP_INLINE_VISIBILITY insert_iterator& operator++(int) {return *this;} -}; - -template -inline _LIBCPP_INLINE_VISIBILITY -insert_iterator<_Container> -inserter(_Container& __x, typename _Container::iterator __i) -{ - return insert_iterator<_Container>(__x, __i); -} - -template , class _Distance = ptrdiff_t> -class _LIBCPP_TEMPLATE_VIS istream_iterator - : public iterator -{ -public: - typedef _CharT char_type; - typedef _Traits traits_type; - typedef basic_istream<_CharT,_Traits> istream_type; -private: - istream_type* __in_stream_; - _Tp __value_; -public: - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR istream_iterator() : __in_stream_(nullptr), __value_() {} - _LIBCPP_INLINE_VISIBILITY istream_iterator(istream_type& __s) : __in_stream_(_VSTD::addressof(__s)) - { - if (!(*__in_stream_ >> __value_)) - __in_stream_ = nullptr; - } - - _LIBCPP_INLINE_VISIBILITY const _Tp& operator*() const {return __value_;} - _LIBCPP_INLINE_VISIBILITY const _Tp* operator->() const {return _VSTD::addressof((operator*()));} - _LIBCPP_INLINE_VISIBILITY istream_iterator& operator++() - { - if (!(*__in_stream_ >> __value_)) - __in_stream_ = nullptr; - return *this; - } - _LIBCPP_INLINE_VISIBILITY istream_iterator operator++(int) - {istream_iterator __t(*this); ++(*this); return __t;} - - template - friend _LIBCPP_INLINE_VISIBILITY - bool - operator==(const istream_iterator<_Up, _CharU, _TraitsU, _DistanceU>& __x, - const istream_iterator<_Up, _CharU, _TraitsU, _DistanceU>& __y); - - template - friend _LIBCPP_INLINE_VISIBILITY - bool - operator==(const istream_iterator<_Up, _CharU, _TraitsU, _DistanceU>& __x, - const istream_iterator<_Up, _CharU, _TraitsU, _DistanceU>& __y); -}; - -template -inline _LIBCPP_INLINE_VISIBILITY -bool -operator==(const istream_iterator<_Tp, _CharT, _Traits, _Distance>& __x, - const istream_iterator<_Tp, _CharT, _Traits, _Distance>& __y) -{ - return __x.__in_stream_ == __y.__in_stream_; -} - -template -inline _LIBCPP_INLINE_VISIBILITY -bool -operator!=(const istream_iterator<_Tp, _CharT, _Traits, _Distance>& __x, - const istream_iterator<_Tp, _CharT, _Traits, _Distance>& __y) -{ - return !(__x == __y); -} - -template > -class _LIBCPP_TEMPLATE_VIS ostream_iterator - : public iterator -{ -public: - typedef output_iterator_tag iterator_category; - typedef void value_type; -#if _LIBCPP_STD_VER > 17 - typedef std::ptrdiff_t difference_type; -#else - typedef void difference_type; -#endif - typedef void pointer; - typedef void reference; - typedef _CharT char_type; - typedef _Traits traits_type; - typedef basic_ostream<_CharT, _Traits> ostream_type; - -private: - ostream_type* __out_stream_; - const char_type* __delim_; -public: - _LIBCPP_INLINE_VISIBILITY ostream_iterator(ostream_type& __s) _NOEXCEPT - : __out_stream_(_VSTD::addressof(__s)), __delim_(nullptr) {} - _LIBCPP_INLINE_VISIBILITY ostream_iterator(ostream_type& __s, const _CharT* __delimiter) _NOEXCEPT - : __out_stream_(_VSTD::addressof(__s)), __delim_(__delimiter) {} - _LIBCPP_INLINE_VISIBILITY ostream_iterator& operator=(const _Tp& __value_) - { - *__out_stream_ << __value_; - if (__delim_) - *__out_stream_ << __delim_; - return *this; - } - - _LIBCPP_INLINE_VISIBILITY ostream_iterator& operator*() {return *this;} - _LIBCPP_INLINE_VISIBILITY ostream_iterator& operator++() {return *this;} - _LIBCPP_INLINE_VISIBILITY ostream_iterator& operator++(int) {return *this;} -}; - -template -class _LIBCPP_TEMPLATE_VIS istreambuf_iterator - : public iterator -{ -public: - typedef _CharT char_type; - typedef _Traits traits_type; - typedef typename _Traits::int_type int_type; - typedef basic_streambuf<_CharT,_Traits> streambuf_type; - typedef basic_istream<_CharT,_Traits> istream_type; -private: - mutable streambuf_type* __sbuf_; - - class __proxy - { - char_type __keep_; - streambuf_type* __sbuf_; - _LIBCPP_INLINE_VISIBILITY __proxy(char_type __c, streambuf_type* __s) - : __keep_(__c), __sbuf_(__s) {} - friend class istreambuf_iterator; - public: - _LIBCPP_INLINE_VISIBILITY char_type operator*() const {return __keep_;} - }; - - _LIBCPP_INLINE_VISIBILITY - bool __test_for_eof() const - { - if (__sbuf_ && traits_type::eq_int_type(__sbuf_->sgetc(), traits_type::eof())) - __sbuf_ = nullptr; - return __sbuf_ == nullptr; - } -public: - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR istreambuf_iterator() _NOEXCEPT : __sbuf_(nullptr) {} - _LIBCPP_INLINE_VISIBILITY istreambuf_iterator(istream_type& __s) _NOEXCEPT - : __sbuf_(__s.rdbuf()) {} - _LIBCPP_INLINE_VISIBILITY istreambuf_iterator(streambuf_type* __s) _NOEXCEPT - : __sbuf_(__s) {} - _LIBCPP_INLINE_VISIBILITY istreambuf_iterator(const __proxy& __p) _NOEXCEPT - : __sbuf_(__p.__sbuf_) {} - - _LIBCPP_INLINE_VISIBILITY char_type operator*() const - {return static_cast(__sbuf_->sgetc());} - _LIBCPP_INLINE_VISIBILITY istreambuf_iterator& operator++() - { - __sbuf_->sbumpc(); - return *this; - } - _LIBCPP_INLINE_VISIBILITY __proxy operator++(int) - { - return __proxy(__sbuf_->sbumpc(), __sbuf_); - } - - _LIBCPP_INLINE_VISIBILITY bool equal(const istreambuf_iterator& __b) const - {return __test_for_eof() == __b.__test_for_eof();} -}; - -template -inline _LIBCPP_INLINE_VISIBILITY -bool operator==(const istreambuf_iterator<_CharT,_Traits>& __a, - const istreambuf_iterator<_CharT,_Traits>& __b) - {return __a.equal(__b);} - -template -inline _LIBCPP_INLINE_VISIBILITY -bool operator!=(const istreambuf_iterator<_CharT,_Traits>& __a, - const istreambuf_iterator<_CharT,_Traits>& __b) - {return !__a.equal(__b);} - -template -class _LIBCPP_TEMPLATE_VIS ostreambuf_iterator - : public iterator -{ -public: - typedef output_iterator_tag iterator_category; - typedef void value_type; -#if _LIBCPP_STD_VER > 17 - typedef std::ptrdiff_t difference_type; -#else - typedef void difference_type; -#endif - typedef void pointer; - typedef void reference; - typedef _CharT char_type; - typedef _Traits traits_type; - typedef basic_streambuf<_CharT, _Traits> streambuf_type; - typedef basic_ostream<_CharT, _Traits> ostream_type; - -private: - streambuf_type* __sbuf_; -public: - _LIBCPP_INLINE_VISIBILITY ostreambuf_iterator(ostream_type& __s) _NOEXCEPT - : __sbuf_(__s.rdbuf()) {} - _LIBCPP_INLINE_VISIBILITY ostreambuf_iterator(streambuf_type* __s) _NOEXCEPT - : __sbuf_(__s) {} - _LIBCPP_INLINE_VISIBILITY ostreambuf_iterator& operator=(_CharT __c) - { - if (__sbuf_ && traits_type::eq_int_type(__sbuf_->sputc(__c), traits_type::eof())) - __sbuf_ = nullptr; - return *this; - } - _LIBCPP_INLINE_VISIBILITY ostreambuf_iterator& operator*() {return *this;} - _LIBCPP_INLINE_VISIBILITY ostreambuf_iterator& operator++() {return *this;} - _LIBCPP_INLINE_VISIBILITY ostreambuf_iterator& operator++(int) {return *this;} - _LIBCPP_INLINE_VISIBILITY bool failed() const _NOEXCEPT {return __sbuf_ == nullptr;} - - template - friend - _LIBCPP_HIDDEN - ostreambuf_iterator<_Ch, _Tr> - __pad_and_output(ostreambuf_iterator<_Ch, _Tr> __s, - const _Ch* __ob, const _Ch* __op, const _Ch* __oe, - ios_base& __iob, _Ch __fl); -}; - -template -class _LIBCPP_TEMPLATE_VIS move_iterator -{ -private: - _Iter __i; -public: - typedef _Iter iterator_type; - typedef typename iterator_traits::iterator_category iterator_category; - typedef typename iterator_traits::value_type value_type; - typedef typename iterator_traits::difference_type difference_type; - typedef iterator_type pointer; -#ifndef _LIBCPP_CXX03_LANG - typedef typename iterator_traits::reference __reference; - typedef typename conditional< - is_reference<__reference>::value, - typename remove_reference<__reference>::type&&, - __reference - >::type reference; -#else - typedef typename iterator_traits::reference reference; -#endif - - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 - move_iterator() : __i() {} - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 - explicit move_iterator(_Iter __x) : __i(__x) {} - template - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 - move_iterator(const move_iterator<_Up>& __u) : __i(__u.base()) {} - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 _Iter base() const {return __i;} - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 - reference operator*() const { return static_cast(*__i); } - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 - pointer operator->() const { return __i;} - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 - move_iterator& operator++() {++__i; return *this;} - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 - move_iterator operator++(int) {move_iterator __tmp(*this); ++__i; return __tmp;} - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 - move_iterator& operator--() {--__i; return *this;} - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 - move_iterator operator--(int) {move_iterator __tmp(*this); --__i; return __tmp;} - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 - move_iterator operator+ (difference_type __n) const {return move_iterator(__i + __n);} - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 - move_iterator& operator+=(difference_type __n) {__i += __n; return *this;} - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 - move_iterator operator- (difference_type __n) const {return move_iterator(__i - __n);} - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 - move_iterator& operator-=(difference_type __n) {__i -= __n; return *this;} - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 - reference operator[](difference_type __n) const { return static_cast(__i[__n]); } -}; - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 -bool -operator==(const move_iterator<_Iter1>& __x, const move_iterator<_Iter2>& __y) -{ - return __x.base() == __y.base(); -} - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 -bool -operator<(const move_iterator<_Iter1>& __x, const move_iterator<_Iter2>& __y) -{ - return __x.base() < __y.base(); -} - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 -bool -operator!=(const move_iterator<_Iter1>& __x, const move_iterator<_Iter2>& __y) -{ - return __x.base() != __y.base(); -} - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 -bool -operator>(const move_iterator<_Iter1>& __x, const move_iterator<_Iter2>& __y) -{ - return __x.base() > __y.base(); -} - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 -bool -operator>=(const move_iterator<_Iter1>& __x, const move_iterator<_Iter2>& __y) -{ - return __x.base() >= __y.base(); -} - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 -bool -operator<=(const move_iterator<_Iter1>& __x, const move_iterator<_Iter2>& __y) -{ - return __x.base() <= __y.base(); -} - -#ifndef _LIBCPP_CXX03_LANG -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 -auto -operator-(const move_iterator<_Iter1>& __x, const move_iterator<_Iter2>& __y) --> decltype(__x.base() - __y.base()) -{ - return __x.base() - __y.base(); -} -#else -template -inline _LIBCPP_INLINE_VISIBILITY -typename move_iterator<_Iter1>::difference_type -operator-(const move_iterator<_Iter1>& __x, const move_iterator<_Iter2>& __y) -{ - return __x.base() - __y.base(); -} -#endif - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 -move_iterator<_Iter> -operator+(typename move_iterator<_Iter>::difference_type __n, const move_iterator<_Iter>& __x) -{ - return move_iterator<_Iter>(__x.base() + __n); -} - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 -move_iterator<_Iter> -make_move_iterator(_Iter __i) -{ - return move_iterator<_Iter>(__i); -} - -// __wrap_iter - -template class __wrap_iter; - -template -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_IF_NODEBUG -bool -operator==(const __wrap_iter<_Iter1>&, const __wrap_iter<_Iter2>&) _NOEXCEPT; - -template -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_IF_NODEBUG -bool -operator<(const __wrap_iter<_Iter1>&, const __wrap_iter<_Iter2>&) _NOEXCEPT; - -template -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_IF_NODEBUG -bool -operator!=(const __wrap_iter<_Iter1>&, const __wrap_iter<_Iter2>&) _NOEXCEPT; - -template -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_IF_NODEBUG -bool -operator>(const __wrap_iter<_Iter1>&, const __wrap_iter<_Iter2>&) _NOEXCEPT; - -template -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_IF_NODEBUG -bool -operator>=(const __wrap_iter<_Iter1>&, const __wrap_iter<_Iter2>&) _NOEXCEPT; - -template -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_IF_NODEBUG -bool -operator<=(const __wrap_iter<_Iter1>&, const __wrap_iter<_Iter2>&) _NOEXCEPT; - -#ifndef _LIBCPP_CXX03_LANG -template -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_IF_NODEBUG -auto -operator-(const __wrap_iter<_Iter1>& __x, const __wrap_iter<_Iter2>& __y) _NOEXCEPT --> decltype(__x.base() - __y.base()); -#else -template -_LIBCPP_INLINE_VISIBILITY -typename __wrap_iter<_Iter1>::difference_type -operator-(const __wrap_iter<_Iter1>&, const __wrap_iter<_Iter2>&) _NOEXCEPT; -#endif - -template -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_IF_NODEBUG -__wrap_iter<_Iter> -operator+(typename __wrap_iter<_Iter>::difference_type, __wrap_iter<_Iter>) _NOEXCEPT; - -template _Op _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 copy(_Ip, _Ip, _Op); -template _B2 _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 copy_backward(_B1, _B1, _B2); -template _Op _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 move(_Ip, _Ip, _Op); -template _B2 _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 move_backward(_B1, _B1, _B2); - -#if _LIBCPP_DEBUG_LEVEL < 2 - -template -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR -typename enable_if -< - is_trivially_copy_assignable<_Tp>::value, - _Tp* ->::type -__unwrap_iter(__wrap_iter<_Tp*>); - -#else - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR -typename enable_if -< - is_trivially_copy_assignable<_Tp>::value, - __wrap_iter<_Tp*> ->::type -__unwrap_iter(__wrap_iter<_Tp*> __i); - -#endif - -template -class __wrap_iter -{ -public: - typedef _Iter iterator_type; - typedef typename iterator_traits::iterator_category iterator_category; - typedef typename iterator_traits::value_type value_type; - typedef typename iterator_traits::difference_type difference_type; - typedef typename iterator_traits::pointer pointer; - typedef typename iterator_traits::reference reference; -private: - iterator_type __i; -public: - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_IF_NODEBUG __wrap_iter() _NOEXCEPT -#if _LIBCPP_STD_VER > 11 - : __i{} -#endif - { -#if _LIBCPP_DEBUG_LEVEL == 2 - __get_db()->__insert_i(this); -#endif - } - template _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_IF_NODEBUG - __wrap_iter(const __wrap_iter<_Up>& __u, - typename enable_if::value>::type* = nullptr) _NOEXCEPT - : __i(__u.base()) - { -#if _LIBCPP_DEBUG_LEVEL == 2 - __get_db()->__iterator_copy(this, &__u); -#endif - } -#if _LIBCPP_DEBUG_LEVEL == 2 - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_IF_NODEBUG - __wrap_iter(const __wrap_iter& __x) - : __i(__x.base()) - { - __get_db()->__iterator_copy(this, &__x); - } - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_IF_NODEBUG - __wrap_iter& operator=(const __wrap_iter& __x) - { - if (this != &__x) - { - __get_db()->__iterator_copy(this, &__x); - __i = __x.__i; - } - return *this; - } - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_IF_NODEBUG - ~__wrap_iter() - { - __get_db()->__erase_i(this); - } -#endif - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_IF_NODEBUG reference operator*() const _NOEXCEPT - { -#if _LIBCPP_DEBUG_LEVEL == 2 - _LIBCPP_ASSERT(__get_const_db()->__dereferenceable(this), - "Attempted to dereference a non-dereferenceable iterator"); -#endif - return *__i; - } - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_IF_NODEBUG pointer operator->() const _NOEXCEPT - { -#if _LIBCPP_DEBUG_LEVEL == 2 - _LIBCPP_ASSERT(__get_const_db()->__dereferenceable(this), - "Attempted to dereference a non-dereferenceable iterator"); -#endif - return (pointer)_VSTD::addressof(*__i); - } - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_IF_NODEBUG __wrap_iter& operator++() _NOEXCEPT - { -#if _LIBCPP_DEBUG_LEVEL == 2 - _LIBCPP_ASSERT(__get_const_db()->__dereferenceable(this), - "Attempted to increment non-incrementable iterator"); -#endif - ++__i; - return *this; - } - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_IF_NODEBUG __wrap_iter operator++(int) _NOEXCEPT - {__wrap_iter __tmp(*this); ++(*this); return __tmp;} - - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_IF_NODEBUG __wrap_iter& operator--() _NOEXCEPT - { -#if _LIBCPP_DEBUG_LEVEL == 2 - _LIBCPP_ASSERT(__get_const_db()->__decrementable(this), - "Attempted to decrement non-decrementable iterator"); -#endif - --__i; - return *this; - } - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_IF_NODEBUG __wrap_iter operator--(int) _NOEXCEPT - {__wrap_iter __tmp(*this); --(*this); return __tmp;} - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_IF_NODEBUG __wrap_iter operator+ (difference_type __n) const _NOEXCEPT - {__wrap_iter __w(*this); __w += __n; return __w;} - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_IF_NODEBUG __wrap_iter& operator+=(difference_type __n) _NOEXCEPT - { -#if _LIBCPP_DEBUG_LEVEL == 2 - _LIBCPP_ASSERT(__get_const_db()->__addable(this, __n), - "Attempted to add/subtract iterator outside of valid range"); -#endif - __i += __n; - return *this; - } - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_IF_NODEBUG __wrap_iter operator- (difference_type __n) const _NOEXCEPT - {return *this + (-__n);} - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_IF_NODEBUG __wrap_iter& operator-=(difference_type __n) _NOEXCEPT - {*this += -__n; return *this;} - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_IF_NODEBUG reference operator[](difference_type __n) const _NOEXCEPT - { -#if _LIBCPP_DEBUG_LEVEL == 2 - _LIBCPP_ASSERT(__get_const_db()->__subscriptable(this, __n), - "Attempted to subscript iterator outside of valid range"); -#endif - return __i[__n]; - } - - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_IF_NODEBUG iterator_type base() const _NOEXCEPT {return __i;} - -private: -#if _LIBCPP_DEBUG_LEVEL == 2 - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_IF_NODEBUG __wrap_iter(const void* __p, iterator_type __x) : __i(__x) - { - __get_db()->__insert_ic(this, __p); - } -#else - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_IF_NODEBUG __wrap_iter(iterator_type __x) _NOEXCEPT : __i(__x) {} -#endif - - template friend class __wrap_iter; - template friend class basic_string; - template friend class _LIBCPP_TEMPLATE_VIS vector; - template friend class _LIBCPP_TEMPLATE_VIS span; - - template - _LIBCPP_CONSTEXPR_IF_NODEBUG friend - bool - operator==(const __wrap_iter<_Iter1>&, const __wrap_iter<_Iter2>&) _NOEXCEPT; - - template - _LIBCPP_CONSTEXPR_IF_NODEBUG friend - bool - operator<(const __wrap_iter<_Iter1>&, const __wrap_iter<_Iter2>&) _NOEXCEPT; - - template - _LIBCPP_CONSTEXPR_IF_NODEBUG friend - bool - operator!=(const __wrap_iter<_Iter1>&, const __wrap_iter<_Iter2>&) _NOEXCEPT; - - template - _LIBCPP_CONSTEXPR_IF_NODEBUG friend - bool - operator>(const __wrap_iter<_Iter1>&, const __wrap_iter<_Iter2>&) _NOEXCEPT; - - template - _LIBCPP_CONSTEXPR_IF_NODEBUG friend - bool - operator>=(const __wrap_iter<_Iter1>&, const __wrap_iter<_Iter2>&) _NOEXCEPT; - - template - _LIBCPP_CONSTEXPR_IF_NODEBUG friend - bool - operator<=(const __wrap_iter<_Iter1>&, const __wrap_iter<_Iter2>&) _NOEXCEPT; - -#ifndef _LIBCPP_CXX03_LANG - template - _LIBCPP_CONSTEXPR_IF_NODEBUG friend - auto - operator-(const __wrap_iter<_Iter1>& __x, const __wrap_iter<_Iter2>& __y) _NOEXCEPT - -> decltype(__x.base() - __y.base()); -#else - template - _LIBCPP_CONSTEXPR_IF_NODEBUG friend - typename __wrap_iter<_Iter1>::difference_type - operator-(const __wrap_iter<_Iter1>&, const __wrap_iter<_Iter2>&) _NOEXCEPT; -#endif - - template - _LIBCPP_CONSTEXPR_IF_NODEBUG friend - __wrap_iter<_Iter1> - operator+(typename __wrap_iter<_Iter1>::difference_type, __wrap_iter<_Iter1>) _NOEXCEPT; - - template friend _LIBCPP_CONSTEXPR_AFTER_CXX17 _Op copy(_Ip, _Ip, _Op); - template friend _LIBCPP_CONSTEXPR_AFTER_CXX17 _B2 copy_backward(_B1, _B1, _B2); - template friend _LIBCPP_CONSTEXPR_AFTER_CXX17 _Op move(_Ip, _Ip, _Op); - template friend _LIBCPP_CONSTEXPR_AFTER_CXX17 _B2 move_backward(_B1, _B1, _B2); - -#if _LIBCPP_DEBUG_LEVEL < 2 - template - _LIBCPP_CONSTEXPR friend - typename enable_if - < - is_trivially_copy_assignable<_Tp>::value, - _Tp* - >::type - __unwrap_iter(__wrap_iter<_Tp*>); -#else - template - inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR friend - typename enable_if - < - is_trivially_copy_assignable<_Tp>::value, - __wrap_iter<_Tp*> - >::type - __unwrap_iter(__wrap_iter<_Tp*> __i); -#endif -}; - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_IF_NODEBUG -bool -operator==(const __wrap_iter<_Iter1>& __x, const __wrap_iter<_Iter2>& __y) _NOEXCEPT -{ - return __x.base() == __y.base(); -} - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_IF_NODEBUG -bool -operator<(const __wrap_iter<_Iter1>& __x, const __wrap_iter<_Iter2>& __y) _NOEXCEPT -{ -#if _LIBCPP_DEBUG_LEVEL == 2 - _LIBCPP_ASSERT(__get_const_db()->__less_than_comparable(&__x, &__y), - "Attempted to compare incomparable iterators"); -#endif - return __x.base() < __y.base(); -} - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_IF_NODEBUG -bool -operator!=(const __wrap_iter<_Iter1>& __x, const __wrap_iter<_Iter2>& __y) _NOEXCEPT -{ - return !(__x == __y); -} - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_IF_NODEBUG -bool -operator>(const __wrap_iter<_Iter1>& __x, const __wrap_iter<_Iter2>& __y) _NOEXCEPT -{ - return __y < __x; -} - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_IF_NODEBUG -bool -operator>=(const __wrap_iter<_Iter1>& __x, const __wrap_iter<_Iter2>& __y) _NOEXCEPT -{ - return !(__x < __y); -} - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_IF_NODEBUG -bool -operator<=(const __wrap_iter<_Iter1>& __x, const __wrap_iter<_Iter2>& __y) _NOEXCEPT -{ - return !(__y < __x); -} - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_IF_NODEBUG -bool -operator!=(const __wrap_iter<_Iter1>& __x, const __wrap_iter<_Iter1>& __y) _NOEXCEPT -{ - return !(__x == __y); -} - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_IF_NODEBUG -bool -operator>(const __wrap_iter<_Iter1>& __x, const __wrap_iter<_Iter1>& __y) _NOEXCEPT -{ - return __y < __x; -} - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_IF_NODEBUG -bool -operator>=(const __wrap_iter<_Iter1>& __x, const __wrap_iter<_Iter1>& __y) _NOEXCEPT -{ - return !(__x < __y); -} - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_IF_NODEBUG -bool -operator<=(const __wrap_iter<_Iter1>& __x, const __wrap_iter<_Iter1>& __y) _NOEXCEPT -{ - return !(__y < __x); -} - -#ifndef _LIBCPP_CXX03_LANG -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_IF_NODEBUG -auto -operator-(const __wrap_iter<_Iter1>& __x, const __wrap_iter<_Iter2>& __y) _NOEXCEPT --> decltype(__x.base() - __y.base()) -{ -#if _LIBCPP_DEBUG_LEVEL == 2 - _LIBCPP_ASSERT(__get_const_db()->__less_than_comparable(&__x, &__y), - "Attempted to subtract incompatible iterators"); -#endif - return __x.base() - __y.base(); -} -#else -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_IF_NODEBUG -typename __wrap_iter<_Iter1>::difference_type -operator-(const __wrap_iter<_Iter1>& __x, const __wrap_iter<_Iter2>& __y) _NOEXCEPT -{ -#if _LIBCPP_DEBUG_LEVEL == 2 - _LIBCPP_ASSERT(__get_const_db()->__less_than_comparable(&__x, &__y), - "Attempted to subtract incompatible iterators"); -#endif - return __x.base() - __y.base(); -} -#endif - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_IF_NODEBUG -__wrap_iter<_Iter> -operator+(typename __wrap_iter<_Iter>::difference_type __n, - __wrap_iter<_Iter> __x) _NOEXCEPT -{ - __x += __n; - return __x; -} - -template -struct __libcpp_is_trivial_iterator - : public _LIBCPP_BOOL_CONSTANT(is_pointer<_Iter>::value) {}; - -template -struct __libcpp_is_trivial_iterator > - : public _LIBCPP_BOOL_CONSTANT(__libcpp_is_trivial_iterator<_Iter>::value) {}; - -template -struct __libcpp_is_trivial_iterator > - : public _LIBCPP_BOOL_CONSTANT(__libcpp_is_trivial_iterator<_Iter>::value) {}; - -template -struct __libcpp_is_trivial_iterator<__wrap_iter<_Iter> > - : public _LIBCPP_BOOL_CONSTANT(__libcpp_is_trivial_iterator<_Iter>::value) {}; - - -template -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 -_Tp* -begin(_Tp (&__array)[_Np]) -{ - return __array; -} - -template -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 -_Tp* -end(_Tp (&__array)[_Np]) -{ - return __array + _Np; -} - -#if !defined(_LIBCPP_CXX03_LANG) - -template -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 -auto -begin(_Cp& __c) -> decltype(__c.begin()) -{ - return __c.begin(); -} - -template -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 -auto -begin(const _Cp& __c) -> decltype(__c.begin()) -{ - return __c.begin(); -} - -template -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 -auto -end(_Cp& __c) -> decltype(__c.end()) -{ - return __c.end(); -} - -template -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 -auto -end(const _Cp& __c) -> decltype(__c.end()) -{ - return __c.end(); -} - -#if _LIBCPP_STD_VER > 11 - -template -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 -reverse_iterator<_Tp*> rbegin(_Tp (&__array)[_Np]) -{ - return reverse_iterator<_Tp*>(__array + _Np); -} - -template -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 -reverse_iterator<_Tp*> rend(_Tp (&__array)[_Np]) -{ - return reverse_iterator<_Tp*>(__array); -} - -template -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 -reverse_iterator rbegin(initializer_list<_Ep> __il) -{ - return reverse_iterator(__il.end()); -} - -template -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 -reverse_iterator rend(initializer_list<_Ep> __il) -{ - return reverse_iterator(__il.begin()); -} - -template -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 -auto cbegin(const _Cp& __c) -> decltype(_VSTD::begin(__c)) -{ - return _VSTD::begin(__c); -} - -template -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 -auto cend(const _Cp& __c) -> decltype(_VSTD::end(__c)) -{ - return _VSTD::end(__c); -} - -template -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 -auto rbegin(_Cp& __c) -> decltype(__c.rbegin()) -{ - return __c.rbegin(); -} - -template -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 -auto rbegin(const _Cp& __c) -> decltype(__c.rbegin()) -{ - return __c.rbegin(); -} - -template -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 -auto rend(_Cp& __c) -> decltype(__c.rend()) -{ - return __c.rend(); -} - -template -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 -auto rend(const _Cp& __c) -> decltype(__c.rend()) -{ - return __c.rend(); -} - -template -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 -auto crbegin(const _Cp& __c) -> decltype(_VSTD::rbegin(__c)) -{ - return _VSTD::rbegin(__c); -} - -template -_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 -auto crend(const _Cp& __c) -> decltype(_VSTD::rend(__c)) -{ - return _VSTD::rend(__c); -} - -#endif - - -#else // defined(_LIBCPP_CXX03_LANG) - -template -_LIBCPP_INLINE_VISIBILITY -typename _Cp::iterator -begin(_Cp& __c) -{ - return __c.begin(); -} - -template -_LIBCPP_INLINE_VISIBILITY -typename _Cp::const_iterator -begin(const _Cp& __c) -{ - return __c.begin(); -} - -template -_LIBCPP_INLINE_VISIBILITY -typename _Cp::iterator -end(_Cp& __c) -{ - return __c.end(); -} - -template -_LIBCPP_INLINE_VISIBILITY -typename _Cp::const_iterator -end(const _Cp& __c) -{ - return __c.end(); -} - -#endif // !defined(_LIBCPP_CXX03_LANG) - -#if _LIBCPP_STD_VER > 14 - -// #if _LIBCPP_STD_VER > 11 -// template <> -// struct _LIBCPP_TEMPLATE_VIS plus -// { -// template -// _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY -// auto operator()(_T1&& __t, _T2&& __u) const -// _NOEXCEPT_(noexcept(_VSTD::forward<_T1>(__t) + _VSTD::forward<_T2>(__u))) -// -> decltype (_VSTD::forward<_T1>(__t) + _VSTD::forward<_T2>(__u)) -// { return _VSTD::forward<_T1>(__t) + _VSTD::forward<_T2>(__u); } -// typedef void is_transparent; -// }; -// #endif - -template -_LIBCPP_INLINE_VISIBILITY -constexpr auto size(const _Cont& __c) -_NOEXCEPT_(noexcept(__c.size())) --> decltype (__c.size()) -{ return __c.size(); } - -template -_LIBCPP_INLINE_VISIBILITY -constexpr size_t size(const _Tp (&)[_Sz]) noexcept { return _Sz; } - -#if _LIBCPP_STD_VER > 17 -template -_LIBCPP_INLINE_VISIBILITY -constexpr auto ssize(const _Cont& __c) -_NOEXCEPT_(noexcept(static_cast>>(__c.size()))) --> common_type_t> -{ return static_cast>>(__c.size()); } - -template -_LIBCPP_INLINE_VISIBILITY -constexpr ptrdiff_t ssize(const _Tp (&)[_Sz]) noexcept { return _Sz; } -#endif - -template -_LIBCPP_NODISCARD_AFTER_CXX17 _LIBCPP_INLINE_VISIBILITY -constexpr auto empty(const _Cont& __c) -_NOEXCEPT_(noexcept(__c.empty())) --> decltype (__c.empty()) -{ return __c.empty(); } - -template -_LIBCPP_NODISCARD_AFTER_CXX17 _LIBCPP_INLINE_VISIBILITY -constexpr bool empty(const _Tp (&)[_Sz]) noexcept { return false; } - -template -_LIBCPP_NODISCARD_AFTER_CXX17 _LIBCPP_INLINE_VISIBILITY -constexpr bool empty(initializer_list<_Ep> __il) noexcept { return __il.size() == 0; } - -template constexpr -_LIBCPP_INLINE_VISIBILITY -auto data(_Cont& __c) -_NOEXCEPT_(noexcept(__c.data())) --> decltype (__c.data()) -{ return __c.data(); } - -template constexpr -_LIBCPP_INLINE_VISIBILITY -auto data(const _Cont& __c) -_NOEXCEPT_(noexcept(__c.data())) --> decltype (__c.data()) -{ return __c.data(); } - -template -_LIBCPP_INLINE_VISIBILITY -constexpr _Tp* data(_Tp (&__array)[_Sz]) noexcept { return __array; } - -template -_LIBCPP_INLINE_VISIBILITY -constexpr const _Ep* data(initializer_list<_Ep> __il) noexcept { return __il.begin(); } -#endif - - -_LIBCPP_END_NAMESPACE_STD - -#endif // _LIBCPP_ITERATOR +#endif // _LIBCPP_ITERATOR diff --git a/lib/libcxx/include/latch b/lib/libcxx/include/latch index a894f8caf..10ae5721f 100644 --- a/lib/libcxx/include/latch +++ b/lib/libcxx/include/latch @@ -40,8 +40,8 @@ namespace std */ -#include <__config> #include <__availability> +#include <__config> #include #if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) diff --git a/lib/libcxx/include/limits b/lib/libcxx/include/limits index 8f97cd10a..a6d517852 100644 --- a/lib/libcxx/include/limits +++ b/lib/libcxx/include/limits @@ -815,4 +815,4 @@ _LIBCPP_END_NAMESPACE_STD _LIBCPP_POP_MACROS -#endif // _LIBCPP_LIMITS +#endif // _LIBCPP_LIMITS diff --git a/lib/libcxx/include/limits.h b/lib/libcxx/include/limits.h index 4a212f8f3..938f8875b 100644 --- a/lib/libcxx/include/limits.h +++ b/lib/libcxx/include/limits.h @@ -61,4 +61,4 @@ Macros: #include_next #endif // __GNUC__ -#endif // _LIBCPP_LIMITS_H +#endif // _LIBCPP_LIMITS_H diff --git a/lib/libcxx/include/list b/lib/libcxx/include/list index a18514d74..23da5fcfc 100644 --- a/lib/libcxx/include/list +++ b/lib/libcxx/include/list @@ -181,17 +181,16 @@ template */ #include <__config> - -#include -#include +#include <__debug> +#include <__utility/forward.h> +#include #include #include -#include +#include +#include #include #include -#include <__debug> - #if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) #pragma GCC system_header #endif @@ -267,7 +266,7 @@ struct __list_node_base }; template -struct __list_node +struct _LIBCPP_STANDALONE_DEBUG __list_node : public __list_node_base<_Tp, _VoidPtr> { _Tp __value_; @@ -351,7 +350,7 @@ public: return *this; } -#endif // _LIBCPP_DEBUG_LEVEL == 2 +#endif // _LIBCPP_DEBUG_LEVEL == 2 _LIBCPP_INLINE_VISIBILITY reference operator*() const @@ -377,7 +376,7 @@ public: { #if _LIBCPP_DEBUG_LEVEL == 2 _LIBCPP_ASSERT(__get_const_db()->__dereferenceable(this), - "Attempted to increment non-incrementable list::iterator"); + "Attempted to increment a non-incrementable list::iterator"); #endif __ptr_ = __ptr_->__next_; return *this; @@ -390,7 +389,7 @@ public: { #if _LIBCPP_DEBUG_LEVEL == 2 _LIBCPP_ASSERT(__get_const_db()->__decrementable(this), - "Attempted to decrement non-decrementable list::iterator"); + "Attempted to decrement a non-decrementable list::iterator"); #endif __ptr_ = __ptr_->__prev_; return *this; @@ -479,7 +478,7 @@ public: return *this; } -#endif // _LIBCPP_DEBUG_LEVEL == 2 +#endif // _LIBCPP_DEBUG_LEVEL == 2 _LIBCPP_INLINE_VISIBILITY reference operator*() const { @@ -504,7 +503,7 @@ public: { #if _LIBCPP_DEBUG_LEVEL == 2 _LIBCPP_ASSERT(__get_const_db()->__dereferenceable(this), - "Attempted to increment non-incrementable list::const_iterator"); + "Attempted to increment a non-incrementable list::const_iterator"); #endif __ptr_ = __ptr_->__next_; return *this; @@ -517,7 +516,7 @@ public: { #if _LIBCPP_DEBUG_LEVEL == 2 _LIBCPP_ASSERT(__get_const_db()->__decrementable(this), - "Attempted to decrement non-decrementable list::const_iterator"); + "Attempted to decrement a non-decrementable list::const_iterator"); #endif __ptr_ = __ptr_->__prev_; return *this; @@ -895,7 +894,7 @@ public: typename enable_if<__is_cpp17_input_iterator<_InpIter>::value>::type* = 0); list(const list& __c); - list(const list& __c, const allocator_type& __a); + list(const list& __c, const __identity_t& __a); _LIBCPP_INLINE_VISIBILITY list& operator=(const list& __c); #ifndef _LIBCPP_CXX03_LANG @@ -906,7 +905,7 @@ public: list(list&& __c) _NOEXCEPT_(is_nothrow_move_constructible<__node_allocator>::value); _LIBCPP_INLINE_VISIBILITY - list(list&& __c, const allocator_type& __a); + list(list&& __c, const __identity_t& __a); _LIBCPP_INLINE_VISIBILITY list& operator=(list&& __c) _NOEXCEPT_( @@ -920,7 +919,7 @@ public: _LIBCPP_INLINE_VISIBILITY void assign(initializer_list __il) {assign(__il.begin(), __il.end());} -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG template void assign(_InpIter __f, _InpIter __l, @@ -1023,7 +1022,7 @@ public: _LIBCPP_INLINE_VISIBILITY iterator insert(const_iterator __p, initializer_list __il) {return insert(__p, __il.begin(), __il.end());} -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG void push_front(const value_type& __x); void push_back(const value_type& __x); @@ -1124,7 +1123,7 @@ public: bool __addable(const const_iterator* __i, ptrdiff_t __n) const; bool __subscriptable(const const_iterator* __i, ptrdiff_t __n) const; -#endif // _LIBCPP_DEBUG_LEVEL == 2 +#endif // _LIBCPP_DEBUG_LEVEL == 2 private: _LIBCPP_INLINE_VISIBILITY @@ -1144,18 +1143,18 @@ private: #ifndef _LIBCPP_HAS_NO_DEDUCTION_GUIDES template::value_type>, - class = typename enable_if<__is_allocator<_Alloc>::value, void>::type + class _Alloc = allocator<__iter_value_type<_InputIterator>>, + class = _EnableIf<__is_allocator<_Alloc>::value> > list(_InputIterator, _InputIterator) - -> list::value_type, _Alloc>; + -> list<__iter_value_type<_InputIterator>, _Alloc>; template::value, void>::type + class = _EnableIf<__is_allocator<_Alloc>::value> > list(_InputIterator, _InputIterator, _Alloc) - -> list::value_type, _Alloc>; + -> list<__iter_value_type<_InputIterator>, _Alloc>; #endif // Link in nodes [__f, __l] just prior to __p @@ -1288,7 +1287,7 @@ list<_Tp, _Alloc>::list(const list& __c) } template -list<_Tp, _Alloc>::list(const list& __c, const allocator_type& __a) +list<_Tp, _Alloc>::list(const list& __c, const __identity_t& __a) : base(__a) { #if _LIBCPP_DEBUG_LEVEL == 2 @@ -1335,7 +1334,7 @@ inline list<_Tp, _Alloc>::list(list&& __c) template inline -list<_Tp, _Alloc>::list(list&& __c, const allocator_type& __a) +list<_Tp, _Alloc>::list(list&& __c, const __identity_t& __a) : base(__a) { #if _LIBCPP_DEBUG_LEVEL == 2 @@ -1386,7 +1385,7 @@ list<_Tp, _Alloc>::__move_assign(list& __c, true_type) splice(end(), __c); } -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG template inline @@ -1495,7 +1494,7 @@ list<_Tp, _Alloc>::insert(const_iterator __p, size_type __n, const value_type& _ #ifndef _LIBCPP_NO_EXCEPTIONS try { -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS for (--__n; __n != 0; --__n, ++__e, ++__ds) { __hold.reset(__node_alloc_traits::allocate(__na, 1)); @@ -1523,7 +1522,7 @@ list<_Tp, _Alloc>::insert(const_iterator __p, size_type __n, const value_type& _ } throw; } -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS __link_nodes(__p.__ptr_, __r.__ptr_, __e.__ptr_); base::__sz() += __ds; } @@ -1561,7 +1560,7 @@ list<_Tp, _Alloc>::insert(const_iterator __p, _InpIter __f, _InpIter __l, #ifndef _LIBCPP_NO_EXCEPTIONS try { -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS for (++__f; __f != __l; ++__f, (void) ++__e, (void) ++__ds) { __hold.reset(__node_alloc_traits::allocate(__na, 1)); @@ -1589,7 +1588,7 @@ list<_Tp, _Alloc>::insert(const_iterator __p, _InpIter __f, _InpIter __l, } throw; } -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS __link_nodes(__p.__ptr_, __r.__ptr_, __e.__ptr_); base::__sz() += __ds; } @@ -1737,7 +1736,7 @@ list<_Tp, _Alloc>::insert(const_iterator __p, value_type&& __x) #endif } -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG template void @@ -1772,7 +1771,7 @@ template void list<_Tp, _Alloc>::pop_back() { - _LIBCPP_ASSERT(!empty(), "list::pop_back() called with empty list"); + _LIBCPP_ASSERT(!empty(), "list::pop_back() called on an empty list"); __node_allocator& __na = base::__node_alloc(); __link_pointer __n = base::__end_.__prev_; base::__unlink_nodes(__n, __n); @@ -1909,7 +1908,7 @@ list<_Tp, _Alloc>::resize(size_type __n) #ifndef _LIBCPP_NO_EXCEPTIONS try { -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS for (--__n; __n != 0; --__n, ++__e, ++__ds) { __hold.reset(__node_alloc_traits::allocate(__na, 1)); @@ -1937,7 +1936,7 @@ list<_Tp, _Alloc>::resize(size_type __n) } throw; } -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS __link_nodes_at_back(__r.__ptr_, __e.__ptr_); base::__sz() += __ds; } @@ -1967,7 +1966,7 @@ list<_Tp, _Alloc>::resize(size_type __n, const value_type& __x) #ifndef _LIBCPP_NO_EXCEPTIONS try { -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS for (--__n; __n != 0; --__n, ++__e, ++__ds) { __hold.reset(__node_alloc_traits::allocate(__na, 1)); @@ -1995,7 +1994,7 @@ list<_Tp, _Alloc>::resize(size_type __n, const value_type& __x) } throw; } -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS __link_nodes(base::__end_as_link(), __r.__ptr_, __e.__ptr_); base::__sz() += __ds; } @@ -2049,14 +2048,14 @@ list<_Tp, _Alloc>::splice(const_iterator __p, list& __c, const_iterator __i) { #if _LIBCPP_DEBUG_LEVEL == 2 _LIBCPP_ASSERT(__get_const_db()->__find_c_from_i(&__p) == this, - "list::splice(iterator, list, iterator) called with first iterator not" - " referring to this list"); + "list::splice(iterator, list, iterator) called with the first iterator" + " not referring to this list"); _LIBCPP_ASSERT(__get_const_db()->__find_c_from_i(&__i) == &__c, - "list::splice(iterator, list, iterator) called with second iterator not" - " referring to list argument"); + "list::splice(iterator, list, iterator) called with the second iterator" + " not referring to the list argument"); _LIBCPP_ASSERT(__get_const_db()->__dereferenceable(&__i), - "list::splice(iterator, list, iterator) called with second iterator not" - " dereferenceable"); + "list::splice(iterator, list, iterator) called with the second iterator" + " not dereferenceable"); #endif if (__p.__ptr_ != __i.__ptr_ && __p.__ptr_ != __i.__ptr_->__next_) { @@ -2098,7 +2097,10 @@ list<_Tp, _Alloc>::splice(const_iterator __p, list& __c, const_iterator __f, con " referring to this list"); _LIBCPP_ASSERT(__get_const_db()->__find_c_from_i(&__f) == &__c, "list::splice(iterator, list, iterator, iterator) called with second iterator not" - " referring to list argument"); + " referring to the list argument"); + _LIBCPP_ASSERT(__get_const_db()->__find_c_from_i(&__l) == &__c, + "list::splice(iterator, list, iterator, iterator) called with third iterator not" + " referring to the list argument"); if (this == &__c) { for (const_iterator __i = __f; __i != __l; ++__i) @@ -2412,7 +2414,7 @@ list<_Tp, _Alloc>::__subscriptable(const const_iterator*, ptrdiff_t) const return false; } -#endif // _LIBCPP_DEBUG_LEVEL == 2 +#endif // _LIBCPP_DEBUG_LEVEL == 2 template inline _LIBCPP_INLINE_VISIBILITY @@ -2489,4 +2491,4 @@ _LIBCPP_END_NAMESPACE_STD _LIBCPP_POP_MACROS -#endif // _LIBCPP_LIST +#endif // _LIBCPP_LIST diff --git a/lib/libcxx/include/locale b/lib/libcxx/include/locale index 9a705c77c..8e584005d 100644 --- a/lib/libcxx/include/locale +++ b/lib/libcxx/include/locale @@ -188,23 +188,28 @@ template class messages_byname; */ #include <__config> -#include <__locale> #include <__debug> +#include <__locale> #include -#include -#include -#include -#include -#include -#include #ifndef __APPLE__ -#include +# include #endif +#include #include #include -#include -#ifdef _LIBCPP_HAS_CATOPEN -#include +#include +#include +#include +#include +#include +#include + +#if defined(__unix__) || (defined(__APPLE__) && defined(__MACH__)) +// Most unix variants have catopen. These are the specific ones that don't. +# if !defined(__BIONIC__) && !defined(_NEWLIB_VERSION) +# define _LIBCPP_HAS_CATOPEN 1 +# include +# endif #endif #ifdef _LIBCPP_LOCALE__L_EXTENSIONS @@ -1453,10 +1458,12 @@ num_put<_CharT, _OutputIterator>::do_put(iter_type __s, ios_base& __iob, char __fmt[6] = {'%', 0}; const char* __len = "l"; this->__format_int(__fmt+1, __len, true, __iob.flags()); - const unsigned __nbuf = (numeric_limits::digits / 3) - + ((numeric_limits::digits % 3) != 0) - + ((__iob.flags() & ios_base::showbase) != 0) - + 2; + // Worst case is octal, with showbase enabled. Note that octal is always + // printed as an unsigned value. + _LIBCPP_CONSTEXPR const unsigned __nbuf + = (numeric_limits::digits / 3) // 1 char per 3 bits + + ((numeric_limits::digits % 3) != 0) // round up + + 2; // base prefix + terminating null character char __nar[__nbuf]; int __nc = __libcpp_snprintf_l(__nar, sizeof(__nar), _LIBCPP_GET_C_LOCALE, __fmt, __v); char* __ne = __nar + __nc; @@ -1480,10 +1487,12 @@ num_put<_CharT, _OutputIterator>::do_put(iter_type __s, ios_base& __iob, char __fmt[8] = {'%', 0}; const char* __len = "ll"; this->__format_int(__fmt+1, __len, true, __iob.flags()); - const unsigned __nbuf = (numeric_limits::digits / 3) - + ((numeric_limits::digits % 3) != 0) - + ((__iob.flags() & ios_base::showbase) != 0) - + 2; + // Worst case is octal, with showbase enabled. Note that octal is always + // printed as an unsigned value. + _LIBCPP_CONSTEXPR const unsigned __nbuf + = (numeric_limits::digits / 3) // 1 char per 3 bits + + ((numeric_limits::digits % 3) != 0) // round up + + 2; // base prefix + terminating null character char __nar[__nbuf]; int __nc = __libcpp_snprintf_l(__nar, sizeof(__nar), _LIBCPP_GET_C_LOCALE, __fmt, __v); char* __ne = __nar + __nc; @@ -1507,10 +1516,11 @@ num_put<_CharT, _OutputIterator>::do_put(iter_type __s, ios_base& __iob, char __fmt[6] = {'%', 0}; const char* __len = "l"; this->__format_int(__fmt+1, __len, false, __iob.flags()); - const unsigned __nbuf = (numeric_limits::digits / 3) - + ((numeric_limits::digits % 3) != 0) - + ((__iob.flags() & ios_base::showbase) != 0) - + 1; + // Worst case is octal, with showbase enabled. + _LIBCPP_CONSTEXPR const unsigned __nbuf + = (numeric_limits::digits / 3) // 1 char per 3 bits + + ((numeric_limits::digits % 3) != 0) // round up + + 2; // base prefix + terminating null character char __nar[__nbuf]; int __nc = __libcpp_snprintf_l(__nar, sizeof(__nar), _LIBCPP_GET_C_LOCALE, __fmt, __v); char* __ne = __nar + __nc; @@ -1534,10 +1544,11 @@ num_put<_CharT, _OutputIterator>::do_put(iter_type __s, ios_base& __iob, char __fmt[8] = {'%', 0}; const char* __len = "ll"; this->__format_int(__fmt+1, __len, false, __iob.flags()); - const unsigned __nbuf = (numeric_limits::digits / 3) - + ((numeric_limits::digits % 3) != 0) - + ((__iob.flags() & ios_base::showbase) != 0) - + 1; + // Worst case is octal, with showbase enabled. + _LIBCPP_CONSTEXPR const unsigned __nbuf + = (numeric_limits::digits / 3) // 1 char per 3 bits + + ((numeric_limits::digits % 3) != 0) // round up + + 2; // base prefix + terminating null character char __nar[__nbuf]; int __nc = __libcpp_snprintf_l(__nar, sizeof(__nar), _LIBCPP_GET_C_LOCALE, __fmt, __v); char* __ne = __nar + __nc; @@ -3567,7 +3578,7 @@ messages<_CharT>::do_open(const basic_string& __nm, const locale&) const __cat = static_cast((static_cast(__cat) >> 1)); return __cat; #else // !_LIBCPP_HAS_CATOPEN - _LIBCPP_UNUSED_VAR(__nm); + (void)__nm; return -1; #endif // _LIBCPP_HAS_CATOPEN } @@ -3591,9 +3602,9 @@ messages<_CharT>::do_get(catalog __c, int __set, int __msgid, __n, __n + _VSTD::strlen(__n)); return __w; #else // !_LIBCPP_HAS_CATOPEN - _LIBCPP_UNUSED_VAR(__c); - _LIBCPP_UNUSED_VAR(__set); - _LIBCPP_UNUSED_VAR(__msgid); + (void)__c; + (void)__set; + (void)__msgid; return __dflt; #endif // _LIBCPP_HAS_CATOPEN } @@ -3608,7 +3619,7 @@ messages<_CharT>::do_close(catalog __c) const nl_catd __cat = (nl_catd)__c; catclose(__cat); #else // !_LIBCPP_HAS_CATOPEN - _LIBCPP_UNUSED_VAR(__c); + (void)__c; #endif // _LIBCPP_HAS_CATOPEN } @@ -3748,7 +3759,7 @@ wstring_convert<_Codecvt, _Elem, _Wide_alloc, _Byte_alloc>:: __wc.__cvtptr_ = nullptr; } -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG template wstring_convert<_Codecvt, _Elem, _Wide_alloc, _Byte_alloc>::~wstring_convert() @@ -4376,4 +4387,4 @@ _LIBCPP_END_NAMESPACE_STD _LIBCPP_POP_MACROS -#endif // _LIBCPP_LOCALE +#endif // _LIBCPP_LOCALE diff --git a/lib/libcxx/include/locale.h b/lib/libcxx/include/locale.h index 81cfcee57..65b2bd5fd 100644 --- a/lib/libcxx/include/locale.h +++ b/lib/libcxx/include/locale.h @@ -45,4 +45,4 @@ Functions: #include_next -#endif // _LIBCPP_LOCALE_H +#endif // _LIBCPP_LOCALE_H diff --git a/lib/libcxx/include/map b/lib/libcxx/include/map index a27002c27..513a04dd7 100644 --- a/lib/libcxx/include/map +++ b/lib/libcxx/include/map @@ -43,7 +43,6 @@ public: typedef INSERT_RETURN_TYPE insert_return_type; // C++17 class value_compare - : public binary_function { friend class map; protected: @@ -51,6 +50,9 @@ public: value_compare(key_compare c); public: + typedef bool result_type; // deprecated in C++17, removed in C++20 + typedef value_type first_argument_type; // deprecated in C++17, removed in C++20 + typedef value_type second_argument_type; // deprecated in C++17, removed in C++20 bool operator()(const value_type& x, const value_type& y) const; }; @@ -191,10 +193,14 @@ public: iterator find(const K& x); // C++14 template const_iterator find(const K& x) const; // C++14 + template size_type count(const K& x) const; // C++14 size_type count(const key_type& k) const; - bool contains(const key_type& x) const; // C++20 + + bool contains(const key_type& x) const; // C++20 + template bool contains(const K& x) const; // C++20 + iterator lower_bound(const key_type& k); const_iterator lower_bound(const key_type& k) const; template @@ -283,13 +289,15 @@ public: typedef unspecified node_type; // C++17 class value_compare - : public binary_function { friend class multimap; protected: key_compare comp; value_compare(key_compare c); public: + typedef bool result_type; // deprecated in C++17, removed in C++20 + typedef value_type first_argument_type; // deprecated in C++17, removed in C++20 + typedef value_type second_argument_type; // deprecated in C++17, removed in C++20 bool operator()(const value_type& x, const value_type& y) const; }; @@ -406,10 +414,14 @@ public: iterator find(const K& x); // C++14 template const_iterator find(const K& x) const; // C++14 + template size_type count(const K& x) const; // C++14 size_type count(const key_type& k) const; - bool contains(const key_type& x) const; // C++20 + + bool contains(const key_type& x) const; // C++20 + template bool contains(const K& x) const; // C++20 + iterator lower_bound(const key_type& k); const_iterator lower_bound(const key_type& k) const; template @@ -478,14 +490,18 @@ erase_if(multimap& c, Predicate pred); // C++20 */ #include <__config> -#include <__tree> +#include <__debug> +#include <__functional/is_transparent.h> #include <__node_handle> -#include -#include -#include +#include <__tree> +#include <__utility/forward.h> +#include #include #include +#include // __libcpp_erase_if_container +#include #include +#include #include #if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) @@ -633,7 +649,7 @@ public: { __x.__value_constructed = false; } -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG _LIBCPP_INLINE_VISIBILITY void operator()(pointer __p) _NOEXCEPT @@ -656,7 +672,7 @@ template class __map_const_iterator; #ifndef _LIBCPP_CXX03_LANG template -struct __value_type +struct _LIBCPP_STANDALONE_DEBUG __value_type { typedef _Key key_type; typedef _Tp mapped_type; @@ -904,23 +920,32 @@ public: typedef _Key key_type; typedef _Tp mapped_type; typedef pair value_type; - typedef typename __identity<_Compare>::type key_compare; - typedef typename __identity<_Allocator>::type allocator_type; + typedef __identity_t<_Compare> key_compare; + typedef __identity_t<_Allocator> allocator_type; typedef value_type& reference; typedef const value_type& const_reference; static_assert((is_same::value), "Allocator::value_type must be same type as value_type"); +_LIBCPP_SUPPRESS_DEPRECATED_PUSH class _LIBCPP_TEMPLATE_VIS value_compare +#if defined(_LIBCPP_ENABLE_CXX20_REMOVED_BINDER_TYPEDEFS) : public binary_function +#endif { +_LIBCPP_SUPPRESS_DEPRECATED_POP friend class map; protected: key_compare comp; _LIBCPP_INLINE_VISIBILITY value_compare(key_compare c) : comp(c) {} public: +#if _LIBCPP_STD_VER <= 17 || defined(_LIBCPP_ENABLE_CXX20_REMOVED_BINDER_TYPEDEFS) + _LIBCPP_DEPRECATED_IN_CXX17 typedef bool result_type; + _LIBCPP_DEPRECATED_IN_CXX17 typedef value_type first_argument_type; + _LIBCPP_DEPRECATED_IN_CXX17 typedef value_type second_argument_type; +#endif _LIBCPP_INLINE_VISIBILITY bool operator()(const value_type& __x, const value_type& __y) const {return comp(__x.first, __y.first);} @@ -1071,7 +1096,7 @@ public: return *this; } -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG _LIBCPP_INLINE_VISIBILITY explicit map(const allocator_type& __a) @@ -1168,7 +1193,7 @@ public: iterator insert(const_iterator __pos, _Pp&& __p) {return __tree_.__insert_unique(__pos.__i_, _VSTD::forward<_Pp>(__p));} -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG _LIBCPP_INLINE_VISIBILITY pair @@ -1404,6 +1429,10 @@ public: #if _LIBCPP_STD_VER > 17 _LIBCPP_INLINE_VISIBILITY bool contains(const key_type& __k) const {return find(__k) != end();} + template + _LIBCPP_INLINE_VISIBILITY + typename enable_if<__is_transparent<_Compare, _K2>::value, bool>::type + contains(const _K2& __k) const { return find(__k) != end(); } #endif // _LIBCPP_STD_VER > 17 _LIBCPP_INLINE_VISIBILITY @@ -1565,7 +1594,7 @@ map<_Key, _Tp, _Compare, _Allocator>::operator[](const key_type& __k) return __r->__value_.__get_value().second; } -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG template _Tp& @@ -1660,7 +1689,7 @@ template ::size_type erase_if(map<_Key, _Tp, _Compare, _Allocator>& __c, _Predicate __pred) { - return __libcpp_erase_if_container(__c, __pred); + return _VSTD::__libcpp_erase_if_container(__c, __pred); } #endif @@ -1674,17 +1703,21 @@ public: typedef _Key key_type; typedef _Tp mapped_type; typedef pair value_type; - typedef typename __identity<_Compare>::type key_compare; - typedef typename __identity<_Allocator>::type allocator_type; + typedef __identity_t<_Compare> key_compare; + typedef __identity_t<_Allocator> allocator_type; typedef value_type& reference; typedef const value_type& const_reference; static_assert((is_same::value), "Allocator::value_type must be same type as value_type"); +_LIBCPP_SUPPRESS_DEPRECATED_PUSH class _LIBCPP_TEMPLATE_VIS value_compare +#if defined(_LIBCPP_ENABLE_CXX20_REMOVED_BINDER_TYPEDEFS) : public binary_function +#endif { +_LIBCPP_SUPPRESS_DEPRECATED_POP friend class multimap; protected: key_compare comp; @@ -1692,6 +1725,11 @@ public: _LIBCPP_INLINE_VISIBILITY value_compare(key_compare c) : comp(c) {} public: +#if _LIBCPP_STD_VER <= 17 || defined(_LIBCPP_ENABLE_CXX20_REMOVED_BINDER_TYPEDEFS) + _LIBCPP_DEPRECATED_IN_CXX17 typedef bool result_type; + _LIBCPP_DEPRECATED_IN_CXX17 typedef value_type first_argument_type; + _LIBCPP_DEPRECATED_IN_CXX17 typedef value_type second_argument_type; +#endif _LIBCPP_INLINE_VISIBILITY bool operator()(const value_type& __x, const value_type& __y) const {return comp(__x.first, __y.first);} @@ -1842,7 +1880,7 @@ public: return *this; } -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG _LIBCPP_INLINE_VISIBILITY explicit multimap(const allocator_type& __a) @@ -1945,7 +1983,7 @@ public: void insert(initializer_list __il) {insert(__il.begin(), __il.end());} -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG _LIBCPP_INLINE_VISIBILITY iterator insert(const value_type& __v) {return __tree_.__insert_multi(__v);} @@ -2070,6 +2108,10 @@ public: #if _LIBCPP_STD_VER > 17 _LIBCPP_INLINE_VISIBILITY bool contains(const key_type& __k) const {return find(__k) != end();} + template + _LIBCPP_INLINE_VISIBILITY + typename enable_if<__is_transparent<_Compare, _K2>::value, bool>::type + contains(const _K2& __k) const { return find(__k) != end(); } #endif // _LIBCPP_STD_VER > 17 _LIBCPP_INLINE_VISIBILITY @@ -2246,10 +2288,10 @@ inline _LIBCPP_INLINE_VISIBILITY typename multimap<_Key, _Tp, _Compare, _Allocator>::size_type erase_if(multimap<_Key, _Tp, _Compare, _Allocator>& __c, _Predicate __pred) { - return __libcpp_erase_if_container(__c, __pred); + return _VSTD::__libcpp_erase_if_container(__c, __pred); } #endif _LIBCPP_END_NAMESPACE_STD -#endif // _LIBCPP_MAP +#endif // _LIBCPP_MAP diff --git a/lib/libcxx/include/math.h b/lib/libcxx/include/math.h index 1603d5748..77762d554 100644 --- a/lib/libcxx/include/math.h +++ b/lib/libcxx/include/math.h @@ -318,7 +318,11 @@ _LIBCPP_INLINE_VISIBILITY bool __libcpp_signbit(_A1 __lcpp_x) _NOEXCEPT { +#if __has_builtin(__builtin_signbit) + return __builtin_signbit(__lcpp_x); +#else return signbit(__lcpp_x); +#endif } #undef signbit @@ -369,7 +373,7 @@ typename std::enable_if< signbit(_A1) _NOEXCEPT { return false; } -#endif // signbit +#endif // signbit // fpclassify @@ -380,7 +384,12 @@ _LIBCPP_INLINE_VISIBILITY int __libcpp_fpclassify(_A1 __lcpp_x) _NOEXCEPT { +#if __has_builtin(__builtin_fpclassify) + return __builtin_fpclassify(FP_NAN, FP_INFINITE, FP_NORMAL, FP_SUBNORMAL, + FP_ZERO, __lcpp_x); +#else return fpclassify(__lcpp_x); +#endif } #undef fpclassify @@ -415,7 +424,7 @@ typename std::enable_if::value, int>::type fpclassify(_A1 __lcpp_x) _NOEXCEPT { return __lcpp_x == 0 ? FP_ZERO : FP_NORMAL; } -#endif // fpclassify +#endif // fpclassify // isfinite @@ -426,7 +435,11 @@ _LIBCPP_INLINE_VISIBILITY bool __libcpp_isfinite(_A1 __lcpp_x) _NOEXCEPT { +#if __has_builtin(__builtin_isfinite) + return __builtin_isfinite(__lcpp_x); +#else return isfinite(__lcpp_x); +#endif } #undef isfinite @@ -449,7 +462,7 @@ typename std::enable_if< isfinite(_A1) _NOEXCEPT { return true; } -#endif // isfinite +#endif // isfinite // isinf @@ -460,7 +473,11 @@ _LIBCPP_INLINE_VISIBILITY bool __libcpp_isinf(_A1 __lcpp_x) _NOEXCEPT { +#if __has_builtin(__builtin_isinf) + return __builtin_isinf(__lcpp_x); +#else return isinf(__lcpp_x); +#endif } #undef isinf @@ -497,7 +514,7 @@ bool isinf(long double __lcpp_x) _NOEXCEPT { return __libcpp_isinf(__lcpp_x); } #endif -#endif // isinf +#endif // isinf // isnan @@ -545,7 +562,7 @@ bool isnan(long double __lcpp_x) _NOEXCEPT { return __libcpp_isnan(__lcpp_x); } #endif -#endif // isnan +#endif // isnan // isnormal @@ -556,7 +573,11 @@ _LIBCPP_INLINE_VISIBILITY bool __libcpp_isnormal(_A1 __lcpp_x) _NOEXCEPT { +#if __has_builtin(__builtin_isnormal) + return __builtin_isnormal(__lcpp_x); +#else return isnormal(__lcpp_x); +#endif } #undef isnormal @@ -575,7 +596,7 @@ typename std::enable_if::value, bool>::type isnormal(_A1 __lcpp_x) _NOEXCEPT { return __lcpp_x != 0; } -#endif // isnormal +#endif // isnormal // isgreater @@ -605,7 +626,7 @@ isgreater(_A1 __lcpp_x, _A2 __lcpp_y) _NOEXCEPT return __libcpp_isgreater((type)__lcpp_x, (type)__lcpp_y); } -#endif // isgreater +#endif // isgreater // isgreaterequal @@ -635,7 +656,7 @@ isgreaterequal(_A1 __lcpp_x, _A2 __lcpp_y) _NOEXCEPT return __libcpp_isgreaterequal((type)__lcpp_x, (type)__lcpp_y); } -#endif // isgreaterequal +#endif // isgreaterequal // isless @@ -665,7 +686,7 @@ isless(_A1 __lcpp_x, _A2 __lcpp_y) _NOEXCEPT return __libcpp_isless((type)__lcpp_x, (type)__lcpp_y); } -#endif // isless +#endif // isless // islessequal @@ -695,7 +716,7 @@ islessequal(_A1 __lcpp_x, _A2 __lcpp_y) _NOEXCEPT return __libcpp_islessequal((type)__lcpp_x, (type)__lcpp_y); } -#endif // islessequal +#endif // islessequal // islessgreater @@ -725,7 +746,7 @@ islessgreater(_A1 __lcpp_x, _A2 __lcpp_y) _NOEXCEPT return __libcpp_islessgreater((type)__lcpp_x, (type)__lcpp_y); } -#endif // islessgreater +#endif // islessgreater // isunordered @@ -755,7 +776,7 @@ isunordered(_A1 __lcpp_x, _A2 __lcpp_y) _NOEXCEPT return __libcpp_isunordered((type)__lcpp_x, (type)__lcpp_y); } -#endif // isunordered +#endif // isunordered // abs // @@ -1099,16 +1120,43 @@ cbrt(_A1 __lcpp_x) _NOEXCEPT {return ::cbrt((double)__lcpp_x);} // copysign -inline _LIBCPP_INLINE_VISIBILITY float copysign(float __lcpp_x, - float __lcpp_y) _NOEXCEPT { +#if __has_builtin(__builtin_copysignf) +_LIBCPP_CONSTEXPR +#endif +inline _LIBCPP_INLINE_VISIBILITY float __libcpp_copysign(float __lcpp_x, float __lcpp_y) _NOEXCEPT { +#if __has_builtin(__builtin_copysignf) + return __builtin_copysignf(__lcpp_x, __lcpp_y); +#else return ::copysignf(__lcpp_x, __lcpp_y); +#endif +} + +#if __has_builtin(__builtin_copysign) +_LIBCPP_CONSTEXPR +#endif +inline _LIBCPP_INLINE_VISIBILITY double __libcpp_copysign(double __lcpp_x, double __lcpp_y) _NOEXCEPT { +#if __has_builtin(__builtin_copysign) + return __builtin_copysign(__lcpp_x, __lcpp_y); +#else + return ::copysign(__lcpp_x, __lcpp_y); +#endif } -inline _LIBCPP_INLINE_VISIBILITY long double -copysign(long double __lcpp_x, long double __lcpp_y) _NOEXCEPT { + +#if __has_builtin(__builtin_copysignl) +_LIBCPP_CONSTEXPR +#endif +inline _LIBCPP_INLINE_VISIBILITY long double __libcpp_copysign(long double __lcpp_x, long double __lcpp_y) _NOEXCEPT { +#if __has_builtin(__builtin_copysignl) + return __builtin_copysignl(__lcpp_x, __lcpp_y); +#else return ::copysignl(__lcpp_x, __lcpp_y); +#endif } template +#if __has_builtin(__builtin_copysign) +_LIBCPP_CONSTEXPR +#endif inline _LIBCPP_INLINE_VISIBILITY typename std::_EnableIf < @@ -1116,12 +1164,35 @@ typename std::_EnableIf std::is_arithmetic<_A2>::value, std::__promote<_A1, _A2> >::type -copysign(_A1 __lcpp_x, _A2 __lcpp_y) _NOEXCEPT -{ +__libcpp_copysign(_A1 __lcpp_x, _A2 __lcpp_y) _NOEXCEPT { typedef typename std::__promote<_A1, _A2>::type __result_type; static_assert((!(std::_IsSame<_A1, __result_type>::value && std::_IsSame<_A2, __result_type>::value)), ""); +#if __has_builtin(__builtin_copysign) + return __builtin_copysign((__result_type)__lcpp_x, (__result_type)__lcpp_y); +#else return ::copysign((__result_type)__lcpp_x, (__result_type)__lcpp_y); +#endif +} + +inline _LIBCPP_INLINE_VISIBILITY float copysign(float __lcpp_x, float __lcpp_y) _NOEXCEPT { + return ::__libcpp_copysign(__lcpp_x, __lcpp_y); +} + +inline _LIBCPP_INLINE_VISIBILITY long double copysign(long double __lcpp_x, long double __lcpp_y) _NOEXCEPT { + return ::__libcpp_copysign(__lcpp_x, __lcpp_y); +} + +template +inline _LIBCPP_INLINE_VISIBILITY +typename std::_EnableIf +< + std::is_arithmetic<_A1>::value && + std::is_arithmetic<_A2>::value, + std::__promote<_A1, _A2> +>::type + copysign(_A1 __lcpp_x, _A2 __lcpp_y) _NOEXCEPT { + return ::__libcpp_copysign(__lcpp_x, __lcpp_y); } // erf @@ -1187,8 +1258,22 @@ fdim(_A1 __lcpp_x, _A2 __lcpp_y) _NOEXCEPT // fma -inline _LIBCPP_INLINE_VISIBILITY float fma(float __lcpp_x, float __lcpp_y, float __lcpp_z) _NOEXCEPT {return ::fmaf(__lcpp_x, __lcpp_y, __lcpp_z);} -inline _LIBCPP_INLINE_VISIBILITY long double fma(long double __lcpp_x, long double __lcpp_y, long double __lcpp_z) _NOEXCEPT {return ::fmal(__lcpp_x, __lcpp_y, __lcpp_z);} +inline _LIBCPP_INLINE_VISIBILITY float fma(float __lcpp_x, float __lcpp_y, float __lcpp_z) _NOEXCEPT +{ +#if __has_builtin(__builtin_fmaf) + return __builtin_fmaf(__lcpp_x, __lcpp_y, __lcpp_z); +#else + return ::fmaf(__lcpp_x, __lcpp_y, __lcpp_z); +#endif +} +inline _LIBCPP_INLINE_VISIBILITY long double fma(long double __lcpp_x, long double __lcpp_y, long double __lcpp_z) _NOEXCEPT +{ +#if __has_builtin(__builtin_fmal) + return __builtin_fmal(__lcpp_x, __lcpp_y, __lcpp_z); +#else + return ::fmal(__lcpp_x, __lcpp_y, __lcpp_z); +#endif +} template inline _LIBCPP_INLINE_VISIBILITY @@ -1205,7 +1290,11 @@ fma(_A1 __lcpp_x, _A2 __lcpp_y, _A3 __lcpp_z) _NOEXCEPT static_assert((!(std::_IsSame<_A1, __result_type>::value && std::_IsSame<_A2, __result_type>::value && std::_IsSame<_A3, __result_type>::value)), ""); +#if __has_builtin(__builtin_fma) + return __builtin_fma((__result_type)__lcpp_x, (__result_type)__lcpp_y, (__result_type)__lcpp_z); +#else return ::fma((__result_type)__lcpp_x, (__result_type)__lcpp_y, (__result_type)__lcpp_z); +#endif } // fmax @@ -1293,23 +1382,65 @@ lgamma(_A1 __lcpp_x) _NOEXCEPT {return ::lgamma((double)__lcpp_x);} // llrint -inline _LIBCPP_INLINE_VISIBILITY long long llrint(float __lcpp_x) _NOEXCEPT {return ::llrintf(__lcpp_x);} -inline _LIBCPP_INLINE_VISIBILITY long long llrint(long double __lcpp_x) _NOEXCEPT {return ::llrintl(__lcpp_x);} +inline _LIBCPP_INLINE_VISIBILITY long long llrint(float __lcpp_x) _NOEXCEPT +{ +#if __has_builtin(__builtin_llrintf) + return __builtin_llrintf(__lcpp_x); +#else + return ::llrintf(__lcpp_x); +#endif +} +inline _LIBCPP_INLINE_VISIBILITY long long llrint(long double __lcpp_x) _NOEXCEPT +{ +#if __has_builtin(__builtin_llrintl) + return __builtin_llrintl(__lcpp_x); +#else + return ::llrintl(__lcpp_x); +#endif +} template inline _LIBCPP_INLINE_VISIBILITY typename std::enable_if::value, long long>::type -llrint(_A1 __lcpp_x) _NOEXCEPT {return ::llrint((double)__lcpp_x);} +llrint(_A1 __lcpp_x) _NOEXCEPT +{ +#if __has_builtin(__builtin_llrint) + return __builtin_llrint((double)__lcpp_x); +#else + return ::llrint((double)__lcpp_x); +#endif +} // llround -inline _LIBCPP_INLINE_VISIBILITY long long llround(float __lcpp_x) _NOEXCEPT {return ::llroundf(__lcpp_x);} -inline _LIBCPP_INLINE_VISIBILITY long long llround(long double __lcpp_x) _NOEXCEPT {return ::llroundl(__lcpp_x);} +inline _LIBCPP_INLINE_VISIBILITY long long llround(float __lcpp_x) _NOEXCEPT +{ +#if __has_builtin(__builtin_llroundf) + return __builtin_llroundf(__lcpp_x); +#else + return ::llroundf(__lcpp_x); +#endif +} +inline _LIBCPP_INLINE_VISIBILITY long long llround(long double __lcpp_x) _NOEXCEPT +{ +#if __has_builtin(__builtin_llroundl) + return __builtin_llroundl(__lcpp_x); +#else + return ::llroundl(__lcpp_x); +#endif +} template inline _LIBCPP_INLINE_VISIBILITY typename std::enable_if::value, long long>::type -llround(_A1 __lcpp_x) _NOEXCEPT {return ::llround((double)__lcpp_x);} +llround(_A1 __lcpp_x) _NOEXCEPT +{ +#if __has_builtin(__builtin_llround) + return __builtin_llround((double)__lcpp_x); +#else + return ::llround((double)__lcpp_x); +#endif +} // log1p @@ -1343,23 +1474,65 @@ logb(_A1 __lcpp_x) _NOEXCEPT {return ::logb((double)__lcpp_x);} // lrint -inline _LIBCPP_INLINE_VISIBILITY long lrint(float __lcpp_x) _NOEXCEPT {return ::lrintf(__lcpp_x);} -inline _LIBCPP_INLINE_VISIBILITY long lrint(long double __lcpp_x) _NOEXCEPT {return ::lrintl(__lcpp_x);} +inline _LIBCPP_INLINE_VISIBILITY long lrint(float __lcpp_x) _NOEXCEPT +{ +#if __has_builtin(__builtin_lrintf) + return __builtin_lrintf(__lcpp_x); +#else + return ::lrintf(__lcpp_x); +#endif +} +inline _LIBCPP_INLINE_VISIBILITY long lrint(long double __lcpp_x) _NOEXCEPT +{ +#if __has_builtin(__builtin_lrintl) + return __builtin_lrintl(__lcpp_x); +#else + return ::lrintl(__lcpp_x); +#endif +} template inline _LIBCPP_INLINE_VISIBILITY typename std::enable_if::value, long>::type -lrint(_A1 __lcpp_x) _NOEXCEPT {return ::lrint((double)__lcpp_x);} +lrint(_A1 __lcpp_x) _NOEXCEPT +{ +#if __has_builtin(__builtin_lrint) + return __builtin_lrint((double)__lcpp_x); +#else + return ::lrint((double)__lcpp_x); +#endif +} // lround -inline _LIBCPP_INLINE_VISIBILITY long lround(float __lcpp_x) _NOEXCEPT {return ::lroundf(__lcpp_x);} -inline _LIBCPP_INLINE_VISIBILITY long lround(long double __lcpp_x) _NOEXCEPT {return ::lroundl(__lcpp_x);} +inline _LIBCPP_INLINE_VISIBILITY long lround(float __lcpp_x) _NOEXCEPT +{ +#if __has_builtin(__builtin_lroundf) + return __builtin_lroundf(__lcpp_x); +#else + return ::lroundf(__lcpp_x); +#endif +} +inline _LIBCPP_INLINE_VISIBILITY long lround(long double __lcpp_x) _NOEXCEPT +{ +#if __has_builtin(__builtin_lroundl) + return __builtin_lroundl(__lcpp_x); +#else + return ::lroundl(__lcpp_x); +#endif +} template inline _LIBCPP_INLINE_VISIBILITY typename std::enable_if::value, long>::type -lround(_A1 __lcpp_x) _NOEXCEPT {return ::lround((double)__lcpp_x);} +lround(_A1 __lcpp_x) _NOEXCEPT +{ +#if __has_builtin(__builtin_lround) + return __builtin_lround((double)__lcpp_x); +#else + return ::lround((double)__lcpp_x); +#endif +} // nan @@ -1448,23 +1621,65 @@ remquo(_A1 __lcpp_x, _A2 __lcpp_y, int* __lcpp_z) _NOEXCEPT // rint -inline _LIBCPP_INLINE_VISIBILITY float rint(float __lcpp_x) _NOEXCEPT {return ::rintf(__lcpp_x);} -inline _LIBCPP_INLINE_VISIBILITY long double rint(long double __lcpp_x) _NOEXCEPT {return ::rintl(__lcpp_x);} +inline _LIBCPP_INLINE_VISIBILITY float rint(float __lcpp_x) _NOEXCEPT +{ +#if __has_builtin(__builtin_rintf) + return __builtin_rintf(__lcpp_x); +#else + return ::rintf(__lcpp_x); +#endif +} +inline _LIBCPP_INLINE_VISIBILITY long double rint(long double __lcpp_x) _NOEXCEPT +{ +#if __has_builtin(__builtin_rintl) + return __builtin_rintl(__lcpp_x); +#else + return ::rintl(__lcpp_x); +#endif +} template inline _LIBCPP_INLINE_VISIBILITY typename std::enable_if::value, double>::type -rint(_A1 __lcpp_x) _NOEXCEPT {return ::rint((double)__lcpp_x);} +rint(_A1 __lcpp_x) _NOEXCEPT +{ +#if __has_builtin(__builtin_rint) + return __builtin_rint((double)__lcpp_x); +#else + return ::rint((double)__lcpp_x); +#endif +} // round -inline _LIBCPP_INLINE_VISIBILITY float round(float __lcpp_x) _NOEXCEPT {return ::roundf(__lcpp_x);} -inline _LIBCPP_INLINE_VISIBILITY long double round(long double __lcpp_x) _NOEXCEPT {return ::roundl(__lcpp_x);} +inline _LIBCPP_INLINE_VISIBILITY float round(float __lcpp_x) _NOEXCEPT +{ +#if __has_builtin(__builtin_round) + return __builtin_round(__lcpp_x); +#else + return ::round(__lcpp_x); +#endif +} +inline _LIBCPP_INLINE_VISIBILITY long double round(long double __lcpp_x) _NOEXCEPT +{ +#if __has_builtin(__builtin_roundl) + return __builtin_roundl(__lcpp_x); +#else + return ::roundl(__lcpp_x); +#endif +} template inline _LIBCPP_INLINE_VISIBILITY typename std::enable_if::value, double>::type -round(_A1 __lcpp_x) _NOEXCEPT {return ::round((double)__lcpp_x);} +round(_A1 __lcpp_x) _NOEXCEPT +{ +#if __has_builtin(__builtin_round) + return __builtin_round((double)__lcpp_x); +#else + return ::round((double)__lcpp_x); +#endif +} // scalbln @@ -1498,13 +1713,34 @@ tgamma(_A1 __lcpp_x) _NOEXCEPT {return ::tgamma((double)__lcpp_x);} // trunc -inline _LIBCPP_INLINE_VISIBILITY float trunc(float __lcpp_x) _NOEXCEPT {return ::truncf(__lcpp_x);} -inline _LIBCPP_INLINE_VISIBILITY long double trunc(long double __lcpp_x) _NOEXCEPT {return ::truncl(__lcpp_x);} +inline _LIBCPP_INLINE_VISIBILITY float trunc(float __lcpp_x) _NOEXCEPT +{ +#if __has_builtin(__builtin_trunc) + return __builtin_trunc(__lcpp_x); +#else + return ::trunc(__lcpp_x); +#endif +} +inline _LIBCPP_INLINE_VISIBILITY long double trunc(long double __lcpp_x) _NOEXCEPT +{ +#if __has_builtin(__builtin_truncl) + return __builtin_truncl(__lcpp_x); +#else + return ::truncl(__lcpp_x); +#endif +} template inline _LIBCPP_INLINE_VISIBILITY typename std::enable_if::value, double>::type -trunc(_A1 __lcpp_x) _NOEXCEPT {return ::trunc((double)__lcpp_x);} +trunc(_A1 __lcpp_x) _NOEXCEPT +{ +#if __has_builtin(__builtin_trunc) + return __builtin_trunc((double)__lcpp_x); +#else + return ::trunc((double)__lcpp_x); +#endif +} } // extern "C++" @@ -1524,4 +1760,4 @@ trunc(_A1 __lcpp_x) _NOEXCEPT {return ::trunc((double)__lcpp_x);} #include_next #endif -#endif // _LIBCPP_MATH_H +#endif // _LIBCPP_MATH_H diff --git a/lib/libcxx/include/memory b/lib/libcxx/include/memory index 39d0f5bee..fdb894353 100644 --- a/lib/libcxx/include/memory +++ b/lib/libcxx/include/memory @@ -99,7 +99,7 @@ struct allocator_traits }; template <> -class allocator // deprecated in C++17, removed in C++20 +class allocator // removed in C++20 { public: typedef void* pointer; @@ -153,14 +153,17 @@ template bool operator!=(const allocator&, const allocator&) noexcept; // constexpr in C++20 template -class raw_storage_iterator - : public iterator // purposefully not C++03 +class raw_storage_iterator // deprecated in C++17, removed in C++20 + : public iterator // until C++17 { public: + typedef output_iterator_tag iterator_category; + typedef void value_type; + typedef void difference_type; // until C++20 + typedef ptrdiff_t difference_type; // since C++20 + typedef void pointer; + typedef void reference; + explicit raw_storage_iterator(OutputIterator x); raw_storage_iterator& operator*(); raw_storage_iterator& operator=(const T& element); @@ -651,12 +654,12 @@ template inline constexpr bool uses_allocator_v = uses_allocator::value; // Pointer safety -enum class pointer_safety { relaxed, preferred, strict }; -void declare_reachable(void *p); -template T *undeclare_reachable(T *p); -void declare_no_pointers(char *p, size_t n); -void undeclare_no_pointers(char *p, size_t n); -pointer_safety get_pointer_safety() noexcept; +enum class pointer_safety { relaxed, preferred, strict }; // since C++11 +void declare_reachable(void *p); // since C++11 +template T *undeclare_reachable(T *p); // since C++11 +void declare_no_pointers(char *p, size_t n); // since C++11 +void undeclare_no_pointers(char *p, size_t n); // since C++11 +pointer_safety get_pointer_safety() noexcept; // since C++11 void* align(size_t alignment, size_t size, void*& ptr, size_t& space); @@ -665,29 +668,40 @@ void* align(size_t alignment, size_t size, void*& ptr, size_t& space); */ #include <__config> -#include <__availability> -#include -#include +#include <__functional_base> +#include <__memory/addressof.h> +#include <__memory/allocation_guard.h> +#include <__memory/allocator.h> +#include <__memory/allocator_arg_t.h> +#include <__memory/allocator_traits.h> +#include <__memory/compressed_pair.h> +#include <__memory/construct_at.h> +#include <__memory/pointer_safety.h> +#include <__memory/pointer_traits.h> +#include <__memory/raw_storage_iterator.h> +#include <__memory/shared_ptr.h> +#include <__memory/temporary_buffer.h> +#include <__memory/uninitialized_algorithms.h> +#include <__memory/unique_ptr.h> +#include <__memory/uses_allocator.h> +#include #include #include -#include -#include -#include -#include -#include <__functional_base> +#include #include -#include +#include +#include #include -#include -#include <__memory/allocator_traits.h> -#include <__memory/base.h> -#include <__memory/pointer_traits.h> -#include <__memory/utilities.h> -#if !defined(_LIBCPP_HAS_NO_ATOMIC_HEADER) -# include -#endif +#include +#include +#include +#include #include +#if _LIBCPP_STD_VER <= 14 || defined(_LIBCPP_ENABLE_CXX17_REMOVED_AUTO_PTR) +# include <__memory/auto_ptr.h> +#endif + #if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) #pragma GCC system_header #endif @@ -698,286 +712,6 @@ _LIBCPP_PUSH_MACROS _LIBCPP_BEGIN_NAMESPACE_STD -template -inline _LIBCPP_INLINE_VISIBILITY -_ValueType __libcpp_relaxed_load(_ValueType const* __value) { -#if !defined(_LIBCPP_HAS_NO_THREADS) && \ - defined(__ATOMIC_RELAXED) && \ - (__has_builtin(__atomic_load_n) || defined(_LIBCPP_COMPILER_GCC)) - return __atomic_load_n(__value, __ATOMIC_RELAXED); -#else - return *__value; -#endif -} - -template -inline _LIBCPP_INLINE_VISIBILITY -_ValueType __libcpp_acquire_load(_ValueType const* __value) { -#if !defined(_LIBCPP_HAS_NO_THREADS) && \ - defined(__ATOMIC_ACQUIRE) && \ - (__has_builtin(__atomic_load_n) || defined(_LIBCPP_COMPILER_GCC)) - return __atomic_load_n(__value, __ATOMIC_ACQUIRE); -#else - return *__value; -#endif -} - -template struct __to_address_helper; - -template <> struct __to_address_helper { - template - using __return_type = decltype(pointer_traits<_Pointer>::to_address(_VSTD::declval())); - - template - _LIBCPP_CONSTEXPR - static __return_type<_Pointer> - __do_it(const _Pointer &__p) _NOEXCEPT { return pointer_traits<_Pointer>::to_address(__p); } -}; - -template -using __choose_to_address = __to_address_helper<_IsValidExpansion<__to_address_helper<_Dummy>::template __return_type, _Pointer>::value>; - - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR -_Tp* -__to_address(_Tp* __p) _NOEXCEPT -{ - static_assert(!is_function<_Tp>::value, "_Tp is a function type"); - return __p; -} - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR -typename __choose_to_address<_Pointer>::template __return_type<_Pointer> -__to_address(const _Pointer& __p) _NOEXCEPT { - return __choose_to_address<_Pointer>::__do_it(__p); -} - -template <> struct __to_address_helper { - template - using __return_type = typename pointer_traits<_Pointer>::element_type*; - - template - _LIBCPP_CONSTEXPR - static __return_type<_Pointer> - __do_it(const _Pointer &__p) _NOEXCEPT { return _VSTD::__to_address(__p.operator->()); } -}; - - -#if _LIBCPP_STD_VER > 17 -template -inline _LIBCPP_INLINE_VISIBILITY constexpr -_Tp* -to_address(_Tp* __p) _NOEXCEPT -{ - static_assert(!is_function_v<_Tp>, "_Tp is a function type"); - return __p; -} - -template -inline _LIBCPP_INLINE_VISIBILITY constexpr -auto -to_address(const _Pointer& __p) _NOEXCEPT -{ - return _VSTD::__to_address(__p); -} -#endif - -template class allocator; - -#if _LIBCPP_STD_VER <= 17 || defined(_LIBCPP_ENABLE_CXX20_REMOVED_ALLOCATOR_MEMBERS) -template <> -class _LIBCPP_TEMPLATE_VIS _LIBCPP_DEPRECATED_IN_CXX17 allocator -{ -public: - typedef void* pointer; - typedef const void* const_pointer; - typedef void value_type; - - template struct rebind {typedef allocator<_Up> other;}; -}; - -template <> -class _LIBCPP_TEMPLATE_VIS _LIBCPP_DEPRECATED_IN_CXX17 allocator -{ -public: - typedef const void* pointer; - typedef const void* const_pointer; - typedef const void value_type; - - template struct rebind {typedef allocator<_Up> other;}; -}; -#endif - -// allocator - -template -class _LIBCPP_TEMPLATE_VIS allocator -{ -public: - typedef size_t size_type; - typedef ptrdiff_t difference_type; - typedef _Tp value_type; - typedef true_type propagate_on_container_move_assignment; - typedef true_type is_always_equal; - - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 - allocator() _NOEXCEPT { } - - template - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 - allocator(const allocator<_Up>&) _NOEXCEPT { } - - _LIBCPP_NODISCARD_AFTER_CXX17 _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 - _Tp* allocate(size_t __n) { - if (__n > allocator_traits::max_size(*this)) - __throw_length_error("allocator::allocate(size_t n)" - " 'n' exceeds maximum supported size"); - if (__libcpp_is_constant_evaluated()) { - return static_cast<_Tp*>(::operator new(__n * sizeof(_Tp))); - } else { - return static_cast<_Tp*>(_VSTD::__libcpp_allocate(__n * sizeof(_Tp), _LIBCPP_ALIGNOF(_Tp))); - } - } - - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 - void deallocate(_Tp* __p, size_t __n) _NOEXCEPT { - if (__libcpp_is_constant_evaluated()) { - ::operator delete(__p); - } else { - _VSTD::__libcpp_deallocate((void*)__p, __n * sizeof(_Tp), _LIBCPP_ALIGNOF(_Tp)); - } - } - - // C++20 Removed members -#if _LIBCPP_STD_VER <= 17 || defined(_LIBCPP_ENABLE_CXX20_REMOVED_ALLOCATOR_MEMBERS) - _LIBCPP_DEPRECATED_IN_CXX17 typedef _Tp* pointer; - _LIBCPP_DEPRECATED_IN_CXX17 typedef const _Tp* const_pointer; - _LIBCPP_DEPRECATED_IN_CXX17 typedef _Tp& reference; - _LIBCPP_DEPRECATED_IN_CXX17 typedef const _Tp& const_reference; - - template - struct _LIBCPP_DEPRECATED_IN_CXX17 rebind { - typedef allocator<_Up> other; - }; - - _LIBCPP_DEPRECATED_IN_CXX17 _LIBCPP_INLINE_VISIBILITY - pointer address(reference __x) const _NOEXCEPT { - return _VSTD::addressof(__x); - } - _LIBCPP_DEPRECATED_IN_CXX17 _LIBCPP_INLINE_VISIBILITY - const_pointer address(const_reference __x) const _NOEXCEPT { - return _VSTD::addressof(__x); - } - - _LIBCPP_NODISCARD_AFTER_CXX17 _LIBCPP_INLINE_VISIBILITY _LIBCPP_DEPRECATED_IN_CXX17 - _Tp* allocate(size_t __n, const void*) { - return allocate(__n); - } - - _LIBCPP_DEPRECATED_IN_CXX17 _LIBCPP_INLINE_VISIBILITY size_type max_size() const _NOEXCEPT { - return size_type(~0) / sizeof(_Tp); - } - - template - _LIBCPP_DEPRECATED_IN_CXX17 _LIBCPP_INLINE_VISIBILITY - void construct(_Up* __p, _Args&&... __args) { - ::new ((void*)__p) _Up(_VSTD::forward<_Args>(__args)...); - } - - _LIBCPP_DEPRECATED_IN_CXX17 _LIBCPP_INLINE_VISIBILITY - void destroy(pointer __p) { - __p->~_Tp(); - } -#endif -}; - -template -class _LIBCPP_TEMPLATE_VIS allocator -{ -public: - typedef size_t size_type; - typedef ptrdiff_t difference_type; - typedef const _Tp value_type; - typedef true_type propagate_on_container_move_assignment; - typedef true_type is_always_equal; - - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 - allocator() _NOEXCEPT { } - - template - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 - allocator(const allocator<_Up>&) _NOEXCEPT { } - - _LIBCPP_NODISCARD_AFTER_CXX17 _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 - const _Tp* allocate(size_t __n) { - if (__n > allocator_traits::max_size(*this)) - __throw_length_error("allocator::allocate(size_t n)" - " 'n' exceeds maximum supported size"); - if (__libcpp_is_constant_evaluated()) { - return static_cast(::operator new(__n * sizeof(_Tp))); - } else { - return static_cast(_VSTD::__libcpp_allocate(__n * sizeof(_Tp), _LIBCPP_ALIGNOF(_Tp))); - } - } - - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 - void deallocate(const _Tp* __p, size_t __n) { - if (__libcpp_is_constant_evaluated()) { - ::operator delete(const_cast<_Tp*>(__p)); - } else { - _VSTD::__libcpp_deallocate((void*) const_cast<_Tp *>(__p), __n * sizeof(_Tp), _LIBCPP_ALIGNOF(_Tp)); - } - } - - // C++20 Removed members -#if _LIBCPP_STD_VER <= 17 || defined(_LIBCPP_ENABLE_CXX20_REMOVED_ALLOCATOR_MEMBERS) - _LIBCPP_DEPRECATED_IN_CXX17 typedef const _Tp* pointer; - _LIBCPP_DEPRECATED_IN_CXX17 typedef const _Tp* const_pointer; - _LIBCPP_DEPRECATED_IN_CXX17 typedef const _Tp& reference; - _LIBCPP_DEPRECATED_IN_CXX17 typedef const _Tp& const_reference; - - template - struct _LIBCPP_DEPRECATED_IN_CXX17 rebind { - typedef allocator<_Up> other; - }; - - _LIBCPP_DEPRECATED_IN_CXX17 _LIBCPP_INLINE_VISIBILITY - const_pointer address(const_reference __x) const _NOEXCEPT { - return _VSTD::addressof(__x); - } - - _LIBCPP_NODISCARD_AFTER_CXX17 _LIBCPP_INLINE_VISIBILITY _LIBCPP_DEPRECATED_IN_CXX17 - const _Tp* allocate(size_t __n, const void*) { - return allocate(__n); - } - - _LIBCPP_DEPRECATED_IN_CXX17 _LIBCPP_INLINE_VISIBILITY size_type max_size() const _NOEXCEPT { - return size_type(~0) / sizeof(_Tp); - } - - template - _LIBCPP_DEPRECATED_IN_CXX17 _LIBCPP_INLINE_VISIBILITY - void construct(_Up* __p, _Args&&... __args) { - ::new ((void*)__p) _Up(_VSTD::forward<_Args>(__args)...); - } - - _LIBCPP_DEPRECATED_IN_CXX17 _LIBCPP_INLINE_VISIBILITY - void destroy(pointer __p) { - __p->~_Tp(); - } -#endif -}; - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -bool operator==(const allocator<_Tp>&, const allocator<_Up>&) _NOEXCEPT {return true;} - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -bool operator!=(const allocator<_Tp>&, const allocator<_Up>&) _NOEXCEPT {return false;} - template _LIBCPP_INLINE_VISIBILITY void __construct_forward_with_exception_guarantees(_Alloc& __a, _Ptr __begin1, _Ptr __end1, _Ptr& __begin2) { @@ -1062,3107 +796,48 @@ void __construct_backward_with_exception_guarantees(_Alloc&, _Tp* __begin1, _Tp* ptrdiff_t _Np = __end1 - __begin1; __end2 -= _Np; if (_Np > 0) - _VSTD::memcpy(__end2, __begin1, _Np * sizeof(_Tp)); -} - -template -class _LIBCPP_TEMPLATE_VIS raw_storage_iterator - : public iterator&> // purposefully not C++03 -{ -private: - _OutputIterator __x_; -public: - _LIBCPP_INLINE_VISIBILITY explicit raw_storage_iterator(_OutputIterator __x) : __x_(__x) {} - _LIBCPP_INLINE_VISIBILITY raw_storage_iterator& operator*() {return *this;} - _LIBCPP_INLINE_VISIBILITY raw_storage_iterator& operator=(const _Tp& __element) - {::new ((void*)_VSTD::addressof(*__x_)) _Tp(__element); return *this;} -#if _LIBCPP_STD_VER >= 14 - _LIBCPP_INLINE_VISIBILITY raw_storage_iterator& operator=(_Tp&& __element) - {::new ((void*)_VSTD::addressof(*__x_)) _Tp(_VSTD::move(__element)); return *this;} -#endif - _LIBCPP_INLINE_VISIBILITY raw_storage_iterator& operator++() {++__x_; return *this;} - _LIBCPP_INLINE_VISIBILITY raw_storage_iterator operator++(int) - {raw_storage_iterator __t(*this); ++__x_; return __t;} -#if _LIBCPP_STD_VER >= 14 - _LIBCPP_INLINE_VISIBILITY _OutputIterator base() const { return __x_; } -#endif -}; - -template -_LIBCPP_NODISCARD_EXT _LIBCPP_NO_CFI -pair<_Tp*, ptrdiff_t> -get_temporary_buffer(ptrdiff_t __n) _NOEXCEPT -{ - pair<_Tp*, ptrdiff_t> __r(0, 0); - const ptrdiff_t __m = (~ptrdiff_t(0) ^ - ptrdiff_t(ptrdiff_t(1) << (sizeof(ptrdiff_t) * __CHAR_BIT__ - 1))) - / sizeof(_Tp); - if (__n > __m) - __n = __m; - while (__n > 0) - { -#if !defined(_LIBCPP_HAS_NO_ALIGNED_ALLOCATION) - if (__is_overaligned_for_new(_LIBCPP_ALIGNOF(_Tp))) - { - align_val_t __al = - align_val_t(alignment_of<_Tp>::value); - __r.first = static_cast<_Tp*>(::operator new( - __n * sizeof(_Tp), __al, nothrow)); - } else { - __r.first = static_cast<_Tp*>(::operator new( - __n * sizeof(_Tp), nothrow)); - } -#else - if (__is_overaligned_for_new(_LIBCPP_ALIGNOF(_Tp))) - { - // Since aligned operator new is unavailable, return an empty - // buffer rather than one with invalid alignment. - return __r; - } - - __r.first = static_cast<_Tp*>(::operator new(__n * sizeof(_Tp), nothrow)); -#endif - - if (__r.first) - { - __r.second = __n; - break; - } - __n /= 2; - } - return __r; -} - -template -inline _LIBCPP_INLINE_VISIBILITY -void return_temporary_buffer(_Tp* __p) _NOEXCEPT -{ - _VSTD::__libcpp_deallocate_unsized((void*)__p, _LIBCPP_ALIGNOF(_Tp)); + _VSTD::memcpy(static_cast(__end2), static_cast(__begin1), _Np * sizeof(_Tp)); } -#if _LIBCPP_STD_VER <= 14 || defined(_LIBCPP_ENABLE_CXX17_REMOVED_AUTO_PTR) -template -struct _LIBCPP_DEPRECATED_IN_CXX11 auto_ptr_ref -{ - _Tp* __ptr_; -}; - -template -class _LIBCPP_TEMPLATE_VIS _LIBCPP_DEPRECATED_IN_CXX11 auto_ptr -{ -private: - _Tp* __ptr_; -public: - typedef _Tp element_type; - - _LIBCPP_INLINE_VISIBILITY explicit auto_ptr(_Tp* __p = 0) _NOEXCEPT : __ptr_(__p) {} - _LIBCPP_INLINE_VISIBILITY auto_ptr(auto_ptr& __p) _NOEXCEPT : __ptr_(__p.release()) {} - template _LIBCPP_INLINE_VISIBILITY auto_ptr(auto_ptr<_Up>& __p) _NOEXCEPT - : __ptr_(__p.release()) {} - _LIBCPP_INLINE_VISIBILITY auto_ptr& operator=(auto_ptr& __p) _NOEXCEPT - {reset(__p.release()); return *this;} - template _LIBCPP_INLINE_VISIBILITY auto_ptr& operator=(auto_ptr<_Up>& __p) _NOEXCEPT - {reset(__p.release()); return *this;} - _LIBCPP_INLINE_VISIBILITY auto_ptr& operator=(auto_ptr_ref<_Tp> __p) _NOEXCEPT - {reset(__p.__ptr_); return *this;} - _LIBCPP_INLINE_VISIBILITY ~auto_ptr() _NOEXCEPT {delete __ptr_;} - - _LIBCPP_INLINE_VISIBILITY _Tp& operator*() const _NOEXCEPT - {return *__ptr_;} - _LIBCPP_INLINE_VISIBILITY _Tp* operator->() const _NOEXCEPT {return __ptr_;} - _LIBCPP_INLINE_VISIBILITY _Tp* get() const _NOEXCEPT {return __ptr_;} - _LIBCPP_INLINE_VISIBILITY _Tp* release() _NOEXCEPT - { - _Tp* __t = __ptr_; - __ptr_ = nullptr; - return __t; - } - _LIBCPP_INLINE_VISIBILITY void reset(_Tp* __p = 0) _NOEXCEPT - { - if (__ptr_ != __p) - delete __ptr_; - __ptr_ = __p; - } - - _LIBCPP_INLINE_VISIBILITY auto_ptr(auto_ptr_ref<_Tp> __p) _NOEXCEPT : __ptr_(__p.__ptr_) {} - template _LIBCPP_INLINE_VISIBILITY operator auto_ptr_ref<_Up>() _NOEXCEPT - {auto_ptr_ref<_Up> __t; __t.__ptr_ = release(); return __t;} - template _LIBCPP_INLINE_VISIBILITY operator auto_ptr<_Up>() _NOEXCEPT - {return auto_ptr<_Up>(release());} -}; - -template <> -class _LIBCPP_TEMPLATE_VIS _LIBCPP_DEPRECATED_IN_CXX11 auto_ptr +struct __destruct_n { -public: - typedef void element_type; -}; -#endif - -// Tag used to default initialize one or both of the pair's elements. -struct __default_init_tag {}; -struct __value_init_tag {}; - -template ::value && !__libcpp_is_final<_Tp>::value> -struct __compressed_pair_elem { - typedef _Tp _ParamT; - typedef _Tp& reference; - typedef const _Tp& const_reference; - - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR - __compressed_pair_elem(__default_init_tag) {} - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR - __compressed_pair_elem(__value_init_tag) : __value_() {} - - template ::type>::value - >::type> - _LIBCPP_INLINE_VISIBILITY - _LIBCPP_CONSTEXPR explicit - __compressed_pair_elem(_Up&& __u) - : __value_(_VSTD::forward<_Up>(__u)) - { - } - - -#ifndef _LIBCPP_CXX03_LANG - template - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 - __compressed_pair_elem(piecewise_construct_t, tuple<_Args...> __args, - __tuple_indices<_Indexes...>) - : __value_(_VSTD::forward<_Args>(_VSTD::get<_Indexes>(__args))...) {} -#endif - - - _LIBCPP_INLINE_VISIBILITY reference __get() _NOEXCEPT { return __value_; } - _LIBCPP_INLINE_VISIBILITY - const_reference __get() const _NOEXCEPT { return __value_; } - private: - _Tp __value_; -}; - -template -struct __compressed_pair_elem<_Tp, _Idx, true> : private _Tp { - typedef _Tp _ParamT; - typedef _Tp& reference; - typedef const _Tp& const_reference; - typedef _Tp __value_type; - - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR __compressed_pair_elem() = default; - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR - __compressed_pair_elem(__default_init_tag) {} - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR - __compressed_pair_elem(__value_init_tag) : __value_type() {} - - template ::type>::value - >::type> - _LIBCPP_INLINE_VISIBILITY - _LIBCPP_CONSTEXPR explicit - __compressed_pair_elem(_Up&& __u) - : __value_type(_VSTD::forward<_Up>(__u)) - {} - -#ifndef _LIBCPP_CXX03_LANG - template - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 - __compressed_pair_elem(piecewise_construct_t, tuple<_Args...> __args, - __tuple_indices<_Indexes...>) - : __value_type(_VSTD::forward<_Args>(_VSTD::get<_Indexes>(__args))...) {} -#endif - - _LIBCPP_INLINE_VISIBILITY reference __get() _NOEXCEPT { return *this; } - _LIBCPP_INLINE_VISIBILITY - const_reference __get() const _NOEXCEPT { return *this; } -}; - -template -class __compressed_pair : private __compressed_pair_elem<_T1, 0>, - private __compressed_pair_elem<_T2, 1> { -public: - // NOTE: This static assert should never fire because __compressed_pair - // is *almost never* used in a scenario where it's possible for T1 == T2. - // (The exception is std::function where it is possible that the function - // object and the allocator have the same type). - static_assert((!is_same<_T1, _T2>::value), - "__compressed_pair cannot be instantiated when T1 and T2 are the same type; " - "The current implementation is NOT ABI-compatible with the previous " - "implementation for this configuration"); - - typedef _LIBCPP_NODEBUG_TYPE __compressed_pair_elem<_T1, 0> _Base1; - typedef _LIBCPP_NODEBUG_TYPE __compressed_pair_elem<_T2, 1> _Base2; - - template , _Dummy>::value && - __dependent_type, _Dummy>::value - >::type - > - _LIBCPP_INLINE_VISIBILITY - _LIBCPP_CONSTEXPR __compressed_pair() : _Base1(__value_init_tag()), _Base2(__value_init_tag()) {} - - template - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR - __compressed_pair(_U1&& __t1, _U2&& __t2) - : _Base1(_VSTD::forward<_U1>(__t1)), _Base2(_VSTD::forward<_U2>(__t2)) {} - -#ifndef _LIBCPP_CXX03_LANG - template - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX14 - __compressed_pair(piecewise_construct_t __pc, tuple<_Args1...> __first_args, - tuple<_Args2...> __second_args) - : _Base1(__pc, _VSTD::move(__first_args), - typename __make_tuple_indices::type()), - _Base2(__pc, _VSTD::move(__second_args), - typename __make_tuple_indices::type()) {} -#endif - - _LIBCPP_INLINE_VISIBILITY - typename _Base1::reference first() _NOEXCEPT { - return static_cast<_Base1&>(*this).__get(); - } - - _LIBCPP_INLINE_VISIBILITY - typename _Base1::const_reference first() const _NOEXCEPT { - return static_cast<_Base1 const&>(*this).__get(); - } - - _LIBCPP_INLINE_VISIBILITY - typename _Base2::reference second() _NOEXCEPT { - return static_cast<_Base2&>(*this).__get(); - } - - _LIBCPP_INLINE_VISIBILITY - typename _Base2::const_reference second() const _NOEXCEPT { - return static_cast<_Base2 const&>(*this).__get(); - } - - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR - static _Base1* __get_first_base(__compressed_pair* __pair) _NOEXCEPT { - return static_cast<_Base1*>(__pair); - } - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR - static _Base2* __get_second_base(__compressed_pair* __pair) _NOEXCEPT { - return static_cast<_Base2*>(__pair); - } - - _LIBCPP_INLINE_VISIBILITY - void swap(__compressed_pair& __x) - _NOEXCEPT_(__is_nothrow_swappable<_T1>::value && - __is_nothrow_swappable<_T2>::value) - { - using _VSTD::swap; - swap(first(), __x.first()); - swap(second(), __x.second()); - } -}; - -template -inline _LIBCPP_INLINE_VISIBILITY -void swap(__compressed_pair<_T1, _T2>& __x, __compressed_pair<_T1, _T2>& __y) - _NOEXCEPT_(__is_nothrow_swappable<_T1>::value && - __is_nothrow_swappable<_T2>::value) { - __x.swap(__y); -} + size_t __size_; -// default_delete + template + _LIBCPP_INLINE_VISIBILITY void __process(_Tp* __p, false_type) _NOEXCEPT + {for (size_t __i = 0; __i < __size_; ++__i, ++__p) __p->~_Tp();} -template -struct _LIBCPP_TEMPLATE_VIS default_delete { - static_assert(!is_function<_Tp>::value, - "default_delete cannot be instantiated for function types"); -#ifndef _LIBCPP_CXX03_LANG - _LIBCPP_INLINE_VISIBILITY constexpr default_delete() _NOEXCEPT = default; -#else - _LIBCPP_INLINE_VISIBILITY default_delete() {} -#endif - template - _LIBCPP_INLINE_VISIBILITY - default_delete(const default_delete<_Up>&, - typename enable_if::value>::type* = - 0) _NOEXCEPT {} - - _LIBCPP_INLINE_VISIBILITY void operator()(_Tp* __ptr) const _NOEXCEPT { - static_assert(sizeof(_Tp) > 0, - "default_delete can not delete incomplete type"); - static_assert(!is_void<_Tp>::value, - "default_delete can not delete incomplete type"); - delete __ptr; - } -}; + template + _LIBCPP_INLINE_VISIBILITY void __process(_Tp*, true_type) _NOEXCEPT + {} -template -struct _LIBCPP_TEMPLATE_VIS default_delete<_Tp[]> { -private: - template - struct _EnableIfConvertible - : enable_if::value> {}; + _LIBCPP_INLINE_VISIBILITY void __incr(false_type) _NOEXCEPT + {++__size_;} + _LIBCPP_INLINE_VISIBILITY void __incr(true_type) _NOEXCEPT + {} + _LIBCPP_INLINE_VISIBILITY void __set(size_t __s, false_type) _NOEXCEPT + {__size_ = __s;} + _LIBCPP_INLINE_VISIBILITY void __set(size_t, true_type) _NOEXCEPT + {} public: -#ifndef _LIBCPP_CXX03_LANG - _LIBCPP_INLINE_VISIBILITY constexpr default_delete() _NOEXCEPT = default; -#else - _LIBCPP_INLINE_VISIBILITY default_delete() {} -#endif - - template - _LIBCPP_INLINE_VISIBILITY - default_delete(const default_delete<_Up[]>&, - typename _EnableIfConvertible<_Up>::type* = 0) _NOEXCEPT {} - - template - _LIBCPP_INLINE_VISIBILITY - typename _EnableIfConvertible<_Up>::type - operator()(_Up* __ptr) const _NOEXCEPT { - static_assert(sizeof(_Tp) > 0, - "default_delete can not delete incomplete type"); - static_assert(!is_void<_Tp>::value, - "default_delete can not delete void type"); - delete[] __ptr; - } -}; + _LIBCPP_INLINE_VISIBILITY explicit __destruct_n(size_t __s) _NOEXCEPT + : __size_(__s) {} -template -struct __unique_ptr_deleter_sfinae { - static_assert(!is_reference<_Deleter>::value, "incorrect specialization"); - typedef const _Deleter& __lval_ref_type; - typedef _Deleter&& __good_rval_ref_type; - typedef true_type __enable_rval_overload; -}; + template + _LIBCPP_INLINE_VISIBILITY void __incr() _NOEXCEPT + {__incr(integral_constant::value>());} -template -struct __unique_ptr_deleter_sfinae<_Deleter const&> { - typedef const _Deleter& __lval_ref_type; - typedef const _Deleter&& __bad_rval_ref_type; - typedef false_type __enable_rval_overload; -}; + template + _LIBCPP_INLINE_VISIBILITY void __set(size_t __s, _Tp*) _NOEXCEPT + {__set(__s, integral_constant::value>());} -template -struct __unique_ptr_deleter_sfinae<_Deleter&> { - typedef _Deleter& __lval_ref_type; - typedef _Deleter&& __bad_rval_ref_type; - typedef false_type __enable_rval_overload; + template + _LIBCPP_INLINE_VISIBILITY void operator()(_Tp* __p) _NOEXCEPT + {__process(__p, integral_constant::value>());} }; -#if defined(_LIBCPP_ABI_ENABLE_UNIQUE_PTR_TRIVIAL_ABI) -# define _LIBCPP_UNIQUE_PTR_TRIVIAL_ABI __attribute__((trivial_abi)) -#else -# define _LIBCPP_UNIQUE_PTR_TRIVIAL_ABI -#endif - -template > -class _LIBCPP_UNIQUE_PTR_TRIVIAL_ABI _LIBCPP_TEMPLATE_VIS unique_ptr { -public: - typedef _Tp element_type; - typedef _Dp deleter_type; - typedef _LIBCPP_NODEBUG_TYPE typename __pointer<_Tp, deleter_type>::type pointer; - - static_assert(!is_rvalue_reference::value, - "the specified deleter type cannot be an rvalue reference"); - -private: - __compressed_pair __ptr_; - - struct __nat { int __for_bool_; }; - - typedef _LIBCPP_NODEBUG_TYPE __unique_ptr_deleter_sfinae<_Dp> _DeleterSFINAE; - - template - using _LValRefType _LIBCPP_NODEBUG_TYPE = - typename __dependent_type<_DeleterSFINAE, _Dummy>::__lval_ref_type; - - template - using _GoodRValRefType _LIBCPP_NODEBUG_TYPE = - typename __dependent_type<_DeleterSFINAE, _Dummy>::__good_rval_ref_type; - - template - using _BadRValRefType _LIBCPP_NODEBUG_TYPE = - typename __dependent_type<_DeleterSFINAE, _Dummy>::__bad_rval_ref_type; - - template , _Dummy>::type> - using _EnableIfDeleterDefaultConstructible _LIBCPP_NODEBUG_TYPE = - typename enable_if::value && - !is_pointer<_Deleter>::value>::type; - - template - using _EnableIfDeleterConstructible _LIBCPP_NODEBUG_TYPE = - typename enable_if::value>::type; - - template - using _EnableIfMoveConvertible _LIBCPP_NODEBUG_TYPE = typename enable_if< - is_convertible::value && - !is_array<_Up>::value - >::type; - - template - using _EnableIfDeleterConvertible _LIBCPP_NODEBUG_TYPE = typename enable_if< - (is_reference<_Dp>::value && is_same<_Dp, _UDel>::value) || - (!is_reference<_Dp>::value && is_convertible<_UDel, _Dp>::value) - >::type; - - template - using _EnableIfDeleterAssignable = typename enable_if< - is_assignable<_Dp&, _UDel&&>::value - >::type; - -public: - template > - _LIBCPP_INLINE_VISIBILITY - _LIBCPP_CONSTEXPR unique_ptr() _NOEXCEPT : __ptr_(pointer(), __default_init_tag()) {} - - template > - _LIBCPP_INLINE_VISIBILITY - _LIBCPP_CONSTEXPR unique_ptr(nullptr_t) _NOEXCEPT : __ptr_(pointer(), __default_init_tag()) {} - - template > - _LIBCPP_INLINE_VISIBILITY - explicit unique_ptr(pointer __p) _NOEXCEPT : __ptr_(__p, __default_init_tag()) {} - - template > > - _LIBCPP_INLINE_VISIBILITY - unique_ptr(pointer __p, _LValRefType<_Dummy> __d) _NOEXCEPT - : __ptr_(__p, __d) {} - - template > > - _LIBCPP_INLINE_VISIBILITY - unique_ptr(pointer __p, _GoodRValRefType<_Dummy> __d) _NOEXCEPT - : __ptr_(__p, _VSTD::move(__d)) { - static_assert(!is_reference::value, - "rvalue deleter bound to reference"); - } - - template > > - _LIBCPP_INLINE_VISIBILITY - unique_ptr(pointer __p, _BadRValRefType<_Dummy> __d) = delete; - - _LIBCPP_INLINE_VISIBILITY - unique_ptr(unique_ptr&& __u) _NOEXCEPT - : __ptr_(__u.release(), _VSTD::forward(__u.get_deleter())) { - } - - template , _Up>, - class = _EnableIfDeleterConvertible<_Ep> - > - _LIBCPP_INLINE_VISIBILITY - unique_ptr(unique_ptr<_Up, _Ep>&& __u) _NOEXCEPT - : __ptr_(__u.release(), _VSTD::forward<_Ep>(__u.get_deleter())) {} - -#if _LIBCPP_STD_VER <= 14 || defined(_LIBCPP_ENABLE_CXX17_REMOVED_AUTO_PTR) - template - _LIBCPP_INLINE_VISIBILITY - unique_ptr(auto_ptr<_Up>&& __p, - typename enable_if::value && - is_same<_Dp, default_delete<_Tp> >::value, - __nat>::type = __nat()) _NOEXCEPT - : __ptr_(__p.release(), __default_init_tag()) {} -#endif - - _LIBCPP_INLINE_VISIBILITY - unique_ptr& operator=(unique_ptr&& __u) _NOEXCEPT { - reset(__u.release()); - __ptr_.second() = _VSTD::forward(__u.get_deleter()); - return *this; - } - - template , _Up>, - class = _EnableIfDeleterAssignable<_Ep> - > - _LIBCPP_INLINE_VISIBILITY - unique_ptr& operator=(unique_ptr<_Up, _Ep>&& __u) _NOEXCEPT { - reset(__u.release()); - __ptr_.second() = _VSTD::forward<_Ep>(__u.get_deleter()); - return *this; - } - -#if _LIBCPP_STD_VER <= 14 || defined(_LIBCPP_ENABLE_CXX17_REMOVED_AUTO_PTR) - template - _LIBCPP_INLINE_VISIBILITY - typename enable_if::value && - is_same<_Dp, default_delete<_Tp> >::value, - unique_ptr&>::type - operator=(auto_ptr<_Up> __p) { - reset(__p.release()); - return *this; - } -#endif - -#ifdef _LIBCPP_CXX03_LANG - unique_ptr(unique_ptr const&) = delete; - unique_ptr& operator=(unique_ptr const&) = delete; -#endif - - - _LIBCPP_INLINE_VISIBILITY - ~unique_ptr() { reset(); } - - _LIBCPP_INLINE_VISIBILITY - unique_ptr& operator=(nullptr_t) _NOEXCEPT { - reset(); - return *this; - } - - _LIBCPP_INLINE_VISIBILITY - typename add_lvalue_reference<_Tp>::type - operator*() const { - return *__ptr_.first(); - } - _LIBCPP_INLINE_VISIBILITY - pointer operator->() const _NOEXCEPT { - return __ptr_.first(); - } - _LIBCPP_INLINE_VISIBILITY - pointer get() const _NOEXCEPT { - return __ptr_.first(); - } - _LIBCPP_INLINE_VISIBILITY - deleter_type& get_deleter() _NOEXCEPT { - return __ptr_.second(); - } - _LIBCPP_INLINE_VISIBILITY - const deleter_type& get_deleter() const _NOEXCEPT { - return __ptr_.second(); - } - _LIBCPP_INLINE_VISIBILITY - _LIBCPP_EXPLICIT operator bool() const _NOEXCEPT { - return __ptr_.first() != nullptr; - } - - _LIBCPP_INLINE_VISIBILITY - pointer release() _NOEXCEPT { - pointer __t = __ptr_.first(); - __ptr_.first() = pointer(); - return __t; - } - - _LIBCPP_INLINE_VISIBILITY - void reset(pointer __p = pointer()) _NOEXCEPT { - pointer __tmp = __ptr_.first(); - __ptr_.first() = __p; - if (__tmp) - __ptr_.second()(__tmp); - } - - _LIBCPP_INLINE_VISIBILITY - void swap(unique_ptr& __u) _NOEXCEPT { - __ptr_.swap(__u.__ptr_); - } -}; - - -template -class _LIBCPP_UNIQUE_PTR_TRIVIAL_ABI _LIBCPP_TEMPLATE_VIS unique_ptr<_Tp[], _Dp> { -public: - typedef _Tp element_type; - typedef _Dp deleter_type; - typedef typename __pointer<_Tp, deleter_type>::type pointer; - -private: - __compressed_pair __ptr_; - - template - struct _CheckArrayPointerConversion : is_same<_From, pointer> {}; - - template - struct _CheckArrayPointerConversion<_FromElem*> - : integral_constant::value || - (is_same::value && - is_convertible<_FromElem(*)[], element_type(*)[]>::value) - > - {}; - - typedef __unique_ptr_deleter_sfinae<_Dp> _DeleterSFINAE; - - template - using _LValRefType _LIBCPP_NODEBUG_TYPE = - typename __dependent_type<_DeleterSFINAE, _Dummy>::__lval_ref_type; - - template - using _GoodRValRefType _LIBCPP_NODEBUG_TYPE = - typename __dependent_type<_DeleterSFINAE, _Dummy>::__good_rval_ref_type; - - template - using _BadRValRefType _LIBCPP_NODEBUG_TYPE = - typename __dependent_type<_DeleterSFINAE, _Dummy>::__bad_rval_ref_type; - - template , _Dummy>::type> - using _EnableIfDeleterDefaultConstructible _LIBCPP_NODEBUG_TYPE = - typename enable_if::value && - !is_pointer<_Deleter>::value>::type; - - template - using _EnableIfDeleterConstructible _LIBCPP_NODEBUG_TYPE = - typename enable_if::value>::type; - - template - using _EnableIfPointerConvertible _LIBCPP_NODEBUG_TYPE = typename enable_if< - _CheckArrayPointerConversion<_Pp>::value - >::type; - - template - using _EnableIfMoveConvertible _LIBCPP_NODEBUG_TYPE = typename enable_if< - is_array<_Up>::value && - is_same::value && - is_same::value && - is_convertible<_ElemT(*)[], element_type(*)[]>::value - >::type; - - template - using _EnableIfDeleterConvertible _LIBCPP_NODEBUG_TYPE = typename enable_if< - (is_reference<_Dp>::value && is_same<_Dp, _UDel>::value) || - (!is_reference<_Dp>::value && is_convertible<_UDel, _Dp>::value) - >::type; - - template - using _EnableIfDeleterAssignable _LIBCPP_NODEBUG_TYPE = typename enable_if< - is_assignable<_Dp&, _UDel&&>::value - >::type; - -public: - template > - _LIBCPP_INLINE_VISIBILITY - _LIBCPP_CONSTEXPR unique_ptr() _NOEXCEPT : __ptr_(pointer(), __default_init_tag()) {} - - template > - _LIBCPP_INLINE_VISIBILITY - _LIBCPP_CONSTEXPR unique_ptr(nullptr_t) _NOEXCEPT : __ptr_(pointer(), __default_init_tag()) {} - - template , - class = _EnableIfPointerConvertible<_Pp> > - _LIBCPP_INLINE_VISIBILITY - explicit unique_ptr(_Pp __p) _NOEXCEPT - : __ptr_(__p, __default_init_tag()) {} - - template >, - class = _EnableIfPointerConvertible<_Pp> > - _LIBCPP_INLINE_VISIBILITY - unique_ptr(_Pp __p, _LValRefType<_Dummy> __d) _NOEXCEPT - : __ptr_(__p, __d) {} - - template > > - _LIBCPP_INLINE_VISIBILITY - unique_ptr(nullptr_t, _LValRefType<_Dummy> __d) _NOEXCEPT - : __ptr_(nullptr, __d) {} - - template >, - class = _EnableIfPointerConvertible<_Pp> > - _LIBCPP_INLINE_VISIBILITY - unique_ptr(_Pp __p, _GoodRValRefType<_Dummy> __d) _NOEXCEPT - : __ptr_(__p, _VSTD::move(__d)) { - static_assert(!is_reference::value, - "rvalue deleter bound to reference"); - } - - template > > - _LIBCPP_INLINE_VISIBILITY - unique_ptr(nullptr_t, _GoodRValRefType<_Dummy> __d) _NOEXCEPT - : __ptr_(nullptr, _VSTD::move(__d)) { - static_assert(!is_reference::value, - "rvalue deleter bound to reference"); - } - - template >, - class = _EnableIfPointerConvertible<_Pp> > - _LIBCPP_INLINE_VISIBILITY - unique_ptr(_Pp __p, _BadRValRefType<_Dummy> __d) = delete; - - _LIBCPP_INLINE_VISIBILITY - unique_ptr(unique_ptr&& __u) _NOEXCEPT - : __ptr_(__u.release(), _VSTD::forward(__u.get_deleter())) { - } - - _LIBCPP_INLINE_VISIBILITY - unique_ptr& operator=(unique_ptr&& __u) _NOEXCEPT { - reset(__u.release()); - __ptr_.second() = _VSTD::forward(__u.get_deleter()); - return *this; - } - - template , _Up>, - class = _EnableIfDeleterConvertible<_Ep> - > - _LIBCPP_INLINE_VISIBILITY - unique_ptr(unique_ptr<_Up, _Ep>&& __u) _NOEXCEPT - : __ptr_(__u.release(), _VSTD::forward<_Ep>(__u.get_deleter())) { - } - - template , _Up>, - class = _EnableIfDeleterAssignable<_Ep> - > - _LIBCPP_INLINE_VISIBILITY - unique_ptr& - operator=(unique_ptr<_Up, _Ep>&& __u) _NOEXCEPT { - reset(__u.release()); - __ptr_.second() = _VSTD::forward<_Ep>(__u.get_deleter()); - return *this; - } - -#ifdef _LIBCPP_CXX03_LANG - unique_ptr(unique_ptr const&) = delete; - unique_ptr& operator=(unique_ptr const&) = delete; -#endif - -public: - _LIBCPP_INLINE_VISIBILITY - ~unique_ptr() { reset(); } - - _LIBCPP_INLINE_VISIBILITY - unique_ptr& operator=(nullptr_t) _NOEXCEPT { - reset(); - return *this; - } - - _LIBCPP_INLINE_VISIBILITY - typename add_lvalue_reference<_Tp>::type - operator[](size_t __i) const { - return __ptr_.first()[__i]; - } - _LIBCPP_INLINE_VISIBILITY - pointer get() const _NOEXCEPT { - return __ptr_.first(); - } - - _LIBCPP_INLINE_VISIBILITY - deleter_type& get_deleter() _NOEXCEPT { - return __ptr_.second(); - } - - _LIBCPP_INLINE_VISIBILITY - const deleter_type& get_deleter() const _NOEXCEPT { - return __ptr_.second(); - } - _LIBCPP_INLINE_VISIBILITY - _LIBCPP_EXPLICIT operator bool() const _NOEXCEPT { - return __ptr_.first() != nullptr; - } - - _LIBCPP_INLINE_VISIBILITY - pointer release() _NOEXCEPT { - pointer __t = __ptr_.first(); - __ptr_.first() = pointer(); - return __t; - } - - template - _LIBCPP_INLINE_VISIBILITY - typename enable_if< - _CheckArrayPointerConversion<_Pp>::value - >::type - reset(_Pp __p) _NOEXCEPT { - pointer __tmp = __ptr_.first(); - __ptr_.first() = __p; - if (__tmp) - __ptr_.second()(__tmp); - } - - _LIBCPP_INLINE_VISIBILITY - void reset(nullptr_t = nullptr) _NOEXCEPT { - pointer __tmp = __ptr_.first(); - __ptr_.first() = nullptr; - if (__tmp) - __ptr_.second()(__tmp); - } - - _LIBCPP_INLINE_VISIBILITY - void swap(unique_ptr& __u) _NOEXCEPT { - __ptr_.swap(__u.__ptr_); - } - -}; - -template -inline _LIBCPP_INLINE_VISIBILITY -typename enable_if< - __is_swappable<_Dp>::value, - void ->::type -swap(unique_ptr<_Tp, _Dp>& __x, unique_ptr<_Tp, _Dp>& __y) _NOEXCEPT {__x.swap(__y);} - -template -inline _LIBCPP_INLINE_VISIBILITY -bool -operator==(const unique_ptr<_T1, _D1>& __x, const unique_ptr<_T2, _D2>& __y) {return __x.get() == __y.get();} - -template -inline _LIBCPP_INLINE_VISIBILITY -bool -operator!=(const unique_ptr<_T1, _D1>& __x, const unique_ptr<_T2, _D2>& __y) {return !(__x == __y);} - -template -inline _LIBCPP_INLINE_VISIBILITY -bool -operator< (const unique_ptr<_T1, _D1>& __x, const unique_ptr<_T2, _D2>& __y) -{ - typedef typename unique_ptr<_T1, _D1>::pointer _P1; - typedef typename unique_ptr<_T2, _D2>::pointer _P2; - typedef typename common_type<_P1, _P2>::type _Vp; - return less<_Vp>()(__x.get(), __y.get()); -} - -template -inline _LIBCPP_INLINE_VISIBILITY -bool -operator> (const unique_ptr<_T1, _D1>& __x, const unique_ptr<_T2, _D2>& __y) {return __y < __x;} - -template -inline _LIBCPP_INLINE_VISIBILITY -bool -operator<=(const unique_ptr<_T1, _D1>& __x, const unique_ptr<_T2, _D2>& __y) {return !(__y < __x);} - -template -inline _LIBCPP_INLINE_VISIBILITY -bool -operator>=(const unique_ptr<_T1, _D1>& __x, const unique_ptr<_T2, _D2>& __y) {return !(__x < __y);} - -template -inline _LIBCPP_INLINE_VISIBILITY -bool -operator==(const unique_ptr<_T1, _D1>& __x, nullptr_t) _NOEXCEPT -{ - return !__x; -} - -template -inline _LIBCPP_INLINE_VISIBILITY -bool -operator==(nullptr_t, const unique_ptr<_T1, _D1>& __x) _NOEXCEPT -{ - return !__x; -} - -template -inline _LIBCPP_INLINE_VISIBILITY -bool -operator!=(const unique_ptr<_T1, _D1>& __x, nullptr_t) _NOEXCEPT -{ - return static_cast(__x); -} - -template -inline _LIBCPP_INLINE_VISIBILITY -bool -operator!=(nullptr_t, const unique_ptr<_T1, _D1>& __x) _NOEXCEPT -{ - return static_cast(__x); -} - -template -inline _LIBCPP_INLINE_VISIBILITY -bool -operator<(const unique_ptr<_T1, _D1>& __x, nullptr_t) -{ - typedef typename unique_ptr<_T1, _D1>::pointer _P1; - return less<_P1>()(__x.get(), nullptr); -} - -template -inline _LIBCPP_INLINE_VISIBILITY -bool -operator<(nullptr_t, const unique_ptr<_T1, _D1>& __x) -{ - typedef typename unique_ptr<_T1, _D1>::pointer _P1; - return less<_P1>()(nullptr, __x.get()); -} - -template -inline _LIBCPP_INLINE_VISIBILITY -bool -operator>(const unique_ptr<_T1, _D1>& __x, nullptr_t) -{ - return nullptr < __x; -} - -template -inline _LIBCPP_INLINE_VISIBILITY -bool -operator>(nullptr_t, const unique_ptr<_T1, _D1>& __x) -{ - return __x < nullptr; -} - -template -inline _LIBCPP_INLINE_VISIBILITY -bool -operator<=(const unique_ptr<_T1, _D1>& __x, nullptr_t) -{ - return !(nullptr < __x); -} - -template -inline _LIBCPP_INLINE_VISIBILITY -bool -operator<=(nullptr_t, const unique_ptr<_T1, _D1>& __x) -{ - return !(__x < nullptr); -} - -template -inline _LIBCPP_INLINE_VISIBILITY -bool -operator>=(const unique_ptr<_T1, _D1>& __x, nullptr_t) -{ - return !(__x < nullptr); -} - -template -inline _LIBCPP_INLINE_VISIBILITY -bool -operator>=(nullptr_t, const unique_ptr<_T1, _D1>& __x) -{ - return !(nullptr < __x); -} - -#if _LIBCPP_STD_VER > 11 - -template -struct __unique_if -{ - typedef unique_ptr<_Tp> __unique_single; -}; - -template -struct __unique_if<_Tp[]> -{ - typedef unique_ptr<_Tp[]> __unique_array_unknown_bound; -}; - -template -struct __unique_if<_Tp[_Np]> -{ - typedef void __unique_array_known_bound; -}; - -template -inline _LIBCPP_INLINE_VISIBILITY -typename __unique_if<_Tp>::__unique_single -make_unique(_Args&&... __args) -{ - return unique_ptr<_Tp>(new _Tp(_VSTD::forward<_Args>(__args)...)); -} - -template -inline _LIBCPP_INLINE_VISIBILITY -typename __unique_if<_Tp>::__unique_array_unknown_bound -make_unique(size_t __n) -{ - typedef typename remove_extent<_Tp>::type _Up; - return unique_ptr<_Tp>(new _Up[__n]()); -} - -template - typename __unique_if<_Tp>::__unique_array_known_bound - make_unique(_Args&&...) = delete; - -#endif // _LIBCPP_STD_VER > 11 - -template -#ifdef _LIBCPP_CXX03_LANG -struct _LIBCPP_TEMPLATE_VIS hash > -#else -struct _LIBCPP_TEMPLATE_VIS hash<__enable_hash_helper< - unique_ptr<_Tp, _Dp>, typename unique_ptr<_Tp, _Dp>::pointer> > -#endif -{ - typedef unique_ptr<_Tp, _Dp> argument_type; - typedef size_t result_type; - _LIBCPP_INLINE_VISIBILITY - result_type operator()(const argument_type& __ptr) const - { - typedef typename argument_type::pointer pointer; - return hash()(__ptr.get()); - } -}; - -struct __destruct_n -{ -private: - size_t __size_; - - template - _LIBCPP_INLINE_VISIBILITY void __process(_Tp* __p, false_type) _NOEXCEPT - {for (size_t __i = 0; __i < __size_; ++__i, ++__p) __p->~_Tp();} - - template - _LIBCPP_INLINE_VISIBILITY void __process(_Tp*, true_type) _NOEXCEPT - {} - - _LIBCPP_INLINE_VISIBILITY void __incr(false_type) _NOEXCEPT - {++__size_;} - _LIBCPP_INLINE_VISIBILITY void __incr(true_type) _NOEXCEPT - {} - - _LIBCPP_INLINE_VISIBILITY void __set(size_t __s, false_type) _NOEXCEPT - {__size_ = __s;} - _LIBCPP_INLINE_VISIBILITY void __set(size_t, true_type) _NOEXCEPT - {} -public: - _LIBCPP_INLINE_VISIBILITY explicit __destruct_n(size_t __s) _NOEXCEPT - : __size_(__s) {} - - template - _LIBCPP_INLINE_VISIBILITY void __incr() _NOEXCEPT - {__incr(integral_constant::value>());} - - template - _LIBCPP_INLINE_VISIBILITY void __set(size_t __s, _Tp*) _NOEXCEPT - {__set(__s, integral_constant::value>());} - - template - _LIBCPP_INLINE_VISIBILITY void operator()(_Tp* __p) _NOEXCEPT - {__process(__p, integral_constant::value>());} -}; - -template -class __allocator_destructor -{ - typedef _LIBCPP_NODEBUG_TYPE allocator_traits<_Alloc> __alloc_traits; -public: - typedef _LIBCPP_NODEBUG_TYPE typename __alloc_traits::pointer pointer; - typedef _LIBCPP_NODEBUG_TYPE typename __alloc_traits::size_type size_type; -private: - _Alloc& __alloc_; - size_type __s_; -public: - _LIBCPP_INLINE_VISIBILITY __allocator_destructor(_Alloc& __a, size_type __s) - _NOEXCEPT - : __alloc_(__a), __s_(__s) {} - _LIBCPP_INLINE_VISIBILITY - void operator()(pointer __p) _NOEXCEPT - {__alloc_traits::deallocate(__alloc_, __p, __s_);} -}; - -template -_ForwardIterator -uninitialized_copy(_InputIterator __f, _InputIterator __l, _ForwardIterator __r) -{ - typedef typename iterator_traits<_ForwardIterator>::value_type value_type; -#ifndef _LIBCPP_NO_EXCEPTIONS - _ForwardIterator __s = __r; - try - { -#endif - for (; __f != __l; ++__f, (void) ++__r) - ::new ((void*)_VSTD::addressof(*__r)) value_type(*__f); -#ifndef _LIBCPP_NO_EXCEPTIONS - } - catch (...) - { - for (; __s != __r; ++__s) - __s->~value_type(); - throw; - } -#endif - return __r; -} - -template -_ForwardIterator -uninitialized_copy_n(_InputIterator __f, _Size __n, _ForwardIterator __r) -{ - typedef typename iterator_traits<_ForwardIterator>::value_type value_type; -#ifndef _LIBCPP_NO_EXCEPTIONS - _ForwardIterator __s = __r; - try - { -#endif - for (; __n > 0; ++__f, (void) ++__r, (void) --__n) - ::new ((void*)_VSTD::addressof(*__r)) value_type(*__f); -#ifndef _LIBCPP_NO_EXCEPTIONS - } - catch (...) - { - for (; __s != __r; ++__s) - __s->~value_type(); - throw; - } -#endif - return __r; -} - -template -void -uninitialized_fill(_ForwardIterator __f, _ForwardIterator __l, const _Tp& __x) -{ - typedef typename iterator_traits<_ForwardIterator>::value_type value_type; -#ifndef _LIBCPP_NO_EXCEPTIONS - _ForwardIterator __s = __f; - try - { -#endif - for (; __f != __l; ++__f) - ::new ((void*)_VSTD::addressof(*__f)) value_type(__x); -#ifndef _LIBCPP_NO_EXCEPTIONS - } - catch (...) - { - for (; __s != __f; ++__s) - __s->~value_type(); - throw; - } -#endif -} - -template -_ForwardIterator -uninitialized_fill_n(_ForwardIterator __f, _Size __n, const _Tp& __x) -{ - typedef typename iterator_traits<_ForwardIterator>::value_type value_type; -#ifndef _LIBCPP_NO_EXCEPTIONS - _ForwardIterator __s = __f; - try - { -#endif - for (; __n > 0; ++__f, (void) --__n) - ::new ((void*)_VSTD::addressof(*__f)) value_type(__x); -#ifndef _LIBCPP_NO_EXCEPTIONS - } - catch (...) - { - for (; __s != __f; ++__s) - __s->~value_type(); - throw; - } -#endif - return __f; -} - -#if _LIBCPP_STD_VER > 14 - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -void destroy(_ForwardIterator __first, _ForwardIterator __last) { - for (; __first != __last; ++__first) - _VSTD::destroy_at(_VSTD::addressof(*__first)); -} - -template -inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 -_ForwardIterator destroy_n(_ForwardIterator __first, _Size __n) { - for (; __n > 0; (void)++__first, --__n) - _VSTD::destroy_at(_VSTD::addressof(*__first)); - return __first; -} - -template -inline _LIBCPP_INLINE_VISIBILITY -void uninitialized_default_construct(_ForwardIterator __first, _ForwardIterator __last) { - using _Vt = typename iterator_traits<_ForwardIterator>::value_type; - auto __idx = __first; -#ifndef _LIBCPP_NO_EXCEPTIONS - try { -#endif - for (; __idx != __last; ++__idx) - ::new ((void*)_VSTD::addressof(*__idx)) _Vt; -#ifndef _LIBCPP_NO_EXCEPTIONS - } catch (...) { - _VSTD::destroy(__first, __idx); - throw; - } -#endif -} - -template -inline _LIBCPP_INLINE_VISIBILITY -_ForwardIterator uninitialized_default_construct_n(_ForwardIterator __first, _Size __n) { - using _Vt = typename iterator_traits<_ForwardIterator>::value_type; - auto __idx = __first; -#ifndef _LIBCPP_NO_EXCEPTIONS - try { -#endif - for (; __n > 0; (void)++__idx, --__n) - ::new ((void*)_VSTD::addressof(*__idx)) _Vt; - return __idx; -#ifndef _LIBCPP_NO_EXCEPTIONS - } catch (...) { - _VSTD::destroy(__first, __idx); - throw; - } -#endif -} - - -template -inline _LIBCPP_INLINE_VISIBILITY -void uninitialized_value_construct(_ForwardIterator __first, _ForwardIterator __last) { - using _Vt = typename iterator_traits<_ForwardIterator>::value_type; - auto __idx = __first; -#ifndef _LIBCPP_NO_EXCEPTIONS - try { -#endif - for (; __idx != __last; ++__idx) - ::new ((void*)_VSTD::addressof(*__idx)) _Vt(); -#ifndef _LIBCPP_NO_EXCEPTIONS - } catch (...) { - _VSTD::destroy(__first, __idx); - throw; - } -#endif -} - -template -inline _LIBCPP_INLINE_VISIBILITY -_ForwardIterator uninitialized_value_construct_n(_ForwardIterator __first, _Size __n) { - using _Vt = typename iterator_traits<_ForwardIterator>::value_type; - auto __idx = __first; -#ifndef _LIBCPP_NO_EXCEPTIONS - try { -#endif - for (; __n > 0; (void)++__idx, --__n) - ::new ((void*)_VSTD::addressof(*__idx)) _Vt(); - return __idx; -#ifndef _LIBCPP_NO_EXCEPTIONS - } catch (...) { - _VSTD::destroy(__first, __idx); - throw; - } -#endif -} - - -template -inline _LIBCPP_INLINE_VISIBILITY -_ForwardIt uninitialized_move(_InputIt __first, _InputIt __last, _ForwardIt __first_res) { - using _Vt = typename iterator_traits<_ForwardIt>::value_type; - auto __idx = __first_res; -#ifndef _LIBCPP_NO_EXCEPTIONS - try { -#endif - for (; __first != __last; (void)++__idx, ++__first) - ::new ((void*)_VSTD::addressof(*__idx)) _Vt(_VSTD::move(*__first)); - return __idx; -#ifndef _LIBCPP_NO_EXCEPTIONS - } catch (...) { - _VSTD::destroy(__first_res, __idx); - throw; - } -#endif -} - -template -inline _LIBCPP_INLINE_VISIBILITY -pair<_InputIt, _ForwardIt> -uninitialized_move_n(_InputIt __first, _Size __n, _ForwardIt __first_res) { - using _Vt = typename iterator_traits<_ForwardIt>::value_type; - auto __idx = __first_res; -#ifndef _LIBCPP_NO_EXCEPTIONS - try { -#endif - for (; __n > 0; ++__idx, (void)++__first, --__n) - ::new ((void*)_VSTD::addressof(*__idx)) _Vt(_VSTD::move(*__first)); - return {__first, __idx}; -#ifndef _LIBCPP_NO_EXCEPTIONS - } catch (...) { - _VSTD::destroy(__first_res, __idx); - throw; - } -#endif -} - - -#endif // _LIBCPP_STD_VER > 14 - -// NOTE: Relaxed and acq/rel atomics (for increment and decrement respectively) -// should be sufficient for thread safety. -// See https://bugs.llvm.org/show_bug.cgi?id=22803 -#if defined(__clang__) && __has_builtin(__atomic_add_fetch) \ - && defined(__ATOMIC_RELAXED) \ - && defined(__ATOMIC_ACQ_REL) -# define _LIBCPP_HAS_BUILTIN_ATOMIC_SUPPORT -#elif defined(_LIBCPP_COMPILER_GCC) -# define _LIBCPP_HAS_BUILTIN_ATOMIC_SUPPORT -#endif - -template -inline _LIBCPP_INLINE_VISIBILITY _Tp -__libcpp_atomic_refcount_increment(_Tp& __t) _NOEXCEPT -{ -#if defined(_LIBCPP_HAS_BUILTIN_ATOMIC_SUPPORT) && !defined(_LIBCPP_HAS_NO_THREADS) - return __atomic_add_fetch(&__t, 1, __ATOMIC_RELAXED); -#else - return __t += 1; -#endif -} - -template -inline _LIBCPP_INLINE_VISIBILITY _Tp -__libcpp_atomic_refcount_decrement(_Tp& __t) _NOEXCEPT -{ -#if defined(_LIBCPP_HAS_BUILTIN_ATOMIC_SUPPORT) && !defined(_LIBCPP_HAS_NO_THREADS) - return __atomic_add_fetch(&__t, -1, __ATOMIC_ACQ_REL); -#else - return __t -= 1; -#endif -} - -class _LIBCPP_EXCEPTION_ABI bad_weak_ptr - : public std::exception -{ -public: - bad_weak_ptr() _NOEXCEPT = default; - bad_weak_ptr(const bad_weak_ptr&) _NOEXCEPT = default; - virtual ~bad_weak_ptr() _NOEXCEPT; - virtual const char* what() const _NOEXCEPT; -}; - -_LIBCPP_NORETURN inline _LIBCPP_INLINE_VISIBILITY -void __throw_bad_weak_ptr() -{ -#ifndef _LIBCPP_NO_EXCEPTIONS - throw bad_weak_ptr(); -#else - _VSTD::abort(); -#endif -} - -template class _LIBCPP_TEMPLATE_VIS weak_ptr; - -class _LIBCPP_TYPE_VIS __shared_count -{ - __shared_count(const __shared_count&); - __shared_count& operator=(const __shared_count&); - -protected: - long __shared_owners_; - virtual ~__shared_count(); -private: - virtual void __on_zero_shared() _NOEXCEPT = 0; - -public: - _LIBCPP_INLINE_VISIBILITY - explicit __shared_count(long __refs = 0) _NOEXCEPT - : __shared_owners_(__refs) {} - -#if defined(_LIBCPP_BUILDING_LIBRARY) && \ - defined(_LIBCPP_DEPRECATED_ABI_LEGACY_LIBRARY_DEFINITIONS_FOR_INLINE_FUNCTIONS) - void __add_shared() _NOEXCEPT; - bool __release_shared() _NOEXCEPT; -#else - _LIBCPP_INLINE_VISIBILITY - void __add_shared() _NOEXCEPT { - __libcpp_atomic_refcount_increment(__shared_owners_); - } - _LIBCPP_INLINE_VISIBILITY - bool __release_shared() _NOEXCEPT { - if (__libcpp_atomic_refcount_decrement(__shared_owners_) == -1) { - __on_zero_shared(); - return true; - } - return false; - } -#endif - _LIBCPP_INLINE_VISIBILITY - long use_count() const _NOEXCEPT { - return __libcpp_relaxed_load(&__shared_owners_) + 1; - } -}; - -class _LIBCPP_TYPE_VIS __shared_weak_count - : private __shared_count -{ - long __shared_weak_owners_; - -public: - _LIBCPP_INLINE_VISIBILITY - explicit __shared_weak_count(long __refs = 0) _NOEXCEPT - : __shared_count(__refs), - __shared_weak_owners_(__refs) {} -protected: - virtual ~__shared_weak_count(); - -public: -#if defined(_LIBCPP_BUILDING_LIBRARY) && \ - defined(_LIBCPP_DEPRECATED_ABI_LEGACY_LIBRARY_DEFINITIONS_FOR_INLINE_FUNCTIONS) - void __add_shared() _NOEXCEPT; - void __add_weak() _NOEXCEPT; - void __release_shared() _NOEXCEPT; -#else - _LIBCPP_INLINE_VISIBILITY - void __add_shared() _NOEXCEPT { - __shared_count::__add_shared(); - } - _LIBCPP_INLINE_VISIBILITY - void __add_weak() _NOEXCEPT { - __libcpp_atomic_refcount_increment(__shared_weak_owners_); - } - _LIBCPP_INLINE_VISIBILITY - void __release_shared() _NOEXCEPT { - if (__shared_count::__release_shared()) - __release_weak(); - } -#endif - void __release_weak() _NOEXCEPT; - _LIBCPP_INLINE_VISIBILITY - long use_count() const _NOEXCEPT {return __shared_count::use_count();} - __shared_weak_count* lock() _NOEXCEPT; - - virtual const void* __get_deleter(const type_info&) const _NOEXCEPT; -private: - virtual void __on_zero_shared_weak() _NOEXCEPT = 0; -}; - -template -class __shared_ptr_pointer - : public __shared_weak_count -{ - __compressed_pair<__compressed_pair<_Tp, _Dp>, _Alloc> __data_; -public: - _LIBCPP_INLINE_VISIBILITY - __shared_ptr_pointer(_Tp __p, _Dp __d, _Alloc __a) - : __data_(__compressed_pair<_Tp, _Dp>(__p, _VSTD::move(__d)), _VSTD::move(__a)) {} - -#ifndef _LIBCPP_NO_RTTI - virtual const void* __get_deleter(const type_info&) const _NOEXCEPT; -#endif - -private: - virtual void __on_zero_shared() _NOEXCEPT; - virtual void __on_zero_shared_weak() _NOEXCEPT; -}; - -#ifndef _LIBCPP_NO_RTTI - -template -const void* -__shared_ptr_pointer<_Tp, _Dp, _Alloc>::__get_deleter(const type_info& __t) const _NOEXCEPT -{ - return __t == typeid(_Dp) ? _VSTD::addressof(__data_.first().second()) : nullptr; -} - -#endif // _LIBCPP_NO_RTTI - -template -void -__shared_ptr_pointer<_Tp, _Dp, _Alloc>::__on_zero_shared() _NOEXCEPT -{ - __data_.first().second()(__data_.first().first()); - __data_.first().second().~_Dp(); -} - -template -void -__shared_ptr_pointer<_Tp, _Dp, _Alloc>::__on_zero_shared_weak() _NOEXCEPT -{ - typedef typename __allocator_traits_rebind<_Alloc, __shared_ptr_pointer>::type _Al; - typedef allocator_traits<_Al> _ATraits; - typedef pointer_traits _PTraits; - - _Al __a(__data_.second()); - __data_.second().~_Alloc(); - __a.deallocate(_PTraits::pointer_to(*this), 1); -} - -template -struct __shared_ptr_emplace - : __shared_weak_count -{ - template - _LIBCPP_HIDE_FROM_ABI - explicit __shared_ptr_emplace(_Alloc __a, _Args&& ...__args) - : __storage_(_VSTD::move(__a)) - { -#if _LIBCPP_STD_VER > 17 - using _TpAlloc = typename __allocator_traits_rebind<_Alloc, _Tp>::type; - _TpAlloc __tmp(*__get_alloc()); - allocator_traits<_TpAlloc>::construct(__tmp, __get_elem(), _VSTD::forward<_Args>(__args)...); -#else - ::new ((void*)__get_elem()) _Tp(_VSTD::forward<_Args>(__args)...); -#endif - } - - _LIBCPP_HIDE_FROM_ABI - _Alloc* __get_alloc() _NOEXCEPT { return __storage_.__get_alloc(); } - - _LIBCPP_HIDE_FROM_ABI - _Tp* __get_elem() _NOEXCEPT { return __storage_.__get_elem(); } - -private: - virtual void __on_zero_shared() _NOEXCEPT { -#if _LIBCPP_STD_VER > 17 - using _TpAlloc = typename __allocator_traits_rebind<_Alloc, _Tp>::type; - _TpAlloc __tmp(*__get_alloc()); - allocator_traits<_TpAlloc>::destroy(__tmp, __get_elem()); -#else - __get_elem()->~_Tp(); -#endif - } - - virtual void __on_zero_shared_weak() _NOEXCEPT { - using _ControlBlockAlloc = typename __allocator_traits_rebind<_Alloc, __shared_ptr_emplace>::type; - using _ControlBlockPointer = typename allocator_traits<_ControlBlockAlloc>::pointer; - _ControlBlockAlloc __tmp(*__get_alloc()); - __storage_.~_Storage(); - allocator_traits<_ControlBlockAlloc>::deallocate(__tmp, - pointer_traits<_ControlBlockPointer>::pointer_to(*this), 1); - } - - // This class implements the control block for non-array shared pointers created - // through `std::allocate_shared` and `std::make_shared`. - // - // In previous versions of the library, we used a compressed pair to store - // both the _Alloc and the _Tp. This implies using EBO, which is incompatible - // with Allocator construction for _Tp. To allow implementing P0674 in C++20, - // we now use a properly aligned char buffer while making sure that we maintain - // the same layout that we had when we used a compressed pair. - using _CompressedPair = __compressed_pair<_Alloc, _Tp>; - struct _ALIGNAS_TYPE(_CompressedPair) _Storage { - char __blob_[sizeof(_CompressedPair)]; - - _LIBCPP_HIDE_FROM_ABI explicit _Storage(_Alloc&& __a) { - ::new ((void*)__get_alloc()) _Alloc(_VSTD::move(__a)); - } - _LIBCPP_HIDE_FROM_ABI ~_Storage() { - __get_alloc()->~_Alloc(); - } - _Alloc* __get_alloc() _NOEXCEPT { - _CompressedPair *__as_pair = reinterpret_cast<_CompressedPair*>(__blob_); - typename _CompressedPair::_Base1* __first = _CompressedPair::__get_first_base(__as_pair); - _Alloc *__alloc = reinterpret_cast<_Alloc*>(__first); - return __alloc; - } - _LIBCPP_NO_CFI _Tp* __get_elem() _NOEXCEPT { - _CompressedPair *__as_pair = reinterpret_cast<_CompressedPair*>(__blob_); - typename _CompressedPair::_Base2* __second = _CompressedPair::__get_second_base(__as_pair); - _Tp *__elem = reinterpret_cast<_Tp*>(__second); - return __elem; - } - }; - - static_assert(_LIBCPP_ALIGNOF(_Storage) == _LIBCPP_ALIGNOF(_CompressedPair), ""); - static_assert(sizeof(_Storage) == sizeof(_CompressedPair), ""); - _Storage __storage_; -}; - -struct __shared_ptr_dummy_rebind_allocator_type; -template <> -class _LIBCPP_TEMPLATE_VIS allocator<__shared_ptr_dummy_rebind_allocator_type> -{ -public: - template - struct rebind - { - typedef allocator<_Other> other; - }; -}; - -template class _LIBCPP_TEMPLATE_VIS enable_shared_from_this; - -template -struct __compatible_with -#if _LIBCPP_STD_VER > 14 - : is_convertible*, remove_extent_t<_Up>*> {}; -#else - : is_convertible<_Tp*, _Up*> {}; -#endif // _LIBCPP_STD_VER > 14 - -#if defined(_LIBCPP_ABI_ENABLE_SHARED_PTR_TRIVIAL_ABI) -# define _LIBCPP_SHARED_PTR_TRIVIAL_ABI __attribute__((trivial_abi)) -#else -# define _LIBCPP_SHARED_PTR_TRIVIAL_ABI -#endif - -template -class _LIBCPP_SHARED_PTR_TRIVIAL_ABI _LIBCPP_TEMPLATE_VIS shared_ptr -{ -public: -#if _LIBCPP_STD_VER > 14 - typedef weak_ptr<_Tp> weak_type; - typedef remove_extent_t<_Tp> element_type; -#else - typedef _Tp element_type; -#endif - -private: - element_type* __ptr_; - __shared_weak_count* __cntrl_; - - struct __nat {int __for_bool_;}; -public: - _LIBCPP_INLINE_VISIBILITY - _LIBCPP_CONSTEXPR shared_ptr() _NOEXCEPT; - _LIBCPP_INLINE_VISIBILITY - _LIBCPP_CONSTEXPR shared_ptr(nullptr_t) _NOEXCEPT; - template - explicit shared_ptr(_Yp* __p, - typename enable_if<__compatible_with<_Yp, element_type>::value, __nat>::type = __nat()); - template - shared_ptr(_Yp* __p, _Dp __d, - typename enable_if<__compatible_with<_Yp, element_type>::value, __nat>::type = __nat()); - template - shared_ptr(_Yp* __p, _Dp __d, _Alloc __a, - typename enable_if<__compatible_with<_Yp, element_type>::value, __nat>::type = __nat()); - template shared_ptr(nullptr_t __p, _Dp __d); - template shared_ptr(nullptr_t __p, _Dp __d, _Alloc __a); - template _LIBCPP_INLINE_VISIBILITY shared_ptr(const shared_ptr<_Yp>& __r, element_type* __p) _NOEXCEPT; - _LIBCPP_INLINE_VISIBILITY - shared_ptr(const shared_ptr& __r) _NOEXCEPT; - template - _LIBCPP_INLINE_VISIBILITY - shared_ptr(const shared_ptr<_Yp>& __r, - typename enable_if<__compatible_with<_Yp, element_type>::value, __nat>::type = __nat()) - _NOEXCEPT; - _LIBCPP_INLINE_VISIBILITY - shared_ptr(shared_ptr&& __r) _NOEXCEPT; - template _LIBCPP_INLINE_VISIBILITY shared_ptr(shared_ptr<_Yp>&& __r, - typename enable_if<__compatible_with<_Yp, element_type>::value, __nat>::type = __nat()) - _NOEXCEPT; - template explicit shared_ptr(const weak_ptr<_Yp>& __r, - typename enable_if::value, __nat>::type= __nat()); -#if _LIBCPP_STD_VER <= 14 || defined(_LIBCPP_ENABLE_CXX17_REMOVED_AUTO_PTR) - template - shared_ptr(auto_ptr<_Yp>&& __r, - typename enable_if::value, __nat>::type = __nat()); -#endif - template - shared_ptr(unique_ptr<_Yp, _Dp>&&, - typename enable_if - < - !is_lvalue_reference<_Dp>::value && - !is_array<_Yp>::value && - is_convertible::pointer, element_type*>::value, - __nat - >::type = __nat()); - template - shared_ptr(unique_ptr<_Yp, _Dp>&&, - typename enable_if - < - is_lvalue_reference<_Dp>::value && - !is_array<_Yp>::value && - is_convertible::pointer, element_type*>::value, - __nat - >::type = __nat()); - - ~shared_ptr(); - - _LIBCPP_INLINE_VISIBILITY - shared_ptr& operator=(const shared_ptr& __r) _NOEXCEPT; - template - typename enable_if - < - __compatible_with<_Yp, element_type>::value, - shared_ptr& - >::type - _LIBCPP_INLINE_VISIBILITY - operator=(const shared_ptr<_Yp>& __r) _NOEXCEPT; - _LIBCPP_INLINE_VISIBILITY - shared_ptr& operator=(shared_ptr&& __r) _NOEXCEPT; - template - typename enable_if - < - __compatible_with<_Yp, element_type>::value, - shared_ptr& - >::type - _LIBCPP_INLINE_VISIBILITY - operator=(shared_ptr<_Yp>&& __r); -#if _LIBCPP_STD_VER <= 14 || defined(_LIBCPP_ENABLE_CXX17_REMOVED_AUTO_PTR) - template - _LIBCPP_INLINE_VISIBILITY - typename enable_if - < - !is_array<_Yp>::value && - is_convertible<_Yp*, element_type*>::value, - shared_ptr - >::type& - operator=(auto_ptr<_Yp>&& __r); -#endif - template - typename enable_if - < - !is_array<_Yp>::value && - is_convertible::pointer, element_type*>::value, - shared_ptr& - >::type - _LIBCPP_INLINE_VISIBILITY - operator=(unique_ptr<_Yp, _Dp>&& __r); - - _LIBCPP_INLINE_VISIBILITY - void swap(shared_ptr& __r) _NOEXCEPT; - _LIBCPP_INLINE_VISIBILITY - void reset() _NOEXCEPT; - template - typename enable_if - < - __compatible_with<_Yp, element_type>::value, - void - >::type - _LIBCPP_INLINE_VISIBILITY - reset(_Yp* __p); - template - typename enable_if - < - __compatible_with<_Yp, element_type>::value, - void - >::type - _LIBCPP_INLINE_VISIBILITY - reset(_Yp* __p, _Dp __d); - template - typename enable_if - < - __compatible_with<_Yp, element_type>::value, - void - >::type - _LIBCPP_INLINE_VISIBILITY - reset(_Yp* __p, _Dp __d, _Alloc __a); - - _LIBCPP_INLINE_VISIBILITY - element_type* get() const _NOEXCEPT {return __ptr_;} - _LIBCPP_INLINE_VISIBILITY - typename add_lvalue_reference::type operator*() const _NOEXCEPT - {return *__ptr_;} - _LIBCPP_INLINE_VISIBILITY - element_type* operator->() const _NOEXCEPT - { - static_assert(!_VSTD::is_array<_Tp>::value, - "std::shared_ptr::operator-> is only valid when T is not an array type."); - return __ptr_; - } - _LIBCPP_INLINE_VISIBILITY - long use_count() const _NOEXCEPT {return __cntrl_ ? __cntrl_->use_count() : 0;} - _LIBCPP_INLINE_VISIBILITY - bool unique() const _NOEXCEPT {return use_count() == 1;} - _LIBCPP_INLINE_VISIBILITY - _LIBCPP_EXPLICIT operator bool() const _NOEXCEPT {return get() != nullptr;} - template - _LIBCPP_INLINE_VISIBILITY - bool owner_before(shared_ptr<_Up> const& __p) const _NOEXCEPT - {return __cntrl_ < __p.__cntrl_;} - template - _LIBCPP_INLINE_VISIBILITY - bool owner_before(weak_ptr<_Up> const& __p) const _NOEXCEPT - {return __cntrl_ < __p.__cntrl_;} - _LIBCPP_INLINE_VISIBILITY - bool - __owner_equivalent(const shared_ptr& __p) const - {return __cntrl_ == __p.__cntrl_;} - -#if _LIBCPP_STD_VER > 14 - typename add_lvalue_reference::type - _LIBCPP_INLINE_VISIBILITY - operator[](ptrdiff_t __i) const - { - static_assert(_VSTD::is_array<_Tp>::value, - "std::shared_ptr::operator[] is only valid when T is an array type."); - return __ptr_[__i]; - } -#endif - -#ifndef _LIBCPP_NO_RTTI - template - _LIBCPP_INLINE_VISIBILITY - _Dp* __get_deleter() const _NOEXCEPT - {return static_cast<_Dp*>(__cntrl_ - ? const_cast(__cntrl_->__get_deleter(typeid(_Dp))) - : nullptr);} -#endif // _LIBCPP_NO_RTTI - - template - static shared_ptr<_Tp> - __create_with_control_block(_Yp* __p, _CntrlBlk* __cntrl) _NOEXCEPT - { - shared_ptr<_Tp> __r; - __r.__ptr_ = __p; - __r.__cntrl_ = __cntrl; - __r.__enable_weak_this(__r.__ptr_, __r.__ptr_); - return __r; - } - -private: - template ::value> - struct __shared_ptr_default_allocator - { - typedef allocator<_Yp> type; - }; - - template - struct __shared_ptr_default_allocator<_Yp, true> - { - typedef allocator<__shared_ptr_dummy_rebind_allocator_type> type; - }; - - template - _LIBCPP_INLINE_VISIBILITY - typename enable_if* - >::value, - void>::type - __enable_weak_this(const enable_shared_from_this<_Yp>* __e, - _OrigPtr* __ptr) _NOEXCEPT - { - typedef typename remove_cv<_Yp>::type _RawYp; - if (__e && __e->__weak_this_.expired()) - { - __e->__weak_this_ = shared_ptr<_RawYp>(*this, - const_cast<_RawYp*>(static_cast(__ptr))); - } - } - - _LIBCPP_INLINE_VISIBILITY void __enable_weak_this(...) _NOEXCEPT {} - - template - struct __shared_ptr_default_delete - : default_delete<_Yp> {}; - - template - struct __shared_ptr_default_delete<_Yp[_Sz], _Un> - : default_delete<_Yp[]> {}; - - template - struct __shared_ptr_default_delete<_Yp[], _Un> - : default_delete<_Yp[]> {}; - - template friend class _LIBCPP_TEMPLATE_VIS shared_ptr; - template friend class _LIBCPP_TEMPLATE_VIS weak_ptr; -}; - -#ifndef _LIBCPP_HAS_NO_DEDUCTION_GUIDES -template -shared_ptr(weak_ptr<_Tp>) -> shared_ptr<_Tp>; -template -shared_ptr(unique_ptr<_Tp, _Dp>) -> shared_ptr<_Tp>; -#endif - -template -inline -_LIBCPP_CONSTEXPR -shared_ptr<_Tp>::shared_ptr() _NOEXCEPT - : __ptr_(nullptr), - __cntrl_(nullptr) -{ -} - -template -inline -_LIBCPP_CONSTEXPR -shared_ptr<_Tp>::shared_ptr(nullptr_t) _NOEXCEPT - : __ptr_(nullptr), - __cntrl_(nullptr) -{ -} - -template -template -shared_ptr<_Tp>::shared_ptr(_Yp* __p, - typename enable_if<__compatible_with<_Yp, element_type>::value, __nat>::type) - : __ptr_(__p) -{ - unique_ptr<_Yp> __hold(__p); - typedef typename __shared_ptr_default_allocator<_Yp>::type _AllocT; - typedef __shared_ptr_pointer<_Yp*, __shared_ptr_default_delete<_Tp, _Yp>, _AllocT > _CntrlBlk; - __cntrl_ = new _CntrlBlk(__p, __shared_ptr_default_delete<_Tp, _Yp>(), _AllocT()); - __hold.release(); - __enable_weak_this(__p, __p); -} - -template -template -shared_ptr<_Tp>::shared_ptr(_Yp* __p, _Dp __d, - typename enable_if<__compatible_with<_Yp, element_type>::value, __nat>::type) - : __ptr_(__p) -{ -#ifndef _LIBCPP_NO_EXCEPTIONS - try - { -#endif // _LIBCPP_NO_EXCEPTIONS - typedef typename __shared_ptr_default_allocator<_Yp>::type _AllocT; - typedef __shared_ptr_pointer<_Yp*, _Dp, _AllocT > _CntrlBlk; - __cntrl_ = new _CntrlBlk(__p, __d, _AllocT()); - __enable_weak_this(__p, __p); -#ifndef _LIBCPP_NO_EXCEPTIONS - } - catch (...) - { - __d(__p); - throw; - } -#endif // _LIBCPP_NO_EXCEPTIONS -} - -template -template -shared_ptr<_Tp>::shared_ptr(nullptr_t __p, _Dp __d) - : __ptr_(nullptr) -{ -#ifndef _LIBCPP_NO_EXCEPTIONS - try - { -#endif // _LIBCPP_NO_EXCEPTIONS - typedef typename __shared_ptr_default_allocator<_Tp>::type _AllocT; - typedef __shared_ptr_pointer _CntrlBlk; - __cntrl_ = new _CntrlBlk(__p, __d, _AllocT()); -#ifndef _LIBCPP_NO_EXCEPTIONS - } - catch (...) - { - __d(__p); - throw; - } -#endif // _LIBCPP_NO_EXCEPTIONS -} - -template -template -shared_ptr<_Tp>::shared_ptr(_Yp* __p, _Dp __d, _Alloc __a, - typename enable_if<__compatible_with<_Yp, element_type>::value, __nat>::type) - : __ptr_(__p) -{ -#ifndef _LIBCPP_NO_EXCEPTIONS - try - { -#endif // _LIBCPP_NO_EXCEPTIONS - typedef __shared_ptr_pointer<_Yp*, _Dp, _Alloc> _CntrlBlk; - typedef typename __allocator_traits_rebind<_Alloc, _CntrlBlk>::type _A2; - typedef __allocator_destructor<_A2> _D2; - _A2 __a2(__a); - unique_ptr<_CntrlBlk, _D2> __hold2(__a2.allocate(1), _D2(__a2, 1)); - ::new ((void*)_VSTD::addressof(*__hold2.get())) _CntrlBlk(__p, __d, __a); - __cntrl_ = _VSTD::addressof(*__hold2.release()); - __enable_weak_this(__p, __p); -#ifndef _LIBCPP_NO_EXCEPTIONS - } - catch (...) - { - __d(__p); - throw; - } -#endif // _LIBCPP_NO_EXCEPTIONS -} - -template -template -shared_ptr<_Tp>::shared_ptr(nullptr_t __p, _Dp __d, _Alloc __a) - : __ptr_(nullptr) -{ -#ifndef _LIBCPP_NO_EXCEPTIONS - try - { -#endif // _LIBCPP_NO_EXCEPTIONS - typedef __shared_ptr_pointer _CntrlBlk; - typedef typename __allocator_traits_rebind<_Alloc, _CntrlBlk>::type _A2; - typedef __allocator_destructor<_A2> _D2; - _A2 __a2(__a); - unique_ptr<_CntrlBlk, _D2> __hold2(__a2.allocate(1), _D2(__a2, 1)); - ::new ((void*)_VSTD::addressof(*__hold2.get())) _CntrlBlk(__p, __d, __a); - __cntrl_ = _VSTD::addressof(*__hold2.release()); -#ifndef _LIBCPP_NO_EXCEPTIONS - } - catch (...) - { - __d(__p); - throw; - } -#endif // _LIBCPP_NO_EXCEPTIONS -} - -template -template -inline -shared_ptr<_Tp>::shared_ptr(const shared_ptr<_Yp>& __r, element_type *__p) _NOEXCEPT - : __ptr_(__p), - __cntrl_(__r.__cntrl_) -{ - if (__cntrl_) - __cntrl_->__add_shared(); -} - -template -inline -shared_ptr<_Tp>::shared_ptr(const shared_ptr& __r) _NOEXCEPT - : __ptr_(__r.__ptr_), - __cntrl_(__r.__cntrl_) -{ - if (__cntrl_) - __cntrl_->__add_shared(); -} - -template -template -inline -shared_ptr<_Tp>::shared_ptr(const shared_ptr<_Yp>& __r, - typename enable_if<__compatible_with<_Yp, element_type>::value, __nat>::type) - _NOEXCEPT - : __ptr_(__r.__ptr_), - __cntrl_(__r.__cntrl_) -{ - if (__cntrl_) - __cntrl_->__add_shared(); -} - -template -inline -shared_ptr<_Tp>::shared_ptr(shared_ptr&& __r) _NOEXCEPT - : __ptr_(__r.__ptr_), - __cntrl_(__r.__cntrl_) -{ - __r.__ptr_ = nullptr; - __r.__cntrl_ = nullptr; -} - -template -template -inline -shared_ptr<_Tp>::shared_ptr(shared_ptr<_Yp>&& __r, - typename enable_if<__compatible_with<_Yp, element_type>::value, __nat>::type) - _NOEXCEPT - : __ptr_(__r.__ptr_), - __cntrl_(__r.__cntrl_) -{ - __r.__ptr_ = nullptr; - __r.__cntrl_ = nullptr; -} - -#if _LIBCPP_STD_VER <= 14 || defined(_LIBCPP_ENABLE_CXX17_REMOVED_AUTO_PTR) -template -template -shared_ptr<_Tp>::shared_ptr(auto_ptr<_Yp>&& __r, - typename enable_if::value, __nat>::type) - : __ptr_(__r.get()) -{ - typedef __shared_ptr_pointer<_Yp*, default_delete<_Yp>, allocator<_Yp> > _CntrlBlk; - __cntrl_ = new _CntrlBlk(__r.get(), default_delete<_Yp>(), allocator<_Yp>()); - __enable_weak_this(__r.get(), __r.get()); - __r.release(); -} -#endif - -template -template -shared_ptr<_Tp>::shared_ptr(unique_ptr<_Yp, _Dp>&& __r, - typename enable_if - < - !is_lvalue_reference<_Dp>::value && - !is_array<_Yp>::value && - is_convertible::pointer, element_type*>::value, - __nat - >::type) - : __ptr_(__r.get()) -{ -#if _LIBCPP_STD_VER > 11 - if (__ptr_ == nullptr) - __cntrl_ = nullptr; - else -#endif - { - typedef typename __shared_ptr_default_allocator<_Yp>::type _AllocT; - typedef __shared_ptr_pointer<_Yp*, _Dp, _AllocT > _CntrlBlk; - __cntrl_ = new _CntrlBlk(__r.get(), __r.get_deleter(), _AllocT()); - __enable_weak_this(__r.get(), __r.get()); - } - __r.release(); -} - -template -template -shared_ptr<_Tp>::shared_ptr(unique_ptr<_Yp, _Dp>&& __r, - typename enable_if - < - is_lvalue_reference<_Dp>::value && - !is_array<_Yp>::value && - is_convertible::pointer, element_type*>::value, - __nat - >::type) - : __ptr_(__r.get()) -{ -#if _LIBCPP_STD_VER > 11 - if (__ptr_ == nullptr) - __cntrl_ = nullptr; - else -#endif - { - typedef typename __shared_ptr_default_allocator<_Yp>::type _AllocT; - typedef __shared_ptr_pointer<_Yp*, - reference_wrapper::type>, - _AllocT > _CntrlBlk; - __cntrl_ = new _CntrlBlk(__r.get(), _VSTD::ref(__r.get_deleter()), _AllocT()); - __enable_weak_this(__r.get(), __r.get()); - } - __r.release(); -} - -template -shared_ptr<_Tp>::~shared_ptr() -{ - if (__cntrl_) - __cntrl_->__release_shared(); -} - -template -inline -shared_ptr<_Tp>& -shared_ptr<_Tp>::operator=(const shared_ptr& __r) _NOEXCEPT -{ - shared_ptr(__r).swap(*this); - return *this; -} - -template -template -inline -typename enable_if -< - __compatible_with<_Yp, typename shared_ptr<_Tp>::element_type>::value, - shared_ptr<_Tp>& ->::type -shared_ptr<_Tp>::operator=(const shared_ptr<_Yp>& __r) _NOEXCEPT -{ - shared_ptr(__r).swap(*this); - return *this; -} - -template -inline -shared_ptr<_Tp>& -shared_ptr<_Tp>::operator=(shared_ptr&& __r) _NOEXCEPT -{ - shared_ptr(_VSTD::move(__r)).swap(*this); - return *this; -} - -template -template -inline -typename enable_if -< - __compatible_with<_Yp, typename shared_ptr<_Tp>::element_type>::value, - shared_ptr<_Tp>& ->::type -shared_ptr<_Tp>::operator=(shared_ptr<_Yp>&& __r) -{ - shared_ptr(_VSTD::move(__r)).swap(*this); - return *this; -} - -#if _LIBCPP_STD_VER <= 14 || defined(_LIBCPP_ENABLE_CXX17_REMOVED_AUTO_PTR) -template -template -inline -typename enable_if -< - !is_array<_Yp>::value && - is_convertible<_Yp*, typename shared_ptr<_Tp>::element_type*>::value, - shared_ptr<_Tp> ->::type& -shared_ptr<_Tp>::operator=(auto_ptr<_Yp>&& __r) -{ - shared_ptr(_VSTD::move(__r)).swap(*this); - return *this; -} -#endif - -template -template -inline -typename enable_if -< - !is_array<_Yp>::value && - is_convertible::pointer, - typename shared_ptr<_Tp>::element_type*>::value, - shared_ptr<_Tp>& ->::type -shared_ptr<_Tp>::operator=(unique_ptr<_Yp, _Dp>&& __r) -{ - shared_ptr(_VSTD::move(__r)).swap(*this); - return *this; -} - -template -inline -void -shared_ptr<_Tp>::swap(shared_ptr& __r) _NOEXCEPT -{ - _VSTD::swap(__ptr_, __r.__ptr_); - _VSTD::swap(__cntrl_, __r.__cntrl_); -} - -template -inline -void -shared_ptr<_Tp>::reset() _NOEXCEPT -{ - shared_ptr().swap(*this); -} - -template -template -inline -typename enable_if -< - __compatible_with<_Yp, typename shared_ptr<_Tp>::element_type>::value, - void ->::type -shared_ptr<_Tp>::reset(_Yp* __p) -{ - shared_ptr(__p).swap(*this); -} - -template -template -inline -typename enable_if -< - __compatible_with<_Yp, typename shared_ptr<_Tp>::element_type>::value, - void ->::type -shared_ptr<_Tp>::reset(_Yp* __p, _Dp __d) -{ - shared_ptr(__p, __d).swap(*this); -} - -template -template -inline -typename enable_if -< - __compatible_with<_Yp, typename shared_ptr<_Tp>::element_type>::value, - void ->::type -shared_ptr<_Tp>::reset(_Yp* __p, _Dp __d, _Alloc __a) -{ - shared_ptr(__p, __d, __a).swap(*this); -} - -// -// std::allocate_shared and std::make_shared -// -template::value> > -_LIBCPP_HIDE_FROM_ABI -shared_ptr<_Tp> allocate_shared(const _Alloc& __a, _Args&& ...__args) -{ - using _ControlBlock = __shared_ptr_emplace<_Tp, _Alloc>; - using _ControlBlockAllocator = typename __allocator_traits_rebind<_Alloc, _ControlBlock>::type; - __allocation_guard<_ControlBlockAllocator> __guard(__a, 1); - ::new ((void*)_VSTD::addressof(*__guard.__get())) _ControlBlock(__a, _VSTD::forward<_Args>(__args)...); - auto __control_block = __guard.__release_ptr(); - return shared_ptr<_Tp>::__create_with_control_block((*__control_block).__get_elem(), _VSTD::addressof(*__control_block)); -} - -template::value> > -_LIBCPP_HIDE_FROM_ABI -shared_ptr<_Tp> make_shared(_Args&& ...__args) -{ - return _VSTD::allocate_shared<_Tp>(allocator<_Tp>(), _VSTD::forward<_Args>(__args)...); -} - -template -inline _LIBCPP_INLINE_VISIBILITY -bool -operator==(const shared_ptr<_Tp>& __x, const shared_ptr<_Up>& __y) _NOEXCEPT -{ - return __x.get() == __y.get(); -} - -template -inline _LIBCPP_INLINE_VISIBILITY -bool -operator!=(const shared_ptr<_Tp>& __x, const shared_ptr<_Up>& __y) _NOEXCEPT -{ - return !(__x == __y); -} - -template -inline _LIBCPP_INLINE_VISIBILITY -bool -operator<(const shared_ptr<_Tp>& __x, const shared_ptr<_Up>& __y) _NOEXCEPT -{ -#if _LIBCPP_STD_VER <= 11 - typedef typename common_type<_Tp*, _Up*>::type _Vp; - return less<_Vp>()(__x.get(), __y.get()); -#else - return less<>()(__x.get(), __y.get()); -#endif - -} - -template -inline _LIBCPP_INLINE_VISIBILITY -bool -operator>(const shared_ptr<_Tp>& __x, const shared_ptr<_Up>& __y) _NOEXCEPT -{ - return __y < __x; -} - -template -inline _LIBCPP_INLINE_VISIBILITY -bool -operator<=(const shared_ptr<_Tp>& __x, const shared_ptr<_Up>& __y) _NOEXCEPT -{ - return !(__y < __x); -} - -template -inline _LIBCPP_INLINE_VISIBILITY -bool -operator>=(const shared_ptr<_Tp>& __x, const shared_ptr<_Up>& __y) _NOEXCEPT -{ - return !(__x < __y); -} - -template -inline _LIBCPP_INLINE_VISIBILITY -bool -operator==(const shared_ptr<_Tp>& __x, nullptr_t) _NOEXCEPT -{ - return !__x; -} - -template -inline _LIBCPP_INLINE_VISIBILITY -bool -operator==(nullptr_t, const shared_ptr<_Tp>& __x) _NOEXCEPT -{ - return !__x; -} - -template -inline _LIBCPP_INLINE_VISIBILITY -bool -operator!=(const shared_ptr<_Tp>& __x, nullptr_t) _NOEXCEPT -{ - return static_cast(__x); -} - -template -inline _LIBCPP_INLINE_VISIBILITY -bool -operator!=(nullptr_t, const shared_ptr<_Tp>& __x) _NOEXCEPT -{ - return static_cast(__x); -} - -template -inline _LIBCPP_INLINE_VISIBILITY -bool -operator<(const shared_ptr<_Tp>& __x, nullptr_t) _NOEXCEPT -{ - return less<_Tp*>()(__x.get(), nullptr); -} - -template -inline _LIBCPP_INLINE_VISIBILITY -bool -operator<(nullptr_t, const shared_ptr<_Tp>& __x) _NOEXCEPT -{ - return less<_Tp*>()(nullptr, __x.get()); -} - -template -inline _LIBCPP_INLINE_VISIBILITY -bool -operator>(const shared_ptr<_Tp>& __x, nullptr_t) _NOEXCEPT -{ - return nullptr < __x; -} - -template -inline _LIBCPP_INLINE_VISIBILITY -bool -operator>(nullptr_t, const shared_ptr<_Tp>& __x) _NOEXCEPT -{ - return __x < nullptr; -} - -template -inline _LIBCPP_INLINE_VISIBILITY -bool -operator<=(const shared_ptr<_Tp>& __x, nullptr_t) _NOEXCEPT -{ - return !(nullptr < __x); -} - -template -inline _LIBCPP_INLINE_VISIBILITY -bool -operator<=(nullptr_t, const shared_ptr<_Tp>& __x) _NOEXCEPT -{ - return !(__x < nullptr); -} - -template -inline _LIBCPP_INLINE_VISIBILITY -bool -operator>=(const shared_ptr<_Tp>& __x, nullptr_t) _NOEXCEPT -{ - return !(__x < nullptr); -} - -template -inline _LIBCPP_INLINE_VISIBILITY -bool -operator>=(nullptr_t, const shared_ptr<_Tp>& __x) _NOEXCEPT -{ - return !(nullptr < __x); -} - -template -inline _LIBCPP_INLINE_VISIBILITY -void -swap(shared_ptr<_Tp>& __x, shared_ptr<_Tp>& __y) _NOEXCEPT -{ - __x.swap(__y); -} - -template -inline _LIBCPP_INLINE_VISIBILITY -shared_ptr<_Tp> -static_pointer_cast(const shared_ptr<_Up>& __r) _NOEXCEPT -{ - return shared_ptr<_Tp>(__r, - static_cast< - typename shared_ptr<_Tp>::element_type*>(__r.get())); -} - -template -inline _LIBCPP_INLINE_VISIBILITY -shared_ptr<_Tp> -dynamic_pointer_cast(const shared_ptr<_Up>& __r) _NOEXCEPT -{ - typedef typename shared_ptr<_Tp>::element_type _ET; - _ET* __p = dynamic_cast<_ET*>(__r.get()); - return __p ? shared_ptr<_Tp>(__r, __p) : shared_ptr<_Tp>(); -} - -template -shared_ptr<_Tp> -const_pointer_cast(const shared_ptr<_Up>& __r) _NOEXCEPT -{ - typedef typename shared_ptr<_Tp>::element_type _RTp; - return shared_ptr<_Tp>(__r, const_cast<_RTp*>(__r.get())); -} - -template -shared_ptr<_Tp> -reinterpret_pointer_cast(const shared_ptr<_Up>& __r) _NOEXCEPT -{ - return shared_ptr<_Tp>(__r, - reinterpret_cast< - typename shared_ptr<_Tp>::element_type*>(__r.get())); -} - -#ifndef _LIBCPP_NO_RTTI - -template -inline _LIBCPP_INLINE_VISIBILITY -_Dp* -get_deleter(const shared_ptr<_Tp>& __p) _NOEXCEPT -{ - return __p.template __get_deleter<_Dp>(); -} - -#endif // _LIBCPP_NO_RTTI - -template -class _LIBCPP_SHARED_PTR_TRIVIAL_ABI _LIBCPP_TEMPLATE_VIS weak_ptr -{ -public: - typedef _Tp element_type; -private: - element_type* __ptr_; - __shared_weak_count* __cntrl_; - -public: - _LIBCPP_INLINE_VISIBILITY - _LIBCPP_CONSTEXPR weak_ptr() _NOEXCEPT; - template _LIBCPP_INLINE_VISIBILITY weak_ptr(shared_ptr<_Yp> const& __r, - typename enable_if::value, __nat*>::type = 0) - _NOEXCEPT; - _LIBCPP_INLINE_VISIBILITY - weak_ptr(weak_ptr const& __r) _NOEXCEPT; - template _LIBCPP_INLINE_VISIBILITY weak_ptr(weak_ptr<_Yp> const& __r, - typename enable_if::value, __nat*>::type = 0) - _NOEXCEPT; - - _LIBCPP_INLINE_VISIBILITY - weak_ptr(weak_ptr&& __r) _NOEXCEPT; - template _LIBCPP_INLINE_VISIBILITY weak_ptr(weak_ptr<_Yp>&& __r, - typename enable_if::value, __nat*>::type = 0) - _NOEXCEPT; - ~weak_ptr(); - - _LIBCPP_INLINE_VISIBILITY - weak_ptr& operator=(weak_ptr const& __r) _NOEXCEPT; - template - typename enable_if - < - is_convertible<_Yp*, element_type*>::value, - weak_ptr& - >::type - _LIBCPP_INLINE_VISIBILITY - operator=(weak_ptr<_Yp> const& __r) _NOEXCEPT; - - _LIBCPP_INLINE_VISIBILITY - weak_ptr& operator=(weak_ptr&& __r) _NOEXCEPT; - template - typename enable_if - < - is_convertible<_Yp*, element_type*>::value, - weak_ptr& - >::type - _LIBCPP_INLINE_VISIBILITY - operator=(weak_ptr<_Yp>&& __r) _NOEXCEPT; - - template - typename enable_if - < - is_convertible<_Yp*, element_type*>::value, - weak_ptr& - >::type - _LIBCPP_INLINE_VISIBILITY - operator=(shared_ptr<_Yp> const& __r) _NOEXCEPT; - - _LIBCPP_INLINE_VISIBILITY - void swap(weak_ptr& __r) _NOEXCEPT; - _LIBCPP_INLINE_VISIBILITY - void reset() _NOEXCEPT; - - _LIBCPP_INLINE_VISIBILITY - long use_count() const _NOEXCEPT - {return __cntrl_ ? __cntrl_->use_count() : 0;} - _LIBCPP_INLINE_VISIBILITY - bool expired() const _NOEXCEPT - {return __cntrl_ == nullptr || __cntrl_->use_count() == 0;} - shared_ptr<_Tp> lock() const _NOEXCEPT; - template - _LIBCPP_INLINE_VISIBILITY - bool owner_before(const shared_ptr<_Up>& __r) const _NOEXCEPT - {return __cntrl_ < __r.__cntrl_;} - template - _LIBCPP_INLINE_VISIBILITY - bool owner_before(const weak_ptr<_Up>& __r) const _NOEXCEPT - {return __cntrl_ < __r.__cntrl_;} - - template friend class _LIBCPP_TEMPLATE_VIS weak_ptr; - template friend class _LIBCPP_TEMPLATE_VIS shared_ptr; -}; - -#ifndef _LIBCPP_HAS_NO_DEDUCTION_GUIDES -template -weak_ptr(shared_ptr<_Tp>) -> weak_ptr<_Tp>; -#endif - -template -inline -_LIBCPP_CONSTEXPR -weak_ptr<_Tp>::weak_ptr() _NOEXCEPT - : __ptr_(nullptr), - __cntrl_(nullptr) -{ -} - -template -inline -weak_ptr<_Tp>::weak_ptr(weak_ptr const& __r) _NOEXCEPT - : __ptr_(__r.__ptr_), - __cntrl_(__r.__cntrl_) -{ - if (__cntrl_) - __cntrl_->__add_weak(); -} - -template -template -inline -weak_ptr<_Tp>::weak_ptr(shared_ptr<_Yp> const& __r, - typename enable_if::value, __nat*>::type) - _NOEXCEPT - : __ptr_(__r.__ptr_), - __cntrl_(__r.__cntrl_) -{ - if (__cntrl_) - __cntrl_->__add_weak(); -} - -template -template -inline -weak_ptr<_Tp>::weak_ptr(weak_ptr<_Yp> const& __r, - typename enable_if::value, __nat*>::type) - _NOEXCEPT - : __ptr_(__r.__ptr_), - __cntrl_(__r.__cntrl_) -{ - if (__cntrl_) - __cntrl_->__add_weak(); -} - -template -inline -weak_ptr<_Tp>::weak_ptr(weak_ptr&& __r) _NOEXCEPT - : __ptr_(__r.__ptr_), - __cntrl_(__r.__cntrl_) -{ - __r.__ptr_ = nullptr; - __r.__cntrl_ = nullptr; -} - -template -template -inline -weak_ptr<_Tp>::weak_ptr(weak_ptr<_Yp>&& __r, - typename enable_if::value, __nat*>::type) - _NOEXCEPT - : __ptr_(__r.__ptr_), - __cntrl_(__r.__cntrl_) -{ - __r.__ptr_ = nullptr; - __r.__cntrl_ = nullptr; -} - -template -weak_ptr<_Tp>::~weak_ptr() -{ - if (__cntrl_) - __cntrl_->__release_weak(); -} - -template -inline -weak_ptr<_Tp>& -weak_ptr<_Tp>::operator=(weak_ptr const& __r) _NOEXCEPT -{ - weak_ptr(__r).swap(*this); - return *this; -} - -template -template -inline -typename enable_if -< - is_convertible<_Yp*, _Tp*>::value, - weak_ptr<_Tp>& ->::type -weak_ptr<_Tp>::operator=(weak_ptr<_Yp> const& __r) _NOEXCEPT -{ - weak_ptr(__r).swap(*this); - return *this; -} - -template -inline -weak_ptr<_Tp>& -weak_ptr<_Tp>::operator=(weak_ptr&& __r) _NOEXCEPT -{ - weak_ptr(_VSTD::move(__r)).swap(*this); - return *this; -} - -template -template -inline -typename enable_if -< - is_convertible<_Yp*, _Tp*>::value, - weak_ptr<_Tp>& ->::type -weak_ptr<_Tp>::operator=(weak_ptr<_Yp>&& __r) _NOEXCEPT -{ - weak_ptr(_VSTD::move(__r)).swap(*this); - return *this; -} - -template -template -inline -typename enable_if -< - is_convertible<_Yp*, _Tp*>::value, - weak_ptr<_Tp>& ->::type -weak_ptr<_Tp>::operator=(shared_ptr<_Yp> const& __r) _NOEXCEPT -{ - weak_ptr(__r).swap(*this); - return *this; -} - -template -inline -void -weak_ptr<_Tp>::swap(weak_ptr& __r) _NOEXCEPT -{ - _VSTD::swap(__ptr_, __r.__ptr_); - _VSTD::swap(__cntrl_, __r.__cntrl_); -} - -template -inline _LIBCPP_INLINE_VISIBILITY -void -swap(weak_ptr<_Tp>& __x, weak_ptr<_Tp>& __y) _NOEXCEPT -{ - __x.swap(__y); -} - -template -inline -void -weak_ptr<_Tp>::reset() _NOEXCEPT -{ - weak_ptr().swap(*this); -} - -template -template -shared_ptr<_Tp>::shared_ptr(const weak_ptr<_Yp>& __r, - typename enable_if::value, __nat>::type) - : __ptr_(__r.__ptr_), - __cntrl_(__r.__cntrl_ ? __r.__cntrl_->lock() : __r.__cntrl_) -{ - if (__cntrl_ == nullptr) - __throw_bad_weak_ptr(); -} - -template -shared_ptr<_Tp> -weak_ptr<_Tp>::lock() const _NOEXCEPT -{ - shared_ptr<_Tp> __r; - __r.__cntrl_ = __cntrl_ ? __cntrl_->lock() : __cntrl_; - if (__r.__cntrl_) - __r.__ptr_ = __ptr_; - return __r; -} - -#if _LIBCPP_STD_VER > 14 -template struct owner_less; -#else -template struct owner_less; -#endif - -template -struct _LIBCPP_TEMPLATE_VIS owner_less > - : binary_function, shared_ptr<_Tp>, bool> -{ - typedef bool result_type; - _LIBCPP_INLINE_VISIBILITY - bool operator()(shared_ptr<_Tp> const& __x, shared_ptr<_Tp> const& __y) const _NOEXCEPT - {return __x.owner_before(__y);} - _LIBCPP_INLINE_VISIBILITY - bool operator()(shared_ptr<_Tp> const& __x, weak_ptr<_Tp> const& __y) const _NOEXCEPT - {return __x.owner_before(__y);} - _LIBCPP_INLINE_VISIBILITY - bool operator()( weak_ptr<_Tp> const& __x, shared_ptr<_Tp> const& __y) const _NOEXCEPT - {return __x.owner_before(__y);} -}; - -template -struct _LIBCPP_TEMPLATE_VIS owner_less > - : binary_function, weak_ptr<_Tp>, bool> -{ - typedef bool result_type; - _LIBCPP_INLINE_VISIBILITY - bool operator()( weak_ptr<_Tp> const& __x, weak_ptr<_Tp> const& __y) const _NOEXCEPT - {return __x.owner_before(__y);} - _LIBCPP_INLINE_VISIBILITY - bool operator()(shared_ptr<_Tp> const& __x, weak_ptr<_Tp> const& __y) const _NOEXCEPT - {return __x.owner_before(__y);} - _LIBCPP_INLINE_VISIBILITY - bool operator()( weak_ptr<_Tp> const& __x, shared_ptr<_Tp> const& __y) const _NOEXCEPT - {return __x.owner_before(__y);} -}; - -#if _LIBCPP_STD_VER > 14 -template <> -struct _LIBCPP_TEMPLATE_VIS owner_less -{ - template - _LIBCPP_INLINE_VISIBILITY - bool operator()( shared_ptr<_Tp> const& __x, shared_ptr<_Up> const& __y) const _NOEXCEPT - {return __x.owner_before(__y);} - template - _LIBCPP_INLINE_VISIBILITY - bool operator()( shared_ptr<_Tp> const& __x, weak_ptr<_Up> const& __y) const _NOEXCEPT - {return __x.owner_before(__y);} - template - _LIBCPP_INLINE_VISIBILITY - bool operator()( weak_ptr<_Tp> const& __x, shared_ptr<_Up> const& __y) const _NOEXCEPT - {return __x.owner_before(__y);} - template - _LIBCPP_INLINE_VISIBILITY - bool operator()( weak_ptr<_Tp> const& __x, weak_ptr<_Up> const& __y) const _NOEXCEPT - {return __x.owner_before(__y);} - typedef void is_transparent; -}; -#endif - -template -class _LIBCPP_TEMPLATE_VIS enable_shared_from_this -{ - mutable weak_ptr<_Tp> __weak_this_; -protected: - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR - enable_shared_from_this() _NOEXCEPT {} - _LIBCPP_INLINE_VISIBILITY - enable_shared_from_this(enable_shared_from_this const&) _NOEXCEPT {} - _LIBCPP_INLINE_VISIBILITY - enable_shared_from_this& operator=(enable_shared_from_this const&) _NOEXCEPT - {return *this;} - _LIBCPP_INLINE_VISIBILITY - ~enable_shared_from_this() {} -public: - _LIBCPP_INLINE_VISIBILITY - shared_ptr<_Tp> shared_from_this() - {return shared_ptr<_Tp>(__weak_this_);} - _LIBCPP_INLINE_VISIBILITY - shared_ptr<_Tp const> shared_from_this() const - {return shared_ptr(__weak_this_);} - -#if _LIBCPP_STD_VER > 14 - _LIBCPP_INLINE_VISIBILITY - weak_ptr<_Tp> weak_from_this() _NOEXCEPT - { return __weak_this_; } - - _LIBCPP_INLINE_VISIBILITY - weak_ptr weak_from_this() const _NOEXCEPT - { return __weak_this_; } -#endif // _LIBCPP_STD_VER > 14 - - template friend class shared_ptr; -}; - -template -struct _LIBCPP_TEMPLATE_VIS hash > -{ - typedef shared_ptr<_Tp> argument_type; - typedef size_t result_type; - - _LIBCPP_INLINE_VISIBILITY - result_type operator()(const argument_type& __ptr) const _NOEXCEPT - { - return hash::element_type*>()(__ptr.get()); - } -}; - -template -inline _LIBCPP_INLINE_VISIBILITY -basic_ostream<_CharT, _Traits>& -operator<<(basic_ostream<_CharT, _Traits>& __os, shared_ptr<_Yp> const& __p); - - -#if !defined(_LIBCPP_HAS_NO_ATOMIC_HEADER) - -class _LIBCPP_TYPE_VIS __sp_mut -{ - void* __lx; -public: - void lock() _NOEXCEPT; - void unlock() _NOEXCEPT; - -private: - _LIBCPP_CONSTEXPR __sp_mut(void*) _NOEXCEPT; - __sp_mut(const __sp_mut&); - __sp_mut& operator=(const __sp_mut&); - - friend _LIBCPP_FUNC_VIS __sp_mut& __get_sp_mut(const void*); -}; - -_LIBCPP_FUNC_VIS _LIBCPP_AVAILABILITY_ATOMIC_SHARED_PTR -__sp_mut& __get_sp_mut(const void*); - -template -inline _LIBCPP_INLINE_VISIBILITY -bool -atomic_is_lock_free(const shared_ptr<_Tp>*) -{ - return false; -} - -template -_LIBCPP_AVAILABILITY_ATOMIC_SHARED_PTR -shared_ptr<_Tp> -atomic_load(const shared_ptr<_Tp>* __p) -{ - __sp_mut& __m = __get_sp_mut(__p); - __m.lock(); - shared_ptr<_Tp> __q = *__p; - __m.unlock(); - return __q; -} - -template -inline _LIBCPP_INLINE_VISIBILITY -_LIBCPP_AVAILABILITY_ATOMIC_SHARED_PTR -shared_ptr<_Tp> -atomic_load_explicit(const shared_ptr<_Tp>* __p, memory_order) -{ - return atomic_load(__p); -} - -template -_LIBCPP_AVAILABILITY_ATOMIC_SHARED_PTR -void -atomic_store(shared_ptr<_Tp>* __p, shared_ptr<_Tp> __r) -{ - __sp_mut& __m = __get_sp_mut(__p); - __m.lock(); - __p->swap(__r); - __m.unlock(); -} - -template -inline _LIBCPP_INLINE_VISIBILITY -_LIBCPP_AVAILABILITY_ATOMIC_SHARED_PTR -void -atomic_store_explicit(shared_ptr<_Tp>* __p, shared_ptr<_Tp> __r, memory_order) -{ - atomic_store(__p, __r); -} - -template -_LIBCPP_AVAILABILITY_ATOMIC_SHARED_PTR -shared_ptr<_Tp> -atomic_exchange(shared_ptr<_Tp>* __p, shared_ptr<_Tp> __r) -{ - __sp_mut& __m = __get_sp_mut(__p); - __m.lock(); - __p->swap(__r); - __m.unlock(); - return __r; -} - -template -inline _LIBCPP_INLINE_VISIBILITY -_LIBCPP_AVAILABILITY_ATOMIC_SHARED_PTR -shared_ptr<_Tp> -atomic_exchange_explicit(shared_ptr<_Tp>* __p, shared_ptr<_Tp> __r, memory_order) -{ - return atomic_exchange(__p, __r); -} - -template -_LIBCPP_AVAILABILITY_ATOMIC_SHARED_PTR -bool -atomic_compare_exchange_strong(shared_ptr<_Tp>* __p, shared_ptr<_Tp>* __v, shared_ptr<_Tp> __w) -{ - shared_ptr<_Tp> __temp; - __sp_mut& __m = __get_sp_mut(__p); - __m.lock(); - if (__p->__owner_equivalent(*__v)) - { - _VSTD::swap(__temp, *__p); - *__p = __w; - __m.unlock(); - return true; - } - _VSTD::swap(__temp, *__v); - *__v = *__p; - __m.unlock(); - return false; -} - -template -inline _LIBCPP_INLINE_VISIBILITY -_LIBCPP_AVAILABILITY_ATOMIC_SHARED_PTR -bool -atomic_compare_exchange_weak(shared_ptr<_Tp>* __p, shared_ptr<_Tp>* __v, shared_ptr<_Tp> __w) -{ - return atomic_compare_exchange_strong(__p, __v, __w); -} - -template -inline _LIBCPP_INLINE_VISIBILITY -_LIBCPP_AVAILABILITY_ATOMIC_SHARED_PTR -bool -atomic_compare_exchange_strong_explicit(shared_ptr<_Tp>* __p, shared_ptr<_Tp>* __v, - shared_ptr<_Tp> __w, memory_order, memory_order) -{ - return atomic_compare_exchange_strong(__p, __v, __w); -} - -template -inline _LIBCPP_INLINE_VISIBILITY -_LIBCPP_AVAILABILITY_ATOMIC_SHARED_PTR -bool -atomic_compare_exchange_weak_explicit(shared_ptr<_Tp>* __p, shared_ptr<_Tp>* __v, - shared_ptr<_Tp> __w, memory_order, memory_order) -{ - return atomic_compare_exchange_weak(__p, __v, __w); -} - -#endif // !defined(_LIBCPP_HAS_NO_ATOMIC_HEADER) - -//enum class -#if defined(_LIBCPP_ABI_POINTER_SAFETY_ENUM_TYPE) -# ifndef _LIBCPP_CXX03_LANG -enum class pointer_safety : unsigned char { - relaxed, - preferred, - strict -}; -# endif -#else -struct _LIBCPP_TYPE_VIS pointer_safety -{ - enum __lx - { - relaxed, - preferred, - strict - }; - - __lx __v_; - - _LIBCPP_INLINE_VISIBILITY - pointer_safety() : __v_() {} - - _LIBCPP_INLINE_VISIBILITY - pointer_safety(__lx __v) : __v_(__v) {} - _LIBCPP_INLINE_VISIBILITY - operator int() const {return __v_;} -}; -#endif - -#if !defined(_LIBCPP_ABI_POINTER_SAFETY_ENUM_TYPE) && \ - defined(_LIBCPP_BUILDING_LIBRARY) -_LIBCPP_FUNC_VIS pointer_safety get_pointer_safety() _NOEXCEPT; -#else -// This function is only offered in C++03 under ABI v1. -# if !defined(_LIBCPP_ABI_POINTER_SAFETY_ENUM_TYPE) || !defined(_LIBCPP_CXX03_LANG) -inline _LIBCPP_INLINE_VISIBILITY -pointer_safety get_pointer_safety() _NOEXCEPT { - return pointer_safety::relaxed; -} -# endif -#endif - - -_LIBCPP_FUNC_VIS void declare_reachable(void* __p); -_LIBCPP_FUNC_VIS void declare_no_pointers(char* __p, size_t __n); -_LIBCPP_FUNC_VIS void undeclare_no_pointers(char* __p, size_t __n); -_LIBCPP_FUNC_VIS void* __undeclare_reachable(void* __p); - -template -inline _LIBCPP_INLINE_VISIBILITY -_Tp* -undeclare_reachable(_Tp* __p) -{ - return static_cast<_Tp*>(__undeclare_reachable(__p)); -} - _LIBCPP_FUNC_VIS void* align(size_t __align, size_t __sz, void*& __ptr, size_t& __space); // --- Helper for container swap -- @@ -4193,7 +868,7 @@ void __swap_allocator(_Alloc & __a1, _Alloc & __a2) #endif { _VSTD::__swap_allocator(__a1, __a2, - integral_constant::propagate_on_container_swap::value>()); + integral_constant::propagate_on_container_swap::value>()); } template > @@ -4233,7 +908,7 @@ struct __is_allocator : false_type {}; template struct __is_allocator<_Alloc, typename __void_t::type, - typename __void_t().allocate(size_t(0)))>::type + typename __void_t().allocate(size_t(0)))>::type > : true_type {}; @@ -4291,4 +966,4 @@ _LIBCPP_POP_MACROS # include <__pstl_memory> #endif -#endif // _LIBCPP_MEMORY +#endif // _LIBCPP_MEMORY diff --git a/lib/libcxx/include/module.modulemap b/lib/libcxx/include/module.modulemap index 750cd3858..b5781b027 100644 --- a/lib/libcxx/include/module.modulemap +++ b/lib/libcxx/include/module.modulemap @@ -217,6 +217,102 @@ module std [system] { header "algorithm" export initializer_list export * + + module __algorithm { + module adjacent_find { private header "__algorithm/adjacent_find.h" } + module all_of { private header "__algorithm/all_of.h" } + module any_of { private header "__algorithm/any_of.h" } + module binary_search { private header "__algorithm/binary_search.h" } + module clamp { private header "__algorithm/clamp.h" } + module comp { private header "__algorithm/comp.h" } + module comp_ref_type { private header "__algorithm/comp_ref_type.h" } + module copy { private header "__algorithm/copy.h" } + module copy_backward { private header "__algorithm/copy_backward.h" } + module copy_if { private header "__algorithm/copy_if.h" } + module copy_n { private header "__algorithm/copy_n.h" } + module count { private header "__algorithm/count.h" } + module count_if { private header "__algorithm/count_if.h" } + module equal { private header "__algorithm/equal.h" } + module equal_range { private header "__algorithm/equal_range.h" } + module fill { private header "__algorithm/fill.h" } + module fill_n { private header "__algorithm/fill_n.h" } + module find { private header "__algorithm/find.h" } + module find_end { private header "__algorithm/find_end.h" } + module find_first_of { private header "__algorithm/find_first_of.h" } + module find_if { private header "__algorithm/find_if.h" } + module find_if_not { private header "__algorithm/find_if_not.h" } + module for_each { private header "__algorithm/for_each.h" } + module for_each_n { private header "__algorithm/for_each_n.h" } + module generate { private header "__algorithm/generate.h" } + module generate_n { private header "__algorithm/generate_n.h" } + module half_positive { private header "__algorithm/half_positive.h" } + module includes { private header "__algorithm/includes.h" } + module inplace_merge { private header "__algorithm/inplace_merge.h" } + module is_heap { private header "__algorithm/is_heap.h" } + module is_heap_until { private header "__algorithm/is_heap_until.h" } + module is_partitioned { private header "__algorithm/is_partitioned.h" } + module is_permutation { private header "__algorithm/is_permutation.h" } + module is_sorted { private header "__algorithm/is_sorted.h" } + module is_sorted_until { private header "__algorithm/is_sorted_until.h" } + module iter_swap { private header "__algorithm/iter_swap.h" } + module lexicographical_compare { private header "__algorithm/lexicographical_compare.h" } + module lower_bound { private header "__algorithm/lower_bound.h" } + module make_heap { private header "__algorithm/make_heap.h" } + module max { private header "__algorithm/max.h" } + module max_element { private header "__algorithm/max_element.h" } + module merge { private header "__algorithm/merge.h" } + module min { private header "__algorithm/min.h" } + module min_element { private header "__algorithm/min_element.h" } + module minmax { private header "__algorithm/minmax.h" } + module minmax_element { private header "__algorithm/minmax_element.h" } + module mismatch { private header "__algorithm/mismatch.h" } + module move { private header "__algorithm/move.h" } + module move_backward { private header "__algorithm/move_backward.h" } + module next_permutation { private header "__algorithm/next_permutation.h" } + module none_of { private header "__algorithm/none_of.h" } + module nth_element { private header "__algorithm/nth_element.h" } + module partial_sort { private header "__algorithm/partial_sort.h" } + module partial_sort_copy { private header "__algorithm/partial_sort_copy.h" } + module partition { private header "__algorithm/partition.h" } + module partition_copy { private header "__algorithm/partition_copy.h" } + module partition_point { private header "__algorithm/partition_point.h" } + module pop_heap { private header "__algorithm/pop_heap.h" } + module prev_permutation { private header "__algorithm/prev_permutation.h" } + module push_heap { private header "__algorithm/push_heap.h" } + module remove { private header "__algorithm/remove.h" } + module remove_copy { private header "__algorithm/remove_copy.h" } + module remove_copy_if { private header "__algorithm/remove_copy_if.h" } + module remove_if { private header "__algorithm/remove_if.h" } + module replace { private header "__algorithm/replace.h" } + module replace_copy { private header "__algorithm/replace_copy.h" } + module replace_copy_if { private header "__algorithm/replace_copy_if.h" } + module replace_if { private header "__algorithm/replace_if.h" } + module reverse { private header "__algorithm/reverse.h" } + module reverse_copy { private header "__algorithm/reverse_copy.h" } + module rotate { private header "__algorithm/rotate.h" } + module rotate_copy { private header "__algorithm/rotate_copy.h" } + module sample { private header "__algorithm/sample.h" } + module search { private header "__algorithm/search.h" } + module search_n { private header "__algorithm/search_n.h" } + module set_difference { private header "__algorithm/set_difference.h" } + module set_intersection { private header "__algorithm/set_intersection.h" } + module set_symmetric_difference { private header "__algorithm/set_symmetric_difference.h" } + module set_union { private header "__algorithm/set_union.h" } + module shift_left { private header "__algorithm/shift_left.h" } + module shift_right { private header "__algorithm/shift_right.h" } + module shuffle { private header "__algorithm/shuffle.h" } + module sift_down { private header "__algorithm/sift_down.h" } + module sort { private header "__algorithm/sort.h" } + module sort_heap { private header "__algorithm/sort_heap.h" } + module stable_partition { private header "__algorithm/stable_partition.h" } + module stable_sort { private header "__algorithm/stable_sort.h" } + module swap_ranges { private header "__algorithm/swap_ranges.h" } + module transform { private header "__algorithm/transform.h" } + module unique { private header "__algorithm/unique.h" } + module unique_copy { private header "__algorithm/unique_copy.h" } + module unwrap_iter { private header "__algorithm/unwrap_iter.h" } + module upper_bound { private header "__algorithm/upper_bound.h" } + } } module any { header "any" @@ -292,6 +388,15 @@ module std [system] { header "filesystem" export * } + module format { + header "format" + export * + + module __format { + module format_error { private header "__format/format_error.h" } + module format_parse_context { private header "__format/format_parse_context.h" } + } + } module forward_list { header "forward_list" export initializer_list @@ -304,6 +409,34 @@ module std [system] { module functional { header "functional" export * + + module __functional { + module binary_function { private header "__functional/binary_function.h" } + module binary_negate { private header "__functional/binary_negate.h" } + module bind { private header "__functional/bind.h" } + module bind_front { private header "__functional/bind_front.h" } + module binder1st { private header "__functional/binder1st.h" } + module binder2nd { private header "__functional/binder2nd.h" } + module default_searcher { private header "__functional/default_searcher.h" } + module function { private header "__functional/function.h" } + module hash { private header "__functional/hash.h" } + module identity { private header "__functional/identity.h" } + module is_transparent { private header "__functional/is_transparent.h" } + module invoke { private header "__functional/invoke.h" } + module mem_fn { private header "__functional/mem_fn.h" } + module mem_fun_ref { private header "__functional/mem_fun_ref.h" } + module not_fn { private header "__functional/not_fn.h" } + module operations { private header "__functional/operations.h" } + module perfect_forward { private header "__functional/perfect_forward.h" } + module pointer_to_binary_function { private header "__functional/pointer_to_binary_function.h" } + module pointer_to_unary_function { private header "__functional/pointer_to_unary_function.h" } + module ranges_operations { private header "__functional/ranges_operations.h" } + module reference_wrapper { private header "__functional/reference_wrapper.h" } + module unary_function { private header "__functional/unary_function.h" } + module unary_negate { private header "__functional/unary_negate.h" } + module unwrap_ref { private header "__functional/unwrap_ref.h" } + module weak_result_type { private header "__functional/weak_result_type.h" } + } } module future { header "future" @@ -342,6 +475,49 @@ module std [system] { module iterator { header "iterator" export * + + module __iterator { + module access { private header "__iterator/access.h" } + module advance { + private header "__iterator/advance.h" + export __function_like + } + module back_insert_iterator { private header "__iterator/back_insert_iterator.h" } + module common_iterator { private header "__iterator/common_iterator.h" } + module concepts { private header "__iterator/concepts.h" } + module counted_iterator { private header "__iterator/counted_iterator.h" } + module data { private header "__iterator/data.h" } + module default_sentinel { private header "__iterator/default_sentinel.h" } + module distance { private header "__iterator/distance.h" } + module empty { private header "__iterator/empty.h" } + module erase_if_container { private header "__iterator/erase_if_container.h" } + module front_insert_iterator { private header "__iterator/front_insert_iterator.h" } + module incrementable_traits { private header "__iterator/incrementable_traits.h" } + module insert_iterator { private header "__iterator/insert_iterator.h" } + module istream_iterator { private header "__iterator/istream_iterator.h" } + module istreambuf_iterator { private header "__iterator/istreambuf_iterator.h" } + module iter_move { private header "__iterator/iter_move.h" } + module iter_swap { private header "__iterator/iter_swap.h" } + module iterator { private header "__iterator/iterator.h" } + module iterator_traits { private header "__iterator/iterator_traits.h" } + module move_iterator { private header "__iterator/move_iterator.h" } + module next { + private header "__iterator/next.h" + export __function_like + } + module ostream_iterator { private header "__iterator/ostream_iterator.h" } + module ostreambuf_iterator { private header "__iterator/ostreambuf_iterator.h" } + module prev { + private header "__iterator/prev.h" + export __function_like + } + module projected { private header "__iterator/projected.h" } + module readable_traits { private header "__iterator/readable_traits.h" } + module reverse_access { private header "__iterator/reverse_access.h" } + module reverse_iterator { private header "__iterator/reverse_iterator.h" } + module size { private header "__iterator/size.h" } + module wrap_iter { private header "__iterator/wrap_iter.h" } + } } module latch { requires cplusplus14 @@ -369,6 +545,25 @@ module std [system] { module memory { header "memory" export * + + module __memory { + module addressof { private header "__memory/addressof.h" } + module allocation_guard { private header "__memory/allocation_guard.h" } + module allocator { private header "__memory/allocator.h" } + module allocator_arg_t { private header "__memory/allocator_arg_t.h" } + module allocator_traits { private header "__memory/allocator_traits.h" } + module auto_ptr { private header "__memory/auto_ptr.h" } + module compressed_pair { private header "__memory/compressed_pair.h" } + module construct_at { private header "__memory/construct_at.h" } + module pointer_safety { private header "__memory/pointer_safety.h" } + module pointer_traits { private header "__memory/pointer_traits.h" } + module raw_storage_iterator { private header "__memory/raw_storage_iterator.h" } + module shared_ptr { private header "__memory/shared_ptr.h" } + module temporary_buffer { private header "__memory/temporary_buffer.h" } + module uninitialized_algorithms { private header "__memory/uninitialized_algorithms.h" } + module unique_ptr { private header "__memory/unique_ptr.h" } + module uses_allocator { private header "__memory/uses_allocator.h" } + } } module mutex { header "mutex" @@ -404,6 +599,38 @@ module std [system] { header "random" export initializer_list export * + + module __random { + module uniform_int_distribution { private header "__random/uniform_int_distribution.h" } + } + } + module ranges { + header "ranges" + export compare + export initializer_list + export iterator + export * + + module __ranges { + module access { private header "__ranges/access.h" } + module all { private header "__ranges/all.h" } + module common_view { private header "__ranges/common_view.h" } + module concepts { private header "__ranges/concepts.h" } + module copyable_box { private header "__ranges/copyable_box.h" } + module dangling { private header "__ranges/dangling.h" } + module data { private header "__ranges/data.h" } + module drop_view { private header "__ranges/drop_view.h" } + module empty { private header "__ranges/empty.h" } + module empty_view { private header "__ranges/empty_view.h" } + module enable_borrowed_range { private header "__ranges/enable_borrowed_range.h" } + module enable_view { private header "__ranges/enable_view.h" } + module non_propagating_cache { private header "__ranges/non_propagating_cache.h" } + module ref_view { private header "__ranges/ref_view.h" } + module size { private header "__ranges/size.h" } + module subrange { private header "__ranges/subrange.h" } + module transform_view { private header "__ranges/transform_view.h" } + module view_interface { private header "__ranges/view_interface.h" } + } } module ratio { header "ratio" @@ -428,6 +655,15 @@ module std [system] { export initializer_list export * } + module shared_mutex { + header "shared_mutex" + export version + } + module span { + header "span" + export ranges.__ranges.enable_borrowed_range + export version + } module sstream { header "sstream" // FIXME: should re-export istream, ostream, ios, streambuf, string? @@ -477,6 +713,7 @@ module std [system] { } module type_traits { header "type_traits" + export functional.__functional.unwrap_ref export * } module typeindex { @@ -501,6 +738,23 @@ module std [system] { header "utility" export initializer_list export * + + module __utility { + module __decay_copy { private header "__utility/__decay_copy.h" } + module as_const { private header "__utility/as_const.h" } + module cmp { private header "__utility/cmp.h" } + module declval { private header "__utility/declval.h" } + module exchange { private header "__utility/exchange.h" } + module forward { private header "__utility/forward.h" } + module in_place { private header "__utility/in_place.h" } + module integer_sequence { private header "__utility/integer_sequence.h" } + module move { private header "__utility/move.h" } + module pair { private header "__utility/pair.h" } + module piecewise_construct { private header "__utility/piecewise_construct.h" } + module rel_ops { private header "__utility/rel_ops.h" } + module swap { private header "__utility/swap.h" } + module to_underlying { private header "__utility/to_underlying.h" } + } } module valarray { header "valarray" @@ -510,6 +764,10 @@ module std [system] { module variant { header "variant" export * + + module __variant { + module monostate { private header "__variant/monostate.h" } + } } module vector { header "vector" @@ -521,23 +779,26 @@ module std [system] { export * } + // __config not modularised due to a bug in Clang // FIXME: These should be private. - module __bits { header "__bits" export * } - module __bit_reference { header "__bit_reference" export * } - module __debug { header "__debug" export * } - module __errc { header "__errc" export * } - module __functional_base { header "__functional_base" export * } - module __hash_table { header "__hash_table" export * } - module __locale { header "__locale" export * } - module __mutex_base { header "__mutex_base" export * } - module __split_buffer { header "__split_buffer" export * } - module __sso_allocator { header "__sso_allocator" export * } - module __std_stream { header "__std_stream" export * } - module __string { header "__string" export * } - module __tree { header "__tree" export * } - module __tuple { header "__tuple" export * } - module __undef_macros { header "__undef_macros" export * } - module __node_handle { header "__node_handle" export * } + module __availability { private header "__availability" export * } + module __bit_reference { private header "__bit_reference" export * } + module __bits { private header "__bits" export * } + module __debug { header "__debug" export * } + module __errc { private header "__errc" export * } + module __function_like { private header "__function_like.h" export * } + module __hash_table { header "__hash_table" export * } + module __locale { private header "__locale" export * } + module __mutex_base { private header "__mutex_base" export * } + module __node_handle { private header "__node_handle" export * } + module __nullptr { header "__nullptr" export * } + module __split_buffer { private header "__split_buffer" export * } + module __std_stream { private header "__std_stream" export * } + module __string { private header "__string" export * } + module __threading_support { header "__threading_support" export * } + module __tree { header "__tree" export * } + module __tuple { private header "__tuple" export * } + module __undef_macros { header "__undef_macros" export * } module experimental { requires cplusplus11 diff --git a/lib/libcxx/include/mutex b/lib/libcxx/include/mutex index f098ccba4..eb8e54ad4 100644 --- a/lib/libcxx/include/mutex +++ b/lib/libcxx/include/mutex @@ -188,14 +188,15 @@ template #include <__config> #include <__mutex_base> +#include <__threading_support> +#include <__utility/forward.h> #include #include #include #ifndef _LIBCPP_CXX03_LANG -#include +# include #endif #include -#include <__threading_support> #if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) #pragma GCC system_header @@ -365,7 +366,7 @@ try_lock(_L0& __l0, _L1& __l1, _L2& __l2, _L3&... __l3) return __r; } -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG template void @@ -469,7 +470,7 @@ void __unlock(_L0& __l0, _L1& __l1, _L2& __l2, _L3&... __l3) { _VSTD::__unlock(__l2, __l3...); } -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG #if _LIBCPP_STD_VER > 14 template @@ -568,7 +569,7 @@ template _LIBCPP_INLINE_VISIBILITY void call_once(once_flag&, const _Callable&); -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG struct _LIBCPP_TEMPLATE_VIS once_flag { @@ -601,7 +602,7 @@ private: template friend void call_once(once_flag&, const _Callable&); -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG }; #ifndef _LIBCPP_CXX03_LANG @@ -702,10 +703,10 @@ call_once(once_flag& __flag, const _Callable& __func) } } -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG _LIBCPP_END_NAMESPACE_STD _LIBCPP_POP_MACROS -#endif // _LIBCPP_MUTEX +#endif // _LIBCPP_MUTEX diff --git a/lib/libcxx/include/new b/lib/libcxx/include/new index 0562cef45..aefc08c16 100644 --- a/lib/libcxx/include/new +++ b/lib/libcxx/include/new @@ -86,8 +86,8 @@ void operator delete[](void* ptr, void*) noexcept; */ -#include <__config> #include <__availability> +#include <__config> #include #include #include @@ -314,7 +314,7 @@ void* __libcpp_aligned_alloc(std::size_t __alignment, std::size_t __size) { return ::_aligned_malloc(__size, __alignment); #else void* __result = nullptr; - ::posix_memalign(&__result, __alignment, __size); + (void)::posix_memalign(&__result, __alignment, __size); // If posix_memalign fails, __result is unmodified so we still return `nullptr`. return __result; #endif @@ -356,4 +356,4 @@ constexpr _Tp* launder(_Tp* __p) noexcept _LIBCPP_END_NAMESPACE_STD -#endif // _LIBCPP_NEW +#endif // _LIBCPP_NEW diff --git a/lib/libcxx/include/numbers b/lib/libcxx/include/numbers index 38dad9955..52c67dd35 100644 --- a/lib/libcxx/include/numbers +++ b/lib/libcxx/include/numbers @@ -59,12 +59,12 @@ namespace std::numbers { */ #include <__config> - -#if _LIBCPP_STD_VER > 17 && defined(__cpp_concepts) && __cpp_concepts >= 201811L - +#include #include #include +#if _LIBCPP_STD_VER > 17 && !defined(_LIBCPP_HAS_NO_CONCEPTS) + #if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) #pragma GCC system_header #endif @@ -99,22 +99,19 @@ template inline constexpr T inv_sqrt3_v = __illformed{}; template inline constexpr T egamma_v = __illformed{}; template inline constexpr T phi_v = __illformed{}; -template -concept __floating_point = is_floating_point_v; - -template <__floating_point T> inline constexpr T e_v = 2.718281828459045235360287471352662; -template <__floating_point T> inline constexpr T log2e_v = 1.442695040888963407359924681001892; -template <__floating_point T> inline constexpr T log10e_v = 0.434294481903251827651128918916605; -template <__floating_point T> inline constexpr T pi_v = 3.141592653589793238462643383279502; -template <__floating_point T> inline constexpr T inv_pi_v = 0.318309886183790671537767526745028; -template <__floating_point T> inline constexpr T inv_sqrtpi_v = 0.564189583547756286948079451560772; -template <__floating_point T> inline constexpr T ln2_v = 0.693147180559945309417232121458176; -template <__floating_point T> inline constexpr T ln10_v = 2.302585092994045684017991454684364; -template <__floating_point T> inline constexpr T sqrt2_v = 1.414213562373095048801688724209698; -template <__floating_point T> inline constexpr T sqrt3_v = 1.732050807568877293527446341505872; -template <__floating_point T> inline constexpr T inv_sqrt3_v = 0.577350269189625764509148780501957; -template <__floating_point T> inline constexpr T egamma_v = 0.577215664901532860606512090082402; -template <__floating_point T> inline constexpr T phi_v = 1.618033988749894848204586834365638; +template inline constexpr T e_v = 2.718281828459045235360287471352662; +template inline constexpr T log2e_v = 1.442695040888963407359924681001892; +template inline constexpr T log10e_v = 0.434294481903251827651128918916605; +template inline constexpr T pi_v = 3.141592653589793238462643383279502; +template inline constexpr T inv_pi_v = 0.318309886183790671537767526745028; +template inline constexpr T inv_sqrtpi_v = 0.564189583547756286948079451560772; +template inline constexpr T ln2_v = 0.693147180559945309417232121458176; +template inline constexpr T ln10_v = 2.302585092994045684017991454684364; +template inline constexpr T sqrt2_v = 1.414213562373095048801688724209698; +template inline constexpr T sqrt3_v = 1.732050807568877293527446341505872; +template inline constexpr T inv_sqrt3_v = 0.577350269189625764509148780501957; +template inline constexpr T egamma_v = 0.577215664901532860606512090082402; +template inline constexpr T phi_v = 1.618033988749894848204586834365638; inline constexpr double e = e_v; inline constexpr double log2e = log2e_v; @@ -136,6 +133,6 @@ _LIBCPP_END_NAMESPACE_STD _LIBCPP_POP_MACROS -#endif //_LIBCPP_STD_VER > 17 && defined(__cpp_concepts) && __cpp_concepts >= 201811L +#endif //_LIBCPP_STD_VER > 17 && !defined(_LIBCPP_HAS_NO_CONCEPTS) #endif // _LIBCPP_NUMBERS diff --git a/lib/libcxx/include/numeric b/lib/libcxx/include/numeric index 4f202bb84..d42cbf97e 100644 --- a/lib/libcxx/include/numeric +++ b/lib/libcxx/include/numeric @@ -145,10 +145,11 @@ template */ #include <__config> +#include <__debug> +#include // for isnormal +#include #include #include // for numeric_limits -#include -#include // for isnormal #include #if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) @@ -613,8 +614,8 @@ _LIBCPP_INLINE_VISIBILITY constexpr enable_if_t, _Fp> midpoint(_Fp __a, _Fp __b) noexcept { - constexpr _Fp __lo = numeric_limits<_Fp>::min()*2; - constexpr _Fp __hi = numeric_limits<_Fp>::max()/2; + constexpr _Fp __lo = numeric_limits<_Fp>::min()*2; + constexpr _Fp __hi = numeric_limits<_Fp>::max()/2; return __fp_abs(__a) <= __hi && __fp_abs(__b) <= __hi ? // typical case: overflow is impossible (__a + __b)/2 : // always correctly rounded __fp_abs(__a) < __lo ? __a + __b/2 : // not safe to halve a @@ -632,4 +633,4 @@ _LIBCPP_POP_MACROS # include <__pstl_numeric> #endif -#endif // _LIBCPP_NUMERIC +#endif // _LIBCPP_NUMERIC diff --git a/lib/libcxx/include/optional b/lib/libcxx/include/optional index 97a0bbe66..118db66a4 100644 --- a/lib/libcxx/include/optional +++ b/lib/libcxx/include/optional @@ -69,7 +69,7 @@ namespace std { template constexpr bool operator>=(const T&, const optional&); // 23.6.9, specialized algorithms - template void swap(optional&, optional&) noexcept(see below ); + template void swap(optional&, optional&) noexcept(see below ); // constexpr in C++20 template constexpr optional make_optional(T&&); template constexpr optional make_optional(Args&&... args); @@ -95,26 +95,26 @@ namespace std { template constexpr EXPLICIT optional(U &&); template - constexpr EXPLICIT optional(const optional &); + EXPLICIT optional(const optional &); // constexpr in C++20 template - constexpr EXPLICIT optional(optional &&); + EXPLICIT optional(optional &&); // constexpr in C++20 // 23.6.3.2, destructor - ~optional(); + ~optional(); // constexpr in C++20 // 23.6.3.3, assignment - optional &operator=(nullopt_t) noexcept; - optional &operator=(const optional &); // constexpr in C++20 - optional &operator=(optional &&) noexcept(see below); // constexpr in C++20 - template optional &operator=(U &&); - template optional &operator=(const optional &); - template optional &operator=(optional &&); - template T& emplace(Args &&...); + optional &operator=(nullopt_t) noexcept; // constexpr in C++20 + optional &operator=(const optional &); // constexpr in C++20 + optional &operator=(optional &&) noexcept(see below); // constexpr in C++20 + template optional &operator=(U &&); // constexpr in C++20 + template optional &operator=(const optional &); // constexpr in C++20 + template optional &operator=(optional &&); // constexpr in C++20 + template T& emplace(Args &&...); // constexpr in C++20 template - T& emplace(initializer_list, Args &&...); + T& emplace(initializer_list, Args &&...); // constexpr in C++20 // 23.6.3.4, swap - void swap(optional &) noexcept(see below ); + void swap(optional &) noexcept(see below ); // constexpr in C++20 // 23.6.3.5, observers constexpr T const *operator->() const; @@ -133,7 +133,7 @@ namespace std { template constexpr T value_or(U &&) &&; // 23.6.3.6, modifiers - void reset() noexcept; + void reset() noexcept; // constexpr in C++20 private: T *val; // exposition only @@ -146,10 +146,11 @@ template */ -#include <__config> #include <__availability> +#include <__config> #include <__debug> #include <__functional_base> +#include #include #include #include @@ -220,7 +221,7 @@ struct __optional_destruct_base<_Tp, false> bool __engaged_; _LIBCPP_INLINE_VISIBILITY - ~__optional_destruct_base() + _LIBCPP_CONSTEXPR_AFTER_CXX17 ~__optional_destruct_base() { if (__engaged_) __val_.~value_type(); @@ -238,7 +239,7 @@ struct __optional_destruct_base<_Tp, false> __engaged_(true) {} _LIBCPP_INLINE_VISIBILITY - void reset() noexcept + _LIBCPP_CONSTEXPR_AFTER_CXX17 void reset() noexcept { if (__engaged_) { @@ -273,7 +274,7 @@ struct __optional_destruct_base<_Tp, true> __engaged_(true) {} _LIBCPP_INLINE_VISIBILITY - void reset() noexcept + _LIBCPP_CONSTEXPR_AFTER_CXX17 void reset() noexcept { if (__engaged_) { @@ -318,16 +319,20 @@ struct __optional_storage_base : __optional_destruct_base<_Tp> template _LIBCPP_INLINE_VISIBILITY - void __construct(_Args&&... __args) + _LIBCPP_CONSTEXPR_AFTER_CXX17 void __construct(_Args&&... __args) { _LIBCPP_ASSERT(!has_value(), "__construct called for engaged __optional_storage"); +#if _LIBCPP_STD_VER > 17 + _VSTD::construct_at(_VSTD::addressof(this->__val_), _VSTD::forward<_Args>(__args)...); +#else ::new ((void*)_VSTD::addressof(this->__val_)) value_type(_VSTD::forward<_Args>(__args)...); +#endif this->__engaged_ = true; } template _LIBCPP_INLINE_VISIBILITY - void __construct_from(_That&& __opt) + _LIBCPP_CONSTEXPR_AFTER_CXX17 void __construct_from(_That&& __opt) { if (__opt.has_value()) __construct(_VSTD::forward<_That>(__opt).__get()); @@ -335,7 +340,7 @@ struct __optional_storage_base : __optional_destruct_base<_Tp> template _LIBCPP_INLINE_VISIBILITY - void __assign_from(_That&& __opt) + _LIBCPP_CONSTEXPR_AFTER_CXX17 void __assign_from(_That&& __opt) { if (this->__engaged_ == __opt.has_value()) { @@ -393,7 +398,7 @@ struct __optional_storage_base<_Tp, true> } _LIBCPP_INLINE_VISIBILITY - void reset() noexcept { __value_ = nullptr; } + _LIBCPP_CONSTEXPR_AFTER_CXX17 void reset() noexcept { __value_ = nullptr; } _LIBCPP_INLINE_VISIBILITY constexpr bool has_value() const noexcept @@ -409,7 +414,7 @@ struct __optional_storage_base<_Tp, true> template _LIBCPP_INLINE_VISIBILITY - void __construct(_UArg&& __val) + _LIBCPP_CONSTEXPR_AFTER_CXX17 void __construct(_UArg&& __val) { _LIBCPP_ASSERT(!has_value(), "__construct called for engaged __optional_storage"); static_assert(__can_bind_reference<_UArg>(), @@ -420,7 +425,7 @@ struct __optional_storage_base<_Tp, true> template _LIBCPP_INLINE_VISIBILITY - void __construct_from(_That&& __opt) + _LIBCPP_CONSTEXPR_AFTER_CXX17 void __construct_from(_That&& __opt) { if (__opt.has_value()) __construct(_VSTD::forward<_That>(__opt).__get()); @@ -428,7 +433,7 @@ struct __optional_storage_base<_Tp, true> template _LIBCPP_INLINE_VISIBILITY - void __assign_from(_That&& __opt) + _LIBCPP_CONSTEXPR_AFTER_CXX17 void __assign_from(_That&& __opt) { if (has_value() == __opt.has_value()) { @@ -460,7 +465,7 @@ struct __optional_copy_base<_Tp, false> : __optional_storage_base<_Tp> __optional_copy_base() = default; _LIBCPP_INLINE_VISIBILITY - __optional_copy_base(const __optional_copy_base& __opt) + _LIBCPP_CONSTEXPR_AFTER_CXX17 __optional_copy_base(const __optional_copy_base& __opt) { this->__construct_from(__opt); } @@ -491,7 +496,7 @@ struct __optional_move_base<_Tp, false> : __optional_copy_base<_Tp> __optional_move_base(const __optional_move_base&) = default; _LIBCPP_INLINE_VISIBILITY - __optional_move_base(__optional_move_base&& __opt) + _LIBCPP_CONSTEXPR_AFTER_CXX17 __optional_move_base(__optional_move_base&& __opt) noexcept(is_nothrow_move_constructible_v) { this->__construct_from(_VSTD::move(__opt)); @@ -525,7 +530,7 @@ struct __optional_copy_assign_base<_Tp, false> : __optional_move_base<_Tp> __optional_copy_assign_base(__optional_copy_assign_base&&) = default; _LIBCPP_INLINE_VISIBILITY - __optional_copy_assign_base& operator=(const __optional_copy_assign_base& __opt) + _LIBCPP_CONSTEXPR_AFTER_CXX17 __optional_copy_assign_base& operator=(const __optional_copy_assign_base& __opt) { this->__assign_from(__opt); return *this; @@ -560,7 +565,7 @@ struct __optional_move_assign_base<_Tp, false> : __optional_copy_assign_base<_Tp __optional_move_assign_base& operator=(const __optional_move_assign_base&) = default; _LIBCPP_INLINE_VISIBILITY - __optional_move_assign_base& operator=(__optional_move_assign_base&& __opt) + _LIBCPP_CONSTEXPR_AFTER_CXX17 __optional_move_assign_base& operator=(__optional_move_assign_base&& __opt) noexcept(is_nothrow_move_assignable_v && is_nothrow_move_constructible_v) { @@ -727,7 +732,7 @@ public: _CheckOptionalLikeCtor<_Up, _Up const&>::template __enable_implicit<_Up>() , int> = 0> _LIBCPP_INLINE_VISIBILITY - optional(const optional<_Up>& __v) + _LIBCPP_CONSTEXPR_AFTER_CXX17 optional(const optional<_Up>& __v) { this->__construct_from(__v); } @@ -735,7 +740,7 @@ public: _CheckOptionalLikeCtor<_Up, _Up const&>::template __enable_explicit<_Up>() , int> = 0> _LIBCPP_INLINE_VISIBILITY - explicit optional(const optional<_Up>& __v) + _LIBCPP_CONSTEXPR_AFTER_CXX17 explicit optional(const optional<_Up>& __v) { this->__construct_from(__v); } @@ -745,7 +750,7 @@ public: _CheckOptionalLikeCtor<_Up, _Up &&>::template __enable_implicit<_Up>() , int> = 0> _LIBCPP_INLINE_VISIBILITY - optional(optional<_Up>&& __v) + _LIBCPP_CONSTEXPR_AFTER_CXX17 optional(optional<_Up>&& __v) { this->__construct_from(_VSTD::move(__v)); } @@ -753,13 +758,13 @@ public: _CheckOptionalLikeCtor<_Up, _Up &&>::template __enable_explicit<_Up>() , int> = 0> _LIBCPP_INLINE_VISIBILITY - explicit optional(optional<_Up>&& __v) + _LIBCPP_CONSTEXPR_AFTER_CXX17 explicit optional(optional<_Up>&& __v) { this->__construct_from(_VSTD::move(__v)); } _LIBCPP_INLINE_VISIBILITY - optional& operator=(nullopt_t) noexcept + _LIBCPP_CONSTEXPR_AFTER_CXX17 optional& operator=(nullopt_t) noexcept { reset(); return *this; @@ -782,7 +787,7 @@ public: >::value> > _LIBCPP_INLINE_VISIBILITY - optional& + _LIBCPP_CONSTEXPR_AFTER_CXX17 optional& operator=(_Up&& __v) { if (this->has_value()) @@ -797,7 +802,7 @@ public: _CheckOptionalLikeAssign<_Up, _Up const&>::template __enable_assign<_Up>() , int> = 0> _LIBCPP_INLINE_VISIBILITY - optional& + _LIBCPP_CONSTEXPR_AFTER_CXX17 optional& operator=(const optional<_Up>& __v) { this->__assign_from(__v); @@ -809,7 +814,7 @@ public: _CheckOptionalLikeCtor<_Up, _Up &&>::template __enable_assign<_Up>() , int> = 0> _LIBCPP_INLINE_VISIBILITY - optional& + _LIBCPP_CONSTEXPR_AFTER_CXX17 optional& operator=(optional<_Up>&& __v) { this->__assign_from(_VSTD::move(__v)); @@ -823,7 +828,7 @@ public: > > _LIBCPP_INLINE_VISIBILITY - _Tp & + _LIBCPP_CONSTEXPR_AFTER_CXX17 _Tp & emplace(_Args&&... __args) { reset(); @@ -838,7 +843,7 @@ public: > > _LIBCPP_INLINE_VISIBILITY - _Tp & + _LIBCPP_CONSTEXPR_AFTER_CXX17 _Tp & emplace(initializer_list<_Up> __il, _Args&&... __args) { reset(); @@ -847,7 +852,7 @@ public: } _LIBCPP_INLINE_VISIBILITY - void swap(optional& __opt) + _LIBCPP_CONSTEXPR_AFTER_CXX17 void swap(optional& __opt) noexcept(is_nothrow_move_constructible_v && is_nothrow_swappable_v) { @@ -877,7 +882,7 @@ public: add_pointer_t operator->() const { - _LIBCPP_ASSERT(this->has_value(), "optional operator-> called for disengaged value"); + _LIBCPP_ASSERT(this->has_value(), "optional operator-> called on a disengaged value"); #ifndef _LIBCPP_HAS_NO_BUILTIN_ADDRESSOF return _VSTD::addressof(this->__get()); #else @@ -890,7 +895,7 @@ public: add_pointer_t operator->() { - _LIBCPP_ASSERT(this->has_value(), "optional operator-> called for disengaged value"); + _LIBCPP_ASSERT(this->has_value(), "optional operator-> called on a disengaged value"); #ifndef _LIBCPP_HAS_NO_BUILTIN_ADDRESSOF return _VSTD::addressof(this->__get()); #else @@ -901,36 +906,36 @@ public: _LIBCPP_INLINE_VISIBILITY constexpr const value_type& - operator*() const& + operator*() const& noexcept { - _LIBCPP_ASSERT(this->has_value(), "optional operator* called for disengaged value"); + _LIBCPP_ASSERT(this->has_value(), "optional operator* called on a disengaged value"); return this->__get(); } _LIBCPP_INLINE_VISIBILITY constexpr value_type& - operator*() & + operator*() & noexcept { - _LIBCPP_ASSERT(this->has_value(), "optional operator* called for disengaged value"); + _LIBCPP_ASSERT(this->has_value(), "optional operator* called on a disengaged value"); return this->__get(); } _LIBCPP_INLINE_VISIBILITY constexpr value_type&& - operator*() && + operator*() && noexcept { - _LIBCPP_ASSERT(this->has_value(), "optional operator* called for disengaged value"); + _LIBCPP_ASSERT(this->has_value(), "optional operator* called on a disengaged value"); return _VSTD::move(this->__get()); } _LIBCPP_INLINE_VISIBILITY constexpr const value_type&& - operator*() const&& + operator*() const&& noexcept { - _LIBCPP_ASSERT(this->has_value(), "optional operator* called for disengaged value"); + _LIBCPP_ASSERT(this->has_value(), "optional operator* called on a disengaged value"); return _VSTD::move(this->__get()); } @@ -1005,7 +1010,7 @@ public: private: template _LIBCPP_INLINE_VISIBILITY - static _Up* + static _LIBCPP_CONSTEXPR_AFTER_CXX17 _Up* __operator_arrow(true_type, _Up& __x) { return _VSTD::addressof(__x); @@ -1029,8 +1034,8 @@ template template _LIBCPP_INLINE_VISIBILITY constexpr _EnableIf< - is_convertible_v() == - _VSTD::declval()), bool>, + is_convertible_v() == + declval()), bool>, bool > operator==(const optional<_Tp>& __x, const optional<_Up>& __y) @@ -1045,8 +1050,8 @@ operator==(const optional<_Tp>& __x, const optional<_Up>& __y) template _LIBCPP_INLINE_VISIBILITY constexpr _EnableIf< - is_convertible_v() != - _VSTD::declval()), bool>, + is_convertible_v() != + declval()), bool>, bool > operator!=(const optional<_Tp>& __x, const optional<_Up>& __y) @@ -1061,8 +1066,8 @@ operator!=(const optional<_Tp>& __x, const optional<_Up>& __y) template _LIBCPP_INLINE_VISIBILITY constexpr _EnableIf< - is_convertible_v() < - _VSTD::declval()), bool>, + is_convertible_v() < + declval()), bool>, bool > operator<(const optional<_Tp>& __x, const optional<_Up>& __y) @@ -1077,8 +1082,8 @@ operator<(const optional<_Tp>& __x, const optional<_Up>& __y) template _LIBCPP_INLINE_VISIBILITY constexpr _EnableIf< - is_convertible_v() > - _VSTD::declval()), bool>, + is_convertible_v() > + declval()), bool>, bool > operator>(const optional<_Tp>& __x, const optional<_Up>& __y) @@ -1093,8 +1098,8 @@ operator>(const optional<_Tp>& __x, const optional<_Up>& __y) template _LIBCPP_INLINE_VISIBILITY constexpr _EnableIf< - is_convertible_v() <= - _VSTD::declval()), bool>, + is_convertible_v() <= + declval()), bool>, bool > operator<=(const optional<_Tp>& __x, const optional<_Up>& __y) @@ -1109,8 +1114,8 @@ operator<=(const optional<_Tp>& __x, const optional<_Up>& __y) template _LIBCPP_INLINE_VISIBILITY constexpr _EnableIf< - is_convertible_v() >= - _VSTD::declval()), bool>, + is_convertible_v() >= + declval()), bool>, bool > operator>=(const optional<_Tp>& __x, const optional<_Up>& __y) @@ -1223,8 +1228,8 @@ operator>=(nullopt_t, const optional<_Tp>& __x) noexcept template _LIBCPP_INLINE_VISIBILITY constexpr _EnableIf< - is_convertible_v() == - _VSTD::declval()), bool>, + is_convertible_v() == + declval()), bool>, bool > operator==(const optional<_Tp>& __x, const _Up& __v) @@ -1235,8 +1240,8 @@ operator==(const optional<_Tp>& __x, const _Up& __v) template _LIBCPP_INLINE_VISIBILITY constexpr _EnableIf< - is_convertible_v() == - _VSTD::declval()), bool>, + is_convertible_v() == + declval()), bool>, bool > operator==(const _Tp& __v, const optional<_Up>& __x) @@ -1247,8 +1252,8 @@ operator==(const _Tp& __v, const optional<_Up>& __x) template _LIBCPP_INLINE_VISIBILITY constexpr _EnableIf< - is_convertible_v() != - _VSTD::declval()), bool>, + is_convertible_v() != + declval()), bool>, bool > operator!=(const optional<_Tp>& __x, const _Up& __v) @@ -1259,8 +1264,8 @@ operator!=(const optional<_Tp>& __x, const _Up& __v) template _LIBCPP_INLINE_VISIBILITY constexpr _EnableIf< - is_convertible_v() != - _VSTD::declval()), bool>, + is_convertible_v() != + declval()), bool>, bool > operator!=(const _Tp& __v, const optional<_Up>& __x) @@ -1271,8 +1276,8 @@ operator!=(const _Tp& __v, const optional<_Up>& __x) template _LIBCPP_INLINE_VISIBILITY constexpr _EnableIf< - is_convertible_v() < - _VSTD::declval()), bool>, + is_convertible_v() < + declval()), bool>, bool > operator<(const optional<_Tp>& __x, const _Up& __v) @@ -1283,8 +1288,8 @@ operator<(const optional<_Tp>& __x, const _Up& __v) template _LIBCPP_INLINE_VISIBILITY constexpr _EnableIf< - is_convertible_v() < - _VSTD::declval()), bool>, + is_convertible_v() < + declval()), bool>, bool > operator<(const _Tp& __v, const optional<_Up>& __x) @@ -1295,8 +1300,8 @@ operator<(const _Tp& __v, const optional<_Up>& __x) template _LIBCPP_INLINE_VISIBILITY constexpr _EnableIf< - is_convertible_v() <= - _VSTD::declval()), bool>, + is_convertible_v() <= + declval()), bool>, bool > operator<=(const optional<_Tp>& __x, const _Up& __v) @@ -1307,8 +1312,8 @@ operator<=(const optional<_Tp>& __x, const _Up& __v) template _LIBCPP_INLINE_VISIBILITY constexpr _EnableIf< - is_convertible_v() <= - _VSTD::declval()), bool>, + is_convertible_v() <= + declval()), bool>, bool > operator<=(const _Tp& __v, const optional<_Up>& __x) @@ -1319,8 +1324,8 @@ operator<=(const _Tp& __v, const optional<_Up>& __x) template _LIBCPP_INLINE_VISIBILITY constexpr _EnableIf< - is_convertible_v() > - _VSTD::declval()), bool>, + is_convertible_v() > + declval()), bool>, bool > operator>(const optional<_Tp>& __x, const _Up& __v) @@ -1331,8 +1336,8 @@ operator>(const optional<_Tp>& __x, const _Up& __v) template _LIBCPP_INLINE_VISIBILITY constexpr _EnableIf< - is_convertible_v() > - _VSTD::declval()), bool>, + is_convertible_v() > + declval()), bool>, bool > operator>(const _Tp& __v, const optional<_Up>& __x) @@ -1343,8 +1348,8 @@ operator>(const _Tp& __v, const optional<_Up>& __x) template _LIBCPP_INLINE_VISIBILITY constexpr _EnableIf< - is_convertible_v() >= - _VSTD::declval()), bool>, + is_convertible_v() >= + declval()), bool>, bool > operator>=(const optional<_Tp>& __x, const _Up& __v) @@ -1355,8 +1360,8 @@ operator>=(const optional<_Tp>& __x, const _Up& __v) template _LIBCPP_INLINE_VISIBILITY constexpr _EnableIf< - is_convertible_v() >= - _VSTD::declval()), bool>, + is_convertible_v() >= + declval()), bool>, bool > operator>=(const _Tp& __v, const optional<_Up>& __x) @@ -1366,7 +1371,7 @@ operator>=(const _Tp& __v, const optional<_Up>& __x) template -inline _LIBCPP_INLINE_VISIBILITY +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 _EnableIf< is_move_constructible_v<_Tp> && is_swappable_v<_Tp>, void @@ -1402,11 +1407,13 @@ struct _LIBCPP_TEMPLATE_VIS hash< __enable_hash_helper, remove_const_t<_Tp>> > { - typedef optional<_Tp> argument_type; - typedef size_t result_type; +#if _LIBCPP_STD_VER <= 17 || defined(_LIBCPP_ENABLE_CXX20_REMOVED_BINDER_TYPEDEFS) + _LIBCPP_DEPRECATED_IN_CXX17 typedef optional<_Tp> argument_type; + _LIBCPP_DEPRECATED_IN_CXX17 typedef size_t result_type; +#endif _LIBCPP_INLINE_VISIBILITY - result_type operator()(const argument_type& __opt) const + size_t operator()(const optional<_Tp>& __opt) const { return static_cast(__opt) ? hash>()(*__opt) : 0; } @@ -1414,8 +1421,8 @@ struct _LIBCPP_TEMPLATE_VIS hash< _LIBCPP_END_NAMESPACE_STD -#endif // _LIBCPP_STD_VER > 14 +#endif // _LIBCPP_STD_VER > 14 _LIBCPP_POP_MACROS -#endif // _LIBCPP_OPTIONAL +#endif // _LIBCPP_OPTIONAL diff --git a/lib/libcxx/include/ostream b/lib/libcxx/include/ostream index f5eb8a894..efeaee253 100644 --- a/lib/libcxx/include/ostream +++ b/lib/libcxx/include/ostream @@ -134,11 +134,11 @@ template */ #include <__config> +#include #include -#include -#include #include -#include +#include +#include #include #if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) @@ -165,27 +165,21 @@ public: { this->init(__sb); } virtual ~basic_ostream(); protected: -#ifndef _LIBCPP_CXX03_LANG inline _LIBCPP_INLINE_VISIBILITY basic_ostream(basic_ostream&& __rhs); // 27.7.2.3 Assign/swap inline _LIBCPP_INLINE_VISIBILITY basic_ostream& operator=(basic_ostream&& __rhs); -#endif + inline _LIBCPP_HIDE_FROM_ABI_AFTER_V1 void swap(basic_ostream& __rhs) { basic_ios::swap(__rhs); } -#ifndef _LIBCPP_CXX03_LANG basic_ostream (const basic_ostream& __rhs) = delete; basic_ostream& operator=(const basic_ostream& __rhs) = delete; -#else - basic_ostream (const basic_ostream& __rhs); // not defined - basic_ostream& operator=(const basic_ostream& __rhs); // not defined -#endif -public: +public: // 27.7.2.4 Prefix/suffix: class _LIBCPP_TEMPLATE_VIS sentry; @@ -254,8 +248,7 @@ public: ~sentry(); _LIBCPP_INLINE_VISIBILITY - _LIBCPP_EXPLICIT - operator bool() const {return __ok_;} + explicit operator bool() const {return __ok_;} }; template @@ -280,7 +273,7 @@ basic_ostream<_CharT, _Traits>::sentry::~sentry() #ifndef _LIBCPP_NO_EXCEPTIONS try { -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS if (__os_.rdbuf()->pubsync() == -1) __os_.setstate(ios_base::badbit); #ifndef _LIBCPP_NO_EXCEPTIONS @@ -288,12 +281,10 @@ basic_ostream<_CharT, _Traits>::sentry::~sentry() catch (...) { } -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS } } -#ifndef _LIBCPP_CXX03_LANG - template basic_ostream<_CharT, _Traits>::basic_ostream(basic_ostream&& __rhs) { @@ -308,8 +299,6 @@ basic_ostream<_CharT, _Traits>::operator=(basic_ostream&& __rhs) return *this; } -#endif // _LIBCPP_CXX03_LANG - template basic_ostream<_CharT, _Traits>::~basic_ostream() { @@ -322,7 +311,7 @@ basic_ostream<_CharT, _Traits>::operator<<(basic_streambuf::operator<<(basic_streambuf _Ip; typedef ostreambuf_iterator<_CharT, _Traits> _Op; _Ip __i(__sb); @@ -352,7 +341,7 @@ basic_ostream<_CharT, _Traits>::operator<<(basic_streambuf__set_failbit_and_consider_rethrow(); } -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS } else this->setstate(ios_base::badbit); @@ -363,7 +352,7 @@ basic_ostream<_CharT, _Traits>::operator<<(basic_streambuf__set_badbit_and_consider_rethrow(); } -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS return *this; } @@ -374,7 +363,7 @@ basic_ostream<_CharT, _Traits>::operator<<(bool __n) #ifndef _LIBCPP_NO_EXCEPTIONS try { -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS sentry __s(*this); if (__s) { @@ -389,7 +378,7 @@ basic_ostream<_CharT, _Traits>::operator<<(bool __n) { this->__set_badbit_and_consider_rethrow(); } -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS return *this; } @@ -400,7 +389,7 @@ basic_ostream<_CharT, _Traits>::operator<<(short __n) #ifndef _LIBCPP_NO_EXCEPTIONS try { -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS sentry __s(*this); if (__s) { @@ -419,7 +408,7 @@ basic_ostream<_CharT, _Traits>::operator<<(short __n) { this->__set_badbit_and_consider_rethrow(); } -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS return *this; } @@ -430,7 +419,7 @@ basic_ostream<_CharT, _Traits>::operator<<(unsigned short __n) #ifndef _LIBCPP_NO_EXCEPTIONS try { -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS sentry __s(*this); if (__s) { @@ -445,7 +434,7 @@ basic_ostream<_CharT, _Traits>::operator<<(unsigned short __n) { this->__set_badbit_and_consider_rethrow(); } -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS return *this; } @@ -456,7 +445,7 @@ basic_ostream<_CharT, _Traits>::operator<<(int __n) #ifndef _LIBCPP_NO_EXCEPTIONS try { -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS sentry __s(*this); if (__s) { @@ -475,7 +464,7 @@ basic_ostream<_CharT, _Traits>::operator<<(int __n) { this->__set_badbit_and_consider_rethrow(); } -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS return *this; } @@ -486,7 +475,7 @@ basic_ostream<_CharT, _Traits>::operator<<(unsigned int __n) #ifndef _LIBCPP_NO_EXCEPTIONS try { -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS sentry __s(*this); if (__s) { @@ -501,7 +490,7 @@ basic_ostream<_CharT, _Traits>::operator<<(unsigned int __n) { this->__set_badbit_and_consider_rethrow(); } -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS return *this; } @@ -512,7 +501,7 @@ basic_ostream<_CharT, _Traits>::operator<<(long __n) #ifndef _LIBCPP_NO_EXCEPTIONS try { -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS sentry __s(*this); if (__s) { @@ -527,7 +516,7 @@ basic_ostream<_CharT, _Traits>::operator<<(long __n) { this->__set_badbit_and_consider_rethrow(); } -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS return *this; } @@ -538,7 +527,7 @@ basic_ostream<_CharT, _Traits>::operator<<(unsigned long __n) #ifndef _LIBCPP_NO_EXCEPTIONS try { -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS sentry __s(*this); if (__s) { @@ -553,7 +542,7 @@ basic_ostream<_CharT, _Traits>::operator<<(unsigned long __n) { this->__set_badbit_and_consider_rethrow(); } -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS return *this; } @@ -564,7 +553,7 @@ basic_ostream<_CharT, _Traits>::operator<<(long long __n) #ifndef _LIBCPP_NO_EXCEPTIONS try { -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS sentry __s(*this); if (__s) { @@ -579,7 +568,7 @@ basic_ostream<_CharT, _Traits>::operator<<(long long __n) { this->__set_badbit_and_consider_rethrow(); } -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS return *this; } @@ -590,7 +579,7 @@ basic_ostream<_CharT, _Traits>::operator<<(unsigned long long __n) #ifndef _LIBCPP_NO_EXCEPTIONS try { -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS sentry __s(*this); if (__s) { @@ -605,7 +594,7 @@ basic_ostream<_CharT, _Traits>::operator<<(unsigned long long __n) { this->__set_badbit_and_consider_rethrow(); } -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS return *this; } @@ -616,7 +605,7 @@ basic_ostream<_CharT, _Traits>::operator<<(float __n) #ifndef _LIBCPP_NO_EXCEPTIONS try { -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS sentry __s(*this); if (__s) { @@ -631,7 +620,7 @@ basic_ostream<_CharT, _Traits>::operator<<(float __n) { this->__set_badbit_and_consider_rethrow(); } -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS return *this; } @@ -642,7 +631,7 @@ basic_ostream<_CharT, _Traits>::operator<<(double __n) #ifndef _LIBCPP_NO_EXCEPTIONS try { -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS sentry __s(*this); if (__s) { @@ -657,7 +646,7 @@ basic_ostream<_CharT, _Traits>::operator<<(double __n) { this->__set_badbit_and_consider_rethrow(); } -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS return *this; } @@ -668,7 +657,7 @@ basic_ostream<_CharT, _Traits>::operator<<(long double __n) #ifndef _LIBCPP_NO_EXCEPTIONS try { -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS sentry __s(*this); if (__s) { @@ -683,7 +672,7 @@ basic_ostream<_CharT, _Traits>::operator<<(long double __n) { this->__set_badbit_and_consider_rethrow(); } -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS return *this; } @@ -694,7 +683,7 @@ basic_ostream<_CharT, _Traits>::operator<<(const void* __n) #ifndef _LIBCPP_NO_EXCEPTIONS try { -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS sentry __s(*this); if (__s) { @@ -709,7 +698,7 @@ basic_ostream<_CharT, _Traits>::operator<<(const void* __n) { this->__set_badbit_and_consider_rethrow(); } -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS return *this; } @@ -721,7 +710,7 @@ __put_character_sequence(basic_ostream<_CharT, _Traits>& __os, #ifndef _LIBCPP_NO_EXCEPTIONS try { -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS typename basic_ostream<_CharT, _Traits>::sentry __s(__os); if (__s) { @@ -742,7 +731,7 @@ __put_character_sequence(basic_ostream<_CharT, _Traits>& __os, { __os.__set_badbit_and_consider_rethrow(); } -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS return __os; } @@ -761,7 +750,7 @@ operator<<(basic_ostream<_CharT, _Traits>& __os, char __cn) #ifndef _LIBCPP_NO_EXCEPTIONS try { -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS typename basic_ostream<_CharT, _Traits>::sentry __s(__os); if (__s) { @@ -783,7 +772,7 @@ operator<<(basic_ostream<_CharT, _Traits>& __os, char __cn) { __os.__set_badbit_and_consider_rethrow(); } -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS return __os; } @@ -822,7 +811,7 @@ operator<<(basic_ostream<_CharT, _Traits>& __os, const char* __strn) #ifndef _LIBCPP_NO_EXCEPTIONS try { -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS typename basic_ostream<_CharT, _Traits>::sentry __s(__os); if (__s) { @@ -857,7 +846,7 @@ operator<<(basic_ostream<_CharT, _Traits>& __os, const char* __strn) { __os.__set_badbit_and_consider_rethrow(); } -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS return __os; } @@ -891,7 +880,7 @@ basic_ostream<_CharT, _Traits>::put(char_type __c) #ifndef _LIBCPP_NO_EXCEPTIONS try { -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS sentry __s(*this); if (__s) { @@ -907,7 +896,7 @@ basic_ostream<_CharT, _Traits>::put(char_type __c) { this->__set_badbit_and_consider_rethrow(); } -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS return *this; } @@ -918,7 +907,7 @@ basic_ostream<_CharT, _Traits>::write(const char_type* __s, streamsize __n) #ifndef _LIBCPP_NO_EXCEPTIONS try { -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS sentry __sen(*this); if (__sen && __n) { @@ -931,7 +920,7 @@ basic_ostream<_CharT, _Traits>::write(const char_type* __s, streamsize __n) { this->__set_badbit_and_consider_rethrow(); } -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS return *this; } @@ -942,7 +931,7 @@ basic_ostream<_CharT, _Traits>::flush() #ifndef _LIBCPP_NO_EXCEPTIONS try { -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS if (this->rdbuf()) { sentry __s(*this); @@ -958,7 +947,7 @@ basic_ostream<_CharT, _Traits>::flush() { this->__set_badbit_and_consider_rethrow(); } -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS return *this; } @@ -1025,19 +1014,17 @@ flush(basic_ostream<_CharT, _Traits>& __os) return __os; } -#ifndef _LIBCPP_CXX03_LANG - template struct __is_ostreamable : false_type { }; template struct __is_ostreamable<_Stream, _Tp, decltype( - _VSTD::declval<_Stream>() << _VSTD::declval<_Tp>(), void() + declval<_Stream>() << declval<_Tp>(), void() )> : true_type { }; template , - __is_ostreamable<_Stream&, const _Tp&>>::value + __is_ostreamable<_Stream&, const _Tp&> >::value >::type> _LIBCPP_INLINE_VISIBILITY _Stream&& operator<<(_Stream&& __os, const _Tp& __x) @@ -1046,8 +1033,6 @@ _Stream&& operator<<(_Stream&& __os, const _Tp& __x) return _VSTD::move(__os); } -#endif // _LIBCPP_CXX03_LANG - template basic_ostream<_CharT, _Traits>& operator<<(basic_ostream<_CharT, _Traits>& __os, @@ -1106,4 +1091,4 @@ _LIBCPP_EXTERN_TEMPLATE(class _LIBCPP_EXTERN_TEMPLATE_TYPE_VIS basic_ostream */ #include <__config> +#include <__memory/uses_allocator.h> +#include <__utility/forward.h> +#include +#include #include -#include #include -#include +#include #if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) #pragma GCC system_header @@ -238,47 +241,42 @@ public: queue& operator=(queue&& __q) _NOEXCEPT_(is_nothrow_move_assignable::value) {c = _VSTD::move(__q.c); return *this;} -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG _LIBCPP_INLINE_VISIBILITY explicit queue(const container_type& __c) : c(__c) {} #ifndef _LIBCPP_CXX03_LANG _LIBCPP_INLINE_VISIBILITY explicit queue(container_type&& __c) : c(_VSTD::move(__c)) {} -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG template _LIBCPP_INLINE_VISIBILITY explicit queue(const _Alloc& __a, - typename enable_if::value>::type* = 0) + _EnableIf::value>* = 0) : c(__a) {} template _LIBCPP_INLINE_VISIBILITY queue(const queue& __q, const _Alloc& __a, - typename enable_if::value>::type* = 0) + _EnableIf::value>* = 0) : c(__q.c, __a) {} template _LIBCPP_INLINE_VISIBILITY queue(const container_type& __c, const _Alloc& __a, - typename enable_if::value>::type* = 0) + _EnableIf::value>* = 0) : c(__c, __a) {} #ifndef _LIBCPP_CXX03_LANG template _LIBCPP_INLINE_VISIBILITY queue(container_type&& __c, const _Alloc& __a, - typename enable_if::value>::type* = 0) + _EnableIf::value>* = 0) : c(_VSTD::move(__c), __a) {} template _LIBCPP_INLINE_VISIBILITY queue(queue&& __q, const _Alloc& __a, - typename enable_if::value>::type* = 0) + _EnableIf::value>* = 0) : c(_VSTD::move(__q.c), __a) {} -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG _LIBCPP_NODISCARD_AFTER_CXX17 _LIBCPP_INLINE_VISIBILITY bool empty() const {return c.empty();} @@ -308,7 +306,7 @@ public: void emplace(_Args&&... __args) { c.emplace_back(_VSTD::forward<_Args>(__args)...);} #endif -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG _LIBCPP_INLINE_VISIBILITY void pop() {c.pop_front();} @@ -335,15 +333,15 @@ public: #ifndef _LIBCPP_HAS_NO_DEDUCTION_GUIDES template::value, nullptr_t>::type + class = _EnableIf::value> > queue(_Container) -> queue; template::value, nullptr_t>::type, - class = typename enable_if< __is_allocator<_Alloc>::value, nullptr_t>::type + class = _EnableIf::value>, + class = _EnableIf::value> > queue(_Container, _Alloc) -> queue; @@ -399,10 +397,7 @@ operator<=(const queue<_Tp, _Container>& __x,const queue<_Tp, _Container>& __y) template inline _LIBCPP_INLINE_VISIBILITY -typename enable_if< - __is_swappable<_Container>::value, - void ->::type +_EnableIf<__is_swappable<_Container>::value, void> swap(queue<_Tp, _Container>& __x, queue<_Tp, _Container>& __y) _NOEXCEPT_(_NOEXCEPT_(__x.swap(__y))) { @@ -458,7 +453,7 @@ public: _NOEXCEPT_(is_nothrow_move_assignable::value && is_nothrow_move_assignable::value) {c = _VSTD::move(__q.c); comp = _VSTD::move(__q.comp); return *this;} -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG _LIBCPP_INLINE_VISIBILITY explicit priority_queue(const value_compare& __comp) @@ -482,41 +477,35 @@ public: _LIBCPP_INLINE_VISIBILITY priority_queue(_InputIter __f, _InputIter __l, const value_compare& __comp, container_type&& __c); -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG template _LIBCPP_INLINE_VISIBILITY explicit priority_queue(const _Alloc& __a, - typename enable_if::value>::type* = 0); + _EnableIf::value>* = 0); template _LIBCPP_INLINE_VISIBILITY priority_queue(const value_compare& __comp, const _Alloc& __a, - typename enable_if::value>::type* = 0); + _EnableIf::value>* = 0); template _LIBCPP_INLINE_VISIBILITY priority_queue(const value_compare& __comp, const container_type& __c, const _Alloc& __a, - typename enable_if::value>::type* = 0); + _EnableIf::value>* = 0); template _LIBCPP_INLINE_VISIBILITY priority_queue(const priority_queue& __q, const _Alloc& __a, - typename enable_if::value>::type* = 0); + _EnableIf::value>* = 0); #ifndef _LIBCPP_CXX03_LANG template _LIBCPP_INLINE_VISIBILITY priority_queue(const value_compare& __comp, container_type&& __c, const _Alloc& __a, - typename enable_if::value>::type* = 0); + _EnableIf::value>* = 0); template _LIBCPP_INLINE_VISIBILITY priority_queue(priority_queue&& __q, const _Alloc& __a, - typename enable_if::value>::type* = 0); -#endif // _LIBCPP_CXX03_LANG + _EnableIf::value>* = 0); +#endif // _LIBCPP_CXX03_LANG _LIBCPP_NODISCARD_AFTER_CXX17 _LIBCPP_INLINE_VISIBILITY bool empty() const {return c.empty();} @@ -533,7 +522,7 @@ public: template _LIBCPP_INLINE_VISIBILITY void emplace(_Args&&... __args); -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG _LIBCPP_INLINE_VISIBILITY void pop(); @@ -546,28 +535,28 @@ public: #ifndef _LIBCPP_HAS_NO_DEDUCTION_GUIDES template ::value, nullptr_t>::type, - class = typename enable_if::value, nullptr_t>::type + class = _EnableIf::value>, + class = _EnableIf::value> > priority_queue(_Compare, _Container) -> priority_queue; template::value_type>, - class _Container = vector::value_type>, - class = typename enable_if< __is_cpp17_input_iterator<_InputIterator>::value, nullptr_t>::type, - class = typename enable_if::value, nullptr_t>::type, - class = typename enable_if::value, nullptr_t>::type + class _Compare = less<__iter_value_type<_InputIterator>>, + class _Container = vector<__iter_value_type<_InputIterator>>, + class = _EnableIf<__is_cpp17_input_iterator<_InputIterator>::value>, + class = _EnableIf::value>, + class = _EnableIf::value> > priority_queue(_InputIterator, _InputIterator, _Compare = _Compare(), _Container = _Container()) - -> priority_queue::value_type, _Container, _Compare>; + -> priority_queue<__iter_value_type<_InputIterator>, _Container, _Compare>; template::value, nullptr_t>::type, - class = typename enable_if::value, nullptr_t>::type, - class = typename enable_if< __is_allocator<_Alloc>::value, nullptr_t>::type + class = _EnableIf::value>, + class = _EnableIf::value>, + class = _EnableIf::value> > priority_queue(_Compare, _Container, _Alloc) -> priority_queue; @@ -595,7 +584,7 @@ priority_queue<_Tp, _Container, _Compare>::priority_queue(const value_compare& _ _VSTD::make_heap(c.begin(), c.end(), comp); } -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG template template @@ -636,14 +625,13 @@ priority_queue<_Tp, _Container, _Compare>::priority_queue(_InputIter __f, _Input _VSTD::make_heap(c.begin(), c.end(), comp); } -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG template template inline priority_queue<_Tp, _Container, _Compare>::priority_queue(const _Alloc& __a, - typename enable_if::value>::type*) + _EnableIf::value>*) : c(__a) { } @@ -653,8 +641,7 @@ template inline priority_queue<_Tp, _Container, _Compare>::priority_queue(const value_compare& __comp, const _Alloc& __a, - typename enable_if::value>::type*) + _EnableIf::value>*) : c(__a), comp(__comp) { @@ -666,8 +653,7 @@ inline priority_queue<_Tp, _Container, _Compare>::priority_queue(const value_compare& __comp, const container_type& __c, const _Alloc& __a, - typename enable_if::value>::type*) + _EnableIf::value>*) : c(__c, __a), comp(__comp) { @@ -679,8 +665,7 @@ template inline priority_queue<_Tp, _Container, _Compare>::priority_queue(const priority_queue& __q, const _Alloc& __a, - typename enable_if::value>::type*) + _EnableIf::value>*) : c(__q.c, __a), comp(__q.comp) { @@ -695,8 +680,7 @@ inline priority_queue<_Tp, _Container, _Compare>::priority_queue(const value_compare& __comp, container_type&& __c, const _Alloc& __a, - typename enable_if::value>::type*) + _EnableIf::value>*) : c(_VSTD::move(__c), __a), comp(__comp) { @@ -708,15 +692,14 @@ template inline priority_queue<_Tp, _Container, _Compare>::priority_queue(priority_queue&& __q, const _Alloc& __a, - typename enable_if::value>::type*) + _EnableIf::value>*) : c(_VSTD::move(__q.c), __a), comp(_VSTD::move(__q.comp)) { _VSTD::make_heap(c.begin(), c.end(), comp); } -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG template inline @@ -748,7 +731,7 @@ priority_queue<_Tp, _Container, _Compare>::emplace(_Args&&... __args) _VSTD::push_heap(c.begin(), c.end(), comp); } -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG template inline @@ -773,11 +756,10 @@ priority_queue<_Tp, _Container, _Compare>::swap(priority_queue& __q) template inline _LIBCPP_INLINE_VISIBILITY -typename enable_if< - __is_swappable<_Container>::value - && __is_swappable<_Compare>::value, +_EnableIf< + __is_swappable<_Container>::value && __is_swappable<_Compare>::value, void ->::type +> swap(priority_queue<_Tp, _Container, _Compare>& __x, priority_queue<_Tp, _Container, _Compare>& __y) _NOEXCEPT_(_NOEXCEPT_(__x.swap(__y))) @@ -793,4 +775,4 @@ struct _LIBCPP_TEMPLATE_VIS uses_allocator +concept uniform_random_bit_generator = see below; // C++20 // Engines @@ -1675,17 +1678,19 @@ class piecewise_linear_distribution */ #include <__config> +#include <__random/uniform_int_distribution.h> +#include +#include +#include #include #include -#include -#include #include +#include #include -#include #include -#include #include -#include +#include +#include #if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) #pragma GCC system_header @@ -1697,6 +1702,20 @@ _LIBCPP_PUSH_MACROS _LIBCPP_BEGIN_NAMESPACE_STD +#if _LIBCPP_STD_VER > 17 && !defined(_LIBCPP_HAS_NO_CONCEPTS) + +// [rand.req.urng] +template +concept uniform_random_bit_generator = + invocable<_Gen&> && unsigned_integral> && + requires { + { _Gen::min() } -> same_as>; + { _Gen::max() } -> same_as>; + requires bool_constant<(_Gen::min() < _Gen::max())>::value; + }; + +#endif // _LIBCPP_STD_VER > 17 && !defined(_LIBCPP_HAS_NO_CONCEPTS) + // __is_seed_sequence template @@ -1712,7 +1731,7 @@ struct __is_seed_sequence template (_Mp-__c)/__a), - bool _OverflowOK = ((__m|__m-1) > __m), // m = 2^n + bool _OverflowOK = ((__m | (__m-1)) > __m), // m = 2^n bool _SchrageOK = (__a != 0 && __m != 0 && __m % __a <= __m / __a)> // r <= q struct __lce_alg_picker { @@ -1901,7 +1920,7 @@ private: static_assert(__m == 0 || __a < __m, "linear_congruential_engine invalid parameters"); static_assert(__m == 0 || __c < __m, "linear_congruential_engine invalid parameters"); - static_assert(_VSTD::is_unsigned<_UIntType>::value, "_UIntType must be unsigned type"); + static_assert(is_unsigned<_UIntType>::value, "_UIntType must be unsigned type"); public: static _LIBCPP_CONSTEXPR const result_type _Min = __c == 0u ? 1u: 0u; static _LIBCPP_CONSTEXPR const result_type _Max = __m - 1u; @@ -2940,7 +2959,7 @@ public: _LIBCPP_INLINE_VISIBILITY explicit discard_block_engine(_Engine&& __e) : __e_(_VSTD::move(__e)), __n_(0) {} -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG _LIBCPP_INLINE_VISIBILITY explicit discard_block_engine(result_type __sd) : __e_(__sd), __n_(0) {} template @@ -3152,7 +3171,7 @@ public: _LIBCPP_INLINE_VISIBILITY explicit independent_bits_engine(_Engine&& __e) : __e_(_VSTD::move(__e)) {} -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG _LIBCPP_INLINE_VISIBILITY explicit independent_bits_engine(result_type __sd) : __e_(__sd) {} template @@ -3382,7 +3401,7 @@ public: _LIBCPP_INLINE_VISIBILITY explicit shuffle_order_engine(_Engine&& __e) : __e_(_VSTD::move(__e)) {__init();} -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG _LIBCPP_INLINE_VISIBILITY explicit shuffle_order_engine(result_type __sd) : __e_(__sd) {__init();} template @@ -3634,7 +3653,7 @@ public: template _LIBCPP_INLINE_VISIBILITY seed_seq(initializer_list<_Tp> __il) {init(__il.begin(), __il.end());} -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG template _LIBCPP_INLINE_VISIBILITY @@ -3755,42 +3774,6 @@ generate_canonical(_URNG& __g) return _Sp / __base; } -// uniform_int_distribution - -// in - -template -basic_ostream<_CharT, _Traits>& -operator<<(basic_ostream<_CharT, _Traits>& __os, - const uniform_int_distribution<_IT>& __x) -{ - __save_flags<_CharT, _Traits> __lx(__os); - typedef basic_ostream<_CharT, _Traits> _Ostream; - __os.flags(_Ostream::dec | _Ostream::left); - _CharT __sp = __os.widen(' '); - __os.fill(__sp); - return __os << __x.a() << __sp << __x.b(); -} - -template -basic_istream<_CharT, _Traits>& -operator>>(basic_istream<_CharT, _Traits>& __is, - uniform_int_distribution<_IT>& __x) -{ - typedef uniform_int_distribution<_IT> _Eng; - typedef typename _Eng::result_type result_type; - typedef typename _Eng::param_type param_type; - __save_flags<_CharT, _Traits> __lx(__is); - typedef basic_istream<_CharT, _Traits> _Istream; - __is.flags(_Istream::dec | _Istream::skipws); - result_type __a; - result_type __b; - __is >> __a >> __b; - if (!__is.fail()) - __x.param(param_type(__a, __b)); - return __is; -} - // uniform_real_distribution template @@ -4142,7 +4125,7 @@ inline _LIBCPP_INLINE_VISIBILITY double __libcpp_lgamma(double __d) { } template -binomial_distribution<_IntType>::param_type::param_type(const result_type __t, const double __p) +binomial_distribution<_IntType>::param_type::param_type(result_type __t, double __p) : __t_(__t), __p_(__p) { if (0 < __p_ && __p_ < 1) @@ -6145,7 +6128,7 @@ public: _LIBCPP_INLINE_VISIBILITY param_type(initializer_list __wl) : __p_(__wl.begin(), __wl.end()) {__init();} -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG template param_type(size_t __nw, double __xmin, double __xmax, _UnaryOperation __fw); @@ -6192,7 +6175,7 @@ public: _LIBCPP_INLINE_VISIBILITY discrete_distribution(initializer_list __wl) : __p_(__wl) {} -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG template _LIBCPP_INLINE_VISIBILITY discrete_distribution(size_t __nw, double __xmin, double __xmax, @@ -6274,8 +6257,7 @@ discrete_distribution<_IntType>::param_type::__init() if (__p_.size() > 1) { double __s = _VSTD::accumulate(__p_.begin(), __p_.end(), 0.0); - for (_VSTD::vector::iterator __i = __p_.begin(), __e = __p_.end(); - __i < __e; ++__i) + for (vector::iterator __i = __p_.begin(), __e = __p_.end(); __i < __e; ++__i) *__i /= __s; vector __t(__p_.size() - 1); _VSTD::partial_sum(__p_.begin(), __p_.end() - 1, __t.begin()); @@ -6294,7 +6276,7 @@ vector discrete_distribution<_IntType>::param_type::probabilities() const { size_t __n = __p_.size(); - _VSTD::vector __p(__n+1); + vector __p(__n+1); _VSTD::adjacent_difference(__p_.begin(), __p_.end(), __p.begin()); if (__n > 0) __p[__n] = 1 - __p_[__n-1]; @@ -6374,7 +6356,7 @@ public: #ifndef _LIBCPP_CXX03_LANG template param_type(initializer_list __bl, _UnaryOperation __fw); -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG template param_type(size_t __nw, result_type __xmin, result_type __xmax, _UnaryOperation __fw); @@ -6431,7 +6413,7 @@ public: piecewise_constant_distribution(initializer_list __bl, _UnaryOperation __fw) : __p_(__bl, __fw) {} -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG template _LIBCPP_INLINE_VISIBILITY @@ -6586,7 +6568,7 @@ piecewise_constant_distribution<_RealType>::param_type::param_type( } } -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG template template @@ -6700,7 +6682,7 @@ public: #ifndef _LIBCPP_CXX03_LANG template param_type(initializer_list __bl, _UnaryOperation __fw); -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG template param_type(size_t __nw, result_type __xmin, result_type __xmax, _UnaryOperation __fw); @@ -6757,7 +6739,7 @@ public: piecewise_linear_distribution(initializer_list __bl, _UnaryOperation __fw) : __p_(__bl, __fw) {} -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG template _LIBCPP_INLINE_VISIBILITY @@ -6916,7 +6898,7 @@ piecewise_linear_distribution<_RealType>::param_type::param_type( } } -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG template template @@ -7022,4 +7004,4 @@ _LIBCPP_END_NAMESPACE_STD _LIBCPP_POP_MACROS -#endif // _LIBCPP_RANDOM +#endif // _LIBCPP_RANDOM diff --git a/lib/libcxx/include/ranges b/lib/libcxx/include/ranges new file mode 100644 index 000000000..5feaf4c32 --- /dev/null +++ b/lib/libcxx/include/ranges @@ -0,0 +1,209 @@ +// -*- C++ -*- +//===--------------------------- ranges -----------------------------------===// +// +// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. +// See https://llvm.org/LICENSE.txt for license information. +// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception +// +//===----------------------------------------------------------------------===// + +#ifndef _LIBCPP_RANGES +#define _LIBCPP_RANGES + +/* + +#include // see [compare.syn] +#include // see [initializer.list.syn] +#include // see [iterator.synopsis] + +namespace std::ranges { + inline namespace unspecified { + // [range.access], range access + inline constexpr unspecified begin = unspecified; + inline constexpr unspecified end = unspecified; + inline constexpr unspecified cbegin = unspecified; + inline constexpr unspecified cend = unspecified; + + inline constexpr unspecified size = unspecified; + inline constexpr unspecified ssize = unspecified; + } + + // [range.range], ranges + template + concept range = see below; + + template + inline constexpr bool enable_borrowed_range = false; + + template + using iterator_t = decltype(ranges::begin(declval())); + template + using sentinel_t = decltype(ranges::end(declval())); + template + using range_difference_t = iter_difference_t>; + template + using range_size_t = decltype(ranges::size(declval())); + template + using range_value_t = iter_value_t>; + template + using range_reference_t = iter_reference_t>; + template + using range_rvalue_reference_t = iter_rvalue_reference_t>; + + // [range.sized], sized ranges + template + inline constexpr bool disable_sized_range = false; + + template + concept sized_range = ...; + + // [range.view], views + template + inline constexpr bool enable_view = ...; + + struct view_base { }; + + template + concept view = ...; + + // [range.refinements], other range refinements + template + concept output_range = see below; + + template + concept input_range = see below; + + template + concept forward_range = see below; + + template + concept bidirectional_range = see below; + + template + concept random_access_range = see below; + + template + concept contiguous_range = see below; + + template + concept common_range = see below; + + template + concept viewable_range = see below; + + // [view.interface], class template view_interface + template + requires is_class_v && same_as> + class view_interface; + + // [range.subrange], sub-ranges + enum class subrange_kind : bool { unsized, sized }; + + template S = I, subrange_kind K = see below> + requires (K == subrange_kind::sized || !sized_sentinel_for) + class subrange; + + template + inline constexpr bool enable_borrowed_range> = true; + + // [range.dangling], dangling iterator handling + struct dangling; + + template + using borrowed_iterator_t = see below; + + template + using borrowed_subrange_t = see below; + + // [range.empty], empty view + template + requires is_object_v + class empty_view; + + // [range.all], all view + namespace views { + inline constexpr unspecified all = unspecified; + + template + using all_t = decltype(all(declval())); + } + + template + requires is_object_v + class ref_view; + + template + inline constexpr bool enable_borrowed_range> = true; + + // [range.drop], drop view + template + class drop_view; + + template + inline constexpr bool enable_borrowed_range> = enable_borrowed_range; + + // [range.transform], transform view + template + requires view && is_object_v && + regular_invocable> && + can-reference>> + class transform_view; + + // [range.common], common view + template + requires (!common_range && copyable>) + class common_view; + + template + inline constexpr bool enable_borrowed_range> = enable_borrowed_range; +} + +*/ + +// Make sure all feature tests macros are always available. +#include +// Only enable the contents of the header when libc++ was build with LIBCXX_ENABLE_INCOMPLETE_FEATURES enabled +#if !defined(_LIBCPP_HAS_NO_INCOMPLETE_RANGES) + +#include <__config> +#include <__ranges/access.h> +#include <__ranges/all.h> +#include <__ranges/common_view.h> +#include <__ranges/concepts.h> +#include <__ranges/dangling.h> +#include <__ranges/data.h> +#include <__ranges/drop_view.h> +#include <__ranges/empty.h> +#include <__ranges/empty_view.h> +#include <__ranges/enable_borrowed_range.h> +#include <__ranges/enable_view.h> +#include <__ranges/ref_view.h> +#include <__ranges/size.h> +#include <__ranges/subrange.h> +#include <__ranges/transform_view.h> +#include <__ranges/view_interface.h> +#include // Required by the standard. +#include // Required by the standard. +#include // Required by the standard. +#include + +#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) +#pragma GCC system_header +#endif + +_LIBCPP_PUSH_MACROS +#include <__undef_macros> + +_LIBCPP_BEGIN_NAMESPACE_STD + +#if _LIBCPP_STD_VER > 17 && !defined(_LIBCPP_HAS_NO_RANGES) + +#endif // _LIBCPP_STD_VER > 17 && !defined(_LIBCPP_HAS_NO_RANGES) + +_LIBCPP_END_NAMESPACE_STD + +_LIBCPP_POP_MACROS + +#endif // !defined(_LIBCPP_HAS_NO_INCOMPLETE_RANGES) + +#endif // _LIBCPP_RANGES diff --git a/lib/libcxx/include/ratio b/lib/libcxx/include/ratio index fa7a4bbb2..091ea53ac 100644 --- a/lib/libcxx/include/ratio +++ b/lib/libcxx/include/ratio @@ -78,8 +78,8 @@ typedef ratio<1000000000000000000000000, 1> yotta; // not supported */ #include <__config> -#include #include +#include #include #if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) @@ -312,7 +312,7 @@ template struct _LIBCPP_TEMPLATE_VIS ratio_multiply : public __ratio_multiply<_R1, _R2>::type {}; -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG template struct __ratio_divide @@ -339,7 +339,7 @@ template struct _LIBCPP_TEMPLATE_VIS ratio_divide : public __ratio_divide<_R1, _R2>::type {}; -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG template struct __ratio_add @@ -374,7 +374,7 @@ template struct _LIBCPP_TEMPLATE_VIS ratio_add : public __ratio_add<_R1, _R2>::type {}; -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG template struct __ratio_subtract @@ -409,7 +409,7 @@ template struct _LIBCPP_TEMPLATE_VIS ratio_subtract : public __ratio_subtract<_R1, _R2>::type {}; -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG // ratio_equal @@ -529,4 +529,4 @@ _LIBCPP_END_NAMESPACE_STD _LIBCPP_POP_MACROS -#endif // _LIBCPP_RATIO +#endif // _LIBCPP_RATIO diff --git a/lib/libcxx/include/regex b/lib/libcxx/include/regex index d78e4888a..e47cf9fca 100644 --- a/lib/libcxx/include/regex +++ b/lib/libcxx/include/regex @@ -763,15 +763,18 @@ typedef regex_token_iterator wsregex_token_iterator; */ #include <__config> -#include +#include <__debug> +#include <__iterator/wrap_iter.h> #include <__locale> +#include +#include #include -#include #include -#include #include +#include +#include +#include #include -#include #include #if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) @@ -2000,14 +2003,14 @@ class __l_anchor_multiline { typedef __owns_one_state<_CharT> base; - bool __multiline; + bool __multiline_; public: typedef _VSTD::__state<_CharT> __state; _LIBCPP_INLINE_VISIBILITY __l_anchor_multiline(bool __multiline, __node<_CharT>* __s) - : base(__s), __multiline(__multiline) {} + : base(__s), __multiline_(__multiline) {} virtual void __exec(__state&) const; }; @@ -2022,7 +2025,7 @@ __l_anchor_multiline<_CharT>::__exec(__state& __s) const __s.__do_ = __state::__accept_but_not_consume; __s.__node_ = this->first(); } - else if (__multiline && + else if (__multiline_ && !__s.__at_first_ && __is_eol(*_VSTD::prev(__s.__current_))) { @@ -2634,7 +2637,7 @@ public: { __init(__il.begin(), __il.end()); } -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG // ~basic_regex() = default; @@ -2647,7 +2650,7 @@ public: _LIBCPP_INLINE_VISIBILITY basic_regex& operator=(initializer_list __il) {return assign(__il);} -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG template _LIBCPP_INLINE_VISIBILITY basic_regex& operator=(const basic_string& __p) @@ -2721,7 +2724,7 @@ public: flag_type __f = regex_constants::ECMAScript) {return assign(__il.begin(), __il.end(), __f);} -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG // const operations: _LIBCPP_INLINE_VISIBILITY @@ -4571,7 +4574,7 @@ basic_regex<_CharT, _Traits>::__parse_character_escape(_ForwardIterator __first, if (__hd == -1) __throw_regex_error(); __sum = 16 * __sum + static_cast(__hd); - // drop through + // fallthrough case 'x': ++__first; if (__first == __last) @@ -5882,7 +5885,6 @@ basic_regex<_CharT, _Traits>::__match_at_start_posix_subs( { vector<__state> __states; __state __best_state; - ptrdiff_t __j = 0; ptrdiff_t __highest_j = 0; ptrdiff_t _Np = _VSTD::distance(__first, __last); __node* __st = __start_.get(); @@ -5903,7 +5905,6 @@ basic_regex<_CharT, _Traits>::__match_at_start_posix_subs( __states.back().__node_ = __st; __states.back().__flags_ = __flags; __states.back().__at_first_ = __at_first; - const _CharT* __current = __first; bool __matched = false; int __counter = 0; int __length = __last - __first; @@ -5943,9 +5944,6 @@ basic_regex<_CharT, _Traits>::__match_at_start_posix_subs( __states.pop_back(); break; case __state::__accept_and_consume: - __j += __s.__current_ - __current; - __current = __s.__current_; - break; case __state::__repeat: case __state::__accept_but_not_consume: break; @@ -6442,7 +6440,7 @@ public: regex_constants::match_flag_type __m = regex_constants::match_default) = delete; #endif -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG template regex_token_iterator(_BidirectionalIterator __a, _BidirectionalIterator __b, @@ -6557,7 +6555,7 @@ regex_token_iterator<_BidirectionalIterator, _CharT, _Traits>:: __init(__a, __b); } -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG template template @@ -6774,4 +6772,4 @@ _LIBCPP_END_NAMESPACE_STD _LIBCPP_POP_MACROS -#endif // _LIBCPP_REGEX +#endif // _LIBCPP_REGEX diff --git a/lib/libcxx/include/scoped_allocator b/lib/libcxx/include/scoped_allocator index 237cd428d..dc24d30b5 100644 --- a/lib/libcxx/include/scoped_allocator +++ b/lib/libcxx/include/scoped_allocator @@ -106,6 +106,7 @@ template */ #include <__config> +#include <__utility/forward.h> #include #include @@ -375,7 +376,7 @@ struct __outermost<_Alloc, true> { typedef typename remove_reference < - decltype(_VSTD::declval<_Alloc>().outer_allocator()) + decltype(declval<_Alloc>().outer_allocator()) >::type _OuterAlloc; typedef typename __outermost<_OuterAlloc>::type type; _LIBCPP_INLINE_VISIBILITY @@ -676,8 +677,8 @@ operator!=(const scoped_allocator_adaptor<_OuterA1, _InnerAllocs...>& __a, return !(__a == __b); } -#endif // !defined(_LIBCPP_CXX03_LANG) +#endif // !defined(_LIBCPP_CXX03_LANG) _LIBCPP_END_NAMESPACE_STD -#endif // _LIBCPP_SCOPED_ALLOCATOR +#endif // _LIBCPP_SCOPED_ALLOCATOR diff --git a/lib/libcxx/include/semaphore b/lib/libcxx/include/semaphore index 0e0434f7f..906f62e0f 100644 --- a/lib/libcxx/include/semaphore +++ b/lib/libcxx/include/semaphore @@ -45,8 +45,8 @@ using binary_semaphore = counting_semaphore<1>; */ -#include <__config> #include <__availability> +#include <__config> #include <__threading_support> #include @@ -98,7 +98,7 @@ public: _LIBCPP_AVAILABILITY_SYNC _LIBCPP_INLINE_VISIBILITY void acquire() { - auto const __test_fn = [=]() -> bool { + auto const __test_fn = [this]() -> bool { auto __old = __a.load(memory_order_relaxed); return (__old != 0) && __a.compare_exchange_strong(__old, __old - 1, memory_order_acquire, memory_order_relaxed); }; @@ -108,7 +108,7 @@ public: _LIBCPP_AVAILABILITY_SYNC _LIBCPP_INLINE_VISIBILITY bool try_acquire_for(chrono::duration const& __rel_time) { - auto const __test_fn = [=]() -> bool { + auto const __test_fn = [this]() -> bool { auto __old = __a.load(memory_order_acquire); while(1) { if (__old == 0) diff --git a/lib/libcxx/include/set b/lib/libcxx/include/set index d58455bfe..21ec8435d 100644 --- a/lib/libcxx/include/set +++ b/lib/libcxx/include/set @@ -154,10 +154,14 @@ public: iterator find(const K& x); template const_iterator find(const K& x) const; // C++14 + template size_type count(const K& x) const; // C++14 size_type count(const key_type& k) const; - bool contains(const key_type& x) const; // C++20 + + bool contains(const key_type& x) const; // C++20 + template bool contains(const K& x) const; // C++20 + iterator lower_bound(const key_type& k); const_iterator lower_bound(const key_type& k) const; template @@ -355,10 +359,14 @@ public: iterator find(const K& x); template const_iterator find(const K& x) const; // C++14 + template size_type count(const K& x) const; // C++14 size_type count(const key_type& k) const; - bool contains(const key_type& x) const; // C++20 + + bool contains(const key_type& x) const; // C++20 + template bool contains(const K& x) const; // C++20 + iterator lower_bound(const key_type& k); const_iterator lower_bound(const key_type& k) const; template @@ -426,9 +434,15 @@ erase_if(multiset& c, Predicate pred); // C++20 */ #include <__config> -#include <__tree> +#include <__debug> +#include <__functional/is_transparent.h> #include <__node_handle> +#include <__tree> +#include <__utility/forward.h> +#include #include +#include +#include // __libcpp_erase_if_container #include #if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) @@ -450,7 +464,7 @@ public: typedef key_type value_type; typedef _Compare key_compare; typedef key_compare value_compare; - typedef typename __identity<_Allocator>::type allocator_type; + typedef __identity_t<_Allocator> allocator_type; typedef value_type& reference; typedef const value_type& const_reference; @@ -546,7 +560,7 @@ public: set(set&& __s) _NOEXCEPT_(is_nothrow_move_constructible<__base>::value) : __tree_(_VSTD::move(__s.__tree_)) {} -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG _LIBCPP_INLINE_VISIBILITY explicit set(const allocator_type& __a) @@ -597,7 +611,7 @@ public: __tree_ = _VSTD::move(__s.__tree_); return *this; } -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG _LIBCPP_INLINE_VISIBILITY ~set() { @@ -652,7 +666,7 @@ public: _LIBCPP_INLINE_VISIBILITY iterator emplace_hint(const_iterator __p, _Args&&... __args) {return __tree_.__emplace_hint_unique(__p, _VSTD::forward<_Args>(__args)...);} -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG _LIBCPP_INLINE_VISIBILITY pair insert(const value_type& __v) @@ -681,7 +695,7 @@ public: _LIBCPP_INLINE_VISIBILITY void insert(initializer_list __il) {insert(__il.begin(), __il.end());} -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG _LIBCPP_INLINE_VISIBILITY iterator erase(const_iterator __p) {return __tree_.erase(__p);} @@ -795,6 +809,10 @@ public: #if _LIBCPP_STD_VER > 17 _LIBCPP_INLINE_VISIBILITY bool contains(const key_type& __k) const {return find(__k) != end();} + template + _LIBCPP_INLINE_VISIBILITY + typename enable_if<__is_transparent<_Compare, _K2>::value, bool>::type + contains(const _K2& __k) const { return find(__k) != end(); } #endif // _LIBCPP_STD_VER > 17 _LIBCPP_INLINE_VISIBILITY @@ -852,12 +870,12 @@ public: #ifndef _LIBCPP_HAS_NO_DEDUCTION_GUIDES template::value_type>, - class _Allocator = allocator::value_type>, + class _Compare = less<__iter_value_type<_InputIterator>>, + class _Allocator = allocator<__iter_value_type<_InputIterator>>, class = _EnableIf<__is_allocator<_Allocator>::value, void>, class = _EnableIf::value, void>> set(_InputIterator, _InputIterator, _Compare = _Compare(), _Allocator = _Allocator()) - -> set::value_type, _Compare, _Allocator>; + -> set<__iter_value_type<_InputIterator>, _Compare, _Allocator>; template, class _Allocator = allocator<_Key>, @@ -869,8 +887,8 @@ set(initializer_list<_Key>, _Compare = _Compare(), _Allocator = _Allocator()) template::value, void>> set(_InputIterator, _InputIterator, _Allocator) - -> set::value_type, - less::value_type>, _Allocator>; + -> set<__iter_value_type<_InputIterator>, + less<__iter_value_type<_InputIterator>>, _Allocator>; template::value, void>> @@ -892,7 +910,7 @@ set<_Key, _Compare, _Allocator>::set(set&& __s, const allocator_type& __a) } } -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG template inline _LIBCPP_INLINE_VISIBILITY @@ -964,7 +982,7 @@ template inline _LIBCPP_INLINE_VISIBILITY typename set<_Key, _Compare, _Allocator>::size_type erase_if(set<_Key, _Compare, _Allocator>& __c, _Predicate __pred) { - return __libcpp_erase_if_container(__c, __pred); + return _VSTD::__libcpp_erase_if_container(__c, __pred); } #endif @@ -974,11 +992,11 @@ class _LIBCPP_TEMPLATE_VIS multiset { public: // types: - typedef _Key key_type; + typedef _Key key_type; typedef key_type value_type; - typedef _Compare key_compare; + typedef _Compare key_compare; typedef key_compare value_compare; - typedef typename __identity<_Allocator>::type allocator_type; + typedef __identity_t<_Allocator> allocator_type; typedef value_type& reference; typedef const value_type& const_reference; @@ -1077,7 +1095,7 @@ public: : __tree_(_VSTD::move(__s.__tree_)) {} multiset(multiset&& __s, const allocator_type& __a); -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG _LIBCPP_INLINE_VISIBILITY explicit multiset(const allocator_type& __a) : __tree_(__a) {} @@ -1124,7 +1142,7 @@ public: __tree_ = _VSTD::move(__s.__tree_); return *this; } -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG _LIBCPP_INLINE_VISIBILITY ~multiset() { @@ -1179,7 +1197,7 @@ public: _LIBCPP_INLINE_VISIBILITY iterator emplace_hint(const_iterator __p, _Args&&... __args) {return __tree_.__emplace_hint_multi(__p, _VSTD::forward<_Args>(__args)...);} -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG _LIBCPP_INLINE_VISIBILITY iterator insert(const value_type& __v) @@ -1208,7 +1226,7 @@ public: _LIBCPP_INLINE_VISIBILITY void insert(initializer_list __il) {insert(__il.begin(), __il.end());} -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG _LIBCPP_INLINE_VISIBILITY iterator erase(const_iterator __p) {return __tree_.erase(__p);} @@ -1301,11 +1319,11 @@ public: #if _LIBCPP_STD_VER > 11 template _LIBCPP_INLINE_VISIBILITY - typename _VSTD::enable_if<_VSTD::__is_transparent<_Compare, _K2>::value,iterator>::type + typename enable_if<__is_transparent<_Compare, _K2>::value,iterator>::type find(const _K2& __k) {return __tree_.find(__k);} template _LIBCPP_INLINE_VISIBILITY - typename _VSTD::enable_if<_VSTD::__is_transparent<_Compare, _K2>::value,const_iterator>::type + typename enable_if<__is_transparent<_Compare, _K2>::value,const_iterator>::type find(const _K2& __k) const {return __tree_.find(__k);} #endif @@ -1322,6 +1340,10 @@ public: #if _LIBCPP_STD_VER > 17 _LIBCPP_INLINE_VISIBILITY bool contains(const key_type& __k) const {return find(__k) != end();} + template + _LIBCPP_INLINE_VISIBILITY + typename enable_if<__is_transparent<_Compare, _K2>::value, bool>::type + contains(const _K2& __k) const { return find(__k) != end(); } #endif // _LIBCPP_STD_VER > 17 _LIBCPP_INLINE_VISIBILITY @@ -1333,12 +1355,12 @@ public: #if _LIBCPP_STD_VER > 11 template _LIBCPP_INLINE_VISIBILITY - typename _VSTD::enable_if<_VSTD::__is_transparent<_Compare, _K2>::value,iterator>::type + typename enable_if<__is_transparent<_Compare, _K2>::value,iterator>::type lower_bound(const _K2& __k) {return __tree_.lower_bound(__k);} template _LIBCPP_INLINE_VISIBILITY - typename _VSTD::enable_if<_VSTD::__is_transparent<_Compare, _K2>::value,const_iterator>::type + typename enable_if<__is_transparent<_Compare, _K2>::value,const_iterator>::type lower_bound(const _K2& __k) const {return __tree_.lower_bound(__k);} #endif @@ -1351,11 +1373,11 @@ public: #if _LIBCPP_STD_VER > 11 template _LIBCPP_INLINE_VISIBILITY - typename _VSTD::enable_if<_VSTD::__is_transparent<_Compare, _K2>::value,iterator>::type + typename enable_if<__is_transparent<_Compare, _K2>::value,iterator>::type upper_bound(const _K2& __k) {return __tree_.upper_bound(__k);} template _LIBCPP_INLINE_VISIBILITY - typename _VSTD::enable_if<_VSTD::__is_transparent<_Compare, _K2>::value,const_iterator>::type + typename enable_if<__is_transparent<_Compare, _K2>::value,const_iterator>::type upper_bound(const _K2& __k) const {return __tree_.upper_bound(__k);} #endif @@ -1368,23 +1390,23 @@ public: #if _LIBCPP_STD_VER > 11 template _LIBCPP_INLINE_VISIBILITY - typename _VSTD::enable_if<_VSTD::__is_transparent<_Compare, _K2>::value,pair>::type + typename enable_if<__is_transparent<_Compare, _K2>::value,pair>::type equal_range(const _K2& __k) {return __tree_.__equal_range_multi(__k);} template _LIBCPP_INLINE_VISIBILITY - typename _VSTD::enable_if<_VSTD::__is_transparent<_Compare, _K2>::value,pair>::type + typename enable_if<__is_transparent<_Compare, _K2>::value,pair>::type equal_range(const _K2& __k) const {return __tree_.__equal_range_multi(__k);} #endif }; #ifndef _LIBCPP_HAS_NO_DEDUCTION_GUIDES template::value_type>, - class _Allocator = allocator::value_type>, + class _Compare = less<__iter_value_type<_InputIterator>>, + class _Allocator = allocator<__iter_value_type<_InputIterator>>, class = _EnableIf<__is_allocator<_Allocator>::value, void>, class = _EnableIf::value, void>> multiset(_InputIterator, _InputIterator, _Compare = _Compare(), _Allocator = _Allocator()) - -> multiset::value_type, _Compare, _Allocator>; + -> multiset<__iter_value_type<_InputIterator>, _Compare, _Allocator>; template, class _Allocator = allocator<_Key>, @@ -1396,8 +1418,8 @@ multiset(initializer_list<_Key>, _Compare = _Compare(), _Allocator = _Allocator( template::value, void>> multiset(_InputIterator, _InputIterator, _Allocator) - -> multiset::value_type, - less::value_type>, _Allocator>; + -> multiset<__iter_value_type<_InputIterator>, + less<__iter_value_type<_InputIterator>>, _Allocator>; template::value, void>> @@ -1419,7 +1441,7 @@ multiset<_Key, _Compare, _Allocator>::multiset(multiset&& __s, const allocator_t } } -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG template inline _LIBCPP_INLINE_VISIBILITY @@ -1490,10 +1512,10 @@ template inline _LIBCPP_INLINE_VISIBILITY typename multiset<_Key, _Compare, _Allocator>::size_type erase_if(multiset<_Key, _Compare, _Allocator>& __c, _Predicate __pred) { - return __libcpp_erase_if_container(__c, __pred); + return _VSTD::__libcpp_erase_if_container(__c, __pred); } #endif _LIBCPP_END_NAMESPACE_STD -#endif // _LIBCPP_SET +#endif // _LIBCPP_SET diff --git a/lib/libcxx/include/setjmp.h b/lib/libcxx/include/setjmp.h index f30a8d401..f9fb2ffe0 100644 --- a/lib/libcxx/include/setjmp.h +++ b/lib/libcxx/include/setjmp.h @@ -41,4 +41,4 @@ void longjmp(jmp_buf env, int val); #endif // __cplusplus -#endif // _LIBCPP_SETJMP_H +#endif // _LIBCPP_SETJMP_H diff --git a/lib/libcxx/include/shared_mutex b/lib/libcxx/include/shared_mutex index 5448d8a80..8205c3e0a 100644 --- a/lib/libcxx/include/shared_mutex +++ b/lib/libcxx/include/shared_mutex @@ -122,8 +122,8 @@ template */ -#include <__config> #include <__availability> +#include <__config> #include _LIBCPP_PUSH_MACROS @@ -500,10 +500,10 @@ swap(shared_lock<_Mutex>& __x, shared_lock<_Mutex>& __y) _NOEXCEPT _LIBCPP_END_NAMESPACE_STD -#endif // !_LIBCPP_HAS_NO_THREADS +#endif // !_LIBCPP_HAS_NO_THREADS -#endif // _LIBCPP_STD_VER > 11 +#endif // _LIBCPP_STD_VER > 11 _LIBCPP_POP_MACROS -#endif // _LIBCPP_SHARED_MUTEX +#endif // _LIBCPP_SHARED_MUTEX diff --git a/lib/libcxx/include/span b/lib/libcxx/include/span index 4f63d0ac4..260a74404 100644 --- a/lib/libcxx/include/span +++ b/lib/libcxx/include/span @@ -22,6 +22,12 @@ inline constexpr size_t dynamic_extent = numeric_limits::max(); template class span; +template + inline constexpr bool ranges::enable_view> = true; + +template + inline constexpr bool ranges::enable_borrowed_range> = true; + // [span.objectrep], views of object representation template span (sizeof(ElementType) * Extent))> as_writable_bytes(span s) noexcept; -namespace std { template class span { public: @@ -123,10 +128,16 @@ template */ #include <__config> +#include <__debug> +#include <__iterator/wrap_iter.h> +#include <__ranges/enable_borrowed_range.h> +#include <__ranges/enable_view.h> #include // for array #include // for byte #include // for iterators +#include #include // for remove_cv, etc +#include #if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) #pragma GCC system_header @@ -197,7 +208,11 @@ public: using const_pointer = const _Tp *; using reference = _Tp &; using const_reference = const _Tp &; - using iterator = __wrap_iter; +#if (_LIBCPP_DEBUG_LEVEL == 2) || defined(_LIBCPP_ABI_SPAN_POINTER_ITERATORS) + using iterator = pointer; +#else + using iterator = __wrap_iter; +#endif using reverse_iterator = _VSTD::reverse_iterator; static constexpr size_type extent = _Extent; @@ -372,7 +387,11 @@ public: using const_pointer = const _Tp *; using reference = _Tp &; using const_reference = const _Tp &; - using iterator = __wrap_iter; +#if (_LIBCPP_DEBUG_LEVEL == 2) || defined(_LIBCPP_ABI_SPAN_POINTER_ITERATORS) + using iterator = pointer; +#else + using iterator = __wrap_iter; +#endif using reverse_iterator = _VSTD::reverse_iterator; static constexpr size_type extent = dynamic_extent; @@ -516,6 +535,14 @@ private: size_type __size; }; +#if !defined(_LIBCPP_HAS_NO_RANGES) +template +inline constexpr bool ranges::enable_borrowed_range > = true; + +template +inline constexpr bool ranges::enable_view> = true; +#endif // !defined(_LIBCPP_HAS_NO_RANGES) + // as_bytes & as_writable_bytes template _LIBCPP_INLINE_VISIBILITY diff --git a/lib/libcxx/include/sstream b/lib/libcxx/include/sstream index 7ce85be6a..fbe5ffcab 100644 --- a/lib/libcxx/include/sstream +++ b/lib/libcxx/include/sstream @@ -181,8 +181,8 @@ typedef basic_stringstream wstringstream; */ #include <__config> -#include #include +#include #include #if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) @@ -219,19 +219,13 @@ private: public: // 30.8.2.1 [stringbuf.cons], constructors -#ifndef _LIBCPP_CXX03_LANG _LIBCPP_INLINE_VISIBILITY - basic_stringbuf() : basic_stringbuf(ios_base::in | ios_base::out) {} + basic_stringbuf() + : __hm_(nullptr), __mode_(ios_base::in | ios_base::out) {} _LIBCPP_INLINE_VISIBILITY explicit basic_stringbuf(ios_base::openmode __wch) : __hm_(nullptr), __mode_(__wch) {} -#else - _LIBCPP_INLINE_VISIBILITY - explicit basic_stringbuf(ios_base::openmode __wch = ios_base::in | - ios_base::out) - : __hm_(nullptr), __mode_(__wch) {} -#endif _LIBCPP_INLINE_VISIBILITY explicit basic_stringbuf(const string_type& __s, @@ -542,7 +536,7 @@ basic_stringbuf<_CharT, _Traits, _Allocator>::overflow(int_type __c) #ifndef _LIBCPP_NO_EXCEPTIONS try { -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS ptrdiff_t __nout = this->pptr() - this->pbase(); ptrdiff_t __hm = __hm_ - this->pbase(); __str_.push_back(char_type()); @@ -557,7 +551,7 @@ basic_stringbuf<_CharT, _Traits, _Allocator>::overflow(int_type __c) { return traits_type::eof(); } -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS } __hm_ = _VSTD::max(this->pptr() + 1, __hm_); if (__mode_ & ios_base::in) @@ -643,18 +637,13 @@ private: public: // 30.8.3.1 [istringstream.cons], constructors -#ifndef _LIBCPP_CXX03_LANG _LIBCPP_INLINE_VISIBILITY - basic_istringstream() : basic_istringstream(ios_base::in) {} + basic_istringstream() + : basic_istream<_CharT, _Traits>(&__sb_), __sb_(ios_base::in) {} _LIBCPP_INLINE_VISIBILITY explicit basic_istringstream(ios_base::openmode __wch) : basic_istream<_CharT, _Traits>(&__sb_), __sb_(__wch | ios_base::in) {} -#else - _LIBCPP_INLINE_VISIBILITY - explicit basic_istringstream(ios_base::openmode __wch = ios_base::in) - : basic_istream<_CharT, _Traits>(&__sb_), __sb_(__wch | ios_base::in) {} -#endif _LIBCPP_INLINE_VISIBILITY explicit basic_istringstream(const string_type& __s, @@ -728,20 +717,13 @@ private: public: // 30.8.4.1 [ostringstream.cons], constructors -#ifndef _LIBCPP_CXX03_LANG _LIBCPP_INLINE_VISIBILITY - basic_ostringstream() : basic_ostringstream(ios_base::out) {} + basic_ostringstream() + : basic_ostream<_CharT, _Traits>(&__sb_), __sb_(ios_base::out) {} _LIBCPP_INLINE_VISIBILITY explicit basic_ostringstream(ios_base::openmode __wch) - : basic_ostream<_CharT, _Traits>(&__sb_), - __sb_(__wch | ios_base::out) {} -#else - _LIBCPP_INLINE_VISIBILITY - explicit basic_ostringstream(ios_base::openmode __wch = ios_base::out) - : basic_ostream<_CharT, _Traits>(&__sb_), - __sb_(__wch | ios_base::out) {} -#endif + : basic_ostream<_CharT, _Traits>(&__sb_), __sb_(__wch | ios_base::out) {} _LIBCPP_INLINE_VISIBILITY explicit basic_ostringstream(const string_type& __s, @@ -816,19 +798,13 @@ private: public: // 30.8.5.1 [stringstream.cons], constructors -#ifndef _LIBCPP_CXX03_LANG _LIBCPP_INLINE_VISIBILITY - basic_stringstream() : basic_stringstream(ios_base::in | ios_base::out) {} + basic_stringstream() + : basic_iostream<_CharT, _Traits>(&__sb_), __sb_(ios_base::in | ios_base::out) {} _LIBCPP_INLINE_VISIBILITY explicit basic_stringstream(ios_base::openmode __wch) : basic_iostream<_CharT, _Traits>(&__sb_), __sb_(__wch) {} -#else - _LIBCPP_INLINE_VISIBILITY - explicit basic_stringstream(ios_base::openmode __wch = ios_base::in | - ios_base::out) - : basic_iostream<_CharT, _Traits>(&__sb_), __sb_(__wch) {} -#endif _LIBCPP_INLINE_VISIBILITY explicit basic_stringstream(const string_type& __s, @@ -892,4 +868,4 @@ _LIBCPP_END_NAMESPACE_STD _LIBCPP_POP_MACROS -#endif // _LIBCPP_SSTREAM +#endif // _LIBCPP_SSTREAM diff --git a/lib/libcxx/include/stack b/lib/libcxx/include/stack index 2a2b35038..aefef31ac 100644 --- a/lib/libcxx/include/stack +++ b/lib/libcxx/include/stack @@ -88,6 +88,8 @@ template */ #include <__config> +#include <__memory/uses_allocator.h> +#include <__utility/forward.h> #include #if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) @@ -148,7 +150,7 @@ public: _LIBCPP_INLINE_VISIBILITY explicit stack(container_type&& __c) : c(_VSTD::move(__c)) {} -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG _LIBCPP_INLINE_VISIBILITY explicit stack(const container_type& __c) : c(__c) {} @@ -156,35 +158,30 @@ public: template _LIBCPP_INLINE_VISIBILITY explicit stack(const _Alloc& __a, - typename enable_if::value>::type* = 0) + _EnableIf::value>* = 0) : c(__a) {} template _LIBCPP_INLINE_VISIBILITY stack(const container_type& __c, const _Alloc& __a, - typename enable_if::value>::type* = 0) + _EnableIf::value>* = 0) : c(__c, __a) {} template _LIBCPP_INLINE_VISIBILITY stack(const stack& __s, const _Alloc& __a, - typename enable_if::value>::type* = 0) + _EnableIf::value>* = 0) : c(__s.c, __a) {} #ifndef _LIBCPP_CXX03_LANG template _LIBCPP_INLINE_VISIBILITY stack(container_type&& __c, const _Alloc& __a, - typename enable_if::value>::type* = 0) + _EnableIf::value>* = 0) : c(_VSTD::move(__c), __a) {} template _LIBCPP_INLINE_VISIBILITY stack(stack&& __s, const _Alloc& __a, - typename enable_if::value>::type* = 0) + _EnableIf::value>* = 0) : c(_VSTD::move(__s.c), __a) {} -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG _LIBCPP_NODISCARD_AFTER_CXX17 _LIBCPP_INLINE_VISIBILITY bool empty() const {return c.empty();} @@ -210,7 +207,7 @@ public: void emplace(_Args&&... __args) { c.emplace_back(_VSTD::forward<_Args>(__args)...);} #endif -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG _LIBCPP_INLINE_VISIBILITY void pop() {c.pop_back();} @@ -236,15 +233,15 @@ public: #ifndef _LIBCPP_HAS_NO_DEDUCTION_GUIDES template::value, nullptr_t>::type + class = _EnableIf::value> > stack(_Container) -> stack; template::value, nullptr_t>::type, - class = typename enable_if< __is_allocator<_Alloc>::value, nullptr_t>::type + class = _EnableIf::value>, + class = _EnableIf::value> > stack(_Container, _Alloc) -> stack; @@ -300,10 +297,7 @@ operator<=(const stack<_Tp, _Container>& __x, const stack<_Tp, _Container>& __y) template inline _LIBCPP_INLINE_VISIBILITY -typename enable_if< - __is_swappable<_Container>::value, - void ->::type +_EnableIf<__is_swappable<_Container>::value, void> swap(stack<_Tp, _Container>& __x, stack<_Tp, _Container>& __y) _NOEXCEPT_(_NOEXCEPT_(__x.swap(__y))) { @@ -318,4 +312,4 @@ struct _LIBCPP_TEMPLATE_VIS uses_allocator, _Alloc> _LIBCPP_END_NAMESPACE_STD -#endif // _LIBCPP_STACK +#endif // _LIBCPP_STACK diff --git a/lib/libcxx/include/stdbool.h b/lib/libcxx/include/stdbool.h index 81a7cb303..69f7719f2 100644 --- a/lib/libcxx/include/stdbool.h +++ b/lib/libcxx/include/stdbool.h @@ -35,4 +35,4 @@ Macros: #define __bool_true_false_are_defined 1 #endif -#endif // _LIBCPP_STDBOOL_H +#endif // _LIBCPP_STDBOOL_H diff --git a/lib/libcxx/include/stddef.h b/lib/libcxx/include/stddef.h index 35c680b4f..2f8c0a5f5 100644 --- a/lib/libcxx/include/stddef.h +++ b/lib/libcxx/include/stddef.h @@ -53,4 +53,4 @@ using std::nullptr_t; #endif -#endif // _LIBCPP_STDDEF_H +#endif // _LIBCPP_STDDEF_H diff --git a/lib/libcxx/include/stdexcept b/lib/libcxx/include/stdexcept index 7b5de7ea3..c0470d1e1 100644 --- a/lib/libcxx/include/stdexcept +++ b/lib/libcxx/include/stdexcept @@ -42,9 +42,9 @@ public: */ #include <__config> -#include #include #include // for string forward decl +#include #if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) #pragma GCC system_header @@ -306,4 +306,4 @@ void __throw_underflow_error(const char*__msg) _LIBCPP_END_NAMESPACE_STD -#endif // _LIBCPP_STDEXCEPT +#endif // _LIBCPP_STDEXCEPT diff --git a/lib/libcxx/include/stdint.h b/lib/libcxx/include/stdint.h index d7b690251..5d64910bf 100644 --- a/lib/libcxx/include/stdint.h +++ b/lib/libcxx/include/stdint.h @@ -122,4 +122,4 @@ Macros: #include_next -#endif // _LIBCPP_STDINT_H +#endif // _LIBCPP_STDINT_H diff --git a/lib/libcxx/include/stdio.h b/lib/libcxx/include/stdio.h index e08e6bc9b..f84122034 100644 --- a/lib/libcxx/include/stdio.h +++ b/lib/libcxx/include/stdio.h @@ -116,4 +116,4 @@ void perror(const char* s); #endif -#endif // _LIBCPP_STDIO_H +#endif // _LIBCPP_STDIO_H diff --git a/lib/libcxx/include/stdlib.h b/lib/libcxx/include/stdlib.h index 812ea1024..242eedc22 100644 --- a/lib/libcxx/include/stdlib.h +++ b/lib/libcxx/include/stdlib.h @@ -103,7 +103,7 @@ extern "C++" { #endif // MSVCRT already has the correct prototype in if __cplusplus is defined -#if !defined(_LIBCPP_MSVCRT) && !defined(__sun__) && !defined(_AIX) +#if !defined(_LIBCPP_MSVCRT) && !defined(__sun__) inline _LIBCPP_INLINE_VISIBILITY long abs(long __x) _NOEXCEPT { return __builtin_labs(__x); } @@ -112,9 +112,9 @@ inline _LIBCPP_INLINE_VISIBILITY long long abs(long long __x) _NOEXCEPT { return __builtin_llabs(__x); } #endif // _LIBCPP_HAS_NO_LONG_LONG -#endif // !defined(_LIBCPP_MSVCRT) && !defined(__sun__) && !defined(_AIX) +#endif // !defined(_LIBCPP_MSVCRT) && !defined(__sun__) -#if !(defined(_AIX) || defined(__sun__)) +#if !defined(__sun__) inline _LIBCPP_INLINE_VISIBILITY float abs(float __lcpp_x) _NOEXCEPT { return __builtin_fabsf(__lcpp_x); // Use builtins to prevent needing math.h } @@ -127,7 +127,7 @@ inline _LIBCPP_INLINE_VISIBILITY long double abs(long double __lcpp_x) _NOEXCEPT { return __builtin_fabsl(__lcpp_x); } -#endif // !(defined(_AIX) || defined(__sun__)) +#endif // !defined(__sun__) // div @@ -138,7 +138,7 @@ abs(long double __lcpp_x) _NOEXCEPT { #endif // MSVCRT already has the correct prototype in if __cplusplus is defined -#if !defined(_LIBCPP_MSVCRT) && !defined(__sun__) && !defined(_AIX) +#if !defined(_LIBCPP_MSVCRT) && !defined(__sun__) inline _LIBCPP_INLINE_VISIBILITY ldiv_t div(long __x, long __y) _NOEXCEPT { return ::ldiv(__x, __y); } @@ -148,8 +148,8 @@ inline _LIBCPP_INLINE_VISIBILITY lldiv_t div(long long __x, return ::lldiv(__x, __y); } #endif // _LIBCPP_HAS_NO_LONG_LONG -#endif // _LIBCPP_MSVCRT / __sun__ / _AIX +#endif // _LIBCPP_MSVCRT / __sun__ } // extern "C++" -#endif // __cplusplus +#endif // __cplusplus -#endif // _LIBCPP_STDLIB_H +#endif // _LIBCPP_STDLIB_H diff --git a/lib/libcxx/include/streambuf b/lib/libcxx/include/streambuf index 0dfa31449..385d48694 100644 --- a/lib/libcxx/include/streambuf +++ b/lib/libcxx/include/streambuf @@ -7,8 +7,8 @@ // //===----------------------------------------------------------------------===// -#ifndef _LIBCPP_STEAMBUF -#define _LIBCPP_STEAMBUF +#ifndef _LIBCPP_STREAMBUF +#define _LIBCPP_STREAMBUF /* streambuf synopsis @@ -108,8 +108,8 @@ protected: */ #include <__config> -#include #include +#include #if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) #pragma GCC system_header @@ -495,4 +495,4 @@ _LIBCPP_END_NAMESPACE_STD _LIBCPP_POP_MACROS -#endif // _LIBCPP_STEAMBUF +#endif // _LIBCPP_STREAMBUF diff --git a/lib/libcxx/include/string b/lib/libcxx/include/string index 687795c79..4159ea580 100644 --- a/lib/libcxx/include/string +++ b/lib/libcxx/include/string @@ -69,6 +69,9 @@ struct char_traits template <> struct char_traits; template <> struct char_traits; +template <> struct char_traits; // C++20 +template <> struct char_traits; +template <> struct char_traits; template, class Allocator = allocator > class basic_string @@ -107,6 +110,7 @@ public: explicit basic_string(const T& t, const Allocator& a = Allocator()); // C++17 basic_string(const value_type* s, const allocator_type& a = allocator_type()); basic_string(const value_type* s, size_type n, const allocator_type& a = allocator_type()); + basic_string(nullptr_t) = delete; // C++2b basic_string(size_type n, value_type c, const allocator_type& a = allocator_type()); template basic_string(InputIterator begin, InputIterator end, @@ -127,6 +131,7 @@ public: allocator_type::propagate_on_container_move_assignment::value || allocator_type::is_always_equal::value ); // C++17 basic_string& operator=(const value_type* s); + basic_string& operator=(nullptr_t) = delete; // C++2b basic_string& operator=(value_type c); basic_string& operator=(initializer_list); @@ -262,49 +267,49 @@ public: size_type find(const basic_string& str, size_type pos = 0) const noexcept; template - size_type find(const T& t, size_type pos = 0) const; // C++17 + size_type find(const T& t, size_type pos = 0) const noexcept; // C++17, noexcept as an extension size_type find(const value_type* s, size_type pos, size_type n) const noexcept; size_type find(const value_type* s, size_type pos = 0) const noexcept; size_type find(value_type c, size_type pos = 0) const noexcept; size_type rfind(const basic_string& str, size_type pos = npos) const noexcept; template - size_type rfind(const T& t, size_type pos = npos) const; // C++17 + size_type rfind(const T& t, size_type pos = npos) const noexcept; // C++17, noexcept as an extension size_type rfind(const value_type* s, size_type pos, size_type n) const noexcept; size_type rfind(const value_type* s, size_type pos = npos) const noexcept; size_type rfind(value_type c, size_type pos = npos) const noexcept; size_type find_first_of(const basic_string& str, size_type pos = 0) const noexcept; template - size_type find_first_of(const T& t, size_type pos = 0) const; // C++17 + size_type find_first_of(const T& t, size_type pos = 0) const noexcept; // C++17, noexcept as an extension size_type find_first_of(const value_type* s, size_type pos, size_type n) const noexcept; size_type find_first_of(const value_type* s, size_type pos = 0) const noexcept; size_type find_first_of(value_type c, size_type pos = 0) const noexcept; size_type find_last_of(const basic_string& str, size_type pos = npos) const noexcept; template - size_type find_last_of(const T& t, size_type pos = npos) const noexcept; // C++17 + size_type find_last_of(const T& t, size_type pos = npos) const noexcept noexcept; // C++17, noexcept as an extension size_type find_last_of(const value_type* s, size_type pos, size_type n) const noexcept; size_type find_last_of(const value_type* s, size_type pos = npos) const noexcept; size_type find_last_of(value_type c, size_type pos = npos) const noexcept; size_type find_first_not_of(const basic_string& str, size_type pos = 0) const noexcept; template - size_type find_first_not_of(const T& t, size_type pos = 0) const; // C++17 + size_type find_first_not_of(const T& t, size_type pos = 0) const noexcept; // C++17, noexcept as an extension size_type find_first_not_of(const value_type* s, size_type pos, size_type n) const noexcept; size_type find_first_not_of(const value_type* s, size_type pos = 0) const noexcept; size_type find_first_not_of(value_type c, size_type pos = 0) const noexcept; size_type find_last_not_of(const basic_string& str, size_type pos = npos) const noexcept; template - size_type find_last_not_of(const T& t, size_type pos = npos) const; // C++17 + size_type find_last_not_of(const T& t, size_type pos = npos) const noexcept; // C++17, noexcept as an extension size_type find_last_not_of(const value_type* s, size_type pos, size_type n) const noexcept; size_type find_last_not_of(const value_type* s, size_type pos = npos) const noexcept; size_type find_last_not_of(value_type c, size_type pos = npos) const noexcept; int compare(const basic_string& str) const noexcept; template - int compare(const T& t) const noexcept; // C++17 + int compare(const T& t) const noexcept; // C++17, noexcept as an extension int compare(size_type pos1, size_type n1, const basic_string& str) const; template int compare(size_type pos1, size_type n1, const T& t) const; // C++17 @@ -450,6 +455,7 @@ erase_if(basic_string& c, Predicate pred); // C++20 typedef basic_string string; typedef basic_string wstring; +typedef basic_string u8string; // C++20 typedef basic_string u16string; typedef basic_string u32string; @@ -494,12 +500,14 @@ wstring to_wstring(double val); wstring to_wstring(long double val); template <> struct hash; +template <> struct hash; // C++20 template <> struct hash; template <> struct hash; template <> struct hash; basic_string operator "" s( const char *str, size_t len ); // C++14 basic_string operator "" s( const wchar_t *str, size_t len ); // C++14 +basic_string operator "" s( const char8_t *str, size_t len ); // C++20 basic_string operator "" s( const char16_t *str, size_t len ); // C++14 basic_string operator "" s( const char32_t *str, size_t len ); // C++14 @@ -508,26 +516,29 @@ basic_string operator "" s( const char32_t *str, size_t len ); // C++1 */ #include <__config> -#include -#include +#include <__debug> +#include <__functional_base> +#include <__iterator/wrap_iter.h> +#include +#include +#include // EOF +#include #include -#include // For EOF. #include -#include +#include +#include #include -#include #include #include +#include #include -#include -#include <__functional_base> +#include #include + #ifndef _LIBCPP_HAS_NO_UNICODE_CHARS -#include +# include #endif -#include <__debug> - #if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) #pragma GCC system_header #endif @@ -625,29 +636,16 @@ __basic_string_common<__b>::__throw_out_of_range() const _LIBCPP_EXTERN_TEMPLATE(class _LIBCPP_EXTERN_TEMPLATE_TYPE_VIS __basic_string_common) -#ifdef _LIBCPP_NO_EXCEPTIONS template -struct __libcpp_string_gets_noexcept_iterator_impl : public true_type {}; -#elif defined(_LIBCPP_HAS_NO_NOEXCEPT) -template -struct __libcpp_string_gets_noexcept_iterator_impl : public false_type {}; -#else -template ::value> -struct __libcpp_string_gets_noexcept_iterator_impl : public _LIBCPP_BOOL_CONSTANT(( - noexcept(++(declval<_Iter&>())) && - is_nothrow_assignable<_Iter&, _Iter>::value && - noexcept(declval<_Iter>() == declval<_Iter>()) && - noexcept(*declval<_Iter>()) -)) {}; - -template -struct __libcpp_string_gets_noexcept_iterator_impl<_Iter, false> : public false_type {}; -#endif +struct __string_is_trivial_iterator : public false_type {}; +template +struct __string_is_trivial_iterator<_Tp*> + : public is_arithmetic<_Tp> {}; template -struct __libcpp_string_gets_noexcept_iterator - : public _LIBCPP_BOOL_CONSTANT(__libcpp_is_trivial_iterator<_Iter>::value || __libcpp_string_gets_noexcept_iterator_impl<_Iter>::value) {}; +struct __string_is_trivial_iterator<__wrap_iter<_Iter> > + : public __string_is_trivial_iterator<_Iter> {}; template struct __can_be_converted_to_string_view : public _BoolConstant< @@ -668,21 +666,21 @@ struct __padding<_CharT, 1> { }; -#endif // _LIBCPP_ABI_ALTERNATE_STRING_LAYOUT +#endif // _LIBCPP_ABI_ALTERNATE_STRING_LAYOUT -#ifndef _LIBCPP_NO_HAS_CHAR8_T +#ifndef _LIBCPP_HAS_NO_CHAR8_T typedef basic_string u8string; #endif #ifndef _LIBCPP_HAS_NO_UNICODE_CHARS typedef basic_string u16string; typedef basic_string u32string; -#endif // _LIBCPP_HAS_NO_UNICODE_CHARS +#endif // _LIBCPP_HAS_NO_UNICODE_CHARS template class _LIBCPP_TEMPLATE_VIS -#ifndef _LIBCPP_NO_HAS_CHAR8_T +#ifndef _LIBCPP_HAS_NO_CHAR8_T _LIBCPP_PREFERRED_NAME(u8string) #endif #ifndef _LIBCPP_HAS_NO_UNICODE_CHARS @@ -736,7 +734,7 @@ private: #else // _LIBCPP_BIG_ENDIAN static const size_type __short_mask = 0x80; static const size_type __long_mask = ~(size_type(~0) >> 1); -#endif // _LIBCPP_BIG_ENDIAN +#endif // _LIBCPP_BIG_ENDIAN enum {__min_cap = (sizeof(__long) - 1)/sizeof(value_type) > 2 ? (sizeof(__long) - 1)/sizeof(value_type) : 2}; @@ -766,7 +764,7 @@ private: #else // _LIBCPP_BIG_ENDIAN static const size_type __short_mask = 0x01; static const size_type __long_mask = 0x1ul; -#endif // _LIBCPP_BIG_ENDIAN +#endif // _LIBCPP_BIG_ENDIAN enum {__min_cap = (sizeof(__long) - 1)/sizeof(value_type) > 2 ? (sizeof(__long) - 1)/sizeof(value_type) : 2}; @@ -781,7 +779,7 @@ private: value_type __data_[__min_cap]; }; -#endif // _LIBCPP_ABI_ALTERNATE_STRING_LAYOUT +#endif // _LIBCPP_ABI_ALTERNATE_STRING_LAYOUT union __ulx{__long __lx; __short __lxx;}; @@ -805,7 +803,7 @@ private: __compressed_pair<__rep, allocator_type> __r_; public: - _LIBCPP_FUNC_VIS + _LIBCPP_TEMPLATE_DATA_VIS static const size_type npos = -1; _LIBCPP_INLINE_VISIBILITY basic_string() @@ -832,7 +830,7 @@ public: _LIBCPP_INLINE_VISIBILITY basic_string(basic_string&& __str, const allocator_type& __a); -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG template ::value, nullptr_t> > _LIBCPP_INLINE_VISIBILITY @@ -848,6 +846,10 @@ public: _LIBCPP_INLINE_VISIBILITY basic_string(const _CharT* __s, const _Allocator& __a); +#if _LIBCPP_STD_VER > 20 + basic_string(nullptr_t) = delete; +#endif + _LIBCPP_INLINE_VISIBILITY basic_string(const _CharT* __s, size_type __n); _LIBCPP_INLINE_VISIBILITY @@ -890,7 +892,7 @@ public: basic_string(initializer_list<_CharT> __il); _LIBCPP_INLINE_VISIBILITY basic_string(initializer_list<_CharT> __il, const _Allocator& __a); -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG inline ~basic_string(); @@ -911,6 +913,9 @@ public: basic_string& operator=(initializer_list __il) {return assign(__il.begin(), __il.size());} #endif _LIBCPP_INLINE_VISIBILITY basic_string& operator=(const value_type* __s) {return assign(__s);} +#if _LIBCPP_STD_VER > 20 + basic_string& operator=(nullptr_t) = delete; +#endif basic_string& operator=(value_type __c); #if _LIBCPP_DEBUG_LEVEL == 2 @@ -939,7 +944,7 @@ public: _LIBCPP_INLINE_VISIBILITY const_iterator end() const _NOEXCEPT {return const_iterator(__get_pointer() + size());} -#endif // _LIBCPP_DEBUG_LEVEL == 2 +#endif // _LIBCPP_DEBUG_LEVEL == 2 _LIBCPP_INLINE_VISIBILITY reverse_iterator rbegin() _NOEXCEPT {return reverse_iterator(end());} @@ -1010,7 +1015,7 @@ public: _LIBCPP_INLINE_VISIBILITY basic_string& operator+=(value_type __c) {push_back(__c); return *this;} #ifndef _LIBCPP_CXX03_LANG _LIBCPP_INLINE_VISIBILITY basic_string& operator+=(initializer_list __il) {return append(__il);} -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG _LIBCPP_INLINE_VISIBILITY basic_string& append(const basic_string& __str); @@ -1041,20 +1046,16 @@ public: _LIBCPP_INLINE_VISIBILITY void __append_default_init(size_type __n); - template - _LIBCPP_METHOD_TEMPLATE_IMPLICIT_INSTANTIATION_VIS - basic_string& __append_forward_unsafe(_ForwardIterator, _ForwardIterator); template _LIBCPP_METHOD_TEMPLATE_IMPLICIT_INSTANTIATION_VIS _EnableIf < - __is_exactly_cpp17_input_iterator<_InputIterator>::value - || !__libcpp_string_gets_noexcept_iterator<_InputIterator>::value, + __is_exactly_cpp17_input_iterator<_InputIterator>::value, basic_string& > _LIBCPP_INLINE_VISIBILITY append(_InputIterator __first, _InputIterator __last) { - const basic_string __temp (__first, __last, __alloc()); + const basic_string __temp(__first, __last, __alloc()); append(__temp.data(), __temp.size()); return *this; } @@ -1062,19 +1063,16 @@ public: _LIBCPP_METHOD_TEMPLATE_IMPLICIT_INSTANTIATION_VIS _EnableIf < - __is_cpp17_forward_iterator<_ForwardIterator>::value - && __libcpp_string_gets_noexcept_iterator<_ForwardIterator>::value, + __is_cpp17_forward_iterator<_ForwardIterator>::value, basic_string& > _LIBCPP_INLINE_VISIBILITY - append(_ForwardIterator __first, _ForwardIterator __last) { - return __append_forward_unsafe(__first, __last); - } + append(_ForwardIterator __first, _ForwardIterator __last); #ifndef _LIBCPP_CXX03_LANG _LIBCPP_INLINE_VISIBILITY basic_string& append(initializer_list __il) {return append(__il.begin(), __il.size());} -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG void push_back(value_type __c); _LIBCPP_INLINE_VISIBILITY @@ -1117,8 +1115,7 @@ public: _LIBCPP_METHOD_TEMPLATE_IMPLICIT_INSTANTIATION_VIS _EnableIf < - __is_exactly_cpp17_input_iterator<_InputIterator>::value - || !__libcpp_string_gets_noexcept_iterator<_InputIterator>::value, + __is_exactly_cpp17_input_iterator<_InputIterator>::value, basic_string& > assign(_InputIterator __first, _InputIterator __last); @@ -1126,15 +1123,14 @@ public: _LIBCPP_METHOD_TEMPLATE_IMPLICIT_INSTANTIATION_VIS _EnableIf < - __is_cpp17_forward_iterator<_ForwardIterator>::value - && __libcpp_string_gets_noexcept_iterator<_ForwardIterator>::value, + __is_cpp17_forward_iterator<_ForwardIterator>::value, basic_string& > assign(_ForwardIterator __first, _ForwardIterator __last); #ifndef _LIBCPP_CXX03_LANG _LIBCPP_INLINE_VISIBILITY basic_string& assign(initializer_list __il) {return assign(__il.begin(), __il.size());} -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG _LIBCPP_INLINE_VISIBILITY basic_string& insert(size_type __pos1, const basic_string& __str); @@ -1168,8 +1164,7 @@ public: _LIBCPP_METHOD_TEMPLATE_IMPLICIT_INSTANTIATION_VIS _EnableIf < - __is_exactly_cpp17_input_iterator<_InputIterator>::value - || !__libcpp_string_gets_noexcept_iterator<_InputIterator>::value, + __is_exactly_cpp17_input_iterator<_InputIterator>::value, iterator > insert(const_iterator __pos, _InputIterator __first, _InputIterator __last); @@ -1177,8 +1172,7 @@ public: _LIBCPP_METHOD_TEMPLATE_IMPLICIT_INSTANTIATION_VIS _EnableIf < - __is_cpp17_forward_iterator<_ForwardIterator>::value - && __libcpp_string_gets_noexcept_iterator<_ForwardIterator>::value, + __is_cpp17_forward_iterator<_ForwardIterator>::value, iterator > insert(const_iterator __pos, _ForwardIterator __first, _ForwardIterator __last); @@ -1186,7 +1180,7 @@ public: _LIBCPP_INLINE_VISIBILITY iterator insert(const_iterator __pos, initializer_list __il) {return insert(__pos, __il.begin(), __il.end());} -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG basic_string& erase(size_type __pos = 0, size_type __n = npos); _LIBCPP_INLINE_VISIBILITY @@ -1247,7 +1241,7 @@ public: _LIBCPP_INLINE_VISIBILITY basic_string& replace(const_iterator __i1, const_iterator __i2, initializer_list __il) {return replace(__i1, __i2, __il.begin(), __il.end());} -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG size_type copy(value_type* __s, size_type __n, size_type __pos = 0) const; _LIBCPP_INLINE_VISIBILITY @@ -1284,7 +1278,7 @@ public: __can_be_converted_to_string_view<_CharT, _Traits, _Tp>::value, size_type > - find(const _Tp& __t, size_type __pos = 0) const; + find(const _Tp& __t, size_type __pos = 0) const _NOEXCEPT; size_type find(const value_type* __s, size_type __pos, size_type __n) const _NOEXCEPT; _LIBCPP_INLINE_VISIBILITY size_type find(const value_type* __s, size_type __pos = 0) const _NOEXCEPT; @@ -1300,7 +1294,7 @@ public: __can_be_converted_to_string_view<_CharT, _Traits, _Tp>::value, size_type > - rfind(const _Tp& __t, size_type __pos = npos) const; + rfind(const _Tp& __t, size_type __pos = npos) const _NOEXCEPT; size_type rfind(const value_type* __s, size_type __pos, size_type __n) const _NOEXCEPT; _LIBCPP_INLINE_VISIBILITY size_type rfind(const value_type* __s, size_type __pos = npos) const _NOEXCEPT; @@ -1316,7 +1310,7 @@ public: __can_be_converted_to_string_view<_CharT, _Traits, _Tp>::value, size_type > - find_first_of(const _Tp& __t, size_type __pos = 0) const; + find_first_of(const _Tp& __t, size_type __pos = 0) const _NOEXCEPT; size_type find_first_of(const value_type* __s, size_type __pos, size_type __n) const _NOEXCEPT; _LIBCPP_INLINE_VISIBILITY size_type find_first_of(const value_type* __s, size_type __pos = 0) const _NOEXCEPT; @@ -1333,7 +1327,7 @@ public: __can_be_converted_to_string_view<_CharT, _Traits, _Tp>::value, size_type > - find_last_of(const _Tp& __t, size_type __pos = npos) const; + find_last_of(const _Tp& __t, size_type __pos = npos) const _NOEXCEPT; size_type find_last_of(const value_type* __s, size_type __pos, size_type __n) const _NOEXCEPT; _LIBCPP_INLINE_VISIBILITY size_type find_last_of(const value_type* __s, size_type __pos = npos) const _NOEXCEPT; @@ -1350,7 +1344,7 @@ public: __can_be_converted_to_string_view<_CharT, _Traits, _Tp>::value, size_type > - find_first_not_of(const _Tp &__t, size_type __pos = 0) const; + find_first_not_of(const _Tp &__t, size_type __pos = 0) const _NOEXCEPT; size_type find_first_not_of(const value_type* __s, size_type __pos, size_type __n) const _NOEXCEPT; _LIBCPP_INLINE_VISIBILITY size_type find_first_not_of(const value_type* __s, size_type __pos = 0) const _NOEXCEPT; @@ -1367,7 +1361,7 @@ public: __can_be_converted_to_string_view<_CharT, _Traits, _Tp>::value, size_type > - find_last_not_of(const _Tp& __t, size_type __pos = npos) const; + find_last_not_of(const _Tp& __t, size_type __pos = npos) const _NOEXCEPT; size_type find_last_not_of(const value_type* __s, size_type __pos, size_type __n) const _NOEXCEPT; _LIBCPP_INLINE_VISIBILITY size_type find_last_not_of(const value_type* __s, size_type __pos = npos) const _NOEXCEPT; @@ -1384,7 +1378,7 @@ public: __can_be_converted_to_string_view<_CharT, _Traits, _Tp>::value, int > - compare(const _Tp &__t) const; + compare(const _Tp &__t) const _NOEXCEPT; template _LIBCPP_METHOD_TEMPLATE_IMPLICIT_INSTANTIATION_VIS @@ -1468,7 +1462,7 @@ public: bool __addable(const const_iterator* __i, ptrdiff_t __n) const; bool __subscriptable(const const_iterator* __i, ptrdiff_t __n) const; -#endif // _LIBCPP_DEBUG_LEVEL == 2 +#endif // _LIBCPP_DEBUG_LEVEL == 2 private: _LIBCPP_INLINE_VISIBILITY @@ -1514,7 +1508,7 @@ private: {return __r_.first().__s.__size_ >> 1;} # endif -#endif // _LIBCPP_ABI_ALTERNATE_STRING_LAYOUT +#endif // _LIBCPP_ABI_ALTERNATE_STRING_LAYOUT _LIBCPP_INLINE_VISIBILITY void __set_long_size(size_type __s) _NOEXCEPT @@ -1714,6 +1708,31 @@ private: _LIBCPP_INLINE_VISIBILITY void __invalidate_all_iterators(); _LIBCPP_INLINE_VISIBILITY void __invalidate_iterators_past(size_type); + template + _LIBCPP_INLINE_VISIBILITY + bool __addr_in_range(_Tp&& __t) const { + const volatile void *__p = _VSTD::addressof(__t); + return data() <= __p && __p <= data() + size(); + } + + _LIBCPP_NORETURN _LIBCPP_HIDE_FROM_ABI + void __throw_length_error() const { +#ifndef _LIBCPP_NO_EXCEPTIONS + __basic_string_common::__throw_length_error(); +#else + _VSTD::abort(); +#endif + } + + _LIBCPP_NORETURN _LIBCPP_HIDE_FROM_ABI + void __throw_out_of_range() const { +#ifndef _LIBCPP_NO_EXCEPTIONS + __basic_string_common::__throw_out_of_range(); +#else + _VSTD::abort(); +#endif + } + friend basic_string operator+<>(const basic_string&, const basic_string&); friend basic_string operator+<>(const value_type*, const basic_string&); friend basic_string operator+<>(value_type, const basic_string&); @@ -1734,7 +1753,7 @@ _LIBCPP_STRING_V1_EXTERN_TEMPLATE_LIST(_LIBCPP_EXTERN_TEMPLATE, wchar_t) #ifndef _LIBCPP_HAS_NO_DEDUCTION_GUIDES template::value_type, + class _CharT = __iter_value_type<_InputIterator>, class _Allocator = allocator<_CharT>, class = _EnableIf<__is_cpp17_input_iterator<_InputIterator>::value>, class = _EnableIf<__is_allocator<_Allocator>::value> @@ -1773,11 +1792,7 @@ basic_string<_CharT, _Traits, _Allocator>::__invalidate_all_iterators() template inline void -basic_string<_CharT, _Traits, _Allocator>::__invalidate_iterators_past(size_type -#if _LIBCPP_DEBUG_LEVEL == 2 - __pos -#endif - ) +basic_string<_CharT, _Traits, _Allocator>::__invalidate_iterators_past(size_type __pos) { #if _LIBCPP_DEBUG_LEVEL == 2 __c_node* __c = __get_db()->__find_c_and_lock(this); @@ -1797,7 +1812,9 @@ basic_string<_CharT, _Traits, _Allocator>::__invalidate_iterators_past(size_type } __get_db()->unlock(); } -#endif // _LIBCPP_DEBUG_LEVEL == 2 +#else + (void)__pos; +#endif // _LIBCPP_DEBUG_LEVEL == 2 } template @@ -2001,7 +2018,7 @@ basic_string<_CharT, _Traits, _Allocator>::basic_string(basic_string&& __str, co #endif } -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG template void @@ -2129,7 +2146,7 @@ basic_string<_CharT, _Traits, _Allocator>::__init(_InputIterator __first, _Input #ifndef _LIBCPP_NO_EXCEPTIONS try { -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS for (; __first != __last; ++__first) push_back(*__first); #ifndef _LIBCPP_NO_EXCEPTIONS @@ -2140,7 +2157,7 @@ basic_string<_CharT, _Traits, _Allocator>::__init(_InputIterator __first, _Input __alloc_traits::deallocate(__alloc(), __get_long_pointer(), __get_long_cap()); throw; } -#endif // _LIBCPP_NO_EXCEPTIONS +#endif // _LIBCPP_NO_EXCEPTIONS } template @@ -2168,9 +2185,23 @@ basic_string<_CharT, _Traits, _Allocator>::__init(_ForwardIterator __first, _For __set_long_cap(__cap+1); __set_long_size(__sz); } + +#ifndef _LIBCPP_NO_EXCEPTIONS + try + { +#endif // _LIBCPP_NO_EXCEPTIONS for (; __first != __last; ++__first, (void) ++__p) traits_type::assign(*__p, *__first); traits_type::assign(*__p, value_type()); +#ifndef _LIBCPP_NO_EXCEPTIONS + } + catch (...) + { + if (__is_long()) + __alloc_traits::deallocate(__alloc(), __get_long_pointer(), __get_long_cap()); + throw; + } +#endif // _LIBCPP_NO_EXCEPTIONS } template @@ -2225,7 +2256,7 @@ basic_string<_CharT, _Traits, _Allocator>::basic_string( #endif } -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG template basic_string<_CharT, _Traits, _Allocator>::~basic_string() @@ -2357,12 +2388,11 @@ basic_string<_CharT, _Traits, _Allocator>::assign(size_type __n, value_type __c) size_type __sz = size(); __grow_by(__cap, __n - __cap, __sz, 0, __sz); } - else - __invalidate_iterators_past(__n); value_type* __p = _VSTD::__to_address(__get_pointer()); traits_type::assign(__p, __n, __c); traits_type::assign(__p[__n], value_type()); __set_size(__n); + __invalidate_iterators_past(__n); return *this; } @@ -2463,8 +2493,7 @@ template template _EnableIf < - __is_exactly_cpp17_input_iterator <_InputIterator>::value - || !__libcpp_string_gets_noexcept_iterator<_InputIterator>::value, + __is_exactly_cpp17_input_iterator<_InputIterator>::value, basic_string<_CharT, _Traits, _Allocator>& > basic_string<_CharT, _Traits, _Allocator>::assign(_InputIterator __first, _InputIterator __last) @@ -2478,26 +2507,35 @@ template template _EnableIf < - __is_cpp17_forward_iterator<_ForwardIterator>::value - && __libcpp_string_gets_noexcept_iterator<_ForwardIterator>::value, + __is_cpp17_forward_iterator<_ForwardIterator>::value, basic_string<_CharT, _Traits, _Allocator>& > basic_string<_CharT, _Traits, _Allocator>::assign(_ForwardIterator __first, _ForwardIterator __last) { - size_type __n = static_cast(_VSTD::distance(__first, __last)); size_type __cap = capacity(); - if (__cap < __n) + size_type __n = __string_is_trivial_iterator<_ForwardIterator>::value ? + static_cast(_VSTD::distance(__first, __last)) : 0; + + if (__string_is_trivial_iterator<_ForwardIterator>::value && + (__cap >= __n || !__addr_in_range(*__first))) { - size_type __sz = size(); - __grow_by(__cap, __n - __cap, __sz, 0, __sz); + if (__cap < __n) + { + size_type __sz = size(); + __grow_by(__cap, __n - __cap, __sz, 0, __sz); + } + pointer __p = __get_pointer(); + for (; __first != __last; ++__first, ++__p) + traits_type::assign(*__p, *__first); + traits_type::assign(*__p, value_type()); + __set_size(__n); + __invalidate_iterators_past(__n); } else - __invalidate_iterators_past(__n); - pointer __p = __get_pointer(); - for (; __first != __last; ++__first, ++__p) - traits_type::assign(*__p, *__first); - traits_type::assign(*__p, value_type()); - __set_size(__n); + { + const basic_string __temp(__first, __last, __alloc()); + assign(__temp.data(), __temp.size()); + } return *this; } @@ -2644,39 +2682,23 @@ basic_string<_CharT, _Traits, _Allocator>::push_back(value_type __c) traits_type::assign(*++__p, value_type()); } -template -bool __ptr_in_range (const _Tp* __p, const _Tp* __first, const _Tp* __last) -{ - return __first <= __p && __p < __last; -} - -template -bool __ptr_in_range (const _Tp1*, const _Tp2*, const _Tp2*) -{ - return false; -} - template template -basic_string<_CharT, _Traits, _Allocator>& -basic_string<_CharT, _Traits, _Allocator>::__append_forward_unsafe( +_EnableIf +< + __is_cpp17_forward_iterator<_ForwardIterator>::value, + basic_string<_CharT, _Traits, _Allocator>& +> +basic_string<_CharT, _Traits, _Allocator>::append( _ForwardIterator __first, _ForwardIterator __last) { - static_assert(__is_cpp17_forward_iterator<_ForwardIterator>::value, - "function requires a ForwardIterator"); size_type __sz = size(); size_type __cap = capacity(); size_type __n = static_cast(_VSTD::distance(__first, __last)); if (__n) { - typedef typename iterator_traits<_ForwardIterator>::reference _CharRef; - _CharRef __tmp_ref = *__first; - if (__ptr_in_range(_VSTD::addressof(__tmp_ref), data(), data() + size())) - { - const basic_string __temp (__first, __last, __alloc()); - append(__temp.data(), __temp.size()); - } - else + if (__string_is_trivial_iterator<_ForwardIterator>::value && + !__addr_in_range(*__first)) { if (__cap - __sz < __n) __grow_by(__cap, __sz + __n - __cap, __sz, __sz, 0); @@ -2686,6 +2708,11 @@ basic_string<_CharT, _Traits, _Allocator>::__append_forward_unsafe( traits_type::assign(*__p, value_type()); __set_size(__sz + __n); } + else + { + const basic_string __temp(__first, __last, __alloc()); + append(__temp.data(), __temp.size()); + } } return *this; } @@ -2801,8 +2828,7 @@ template template _EnableIf < - __is_exactly_cpp17_input_iterator<_InputIterator>::value - || !__libcpp_string_gets_noexcept_iterator<_InputIterator>::value, + __is_exactly_cpp17_input_iterator<_InputIterator>::value, typename basic_string<_CharT, _Traits, _Allocator>::iterator > basic_string<_CharT, _Traits, _Allocator>::insert(const_iterator __pos, _InputIterator __first, _InputIterator __last) @@ -2820,8 +2846,7 @@ template template _EnableIf < - __is_cpp17_forward_iterator<_ForwardIterator>::value - && __libcpp_string_gets_noexcept_iterator<_ForwardIterator>::value, + __is_cpp17_forward_iterator<_ForwardIterator>::value, typename basic_string<_CharT, _Traits, _Allocator>::iterator > basic_string<_CharT, _Traits, _Allocator>::insert(const_iterator __pos, _ForwardIterator __first, _ForwardIterator __last) @@ -2835,34 +2860,35 @@ basic_string<_CharT, _Traits, _Allocator>::insert(const_iterator __pos, _Forward size_type __n = static_cast(_VSTD::distance(__first, __last)); if (__n) { - typedef typename iterator_traits<_ForwardIterator>::reference _CharRef; - _CharRef __tmp_char = *__first; - if (__ptr_in_range(_VSTD::addressof(__tmp_char), data(), data() + size())) - { - const basic_string __temp(__first, __last, __alloc()); - return insert(__pos, __temp.data(), __temp.data() + __temp.size()); - } - - size_type __sz = size(); - size_type __cap = capacity(); - value_type* __p; - if (__cap - __sz >= __n) + if (__string_is_trivial_iterator<_ForwardIterator>::value && + !__addr_in_range(*__first)) { - __p = _VSTD::__to_address(__get_pointer()); - size_type __n_move = __sz - __ip; - if (__n_move != 0) - traits_type::move(__p + __ip + __n, __p + __ip, __n_move); + size_type __sz = size(); + size_type __cap = capacity(); + value_type* __p; + if (__cap - __sz >= __n) + { + __p = _VSTD::__to_address(__get_pointer()); + size_type __n_move = __sz - __ip; + if (__n_move != 0) + traits_type::move(__p + __ip + __n, __p + __ip, __n_move); + } + else + { + __grow_by(__cap, __sz + __n - __cap, __sz, __ip, 0, __n); + __p = _VSTD::__to_address(__get_long_pointer()); + } + __sz += __n; + __set_size(__sz); + traits_type::assign(__p[__sz], value_type()); + for (__p += __ip; __first != __last; ++__p, ++__first) + traits_type::assign(*__p, *__first); } else { - __grow_by(__cap, __sz + __n - __cap, __sz, __ip, 0, __n); - __p = _VSTD::__to_address(__get_long_pointer()); + const basic_string __temp(__first, __last, __alloc()); + return insert(__pos, __temp.data(), __temp.data() + __temp.size()); } - __sz += __n; - __set_size(__sz); - traits_type::assign(__p[__sz], value_type()); - for (__p += __ip; __first != __last; ++__p, ++__first) - traits_type::assign(*__p, *__first); } return begin() + __ip; } @@ -3353,7 +3379,7 @@ basic_string<_CharT, _Traits, _Allocator>::__shrink_or_extend(size_type __target #ifndef _LIBCPP_NO_EXCEPTIONS try { - #endif // _LIBCPP_NO_EXCEPTIONS + #endif // _LIBCPP_NO_EXCEPTIONS __new_data = __alloc_traits::allocate(__alloc(), __target_capacity+1); #ifndef _LIBCPP_NO_EXCEPTIONS } @@ -3364,7 +3390,7 @@ basic_string<_CharT, _Traits, _Allocator>::__shrink_or_extend(size_type __target #else // _LIBCPP_NO_EXCEPTIONS if (__new_data == nullptr) return; - #endif // _LIBCPP_NO_EXCEPTIONS + #endif // _LIBCPP_NO_EXCEPTIONS } __now_long = true; __was_long = __is_long(); @@ -3543,7 +3569,7 @@ _EnableIf typename basic_string<_CharT, _Traits, _Allocator>::size_type > basic_string<_CharT, _Traits, _Allocator>::find(const _Tp &__t, - size_type __pos) const + size_type __pos) const _NOEXCEPT { __self_view __sv = __t; return __str_find @@ -3601,7 +3627,7 @@ _EnableIf typename basic_string<_CharT, _Traits, _Allocator>::size_type > basic_string<_CharT, _Traits, _Allocator>::rfind(const _Tp& __t, - size_type __pos) const + size_type __pos) const _NOEXCEPT { __self_view __sv = __t; return __str_rfind @@ -3659,7 +3685,7 @@ _EnableIf typename basic_string<_CharT, _Traits, _Allocator>::size_type > basic_string<_CharT, _Traits, _Allocator>::find_first_of(const _Tp& __t, - size_type __pos) const + size_type __pos) const _NOEXCEPT { __self_view __sv = __t; return __str_find_first_of @@ -3717,7 +3743,7 @@ _EnableIf typename basic_string<_CharT, _Traits, _Allocator>::size_type > basic_string<_CharT, _Traits, _Allocator>::find_last_of(const _Tp& __t, - size_type __pos) const + size_type __pos) const _NOEXCEPT { __self_view __sv = __t; return __str_find_last_of @@ -3775,7 +3801,7 @@ _EnableIf typename basic_string<_CharT, _Traits, _Allocator>::size_type > basic_string<_CharT, _Traits, _Allocator>::find_first_not_of(const _Tp& __t, - size_type __pos) const + size_type __pos) const _NOEXCEPT { __self_view __sv = __t; return __str_find_first_not_of @@ -3834,7 +3860,7 @@ _EnableIf typename basic_string<_CharT, _Traits, _Allocator>::size_type > basic_string<_CharT, _Traits, _Allocator>::find_last_not_of(const _Tp& __t, - size_type __pos) const + size_type __pos) const _NOEXCEPT { __self_view __sv = __t; return __str_find_last_not_of @@ -3871,7 +3897,7 @@ _EnableIf __can_be_converted_to_string_view<_CharT, _Traits, _Tp>::value, int > -basic_string<_CharT, _Traits, _Allocator>::compare(const _Tp& __t) const +basic_string<_CharT, _Traits, _Allocator>::compare(const _Tp& __t) const _NOEXCEPT { __self_view __sv = __t; size_t __lhs_sz = size(); @@ -4349,7 +4375,7 @@ operator+(basic_string<_CharT, _Traits, _Allocator>&& __lhs, _CharT __rhs) return _VSTD::move(__lhs); } -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG // swap @@ -4404,7 +4430,7 @@ _LIBCPP_FUNC_VIS wstring to_wstring(double __val); _LIBCPP_FUNC_VIS wstring to_wstring(long double __val); template -_LIBCPP_FUNC_VIS +_LIBCPP_TEMPLATE_DATA_VIS const typename basic_string<_CharT, _Traits, _Allocator>::size_type basic_string<_CharT, _Traits, _Allocator>::npos; @@ -4441,8 +4467,6 @@ basic_istream<_CharT, _Traits>& getline(basic_istream<_CharT, _Traits>& __is, basic_string<_CharT, _Traits, _Allocator>& __str); -#ifndef _LIBCPP_CXX03_LANG - template inline _LIBCPP_INLINE_VISIBILITY basic_istream<_CharT, _Traits>& @@ -4455,8 +4479,6 @@ basic_istream<_CharT, _Traits>& getline(basic_istream<_CharT, _Traits>&& __is, basic_string<_CharT, _Traits, _Allocator>& __str); -#endif // _LIBCPP_CXX03_LANG - #if _LIBCPP_STD_VER > 17 template inline _LIBCPP_INLINE_VISIBILITY @@ -4513,7 +4535,7 @@ basic_string<_CharT, _Traits, _Allocator>::__subscriptable(const const_iterator* return this->data() <= __p && __p < this->data() + this->size(); } -#endif // _LIBCPP_DEBUG_LEVEL == 2 +#endif // _LIBCPP_DEBUG_LEVEL == 2 #if _LIBCPP_STD_VER > 11 // Literal suffixes for basic_string [basic.string.literals] @@ -4533,7 +4555,7 @@ inline namespace literals return basic_string (__str, __len); } -#ifndef _LIBCPP_NO_HAS_CHAR8_T +#ifndef _LIBCPP_HAS_NO_CHAR8_T inline _LIBCPP_INLINE_VISIBILITY basic_string operator "" s(const char8_t *__str, size_t __len) _NOEXCEPT { @@ -4560,4 +4582,4 @@ _LIBCPP_END_NAMESPACE_STD _LIBCPP_POP_MACROS -#endif // _LIBCPP_STRING +#endif // _LIBCPP_STRING diff --git a/lib/libcxx/include/string.h b/lib/libcxx/include/string.h index e09251d06..2d563444b 100644 --- a/lib/libcxx/include/string.h +++ b/lib/libcxx/include/string.h @@ -106,4 +106,4 @@ inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_PREFERRED_OVERLOAD } #endif -#endif // _LIBCPP_STRING_H +#endif // _LIBCPP_STRING_H diff --git a/lib/libcxx/include/string_view b/lib/libcxx/include/string_view index bc0245cf2..2c94cb852 100644 --- a/lib/libcxx/include/string_view +++ b/lib/libcxx/include/string_view @@ -19,6 +19,12 @@ namespace std { template> class basic_string_view; + template + inline constexpr bool ranges::enable_view> = true; + + template + inline constexpr bool ranges::enable_borrowed_range> = true; // C++20 + // 7.9, basic_string_view non-member comparison functions template constexpr bool operator==(basic_string_view x, @@ -48,6 +54,7 @@ namespace std { // basic_string_view typedef names typedef basic_string_view string_view; + typedef basic_string_view u8string_view; // C++20 typedef basic_string_view u16string_view; typedef basic_string_view u32string_view; typedef basic_string_view wstring_view; @@ -76,6 +83,7 @@ namespace std { basic_string_view& operator=(const basic_string_view&) noexcept = default; template constexpr basic_string_view(const charT* str); + basic_string_view(nullptr_t) = delete; // C++2b constexpr basic_string_view(const charT* str, size_type len); // 7.4, basic_string_view iterator support @@ -106,7 +114,7 @@ namespace std { constexpr void remove_suffix(size_type n); constexpr void swap(basic_string_view& s) noexcept; - size_type copy(charT* s, size_type n, size_type pos = 0) const; + size_type copy(charT* s, size_type n, size_type pos = 0) const; // constexpr in C++20 constexpr basic_string_view substr(size_type pos = 0, size_type n = npos) const; constexpr int compare(basic_string_view s) const noexcept; @@ -119,28 +127,28 @@ namespace std { const charT* s, size_type n2) const; constexpr size_type find(basic_string_view s, size_type pos = 0) const noexcept; constexpr size_type find(charT c, size_type pos = 0) const noexcept; - constexpr size_type find(const charT* s, size_type pos, size_type n) const; - constexpr size_type find(const charT* s, size_type pos = 0) const; + constexpr size_type find(const charT* s, size_type pos, size_type n) const noexcept; // noexcept as an extension + constexpr size_type find(const charT* s, size_type pos = 0) const noexcept; // noexcept as an extension constexpr size_type rfind(basic_string_view s, size_type pos = npos) const noexcept; constexpr size_type rfind(charT c, size_type pos = npos) const noexcept; - constexpr size_type rfind(const charT* s, size_type pos, size_type n) const; - constexpr size_type rfind(const charT* s, size_type pos = npos) const; + constexpr size_type rfind(const charT* s, size_type pos, size_type n) const noexcept; // noexcept as an extension + constexpr size_type rfind(const charT* s, size_type pos = npos) const noexcept; // noexcept as an extension constexpr size_type find_first_of(basic_string_view s, size_type pos = 0) const noexcept; constexpr size_type find_first_of(charT c, size_type pos = 0) const noexcept; - constexpr size_type find_first_of(const charT* s, size_type pos, size_type n) const; - constexpr size_type find_first_of(const charT* s, size_type pos = 0) const; + constexpr size_type find_first_of(const charT* s, size_type pos, size_type n) const noexcept; // noexcept as an extension + constexpr size_type find_first_of(const charT* s, size_type pos = 0) const noexcept; // noexcept as an extension constexpr size_type find_last_of(basic_string_view s, size_type pos = npos) const noexcept; constexpr size_type find_last_of(charT c, size_type pos = npos) const noexcept; - constexpr size_type find_last_of(const charT* s, size_type pos, size_type n) const; - constexpr size_type find_last_of(const charT* s, size_type pos = npos) const; + constexpr size_type find_last_of(const charT* s, size_type pos, size_type n) const noexcept; // noexcept as an extension + constexpr size_type find_last_of(const charT* s, size_type pos = npos) const noexcept; // noexcept as an extension constexpr size_type find_first_not_of(basic_string_view s, size_type pos = 0) const noexcept; constexpr size_type find_first_not_of(charT c, size_type pos = 0) const noexcept; - constexpr size_type find_first_not_of(const charT* s, size_type pos, size_type n) const; - constexpr size_type find_first_not_of(const charT* s, size_type pos = 0) const; + constexpr size_type find_first_not_of(const charT* s, size_type pos, size_type n) const noexcept; // noexcept as an extension + constexpr size_type find_first_not_of(const charT* s, size_type pos = 0) const noexcept; // noexcept as an extension constexpr size_type find_last_not_of(basic_string_view s, size_type pos = npos) const noexcept; constexpr size_type find_last_not_of(charT c, size_type pos = npos) const noexcept; - constexpr size_type find_last_not_of(const charT* s, size_type pos, size_type n) const; - constexpr size_type find_last_not_of(const charT* s, size_type pos = npos) const; + constexpr size_type find_last_not_of(const charT* s, size_type pos, size_type n) const noexcept; // noexcept as an extension + constexpr size_type find_last_not_of(const charT* s, size_type pos = npos) const noexcept; // noexcept as an extension constexpr bool starts_with(basic_string_view s) const noexcept; // C++20 constexpr bool starts_with(charT c) const noexcept; // C++20 @@ -161,12 +169,14 @@ namespace std { // 7.11, Hash support template struct hash; template <> struct hash; + template <> struct hash; // C++20 template <> struct hash; template <> struct hash; template <> struct hash; constexpr basic_string_view operator "" sv( const char *str, size_t len ) noexcept; constexpr basic_string_view operator "" sv( const wchar_t *str, size_t len ) noexcept; + constexpr basic_string_view operator "" sv( const char8_t *str, size_t len ) noexcept; // C++20 constexpr basic_string_view operator "" sv( const char16_t *str, size_t len ) noexcept; constexpr basic_string_view operator "" sv( const char32_t *str, size_t len ) noexcept; @@ -176,14 +186,17 @@ namespace std { */ #include <__config> +#include <__debug> +#include <__ranges/enable_borrowed_range.h> +#include <__ranges/enable_view.h> #include <__string> -#include #include +#include +#include #include #include #include #include -#include <__debug> #if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) #pragma GCC system_header @@ -199,7 +212,7 @@ template > class _LIBCPP_TEMPLATE_VIS basic_string_view; typedef basic_string_view string_view; -#ifndef _LIBCPP_NO_HAS_CHAR8_T +#ifndef _LIBCPP_HAS_NO_CHAR8_T typedef basic_string_view u8string_view; #endif typedef basic_string_view u16string_view; @@ -209,7 +222,7 @@ typedef basic_string_view wstring_view; template class _LIBCPP_PREFERRED_NAME(string_view) -#ifndef _LIBCPP_NO_HAS_CHAR8_T +#ifndef _LIBCPP_HAS_NO_CHAR8_T _LIBCPP_PREFERRED_NAME(u8string_view) #endif _LIBCPP_PREFERRED_NAME(u16string_view) @@ -261,6 +274,10 @@ public: basic_string_view(const _CharT* __s) : __data(__s), __size(_VSTD::__char_traits_length_checked<_Traits>(__s)) {} +#if _LIBCPP_STD_VER > 20 + basic_string_view(nullptr_t) = delete; +#endif + // [string.view.iterators], iterators _LIBCPP_CONSTEXPR _LIBCPP_INLINE_VISIBILITY const_iterator begin() const _NOEXCEPT { return cbegin(); } @@ -356,7 +373,7 @@ public: __other.__size = __sz; } - _LIBCPP_INLINE_VISIBILITY + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 size_type copy(_CharT* __s, size_type __n, size_type __pos = 0) const { if (__pos > size()) @@ -431,7 +448,7 @@ public: } _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY - size_type find(const _CharT* __s, size_type __pos, size_type __n) const + size_type find(const _CharT* __s, size_type __pos, size_type __n) const _NOEXCEPT { _LIBCPP_ASSERT(__n == 0 || __s != nullptr, "string_view::find(): received nullptr"); return __str_find @@ -439,7 +456,7 @@ public: } _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY - size_type find(const _CharT* __s, size_type __pos = 0) const + size_type find(const _CharT* __s, size_type __pos = 0) const _NOEXCEPT { _LIBCPP_ASSERT(__s != nullptr, "string_view::find(): received nullptr"); return __str_find @@ -463,7 +480,7 @@ public: } _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY - size_type rfind(const _CharT* __s, size_type __pos, size_type __n) const + size_type rfind(const _CharT* __s, size_type __pos, size_type __n) const _NOEXCEPT { _LIBCPP_ASSERT(__n == 0 || __s != nullptr, "string_view::rfind(): received nullptr"); return __str_rfind @@ -471,7 +488,7 @@ public: } _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY - size_type rfind(const _CharT* __s, size_type __pos=npos) const + size_type rfind(const _CharT* __s, size_type __pos=npos) const _NOEXCEPT { _LIBCPP_ASSERT(__s != nullptr, "string_view::rfind(): received nullptr"); return __str_rfind @@ -492,7 +509,7 @@ public: { return find(__c, __pos); } _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY - size_type find_first_of(const _CharT* __s, size_type __pos, size_type __n) const + size_type find_first_of(const _CharT* __s, size_type __pos, size_type __n) const _NOEXCEPT { _LIBCPP_ASSERT(__n == 0 || __s != nullptr, "string_view::find_first_of(): received nullptr"); return __str_find_first_of @@ -500,7 +517,7 @@ public: } _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY - size_type find_first_of(const _CharT* __s, size_type __pos=0) const + size_type find_first_of(const _CharT* __s, size_type __pos=0) const _NOEXCEPT { _LIBCPP_ASSERT(__s != nullptr, "string_view::find_first_of(): received nullptr"); return __str_find_first_of @@ -521,7 +538,7 @@ public: { return rfind(__c, __pos); } _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY - size_type find_last_of(const _CharT* __s, size_type __pos, size_type __n) const + size_type find_last_of(const _CharT* __s, size_type __pos, size_type __n) const _NOEXCEPT { _LIBCPP_ASSERT(__n == 0 || __s != nullptr, "string_view::find_last_of(): received nullptr"); return __str_find_last_of @@ -529,7 +546,7 @@ public: } _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY - size_type find_last_of(const _CharT* __s, size_type __pos=npos) const + size_type find_last_of(const _CharT* __s, size_type __pos=npos) const _NOEXCEPT { _LIBCPP_ASSERT(__s != nullptr, "string_view::find_last_of(): received nullptr"); return __str_find_last_of @@ -553,7 +570,7 @@ public: } _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY - size_type find_first_not_of(const _CharT* __s, size_type __pos, size_type __n) const + size_type find_first_not_of(const _CharT* __s, size_type __pos, size_type __n) const _NOEXCEPT { _LIBCPP_ASSERT(__n == 0 || __s != nullptr, "string_view::find_first_not_of(): received nullptr"); return __str_find_first_not_of @@ -561,7 +578,7 @@ public: } _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY - size_type find_first_not_of(const _CharT* __s, size_type __pos=0) const + size_type find_first_not_of(const _CharT* __s, size_type __pos=0) const _NOEXCEPT { _LIBCPP_ASSERT(__s != nullptr, "string_view::find_first_not_of(): received nullptr"); return __str_find_first_not_of @@ -585,7 +602,7 @@ public: } _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY - size_type find_last_not_of(const _CharT* __s, size_type __pos, size_type __n) const + size_type find_last_not_of(const _CharT* __s, size_type __pos, size_type __n) const _NOEXCEPT { _LIBCPP_ASSERT(__n == 0 || __s != nullptr, "string_view::find_last_not_of(): received nullptr"); return __str_find_last_not_of @@ -593,7 +610,7 @@ public: } _LIBCPP_CONSTEXPR_AFTER_CXX11 _LIBCPP_INLINE_VISIBILITY - size_type find_last_not_of(const _CharT* __s, size_type __pos=npos) const + size_type find_last_not_of(const _CharT* __s, size_type __pos=npos) const _NOEXCEPT { _LIBCPP_ASSERT(__s != nullptr, "string_view::find_last_not_of(): received nullptr"); return __str_find_last_not_of @@ -645,6 +662,13 @@ private: size_type __size; }; +#if _LIBCPP_STD_VER > 17 && !defined(_LIBCPP_HAS_NO_RANGES) +template +inline constexpr bool ranges::enable_view> = true; + +template +inline constexpr bool ranges::enable_borrowed_range > = true; +#endif // _LIBCPP_STD_VER > 17 && !defined(_LIBCPP_HAS_NO_RANGES) // [string.view.comparison] // operator == @@ -842,7 +866,7 @@ inline namespace literals return basic_string_view (__str, __len); } -#ifndef _LIBCPP_NO_HAS_CHAR8_T +#ifndef _LIBCPP_HAS_NO_CHAR8_T inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR basic_string_view operator "" sv(const char8_t *__str, size_t __len) _NOEXCEPT { diff --git a/lib/libcxx/include/strstream b/lib/libcxx/include/strstream index 0062777cd..ca837aef6 100644 --- a/lib/libcxx/include/strstream +++ b/lib/libcxx/include/strstream @@ -130,8 +130,8 @@ private: */ #include <__config> -#include #include +#include #if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) #pragma GCC system_header @@ -163,7 +163,7 @@ public: strstreambuf(strstreambuf&& __rhs); _LIBCPP_INLINE_VISIBILITY strstreambuf& operator=(strstreambuf&& __rhs); -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG virtual ~strstreambuf(); @@ -233,7 +233,7 @@ strstreambuf::operator=(strstreambuf&& __rhs) return *this; } -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG class _LIBCPP_TYPE_VIS istrstream : public istream @@ -268,7 +268,7 @@ public: __sb_ = _VSTD::move(__rhs.__sb_); return *this; } -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG virtual ~istrstream(); @@ -317,7 +317,7 @@ public: __sb_ = _VSTD::move(__rhs.__sb_); return *this; } -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG virtual ~ostrstream(); @@ -377,7 +377,7 @@ public: __sb_ = _VSTD::move(__rhs.__sb_); return *this; } -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG virtual ~strstream(); @@ -404,4 +404,4 @@ private: _LIBCPP_END_NAMESPACE_STD -#endif // _LIBCPP_STRSTREAM +#endif // _LIBCPP_STRSTREAM diff --git a/lib/libcxx/include/system_error b/lib/libcxx/include/system_error index b714e1d42..aab976811 100644 --- a/lib/libcxx/include/system_error +++ b/lib/libcxx/include/system_error @@ -142,11 +142,14 @@ template <> struct hash; */ +#include <__config> #include <__errc> -#include -#include +#include <__functional/unary_function.h> #include <__functional_base> +#include +#include #include +#include #if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) #pragma GCC system_header @@ -289,8 +292,7 @@ public: string message() const; _LIBCPP_INLINE_VISIBILITY - _LIBCPP_EXPLICIT - operator bool() const _NOEXCEPT {return __val_ != 0;} + explicit operator bool() const _NOEXCEPT {return __val_ != 0;} }; inline _LIBCPP_INLINE_VISIBILITY @@ -366,8 +368,7 @@ public: string message() const; _LIBCPP_INLINE_VISIBILITY - _LIBCPP_EXPLICIT - operator bool() const _NOEXCEPT {return __val_ != 0;} + explicit operator bool() const _NOEXCEPT {return __val_ != 0;} }; inline _LIBCPP_INLINE_VISIBILITY @@ -484,4 +485,4 @@ void __throw_system_error(int ev, const char* what_arg); _LIBCPP_END_NAMESPACE_STD -#endif // _LIBCPP_SYSTEM_ERROR +#endif // _LIBCPP_SYSTEM_ERROR diff --git a/lib/libcxx/include/tgmath.h b/lib/libcxx/include/tgmath.h index 2dc1f771d..ba9396e6a 100644 --- a/lib/libcxx/include/tgmath.h +++ b/lib/libcxx/include/tgmath.h @@ -31,6 +31,6 @@ #include_next -#endif // __cplusplus +#endif // __cplusplus -#endif // _LIBCPP_TGMATH_H +#endif // _LIBCPP_TGMATH_H diff --git a/lib/libcxx/include/thread b/lib/libcxx/include/thread index 34e0c2a23..acfc20bce 100644 --- a/lib/libcxx/include/thread +++ b/lib/libcxx/include/thread @@ -83,20 +83,20 @@ void sleep_for(const chrono::duration& rel_time); */ #include <__config> -#include +#include <__debug> #include <__functional_base> -#include +#include <__mutex_base> +#include <__threading_support> +#include <__utility/__decay_copy.h> +#include <__utility/forward.h> +#include #include #include +#include #include #include -#include -#include <__mutex_base> -#ifndef _LIBCPP_CXX03_LANG #include -#endif -#include <__threading_support> -#include <__debug> +#include #if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER) #pragma GCC system_header @@ -346,7 +346,7 @@ thread::thread(_Fp __f) __throw_system_error(__ec, "thread constructor failed"); } -#endif // _LIBCPP_CXX03_LANG +#endif // _LIBCPP_CXX03_LANG inline _LIBCPP_INLINE_VISIBILITY void swap(thread& __x, thread& __y) _NOEXCEPT {__x.swap(__y);} @@ -362,12 +362,11 @@ sleep_for(const chrono::duration<_Rep, _Period>& __d) { if (__d > chrono::duration<_Rep, _Period>::zero()) { -#if defined(_LIBCPP_COMPILER_GCC) && (__powerpc__ || __POWERPC__) - // GCC's long double const folding is incomplete for IBM128 long doubles. - _LIBCPP_CONSTEXPR chrono::duration _Max = chrono::duration(ULLONG_MAX/1000000000ULL) ; -#else - _LIBCPP_CONSTEXPR chrono::duration _Max = chrono::nanoseconds::max(); -#endif + // The standard guarantees a 64bit signed integer resolution for nanoseconds, + // so use INT64_MAX / 1e9 as cut-off point. Use a constant to avoid + // and issues with long double folding on PowerPC with GCC. + _LIBCPP_CONSTEXPR chrono::duration _Max = + chrono::duration(9223372036.0L); chrono::nanoseconds __ns; if (__d < _Max) { @@ -411,4 +410,4 @@ _LIBCPP_END_NAMESPACE_STD _LIBCPP_POP_MACROS -#endif // _LIBCPP_THREAD +#endif // _LIBCPP_THREAD diff --git a/lib/libcxx/include/tuple b/lib/libcxx/include/tuple index c3c7db5ff..e1019ef99 100644 --- a/lib/libcxx/include/tuple +++ b/lib/libcxx/include/tuple @@ -38,35 +38,39 @@ public: template tuple(allocator_arg_t, const Alloc& a); template - explicit(see-below) tuple(allocator_arg_t, const Alloc& a, const T&...); + explicit(see-below) tuple(allocator_arg_t, const Alloc& a, const T&...); // constexpr in C++20 template - explicit(see-below) tuple(allocator_arg_t, const Alloc& a, U&&...); + explicit(see-below) tuple(allocator_arg_t, const Alloc& a, U&&...); // constexpr in C++20 template - tuple(allocator_arg_t, const Alloc& a, const tuple&); + tuple(allocator_arg_t, const Alloc& a, const tuple&); // constexpr in C++20 template - tuple(allocator_arg_t, const Alloc& a, tuple&&); + tuple(allocator_arg_t, const Alloc& a, tuple&&); // constexpr in C++20 template - explicit(see-below) tuple(allocator_arg_t, const Alloc& a, const tuple&); + explicit(see-below) tuple(allocator_arg_t, const Alloc& a, const tuple&); // constexpr in C++20 template - explicit(see-below) tuple(allocator_arg_t, const Alloc& a, tuple&&); + explicit(see-below) tuple(allocator_arg_t, const Alloc& a, tuple&&); // constexpr in C++20 template - explicit(see-below) tuple(allocator_arg_t, const Alloc& a, const pair&); + explicit(see-below) tuple(allocator_arg_t, const Alloc& a, const pair&); // constexpr in C++20 template - explicit(see-below) tuple(allocator_arg_t, const Alloc& a, pair&&); + explicit(see-below) tuple(allocator_arg_t, const Alloc& a, pair&&); // constexpr in C++20 - tuple& operator=(const tuple&); - tuple& - operator=(tuple&&) noexcept(AND(is_nothrow_move_assignable::value ...)); + tuple& operator=(const tuple&); // constexpr in C++20 + tuple& operator=(tuple&&) noexcept(is_nothrow_move_assignable_v && ...); // constexpr in C++20 template - tuple& operator=(const tuple&); + tuple& operator=(const tuple&); // constexpr in C++20 template - tuple& operator=(tuple&&); + tuple& operator=(tuple&&); // constexpr in C++20 template - tuple& operator=(const pair&); // iff sizeof...(T) == 2 + tuple& operator=(const pair&); // iff sizeof...(T) == 2 // constexpr in C++20 template - tuple& operator=(pair&&); // iff sizeof...(T) == 2 + tuple& operator=(pair&&); // iff sizeof...(T) == 2 // constexpr in C++20 - void swap(tuple&) noexcept(AND(swap(declval(), declval())...)); + template + tuple& operator=(array const&) // iff sizeof...(T) == N, EXTENSION + template + tuple& operator=(array&&) // iff sizeof...(T) == N, EXTENSION + + void swap(tuple&) noexcept(AND(swap(declval(), declval())...)); // constexpr in C++20 }; template @@ -146,10 +150,16 @@ template */ #include <__config> +#include <__functional/unwrap_ref.h> +#include <__functional_base> +#include <__memory/allocator_arg_t.h> +#include <__memory/uses_allocator.h> #include <__tuple> +#include <__utility/forward.h> +#include <__utility/move.h> +#include #include #include -#include <__functional_base> #include #include @@ -170,7 +180,7 @@ template -inline _LIBCPP_INLINE_VISIBILITY +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 void swap(__tuple_leaf<_Ip, _Hp, _Ep>& __x, __tuple_leaf<_Ip, _Hp, _Ep>& __y) _NOEXCEPT_(__is_nothrow_swappable<_Hp>::value) { @@ -191,29 +201,30 @@ class __tuple_leaf #endif } + _LIBCPP_CONSTEXPR_AFTER_CXX11 __tuple_leaf& operator=(const __tuple_leaf&); public: - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR __tuple_leaf() + _LIBCPP_INLINE_VISIBILITY constexpr __tuple_leaf() _NOEXCEPT_(is_nothrow_default_constructible<_Hp>::value) : __value_() {static_assert(!is_reference<_Hp>::value, "Attempted to default construct a reference element in a tuple");} template - _LIBCPP_INLINE_VISIBILITY + _LIBCPP_INLINE_VISIBILITY constexpr __tuple_leaf(integral_constant, const _Alloc&) : __value_() {static_assert(!is_reference<_Hp>::value, "Attempted to default construct a reference element in a tuple");} template - _LIBCPP_INLINE_VISIBILITY + _LIBCPP_INLINE_VISIBILITY constexpr __tuple_leaf(integral_constant, const _Alloc& __a) : __value_(allocator_arg_t(), __a) {static_assert(!is_reference<_Hp>::value, "Attempted to default construct a reference element in a tuple");} template - _LIBCPP_INLINE_VISIBILITY + _LIBCPP_INLINE_VISIBILITY constexpr __tuple_leaf(integral_constant, const _Alloc& __a) : __value_(__a) {static_assert(!is_reference<_Hp>::value, @@ -234,21 +245,21 @@ public: "Attempted construction of reference element binds to a temporary whose lifetime has ended");} template - _LIBCPP_INLINE_VISIBILITY + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 explicit __tuple_leaf(integral_constant, const _Alloc&, _Tp&& __t) : __value_(_VSTD::forward<_Tp>(__t)) {static_assert(__can_bind_reference<_Tp&&>(), "Attempted construction of reference element binds to a temporary whose lifetime has ended");} template - _LIBCPP_INLINE_VISIBILITY + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 explicit __tuple_leaf(integral_constant, const _Alloc& __a, _Tp&& __t) : __value_(allocator_arg_t(), __a, _VSTD::forward<_Tp>(__t)) {static_assert(!is_reference<_Hp>::value, "Attempted to uses-allocator construct a reference element in a tuple");} template - _LIBCPP_INLINE_VISIBILITY + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 explicit __tuple_leaf(integral_constant, const _Alloc& __a, _Tp&& __t) : __value_(_VSTD::forward<_Tp>(__t), __a) {static_assert(!is_reference<_Hp>::value, @@ -257,16 +268,7 @@ public: __tuple_leaf(const __tuple_leaf& __t) = default; __tuple_leaf(__tuple_leaf&& __t) = default; - template - _LIBCPP_INLINE_VISIBILITY - __tuple_leaf& - operator=(_Tp&& __t) _NOEXCEPT_((is_nothrow_assignable<_Hp&, _Tp>::value)) - { - __value_ = _VSTD::forward<_Tp>(__t); - return *this; - } - - _LIBCPP_INLINE_VISIBILITY + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 int swap(__tuple_leaf& __t) _NOEXCEPT_(__is_nothrow_swappable<__tuple_leaf>::value) { _VSTD::swap(*this, __t); @@ -281,23 +283,23 @@ template class __tuple_leaf<_Ip, _Hp, true> : private _Hp { - + _LIBCPP_CONSTEXPR_AFTER_CXX11 __tuple_leaf& operator=(const __tuple_leaf&); public: - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR __tuple_leaf() + _LIBCPP_INLINE_VISIBILITY constexpr __tuple_leaf() _NOEXCEPT_(is_nothrow_default_constructible<_Hp>::value) {} template - _LIBCPP_INLINE_VISIBILITY + _LIBCPP_INLINE_VISIBILITY constexpr __tuple_leaf(integral_constant, const _Alloc&) {} template - _LIBCPP_INLINE_VISIBILITY + _LIBCPP_INLINE_VISIBILITY constexpr __tuple_leaf(integral_constant, const _Alloc& __a) : _Hp(allocator_arg_t(), __a) {} template - _LIBCPP_INLINE_VISIBILITY + _LIBCPP_INLINE_VISIBILITY constexpr __tuple_leaf(integral_constant, const _Alloc& __a) : _Hp(__a) {} @@ -314,33 +316,24 @@ public: : _Hp(_VSTD::forward<_Tp>(__t)) {} template - _LIBCPP_INLINE_VISIBILITY + _LIBCPP_INLINE_VISIBILITY constexpr explicit __tuple_leaf(integral_constant, const _Alloc&, _Tp&& __t) : _Hp(_VSTD::forward<_Tp>(__t)) {} template - _LIBCPP_INLINE_VISIBILITY + _LIBCPP_INLINE_VISIBILITY constexpr explicit __tuple_leaf(integral_constant, const _Alloc& __a, _Tp&& __t) : _Hp(allocator_arg_t(), __a, _VSTD::forward<_Tp>(__t)) {} template - _LIBCPP_INLINE_VISIBILITY + _LIBCPP_INLINE_VISIBILITY constexpr explicit __tuple_leaf(integral_constant, const _Alloc& __a, _Tp&& __t) : _Hp(_VSTD::forward<_Tp>(__t), __a) {} __tuple_leaf(__tuple_leaf const &) = default; __tuple_leaf(__tuple_leaf &&) = default; - template - _LIBCPP_INLINE_VISIBILITY - __tuple_leaf& - operator=(_Tp&& __t) _NOEXCEPT_((is_nothrow_assignable<_Hp&, _Tp>::value)) - { - _Hp::operator=(_VSTD::forward<_Tp>(__t)); - return *this; - } - - _LIBCPP_INLINE_VISIBILITY + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 int swap(__tuple_leaf& __t) _NOEXCEPT_(__is_nothrow_swappable<__tuple_leaf>::value) { @@ -353,7 +346,7 @@ public: }; template -_LIBCPP_INLINE_VISIBILITY +_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 void __swallow(_Tp&&...) _NOEXCEPT {} template @@ -373,7 +366,7 @@ struct _LIBCPP_DECLSPEC_EMPTY_BASES __tuple_impl<__tuple_indices<_Indx...>, _Tp. : public __tuple_leaf<_Indx, _Tp>... { _LIBCPP_INLINE_VISIBILITY - _LIBCPP_CONSTEXPR __tuple_impl() + constexpr __tuple_impl() _NOEXCEPT_(__all::value...>::value) {} template , _Tp. template - _LIBCPP_INLINE_VISIBILITY + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 explicit __tuple_impl(allocator_arg_t, const _Alloc& __a, __tuple_indices<_Uf...>, __tuple_types<_Tf...>, @@ -421,7 +414,7 @@ struct _LIBCPP_DECLSPEC_EMPTY_BASES __tuple_impl<__tuple_indices<_Indx...>, _Tp. __tuple_constructible<_Tuple, tuple<_Tp...> >::value >::type > - _LIBCPP_INLINE_VISIBILITY + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 __tuple_impl(allocator_arg_t, const _Alloc& __a, _Tuple&& __t) : __tuple_leaf<_Indx, _Tp>(__uses_alloc_ctor<_Tp, _Alloc, typename tuple_element<_Indx, typename __make_tuple_types<_Tuple>::type>::type>(), __a, @@ -429,49 +422,30 @@ struct _LIBCPP_DECLSPEC_EMPTY_BASES __tuple_impl<__tuple_indices<_Indx...>, _Tp. typename __make_tuple_types<_Tuple>::type>::type>(_VSTD::get<_Indx>(__t)))... {} - template - _LIBCPP_INLINE_VISIBILITY - typename enable_if - < - __tuple_assignable<_Tuple, tuple<_Tp...> >::value, - __tuple_impl& - >::type - operator=(_Tuple&& __t) _NOEXCEPT_((__all::type>::type>::value...>::value)) - { - __swallow(__tuple_leaf<_Indx, _Tp>::operator=(_VSTD::forward::type>::type>(_VSTD::get<_Indx>(__t)))...); - return *this; - } - __tuple_impl(const __tuple_impl&) = default; __tuple_impl(__tuple_impl&&) = default; - _LIBCPP_INLINE_VISIBILITY - __tuple_impl& - operator=(const __tuple_impl& __t) _NOEXCEPT_((__all::value...>::value)) - { - __swallow(__tuple_leaf<_Indx, _Tp>::operator=(static_cast&>(__t).get())...); - return *this; - } - - _LIBCPP_INLINE_VISIBILITY - __tuple_impl& - operator=(__tuple_impl&& __t) _NOEXCEPT_((__all::value...>::value)) - { - __swallow(__tuple_leaf<_Indx, _Tp>::operator=(_VSTD::forward<_Tp>(static_cast<__tuple_leaf<_Indx, _Tp>&>(__t).get()))...); - return *this; - } - - _LIBCPP_INLINE_VISIBILITY + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 void swap(__tuple_impl& __t) _NOEXCEPT_(__all<__is_nothrow_swappable<_Tp>::value...>::value) { - __swallow(__tuple_leaf<_Indx, _Tp>::swap(static_cast<__tuple_leaf<_Indx, _Tp>&>(__t))...); + _VSTD::__swallow(__tuple_leaf<_Indx, _Tp>::swap(static_cast<__tuple_leaf<_Indx, _Tp>&>(__t))...); } }; +template +_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 +void __memberwise_copy_assign(_Dest& __dest, _Source const& __source, __tuple_indices<_Np...>) { + _VSTD::__swallow(((_VSTD::get<_Np>(__dest) = _VSTD::get<_Np>(__source)), void(), 0)...); +} +template +_LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 +void __memberwise_forward_assign(_Dest& __dest, _Source&& __source, __tuple_types<_Up...>, __tuple_indices<_Np...>) { + _VSTD::__swallow((( + _VSTD::get<_Np>(__dest) = _VSTD::forward<_Up>(_VSTD::get<_Np>(__source)) + ), void(), 0)...); +} template class _LIBCPP_TEMPLATE_VIS tuple @@ -480,165 +454,6 @@ class _LIBCPP_TEMPLATE_VIS tuple _BaseT __base_; -#if defined(_LIBCPP_ENABLE_TUPLE_IMPLICIT_REDUCED_ARITY_EXTENSION) - static constexpr bool _EnableImplicitReducedArityExtension = true; -#else - static constexpr bool _EnableImplicitReducedArityExtension = false; -#endif - - template - struct _PackExpandsToThisTuple : false_type {}; - - template - struct _PackExpandsToThisTuple<_Arg> - : is_same::type, tuple> {}; - - template - struct _CheckArgsConstructor : __check_tuple_constructor_fail {}; - - template - struct _CheckArgsConstructor - { - template - static constexpr bool __enable_implicit_default() { - return __all<__is_implicitly_default_constructible<_Tp>::value... >::value; - } - - template - static constexpr bool __enable_explicit_default() { - return - __all::value...>::value && - !__enable_implicit_default< >(); - } - - - template - static constexpr bool __enable_explicit() { - return - __tuple_constructible< - tuple<_Args...>, - typename __make_tuple_types::type - >::value && - !__tuple_convertible< - tuple<_Args...>, - typename __make_tuple_types::type - >::value && - __all_default_constructible< - typename __make_tuple_types::type - >::value; - } - - template - static constexpr bool __enable_implicit() { - return - __tuple_constructible< - tuple<_Args...>, - typename __make_tuple_types::type - >::value && - __tuple_convertible< - tuple<_Args...>, - typename __make_tuple_types::type - >::value && - __all_default_constructible< - typename __make_tuple_types::type - >::value; - } - }; - - template - struct _CheckTupleLikeConstructor : __check_tuple_constructor_fail {}; - - template - struct _CheckTupleLikeConstructor - { - template - static constexpr bool __enable_implicit() { - return __tuple_constructible<_Tuple, tuple>::value - && __tuple_convertible<_Tuple, tuple>::value; - } - - template - static constexpr bool __enable_explicit() { - return __tuple_constructible<_Tuple, tuple>::value - && !__tuple_convertible<_Tuple, tuple>::value; - } - }; - - template - struct _CheckTupleLikeConstructor - { - // This trait is used to disable the tuple-like constructor when - // the UTypes... constructor should be selected instead. - // See LWG issue #2549. - template - using _PreferTupleLikeConstructor = _Or< - // Don't attempt the two checks below if the tuple we are given - // has the same type as this tuple. - _IsSame<__uncvref_t<_Tuple>, tuple>, - _Lazy<_And, - _Not>, - _Not> - > - >; - - template - static constexpr bool __enable_implicit() { - return _And< - __tuple_constructible<_Tuple, tuple>, - __tuple_convertible<_Tuple, tuple>, - _PreferTupleLikeConstructor<_Tuple> - >::value; - } - - template - static constexpr bool __enable_explicit() { - return _And< - __tuple_constructible<_Tuple, tuple>, - _PreferTupleLikeConstructor<_Tuple>, - _Not<__tuple_convertible<_Tuple, tuple>> - >::value; - } - }; - - template - using _EnableImplicitTupleLikeConstructor = _EnableIf< - _CheckTupleLikeConstructor< - __tuple_like_with_size<_Tuple, sizeof...(_Tp)>::value - && !_PackExpandsToThisTuple<_Tuple>::value - && (!is_lvalue_reference<_Tuple>::value || !_DisableIfLValue) - >::template __enable_implicit<_Tuple>(), - bool - >; - - template - using _EnableExplicitTupleLikeConstructor = _EnableIf< - _CheckTupleLikeConstructor< - __tuple_like_with_size<_Tuple, sizeof...(_Tp)>::value - && !_PackExpandsToThisTuple<_Tuple>::value - && (!is_lvalue_reference<_Tuple>::value || !_DisableIfLValue) - >::template __enable_explicit<_Tuple>(), - bool - >; template friend _LIBCPP_CONSTEXPR_AFTER_CXX11 typename tuple_element<_Jp, tuple<_Up...> >::type& get(tuple<_Up...>&) _NOEXCEPT; template friend _LIBCPP_CONSTEXPR_AFTER_CXX11 @@ -648,57 +463,69 @@ class _LIBCPP_TEMPLATE_VIS tuple template friend _LIBCPP_CONSTEXPR_AFTER_CXX11 const typename tuple_element<_Jp, tuple<_Up...> >::type&& get(const tuple<_Up...>&&) _NOEXCEPT; public: - - template ::__enable_implicit_default() - , void*> = nullptr> + // [tuple.cnstr] + + // tuple() constructors (including allocator_arg_t variants) + template class _IsImpDefault = __is_implicitly_default_constructible, _EnableIf< + _And< + _IsImpDefault<_Tp>... // explicit check + >::value + , int> = 0> _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR tuple() - _NOEXCEPT_(__all::value...>::value) {} - - template ::__enable_explicit_default() - , void*> = nullptr> - explicit _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR - tuple() - _NOEXCEPT_(__all::value...>::value) {} - - tuple(tuple const&) = default; - tuple(tuple&&) = default; - - template ::value >::__enable_implicit_default() - , void*> = nullptr - > - _LIBCPP_INLINE_VISIBILITY - tuple(_AllocArgT, _Alloc const& __a) + _NOEXCEPT_(_And...>::value) + { } + + template class _IsImpDefault = __is_implicitly_default_constructible, + template class _IsDefault = is_default_constructible, _EnableIf< + _And< + _IsDefault<_Tp>..., + _Not<_Lazy<_And, _IsImpDefault<_Tp>...> > // explicit check + >::value + , int> = 0> + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR + explicit tuple() + _NOEXCEPT_(_And...>::value) + { } + + template class _IsImpDefault = __is_implicitly_default_constructible, _EnableIf< + _And< + _IsImpDefault<_Tp>... // explicit check + >::value + , int> = 0> + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 + tuple(allocator_arg_t, _Alloc const& __a) : __base_(allocator_arg_t(), __a, __tuple_indices<>(), __tuple_types<>(), typename __make_tuple_indices::type(), __tuple_types<_Tp...>()) {} - template ::value>::__enable_explicit_default() - , void*> = nullptr - > - explicit _LIBCPP_INLINE_VISIBILITY - tuple(_AllocArgT, _Alloc const& __a) + template class _IsImpDefault = __is_implicitly_default_constructible, + template class _IsDefault = is_default_constructible, _EnableIf< + _And< + _IsDefault<_Tp>..., + _Not<_Lazy<_And, _IsImpDefault<_Tp>...> > // explicit check + >::value + , int> = 0> + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 + explicit tuple(allocator_arg_t, _Alloc const& __a) : __base_(allocator_arg_t(), __a, __tuple_indices<>(), __tuple_types<>(), typename __make_tuple_indices::type(), __tuple_types<_Tp...>()) {} - template ::template __enable_implicit<_Tp const&...>(), - bool - >::type = false - > + // tuple(const T&...) constructors (including allocator_arg_t variants) + template class _And = _And, _EnableIf< + _And< + _BoolConstant= 1>, + is_copy_constructible<_Tp>..., + is_convertible... // explicit check + >::value + , int> = 0> _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 - tuple(const _Tp& ... __t) _NOEXCEPT_((__all::value...>::value)) + tuple(const _Tp& ... __t) + _NOEXCEPT_(_And...>::value) : __base_(typename __make_tuple_indices::type(), typename __make_tuple_types::type(), typename __make_tuple_indices<0>::type(), @@ -706,17 +533,16 @@ public: __t... ) {} - template ::template __enable_explicit<_Tp const&...>(), - bool - >::type = false - > + template class _And = _And, _EnableIf< + _And< + _BoolConstant= 1>, + is_copy_constructible<_Tp>..., + _Not<_Lazy<_And, is_convertible...> > // explicit check + >::value + , int> = 0> _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 - explicit tuple(const _Tp& ... __t) _NOEXCEPT_((__all::value...>::value)) + explicit tuple(const _Tp& ... __t) + _NOEXCEPT_(_And...>::value) : __base_(typename __make_tuple_indices::type(), typename __make_tuple_types::type(), typename __make_tuple_indices<0>::type(), @@ -724,17 +550,15 @@ public: __t... ) {} - template ::template __enable_implicit<_Tp const&...>(), - bool - >::type = false - > - _LIBCPP_INLINE_VISIBILITY - tuple(allocator_arg_t, const _Alloc& __a, const _Tp& ... __t) + template class _And = _And, _EnableIf< + _And< + _BoolConstant= 1>, + is_copy_constructible<_Tp>..., + is_convertible... // explicit check + >::value + , int> = 0> + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 + tuple(allocator_arg_t, const _Alloc& __a, const _Tp& ... __t) : __base_(allocator_arg_t(), __a, typename __make_tuple_indices::type(), typename __make_tuple_types::type(), @@ -743,18 +567,15 @@ public: __t... ) {} - template ::template __enable_explicit<_Tp const&...>(), - bool - >::type = false - > - _LIBCPP_INLINE_VISIBILITY - explicit - tuple(allocator_arg_t, const _Alloc& __a, const _Tp& ... __t) + template class _And = _And, _EnableIf< + _And< + _BoolConstant= 1>, + is_copy_constructible<_Tp>..., + _Not<_Lazy<_And, is_convertible...> > // explicit check + >::value + , int> = 0> + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 + explicit tuple(allocator_arg_t, const _Alloc& __a, const _Tp& ... __t) : __base_(allocator_arg_t(), __a, typename __make_tuple_indices::type(), typename __make_tuple_types::type(), @@ -763,193 +584,493 @@ public: __t... ) {} - template ::value, - typename enable_if - < - _CheckArgsConstructor< - sizeof...(_Up) == sizeof...(_Tp) - && !_PackIsTuple - >::template __enable_implicit<_Up...>() || - _CheckArgsConstructor< - _EnableImplicitReducedArityExtension - && sizeof...(_Up) < sizeof...(_Tp) - && !_PackIsTuple - >::template __enable_implicit<_Up...>(), - bool - >::type = false - > - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 - tuple(_Up&&... __u) - _NOEXCEPT_(( - is_nothrow_constructible<_BaseT, - typename __make_tuple_indices::type, - typename __make_tuple_types::type, - typename __make_tuple_indices::type, - typename __make_tuple_types::type, - _Up... - >::value - )) - : __base_(typename __make_tuple_indices::type(), + // tuple(U&& ...) constructors (including allocator_arg_t variants) + template struct _IsThisTuple : false_type { }; + template struct _IsThisTuple<_Up> : is_same<__uncvref_t<_Up>, tuple> { }; + + template + struct _EnableUTypesCtor : _And< + _BoolConstant= 1>, + _Not<_IsThisTuple<_Up...> >, // extension to allow mis-behaved user constructors + is_constructible<_Tp, _Up>... + > { }; + + template , + _EnableUTypesCtor<_Up...>, + is_convertible<_Up, _Tp>... // explicit check + >::value + , int> = 0> + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 + tuple(_Up&&... __u) + _NOEXCEPT_((_And...>::value)) + : __base_(typename __make_tuple_indices::type(), typename __make_tuple_types::type(), typename __make_tuple_indices::type(), typename __make_tuple_types::type(), _VSTD::forward<_Up>(__u)...) {} - template ::value - >::template __enable_explicit<_Up...>() || - _CheckArgsConstructor< - !_EnableImplicitReducedArityExtension - && sizeof...(_Up) < sizeof...(_Tp) - && !_PackExpandsToThisTuple<_Up...>::value - >::template __enable_implicit<_Up...>(), - bool - >::type = false - > - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 - explicit - tuple(_Up&&... __u) - _NOEXCEPT_(( - is_nothrow_constructible<_BaseT, - typename __make_tuple_indices::type, - typename __make_tuple_types::type, - typename __make_tuple_indices::type, - typename __make_tuple_types::type, - _Up... - >::value - )) - : __base_(typename __make_tuple_indices::type(), + template , + _EnableUTypesCtor<_Up...>, + _Not<_Lazy<_And, is_convertible<_Up, _Tp>...> > // explicit check + >::value + , int> = 0> + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 + explicit tuple(_Up&&... __u) + _NOEXCEPT_((_And...>::value)) + : __base_(typename __make_tuple_indices::type(), typename __make_tuple_types::type(), typename __make_tuple_indices::type(), typename __make_tuple_types::type(), _VSTD::forward<_Up>(__u)...) {} - template ::value - >::template __enable_implicit<_Up...>(), - bool - >::type = false - > - _LIBCPP_INLINE_VISIBILITY - tuple(allocator_arg_t, const _Alloc& __a, _Up&&... __u) - : __base_(allocator_arg_t(), __a, + template , + _EnableUTypesCtor<_Up...>, + is_convertible<_Up, _Tp>... // explicit check + >::value + , int> = 0> + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 + tuple(allocator_arg_t, const _Alloc& __a, _Up&&... __u) + : __base_(allocator_arg_t(), __a, typename __make_tuple_indices::type(), typename __make_tuple_types::type(), typename __make_tuple_indices::type(), typename __make_tuple_types::type(), _VSTD::forward<_Up>(__u)...) {} - template ::value - >::template __enable_explicit<_Up...>(), - bool - >::type = false - > - _LIBCPP_INLINE_VISIBILITY - explicit - tuple(allocator_arg_t, const _Alloc& __a, _Up&&... __u) - : __base_(allocator_arg_t(), __a, + template , + _EnableUTypesCtor<_Up...>, + _Not<_Lazy<_And, is_convertible<_Up, _Tp>...> > // explicit check + >::value + , int> = 0> + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 + explicit tuple(allocator_arg_t, const _Alloc& __a, _Up&&... __u) + : __base_(allocator_arg_t(), __a, typename __make_tuple_indices::type(), typename __make_tuple_types::type(), typename __make_tuple_indices::type(), typename __make_tuple_types::type(), _VSTD::forward<_Up>(__u)...) {} - template = false> - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 - tuple(_Tuple&& __t) _NOEXCEPT_((is_nothrow_constructible<_BaseT, _Tuple>::value)) - : __base_(_VSTD::forward<_Tuple>(__t)) {} + // Copy and move constructors (including the allocator_arg_t variants) + tuple(const tuple&) = default; + tuple(tuple&&) = default; - template = false> - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 - tuple(const _Tuple& __t) _NOEXCEPT_((is_nothrow_constructible<_BaseT, const _Tuple&>::value)) - : __base_(__t) {} - template = false> - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 - explicit - tuple(_Tuple&& __t) _NOEXCEPT_((is_nothrow_constructible<_BaseT, _Tuple>::value)) - : __base_(_VSTD::forward<_Tuple>(__t)) {} + template class _And = _And, _EnableIf< + _And...>::value + , int> = 0> + tuple(allocator_arg_t, const _Alloc& __alloc, const tuple& __t) + : __base_(allocator_arg_t(), __alloc, __t) + { } + + template class _And = _And, _EnableIf< + _And...>::value + , int> = 0> + tuple(allocator_arg_t, const _Alloc& __alloc, tuple&& __t) + : __base_(allocator_arg_t(), __alloc, _VSTD::move(__t)) + { } + + // tuple(const tuple&) constructors (including allocator_arg_t variants) + template + struct _EnableCopyFromOtherTuple : _And< + _Not, tuple<_Up...> > >, + _Lazy<_Or, + _BoolConstant, + // _Tp and _Up are 1-element packs - the pack expansions look + // weird to avoid tripping up the type traits in degenerate cases + _Lazy<_And, + _Not&, _Tp> >..., + _Not&> >... + > + >, + is_constructible<_Tp, const _Up&>... + > { }; + + template , + _EnableCopyFromOtherTuple<_Up...>, + is_convertible... // explicit check + >::value + , int> = 0> + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 + tuple(const tuple<_Up...>& __t) + _NOEXCEPT_((_And...>::value)) + : __base_(__t) + { } + + template , + _EnableCopyFromOtherTuple<_Up...>, + _Not<_Lazy<_And, is_convertible...> > // explicit check + >::value + , int> = 0> + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 + explicit tuple(const tuple<_Up...>& __t) + _NOEXCEPT_((_And...>::value)) + : __base_(__t) + { } + + template , + _EnableCopyFromOtherTuple<_Up...>, + is_convertible... // explicit check + >::value + , int> = 0> + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 + tuple(allocator_arg_t, const _Alloc& __a, const tuple<_Up...>& __t) + : __base_(allocator_arg_t(), __a, __t) + { } + + template , + _EnableCopyFromOtherTuple<_Up...>, + _Not<_Lazy<_And, is_convertible...> > // explicit check + >::value + , int> = 0> + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 + explicit tuple(allocator_arg_t, const _Alloc& __a, const tuple<_Up...>& __t) + : __base_(allocator_arg_t(), __a, __t) + { } + + // tuple(tuple&&) constructors (including allocator_arg_t variants) + template + struct _EnableMoveFromOtherTuple : _And< + _Not, tuple<_Up...> > >, + _Lazy<_Or, + _BoolConstant, + // _Tp and _Up are 1-element packs - the pack expansions look + // weird to avoid tripping up the type traits in degenerate cases + _Lazy<_And, + _Not, _Tp> >..., + _Not > >... + > + >, + is_constructible<_Tp, _Up>... + > { }; + + template , + _EnableMoveFromOtherTuple<_Up...>, + is_convertible<_Up, _Tp>... // explicit check + >::value + , int> = 0> + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 + tuple(tuple<_Up...>&& __t) + _NOEXCEPT_((_And...>::value)) + : __base_(_VSTD::move(__t)) + { } + + template , + _EnableMoveFromOtherTuple<_Up...>, + _Not<_Lazy<_And, is_convertible<_Up, _Tp>...> > // explicit check + >::value + , int> = 0> + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 + explicit tuple(tuple<_Up...>&& __t) + _NOEXCEPT_((_And...>::value)) + : __base_(_VSTD::move(__t)) + { } + + template , + _EnableMoveFromOtherTuple<_Up...>, + is_convertible<_Up, _Tp>... // explicit check + >::value + , int> = 0> + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 + tuple(allocator_arg_t, const _Alloc& __a, tuple<_Up...>&& __t) + : __base_(allocator_arg_t(), __a, _VSTD::move(__t)) + { } + + template , + _EnableMoveFromOtherTuple<_Up...>, + _Not<_Lazy<_And, is_convertible<_Up, _Tp>...> > // explicit check + >::value + , int> = 0> + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 + explicit tuple(allocator_arg_t, const _Alloc& __a, tuple<_Up...>&& __t) + : __base_(allocator_arg_t(), __a, _VSTD::move(__t)) + { } + + // tuple(const pair&) constructors (including allocator_arg_t variants) + template + struct _EnableImplicitCopyFromPair : _And< + is_constructible<_FirstType<_DependentTp...>, const _Up1&>, + is_constructible<_SecondType<_DependentTp...>, const _Up2&>, + is_convertible >, // explicit check + is_convertible > + > { }; + + template + struct _EnableExplicitCopyFromPair : _And< + is_constructible<_FirstType<_DependentTp...>, const _Up1&>, + is_constructible<_SecondType<_DependentTp...>, const _Up2&>, + _Not > >, // explicit check + _Not > > + > { }; + + template class _And = _And, _EnableIf< + _And< + _BoolConstant, + _EnableImplicitCopyFromPair<_Up1, _Up2, _Tp...> + >::value + , int> = 0> + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 + tuple(const pair<_Up1, _Up2>& __p) + _NOEXCEPT_((_And< + is_nothrow_constructible<_FirstType<_Tp...>, const _Up1&>, + is_nothrow_constructible<_SecondType<_Tp...>, const _Up2&> + >::value)) + : __base_(__p) + { } + + template class _And = _And, _EnableIf< + _And< + _BoolConstant, + _EnableExplicitCopyFromPair<_Up1, _Up2, _Tp...> + >::value + , int> = 0> + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 + explicit tuple(const pair<_Up1, _Up2>& __p) + _NOEXCEPT_((_And< + is_nothrow_constructible<_FirstType<_Tp...>, const _Up1&>, + is_nothrow_constructible<_SecondType<_Tp...>, const _Up2&> + >::value)) + : __base_(__p) + { } + + template class _And = _And, _EnableIf< + _And< + _BoolConstant, + _EnableImplicitCopyFromPair<_Up1, _Up2, _Tp...> + >::value + , int> = 0> + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 + tuple(allocator_arg_t, const _Alloc& __a, const pair<_Up1, _Up2>& __p) + : __base_(allocator_arg_t(), __a, __p) + { } + + template class _And = _And, _EnableIf< + _And< + _BoolConstant, + _EnableExplicitCopyFromPair<_Up1, _Up2, _Tp...> + >::value + , int> = 0> + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 + explicit tuple(allocator_arg_t, const _Alloc& __a, const pair<_Up1, _Up2>& __p) + : __base_(allocator_arg_t(), __a, __p) + { } + + // tuple(pair&&) constructors (including allocator_arg_t variants) + template + struct _EnableImplicitMoveFromPair : _And< + is_constructible<_FirstType<_DependentTp...>, _Up1>, + is_constructible<_SecondType<_DependentTp...>, _Up2>, + is_convertible<_Up1, _FirstType<_DependentTp...> >, // explicit check + is_convertible<_Up2, _SecondType<_DependentTp...> > + > { }; + + template + struct _EnableExplicitMoveFromPair : _And< + is_constructible<_FirstType<_DependentTp...>, _Up1>, + is_constructible<_SecondType<_DependentTp...>, _Up2>, + _Not > >, // explicit check + _Not > > + > { }; + + template class _And = _And, _EnableIf< + _And< + _BoolConstant, + _EnableImplicitMoveFromPair<_Up1, _Up2, _Tp...> + >::value + , int> = 0> + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 + tuple(pair<_Up1, _Up2>&& __p) + _NOEXCEPT_((_And< + is_nothrow_constructible<_FirstType<_Tp...>, _Up1>, + is_nothrow_constructible<_SecondType<_Tp...>, _Up2> + >::value)) + : __base_(_VSTD::move(__p)) + { } + + template class _And = _And, _EnableIf< + _And< + _BoolConstant, + _EnableExplicitMoveFromPair<_Up1, _Up2, _Tp...> + >::value + , int> = 0> + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 + explicit tuple(pair<_Up1, _Up2>&& __p) + _NOEXCEPT_((_And< + is_nothrow_constructible<_FirstType<_Tp...>, _Up1>, + is_nothrow_constructible<_SecondType<_Tp...>, _Up2> + >::value)) + : __base_(_VSTD::move(__p)) + { } + + template class _And = _And, _EnableIf< + _And< + _BoolConstant, + _EnableImplicitMoveFromPair<_Up1, _Up2, _Tp...> + >::value + , int> = 0> + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 + tuple(allocator_arg_t, const _Alloc& __a, pair<_Up1, _Up2>&& __p) + : __base_(allocator_arg_t(), __a, _VSTD::move(__p)) + { } + + template class _And = _And, _EnableIf< + _And< + _BoolConstant, + _EnableExplicitMoveFromPair<_Up1, _Up2, _Tp...> + >::value + , int> = 0> + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 + explicit tuple(allocator_arg_t, const _Alloc& __a, pair<_Up1, _Up2>&& __p) + : __base_(allocator_arg_t(), __a, _VSTD::move(__p)) + { } + + // [tuple.assign] + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 + tuple& operator=(_If<_And...>::value, tuple, __nat> const& __tuple) + _NOEXCEPT_((_And...>::value)) + { + _VSTD::__memberwise_copy_assign(*this, __tuple, + typename __make_tuple_indices::type()); + return *this; + } - template = false> - _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX11 - explicit - tuple(const _Tuple& __t) _NOEXCEPT_((is_nothrow_constructible<_BaseT, const _Tuple&>::value)) - : __base_(__t) {} + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 + tuple& operator=(_If<_And...>::value, tuple, __nat>&& __tuple) + _NOEXCEPT_((_And...>::value)) + { + _VSTD::__memberwise_forward_assign(*this, _VSTD::move(__tuple), + __tuple_types<_Tp...>(), + typename __make_tuple_indices::type()); + return *this; + } - template ::value - >::template __enable_implicit<_Tuple>(), - bool - >::type = false - > - _LIBCPP_INLINE_VISIBILITY - tuple(allocator_arg_t, const _Alloc& __a, _Tuple&& __t) - : __base_(allocator_arg_t(), __a, _VSTD::forward<_Tuple>(__t)) {} + template, + is_assignable<_Tp&, _Up const&>... + >::value + ,int> = 0> + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 + tuple& operator=(tuple<_Up...> const& __tuple) + _NOEXCEPT_((_And...>::value)) + { + _VSTD::__memberwise_copy_assign(*this, __tuple, + typename __make_tuple_indices::type()); + return *this; + } - template ::value - >::template __enable_explicit<_Tuple>(), - bool - >::type = false - > - _LIBCPP_INLINE_VISIBILITY - explicit - tuple(allocator_arg_t, const _Alloc& __a, _Tuple&& __t) - : __base_(allocator_arg_t(), __a, _VSTD::forward<_Tuple>(__t)) {} + template, + is_assignable<_Tp&, _Up>... + >::value + ,int> = 0> + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 + tuple& operator=(tuple<_Up...>&& __tuple) + _NOEXCEPT_((_And...>::value)) + { + _VSTD::__memberwise_forward_assign(*this, _VSTD::move(__tuple), + __tuple_types<_Up...>(), + typename __make_tuple_indices::type()); + return *this; + } - using _CanCopyAssign = __all::value...>; - using _CanMoveAssign = __all::value...>; + template, + is_assignable<_FirstType<_Tp..., _Dep>&, _Up1 const&>, + is_assignable<_SecondType<_Tp..., _Dep>&, _Up2 const&> + >::value + ,int> = 0> + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 + tuple& operator=(pair<_Up1, _Up2> const& __pair) + _NOEXCEPT_((_And< + is_nothrow_assignable<_FirstType<_Tp...>&, _Up1 const&>, + is_nothrow_assignable<_SecondType<_Tp...>&, _Up2 const&> + >::value)) + { + _VSTD::get<0>(*this) = __pair.first; + _VSTD::get<1>(*this) = __pair.second; + return *this; + } - _LIBCPP_INLINE_VISIBILITY - tuple& operator=(typename conditional<_CanCopyAssign::value, tuple, __nat>::type const& __t) - _NOEXCEPT_((__all::value...>::value)) + template, + is_assignable<_FirstType<_Tp..., _Dep>&, _Up1>, + is_assignable<_SecondType<_Tp..., _Dep>&, _Up2> + >::value + ,int> = 0> + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 + tuple& operator=(pair<_Up1, _Up2>&& __pair) + _NOEXCEPT_((_And< + is_nothrow_assignable<_FirstType<_Tp...>&, _Up1>, + is_nothrow_assignable<_SecondType<_Tp...>&, _Up2> + >::value)) { - __base_.operator=(__t.__base_); + _VSTD::get<0>(*this) = _VSTD::forward<_Up1>(__pair.first); + _VSTD::get<1>(*this) = _VSTD::forward<_Up2>(__pair.second); return *this; } - _LIBCPP_INLINE_VISIBILITY - tuple& operator=(typename conditional<_CanMoveAssign::value, tuple, __nat>::type&& __t) - _NOEXCEPT_((__all::value...>::value)) + // EXTENSION + template, + is_assignable<_Tp&, _Up const&>... + >::value + > > + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 + tuple& operator=(array<_Up, _Np> const& __array) + _NOEXCEPT_((_And...>::value)) { - __base_.operator=(static_cast<_BaseT&&>(__t.__base_)); + _VSTD::__memberwise_copy_assign(*this, __array, + typename __make_tuple_indices::type()); return *this; } - template ::value - >::type - > - _LIBCPP_INLINE_VISIBILITY - tuple& - operator=(_Tuple&& __t) _NOEXCEPT_((is_nothrow_assignable<_BaseT&, _Tuple>::value)) - { - __base_.operator=(_VSTD::forward<_Tuple>(__t)); - return *this; - } + // EXTENSION + template, + is_assignable<_Tp&, _Up>... + >::value + > > + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 + tuple& operator=(array<_Up, _Np>&& __array) + _NOEXCEPT_((_And...>::value)) + { + _VSTD::__memberwise_forward_assign(*this, _VSTD::move(__array), + __tuple_types<_If...>(), + typename __make_tuple_indices::type()); + return *this; + } - _LIBCPP_INLINE_VISIBILITY + // [tuple.swap] + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 void swap(tuple& __t) _NOEXCEPT_(__all<__is_nothrow_swappable<_Tp>::value...>::value) {__base_.swap(__t.__base_);} }; @@ -958,21 +1079,21 @@ template <> class _LIBCPP_TEMPLATE_VIS tuple<> { public: - _LIBCPP_INLINE_VISIBILITY - _LIBCPP_CONSTEXPR tuple() _NOEXCEPT = default; + _LIBCPP_INLINE_VISIBILITY constexpr + tuple() _NOEXCEPT = default; template - _LIBCPP_INLINE_VISIBILITY + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 tuple(allocator_arg_t, const _Alloc&) _NOEXCEPT {} template - _LIBCPP_INLINE_VISIBILITY + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 tuple(allocator_arg_t, const _Alloc&, const tuple&) _NOEXCEPT {} template - _LIBCPP_INLINE_VISIBILITY + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 tuple(array<_Up, 0>) _NOEXCEPT {} template - _LIBCPP_INLINE_VISIBILITY + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 tuple(allocator_arg_t, const _Alloc&, array<_Up, 0>) _NOEXCEPT {} - _LIBCPP_INLINE_VISIBILITY + _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 void swap(tuple&) _NOEXCEPT {} }; @@ -990,7 +1111,7 @@ tuple(allocator_arg_t, _Alloc, tuple<_Tp...>) -> tuple<_Tp...>; #endif template -inline _LIBCPP_INLINE_VISIBILITY +inline _LIBCPP_INLINE_VISIBILITY _LIBCPP_CONSTEXPR_AFTER_CXX17 typename enable_if < __all<__is_swappable<_Tp>::value...>::value, @@ -1044,7 +1165,7 @@ get(const tuple<_Tp...>&& __t) _NOEXCEPT namespace __find_detail { -static constexpr size_t __not_found = -1; +static constexpr size_t __not_found = static_cast(-1); static constexpr size_t __ambiguous = __not_found - 1; inline _LIBCPP_INLINE_VISIBILITY @@ -1450,4 +1571,4 @@ _LIBCPP_NOEXCEPT_RETURN( _LIBCPP_END_NAMESPACE_STD -#endif // _LIBCPP_TUPLE +#endif // _LIBCPP_TUPLE diff --git a/lib/libcxx/include/type_traits b/lib/libcxx/include/type_traits index 59dfd1e9b..b15c7a2a5 100644 --- a/lib/libcxx/include/type_traits +++ b/lib/libcxx/include/type_traits @@ -99,7 +99,7 @@ namespace std template struct is_trivial; template struct is_trivially_copyable; template struct is_standard_layout; - template struct is_literal_type; + template struct is_literal_type; // Deprecated in C++17; removed in C++20 template struct is_empty; template struct is_polymorphic; template struct is_abstract; @@ -165,8 +165,8 @@ namespace std template struct decay; template struct common_type; template struct underlying_type; - template class result_of; // undefined - template class result_of; + template class result_of; // undefined; deprecated in C++17; removed in C++20 + template class result_of; // deprecated in C++17; removed in C++20 template struct invoke_result; // C++17 // const-volatile modifications: @@ -216,9 +216,9 @@ namespace std using add_pointer_t = typename add_pointer::type; // C++14 // other transformations: - template + template using aligned_storage_t = typename aligned_storage::type; // C++14 - template + template using aligned_union_t = typename aligned_union::type; // C++14 template using remove_cvref_t = typename remove_cvref::type; // C++20 @@ -233,7 +233,7 @@ namespace std template using underlying_type_t = typename underlying_type::type; // C++14 template - using result_of_t = typename result_of::type; // C++14 + using result_of_t = typename result_of::type; // C++14; deprecated in C++17; removed in C++20 template using invoke_result_t = typename invoke_result::type; // C++17 @@ -302,7 +302,7 @@ namespace std template inline constexpr bool is_pod_v = is_pod::value; // C++17 template inline constexpr bool is_literal_type_v - = is_literal_type::value; // C++17 + = is_literal_type::value; // C++17; deprecated in C++17; removed in C++20 template inline constexpr bool is_empty_v = is_empty::value; // C++17 template inline constexpr bool is_polymorphic_v @@ -476,8 +476,6 @@ struct _MetaBase { using _EnableIfImpl _LIBCPP_NODEBUG_TYPE = _Tp; template using _OrImpl _LIBCPP_NODEBUG_TYPE = typename _MetaBase<_First::value != true && sizeof...(_Rest) != 0>::template _OrImpl<_First, _Rest...>; - template - using _AndImpl _LIBCPP_NODEBUG_TYPE = typename _MetaBase<_First::value == true && sizeof...(_Rest) != 0>::template _AndImpl<_First, _Rest...>; }; template <> @@ -488,8 +486,6 @@ struct _MetaBase { using _SelectApplyImpl _LIBCPP_NODEBUG_TYPE = _SecondFn<_Args...>; template using _OrImpl _LIBCPP_NODEBUG_TYPE = _Result; - template - using _AndImpl _LIBCPP_NODEBUG_TYPE = _Result; }; template using _EnableIf _LIBCPP_NODEBUG_TYPE = typename _MetaBase<_Cond>::template _EnableIfImpl<_Ret>; @@ -497,8 +493,6 @@ template using _If _LIBCPP_NODEBUG_TYPE = typename _MetaBase<_Cond>::template _SelectImpl<_IfRes, _ElseRes>; template using _Or _LIBCPP_NODEBUG_TYPE = typename _MetaBase< sizeof...(_Rest) != 0 >::template _OrImpl; -template -using _And _LIBCPP_NODEBUG_TYPE = typename _MetaBase< sizeof...(_Rest) != 0 >::template _AndImpl; template struct _Not : _BoolConstant {}; template @@ -506,6 +500,14 @@ using _FirstType _LIBCPP_NODEBUG_TYPE = typename _MetaBase<(sizeof...(_Args) >= template using _SecondType _LIBCPP_NODEBUG_TYPE = typename _MetaBase<(sizeof...(_Args) >= 2)>::template _SecondImpl<_Args...>; +template using __expand_to_true = true_type; +template +__expand_to_true<_EnableIf<_Pred::value>...> __and_helper(int); +template +false_type __and_helper(...); +template +using _And _LIBCPP_NODEBUG_TYPE = decltype(__and_helper<_Pred...>(0)); + template